aboutsummaryrefslogtreecommitdiffstats
path: root/drivers/net
diff options
context:
space:
mode:
authorLinus Torvalds <torvalds@linux-foundation.org>2013-11-13 03:40:34 -0500
committerLinus Torvalds <torvalds@linux-foundation.org>2013-11-13 03:40:34 -0500
commit42a2d923cc349583ebf6fdd52a7d35e1c2f7e6bd (patch)
tree2b2b0c03b5389c1301800119333967efafd994ca /drivers/net
parent5cbb3d216e2041700231bcfc383ee5f8b7fc8b74 (diff)
parent75ecab1df14d90e86cebef9ec5c76befde46e65f (diff)
Merge git://git.kernel.org/pub/scm/linux/kernel/git/davem/net-next
Pull networking updates from David Miller: 1) The addition of nftables. No longer will we need protocol aware firewall filtering modules, it can all live in userspace. At the core of nftables is a, for lack of a better term, virtual machine that executes byte codes to inspect packet or metadata (arriving interface index, etc.) and make verdict decisions. Besides support for loading packet contents and comparing them, the interpreter supports lookups in various datastructures as fundamental operations. For example sets are supports, and therefore one could create a set of whitelist IP address entries which have ACCEPT verdicts attached to them, and use the appropriate byte codes to do such lookups. Since the interpreted code is composed in userspace, userspace can do things like optimize things before giving it to the kernel. Another major improvement is the capability of atomically updating portions of the ruleset. In the existing netfilter implementation, one has to update the entire rule set in order to make a change and this is very expensive. Userspace tools exist to create nftables rules using existing netfilter rule sets, but both kernel implementations will need to co-exist for quite some time as we transition from the old to the new stuff. Kudos to Patrick McHardy, Pablo Neira Ayuso, and others who have worked so hard on this. 2) Daniel Borkmann and Hannes Frederic Sowa made several improvements to our pseudo-random number generator, mostly used for things like UDP port randomization and netfitler, amongst other things. In particular the taus88 generater is updated to taus113, and test cases are added. 3) Support 64-bit rates in HTB and TBF schedulers, from Eric Dumazet and Yang Yingliang. 4) Add support for new 577xx tigon3 chips to tg3 driver, from Nithin Sujir. 5) Fix two fatal flaws in TCP dynamic right sizing, from Eric Dumazet, Neal Cardwell, and Yuchung Cheng. 6) Allow IP_TOS and IP_TTL to be specified in sendmsg() ancillary control message data, much like other socket option attributes. From Francesco Fusco. 7) Allow applications to specify a cap on the rate computed automatically by the kernel for pacing flows, via a new SO_MAX_PACING_RATE socket option. From Eric Dumazet. 8) Make the initial autotuned send buffer sizing in TCP more closely reflect actual needs, from Eric Dumazet. 9) Currently early socket demux only happens for TCP sockets, but we can do it for connected UDP sockets too. Implementation from Shawn Bohrer. 10) Refactor inet socket demux with the goal of improving hash demux performance for listening sockets. With the main goals being able to use RCU lookups on even request sockets, and eliminating the listening lock contention. From Eric Dumazet. 11) The bonding layer has many demuxes in it's fast path, and an RCU conversion was started back in 3.11, several changes here extend the RCU usage to even more locations. From Ding Tianhong and Wang Yufen, based upon suggestions by Nikolay Aleksandrov and Veaceslav Falico. 12) Allow stackability of segmentation offloads to, in particular, allow segmentation offloading over tunnels. From Eric Dumazet. 13) Significantly improve the handling of secret keys we input into the various hash functions in the inet hashtables, TCP fast open, as well as syncookies. From Hannes Frederic Sowa. The key fundamental operation is "net_get_random_once()" which uses static keys. Hannes even extended this to ipv4/ipv6 fragmentation handling and our generic flow dissector. 14) The generic driver layer takes care now to set the driver data to NULL on device removal, so it's no longer necessary for drivers to explicitly set it to NULL any more. Many drivers have been cleaned up in this way, from Jingoo Han. 15) Add a BPF based packet scheduler classifier, from Daniel Borkmann. 16) Improve CRC32 interfaces and generic SKB checksum iterators so that SCTP's checksumming can more cleanly be handled. Also from Daniel Borkmann. 17) Add a new PMTU discovery mode, IP_PMTUDISC_INTERFACE, which forces using the interface MTU value. This helps avoid PMTU attacks, particularly on DNS servers. From Hannes Frederic Sowa. 18) Use generic XPS for transmit queue steering rather than internal (re-)implementation in virtio-net. From Jason Wang. * git://git.kernel.org/pub/scm/linux/kernel/git/davem/net-next: (1622 commits) random32: add test cases for taus113 implementation random32: upgrade taus88 generator to taus113 from errata paper random32: move rnd_state to linux/random.h random32: add prandom_reseed_late() and call when nonblocking pool becomes initialized random32: add periodic reseeding random32: fix off-by-one in seeding requirement PHY: Add RTL8201CP phy_driver to realtek xtsonic: add missing platform_set_drvdata() in xtsonic_probe() macmace: add missing platform_set_drvdata() in mace_probe() ethernet/arc/arc_emac: add missing platform_set_drvdata() in arc_emac_probe() ipv6: protect for_each_sk_fl_rcu in mem_check with rcu_read_lock_bh vlan: Implement vlan_dev_get_egress_qos_mask as an inline. ixgbe: add warning when max_vfs is out of range. igb: Update link modes display in ethtool netfilter: push reasm skb through instead of original frag skbs ip6_output: fragment outgoing reassembled skb properly MAINTAINERS: mv643xx_eth: take over maintainership from Lennart net_sched: tbf: support of 64bit rates ixgbe: deleting dfwd stations out of order can cause null ptr deref ixgbe: fix build err, num_rx_queues is only available with CONFIG_RPS ...
Diffstat (limited to 'drivers/net')
-rw-r--r--drivers/net/bonding/Makefile2
-rw-r--r--drivers/net/bonding/bond_3ad.c199
-rw-r--r--drivers/net/bonding/bond_alb.c150
-rw-r--r--drivers/net/bonding/bond_alb.h4
-rw-r--r--drivers/net/bonding/bond_main.c715
-rw-r--r--drivers/net/bonding/bond_netlink.c131
-rw-r--r--drivers/net/bonding/bond_options.c142
-rw-r--r--drivers/net/bonding/bond_procfs.c21
-rw-r--r--drivers/net/bonding/bond_sysfs.c264
-rw-r--r--drivers/net/bonding/bonding.h113
-rw-r--r--drivers/net/can/at91_can.c2
-rw-r--r--drivers/net/can/bfin_can.c2
-rw-r--r--drivers/net/can/c_can/c_can_pci.c2
-rw-r--r--drivers/net/can/c_can/c_can_platform.c2
-rw-r--r--drivers/net/can/cc770/cc770_platform.c4
-rw-r--r--drivers/net/can/dev.c67
-rw-r--r--drivers/net/can/flexcan.c2
-rw-r--r--drivers/net/can/janz-ican3.c2
-rw-r--r--drivers/net/can/mcp251x.c2
-rw-r--r--drivers/net/can/mscan/mscan.h6
-rw-r--r--drivers/net/can/pch_can.c1
-rw-r--r--drivers/net/can/sja1000/ems_pci.c1
-rw-r--r--drivers/net/can/sja1000/kvaser_pci.c1
-rw-r--r--drivers/net/can/sja1000/peak_pci.c2
-rw-r--r--drivers/net/can/sja1000/plx_pci.c1
-rw-r--r--drivers/net/can/sja1000/sja1000_platform.c2
-rw-r--r--drivers/net/can/softing/softing.h24
-rw-r--r--drivers/net/can/softing/softing_main.c2
-rw-r--r--drivers/net/can/ti_hecc.c12
-rw-r--r--drivers/net/ethernet/3com/Kconfig4
-rw-r--r--drivers/net/ethernet/3com/typhoon.c1
-rw-r--r--drivers/net/ethernet/8390/8390.h40
-rw-r--r--drivers/net/ethernet/8390/ax88796.c2
-rw-r--r--drivers/net/ethernet/8390/ne2k-pci.c3
-rw-r--r--drivers/net/ethernet/adaptec/starfire.c2
-rw-r--r--drivers/net/ethernet/adi/bfin_mac.h2
-rw-r--r--drivers/net/ethernet/amd/7990.h12
-rw-r--r--drivers/net/ethernet/amd/amd8111e.c2
-rw-r--r--drivers/net/ethernet/amd/atarilance.c4
-rw-r--r--drivers/net/ethernet/amd/au1000_eth.c2
-rw-r--r--drivers/net/ethernet/amd/declance.c16
-rw-r--r--drivers/net/ethernet/amd/lance.c2
-rw-r--r--drivers/net/ethernet/amd/pcnet32.c3
-rw-r--r--drivers/net/ethernet/apple/bmac.c4
-rw-r--r--drivers/net/ethernet/apple/macmace.c1
-rw-r--r--drivers/net/ethernet/arc/emac_main.c7
-rw-r--r--drivers/net/ethernet/atheros/alx/main.c1
-rw-r--r--drivers/net/ethernet/atheros/atl1c/atl1c.h6
-rw-r--r--drivers/net/ethernet/atheros/atl1c/atl1c_hw.c2
-rw-r--r--drivers/net/ethernet/atheros/atl1e/atl1e.h12
-rw-r--r--drivers/net/ethernet/atheros/atl1e/atl1e_main.c46
-rw-r--r--drivers/net/ethernet/atheros/atlx/atl2.h2
-rw-r--r--drivers/net/ethernet/broadcom/b44.c12
-rw-r--r--drivers/net/ethernet/broadcom/bgmac.c117
-rw-r--r--drivers/net/ethernet/broadcom/bnx2.c9
-rw-r--r--drivers/net/ethernet/broadcom/bnx2x/bnx2x.h11
-rw-r--r--drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c18
-rw-r--r--drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c22
-rw-r--r--drivers/net/ethernet/broadcom/bnx2x/bnx2x_hsi.h1
-rw-r--r--drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c25
-rw-r--r--drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c80
-rw-r--r--drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c30
-rw-r--r--drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h2
-rw-r--r--drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c149
-rw-r--r--drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.h7
-rw-r--r--drivers/net/ethernet/broadcom/cnic.c4
-rw-r--r--drivers/net/ethernet/broadcom/cnic_if.h4
-rw-r--r--drivers/net/ethernet/broadcom/tg3.c154
-rw-r--r--drivers/net/ethernet/broadcom/tg3.h3
-rw-r--r--drivers/net/ethernet/brocade/bna/bnad.c1
-rw-r--r--drivers/net/ethernet/brocade/bna/bnad.h43
-rw-r--r--drivers/net/ethernet/calxeda/xgmac.c10
-rw-r--r--drivers/net/ethernet/chelsio/cxgb/common.h46
-rw-r--r--drivers/net/ethernet/chelsio/cxgb/cxgb2.c2
-rw-r--r--drivers/net/ethernet/chelsio/cxgb/pm3393.c4
-rw-r--r--drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c2
-rw-r--r--drivers/net/ethernet/chelsio/cxgb3/regs.h35
-rw-r--r--drivers/net/ethernet/chelsio/cxgb4/cxgb4.h1
-rw-r--r--drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c5
-rw-r--r--drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c5
-rw-r--r--drivers/net/ethernet/chelsio/cxgb4vf/sge.c11
-rw-r--r--drivers/net/ethernet/cisco/enic/enic_main.c2
-rw-r--r--drivers/net/ethernet/davicom/dm9000.c2
-rw-r--r--drivers/net/ethernet/dec/tulip/de2104x.c1
-rw-r--r--drivers/net/ethernet/dec/tulip/de4x5.c2
-rw-r--r--drivers/net/ethernet/dec/tulip/dmfe.c3
-rw-r--r--drivers/net/ethernet/dec/tulip/tulip_core.c1
-rw-r--r--drivers/net/ethernet/dec/tulip/uli526x.c2
-rw-r--r--drivers/net/ethernet/dec/tulip/winbond-840.c3
-rw-r--r--drivers/net/ethernet/dec/tulip/xircom_cb.c2
-rw-r--r--drivers/net/ethernet/dlink/dl2k.c1
-rw-r--r--drivers/net/ethernet/dlink/sundance.c2
-rw-r--r--drivers/net/ethernet/emulex/benet/be.h177
-rw-r--r--drivers/net/ethernet/emulex/benet/be_cmds.c27
-rw-r--r--drivers/net/ethernet/emulex/benet/be_cmds.h332
-rw-r--r--drivers/net/ethernet/emulex/benet/be_ethtool.c40
-rw-r--r--drivers/net/ethernet/emulex/benet/be_main.c315
-rw-r--r--drivers/net/ethernet/fealnx.c1
-rw-r--r--drivers/net/ethernet/freescale/fs_enet/fs_enet-main.c2
-rw-r--r--drivers/net/ethernet/freescale/gianfar.c87
-rw-r--r--drivers/net/ethernet/freescale/gianfar.h26
-rw-r--r--drivers/net/ethernet/freescale/ucc_geth.c2
-rw-r--r--drivers/net/ethernet/fujitsu/Kconfig2
-rw-r--r--drivers/net/ethernet/hp/hp100.c2
-rw-r--r--drivers/net/ethernet/i825xx/82596.c8
-rw-r--r--drivers/net/ethernet/i825xx/lib82596.c6
-rw-r--r--drivers/net/ethernet/ibm/emac/core.c2
-rw-r--r--drivers/net/ethernet/ibm/emac/debug.h14
-rw-r--r--drivers/net/ethernet/ibm/emac/rgmii.h18
-rw-r--r--drivers/net/ethernet/ibm/emac/tah.h14
-rw-r--r--drivers/net/ethernet/ibm/emac/zmii.h18
-rw-r--r--drivers/net/ethernet/ibm/ibmveth.c4
-rw-r--r--drivers/net/ethernet/icplus/ipg.c1
-rw-r--r--drivers/net/ethernet/intel/e100.c2
-rw-r--r--drivers/net/ethernet/intel/e1000/e1000.h32
-rw-r--r--drivers/net/ethernet/intel/e1000/e1000_main.c3
-rw-r--r--drivers/net/ethernet/intel/e1000e/e1000.h45
-rw-r--r--drivers/net/ethernet/intel/i40e/i40e.h13
-rw-r--r--drivers/net/ethernet/intel/i40e/i40e_debugfs.c338
-rw-r--r--drivers/net/ethernet/intel/i40e/i40e_ethtool.c69
-rw-r--r--drivers/net/ethernet/intel/i40e/i40e_main.c484
-rw-r--r--drivers/net/ethernet/intel/i40e/i40e_txrx.c433
-rw-r--r--drivers/net/ethernet/intel/i40e/i40e_txrx.h35
-rw-r--r--drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c4
-rw-r--r--drivers/net/ethernet/intel/igb/e1000_82575.h16
-rw-r--r--drivers/net/ethernet/intel/igb/e1000_hw.h6
-rw-r--r--drivers/net/ethernet/intel/igb/e1000_i210.h38
-rw-r--r--drivers/net/ethernet/intel/igb/e1000_mac.h2
-rw-r--r--drivers/net/ethernet/intel/igb/e1000_phy.c8
-rw-r--r--drivers/net/ethernet/intel/igb/igb.h74
-rw-r--r--drivers/net/ethernet/intel/igb/igb_ethtool.c145
-rw-r--r--drivers/net/ethernet/intel/igb/igb_main.c61
-rw-r--r--drivers/net/ethernet/intel/igbvf/igbvf.h22
-rw-r--r--drivers/net/ethernet/intel/igbvf/netdev.c8
-rw-r--r--drivers/net/ethernet/intel/igbvf/vf.c4
-rw-r--r--drivers/net/ethernet/intel/ixgb/ixgb.h22
-rw-r--r--drivers/net/ethernet/intel/ixgb/ixgb_hw.h25
-rw-r--r--drivers/net/ethernet/intel/ixgbe/ixgbe.h258
-rw-r--r--drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c109
-rw-r--r--drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c15
-rw-r--r--drivers/net/ethernet/intel/ixgbe/ixgbe_main.c589
-rw-r--r--drivers/net/ethernet/intel/ixgbe/ixgbe_phy.h40
-rw-r--r--drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c30
-rw-r--r--drivers/net/ethernet/intel/ixgbe/ixgbe_type.h5
-rw-r--r--drivers/net/ethernet/intel/ixgbe/ixgbe_x540.c12
-rw-r--r--drivers/net/ethernet/intel/ixgbevf/ethtool.c234
-rw-r--r--drivers/net/ethernet/intel/ixgbevf/ixgbevf.h178
-rw-r--r--drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c255
-rw-r--r--drivers/net/ethernet/intel/ixgbevf/vf.c4
-rw-r--r--drivers/net/ethernet/jme.c6
-rw-r--r--drivers/net/ethernet/jme.h1
-rw-r--r--drivers/net/ethernet/korina.c2
-rw-r--r--drivers/net/ethernet/marvell/mv643xx_eth.c5
-rw-r--r--drivers/net/ethernet/marvell/mvmdio.c110
-rw-r--r--drivers/net/ethernet/marvell/mvneta.c2
-rw-r--r--drivers/net/ethernet/marvell/skge.c2
-rw-r--r--drivers/net/ethernet/marvell/sky2.c3
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/cmd.c41
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/cq.c6
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/en_cq.c45
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/en_ethtool.c36
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/en_main.c8
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/en_netdev.c152
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/en_port.c17
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/en_rx.c68
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/en_selftest.c2
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/en_tx.c54
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/eq.c1
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/fw.c110
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/icm.c42
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/main.c62
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/mcg.c8
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/mlx4.h21
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/mlx4_en.h33
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/mr.c11
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/pd.c11
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/port.c98
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/qp.c3
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/resource_tracker.c589
-rw-r--r--drivers/net/ethernet/mellanox/mlx4/srq.c3
-rw-r--r--drivers/net/ethernet/micrel/ks8851_mll.c4
-rw-r--r--drivers/net/ethernet/micrel/ksz884x.c4
-rw-r--r--drivers/net/ethernet/moxa/moxart_ether.c1
-rw-r--r--drivers/net/ethernet/myricom/myri10ge/myri10ge.c9
-rw-r--r--drivers/net/ethernet/natsemi/natsemi.c2
-rw-r--r--drivers/net/ethernet/natsemi/xtsonic.c1
-rw-r--r--drivers/net/ethernet/neterion/s2io.c2
-rw-r--r--drivers/net/ethernet/neterion/vxge/vxge-main.c2
-rw-r--r--drivers/net/ethernet/octeon/octeon_mgmt.c2
-rw-r--r--drivers/net/ethernet/oki-semi/pch_gbe/pch_gbe.h56
-rw-r--r--drivers/net/ethernet/packetengines/hamachi.c1
-rw-r--r--drivers/net/ethernet/packetengines/yellowfin.c2
-rw-r--r--drivers/net/ethernet/pasemi/pasemi_mac.c1
-rw-r--r--drivers/net/ethernet/qlogic/netxen/netxen_nic.h9
-rw-r--r--drivers/net/ethernet/qlogic/netxen/netxen_nic_hdr.h1
-rw-r--r--drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c2
-rw-r--r--drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c30
-rw-r--r--drivers/net/ethernet/qlogic/qla3xxx.c2
-rw-r--r--drivers/net/ethernet/qlogic/qlcnic/qlcnic.h181
-rw-r--r--drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c145
-rw-r--r--drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.h18
-rw-r--r--drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c78
-rw-r--r--drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_vnic.c47
-rw-r--r--drivers/net/ethernet/qlogic/qlcnic/qlcnic_ctx.c20
-rw-r--r--drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c184
-rw-r--r--drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.h109
-rw-r--r--drivers/net/ethernet/qlogic/qlcnic/qlcnic_ethtool.c260
-rw-r--r--drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c2
-rw-r--r--drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.h8
-rw-r--r--drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c2
-rw-r--r--drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c67
-rw-r--r--drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c434
-rw-r--r--drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c41
-rw-r--r--drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c15
-rw-r--r--drivers/net/ethernet/qlogic/qlcnic/qlcnic_sysfs.c12
-rw-r--r--drivers/net/ethernet/qlogic/qlge/qlge.h60
-rw-r--r--drivers/net/ethernet/qlogic/qlge/qlge_main.c127
-rw-r--r--drivers/net/ethernet/rdc/r6040.c2
-rw-r--r--drivers/net/ethernet/realtek/8139cp.c1
-rw-r--r--drivers/net/ethernet/realtek/8139too.c1
-rw-r--r--drivers/net/ethernet/realtek/r8169.c1
-rw-r--r--drivers/net/ethernet/renesas/sh_eth.c14
-rw-r--r--drivers/net/ethernet/renesas/sh_eth.h3
-rw-r--r--drivers/net/ethernet/sfc/ef10.c319
-rw-r--r--drivers/net/ethernet/sfc/ef10_regs.h1
-rw-r--r--drivers/net/ethernet/sfc/efx.h105
-rw-r--r--drivers/net/ethernet/sfc/ethtool.c5
-rw-r--r--drivers/net/ethernet/sfc/io.h5
-rw-r--r--drivers/net/ethernet/sfc/mcdi.h120
-rw-r--r--drivers/net/ethernet/sfc/mdio_10g.h26
-rw-r--r--drivers/net/ethernet/sfc/net_driver.h10
-rw-r--r--drivers/net/ethernet/sfc/nic.c73
-rw-r--r--drivers/net/ethernet/sfc/nic.h256
-rw-r--r--drivers/net/ethernet/sfc/phy.h8
-rw-r--r--drivers/net/ethernet/sfc/rx.c90
-rw-r--r--drivers/net/ethernet/sfc/selftest.h15
-rw-r--r--drivers/net/ethernet/sfc/tx.c426
-rw-r--r--drivers/net/ethernet/sgi/meth.c2
-rw-r--r--drivers/net/ethernet/sis/sis190.c1
-rw-r--r--drivers/net/ethernet/smsc/Kconfig4
-rw-r--r--drivers/net/ethernet/smsc/epic100.c126
-rw-r--r--drivers/net/ethernet/smsc/smc911x.c331
-rw-r--r--drivers/net/ethernet/smsc/smc911x.h2
-rw-r--r--drivers/net/ethernet/smsc/smc9194.c56
-rw-r--r--drivers/net/ethernet/smsc/smc91c92_cs.c43
-rw-r--r--drivers/net/ethernet/smsc/smc91x.c237
-rw-r--r--drivers/net/ethernet/smsc/smc91x.h4
-rw-r--r--drivers/net/ethernet/smsc/smsc911x.c4
-rw-r--r--drivers/net/ethernet/smsc/smsc9420.c169
-rw-r--r--drivers/net/ethernet/stmicro/stmmac/common.h12
-rw-r--r--drivers/net/ethernet/stmicro/stmmac/dwmac_dma.h17
-rw-r--r--drivers/net/ethernet/stmicro/stmmac/mmc.h6
-rw-r--r--drivers/net/ethernet/stmicro/stmmac/stmmac.h10
-rw-r--r--drivers/net/ethernet/stmicro/stmmac/stmmac_pci.c1
-rw-r--r--drivers/net/ethernet/sun/cassini.c4
-rw-r--r--drivers/net/ethernet/sun/niu.c2
-rw-r--r--drivers/net/ethernet/sun/sungem.c4
-rw-r--r--drivers/net/ethernet/sun/sunhme.c12
-rw-r--r--drivers/net/ethernet/sun/sunqe.c2
-rw-r--r--drivers/net/ethernet/tehuti/tehuti.c1
-rw-r--r--drivers/net/ethernet/ti/Kconfig8
-rw-r--r--drivers/net/ethernet/ti/Makefile1
-rw-r--r--drivers/net/ethernet/ti/cpsw-phy-sel.c161
-rw-r--r--drivers/net/ethernet/ti/cpsw.c159
-rw-r--r--drivers/net/ethernet/ti/cpsw.h2
-rw-r--r--drivers/net/ethernet/ti/cpts.h9
-rw-r--r--drivers/net/ethernet/ti/davinci_emac.c2
-rw-r--r--drivers/net/ethernet/ti/tlan.c1
-rw-r--r--drivers/net/ethernet/tile/tilegx.c2
-rw-r--r--drivers/net/ethernet/toshiba/ps3_gelic_net.h29
-rw-r--r--drivers/net/ethernet/toshiba/ps3_gelic_wireless.h6
-rw-r--r--drivers/net/ethernet/toshiba/spider_net.c1
-rw-r--r--drivers/net/ethernet/toshiba/spider_net.h4
-rw-r--r--drivers/net/ethernet/toshiba/tc35815.c1
-rw-r--r--drivers/net/ethernet/via/via-rhine.c1
-rw-r--r--drivers/net/ethernet/xilinx/xilinx_emaclite.c41
-rw-r--r--drivers/net/fddi/skfp/fplustm.c2
-rw-r--r--drivers/net/fddi/skfp/h/smc.h28
-rw-r--r--drivers/net/fddi/skfp/skfddi.c6
-rw-r--r--drivers/net/hamradio/baycom_ser_fdx.c2
-rw-r--r--drivers/net/hamradio/baycom_ser_hdx.c2
-rw-r--r--drivers/net/hamradio/scc.c2
-rw-r--r--drivers/net/hamradio/yam.c2
-rw-r--r--drivers/net/irda/bfin_sir.c4
-rw-r--r--drivers/net/irda/donauboe.c4
-rw-r--r--drivers/net/irda/sh_irda.c2
-rw-r--r--drivers/net/irda/sh_sir.c2
-rw-r--r--drivers/net/irda/sir-dev.h29
-rw-r--r--drivers/net/macvlan.c47
-rw-r--r--drivers/net/netconsole.c57
-rw-r--r--drivers/net/phy/Kconfig7
-rw-r--r--drivers/net/phy/Makefile1
-rw-r--r--drivers/net/phy/at803x.c57
-rw-r--r--drivers/net/phy/marvell.c4
-rw-r--r--drivers/net/phy/mdio-moxart.c201
-rw-r--r--drivers/net/phy/micrel.c24
-rw-r--r--drivers/net/phy/realtek.c15
-rw-r--r--drivers/net/plip/plip.c2
-rw-r--r--drivers/net/usb/Kconfig15
-rw-r--r--drivers/net/usb/Makefile1
-rw-r--r--drivers/net/usb/catc.c8
-rw-r--r--drivers/net/usb/cdc-phonet.c2
-rw-r--r--drivers/net/usb/cdc_mbim.c104
-rw-r--r--drivers/net/usb/cdc_ncm.c507
-rw-r--r--drivers/net/usb/huawei_cdc_ncm.c230
-rw-r--r--drivers/net/usb/qmi_wwan.c69
-rw-r--r--drivers/net/veth.c9
-rw-r--r--drivers/net/virtio_net.c219
-rw-r--r--drivers/net/vmxnet3/vmxnet3_int.h4
-rw-r--r--drivers/net/vxlan.c37
-rw-r--r--drivers/net/wan/hostess_sv11.c2
-rw-r--r--drivers/net/wan/sealevel.c2
-rw-r--r--drivers/net/wan/x25_asy.h2
-rw-r--r--drivers/net/wan/z85230.h27
-rw-r--r--drivers/net/wimax/i2400m/i2400m-usb.h27
-rw-r--r--drivers/net/wimax/i2400m/i2400m.h117
-rw-r--r--drivers/net/wireless/adm8211.c1
-rw-r--r--drivers/net/wireless/airo.c1
-rw-r--r--drivers/net/wireless/ath/Kconfig18
-rw-r--r--drivers/net/wireless/ath/Makefile5
-rw-r--r--drivers/net/wireless/ath/ar5523/ar5523.c1
-rw-r--r--drivers/net/wireless/ath/ath10k/bmi.c42
-rw-r--r--drivers/net/wireless/ath/ath10k/ce.c397
-rw-r--r--drivers/net/wireless/ath/ath10k/ce.h126
-rw-r--r--drivers/net/wireless/ath/ath10k/core.c355
-rw-r--r--drivers/net/wireless/ath/ath10k/core.h80
-rw-r--r--drivers/net/wireless/ath/ath10k/debug.c157
-rw-r--r--drivers/net/wireless/ath/ath10k/debug.h27
-rw-r--r--drivers/net/wireless/ath/ath10k/htc.c241
-rw-r--r--drivers/net/wireless/ath/ath10k/htc.h5
-rw-r--r--drivers/net/wireless/ath/ath10k/htt.c19
-rw-r--r--drivers/net/wireless/ath/ath10k/htt.h13
-rw-r--r--drivers/net/wireless/ath/ath10k/htt_rx.c314
-rw-r--r--drivers/net/wireless/ath/ath10k/htt_tx.c287
-rw-r--r--drivers/net/wireless/ath/ath10k/hw.h79
-rw-r--r--drivers/net/wireless/ath/ath10k/mac.c732
-rw-r--r--drivers/net/wireless/ath/ath10k/mac.h2
-rw-r--r--drivers/net/wireless/ath/ath10k/pci.c465
-rw-r--r--drivers/net/wireless/ath/ath10k/pci.h76
-rw-r--r--drivers/net/wireless/ath/ath10k/rx_desc.h24
-rw-r--r--drivers/net/wireless/ath/ath10k/trace.h32
-rw-r--r--drivers/net/wireless/ath/ath10k/txrx.c67
-rw-r--r--drivers/net/wireless/ath/ath10k/txrx.h5
-rw-r--r--drivers/net/wireless/ath/ath10k/wmi.c1277
-rw-r--r--drivers/net/wireless/ath/ath10k/wmi.h1037
-rw-r--r--drivers/net/wireless/ath/ath5k/ahb.c15
-rw-r--r--drivers/net/wireless/ath/ath5k/base.c6
-rw-r--r--drivers/net/wireless/ath/ath6kl/common.h3
-rw-r--r--drivers/net/wireless/ath/ath6kl/debug.h9
-rw-r--r--drivers/net/wireless/ath/ath6kl/htc.h2
-rw-r--r--drivers/net/wireless/ath/ath9k/Kconfig20
-rw-r--r--drivers/net/wireless/ath/ath9k/Makefile4
-rw-r--r--drivers/net/wireless/ath/ath9k/ahb.c4
-rw-r--r--drivers/net/wireless/ath/ath9k/ani.c6
-rw-r--r--drivers/net/wireless/ath/ath9k/antenna.c36
-rw-r--r--drivers/net/wireless/ath/ath9k/ar5008_phy.c48
-rw-r--r--drivers/net/wireless/ath/ath9k/ar9002_calib.c11
-rw-r--r--drivers/net/wireless/ath/ath9k/ar9002_hw.c26
-rw-r--r--drivers/net/wireless/ath/ath9k/ar9002_phy.c25
-rw-r--r--drivers/net/wireless/ath/ath9k/ar9003_calib.c92
-rw-r--r--drivers/net/wireless/ath/ath9k/ar9003_eeprom.c34
-rw-r--r--drivers/net/wireless/ath/ath9k/ar9003_eeprom.h2
-rw-r--r--drivers/net/wireless/ath/ath9k/ar9003_hw.c5
-rw-r--r--drivers/net/wireless/ath/ath9k/ar9003_mci.c6
-rw-r--r--drivers/net/wireless/ath/ath9k/ar9003_phy.c240
-rw-r--r--drivers/net/wireless/ath/ath9k/ar9003_phy.h4
-rw-r--r--drivers/net/wireless/ath/ath9k/ar9003_rtt.c58
-rw-r--r--drivers/net/wireless/ath/ath9k/ar9485_initvals.h218
-rw-r--r--drivers/net/wireless/ath/ath9k/ar9565_1p0_initvals.h24
-rw-r--r--drivers/net/wireless/ath/ath9k/ath9k.h73
-rw-r--r--drivers/net/wireless/ath/ath9k/beacon.c2
-rw-r--r--drivers/net/wireless/ath/ath9k/calib.c33
-rw-r--r--drivers/net/wireless/ath/ath9k/calib.h3
-rw-r--r--drivers/net/wireless/ath/ath9k/common.c91
-rw-r--r--drivers/net/wireless/ath/ath9k/common.h7
-rw-r--r--drivers/net/wireless/ath/ath9k/debug.c564
-rw-r--r--drivers/net/wireless/ath/ath9k/debug.h12
-rw-r--r--drivers/net/wireless/ath/ath9k/dfs.h2
-rw-r--r--drivers/net/wireless/ath/ath9k/dfs_debug.c32
-rw-r--r--drivers/net/wireless/ath/ath9k/dfs_debug.h16
-rw-r--r--drivers/net/wireless/ath/ath9k/eeprom_4k.c10
-rw-r--r--drivers/net/wireless/ath/ath9k/eeprom_9287.c8
-rw-r--r--drivers/net/wireless/ath/ath9k/eeprom_def.c12
-rw-r--r--drivers/net/wireless/ath/ath9k/gpio.c22
-rw-r--r--drivers/net/wireless/ath/ath9k/htc_drv_debug.c456
-rw-r--r--drivers/net/wireless/ath/ath9k/htc_drv_main.c32
-rw-r--r--drivers/net/wireless/ath/ath9k/hw-ops.h16
-rw-r--r--drivers/net/wireless/ath/ath9k/hw.c128
-rw-r--r--drivers/net/wireless/ath/ath9k/hw.h112
-rw-r--r--drivers/net/wireless/ath/ath9k/init.c140
-rw-r--r--drivers/net/wireless/ath/ath9k/link.c27
-rw-r--r--drivers/net/wireless/ath/ath9k/mac.c6
-rw-r--r--drivers/net/wireless/ath/ath9k/mac.h2
-rw-r--r--drivers/net/wireless/ath/ath9k/main.c327
-rw-r--r--drivers/net/wireless/ath/ath9k/mci.c8
-rw-r--r--drivers/net/wireless/ath/ath9k/pci.c195
-rw-r--r--drivers/net/wireless/ath/ath9k/rc.c32
-rw-r--r--drivers/net/wireless/ath/ath9k/recv.c197
-rw-r--r--drivers/net/wireless/ath/ath9k/wmi.h2
-rw-r--r--drivers/net/wireless/ath/ath9k/xmit.c70
-rw-r--r--drivers/net/wireless/ath/dfs_pattern_detector.c (renamed from drivers/net/wireless/ath/ath9k/dfs_pattern_detector.c)23
-rw-r--r--drivers/net/wireless/ath/dfs_pattern_detector.h (renamed from drivers/net/wireless/ath/ath9k/dfs_pattern_detector.h)28
-rw-r--r--drivers/net/wireless/ath/dfs_pri_detector.c (renamed from drivers/net/wireless/ath/ath9k/dfs_pri_detector.c)10
-rw-r--r--drivers/net/wireless/ath/dfs_pri_detector.h (renamed from drivers/net/wireless/ath/ath9k/dfs_pri_detector.h)2
-rw-r--r--drivers/net/wireless/ath/regd.c140
-rw-r--r--drivers/net/wireless/ath/wcn36xx/Kconfig16
-rw-r--r--drivers/net/wireless/ath/wcn36xx/Makefile7
-rw-r--r--drivers/net/wireless/ath/wcn36xx/debug.c181
-rw-r--r--drivers/net/wireless/ath/wcn36xx/debug.h49
-rw-r--r--drivers/net/wireless/ath/wcn36xx/dxe.c805
-rw-r--r--drivers/net/wireless/ath/wcn36xx/dxe.h284
-rw-r--r--drivers/net/wireless/ath/wcn36xx/hal.h4657
-rw-r--r--drivers/net/wireless/ath/wcn36xx/main.c1036
-rw-r--r--drivers/net/wireless/ath/wcn36xx/pmc.c62
-rw-r--r--drivers/net/wireless/ath/wcn36xx/pmc.h33
-rw-r--r--drivers/net/wireless/ath/wcn36xx/smd.c2126
-rw-r--r--drivers/net/wireless/ath/wcn36xx/smd.h127
-rw-r--r--drivers/net/wireless/ath/wcn36xx/txrx.c284
-rw-r--r--drivers/net/wireless/ath/wcn36xx/txrx.h160
-rw-r--r--drivers/net/wireless/ath/wcn36xx/wcn36xx.h238
-rw-r--r--drivers/net/wireless/ath/wil6210/cfg80211.c4
-rw-r--r--drivers/net/wireless/ath/wil6210/pcie_bus.c1
-rw-r--r--drivers/net/wireless/atmel.c94
-rw-r--r--drivers/net/wireless/b43/phy_n.c3
-rw-r--r--drivers/net/wireless/b43/xmit.c2
-rw-r--r--drivers/net/wireless/b43legacy/xmit.c2
-rw-r--r--drivers/net/wireless/brcm80211/brcmfmac/bcmsdh.c186
-rw-r--r--drivers/net/wireless/brcm80211/brcmfmac/bcmsdh_sdmmc.c30
-rw-r--r--drivers/net/wireless/brcm80211/brcmfmac/dhd.h32
-rw-r--r--drivers/net/wireless/brcm80211/brcmfmac/dhd_bus.h29
-rw-r--r--drivers/net/wireless/brcm80211/brcmfmac/dhd_linux.c38
-rw-r--r--drivers/net/wireless/brcm80211/brcmfmac/dhd_proto.h12
-rw-r--r--drivers/net/wireless/brcm80211/brcmfmac/dhd_sdio.c343
-rw-r--r--drivers/net/wireless/brcm80211/brcmfmac/fweh.h5
-rw-r--r--drivers/net/wireless/brcm80211/brcmfmac/fwsignal.c2
-rw-r--r--drivers/net/wireless/brcm80211/brcmfmac/sdio_chip.c28
-rw-r--r--drivers/net/wireless/brcm80211/brcmfmac/sdio_chip.h31
-rw-r--r--drivers/net/wireless/brcm80211/brcmfmac/sdio_host.h96
-rw-r--r--drivers/net/wireless/brcm80211/brcmfmac/tracepoint.h21
-rw-r--r--drivers/net/wireless/brcm80211/brcmfmac/usb.c5
-rw-r--r--drivers/net/wireless/brcm80211/brcmsmac/aiutils.h18
-rw-r--r--drivers/net/wireless/brcm80211/brcmsmac/ampdu.h22
-rw-r--r--drivers/net/wireless/brcm80211/brcmsmac/antsel.h14
-rw-r--r--drivers/net/wireless/brcm80211/brcmsmac/channel.h20
-rw-r--r--drivers/net/wireless/brcm80211/brcmsmac/mac80211_if.h38
-rw-r--r--drivers/net/wireless/brcm80211/brcmsmac/main.c8
-rw-r--r--drivers/net/wireless/brcm80211/brcmsmac/main.h110
-rw-r--r--drivers/net/wireless/brcm80211/brcmsmac/phy/phy_hal.h219
-rw-r--r--drivers/net/wireless/brcm80211/brcmsmac/phy/phy_int.h371
-rw-r--r--drivers/net/wireless/brcm80211/brcmsmac/phy_shim.h91
-rw-r--r--drivers/net/wireless/brcm80211/brcmsmac/pmu.h4
-rw-r--r--drivers/net/wireless/brcm80211/brcmsmac/pub.h145
-rw-r--r--drivers/net/wireless/brcm80211/brcmsmac/rate.h48
-rw-r--r--drivers/net/wireless/brcm80211/brcmsmac/stf.h31
-rw-r--r--drivers/net/wireless/brcm80211/brcmsmac/ucode_loader.h16
-rw-r--r--drivers/net/wireless/brcm80211/include/brcm_hw_ids.h1
-rw-r--r--drivers/net/wireless/brcm80211/include/brcmu_d11.h2
-rw-r--r--drivers/net/wireless/brcm80211/include/brcmu_utils.h44
-rw-r--r--drivers/net/wireless/cw1200/cw1200_spi.c4
-rw-r--r--drivers/net/wireless/hostap/hostap_info.c2
-rw-r--r--drivers/net/wireless/ipw2x00/ipw2200.c4
-rw-r--r--drivers/net/wireless/ipw2x00/libipw.h87
-rw-r--r--drivers/net/wireless/iwlegacy/3945-mac.c2
-rw-r--r--drivers/net/wireless/iwlegacy/3945.h82
-rw-r--r--drivers/net/wireless/iwlegacy/4965-mac.c2
-rw-r--r--drivers/net/wireless/iwlegacy/4965.h2
-rw-r--r--drivers/net/wireless/iwlegacy/common.h66
-rw-r--r--drivers/net/wireless/iwlwifi/dvm/agn.h2
-rw-r--r--drivers/net/wireless/iwlwifi/dvm/dev.h2
-rw-r--r--drivers/net/wireless/iwlwifi/dvm/rs.h8
-rw-r--r--drivers/net/wireless/iwlwifi/dvm/tx.c14
-rw-r--r--drivers/net/wireless/iwlwifi/dvm/ucode.c9
-rw-r--r--drivers/net/wireless/iwlwifi/iwl-7000.c14
-rw-r--r--drivers/net/wireless/iwlwifi/iwl-config.h1
-rw-r--r--drivers/net/wireless/iwlwifi/iwl-csr.h32
-rw-r--r--drivers/net/wireless/iwlwifi/iwl-drv.c37
-rw-r--r--drivers/net/wireless/iwlwifi/iwl-fw-file.h4
-rw-r--r--drivers/net/wireless/iwlwifi/iwl-fw.h29
-rw-r--r--drivers/net/wireless/iwlwifi/iwl-io.c1
-rw-r--r--drivers/net/wireless/iwlwifi/iwl-prph.h2
-rw-r--r--drivers/net/wireless/iwlwifi/iwl-trans.h24
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/bt-coex.c638
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/constants.h4
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/d3.c515
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/debugfs.c207
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/fw-api-bt-coex.h149
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/fw-api-d3.h69
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/fw-api-mac.h11
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/fw-api-power.h29
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/fw-api-rs.h21
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/fw-api-scan.h34
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/fw-api-sta.h55
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/fw-api.h16
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/fw.c31
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/mac-ctxt.c81
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/mac80211.c256
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/mvm.h89
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/nvm.c101
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/ops.c62
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/power.c70
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/quota.c42
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/rs.c793
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/rs.h163
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/rx.c27
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/scan.c462
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/sta.c206
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/sta.h4
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/testmode.h95
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/time-event.c5
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/time-event.h4
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/tx.c49
-rw-r--r--drivers/net/wireless/iwlwifi/mvm/utils.c2
-rw-r--r--drivers/net/wireless/iwlwifi/pcie/drv.c8
-rw-r--r--drivers/net/wireless/iwlwifi/pcie/trans.c127
-rw-r--r--drivers/net/wireless/iwlwifi/pcie/tx.c43
-rw-r--r--drivers/net/wireless/libertas/firmware.c5
-rw-r--r--drivers/net/wireless/libertas/if_cs.c8
-rw-r--r--drivers/net/wireless/libertas/if_sdio.c8
-rw-r--r--drivers/net/wireless/libertas/if_spi.c6
-rw-r--r--drivers/net/wireless/libertas/if_usb.c17
-rw-r--r--drivers/net/wireless/mac80211_hwsim.c27
-rw-r--r--drivers/net/wireless/mwifiex/cmdevt.c2
-rw-r--r--drivers/net/wireless/mwifiex/join.c2
-rw-r--r--drivers/net/wireless/mwifiex/main.c4
-rw-r--r--drivers/net/wireless/mwifiex/pcie.c6
-rw-r--r--drivers/net/wireless/mwifiex/sta_cmd.c2
-rw-r--r--drivers/net/wireless/mwifiex/wmm.c2
-rw-r--r--drivers/net/wireless/mwifiex/wmm.h24
-rw-r--r--drivers/net/wireless/mwl8k.c2
-rw-r--r--drivers/net/wireless/orinoco/orinoco.h31
-rw-r--r--drivers/net/wireless/orinoco/orinoco_nortel.c2
-rw-r--r--drivers/net/wireless/orinoco/orinoco_pci.c2
-rw-r--r--drivers/net/wireless/orinoco/orinoco_plx.c2
-rw-r--r--drivers/net/wireless/orinoco/orinoco_tmd.c2
-rw-r--r--drivers/net/wireless/p54/p54pci.c1
-rw-r--r--drivers/net/wireless/p54/p54spi.c2
-rw-r--r--drivers/net/wireless/prism54/isl_ioctl.c10
-rw-r--r--drivers/net/wireless/prism54/islpci_dev.c2
-rw-r--r--drivers/net/wireless/prism54/oid_mgt.c2
-rw-r--r--drivers/net/wireless/rt2x00/Kconfig28
-rw-r--r--drivers/net/wireless/rt2x00/Makefile2
-rw-r--r--drivers/net/wireless/rt2x00/rt2400pci.c2
-rw-r--r--drivers/net/wireless/rt2x00/rt2800.h44
-rw-r--r--drivers/net/wireless/rt2x00/rt2800lib.c289
-rw-r--r--drivers/net/wireless/rt2x00/rt2800mmio.c873
-rw-r--r--drivers/net/wireless/rt2x00/rt2800mmio.h165
-rw-r--r--drivers/net/wireless/rt2x00/rt2800pci.c951
-rw-r--r--drivers/net/wireless/rt2x00/rt2800pci.h97
-rw-r--r--drivers/net/wireless/rt2x00/rt2800soc.c263
-rw-r--r--drivers/net/wireless/rt2x00/rt2800usb.c29
-rw-r--r--drivers/net/wireless/rt2x00/rt2x00.h103
-rw-r--r--drivers/net/wireless/rt2x00/rt2x00crypto.c4
-rw-r--r--drivers/net/wireless/rt2x00/rt2x00debug.c2
-rw-r--r--drivers/net/wireless/rt2x00/rt2x00dev.c8
-rw-r--r--drivers/net/wireless/rt2x00/rt2x00link.c74
-rw-r--r--drivers/net/wireless/rt2x00/rt2x00mac.c9
-rw-r--r--drivers/net/wireless/rt2x00/rt2x00pci.c2
-rw-r--r--drivers/net/wireless/rt2x00/rt2x00queue.c39
-rw-r--r--drivers/net/wireless/rt2x00/rt2x00usb.c2
-rw-r--r--drivers/net/wireless/rt2x00/rt61pci.c20
-rw-r--r--drivers/net/wireless/rt2x00/rt73usb.c18
-rw-r--r--drivers/net/wireless/rtl818x/rtl8180/dev.c1
-rw-r--r--drivers/net/wireless/rtlwifi/base.c29
-rw-r--r--drivers/net/wireless/rtlwifi/base.h2
-rw-r--r--drivers/net/wireless/rtlwifi/cam.h10
-rw-r--r--drivers/net/wireless/rtlwifi/core.c10
-rw-r--r--drivers/net/wireless/rtlwifi/efuse.c18
-rw-r--r--drivers/net/wireless/rtlwifi/efuse.h29
-rw-r--r--drivers/net/wireless/rtlwifi/pci.c4
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8188ee/hw.c1
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8188ee/phy.c28
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8188ee/phy.h52
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8188ee/sw.c3
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8188ee/trx.c1
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192c/dm_common.c25
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192c/phy_common.c30
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192c/phy_common.h4
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192ce/def.h2
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192ce/phy.h52
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192ce/reg.h20
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192ce/rf.h13
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192ce/sw.c3
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192ce/trx.c1
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192cu/mac.c187
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192cu/rf.h13
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192cu/sw.c3
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192cu/trx.c2
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192de/dm.c8
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192de/hw.c18
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192de/hw.h7
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192de/phy.c28
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192de/phy.h49
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192de/rf.h18
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192de/sw.c3
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192de/trx.c1
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192se/reg.h5
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8192se/trx.c1
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8723ae/phy.c29
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8723ae/phy.h62
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8723ae/rf.h13
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8723ae/sw.c3
-rw-r--r--drivers/net/wireless/rtlwifi/rtl8723ae/trx.c1
-rw-r--r--drivers/net/wireless/rtlwifi/usb.c6
-rw-r--r--drivers/net/wireless/rtlwifi/wifi.h2
-rw-r--r--drivers/net/wireless/ti/wl1251/spi.c2
-rw-r--r--drivers/net/wireless/ti/wl1251/wl1251.h4
-rw-r--r--drivers/net/wireless/ti/wl12xx/main.c18
-rw-r--r--drivers/net/wireless/ti/wl18xx/main.c127
-rw-r--r--drivers/net/wireless/ti/wl18xx/reg.h33
-rw-r--r--drivers/net/wireless/ti/wlcore/acx.c10
-rw-r--r--drivers/net/wireless/ti/wlcore/cmd.c70
-rw-r--r--drivers/net/wireless/ti/wlcore/cmd.h3
-rw-r--r--drivers/net/wireless/ti/wlcore/conf.h5
-rw-r--r--drivers/net/wireless/ti/wlcore/debugfs.c18
-rw-r--r--drivers/net/wireless/ti/wlcore/event.c1
-rw-r--r--drivers/net/wireless/ti/wlcore/hw_ops.h9
-rw-r--r--drivers/net/wireless/ti/wlcore/init.c6
-rw-r--r--drivers/net/wireless/ti/wlcore/io.h4
-rw-r--r--drivers/net/wireless/ti/wlcore/main.c232
-rw-r--r--drivers/net/wireless/ti/wlcore/ps.c4
-rw-r--r--drivers/net/wireless/ti/wlcore/scan.c51
-rw-r--r--drivers/net/wireless/ti/wlcore/spi.c2
-rw-r--r--drivers/net/wireless/ti/wlcore/testmode.c16
-rw-r--r--drivers/net/wireless/ti/wlcore/tx.c27
-rw-r--r--drivers/net/wireless/ti/wlcore/tx.h3
-rw-r--r--drivers/net/wireless/ti/wlcore/wlcore.h11
-rw-r--r--drivers/net/wireless/ti/wlcore/wlcore_i.h13
-rw-r--r--drivers/net/xen-netback/common.h12
-rw-r--r--drivers/net/xen-netback/interface.c16
-rw-r--r--drivers/net/xen-netback/netback.c294
-rw-r--r--drivers/net/xen-netback/xenbus.c52
-rw-r--r--drivers/net/xen-netfront.c4
632 files changed, 34179 insertions, 13550 deletions
diff --git a/drivers/net/bonding/Makefile b/drivers/net/bonding/Makefile
index 4c21bf6b8b2f..5a5d720da929 100644
--- a/drivers/net/bonding/Makefile
+++ b/drivers/net/bonding/Makefile
@@ -4,7 +4,7 @@
4 4
5obj-$(CONFIG_BONDING) += bonding.o 5obj-$(CONFIG_BONDING) += bonding.o
6 6
7bonding-objs := bond_main.o bond_3ad.o bond_alb.o bond_sysfs.o bond_debugfs.o 7bonding-objs := bond_main.o bond_3ad.o bond_alb.o bond_sysfs.o bond_debugfs.o bond_netlink.o bond_options.o
8 8
9proc-$(CONFIG_PROC_FS) += bond_procfs.o 9proc-$(CONFIG_PROC_FS) += bond_procfs.o
10bonding-objs += $(proc-y) 10bonding-objs += $(proc-y)
diff --git a/drivers/net/bonding/bond_3ad.c b/drivers/net/bonding/bond_3ad.c
index 0d8f427ade93..187b1b7772ef 100644
--- a/drivers/net/bonding/bond_3ad.c
+++ b/drivers/net/bonding/bond_3ad.c
@@ -136,41 +136,6 @@ static inline struct bonding *__get_bond_by_port(struct port *port)
136} 136}
137 137
138/** 138/**
139 * __get_first_port - get the first port in the bond
140 * @bond: the bond we're looking at
141 *
142 * Return the port of the first slave in @bond, or %NULL if it can't be found.
143 */
144static inline struct port *__get_first_port(struct bonding *bond)
145{
146 struct slave *first_slave = bond_first_slave(bond);
147
148 return first_slave ? &(SLAVE_AD_INFO(first_slave).port) : NULL;
149}
150
151/**
152 * __get_next_port - get the next port in the bond
153 * @port: the port we're looking at
154 *
155 * Return the port of the slave that is next in line of @port's slave in the
156 * bond, or %NULL if it can't be found.
157 */
158static inline struct port *__get_next_port(struct port *port)
159{
160 struct bonding *bond = __get_bond_by_port(port);
161 struct slave *slave = port->slave, *slave_next;
162
163 // If there's no bond for this port, or this is the last slave
164 if (bond == NULL)
165 return NULL;
166 slave_next = bond_next_slave(bond, slave);
167 if (!slave_next || bond_is_first_slave(bond, slave_next))
168 return NULL;
169
170 return &(SLAVE_AD_INFO(slave_next).port);
171}
172
173/**
174 * __get_first_agg - get the first aggregator in the bond 139 * __get_first_agg - get the first aggregator in the bond
175 * @bond: the bond we're looking at 140 * @bond: the bond we're looking at
176 * 141 *
@@ -190,28 +155,6 @@ static inline struct aggregator *__get_first_agg(struct port *port)
190 return first_slave ? &(SLAVE_AD_INFO(first_slave).aggregator) : NULL; 155 return first_slave ? &(SLAVE_AD_INFO(first_slave).aggregator) : NULL;
191} 156}
192 157
193/**
194 * __get_next_agg - get the next aggregator in the bond
195 * @aggregator: the aggregator we're looking at
196 *
197 * Return the aggregator of the slave that is next in line of @aggregator's
198 * slave in the bond, or %NULL if it can't be found.
199 */
200static inline struct aggregator *__get_next_agg(struct aggregator *aggregator)
201{
202 struct slave *slave = aggregator->slave, *slave_next;
203 struct bonding *bond = bond_get_bond_by_slave(slave);
204
205 // If there's no bond for this aggregator, or this is the last slave
206 if (bond == NULL)
207 return NULL;
208 slave_next = bond_next_slave(bond, slave);
209 if (!slave_next || bond_is_first_slave(bond, slave_next))
210 return NULL;
211
212 return &(SLAVE_AD_INFO(slave_next).aggregator);
213}
214
215/* 158/*
216 * __agg_has_partner 159 * __agg_has_partner
217 * 160 *
@@ -755,16 +698,15 @@ static u32 __get_agg_bandwidth(struct aggregator *aggregator)
755 */ 698 */
756static struct aggregator *__get_active_agg(struct aggregator *aggregator) 699static struct aggregator *__get_active_agg(struct aggregator *aggregator)
757{ 700{
758 struct aggregator *retval = NULL; 701 struct bonding *bond = aggregator->slave->bond;
702 struct list_head *iter;
703 struct slave *slave;
759 704
760 for (; aggregator; aggregator = __get_next_agg(aggregator)) { 705 bond_for_each_slave(bond, slave, iter)
761 if (aggregator->is_active) { 706 if (SLAVE_AD_INFO(slave).aggregator.is_active)
762 retval = aggregator; 707 return &(SLAVE_AD_INFO(slave).aggregator);
763 break;
764 }
765 }
766 708
767 return retval; 709 return NULL;
768} 710}
769 711
770/** 712/**
@@ -1274,12 +1216,17 @@ static void ad_port_selection_logic(struct port *port)
1274{ 1216{
1275 struct aggregator *aggregator, *free_aggregator = NULL, *temp_aggregator; 1217 struct aggregator *aggregator, *free_aggregator = NULL, *temp_aggregator;
1276 struct port *last_port = NULL, *curr_port; 1218 struct port *last_port = NULL, *curr_port;
1219 struct list_head *iter;
1220 struct bonding *bond;
1221 struct slave *slave;
1277 int found = 0; 1222 int found = 0;
1278 1223
1279 // if the port is already Selected, do nothing 1224 // if the port is already Selected, do nothing
1280 if (port->sm_vars & AD_PORT_SELECTED) 1225 if (port->sm_vars & AD_PORT_SELECTED)
1281 return; 1226 return;
1282 1227
1228 bond = __get_bond_by_port(port);
1229
1283 // if the port is connected to other aggregator, detach it 1230 // if the port is connected to other aggregator, detach it
1284 if (port->aggregator) { 1231 if (port->aggregator) {
1285 // detach the port from its former aggregator 1232 // detach the port from its former aggregator
@@ -1320,8 +1267,8 @@ static void ad_port_selection_logic(struct port *port)
1320 } 1267 }
1321 } 1268 }
1322 // search on all aggregators for a suitable aggregator for this port 1269 // search on all aggregators for a suitable aggregator for this port
1323 for (aggregator = __get_first_agg(port); aggregator; 1270 bond_for_each_slave(bond, slave, iter) {
1324 aggregator = __get_next_agg(aggregator)) { 1271 aggregator = &(SLAVE_AD_INFO(slave).aggregator);
1325 1272
1326 // keep a free aggregator for later use(if needed) 1273 // keep a free aggregator for later use(if needed)
1327 if (!aggregator->lag_ports) { 1274 if (!aggregator->lag_ports) {
@@ -1515,19 +1462,23 @@ static int agg_device_up(const struct aggregator *agg)
1515static void ad_agg_selection_logic(struct aggregator *agg) 1462static void ad_agg_selection_logic(struct aggregator *agg)
1516{ 1463{
1517 struct aggregator *best, *active, *origin; 1464 struct aggregator *best, *active, *origin;
1465 struct bonding *bond = agg->slave->bond;
1466 struct list_head *iter;
1467 struct slave *slave;
1518 struct port *port; 1468 struct port *port;
1519 1469
1520 origin = agg; 1470 origin = agg;
1521 active = __get_active_agg(agg); 1471 active = __get_active_agg(agg);
1522 best = (active && agg_device_up(active)) ? active : NULL; 1472 best = (active && agg_device_up(active)) ? active : NULL;
1523 1473
1524 do { 1474 bond_for_each_slave(bond, slave, iter) {
1475 agg = &(SLAVE_AD_INFO(slave).aggregator);
1476
1525 agg->is_active = 0; 1477 agg->is_active = 0;
1526 1478
1527 if (agg->num_of_ports && agg_device_up(agg)) 1479 if (agg->num_of_ports && agg_device_up(agg))
1528 best = ad_agg_selection_test(best, agg); 1480 best = ad_agg_selection_test(best, agg);
1529 1481 }
1530 } while ((agg = __get_next_agg(agg)));
1531 1482
1532 if (best && 1483 if (best &&
1533 __get_agg_selection_mode(best->lag_ports) == BOND_AD_STABLE) { 1484 __get_agg_selection_mode(best->lag_ports) == BOND_AD_STABLE) {
@@ -1565,8 +1516,8 @@ static void ad_agg_selection_logic(struct aggregator *agg)
1565 best->lag_ports, best->slave, 1516 best->lag_ports, best->slave,
1566 best->slave ? best->slave->dev->name : "NULL"); 1517 best->slave ? best->slave->dev->name : "NULL");
1567 1518
1568 for (agg = __get_first_agg(best->lag_ports); agg; 1519 bond_for_each_slave(bond, slave, iter) {
1569 agg = __get_next_agg(agg)) { 1520 agg = &(SLAVE_AD_INFO(slave).aggregator);
1570 1521
1571 pr_debug("Agg=%d; P=%d; a k=%d; p k=%d; Ind=%d; Act=%d\n", 1522 pr_debug("Agg=%d; P=%d; a k=%d; p k=%d; Ind=%d; Act=%d\n",
1572 agg->aggregator_identifier, agg->num_of_ports, 1523 agg->aggregator_identifier, agg->num_of_ports,
@@ -1614,13 +1565,7 @@ static void ad_agg_selection_logic(struct aggregator *agg)
1614 } 1565 }
1615 } 1566 }
1616 1567
1617 if (origin->slave) { 1568 bond_3ad_set_carrier(bond);
1618 struct bonding *bond;
1619
1620 bond = bond_get_bond_by_slave(origin->slave);
1621 if (bond)
1622 bond_3ad_set_carrier(bond);
1623 }
1624} 1569}
1625 1570
1626/** 1571/**
@@ -1969,6 +1914,9 @@ void bond_3ad_unbind_slave(struct slave *slave)
1969 struct port *port, *prev_port, *temp_port; 1914 struct port *port, *prev_port, *temp_port;
1970 struct aggregator *aggregator, *new_aggregator, *temp_aggregator; 1915 struct aggregator *aggregator, *new_aggregator, *temp_aggregator;
1971 int select_new_active_agg = 0; 1916 int select_new_active_agg = 0;
1917 struct bonding *bond = slave->bond;
1918 struct slave *slave_iter;
1919 struct list_head *iter;
1972 1920
1973 // find the aggregator related to this slave 1921 // find the aggregator related to this slave
1974 aggregator = &(SLAVE_AD_INFO(slave).aggregator); 1922 aggregator = &(SLAVE_AD_INFO(slave).aggregator);
@@ -1998,14 +1946,16 @@ void bond_3ad_unbind_slave(struct slave *slave)
1998 // reason to search for new aggregator, and that we will find one 1946 // reason to search for new aggregator, and that we will find one
1999 if ((aggregator->lag_ports != port) || (aggregator->lag_ports->next_port_in_aggregator)) { 1947 if ((aggregator->lag_ports != port) || (aggregator->lag_ports->next_port_in_aggregator)) {
2000 // find new aggregator for the related port(s) 1948 // find new aggregator for the related port(s)
2001 new_aggregator = __get_first_agg(port); 1949 bond_for_each_slave(bond, slave_iter, iter) {
2002 for (; new_aggregator; new_aggregator = __get_next_agg(new_aggregator)) { 1950 new_aggregator = &(SLAVE_AD_INFO(slave_iter).aggregator);
2003 // if the new aggregator is empty, or it is connected to our port only 1951 // if the new aggregator is empty, or it is connected to our port only
2004 if (!new_aggregator->lag_ports 1952 if (!new_aggregator->lag_ports
2005 || ((new_aggregator->lag_ports == port) 1953 || ((new_aggregator->lag_ports == port)
2006 && !new_aggregator->lag_ports->next_port_in_aggregator)) 1954 && !new_aggregator->lag_ports->next_port_in_aggregator))
2007 break; 1955 break;
2008 } 1956 }
1957 if (!slave_iter)
1958 new_aggregator = NULL;
2009 // if new aggregator found, copy the aggregator's parameters 1959 // if new aggregator found, copy the aggregator's parameters
2010 // and connect the related lag_ports to the new aggregator 1960 // and connect the related lag_ports to the new aggregator
2011 if ((new_aggregator) && ((!new_aggregator->lag_ports) || ((new_aggregator->lag_ports == port) && !new_aggregator->lag_ports->next_port_in_aggregator))) { 1961 if ((new_aggregator) && ((!new_aggregator->lag_ports) || ((new_aggregator->lag_ports == port) && !new_aggregator->lag_ports->next_port_in_aggregator))) {
@@ -2056,15 +2006,17 @@ void bond_3ad_unbind_slave(struct slave *slave)
2056 pr_info("%s: Removing an active aggregator\n", 2006 pr_info("%s: Removing an active aggregator\n",
2057 slave->bond->dev->name); 2007 slave->bond->dev->name);
2058 // select new active aggregator 2008 // select new active aggregator
2059 ad_agg_selection_logic(__get_first_agg(port)); 2009 temp_aggregator = __get_first_agg(port);
2010 if (temp_aggregator)
2011 ad_agg_selection_logic(temp_aggregator);
2060 } 2012 }
2061 } 2013 }
2062 } 2014 }
2063 2015
2064 pr_debug("Unbinding port %d\n", port->actor_port_number); 2016 pr_debug("Unbinding port %d\n", port->actor_port_number);
2065 // find the aggregator that this port is connected to 2017 // find the aggregator that this port is connected to
2066 temp_aggregator = __get_first_agg(port); 2018 bond_for_each_slave(bond, slave_iter, iter) {
2067 for (; temp_aggregator; temp_aggregator = __get_next_agg(temp_aggregator)) { 2019 temp_aggregator = &(SLAVE_AD_INFO(slave_iter).aggregator);
2068 prev_port = NULL; 2020 prev_port = NULL;
2069 // search the port in the aggregator's related ports 2021 // search the port in the aggregator's related ports
2070 for (temp_port = temp_aggregator->lag_ports; temp_port; 2022 for (temp_port = temp_aggregator->lag_ports; temp_port;
@@ -2111,19 +2063,24 @@ void bond_3ad_state_machine_handler(struct work_struct *work)
2111{ 2063{
2112 struct bonding *bond = container_of(work, struct bonding, 2064 struct bonding *bond = container_of(work, struct bonding,
2113 ad_work.work); 2065 ad_work.work);
2114 struct port *port;
2115 struct aggregator *aggregator; 2066 struct aggregator *aggregator;
2067 struct list_head *iter;
2068 struct slave *slave;
2069 struct port *port;
2116 2070
2117 read_lock(&bond->lock); 2071 read_lock(&bond->lock);
2118 2072
2119 //check if there are any slaves 2073 //check if there are any slaves
2120 if (list_empty(&bond->slave_list)) 2074 if (!bond_has_slaves(bond))
2121 goto re_arm; 2075 goto re_arm;
2122 2076
2123 // check if agg_select_timer timer after initialize is timed out 2077 // check if agg_select_timer timer after initialize is timed out
2124 if (BOND_AD_INFO(bond).agg_select_timer && !(--BOND_AD_INFO(bond).agg_select_timer)) { 2078 if (BOND_AD_INFO(bond).agg_select_timer && !(--BOND_AD_INFO(bond).agg_select_timer)) {
2079 slave = bond_first_slave(bond);
2080 port = slave ? &(SLAVE_AD_INFO(slave).port) : NULL;
2081
2125 // select the active aggregator for the bond 2082 // select the active aggregator for the bond
2126 if ((port = __get_first_port(bond))) { 2083 if (port) {
2127 if (!port->slave) { 2084 if (!port->slave) {
2128 pr_warning("%s: Warning: bond's first port is uninitialized\n", 2085 pr_warning("%s: Warning: bond's first port is uninitialized\n",
2129 bond->dev->name); 2086 bond->dev->name);
@@ -2137,7 +2094,8 @@ void bond_3ad_state_machine_handler(struct work_struct *work)
2137 } 2094 }
2138 2095
2139 // for each port run the state machines 2096 // for each port run the state machines
2140 for (port = __get_first_port(bond); port; port = __get_next_port(port)) { 2097 bond_for_each_slave(bond, slave, iter) {
2098 port = &(SLAVE_AD_INFO(slave).port);
2141 if (!port->slave) { 2099 if (!port->slave) {
2142 pr_warning("%s: Warning: Found an uninitialized port\n", 2100 pr_warning("%s: Warning: Found an uninitialized port\n",
2143 bond->dev->name); 2101 bond->dev->name);
@@ -2382,9 +2340,12 @@ int __bond_3ad_get_active_agg_info(struct bonding *bond,
2382 struct ad_info *ad_info) 2340 struct ad_info *ad_info)
2383{ 2341{
2384 struct aggregator *aggregator = NULL; 2342 struct aggregator *aggregator = NULL;
2343 struct list_head *iter;
2344 struct slave *slave;
2385 struct port *port; 2345 struct port *port;
2386 2346
2387 for (port = __get_first_port(bond); port; port = __get_next_port(port)) { 2347 bond_for_each_slave_rcu(bond, slave, iter) {
2348 port = &(SLAVE_AD_INFO(slave).port);
2388 if (port->aggregator && port->aggregator->is_active) { 2349 if (port->aggregator && port->aggregator->is_active) {
2389 aggregator = port->aggregator; 2350 aggregator = port->aggregator;
2390 break; 2351 break;
@@ -2408,25 +2369,25 @@ int bond_3ad_get_active_agg_info(struct bonding *bond, struct ad_info *ad_info)
2408{ 2369{
2409 int ret; 2370 int ret;
2410 2371
2411 read_lock(&bond->lock); 2372 rcu_read_lock();
2412 ret = __bond_3ad_get_active_agg_info(bond, ad_info); 2373 ret = __bond_3ad_get_active_agg_info(bond, ad_info);
2413 read_unlock(&bond->lock); 2374 rcu_read_unlock();
2414 2375
2415 return ret; 2376 return ret;
2416} 2377}
2417 2378
2418int bond_3ad_xmit_xor(struct sk_buff *skb, struct net_device *dev) 2379int bond_3ad_xmit_xor(struct sk_buff *skb, struct net_device *dev)
2419{ 2380{
2420 struct slave *slave, *start_at;
2421 struct bonding *bond = netdev_priv(dev); 2381 struct bonding *bond = netdev_priv(dev);
2422 int slave_agg_no; 2382 struct slave *slave, *first_ok_slave;
2423 int slaves_in_agg; 2383 struct aggregator *agg;
2424 int agg_id;
2425 int i;
2426 struct ad_info ad_info; 2384 struct ad_info ad_info;
2385 struct list_head *iter;
2386 int slaves_in_agg;
2387 int slave_agg_no;
2427 int res = 1; 2388 int res = 1;
2389 int agg_id;
2428 2390
2429 read_lock(&bond->lock);
2430 if (__bond_3ad_get_active_agg_info(bond, &ad_info)) { 2391 if (__bond_3ad_get_active_agg_info(bond, &ad_info)) {
2431 pr_debug("%s: Error: __bond_3ad_get_active_agg_info failed\n", 2392 pr_debug("%s: Error: __bond_3ad_get_active_agg_info failed\n",
2432 dev->name); 2393 dev->name);
@@ -2437,20 +2398,28 @@ int bond_3ad_xmit_xor(struct sk_buff *skb, struct net_device *dev)
2437 agg_id = ad_info.aggregator_id; 2398 agg_id = ad_info.aggregator_id;
2438 2399
2439 if (slaves_in_agg == 0) { 2400 if (slaves_in_agg == 0) {
2440 /*the aggregator is empty*/
2441 pr_debug("%s: Error: active aggregator is empty\n", dev->name); 2401 pr_debug("%s: Error: active aggregator is empty\n", dev->name);
2442 goto out; 2402 goto out;
2443 } 2403 }
2444 2404
2445 slave_agg_no = bond->xmit_hash_policy(skb, slaves_in_agg); 2405 slave_agg_no = bond_xmit_hash(bond, skb, slaves_in_agg);
2406 first_ok_slave = NULL;
2446 2407
2447 bond_for_each_slave(bond, slave) { 2408 bond_for_each_slave_rcu(bond, slave, iter) {
2448 struct aggregator *agg = SLAVE_AD_INFO(slave).port.aggregator; 2409 agg = SLAVE_AD_INFO(slave).port.aggregator;
2410 if (!agg || agg->aggregator_identifier != agg_id)
2411 continue;
2449 2412
2450 if (agg && (agg->aggregator_identifier == agg_id)) { 2413 if (slave_agg_no >= 0) {
2414 if (!first_ok_slave && SLAVE_IS_OK(slave))
2415 first_ok_slave = slave;
2451 slave_agg_no--; 2416 slave_agg_no--;
2452 if (slave_agg_no < 0) 2417 continue;
2453 break; 2418 }
2419
2420 if (SLAVE_IS_OK(slave)) {
2421 res = bond_dev_queue_xmit(bond, skb, slave->dev);
2422 goto out;
2454 } 2423 }
2455 } 2424 }
2456 2425
@@ -2460,23 +2429,12 @@ int bond_3ad_xmit_xor(struct sk_buff *skb, struct net_device *dev)
2460 goto out; 2429 goto out;
2461 } 2430 }
2462 2431
2463 start_at = slave; 2432 /* we couldn't find any suitable slave after the agg_no, so use the
2464 2433 * first suitable found, if found. */
2465 bond_for_each_slave_from(bond, slave, i, start_at) { 2434 if (first_ok_slave)
2466 int slave_agg_id = 0; 2435 res = bond_dev_queue_xmit(bond, skb, first_ok_slave->dev);
2467 struct aggregator *agg = SLAVE_AD_INFO(slave).port.aggregator;
2468
2469 if (agg)
2470 slave_agg_id = agg->aggregator_identifier;
2471
2472 if (SLAVE_IS_OK(slave) && agg && (slave_agg_id == agg_id)) {
2473 res = bond_dev_queue_xmit(bond, skb, slave->dev);
2474 break;
2475 }
2476 }
2477 2436
2478out: 2437out:
2479 read_unlock(&bond->lock);
2480 if (res) { 2438 if (res) {
2481 /* no suitable interface, frame not sent */ 2439 /* no suitable interface, frame not sent */
2482 kfree_skb(skb); 2440 kfree_skb(skb);
@@ -2515,11 +2473,12 @@ int bond_3ad_lacpdu_recv(const struct sk_buff *skb, struct bonding *bond,
2515void bond_3ad_update_lacp_rate(struct bonding *bond) 2473void bond_3ad_update_lacp_rate(struct bonding *bond)
2516{ 2474{
2517 struct port *port = NULL; 2475 struct port *port = NULL;
2476 struct list_head *iter;
2518 struct slave *slave; 2477 struct slave *slave;
2519 int lacp_fast; 2478 int lacp_fast;
2520 2479
2521 lacp_fast = bond->params.lacp_fast; 2480 lacp_fast = bond->params.lacp_fast;
2522 bond_for_each_slave(bond, slave) { 2481 bond_for_each_slave(bond, slave, iter) {
2523 port = &(SLAVE_AD_INFO(slave).port); 2482 port = &(SLAVE_AD_INFO(slave).port);
2524 __get_state_machine_lock(port); 2483 __get_state_machine_lock(port);
2525 if (lacp_fast) 2484 if (lacp_fast)
diff --git a/drivers/net/bonding/bond_alb.c b/drivers/net/bonding/bond_alb.c
index f428ef574372..02872405d35d 100644
--- a/drivers/net/bonding/bond_alb.c
+++ b/drivers/net/bonding/bond_alb.c
@@ -223,13 +223,14 @@ static long long compute_gap(struct slave *slave)
223static struct slave *tlb_get_least_loaded_slave(struct bonding *bond) 223static struct slave *tlb_get_least_loaded_slave(struct bonding *bond)
224{ 224{
225 struct slave *slave, *least_loaded; 225 struct slave *slave, *least_loaded;
226 struct list_head *iter;
226 long long max_gap; 227 long long max_gap;
227 228
228 least_loaded = NULL; 229 least_loaded = NULL;
229 max_gap = LLONG_MIN; 230 max_gap = LLONG_MIN;
230 231
231 /* Find the slave with the largest gap */ 232 /* Find the slave with the largest gap */
232 bond_for_each_slave(bond, slave) { 233 bond_for_each_slave_rcu(bond, slave, iter) {
233 if (SLAVE_IS_OK(slave)) { 234 if (SLAVE_IS_OK(slave)) {
234 long long gap = compute_gap(slave); 235 long long gap = compute_gap(slave);
235 236
@@ -382,30 +383,64 @@ out:
382static struct slave *rlb_next_rx_slave(struct bonding *bond) 383static struct slave *rlb_next_rx_slave(struct bonding *bond)
383{ 384{
384 struct alb_bond_info *bond_info = &(BOND_ALB_INFO(bond)); 385 struct alb_bond_info *bond_info = &(BOND_ALB_INFO(bond));
385 struct slave *rx_slave, *slave, *start_at; 386 struct slave *before = NULL, *rx_slave = NULL, *slave;
386 int i = 0; 387 struct list_head *iter;
388 bool found = false;
387 389
388 if (bond_info->next_rx_slave) 390 bond_for_each_slave(bond, slave, iter) {
389 start_at = bond_info->next_rx_slave; 391 if (!SLAVE_IS_OK(slave))
390 else 392 continue;
391 start_at = bond_first_slave(bond); 393 if (!found) {
394 if (!before || before->speed < slave->speed)
395 before = slave;
396 } else {
397 if (!rx_slave || rx_slave->speed < slave->speed)
398 rx_slave = slave;
399 }
400 if (slave == bond_info->rx_slave)
401 found = true;
402 }
403 /* we didn't find anything after the current or we have something
404 * better before and up to the current slave
405 */
406 if (!rx_slave || (before && rx_slave->speed < before->speed))
407 rx_slave = before;
392 408
393 rx_slave = NULL; 409 if (rx_slave)
410 bond_info->rx_slave = rx_slave;
394 411
395 bond_for_each_slave_from(bond, slave, i, start_at) { 412 return rx_slave;
396 if (SLAVE_IS_OK(slave)) { 413}
397 if (!rx_slave) { 414
398 rx_slave = slave; 415/* Caller must hold rcu_read_lock() for read */
399 } else if (slave->speed > rx_slave->speed) { 416static struct slave *__rlb_next_rx_slave(struct bonding *bond)
417{
418 struct alb_bond_info *bond_info = &(BOND_ALB_INFO(bond));
419 struct slave *before = NULL, *rx_slave = NULL, *slave;
420 struct list_head *iter;
421 bool found = false;
422
423 bond_for_each_slave_rcu(bond, slave, iter) {
424 if (!SLAVE_IS_OK(slave))
425 continue;
426 if (!found) {
427 if (!before || before->speed < slave->speed)
428 before = slave;
429 } else {
430 if (!rx_slave || rx_slave->speed < slave->speed)
400 rx_slave = slave; 431 rx_slave = slave;
401 }
402 } 432 }
433 if (slave == bond_info->rx_slave)
434 found = true;
403 } 435 }
436 /* we didn't find anything after the current or we have something
437 * better before and up to the current slave
438 */
439 if (!rx_slave || (before && rx_slave->speed < before->speed))
440 rx_slave = before;
404 441
405 if (rx_slave) { 442 if (rx_slave)
406 slave = bond_next_slave(bond, rx_slave); 443 bond_info->rx_slave = rx_slave;
407 bond_info->next_rx_slave = slave;
408 }
409 444
410 return rx_slave; 445 return rx_slave;
411} 446}
@@ -626,12 +661,14 @@ static struct slave *rlb_choose_channel(struct sk_buff *skb, struct bonding *bon
626{ 661{
627 struct alb_bond_info *bond_info = &(BOND_ALB_INFO(bond)); 662 struct alb_bond_info *bond_info = &(BOND_ALB_INFO(bond));
628 struct arp_pkt *arp = arp_pkt(skb); 663 struct arp_pkt *arp = arp_pkt(skb);
629 struct slave *assigned_slave; 664 struct slave *assigned_slave, *curr_active_slave;
630 struct rlb_client_info *client_info; 665 struct rlb_client_info *client_info;
631 u32 hash_index = 0; 666 u32 hash_index = 0;
632 667
633 _lock_rx_hashtbl(bond); 668 _lock_rx_hashtbl(bond);
634 669
670 curr_active_slave = rcu_dereference(bond->curr_active_slave);
671
635 hash_index = _simple_hash((u8 *)&arp->ip_dst, sizeof(arp->ip_dst)); 672 hash_index = _simple_hash((u8 *)&arp->ip_dst, sizeof(arp->ip_dst));
636 client_info = &(bond_info->rx_hashtbl[hash_index]); 673 client_info = &(bond_info->rx_hashtbl[hash_index]);
637 674
@@ -656,14 +693,14 @@ static struct slave *rlb_choose_channel(struct sk_buff *skb, struct bonding *bon
656 * that the new client can be assigned to this entry. 693 * that the new client can be assigned to this entry.
657 */ 694 */
658 if (bond->curr_active_slave && 695 if (bond->curr_active_slave &&
659 client_info->slave != bond->curr_active_slave) { 696 client_info->slave != curr_active_slave) {
660 client_info->slave = bond->curr_active_slave; 697 client_info->slave = curr_active_slave;
661 rlb_update_client(client_info); 698 rlb_update_client(client_info);
662 } 699 }
663 } 700 }
664 } 701 }
665 /* assign a new slave */ 702 /* assign a new slave */
666 assigned_slave = rlb_next_rx_slave(bond); 703 assigned_slave = __rlb_next_rx_slave(bond);
667 704
668 if (assigned_slave) { 705 if (assigned_slave) {
669 if (!(client_info->assigned && 706 if (!(client_info->assigned &&
@@ -726,7 +763,7 @@ static struct slave *rlb_arp_xmit(struct sk_buff *skb, struct bonding *bond)
726 /* Don't modify or load balance ARPs that do not originate locally 763 /* Don't modify or load balance ARPs that do not originate locally
727 * (e.g.,arrive via a bridge). 764 * (e.g.,arrive via a bridge).
728 */ 765 */
729 if (!bond_slave_has_mac(bond, arp->mac_src)) 766 if (!bond_slave_has_mac_rcu(bond, arp->mac_src))
730 return NULL; 767 return NULL;
731 768
732 if (arp->op_code == htons(ARPOP_REPLY)) { 769 if (arp->op_code == htons(ARPOP_REPLY)) {
@@ -1019,7 +1056,7 @@ static void alb_send_learning_packets(struct slave *slave, u8 mac_addr[])
1019 1056
1020 /* loop through vlans and send one packet for each */ 1057 /* loop through vlans and send one packet for each */
1021 rcu_read_lock(); 1058 rcu_read_lock();
1022 netdev_for_each_upper_dev_rcu(bond->dev, upper, iter) { 1059 netdev_for_each_all_upper_dev_rcu(bond->dev, upper, iter) {
1023 if (upper->priv_flags & IFF_802_1Q_VLAN) 1060 if (upper->priv_flags & IFF_802_1Q_VLAN)
1024 alb_send_lp_vid(slave, mac_addr, 1061 alb_send_lp_vid(slave, mac_addr,
1025 vlan_dev_vlan_id(upper)); 1062 vlan_dev_vlan_id(upper));
@@ -1172,10 +1209,11 @@ static void alb_change_hw_addr_on_detach(struct bonding *bond, struct slave *sla
1172 */ 1209 */
1173static int alb_handle_addr_collision_on_attach(struct bonding *bond, struct slave *slave) 1210static int alb_handle_addr_collision_on_attach(struct bonding *bond, struct slave *slave)
1174{ 1211{
1175 struct slave *tmp_slave1, *free_mac_slave = NULL;
1176 struct slave *has_bond_addr = bond->curr_active_slave; 1212 struct slave *has_bond_addr = bond->curr_active_slave;
1213 struct slave *tmp_slave1, *free_mac_slave = NULL;
1214 struct list_head *iter;
1177 1215
1178 if (list_empty(&bond->slave_list)) { 1216 if (!bond_has_slaves(bond)) {
1179 /* this is the first slave */ 1217 /* this is the first slave */
1180 return 0; 1218 return 0;
1181 } 1219 }
@@ -1196,7 +1234,7 @@ static int alb_handle_addr_collision_on_attach(struct bonding *bond, struct slav
1196 /* The slave's address is equal to the address of the bond. 1234 /* The slave's address is equal to the address of the bond.
1197 * Search for a spare address in the bond for this slave. 1235 * Search for a spare address in the bond for this slave.
1198 */ 1236 */
1199 bond_for_each_slave(bond, tmp_slave1) { 1237 bond_for_each_slave(bond, tmp_slave1, iter) {
1200 if (!bond_slave_has_mac(bond, tmp_slave1->perm_hwaddr)) { 1238 if (!bond_slave_has_mac(bond, tmp_slave1->perm_hwaddr)) {
1201 /* no slave has tmp_slave1's perm addr 1239 /* no slave has tmp_slave1's perm addr
1202 * as its curr addr 1240 * as its curr addr
@@ -1246,15 +1284,16 @@ static int alb_handle_addr_collision_on_attach(struct bonding *bond, struct slav
1246 */ 1284 */
1247static int alb_set_mac_address(struct bonding *bond, void *addr) 1285static int alb_set_mac_address(struct bonding *bond, void *addr)
1248{ 1286{
1249 char tmp_addr[ETH_ALEN]; 1287 struct slave *slave, *rollback_slave;
1250 struct slave *slave; 1288 struct list_head *iter;
1251 struct sockaddr sa; 1289 struct sockaddr sa;
1290 char tmp_addr[ETH_ALEN];
1252 int res; 1291 int res;
1253 1292
1254 if (bond->alb_info.rlb_enabled) 1293 if (bond->alb_info.rlb_enabled)
1255 return 0; 1294 return 0;
1256 1295
1257 bond_for_each_slave(bond, slave) { 1296 bond_for_each_slave(bond, slave, iter) {
1258 /* save net_device's current hw address */ 1297 /* save net_device's current hw address */
1259 memcpy(tmp_addr, slave->dev->dev_addr, ETH_ALEN); 1298 memcpy(tmp_addr, slave->dev->dev_addr, ETH_ALEN);
1260 1299
@@ -1274,10 +1313,12 @@ unwind:
1274 sa.sa_family = bond->dev->type; 1313 sa.sa_family = bond->dev->type;
1275 1314
1276 /* unwind from head to the slave that failed */ 1315 /* unwind from head to the slave that failed */
1277 bond_for_each_slave_continue_reverse(bond, slave) { 1316 bond_for_each_slave(bond, rollback_slave, iter) {
1278 memcpy(tmp_addr, slave->dev->dev_addr, ETH_ALEN); 1317 if (rollback_slave == slave)
1279 dev_set_mac_address(slave->dev, &sa); 1318 break;
1280 memcpy(slave->dev->dev_addr, tmp_addr, ETH_ALEN); 1319 memcpy(tmp_addr, rollback_slave->dev->dev_addr, ETH_ALEN);
1320 dev_set_mac_address(rollback_slave->dev, &sa);
1321 memcpy(rollback_slave->dev->dev_addr, tmp_addr, ETH_ALEN);
1281 } 1322 }
1282 1323
1283 return res; 1324 return res;
@@ -1337,11 +1378,6 @@ int bond_alb_xmit(struct sk_buff *skb, struct net_device *bond_dev)
1337 skb_reset_mac_header(skb); 1378 skb_reset_mac_header(skb);
1338 eth_data = eth_hdr(skb); 1379 eth_data = eth_hdr(skb);
1339 1380
1340 /* make sure that the curr_active_slave do not change during tx
1341 */
1342 read_lock(&bond->lock);
1343 read_lock(&bond->curr_slave_lock);
1344
1345 switch (ntohs(skb->protocol)) { 1381 switch (ntohs(skb->protocol)) {
1346 case ETH_P_IP: { 1382 case ETH_P_IP: {
1347 const struct iphdr *iph = ip_hdr(skb); 1383 const struct iphdr *iph = ip_hdr(skb);
@@ -1423,12 +1459,12 @@ int bond_alb_xmit(struct sk_buff *skb, struct net_device *bond_dev)
1423 1459
1424 if (!tx_slave) { 1460 if (!tx_slave) {
1425 /* unbalanced or unassigned, send through primary */ 1461 /* unbalanced or unassigned, send through primary */
1426 tx_slave = bond->curr_active_slave; 1462 tx_slave = rcu_dereference(bond->curr_active_slave);
1427 bond_info->unbalanced_load += skb->len; 1463 bond_info->unbalanced_load += skb->len;
1428 } 1464 }
1429 1465
1430 if (tx_slave && SLAVE_IS_OK(tx_slave)) { 1466 if (tx_slave && SLAVE_IS_OK(tx_slave)) {
1431 if (tx_slave != bond->curr_active_slave) { 1467 if (tx_slave != rcu_dereference(bond->curr_active_slave)) {
1432 memcpy(eth_data->h_source, 1468 memcpy(eth_data->h_source,
1433 tx_slave->dev->dev_addr, 1469 tx_slave->dev->dev_addr,
1434 ETH_ALEN); 1470 ETH_ALEN);
@@ -1443,8 +1479,6 @@ int bond_alb_xmit(struct sk_buff *skb, struct net_device *bond_dev)
1443 } 1479 }
1444 } 1480 }
1445 1481
1446 read_unlock(&bond->curr_slave_lock);
1447 read_unlock(&bond->lock);
1448 if (res) { 1482 if (res) {
1449 /* no suitable interface, frame not sent */ 1483 /* no suitable interface, frame not sent */
1450 kfree_skb(skb); 1484 kfree_skb(skb);
@@ -1458,11 +1492,12 @@ void bond_alb_monitor(struct work_struct *work)
1458 struct bonding *bond = container_of(work, struct bonding, 1492 struct bonding *bond = container_of(work, struct bonding,
1459 alb_work.work); 1493 alb_work.work);
1460 struct alb_bond_info *bond_info = &(BOND_ALB_INFO(bond)); 1494 struct alb_bond_info *bond_info = &(BOND_ALB_INFO(bond));
1495 struct list_head *iter;
1461 struct slave *slave; 1496 struct slave *slave;
1462 1497
1463 read_lock(&bond->lock); 1498 read_lock(&bond->lock);
1464 1499
1465 if (list_empty(&bond->slave_list)) { 1500 if (!bond_has_slaves(bond)) {
1466 bond_info->tx_rebalance_counter = 0; 1501 bond_info->tx_rebalance_counter = 0;
1467 bond_info->lp_counter = 0; 1502 bond_info->lp_counter = 0;
1468 goto re_arm; 1503 goto re_arm;
@@ -1480,7 +1515,7 @@ void bond_alb_monitor(struct work_struct *work)
1480 */ 1515 */
1481 read_lock(&bond->curr_slave_lock); 1516 read_lock(&bond->curr_slave_lock);
1482 1517
1483 bond_for_each_slave(bond, slave) 1518 bond_for_each_slave(bond, slave, iter)
1484 alb_send_learning_packets(slave, slave->dev->dev_addr); 1519 alb_send_learning_packets(slave, slave->dev->dev_addr);
1485 1520
1486 read_unlock(&bond->curr_slave_lock); 1521 read_unlock(&bond->curr_slave_lock);
@@ -1493,7 +1528,7 @@ void bond_alb_monitor(struct work_struct *work)
1493 1528
1494 read_lock(&bond->curr_slave_lock); 1529 read_lock(&bond->curr_slave_lock);
1495 1530
1496 bond_for_each_slave(bond, slave) { 1531 bond_for_each_slave(bond, slave, iter) {
1497 tlb_clear_slave(bond, slave, 1); 1532 tlb_clear_slave(bond, slave, 1);
1498 if (slave == bond->curr_active_slave) { 1533 if (slave == bond->curr_active_slave) {
1499 SLAVE_TLB_INFO(slave).load = 1534 SLAVE_TLB_INFO(slave).load =
@@ -1599,13 +1634,13 @@ int bond_alb_init_slave(struct bonding *bond, struct slave *slave)
1599 */ 1634 */
1600void bond_alb_deinit_slave(struct bonding *bond, struct slave *slave) 1635void bond_alb_deinit_slave(struct bonding *bond, struct slave *slave)
1601{ 1636{
1602 if (!list_empty(&bond->slave_list)) 1637 if (bond_has_slaves(bond))
1603 alb_change_hw_addr_on_detach(bond, slave); 1638 alb_change_hw_addr_on_detach(bond, slave);
1604 1639
1605 tlb_clear_slave(bond, slave, 0); 1640 tlb_clear_slave(bond, slave, 0);
1606 1641
1607 if (bond->alb_info.rlb_enabled) { 1642 if (bond->alb_info.rlb_enabled) {
1608 bond->alb_info.next_rx_slave = NULL; 1643 bond->alb_info.rx_slave = NULL;
1609 rlb_clear_slave(bond, slave); 1644 rlb_clear_slave(bond, slave);
1610 } 1645 }
1611} 1646}
@@ -1669,7 +1704,7 @@ void bond_alb_handle_active_change(struct bonding *bond, struct slave *new_slave
1669 swap_slave = bond->curr_active_slave; 1704 swap_slave = bond->curr_active_slave;
1670 rcu_assign_pointer(bond->curr_active_slave, new_slave); 1705 rcu_assign_pointer(bond->curr_active_slave, new_slave);
1671 1706
1672 if (!new_slave || list_empty(&bond->slave_list)) 1707 if (!new_slave || !bond_has_slaves(bond))
1673 return; 1708 return;
1674 1709
1675 /* set the new curr_active_slave to the bonds mac address 1710 /* set the new curr_active_slave to the bonds mac address
@@ -1692,6 +1727,23 @@ void bond_alb_handle_active_change(struct bonding *bond, struct slave *new_slave
1692 1727
1693 ASSERT_RTNL(); 1728 ASSERT_RTNL();
1694 1729
1730 /* in TLB mode, the slave might flip down/up with the old dev_addr,
1731 * and thus filter bond->dev_addr's packets, so force bond's mac
1732 */
1733 if (bond->params.mode == BOND_MODE_TLB) {
1734 struct sockaddr sa;
1735 u8 tmp_addr[ETH_ALEN];
1736
1737 memcpy(tmp_addr, new_slave->dev->dev_addr, ETH_ALEN);
1738
1739 memcpy(sa.sa_data, bond->dev->dev_addr, bond->dev->addr_len);
1740 sa.sa_family = bond->dev->type;
1741 /* we don't care if it can't change its mac, best effort */
1742 dev_set_mac_address(new_slave->dev, &sa);
1743
1744 memcpy(new_slave->dev->dev_addr, tmp_addr, ETH_ALEN);
1745 }
1746
1695 /* curr_active_slave must be set before calling alb_swap_mac_addr */ 1747 /* curr_active_slave must be set before calling alb_swap_mac_addr */
1696 if (swap_slave) { 1748 if (swap_slave) {
1697 /* swap mac address */ 1749 /* swap mac address */
diff --git a/drivers/net/bonding/bond_alb.h b/drivers/net/bonding/bond_alb.h
index c5eff5dafdfe..4226044efd08 100644
--- a/drivers/net/bonding/bond_alb.h
+++ b/drivers/net/bonding/bond_alb.h
@@ -154,9 +154,7 @@ struct alb_bond_info {
154 u8 rx_ntt; /* flag - need to transmit 154 u8 rx_ntt; /* flag - need to transmit
155 * to all rx clients 155 * to all rx clients
156 */ 156 */
157 struct slave *next_rx_slave;/* next slave to be assigned 157 struct slave *rx_slave;/* last slave to xmit from */
158 * to a new rx client for
159 */
160 u8 primary_is_promisc; /* boolean */ 158 u8 primary_is_promisc; /* boolean */
161 u32 rlb_promisc_timeout_counter;/* counts primary 159 u32 rlb_promisc_timeout_counter;/* counts primary
162 * promiscuity time 160 * promiscuity time
diff --git a/drivers/net/bonding/bond_main.c b/drivers/net/bonding/bond_main.c
index e883bfe2e727..4dd5ee2a34cc 100644
--- a/drivers/net/bonding/bond_main.c
+++ b/drivers/net/bonding/bond_main.c
@@ -78,6 +78,8 @@
78#include <net/netns/generic.h> 78#include <net/netns/generic.h>
79#include <net/pkt_sched.h> 79#include <net/pkt_sched.h>
80#include <linux/rculist.h> 80#include <linux/rculist.h>
81#include <net/flow_keys.h>
82#include <linux/reciprocal_div.h>
81#include "bonding.h" 83#include "bonding.h"
82#include "bond_3ad.h" 84#include "bond_3ad.h"
83#include "bond_alb.h" 85#include "bond_alb.h"
@@ -110,6 +112,7 @@ static char *fail_over_mac;
110static int all_slaves_active; 112static int all_slaves_active;
111static struct bond_params bonding_defaults; 113static struct bond_params bonding_defaults;
112static int resend_igmp = BOND_DEFAULT_RESEND_IGMP; 114static int resend_igmp = BOND_DEFAULT_RESEND_IGMP;
115static int packets_per_slave = 1;
113 116
114module_param(max_bonds, int, 0); 117module_param(max_bonds, int, 0);
115MODULE_PARM_DESC(max_bonds, "Max number of bonded devices"); 118MODULE_PARM_DESC(max_bonds, "Max number of bonded devices");
@@ -159,7 +162,8 @@ MODULE_PARM_DESC(min_links, "Minimum number of available links before turning on
159module_param(xmit_hash_policy, charp, 0); 162module_param(xmit_hash_policy, charp, 0);
160MODULE_PARM_DESC(xmit_hash_policy, "balance-xor and 802.3ad hashing method; " 163MODULE_PARM_DESC(xmit_hash_policy, "balance-xor and 802.3ad hashing method; "
161 "0 for layer 2 (default), 1 for layer 3+4, " 164 "0 for layer 2 (default), 1 for layer 3+4, "
162 "2 for layer 2+3"); 165 "2 for layer 2+3, 3 for encap layer 2+3, "
166 "4 for encap layer 3+4");
163module_param(arp_interval, int, 0); 167module_param(arp_interval, int, 0);
164MODULE_PARM_DESC(arp_interval, "arp interval in milliseconds"); 168MODULE_PARM_DESC(arp_interval, "arp interval in milliseconds");
165module_param_array(arp_ip_target, charp, NULL, 0); 169module_param_array(arp_ip_target, charp, NULL, 0);
@@ -181,6 +185,10 @@ MODULE_PARM_DESC(all_slaves_active, "Keep all frames received on an interface"
181module_param(resend_igmp, int, 0); 185module_param(resend_igmp, int, 0);
182MODULE_PARM_DESC(resend_igmp, "Number of IGMP membership reports to send on " 186MODULE_PARM_DESC(resend_igmp, "Number of IGMP membership reports to send on "
183 "link failure"); 187 "link failure");
188module_param(packets_per_slave, int, 0);
189MODULE_PARM_DESC(packets_per_slave, "Packets to send per slave in balance-rr "
190 "mode; 0 for a random slave, 1 packet per "
191 "slave (default), >1 packets per slave.");
184 192
185/*----------------------------- Global variables ----------------------------*/ 193/*----------------------------- Global variables ----------------------------*/
186 194
@@ -217,6 +225,8 @@ const struct bond_parm_tbl xmit_hashtype_tbl[] = {
217{ "layer2", BOND_XMIT_POLICY_LAYER2}, 225{ "layer2", BOND_XMIT_POLICY_LAYER2},
218{ "layer3+4", BOND_XMIT_POLICY_LAYER34}, 226{ "layer3+4", BOND_XMIT_POLICY_LAYER34},
219{ "layer2+3", BOND_XMIT_POLICY_LAYER23}, 227{ "layer2+3", BOND_XMIT_POLICY_LAYER23},
228{ "encap2+3", BOND_XMIT_POLICY_ENCAP23},
229{ "encap3+4", BOND_XMIT_POLICY_ENCAP34},
220{ NULL, -1}, 230{ NULL, -1},
221}; 231};
222 232
@@ -332,10 +342,11 @@ static int bond_vlan_rx_add_vid(struct net_device *bond_dev,
332 __be16 proto, u16 vid) 342 __be16 proto, u16 vid)
333{ 343{
334 struct bonding *bond = netdev_priv(bond_dev); 344 struct bonding *bond = netdev_priv(bond_dev);
335 struct slave *slave; 345 struct slave *slave, *rollback_slave;
346 struct list_head *iter;
336 int res; 347 int res;
337 348
338 bond_for_each_slave(bond, slave) { 349 bond_for_each_slave(bond, slave, iter) {
339 res = vlan_vid_add(slave->dev, proto, vid); 350 res = vlan_vid_add(slave->dev, proto, vid);
340 if (res) 351 if (res)
341 goto unwind; 352 goto unwind;
@@ -344,9 +355,13 @@ static int bond_vlan_rx_add_vid(struct net_device *bond_dev,
344 return 0; 355 return 0;
345 356
346unwind: 357unwind:
347 /* unwind from the slave that failed */ 358 /* unwind to the slave that failed */
348 bond_for_each_slave_continue_reverse(bond, slave) 359 bond_for_each_slave(bond, rollback_slave, iter) {
349 vlan_vid_del(slave->dev, proto, vid); 360 if (rollback_slave == slave)
361 break;
362
363 vlan_vid_del(rollback_slave->dev, proto, vid);
364 }
350 365
351 return res; 366 return res;
352} 367}
@@ -360,9 +375,10 @@ static int bond_vlan_rx_kill_vid(struct net_device *bond_dev,
360 __be16 proto, u16 vid) 375 __be16 proto, u16 vid)
361{ 376{
362 struct bonding *bond = netdev_priv(bond_dev); 377 struct bonding *bond = netdev_priv(bond_dev);
378 struct list_head *iter;
363 struct slave *slave; 379 struct slave *slave;
364 380
365 bond_for_each_slave(bond, slave) 381 bond_for_each_slave(bond, slave, iter)
366 vlan_vid_del(slave->dev, proto, vid); 382 vlan_vid_del(slave->dev, proto, vid);
367 383
368 if (bond_is_lb(bond)) 384 if (bond_is_lb(bond))
@@ -382,15 +398,16 @@ static int bond_vlan_rx_kill_vid(struct net_device *bond_dev,
382 */ 398 */
383static int bond_set_carrier(struct bonding *bond) 399static int bond_set_carrier(struct bonding *bond)
384{ 400{
401 struct list_head *iter;
385 struct slave *slave; 402 struct slave *slave;
386 403
387 if (list_empty(&bond->slave_list)) 404 if (!bond_has_slaves(bond))
388 goto down; 405 goto down;
389 406
390 if (bond->params.mode == BOND_MODE_8023AD) 407 if (bond->params.mode == BOND_MODE_8023AD)
391 return bond_3ad_set_carrier(bond); 408 return bond_3ad_set_carrier(bond);
392 409
393 bond_for_each_slave(bond, slave) { 410 bond_for_each_slave(bond, slave, iter) {
394 if (slave->link == BOND_LINK_UP) { 411 if (slave->link == BOND_LINK_UP) {
395 if (!netif_carrier_ok(bond->dev)) { 412 if (!netif_carrier_ok(bond->dev)) {
396 netif_carrier_on(bond->dev); 413 netif_carrier_on(bond->dev);
@@ -522,7 +539,9 @@ static int bond_check_dev_link(struct bonding *bond,
522 */ 539 */
523static int bond_set_promiscuity(struct bonding *bond, int inc) 540static int bond_set_promiscuity(struct bonding *bond, int inc)
524{ 541{
542 struct list_head *iter;
525 int err = 0; 543 int err = 0;
544
526 if (USES_PRIMARY(bond->params.mode)) { 545 if (USES_PRIMARY(bond->params.mode)) {
527 /* write lock already acquired */ 546 /* write lock already acquired */
528 if (bond->curr_active_slave) { 547 if (bond->curr_active_slave) {
@@ -532,7 +551,7 @@ static int bond_set_promiscuity(struct bonding *bond, int inc)
532 } else { 551 } else {
533 struct slave *slave; 552 struct slave *slave;
534 553
535 bond_for_each_slave(bond, slave) { 554 bond_for_each_slave(bond, slave, iter) {
536 err = dev_set_promiscuity(slave->dev, inc); 555 err = dev_set_promiscuity(slave->dev, inc);
537 if (err) 556 if (err)
538 return err; 557 return err;
@@ -546,7 +565,9 @@ static int bond_set_promiscuity(struct bonding *bond, int inc)
546 */ 565 */
547static int bond_set_allmulti(struct bonding *bond, int inc) 566static int bond_set_allmulti(struct bonding *bond, int inc)
548{ 567{
568 struct list_head *iter;
549 int err = 0; 569 int err = 0;
570
550 if (USES_PRIMARY(bond->params.mode)) { 571 if (USES_PRIMARY(bond->params.mode)) {
551 /* write lock already acquired */ 572 /* write lock already acquired */
552 if (bond->curr_active_slave) { 573 if (bond->curr_active_slave) {
@@ -556,7 +577,7 @@ static int bond_set_allmulti(struct bonding *bond, int inc)
556 } else { 577 } else {
557 struct slave *slave; 578 struct slave *slave;
558 579
559 bond_for_each_slave(bond, slave) { 580 bond_for_each_slave(bond, slave, iter) {
560 err = dev_set_allmulti(slave->dev, inc); 581 err = dev_set_allmulti(slave->dev, inc);
561 if (err) 582 if (err)
562 return err; 583 return err;
@@ -774,43 +795,24 @@ static bool bond_should_change_active(struct bonding *bond)
774/** 795/**
775 * find_best_interface - select the best available slave to be the active one 796 * find_best_interface - select the best available slave to be the active one
776 * @bond: our bonding struct 797 * @bond: our bonding struct
777 *
778 * Warning: Caller must hold curr_slave_lock for writing.
779 */ 798 */
780static struct slave *bond_find_best_slave(struct bonding *bond) 799static struct slave *bond_find_best_slave(struct bonding *bond)
781{ 800{
782 struct slave *new_active, *old_active; 801 struct slave *slave, *bestslave = NULL;
783 struct slave *bestslave = NULL; 802 struct list_head *iter;
784 int mintime = bond->params.updelay; 803 int mintime = bond->params.updelay;
785 int i;
786 804
787 new_active = bond->curr_active_slave; 805 if (bond->primary_slave && bond->primary_slave->link == BOND_LINK_UP &&
788 806 bond_should_change_active(bond))
789 if (!new_active) { /* there were no active slaves left */ 807 return bond->primary_slave;
790 new_active = bond_first_slave(bond); 808
791 if (!new_active) 809 bond_for_each_slave(bond, slave, iter) {
792 return NULL; /* still no slave, return NULL */ 810 if (slave->link == BOND_LINK_UP)
793 } 811 return slave;
794 812 if (slave->link == BOND_LINK_BACK && IS_UP(slave->dev) &&
795 if ((bond->primary_slave) && 813 slave->delay < mintime) {
796 bond->primary_slave->link == BOND_LINK_UP && 814 mintime = slave->delay;
797 bond_should_change_active(bond)) { 815 bestslave = slave;
798 new_active = bond->primary_slave;
799 }
800
801 /* remember where to stop iterating over the slaves */
802 old_active = new_active;
803
804 bond_for_each_slave_from(bond, new_active, i, old_active) {
805 if (new_active->link == BOND_LINK_UP) {
806 return new_active;
807 } else if (new_active->link == BOND_LINK_BACK &&
808 IS_UP(new_active->dev)) {
809 /* link up, but waiting for stabilization */
810 if (new_active->delay < mintime) {
811 mintime = new_active->delay;
812 bestslave = new_active;
813 }
814 } 816 }
815 } 817 }
816 818
@@ -971,35 +973,6 @@ void bond_select_active_slave(struct bonding *bond)
971 } 973 }
972} 974}
973 975
974/*--------------------------- slave list handling ---------------------------*/
975
976/*
977 * This function attaches the slave to the end of list.
978 *
979 * bond->lock held for writing by caller.
980 */
981static void bond_attach_slave(struct bonding *bond, struct slave *new_slave)
982{
983 list_add_tail_rcu(&new_slave->list, &bond->slave_list);
984 bond->slave_cnt++;
985}
986
987/*
988 * This function detaches the slave from the list.
989 * WARNING: no check is made to verify if the slave effectively
990 * belongs to <bond>.
991 * Nothing is freed on return, structures are just unchained.
992 * If any slave pointer in bond was pointing to <slave>,
993 * it should be changed by the calling function.
994 *
995 * bond->lock held for writing by caller.
996 */
997static void bond_detach_slave(struct bonding *bond, struct slave *slave)
998{
999 list_del_rcu(&slave->list);
1000 bond->slave_cnt--;
1001}
1002
1003#ifdef CONFIG_NET_POLL_CONTROLLER 976#ifdef CONFIG_NET_POLL_CONTROLLER
1004static inline int slave_enable_netpoll(struct slave *slave) 977static inline int slave_enable_netpoll(struct slave *slave)
1005{ 978{
@@ -1046,9 +1019,10 @@ static void bond_poll_controller(struct net_device *bond_dev)
1046static void bond_netpoll_cleanup(struct net_device *bond_dev) 1019static void bond_netpoll_cleanup(struct net_device *bond_dev)
1047{ 1020{
1048 struct bonding *bond = netdev_priv(bond_dev); 1021 struct bonding *bond = netdev_priv(bond_dev);
1022 struct list_head *iter;
1049 struct slave *slave; 1023 struct slave *slave;
1050 1024
1051 bond_for_each_slave(bond, slave) 1025 bond_for_each_slave(bond, slave, iter)
1052 if (IS_UP(slave->dev)) 1026 if (IS_UP(slave->dev))
1053 slave_disable_netpoll(slave); 1027 slave_disable_netpoll(slave);
1054} 1028}
@@ -1056,10 +1030,11 @@ static void bond_netpoll_cleanup(struct net_device *bond_dev)
1056static int bond_netpoll_setup(struct net_device *dev, struct netpoll_info *ni, gfp_t gfp) 1030static int bond_netpoll_setup(struct net_device *dev, struct netpoll_info *ni, gfp_t gfp)
1057{ 1031{
1058 struct bonding *bond = netdev_priv(dev); 1032 struct bonding *bond = netdev_priv(dev);
1033 struct list_head *iter;
1059 struct slave *slave; 1034 struct slave *slave;
1060 int err = 0; 1035 int err = 0;
1061 1036
1062 bond_for_each_slave(bond, slave) { 1037 bond_for_each_slave(bond, slave, iter) {
1063 err = slave_enable_netpoll(slave); 1038 err = slave_enable_netpoll(slave);
1064 if (err) { 1039 if (err) {
1065 bond_netpoll_cleanup(dev); 1040 bond_netpoll_cleanup(dev);
@@ -1087,10 +1062,11 @@ static netdev_features_t bond_fix_features(struct net_device *dev,
1087 netdev_features_t features) 1062 netdev_features_t features)
1088{ 1063{
1089 struct bonding *bond = netdev_priv(dev); 1064 struct bonding *bond = netdev_priv(dev);
1065 struct list_head *iter;
1090 netdev_features_t mask; 1066 netdev_features_t mask;
1091 struct slave *slave; 1067 struct slave *slave;
1092 1068
1093 if (list_empty(&bond->slave_list)) { 1069 if (!bond_has_slaves(bond)) {
1094 /* Disable adding VLANs to empty bond. But why? --mq */ 1070 /* Disable adding VLANs to empty bond. But why? --mq */
1095 features |= NETIF_F_VLAN_CHALLENGED; 1071 features |= NETIF_F_VLAN_CHALLENGED;
1096 return features; 1072 return features;
@@ -1100,7 +1076,7 @@ static netdev_features_t bond_fix_features(struct net_device *dev,
1100 features &= ~NETIF_F_ONE_FOR_ALL; 1076 features &= ~NETIF_F_ONE_FOR_ALL;
1101 features |= NETIF_F_ALL_FOR_ALL; 1077 features |= NETIF_F_ALL_FOR_ALL;
1102 1078
1103 bond_for_each_slave(bond, slave) { 1079 bond_for_each_slave(bond, slave, iter) {
1104 features = netdev_increment_features(features, 1080 features = netdev_increment_features(features,
1105 slave->dev->features, 1081 slave->dev->features,
1106 mask); 1082 mask);
@@ -1118,16 +1094,17 @@ static void bond_compute_features(struct bonding *bond)
1118{ 1094{
1119 unsigned int flags, dst_release_flag = IFF_XMIT_DST_RELEASE; 1095 unsigned int flags, dst_release_flag = IFF_XMIT_DST_RELEASE;
1120 netdev_features_t vlan_features = BOND_VLAN_FEATURES; 1096 netdev_features_t vlan_features = BOND_VLAN_FEATURES;
1097 struct net_device *bond_dev = bond->dev;
1098 struct list_head *iter;
1099 struct slave *slave;
1121 unsigned short max_hard_header_len = ETH_HLEN; 1100 unsigned short max_hard_header_len = ETH_HLEN;
1122 unsigned int gso_max_size = GSO_MAX_SIZE; 1101 unsigned int gso_max_size = GSO_MAX_SIZE;
1123 struct net_device *bond_dev = bond->dev;
1124 u16 gso_max_segs = GSO_MAX_SEGS; 1102 u16 gso_max_segs = GSO_MAX_SEGS;
1125 struct slave *slave;
1126 1103
1127 if (list_empty(&bond->slave_list)) 1104 if (!bond_has_slaves(bond))
1128 goto done; 1105 goto done;
1129 1106
1130 bond_for_each_slave(bond, slave) { 1107 bond_for_each_slave(bond, slave, iter) {
1131 vlan_features = netdev_increment_features(vlan_features, 1108 vlan_features = netdev_increment_features(vlan_features,
1132 slave->dev->vlan_features, BOND_VLAN_FEATURES); 1109 slave->dev->vlan_features, BOND_VLAN_FEATURES);
1133 1110
@@ -1233,15 +1210,16 @@ static rx_handler_result_t bond_handle_frame(struct sk_buff **pskb)
1233} 1210}
1234 1211
1235static int bond_master_upper_dev_link(struct net_device *bond_dev, 1212static int bond_master_upper_dev_link(struct net_device *bond_dev,
1236 struct net_device *slave_dev) 1213 struct net_device *slave_dev,
1214 struct slave *slave)
1237{ 1215{
1238 int err; 1216 int err;
1239 1217
1240 err = netdev_master_upper_dev_link(slave_dev, bond_dev); 1218 err = netdev_master_upper_dev_link_private(slave_dev, bond_dev, slave);
1241 if (err) 1219 if (err)
1242 return err; 1220 return err;
1243 slave_dev->flags |= IFF_SLAVE; 1221 slave_dev->flags |= IFF_SLAVE;
1244 rtmsg_ifinfo(RTM_NEWLINK, slave_dev, IFF_SLAVE); 1222 rtmsg_ifinfo(RTM_NEWLINK, slave_dev, IFF_SLAVE, GFP_KERNEL);
1245 return 0; 1223 return 0;
1246} 1224}
1247 1225
@@ -1250,7 +1228,7 @@ static void bond_upper_dev_unlink(struct net_device *bond_dev,
1250{ 1228{
1251 netdev_upper_dev_unlink(slave_dev, bond_dev); 1229 netdev_upper_dev_unlink(slave_dev, bond_dev);
1252 slave_dev->flags &= ~IFF_SLAVE; 1230 slave_dev->flags &= ~IFF_SLAVE;
1253 rtmsg_ifinfo(RTM_NEWLINK, slave_dev, IFF_SLAVE); 1231 rtmsg_ifinfo(RTM_NEWLINK, slave_dev, IFF_SLAVE, GFP_KERNEL);
1254} 1232}
1255 1233
1256/* enslave device <slave> to bond device <master> */ 1234/* enslave device <slave> to bond device <master> */
@@ -1258,7 +1236,7 @@ int bond_enslave(struct net_device *bond_dev, struct net_device *slave_dev)
1258{ 1236{
1259 struct bonding *bond = netdev_priv(bond_dev); 1237 struct bonding *bond = netdev_priv(bond_dev);
1260 const struct net_device_ops *slave_ops = slave_dev->netdev_ops; 1238 const struct net_device_ops *slave_ops = slave_dev->netdev_ops;
1261 struct slave *new_slave = NULL; 1239 struct slave *new_slave = NULL, *prev_slave;
1262 struct sockaddr addr; 1240 struct sockaddr addr;
1263 int link_reporting; 1241 int link_reporting;
1264 int res = 0, i; 1242 int res = 0, i;
@@ -1313,7 +1291,7 @@ int bond_enslave(struct net_device *bond_dev, struct net_device *slave_dev)
1313 * bond ether type mutual exclusion - don't allow slaves of dissimilar 1291 * bond ether type mutual exclusion - don't allow slaves of dissimilar
1314 * ether type (eg ARPHRD_ETHER and ARPHRD_INFINIBAND) share the same bond 1292 * ether type (eg ARPHRD_ETHER and ARPHRD_INFINIBAND) share the same bond
1315 */ 1293 */
1316 if (list_empty(&bond->slave_list)) { 1294 if (!bond_has_slaves(bond)) {
1317 if (bond_dev->type != slave_dev->type) { 1295 if (bond_dev->type != slave_dev->type) {
1318 pr_debug("%s: change device type from %d to %d\n", 1296 pr_debug("%s: change device type from %d to %d\n",
1319 bond_dev->name, 1297 bond_dev->name,
@@ -1352,7 +1330,7 @@ int bond_enslave(struct net_device *bond_dev, struct net_device *slave_dev)
1352 } 1330 }
1353 1331
1354 if (slave_ops->ndo_set_mac_address == NULL) { 1332 if (slave_ops->ndo_set_mac_address == NULL) {
1355 if (list_empty(&bond->slave_list)) { 1333 if (!bond_has_slaves(bond)) {
1356 pr_warning("%s: Warning: The first slave device specified does not support setting the MAC address. Setting fail_over_mac to active.", 1334 pr_warning("%s: Warning: The first slave device specified does not support setting the MAC address. Setting fail_over_mac to active.",
1357 bond_dev->name); 1335 bond_dev->name);
1358 bond->params.fail_over_mac = BOND_FOM_ACTIVE; 1336 bond->params.fail_over_mac = BOND_FOM_ACTIVE;
@@ -1368,7 +1346,7 @@ int bond_enslave(struct net_device *bond_dev, struct net_device *slave_dev)
1368 1346
1369 /* If this is the first slave, then we need to set the master's hardware 1347 /* If this is the first slave, then we need to set the master's hardware
1370 * address to be the same as the slave's. */ 1348 * address to be the same as the slave's. */
1371 if (list_empty(&bond->slave_list) && 1349 if (!bond_has_slaves(bond) &&
1372 bond->dev->addr_assign_type == NET_ADDR_RANDOM) 1350 bond->dev->addr_assign_type == NET_ADDR_RANDOM)
1373 bond_set_dev_addr(bond->dev, slave_dev); 1351 bond_set_dev_addr(bond->dev, slave_dev);
1374 1352
@@ -1377,7 +1355,6 @@ int bond_enslave(struct net_device *bond_dev, struct net_device *slave_dev)
1377 res = -ENOMEM; 1355 res = -ENOMEM;
1378 goto err_undo_flags; 1356 goto err_undo_flags;
1379 } 1357 }
1380 INIT_LIST_HEAD(&new_slave->list);
1381 /* 1358 /*
1382 * Set the new_slave's queue_id to be zero. Queue ID mapping 1359 * Set the new_slave's queue_id to be zero. Queue ID mapping
1383 * is set via sysfs or module option if desired. 1360 * is set via sysfs or module option if desired.
@@ -1413,17 +1390,11 @@ int bond_enslave(struct net_device *bond_dev, struct net_device *slave_dev)
1413 } 1390 }
1414 } 1391 }
1415 1392
1416 res = bond_master_upper_dev_link(bond_dev, slave_dev);
1417 if (res) {
1418 pr_debug("Error %d calling bond_master_upper_dev_link\n", res);
1419 goto err_restore_mac;
1420 }
1421
1422 /* open the slave since the application closed it */ 1393 /* open the slave since the application closed it */
1423 res = dev_open(slave_dev); 1394 res = dev_open(slave_dev);
1424 if (res) { 1395 if (res) {
1425 pr_debug("Opening slave %s failed\n", slave_dev->name); 1396 pr_debug("Opening slave %s failed\n", slave_dev->name);
1426 goto err_unset_master; 1397 goto err_restore_mac;
1427 } 1398 }
1428 1399
1429 new_slave->bond = bond; 1400 new_slave->bond = bond;
@@ -1479,21 +1450,13 @@ int bond_enslave(struct net_device *bond_dev, struct net_device *slave_dev)
1479 goto err_close; 1450 goto err_close;
1480 } 1451 }
1481 1452
1482 write_lock_bh(&bond->lock); 1453 prev_slave = bond_last_slave(bond);
1483
1484 bond_attach_slave(bond, new_slave);
1485 1454
1486 new_slave->delay = 0; 1455 new_slave->delay = 0;
1487 new_slave->link_failure_count = 0; 1456 new_slave->link_failure_count = 0;
1488 1457
1489 write_unlock_bh(&bond->lock);
1490
1491 bond_compute_features(bond);
1492
1493 bond_update_speed_duplex(new_slave); 1458 bond_update_speed_duplex(new_slave);
1494 1459
1495 read_lock(&bond->lock);
1496
1497 new_slave->last_arp_rx = jiffies - 1460 new_slave->last_arp_rx = jiffies -
1498 (msecs_to_jiffies(bond->params.arp_interval) + 1); 1461 (msecs_to_jiffies(bond->params.arp_interval) + 1);
1499 for (i = 0; i < BOND_MAX_ARP_TARGETS; i++) 1462 for (i = 0; i < BOND_MAX_ARP_TARGETS; i++)
@@ -1554,12 +1517,9 @@ int bond_enslave(struct net_device *bond_dev, struct net_device *slave_dev)
1554 } 1517 }
1555 } 1518 }
1556 1519
1557 write_lock_bh(&bond->curr_slave_lock);
1558
1559 switch (bond->params.mode) { 1520 switch (bond->params.mode) {
1560 case BOND_MODE_ACTIVEBACKUP: 1521 case BOND_MODE_ACTIVEBACKUP:
1561 bond_set_slave_inactive_flags(new_slave); 1522 bond_set_slave_inactive_flags(new_slave);
1562 bond_select_active_slave(bond);
1563 break; 1523 break;
1564 case BOND_MODE_8023AD: 1524 case BOND_MODE_8023AD:
1565 /* in 802.3ad mode, the internal mechanism 1525 /* in 802.3ad mode, the internal mechanism
@@ -1568,16 +1528,13 @@ int bond_enslave(struct net_device *bond_dev, struct net_device *slave_dev)
1568 */ 1528 */
1569 bond_set_slave_inactive_flags(new_slave); 1529 bond_set_slave_inactive_flags(new_slave);
1570 /* if this is the first slave */ 1530 /* if this is the first slave */
1571 if (bond_first_slave(bond) == new_slave) { 1531 if (!prev_slave) {
1572 SLAVE_AD_INFO(new_slave).id = 1; 1532 SLAVE_AD_INFO(new_slave).id = 1;
1573 /* Initialize AD with the number of times that the AD timer is called in 1 second 1533 /* Initialize AD with the number of times that the AD timer is called in 1 second
1574 * can be called only after the mac address of the bond is set 1534 * can be called only after the mac address of the bond is set
1575 */ 1535 */
1576 bond_3ad_initialize(bond, 1000/AD_TIMER_INTERVAL); 1536 bond_3ad_initialize(bond, 1000/AD_TIMER_INTERVAL);
1577 } else { 1537 } else {
1578 struct slave *prev_slave;
1579
1580 prev_slave = bond_prev_slave(bond, new_slave);
1581 SLAVE_AD_INFO(new_slave).id = 1538 SLAVE_AD_INFO(new_slave).id =
1582 SLAVE_AD_INFO(prev_slave).id + 1; 1539 SLAVE_AD_INFO(prev_slave).id + 1;
1583 } 1540 }
@@ -1588,7 +1545,6 @@ int bond_enslave(struct net_device *bond_dev, struct net_device *slave_dev)
1588 case BOND_MODE_ALB: 1545 case BOND_MODE_ALB:
1589 bond_set_active_slave(new_slave); 1546 bond_set_active_slave(new_slave);
1590 bond_set_slave_inactive_flags(new_slave); 1547 bond_set_slave_inactive_flags(new_slave);
1591 bond_select_active_slave(bond);
1592 break; 1548 break;
1593 default: 1549 default:
1594 pr_debug("This slave is always active in trunk mode\n"); 1550 pr_debug("This slave is always active in trunk mode\n");
@@ -1606,10 +1562,6 @@ int bond_enslave(struct net_device *bond_dev, struct net_device *slave_dev)
1606 break; 1562 break;
1607 } /* switch(bond_mode) */ 1563 } /* switch(bond_mode) */
1608 1564
1609 write_unlock_bh(&bond->curr_slave_lock);
1610
1611 bond_set_carrier(bond);
1612
1613#ifdef CONFIG_NET_POLL_CONTROLLER 1565#ifdef CONFIG_NET_POLL_CONTROLLER
1614 slave_dev->npinfo = bond->dev->npinfo; 1566 slave_dev->npinfo = bond->dev->npinfo;
1615 if (slave_dev->npinfo) { 1567 if (slave_dev->npinfo) {
@@ -1624,17 +1576,29 @@ int bond_enslave(struct net_device *bond_dev, struct net_device *slave_dev)
1624 } 1576 }
1625#endif 1577#endif
1626 1578
1627 read_unlock(&bond->lock);
1628
1629 res = bond_create_slave_symlinks(bond_dev, slave_dev);
1630 if (res)
1631 goto err_detach;
1632
1633 res = netdev_rx_handler_register(slave_dev, bond_handle_frame, 1579 res = netdev_rx_handler_register(slave_dev, bond_handle_frame,
1634 new_slave); 1580 new_slave);
1635 if (res) { 1581 if (res) {
1636 pr_debug("Error %d calling netdev_rx_handler_register\n", res); 1582 pr_debug("Error %d calling netdev_rx_handler_register\n", res);
1637 goto err_dest_symlinks; 1583 goto err_detach;
1584 }
1585
1586 res = bond_master_upper_dev_link(bond_dev, slave_dev, new_slave);
1587 if (res) {
1588 pr_debug("Error %d calling bond_master_upper_dev_link\n", res);
1589 goto err_unregister;
1590 }
1591
1592 bond->slave_cnt++;
1593 bond_compute_features(bond);
1594 bond_set_carrier(bond);
1595
1596 if (USES_PRIMARY(bond->params.mode)) {
1597 read_lock(&bond->lock);
1598 write_lock_bh(&bond->curr_slave_lock);
1599 bond_select_active_slave(bond);
1600 write_unlock_bh(&bond->curr_slave_lock);
1601 read_unlock(&bond->lock);
1638 } 1602 }
1639 1603
1640 pr_info("%s: enslaving %s as a%s interface with a%s link.\n", 1604 pr_info("%s: enslaving %s as a%s interface with a%s link.\n",
@@ -1646,8 +1610,8 @@ int bond_enslave(struct net_device *bond_dev, struct net_device *slave_dev)
1646 return 0; 1610 return 0;
1647 1611
1648/* Undo stages on error */ 1612/* Undo stages on error */
1649err_dest_symlinks: 1613err_unregister:
1650 bond_destroy_slave_symlinks(bond_dev, slave_dev); 1614 netdev_rx_handler_unregister(slave_dev);
1651 1615
1652err_detach: 1616err_detach:
1653 if (!USES_PRIMARY(bond->params.mode)) 1617 if (!USES_PRIMARY(bond->params.mode))
@@ -1655,7 +1619,6 @@ err_detach:
1655 1619
1656 vlan_vids_del_by_dev(slave_dev, bond_dev); 1620 vlan_vids_del_by_dev(slave_dev, bond_dev);
1657 write_lock_bh(&bond->lock); 1621 write_lock_bh(&bond->lock);
1658 bond_detach_slave(bond, new_slave);
1659 if (bond->primary_slave == new_slave) 1622 if (bond->primary_slave == new_slave)
1660 bond->primary_slave = NULL; 1623 bond->primary_slave = NULL;
1661 if (bond->curr_active_slave == new_slave) { 1624 if (bond->curr_active_slave == new_slave) {
@@ -1675,9 +1638,6 @@ err_close:
1675 slave_dev->priv_flags &= ~IFF_BONDING; 1638 slave_dev->priv_flags &= ~IFF_BONDING;
1676 dev_close(slave_dev); 1639 dev_close(slave_dev);
1677 1640
1678err_unset_master:
1679 bond_upper_dev_unlink(bond_dev, slave_dev);
1680
1681err_restore_mac: 1641err_restore_mac:
1682 if (!bond->params.fail_over_mac) { 1642 if (!bond->params.fail_over_mac) {
1683 /* XXX TODO - fom follow mode needs to change master's 1643 /* XXX TODO - fom follow mode needs to change master's
@@ -1696,9 +1656,8 @@ err_free:
1696 kfree(new_slave); 1656 kfree(new_slave);
1697 1657
1698err_undo_flags: 1658err_undo_flags:
1699 bond_compute_features(bond);
1700 /* Enslave of first slave has failed and we need to fix master's mac */ 1659 /* Enslave of first slave has failed and we need to fix master's mac */
1701 if (list_empty(&bond->slave_list) && 1660 if (!bond_has_slaves(bond) &&
1702 ether_addr_equal(bond_dev->dev_addr, slave_dev->dev_addr)) 1661 ether_addr_equal(bond_dev->dev_addr, slave_dev->dev_addr))
1703 eth_hw_addr_random(bond_dev); 1662 eth_hw_addr_random(bond_dev);
1704 1663
@@ -1749,6 +1708,11 @@ static int __bond_release_one(struct net_device *bond_dev,
1749 } 1708 }
1750 1709
1751 write_unlock_bh(&bond->lock); 1710 write_unlock_bh(&bond->lock);
1711
1712 /* release the slave from its bond */
1713 bond->slave_cnt--;
1714
1715 bond_upper_dev_unlink(bond_dev, slave_dev);
1752 /* unregister rx_handler early so bond_handle_frame wouldn't be called 1716 /* unregister rx_handler early so bond_handle_frame wouldn't be called
1753 * for this slave anymore. 1717 * for this slave anymore.
1754 */ 1718 */
@@ -1772,12 +1736,9 @@ static int __bond_release_one(struct net_device *bond_dev,
1772 1736
1773 bond->current_arp_slave = NULL; 1737 bond->current_arp_slave = NULL;
1774 1738
1775 /* release the slave from its bond */
1776 bond_detach_slave(bond, slave);
1777
1778 if (!all && !bond->params.fail_over_mac) { 1739 if (!all && !bond->params.fail_over_mac) {
1779 if (ether_addr_equal(bond_dev->dev_addr, slave->perm_hwaddr) && 1740 if (ether_addr_equal(bond_dev->dev_addr, slave->perm_hwaddr) &&
1780 !list_empty(&bond->slave_list)) 1741 bond_has_slaves(bond))
1781 pr_warn("%s: Warning: the permanent HWaddr of %s - %pM - is still in use by %s. Set the HWaddr of %s to a different address to avoid conflicts.\n", 1742 pr_warn("%s: Warning: the permanent HWaddr of %s - %pM - is still in use by %s. Set the HWaddr of %s to a different address to avoid conflicts.\n",
1782 bond_dev->name, slave_dev->name, 1743 bond_dev->name, slave_dev->name,
1783 slave->perm_hwaddr, 1744 slave->perm_hwaddr,
@@ -1820,7 +1781,7 @@ static int __bond_release_one(struct net_device *bond_dev,
1820 write_lock_bh(&bond->lock); 1781 write_lock_bh(&bond->lock);
1821 } 1782 }
1822 1783
1823 if (list_empty(&bond->slave_list)) { 1784 if (!bond_has_slaves(bond)) {
1824 bond_set_carrier(bond); 1785 bond_set_carrier(bond);
1825 eth_hw_addr_random(bond_dev); 1786 eth_hw_addr_random(bond_dev);
1826 1787
@@ -1836,7 +1797,7 @@ static int __bond_release_one(struct net_device *bond_dev,
1836 unblock_netpoll_tx(); 1797 unblock_netpoll_tx();
1837 synchronize_rcu(); 1798 synchronize_rcu();
1838 1799
1839 if (list_empty(&bond->slave_list)) { 1800 if (!bond_has_slaves(bond)) {
1840 call_netdevice_notifiers(NETDEV_CHANGEADDR, bond->dev); 1801 call_netdevice_notifiers(NETDEV_CHANGEADDR, bond->dev);
1841 call_netdevice_notifiers(NETDEV_RELEASE, bond->dev); 1802 call_netdevice_notifiers(NETDEV_RELEASE, bond->dev);
1842 } 1803 }
@@ -1848,8 +1809,6 @@ static int __bond_release_one(struct net_device *bond_dev,
1848 bond_dev->name, slave_dev->name, bond_dev->name); 1809 bond_dev->name, slave_dev->name, bond_dev->name);
1849 1810
1850 /* must do this from outside any spinlocks */ 1811 /* must do this from outside any spinlocks */
1851 bond_destroy_slave_symlinks(bond_dev, slave_dev);
1852
1853 vlan_vids_del_by_dev(slave_dev, bond_dev); 1812 vlan_vids_del_by_dev(slave_dev, bond_dev);
1854 1813
1855 /* If the mode USES_PRIMARY, then this cases was handled above by 1814 /* If the mode USES_PRIMARY, then this cases was handled above by
@@ -1873,8 +1832,6 @@ static int __bond_release_one(struct net_device *bond_dev,
1873 bond_hw_addr_flush(bond_dev, slave_dev); 1832 bond_hw_addr_flush(bond_dev, slave_dev);
1874 } 1833 }
1875 1834
1876 bond_upper_dev_unlink(bond_dev, slave_dev);
1877
1878 slave_disable_netpoll(slave); 1835 slave_disable_netpoll(slave);
1879 1836
1880 /* close slave before restoring its mac address */ 1837 /* close slave before restoring its mac address */
@@ -1913,7 +1870,7 @@ static int bond_release_and_destroy(struct net_device *bond_dev,
1913 int ret; 1870 int ret;
1914 1871
1915 ret = bond_release(bond_dev, slave_dev); 1872 ret = bond_release(bond_dev, slave_dev);
1916 if (ret == 0 && list_empty(&bond->slave_list)) { 1873 if (ret == 0 && !bond_has_slaves(bond)) {
1917 bond_dev->priv_flags |= IFF_DISABLE_NETPOLL; 1874 bond_dev->priv_flags |= IFF_DISABLE_NETPOLL;
1918 pr_info("%s: destroying bond %s.\n", 1875 pr_info("%s: destroying bond %s.\n",
1919 bond_dev->name, bond_dev->name); 1876 bond_dev->name, bond_dev->name);
@@ -1922,61 +1879,6 @@ static int bond_release_and_destroy(struct net_device *bond_dev,
1922 return ret; 1879 return ret;
1923} 1880}
1924 1881
1925/*
1926 * This function changes the active slave to slave <slave_dev>.
1927 * It returns -EINVAL in the following cases.
1928 * - <slave_dev> is not found in the list.
1929 * - There is not active slave now.
1930 * - <slave_dev> is already active.
1931 * - The link state of <slave_dev> is not BOND_LINK_UP.
1932 * - <slave_dev> is not running.
1933 * In these cases, this function does nothing.
1934 * In the other cases, current_slave pointer is changed and 0 is returned.
1935 */
1936static int bond_ioctl_change_active(struct net_device *bond_dev, struct net_device *slave_dev)
1937{
1938 struct bonding *bond = netdev_priv(bond_dev);
1939 struct slave *old_active = NULL;
1940 struct slave *new_active = NULL;
1941 int res = 0;
1942
1943 if (!USES_PRIMARY(bond->params.mode))
1944 return -EINVAL;
1945
1946 /* Verify that bond_dev is indeed the master of slave_dev */
1947 if (!(slave_dev->flags & IFF_SLAVE) ||
1948 !netdev_has_upper_dev(slave_dev, bond_dev))
1949 return -EINVAL;
1950
1951 read_lock(&bond->lock);
1952
1953 old_active = bond->curr_active_slave;
1954 new_active = bond_get_slave_by_dev(bond, slave_dev);
1955 /*
1956 * Changing to the current active: do nothing; return success.
1957 */
1958 if (new_active && new_active == old_active) {
1959 read_unlock(&bond->lock);
1960 return 0;
1961 }
1962
1963 if (new_active &&
1964 old_active &&
1965 new_active->link == BOND_LINK_UP &&
1966 IS_UP(new_active->dev)) {
1967 block_netpoll_tx();
1968 write_lock_bh(&bond->curr_slave_lock);
1969 bond_change_active_slave(bond, new_active);
1970 write_unlock_bh(&bond->curr_slave_lock);
1971 unblock_netpoll_tx();
1972 } else
1973 res = -EINVAL;
1974
1975 read_unlock(&bond->lock);
1976
1977 return res;
1978}
1979
1980static int bond_info_query(struct net_device *bond_dev, struct ifbond *info) 1882static int bond_info_query(struct net_device *bond_dev, struct ifbond *info)
1981{ 1883{
1982 struct bonding *bond = netdev_priv(bond_dev); 1884 struct bonding *bond = netdev_priv(bond_dev);
@@ -1994,11 +1896,12 @@ static int bond_info_query(struct net_device *bond_dev, struct ifbond *info)
1994static int bond_slave_info_query(struct net_device *bond_dev, struct ifslave *info) 1896static int bond_slave_info_query(struct net_device *bond_dev, struct ifslave *info)
1995{ 1897{
1996 struct bonding *bond = netdev_priv(bond_dev); 1898 struct bonding *bond = netdev_priv(bond_dev);
1899 struct list_head *iter;
1997 int i = 0, res = -ENODEV; 1900 int i = 0, res = -ENODEV;
1998 struct slave *slave; 1901 struct slave *slave;
1999 1902
2000 read_lock(&bond->lock); 1903 read_lock(&bond->lock);
2001 bond_for_each_slave(bond, slave) { 1904 bond_for_each_slave(bond, slave, iter) {
2002 if (i++ == (int)info->slave_id) { 1905 if (i++ == (int)info->slave_id) {
2003 res = 0; 1906 res = 0;
2004 strcpy(info->slave_name, slave->dev->name); 1907 strcpy(info->slave_name, slave->dev->name);
@@ -2019,12 +1922,13 @@ static int bond_slave_info_query(struct net_device *bond_dev, struct ifslave *in
2019static int bond_miimon_inspect(struct bonding *bond) 1922static int bond_miimon_inspect(struct bonding *bond)
2020{ 1923{
2021 int link_state, commit = 0; 1924 int link_state, commit = 0;
1925 struct list_head *iter;
2022 struct slave *slave; 1926 struct slave *slave;
2023 bool ignore_updelay; 1927 bool ignore_updelay;
2024 1928
2025 ignore_updelay = !bond->curr_active_slave ? true : false; 1929 ignore_updelay = !bond->curr_active_slave ? true : false;
2026 1930
2027 bond_for_each_slave(bond, slave) { 1931 bond_for_each_slave(bond, slave, iter) {
2028 slave->new_link = BOND_LINK_NOCHANGE; 1932 slave->new_link = BOND_LINK_NOCHANGE;
2029 1933
2030 link_state = bond_check_dev_link(bond, slave->dev, 0); 1934 link_state = bond_check_dev_link(bond, slave->dev, 0);
@@ -2118,9 +2022,10 @@ static int bond_miimon_inspect(struct bonding *bond)
2118 2022
2119static void bond_miimon_commit(struct bonding *bond) 2023static void bond_miimon_commit(struct bonding *bond)
2120{ 2024{
2025 struct list_head *iter;
2121 struct slave *slave; 2026 struct slave *slave;
2122 2027
2123 bond_for_each_slave(bond, slave) { 2028 bond_for_each_slave(bond, slave, iter) {
2124 switch (slave->new_link) { 2029 switch (slave->new_link) {
2125 case BOND_LINK_NOCHANGE: 2030 case BOND_LINK_NOCHANGE:
2126 continue; 2031 continue;
@@ -2225,7 +2130,7 @@ void bond_mii_monitor(struct work_struct *work)
2225 2130
2226 delay = msecs_to_jiffies(bond->params.miimon); 2131 delay = msecs_to_jiffies(bond->params.miimon);
2227 2132
2228 if (list_empty(&bond->slave_list)) 2133 if (!bond_has_slaves(bond))
2229 goto re_arm; 2134 goto re_arm;
2230 2135
2231 should_notify_peers = bond_should_notify_peers(bond); 2136 should_notify_peers = bond_should_notify_peers(bond);
@@ -2274,7 +2179,7 @@ static bool bond_has_this_ip(struct bonding *bond, __be32 ip)
2274 return true; 2179 return true;
2275 2180
2276 rcu_read_lock(); 2181 rcu_read_lock();
2277 netdev_for_each_upper_dev_rcu(bond->dev, upper, iter) { 2182 netdev_for_each_all_upper_dev_rcu(bond->dev, upper, iter) {
2278 if (ip == bond_confirm_addr(upper, 0, ip)) { 2183 if (ip == bond_confirm_addr(upper, 0, ip)) {
2279 ret = true; 2184 ret = true;
2280 break; 2185 break;
@@ -2349,10 +2254,12 @@ static void bond_arp_send_all(struct bonding *bond, struct slave *slave)
2349 * 2254 *
2350 * TODO: QinQ? 2255 * TODO: QinQ?
2351 */ 2256 */
2352 netdev_for_each_upper_dev_rcu(bond->dev, vlan_upper, vlan_iter) { 2257 netdev_for_each_all_upper_dev_rcu(bond->dev, vlan_upper,
2258 vlan_iter) {
2353 if (!is_vlan_dev(vlan_upper)) 2259 if (!is_vlan_dev(vlan_upper))
2354 continue; 2260 continue;
2355 netdev_for_each_upper_dev_rcu(vlan_upper, upper, iter) { 2261 netdev_for_each_all_upper_dev_rcu(vlan_upper, upper,
2262 iter) {
2356 if (upper == rt->dst.dev) { 2263 if (upper == rt->dst.dev) {
2357 vlan_id = vlan_dev_vlan_id(vlan_upper); 2264 vlan_id = vlan_dev_vlan_id(vlan_upper);
2358 rcu_read_unlock(); 2265 rcu_read_unlock();
@@ -2365,7 +2272,7 @@ static void bond_arp_send_all(struct bonding *bond, struct slave *slave)
2365 * our upper vlans, then just search for any dev that 2272 * our upper vlans, then just search for any dev that
2366 * matches, and in case it's a vlan - save the id 2273 * matches, and in case it's a vlan - save the id
2367 */ 2274 */
2368 netdev_for_each_upper_dev_rcu(bond->dev, upper, iter) { 2275 netdev_for_each_all_upper_dev_rcu(bond->dev, upper, iter) {
2369 if (upper == rt->dst.dev) { 2276 if (upper == rt->dst.dev) {
2370 /* if it's a vlan - get its VID */ 2277 /* if it's a vlan - get its VID */
2371 if (is_vlan_dev(upper)) 2278 if (is_vlan_dev(upper))
@@ -2512,11 +2419,12 @@ void bond_loadbalance_arp_mon(struct work_struct *work)
2512 struct bonding *bond = container_of(work, struct bonding, 2419 struct bonding *bond = container_of(work, struct bonding,
2513 arp_work.work); 2420 arp_work.work);
2514 struct slave *slave, *oldcurrent; 2421 struct slave *slave, *oldcurrent;
2422 struct list_head *iter;
2515 int do_failover = 0; 2423 int do_failover = 0;
2516 2424
2517 read_lock(&bond->lock); 2425 read_lock(&bond->lock);
2518 2426
2519 if (list_empty(&bond->slave_list)) 2427 if (!bond_has_slaves(bond))
2520 goto re_arm; 2428 goto re_arm;
2521 2429
2522 oldcurrent = bond->curr_active_slave; 2430 oldcurrent = bond->curr_active_slave;
@@ -2528,7 +2436,7 @@ void bond_loadbalance_arp_mon(struct work_struct *work)
2528 * TODO: what about up/down delay in arp mode? it wasn't here before 2436 * TODO: what about up/down delay in arp mode? it wasn't here before
2529 * so it can wait 2437 * so it can wait
2530 */ 2438 */
2531 bond_for_each_slave(bond, slave) { 2439 bond_for_each_slave(bond, slave, iter) {
2532 unsigned long trans_start = dev_trans_start(slave->dev); 2440 unsigned long trans_start = dev_trans_start(slave->dev);
2533 2441
2534 if (slave->link != BOND_LINK_UP) { 2442 if (slave->link != BOND_LINK_UP) {
@@ -2619,10 +2527,11 @@ re_arm:
2619static int bond_ab_arp_inspect(struct bonding *bond) 2527static int bond_ab_arp_inspect(struct bonding *bond)
2620{ 2528{
2621 unsigned long trans_start, last_rx; 2529 unsigned long trans_start, last_rx;
2530 struct list_head *iter;
2622 struct slave *slave; 2531 struct slave *slave;
2623 int commit = 0; 2532 int commit = 0;
2624 2533
2625 bond_for_each_slave(bond, slave) { 2534 bond_for_each_slave(bond, slave, iter) {
2626 slave->new_link = BOND_LINK_NOCHANGE; 2535 slave->new_link = BOND_LINK_NOCHANGE;
2627 last_rx = slave_last_rx(bond, slave); 2536 last_rx = slave_last_rx(bond, slave);
2628 2537
@@ -2689,9 +2598,10 @@ static int bond_ab_arp_inspect(struct bonding *bond)
2689static void bond_ab_arp_commit(struct bonding *bond) 2598static void bond_ab_arp_commit(struct bonding *bond)
2690{ 2599{
2691 unsigned long trans_start; 2600 unsigned long trans_start;
2601 struct list_head *iter;
2692 struct slave *slave; 2602 struct slave *slave;
2693 2603
2694 bond_for_each_slave(bond, slave) { 2604 bond_for_each_slave(bond, slave, iter) {
2695 switch (slave->new_link) { 2605 switch (slave->new_link) {
2696 case BOND_LINK_NOCHANGE: 2606 case BOND_LINK_NOCHANGE:
2697 continue; 2607 continue;
@@ -2762,8 +2672,9 @@ do_failover:
2762 */ 2672 */
2763static void bond_ab_arp_probe(struct bonding *bond) 2673static void bond_ab_arp_probe(struct bonding *bond)
2764{ 2674{
2765 struct slave *slave, *next_slave; 2675 struct slave *slave, *before = NULL, *new_slave = NULL;
2766 int i; 2676 struct list_head *iter;
2677 bool found = false;
2767 2678
2768 read_lock(&bond->curr_slave_lock); 2679 read_lock(&bond->curr_slave_lock);
2769 2680
@@ -2793,18 +2704,12 @@ static void bond_ab_arp_probe(struct bonding *bond)
2793 2704
2794 bond_set_slave_inactive_flags(bond->current_arp_slave); 2705 bond_set_slave_inactive_flags(bond->current_arp_slave);
2795 2706
2796 /* search for next candidate */ 2707 bond_for_each_slave(bond, slave, iter) {
2797 next_slave = bond_next_slave(bond, bond->current_arp_slave); 2708 if (!found && !before && IS_UP(slave->dev))
2798 bond_for_each_slave_from(bond, slave, i, next_slave) { 2709 before = slave;
2799 if (IS_UP(slave->dev)) {
2800 slave->link = BOND_LINK_BACK;
2801 bond_set_slave_active_flags(slave);
2802 bond_arp_send_all(bond, slave);
2803 slave->jiffies = jiffies;
2804 bond->current_arp_slave = slave;
2805 break;
2806 }
2807 2710
2711 if (found && !new_slave && IS_UP(slave->dev))
2712 new_slave = slave;
2808 /* if the link state is up at this point, we 2713 /* if the link state is up at this point, we
2809 * mark it down - this can happen if we have 2714 * mark it down - this can happen if we have
2810 * simultaneous link failures and 2715 * simultaneous link failures and
@@ -2812,7 +2717,7 @@ static void bond_ab_arp_probe(struct bonding *bond)
2812 * one the current slave so it is still marked 2717 * one the current slave so it is still marked
2813 * up when it is actually down 2718 * up when it is actually down
2814 */ 2719 */
2815 if (slave->link == BOND_LINK_UP) { 2720 if (!IS_UP(slave->dev) && slave->link == BOND_LINK_UP) {
2816 slave->link = BOND_LINK_DOWN; 2721 slave->link = BOND_LINK_DOWN;
2817 if (slave->link_failure_count < UINT_MAX) 2722 if (slave->link_failure_count < UINT_MAX)
2818 slave->link_failure_count++; 2723 slave->link_failure_count++;
@@ -2822,7 +2727,22 @@ static void bond_ab_arp_probe(struct bonding *bond)
2822 pr_info("%s: backup interface %s is now down.\n", 2727 pr_info("%s: backup interface %s is now down.\n",
2823 bond->dev->name, slave->dev->name); 2728 bond->dev->name, slave->dev->name);
2824 } 2729 }
2730 if (slave == bond->current_arp_slave)
2731 found = true;
2825 } 2732 }
2733
2734 if (!new_slave && before)
2735 new_slave = before;
2736
2737 if (!new_slave)
2738 return;
2739
2740 new_slave->link = BOND_LINK_BACK;
2741 bond_set_slave_active_flags(new_slave);
2742 bond_arp_send_all(bond, new_slave);
2743 new_slave->jiffies = jiffies;
2744 bond->current_arp_slave = new_slave;
2745
2826} 2746}
2827 2747
2828void bond_activebackup_arp_mon(struct work_struct *work) 2748void bond_activebackup_arp_mon(struct work_struct *work)
@@ -2836,7 +2756,7 @@ void bond_activebackup_arp_mon(struct work_struct *work)
2836 2756
2837 delta_in_ticks = msecs_to_jiffies(bond->params.arp_interval); 2757 delta_in_ticks = msecs_to_jiffies(bond->params.arp_interval);
2838 2758
2839 if (list_empty(&bond->slave_list)) 2759 if (!bond_has_slaves(bond))
2840 goto re_arm; 2760 goto re_arm;
2841 2761
2842 should_notify_peers = bond_should_notify_peers(bond); 2762 should_notify_peers = bond_should_notify_peers(bond);
@@ -3033,99 +2953,85 @@ static struct notifier_block bond_netdev_notifier = {
3033 2953
3034/*---------------------------- Hashing Policies -----------------------------*/ 2954/*---------------------------- Hashing Policies -----------------------------*/
3035 2955
3036/* 2956/* L2 hash helper */
3037 * Hash for the output device based upon layer 2 data 2957static inline u32 bond_eth_hash(struct sk_buff *skb)
3038 */
3039static int bond_xmit_hash_policy_l2(struct sk_buff *skb, int count)
3040{ 2958{
3041 struct ethhdr *data = (struct ethhdr *)skb->data; 2959 struct ethhdr *data = (struct ethhdr *)skb->data;
3042 2960
3043 if (skb_headlen(skb) >= offsetof(struct ethhdr, h_proto)) 2961 if (skb_headlen(skb) >= offsetof(struct ethhdr, h_proto))
3044 return (data->h_dest[5] ^ data->h_source[5]) % count; 2962 return data->h_dest[5] ^ data->h_source[5];
3045 2963
3046 return 0; 2964 return 0;
3047} 2965}
3048 2966
3049/* 2967/* Extract the appropriate headers based on bond's xmit policy */
3050 * Hash for the output device based upon layer 2 and layer 3 data. If 2968static bool bond_flow_dissect(struct bonding *bond, struct sk_buff *skb,
3051 * the packet is not IP, fall back on bond_xmit_hash_policy_l2() 2969 struct flow_keys *fk)
3052 */
3053static int bond_xmit_hash_policy_l23(struct sk_buff *skb, int count)
3054{ 2970{
3055 const struct ethhdr *data; 2971 const struct ipv6hdr *iph6;
3056 const struct iphdr *iph; 2972 const struct iphdr *iph;
3057 const struct ipv6hdr *ipv6h; 2973 int noff, proto = -1;
3058 u32 v6hash; 2974
3059 const __be32 *s, *d; 2975 if (bond->params.xmit_policy > BOND_XMIT_POLICY_LAYER23)
2976 return skb_flow_dissect(skb, fk);
3060 2977
3061 if (skb->protocol == htons(ETH_P_IP) && 2978 fk->ports = 0;
3062 pskb_network_may_pull(skb, sizeof(*iph))) { 2979 noff = skb_network_offset(skb);
2980 if (skb->protocol == htons(ETH_P_IP)) {
2981 if (!pskb_may_pull(skb, noff + sizeof(*iph)))
2982 return false;
3063 iph = ip_hdr(skb); 2983 iph = ip_hdr(skb);
3064 data = (struct ethhdr *)skb->data; 2984 fk->src = iph->saddr;
3065 return ((ntohl(iph->saddr ^ iph->daddr) & 0xffff) ^ 2985 fk->dst = iph->daddr;
3066 (data->h_dest[5] ^ data->h_source[5])) % count; 2986 noff += iph->ihl << 2;
3067 } else if (skb->protocol == htons(ETH_P_IPV6) && 2987 if (!ip_is_fragment(iph))
3068 pskb_network_may_pull(skb, sizeof(*ipv6h))) { 2988 proto = iph->protocol;
3069 ipv6h = ipv6_hdr(skb); 2989 } else if (skb->protocol == htons(ETH_P_IPV6)) {
3070 data = (struct ethhdr *)skb->data; 2990 if (!pskb_may_pull(skb, noff + sizeof(*iph6)))
3071 s = &ipv6h->saddr.s6_addr32[0]; 2991 return false;
3072 d = &ipv6h->daddr.s6_addr32[0]; 2992 iph6 = ipv6_hdr(skb);
3073 v6hash = (s[1] ^ d[1]) ^ (s[2] ^ d[2]) ^ (s[3] ^ d[3]); 2993 fk->src = (__force __be32)ipv6_addr_hash(&iph6->saddr);
3074 v6hash ^= (v6hash >> 24) ^ (v6hash >> 16) ^ (v6hash >> 8); 2994 fk->dst = (__force __be32)ipv6_addr_hash(&iph6->daddr);
3075 return (v6hash ^ data->h_dest[5] ^ data->h_source[5]) % count; 2995 noff += sizeof(*iph6);
3076 } 2996 proto = iph6->nexthdr;
3077 2997 } else {
3078 return bond_xmit_hash_policy_l2(skb, count); 2998 return false;
2999 }
3000 if (bond->params.xmit_policy == BOND_XMIT_POLICY_LAYER34 && proto >= 0)
3001 fk->ports = skb_flow_get_ports(skb, noff, proto);
3002
3003 return true;
3079} 3004}
3080 3005
3081/* 3006/**
3082 * Hash for the output device based upon layer 3 and layer 4 data. If 3007 * bond_xmit_hash - generate a hash value based on the xmit policy
3083 * the packet is a frag or not TCP or UDP, just use layer 3 data. If it is 3008 * @bond: bonding device
3084 * altogether not IP, fall back on bond_xmit_hash_policy_l2() 3009 * @skb: buffer to use for headers
3010 * @count: modulo value
3011 *
3012 * This function will extract the necessary headers from the skb buffer and use
3013 * them to generate a hash based on the xmit_policy set in the bonding device
3014 * which will be reduced modulo count before returning.
3085 */ 3015 */
3086static int bond_xmit_hash_policy_l34(struct sk_buff *skb, int count) 3016int bond_xmit_hash(struct bonding *bond, struct sk_buff *skb, int count)
3087{ 3017{
3088 u32 layer4_xor = 0; 3018 struct flow_keys flow;
3089 const struct iphdr *iph; 3019 u32 hash;
3090 const struct ipv6hdr *ipv6h;
3091 const __be32 *s, *d;
3092 const __be16 *l4 = NULL;
3093 __be16 _l4[2];
3094 int noff = skb_network_offset(skb);
3095 int poff;
3096
3097 if (skb->protocol == htons(ETH_P_IP) &&
3098 pskb_may_pull(skb, noff + sizeof(*iph))) {
3099 iph = ip_hdr(skb);
3100 poff = proto_ports_offset(iph->protocol);
3101 3020
3102 if (!ip_is_fragment(iph) && poff >= 0) { 3021 if (bond->params.xmit_policy == BOND_XMIT_POLICY_LAYER2 ||
3103 l4 = skb_header_pointer(skb, noff + (iph->ihl << 2) + poff, 3022 !bond_flow_dissect(bond, skb, &flow))
3104 sizeof(_l4), &_l4); 3023 return bond_eth_hash(skb) % count;
3105 if (l4) 3024
3106 layer4_xor = ntohs(l4[0] ^ l4[1]); 3025 if (bond->params.xmit_policy == BOND_XMIT_POLICY_LAYER23 ||
3107 } 3026 bond->params.xmit_policy == BOND_XMIT_POLICY_ENCAP23)
3108 return (layer4_xor ^ 3027 hash = bond_eth_hash(skb);
3109 ((ntohl(iph->saddr ^ iph->daddr)) & 0xffff)) % count; 3028 else
3110 } else if (skb->protocol == htons(ETH_P_IPV6) && 3029 hash = (__force u32)flow.ports;
3111 pskb_may_pull(skb, noff + sizeof(*ipv6h))) { 3030 hash ^= (__force u32)flow.dst ^ (__force u32)flow.src;
3112 ipv6h = ipv6_hdr(skb); 3031 hash ^= (hash >> 16);
3113 poff = proto_ports_offset(ipv6h->nexthdr); 3032 hash ^= (hash >> 8);
3114 if (poff >= 0) {
3115 l4 = skb_header_pointer(skb, noff + sizeof(*ipv6h) + poff,
3116 sizeof(_l4), &_l4);
3117 if (l4)
3118 layer4_xor = ntohs(l4[0] ^ l4[1]);
3119 }
3120 s = &ipv6h->saddr.s6_addr32[0];
3121 d = &ipv6h->daddr.s6_addr32[0];
3122 layer4_xor ^= (s[1] ^ d[1]) ^ (s[2] ^ d[2]) ^ (s[3] ^ d[3]);
3123 layer4_xor ^= (layer4_xor >> 24) ^ (layer4_xor >> 16) ^
3124 (layer4_xor >> 8);
3125 return layer4_xor % count;
3126 }
3127 3033
3128 return bond_xmit_hash_policy_l2(skb, count); 3034 return hash % count;
3129} 3035}
3130 3036
3131/*-------------------------- Device entry points ----------------------------*/ 3037/*-------------------------- Device entry points ----------------------------*/
@@ -3155,13 +3061,14 @@ static void bond_work_cancel_all(struct bonding *bond)
3155static int bond_open(struct net_device *bond_dev) 3061static int bond_open(struct net_device *bond_dev)
3156{ 3062{
3157 struct bonding *bond = netdev_priv(bond_dev); 3063 struct bonding *bond = netdev_priv(bond_dev);
3064 struct list_head *iter;
3158 struct slave *slave; 3065 struct slave *slave;
3159 3066
3160 /* reset slave->backup and slave->inactive */ 3067 /* reset slave->backup and slave->inactive */
3161 read_lock(&bond->lock); 3068 read_lock(&bond->lock);
3162 if (!list_empty(&bond->slave_list)) { 3069 if (bond_has_slaves(bond)) {
3163 read_lock(&bond->curr_slave_lock); 3070 read_lock(&bond->curr_slave_lock);
3164 bond_for_each_slave(bond, slave) { 3071 bond_for_each_slave(bond, slave, iter) {
3165 if ((bond->params.mode == BOND_MODE_ACTIVEBACKUP) 3072 if ((bond->params.mode == BOND_MODE_ACTIVEBACKUP)
3166 && (slave != bond->curr_active_slave)) { 3073 && (slave != bond->curr_active_slave)) {
3167 bond_set_slave_inactive_flags(slave); 3074 bond_set_slave_inactive_flags(slave);
@@ -3221,12 +3128,13 @@ static struct rtnl_link_stats64 *bond_get_stats(struct net_device *bond_dev,
3221{ 3128{
3222 struct bonding *bond = netdev_priv(bond_dev); 3129 struct bonding *bond = netdev_priv(bond_dev);
3223 struct rtnl_link_stats64 temp; 3130 struct rtnl_link_stats64 temp;
3131 struct list_head *iter;
3224 struct slave *slave; 3132 struct slave *slave;
3225 3133
3226 memset(stats, 0, sizeof(*stats)); 3134 memset(stats, 0, sizeof(*stats));
3227 3135
3228 read_lock_bh(&bond->lock); 3136 read_lock_bh(&bond->lock);
3229 bond_for_each_slave(bond, slave) { 3137 bond_for_each_slave(bond, slave, iter) {
3230 const struct rtnl_link_stats64 *sstats = 3138 const struct rtnl_link_stats64 *sstats =
3231 dev_get_stats(slave->dev, &temp); 3139 dev_get_stats(slave->dev, &temp);
3232 3140
@@ -3263,6 +3171,7 @@ static struct rtnl_link_stats64 *bond_get_stats(struct net_device *bond_dev,
3263 3171
3264static int bond_do_ioctl(struct net_device *bond_dev, struct ifreq *ifr, int cmd) 3172static int bond_do_ioctl(struct net_device *bond_dev, struct ifreq *ifr, int cmd)
3265{ 3173{
3174 struct bonding *bond = netdev_priv(bond_dev);
3266 struct net_device *slave_dev = NULL; 3175 struct net_device *slave_dev = NULL;
3267 struct ifbond k_binfo; 3176 struct ifbond k_binfo;
3268 struct ifbond __user *u_binfo = NULL; 3177 struct ifbond __user *u_binfo = NULL;
@@ -3293,7 +3202,6 @@ static int bond_do_ioctl(struct net_device *bond_dev, struct ifreq *ifr, int cmd
3293 3202
3294 3203
3295 if (mii->reg_num == 1) { 3204 if (mii->reg_num == 1) {
3296 struct bonding *bond = netdev_priv(bond_dev);
3297 mii->val_out = 0; 3205 mii->val_out = 0;
3298 read_lock(&bond->lock); 3206 read_lock(&bond->lock);
3299 read_lock(&bond->curr_slave_lock); 3207 read_lock(&bond->curr_slave_lock);
@@ -3365,7 +3273,7 @@ static int bond_do_ioctl(struct net_device *bond_dev, struct ifreq *ifr, int cmd
3365 break; 3273 break;
3366 case BOND_CHANGE_ACTIVE_OLD: 3274 case BOND_CHANGE_ACTIVE_OLD:
3367 case SIOCBONDCHANGEACTIVE: 3275 case SIOCBONDCHANGEACTIVE:
3368 res = bond_ioctl_change_active(bond_dev, slave_dev); 3276 res = bond_option_active_slave_set(bond, slave_dev);
3369 break; 3277 break;
3370 default: 3278 default:
3371 res = -EOPNOTSUPP; 3279 res = -EOPNOTSUPP;
@@ -3393,22 +3301,24 @@ static void bond_change_rx_flags(struct net_device *bond_dev, int change)
3393static void bond_set_rx_mode(struct net_device *bond_dev) 3301static void bond_set_rx_mode(struct net_device *bond_dev)
3394{ 3302{
3395 struct bonding *bond = netdev_priv(bond_dev); 3303 struct bonding *bond = netdev_priv(bond_dev);
3304 struct list_head *iter;
3396 struct slave *slave; 3305 struct slave *slave;
3397 3306
3398 ASSERT_RTNL();
3399 3307
3308 rcu_read_lock();
3400 if (USES_PRIMARY(bond->params.mode)) { 3309 if (USES_PRIMARY(bond->params.mode)) {
3401 slave = rtnl_dereference(bond->curr_active_slave); 3310 slave = rcu_dereference(bond->curr_active_slave);
3402 if (slave) { 3311 if (slave) {
3403 dev_uc_sync(slave->dev, bond_dev); 3312 dev_uc_sync(slave->dev, bond_dev);
3404 dev_mc_sync(slave->dev, bond_dev); 3313 dev_mc_sync(slave->dev, bond_dev);
3405 } 3314 }
3406 } else { 3315 } else {
3407 bond_for_each_slave(bond, slave) { 3316 bond_for_each_slave_rcu(bond, slave, iter) {
3408 dev_uc_sync_multiple(slave->dev, bond_dev); 3317 dev_uc_sync_multiple(slave->dev, bond_dev);
3409 dev_mc_sync_multiple(slave->dev, bond_dev); 3318 dev_mc_sync_multiple(slave->dev, bond_dev);
3410 } 3319 }
3411 } 3320 }
3321 rcu_read_unlock();
3412} 3322}
3413 3323
3414static int bond_neigh_init(struct neighbour *n) 3324static int bond_neigh_init(struct neighbour *n)
@@ -3471,7 +3381,8 @@ static int bond_neigh_setup(struct net_device *dev,
3471static int bond_change_mtu(struct net_device *bond_dev, int new_mtu) 3381static int bond_change_mtu(struct net_device *bond_dev, int new_mtu)
3472{ 3382{
3473 struct bonding *bond = netdev_priv(bond_dev); 3383 struct bonding *bond = netdev_priv(bond_dev);
3474 struct slave *slave; 3384 struct slave *slave, *rollback_slave;
3385 struct list_head *iter;
3475 int res = 0; 3386 int res = 0;
3476 3387
3477 pr_debug("bond=%p, name=%s, new_mtu=%d\n", bond, 3388 pr_debug("bond=%p, name=%s, new_mtu=%d\n", bond,
@@ -3492,10 +3403,9 @@ static int bond_change_mtu(struct net_device *bond_dev, int new_mtu)
3492 * call to the base driver. 3403 * call to the base driver.
3493 */ 3404 */
3494 3405
3495 bond_for_each_slave(bond, slave) { 3406 bond_for_each_slave(bond, slave, iter) {
3496 pr_debug("s %p s->p %p c_m %p\n", 3407 pr_debug("s %p c_m %p\n",
3497 slave, 3408 slave,
3498 bond_prev_slave(bond, slave),
3499 slave->dev->netdev_ops->ndo_change_mtu); 3409 slave->dev->netdev_ops->ndo_change_mtu);
3500 3410
3501 res = dev_set_mtu(slave->dev, new_mtu); 3411 res = dev_set_mtu(slave->dev, new_mtu);
@@ -3520,13 +3430,16 @@ static int bond_change_mtu(struct net_device *bond_dev, int new_mtu)
3520 3430
3521unwind: 3431unwind:
3522 /* unwind from head to the slave that failed */ 3432 /* unwind from head to the slave that failed */
3523 bond_for_each_slave_continue_reverse(bond, slave) { 3433 bond_for_each_slave(bond, rollback_slave, iter) {
3524 int tmp_res; 3434 int tmp_res;
3525 3435
3526 tmp_res = dev_set_mtu(slave->dev, bond_dev->mtu); 3436 if (rollback_slave == slave)
3437 break;
3438
3439 tmp_res = dev_set_mtu(rollback_slave->dev, bond_dev->mtu);
3527 if (tmp_res) { 3440 if (tmp_res) {
3528 pr_debug("unwind err %d dev %s\n", 3441 pr_debug("unwind err %d dev %s\n",
3529 tmp_res, slave->dev->name); 3442 tmp_res, rollback_slave->dev->name);
3530 } 3443 }
3531 } 3444 }
3532 3445
@@ -3543,8 +3456,9 @@ unwind:
3543static int bond_set_mac_address(struct net_device *bond_dev, void *addr) 3456static int bond_set_mac_address(struct net_device *bond_dev, void *addr)
3544{ 3457{
3545 struct bonding *bond = netdev_priv(bond_dev); 3458 struct bonding *bond = netdev_priv(bond_dev);
3459 struct slave *slave, *rollback_slave;
3546 struct sockaddr *sa = addr, tmp_sa; 3460 struct sockaddr *sa = addr, tmp_sa;
3547 struct slave *slave; 3461 struct list_head *iter;
3548 int res = 0; 3462 int res = 0;
3549 3463
3550 if (bond->params.mode == BOND_MODE_ALB) 3464 if (bond->params.mode == BOND_MODE_ALB)
@@ -3578,7 +3492,7 @@ static int bond_set_mac_address(struct net_device *bond_dev, void *addr)
3578 * call to the base driver. 3492 * call to the base driver.
3579 */ 3493 */
3580 3494
3581 bond_for_each_slave(bond, slave) { 3495 bond_for_each_slave(bond, slave, iter) {
3582 const struct net_device_ops *slave_ops = slave->dev->netdev_ops; 3496 const struct net_device_ops *slave_ops = slave->dev->netdev_ops;
3583 pr_debug("slave %p %s\n", slave, slave->dev->name); 3497 pr_debug("slave %p %s\n", slave, slave->dev->name);
3584 3498
@@ -3610,13 +3524,16 @@ unwind:
3610 tmp_sa.sa_family = bond_dev->type; 3524 tmp_sa.sa_family = bond_dev->type;
3611 3525
3612 /* unwind from head to the slave that failed */ 3526 /* unwind from head to the slave that failed */
3613 bond_for_each_slave_continue_reverse(bond, slave) { 3527 bond_for_each_slave(bond, rollback_slave, iter) {
3614 int tmp_res; 3528 int tmp_res;
3615 3529
3616 tmp_res = dev_set_mac_address(slave->dev, &tmp_sa); 3530 if (rollback_slave == slave)
3531 break;
3532
3533 tmp_res = dev_set_mac_address(rollback_slave->dev, &tmp_sa);
3617 if (tmp_res) { 3534 if (tmp_res) {
3618 pr_debug("unwind err %d dev %s\n", 3535 pr_debug("unwind err %d dev %s\n",
3619 tmp_res, slave->dev->name); 3536 tmp_res, rollback_slave->dev->name);
3620 } 3537 }
3621 } 3538 }
3622 3539
@@ -3635,11 +3552,12 @@ unwind:
3635 */ 3552 */
3636void bond_xmit_slave_id(struct bonding *bond, struct sk_buff *skb, int slave_id) 3553void bond_xmit_slave_id(struct bonding *bond, struct sk_buff *skb, int slave_id)
3637{ 3554{
3555 struct list_head *iter;
3638 struct slave *slave; 3556 struct slave *slave;
3639 int i = slave_id; 3557 int i = slave_id;
3640 3558
3641 /* Here we start from the slave with slave_id */ 3559 /* Here we start from the slave with slave_id */
3642 bond_for_each_slave_rcu(bond, slave) { 3560 bond_for_each_slave_rcu(bond, slave, iter) {
3643 if (--i < 0) { 3561 if (--i < 0) {
3644 if (slave_can_tx(slave)) { 3562 if (slave_can_tx(slave)) {
3645 bond_dev_queue_xmit(bond, skb, slave->dev); 3563 bond_dev_queue_xmit(bond, skb, slave->dev);
@@ -3650,7 +3568,7 @@ void bond_xmit_slave_id(struct bonding *bond, struct sk_buff *skb, int slave_id)
3650 3568
3651 /* Here we start from the first slave up to slave_id */ 3569 /* Here we start from the first slave up to slave_id */
3652 i = slave_id; 3570 i = slave_id;
3653 bond_for_each_slave_rcu(bond, slave) { 3571 bond_for_each_slave_rcu(bond, slave, iter) {
3654 if (--i < 0) 3572 if (--i < 0)
3655 break; 3573 break;
3656 if (slave_can_tx(slave)) { 3574 if (slave_can_tx(slave)) {
@@ -3662,14 +3580,44 @@ void bond_xmit_slave_id(struct bonding *bond, struct sk_buff *skb, int slave_id)
3662 kfree_skb(skb); 3580 kfree_skb(skb);
3663} 3581}
3664 3582
3583/**
3584 * bond_rr_gen_slave_id - generate slave id based on packets_per_slave
3585 * @bond: bonding device to use
3586 *
3587 * Based on the value of the bonding device's packets_per_slave parameter
3588 * this function generates a slave id, which is usually used as the next
3589 * slave to transmit through.
3590 */
3591static u32 bond_rr_gen_slave_id(struct bonding *bond)
3592{
3593 int packets_per_slave = bond->params.packets_per_slave;
3594 u32 slave_id;
3595
3596 switch (packets_per_slave) {
3597 case 0:
3598 slave_id = prandom_u32();
3599 break;
3600 case 1:
3601 slave_id = bond->rr_tx_counter;
3602 break;
3603 default:
3604 slave_id = reciprocal_divide(bond->rr_tx_counter,
3605 packets_per_slave);
3606 break;
3607 }
3608 bond->rr_tx_counter++;
3609
3610 return slave_id;
3611}
3612
3665static int bond_xmit_roundrobin(struct sk_buff *skb, struct net_device *bond_dev) 3613static int bond_xmit_roundrobin(struct sk_buff *skb, struct net_device *bond_dev)
3666{ 3614{
3667 struct bonding *bond = netdev_priv(bond_dev); 3615 struct bonding *bond = netdev_priv(bond_dev);
3668 struct iphdr *iph = ip_hdr(skb); 3616 struct iphdr *iph = ip_hdr(skb);
3669 struct slave *slave; 3617 struct slave *slave;
3618 u32 slave_id;
3670 3619
3671 /* 3620 /* Start with the curr_active_slave that joined the bond as the
3672 * Start with the curr_active_slave that joined the bond as the
3673 * default for sending IGMP traffic. For failover purposes one 3621 * default for sending IGMP traffic. For failover purposes one
3674 * needs to maintain some consistency for the interface that will 3622 * needs to maintain some consistency for the interface that will
3675 * send the join/membership reports. The curr_active_slave found 3623 * send the join/membership reports. The curr_active_slave found
@@ -3682,8 +3630,8 @@ static int bond_xmit_roundrobin(struct sk_buff *skb, struct net_device *bond_dev
3682 else 3630 else
3683 bond_xmit_slave_id(bond, skb, 0); 3631 bond_xmit_slave_id(bond, skb, 0);
3684 } else { 3632 } else {
3685 bond_xmit_slave_id(bond, skb, 3633 slave_id = bond_rr_gen_slave_id(bond);
3686 bond->rr_tx_counter++ % bond->slave_cnt); 3634 bond_xmit_slave_id(bond, skb, slave_id % bond->slave_cnt);
3687 } 3635 }
3688 3636
3689 return NETDEV_TX_OK; 3637 return NETDEV_TX_OK;
@@ -3707,8 +3655,7 @@ static int bond_xmit_activebackup(struct sk_buff *skb, struct net_device *bond_d
3707 return NETDEV_TX_OK; 3655 return NETDEV_TX_OK;
3708} 3656}
3709 3657
3710/* 3658/* In bond_xmit_xor() , we determine the output device by using a pre-
3711 * In bond_xmit_xor() , we determine the output device by using a pre-
3712 * determined xmit_hash_policy(), If the selected device is not enabled, 3659 * determined xmit_hash_policy(), If the selected device is not enabled,
3713 * find the next active slave. 3660 * find the next active slave.
3714 */ 3661 */
@@ -3716,8 +3663,7 @@ static int bond_xmit_xor(struct sk_buff *skb, struct net_device *bond_dev)
3716{ 3663{
3717 struct bonding *bond = netdev_priv(bond_dev); 3664 struct bonding *bond = netdev_priv(bond_dev);
3718 3665
3719 bond_xmit_slave_id(bond, skb, 3666 bond_xmit_slave_id(bond, skb, bond_xmit_hash(bond, skb, bond->slave_cnt));
3720 bond->xmit_hash_policy(skb, bond->slave_cnt));
3721 3667
3722 return NETDEV_TX_OK; 3668 return NETDEV_TX_OK;
3723} 3669}
@@ -3727,8 +3673,9 @@ static int bond_xmit_broadcast(struct sk_buff *skb, struct net_device *bond_dev)
3727{ 3673{
3728 struct bonding *bond = netdev_priv(bond_dev); 3674 struct bonding *bond = netdev_priv(bond_dev);
3729 struct slave *slave = NULL; 3675 struct slave *slave = NULL;
3676 struct list_head *iter;
3730 3677
3731 bond_for_each_slave_rcu(bond, slave) { 3678 bond_for_each_slave_rcu(bond, slave, iter) {
3732 if (bond_is_last_slave(bond, slave)) 3679 if (bond_is_last_slave(bond, slave))
3733 break; 3680 break;
3734 if (IS_UP(slave->dev) && slave->link == BOND_LINK_UP) { 3681 if (IS_UP(slave->dev) && slave->link == BOND_LINK_UP) {
@@ -3753,22 +3700,6 @@ static int bond_xmit_broadcast(struct sk_buff *skb, struct net_device *bond_dev)
3753 3700
3754/*------------------------- Device initialization ---------------------------*/ 3701/*------------------------- Device initialization ---------------------------*/
3755 3702
3756static void bond_set_xmit_hash_policy(struct bonding *bond)
3757{
3758 switch (bond->params.xmit_policy) {
3759 case BOND_XMIT_POLICY_LAYER23:
3760 bond->xmit_hash_policy = bond_xmit_hash_policy_l23;
3761 break;
3762 case BOND_XMIT_POLICY_LAYER34:
3763 bond->xmit_hash_policy = bond_xmit_hash_policy_l34;
3764 break;
3765 case BOND_XMIT_POLICY_LAYER2:
3766 default:
3767 bond->xmit_hash_policy = bond_xmit_hash_policy_l2;
3768 break;
3769 }
3770}
3771
3772/* 3703/*
3773 * Lookup the slave that corresponds to a qid 3704 * Lookup the slave that corresponds to a qid
3774 */ 3705 */
@@ -3777,13 +3708,14 @@ static inline int bond_slave_override(struct bonding *bond,
3777{ 3708{
3778 struct slave *slave = NULL; 3709 struct slave *slave = NULL;
3779 struct slave *check_slave; 3710 struct slave *check_slave;
3711 struct list_head *iter;
3780 int res = 1; 3712 int res = 1;
3781 3713
3782 if (!skb->queue_mapping) 3714 if (!skb->queue_mapping)
3783 return 1; 3715 return 1;
3784 3716
3785 /* Find out if any slaves have the same mapping as this skb. */ 3717 /* Find out if any slaves have the same mapping as this skb. */
3786 bond_for_each_slave_rcu(bond, check_slave) { 3718 bond_for_each_slave_rcu(bond, check_slave, iter) {
3787 if (check_slave->queue_id == skb->queue_mapping) { 3719 if (check_slave->queue_id == skb->queue_mapping) {
3788 slave = check_slave; 3720 slave = check_slave;
3789 break; 3721 break;
@@ -3869,7 +3801,7 @@ static netdev_tx_t bond_start_xmit(struct sk_buff *skb, struct net_device *dev)
3869 return NETDEV_TX_BUSY; 3801 return NETDEV_TX_BUSY;
3870 3802
3871 rcu_read_lock(); 3803 rcu_read_lock();
3872 if (!list_empty(&bond->slave_list)) 3804 if (bond_has_slaves(bond))
3873 ret = __bond_start_xmit(skb, dev); 3805 ret = __bond_start_xmit(skb, dev);
3874 else 3806 else
3875 kfree_skb(skb); 3807 kfree_skb(skb);
@@ -3878,43 +3810,12 @@ static netdev_tx_t bond_start_xmit(struct sk_buff *skb, struct net_device *dev)
3878 return ret; 3810 return ret;
3879} 3811}
3880 3812
3881/*
3882 * set bond mode specific net device operations
3883 */
3884void bond_set_mode_ops(struct bonding *bond, int mode)
3885{
3886 struct net_device *bond_dev = bond->dev;
3887
3888 switch (mode) {
3889 case BOND_MODE_ROUNDROBIN:
3890 break;
3891 case BOND_MODE_ACTIVEBACKUP:
3892 break;
3893 case BOND_MODE_XOR:
3894 bond_set_xmit_hash_policy(bond);
3895 break;
3896 case BOND_MODE_BROADCAST:
3897 break;
3898 case BOND_MODE_8023AD:
3899 bond_set_xmit_hash_policy(bond);
3900 break;
3901 case BOND_MODE_ALB:
3902 /* FALLTHRU */
3903 case BOND_MODE_TLB:
3904 break;
3905 default:
3906 /* Should never happen, mode already checked */
3907 pr_err("%s: Error: Unknown bonding mode %d\n",
3908 bond_dev->name, mode);
3909 break;
3910 }
3911}
3912
3913static int bond_ethtool_get_settings(struct net_device *bond_dev, 3813static int bond_ethtool_get_settings(struct net_device *bond_dev,
3914 struct ethtool_cmd *ecmd) 3814 struct ethtool_cmd *ecmd)
3915{ 3815{
3916 struct bonding *bond = netdev_priv(bond_dev); 3816 struct bonding *bond = netdev_priv(bond_dev);
3917 unsigned long speed = 0; 3817 unsigned long speed = 0;
3818 struct list_head *iter;
3918 struct slave *slave; 3819 struct slave *slave;
3919 3820
3920 ecmd->duplex = DUPLEX_UNKNOWN; 3821 ecmd->duplex = DUPLEX_UNKNOWN;
@@ -3926,7 +3827,7 @@ static int bond_ethtool_get_settings(struct net_device *bond_dev,
3926 * this is an accurate maximum. 3827 * this is an accurate maximum.
3927 */ 3828 */
3928 read_lock(&bond->lock); 3829 read_lock(&bond->lock);
3929 bond_for_each_slave(bond, slave) { 3830 bond_for_each_slave(bond, slave, iter) {
3930 if (SLAVE_IS_OK(slave)) { 3831 if (SLAVE_IS_OK(slave)) {
3931 if (slave->speed != SPEED_UNKNOWN) 3832 if (slave->speed != SPEED_UNKNOWN)
3932 speed += slave->speed; 3833 speed += slave->speed;
@@ -3994,14 +3895,13 @@ static void bond_destructor(struct net_device *bond_dev)
3994 free_netdev(bond_dev); 3895 free_netdev(bond_dev);
3995} 3896}
3996 3897
3997static void bond_setup(struct net_device *bond_dev) 3898void bond_setup(struct net_device *bond_dev)
3998{ 3899{
3999 struct bonding *bond = netdev_priv(bond_dev); 3900 struct bonding *bond = netdev_priv(bond_dev);
4000 3901
4001 /* initialize rwlocks */ 3902 /* initialize rwlocks */
4002 rwlock_init(&bond->lock); 3903 rwlock_init(&bond->lock);
4003 rwlock_init(&bond->curr_slave_lock); 3904 rwlock_init(&bond->curr_slave_lock);
4004 INIT_LIST_HEAD(&bond->slave_list);
4005 bond->params = bonding_defaults; 3905 bond->params = bonding_defaults;
4006 3906
4007 /* Initialize pointers */ 3907 /* Initialize pointers */
@@ -4011,7 +3911,6 @@ static void bond_setup(struct net_device *bond_dev)
4011 ether_setup(bond_dev); 3911 ether_setup(bond_dev);
4012 bond_dev->netdev_ops = &bond_netdev_ops; 3912 bond_dev->netdev_ops = &bond_netdev_ops;
4013 bond_dev->ethtool_ops = &bond_ethtool_ops; 3913 bond_dev->ethtool_ops = &bond_ethtool_ops;
4014 bond_set_mode_ops(bond, bond->params.mode);
4015 3914
4016 bond_dev->destructor = bond_destructor; 3915 bond_dev->destructor = bond_destructor;
4017 3916
@@ -4057,12 +3956,13 @@ static void bond_setup(struct net_device *bond_dev)
4057static void bond_uninit(struct net_device *bond_dev) 3956static void bond_uninit(struct net_device *bond_dev)
4058{ 3957{
4059 struct bonding *bond = netdev_priv(bond_dev); 3958 struct bonding *bond = netdev_priv(bond_dev);
4060 struct slave *slave, *tmp_slave; 3959 struct list_head *iter;
3960 struct slave *slave;
4061 3961
4062 bond_netpoll_cleanup(bond_dev); 3962 bond_netpoll_cleanup(bond_dev);
4063 3963
4064 /* Release the bonded slaves */ 3964 /* Release the bonded slaves */
4065 list_for_each_entry_safe(slave, tmp_slave, &bond->slave_list, list) 3965 bond_for_each_slave(bond, slave, iter)
4066 __bond_release_one(bond_dev, slave->dev, true); 3966 __bond_release_one(bond_dev, slave->dev, true);
4067 pr_info("%s: released all slaves\n", bond_dev->name); 3967 pr_info("%s: released all slaves\n", bond_dev->name);
4068 3968
@@ -4235,6 +4135,12 @@ static int bond_check_params(struct bond_params *params)
4235 resend_igmp = BOND_DEFAULT_RESEND_IGMP; 4135 resend_igmp = BOND_DEFAULT_RESEND_IGMP;
4236 } 4136 }
4237 4137
4138 if (packets_per_slave < 0 || packets_per_slave > USHRT_MAX) {
4139 pr_warn("Warning: packets_per_slave (%d) should be between 0 and %u resetting to 1\n",
4140 packets_per_slave, USHRT_MAX);
4141 packets_per_slave = 1;
4142 }
4143
4238 /* reset values for TLB/ALB */ 4144 /* reset values for TLB/ALB */
4239 if ((bond_mode == BOND_MODE_TLB) || 4145 if ((bond_mode == BOND_MODE_TLB) ||
4240 (bond_mode == BOND_MODE_ALB)) { 4146 (bond_mode == BOND_MODE_ALB)) {
@@ -4424,7 +4330,10 @@ static int bond_check_params(struct bond_params *params)
4424 params->resend_igmp = resend_igmp; 4330 params->resend_igmp = resend_igmp;
4425 params->min_links = min_links; 4331 params->min_links = min_links;
4426 params->lp_interval = BOND_ALB_DEFAULT_LP_INTERVAL; 4332 params->lp_interval = BOND_ALB_DEFAULT_LP_INTERVAL;
4427 4333 if (packets_per_slave > 1)
4334 params->packets_per_slave = reciprocal_value(packets_per_slave);
4335 else
4336 params->packets_per_slave = packets_per_slave;
4428 if (primary) { 4337 if (primary) {
4429 strncpy(params->primary, primary, IFNAMSIZ); 4338 strncpy(params->primary, primary, IFNAMSIZ);
4430 params->primary[IFNAMSIZ - 1] = 0; 4339 params->primary[IFNAMSIZ - 1] = 0;
@@ -4495,32 +4404,11 @@ static int bond_init(struct net_device *bond_dev)
4495 return 0; 4404 return 0;
4496} 4405}
4497 4406
4498static int bond_validate(struct nlattr *tb[], struct nlattr *data[]) 4407unsigned int bond_get_num_tx_queues(void)
4499{
4500 if (tb[IFLA_ADDRESS]) {
4501 if (nla_len(tb[IFLA_ADDRESS]) != ETH_ALEN)
4502 return -EINVAL;
4503 if (!is_valid_ether_addr(nla_data(tb[IFLA_ADDRESS])))
4504 return -EADDRNOTAVAIL;
4505 }
4506 return 0;
4507}
4508
4509static unsigned int bond_get_num_tx_queues(void)
4510{ 4408{
4511 return tx_queues; 4409 return tx_queues;
4512} 4410}
4513 4411
4514static struct rtnl_link_ops bond_link_ops __read_mostly = {
4515 .kind = "bond",
4516 .priv_size = sizeof(struct bonding),
4517 .setup = bond_setup,
4518 .validate = bond_validate,
4519 .get_num_tx_queues = bond_get_num_tx_queues,
4520 .get_num_rx_queues = bond_get_num_tx_queues, /* Use the same number
4521 as for TX queues */
4522};
4523
4524/* Create a new bond based on the specified name and bonding parameters. 4412/* Create a new bond based on the specified name and bonding parameters.
4525 * If name is NULL, obtain a suitable "bond%d" name for us. 4413 * If name is NULL, obtain a suitable "bond%d" name for us.
4526 * Caller must NOT hold rtnl_lock; we need to release it here before we 4414 * Caller must NOT hold rtnl_lock; we need to release it here before we
@@ -4607,7 +4495,7 @@ static int __init bonding_init(void)
4607 if (res) 4495 if (res)
4608 goto out; 4496 goto out;
4609 4497
4610 res = rtnl_link_register(&bond_link_ops); 4498 res = bond_netlink_init();
4611 if (res) 4499 if (res)
4612 goto err_link; 4500 goto err_link;
4613 4501
@@ -4623,7 +4511,7 @@ static int __init bonding_init(void)
4623out: 4511out:
4624 return res; 4512 return res;
4625err: 4513err:
4626 rtnl_link_unregister(&bond_link_ops); 4514 bond_netlink_fini();
4627err_link: 4515err_link:
4628 unregister_pernet_subsys(&bond_net_ops); 4516 unregister_pernet_subsys(&bond_net_ops);
4629 goto out; 4517 goto out;
@@ -4636,7 +4524,7 @@ static void __exit bonding_exit(void)
4636 4524
4637 bond_destroy_debugfs(); 4525 bond_destroy_debugfs();
4638 4526
4639 rtnl_link_unregister(&bond_link_ops); 4527 bond_netlink_fini();
4640 unregister_pernet_subsys(&bond_net_ops); 4528 unregister_pernet_subsys(&bond_net_ops);
4641 4529
4642#ifdef CONFIG_NET_POLL_CONTROLLER 4530#ifdef CONFIG_NET_POLL_CONTROLLER
@@ -4653,4 +4541,3 @@ MODULE_LICENSE("GPL");
4653MODULE_VERSION(DRV_VERSION); 4541MODULE_VERSION(DRV_VERSION);
4654MODULE_DESCRIPTION(DRV_DESCRIPTION ", v" DRV_VERSION); 4542MODULE_DESCRIPTION(DRV_DESCRIPTION ", v" DRV_VERSION);
4655MODULE_AUTHOR("Thomas Davis, tadavis@lbl.gov and many others"); 4543MODULE_AUTHOR("Thomas Davis, tadavis@lbl.gov and many others");
4656MODULE_ALIAS_RTNL_LINK("bond");
diff --git a/drivers/net/bonding/bond_netlink.c b/drivers/net/bonding/bond_netlink.c
new file mode 100644
index 000000000000..40e7b1cb4aea
--- /dev/null
+++ b/drivers/net/bonding/bond_netlink.c
@@ -0,0 +1,131 @@
1/*
2 * drivers/net/bond/bond_netlink.c - Netlink interface for bonding
3 * Copyright (c) 2013 Jiri Pirko <jiri@resnulli.us>
4 *
5 * This program is free software; you can redistribute it and/or modify
6 * it under the terms of the GNU General Public License as published by
7 * the Free Software Foundation; either version 2 of the License, or
8 * (at your option) any later version.
9 */
10
11#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
12
13#include <linux/module.h>
14#include <linux/errno.h>
15#include <linux/netdevice.h>
16#include <linux/etherdevice.h>
17#include <linux/if_link.h>
18#include <linux/if_ether.h>
19#include <net/netlink.h>
20#include <net/rtnetlink.h>
21#include "bonding.h"
22
23static const struct nla_policy bond_policy[IFLA_BOND_MAX + 1] = {
24 [IFLA_BOND_MODE] = { .type = NLA_U8 },
25 [IFLA_BOND_ACTIVE_SLAVE] = { .type = NLA_U32 },
26};
27
28static int bond_validate(struct nlattr *tb[], struct nlattr *data[])
29{
30 if (tb[IFLA_ADDRESS]) {
31 if (nla_len(tb[IFLA_ADDRESS]) != ETH_ALEN)
32 return -EINVAL;
33 if (!is_valid_ether_addr(nla_data(tb[IFLA_ADDRESS])))
34 return -EADDRNOTAVAIL;
35 }
36 return 0;
37}
38
39static int bond_changelink(struct net_device *bond_dev,
40 struct nlattr *tb[], struct nlattr *data[])
41{
42 struct bonding *bond = netdev_priv(bond_dev);
43 int err;
44
45 if (data && data[IFLA_BOND_MODE]) {
46 int mode = nla_get_u8(data[IFLA_BOND_MODE]);
47
48 err = bond_option_mode_set(bond, mode);
49 if (err)
50 return err;
51 }
52 if (data && data[IFLA_BOND_ACTIVE_SLAVE]) {
53 int ifindex = nla_get_u32(data[IFLA_BOND_ACTIVE_SLAVE]);
54 struct net_device *slave_dev;
55
56 if (ifindex == 0) {
57 slave_dev = NULL;
58 } else {
59 slave_dev = __dev_get_by_index(dev_net(bond_dev),
60 ifindex);
61 if (!slave_dev)
62 return -ENODEV;
63 }
64 err = bond_option_active_slave_set(bond, slave_dev);
65 if (err)
66 return err;
67 }
68 return 0;
69}
70
71static int bond_newlink(struct net *src_net, struct net_device *bond_dev,
72 struct nlattr *tb[], struct nlattr *data[])
73{
74 int err;
75
76 err = bond_changelink(bond_dev, tb, data);
77 if (err < 0)
78 return err;
79
80 return register_netdevice(bond_dev);
81}
82
83static size_t bond_get_size(const struct net_device *bond_dev)
84{
85 return nla_total_size(sizeof(u8)) + /* IFLA_BOND_MODE */
86 nla_total_size(sizeof(u32)); /* IFLA_BOND_ACTIVE_SLAVE */
87}
88
89static int bond_fill_info(struct sk_buff *skb,
90 const struct net_device *bond_dev)
91{
92 struct bonding *bond = netdev_priv(bond_dev);
93 struct net_device *slave_dev = bond_option_active_slave_get(bond);
94
95 if (nla_put_u8(skb, IFLA_BOND_MODE, bond->params.mode) ||
96 (slave_dev &&
97 nla_put_u32(skb, IFLA_BOND_ACTIVE_SLAVE, slave_dev->ifindex)))
98 goto nla_put_failure;
99 return 0;
100
101nla_put_failure:
102 return -EMSGSIZE;
103}
104
105struct rtnl_link_ops bond_link_ops __read_mostly = {
106 .kind = "bond",
107 .priv_size = sizeof(struct bonding),
108 .setup = bond_setup,
109 .maxtype = IFLA_BOND_MAX,
110 .policy = bond_policy,
111 .validate = bond_validate,
112 .newlink = bond_newlink,
113 .changelink = bond_changelink,
114 .get_size = bond_get_size,
115 .fill_info = bond_fill_info,
116 .get_num_tx_queues = bond_get_num_tx_queues,
117 .get_num_rx_queues = bond_get_num_tx_queues, /* Use the same number
118 as for TX queues */
119};
120
121int __init bond_netlink_init(void)
122{
123 return rtnl_link_register(&bond_link_ops);
124}
125
126void bond_netlink_fini(void)
127{
128 rtnl_link_unregister(&bond_link_ops);
129}
130
131MODULE_ALIAS_RTNL_LINK("bond");
diff --git a/drivers/net/bonding/bond_options.c b/drivers/net/bonding/bond_options.c
new file mode 100644
index 000000000000..9a5223c7b4d1
--- /dev/null
+++ b/drivers/net/bonding/bond_options.c
@@ -0,0 +1,142 @@
1/*
2 * drivers/net/bond/bond_options.c - bonding options
3 * Copyright (c) 2013 Jiri Pirko <jiri@resnulli.us>
4 *
5 * This program is free software; you can redistribute it and/or modify
6 * it under the terms of the GNU General Public License as published by
7 * the Free Software Foundation; either version 2 of the License, or
8 * (at your option) any later version.
9 */
10
11#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
12
13#include <linux/errno.h>
14#include <linux/if.h>
15#include <linux/netdevice.h>
16#include <linux/rwlock.h>
17#include <linux/rcupdate.h>
18#include "bonding.h"
19
20static bool bond_mode_is_valid(int mode)
21{
22 int i;
23
24 for (i = 0; bond_mode_tbl[i].modename; i++);
25
26 return mode >= 0 && mode < i;
27}
28
29int bond_option_mode_set(struct bonding *bond, int mode)
30{
31 if (!bond_mode_is_valid(mode)) {
32 pr_err("invalid mode value %d.\n", mode);
33 return -EINVAL;
34 }
35
36 if (bond->dev->flags & IFF_UP) {
37 pr_err("%s: unable to update mode because interface is up.\n",
38 bond->dev->name);
39 return -EPERM;
40 }
41
42 if (bond_has_slaves(bond)) {
43 pr_err("%s: unable to update mode because bond has slaves.\n",
44 bond->dev->name);
45 return -EPERM;
46 }
47
48 if (BOND_MODE_IS_LB(mode) && bond->params.arp_interval) {
49 pr_err("%s: %s mode is incompatible with arp monitoring.\n",
50 bond->dev->name, bond_mode_tbl[mode].modename);
51 return -EINVAL;
52 }
53
54 /* don't cache arp_validate between modes */
55 bond->params.arp_validate = BOND_ARP_VALIDATE_NONE;
56 bond->params.mode = mode;
57 return 0;
58}
59
60static struct net_device *__bond_option_active_slave_get(struct bonding *bond,
61 struct slave *slave)
62{
63 return USES_PRIMARY(bond->params.mode) && slave ? slave->dev : NULL;
64}
65
66struct net_device *bond_option_active_slave_get_rcu(struct bonding *bond)
67{
68 struct slave *slave = rcu_dereference(bond->curr_active_slave);
69
70 return __bond_option_active_slave_get(bond, slave);
71}
72
73struct net_device *bond_option_active_slave_get(struct bonding *bond)
74{
75 return __bond_option_active_slave_get(bond, bond->curr_active_slave);
76}
77
78int bond_option_active_slave_set(struct bonding *bond,
79 struct net_device *slave_dev)
80{
81 int ret = 0;
82
83 if (slave_dev) {
84 if (!netif_is_bond_slave(slave_dev)) {
85 pr_err("Device %s is not bonding slave.\n",
86 slave_dev->name);
87 return -EINVAL;
88 }
89
90 if (bond->dev != netdev_master_upper_dev_get(slave_dev)) {
91 pr_err("%s: Device %s is not our slave.\n",
92 bond->dev->name, slave_dev->name);
93 return -EINVAL;
94 }
95 }
96
97 if (!USES_PRIMARY(bond->params.mode)) {
98 pr_err("%s: Unable to change active slave; %s is in mode %d\n",
99 bond->dev->name, bond->dev->name, bond->params.mode);
100 return -EINVAL;
101 }
102
103 block_netpoll_tx();
104 read_lock(&bond->lock);
105 write_lock_bh(&bond->curr_slave_lock);
106
107 /* check to see if we are clearing active */
108 if (!slave_dev) {
109 pr_info("%s: Clearing current active slave.\n",
110 bond->dev->name);
111 rcu_assign_pointer(bond->curr_active_slave, NULL);
112 bond_select_active_slave(bond);
113 } else {
114 struct slave *old_active = bond->curr_active_slave;
115 struct slave *new_active = bond_slave_get_rtnl(slave_dev);
116
117 BUG_ON(!new_active);
118
119 if (new_active == old_active) {
120 /* do nothing */
121 pr_info("%s: %s is already the current active slave.\n",
122 bond->dev->name, new_active->dev->name);
123 } else {
124 if (old_active && (new_active->link == BOND_LINK_UP) &&
125 IS_UP(new_active->dev)) {
126 pr_info("%s: Setting %s as active slave.\n",
127 bond->dev->name, new_active->dev->name);
128 bond_change_active_slave(bond, new_active);
129 } else {
130 pr_err("%s: Could not set %s as active slave; either %s is down or the link is down.\n",
131 bond->dev->name, new_active->dev->name,
132 new_active->dev->name);
133 ret = -EINVAL;
134 }
135 }
136 }
137
138 write_unlock_bh(&bond->curr_slave_lock);
139 read_unlock(&bond->lock);
140 unblock_netpoll_tx();
141 return ret;
142}
diff --git a/drivers/net/bonding/bond_procfs.c b/drivers/net/bonding/bond_procfs.c
index 20a6ee25bb63..fb868d6c22da 100644
--- a/drivers/net/bonding/bond_procfs.c
+++ b/drivers/net/bonding/bond_procfs.c
@@ -10,8 +10,9 @@ static void *bond_info_seq_start(struct seq_file *seq, loff_t *pos)
10 __acquires(&bond->lock) 10 __acquires(&bond->lock)
11{ 11{
12 struct bonding *bond = seq->private; 12 struct bonding *bond = seq->private;
13 loff_t off = 0; 13 struct list_head *iter;
14 struct slave *slave; 14 struct slave *slave;
15 loff_t off = 0;
15 16
16 /* make sure the bond won't be taken away */ 17 /* make sure the bond won't be taken away */
17 rcu_read_lock(); 18 rcu_read_lock();
@@ -20,7 +21,7 @@ static void *bond_info_seq_start(struct seq_file *seq, loff_t *pos)
20 if (*pos == 0) 21 if (*pos == 0)
21 return SEQ_START_TOKEN; 22 return SEQ_START_TOKEN;
22 23
23 bond_for_each_slave(bond, slave) 24 bond_for_each_slave(bond, slave, iter)
24 if (++off == *pos) 25 if (++off == *pos)
25 return slave; 26 return slave;
26 27
@@ -30,17 +31,25 @@ static void *bond_info_seq_start(struct seq_file *seq, loff_t *pos)
30static void *bond_info_seq_next(struct seq_file *seq, void *v, loff_t *pos) 31static void *bond_info_seq_next(struct seq_file *seq, void *v, loff_t *pos)
31{ 32{
32 struct bonding *bond = seq->private; 33 struct bonding *bond = seq->private;
33 struct slave *slave = v; 34 struct list_head *iter;
35 struct slave *slave;
36 bool found = false;
34 37
35 ++*pos; 38 ++*pos;
36 if (v == SEQ_START_TOKEN) 39 if (v == SEQ_START_TOKEN)
37 return bond_first_slave(bond); 40 return bond_first_slave(bond);
38 41
39 if (bond_is_last_slave(bond, slave)) 42 if (bond_is_last_slave(bond, v))
40 return NULL; 43 return NULL;
41 slave = bond_next_slave(bond, slave);
42 44
43 return slave; 45 bond_for_each_slave(bond, slave, iter) {
46 if (found)
47 return slave;
48 if (slave == v)
49 found = true;
50 }
51
52 return NULL;
44} 53}
45 54
46static void bond_info_seq_stop(struct seq_file *seq, void *v) 55static void bond_info_seq_stop(struct seq_file *seq, void *v)
diff --git a/drivers/net/bonding/bond_sysfs.c b/drivers/net/bonding/bond_sysfs.c
index ec9b6460a38d..bc8fd362a5aa 100644
--- a/drivers/net/bonding/bond_sysfs.c
+++ b/drivers/net/bonding/bond_sysfs.c
@@ -40,6 +40,7 @@
40#include <net/net_namespace.h> 40#include <net/net_namespace.h>
41#include <net/netns/generic.h> 41#include <net/netns/generic.h>
42#include <linux/nsproxy.h> 42#include <linux/nsproxy.h>
43#include <linux/reciprocal_div.h>
43 44
44#include "bonding.h" 45#include "bonding.h"
45 46
@@ -159,41 +160,6 @@ static const struct class_attribute class_attr_bonding_masters = {
159 .store = bonding_store_bonds, 160 .store = bonding_store_bonds,
160}; 161};
161 162
162int bond_create_slave_symlinks(struct net_device *master,
163 struct net_device *slave)
164{
165 char linkname[IFNAMSIZ+7];
166 int ret = 0;
167
168 /* first, create a link from the slave back to the master */
169 ret = sysfs_create_link(&(slave->dev.kobj), &(master->dev.kobj),
170 "master");
171 if (ret)
172 return ret;
173 /* next, create a link from the master to the slave */
174 sprintf(linkname, "slave_%s", slave->name);
175 ret = sysfs_create_link(&(master->dev.kobj), &(slave->dev.kobj),
176 linkname);
177
178 /* free the master link created earlier in case of error */
179 if (ret)
180 sysfs_remove_link(&(slave->dev.kobj), "master");
181
182 return ret;
183
184}
185
186void bond_destroy_slave_symlinks(struct net_device *master,
187 struct net_device *slave)
188{
189 char linkname[IFNAMSIZ+7];
190
191 sysfs_remove_link(&(slave->dev.kobj), "master");
192 sprintf(linkname, "slave_%s", slave->name);
193 sysfs_remove_link(&(master->dev.kobj), linkname);
194}
195
196
197/* 163/*
198 * Show the slaves in the current bond. 164 * Show the slaves in the current bond.
199 */ 165 */
@@ -201,11 +167,14 @@ static ssize_t bonding_show_slaves(struct device *d,
201 struct device_attribute *attr, char *buf) 167 struct device_attribute *attr, char *buf)
202{ 168{
203 struct bonding *bond = to_bond(d); 169 struct bonding *bond = to_bond(d);
170 struct list_head *iter;
204 struct slave *slave; 171 struct slave *slave;
205 int res = 0; 172 int res = 0;
206 173
207 read_lock(&bond->lock); 174 if (!rtnl_trylock())
208 bond_for_each_slave(bond, slave) { 175 return restart_syscall();
176
177 bond_for_each_slave(bond, slave, iter) {
209 if (res > (PAGE_SIZE - IFNAMSIZ)) { 178 if (res > (PAGE_SIZE - IFNAMSIZ)) {
210 /* not enough space for another interface name */ 179 /* not enough space for another interface name */
211 if ((PAGE_SIZE - res) > 10) 180 if ((PAGE_SIZE - res) > 10)
@@ -215,7 +184,9 @@ static ssize_t bonding_show_slaves(struct device *d,
215 } 184 }
216 res += sprintf(buf + res, "%s ", slave->dev->name); 185 res += sprintf(buf + res, "%s ", slave->dev->name);
217 } 186 }
218 read_unlock(&bond->lock); 187
188 rtnl_unlock();
189
219 if (res) 190 if (res)
220 buf[res-1] = '\n'; /* eat the leftover space */ 191 buf[res-1] = '\n'; /* eat the leftover space */
221 192
@@ -304,50 +275,26 @@ static ssize_t bonding_store_mode(struct device *d,
304 struct device_attribute *attr, 275 struct device_attribute *attr,
305 const char *buf, size_t count) 276 const char *buf, size_t count)
306{ 277{
307 int new_value, ret = count; 278 int new_value, ret;
308 struct bonding *bond = to_bond(d); 279 struct bonding *bond = to_bond(d);
309 280
310 if (!rtnl_trylock())
311 return restart_syscall();
312
313 if (bond->dev->flags & IFF_UP) {
314 pr_err("unable to update mode of %s because interface is up.\n",
315 bond->dev->name);
316 ret = -EPERM;
317 goto out;
318 }
319
320 if (!list_empty(&bond->slave_list)) {
321 pr_err("unable to update mode of %s because it has slaves.\n",
322 bond->dev->name);
323 ret = -EPERM;
324 goto out;
325 }
326
327 new_value = bond_parse_parm(buf, bond_mode_tbl); 281 new_value = bond_parse_parm(buf, bond_mode_tbl);
328 if (new_value < 0) { 282 if (new_value < 0) {
329 pr_err("%s: Ignoring invalid mode value %.*s.\n", 283 pr_err("%s: Ignoring invalid mode value %.*s.\n",
330 bond->dev->name, (int)strlen(buf) - 1, buf); 284 bond->dev->name, (int)strlen(buf) - 1, buf);
331 ret = -EINVAL; 285 return -EINVAL;
332 goto out;
333 } 286 }
334 if ((new_value == BOND_MODE_ALB || 287 if (!rtnl_trylock())
335 new_value == BOND_MODE_TLB) && 288 return restart_syscall();
336 bond->params.arp_interval) { 289
337 pr_err("%s: %s mode is incompatible with arp monitoring.\n", 290 ret = bond_option_mode_set(bond, new_value);
338 bond->dev->name, bond_mode_tbl[new_value].modename); 291 if (!ret) {
339 ret = -EINVAL; 292 pr_info("%s: setting mode to %s (%d).\n",
340 goto out; 293 bond->dev->name, bond_mode_tbl[new_value].modename,
294 new_value);
295 ret = count;
341 } 296 }
342 297
343 /* don't cache arp_validate between modes */
344 bond->params.arp_validate = BOND_ARP_VALIDATE_NONE;
345 bond->params.mode = new_value;
346 bond_set_mode_ops(bond, bond->params.mode);
347 pr_info("%s: setting mode to %s (%d).\n",
348 bond->dev->name, bond_mode_tbl[new_value].modename,
349 new_value);
350out:
351 rtnl_unlock(); 298 rtnl_unlock();
352 return ret; 299 return ret;
353} 300}
@@ -383,7 +330,6 @@ static ssize_t bonding_store_xmit_hash(struct device *d,
383 ret = -EINVAL; 330 ret = -EINVAL;
384 } else { 331 } else {
385 bond->params.xmit_policy = new_value; 332 bond->params.xmit_policy = new_value;
386 bond_set_mode_ops(bond, bond->params.mode);
387 pr_info("%s: setting xmit hash policy to %s (%d).\n", 333 pr_info("%s: setting xmit hash policy to %s (%d).\n",
388 bond->dev->name, 334 bond->dev->name,
389 xmit_hashtype_tbl[new_value].modename, new_value); 335 xmit_hashtype_tbl[new_value].modename, new_value);
@@ -513,7 +459,7 @@ static ssize_t bonding_store_fail_over_mac(struct device *d,
513 if (!rtnl_trylock()) 459 if (!rtnl_trylock())
514 return restart_syscall(); 460 return restart_syscall();
515 461
516 if (!list_empty(&bond->slave_list)) { 462 if (bond_has_slaves(bond)) {
517 pr_err("%s: Can't alter fail_over_mac with slaves in bond.\n", 463 pr_err("%s: Can't alter fail_over_mac with slaves in bond.\n",
518 bond->dev->name); 464 bond->dev->name);
519 ret = -EPERM; 465 ret = -EPERM;
@@ -647,11 +593,15 @@ static ssize_t bonding_store_arp_targets(struct device *d,
647 const char *buf, size_t count) 593 const char *buf, size_t count)
648{ 594{
649 struct bonding *bond = to_bond(d); 595 struct bonding *bond = to_bond(d);
596 struct list_head *iter;
650 struct slave *slave; 597 struct slave *slave;
651 __be32 newtarget, *targets; 598 __be32 newtarget, *targets;
652 unsigned long *targets_rx; 599 unsigned long *targets_rx;
653 int ind, i, j, ret = -EINVAL; 600 int ind, i, j, ret = -EINVAL;
654 601
602 if (!rtnl_trylock())
603 return restart_syscall();
604
655 targets = bond->params.arp_targets; 605 targets = bond->params.arp_targets;
656 newtarget = in_aton(buf + 1); 606 newtarget = in_aton(buf + 1);
657 /* look for adds */ 607 /* look for adds */
@@ -679,7 +629,7 @@ static ssize_t bonding_store_arp_targets(struct device *d,
679 &newtarget); 629 &newtarget);
680 /* not to race with bond_arp_rcv */ 630 /* not to race with bond_arp_rcv */
681 write_lock_bh(&bond->lock); 631 write_lock_bh(&bond->lock);
682 bond_for_each_slave(bond, slave) 632 bond_for_each_slave(bond, slave, iter)
683 slave->target_last_arp_rx[ind] = jiffies; 633 slave->target_last_arp_rx[ind] = jiffies;
684 targets[ind] = newtarget; 634 targets[ind] = newtarget;
685 write_unlock_bh(&bond->lock); 635 write_unlock_bh(&bond->lock);
@@ -705,7 +655,7 @@ static ssize_t bonding_store_arp_targets(struct device *d,
705 &newtarget); 655 &newtarget);
706 656
707 write_lock_bh(&bond->lock); 657 write_lock_bh(&bond->lock);
708 bond_for_each_slave(bond, slave) { 658 bond_for_each_slave(bond, slave, iter) {
709 targets_rx = slave->target_last_arp_rx; 659 targets_rx = slave->target_last_arp_rx;
710 j = ind; 660 j = ind;
711 for (; (j < BOND_MAX_ARP_TARGETS-1) && targets[j+1]; j++) 661 for (; (j < BOND_MAX_ARP_TARGETS-1) && targets[j+1]; j++)
@@ -725,6 +675,7 @@ static ssize_t bonding_store_arp_targets(struct device *d,
725 675
726 ret = count; 676 ret = count;
727out: 677out:
678 rtnl_unlock();
728 return ret; 679 return ret;
729} 680}
730static DEVICE_ATTR(arp_ip_target, S_IRUGO | S_IWUSR , bonding_show_arp_targets, bonding_store_arp_targets); 681static DEVICE_ATTR(arp_ip_target, S_IRUGO | S_IWUSR , bonding_show_arp_targets, bonding_store_arp_targets);
@@ -1102,6 +1053,7 @@ static ssize_t bonding_store_primary(struct device *d,
1102 const char *buf, size_t count) 1053 const char *buf, size_t count)
1103{ 1054{
1104 struct bonding *bond = to_bond(d); 1055 struct bonding *bond = to_bond(d);
1056 struct list_head *iter;
1105 char ifname[IFNAMSIZ]; 1057 char ifname[IFNAMSIZ];
1106 struct slave *slave; 1058 struct slave *slave;
1107 1059
@@ -1129,7 +1081,7 @@ static ssize_t bonding_store_primary(struct device *d,
1129 goto out; 1081 goto out;
1130 } 1082 }
1131 1083
1132 bond_for_each_slave(bond, slave) { 1084 bond_for_each_slave(bond, slave, iter) {
1133 if (strncmp(slave->dev->name, ifname, IFNAMSIZ) == 0) { 1085 if (strncmp(slave->dev->name, ifname, IFNAMSIZ) == 0) {
1134 pr_info("%s: Setting %s as primary slave.\n", 1086 pr_info("%s: Setting %s as primary slave.\n",
1135 bond->dev->name, slave->dev->name); 1087 bond->dev->name, slave->dev->name);
@@ -1259,13 +1211,13 @@ static ssize_t bonding_show_active_slave(struct device *d,
1259 char *buf) 1211 char *buf)
1260{ 1212{
1261 struct bonding *bond = to_bond(d); 1213 struct bonding *bond = to_bond(d);
1262 struct slave *curr; 1214 struct net_device *slave_dev;
1263 int count = 0; 1215 int count = 0;
1264 1216
1265 rcu_read_lock(); 1217 rcu_read_lock();
1266 curr = rcu_dereference(bond->curr_active_slave); 1218 slave_dev = bond_option_active_slave_get_rcu(bond);
1267 if (USES_PRIMARY(bond->params.mode) && curr) 1219 if (slave_dev)
1268 count = sprintf(buf, "%s\n", curr->dev->name); 1220 count = sprintf(buf, "%s\n", slave_dev->name);
1269 rcu_read_unlock(); 1221 rcu_read_unlock();
1270 1222
1271 return count; 1223 return count;
@@ -1275,80 +1227,33 @@ static ssize_t bonding_store_active_slave(struct device *d,
1275 struct device_attribute *attr, 1227 struct device_attribute *attr,
1276 const char *buf, size_t count) 1228 const char *buf, size_t count)
1277{ 1229{
1278 struct slave *slave, *old_active, *new_active; 1230 int ret;
1279 struct bonding *bond = to_bond(d); 1231 struct bonding *bond = to_bond(d);
1280 char ifname[IFNAMSIZ]; 1232 char ifname[IFNAMSIZ];
1233 struct net_device *dev;
1281 1234
1282 if (!rtnl_trylock()) 1235 if (!rtnl_trylock())
1283 return restart_syscall(); 1236 return restart_syscall();
1284 1237
1285 old_active = new_active = NULL;
1286 block_netpoll_tx();
1287 read_lock(&bond->lock);
1288 write_lock_bh(&bond->curr_slave_lock);
1289
1290 if (!USES_PRIMARY(bond->params.mode)) {
1291 pr_info("%s: Unable to change active slave; %s is in mode %d\n",
1292 bond->dev->name, bond->dev->name, bond->params.mode);
1293 goto out;
1294 }
1295
1296 sscanf(buf, "%15s", ifname); /* IFNAMSIZ */ 1238 sscanf(buf, "%15s", ifname); /* IFNAMSIZ */
1297
1298 /* check to see if we are clearing active */
1299 if (!strlen(ifname) || buf[0] == '\n') { 1239 if (!strlen(ifname) || buf[0] == '\n') {
1300 pr_info("%s: Clearing current active slave.\n", 1240 dev = NULL;
1301 bond->dev->name); 1241 } else {
1302 rcu_assign_pointer(bond->curr_active_slave, NULL); 1242 dev = __dev_get_by_name(dev_net(bond->dev), ifname);
1303 bond_select_active_slave(bond); 1243 if (!dev) {
1304 goto out; 1244 ret = -ENODEV;
1305 } 1245 goto out;
1306
1307 bond_for_each_slave(bond, slave) {
1308 if (strncmp(slave->dev->name, ifname, IFNAMSIZ) == 0) {
1309 old_active = bond->curr_active_slave;
1310 new_active = slave;
1311 if (new_active == old_active) {
1312 /* do nothing */
1313 pr_info("%s: %s is already the current"
1314 " active slave.\n",
1315 bond->dev->name,
1316 slave->dev->name);
1317 goto out;
1318 } else {
1319 if ((new_active) &&
1320 (old_active) &&
1321 (new_active->link == BOND_LINK_UP) &&
1322 IS_UP(new_active->dev)) {
1323 pr_info("%s: Setting %s as active"
1324 " slave.\n",
1325 bond->dev->name,
1326 slave->dev->name);
1327 bond_change_active_slave(bond,
1328 new_active);
1329 } else {
1330 pr_info("%s: Could not set %s as"
1331 " active slave; either %s is"
1332 " down or the link is down.\n",
1333 bond->dev->name,
1334 slave->dev->name,
1335 slave->dev->name);
1336 }
1337 goto out;
1338 }
1339 } 1246 }
1340 } 1247 }
1341 1248
1342 pr_info("%s: Unable to set %.*s as active slave.\n", 1249 ret = bond_option_active_slave_set(bond, dev);
1343 bond->dev->name, (int)strlen(buf) - 1, buf); 1250 if (!ret)
1344 out: 1251 ret = count;
1345 write_unlock_bh(&bond->curr_slave_lock);
1346 read_unlock(&bond->lock);
1347 unblock_netpoll_tx();
1348 1252
1253 out:
1349 rtnl_unlock(); 1254 rtnl_unlock();
1350 1255
1351 return count; 1256 return ret;
1352 1257
1353} 1258}
1354static DEVICE_ATTR(active_slave, S_IRUGO | S_IWUSR, 1259static DEVICE_ATTR(active_slave, S_IRUGO | S_IWUSR,
@@ -1484,14 +1389,14 @@ static ssize_t bonding_show_queue_id(struct device *d,
1484 char *buf) 1389 char *buf)
1485{ 1390{
1486 struct bonding *bond = to_bond(d); 1391 struct bonding *bond = to_bond(d);
1392 struct list_head *iter;
1487 struct slave *slave; 1393 struct slave *slave;
1488 int res = 0; 1394 int res = 0;
1489 1395
1490 if (!rtnl_trylock()) 1396 if (!rtnl_trylock())
1491 return restart_syscall(); 1397 return restart_syscall();
1492 1398
1493 read_lock(&bond->lock); 1399 bond_for_each_slave(bond, slave, iter) {
1494 bond_for_each_slave(bond, slave) {
1495 if (res > (PAGE_SIZE - IFNAMSIZ - 6)) { 1400 if (res > (PAGE_SIZE - IFNAMSIZ - 6)) {
1496 /* not enough space for another interface_name:queue_id pair */ 1401 /* not enough space for another interface_name:queue_id pair */
1497 if ((PAGE_SIZE - res) > 10) 1402 if ((PAGE_SIZE - res) > 10)
@@ -1502,9 +1407,9 @@ static ssize_t bonding_show_queue_id(struct device *d,
1502 res += sprintf(buf + res, "%s:%d ", 1407 res += sprintf(buf + res, "%s:%d ",
1503 slave->dev->name, slave->queue_id); 1408 slave->dev->name, slave->queue_id);
1504 } 1409 }
1505 read_unlock(&bond->lock);
1506 if (res) 1410 if (res)
1507 buf[res-1] = '\n'; /* eat the leftover space */ 1411 buf[res-1] = '\n'; /* eat the leftover space */
1412
1508 rtnl_unlock(); 1413 rtnl_unlock();
1509 1414
1510 return res; 1415 return res;
@@ -1520,6 +1425,7 @@ static ssize_t bonding_store_queue_id(struct device *d,
1520{ 1425{
1521 struct slave *slave, *update_slave; 1426 struct slave *slave, *update_slave;
1522 struct bonding *bond = to_bond(d); 1427 struct bonding *bond = to_bond(d);
1428 struct list_head *iter;
1523 u16 qid; 1429 u16 qid;
1524 int ret = count; 1430 int ret = count;
1525 char *delim; 1431 char *delim;
@@ -1552,11 +1458,9 @@ static ssize_t bonding_store_queue_id(struct device *d,
1552 if (!sdev) 1458 if (!sdev)
1553 goto err_no_cmd; 1459 goto err_no_cmd;
1554 1460
1555 read_lock(&bond->lock);
1556
1557 /* Search for thes slave and check for duplicate qids */ 1461 /* Search for thes slave and check for duplicate qids */
1558 update_slave = NULL; 1462 update_slave = NULL;
1559 bond_for_each_slave(bond, slave) { 1463 bond_for_each_slave(bond, slave, iter) {
1560 if (sdev == slave->dev) 1464 if (sdev == slave->dev)
1561 /* 1465 /*
1562 * We don't need to check the matching 1466 * We don't need to check the matching
@@ -1564,23 +1468,20 @@ static ssize_t bonding_store_queue_id(struct device *d,
1564 */ 1468 */
1565 update_slave = slave; 1469 update_slave = slave;
1566 else if (qid && qid == slave->queue_id) { 1470 else if (qid && qid == slave->queue_id) {
1567 goto err_no_cmd_unlock; 1471 goto err_no_cmd;
1568 } 1472 }
1569 } 1473 }
1570 1474
1571 if (!update_slave) 1475 if (!update_slave)
1572 goto err_no_cmd_unlock; 1476 goto err_no_cmd;
1573 1477
1574 /* Actually set the qids for the slave */ 1478 /* Actually set the qids for the slave */
1575 update_slave->queue_id = qid; 1479 update_slave->queue_id = qid;
1576 1480
1577 read_unlock(&bond->lock);
1578out: 1481out:
1579 rtnl_unlock(); 1482 rtnl_unlock();
1580 return ret; 1483 return ret;
1581 1484
1582err_no_cmd_unlock:
1583 read_unlock(&bond->lock);
1584err_no_cmd: 1485err_no_cmd:
1585 pr_info("invalid input for queue_id set for %s.\n", 1486 pr_info("invalid input for queue_id set for %s.\n",
1586 bond->dev->name); 1487 bond->dev->name);
@@ -1610,8 +1511,12 @@ static ssize_t bonding_store_slaves_active(struct device *d,
1610{ 1511{
1611 struct bonding *bond = to_bond(d); 1512 struct bonding *bond = to_bond(d);
1612 int new_value, ret = count; 1513 int new_value, ret = count;
1514 struct list_head *iter;
1613 struct slave *slave; 1515 struct slave *slave;
1614 1516
1517 if (!rtnl_trylock())
1518 return restart_syscall();
1519
1615 if (sscanf(buf, "%d", &new_value) != 1) { 1520 if (sscanf(buf, "%d", &new_value) != 1) {
1616 pr_err("%s: no all_slaves_active value specified.\n", 1521 pr_err("%s: no all_slaves_active value specified.\n",
1617 bond->dev->name); 1522 bond->dev->name);
@@ -1631,8 +1536,7 @@ static ssize_t bonding_store_slaves_active(struct device *d,
1631 goto out; 1536 goto out;
1632 } 1537 }
1633 1538
1634 read_lock(&bond->lock); 1539 bond_for_each_slave(bond, slave, iter) {
1635 bond_for_each_slave(bond, slave) {
1636 if (!bond_is_active_slave(slave)) { 1540 if (!bond_is_active_slave(slave)) {
1637 if (new_value) 1541 if (new_value)
1638 slave->inactive = 0; 1542 slave->inactive = 0;
@@ -1640,8 +1544,8 @@ static ssize_t bonding_store_slaves_active(struct device *d,
1640 slave->inactive = 1; 1544 slave->inactive = 1;
1641 } 1545 }
1642 } 1546 }
1643 read_unlock(&bond->lock);
1644out: 1547out:
1548 rtnl_unlock();
1645 return ret; 1549 return ret;
1646} 1550}
1647static DEVICE_ATTR(all_slaves_active, S_IRUGO | S_IWUSR, 1551static DEVICE_ATTR(all_slaves_active, S_IRUGO | S_IWUSR,
@@ -1728,6 +1632,53 @@ out:
1728static DEVICE_ATTR(lp_interval, S_IRUGO | S_IWUSR, 1632static DEVICE_ATTR(lp_interval, S_IRUGO | S_IWUSR,
1729 bonding_show_lp_interval, bonding_store_lp_interval); 1633 bonding_show_lp_interval, bonding_store_lp_interval);
1730 1634
1635static ssize_t bonding_show_packets_per_slave(struct device *d,
1636 struct device_attribute *attr,
1637 char *buf)
1638{
1639 struct bonding *bond = to_bond(d);
1640 int packets_per_slave = bond->params.packets_per_slave;
1641
1642 if (packets_per_slave > 1)
1643 packets_per_slave = reciprocal_value(packets_per_slave);
1644
1645 return sprintf(buf, "%d\n", packets_per_slave);
1646}
1647
1648static ssize_t bonding_store_packets_per_slave(struct device *d,
1649 struct device_attribute *attr,
1650 const char *buf, size_t count)
1651{
1652 struct bonding *bond = to_bond(d);
1653 int new_value, ret = count;
1654
1655 if (sscanf(buf, "%d", &new_value) != 1) {
1656 pr_err("%s: no packets_per_slave value specified.\n",
1657 bond->dev->name);
1658 ret = -EINVAL;
1659 goto out;
1660 }
1661 if (new_value < 0 || new_value > USHRT_MAX) {
1662 pr_err("%s: packets_per_slave must be between 0 and %u\n",
1663 bond->dev->name, USHRT_MAX);
1664 ret = -EINVAL;
1665 goto out;
1666 }
1667 if (bond->params.mode != BOND_MODE_ROUNDROBIN)
1668 pr_warn("%s: Warning: packets_per_slave has effect only in balance-rr mode\n",
1669 bond->dev->name);
1670 if (new_value > 1)
1671 bond->params.packets_per_slave = reciprocal_value(new_value);
1672 else
1673 bond->params.packets_per_slave = new_value;
1674out:
1675 return ret;
1676}
1677
1678static DEVICE_ATTR(packets_per_slave, S_IRUGO | S_IWUSR,
1679 bonding_show_packets_per_slave,
1680 bonding_store_packets_per_slave);
1681
1731static struct attribute *per_bond_attrs[] = { 1682static struct attribute *per_bond_attrs[] = {
1732 &dev_attr_slaves.attr, 1683 &dev_attr_slaves.attr,
1733 &dev_attr_mode.attr, 1684 &dev_attr_mode.attr,
@@ -1759,6 +1710,7 @@ static struct attribute *per_bond_attrs[] = {
1759 &dev_attr_resend_igmp.attr, 1710 &dev_attr_resend_igmp.attr,
1760 &dev_attr_min_links.attr, 1711 &dev_attr_min_links.attr,
1761 &dev_attr_lp_interval.attr, 1712 &dev_attr_lp_interval.attr,
1713 &dev_attr_packets_per_slave.attr,
1762 NULL, 1714 NULL,
1763}; 1715};
1764 1716
diff --git a/drivers/net/bonding/bonding.h b/drivers/net/bonding/bonding.h
index 03cf3fd14490..77a07a12e77f 100644
--- a/drivers/net/bonding/bonding.h
+++ b/drivers/net/bonding/bonding.h
@@ -58,6 +58,11 @@
58#define TX_QUEUE_OVERRIDE(mode) \ 58#define TX_QUEUE_OVERRIDE(mode) \
59 (((mode) == BOND_MODE_ACTIVEBACKUP) || \ 59 (((mode) == BOND_MODE_ACTIVEBACKUP) || \
60 ((mode) == BOND_MODE_ROUNDROBIN)) 60 ((mode) == BOND_MODE_ROUNDROBIN))
61
62#define BOND_MODE_IS_LB(mode) \
63 (((mode) == BOND_MODE_TLB) || \
64 ((mode) == BOND_MODE_ALB))
65
61/* 66/*
62 * Less bad way to call ioctl from within the kernel; this needs to be 67 * Less bad way to call ioctl from within the kernel; this needs to be
63 * done some other way to get the call out of interrupt context. 68 * done some other way to get the call out of interrupt context.
@@ -72,63 +77,37 @@
72 res; }) 77 res; })
73 78
74/* slave list primitives */ 79/* slave list primitives */
75#define bond_to_slave(ptr) list_entry(ptr, struct slave, list) 80#define bond_slave_list(bond) (&(bond)->dev->adj_list.lower)
81
82#define bond_has_slaves(bond) !list_empty(bond_slave_list(bond))
76 83
77/* IMPORTANT: bond_first/last_slave can return NULL in case of an empty list */ 84/* IMPORTANT: bond_first/last_slave can return NULL in case of an empty list */
78#define bond_first_slave(bond) \ 85#define bond_first_slave(bond) \
79 list_first_entry_or_null(&(bond)->slave_list, struct slave, list) 86 (bond_has_slaves(bond) ? \
87 netdev_adjacent_get_private(bond_slave_list(bond)->next) : \
88 NULL)
80#define bond_last_slave(bond) \ 89#define bond_last_slave(bond) \
81 (list_empty(&(bond)->slave_list) ? NULL : \ 90 (bond_has_slaves(bond) ? \
82 bond_to_slave((bond)->slave_list.prev)) 91 netdev_adjacent_get_private(bond_slave_list(bond)->prev) : \
92 NULL)
83 93
84#define bond_is_first_slave(bond, pos) ((pos)->list.prev == &(bond)->slave_list) 94#define bond_is_first_slave(bond, pos) (pos == bond_first_slave(bond))
85#define bond_is_last_slave(bond, pos) ((pos)->list.next == &(bond)->slave_list) 95#define bond_is_last_slave(bond, pos) (pos == bond_last_slave(bond))
86
87/* Since bond_first/last_slave can return NULL, these can return NULL too */
88#define bond_next_slave(bond, pos) \
89 (bond_is_last_slave(bond, pos) ? bond_first_slave(bond) : \
90 bond_to_slave((pos)->list.next))
91
92#define bond_prev_slave(bond, pos) \
93 (bond_is_first_slave(bond, pos) ? bond_last_slave(bond) : \
94 bond_to_slave((pos)->list.prev))
95
96/**
97 * bond_for_each_slave_from - iterate the slaves list from a starting point
98 * @bond: the bond holding this list.
99 * @pos: current slave.
100 * @cnt: counter for max number of moves
101 * @start: starting point.
102 *
103 * Caller must hold bond->lock
104 */
105#define bond_for_each_slave_from(bond, pos, cnt, start) \
106 for (cnt = 0, pos = start; pos && cnt < (bond)->slave_cnt; \
107 cnt++, pos = bond_next_slave(bond, pos))
108 96
109/** 97/**
110 * bond_for_each_slave - iterate over all slaves 98 * bond_for_each_slave - iterate over all slaves
111 * @bond: the bond holding this list 99 * @bond: the bond holding this list
112 * @pos: current slave 100 * @pos: current slave
101 * @iter: list_head * iterator
113 * 102 *
114 * Caller must hold bond->lock 103 * Caller must hold bond->lock
115 */ 104 */
116#define bond_for_each_slave(bond, pos) \ 105#define bond_for_each_slave(bond, pos, iter) \
117 list_for_each_entry(pos, &(bond)->slave_list, list) 106 netdev_for_each_lower_private((bond)->dev, pos, iter)
118 107
119/* Caller must have rcu_read_lock */ 108/* Caller must have rcu_read_lock */
120#define bond_for_each_slave_rcu(bond, pos) \ 109#define bond_for_each_slave_rcu(bond, pos, iter) \
121 list_for_each_entry_rcu(pos, &(bond)->slave_list, list) 110 netdev_for_each_lower_private_rcu((bond)->dev, pos, iter)
122
123/**
124 * bond_for_each_slave_reverse - iterate in reverse from a given position
125 * @bond: the bond holding this list
126 * @pos: slave to continue from
127 *
128 * Caller must hold bond->lock
129 */
130#define bond_for_each_slave_continue_reverse(bond, pos) \
131 list_for_each_entry_continue_reverse(pos, &(bond)->slave_list, list)
132 111
133#ifdef CONFIG_NET_POLL_CONTROLLER 112#ifdef CONFIG_NET_POLL_CONTROLLER
134extern atomic_t netpoll_block_tx; 113extern atomic_t netpoll_block_tx;
@@ -177,6 +156,7 @@ struct bond_params {
177 int all_slaves_active; 156 int all_slaves_active;
178 int resend_igmp; 157 int resend_igmp;
179 int lp_interval; 158 int lp_interval;
159 int packets_per_slave;
180}; 160};
181 161
182struct bond_parm_tbl { 162struct bond_parm_tbl {
@@ -188,7 +168,6 @@ struct bond_parm_tbl {
188 168
189struct slave { 169struct slave {
190 struct net_device *dev; /* first - useful for panic debug */ 170 struct net_device *dev; /* first - useful for panic debug */
191 struct list_head list;
192 struct bonding *bond; /* our master */ 171 struct bonding *bond; /* our master */
193 int delay; 172 int delay;
194 unsigned long jiffies; 173 unsigned long jiffies;
@@ -228,7 +207,6 @@ struct slave {
228 */ 207 */
229struct bonding { 208struct bonding {
230 struct net_device *dev; /* first - useful for panic debug */ 209 struct net_device *dev; /* first - useful for panic debug */
231 struct list_head slave_list;
232 struct slave *curr_active_slave; 210 struct slave *curr_active_slave;
233 struct slave *current_arp_slave; 211 struct slave *current_arp_slave;
234 struct slave *primary_slave; 212 struct slave *primary_slave;
@@ -245,8 +223,7 @@ struct bonding {
245 char proc_file_name[IFNAMSIZ]; 223 char proc_file_name[IFNAMSIZ];
246#endif /* CONFIG_PROC_FS */ 224#endif /* CONFIG_PROC_FS */
247 struct list_head bond_list; 225 struct list_head bond_list;
248 int (*xmit_hash_policy)(struct sk_buff *, int); 226 u32 rr_tx_counter;
249 u16 rr_tx_counter;
250 struct ad_bond_info ad_info; 227 struct ad_bond_info ad_info;
251 struct alb_bond_info alb_info; 228 struct alb_bond_info alb_info;
252 struct bond_params params; 229 struct bond_params params;
@@ -276,13 +253,7 @@ struct bonding {
276static inline struct slave *bond_get_slave_by_dev(struct bonding *bond, 253static inline struct slave *bond_get_slave_by_dev(struct bonding *bond,
277 struct net_device *slave_dev) 254 struct net_device *slave_dev)
278{ 255{
279 struct slave *slave = NULL; 256 return netdev_lower_dev_get_private(bond->dev, slave_dev);
280
281 bond_for_each_slave(bond, slave)
282 if (slave->dev == slave_dev)
283 return slave;
284
285 return NULL;
286} 257}
287 258
288static inline struct bonding *bond_get_bond_by_slave(struct slave *slave) 259static inline struct bonding *bond_get_bond_by_slave(struct slave *slave)
@@ -294,8 +265,7 @@ static inline struct bonding *bond_get_bond_by_slave(struct slave *slave)
294 265
295static inline bool bond_is_lb(const struct bonding *bond) 266static inline bool bond_is_lb(const struct bonding *bond)
296{ 267{
297 return (bond->params.mode == BOND_MODE_TLB || 268 return BOND_MODE_IS_LB(bond->params.mode);
298 bond->params.mode == BOND_MODE_ALB);
299} 269}
300 270
301static inline void bond_set_active_slave(struct slave *slave) 271static inline void bond_set_active_slave(struct slave *slave)
@@ -432,21 +402,18 @@ static inline bool slave_can_tx(struct slave *slave)
432struct bond_net; 402struct bond_net;
433 403
434int bond_arp_rcv(const struct sk_buff *skb, struct bonding *bond, struct slave *slave); 404int bond_arp_rcv(const struct sk_buff *skb, struct bonding *bond, struct slave *slave);
435struct vlan_entry *bond_next_vlan(struct bonding *bond, struct vlan_entry *curr);
436int bond_dev_queue_xmit(struct bonding *bond, struct sk_buff *skb, struct net_device *slave_dev); 405int bond_dev_queue_xmit(struct bonding *bond, struct sk_buff *skb, struct net_device *slave_dev);
437void bond_xmit_slave_id(struct bonding *bond, struct sk_buff *skb, int slave_id); 406void bond_xmit_slave_id(struct bonding *bond, struct sk_buff *skb, int slave_id);
438int bond_create(struct net *net, const char *name); 407int bond_create(struct net *net, const char *name);
439int bond_create_sysfs(struct bond_net *net); 408int bond_create_sysfs(struct bond_net *net);
440void bond_destroy_sysfs(struct bond_net *net); 409void bond_destroy_sysfs(struct bond_net *net);
441void bond_prepare_sysfs_group(struct bonding *bond); 410void bond_prepare_sysfs_group(struct bonding *bond);
442int bond_create_slave_symlinks(struct net_device *master, struct net_device *slave);
443void bond_destroy_slave_symlinks(struct net_device *master, struct net_device *slave);
444int bond_enslave(struct net_device *bond_dev, struct net_device *slave_dev); 411int bond_enslave(struct net_device *bond_dev, struct net_device *slave_dev);
445int bond_release(struct net_device *bond_dev, struct net_device *slave_dev); 412int bond_release(struct net_device *bond_dev, struct net_device *slave_dev);
446void bond_mii_monitor(struct work_struct *); 413void bond_mii_monitor(struct work_struct *);
447void bond_loadbalance_arp_mon(struct work_struct *); 414void bond_loadbalance_arp_mon(struct work_struct *);
448void bond_activebackup_arp_mon(struct work_struct *); 415void bond_activebackup_arp_mon(struct work_struct *);
449void bond_set_mode_ops(struct bonding *bond, int mode); 416int bond_xmit_hash(struct bonding *bond, struct sk_buff *skb, int count);
450int bond_parse_parm(const char *mode_arg, const struct bond_parm_tbl *tbl); 417int bond_parse_parm(const char *mode_arg, const struct bond_parm_tbl *tbl);
451void bond_select_active_slave(struct bonding *bond); 418void bond_select_active_slave(struct bonding *bond);
452void bond_change_active_slave(struct bonding *bond, struct slave *new_active); 419void bond_change_active_slave(struct bonding *bond, struct slave *new_active);
@@ -456,6 +423,14 @@ void bond_debug_register(struct bonding *bond);
456void bond_debug_unregister(struct bonding *bond); 423void bond_debug_unregister(struct bonding *bond);
457void bond_debug_reregister(struct bonding *bond); 424void bond_debug_reregister(struct bonding *bond);
458const char *bond_mode_name(int mode); 425const char *bond_mode_name(int mode);
426void bond_setup(struct net_device *bond_dev);
427unsigned int bond_get_num_tx_queues(void);
428int bond_netlink_init(void);
429void bond_netlink_fini(void);
430int bond_option_mode_set(struct bonding *bond, int mode);
431int bond_option_active_slave_set(struct bonding *bond, struct net_device *slave_dev);
432struct net_device *bond_option_active_slave_get_rcu(struct bonding *bond);
433struct net_device *bond_option_active_slave_get(struct bonding *bond);
459 434
460struct bond_net { 435struct bond_net {
461 struct net * net; /* Associated network namespace */ 436 struct net * net; /* Associated network namespace */
@@ -492,9 +467,24 @@ static inline void bond_destroy_proc_dir(struct bond_net *bn)
492static inline struct slave *bond_slave_has_mac(struct bonding *bond, 467static inline struct slave *bond_slave_has_mac(struct bonding *bond,
493 const u8 *mac) 468 const u8 *mac)
494{ 469{
470 struct list_head *iter;
495 struct slave *tmp; 471 struct slave *tmp;
496 472
497 bond_for_each_slave(bond, tmp) 473 bond_for_each_slave(bond, tmp, iter)
474 if (ether_addr_equal_64bits(mac, tmp->dev->dev_addr))
475 return tmp;
476
477 return NULL;
478}
479
480/* Caller must hold rcu_read_lock() for read */
481static inline struct slave *bond_slave_has_mac_rcu(struct bonding *bond,
482 const u8 *mac)
483{
484 struct list_head *iter;
485 struct slave *tmp;
486
487 bond_for_each_slave_rcu(bond, tmp, iter)
498 if (ether_addr_equal_64bits(mac, tmp->dev->dev_addr)) 488 if (ether_addr_equal_64bits(mac, tmp->dev->dev_addr))
499 return tmp; 489 return tmp;
500 490
@@ -528,4 +518,7 @@ extern const struct bond_parm_tbl fail_over_mac_tbl[];
528extern const struct bond_parm_tbl pri_reselect_tbl[]; 518extern const struct bond_parm_tbl pri_reselect_tbl[];
529extern struct bond_parm_tbl ad_select_tbl[]; 519extern struct bond_parm_tbl ad_select_tbl[];
530 520
521/* exported from bond_netlink.c */
522extern struct rtnl_link_ops bond_link_ops;
523
531#endif /* _LINUX_BONDING_H */ 524#endif /* _LINUX_BONDING_H */
diff --git a/drivers/net/can/at91_can.c b/drivers/net/can/at91_can.c
index 693d8ffe4653..cf0f63e14e53 100644
--- a/drivers/net/can/at91_can.c
+++ b/drivers/net/can/at91_can.c
@@ -1347,7 +1347,7 @@ static int at91_can_probe(struct platform_device *pdev)
1347 priv->reg_base = addr; 1347 priv->reg_base = addr;
1348 priv->devtype_data = *devtype_data; 1348 priv->devtype_data = *devtype_data;
1349 priv->clk = clk; 1349 priv->clk = clk;
1350 priv->pdata = pdev->dev.platform_data; 1350 priv->pdata = dev_get_platdata(&pdev->dev);
1351 priv->mb0_id = 0x7ff; 1351 priv->mb0_id = 0x7ff;
1352 1352
1353 netif_napi_add(dev, &priv->napi, at91_poll, get_mb_rx_num(priv)); 1353 netif_napi_add(dev, &priv->napi, at91_poll, get_mb_rx_num(priv));
diff --git a/drivers/net/can/bfin_can.c b/drivers/net/can/bfin_can.c
index a2700d25ff0e..8a0b515b33ea 100644
--- a/drivers/net/can/bfin_can.c
+++ b/drivers/net/can/bfin_can.c
@@ -539,7 +539,7 @@ static int bfin_can_probe(struct platform_device *pdev)
539 struct resource *res_mem, *rx_irq, *tx_irq, *err_irq; 539 struct resource *res_mem, *rx_irq, *tx_irq, *err_irq;
540 unsigned short *pdata; 540 unsigned short *pdata;
541 541
542 pdata = pdev->dev.platform_data; 542 pdata = dev_get_platdata(&pdev->dev);
543 if (!pdata) { 543 if (!pdata) {
544 dev_err(&pdev->dev, "No platform data provided!\n"); 544 dev_err(&pdev->dev, "No platform data provided!\n");
545 err = -EINVAL; 545 err = -EINVAL;
diff --git a/drivers/net/can/c_can/c_can_pci.c b/drivers/net/can/c_can/c_can_pci.c
index b374be7891a2..bce0be54c2f5 100644
--- a/drivers/net/can/c_can/c_can_pci.c
+++ b/drivers/net/can/c_can/c_can_pci.c
@@ -160,7 +160,6 @@ static int c_can_pci_probe(struct pci_dev *pdev,
160 return 0; 160 return 0;
161 161
162out_free_c_can: 162out_free_c_can:
163 pci_set_drvdata(pdev, NULL);
164 free_c_can_dev(dev); 163 free_c_can_dev(dev);
165out_iounmap: 164out_iounmap:
166 pci_iounmap(pdev, addr); 165 pci_iounmap(pdev, addr);
@@ -181,7 +180,6 @@ static void c_can_pci_remove(struct pci_dev *pdev)
181 180
182 unregister_c_can_dev(dev); 181 unregister_c_can_dev(dev);
183 182
184 pci_set_drvdata(pdev, NULL);
185 free_c_can_dev(dev); 183 free_c_can_dev(dev);
186 184
187 pci_iounmap(pdev, priv->base); 185 pci_iounmap(pdev, priv->base);
diff --git a/drivers/net/can/c_can/c_can_platform.c b/drivers/net/can/c_can/c_can_platform.c
index 294ced3cc227..d66ac265269c 100644
--- a/drivers/net/can/c_can/c_can_platform.c
+++ b/drivers/net/can/c_can/c_can_platform.c
@@ -322,7 +322,7 @@ static struct platform_driver c_can_plat_driver = {
322 .driver = { 322 .driver = {
323 .name = KBUILD_MODNAME, 323 .name = KBUILD_MODNAME,
324 .owner = THIS_MODULE, 324 .owner = THIS_MODULE,
325 .of_match_table = of_match_ptr(c_can_of_table), 325 .of_match_table = c_can_of_table,
326 }, 326 },
327 .probe = c_can_plat_probe, 327 .probe = c_can_plat_probe,
328 .remove = c_can_plat_remove, 328 .remove = c_can_plat_remove,
diff --git a/drivers/net/can/cc770/cc770_platform.c b/drivers/net/can/cc770/cc770_platform.c
index 034bdd816a60..ad76734b3ecc 100644
--- a/drivers/net/can/cc770/cc770_platform.c
+++ b/drivers/net/can/cc770/cc770_platform.c
@@ -152,7 +152,7 @@ static int cc770_get_platform_data(struct platform_device *pdev,
152 struct cc770_priv *priv) 152 struct cc770_priv *priv)
153{ 153{
154 154
155 struct cc770_platform_data *pdata = pdev->dev.platform_data; 155 struct cc770_platform_data *pdata = dev_get_platdata(&pdev->dev);
156 156
157 priv->can.clock.freq = pdata->osc_freq; 157 priv->can.clock.freq = pdata->osc_freq;
158 if (priv->cpu_interface & CPUIF_DSC) 158 if (priv->cpu_interface & CPUIF_DSC)
@@ -203,7 +203,7 @@ static int cc770_platform_probe(struct platform_device *pdev)
203 203
204 if (pdev->dev.of_node) 204 if (pdev->dev.of_node)
205 err = cc770_get_of_node_data(pdev, priv); 205 err = cc770_get_of_node_data(pdev, priv);
206 else if (pdev->dev.platform_data) 206 else if (dev_get_platdata(&pdev->dev))
207 err = cc770_get_platform_data(pdev, priv); 207 err = cc770_get_platform_data(pdev, priv);
208 else 208 else
209 err = -ENODEV; 209 err = -ENODEV;
diff --git a/drivers/net/can/dev.c b/drivers/net/can/dev.c
index 1870c4731a57..bda1888cae9a 100644
--- a/drivers/net/can/dev.c
+++ b/drivers/net/can/dev.c
@@ -645,19 +645,6 @@ static int can_changelink(struct net_device *dev,
645 /* We need synchronization with dev->stop() */ 645 /* We need synchronization with dev->stop() */
646 ASSERT_RTNL(); 646 ASSERT_RTNL();
647 647
648 if (data[IFLA_CAN_CTRLMODE]) {
649 struct can_ctrlmode *cm;
650
651 /* Do not allow changing controller mode while running */
652 if (dev->flags & IFF_UP)
653 return -EBUSY;
654 cm = nla_data(data[IFLA_CAN_CTRLMODE]);
655 if (cm->flags & ~priv->ctrlmode_supported)
656 return -EOPNOTSUPP;
657 priv->ctrlmode &= ~cm->mask;
658 priv->ctrlmode |= cm->flags;
659 }
660
661 if (data[IFLA_CAN_BITTIMING]) { 648 if (data[IFLA_CAN_BITTIMING]) {
662 struct can_bittiming bt; 649 struct can_bittiming bt;
663 650
@@ -680,6 +667,19 @@ static int can_changelink(struct net_device *dev,
680 } 667 }
681 } 668 }
682 669
670 if (data[IFLA_CAN_CTRLMODE]) {
671 struct can_ctrlmode *cm;
672
673 /* Do not allow changing controller mode while running */
674 if (dev->flags & IFF_UP)
675 return -EBUSY;
676 cm = nla_data(data[IFLA_CAN_CTRLMODE]);
677 if (cm->flags & ~priv->ctrlmode_supported)
678 return -EOPNOTSUPP;
679 priv->ctrlmode &= ~cm->mask;
680 priv->ctrlmode |= cm->flags;
681 }
682
683 if (data[IFLA_CAN_RESTART_MS]) { 683 if (data[IFLA_CAN_RESTART_MS]) {
684 /* Do not allow changing restart delay while running */ 684 /* Do not allow changing restart delay while running */
685 if (dev->flags & IFF_UP) 685 if (dev->flags & IFF_UP)
@@ -702,17 +702,17 @@ static int can_changelink(struct net_device *dev,
702static size_t can_get_size(const struct net_device *dev) 702static size_t can_get_size(const struct net_device *dev)
703{ 703{
704 struct can_priv *priv = netdev_priv(dev); 704 struct can_priv *priv = netdev_priv(dev);
705 size_t size; 705 size_t size = 0;
706 706
707 size = nla_total_size(sizeof(u32)); /* IFLA_CAN_STATE */ 707 size += nla_total_size(sizeof(struct can_bittiming)); /* IFLA_CAN_BITTIMING */
708 size += nla_total_size(sizeof(struct can_ctrlmode)); /* IFLA_CAN_CTRLMODE */ 708 if (priv->bittiming_const) /* IFLA_CAN_BITTIMING_CONST */
709 size += nla_total_size(sizeof(u32)); /* IFLA_CAN_RESTART_MS */
710 size += nla_total_size(sizeof(struct can_bittiming)); /* IFLA_CAN_BITTIMING */
711 size += nla_total_size(sizeof(struct can_clock)); /* IFLA_CAN_CLOCK */
712 if (priv->do_get_berr_counter) /* IFLA_CAN_BERR_COUNTER */
713 size += nla_total_size(sizeof(struct can_berr_counter));
714 if (priv->bittiming_const) /* IFLA_CAN_BITTIMING_CONST */
715 size += nla_total_size(sizeof(struct can_bittiming_const)); 709 size += nla_total_size(sizeof(struct can_bittiming_const));
710 size += nla_total_size(sizeof(struct can_clock)); /* IFLA_CAN_CLOCK */
711 size += nla_total_size(sizeof(u32)); /* IFLA_CAN_STATE */
712 size += nla_total_size(sizeof(struct can_ctrlmode)); /* IFLA_CAN_CTRLMODE */
713 size += nla_total_size(sizeof(u32)); /* IFLA_CAN_RESTART_MS */
714 if (priv->do_get_berr_counter) /* IFLA_CAN_BERR_COUNTER */
715 size += nla_total_size(sizeof(struct can_berr_counter));
716 716
717 return size; 717 return size;
718} 718}
@@ -726,23 +726,20 @@ static int can_fill_info(struct sk_buff *skb, const struct net_device *dev)
726 726
727 if (priv->do_get_state) 727 if (priv->do_get_state)
728 priv->do_get_state(dev, &state); 728 priv->do_get_state(dev, &state);
729 if (nla_put_u32(skb, IFLA_CAN_STATE, state) || 729 if (nla_put(skb, IFLA_CAN_BITTIMING,
730 nla_put(skb, IFLA_CAN_CTRLMODE, sizeof(cm), &cm) ||
731 nla_put_u32(skb, IFLA_CAN_RESTART_MS, priv->restart_ms) ||
732 nla_put(skb, IFLA_CAN_BITTIMING,
733 sizeof(priv->bittiming), &priv->bittiming) || 730 sizeof(priv->bittiming), &priv->bittiming) ||
731 (priv->bittiming_const &&
732 nla_put(skb, IFLA_CAN_BITTIMING_CONST,
733 sizeof(*priv->bittiming_const), priv->bittiming_const)) ||
734 nla_put(skb, IFLA_CAN_CLOCK, sizeof(cm), &priv->clock) || 734 nla_put(skb, IFLA_CAN_CLOCK, sizeof(cm), &priv->clock) ||
735 nla_put_u32(skb, IFLA_CAN_STATE, state) ||
736 nla_put(skb, IFLA_CAN_CTRLMODE, sizeof(cm), &cm) ||
737 nla_put_u32(skb, IFLA_CAN_RESTART_MS, priv->restart_ms) ||
735 (priv->do_get_berr_counter && 738 (priv->do_get_berr_counter &&
736 !priv->do_get_berr_counter(dev, &bec) && 739 !priv->do_get_berr_counter(dev, &bec) &&
737 nla_put(skb, IFLA_CAN_BERR_COUNTER, sizeof(bec), &bec)) || 740 nla_put(skb, IFLA_CAN_BERR_COUNTER, sizeof(bec), &bec)))
738 (priv->bittiming_const && 741 return -EMSGSIZE;
739 nla_put(skb, IFLA_CAN_BITTIMING_CONST,
740 sizeof(*priv->bittiming_const), priv->bittiming_const)))
741 goto nla_put_failure;
742 return 0; 742 return 0;
743
744nla_put_failure:
745 return -EMSGSIZE;
746} 743}
747 744
748static size_t can_get_xstats_size(const struct net_device *dev) 745static size_t can_get_xstats_size(const struct net_device *dev)
diff --git a/drivers/net/can/flexcan.c b/drivers/net/can/flexcan.c
index 8f5ce747feb5..ae08cf129ebb 100644
--- a/drivers/net/can/flexcan.c
+++ b/drivers/net/can/flexcan.c
@@ -1068,7 +1068,7 @@ static int flexcan_probe(struct platform_device *pdev)
1068 priv->dev = dev; 1068 priv->dev = dev;
1069 priv->clk_ipg = clk_ipg; 1069 priv->clk_ipg = clk_ipg;
1070 priv->clk_per = clk_per; 1070 priv->clk_per = clk_per;
1071 priv->pdata = pdev->dev.platform_data; 1071 priv->pdata = dev_get_platdata(&pdev->dev);
1072 priv->devtype_data = devtype_data; 1072 priv->devtype_data = devtype_data;
1073 1073
1074 priv->reg_xceiver = devm_regulator_get(&pdev->dev, "xceiver"); 1074 priv->reg_xceiver = devm_regulator_get(&pdev->dev, "xceiver");
diff --git a/drivers/net/can/janz-ican3.c b/drivers/net/can/janz-ican3.c
index 36bd6fa1c7f3..ab5909a7bae9 100644
--- a/drivers/net/can/janz-ican3.c
+++ b/drivers/net/can/janz-ican3.c
@@ -1769,7 +1769,7 @@ static int ican3_probe(struct platform_device *pdev)
1769 struct device *dev; 1769 struct device *dev;
1770 int ret; 1770 int ret;
1771 1771
1772 pdata = pdev->dev.platform_data; 1772 pdata = dev_get_platdata(&pdev->dev);
1773 if (!pdata) 1773 if (!pdata)
1774 return -ENXIO; 1774 return -ENXIO;
1775 1775
diff --git a/drivers/net/can/mcp251x.c b/drivers/net/can/mcp251x.c
index fe7dd696957e..08ac401e0214 100644
--- a/drivers/net/can/mcp251x.c
+++ b/drivers/net/can/mcp251x.c
@@ -999,7 +999,7 @@ static int mcp251x_can_probe(struct spi_device *spi)
999{ 999{
1000 struct net_device *net; 1000 struct net_device *net;
1001 struct mcp251x_priv *priv; 1001 struct mcp251x_priv *priv;
1002 struct mcp251x_platform_data *pdata = spi->dev.platform_data; 1002 struct mcp251x_platform_data *pdata = dev_get_platdata(&spi->dev);
1003 int ret = -ENODEV; 1003 int ret = -ENODEV;
1004 1004
1005 if (!pdata) 1005 if (!pdata)
diff --git a/drivers/net/can/mscan/mscan.h b/drivers/net/can/mscan/mscan.h
index 9c24d60a23b1..e98abb97a050 100644
--- a/drivers/net/can/mscan/mscan.h
+++ b/drivers/net/can/mscan/mscan.h
@@ -297,8 +297,8 @@ struct mscan_priv {
297 struct napi_struct napi; 297 struct napi_struct napi;
298}; 298};
299 299
300extern struct net_device *alloc_mscandev(void); 300struct net_device *alloc_mscandev(void);
301extern int register_mscandev(struct net_device *dev, int mscan_clksrc); 301int register_mscandev(struct net_device *dev, int mscan_clksrc);
302extern void unregister_mscandev(struct net_device *dev); 302void unregister_mscandev(struct net_device *dev);
303 303
304#endif /* __MSCAN_H__ */ 304#endif /* __MSCAN_H__ */
diff --git a/drivers/net/can/pch_can.c b/drivers/net/can/pch_can.c
index 5c314a961970..5f0e9b3bfa7b 100644
--- a/drivers/net/can/pch_can.c
+++ b/drivers/net/can/pch_can.c
@@ -964,7 +964,6 @@ static void pch_can_remove(struct pci_dev *pdev)
964 pci_disable_msi(priv->dev); 964 pci_disable_msi(priv->dev);
965 pci_release_regions(pdev); 965 pci_release_regions(pdev);
966 pci_disable_device(pdev); 966 pci_disable_device(pdev);
967 pci_set_drvdata(pdev, NULL);
968 pch_can_reset(priv); 967 pch_can_reset(priv);
969 pci_iounmap(pdev, priv->regs); 968 pci_iounmap(pdev, priv->regs);
970 free_candev(priv->ndev); 969 free_candev(priv->ndev);
diff --git a/drivers/net/can/sja1000/ems_pci.c b/drivers/net/can/sja1000/ems_pci.c
index 3752342a678a..835921388e7b 100644
--- a/drivers/net/can/sja1000/ems_pci.c
+++ b/drivers/net/can/sja1000/ems_pci.c
@@ -207,7 +207,6 @@ static void ems_pci_del_card(struct pci_dev *pdev)
207 kfree(card); 207 kfree(card);
208 208
209 pci_disable_device(pdev); 209 pci_disable_device(pdev);
210 pci_set_drvdata(pdev, NULL);
211} 210}
212 211
213static void ems_pci_card_reset(struct ems_pci_card *card) 212static void ems_pci_card_reset(struct ems_pci_card *card)
diff --git a/drivers/net/can/sja1000/kvaser_pci.c b/drivers/net/can/sja1000/kvaser_pci.c
index 217585b97cd3..087b13bd300e 100644
--- a/drivers/net/can/sja1000/kvaser_pci.c
+++ b/drivers/net/can/sja1000/kvaser_pci.c
@@ -387,7 +387,6 @@ static void kvaser_pci_remove_one(struct pci_dev *pdev)
387 387
388 pci_release_regions(pdev); 388 pci_release_regions(pdev);
389 pci_disable_device(pdev); 389 pci_disable_device(pdev);
390 pci_set_drvdata(pdev, NULL);
391} 390}
392 391
393static struct pci_driver kvaser_pci_driver = { 392static struct pci_driver kvaser_pci_driver = {
diff --git a/drivers/net/can/sja1000/peak_pci.c b/drivers/net/can/sja1000/peak_pci.c
index 6b6f0ad75090..065ca49eb45e 100644
--- a/drivers/net/can/sja1000/peak_pci.c
+++ b/drivers/net/can/sja1000/peak_pci.c
@@ -744,8 +744,6 @@ static void peak_pci_remove(struct pci_dev *pdev)
744 pci_iounmap(pdev, cfg_base); 744 pci_iounmap(pdev, cfg_base);
745 pci_release_regions(pdev); 745 pci_release_regions(pdev);
746 pci_disable_device(pdev); 746 pci_disable_device(pdev);
747
748 pci_set_drvdata(pdev, NULL);
749} 747}
750 748
751static struct pci_driver peak_pci_driver = { 749static struct pci_driver peak_pci_driver = {
diff --git a/drivers/net/can/sja1000/plx_pci.c b/drivers/net/can/sja1000/plx_pci.c
index c52c1e96bf90..f9b4f81cd86a 100644
--- a/drivers/net/can/sja1000/plx_pci.c
+++ b/drivers/net/can/sja1000/plx_pci.c
@@ -477,7 +477,6 @@ static void plx_pci_del_card(struct pci_dev *pdev)
477 kfree(card); 477 kfree(card);
478 478
479 pci_disable_device(pdev); 479 pci_disable_device(pdev);
480 pci_set_drvdata(pdev, NULL);
481} 480}
482 481
483/* 482/*
diff --git a/drivers/net/can/sja1000/sja1000_platform.c b/drivers/net/can/sja1000/sja1000_platform.c
index 8e259c541036..29f9b6321187 100644
--- a/drivers/net/can/sja1000/sja1000_platform.c
+++ b/drivers/net/can/sja1000/sja1000_platform.c
@@ -76,7 +76,7 @@ static int sp_probe(struct platform_device *pdev)
76 struct resource *res_mem, *res_irq; 76 struct resource *res_mem, *res_irq;
77 struct sja1000_platform_data *pdata; 77 struct sja1000_platform_data *pdata;
78 78
79 pdata = pdev->dev.platform_data; 79 pdata = dev_get_platdata(&pdev->dev);
80 if (!pdata) { 80 if (!pdata) {
81 dev_err(&pdev->dev, "No platform data provided!\n"); 81 dev_err(&pdev->dev, "No platform data provided!\n");
82 err = -ENODEV; 82 err = -ENODEV;
diff --git a/drivers/net/can/softing/softing.h b/drivers/net/can/softing/softing.h
index afd7d85b6915..35f062282dbd 100644
--- a/drivers/net/can/softing/softing.h
+++ b/drivers/net/can/softing/softing.h
@@ -71,34 +71,34 @@ struct softing {
71 } id; 71 } id;
72}; 72};
73 73
74extern int softing_default_output(struct net_device *netdev); 74int softing_default_output(struct net_device *netdev);
75 75
76extern ktime_t softing_raw2ktime(struct softing *card, u32 raw); 76ktime_t softing_raw2ktime(struct softing *card, u32 raw);
77 77
78extern int softing_chip_poweron(struct softing *card); 78int softing_chip_poweron(struct softing *card);
79 79
80extern int softing_bootloader_command(struct softing *card, int16_t cmd, 80int softing_bootloader_command(struct softing *card, int16_t cmd,
81 const char *msg); 81 const char *msg);
82 82
83/* Load firmware after reset */ 83/* Load firmware after reset */
84extern int softing_load_fw(const char *file, struct softing *card, 84int softing_load_fw(const char *file, struct softing *card,
85 __iomem uint8_t *virt, unsigned int size, int offset); 85 __iomem uint8_t *virt, unsigned int size, int offset);
86 86
87/* Load final application firmware after bootloader */ 87/* Load final application firmware after bootloader */
88extern int softing_load_app_fw(const char *file, struct softing *card); 88int softing_load_app_fw(const char *file, struct softing *card);
89 89
90/* 90/*
91 * enable or disable irq 91 * enable or disable irq
92 * only called with fw.lock locked 92 * only called with fw.lock locked
93 */ 93 */
94extern int softing_enable_irq(struct softing *card, int enable); 94int softing_enable_irq(struct softing *card, int enable);
95 95
96/* start/stop 1 bus on card */ 96/* start/stop 1 bus on card */
97extern int softing_startstop(struct net_device *netdev, int up); 97int softing_startstop(struct net_device *netdev, int up);
98 98
99/* netif_rx() */ 99/* netif_rx() */
100extern int softing_netdev_rx(struct net_device *netdev, 100int softing_netdev_rx(struct net_device *netdev, const struct can_frame *msg,
101 const struct can_frame *msg, ktime_t ktime); 101 ktime_t ktime);
102 102
103/* SOFTING DPRAM mappings */ 103/* SOFTING DPRAM mappings */
104#define DPRAM_RX 0x0000 104#define DPRAM_RX 0x0000
diff --git a/drivers/net/can/softing/softing_main.c b/drivers/net/can/softing/softing_main.c
index 65eef1eea2e2..6cd5c01b624d 100644
--- a/drivers/net/can/softing/softing_main.c
+++ b/drivers/net/can/softing/softing_main.c
@@ -768,7 +768,7 @@ static int softing_pdev_remove(struct platform_device *pdev)
768 768
769static int softing_pdev_probe(struct platform_device *pdev) 769static int softing_pdev_probe(struct platform_device *pdev)
770{ 770{
771 const struct softing_platform_data *pdat = pdev->dev.platform_data; 771 const struct softing_platform_data *pdat = dev_get_platdata(&pdev->dev);
772 struct softing *card; 772 struct softing *card;
773 struct net_device *netdev; 773 struct net_device *netdev;
774 struct softing_priv *priv; 774 struct softing_priv *priv;
diff --git a/drivers/net/can/ti_hecc.c b/drivers/net/can/ti_hecc.c
index 3a349a22d5bc..60d95b44d0f7 100644
--- a/drivers/net/can/ti_hecc.c
+++ b/drivers/net/can/ti_hecc.c
@@ -286,15 +286,6 @@ static inline u32 hecc_get_bit(struct ti_hecc_priv *priv, int reg, u32 bit_mask)
286 return (hecc_read(priv, reg) & bit_mask) ? 1 : 0; 286 return (hecc_read(priv, reg) & bit_mask) ? 1 : 0;
287} 287}
288 288
289static int ti_hecc_get_state(const struct net_device *ndev,
290 enum can_state *state)
291{
292 struct ti_hecc_priv *priv = netdev_priv(ndev);
293
294 *state = priv->can.state;
295 return 0;
296}
297
298static int ti_hecc_set_btc(struct ti_hecc_priv *priv) 289static int ti_hecc_set_btc(struct ti_hecc_priv *priv)
299{ 290{
300 struct can_bittiming *bit_timing = &priv->can.bittiming; 291 struct can_bittiming *bit_timing = &priv->can.bittiming;
@@ -894,7 +885,7 @@ static int ti_hecc_probe(struct platform_device *pdev)
894 void __iomem *addr; 885 void __iomem *addr;
895 int err = -ENODEV; 886 int err = -ENODEV;
896 887
897 pdata = pdev->dev.platform_data; 888 pdata = dev_get_platdata(&pdev->dev);
898 if (!pdata) { 889 if (!pdata) {
899 dev_err(&pdev->dev, "No platform data\n"); 890 dev_err(&pdev->dev, "No platform data\n");
900 goto probe_exit; 891 goto probe_exit;
@@ -940,7 +931,6 @@ static int ti_hecc_probe(struct platform_device *pdev)
940 931
941 priv->can.bittiming_const = &ti_hecc_bittiming_const; 932 priv->can.bittiming_const = &ti_hecc_bittiming_const;
942 priv->can.do_set_mode = ti_hecc_do_set_mode; 933 priv->can.do_set_mode = ti_hecc_do_set_mode;
943 priv->can.do_get_state = ti_hecc_get_state;
944 priv->can.do_get_berr_counter = ti_hecc_get_berr_counter; 934 priv->can.do_get_berr_counter = ti_hecc_get_berr_counter;
945 priv->can.ctrlmode_supported = CAN_CTRLMODE_3_SAMPLES; 935 priv->can.ctrlmode_supported = CAN_CTRLMODE_3_SAMPLES;
946 936
diff --git a/drivers/net/ethernet/3com/Kconfig b/drivers/net/ethernet/3com/Kconfig
index f00c76377b44..65b735d4a6ad 100644
--- a/drivers/net/ethernet/3com/Kconfig
+++ b/drivers/net/ethernet/3com/Kconfig
@@ -35,7 +35,7 @@ config EL3
35 35
36config 3C515 36config 3C515
37 tristate "3c515 ISA \"Fast EtherLink\"" 37 tristate "3c515 ISA \"Fast EtherLink\""
38 depends on (ISA || EISA) && ISA_DMA_API 38 depends on ISA && ISA_DMA_API
39 ---help--- 39 ---help---
40 If you have a 3Com ISA EtherLink XL "Corkscrew" 3c515 Fast Ethernet 40 If you have a 3Com ISA EtherLink XL "Corkscrew" 3c515 Fast Ethernet
41 network card, say Y and read the Ethernet-HOWTO, available from 41 network card, say Y and read the Ethernet-HOWTO, available from
@@ -70,7 +70,7 @@ config VORTEX
70 select MII 70 select MII
71 ---help--- 71 ---help---
72 This option enables driver support for a large number of 10Mbps and 72 This option enables driver support for a large number of 10Mbps and
73 10/100Mbps EISA, PCI and PCMCIA 3Com network cards: 73 10/100Mbps EISA, PCI and Cardbus 3Com network cards:
74 74
75 "Vortex" (Fast EtherLink 3c590/3c592/3c595/3c597) EISA and PCI 75 "Vortex" (Fast EtherLink 3c590/3c592/3c595/3c597) EISA and PCI
76 "Boomerang" (EtherLink XL 3c900 or 3c905) PCI 76 "Boomerang" (EtherLink XL 3c900 or 3c905) PCI
diff --git a/drivers/net/ethernet/3com/typhoon.c b/drivers/net/ethernet/3com/typhoon.c
index 144942f6372b..465cc7108d8a 100644
--- a/drivers/net/ethernet/3com/typhoon.c
+++ b/drivers/net/ethernet/3com/typhoon.c
@@ -2525,7 +2525,6 @@ typhoon_remove_one(struct pci_dev *pdev)
2525 pci_release_regions(pdev); 2525 pci_release_regions(pdev);
2526 pci_clear_mwi(pdev); 2526 pci_clear_mwi(pdev);
2527 pci_disable_device(pdev); 2527 pci_disable_device(pdev);
2528 pci_set_drvdata(pdev, NULL);
2529 free_netdev(dev); 2528 free_netdev(dev);
2530} 2529}
2531 2530
diff --git a/drivers/net/ethernet/8390/8390.h b/drivers/net/ethernet/8390/8390.h
index ef325ffa1b5a..2923c51bb351 100644
--- a/drivers/net/ethernet/8390/8390.h
+++ b/drivers/net/ethernet/8390/8390.h
@@ -28,42 +28,42 @@ extern int ei_debug;
28#endif 28#endif
29 29
30#ifdef CONFIG_NET_POLL_CONTROLLER 30#ifdef CONFIG_NET_POLL_CONTROLLER
31extern void ei_poll(struct net_device *dev); 31void ei_poll(struct net_device *dev);
32extern void eip_poll(struct net_device *dev); 32void eip_poll(struct net_device *dev);
33#endif 33#endif
34 34
35 35
36/* Without I/O delay - non ISA or later chips */ 36/* Without I/O delay - non ISA or later chips */
37extern void NS8390_init(struct net_device *dev, int startp); 37void NS8390_init(struct net_device *dev, int startp);
38extern int ei_open(struct net_device *dev); 38int ei_open(struct net_device *dev);
39extern int ei_close(struct net_device *dev); 39int ei_close(struct net_device *dev);
40extern irqreturn_t ei_interrupt(int irq, void *dev_id); 40irqreturn_t ei_interrupt(int irq, void *dev_id);
41extern void ei_tx_timeout(struct net_device *dev); 41void ei_tx_timeout(struct net_device *dev);
42extern netdev_tx_t ei_start_xmit(struct sk_buff *skb, struct net_device *dev); 42netdev_tx_t ei_start_xmit(struct sk_buff *skb, struct net_device *dev);
43extern void ei_set_multicast_list(struct net_device *dev); 43void ei_set_multicast_list(struct net_device *dev);
44extern struct net_device_stats *ei_get_stats(struct net_device *dev); 44struct net_device_stats *ei_get_stats(struct net_device *dev);
45 45
46extern const struct net_device_ops ei_netdev_ops; 46extern const struct net_device_ops ei_netdev_ops;
47 47
48extern struct net_device *__alloc_ei_netdev(int size); 48struct net_device *__alloc_ei_netdev(int size);
49static inline struct net_device *alloc_ei_netdev(void) 49static inline struct net_device *alloc_ei_netdev(void)
50{ 50{
51 return __alloc_ei_netdev(0); 51 return __alloc_ei_netdev(0);
52} 52}
53 53
54/* With I/O delay form */ 54/* With I/O delay form */
55extern void NS8390p_init(struct net_device *dev, int startp); 55void NS8390p_init(struct net_device *dev, int startp);
56extern int eip_open(struct net_device *dev); 56int eip_open(struct net_device *dev);
57extern int eip_close(struct net_device *dev); 57int eip_close(struct net_device *dev);
58extern irqreturn_t eip_interrupt(int irq, void *dev_id); 58irqreturn_t eip_interrupt(int irq, void *dev_id);
59extern void eip_tx_timeout(struct net_device *dev); 59void eip_tx_timeout(struct net_device *dev);
60extern netdev_tx_t eip_start_xmit(struct sk_buff *skb, struct net_device *dev); 60netdev_tx_t eip_start_xmit(struct sk_buff *skb, struct net_device *dev);
61extern void eip_set_multicast_list(struct net_device *dev); 61void eip_set_multicast_list(struct net_device *dev);
62extern struct net_device_stats *eip_get_stats(struct net_device *dev); 62struct net_device_stats *eip_get_stats(struct net_device *dev);
63 63
64extern const struct net_device_ops eip_netdev_ops; 64extern const struct net_device_ops eip_netdev_ops;
65 65
66extern struct net_device *__alloc_eip_netdev(int size); 66struct net_device *__alloc_eip_netdev(int size);
67static inline struct net_device *alloc_eip_netdev(void) 67static inline struct net_device *alloc_eip_netdev(void)
68{ 68{
69 return __alloc_eip_netdev(0); 69 return __alloc_eip_netdev(0);
diff --git a/drivers/net/ethernet/8390/ax88796.c b/drivers/net/ethernet/8390/ax88796.c
index f92f001551da..36fa577970bb 100644
--- a/drivers/net/ethernet/8390/ax88796.c
+++ b/drivers/net/ethernet/8390/ax88796.c
@@ -702,7 +702,7 @@ static int ax_init_dev(struct net_device *dev)
702 for (i = 0; i < 16; i++) 702 for (i = 0; i < 16; i++)
703 SA_prom[i] = SA_prom[i+i]; 703 SA_prom[i] = SA_prom[i+i];
704 704
705 memcpy(dev->dev_addr, SA_prom, 6); 705 memcpy(dev->dev_addr, SA_prom, ETH_ALEN);
706 } 706 }
707 707
708#ifdef CONFIG_AX88796_93CX6 708#ifdef CONFIG_AX88796_93CX6
diff --git a/drivers/net/ethernet/8390/ne2k-pci.c b/drivers/net/ethernet/8390/ne2k-pci.c
index 92201080e07a..fc14a85e4d5f 100644
--- a/drivers/net/ethernet/8390/ne2k-pci.c
+++ b/drivers/net/ethernet/8390/ne2k-pci.c
@@ -389,9 +389,7 @@ err_out_free_netdev:
389 free_netdev (dev); 389 free_netdev (dev);
390err_out_free_res: 390err_out_free_res:
391 release_region (ioaddr, NE_IO_EXTENT); 391 release_region (ioaddr, NE_IO_EXTENT);
392 pci_set_drvdata (pdev, NULL);
393 return -ENODEV; 392 return -ENODEV;
394
395} 393}
396 394
397/* 395/*
@@ -655,7 +653,6 @@ static void ne2k_pci_remove_one(struct pci_dev *pdev)
655 release_region(dev->base_addr, NE_IO_EXTENT); 653 release_region(dev->base_addr, NE_IO_EXTENT);
656 free_netdev(dev); 654 free_netdev(dev);
657 pci_disable_device(pdev); 655 pci_disable_device(pdev);
658 pci_set_drvdata(pdev, NULL);
659} 656}
660 657
661#ifdef CONFIG_PM 658#ifdef CONFIG_PM
diff --git a/drivers/net/ethernet/adaptec/starfire.c b/drivers/net/ethernet/adaptec/starfire.c
index 8b04bfc20cfb..171d73c1d3c2 100644
--- a/drivers/net/ethernet/adaptec/starfire.c
+++ b/drivers/net/ethernet/adaptec/starfire.c
@@ -835,7 +835,6 @@ static int starfire_init_one(struct pci_dev *pdev,
835 return 0; 835 return 0;
836 836
837err_out_cleardev: 837err_out_cleardev:
838 pci_set_drvdata(pdev, NULL);
839 iounmap(base); 838 iounmap(base);
840err_out_free_res: 839err_out_free_res:
841 pci_release_regions (pdev); 840 pci_release_regions (pdev);
@@ -2012,7 +2011,6 @@ static void starfire_remove_one(struct pci_dev *pdev)
2012 iounmap(np->base); 2011 iounmap(np->base);
2013 pci_release_regions(pdev); 2012 pci_release_regions(pdev);
2014 2013
2015 pci_set_drvdata(pdev, NULL);
2016 free_netdev(dev); /* Will also free np!! */ 2014 free_netdev(dev); /* Will also free np!! */
2017} 2015}
2018 2016
diff --git a/drivers/net/ethernet/adi/bfin_mac.h b/drivers/net/ethernet/adi/bfin_mac.h
index 7a07ee07906b..6dec86ac97cd 100644
--- a/drivers/net/ethernet/adi/bfin_mac.h
+++ b/drivers/net/ethernet/adi/bfin_mac.h
@@ -104,6 +104,6 @@ struct bfin_mac_local {
104#endif 104#endif
105}; 105};
106 106
107extern int bfin_get_ether_addr(char *addr); 107int bfin_get_ether_addr(char *addr);
108 108
109#endif 109#endif
diff --git a/drivers/net/ethernet/amd/7990.h b/drivers/net/ethernet/amd/7990.h
index 0a5837b96421..ae33a99bf476 100644
--- a/drivers/net/ethernet/amd/7990.h
+++ b/drivers/net/ethernet/amd/7990.h
@@ -242,13 +242,13 @@ struct lance_private
242#define LANCE_ADDR(x) ((int)(x) & ~0xff000000) 242#define LANCE_ADDR(x) ((int)(x) & ~0xff000000)
243 243
244/* Now the prototypes we export */ 244/* Now the prototypes we export */
245extern int lance_open(struct net_device *dev); 245int lance_open(struct net_device *dev);
246extern int lance_close (struct net_device *dev); 246int lance_close (struct net_device *dev);
247extern int lance_start_xmit (struct sk_buff *skb, struct net_device *dev); 247int lance_start_xmit (struct sk_buff *skb, struct net_device *dev);
248extern void lance_set_multicast (struct net_device *dev); 248void lance_set_multicast (struct net_device *dev);
249extern void lance_tx_timeout(struct net_device *dev); 249void lance_tx_timeout(struct net_device *dev);
250#ifdef CONFIG_NET_POLL_CONTROLLER 250#ifdef CONFIG_NET_POLL_CONTROLLER
251extern void lance_poll(struct net_device *dev); 251void lance_poll(struct net_device *dev);
252#endif 252#endif
253 253
254#endif /* ndef _7990_H */ 254#endif /* ndef _7990_H */
diff --git a/drivers/net/ethernet/amd/amd8111e.c b/drivers/net/ethernet/amd/amd8111e.c
index 1b1429d5d5c2..d042511bdc13 100644
--- a/drivers/net/ethernet/amd/amd8111e.c
+++ b/drivers/net/ethernet/amd/amd8111e.c
@@ -1711,7 +1711,6 @@ static void amd8111e_remove_one(struct pci_dev *pdev)
1711 free_netdev(dev); 1711 free_netdev(dev);
1712 pci_release_regions(pdev); 1712 pci_release_regions(pdev);
1713 pci_disable_device(pdev); 1713 pci_disable_device(pdev);
1714 pci_set_drvdata(pdev, NULL);
1715 } 1714 }
1716} 1715}
1717static void amd8111e_config_ipg(struct net_device* dev) 1716static void amd8111e_config_ipg(struct net_device* dev)
@@ -1967,7 +1966,6 @@ err_free_reg:
1967 1966
1968err_disable_pdev: 1967err_disable_pdev:
1969 pci_disable_device(pdev); 1968 pci_disable_device(pdev);
1970 pci_set_drvdata(pdev, NULL);
1971 return err; 1969 return err;
1972 1970
1973} 1971}
diff --git a/drivers/net/ethernet/amd/atarilance.c b/drivers/net/ethernet/amd/atarilance.c
index 10ceca523fc0..e07ce5ff2d48 100644
--- a/drivers/net/ethernet/amd/atarilance.c
+++ b/drivers/net/ethernet/amd/atarilance.c
@@ -586,10 +586,10 @@ static unsigned long __init lance_probe1( struct net_device *dev,
586 switch( lp->cardtype ) { 586 switch( lp->cardtype ) {
587 case OLD_RIEBL: 587 case OLD_RIEBL:
588 /* No ethernet address! (Set some default address) */ 588 /* No ethernet address! (Set some default address) */
589 memcpy( dev->dev_addr, OldRieblDefHwaddr, 6 ); 589 memcpy(dev->dev_addr, OldRieblDefHwaddr, ETH_ALEN);
590 break; 590 break;
591 case NEW_RIEBL: 591 case NEW_RIEBL:
592 lp->memcpy_f( dev->dev_addr, RIEBL_HWADDR_ADDR, 6 ); 592 lp->memcpy_f(dev->dev_addr, RIEBL_HWADDR_ADDR, ETH_ALEN);
593 break; 593 break;
594 case PAM_CARD: 594 case PAM_CARD:
595 i = IO->eeprom; 595 i = IO->eeprom;
diff --git a/drivers/net/ethernet/amd/au1000_eth.c b/drivers/net/ethernet/amd/au1000_eth.c
index 91d52b495848..427c148bb643 100644
--- a/drivers/net/ethernet/amd/au1000_eth.c
+++ b/drivers/net/ethernet/amd/au1000_eth.c
@@ -1138,7 +1138,7 @@ static int au1000_probe(struct platform_device *pdev)
1138 aup->phy1_search_mac0 = 1; 1138 aup->phy1_search_mac0 = 1;
1139 } else { 1139 } else {
1140 if (is_valid_ether_addr(pd->mac)) { 1140 if (is_valid_ether_addr(pd->mac)) {
1141 memcpy(dev->dev_addr, pd->mac, 6); 1141 memcpy(dev->dev_addr, pd->mac, ETH_ALEN);
1142 } else { 1142 } else {
1143 /* Set a random MAC since no valid provided by platform_data. */ 1143 /* Set a random MAC since no valid provided by platform_data. */
1144 eth_hw_addr_random(dev); 1144 eth_hw_addr_random(dev);
diff --git a/drivers/net/ethernet/amd/declance.c b/drivers/net/ethernet/amd/declance.c
index 1b89f1a9127e..57397295887c 100644
--- a/drivers/net/ethernet/amd/declance.c
+++ b/drivers/net/ethernet/amd/declance.c
@@ -344,8 +344,8 @@ static void cp_to_buf(const int type, void *to, const void *from, int len)
344 } 344 }
345 345
346 clen = len & 1; 346 clen = len & 1;
347 rtp = tp; 347 rtp = (unsigned char *)tp;
348 rfp = fp; 348 rfp = (const unsigned char *)fp;
349 while (clen--) { 349 while (clen--) {
350 *rtp++ = *rfp++; 350 *rtp++ = *rfp++;
351 } 351 }
@@ -372,8 +372,8 @@ static void cp_to_buf(const int type, void *to, const void *from, int len)
372 * do the rest, if any. 372 * do the rest, if any.
373 */ 373 */
374 clen = len & 15; 374 clen = len & 15;
375 rtp = (unsigned char *) tp; 375 rtp = (unsigned char *)tp;
376 rfp = (unsigned char *) fp; 376 rfp = (const unsigned char *)fp;
377 while (clen--) { 377 while (clen--) {
378 *rtp++ = *rfp++; 378 *rtp++ = *rfp++;
379 } 379 }
@@ -403,8 +403,8 @@ static void cp_from_buf(const int type, void *to, const void *from, int len)
403 403
404 clen = len & 1; 404 clen = len & 1;
405 405
406 rtp = tp; 406 rtp = (unsigned char *)tp;
407 rfp = fp; 407 rfp = (const unsigned char *)fp;
408 408
409 while (clen--) { 409 while (clen--) {
410 *rtp++ = *rfp++; 410 *rtp++ = *rfp++;
@@ -433,8 +433,8 @@ static void cp_from_buf(const int type, void *to, const void *from, int len)
433 * do the rest, if any. 433 * do the rest, if any.
434 */ 434 */
435 clen = len & 15; 435 clen = len & 15;
436 rtp = (unsigned char *) tp; 436 rtp = (unsigned char *)tp;
437 rfp = (unsigned char *) fp; 437 rfp = (const unsigned char *)fp;
438 while (clen--) { 438 while (clen--) {
439 *rtp++ = *rfp++; 439 *rtp++ = *rfp++;
440 } 440 }
diff --git a/drivers/net/ethernet/amd/lance.c b/drivers/net/ethernet/amd/lance.c
index 5c728436b85e..256f590f6bb1 100644
--- a/drivers/net/ethernet/amd/lance.c
+++ b/drivers/net/ethernet/amd/lance.c
@@ -754,7 +754,7 @@ lance_open(struct net_device *dev)
754 int i; 754 int i;
755 755
756 if (dev->irq == 0 || 756 if (dev->irq == 0 ||
757 request_irq(dev->irq, lance_interrupt, 0, lp->name, dev)) { 757 request_irq(dev->irq, lance_interrupt, 0, dev->name, dev)) {
758 return -EAGAIN; 758 return -EAGAIN;
759 } 759 }
760 760
diff --git a/drivers/net/ethernet/amd/pcnet32.c b/drivers/net/ethernet/amd/pcnet32.c
index 2d8e28819779..38492e0b704e 100644
--- a/drivers/net/ethernet/amd/pcnet32.c
+++ b/drivers/net/ethernet/amd/pcnet32.c
@@ -1675,7 +1675,7 @@ pcnet32_probe1(unsigned long ioaddr, int shared, struct pci_dev *pdev)
1675 pr_cont(" warning: CSR address invalid,\n"); 1675 pr_cont(" warning: CSR address invalid,\n");
1676 pr_info(" using instead PROM address of"); 1676 pr_info(" using instead PROM address of");
1677 } 1677 }
1678 memcpy(dev->dev_addr, promaddr, 6); 1678 memcpy(dev->dev_addr, promaddr, ETH_ALEN);
1679 } 1679 }
1680 } 1680 }
1681 1681
@@ -2818,7 +2818,6 @@ static void pcnet32_remove_one(struct pci_dev *pdev)
2818 lp->init_block, lp->init_dma_addr); 2818 lp->init_block, lp->init_dma_addr);
2819 free_netdev(dev); 2819 free_netdev(dev);
2820 pci_disable_device(pdev); 2820 pci_disable_device(pdev);
2821 pci_set_drvdata(pdev, NULL);
2822 } 2821 }
2823} 2822}
2824 2823
diff --git a/drivers/net/ethernet/apple/bmac.c b/drivers/net/ethernet/apple/bmac.c
index a597b766f080..daae0e016253 100644
--- a/drivers/net/ethernet/apple/bmac.c
+++ b/drivers/net/ethernet/apple/bmac.c
@@ -1220,8 +1220,8 @@ static void bmac_reset_and_enable(struct net_device *dev)
1220 if (skb != NULL) { 1220 if (skb != NULL) {
1221 data = skb_put(skb, ETHERMINPACKET); 1221 data = skb_put(skb, ETHERMINPACKET);
1222 memset(data, 0, ETHERMINPACKET); 1222 memset(data, 0, ETHERMINPACKET);
1223 memcpy(data, dev->dev_addr, 6); 1223 memcpy(data, dev->dev_addr, ETH_ALEN);
1224 memcpy(data+6, dev->dev_addr, 6); 1224 memcpy(data + ETH_ALEN, dev->dev_addr, ETH_ALEN);
1225 bmac_transmit_packet(skb, dev); 1225 bmac_transmit_packet(skb, dev);
1226 } 1226 }
1227 spin_unlock_irqrestore(&bp->lock, flags); 1227 spin_unlock_irqrestore(&bp->lock, flags);
diff --git a/drivers/net/ethernet/apple/macmace.c b/drivers/net/ethernet/apple/macmace.c
index 4ce8ceb62205..58a200df4c35 100644
--- a/drivers/net/ethernet/apple/macmace.c
+++ b/drivers/net/ethernet/apple/macmace.c
@@ -211,6 +211,7 @@ static int mace_probe(struct platform_device *pdev)
211 mp = netdev_priv(dev); 211 mp = netdev_priv(dev);
212 212
213 mp->device = &pdev->dev; 213 mp->device = &pdev->dev;
214 platform_set_drvdata(pdev, dev);
214 SET_NETDEV_DEV(dev, &pdev->dev); 215 SET_NETDEV_DEV(dev, &pdev->dev);
215 216
216 dev->base_addr = (u32)MACE_BASE; 217 dev->base_addr = (u32)MACE_BASE;
diff --git a/drivers/net/ethernet/arc/emac_main.c b/drivers/net/ethernet/arc/emac_main.c
index d0878526c0c8..b2ffad1304d2 100644
--- a/drivers/net/ethernet/arc/emac_main.c
+++ b/drivers/net/ethernet/arc/emac_main.c
@@ -671,6 +671,7 @@ static int arc_emac_probe(struct platform_device *pdev)
671 if (!ndev) 671 if (!ndev)
672 return -ENOMEM; 672 return -ENOMEM;
673 673
674 platform_set_drvdata(pdev, ndev);
674 SET_NETDEV_DEV(ndev, &pdev->dev); 675 SET_NETDEV_DEV(ndev, &pdev->dev);
675 676
676 ndev->netdev_ops = &arc_emac_netdev_ops; 677 ndev->netdev_ops = &arc_emac_netdev_ops;
@@ -725,10 +726,10 @@ static int arc_emac_probe(struct platform_device *pdev)
725 /* Get MAC address from device tree */ 726 /* Get MAC address from device tree */
726 mac_addr = of_get_mac_address(pdev->dev.of_node); 727 mac_addr = of_get_mac_address(pdev->dev.of_node);
727 728
728 if (!mac_addr || !is_valid_ether_addr(mac_addr)) 729 if (mac_addr)
729 eth_hw_addr_random(ndev);
730 else
731 memcpy(ndev->dev_addr, mac_addr, ETH_ALEN); 730 memcpy(ndev->dev_addr, mac_addr, ETH_ALEN);
731 else
732 eth_hw_addr_random(ndev);
732 733
733 dev_info(&pdev->dev, "MAC address is now %pM\n", ndev->dev_addr); 734 dev_info(&pdev->dev, "MAC address is now %pM\n", ndev->dev_addr);
734 735
diff --git a/drivers/net/ethernet/atheros/alx/main.c b/drivers/net/ethernet/atheros/alx/main.c
index fc95b235e210..5aa5e8146496 100644
--- a/drivers/net/ethernet/atheros/alx/main.c
+++ b/drivers/net/ethernet/atheros/alx/main.c
@@ -1367,7 +1367,6 @@ static void alx_remove(struct pci_dev *pdev)
1367 1367
1368 pci_disable_pcie_error_reporting(pdev); 1368 pci_disable_pcie_error_reporting(pdev);
1369 pci_disable_device(pdev); 1369 pci_disable_device(pdev);
1370 pci_set_drvdata(pdev, NULL);
1371 1370
1372 free_netdev(alx->dev); 1371 free_netdev(alx->dev);
1373} 1372}
diff --git a/drivers/net/ethernet/atheros/atl1c/atl1c.h b/drivers/net/ethernet/atheros/atl1c/atl1c.h
index 0f0556526ba9..7f9369a3b378 100644
--- a/drivers/net/ethernet/atheros/atl1c/atl1c.h
+++ b/drivers/net/ethernet/atheros/atl1c/atl1c.h
@@ -600,7 +600,7 @@ struct atl1c_adapter {
600extern char atl1c_driver_name[]; 600extern char atl1c_driver_name[];
601extern char atl1c_driver_version[]; 601extern char atl1c_driver_version[];
602 602
603extern void atl1c_reinit_locked(struct atl1c_adapter *adapter); 603void atl1c_reinit_locked(struct atl1c_adapter *adapter);
604extern s32 atl1c_reset_hw(struct atl1c_hw *hw); 604s32 atl1c_reset_hw(struct atl1c_hw *hw);
605extern void atl1c_set_ethtool_ops(struct net_device *netdev); 605void atl1c_set_ethtool_ops(struct net_device *netdev);
606#endif /* _ATL1C_H_ */ 606#endif /* _ATL1C_H_ */
diff --git a/drivers/net/ethernet/atheros/atl1c/atl1c_hw.c b/drivers/net/ethernet/atheros/atl1c/atl1c_hw.c
index 3ef7092e3f1c..1cda49a28f7f 100644
--- a/drivers/net/ethernet/atheros/atl1c/atl1c_hw.c
+++ b/drivers/net/ethernet/atheros/atl1c/atl1c_hw.c
@@ -153,7 +153,7 @@ static int atl1c_get_permanent_address(struct atl1c_hw *hw)
153bool atl1c_read_eeprom(struct atl1c_hw *hw, u32 offset, u32 *p_value) 153bool atl1c_read_eeprom(struct atl1c_hw *hw, u32 offset, u32 *p_value)
154{ 154{
155 int i; 155 int i;
156 int ret = false; 156 bool ret = false;
157 u32 otp_ctrl_data; 157 u32 otp_ctrl_data;
158 u32 control; 158 u32 control;
159 u32 data; 159 u32 data;
diff --git a/drivers/net/ethernet/atheros/atl1e/atl1e.h b/drivers/net/ethernet/atheros/atl1e/atl1e.h
index b5fd934585e9..1b0fe2d04a0e 100644
--- a/drivers/net/ethernet/atheros/atl1e/atl1e.h
+++ b/drivers/net/ethernet/atheros/atl1e/atl1e.h
@@ -499,10 +499,10 @@ struct atl1e_adapter {
499extern char atl1e_driver_name[]; 499extern char atl1e_driver_name[];
500extern char atl1e_driver_version[]; 500extern char atl1e_driver_version[];
501 501
502extern void atl1e_check_options(struct atl1e_adapter *adapter); 502void atl1e_check_options(struct atl1e_adapter *adapter);
503extern int atl1e_up(struct atl1e_adapter *adapter); 503int atl1e_up(struct atl1e_adapter *adapter);
504extern void atl1e_down(struct atl1e_adapter *adapter); 504void atl1e_down(struct atl1e_adapter *adapter);
505extern void atl1e_reinit_locked(struct atl1e_adapter *adapter); 505void atl1e_reinit_locked(struct atl1e_adapter *adapter);
506extern s32 atl1e_reset_hw(struct atl1e_hw *hw); 506s32 atl1e_reset_hw(struct atl1e_hw *hw);
507extern void atl1e_set_ethtool_ops(struct net_device *netdev); 507void atl1e_set_ethtool_ops(struct net_device *netdev);
508#endif /* _ATL1_E_H_ */ 508#endif /* _ATL1_E_H_ */
diff --git a/drivers/net/ethernet/atheros/atl1e/atl1e_main.c b/drivers/net/ethernet/atheros/atl1e/atl1e_main.c
index 1966444590f6..7a73f3a9fcb5 100644
--- a/drivers/net/ethernet/atheros/atl1e/atl1e_main.c
+++ b/drivers/net/ethernet/atheros/atl1e/atl1e_main.c
@@ -313,6 +313,34 @@ static void atl1e_set_multi(struct net_device *netdev)
313 } 313 }
314} 314}
315 315
316static void __atl1e_rx_mode(netdev_features_t features, u32 *mac_ctrl_data)
317{
318
319 if (features & NETIF_F_RXALL) {
320 /* enable RX of ALL frames */
321 *mac_ctrl_data |= MAC_CTRL_DBG;
322 } else {
323 /* disable RX of ALL frames */
324 *mac_ctrl_data &= ~MAC_CTRL_DBG;
325 }
326}
327
328static void atl1e_rx_mode(struct net_device *netdev,
329 netdev_features_t features)
330{
331 struct atl1e_adapter *adapter = netdev_priv(netdev);
332 u32 mac_ctrl_data = 0;
333
334 netdev_dbg(adapter->netdev, "%s\n", __func__);
335
336 atl1e_irq_disable(adapter);
337 mac_ctrl_data = AT_READ_REG(&adapter->hw, REG_MAC_CTRL);
338 __atl1e_rx_mode(features, &mac_ctrl_data);
339 AT_WRITE_REG(&adapter->hw, REG_MAC_CTRL, mac_ctrl_data);
340 atl1e_irq_enable(adapter);
341}
342
343
316static void __atl1e_vlan_mode(netdev_features_t features, u32 *mac_ctrl_data) 344static void __atl1e_vlan_mode(netdev_features_t features, u32 *mac_ctrl_data)
317{ 345{
318 if (features & NETIF_F_HW_VLAN_CTAG_RX) { 346 if (features & NETIF_F_HW_VLAN_CTAG_RX) {
@@ -394,6 +422,10 @@ static int atl1e_set_features(struct net_device *netdev,
394 if (changed & NETIF_F_HW_VLAN_CTAG_RX) 422 if (changed & NETIF_F_HW_VLAN_CTAG_RX)
395 atl1e_vlan_mode(netdev, features); 423 atl1e_vlan_mode(netdev, features);
396 424
425 if (changed & NETIF_F_RXALL)
426 atl1e_rx_mode(netdev, features);
427
428
397 return 0; 429 return 0;
398} 430}
399 431
@@ -1057,7 +1089,8 @@ static void atl1e_setup_mac_ctrl(struct atl1e_adapter *adapter)
1057 value |= MAC_CTRL_PROMIS_EN; 1089 value |= MAC_CTRL_PROMIS_EN;
1058 if (netdev->flags & IFF_ALLMULTI) 1090 if (netdev->flags & IFF_ALLMULTI)
1059 value |= MAC_CTRL_MC_ALL_EN; 1091 value |= MAC_CTRL_MC_ALL_EN;
1060 1092 if (netdev->features & NETIF_F_RXALL)
1093 value |= MAC_CTRL_DBG;
1061 AT_WRITE_REG(hw, REG_MAC_CTRL, value); 1094 AT_WRITE_REG(hw, REG_MAC_CTRL, value);
1062} 1095}
1063 1096
@@ -1405,7 +1438,8 @@ static void atl1e_clean_rx_irq(struct atl1e_adapter *adapter, u8 que,
1405 rx_page_desc[que].rx_nxseq++; 1438 rx_page_desc[que].rx_nxseq++;
1406 1439
1407 /* error packet */ 1440 /* error packet */
1408 if (prrs->pkt_flag & RRS_IS_ERR_FRAME) { 1441 if ((prrs->pkt_flag & RRS_IS_ERR_FRAME) &&
1442 !(netdev->features & NETIF_F_RXALL)) {
1409 if (prrs->err_flag & (RRS_ERR_BAD_CRC | 1443 if (prrs->err_flag & (RRS_ERR_BAD_CRC |
1410 RRS_ERR_DRIBBLE | RRS_ERR_CODE | 1444 RRS_ERR_DRIBBLE | RRS_ERR_CODE |
1411 RRS_ERR_TRUNC)) { 1445 RRS_ERR_TRUNC)) {
@@ -1418,7 +1452,10 @@ static void atl1e_clean_rx_irq(struct atl1e_adapter *adapter, u8 que,
1418 } 1452 }
1419 1453
1420 packet_size = ((prrs->word1 >> RRS_PKT_SIZE_SHIFT) & 1454 packet_size = ((prrs->word1 >> RRS_PKT_SIZE_SHIFT) &
1421 RRS_PKT_SIZE_MASK) - 4; /* CRC */ 1455 RRS_PKT_SIZE_MASK);
1456 if (likely(!(netdev->features & NETIF_F_RXFCS)))
1457 packet_size -= 4; /* CRC */
1458
1422 skb = netdev_alloc_skb_ip_align(netdev, packet_size); 1459 skb = netdev_alloc_skb_ip_align(netdev, packet_size);
1423 if (skb == NULL) 1460 if (skb == NULL)
1424 goto skip_pkt; 1461 goto skip_pkt;
@@ -2245,7 +2282,8 @@ static int atl1e_init_netdev(struct net_device *netdev, struct pci_dev *pdev)
2245 NETIF_F_HW_VLAN_CTAG_RX; 2282 NETIF_F_HW_VLAN_CTAG_RX;
2246 netdev->features = netdev->hw_features | NETIF_F_LLTX | 2283 netdev->features = netdev->hw_features | NETIF_F_LLTX |
2247 NETIF_F_HW_VLAN_CTAG_TX; 2284 NETIF_F_HW_VLAN_CTAG_TX;
2248 2285 /* not enabled by default */
2286 netdev->hw_features |= NETIF_F_RXALL | NETIF_F_RXFCS;
2249 return 0; 2287 return 0;
2250} 2288}
2251 2289
diff --git a/drivers/net/ethernet/atheros/atlx/atl2.h b/drivers/net/ethernet/atheros/atlx/atl2.h
index 3ebe19f7242b..2f27d4c4c3ad 100644
--- a/drivers/net/ethernet/atheros/atlx/atl2.h
+++ b/drivers/net/ethernet/atheros/atlx/atl2.h
@@ -42,7 +42,7 @@
42#include "atlx.h" 42#include "atlx.h"
43 43
44#ifdef ETHTOOL_OPS_COMPAT 44#ifdef ETHTOOL_OPS_COMPAT
45extern int ethtool_ioctl(struct ifreq *ifr); 45int ethtool_ioctl(struct ifreq *ifr);
46#endif 46#endif
47 47
48#define PCI_COMMAND_REGISTER PCI_COMMAND 48#define PCI_COMMAND_REGISTER PCI_COMMAND
diff --git a/drivers/net/ethernet/broadcom/b44.c b/drivers/net/ethernet/broadcom/b44.c
index 9b017d9c58e9..079a597fa20c 100644
--- a/drivers/net/ethernet/broadcom/b44.c
+++ b/drivers/net/ethernet/broadcom/b44.c
@@ -596,6 +596,7 @@ static void b44_timer(unsigned long __opaque)
596static void b44_tx(struct b44 *bp) 596static void b44_tx(struct b44 *bp)
597{ 597{
598 u32 cur, cons; 598 u32 cur, cons;
599 unsigned bytes_compl = 0, pkts_compl = 0;
599 600
600 cur = br32(bp, B44_DMATX_STAT) & DMATX_STAT_CDMASK; 601 cur = br32(bp, B44_DMATX_STAT) & DMATX_STAT_CDMASK;
601 cur /= sizeof(struct dma_desc); 602 cur /= sizeof(struct dma_desc);
@@ -612,9 +613,14 @@ static void b44_tx(struct b44 *bp)
612 skb->len, 613 skb->len,
613 DMA_TO_DEVICE); 614 DMA_TO_DEVICE);
614 rp->skb = NULL; 615 rp->skb = NULL;
616
617 bytes_compl += skb->len;
618 pkts_compl++;
619
615 dev_kfree_skb_irq(skb); 620 dev_kfree_skb_irq(skb);
616 } 621 }
617 622
623 netdev_completed_queue(bp->dev, pkts_compl, bytes_compl);
618 bp->tx_cons = cons; 624 bp->tx_cons = cons;
619 if (netif_queue_stopped(bp->dev) && 625 if (netif_queue_stopped(bp->dev) &&
620 TX_BUFFS_AVAIL(bp) > B44_TX_WAKEUP_THRESH) 626 TX_BUFFS_AVAIL(bp) > B44_TX_WAKEUP_THRESH)
@@ -1018,6 +1024,8 @@ static netdev_tx_t b44_start_xmit(struct sk_buff *skb, struct net_device *dev)
1018 if (bp->flags & B44_FLAG_REORDER_BUG) 1024 if (bp->flags & B44_FLAG_REORDER_BUG)
1019 br32(bp, B44_DMATX_PTR); 1025 br32(bp, B44_DMATX_PTR);
1020 1026
1027 netdev_sent_queue(dev, skb->len);
1028
1021 if (TX_BUFFS_AVAIL(bp) < 1) 1029 if (TX_BUFFS_AVAIL(bp) < 1)
1022 netif_stop_queue(dev); 1030 netif_stop_queue(dev);
1023 1031
@@ -1416,6 +1424,8 @@ static void b44_init_hw(struct b44 *bp, int reset_kind)
1416 1424
1417 val = br32(bp, B44_ENET_CTRL); 1425 val = br32(bp, B44_ENET_CTRL);
1418 bw32(bp, B44_ENET_CTRL, (val | ENET_CTRL_ENABLE)); 1426 bw32(bp, B44_ENET_CTRL, (val | ENET_CTRL_ENABLE));
1427
1428 netdev_reset_queue(bp->dev);
1419} 1429}
1420 1430
1421static int b44_open(struct net_device *dev) 1431static int b44_open(struct net_device *dev)
@@ -2101,7 +2111,7 @@ static int b44_get_invariants(struct b44 *bp)
2101 * valid PHY address. */ 2111 * valid PHY address. */
2102 bp->phy_addr &= 0x1F; 2112 bp->phy_addr &= 0x1F;
2103 2113
2104 memcpy(bp->dev->dev_addr, addr, 6); 2114 memcpy(bp->dev->dev_addr, addr, ETH_ALEN);
2105 2115
2106 if (!is_valid_ether_addr(&bp->dev->dev_addr[0])){ 2116 if (!is_valid_ether_addr(&bp->dev->dev_addr[0])){
2107 pr_err("Invalid MAC address found in EEPROM\n"); 2117 pr_err("Invalid MAC address found in EEPROM\n");
diff --git a/drivers/net/ethernet/broadcom/bgmac.c b/drivers/net/ethernet/broadcom/bgmac.c
index 9e8a3e024e01..e2aa09ce6af7 100644
--- a/drivers/net/ethernet/broadcom/bgmac.c
+++ b/drivers/net/ethernet/broadcom/bgmac.c
@@ -149,6 +149,8 @@ static netdev_tx_t bgmac_dma_tx_add(struct bgmac *bgmac,
149 dma_desc->ctl0 = cpu_to_le32(ctl0); 149 dma_desc->ctl0 = cpu_to_le32(ctl0);
150 dma_desc->ctl1 = cpu_to_le32(ctl1); 150 dma_desc->ctl1 = cpu_to_le32(ctl1);
151 151
152 netdev_sent_queue(net_dev, skb->len);
153
152 wmb(); 154 wmb();
153 155
154 /* Increase ring->end to point empty slot. We tell hardware the first 156 /* Increase ring->end to point empty slot. We tell hardware the first
@@ -178,6 +180,7 @@ static void bgmac_dma_tx_free(struct bgmac *bgmac, struct bgmac_dma_ring *ring)
178 struct device *dma_dev = bgmac->core->dma_dev; 180 struct device *dma_dev = bgmac->core->dma_dev;
179 int empty_slot; 181 int empty_slot;
180 bool freed = false; 182 bool freed = false;
183 unsigned bytes_compl = 0, pkts_compl = 0;
181 184
182 /* The last slot that hardware didn't consume yet */ 185 /* The last slot that hardware didn't consume yet */
183 empty_slot = bgmac_read(bgmac, ring->mmio_base + BGMAC_DMA_TX_STATUS); 186 empty_slot = bgmac_read(bgmac, ring->mmio_base + BGMAC_DMA_TX_STATUS);
@@ -195,6 +198,9 @@ static void bgmac_dma_tx_free(struct bgmac *bgmac, struct bgmac_dma_ring *ring)
195 slot->skb->len, DMA_TO_DEVICE); 198 slot->skb->len, DMA_TO_DEVICE);
196 slot->dma_addr = 0; 199 slot->dma_addr = 0;
197 200
201 bytes_compl += slot->skb->len;
202 pkts_compl++;
203
198 /* Free memory! :) */ 204 /* Free memory! :) */
199 dev_kfree_skb(slot->skb); 205 dev_kfree_skb(slot->skb);
200 slot->skb = NULL; 206 slot->skb = NULL;
@@ -208,6 +214,8 @@ static void bgmac_dma_tx_free(struct bgmac *bgmac, struct bgmac_dma_ring *ring)
208 freed = true; 214 freed = true;
209 } 215 }
210 216
217 netdev_completed_queue(bgmac->net_dev, pkts_compl, bytes_compl);
218
211 if (freed && netif_queue_stopped(bgmac->net_dev)) 219 if (freed && netif_queue_stopped(bgmac->net_dev))
212 netif_wake_queue(bgmac->net_dev); 220 netif_wake_queue(bgmac->net_dev);
213} 221}
@@ -277,6 +285,26 @@ static int bgmac_dma_rx_skb_for_slot(struct bgmac *bgmac,
277 return 0; 285 return 0;
278} 286}
279 287
288static void bgmac_dma_rx_setup_desc(struct bgmac *bgmac,
289 struct bgmac_dma_ring *ring, int desc_idx)
290{
291 struct bgmac_dma_desc *dma_desc = ring->cpu_base + desc_idx;
292 u32 ctl0 = 0, ctl1 = 0;
293
294 if (desc_idx == ring->num_slots - 1)
295 ctl0 |= BGMAC_DESC_CTL0_EOT;
296 ctl1 |= BGMAC_RX_BUF_SIZE & BGMAC_DESC_CTL1_LEN;
297 /* Is there any BGMAC device that requires extension? */
298 /* ctl1 |= (addrext << B43_DMA64_DCTL1_ADDREXT_SHIFT) &
299 * B43_DMA64_DCTL1_ADDREXT_MASK;
300 */
301
302 dma_desc->addr_low = cpu_to_le32(lower_32_bits(ring->slots[desc_idx].dma_addr));
303 dma_desc->addr_high = cpu_to_le32(upper_32_bits(ring->slots[desc_idx].dma_addr));
304 dma_desc->ctl0 = cpu_to_le32(ctl0);
305 dma_desc->ctl1 = cpu_to_le32(ctl1);
306}
307
280static int bgmac_dma_rx_read(struct bgmac *bgmac, struct bgmac_dma_ring *ring, 308static int bgmac_dma_rx_read(struct bgmac *bgmac, struct bgmac_dma_ring *ring,
281 int weight) 309 int weight)
282{ 310{
@@ -295,7 +323,6 @@ static int bgmac_dma_rx_read(struct bgmac *bgmac, struct bgmac_dma_ring *ring,
295 struct device *dma_dev = bgmac->core->dma_dev; 323 struct device *dma_dev = bgmac->core->dma_dev;
296 struct bgmac_slot_info *slot = &ring->slots[ring->start]; 324 struct bgmac_slot_info *slot = &ring->slots[ring->start];
297 struct sk_buff *skb = slot->skb; 325 struct sk_buff *skb = slot->skb;
298 struct sk_buff *new_skb;
299 struct bgmac_rx_header *rx; 326 struct bgmac_rx_header *rx;
300 u16 len, flags; 327 u16 len, flags;
301 328
@@ -308,38 +335,51 @@ static int bgmac_dma_rx_read(struct bgmac *bgmac, struct bgmac_dma_ring *ring,
308 len = le16_to_cpu(rx->len); 335 len = le16_to_cpu(rx->len);
309 flags = le16_to_cpu(rx->flags); 336 flags = le16_to_cpu(rx->flags);
310 337
311 /* Check for poison and drop or pass the packet */ 338 do {
312 if (len == 0xdead && flags == 0xbeef) { 339 dma_addr_t old_dma_addr = slot->dma_addr;
313 bgmac_err(bgmac, "Found poisoned packet at slot %d, DMA issue!\n", 340 int err;
314 ring->start); 341
315 } else { 342 /* Check for poison and drop or pass the packet */
343 if (len == 0xdead && flags == 0xbeef) {
344 bgmac_err(bgmac, "Found poisoned packet at slot %d, DMA issue!\n",
345 ring->start);
346 dma_sync_single_for_device(dma_dev,
347 slot->dma_addr,
348 BGMAC_RX_BUF_SIZE,
349 DMA_FROM_DEVICE);
350 break;
351 }
352
316 /* Omit CRC. */ 353 /* Omit CRC. */
317 len -= ETH_FCS_LEN; 354 len -= ETH_FCS_LEN;
318 355
319 new_skb = netdev_alloc_skb_ip_align(bgmac->net_dev, len); 356 /* Prepare new skb as replacement */
320 if (new_skb) { 357 err = bgmac_dma_rx_skb_for_slot(bgmac, slot);
321 skb_put(new_skb, len); 358 if (err) {
322 skb_copy_from_linear_data_offset(skb, BGMAC_RX_FRAME_OFFSET, 359 /* Poison the old skb */
323 new_skb->data, 360 rx->len = cpu_to_le16(0xdead);
324 len); 361 rx->flags = cpu_to_le16(0xbeef);
325 skb_checksum_none_assert(skb); 362
326 new_skb->protocol = 363 dma_sync_single_for_device(dma_dev,
327 eth_type_trans(new_skb, bgmac->net_dev); 364 slot->dma_addr,
328 netif_receive_skb(new_skb); 365 BGMAC_RX_BUF_SIZE,
329 handled++; 366 DMA_FROM_DEVICE);
330 } else { 367 break;
331 bgmac->net_dev->stats.rx_dropped++;
332 bgmac_err(bgmac, "Allocation of skb for copying packet failed!\n");
333 } 368 }
369 bgmac_dma_rx_setup_desc(bgmac, ring, ring->start);
334 370
335 /* Poison the old skb */ 371 /* Unmap old skb, we'll pass it to the netfif */
336 rx->len = cpu_to_le16(0xdead); 372 dma_unmap_single(dma_dev, old_dma_addr,
337 rx->flags = cpu_to_le16(0xbeef); 373 BGMAC_RX_BUF_SIZE, DMA_FROM_DEVICE);
338 }
339 374
340 /* Make it back accessible to the hardware */ 375 skb_put(skb, BGMAC_RX_FRAME_OFFSET + len);
341 dma_sync_single_for_device(dma_dev, slot->dma_addr, 376 skb_pull(skb, BGMAC_RX_FRAME_OFFSET);
342 BGMAC_RX_BUF_SIZE, DMA_FROM_DEVICE); 377
378 skb_checksum_none_assert(skb);
379 skb->protocol = eth_type_trans(skb, bgmac->net_dev);
380 netif_receive_skb(skb);
381 handled++;
382 } while (0);
343 383
344 if (++ring->start >= BGMAC_RX_RING_SLOTS) 384 if (++ring->start >= BGMAC_RX_RING_SLOTS)
345 ring->start = 0; 385 ring->start = 0;
@@ -503,8 +543,6 @@ err_dma_free:
503static void bgmac_dma_init(struct bgmac *bgmac) 543static void bgmac_dma_init(struct bgmac *bgmac)
504{ 544{
505 struct bgmac_dma_ring *ring; 545 struct bgmac_dma_ring *ring;
506 struct bgmac_dma_desc *dma_desc;
507 u32 ctl0, ctl1;
508 int i; 546 int i;
509 547
510 for (i = 0; i < BGMAC_MAX_TX_RINGS; i++) { 548 for (i = 0; i < BGMAC_MAX_TX_RINGS; i++) {
@@ -537,23 +575,8 @@ static void bgmac_dma_init(struct bgmac *bgmac)
537 if (ring->unaligned) 575 if (ring->unaligned)
538 bgmac_dma_rx_enable(bgmac, ring); 576 bgmac_dma_rx_enable(bgmac, ring);
539 577
540 for (j = 0, dma_desc = ring->cpu_base; j < ring->num_slots; 578 for (j = 0; j < ring->num_slots; j++)
541 j++, dma_desc++) { 579 bgmac_dma_rx_setup_desc(bgmac, ring, j);
542 ctl0 = ctl1 = 0;
543
544 if (j == ring->num_slots - 1)
545 ctl0 |= BGMAC_DESC_CTL0_EOT;
546 ctl1 |= BGMAC_RX_BUF_SIZE & BGMAC_DESC_CTL1_LEN;
547 /* Is there any BGMAC device that requires extension? */
548 /* ctl1 |= (addrext << B43_DMA64_DCTL1_ADDREXT_SHIFT) &
549 * B43_DMA64_DCTL1_ADDREXT_MASK;
550 */
551
552 dma_desc->addr_low = cpu_to_le32(lower_32_bits(ring->slots[j].dma_addr));
553 dma_desc->addr_high = cpu_to_le32(upper_32_bits(ring->slots[j].dma_addr));
554 dma_desc->ctl0 = cpu_to_le32(ctl0);
555 dma_desc->ctl1 = cpu_to_le32(ctl1);
556 }
557 580
558 bgmac_write(bgmac, ring->mmio_base + BGMAC_DMA_RX_INDEX, 581 bgmac_write(bgmac, ring->mmio_base + BGMAC_DMA_RX_INDEX,
559 ring->index_base + 582 ring->index_base +
@@ -996,6 +1019,8 @@ static void bgmac_chip_reset(struct bgmac *bgmac)
996 bgmac_miiconfig(bgmac); 1019 bgmac_miiconfig(bgmac);
997 bgmac_phy_init(bgmac); 1020 bgmac_phy_init(bgmac);
998 1021
1022 netdev_reset_queue(bgmac->net_dev);
1023
999 bgmac->int_status = 0; 1024 bgmac->int_status = 0;
1000} 1025}
1001 1026
diff --git a/drivers/net/ethernet/broadcom/bnx2.c b/drivers/net/ethernet/broadcom/bnx2.c
index e838a3f74b69..d9980ad00b4b 100644
--- a/drivers/net/ethernet/broadcom/bnx2.c
+++ b/drivers/net/ethernet/broadcom/bnx2.c
@@ -5761,8 +5761,8 @@ bnx2_run_loopback(struct bnx2 *bp, int loopback_mode)
5761 if (!skb) 5761 if (!skb)
5762 return -ENOMEM; 5762 return -ENOMEM;
5763 packet = skb_put(skb, pkt_size); 5763 packet = skb_put(skb, pkt_size);
5764 memcpy(packet, bp->dev->dev_addr, 6); 5764 memcpy(packet, bp->dev->dev_addr, ETH_ALEN);
5765 memset(packet + 6, 0x0, 8); 5765 memset(packet + ETH_ALEN, 0x0, 8);
5766 for (i = 14; i < pkt_size; i++) 5766 for (i = 14; i < pkt_size; i++)
5767 packet[i] = (unsigned char) (i & 0xff); 5767 packet[i] = (unsigned char) (i & 0xff);
5768 5768
@@ -8413,7 +8413,6 @@ err_out_release:
8413 8413
8414err_out_disable: 8414err_out_disable:
8415 pci_disable_device(pdev); 8415 pci_disable_device(pdev);
8416 pci_set_drvdata(pdev, NULL);
8417 8416
8418err_out: 8417err_out:
8419 return rc; 8418 return rc;
@@ -8514,7 +8513,7 @@ bnx2_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
8514 8513
8515 pci_set_drvdata(pdev, dev); 8514 pci_set_drvdata(pdev, dev);
8516 8515
8517 memcpy(dev->dev_addr, bp->mac_addr, 6); 8516 memcpy(dev->dev_addr, bp->mac_addr, ETH_ALEN);
8518 8517
8519 dev->hw_features = NETIF_F_IP_CSUM | NETIF_F_SG | 8518 dev->hw_features = NETIF_F_IP_CSUM | NETIF_F_SG |
8520 NETIF_F_TSO | NETIF_F_TSO_ECN | 8519 NETIF_F_TSO | NETIF_F_TSO_ECN |
@@ -8546,7 +8545,6 @@ error:
8546 pci_iounmap(pdev, bp->regview); 8545 pci_iounmap(pdev, bp->regview);
8547 pci_release_regions(pdev); 8546 pci_release_regions(pdev);
8548 pci_disable_device(pdev); 8547 pci_disable_device(pdev);
8549 pci_set_drvdata(pdev, NULL);
8550err_free: 8548err_free:
8551 free_netdev(dev); 8549 free_netdev(dev);
8552 return rc; 8550 return rc;
@@ -8578,7 +8576,6 @@ bnx2_remove_one(struct pci_dev *pdev)
8578 8576
8579 pci_release_regions(pdev); 8577 pci_release_regions(pdev);
8580 pci_disable_device(pdev); 8578 pci_disable_device(pdev);
8581 pci_set_drvdata(pdev, NULL);
8582} 8579}
8583 8580
8584static int 8581static int
diff --git a/drivers/net/ethernet/broadcom/bnx2x/bnx2x.h b/drivers/net/ethernet/broadcom/bnx2x/bnx2x.h
index c5e375ddd6c0..4e01c57d8c8d 100644
--- a/drivers/net/ethernet/broadcom/bnx2x/bnx2x.h
+++ b/drivers/net/ethernet/broadcom/bnx2x/bnx2x.h
@@ -1546,6 +1546,7 @@ struct bnx2x {
1546#define IS_VF_FLAG (1 << 22) 1546#define IS_VF_FLAG (1 << 22)
1547#define INTERRUPTS_ENABLED_FLAG (1 << 23) 1547#define INTERRUPTS_ENABLED_FLAG (1 << 23)
1548#define BC_SUPPORTS_RMMOD_CMD (1 << 24) 1548#define BC_SUPPORTS_RMMOD_CMD (1 << 24)
1549#define HAS_PHYS_PORT_ID (1 << 25)
1549 1550
1550#define BP_NOMCP(bp) ((bp)->flags & NO_MCP_FLAG) 1551#define BP_NOMCP(bp) ((bp)->flags & NO_MCP_FLAG)
1551 1552
@@ -1876,6 +1877,8 @@ struct bnx2x {
1876 u32 dump_preset_idx; 1877 u32 dump_preset_idx;
1877 bool stats_started; 1878 bool stats_started;
1878 struct semaphore stats_sema; 1879 struct semaphore stats_sema;
1880
1881 u8 phys_port_id[ETH_ALEN];
1879}; 1882};
1880 1883
1881/* Tx queues may be less or equal to Rx queues */ 1884/* Tx queues may be less or equal to Rx queues */
@@ -2232,7 +2235,7 @@ void bnx2x_igu_clear_sb_gen(struct bnx2x *bp, u8 func, u8 idu_sb_id,
2232#define BNX2X_NUM_TESTS_SF 7 2235#define BNX2X_NUM_TESTS_SF 7
2233#define BNX2X_NUM_TESTS_MF 3 2236#define BNX2X_NUM_TESTS_MF 3
2234#define BNX2X_NUM_TESTS(bp) (IS_MF(bp) ? BNX2X_NUM_TESTS_MF : \ 2237#define BNX2X_NUM_TESTS(bp) (IS_MF(bp) ? BNX2X_NUM_TESTS_MF : \
2235 BNX2X_NUM_TESTS_SF) 2238 IS_VF(bp) ? 0 : BNX2X_NUM_TESTS_SF)
2236 2239
2237#define BNX2X_PHY_LOOPBACK 0 2240#define BNX2X_PHY_LOOPBACK 0
2238#define BNX2X_MAC_LOOPBACK 1 2241#define BNX2X_MAC_LOOPBACK 1
@@ -2492,12 +2495,6 @@ enum {
2492 2495
2493#define NUM_MACS 8 2496#define NUM_MACS 8
2494 2497
2495enum bnx2x_pci_bus_speed {
2496 BNX2X_PCI_LINK_SPEED_2500 = 2500,
2497 BNX2X_PCI_LINK_SPEED_5000 = 5000,
2498 BNX2X_PCI_LINK_SPEED_8000 = 8000
2499};
2500
2501void bnx2x_set_local_cmng(struct bnx2x *bp); 2498void bnx2x_set_local_cmng(struct bnx2x *bp);
2502 2499
2503#define MCPR_SCRATCH_BASE(bp) \ 2500#define MCPR_SCRATCH_BASE(bp) \
diff --git a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
index 74d6486fccfd..dcafbda3e5be 100644
--- a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
+++ b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
@@ -3256,14 +3256,16 @@ static u32 bnx2x_xmit_type(struct bnx2x *bp, struct sk_buff *skb)
3256 if (prot == IPPROTO_TCP) 3256 if (prot == IPPROTO_TCP)
3257 rc |= XMIT_CSUM_TCP; 3257 rc |= XMIT_CSUM_TCP;
3258 3258
3259 if (skb_is_gso_v6(skb)) { 3259 if (skb_is_gso(skb)) {
3260 rc |= (XMIT_GSO_V6 | XMIT_CSUM_TCP); 3260 if (skb_is_gso_v6(skb)) {
3261 if (rc & XMIT_CSUM_ENC) 3261 rc |= (XMIT_GSO_V6 | XMIT_CSUM_TCP);
3262 rc |= XMIT_GSO_ENC_V6; 3262 if (rc & XMIT_CSUM_ENC)
3263 } else if (skb_is_gso(skb)) { 3263 rc |= XMIT_GSO_ENC_V6;
3264 rc |= (XMIT_GSO_V4 | XMIT_CSUM_TCP); 3264 } else {
3265 if (rc & XMIT_CSUM_ENC) 3265 rc |= (XMIT_GSO_V4 | XMIT_CSUM_TCP);
3266 rc |= XMIT_GSO_ENC_V4; 3266 if (rc & XMIT_CSUM_ENC)
3267 rc |= XMIT_GSO_ENC_V4;
3268 }
3267 } 3269 }
3268 3270
3269 return rc; 3271 return rc;
diff --git a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c
index e8efa1c93ffe..32d0f1435fb4 100644
--- a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c
+++ b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_ethtool.c
@@ -639,6 +639,9 @@ static int bnx2x_get_regs_len(struct net_device *dev)
639 struct bnx2x *bp = netdev_priv(dev); 639 struct bnx2x *bp = netdev_priv(dev);
640 int regdump_len = 0; 640 int regdump_len = 0;
641 641
642 if (IS_VF(bp))
643 return 0;
644
642 regdump_len = __bnx2x_get_regs_len(bp); 645 regdump_len = __bnx2x_get_regs_len(bp);
643 regdump_len *= 4; 646 regdump_len *= 4;
644 regdump_len += sizeof(struct dump_header); 647 regdump_len += sizeof(struct dump_header);
@@ -2864,9 +2867,16 @@ static void bnx2x_self_test(struct net_device *dev,
2864 2867
2865 memset(buf, 0, sizeof(u64) * BNX2X_NUM_TESTS(bp)); 2868 memset(buf, 0, sizeof(u64) * BNX2X_NUM_TESTS(bp));
2866 2869
2870 if (bnx2x_test_nvram(bp) != 0) {
2871 if (!IS_MF(bp))
2872 buf[4] = 1;
2873 else
2874 buf[0] = 1;
2875 etest->flags |= ETH_TEST_FL_FAILED;
2876 }
2877
2867 if (!netif_running(dev)) { 2878 if (!netif_running(dev)) {
2868 DP(BNX2X_MSG_ETHTOOL, 2879 DP(BNX2X_MSG_ETHTOOL, "Interface is down\n");
2869 "Can't perform self-test when interface is down\n");
2870 return; 2880 return;
2871 } 2881 }
2872 2882
@@ -2928,13 +2938,7 @@ static void bnx2x_self_test(struct net_device *dev,
2928 /* wait until link state is restored */ 2938 /* wait until link state is restored */
2929 bnx2x_wait_for_link(bp, link_up, is_serdes); 2939 bnx2x_wait_for_link(bp, link_up, is_serdes);
2930 } 2940 }
2931 if (bnx2x_test_nvram(bp) != 0) { 2941
2932 if (!IS_MF(bp))
2933 buf[4] = 1;
2934 else
2935 buf[0] = 1;
2936 etest->flags |= ETH_TEST_FL_FAILED;
2937 }
2938 if (bnx2x_test_intr(bp) != 0) { 2942 if (bnx2x_test_intr(bp) != 0) {
2939 if (!IS_MF(bp)) 2943 if (!IS_MF(bp))
2940 buf[5] = 1; 2944 buf[5] = 1;
diff --git a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_hsi.h b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_hsi.h
index 32767f6aa33f..cf1df8b62e2c 100644
--- a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_hsi.h
+++ b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_hsi.h
@@ -172,6 +172,7 @@ struct shared_hw_cfg { /* NVRAM Offset */
172 #define SHARED_HW_CFG_LED_MAC4 0x000c0000 172 #define SHARED_HW_CFG_LED_MAC4 0x000c0000
173 #define SHARED_HW_CFG_LED_PHY8 0x000d0000 173 #define SHARED_HW_CFG_LED_PHY8 0x000d0000
174 #define SHARED_HW_CFG_LED_EXTPHY1 0x000e0000 174 #define SHARED_HW_CFG_LED_EXTPHY1 0x000e0000
175 #define SHARED_HW_CFG_LED_EXTPHY2 0x000f0000
175 176
176 177
177 #define SHARED_HW_CFG_AN_ENABLE_MASK 0x3f000000 178 #define SHARED_HW_CFG_AN_ENABLE_MASK 0x3f000000
diff --git a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c
index 51468227bf3b..20dcc02431ca 100644
--- a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c
+++ b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c
@@ -3122,7 +3122,7 @@ static void bnx2x_bsc_module_sel(struct link_params *params)
3122} 3122}
3123 3123
3124static int bnx2x_bsc_read(struct link_params *params, 3124static int bnx2x_bsc_read(struct link_params *params,
3125 struct bnx2x_phy *phy, 3125 struct bnx2x *bp,
3126 u8 sl_devid, 3126 u8 sl_devid,
3127 u16 sl_addr, 3127 u16 sl_addr,
3128 u8 lc_addr, 3128 u8 lc_addr,
@@ -3131,7 +3131,6 @@ static int bnx2x_bsc_read(struct link_params *params,
3131{ 3131{
3132 u32 val, i; 3132 u32 val, i;
3133 int rc = 0; 3133 int rc = 0;
3134 struct bnx2x *bp = params->bp;
3135 3134
3136 if (xfer_cnt > 16) { 3135 if (xfer_cnt > 16) {
3137 DP(NETIF_MSG_LINK, "invalid xfer_cnt %d. Max is 16 bytes\n", 3136 DP(NETIF_MSG_LINK, "invalid xfer_cnt %d. Max is 16 bytes\n",
@@ -6371,9 +6370,15 @@ int bnx2x_set_led(struct link_params *params,
6371 * intended override. 6370 * intended override.
6372 */ 6371 */
6373 break; 6372 break;
6374 } else 6373 } else {
6374 u32 nig_led_mode = ((params->hw_led_mode <<
6375 SHARED_HW_CFG_LED_MODE_SHIFT) ==
6376 SHARED_HW_CFG_LED_EXTPHY2) ?
6377 (SHARED_HW_CFG_LED_PHY1 >>
6378 SHARED_HW_CFG_LED_MODE_SHIFT) : hw_led_mode;
6375 REG_WR(bp, NIG_REG_LED_MODE_P0 + port*4, 6379 REG_WR(bp, NIG_REG_LED_MODE_P0 + port*4,
6376 hw_led_mode); 6380 nig_led_mode);
6381 }
6377 6382
6378 REG_WR(bp, NIG_REG_LED_CONTROL_OVERRIDE_TRAFFIC_P0 + port*4, 0); 6383 REG_WR(bp, NIG_REG_LED_CONTROL_OVERRIDE_TRAFFIC_P0 + port*4, 0);
6379 /* Set blinking rate to ~15.9Hz */ 6384 /* Set blinking rate to ~15.9Hz */
@@ -7917,7 +7922,7 @@ static int bnx2x_warpcore_read_sfp_module_eeprom(struct bnx2x_phy *phy,
7917 usleep_range(1000, 2000); 7922 usleep_range(1000, 2000);
7918 bnx2x_warpcore_power_module(params, 1); 7923 bnx2x_warpcore_power_module(params, 1);
7919 } 7924 }
7920 rc = bnx2x_bsc_read(params, phy, dev_addr, addr32, 0, byte_cnt, 7925 rc = bnx2x_bsc_read(params, bp, dev_addr, addr32, 0, byte_cnt,
7921 data_array); 7926 data_array);
7922 } while ((rc != 0) && (++cnt < I2C_WA_RETRY_CNT)); 7927 } while ((rc != 0) && (++cnt < I2C_WA_RETRY_CNT));
7923 7928
@@ -10653,10 +10658,18 @@ static void bnx2x_848xx_set_link_led(struct bnx2x_phy *phy,
10653 0x40); 10658 0x40);
10654 10659
10655 } else { 10660 } else {
10661 /* EXTPHY2 LED mode indicate that the 100M/1G/10G LED
10662 * sources are all wired through LED1, rather than only
10663 * 10G in other modes.
10664 */
10665 val = ((params->hw_led_mode <<
10666 SHARED_HW_CFG_LED_MODE_SHIFT) ==
10667 SHARED_HW_CFG_LED_EXTPHY2) ? 0x98 : 0x80;
10668
10656 bnx2x_cl45_write(bp, phy, 10669 bnx2x_cl45_write(bp, phy,
10657 MDIO_PMA_DEVAD, 10670 MDIO_PMA_DEVAD,
10658 MDIO_PMA_REG_8481_LED1_MASK, 10671 MDIO_PMA_REG_8481_LED1_MASK,
10659 0x80); 10672 val);
10660 10673
10661 /* Tell LED3 to blink on source */ 10674 /* Tell LED3 to blink on source */
10662 bnx2x_cl45_read(bp, phy, 10675 bnx2x_cl45_read(bp, phy,
diff --git a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
index b42f89ce02ef..bb2f20291509 100644
--- a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
+++ b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
@@ -9916,7 +9916,7 @@ static int bnx2x_prev_path_mark_eeh(struct bnx2x *bp)
9916static bool bnx2x_prev_is_path_marked(struct bnx2x *bp) 9916static bool bnx2x_prev_is_path_marked(struct bnx2x *bp)
9917{ 9917{
9918 struct bnx2x_prev_path_list *tmp_list; 9918 struct bnx2x_prev_path_list *tmp_list;
9919 int rc = false; 9919 bool rc = false;
9920 9920
9921 if (down_trylock(&bnx2x_prev_sem)) 9921 if (down_trylock(&bnx2x_prev_sem))
9922 return false; 9922 return false;
@@ -11186,6 +11186,14 @@ static void bnx2x_get_mac_hwinfo(struct bnx2x *bp)
11186 bnx2x_get_cnic_mac_hwinfo(bp); 11186 bnx2x_get_cnic_mac_hwinfo(bp);
11187 } 11187 }
11188 11188
11189 if (!BP_NOMCP(bp)) {
11190 /* Read physical port identifier from shmem */
11191 val2 = SHMEM_RD(bp, dev_info.port_hw_config[port].mac_upper);
11192 val = SHMEM_RD(bp, dev_info.port_hw_config[port].mac_lower);
11193 bnx2x_set_mac_buf(bp->phys_port_id, val, val2);
11194 bp->flags |= HAS_PHYS_PORT_ID;
11195 }
11196
11189 memcpy(bp->link_params.mac_addr, bp->dev->dev_addr, ETH_ALEN); 11197 memcpy(bp->link_params.mac_addr, bp->dev->dev_addr, ETH_ALEN);
11190 11198
11191 if (!bnx2x_is_valid_ether_addr(bp, bp->dev->dev_addr)) 11199 if (!bnx2x_is_valid_ether_addr(bp, bp->dev->dev_addr))
@@ -11784,7 +11792,7 @@ static int bnx2x_open(struct net_device *dev)
11784 rc = bnx2x_nic_load(bp, LOAD_OPEN); 11792 rc = bnx2x_nic_load(bp, LOAD_OPEN);
11785 if (rc) 11793 if (rc)
11786 return rc; 11794 return rc;
11787 return bnx2x_open_epilog(bp); 11795 return 0;
11788} 11796}
11789 11797
11790/* called with rtnl_lock */ 11798/* called with rtnl_lock */
@@ -12082,6 +12090,20 @@ static int bnx2x_validate_addr(struct net_device *dev)
12082 return 0; 12090 return 0;
12083} 12091}
12084 12092
12093static int bnx2x_get_phys_port_id(struct net_device *netdev,
12094 struct netdev_phys_port_id *ppid)
12095{
12096 struct bnx2x *bp = netdev_priv(netdev);
12097
12098 if (!(bp->flags & HAS_PHYS_PORT_ID))
12099 return -EOPNOTSUPP;
12100
12101 ppid->id_len = sizeof(bp->phys_port_id);
12102 memcpy(ppid->id, bp->phys_port_id, ppid->id_len);
12103
12104 return 0;
12105}
12106
12085static const struct net_device_ops bnx2x_netdev_ops = { 12107static const struct net_device_ops bnx2x_netdev_ops = {
12086 .ndo_open = bnx2x_open, 12108 .ndo_open = bnx2x_open,
12087 .ndo_stop = bnx2x_close, 12109 .ndo_stop = bnx2x_close,
@@ -12111,6 +12133,7 @@ static const struct net_device_ops bnx2x_netdev_ops = {
12111#ifdef CONFIG_NET_RX_BUSY_POLL 12133#ifdef CONFIG_NET_RX_BUSY_POLL
12112 .ndo_busy_poll = bnx2x_low_latency_recv, 12134 .ndo_busy_poll = bnx2x_low_latency_recv,
12113#endif 12135#endif
12136 .ndo_get_phys_port_id = bnx2x_get_phys_port_id,
12114}; 12137};
12115 12138
12116static int bnx2x_set_coherency_mask(struct bnx2x *bp) 12139static int bnx2x_set_coherency_mask(struct bnx2x *bp)
@@ -12274,10 +12297,13 @@ static int bnx2x_init_dev(struct bnx2x *bp, struct pci_dev *pdev,
12274 NETIF_F_RXCSUM | NETIF_F_LRO | NETIF_F_GRO | 12297 NETIF_F_RXCSUM | NETIF_F_LRO | NETIF_F_GRO |
12275 NETIF_F_RXHASH | NETIF_F_HW_VLAN_CTAG_TX; 12298 NETIF_F_RXHASH | NETIF_F_HW_VLAN_CTAG_TX;
12276 if (!CHIP_IS_E1x(bp)) { 12299 if (!CHIP_IS_E1x(bp)) {
12277 dev->hw_features |= NETIF_F_GSO_GRE | NETIF_F_GSO_UDP_TUNNEL; 12300 dev->hw_features |= NETIF_F_GSO_GRE | NETIF_F_GSO_UDP_TUNNEL |
12301 NETIF_F_GSO_IPIP | NETIF_F_GSO_SIT;
12278 dev->hw_enc_features = 12302 dev->hw_enc_features =
12279 NETIF_F_IP_CSUM | NETIF_F_IPV6_CSUM | NETIF_F_SG | 12303 NETIF_F_IP_CSUM | NETIF_F_IPV6_CSUM | NETIF_F_SG |
12280 NETIF_F_TSO | NETIF_F_TSO_ECN | NETIF_F_TSO6 | 12304 NETIF_F_TSO | NETIF_F_TSO_ECN | NETIF_F_TSO6 |
12305 NETIF_F_GSO_IPIP |
12306 NETIF_F_GSO_SIT |
12281 NETIF_F_GSO_GRE | NETIF_F_GSO_UDP_TUNNEL; 12307 NETIF_F_GSO_GRE | NETIF_F_GSO_UDP_TUNNEL;
12282 } 12308 }
12283 12309
@@ -12310,34 +12336,11 @@ err_out_release:
12310 12336
12311err_out_disable: 12337err_out_disable:
12312 pci_disable_device(pdev); 12338 pci_disable_device(pdev);
12313 pci_set_drvdata(pdev, NULL);
12314 12339
12315err_out: 12340err_out:
12316 return rc; 12341 return rc;
12317} 12342}
12318 12343
12319static void bnx2x_get_pcie_width_speed(struct bnx2x *bp, int *width,
12320 enum bnx2x_pci_bus_speed *speed)
12321{
12322 u32 link_speed, val = 0;
12323
12324 pci_read_config_dword(bp->pdev, PCICFG_LINK_CONTROL, &val);
12325 *width = (val & PCICFG_LINK_WIDTH) >> PCICFG_LINK_WIDTH_SHIFT;
12326
12327 link_speed = (val & PCICFG_LINK_SPEED) >> PCICFG_LINK_SPEED_SHIFT;
12328
12329 switch (link_speed) {
12330 case 3:
12331 *speed = BNX2X_PCI_LINK_SPEED_8000;
12332 break;
12333 case 2:
12334 *speed = BNX2X_PCI_LINK_SPEED_5000;
12335 break;
12336 default:
12337 *speed = BNX2X_PCI_LINK_SPEED_2500;
12338 }
12339}
12340
12341static int bnx2x_check_firmware(struct bnx2x *bp) 12344static int bnx2x_check_firmware(struct bnx2x *bp)
12342{ 12345{
12343 const struct firmware *firmware = bp->firmware; 12346 const struct firmware *firmware = bp->firmware;
@@ -12694,8 +12697,8 @@ static int bnx2x_init_one(struct pci_dev *pdev,
12694{ 12697{
12695 struct net_device *dev = NULL; 12698 struct net_device *dev = NULL;
12696 struct bnx2x *bp; 12699 struct bnx2x *bp;
12697 int pcie_width; 12700 enum pcie_link_width pcie_width;
12698 enum bnx2x_pci_bus_speed pcie_speed; 12701 enum pci_bus_speed pcie_speed;
12699 int rc, max_non_def_sbs; 12702 int rc, max_non_def_sbs;
12700 int rx_count, tx_count, rss_count, doorbell_size; 12703 int rx_count, tx_count, rss_count, doorbell_size;
12701 int max_cos_est; 12704 int max_cos_est;
@@ -12844,18 +12847,19 @@ static int bnx2x_init_one(struct pci_dev *pdev,
12844 dev_addr_add(bp->dev, bp->fip_mac, NETDEV_HW_ADDR_T_SAN); 12847 dev_addr_add(bp->dev, bp->fip_mac, NETDEV_HW_ADDR_T_SAN);
12845 rtnl_unlock(); 12848 rtnl_unlock();
12846 } 12849 }
12847 12850 if (pcie_get_minimum_link(bp->pdev, &pcie_speed, &pcie_width) ||
12848 bnx2x_get_pcie_width_speed(bp, &pcie_width, &pcie_speed); 12851 pcie_speed == PCI_SPEED_UNKNOWN ||
12849 BNX2X_DEV_INFO("got pcie width %d and speed %d\n", 12852 pcie_width == PCIE_LNK_WIDTH_UNKNOWN)
12850 pcie_width, pcie_speed); 12853 BNX2X_DEV_INFO("Failed to determine PCI Express Bandwidth\n");
12851 12854 else
12852 BNX2X_DEV_INFO("%s (%c%d) PCI-E x%d %s found at mem %lx, IRQ %d, node addr %pM\n", 12855 BNX2X_DEV_INFO(
12856 "%s (%c%d) PCI-E x%d %s found at mem %lx, IRQ %d, node addr %pM\n",
12853 board_info[ent->driver_data].name, 12857 board_info[ent->driver_data].name,
12854 (CHIP_REV(bp) >> 12) + 'A', (CHIP_METAL(bp) >> 4), 12858 (CHIP_REV(bp) >> 12) + 'A', (CHIP_METAL(bp) >> 4),
12855 pcie_width, 12859 pcie_width,
12856 pcie_speed == BNX2X_PCI_LINK_SPEED_2500 ? "2.5GHz" : 12860 pcie_speed == PCIE_SPEED_2_5GT ? "2.5GHz" :
12857 pcie_speed == BNX2X_PCI_LINK_SPEED_5000 ? "5.0GHz" : 12861 pcie_speed == PCIE_SPEED_5_0GT ? "5.0GHz" :
12858 pcie_speed == BNX2X_PCI_LINK_SPEED_8000 ? "8.0GHz" : 12862 pcie_speed == PCIE_SPEED_8_0GT ? "8.0GHz" :
12859 "Unknown", 12863 "Unknown",
12860 dev->base_addr, bp->pdev->irq, dev->dev_addr); 12864 dev->base_addr, bp->pdev->irq, dev->dev_addr);
12861 12865
@@ -12874,7 +12878,6 @@ init_one_exit:
12874 pci_release_regions(pdev); 12878 pci_release_regions(pdev);
12875 12879
12876 pci_disable_device(pdev); 12880 pci_disable_device(pdev);
12877 pci_set_drvdata(pdev, NULL);
12878 12881
12879 return rc; 12882 return rc;
12880} 12883}
@@ -12957,7 +12960,6 @@ static void __bnx2x_remove(struct pci_dev *pdev,
12957 pci_release_regions(pdev); 12960 pci_release_regions(pdev);
12958 12961
12959 pci_disable_device(pdev); 12962 pci_disable_device(pdev);
12960 pci_set_drvdata(pdev, NULL);
12961} 12963}
12962 12964
12963static void bnx2x_remove_one(struct pci_dev *pdev) 12965static void bnx2x_remove_one(struct pci_dev *pdev)
diff --git a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c
index 5e07efb6ec13..0216d592d0ce 100644
--- a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c
+++ b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c
@@ -2816,7 +2816,7 @@ struct set_vf_state_cookie {
2816 u8 state; 2816 u8 state;
2817}; 2817};
2818 2818
2819void bnx2x_set_vf_state(void *cookie) 2819static void bnx2x_set_vf_state(void *cookie)
2820{ 2820{
2821 struct set_vf_state_cookie *p = (struct set_vf_state_cookie *)cookie; 2821 struct set_vf_state_cookie *p = (struct set_vf_state_cookie *)cookie;
2822 2822
@@ -3239,8 +3239,9 @@ void bnx2x_disable_sriov(struct bnx2x *bp)
3239 pci_disable_sriov(bp->pdev); 3239 pci_disable_sriov(bp->pdev);
3240} 3240}
3241 3241
3242int bnx2x_vf_ndo_prep(struct bnx2x *bp, int vfidx, struct bnx2x_virtf **vf, 3242static int bnx2x_vf_ndo_prep(struct bnx2x *bp, int vfidx,
3243 struct pf_vf_bulletin_content **bulletin) 3243 struct bnx2x_virtf **vf,
3244 struct pf_vf_bulletin_content **bulletin)
3244{ 3245{
3245 if (bp->state != BNX2X_STATE_OPEN) { 3246 if (bp->state != BNX2X_STATE_OPEN) {
3246 BNX2X_ERR("vf ndo called though PF is down\n"); 3247 BNX2X_ERR("vf ndo called though PF is down\n");
@@ -3656,29 +3657,6 @@ alloc_mem_err:
3656 return -ENOMEM; 3657 return -ENOMEM;
3657} 3658}
3658 3659
3659int bnx2x_open_epilog(struct bnx2x *bp)
3660{
3661 /* Enable sriov via delayed work. This must be done via delayed work
3662 * because it causes the probe of the vf devices to be run, which invoke
3663 * register_netdevice which must have rtnl lock taken. As we are holding
3664 * the lock right now, that could only work if the probe would not take
3665 * the lock. However, as the probe of the vf may be called from other
3666 * contexts as well (such as passthrough to vm fails) it can't assume
3667 * the lock is being held for it. Using delayed work here allows the
3668 * probe code to simply take the lock (i.e. wait for it to be released
3669 * if it is being held). We only want to do this if the number of VFs
3670 * was set before PF driver was loaded.
3671 */
3672 if (IS_SRIOV(bp) && BNX2X_NR_VIRTFN(bp)) {
3673 smp_mb__before_clear_bit();
3674 set_bit(BNX2X_SP_RTNL_ENABLE_SRIOV, &bp->sp_rtnl_state);
3675 smp_mb__after_clear_bit();
3676 schedule_delayed_work(&bp->sp_rtnl_task, 0);
3677 }
3678
3679 return 0;
3680}
3681
3682void bnx2x_iov_channel_down(struct bnx2x *bp) 3660void bnx2x_iov_channel_down(struct bnx2x *bp)
3683{ 3661{
3684 int vf_idx; 3662 int vf_idx;
diff --git a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h
index 059f0d460af2..1ff6a9366629 100644
--- a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h
+++ b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h
@@ -782,7 +782,6 @@ static inline int bnx2x_vf_headroom(struct bnx2x *bp)
782void bnx2x_pf_set_vfs_vlan(struct bnx2x *bp); 782void bnx2x_pf_set_vfs_vlan(struct bnx2x *bp);
783int bnx2x_sriov_configure(struct pci_dev *dev, int num_vfs); 783int bnx2x_sriov_configure(struct pci_dev *dev, int num_vfs);
784void bnx2x_iov_channel_down(struct bnx2x *bp); 784void bnx2x_iov_channel_down(struct bnx2x *bp);
785int bnx2x_open_epilog(struct bnx2x *bp);
786 785
787#else /* CONFIG_BNX2X_SRIOV */ 786#else /* CONFIG_BNX2X_SRIOV */
788 787
@@ -842,7 +841,6 @@ static inline int bnx2x_vf_pci_alloc(struct bnx2x *bp) {return 0; }
842static inline void bnx2x_pf_set_vfs_vlan(struct bnx2x *bp) {} 841static inline void bnx2x_pf_set_vfs_vlan(struct bnx2x *bp) {}
843static inline int bnx2x_sriov_configure(struct pci_dev *dev, int num_vfs) {return 0; } 842static inline int bnx2x_sriov_configure(struct pci_dev *dev, int num_vfs) {return 0; }
844static inline void bnx2x_iov_channel_down(struct bnx2x *bp) {} 843static inline void bnx2x_iov_channel_down(struct bnx2x *bp) {}
845static inline int bnx2x_open_epilog(struct bnx2x *bp) {return 0; }
846 844
847#endif /* CONFIG_BNX2X_SRIOV */ 845#endif /* CONFIG_BNX2X_SRIOV */
848#endif /* bnx2x_sriov.h */ 846#endif /* bnx2x_sriov.h */
diff --git a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c
index 28757dfacf0d..9199adf32d33 100644
--- a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c
+++ b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c
@@ -60,6 +60,30 @@ void bnx2x_vfpf_finalize(struct bnx2x *bp, struct vfpf_first_tlv *first_tlv)
60 mutex_unlock(&bp->vf2pf_mutex); 60 mutex_unlock(&bp->vf2pf_mutex);
61} 61}
62 62
63/* Finds a TLV by type in a TLV buffer; If found, returns pointer to the TLV */
64static void *bnx2x_search_tlv_list(struct bnx2x *bp, void *tlvs_list,
65 enum channel_tlvs req_tlv)
66{
67 struct channel_tlv *tlv = (struct channel_tlv *)tlvs_list;
68
69 do {
70 if (tlv->type == req_tlv)
71 return tlv;
72
73 if (!tlv->length) {
74 BNX2X_ERR("Found TLV with length 0\n");
75 return NULL;
76 }
77
78 tlvs_list += tlv->length;
79 tlv = (struct channel_tlv *)tlvs_list;
80 } while (tlv->type != CHANNEL_TLV_LIST_END);
81
82 DP(BNX2X_MSG_IOV, "TLV list does not contain %d TLV\n", req_tlv);
83
84 return NULL;
85}
86
63/* list the types and lengths of the tlvs on the buffer */ 87/* list the types and lengths of the tlvs on the buffer */
64void bnx2x_dp_tlv_list(struct bnx2x *bp, void *tlvs_list) 88void bnx2x_dp_tlv_list(struct bnx2x *bp, void *tlvs_list)
65{ 89{
@@ -196,6 +220,7 @@ int bnx2x_vfpf_acquire(struct bnx2x *bp, u8 tx_count, u8 rx_count)
196 int rc = 0, attempts = 0; 220 int rc = 0, attempts = 0;
197 struct vfpf_acquire_tlv *req = &bp->vf2pf_mbox->req.acquire; 221 struct vfpf_acquire_tlv *req = &bp->vf2pf_mbox->req.acquire;
198 struct pfvf_acquire_resp_tlv *resp = &bp->vf2pf_mbox->resp.acquire_resp; 222 struct pfvf_acquire_resp_tlv *resp = &bp->vf2pf_mbox->resp.acquire_resp;
223 struct vfpf_port_phys_id_resp_tlv *phys_port_resp;
199 u32 vf_id; 224 u32 vf_id;
200 bool resources_acquired = false; 225 bool resources_acquired = false;
201 226
@@ -219,8 +244,14 @@ int bnx2x_vfpf_acquire(struct bnx2x *bp, u8 tx_count, u8 rx_count)
219 /* pf 2 vf bulletin board address */ 244 /* pf 2 vf bulletin board address */
220 req->bulletin_addr = bp->pf2vf_bulletin_mapping; 245 req->bulletin_addr = bp->pf2vf_bulletin_mapping;
221 246
247 /* Request physical port identifier */
248 bnx2x_add_tlv(bp, req, req->first_tlv.tl.length,
249 CHANNEL_TLV_PHYS_PORT_ID, sizeof(struct channel_tlv));
250
222 /* add list termination tlv */ 251 /* add list termination tlv */
223 bnx2x_add_tlv(bp, req, req->first_tlv.tl.length, CHANNEL_TLV_LIST_END, 252 bnx2x_add_tlv(bp, req,
253 req->first_tlv.tl.length + sizeof(struct channel_tlv),
254 CHANNEL_TLV_LIST_END,
224 sizeof(struct channel_list_end_tlv)); 255 sizeof(struct channel_list_end_tlv));
225 256
226 /* output tlvs list */ 257 /* output tlvs list */
@@ -287,6 +318,15 @@ int bnx2x_vfpf_acquire(struct bnx2x *bp, u8 tx_count, u8 rx_count)
287 } 318 }
288 } 319 }
289 320
321 /* Retrieve physical port id (if possible) */
322 phys_port_resp = (struct vfpf_port_phys_id_resp_tlv *)
323 bnx2x_search_tlv_list(bp, resp,
324 CHANNEL_TLV_PHYS_PORT_ID);
325 if (phys_port_resp) {
326 memcpy(bp->phys_port_id, phys_port_resp->id, ETH_ALEN);
327 bp->flags |= HAS_PHYS_PORT_ID;
328 }
329
290 /* get HW info */ 330 /* get HW info */
291 bp->common.chip_id |= (bp->acquire_resp.pfdev_info.chip_num & 0xffff); 331 bp->common.chip_id |= (bp->acquire_resp.pfdev_info.chip_num & 0xffff);
292 bp->link_params.chip_id = bp->common.chip_id; 332 bp->link_params.chip_id = bp->common.chip_id;
@@ -983,53 +1023,59 @@ static int bnx2x_copy32_vf_dmae(struct bnx2x *bp, u8 from_vf,
983 return bnx2x_issue_dmae_with_comp(bp, &dmae, bnx2x_sp(bp, wb_comp)); 1023 return bnx2x_issue_dmae_with_comp(bp, &dmae, bnx2x_sp(bp, wb_comp));
984} 1024}
985 1025
986static void bnx2x_vf_mbx_resp(struct bnx2x *bp, struct bnx2x_virtf *vf) 1026static void bnx2x_vf_mbx_resp_single_tlv(struct bnx2x *bp,
1027 struct bnx2x_virtf *vf)
987{ 1028{
988 struct bnx2x_vf_mbx *mbx = BP_VF_MBX(bp, vf->index); 1029 struct bnx2x_vf_mbx *mbx = BP_VF_MBX(bp, vf->index);
989 u64 vf_addr;
990 dma_addr_t pf_addr;
991 u16 length, type; 1030 u16 length, type;
992 int rc;
993 struct pfvf_general_resp_tlv *resp = &mbx->msg->resp.general_resp;
994 1031
995 /* prepare response */ 1032 /* prepare response */
996 type = mbx->first_tlv.tl.type; 1033 type = mbx->first_tlv.tl.type;
997 length = type == CHANNEL_TLV_ACQUIRE ? 1034 length = type == CHANNEL_TLV_ACQUIRE ?
998 sizeof(struct pfvf_acquire_resp_tlv) : 1035 sizeof(struct pfvf_acquire_resp_tlv) :
999 sizeof(struct pfvf_general_resp_tlv); 1036 sizeof(struct pfvf_general_resp_tlv);
1000 bnx2x_add_tlv(bp, resp, 0, type, length); 1037 bnx2x_add_tlv(bp, &mbx->msg->resp, 0, type, length);
1001 resp->hdr.status = bnx2x_pfvf_status_codes(vf->op_rc); 1038 bnx2x_add_tlv(bp, &mbx->msg->resp, length, CHANNEL_TLV_LIST_END,
1002 bnx2x_add_tlv(bp, resp, length, CHANNEL_TLV_LIST_END,
1003 sizeof(struct channel_list_end_tlv)); 1039 sizeof(struct channel_list_end_tlv));
1040}
1041
1042static void bnx2x_vf_mbx_resp_send_msg(struct bnx2x *bp,
1043 struct bnx2x_virtf *vf)
1044{
1045 struct bnx2x_vf_mbx *mbx = BP_VF_MBX(bp, vf->index);
1046 struct pfvf_general_resp_tlv *resp = &mbx->msg->resp.general_resp;
1047 dma_addr_t pf_addr;
1048 u64 vf_addr;
1049 int rc;
1050
1004 bnx2x_dp_tlv_list(bp, resp); 1051 bnx2x_dp_tlv_list(bp, resp);
1005 DP(BNX2X_MSG_IOV, "mailbox vf address hi 0x%x, lo 0x%x, offset 0x%x\n", 1052 DP(BNX2X_MSG_IOV, "mailbox vf address hi 0x%x, lo 0x%x, offset 0x%x\n",
1006 mbx->vf_addr_hi, mbx->vf_addr_lo, mbx->first_tlv.resp_msg_offset); 1053 mbx->vf_addr_hi, mbx->vf_addr_lo, mbx->first_tlv.resp_msg_offset);
1007 1054
1055 resp->hdr.status = bnx2x_pfvf_status_codes(vf->op_rc);
1056
1008 /* send response */ 1057 /* send response */
1009 vf_addr = HILO_U64(mbx->vf_addr_hi, mbx->vf_addr_lo) + 1058 vf_addr = HILO_U64(mbx->vf_addr_hi, mbx->vf_addr_lo) +
1010 mbx->first_tlv.resp_msg_offset; 1059 mbx->first_tlv.resp_msg_offset;
1011 pf_addr = mbx->msg_mapping + 1060 pf_addr = mbx->msg_mapping +
1012 offsetof(struct bnx2x_vf_mbx_msg, resp); 1061 offsetof(struct bnx2x_vf_mbx_msg, resp);
1013 1062
1014 /* copy the response body, if there is one, before the header, as the vf 1063 /* Copy the response buffer. The first u64 is written afterwards, as
1015 * is sensitive to the header being written 1064 * the vf is sensitive to the header being written
1016 */ 1065 */
1017 if (resp->hdr.tl.length > sizeof(u64)) { 1066 vf_addr += sizeof(u64);
1018 length = resp->hdr.tl.length - sizeof(u64); 1067 pf_addr += sizeof(u64);
1019 vf_addr += sizeof(u64); 1068 rc = bnx2x_copy32_vf_dmae(bp, false, pf_addr, vf->abs_vfid,
1020 pf_addr += sizeof(u64); 1069 U64_HI(vf_addr),
1021 rc = bnx2x_copy32_vf_dmae(bp, false, pf_addr, vf->abs_vfid, 1070 U64_LO(vf_addr),
1022 U64_HI(vf_addr), 1071 (sizeof(union pfvf_tlvs) - sizeof(u64))/4);
1023 U64_LO(vf_addr), 1072 if (rc) {
1024 length/4); 1073 BNX2X_ERR("Failed to copy response body to VF %d\n",
1025 if (rc) { 1074 vf->abs_vfid);
1026 BNX2X_ERR("Failed to copy response body to VF %d\n", 1075 goto mbx_error;
1027 vf->abs_vfid);
1028 goto mbx_error;
1029 }
1030 vf_addr -= sizeof(u64);
1031 pf_addr -= sizeof(u64);
1032 } 1076 }
1077 vf_addr -= sizeof(u64);
1078 pf_addr -= sizeof(u64);
1033 1079
1034 /* ack the FW */ 1080 /* ack the FW */
1035 storm_memset_vf_mbx_ack(bp, vf->abs_vfid); 1081 storm_memset_vf_mbx_ack(bp, vf->abs_vfid);
@@ -1060,6 +1106,36 @@ mbx_error:
1060 bnx2x_vf_release(bp, vf, false); /* non blocking */ 1106 bnx2x_vf_release(bp, vf, false); /* non blocking */
1061} 1107}
1062 1108
1109static void bnx2x_vf_mbx_resp(struct bnx2x *bp,
1110 struct bnx2x_virtf *vf)
1111{
1112 bnx2x_vf_mbx_resp_single_tlv(bp, vf);
1113 bnx2x_vf_mbx_resp_send_msg(bp, vf);
1114}
1115
1116static void bnx2x_vf_mbx_resp_phys_port(struct bnx2x *bp,
1117 struct bnx2x_virtf *vf,
1118 void *buffer,
1119 u16 *offset)
1120{
1121 struct vfpf_port_phys_id_resp_tlv *port_id;
1122
1123 if (!(bp->flags & HAS_PHYS_PORT_ID))
1124 return;
1125
1126 bnx2x_add_tlv(bp, buffer, *offset, CHANNEL_TLV_PHYS_PORT_ID,
1127 sizeof(struct vfpf_port_phys_id_resp_tlv));
1128
1129 port_id = (struct vfpf_port_phys_id_resp_tlv *)
1130 (((u8 *)buffer) + *offset);
1131 memcpy(port_id->id, bp->phys_port_id, ETH_ALEN);
1132
1133 /* Offset should continue representing the offset to the tail
1134 * of TLV data (outside this function scope)
1135 */
1136 *offset += sizeof(struct vfpf_port_phys_id_resp_tlv);
1137}
1138
1063static void bnx2x_vf_mbx_acquire_resp(struct bnx2x *bp, struct bnx2x_virtf *vf, 1139static void bnx2x_vf_mbx_acquire_resp(struct bnx2x *bp, struct bnx2x_virtf *vf,
1064 struct bnx2x_vf_mbx *mbx, int vfop_status) 1140 struct bnx2x_vf_mbx *mbx, int vfop_status)
1065{ 1141{
@@ -1067,6 +1143,7 @@ static void bnx2x_vf_mbx_acquire_resp(struct bnx2x *bp, struct bnx2x_virtf *vf,
1067 struct pfvf_acquire_resp_tlv *resp = &mbx->msg->resp.acquire_resp; 1143 struct pfvf_acquire_resp_tlv *resp = &mbx->msg->resp.acquire_resp;
1068 struct pf_vf_resc *resc = &resp->resc; 1144 struct pf_vf_resc *resc = &resp->resc;
1069 u8 status = bnx2x_pfvf_status_codes(vfop_status); 1145 u8 status = bnx2x_pfvf_status_codes(vfop_status);
1146 u16 length;
1070 1147
1071 memset(resp, 0, sizeof(*resp)); 1148 memset(resp, 0, sizeof(*resp));
1072 1149
@@ -1140,9 +1217,24 @@ static void bnx2x_vf_mbx_acquire_resp(struct bnx2x *bp, struct bnx2x_virtf *vf,
1140 resc->hw_sbs[i].sb_qid); 1217 resc->hw_sbs[i].sb_qid);
1141 DP_CONT(BNX2X_MSG_IOV, "]\n"); 1218 DP_CONT(BNX2X_MSG_IOV, "]\n");
1142 1219
1220 /* prepare response */
1221 length = sizeof(struct pfvf_acquire_resp_tlv);
1222 bnx2x_add_tlv(bp, &mbx->msg->resp, 0, CHANNEL_TLV_ACQUIRE, length);
1223
1224 /* Handle possible VF requests for physical port identifiers.
1225 * 'length' should continue to indicate the offset of the first empty
1226 * place in the buffer (i.e., where next TLV should be inserted)
1227 */
1228 if (bnx2x_search_tlv_list(bp, &mbx->msg->req,
1229 CHANNEL_TLV_PHYS_PORT_ID))
1230 bnx2x_vf_mbx_resp_phys_port(bp, vf, &mbx->msg->resp, &length);
1231
1232 bnx2x_add_tlv(bp, &mbx->msg->resp, length, CHANNEL_TLV_LIST_END,
1233 sizeof(struct channel_list_end_tlv));
1234
1143 /* send the response */ 1235 /* send the response */
1144 vf->op_rc = vfop_status; 1236 vf->op_rc = vfop_status;
1145 bnx2x_vf_mbx_resp(bp, vf); 1237 bnx2x_vf_mbx_resp_send_msg(bp, vf);
1146} 1238}
1147 1239
1148static void bnx2x_vf_mbx_acquire(struct bnx2x *bp, struct bnx2x_virtf *vf, 1240static void bnx2x_vf_mbx_acquire(struct bnx2x *bp, struct bnx2x_virtf *vf,
@@ -1874,6 +1966,9 @@ void bnx2x_vf_mbx(struct bnx2x *bp, struct vf_pf_event_data *vfpf_event)
1874 /* process the VF message header */ 1966 /* process the VF message header */
1875 mbx->first_tlv = mbx->msg->req.first_tlv; 1967 mbx->first_tlv = mbx->msg->req.first_tlv;
1876 1968
1969 /* Clean response buffer to refrain from falsely seeing chains */
1970 memset(&mbx->msg->resp, 0, sizeof(union pfvf_tlvs));
1971
1877 /* dispatch the request (will prepare the response) */ 1972 /* dispatch the request (will prepare the response) */
1878 bnx2x_vf_mbx_request(bp, vf, mbx); 1973 bnx2x_vf_mbx_request(bp, vf, mbx);
1879 goto mbx_done; 1974 goto mbx_done;
diff --git a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.h b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.h
index 1179fe06d0c7..208568bc7a71 100644
--- a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.h
+++ b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.h
@@ -188,6 +188,12 @@ struct pfvf_acquire_resp_tlv {
188 } resc; 188 } resc;
189}; 189};
190 190
191struct vfpf_port_phys_id_resp_tlv {
192 struct channel_tlv tl;
193 u8 id[ETH_ALEN];
194 u8 padding[2];
195};
196
191#define VFPF_INIT_FLG_STATS_COALESCE (1 << 0) /* when set the VFs queues 197#define VFPF_INIT_FLG_STATS_COALESCE (1 << 0) /* when set the VFs queues
192 * stats will be coalesced on 198 * stats will be coalesced on
193 * the leading RSS queue 199 * the leading RSS queue
@@ -398,6 +404,7 @@ enum channel_tlvs {
398 CHANNEL_TLV_PF_SET_MAC, 404 CHANNEL_TLV_PF_SET_MAC,
399 CHANNEL_TLV_PF_SET_VLAN, 405 CHANNEL_TLV_PF_SET_VLAN,
400 CHANNEL_TLV_UPDATE_RSS, 406 CHANNEL_TLV_UPDATE_RSS,
407 CHANNEL_TLV_PHYS_PORT_ID,
401 CHANNEL_TLV_MAX 408 CHANNEL_TLV_MAX
402}; 409};
403 410
diff --git a/drivers/net/ethernet/broadcom/cnic.c b/drivers/net/ethernet/broadcom/cnic.c
index 99394bd49a13..f58a8b80302d 100644
--- a/drivers/net/ethernet/broadcom/cnic.c
+++ b/drivers/net/ethernet/broadcom/cnic.c
@@ -393,7 +393,7 @@ static int cnic_iscsi_nl_msg_recv(struct cnic_dev *dev, u32 msg_type,
393 393
394 csk->vlan_id = path_resp->vlan_id; 394 csk->vlan_id = path_resp->vlan_id;
395 395
396 memcpy(csk->ha, path_resp->mac_addr, 6); 396 memcpy(csk->ha, path_resp->mac_addr, ETH_ALEN);
397 if (test_bit(SK_F_IPV6, &csk->flags)) 397 if (test_bit(SK_F_IPV6, &csk->flags))
398 memcpy(&csk->src_ip[0], &path_resp->src.v6_addr, 398 memcpy(&csk->src_ip[0], &path_resp->src.v6_addr,
399 sizeof(struct in6_addr)); 399 sizeof(struct in6_addr));
@@ -5572,7 +5572,7 @@ static struct cnic_dev *init_bnx2x_cnic(struct net_device *dev)
5572 if (cdev->max_fcoe_conn > BNX2X_FCOE_NUM_CONNECTIONS) 5572 if (cdev->max_fcoe_conn > BNX2X_FCOE_NUM_CONNECTIONS)
5573 cdev->max_fcoe_conn = BNX2X_FCOE_NUM_CONNECTIONS; 5573 cdev->max_fcoe_conn = BNX2X_FCOE_NUM_CONNECTIONS;
5574 5574
5575 memcpy(cdev->mac_addr, ethdev->iscsi_mac, 6); 5575 memcpy(cdev->mac_addr, ethdev->iscsi_mac, ETH_ALEN);
5576 5576
5577 cp->cnic_ops = &cnic_bnx2x_ops; 5577 cp->cnic_ops = &cnic_bnx2x_ops;
5578 cp->start_hw = cnic_start_bnx2x_hw; 5578 cp->start_hw = cnic_start_bnx2x_hw;
diff --git a/drivers/net/ethernet/broadcom/cnic_if.h b/drivers/net/ethernet/broadcom/cnic_if.h
index 0658b43e148c..ebbfe25acaa6 100644
--- a/drivers/net/ethernet/broadcom/cnic_if.h
+++ b/drivers/net/ethernet/broadcom/cnic_if.h
@@ -353,8 +353,8 @@ struct cnic_ulp_ops {
353 atomic_t ref_count; 353 atomic_t ref_count;
354}; 354};
355 355
356extern int cnic_register_driver(int ulp_type, struct cnic_ulp_ops *ulp_ops); 356int cnic_register_driver(int ulp_type, struct cnic_ulp_ops *ulp_ops);
357 357
358extern int cnic_unregister_driver(int ulp_type); 358int cnic_unregister_driver(int ulp_type);
359 359
360#endif 360#endif
diff --git a/drivers/net/ethernet/broadcom/tg3.c b/drivers/net/ethernet/broadcom/tg3.c
index 12d961c4ebca..00c5be8c55b8 100644
--- a/drivers/net/ethernet/broadcom/tg3.c
+++ b/drivers/net/ethernet/broadcom/tg3.c
@@ -94,10 +94,10 @@ static inline void _tg3_flag_clear(enum TG3_FLAGS flag, unsigned long *bits)
94 94
95#define DRV_MODULE_NAME "tg3" 95#define DRV_MODULE_NAME "tg3"
96#define TG3_MAJ_NUM 3 96#define TG3_MAJ_NUM 3
97#define TG3_MIN_NUM 133 97#define TG3_MIN_NUM 134
98#define DRV_MODULE_VERSION \ 98#define DRV_MODULE_VERSION \
99 __stringify(TG3_MAJ_NUM) "." __stringify(TG3_MIN_NUM) 99 __stringify(TG3_MAJ_NUM) "." __stringify(TG3_MIN_NUM)
100#define DRV_MODULE_RELDATE "Jul 29, 2013" 100#define DRV_MODULE_RELDATE "Sep 16, 2013"
101 101
102#define RESET_KIND_SHUTDOWN 0 102#define RESET_KIND_SHUTDOWN 0
103#define RESET_KIND_INIT 1 103#define RESET_KIND_INIT 1
@@ -337,6 +337,11 @@ static DEFINE_PCI_DEVICE_TABLE(tg3_pci_tbl) = {
337 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, TG3PCI_DEVICE_TIGON3_5762)}, 337 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, TG3PCI_DEVICE_TIGON3_5762)},
338 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, TG3PCI_DEVICE_TIGON3_5725)}, 338 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, TG3PCI_DEVICE_TIGON3_5725)},
339 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, TG3PCI_DEVICE_TIGON3_5727)}, 339 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, TG3PCI_DEVICE_TIGON3_5727)},
340 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, TG3PCI_DEVICE_TIGON3_57764)},
341 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, TG3PCI_DEVICE_TIGON3_57767)},
342 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, TG3PCI_DEVICE_TIGON3_57787)},
343 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, TG3PCI_DEVICE_TIGON3_57782)},
344 {PCI_DEVICE(PCI_VENDOR_ID_BROADCOM, TG3PCI_DEVICE_TIGON3_57786)},
340 {PCI_DEVICE(PCI_VENDOR_ID_SYSKONNECT, PCI_DEVICE_ID_SYSKONNECT_9DXX)}, 345 {PCI_DEVICE(PCI_VENDOR_ID_SYSKONNECT, PCI_DEVICE_ID_SYSKONNECT_9DXX)},
341 {PCI_DEVICE(PCI_VENDOR_ID_SYSKONNECT, PCI_DEVICE_ID_SYSKONNECT_9MXX)}, 346 {PCI_DEVICE(PCI_VENDOR_ID_SYSKONNECT, PCI_DEVICE_ID_SYSKONNECT_9MXX)},
342 {PCI_DEVICE(PCI_VENDOR_ID_ALTIMA, PCI_DEVICE_ID_ALTIMA_AC1000)}, 347 {PCI_DEVICE(PCI_VENDOR_ID_ALTIMA, PCI_DEVICE_ID_ALTIMA_AC1000)},
@@ -1326,6 +1331,12 @@ static int tg3_phy_toggle_auxctl_smdsp(struct tg3 *tp, bool enable)
1326 return err; 1331 return err;
1327} 1332}
1328 1333
1334static int tg3_phy_shdw_write(struct tg3 *tp, int reg, u32 val)
1335{
1336 return tg3_writephy(tp, MII_TG3_MISC_SHDW,
1337 reg | val | MII_TG3_MISC_SHDW_WREN);
1338}
1339
1329static int tg3_bmcr_reset(struct tg3 *tp) 1340static int tg3_bmcr_reset(struct tg3 *tp)
1330{ 1341{
1331 u32 phy_control; 1342 u32 phy_control;
@@ -1364,7 +1375,7 @@ static int tg3_mdio_read(struct mii_bus *bp, int mii_id, int reg)
1364 1375
1365 spin_lock_bh(&tp->lock); 1376 spin_lock_bh(&tp->lock);
1366 1377
1367 if (tg3_readphy(tp, reg, &val)) 1378 if (__tg3_readphy(tp, mii_id, reg, &val))
1368 val = -EIO; 1379 val = -EIO;
1369 1380
1370 spin_unlock_bh(&tp->lock); 1381 spin_unlock_bh(&tp->lock);
@@ -1379,7 +1390,7 @@ static int tg3_mdio_write(struct mii_bus *bp, int mii_id, int reg, u16 val)
1379 1390
1380 spin_lock_bh(&tp->lock); 1391 spin_lock_bh(&tp->lock);
1381 1392
1382 if (tg3_writephy(tp, reg, val)) 1393 if (__tg3_writephy(tp, mii_id, reg, val))
1383 ret = -EIO; 1394 ret = -EIO;
1384 1395
1385 spin_unlock_bh(&tp->lock); 1396 spin_unlock_bh(&tp->lock);
@@ -1397,7 +1408,7 @@ static void tg3_mdio_config_5785(struct tg3 *tp)
1397 u32 val; 1408 u32 val;
1398 struct phy_device *phydev; 1409 struct phy_device *phydev;
1399 1410
1400 phydev = tp->mdio_bus->phy_map[TG3_PHY_MII_ADDR]; 1411 phydev = tp->mdio_bus->phy_map[tp->phy_addr];
1401 switch (phydev->drv->phy_id & phydev->drv->phy_id_mask) { 1412 switch (phydev->drv->phy_id & phydev->drv->phy_id_mask) {
1402 case PHY_ID_BCM50610: 1413 case PHY_ID_BCM50610:
1403 case PHY_ID_BCM50610M: 1414 case PHY_ID_BCM50610M:
@@ -1502,6 +1513,13 @@ static int tg3_mdio_init(struct tg3 *tp)
1502 TG3_CPMU_PHY_STRAP_IS_SERDES; 1513 TG3_CPMU_PHY_STRAP_IS_SERDES;
1503 if (is_serdes) 1514 if (is_serdes)
1504 tp->phy_addr += 7; 1515 tp->phy_addr += 7;
1516 } else if (tg3_flag(tp, IS_SSB_CORE) && tg3_flag(tp, ROBOSWITCH)) {
1517 int addr;
1518
1519 addr = ssb_gige_get_phyaddr(tp->pdev);
1520 if (addr < 0)
1521 return addr;
1522 tp->phy_addr = addr;
1505 } else 1523 } else
1506 tp->phy_addr = TG3_PHY_MII_ADDR; 1524 tp->phy_addr = TG3_PHY_MII_ADDR;
1507 1525
@@ -1522,7 +1540,7 @@ static int tg3_mdio_init(struct tg3 *tp)
1522 tp->mdio_bus->read = &tg3_mdio_read; 1540 tp->mdio_bus->read = &tg3_mdio_read;
1523 tp->mdio_bus->write = &tg3_mdio_write; 1541 tp->mdio_bus->write = &tg3_mdio_write;
1524 tp->mdio_bus->reset = &tg3_mdio_reset; 1542 tp->mdio_bus->reset = &tg3_mdio_reset;
1525 tp->mdio_bus->phy_mask = ~(1 << TG3_PHY_MII_ADDR); 1543 tp->mdio_bus->phy_mask = ~(1 << tp->phy_addr);
1526 tp->mdio_bus->irq = &tp->mdio_irq[0]; 1544 tp->mdio_bus->irq = &tp->mdio_irq[0];
1527 1545
1528 for (i = 0; i < PHY_MAX_ADDR; i++) 1546 for (i = 0; i < PHY_MAX_ADDR; i++)
@@ -1543,7 +1561,7 @@ static int tg3_mdio_init(struct tg3 *tp)
1543 return i; 1561 return i;
1544 } 1562 }
1545 1563
1546 phydev = tp->mdio_bus->phy_map[TG3_PHY_MII_ADDR]; 1564 phydev = tp->mdio_bus->phy_map[tp->phy_addr];
1547 1565
1548 if (!phydev || !phydev->drv) { 1566 if (!phydev || !phydev->drv) {
1549 dev_warn(&tp->pdev->dev, "No PHY devices\n"); 1567 dev_warn(&tp->pdev->dev, "No PHY devices\n");
@@ -1953,7 +1971,7 @@ static void tg3_setup_flow_control(struct tg3 *tp, u32 lcladv, u32 rmtadv)
1953 u32 old_tx_mode = tp->tx_mode; 1971 u32 old_tx_mode = tp->tx_mode;
1954 1972
1955 if (tg3_flag(tp, USE_PHYLIB)) 1973 if (tg3_flag(tp, USE_PHYLIB))
1956 autoneg = tp->mdio_bus->phy_map[TG3_PHY_MII_ADDR]->autoneg; 1974 autoneg = tp->mdio_bus->phy_map[tp->phy_addr]->autoneg;
1957 else 1975 else
1958 autoneg = tp->link_config.autoneg; 1976 autoneg = tp->link_config.autoneg;
1959 1977
@@ -1989,7 +2007,7 @@ static void tg3_adjust_link(struct net_device *dev)
1989 u8 oldflowctrl, linkmesg = 0; 2007 u8 oldflowctrl, linkmesg = 0;
1990 u32 mac_mode, lcl_adv, rmt_adv; 2008 u32 mac_mode, lcl_adv, rmt_adv;
1991 struct tg3 *tp = netdev_priv(dev); 2009 struct tg3 *tp = netdev_priv(dev);
1992 struct phy_device *phydev = tp->mdio_bus->phy_map[TG3_PHY_MII_ADDR]; 2010 struct phy_device *phydev = tp->mdio_bus->phy_map[tp->phy_addr];
1993 2011
1994 spin_lock_bh(&tp->lock); 2012 spin_lock_bh(&tp->lock);
1995 2013
@@ -2078,7 +2096,7 @@ static int tg3_phy_init(struct tg3 *tp)
2078 /* Bring the PHY back to a known state. */ 2096 /* Bring the PHY back to a known state. */
2079 tg3_bmcr_reset(tp); 2097 tg3_bmcr_reset(tp);
2080 2098
2081 phydev = tp->mdio_bus->phy_map[TG3_PHY_MII_ADDR]; 2099 phydev = tp->mdio_bus->phy_map[tp->phy_addr];
2082 2100
2083 /* Attach the MAC to the PHY. */ 2101 /* Attach the MAC to the PHY. */
2084 phydev = phy_connect(tp->dev, dev_name(&phydev->dev), 2102 phydev = phy_connect(tp->dev, dev_name(&phydev->dev),
@@ -2105,7 +2123,7 @@ static int tg3_phy_init(struct tg3 *tp)
2105 SUPPORTED_Asym_Pause); 2123 SUPPORTED_Asym_Pause);
2106 break; 2124 break;
2107 default: 2125 default:
2108 phy_disconnect(tp->mdio_bus->phy_map[TG3_PHY_MII_ADDR]); 2126 phy_disconnect(tp->mdio_bus->phy_map[tp->phy_addr]);
2109 return -EINVAL; 2127 return -EINVAL;
2110 } 2128 }
2111 2129
@@ -2123,7 +2141,7 @@ static void tg3_phy_start(struct tg3 *tp)
2123 if (!(tp->phy_flags & TG3_PHYFLG_IS_CONNECTED)) 2141 if (!(tp->phy_flags & TG3_PHYFLG_IS_CONNECTED))
2124 return; 2142 return;
2125 2143
2126 phydev = tp->mdio_bus->phy_map[TG3_PHY_MII_ADDR]; 2144 phydev = tp->mdio_bus->phy_map[tp->phy_addr];
2127 2145
2128 if (tp->phy_flags & TG3_PHYFLG_IS_LOW_POWER) { 2146 if (tp->phy_flags & TG3_PHYFLG_IS_LOW_POWER) {
2129 tp->phy_flags &= ~TG3_PHYFLG_IS_LOW_POWER; 2147 tp->phy_flags &= ~TG3_PHYFLG_IS_LOW_POWER;
@@ -2143,13 +2161,13 @@ static void tg3_phy_stop(struct tg3 *tp)
2143 if (!(tp->phy_flags & TG3_PHYFLG_IS_CONNECTED)) 2161 if (!(tp->phy_flags & TG3_PHYFLG_IS_CONNECTED))
2144 return; 2162 return;
2145 2163
2146 phy_stop(tp->mdio_bus->phy_map[TG3_PHY_MII_ADDR]); 2164 phy_stop(tp->mdio_bus->phy_map[tp->phy_addr]);
2147} 2165}
2148 2166
2149static void tg3_phy_fini(struct tg3 *tp) 2167static void tg3_phy_fini(struct tg3 *tp)
2150{ 2168{
2151 if (tp->phy_flags & TG3_PHYFLG_IS_CONNECTED) { 2169 if (tp->phy_flags & TG3_PHYFLG_IS_CONNECTED) {
2152 phy_disconnect(tp->mdio_bus->phy_map[TG3_PHY_MII_ADDR]); 2170 phy_disconnect(tp->mdio_bus->phy_map[tp->phy_addr]);
2153 tp->phy_flags &= ~TG3_PHYFLG_IS_CONNECTED; 2171 tp->phy_flags &= ~TG3_PHYFLG_IS_CONNECTED;
2154 } 2172 }
2155} 2173}
@@ -2218,25 +2236,21 @@ static void tg3_phy_toggle_apd(struct tg3 *tp, bool enable)
2218 return; 2236 return;
2219 } 2237 }
2220 2238
2221 reg = MII_TG3_MISC_SHDW_WREN | 2239 reg = MII_TG3_MISC_SHDW_SCR5_LPED |
2222 MII_TG3_MISC_SHDW_SCR5_SEL |
2223 MII_TG3_MISC_SHDW_SCR5_LPED |
2224 MII_TG3_MISC_SHDW_SCR5_DLPTLM | 2240 MII_TG3_MISC_SHDW_SCR5_DLPTLM |
2225 MII_TG3_MISC_SHDW_SCR5_SDTL | 2241 MII_TG3_MISC_SHDW_SCR5_SDTL |
2226 MII_TG3_MISC_SHDW_SCR5_C125OE; 2242 MII_TG3_MISC_SHDW_SCR5_C125OE;
2227 if (tg3_asic_rev(tp) != ASIC_REV_5784 || !enable) 2243 if (tg3_asic_rev(tp) != ASIC_REV_5784 || !enable)
2228 reg |= MII_TG3_MISC_SHDW_SCR5_DLLAPD; 2244 reg |= MII_TG3_MISC_SHDW_SCR5_DLLAPD;
2229 2245
2230 tg3_writephy(tp, MII_TG3_MISC_SHDW, reg); 2246 tg3_phy_shdw_write(tp, MII_TG3_MISC_SHDW_SCR5_SEL, reg);
2231 2247
2232 2248
2233 reg = MII_TG3_MISC_SHDW_WREN | 2249 reg = MII_TG3_MISC_SHDW_APD_WKTM_84MS;
2234 MII_TG3_MISC_SHDW_APD_SEL |
2235 MII_TG3_MISC_SHDW_APD_WKTM_84MS;
2236 if (enable) 2250 if (enable)
2237 reg |= MII_TG3_MISC_SHDW_APD_ENABLE; 2251 reg |= MII_TG3_MISC_SHDW_APD_ENABLE;
2238 2252
2239 tg3_writephy(tp, MII_TG3_MISC_SHDW, reg); 2253 tg3_phy_shdw_write(tp, MII_TG3_MISC_SHDW_APD_SEL, reg);
2240} 2254}
2241 2255
2242static void tg3_phy_toggle_automdix(struct tg3 *tp, bool enable) 2256static void tg3_phy_toggle_automdix(struct tg3 *tp, bool enable)
@@ -4027,7 +4041,7 @@ static int tg3_power_down_prepare(struct tg3 *tp)
4027 struct phy_device *phydev; 4041 struct phy_device *phydev;
4028 u32 phyid, advertising; 4042 u32 phyid, advertising;
4029 4043
4030 phydev = tp->mdio_bus->phy_map[TG3_PHY_MII_ADDR]; 4044 phydev = tp->mdio_bus->phy_map[tp->phy_addr];
4031 4045
4032 tp->phy_flags |= TG3_PHYFLG_IS_LOW_POWER; 4046 tp->phy_flags |= TG3_PHYFLG_IS_LOW_POWER;
4033 4047
@@ -6848,12 +6862,6 @@ static int tg3_rx(struct tg3_napi *tnapi, int budget)
6848 pci_unmap_single(tp->pdev, dma_addr, skb_size, 6862 pci_unmap_single(tp->pdev, dma_addr, skb_size,
6849 PCI_DMA_FROMDEVICE); 6863 PCI_DMA_FROMDEVICE);
6850 6864
6851 skb = build_skb(data, frag_size);
6852 if (!skb) {
6853 tg3_frag_free(frag_size != 0, data);
6854 goto drop_it_no_recycle;
6855 }
6856 skb_reserve(skb, TG3_RX_OFFSET(tp));
6857 /* Ensure that the update to the data happens 6865 /* Ensure that the update to the data happens
6858 * after the usage of the old DMA mapping. 6866 * after the usage of the old DMA mapping.
6859 */ 6867 */
@@ -6861,6 +6869,12 @@ static int tg3_rx(struct tg3_napi *tnapi, int budget)
6861 6869
6862 ri->data = NULL; 6870 ri->data = NULL;
6863 6871
6872 skb = build_skb(data, frag_size);
6873 if (!skb) {
6874 tg3_frag_free(frag_size != 0, data);
6875 goto drop_it_no_recycle;
6876 }
6877 skb_reserve(skb, TG3_RX_OFFSET(tp));
6864 } else { 6878 } else {
6865 tg3_recycle_rx(tnapi, tpr, opaque_key, 6879 tg3_recycle_rx(tnapi, tpr, opaque_key,
6866 desc_idx, *post_ptr); 6880 desc_idx, *post_ptr);
@@ -9196,10 +9210,7 @@ static int tg3_halt(struct tg3 *tp, int kind, bool silent)
9196 memset(tp->hw_stats, 0, sizeof(struct tg3_hw_stats)); 9210 memset(tp->hw_stats, 0, sizeof(struct tg3_hw_stats));
9197 } 9211 }
9198 9212
9199 if (err) 9213 return err;
9200 return err;
9201
9202 return 0;
9203} 9214}
9204 9215
9205static int tg3_set_mac_addr(struct net_device *dev, void *p) 9216static int tg3_set_mac_addr(struct net_device *dev, void *p)
@@ -11035,7 +11046,18 @@ static int tg3_request_irq(struct tg3 *tp, int irq_num)
11035 name = tp->dev->name; 11046 name = tp->dev->name;
11036 else { 11047 else {
11037 name = &tnapi->irq_lbl[0]; 11048 name = &tnapi->irq_lbl[0];
11038 snprintf(name, IFNAMSIZ, "%s-%d", tp->dev->name, irq_num); 11049 if (tnapi->tx_buffers && tnapi->rx_rcb)
11050 snprintf(name, IFNAMSIZ,
11051 "%s-txrx-%d", tp->dev->name, irq_num);
11052 else if (tnapi->tx_buffers)
11053 snprintf(name, IFNAMSIZ,
11054 "%s-tx-%d", tp->dev->name, irq_num);
11055 else if (tnapi->rx_rcb)
11056 snprintf(name, IFNAMSIZ,
11057 "%s-rx-%d", tp->dev->name, irq_num);
11058 else
11059 snprintf(name, IFNAMSIZ,
11060 "%s-%d", tp->dev->name, irq_num);
11039 name[IFNAMSIZ-1] = 0; 11061 name[IFNAMSIZ-1] = 0;
11040 } 11062 }
11041 11063
@@ -11907,7 +11929,7 @@ static int tg3_get_settings(struct net_device *dev, struct ethtool_cmd *cmd)
11907 struct phy_device *phydev; 11929 struct phy_device *phydev;
11908 if (!(tp->phy_flags & TG3_PHYFLG_IS_CONNECTED)) 11930 if (!(tp->phy_flags & TG3_PHYFLG_IS_CONNECTED))
11909 return -EAGAIN; 11931 return -EAGAIN;
11910 phydev = tp->mdio_bus->phy_map[TG3_PHY_MII_ADDR]; 11932 phydev = tp->mdio_bus->phy_map[tp->phy_addr];
11911 return phy_ethtool_gset(phydev, cmd); 11933 return phy_ethtool_gset(phydev, cmd);
11912 } 11934 }
11913 11935
@@ -11974,7 +11996,7 @@ static int tg3_set_settings(struct net_device *dev, struct ethtool_cmd *cmd)
11974 struct phy_device *phydev; 11996 struct phy_device *phydev;
11975 if (!(tp->phy_flags & TG3_PHYFLG_IS_CONNECTED)) 11997 if (!(tp->phy_flags & TG3_PHYFLG_IS_CONNECTED))
11976 return -EAGAIN; 11998 return -EAGAIN;
11977 phydev = tp->mdio_bus->phy_map[TG3_PHY_MII_ADDR]; 11999 phydev = tp->mdio_bus->phy_map[tp->phy_addr];
11978 return phy_ethtool_sset(phydev, cmd); 12000 return phy_ethtool_sset(phydev, cmd);
11979 } 12001 }
11980 12002
@@ -12093,12 +12115,10 @@ static int tg3_set_wol(struct net_device *dev, struct ethtool_wolinfo *wol)
12093 12115
12094 device_set_wakeup_enable(dp, wol->wolopts & WAKE_MAGIC); 12116 device_set_wakeup_enable(dp, wol->wolopts & WAKE_MAGIC);
12095 12117
12096 spin_lock_bh(&tp->lock);
12097 if (device_may_wakeup(dp)) 12118 if (device_may_wakeup(dp))
12098 tg3_flag_set(tp, WOL_ENABLE); 12119 tg3_flag_set(tp, WOL_ENABLE);
12099 else 12120 else
12100 tg3_flag_clear(tp, WOL_ENABLE); 12121 tg3_flag_clear(tp, WOL_ENABLE);
12101 spin_unlock_bh(&tp->lock);
12102 12122
12103 return 0; 12123 return 0;
12104} 12124}
@@ -12131,7 +12151,7 @@ static int tg3_nway_reset(struct net_device *dev)
12131 if (tg3_flag(tp, USE_PHYLIB)) { 12151 if (tg3_flag(tp, USE_PHYLIB)) {
12132 if (!(tp->phy_flags & TG3_PHYFLG_IS_CONNECTED)) 12152 if (!(tp->phy_flags & TG3_PHYFLG_IS_CONNECTED))
12133 return -EAGAIN; 12153 return -EAGAIN;
12134 r = phy_start_aneg(tp->mdio_bus->phy_map[TG3_PHY_MII_ADDR]); 12154 r = phy_start_aneg(tp->mdio_bus->phy_map[tp->phy_addr]);
12135 } else { 12155 } else {
12136 u32 bmcr; 12156 u32 bmcr;
12137 12157
@@ -12247,7 +12267,7 @@ static int tg3_set_pauseparam(struct net_device *dev, struct ethtool_pauseparam
12247 u32 newadv; 12267 u32 newadv;
12248 struct phy_device *phydev; 12268 struct phy_device *phydev;
12249 12269
12250 phydev = tp->mdio_bus->phy_map[TG3_PHY_MII_ADDR]; 12270 phydev = tp->mdio_bus->phy_map[tp->phy_addr];
12251 12271
12252 if (!(phydev->supported & SUPPORTED_Pause) || 12272 if (!(phydev->supported & SUPPORTED_Pause) ||
12253 (!(phydev->supported & SUPPORTED_Asym_Pause) && 12273 (!(phydev->supported & SUPPORTED_Asym_Pause) &&
@@ -13194,8 +13214,8 @@ static int tg3_run_loopback(struct tg3 *tp, u32 pktsz, bool tso_loopback)
13194 return -ENOMEM; 13214 return -ENOMEM;
13195 13215
13196 tx_data = skb_put(skb, tx_len); 13216 tx_data = skb_put(skb, tx_len);
13197 memcpy(tx_data, tp->dev->dev_addr, 6); 13217 memcpy(tx_data, tp->dev->dev_addr, ETH_ALEN);
13198 memset(tx_data + 6, 0x0, 8); 13218 memset(tx_data + ETH_ALEN, 0x0, 8);
13199 13219
13200 tw32(MAC_RX_MTU_SIZE, tx_len + ETH_FCS_LEN); 13220 tw32(MAC_RX_MTU_SIZE, tx_len + ETH_FCS_LEN);
13201 13221
@@ -13683,7 +13703,7 @@ static int tg3_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd)
13683 struct phy_device *phydev; 13703 struct phy_device *phydev;
13684 if (!(tp->phy_flags & TG3_PHYFLG_IS_CONNECTED)) 13704 if (!(tp->phy_flags & TG3_PHYFLG_IS_CONNECTED))
13685 return -EAGAIN; 13705 return -EAGAIN;
13686 phydev = tp->mdio_bus->phy_map[TG3_PHY_MII_ADDR]; 13706 phydev = tp->mdio_bus->phy_map[tp->phy_addr];
13687 return phy_mii_ioctl(phydev, ifr, cmd); 13707 return phy_mii_ioctl(phydev, ifr, cmd);
13688 } 13708 }
13689 13709
@@ -14921,6 +14941,12 @@ static void tg3_get_eeprom_hw_cfg(struct tg3 *tp)
14921 tg3_chip_rev_id(tp) != CHIPREV_ID_5750_A1) 14941 tg3_chip_rev_id(tp) != CHIPREV_ID_5750_A1)
14922 tp->led_ctrl |= (LED_CTRL_MODE_PHY_1 | 14942 tp->led_ctrl |= (LED_CTRL_MODE_PHY_1 |
14923 LED_CTRL_MODE_PHY_2); 14943 LED_CTRL_MODE_PHY_2);
14944
14945 if (tg3_flag(tp, 5717_PLUS) ||
14946 tg3_asic_rev(tp) == ASIC_REV_5762)
14947 tp->led_ctrl |= LED_CTRL_BLINK_RATE_OVERRIDE |
14948 LED_CTRL_BLINK_RATE_MASK;
14949
14924 break; 14950 break;
14925 14951
14926 case SHASTA_EXT_LED_MAC: 14952 case SHASTA_EXT_LED_MAC:
@@ -15759,9 +15785,12 @@ static void tg3_detect_asic_rev(struct tg3 *tp, u32 misc_ctrl_reg)
15759 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5718 || 15785 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5718 ||
15760 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5719 || 15786 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5719 ||
15761 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5720 || 15787 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5720 ||
15788 tp->pdev->device == TG3PCI_DEVICE_TIGON3_57767 ||
15789 tp->pdev->device == TG3PCI_DEVICE_TIGON3_57764 ||
15762 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5762 || 15790 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5762 ||
15763 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5725 || 15791 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5725 ||
15764 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5727) 15792 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5727 ||
15793 tp->pdev->device == TG3PCI_DEVICE_TIGON3_57787)
15765 reg = TG3PCI_GEN2_PRODID_ASICREV; 15794 reg = TG3PCI_GEN2_PRODID_ASICREV;
15766 else if (tp->pdev->device == TG3PCI_DEVICE_TIGON3_57781 || 15795 else if (tp->pdev->device == TG3PCI_DEVICE_TIGON3_57781 ||
15767 tp->pdev->device == TG3PCI_DEVICE_TIGON3_57785 || 15796 tp->pdev->device == TG3PCI_DEVICE_TIGON3_57785 ||
@@ -16632,8 +16661,8 @@ static int tg3_get_macaddr_sparc(struct tg3 *tp)
16632 int len; 16661 int len;
16633 16662
16634 addr = of_get_property(dp, "local-mac-address", &len); 16663 addr = of_get_property(dp, "local-mac-address", &len);
16635 if (addr && len == 6) { 16664 if (addr && len == ETH_ALEN) {
16636 memcpy(dev->dev_addr, addr, 6); 16665 memcpy(dev->dev_addr, addr, ETH_ALEN);
16637 return 0; 16666 return 0;
16638 } 16667 }
16639 return -ENODEV; 16668 return -ENODEV;
@@ -16643,7 +16672,7 @@ static int tg3_get_default_macaddr_sparc(struct tg3 *tp)
16643{ 16672{
16644 struct net_device *dev = tp->dev; 16673 struct net_device *dev = tp->dev;
16645 16674
16646 memcpy(dev->dev_addr, idprom->id_ethaddr, 6); 16675 memcpy(dev->dev_addr, idprom->id_ethaddr, ETH_ALEN);
16647 return 0; 16676 return 0;
16648} 16677}
16649#endif 16678#endif
@@ -17052,10 +17081,6 @@ static int tg3_test_dma(struct tg3 *tp)
17052 17081
17053 tw32(TG3PCI_DMA_RW_CTRL, tp->dma_rwctrl); 17082 tw32(TG3PCI_DMA_RW_CTRL, tp->dma_rwctrl);
17054 17083
17055#if 0
17056 /* Unneeded, already done by tg3_get_invariants. */
17057 tg3_switch_clocks(tp);
17058#endif
17059 17084
17060 if (tg3_asic_rev(tp) != ASIC_REV_5700 && 17085 if (tg3_asic_rev(tp) != ASIC_REV_5700 &&
17061 tg3_asic_rev(tp) != ASIC_REV_5701) 17086 tg3_asic_rev(tp) != ASIC_REV_5701)
@@ -17083,20 +17108,6 @@ static int tg3_test_dma(struct tg3 *tp)
17083 break; 17108 break;
17084 } 17109 }
17085 17110
17086#if 0
17087 /* validate data reached card RAM correctly. */
17088 for (i = 0; i < TEST_BUFFER_SIZE / sizeof(u32); i++) {
17089 u32 val;
17090 tg3_read_mem(tp, 0x2100 + (i*4), &val);
17091 if (le32_to_cpu(val) != p[i]) {
17092 dev_err(&tp->pdev->dev,
17093 "%s: Buffer corrupted on device! "
17094 "(%d != %d)\n", __func__, val, i);
17095 /* ret = -ENODEV here? */
17096 }
17097 p[i] = 0;
17098 }
17099#endif
17100 /* Now read it back. */ 17111 /* Now read it back. */
17101 ret = tg3_do_test_dma(tp, buf, buf_dma, TEST_BUFFER_SIZE, false); 17112 ret = tg3_do_test_dma(tp, buf, buf_dma, TEST_BUFFER_SIZE, false);
17102 if (ret) { 17113 if (ret) {
@@ -17362,8 +17373,10 @@ static int tg3_init_one(struct pci_dev *pdev,
17362 tg3_flag_set(tp, FLUSH_POSTED_WRITES); 17373 tg3_flag_set(tp, FLUSH_POSTED_WRITES);
17363 if (ssb_gige_one_dma_at_once(pdev)) 17374 if (ssb_gige_one_dma_at_once(pdev))
17364 tg3_flag_set(tp, ONE_DMA_AT_ONCE); 17375 tg3_flag_set(tp, ONE_DMA_AT_ONCE);
17365 if (ssb_gige_have_roboswitch(pdev)) 17376 if (ssb_gige_have_roboswitch(pdev)) {
17377 tg3_flag_set(tp, USE_PHYLIB);
17366 tg3_flag_set(tp, ROBOSWITCH); 17378 tg3_flag_set(tp, ROBOSWITCH);
17379 }
17367 if (ssb_gige_is_rgmii(pdev)) 17380 if (ssb_gige_is_rgmii(pdev))
17368 tg3_flag_set(tp, RGMII_MODE); 17381 tg3_flag_set(tp, RGMII_MODE);
17369 } 17382 }
@@ -17409,9 +17422,12 @@ static int tg3_init_one(struct pci_dev *pdev,
17409 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5718 || 17422 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5718 ||
17410 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5719 || 17423 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5719 ||
17411 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5720 || 17424 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5720 ||
17425 tp->pdev->device == TG3PCI_DEVICE_TIGON3_57767 ||
17426 tp->pdev->device == TG3PCI_DEVICE_TIGON3_57764 ||
17412 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5762 || 17427 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5762 ||
17413 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5725 || 17428 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5725 ||
17414 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5727) { 17429 tp->pdev->device == TG3PCI_DEVICE_TIGON3_5727 ||
17430 tp->pdev->device == TG3PCI_DEVICE_TIGON3_57787) {
17415 tg3_flag_set(tp, ENABLE_APE); 17431 tg3_flag_set(tp, ENABLE_APE);
17416 tp->aperegs = pci_ioremap_bar(pdev, BAR_2); 17432 tp->aperegs = pci_ioremap_bar(pdev, BAR_2);
17417 if (!tp->aperegs) { 17433 if (!tp->aperegs) {
@@ -17628,7 +17644,7 @@ static int tg3_init_one(struct pci_dev *pdev,
17628 17644
17629 if (tp->phy_flags & TG3_PHYFLG_IS_CONNECTED) { 17645 if (tp->phy_flags & TG3_PHYFLG_IS_CONNECTED) {
17630 struct phy_device *phydev; 17646 struct phy_device *phydev;
17631 phydev = tp->mdio_bus->phy_map[TG3_PHY_MII_ADDR]; 17647 phydev = tp->mdio_bus->phy_map[tp->phy_addr];
17632 netdev_info(dev, 17648 netdev_info(dev,
17633 "attached PHY driver [%s] (mii_bus:phy_addr=%s)\n", 17649 "attached PHY driver [%s] (mii_bus:phy_addr=%s)\n",
17634 phydev->drv->name, dev_name(&phydev->dev)); 17650 phydev->drv->name, dev_name(&phydev->dev));
@@ -17685,7 +17701,6 @@ err_out_free_res:
17685err_out_disable_pdev: 17701err_out_disable_pdev:
17686 if (pci_is_enabled(pdev)) 17702 if (pci_is_enabled(pdev))
17687 pci_disable_device(pdev); 17703 pci_disable_device(pdev);
17688 pci_set_drvdata(pdev, NULL);
17689 return err; 17704 return err;
17690} 17705}
17691 17706
@@ -17717,7 +17732,6 @@ static void tg3_remove_one(struct pci_dev *pdev)
17717 free_netdev(dev); 17732 free_netdev(dev);
17718 pci_release_regions(pdev); 17733 pci_release_regions(pdev);
17719 pci_disable_device(pdev); 17734 pci_disable_device(pdev);
17720 pci_set_drvdata(pdev, NULL);
17721 } 17735 }
17722} 17736}
17723 17737
diff --git a/drivers/net/ethernet/broadcom/tg3.h b/drivers/net/ethernet/broadcom/tg3.h
index 70257808aa37..5c3835aa1e1b 100644
--- a/drivers/net/ethernet/broadcom/tg3.h
+++ b/drivers/net/ethernet/broadcom/tg3.h
@@ -68,6 +68,9 @@
68#define TG3PCI_DEVICE_TIGON3_5762 0x1687 68#define TG3PCI_DEVICE_TIGON3_5762 0x1687
69#define TG3PCI_DEVICE_TIGON3_5725 0x1643 69#define TG3PCI_DEVICE_TIGON3_5725 0x1643
70#define TG3PCI_DEVICE_TIGON3_5727 0x16f3 70#define TG3PCI_DEVICE_TIGON3_5727 0x16f3
71#define TG3PCI_DEVICE_TIGON3_57764 0x1642
72#define TG3PCI_DEVICE_TIGON3_57767 0x1683
73#define TG3PCI_DEVICE_TIGON3_57787 0x1641
71/* 0x04 --> 0x2c unused */ 74/* 0x04 --> 0x2c unused */
72#define TG3PCI_SUBVENDOR_ID_BROADCOM PCI_VENDOR_ID_BROADCOM 75#define TG3PCI_SUBVENDOR_ID_BROADCOM PCI_VENDOR_ID_BROADCOM
73#define TG3PCI_SUBDEVICE_ID_BROADCOM_95700A6 0x1644 76#define TG3PCI_SUBDEVICE_ID_BROADCOM_95700A6 0x1644
diff --git a/drivers/net/ethernet/brocade/bna/bnad.c b/drivers/net/ethernet/brocade/bna/bnad.c
index b78e69e0e52a..f276433d37ce 100644
--- a/drivers/net/ethernet/brocade/bna/bnad.c
+++ b/drivers/net/ethernet/brocade/bna/bnad.c
@@ -3212,7 +3212,6 @@ bnad_init(struct bnad *bnad,
3212 bnad->bar0 = ioremap_nocache(bnad->mmio_start, bnad->mmio_len); 3212 bnad->bar0 = ioremap_nocache(bnad->mmio_start, bnad->mmio_len);
3213 if (!bnad->bar0) { 3213 if (!bnad->bar0) {
3214 dev_err(&pdev->dev, "ioremap for bar0 failed\n"); 3214 dev_err(&pdev->dev, "ioremap for bar0 failed\n");
3215 pci_set_drvdata(pdev, NULL);
3216 return -ENOMEM; 3215 return -ENOMEM;
3217 } 3216 }
3218 pr_info("bar0 mapped to %p, len %llu\n", bnad->bar0, 3217 pr_info("bar0 mapped to %p, len %llu\n", bnad->bar0,
diff --git a/drivers/net/ethernet/brocade/bna/bnad.h b/drivers/net/ethernet/brocade/bna/bnad.h
index aefee77523f2..f7e033f8a00e 100644
--- a/drivers/net/ethernet/brocade/bna/bnad.h
+++ b/drivers/net/ethernet/brocade/bna/bnad.h
@@ -372,38 +372,37 @@ extern u32 bnad_rxqs_per_cq;
372/* 372/*
373 * EXTERN PROTOTYPES 373 * EXTERN PROTOTYPES
374 */ 374 */
375extern u32 *cna_get_firmware_buf(struct pci_dev *pdev); 375u32 *cna_get_firmware_buf(struct pci_dev *pdev);
376/* Netdev entry point prototypes */ 376/* Netdev entry point prototypes */
377extern void bnad_set_rx_mode(struct net_device *netdev); 377void bnad_set_rx_mode(struct net_device *netdev);
378extern struct net_device_stats *bnad_get_netdev_stats( 378struct net_device_stats *bnad_get_netdev_stats(struct net_device *netdev);
379 struct net_device *netdev); 379int bnad_mac_addr_set_locked(struct bnad *bnad, u8 *mac_addr);
380extern int bnad_mac_addr_set_locked(struct bnad *bnad, u8 *mac_addr); 380int bnad_enable_default_bcast(struct bnad *bnad);
381extern int bnad_enable_default_bcast(struct bnad *bnad); 381void bnad_restore_vlans(struct bnad *bnad, u32 rx_id);
382extern void bnad_restore_vlans(struct bnad *bnad, u32 rx_id); 382void bnad_set_ethtool_ops(struct net_device *netdev);
383extern void bnad_set_ethtool_ops(struct net_device *netdev); 383void bnad_cb_completion(void *arg, enum bfa_status status);
384extern void bnad_cb_completion(void *arg, enum bfa_status status);
385 384
386/* Configuration & setup */ 385/* Configuration & setup */
387extern void bnad_tx_coalescing_timeo_set(struct bnad *bnad); 386void bnad_tx_coalescing_timeo_set(struct bnad *bnad);
388extern void bnad_rx_coalescing_timeo_set(struct bnad *bnad); 387void bnad_rx_coalescing_timeo_set(struct bnad *bnad);
389 388
390extern int bnad_setup_rx(struct bnad *bnad, u32 rx_id); 389int bnad_setup_rx(struct bnad *bnad, u32 rx_id);
391extern int bnad_setup_tx(struct bnad *bnad, u32 tx_id); 390int bnad_setup_tx(struct bnad *bnad, u32 tx_id);
392extern void bnad_destroy_tx(struct bnad *bnad, u32 tx_id); 391void bnad_destroy_tx(struct bnad *bnad, u32 tx_id);
393extern void bnad_destroy_rx(struct bnad *bnad, u32 rx_id); 392void bnad_destroy_rx(struct bnad *bnad, u32 rx_id);
394 393
395/* Timer start/stop protos */ 394/* Timer start/stop protos */
396extern void bnad_dim_timer_start(struct bnad *bnad); 395void bnad_dim_timer_start(struct bnad *bnad);
397 396
398/* Statistics */ 397/* Statistics */
399extern void bnad_netdev_qstats_fill(struct bnad *bnad, 398void bnad_netdev_qstats_fill(struct bnad *bnad,
400 struct rtnl_link_stats64 *stats); 399 struct rtnl_link_stats64 *stats);
401extern void bnad_netdev_hwstats_fill(struct bnad *bnad, 400void bnad_netdev_hwstats_fill(struct bnad *bnad,
402 struct rtnl_link_stats64 *stats); 401 struct rtnl_link_stats64 *stats);
403 402
404/* Debugfs */ 403/* Debugfs */
405void bnad_debugfs_init(struct bnad *bnad); 404void bnad_debugfs_init(struct bnad *bnad);
406void bnad_debugfs_uninit(struct bnad *bnad); 405void bnad_debugfs_uninit(struct bnad *bnad);
407 406
408/* MACROS */ 407/* MACROS */
409/* To set & get the stats counters */ 408/* To set & get the stats counters */
diff --git a/drivers/net/ethernet/calxeda/xgmac.c b/drivers/net/ethernet/calxeda/xgmac.c
index 48f52882a22b..4fc5c8ef5121 100644
--- a/drivers/net/ethernet/calxeda/xgmac.c
+++ b/drivers/net/ethernet/calxeda/xgmac.c
@@ -1060,13 +1060,13 @@ static int xgmac_stop(struct net_device *dev)
1060{ 1060{
1061 struct xgmac_priv *priv = netdev_priv(dev); 1061 struct xgmac_priv *priv = netdev_priv(dev);
1062 1062
1063 netif_stop_queue(dev);
1064
1065 if (readl(priv->base + XGMAC_DMA_INTR_ENA)) 1063 if (readl(priv->base + XGMAC_DMA_INTR_ENA))
1066 napi_disable(&priv->napi); 1064 napi_disable(&priv->napi);
1067 1065
1068 writel(0, priv->base + XGMAC_DMA_INTR_ENA); 1066 writel(0, priv->base + XGMAC_DMA_INTR_ENA);
1069 1067
1068 netif_tx_disable(dev);
1069
1070 /* Disable the MAC core */ 1070 /* Disable the MAC core */
1071 xgmac_mac_disable(priv->base); 1071 xgmac_mac_disable(priv->base);
1072 1072
@@ -1370,11 +1370,8 @@ static int xgmac_change_mtu(struct net_device *dev, int new_mtu)
1370 } 1370 }
1371 1371
1372 old_mtu = dev->mtu; 1372 old_mtu = dev->mtu;
1373 dev->mtu = new_mtu;
1374 1373
1375 /* return early if the buffer sizes will not change */ 1374 /* return early if the buffer sizes will not change */
1376 if (old_mtu <= ETH_DATA_LEN && new_mtu <= ETH_DATA_LEN)
1377 return 0;
1378 if (old_mtu == new_mtu) 1375 if (old_mtu == new_mtu)
1379 return 0; 1376 return 0;
1380 1377
@@ -1382,8 +1379,9 @@ static int xgmac_change_mtu(struct net_device *dev, int new_mtu)
1382 if (!netif_running(dev)) 1379 if (!netif_running(dev))
1383 return 0; 1380 return 0;
1384 1381
1385 /* Bring the interface down and then back up */ 1382 /* Bring interface down, change mtu and bring interface back up */
1386 xgmac_stop(dev); 1383 xgmac_stop(dev);
1384 dev->mtu = new_mtu;
1387 return xgmac_open(dev); 1385 return xgmac_open(dev);
1388} 1386}
1389 1387
diff --git a/drivers/net/ethernet/chelsio/cxgb/common.h b/drivers/net/ethernet/chelsio/cxgb/common.h
index 5ccbed1784d2..8abb46b39032 100644
--- a/drivers/net/ethernet/chelsio/cxgb/common.h
+++ b/drivers/net/ethernet/chelsio/cxgb/common.h
@@ -324,30 +324,30 @@ static inline unsigned int core_ticks_per_usec(const adapter_t *adap)
324 return board_info(adap)->clock_core / 1000000; 324 return board_info(adap)->clock_core / 1000000;
325} 325}
326 326
327extern int __t1_tpi_read(adapter_t *adapter, u32 addr, u32 *valp); 327int __t1_tpi_read(adapter_t *adapter, u32 addr, u32 *valp);
328extern int __t1_tpi_write(adapter_t *adapter, u32 addr, u32 value); 328int __t1_tpi_write(adapter_t *adapter, u32 addr, u32 value);
329extern int t1_tpi_write(adapter_t *adapter, u32 addr, u32 value); 329int t1_tpi_write(adapter_t *adapter, u32 addr, u32 value);
330extern int t1_tpi_read(adapter_t *adapter, u32 addr, u32 *value); 330int t1_tpi_read(adapter_t *adapter, u32 addr, u32 *value);
331 331
332extern void t1_interrupts_enable(adapter_t *adapter); 332void t1_interrupts_enable(adapter_t *adapter);
333extern void t1_interrupts_disable(adapter_t *adapter); 333void t1_interrupts_disable(adapter_t *adapter);
334extern void t1_interrupts_clear(adapter_t *adapter); 334void t1_interrupts_clear(adapter_t *adapter);
335extern int t1_elmer0_ext_intr_handler(adapter_t *adapter); 335int t1_elmer0_ext_intr_handler(adapter_t *adapter);
336extern void t1_elmer0_ext_intr(adapter_t *adapter); 336void t1_elmer0_ext_intr(adapter_t *adapter);
337extern int t1_slow_intr_handler(adapter_t *adapter); 337int t1_slow_intr_handler(adapter_t *adapter);
338 338
339extern int t1_link_start(struct cphy *phy, struct cmac *mac, struct link_config *lc); 339int t1_link_start(struct cphy *phy, struct cmac *mac, struct link_config *lc);
340extern const struct board_info *t1_get_board_info(unsigned int board_id); 340const struct board_info *t1_get_board_info(unsigned int board_id);
341extern const struct board_info *t1_get_board_info_from_ids(unsigned int devid, 341const struct board_info *t1_get_board_info_from_ids(unsigned int devid,
342 unsigned short ssid); 342 unsigned short ssid);
343extern int t1_seeprom_read(adapter_t *adapter, u32 addr, __le32 *data); 343int t1_seeprom_read(adapter_t *adapter, u32 addr, __le32 *data);
344extern int t1_get_board_rev(adapter_t *adapter, const struct board_info *bi, 344int t1_get_board_rev(adapter_t *adapter, const struct board_info *bi,
345 struct adapter_params *p); 345 struct adapter_params *p);
346extern int t1_init_hw_modules(adapter_t *adapter); 346int t1_init_hw_modules(adapter_t *adapter);
347extern int t1_init_sw_modules(adapter_t *adapter, const struct board_info *bi); 347int t1_init_sw_modules(adapter_t *adapter, const struct board_info *bi);
348extern void t1_free_sw_modules(adapter_t *adapter); 348void t1_free_sw_modules(adapter_t *adapter);
349extern void t1_fatal_err(adapter_t *adapter); 349void t1_fatal_err(adapter_t *adapter);
350extern void t1_link_changed(adapter_t *adapter, int port_id); 350void t1_link_changed(adapter_t *adapter, int port_id);
351extern void t1_link_negotiated(adapter_t *adapter, int port_id, int link_stat, 351void t1_link_negotiated(adapter_t *adapter, int port_id, int link_stat,
352 int speed, int duplex, int pause); 352 int speed, int duplex, int pause);
353#endif /* _CXGB_COMMON_H_ */ 353#endif /* _CXGB_COMMON_H_ */
diff --git a/drivers/net/ethernet/chelsio/cxgb/cxgb2.c b/drivers/net/ethernet/chelsio/cxgb/cxgb2.c
index d7048db9863d..1d021059f097 100644
--- a/drivers/net/ethernet/chelsio/cxgb/cxgb2.c
+++ b/drivers/net/ethernet/chelsio/cxgb/cxgb2.c
@@ -1168,7 +1168,6 @@ out_free_dev:
1168 pci_release_regions(pdev); 1168 pci_release_regions(pdev);
1169out_disable_pdev: 1169out_disable_pdev:
1170 pci_disable_device(pdev); 1170 pci_disable_device(pdev);
1171 pci_set_drvdata(pdev, NULL);
1172 return err; 1171 return err;
1173} 1172}
1174 1173
@@ -1347,7 +1346,6 @@ static void remove_one(struct pci_dev *pdev)
1347 1346
1348 pci_release_regions(pdev); 1347 pci_release_regions(pdev);
1349 pci_disable_device(pdev); 1348 pci_disable_device(pdev);
1350 pci_set_drvdata(pdev, NULL);
1351 t1_sw_reset(pdev); 1349 t1_sw_reset(pdev);
1352} 1350}
1353 1351
diff --git a/drivers/net/ethernet/chelsio/cxgb/pm3393.c b/drivers/net/ethernet/chelsio/cxgb/pm3393.c
index 40c7b93ababc..eb33a31b08a0 100644
--- a/drivers/net/ethernet/chelsio/cxgb/pm3393.c
+++ b/drivers/net/ethernet/chelsio/cxgb/pm3393.c
@@ -499,7 +499,7 @@ static const struct cmac_statistics *pm3393_update_statistics(struct cmac *mac,
499 499
500static int pm3393_macaddress_get(struct cmac *cmac, u8 mac_addr[6]) 500static int pm3393_macaddress_get(struct cmac *cmac, u8 mac_addr[6])
501{ 501{
502 memcpy(mac_addr, cmac->instance->mac_addr, 6); 502 memcpy(mac_addr, cmac->instance->mac_addr, ETH_ALEN);
503 return 0; 503 return 0;
504} 504}
505 505
@@ -526,7 +526,7 @@ static int pm3393_macaddress_set(struct cmac *cmac, u8 ma[6])
526 */ 526 */
527 527
528 /* Store local copy */ 528 /* Store local copy */
529 memcpy(cmac->instance->mac_addr, ma, 6); 529 memcpy(cmac->instance->mac_addr, ma, ETH_ALEN);
530 530
531 lo = ((u32) ma[1] << 8) | (u32) ma[0]; 531 lo = ((u32) ma[1] << 8) | (u32) ma[0];
532 mid = ((u32) ma[3] << 8) | (u32) ma[2]; 532 mid = ((u32) ma[3] << 8) | (u32) ma[2];
diff --git a/drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c b/drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
index b650951791dd..45d77334d7d9 100644
--- a/drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
+++ b/drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
@@ -3374,7 +3374,6 @@ out_release_regions:
3374 pci_release_regions(pdev); 3374 pci_release_regions(pdev);
3375out_disable_device: 3375out_disable_device:
3376 pci_disable_device(pdev); 3376 pci_disable_device(pdev);
3377 pci_set_drvdata(pdev, NULL);
3378out: 3377out:
3379 return err; 3378 return err;
3380} 3379}
@@ -3415,7 +3414,6 @@ static void remove_one(struct pci_dev *pdev)
3415 kfree(adapter); 3414 kfree(adapter);
3416 pci_release_regions(pdev); 3415 pci_release_regions(pdev);
3417 pci_disable_device(pdev); 3416 pci_disable_device(pdev);
3418 pci_set_drvdata(pdev, NULL);
3419 } 3417 }
3420} 3418}
3421 3419
diff --git a/drivers/net/ethernet/chelsio/cxgb3/regs.h b/drivers/net/ethernet/chelsio/cxgb3/regs.h
index 6990f6c65221..81029b872bdd 100644
--- a/drivers/net/ethernet/chelsio/cxgb3/regs.h
+++ b/drivers/net/ethernet/chelsio/cxgb3/regs.h
@@ -685,10 +685,6 @@
685#define V_BUSY(x) ((x) << S_BUSY) 685#define V_BUSY(x) ((x) << S_BUSY)
686#define F_BUSY V_BUSY(1U) 686#define F_BUSY V_BUSY(1U)
687 687
688#define S_BUSY 31
689#define V_BUSY(x) ((x) << S_BUSY)
690#define F_BUSY V_BUSY(1U)
691
692#define A_MC7_EXT_MODE1 0x108 688#define A_MC7_EXT_MODE1 0x108
693 689
694#define A_MC7_EXT_MODE2 0x10c 690#define A_MC7_EXT_MODE2 0x10c
@@ -749,14 +745,6 @@
749 745
750#define A_MC7_CAL 0x128 746#define A_MC7_CAL 0x128
751 747
752#define S_BUSY 31
753#define V_BUSY(x) ((x) << S_BUSY)
754#define F_BUSY V_BUSY(1U)
755
756#define S_BUSY 31
757#define V_BUSY(x) ((x) << S_BUSY)
758#define F_BUSY V_BUSY(1U)
759
760#define S_CAL_FAULT 30 748#define S_CAL_FAULT 30
761#define V_CAL_FAULT(x) ((x) << S_CAL_FAULT) 749#define V_CAL_FAULT(x) ((x) << S_CAL_FAULT)
762#define F_CAL_FAULT V_CAL_FAULT(1U) 750#define F_CAL_FAULT V_CAL_FAULT(1U)
@@ -815,9 +803,6 @@
815#define V_OP(x) ((x) << S_OP) 803#define V_OP(x) ((x) << S_OP)
816#define F_OP V_OP(1U) 804#define F_OP V_OP(1U)
817 805
818#define F_OP V_OP(1U)
819#define A_SF_OP 0x6dc
820
821#define A_MC7_BIST_ADDR_BEG 0x168 806#define A_MC7_BIST_ADDR_BEG 0x168
822 807
823#define A_MC7_BIST_ADDR_END 0x16c 808#define A_MC7_BIST_ADDR_END 0x16c
@@ -830,8 +815,6 @@
830#define V_CONT(x) ((x) << S_CONT) 815#define V_CONT(x) ((x) << S_CONT)
831#define F_CONT V_CONT(1U) 816#define F_CONT V_CONT(1U)
832 817
833#define F_CONT V_CONT(1U)
834
835#define A_MC7_INT_ENABLE 0x178 818#define A_MC7_INT_ENABLE 0x178
836 819
837#define S_AE 17 820#define S_AE 17
@@ -1017,8 +1000,6 @@
1017#define V_NICMODE(x) ((x) << S_NICMODE) 1000#define V_NICMODE(x) ((x) << S_NICMODE)
1018#define F_NICMODE V_NICMODE(1U) 1001#define F_NICMODE V_NICMODE(1U)
1019 1002
1020#define F_NICMODE V_NICMODE(1U)
1021
1022#define S_IPV6ENABLE 15 1003#define S_IPV6ENABLE 15
1023#define V_IPV6ENABLE(x) ((x) << S_IPV6ENABLE) 1004#define V_IPV6ENABLE(x) ((x) << S_IPV6ENABLE)
1024#define F_IPV6ENABLE V_IPV6ENABLE(1U) 1005#define F_IPV6ENABLE V_IPV6ENABLE(1U)
@@ -1562,27 +1543,15 @@
1562#define A_ULPRX_STAG_ULIMIT 0x530 1543#define A_ULPRX_STAG_ULIMIT 0x530
1563 1544
1564#define A_ULPRX_RQ_LLIMIT 0x534 1545#define A_ULPRX_RQ_LLIMIT 0x534
1565#define A_ULPRX_RQ_LLIMIT 0x534
1566 1546
1567#define A_ULPRX_RQ_ULIMIT 0x538 1547#define A_ULPRX_RQ_ULIMIT 0x538
1568#define A_ULPRX_RQ_ULIMIT 0x538
1569 1548
1570#define A_ULPRX_PBL_LLIMIT 0x53c 1549#define A_ULPRX_PBL_LLIMIT 0x53c
1571 1550
1572#define A_ULPRX_PBL_ULIMIT 0x540 1551#define A_ULPRX_PBL_ULIMIT 0x540
1573#define A_ULPRX_PBL_ULIMIT 0x540
1574 1552
1575#define A_ULPRX_TDDP_TAGMASK 0x524 1553#define A_ULPRX_TDDP_TAGMASK 0x524
1576 1554
1577#define A_ULPRX_RQ_LLIMIT 0x534
1578#define A_ULPRX_RQ_LLIMIT 0x534
1579
1580#define A_ULPRX_RQ_ULIMIT 0x538
1581#define A_ULPRX_RQ_ULIMIT 0x538
1582
1583#define A_ULPRX_PBL_ULIMIT 0x540
1584#define A_ULPRX_PBL_ULIMIT 0x540
1585
1586#define A_ULPTX_CONFIG 0x580 1555#define A_ULPTX_CONFIG 0x580
1587 1556
1588#define S_CFG_CQE_SOP_MASK 1 1557#define S_CFG_CQE_SOP_MASK 1
@@ -2053,8 +2022,6 @@
2053#define V_TMMODE(x) ((x) << S_TMMODE) 2022#define V_TMMODE(x) ((x) << S_TMMODE)
2054#define F_TMMODE V_TMMODE(1U) 2023#define F_TMMODE V_TMMODE(1U)
2055 2024
2056#define F_TMMODE V_TMMODE(1U)
2057
2058#define A_MC5_DB_ROUTING_TABLE_INDEX 0x70c 2025#define A_MC5_DB_ROUTING_TABLE_INDEX 0x70c
2059 2026
2060#define A_MC5_DB_FILTER_TABLE 0x710 2027#define A_MC5_DB_FILTER_TABLE 0x710
@@ -2454,8 +2421,6 @@
2454#define V_TXACTENABLE(x) ((x) << S_TXACTENABLE) 2421#define V_TXACTENABLE(x) ((x) << S_TXACTENABLE)
2455#define F_TXACTENABLE V_TXACTENABLE(1U) 2422#define F_TXACTENABLE V_TXACTENABLE(1U)
2456 2423
2457#define A_XGM_SERDES_CTRL0 0x8e0
2458
2459#define S_RESET3 23 2424#define S_RESET3 23
2460#define V_RESET3(x) ((x) << S_RESET3) 2425#define V_RESET3(x) ((x) << S_RESET3)
2461#define F_RESET3 V_RESET3(1U) 2426#define F_RESET3 V_RESET3(1U)
diff --git a/drivers/net/ethernet/chelsio/cxgb4/cxgb4.h b/drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
index dfd1e36f5753..ecd2fb3ef695 100644
--- a/drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
+++ b/drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
@@ -48,7 +48,6 @@
48#include <linux/vmalloc.h> 48#include <linux/vmalloc.h>
49#include <asm/io.h> 49#include <asm/io.h>
50#include "cxgb4_uld.h" 50#include "cxgb4_uld.h"
51#include "t4_hw.h"
52 51
53#define FW_VERSION_MAJOR 1 52#define FW_VERSION_MAJOR 1
54#define FW_VERSION_MINOR 4 53#define FW_VERSION_MINOR 4
diff --git a/drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c b/drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
index c73cabdbd4c0..8b929eeecd2d 100644
--- a/drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
+++ b/drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
@@ -3983,6 +3983,7 @@ static int cxgb4_inet6addr_handler(struct notifier_block *this,
3983 struct net_device *event_dev; 3983 struct net_device *event_dev;
3984 int ret = NOTIFY_DONE; 3984 int ret = NOTIFY_DONE;
3985 struct bonding *bond = netdev_priv(ifa->idev->dev); 3985 struct bonding *bond = netdev_priv(ifa->idev->dev);
3986 struct list_head *iter;
3986 struct slave *slave; 3987 struct slave *slave;
3987 struct pci_dev *first_pdev = NULL; 3988 struct pci_dev *first_pdev = NULL;
3988 3989
@@ -3995,7 +3996,7 @@ static int cxgb4_inet6addr_handler(struct notifier_block *this,
3995 * in all of them only once. 3996 * in all of them only once.
3996 */ 3997 */
3997 read_lock(&bond->lock); 3998 read_lock(&bond->lock);
3998 bond_for_each_slave(bond, slave) { 3999 bond_for_each_slave(bond, slave, iter) {
3999 if (!first_pdev) { 4000 if (!first_pdev) {
4000 ret = clip_add(slave->dev, ifa, event); 4001 ret = clip_add(slave->dev, ifa, event);
4001 /* If clip_add is success then only initialize 4002 /* If clip_add is success then only initialize
@@ -6074,7 +6075,6 @@ sriov:
6074 pci_disable_device(pdev); 6075 pci_disable_device(pdev);
6075 out_release_regions: 6076 out_release_regions:
6076 pci_release_regions(pdev); 6077 pci_release_regions(pdev);
6077 pci_set_drvdata(pdev, NULL);
6078 return err; 6078 return err;
6079} 6079}
6080 6080
@@ -6122,7 +6122,6 @@ static void remove_one(struct pci_dev *pdev)
6122 pci_disable_pcie_error_reporting(pdev); 6122 pci_disable_pcie_error_reporting(pdev);
6123 pci_disable_device(pdev); 6123 pci_disable_device(pdev);
6124 pci_release_regions(pdev); 6124 pci_release_regions(pdev);
6125 pci_set_drvdata(pdev, NULL);
6126 } else 6125 } else
6127 pci_release_regions(pdev); 6126 pci_release_regions(pdev);
6128} 6127}
diff --git a/drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c b/drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
index 40c22e7de15c..5f90ec5f7519 100644
--- a/drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
+++ b/drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
@@ -2782,11 +2782,9 @@ err_unmap_bar:
2782 2782
2783err_free_adapter: 2783err_free_adapter:
2784 kfree(adapter); 2784 kfree(adapter);
2785 pci_set_drvdata(pdev, NULL);
2786 2785
2787err_release_regions: 2786err_release_regions:
2788 pci_release_regions(pdev); 2787 pci_release_regions(pdev);
2789 pci_set_drvdata(pdev, NULL);
2790 pci_clear_master(pdev); 2788 pci_clear_master(pdev);
2791 2789
2792err_disable_device: 2790err_disable_device:
@@ -2851,7 +2849,6 @@ static void cxgb4vf_pci_remove(struct pci_dev *pdev)
2851 } 2849 }
2852 iounmap(adapter->regs); 2850 iounmap(adapter->regs);
2853 kfree(adapter); 2851 kfree(adapter);
2854 pci_set_drvdata(pdev, NULL);
2855 } 2852 }
2856 2853
2857 /* 2854 /*
@@ -2908,7 +2905,7 @@ static void cxgb4vf_pci_shutdown(struct pci_dev *pdev)
2908#define CH_DEVICE(devid, idx) \ 2905#define CH_DEVICE(devid, idx) \
2909 { PCI_VENDOR_ID_CHELSIO, devid, PCI_ANY_ID, PCI_ANY_ID, 0, 0, idx } 2906 { PCI_VENDOR_ID_CHELSIO, devid, PCI_ANY_ID, PCI_ANY_ID, 0, 0, idx }
2910 2907
2911static struct pci_device_id cxgb4vf_pci_tbl[] = { 2908static DEFINE_PCI_DEVICE_TABLE(cxgb4vf_pci_tbl) = {
2912 CH_DEVICE(0xb000, 0), /* PE10K FPGA */ 2909 CH_DEVICE(0xb000, 0), /* PE10K FPGA */
2913 CH_DEVICE(0x4800, 0), /* T440-dbg */ 2910 CH_DEVICE(0x4800, 0), /* T440-dbg */
2914 CH_DEVICE(0x4801, 0), /* T420-cr */ 2911 CH_DEVICE(0x4801, 0), /* T420-cr */
diff --git a/drivers/net/ethernet/chelsio/cxgb4vf/sge.c b/drivers/net/ethernet/chelsio/cxgb4vf/sge.c
index df296af20bd5..8475c4cda9e4 100644
--- a/drivers/net/ethernet/chelsio/cxgb4vf/sge.c
+++ b/drivers/net/ethernet/chelsio/cxgb4vf/sge.c
@@ -1396,8 +1396,9 @@ static inline void copy_frags(struct sk_buff *skb,
1396 * Builds an sk_buff from the given packet gather list. Returns the 1396 * Builds an sk_buff from the given packet gather list. Returns the
1397 * sk_buff or %NULL if sk_buff allocation failed. 1397 * sk_buff or %NULL if sk_buff allocation failed.
1398 */ 1398 */
1399struct sk_buff *t4vf_pktgl_to_skb(const struct pkt_gl *gl, 1399static struct sk_buff *t4vf_pktgl_to_skb(const struct pkt_gl *gl,
1400 unsigned int skb_len, unsigned int pull_len) 1400 unsigned int skb_len,
1401 unsigned int pull_len)
1401{ 1402{
1402 struct sk_buff *skb; 1403 struct sk_buff *skb;
1403 1404
@@ -1443,7 +1444,7 @@ out:
1443 * Releases the pages of a packet gather list. We do not own the last 1444 * Releases the pages of a packet gather list. We do not own the last
1444 * page on the list and do not free it. 1445 * page on the list and do not free it.
1445 */ 1446 */
1446void t4vf_pktgl_free(const struct pkt_gl *gl) 1447static void t4vf_pktgl_free(const struct pkt_gl *gl)
1447{ 1448{
1448 int frag; 1449 int frag;
1449 1450
@@ -1640,7 +1641,7 @@ static inline void rspq_next(struct sge_rspq *rspq)
1640 * on this queue. If the system is under memory shortage use a fairly 1641 * on this queue. If the system is under memory shortage use a fairly
1641 * long delay to help recovery. 1642 * long delay to help recovery.
1642 */ 1643 */
1643int process_responses(struct sge_rspq *rspq, int budget) 1644static int process_responses(struct sge_rspq *rspq, int budget)
1644{ 1645{
1645 struct sge_eth_rxq *rxq = container_of(rspq, struct sge_eth_rxq, rspq); 1646 struct sge_eth_rxq *rxq = container_of(rspq, struct sge_eth_rxq, rspq);
1646 int budget_left = budget; 1647 int budget_left = budget;
@@ -1893,7 +1894,7 @@ static unsigned int process_intrq(struct adapter *adapter)
1893 * The MSI interrupt handler handles data events from SGE response queues as 1894 * The MSI interrupt handler handles data events from SGE response queues as
1894 * well as error and other async events as they all use the same MSI vector. 1895 * well as error and other async events as they all use the same MSI vector.
1895 */ 1896 */
1896irqreturn_t t4vf_intr_msi(int irq, void *cookie) 1897static irqreturn_t t4vf_intr_msi(int irq, void *cookie)
1897{ 1898{
1898 struct adapter *adapter = cookie; 1899 struct adapter *adapter = cookie;
1899 1900
diff --git a/drivers/net/ethernet/cisco/enic/enic_main.c b/drivers/net/ethernet/cisco/enic/enic_main.c
index 7b756cf9474a..ff78dfaec508 100644
--- a/drivers/net/ethernet/cisco/enic/enic_main.c
+++ b/drivers/net/ethernet/cisco/enic/enic_main.c
@@ -2309,7 +2309,6 @@ err_out_release_regions:
2309err_out_disable_device: 2309err_out_disable_device:
2310 pci_disable_device(pdev); 2310 pci_disable_device(pdev);
2311err_out_free_netdev: 2311err_out_free_netdev:
2312 pci_set_drvdata(pdev, NULL);
2313 free_netdev(netdev); 2312 free_netdev(netdev);
2314 2313
2315 return err; 2314 return err;
@@ -2338,7 +2337,6 @@ static void enic_remove(struct pci_dev *pdev)
2338 enic_iounmap(enic); 2337 enic_iounmap(enic);
2339 pci_release_regions(pdev); 2338 pci_release_regions(pdev);
2340 pci_disable_device(pdev); 2339 pci_disable_device(pdev);
2341 pci_set_drvdata(pdev, NULL);
2342 free_netdev(netdev); 2340 free_netdev(netdev);
2343 } 2341 }
2344} 2342}
diff --git a/drivers/net/ethernet/davicom/dm9000.c b/drivers/net/ethernet/davicom/dm9000.c
index a7a941b1a655..7080ad6c4014 100644
--- a/drivers/net/ethernet/davicom/dm9000.c
+++ b/drivers/net/ethernet/davicom/dm9000.c
@@ -1623,7 +1623,7 @@ dm9000_probe(struct platform_device *pdev)
1623 1623
1624 if (!is_valid_ether_addr(ndev->dev_addr) && pdata != NULL) { 1624 if (!is_valid_ether_addr(ndev->dev_addr) && pdata != NULL) {
1625 mac_src = "platform data"; 1625 mac_src = "platform data";
1626 memcpy(ndev->dev_addr, pdata->dev_addr, 6); 1626 memcpy(ndev->dev_addr, pdata->dev_addr, ETH_ALEN);
1627 } 1627 }
1628 1628
1629 if (!is_valid_ether_addr(ndev->dev_addr)) { 1629 if (!is_valid_ether_addr(ndev->dev_addr)) {
diff --git a/drivers/net/ethernet/dec/tulip/de2104x.c b/drivers/net/ethernet/dec/tulip/de2104x.c
index eaab73cf27ca..38148b0e3a95 100644
--- a/drivers/net/ethernet/dec/tulip/de2104x.c
+++ b/drivers/net/ethernet/dec/tulip/de2104x.c
@@ -2110,7 +2110,6 @@ static void de_remove_one(struct pci_dev *pdev)
2110 iounmap(de->regs); 2110 iounmap(de->regs);
2111 pci_release_regions(pdev); 2111 pci_release_regions(pdev);
2112 pci_disable_device(pdev); 2112 pci_disable_device(pdev);
2113 pci_set_drvdata(pdev, NULL);
2114 free_netdev(dev); 2113 free_netdev(dev);
2115} 2114}
2116 2115
diff --git a/drivers/net/ethernet/dec/tulip/de4x5.c b/drivers/net/ethernet/dec/tulip/de4x5.c
index 263b92c00cbf..c05b66dfcc30 100644
--- a/drivers/net/ethernet/dec/tulip/de4x5.c
+++ b/drivers/net/ethernet/dec/tulip/de4x5.c
@@ -2328,7 +2328,7 @@ static void de4x5_pci_remove(struct pci_dev *pdev)
2328 pci_disable_device (pdev); 2328 pci_disable_device (pdev);
2329} 2329}
2330 2330
2331static struct pci_device_id de4x5_pci_tbl[] = { 2331static DEFINE_PCI_DEVICE_TABLE(de4x5_pci_tbl) = {
2332 { PCI_VENDOR_ID_DEC, PCI_DEVICE_ID_DEC_TULIP, 2332 { PCI_VENDOR_ID_DEC, PCI_DEVICE_ID_DEC_TULIP,
2333 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0 }, 2333 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0 },
2334 { PCI_VENDOR_ID_DEC, PCI_DEVICE_ID_DEC_TULIP_PLUS, 2334 { PCI_VENDOR_ID_DEC, PCI_DEVICE_ID_DEC_TULIP_PLUS,
diff --git a/drivers/net/ethernet/dec/tulip/dmfe.c b/drivers/net/ethernet/dec/tulip/dmfe.c
index 83139307861c..5ad9e3e3c0b8 100644
--- a/drivers/net/ethernet/dec/tulip/dmfe.c
+++ b/drivers/net/ethernet/dec/tulip/dmfe.c
@@ -523,7 +523,6 @@ err_out_res:
523err_out_disable: 523err_out_disable:
524 pci_disable_device(pdev); 524 pci_disable_device(pdev);
525err_out_free: 525err_out_free:
526 pci_set_drvdata(pdev, NULL);
527 free_netdev(dev); 526 free_netdev(dev);
528 527
529 return err; 528 return err;
@@ -548,8 +547,6 @@ static void dmfe_remove_one(struct pci_dev *pdev)
548 db->buf_pool_ptr, db->buf_pool_dma_ptr); 547 db->buf_pool_ptr, db->buf_pool_dma_ptr);
549 pci_release_regions(pdev); 548 pci_release_regions(pdev);
550 free_netdev(dev); /* free board information */ 549 free_netdev(dev); /* free board information */
551
552 pci_set_drvdata(pdev, NULL);
553 } 550 }
554 551
555 DMFE_DBUG(0, "dmfe_remove_one() exit", 0); 552 DMFE_DBUG(0, "dmfe_remove_one() exit", 0);
diff --git a/drivers/net/ethernet/dec/tulip/tulip_core.c b/drivers/net/ethernet/dec/tulip/tulip_core.c
index 4e8cfa2ac803..add05f14b38b 100644
--- a/drivers/net/ethernet/dec/tulip/tulip_core.c
+++ b/drivers/net/ethernet/dec/tulip/tulip_core.c
@@ -1939,7 +1939,6 @@ static void tulip_remove_one(struct pci_dev *pdev)
1939 pci_iounmap(pdev, tp->base_addr); 1939 pci_iounmap(pdev, tp->base_addr);
1940 free_netdev (dev); 1940 free_netdev (dev);
1941 pci_release_regions (pdev); 1941 pci_release_regions (pdev);
1942 pci_set_drvdata (pdev, NULL);
1943 1942
1944 /* pci_power_off (pdev, -1); */ 1943 /* pci_power_off (pdev, -1); */
1945} 1944}
diff --git a/drivers/net/ethernet/dec/tulip/uli526x.c b/drivers/net/ethernet/dec/tulip/uli526x.c
index 93845afe1cea..a5397b130724 100644
--- a/drivers/net/ethernet/dec/tulip/uli526x.c
+++ b/drivers/net/ethernet/dec/tulip/uli526x.c
@@ -429,7 +429,6 @@ err_out_release:
429err_out_disable: 429err_out_disable:
430 pci_disable_device(pdev); 430 pci_disable_device(pdev);
431err_out_free: 431err_out_free:
432 pci_set_drvdata(pdev, NULL);
433 free_netdev(dev); 432 free_netdev(dev);
434 433
435 return err; 434 return err;
@@ -450,7 +449,6 @@ static void uli526x_remove_one(struct pci_dev *pdev)
450 db->buf_pool_ptr, db->buf_pool_dma_ptr); 449 db->buf_pool_ptr, db->buf_pool_dma_ptr);
451 pci_release_regions(pdev); 450 pci_release_regions(pdev);
452 pci_disable_device(pdev); 451 pci_disable_device(pdev);
453 pci_set_drvdata(pdev, NULL);
454 free_netdev(dev); 452 free_netdev(dev);
455} 453}
456 454
diff --git a/drivers/net/ethernet/dec/tulip/winbond-840.c b/drivers/net/ethernet/dec/tulip/winbond-840.c
index c7b04ecf5b49..62fe512bb216 100644
--- a/drivers/net/ethernet/dec/tulip/winbond-840.c
+++ b/drivers/net/ethernet/dec/tulip/winbond-840.c
@@ -468,7 +468,6 @@ static int w840_probe1(struct pci_dev *pdev, const struct pci_device_id *ent)
468 return 0; 468 return 0;
469 469
470err_out_cleardev: 470err_out_cleardev:
471 pci_set_drvdata(pdev, NULL);
472 pci_iounmap(pdev, ioaddr); 471 pci_iounmap(pdev, ioaddr);
473err_out_free_res: 472err_out_free_res:
474 pci_release_regions(pdev); 473 pci_release_regions(pdev);
@@ -1542,8 +1541,6 @@ static void w840_remove1(struct pci_dev *pdev)
1542 pci_iounmap(pdev, np->base_addr); 1541 pci_iounmap(pdev, np->base_addr);
1543 free_netdev(dev); 1542 free_netdev(dev);
1544 } 1543 }
1545
1546 pci_set_drvdata(pdev, NULL);
1547} 1544}
1548 1545
1549#ifdef CONFIG_PM 1546#ifdef CONFIG_PM
diff --git a/drivers/net/ethernet/dec/tulip/xircom_cb.c b/drivers/net/ethernet/dec/tulip/xircom_cb.c
index 9b84cb04fe5f..ab7ebac6fbea 100644
--- a/drivers/net/ethernet/dec/tulip/xircom_cb.c
+++ b/drivers/net/ethernet/dec/tulip/xircom_cb.c
@@ -289,7 +289,6 @@ out:
289err_unmap: 289err_unmap:
290 pci_iounmap(pdev, private->ioaddr); 290 pci_iounmap(pdev, private->ioaddr);
291reg_fail: 291reg_fail:
292 pci_set_drvdata(pdev, NULL);
293 dma_free_coherent(d, 8192, private->tx_buffer, private->tx_dma_handle); 292 dma_free_coherent(d, 8192, private->tx_buffer, private->tx_dma_handle);
294tx_buf_fail: 293tx_buf_fail:
295 dma_free_coherent(d, 8192, private->rx_buffer, private->rx_dma_handle); 294 dma_free_coherent(d, 8192, private->rx_buffer, private->rx_dma_handle);
@@ -317,7 +316,6 @@ static void xircom_remove(struct pci_dev *pdev)
317 316
318 unregister_netdev(dev); 317 unregister_netdev(dev);
319 pci_iounmap(pdev, card->ioaddr); 318 pci_iounmap(pdev, card->ioaddr);
320 pci_set_drvdata(pdev, NULL);
321 dma_free_coherent(d, 8192, card->tx_buffer, card->tx_dma_handle); 319 dma_free_coherent(d, 8192, card->tx_buffer, card->tx_dma_handle);
322 dma_free_coherent(d, 8192, card->rx_buffer, card->rx_dma_handle); 320 dma_free_coherent(d, 8192, card->rx_buffer, card->rx_dma_handle);
323 free_netdev(dev); 321 free_netdev(dev);
diff --git a/drivers/net/ethernet/dlink/dl2k.c b/drivers/net/ethernet/dlink/dl2k.c
index afa8e3af2c4d..4fb756d219f7 100644
--- a/drivers/net/ethernet/dlink/dl2k.c
+++ b/drivers/net/ethernet/dlink/dl2k.c
@@ -1746,7 +1746,6 @@ rio_remove1 (struct pci_dev *pdev)
1746 pci_release_regions (pdev); 1746 pci_release_regions (pdev);
1747 pci_disable_device (pdev); 1747 pci_disable_device (pdev);
1748 } 1748 }
1749 pci_set_drvdata (pdev, NULL);
1750} 1749}
1751 1750
1752static struct pci_driver rio_driver = { 1751static struct pci_driver rio_driver = {
diff --git a/drivers/net/ethernet/dlink/sundance.c b/drivers/net/ethernet/dlink/sundance.c
index bf3bf6f22c99..113cd799a131 100644
--- a/drivers/net/ethernet/dlink/sundance.c
+++ b/drivers/net/ethernet/dlink/sundance.c
@@ -703,7 +703,6 @@ err_out_unmap_tx:
703 dma_free_coherent(&pdev->dev, TX_TOTAL_SIZE, 703 dma_free_coherent(&pdev->dev, TX_TOTAL_SIZE,
704 np->tx_ring, np->tx_ring_dma); 704 np->tx_ring, np->tx_ring_dma);
705err_out_cleardev: 705err_out_cleardev:
706 pci_set_drvdata(pdev, NULL);
707 pci_iounmap(pdev, ioaddr); 706 pci_iounmap(pdev, ioaddr);
708err_out_res: 707err_out_res:
709 pci_release_regions(pdev); 708 pci_release_regions(pdev);
@@ -1941,7 +1940,6 @@ static void sundance_remove1(struct pci_dev *pdev)
1941 pci_iounmap(pdev, np->base); 1940 pci_iounmap(pdev, np->base);
1942 pci_release_regions(pdev); 1941 pci_release_regions(pdev);
1943 free_netdev(dev); 1942 free_netdev(dev);
1944 pci_set_drvdata(pdev, NULL);
1945 } 1943 }
1946} 1944}
1947 1945
diff --git a/drivers/net/ethernet/emulex/benet/be.h b/drivers/net/ethernet/emulex/benet/be.h
index c99dac6a9ddf..f4825db5d179 100644
--- a/drivers/net/ethernet/emulex/benet/be.h
+++ b/drivers/net/ethernet/emulex/benet/be.h
@@ -34,7 +34,7 @@
34#include "be_hw.h" 34#include "be_hw.h"
35#include "be_roce.h" 35#include "be_roce.h"
36 36
37#define DRV_VER "4.9.134.0u" 37#define DRV_VER "4.9.224.0u"
38#define DRV_NAME "be2net" 38#define DRV_NAME "be2net"
39#define BE_NAME "Emulex BladeEngine2" 39#define BE_NAME "Emulex BladeEngine2"
40#define BE3_NAME "Emulex BladeEngine3" 40#define BE3_NAME "Emulex BladeEngine3"
@@ -89,7 +89,7 @@ static inline char *nic_name(struct pci_dev *pdev)
89 89
90#define BE_NUM_VLANS_SUPPORTED 64 90#define BE_NUM_VLANS_SUPPORTED 64
91#define BE_UMC_NUM_VLANS_SUPPORTED 15 91#define BE_UMC_NUM_VLANS_SUPPORTED 15
92#define BE_MAX_EQD 96u 92#define BE_MAX_EQD 128u
93#define BE_MAX_TX_FRAG_COUNT 30 93#define BE_MAX_TX_FRAG_COUNT 30
94 94
95#define EVNT_Q_LEN 1024 95#define EVNT_Q_LEN 1024
@@ -199,8 +199,37 @@ struct be_eq_obj {
199 u16 spurious_intr; 199 u16 spurious_intr;
200 struct napi_struct napi; 200 struct napi_struct napi;
201 struct be_adapter *adapter; 201 struct be_adapter *adapter;
202
203#ifdef CONFIG_NET_RX_BUSY_POLL
204#define BE_EQ_IDLE 0
205#define BE_EQ_NAPI 1 /* napi owns this EQ */
206#define BE_EQ_POLL 2 /* poll owns this EQ */
207#define BE_EQ_LOCKED (BE_EQ_NAPI | BE_EQ_POLL)
208#define BE_EQ_NAPI_YIELD 4 /* napi yielded this EQ */
209#define BE_EQ_POLL_YIELD 8 /* poll yielded this EQ */
210#define BE_EQ_YIELD (BE_EQ_NAPI_YIELD | BE_EQ_POLL_YIELD)
211#define BE_EQ_USER_PEND (BE_EQ_POLL | BE_EQ_POLL_YIELD)
212 unsigned int state;
213 spinlock_t lock; /* lock to serialize napi and busy-poll */
214#endif /* CONFIG_NET_RX_BUSY_POLL */
202} ____cacheline_aligned_in_smp; 215} ____cacheline_aligned_in_smp;
203 216
217struct be_aic_obj { /* Adaptive interrupt coalescing (AIC) info */
218 bool enable;
219 u32 min_eqd; /* in usecs */
220 u32 max_eqd; /* in usecs */
221 u32 prev_eqd; /* in usecs */
222 u32 et_eqd; /* configured val when aic is off */
223 ulong jiffies;
224 u64 rx_pkts_prev; /* Used to calculate RX pps */
225 u64 tx_reqs_prev; /* Used to calculate TX pps */
226};
227
228enum {
229 NAPI_POLLING,
230 BUSY_POLLING
231};
232
204struct be_mcc_obj { 233struct be_mcc_obj {
205 struct be_queue_info q; 234 struct be_queue_info q;
206 struct be_queue_info cq; 235 struct be_queue_info cq;
@@ -215,6 +244,7 @@ struct be_tx_stats {
215 u64 tx_compl; 244 u64 tx_compl;
216 ulong tx_jiffies; 245 ulong tx_jiffies;
217 u32 tx_stops; 246 u32 tx_stops;
247 u32 tx_drv_drops; /* pkts dropped by driver */
218 struct u64_stats_sync sync; 248 struct u64_stats_sync sync;
219 struct u64_stats_sync sync_compl; 249 struct u64_stats_sync sync_compl;
220}; 250};
@@ -239,15 +269,12 @@ struct be_rx_page_info {
239struct be_rx_stats { 269struct be_rx_stats {
240 u64 rx_bytes; 270 u64 rx_bytes;
241 u64 rx_pkts; 271 u64 rx_pkts;
242 u64 rx_pkts_prev;
243 ulong rx_jiffies;
244 u32 rx_drops_no_skbs; /* skb allocation errors */ 272 u32 rx_drops_no_skbs; /* skb allocation errors */
245 u32 rx_drops_no_frags; /* HW has no fetched frags */ 273 u32 rx_drops_no_frags; /* HW has no fetched frags */
246 u32 rx_post_fail; /* page post alloc failures */ 274 u32 rx_post_fail; /* page post alloc failures */
247 u32 rx_compl; 275 u32 rx_compl;
248 u32 rx_mcast_pkts; 276 u32 rx_mcast_pkts;
249 u32 rx_compl_err; /* completions with err set */ 277 u32 rx_compl_err; /* completions with err set */
250 u32 rx_pps; /* pkts per second */
251 struct u64_stats_sync sync; 278 struct u64_stats_sync sync;
252}; 279};
253 280
@@ -316,6 +343,11 @@ struct be_drv_stats {
316 u32 rx_input_fifo_overflow_drop; 343 u32 rx_input_fifo_overflow_drop;
317 u32 pmem_fifo_overflow_drop; 344 u32 pmem_fifo_overflow_drop;
318 u32 jabber_events; 345 u32 jabber_events;
346 u32 rx_roce_bytes_lsd;
347 u32 rx_roce_bytes_msd;
348 u32 rx_roce_frames;
349 u32 roce_drops_payload_len;
350 u32 roce_drops_crc;
319}; 351};
320 352
321struct be_vf_cfg { 353struct be_vf_cfg {
@@ -405,6 +437,7 @@ struct be_adapter {
405 u32 big_page_size; /* Compounded page size shared by rx wrbs */ 437 u32 big_page_size; /* Compounded page size shared by rx wrbs */
406 438
407 struct be_drv_stats drv_stats; 439 struct be_drv_stats drv_stats;
440 struct be_aic_obj aic_obj[MAX_EVT_QS];
408 u16 vlans_added; 441 u16 vlans_added;
409 u8 vlan_tag[VLAN_N_VID]; 442 u8 vlan_tag[VLAN_N_VID];
410 u8 vlan_prio_bmap; /* Available Priority BitMap */ 443 u8 vlan_prio_bmap; /* Available Priority BitMap */
@@ -437,7 +470,6 @@ struct be_adapter {
437 u32 rx_fc; /* Rx flow control */ 470 u32 rx_fc; /* Rx flow control */
438 u32 tx_fc; /* Tx flow control */ 471 u32 tx_fc; /* Tx flow control */
439 bool stats_cmd_sent; 472 bool stats_cmd_sent;
440 u32 if_type;
441 struct { 473 struct {
442 u32 size; 474 u32 size;
443 u32 total_size; 475 u32 total_size;
@@ -472,8 +504,8 @@ struct be_adapter {
472 504
473#define be_physfn(adapter) (!adapter->virtfn) 505#define be_physfn(adapter) (!adapter->virtfn)
474#define sriov_enabled(adapter) (adapter->num_vfs > 0) 506#define sriov_enabled(adapter) (adapter->num_vfs > 0)
475#define sriov_want(adapter) (be_max_vfs(adapter) && num_vfs && \ 507#define sriov_want(adapter) (be_physfn(adapter) && \
476 be_physfn(adapter)) 508 (num_vfs || pci_num_vf(adapter->pdev)))
477#define for_all_vfs(adapter, vf_cfg, i) \ 509#define for_all_vfs(adapter, vf_cfg, i) \
478 for (i = 0, vf_cfg = &adapter->vf_cfg[i]; i < adapter->num_vfs; \ 510 for (i = 0, vf_cfg = &adapter->vf_cfg[i]; i < adapter->num_vfs; \
479 i++, vf_cfg++) 511 i++, vf_cfg++)
@@ -546,6 +578,10 @@ extern const struct ethtool_ops be_ethtool_ops;
546 for (i = 0, eqo = &adapter->eq_obj[i]; i < adapter->num_evt_qs; \ 578 for (i = 0, eqo = &adapter->eq_obj[i]; i < adapter->num_evt_qs; \
547 i++, eqo++) 579 i++, eqo++)
548 580
581#define for_all_rx_queues_on_eq(adapter, eqo, rxo, i) \
582 for (i = eqo->idx, rxo = &adapter->rx_obj[i]; i < adapter->num_rx_qs;\
583 i += adapter->num_evt_qs, rxo += adapter->num_evt_qs)
584
549#define is_mcc_eqo(eqo) (eqo->idx == 0) 585#define is_mcc_eqo(eqo) (eqo->idx == 0)
550#define mcc_eqo(adapter) (&adapter->eq_obj[0]) 586#define mcc_eqo(adapter) (&adapter->eq_obj[0])
551 587
@@ -696,6 +732,115 @@ static inline int qnq_async_evt_rcvd(struct be_adapter *adapter)
696 return adapter->flags & BE_FLAGS_QNQ_ASYNC_EVT_RCVD; 732 return adapter->flags & BE_FLAGS_QNQ_ASYNC_EVT_RCVD;
697} 733}
698 734
735#ifdef CONFIG_NET_RX_BUSY_POLL
736static inline bool be_lock_napi(struct be_eq_obj *eqo)
737{
738 bool status = true;
739
740 spin_lock(&eqo->lock); /* BH is already disabled */
741 if (eqo->state & BE_EQ_LOCKED) {
742 WARN_ON(eqo->state & BE_EQ_NAPI);
743 eqo->state |= BE_EQ_NAPI_YIELD;
744 status = false;
745 } else {
746 eqo->state = BE_EQ_NAPI;
747 }
748 spin_unlock(&eqo->lock);
749 return status;
750}
751
752static inline void be_unlock_napi(struct be_eq_obj *eqo)
753{
754 spin_lock(&eqo->lock); /* BH is already disabled */
755
756 WARN_ON(eqo->state & (BE_EQ_POLL | BE_EQ_NAPI_YIELD));
757 eqo->state = BE_EQ_IDLE;
758
759 spin_unlock(&eqo->lock);
760}
761
762static inline bool be_lock_busy_poll(struct be_eq_obj *eqo)
763{
764 bool status = true;
765
766 spin_lock_bh(&eqo->lock);
767 if (eqo->state & BE_EQ_LOCKED) {
768 eqo->state |= BE_EQ_POLL_YIELD;
769 status = false;
770 } else {
771 eqo->state |= BE_EQ_POLL;
772 }
773 spin_unlock_bh(&eqo->lock);
774 return status;
775}
776
777static inline void be_unlock_busy_poll(struct be_eq_obj *eqo)
778{
779 spin_lock_bh(&eqo->lock);
780
781 WARN_ON(eqo->state & (BE_EQ_NAPI));
782 eqo->state = BE_EQ_IDLE;
783
784 spin_unlock_bh(&eqo->lock);
785}
786
787static inline void be_enable_busy_poll(struct be_eq_obj *eqo)
788{
789 spin_lock_init(&eqo->lock);
790 eqo->state = BE_EQ_IDLE;
791}
792
793static inline void be_disable_busy_poll(struct be_eq_obj *eqo)
794{
795 local_bh_disable();
796
797 /* It's enough to just acquire napi lock on the eqo to stop
798 * be_busy_poll() from processing any queueus.
799 */
800 while (!be_lock_napi(eqo))
801 mdelay(1);
802
803 local_bh_enable();
804}
805
806#else /* CONFIG_NET_RX_BUSY_POLL */
807
808static inline bool be_lock_napi(struct be_eq_obj *eqo)
809{
810 return true;
811}
812
813static inline void be_unlock_napi(struct be_eq_obj *eqo)
814{
815}
816
817static inline bool be_lock_busy_poll(struct be_eq_obj *eqo)
818{
819 return false;
820}
821
822static inline void be_unlock_busy_poll(struct be_eq_obj *eqo)
823{
824}
825
826static inline void be_enable_busy_poll(struct be_eq_obj *eqo)
827{
828}
829
830static inline void be_disable_busy_poll(struct be_eq_obj *eqo)
831{
832}
833#endif /* CONFIG_NET_RX_BUSY_POLL */
834
835void be_cq_notify(struct be_adapter *adapter, u16 qid, bool arm,
836 u16 num_popped);
837void be_link_status_update(struct be_adapter *adapter, u8 link_status);
838void be_parse_stats(struct be_adapter *adapter);
839int be_load_fw(struct be_adapter *adapter, u8 *func);
840bool be_is_wol_supported(struct be_adapter *adapter);
841bool be_pause_supported(struct be_adapter *adapter);
842u32 be_get_fw_log_level(struct be_adapter *adapter);
843
699static inline int fw_major_num(const char *fw_ver) 844static inline int fw_major_num(const char *fw_ver)
700{ 845{
701 int fw_major = 0; 846 int fw_major = 0;
@@ -705,27 +850,19 @@ static inline int fw_major_num(const char *fw_ver)
705 return fw_major; 850 return fw_major;
706} 851}
707 852
708extern void be_cq_notify(struct be_adapter *adapter, u16 qid, bool arm,
709 u16 num_popped);
710extern void be_link_status_update(struct be_adapter *adapter, u8 link_status);
711extern void be_parse_stats(struct be_adapter *adapter);
712extern int be_load_fw(struct be_adapter *adapter, u8 *func);
713extern bool be_is_wol_supported(struct be_adapter *adapter);
714extern bool be_pause_supported(struct be_adapter *adapter);
715extern u32 be_get_fw_log_level(struct be_adapter *adapter);
716int be_update_queues(struct be_adapter *adapter); 853int be_update_queues(struct be_adapter *adapter);
717int be_poll(struct napi_struct *napi, int budget); 854int be_poll(struct napi_struct *napi, int budget);
718 855
719/* 856/*
720 * internal function to initialize-cleanup roce device. 857 * internal function to initialize-cleanup roce device.
721 */ 858 */
722extern void be_roce_dev_add(struct be_adapter *); 859void be_roce_dev_add(struct be_adapter *);
723extern void be_roce_dev_remove(struct be_adapter *); 860void be_roce_dev_remove(struct be_adapter *);
724 861
725/* 862/*
726 * internal function to open-close roce device during ifup-ifdown. 863 * internal function to open-close roce device during ifup-ifdown.
727 */ 864 */
728extern void be_roce_dev_open(struct be_adapter *); 865void be_roce_dev_open(struct be_adapter *);
729extern void be_roce_dev_close(struct be_adapter *); 866void be_roce_dev_close(struct be_adapter *);
730 867
731#endif /* BE_H */ 868#endif /* BE_H */
diff --git a/drivers/net/ethernet/emulex/benet/be_cmds.c b/drivers/net/ethernet/emulex/benet/be_cmds.c
index c08fd32bb8e5..7fb0edfe3d24 100644
--- a/drivers/net/ethernet/emulex/benet/be_cmds.c
+++ b/drivers/net/ethernet/emulex/benet/be_cmds.c
@@ -522,7 +522,7 @@ static u16 be_POST_stage_get(struct be_adapter *adapter)
522 return sem & POST_STAGE_MASK; 522 return sem & POST_STAGE_MASK;
523} 523}
524 524
525int lancer_wait_ready(struct be_adapter *adapter) 525static int lancer_wait_ready(struct be_adapter *adapter)
526{ 526{
527#define SLIPORT_READY_TIMEOUT 30 527#define SLIPORT_READY_TIMEOUT 30
528 u32 sliport_status; 528 u32 sliport_status;
@@ -1436,8 +1436,12 @@ int be_cmd_get_stats(struct be_adapter *adapter, struct be_dma_mem *nonemb_cmd)
1436 OPCODE_ETH_GET_STATISTICS, nonemb_cmd->size, wrb, nonemb_cmd); 1436 OPCODE_ETH_GET_STATISTICS, nonemb_cmd->size, wrb, nonemb_cmd);
1437 1437
1438 /* version 1 of the cmd is not supported only by BE2 */ 1438 /* version 1 of the cmd is not supported only by BE2 */
1439 if (!BE2_chip(adapter)) 1439 if (BE2_chip(adapter))
1440 hdr->version = 0;
1441 if (BE3_chip(adapter) || lancer_chip(adapter))
1440 hdr->version = 1; 1442 hdr->version = 1;
1443 else
1444 hdr->version = 2;
1441 1445
1442 be_mcc_notify(adapter); 1446 be_mcc_notify(adapter);
1443 adapter->stats_cmd_sent = true; 1447 adapter->stats_cmd_sent = true;
@@ -1719,11 +1723,12 @@ err:
1719/* set the EQ delay interval of an EQ to specified value 1723/* set the EQ delay interval of an EQ to specified value
1720 * Uses async mcc 1724 * Uses async mcc
1721 */ 1725 */
1722int be_cmd_modify_eqd(struct be_adapter *adapter, u32 eq_id, u32 eqd) 1726int be_cmd_modify_eqd(struct be_adapter *adapter, struct be_set_eqd *set_eqd,
1727 int num)
1723{ 1728{
1724 struct be_mcc_wrb *wrb; 1729 struct be_mcc_wrb *wrb;
1725 struct be_cmd_req_modify_eq_delay *req; 1730 struct be_cmd_req_modify_eq_delay *req;
1726 int status = 0; 1731 int status = 0, i;
1727 1732
1728 spin_lock_bh(&adapter->mcc_lock); 1733 spin_lock_bh(&adapter->mcc_lock);
1729 1734
@@ -1737,13 +1742,15 @@ int be_cmd_modify_eqd(struct be_adapter *adapter, u32 eq_id, u32 eqd)
1737 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON, 1742 be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1738 OPCODE_COMMON_MODIFY_EQ_DELAY, sizeof(*req), wrb, NULL); 1743 OPCODE_COMMON_MODIFY_EQ_DELAY, sizeof(*req), wrb, NULL);
1739 1744
1740 req->num_eq = cpu_to_le32(1); 1745 req->num_eq = cpu_to_le32(num);
1741 req->delay[0].eq_id = cpu_to_le32(eq_id); 1746 for (i = 0; i < num; i++) {
1742 req->delay[0].phase = 0; 1747 req->set_eqd[i].eq_id = cpu_to_le32(set_eqd[i].eq_id);
1743 req->delay[0].delay_multiplier = cpu_to_le32(eqd); 1748 req->set_eqd[i].phase = 0;
1749 req->set_eqd[i].delay_multiplier =
1750 cpu_to_le32(set_eqd[i].delay_multiplier);
1751 }
1744 1752
1745 be_mcc_notify(adapter); 1753 be_mcc_notify(adapter);
1746
1747err: 1754err:
1748 spin_unlock_bh(&adapter->mcc_lock); 1755 spin_unlock_bh(&adapter->mcc_lock);
1749 return status; 1756 return status;
@@ -3520,7 +3527,7 @@ int be_cmd_enable_vf(struct be_adapter *adapter, u8 domain)
3520 struct be_cmd_enable_disable_vf *req; 3527 struct be_cmd_enable_disable_vf *req;
3521 int status; 3528 int status;
3522 3529
3523 if (!lancer_chip(adapter)) 3530 if (BEx_chip(adapter))
3524 return 0; 3531 return 0;
3525 3532
3526 spin_lock_bh(&adapter->mcc_lock); 3533 spin_lock_bh(&adapter->mcc_lock);
diff --git a/drivers/net/ethernet/emulex/benet/be_cmds.h b/drivers/net/ethernet/emulex/benet/be_cmds.h
index 108ca8abf0af..edf3e8a0ff83 100644
--- a/drivers/net/ethernet/emulex/benet/be_cmds.h
+++ b/drivers/net/ethernet/emulex/benet/be_cmds.h
@@ -1057,14 +1057,16 @@ struct be_cmd_resp_get_flow_control {
1057} __packed; 1057} __packed;
1058 1058
1059/******************** Modify EQ Delay *******************/ 1059/******************** Modify EQ Delay *******************/
1060struct be_set_eqd {
1061 u32 eq_id;
1062 u32 phase;
1063 u32 delay_multiplier;
1064};
1065
1060struct be_cmd_req_modify_eq_delay { 1066struct be_cmd_req_modify_eq_delay {
1061 struct be_cmd_req_hdr hdr; 1067 struct be_cmd_req_hdr hdr;
1062 u32 num_eq; 1068 u32 num_eq;
1063 struct { 1069 struct be_set_eqd set_eqd[MAX_EVT_QS];
1064 u32 eq_id;
1065 u32 phase;
1066 u32 delay_multiplier;
1067 } delay[8];
1068} __packed; 1070} __packed;
1069 1071
1070struct be_cmd_resp_modify_eq_delay { 1072struct be_cmd_resp_modify_eq_delay {
@@ -1660,6 +1662,67 @@ struct be_erx_stats_v1 {
1660 u32 rsvd[4]; 1662 u32 rsvd[4];
1661}; 1663};
1662 1664
1665struct be_port_rxf_stats_v2 {
1666 u32 rsvd0[10];
1667 u32 roce_bytes_received_lsd;
1668 u32 roce_bytes_received_msd;
1669 u32 rsvd1[5];
1670 u32 roce_frames_received;
1671 u32 rx_crc_errors;
1672 u32 rx_alignment_symbol_errors;
1673 u32 rx_pause_frames;
1674 u32 rx_priority_pause_frames;
1675 u32 rx_control_frames;
1676 u32 rx_in_range_errors;
1677 u32 rx_out_range_errors;
1678 u32 rx_frame_too_long;
1679 u32 rx_address_filtered;
1680 u32 rx_dropped_too_small;
1681 u32 rx_dropped_too_short;
1682 u32 rx_dropped_header_too_small;
1683 u32 rx_dropped_tcp_length;
1684 u32 rx_dropped_runt;
1685 u32 rsvd2[10];
1686 u32 rx_ip_checksum_errs;
1687 u32 rx_tcp_checksum_errs;
1688 u32 rx_udp_checksum_errs;
1689 u32 rsvd3[7];
1690 u32 rx_switched_unicast_packets;
1691 u32 rx_switched_multicast_packets;
1692 u32 rx_switched_broadcast_packets;
1693 u32 rsvd4[3];
1694 u32 tx_pauseframes;
1695 u32 tx_priority_pauseframes;
1696 u32 tx_controlframes;
1697 u32 rsvd5[10];
1698 u32 rxpp_fifo_overflow_drop;
1699 u32 rx_input_fifo_overflow_drop;
1700 u32 pmem_fifo_overflow_drop;
1701 u32 jabber_events;
1702 u32 rsvd6[3];
1703 u32 rx_drops_payload_size;
1704 u32 rx_drops_clipped_header;
1705 u32 rx_drops_crc;
1706 u32 roce_drops_payload_len;
1707 u32 roce_drops_crc;
1708 u32 rsvd7[19];
1709};
1710
1711struct be_rxf_stats_v2 {
1712 struct be_port_rxf_stats_v2 port[4];
1713 u32 rsvd0[2];
1714 u32 rx_drops_no_pbuf;
1715 u32 rx_drops_no_txpb;
1716 u32 rx_drops_no_erx_descr;
1717 u32 rx_drops_no_tpre_descr;
1718 u32 rsvd1[6];
1719 u32 rx_drops_too_many_frags;
1720 u32 rx_drops_invalid_ring;
1721 u32 forwarded_packets;
1722 u32 rx_drops_mtu;
1723 u32 rsvd2[35];
1724};
1725
1663struct be_hw_stats_v1 { 1726struct be_hw_stats_v1 {
1664 struct be_rxf_stats_v1 rxf; 1727 struct be_rxf_stats_v1 rxf;
1665 u32 rsvd0[BE_TXP_SW_SZ]; 1728 u32 rsvd0[BE_TXP_SW_SZ];
@@ -1678,6 +1741,29 @@ struct be_cmd_resp_get_stats_v1 {
1678 struct be_hw_stats_v1 hw_stats; 1741 struct be_hw_stats_v1 hw_stats;
1679}; 1742};
1680 1743
1744struct be_erx_stats_v2 {
1745 u32 rx_drops_no_fragments[136]; /* dwordS 0 to 135*/
1746 u32 rsvd[3];
1747};
1748
1749struct be_hw_stats_v2 {
1750 struct be_rxf_stats_v2 rxf;
1751 u32 rsvd0[BE_TXP_SW_SZ];
1752 struct be_erx_stats_v2 erx;
1753 struct be_pmem_stats pmem;
1754 u32 rsvd1[18];
1755};
1756
1757struct be_cmd_req_get_stats_v2 {
1758 struct be_cmd_req_hdr hdr;
1759 u8 rsvd[sizeof(struct be_hw_stats_v2)];
1760};
1761
1762struct be_cmd_resp_get_stats_v2 {
1763 struct be_cmd_resp_hdr hdr;
1764 struct be_hw_stats_v2 hw_stats;
1765};
1766
1681/************** get fat capabilites *******************/ 1767/************** get fat capabilites *******************/
1682#define MAX_MODULES 27 1768#define MAX_MODULES 27
1683#define MAX_MODES 4 1769#define MAX_MODES 4
@@ -1865,137 +1951,119 @@ struct be_cmd_resp_get_iface_list {
1865 struct be_if_desc if_desc; 1951 struct be_if_desc if_desc;
1866}; 1952};
1867 1953
1868extern int be_pci_fnum_get(struct be_adapter *adapter); 1954int be_pci_fnum_get(struct be_adapter *adapter);
1869extern int be_fw_wait_ready(struct be_adapter *adapter); 1955int be_fw_wait_ready(struct be_adapter *adapter);
1870extern int be_cmd_mac_addr_query(struct be_adapter *adapter, u8 *mac_addr, 1956int be_cmd_mac_addr_query(struct be_adapter *adapter, u8 *mac_addr,
1871 bool permanent, u32 if_handle, u32 pmac_id); 1957 bool permanent, u32 if_handle, u32 pmac_id);
1872extern int be_cmd_pmac_add(struct be_adapter *adapter, u8 *mac_addr, 1958int be_cmd_pmac_add(struct be_adapter *adapter, u8 *mac_addr, u32 if_id,
1873 u32 if_id, u32 *pmac_id, u32 domain); 1959 u32 *pmac_id, u32 domain);
1874extern int be_cmd_pmac_del(struct be_adapter *adapter, u32 if_id, 1960int be_cmd_pmac_del(struct be_adapter *adapter, u32 if_id, int pmac_id,
1875 int pmac_id, u32 domain); 1961 u32 domain);
1876extern int be_cmd_if_create(struct be_adapter *adapter, u32 cap_flags, 1962int be_cmd_if_create(struct be_adapter *adapter, u32 cap_flags, u32 en_flags,
1877 u32 en_flags, u32 *if_handle, u32 domain); 1963 u32 *if_handle, u32 domain);
1878extern int be_cmd_if_destroy(struct be_adapter *adapter, int if_handle, 1964int be_cmd_if_destroy(struct be_adapter *adapter, int if_handle, u32 domain);
1879 u32 domain); 1965int be_cmd_eq_create(struct be_adapter *adapter, struct be_eq_obj *eqo);
1880extern int be_cmd_eq_create(struct be_adapter *adapter, struct be_eq_obj *eqo); 1966int be_cmd_cq_create(struct be_adapter *adapter, struct be_queue_info *cq,
1881extern int be_cmd_cq_create(struct be_adapter *adapter, 1967 struct be_queue_info *eq, bool no_delay,
1882 struct be_queue_info *cq, struct be_queue_info *eq, 1968 int num_cqe_dma_coalesce);
1883 bool no_delay, int num_cqe_dma_coalesce); 1969int be_cmd_mccq_create(struct be_adapter *adapter, struct be_queue_info *mccq,
1884extern int be_cmd_mccq_create(struct be_adapter *adapter, 1970 struct be_queue_info *cq);
1885 struct be_queue_info *mccq, 1971int be_cmd_txq_create(struct be_adapter *adapter, struct be_tx_obj *txo);
1886 struct be_queue_info *cq); 1972int be_cmd_rxq_create(struct be_adapter *adapter, struct be_queue_info *rxq,
1887extern int be_cmd_txq_create(struct be_adapter *adapter, 1973 u16 cq_id, u16 frag_size, u32 if_id, u32 rss, u8 *rss_id);
1888 struct be_tx_obj *txo); 1974int be_cmd_q_destroy(struct be_adapter *adapter, struct be_queue_info *q,
1889extern int be_cmd_rxq_create(struct be_adapter *adapter, 1975 int type);
1890 struct be_queue_info *rxq, u16 cq_id, 1976int be_cmd_rxq_destroy(struct be_adapter *adapter, struct be_queue_info *q);
1891 u16 frag_size, u32 if_id, u32 rss, u8 *rss_id); 1977int be_cmd_link_status_query(struct be_adapter *adapter, u16 *link_speed,
1892extern int be_cmd_q_destroy(struct be_adapter *adapter, struct be_queue_info *q, 1978 u8 *link_status, u32 dom);
1893 int type); 1979int be_cmd_reset(struct be_adapter *adapter);
1894extern int be_cmd_rxq_destroy(struct be_adapter *adapter, 1980int be_cmd_get_stats(struct be_adapter *adapter, struct be_dma_mem *nonemb_cmd);
1895 struct be_queue_info *q); 1981int lancer_cmd_get_pport_stats(struct be_adapter *adapter,
1896extern int be_cmd_link_status_query(struct be_adapter *adapter, u16 *link_speed, 1982 struct be_dma_mem *nonemb_cmd);
1897 u8 *link_status, u32 dom); 1983int be_cmd_get_fw_ver(struct be_adapter *adapter, char *fw_ver,
1898extern int be_cmd_reset(struct be_adapter *adapter); 1984 char *fw_on_flash);
1899extern int be_cmd_get_stats(struct be_adapter *adapter, 1985int be_cmd_modify_eqd(struct be_adapter *adapter, struct be_set_eqd *, int num);
1900 struct be_dma_mem *nonemb_cmd); 1986int be_cmd_vlan_config(struct be_adapter *adapter, u32 if_id, u16 *vtag_array,
1901extern int lancer_cmd_get_pport_stats(struct be_adapter *adapter, 1987 u32 num, bool untagged, bool promiscuous);
1902 struct be_dma_mem *nonemb_cmd); 1988int be_cmd_rx_filter(struct be_adapter *adapter, u32 flags, u32 status);
1903extern int be_cmd_get_fw_ver(struct be_adapter *adapter, char *fw_ver, 1989int be_cmd_set_flow_control(struct be_adapter *adapter, u32 tx_fc, u32 rx_fc);
1904 char *fw_on_flash); 1990int be_cmd_get_flow_control(struct be_adapter *adapter, u32 *tx_fc, u32 *rx_fc);
1905 1991int be_cmd_query_fw_cfg(struct be_adapter *adapter, u32 *port_num,
1906extern int be_cmd_modify_eqd(struct be_adapter *adapter, u32 eq_id, u32 eqd);
1907extern int be_cmd_vlan_config(struct be_adapter *adapter, u32 if_id,
1908 u16 *vtag_array, u32 num, bool untagged,
1909 bool promiscuous);
1910extern int be_cmd_rx_filter(struct be_adapter *adapter, u32 flags, u32 status);
1911extern int be_cmd_set_flow_control(struct be_adapter *adapter,
1912 u32 tx_fc, u32 rx_fc);
1913extern int be_cmd_get_flow_control(struct be_adapter *adapter,
1914 u32 *tx_fc, u32 *rx_fc);
1915extern int be_cmd_query_fw_cfg(struct be_adapter *adapter, u32 *port_num,
1916 u32 *function_mode, u32 *function_caps, u16 *asic_rev); 1992 u32 *function_mode, u32 *function_caps, u16 *asic_rev);
1917extern int be_cmd_reset_function(struct be_adapter *adapter); 1993int be_cmd_reset_function(struct be_adapter *adapter);
1918extern int be_cmd_rss_config(struct be_adapter *adapter, u8 *rsstable, 1994int be_cmd_rss_config(struct be_adapter *adapter, u8 *rsstable,
1919 u32 rss_hash_opts, u16 table_size); 1995 u32 rss_hash_opts, u16 table_size);
1920extern int be_process_mcc(struct be_adapter *adapter); 1996int be_process_mcc(struct be_adapter *adapter);
1921extern int be_cmd_set_beacon_state(struct be_adapter *adapter, 1997int be_cmd_set_beacon_state(struct be_adapter *adapter, u8 port_num, u8 beacon,
1922 u8 port_num, u8 beacon, u8 status, u8 state); 1998 u8 status, u8 state);
1923extern int be_cmd_get_beacon_state(struct be_adapter *adapter, 1999int be_cmd_get_beacon_state(struct be_adapter *adapter, u8 port_num,
1924 u8 port_num, u32 *state); 2000 u32 *state);
1925extern int be_cmd_write_flashrom(struct be_adapter *adapter, 2001int be_cmd_write_flashrom(struct be_adapter *adapter, struct be_dma_mem *cmd,
1926 struct be_dma_mem *cmd, u32 flash_oper, 2002 u32 flash_oper, u32 flash_opcode, u32 buf_size);
1927 u32 flash_opcode, u32 buf_size); 2003int lancer_cmd_write_object(struct be_adapter *adapter, struct be_dma_mem *cmd,
1928extern int lancer_cmd_write_object(struct be_adapter *adapter, 2004 u32 data_size, u32 data_offset,
1929 struct be_dma_mem *cmd, 2005 const char *obj_name, u32 *data_written,
1930 u32 data_size, u32 data_offset, 2006 u8 *change_status, u8 *addn_status);
1931 const char *obj_name,
1932 u32 *data_written, u8 *change_status,
1933 u8 *addn_status);
1934int lancer_cmd_read_object(struct be_adapter *adapter, struct be_dma_mem *cmd, 2007int lancer_cmd_read_object(struct be_adapter *adapter, struct be_dma_mem *cmd,
1935 u32 data_size, u32 data_offset, const char *obj_name, 2008 u32 data_size, u32 data_offset, const char *obj_name,
1936 u32 *data_read, u32 *eof, u8 *addn_status); 2009 u32 *data_read, u32 *eof, u8 *addn_status);
1937int be_cmd_get_flash_crc(struct be_adapter *adapter, u8 *flashed_crc, 2010int be_cmd_get_flash_crc(struct be_adapter *adapter, u8 *flashed_crc,
1938 int offset); 2011 int offset);
1939extern int be_cmd_enable_magic_wol(struct be_adapter *adapter, u8 *mac, 2012int be_cmd_enable_magic_wol(struct be_adapter *adapter, u8 *mac,
1940 struct be_dma_mem *nonemb_cmd); 2013 struct be_dma_mem *nonemb_cmd);
1941extern int be_cmd_fw_init(struct be_adapter *adapter); 2014int be_cmd_fw_init(struct be_adapter *adapter);
1942extern int be_cmd_fw_clean(struct be_adapter *adapter); 2015int be_cmd_fw_clean(struct be_adapter *adapter);
1943extern void be_async_mcc_enable(struct be_adapter *adapter); 2016void be_async_mcc_enable(struct be_adapter *adapter);
1944extern void be_async_mcc_disable(struct be_adapter *adapter); 2017void be_async_mcc_disable(struct be_adapter *adapter);
1945extern int be_cmd_loopback_test(struct be_adapter *adapter, u32 port_num, 2018int be_cmd_loopback_test(struct be_adapter *adapter, u32 port_num,
1946 u32 loopback_type, u32 pkt_size, 2019 u32 loopback_type, u32 pkt_size, u32 num_pkts,
1947 u32 num_pkts, u64 pattern); 2020 u64 pattern);
1948extern int be_cmd_ddr_dma_test(struct be_adapter *adapter, u64 pattern, 2021int be_cmd_ddr_dma_test(struct be_adapter *adapter, u64 pattern, u32 byte_cnt,
1949 u32 byte_cnt, struct be_dma_mem *cmd); 2022 struct be_dma_mem *cmd);
1950extern int be_cmd_get_seeprom_data(struct be_adapter *adapter, 2023int be_cmd_get_seeprom_data(struct be_adapter *adapter,
1951 struct be_dma_mem *nonemb_cmd); 2024 struct be_dma_mem *nonemb_cmd);
1952extern int be_cmd_set_loopback(struct be_adapter *adapter, u8 port_num, 2025int be_cmd_set_loopback(struct be_adapter *adapter, u8 port_num,
1953 u8 loopback_type, u8 enable); 2026 u8 loopback_type, u8 enable);
1954extern int be_cmd_get_phy_info(struct be_adapter *adapter); 2027int be_cmd_get_phy_info(struct be_adapter *adapter);
1955extern int be_cmd_set_qos(struct be_adapter *adapter, u32 bps, u32 domain); 2028int be_cmd_set_qos(struct be_adapter *adapter, u32 bps, u32 domain);
1956extern void be_detect_error(struct be_adapter *adapter); 2029void be_detect_error(struct be_adapter *adapter);
1957extern int be_cmd_get_die_temperature(struct be_adapter *adapter); 2030int be_cmd_get_die_temperature(struct be_adapter *adapter);
1958extern int be_cmd_get_cntl_attributes(struct be_adapter *adapter); 2031int be_cmd_get_cntl_attributes(struct be_adapter *adapter);
1959extern int be_cmd_req_native_mode(struct be_adapter *adapter); 2032int be_cmd_req_native_mode(struct be_adapter *adapter);
1960extern int be_cmd_get_reg_len(struct be_adapter *adapter, u32 *log_size); 2033int be_cmd_get_reg_len(struct be_adapter *adapter, u32 *log_size);
1961extern void be_cmd_get_regs(struct be_adapter *adapter, u32 buf_len, void *buf); 2034void be_cmd_get_regs(struct be_adapter *adapter, u32 buf_len, void *buf);
1962extern int be_cmd_get_fn_privileges(struct be_adapter *adapter, 2035int be_cmd_get_fn_privileges(struct be_adapter *adapter, u32 *privilege,
1963 u32 *privilege, u32 domain); 2036 u32 domain);
1964extern int be_cmd_set_fn_privileges(struct be_adapter *adapter, 2037int be_cmd_set_fn_privileges(struct be_adapter *adapter, u32 privileges,
1965 u32 privileges, u32 vf_num); 2038 u32 vf_num);
1966extern int be_cmd_get_mac_from_list(struct be_adapter *adapter, u8 *mac, 2039int be_cmd_get_mac_from_list(struct be_adapter *adapter, u8 *mac,
1967 bool *pmac_id_active, u32 *pmac_id, 2040 bool *pmac_id_active, u32 *pmac_id, u8 domain);
1968 u8 domain); 2041int be_cmd_get_active_mac(struct be_adapter *adapter, u32 pmac_id, u8 *mac);
1969extern int be_cmd_get_active_mac(struct be_adapter *adapter, u32 pmac_id, 2042int be_cmd_get_perm_mac(struct be_adapter *adapter, u8 *mac);
1970 u8 *mac); 2043int be_cmd_set_mac_list(struct be_adapter *adapter, u8 *mac_array, u8 mac_count,
1971extern int be_cmd_get_perm_mac(struct be_adapter *adapter, u8 *mac); 2044 u32 domain);
1972extern int be_cmd_set_mac_list(struct be_adapter *adapter, u8 *mac_array, 2045int be_cmd_set_mac(struct be_adapter *adapter, u8 *mac, int if_id, u32 dom);
1973 u8 mac_count, u32 domain); 2046int be_cmd_set_hsw_config(struct be_adapter *adapter, u16 pvid, u32 domain,
1974extern int be_cmd_set_mac(struct be_adapter *adapter, u8 *mac, int if_id, 2047 u16 intf_id, u16 hsw_mode);
1975 u32 dom); 2048int be_cmd_get_hsw_config(struct be_adapter *adapter, u16 *pvid, u32 domain,
1976extern int be_cmd_set_hsw_config(struct be_adapter *adapter, u16 pvid, 2049 u16 intf_id, u8 *mode);
1977 u32 domain, u16 intf_id, u16 hsw_mode); 2050int be_cmd_get_acpi_wol_cap(struct be_adapter *adapter);
1978extern int be_cmd_get_hsw_config(struct be_adapter *adapter, u16 *pvid, 2051int be_cmd_get_ext_fat_capabilites(struct be_adapter *adapter,
1979 u32 domain, u16 intf_id, u8 *mode); 2052 struct be_dma_mem *cmd);
1980extern int be_cmd_get_acpi_wol_cap(struct be_adapter *adapter); 2053int be_cmd_set_ext_fat_capabilites(struct be_adapter *adapter,
1981extern int be_cmd_get_ext_fat_capabilites(struct be_adapter *adapter, 2054 struct be_dma_mem *cmd,
1982 struct be_dma_mem *cmd); 2055 struct be_fat_conf_params *cfgs);
1983extern int be_cmd_set_ext_fat_capabilites(struct be_adapter *adapter, 2056int lancer_physdev_ctrl(struct be_adapter *adapter, u32 mask);
1984 struct be_dma_mem *cmd, 2057int lancer_initiate_dump(struct be_adapter *adapter);
1985 struct be_fat_conf_params *cfgs); 2058bool dump_present(struct be_adapter *adapter);
1986extern int lancer_wait_ready(struct be_adapter *adapter); 2059int lancer_test_and_set_rdy_state(struct be_adapter *adapter);
1987extern int lancer_physdev_ctrl(struct be_adapter *adapter, u32 mask); 2060int be_cmd_query_port_name(struct be_adapter *adapter, u8 *port_name);
1988extern int lancer_initiate_dump(struct be_adapter *adapter);
1989extern bool dump_present(struct be_adapter *adapter);
1990extern int lancer_test_and_set_rdy_state(struct be_adapter *adapter);
1991extern int be_cmd_query_port_name(struct be_adapter *adapter, u8 *port_name);
1992int be_cmd_get_func_config(struct be_adapter *adapter, 2061int be_cmd_get_func_config(struct be_adapter *adapter,
1993 struct be_resources *res); 2062 struct be_resources *res);
1994int be_cmd_get_profile_config(struct be_adapter *adapter, 2063int be_cmd_get_profile_config(struct be_adapter *adapter,
1995 struct be_resources *res, u8 domain); 2064 struct be_resources *res, u8 domain);
1996extern int be_cmd_set_profile_config(struct be_adapter *adapter, u32 bps, 2065int be_cmd_set_profile_config(struct be_adapter *adapter, u32 bps, u8 domain);
1997 u8 domain); 2066int be_cmd_get_if_id(struct be_adapter *adapter, struct be_vf_cfg *vf_cfg,
1998extern int be_cmd_get_if_id(struct be_adapter *adapter, 2067 int vf_num);
1999 struct be_vf_cfg *vf_cfg, int vf_num); 2068int be_cmd_enable_vf(struct be_adapter *adapter, u8 domain);
2000extern int be_cmd_enable_vf(struct be_adapter *adapter, u8 domain); 2069int be_cmd_intr_set(struct be_adapter *adapter, bool intr_enable);
2001extern int be_cmd_intr_set(struct be_adapter *adapter, bool intr_enable);
diff --git a/drivers/net/ethernet/emulex/benet/be_ethtool.c b/drivers/net/ethernet/emulex/benet/be_ethtool.c
index b440a1fac77b..08330034d9ef 100644
--- a/drivers/net/ethernet/emulex/benet/be_ethtool.c
+++ b/drivers/net/ethernet/emulex/benet/be_ethtool.c
@@ -116,7 +116,12 @@ static const struct be_ethtool_stat et_stats[] = {
116 {DRVSTAT_INFO(rx_drops_mtu)}, 116 {DRVSTAT_INFO(rx_drops_mtu)},
117 /* Number of packets dropped due to random early drop function */ 117 /* Number of packets dropped due to random early drop function */
118 {DRVSTAT_INFO(eth_red_drops)}, 118 {DRVSTAT_INFO(eth_red_drops)},
119 {DRVSTAT_INFO(be_on_die_temperature)} 119 {DRVSTAT_INFO(be_on_die_temperature)},
120 {DRVSTAT_INFO(rx_roce_bytes_lsd)},
121 {DRVSTAT_INFO(rx_roce_bytes_msd)},
122 {DRVSTAT_INFO(rx_roce_frames)},
123 {DRVSTAT_INFO(roce_drops_payload_len)},
124 {DRVSTAT_INFO(roce_drops_crc)}
120}; 125};
121#define ETHTOOL_STATS_NUM ARRAY_SIZE(et_stats) 126#define ETHTOOL_STATS_NUM ARRAY_SIZE(et_stats)
122 127
@@ -155,7 +160,9 @@ static const struct be_ethtool_stat et_tx_stats[] = {
155 /* Number of times the TX queue was stopped due to lack 160 /* Number of times the TX queue was stopped due to lack
156 * of spaces in the TXQ. 161 * of spaces in the TXQ.
157 */ 162 */
158 {DRVSTAT_TX_INFO(tx_stops)} 163 {DRVSTAT_TX_INFO(tx_stops)},
164 /* Pkts dropped in the driver's transmit path */
165 {DRVSTAT_TX_INFO(tx_drv_drops)}
159}; 166};
160#define ETHTOOL_TXSTATS_NUM (ARRAY_SIZE(et_tx_stats)) 167#define ETHTOOL_TXSTATS_NUM (ARRAY_SIZE(et_tx_stats))
161 168
@@ -290,19 +297,19 @@ static int be_get_coalesce(struct net_device *netdev,
290 struct ethtool_coalesce *et) 297 struct ethtool_coalesce *et)
291{ 298{
292 struct be_adapter *adapter = netdev_priv(netdev); 299 struct be_adapter *adapter = netdev_priv(netdev);
293 struct be_eq_obj *eqo = &adapter->eq_obj[0]; 300 struct be_aic_obj *aic = &adapter->aic_obj[0];
294 301
295 302
296 et->rx_coalesce_usecs = eqo->cur_eqd; 303 et->rx_coalesce_usecs = aic->prev_eqd;
297 et->rx_coalesce_usecs_high = eqo->max_eqd; 304 et->rx_coalesce_usecs_high = aic->max_eqd;
298 et->rx_coalesce_usecs_low = eqo->min_eqd; 305 et->rx_coalesce_usecs_low = aic->min_eqd;
299 306
300 et->tx_coalesce_usecs = eqo->cur_eqd; 307 et->tx_coalesce_usecs = aic->prev_eqd;
301 et->tx_coalesce_usecs_high = eqo->max_eqd; 308 et->tx_coalesce_usecs_high = aic->max_eqd;
302 et->tx_coalesce_usecs_low = eqo->min_eqd; 309 et->tx_coalesce_usecs_low = aic->min_eqd;
303 310
304 et->use_adaptive_rx_coalesce = eqo->enable_aic; 311 et->use_adaptive_rx_coalesce = aic->enable;
305 et->use_adaptive_tx_coalesce = eqo->enable_aic; 312 et->use_adaptive_tx_coalesce = aic->enable;
306 313
307 return 0; 314 return 0;
308} 315}
@@ -314,14 +321,17 @@ static int be_set_coalesce(struct net_device *netdev,
314 struct ethtool_coalesce *et) 321 struct ethtool_coalesce *et)
315{ 322{
316 struct be_adapter *adapter = netdev_priv(netdev); 323 struct be_adapter *adapter = netdev_priv(netdev);
324 struct be_aic_obj *aic = &adapter->aic_obj[0];
317 struct be_eq_obj *eqo; 325 struct be_eq_obj *eqo;
318 int i; 326 int i;
319 327
320 for_all_evt_queues(adapter, eqo, i) { 328 for_all_evt_queues(adapter, eqo, i) {
321 eqo->enable_aic = et->use_adaptive_rx_coalesce; 329 aic->enable = et->use_adaptive_rx_coalesce;
322 eqo->max_eqd = min(et->rx_coalesce_usecs_high, BE_MAX_EQD); 330 aic->max_eqd = min(et->rx_coalesce_usecs_high, BE_MAX_EQD);
323 eqo->min_eqd = min(et->rx_coalesce_usecs_low, eqo->max_eqd); 331 aic->min_eqd = min(et->rx_coalesce_usecs_low, aic->max_eqd);
324 eqo->eqd = et->rx_coalesce_usecs; 332 aic->et_eqd = min(et->rx_coalesce_usecs, aic->max_eqd);
333 aic->et_eqd = max(aic->et_eqd, aic->min_eqd);
334 aic++;
325 } 335 }
326 336
327 return 0; 337 return 0;
diff --git a/drivers/net/ethernet/emulex/benet/be_main.c b/drivers/net/ethernet/emulex/benet/be_main.c
index 53ed58b492c8..741d3bff5ae7 100644
--- a/drivers/net/ethernet/emulex/benet/be_main.c
+++ b/drivers/net/ethernet/emulex/benet/be_main.c
@@ -22,6 +22,7 @@
22#include <asm/div64.h> 22#include <asm/div64.h>
23#include <linux/aer.h> 23#include <linux/aer.h>
24#include <linux/if_bridge.h> 24#include <linux/if_bridge.h>
25#include <net/busy_poll.h>
25 26
26MODULE_VERSION(DRV_VER); 27MODULE_VERSION(DRV_VER);
27MODULE_DEVICE_TABLE(pci, be_dev_ids); 28MODULE_DEVICE_TABLE(pci, be_dev_ids);
@@ -306,10 +307,14 @@ static void *hw_stats_from_cmd(struct be_adapter *adapter)
306 struct be_cmd_resp_get_stats_v0 *cmd = adapter->stats_cmd.va; 307 struct be_cmd_resp_get_stats_v0 *cmd = adapter->stats_cmd.va;
307 308
308 return &cmd->hw_stats; 309 return &cmd->hw_stats;
309 } else { 310 } else if (BE3_chip(adapter)) {
310 struct be_cmd_resp_get_stats_v1 *cmd = adapter->stats_cmd.va; 311 struct be_cmd_resp_get_stats_v1 *cmd = adapter->stats_cmd.va;
311 312
312 return &cmd->hw_stats; 313 return &cmd->hw_stats;
314 } else {
315 struct be_cmd_resp_get_stats_v2 *cmd = adapter->stats_cmd.va;
316
317 return &cmd->hw_stats;
313 } 318 }
314} 319}
315 320
@@ -320,10 +325,14 @@ static void *be_erx_stats_from_cmd(struct be_adapter *adapter)
320 struct be_hw_stats_v0 *hw_stats = hw_stats_from_cmd(adapter); 325 struct be_hw_stats_v0 *hw_stats = hw_stats_from_cmd(adapter);
321 326
322 return &hw_stats->erx; 327 return &hw_stats->erx;
323 } else { 328 } else if (BE3_chip(adapter)) {
324 struct be_hw_stats_v1 *hw_stats = hw_stats_from_cmd(adapter); 329 struct be_hw_stats_v1 *hw_stats = hw_stats_from_cmd(adapter);
325 330
326 return &hw_stats->erx; 331 return &hw_stats->erx;
332 } else {
333 struct be_hw_stats_v2 *hw_stats = hw_stats_from_cmd(adapter);
334
335 return &hw_stats->erx;
327 } 336 }
328} 337}
329 338
@@ -422,6 +431,60 @@ static void populate_be_v1_stats(struct be_adapter *adapter)
422 adapter->drv_stats.eth_red_drops = pmem_sts->eth_red_drops; 431 adapter->drv_stats.eth_red_drops = pmem_sts->eth_red_drops;
423} 432}
424 433
434static void populate_be_v2_stats(struct be_adapter *adapter)
435{
436 struct be_hw_stats_v2 *hw_stats = hw_stats_from_cmd(adapter);
437 struct be_pmem_stats *pmem_sts = &hw_stats->pmem;
438 struct be_rxf_stats_v2 *rxf_stats = &hw_stats->rxf;
439 struct be_port_rxf_stats_v2 *port_stats =
440 &rxf_stats->port[adapter->port_num];
441 struct be_drv_stats *drvs = &adapter->drv_stats;
442
443 be_dws_le_to_cpu(hw_stats, sizeof(*hw_stats));
444 drvs->pmem_fifo_overflow_drop = port_stats->pmem_fifo_overflow_drop;
445 drvs->rx_priority_pause_frames = port_stats->rx_priority_pause_frames;
446 drvs->rx_pause_frames = port_stats->rx_pause_frames;
447 drvs->rx_crc_errors = port_stats->rx_crc_errors;
448 drvs->rx_control_frames = port_stats->rx_control_frames;
449 drvs->rx_in_range_errors = port_stats->rx_in_range_errors;
450 drvs->rx_frame_too_long = port_stats->rx_frame_too_long;
451 drvs->rx_dropped_runt = port_stats->rx_dropped_runt;
452 drvs->rx_ip_checksum_errs = port_stats->rx_ip_checksum_errs;
453 drvs->rx_tcp_checksum_errs = port_stats->rx_tcp_checksum_errs;
454 drvs->rx_udp_checksum_errs = port_stats->rx_udp_checksum_errs;
455 drvs->rx_dropped_tcp_length = port_stats->rx_dropped_tcp_length;
456 drvs->rx_dropped_too_small = port_stats->rx_dropped_too_small;
457 drvs->rx_dropped_too_short = port_stats->rx_dropped_too_short;
458 drvs->rx_out_range_errors = port_stats->rx_out_range_errors;
459 drvs->rx_dropped_header_too_small =
460 port_stats->rx_dropped_header_too_small;
461 drvs->rx_input_fifo_overflow_drop =
462 port_stats->rx_input_fifo_overflow_drop;
463 drvs->rx_address_filtered = port_stats->rx_address_filtered;
464 drvs->rx_alignment_symbol_errors =
465 port_stats->rx_alignment_symbol_errors;
466 drvs->rxpp_fifo_overflow_drop = port_stats->rxpp_fifo_overflow_drop;
467 drvs->tx_pauseframes = port_stats->tx_pauseframes;
468 drvs->tx_controlframes = port_stats->tx_controlframes;
469 drvs->tx_priority_pauseframes = port_stats->tx_priority_pauseframes;
470 drvs->jabber_events = port_stats->jabber_events;
471 drvs->rx_drops_no_pbuf = rxf_stats->rx_drops_no_pbuf;
472 drvs->rx_drops_no_erx_descr = rxf_stats->rx_drops_no_erx_descr;
473 drvs->forwarded_packets = rxf_stats->forwarded_packets;
474 drvs->rx_drops_mtu = rxf_stats->rx_drops_mtu;
475 drvs->rx_drops_no_tpre_descr = rxf_stats->rx_drops_no_tpre_descr;
476 drvs->rx_drops_too_many_frags = rxf_stats->rx_drops_too_many_frags;
477 adapter->drv_stats.eth_red_drops = pmem_sts->eth_red_drops;
478 if (be_roce_supported(adapter)) {
479 drvs->rx_roce_bytes_lsd = port_stats->roce_bytes_received_lsd;
480 drvs->rx_roce_bytes_msd = port_stats->roce_bytes_received_msd;
481 drvs->rx_roce_frames = port_stats->roce_frames_received;
482 drvs->roce_drops_crc = port_stats->roce_drops_crc;
483 drvs->roce_drops_payload_len =
484 port_stats->roce_drops_payload_len;
485 }
486}
487
425static void populate_lancer_stats(struct be_adapter *adapter) 488static void populate_lancer_stats(struct be_adapter *adapter)
426{ 489{
427 490
@@ -489,7 +552,7 @@ static void populate_erx_stats(struct be_adapter *adapter,
489 552
490void be_parse_stats(struct be_adapter *adapter) 553void be_parse_stats(struct be_adapter *adapter)
491{ 554{
492 struct be_erx_stats_v1 *erx = be_erx_stats_from_cmd(adapter); 555 struct be_erx_stats_v2 *erx = be_erx_stats_from_cmd(adapter);
493 struct be_rx_obj *rxo; 556 struct be_rx_obj *rxo;
494 int i; 557 int i;
495 u32 erx_stat; 558 u32 erx_stat;
@@ -499,11 +562,13 @@ void be_parse_stats(struct be_adapter *adapter)
499 } else { 562 } else {
500 if (BE2_chip(adapter)) 563 if (BE2_chip(adapter))
501 populate_be_v0_stats(adapter); 564 populate_be_v0_stats(adapter);
502 else 565 else if (BE3_chip(adapter))
503 /* for BE3 and Skyhawk */ 566 /* for BE3 */
504 populate_be_v1_stats(adapter); 567 populate_be_v1_stats(adapter);
568 else
569 populate_be_v2_stats(adapter);
505 570
506 /* as erx_v1 is longer than v0, ok to use v1 for v0 access */ 571 /* erx_v2 is longer than v0, v1. use v2 for v0, v1 access */
507 for_all_rx_queues(adapter, rxo, i) { 572 for_all_rx_queues(adapter, rxo, i) {
508 erx_stat = erx->rx_drops_no_fragments[rxo->q.id]; 573 erx_stat = erx->rx_drops_no_fragments[rxo->q.id];
509 populate_erx_stats(adapter, rxo, erx_stat); 574 populate_erx_stats(adapter, rxo, erx_stat);
@@ -935,8 +1000,10 @@ static netdev_tx_t be_xmit(struct sk_buff *skb, struct net_device *netdev)
935 u32 start = txq->head; 1000 u32 start = txq->head;
936 1001
937 skb = be_xmit_workarounds(adapter, skb, &skip_hw_vlan); 1002 skb = be_xmit_workarounds(adapter, skb, &skip_hw_vlan);
938 if (!skb) 1003 if (!skb) {
1004 tx_stats(txo)->tx_drv_drops++;
939 return NETDEV_TX_OK; 1005 return NETDEV_TX_OK;
1006 }
940 1007
941 wrb_cnt = wrb_cnt_for_skb(adapter, skb, &dummy_wrb); 1008 wrb_cnt = wrb_cnt_for_skb(adapter, skb, &dummy_wrb);
942 1009
@@ -965,6 +1032,7 @@ static netdev_tx_t be_xmit(struct sk_buff *skb, struct net_device *netdev)
965 be_tx_stats_update(txo, wrb_cnt, copied, gso_segs, stopped); 1032 be_tx_stats_update(txo, wrb_cnt, copied, gso_segs, stopped);
966 } else { 1033 } else {
967 txq->head = start; 1034 txq->head = start;
1035 tx_stats(txo)->tx_drv_drops++;
968 dev_kfree_skb_any(skb); 1036 dev_kfree_skb_any(skb);
969 } 1037 }
970 return NETDEV_TX_OK; 1038 return NETDEV_TX_OK;
@@ -1275,53 +1343,79 @@ static int be_set_vf_tx_rate(struct net_device *netdev,
1275 return status; 1343 return status;
1276} 1344}
1277 1345
1278static void be_eqd_update(struct be_adapter *adapter, struct be_eq_obj *eqo) 1346static void be_aic_update(struct be_aic_obj *aic, u64 rx_pkts, u64 tx_pkts,
1347 ulong now)
1279{ 1348{
1280 struct be_rx_stats *stats = rx_stats(&adapter->rx_obj[eqo->idx]); 1349 aic->rx_pkts_prev = rx_pkts;
1281 ulong now = jiffies; 1350 aic->tx_reqs_prev = tx_pkts;
1282 ulong delta = now - stats->rx_jiffies; 1351 aic->jiffies = now;
1283 u64 pkts; 1352}
1284 unsigned int start, eqd;
1285 1353
1286 if (!eqo->enable_aic) { 1354static void be_eqd_update(struct be_adapter *adapter)
1287 eqd = eqo->eqd; 1355{
1288 goto modify_eqd; 1356 struct be_set_eqd set_eqd[MAX_EVT_QS];
1289 } 1357 int eqd, i, num = 0, start;
1358 struct be_aic_obj *aic;
1359 struct be_eq_obj *eqo;
1360 struct be_rx_obj *rxo;
1361 struct be_tx_obj *txo;
1362 u64 rx_pkts, tx_pkts;
1363 ulong now;
1364 u32 pps, delta;
1290 1365
1291 if (eqo->idx >= adapter->num_rx_qs) 1366 for_all_evt_queues(adapter, eqo, i) {
1292 return; 1367 aic = &adapter->aic_obj[eqo->idx];
1368 if (!aic->enable) {
1369 if (aic->jiffies)
1370 aic->jiffies = 0;
1371 eqd = aic->et_eqd;
1372 goto modify_eqd;
1373 }
1293 1374
1294 stats = rx_stats(&adapter->rx_obj[eqo->idx]); 1375 rxo = &adapter->rx_obj[eqo->idx];
1376 do {
1377 start = u64_stats_fetch_begin_bh(&rxo->stats.sync);
1378 rx_pkts = rxo->stats.rx_pkts;
1379 } while (u64_stats_fetch_retry_bh(&rxo->stats.sync, start));
1295 1380
1296 /* Wrapped around */ 1381 txo = &adapter->tx_obj[eqo->idx];
1297 if (time_before(now, stats->rx_jiffies)) { 1382 do {
1298 stats->rx_jiffies = now; 1383 start = u64_stats_fetch_begin_bh(&txo->stats.sync);
1299 return; 1384 tx_pkts = txo->stats.tx_reqs;
1300 } 1385 } while (u64_stats_fetch_retry_bh(&txo->stats.sync, start));
1301 1386
1302 /* Update once a second */
1303 if (delta < HZ)
1304 return;
1305 1387
1306 do { 1388 /* Skip, if wrapped around or first calculation */
1307 start = u64_stats_fetch_begin_bh(&stats->sync); 1389 now = jiffies;
1308 pkts = stats->rx_pkts; 1390 if (!aic->jiffies || time_before(now, aic->jiffies) ||
1309 } while (u64_stats_fetch_retry_bh(&stats->sync, start)); 1391 rx_pkts < aic->rx_pkts_prev ||
1310 1392 tx_pkts < aic->tx_reqs_prev) {
1311 stats->rx_pps = (unsigned long)(pkts - stats->rx_pkts_prev) / (delta / HZ); 1393 be_aic_update(aic, rx_pkts, tx_pkts, now);
1312 stats->rx_pkts_prev = pkts; 1394 continue;
1313 stats->rx_jiffies = now; 1395 }
1314 eqd = (stats->rx_pps / 110000) << 3; 1396
1315 eqd = min(eqd, eqo->max_eqd); 1397 delta = jiffies_to_msecs(now - aic->jiffies);
1316 eqd = max(eqd, eqo->min_eqd); 1398 pps = (((u32)(rx_pkts - aic->rx_pkts_prev) * 1000) / delta) +
1317 if (eqd < 10) 1399 (((u32)(tx_pkts - aic->tx_reqs_prev) * 1000) / delta);
1318 eqd = 0; 1400 eqd = (pps / 15000) << 2;
1401
1402 if (eqd < 8)
1403 eqd = 0;
1404 eqd = min_t(u32, eqd, aic->max_eqd);
1405 eqd = max_t(u32, eqd, aic->min_eqd);
1319 1406
1407 be_aic_update(aic, rx_pkts, tx_pkts, now);
1320modify_eqd: 1408modify_eqd:
1321 if (eqd != eqo->cur_eqd) { 1409 if (eqd != aic->prev_eqd) {
1322 be_cmd_modify_eqd(adapter, eqo->q.id, eqd); 1410 set_eqd[num].delay_multiplier = (eqd * 65)/100;
1323 eqo->cur_eqd = eqd; 1411 set_eqd[num].eq_id = eqo->q.id;
1412 aic->prev_eqd = eqd;
1413 num++;
1414 }
1324 } 1415 }
1416
1417 if (num)
1418 be_cmd_modify_eqd(adapter, set_eqd, num);
1325} 1419}
1326 1420
1327static void be_rx_stats_update(struct be_rx_obj *rxo, 1421static void be_rx_stats_update(struct be_rx_obj *rxo,
@@ -1463,7 +1557,7 @@ static void skb_fill_rx_data(struct be_rx_obj *rxo, struct sk_buff *skb,
1463} 1557}
1464 1558
1465/* Process the RX completion indicated by rxcp when GRO is disabled */ 1559/* Process the RX completion indicated by rxcp when GRO is disabled */
1466static void be_rx_compl_process(struct be_rx_obj *rxo, 1560static void be_rx_compl_process(struct be_rx_obj *rxo, struct napi_struct *napi,
1467 struct be_rx_compl_info *rxcp) 1561 struct be_rx_compl_info *rxcp)
1468{ 1562{
1469 struct be_adapter *adapter = rxo->adapter; 1563 struct be_adapter *adapter = rxo->adapter;
@@ -1488,7 +1582,7 @@ static void be_rx_compl_process(struct be_rx_obj *rxo,
1488 skb_record_rx_queue(skb, rxo - &adapter->rx_obj[0]); 1582 skb_record_rx_queue(skb, rxo - &adapter->rx_obj[0]);
1489 if (netdev->features & NETIF_F_RXHASH) 1583 if (netdev->features & NETIF_F_RXHASH)
1490 skb->rxhash = rxcp->rss_hash; 1584 skb->rxhash = rxcp->rss_hash;
1491 1585 skb_mark_napi_id(skb, napi);
1492 1586
1493 if (rxcp->vlanf) 1587 if (rxcp->vlanf)
1494 __vlan_hwaccel_put_tag(skb, htons(ETH_P_8021Q), rxcp->vlan_tag); 1588 __vlan_hwaccel_put_tag(skb, htons(ETH_P_8021Q), rxcp->vlan_tag);
@@ -1546,6 +1640,7 @@ static void be_rx_compl_process_gro(struct be_rx_obj *rxo,
1546 skb_record_rx_queue(skb, rxo - &adapter->rx_obj[0]); 1640 skb_record_rx_queue(skb, rxo - &adapter->rx_obj[0]);
1547 if (adapter->netdev->features & NETIF_F_RXHASH) 1641 if (adapter->netdev->features & NETIF_F_RXHASH)
1548 skb->rxhash = rxcp->rss_hash; 1642 skb->rxhash = rxcp->rss_hash;
1643 skb_mark_napi_id(skb, napi);
1549 1644
1550 if (rxcp->vlanf) 1645 if (rxcp->vlanf)
1551 __vlan_hwaccel_put_tag(skb, htons(ETH_P_8021Q), rxcp->vlan_tag); 1646 __vlan_hwaccel_put_tag(skb, htons(ETH_P_8021Q), rxcp->vlan_tag);
@@ -1726,6 +1821,8 @@ static void be_post_rx_frags(struct be_rx_obj *rxo, gfp_t gfp)
1726 1821
1727 if (posted) { 1822 if (posted) {
1728 atomic_add(posted, &rxq->used); 1823 atomic_add(posted, &rxq->used);
1824 if (rxo->rx_post_starved)
1825 rxo->rx_post_starved = false;
1729 be_rxq_notify(adapter, rxq->id, posted); 1826 be_rxq_notify(adapter, rxq->id, posted);
1730 } else if (atomic_read(&rxq->used) == 0) { 1827 } else if (atomic_read(&rxq->used) == 0) {
1731 /* Let be_worker replenish when memory is available */ 1828 /* Let be_worker replenish when memory is available */
@@ -1928,6 +2025,7 @@ static void be_evt_queues_destroy(struct be_adapter *adapter)
1928 if (eqo->q.created) { 2025 if (eqo->q.created) {
1929 be_eq_clean(eqo); 2026 be_eq_clean(eqo);
1930 be_cmd_q_destroy(adapter, &eqo->q, QTYPE_EQ); 2027 be_cmd_q_destroy(adapter, &eqo->q, QTYPE_EQ);
2028 napi_hash_del(&eqo->napi);
1931 netif_napi_del(&eqo->napi); 2029 netif_napi_del(&eqo->napi);
1932 } 2030 }
1933 be_queue_free(adapter, &eqo->q); 2031 be_queue_free(adapter, &eqo->q);
@@ -1938,6 +2036,7 @@ static int be_evt_queues_create(struct be_adapter *adapter)
1938{ 2036{
1939 struct be_queue_info *eq; 2037 struct be_queue_info *eq;
1940 struct be_eq_obj *eqo; 2038 struct be_eq_obj *eqo;
2039 struct be_aic_obj *aic;
1941 int i, rc; 2040 int i, rc;
1942 2041
1943 adapter->num_evt_qs = min_t(u16, num_irqs(adapter), 2042 adapter->num_evt_qs = min_t(u16, num_irqs(adapter),
@@ -1946,11 +2045,13 @@ static int be_evt_queues_create(struct be_adapter *adapter)
1946 for_all_evt_queues(adapter, eqo, i) { 2045 for_all_evt_queues(adapter, eqo, i) {
1947 netif_napi_add(adapter->netdev, &eqo->napi, be_poll, 2046 netif_napi_add(adapter->netdev, &eqo->napi, be_poll,
1948 BE_NAPI_WEIGHT); 2047 BE_NAPI_WEIGHT);
2048 napi_hash_add(&eqo->napi);
2049 aic = &adapter->aic_obj[i];
1949 eqo->adapter = adapter; 2050 eqo->adapter = adapter;
1950 eqo->tx_budget = BE_TX_BUDGET; 2051 eqo->tx_budget = BE_TX_BUDGET;
1951 eqo->idx = i; 2052 eqo->idx = i;
1952 eqo->max_eqd = BE_MAX_EQD; 2053 aic->max_eqd = BE_MAX_EQD;
1953 eqo->enable_aic = true; 2054 aic->enable = true;
1954 2055
1955 eq = &eqo->q; 2056 eq = &eqo->q;
1956 rc = be_queue_alloc(adapter, eq, EVNT_Q_LEN, 2057 rc = be_queue_alloc(adapter, eq, EVNT_Q_LEN,
@@ -2167,7 +2268,7 @@ static inline bool do_gro(struct be_rx_compl_info *rxcp)
2167} 2268}
2168 2269
2169static int be_process_rx(struct be_rx_obj *rxo, struct napi_struct *napi, 2270static int be_process_rx(struct be_rx_obj *rxo, struct napi_struct *napi,
2170 int budget) 2271 int budget, int polling)
2171{ 2272{
2172 struct be_adapter *adapter = rxo->adapter; 2273 struct be_adapter *adapter = rxo->adapter;
2173 struct be_queue_info *rx_cq = &rxo->cq; 2274 struct be_queue_info *rx_cq = &rxo->cq;
@@ -2198,10 +2299,12 @@ static int be_process_rx(struct be_rx_obj *rxo, struct napi_struct *napi,
2198 goto loop_continue; 2299 goto loop_continue;
2199 } 2300 }
2200 2301
2201 if (do_gro(rxcp)) 2302 /* Don't do gro when we're busy_polling */
2303 if (do_gro(rxcp) && polling != BUSY_POLLING)
2202 be_rx_compl_process_gro(rxo, napi, rxcp); 2304 be_rx_compl_process_gro(rxo, napi, rxcp);
2203 else 2305 else
2204 be_rx_compl_process(rxo, rxcp); 2306 be_rx_compl_process(rxo, napi, rxcp);
2307
2205loop_continue: 2308loop_continue:
2206 be_rx_stats_update(rxo, rxcp); 2309 be_rx_stats_update(rxo, rxcp);
2207 } 2310 }
@@ -2209,7 +2312,11 @@ loop_continue:
2209 if (work_done) { 2312 if (work_done) {
2210 be_cq_notify(adapter, rx_cq->id, true, work_done); 2313 be_cq_notify(adapter, rx_cq->id, true, work_done);
2211 2314
2212 if (atomic_read(&rxo->q.used) < RX_FRAGS_REFILL_WM) 2315 /* When an rx-obj gets into post_starved state, just
2316 * let be_worker do the posting.
2317 */
2318 if (atomic_read(&rxo->q.used) < RX_FRAGS_REFILL_WM &&
2319 !rxo->rx_post_starved)
2213 be_post_rx_frags(rxo, GFP_ATOMIC); 2320 be_post_rx_frags(rxo, GFP_ATOMIC);
2214 } 2321 }
2215 2322
@@ -2254,6 +2361,7 @@ int be_poll(struct napi_struct *napi, int budget)
2254 struct be_eq_obj *eqo = container_of(napi, struct be_eq_obj, napi); 2361 struct be_eq_obj *eqo = container_of(napi, struct be_eq_obj, napi);
2255 struct be_adapter *adapter = eqo->adapter; 2362 struct be_adapter *adapter = eqo->adapter;
2256 int max_work = 0, work, i, num_evts; 2363 int max_work = 0, work, i, num_evts;
2364 struct be_rx_obj *rxo;
2257 bool tx_done; 2365 bool tx_done;
2258 2366
2259 num_evts = events_get(eqo); 2367 num_evts = events_get(eqo);
@@ -2266,13 +2374,18 @@ int be_poll(struct napi_struct *napi, int budget)
2266 max_work = budget; 2374 max_work = budget;
2267 } 2375 }
2268 2376
2269 /* This loop will iterate twice for EQ0 in which 2377 if (be_lock_napi(eqo)) {
2270 * completions of the last RXQ (default one) are also processed 2378 /* This loop will iterate twice for EQ0 in which
2271 * For other EQs the loop iterates only once 2379 * completions of the last RXQ (default one) are also processed
2272 */ 2380 * For other EQs the loop iterates only once
2273 for (i = eqo->idx; i < adapter->num_rx_qs; i += adapter->num_evt_qs) { 2381 */
2274 work = be_process_rx(&adapter->rx_obj[i], napi, budget); 2382 for_all_rx_queues_on_eq(adapter, eqo, rxo, i) {
2275 max_work = max(work, max_work); 2383 work = be_process_rx(rxo, napi, budget, NAPI_POLLING);
2384 max_work = max(work, max_work);
2385 }
2386 be_unlock_napi(eqo);
2387 } else {
2388 max_work = budget;
2276 } 2389 }
2277 2390
2278 if (is_mcc_eqo(eqo)) 2391 if (is_mcc_eqo(eqo))
@@ -2288,6 +2401,28 @@ int be_poll(struct napi_struct *napi, int budget)
2288 return max_work; 2401 return max_work;
2289} 2402}
2290 2403
2404#ifdef CONFIG_NET_RX_BUSY_POLL
2405static int be_busy_poll(struct napi_struct *napi)
2406{
2407 struct be_eq_obj *eqo = container_of(napi, struct be_eq_obj, napi);
2408 struct be_adapter *adapter = eqo->adapter;
2409 struct be_rx_obj *rxo;
2410 int i, work = 0;
2411
2412 if (!be_lock_busy_poll(eqo))
2413 return LL_FLUSH_BUSY;
2414
2415 for_all_rx_queues_on_eq(adapter, eqo, rxo, i) {
2416 work = be_process_rx(rxo, napi, 4, BUSY_POLLING);
2417 if (work)
2418 break;
2419 }
2420
2421 be_unlock_busy_poll(eqo);
2422 return work;
2423}
2424#endif
2425
2291void be_detect_error(struct be_adapter *adapter) 2426void be_detect_error(struct be_adapter *adapter)
2292{ 2427{
2293 u32 ue_lo = 0, ue_hi = 0, ue_lo_mask = 0, ue_hi_mask = 0; 2428 u32 ue_lo = 0, ue_hi = 0, ue_lo_mask = 0, ue_hi_mask = 0;
@@ -2519,9 +2654,11 @@ static int be_close(struct net_device *netdev)
2519 2654
2520 be_roce_dev_close(adapter); 2655 be_roce_dev_close(adapter);
2521 2656
2522 if (adapter->flags & BE_FLAGS_NAPI_ENABLED) { 2657 for_all_evt_queues(adapter, eqo, i) {
2523 for_all_evt_queues(adapter, eqo, i) 2658 if (adapter->flags & BE_FLAGS_NAPI_ENABLED) {
2524 napi_disable(&eqo->napi); 2659 napi_disable(&eqo->napi);
2660 be_disable_busy_poll(eqo);
2661 }
2525 adapter->flags &= ~BE_FLAGS_NAPI_ENABLED; 2662 adapter->flags &= ~BE_FLAGS_NAPI_ENABLED;
2526 } 2663 }
2527 2664
@@ -2632,6 +2769,7 @@ static int be_open(struct net_device *netdev)
2632 2769
2633 for_all_evt_queues(adapter, eqo, i) { 2770 for_all_evt_queues(adapter, eqo, i) {
2634 napi_enable(&eqo->napi); 2771 napi_enable(&eqo->napi);
2772 be_enable_busy_poll(eqo);
2635 be_eq_notify(adapter, eqo->q.id, true, false, 0); 2773 be_eq_notify(adapter, eqo->q.id, true, false, 0);
2636 } 2774 }
2637 adapter->flags |= BE_FLAGS_NAPI_ENABLED; 2775 adapter->flags |= BE_FLAGS_NAPI_ENABLED;
@@ -2937,7 +3075,8 @@ static int be_vf_setup(struct be_adapter *adapter)
2937 goto err; 3075 goto err;
2938 vf_cfg->def_vid = def_vlan; 3076 vf_cfg->def_vid = def_vlan;
2939 3077
2940 be_cmd_enable_vf(adapter, vf + 1); 3078 if (!old_vfs)
3079 be_cmd_enable_vf(adapter, vf + 1);
2941 } 3080 }
2942 3081
2943 if (!old_vfs) { 3082 if (!old_vfs) {
@@ -2962,12 +3101,12 @@ static void BEx_get_resources(struct be_adapter *adapter,
2962 struct pci_dev *pdev = adapter->pdev; 3101 struct pci_dev *pdev = adapter->pdev;
2963 bool use_sriov = false; 3102 bool use_sriov = false;
2964 3103
2965 if (BE3_chip(adapter) && be_physfn(adapter)) { 3104 if (BE3_chip(adapter) && sriov_want(adapter)) {
2966 int max_vfs; 3105 int max_vfs;
2967 3106
2968 max_vfs = pci_sriov_get_totalvfs(pdev); 3107 max_vfs = pci_sriov_get_totalvfs(pdev);
2969 res->max_vfs = max_vfs > 0 ? min(MAX_VFS, max_vfs) : 0; 3108 res->max_vfs = max_vfs > 0 ? min(MAX_VFS, max_vfs) : 0;
2970 use_sriov = res->max_vfs && num_vfs; 3109 use_sriov = res->max_vfs;
2971 } 3110 }
2972 3111
2973 if (be_physfn(adapter)) 3112 if (be_physfn(adapter))
@@ -2983,8 +3122,9 @@ static void BEx_get_resources(struct be_adapter *adapter,
2983 res->max_vlans = BE_NUM_VLANS_SUPPORTED; 3122 res->max_vlans = BE_NUM_VLANS_SUPPORTED;
2984 res->max_mcast_mac = BE_MAX_MC; 3123 res->max_mcast_mac = BE_MAX_MC;
2985 3124
3125 /* For BE3 1Gb ports, F/W does not properly support multiple TXQs */
2986 if (BE2_chip(adapter) || use_sriov || be_is_mc(adapter) || 3126 if (BE2_chip(adapter) || use_sriov || be_is_mc(adapter) ||
2987 !be_physfn(adapter)) 3127 !be_physfn(adapter) || (adapter->port_num > 1))
2988 res->max_tx_qs = 1; 3128 res->max_tx_qs = 1;
2989 else 3129 else
2990 res->max_tx_qs = BE3_MAX_TX_QS; 3130 res->max_tx_qs = BE3_MAX_TX_QS;
@@ -3026,14 +3166,6 @@ static int be_get_resources(struct be_adapter *adapter)
3026 adapter->res = res; 3166 adapter->res = res;
3027 } 3167 }
3028 3168
3029 /* For BE3 only check if FW suggests a different max-txqs value */
3030 if (BE3_chip(adapter)) {
3031 status = be_cmd_get_profile_config(adapter, &res, 0);
3032 if (!status && res.max_tx_qs)
3033 adapter->res.max_tx_qs =
3034 min(adapter->res.max_tx_qs, res.max_tx_qs);
3035 }
3036
3037 /* For Lancer, SH etc read per-function resource limits from FW. 3169 /* For Lancer, SH etc read per-function resource limits from FW.
3038 * GET_FUNC_CONFIG returns per function guaranteed limits. 3170 * GET_FUNC_CONFIG returns per function guaranteed limits.
3039 * GET_PROFILE_CONFIG returns PCI-E related limits PF-pool limits 3171 * GET_PROFILE_CONFIG returns PCI-E related limits PF-pool limits
@@ -3264,7 +3396,7 @@ static int be_setup(struct be_adapter *adapter)
3264 be_cmd_set_flow_control(adapter, adapter->tx_fc, 3396 be_cmd_set_flow_control(adapter, adapter->tx_fc,
3265 adapter->rx_fc); 3397 adapter->rx_fc);
3266 3398
3267 if (be_physfn(adapter) && num_vfs) { 3399 if (sriov_want(adapter)) {
3268 if (be_max_vfs(adapter)) 3400 if (be_max_vfs(adapter))
3269 be_vf_setup(adapter); 3401 be_vf_setup(adapter);
3270 else 3402 else
@@ -3906,6 +4038,9 @@ static const struct net_device_ops be_netdev_ops = {
3906#endif 4038#endif
3907 .ndo_bridge_setlink = be_ndo_bridge_setlink, 4039 .ndo_bridge_setlink = be_ndo_bridge_setlink,
3908 .ndo_bridge_getlink = be_ndo_bridge_getlink, 4040 .ndo_bridge_getlink = be_ndo_bridge_getlink,
4041#ifdef CONFIG_NET_RX_BUSY_POLL
4042 .ndo_busy_poll = be_busy_poll
4043#endif
3909}; 4044};
3910 4045
3911static void be_netdev_init(struct net_device *netdev) 4046static void be_netdev_init(struct net_device *netdev)
@@ -3966,11 +4101,6 @@ static int be_roce_map_pci_bars(struct be_adapter *adapter)
3966static int be_map_pci_bars(struct be_adapter *adapter) 4101static int be_map_pci_bars(struct be_adapter *adapter)
3967{ 4102{
3968 u8 __iomem *addr; 4103 u8 __iomem *addr;
3969 u32 sli_intf;
3970
3971 pci_read_config_dword(adapter->pdev, SLI_INTF_REG_OFFSET, &sli_intf);
3972 adapter->if_type = (sli_intf & SLI_INTF_IF_TYPE_MASK) >>
3973 SLI_INTF_IF_TYPE_SHIFT;
3974 4104
3975 if (BEx_chip(adapter) && be_physfn(adapter)) { 4105 if (BEx_chip(adapter) && be_physfn(adapter)) {
3976 adapter->csr = pci_iomap(adapter->pdev, 2, 0); 4106 adapter->csr = pci_iomap(adapter->pdev, 2, 0);
@@ -4083,9 +4213,11 @@ static int be_stats_init(struct be_adapter *adapter)
4083 cmd->size = sizeof(struct lancer_cmd_req_pport_stats); 4213 cmd->size = sizeof(struct lancer_cmd_req_pport_stats);
4084 else if (BE2_chip(adapter)) 4214 else if (BE2_chip(adapter))
4085 cmd->size = sizeof(struct be_cmd_req_get_stats_v0); 4215 cmd->size = sizeof(struct be_cmd_req_get_stats_v0);
4086 else 4216 else if (BE3_chip(adapter))
4087 /* BE3 and Skyhawk */
4088 cmd->size = sizeof(struct be_cmd_req_get_stats_v1); 4217 cmd->size = sizeof(struct be_cmd_req_get_stats_v1);
4218 else
4219 /* ALL non-BE ASICs */
4220 cmd->size = sizeof(struct be_cmd_req_get_stats_v2);
4089 4221
4090 cmd->va = dma_zalloc_coherent(&adapter->pdev->dev, cmd->size, &cmd->dma, 4222 cmd->va = dma_zalloc_coherent(&adapter->pdev->dev, cmd->size, &cmd->dma,
4091 GFP_KERNEL); 4223 GFP_KERNEL);
@@ -4119,7 +4251,6 @@ static void be_remove(struct pci_dev *pdev)
4119 4251
4120 pci_disable_pcie_error_reporting(pdev); 4252 pci_disable_pcie_error_reporting(pdev);
4121 4253
4122 pci_set_drvdata(pdev, NULL);
4123 pci_release_regions(pdev); 4254 pci_release_regions(pdev);
4124 pci_disable_device(pdev); 4255 pci_disable_device(pdev);
4125 4256
@@ -4268,7 +4399,6 @@ static void be_worker(struct work_struct *work)
4268 struct be_adapter *adapter = 4399 struct be_adapter *adapter =
4269 container_of(work, struct be_adapter, work.work); 4400 container_of(work, struct be_adapter, work.work);
4270 struct be_rx_obj *rxo; 4401 struct be_rx_obj *rxo;
4271 struct be_eq_obj *eqo;
4272 int i; 4402 int i;
4273 4403
4274 /* when interrupts are not yet enabled, just reap any pending 4404 /* when interrupts are not yet enabled, just reap any pending
@@ -4293,14 +4423,14 @@ static void be_worker(struct work_struct *work)
4293 be_cmd_get_die_temperature(adapter); 4423 be_cmd_get_die_temperature(adapter);
4294 4424
4295 for_all_rx_queues(adapter, rxo, i) { 4425 for_all_rx_queues(adapter, rxo, i) {
4296 if (rxo->rx_post_starved) { 4426 /* Replenish RX-queues starved due to memory
4297 rxo->rx_post_starved = false; 4427 * allocation failures.
4428 */
4429 if (rxo->rx_post_starved)
4298 be_post_rx_frags(rxo, GFP_KERNEL); 4430 be_post_rx_frags(rxo, GFP_KERNEL);
4299 }
4300 } 4431 }
4301 4432
4302 for_all_evt_queues(adapter, eqo, i) 4433 be_eqd_update(adapter);
4303 be_eqd_update(adapter, eqo);
4304 4434
4305reschedule: 4435reschedule:
4306 adapter->work_counter++; 4436 adapter->work_counter++;
@@ -4376,9 +4506,11 @@ static int be_probe(struct pci_dev *pdev, const struct pci_device_id *pdev_id)
4376 } 4506 }
4377 } 4507 }
4378 4508
4379 status = pci_enable_pcie_error_reporting(pdev); 4509 if (be_physfn(adapter)) {
4380 if (status) 4510 status = pci_enable_pcie_error_reporting(pdev);
4381 dev_info(&pdev->dev, "Could not use PCIe error reporting\n"); 4511 if (!status)
4512 dev_info(&pdev->dev, "PCIe error reporting enabled\n");
4513 }
4382 4514
4383 status = be_ctrl_init(adapter); 4515 status = be_ctrl_init(adapter);
4384 if (status) 4516 if (status)
@@ -4449,7 +4581,6 @@ ctrl_clean:
4449 be_ctrl_cleanup(adapter); 4581 be_ctrl_cleanup(adapter);
4450free_netdev: 4582free_netdev:
4451 free_netdev(netdev); 4583 free_netdev(netdev);
4452 pci_set_drvdata(pdev, NULL);
4453rel_reg: 4584rel_reg:
4454 pci_release_regions(pdev); 4585 pci_release_regions(pdev);
4455disable_dev: 4586disable_dev:
diff --git a/drivers/net/ethernet/fealnx.c b/drivers/net/ethernet/fealnx.c
index c706b7a9397e..4b22a9579f85 100644
--- a/drivers/net/ethernet/fealnx.c
+++ b/drivers/net/ethernet/fealnx.c
@@ -699,7 +699,6 @@ static void fealnx_remove_one(struct pci_dev *pdev)
699 pci_iounmap(pdev, np->mem); 699 pci_iounmap(pdev, np->mem);
700 free_netdev(dev); 700 free_netdev(dev);
701 pci_release_regions(pdev); 701 pci_release_regions(pdev);
702 pci_set_drvdata(pdev, NULL);
703 } else 702 } else
704 printk(KERN_ERR "fealnx: remove for unknown device\n"); 703 printk(KERN_ERR "fealnx: remove for unknown device\n");
705} 704}
diff --git a/drivers/net/ethernet/freescale/fs_enet/fs_enet-main.c b/drivers/net/ethernet/freescale/fs_enet/fs_enet-main.c
index 6b60582ce8cf..56f2f608a9f4 100644
--- a/drivers/net/ethernet/freescale/fs_enet/fs_enet-main.c
+++ b/drivers/net/ethernet/freescale/fs_enet/fs_enet-main.c
@@ -1083,7 +1083,7 @@ static int fs_enet_probe(struct platform_device *ofdev)
1083 1083
1084 mac_addr = of_get_mac_address(ofdev->dev.of_node); 1084 mac_addr = of_get_mac_address(ofdev->dev.of_node);
1085 if (mac_addr) 1085 if (mac_addr)
1086 memcpy(ndev->dev_addr, mac_addr, 6); 1086 memcpy(ndev->dev_addr, mac_addr, ETH_ALEN);
1087 1087
1088 ret = fep->ops->allocate_bd(ndev); 1088 ret = fep->ops->allocate_bd(ndev);
1089 if (ret) 1089 if (ret)
diff --git a/drivers/net/ethernet/freescale/gianfar.c b/drivers/net/ethernet/freescale/gianfar.c
index ef95c3ae53e6..b14d7904a075 100644
--- a/drivers/net/ethernet/freescale/gianfar.c
+++ b/drivers/net/ethernet/freescale/gianfar.c
@@ -2920,7 +2920,7 @@ static int gfar_poll(struct napi_struct *napi, int budget)
2920 struct gfar_priv_rx_q *rx_queue = NULL; 2920 struct gfar_priv_rx_q *rx_queue = NULL;
2921 int work_done = 0, work_done_per_q = 0; 2921 int work_done = 0, work_done_per_q = 0;
2922 int i, budget_per_q = 0; 2922 int i, budget_per_q = 0;
2923 int has_tx_work; 2923 int has_tx_work = 0;
2924 unsigned long rstat_rxf; 2924 unsigned long rstat_rxf;
2925 int num_act_queues; 2925 int num_act_queues;
2926 2926
@@ -2935,62 +2935,51 @@ static int gfar_poll(struct napi_struct *napi, int budget)
2935 if (num_act_queues) 2935 if (num_act_queues)
2936 budget_per_q = budget/num_act_queues; 2936 budget_per_q = budget/num_act_queues;
2937 2937
2938 while (1) { 2938 for_each_set_bit(i, &gfargrp->tx_bit_map, priv->num_tx_queues) {
2939 has_tx_work = 0; 2939 tx_queue = priv->tx_queue[i];
2940 for_each_set_bit(i, &gfargrp->tx_bit_map, priv->num_tx_queues) { 2940 /* run Tx cleanup to completion */
2941 tx_queue = priv->tx_queue[i]; 2941 if (tx_queue->tx_skbuff[tx_queue->skb_dirtytx]) {
2942 /* run Tx cleanup to completion */ 2942 gfar_clean_tx_ring(tx_queue);
2943 if (tx_queue->tx_skbuff[tx_queue->skb_dirtytx]) { 2943 has_tx_work = 1;
2944 gfar_clean_tx_ring(tx_queue);
2945 has_tx_work = 1;
2946 }
2947 } 2944 }
2945 }
2948 2946
2949 for_each_set_bit(i, &gfargrp->rx_bit_map, priv->num_rx_queues) { 2947 for_each_set_bit(i, &gfargrp->rx_bit_map, priv->num_rx_queues) {
2950 /* skip queue if not active */ 2948 /* skip queue if not active */
2951 if (!(rstat_rxf & (RSTAT_CLEAR_RXF0 >> i))) 2949 if (!(rstat_rxf & (RSTAT_CLEAR_RXF0 >> i)))
2952 continue; 2950 continue;
2953
2954 rx_queue = priv->rx_queue[i];
2955 work_done_per_q =
2956 gfar_clean_rx_ring(rx_queue, budget_per_q);
2957 work_done += work_done_per_q;
2958
2959 /* finished processing this queue */
2960 if (work_done_per_q < budget_per_q) {
2961 /* clear active queue hw indication */
2962 gfar_write(&regs->rstat,
2963 RSTAT_CLEAR_RXF0 >> i);
2964 rstat_rxf &= ~(RSTAT_CLEAR_RXF0 >> i);
2965 num_act_queues--;
2966
2967 if (!num_act_queues)
2968 break;
2969 /* recompute budget per Rx queue */
2970 budget_per_q =
2971 (budget - work_done) / num_act_queues;
2972 }
2973 }
2974 2951
2975 if (work_done >= budget) 2952 rx_queue = priv->rx_queue[i];
2976 break; 2953 work_done_per_q =
2954 gfar_clean_rx_ring(rx_queue, budget_per_q);
2955 work_done += work_done_per_q;
2956
2957 /* finished processing this queue */
2958 if (work_done_per_q < budget_per_q) {
2959 /* clear active queue hw indication */
2960 gfar_write(&regs->rstat,
2961 RSTAT_CLEAR_RXF0 >> i);
2962 num_act_queues--;
2963
2964 if (!num_act_queues)
2965 break;
2966 }
2967 }
2977 2968
2978 if (!num_act_queues && !has_tx_work) { 2969 if (!num_act_queues && !has_tx_work) {
2979 2970
2980 napi_complete(napi); 2971 napi_complete(napi);
2981 2972
2982 /* Clear the halt bit in RSTAT */ 2973 /* Clear the halt bit in RSTAT */
2983 gfar_write(&regs->rstat, gfargrp->rstat); 2974 gfar_write(&regs->rstat, gfargrp->rstat);
2984 2975
2985 gfar_write(&regs->imask, IMASK_DEFAULT); 2976 gfar_write(&regs->imask, IMASK_DEFAULT);
2986 2977
2987 /* If we are coalescing interrupts, update the timer 2978 /* If we are coalescing interrupts, update the timer
2988 * Otherwise, clear it 2979 * Otherwise, clear it
2989 */ 2980 */
2990 gfar_configure_coalescing(priv, gfargrp->rx_bit_map, 2981 gfar_configure_coalescing(priv, gfargrp->rx_bit_map,
2991 gfargrp->tx_bit_map); 2982 gfargrp->tx_bit_map);
2992 break;
2993 }
2994 } 2983 }
2995 2984
2996 return work_done; 2985 return work_done;
diff --git a/drivers/net/ethernet/freescale/gianfar.h b/drivers/net/ethernet/freescale/gianfar.h
index 04112b98ff5d..114c58f9d8d2 100644
--- a/drivers/net/ethernet/freescale/gianfar.h
+++ b/drivers/net/ethernet/freescale/gianfar.h
@@ -1177,21 +1177,21 @@ static inline void gfar_read_filer(struct gfar_private *priv,
1177 *fpr = gfar_read(&regs->rqfpr); 1177 *fpr = gfar_read(&regs->rqfpr);
1178} 1178}
1179 1179
1180extern void lock_rx_qs(struct gfar_private *priv); 1180void lock_rx_qs(struct gfar_private *priv);
1181extern void lock_tx_qs(struct gfar_private *priv); 1181void lock_tx_qs(struct gfar_private *priv);
1182extern void unlock_rx_qs(struct gfar_private *priv); 1182void unlock_rx_qs(struct gfar_private *priv);
1183extern void unlock_tx_qs(struct gfar_private *priv); 1183void unlock_tx_qs(struct gfar_private *priv);
1184extern irqreturn_t gfar_receive(int irq, void *dev_id); 1184irqreturn_t gfar_receive(int irq, void *dev_id);
1185extern int startup_gfar(struct net_device *dev); 1185int startup_gfar(struct net_device *dev);
1186extern void stop_gfar(struct net_device *dev); 1186void stop_gfar(struct net_device *dev);
1187extern void gfar_halt(struct net_device *dev); 1187void gfar_halt(struct net_device *dev);
1188extern void gfar_phy_test(struct mii_bus *bus, struct phy_device *phydev, 1188void gfar_phy_test(struct mii_bus *bus, struct phy_device *phydev, int enable,
1189 int enable, u32 regnum, u32 read); 1189 u32 regnum, u32 read);
1190extern void gfar_configure_coalescing_all(struct gfar_private *priv); 1190void gfar_configure_coalescing_all(struct gfar_private *priv);
1191void gfar_init_sysfs(struct net_device *dev); 1191void gfar_init_sysfs(struct net_device *dev);
1192int gfar_set_features(struct net_device *dev, netdev_features_t features); 1192int gfar_set_features(struct net_device *dev, netdev_features_t features);
1193extern void gfar_check_rx_parser_mode(struct gfar_private *priv); 1193void gfar_check_rx_parser_mode(struct gfar_private *priv);
1194extern void gfar_vlan_mode(struct net_device *dev, netdev_features_t features); 1194void gfar_vlan_mode(struct net_device *dev, netdev_features_t features);
1195 1195
1196extern const struct ethtool_ops gfar_ethtool_ops; 1196extern const struct ethtool_ops gfar_ethtool_ops;
1197 1197
diff --git a/drivers/net/ethernet/freescale/ucc_geth.c b/drivers/net/ethernet/freescale/ucc_geth.c
index 64b329fecf3d..5548b6d00c31 100644
--- a/drivers/net/ethernet/freescale/ucc_geth.c
+++ b/drivers/net/ethernet/freescale/ucc_geth.c
@@ -3901,7 +3901,7 @@ static int ucc_geth_probe(struct platform_device* ofdev)
3901 3901
3902 mac_addr = of_get_mac_address(np); 3902 mac_addr = of_get_mac_address(np);
3903 if (mac_addr) 3903 if (mac_addr)
3904 memcpy(dev->dev_addr, mac_addr, 6); 3904 memcpy(dev->dev_addr, mac_addr, ETH_ALEN);
3905 3905
3906 ugeth->ug_info = ug_info; 3906 ugeth->ug_info = ug_info;
3907 ugeth->dev = device; 3907 ugeth->dev = device;
diff --git a/drivers/net/ethernet/fujitsu/Kconfig b/drivers/net/ethernet/fujitsu/Kconfig
index 6231bc02b964..1085257385d2 100644
--- a/drivers/net/ethernet/fujitsu/Kconfig
+++ b/drivers/net/ethernet/fujitsu/Kconfig
@@ -5,7 +5,7 @@
5config NET_VENDOR_FUJITSU 5config NET_VENDOR_FUJITSU
6 bool "Fujitsu devices" 6 bool "Fujitsu devices"
7 default y 7 default y
8 depends on ISA || PCMCIA 8 depends on PCMCIA
9 ---help--- 9 ---help---
10 If you have a network (Ethernet) card belonging to this class, say Y 10 If you have a network (Ethernet) card belonging to this class, say Y
11 and read the Ethernet-HOWTO, available from 11 and read the Ethernet-HOWTO, available from
diff --git a/drivers/net/ethernet/hp/hp100.c b/drivers/net/ethernet/hp/hp100.c
index 91227d03274e..37860096f744 100644
--- a/drivers/net/ethernet/hp/hp100.c
+++ b/drivers/net/ethernet/hp/hp100.c
@@ -1098,7 +1098,7 @@ static int hp100_open(struct net_device *dev)
1098 if (request_irq(dev->irq, hp100_interrupt, 1098 if (request_irq(dev->irq, hp100_interrupt,
1099 lp->bus == HP100_BUS_PCI || lp->bus == 1099 lp->bus == HP100_BUS_PCI || lp->bus ==
1100 HP100_BUS_EISA ? IRQF_SHARED : 0, 1100 HP100_BUS_EISA ? IRQF_SHARED : 0,
1101 "hp100", dev)) { 1101 dev->name, dev)) {
1102 printk("hp100: %s: unable to get IRQ %d\n", dev->name, dev->irq); 1102 printk("hp100: %s: unable to get IRQ %d\n", dev->name, dev->irq);
1103 return -EAGAIN; 1103 return -EAGAIN;
1104 } 1104 }
diff --git a/drivers/net/ethernet/i825xx/82596.c b/drivers/net/ethernet/i825xx/82596.c
index e38816145395..7ce6379fd1a3 100644
--- a/drivers/net/ethernet/i825xx/82596.c
+++ b/drivers/net/ethernet/i825xx/82596.c
@@ -711,7 +711,7 @@ static int init_i596_mem(struct net_device *dev)
711 i596_add_cmd(dev, &lp->cf_cmd.cmd); 711 i596_add_cmd(dev, &lp->cf_cmd.cmd);
712 712
713 DEB(DEB_INIT,printk(KERN_DEBUG "%s: queuing CmdSASetup\n", dev->name)); 713 DEB(DEB_INIT,printk(KERN_DEBUG "%s: queuing CmdSASetup\n", dev->name));
714 memcpy(lp->sa_cmd.eth_addr, dev->dev_addr, 6); 714 memcpy(lp->sa_cmd.eth_addr, dev->dev_addr, ETH_ALEN);
715 lp->sa_cmd.cmd.command = CmdSASetup; 715 lp->sa_cmd.cmd.command = CmdSASetup;
716 i596_add_cmd(dev, &lp->sa_cmd.cmd); 716 i596_add_cmd(dev, &lp->sa_cmd.cmd);
717 717
@@ -1155,7 +1155,7 @@ struct net_device * __init i82596_probe(int unit)
1155 err = -ENODEV; 1155 err = -ENODEV;
1156 goto out; 1156 goto out;
1157 } 1157 }
1158 memcpy(eth_addr, (void *) 0xfffc1f2c, 6); /* YUCK! Get addr from NOVRAM */ 1158 memcpy(eth_addr, (void *) 0xfffc1f2c, ETH_ALEN); /* YUCK! Get addr from NOVRAM */
1159 dev->base_addr = MVME_I596_BASE; 1159 dev->base_addr = MVME_I596_BASE;
1160 dev->irq = (unsigned) MVME16x_IRQ_I596; 1160 dev->irq = (unsigned) MVME16x_IRQ_I596;
1161 goto found; 1161 goto found;
@@ -1527,9 +1527,7 @@ int __init init_module(void)
1527 if (debug >= 0) 1527 if (debug >= 0)
1528 i596_debug = debug; 1528 i596_debug = debug;
1529 dev_82596 = i82596_probe(-1); 1529 dev_82596 = i82596_probe(-1);
1530 if (IS_ERR(dev_82596)) 1530 return PTR_ERR_OR_ZERO(dev_82596);
1531 return PTR_ERR(dev_82596);
1532 return 0;
1533} 1531}
1534 1532
1535void __exit cleanup_module(void) 1533void __exit cleanup_module(void)
diff --git a/drivers/net/ethernet/i825xx/lib82596.c b/drivers/net/ethernet/i825xx/lib82596.c
index d653bac4cfc4..861fa15e1e81 100644
--- a/drivers/net/ethernet/i825xx/lib82596.c
+++ b/drivers/net/ethernet/i825xx/lib82596.c
@@ -607,7 +607,7 @@ static int init_i596_mem(struct net_device *dev)
607 i596_add_cmd(dev, &dma->cf_cmd.cmd); 607 i596_add_cmd(dev, &dma->cf_cmd.cmd);
608 608
609 DEB(DEB_INIT, printk(KERN_DEBUG "%s: queuing CmdSASetup\n", dev->name)); 609 DEB(DEB_INIT, printk(KERN_DEBUG "%s: queuing CmdSASetup\n", dev->name));
610 memcpy(dma->sa_cmd.eth_addr, dev->dev_addr, 6); 610 memcpy(dma->sa_cmd.eth_addr, dev->dev_addr, ETH_ALEN);
611 dma->sa_cmd.cmd.command = SWAP16(CmdSASetup); 611 dma->sa_cmd.cmd.command = SWAP16(CmdSASetup);
612 DMA_WBACK(dev, &(dma->sa_cmd), sizeof(struct sa_cmd)); 612 DMA_WBACK(dev, &(dma->sa_cmd), sizeof(struct sa_cmd));
613 i596_add_cmd(dev, &dma->sa_cmd.cmd); 613 i596_add_cmd(dev, &dma->sa_cmd.cmd);
@@ -1396,13 +1396,13 @@ static void set_multicast_list(struct net_device *dev)
1396 netdev_for_each_mc_addr(ha, dev) { 1396 netdev_for_each_mc_addr(ha, dev) {
1397 if (!cnt--) 1397 if (!cnt--)
1398 break; 1398 break;
1399 memcpy(cp, ha->addr, 6); 1399 memcpy(cp, ha->addr, ETH_ALEN);
1400 if (i596_debug > 1) 1400 if (i596_debug > 1)
1401 DEB(DEB_MULTI, 1401 DEB(DEB_MULTI,
1402 printk(KERN_DEBUG 1402 printk(KERN_DEBUG
1403 "%s: Adding address %pM\n", 1403 "%s: Adding address %pM\n",
1404 dev->name, cp)); 1404 dev->name, cp));
1405 cp += 6; 1405 cp += ETH_ALEN;
1406 } 1406 }
1407 DMA_WBACK_INV(dev, &dma->mc_cmd, sizeof(struct mc_cmd)); 1407 DMA_WBACK_INV(dev, &dma->mc_cmd, sizeof(struct mc_cmd));
1408 i596_add_cmd(dev, &cmd->cmd); 1408 i596_add_cmd(dev, &cmd->cmd);
diff --git a/drivers/net/ethernet/ibm/emac/core.c b/drivers/net/ethernet/ibm/emac/core.c
index cdf2321f1f0f..ae342fdb42c8 100644
--- a/drivers/net/ethernet/ibm/emac/core.c
+++ b/drivers/net/ethernet/ibm/emac/core.c
@@ -2678,7 +2678,7 @@ static int emac_init_config(struct emac_instance *dev)
2678 np->full_name); 2678 np->full_name);
2679 return -ENXIO; 2679 return -ENXIO;
2680 } 2680 }
2681 memcpy(dev->ndev->dev_addr, p, 6); 2681 memcpy(dev->ndev->dev_addr, p, ETH_ALEN);
2682 2682
2683 /* IAHT and GAHT filter parameterization */ 2683 /* IAHT and GAHT filter parameterization */
2684 if (emac_has_feature(dev, EMAC_FTR_EMAC4SYNC)) { 2684 if (emac_has_feature(dev, EMAC_FTR_EMAC4SYNC)) {
diff --git a/drivers/net/ethernet/ibm/emac/debug.h b/drivers/net/ethernet/ibm/emac/debug.h
index 59a92d5870b5..9c45efe4c8fe 100644
--- a/drivers/net/ethernet/ibm/emac/debug.h
+++ b/drivers/net/ethernet/ibm/emac/debug.h
@@ -29,13 +29,13 @@
29struct emac_instance; 29struct emac_instance;
30struct mal_instance; 30struct mal_instance;
31 31
32extern void emac_dbg_register(struct emac_instance *dev); 32void emac_dbg_register(struct emac_instance *dev);
33extern void emac_dbg_unregister(struct emac_instance *dev); 33void emac_dbg_unregister(struct emac_instance *dev);
34extern void mal_dbg_register(struct mal_instance *mal); 34void mal_dbg_register(struct mal_instance *mal);
35extern void mal_dbg_unregister(struct mal_instance *mal); 35void mal_dbg_unregister(struct mal_instance *mal);
36extern int emac_init_debug(void) __init; 36int emac_init_debug(void) __init;
37extern void emac_fini_debug(void) __exit; 37void emac_fini_debug(void) __exit;
38extern void emac_dbg_dump_all(void); 38void emac_dbg_dump_all(void);
39 39
40# define DBG_LEVEL 1 40# define DBG_LEVEL 1
41 41
diff --git a/drivers/net/ethernet/ibm/emac/rgmii.h b/drivers/net/ethernet/ibm/emac/rgmii.h
index 668bceeff4a2..d4f1374d1900 100644
--- a/drivers/net/ethernet/ibm/emac/rgmii.h
+++ b/drivers/net/ethernet/ibm/emac/rgmii.h
@@ -56,15 +56,15 @@ struct rgmii_instance {
56 56
57#ifdef CONFIG_IBM_EMAC_RGMII 57#ifdef CONFIG_IBM_EMAC_RGMII
58 58
59extern int rgmii_init(void); 59int rgmii_init(void);
60extern void rgmii_exit(void); 60void rgmii_exit(void);
61extern int rgmii_attach(struct platform_device *ofdev, int input, int mode); 61int rgmii_attach(struct platform_device *ofdev, int input, int mode);
62extern void rgmii_detach(struct platform_device *ofdev, int input); 62void rgmii_detach(struct platform_device *ofdev, int input);
63extern void rgmii_get_mdio(struct platform_device *ofdev, int input); 63void rgmii_get_mdio(struct platform_device *ofdev, int input);
64extern void rgmii_put_mdio(struct platform_device *ofdev, int input); 64void rgmii_put_mdio(struct platform_device *ofdev, int input);
65extern void rgmii_set_speed(struct platform_device *ofdev, int input, int speed); 65void rgmii_set_speed(struct platform_device *ofdev, int input, int speed);
66extern int rgmii_get_regs_len(struct platform_device *ofdev); 66int rgmii_get_regs_len(struct platform_device *ofdev);
67extern void *rgmii_dump_regs(struct platform_device *ofdev, void *buf); 67void *rgmii_dump_regs(struct platform_device *ofdev, void *buf);
68 68
69#else 69#else
70 70
diff --git a/drivers/net/ethernet/ibm/emac/tah.h b/drivers/net/ethernet/ibm/emac/tah.h
index 350b7096a041..4d5f336f07b3 100644
--- a/drivers/net/ethernet/ibm/emac/tah.h
+++ b/drivers/net/ethernet/ibm/emac/tah.h
@@ -72,13 +72,13 @@ struct tah_instance {
72 72
73#ifdef CONFIG_IBM_EMAC_TAH 73#ifdef CONFIG_IBM_EMAC_TAH
74 74
75extern int tah_init(void); 75int tah_init(void);
76extern void tah_exit(void); 76void tah_exit(void);
77extern int tah_attach(struct platform_device *ofdev, int channel); 77int tah_attach(struct platform_device *ofdev, int channel);
78extern void tah_detach(struct platform_device *ofdev, int channel); 78void tah_detach(struct platform_device *ofdev, int channel);
79extern void tah_reset(struct platform_device *ofdev); 79void tah_reset(struct platform_device *ofdev);
80extern int tah_get_regs_len(struct platform_device *ofdev); 80int tah_get_regs_len(struct platform_device *ofdev);
81extern void *tah_dump_regs(struct platform_device *ofdev, void *buf); 81void *tah_dump_regs(struct platform_device *ofdev, void *buf);
82 82
83#else 83#else
84 84
diff --git a/drivers/net/ethernet/ibm/emac/zmii.h b/drivers/net/ethernet/ibm/emac/zmii.h
index 455bfb085493..0959c55b1459 100644
--- a/drivers/net/ethernet/ibm/emac/zmii.h
+++ b/drivers/net/ethernet/ibm/emac/zmii.h
@@ -53,15 +53,15 @@ struct zmii_instance {
53 53
54#ifdef CONFIG_IBM_EMAC_ZMII 54#ifdef CONFIG_IBM_EMAC_ZMII
55 55
56extern int zmii_init(void); 56int zmii_init(void);
57extern void zmii_exit(void); 57void zmii_exit(void);
58extern int zmii_attach(struct platform_device *ofdev, int input, int *mode); 58int zmii_attach(struct platform_device *ofdev, int input, int *mode);
59extern void zmii_detach(struct platform_device *ofdev, int input); 59void zmii_detach(struct platform_device *ofdev, int input);
60extern void zmii_get_mdio(struct platform_device *ofdev, int input); 60void zmii_get_mdio(struct platform_device *ofdev, int input);
61extern void zmii_put_mdio(struct platform_device *ofdev, int input); 61void zmii_put_mdio(struct platform_device *ofdev, int input);
62extern void zmii_set_speed(struct platform_device *ofdev, int input, int speed); 62void zmii_set_speed(struct platform_device *ofdev, int input, int speed);
63extern int zmii_get_regs_len(struct platform_device *ocpdev); 63int zmii_get_regs_len(struct platform_device *ocpdev);
64extern void *zmii_dump_regs(struct platform_device *ofdev, void *buf); 64void *zmii_dump_regs(struct platform_device *ofdev, void *buf);
65 65
66#else 66#else
67# define zmii_init() 0 67# define zmii_init() 0
diff --git a/drivers/net/ethernet/ibm/ibmveth.c b/drivers/net/ethernet/ibm/ibmveth.c
index 5d41aee69d16..952d795230a4 100644
--- a/drivers/net/ethernet/ibm/ibmveth.c
+++ b/drivers/net/ethernet/ibm/ibmveth.c
@@ -1185,7 +1185,7 @@ static void ibmveth_set_multicast_list(struct net_device *netdev)
1185 netdev_for_each_mc_addr(ha, netdev) { 1185 netdev_for_each_mc_addr(ha, netdev) {
1186 /* add the multicast address to the filter table */ 1186 /* add the multicast address to the filter table */
1187 unsigned long mcast_addr = 0; 1187 unsigned long mcast_addr = 0;
1188 memcpy(((char *)&mcast_addr)+2, ha->addr, 6); 1188 memcpy(((char *)&mcast_addr)+2, ha->addr, ETH_ALEN);
1189 lpar_rc = h_multicast_ctrl(adapter->vdev->unit_address, 1189 lpar_rc = h_multicast_ctrl(adapter->vdev->unit_address,
1190 IbmVethMcastAddFilter, 1190 IbmVethMcastAddFilter,
1191 mcast_addr); 1191 mcast_addr);
@@ -1370,7 +1370,7 @@ static int ibmveth_probe(struct vio_dev *dev, const struct vio_device_id *id)
1370 netif_napi_add(netdev, &adapter->napi, ibmveth_poll, 16); 1370 netif_napi_add(netdev, &adapter->napi, ibmveth_poll, 16);
1371 1371
1372 adapter->mac_addr = 0; 1372 adapter->mac_addr = 0;
1373 memcpy(&adapter->mac_addr, mac_addr_p, 6); 1373 memcpy(&adapter->mac_addr, mac_addr_p, ETH_ALEN);
1374 1374
1375 netdev->irq = dev->irq; 1375 netdev->irq = dev->irq;
1376 netdev->netdev_ops = &ibmveth_netdev_ops; 1376 netdev->netdev_ops = &ibmveth_netdev_ops;
diff --git a/drivers/net/ethernet/icplus/ipg.c b/drivers/net/ethernet/icplus/ipg.c
index bdf5023724e7..25045ae07171 100644
--- a/drivers/net/ethernet/icplus/ipg.c
+++ b/drivers/net/ethernet/icplus/ipg.c
@@ -2183,7 +2183,6 @@ static void ipg_remove(struct pci_dev *pdev)
2183 2183
2184 free_netdev(dev); 2184 free_netdev(dev);
2185 pci_disable_device(pdev); 2185 pci_disable_device(pdev);
2186 pci_set_drvdata(pdev, NULL);
2187} 2186}
2188 2187
2189static const struct net_device_ops ipg_netdev_ops = { 2188static const struct net_device_ops ipg_netdev_ops = {
diff --git a/drivers/net/ethernet/intel/e100.c b/drivers/net/ethernet/intel/e100.c
index ada6e210279f..cbaba4442d4b 100644
--- a/drivers/net/ethernet/intel/e100.c
+++ b/drivers/net/ethernet/intel/e100.c
@@ -2985,7 +2985,6 @@ err_out_free_res:
2985err_out_disable_pdev: 2985err_out_disable_pdev:
2986 pci_disable_device(pdev); 2986 pci_disable_device(pdev);
2987err_out_free_dev: 2987err_out_free_dev:
2988 pci_set_drvdata(pdev, NULL);
2989 free_netdev(netdev); 2988 free_netdev(netdev);
2990 return err; 2989 return err;
2991} 2990}
@@ -3003,7 +3002,6 @@ static void e100_remove(struct pci_dev *pdev)
3003 free_netdev(netdev); 3002 free_netdev(netdev);
3004 pci_release_regions(pdev); 3003 pci_release_regions(pdev);
3005 pci_disable_device(pdev); 3004 pci_disable_device(pdev);
3006 pci_set_drvdata(pdev, NULL);
3007 } 3005 }
3008} 3006}
3009 3007
diff --git a/drivers/net/ethernet/intel/e1000/e1000.h b/drivers/net/ethernet/intel/e1000/e1000.h
index 26d9cd59ec75..58c147271a36 100644
--- a/drivers/net/ethernet/intel/e1000/e1000.h
+++ b/drivers/net/ethernet/intel/e1000/e1000.h
@@ -325,7 +325,7 @@ enum e1000_state_t {
325#undef pr_fmt 325#undef pr_fmt
326#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt 326#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
327 327
328extern struct net_device *e1000_get_hw_dev(struct e1000_hw *hw); 328struct net_device *e1000_get_hw_dev(struct e1000_hw *hw);
329#define e_dbg(format, arg...) \ 329#define e_dbg(format, arg...) \
330 netdev_dbg(e1000_get_hw_dev(hw), format, ## arg) 330 netdev_dbg(e1000_get_hw_dev(hw), format, ## arg)
331#define e_err(msglvl, format, arg...) \ 331#define e_err(msglvl, format, arg...) \
@@ -346,20 +346,20 @@ extern struct net_device *e1000_get_hw_dev(struct e1000_hw *hw);
346extern char e1000_driver_name[]; 346extern char e1000_driver_name[];
347extern const char e1000_driver_version[]; 347extern const char e1000_driver_version[];
348 348
349extern int e1000_up(struct e1000_adapter *adapter); 349int e1000_up(struct e1000_adapter *adapter);
350extern void e1000_down(struct e1000_adapter *adapter); 350void e1000_down(struct e1000_adapter *adapter);
351extern void e1000_reinit_locked(struct e1000_adapter *adapter); 351void e1000_reinit_locked(struct e1000_adapter *adapter);
352extern void e1000_reset(struct e1000_adapter *adapter); 352void e1000_reset(struct e1000_adapter *adapter);
353extern int e1000_set_spd_dplx(struct e1000_adapter *adapter, u32 spd, u8 dplx); 353int e1000_set_spd_dplx(struct e1000_adapter *adapter, u32 spd, u8 dplx);
354extern int e1000_setup_all_rx_resources(struct e1000_adapter *adapter); 354int e1000_setup_all_rx_resources(struct e1000_adapter *adapter);
355extern int e1000_setup_all_tx_resources(struct e1000_adapter *adapter); 355int e1000_setup_all_tx_resources(struct e1000_adapter *adapter);
356extern void e1000_free_all_rx_resources(struct e1000_adapter *adapter); 356void e1000_free_all_rx_resources(struct e1000_adapter *adapter);
357extern void e1000_free_all_tx_resources(struct e1000_adapter *adapter); 357void e1000_free_all_tx_resources(struct e1000_adapter *adapter);
358extern void e1000_update_stats(struct e1000_adapter *adapter); 358void e1000_update_stats(struct e1000_adapter *adapter);
359extern bool e1000_has_link(struct e1000_adapter *adapter); 359bool e1000_has_link(struct e1000_adapter *adapter);
360extern void e1000_power_up_phy(struct e1000_adapter *); 360void e1000_power_up_phy(struct e1000_adapter *);
361extern void e1000_set_ethtool_ops(struct net_device *netdev); 361void e1000_set_ethtool_ops(struct net_device *netdev);
362extern void e1000_check_options(struct e1000_adapter *adapter); 362void e1000_check_options(struct e1000_adapter *adapter);
363extern char *e1000_get_hw_dev_name(struct e1000_hw *hw); 363char *e1000_get_hw_dev_name(struct e1000_hw *hw);
364 364
365#endif /* _E1000_H_ */ 365#endif /* _E1000_H_ */
diff --git a/drivers/net/ethernet/intel/e1000/e1000_main.c b/drivers/net/ethernet/intel/e1000/e1000_main.c
index 59ad007dd5aa..ad6800ad1bfc 100644
--- a/drivers/net/ethernet/intel/e1000/e1000_main.c
+++ b/drivers/net/ethernet/intel/e1000/e1000_main.c
@@ -3917,8 +3917,7 @@ static bool e1000_clean_tx_irq(struct e1000_adapter *adapter,
3917 " next_to_watch <%x>\n" 3917 " next_to_watch <%x>\n"
3918 " jiffies <%lx>\n" 3918 " jiffies <%lx>\n"
3919 " next_to_watch.status <%x>\n", 3919 " next_to_watch.status <%x>\n",
3920 (unsigned long)((tx_ring - adapter->tx_ring) / 3920 (unsigned long)(tx_ring - adapter->tx_ring),
3921 sizeof(struct e1000_tx_ring)),
3922 readl(hw->hw_addr + tx_ring->tdh), 3921 readl(hw->hw_addr + tx_ring->tdh),
3923 readl(hw->hw_addr + tx_ring->tdt), 3922 readl(hw->hw_addr + tx_ring->tdt),
3924 tx_ring->next_to_use, 3923 tx_ring->next_to_use,
diff --git a/drivers/net/ethernet/intel/e1000e/e1000.h b/drivers/net/ethernet/intel/e1000e/e1000.h
index ad0edd11015d..0150f7fc893d 100644
--- a/drivers/net/ethernet/intel/e1000e/e1000.h
+++ b/drivers/net/ethernet/intel/e1000e/e1000.h
@@ -472,26 +472,25 @@ enum latency_range {
472extern char e1000e_driver_name[]; 472extern char e1000e_driver_name[];
473extern const char e1000e_driver_version[]; 473extern const char e1000e_driver_version[];
474 474
475extern void e1000e_check_options(struct e1000_adapter *adapter); 475void e1000e_check_options(struct e1000_adapter *adapter);
476extern void e1000e_set_ethtool_ops(struct net_device *netdev); 476void e1000e_set_ethtool_ops(struct net_device *netdev);
477 477
478extern int e1000e_up(struct e1000_adapter *adapter); 478int e1000e_up(struct e1000_adapter *adapter);
479extern void e1000e_down(struct e1000_adapter *adapter); 479void e1000e_down(struct e1000_adapter *adapter);
480extern void e1000e_reinit_locked(struct e1000_adapter *adapter); 480void e1000e_reinit_locked(struct e1000_adapter *adapter);
481extern void e1000e_reset(struct e1000_adapter *adapter); 481void e1000e_reset(struct e1000_adapter *adapter);
482extern void e1000e_power_up_phy(struct e1000_adapter *adapter); 482void e1000e_power_up_phy(struct e1000_adapter *adapter);
483extern int e1000e_setup_rx_resources(struct e1000_ring *ring); 483int e1000e_setup_rx_resources(struct e1000_ring *ring);
484extern int e1000e_setup_tx_resources(struct e1000_ring *ring); 484int e1000e_setup_tx_resources(struct e1000_ring *ring);
485extern void e1000e_free_rx_resources(struct e1000_ring *ring); 485void e1000e_free_rx_resources(struct e1000_ring *ring);
486extern void e1000e_free_tx_resources(struct e1000_ring *ring); 486void e1000e_free_tx_resources(struct e1000_ring *ring);
487extern struct rtnl_link_stats64 *e1000e_get_stats64(struct net_device *netdev, 487struct rtnl_link_stats64 *e1000e_get_stats64(struct net_device *netdev,
488 struct rtnl_link_stats64 488 struct rtnl_link_stats64 *stats);
489 *stats); 489void e1000e_set_interrupt_capability(struct e1000_adapter *adapter);
490extern void e1000e_set_interrupt_capability(struct e1000_adapter *adapter); 490void e1000e_reset_interrupt_capability(struct e1000_adapter *adapter);
491extern void e1000e_reset_interrupt_capability(struct e1000_adapter *adapter); 491void e1000e_get_hw_control(struct e1000_adapter *adapter);
492extern void e1000e_get_hw_control(struct e1000_adapter *adapter); 492void e1000e_release_hw_control(struct e1000_adapter *adapter);
493extern void e1000e_release_hw_control(struct e1000_adapter *adapter); 493void e1000e_write_itr(struct e1000_adapter *adapter, u32 itr);
494extern void e1000e_write_itr(struct e1000_adapter *adapter, u32 itr);
495 494
496extern unsigned int copybreak; 495extern unsigned int copybreak;
497 496
@@ -508,8 +507,8 @@ extern const struct e1000_info e1000_pch2_info;
508extern const struct e1000_info e1000_pch_lpt_info; 507extern const struct e1000_info e1000_pch_lpt_info;
509extern const struct e1000_info e1000_es2_info; 508extern const struct e1000_info e1000_es2_info;
510 509
511extern void e1000e_ptp_init(struct e1000_adapter *adapter); 510void e1000e_ptp_init(struct e1000_adapter *adapter);
512extern void e1000e_ptp_remove(struct e1000_adapter *adapter); 511void e1000e_ptp_remove(struct e1000_adapter *adapter);
513 512
514static inline s32 e1000_phy_hw_reset(struct e1000_hw *hw) 513static inline s32 e1000_phy_hw_reset(struct e1000_hw *hw)
515{ 514{
@@ -536,7 +535,7 @@ static inline s32 e1e_wphy_locked(struct e1000_hw *hw, u32 offset, u16 data)
536 return hw->phy.ops.write_reg_locked(hw, offset, data); 535 return hw->phy.ops.write_reg_locked(hw, offset, data);
537} 536}
538 537
539extern void e1000e_reload_nvm_generic(struct e1000_hw *hw); 538void e1000e_reload_nvm_generic(struct e1000_hw *hw);
540 539
541static inline s32 e1000e_read_mac_addr(struct e1000_hw *hw) 540static inline s32 e1000e_read_mac_addr(struct e1000_hw *hw)
542{ 541{
diff --git a/drivers/net/ethernet/intel/i40e/i40e.h b/drivers/net/ethernet/intel/i40e/i40e.h
index b5252eb8a6c7..1ca9834cdfda 100644
--- a/drivers/net/ethernet/intel/i40e/i40e.h
+++ b/drivers/net/ethernet/intel/i40e/i40e.h
@@ -46,7 +46,6 @@
46#include <linux/sctp.h> 46#include <linux/sctp.h>
47#include <linux/pkt_sched.h> 47#include <linux/pkt_sched.h>
48#include <linux/ipv6.h> 48#include <linux/ipv6.h>
49#include <linux/version.h>
50#include <net/checksum.h> 49#include <net/checksum.h>
51#include <net/ip6_checksum.h> 50#include <net/ip6_checksum.h>
52#include <linux/ethtool.h> 51#include <linux/ethtool.h>
@@ -347,9 +346,9 @@ struct i40e_vsi {
347 u32 rx_buf_failed; 346 u32 rx_buf_failed;
348 u32 rx_page_failed; 347 u32 rx_page_failed;
349 348
350 /* These are arrays of rings, allocated at run-time */ 349 /* These are containers of ring pointers, allocated at run-time */
351 struct i40e_ring *rx_rings; 350 struct i40e_ring **rx_rings;
352 struct i40e_ring *tx_rings; 351 struct i40e_ring **tx_rings;
353 352
354 u16 work_limit; 353 u16 work_limit;
355 /* high bit set means dynamic, use accessor routines to read/write. 354 /* high bit set means dynamic, use accessor routines to read/write.
@@ -366,7 +365,7 @@ struct i40e_vsi {
366 u8 dtype; 365 u8 dtype;
367 366
368 /* List of q_vectors allocated to this VSI */ 367 /* List of q_vectors allocated to this VSI */
369 struct i40e_q_vector *q_vectors; 368 struct i40e_q_vector **q_vectors;
370 int num_q_vectors; 369 int num_q_vectors;
371 int base_vector; 370 int base_vector;
372 371
@@ -422,8 +421,9 @@ struct i40e_q_vector {
422 421
423 u8 num_ringpairs; /* total number of ring pairs in vector */ 422 u8 num_ringpairs; /* total number of ring pairs in vector */
424 423
425 char name[IFNAMSIZ + 9];
426 cpumask_t affinity_mask; 424 cpumask_t affinity_mask;
425 struct rcu_head rcu; /* to avoid race with update stats on free */
426 char name[IFNAMSIZ + 9];
427} ____cacheline_internodealigned_in_smp; 427} ____cacheline_internodealigned_in_smp;
428 428
429/* lan device */ 429/* lan device */
@@ -544,6 +544,7 @@ static inline void i40e_dbg_init(void) {}
544static inline void i40e_dbg_exit(void) {} 544static inline void i40e_dbg_exit(void) {}
545#endif /* CONFIG_DEBUG_FS*/ 545#endif /* CONFIG_DEBUG_FS*/
546void i40e_irq_dynamic_enable(struct i40e_vsi *vsi, int vector); 546void i40e_irq_dynamic_enable(struct i40e_vsi *vsi, int vector);
547void i40e_irq_dynamic_enable_icr0(struct i40e_pf *pf);
547int i40e_ioctl(struct net_device *netdev, struct ifreq *ifr, int cmd); 548int i40e_ioctl(struct net_device *netdev, struct ifreq *ifr, int cmd);
548void i40e_vlan_stripping_disable(struct i40e_vsi *vsi); 549void i40e_vlan_stripping_disable(struct i40e_vsi *vsi);
549int i40e_vsi_add_vlan(struct i40e_vsi *vsi, s16 vid); 550int i40e_vsi_add_vlan(struct i40e_vsi *vsi, s16 vid);
diff --git a/drivers/net/ethernet/intel/i40e/i40e_debugfs.c b/drivers/net/ethernet/intel/i40e/i40e_debugfs.c
index 8dbd91f64b74..ef4cb1cf31f2 100644
--- a/drivers/net/ethernet/intel/i40e/i40e_debugfs.c
+++ b/drivers/net/ethernet/intel/i40e/i40e_debugfs.c
@@ -151,9 +151,7 @@ static ssize_t i40e_dbg_dump_write(struct file *filp,
151 size_t count, loff_t *ppos) 151 size_t count, loff_t *ppos)
152{ 152{
153 struct i40e_pf *pf = filp->private_data; 153 struct i40e_pf *pf = filp->private_data;
154 char dump_request_buf[16];
155 bool seid_found = false; 154 bool seid_found = false;
156 int bytes_not_copied;
157 long seid = -1; 155 long seid = -1;
158 int buflen = 0; 156 int buflen = 0;
159 int i, ret; 157 int i, ret;
@@ -163,21 +161,12 @@ static ssize_t i40e_dbg_dump_write(struct file *filp,
163 /* don't allow partial writes */ 161 /* don't allow partial writes */
164 if (*ppos != 0) 162 if (*ppos != 0)
165 return 0; 163 return 0;
166 if (count >= sizeof(dump_request_buf))
167 return -ENOSPC;
168
169 bytes_not_copied = copy_from_user(dump_request_buf, buffer, count);
170 if (bytes_not_copied < 0)
171 return bytes_not_copied;
172 if (bytes_not_copied > 0)
173 count -= bytes_not_copied;
174 dump_request_buf[count] = '\0';
175 164
176 /* decode the SEID given to be dumped */ 165 /* decode the SEID given to be dumped */
177 ret = kstrtol(dump_request_buf, 0, &seid); 166 ret = kstrtol_from_user(buffer, count, 0, &seid);
178 if (ret < 0) { 167
179 dev_info(&pf->pdev->dev, "bad seid value '%s'\n", 168 if (ret) {
180 dump_request_buf); 169 dev_info(&pf->pdev->dev, "bad seid value\n");
181 } else if (seid == 0) { 170 } else if (seid == 0) {
182 seid_found = true; 171 seid_found = true;
183 172
@@ -245,26 +234,33 @@ static ssize_t i40e_dbg_dump_write(struct file *filp,
245 memcpy(p, vsi, len); 234 memcpy(p, vsi, len);
246 p += len; 235 p += len;
247 236
248 len = (sizeof(struct i40e_q_vector) 237 if (vsi->num_q_vectors) {
249 * vsi->num_q_vectors); 238 len = (sizeof(struct i40e_q_vector)
250 memcpy(p, vsi->q_vectors, len); 239 * vsi->num_q_vectors);
251 p += len; 240 memcpy(p, vsi->q_vectors, len);
252 241 p += len;
253 len = (sizeof(struct i40e_ring) * vsi->num_queue_pairs); 242 }
254 memcpy(p, vsi->tx_rings, len);
255 p += len;
256 memcpy(p, vsi->rx_rings, len);
257 p += len;
258 243
259 for (i = 0; i < vsi->num_queue_pairs; i++) { 244 if (vsi->num_queue_pairs) {
260 len = sizeof(struct i40e_tx_buffer); 245 len = (sizeof(struct i40e_ring) *
261 memcpy(p, vsi->tx_rings[i].tx_bi, len); 246 vsi->num_queue_pairs);
247 memcpy(p, vsi->tx_rings, len);
248 p += len;
249 memcpy(p, vsi->rx_rings, len);
262 p += len; 250 p += len;
263 } 251 }
264 for (i = 0; i < vsi->num_queue_pairs; i++) { 252
253 if (vsi->tx_rings[0]) {
254 len = sizeof(struct i40e_tx_buffer);
255 for (i = 0; i < vsi->num_queue_pairs; i++) {
256 memcpy(p, vsi->tx_rings[i]->tx_bi, len);
257 p += len;
258 }
265 len = sizeof(struct i40e_rx_buffer); 259 len = sizeof(struct i40e_rx_buffer);
266 memcpy(p, vsi->rx_rings[i].rx_bi, len); 260 for (i = 0; i < vsi->num_queue_pairs; i++) {
267 p += len; 261 memcpy(p, vsi->rx_rings[i]->rx_bi, len);
262 p += len;
263 }
268 } 264 }
269 265
270 /* macvlan filter list */ 266 /* macvlan filter list */
@@ -484,100 +480,104 @@ static void i40e_dbg_dump_vsi_seid(struct i40e_pf *pf, int seid)
484 " tx_restart = %d, tx_busy = %d, rx_buf_failed = %d, rx_page_failed = %d\n", 480 " tx_restart = %d, tx_busy = %d, rx_buf_failed = %d, rx_page_failed = %d\n",
485 vsi->tx_restart, vsi->tx_busy, 481 vsi->tx_restart, vsi->tx_busy,
486 vsi->rx_buf_failed, vsi->rx_page_failed); 482 vsi->rx_buf_failed, vsi->rx_page_failed);
487 if (vsi->rx_rings) { 483 rcu_read_lock();
488 for (i = 0; i < vsi->num_queue_pairs; i++) { 484 for (i = 0; i < vsi->num_queue_pairs; i++) {
489 dev_info(&pf->pdev->dev, 485 struct i40e_ring *rx_ring = ACCESS_ONCE(vsi->rx_rings[i]);
490 " rx_rings[%i]: desc = %p\n", 486 if (!rx_ring)
491 i, vsi->rx_rings[i].desc); 487 continue;
492 dev_info(&pf->pdev->dev, 488
493 " rx_rings[%i]: dev = %p, netdev = %p, rx_bi = %p\n", 489 dev_info(&pf->pdev->dev,
494 i, vsi->rx_rings[i].dev, 490 " rx_rings[%i]: desc = %p\n",
495 vsi->rx_rings[i].netdev, 491 i, rx_ring->desc);
496 vsi->rx_rings[i].rx_bi); 492 dev_info(&pf->pdev->dev,
497 dev_info(&pf->pdev->dev, 493 " rx_rings[%i]: dev = %p, netdev = %p, rx_bi = %p\n",
498 " rx_rings[%i]: state = %li, queue_index = %d, reg_idx = %d\n", 494 i, rx_ring->dev,
499 i, vsi->rx_rings[i].state, 495 rx_ring->netdev,
500 vsi->rx_rings[i].queue_index, 496 rx_ring->rx_bi);
501 vsi->rx_rings[i].reg_idx); 497 dev_info(&pf->pdev->dev,
502 dev_info(&pf->pdev->dev, 498 " rx_rings[%i]: state = %li, queue_index = %d, reg_idx = %d\n",
503 " rx_rings[%i]: rx_hdr_len = %d, rx_buf_len = %d, dtype = %d\n", 499 i, rx_ring->state,
504 i, vsi->rx_rings[i].rx_hdr_len, 500 rx_ring->queue_index,
505 vsi->rx_rings[i].rx_buf_len, 501 rx_ring->reg_idx);
506 vsi->rx_rings[i].dtype); 502 dev_info(&pf->pdev->dev,
507 dev_info(&pf->pdev->dev, 503 " rx_rings[%i]: rx_hdr_len = %d, rx_buf_len = %d, dtype = %d\n",
508 " rx_rings[%i]: hsplit = %d, next_to_use = %d, next_to_clean = %d, ring_active = %i\n", 504 i, rx_ring->rx_hdr_len,
509 i, vsi->rx_rings[i].hsplit, 505 rx_ring->rx_buf_len,
510 vsi->rx_rings[i].next_to_use, 506 rx_ring->dtype);
511 vsi->rx_rings[i].next_to_clean, 507 dev_info(&pf->pdev->dev,
512 vsi->rx_rings[i].ring_active); 508 " rx_rings[%i]: hsplit = %d, next_to_use = %d, next_to_clean = %d, ring_active = %i\n",
513 dev_info(&pf->pdev->dev, 509 i, rx_ring->hsplit,
514 " rx_rings[%i]: rx_stats: packets = %lld, bytes = %lld, non_eop_descs = %lld\n", 510 rx_ring->next_to_use,
515 i, vsi->rx_rings[i].rx_stats.packets, 511 rx_ring->next_to_clean,
516 vsi->rx_rings[i].rx_stats.bytes, 512 rx_ring->ring_active);
517 vsi->rx_rings[i].rx_stats.non_eop_descs); 513 dev_info(&pf->pdev->dev,
518 dev_info(&pf->pdev->dev, 514 " rx_rings[%i]: rx_stats: packets = %lld, bytes = %lld, non_eop_descs = %lld\n",
519 " rx_rings[%i]: rx_stats: alloc_rx_page_failed = %lld, alloc_rx_buff_failed = %lld\n", 515 i, rx_ring->stats.packets,
520 i, 516 rx_ring->stats.bytes,
521 vsi->rx_rings[i].rx_stats.alloc_rx_page_failed, 517 rx_ring->rx_stats.non_eop_descs);
522 vsi->rx_rings[i].rx_stats.alloc_rx_buff_failed); 518 dev_info(&pf->pdev->dev,
523 dev_info(&pf->pdev->dev, 519 " rx_rings[%i]: rx_stats: alloc_rx_page_failed = %lld, alloc_rx_buff_failed = %lld\n",
524 " rx_rings[%i]: size = %i, dma = 0x%08lx\n", 520 i,
525 i, vsi->rx_rings[i].size, 521 rx_ring->rx_stats.alloc_rx_page_failed,
526 (long unsigned int)vsi->rx_rings[i].dma); 522 rx_ring->rx_stats.alloc_rx_buff_failed);
527 dev_info(&pf->pdev->dev, 523 dev_info(&pf->pdev->dev,
528 " rx_rings[%i]: vsi = %p, q_vector = %p\n", 524 " rx_rings[%i]: size = %i, dma = 0x%08lx\n",
529 i, vsi->rx_rings[i].vsi, 525 i, rx_ring->size,
530 vsi->rx_rings[i].q_vector); 526 (long unsigned int)rx_ring->dma);
531 } 527 dev_info(&pf->pdev->dev,
528 " rx_rings[%i]: vsi = %p, q_vector = %p\n",
529 i, rx_ring->vsi,
530 rx_ring->q_vector);
532 } 531 }
533 if (vsi->tx_rings) { 532 for (i = 0; i < vsi->num_queue_pairs; i++) {
534 for (i = 0; i < vsi->num_queue_pairs; i++) { 533 struct i40e_ring *tx_ring = ACCESS_ONCE(vsi->tx_rings[i]);
535 dev_info(&pf->pdev->dev, 534 if (!tx_ring)
536 " tx_rings[%i]: desc = %p\n", 535 continue;
537 i, vsi->tx_rings[i].desc); 536 dev_info(&pf->pdev->dev,
538 dev_info(&pf->pdev->dev, 537 " tx_rings[%i]: desc = %p\n",
539 " tx_rings[%i]: dev = %p, netdev = %p, tx_bi = %p\n", 538 i, tx_ring->desc);
540 i, vsi->tx_rings[i].dev, 539 dev_info(&pf->pdev->dev,
541 vsi->tx_rings[i].netdev, 540 " tx_rings[%i]: dev = %p, netdev = %p, tx_bi = %p\n",
542 vsi->tx_rings[i].tx_bi); 541 i, tx_ring->dev,
543 dev_info(&pf->pdev->dev, 542 tx_ring->netdev,
544 " tx_rings[%i]: state = %li, queue_index = %d, reg_idx = %d\n", 543 tx_ring->tx_bi);
545 i, vsi->tx_rings[i].state, 544 dev_info(&pf->pdev->dev,
546 vsi->tx_rings[i].queue_index, 545 " tx_rings[%i]: state = %li, queue_index = %d, reg_idx = %d\n",
547 vsi->tx_rings[i].reg_idx); 546 i, tx_ring->state,
548 dev_info(&pf->pdev->dev, 547 tx_ring->queue_index,
549 " tx_rings[%i]: dtype = %d\n", 548 tx_ring->reg_idx);
550 i, vsi->tx_rings[i].dtype); 549 dev_info(&pf->pdev->dev,
551 dev_info(&pf->pdev->dev, 550 " tx_rings[%i]: dtype = %d\n",
552 " tx_rings[%i]: hsplit = %d, next_to_use = %d, next_to_clean = %d, ring_active = %i\n", 551 i, tx_ring->dtype);
553 i, vsi->tx_rings[i].hsplit, 552 dev_info(&pf->pdev->dev,
554 vsi->tx_rings[i].next_to_use, 553 " tx_rings[%i]: hsplit = %d, next_to_use = %d, next_to_clean = %d, ring_active = %i\n",
555 vsi->tx_rings[i].next_to_clean, 554 i, tx_ring->hsplit,
556 vsi->tx_rings[i].ring_active); 555 tx_ring->next_to_use,
557 dev_info(&pf->pdev->dev, 556 tx_ring->next_to_clean,
558 " tx_rings[%i]: tx_stats: packets = %lld, bytes = %lld, restart_queue = %lld\n", 557 tx_ring->ring_active);
559 i, vsi->tx_rings[i].tx_stats.packets, 558 dev_info(&pf->pdev->dev,
560 vsi->tx_rings[i].tx_stats.bytes, 559 " tx_rings[%i]: tx_stats: packets = %lld, bytes = %lld, restart_queue = %lld\n",
561 vsi->tx_rings[i].tx_stats.restart_queue); 560 i, tx_ring->stats.packets,
562 dev_info(&pf->pdev->dev, 561 tx_ring->stats.bytes,
563 " tx_rings[%i]: tx_stats: tx_busy = %lld, completed = %lld, tx_done_old = %lld\n", 562 tx_ring->tx_stats.restart_queue);
564 i, 563 dev_info(&pf->pdev->dev,
565 vsi->tx_rings[i].tx_stats.tx_busy, 564 " tx_rings[%i]: tx_stats: tx_busy = %lld, tx_done_old = %lld\n",
566 vsi->tx_rings[i].tx_stats.completed, 565 i,
567 vsi->tx_rings[i].tx_stats.tx_done_old); 566 tx_ring->tx_stats.tx_busy,
568 dev_info(&pf->pdev->dev, 567 tx_ring->tx_stats.tx_done_old);
569 " tx_rings[%i]: size = %i, dma = 0x%08lx\n", 568 dev_info(&pf->pdev->dev,
570 i, vsi->tx_rings[i].size, 569 " tx_rings[%i]: size = %i, dma = 0x%08lx\n",
571 (long unsigned int)vsi->tx_rings[i].dma); 570 i, tx_ring->size,
572 dev_info(&pf->pdev->dev, 571 (long unsigned int)tx_ring->dma);
573 " tx_rings[%i]: vsi = %p, q_vector = %p\n", 572 dev_info(&pf->pdev->dev,
574 i, vsi->tx_rings[i].vsi, 573 " tx_rings[%i]: vsi = %p, q_vector = %p\n",
575 vsi->tx_rings[i].q_vector); 574 i, tx_ring->vsi,
576 dev_info(&pf->pdev->dev, 575 tx_ring->q_vector);
577 " tx_rings[%i]: DCB tc = %d\n", 576 dev_info(&pf->pdev->dev,
578 i, vsi->tx_rings[i].dcb_tc); 577 " tx_rings[%i]: DCB tc = %d\n",
579 } 578 i, tx_ring->dcb_tc);
580 } 579 }
580 rcu_read_unlock();
581 dev_info(&pf->pdev->dev, 581 dev_info(&pf->pdev->dev,
582 " work_limit = %d, rx_itr_setting = %d (%s), tx_itr_setting = %d (%s)\n", 582 " work_limit = %d, rx_itr_setting = %d (%s), tx_itr_setting = %d (%s)\n",
583 vsi->work_limit, vsi->rx_itr_setting, 583 vsi->work_limit, vsi->rx_itr_setting,
@@ -587,15 +587,6 @@ static void i40e_dbg_dump_vsi_seid(struct i40e_pf *pf, int seid)
587 dev_info(&pf->pdev->dev, 587 dev_info(&pf->pdev->dev,
588 " max_frame = %d, rx_hdr_len = %d, rx_buf_len = %d dtype = %d\n", 588 " max_frame = %d, rx_hdr_len = %d, rx_buf_len = %d dtype = %d\n",
589 vsi->max_frame, vsi->rx_hdr_len, vsi->rx_buf_len, vsi->dtype); 589 vsi->max_frame, vsi->rx_hdr_len, vsi->rx_buf_len, vsi->dtype);
590 if (vsi->q_vectors) {
591 for (i = 0; i < vsi->num_q_vectors; i++) {
592 dev_info(&pf->pdev->dev,
593 " q_vectors[%i]: base index = %ld\n",
594 i, ((long int)*vsi->q_vectors[i].rx.ring-
595 (long int)*vsi->q_vectors[0].rx.ring)/
596 sizeof(struct i40e_ring));
597 }
598 }
599 dev_info(&pf->pdev->dev, 590 dev_info(&pf->pdev->dev,
600 " num_q_vectors = %i, base_vector = %i\n", 591 " num_q_vectors = %i, base_vector = %i\n",
601 vsi->num_q_vectors, vsi->base_vector); 592 vsi->num_q_vectors, vsi->base_vector);
@@ -792,9 +783,9 @@ static void i40e_dbg_dump_desc(int cnt, int vsi_seid, int ring_id, int desc_n,
792 return; 783 return;
793 } 784 }
794 if (is_rx_ring) 785 if (is_rx_ring)
795 ring = vsi->rx_rings[ring_id]; 786 ring = *vsi->rx_rings[ring_id];
796 else 787 else
797 ring = vsi->tx_rings[ring_id]; 788 ring = *vsi->tx_rings[ring_id];
798 if (cnt == 2) { 789 if (cnt == 2) {
799 dev_info(&pf->pdev->dev, "vsi = %02i %s ring = %02i\n", 790 dev_info(&pf->pdev->dev, "vsi = %02i %s ring = %02i\n",
800 vsi_seid, is_rx_ring ? "rx" : "tx", ring_id); 791 vsi_seid, is_rx_ring ? "rx" : "tx", ring_id);
@@ -1028,11 +1019,11 @@ static ssize_t i40e_dbg_command_write(struct file *filp,
1028 size_t count, loff_t *ppos) 1019 size_t count, loff_t *ppos)
1029{ 1020{
1030 struct i40e_pf *pf = filp->private_data; 1021 struct i40e_pf *pf = filp->private_data;
1022 char *cmd_buf, *cmd_buf_tmp;
1031 int bytes_not_copied; 1023 int bytes_not_copied;
1032 struct i40e_vsi *vsi; 1024 struct i40e_vsi *vsi;
1033 u8 *print_buf_start; 1025 u8 *print_buf_start;
1034 u8 *print_buf; 1026 u8 *print_buf;
1035 char *cmd_buf;
1036 int vsi_seid; 1027 int vsi_seid;
1037 int veb_seid; 1028 int veb_seid;
1038 int cnt; 1029 int cnt;
@@ -1051,6 +1042,12 @@ static ssize_t i40e_dbg_command_write(struct file *filp,
1051 count -= bytes_not_copied; 1042 count -= bytes_not_copied;
1052 cmd_buf[count] = '\0'; 1043 cmd_buf[count] = '\0';
1053 1044
1045 cmd_buf_tmp = strchr(cmd_buf, '\n');
1046 if (cmd_buf_tmp) {
1047 *cmd_buf_tmp = '\0';
1048 count = cmd_buf_tmp - cmd_buf + 1;
1049 }
1050
1054 print_buf_start = kzalloc(I40E_MAX_DEBUG_OUT_BUFFER, GFP_KERNEL); 1051 print_buf_start = kzalloc(I40E_MAX_DEBUG_OUT_BUFFER, GFP_KERNEL);
1055 if (!print_buf_start) 1052 if (!print_buf_start)
1056 goto command_write_done; 1053 goto command_write_done;
@@ -1157,9 +1154,9 @@ static ssize_t i40e_dbg_command_write(struct file *filp,
1157 i40e_veb_release(pf->veb[i]); 1154 i40e_veb_release(pf->veb[i]);
1158 1155
1159 } else if (strncmp(cmd_buf, "add macaddr", 11) == 0) { 1156 } else if (strncmp(cmd_buf, "add macaddr", 11) == 0) {
1160 u8 ma[6];
1161 int vlan = 0;
1162 struct i40e_mac_filter *f; 1157 struct i40e_mac_filter *f;
1158 int vlan = 0;
1159 u8 ma[6];
1163 int ret; 1160 int ret;
1164 1161
1165 cnt = sscanf(&cmd_buf[11], 1162 cnt = sscanf(&cmd_buf[11],
@@ -1195,8 +1192,8 @@ static ssize_t i40e_dbg_command_write(struct file *filp,
1195 ma, vlan, vsi_seid, f, ret); 1192 ma, vlan, vsi_seid, f, ret);
1196 1193
1197 } else if (strncmp(cmd_buf, "del macaddr", 11) == 0) { 1194 } else if (strncmp(cmd_buf, "del macaddr", 11) == 0) {
1198 u8 ma[6];
1199 int vlan = 0; 1195 int vlan = 0;
1196 u8 ma[6];
1200 int ret; 1197 int ret;
1201 1198
1202 cnt = sscanf(&cmd_buf[11], 1199 cnt = sscanf(&cmd_buf[11],
@@ -1232,9 +1229,9 @@ static ssize_t i40e_dbg_command_write(struct file *filp,
1232 ma, vlan, vsi_seid, ret); 1229 ma, vlan, vsi_seid, ret);
1233 1230
1234 } else if (strncmp(cmd_buf, "add pvid", 8) == 0) { 1231 } else if (strncmp(cmd_buf, "add pvid", 8) == 0) {
1235 int v;
1236 u16 vid;
1237 i40e_status ret; 1232 i40e_status ret;
1233 u16 vid;
1234 int v;
1238 1235
1239 cnt = sscanf(&cmd_buf[8], "%i %u", &vsi_seid, &v); 1236 cnt = sscanf(&cmd_buf[8], "%i %u", &vsi_seid, &v);
1240 if (cnt != 2) { 1237 if (cnt != 2) {
@@ -1545,10 +1542,10 @@ static ssize_t i40e_dbg_command_write(struct file *filp,
1545 } else if ((strncmp(cmd_buf, "add fd_filter", 13) == 0) || 1542 } else if ((strncmp(cmd_buf, "add fd_filter", 13) == 0) ||
1546 (strncmp(cmd_buf, "rem fd_filter", 13) == 0)) { 1543 (strncmp(cmd_buf, "rem fd_filter", 13) == 0)) {
1547 struct i40e_fdir_data fd_data; 1544 struct i40e_fdir_data fd_data;
1548 int ret;
1549 u16 packet_len, i, j = 0; 1545 u16 packet_len, i, j = 0;
1550 char *asc_packet; 1546 char *asc_packet;
1551 bool add = false; 1547 bool add = false;
1548 int ret;
1552 1549
1553 asc_packet = kzalloc(I40E_FDIR_MAX_RAW_PACKET_LOOKUP, 1550 asc_packet = kzalloc(I40E_FDIR_MAX_RAW_PACKET_LOOKUP,
1554 GFP_KERNEL); 1551 GFP_KERNEL);
@@ -1636,9 +1633,9 @@ static ssize_t i40e_dbg_command_write(struct file *filp,
1636 } 1633 }
1637 } else if (strncmp(&cmd_buf[5], 1634 } else if (strncmp(&cmd_buf[5],
1638 "get local", 9) == 0) { 1635 "get local", 9) == 0) {
1636 u16 llen, rlen;
1639 int ret, i; 1637 int ret, i;
1640 u8 *buff; 1638 u8 *buff;
1641 u16 llen, rlen;
1642 buff = kzalloc(I40E_LLDPDU_SIZE, GFP_KERNEL); 1639 buff = kzalloc(I40E_LLDPDU_SIZE, GFP_KERNEL);
1643 if (!buff) 1640 if (!buff)
1644 goto command_write_done; 1641 goto command_write_done;
@@ -1669,9 +1666,9 @@ static ssize_t i40e_dbg_command_write(struct file *filp,
1669 kfree(buff); 1666 kfree(buff);
1670 buff = NULL; 1667 buff = NULL;
1671 } else if (strncmp(&cmd_buf[5], "get remote", 10) == 0) { 1668 } else if (strncmp(&cmd_buf[5], "get remote", 10) == 0) {
1669 u16 llen, rlen;
1672 int ret, i; 1670 int ret, i;
1673 u8 *buff; 1671 u8 *buff;
1674 u16 llen, rlen;
1675 buff = kzalloc(I40E_LLDPDU_SIZE, GFP_KERNEL); 1672 buff = kzalloc(I40E_LLDPDU_SIZE, GFP_KERNEL);
1676 if (!buff) 1673 if (!buff)
1677 goto command_write_done; 1674 goto command_write_done;
@@ -1747,11 +1744,13 @@ static ssize_t i40e_dbg_command_write(struct file *filp,
1747 goto command_write_done; 1744 goto command_write_done;
1748 } 1745 }
1749 1746
1750 /* Read at least 512 words */ 1747 /* set the max length */
1751 if (buffer_len == 0) 1748 buffer_len = min_t(u16, buffer_len, I40E_MAX_AQ_BUF_SIZE/2);
1752 buffer_len = 512;
1753 1749
1754 bytes = 2 * buffer_len; 1750 bytes = 2 * buffer_len;
1751
1752 /* read at least 1k bytes, no more than 4kB */
1753 bytes = clamp(bytes, (u16)1024, (u16)I40E_MAX_AQ_BUF_SIZE);
1755 buff = kzalloc(bytes, GFP_KERNEL); 1754 buff = kzalloc(bytes, GFP_KERNEL);
1756 if (!buff) 1755 if (!buff)
1757 goto command_write_done; 1756 goto command_write_done;
@@ -1903,6 +1902,7 @@ static ssize_t i40e_dbg_netdev_ops_write(struct file *filp,
1903 struct i40e_pf *pf = filp->private_data; 1902 struct i40e_pf *pf = filp->private_data;
1904 int bytes_not_copied; 1903 int bytes_not_copied;
1905 struct i40e_vsi *vsi; 1904 struct i40e_vsi *vsi;
1905 char *buf_tmp;
1906 int vsi_seid; 1906 int vsi_seid;
1907 int i, cnt; 1907 int i, cnt;
1908 1908
@@ -1921,6 +1921,12 @@ static ssize_t i40e_dbg_netdev_ops_write(struct file *filp,
1921 count -= bytes_not_copied; 1921 count -= bytes_not_copied;
1922 i40e_dbg_netdev_ops_buf[count] = '\0'; 1922 i40e_dbg_netdev_ops_buf[count] = '\0';
1923 1923
1924 buf_tmp = strchr(i40e_dbg_netdev_ops_buf, '\n');
1925 if (buf_tmp) {
1926 *buf_tmp = '\0';
1927 count = buf_tmp - i40e_dbg_netdev_ops_buf + 1;
1928 }
1929
1924 if (strncmp(i40e_dbg_netdev_ops_buf, "tx_timeout", 10) == 0) { 1930 if (strncmp(i40e_dbg_netdev_ops_buf, "tx_timeout", 10) == 0) {
1925 cnt = sscanf(&i40e_dbg_netdev_ops_buf[11], "%i", &vsi_seid); 1931 cnt = sscanf(&i40e_dbg_netdev_ops_buf[11], "%i", &vsi_seid);
1926 if (cnt != 1) { 1932 if (cnt != 1) {
@@ -1996,7 +2002,7 @@ static ssize_t i40e_dbg_netdev_ops_write(struct file *filp,
1996 goto netdev_ops_write_done; 2002 goto netdev_ops_write_done;
1997 } 2003 }
1998 for (i = 0; i < vsi->num_q_vectors; i++) 2004 for (i = 0; i < vsi->num_q_vectors; i++)
1999 napi_schedule(&vsi->q_vectors[i].napi); 2005 napi_schedule(&vsi->q_vectors[i]->napi);
2000 dev_info(&pf->pdev->dev, "napi called\n"); 2006 dev_info(&pf->pdev->dev, "napi called\n");
2001 } else { 2007 } else {
2002 dev_info(&pf->pdev->dev, "unknown command '%s'\n", 2008 dev_info(&pf->pdev->dev, "unknown command '%s'\n",
@@ -2024,21 +2030,35 @@ static const struct file_operations i40e_dbg_netdev_ops_fops = {
2024 **/ 2030 **/
2025void i40e_dbg_pf_init(struct i40e_pf *pf) 2031void i40e_dbg_pf_init(struct i40e_pf *pf)
2026{ 2032{
2027 struct dentry *pfile __attribute__((unused)); 2033 struct dentry *pfile;
2028 const char *name = pci_name(pf->pdev); 2034 const char *name = pci_name(pf->pdev);
2035 const struct device *dev = &pf->pdev->dev;
2029 2036
2030 pf->i40e_dbg_pf = debugfs_create_dir(name, i40e_dbg_root); 2037 pf->i40e_dbg_pf = debugfs_create_dir(name, i40e_dbg_root);
2031 if (pf->i40e_dbg_pf) { 2038 if (!pf->i40e_dbg_pf)
2032 pfile = debugfs_create_file("command", 0600, pf->i40e_dbg_pf, 2039 return;
2033 pf, &i40e_dbg_command_fops); 2040
2034 pfile = debugfs_create_file("dump", 0600, pf->i40e_dbg_pf, pf, 2041 pfile = debugfs_create_file("command", 0600, pf->i40e_dbg_pf, pf,
2035 &i40e_dbg_dump_fops); 2042 &i40e_dbg_command_fops);
2036 pfile = debugfs_create_file("netdev_ops", 0600, pf->i40e_dbg_pf, 2043 if (!pfile)
2037 pf, &i40e_dbg_netdev_ops_fops); 2044 goto create_failed;
2038 } else { 2045
2039 dev_info(&pf->pdev->dev, 2046 pfile = debugfs_create_file("dump", 0600, pf->i40e_dbg_pf, pf,
2040 "debugfs entry for %s failed\n", name); 2047 &i40e_dbg_dump_fops);
2041 } 2048 if (!pfile)
2049 goto create_failed;
2050
2051 pfile = debugfs_create_file("netdev_ops", 0600, pf->i40e_dbg_pf, pf,
2052 &i40e_dbg_netdev_ops_fops);
2053 if (!pfile)
2054 goto create_failed;
2055
2056 return;
2057
2058create_failed:
2059 dev_info(dev, "debugfs dir/file for %s failed\n", name);
2060 debugfs_remove_recursive(pf->i40e_dbg_pf);
2061 return;
2042} 2062}
2043 2063
2044/** 2064/**
diff --git a/drivers/net/ethernet/intel/i40e/i40e_ethtool.c b/drivers/net/ethernet/intel/i40e/i40e_ethtool.c
index 9a76b8cec76c..1b86138fa9e1 100644
--- a/drivers/net/ethernet/intel/i40e/i40e_ethtool.c
+++ b/drivers/net/ethernet/intel/i40e/i40e_ethtool.c
@@ -399,8 +399,8 @@ static void i40e_get_ringparam(struct net_device *netdev,
399 ring->tx_max_pending = I40E_MAX_NUM_DESCRIPTORS; 399 ring->tx_max_pending = I40E_MAX_NUM_DESCRIPTORS;
400 ring->rx_mini_max_pending = 0; 400 ring->rx_mini_max_pending = 0;
401 ring->rx_jumbo_max_pending = 0; 401 ring->rx_jumbo_max_pending = 0;
402 ring->rx_pending = vsi->rx_rings[0].count; 402 ring->rx_pending = vsi->rx_rings[0]->count;
403 ring->tx_pending = vsi->tx_rings[0].count; 403 ring->tx_pending = vsi->tx_rings[0]->count;
404 ring->rx_mini_pending = 0; 404 ring->rx_mini_pending = 0;
405 ring->rx_jumbo_pending = 0; 405 ring->rx_jumbo_pending = 0;
406} 406}
@@ -429,8 +429,8 @@ static int i40e_set_ringparam(struct net_device *netdev,
429 new_rx_count = ALIGN(new_rx_count, I40E_REQ_DESCRIPTOR_MULTIPLE); 429 new_rx_count = ALIGN(new_rx_count, I40E_REQ_DESCRIPTOR_MULTIPLE);
430 430
431 /* if nothing to do return success */ 431 /* if nothing to do return success */
432 if ((new_tx_count == vsi->tx_rings[0].count) && 432 if ((new_tx_count == vsi->tx_rings[0]->count) &&
433 (new_rx_count == vsi->rx_rings[0].count)) 433 (new_rx_count == vsi->rx_rings[0]->count))
434 return 0; 434 return 0;
435 435
436 while (test_and_set_bit(__I40E_CONFIG_BUSY, &pf->state)) 436 while (test_and_set_bit(__I40E_CONFIG_BUSY, &pf->state))
@@ -439,8 +439,8 @@ static int i40e_set_ringparam(struct net_device *netdev,
439 if (!netif_running(vsi->netdev)) { 439 if (!netif_running(vsi->netdev)) {
440 /* simple case - set for the next time the netdev is started */ 440 /* simple case - set for the next time the netdev is started */
441 for (i = 0; i < vsi->num_queue_pairs; i++) { 441 for (i = 0; i < vsi->num_queue_pairs; i++) {
442 vsi->tx_rings[i].count = new_tx_count; 442 vsi->tx_rings[i]->count = new_tx_count;
443 vsi->rx_rings[i].count = new_rx_count; 443 vsi->rx_rings[i]->count = new_rx_count;
444 } 444 }
445 goto done; 445 goto done;
446 } 446 }
@@ -451,10 +451,10 @@ static int i40e_set_ringparam(struct net_device *netdev,
451 */ 451 */
452 452
453 /* alloc updated Tx resources */ 453 /* alloc updated Tx resources */
454 if (new_tx_count != vsi->tx_rings[0].count) { 454 if (new_tx_count != vsi->tx_rings[0]->count) {
455 netdev_info(netdev, 455 netdev_info(netdev,
456 "Changing Tx descriptor count from %d to %d.\n", 456 "Changing Tx descriptor count from %d to %d.\n",
457 vsi->tx_rings[0].count, new_tx_count); 457 vsi->tx_rings[0]->count, new_tx_count);
458 tx_rings = kcalloc(vsi->alloc_queue_pairs, 458 tx_rings = kcalloc(vsi->alloc_queue_pairs,
459 sizeof(struct i40e_ring), GFP_KERNEL); 459 sizeof(struct i40e_ring), GFP_KERNEL);
460 if (!tx_rings) { 460 if (!tx_rings) {
@@ -464,7 +464,7 @@ static int i40e_set_ringparam(struct net_device *netdev,
464 464
465 for (i = 0; i < vsi->num_queue_pairs; i++) { 465 for (i = 0; i < vsi->num_queue_pairs; i++) {
466 /* clone ring and setup updated count */ 466 /* clone ring and setup updated count */
467 tx_rings[i] = vsi->tx_rings[i]; 467 tx_rings[i] = *vsi->tx_rings[i];
468 tx_rings[i].count = new_tx_count; 468 tx_rings[i].count = new_tx_count;
469 err = i40e_setup_tx_descriptors(&tx_rings[i]); 469 err = i40e_setup_tx_descriptors(&tx_rings[i]);
470 if (err) { 470 if (err) {
@@ -481,10 +481,10 @@ static int i40e_set_ringparam(struct net_device *netdev,
481 } 481 }
482 482
483 /* alloc updated Rx resources */ 483 /* alloc updated Rx resources */
484 if (new_rx_count != vsi->rx_rings[0].count) { 484 if (new_rx_count != vsi->rx_rings[0]->count) {
485 netdev_info(netdev, 485 netdev_info(netdev,
486 "Changing Rx descriptor count from %d to %d\n", 486 "Changing Rx descriptor count from %d to %d\n",
487 vsi->rx_rings[0].count, new_rx_count); 487 vsi->rx_rings[0]->count, new_rx_count);
488 rx_rings = kcalloc(vsi->alloc_queue_pairs, 488 rx_rings = kcalloc(vsi->alloc_queue_pairs,
489 sizeof(struct i40e_ring), GFP_KERNEL); 489 sizeof(struct i40e_ring), GFP_KERNEL);
490 if (!rx_rings) { 490 if (!rx_rings) {
@@ -494,7 +494,7 @@ static int i40e_set_ringparam(struct net_device *netdev,
494 494
495 for (i = 0; i < vsi->num_queue_pairs; i++) { 495 for (i = 0; i < vsi->num_queue_pairs; i++) {
496 /* clone ring and setup updated count */ 496 /* clone ring and setup updated count */
497 rx_rings[i] = vsi->rx_rings[i]; 497 rx_rings[i] = *vsi->rx_rings[i];
498 rx_rings[i].count = new_rx_count; 498 rx_rings[i].count = new_rx_count;
499 err = i40e_setup_rx_descriptors(&rx_rings[i]); 499 err = i40e_setup_rx_descriptors(&rx_rings[i]);
500 if (err) { 500 if (err) {
@@ -517,8 +517,8 @@ static int i40e_set_ringparam(struct net_device *netdev,
517 517
518 if (tx_rings) { 518 if (tx_rings) {
519 for (i = 0; i < vsi->num_queue_pairs; i++) { 519 for (i = 0; i < vsi->num_queue_pairs; i++) {
520 i40e_free_tx_resources(&vsi->tx_rings[i]); 520 i40e_free_tx_resources(vsi->tx_rings[i]);
521 vsi->tx_rings[i] = tx_rings[i]; 521 *vsi->tx_rings[i] = tx_rings[i];
522 } 522 }
523 kfree(tx_rings); 523 kfree(tx_rings);
524 tx_rings = NULL; 524 tx_rings = NULL;
@@ -526,8 +526,8 @@ static int i40e_set_ringparam(struct net_device *netdev,
526 526
527 if (rx_rings) { 527 if (rx_rings) {
528 for (i = 0; i < vsi->num_queue_pairs; i++) { 528 for (i = 0; i < vsi->num_queue_pairs; i++) {
529 i40e_free_rx_resources(&vsi->rx_rings[i]); 529 i40e_free_rx_resources(vsi->rx_rings[i]);
530 vsi->rx_rings[i] = rx_rings[i]; 530 *vsi->rx_rings[i] = rx_rings[i];
531 } 531 }
532 kfree(rx_rings); 532 kfree(rx_rings);
533 rx_rings = NULL; 533 rx_rings = NULL;
@@ -579,6 +579,7 @@ static void i40e_get_ethtool_stats(struct net_device *netdev,
579 char *p; 579 char *p;
580 int j; 580 int j;
581 struct rtnl_link_stats64 *net_stats = i40e_get_vsi_stats_struct(vsi); 581 struct rtnl_link_stats64 *net_stats = i40e_get_vsi_stats_struct(vsi);
582 unsigned int start;
582 583
583 i40e_update_stats(vsi); 584 i40e_update_stats(vsi);
584 585
@@ -587,14 +588,30 @@ static void i40e_get_ethtool_stats(struct net_device *netdev,
587 data[i++] = (i40e_gstrings_net_stats[j].sizeof_stat == 588 data[i++] = (i40e_gstrings_net_stats[j].sizeof_stat ==
588 sizeof(u64)) ? *(u64 *)p : *(u32 *)p; 589 sizeof(u64)) ? *(u64 *)p : *(u32 *)p;
589 } 590 }
590 for (j = 0; j < vsi->num_queue_pairs; j++) { 591 rcu_read_lock();
591 data[i++] = vsi->tx_rings[j].tx_stats.packets; 592 for (j = 0; j < vsi->num_queue_pairs; j++, i += 4) {
592 data[i++] = vsi->tx_rings[j].tx_stats.bytes; 593 struct i40e_ring *tx_ring = ACCESS_ONCE(vsi->tx_rings[j]);
593 } 594 struct i40e_ring *rx_ring;
594 for (j = 0; j < vsi->num_queue_pairs; j++) { 595
595 data[i++] = vsi->rx_rings[j].rx_stats.packets; 596 if (!tx_ring)
596 data[i++] = vsi->rx_rings[j].rx_stats.bytes; 597 continue;
598
599 /* process Tx ring statistics */
600 do {
601 start = u64_stats_fetch_begin_bh(&tx_ring->syncp);
602 data[i] = tx_ring->stats.packets;
603 data[i + 1] = tx_ring->stats.bytes;
604 } while (u64_stats_fetch_retry_bh(&tx_ring->syncp, start));
605
606 /* Rx ring is the 2nd half of the queue pair */
607 rx_ring = &tx_ring[1];
608 do {
609 start = u64_stats_fetch_begin_bh(&rx_ring->syncp);
610 data[i + 2] = rx_ring->stats.packets;
611 data[i + 3] = rx_ring->stats.bytes;
612 } while (u64_stats_fetch_retry_bh(&rx_ring->syncp, start));
597 } 613 }
614 rcu_read_unlock();
598 if (vsi == pf->vsi[pf->lan_vsi]) { 615 if (vsi == pf->vsi[pf->lan_vsi]) {
599 for (j = 0; j < I40E_GLOBAL_STATS_LEN; j++) { 616 for (j = 0; j < I40E_GLOBAL_STATS_LEN; j++) {
600 p = (char *)pf + i40e_gstrings_stats[j].stat_offset; 617 p = (char *)pf + i40e_gstrings_stats[j].stat_offset;
@@ -641,8 +658,6 @@ static void i40e_get_strings(struct net_device *netdev, u32 stringset,
641 p += ETH_GSTRING_LEN; 658 p += ETH_GSTRING_LEN;
642 snprintf(p, ETH_GSTRING_LEN, "tx-%u.tx_bytes", i); 659 snprintf(p, ETH_GSTRING_LEN, "tx-%u.tx_bytes", i);
643 p += ETH_GSTRING_LEN; 660 p += ETH_GSTRING_LEN;
644 }
645 for (i = 0; i < vsi->num_queue_pairs; i++) {
646 snprintf(p, ETH_GSTRING_LEN, "rx-%u.rx_packets", i); 661 snprintf(p, ETH_GSTRING_LEN, "rx-%u.rx_packets", i);
647 p += ETH_GSTRING_LEN; 662 p += ETH_GSTRING_LEN;
648 snprintf(p, ETH_GSTRING_LEN, "rx-%u.rx_bytes", i); 663 snprintf(p, ETH_GSTRING_LEN, "rx-%u.rx_bytes", i);
@@ -910,8 +925,8 @@ static int i40e_set_coalesce(struct net_device *netdev,
910 } 925 }
911 926
912 vector = vsi->base_vector; 927 vector = vsi->base_vector;
913 q_vector = vsi->q_vectors; 928 for (i = 0; i < vsi->num_q_vectors; i++, vector++) {
914 for (i = 0; i < vsi->num_q_vectors; i++, vector++, q_vector++) { 929 q_vector = vsi->q_vectors[i];
915 q_vector->rx.itr = ITR_TO_REG(vsi->rx_itr_setting); 930 q_vector->rx.itr = ITR_TO_REG(vsi->rx_itr_setting);
916 wr32(hw, I40E_PFINT_ITRN(0, vector - 1), q_vector->rx.itr); 931 wr32(hw, I40E_PFINT_ITRN(0, vector - 1), q_vector->rx.itr);
917 q_vector->tx.itr = ITR_TO_REG(vsi->tx_itr_setting); 932 q_vector->tx.itr = ITR_TO_REG(vsi->tx_itr_setting);
diff --git a/drivers/net/ethernet/intel/i40e/i40e_main.c b/drivers/net/ethernet/intel/i40e/i40e_main.c
index 221aa4795017..be15938ba213 100644
--- a/drivers/net/ethernet/intel/i40e/i40e_main.c
+++ b/drivers/net/ethernet/intel/i40e/i40e_main.c
@@ -36,7 +36,7 @@ static const char i40e_driver_string[] =
36 36
37#define DRV_VERSION_MAJOR 0 37#define DRV_VERSION_MAJOR 0
38#define DRV_VERSION_MINOR 3 38#define DRV_VERSION_MINOR 3
39#define DRV_VERSION_BUILD 9 39#define DRV_VERSION_BUILD 11
40#define DRV_VERSION __stringify(DRV_VERSION_MAJOR) "." \ 40#define DRV_VERSION __stringify(DRV_VERSION_MAJOR) "." \
41 __stringify(DRV_VERSION_MINOR) "." \ 41 __stringify(DRV_VERSION_MINOR) "." \
42 __stringify(DRV_VERSION_BUILD) DRV_KERN 42 __stringify(DRV_VERSION_BUILD) DRV_KERN
@@ -347,14 +347,53 @@ struct rtnl_link_stats64 *i40e_get_vsi_stats_struct(struct i40e_vsi *vsi)
347 **/ 347 **/
348static struct rtnl_link_stats64 *i40e_get_netdev_stats_struct( 348static struct rtnl_link_stats64 *i40e_get_netdev_stats_struct(
349 struct net_device *netdev, 349 struct net_device *netdev,
350 struct rtnl_link_stats64 *storage) 350 struct rtnl_link_stats64 *stats)
351{ 351{
352 struct i40e_netdev_priv *np = netdev_priv(netdev); 352 struct i40e_netdev_priv *np = netdev_priv(netdev);
353 struct i40e_vsi *vsi = np->vsi; 353 struct i40e_vsi *vsi = np->vsi;
354 struct rtnl_link_stats64 *vsi_stats = i40e_get_vsi_stats_struct(vsi);
355 int i;
356
357 rcu_read_lock();
358 for (i = 0; i < vsi->num_queue_pairs; i++) {
359 struct i40e_ring *tx_ring, *rx_ring;
360 u64 bytes, packets;
361 unsigned int start;
362
363 tx_ring = ACCESS_ONCE(vsi->tx_rings[i]);
364 if (!tx_ring)
365 continue;
354 366
355 *storage = *i40e_get_vsi_stats_struct(vsi); 367 do {
368 start = u64_stats_fetch_begin_bh(&tx_ring->syncp);
369 packets = tx_ring->stats.packets;
370 bytes = tx_ring->stats.bytes;
371 } while (u64_stats_fetch_retry_bh(&tx_ring->syncp, start));
372
373 stats->tx_packets += packets;
374 stats->tx_bytes += bytes;
375 rx_ring = &tx_ring[1];
356 376
357 return storage; 377 do {
378 start = u64_stats_fetch_begin_bh(&rx_ring->syncp);
379 packets = rx_ring->stats.packets;
380 bytes = rx_ring->stats.bytes;
381 } while (u64_stats_fetch_retry_bh(&rx_ring->syncp, start));
382
383 stats->rx_packets += packets;
384 stats->rx_bytes += bytes;
385 }
386 rcu_read_unlock();
387
388 /* following stats updated by ixgbe_watchdog_task() */
389 stats->multicast = vsi_stats->multicast;
390 stats->tx_errors = vsi_stats->tx_errors;
391 stats->tx_dropped = vsi_stats->tx_dropped;
392 stats->rx_errors = vsi_stats->rx_errors;
393 stats->rx_crc_errors = vsi_stats->rx_crc_errors;
394 stats->rx_length_errors = vsi_stats->rx_length_errors;
395
396 return stats;
358} 397}
359 398
360/** 399/**
@@ -376,10 +415,14 @@ void i40e_vsi_reset_stats(struct i40e_vsi *vsi)
376 memset(&vsi->eth_stats_offsets, 0, sizeof(vsi->eth_stats_offsets)); 415 memset(&vsi->eth_stats_offsets, 0, sizeof(vsi->eth_stats_offsets));
377 if (vsi->rx_rings) 416 if (vsi->rx_rings)
378 for (i = 0; i < vsi->num_queue_pairs; i++) { 417 for (i = 0; i < vsi->num_queue_pairs; i++) {
379 memset(&vsi->rx_rings[i].rx_stats, 0 , 418 memset(&vsi->rx_rings[i]->stats, 0 ,
380 sizeof(vsi->rx_rings[i].rx_stats)); 419 sizeof(vsi->rx_rings[i]->stats));
381 memset(&vsi->tx_rings[i].tx_stats, 0, 420 memset(&vsi->rx_rings[i]->rx_stats, 0 ,
382 sizeof(vsi->tx_rings[i].tx_stats)); 421 sizeof(vsi->rx_rings[i]->rx_stats));
422 memset(&vsi->tx_rings[i]->stats, 0 ,
423 sizeof(vsi->tx_rings[i]->stats));
424 memset(&vsi->tx_rings[i]->tx_stats, 0,
425 sizeof(vsi->tx_rings[i]->tx_stats));
383 } 426 }
384 vsi->stat_offsets_loaded = false; 427 vsi->stat_offsets_loaded = false;
385} 428}
@@ -598,7 +641,7 @@ static void i40e_update_link_xoff_rx(struct i40e_pf *pf)
598 continue; 641 continue;
599 642
600 for (i = 0; i < vsi->num_queue_pairs; i++) { 643 for (i = 0; i < vsi->num_queue_pairs; i++) {
601 struct i40e_ring *ring = &vsi->tx_rings[i]; 644 struct i40e_ring *ring = vsi->tx_rings[i];
602 clear_bit(__I40E_HANG_CHECK_ARMED, &ring->state); 645 clear_bit(__I40E_HANG_CHECK_ARMED, &ring->state);
603 } 646 }
604 } 647 }
@@ -652,7 +695,7 @@ static void i40e_update_prio_xoff_rx(struct i40e_pf *pf)
652 continue; 695 continue;
653 696
654 for (i = 0; i < vsi->num_queue_pairs; i++) { 697 for (i = 0; i < vsi->num_queue_pairs; i++) {
655 struct i40e_ring *ring = &vsi->tx_rings[i]; 698 struct i40e_ring *ring = vsi->tx_rings[i];
656 699
657 tc = ring->dcb_tc; 700 tc = ring->dcb_tc;
658 if (xoff[tc]) 701 if (xoff[tc])
@@ -704,21 +747,38 @@ void i40e_update_stats(struct i40e_vsi *vsi)
704 tx_restart = tx_busy = 0; 747 tx_restart = tx_busy = 0;
705 rx_page = 0; 748 rx_page = 0;
706 rx_buf = 0; 749 rx_buf = 0;
750 rcu_read_lock();
707 for (q = 0; q < vsi->num_queue_pairs; q++) { 751 for (q = 0; q < vsi->num_queue_pairs; q++) {
708 struct i40e_ring *p; 752 struct i40e_ring *p;
753 u64 bytes, packets;
754 unsigned int start;
709 755
710 p = &vsi->rx_rings[q]; 756 /* locate Tx ring */
711 rx_b += p->rx_stats.bytes; 757 p = ACCESS_ONCE(vsi->tx_rings[q]);
712 rx_p += p->rx_stats.packets;
713 rx_buf += p->rx_stats.alloc_rx_buff_failed;
714 rx_page += p->rx_stats.alloc_rx_page_failed;
715 758
716 p = &vsi->tx_rings[q]; 759 do {
717 tx_b += p->tx_stats.bytes; 760 start = u64_stats_fetch_begin_bh(&p->syncp);
718 tx_p += p->tx_stats.packets; 761 packets = p->stats.packets;
762 bytes = p->stats.bytes;
763 } while (u64_stats_fetch_retry_bh(&p->syncp, start));
764 tx_b += bytes;
765 tx_p += packets;
719 tx_restart += p->tx_stats.restart_queue; 766 tx_restart += p->tx_stats.restart_queue;
720 tx_busy += p->tx_stats.tx_busy; 767 tx_busy += p->tx_stats.tx_busy;
768
769 /* Rx queue is part of the same block as Tx queue */
770 p = &p[1];
771 do {
772 start = u64_stats_fetch_begin_bh(&p->syncp);
773 packets = p->stats.packets;
774 bytes = p->stats.bytes;
775 } while (u64_stats_fetch_retry_bh(&p->syncp, start));
776 rx_b += bytes;
777 rx_p += packets;
778 rx_buf += p->rx_stats.alloc_rx_buff_failed;
779 rx_page += p->rx_stats.alloc_rx_page_failed;
721 } 780 }
781 rcu_read_unlock();
722 vsi->tx_restart = tx_restart; 782 vsi->tx_restart = tx_restart;
723 vsi->tx_busy = tx_busy; 783 vsi->tx_busy = tx_busy;
724 vsi->rx_page_failed = rx_page; 784 vsi->rx_page_failed = rx_page;
@@ -1988,7 +2048,7 @@ static int i40e_vsi_setup_tx_resources(struct i40e_vsi *vsi)
1988 int i, err = 0; 2048 int i, err = 0;
1989 2049
1990 for (i = 0; i < vsi->num_queue_pairs && !err; i++) 2050 for (i = 0; i < vsi->num_queue_pairs && !err; i++)
1991 err = i40e_setup_tx_descriptors(&vsi->tx_rings[i]); 2051 err = i40e_setup_tx_descriptors(vsi->tx_rings[i]);
1992 2052
1993 return err; 2053 return err;
1994} 2054}
@@ -2004,8 +2064,8 @@ static void i40e_vsi_free_tx_resources(struct i40e_vsi *vsi)
2004 int i; 2064 int i;
2005 2065
2006 for (i = 0; i < vsi->num_queue_pairs; i++) 2066 for (i = 0; i < vsi->num_queue_pairs; i++)
2007 if (vsi->tx_rings[i].desc) 2067 if (vsi->tx_rings[i]->desc)
2008 i40e_free_tx_resources(&vsi->tx_rings[i]); 2068 i40e_free_tx_resources(vsi->tx_rings[i]);
2009} 2069}
2010 2070
2011/** 2071/**
@@ -2023,7 +2083,7 @@ static int i40e_vsi_setup_rx_resources(struct i40e_vsi *vsi)
2023 int i, err = 0; 2083 int i, err = 0;
2024 2084
2025 for (i = 0; i < vsi->num_queue_pairs && !err; i++) 2085 for (i = 0; i < vsi->num_queue_pairs && !err; i++)
2026 err = i40e_setup_rx_descriptors(&vsi->rx_rings[i]); 2086 err = i40e_setup_rx_descriptors(vsi->rx_rings[i]);
2027 return err; 2087 return err;
2028} 2088}
2029 2089
@@ -2038,8 +2098,8 @@ static void i40e_vsi_free_rx_resources(struct i40e_vsi *vsi)
2038 int i; 2098 int i;
2039 2099
2040 for (i = 0; i < vsi->num_queue_pairs; i++) 2100 for (i = 0; i < vsi->num_queue_pairs; i++)
2041 if (vsi->rx_rings[i].desc) 2101 if (vsi->rx_rings[i]->desc)
2042 i40e_free_rx_resources(&vsi->rx_rings[i]); 2102 i40e_free_rx_resources(vsi->rx_rings[i]);
2043} 2103}
2044 2104
2045/** 2105/**
@@ -2114,8 +2174,8 @@ static int i40e_configure_tx_ring(struct i40e_ring *ring)
2114 2174
2115 /* Now associate this queue with this PCI function */ 2175 /* Now associate this queue with this PCI function */
2116 qtx_ctl = I40E_QTX_CTL_PF_QUEUE; 2176 qtx_ctl = I40E_QTX_CTL_PF_QUEUE;
2117 qtx_ctl |= ((hw->hmc.hmc_fn_id << I40E_QTX_CTL_PF_INDX_SHIFT) 2177 qtx_ctl |= ((hw->pf_id << I40E_QTX_CTL_PF_INDX_SHIFT) &
2118 & I40E_QTX_CTL_PF_INDX_MASK); 2178 I40E_QTX_CTL_PF_INDX_MASK);
2119 wr32(hw, I40E_QTX_CTL(pf_q), qtx_ctl); 2179 wr32(hw, I40E_QTX_CTL(pf_q), qtx_ctl);
2120 i40e_flush(hw); 2180 i40e_flush(hw);
2121 2181
@@ -2223,8 +2283,8 @@ static int i40e_vsi_configure_tx(struct i40e_vsi *vsi)
2223 int err = 0; 2283 int err = 0;
2224 u16 i; 2284 u16 i;
2225 2285
2226 for (i = 0; (i < vsi->num_queue_pairs) && (!err); i++) 2286 for (i = 0; (i < vsi->num_queue_pairs) && !err; i++)
2227 err = i40e_configure_tx_ring(&vsi->tx_rings[i]); 2287 err = i40e_configure_tx_ring(vsi->tx_rings[i]);
2228 2288
2229 return err; 2289 return err;
2230} 2290}
@@ -2274,7 +2334,7 @@ static int i40e_vsi_configure_rx(struct i40e_vsi *vsi)
2274 2334
2275 /* set up individual rings */ 2335 /* set up individual rings */
2276 for (i = 0; i < vsi->num_queue_pairs && !err; i++) 2336 for (i = 0; i < vsi->num_queue_pairs && !err; i++)
2277 err = i40e_configure_rx_ring(&vsi->rx_rings[i]); 2337 err = i40e_configure_rx_ring(vsi->rx_rings[i]);
2278 2338
2279 return err; 2339 return err;
2280} 2340}
@@ -2298,8 +2358,8 @@ static void i40e_vsi_config_dcb_rings(struct i40e_vsi *vsi)
2298 qoffset = vsi->tc_config.tc_info[n].qoffset; 2358 qoffset = vsi->tc_config.tc_info[n].qoffset;
2299 qcount = vsi->tc_config.tc_info[n].qcount; 2359 qcount = vsi->tc_config.tc_info[n].qcount;
2300 for (i = qoffset; i < (qoffset + qcount); i++) { 2360 for (i = qoffset; i < (qoffset + qcount); i++) {
2301 struct i40e_ring *rx_ring = &vsi->rx_rings[i]; 2361 struct i40e_ring *rx_ring = vsi->rx_rings[i];
2302 struct i40e_ring *tx_ring = &vsi->tx_rings[i]; 2362 struct i40e_ring *tx_ring = vsi->tx_rings[i];
2303 rx_ring->dcb_tc = n; 2363 rx_ring->dcb_tc = n;
2304 tx_ring->dcb_tc = n; 2364 tx_ring->dcb_tc = n;
2305 } 2365 }
@@ -2354,8 +2414,8 @@ static void i40e_vsi_configure_msix(struct i40e_vsi *vsi)
2354 */ 2414 */
2355 qp = vsi->base_queue; 2415 qp = vsi->base_queue;
2356 vector = vsi->base_vector; 2416 vector = vsi->base_vector;
2357 q_vector = vsi->q_vectors; 2417 for (i = 0; i < vsi->num_q_vectors; i++, vector++) {
2358 for (i = 0; i < vsi->num_q_vectors; i++, q_vector++, vector++) { 2418 q_vector = vsi->q_vectors[i];
2359 q_vector->rx.itr = ITR_TO_REG(vsi->rx_itr_setting); 2419 q_vector->rx.itr = ITR_TO_REG(vsi->rx_itr_setting);
2360 q_vector->rx.latency_range = I40E_LOW_LATENCY; 2420 q_vector->rx.latency_range = I40E_LOW_LATENCY;
2361 wr32(hw, I40E_PFINT_ITRN(I40E_RX_ITR, vector - 1), 2421 wr32(hw, I40E_PFINT_ITRN(I40E_RX_ITR, vector - 1),
@@ -2435,7 +2495,7 @@ static void i40e_enable_misc_int_causes(struct i40e_hw *hw)
2435 **/ 2495 **/
2436static void i40e_configure_msi_and_legacy(struct i40e_vsi *vsi) 2496static void i40e_configure_msi_and_legacy(struct i40e_vsi *vsi)
2437{ 2497{
2438 struct i40e_q_vector *q_vector = vsi->q_vectors; 2498 struct i40e_q_vector *q_vector = vsi->q_vectors[0];
2439 struct i40e_pf *pf = vsi->back; 2499 struct i40e_pf *pf = vsi->back;
2440 struct i40e_hw *hw = &pf->hw; 2500 struct i40e_hw *hw = &pf->hw;
2441 u32 val; 2501 u32 val;
@@ -2472,7 +2532,7 @@ static void i40e_configure_msi_and_legacy(struct i40e_vsi *vsi)
2472 * i40e_irq_dynamic_enable_icr0 - Enable default interrupt generation for icr0 2532 * i40e_irq_dynamic_enable_icr0 - Enable default interrupt generation for icr0
2473 * @pf: board private structure 2533 * @pf: board private structure
2474 **/ 2534 **/
2475static void i40e_irq_dynamic_enable_icr0(struct i40e_pf *pf) 2535void i40e_irq_dynamic_enable_icr0(struct i40e_pf *pf)
2476{ 2536{
2477 struct i40e_hw *hw = &pf->hw; 2537 struct i40e_hw *hw = &pf->hw;
2478 u32 val; 2538 u32 val;
@@ -2500,7 +2560,7 @@ void i40e_irq_dynamic_enable(struct i40e_vsi *vsi, int vector)
2500 I40E_PFINT_DYN_CTLN_CLEARPBA_MASK | 2560 I40E_PFINT_DYN_CTLN_CLEARPBA_MASK |
2501 (I40E_ITR_NONE << I40E_PFINT_DYN_CTLN_ITR_INDX_SHIFT); 2561 (I40E_ITR_NONE << I40E_PFINT_DYN_CTLN_ITR_INDX_SHIFT);
2502 wr32(hw, I40E_PFINT_DYN_CTLN(vector - 1), val); 2562 wr32(hw, I40E_PFINT_DYN_CTLN(vector - 1), val);
2503 i40e_flush(hw); 2563 /* skip the flush */
2504} 2564}
2505 2565
2506/** 2566/**
@@ -2512,7 +2572,7 @@ static irqreturn_t i40e_msix_clean_rings(int irq, void *data)
2512{ 2572{
2513 struct i40e_q_vector *q_vector = data; 2573 struct i40e_q_vector *q_vector = data;
2514 2574
2515 if (!q_vector->tx.ring[0] && !q_vector->rx.ring[0]) 2575 if (!q_vector->tx.ring && !q_vector->rx.ring)
2516 return IRQ_HANDLED; 2576 return IRQ_HANDLED;
2517 2577
2518 napi_schedule(&q_vector->napi); 2578 napi_schedule(&q_vector->napi);
@@ -2529,7 +2589,7 @@ static irqreturn_t i40e_fdir_clean_rings(int irq, void *data)
2529{ 2589{
2530 struct i40e_q_vector *q_vector = data; 2590 struct i40e_q_vector *q_vector = data;
2531 2591
2532 if (!q_vector->tx.ring[0] && !q_vector->rx.ring[0]) 2592 if (!q_vector->tx.ring && !q_vector->rx.ring)
2533 return IRQ_HANDLED; 2593 return IRQ_HANDLED;
2534 2594
2535 pr_info("fdir ring cleaning needed\n"); 2595 pr_info("fdir ring cleaning needed\n");
@@ -2554,16 +2614,16 @@ static int i40e_vsi_request_irq_msix(struct i40e_vsi *vsi, char *basename)
2554 int vector, err; 2614 int vector, err;
2555 2615
2556 for (vector = 0; vector < q_vectors; vector++) { 2616 for (vector = 0; vector < q_vectors; vector++) {
2557 struct i40e_q_vector *q_vector = &(vsi->q_vectors[vector]); 2617 struct i40e_q_vector *q_vector = vsi->q_vectors[vector];
2558 2618
2559 if (q_vector->tx.ring[0] && q_vector->rx.ring[0]) { 2619 if (q_vector->tx.ring && q_vector->rx.ring) {
2560 snprintf(q_vector->name, sizeof(q_vector->name) - 1, 2620 snprintf(q_vector->name, sizeof(q_vector->name) - 1,
2561 "%s-%s-%d", basename, "TxRx", rx_int_idx++); 2621 "%s-%s-%d", basename, "TxRx", rx_int_idx++);
2562 tx_int_idx++; 2622 tx_int_idx++;
2563 } else if (q_vector->rx.ring[0]) { 2623 } else if (q_vector->rx.ring) {
2564 snprintf(q_vector->name, sizeof(q_vector->name) - 1, 2624 snprintf(q_vector->name, sizeof(q_vector->name) - 1,
2565 "%s-%s-%d", basename, "rx", rx_int_idx++); 2625 "%s-%s-%d", basename, "rx", rx_int_idx++);
2566 } else if (q_vector->tx.ring[0]) { 2626 } else if (q_vector->tx.ring) {
2567 snprintf(q_vector->name, sizeof(q_vector->name) - 1, 2627 snprintf(q_vector->name, sizeof(q_vector->name) - 1,
2568 "%s-%s-%d", basename, "tx", tx_int_idx++); 2628 "%s-%s-%d", basename, "tx", tx_int_idx++);
2569 } else { 2629 } else {
@@ -2611,8 +2671,8 @@ static void i40e_vsi_disable_irq(struct i40e_vsi *vsi)
2611 int i; 2671 int i;
2612 2672
2613 for (i = 0; i < vsi->num_queue_pairs; i++) { 2673 for (i = 0; i < vsi->num_queue_pairs; i++) {
2614 wr32(hw, I40E_QINT_TQCTL(vsi->tx_rings[i].reg_idx), 0); 2674 wr32(hw, I40E_QINT_TQCTL(vsi->tx_rings[i]->reg_idx), 0);
2615 wr32(hw, I40E_QINT_RQCTL(vsi->rx_rings[i].reg_idx), 0); 2675 wr32(hw, I40E_QINT_RQCTL(vsi->rx_rings[i]->reg_idx), 0);
2616 } 2676 }
2617 2677
2618 if (pf->flags & I40E_FLAG_MSIX_ENABLED) { 2678 if (pf->flags & I40E_FLAG_MSIX_ENABLED) {
@@ -2649,6 +2709,7 @@ static int i40e_vsi_enable_irq(struct i40e_vsi *vsi)
2649 i40e_irq_dynamic_enable_icr0(pf); 2709 i40e_irq_dynamic_enable_icr0(pf);
2650 } 2710 }
2651 2711
2712 i40e_flush(&pf->hw);
2652 return 0; 2713 return 0;
2653} 2714}
2654 2715
@@ -2681,14 +2742,14 @@ static irqreturn_t i40e_intr(int irq, void *data)
2681 2742
2682 icr0 = rd32(hw, I40E_PFINT_ICR0); 2743 icr0 = rd32(hw, I40E_PFINT_ICR0);
2683 2744
2684 /* if sharing a legacy IRQ, we might get called w/o an intr pending */
2685 if ((icr0 & I40E_PFINT_ICR0_INTEVENT_MASK) == 0)
2686 return IRQ_NONE;
2687
2688 val = rd32(hw, I40E_PFINT_DYN_CTL0); 2745 val = rd32(hw, I40E_PFINT_DYN_CTL0);
2689 val = val | I40E_PFINT_DYN_CTL0_CLEARPBA_MASK; 2746 val = val | I40E_PFINT_DYN_CTL0_CLEARPBA_MASK;
2690 wr32(hw, I40E_PFINT_DYN_CTL0, val); 2747 wr32(hw, I40E_PFINT_DYN_CTL0, val);
2691 2748
2749 /* if sharing a legacy IRQ, we might get called w/o an intr pending */
2750 if ((icr0 & I40E_PFINT_ICR0_INTEVENT_MASK) == 0)
2751 return IRQ_NONE;
2752
2692 ena_mask = rd32(hw, I40E_PFINT_ICR0_ENA); 2753 ena_mask = rd32(hw, I40E_PFINT_ICR0_ENA);
2693 2754
2694 /* only q0 is used in MSI/Legacy mode, and none are used in MSIX */ 2755 /* only q0 is used in MSI/Legacy mode, and none are used in MSIX */
@@ -2702,10 +2763,9 @@ static irqreturn_t i40e_intr(int irq, void *data)
2702 qval = rd32(hw, I40E_QINT_TQCTL(0)); 2763 qval = rd32(hw, I40E_QINT_TQCTL(0));
2703 qval &= ~I40E_QINT_TQCTL_CAUSE_ENA_MASK; 2764 qval &= ~I40E_QINT_TQCTL_CAUSE_ENA_MASK;
2704 wr32(hw, I40E_QINT_TQCTL(0), qval); 2765 wr32(hw, I40E_QINT_TQCTL(0), qval);
2705 i40e_flush(hw);
2706 2766
2707 if (!test_bit(__I40E_DOWN, &pf->state)) 2767 if (!test_bit(__I40E_DOWN, &pf->state))
2708 napi_schedule(&pf->vsi[pf->lan_vsi]->q_vectors[0].napi); 2768 napi_schedule(&pf->vsi[pf->lan_vsi]->q_vectors[0]->napi);
2709 } 2769 }
2710 2770
2711 if (icr0 & I40E_PFINT_ICR0_ADMINQ_MASK) { 2771 if (icr0 & I40E_PFINT_ICR0_ADMINQ_MASK) {
@@ -2764,7 +2824,6 @@ static irqreturn_t i40e_intr(int irq, void *data)
2764 2824
2765 /* re-enable interrupt causes */ 2825 /* re-enable interrupt causes */
2766 wr32(hw, I40E_PFINT_ICR0_ENA, ena_mask); 2826 wr32(hw, I40E_PFINT_ICR0_ENA, ena_mask);
2767 i40e_flush(hw);
2768 if (!test_bit(__I40E_DOWN, &pf->state)) { 2827 if (!test_bit(__I40E_DOWN, &pf->state)) {
2769 i40e_service_event_schedule(pf); 2828 i40e_service_event_schedule(pf);
2770 i40e_irq_dynamic_enable_icr0(pf); 2829 i40e_irq_dynamic_enable_icr0(pf);
@@ -2774,40 +2833,26 @@ static irqreturn_t i40e_intr(int irq, void *data)
2774} 2833}
2775 2834
2776/** 2835/**
2777 * i40e_map_vector_to_rxq - Assigns the Rx queue to the vector 2836 * i40e_map_vector_to_qp - Assigns the queue pair to the vector
2778 * @vsi: the VSI being configured 2837 * @vsi: the VSI being configured
2779 * @v_idx: vector index 2838 * @v_idx: vector index
2780 * @r_idx: rx queue index 2839 * @qp_idx: queue pair index
2781 **/ 2840 **/
2782static void map_vector_to_rxq(struct i40e_vsi *vsi, int v_idx, int r_idx) 2841static void map_vector_to_qp(struct i40e_vsi *vsi, int v_idx, int qp_idx)
2783{ 2842{
2784 struct i40e_q_vector *q_vector = &(vsi->q_vectors[v_idx]); 2843 struct i40e_q_vector *q_vector = vsi->q_vectors[v_idx];
2785 struct i40e_ring *rx_ring = &(vsi->rx_rings[r_idx]); 2844 struct i40e_ring *tx_ring = vsi->tx_rings[qp_idx];
2786 2845 struct i40e_ring *rx_ring = vsi->rx_rings[qp_idx];
2787 rx_ring->q_vector = q_vector;
2788 q_vector->rx.ring[q_vector->rx.count] = rx_ring;
2789 q_vector->rx.count++;
2790 q_vector->rx.latency_range = I40E_LOW_LATENCY;
2791 q_vector->vsi = vsi;
2792}
2793
2794/**
2795 * i40e_map_vector_to_txq - Assigns the Tx queue to the vector
2796 * @vsi: the VSI being configured
2797 * @v_idx: vector index
2798 * @t_idx: tx queue index
2799 **/
2800static void map_vector_to_txq(struct i40e_vsi *vsi, int v_idx, int t_idx)
2801{
2802 struct i40e_q_vector *q_vector = &(vsi->q_vectors[v_idx]);
2803 struct i40e_ring *tx_ring = &(vsi->tx_rings[t_idx]);
2804 2846
2805 tx_ring->q_vector = q_vector; 2847 tx_ring->q_vector = q_vector;
2806 q_vector->tx.ring[q_vector->tx.count] = tx_ring; 2848 tx_ring->next = q_vector->tx.ring;
2849 q_vector->tx.ring = tx_ring;
2807 q_vector->tx.count++; 2850 q_vector->tx.count++;
2808 q_vector->tx.latency_range = I40E_LOW_LATENCY; 2851
2809 q_vector->num_ringpairs++; 2852 rx_ring->q_vector = q_vector;
2810 q_vector->vsi = vsi; 2853 rx_ring->next = q_vector->rx.ring;
2854 q_vector->rx.ring = rx_ring;
2855 q_vector->rx.count++;
2811} 2856}
2812 2857
2813/** 2858/**
@@ -2823,7 +2868,7 @@ static void i40e_vsi_map_rings_to_vectors(struct i40e_vsi *vsi)
2823{ 2868{
2824 int qp_remaining = vsi->num_queue_pairs; 2869 int qp_remaining = vsi->num_queue_pairs;
2825 int q_vectors = vsi->num_q_vectors; 2870 int q_vectors = vsi->num_q_vectors;
2826 int qp_per_vector; 2871 int num_ringpairs;
2827 int v_start = 0; 2872 int v_start = 0;
2828 int qp_idx = 0; 2873 int qp_idx = 0;
2829 2874
@@ -2831,11 +2876,21 @@ static void i40e_vsi_map_rings_to_vectors(struct i40e_vsi *vsi)
2831 * group them so there are multiple queues per vector. 2876 * group them so there are multiple queues per vector.
2832 */ 2877 */
2833 for (; v_start < q_vectors && qp_remaining; v_start++) { 2878 for (; v_start < q_vectors && qp_remaining; v_start++) {
2834 qp_per_vector = DIV_ROUND_UP(qp_remaining, q_vectors - v_start); 2879 struct i40e_q_vector *q_vector = vsi->q_vectors[v_start];
2835 for (; qp_per_vector; 2880
2836 qp_per_vector--, qp_idx++, qp_remaining--) { 2881 num_ringpairs = DIV_ROUND_UP(qp_remaining, q_vectors - v_start);
2837 map_vector_to_rxq(vsi, v_start, qp_idx); 2882
2838 map_vector_to_txq(vsi, v_start, qp_idx); 2883 q_vector->num_ringpairs = num_ringpairs;
2884
2885 q_vector->rx.count = 0;
2886 q_vector->tx.count = 0;
2887 q_vector->rx.ring = NULL;
2888 q_vector->tx.ring = NULL;
2889
2890 while (num_ringpairs--) {
2891 map_vector_to_qp(vsi, v_start, qp_idx);
2892 qp_idx++;
2893 qp_remaining--;
2839 } 2894 }
2840 } 2895 }
2841} 2896}
@@ -2887,7 +2942,7 @@ static void i40e_netpoll(struct net_device *netdev)
2887 pf->flags |= I40E_FLAG_IN_NETPOLL; 2942 pf->flags |= I40E_FLAG_IN_NETPOLL;
2888 if (pf->flags & I40E_FLAG_MSIX_ENABLED) { 2943 if (pf->flags & I40E_FLAG_MSIX_ENABLED) {
2889 for (i = 0; i < vsi->num_q_vectors; i++) 2944 for (i = 0; i < vsi->num_q_vectors; i++)
2890 i40e_msix_clean_rings(0, &vsi->q_vectors[i]); 2945 i40e_msix_clean_rings(0, vsi->q_vectors[i]);
2891 } else { 2946 } else {
2892 i40e_intr(pf->pdev->irq, netdev); 2947 i40e_intr(pf->pdev->irq, netdev);
2893 } 2948 }
@@ -3073,14 +3128,14 @@ static void i40e_vsi_free_irq(struct i40e_vsi *vsi)
3073 u16 vector = i + base; 3128 u16 vector = i + base;
3074 3129
3075 /* free only the irqs that were actually requested */ 3130 /* free only the irqs that were actually requested */
3076 if (vsi->q_vectors[i].num_ringpairs == 0) 3131 if (vsi->q_vectors[i]->num_ringpairs == 0)
3077 continue; 3132 continue;
3078 3133
3079 /* clear the affinity_mask in the IRQ descriptor */ 3134 /* clear the affinity_mask in the IRQ descriptor */
3080 irq_set_affinity_hint(pf->msix_entries[vector].vector, 3135 irq_set_affinity_hint(pf->msix_entries[vector].vector,
3081 NULL); 3136 NULL);
3082 free_irq(pf->msix_entries[vector].vector, 3137 free_irq(pf->msix_entries[vector].vector,
3083 &vsi->q_vectors[i]); 3138 vsi->q_vectors[i]);
3084 3139
3085 /* Tear down the interrupt queue link list 3140 /* Tear down the interrupt queue link list
3086 * 3141 *
@@ -3164,6 +3219,39 @@ static void i40e_vsi_free_irq(struct i40e_vsi *vsi)
3164} 3219}
3165 3220
3166/** 3221/**
3222 * i40e_free_q_vector - Free memory allocated for specific interrupt vector
3223 * @vsi: the VSI being configured
3224 * @v_idx: Index of vector to be freed
3225 *
3226 * This function frees the memory allocated to the q_vector. In addition if
3227 * NAPI is enabled it will delete any references to the NAPI struct prior
3228 * to freeing the q_vector.
3229 **/
3230static void i40e_free_q_vector(struct i40e_vsi *vsi, int v_idx)
3231{
3232 struct i40e_q_vector *q_vector = vsi->q_vectors[v_idx];
3233 struct i40e_ring *ring;
3234
3235 if (!q_vector)
3236 return;
3237
3238 /* disassociate q_vector from rings */
3239 i40e_for_each_ring(ring, q_vector->tx)
3240 ring->q_vector = NULL;
3241
3242 i40e_for_each_ring(ring, q_vector->rx)
3243 ring->q_vector = NULL;
3244
3245 /* only VSI w/ an associated netdev is set up w/ NAPI */
3246 if (vsi->netdev)
3247 netif_napi_del(&q_vector->napi);
3248
3249 vsi->q_vectors[v_idx] = NULL;
3250
3251 kfree_rcu(q_vector, rcu);
3252}
3253
3254/**
3167 * i40e_vsi_free_q_vectors - Free memory allocated for interrupt vectors 3255 * i40e_vsi_free_q_vectors - Free memory allocated for interrupt vectors
3168 * @vsi: the VSI being un-configured 3256 * @vsi: the VSI being un-configured
3169 * 3257 *
@@ -3174,24 +3262,8 @@ static void i40e_vsi_free_q_vectors(struct i40e_vsi *vsi)
3174{ 3262{
3175 int v_idx; 3263 int v_idx;
3176 3264
3177 for (v_idx = 0; v_idx < vsi->num_q_vectors; v_idx++) { 3265 for (v_idx = 0; v_idx < vsi->num_q_vectors; v_idx++)
3178 struct i40e_q_vector *q_vector = &vsi->q_vectors[v_idx]; 3266 i40e_free_q_vector(vsi, v_idx);
3179 int r_idx;
3180
3181 if (!q_vector)
3182 continue;
3183
3184 /* disassociate q_vector from rings */
3185 for (r_idx = 0; r_idx < q_vector->tx.count; r_idx++)
3186 q_vector->tx.ring[r_idx]->q_vector = NULL;
3187 for (r_idx = 0; r_idx < q_vector->rx.count; r_idx++)
3188 q_vector->rx.ring[r_idx]->q_vector = NULL;
3189
3190 /* only VSI w/ an associated netdev is set up w/ NAPI */
3191 if (vsi->netdev)
3192 netif_napi_del(&q_vector->napi);
3193 }
3194 kfree(vsi->q_vectors);
3195} 3267}
3196 3268
3197/** 3269/**
@@ -3241,7 +3313,7 @@ static void i40e_napi_enable_all(struct i40e_vsi *vsi)
3241 return; 3313 return;
3242 3314
3243 for (q_idx = 0; q_idx < vsi->num_q_vectors; q_idx++) 3315 for (q_idx = 0; q_idx < vsi->num_q_vectors; q_idx++)
3244 napi_enable(&vsi->q_vectors[q_idx].napi); 3316 napi_enable(&vsi->q_vectors[q_idx]->napi);
3245} 3317}
3246 3318
3247/** 3319/**
@@ -3256,7 +3328,7 @@ static void i40e_napi_disable_all(struct i40e_vsi *vsi)
3256 return; 3328 return;
3257 3329
3258 for (q_idx = 0; q_idx < vsi->num_q_vectors; q_idx++) 3330 for (q_idx = 0; q_idx < vsi->num_q_vectors; q_idx++)
3259 napi_disable(&vsi->q_vectors[q_idx].napi); 3331 napi_disable(&vsi->q_vectors[q_idx]->napi);
3260} 3332}
3261 3333
3262/** 3334/**
@@ -3703,8 +3775,11 @@ static int i40e_up_complete(struct i40e_vsi *vsi)
3703 3775
3704 if ((pf->hw.phy.link_info.link_info & I40E_AQ_LINK_UP) && 3776 if ((pf->hw.phy.link_info.link_info & I40E_AQ_LINK_UP) &&
3705 (vsi->netdev)) { 3777 (vsi->netdev)) {
3778 netdev_info(vsi->netdev, "NIC Link is Up\n");
3706 netif_tx_start_all_queues(vsi->netdev); 3779 netif_tx_start_all_queues(vsi->netdev);
3707 netif_carrier_on(vsi->netdev); 3780 netif_carrier_on(vsi->netdev);
3781 } else if (vsi->netdev) {
3782 netdev_info(vsi->netdev, "NIC Link is Down\n");
3708 } 3783 }
3709 i40e_service_event_schedule(pf); 3784 i40e_service_event_schedule(pf);
3710 3785
@@ -3772,8 +3847,8 @@ void i40e_down(struct i40e_vsi *vsi)
3772 i40e_napi_disable_all(vsi); 3847 i40e_napi_disable_all(vsi);
3773 3848
3774 for (i = 0; i < vsi->num_queue_pairs; i++) { 3849 for (i = 0; i < vsi->num_queue_pairs; i++) {
3775 i40e_clean_tx_ring(&vsi->tx_rings[i]); 3850 i40e_clean_tx_ring(vsi->tx_rings[i]);
3776 i40e_clean_rx_ring(&vsi->rx_rings[i]); 3851 i40e_clean_rx_ring(vsi->rx_rings[i]);
3777 } 3852 }
3778} 3853}
3779 3854
@@ -4153,8 +4228,9 @@ static void i40e_link_event(struct i40e_pf *pf)
4153 if (new_link == old_link) 4228 if (new_link == old_link)
4154 return; 4229 return;
4155 4230
4156 netdev_info(pf->vsi[pf->lan_vsi]->netdev, 4231 if (!test_bit(__I40E_DOWN, &pf->vsi[pf->lan_vsi]->state))
4157 "NIC Link is %s\n", (new_link ? "Up" : "Down")); 4232 netdev_info(pf->vsi[pf->lan_vsi]->netdev,
4233 "NIC Link is %s\n", (new_link ? "Up" : "Down"));
4158 4234
4159 /* Notify the base of the switch tree connected to 4235 /* Notify the base of the switch tree connected to
4160 * the link. Floating VEBs are not notified. 4236 * the link. Floating VEBs are not notified.
@@ -4199,9 +4275,9 @@ static void i40e_check_hang_subtask(struct i40e_pf *pf)
4199 continue; 4275 continue;
4200 4276
4201 for (i = 0; i < vsi->num_queue_pairs; i++) { 4277 for (i = 0; i < vsi->num_queue_pairs; i++) {
4202 set_check_for_tx_hang(&vsi->tx_rings[i]); 4278 set_check_for_tx_hang(vsi->tx_rings[i]);
4203 if (test_bit(__I40E_HANG_CHECK_ARMED, 4279 if (test_bit(__I40E_HANG_CHECK_ARMED,
4204 &vsi->tx_rings[i].state)) 4280 &vsi->tx_rings[i]->state))
4205 armed++; 4281 armed++;
4206 } 4282 }
4207 4283
@@ -4537,7 +4613,8 @@ static void i40e_fdir_setup(struct i40e_pf *pf)
4537 bool new_vsi = false; 4613 bool new_vsi = false;
4538 int err, i; 4614 int err, i;
4539 4615
4540 if (!(pf->flags & (I40E_FLAG_FDIR_ENABLED|I40E_FLAG_FDIR_ATR_ENABLED))) 4616 if (!(pf->flags & (I40E_FLAG_FDIR_ENABLED |
4617 I40E_FLAG_FDIR_ATR_ENABLED)))
4541 return; 4618 return;
4542 4619
4543 pf->atr_sample_rate = I40E_DEFAULT_ATR_SAMPLE_RATE; 4620 pf->atr_sample_rate = I40E_DEFAULT_ATR_SAMPLE_RATE;
@@ -4937,6 +5014,8 @@ static int i40e_vsi_mem_alloc(struct i40e_pf *pf, enum i40e_vsi_type type)
4937{ 5014{
4938 int ret = -ENODEV; 5015 int ret = -ENODEV;
4939 struct i40e_vsi *vsi; 5016 struct i40e_vsi *vsi;
5017 int sz_vectors;
5018 int sz_rings;
4940 int vsi_idx; 5019 int vsi_idx;
4941 int i; 5020 int i;
4942 5021
@@ -4962,14 +5041,14 @@ static int i40e_vsi_mem_alloc(struct i40e_pf *pf, enum i40e_vsi_type type)
4962 vsi_idx = i; /* Found one! */ 5041 vsi_idx = i; /* Found one! */
4963 } else { 5042 } else {
4964 ret = -ENODEV; 5043 ret = -ENODEV;
4965 goto err_alloc_vsi; /* out of VSI slots! */ 5044 goto unlock_pf; /* out of VSI slots! */
4966 } 5045 }
4967 pf->next_vsi = ++i; 5046 pf->next_vsi = ++i;
4968 5047
4969 vsi = kzalloc(sizeof(*vsi), GFP_KERNEL); 5048 vsi = kzalloc(sizeof(*vsi), GFP_KERNEL);
4970 if (!vsi) { 5049 if (!vsi) {
4971 ret = -ENOMEM; 5050 ret = -ENOMEM;
4972 goto err_alloc_vsi; 5051 goto unlock_pf;
4973 } 5052 }
4974 vsi->type = type; 5053 vsi->type = type;
4975 vsi->back = pf; 5054 vsi->back = pf;
@@ -4982,14 +5061,40 @@ static int i40e_vsi_mem_alloc(struct i40e_pf *pf, enum i40e_vsi_type type)
4982 vsi->work_limit = I40E_DEFAULT_IRQ_WORK; 5061 vsi->work_limit = I40E_DEFAULT_IRQ_WORK;
4983 INIT_LIST_HEAD(&vsi->mac_filter_list); 5062 INIT_LIST_HEAD(&vsi->mac_filter_list);
4984 5063
4985 i40e_set_num_rings_in_vsi(vsi); 5064 ret = i40e_set_num_rings_in_vsi(vsi);
5065 if (ret)
5066 goto err_rings;
5067
5068 /* allocate memory for ring pointers */
5069 sz_rings = sizeof(struct i40e_ring *) * vsi->alloc_queue_pairs * 2;
5070 vsi->tx_rings = kzalloc(sz_rings, GFP_KERNEL);
5071 if (!vsi->tx_rings) {
5072 ret = -ENOMEM;
5073 goto err_rings;
5074 }
5075 vsi->rx_rings = &vsi->tx_rings[vsi->alloc_queue_pairs];
5076
5077 /* allocate memory for q_vector pointers */
5078 sz_vectors = sizeof(struct i40e_q_vectors *) * vsi->num_q_vectors;
5079 vsi->q_vectors = kzalloc(sz_vectors, GFP_KERNEL);
5080 if (!vsi->q_vectors) {
5081 ret = -ENOMEM;
5082 goto err_vectors;
5083 }
4986 5084
4987 /* Setup default MSIX irq handler for VSI */ 5085 /* Setup default MSIX irq handler for VSI */
4988 i40e_vsi_setup_irqhandler(vsi, i40e_msix_clean_rings); 5086 i40e_vsi_setup_irqhandler(vsi, i40e_msix_clean_rings);
4989 5087
4990 pf->vsi[vsi_idx] = vsi; 5088 pf->vsi[vsi_idx] = vsi;
4991 ret = vsi_idx; 5089 ret = vsi_idx;
4992err_alloc_vsi: 5090 goto unlock_pf;
5091
5092err_vectors:
5093 kfree(vsi->tx_rings);
5094err_rings:
5095 pf->next_vsi = i - 1;
5096 kfree(vsi);
5097unlock_pf:
4993 mutex_unlock(&pf->switch_mutex); 5098 mutex_unlock(&pf->switch_mutex);
4994 return ret; 5099 return ret;
4995} 5100}
@@ -5030,6 +5135,10 @@ static int i40e_vsi_clear(struct i40e_vsi *vsi)
5030 i40e_put_lump(pf->qp_pile, vsi->base_queue, vsi->idx); 5135 i40e_put_lump(pf->qp_pile, vsi->base_queue, vsi->idx);
5031 i40e_put_lump(pf->irq_pile, vsi->base_vector, vsi->idx); 5136 i40e_put_lump(pf->irq_pile, vsi->base_vector, vsi->idx);
5032 5137
5138 /* free the ring and vector containers */
5139 kfree(vsi->q_vectors);
5140 kfree(vsi->tx_rings);
5141
5033 pf->vsi[vsi->idx] = NULL; 5142 pf->vsi[vsi->idx] = NULL;
5034 if (vsi->idx < pf->next_vsi) 5143 if (vsi->idx < pf->next_vsi)
5035 pf->next_vsi = vsi->idx; 5144 pf->next_vsi = vsi->idx;
@@ -5043,34 +5152,40 @@ free_vsi:
5043} 5152}
5044 5153
5045/** 5154/**
5155 * i40e_vsi_clear_rings - Deallocates the Rx and Tx rings for the provided VSI
5156 * @vsi: the VSI being cleaned
5157 **/
5158static s32 i40e_vsi_clear_rings(struct i40e_vsi *vsi)
5159{
5160 int i;
5161
5162 if (vsi->tx_rings[0])
5163 for (i = 0; i < vsi->alloc_queue_pairs; i++) {
5164 kfree_rcu(vsi->tx_rings[i], rcu);
5165 vsi->tx_rings[i] = NULL;
5166 vsi->rx_rings[i] = NULL;
5167 }
5168
5169 return 0;
5170}
5171
5172/**
5046 * i40e_alloc_rings - Allocates the Rx and Tx rings for the provided VSI 5173 * i40e_alloc_rings - Allocates the Rx and Tx rings for the provided VSI
5047 * @vsi: the VSI being configured 5174 * @vsi: the VSI being configured
5048 **/ 5175 **/
5049static int i40e_alloc_rings(struct i40e_vsi *vsi) 5176static int i40e_alloc_rings(struct i40e_vsi *vsi)
5050{ 5177{
5051 struct i40e_pf *pf = vsi->back; 5178 struct i40e_pf *pf = vsi->back;
5052 int ret = 0;
5053 int i; 5179 int i;
5054 5180
5055 vsi->rx_rings = kcalloc(vsi->alloc_queue_pairs,
5056 sizeof(struct i40e_ring), GFP_KERNEL);
5057 if (!vsi->rx_rings) {
5058 ret = -ENOMEM;
5059 goto err_alloc_rings;
5060 }
5061
5062 vsi->tx_rings = kcalloc(vsi->alloc_queue_pairs,
5063 sizeof(struct i40e_ring), GFP_KERNEL);
5064 if (!vsi->tx_rings) {
5065 ret = -ENOMEM;
5066 kfree(vsi->rx_rings);
5067 goto err_alloc_rings;
5068 }
5069
5070 /* Set basic values in the rings to be used later during open() */ 5181 /* Set basic values in the rings to be used later during open() */
5071 for (i = 0; i < vsi->alloc_queue_pairs; i++) { 5182 for (i = 0; i < vsi->alloc_queue_pairs; i++) {
5072 struct i40e_ring *rx_ring = &vsi->rx_rings[i]; 5183 struct i40e_ring *tx_ring;
5073 struct i40e_ring *tx_ring = &vsi->tx_rings[i]; 5184 struct i40e_ring *rx_ring;
5185
5186 tx_ring = kzalloc(sizeof(struct i40e_ring) * 2, GFP_KERNEL);
5187 if (!tx_ring)
5188 goto err_out;
5074 5189
5075 tx_ring->queue_index = i; 5190 tx_ring->queue_index = i;
5076 tx_ring->reg_idx = vsi->base_queue + i; 5191 tx_ring->reg_idx = vsi->base_queue + i;
@@ -5081,7 +5196,9 @@ static int i40e_alloc_rings(struct i40e_vsi *vsi)
5081 tx_ring->count = vsi->num_desc; 5196 tx_ring->count = vsi->num_desc;
5082 tx_ring->size = 0; 5197 tx_ring->size = 0;
5083 tx_ring->dcb_tc = 0; 5198 tx_ring->dcb_tc = 0;
5199 vsi->tx_rings[i] = tx_ring;
5084 5200
5201 rx_ring = &tx_ring[1];
5085 rx_ring->queue_index = i; 5202 rx_ring->queue_index = i;
5086 rx_ring->reg_idx = vsi->base_queue + i; 5203 rx_ring->reg_idx = vsi->base_queue + i;
5087 rx_ring->ring_active = false; 5204 rx_ring->ring_active = false;
@@ -5095,24 +5212,14 @@ static int i40e_alloc_rings(struct i40e_vsi *vsi)
5095 set_ring_16byte_desc_enabled(rx_ring); 5212 set_ring_16byte_desc_enabled(rx_ring);
5096 else 5213 else
5097 clear_ring_16byte_desc_enabled(rx_ring); 5214 clear_ring_16byte_desc_enabled(rx_ring);
5098 } 5215 vsi->rx_rings[i] = rx_ring;
5099
5100err_alloc_rings:
5101 return ret;
5102}
5103
5104/**
5105 * i40e_vsi_clear_rings - Deallocates the Rx and Tx rings for the provided VSI
5106 * @vsi: the VSI being cleaned
5107 **/
5108static int i40e_vsi_clear_rings(struct i40e_vsi *vsi)
5109{
5110 if (vsi) {
5111 kfree(vsi->rx_rings);
5112 kfree(vsi->tx_rings);
5113 } 5216 }
5114 5217
5115 return 0; 5218 return 0;
5219
5220err_out:
5221 i40e_vsi_clear_rings(vsi);
5222 return -ENOMEM;
5116} 5223}
5117 5224
5118/** 5225/**
@@ -5249,6 +5356,38 @@ static int i40e_init_msix(struct i40e_pf *pf)
5249} 5356}
5250 5357
5251/** 5358/**
5359 * i40e_alloc_q_vector - Allocate memory for a single interrupt vector
5360 * @vsi: the VSI being configured
5361 * @v_idx: index of the vector in the vsi struct
5362 *
5363 * We allocate one q_vector. If allocation fails we return -ENOMEM.
5364 **/
5365static int i40e_alloc_q_vector(struct i40e_vsi *vsi, int v_idx)
5366{
5367 struct i40e_q_vector *q_vector;
5368
5369 /* allocate q_vector */
5370 q_vector = kzalloc(sizeof(struct i40e_q_vector), GFP_KERNEL);
5371 if (!q_vector)
5372 return -ENOMEM;
5373
5374 q_vector->vsi = vsi;
5375 q_vector->v_idx = v_idx;
5376 cpumask_set_cpu(v_idx, &q_vector->affinity_mask);
5377 if (vsi->netdev)
5378 netif_napi_add(vsi->netdev, &q_vector->napi,
5379 i40e_napi_poll, vsi->work_limit);
5380
5381 q_vector->rx.latency_range = I40E_LOW_LATENCY;
5382 q_vector->tx.latency_range = I40E_LOW_LATENCY;
5383
5384 /* tie q_vector and vsi together */
5385 vsi->q_vectors[v_idx] = q_vector;
5386
5387 return 0;
5388}
5389
5390/**
5252 * i40e_alloc_q_vectors - Allocate memory for interrupt vectors 5391 * i40e_alloc_q_vectors - Allocate memory for interrupt vectors
5253 * @vsi: the VSI being configured 5392 * @vsi: the VSI being configured
5254 * 5393 *
@@ -5259,6 +5398,7 @@ static int i40e_alloc_q_vectors(struct i40e_vsi *vsi)
5259{ 5398{
5260 struct i40e_pf *pf = vsi->back; 5399 struct i40e_pf *pf = vsi->back;
5261 int v_idx, num_q_vectors; 5400 int v_idx, num_q_vectors;
5401 int err;
5262 5402
5263 /* if not MSIX, give the one vector only to the LAN VSI */ 5403 /* if not MSIX, give the one vector only to the LAN VSI */
5264 if (pf->flags & I40E_FLAG_MSIX_ENABLED) 5404 if (pf->flags & I40E_FLAG_MSIX_ENABLED)
@@ -5268,22 +5408,19 @@ static int i40e_alloc_q_vectors(struct i40e_vsi *vsi)
5268 else 5408 else
5269 return -EINVAL; 5409 return -EINVAL;
5270 5410
5271 vsi->q_vectors = kcalloc(num_q_vectors,
5272 sizeof(struct i40e_q_vector),
5273 GFP_KERNEL);
5274 if (!vsi->q_vectors)
5275 return -ENOMEM;
5276
5277 for (v_idx = 0; v_idx < num_q_vectors; v_idx++) { 5411 for (v_idx = 0; v_idx < num_q_vectors; v_idx++) {
5278 vsi->q_vectors[v_idx].vsi = vsi; 5412 err = i40e_alloc_q_vector(vsi, v_idx);
5279 vsi->q_vectors[v_idx].v_idx = v_idx; 5413 if (err)
5280 cpumask_set_cpu(v_idx, &vsi->q_vectors[v_idx].affinity_mask); 5414 goto err_out;
5281 if (vsi->netdev)
5282 netif_napi_add(vsi->netdev, &vsi->q_vectors[v_idx].napi,
5283 i40e_napi_poll, vsi->work_limit);
5284 } 5415 }
5285 5416
5286 return 0; 5417 return 0;
5418
5419err_out:
5420 while (v_idx--)
5421 i40e_free_q_vector(vsi, v_idx);
5422
5423 return err;
5287} 5424}
5288 5425
5289/** 5426/**
@@ -5297,7 +5434,8 @@ static void i40e_init_interrupt_scheme(struct i40e_pf *pf)
5297 if (pf->flags & I40E_FLAG_MSIX_ENABLED) { 5434 if (pf->flags & I40E_FLAG_MSIX_ENABLED) {
5298 err = i40e_init_msix(pf); 5435 err = i40e_init_msix(pf);
5299 if (err) { 5436 if (err) {
5300 pf->flags &= ~(I40E_FLAG_RSS_ENABLED | 5437 pf->flags &= ~(I40E_FLAG_MSIX_ENABLED |
5438 I40E_FLAG_RSS_ENABLED |
5301 I40E_FLAG_MQ_ENABLED | 5439 I40E_FLAG_MQ_ENABLED |
5302 I40E_FLAG_DCB_ENABLED | 5440 I40E_FLAG_DCB_ENABLED |
5303 I40E_FLAG_SRIOV_ENABLED | 5441 I40E_FLAG_SRIOV_ENABLED |
@@ -5312,14 +5450,17 @@ static void i40e_init_interrupt_scheme(struct i40e_pf *pf)
5312 5450
5313 if (!(pf->flags & I40E_FLAG_MSIX_ENABLED) && 5451 if (!(pf->flags & I40E_FLAG_MSIX_ENABLED) &&
5314 (pf->flags & I40E_FLAG_MSI_ENABLED)) { 5452 (pf->flags & I40E_FLAG_MSI_ENABLED)) {
5453 dev_info(&pf->pdev->dev, "MSIX not available, trying MSI\n");
5315 err = pci_enable_msi(pf->pdev); 5454 err = pci_enable_msi(pf->pdev);
5316 if (err) { 5455 if (err) {
5317 dev_info(&pf->pdev->dev, 5456 dev_info(&pf->pdev->dev, "MSI init failed - %d\n", err);
5318 "MSI init failed (%d), trying legacy.\n", err);
5319 pf->flags &= ~I40E_FLAG_MSI_ENABLED; 5457 pf->flags &= ~I40E_FLAG_MSI_ENABLED;
5320 } 5458 }
5321 } 5459 }
5322 5460
5461 if (!(pf->flags & (I40E_FLAG_MSIX_ENABLED | I40E_FLAG_MSI_ENABLED)))
5462 dev_info(&pf->pdev->dev, "MSIX and MSI not available, falling back to Legacy IRQ\n");
5463
5323 /* track first vector for misc interrupts */ 5464 /* track first vector for misc interrupts */
5324 err = i40e_get_lump(pf, pf->irq_pile, 1, I40E_PILE_VALID_BIT-1); 5465 err = i40e_get_lump(pf, pf->irq_pile, 1, I40E_PILE_VALID_BIT-1);
5325} 5466}
@@ -5950,7 +6091,7 @@ static int i40e_vsi_setup_vectors(struct i40e_vsi *vsi)
5950 int ret = -ENOENT; 6091 int ret = -ENOENT;
5951 struct i40e_pf *pf = vsi->back; 6092 struct i40e_pf *pf = vsi->back;
5952 6093
5953 if (vsi->q_vectors) { 6094 if (vsi->q_vectors[0]) {
5954 dev_info(&pf->pdev->dev, "VSI %d has existing q_vectors\n", 6095 dev_info(&pf->pdev->dev, "VSI %d has existing q_vectors\n",
5955 vsi->seid); 6096 vsi->seid);
5956 return -EEXIST; 6097 return -EEXIST;
@@ -5972,8 +6113,9 @@ static int i40e_vsi_setup_vectors(struct i40e_vsi *vsi)
5972 goto vector_setup_out; 6113 goto vector_setup_out;
5973 } 6114 }
5974 6115
5975 vsi->base_vector = i40e_get_lump(pf, pf->irq_pile, 6116 if (vsi->num_q_vectors)
5976 vsi->num_q_vectors, vsi->idx); 6117 vsi->base_vector = i40e_get_lump(pf, pf->irq_pile,
6118 vsi->num_q_vectors, vsi->idx);
5977 if (vsi->base_vector < 0) { 6119 if (vsi->base_vector < 0) {
5978 dev_info(&pf->pdev->dev, 6120 dev_info(&pf->pdev->dev,
5979 "failed to get q tracking for VSI %d, err=%d\n", 6121 "failed to get q tracking for VSI %d, err=%d\n",
@@ -7062,8 +7204,10 @@ static int i40e_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
7062 */ 7204 */
7063 len = sizeof(struct i40e_vsi *) * pf->hw.func_caps.num_vsis; 7205 len = sizeof(struct i40e_vsi *) * pf->hw.func_caps.num_vsis;
7064 pf->vsi = kzalloc(len, GFP_KERNEL); 7206 pf->vsi = kzalloc(len, GFP_KERNEL);
7065 if (!pf->vsi) 7207 if (!pf->vsi) {
7208 err = -ENOMEM;
7066 goto err_switch_setup; 7209 goto err_switch_setup;
7210 }
7067 7211
7068 err = i40e_setup_pf_switch(pf); 7212 err = i40e_setup_pf_switch(pf);
7069 if (err) { 7213 if (err) {
diff --git a/drivers/net/ethernet/intel/i40e/i40e_txrx.c b/drivers/net/ethernet/intel/i40e/i40e_txrx.c
index 49d2cfa9b0cc..f1f03bc5c729 100644
--- a/drivers/net/ethernet/intel/i40e/i40e_txrx.c
+++ b/drivers/net/ethernet/intel/i40e/i40e_txrx.c
@@ -37,6 +37,7 @@ static inline __le64 build_ctob(u32 td_cmd, u32 td_offset, unsigned int size,
37 ((u64)td_tag << I40E_TXD_QW1_L2TAG1_SHIFT)); 37 ((u64)td_tag << I40E_TXD_QW1_L2TAG1_SHIFT));
38} 38}
39 39
40#define I40E_TXD_CMD (I40E_TX_DESC_CMD_EOP | I40E_TX_DESC_CMD_RS)
40/** 41/**
41 * i40e_program_fdir_filter - Program a Flow Director filter 42 * i40e_program_fdir_filter - Program a Flow Director filter
42 * @fdir_input: Packet data that will be filter parameters 43 * @fdir_input: Packet data that will be filter parameters
@@ -50,6 +51,7 @@ int i40e_program_fdir_filter(struct i40e_fdir_data *fdir_data,
50 struct i40e_tx_buffer *tx_buf; 51 struct i40e_tx_buffer *tx_buf;
51 struct i40e_tx_desc *tx_desc; 52 struct i40e_tx_desc *tx_desc;
52 struct i40e_ring *tx_ring; 53 struct i40e_ring *tx_ring;
54 unsigned int fpt, dcc;
53 struct i40e_vsi *vsi; 55 struct i40e_vsi *vsi;
54 struct device *dev; 56 struct device *dev;
55 dma_addr_t dma; 57 dma_addr_t dma;
@@ -64,93 +66,78 @@ int i40e_program_fdir_filter(struct i40e_fdir_data *fdir_data,
64 if (!vsi) 66 if (!vsi)
65 return -ENOENT; 67 return -ENOENT;
66 68
67 tx_ring = &vsi->tx_rings[0]; 69 tx_ring = vsi->tx_rings[0];
68 dev = tx_ring->dev; 70 dev = tx_ring->dev;
69 71
70 dma = dma_map_single(dev, fdir_data->raw_packet, 72 dma = dma_map_single(dev, fdir_data->raw_packet,
71 I40E_FDIR_MAX_RAW_PACKET_LOOKUP, DMA_TO_DEVICE); 73 I40E_FDIR_MAX_RAW_PACKET_LOOKUP, DMA_TO_DEVICE);
72 if (dma_mapping_error(dev, dma)) 74 if (dma_mapping_error(dev, dma))
73 goto dma_fail; 75 goto dma_fail;
74 76
75 /* grab the next descriptor */ 77 /* grab the next descriptor */
76 fdir_desc = I40E_TX_FDIRDESC(tx_ring, tx_ring->next_to_use); 78 i = tx_ring->next_to_use;
77 tx_buf = &tx_ring->tx_bi[tx_ring->next_to_use]; 79 fdir_desc = I40E_TX_FDIRDESC(tx_ring, i);
78 tx_ring->next_to_use++; 80 tx_buf = &tx_ring->tx_bi[i];
79 if (tx_ring->next_to_use == tx_ring->count) 81
80 tx_ring->next_to_use = 0; 82 tx_ring->next_to_use = (i + 1 < tx_ring->count) ? i + 1 : 0;
81 83
82 fdir_desc->qindex_flex_ptype_vsi = cpu_to_le32((fdir_data->q_index 84 fpt = (fdir_data->q_index << I40E_TXD_FLTR_QW0_QINDEX_SHIFT) &
83 << I40E_TXD_FLTR_QW0_QINDEX_SHIFT) 85 I40E_TXD_FLTR_QW0_QINDEX_MASK;
84 & I40E_TXD_FLTR_QW0_QINDEX_MASK);
85 86
86 fdir_desc->qindex_flex_ptype_vsi |= cpu_to_le32((fdir_data->flex_off 87 fpt |= (fdir_data->flex_off << I40E_TXD_FLTR_QW0_FLEXOFF_SHIFT) &
87 << I40E_TXD_FLTR_QW0_FLEXOFF_SHIFT) 88 I40E_TXD_FLTR_QW0_FLEXOFF_MASK;
88 & I40E_TXD_FLTR_QW0_FLEXOFF_MASK);
89 89
90 fdir_desc->qindex_flex_ptype_vsi |= cpu_to_le32((fdir_data->pctype 90 fpt |= (fdir_data->pctype << I40E_TXD_FLTR_QW0_PCTYPE_SHIFT) &
91 << I40E_TXD_FLTR_QW0_PCTYPE_SHIFT) 91 I40E_TXD_FLTR_QW0_PCTYPE_MASK;
92 & I40E_TXD_FLTR_QW0_PCTYPE_MASK);
93 92
94 /* Use LAN VSI Id if not programmed by user */ 93 /* Use LAN VSI Id if not programmed by user */
95 if (fdir_data->dest_vsi == 0) 94 if (fdir_data->dest_vsi == 0)
96 fdir_desc->qindex_flex_ptype_vsi |= 95 fpt |= (pf->vsi[pf->lan_vsi]->id) <<
97 cpu_to_le32((pf->vsi[pf->lan_vsi]->id) 96 I40E_TXD_FLTR_QW0_DEST_VSI_SHIFT;
98 << I40E_TXD_FLTR_QW0_DEST_VSI_SHIFT);
99 else 97 else
100 fdir_desc->qindex_flex_ptype_vsi |= 98 fpt |= ((u32)fdir_data->dest_vsi <<
101 cpu_to_le32((fdir_data->dest_vsi 99 I40E_TXD_FLTR_QW0_DEST_VSI_SHIFT) &
102 << I40E_TXD_FLTR_QW0_DEST_VSI_SHIFT) 100 I40E_TXD_FLTR_QW0_DEST_VSI_MASK;
103 & I40E_TXD_FLTR_QW0_DEST_VSI_MASK);
104 101
105 fdir_desc->dtype_cmd_cntindex = 102 fdir_desc->qindex_flex_ptype_vsi = cpu_to_le32(fpt);
106 cpu_to_le32(I40E_TX_DESC_DTYPE_FILTER_PROG); 103
104 dcc = I40E_TX_DESC_DTYPE_FILTER_PROG;
107 105
108 if (add) 106 if (add)
109 fdir_desc->dtype_cmd_cntindex |= cpu_to_le32( 107 dcc |= I40E_FILTER_PROGRAM_DESC_PCMD_ADD_UPDATE <<
110 I40E_FILTER_PROGRAM_DESC_PCMD_ADD_UPDATE 108 I40E_TXD_FLTR_QW1_PCMD_SHIFT;
111 << I40E_TXD_FLTR_QW1_PCMD_SHIFT);
112 else 109 else
113 fdir_desc->dtype_cmd_cntindex |= cpu_to_le32( 110 dcc |= I40E_FILTER_PROGRAM_DESC_PCMD_REMOVE <<
114 I40E_FILTER_PROGRAM_DESC_PCMD_REMOVE 111 I40E_TXD_FLTR_QW1_PCMD_SHIFT;
115 << I40E_TXD_FLTR_QW1_PCMD_SHIFT);
116 112
117 fdir_desc->dtype_cmd_cntindex |= cpu_to_le32((fdir_data->dest_ctl 113 dcc |= (fdir_data->dest_ctl << I40E_TXD_FLTR_QW1_DEST_SHIFT) &
118 << I40E_TXD_FLTR_QW1_DEST_SHIFT) 114 I40E_TXD_FLTR_QW1_DEST_MASK;
119 & I40E_TXD_FLTR_QW1_DEST_MASK);
120 115
121 fdir_desc->dtype_cmd_cntindex |= cpu_to_le32( 116 dcc |= (fdir_data->fd_status << I40E_TXD_FLTR_QW1_FD_STATUS_SHIFT) &
122 (fdir_data->fd_status << I40E_TXD_FLTR_QW1_FD_STATUS_SHIFT) 117 I40E_TXD_FLTR_QW1_FD_STATUS_MASK;
123 & I40E_TXD_FLTR_QW1_FD_STATUS_MASK);
124 118
125 if (fdir_data->cnt_index != 0) { 119 if (fdir_data->cnt_index != 0) {
126 fdir_desc->dtype_cmd_cntindex |= 120 dcc |= I40E_TXD_FLTR_QW1_CNT_ENA_MASK;
127 cpu_to_le32(I40E_TXD_FLTR_QW1_CNT_ENA_MASK); 121 dcc |= ((u32)fdir_data->cnt_index <<
128 fdir_desc->dtype_cmd_cntindex |= 122 I40E_TXD_FLTR_QW1_CNTINDEX_SHIFT) &
129 cpu_to_le32((fdir_data->cnt_index 123 I40E_TXD_FLTR_QW1_CNTINDEX_MASK;
130 << I40E_TXD_FLTR_QW1_CNTINDEX_SHIFT)
131 & I40E_TXD_FLTR_QW1_CNTINDEX_MASK);
132 } 124 }
133 125
126 fdir_desc->dtype_cmd_cntindex = cpu_to_le32(dcc);
134 fdir_desc->fd_id = cpu_to_le32(fdir_data->fd_id); 127 fdir_desc->fd_id = cpu_to_le32(fdir_data->fd_id);
135 128
136 /* Now program a dummy descriptor */ 129 /* Now program a dummy descriptor */
137 tx_desc = I40E_TX_DESC(tx_ring, tx_ring->next_to_use); 130 i = tx_ring->next_to_use;
138 tx_buf = &tx_ring->tx_bi[tx_ring->next_to_use]; 131 tx_desc = I40E_TX_DESC(tx_ring, i);
139 tx_ring->next_to_use++; 132
140 if (tx_ring->next_to_use == tx_ring->count) 133 tx_ring->next_to_use = (i + 1 < tx_ring->count) ? i + 1 : 0;
141 tx_ring->next_to_use = 0;
142 134
143 tx_desc->buffer_addr = cpu_to_le64(dma); 135 tx_desc->buffer_addr = cpu_to_le64(dma);
144 td_cmd = I40E_TX_DESC_CMD_EOP | 136 td_cmd = I40E_TXD_CMD | I40E_TX_DESC_CMD_DUMMY;
145 I40E_TX_DESC_CMD_RS |
146 I40E_TX_DESC_CMD_DUMMY;
147 137
148 tx_desc->cmd_type_offset_bsz = 138 tx_desc->cmd_type_offset_bsz =
149 build_ctob(td_cmd, 0, I40E_FDIR_MAX_RAW_PACKET_LOOKUP, 0); 139 build_ctob(td_cmd, 0, I40E_FDIR_MAX_RAW_PACKET_LOOKUP, 0);
150 140
151 /* Mark the data descriptor to be watched */
152 tx_buf->next_to_watch = tx_desc;
153
154 /* Force memory writes to complete before letting h/w 141 /* Force memory writes to complete before letting h/w
155 * know there are new descriptors to fetch. (Only 142 * know there are new descriptors to fetch. (Only
156 * applicable for weak-ordered memory model archs, 143 * applicable for weak-ordered memory model archs,
@@ -158,6 +145,9 @@ int i40e_program_fdir_filter(struct i40e_fdir_data *fdir_data,
158 */ 145 */
159 wmb(); 146 wmb();
160 147
148 /* Mark the data descriptor to be watched */
149 tx_buf->next_to_watch = tx_desc;
150
161 writel(tx_ring->next_to_use, tx_ring->tail); 151 writel(tx_ring->next_to_use, tx_ring->tail);
162 return 0; 152 return 0;
163 153
@@ -188,27 +178,30 @@ static void i40e_fd_handle_status(struct i40e_ring *rx_ring, u32 qw, u8 prog_id)
188} 178}
189 179
190/** 180/**
191 * i40e_unmap_tx_resource - Release a Tx buffer 181 * i40e_unmap_and_free_tx_resource - Release a Tx buffer
192 * @ring: the ring that owns the buffer 182 * @ring: the ring that owns the buffer
193 * @tx_buffer: the buffer to free 183 * @tx_buffer: the buffer to free
194 **/ 184 **/
195static inline void i40e_unmap_tx_resource(struct i40e_ring *ring, 185static void i40e_unmap_and_free_tx_resource(struct i40e_ring *ring,
196 struct i40e_tx_buffer *tx_buffer) 186 struct i40e_tx_buffer *tx_buffer)
197{ 187{
198 if (tx_buffer->dma) { 188 if (tx_buffer->skb) {
199 if (tx_buffer->tx_flags & I40E_TX_FLAGS_MAPPED_AS_PAGE) 189 dev_kfree_skb_any(tx_buffer->skb);
200 dma_unmap_page(ring->dev, 190 if (dma_unmap_len(tx_buffer, len))
201 tx_buffer->dma,
202 tx_buffer->length,
203 DMA_TO_DEVICE);
204 else
205 dma_unmap_single(ring->dev, 191 dma_unmap_single(ring->dev,
206 tx_buffer->dma, 192 dma_unmap_addr(tx_buffer, dma),
207 tx_buffer->length, 193 dma_unmap_len(tx_buffer, len),
208 DMA_TO_DEVICE); 194 DMA_TO_DEVICE);
195 } else if (dma_unmap_len(tx_buffer, len)) {
196 dma_unmap_page(ring->dev,
197 dma_unmap_addr(tx_buffer, dma),
198 dma_unmap_len(tx_buffer, len),
199 DMA_TO_DEVICE);
209 } 200 }
210 tx_buffer->dma = 0; 201 tx_buffer->next_to_watch = NULL;
211 tx_buffer->time_stamp = 0; 202 tx_buffer->skb = NULL;
203 dma_unmap_len_set(tx_buffer, len, 0);
204 /* tx_buffer must be completely set up in the transmit path */
212} 205}
213 206
214/** 207/**
@@ -217,7 +210,6 @@ static inline void i40e_unmap_tx_resource(struct i40e_ring *ring,
217 **/ 210 **/
218void i40e_clean_tx_ring(struct i40e_ring *tx_ring) 211void i40e_clean_tx_ring(struct i40e_ring *tx_ring)
219{ 212{
220 struct i40e_tx_buffer *tx_buffer;
221 unsigned long bi_size; 213 unsigned long bi_size;
222 u16 i; 214 u16 i;
223 215
@@ -226,13 +218,8 @@ void i40e_clean_tx_ring(struct i40e_ring *tx_ring)
226 return; 218 return;
227 219
228 /* Free all the Tx ring sk_buffs */ 220 /* Free all the Tx ring sk_buffs */
229 for (i = 0; i < tx_ring->count; i++) { 221 for (i = 0; i < tx_ring->count; i++)
230 tx_buffer = &tx_ring->tx_bi[i]; 222 i40e_unmap_and_free_tx_resource(tx_ring, &tx_ring->tx_bi[i]);
231 i40e_unmap_tx_resource(tx_ring, tx_buffer);
232 if (tx_buffer->skb)
233 dev_kfree_skb_any(tx_buffer->skb);
234 tx_buffer->skb = NULL;
235 }
236 223
237 bi_size = sizeof(struct i40e_tx_buffer) * tx_ring->count; 224 bi_size = sizeof(struct i40e_tx_buffer) * tx_ring->count;
238 memset(tx_ring->tx_bi, 0, bi_size); 225 memset(tx_ring->tx_bi, 0, bi_size);
@@ -242,6 +229,13 @@ void i40e_clean_tx_ring(struct i40e_ring *tx_ring)
242 229
243 tx_ring->next_to_use = 0; 230 tx_ring->next_to_use = 0;
244 tx_ring->next_to_clean = 0; 231 tx_ring->next_to_clean = 0;
232
233 if (!tx_ring->netdev)
234 return;
235
236 /* cleanup Tx queue statistics */
237 netdev_tx_reset_queue(netdev_get_tx_queue(tx_ring->netdev,
238 tx_ring->queue_index));
245} 239}
246 240
247/** 241/**
@@ -300,14 +294,14 @@ static bool i40e_check_tx_hang(struct i40e_ring *tx_ring)
300 * run the check_tx_hang logic with a transmit completion 294 * run the check_tx_hang logic with a transmit completion
301 * pending but without time to complete it yet. 295 * pending but without time to complete it yet.
302 */ 296 */
303 if ((tx_ring->tx_stats.tx_done_old == tx_ring->tx_stats.packets) && 297 if ((tx_ring->tx_stats.tx_done_old == tx_ring->stats.packets) &&
304 tx_pending) { 298 tx_pending) {
305 /* make sure it is true for two checks in a row */ 299 /* make sure it is true for two checks in a row */
306 ret = test_and_set_bit(__I40E_HANG_CHECK_ARMED, 300 ret = test_and_set_bit(__I40E_HANG_CHECK_ARMED,
307 &tx_ring->state); 301 &tx_ring->state);
308 } else { 302 } else {
309 /* update completed stats and disarm the hang check */ 303 /* update completed stats and disarm the hang check */
310 tx_ring->tx_stats.tx_done_old = tx_ring->tx_stats.packets; 304 tx_ring->tx_stats.tx_done_old = tx_ring->stats.packets;
311 clear_bit(__I40E_HANG_CHECK_ARMED, &tx_ring->state); 305 clear_bit(__I40E_HANG_CHECK_ARMED, &tx_ring->state);
312 } 306 }
313 307
@@ -331,62 +325,88 @@ static bool i40e_clean_tx_irq(struct i40e_ring *tx_ring, int budget)
331 325
332 tx_buf = &tx_ring->tx_bi[i]; 326 tx_buf = &tx_ring->tx_bi[i];
333 tx_desc = I40E_TX_DESC(tx_ring, i); 327 tx_desc = I40E_TX_DESC(tx_ring, i);
328 i -= tx_ring->count;
334 329
335 for (; budget; budget--) { 330 do {
336 struct i40e_tx_desc *eop_desc; 331 struct i40e_tx_desc *eop_desc = tx_buf->next_to_watch;
337
338 eop_desc = tx_buf->next_to_watch;
339 332
340 /* if next_to_watch is not set then there is no work pending */ 333 /* if next_to_watch is not set then there is no work pending */
341 if (!eop_desc) 334 if (!eop_desc)
342 break; 335 break;
343 336
337 /* prevent any other reads prior to eop_desc */
338 read_barrier_depends();
339
344 /* if the descriptor isn't done, no work yet to do */ 340 /* if the descriptor isn't done, no work yet to do */
345 if (!(eop_desc->cmd_type_offset_bsz & 341 if (!(eop_desc->cmd_type_offset_bsz &
346 cpu_to_le64(I40E_TX_DESC_DTYPE_DESC_DONE))) 342 cpu_to_le64(I40E_TX_DESC_DTYPE_DESC_DONE)))
347 break; 343 break;
348 344
349 /* count the packet as being completed */ 345 /* clear next_to_watch to prevent false hangs */
350 tx_ring->tx_stats.completed++;
351 tx_buf->next_to_watch = NULL; 346 tx_buf->next_to_watch = NULL;
352 tx_buf->time_stamp = 0;
353
354 /* set memory barrier before eop_desc is verified */
355 rmb();
356 347
357 do { 348 /* update the statistics for this packet */
358 i40e_unmap_tx_resource(tx_ring, tx_buf); 349 total_bytes += tx_buf->bytecount;
350 total_packets += tx_buf->gso_segs;
359 351
360 /* clear dtype status */ 352 /* free the skb */
361 tx_desc->cmd_type_offset_bsz &= 353 dev_kfree_skb_any(tx_buf->skb);
362 ~cpu_to_le64(I40E_TXD_QW1_DTYPE_MASK);
363 354
364 if (likely(tx_desc == eop_desc)) { 355 /* unmap skb header data */
365 eop_desc = NULL; 356 dma_unmap_single(tx_ring->dev,
357 dma_unmap_addr(tx_buf, dma),
358 dma_unmap_len(tx_buf, len),
359 DMA_TO_DEVICE);
366 360
367 dev_kfree_skb_any(tx_buf->skb); 361 /* clear tx_buffer data */
368 tx_buf->skb = NULL; 362 tx_buf->skb = NULL;
363 dma_unmap_len_set(tx_buf, len, 0);
369 364
370 total_bytes += tx_buf->bytecount; 365 /* unmap remaining buffers */
371 total_packets += tx_buf->gso_segs; 366 while (tx_desc != eop_desc) {
372 }
373 367
374 tx_buf++; 368 tx_buf++;
375 tx_desc++; 369 tx_desc++;
376 i++; 370 i++;
377 if (unlikely(i == tx_ring->count)) { 371 if (unlikely(!i)) {
378 i = 0; 372 i -= tx_ring->count;
379 tx_buf = tx_ring->tx_bi; 373 tx_buf = tx_ring->tx_bi;
380 tx_desc = I40E_TX_DESC(tx_ring, 0); 374 tx_desc = I40E_TX_DESC(tx_ring, 0);
381 } 375 }
382 } while (eop_desc);
383 }
384 376
377 /* unmap any remaining paged data */
378 if (dma_unmap_len(tx_buf, len)) {
379 dma_unmap_page(tx_ring->dev,
380 dma_unmap_addr(tx_buf, dma),
381 dma_unmap_len(tx_buf, len),
382 DMA_TO_DEVICE);
383 dma_unmap_len_set(tx_buf, len, 0);
384 }
385 }
386
387 /* move us one more past the eop_desc for start of next pkt */
388 tx_buf++;
389 tx_desc++;
390 i++;
391 if (unlikely(!i)) {
392 i -= tx_ring->count;
393 tx_buf = tx_ring->tx_bi;
394 tx_desc = I40E_TX_DESC(tx_ring, 0);
395 }
396
397 /* update budget accounting */
398 budget--;
399 } while (likely(budget));
400
401 i += tx_ring->count;
385 tx_ring->next_to_clean = i; 402 tx_ring->next_to_clean = i;
386 tx_ring->tx_stats.bytes += total_bytes; 403 u64_stats_update_begin(&tx_ring->syncp);
387 tx_ring->tx_stats.packets += total_packets; 404 tx_ring->stats.bytes += total_bytes;
405 tx_ring->stats.packets += total_packets;
406 u64_stats_update_end(&tx_ring->syncp);
388 tx_ring->q_vector->tx.total_bytes += total_bytes; 407 tx_ring->q_vector->tx.total_bytes += total_bytes;
389 tx_ring->q_vector->tx.total_packets += total_packets; 408 tx_ring->q_vector->tx.total_packets += total_packets;
409
390 if (check_for_tx_hang(tx_ring) && i40e_check_tx_hang(tx_ring)) { 410 if (check_for_tx_hang(tx_ring) && i40e_check_tx_hang(tx_ring)) {
391 /* schedule immediate reset if we believe we hung */ 411 /* schedule immediate reset if we believe we hung */
392 dev_info(tx_ring->dev, "Detected Tx Unit Hang\n" 412 dev_info(tx_ring->dev, "Detected Tx Unit Hang\n"
@@ -414,6 +434,10 @@ static bool i40e_clean_tx_irq(struct i40e_ring *tx_ring, int budget)
414 return true; 434 return true;
415 } 435 }
416 436
437 netdev_tx_completed_queue(netdev_get_tx_queue(tx_ring->netdev,
438 tx_ring->queue_index),
439 total_packets, total_bytes);
440
417#define TX_WAKE_THRESHOLD (DESC_NEEDED * 2) 441#define TX_WAKE_THRESHOLD (DESC_NEEDED * 2)
418 if (unlikely(total_packets && netif_carrier_ok(tx_ring->netdev) && 442 if (unlikely(total_packets && netif_carrier_ok(tx_ring->netdev) &&
419 (I40E_DESC_UNUSED(tx_ring) >= TX_WAKE_THRESHOLD))) { 443 (I40E_DESC_UNUSED(tx_ring) >= TX_WAKE_THRESHOLD))) {
@@ -524,8 +548,6 @@ static void i40e_update_dynamic_itr(struct i40e_q_vector *q_vector)
524 i40e_set_new_dynamic_itr(&q_vector->tx); 548 i40e_set_new_dynamic_itr(&q_vector->tx);
525 if (old_itr != q_vector->tx.itr) 549 if (old_itr != q_vector->tx.itr)
526 wr32(hw, reg_addr, q_vector->tx.itr); 550 wr32(hw, reg_addr, q_vector->tx.itr);
527
528 i40e_flush(hw);
529} 551}
530 552
531/** 553/**
@@ -1042,8 +1064,10 @@ next_desc:
1042 } 1064 }
1043 1065
1044 rx_ring->next_to_clean = i; 1066 rx_ring->next_to_clean = i;
1045 rx_ring->rx_stats.packets += total_rx_packets; 1067 u64_stats_update_begin(&rx_ring->syncp);
1046 rx_ring->rx_stats.bytes += total_rx_bytes; 1068 rx_ring->stats.packets += total_rx_packets;
1069 rx_ring->stats.bytes += total_rx_bytes;
1070 u64_stats_update_end(&rx_ring->syncp);
1047 rx_ring->q_vector->rx.total_packets += total_rx_packets; 1071 rx_ring->q_vector->rx.total_packets += total_rx_packets;
1048 rx_ring->q_vector->rx.total_bytes += total_rx_bytes; 1072 rx_ring->q_vector->rx.total_bytes += total_rx_bytes;
1049 1073
@@ -1067,27 +1091,28 @@ int i40e_napi_poll(struct napi_struct *napi, int budget)
1067 struct i40e_q_vector *q_vector = 1091 struct i40e_q_vector *q_vector =
1068 container_of(napi, struct i40e_q_vector, napi); 1092 container_of(napi, struct i40e_q_vector, napi);
1069 struct i40e_vsi *vsi = q_vector->vsi; 1093 struct i40e_vsi *vsi = q_vector->vsi;
1094 struct i40e_ring *ring;
1070 bool clean_complete = true; 1095 bool clean_complete = true;
1071 int budget_per_ring; 1096 int budget_per_ring;
1072 int i;
1073 1097
1074 if (test_bit(__I40E_DOWN, &vsi->state)) { 1098 if (test_bit(__I40E_DOWN, &vsi->state)) {
1075 napi_complete(napi); 1099 napi_complete(napi);
1076 return 0; 1100 return 0;
1077 } 1101 }
1078 1102
1103 /* Since the actual Tx work is minimal, we can give the Tx a larger
1104 * budget and be more aggressive about cleaning up the Tx descriptors.
1105 */
1106 i40e_for_each_ring(ring, q_vector->tx)
1107 clean_complete &= i40e_clean_tx_irq(ring, vsi->work_limit);
1108
1079 /* We attempt to distribute budget to each Rx queue fairly, but don't 1109 /* We attempt to distribute budget to each Rx queue fairly, but don't
1080 * allow the budget to go below 1 because that would exit polling early. 1110 * allow the budget to go below 1 because that would exit polling early.
1081 * Since the actual Tx work is minimal, we can give the Tx a larger
1082 * budget and be more aggressive about cleaning up the Tx descriptors.
1083 */ 1111 */
1084 budget_per_ring = max(budget/q_vector->num_ringpairs, 1); 1112 budget_per_ring = max(budget/q_vector->num_ringpairs, 1);
1085 for (i = 0; i < q_vector->num_ringpairs; i++) { 1113
1086 clean_complete &= i40e_clean_tx_irq(q_vector->tx.ring[i], 1114 i40e_for_each_ring(ring, q_vector->rx)
1087 vsi->work_limit); 1115 clean_complete &= i40e_clean_rx_irq(ring, budget_per_ring);
1088 clean_complete &= i40e_clean_rx_irq(q_vector->rx.ring[i],
1089 budget_per_ring);
1090 }
1091 1116
1092 /* If work not completed, return budget and polling will return */ 1117 /* If work not completed, return budget and polling will return */
1093 if (!clean_complete) 1118 if (!clean_complete)
@@ -1117,7 +1142,8 @@ int i40e_napi_poll(struct napi_struct *napi, int budget)
1117 qval = rd32(hw, I40E_QINT_TQCTL(0)); 1142 qval = rd32(hw, I40E_QINT_TQCTL(0));
1118 qval |= I40E_QINT_TQCTL_CAUSE_ENA_MASK; 1143 qval |= I40E_QINT_TQCTL_CAUSE_ENA_MASK;
1119 wr32(hw, I40E_QINT_TQCTL(0), qval); 1144 wr32(hw, I40E_QINT_TQCTL(0), qval);
1120 i40e_flush(hw); 1145
1146 i40e_irq_dynamic_enable_icr0(vsi->back);
1121 } 1147 }
1122 } 1148 }
1123 1149
@@ -1144,6 +1170,7 @@ static void i40e_atr(struct i40e_ring *tx_ring, struct sk_buff *skb,
1144 struct tcphdr *th; 1170 struct tcphdr *th;
1145 unsigned int hlen; 1171 unsigned int hlen;
1146 u32 flex_ptype, dtype_cmd; 1172 u32 flex_ptype, dtype_cmd;
1173 u16 i;
1147 1174
1148 /* make sure ATR is enabled */ 1175 /* make sure ATR is enabled */
1149 if (!(pf->flags & I40E_FLAG_FDIR_ATR_ENABLED)) 1176 if (!(pf->flags & I40E_FLAG_FDIR_ATR_ENABLED))
@@ -1183,10 +1210,11 @@ static void i40e_atr(struct i40e_ring *tx_ring, struct sk_buff *skb,
1183 tx_ring->atr_count = 0; 1210 tx_ring->atr_count = 0;
1184 1211
1185 /* grab the next descriptor */ 1212 /* grab the next descriptor */
1186 fdir_desc = I40E_TX_FDIRDESC(tx_ring, tx_ring->next_to_use); 1213 i = tx_ring->next_to_use;
1187 tx_ring->next_to_use++; 1214 fdir_desc = I40E_TX_FDIRDESC(tx_ring, i);
1188 if (tx_ring->next_to_use == tx_ring->count) 1215
1189 tx_ring->next_to_use = 0; 1216 i++;
1217 tx_ring->next_to_use = (i < tx_ring->count) ? i : 0;
1190 1218
1191 flex_ptype = (tx_ring->queue_index << I40E_TXD_FLTR_QW0_QINDEX_SHIFT) & 1219 flex_ptype = (tx_ring->queue_index << I40E_TXD_FLTR_QW0_QINDEX_SHIFT) &
1192 I40E_TXD_FLTR_QW0_QINDEX_MASK; 1220 I40E_TXD_FLTR_QW0_QINDEX_MASK;
@@ -1216,7 +1244,6 @@ static void i40e_atr(struct i40e_ring *tx_ring, struct sk_buff *skb,
1216 fdir_desc->dtype_cmd_cntindex = cpu_to_le32(dtype_cmd); 1244 fdir_desc->dtype_cmd_cntindex = cpu_to_le32(dtype_cmd);
1217} 1245}
1218 1246
1219#define I40E_TXD_CMD (I40E_TX_DESC_CMD_EOP | I40E_TX_DESC_CMD_RS)
1220/** 1247/**
1221 * i40e_tx_prepare_vlan_flags - prepare generic TX VLAN tagging flags for HW 1248 * i40e_tx_prepare_vlan_flags - prepare generic TX VLAN tagging flags for HW
1222 * @skb: send buffer 1249 * @skb: send buffer
@@ -1276,27 +1303,6 @@ static int i40e_tx_prepare_vlan_flags(struct sk_buff *skb,
1276} 1303}
1277 1304
1278/** 1305/**
1279 * i40e_tx_csum - is checksum offload requested
1280 * @tx_ring: ptr to the ring to send
1281 * @skb: ptr to the skb we're sending
1282 * @tx_flags: the collected send information
1283 * @protocol: the send protocol
1284 *
1285 * Returns true if checksum offload is requested
1286 **/
1287static bool i40e_tx_csum(struct i40e_ring *tx_ring, struct sk_buff *skb,
1288 u32 tx_flags, __be16 protocol)
1289{
1290 if ((skb->ip_summed != CHECKSUM_PARTIAL) &&
1291 !(tx_flags & I40E_TX_FLAGS_TXSW)) {
1292 if (!(tx_flags & I40E_TX_FLAGS_HW_VLAN))
1293 return false;
1294 }
1295
1296 return skb->ip_summed == CHECKSUM_PARTIAL;
1297}
1298
1299/**
1300 * i40e_tso - set up the tso context descriptor 1306 * i40e_tso - set up the tso context descriptor
1301 * @tx_ring: ptr to the ring to send 1307 * @tx_ring: ptr to the ring to send
1302 * @skb: ptr to the skb we're sending 1308 * @skb: ptr to the skb we're sending
@@ -1482,15 +1488,16 @@ static void i40e_create_tx_ctx(struct i40e_ring *tx_ring,
1482 const u32 cd_tunneling, const u32 cd_l2tag2) 1488 const u32 cd_tunneling, const u32 cd_l2tag2)
1483{ 1489{
1484 struct i40e_tx_context_desc *context_desc; 1490 struct i40e_tx_context_desc *context_desc;
1491 int i = tx_ring->next_to_use;
1485 1492
1486 if (!cd_type_cmd_tso_mss && !cd_tunneling && !cd_l2tag2) 1493 if (!cd_type_cmd_tso_mss && !cd_tunneling && !cd_l2tag2)
1487 return; 1494 return;
1488 1495
1489 /* grab the next descriptor */ 1496 /* grab the next descriptor */
1490 context_desc = I40E_TX_CTXTDESC(tx_ring, tx_ring->next_to_use); 1497 context_desc = I40E_TX_CTXTDESC(tx_ring, i);
1491 tx_ring->next_to_use++; 1498
1492 if (tx_ring->next_to_use == tx_ring->count) 1499 i++;
1493 tx_ring->next_to_use = 0; 1500 tx_ring->next_to_use = (i < tx_ring->count) ? i : 0;
1494 1501
1495 /* cpu_to_le32 and assign to struct fields */ 1502 /* cpu_to_le32 and assign to struct fields */
1496 context_desc->tunneling_params = cpu_to_le32(cd_tunneling); 1503 context_desc->tunneling_params = cpu_to_le32(cd_tunneling);
@@ -1512,68 +1519,71 @@ static void i40e_tx_map(struct i40e_ring *tx_ring, struct sk_buff *skb,
1512 struct i40e_tx_buffer *first, u32 tx_flags, 1519 struct i40e_tx_buffer *first, u32 tx_flags,
1513 const u8 hdr_len, u32 td_cmd, u32 td_offset) 1520 const u8 hdr_len, u32 td_cmd, u32 td_offset)
1514{ 1521{
1515 struct skb_frag_struct *frag = &skb_shinfo(skb)->frags[0];
1516 unsigned int data_len = skb->data_len; 1522 unsigned int data_len = skb->data_len;
1517 unsigned int size = skb_headlen(skb); 1523 unsigned int size = skb_headlen(skb);
1518 struct device *dev = tx_ring->dev; 1524 struct skb_frag_struct *frag;
1519 u32 paylen = skb->len - hdr_len;
1520 u16 i = tx_ring->next_to_use;
1521 struct i40e_tx_buffer *tx_bi; 1525 struct i40e_tx_buffer *tx_bi;
1522 struct i40e_tx_desc *tx_desc; 1526 struct i40e_tx_desc *tx_desc;
1523 u32 buf_offset = 0; 1527 u16 i = tx_ring->next_to_use;
1524 u32 td_tag = 0; 1528 u32 td_tag = 0;
1525 dma_addr_t dma; 1529 dma_addr_t dma;
1526 u16 gso_segs; 1530 u16 gso_segs;
1527 1531
1528 dma = dma_map_single(dev, skb->data, size, DMA_TO_DEVICE);
1529 if (dma_mapping_error(dev, dma))
1530 goto dma_error;
1531
1532 if (tx_flags & I40E_TX_FLAGS_HW_VLAN) { 1532 if (tx_flags & I40E_TX_FLAGS_HW_VLAN) {
1533 td_cmd |= I40E_TX_DESC_CMD_IL2TAG1; 1533 td_cmd |= I40E_TX_DESC_CMD_IL2TAG1;
1534 td_tag = (tx_flags & I40E_TX_FLAGS_VLAN_MASK) >> 1534 td_tag = (tx_flags & I40E_TX_FLAGS_VLAN_MASK) >>
1535 I40E_TX_FLAGS_VLAN_SHIFT; 1535 I40E_TX_FLAGS_VLAN_SHIFT;
1536 } 1536 }
1537 1537
1538 if (tx_flags & (I40E_TX_FLAGS_TSO | I40E_TX_FLAGS_FSO))
1539 gso_segs = skb_shinfo(skb)->gso_segs;
1540 else
1541 gso_segs = 1;
1542
1543 /* multiply data chunks by size of headers */
1544 first->bytecount = skb->len - hdr_len + (gso_segs * hdr_len);
1545 first->gso_segs = gso_segs;
1546 first->skb = skb;
1547 first->tx_flags = tx_flags;
1548
1549 dma = dma_map_single(tx_ring->dev, skb->data, size, DMA_TO_DEVICE);
1550
1538 tx_desc = I40E_TX_DESC(tx_ring, i); 1551 tx_desc = I40E_TX_DESC(tx_ring, i);
1539 for (;;) { 1552 tx_bi = first;
1540 while (size > I40E_MAX_DATA_PER_TXD) { 1553
1541 tx_desc->buffer_addr = cpu_to_le64(dma + buf_offset); 1554 for (frag = &skb_shinfo(skb)->frags[0];; frag++) {
1555 if (dma_mapping_error(tx_ring->dev, dma))
1556 goto dma_error;
1557
1558 /* record length, and DMA address */
1559 dma_unmap_len_set(tx_bi, len, size);
1560 dma_unmap_addr_set(tx_bi, dma, dma);
1561
1562 tx_desc->buffer_addr = cpu_to_le64(dma);
1563
1564 while (unlikely(size > I40E_MAX_DATA_PER_TXD)) {
1542 tx_desc->cmd_type_offset_bsz = 1565 tx_desc->cmd_type_offset_bsz =
1543 build_ctob(td_cmd, td_offset, 1566 build_ctob(td_cmd, td_offset,
1544 I40E_MAX_DATA_PER_TXD, td_tag); 1567 I40E_MAX_DATA_PER_TXD, td_tag);
1545 1568
1546 buf_offset += I40E_MAX_DATA_PER_TXD;
1547 size -= I40E_MAX_DATA_PER_TXD;
1548
1549 tx_desc++; 1569 tx_desc++;
1550 i++; 1570 i++;
1551 if (i == tx_ring->count) { 1571 if (i == tx_ring->count) {
1552 tx_desc = I40E_TX_DESC(tx_ring, 0); 1572 tx_desc = I40E_TX_DESC(tx_ring, 0);
1553 i = 0; 1573 i = 0;
1554 } 1574 }
1555 }
1556 1575
1557 tx_bi = &tx_ring->tx_bi[i]; 1576 dma += I40E_MAX_DATA_PER_TXD;
1558 tx_bi->length = buf_offset + size; 1577 size -= I40E_MAX_DATA_PER_TXD;
1559 tx_bi->tx_flags = tx_flags;
1560 tx_bi->dma = dma;
1561 1578
1562 tx_desc->buffer_addr = cpu_to_le64(dma + buf_offset); 1579 tx_desc->buffer_addr = cpu_to_le64(dma);
1563 tx_desc->cmd_type_offset_bsz = build_ctob(td_cmd, td_offset, 1580 }
1564 size, td_tag);
1565 1581
1566 if (likely(!data_len)) 1582 if (likely(!data_len))
1567 break; 1583 break;
1568 1584
1569 size = skb_frag_size(frag); 1585 tx_desc->cmd_type_offset_bsz = build_ctob(td_cmd, td_offset,
1570 data_len -= size; 1586 size, td_tag);
1571 buf_offset = 0;
1572 tx_flags |= I40E_TX_FLAGS_MAPPED_AS_PAGE;
1573
1574 dma = skb_frag_dma_map(dev, frag, 0, size, DMA_TO_DEVICE);
1575 if (dma_mapping_error(dev, dma))
1576 goto dma_error;
1577 1587
1578 tx_desc++; 1588 tx_desc++;
1579 i++; 1589 i++;
@@ -1582,31 +1592,25 @@ static void i40e_tx_map(struct i40e_ring *tx_ring, struct sk_buff *skb,
1582 i = 0; 1592 i = 0;
1583 } 1593 }
1584 1594
1585 frag++; 1595 size = skb_frag_size(frag);
1586 } 1596 data_len -= size;
1587
1588 tx_desc->cmd_type_offset_bsz |=
1589 cpu_to_le64((u64)I40E_TXD_CMD << I40E_TXD_QW1_CMD_SHIFT);
1590 1597
1591 i++; 1598 dma = skb_frag_dma_map(tx_ring->dev, frag, 0, size,
1592 if (i == tx_ring->count) 1599 DMA_TO_DEVICE);
1593 i = 0;
1594 1600
1595 tx_ring->next_to_use = i; 1601 tx_bi = &tx_ring->tx_bi[i];
1602 }
1596 1603
1597 if (tx_flags & (I40E_TX_FLAGS_TSO | I40E_TX_FLAGS_FSO)) 1604 tx_desc->cmd_type_offset_bsz =
1598 gso_segs = skb_shinfo(skb)->gso_segs; 1605 build_ctob(td_cmd, td_offset, size, td_tag) |
1599 else 1606 cpu_to_le64((u64)I40E_TXD_CMD << I40E_TXD_QW1_CMD_SHIFT);
1600 gso_segs = 1;
1601 1607
1602 /* multiply data chunks by size of headers */ 1608 netdev_tx_sent_queue(netdev_get_tx_queue(tx_ring->netdev,
1603 tx_bi->bytecount = paylen + (gso_segs * hdr_len); 1609 tx_ring->queue_index),
1604 tx_bi->gso_segs = gso_segs; 1610 first->bytecount);
1605 tx_bi->skb = skb;
1606 1611
1607 /* set the timestamp and next to watch values */ 1612 /* set the timestamp */
1608 first->time_stamp = jiffies; 1613 first->time_stamp = jiffies;
1609 first->next_to_watch = tx_desc;
1610 1614
1611 /* Force memory writes to complete before letting h/w 1615 /* Force memory writes to complete before letting h/w
1612 * know there are new descriptors to fetch. (Only 1616 * know there are new descriptors to fetch. (Only
@@ -1615,16 +1619,27 @@ static void i40e_tx_map(struct i40e_ring *tx_ring, struct sk_buff *skb,
1615 */ 1619 */
1616 wmb(); 1620 wmb();
1617 1621
1622 /* set next_to_watch value indicating a packet is present */
1623 first->next_to_watch = tx_desc;
1624
1625 i++;
1626 if (i == tx_ring->count)
1627 i = 0;
1628
1629 tx_ring->next_to_use = i;
1630
1631 /* notify HW of packet */
1618 writel(i, tx_ring->tail); 1632 writel(i, tx_ring->tail);
1633
1619 return; 1634 return;
1620 1635
1621dma_error: 1636dma_error:
1622 dev_info(dev, "TX DMA map failed\n"); 1637 dev_info(tx_ring->dev, "TX DMA map failed\n");
1623 1638
1624 /* clear dma mappings for failed tx_bi map */ 1639 /* clear dma mappings for failed tx_bi map */
1625 for (;;) { 1640 for (;;) {
1626 tx_bi = &tx_ring->tx_bi[i]; 1641 tx_bi = &tx_ring->tx_bi[i];
1627 i40e_unmap_tx_resource(tx_ring, tx_bi); 1642 i40e_unmap_and_free_tx_resource(tx_ring, tx_bi);
1628 if (tx_bi == first) 1643 if (tx_bi == first)
1629 break; 1644 break;
1630 if (i == 0) 1645 if (i == 0)
@@ -1632,8 +1647,6 @@ dma_error:
1632 i--; 1647 i--;
1633 } 1648 }
1634 1649
1635 dev_kfree_skb_any(skb);
1636
1637 tx_ring->next_to_use = i; 1650 tx_ring->next_to_use = i;
1638} 1651}
1639 1652
@@ -1758,16 +1771,16 @@ static netdev_tx_t i40e_xmit_frame_ring(struct sk_buff *skb,
1758 1771
1759 skb_tx_timestamp(skb); 1772 skb_tx_timestamp(skb);
1760 1773
1774 /* always enable CRC insertion offload */
1775 td_cmd |= I40E_TX_DESC_CMD_ICRC;
1776
1761 /* Always offload the checksum, since it's in the data descriptor */ 1777 /* Always offload the checksum, since it's in the data descriptor */
1762 if (i40e_tx_csum(tx_ring, skb, tx_flags, protocol)) 1778 if (skb->ip_summed == CHECKSUM_PARTIAL) {
1763 tx_flags |= I40E_TX_FLAGS_CSUM; 1779 tx_flags |= I40E_TX_FLAGS_CSUM;
1764 1780
1765 /* always enable offload insertion */
1766 td_cmd |= I40E_TX_DESC_CMD_ICRC;
1767
1768 if (tx_flags & I40E_TX_FLAGS_CSUM)
1769 i40e_tx_enable_csum(skb, tx_flags, &td_cmd, &td_offset, 1781 i40e_tx_enable_csum(skb, tx_flags, &td_cmd, &td_offset,
1770 tx_ring, &cd_tunneling); 1782 tx_ring, &cd_tunneling);
1783 }
1771 1784
1772 i40e_create_tx_ctx(tx_ring, cd_type_cmd_tso_mss, 1785 i40e_create_tx_ctx(tx_ring, cd_type_cmd_tso_mss,
1773 cd_tunneling, cd_l2tag2); 1786 cd_tunneling, cd_l2tag2);
@@ -1801,7 +1814,7 @@ netdev_tx_t i40e_lan_xmit_frame(struct sk_buff *skb, struct net_device *netdev)
1801{ 1814{
1802 struct i40e_netdev_priv *np = netdev_priv(netdev); 1815 struct i40e_netdev_priv *np = netdev_priv(netdev);
1803 struct i40e_vsi *vsi = np->vsi; 1816 struct i40e_vsi *vsi = np->vsi;
1804 struct i40e_ring *tx_ring = &vsi->tx_rings[skb->queue_mapping]; 1817 struct i40e_ring *tx_ring = vsi->tx_rings[skb->queue_mapping];
1805 1818
1806 /* hardware can't handle really short frames, hardware padding works 1819 /* hardware can't handle really short frames, hardware padding works
1807 * beyond this point 1820 * beyond this point
diff --git a/drivers/net/ethernet/intel/i40e/i40e_txrx.h b/drivers/net/ethernet/intel/i40e/i40e_txrx.h
index b1d7722d98a7..db55d9947f15 100644
--- a/drivers/net/ethernet/intel/i40e/i40e_txrx.h
+++ b/drivers/net/ethernet/intel/i40e/i40e_txrx.h
@@ -102,23 +102,20 @@
102#define I40E_TX_FLAGS_IPV6 (u32)(1 << 5) 102#define I40E_TX_FLAGS_IPV6 (u32)(1 << 5)
103#define I40E_TX_FLAGS_FCCRC (u32)(1 << 6) 103#define I40E_TX_FLAGS_FCCRC (u32)(1 << 6)
104#define I40E_TX_FLAGS_FSO (u32)(1 << 7) 104#define I40E_TX_FLAGS_FSO (u32)(1 << 7)
105#define I40E_TX_FLAGS_TXSW (u32)(1 << 8)
106#define I40E_TX_FLAGS_MAPPED_AS_PAGE (u32)(1 << 9)
107#define I40E_TX_FLAGS_VLAN_MASK 0xffff0000 105#define I40E_TX_FLAGS_VLAN_MASK 0xffff0000
108#define I40E_TX_FLAGS_VLAN_PRIO_MASK 0xe0000000 106#define I40E_TX_FLAGS_VLAN_PRIO_MASK 0xe0000000
109#define I40E_TX_FLAGS_VLAN_PRIO_SHIFT 29 107#define I40E_TX_FLAGS_VLAN_PRIO_SHIFT 29
110#define I40E_TX_FLAGS_VLAN_SHIFT 16 108#define I40E_TX_FLAGS_VLAN_SHIFT 16
111 109
112struct i40e_tx_buffer { 110struct i40e_tx_buffer {
113 struct sk_buff *skb;
114 dma_addr_t dma;
115 unsigned long time_stamp;
116 u16 length;
117 u32 tx_flags;
118 struct i40e_tx_desc *next_to_watch; 111 struct i40e_tx_desc *next_to_watch;
112 unsigned long time_stamp;
113 struct sk_buff *skb;
119 unsigned int bytecount; 114 unsigned int bytecount;
120 u16 gso_segs; 115 unsigned short gso_segs;
121 u8 mapped_as_page; 116 DEFINE_DMA_UNMAP_ADDR(dma);
117 DEFINE_DMA_UNMAP_LEN(len);
118 u32 tx_flags;
122}; 119};
123 120
124struct i40e_rx_buffer { 121struct i40e_rx_buffer {
@@ -129,18 +126,18 @@ struct i40e_rx_buffer {
129 unsigned int page_offset; 126 unsigned int page_offset;
130}; 127};
131 128
132struct i40e_tx_queue_stats { 129struct i40e_queue_stats {
133 u64 packets; 130 u64 packets;
134 u64 bytes; 131 u64 bytes;
132};
133
134struct i40e_tx_queue_stats {
135 u64 restart_queue; 135 u64 restart_queue;
136 u64 tx_busy; 136 u64 tx_busy;
137 u64 completed;
138 u64 tx_done_old; 137 u64 tx_done_old;
139}; 138};
140 139
141struct i40e_rx_queue_stats { 140struct i40e_rx_queue_stats {
142 u64 packets;
143 u64 bytes;
144 u64 non_eop_descs; 141 u64 non_eop_descs;
145 u64 alloc_rx_page_failed; 142 u64 alloc_rx_page_failed;
146 u64 alloc_rx_buff_failed; 143 u64 alloc_rx_buff_failed;
@@ -183,6 +180,7 @@ enum i40e_ring_state_t {
183 180
184/* struct that defines a descriptor ring, associated with a VSI */ 181/* struct that defines a descriptor ring, associated with a VSI */
185struct i40e_ring { 182struct i40e_ring {
183 struct i40e_ring *next; /* pointer to next ring in q_vector */
186 void *desc; /* Descriptor ring memory */ 184 void *desc; /* Descriptor ring memory */
187 struct device *dev; /* Used for DMA mapping */ 185 struct device *dev; /* Used for DMA mapping */
188 struct net_device *netdev; /* netdev ring maps to */ 186 struct net_device *netdev; /* netdev ring maps to */
@@ -219,6 +217,8 @@ struct i40e_ring {
219 bool ring_active; /* is ring online or not */ 217 bool ring_active; /* is ring online or not */
220 218
221 /* stats structs */ 219 /* stats structs */
220 struct i40e_queue_stats stats;
221 struct u64_stats_sync syncp;
222 union { 222 union {
223 struct i40e_tx_queue_stats tx_stats; 223 struct i40e_tx_queue_stats tx_stats;
224 struct i40e_rx_queue_stats rx_stats; 224 struct i40e_rx_queue_stats rx_stats;
@@ -229,6 +229,8 @@ struct i40e_ring {
229 229
230 struct i40e_vsi *vsi; /* Backreference to associated VSI */ 230 struct i40e_vsi *vsi; /* Backreference to associated VSI */
231 struct i40e_q_vector *q_vector; /* Backreference to associated vector */ 231 struct i40e_q_vector *q_vector; /* Backreference to associated vector */
232
233 struct rcu_head rcu; /* to avoid race on free */
232} ____cacheline_internodealigned_in_smp; 234} ____cacheline_internodealigned_in_smp;
233 235
234enum i40e_latency_range { 236enum i40e_latency_range {
@@ -238,9 +240,8 @@ enum i40e_latency_range {
238}; 240};
239 241
240struct i40e_ring_container { 242struct i40e_ring_container {
241#define I40E_MAX_RINGPAIR_PER_VECTOR 8
242 /* array of pointers to rings */ 243 /* array of pointers to rings */
243 struct i40e_ring *ring[I40E_MAX_RINGPAIR_PER_VECTOR]; 244 struct i40e_ring *ring;
244 unsigned int total_bytes; /* total bytes processed this int */ 245 unsigned int total_bytes; /* total bytes processed this int */
245 unsigned int total_packets; /* total packets processed this int */ 246 unsigned int total_packets; /* total packets processed this int */
246 u16 count; 247 u16 count;
@@ -248,6 +249,10 @@ struct i40e_ring_container {
248 u16 itr; 249 u16 itr;
249}; 250};
250 251
252/* iterator for handling rings in ring container */
253#define i40e_for_each_ring(pos, head) \
254 for (pos = (head).ring; pos != NULL; pos = pos->next)
255
251void i40e_alloc_rx_buffers(struct i40e_ring *rxr, u16 cleaned_count); 256void i40e_alloc_rx_buffers(struct i40e_ring *rxr, u16 cleaned_count);
252netdev_tx_t i40e_lan_xmit_frame(struct sk_buff *skb, struct net_device *netdev); 257netdev_tx_t i40e_lan_xmit_frame(struct sk_buff *skb, struct net_device *netdev);
253void i40e_clean_tx_ring(struct i40e_ring *tx_ring); 258void i40e_clean_tx_ring(struct i40e_ring *tx_ring);
diff --git a/drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c b/drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
index 8967e58e2408..07596982a477 100644
--- a/drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
+++ b/drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
@@ -251,7 +251,7 @@ static void i40e_config_irq_link_list(struct i40e_vf *vf, u16 vsi_idx,
251 reg_idx = I40E_VPINT_LNKLST0(vf->vf_id); 251 reg_idx = I40E_VPINT_LNKLST0(vf->vf_id);
252 else 252 else
253 reg_idx = I40E_VPINT_LNKLSTN( 253 reg_idx = I40E_VPINT_LNKLSTN(
254 ((pf->hw.func_caps.num_msix_vectors_vf - 1) 254 (pf->hw.func_caps.num_msix_vectors_vf
255 * vf->vf_id) + (vector_id - 1)); 255 * vf->vf_id) + (vector_id - 1));
256 256
257 if (vecmap->rxq_map == 0 && vecmap->txq_map == 0) { 257 if (vecmap->rxq_map == 0 && vecmap->txq_map == 0) {
@@ -383,7 +383,7 @@ static int i40e_config_vsi_tx_queue(struct i40e_vf *vf, u16 vsi_idx,
383 383
384 /* associate this queue with the PCI VF function */ 384 /* associate this queue with the PCI VF function */
385 qtx_ctl = I40E_QTX_CTL_VF_QUEUE; 385 qtx_ctl = I40E_QTX_CTL_VF_QUEUE;
386 qtx_ctl |= ((hw->hmc.hmc_fn_id << I40E_QTX_CTL_PF_INDX_SHIFT) 386 qtx_ctl |= ((hw->pf_id << I40E_QTX_CTL_PF_INDX_SHIFT)
387 & I40E_QTX_CTL_PF_INDX_MASK); 387 & I40E_QTX_CTL_PF_INDX_MASK);
388 qtx_ctl |= (((vf->vf_id + hw->func_caps.vf_base_id) 388 qtx_ctl |= (((vf->vf_id + hw->func_caps.vf_base_id)
389 << I40E_QTX_CTL_VFVM_INDX_SHIFT) 389 << I40E_QTX_CTL_VFVM_INDX_SHIFT)
diff --git a/drivers/net/ethernet/intel/igb/e1000_82575.h b/drivers/net/ethernet/intel/igb/e1000_82575.h
index 74a1506b4235..8c2437722aad 100644
--- a/drivers/net/ethernet/intel/igb/e1000_82575.h
+++ b/drivers/net/ethernet/intel/igb/e1000_82575.h
@@ -28,14 +28,14 @@
28#ifndef _E1000_82575_H_ 28#ifndef _E1000_82575_H_
29#define _E1000_82575_H_ 29#define _E1000_82575_H_
30 30
31extern void igb_shutdown_serdes_link_82575(struct e1000_hw *hw); 31void igb_shutdown_serdes_link_82575(struct e1000_hw *hw);
32extern void igb_power_up_serdes_link_82575(struct e1000_hw *hw); 32void igb_power_up_serdes_link_82575(struct e1000_hw *hw);
33extern void igb_power_down_phy_copper_82575(struct e1000_hw *hw); 33void igb_power_down_phy_copper_82575(struct e1000_hw *hw);
34extern void igb_rx_fifo_flush_82575(struct e1000_hw *hw); 34void igb_rx_fifo_flush_82575(struct e1000_hw *hw);
35extern s32 igb_read_i2c_byte(struct e1000_hw *hw, u8 byte_offset, 35s32 igb_read_i2c_byte(struct e1000_hw *hw, u8 byte_offset, u8 dev_addr,
36 u8 dev_addr, u8 *data); 36 u8 *data);
37extern s32 igb_write_i2c_byte(struct e1000_hw *hw, u8 byte_offset, 37s32 igb_write_i2c_byte(struct e1000_hw *hw, u8 byte_offset, u8 dev_addr,
38 u8 dev_addr, u8 data); 38 u8 data);
39 39
40#define ID_LED_DEFAULT_82575_SERDES ((ID_LED_DEF1_DEF2 << 12) | \ 40#define ID_LED_DEFAULT_82575_SERDES ((ID_LED_DEF1_DEF2 << 12) | \
41 (ID_LED_DEF1_DEF2 << 8) | \ 41 (ID_LED_DEF1_DEF2 << 8) | \
diff --git a/drivers/net/ethernet/intel/igb/e1000_hw.h b/drivers/net/ethernet/intel/igb/e1000_hw.h
index 37a9c06a6c68..2e166b22d52b 100644
--- a/drivers/net/ethernet/intel/igb/e1000_hw.h
+++ b/drivers/net/ethernet/intel/igb/e1000_hw.h
@@ -562,11 +562,11 @@ struct e1000_hw {
562 u8 revision_id; 562 u8 revision_id;
563}; 563};
564 564
565extern struct net_device *igb_get_hw_dev(struct e1000_hw *hw); 565struct net_device *igb_get_hw_dev(struct e1000_hw *hw);
566#define hw_dbg(format, arg...) \ 566#define hw_dbg(format, arg...) \
567 netdev_dbg(igb_get_hw_dev(hw), format, ##arg) 567 netdev_dbg(igb_get_hw_dev(hw), format, ##arg)
568 568
569/* These functions must be implemented by drivers */ 569/* These functions must be implemented by drivers */
570s32 igb_read_pcie_cap_reg(struct e1000_hw *hw, u32 reg, u16 *value); 570s32 igb_read_pcie_cap_reg(struct e1000_hw *hw, u32 reg, u16 *value);
571s32 igb_write_pcie_cap_reg(struct e1000_hw *hw, u32 reg, u16 *value); 571s32 igb_write_pcie_cap_reg(struct e1000_hw *hw, u32 reg, u16 *value);
572#endif /* _E1000_HW_H_ */ 572#endif /* _E1000_HW_H_ */
diff --git a/drivers/net/ethernet/intel/igb/e1000_i210.h b/drivers/net/ethernet/intel/igb/e1000_i210.h
index dde3c4b7ea99..2d913716573a 100644
--- a/drivers/net/ethernet/intel/igb/e1000_i210.h
+++ b/drivers/net/ethernet/intel/igb/e1000_i210.h
@@ -28,26 +28,24 @@
28#ifndef _E1000_I210_H_ 28#ifndef _E1000_I210_H_
29#define _E1000_I210_H_ 29#define _E1000_I210_H_
30 30
31extern s32 igb_update_flash_i210(struct e1000_hw *hw); 31s32 igb_update_flash_i210(struct e1000_hw *hw);
32extern s32 igb_update_nvm_checksum_i210(struct e1000_hw *hw); 32s32 igb_update_nvm_checksum_i210(struct e1000_hw *hw);
33extern s32 igb_validate_nvm_checksum_i210(struct e1000_hw *hw); 33s32 igb_validate_nvm_checksum_i210(struct e1000_hw *hw);
34extern s32 igb_write_nvm_srwr_i210(struct e1000_hw *hw, u16 offset, 34s32 igb_write_nvm_srwr_i210(struct e1000_hw *hw, u16 offset, u16 words,
35 u16 words, u16 *data); 35 u16 *data);
36extern s32 igb_read_nvm_srrd_i210(struct e1000_hw *hw, u16 offset, 36s32 igb_read_nvm_srrd_i210(struct e1000_hw *hw, u16 offset, u16 words,
37 u16 words, u16 *data); 37 u16 *data);
38extern s32 igb_acquire_swfw_sync_i210(struct e1000_hw *hw, u16 mask); 38s32 igb_acquire_swfw_sync_i210(struct e1000_hw *hw, u16 mask);
39extern void igb_release_swfw_sync_i210(struct e1000_hw *hw, u16 mask); 39void igb_release_swfw_sync_i210(struct e1000_hw *hw, u16 mask);
40extern s32 igb_acquire_nvm_i210(struct e1000_hw *hw); 40s32 igb_acquire_nvm_i210(struct e1000_hw *hw);
41extern void igb_release_nvm_i210(struct e1000_hw *hw); 41void igb_release_nvm_i210(struct e1000_hw *hw);
42extern s32 igb_valid_led_default_i210(struct e1000_hw *hw, u16 *data); 42s32 igb_valid_led_default_i210(struct e1000_hw *hw, u16 *data);
43extern s32 igb_read_invm_version(struct e1000_hw *hw, 43s32 igb_read_invm_version(struct e1000_hw *hw,
44 struct e1000_fw_version *invm_ver); 44 struct e1000_fw_version *invm_ver);
45extern s32 igb_read_xmdio_reg(struct e1000_hw *hw, u16 addr, u8 dev_addr, 45s32 igb_read_xmdio_reg(struct e1000_hw *hw, u16 addr, u8 dev_addr, u16 *data);
46 u16 *data); 46s32 igb_write_xmdio_reg(struct e1000_hw *hw, u16 addr, u8 dev_addr, u16 data);
47extern s32 igb_write_xmdio_reg(struct e1000_hw *hw, u16 addr, u8 dev_addr, 47s32 igb_init_nvm_params_i210(struct e1000_hw *hw);
48 u16 data); 48bool igb_get_flash_presence_i210(struct e1000_hw *hw);
49extern s32 igb_init_nvm_params_i210(struct e1000_hw *hw);
50extern bool igb_get_flash_presence_i210(struct e1000_hw *hw);
51 49
52#define E1000_STM_OPCODE 0xDB00 50#define E1000_STM_OPCODE 0xDB00
53#define E1000_EEPROM_FLASH_SIZE_WORD 0x11 51#define E1000_EEPROM_FLASH_SIZE_WORD 0x11
diff --git a/drivers/net/ethernet/intel/igb/e1000_mac.h b/drivers/net/ethernet/intel/igb/e1000_mac.h
index 5e13e83cc608..e4cbe8ef67b3 100644
--- a/drivers/net/ethernet/intel/igb/e1000_mac.h
+++ b/drivers/net/ethernet/intel/igb/e1000_mac.h
@@ -86,6 +86,6 @@ enum e1000_mng_mode {
86 86
87#define E1000_MNG_DHCP_COOKIE_STATUS_VLAN 0x2 87#define E1000_MNG_DHCP_COOKIE_STATUS_VLAN 0x2
88 88
89extern void e1000_init_function_pointers_82575(struct e1000_hw *hw); 89void e1000_init_function_pointers_82575(struct e1000_hw *hw);
90 90
91#endif 91#endif
diff --git a/drivers/net/ethernet/intel/igb/e1000_phy.c b/drivers/net/ethernet/intel/igb/e1000_phy.c
index e7266759a10b..c4c4fe332c7e 100644
--- a/drivers/net/ethernet/intel/igb/e1000_phy.c
+++ b/drivers/net/ethernet/intel/igb/e1000_phy.c
@@ -708,11 +708,6 @@ s32 igb_copper_link_setup_m88(struct e1000_hw *hw)
708 hw_dbg("Error committing the PHY changes\n"); 708 hw_dbg("Error committing the PHY changes\n");
709 goto out; 709 goto out;
710 } 710 }
711 if (phy->type == e1000_phy_i210) {
712 ret_val = igb_set_master_slave_mode(hw);
713 if (ret_val)
714 return ret_val;
715 }
716 711
717out: 712out:
718 return ret_val; 713 return ret_val;
@@ -806,6 +801,9 @@ s32 igb_copper_link_setup_m88_gen2(struct e1000_hw *hw)
806 hw_dbg("Error committing the PHY changes\n"); 801 hw_dbg("Error committing the PHY changes\n");
807 return ret_val; 802 return ret_val;
808 } 803 }
804 ret_val = igb_set_master_slave_mode(hw);
805 if (ret_val)
806 return ret_val;
809 807
810 return 0; 808 return 0;
811} 809}
diff --git a/drivers/net/ethernet/intel/igb/igb.h b/drivers/net/ethernet/intel/igb/igb.h
index 6807b098edae..5e9ed89403aa 100644
--- a/drivers/net/ethernet/intel/igb/igb.h
+++ b/drivers/net/ethernet/intel/igb/igb.h
@@ -483,40 +483,38 @@ enum igb_boards {
483extern char igb_driver_name[]; 483extern char igb_driver_name[];
484extern char igb_driver_version[]; 484extern char igb_driver_version[];
485 485
486extern int igb_up(struct igb_adapter *); 486int igb_up(struct igb_adapter *);
487extern void igb_down(struct igb_adapter *); 487void igb_down(struct igb_adapter *);
488extern void igb_reinit_locked(struct igb_adapter *); 488void igb_reinit_locked(struct igb_adapter *);
489extern void igb_reset(struct igb_adapter *); 489void igb_reset(struct igb_adapter *);
490extern void igb_write_rss_indir_tbl(struct igb_adapter *); 490int igb_reinit_queues(struct igb_adapter *);
491extern int igb_set_spd_dplx(struct igb_adapter *, u32, u8); 491void igb_write_rss_indir_tbl(struct igb_adapter *);
492extern int igb_setup_tx_resources(struct igb_ring *); 492int igb_set_spd_dplx(struct igb_adapter *, u32, u8);
493extern int igb_setup_rx_resources(struct igb_ring *); 493int igb_setup_tx_resources(struct igb_ring *);
494extern void igb_free_tx_resources(struct igb_ring *); 494int igb_setup_rx_resources(struct igb_ring *);
495extern void igb_free_rx_resources(struct igb_ring *); 495void igb_free_tx_resources(struct igb_ring *);
496extern void igb_configure_tx_ring(struct igb_adapter *, struct igb_ring *); 496void igb_free_rx_resources(struct igb_ring *);
497extern void igb_configure_rx_ring(struct igb_adapter *, struct igb_ring *); 497void igb_configure_tx_ring(struct igb_adapter *, struct igb_ring *);
498extern void igb_setup_tctl(struct igb_adapter *); 498void igb_configure_rx_ring(struct igb_adapter *, struct igb_ring *);
499extern void igb_setup_rctl(struct igb_adapter *); 499void igb_setup_tctl(struct igb_adapter *);
500extern netdev_tx_t igb_xmit_frame_ring(struct sk_buff *, struct igb_ring *); 500void igb_setup_rctl(struct igb_adapter *);
501extern void igb_unmap_and_free_tx_resource(struct igb_ring *, 501netdev_tx_t igb_xmit_frame_ring(struct sk_buff *, struct igb_ring *);
502 struct igb_tx_buffer *); 502void igb_unmap_and_free_tx_resource(struct igb_ring *, struct igb_tx_buffer *);
503extern void igb_alloc_rx_buffers(struct igb_ring *, u16); 503void igb_alloc_rx_buffers(struct igb_ring *, u16);
504extern void igb_update_stats(struct igb_adapter *, struct rtnl_link_stats64 *); 504void igb_update_stats(struct igb_adapter *, struct rtnl_link_stats64 *);
505extern bool igb_has_link(struct igb_adapter *adapter); 505bool igb_has_link(struct igb_adapter *adapter);
506extern void igb_set_ethtool_ops(struct net_device *); 506void igb_set_ethtool_ops(struct net_device *);
507extern void igb_power_up_link(struct igb_adapter *); 507void igb_power_up_link(struct igb_adapter *);
508extern void igb_set_fw_version(struct igb_adapter *); 508void igb_set_fw_version(struct igb_adapter *);
509extern void igb_ptp_init(struct igb_adapter *adapter); 509void igb_ptp_init(struct igb_adapter *adapter);
510extern void igb_ptp_stop(struct igb_adapter *adapter); 510void igb_ptp_stop(struct igb_adapter *adapter);
511extern void igb_ptp_reset(struct igb_adapter *adapter); 511void igb_ptp_reset(struct igb_adapter *adapter);
512extern void igb_ptp_tx_work(struct work_struct *work); 512void igb_ptp_tx_work(struct work_struct *work);
513extern void igb_ptp_rx_hang(struct igb_adapter *adapter); 513void igb_ptp_rx_hang(struct igb_adapter *adapter);
514extern void igb_ptp_tx_hwtstamp(struct igb_adapter *adapter); 514void igb_ptp_tx_hwtstamp(struct igb_adapter *adapter);
515extern void igb_ptp_rx_rgtstamp(struct igb_q_vector *q_vector, 515void igb_ptp_rx_rgtstamp(struct igb_q_vector *q_vector, struct sk_buff *skb);
516 struct sk_buff *skb); 516void igb_ptp_rx_pktstamp(struct igb_q_vector *q_vector, unsigned char *va,
517extern void igb_ptp_rx_pktstamp(struct igb_q_vector *q_vector, 517 struct sk_buff *skb);
518 unsigned char *va,
519 struct sk_buff *skb);
520static inline void igb_ptp_rx_hwtstamp(struct igb_ring *rx_ring, 518static inline void igb_ptp_rx_hwtstamp(struct igb_ring *rx_ring,
521 union e1000_adv_rx_desc *rx_desc, 519 union e1000_adv_rx_desc *rx_desc,
522 struct sk_buff *skb) 520 struct sk_buff *skb)
@@ -531,11 +529,11 @@ static inline void igb_ptp_rx_hwtstamp(struct igb_ring *rx_ring,
531 rx_ring->last_rx_timestamp = jiffies; 529 rx_ring->last_rx_timestamp = jiffies;
532} 530}
533 531
534extern int igb_ptp_hwtstamp_ioctl(struct net_device *netdev, 532int igb_ptp_hwtstamp_ioctl(struct net_device *netdev, struct ifreq *ifr,
535 struct ifreq *ifr, int cmd); 533 int cmd);
536#ifdef CONFIG_IGB_HWMON 534#ifdef CONFIG_IGB_HWMON
537extern void igb_sysfs_exit(struct igb_adapter *adapter); 535void igb_sysfs_exit(struct igb_adapter *adapter);
538extern int igb_sysfs_init(struct igb_adapter *adapter); 536int igb_sysfs_init(struct igb_adapter *adapter);
539#endif 537#endif
540static inline s32 igb_reset_phy(struct e1000_hw *hw) 538static inline s32 igb_reset_phy(struct e1000_hw *hw)
541{ 539{
diff --git a/drivers/net/ethernet/intel/igb/igb_ethtool.c b/drivers/net/ethernet/intel/igb/igb_ethtool.c
index 151e00cad113..b0f3666b1d7f 100644
--- a/drivers/net/ethernet/intel/igb/igb_ethtool.c
+++ b/drivers/net/ethernet/intel/igb/igb_ethtool.c
@@ -146,6 +146,7 @@ static int igb_get_settings(struct net_device *netdev, struct ethtool_cmd *ecmd)
146 struct e1000_sfp_flags *eth_flags = &dev_spec->eth_flags; 146 struct e1000_sfp_flags *eth_flags = &dev_spec->eth_flags;
147 u32 status; 147 u32 status;
148 148
149 status = rd32(E1000_STATUS);
149 if (hw->phy.media_type == e1000_media_type_copper) { 150 if (hw->phy.media_type == e1000_media_type_copper) {
150 151
151 ecmd->supported = (SUPPORTED_10baseT_Half | 152 ecmd->supported = (SUPPORTED_10baseT_Half |
@@ -169,13 +170,22 @@ static int igb_get_settings(struct net_device *netdev, struct ethtool_cmd *ecmd)
169 ecmd->transceiver = XCVR_INTERNAL; 170 ecmd->transceiver = XCVR_INTERNAL;
170 } else { 171 } else {
171 ecmd->supported = (SUPPORTED_FIBRE | 172 ecmd->supported = (SUPPORTED_FIBRE |
173 SUPPORTED_1000baseKX_Full |
172 SUPPORTED_Autoneg | 174 SUPPORTED_Autoneg |
173 SUPPORTED_Pause); 175 SUPPORTED_Pause);
174 ecmd->advertising = ADVERTISED_FIBRE; 176 ecmd->advertising = (ADVERTISED_FIBRE |
175 177 ADVERTISED_1000baseKX_Full);
176 if ((eth_flags->e1000_base_lx) || (eth_flags->e1000_base_sx)) { 178 if (hw->mac.type == e1000_i354) {
177 ecmd->supported |= SUPPORTED_1000baseT_Full; 179 if ((hw->device_id ==
178 ecmd->advertising |= ADVERTISED_1000baseT_Full; 180 E1000_DEV_ID_I354_BACKPLANE_2_5GBPS) &&
181 !(status & E1000_STATUS_2P5_SKU_OVER)) {
182 ecmd->supported |= SUPPORTED_2500baseX_Full;
183 ecmd->supported &=
184 ~SUPPORTED_1000baseKX_Full;
185 ecmd->advertising |= ADVERTISED_2500baseX_Full;
186 ecmd->advertising &=
187 ~ADVERTISED_1000baseKX_Full;
188 }
179 } 189 }
180 if (eth_flags->e100_base_fx) { 190 if (eth_flags->e100_base_fx) {
181 ecmd->supported |= SUPPORTED_100baseT_Full; 191 ecmd->supported |= SUPPORTED_100baseT_Full;
@@ -187,35 +197,29 @@ static int igb_get_settings(struct net_device *netdev, struct ethtool_cmd *ecmd)
187 ecmd->port = PORT_FIBRE; 197 ecmd->port = PORT_FIBRE;
188 ecmd->transceiver = XCVR_EXTERNAL; 198 ecmd->transceiver = XCVR_EXTERNAL;
189 } 199 }
190
191 if (hw->mac.autoneg != 1) 200 if (hw->mac.autoneg != 1)
192 ecmd->advertising &= ~(ADVERTISED_Pause | 201 ecmd->advertising &= ~(ADVERTISED_Pause |
193 ADVERTISED_Asym_Pause); 202 ADVERTISED_Asym_Pause);
194 203
195 if (hw->fc.requested_mode == e1000_fc_full) 204 switch (hw->fc.requested_mode) {
205 case e1000_fc_full:
196 ecmd->advertising |= ADVERTISED_Pause; 206 ecmd->advertising |= ADVERTISED_Pause;
197 else if (hw->fc.requested_mode == e1000_fc_rx_pause) 207 break;
208 case e1000_fc_rx_pause:
198 ecmd->advertising |= (ADVERTISED_Pause | 209 ecmd->advertising |= (ADVERTISED_Pause |
199 ADVERTISED_Asym_Pause); 210 ADVERTISED_Asym_Pause);
200 else if (hw->fc.requested_mode == e1000_fc_tx_pause) 211 break;
212 case e1000_fc_tx_pause:
201 ecmd->advertising |= ADVERTISED_Asym_Pause; 213 ecmd->advertising |= ADVERTISED_Asym_Pause;
202 else 214 break;
215 default:
203 ecmd->advertising &= ~(ADVERTISED_Pause | 216 ecmd->advertising &= ~(ADVERTISED_Pause |
204 ADVERTISED_Asym_Pause); 217 ADVERTISED_Asym_Pause);
205 218 }
206 status = rd32(E1000_STATUS);
207
208 if (status & E1000_STATUS_LU) { 219 if (status & E1000_STATUS_LU) {
209 if (hw->mac.type == e1000_i354) { 220 if ((status & E1000_STATUS_2P5_SKU) &&
210 if ((status & E1000_STATUS_2P5_SKU) && 221 !(status & E1000_STATUS_2P5_SKU_OVER)) {
211 !(status & E1000_STATUS_2P5_SKU_OVER)) { 222 ecmd->speed = SPEED_2500;
212 ecmd->supported = SUPPORTED_2500baseX_Full;
213 ecmd->advertising = ADVERTISED_2500baseX_Full;
214 ecmd->speed = SPEED_2500;
215 } else {
216 ecmd->supported = SUPPORTED_1000baseT_Full;
217 ecmd->advertising = ADVERTISED_1000baseT_Full;
218 }
219 } else if (status & E1000_STATUS_SPEED_1000) { 223 } else if (status & E1000_STATUS_SPEED_1000) {
220 ecmd->speed = SPEED_1000; 224 ecmd->speed = SPEED_1000;
221 } else if (status & E1000_STATUS_SPEED_100) { 225 } else if (status & E1000_STATUS_SPEED_100) {
@@ -232,7 +236,6 @@ static int igb_get_settings(struct net_device *netdev, struct ethtool_cmd *ecmd)
232 ecmd->speed = -1; 236 ecmd->speed = -1;
233 ecmd->duplex = -1; 237 ecmd->duplex = -1;
234 } 238 }
235
236 if ((hw->phy.media_type == e1000_media_type_fiber) || 239 if ((hw->phy.media_type == e1000_media_type_fiber) ||
237 hw->mac.autoneg) 240 hw->mac.autoneg)
238 ecmd->autoneg = AUTONEG_ENABLE; 241 ecmd->autoneg = AUTONEG_ENABLE;
@@ -771,8 +774,10 @@ static int igb_set_eeprom(struct net_device *netdev,
771 if (eeprom->len == 0) 774 if (eeprom->len == 0)
772 return -EOPNOTSUPP; 775 return -EOPNOTSUPP;
773 776
774 if (hw->mac.type == e1000_i211) 777 if ((hw->mac.type >= e1000_i210) &&
778 !igb_get_flash_presence_i210(hw)) {
775 return -EOPNOTSUPP; 779 return -EOPNOTSUPP;
780 }
776 781
777 if (eeprom->magic != (hw->vendor_id | (hw->device_id << 16))) 782 if (eeprom->magic != (hw->vendor_id | (hw->device_id << 16)))
778 return -EFAULT; 783 return -EFAULT;
@@ -1659,7 +1664,8 @@ static int igb_setup_loopback_test(struct igb_adapter *adapter)
1659 if ((hw->device_id == E1000_DEV_ID_DH89XXCC_SGMII) || 1664 if ((hw->device_id == E1000_DEV_ID_DH89XXCC_SGMII) ||
1660 (hw->device_id == E1000_DEV_ID_DH89XXCC_SERDES) || 1665 (hw->device_id == E1000_DEV_ID_DH89XXCC_SERDES) ||
1661 (hw->device_id == E1000_DEV_ID_DH89XXCC_BACKPLANE) || 1666 (hw->device_id == E1000_DEV_ID_DH89XXCC_BACKPLANE) ||
1662 (hw->device_id == E1000_DEV_ID_DH89XXCC_SFP)) { 1667 (hw->device_id == E1000_DEV_ID_DH89XXCC_SFP) ||
1668 (hw->device_id == E1000_DEV_ID_I354_SGMII)) {
1663 1669
1664 /* Enable DH89xxCC MPHY for near end loopback */ 1670 /* Enable DH89xxCC MPHY for near end loopback */
1665 reg = rd32(E1000_MPHY_ADDR_CTL); 1671 reg = rd32(E1000_MPHY_ADDR_CTL);
@@ -1725,7 +1731,8 @@ static void igb_loopback_cleanup(struct igb_adapter *adapter)
1725 if ((hw->device_id == E1000_DEV_ID_DH89XXCC_SGMII) || 1731 if ((hw->device_id == E1000_DEV_ID_DH89XXCC_SGMII) ||
1726 (hw->device_id == E1000_DEV_ID_DH89XXCC_SERDES) || 1732 (hw->device_id == E1000_DEV_ID_DH89XXCC_SERDES) ||
1727 (hw->device_id == E1000_DEV_ID_DH89XXCC_BACKPLANE) || 1733 (hw->device_id == E1000_DEV_ID_DH89XXCC_BACKPLANE) ||
1728 (hw->device_id == E1000_DEV_ID_DH89XXCC_SFP)) { 1734 (hw->device_id == E1000_DEV_ID_DH89XXCC_SFP) ||
1735 (hw->device_id == E1000_DEV_ID_I354_SGMII)) {
1729 u32 reg; 1736 u32 reg;
1730 1737
1731 /* Disable near end loopback on DH89xxCC */ 1738 /* Disable near end loopback on DH89xxCC */
@@ -2877,6 +2884,88 @@ static int igb_set_rxfh_indir(struct net_device *netdev, const u32 *indir)
2877 return 0; 2884 return 0;
2878} 2885}
2879 2886
2887static unsigned int igb_max_channels(struct igb_adapter *adapter)
2888{
2889 struct e1000_hw *hw = &adapter->hw;
2890 unsigned int max_combined = 0;
2891
2892 switch (hw->mac.type) {
2893 case e1000_i211:
2894 max_combined = IGB_MAX_RX_QUEUES_I211;
2895 break;
2896 case e1000_82575:
2897 case e1000_i210:
2898 max_combined = IGB_MAX_RX_QUEUES_82575;
2899 break;
2900 case e1000_i350:
2901 if (!!adapter->vfs_allocated_count) {
2902 max_combined = 1;
2903 break;
2904 }
2905 /* fall through */
2906 case e1000_82576:
2907 if (!!adapter->vfs_allocated_count) {
2908 max_combined = 2;
2909 break;
2910 }
2911 /* fall through */
2912 case e1000_82580:
2913 case e1000_i354:
2914 default:
2915 max_combined = IGB_MAX_RX_QUEUES;
2916 break;
2917 }
2918
2919 return max_combined;
2920}
2921
2922static void igb_get_channels(struct net_device *netdev,
2923 struct ethtool_channels *ch)
2924{
2925 struct igb_adapter *adapter = netdev_priv(netdev);
2926
2927 /* Report maximum channels */
2928 ch->max_combined = igb_max_channels(adapter);
2929
2930 /* Report info for other vector */
2931 if (adapter->msix_entries) {
2932 ch->max_other = NON_Q_VECTORS;
2933 ch->other_count = NON_Q_VECTORS;
2934 }
2935
2936 ch->combined_count = adapter->rss_queues;
2937}
2938
2939static int igb_set_channels(struct net_device *netdev,
2940 struct ethtool_channels *ch)
2941{
2942 struct igb_adapter *adapter = netdev_priv(netdev);
2943 unsigned int count = ch->combined_count;
2944
2945 /* Verify they are not requesting separate vectors */
2946 if (!count || ch->rx_count || ch->tx_count)
2947 return -EINVAL;
2948
2949 /* Verify other_count is valid and has not been changed */
2950 if (ch->other_count != NON_Q_VECTORS)
2951 return -EINVAL;
2952
2953 /* Verify the number of channels doesn't exceed hw limits */
2954 if (count > igb_max_channels(adapter))
2955 return -EINVAL;
2956
2957 if (count != adapter->rss_queues) {
2958 adapter->rss_queues = count;
2959
2960 /* Hardware has to reinitialize queues and interrupts to
2961 * match the new configuration.
2962 */
2963 return igb_reinit_queues(adapter);
2964 }
2965
2966 return 0;
2967}
2968
2880static const struct ethtool_ops igb_ethtool_ops = { 2969static const struct ethtool_ops igb_ethtool_ops = {
2881 .get_settings = igb_get_settings, 2970 .get_settings = igb_get_settings,
2882 .set_settings = igb_set_settings, 2971 .set_settings = igb_set_settings,
@@ -2913,6 +3002,8 @@ static const struct ethtool_ops igb_ethtool_ops = {
2913 .get_rxfh_indir_size = igb_get_rxfh_indir_size, 3002 .get_rxfh_indir_size = igb_get_rxfh_indir_size,
2914 .get_rxfh_indir = igb_get_rxfh_indir, 3003 .get_rxfh_indir = igb_get_rxfh_indir,
2915 .set_rxfh_indir = igb_set_rxfh_indir, 3004 .set_rxfh_indir = igb_set_rxfh_indir,
3005 .get_channels = igb_get_channels,
3006 .set_channels = igb_set_channels,
2916 .begin = igb_ethtool_begin, 3007 .begin = igb_ethtool_begin,
2917 .complete = igb_ethtool_complete, 3008 .complete = igb_ethtool_complete,
2918}; 3009};
diff --git a/drivers/net/ethernet/intel/igb/igb_main.c b/drivers/net/ethernet/intel/igb/igb_main.c
index 8cf44f2a8ccd..ebe6370c4b18 100644
--- a/drivers/net/ethernet/intel/igb/igb_main.c
+++ b/drivers/net/ethernet/intel/igb/igb_main.c
@@ -182,6 +182,7 @@ static void igb_check_vf_rate_limit(struct igb_adapter *);
182 182
183#ifdef CONFIG_PCI_IOV 183#ifdef CONFIG_PCI_IOV
184static int igb_vf_configure(struct igb_adapter *adapter, int vf); 184static int igb_vf_configure(struct igb_adapter *adapter, int vf);
185static int igb_pci_enable_sriov(struct pci_dev *dev, int num_vfs);
185#endif 186#endif
186 187
187#ifdef CONFIG_PM 188#ifdef CONFIG_PM
@@ -2429,7 +2430,7 @@ err_dma:
2429} 2430}
2430 2431
2431#ifdef CONFIG_PCI_IOV 2432#ifdef CONFIG_PCI_IOV
2432static int igb_disable_sriov(struct pci_dev *pdev) 2433static int igb_disable_sriov(struct pci_dev *pdev)
2433{ 2434{
2434 struct net_device *netdev = pci_get_drvdata(pdev); 2435 struct net_device *netdev = pci_get_drvdata(pdev);
2435 struct igb_adapter *adapter = netdev_priv(netdev); 2436 struct igb_adapter *adapter = netdev_priv(netdev);
@@ -2470,27 +2471,19 @@ static int igb_enable_sriov(struct pci_dev *pdev, int num_vfs)
2470 int err = 0; 2471 int err = 0;
2471 int i; 2472 int i;
2472 2473
2473 if (!adapter->msix_entries) { 2474 if (!adapter->msix_entries || num_vfs > 7) {
2474 err = -EPERM; 2475 err = -EPERM;
2475 goto out; 2476 goto out;
2476 } 2477 }
2477
2478 if (!num_vfs) 2478 if (!num_vfs)
2479 goto out; 2479 goto out;
2480 else if (old_vfs && old_vfs == num_vfs)
2481 goto out;
2482 else if (old_vfs && old_vfs != num_vfs)
2483 err = igb_disable_sriov(pdev);
2484
2485 if (err)
2486 goto out;
2487
2488 if (num_vfs > 7) {
2489 err = -EPERM;
2490 goto out;
2491 }
2492 2480
2493 adapter->vfs_allocated_count = num_vfs; 2481 if (old_vfs) {
2482 dev_info(&pdev->dev, "%d pre-allocated VFs found - override max_vfs setting of %d\n",
2483 old_vfs, max_vfs);
2484 adapter->vfs_allocated_count = old_vfs;
2485 } else
2486 adapter->vfs_allocated_count = num_vfs;
2494 2487
2495 adapter->vf_data = kcalloc(adapter->vfs_allocated_count, 2488 adapter->vf_data = kcalloc(adapter->vfs_allocated_count,
2496 sizeof(struct vf_data_storage), GFP_KERNEL); 2489 sizeof(struct vf_data_storage), GFP_KERNEL);
@@ -2504,10 +2497,12 @@ static int igb_enable_sriov(struct pci_dev *pdev, int num_vfs)
2504 goto out; 2497 goto out;
2505 } 2498 }
2506 2499
2507 err = pci_enable_sriov(pdev, adapter->vfs_allocated_count); 2500 /* only call pci_enable_sriov() if no VFs are allocated already */
2508 if (err) 2501 if (!old_vfs) {
2509 goto err_out; 2502 err = pci_enable_sriov(pdev, adapter->vfs_allocated_count);
2510 2503 if (err)
2504 goto err_out;
2505 }
2511 dev_info(&pdev->dev, "%d VFs allocated\n", 2506 dev_info(&pdev->dev, "%d VFs allocated\n",
2512 adapter->vfs_allocated_count); 2507 adapter->vfs_allocated_count);
2513 for (i = 0; i < adapter->vfs_allocated_count; i++) 2508 for (i = 0; i < adapter->vfs_allocated_count; i++)
@@ -2623,7 +2618,7 @@ static void igb_probe_vfs(struct igb_adapter *adapter)
2623 return; 2618 return;
2624 2619
2625 pci_sriov_set_totalvfs(pdev, 7); 2620 pci_sriov_set_totalvfs(pdev, 7);
2626 igb_enable_sriov(pdev, max_vfs); 2621 igb_pci_enable_sriov(pdev, max_vfs);
2627 2622
2628#endif /* CONFIG_PCI_IOV */ 2623#endif /* CONFIG_PCI_IOV */
2629} 2624}
@@ -5708,7 +5703,7 @@ static void igb_vf_reset_msg(struct igb_adapter *adapter, u32 vf)
5708 5703
5709 /* reply to reset with ack and vf mac address */ 5704 /* reply to reset with ack and vf mac address */
5710 msgbuf[0] = E1000_VF_RESET | E1000_VT_MSGTYPE_ACK; 5705 msgbuf[0] = E1000_VF_RESET | E1000_VT_MSGTYPE_ACK;
5711 memcpy(addr, vf_mac, 6); 5706 memcpy(addr, vf_mac, ETH_ALEN);
5712 igb_write_mbx(hw, msgbuf, 3, vf); 5707 igb_write_mbx(hw, msgbuf, 3, vf);
5713} 5708}
5714 5709
@@ -7838,4 +7833,26 @@ s32 igb_write_i2c_byte(struct e1000_hw *hw, u8 byte_offset,
7838 return E1000_SUCCESS; 7833 return E1000_SUCCESS;
7839 7834
7840} 7835}
7836
7837int igb_reinit_queues(struct igb_adapter *adapter)
7838{
7839 struct net_device *netdev = adapter->netdev;
7840 struct pci_dev *pdev = adapter->pdev;
7841 int err = 0;
7842
7843 if (netif_running(netdev))
7844 igb_close(netdev);
7845
7846 igb_clear_interrupt_scheme(adapter);
7847
7848 if (igb_init_interrupt_scheme(adapter, true)) {
7849 dev_err(&pdev->dev, "Unable to allocate memory for queues\n");
7850 return -ENOMEM;
7851 }
7852
7853 if (netif_running(netdev))
7854 err = igb_open(netdev);
7855
7856 return err;
7857}
7841/* igb_main.c */ 7858/* igb_main.c */
diff --git a/drivers/net/ethernet/intel/igbvf/igbvf.h b/drivers/net/ethernet/intel/igbvf/igbvf.h
index a1463e3d14c0..7d6a25c8f889 100644
--- a/drivers/net/ethernet/intel/igbvf/igbvf.h
+++ b/drivers/net/ethernet/intel/igbvf/igbvf.h
@@ -312,17 +312,17 @@ enum igbvf_state_t {
312extern char igbvf_driver_name[]; 312extern char igbvf_driver_name[];
313extern const char igbvf_driver_version[]; 313extern const char igbvf_driver_version[];
314 314
315extern void igbvf_check_options(struct igbvf_adapter *); 315void igbvf_check_options(struct igbvf_adapter *);
316extern void igbvf_set_ethtool_ops(struct net_device *); 316void igbvf_set_ethtool_ops(struct net_device *);
317 317
318extern int igbvf_up(struct igbvf_adapter *); 318int igbvf_up(struct igbvf_adapter *);
319extern void igbvf_down(struct igbvf_adapter *); 319void igbvf_down(struct igbvf_adapter *);
320extern void igbvf_reinit_locked(struct igbvf_adapter *); 320void igbvf_reinit_locked(struct igbvf_adapter *);
321extern int igbvf_setup_rx_resources(struct igbvf_adapter *, struct igbvf_ring *); 321int igbvf_setup_rx_resources(struct igbvf_adapter *, struct igbvf_ring *);
322extern int igbvf_setup_tx_resources(struct igbvf_adapter *, struct igbvf_ring *); 322int igbvf_setup_tx_resources(struct igbvf_adapter *, struct igbvf_ring *);
323extern void igbvf_free_rx_resources(struct igbvf_ring *); 323void igbvf_free_rx_resources(struct igbvf_ring *);
324extern void igbvf_free_tx_resources(struct igbvf_ring *); 324void igbvf_free_tx_resources(struct igbvf_ring *);
325extern void igbvf_update_stats(struct igbvf_adapter *); 325void igbvf_update_stats(struct igbvf_adapter *);
326 326
327extern unsigned int copybreak; 327extern unsigned int copybreak;
328 328
diff --git a/drivers/net/ethernet/intel/igbvf/netdev.c b/drivers/net/ethernet/intel/igbvf/netdev.c
index 93eb7ee06d3e..9fadbb28cf08 100644
--- a/drivers/net/ethernet/intel/igbvf/netdev.c
+++ b/drivers/net/ethernet/intel/igbvf/netdev.c
@@ -2343,10 +2343,9 @@ static int igbvf_change_mtu(struct net_device *netdev, int new_mtu)
2343 struct igbvf_adapter *adapter = netdev_priv(netdev); 2343 struct igbvf_adapter *adapter = netdev_priv(netdev);
2344 int max_frame = new_mtu + ETH_HLEN + ETH_FCS_LEN; 2344 int max_frame = new_mtu + ETH_HLEN + ETH_FCS_LEN;
2345 2345
2346 if ((new_mtu < 68) || (max_frame > MAX_JUMBO_FRAME_SIZE)) { 2346 if (new_mtu < 68 || new_mtu > INT_MAX - ETH_HLEN - ETH_FCS_LEN ||
2347 dev_err(&adapter->pdev->dev, "Invalid MTU setting\n"); 2347 max_frame > MAX_JUMBO_FRAME_SIZE)
2348 return -EINVAL; 2348 return -EINVAL;
2349 }
2350 2349
2351#define MAX_STD_JUMBO_FRAME_SIZE 9234 2350#define MAX_STD_JUMBO_FRAME_SIZE 9234
2352 if (max_frame > MAX_STD_JUMBO_FRAME_SIZE) { 2351 if (max_frame > MAX_STD_JUMBO_FRAME_SIZE) {
@@ -2699,7 +2698,7 @@ static int igbvf_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
2699 if (ei->get_variants) { 2698 if (ei->get_variants) {
2700 err = ei->get_variants(adapter); 2699 err = ei->get_variants(adapter);
2701 if (err) 2700 if (err)
2702 goto err_ioremap; 2701 goto err_get_variants;
2703 } 2702 }
2704 2703
2705 /* setup adapter struct */ 2704 /* setup adapter struct */
@@ -2796,6 +2795,7 @@ err_hw_init:
2796 kfree(adapter->rx_ring); 2795 kfree(adapter->rx_ring);
2797err_sw_init: 2796err_sw_init:
2798 igbvf_reset_interrupt_capability(adapter); 2797 igbvf_reset_interrupt_capability(adapter);
2798err_get_variants:
2799 iounmap(adapter->hw.hw_addr); 2799 iounmap(adapter->hw.hw_addr);
2800err_ioremap: 2800err_ioremap:
2801 free_netdev(netdev); 2801 free_netdev(netdev);
diff --git a/drivers/net/ethernet/intel/igbvf/vf.c b/drivers/net/ethernet/intel/igbvf/vf.c
index eea0e10ce12f..955ad8c2c534 100644
--- a/drivers/net/ethernet/intel/igbvf/vf.c
+++ b/drivers/net/ethernet/intel/igbvf/vf.c
@@ -154,7 +154,7 @@ static s32 e1000_reset_hw_vf(struct e1000_hw *hw)
154 ret_val = mbx->ops.read_posted(hw, msgbuf, 3); 154 ret_val = mbx->ops.read_posted(hw, msgbuf, 3);
155 if (!ret_val) { 155 if (!ret_val) {
156 if (msgbuf[0] == (E1000_VF_RESET | E1000_VT_MSGTYPE_ACK)) 156 if (msgbuf[0] == (E1000_VF_RESET | E1000_VT_MSGTYPE_ACK))
157 memcpy(hw->mac.perm_addr, addr, 6); 157 memcpy(hw->mac.perm_addr, addr, ETH_ALEN);
158 else 158 else
159 ret_val = -E1000_ERR_MAC_INIT; 159 ret_val = -E1000_ERR_MAC_INIT;
160 } 160 }
@@ -314,7 +314,7 @@ static void e1000_rar_set_vf(struct e1000_hw *hw, u8 * addr, u32 index)
314 314
315 memset(msgbuf, 0, 12); 315 memset(msgbuf, 0, 12);
316 msgbuf[0] = E1000_VF_SET_MAC_ADDR; 316 msgbuf[0] = E1000_VF_SET_MAC_ADDR;
317 memcpy(msg_addr, addr, 6); 317 memcpy(msg_addr, addr, ETH_ALEN);
318 ret_val = mbx->ops.write_posted(hw, msgbuf, 3); 318 ret_val = mbx->ops.write_posted(hw, msgbuf, 3);
319 319
320 if (!ret_val) 320 if (!ret_val)
diff --git a/drivers/net/ethernet/intel/ixgb/ixgb.h b/drivers/net/ethernet/intel/ixgb/ixgb.h
index 4d2ae97ff1b3..2224cc2edf13 100644
--- a/drivers/net/ethernet/intel/ixgb/ixgb.h
+++ b/drivers/net/ethernet/intel/ixgb/ixgb.h
@@ -187,21 +187,21 @@ enum ixgb_state_t {
187}; 187};
188 188
189/* Exported from other modules */ 189/* Exported from other modules */
190extern void ixgb_check_options(struct ixgb_adapter *adapter); 190void ixgb_check_options(struct ixgb_adapter *adapter);
191extern void ixgb_set_ethtool_ops(struct net_device *netdev); 191void ixgb_set_ethtool_ops(struct net_device *netdev);
192extern char ixgb_driver_name[]; 192extern char ixgb_driver_name[];
193extern const char ixgb_driver_version[]; 193extern const char ixgb_driver_version[];
194 194
195extern void ixgb_set_speed_duplex(struct net_device *netdev); 195void ixgb_set_speed_duplex(struct net_device *netdev);
196 196
197extern int ixgb_up(struct ixgb_adapter *adapter); 197int ixgb_up(struct ixgb_adapter *adapter);
198extern void ixgb_down(struct ixgb_adapter *adapter, bool kill_watchdog); 198void ixgb_down(struct ixgb_adapter *adapter, bool kill_watchdog);
199extern void ixgb_reset(struct ixgb_adapter *adapter); 199void ixgb_reset(struct ixgb_adapter *adapter);
200extern int ixgb_setup_rx_resources(struct ixgb_adapter *adapter); 200int ixgb_setup_rx_resources(struct ixgb_adapter *adapter);
201extern int ixgb_setup_tx_resources(struct ixgb_adapter *adapter); 201int ixgb_setup_tx_resources(struct ixgb_adapter *adapter);
202extern void ixgb_free_rx_resources(struct ixgb_adapter *adapter); 202void ixgb_free_rx_resources(struct ixgb_adapter *adapter);
203extern void ixgb_free_tx_resources(struct ixgb_adapter *adapter); 203void ixgb_free_tx_resources(struct ixgb_adapter *adapter);
204extern void ixgb_update_stats(struct ixgb_adapter *adapter); 204void ixgb_update_stats(struct ixgb_adapter *adapter);
205 205
206 206
207#endif /* _IXGB_H_ */ 207#endif /* _IXGB_H_ */
diff --git a/drivers/net/ethernet/intel/ixgb/ixgb_hw.h b/drivers/net/ethernet/intel/ixgb/ixgb_hw.h
index 2a99a35c33aa..0bd5d72e1af5 100644
--- a/drivers/net/ethernet/intel/ixgb/ixgb_hw.h
+++ b/drivers/net/ethernet/intel/ixgb/ixgb_hw.h
@@ -759,27 +759,20 @@ struct ixgb_hw_stats {
759}; 759};
760 760
761/* Function Prototypes */ 761/* Function Prototypes */
762extern bool ixgb_adapter_stop(struct ixgb_hw *hw); 762bool ixgb_adapter_stop(struct ixgb_hw *hw);
763extern bool ixgb_init_hw(struct ixgb_hw *hw); 763bool ixgb_init_hw(struct ixgb_hw *hw);
764extern bool ixgb_adapter_start(struct ixgb_hw *hw); 764bool ixgb_adapter_start(struct ixgb_hw *hw);
765extern void ixgb_check_for_link(struct ixgb_hw *hw); 765void ixgb_check_for_link(struct ixgb_hw *hw);
766extern bool ixgb_check_for_bad_link(struct ixgb_hw *hw); 766bool ixgb_check_for_bad_link(struct ixgb_hw *hw);
767
768extern void ixgb_rar_set(struct ixgb_hw *hw,
769 u8 *addr,
770 u32 index);
771 767
768void ixgb_rar_set(struct ixgb_hw *hw, u8 *addr, u32 index);
772 769
773/* Filters (multicast, vlan, receive) */ 770/* Filters (multicast, vlan, receive) */
774extern void ixgb_mc_addr_list_update(struct ixgb_hw *hw, 771void ixgb_mc_addr_list_update(struct ixgb_hw *hw, u8 *mc_addr_list,
775 u8 *mc_addr_list, 772 u32 mc_addr_count, u32 pad);
776 u32 mc_addr_count,
777 u32 pad);
778 773
779/* Vfta functions */ 774/* Vfta functions */
780extern void ixgb_write_vfta(struct ixgb_hw *hw, 775void ixgb_write_vfta(struct ixgb_hw *hw, u32 offset, u32 value);
781 u32 offset,
782 u32 value);
783 776
784/* Access functions to eeprom data */ 777/* Access functions to eeprom data */
785void ixgb_get_ee_mac_addr(struct ixgb_hw *hw, u8 *mac_addr); 778void ixgb_get_ee_mac_addr(struct ixgb_hw *hw, u8 *mac_addr);
diff --git a/drivers/net/ethernet/intel/ixgbe/ixgbe.h b/drivers/net/ethernet/intel/ixgbe/ixgbe.h
index 0ac6b11c6e4e..f38fc0a343a2 100644
--- a/drivers/net/ethernet/intel/ixgbe/ixgbe.h
+++ b/drivers/net/ethernet/intel/ixgbe/ixgbe.h
@@ -55,7 +55,7 @@
55#include <net/busy_poll.h> 55#include <net/busy_poll.h>
56 56
57#ifdef CONFIG_NET_RX_BUSY_POLL 57#ifdef CONFIG_NET_RX_BUSY_POLL
58#define LL_EXTENDED_STATS 58#define BP_EXTENDED_STATS
59#endif 59#endif
60/* common prefix used by pr_<> macros */ 60/* common prefix used by pr_<> macros */
61#undef pr_fmt 61#undef pr_fmt
@@ -67,7 +67,11 @@
67#define IXGBE_MAX_TXD 4096 67#define IXGBE_MAX_TXD 4096
68#define IXGBE_MIN_TXD 64 68#define IXGBE_MIN_TXD 64
69 69
70#if (PAGE_SIZE < 8192)
70#define IXGBE_DEFAULT_RXD 512 71#define IXGBE_DEFAULT_RXD 512
72#else
73#define IXGBE_DEFAULT_RXD 128
74#endif
71#define IXGBE_MAX_RXD 4096 75#define IXGBE_MAX_RXD 4096
72#define IXGBE_MIN_RXD 64 76#define IXGBE_MIN_RXD 64
73 77
@@ -187,11 +191,11 @@ struct ixgbe_rx_buffer {
187struct ixgbe_queue_stats { 191struct ixgbe_queue_stats {
188 u64 packets; 192 u64 packets;
189 u64 bytes; 193 u64 bytes;
190#ifdef LL_EXTENDED_STATS 194#ifdef BP_EXTENDED_STATS
191 u64 yields; 195 u64 yields;
192 u64 misses; 196 u64 misses;
193 u64 cleaned; 197 u64 cleaned;
194#endif /* LL_EXTENDED_STATS */ 198#endif /* BP_EXTENDED_STATS */
195}; 199};
196 200
197struct ixgbe_tx_queue_stats { 201struct ixgbe_tx_queue_stats {
@@ -219,6 +223,15 @@ enum ixgbe_ring_state_t {
219 __IXGBE_RX_FCOE, 223 __IXGBE_RX_FCOE,
220}; 224};
221 225
226struct ixgbe_fwd_adapter {
227 unsigned long active_vlans[BITS_TO_LONGS(VLAN_N_VID)];
228 struct net_device *netdev;
229 struct ixgbe_adapter *real_adapter;
230 unsigned int tx_base_queue;
231 unsigned int rx_base_queue;
232 int pool;
233};
234
222#define check_for_tx_hang(ring) \ 235#define check_for_tx_hang(ring) \
223 test_bit(__IXGBE_TX_DETECT_HANG, &(ring)->state) 236 test_bit(__IXGBE_TX_DETECT_HANG, &(ring)->state)
224#define set_check_for_tx_hang(ring) \ 237#define set_check_for_tx_hang(ring) \
@@ -236,6 +249,7 @@ struct ixgbe_ring {
236 struct ixgbe_q_vector *q_vector; /* backpointer to host q_vector */ 249 struct ixgbe_q_vector *q_vector; /* backpointer to host q_vector */
237 struct net_device *netdev; /* netdev ring belongs to */ 250 struct net_device *netdev; /* netdev ring belongs to */
238 struct device *dev; /* device for DMA mapping */ 251 struct device *dev; /* device for DMA mapping */
252 struct ixgbe_fwd_adapter *l2_accel_priv;
239 void *desc; /* descriptor ring memory */ 253 void *desc; /* descriptor ring memory */
240 union { 254 union {
241 struct ixgbe_tx_buffer *tx_buffer_info; 255 struct ixgbe_tx_buffer *tx_buffer_info;
@@ -293,6 +307,12 @@ enum ixgbe_ring_f_enum {
293#define IXGBE_MAX_FCOE_INDICES 8 307#define IXGBE_MAX_FCOE_INDICES 8
294#define MAX_RX_QUEUES (IXGBE_MAX_FDIR_INDICES + 1) 308#define MAX_RX_QUEUES (IXGBE_MAX_FDIR_INDICES + 1)
295#define MAX_TX_QUEUES (IXGBE_MAX_FDIR_INDICES + 1) 309#define MAX_TX_QUEUES (IXGBE_MAX_FDIR_INDICES + 1)
310#define IXGBE_MAX_L2A_QUEUES 4
311#define IXGBE_MAX_L2A_QUEUES 4
312#define IXGBE_BAD_L2A_QUEUE 3
313#define IXGBE_MAX_MACVLANS 31
314#define IXGBE_MAX_DCBMACVLANS 8
315
296struct ixgbe_ring_feature { 316struct ixgbe_ring_feature {
297 u16 limit; /* upper limit on feature indices */ 317 u16 limit; /* upper limit on feature indices */
298 u16 indices; /* current value of indices */ 318 u16 indices; /* current value of indices */
@@ -369,11 +389,13 @@ struct ixgbe_q_vector {
369#ifdef CONFIG_NET_RX_BUSY_POLL 389#ifdef CONFIG_NET_RX_BUSY_POLL
370 unsigned int state; 390 unsigned int state;
371#define IXGBE_QV_STATE_IDLE 0 391#define IXGBE_QV_STATE_IDLE 0
372#define IXGBE_QV_STATE_NAPI 1 /* NAPI owns this QV */ 392#define IXGBE_QV_STATE_NAPI 1 /* NAPI owns this QV */
373#define IXGBE_QV_STATE_POLL 2 /* poll owns this QV */ 393#define IXGBE_QV_STATE_POLL 2 /* poll owns this QV */
374#define IXGBE_QV_LOCKED (IXGBE_QV_STATE_NAPI | IXGBE_QV_STATE_POLL) 394#define IXGBE_QV_STATE_DISABLED 4 /* QV is disabled */
375#define IXGBE_QV_STATE_NAPI_YIELD 4 /* NAPI yielded this QV */ 395#define IXGBE_QV_OWNED (IXGBE_QV_STATE_NAPI | IXGBE_QV_STATE_POLL)
376#define IXGBE_QV_STATE_POLL_YIELD 8 /* poll yielded this QV */ 396#define IXGBE_QV_LOCKED (IXGBE_QV_OWNED | IXGBE_QV_STATE_DISABLED)
397#define IXGBE_QV_STATE_NAPI_YIELD 8 /* NAPI yielded this QV */
398#define IXGBE_QV_STATE_POLL_YIELD 16 /* poll yielded this QV */
377#define IXGBE_QV_YIELD (IXGBE_QV_STATE_NAPI_YIELD | IXGBE_QV_STATE_POLL_YIELD) 399#define IXGBE_QV_YIELD (IXGBE_QV_STATE_NAPI_YIELD | IXGBE_QV_STATE_POLL_YIELD)
378#define IXGBE_QV_USER_PEND (IXGBE_QV_STATE_POLL | IXGBE_QV_STATE_POLL_YIELD) 400#define IXGBE_QV_USER_PEND (IXGBE_QV_STATE_POLL | IXGBE_QV_STATE_POLL_YIELD)
379 spinlock_t lock; 401 spinlock_t lock;
@@ -394,18 +416,18 @@ static inline void ixgbe_qv_init_lock(struct ixgbe_q_vector *q_vector)
394static inline bool ixgbe_qv_lock_napi(struct ixgbe_q_vector *q_vector) 416static inline bool ixgbe_qv_lock_napi(struct ixgbe_q_vector *q_vector)
395{ 417{
396 int rc = true; 418 int rc = true;
397 spin_lock(&q_vector->lock); 419 spin_lock_bh(&q_vector->lock);
398 if (q_vector->state & IXGBE_QV_LOCKED) { 420 if (q_vector->state & IXGBE_QV_LOCKED) {
399 WARN_ON(q_vector->state & IXGBE_QV_STATE_NAPI); 421 WARN_ON(q_vector->state & IXGBE_QV_STATE_NAPI);
400 q_vector->state |= IXGBE_QV_STATE_NAPI_YIELD; 422 q_vector->state |= IXGBE_QV_STATE_NAPI_YIELD;
401 rc = false; 423 rc = false;
402#ifdef LL_EXTENDED_STATS 424#ifdef BP_EXTENDED_STATS
403 q_vector->tx.ring->stats.yields++; 425 q_vector->tx.ring->stats.yields++;
404#endif 426#endif
405 } else 427 } else
406 /* we don't care if someone yielded */ 428 /* we don't care if someone yielded */
407 q_vector->state = IXGBE_QV_STATE_NAPI; 429 q_vector->state = IXGBE_QV_STATE_NAPI;
408 spin_unlock(&q_vector->lock); 430 spin_unlock_bh(&q_vector->lock);
409 return rc; 431 return rc;
410} 432}
411 433
@@ -413,14 +435,15 @@ static inline bool ixgbe_qv_lock_napi(struct ixgbe_q_vector *q_vector)
413static inline bool ixgbe_qv_unlock_napi(struct ixgbe_q_vector *q_vector) 435static inline bool ixgbe_qv_unlock_napi(struct ixgbe_q_vector *q_vector)
414{ 436{
415 int rc = false; 437 int rc = false;
416 spin_lock(&q_vector->lock); 438 spin_lock_bh(&q_vector->lock);
417 WARN_ON(q_vector->state & (IXGBE_QV_STATE_POLL | 439 WARN_ON(q_vector->state & (IXGBE_QV_STATE_POLL |
418 IXGBE_QV_STATE_NAPI_YIELD)); 440 IXGBE_QV_STATE_NAPI_YIELD));
419 441
420 if (q_vector->state & IXGBE_QV_STATE_POLL_YIELD) 442 if (q_vector->state & IXGBE_QV_STATE_POLL_YIELD)
421 rc = true; 443 rc = true;
422 q_vector->state = IXGBE_QV_STATE_IDLE; 444 /* will reset state to idle, unless QV is disabled */
423 spin_unlock(&q_vector->lock); 445 q_vector->state &= IXGBE_QV_STATE_DISABLED;
446 spin_unlock_bh(&q_vector->lock);
424 return rc; 447 return rc;
425} 448}
426 449
@@ -432,7 +455,7 @@ static inline bool ixgbe_qv_lock_poll(struct ixgbe_q_vector *q_vector)
432 if ((q_vector->state & IXGBE_QV_LOCKED)) { 455 if ((q_vector->state & IXGBE_QV_LOCKED)) {
433 q_vector->state |= IXGBE_QV_STATE_POLL_YIELD; 456 q_vector->state |= IXGBE_QV_STATE_POLL_YIELD;
434 rc = false; 457 rc = false;
435#ifdef LL_EXTENDED_STATS 458#ifdef BP_EXTENDED_STATS
436 q_vector->rx.ring->stats.yields++; 459 q_vector->rx.ring->stats.yields++;
437#endif 460#endif
438 } else 461 } else
@@ -451,17 +474,32 @@ static inline bool ixgbe_qv_unlock_poll(struct ixgbe_q_vector *q_vector)
451 474
452 if (q_vector->state & IXGBE_QV_STATE_POLL_YIELD) 475 if (q_vector->state & IXGBE_QV_STATE_POLL_YIELD)
453 rc = true; 476 rc = true;
454 q_vector->state = IXGBE_QV_STATE_IDLE; 477 /* will reset state to idle, unless QV is disabled */
478 q_vector->state &= IXGBE_QV_STATE_DISABLED;
455 spin_unlock_bh(&q_vector->lock); 479 spin_unlock_bh(&q_vector->lock);
456 return rc; 480 return rc;
457} 481}
458 482
459/* true if a socket is polling, even if it did not get the lock */ 483/* true if a socket is polling, even if it did not get the lock */
460static inline bool ixgbe_qv_ll_polling(struct ixgbe_q_vector *q_vector) 484static inline bool ixgbe_qv_busy_polling(struct ixgbe_q_vector *q_vector)
461{ 485{
462 WARN_ON(!(q_vector->state & IXGBE_QV_LOCKED)); 486 WARN_ON(!(q_vector->state & IXGBE_QV_OWNED));
463 return q_vector->state & IXGBE_QV_USER_PEND; 487 return q_vector->state & IXGBE_QV_USER_PEND;
464} 488}
489
490/* false if QV is currently owned */
491static inline bool ixgbe_qv_disable(struct ixgbe_q_vector *q_vector)
492{
493 int rc = true;
494 spin_lock_bh(&q_vector->lock);
495 if (q_vector->state & IXGBE_QV_OWNED)
496 rc = false;
497 q_vector->state |= IXGBE_QV_STATE_DISABLED;
498 spin_unlock_bh(&q_vector->lock);
499
500 return rc;
501}
502
465#else /* CONFIG_NET_RX_BUSY_POLL */ 503#else /* CONFIG_NET_RX_BUSY_POLL */
466static inline void ixgbe_qv_init_lock(struct ixgbe_q_vector *q_vector) 504static inline void ixgbe_qv_init_lock(struct ixgbe_q_vector *q_vector)
467{ 505{
@@ -487,10 +525,16 @@ static inline bool ixgbe_qv_unlock_poll(struct ixgbe_q_vector *q_vector)
487 return false; 525 return false;
488} 526}
489 527
490static inline bool ixgbe_qv_ll_polling(struct ixgbe_q_vector *q_vector) 528static inline bool ixgbe_qv_busy_polling(struct ixgbe_q_vector *q_vector)
491{ 529{
492 return false; 530 return false;
493} 531}
532
533static inline bool ixgbe_qv_disable(struct ixgbe_q_vector *q_vector)
534{
535 return true;
536}
537
494#endif /* CONFIG_NET_RX_BUSY_POLL */ 538#endif /* CONFIG_NET_RX_BUSY_POLL */
495 539
496#ifdef CONFIG_IXGBE_HWMON 540#ifdef CONFIG_IXGBE_HWMON
@@ -738,6 +782,7 @@ struct ixgbe_adapter {
738#endif /*CONFIG_DEBUG_FS*/ 782#endif /*CONFIG_DEBUG_FS*/
739 783
740 u8 default_up; 784 u8 default_up;
785 unsigned long fwd_bitmask; /* Bitmask indicating in use pools */
741}; 786};
742 787
743struct ixgbe_fdir_filter { 788struct ixgbe_fdir_filter {
@@ -786,93 +831,89 @@ extern const char ixgbe_driver_version[];
786extern char ixgbe_default_device_descr[]; 831extern char ixgbe_default_device_descr[];
787#endif /* IXGBE_FCOE */ 832#endif /* IXGBE_FCOE */
788 833
789extern void ixgbe_up(struct ixgbe_adapter *adapter); 834void ixgbe_up(struct ixgbe_adapter *adapter);
790extern void ixgbe_down(struct ixgbe_adapter *adapter); 835void ixgbe_down(struct ixgbe_adapter *adapter);
791extern void ixgbe_reinit_locked(struct ixgbe_adapter *adapter); 836void ixgbe_reinit_locked(struct ixgbe_adapter *adapter);
792extern void ixgbe_reset(struct ixgbe_adapter *adapter); 837void ixgbe_reset(struct ixgbe_adapter *adapter);
793extern void ixgbe_set_ethtool_ops(struct net_device *netdev); 838void ixgbe_set_ethtool_ops(struct net_device *netdev);
794extern int ixgbe_setup_rx_resources(struct ixgbe_ring *); 839int ixgbe_setup_rx_resources(struct ixgbe_ring *);
795extern int ixgbe_setup_tx_resources(struct ixgbe_ring *); 840int ixgbe_setup_tx_resources(struct ixgbe_ring *);
796extern void ixgbe_free_rx_resources(struct ixgbe_ring *); 841void ixgbe_free_rx_resources(struct ixgbe_ring *);
797extern void ixgbe_free_tx_resources(struct ixgbe_ring *); 842void ixgbe_free_tx_resources(struct ixgbe_ring *);
798extern void ixgbe_configure_rx_ring(struct ixgbe_adapter *,struct ixgbe_ring *); 843void ixgbe_configure_rx_ring(struct ixgbe_adapter *, struct ixgbe_ring *);
799extern void ixgbe_configure_tx_ring(struct ixgbe_adapter *,struct ixgbe_ring *); 844void ixgbe_configure_tx_ring(struct ixgbe_adapter *, struct ixgbe_ring *);
800extern void ixgbe_disable_rx_queue(struct ixgbe_adapter *adapter, 845void ixgbe_disable_rx_queue(struct ixgbe_adapter *adapter, struct ixgbe_ring *);
801 struct ixgbe_ring *); 846void ixgbe_update_stats(struct ixgbe_adapter *adapter);
802extern void ixgbe_update_stats(struct ixgbe_adapter *adapter); 847int ixgbe_init_interrupt_scheme(struct ixgbe_adapter *adapter);
803extern int ixgbe_init_interrupt_scheme(struct ixgbe_adapter *adapter); 848int ixgbe_wol_supported(struct ixgbe_adapter *adapter, u16 device_id,
804extern int ixgbe_wol_supported(struct ixgbe_adapter *adapter, u16 device_id,
805 u16 subdevice_id); 849 u16 subdevice_id);
806extern void ixgbe_clear_interrupt_scheme(struct ixgbe_adapter *adapter); 850void ixgbe_clear_interrupt_scheme(struct ixgbe_adapter *adapter);
807extern netdev_tx_t ixgbe_xmit_frame_ring(struct sk_buff *, 851netdev_tx_t ixgbe_xmit_frame_ring(struct sk_buff *, struct ixgbe_adapter *,
808 struct ixgbe_adapter *, 852 struct ixgbe_ring *);
809 struct ixgbe_ring *); 853void ixgbe_unmap_and_free_tx_resource(struct ixgbe_ring *,
810extern void ixgbe_unmap_and_free_tx_resource(struct ixgbe_ring *, 854 struct ixgbe_tx_buffer *);
811 struct ixgbe_tx_buffer *); 855void ixgbe_alloc_rx_buffers(struct ixgbe_ring *, u16);
812extern void ixgbe_alloc_rx_buffers(struct ixgbe_ring *, u16); 856void ixgbe_write_eitr(struct ixgbe_q_vector *);
813extern void ixgbe_write_eitr(struct ixgbe_q_vector *); 857int ixgbe_poll(struct napi_struct *napi, int budget);
814extern int ixgbe_poll(struct napi_struct *napi, int budget); 858int ethtool_ioctl(struct ifreq *ifr);
815extern int ethtool_ioctl(struct ifreq *ifr); 859s32 ixgbe_reinit_fdir_tables_82599(struct ixgbe_hw *hw);
816extern s32 ixgbe_reinit_fdir_tables_82599(struct ixgbe_hw *hw); 860s32 ixgbe_init_fdir_signature_82599(struct ixgbe_hw *hw, u32 fdirctrl);
817extern s32 ixgbe_init_fdir_signature_82599(struct ixgbe_hw *hw, u32 fdirctrl); 861s32 ixgbe_init_fdir_perfect_82599(struct ixgbe_hw *hw, u32 fdirctrl);
818extern s32 ixgbe_init_fdir_perfect_82599(struct ixgbe_hw *hw, u32 fdirctrl); 862s32 ixgbe_fdir_add_signature_filter_82599(struct ixgbe_hw *hw,
819extern s32 ixgbe_fdir_add_signature_filter_82599(struct ixgbe_hw *hw, 863 union ixgbe_atr_hash_dword input,
820 union ixgbe_atr_hash_dword input, 864 union ixgbe_atr_hash_dword common,
821 union ixgbe_atr_hash_dword common, 865 u8 queue);
822 u8 queue); 866s32 ixgbe_fdir_set_input_mask_82599(struct ixgbe_hw *hw,
823extern s32 ixgbe_fdir_set_input_mask_82599(struct ixgbe_hw *hw, 867 union ixgbe_atr_input *input_mask);
824 union ixgbe_atr_input *input_mask); 868s32 ixgbe_fdir_write_perfect_filter_82599(struct ixgbe_hw *hw,
825extern s32 ixgbe_fdir_write_perfect_filter_82599(struct ixgbe_hw *hw, 869 union ixgbe_atr_input *input,
826 union ixgbe_atr_input *input, 870 u16 soft_id, u8 queue);
827 u16 soft_id, u8 queue); 871s32 ixgbe_fdir_erase_perfect_filter_82599(struct ixgbe_hw *hw,
828extern s32 ixgbe_fdir_erase_perfect_filter_82599(struct ixgbe_hw *hw, 872 union ixgbe_atr_input *input,
829 union ixgbe_atr_input *input, 873 u16 soft_id);
830 u16 soft_id); 874void ixgbe_atr_compute_perfect_hash_82599(union ixgbe_atr_input *input,
831extern void ixgbe_atr_compute_perfect_hash_82599(union ixgbe_atr_input *input, 875 union ixgbe_atr_input *mask);
832 union ixgbe_atr_input *mask); 876bool ixgbe_verify_lesm_fw_enabled_82599(struct ixgbe_hw *hw);
833extern bool ixgbe_verify_lesm_fw_enabled_82599(struct ixgbe_hw *hw); 877void ixgbe_set_rx_mode(struct net_device *netdev);
834extern void ixgbe_set_rx_mode(struct net_device *netdev);
835#ifdef CONFIG_IXGBE_DCB 878#ifdef CONFIG_IXGBE_DCB
836extern void ixgbe_set_rx_drop_en(struct ixgbe_adapter *adapter); 879void ixgbe_set_rx_drop_en(struct ixgbe_adapter *adapter);
837#endif 880#endif
838extern int ixgbe_setup_tc(struct net_device *dev, u8 tc); 881int ixgbe_setup_tc(struct net_device *dev, u8 tc);
839extern void ixgbe_tx_ctxtdesc(struct ixgbe_ring *, u32, u32, u32, u32); 882void ixgbe_tx_ctxtdesc(struct ixgbe_ring *, u32, u32, u32, u32);
840extern void ixgbe_do_reset(struct net_device *netdev); 883void ixgbe_do_reset(struct net_device *netdev);
841#ifdef CONFIG_IXGBE_HWMON 884#ifdef CONFIG_IXGBE_HWMON
842extern void ixgbe_sysfs_exit(struct ixgbe_adapter *adapter); 885void ixgbe_sysfs_exit(struct ixgbe_adapter *adapter);
843extern int ixgbe_sysfs_init(struct ixgbe_adapter *adapter); 886int ixgbe_sysfs_init(struct ixgbe_adapter *adapter);
844#endif /* CONFIG_IXGBE_HWMON */ 887#endif /* CONFIG_IXGBE_HWMON */
845#ifdef IXGBE_FCOE 888#ifdef IXGBE_FCOE
846extern void ixgbe_configure_fcoe(struct ixgbe_adapter *adapter); 889void ixgbe_configure_fcoe(struct ixgbe_adapter *adapter);
847extern int ixgbe_fso(struct ixgbe_ring *tx_ring, 890int ixgbe_fso(struct ixgbe_ring *tx_ring, struct ixgbe_tx_buffer *first,
848 struct ixgbe_tx_buffer *first, 891 u8 *hdr_len);
849 u8 *hdr_len); 892int ixgbe_fcoe_ddp(struct ixgbe_adapter *adapter,
850extern int ixgbe_fcoe_ddp(struct ixgbe_adapter *adapter, 893 union ixgbe_adv_rx_desc *rx_desc, struct sk_buff *skb);
851 union ixgbe_adv_rx_desc *rx_desc, 894int ixgbe_fcoe_ddp_get(struct net_device *netdev, u16 xid,
852 struct sk_buff *skb); 895 struct scatterlist *sgl, unsigned int sgc);
853extern int ixgbe_fcoe_ddp_get(struct net_device *netdev, u16 xid, 896int ixgbe_fcoe_ddp_target(struct net_device *netdev, u16 xid,
854 struct scatterlist *sgl, unsigned int sgc); 897 struct scatterlist *sgl, unsigned int sgc);
855extern int ixgbe_fcoe_ddp_target(struct net_device *netdev, u16 xid, 898int ixgbe_fcoe_ddp_put(struct net_device *netdev, u16 xid);
856 struct scatterlist *sgl, unsigned int sgc); 899int ixgbe_setup_fcoe_ddp_resources(struct ixgbe_adapter *adapter);
857extern int ixgbe_fcoe_ddp_put(struct net_device *netdev, u16 xid); 900void ixgbe_free_fcoe_ddp_resources(struct ixgbe_adapter *adapter);
858extern int ixgbe_setup_fcoe_ddp_resources(struct ixgbe_adapter *adapter); 901int ixgbe_fcoe_enable(struct net_device *netdev);
859extern void ixgbe_free_fcoe_ddp_resources(struct ixgbe_adapter *adapter); 902int ixgbe_fcoe_disable(struct net_device *netdev);
860extern int ixgbe_fcoe_enable(struct net_device *netdev);
861extern int ixgbe_fcoe_disable(struct net_device *netdev);
862#ifdef CONFIG_IXGBE_DCB 903#ifdef CONFIG_IXGBE_DCB
863extern u8 ixgbe_fcoe_getapp(struct ixgbe_adapter *adapter); 904u8 ixgbe_fcoe_getapp(struct ixgbe_adapter *adapter);
864extern u8 ixgbe_fcoe_setapp(struct ixgbe_adapter *adapter, u8 up); 905u8 ixgbe_fcoe_setapp(struct ixgbe_adapter *adapter, u8 up);
865#endif /* CONFIG_IXGBE_DCB */ 906#endif /* CONFIG_IXGBE_DCB */
866extern int ixgbe_fcoe_get_wwn(struct net_device *netdev, u64 *wwn, int type); 907int ixgbe_fcoe_get_wwn(struct net_device *netdev, u64 *wwn, int type);
867extern int ixgbe_fcoe_get_hbainfo(struct net_device *netdev, 908int ixgbe_fcoe_get_hbainfo(struct net_device *netdev,
868 struct netdev_fcoe_hbainfo *info); 909 struct netdev_fcoe_hbainfo *info);
869extern u8 ixgbe_fcoe_get_tc(struct ixgbe_adapter *adapter); 910u8 ixgbe_fcoe_get_tc(struct ixgbe_adapter *adapter);
870#endif /* IXGBE_FCOE */ 911#endif /* IXGBE_FCOE */
871#ifdef CONFIG_DEBUG_FS 912#ifdef CONFIG_DEBUG_FS
872extern void ixgbe_dbg_adapter_init(struct ixgbe_adapter *adapter); 913void ixgbe_dbg_adapter_init(struct ixgbe_adapter *adapter);
873extern void ixgbe_dbg_adapter_exit(struct ixgbe_adapter *adapter); 914void ixgbe_dbg_adapter_exit(struct ixgbe_adapter *adapter);
874extern void ixgbe_dbg_init(void); 915void ixgbe_dbg_init(void);
875extern void ixgbe_dbg_exit(void); 916void ixgbe_dbg_exit(void);
876#else 917#else
877static inline void ixgbe_dbg_adapter_init(struct ixgbe_adapter *adapter) {} 918static inline void ixgbe_dbg_adapter_init(struct ixgbe_adapter *adapter) {}
878static inline void ixgbe_dbg_adapter_exit(struct ixgbe_adapter *adapter) {} 919static inline void ixgbe_dbg_adapter_exit(struct ixgbe_adapter *adapter) {}
@@ -884,12 +925,12 @@ static inline struct netdev_queue *txring_txq(const struct ixgbe_ring *ring)
884 return netdev_get_tx_queue(ring->netdev, ring->queue_index); 925 return netdev_get_tx_queue(ring->netdev, ring->queue_index);
885} 926}
886 927
887extern void ixgbe_ptp_init(struct ixgbe_adapter *adapter); 928void ixgbe_ptp_init(struct ixgbe_adapter *adapter);
888extern void ixgbe_ptp_stop(struct ixgbe_adapter *adapter); 929void ixgbe_ptp_stop(struct ixgbe_adapter *adapter);
889extern void ixgbe_ptp_overflow_check(struct ixgbe_adapter *adapter); 930void ixgbe_ptp_overflow_check(struct ixgbe_adapter *adapter);
890extern void ixgbe_ptp_rx_hang(struct ixgbe_adapter *adapter); 931void ixgbe_ptp_rx_hang(struct ixgbe_adapter *adapter);
891extern void __ixgbe_ptp_rx_hwtstamp(struct ixgbe_q_vector *q_vector, 932void __ixgbe_ptp_rx_hwtstamp(struct ixgbe_q_vector *q_vector,
892 struct sk_buff *skb); 933 struct sk_buff *skb);
893static inline void ixgbe_ptp_rx_hwtstamp(struct ixgbe_ring *rx_ring, 934static inline void ixgbe_ptp_rx_hwtstamp(struct ixgbe_ring *rx_ring,
894 union ixgbe_adv_rx_desc *rx_desc, 935 union ixgbe_adv_rx_desc *rx_desc,
895 struct sk_buff *skb) 936 struct sk_buff *skb)
@@ -906,13 +947,16 @@ static inline void ixgbe_ptp_rx_hwtstamp(struct ixgbe_ring *rx_ring,
906 rx_ring->last_rx_timestamp = jiffies; 947 rx_ring->last_rx_timestamp = jiffies;
907} 948}
908 949
909extern int ixgbe_ptp_hwtstamp_ioctl(struct ixgbe_adapter *adapter, 950int ixgbe_ptp_hwtstamp_ioctl(struct ixgbe_adapter *adapter, struct ifreq *ifr,
910 struct ifreq *ifr, int cmd); 951 int cmd);
911extern void ixgbe_ptp_start_cyclecounter(struct ixgbe_adapter *adapter); 952void ixgbe_ptp_start_cyclecounter(struct ixgbe_adapter *adapter);
912extern void ixgbe_ptp_reset(struct ixgbe_adapter *adapter); 953void ixgbe_ptp_reset(struct ixgbe_adapter *adapter);
913extern void ixgbe_ptp_check_pps_event(struct ixgbe_adapter *adapter, u32 eicr); 954void ixgbe_ptp_check_pps_event(struct ixgbe_adapter *adapter, u32 eicr);
914#ifdef CONFIG_PCI_IOV 955#ifdef CONFIG_PCI_IOV
915void ixgbe_sriov_reinit(struct ixgbe_adapter *adapter); 956void ixgbe_sriov_reinit(struct ixgbe_adapter *adapter);
916#endif 957#endif
917 958
959netdev_tx_t ixgbe_xmit_frame_ring(struct sk_buff *skb,
960 struct ixgbe_adapter *adapter,
961 struct ixgbe_ring *tx_ring);
918#endif /* _IXGBE_H_ */ 962#endif /* _IXGBE_H_ */
diff --git a/drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c b/drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c
index e8649abf97c0..4e7c9b098b58 100644
--- a/drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c
+++ b/drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c
@@ -442,7 +442,7 @@ static void ixgbe_set_msglevel(struct net_device *netdev, u32 data)
442 442
443static int ixgbe_get_regs_len(struct net_device *netdev) 443static int ixgbe_get_regs_len(struct net_device *netdev)
444{ 444{
445#define IXGBE_REGS_LEN 1129 445#define IXGBE_REGS_LEN 1139
446 return IXGBE_REGS_LEN * sizeof(u32); 446 return IXGBE_REGS_LEN * sizeof(u32);
447} 447}
448 448
@@ -602,22 +602,53 @@ static void ixgbe_get_regs(struct net_device *netdev,
602 regs_buff[828] = IXGBE_READ_REG(hw, IXGBE_FHFT(0)); 602 regs_buff[828] = IXGBE_READ_REG(hw, IXGBE_FHFT(0));
603 603
604 /* DCB */ 604 /* DCB */
605 regs_buff[829] = IXGBE_READ_REG(hw, IXGBE_RMCS); 605 regs_buff[829] = IXGBE_READ_REG(hw, IXGBE_RMCS); /* same as FCCFG */
606 regs_buff[830] = IXGBE_READ_REG(hw, IXGBE_DPMCS); 606 regs_buff[831] = IXGBE_READ_REG(hw, IXGBE_PDPMCS); /* same as RTTPCS */
607 regs_buff[831] = IXGBE_READ_REG(hw, IXGBE_PDPMCS); 607
608 regs_buff[832] = IXGBE_READ_REG(hw, IXGBE_RUPPBMR); 608 switch (hw->mac.type) {
609 for (i = 0; i < 8; i++) 609 case ixgbe_mac_82598EB:
610 regs_buff[833 + i] = IXGBE_READ_REG(hw, IXGBE_RT2CR(i)); 610 regs_buff[830] = IXGBE_READ_REG(hw, IXGBE_DPMCS);
611 for (i = 0; i < 8; i++) 611 regs_buff[832] = IXGBE_READ_REG(hw, IXGBE_RUPPBMR);
612 regs_buff[841 + i] = IXGBE_READ_REG(hw, IXGBE_RT2SR(i)); 612 for (i = 0; i < 8; i++)
613 for (i = 0; i < 8; i++) 613 regs_buff[833 + i] =
614 regs_buff[849 + i] = IXGBE_READ_REG(hw, IXGBE_TDTQ2TCCR(i)); 614 IXGBE_READ_REG(hw, IXGBE_RT2CR(i));
615 for (i = 0; i < 8; i++) 615 for (i = 0; i < 8; i++)
616 regs_buff[857 + i] = IXGBE_READ_REG(hw, IXGBE_TDTQ2TCSR(i)); 616 regs_buff[841 + i] =
617 IXGBE_READ_REG(hw, IXGBE_RT2SR(i));
618 for (i = 0; i < 8; i++)
619 regs_buff[849 + i] =
620 IXGBE_READ_REG(hw, IXGBE_TDTQ2TCCR(i));
621 for (i = 0; i < 8; i++)
622 regs_buff[857 + i] =
623 IXGBE_READ_REG(hw, IXGBE_TDTQ2TCSR(i));
624 break;
625 case ixgbe_mac_82599EB:
626 case ixgbe_mac_X540:
627 regs_buff[830] = IXGBE_READ_REG(hw, IXGBE_RTTDCS);
628 regs_buff[832] = IXGBE_READ_REG(hw, IXGBE_RTRPCS);
629 for (i = 0; i < 8; i++)
630 regs_buff[833 + i] =
631 IXGBE_READ_REG(hw, IXGBE_RTRPT4C(i));
632 for (i = 0; i < 8; i++)
633 regs_buff[841 + i] =
634 IXGBE_READ_REG(hw, IXGBE_RTRPT4S(i));
635 for (i = 0; i < 8; i++)
636 regs_buff[849 + i] =
637 IXGBE_READ_REG(hw, IXGBE_RTTDT2C(i));
638 for (i = 0; i < 8; i++)
639 regs_buff[857 + i] =
640 IXGBE_READ_REG(hw, IXGBE_RTTDT2S(i));
641 break;
642 default:
643 break;
644 }
645
617 for (i = 0; i < 8; i++) 646 for (i = 0; i < 8; i++)
618 regs_buff[865 + i] = IXGBE_READ_REG(hw, IXGBE_TDPT2TCCR(i)); 647 regs_buff[865 + i] =
648 IXGBE_READ_REG(hw, IXGBE_TDPT2TCCR(i)); /* same as RTTPT2C */
619 for (i = 0; i < 8; i++) 649 for (i = 0; i < 8; i++)
620 regs_buff[873 + i] = IXGBE_READ_REG(hw, IXGBE_TDPT2TCSR(i)); 650 regs_buff[873 + i] =
651 IXGBE_READ_REG(hw, IXGBE_TDPT2TCSR(i)); /* same as RTTPT2S */
621 652
622 /* Statistics */ 653 /* Statistics */
623 regs_buff[881] = IXGBE_GET_STAT(adapter, crcerrs); 654 regs_buff[881] = IXGBE_GET_STAT(adapter, crcerrs);
@@ -757,6 +788,20 @@ static void ixgbe_get_regs(struct net_device *netdev,
757 788
758 /* 82599 X540 specific registers */ 789 /* 82599 X540 specific registers */
759 regs_buff[1128] = IXGBE_READ_REG(hw, IXGBE_MFLCN); 790 regs_buff[1128] = IXGBE_READ_REG(hw, IXGBE_MFLCN);
791
792 /* 82599 X540 specific DCB registers */
793 regs_buff[1129] = IXGBE_READ_REG(hw, IXGBE_RTRUP2TC);
794 regs_buff[1130] = IXGBE_READ_REG(hw, IXGBE_RTTUP2TC);
795 for (i = 0; i < 4; i++)
796 regs_buff[1131 + i] = IXGBE_READ_REG(hw, IXGBE_TXLLQ(i));
797 regs_buff[1135] = IXGBE_READ_REG(hw, IXGBE_RTTBCNRM);
798 /* same as RTTQCNRM */
799 regs_buff[1136] = IXGBE_READ_REG(hw, IXGBE_RTTBCNRD);
800 /* same as RTTQCNRR */
801
802 /* X540 specific DCB registers */
803 regs_buff[1137] = IXGBE_READ_REG(hw, IXGBE_RTTQCNCR);
804 regs_buff[1138] = IXGBE_READ_REG(hw, IXGBE_RTTQCNTG);
760} 805}
761 806
762static int ixgbe_get_eeprom_len(struct net_device *netdev) 807static int ixgbe_get_eeprom_len(struct net_device *netdev)
@@ -1072,7 +1117,7 @@ static void ixgbe_get_ethtool_stats(struct net_device *netdev,
1072 data[i] = 0; 1117 data[i] = 0;
1073 data[i+1] = 0; 1118 data[i+1] = 0;
1074 i += 2; 1119 i += 2;
1075#ifdef LL_EXTENDED_STATS 1120#ifdef BP_EXTENDED_STATS
1076 data[i] = 0; 1121 data[i] = 0;
1077 data[i+1] = 0; 1122 data[i+1] = 0;
1078 data[i+2] = 0; 1123 data[i+2] = 0;
@@ -1087,7 +1132,7 @@ static void ixgbe_get_ethtool_stats(struct net_device *netdev,
1087 data[i+1] = ring->stats.bytes; 1132 data[i+1] = ring->stats.bytes;
1088 } while (u64_stats_fetch_retry_bh(&ring->syncp, start)); 1133 } while (u64_stats_fetch_retry_bh(&ring->syncp, start));
1089 i += 2; 1134 i += 2;
1090#ifdef LL_EXTENDED_STATS 1135#ifdef BP_EXTENDED_STATS
1091 data[i] = ring->stats.yields; 1136 data[i] = ring->stats.yields;
1092 data[i+1] = ring->stats.misses; 1137 data[i+1] = ring->stats.misses;
1093 data[i+2] = ring->stats.cleaned; 1138 data[i+2] = ring->stats.cleaned;
@@ -1100,7 +1145,7 @@ static void ixgbe_get_ethtool_stats(struct net_device *netdev,
1100 data[i] = 0; 1145 data[i] = 0;
1101 data[i+1] = 0; 1146 data[i+1] = 0;
1102 i += 2; 1147 i += 2;
1103#ifdef LL_EXTENDED_STATS 1148#ifdef BP_EXTENDED_STATS
1104 data[i] = 0; 1149 data[i] = 0;
1105 data[i+1] = 0; 1150 data[i+1] = 0;
1106 data[i+2] = 0; 1151 data[i+2] = 0;
@@ -1115,7 +1160,7 @@ static void ixgbe_get_ethtool_stats(struct net_device *netdev,
1115 data[i+1] = ring->stats.bytes; 1160 data[i+1] = ring->stats.bytes;
1116 } while (u64_stats_fetch_retry_bh(&ring->syncp, start)); 1161 } while (u64_stats_fetch_retry_bh(&ring->syncp, start));
1117 i += 2; 1162 i += 2;
1118#ifdef LL_EXTENDED_STATS 1163#ifdef BP_EXTENDED_STATS
1119 data[i] = ring->stats.yields; 1164 data[i] = ring->stats.yields;
1120 data[i+1] = ring->stats.misses; 1165 data[i+1] = ring->stats.misses;
1121 data[i+2] = ring->stats.cleaned; 1166 data[i+2] = ring->stats.cleaned;
@@ -1157,28 +1202,28 @@ static void ixgbe_get_strings(struct net_device *netdev, u32 stringset,
1157 p += ETH_GSTRING_LEN; 1202 p += ETH_GSTRING_LEN;
1158 sprintf(p, "tx_queue_%u_bytes", i); 1203 sprintf(p, "tx_queue_%u_bytes", i);
1159 p += ETH_GSTRING_LEN; 1204 p += ETH_GSTRING_LEN;
1160#ifdef LL_EXTENDED_STATS 1205#ifdef BP_EXTENDED_STATS
1161 sprintf(p, "tx_queue_%u_ll_napi_yield", i); 1206 sprintf(p, "tx_queue_%u_bp_napi_yield", i);
1162 p += ETH_GSTRING_LEN; 1207 p += ETH_GSTRING_LEN;
1163 sprintf(p, "tx_queue_%u_ll_misses", i); 1208 sprintf(p, "tx_queue_%u_bp_misses", i);
1164 p += ETH_GSTRING_LEN; 1209 p += ETH_GSTRING_LEN;
1165 sprintf(p, "tx_queue_%u_ll_cleaned", i); 1210 sprintf(p, "tx_queue_%u_bp_cleaned", i);
1166 p += ETH_GSTRING_LEN; 1211 p += ETH_GSTRING_LEN;
1167#endif /* LL_EXTENDED_STATS */ 1212#endif /* BP_EXTENDED_STATS */
1168 } 1213 }
1169 for (i = 0; i < IXGBE_NUM_RX_QUEUES; i++) { 1214 for (i = 0; i < IXGBE_NUM_RX_QUEUES; i++) {
1170 sprintf(p, "rx_queue_%u_packets", i); 1215 sprintf(p, "rx_queue_%u_packets", i);
1171 p += ETH_GSTRING_LEN; 1216 p += ETH_GSTRING_LEN;
1172 sprintf(p, "rx_queue_%u_bytes", i); 1217 sprintf(p, "rx_queue_%u_bytes", i);
1173 p += ETH_GSTRING_LEN; 1218 p += ETH_GSTRING_LEN;
1174#ifdef LL_EXTENDED_STATS 1219#ifdef BP_EXTENDED_STATS
1175 sprintf(p, "rx_queue_%u_ll_poll_yield", i); 1220 sprintf(p, "rx_queue_%u_bp_poll_yield", i);
1176 p += ETH_GSTRING_LEN; 1221 p += ETH_GSTRING_LEN;
1177 sprintf(p, "rx_queue_%u_ll_misses", i); 1222 sprintf(p, "rx_queue_%u_bp_misses", i);
1178 p += ETH_GSTRING_LEN; 1223 p += ETH_GSTRING_LEN;
1179 sprintf(p, "rx_queue_%u_ll_cleaned", i); 1224 sprintf(p, "rx_queue_%u_bp_cleaned", i);
1180 p += ETH_GSTRING_LEN; 1225 p += ETH_GSTRING_LEN;
1181#endif /* LL_EXTENDED_STATS */ 1226#endif /* BP_EXTENDED_STATS */
1182 } 1227 }
1183 for (i = 0; i < IXGBE_MAX_PACKET_BUFFERS; i++) { 1228 for (i = 0; i < IXGBE_MAX_PACKET_BUFFERS; i++) {
1184 sprintf(p, "tx_pb_%u_pxon", i); 1229 sprintf(p, "tx_pb_%u_pxon", i);
@@ -2212,13 +2257,13 @@ static int ixgbe_set_coalesce(struct net_device *netdev,
2212 2257
2213#if IS_ENABLED(CONFIG_BQL) 2258#if IS_ENABLED(CONFIG_BQL)
2214 /* detect ITR changes that require update of TXDCTL.WTHRESH */ 2259 /* detect ITR changes that require update of TXDCTL.WTHRESH */
2215 if ((adapter->tx_itr_setting > 1) && 2260 if ((adapter->tx_itr_setting != 1) &&
2216 (adapter->tx_itr_setting < IXGBE_100K_ITR)) { 2261 (adapter->tx_itr_setting < IXGBE_100K_ITR)) {
2217 if ((tx_itr_prev == 1) || 2262 if ((tx_itr_prev == 1) ||
2218 (tx_itr_prev > IXGBE_100K_ITR)) 2263 (tx_itr_prev >= IXGBE_100K_ITR))
2219 need_reset = true; 2264 need_reset = true;
2220 } else { 2265 } else {
2221 if ((tx_itr_prev > 1) && 2266 if ((tx_itr_prev != 1) &&
2222 (tx_itr_prev < IXGBE_100K_ITR)) 2267 (tx_itr_prev < IXGBE_100K_ITR))
2223 need_reset = true; 2268 need_reset = true;
2224 } 2269 }
diff --git a/drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c b/drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
index 90b4e1089ecc..32e3eaaa160a 100644
--- a/drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
+++ b/drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
@@ -498,6 +498,7 @@ static bool ixgbe_set_sriov_queues(struct ixgbe_adapter *adapter)
498#ifdef IXGBE_FCOE 498#ifdef IXGBE_FCOE
499 u16 fcoe_i = 0; 499 u16 fcoe_i = 0;
500#endif 500#endif
501 bool pools = (find_first_zero_bit(&adapter->fwd_bitmask, 32) > 1);
501 502
502 /* only proceed if SR-IOV is enabled */ 503 /* only proceed if SR-IOV is enabled */
503 if (!(adapter->flags & IXGBE_FLAG_SRIOV_ENABLED)) 504 if (!(adapter->flags & IXGBE_FLAG_SRIOV_ENABLED))
@@ -510,7 +511,7 @@ static bool ixgbe_set_sriov_queues(struct ixgbe_adapter *adapter)
510 vmdq_i = min_t(u16, IXGBE_MAX_VMDQ_INDICES, vmdq_i); 511 vmdq_i = min_t(u16, IXGBE_MAX_VMDQ_INDICES, vmdq_i);
511 512
512 /* 64 pool mode with 2 queues per pool */ 513 /* 64 pool mode with 2 queues per pool */
513 if ((vmdq_i > 32) || (rss_i < 4)) { 514 if ((vmdq_i > 32) || (rss_i < 4) || (vmdq_i > 16 && pools)) {
514 vmdq_m = IXGBE_82599_VMDQ_2Q_MASK; 515 vmdq_m = IXGBE_82599_VMDQ_2Q_MASK;
515 rss_m = IXGBE_RSS_2Q_MASK; 516 rss_m = IXGBE_RSS_2Q_MASK;
516 rss_i = min_t(u16, rss_i, 2); 517 rss_i = min_t(u16, rss_i, 2);
@@ -852,7 +853,11 @@ static int ixgbe_alloc_q_vector(struct ixgbe_adapter *adapter,
852 853
853 /* apply Tx specific ring traits */ 854 /* apply Tx specific ring traits */
854 ring->count = adapter->tx_ring_count; 855 ring->count = adapter->tx_ring_count;
855 ring->queue_index = txr_idx; 856 if (adapter->num_rx_pools > 1)
857 ring->queue_index =
858 txr_idx % adapter->num_rx_queues_per_pool;
859 else
860 ring->queue_index = txr_idx;
856 861
857 /* assign ring to adapter */ 862 /* assign ring to adapter */
858 adapter->tx_ring[txr_idx] = ring; 863 adapter->tx_ring[txr_idx] = ring;
@@ -895,7 +900,11 @@ static int ixgbe_alloc_q_vector(struct ixgbe_adapter *adapter,
895#endif /* IXGBE_FCOE */ 900#endif /* IXGBE_FCOE */
896 /* apply Rx specific ring traits */ 901 /* apply Rx specific ring traits */
897 ring->count = adapter->rx_ring_count; 902 ring->count = adapter->rx_ring_count;
898 ring->queue_index = rxr_idx; 903 if (adapter->num_rx_pools > 1)
904 ring->queue_index =
905 rxr_idx % adapter->num_rx_queues_per_pool;
906 else
907 ring->queue_index = rxr_idx;
899 908
900 /* assign ring to adapter */ 909 /* assign ring to adapter */
901 adapter->rx_ring[rxr_idx] = ring; 910 adapter->rx_ring[rxr_idx] = ring;
diff --git a/drivers/net/ethernet/intel/ixgbe/ixgbe_main.c b/drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
index 0ade0cd5ef53..bd8f5239dfe6 100644
--- a/drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
+++ b/drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
@@ -44,6 +44,7 @@
44#include <linux/ethtool.h> 44#include <linux/ethtool.h>
45#include <linux/if.h> 45#include <linux/if.h>
46#include <linux/if_vlan.h> 46#include <linux/if_vlan.h>
47#include <linux/if_macvlan.h>
47#include <linux/if_bridge.h> 48#include <linux/if_bridge.h>
48#include <linux/prefetch.h> 49#include <linux/prefetch.h>
49#include <scsi/fc/fc_fcoe.h> 50#include <scsi/fc/fc_fcoe.h>
@@ -132,7 +133,7 @@ static struct notifier_block dca_notifier = {
132static unsigned int max_vfs; 133static unsigned int max_vfs;
133module_param(max_vfs, uint, 0); 134module_param(max_vfs, uint, 0);
134MODULE_PARM_DESC(max_vfs, 135MODULE_PARM_DESC(max_vfs,
135 "Maximum number of virtual functions to allocate per physical function - default is zero and maximum value is 63"); 136 "Maximum number of virtual functions to allocate per physical function - default is zero and maximum value is 63. (Deprecated)");
136#endif /* CONFIG_PCI_IOV */ 137#endif /* CONFIG_PCI_IOV */
137 138
138static unsigned int allow_unsupported_sfp; 139static unsigned int allow_unsupported_sfp;
@@ -153,7 +154,6 @@ MODULE_VERSION(DRV_VERSION);
153static int ixgbe_read_pci_cfg_word_parent(struct ixgbe_adapter *adapter, 154static int ixgbe_read_pci_cfg_word_parent(struct ixgbe_adapter *adapter,
154 u32 reg, u16 *value) 155 u32 reg, u16 *value)
155{ 156{
156 int pos = 0;
157 struct pci_dev *parent_dev; 157 struct pci_dev *parent_dev;
158 struct pci_bus *parent_bus; 158 struct pci_bus *parent_bus;
159 159
@@ -165,11 +165,10 @@ static int ixgbe_read_pci_cfg_word_parent(struct ixgbe_adapter *adapter,
165 if (!parent_dev) 165 if (!parent_dev)
166 return -1; 166 return -1;
167 167
168 pos = pci_find_capability(parent_dev, PCI_CAP_ID_EXP); 168 if (!pci_is_pcie(parent_dev))
169 if (!pos)
170 return -1; 169 return -1;
171 170
172 pci_read_config_word(parent_dev, pos + reg, value); 171 pcie_capability_read_word(parent_dev, reg, value);
173 return 0; 172 return 0;
174} 173}
175 174
@@ -247,7 +246,7 @@ static void ixgbe_check_minimum_link(struct ixgbe_adapter *adapter,
247 max_gts = 4 * width; 246 max_gts = 4 * width;
248 break; 247 break;
249 case PCIE_SPEED_8_0GT: 248 case PCIE_SPEED_8_0GT:
250 /* 128b/130b encoding only reduces throughput by 1% */ 249 /* 128b/130b encoding reduces throughput by less than 2% */
251 max_gts = 8 * width; 250 max_gts = 8 * width;
252 break; 251 break;
253 default: 252 default:
@@ -265,7 +264,7 @@ static void ixgbe_check_minimum_link(struct ixgbe_adapter *adapter,
265 width, 264 width,
266 (speed == PCIE_SPEED_2_5GT ? "20%" : 265 (speed == PCIE_SPEED_2_5GT ? "20%" :
267 speed == PCIE_SPEED_5_0GT ? "20%" : 266 speed == PCIE_SPEED_5_0GT ? "20%" :
268 speed == PCIE_SPEED_8_0GT ? "N/a" : 267 speed == PCIE_SPEED_8_0GT ? "<2%" :
269 "Unknown")); 268 "Unknown"));
270 269
271 if (max_gts < expected_gts) { 270 if (max_gts < expected_gts) {
@@ -872,11 +871,18 @@ static u64 ixgbe_get_tx_completed(struct ixgbe_ring *ring)
872 871
873static u64 ixgbe_get_tx_pending(struct ixgbe_ring *ring) 872static u64 ixgbe_get_tx_pending(struct ixgbe_ring *ring)
874{ 873{
875 struct ixgbe_adapter *adapter = netdev_priv(ring->netdev); 874 struct ixgbe_adapter *adapter;
876 struct ixgbe_hw *hw = &adapter->hw; 875 struct ixgbe_hw *hw;
876 u32 head, tail;
877 877
878 u32 head = IXGBE_READ_REG(hw, IXGBE_TDH(ring->reg_idx)); 878 if (ring->l2_accel_priv)
879 u32 tail = IXGBE_READ_REG(hw, IXGBE_TDT(ring->reg_idx)); 879 adapter = ring->l2_accel_priv->real_adapter;
880 else
881 adapter = netdev_priv(ring->netdev);
882
883 hw = &adapter->hw;
884 head = IXGBE_READ_REG(hw, IXGBE_TDH(ring->reg_idx));
885 tail = IXGBE_READ_REG(hw, IXGBE_TDT(ring->reg_idx));
880 886
881 if (head != tail) 887 if (head != tail)
882 return (head < tail) ? 888 return (head < tail) ?
@@ -1585,7 +1591,7 @@ static void ixgbe_rx_skb(struct ixgbe_q_vector *q_vector,
1585{ 1591{
1586 struct ixgbe_adapter *adapter = q_vector->adapter; 1592 struct ixgbe_adapter *adapter = q_vector->adapter;
1587 1593
1588 if (ixgbe_qv_ll_polling(q_vector)) 1594 if (ixgbe_qv_busy_polling(q_vector))
1589 netif_receive_skb(skb); 1595 netif_receive_skb(skb);
1590 else if (!(adapter->flags & IXGBE_FLAG_IN_NETPOLL)) 1596 else if (!(adapter->flags & IXGBE_FLAG_IN_NETPOLL))
1591 napi_gro_receive(&q_vector->napi, skb); 1597 napi_gro_receive(&q_vector->napi, skb);
@@ -2097,7 +2103,7 @@ static int ixgbe_low_latency_recv(struct napi_struct *napi)
2097 2103
2098 ixgbe_for_each_ring(ring, q_vector->rx) { 2104 ixgbe_for_each_ring(ring, q_vector->rx) {
2099 found = ixgbe_clean_rx_irq(q_vector, ring, 4); 2105 found = ixgbe_clean_rx_irq(q_vector, ring, 4);
2100#ifdef LL_EXTENDED_STATS 2106#ifdef BP_EXTENDED_STATS
2101 if (found) 2107 if (found)
2102 ring->stats.cleaned += found; 2108 ring->stats.cleaned += found;
2103 else 2109 else
@@ -3005,7 +3011,7 @@ void ixgbe_configure_tx_ring(struct ixgbe_adapter *adapter,
3005 struct ixgbe_q_vector *q_vector = ring->q_vector; 3011 struct ixgbe_q_vector *q_vector = ring->q_vector;
3006 3012
3007 if (q_vector) 3013 if (q_vector)
3008 netif_set_xps_queue(adapter->netdev, 3014 netif_set_xps_queue(ring->netdev,
3009 &q_vector->affinity_mask, 3015 &q_vector->affinity_mask,
3010 ring->queue_index); 3016 ring->queue_index);
3011 } 3017 }
@@ -3395,7 +3401,7 @@ static void ixgbe_setup_psrtype(struct ixgbe_adapter *adapter)
3395{ 3401{
3396 struct ixgbe_hw *hw = &adapter->hw; 3402 struct ixgbe_hw *hw = &adapter->hw;
3397 int rss_i = adapter->ring_feature[RING_F_RSS].indices; 3403 int rss_i = adapter->ring_feature[RING_F_RSS].indices;
3398 int p; 3404 u16 pool;
3399 3405
3400 /* PSRTYPE must be initialized in non 82598 adapters */ 3406 /* PSRTYPE must be initialized in non 82598 adapters */
3401 u32 psrtype = IXGBE_PSRTYPE_TCPHDR | 3407 u32 psrtype = IXGBE_PSRTYPE_TCPHDR |
@@ -3412,9 +3418,8 @@ static void ixgbe_setup_psrtype(struct ixgbe_adapter *adapter)
3412 else if (rss_i > 1) 3418 else if (rss_i > 1)
3413 psrtype |= 1 << 29; 3419 psrtype |= 1 << 29;
3414 3420
3415 for (p = 0; p < adapter->num_rx_pools; p++) 3421 for_each_set_bit(pool, &adapter->fwd_bitmask, 32)
3416 IXGBE_WRITE_REG(hw, IXGBE_PSRTYPE(VMDQ_P(p)), 3422 IXGBE_WRITE_REG(hw, IXGBE_PSRTYPE(VMDQ_P(pool)), psrtype);
3417 psrtype);
3418} 3423}
3419 3424
3420static void ixgbe_configure_virtualization(struct ixgbe_adapter *adapter) 3425static void ixgbe_configure_virtualization(struct ixgbe_adapter *adapter)
@@ -3683,7 +3688,11 @@ static void ixgbe_vlan_strip_disable(struct ixgbe_adapter *adapter)
3683 case ixgbe_mac_82599EB: 3688 case ixgbe_mac_82599EB:
3684 case ixgbe_mac_X540: 3689 case ixgbe_mac_X540:
3685 for (i = 0; i < adapter->num_rx_queues; i++) { 3690 for (i = 0; i < adapter->num_rx_queues; i++) {
3686 j = adapter->rx_ring[i]->reg_idx; 3691 struct ixgbe_ring *ring = adapter->rx_ring[i];
3692
3693 if (ring->l2_accel_priv)
3694 continue;
3695 j = ring->reg_idx;
3687 vlnctrl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(j)); 3696 vlnctrl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(j));
3688 vlnctrl &= ~IXGBE_RXDCTL_VME; 3697 vlnctrl &= ~IXGBE_RXDCTL_VME;
3689 IXGBE_WRITE_REG(hw, IXGBE_RXDCTL(j), vlnctrl); 3698 IXGBE_WRITE_REG(hw, IXGBE_RXDCTL(j), vlnctrl);
@@ -3713,7 +3722,11 @@ static void ixgbe_vlan_strip_enable(struct ixgbe_adapter *adapter)
3713 case ixgbe_mac_82599EB: 3722 case ixgbe_mac_82599EB:
3714 case ixgbe_mac_X540: 3723 case ixgbe_mac_X540:
3715 for (i = 0; i < adapter->num_rx_queues; i++) { 3724 for (i = 0; i < adapter->num_rx_queues; i++) {
3716 j = adapter->rx_ring[i]->reg_idx; 3725 struct ixgbe_ring *ring = adapter->rx_ring[i];
3726
3727 if (ring->l2_accel_priv)
3728 continue;
3729 j = ring->reg_idx;
3717 vlnctrl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(j)); 3730 vlnctrl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(j));
3718 vlnctrl |= IXGBE_RXDCTL_VME; 3731 vlnctrl |= IXGBE_RXDCTL_VME;
3719 IXGBE_WRITE_REG(hw, IXGBE_RXDCTL(j), vlnctrl); 3732 IXGBE_WRITE_REG(hw, IXGBE_RXDCTL(j), vlnctrl);
@@ -3750,7 +3763,7 @@ static int ixgbe_write_uc_addr_list(struct net_device *netdev)
3750 unsigned int rar_entries = hw->mac.num_rar_entries - 1; 3763 unsigned int rar_entries = hw->mac.num_rar_entries - 1;
3751 int count = 0; 3764 int count = 0;
3752 3765
3753 /* In SR-IOV mode significantly less RAR entries are available */ 3766 /* In SR-IOV/VMDQ modes significantly less RAR entries are available */
3754 if (adapter->flags & IXGBE_FLAG_SRIOV_ENABLED) 3767 if (adapter->flags & IXGBE_FLAG_SRIOV_ENABLED)
3755 rar_entries = IXGBE_MAX_PF_MACVLANS - 1; 3768 rar_entries = IXGBE_MAX_PF_MACVLANS - 1;
3756 3769
@@ -3825,14 +3838,6 @@ void ixgbe_set_rx_mode(struct net_device *netdev)
3825 if (netdev->flags & IFF_ALLMULTI) { 3838 if (netdev->flags & IFF_ALLMULTI) {
3826 fctrl |= IXGBE_FCTRL_MPE; 3839 fctrl |= IXGBE_FCTRL_MPE;
3827 vmolr |= IXGBE_VMOLR_MPE; 3840 vmolr |= IXGBE_VMOLR_MPE;
3828 } else {
3829 /*
3830 * Write addresses to the MTA, if the attempt fails
3831 * then we should just turn on promiscuous mode so
3832 * that we can at least receive multicast traffic
3833 */
3834 hw->mac.ops.update_mc_addr_list(hw, netdev);
3835 vmolr |= IXGBE_VMOLR_ROMPE;
3836 } 3841 }
3837 ixgbe_vlan_filter_enable(adapter); 3842 ixgbe_vlan_filter_enable(adapter);
3838 hw->addr_ctrl.user_set_promisc = false; 3843 hw->addr_ctrl.user_set_promisc = false;
@@ -3849,6 +3854,13 @@ void ixgbe_set_rx_mode(struct net_device *netdev)
3849 vmolr |= IXGBE_VMOLR_ROPE; 3854 vmolr |= IXGBE_VMOLR_ROPE;
3850 } 3855 }
3851 3856
3857 /* Write addresses to the MTA, if the attempt fails
3858 * then we should just turn on promiscuous mode so
3859 * that we can at least receive multicast traffic
3860 */
3861 hw->mac.ops.update_mc_addr_list(hw, netdev);
3862 vmolr |= IXGBE_VMOLR_ROMPE;
3863
3852 if (adapter->num_vfs) 3864 if (adapter->num_vfs)
3853 ixgbe_restore_vf_multicasts(adapter); 3865 ixgbe_restore_vf_multicasts(adapter);
3854 3866
@@ -3893,15 +3905,13 @@ static void ixgbe_napi_disable_all(struct ixgbe_adapter *adapter)
3893{ 3905{
3894 int q_idx; 3906 int q_idx;
3895 3907
3896 local_bh_disable(); /* for ixgbe_qv_lock_napi() */
3897 for (q_idx = 0; q_idx < adapter->num_q_vectors; q_idx++) { 3908 for (q_idx = 0; q_idx < adapter->num_q_vectors; q_idx++) {
3898 napi_disable(&adapter->q_vector[q_idx]->napi); 3909 napi_disable(&adapter->q_vector[q_idx]->napi);
3899 while (!ixgbe_qv_lock_napi(adapter->q_vector[q_idx])) { 3910 while (!ixgbe_qv_disable(adapter->q_vector[q_idx])) {
3900 pr_info("QV %d locked\n", q_idx); 3911 pr_info("QV %d locked\n", q_idx);
3901 mdelay(1); 3912 usleep_range(1000, 20000);
3902 } 3913 }
3903 } 3914 }
3904 local_bh_enable();
3905} 3915}
3906 3916
3907#ifdef CONFIG_IXGBE_DCB 3917#ifdef CONFIG_IXGBE_DCB
@@ -4118,6 +4128,228 @@ static void ixgbe_fdir_filter_restore(struct ixgbe_adapter *adapter)
4118 spin_unlock(&adapter->fdir_perfect_lock); 4128 spin_unlock(&adapter->fdir_perfect_lock);
4119} 4129}
4120 4130
4131static void ixgbe_macvlan_set_rx_mode(struct net_device *dev, unsigned int pool,
4132 struct ixgbe_adapter *adapter)
4133{
4134 struct ixgbe_hw *hw = &adapter->hw;
4135 u32 vmolr;
4136
4137 /* No unicast promiscuous support for VMDQ devices. */
4138 vmolr = IXGBE_READ_REG(hw, IXGBE_VMOLR(pool));
4139 vmolr |= (IXGBE_VMOLR_ROMPE | IXGBE_VMOLR_BAM | IXGBE_VMOLR_AUPE);
4140
4141 /* clear the affected bit */
4142 vmolr &= ~IXGBE_VMOLR_MPE;
4143
4144 if (dev->flags & IFF_ALLMULTI) {
4145 vmolr |= IXGBE_VMOLR_MPE;
4146 } else {
4147 vmolr |= IXGBE_VMOLR_ROMPE;
4148 hw->mac.ops.update_mc_addr_list(hw, dev);
4149 }
4150 ixgbe_write_uc_addr_list(adapter->netdev);
4151 IXGBE_WRITE_REG(hw, IXGBE_VMOLR(pool), vmolr);
4152}
4153
4154static void ixgbe_add_mac_filter(struct ixgbe_adapter *adapter,
4155 u8 *addr, u16 pool)
4156{
4157 struct ixgbe_hw *hw = &adapter->hw;
4158 unsigned int entry;
4159
4160 entry = hw->mac.num_rar_entries - pool;
4161 hw->mac.ops.set_rar(hw, entry, addr, VMDQ_P(pool), IXGBE_RAH_AV);
4162}
4163
4164static void ixgbe_fwd_psrtype(struct ixgbe_fwd_adapter *vadapter)
4165{
4166 struct ixgbe_adapter *adapter = vadapter->real_adapter;
4167 int rss_i = adapter->num_rx_queues_per_pool;
4168 struct ixgbe_hw *hw = &adapter->hw;
4169 u16 pool = vadapter->pool;
4170 u32 psrtype = IXGBE_PSRTYPE_TCPHDR |
4171 IXGBE_PSRTYPE_UDPHDR |
4172 IXGBE_PSRTYPE_IPV4HDR |
4173 IXGBE_PSRTYPE_L2HDR |
4174 IXGBE_PSRTYPE_IPV6HDR;
4175
4176 if (hw->mac.type == ixgbe_mac_82598EB)
4177 return;
4178
4179 if (rss_i > 3)
4180 psrtype |= 2 << 29;
4181 else if (rss_i > 1)
4182 psrtype |= 1 << 29;
4183
4184 IXGBE_WRITE_REG(hw, IXGBE_PSRTYPE(VMDQ_P(pool)), psrtype);
4185}
4186
4187/**
4188 * ixgbe_clean_rx_ring - Free Rx Buffers per Queue
4189 * @rx_ring: ring to free buffers from
4190 **/
4191static void ixgbe_clean_rx_ring(struct ixgbe_ring *rx_ring)
4192{
4193 struct device *dev = rx_ring->dev;
4194 unsigned long size;
4195 u16 i;
4196
4197 /* ring already cleared, nothing to do */
4198 if (!rx_ring->rx_buffer_info)
4199 return;
4200
4201 /* Free all the Rx ring sk_buffs */
4202 for (i = 0; i < rx_ring->count; i++) {
4203 struct ixgbe_rx_buffer *rx_buffer;
4204
4205 rx_buffer = &rx_ring->rx_buffer_info[i];
4206 if (rx_buffer->skb) {
4207 struct sk_buff *skb = rx_buffer->skb;
4208 if (IXGBE_CB(skb)->page_released) {
4209 dma_unmap_page(dev,
4210 IXGBE_CB(skb)->dma,
4211 ixgbe_rx_bufsz(rx_ring),
4212 DMA_FROM_DEVICE);
4213 IXGBE_CB(skb)->page_released = false;
4214 }
4215 dev_kfree_skb(skb);
4216 }
4217 rx_buffer->skb = NULL;
4218 if (rx_buffer->dma)
4219 dma_unmap_page(dev, rx_buffer->dma,
4220 ixgbe_rx_pg_size(rx_ring),
4221 DMA_FROM_DEVICE);
4222 rx_buffer->dma = 0;
4223 if (rx_buffer->page)
4224 __free_pages(rx_buffer->page,
4225 ixgbe_rx_pg_order(rx_ring));
4226 rx_buffer->page = NULL;
4227 }
4228
4229 size = sizeof(struct ixgbe_rx_buffer) * rx_ring->count;
4230 memset(rx_ring->rx_buffer_info, 0, size);
4231
4232 /* Zero out the descriptor ring */
4233 memset(rx_ring->desc, 0, rx_ring->size);
4234
4235 rx_ring->next_to_alloc = 0;
4236 rx_ring->next_to_clean = 0;
4237 rx_ring->next_to_use = 0;
4238}
4239
4240static void ixgbe_disable_fwd_ring(struct ixgbe_fwd_adapter *vadapter,
4241 struct ixgbe_ring *rx_ring)
4242{
4243 struct ixgbe_adapter *adapter = vadapter->real_adapter;
4244 int index = rx_ring->queue_index + vadapter->rx_base_queue;
4245
4246 /* shutdown specific queue receive and wait for dma to settle */
4247 ixgbe_disable_rx_queue(adapter, rx_ring);
4248 usleep_range(10000, 20000);
4249 ixgbe_irq_disable_queues(adapter, ((u64)1 << index));
4250 ixgbe_clean_rx_ring(rx_ring);
4251 rx_ring->l2_accel_priv = NULL;
4252}
4253
4254int ixgbe_fwd_ring_down(struct net_device *vdev,
4255 struct ixgbe_fwd_adapter *accel)
4256{
4257 struct ixgbe_adapter *adapter = accel->real_adapter;
4258 unsigned int rxbase = accel->rx_base_queue;
4259 unsigned int txbase = accel->tx_base_queue;
4260 int i;
4261
4262 netif_tx_stop_all_queues(vdev);
4263
4264 for (i = 0; i < adapter->num_rx_queues_per_pool; i++) {
4265 ixgbe_disable_fwd_ring(accel, adapter->rx_ring[rxbase + i]);
4266 adapter->rx_ring[rxbase + i]->netdev = adapter->netdev;
4267 }
4268
4269 for (i = 0; i < adapter->num_rx_queues_per_pool; i++) {
4270 adapter->tx_ring[txbase + i]->l2_accel_priv = NULL;
4271 adapter->tx_ring[txbase + i]->netdev = adapter->netdev;
4272 }
4273
4274
4275 return 0;
4276}
4277
4278static int ixgbe_fwd_ring_up(struct net_device *vdev,
4279 struct ixgbe_fwd_adapter *accel)
4280{
4281 struct ixgbe_adapter *adapter = accel->real_adapter;
4282 unsigned int rxbase, txbase, queues;
4283 int i, baseq, err = 0;
4284
4285 if (!test_bit(accel->pool, &adapter->fwd_bitmask))
4286 return 0;
4287
4288 baseq = accel->pool * adapter->num_rx_queues_per_pool;
4289 netdev_dbg(vdev, "pool %i:%i queues %i:%i VSI bitmask %lx\n",
4290 accel->pool, adapter->num_rx_pools,
4291 baseq, baseq + adapter->num_rx_queues_per_pool,
4292 adapter->fwd_bitmask);
4293
4294 accel->netdev = vdev;
4295 accel->rx_base_queue = rxbase = baseq;
4296 accel->tx_base_queue = txbase = baseq;
4297
4298 for (i = 0; i < adapter->num_rx_queues_per_pool; i++)
4299 ixgbe_disable_fwd_ring(accel, adapter->rx_ring[rxbase + i]);
4300
4301 for (i = 0; i < adapter->num_rx_queues_per_pool; i++) {
4302 adapter->rx_ring[rxbase + i]->netdev = vdev;
4303 adapter->rx_ring[rxbase + i]->l2_accel_priv = accel;
4304 ixgbe_configure_rx_ring(adapter, adapter->rx_ring[rxbase + i]);
4305 }
4306
4307 for (i = 0; i < adapter->num_rx_queues_per_pool; i++) {
4308 adapter->tx_ring[txbase + i]->netdev = vdev;
4309 adapter->tx_ring[txbase + i]->l2_accel_priv = accel;
4310 }
4311
4312 queues = min_t(unsigned int,
4313 adapter->num_rx_queues_per_pool, vdev->num_tx_queues);
4314 err = netif_set_real_num_tx_queues(vdev, queues);
4315 if (err)
4316 goto fwd_queue_err;
4317
4318 err = netif_set_real_num_rx_queues(vdev, queues);
4319 if (err)
4320 goto fwd_queue_err;
4321
4322 if (is_valid_ether_addr(vdev->dev_addr))
4323 ixgbe_add_mac_filter(adapter, vdev->dev_addr, accel->pool);
4324
4325 ixgbe_fwd_psrtype(accel);
4326 ixgbe_macvlan_set_rx_mode(vdev, accel->pool, adapter);
4327 return err;
4328fwd_queue_err:
4329 ixgbe_fwd_ring_down(vdev, accel);
4330 return err;
4331}
4332
4333static void ixgbe_configure_dfwd(struct ixgbe_adapter *adapter)
4334{
4335 struct net_device *upper;
4336 struct list_head *iter;
4337 int err;
4338
4339 netdev_for_each_all_upper_dev_rcu(adapter->netdev, upper, iter) {
4340 if (netif_is_macvlan(upper)) {
4341 struct macvlan_dev *dfwd = netdev_priv(upper);
4342 struct ixgbe_fwd_adapter *vadapter = dfwd->fwd_priv;
4343
4344 if (dfwd->fwd_priv) {
4345 err = ixgbe_fwd_ring_up(upper, vadapter);
4346 if (err)
4347 continue;
4348 }
4349 }
4350 }
4351}
4352
4121static void ixgbe_configure(struct ixgbe_adapter *adapter) 4353static void ixgbe_configure(struct ixgbe_adapter *adapter)
4122{ 4354{
4123 struct ixgbe_hw *hw = &adapter->hw; 4355 struct ixgbe_hw *hw = &adapter->hw;
@@ -4169,6 +4401,7 @@ static void ixgbe_configure(struct ixgbe_adapter *adapter)
4169#endif /* IXGBE_FCOE */ 4401#endif /* IXGBE_FCOE */
4170 ixgbe_configure_tx(adapter); 4402 ixgbe_configure_tx(adapter);
4171 ixgbe_configure_rx(adapter); 4403 ixgbe_configure_rx(adapter);
4404 ixgbe_configure_dfwd(adapter);
4172} 4405}
4173 4406
4174static inline bool ixgbe_is_sfp(struct ixgbe_hw *hw) 4407static inline bool ixgbe_is_sfp(struct ixgbe_hw *hw)
@@ -4322,6 +4555,8 @@ static void ixgbe_setup_gpie(struct ixgbe_adapter *adapter)
4322static void ixgbe_up_complete(struct ixgbe_adapter *adapter) 4555static void ixgbe_up_complete(struct ixgbe_adapter *adapter)
4323{ 4556{
4324 struct ixgbe_hw *hw = &adapter->hw; 4557 struct ixgbe_hw *hw = &adapter->hw;
4558 struct net_device *upper;
4559 struct list_head *iter;
4325 int err; 4560 int err;
4326 u32 ctrl_ext; 4561 u32 ctrl_ext;
4327 4562
@@ -4365,6 +4600,16 @@ static void ixgbe_up_complete(struct ixgbe_adapter *adapter)
4365 /* enable transmits */ 4600 /* enable transmits */
4366 netif_tx_start_all_queues(adapter->netdev); 4601 netif_tx_start_all_queues(adapter->netdev);
4367 4602
4603 /* enable any upper devices */
4604 netdev_for_each_all_upper_dev_rcu(adapter->netdev, upper, iter) {
4605 if (netif_is_macvlan(upper)) {
4606 struct macvlan_dev *vlan = netdev_priv(upper);
4607
4608 if (vlan->fwd_priv)
4609 netif_tx_start_all_queues(upper);
4610 }
4611 }
4612
4368 /* bring the link up in the watchdog, this could race with our first 4613 /* bring the link up in the watchdog, this could race with our first
4369 * link up interrupt but shouldn't be a problem */ 4614 * link up interrupt but shouldn't be a problem */
4370 adapter->flags |= IXGBE_FLAG_NEED_LINK_UPDATE; 4615 adapter->flags |= IXGBE_FLAG_NEED_LINK_UPDATE;
@@ -4456,59 +4701,6 @@ void ixgbe_reset(struct ixgbe_adapter *adapter)
4456} 4701}
4457 4702
4458/** 4703/**
4459 * ixgbe_clean_rx_ring - Free Rx Buffers per Queue
4460 * @rx_ring: ring to free buffers from
4461 **/
4462static void ixgbe_clean_rx_ring(struct ixgbe_ring *rx_ring)
4463{
4464 struct device *dev = rx_ring->dev;
4465 unsigned long size;
4466 u16 i;
4467
4468 /* ring already cleared, nothing to do */
4469 if (!rx_ring->rx_buffer_info)
4470 return;
4471
4472 /* Free all the Rx ring sk_buffs */
4473 for (i = 0; i < rx_ring->count; i++) {
4474 struct ixgbe_rx_buffer *rx_buffer;
4475
4476 rx_buffer = &rx_ring->rx_buffer_info[i];
4477 if (rx_buffer->skb) {
4478 struct sk_buff *skb = rx_buffer->skb;
4479 if (IXGBE_CB(skb)->page_released) {
4480 dma_unmap_page(dev,
4481 IXGBE_CB(skb)->dma,
4482 ixgbe_rx_bufsz(rx_ring),
4483 DMA_FROM_DEVICE);
4484 IXGBE_CB(skb)->page_released = false;
4485 }
4486 dev_kfree_skb(skb);
4487 }
4488 rx_buffer->skb = NULL;
4489 if (rx_buffer->dma)
4490 dma_unmap_page(dev, rx_buffer->dma,
4491 ixgbe_rx_pg_size(rx_ring),
4492 DMA_FROM_DEVICE);
4493 rx_buffer->dma = 0;
4494 if (rx_buffer->page)
4495 __free_pages(rx_buffer->page,
4496 ixgbe_rx_pg_order(rx_ring));
4497 rx_buffer->page = NULL;
4498 }
4499
4500 size = sizeof(struct ixgbe_rx_buffer) * rx_ring->count;
4501 memset(rx_ring->rx_buffer_info, 0, size);
4502
4503 /* Zero out the descriptor ring */
4504 memset(rx_ring->desc, 0, rx_ring->size);
4505
4506 rx_ring->next_to_alloc = 0;
4507 rx_ring->next_to_clean = 0;
4508 rx_ring->next_to_use = 0;
4509}
4510
4511/**
4512 * ixgbe_clean_tx_ring - Free Tx Buffers 4704 * ixgbe_clean_tx_ring - Free Tx Buffers
4513 * @tx_ring: ring to be cleaned 4705 * @tx_ring: ring to be cleaned
4514 **/ 4706 **/
@@ -4585,6 +4777,8 @@ void ixgbe_down(struct ixgbe_adapter *adapter)
4585{ 4777{
4586 struct net_device *netdev = adapter->netdev; 4778 struct net_device *netdev = adapter->netdev;
4587 struct ixgbe_hw *hw = &adapter->hw; 4779 struct ixgbe_hw *hw = &adapter->hw;
4780 struct net_device *upper;
4781 struct list_head *iter;
4588 u32 rxctrl; 4782 u32 rxctrl;
4589 int i; 4783 int i;
4590 4784
@@ -4608,6 +4802,19 @@ void ixgbe_down(struct ixgbe_adapter *adapter)
4608 netif_carrier_off(netdev); 4802 netif_carrier_off(netdev);
4609 netif_tx_disable(netdev); 4803 netif_tx_disable(netdev);
4610 4804
4805 /* disable any upper devices */
4806 netdev_for_each_all_upper_dev_rcu(adapter->netdev, upper, iter) {
4807 if (netif_is_macvlan(upper)) {
4808 struct macvlan_dev *vlan = netdev_priv(upper);
4809
4810 if (vlan->fwd_priv) {
4811 netif_tx_stop_all_queues(upper);
4812 netif_carrier_off(upper);
4813 netif_tx_disable(upper);
4814 }
4815 }
4816 }
4817
4611 ixgbe_irq_disable(adapter); 4818 ixgbe_irq_disable(adapter);
4612 4819
4613 ixgbe_napi_disable_all(adapter); 4820 ixgbe_napi_disable_all(adapter);
@@ -4816,11 +5023,20 @@ static int ixgbe_sw_init(struct ixgbe_adapter *adapter)
4816 hw->fc.disable_fc_autoneg = ixgbe_device_supports_autoneg_fc(hw); 5023 hw->fc.disable_fc_autoneg = ixgbe_device_supports_autoneg_fc(hw);
4817 5024
4818#ifdef CONFIG_PCI_IOV 5025#ifdef CONFIG_PCI_IOV
5026 if (max_vfs > 0)
5027 e_dev_warn("Enabling SR-IOV VFs using the max_vfs module parameter is deprecated - please use the pci sysfs interface instead.\n");
5028
4819 /* assign number of SR-IOV VFs */ 5029 /* assign number of SR-IOV VFs */
4820 if (hw->mac.type != ixgbe_mac_82598EB) 5030 if (hw->mac.type != ixgbe_mac_82598EB) {
4821 adapter->num_vfs = (max_vfs > 63) ? 0 : max_vfs; 5031 if (max_vfs > 63) {
5032 adapter->num_vfs = 0;
5033 e_dev_warn("max_vfs parameter out of range. Not assigning any SR-IOV VFs\n");
5034 } else {
5035 adapter->num_vfs = max_vfs;
5036 }
5037 }
5038#endif /* CONFIG_PCI_IOV */
4822 5039
4823#endif
4824 /* enable itr by default in dynamic mode */ 5040 /* enable itr by default in dynamic mode */
4825 adapter->rx_itr_setting = 1; 5041 adapter->rx_itr_setting = 1;
4826 adapter->tx_itr_setting = 1; 5042 adapter->tx_itr_setting = 1;
@@ -4838,6 +5054,8 @@ static int ixgbe_sw_init(struct ixgbe_adapter *adapter)
4838 return -EIO; 5054 return -EIO;
4839 } 5055 }
4840 5056
5057 /* PF holds first pool slot */
5058 set_bit(0, &adapter->fwd_bitmask);
4841 set_bit(__IXGBE_DOWN, &adapter->state); 5059 set_bit(__IXGBE_DOWN, &adapter->state);
4842 5060
4843 return 0; 5061 return 0;
@@ -5143,7 +5361,7 @@ static int ixgbe_change_mtu(struct net_device *netdev, int new_mtu)
5143static int ixgbe_open(struct net_device *netdev) 5361static int ixgbe_open(struct net_device *netdev)
5144{ 5362{
5145 struct ixgbe_adapter *adapter = netdev_priv(netdev); 5363 struct ixgbe_adapter *adapter = netdev_priv(netdev);
5146 int err; 5364 int err, queues;
5147 5365
5148 /* disallow open during test */ 5366 /* disallow open during test */
5149 if (test_bit(__IXGBE_TESTING, &adapter->state)) 5367 if (test_bit(__IXGBE_TESTING, &adapter->state))
@@ -5168,16 +5386,21 @@ static int ixgbe_open(struct net_device *netdev)
5168 goto err_req_irq; 5386 goto err_req_irq;
5169 5387
5170 /* Notify the stack of the actual queue counts. */ 5388 /* Notify the stack of the actual queue counts. */
5171 err = netif_set_real_num_tx_queues(netdev, 5389 if (adapter->num_rx_pools > 1)
5172 adapter->num_rx_pools > 1 ? 1 : 5390 queues = adapter->num_rx_queues_per_pool;
5173 adapter->num_tx_queues); 5391 else
5392 queues = adapter->num_tx_queues;
5393
5394 err = netif_set_real_num_tx_queues(netdev, queues);
5174 if (err) 5395 if (err)
5175 goto err_set_queues; 5396 goto err_set_queues;
5176 5397
5177 5398 if (adapter->num_rx_pools > 1 &&
5178 err = netif_set_real_num_rx_queues(netdev, 5399 adapter->num_rx_queues > IXGBE_MAX_L2A_QUEUES)
5179 adapter->num_rx_pools > 1 ? 1 : 5400 queues = IXGBE_MAX_L2A_QUEUES;
5180 adapter->num_rx_queues); 5401 else
5402 queues = adapter->num_rx_queues;
5403 err = netif_set_real_num_rx_queues(netdev, queues);
5181 if (err) 5404 if (err)
5182 goto err_set_queues; 5405 goto err_set_queues;
5183 5406
@@ -6767,8 +6990,9 @@ out_drop:
6767 return NETDEV_TX_OK; 6990 return NETDEV_TX_OK;
6768} 6991}
6769 6992
6770static netdev_tx_t ixgbe_xmit_frame(struct sk_buff *skb, 6993static netdev_tx_t __ixgbe_xmit_frame(struct sk_buff *skb,
6771 struct net_device *netdev) 6994 struct net_device *netdev,
6995 struct ixgbe_ring *ring)
6772{ 6996{
6773 struct ixgbe_adapter *adapter = netdev_priv(netdev); 6997 struct ixgbe_adapter *adapter = netdev_priv(netdev);
6774 struct ixgbe_ring *tx_ring; 6998 struct ixgbe_ring *tx_ring;
@@ -6784,10 +7008,17 @@ static netdev_tx_t ixgbe_xmit_frame(struct sk_buff *skb,
6784 skb_set_tail_pointer(skb, 17); 7008 skb_set_tail_pointer(skb, 17);
6785 } 7009 }
6786 7010
6787 tx_ring = adapter->tx_ring[skb->queue_mapping]; 7011 tx_ring = ring ? ring : adapter->tx_ring[skb->queue_mapping];
7012
6788 return ixgbe_xmit_frame_ring(skb, adapter, tx_ring); 7013 return ixgbe_xmit_frame_ring(skb, adapter, tx_ring);
6789} 7014}
6790 7015
7016static netdev_tx_t ixgbe_xmit_frame(struct sk_buff *skb,
7017 struct net_device *netdev)
7018{
7019 return __ixgbe_xmit_frame(skb, netdev, NULL);
7020}
7021
6791/** 7022/**
6792 * ixgbe_set_mac - Change the Ethernet Address of the NIC 7023 * ixgbe_set_mac - Change the Ethernet Address of the NIC
6793 * @netdev: network interface device structure 7024 * @netdev: network interface device structure
@@ -7044,6 +7275,7 @@ int ixgbe_setup_tc(struct net_device *dev, u8 tc)
7044{ 7275{
7045 struct ixgbe_adapter *adapter = netdev_priv(dev); 7276 struct ixgbe_adapter *adapter = netdev_priv(dev);
7046 struct ixgbe_hw *hw = &adapter->hw; 7277 struct ixgbe_hw *hw = &adapter->hw;
7278 bool pools;
7047 7279
7048 /* Hardware supports up to 8 traffic classes */ 7280 /* Hardware supports up to 8 traffic classes */
7049 if (tc > adapter->dcb_cfg.num_tcs.pg_tcs || 7281 if (tc > adapter->dcb_cfg.num_tcs.pg_tcs ||
@@ -7051,6 +7283,10 @@ int ixgbe_setup_tc(struct net_device *dev, u8 tc)
7051 tc < MAX_TRAFFIC_CLASS)) 7283 tc < MAX_TRAFFIC_CLASS))
7052 return -EINVAL; 7284 return -EINVAL;
7053 7285
7286 pools = (find_first_zero_bit(&adapter->fwd_bitmask, 32) > 1);
7287 if (tc && pools && adapter->num_rx_pools > IXGBE_MAX_DCBMACVLANS)
7288 return -EBUSY;
7289
7054 /* Hardware has to reinitialize queues and interrupts to 7290 /* Hardware has to reinitialize queues and interrupts to
7055 * match packet buffer alignment. Unfortunately, the 7291 * match packet buffer alignment. Unfortunately, the
7056 * hardware is not flexible enough to do this dynamically. 7292 * hardware is not flexible enough to do this dynamically.
@@ -7305,6 +7541,104 @@ static int ixgbe_ndo_bridge_getlink(struct sk_buff *skb, u32 pid, u32 seq,
7305 return ndo_dflt_bridge_getlink(skb, pid, seq, dev, mode); 7541 return ndo_dflt_bridge_getlink(skb, pid, seq, dev, mode);
7306} 7542}
7307 7543
7544static void *ixgbe_fwd_add(struct net_device *pdev, struct net_device *vdev)
7545{
7546 struct ixgbe_fwd_adapter *fwd_adapter = NULL;
7547 struct ixgbe_adapter *adapter = netdev_priv(pdev);
7548 unsigned int limit;
7549 int pool, err;
7550
7551#ifdef CONFIG_RPS
7552 if (vdev->num_rx_queues != vdev->num_tx_queues) {
7553 netdev_info(pdev, "%s: Only supports a single queue count for TX and RX\n",
7554 vdev->name);
7555 return ERR_PTR(-EINVAL);
7556 }
7557#endif
7558 /* Check for hardware restriction on number of rx/tx queues */
7559 if (vdev->num_tx_queues > IXGBE_MAX_L2A_QUEUES ||
7560 vdev->num_tx_queues == IXGBE_BAD_L2A_QUEUE) {
7561 netdev_info(pdev,
7562 "%s: Supports RX/TX Queue counts 1,2, and 4\n",
7563 pdev->name);
7564 return ERR_PTR(-EINVAL);
7565 }
7566
7567 if (((adapter->flags & IXGBE_FLAG_DCB_ENABLED) &&
7568 adapter->num_rx_pools > IXGBE_MAX_DCBMACVLANS - 1) ||
7569 (adapter->num_rx_pools > IXGBE_MAX_MACVLANS))
7570 return ERR_PTR(-EBUSY);
7571
7572 fwd_adapter = kcalloc(1, sizeof(struct ixgbe_fwd_adapter), GFP_KERNEL);
7573 if (!fwd_adapter)
7574 return ERR_PTR(-ENOMEM);
7575
7576 pool = find_first_zero_bit(&adapter->fwd_bitmask, 32);
7577 adapter->num_rx_pools++;
7578 set_bit(pool, &adapter->fwd_bitmask);
7579 limit = find_last_bit(&adapter->fwd_bitmask, 32);
7580
7581 /* Enable VMDq flag so device will be set in VM mode */
7582 adapter->flags |= IXGBE_FLAG_VMDQ_ENABLED | IXGBE_FLAG_SRIOV_ENABLED;
7583 adapter->ring_feature[RING_F_VMDQ].limit = limit + 1;
7584 adapter->ring_feature[RING_F_RSS].limit = vdev->num_tx_queues;
7585
7586 /* Force reinit of ring allocation with VMDQ enabled */
7587 err = ixgbe_setup_tc(pdev, netdev_get_num_tc(pdev));
7588 if (err)
7589 goto fwd_add_err;
7590 fwd_adapter->pool = pool;
7591 fwd_adapter->real_adapter = adapter;
7592 err = ixgbe_fwd_ring_up(vdev, fwd_adapter);
7593 if (err)
7594 goto fwd_add_err;
7595 netif_tx_start_all_queues(vdev);
7596 return fwd_adapter;
7597fwd_add_err:
7598 /* unwind counter and free adapter struct */
7599 netdev_info(pdev,
7600 "%s: dfwd hardware acceleration failed\n", vdev->name);
7601 clear_bit(pool, &adapter->fwd_bitmask);
7602 adapter->num_rx_pools--;
7603 kfree(fwd_adapter);
7604 return ERR_PTR(err);
7605}
7606
7607static void ixgbe_fwd_del(struct net_device *pdev, void *priv)
7608{
7609 struct ixgbe_fwd_adapter *fwd_adapter = priv;
7610 struct ixgbe_adapter *adapter = fwd_adapter->real_adapter;
7611 unsigned int limit;
7612
7613 clear_bit(fwd_adapter->pool, &adapter->fwd_bitmask);
7614 adapter->num_rx_pools--;
7615
7616 limit = find_last_bit(&adapter->fwd_bitmask, 32);
7617 adapter->ring_feature[RING_F_VMDQ].limit = limit + 1;
7618 ixgbe_fwd_ring_down(fwd_adapter->netdev, fwd_adapter);
7619 ixgbe_setup_tc(pdev, netdev_get_num_tc(pdev));
7620 netdev_dbg(pdev, "pool %i:%i queues %i:%i VSI bitmask %lx\n",
7621 fwd_adapter->pool, adapter->num_rx_pools,
7622 fwd_adapter->rx_base_queue,
7623 fwd_adapter->rx_base_queue + adapter->num_rx_queues_per_pool,
7624 adapter->fwd_bitmask);
7625 kfree(fwd_adapter);
7626}
7627
7628static netdev_tx_t ixgbe_fwd_xmit(struct sk_buff *skb,
7629 struct net_device *dev,
7630 void *priv)
7631{
7632 struct ixgbe_fwd_adapter *fwd_adapter = priv;
7633 unsigned int queue;
7634 struct ixgbe_ring *tx_ring;
7635
7636 queue = skb->queue_mapping + fwd_adapter->tx_base_queue;
7637 tx_ring = fwd_adapter->real_adapter->tx_ring[queue];
7638
7639 return __ixgbe_xmit_frame(skb, dev, tx_ring);
7640}
7641
7308static const struct net_device_ops ixgbe_netdev_ops = { 7642static const struct net_device_ops ixgbe_netdev_ops = {
7309 .ndo_open = ixgbe_open, 7643 .ndo_open = ixgbe_open,
7310 .ndo_stop = ixgbe_close, 7644 .ndo_stop = ixgbe_close,
@@ -7349,6 +7683,9 @@ static const struct net_device_ops ixgbe_netdev_ops = {
7349 .ndo_fdb_add = ixgbe_ndo_fdb_add, 7683 .ndo_fdb_add = ixgbe_ndo_fdb_add,
7350 .ndo_bridge_setlink = ixgbe_ndo_bridge_setlink, 7684 .ndo_bridge_setlink = ixgbe_ndo_bridge_setlink,
7351 .ndo_bridge_getlink = ixgbe_ndo_bridge_getlink, 7685 .ndo_bridge_getlink = ixgbe_ndo_bridge_getlink,
7686 .ndo_dfwd_add_station = ixgbe_fwd_add,
7687 .ndo_dfwd_del_station = ixgbe_fwd_del,
7688 .ndo_dfwd_start_xmit = ixgbe_fwd_xmit,
7352}; 7689};
7353 7690
7354/** 7691/**
@@ -7362,19 +7699,16 @@ static const struct net_device_ops ixgbe_netdev_ops = {
7362 **/ 7699 **/
7363static inline int ixgbe_enumerate_functions(struct ixgbe_adapter *adapter) 7700static inline int ixgbe_enumerate_functions(struct ixgbe_adapter *adapter)
7364{ 7701{
7365 struct ixgbe_hw *hw = &adapter->hw;
7366 struct list_head *entry; 7702 struct list_head *entry;
7367 int physfns = 0; 7703 int physfns = 0;
7368 7704
7369 /* Some cards can not use the generic count PCIe functions method, and 7705 /* Some cards can not use the generic count PCIe functions method,
7370 * so must be hardcoded to the correct value. 7706 * because they are behind a parent switch, so we hardcode these with
7707 * the correct number of functions.
7371 */ 7708 */
7372 switch (hw->device_id) { 7709 if (ixgbe_pcie_from_parent(&adapter->hw)) {
7373 case IXGBE_DEV_ID_82599_SFP_SF_QP:
7374 case IXGBE_DEV_ID_82599_QSFP_SF_QP:
7375 physfns = 4; 7710 physfns = 4;
7376 break; 7711 } else {
7377 default:
7378 list_for_each(entry, &adapter->pdev->bus_list) { 7712 list_for_each(entry, &adapter->pdev->bus_list) {
7379 struct pci_dev *pdev = 7713 struct pci_dev *pdev =
7380 list_entry(entry, struct pci_dev, bus_list); 7714 list_entry(entry, struct pci_dev, bus_list);
@@ -7653,7 +7987,8 @@ skip_sriov:
7653 NETIF_F_TSO | 7987 NETIF_F_TSO |
7654 NETIF_F_TSO6 | 7988 NETIF_F_TSO6 |
7655 NETIF_F_RXHASH | 7989 NETIF_F_RXHASH |
7656 NETIF_F_RXCSUM; 7990 NETIF_F_RXCSUM |
7991 NETIF_F_HW_L2FW_DOFFLOAD;
7657 7992
7658 netdev->hw_features = netdev->features; 7993 netdev->hw_features = netdev->features;
7659 7994
@@ -7759,29 +8094,6 @@ skip_sriov:
7759 if (ixgbe_pcie_from_parent(hw)) 8094 if (ixgbe_pcie_from_parent(hw))
7760 ixgbe_get_parent_bus_info(adapter); 8095 ixgbe_get_parent_bus_info(adapter);
7761 8096
7762 /* print bus type/speed/width info */
7763 e_dev_info("(PCI Express:%s:%s) %pM\n",
7764 (hw->bus.speed == ixgbe_bus_speed_8000 ? "8.0GT/s" :
7765 hw->bus.speed == ixgbe_bus_speed_5000 ? "5.0GT/s" :
7766 hw->bus.speed == ixgbe_bus_speed_2500 ? "2.5GT/s" :
7767 "Unknown"),
7768 (hw->bus.width == ixgbe_bus_width_pcie_x8 ? "Width x8" :
7769 hw->bus.width == ixgbe_bus_width_pcie_x4 ? "Width x4" :
7770 hw->bus.width == ixgbe_bus_width_pcie_x1 ? "Width x1" :
7771 "Unknown"),
7772 netdev->dev_addr);
7773
7774 err = ixgbe_read_pba_string_generic(hw, part_str, IXGBE_PBANUM_LENGTH);
7775 if (err)
7776 strncpy(part_str, "Unknown", IXGBE_PBANUM_LENGTH);
7777 if (ixgbe_is_sfp(hw) && hw->phy.sfp_type != ixgbe_sfp_type_not_present)
7778 e_dev_info("MAC: %d, PHY: %d, SFP+: %d, PBA No: %s\n",
7779 hw->mac.type, hw->phy.type, hw->phy.sfp_type,
7780 part_str);
7781 else
7782 e_dev_info("MAC: %d, PHY: %d, PBA No: %s\n",
7783 hw->mac.type, hw->phy.type, part_str);
7784
7785 /* calculate the expected PCIe bandwidth required for optimal 8097 /* calculate the expected PCIe bandwidth required for optimal
7786 * performance. Note that some older parts will never have enough 8098 * performance. Note that some older parts will never have enough
7787 * bandwidth due to being older generation PCIe parts. We clamp these 8099 * bandwidth due to being older generation PCIe parts. We clamp these
@@ -7797,6 +8109,19 @@ skip_sriov:
7797 } 8109 }
7798 ixgbe_check_minimum_link(adapter, expected_gts); 8110 ixgbe_check_minimum_link(adapter, expected_gts);
7799 8111
8112 err = ixgbe_read_pba_string_generic(hw, part_str, IXGBE_PBANUM_LENGTH);
8113 if (err)
8114 strncpy(part_str, "Unknown", IXGBE_PBANUM_LENGTH);
8115 if (ixgbe_is_sfp(hw) && hw->phy.sfp_type != ixgbe_sfp_type_not_present)
8116 e_dev_info("MAC: %d, PHY: %d, SFP+: %d, PBA No: %s\n",
8117 hw->mac.type, hw->phy.type, hw->phy.sfp_type,
8118 part_str);
8119 else
8120 e_dev_info("MAC: %d, PHY: %d, PBA No: %s\n",
8121 hw->mac.type, hw->phy.type, part_str);
8122
8123 e_dev_info("%pM\n", netdev->dev_addr);
8124
7800 /* reset the hardware with the new settings */ 8125 /* reset the hardware with the new settings */
7801 err = hw->mac.ops.start_hw(hw); 8126 err = hw->mac.ops.start_hw(hw);
7802 if (err == IXGBE_ERR_EEPROM_VERSION) { 8127 if (err == IXGBE_ERR_EEPROM_VERSION) {
diff --git a/drivers/net/ethernet/intel/ixgbe/ixgbe_phy.h b/drivers/net/ethernet/intel/ixgbe/ixgbe_phy.h
index 24af12e3719e..aae900a256da 100644
--- a/drivers/net/ethernet/intel/ixgbe/ixgbe_phy.h
+++ b/drivers/net/ethernet/intel/ixgbe/ixgbe_phy.h
@@ -57,28 +57,28 @@
57#define IXGBE_SFF_QSFP_DEVICE_TECH 0x93 57#define IXGBE_SFF_QSFP_DEVICE_TECH 0x93
58 58
59/* Bitmasks */ 59/* Bitmasks */
60#define IXGBE_SFF_DA_PASSIVE_CABLE 0x4 60#define IXGBE_SFF_DA_PASSIVE_CABLE 0x4
61#define IXGBE_SFF_DA_ACTIVE_CABLE 0x8 61#define IXGBE_SFF_DA_ACTIVE_CABLE 0x8
62#define IXGBE_SFF_DA_SPEC_ACTIVE_LIMITING 0x4 62#define IXGBE_SFF_DA_SPEC_ACTIVE_LIMITING 0x4
63#define IXGBE_SFF_1GBASESX_CAPABLE 0x1 63#define IXGBE_SFF_1GBASESX_CAPABLE 0x1
64#define IXGBE_SFF_1GBASELX_CAPABLE 0x2 64#define IXGBE_SFF_1GBASELX_CAPABLE 0x2
65#define IXGBE_SFF_1GBASET_CAPABLE 0x8 65#define IXGBE_SFF_1GBASET_CAPABLE 0x8
66#define IXGBE_SFF_10GBASESR_CAPABLE 0x10 66#define IXGBE_SFF_10GBASESR_CAPABLE 0x10
67#define IXGBE_SFF_10GBASELR_CAPABLE 0x20 67#define IXGBE_SFF_10GBASELR_CAPABLE 0x20
68#define IXGBE_SFF_SOFT_RS_SELECT_MASK 0x8 68#define IXGBE_SFF_SOFT_RS_SELECT_MASK 0x8
69#define IXGBE_SFF_SOFT_RS_SELECT_10G 0x8 69#define IXGBE_SFF_SOFT_RS_SELECT_10G 0x8
70#define IXGBE_SFF_SOFT_RS_SELECT_1G 0x0 70#define IXGBE_SFF_SOFT_RS_SELECT_1G 0x0
71#define IXGBE_SFF_ADDRESSING_MODE 0x4 71#define IXGBE_SFF_ADDRESSING_MODE 0x4
72#define IXGBE_SFF_QSFP_DA_ACTIVE_CABLE 0x1 72#define IXGBE_SFF_QSFP_DA_ACTIVE_CABLE 0x1
73#define IXGBE_SFF_QSFP_DA_PASSIVE_CABLE 0x8 73#define IXGBE_SFF_QSFP_DA_PASSIVE_CABLE 0x8
74#define IXGBE_SFF_QSFP_CONNECTOR_NOT_SEPARABLE 0x23 74#define IXGBE_SFF_QSFP_CONNECTOR_NOT_SEPARABLE 0x23
75#define IXGBE_SFF_QSFP_TRANSMITER_850NM_VCSEL 0x0 75#define IXGBE_SFF_QSFP_TRANSMITER_850NM_VCSEL 0x0
76#define IXGBE_I2C_EEPROM_READ_MASK 0x100 76#define IXGBE_I2C_EEPROM_READ_MASK 0x100
77#define IXGBE_I2C_EEPROM_STATUS_MASK 0x3 77#define IXGBE_I2C_EEPROM_STATUS_MASK 0x3
78#define IXGBE_I2C_EEPROM_STATUS_NO_OPERATION 0x0 78#define IXGBE_I2C_EEPROM_STATUS_NO_OPERATION 0x0
79#define IXGBE_I2C_EEPROM_STATUS_PASS 0x1 79#define IXGBE_I2C_EEPROM_STATUS_PASS 0x1
80#define IXGBE_I2C_EEPROM_STATUS_FAIL 0x2 80#define IXGBE_I2C_EEPROM_STATUS_FAIL 0x2
81#define IXGBE_I2C_EEPROM_STATUS_IN_PROGRESS 0x3 81#define IXGBE_I2C_EEPROM_STATUS_IN_PROGRESS 0x3
82 82
83/* Flow control defines */ 83/* Flow control defines */
84#define IXGBE_TAF_SYM_PAUSE 0x400 84#define IXGBE_TAF_SYM_PAUSE 0x400
diff --git a/drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c b/drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c
index 276d7b135332..d6f0c0d8cf11 100644
--- a/drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c
+++ b/drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c
@@ -129,10 +129,6 @@ void ixgbe_enable_sriov(struct ixgbe_adapter *adapter)
129 if (!pre_existing_vfs && !adapter->num_vfs) 129 if (!pre_existing_vfs && !adapter->num_vfs)
130 return; 130 return;
131 131
132 if (!pre_existing_vfs)
133 dev_warn(&adapter->pdev->dev,
134 "Enabling SR-IOV VFs using the module parameter is deprecated - please use the pci sysfs interface.\n");
135
136 /* If there are pre-existing VFs then we have to force 132 /* If there are pre-existing VFs then we have to force
137 * use of that many - over ride any module parameter value. 133 * use of that many - over ride any module parameter value.
138 * This may result from the user unloading the PF driver 134 * This may result from the user unloading the PF driver
@@ -223,17 +219,19 @@ int ixgbe_disable_sriov(struct ixgbe_adapter *adapter)
223 IXGBE_WRITE_FLUSH(hw); 219 IXGBE_WRITE_FLUSH(hw);
224 220
225 /* Disable VMDq flag so device will be set in VM mode */ 221 /* Disable VMDq flag so device will be set in VM mode */
226 if (adapter->ring_feature[RING_F_VMDQ].limit == 1) 222 if (adapter->ring_feature[RING_F_VMDQ].limit == 1) {
227 adapter->flags &= ~IXGBE_FLAG_VMDQ_ENABLED; 223 adapter->flags &= ~IXGBE_FLAG_VMDQ_ENABLED;
228 adapter->ring_feature[RING_F_VMDQ].offset = 0; 224 adapter->flags &= ~IXGBE_FLAG_SRIOV_ENABLED;
225 rss = min_t(int, IXGBE_MAX_RSS_INDICES, num_online_cpus());
226 } else {
227 rss = min_t(int, IXGBE_MAX_L2A_QUEUES, num_online_cpus());
228 }
229 229
230 rss = min_t(int, IXGBE_MAX_RSS_INDICES, num_online_cpus()); 230 adapter->ring_feature[RING_F_VMDQ].offset = 0;
231 adapter->ring_feature[RING_F_RSS].limit = rss; 231 adapter->ring_feature[RING_F_RSS].limit = rss;
232 232
233 /* take a breather then clean up driver data */ 233 /* take a breather then clean up driver data */
234 msleep(100); 234 msleep(100);
235
236 adapter->flags &= ~IXGBE_FLAG_SRIOV_ENABLED;
237 return 0; 235 return 0;
238} 236}
239 237
@@ -298,13 +296,10 @@ static int ixgbe_pci_sriov_disable(struct pci_dev *dev)
298 err = ixgbe_disable_sriov(adapter); 296 err = ixgbe_disable_sriov(adapter);
299 297
300 /* Only reinit if no error and state changed */ 298 /* Only reinit if no error and state changed */
301 if (!err && current_flags != adapter->flags) {
302 /* ixgbe_disable_sriov() doesn't clear VMDQ flag */
303 adapter->flags &= ~IXGBE_FLAG_VMDQ_ENABLED;
304#ifdef CONFIG_PCI_IOV 299#ifdef CONFIG_PCI_IOV
300 if (!err && current_flags != adapter->flags)
305 ixgbe_sriov_reinit(adapter); 301 ixgbe_sriov_reinit(adapter);
306#endif 302#endif
307 }
308 303
309 return err; 304 return err;
310} 305}
@@ -558,7 +553,7 @@ static int ixgbe_set_vf_mac(struct ixgbe_adapter *adapter,
558 struct ixgbe_hw *hw = &adapter->hw; 553 struct ixgbe_hw *hw = &adapter->hw;
559 int rar_entry = hw->mac.num_rar_entries - (vf + 1); 554 int rar_entry = hw->mac.num_rar_entries - (vf + 1);
560 555
561 memcpy(adapter->vfinfo[vf].vf_mac_addresses, mac_addr, 6); 556 memcpy(adapter->vfinfo[vf].vf_mac_addresses, mac_addr, ETH_ALEN);
562 hw->mac.ops.set_rar(hw, rar_entry, mac_addr, vf, IXGBE_RAH_AV); 557 hw->mac.ops.set_rar(hw, rar_entry, mac_addr, vf, IXGBE_RAH_AV);
563 558
564 return 0; 559 return 0;
@@ -621,16 +616,13 @@ static int ixgbe_set_vf_macvlan(struct ixgbe_adapter *adapter,
621 616
622int ixgbe_vf_configuration(struct pci_dev *pdev, unsigned int event_mask) 617int ixgbe_vf_configuration(struct pci_dev *pdev, unsigned int event_mask)
623{ 618{
624 unsigned char vf_mac_addr[6];
625 struct ixgbe_adapter *adapter = pci_get_drvdata(pdev); 619 struct ixgbe_adapter *adapter = pci_get_drvdata(pdev);
626 unsigned int vfn = (event_mask & 0x3f); 620 unsigned int vfn = (event_mask & 0x3f);
627 621
628 bool enable = ((event_mask & 0x10000000U) != 0); 622 bool enable = ((event_mask & 0x10000000U) != 0);
629 623
630 if (enable) { 624 if (enable)
631 eth_zero_addr(vf_mac_addr); 625 eth_zero_addr(adapter->vfinfo[vfn].vf_mac_addresses);
632 memcpy(adapter->vfinfo[vfn].vf_mac_addresses, vf_mac_addr, 6);
633 }
634 626
635 return 0; 627 return 0;
636} 628}
diff --git a/drivers/net/ethernet/intel/ixgbe/ixgbe_type.h b/drivers/net/ethernet/intel/ixgbe/ixgbe_type.h
index 10775cb9b6d8..7c19e969576f 100644
--- a/drivers/net/ethernet/intel/ixgbe/ixgbe_type.h
+++ b/drivers/net/ethernet/intel/ixgbe/ixgbe_type.h
@@ -561,6 +561,10 @@ struct ixgbe_thermal_sensor_data {
561#define IXGBE_RTTDQSEL 0x04904 561#define IXGBE_RTTDQSEL 0x04904
562#define IXGBE_RTTDT1C 0x04908 562#define IXGBE_RTTDT1C 0x04908
563#define IXGBE_RTTDT1S 0x0490C 563#define IXGBE_RTTDT1S 0x0490C
564#define IXGBE_RTTQCNCR 0x08B00
565#define IXGBE_RTTQCNTG 0x04A90
566#define IXGBE_RTTBCNRD 0x0498C
567#define IXGBE_RTTQCNRR 0x0498C
564#define IXGBE_RTTDTECC 0x04990 568#define IXGBE_RTTDTECC 0x04990
565#define IXGBE_RTTDTECC_NO_BCN 0x00000100 569#define IXGBE_RTTDTECC_NO_BCN 0x00000100
566#define IXGBE_RTTBCNRC 0x04984 570#define IXGBE_RTTBCNRC 0x04984
@@ -570,6 +574,7 @@ struct ixgbe_thermal_sensor_data {
570#define IXGBE_RTTBCNRC_RF_INT_MASK \ 574#define IXGBE_RTTBCNRC_RF_INT_MASK \
571 (IXGBE_RTTBCNRC_RF_DEC_MASK << IXGBE_RTTBCNRC_RF_INT_SHIFT) 575 (IXGBE_RTTBCNRC_RF_DEC_MASK << IXGBE_RTTBCNRC_RF_INT_SHIFT)
572#define IXGBE_RTTBCNRM 0x04980 576#define IXGBE_RTTBCNRM 0x04980
577#define IXGBE_RTTQCNRM 0x04980
573 578
574/* FCoE DMA Context Registers */ 579/* FCoE DMA Context Registers */
575#define IXGBE_FCPTRL 0x02410 /* FC User Desc. PTR Low */ 580#define IXGBE_FCPTRL 0x02410 /* FC User Desc. PTR Low */
diff --git a/drivers/net/ethernet/intel/ixgbe/ixgbe_x540.c b/drivers/net/ethernet/intel/ixgbe/ixgbe_x540.c
index 389324f5929a..24b80a6cfca4 100644
--- a/drivers/net/ethernet/intel/ixgbe/ixgbe_x540.c
+++ b/drivers/net/ethernet/intel/ixgbe/ixgbe_x540.c
@@ -32,12 +32,12 @@
32#include "ixgbe.h" 32#include "ixgbe.h"
33#include "ixgbe_phy.h" 33#include "ixgbe_phy.h"
34 34
35#define IXGBE_X540_MAX_TX_QUEUES 128 35#define IXGBE_X540_MAX_TX_QUEUES 128
36#define IXGBE_X540_MAX_RX_QUEUES 128 36#define IXGBE_X540_MAX_RX_QUEUES 128
37#define IXGBE_X540_RAR_ENTRIES 128 37#define IXGBE_X540_RAR_ENTRIES 128
38#define IXGBE_X540_MC_TBL_SIZE 128 38#define IXGBE_X540_MC_TBL_SIZE 128
39#define IXGBE_X540_VFT_TBL_SIZE 128 39#define IXGBE_X540_VFT_TBL_SIZE 128
40#define IXGBE_X540_RX_PB_SIZE 384 40#define IXGBE_X540_RX_PB_SIZE 384
41 41
42static s32 ixgbe_update_flash_X540(struct ixgbe_hw *hw); 42static s32 ixgbe_update_flash_X540(struct ixgbe_hw *hw);
43static s32 ixgbe_poll_flash_update_done_X540(struct ixgbe_hw *hw); 43static s32 ixgbe_poll_flash_update_done_X540(struct ixgbe_hw *hw);
diff --git a/drivers/net/ethernet/intel/ixgbevf/ethtool.c b/drivers/net/ethernet/intel/ixgbevf/ethtool.c
index c9d0c12d6f04..54d9acef9c4e 100644
--- a/drivers/net/ethernet/intel/ixgbevf/ethtool.c
+++ b/drivers/net/ethernet/intel/ixgbevf/ethtool.c
@@ -45,16 +45,27 @@
45 45
46struct ixgbe_stats { 46struct ixgbe_stats {
47 char stat_string[ETH_GSTRING_LEN]; 47 char stat_string[ETH_GSTRING_LEN];
48 int sizeof_stat; 48 struct {
49 int stat_offset; 49 int sizeof_stat;
50 int base_stat_offset; 50 int stat_offset;
51 int saved_reset_offset; 51 int base_stat_offset;
52 int saved_reset_offset;
53 };
52}; 54};
53 55
54#define IXGBEVF_STAT(m, b, r) sizeof(((struct ixgbevf_adapter *)0)->m), \ 56#define IXGBEVF_STAT(m, b, r) { \
55 offsetof(struct ixgbevf_adapter, m), \ 57 .sizeof_stat = FIELD_SIZEOF(struct ixgbevf_adapter, m), \
56 offsetof(struct ixgbevf_adapter, b), \ 58 .stat_offset = offsetof(struct ixgbevf_adapter, m), \
57 offsetof(struct ixgbevf_adapter, r) 59 .base_stat_offset = offsetof(struct ixgbevf_adapter, b), \
60 .saved_reset_offset = offsetof(struct ixgbevf_adapter, r) \
61}
62
63#define IXGBEVF_ZSTAT(m) { \
64 .sizeof_stat = FIELD_SIZEOF(struct ixgbevf_adapter, m), \
65 .stat_offset = offsetof(struct ixgbevf_adapter, m), \
66 .base_stat_offset = -1, \
67 .saved_reset_offset = -1 \
68}
58 69
59static const struct ixgbe_stats ixgbe_gstrings_stats[] = { 70static const struct ixgbe_stats ixgbe_gstrings_stats[] = {
60 {"rx_packets", IXGBEVF_STAT(stats.vfgprc, stats.base_vfgprc, 71 {"rx_packets", IXGBEVF_STAT(stats.vfgprc, stats.base_vfgprc,
@@ -65,15 +76,20 @@ static const struct ixgbe_stats ixgbe_gstrings_stats[] = {
65 stats.saved_reset_vfgorc)}, 76 stats.saved_reset_vfgorc)},
66 {"tx_bytes", IXGBEVF_STAT(stats.vfgotc, stats.base_vfgotc, 77 {"tx_bytes", IXGBEVF_STAT(stats.vfgotc, stats.base_vfgotc,
67 stats.saved_reset_vfgotc)}, 78 stats.saved_reset_vfgotc)},
68 {"tx_busy", IXGBEVF_STAT(tx_busy, zero_base, zero_base)}, 79 {"tx_busy", IXGBEVF_ZSTAT(tx_busy)},
69 {"multicast", IXGBEVF_STAT(stats.vfmprc, stats.base_vfmprc, 80 {"multicast", IXGBEVF_STAT(stats.vfmprc, stats.base_vfmprc,
70 stats.saved_reset_vfmprc)}, 81 stats.saved_reset_vfmprc)},
71 {"rx_csum_offload_good", IXGBEVF_STAT(hw_csum_rx_good, zero_base, 82 {"rx_csum_offload_good", IXGBEVF_ZSTAT(hw_csum_rx_good)},
72 zero_base)}, 83 {"rx_csum_offload_errors", IXGBEVF_ZSTAT(hw_csum_rx_error)},
73 {"rx_csum_offload_errors", IXGBEVF_STAT(hw_csum_rx_error, zero_base, 84 {"tx_csum_offload_ctxt", IXGBEVF_ZSTAT(hw_csum_tx_good)},
74 zero_base)}, 85#ifdef BP_EXTENDED_STATS
75 {"tx_csum_offload_ctxt", IXGBEVF_STAT(hw_csum_tx_good, zero_base, 86 {"rx_bp_poll_yield", IXGBEVF_ZSTAT(bp_rx_yields)},
76 zero_base)}, 87 {"rx_bp_cleaned", IXGBEVF_ZSTAT(bp_rx_cleaned)},
88 {"rx_bp_misses", IXGBEVF_ZSTAT(bp_rx_missed)},
89 {"tx_bp_napi_yield", IXGBEVF_ZSTAT(bp_tx_yields)},
90 {"tx_bp_cleaned", IXGBEVF_ZSTAT(bp_tx_cleaned)},
91 {"tx_bp_misses", IXGBEVF_ZSTAT(bp_tx_missed)},
92#endif
77}; 93};
78 94
79#define IXGBE_QUEUE_STATS_LEN 0 95#define IXGBE_QUEUE_STATS_LEN 0
@@ -140,58 +156,10 @@ static void ixgbevf_set_msglevel(struct net_device *netdev, u32 data)
140 156
141#define IXGBE_GET_STAT(_A_, _R_) (_A_->stats._R_) 157#define IXGBE_GET_STAT(_A_, _R_) (_A_->stats._R_)
142 158
143static char *ixgbevf_reg_names[] = {
144 "IXGBE_VFCTRL",
145 "IXGBE_VFSTATUS",
146 "IXGBE_VFLINKS",
147 "IXGBE_VFRXMEMWRAP",
148 "IXGBE_VFFRTIMER",
149 "IXGBE_VTEICR",
150 "IXGBE_VTEICS",
151 "IXGBE_VTEIMS",
152 "IXGBE_VTEIMC",
153 "IXGBE_VTEIAC",
154 "IXGBE_VTEIAM",
155 "IXGBE_VTEITR",
156 "IXGBE_VTIVAR",
157 "IXGBE_VTIVAR_MISC",
158 "IXGBE_VFRDBAL0",
159 "IXGBE_VFRDBAL1",
160 "IXGBE_VFRDBAH0",
161 "IXGBE_VFRDBAH1",
162 "IXGBE_VFRDLEN0",
163 "IXGBE_VFRDLEN1",
164 "IXGBE_VFRDH0",
165 "IXGBE_VFRDH1",
166 "IXGBE_VFRDT0",
167 "IXGBE_VFRDT1",
168 "IXGBE_VFRXDCTL0",
169 "IXGBE_VFRXDCTL1",
170 "IXGBE_VFSRRCTL0",
171 "IXGBE_VFSRRCTL1",
172 "IXGBE_VFPSRTYPE",
173 "IXGBE_VFTDBAL0",
174 "IXGBE_VFTDBAL1",
175 "IXGBE_VFTDBAH0",
176 "IXGBE_VFTDBAH1",
177 "IXGBE_VFTDLEN0",
178 "IXGBE_VFTDLEN1",
179 "IXGBE_VFTDH0",
180 "IXGBE_VFTDH1",
181 "IXGBE_VFTDT0",
182 "IXGBE_VFTDT1",
183 "IXGBE_VFTXDCTL0",
184 "IXGBE_VFTXDCTL1",
185 "IXGBE_VFTDWBAL0",
186 "IXGBE_VFTDWBAL1",
187 "IXGBE_VFTDWBAH0",
188 "IXGBE_VFTDWBAH1"
189};
190
191
192static int ixgbevf_get_regs_len(struct net_device *netdev) 159static int ixgbevf_get_regs_len(struct net_device *netdev)
193{ 160{
194 return (ARRAY_SIZE(ixgbevf_reg_names)) * sizeof(u32); 161#define IXGBE_REGS_LEN 45
162 return IXGBE_REGS_LEN * sizeof(u32);
195} 163}
196 164
197static void ixgbevf_get_regs(struct net_device *netdev, 165static void ixgbevf_get_regs(struct net_device *netdev,
@@ -264,9 +232,6 @@ static void ixgbevf_get_regs(struct net_device *netdev,
264 regs_buff[41 + i] = IXGBE_READ_REG(hw, IXGBE_VFTDWBAL(i)); 232 regs_buff[41 + i] = IXGBE_READ_REG(hw, IXGBE_VFTDWBAL(i));
265 for (i = 0; i < 2; i++) 233 for (i = 0; i < 2; i++)
266 regs_buff[43 + i] = IXGBE_READ_REG(hw, IXGBE_VFTDWBAH(i)); 234 regs_buff[43 + i] = IXGBE_READ_REG(hw, IXGBE_VFTDWBAH(i));
267
268 for (i = 0; i < ARRAY_SIZE(ixgbevf_reg_names); i++)
269 hw_dbg(hw, "%s\t%8.8x\n", ixgbevf_reg_names[i], regs_buff[i]);
270} 235}
271 236
272static void ixgbevf_get_drvinfo(struct net_device *netdev, 237static void ixgbevf_get_drvinfo(struct net_device *netdev,
@@ -441,22 +406,50 @@ static void ixgbevf_get_ethtool_stats(struct net_device *netdev,
441 struct ethtool_stats *stats, u64 *data) 406 struct ethtool_stats *stats, u64 *data)
442{ 407{
443 struct ixgbevf_adapter *adapter = netdev_priv(netdev); 408 struct ixgbevf_adapter *adapter = netdev_priv(netdev);
409 char *base = (char *) adapter;
444 int i; 410 int i;
411#ifdef BP_EXTENDED_STATS
412 u64 rx_yields = 0, rx_cleaned = 0, rx_missed = 0,
413 tx_yields = 0, tx_cleaned = 0, tx_missed = 0;
414
415 for (i = 0; i < adapter->num_rx_queues; i++) {
416 rx_yields += adapter->rx_ring[i].bp_yields;
417 rx_cleaned += adapter->rx_ring[i].bp_cleaned;
418 rx_yields += adapter->rx_ring[i].bp_yields;
419 }
420
421 for (i = 0; i < adapter->num_tx_queues; i++) {
422 tx_yields += adapter->tx_ring[i].bp_yields;
423 tx_cleaned += adapter->tx_ring[i].bp_cleaned;
424 tx_yields += adapter->tx_ring[i].bp_yields;
425 }
426
427 adapter->bp_rx_yields = rx_yields;
428 adapter->bp_rx_cleaned = rx_cleaned;
429 adapter->bp_rx_missed = rx_missed;
430
431 adapter->bp_tx_yields = tx_yields;
432 adapter->bp_tx_cleaned = tx_cleaned;
433 adapter->bp_tx_missed = tx_missed;
434#endif
445 435
446 ixgbevf_update_stats(adapter); 436 ixgbevf_update_stats(adapter);
447 for (i = 0; i < IXGBE_GLOBAL_STATS_LEN; i++) { 437 for (i = 0; i < IXGBE_GLOBAL_STATS_LEN; i++) {
448 char *p = (char *)adapter + 438 char *p = base + ixgbe_gstrings_stats[i].stat_offset;
449 ixgbe_gstrings_stats[i].stat_offset; 439 char *b = base + ixgbe_gstrings_stats[i].base_stat_offset;
450 char *b = (char *)adapter + 440 char *r = base + ixgbe_gstrings_stats[i].saved_reset_offset;
451 ixgbe_gstrings_stats[i].base_stat_offset; 441
452 char *r = (char *)adapter + 442 if (ixgbe_gstrings_stats[i].sizeof_stat == sizeof(u64)) {
453 ixgbe_gstrings_stats[i].saved_reset_offset; 443 if (ixgbe_gstrings_stats[i].base_stat_offset >= 0)
454 data[i] = ((ixgbe_gstrings_stats[i].sizeof_stat == 444 data[i] = *(u64 *)p - *(u64 *)b + *(u64 *)r;
455 sizeof(u64)) ? *(u64 *)p : *(u32 *)p) - 445 else
456 ((ixgbe_gstrings_stats[i].sizeof_stat == 446 data[i] = *(u64 *)p;
457 sizeof(u64)) ? *(u64 *)b : *(u32 *)b) + 447 } else {
458 ((ixgbe_gstrings_stats[i].sizeof_stat == 448 if (ixgbe_gstrings_stats[i].base_stat_offset >= 0)
459 sizeof(u64)) ? *(u64 *)r : *(u32 *)r); 449 data[i] = *(u32 *)p - *(u32 *)b + *(u32 *)r;
450 else
451 data[i] = *(u32 *)p;
452 }
460 } 453 }
461} 454}
462 455
@@ -685,6 +678,85 @@ static int ixgbevf_nway_reset(struct net_device *netdev)
685 return 0; 678 return 0;
686} 679}
687 680
681static int ixgbevf_get_coalesce(struct net_device *netdev,
682 struct ethtool_coalesce *ec)
683{
684 struct ixgbevf_adapter *adapter = netdev_priv(netdev);
685
686 /* only valid if in constant ITR mode */
687 if (adapter->rx_itr_setting <= 1)
688 ec->rx_coalesce_usecs = adapter->rx_itr_setting;
689 else
690 ec->rx_coalesce_usecs = adapter->rx_itr_setting >> 2;
691
692 /* if in mixed tx/rx queues per vector mode, report only rx settings */
693 if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count)
694 return 0;
695
696 /* only valid if in constant ITR mode */
697 if (adapter->tx_itr_setting <= 1)
698 ec->tx_coalesce_usecs = adapter->tx_itr_setting;
699 else
700 ec->tx_coalesce_usecs = adapter->tx_itr_setting >> 2;
701
702 return 0;
703}
704
705static int ixgbevf_set_coalesce(struct net_device *netdev,
706 struct ethtool_coalesce *ec)
707{
708 struct ixgbevf_adapter *adapter = netdev_priv(netdev);
709 struct ixgbevf_q_vector *q_vector;
710 int num_vectors, i;
711 u16 tx_itr_param, rx_itr_param;
712
713 /* don't accept tx specific changes if we've got mixed RxTx vectors */
714 if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count
715 && ec->tx_coalesce_usecs)
716 return -EINVAL;
717
718
719 if ((ec->rx_coalesce_usecs > (IXGBE_MAX_EITR >> 2)) ||
720 (ec->tx_coalesce_usecs > (IXGBE_MAX_EITR >> 2)))
721 return -EINVAL;
722
723 if (ec->rx_coalesce_usecs > 1)
724 adapter->rx_itr_setting = ec->rx_coalesce_usecs << 2;
725 else
726 adapter->rx_itr_setting = ec->rx_coalesce_usecs;
727
728 if (adapter->rx_itr_setting == 1)
729 rx_itr_param = IXGBE_20K_ITR;
730 else
731 rx_itr_param = adapter->rx_itr_setting;
732
733
734 if (ec->tx_coalesce_usecs > 1)
735 adapter->tx_itr_setting = ec->tx_coalesce_usecs << 2;
736 else
737 adapter->tx_itr_setting = ec->tx_coalesce_usecs;
738
739 if (adapter->tx_itr_setting == 1)
740 tx_itr_param = IXGBE_10K_ITR;
741 else
742 tx_itr_param = adapter->tx_itr_setting;
743
744 num_vectors = adapter->num_msix_vectors - NON_Q_VECTORS;
745
746 for (i = 0; i < num_vectors; i++) {
747 q_vector = adapter->q_vector[i];
748 if (q_vector->tx.count && !q_vector->rx.count)
749 /* tx only */
750 q_vector->itr = tx_itr_param;
751 else
752 /* rx only or mixed */
753 q_vector->itr = rx_itr_param;
754 ixgbevf_write_eitr(q_vector);
755 }
756
757 return 0;
758}
759
688static const struct ethtool_ops ixgbevf_ethtool_ops = { 760static const struct ethtool_ops ixgbevf_ethtool_ops = {
689 .get_settings = ixgbevf_get_settings, 761 .get_settings = ixgbevf_get_settings,
690 .get_drvinfo = ixgbevf_get_drvinfo, 762 .get_drvinfo = ixgbevf_get_drvinfo,
@@ -700,6 +772,8 @@ static const struct ethtool_ops ixgbevf_ethtool_ops = {
700 .get_sset_count = ixgbevf_get_sset_count, 772 .get_sset_count = ixgbevf_get_sset_count,
701 .get_strings = ixgbevf_get_strings, 773 .get_strings = ixgbevf_get_strings,
702 .get_ethtool_stats = ixgbevf_get_ethtool_stats, 774 .get_ethtool_stats = ixgbevf_get_ethtool_stats,
775 .get_coalesce = ixgbevf_get_coalesce,
776 .set_coalesce = ixgbevf_set_coalesce,
703}; 777};
704 778
705void ixgbevf_set_ethtool_ops(struct net_device *netdev) 779void ixgbevf_set_ethtool_ops(struct net_device *netdev)
diff --git a/drivers/net/ethernet/intel/ixgbevf/ixgbevf.h b/drivers/net/ethernet/intel/ixgbevf/ixgbevf.h
index fff0d9867529..8971e2d0a984 100644
--- a/drivers/net/ethernet/intel/ixgbevf/ixgbevf.h
+++ b/drivers/net/ethernet/intel/ixgbevf/ixgbevf.h
@@ -38,6 +38,11 @@
38 38
39#include "vf.h" 39#include "vf.h"
40 40
41#ifdef CONFIG_NET_RX_BUSY_POLL
42#include <net/busy_poll.h>
43#define BP_EXTENDED_STATS
44#endif
45
41/* wrapper around a pointer to a socket buffer, 46/* wrapper around a pointer to a socket buffer,
42 * so a DMA handle can be stored along with the buffer */ 47 * so a DMA handle can be stored along with the buffer */
43struct ixgbevf_tx_buffer { 48struct ixgbevf_tx_buffer {
@@ -76,6 +81,11 @@ struct ixgbevf_ring {
76 struct u64_stats_sync syncp; 81 struct u64_stats_sync syncp;
77 u64 hw_csum_rx_error; 82 u64 hw_csum_rx_error;
78 u64 hw_csum_rx_good; 83 u64 hw_csum_rx_good;
84#ifdef BP_EXTENDED_STATS
85 u64 bp_yields;
86 u64 bp_misses;
87 u64 bp_cleaned;
88#endif
79 89
80 u16 head; 90 u16 head;
81 u16 tail; 91 u16 tail;
@@ -145,7 +155,118 @@ struct ixgbevf_q_vector {
145 struct napi_struct napi; 155 struct napi_struct napi;
146 struct ixgbevf_ring_container rx, tx; 156 struct ixgbevf_ring_container rx, tx;
147 char name[IFNAMSIZ + 9]; 157 char name[IFNAMSIZ + 9];
158#ifdef CONFIG_NET_RX_BUSY_POLL
159 unsigned int state;
160#define IXGBEVF_QV_STATE_IDLE 0
161#define IXGBEVF_QV_STATE_NAPI 1 /* NAPI owns this QV */
162#define IXGBEVF_QV_STATE_POLL 2 /* poll owns this QV */
163#define IXGBEVF_QV_STATE_DISABLED 4 /* QV is disabled */
164#define IXGBEVF_QV_OWNED (IXGBEVF_QV_STATE_NAPI | IXGBEVF_QV_STATE_POLL)
165#define IXGBEVF_QV_LOCKED (IXGBEVF_QV_OWNED | IXGBEVF_QV_STATE_DISABLED)
166#define IXGBEVF_QV_STATE_NAPI_YIELD 8 /* NAPI yielded this QV */
167#define IXGBEVF_QV_STATE_POLL_YIELD 16 /* poll yielded this QV */
168#define IXGBEVF_QV_YIELD (IXGBEVF_QV_STATE_NAPI_YIELD | IXGBEVF_QV_STATE_POLL_YIELD)
169#define IXGBEVF_QV_USER_PEND (IXGBEVF_QV_STATE_POLL | IXGBEVF_QV_STATE_POLL_YIELD)
170 spinlock_t lock;
171#endif /* CONFIG_NET_RX_BUSY_POLL */
148}; 172};
173#ifdef CONFIG_NET_RX_BUSY_POLL
174static inline void ixgbevf_qv_init_lock(struct ixgbevf_q_vector *q_vector)
175{
176
177 spin_lock_init(&q_vector->lock);
178 q_vector->state = IXGBEVF_QV_STATE_IDLE;
179}
180
181/* called from the device poll routine to get ownership of a q_vector */
182static inline bool ixgbevf_qv_lock_napi(struct ixgbevf_q_vector *q_vector)
183{
184 int rc = true;
185 spin_lock_bh(&q_vector->lock);
186 if (q_vector->state & IXGBEVF_QV_LOCKED) {
187 WARN_ON(q_vector->state & IXGBEVF_QV_STATE_NAPI);
188 q_vector->state |= IXGBEVF_QV_STATE_NAPI_YIELD;
189 rc = false;
190#ifdef BP_EXTENDED_STATS
191 q_vector->tx.ring->bp_yields++;
192#endif
193 } else {
194 /* we don't care if someone yielded */
195 q_vector->state = IXGBEVF_QV_STATE_NAPI;
196 }
197 spin_unlock_bh(&q_vector->lock);
198 return rc;
199}
200
201/* returns true is someone tried to get the qv while napi had it */
202static inline bool ixgbevf_qv_unlock_napi(struct ixgbevf_q_vector *q_vector)
203{
204 int rc = false;
205 spin_lock_bh(&q_vector->lock);
206 WARN_ON(q_vector->state & (IXGBEVF_QV_STATE_POLL |
207 IXGBEVF_QV_STATE_NAPI_YIELD));
208
209 if (q_vector->state & IXGBEVF_QV_STATE_POLL_YIELD)
210 rc = true;
211 /* reset state to idle, unless QV is disabled */
212 q_vector->state &= IXGBEVF_QV_STATE_DISABLED;
213 spin_unlock_bh(&q_vector->lock);
214 return rc;
215}
216
217/* called from ixgbevf_low_latency_poll() */
218static inline bool ixgbevf_qv_lock_poll(struct ixgbevf_q_vector *q_vector)
219{
220 int rc = true;
221 spin_lock_bh(&q_vector->lock);
222 if ((q_vector->state & IXGBEVF_QV_LOCKED)) {
223 q_vector->state |= IXGBEVF_QV_STATE_POLL_YIELD;
224 rc = false;
225#ifdef BP_EXTENDED_STATS
226 q_vector->rx.ring->bp_yields++;
227#endif
228 } else {
229 /* preserve yield marks */
230 q_vector->state |= IXGBEVF_QV_STATE_POLL;
231 }
232 spin_unlock_bh(&q_vector->lock);
233 return rc;
234}
235
236/* returns true if someone tried to get the qv while it was locked */
237static inline bool ixgbevf_qv_unlock_poll(struct ixgbevf_q_vector *q_vector)
238{
239 int rc = false;
240 spin_lock_bh(&q_vector->lock);
241 WARN_ON(q_vector->state & (IXGBEVF_QV_STATE_NAPI));
242
243 if (q_vector->state & IXGBEVF_QV_STATE_POLL_YIELD)
244 rc = true;
245 /* reset state to idle, unless QV is disabled */
246 q_vector->state &= IXGBEVF_QV_STATE_DISABLED;
247 spin_unlock_bh(&q_vector->lock);
248 return rc;
249}
250
251/* true if a socket is polling, even if it did not get the lock */
252static inline bool ixgbevf_qv_busy_polling(struct ixgbevf_q_vector *q_vector)
253{
254 WARN_ON(!(q_vector->state & IXGBEVF_QV_OWNED));
255 return q_vector->state & IXGBEVF_QV_USER_PEND;
256}
257
258/* false if QV is currently owned */
259static inline bool ixgbevf_qv_disable(struct ixgbevf_q_vector *q_vector)
260{
261 int rc = true;
262 spin_lock_bh(&q_vector->lock);
263 if (q_vector->state & IXGBEVF_QV_OWNED)
264 rc = false;
265 spin_unlock_bh(&q_vector->lock);
266 return rc;
267}
268
269#endif /* CONFIG_NET_RX_BUSY_POLL */
149 270
150/* 271/*
151 * microsecond values for various ITR rates shifted by 2 to fit itr register 272 * microsecond values for various ITR rates shifted by 2 to fit itr register
@@ -165,9 +286,13 @@ struct ixgbevf_q_vector {
165 ((_eitr) ? (1000000000 / ((_eitr) * 256)) : 8) 286 ((_eitr) ? (1000000000 / ((_eitr) * 256)) : 8)
166#define EITR_REG_TO_INTS_PER_SEC EITR_INTS_PER_SEC_TO_REG 287#define EITR_REG_TO_INTS_PER_SEC EITR_INTS_PER_SEC_TO_REG
167 288
168#define IXGBE_DESC_UNUSED(R) \ 289static inline u16 ixgbevf_desc_unused(struct ixgbevf_ring *ring)
169 ((((R)->next_to_clean > (R)->next_to_use) ? 0 : (R)->count) + \ 290{
170 (R)->next_to_clean - (R)->next_to_use - 1) 291 u16 ntc = ring->next_to_clean;
292 u16 ntu = ring->next_to_use;
293
294 return ((ntc > ntu) ? 0 : ring->count) + ntc - ntu - 1;
295}
171 296
172#define IXGBEVF_RX_DESC(R, i) \ 297#define IXGBEVF_RX_DESC(R, i) \
173 (&(((union ixgbe_adv_rx_desc *)((R)->desc))[i])) 298 (&(((union ixgbe_adv_rx_desc *)((R)->desc))[i]))
@@ -240,7 +365,6 @@ struct ixgbevf_adapter {
240 struct ixgbe_hw hw; 365 struct ixgbe_hw hw;
241 u16 msg_enable; 366 u16 msg_enable;
242 struct ixgbevf_hw_stats stats; 367 struct ixgbevf_hw_stats stats;
243 u64 zero_base;
244 /* Interrupt Throttle Rate */ 368 /* Interrupt Throttle Rate */
245 u32 eitr_param; 369 u32 eitr_param;
246 370
@@ -249,6 +373,16 @@ struct ixgbevf_adapter {
249 unsigned int tx_ring_count; 373 unsigned int tx_ring_count;
250 unsigned int rx_ring_count; 374 unsigned int rx_ring_count;
251 375
376#ifdef BP_EXTENDED_STATS
377 u64 bp_rx_yields;
378 u64 bp_rx_cleaned;
379 u64 bp_rx_missed;
380
381 u64 bp_tx_yields;
382 u64 bp_tx_cleaned;
383 u64 bp_tx_missed;
384#endif
385
252 u32 link_speed; 386 u32 link_speed;
253 bool link_up; 387 bool link_up;
254 388
@@ -281,27 +415,25 @@ extern const struct ixgbe_mbx_operations ixgbevf_mbx_ops;
281extern const char ixgbevf_driver_name[]; 415extern const char ixgbevf_driver_name[];
282extern const char ixgbevf_driver_version[]; 416extern const char ixgbevf_driver_version[];
283 417
284extern void ixgbevf_up(struct ixgbevf_adapter *adapter); 418void ixgbevf_up(struct ixgbevf_adapter *adapter);
285extern void ixgbevf_down(struct ixgbevf_adapter *adapter); 419void ixgbevf_down(struct ixgbevf_adapter *adapter);
286extern void ixgbevf_reinit_locked(struct ixgbevf_adapter *adapter); 420void ixgbevf_reinit_locked(struct ixgbevf_adapter *adapter);
287extern void ixgbevf_reset(struct ixgbevf_adapter *adapter); 421void ixgbevf_reset(struct ixgbevf_adapter *adapter);
288extern void ixgbevf_set_ethtool_ops(struct net_device *netdev); 422void ixgbevf_set_ethtool_ops(struct net_device *netdev);
289extern int ixgbevf_setup_rx_resources(struct ixgbevf_adapter *, 423int ixgbevf_setup_rx_resources(struct ixgbevf_adapter *, struct ixgbevf_ring *);
290 struct ixgbevf_ring *); 424int ixgbevf_setup_tx_resources(struct ixgbevf_adapter *, struct ixgbevf_ring *);
291extern int ixgbevf_setup_tx_resources(struct ixgbevf_adapter *, 425void ixgbevf_free_rx_resources(struct ixgbevf_adapter *, struct ixgbevf_ring *);
292 struct ixgbevf_ring *); 426void ixgbevf_free_tx_resources(struct ixgbevf_adapter *, struct ixgbevf_ring *);
293extern void ixgbevf_free_rx_resources(struct ixgbevf_adapter *, 427void ixgbevf_update_stats(struct ixgbevf_adapter *adapter);
294 struct ixgbevf_ring *); 428int ethtool_ioctl(struct ifreq *ifr);
295extern void ixgbevf_free_tx_resources(struct ixgbevf_adapter *, 429
296 struct ixgbevf_ring *); 430extern void ixgbevf_write_eitr(struct ixgbevf_q_vector *q_vector);
297extern void ixgbevf_update_stats(struct ixgbevf_adapter *adapter); 431
298extern int ethtool_ioctl(struct ifreq *ifr); 432void ixgbe_napi_add_all(struct ixgbevf_adapter *adapter);
299 433void ixgbe_napi_del_all(struct ixgbevf_adapter *adapter);
300extern void ixgbe_napi_add_all(struct ixgbevf_adapter *adapter);
301extern void ixgbe_napi_del_all(struct ixgbevf_adapter *adapter);
302 434
303#ifdef DEBUG 435#ifdef DEBUG
304extern char *ixgbevf_get_hw_dev_name(struct ixgbe_hw *hw); 436char *ixgbevf_get_hw_dev_name(struct ixgbe_hw *hw);
305#define hw_dbg(hw, format, arg...) \ 437#define hw_dbg(hw, format, arg...) \
306 printk(KERN_DEBUG "%s: " format, ixgbevf_get_hw_dev_name(hw), ##arg) 438 printk(KERN_DEBUG "%s: " format, ixgbevf_get_hw_dev_name(hw), ##arg)
307#else 439#else
diff --git a/drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c b/drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
index 59a62bbfb371..038bfc8b7616 100644
--- a/drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
+++ b/drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
@@ -58,7 +58,7 @@ const char ixgbevf_driver_name[] = "ixgbevf";
58static const char ixgbevf_driver_string[] = 58static const char ixgbevf_driver_string[] =
59 "Intel(R) 10 Gigabit PCI Express Virtual Function Network Driver"; 59 "Intel(R) 10 Gigabit PCI Express Virtual Function Network Driver";
60 60
61#define DRV_VERSION "2.7.12-k" 61#define DRV_VERSION "2.11.3-k"
62const char ixgbevf_driver_version[] = DRV_VERSION; 62const char ixgbevf_driver_version[] = DRV_VERSION;
63static char ixgbevf_copyright[] = 63static char ixgbevf_copyright[] =
64 "Copyright (c) 2009 - 2012 Intel Corporation."; 64 "Copyright (c) 2009 - 2012 Intel Corporation.";
@@ -251,7 +251,7 @@ static bool ixgbevf_clean_tx_irq(struct ixgbevf_q_vector *q_vector,
251 251
252#define TX_WAKE_THRESHOLD (DESC_NEEDED * 2) 252#define TX_WAKE_THRESHOLD (DESC_NEEDED * 2)
253 if (unlikely(count && netif_carrier_ok(tx_ring->netdev) && 253 if (unlikely(count && netif_carrier_ok(tx_ring->netdev) &&
254 (IXGBE_DESC_UNUSED(tx_ring) >= TX_WAKE_THRESHOLD))) { 254 (ixgbevf_desc_unused(tx_ring) >= TX_WAKE_THRESHOLD))) {
255 /* Make sure that anybody stopping the queue after this 255 /* Make sure that anybody stopping the queue after this
256 * sees the new next_to_clean. 256 * sees the new next_to_clean.
257 */ 257 */
@@ -300,6 +300,30 @@ static void ixgbevf_receive_skb(struct ixgbevf_q_vector *q_vector,
300} 300}
301 301
302/** 302/**
303 * ixgbevf_rx_skb - Helper function to determine proper Rx method
304 * @q_vector: structure containing interrupt and ring information
305 * @skb: packet to send up
306 * @status: hardware indication of status of receive
307 * @rx_desc: rx descriptor
308 **/
309static void ixgbevf_rx_skb(struct ixgbevf_q_vector *q_vector,
310 struct sk_buff *skb, u8 status,
311 union ixgbe_adv_rx_desc *rx_desc)
312{
313#ifdef CONFIG_NET_RX_BUSY_POLL
314 skb_mark_napi_id(skb, &q_vector->napi);
315
316 if (ixgbevf_qv_busy_polling(q_vector)) {
317 netif_receive_skb(skb);
318 /* exit early if we busy polled */
319 return;
320 }
321#endif /* CONFIG_NET_RX_BUSY_POLL */
322
323 ixgbevf_receive_skb(q_vector, skb, status, rx_desc);
324}
325
326/**
303 * ixgbevf_rx_checksum - indicate in skb if hw indicated a good cksum 327 * ixgbevf_rx_checksum - indicate in skb if hw indicated a good cksum
304 * @ring: pointer to Rx descriptor ring structure 328 * @ring: pointer to Rx descriptor ring structure
305 * @status_err: hardware indication of status of receive 329 * @status_err: hardware indication of status of receive
@@ -396,9 +420,9 @@ static inline void ixgbevf_irq_enable_queues(struct ixgbevf_adapter *adapter,
396 IXGBE_WRITE_REG(hw, IXGBE_VTEIMS, qmask); 420 IXGBE_WRITE_REG(hw, IXGBE_VTEIMS, qmask);
397} 421}
398 422
399static bool ixgbevf_clean_rx_irq(struct ixgbevf_q_vector *q_vector, 423static int ixgbevf_clean_rx_irq(struct ixgbevf_q_vector *q_vector,
400 struct ixgbevf_ring *rx_ring, 424 struct ixgbevf_ring *rx_ring,
401 int budget) 425 int budget)
402{ 426{
403 struct ixgbevf_adapter *adapter = q_vector->adapter; 427 struct ixgbevf_adapter *adapter = q_vector->adapter;
404 struct pci_dev *pdev = adapter->pdev; 428 struct pci_dev *pdev = adapter->pdev;
@@ -473,15 +497,6 @@ static bool ixgbevf_clean_rx_irq(struct ixgbevf_q_vector *q_vector,
473 total_rx_bytes += skb->len; 497 total_rx_bytes += skb->len;
474 total_rx_packets++; 498 total_rx_packets++;
475 499
476 /*
477 * Work around issue of some types of VM to VM loop back
478 * packets not getting split correctly
479 */
480 if (staterr & IXGBE_RXD_STAT_LB) {
481 u32 header_fixup_len = skb_headlen(skb);
482 if (header_fixup_len < 14)
483 skb_push(skb, header_fixup_len);
484 }
485 skb->protocol = eth_type_trans(skb, rx_ring->netdev); 500 skb->protocol = eth_type_trans(skb, rx_ring->netdev);
486 501
487 /* Workaround hardware that can't do proper VEPA multicast 502 /* Workaround hardware that can't do proper VEPA multicast
@@ -494,7 +509,7 @@ static bool ixgbevf_clean_rx_irq(struct ixgbevf_q_vector *q_vector,
494 goto next_desc; 509 goto next_desc;
495 } 510 }
496 511
497 ixgbevf_receive_skb(q_vector, skb, staterr, rx_desc); 512 ixgbevf_rx_skb(q_vector, skb, staterr, rx_desc);
498 513
499next_desc: 514next_desc:
500 rx_desc->wb.upper.status_error = 0; 515 rx_desc->wb.upper.status_error = 0;
@@ -514,7 +529,7 @@ next_desc:
514 } 529 }
515 530
516 rx_ring->next_to_clean = i; 531 rx_ring->next_to_clean = i;
517 cleaned_count = IXGBE_DESC_UNUSED(rx_ring); 532 cleaned_count = ixgbevf_desc_unused(rx_ring);
518 533
519 if (cleaned_count) 534 if (cleaned_count)
520 ixgbevf_alloc_rx_buffers(adapter, rx_ring, cleaned_count); 535 ixgbevf_alloc_rx_buffers(adapter, rx_ring, cleaned_count);
@@ -526,7 +541,7 @@ next_desc:
526 q_vector->rx.total_packets += total_rx_packets; 541 q_vector->rx.total_packets += total_rx_packets;
527 q_vector->rx.total_bytes += total_rx_bytes; 542 q_vector->rx.total_bytes += total_rx_bytes;
528 543
529 return !!budget; 544 return total_rx_packets;
530} 545}
531 546
532/** 547/**
@@ -549,6 +564,11 @@ static int ixgbevf_poll(struct napi_struct *napi, int budget)
549 ixgbevf_for_each_ring(ring, q_vector->tx) 564 ixgbevf_for_each_ring(ring, q_vector->tx)
550 clean_complete &= ixgbevf_clean_tx_irq(q_vector, ring); 565 clean_complete &= ixgbevf_clean_tx_irq(q_vector, ring);
551 566
567#ifdef CONFIG_NET_RX_BUSY_POLL
568 if (!ixgbevf_qv_lock_napi(q_vector))
569 return budget;
570#endif
571
552 /* attempt to distribute budget to each queue fairly, but don't allow 572 /* attempt to distribute budget to each queue fairly, but don't allow
553 * the budget to go below 1 because we'll exit polling */ 573 * the budget to go below 1 because we'll exit polling */
554 if (q_vector->rx.count > 1) 574 if (q_vector->rx.count > 1)
@@ -558,10 +578,15 @@ static int ixgbevf_poll(struct napi_struct *napi, int budget)
558 578
559 adapter->flags |= IXGBE_FLAG_IN_NETPOLL; 579 adapter->flags |= IXGBE_FLAG_IN_NETPOLL;
560 ixgbevf_for_each_ring(ring, q_vector->rx) 580 ixgbevf_for_each_ring(ring, q_vector->rx)
561 clean_complete &= ixgbevf_clean_rx_irq(q_vector, ring, 581 clean_complete &= (ixgbevf_clean_rx_irq(q_vector, ring,
562 per_ring_budget); 582 per_ring_budget)
583 < per_ring_budget);
563 adapter->flags &= ~IXGBE_FLAG_IN_NETPOLL; 584 adapter->flags &= ~IXGBE_FLAG_IN_NETPOLL;
564 585
586#ifdef CONFIG_NET_RX_BUSY_POLL
587 ixgbevf_qv_unlock_napi(q_vector);
588#endif
589
565 /* If all work not completed, return budget and keep polling */ 590 /* If all work not completed, return budget and keep polling */
566 if (!clean_complete) 591 if (!clean_complete)
567 return budget; 592 return budget;
@@ -580,7 +605,7 @@ static int ixgbevf_poll(struct napi_struct *napi, int budget)
580 * ixgbevf_write_eitr - write VTEITR register in hardware specific way 605 * ixgbevf_write_eitr - write VTEITR register in hardware specific way
581 * @q_vector: structure containing interrupt and ring information 606 * @q_vector: structure containing interrupt and ring information
582 */ 607 */
583static void ixgbevf_write_eitr(struct ixgbevf_q_vector *q_vector) 608void ixgbevf_write_eitr(struct ixgbevf_q_vector *q_vector)
584{ 609{
585 struct ixgbevf_adapter *adapter = q_vector->adapter; 610 struct ixgbevf_adapter *adapter = q_vector->adapter;
586 struct ixgbe_hw *hw = &adapter->hw; 611 struct ixgbe_hw *hw = &adapter->hw;
@@ -596,6 +621,40 @@ static void ixgbevf_write_eitr(struct ixgbevf_q_vector *q_vector)
596 IXGBE_WRITE_REG(hw, IXGBE_VTEITR(v_idx), itr_reg); 621 IXGBE_WRITE_REG(hw, IXGBE_VTEITR(v_idx), itr_reg);
597} 622}
598 623
624#ifdef CONFIG_NET_RX_BUSY_POLL
625/* must be called with local_bh_disable()d */
626static int ixgbevf_busy_poll_recv(struct napi_struct *napi)
627{
628 struct ixgbevf_q_vector *q_vector =
629 container_of(napi, struct ixgbevf_q_vector, napi);
630 struct ixgbevf_adapter *adapter = q_vector->adapter;
631 struct ixgbevf_ring *ring;
632 int found = 0;
633
634 if (test_bit(__IXGBEVF_DOWN, &adapter->state))
635 return LL_FLUSH_FAILED;
636
637 if (!ixgbevf_qv_lock_poll(q_vector))
638 return LL_FLUSH_BUSY;
639
640 ixgbevf_for_each_ring(ring, q_vector->rx) {
641 found = ixgbevf_clean_rx_irq(q_vector, ring, 4);
642#ifdef BP_EXTENDED_STATS
643 if (found)
644 ring->bp_cleaned += found;
645 else
646 ring->bp_misses++;
647#endif
648 if (found)
649 break;
650 }
651
652 ixgbevf_qv_unlock_poll(q_vector);
653
654 return found;
655}
656#endif /* CONFIG_NET_RX_BUSY_POLL */
657
599/** 658/**
600 * ixgbevf_configure_msix - Configure MSI-X hardware 659 * ixgbevf_configure_msix - Configure MSI-X hardware
601 * @adapter: board private structure 660 * @adapter: board private structure
@@ -756,37 +815,12 @@ static void ixgbevf_set_itr(struct ixgbevf_q_vector *q_vector)
756static irqreturn_t ixgbevf_msix_other(int irq, void *data) 815static irqreturn_t ixgbevf_msix_other(int irq, void *data)
757{ 816{
758 struct ixgbevf_adapter *adapter = data; 817 struct ixgbevf_adapter *adapter = data;
759 struct pci_dev *pdev = adapter->pdev;
760 struct ixgbe_hw *hw = &adapter->hw; 818 struct ixgbe_hw *hw = &adapter->hw;
761 u32 msg;
762 bool got_ack = false;
763 819
764 hw->mac.get_link_status = 1; 820 hw->mac.get_link_status = 1;
765 if (!hw->mbx.ops.check_for_ack(hw))
766 got_ack = true;
767
768 if (!hw->mbx.ops.check_for_msg(hw)) {
769 hw->mbx.ops.read(hw, &msg, 1);
770
771 if ((msg & IXGBE_MBVFICR_VFREQ_MASK) == IXGBE_PF_CONTROL_MSG) {
772 mod_timer(&adapter->watchdog_timer,
773 round_jiffies(jiffies + 1));
774 adapter->link_up = false;
775 }
776 821
777 if (msg & IXGBE_VT_MSGTYPE_NACK) 822 if (!test_bit(__IXGBEVF_DOWN, &adapter->state))
778 dev_info(&pdev->dev, 823 mod_timer(&adapter->watchdog_timer, jiffies);
779 "Last Request of type %2.2x to PF Nacked\n",
780 msg & 0xFF);
781 hw->mbx.v2p_mailbox |= IXGBE_VFMAILBOX_PFSTS;
782 }
783
784 /* checking for the ack clears the PFACK bit. Place
785 * it back in the v2p_mailbox cache so that anyone
786 * polling for an ack will not miss it
787 */
788 if (got_ack)
789 hw->mbx.v2p_mailbox |= IXGBE_VFMAILBOX_PFACK;
790 824
791 IXGBE_WRITE_REG(hw, IXGBE_VTEIMS, adapter->eims_other); 825 IXGBE_WRITE_REG(hw, IXGBE_VTEIMS, adapter->eims_other);
792 826
@@ -1107,6 +1141,21 @@ static void ixgbevf_configure_srrctl(struct ixgbevf_adapter *adapter, int index)
1107 IXGBE_WRITE_REG(hw, IXGBE_VFSRRCTL(index), srrctl); 1141 IXGBE_WRITE_REG(hw, IXGBE_VFSRRCTL(index), srrctl);
1108} 1142}
1109 1143
1144static void ixgbevf_setup_psrtype(struct ixgbevf_adapter *adapter)
1145{
1146 struct ixgbe_hw *hw = &adapter->hw;
1147
1148 /* PSRTYPE must be initialized in 82599 */
1149 u32 psrtype = IXGBE_PSRTYPE_TCPHDR | IXGBE_PSRTYPE_UDPHDR |
1150 IXGBE_PSRTYPE_IPV4HDR | IXGBE_PSRTYPE_IPV6HDR |
1151 IXGBE_PSRTYPE_L2HDR;
1152
1153 if (adapter->num_rx_queues > 1)
1154 psrtype |= 1 << 29;
1155
1156 IXGBE_WRITE_REG(hw, IXGBE_VFPSRTYPE, psrtype);
1157}
1158
1110static void ixgbevf_set_rx_buffer_len(struct ixgbevf_adapter *adapter) 1159static void ixgbevf_set_rx_buffer_len(struct ixgbevf_adapter *adapter)
1111{ 1160{
1112 struct ixgbe_hw *hw = &adapter->hw; 1161 struct ixgbe_hw *hw = &adapter->hw;
@@ -1154,8 +1203,7 @@ static void ixgbevf_configure_rx(struct ixgbevf_adapter *adapter)
1154 int i, j; 1203 int i, j;
1155 u32 rdlen; 1204 u32 rdlen;
1156 1205
1157 /* PSRTYPE must be initialized in 82599 */ 1206 ixgbevf_setup_psrtype(adapter);
1158 IXGBE_WRITE_REG(hw, IXGBE_VFPSRTYPE, 0);
1159 1207
1160 /* set_rx_buffer_len must be called before ring initialization */ 1208 /* set_rx_buffer_len must be called before ring initialization */
1161 ixgbevf_set_rx_buffer_len(adapter); 1209 ixgbevf_set_rx_buffer_len(adapter);
@@ -1293,6 +1341,9 @@ static void ixgbevf_napi_enable_all(struct ixgbevf_adapter *adapter)
1293 1341
1294 for (q_idx = 0; q_idx < q_vectors; q_idx++) { 1342 for (q_idx = 0; q_idx < q_vectors; q_idx++) {
1295 q_vector = adapter->q_vector[q_idx]; 1343 q_vector = adapter->q_vector[q_idx];
1344#ifdef CONFIG_NET_RX_BUSY_POLL
1345 ixgbevf_qv_init_lock(adapter->q_vector[q_idx]);
1346#endif
1296 napi_enable(&q_vector->napi); 1347 napi_enable(&q_vector->napi);
1297 } 1348 }
1298} 1349}
@@ -1306,6 +1357,12 @@ static void ixgbevf_napi_disable_all(struct ixgbevf_adapter *adapter)
1306 for (q_idx = 0; q_idx < q_vectors; q_idx++) { 1357 for (q_idx = 0; q_idx < q_vectors; q_idx++) {
1307 q_vector = adapter->q_vector[q_idx]; 1358 q_vector = adapter->q_vector[q_idx];
1308 napi_disable(&q_vector->napi); 1359 napi_disable(&q_vector->napi);
1360#ifdef CONFIG_NET_RX_BUSY_POLL
1361 while (!ixgbevf_qv_disable(adapter->q_vector[q_idx])) {
1362 pr_info("QV %d locked\n", q_idx);
1363 usleep_range(1000, 20000);
1364 }
1365#endif /* CONFIG_NET_RX_BUSY_POLL */
1309 } 1366 }
1310} 1367}
1311 1368
@@ -1323,31 +1380,55 @@ static void ixgbevf_configure(struct ixgbevf_adapter *adapter)
1323 for (i = 0; i < adapter->num_rx_queues; i++) { 1380 for (i = 0; i < adapter->num_rx_queues; i++) {
1324 struct ixgbevf_ring *ring = &adapter->rx_ring[i]; 1381 struct ixgbevf_ring *ring = &adapter->rx_ring[i];
1325 ixgbevf_alloc_rx_buffers(adapter, ring, 1382 ixgbevf_alloc_rx_buffers(adapter, ring,
1326 IXGBE_DESC_UNUSED(ring)); 1383 ixgbevf_desc_unused(ring));
1327 } 1384 }
1328} 1385}
1329 1386
1330#define IXGBE_MAX_RX_DESC_POLL 10 1387#define IXGBEVF_MAX_RX_DESC_POLL 10
1331static inline void ixgbevf_rx_desc_queue_enable(struct ixgbevf_adapter *adapter, 1388static void ixgbevf_rx_desc_queue_enable(struct ixgbevf_adapter *adapter,
1332 int rxr) 1389 int rxr)
1333{ 1390{
1334 struct ixgbe_hw *hw = &adapter->hw; 1391 struct ixgbe_hw *hw = &adapter->hw;
1392 int wait_loop = IXGBEVF_MAX_RX_DESC_POLL;
1393 u32 rxdctl;
1335 int j = adapter->rx_ring[rxr].reg_idx; 1394 int j = adapter->rx_ring[rxr].reg_idx;
1336 int k;
1337 1395
1338 for (k = 0; k < IXGBE_MAX_RX_DESC_POLL; k++) { 1396 do {
1339 if (IXGBE_READ_REG(hw, IXGBE_VFRXDCTL(j)) & IXGBE_RXDCTL_ENABLE) 1397 usleep_range(1000, 2000);
1340 break; 1398 rxdctl = IXGBE_READ_REG(hw, IXGBE_VFRXDCTL(j));
1341 else 1399 } while (--wait_loop && !(rxdctl & IXGBE_RXDCTL_ENABLE));
1342 msleep(1); 1400
1343 } 1401 if (!wait_loop)
1344 if (k >= IXGBE_MAX_RX_DESC_POLL) { 1402 hw_dbg(hw, "RXDCTL.ENABLE queue %d not set while polling\n",
1345 hw_dbg(hw, "RXDCTL.ENABLE on Rx queue %d " 1403 rxr);
1346 "not set within the polling period\n", rxr); 1404
1347 } 1405 ixgbevf_release_rx_desc(&adapter->hw, &adapter->rx_ring[rxr],
1406 (adapter->rx_ring[rxr].count - 1));
1407}
1408
1409static void ixgbevf_disable_rx_queue(struct ixgbevf_adapter *adapter,
1410 struct ixgbevf_ring *ring)
1411{
1412 struct ixgbe_hw *hw = &adapter->hw;
1413 int wait_loop = IXGBEVF_MAX_RX_DESC_POLL;
1414 u32 rxdctl;
1415 u8 reg_idx = ring->reg_idx;
1416
1417 rxdctl = IXGBE_READ_REG(hw, IXGBE_VFRXDCTL(reg_idx));
1418 rxdctl &= ~IXGBE_RXDCTL_ENABLE;
1419
1420 /* write value back with RXDCTL.ENABLE bit cleared */
1421 IXGBE_WRITE_REG(hw, IXGBE_VFRXDCTL(reg_idx), rxdctl);
1348 1422
1349 ixgbevf_release_rx_desc(hw, &adapter->rx_ring[rxr], 1423 /* the hardware may take up to 100us to really disable the rx queue */
1350 adapter->rx_ring[rxr].count - 1); 1424 do {
1425 udelay(10);
1426 rxdctl = IXGBE_READ_REG(hw, IXGBE_VFRXDCTL(reg_idx));
1427 } while (--wait_loop && (rxdctl & IXGBE_RXDCTL_ENABLE));
1428
1429 if (!wait_loop)
1430 hw_dbg(hw, "RXDCTL.ENABLE queue %d not cleared while polling\n",
1431 reg_idx);
1351} 1432}
1352 1433
1353static void ixgbevf_save_reset_stats(struct ixgbevf_adapter *adapter) 1434static void ixgbevf_save_reset_stats(struct ixgbevf_adapter *adapter)
@@ -1545,8 +1626,6 @@ void ixgbevf_up(struct ixgbevf_adapter *adapter)
1545{ 1626{
1546 struct ixgbe_hw *hw = &adapter->hw; 1627 struct ixgbe_hw *hw = &adapter->hw;
1547 1628
1548 ixgbevf_negotiate_api(adapter);
1549
1550 ixgbevf_reset_queues(adapter); 1629 ixgbevf_reset_queues(adapter);
1551 1630
1552 ixgbevf_configure(adapter); 1631 ixgbevf_configure(adapter);
@@ -1679,7 +1758,10 @@ void ixgbevf_down(struct ixgbevf_adapter *adapter)
1679 1758
1680 /* signal that we are down to the interrupt handler */ 1759 /* signal that we are down to the interrupt handler */
1681 set_bit(__IXGBEVF_DOWN, &adapter->state); 1760 set_bit(__IXGBEVF_DOWN, &adapter->state);
1682 /* disable receives */ 1761
1762 /* disable all enabled rx queues */
1763 for (i = 0; i < adapter->num_rx_queues; i++)
1764 ixgbevf_disable_rx_queue(adapter, &adapter->rx_ring[i]);
1683 1765
1684 netif_tx_disable(netdev); 1766 netif_tx_disable(netdev);
1685 1767
@@ -1733,10 +1815,12 @@ void ixgbevf_reset(struct ixgbevf_adapter *adapter)
1733 struct ixgbe_hw *hw = &adapter->hw; 1815 struct ixgbe_hw *hw = &adapter->hw;
1734 struct net_device *netdev = adapter->netdev; 1816 struct net_device *netdev = adapter->netdev;
1735 1817
1736 if (hw->mac.ops.reset_hw(hw)) 1818 if (hw->mac.ops.reset_hw(hw)) {
1737 hw_dbg(hw, "PF still resetting\n"); 1819 hw_dbg(hw, "PF still resetting\n");
1738 else 1820 } else {
1739 hw->mac.ops.init_hw(hw); 1821 hw->mac.ops.init_hw(hw);
1822 ixgbevf_negotiate_api(adapter);
1823 }
1740 1824
1741 if (is_valid_ether_addr(adapter->hw.mac.addr)) { 1825 if (is_valid_ether_addr(adapter->hw.mac.addr)) {
1742 memcpy(netdev->dev_addr, adapter->hw.mac.addr, 1826 memcpy(netdev->dev_addr, adapter->hw.mac.addr,
@@ -1929,6 +2013,9 @@ static int ixgbevf_alloc_q_vectors(struct ixgbevf_adapter *adapter)
1929 q_vector->v_idx = q_idx; 2013 q_vector->v_idx = q_idx;
1930 netif_napi_add(adapter->netdev, &q_vector->napi, 2014 netif_napi_add(adapter->netdev, &q_vector->napi,
1931 ixgbevf_poll, 64); 2015 ixgbevf_poll, 64);
2016#ifdef CONFIG_NET_RX_BUSY_POLL
2017 napi_hash_add(&q_vector->napi);
2018#endif
1932 adapter->q_vector[q_idx] = q_vector; 2019 adapter->q_vector[q_idx] = q_vector;
1933 } 2020 }
1934 2021
@@ -1938,6 +2025,9 @@ err_out:
1938 while (q_idx) { 2025 while (q_idx) {
1939 q_idx--; 2026 q_idx--;
1940 q_vector = adapter->q_vector[q_idx]; 2027 q_vector = adapter->q_vector[q_idx];
2028#ifdef CONFIG_NET_RX_BUSY_POLL
2029 napi_hash_del(&q_vector->napi);
2030#endif
1941 netif_napi_del(&q_vector->napi); 2031 netif_napi_del(&q_vector->napi);
1942 kfree(q_vector); 2032 kfree(q_vector);
1943 adapter->q_vector[q_idx] = NULL; 2033 adapter->q_vector[q_idx] = NULL;
@@ -1961,6 +2051,9 @@ static void ixgbevf_free_q_vectors(struct ixgbevf_adapter *adapter)
1961 struct ixgbevf_q_vector *q_vector = adapter->q_vector[q_idx]; 2051 struct ixgbevf_q_vector *q_vector = adapter->q_vector[q_idx];
1962 2052
1963 adapter->q_vector[q_idx] = NULL; 2053 adapter->q_vector[q_idx] = NULL;
2054#ifdef CONFIG_NET_RX_BUSY_POLL
2055 napi_hash_del(&q_vector->napi);
2056#endif
1964 netif_napi_del(&q_vector->napi); 2057 netif_napi_del(&q_vector->napi);
1965 kfree(q_vector); 2058 kfree(q_vector);
1966 } 2059 }
@@ -2072,6 +2165,9 @@ static int ixgbevf_sw_init(struct ixgbevf_adapter *adapter)
2072 hw->mac.max_tx_queues = 2; 2165 hw->mac.max_tx_queues = 2;
2073 hw->mac.max_rx_queues = 2; 2166 hw->mac.max_rx_queues = 2;
2074 2167
2168 /* lock to protect mailbox accesses */
2169 spin_lock_init(&adapter->mbx_lock);
2170
2075 err = hw->mac.ops.reset_hw(hw); 2171 err = hw->mac.ops.reset_hw(hw);
2076 if (err) { 2172 if (err) {
2077 dev_info(&pdev->dev, 2173 dev_info(&pdev->dev,
@@ -2082,6 +2178,7 @@ static int ixgbevf_sw_init(struct ixgbevf_adapter *adapter)
2082 pr_err("init_shared_code failed: %d\n", err); 2178 pr_err("init_shared_code failed: %d\n", err);
2083 goto out; 2179 goto out;
2084 } 2180 }
2181 ixgbevf_negotiate_api(adapter);
2085 err = hw->mac.ops.get_mac_addr(hw, hw->mac.addr); 2182 err = hw->mac.ops.get_mac_addr(hw, hw->mac.addr);
2086 if (err) 2183 if (err)
2087 dev_info(&pdev->dev, "Error reading MAC address\n"); 2184 dev_info(&pdev->dev, "Error reading MAC address\n");
@@ -2097,9 +2194,6 @@ static int ixgbevf_sw_init(struct ixgbevf_adapter *adapter)
2097 memcpy(hw->mac.addr, netdev->dev_addr, netdev->addr_len); 2194 memcpy(hw->mac.addr, netdev->dev_addr, netdev->addr_len);
2098 } 2195 }
2099 2196
2100 /* lock to protect mailbox accesses */
2101 spin_lock_init(&adapter->mbx_lock);
2102
2103 /* Enable dynamic interrupt throttling rates */ 2197 /* Enable dynamic interrupt throttling rates */
2104 adapter->rx_itr_setting = 1; 2198 adapter->rx_itr_setting = 1;
2105 adapter->tx_itr_setting = 1; 2199 adapter->tx_itr_setting = 1;
@@ -2620,8 +2714,6 @@ static int ixgbevf_open(struct net_device *netdev)
2620 } 2714 }
2621 } 2715 }
2622 2716
2623 ixgbevf_negotiate_api(adapter);
2624
2625 /* setup queue reg_idx and Rx queue count */ 2717 /* setup queue reg_idx and Rx queue count */
2626 err = ixgbevf_setup_queues(adapter); 2718 err = ixgbevf_setup_queues(adapter);
2627 if (err) 2719 if (err)
@@ -3010,7 +3102,7 @@ static int __ixgbevf_maybe_stop_tx(struct ixgbevf_ring *tx_ring, int size)
3010 3102
3011 /* We need to check again in a case another CPU has just 3103 /* We need to check again in a case another CPU has just
3012 * made room available. */ 3104 * made room available. */
3013 if (likely(IXGBE_DESC_UNUSED(tx_ring) < size)) 3105 if (likely(ixgbevf_desc_unused(tx_ring) < size))
3014 return -EBUSY; 3106 return -EBUSY;
3015 3107
3016 /* A reprieve! - use start_queue because it doesn't call schedule */ 3108 /* A reprieve! - use start_queue because it doesn't call schedule */
@@ -3021,7 +3113,7 @@ static int __ixgbevf_maybe_stop_tx(struct ixgbevf_ring *tx_ring, int size)
3021 3113
3022static int ixgbevf_maybe_stop_tx(struct ixgbevf_ring *tx_ring, int size) 3114static int ixgbevf_maybe_stop_tx(struct ixgbevf_ring *tx_ring, int size)
3023{ 3115{
3024 if (likely(IXGBE_DESC_UNUSED(tx_ring) >= size)) 3116 if (likely(ixgbevf_desc_unused(tx_ring) >= size))
3025 return 0; 3117 return 0;
3026 return __ixgbevf_maybe_stop_tx(tx_ring, size); 3118 return __ixgbevf_maybe_stop_tx(tx_ring, size);
3027} 3119}
@@ -3216,6 +3308,8 @@ static int ixgbevf_resume(struct pci_dev *pdev)
3216 } 3308 }
3217 pci_set_master(pdev); 3309 pci_set_master(pdev);
3218 3310
3311 ixgbevf_reset(adapter);
3312
3219 rtnl_lock(); 3313 rtnl_lock();
3220 err = ixgbevf_init_interrupt_scheme(adapter); 3314 err = ixgbevf_init_interrupt_scheme(adapter);
3221 rtnl_unlock(); 3315 rtnl_unlock();
@@ -3224,8 +3318,6 @@ static int ixgbevf_resume(struct pci_dev *pdev)
3224 return err; 3318 return err;
3225 } 3319 }
3226 3320
3227 ixgbevf_reset(adapter);
3228
3229 if (netif_running(netdev)) { 3321 if (netif_running(netdev)) {
3230 err = ixgbevf_open(netdev); 3322 err = ixgbevf_open(netdev);
3231 if (err) 3323 if (err)
@@ -3293,6 +3385,9 @@ static const struct net_device_ops ixgbevf_netdev_ops = {
3293 .ndo_tx_timeout = ixgbevf_tx_timeout, 3385 .ndo_tx_timeout = ixgbevf_tx_timeout,
3294 .ndo_vlan_rx_add_vid = ixgbevf_vlan_rx_add_vid, 3386 .ndo_vlan_rx_add_vid = ixgbevf_vlan_rx_add_vid,
3295 .ndo_vlan_rx_kill_vid = ixgbevf_vlan_rx_kill_vid, 3387 .ndo_vlan_rx_kill_vid = ixgbevf_vlan_rx_kill_vid,
3388#ifdef CONFIG_NET_RX_BUSY_POLL
3389 .ndo_busy_poll = ixgbevf_busy_poll_recv,
3390#endif
3296}; 3391};
3297 3392
3298static void ixgbevf_assign_netdev_ops(struct net_device *dev) 3393static void ixgbevf_assign_netdev_ops(struct net_device *dev)
diff --git a/drivers/net/ethernet/intel/ixgbevf/vf.c b/drivers/net/ethernet/intel/ixgbevf/vf.c
index 387b52635bc0..4d44d64ae387 100644
--- a/drivers/net/ethernet/intel/ixgbevf/vf.c
+++ b/drivers/net/ethernet/intel/ixgbevf/vf.c
@@ -242,7 +242,7 @@ static s32 ixgbevf_set_uc_addr_vf(struct ixgbe_hw *hw, u32 index, u8 *addr)
242 msgbuf[0] |= index << IXGBE_VT_MSGINFO_SHIFT; 242 msgbuf[0] |= index << IXGBE_VT_MSGINFO_SHIFT;
243 msgbuf[0] |= IXGBE_VF_SET_MACVLAN; 243 msgbuf[0] |= IXGBE_VF_SET_MACVLAN;
244 if (addr) 244 if (addr)
245 memcpy(msg_addr, addr, 6); 245 memcpy(msg_addr, addr, ETH_ALEN);
246 ret_val = mbx->ops.write_posted(hw, msgbuf, 3); 246 ret_val = mbx->ops.write_posted(hw, msgbuf, 3);
247 247
248 if (!ret_val) 248 if (!ret_val)
@@ -275,7 +275,7 @@ static s32 ixgbevf_set_rar_vf(struct ixgbe_hw *hw, u32 index, u8 *addr,
275 275
276 memset(msgbuf, 0, sizeof(msgbuf)); 276 memset(msgbuf, 0, sizeof(msgbuf));
277 msgbuf[0] = IXGBE_VF_SET_MAC_ADDR; 277 msgbuf[0] = IXGBE_VF_SET_MAC_ADDR;
278 memcpy(msg_addr, addr, 6); 278 memcpy(msg_addr, addr, ETH_ALEN);
279 ret_val = mbx->ops.write_posted(hw, msgbuf, 3); 279 ret_val = mbx->ops.write_posted(hw, msgbuf, 3);
280 280
281 if (!ret_val) 281 if (!ret_val)
diff --git a/drivers/net/ethernet/jme.c b/drivers/net/ethernet/jme.c
index 23de82a9da82..f5685c0d0579 100644
--- a/drivers/net/ethernet/jme.c
+++ b/drivers/net/ethernet/jme.c
@@ -309,7 +309,7 @@ static void
309jme_load_macaddr(struct net_device *netdev) 309jme_load_macaddr(struct net_device *netdev)
310{ 310{
311 struct jme_adapter *jme = netdev_priv(netdev); 311 struct jme_adapter *jme = netdev_priv(netdev);
312 unsigned char macaddr[6]; 312 unsigned char macaddr[ETH_ALEN];
313 u32 val; 313 u32 val;
314 314
315 spin_lock_bh(&jme->macaddr_lock); 315 spin_lock_bh(&jme->macaddr_lock);
@@ -321,7 +321,7 @@ jme_load_macaddr(struct net_device *netdev)
321 val = jread32(jme, JME_RXUMA_HI); 321 val = jread32(jme, JME_RXUMA_HI);
322 macaddr[4] = (val >> 0) & 0xFF; 322 macaddr[4] = (val >> 0) & 0xFF;
323 macaddr[5] = (val >> 8) & 0xFF; 323 macaddr[5] = (val >> 8) & 0xFF;
324 memcpy(netdev->dev_addr, macaddr, 6); 324 memcpy(netdev->dev_addr, macaddr, ETH_ALEN);
325 spin_unlock_bh(&jme->macaddr_lock); 325 spin_unlock_bh(&jme->macaddr_lock);
326} 326}
327 327
@@ -3192,7 +3192,6 @@ jme_init_one(struct pci_dev *pdev,
3192err_out_unmap: 3192err_out_unmap:
3193 iounmap(jme->regs); 3193 iounmap(jme->regs);
3194err_out_free_netdev: 3194err_out_free_netdev:
3195 pci_set_drvdata(pdev, NULL);
3196 free_netdev(netdev); 3195 free_netdev(netdev);
3197err_out_release_regions: 3196err_out_release_regions:
3198 pci_release_regions(pdev); 3197 pci_release_regions(pdev);
@@ -3210,7 +3209,6 @@ jme_remove_one(struct pci_dev *pdev)
3210 3209
3211 unregister_netdev(netdev); 3210 unregister_netdev(netdev);
3212 iounmap(jme->regs); 3211 iounmap(jme->regs);
3213 pci_set_drvdata(pdev, NULL);
3214 free_netdev(netdev); 3212 free_netdev(netdev);
3215 pci_release_regions(pdev); 3213 pci_release_regions(pdev);
3216 pci_disable_device(pdev); 3214 pci_disable_device(pdev);
diff --git a/drivers/net/ethernet/jme.h b/drivers/net/ethernet/jme.h
index 3efc897c9913..58cd67c0c8e4 100644
--- a/drivers/net/ethernet/jme.h
+++ b/drivers/net/ethernet/jme.h
@@ -28,7 +28,6 @@
28 28
29#define DRV_NAME "jme" 29#define DRV_NAME "jme"
30#define DRV_VERSION "1.0.8" 30#define DRV_VERSION "1.0.8"
31#define PFX DRV_NAME ": "
32 31
33#define PCI_DEVICE_ID_JMICRON_JMC250 0x0250 32#define PCI_DEVICE_ID_JMICRON_JMC250 0x0250
34#define PCI_DEVICE_ID_JMICRON_JMC260 0x0260 33#define PCI_DEVICE_ID_JMICRON_JMC260 0x0260
diff --git a/drivers/net/ethernet/korina.c b/drivers/net/ethernet/korina.c
index a36fa80968eb..4a5e3b0f712e 100644
--- a/drivers/net/ethernet/korina.c
+++ b/drivers/net/ethernet/korina.c
@@ -1110,7 +1110,7 @@ static int korina_probe(struct platform_device *pdev)
1110 lp = netdev_priv(dev); 1110 lp = netdev_priv(dev);
1111 1111
1112 bif->dev = dev; 1112 bif->dev = dev;
1113 memcpy(dev->dev_addr, bif->mac, 6); 1113 memcpy(dev->dev_addr, bif->mac, ETH_ALEN);
1114 1114
1115 lp->rx_irq = platform_get_irq_byname(pdev, "korina_rx"); 1115 lp->rx_irq = platform_get_irq_byname(pdev, "korina_rx");
1116 lp->tx_irq = platform_get_irq_byname(pdev, "korina_tx"); 1116 lp->tx_irq = platform_get_irq_byname(pdev, "korina_tx");
diff --git a/drivers/net/ethernet/marvell/mv643xx_eth.c b/drivers/net/ethernet/marvell/mv643xx_eth.c
index 2c210ec35d59..00cd36e08601 100644
--- a/drivers/net/ethernet/marvell/mv643xx_eth.c
+++ b/drivers/net/ethernet/marvell/mv643xx_eth.c
@@ -2513,7 +2513,7 @@ static int mv643xx_eth_shared_of_add_port(struct platform_device *pdev,
2513 2513
2514 mac_addr = of_get_mac_address(pnp); 2514 mac_addr = of_get_mac_address(pnp);
2515 if (mac_addr) 2515 if (mac_addr)
2516 memcpy(ppd.mac_addr, mac_addr, 6); 2516 memcpy(ppd.mac_addr, mac_addr, ETH_ALEN);
2517 2517
2518 mv643xx_eth_property(pnp, "tx-queue-size", ppd.tx_queue_size); 2518 mv643xx_eth_property(pnp, "tx-queue-size", ppd.tx_queue_size);
2519 mv643xx_eth_property(pnp, "tx-sram-addr", ppd.tx_sram_addr); 2519 mv643xx_eth_property(pnp, "tx-sram-addr", ppd.tx_sram_addr);
@@ -2696,7 +2696,7 @@ static void set_params(struct mv643xx_eth_private *mp,
2696 struct net_device *dev = mp->dev; 2696 struct net_device *dev = mp->dev;
2697 2697
2698 if (is_valid_ether_addr(pd->mac_addr)) 2698 if (is_valid_ether_addr(pd->mac_addr))
2699 memcpy(dev->dev_addr, pd->mac_addr, 6); 2699 memcpy(dev->dev_addr, pd->mac_addr, ETH_ALEN);
2700 else 2700 else
2701 uc_addr_get(mp, dev->dev_addr); 2701 uc_addr_get(mp, dev->dev_addr);
2702 2702
@@ -2890,6 +2890,7 @@ static int mv643xx_eth_probe(struct platform_device *pdev)
2890 PHY_INTERFACE_MODE_GMII); 2890 PHY_INTERFACE_MODE_GMII);
2891 if (!mp->phy) 2891 if (!mp->phy)
2892 err = -ENODEV; 2892 err = -ENODEV;
2893 phy_addr_set(mp, mp->phy->addr);
2893 } else if (pd->phy_addr != MV643XX_ETH_PHY_NONE) { 2894 } else if (pd->phy_addr != MV643XX_ETH_PHY_NONE) {
2894 mp->phy = phy_scan(mp, pd->phy_addr); 2895 mp->phy = phy_scan(mp, pd->phy_addr);
2895 2896
diff --git a/drivers/net/ethernet/marvell/mvmdio.c b/drivers/net/ethernet/marvell/mvmdio.c
index e2f662660313..7354960b583b 100644
--- a/drivers/net/ethernet/marvell/mvmdio.c
+++ b/drivers/net/ethernet/marvell/mvmdio.c
@@ -4,11 +4,9 @@
4 * Since the MDIO interface of Marvell network interfaces is shared 4 * Since the MDIO interface of Marvell network interfaces is shared
5 * between all network interfaces, having a single driver allows to 5 * between all network interfaces, having a single driver allows to
6 * handle concurrent accesses properly (you may have four Ethernet 6 * handle concurrent accesses properly (you may have four Ethernet
7 * ports, but they in fact share the same SMI interface to access the 7 * ports, but they in fact share the same SMI interface to access
8 * MDIO bus). Moreover, this MDIO interface code is similar between 8 * the MDIO bus). This driver is currently used by the mvneta and
9 * the mv643xx_eth driver and the mvneta driver. For now, it is only 9 * mv643xx_eth drivers.
10 * used by the mvneta driver, but it could later be used by the
11 * mv643xx_eth driver as well.
12 * 10 *
13 * Copyright (C) 2012 Marvell 11 * Copyright (C) 2012 Marvell
14 * 12 *
@@ -44,6 +42,15 @@
44#define MVMDIO_ERR_INT_SMI_DONE 0x00000010 42#define MVMDIO_ERR_INT_SMI_DONE 0x00000010
45#define MVMDIO_ERR_INT_MASK 0x0080 43#define MVMDIO_ERR_INT_MASK 0x0080
46 44
45/*
46 * SMI Timeout measurements:
47 * - Kirkwood 88F6281 (Globalscale Dreamplug): 45us to 95us (Interrupt)
48 * - Armada 370 (Globalscale Mirabox): 41us to 43us (Polled)
49 */
50#define MVMDIO_SMI_TIMEOUT 1000 /* 1000us = 1ms */
51#define MVMDIO_SMI_POLL_INTERVAL_MIN 45
52#define MVMDIO_SMI_POLL_INTERVAL_MAX 55
53
47struct orion_mdio_dev { 54struct orion_mdio_dev {
48 struct mutex lock; 55 struct mutex lock;
49 void __iomem *regs; 56 void __iomem *regs;
@@ -68,77 +75,68 @@ static int orion_mdio_smi_is_done(struct orion_mdio_dev *dev)
68static int orion_mdio_wait_ready(struct mii_bus *bus) 75static int orion_mdio_wait_ready(struct mii_bus *bus)
69{ 76{
70 struct orion_mdio_dev *dev = bus->priv; 77 struct orion_mdio_dev *dev = bus->priv;
71 int count; 78 unsigned long timeout = usecs_to_jiffies(MVMDIO_SMI_TIMEOUT);
72 79 unsigned long end = jiffies + timeout;
73 if (dev->err_interrupt <= 0) { 80 int timedout = 0;
74 count = 0; 81
75 while (1) { 82 while (1) {
76 if (orion_mdio_smi_is_done(dev)) 83 if (orion_mdio_smi_is_done(dev))
77 break; 84 return 0;
78 85 else if (timedout)
79 if (count > 100) { 86 break;
80 dev_err(bus->parent, 87
81 "Timeout: SMI busy for too long\n"); 88 if (dev->err_interrupt <= 0) {
82 return -ETIMEDOUT; 89 usleep_range(MVMDIO_SMI_POLL_INTERVAL_MIN,
83 } 90 MVMDIO_SMI_POLL_INTERVAL_MAX);
84 91
85 udelay(10); 92 if (time_is_before_jiffies(end))
86 count++; 93 ++timedout;
87 } 94 } else {
88 } else {
89 if (!orion_mdio_smi_is_done(dev)) {
90 wait_event_timeout(dev->smi_busy_wait, 95 wait_event_timeout(dev->smi_busy_wait,
91 orion_mdio_smi_is_done(dev), 96 orion_mdio_smi_is_done(dev),
92 msecs_to_jiffies(100)); 97 timeout);
93 if (!orion_mdio_smi_is_done(dev)) 98
94 return -ETIMEDOUT; 99 ++timedout;
95 } 100 }
96 } 101 }
97 102
98 return 0; 103 dev_err(bus->parent, "Timeout: SMI busy for too long\n");
104 return -ETIMEDOUT;
99} 105}
100 106
101static int orion_mdio_read(struct mii_bus *bus, int mii_id, 107static int orion_mdio_read(struct mii_bus *bus, int mii_id,
102 int regnum) 108 int regnum)
103{ 109{
104 struct orion_mdio_dev *dev = bus->priv; 110 struct orion_mdio_dev *dev = bus->priv;
105 int count;
106 u32 val; 111 u32 val;
107 int ret; 112 int ret;
108 113
109 mutex_lock(&dev->lock); 114 mutex_lock(&dev->lock);
110 115
111 ret = orion_mdio_wait_ready(bus); 116 ret = orion_mdio_wait_ready(bus);
112 if (ret < 0) { 117 if (ret < 0)
113 mutex_unlock(&dev->lock); 118 goto out;
114 return ret;
115 }
116 119
117 writel(((mii_id << MVMDIO_SMI_PHY_ADDR_SHIFT) | 120 writel(((mii_id << MVMDIO_SMI_PHY_ADDR_SHIFT) |
118 (regnum << MVMDIO_SMI_PHY_REG_SHIFT) | 121 (regnum << MVMDIO_SMI_PHY_REG_SHIFT) |
119 MVMDIO_SMI_READ_OPERATION), 122 MVMDIO_SMI_READ_OPERATION),
120 dev->regs); 123 dev->regs);
121 124
122 /* Wait for the value to become available */ 125 ret = orion_mdio_wait_ready(bus);
123 count = 0; 126 if (ret < 0)
124 while (1) { 127 goto out;
125 val = readl(dev->regs);
126 if (val & MVMDIO_SMI_READ_VALID)
127 break;
128
129 if (count > 100) {
130 dev_err(bus->parent, "Timeout when reading PHY\n");
131 mutex_unlock(&dev->lock);
132 return -ETIMEDOUT;
133 }
134 128
135 udelay(10); 129 val = readl(dev->regs);
136 count++; 130 if (!(val & MVMDIO_SMI_READ_VALID)) {
131 dev_err(bus->parent, "SMI bus read not valid\n");
132 ret = -ENODEV;
133 goto out;
137 } 134 }
138 135
136 ret = val & 0xFFFF;
137out:
139 mutex_unlock(&dev->lock); 138 mutex_unlock(&dev->lock);
140 139 return ret;
141 return val & 0xFFFF;
142} 140}
143 141
144static int orion_mdio_write(struct mii_bus *bus, int mii_id, 142static int orion_mdio_write(struct mii_bus *bus, int mii_id,
@@ -150,10 +148,8 @@ static int orion_mdio_write(struct mii_bus *bus, int mii_id,
150 mutex_lock(&dev->lock); 148 mutex_lock(&dev->lock);
151 149
152 ret = orion_mdio_wait_ready(bus); 150 ret = orion_mdio_wait_ready(bus);
153 if (ret < 0) { 151 if (ret < 0)
154 mutex_unlock(&dev->lock); 152 goto out;
155 return ret;
156 }
157 153
158 writel(((mii_id << MVMDIO_SMI_PHY_ADDR_SHIFT) | 154 writel(((mii_id << MVMDIO_SMI_PHY_ADDR_SHIFT) |
159 (regnum << MVMDIO_SMI_PHY_REG_SHIFT) | 155 (regnum << MVMDIO_SMI_PHY_REG_SHIFT) |
@@ -161,9 +157,9 @@ static int orion_mdio_write(struct mii_bus *bus, int mii_id,
161 (value << MVMDIO_SMI_DATA_SHIFT)), 157 (value << MVMDIO_SMI_DATA_SHIFT)),
162 dev->regs); 158 dev->regs);
163 159
160out:
164 mutex_unlock(&dev->lock); 161 mutex_unlock(&dev->lock);
165 162 return ret;
166 return 0;
167} 163}
168 164
169static int orion_mdio_reset(struct mii_bus *bus) 165static int orion_mdio_reset(struct mii_bus *bus)
diff --git a/drivers/net/ethernet/marvell/mvneta.c b/drivers/net/ethernet/marvell/mvneta.c
index e35bac7cfdf1..7d99e695a110 100644
--- a/drivers/net/ethernet/marvell/mvneta.c
+++ b/drivers/net/ethernet/marvell/mvneta.c
@@ -2811,7 +2811,7 @@ static int mvneta_probe(struct platform_device *pdev)
2811 } 2811 }
2812 2812
2813 dt_mac_addr = of_get_mac_address(dn); 2813 dt_mac_addr = of_get_mac_address(dn);
2814 if (dt_mac_addr && is_valid_ether_addr(dt_mac_addr)) { 2814 if (dt_mac_addr) {
2815 mac_from = "device tree"; 2815 mac_from = "device tree";
2816 memcpy(dev->dev_addr, dt_mac_addr, ETH_ALEN); 2816 memcpy(dev->dev_addr, dt_mac_addr, ETH_ALEN);
2817 } else { 2817 } else {
diff --git a/drivers/net/ethernet/marvell/skge.c b/drivers/net/ethernet/marvell/skge.c
index ecc7f7b696b8..597846193869 100644
--- a/drivers/net/ethernet/marvell/skge.c
+++ b/drivers/net/ethernet/marvell/skge.c
@@ -4046,7 +4046,6 @@ err_out_free_regions:
4046 pci_release_regions(pdev); 4046 pci_release_regions(pdev);
4047err_out_disable_pdev: 4047err_out_disable_pdev:
4048 pci_disable_device(pdev); 4048 pci_disable_device(pdev);
4049 pci_set_drvdata(pdev, NULL);
4050err_out: 4049err_out:
4051 return err; 4050 return err;
4052} 4051}
@@ -4090,7 +4089,6 @@ static void skge_remove(struct pci_dev *pdev)
4090 4089
4091 iounmap(hw->regs); 4090 iounmap(hw->regs);
4092 kfree(hw); 4091 kfree(hw);
4093 pci_set_drvdata(pdev, NULL);
4094} 4092}
4095 4093
4096#ifdef CONFIG_PM_SLEEP 4094#ifdef CONFIG_PM_SLEEP
diff --git a/drivers/net/ethernet/marvell/sky2.c b/drivers/net/ethernet/marvell/sky2.c
index e09a8c6f8536..a7df981d2123 100644
--- a/drivers/net/ethernet/marvell/sky2.c
+++ b/drivers/net/ethernet/marvell/sky2.c
@@ -5081,7 +5081,6 @@ err_out_free_regions:
5081err_out_disable: 5081err_out_disable:
5082 pci_disable_device(pdev); 5082 pci_disable_device(pdev);
5083err_out: 5083err_out:
5084 pci_set_drvdata(pdev, NULL);
5085 return err; 5084 return err;
5086} 5085}
5087 5086
@@ -5124,8 +5123,6 @@ static void sky2_remove(struct pci_dev *pdev)
5124 5123
5125 iounmap(hw->regs); 5124 iounmap(hw->regs);
5126 kfree(hw); 5125 kfree(hw);
5127
5128 pci_set_drvdata(pdev, NULL);
5129} 5126}
5130 5127
5131static int sky2_suspend(struct device *dev) 5128static int sky2_suspend(struct device *dev)
diff --git a/drivers/net/ethernet/mellanox/mlx4/cmd.c b/drivers/net/ethernet/mellanox/mlx4/cmd.c
index bb11624a1f39..1e9970d2f0f3 100644
--- a/drivers/net/ethernet/mellanox/mlx4/cmd.c
+++ b/drivers/net/ethernet/mellanox/mlx4/cmd.c
@@ -1539,11 +1539,6 @@ out:
1539 return ret; 1539 return ret;
1540} 1540}
1541 1541
1542static int calculate_transition(u16 oper_vlan, u16 admin_vlan)
1543{
1544 return (2 * (oper_vlan == MLX4_VGT) + (admin_vlan == MLX4_VGT));
1545}
1546
1547static int mlx4_master_immediate_activate_vlan_qos(struct mlx4_priv *priv, 1542static int mlx4_master_immediate_activate_vlan_qos(struct mlx4_priv *priv,
1548 int slave, int port) 1543 int slave, int port)
1549{ 1544{
@@ -1553,7 +1548,6 @@ static int mlx4_master_immediate_activate_vlan_qos(struct mlx4_priv *priv,
1553 struct mlx4_dev *dev = &(priv->dev); 1548 struct mlx4_dev *dev = &(priv->dev);
1554 int err; 1549 int err;
1555 int admin_vlan_ix = NO_INDX; 1550 int admin_vlan_ix = NO_INDX;
1556 enum mlx4_vlan_transition vlan_trans;
1557 1551
1558 vp_oper = &priv->mfunc.master.vf_oper[slave].vport[port]; 1552 vp_oper = &priv->mfunc.master.vf_oper[slave].vport[port];
1559 vp_admin = &priv->mfunc.master.vf_admin[slave].vport[port]; 1553 vp_admin = &priv->mfunc.master.vf_admin[slave].vport[port];
@@ -1563,12 +1557,8 @@ static int mlx4_master_immediate_activate_vlan_qos(struct mlx4_priv *priv,
1563 vp_oper->state.link_state == vp_admin->link_state) 1557 vp_oper->state.link_state == vp_admin->link_state)
1564 return 0; 1558 return 0;
1565 1559
1566 vlan_trans = calculate_transition(vp_oper->state.default_vlan,
1567 vp_admin->default_vlan);
1568
1569 if (!(priv->mfunc.master.slave_state[slave].active && 1560 if (!(priv->mfunc.master.slave_state[slave].active &&
1570 dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_UPDATE_QP && 1561 dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_UPDATE_QP)) {
1571 vlan_trans == MLX4_VLAN_TRANSITION_VST_VST)) {
1572 /* even if the UPDATE_QP command isn't supported, we still want 1562 /* even if the UPDATE_QP command isn't supported, we still want
1573 * to set this VF link according to the admin directive 1563 * to set this VF link according to the admin directive
1574 */ 1564 */
@@ -1586,15 +1576,19 @@ static int mlx4_master_immediate_activate_vlan_qos(struct mlx4_priv *priv,
1586 return -ENOMEM; 1576 return -ENOMEM;
1587 1577
1588 if (vp_oper->state.default_vlan != vp_admin->default_vlan) { 1578 if (vp_oper->state.default_vlan != vp_admin->default_vlan) {
1589 err = __mlx4_register_vlan(&priv->dev, port, 1579 if (MLX4_VGT != vp_admin->default_vlan) {
1590 vp_admin->default_vlan, 1580 err = __mlx4_register_vlan(&priv->dev, port,
1591 &admin_vlan_ix); 1581 vp_admin->default_vlan,
1592 if (err) { 1582 &admin_vlan_ix);
1593 kfree(work); 1583 if (err) {
1594 mlx4_warn((&priv->dev), 1584 kfree(work);
1595 "No vlan resources slave %d, port %d\n", 1585 mlx4_warn((&priv->dev),
1596 slave, port); 1586 "No vlan resources slave %d, port %d\n",
1597 return err; 1587 slave, port);
1588 return err;
1589 }
1590 } else {
1591 admin_vlan_ix = NO_INDX;
1598 } 1592 }
1599 work->flags |= MLX4_VF_IMMED_VLAN_FLAG_VLAN; 1593 work->flags |= MLX4_VF_IMMED_VLAN_FLAG_VLAN;
1600 mlx4_dbg((&(priv->dev)), 1594 mlx4_dbg((&(priv->dev)),
@@ -1687,7 +1681,7 @@ static void mlx4_master_deactivate_admin_state(struct mlx4_priv *priv, int slave
1687 vp_oper = &priv->mfunc.master.vf_oper[slave].vport[port]; 1681 vp_oper = &priv->mfunc.master.vf_oper[slave].vport[port];
1688 if (NO_INDX != vp_oper->vlan_idx) { 1682 if (NO_INDX != vp_oper->vlan_idx) {
1689 __mlx4_unregister_vlan(&priv->dev, 1683 __mlx4_unregister_vlan(&priv->dev,
1690 port, vp_oper->vlan_idx); 1684 port, vp_oper->state.default_vlan);
1691 vp_oper->vlan_idx = NO_INDX; 1685 vp_oper->vlan_idx = NO_INDX;
1692 } 1686 }
1693 if (NO_INDX != vp_oper->mac_idx) { 1687 if (NO_INDX != vp_oper->mac_idx) {
@@ -1718,6 +1712,7 @@ static void mlx4_master_do_cmd(struct mlx4_dev *dev, int slave, u8 cmd,
1718 if (cmd == MLX4_COMM_CMD_RESET) { 1712 if (cmd == MLX4_COMM_CMD_RESET) {
1719 mlx4_warn(dev, "Received reset from slave:%d\n", slave); 1713 mlx4_warn(dev, "Received reset from slave:%d\n", slave);
1720 slave_state[slave].active = false; 1714 slave_state[slave].active = false;
1715 slave_state[slave].old_vlan_api = false;
1721 mlx4_master_deactivate_admin_state(priv, slave); 1716 mlx4_master_deactivate_admin_state(priv, slave);
1722 for (i = 0; i < MLX4_EVENT_TYPES_NUM; ++i) { 1717 for (i = 0; i < MLX4_EVENT_TYPES_NUM; ++i) {
1723 slave_state[slave].event_eq[i].eqn = -1; 1718 slave_state[slave].event_eq[i].eqn = -1;
@@ -2198,6 +2193,8 @@ struct mlx4_cmd_mailbox *mlx4_alloc_cmd_mailbox(struct mlx4_dev *dev)
2198 return ERR_PTR(-ENOMEM); 2193 return ERR_PTR(-ENOMEM);
2199 } 2194 }
2200 2195
2196 memset(mailbox->buf, 0, MLX4_MAILBOX_SIZE);
2197
2201 return mailbox; 2198 return mailbox;
2202} 2199}
2203EXPORT_SYMBOL_GPL(mlx4_alloc_cmd_mailbox); 2200EXPORT_SYMBOL_GPL(mlx4_alloc_cmd_mailbox);
@@ -2253,7 +2250,6 @@ EXPORT_SYMBOL_GPL(mlx4_set_vf_mac);
2253int mlx4_set_vf_vlan(struct mlx4_dev *dev, int port, int vf, u16 vlan, u8 qos) 2250int mlx4_set_vf_vlan(struct mlx4_dev *dev, int port, int vf, u16 vlan, u8 qos)
2254{ 2251{
2255 struct mlx4_priv *priv = mlx4_priv(dev); 2252 struct mlx4_priv *priv = mlx4_priv(dev);
2256 struct mlx4_vport_oper_state *vf_oper;
2257 struct mlx4_vport_state *vf_admin; 2253 struct mlx4_vport_state *vf_admin;
2258 int slave; 2254 int slave;
2259 2255
@@ -2269,7 +2265,6 @@ int mlx4_set_vf_vlan(struct mlx4_dev *dev, int port, int vf, u16 vlan, u8 qos)
2269 return -EINVAL; 2265 return -EINVAL;
2270 2266
2271 vf_admin = &priv->mfunc.master.vf_admin[slave].vport[port]; 2267 vf_admin = &priv->mfunc.master.vf_admin[slave].vport[port];
2272 vf_oper = &priv->mfunc.master.vf_oper[slave].vport[port];
2273 2268
2274 if ((0 == vlan) && (0 == qos)) 2269 if ((0 == vlan) && (0 == qos))
2275 vf_admin->default_vlan = MLX4_VGT; 2270 vf_admin->default_vlan = MLX4_VGT;
diff --git a/drivers/net/ethernet/mellanox/mlx4/cq.c b/drivers/net/ethernet/mellanox/mlx4/cq.c
index 004e4231af67..22fcbe78311c 100644
--- a/drivers/net/ethernet/mellanox/mlx4/cq.c
+++ b/drivers/net/ethernet/mellanox/mlx4/cq.c
@@ -128,8 +128,6 @@ int mlx4_cq_modify(struct mlx4_dev *dev, struct mlx4_cq *cq,
128 return PTR_ERR(mailbox); 128 return PTR_ERR(mailbox);
129 129
130 cq_context = mailbox->buf; 130 cq_context = mailbox->buf;
131 memset(cq_context, 0, sizeof *cq_context);
132
133 cq_context->cq_max_count = cpu_to_be16(count); 131 cq_context->cq_max_count = cpu_to_be16(count);
134 cq_context->cq_period = cpu_to_be16(period); 132 cq_context->cq_period = cpu_to_be16(period);
135 133
@@ -153,8 +151,6 @@ int mlx4_cq_resize(struct mlx4_dev *dev, struct mlx4_cq *cq,
153 return PTR_ERR(mailbox); 151 return PTR_ERR(mailbox);
154 152
155 cq_context = mailbox->buf; 153 cq_context = mailbox->buf;
156 memset(cq_context, 0, sizeof *cq_context);
157
158 cq_context->logsize_usrpage = cpu_to_be32(ilog2(entries) << 24); 154 cq_context->logsize_usrpage = cpu_to_be32(ilog2(entries) << 24);
159 cq_context->log_page_size = mtt->page_shift - 12; 155 cq_context->log_page_size = mtt->page_shift - 12;
160 mtt_addr = mlx4_mtt_addr(dev, mtt); 156 mtt_addr = mlx4_mtt_addr(dev, mtt);
@@ -274,8 +270,6 @@ int mlx4_cq_alloc(struct mlx4_dev *dev, int nent,
274 } 270 }
275 271
276 cq_context = mailbox->buf; 272 cq_context = mailbox->buf;
277 memset(cq_context, 0, sizeof *cq_context);
278
279 cq_context->flags = cpu_to_be32(!!collapsed << 18); 273 cq_context->flags = cpu_to_be32(!!collapsed << 18);
280 if (timestamp_en) 274 if (timestamp_en)
281 cq_context->flags |= cpu_to_be32(1 << 19); 275 cq_context->flags |= cpu_to_be32(1 << 19);
diff --git a/drivers/net/ethernet/mellanox/mlx4/en_cq.c b/drivers/net/ethernet/mellanox/mlx4/en_cq.c
index 3e2d5047cdb3..3a098cc4d349 100644
--- a/drivers/net/ethernet/mellanox/mlx4/en_cq.c
+++ b/drivers/net/ethernet/mellanox/mlx4/en_cq.c
@@ -44,12 +44,23 @@ static void mlx4_en_cq_event(struct mlx4_cq *cq, enum mlx4_event event)
44 44
45 45
46int mlx4_en_create_cq(struct mlx4_en_priv *priv, 46int mlx4_en_create_cq(struct mlx4_en_priv *priv,
47 struct mlx4_en_cq *cq, 47 struct mlx4_en_cq **pcq,
48 int entries, int ring, enum cq_type mode) 48 int entries, int ring, enum cq_type mode,
49 int node)
49{ 50{
50 struct mlx4_en_dev *mdev = priv->mdev; 51 struct mlx4_en_dev *mdev = priv->mdev;
52 struct mlx4_en_cq *cq;
51 int err; 53 int err;
52 54
55 cq = kzalloc_node(sizeof(*cq), GFP_KERNEL, node);
56 if (!cq) {
57 cq = kzalloc(sizeof(*cq), GFP_KERNEL);
58 if (!cq) {
59 en_err(priv, "Failed to allocate CQ structure\n");
60 return -ENOMEM;
61 }
62 }
63
53 cq->size = entries; 64 cq->size = entries;
54 cq->buf_size = cq->size * mdev->dev->caps.cqe_size; 65 cq->buf_size = cq->size * mdev->dev->caps.cqe_size;
55 66
@@ -57,17 +68,30 @@ int mlx4_en_create_cq(struct mlx4_en_priv *priv,
57 cq->is_tx = mode; 68 cq->is_tx = mode;
58 spin_lock_init(&cq->lock); 69 spin_lock_init(&cq->lock);
59 70
71 /* Allocate HW buffers on provided NUMA node.
72 * dev->numa_node is used in mtt range allocation flow.
73 */
74 set_dev_node(&mdev->dev->pdev->dev, node);
60 err = mlx4_alloc_hwq_res(mdev->dev, &cq->wqres, 75 err = mlx4_alloc_hwq_res(mdev->dev, &cq->wqres,
61 cq->buf_size, 2 * PAGE_SIZE); 76 cq->buf_size, 2 * PAGE_SIZE);
77 set_dev_node(&mdev->dev->pdev->dev, mdev->dev->numa_node);
62 if (err) 78 if (err)
63 return err; 79 goto err_cq;
64 80
65 err = mlx4_en_map_buffer(&cq->wqres.buf); 81 err = mlx4_en_map_buffer(&cq->wqres.buf);
66 if (err) 82 if (err)
67 mlx4_free_hwq_res(mdev->dev, &cq->wqres, cq->buf_size); 83 goto err_res;
68 else
69 cq->buf = (struct mlx4_cqe *) cq->wqres.buf.direct.buf;
70 84
85 cq->buf = (struct mlx4_cqe *)cq->wqres.buf.direct.buf;
86 *pcq = cq;
87
88 return 0;
89
90err_res:
91 mlx4_free_hwq_res(mdev->dev, &cq->wqres, cq->buf_size);
92err_cq:
93 kfree(cq);
94 *pcq = NULL;
71 return err; 95 return err;
72} 96}
73 97
@@ -117,12 +141,12 @@ int mlx4_en_activate_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq,
117 struct mlx4_en_cq *rx_cq; 141 struct mlx4_en_cq *rx_cq;
118 142
119 cq_idx = cq_idx % priv->rx_ring_num; 143 cq_idx = cq_idx % priv->rx_ring_num;
120 rx_cq = &priv->rx_cq[cq_idx]; 144 rx_cq = priv->rx_cq[cq_idx];
121 cq->vector = rx_cq->vector; 145 cq->vector = rx_cq->vector;
122 } 146 }
123 147
124 if (!cq->is_tx) 148 if (!cq->is_tx)
125 cq->size = priv->rx_ring[cq->ring].actual_size; 149 cq->size = priv->rx_ring[cq->ring]->actual_size;
126 150
127 if ((cq->is_tx && priv->hwtstamp_config.tx_type) || 151 if ((cq->is_tx && priv->hwtstamp_config.tx_type) ||
128 (!cq->is_tx && priv->hwtstamp_config.rx_filter)) 152 (!cq->is_tx && priv->hwtstamp_config.rx_filter))
@@ -146,9 +170,10 @@ int mlx4_en_activate_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq,
146 return 0; 170 return 0;
147} 171}
148 172
149void mlx4_en_destroy_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq) 173void mlx4_en_destroy_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq **pcq)
150{ 174{
151 struct mlx4_en_dev *mdev = priv->mdev; 175 struct mlx4_en_dev *mdev = priv->mdev;
176 struct mlx4_en_cq *cq = *pcq;
152 177
153 mlx4_en_unmap_buffer(&cq->wqres.buf); 178 mlx4_en_unmap_buffer(&cq->wqres.buf);
154 mlx4_free_hwq_res(mdev->dev, &cq->wqres, cq->buf_size); 179 mlx4_free_hwq_res(mdev->dev, &cq->wqres, cq->buf_size);
@@ -157,6 +182,8 @@ void mlx4_en_destroy_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq)
157 cq->vector = 0; 182 cq->vector = 0;
158 cq->buf_size = 0; 183 cq->buf_size = 0;
159 cq->buf = NULL; 184 cq->buf = NULL;
185 kfree(cq);
186 *pcq = NULL;
160} 187}
161 188
162void mlx4_en_deactivate_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq) 189void mlx4_en_deactivate_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq)
diff --git a/drivers/net/ethernet/mellanox/mlx4/en_ethtool.c b/drivers/net/ethernet/mellanox/mlx4/en_ethtool.c
index 0c750985f47e..0596f9f85a0e 100644
--- a/drivers/net/ethernet/mellanox/mlx4/en_ethtool.c
+++ b/drivers/net/ethernet/mellanox/mlx4/en_ethtool.c
@@ -51,10 +51,10 @@ static int mlx4_en_moderation_update(struct mlx4_en_priv *priv)
51 int err = 0; 51 int err = 0;
52 52
53 for (i = 0; i < priv->tx_ring_num; i++) { 53 for (i = 0; i < priv->tx_ring_num; i++) {
54 priv->tx_cq[i].moder_cnt = priv->tx_frames; 54 priv->tx_cq[i]->moder_cnt = priv->tx_frames;
55 priv->tx_cq[i].moder_time = priv->tx_usecs; 55 priv->tx_cq[i]->moder_time = priv->tx_usecs;
56 if (priv->port_up) { 56 if (priv->port_up) {
57 err = mlx4_en_set_cq_moder(priv, &priv->tx_cq[i]); 57 err = mlx4_en_set_cq_moder(priv, priv->tx_cq[i]);
58 if (err) 58 if (err)
59 return err; 59 return err;
60 } 60 }
@@ -64,11 +64,11 @@ static int mlx4_en_moderation_update(struct mlx4_en_priv *priv)
64 return 0; 64 return 0;
65 65
66 for (i = 0; i < priv->rx_ring_num; i++) { 66 for (i = 0; i < priv->rx_ring_num; i++) {
67 priv->rx_cq[i].moder_cnt = priv->rx_frames; 67 priv->rx_cq[i]->moder_cnt = priv->rx_frames;
68 priv->rx_cq[i].moder_time = priv->rx_usecs; 68 priv->rx_cq[i]->moder_time = priv->rx_usecs;
69 priv->last_moder_time[i] = MLX4_EN_AUTO_CONF; 69 priv->last_moder_time[i] = MLX4_EN_AUTO_CONF;
70 if (priv->port_up) { 70 if (priv->port_up) {
71 err = mlx4_en_set_cq_moder(priv, &priv->rx_cq[i]); 71 err = mlx4_en_set_cq_moder(priv, priv->rx_cq[i]);
72 if (err) 72 if (err)
73 return err; 73 return err;
74 } 74 }
@@ -274,16 +274,16 @@ static void mlx4_en_get_ethtool_stats(struct net_device *dev,
274 } 274 }
275 } 275 }
276 for (i = 0; i < priv->tx_ring_num; i++) { 276 for (i = 0; i < priv->tx_ring_num; i++) {
277 data[index++] = priv->tx_ring[i].packets; 277 data[index++] = priv->tx_ring[i]->packets;
278 data[index++] = priv->tx_ring[i].bytes; 278 data[index++] = priv->tx_ring[i]->bytes;
279 } 279 }
280 for (i = 0; i < priv->rx_ring_num; i++) { 280 for (i = 0; i < priv->rx_ring_num; i++) {
281 data[index++] = priv->rx_ring[i].packets; 281 data[index++] = priv->rx_ring[i]->packets;
282 data[index++] = priv->rx_ring[i].bytes; 282 data[index++] = priv->rx_ring[i]->bytes;
283#ifdef CONFIG_NET_RX_BUSY_POLL 283#ifdef CONFIG_NET_RX_BUSY_POLL
284 data[index++] = priv->rx_ring[i].yields; 284 data[index++] = priv->rx_ring[i]->yields;
285 data[index++] = priv->rx_ring[i].misses; 285 data[index++] = priv->rx_ring[i]->misses;
286 data[index++] = priv->rx_ring[i].cleaned; 286 data[index++] = priv->rx_ring[i]->cleaned;
287#endif 287#endif
288 } 288 }
289 spin_unlock_bh(&priv->stats_lock); 289 spin_unlock_bh(&priv->stats_lock);
@@ -510,9 +510,9 @@ static int mlx4_en_set_ringparam(struct net_device *dev,
510 tx_size = max_t(u32, tx_size, MLX4_EN_MIN_TX_SIZE); 510 tx_size = max_t(u32, tx_size, MLX4_EN_MIN_TX_SIZE);
511 tx_size = min_t(u32, tx_size, MLX4_EN_MAX_TX_SIZE); 511 tx_size = min_t(u32, tx_size, MLX4_EN_MAX_TX_SIZE);
512 512
513 if (rx_size == (priv->port_up ? priv->rx_ring[0].actual_size : 513 if (rx_size == (priv->port_up ? priv->rx_ring[0]->actual_size :
514 priv->rx_ring[0].size) && 514 priv->rx_ring[0]->size) &&
515 tx_size == priv->tx_ring[0].size) 515 tx_size == priv->tx_ring[0]->size)
516 return 0; 516 return 0;
517 517
518 mutex_lock(&mdev->state_lock); 518 mutex_lock(&mdev->state_lock);
@@ -553,8 +553,8 @@ static void mlx4_en_get_ringparam(struct net_device *dev,
553 param->rx_max_pending = MLX4_EN_MAX_RX_SIZE; 553 param->rx_max_pending = MLX4_EN_MAX_RX_SIZE;
554 param->tx_max_pending = MLX4_EN_MAX_TX_SIZE; 554 param->tx_max_pending = MLX4_EN_MAX_TX_SIZE;
555 param->rx_pending = priv->port_up ? 555 param->rx_pending = priv->port_up ?
556 priv->rx_ring[0].actual_size : priv->rx_ring[0].size; 556 priv->rx_ring[0]->actual_size : priv->rx_ring[0]->size;
557 param->tx_pending = priv->tx_ring[0].size; 557 param->tx_pending = priv->tx_ring[0]->size;
558} 558}
559 559
560static u32 mlx4_en_get_rxfh_indir_size(struct net_device *dev) 560static u32 mlx4_en_get_rxfh_indir_size(struct net_device *dev)
diff --git a/drivers/net/ethernet/mellanox/mlx4/en_main.c b/drivers/net/ethernet/mellanox/mlx4/en_main.c
index a071cda2dd04..0d087b03a7b0 100644
--- a/drivers/net/ethernet/mellanox/mlx4/en_main.c
+++ b/drivers/net/ethernet/mellanox/mlx4/en_main.c
@@ -264,6 +264,10 @@ static void *mlx4_en_add(struct mlx4_dev *dev)
264 mlx4_foreach_port(i, dev, MLX4_PORT_TYPE_ETH) 264 mlx4_foreach_port(i, dev, MLX4_PORT_TYPE_ETH)
265 mdev->port_cnt++; 265 mdev->port_cnt++;
266 266
267 /* Initialize time stamp mechanism */
268 if (mdev->dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_TS)
269 mlx4_en_init_timestamp(mdev);
270
267 mlx4_foreach_port(i, dev, MLX4_PORT_TYPE_ETH) { 271 mlx4_foreach_port(i, dev, MLX4_PORT_TYPE_ETH) {
268 if (!dev->caps.comp_pool) { 272 if (!dev->caps.comp_pool) {
269 mdev->profile.prof[i].rx_ring_num = 273 mdev->profile.prof[i].rx_ring_num =
@@ -301,10 +305,6 @@ static void *mlx4_en_add(struct mlx4_dev *dev)
301 mdev->pndev[i] = NULL; 305 mdev->pndev[i] = NULL;
302 } 306 }
303 307
304 /* Initialize time stamp mechanism */
305 if (mdev->dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_TS)
306 mlx4_en_init_timestamp(mdev);
307
308 return mdev; 308 return mdev;
309 309
310err_mr: 310err_mr:
diff --git a/drivers/net/ethernet/mellanox/mlx4/en_netdev.c b/drivers/net/ethernet/mellanox/mlx4/en_netdev.c
index fa37b7a61213..e72d8a112a6b 100644
--- a/drivers/net/ethernet/mellanox/mlx4/en_netdev.c
+++ b/drivers/net/ethernet/mellanox/mlx4/en_netdev.c
@@ -75,7 +75,7 @@ static int mlx4_en_low_latency_recv(struct napi_struct *napi)
75 struct mlx4_en_cq *cq = container_of(napi, struct mlx4_en_cq, napi); 75 struct mlx4_en_cq *cq = container_of(napi, struct mlx4_en_cq, napi);
76 struct net_device *dev = cq->dev; 76 struct net_device *dev = cq->dev;
77 struct mlx4_en_priv *priv = netdev_priv(dev); 77 struct mlx4_en_priv *priv = netdev_priv(dev);
78 struct mlx4_en_rx_ring *rx_ring = &priv->rx_ring[cq->ring]; 78 struct mlx4_en_rx_ring *rx_ring = priv->rx_ring[cq->ring];
79 int done; 79 int done;
80 80
81 if (!priv->port_up) 81 if (!priv->port_up)
@@ -102,6 +102,7 @@ struct mlx4_en_filter {
102 struct list_head next; 102 struct list_head next;
103 struct work_struct work; 103 struct work_struct work;
104 104
105 u8 ip_proto;
105 __be32 src_ip; 106 __be32 src_ip;
106 __be32 dst_ip; 107 __be32 dst_ip;
107 __be16 src_port; 108 __be16 src_port;
@@ -120,14 +121,26 @@ struct mlx4_en_filter {
120 121
121static void mlx4_en_filter_rfs_expire(struct mlx4_en_priv *priv); 122static void mlx4_en_filter_rfs_expire(struct mlx4_en_priv *priv);
122 123
124static enum mlx4_net_trans_rule_id mlx4_ip_proto_to_trans_rule_id(u8 ip_proto)
125{
126 switch (ip_proto) {
127 case IPPROTO_UDP:
128 return MLX4_NET_TRANS_RULE_ID_UDP;
129 case IPPROTO_TCP:
130 return MLX4_NET_TRANS_RULE_ID_TCP;
131 default:
132 return -EPROTONOSUPPORT;
133 }
134};
135
123static void mlx4_en_filter_work(struct work_struct *work) 136static void mlx4_en_filter_work(struct work_struct *work)
124{ 137{
125 struct mlx4_en_filter *filter = container_of(work, 138 struct mlx4_en_filter *filter = container_of(work,
126 struct mlx4_en_filter, 139 struct mlx4_en_filter,
127 work); 140 work);
128 struct mlx4_en_priv *priv = filter->priv; 141 struct mlx4_en_priv *priv = filter->priv;
129 struct mlx4_spec_list spec_tcp = { 142 struct mlx4_spec_list spec_tcp_udp = {
130 .id = MLX4_NET_TRANS_RULE_ID_TCP, 143 .id = mlx4_ip_proto_to_trans_rule_id(filter->ip_proto),
131 { 144 {
132 .tcp_udp = { 145 .tcp_udp = {
133 .dst_port = filter->dst_port, 146 .dst_port = filter->dst_port,
@@ -163,9 +176,14 @@ static void mlx4_en_filter_work(struct work_struct *work)
163 int rc; 176 int rc;
164 __be64 mac_mask = cpu_to_be64(MLX4_MAC_MASK << 16); 177 __be64 mac_mask = cpu_to_be64(MLX4_MAC_MASK << 16);
165 178
179 if (spec_tcp_udp.id < 0) {
180 en_warn(priv, "RFS: ignoring unsupported ip protocol (%d)\n",
181 filter->ip_proto);
182 goto ignore;
183 }
166 list_add_tail(&spec_eth.list, &rule.list); 184 list_add_tail(&spec_eth.list, &rule.list);
167 list_add_tail(&spec_ip.list, &rule.list); 185 list_add_tail(&spec_ip.list, &rule.list);
168 list_add_tail(&spec_tcp.list, &rule.list); 186 list_add_tail(&spec_tcp_udp.list, &rule.list);
169 187
170 rule.qpn = priv->rss_map.qps[filter->rxq_index].qpn; 188 rule.qpn = priv->rss_map.qps[filter->rxq_index].qpn;
171 memcpy(spec_eth.eth.dst_mac, priv->dev->dev_addr, ETH_ALEN); 189 memcpy(spec_eth.eth.dst_mac, priv->dev->dev_addr, ETH_ALEN);
@@ -183,6 +201,7 @@ static void mlx4_en_filter_work(struct work_struct *work)
183 if (rc) 201 if (rc)
184 en_err(priv, "Error attaching flow. err = %d\n", rc); 202 en_err(priv, "Error attaching flow. err = %d\n", rc);
185 203
204ignore:
186 mlx4_en_filter_rfs_expire(priv); 205 mlx4_en_filter_rfs_expire(priv);
187 206
188 filter->activated = 1; 207 filter->activated = 1;
@@ -206,8 +225,8 @@ filter_hash_bucket(struct mlx4_en_priv *priv, __be32 src_ip, __be32 dst_ip,
206 225
207static struct mlx4_en_filter * 226static struct mlx4_en_filter *
208mlx4_en_filter_alloc(struct mlx4_en_priv *priv, int rxq_index, __be32 src_ip, 227mlx4_en_filter_alloc(struct mlx4_en_priv *priv, int rxq_index, __be32 src_ip,
209 __be32 dst_ip, __be16 src_port, __be16 dst_port, 228 __be32 dst_ip, u8 ip_proto, __be16 src_port,
210 u32 flow_id) 229 __be16 dst_port, u32 flow_id)
211{ 230{
212 struct mlx4_en_filter *filter = NULL; 231 struct mlx4_en_filter *filter = NULL;
213 232
@@ -221,6 +240,7 @@ mlx4_en_filter_alloc(struct mlx4_en_priv *priv, int rxq_index, __be32 src_ip,
221 240
222 filter->src_ip = src_ip; 241 filter->src_ip = src_ip;
223 filter->dst_ip = dst_ip; 242 filter->dst_ip = dst_ip;
243 filter->ip_proto = ip_proto;
224 filter->src_port = src_port; 244 filter->src_port = src_port;
225 filter->dst_port = dst_port; 245 filter->dst_port = dst_port;
226 246
@@ -252,7 +272,7 @@ static void mlx4_en_filter_free(struct mlx4_en_filter *filter)
252 272
253static inline struct mlx4_en_filter * 273static inline struct mlx4_en_filter *
254mlx4_en_filter_find(struct mlx4_en_priv *priv, __be32 src_ip, __be32 dst_ip, 274mlx4_en_filter_find(struct mlx4_en_priv *priv, __be32 src_ip, __be32 dst_ip,
255 __be16 src_port, __be16 dst_port) 275 u8 ip_proto, __be16 src_port, __be16 dst_port)
256{ 276{
257 struct mlx4_en_filter *filter; 277 struct mlx4_en_filter *filter;
258 struct mlx4_en_filter *ret = NULL; 278 struct mlx4_en_filter *ret = NULL;
@@ -263,6 +283,7 @@ mlx4_en_filter_find(struct mlx4_en_priv *priv, __be32 src_ip, __be32 dst_ip,
263 filter_chain) { 283 filter_chain) {
264 if (filter->src_ip == src_ip && 284 if (filter->src_ip == src_ip &&
265 filter->dst_ip == dst_ip && 285 filter->dst_ip == dst_ip &&
286 filter->ip_proto == ip_proto &&
266 filter->src_port == src_port && 287 filter->src_port == src_port &&
267 filter->dst_port == dst_port) { 288 filter->dst_port == dst_port) {
268 ret = filter; 289 ret = filter;
@@ -281,6 +302,7 @@ mlx4_en_filter_rfs(struct net_device *net_dev, const struct sk_buff *skb,
281 struct mlx4_en_filter *filter; 302 struct mlx4_en_filter *filter;
282 const struct iphdr *ip; 303 const struct iphdr *ip;
283 const __be16 *ports; 304 const __be16 *ports;
305 u8 ip_proto;
284 __be32 src_ip; 306 __be32 src_ip;
285 __be32 dst_ip; 307 __be32 dst_ip;
286 __be16 src_port; 308 __be16 src_port;
@@ -295,18 +317,19 @@ mlx4_en_filter_rfs(struct net_device *net_dev, const struct sk_buff *skb,
295 if (ip_is_fragment(ip)) 317 if (ip_is_fragment(ip))
296 return -EPROTONOSUPPORT; 318 return -EPROTONOSUPPORT;
297 319
320 if ((ip->protocol != IPPROTO_TCP) && (ip->protocol != IPPROTO_UDP))
321 return -EPROTONOSUPPORT;
298 ports = (const __be16 *)(skb->data + nhoff + 4 * ip->ihl); 322 ports = (const __be16 *)(skb->data + nhoff + 4 * ip->ihl);
299 323
324 ip_proto = ip->protocol;
300 src_ip = ip->saddr; 325 src_ip = ip->saddr;
301 dst_ip = ip->daddr; 326 dst_ip = ip->daddr;
302 src_port = ports[0]; 327 src_port = ports[0];
303 dst_port = ports[1]; 328 dst_port = ports[1];
304 329
305 if (ip->protocol != IPPROTO_TCP)
306 return -EPROTONOSUPPORT;
307
308 spin_lock_bh(&priv->filters_lock); 330 spin_lock_bh(&priv->filters_lock);
309 filter = mlx4_en_filter_find(priv, src_ip, dst_ip, src_port, dst_port); 331 filter = mlx4_en_filter_find(priv, src_ip, dst_ip, ip_proto,
332 src_port, dst_port);
310 if (filter) { 333 if (filter) {
311 if (filter->rxq_index == rxq_index) 334 if (filter->rxq_index == rxq_index)
312 goto out; 335 goto out;
@@ -314,7 +337,7 @@ mlx4_en_filter_rfs(struct net_device *net_dev, const struct sk_buff *skb,
314 filter->rxq_index = rxq_index; 337 filter->rxq_index = rxq_index;
315 } else { 338 } else {
316 filter = mlx4_en_filter_alloc(priv, rxq_index, 339 filter = mlx4_en_filter_alloc(priv, rxq_index,
317 src_ip, dst_ip, 340 src_ip, dst_ip, ip_proto,
318 src_port, dst_port, flow_id); 341 src_port, dst_port, flow_id);
319 if (!filter) { 342 if (!filter) {
320 ret = -ENOMEM; 343 ret = -ENOMEM;
@@ -332,8 +355,7 @@ err:
332 return ret; 355 return ret;
333} 356}
334 357
335void mlx4_en_cleanup_filters(struct mlx4_en_priv *priv, 358void mlx4_en_cleanup_filters(struct mlx4_en_priv *priv)
336 struct mlx4_en_rx_ring *rx_ring)
337{ 359{
338 struct mlx4_en_filter *filter, *tmp; 360 struct mlx4_en_filter *filter, *tmp;
339 LIST_HEAD(del_list); 361 LIST_HEAD(del_list);
@@ -417,7 +439,6 @@ static int mlx4_en_vlan_rx_kill_vid(struct net_device *dev,
417 struct mlx4_en_priv *priv = netdev_priv(dev); 439 struct mlx4_en_priv *priv = netdev_priv(dev);
418 struct mlx4_en_dev *mdev = priv->mdev; 440 struct mlx4_en_dev *mdev = priv->mdev;
419 int err; 441 int err;
420 int idx;
421 442
422 en_dbg(HW, priv, "Killing VID:%d\n", vid); 443 en_dbg(HW, priv, "Killing VID:%d\n", vid);
423 444
@@ -425,10 +446,7 @@ static int mlx4_en_vlan_rx_kill_vid(struct net_device *dev,
425 446
426 /* Remove VID from port VLAN filter */ 447 /* Remove VID from port VLAN filter */
427 mutex_lock(&mdev->state_lock); 448 mutex_lock(&mdev->state_lock);
428 if (!mlx4_find_cached_vlan(mdev->dev, priv->port, vid, &idx)) 449 mlx4_unregister_vlan(mdev->dev, priv->port, vid);
429 mlx4_unregister_vlan(mdev->dev, priv->port, idx);
430 else
431 en_dbg(HW, priv, "could not find vid %d in cache\n", vid);
432 450
433 if (mdev->device_up && priv->port_up) { 451 if (mdev->device_up && priv->port_up) {
434 err = mlx4_SET_VLAN_FLTR(mdev->dev, priv); 452 err = mlx4_SET_VLAN_FLTR(mdev->dev, priv);
@@ -1223,7 +1241,7 @@ static void mlx4_en_netpoll(struct net_device *dev)
1223 int i; 1241 int i;
1224 1242
1225 for (i = 0; i < priv->rx_ring_num; i++) { 1243 for (i = 0; i < priv->rx_ring_num; i++) {
1226 cq = &priv->rx_cq[i]; 1244 cq = priv->rx_cq[i];
1227 spin_lock_irqsave(&cq->lock, flags); 1245 spin_lock_irqsave(&cq->lock, flags);
1228 napi_synchronize(&cq->napi); 1246 napi_synchronize(&cq->napi);
1229 mlx4_en_process_rx_cq(dev, cq, 0); 1247 mlx4_en_process_rx_cq(dev, cq, 0);
@@ -1245,8 +1263,8 @@ static void mlx4_en_tx_timeout(struct net_device *dev)
1245 if (!netif_tx_queue_stopped(netdev_get_tx_queue(dev, i))) 1263 if (!netif_tx_queue_stopped(netdev_get_tx_queue(dev, i)))
1246 continue; 1264 continue;
1247 en_warn(priv, "TX timeout on queue: %d, QP: 0x%x, CQ: 0x%x, Cons: 0x%x, Prod: 0x%x\n", 1265 en_warn(priv, "TX timeout on queue: %d, QP: 0x%x, CQ: 0x%x, Cons: 0x%x, Prod: 0x%x\n",
1248 i, priv->tx_ring[i].qpn, priv->tx_ring[i].cqn, 1266 i, priv->tx_ring[i]->qpn, priv->tx_ring[i]->cqn,
1249 priv->tx_ring[i].cons, priv->tx_ring[i].prod); 1267 priv->tx_ring[i]->cons, priv->tx_ring[i]->prod);
1250 } 1268 }
1251 1269
1252 priv->port_stats.tx_timeout++; 1270 priv->port_stats.tx_timeout++;
@@ -1286,7 +1304,7 @@ static void mlx4_en_set_default_moderation(struct mlx4_en_priv *priv)
1286 1304
1287 /* Setup cq moderation params */ 1305 /* Setup cq moderation params */
1288 for (i = 0; i < priv->rx_ring_num; i++) { 1306 for (i = 0; i < priv->rx_ring_num; i++) {
1289 cq = &priv->rx_cq[i]; 1307 cq = priv->rx_cq[i];
1290 cq->moder_cnt = priv->rx_frames; 1308 cq->moder_cnt = priv->rx_frames;
1291 cq->moder_time = priv->rx_usecs; 1309 cq->moder_time = priv->rx_usecs;
1292 priv->last_moder_time[i] = MLX4_EN_AUTO_CONF; 1310 priv->last_moder_time[i] = MLX4_EN_AUTO_CONF;
@@ -1295,7 +1313,7 @@ static void mlx4_en_set_default_moderation(struct mlx4_en_priv *priv)
1295 } 1313 }
1296 1314
1297 for (i = 0; i < priv->tx_ring_num; i++) { 1315 for (i = 0; i < priv->tx_ring_num; i++) {
1298 cq = &priv->tx_cq[i]; 1316 cq = priv->tx_cq[i];
1299 cq->moder_cnt = priv->tx_frames; 1317 cq->moder_cnt = priv->tx_frames;
1300 cq->moder_time = priv->tx_usecs; 1318 cq->moder_time = priv->tx_usecs;
1301 } 1319 }
@@ -1329,8 +1347,8 @@ static void mlx4_en_auto_moderation(struct mlx4_en_priv *priv)
1329 1347
1330 for (ring = 0; ring < priv->rx_ring_num; ring++) { 1348 for (ring = 0; ring < priv->rx_ring_num; ring++) {
1331 spin_lock_bh(&priv->stats_lock); 1349 spin_lock_bh(&priv->stats_lock);
1332 rx_packets = priv->rx_ring[ring].packets; 1350 rx_packets = priv->rx_ring[ring]->packets;
1333 rx_bytes = priv->rx_ring[ring].bytes; 1351 rx_bytes = priv->rx_ring[ring]->bytes;
1334 spin_unlock_bh(&priv->stats_lock); 1352 spin_unlock_bh(&priv->stats_lock);
1335 1353
1336 rx_pkt_diff = ((unsigned long) (rx_packets - 1354 rx_pkt_diff = ((unsigned long) (rx_packets -
@@ -1359,7 +1377,7 @@ static void mlx4_en_auto_moderation(struct mlx4_en_priv *priv)
1359 1377
1360 if (moder_time != priv->last_moder_time[ring]) { 1378 if (moder_time != priv->last_moder_time[ring]) {
1361 priv->last_moder_time[ring] = moder_time; 1379 priv->last_moder_time[ring] = moder_time;
1362 cq = &priv->rx_cq[ring]; 1380 cq = priv->rx_cq[ring];
1363 cq->moder_time = moder_time; 1381 cq->moder_time = moder_time;
1364 cq->moder_cnt = priv->rx_frames; 1382 cq->moder_cnt = priv->rx_frames;
1365 err = mlx4_en_set_cq_moder(priv, cq); 1383 err = mlx4_en_set_cq_moder(priv, cq);
@@ -1482,7 +1500,7 @@ int mlx4_en_start_port(struct net_device *dev)
1482 return err; 1500 return err;
1483 } 1501 }
1484 for (i = 0; i < priv->rx_ring_num; i++) { 1502 for (i = 0; i < priv->rx_ring_num; i++) {
1485 cq = &priv->rx_cq[i]; 1503 cq = priv->rx_cq[i];
1486 1504
1487 mlx4_en_cq_init_lock(cq); 1505 mlx4_en_cq_init_lock(cq);
1488 1506
@@ -1500,7 +1518,7 @@ int mlx4_en_start_port(struct net_device *dev)
1500 goto cq_err; 1518 goto cq_err;
1501 } 1519 }
1502 mlx4_en_arm_cq(priv, cq); 1520 mlx4_en_arm_cq(priv, cq);
1503 priv->rx_ring[i].cqn = cq->mcq.cqn; 1521 priv->rx_ring[i]->cqn = cq->mcq.cqn;
1504 ++rx_index; 1522 ++rx_index;
1505 } 1523 }
1506 1524
@@ -1526,7 +1544,7 @@ int mlx4_en_start_port(struct net_device *dev)
1526 /* Configure tx cq's and rings */ 1544 /* Configure tx cq's and rings */
1527 for (i = 0; i < priv->tx_ring_num; i++) { 1545 for (i = 0; i < priv->tx_ring_num; i++) {
1528 /* Configure cq */ 1546 /* Configure cq */
1529 cq = &priv->tx_cq[i]; 1547 cq = priv->tx_cq[i];
1530 err = mlx4_en_activate_cq(priv, cq, i); 1548 err = mlx4_en_activate_cq(priv, cq, i);
1531 if (err) { 1549 if (err) {
1532 en_err(priv, "Failed allocating Tx CQ\n"); 1550 en_err(priv, "Failed allocating Tx CQ\n");
@@ -1542,7 +1560,7 @@ int mlx4_en_start_port(struct net_device *dev)
1542 cq->buf->wqe_index = cpu_to_be16(0xffff); 1560 cq->buf->wqe_index = cpu_to_be16(0xffff);
1543 1561
1544 /* Configure ring */ 1562 /* Configure ring */
1545 tx_ring = &priv->tx_ring[i]; 1563 tx_ring = priv->tx_ring[i];
1546 err = mlx4_en_activate_tx_ring(priv, tx_ring, cq->mcq.cqn, 1564 err = mlx4_en_activate_tx_ring(priv, tx_ring, cq->mcq.cqn,
1547 i / priv->num_tx_rings_p_up); 1565 i / priv->num_tx_rings_p_up);
1548 if (err) { 1566 if (err) {
@@ -1612,8 +1630,8 @@ int mlx4_en_start_port(struct net_device *dev)
1612 1630
1613tx_err: 1631tx_err:
1614 while (tx_index--) { 1632 while (tx_index--) {
1615 mlx4_en_deactivate_tx_ring(priv, &priv->tx_ring[tx_index]); 1633 mlx4_en_deactivate_tx_ring(priv, priv->tx_ring[tx_index]);
1616 mlx4_en_deactivate_cq(priv, &priv->tx_cq[tx_index]); 1634 mlx4_en_deactivate_cq(priv, priv->tx_cq[tx_index]);
1617 } 1635 }
1618 mlx4_en_destroy_drop_qp(priv); 1636 mlx4_en_destroy_drop_qp(priv);
1619rss_err: 1637rss_err:
@@ -1622,9 +1640,9 @@ mac_err:
1622 mlx4_en_put_qp(priv); 1640 mlx4_en_put_qp(priv);
1623cq_err: 1641cq_err:
1624 while (rx_index--) 1642 while (rx_index--)
1625 mlx4_en_deactivate_cq(priv, &priv->rx_cq[rx_index]); 1643 mlx4_en_deactivate_cq(priv, priv->rx_cq[rx_index]);
1626 for (i = 0; i < priv->rx_ring_num; i++) 1644 for (i = 0; i < priv->rx_ring_num; i++)
1627 mlx4_en_deactivate_rx_ring(priv, &priv->rx_ring[i]); 1645 mlx4_en_deactivate_rx_ring(priv, priv->rx_ring[i]);
1628 1646
1629 return err; /* need to close devices */ 1647 return err; /* need to close devices */
1630} 1648}
@@ -1720,25 +1738,25 @@ void mlx4_en_stop_port(struct net_device *dev, int detach)
1720 1738
1721 /* Free TX Rings */ 1739 /* Free TX Rings */
1722 for (i = 0; i < priv->tx_ring_num; i++) { 1740 for (i = 0; i < priv->tx_ring_num; i++) {
1723 mlx4_en_deactivate_tx_ring(priv, &priv->tx_ring[i]); 1741 mlx4_en_deactivate_tx_ring(priv, priv->tx_ring[i]);
1724 mlx4_en_deactivate_cq(priv, &priv->tx_cq[i]); 1742 mlx4_en_deactivate_cq(priv, priv->tx_cq[i]);
1725 } 1743 }
1726 msleep(10); 1744 msleep(10);
1727 1745
1728 for (i = 0; i < priv->tx_ring_num; i++) 1746 for (i = 0; i < priv->tx_ring_num; i++)
1729 mlx4_en_free_tx_buf(dev, &priv->tx_ring[i]); 1747 mlx4_en_free_tx_buf(dev, priv->tx_ring[i]);
1730 1748
1731 /* Free RSS qps */ 1749 /* Free RSS qps */
1732 mlx4_en_release_rss_steer(priv); 1750 mlx4_en_release_rss_steer(priv);
1733 1751
1734 /* Unregister Mac address for the port */ 1752 /* Unregister Mac address for the port */
1735 mlx4_en_put_qp(priv); 1753 mlx4_en_put_qp(priv);
1736 if (!(mdev->dev->caps.flags2 & MLX4_DEV_CAP_FLAGS2_REASSIGN_MAC_EN)) 1754 if (!(mdev->dev->caps.flags2 & MLX4_DEV_CAP_FLAG2_REASSIGN_MAC_EN))
1737 mdev->mac_removed[priv->port] = 1; 1755 mdev->mac_removed[priv->port] = 1;
1738 1756
1739 /* Free RX Rings */ 1757 /* Free RX Rings */
1740 for (i = 0; i < priv->rx_ring_num; i++) { 1758 for (i = 0; i < priv->rx_ring_num; i++) {
1741 struct mlx4_en_cq *cq = &priv->rx_cq[i]; 1759 struct mlx4_en_cq *cq = priv->rx_cq[i];
1742 1760
1743 local_bh_disable(); 1761 local_bh_disable();
1744 while (!mlx4_en_cq_lock_napi(cq)) { 1762 while (!mlx4_en_cq_lock_napi(cq)) {
@@ -1749,7 +1767,7 @@ void mlx4_en_stop_port(struct net_device *dev, int detach)
1749 1767
1750 while (test_bit(NAPI_STATE_SCHED, &cq->napi.state)) 1768 while (test_bit(NAPI_STATE_SCHED, &cq->napi.state))
1751 msleep(1); 1769 msleep(1);
1752 mlx4_en_deactivate_rx_ring(priv, &priv->rx_ring[i]); 1770 mlx4_en_deactivate_rx_ring(priv, priv->rx_ring[i]);
1753 mlx4_en_deactivate_cq(priv, cq); 1771 mlx4_en_deactivate_cq(priv, cq);
1754 } 1772 }
1755} 1773}
@@ -1787,15 +1805,15 @@ static void mlx4_en_clear_stats(struct net_device *dev)
1787 memset(&priv->port_stats, 0, sizeof(priv->port_stats)); 1805 memset(&priv->port_stats, 0, sizeof(priv->port_stats));
1788 1806
1789 for (i = 0; i < priv->tx_ring_num; i++) { 1807 for (i = 0; i < priv->tx_ring_num; i++) {
1790 priv->tx_ring[i].bytes = 0; 1808 priv->tx_ring[i]->bytes = 0;
1791 priv->tx_ring[i].packets = 0; 1809 priv->tx_ring[i]->packets = 0;
1792 priv->tx_ring[i].tx_csum = 0; 1810 priv->tx_ring[i]->tx_csum = 0;
1793 } 1811 }
1794 for (i = 0; i < priv->rx_ring_num; i++) { 1812 for (i = 0; i < priv->rx_ring_num; i++) {
1795 priv->rx_ring[i].bytes = 0; 1813 priv->rx_ring[i]->bytes = 0;
1796 priv->rx_ring[i].packets = 0; 1814 priv->rx_ring[i]->packets = 0;
1797 priv->rx_ring[i].csum_ok = 0; 1815 priv->rx_ring[i]->csum_ok = 0;
1798 priv->rx_ring[i].csum_none = 0; 1816 priv->rx_ring[i]->csum_none = 0;
1799 } 1817 }
1800} 1818}
1801 1819
@@ -1852,17 +1870,17 @@ void mlx4_en_free_resources(struct mlx4_en_priv *priv)
1852#endif 1870#endif
1853 1871
1854 for (i = 0; i < priv->tx_ring_num; i++) { 1872 for (i = 0; i < priv->tx_ring_num; i++) {
1855 if (priv->tx_ring[i].tx_info) 1873 if (priv->tx_ring && priv->tx_ring[i])
1856 mlx4_en_destroy_tx_ring(priv, &priv->tx_ring[i]); 1874 mlx4_en_destroy_tx_ring(priv, &priv->tx_ring[i]);
1857 if (priv->tx_cq[i].buf) 1875 if (priv->tx_cq && priv->tx_cq[i])
1858 mlx4_en_destroy_cq(priv, &priv->tx_cq[i]); 1876 mlx4_en_destroy_cq(priv, &priv->tx_cq[i]);
1859 } 1877 }
1860 1878
1861 for (i = 0; i < priv->rx_ring_num; i++) { 1879 for (i = 0; i < priv->rx_ring_num; i++) {
1862 if (priv->rx_ring[i].rx_info) 1880 if (priv->rx_ring[i])
1863 mlx4_en_destroy_rx_ring(priv, &priv->rx_ring[i], 1881 mlx4_en_destroy_rx_ring(priv, &priv->rx_ring[i],
1864 priv->prof->rx_ring_size, priv->stride); 1882 priv->prof->rx_ring_size, priv->stride);
1865 if (priv->rx_cq[i].buf) 1883 if (priv->rx_cq[i])
1866 mlx4_en_destroy_cq(priv, &priv->rx_cq[i]); 1884 mlx4_en_destroy_cq(priv, &priv->rx_cq[i]);
1867 } 1885 }
1868 1886
@@ -1877,6 +1895,7 @@ int mlx4_en_alloc_resources(struct mlx4_en_priv *priv)
1877 struct mlx4_en_port_profile *prof = priv->prof; 1895 struct mlx4_en_port_profile *prof = priv->prof;
1878 int i; 1896 int i;
1879 int err; 1897 int err;
1898 int node;
1880 1899
1881 err = mlx4_qp_reserve_range(priv->mdev->dev, priv->tx_ring_num, 256, &priv->base_tx_qpn); 1900 err = mlx4_qp_reserve_range(priv->mdev->dev, priv->tx_ring_num, 256, &priv->base_tx_qpn);
1882 if (err) { 1901 if (err) {
@@ -1886,23 +1905,26 @@ int mlx4_en_alloc_resources(struct mlx4_en_priv *priv)
1886 1905
1887 /* Create tx Rings */ 1906 /* Create tx Rings */
1888 for (i = 0; i < priv->tx_ring_num; i++) { 1907 for (i = 0; i < priv->tx_ring_num; i++) {
1908 node = cpu_to_node(i % num_online_cpus());
1889 if (mlx4_en_create_cq(priv, &priv->tx_cq[i], 1909 if (mlx4_en_create_cq(priv, &priv->tx_cq[i],
1890 prof->tx_ring_size, i, TX)) 1910 prof->tx_ring_size, i, TX, node))
1891 goto err; 1911 goto err;
1892 1912
1893 if (mlx4_en_create_tx_ring(priv, &priv->tx_ring[i], priv->base_tx_qpn + i, 1913 if (mlx4_en_create_tx_ring(priv, &priv->tx_ring[i], priv->base_tx_qpn + i,
1894 prof->tx_ring_size, TXBB_SIZE)) 1914 prof->tx_ring_size, TXBB_SIZE, node))
1895 goto err; 1915 goto err;
1896 } 1916 }
1897 1917
1898 /* Create rx Rings */ 1918 /* Create rx Rings */
1899 for (i = 0; i < priv->rx_ring_num; i++) { 1919 for (i = 0; i < priv->rx_ring_num; i++) {
1920 node = cpu_to_node(i % num_online_cpus());
1900 if (mlx4_en_create_cq(priv, &priv->rx_cq[i], 1921 if (mlx4_en_create_cq(priv, &priv->rx_cq[i],
1901 prof->rx_ring_size, i, RX)) 1922 prof->rx_ring_size, i, RX, node))
1902 goto err; 1923 goto err;
1903 1924
1904 if (mlx4_en_create_rx_ring(priv, &priv->rx_ring[i], 1925 if (mlx4_en_create_rx_ring(priv, &priv->rx_ring[i],
1905 prof->rx_ring_size, priv->stride)) 1926 prof->rx_ring_size, priv->stride,
1927 node))
1906 goto err; 1928 goto err;
1907 } 1929 }
1908 1930
@@ -1918,6 +1940,20 @@ int mlx4_en_alloc_resources(struct mlx4_en_priv *priv)
1918 1940
1919err: 1941err:
1920 en_err(priv, "Failed to allocate NIC resources\n"); 1942 en_err(priv, "Failed to allocate NIC resources\n");
1943 for (i = 0; i < priv->rx_ring_num; i++) {
1944 if (priv->rx_ring[i])
1945 mlx4_en_destroy_rx_ring(priv, &priv->rx_ring[i],
1946 prof->rx_ring_size,
1947 priv->stride);
1948 if (priv->rx_cq[i])
1949 mlx4_en_destroy_cq(priv, &priv->rx_cq[i]);
1950 }
1951 for (i = 0; i < priv->tx_ring_num; i++) {
1952 if (priv->tx_ring[i])
1953 mlx4_en_destroy_tx_ring(priv, &priv->tx_ring[i]);
1954 if (priv->tx_cq[i])
1955 mlx4_en_destroy_cq(priv, &priv->tx_cq[i]);
1956 }
1921 return -ENOMEM; 1957 return -ENOMEM;
1922} 1958}
1923 1959
@@ -2211,13 +2247,13 @@ int mlx4_en_init_netdev(struct mlx4_en_dev *mdev, int port,
2211 priv->num_tx_rings_p_up = mdev->profile.num_tx_rings_p_up; 2247 priv->num_tx_rings_p_up = mdev->profile.num_tx_rings_p_up;
2212 priv->tx_ring_num = prof->tx_ring_num; 2248 priv->tx_ring_num = prof->tx_ring_num;
2213 2249
2214 priv->tx_ring = kzalloc(sizeof(struct mlx4_en_tx_ring) * MAX_TX_RINGS, 2250 priv->tx_ring = kzalloc(sizeof(struct mlx4_en_tx_ring *) * MAX_TX_RINGS,
2215 GFP_KERNEL); 2251 GFP_KERNEL);
2216 if (!priv->tx_ring) { 2252 if (!priv->tx_ring) {
2217 err = -ENOMEM; 2253 err = -ENOMEM;
2218 goto out; 2254 goto out;
2219 } 2255 }
2220 priv->tx_cq = kzalloc(sizeof(struct mlx4_en_cq) * MAX_TX_RINGS, 2256 priv->tx_cq = kzalloc(sizeof(struct mlx4_en_cq *) * MAX_TX_RINGS,
2221 GFP_KERNEL); 2257 GFP_KERNEL);
2222 if (!priv->tx_cq) { 2258 if (!priv->tx_cq) {
2223 err = -ENOMEM; 2259 err = -ENOMEM;
diff --git a/drivers/net/ethernet/mellanox/mlx4/en_port.c b/drivers/net/ethernet/mellanox/mlx4/en_port.c
index 331791467a22..dae1a1f4ae55 100644
--- a/drivers/net/ethernet/mellanox/mlx4/en_port.c
+++ b/drivers/net/ethernet/mellanox/mlx4/en_port.c
@@ -56,7 +56,6 @@ int mlx4_SET_VLAN_FLTR(struct mlx4_dev *dev, struct mlx4_en_priv *priv)
56 return PTR_ERR(mailbox); 56 return PTR_ERR(mailbox);
57 57
58 filter = mailbox->buf; 58 filter = mailbox->buf;
59 memset(filter, 0, sizeof(*filter));
60 for (i = VLAN_FLTR_SIZE - 1; i >= 0; i--) { 59 for (i = VLAN_FLTR_SIZE - 1; i >= 0; i--) {
61 entry = 0; 60 entry = 0;
62 for (j = 0; j < 32; j++) 61 for (j = 0; j < 32; j++)
@@ -81,7 +80,6 @@ int mlx4_en_QUERY_PORT(struct mlx4_en_dev *mdev, u8 port)
81 mailbox = mlx4_alloc_cmd_mailbox(mdev->dev); 80 mailbox = mlx4_alloc_cmd_mailbox(mdev->dev);
82 if (IS_ERR(mailbox)) 81 if (IS_ERR(mailbox))
83 return PTR_ERR(mailbox); 82 return PTR_ERR(mailbox);
84 memset(mailbox->buf, 0, sizeof(*qport_context));
85 err = mlx4_cmd_box(mdev->dev, 0, mailbox->dma, port, 0, 83 err = mlx4_cmd_box(mdev->dev, 0, mailbox->dma, port, 0,
86 MLX4_CMD_QUERY_PORT, MLX4_CMD_TIME_CLASS_B, 84 MLX4_CMD_QUERY_PORT, MLX4_CMD_TIME_CLASS_B,
87 MLX4_CMD_WRAPPED); 85 MLX4_CMD_WRAPPED);
@@ -127,7 +125,6 @@ int mlx4_en_DUMP_ETH_STATS(struct mlx4_en_dev *mdev, u8 port, u8 reset)
127 mailbox = mlx4_alloc_cmd_mailbox(mdev->dev); 125 mailbox = mlx4_alloc_cmd_mailbox(mdev->dev);
128 if (IS_ERR(mailbox)) 126 if (IS_ERR(mailbox))
129 return PTR_ERR(mailbox); 127 return PTR_ERR(mailbox);
130 memset(mailbox->buf, 0, sizeof(*mlx4_en_stats));
131 err = mlx4_cmd_box(mdev->dev, 0, mailbox->dma, in_mod, 0, 128 err = mlx4_cmd_box(mdev->dev, 0, mailbox->dma, in_mod, 0,
132 MLX4_CMD_DUMP_ETH_STATS, MLX4_CMD_TIME_CLASS_B, 129 MLX4_CMD_DUMP_ETH_STATS, MLX4_CMD_TIME_CLASS_B,
133 MLX4_CMD_WRAPPED); 130 MLX4_CMD_WRAPPED);
@@ -143,18 +140,18 @@ int mlx4_en_DUMP_ETH_STATS(struct mlx4_en_dev *mdev, u8 port, u8 reset)
143 priv->port_stats.rx_chksum_good = 0; 140 priv->port_stats.rx_chksum_good = 0;
144 priv->port_stats.rx_chksum_none = 0; 141 priv->port_stats.rx_chksum_none = 0;
145 for (i = 0; i < priv->rx_ring_num; i++) { 142 for (i = 0; i < priv->rx_ring_num; i++) {
146 stats->rx_packets += priv->rx_ring[i].packets; 143 stats->rx_packets += priv->rx_ring[i]->packets;
147 stats->rx_bytes += priv->rx_ring[i].bytes; 144 stats->rx_bytes += priv->rx_ring[i]->bytes;
148 priv->port_stats.rx_chksum_good += priv->rx_ring[i].csum_ok; 145 priv->port_stats.rx_chksum_good += priv->rx_ring[i]->csum_ok;
149 priv->port_stats.rx_chksum_none += priv->rx_ring[i].csum_none; 146 priv->port_stats.rx_chksum_none += priv->rx_ring[i]->csum_none;
150 } 147 }
151 stats->tx_packets = 0; 148 stats->tx_packets = 0;
152 stats->tx_bytes = 0; 149 stats->tx_bytes = 0;
153 priv->port_stats.tx_chksum_offload = 0; 150 priv->port_stats.tx_chksum_offload = 0;
154 for (i = 0; i < priv->tx_ring_num; i++) { 151 for (i = 0; i < priv->tx_ring_num; i++) {
155 stats->tx_packets += priv->tx_ring[i].packets; 152 stats->tx_packets += priv->tx_ring[i]->packets;
156 stats->tx_bytes += priv->tx_ring[i].bytes; 153 stats->tx_bytes += priv->tx_ring[i]->bytes;
157 priv->port_stats.tx_chksum_offload += priv->tx_ring[i].tx_csum; 154 priv->port_stats.tx_chksum_offload += priv->tx_ring[i]->tx_csum;
158 } 155 }
159 156
160 stats->rx_errors = be64_to_cpu(mlx4_en_stats->PCS) + 157 stats->rx_errors = be64_to_cpu(mlx4_en_stats->PCS) +
diff --git a/drivers/net/ethernet/mellanox/mlx4/en_rx.c b/drivers/net/ethernet/mellanox/mlx4/en_rx.c
index afe2efa69c86..07a1d0fbae47 100644
--- a/drivers/net/ethernet/mellanox/mlx4/en_rx.c
+++ b/drivers/net/ethernet/mellanox/mlx4/en_rx.c
@@ -264,7 +264,7 @@ static int mlx4_en_fill_rx_buffers(struct mlx4_en_priv *priv)
264 264
265 for (buf_ind = 0; buf_ind < priv->prof->rx_ring_size; buf_ind++) { 265 for (buf_ind = 0; buf_ind < priv->prof->rx_ring_size; buf_ind++) {
266 for (ring_ind = 0; ring_ind < priv->rx_ring_num; ring_ind++) { 266 for (ring_ind = 0; ring_ind < priv->rx_ring_num; ring_ind++) {
267 ring = &priv->rx_ring[ring_ind]; 267 ring = priv->rx_ring[ring_ind];
268 268
269 if (mlx4_en_prepare_rx_desc(priv, ring, 269 if (mlx4_en_prepare_rx_desc(priv, ring,
270 ring->actual_size, 270 ring->actual_size,
@@ -289,7 +289,7 @@ static int mlx4_en_fill_rx_buffers(struct mlx4_en_priv *priv)
289 289
290reduce_rings: 290reduce_rings:
291 for (ring_ind = 0; ring_ind < priv->rx_ring_num; ring_ind++) { 291 for (ring_ind = 0; ring_ind < priv->rx_ring_num; ring_ind++) {
292 ring = &priv->rx_ring[ring_ind]; 292 ring = priv->rx_ring[ring_ind];
293 while (ring->actual_size > new_size) { 293 while (ring->actual_size > new_size) {
294 ring->actual_size--; 294 ring->actual_size--;
295 ring->prod--; 295 ring->prod--;
@@ -319,12 +319,23 @@ static void mlx4_en_free_rx_buf(struct mlx4_en_priv *priv,
319} 319}
320 320
321int mlx4_en_create_rx_ring(struct mlx4_en_priv *priv, 321int mlx4_en_create_rx_ring(struct mlx4_en_priv *priv,
322 struct mlx4_en_rx_ring *ring, u32 size, u16 stride) 322 struct mlx4_en_rx_ring **pring,
323 u32 size, u16 stride, int node)
323{ 324{
324 struct mlx4_en_dev *mdev = priv->mdev; 325 struct mlx4_en_dev *mdev = priv->mdev;
326 struct mlx4_en_rx_ring *ring;
325 int err = -ENOMEM; 327 int err = -ENOMEM;
326 int tmp; 328 int tmp;
327 329
330 ring = kzalloc_node(sizeof(*ring), GFP_KERNEL, node);
331 if (!ring) {
332 ring = kzalloc(sizeof(*ring), GFP_KERNEL);
333 if (!ring) {
334 en_err(priv, "Failed to allocate RX ring structure\n");
335 return -ENOMEM;
336 }
337 }
338
328 ring->prod = 0; 339 ring->prod = 0;
329 ring->cons = 0; 340 ring->cons = 0;
330 ring->size = size; 341 ring->size = size;
@@ -335,17 +346,25 @@ int mlx4_en_create_rx_ring(struct mlx4_en_priv *priv,
335 346
336 tmp = size * roundup_pow_of_two(MLX4_EN_MAX_RX_FRAGS * 347 tmp = size * roundup_pow_of_two(MLX4_EN_MAX_RX_FRAGS *
337 sizeof(struct mlx4_en_rx_alloc)); 348 sizeof(struct mlx4_en_rx_alloc));
338 ring->rx_info = vmalloc(tmp); 349 ring->rx_info = vmalloc_node(tmp, node);
339 if (!ring->rx_info) 350 if (!ring->rx_info) {
340 return -ENOMEM; 351 ring->rx_info = vmalloc(tmp);
352 if (!ring->rx_info) {
353 err = -ENOMEM;
354 goto err_ring;
355 }
356 }
341 357
342 en_dbg(DRV, priv, "Allocated rx_info ring at addr:%p size:%d\n", 358 en_dbg(DRV, priv, "Allocated rx_info ring at addr:%p size:%d\n",
343 ring->rx_info, tmp); 359 ring->rx_info, tmp);
344 360
361 /* Allocate HW buffers on provided NUMA node */
362 set_dev_node(&mdev->dev->pdev->dev, node);
345 err = mlx4_alloc_hwq_res(mdev->dev, &ring->wqres, 363 err = mlx4_alloc_hwq_res(mdev->dev, &ring->wqres,
346 ring->buf_size, 2 * PAGE_SIZE); 364 ring->buf_size, 2 * PAGE_SIZE);
365 set_dev_node(&mdev->dev->pdev->dev, mdev->dev->numa_node);
347 if (err) 366 if (err)
348 goto err_ring; 367 goto err_info;
349 368
350 err = mlx4_en_map_buffer(&ring->wqres.buf); 369 err = mlx4_en_map_buffer(&ring->wqres.buf);
351 if (err) { 370 if (err) {
@@ -356,13 +375,18 @@ int mlx4_en_create_rx_ring(struct mlx4_en_priv *priv,
356 375
357 ring->hwtstamp_rx_filter = priv->hwtstamp_config.rx_filter; 376 ring->hwtstamp_rx_filter = priv->hwtstamp_config.rx_filter;
358 377
378 *pring = ring;
359 return 0; 379 return 0;
360 380
361err_hwq: 381err_hwq:
362 mlx4_free_hwq_res(mdev->dev, &ring->wqres, ring->buf_size); 382 mlx4_free_hwq_res(mdev->dev, &ring->wqres, ring->buf_size);
363err_ring: 383err_info:
364 vfree(ring->rx_info); 384 vfree(ring->rx_info);
365 ring->rx_info = NULL; 385 ring->rx_info = NULL;
386err_ring:
387 kfree(ring);
388 *pring = NULL;
389
366 return err; 390 return err;
367} 391}
368 392
@@ -376,12 +400,12 @@ int mlx4_en_activate_rx_rings(struct mlx4_en_priv *priv)
376 DS_SIZE * priv->num_frags); 400 DS_SIZE * priv->num_frags);
377 401
378 for (ring_ind = 0; ring_ind < priv->rx_ring_num; ring_ind++) { 402 for (ring_ind = 0; ring_ind < priv->rx_ring_num; ring_ind++) {
379 ring = &priv->rx_ring[ring_ind]; 403 ring = priv->rx_ring[ring_ind];
380 404
381 ring->prod = 0; 405 ring->prod = 0;
382 ring->cons = 0; 406 ring->cons = 0;
383 ring->actual_size = 0; 407 ring->actual_size = 0;
384 ring->cqn = priv->rx_cq[ring_ind].mcq.cqn; 408 ring->cqn = priv->rx_cq[ring_ind]->mcq.cqn;
385 409
386 ring->stride = stride; 410 ring->stride = stride;
387 if (ring->stride <= TXBB_SIZE) 411 if (ring->stride <= TXBB_SIZE)
@@ -412,7 +436,7 @@ int mlx4_en_activate_rx_rings(struct mlx4_en_priv *priv)
412 goto err_buffers; 436 goto err_buffers;
413 437
414 for (ring_ind = 0; ring_ind < priv->rx_ring_num; ring_ind++) { 438 for (ring_ind = 0; ring_ind < priv->rx_ring_num; ring_ind++) {
415 ring = &priv->rx_ring[ring_ind]; 439 ring = priv->rx_ring[ring_ind];
416 440
417 ring->size_mask = ring->actual_size - 1; 441 ring->size_mask = ring->actual_size - 1;
418 mlx4_en_update_rx_prod_db(ring); 442 mlx4_en_update_rx_prod_db(ring);
@@ -422,30 +446,34 @@ int mlx4_en_activate_rx_rings(struct mlx4_en_priv *priv)
422 446
423err_buffers: 447err_buffers:
424 for (ring_ind = 0; ring_ind < priv->rx_ring_num; ring_ind++) 448 for (ring_ind = 0; ring_ind < priv->rx_ring_num; ring_ind++)
425 mlx4_en_free_rx_buf(priv, &priv->rx_ring[ring_ind]); 449 mlx4_en_free_rx_buf(priv, priv->rx_ring[ring_ind]);
426 450
427 ring_ind = priv->rx_ring_num - 1; 451 ring_ind = priv->rx_ring_num - 1;
428err_allocator: 452err_allocator:
429 while (ring_ind >= 0) { 453 while (ring_ind >= 0) {
430 if (priv->rx_ring[ring_ind].stride <= TXBB_SIZE) 454 if (priv->rx_ring[ring_ind]->stride <= TXBB_SIZE)
431 priv->rx_ring[ring_ind].buf -= TXBB_SIZE; 455 priv->rx_ring[ring_ind]->buf -= TXBB_SIZE;
432 mlx4_en_destroy_allocator(priv, &priv->rx_ring[ring_ind]); 456 mlx4_en_destroy_allocator(priv, priv->rx_ring[ring_ind]);
433 ring_ind--; 457 ring_ind--;
434 } 458 }
435 return err; 459 return err;
436} 460}
437 461
438void mlx4_en_destroy_rx_ring(struct mlx4_en_priv *priv, 462void mlx4_en_destroy_rx_ring(struct mlx4_en_priv *priv,
439 struct mlx4_en_rx_ring *ring, u32 size, u16 stride) 463 struct mlx4_en_rx_ring **pring,
464 u32 size, u16 stride)
440{ 465{
441 struct mlx4_en_dev *mdev = priv->mdev; 466 struct mlx4_en_dev *mdev = priv->mdev;
467 struct mlx4_en_rx_ring *ring = *pring;
442 468
443 mlx4_en_unmap_buffer(&ring->wqres.buf); 469 mlx4_en_unmap_buffer(&ring->wqres.buf);
444 mlx4_free_hwq_res(mdev->dev, &ring->wqres, size * stride + TXBB_SIZE); 470 mlx4_free_hwq_res(mdev->dev, &ring->wqres, size * stride + TXBB_SIZE);
445 vfree(ring->rx_info); 471 vfree(ring->rx_info);
446 ring->rx_info = NULL; 472 ring->rx_info = NULL;
473 kfree(ring);
474 *pring = NULL;
447#ifdef CONFIG_RFS_ACCEL 475#ifdef CONFIG_RFS_ACCEL
448 mlx4_en_cleanup_filters(priv, ring); 476 mlx4_en_cleanup_filters(priv);
449#endif 477#endif
450} 478}
451 479
@@ -592,7 +620,7 @@ int mlx4_en_process_rx_cq(struct net_device *dev, struct mlx4_en_cq *cq, int bud
592 struct mlx4_en_priv *priv = netdev_priv(dev); 620 struct mlx4_en_priv *priv = netdev_priv(dev);
593 struct mlx4_en_dev *mdev = priv->mdev; 621 struct mlx4_en_dev *mdev = priv->mdev;
594 struct mlx4_cqe *cqe; 622 struct mlx4_cqe *cqe;
595 struct mlx4_en_rx_ring *ring = &priv->rx_ring[cq->ring]; 623 struct mlx4_en_rx_ring *ring = priv->rx_ring[cq->ring];
596 struct mlx4_en_rx_alloc *frags; 624 struct mlx4_en_rx_alloc *frags;
597 struct mlx4_en_rx_desc *rx_desc; 625 struct mlx4_en_rx_desc *rx_desc;
598 struct sk_buff *skb; 626 struct sk_buff *skb;
@@ -991,7 +1019,7 @@ int mlx4_en_config_rss_steer(struct mlx4_en_priv *priv)
991 1019
992 for (i = 0; i < priv->rx_ring_num; i++) { 1020 for (i = 0; i < priv->rx_ring_num; i++) {
993 qpn = rss_map->base_qpn + i; 1021 qpn = rss_map->base_qpn + i;
994 err = mlx4_en_config_rss_qp(priv, qpn, &priv->rx_ring[i], 1022 err = mlx4_en_config_rss_qp(priv, qpn, priv->rx_ring[i],
995 &rss_map->state[i], 1023 &rss_map->state[i],
996 &rss_map->qps[i]); 1024 &rss_map->qps[i]);
997 if (err) 1025 if (err)
@@ -1008,7 +1036,7 @@ int mlx4_en_config_rss_steer(struct mlx4_en_priv *priv)
1008 } 1036 }
1009 rss_map->indir_qp.event = mlx4_en_sqp_event; 1037 rss_map->indir_qp.event = mlx4_en_sqp_event;
1010 mlx4_en_fill_qp_context(priv, 0, 0, 0, 1, priv->base_qpn, 1038 mlx4_en_fill_qp_context(priv, 0, 0, 0, 1, priv->base_qpn,
1011 priv->rx_ring[0].cqn, -1, &context); 1039 priv->rx_ring[0]->cqn, -1, &context);
1012 1040
1013 if (!priv->prof->rss_rings || priv->prof->rss_rings > priv->rx_ring_num) 1041 if (!priv->prof->rss_rings || priv->prof->rss_rings > priv->rx_ring_num)
1014 rss_rings = priv->rx_ring_num; 1042 rss_rings = priv->rx_ring_num;
diff --git a/drivers/net/ethernet/mellanox/mlx4/en_selftest.c b/drivers/net/ethernet/mellanox/mlx4/en_selftest.c
index 2448f0d669e6..40626690e8a8 100644
--- a/drivers/net/ethernet/mellanox/mlx4/en_selftest.c
+++ b/drivers/net/ethernet/mellanox/mlx4/en_selftest.c
@@ -156,7 +156,7 @@ retry_tx:
156 * since we turned the carrier off */ 156 * since we turned the carrier off */
157 msleep(200); 157 msleep(200);
158 for (i = 0; i < priv->tx_ring_num && carrier_ok; i++) { 158 for (i = 0; i < priv->tx_ring_num && carrier_ok; i++) {
159 tx_ring = &priv->tx_ring[i]; 159 tx_ring = priv->tx_ring[i];
160 if (tx_ring->prod != (tx_ring->cons + tx_ring->last_nr_txbb)) 160 if (tx_ring->prod != (tx_ring->cons + tx_ring->last_nr_txbb))
161 goto retry_tx; 161 goto retry_tx;
162 } 162 }
diff --git a/drivers/net/ethernet/mellanox/mlx4/en_tx.c b/drivers/net/ethernet/mellanox/mlx4/en_tx.c
index 0698c82d6ff1..f54ebd5a1702 100644
--- a/drivers/net/ethernet/mellanox/mlx4/en_tx.c
+++ b/drivers/net/ethernet/mellanox/mlx4/en_tx.c
@@ -54,13 +54,23 @@ module_param_named(inline_thold, inline_thold, int, 0444);
54MODULE_PARM_DESC(inline_thold, "threshold for using inline data"); 54MODULE_PARM_DESC(inline_thold, "threshold for using inline data");
55 55
56int mlx4_en_create_tx_ring(struct mlx4_en_priv *priv, 56int mlx4_en_create_tx_ring(struct mlx4_en_priv *priv,
57 struct mlx4_en_tx_ring *ring, int qpn, u32 size, 57 struct mlx4_en_tx_ring **pring, int qpn, u32 size,
58 u16 stride) 58 u16 stride, int node)
59{ 59{
60 struct mlx4_en_dev *mdev = priv->mdev; 60 struct mlx4_en_dev *mdev = priv->mdev;
61 struct mlx4_en_tx_ring *ring;
61 int tmp; 62 int tmp;
62 int err; 63 int err;
63 64
65 ring = kzalloc_node(sizeof(*ring), GFP_KERNEL, node);
66 if (!ring) {
67 ring = kzalloc(sizeof(*ring), GFP_KERNEL);
68 if (!ring) {
69 en_err(priv, "Failed allocating TX ring\n");
70 return -ENOMEM;
71 }
72 }
73
64 ring->size = size; 74 ring->size = size;
65 ring->size_mask = size - 1; 75 ring->size_mask = size - 1;
66 ring->stride = stride; 76 ring->stride = stride;
@@ -68,22 +78,33 @@ int mlx4_en_create_tx_ring(struct mlx4_en_priv *priv,
68 inline_thold = min(inline_thold, MAX_INLINE); 78 inline_thold = min(inline_thold, MAX_INLINE);
69 79
70 tmp = size * sizeof(struct mlx4_en_tx_info); 80 tmp = size * sizeof(struct mlx4_en_tx_info);
71 ring->tx_info = vmalloc(tmp); 81 ring->tx_info = vmalloc_node(tmp, node);
72 if (!ring->tx_info) 82 if (!ring->tx_info) {
73 return -ENOMEM; 83 ring->tx_info = vmalloc(tmp);
84 if (!ring->tx_info) {
85 err = -ENOMEM;
86 goto err_ring;
87 }
88 }
74 89
75 en_dbg(DRV, priv, "Allocated tx_info ring at addr:%p size:%d\n", 90 en_dbg(DRV, priv, "Allocated tx_info ring at addr:%p size:%d\n",
76 ring->tx_info, tmp); 91 ring->tx_info, tmp);
77 92
78 ring->bounce_buf = kmalloc(MAX_DESC_SIZE, GFP_KERNEL); 93 ring->bounce_buf = kmalloc_node(MAX_DESC_SIZE, GFP_KERNEL, node);
79 if (!ring->bounce_buf) { 94 if (!ring->bounce_buf) {
80 err = -ENOMEM; 95 ring->bounce_buf = kmalloc(MAX_DESC_SIZE, GFP_KERNEL);
81 goto err_tx; 96 if (!ring->bounce_buf) {
97 err = -ENOMEM;
98 goto err_info;
99 }
82 } 100 }
83 ring->buf_size = ALIGN(size * ring->stride, MLX4_EN_PAGE_SIZE); 101 ring->buf_size = ALIGN(size * ring->stride, MLX4_EN_PAGE_SIZE);
84 102
103 /* Allocate HW buffers on provided NUMA node */
104 set_dev_node(&mdev->dev->pdev->dev, node);
85 err = mlx4_alloc_hwq_res(mdev->dev, &ring->wqres, ring->buf_size, 105 err = mlx4_alloc_hwq_res(mdev->dev, &ring->wqres, ring->buf_size,
86 2 * PAGE_SIZE); 106 2 * PAGE_SIZE);
107 set_dev_node(&mdev->dev->pdev->dev, mdev->dev->numa_node);
87 if (err) { 108 if (err) {
88 en_err(priv, "Failed allocating hwq resources\n"); 109 en_err(priv, "Failed allocating hwq resources\n");
89 goto err_bounce; 110 goto err_bounce;
@@ -109,7 +130,7 @@ int mlx4_en_create_tx_ring(struct mlx4_en_priv *priv,
109 } 130 }
110 ring->qp.event = mlx4_en_sqp_event; 131 ring->qp.event = mlx4_en_sqp_event;
111 132
112 err = mlx4_bf_alloc(mdev->dev, &ring->bf); 133 err = mlx4_bf_alloc(mdev->dev, &ring->bf, node);
113 if (err) { 134 if (err) {
114 en_dbg(DRV, priv, "working without blueflame (%d)", err); 135 en_dbg(DRV, priv, "working without blueflame (%d)", err);
115 ring->bf.uar = &mdev->priv_uar; 136 ring->bf.uar = &mdev->priv_uar;
@@ -120,6 +141,7 @@ int mlx4_en_create_tx_ring(struct mlx4_en_priv *priv,
120 141
121 ring->hwtstamp_tx_type = priv->hwtstamp_config.tx_type; 142 ring->hwtstamp_tx_type = priv->hwtstamp_config.tx_type;
122 143
144 *pring = ring;
123 return 0; 145 return 0;
124 146
125err_map: 147err_map:
@@ -129,16 +151,20 @@ err_hwq_res:
129err_bounce: 151err_bounce:
130 kfree(ring->bounce_buf); 152 kfree(ring->bounce_buf);
131 ring->bounce_buf = NULL; 153 ring->bounce_buf = NULL;
132err_tx: 154err_info:
133 vfree(ring->tx_info); 155 vfree(ring->tx_info);
134 ring->tx_info = NULL; 156 ring->tx_info = NULL;
157err_ring:
158 kfree(ring);
159 *pring = NULL;
135 return err; 160 return err;
136} 161}
137 162
138void mlx4_en_destroy_tx_ring(struct mlx4_en_priv *priv, 163void mlx4_en_destroy_tx_ring(struct mlx4_en_priv *priv,
139 struct mlx4_en_tx_ring *ring) 164 struct mlx4_en_tx_ring **pring)
140{ 165{
141 struct mlx4_en_dev *mdev = priv->mdev; 166 struct mlx4_en_dev *mdev = priv->mdev;
167 struct mlx4_en_tx_ring *ring = *pring;
142 en_dbg(DRV, priv, "Destroying tx ring, qpn: %d\n", ring->qpn); 168 en_dbg(DRV, priv, "Destroying tx ring, qpn: %d\n", ring->qpn);
143 169
144 if (ring->bf_enabled) 170 if (ring->bf_enabled)
@@ -151,6 +177,8 @@ void mlx4_en_destroy_tx_ring(struct mlx4_en_priv *priv,
151 ring->bounce_buf = NULL; 177 ring->bounce_buf = NULL;
152 vfree(ring->tx_info); 178 vfree(ring->tx_info);
153 ring->tx_info = NULL; 179 ring->tx_info = NULL;
180 kfree(ring);
181 *pring = NULL;
154} 182}
155 183
156int mlx4_en_activate_tx_ring(struct mlx4_en_priv *priv, 184int mlx4_en_activate_tx_ring(struct mlx4_en_priv *priv,
@@ -330,7 +358,7 @@ static void mlx4_en_process_tx_cq(struct net_device *dev, struct mlx4_en_cq *cq)
330{ 358{
331 struct mlx4_en_priv *priv = netdev_priv(dev); 359 struct mlx4_en_priv *priv = netdev_priv(dev);
332 struct mlx4_cq *mcq = &cq->mcq; 360 struct mlx4_cq *mcq = &cq->mcq;
333 struct mlx4_en_tx_ring *ring = &priv->tx_ring[cq->ring]; 361 struct mlx4_en_tx_ring *ring = priv->tx_ring[cq->ring];
334 struct mlx4_cqe *cqe; 362 struct mlx4_cqe *cqe;
335 u16 index; 363 u16 index;
336 u16 new_index, ring_index, stamp_index; 364 u16 new_index, ring_index, stamp_index;
@@ -622,7 +650,7 @@ netdev_tx_t mlx4_en_xmit(struct sk_buff *skb, struct net_device *dev)
622 } 650 }
623 651
624 tx_ind = skb->queue_mapping; 652 tx_ind = skb->queue_mapping;
625 ring = &priv->tx_ring[tx_ind]; 653 ring = priv->tx_ring[tx_ind];
626 if (vlan_tx_tag_present(skb)) 654 if (vlan_tx_tag_present(skb))
627 vlan_tag = vlan_tx_tag_get(skb); 655 vlan_tag = vlan_tx_tag_get(skb);
628 656
diff --git a/drivers/net/ethernet/mellanox/mlx4/eq.c b/drivers/net/ethernet/mellanox/mlx4/eq.c
index 0416c5b3b35c..c9cdb2a2c596 100644
--- a/drivers/net/ethernet/mellanox/mlx4/eq.c
+++ b/drivers/net/ethernet/mellanox/mlx4/eq.c
@@ -936,7 +936,6 @@ static int mlx4_create_eq(struct mlx4_dev *dev, int nent,
936 if (err) 936 if (err)
937 goto err_out_free_mtt; 937 goto err_out_free_mtt;
938 938
939 memset(eq_context, 0, sizeof *eq_context);
940 eq_context->flags = cpu_to_be32(MLX4_EQ_STATUS_OK | 939 eq_context->flags = cpu_to_be32(MLX4_EQ_STATUS_OK |
941 MLX4_EQ_STATE_ARMED); 940 MLX4_EQ_STATE_ARMED);
942 eq_context->log_eq_size = ilog2(eq->nent); 941 eq_context->log_eq_size = ilog2(eq->nent);
diff --git a/drivers/net/ethernet/mellanox/mlx4/fw.c b/drivers/net/ethernet/mellanox/mlx4/fw.c
index 0d63daa2f422..fda26679f7d5 100644
--- a/drivers/net/ethernet/mellanox/mlx4/fw.c
+++ b/drivers/net/ethernet/mellanox/mlx4/fw.c
@@ -159,8 +159,6 @@ int mlx4_MOD_STAT_CFG(struct mlx4_dev *dev, struct mlx4_mod_stat_cfg *cfg)
159 return PTR_ERR(mailbox); 159 return PTR_ERR(mailbox);
160 inbox = mailbox->buf; 160 inbox = mailbox->buf;
161 161
162 memset(inbox, 0, MOD_STAT_CFG_IN_SIZE);
163
164 MLX4_PUT(inbox, cfg->log_pg_sz, MOD_STAT_CFG_PG_SZ_OFFSET); 162 MLX4_PUT(inbox, cfg->log_pg_sz, MOD_STAT_CFG_PG_SZ_OFFSET);
165 MLX4_PUT(inbox, cfg->log_pg_sz_m, MOD_STAT_CFG_PG_SZ_M_OFFSET); 163 MLX4_PUT(inbox, cfg->log_pg_sz_m, MOD_STAT_CFG_PG_SZ_M_OFFSET);
166 164
@@ -177,6 +175,7 @@ int mlx4_QUERY_FUNC_CAP_wrapper(struct mlx4_dev *dev, int slave,
177 struct mlx4_cmd_mailbox *outbox, 175 struct mlx4_cmd_mailbox *outbox,
178 struct mlx4_cmd_info *cmd) 176 struct mlx4_cmd_info *cmd)
179{ 177{
178 struct mlx4_priv *priv = mlx4_priv(dev);
180 u8 field; 179 u8 field;
181 u32 size; 180 u32 size;
182 int err = 0; 181 int err = 0;
@@ -185,18 +184,26 @@ int mlx4_QUERY_FUNC_CAP_wrapper(struct mlx4_dev *dev, int slave,
185#define QUERY_FUNC_CAP_NUM_PORTS_OFFSET 0x1 184#define QUERY_FUNC_CAP_NUM_PORTS_OFFSET 0x1
186#define QUERY_FUNC_CAP_PF_BHVR_OFFSET 0x4 185#define QUERY_FUNC_CAP_PF_BHVR_OFFSET 0x4
187#define QUERY_FUNC_CAP_FMR_OFFSET 0x8 186#define QUERY_FUNC_CAP_FMR_OFFSET 0x8
188#define QUERY_FUNC_CAP_QP_QUOTA_OFFSET 0x10 187#define QUERY_FUNC_CAP_QP_QUOTA_OFFSET_DEP 0x10
189#define QUERY_FUNC_CAP_CQ_QUOTA_OFFSET 0x14 188#define QUERY_FUNC_CAP_CQ_QUOTA_OFFSET_DEP 0x14
190#define QUERY_FUNC_CAP_SRQ_QUOTA_OFFSET 0x18 189#define QUERY_FUNC_CAP_SRQ_QUOTA_OFFSET_DEP 0x18
191#define QUERY_FUNC_CAP_MPT_QUOTA_OFFSET 0x20 190#define QUERY_FUNC_CAP_MPT_QUOTA_OFFSET_DEP 0x20
192#define QUERY_FUNC_CAP_MTT_QUOTA_OFFSET 0x24 191#define QUERY_FUNC_CAP_MTT_QUOTA_OFFSET_DEP 0x24
193#define QUERY_FUNC_CAP_MCG_QUOTA_OFFSET 0x28 192#define QUERY_FUNC_CAP_MCG_QUOTA_OFFSET_DEP 0x28
194#define QUERY_FUNC_CAP_MAX_EQ_OFFSET 0x2c 193#define QUERY_FUNC_CAP_MAX_EQ_OFFSET 0x2c
195#define QUERY_FUNC_CAP_RESERVED_EQ_OFFSET 0x30 194#define QUERY_FUNC_CAP_RESERVED_EQ_OFFSET 0x30
196 195
196#define QUERY_FUNC_CAP_QP_QUOTA_OFFSET 0x50
197#define QUERY_FUNC_CAP_CQ_QUOTA_OFFSET 0x54
198#define QUERY_FUNC_CAP_SRQ_QUOTA_OFFSET 0x58
199#define QUERY_FUNC_CAP_MPT_QUOTA_OFFSET 0x60
200#define QUERY_FUNC_CAP_MTT_QUOTA_OFFSET 0x64
201#define QUERY_FUNC_CAP_MCG_QUOTA_OFFSET 0x68
202
197#define QUERY_FUNC_CAP_FMR_FLAG 0x80 203#define QUERY_FUNC_CAP_FMR_FLAG 0x80
198#define QUERY_FUNC_CAP_FLAG_RDMA 0x40 204#define QUERY_FUNC_CAP_FLAG_RDMA 0x40
199#define QUERY_FUNC_CAP_FLAG_ETH 0x80 205#define QUERY_FUNC_CAP_FLAG_ETH 0x80
206#define QUERY_FUNC_CAP_FLAG_QUOTAS 0x10
200 207
201/* when opcode modifier = 1 */ 208/* when opcode modifier = 1 */
202#define QUERY_FUNC_CAP_PHYS_PORT_OFFSET 0x3 209#define QUERY_FUNC_CAP_PHYS_PORT_OFFSET 0x3
@@ -237,8 +244,9 @@ int mlx4_QUERY_FUNC_CAP_wrapper(struct mlx4_dev *dev, int slave,
237 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_QP1_PROXY); 244 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_QP1_PROXY);
238 245
239 } else if (vhcr->op_modifier == 0) { 246 } else if (vhcr->op_modifier == 0) {
240 /* enable rdma and ethernet interfaces */ 247 /* enable rdma and ethernet interfaces, and new quota locations */
241 field = (QUERY_FUNC_CAP_FLAG_ETH | QUERY_FUNC_CAP_FLAG_RDMA); 248 field = (QUERY_FUNC_CAP_FLAG_ETH | QUERY_FUNC_CAP_FLAG_RDMA |
249 QUERY_FUNC_CAP_FLAG_QUOTAS);
242 MLX4_PUT(outbox->buf, field, QUERY_FUNC_CAP_FLAGS_OFFSET); 250 MLX4_PUT(outbox->buf, field, QUERY_FUNC_CAP_FLAGS_OFFSET);
243 251
244 field = dev->caps.num_ports; 252 field = dev->caps.num_ports;
@@ -250,14 +258,20 @@ int mlx4_QUERY_FUNC_CAP_wrapper(struct mlx4_dev *dev, int slave,
250 field = 0; /* protected FMR support not available as yet */ 258 field = 0; /* protected FMR support not available as yet */
251 MLX4_PUT(outbox->buf, field, QUERY_FUNC_CAP_FMR_OFFSET); 259 MLX4_PUT(outbox->buf, field, QUERY_FUNC_CAP_FMR_OFFSET);
252 260
253 size = dev->caps.num_qps; 261 size = priv->mfunc.master.res_tracker.res_alloc[RES_QP].quota[slave];
254 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_QP_QUOTA_OFFSET); 262 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_QP_QUOTA_OFFSET);
263 size = dev->caps.num_qps;
264 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_QP_QUOTA_OFFSET_DEP);
255 265
256 size = dev->caps.num_srqs; 266 size = priv->mfunc.master.res_tracker.res_alloc[RES_SRQ].quota[slave];
257 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_SRQ_QUOTA_OFFSET); 267 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_SRQ_QUOTA_OFFSET);
268 size = dev->caps.num_srqs;
269 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_SRQ_QUOTA_OFFSET_DEP);
258 270
259 size = dev->caps.num_cqs; 271 size = priv->mfunc.master.res_tracker.res_alloc[RES_CQ].quota[slave];
260 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_CQ_QUOTA_OFFSET); 272 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_CQ_QUOTA_OFFSET);
273 size = dev->caps.num_cqs;
274 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_CQ_QUOTA_OFFSET_DEP);
261 275
262 size = dev->caps.num_eqs; 276 size = dev->caps.num_eqs;
263 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_MAX_EQ_OFFSET); 277 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_MAX_EQ_OFFSET);
@@ -265,14 +279,19 @@ int mlx4_QUERY_FUNC_CAP_wrapper(struct mlx4_dev *dev, int slave,
265 size = dev->caps.reserved_eqs; 279 size = dev->caps.reserved_eqs;
266 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_RESERVED_EQ_OFFSET); 280 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_RESERVED_EQ_OFFSET);
267 281
268 size = dev->caps.num_mpts; 282 size = priv->mfunc.master.res_tracker.res_alloc[RES_MPT].quota[slave];
269 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_MPT_QUOTA_OFFSET); 283 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_MPT_QUOTA_OFFSET);
284 size = dev->caps.num_mpts;
285 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_MPT_QUOTA_OFFSET_DEP);
270 286
271 size = dev->caps.num_mtts; 287 size = priv->mfunc.master.res_tracker.res_alloc[RES_MTT].quota[slave];
272 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_MTT_QUOTA_OFFSET); 288 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_MTT_QUOTA_OFFSET);
289 size = dev->caps.num_mtts;
290 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_MTT_QUOTA_OFFSET_DEP);
273 291
274 size = dev->caps.num_mgms + dev->caps.num_amgms; 292 size = dev->caps.num_mgms + dev->caps.num_amgms;
275 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_MCG_QUOTA_OFFSET); 293 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_MCG_QUOTA_OFFSET);
294 MLX4_PUT(outbox->buf, size, QUERY_FUNC_CAP_MCG_QUOTA_OFFSET_DEP);
276 295
277 } else 296 } else
278 err = -EINVAL; 297 err = -EINVAL;
@@ -287,7 +306,7 @@ int mlx4_QUERY_FUNC_CAP(struct mlx4_dev *dev, u32 gen_or_port,
287 u32 *outbox; 306 u32 *outbox;
288 u8 field, op_modifier; 307 u8 field, op_modifier;
289 u32 size; 308 u32 size;
290 int err = 0; 309 int err = 0, quotas = 0;
291 310
292 op_modifier = !!gen_or_port; /* 0 = general, 1 = logical port */ 311 op_modifier = !!gen_or_port; /* 0 = general, 1 = logical port */
293 312
@@ -311,6 +330,7 @@ int mlx4_QUERY_FUNC_CAP(struct mlx4_dev *dev, u32 gen_or_port,
311 goto out; 330 goto out;
312 } 331 }
313 func_cap->flags = field; 332 func_cap->flags = field;
333 quotas = !!(func_cap->flags & QUERY_FUNC_CAP_FLAG_QUOTAS);
314 334
315 MLX4_GET(field, outbox, QUERY_FUNC_CAP_NUM_PORTS_OFFSET); 335 MLX4_GET(field, outbox, QUERY_FUNC_CAP_NUM_PORTS_OFFSET);
316 func_cap->num_ports = field; 336 func_cap->num_ports = field;
@@ -318,29 +338,50 @@ int mlx4_QUERY_FUNC_CAP(struct mlx4_dev *dev, u32 gen_or_port,
318 MLX4_GET(size, outbox, QUERY_FUNC_CAP_PF_BHVR_OFFSET); 338 MLX4_GET(size, outbox, QUERY_FUNC_CAP_PF_BHVR_OFFSET);
319 func_cap->pf_context_behaviour = size; 339 func_cap->pf_context_behaviour = size;
320 340
321 MLX4_GET(size, outbox, QUERY_FUNC_CAP_QP_QUOTA_OFFSET); 341 if (quotas) {
322 func_cap->qp_quota = size & 0xFFFFFF; 342 MLX4_GET(size, outbox, QUERY_FUNC_CAP_QP_QUOTA_OFFSET);
343 func_cap->qp_quota = size & 0xFFFFFF;
323 344
324 MLX4_GET(size, outbox, QUERY_FUNC_CAP_SRQ_QUOTA_OFFSET); 345 MLX4_GET(size, outbox, QUERY_FUNC_CAP_SRQ_QUOTA_OFFSET);
325 func_cap->srq_quota = size & 0xFFFFFF; 346 func_cap->srq_quota = size & 0xFFFFFF;
326 347
327 MLX4_GET(size, outbox, QUERY_FUNC_CAP_CQ_QUOTA_OFFSET); 348 MLX4_GET(size, outbox, QUERY_FUNC_CAP_CQ_QUOTA_OFFSET);
328 func_cap->cq_quota = size & 0xFFFFFF; 349 func_cap->cq_quota = size & 0xFFFFFF;
329 350
351 MLX4_GET(size, outbox, QUERY_FUNC_CAP_MPT_QUOTA_OFFSET);
352 func_cap->mpt_quota = size & 0xFFFFFF;
353
354 MLX4_GET(size, outbox, QUERY_FUNC_CAP_MTT_QUOTA_OFFSET);
355 func_cap->mtt_quota = size & 0xFFFFFF;
356
357 MLX4_GET(size, outbox, QUERY_FUNC_CAP_MCG_QUOTA_OFFSET);
358 func_cap->mcg_quota = size & 0xFFFFFF;
359
360 } else {
361 MLX4_GET(size, outbox, QUERY_FUNC_CAP_QP_QUOTA_OFFSET_DEP);
362 func_cap->qp_quota = size & 0xFFFFFF;
363
364 MLX4_GET(size, outbox, QUERY_FUNC_CAP_SRQ_QUOTA_OFFSET_DEP);
365 func_cap->srq_quota = size & 0xFFFFFF;
366
367 MLX4_GET(size, outbox, QUERY_FUNC_CAP_CQ_QUOTA_OFFSET_DEP);
368 func_cap->cq_quota = size & 0xFFFFFF;
369
370 MLX4_GET(size, outbox, QUERY_FUNC_CAP_MPT_QUOTA_OFFSET_DEP);
371 func_cap->mpt_quota = size & 0xFFFFFF;
372
373 MLX4_GET(size, outbox, QUERY_FUNC_CAP_MTT_QUOTA_OFFSET_DEP);
374 func_cap->mtt_quota = size & 0xFFFFFF;
375
376 MLX4_GET(size, outbox, QUERY_FUNC_CAP_MCG_QUOTA_OFFSET_DEP);
377 func_cap->mcg_quota = size & 0xFFFFFF;
378 }
330 MLX4_GET(size, outbox, QUERY_FUNC_CAP_MAX_EQ_OFFSET); 379 MLX4_GET(size, outbox, QUERY_FUNC_CAP_MAX_EQ_OFFSET);
331 func_cap->max_eq = size & 0xFFFFFF; 380 func_cap->max_eq = size & 0xFFFFFF;
332 381
333 MLX4_GET(size, outbox, QUERY_FUNC_CAP_RESERVED_EQ_OFFSET); 382 MLX4_GET(size, outbox, QUERY_FUNC_CAP_RESERVED_EQ_OFFSET);
334 func_cap->reserved_eq = size & 0xFFFFFF; 383 func_cap->reserved_eq = size & 0xFFFFFF;
335 384
336 MLX4_GET(size, outbox, QUERY_FUNC_CAP_MPT_QUOTA_OFFSET);
337 func_cap->mpt_quota = size & 0xFFFFFF;
338
339 MLX4_GET(size, outbox, QUERY_FUNC_CAP_MTT_QUOTA_OFFSET);
340 func_cap->mtt_quota = size & 0xFFFFFF;
341
342 MLX4_GET(size, outbox, QUERY_FUNC_CAP_MCG_QUOTA_OFFSET);
343 func_cap->mcg_quota = size & 0xFFFFFF;
344 goto out; 385 goto out;
345 } 386 }
346 387
@@ -652,7 +693,7 @@ int mlx4_QUERY_DEV_CAP(struct mlx4_dev *dev, struct mlx4_dev_cap *dev_cap)
652 QUERY_DEV_CAP_RSVD_LKEY_OFFSET); 693 QUERY_DEV_CAP_RSVD_LKEY_OFFSET);
653 MLX4_GET(field, outbox, QUERY_DEV_CAP_FW_REASSIGN_MAC); 694 MLX4_GET(field, outbox, QUERY_DEV_CAP_FW_REASSIGN_MAC);
654 if (field & 1<<6) 695 if (field & 1<<6)
655 dev_cap->flags2 |= MLX4_DEV_CAP_FLAGS2_REASSIGN_MAC_EN; 696 dev_cap->flags2 |= MLX4_DEV_CAP_FLAG2_REASSIGN_MAC_EN;
656 MLX4_GET(dev_cap->max_icm_sz, outbox, 697 MLX4_GET(dev_cap->max_icm_sz, outbox,
657 QUERY_DEV_CAP_MAX_ICM_SZ_OFFSET); 698 QUERY_DEV_CAP_MAX_ICM_SZ_OFFSET);
658 if (dev_cap->flags & MLX4_DEV_CAP_FLAG_COUNTERS) 699 if (dev_cap->flags & MLX4_DEV_CAP_FLAG_COUNTERS)
@@ -924,7 +965,6 @@ int mlx4_map_cmd(struct mlx4_dev *dev, u16 op, struct mlx4_icm *icm, u64 virt)
924 mailbox = mlx4_alloc_cmd_mailbox(dev); 965 mailbox = mlx4_alloc_cmd_mailbox(dev);
925 if (IS_ERR(mailbox)) 966 if (IS_ERR(mailbox))
926 return PTR_ERR(mailbox); 967 return PTR_ERR(mailbox);
927 memset(mailbox->buf, 0, MLX4_MAILBOX_SIZE);
928 pages = mailbox->buf; 968 pages = mailbox->buf;
929 969
930 for (mlx4_icm_first(icm, &iter); 970 for (mlx4_icm_first(icm, &iter);
@@ -1273,8 +1313,6 @@ int mlx4_INIT_HCA(struct mlx4_dev *dev, struct mlx4_init_hca_param *param)
1273 return PTR_ERR(mailbox); 1313 return PTR_ERR(mailbox);
1274 inbox = mailbox->buf; 1314 inbox = mailbox->buf;
1275 1315
1276 memset(inbox, 0, INIT_HCA_IN_SIZE);
1277
1278 *((u8 *) mailbox->buf + INIT_HCA_VERSION_OFFSET) = INIT_HCA_VERSION; 1316 *((u8 *) mailbox->buf + INIT_HCA_VERSION_OFFSET) = INIT_HCA_VERSION;
1279 1317
1280 *((u8 *) mailbox->buf + INIT_HCA_CACHELINE_SZ_OFFSET) = 1318 *((u8 *) mailbox->buf + INIT_HCA_CACHELINE_SZ_OFFSET) =
@@ -1573,8 +1611,6 @@ int mlx4_INIT_PORT(struct mlx4_dev *dev, int port)
1573 return PTR_ERR(mailbox); 1611 return PTR_ERR(mailbox);
1574 inbox = mailbox->buf; 1612 inbox = mailbox->buf;
1575 1613
1576 memset(inbox, 0, INIT_PORT_IN_SIZE);
1577
1578 flags = 0; 1614 flags = 0;
1579 flags |= (dev->caps.vl_cap[port] & 0xf) << INIT_PORT_VL_SHIFT; 1615 flags |= (dev->caps.vl_cap[port] & 0xf) << INIT_PORT_VL_SHIFT;
1580 flags |= (dev->caps.port_width_cap[port] & 0xf) << INIT_PORT_PORT_WIDTH_SHIFT; 1616 flags |= (dev->caps.port_width_cap[port] & 0xf) << INIT_PORT_PORT_WIDTH_SHIFT;
@@ -1713,7 +1749,6 @@ void mlx4_opreq_action(struct work_struct *work)
1713 u32 *outbox; 1749 u32 *outbox;
1714 u32 modifier; 1750 u32 modifier;
1715 u16 token; 1751 u16 token;
1716 u16 type_m;
1717 u16 type; 1752 u16 type;
1718 int err; 1753 int err;
1719 u32 num_qps; 1754 u32 num_qps;
@@ -1746,7 +1781,6 @@ void mlx4_opreq_action(struct work_struct *work)
1746 MLX4_GET(modifier, outbox, GET_OP_REQ_MODIFIER_OFFSET); 1781 MLX4_GET(modifier, outbox, GET_OP_REQ_MODIFIER_OFFSET);
1747 MLX4_GET(token, outbox, GET_OP_REQ_TOKEN_OFFSET); 1782 MLX4_GET(token, outbox, GET_OP_REQ_TOKEN_OFFSET);
1748 MLX4_GET(type, outbox, GET_OP_REQ_TYPE_OFFSET); 1783 MLX4_GET(type, outbox, GET_OP_REQ_TYPE_OFFSET);
1749 type_m = type >> 12;
1750 type &= 0xfff; 1784 type &= 0xfff;
1751 1785
1752 switch (type) { 1786 switch (type) {
diff --git a/drivers/net/ethernet/mellanox/mlx4/icm.c b/drivers/net/ethernet/mellanox/mlx4/icm.c
index 31d02649be41..5fbf4924c272 100644
--- a/drivers/net/ethernet/mellanox/mlx4/icm.c
+++ b/drivers/net/ethernet/mellanox/mlx4/icm.c
@@ -93,13 +93,17 @@ void mlx4_free_icm(struct mlx4_dev *dev, struct mlx4_icm *icm, int coherent)
93 kfree(icm); 93 kfree(icm);
94} 94}
95 95
96static int mlx4_alloc_icm_pages(struct scatterlist *mem, int order, gfp_t gfp_mask) 96static int mlx4_alloc_icm_pages(struct scatterlist *mem, int order,
97 gfp_t gfp_mask, int node)
97{ 98{
98 struct page *page; 99 struct page *page;
99 100
100 page = alloc_pages(gfp_mask, order); 101 page = alloc_pages_node(node, gfp_mask, order);
101 if (!page) 102 if (!page) {
102 return -ENOMEM; 103 page = alloc_pages(gfp_mask, order);
104 if (!page)
105 return -ENOMEM;
106 }
103 107
104 sg_set_page(mem, page, PAGE_SIZE << order, 0); 108 sg_set_page(mem, page, PAGE_SIZE << order, 0);
105 return 0; 109 return 0;
@@ -130,9 +134,15 @@ struct mlx4_icm *mlx4_alloc_icm(struct mlx4_dev *dev, int npages,
130 /* We use sg_set_buf for coherent allocs, which assumes low memory */ 134 /* We use sg_set_buf for coherent allocs, which assumes low memory */
131 BUG_ON(coherent && (gfp_mask & __GFP_HIGHMEM)); 135 BUG_ON(coherent && (gfp_mask & __GFP_HIGHMEM));
132 136
133 icm = kmalloc(sizeof *icm, gfp_mask & ~(__GFP_HIGHMEM | __GFP_NOWARN)); 137 icm = kmalloc_node(sizeof(*icm),
134 if (!icm) 138 gfp_mask & ~(__GFP_HIGHMEM | __GFP_NOWARN),
135 return NULL; 139 dev->numa_node);
140 if (!icm) {
141 icm = kmalloc(sizeof(*icm),
142 gfp_mask & ~(__GFP_HIGHMEM | __GFP_NOWARN));
143 if (!icm)
144 return NULL;
145 }
136 146
137 icm->refcount = 0; 147 icm->refcount = 0;
138 INIT_LIST_HEAD(&icm->chunk_list); 148 INIT_LIST_HEAD(&icm->chunk_list);
@@ -141,10 +151,17 @@ struct mlx4_icm *mlx4_alloc_icm(struct mlx4_dev *dev, int npages,
141 151
142 while (npages > 0) { 152 while (npages > 0) {
143 if (!chunk) { 153 if (!chunk) {
144 chunk = kmalloc(sizeof *chunk, 154 chunk = kmalloc_node(sizeof(*chunk),
145 gfp_mask & ~(__GFP_HIGHMEM | __GFP_NOWARN)); 155 gfp_mask & ~(__GFP_HIGHMEM |
146 if (!chunk) 156 __GFP_NOWARN),
147 goto fail; 157 dev->numa_node);
158 if (!chunk) {
159 chunk = kmalloc(sizeof(*chunk),
160 gfp_mask & ~(__GFP_HIGHMEM |
161 __GFP_NOWARN));
162 if (!chunk)
163 goto fail;
164 }
148 165
149 sg_init_table(chunk->mem, MLX4_ICM_CHUNK_LEN); 166 sg_init_table(chunk->mem, MLX4_ICM_CHUNK_LEN);
150 chunk->npages = 0; 167 chunk->npages = 0;
@@ -161,7 +178,8 @@ struct mlx4_icm *mlx4_alloc_icm(struct mlx4_dev *dev, int npages,
161 cur_order, gfp_mask); 178 cur_order, gfp_mask);
162 else 179 else
163 ret = mlx4_alloc_icm_pages(&chunk->mem[chunk->npages], 180 ret = mlx4_alloc_icm_pages(&chunk->mem[chunk->npages],
164 cur_order, gfp_mask); 181 cur_order, gfp_mask,
182 dev->numa_node);
165 183
166 if (ret) { 184 if (ret) {
167 if (--cur_order < 0) 185 if (--cur_order < 0)
diff --git a/drivers/net/ethernet/mellanox/mlx4/main.c b/drivers/net/ethernet/mellanox/mlx4/main.c
index 60c9f4f103fc..5789ea2c934d 100644
--- a/drivers/net/ethernet/mellanox/mlx4/main.c
+++ b/drivers/net/ethernet/mellanox/mlx4/main.c
@@ -42,6 +42,7 @@
42#include <linux/io-mapping.h> 42#include <linux/io-mapping.h>
43#include <linux/delay.h> 43#include <linux/delay.h>
44#include <linux/netdevice.h> 44#include <linux/netdevice.h>
45#include <linux/kmod.h>
45 46
46#include <linux/mlx4/device.h> 47#include <linux/mlx4/device.h>
47#include <linux/mlx4/doorbell.h> 48#include <linux/mlx4/doorbell.h>
@@ -561,13 +562,17 @@ static int mlx4_slave_cap(struct mlx4_dev *dev)
561 } 562 }
562 563
563 dev->caps.num_ports = func_cap.num_ports; 564 dev->caps.num_ports = func_cap.num_ports;
564 dev->caps.num_qps = func_cap.qp_quota; 565 dev->quotas.qp = func_cap.qp_quota;
565 dev->caps.num_srqs = func_cap.srq_quota; 566 dev->quotas.srq = func_cap.srq_quota;
566 dev->caps.num_cqs = func_cap.cq_quota; 567 dev->quotas.cq = func_cap.cq_quota;
567 dev->caps.num_eqs = func_cap.max_eq; 568 dev->quotas.mpt = func_cap.mpt_quota;
568 dev->caps.reserved_eqs = func_cap.reserved_eq; 569 dev->quotas.mtt = func_cap.mtt_quota;
569 dev->caps.num_mpts = func_cap.mpt_quota; 570 dev->caps.num_qps = 1 << hca_param.log_num_qps;
570 dev->caps.num_mtts = func_cap.mtt_quota; 571 dev->caps.num_srqs = 1 << hca_param.log_num_srqs;
572 dev->caps.num_cqs = 1 << hca_param.log_num_cqs;
573 dev->caps.num_mpts = 1 << hca_param.log_mpt_sz;
574 dev->caps.num_eqs = func_cap.max_eq;
575 dev->caps.reserved_eqs = func_cap.reserved_eq;
571 dev->caps.num_pds = MLX4_NUM_PDS; 576 dev->caps.num_pds = MLX4_NUM_PDS;
572 dev->caps.num_mgms = 0; 577 dev->caps.num_mgms = 0;
573 dev->caps.num_amgms = 0; 578 dev->caps.num_amgms = 0;
@@ -650,6 +655,27 @@ err_mem:
650 return err; 655 return err;
651} 656}
652 657
658static void mlx4_request_modules(struct mlx4_dev *dev)
659{
660 int port;
661 int has_ib_port = false;
662 int has_eth_port = false;
663#define EN_DRV_NAME "mlx4_en"
664#define IB_DRV_NAME "mlx4_ib"
665
666 for (port = 1; port <= dev->caps.num_ports; port++) {
667 if (dev->caps.port_type[port] == MLX4_PORT_TYPE_IB)
668 has_ib_port = true;
669 else if (dev->caps.port_type[port] == MLX4_PORT_TYPE_ETH)
670 has_eth_port = true;
671 }
672
673 if (has_ib_port)
674 request_module_nowait(IB_DRV_NAME);
675 if (has_eth_port)
676 request_module_nowait(EN_DRV_NAME);
677}
678
653/* 679/*
654 * Change the port configuration of the device. 680 * Change the port configuration of the device.
655 * Every user of this function must hold the port mutex. 681 * Every user of this function must hold the port mutex.
@@ -681,6 +707,11 @@ int mlx4_change_port_types(struct mlx4_dev *dev,
681 } 707 }
682 mlx4_set_port_mask(dev); 708 mlx4_set_port_mask(dev);
683 err = mlx4_register_device(dev); 709 err = mlx4_register_device(dev);
710 if (err) {
711 mlx4_err(dev, "Failed to register device\n");
712 goto out;
713 }
714 mlx4_request_modules(dev);
684 } 715 }
685 716
686out: 717out:
@@ -2075,9 +2106,15 @@ static int __mlx4_init_one(struct pci_dev *pdev, int pci_dev_data)
2075 "aborting.\n"); 2106 "aborting.\n");
2076 return err; 2107 return err;
2077 } 2108 }
2078 if (num_vfs > MLX4_MAX_NUM_VF) { 2109
2079 printk(KERN_ERR "There are more VF's (%d) than allowed(%d)\n", 2110 /* Due to requirement that all VFs and the PF are *guaranteed* 2 MACS
2080 num_vfs, MLX4_MAX_NUM_VF); 2111 * per port, we must limit the number of VFs to 63 (since their are
2112 * 128 MACs)
2113 */
2114 if (num_vfs >= MLX4_MAX_NUM_VF) {
2115 dev_err(&pdev->dev,
2116 "Requested more VF's (%d) than allowed (%d)\n",
2117 num_vfs, MLX4_MAX_NUM_VF - 1);
2081 return -EINVAL; 2118 return -EINVAL;
2082 } 2119 }
2083 2120
@@ -2154,6 +2191,7 @@ static int __mlx4_init_one(struct pci_dev *pdev, int pci_dev_data)
2154 mutex_init(&priv->bf_mutex); 2191 mutex_init(&priv->bf_mutex);
2155 2192
2156 dev->rev_id = pdev->revision; 2193 dev->rev_id = pdev->revision;
2194 dev->numa_node = dev_to_node(&pdev->dev);
2157 /* Detect if this device is a virtual function */ 2195 /* Detect if this device is a virtual function */
2158 if (pci_dev_data & MLX4_PCI_DEV_IS_VF) { 2196 if (pci_dev_data & MLX4_PCI_DEV_IS_VF) {
2159 /* When acting as pf, we normally skip vfs unless explicitly 2197 /* When acting as pf, we normally skip vfs unless explicitly
@@ -2295,6 +2333,8 @@ slave_start:
2295 if (err) 2333 if (err)
2296 goto err_steer; 2334 goto err_steer;
2297 2335
2336 mlx4_init_quotas(dev);
2337
2298 for (port = 1; port <= dev->caps.num_ports; port++) { 2338 for (port = 1; port <= dev->caps.num_ports; port++) {
2299 err = mlx4_init_port_info(dev, port); 2339 err = mlx4_init_port_info(dev, port);
2300 if (err) 2340 if (err)
@@ -2305,6 +2345,8 @@ slave_start:
2305 if (err) 2345 if (err)
2306 goto err_port; 2346 goto err_port;
2307 2347
2348 mlx4_request_modules(dev);
2349
2308 mlx4_sense_init(dev); 2350 mlx4_sense_init(dev);
2309 mlx4_start_sense(dev); 2351 mlx4_start_sense(dev);
2310 2352
diff --git a/drivers/net/ethernet/mellanox/mlx4/mcg.c b/drivers/net/ethernet/mellanox/mlx4/mcg.c
index 55f6245efb6c..acf9d5f1f922 100644
--- a/drivers/net/ethernet/mellanox/mlx4/mcg.c
+++ b/drivers/net/ethernet/mellanox/mlx4/mcg.c
@@ -506,7 +506,6 @@ static int remove_promisc_qp(struct mlx4_dev *dev, u8 port,
506 goto out_list; 506 goto out_list;
507 } 507 }
508 mgm = mailbox->buf; 508 mgm = mailbox->buf;
509 memset(mgm, 0, sizeof *mgm);
510 members_count = 0; 509 members_count = 0;
511 list_for_each_entry(dqp, &s_steer->promisc_qps[steer], list) 510 list_for_each_entry(dqp, &s_steer->promisc_qps[steer], list)
512 mgm->qp[members_count++] = cpu_to_be32(dqp->qpn & MGM_QPN_MASK); 511 mgm->qp[members_count++] = cpu_to_be32(dqp->qpn & MGM_QPN_MASK);
@@ -645,7 +644,7 @@ static const u8 __promisc_mode[] = {
645int mlx4_map_sw_to_hw_steering_mode(struct mlx4_dev *dev, 644int mlx4_map_sw_to_hw_steering_mode(struct mlx4_dev *dev,
646 enum mlx4_net_trans_promisc_mode flow_type) 645 enum mlx4_net_trans_promisc_mode flow_type)
647{ 646{
648 if (flow_type >= MLX4_FS_MODE_NUM || flow_type < 0) { 647 if (flow_type >= MLX4_FS_MODE_NUM) {
649 mlx4_err(dev, "Invalid flow type. type = %d\n", flow_type); 648 mlx4_err(dev, "Invalid flow type. type = %d\n", flow_type);
650 return -EINVAL; 649 return -EINVAL;
651 } 650 }
@@ -681,7 +680,7 @@ const u16 __sw_id_hw[] = {
681int mlx4_map_sw_to_hw_steering_id(struct mlx4_dev *dev, 680int mlx4_map_sw_to_hw_steering_id(struct mlx4_dev *dev,
682 enum mlx4_net_trans_rule_id id) 681 enum mlx4_net_trans_rule_id id)
683{ 682{
684 if (id >= MLX4_NET_TRANS_RULE_NUM || id < 0) { 683 if (id >= MLX4_NET_TRANS_RULE_NUM) {
685 mlx4_err(dev, "Invalid network rule id. id = %d\n", id); 684 mlx4_err(dev, "Invalid network rule id. id = %d\n", id);
686 return -EINVAL; 685 return -EINVAL;
687 } 686 }
@@ -706,7 +705,7 @@ static const int __rule_hw_sz[] = {
706int mlx4_hw_rule_sz(struct mlx4_dev *dev, 705int mlx4_hw_rule_sz(struct mlx4_dev *dev,
707 enum mlx4_net_trans_rule_id id) 706 enum mlx4_net_trans_rule_id id)
708{ 707{
709 if (id >= MLX4_NET_TRANS_RULE_NUM || id < 0) { 708 if (id >= MLX4_NET_TRANS_RULE_NUM) {
710 mlx4_err(dev, "Invalid network rule id. id = %d\n", id); 709 mlx4_err(dev, "Invalid network rule id. id = %d\n", id);
711 return -EINVAL; 710 return -EINVAL;
712 } 711 }
@@ -857,7 +856,6 @@ int mlx4_flow_attach(struct mlx4_dev *dev,
857 if (IS_ERR(mailbox)) 856 if (IS_ERR(mailbox))
858 return PTR_ERR(mailbox); 857 return PTR_ERR(mailbox);
859 858
860 memset(mailbox->buf, 0, sizeof(struct mlx4_net_trans_rule_hw_ctrl));
861 trans_rule_ctrl_to_hw(rule, mailbox->buf); 859 trans_rule_ctrl_to_hw(rule, mailbox->buf);
862 860
863 size += sizeof(struct mlx4_net_trans_rule_hw_ctrl); 861 size += sizeof(struct mlx4_net_trans_rule_hw_ctrl);
diff --git a/drivers/net/ethernet/mellanox/mlx4/mlx4.h b/drivers/net/ethernet/mellanox/mlx4/mlx4.h
index 348bb8c7d9a7..e582a41a802b 100644
--- a/drivers/net/ethernet/mellanox/mlx4/mlx4.h
+++ b/drivers/net/ethernet/mellanox/mlx4/mlx4.h
@@ -455,6 +455,7 @@ struct mlx4_slave_state {
455 u8 last_cmd; 455 u8 last_cmd;
456 u8 init_port_mask; 456 u8 init_port_mask;
457 bool active; 457 bool active;
458 bool old_vlan_api;
458 u8 function; 459 u8 function;
459 dma_addr_t vhcr_dma; 460 dma_addr_t vhcr_dma;
460 u16 mtu[MLX4_MAX_PORTS + 1]; 461 u16 mtu[MLX4_MAX_PORTS + 1];
@@ -503,12 +504,28 @@ struct slave_list {
503 struct list_head res_list[MLX4_NUM_OF_RESOURCE_TYPE]; 504 struct list_head res_list[MLX4_NUM_OF_RESOURCE_TYPE];
504}; 505};
505 506
507struct resource_allocator {
508 spinlock_t alloc_lock; /* protect quotas */
509 union {
510 int res_reserved;
511 int res_port_rsvd[MLX4_MAX_PORTS];
512 };
513 union {
514 int res_free;
515 int res_port_free[MLX4_MAX_PORTS];
516 };
517 int *quota;
518 int *allocated;
519 int *guaranteed;
520};
521
506struct mlx4_resource_tracker { 522struct mlx4_resource_tracker {
507 spinlock_t lock; 523 spinlock_t lock;
508 /* tree for each resources */ 524 /* tree for each resources */
509 struct rb_root res_tree[MLX4_NUM_OF_RESOURCE_TYPE]; 525 struct rb_root res_tree[MLX4_NUM_OF_RESOURCE_TYPE];
510 /* num_of_slave's lists, one per slave */ 526 /* num_of_slave's lists, one per slave */
511 struct slave_list *slave_list; 527 struct slave_list *slave_list;
528 struct resource_allocator res_alloc[MLX4_NUM_OF_RESOURCE_TYPE];
512}; 529};
513 530
514#define SLAVE_EVENT_EQ_SIZE 128 531#define SLAVE_EVENT_EQ_SIZE 128
@@ -1111,7 +1128,7 @@ int mlx4_change_port_types(struct mlx4_dev *dev,
1111 1128
1112void mlx4_init_mac_table(struct mlx4_dev *dev, struct mlx4_mac_table *table); 1129void mlx4_init_mac_table(struct mlx4_dev *dev, struct mlx4_mac_table *table);
1113void mlx4_init_vlan_table(struct mlx4_dev *dev, struct mlx4_vlan_table *table); 1130void mlx4_init_vlan_table(struct mlx4_dev *dev, struct mlx4_vlan_table *table);
1114void __mlx4_unregister_vlan(struct mlx4_dev *dev, u8 port, int index); 1131void __mlx4_unregister_vlan(struct mlx4_dev *dev, u8 port, u16 vlan);
1115int __mlx4_register_vlan(struct mlx4_dev *dev, u8 port, u16 vlan, int *index); 1132int __mlx4_register_vlan(struct mlx4_dev *dev, u8 port, u16 vlan, int *index);
1116 1133
1117int mlx4_SET_PORT(struct mlx4_dev *dev, u8 port, int pkey_tbl_sz); 1134int mlx4_SET_PORT(struct mlx4_dev *dev, u8 port, int pkey_tbl_sz);
@@ -1252,4 +1269,6 @@ static inline spinlock_t *mlx4_tlock(struct mlx4_dev *dev)
1252 1269
1253void mlx4_vf_immed_vlan_work_handler(struct work_struct *_work); 1270void mlx4_vf_immed_vlan_work_handler(struct work_struct *_work);
1254 1271
1272void mlx4_init_quotas(struct mlx4_dev *dev);
1273
1255#endif /* MLX4_H */ 1274#endif /* MLX4_H */
diff --git a/drivers/net/ethernet/mellanox/mlx4/mlx4_en.h b/drivers/net/ethernet/mellanox/mlx4/mlx4_en.h
index bf06e3610d27..f3758de59c05 100644
--- a/drivers/net/ethernet/mellanox/mlx4/mlx4_en.h
+++ b/drivers/net/ethernet/mellanox/mlx4/mlx4_en.h
@@ -530,10 +530,10 @@ struct mlx4_en_priv {
530 u16 num_frags; 530 u16 num_frags;
531 u16 log_rx_info; 531 u16 log_rx_info;
532 532
533 struct mlx4_en_tx_ring *tx_ring; 533 struct mlx4_en_tx_ring **tx_ring;
534 struct mlx4_en_rx_ring rx_ring[MAX_RX_RINGS]; 534 struct mlx4_en_rx_ring *rx_ring[MAX_RX_RINGS];
535 struct mlx4_en_cq *tx_cq; 535 struct mlx4_en_cq **tx_cq;
536 struct mlx4_en_cq rx_cq[MAX_RX_RINGS]; 536 struct mlx4_en_cq *rx_cq[MAX_RX_RINGS];
537 struct mlx4_qp drop_qp; 537 struct mlx4_qp drop_qp;
538 struct work_struct rx_mode_task; 538 struct work_struct rx_mode_task;
539 struct work_struct watchdog_task; 539 struct work_struct watchdog_task;
@@ -626,7 +626,7 @@ static inline bool mlx4_en_cq_lock_poll(struct mlx4_en_cq *cq)
626 if ((cq->state & MLX4_CQ_LOCKED)) { 626 if ((cq->state & MLX4_CQ_LOCKED)) {
627 struct net_device *dev = cq->dev; 627 struct net_device *dev = cq->dev;
628 struct mlx4_en_priv *priv = netdev_priv(dev); 628 struct mlx4_en_priv *priv = netdev_priv(dev);
629 struct mlx4_en_rx_ring *rx_ring = &priv->rx_ring[cq->ring]; 629 struct mlx4_en_rx_ring *rx_ring = priv->rx_ring[cq->ring];
630 630
631 cq->state |= MLX4_EN_CQ_STATE_POLL_YIELD; 631 cq->state |= MLX4_EN_CQ_STATE_POLL_YIELD;
632 rc = false; 632 rc = false;
@@ -704,9 +704,9 @@ void mlx4_en_stop_port(struct net_device *dev, int detach);
704void mlx4_en_free_resources(struct mlx4_en_priv *priv); 704void mlx4_en_free_resources(struct mlx4_en_priv *priv);
705int mlx4_en_alloc_resources(struct mlx4_en_priv *priv); 705int mlx4_en_alloc_resources(struct mlx4_en_priv *priv);
706 706
707int mlx4_en_create_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq, 707int mlx4_en_create_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq **pcq,
708 int entries, int ring, enum cq_type mode); 708 int entries, int ring, enum cq_type mode, int node);
709void mlx4_en_destroy_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq); 709void mlx4_en_destroy_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq **pcq);
710int mlx4_en_activate_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq, 710int mlx4_en_activate_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq,
711 int cq_idx); 711 int cq_idx);
712void mlx4_en_deactivate_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq); 712void mlx4_en_deactivate_cq(struct mlx4_en_priv *priv, struct mlx4_en_cq *cq);
@@ -717,9 +717,11 @@ void mlx4_en_tx_irq(struct mlx4_cq *mcq);
717u16 mlx4_en_select_queue(struct net_device *dev, struct sk_buff *skb); 717u16 mlx4_en_select_queue(struct net_device *dev, struct sk_buff *skb);
718netdev_tx_t mlx4_en_xmit(struct sk_buff *skb, struct net_device *dev); 718netdev_tx_t mlx4_en_xmit(struct sk_buff *skb, struct net_device *dev);
719 719
720int mlx4_en_create_tx_ring(struct mlx4_en_priv *priv, struct mlx4_en_tx_ring *ring, 720int mlx4_en_create_tx_ring(struct mlx4_en_priv *priv,
721 int qpn, u32 size, u16 stride); 721 struct mlx4_en_tx_ring **pring,
722void mlx4_en_destroy_tx_ring(struct mlx4_en_priv *priv, struct mlx4_en_tx_ring *ring); 722 int qpn, u32 size, u16 stride, int node);
723void mlx4_en_destroy_tx_ring(struct mlx4_en_priv *priv,
724 struct mlx4_en_tx_ring **pring);
723int mlx4_en_activate_tx_ring(struct mlx4_en_priv *priv, 725int mlx4_en_activate_tx_ring(struct mlx4_en_priv *priv,
724 struct mlx4_en_tx_ring *ring, 726 struct mlx4_en_tx_ring *ring,
725 int cq, int user_prio); 727 int cq, int user_prio);
@@ -727,10 +729,10 @@ void mlx4_en_deactivate_tx_ring(struct mlx4_en_priv *priv,
727 struct mlx4_en_tx_ring *ring); 729 struct mlx4_en_tx_ring *ring);
728 730
729int mlx4_en_create_rx_ring(struct mlx4_en_priv *priv, 731int mlx4_en_create_rx_ring(struct mlx4_en_priv *priv,
730 struct mlx4_en_rx_ring *ring, 732 struct mlx4_en_rx_ring **pring,
731 u32 size, u16 stride); 733 u32 size, u16 stride, int node);
732void mlx4_en_destroy_rx_ring(struct mlx4_en_priv *priv, 734void mlx4_en_destroy_rx_ring(struct mlx4_en_priv *priv,
733 struct mlx4_en_rx_ring *ring, 735 struct mlx4_en_rx_ring **pring,
734 u32 size, u16 stride); 736 u32 size, u16 stride);
735int mlx4_en_activate_rx_rings(struct mlx4_en_priv *priv); 737int mlx4_en_activate_rx_rings(struct mlx4_en_priv *priv);
736void mlx4_en_deactivate_rx_ring(struct mlx4_en_priv *priv, 738void mlx4_en_deactivate_rx_ring(struct mlx4_en_priv *priv,
@@ -768,8 +770,7 @@ extern const struct dcbnl_rtnl_ops mlx4_en_dcbnl_pfc_ops;
768int mlx4_en_setup_tc(struct net_device *dev, u8 up); 770int mlx4_en_setup_tc(struct net_device *dev, u8 up);
769 771
770#ifdef CONFIG_RFS_ACCEL 772#ifdef CONFIG_RFS_ACCEL
771void mlx4_en_cleanup_filters(struct mlx4_en_priv *priv, 773void mlx4_en_cleanup_filters(struct mlx4_en_priv *priv);
772 struct mlx4_en_rx_ring *rx_ring);
773#endif 774#endif
774 775
775#define MLX4_EN_NUM_SELF_TEST 5 776#define MLX4_EN_NUM_SELF_TEST 5
diff --git a/drivers/net/ethernet/mellanox/mlx4/mr.c b/drivers/net/ethernet/mellanox/mlx4/mr.c
index f91719a08cba..b3ee9bafff5e 100644
--- a/drivers/net/ethernet/mellanox/mlx4/mr.c
+++ b/drivers/net/ethernet/mellanox/mlx4/mr.c
@@ -480,9 +480,6 @@ int mlx4_mr_enable(struct mlx4_dev *dev, struct mlx4_mr *mr)
480 goto err_table; 480 goto err_table;
481 } 481 }
482 mpt_entry = mailbox->buf; 482 mpt_entry = mailbox->buf;
483
484 memset(mpt_entry, 0, sizeof *mpt_entry);
485
486 mpt_entry->flags = cpu_to_be32(MLX4_MPT_FLAG_MIO | 483 mpt_entry->flags = cpu_to_be32(MLX4_MPT_FLAG_MIO |
487 MLX4_MPT_FLAG_REGION | 484 MLX4_MPT_FLAG_REGION |
488 mr->access); 485 mr->access);
@@ -695,8 +692,6 @@ int mlx4_mw_enable(struct mlx4_dev *dev, struct mlx4_mw *mw)
695 } 692 }
696 mpt_entry = mailbox->buf; 693 mpt_entry = mailbox->buf;
697 694
698 memset(mpt_entry, 0, sizeof(*mpt_entry));
699
700 /* Note that the MLX4_MPT_FLAG_REGION bit in mpt_entry->flags is turned 695 /* Note that the MLX4_MPT_FLAG_REGION bit in mpt_entry->flags is turned
701 * off, thus creating a memory window and not a memory region. 696 * off, thus creating a memory window and not a memory region.
702 */ 697 */
@@ -755,14 +750,14 @@ int mlx4_init_mr_table(struct mlx4_dev *dev)
755 struct mlx4_mr_table *mr_table = &priv->mr_table; 750 struct mlx4_mr_table *mr_table = &priv->mr_table;
756 int err; 751 int err;
757 752
758 if (!is_power_of_2(dev->caps.num_mpts))
759 return -EINVAL;
760
761 /* Nothing to do for slaves - all MR handling is forwarded 753 /* Nothing to do for slaves - all MR handling is forwarded
762 * to the master */ 754 * to the master */
763 if (mlx4_is_slave(dev)) 755 if (mlx4_is_slave(dev))
764 return 0; 756 return 0;
765 757
758 if (!is_power_of_2(dev->caps.num_mpts))
759 return -EINVAL;
760
766 err = mlx4_bitmap_init(&mr_table->mpt_bitmap, dev->caps.num_mpts, 761 err = mlx4_bitmap_init(&mr_table->mpt_bitmap, dev->caps.num_mpts,
767 ~0, dev->caps.reserved_mrws, 0); 762 ~0, dev->caps.reserved_mrws, 0);
768 if (err) 763 if (err)
diff --git a/drivers/net/ethernet/mellanox/mlx4/pd.c b/drivers/net/ethernet/mellanox/mlx4/pd.c
index 00f223acada7..84cfb40bf451 100644
--- a/drivers/net/ethernet/mellanox/mlx4/pd.c
+++ b/drivers/net/ethernet/mellanox/mlx4/pd.c
@@ -168,7 +168,7 @@ void mlx4_uar_free(struct mlx4_dev *dev, struct mlx4_uar *uar)
168} 168}
169EXPORT_SYMBOL_GPL(mlx4_uar_free); 169EXPORT_SYMBOL_GPL(mlx4_uar_free);
170 170
171int mlx4_bf_alloc(struct mlx4_dev *dev, struct mlx4_bf *bf) 171int mlx4_bf_alloc(struct mlx4_dev *dev, struct mlx4_bf *bf, int node)
172{ 172{
173 struct mlx4_priv *priv = mlx4_priv(dev); 173 struct mlx4_priv *priv = mlx4_priv(dev);
174 struct mlx4_uar *uar; 174 struct mlx4_uar *uar;
@@ -186,10 +186,13 @@ int mlx4_bf_alloc(struct mlx4_dev *dev, struct mlx4_bf *bf)
186 err = -ENOMEM; 186 err = -ENOMEM;
187 goto out; 187 goto out;
188 } 188 }
189 uar = kmalloc(sizeof *uar, GFP_KERNEL); 189 uar = kmalloc_node(sizeof(*uar), GFP_KERNEL, node);
190 if (!uar) { 190 if (!uar) {
191 err = -ENOMEM; 191 uar = kmalloc(sizeof(*uar), GFP_KERNEL);
192 goto out; 192 if (!uar) {
193 err = -ENOMEM;
194 goto out;
195 }
193 } 196 }
194 err = mlx4_uar_alloc(dev, uar); 197 err = mlx4_uar_alloc(dev, uar);
195 if (err) 198 if (err)
diff --git a/drivers/net/ethernet/mellanox/mlx4/port.c b/drivers/net/ethernet/mellanox/mlx4/port.c
index 946e0af5faef..97d342fa5032 100644
--- a/drivers/net/ethernet/mellanox/mlx4/port.c
+++ b/drivers/net/ethernet/mellanox/mlx4/port.c
@@ -178,13 +178,24 @@ EXPORT_SYMBOL_GPL(__mlx4_register_mac);
178int mlx4_register_mac(struct mlx4_dev *dev, u8 port, u64 mac) 178int mlx4_register_mac(struct mlx4_dev *dev, u8 port, u64 mac)
179{ 179{
180 u64 out_param = 0; 180 u64 out_param = 0;
181 int err; 181 int err = -EINVAL;
182 182
183 if (mlx4_is_mfunc(dev)) { 183 if (mlx4_is_mfunc(dev)) {
184 set_param_l(&out_param, port); 184 if (!(dev->flags & MLX4_FLAG_OLD_REG_MAC)) {
185 err = mlx4_cmd_imm(dev, mac, &out_param, RES_MAC, 185 err = mlx4_cmd_imm(dev, mac, &out_param,
186 RES_OP_RESERVE_AND_MAP, MLX4_CMD_ALLOC_RES, 186 ((u32) port) << 8 | (u32) RES_MAC,
187 MLX4_CMD_TIME_CLASS_A, MLX4_CMD_WRAPPED); 187 RES_OP_RESERVE_AND_MAP, MLX4_CMD_ALLOC_RES,
188 MLX4_CMD_TIME_CLASS_A, MLX4_CMD_WRAPPED);
189 }
190 if (err && err == -EINVAL && mlx4_is_slave(dev)) {
191 /* retry using old REG_MAC format */
192 set_param_l(&out_param, port);
193 err = mlx4_cmd_imm(dev, mac, &out_param, RES_MAC,
194 RES_OP_RESERVE_AND_MAP, MLX4_CMD_ALLOC_RES,
195 MLX4_CMD_TIME_CLASS_A, MLX4_CMD_WRAPPED);
196 if (!err)
197 dev->flags |= MLX4_FLAG_OLD_REG_MAC;
198 }
188 if (err) 199 if (err)
189 return err; 200 return err;
190 201
@@ -231,10 +242,18 @@ void mlx4_unregister_mac(struct mlx4_dev *dev, u8 port, u64 mac)
231 u64 out_param = 0; 242 u64 out_param = 0;
232 243
233 if (mlx4_is_mfunc(dev)) { 244 if (mlx4_is_mfunc(dev)) {
234 set_param_l(&out_param, port); 245 if (!(dev->flags & MLX4_FLAG_OLD_REG_MAC)) {
235 (void) mlx4_cmd_imm(dev, mac, &out_param, RES_MAC, 246 (void) mlx4_cmd_imm(dev, mac, &out_param,
236 RES_OP_RESERVE_AND_MAP, MLX4_CMD_FREE_RES, 247 ((u32) port) << 8 | (u32) RES_MAC,
237 MLX4_CMD_TIME_CLASS_A, MLX4_CMD_WRAPPED); 248 RES_OP_RESERVE_AND_MAP, MLX4_CMD_FREE_RES,
249 MLX4_CMD_TIME_CLASS_A, MLX4_CMD_WRAPPED);
250 } else {
251 /* use old unregister mac format */
252 set_param_l(&out_param, port);
253 (void) mlx4_cmd_imm(dev, mac, &out_param, RES_MAC,
254 RES_OP_RESERVE_AND_MAP, MLX4_CMD_FREE_RES,
255 MLX4_CMD_TIME_CLASS_A, MLX4_CMD_WRAPPED);
256 }
238 return; 257 return;
239 } 258 }
240 __mlx4_unregister_mac(dev, port, mac); 259 __mlx4_unregister_mac(dev, port, mac);
@@ -284,7 +303,7 @@ static int mlx4_set_port_vlan_table(struct mlx4_dev *dev, u8 port,
284 memcpy(mailbox->buf, entries, MLX4_VLAN_TABLE_SIZE); 303 memcpy(mailbox->buf, entries, MLX4_VLAN_TABLE_SIZE);
285 in_mod = MLX4_SET_PORT_VLAN_TABLE << 8 | port; 304 in_mod = MLX4_SET_PORT_VLAN_TABLE << 8 | port;
286 err = mlx4_cmd(dev, mailbox->dma, in_mod, 1, MLX4_CMD_SET_PORT, 305 err = mlx4_cmd(dev, mailbox->dma, in_mod, 1, MLX4_CMD_SET_PORT,
287 MLX4_CMD_TIME_CLASS_B, MLX4_CMD_WRAPPED); 306 MLX4_CMD_TIME_CLASS_B, MLX4_CMD_NATIVE);
288 307
289 mlx4_free_cmd_mailbox(dev, mailbox); 308 mlx4_free_cmd_mailbox(dev, mailbox);
290 309
@@ -370,9 +389,12 @@ int mlx4_register_vlan(struct mlx4_dev *dev, u8 port, u16 vlan, int *index)
370 u64 out_param = 0; 389 u64 out_param = 0;
371 int err; 390 int err;
372 391
392 if (vlan > 4095)
393 return -EINVAL;
394
373 if (mlx4_is_mfunc(dev)) { 395 if (mlx4_is_mfunc(dev)) {
374 set_param_l(&out_param, port); 396 err = mlx4_cmd_imm(dev, vlan, &out_param,
375 err = mlx4_cmd_imm(dev, vlan, &out_param, RES_VLAN, 397 ((u32) port) << 8 | (u32) RES_VLAN,
376 RES_OP_RESERVE_AND_MAP, MLX4_CMD_ALLOC_RES, 398 RES_OP_RESERVE_AND_MAP, MLX4_CMD_ALLOC_RES,
377 MLX4_CMD_TIME_CLASS_A, MLX4_CMD_WRAPPED); 399 MLX4_CMD_TIME_CLASS_A, MLX4_CMD_WRAPPED);
378 if (!err) 400 if (!err)
@@ -384,23 +406,26 @@ int mlx4_register_vlan(struct mlx4_dev *dev, u8 port, u16 vlan, int *index)
384} 406}
385EXPORT_SYMBOL_GPL(mlx4_register_vlan); 407EXPORT_SYMBOL_GPL(mlx4_register_vlan);
386 408
387void __mlx4_unregister_vlan(struct mlx4_dev *dev, u8 port, int index) 409void __mlx4_unregister_vlan(struct mlx4_dev *dev, u8 port, u16 vlan)
388{ 410{
389 struct mlx4_vlan_table *table = &mlx4_priv(dev)->port[port].vlan_table; 411 struct mlx4_vlan_table *table = &mlx4_priv(dev)->port[port].vlan_table;
412 int index;
390 413
391 if (index < MLX4_VLAN_REGULAR) { 414 mutex_lock(&table->mutex);
392 mlx4_warn(dev, "Trying to free special vlan index %d\n", index); 415 if (mlx4_find_cached_vlan(dev, port, vlan, &index)) {
393 return; 416 mlx4_warn(dev, "vlan 0x%x is not in the vlan table\n", vlan);
417 goto out;
394 } 418 }
395 419
396 mutex_lock(&table->mutex); 420 if (index < MLX4_VLAN_REGULAR) {
397 if (!table->refs[index]) { 421 mlx4_warn(dev, "Trying to free special vlan index %d\n", index);
398 mlx4_warn(dev, "No vlan entry for index %d\n", index);
399 goto out; 422 goto out;
400 } 423 }
424
401 if (--table->refs[index]) { 425 if (--table->refs[index]) {
402 mlx4_dbg(dev, "Have more references for index %d," 426 mlx4_dbg(dev, "Have %d more references for index %d,"
403 "no need to modify vlan table\n", index); 427 "no need to modify vlan table\n", table->refs[index],
428 index);
404 goto out; 429 goto out;
405 } 430 }
406 table->entries[index] = 0; 431 table->entries[index] = 0;
@@ -410,23 +435,19 @@ out:
410 mutex_unlock(&table->mutex); 435 mutex_unlock(&table->mutex);
411} 436}
412 437
413void mlx4_unregister_vlan(struct mlx4_dev *dev, u8 port, int index) 438void mlx4_unregister_vlan(struct mlx4_dev *dev, u8 port, u16 vlan)
414{ 439{
415 u64 in_param = 0; 440 u64 out_param = 0;
416 int err;
417 441
418 if (mlx4_is_mfunc(dev)) { 442 if (mlx4_is_mfunc(dev)) {
419 set_param_l(&in_param, port); 443 (void) mlx4_cmd_imm(dev, vlan, &out_param,
420 err = mlx4_cmd(dev, in_param, RES_VLAN, RES_OP_RESERVE_AND_MAP, 444 ((u32) port) << 8 | (u32) RES_VLAN,
421 MLX4_CMD_FREE_RES, MLX4_CMD_TIME_CLASS_A, 445 RES_OP_RESERVE_AND_MAP,
422 MLX4_CMD_WRAPPED); 446 MLX4_CMD_FREE_RES, MLX4_CMD_TIME_CLASS_A,
423 if (!err) 447 MLX4_CMD_WRAPPED);
424 mlx4_warn(dev, "Failed freeing vlan at index:%d\n",
425 index);
426
427 return; 448 return;
428 } 449 }
429 __mlx4_unregister_vlan(dev, port, index); 450 __mlx4_unregister_vlan(dev, port, vlan);
430} 451}
431EXPORT_SYMBOL_GPL(mlx4_unregister_vlan); 452EXPORT_SYMBOL_GPL(mlx4_unregister_vlan);
432 453
@@ -448,8 +469,6 @@ int mlx4_get_port_ib_caps(struct mlx4_dev *dev, u8 port, __be32 *caps)
448 469
449 inbuf = inmailbox->buf; 470 inbuf = inmailbox->buf;
450 outbuf = outmailbox->buf; 471 outbuf = outmailbox->buf;
451 memset(inbuf, 0, 256);
452 memset(outbuf, 0, 256);
453 inbuf[0] = 1; 472 inbuf[0] = 1;
454 inbuf[1] = 1; 473 inbuf[1] = 1;
455 inbuf[2] = 1; 474 inbuf[2] = 1;
@@ -632,8 +651,6 @@ int mlx4_SET_PORT(struct mlx4_dev *dev, u8 port, int pkey_tbl_sz)
632 if (IS_ERR(mailbox)) 651 if (IS_ERR(mailbox))
633 return PTR_ERR(mailbox); 652 return PTR_ERR(mailbox);
634 653
635 memset(mailbox->buf, 0, 256);
636
637 ((__be32 *) mailbox->buf)[1] = dev->caps.ib_port_def_cap[port]; 654 ((__be32 *) mailbox->buf)[1] = dev->caps.ib_port_def_cap[port];
638 655
639 if (pkey_tbl_sz >= 0 && mlx4_is_master(dev)) { 656 if (pkey_tbl_sz >= 0 && mlx4_is_master(dev)) {
@@ -671,8 +688,6 @@ int mlx4_SET_PORT_general(struct mlx4_dev *dev, u8 port, int mtu,
671 if (IS_ERR(mailbox)) 688 if (IS_ERR(mailbox))
672 return PTR_ERR(mailbox); 689 return PTR_ERR(mailbox);
673 context = mailbox->buf; 690 context = mailbox->buf;
674 memset(context, 0, sizeof *context);
675
676 context->flags = SET_PORT_GEN_ALL_VALID; 691 context->flags = SET_PORT_GEN_ALL_VALID;
677 context->mtu = cpu_to_be16(mtu); 692 context->mtu = cpu_to_be16(mtu);
678 context->pptx = (pptx * (!pfctx)) << 7; 693 context->pptx = (pptx * (!pfctx)) << 7;
@@ -706,8 +721,6 @@ int mlx4_SET_PORT_qpn_calc(struct mlx4_dev *dev, u8 port, u32 base_qpn,
706 if (IS_ERR(mailbox)) 721 if (IS_ERR(mailbox))
707 return PTR_ERR(mailbox); 722 return PTR_ERR(mailbox);
708 context = mailbox->buf; 723 context = mailbox->buf;
709 memset(context, 0, sizeof *context);
710
711 context->base_qpn = cpu_to_be32(base_qpn); 724 context->base_qpn = cpu_to_be32(base_qpn);
712 context->n_mac = dev->caps.log_num_macs; 725 context->n_mac = dev->caps.log_num_macs;
713 context->promisc = cpu_to_be32(promisc << SET_PORT_PROMISC_SHIFT | 726 context->promisc = cpu_to_be32(promisc << SET_PORT_PROMISC_SHIFT |
@@ -740,8 +753,6 @@ int mlx4_SET_PORT_PRIO2TC(struct mlx4_dev *dev, u8 port, u8 *prio2tc)
740 if (IS_ERR(mailbox)) 753 if (IS_ERR(mailbox))
741 return PTR_ERR(mailbox); 754 return PTR_ERR(mailbox);
742 context = mailbox->buf; 755 context = mailbox->buf;
743 memset(context, 0, sizeof *context);
744
745 for (i = 0; i < MLX4_NUM_UP; i += 2) 756 for (i = 0; i < MLX4_NUM_UP; i += 2)
746 context->prio2tc[i >> 1] = prio2tc[i] << 4 | prio2tc[i + 1]; 757 context->prio2tc[i >> 1] = prio2tc[i] << 4 | prio2tc[i + 1];
747 758
@@ -767,7 +778,6 @@ int mlx4_SET_PORT_SCHEDULER(struct mlx4_dev *dev, u8 port, u8 *tc_tx_bw,
767 if (IS_ERR(mailbox)) 778 if (IS_ERR(mailbox))
768 return PTR_ERR(mailbox); 779 return PTR_ERR(mailbox);
769 context = mailbox->buf; 780 context = mailbox->buf;
770 memset(context, 0, sizeof *context);
771 781
772 for (i = 0; i < MLX4_NUM_TC; i++) { 782 for (i = 0; i < MLX4_NUM_TC; i++) {
773 struct mlx4_port_scheduler_tc_cfg_be *tc = &context->tc[i]; 783 struct mlx4_port_scheduler_tc_cfg_be *tc = &context->tc[i];
diff --git a/drivers/net/ethernet/mellanox/mlx4/qp.c b/drivers/net/ethernet/mellanox/mlx4/qp.c
index e891b058c1be..2715e61dbb74 100644
--- a/drivers/net/ethernet/mellanox/mlx4/qp.c
+++ b/drivers/net/ethernet/mellanox/mlx4/qp.c
@@ -480,8 +480,7 @@ int mlx4_init_qp_table(struct mlx4_dev *dev)
480 */ 480 */
481 481
482 err = mlx4_bitmap_init(&qp_table->bitmap, dev->caps.num_qps, 482 err = mlx4_bitmap_init(&qp_table->bitmap, dev->caps.num_qps,
483 (1 << 23) - 1, dev->phys_caps.base_sqpn + 8 + 483 (1 << 23) - 1, mlx4_num_reserved_sqps(dev),
484 16 * MLX4_MFUNC_MAX * !!mlx4_is_master(dev),
485 reserved_from_top); 484 reserved_from_top);
486 if (err) 485 if (err)
487 return err; 486 return err;
diff --git a/drivers/net/ethernet/mellanox/mlx4/resource_tracker.c b/drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
index dd6876321116..2f3f2bc7f283 100644
--- a/drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
+++ b/drivers/net/ethernet/mellanox/mlx4/resource_tracker.c
@@ -55,6 +55,14 @@ struct mac_res {
55 u8 port; 55 u8 port;
56}; 56};
57 57
58struct vlan_res {
59 struct list_head list;
60 u16 vlan;
61 int ref_count;
62 int vlan_index;
63 u8 port;
64};
65
58struct res_common { 66struct res_common {
59 struct list_head list; 67 struct list_head list;
60 struct rb_node node; 68 struct rb_node node;
@@ -102,7 +110,14 @@ struct res_qp {
102 int local_qpn; 110 int local_qpn;
103 atomic_t ref_count; 111 atomic_t ref_count;
104 u32 qpc_flags; 112 u32 qpc_flags;
113 /* saved qp params before VST enforcement in order to restore on VGT */
105 u8 sched_queue; 114 u8 sched_queue;
115 __be32 param3;
116 u8 vlan_control;
117 u8 fvl_rx;
118 u8 pri_path_fl;
119 u8 vlan_index;
120 u8 feup;
106}; 121};
107 122
108enum res_mtt_states { 123enum res_mtt_states {
@@ -266,6 +281,7 @@ static const char *ResourceType(enum mlx4_resource rt)
266 case RES_MPT: return "RES_MPT"; 281 case RES_MPT: return "RES_MPT";
267 case RES_MTT: return "RES_MTT"; 282 case RES_MTT: return "RES_MTT";
268 case RES_MAC: return "RES_MAC"; 283 case RES_MAC: return "RES_MAC";
284 case RES_VLAN: return "RES_VLAN";
269 case RES_EQ: return "RES_EQ"; 285 case RES_EQ: return "RES_EQ";
270 case RES_COUNTER: return "RES_COUNTER"; 286 case RES_COUNTER: return "RES_COUNTER";
271 case RES_FS_RULE: return "RES_FS_RULE"; 287 case RES_FS_RULE: return "RES_FS_RULE";
@@ -274,10 +290,139 @@ static const char *ResourceType(enum mlx4_resource rt)
274 }; 290 };
275} 291}
276 292
293static void rem_slave_vlans(struct mlx4_dev *dev, int slave);
294static inline int mlx4_grant_resource(struct mlx4_dev *dev, int slave,
295 enum mlx4_resource res_type, int count,
296 int port)
297{
298 struct mlx4_priv *priv = mlx4_priv(dev);
299 struct resource_allocator *res_alloc =
300 &priv->mfunc.master.res_tracker.res_alloc[res_type];
301 int err = -EINVAL;
302 int allocated, free, reserved, guaranteed, from_free;
303
304 if (slave > dev->num_vfs)
305 return -EINVAL;
306
307 spin_lock(&res_alloc->alloc_lock);
308 allocated = (port > 0) ?
309 res_alloc->allocated[(port - 1) * (dev->num_vfs + 1) + slave] :
310 res_alloc->allocated[slave];
311 free = (port > 0) ? res_alloc->res_port_free[port - 1] :
312 res_alloc->res_free;
313 reserved = (port > 0) ? res_alloc->res_port_rsvd[port - 1] :
314 res_alloc->res_reserved;
315 guaranteed = res_alloc->guaranteed[slave];
316
317 if (allocated + count > res_alloc->quota[slave])
318 goto out;
319
320 if (allocated + count <= guaranteed) {
321 err = 0;
322 } else {
323 /* portion may need to be obtained from free area */
324 if (guaranteed - allocated > 0)
325 from_free = count - (guaranteed - allocated);
326 else
327 from_free = count;
328
329 if (free - from_free > reserved)
330 err = 0;
331 }
332
333 if (!err) {
334 /* grant the request */
335 if (port > 0) {
336 res_alloc->allocated[(port - 1) * (dev->num_vfs + 1) + slave] += count;
337 res_alloc->res_port_free[port - 1] -= count;
338 } else {
339 res_alloc->allocated[slave] += count;
340 res_alloc->res_free -= count;
341 }
342 }
343
344out:
345 spin_unlock(&res_alloc->alloc_lock);
346 return err;
347}
348
349static inline void mlx4_release_resource(struct mlx4_dev *dev, int slave,
350 enum mlx4_resource res_type, int count,
351 int port)
352{
353 struct mlx4_priv *priv = mlx4_priv(dev);
354 struct resource_allocator *res_alloc =
355 &priv->mfunc.master.res_tracker.res_alloc[res_type];
356
357 if (slave > dev->num_vfs)
358 return;
359
360 spin_lock(&res_alloc->alloc_lock);
361 if (port > 0) {
362 res_alloc->allocated[(port - 1) * (dev->num_vfs + 1) + slave] -= count;
363 res_alloc->res_port_free[port - 1] += count;
364 } else {
365 res_alloc->allocated[slave] -= count;
366 res_alloc->res_free += count;
367 }
368
369 spin_unlock(&res_alloc->alloc_lock);
370 return;
371}
372
373static inline void initialize_res_quotas(struct mlx4_dev *dev,
374 struct resource_allocator *res_alloc,
375 enum mlx4_resource res_type,
376 int vf, int num_instances)
377{
378 res_alloc->guaranteed[vf] = num_instances / (2 * (dev->num_vfs + 1));
379 res_alloc->quota[vf] = (num_instances / 2) + res_alloc->guaranteed[vf];
380 if (vf == mlx4_master_func_num(dev)) {
381 res_alloc->res_free = num_instances;
382 if (res_type == RES_MTT) {
383 /* reserved mtts will be taken out of the PF allocation */
384 res_alloc->res_free += dev->caps.reserved_mtts;
385 res_alloc->guaranteed[vf] += dev->caps.reserved_mtts;
386 res_alloc->quota[vf] += dev->caps.reserved_mtts;
387 }
388 }
389}
390
391void mlx4_init_quotas(struct mlx4_dev *dev)
392{
393 struct mlx4_priv *priv = mlx4_priv(dev);
394 int pf;
395
396 /* quotas for VFs are initialized in mlx4_slave_cap */
397 if (mlx4_is_slave(dev))
398 return;
399
400 if (!mlx4_is_mfunc(dev)) {
401 dev->quotas.qp = dev->caps.num_qps - dev->caps.reserved_qps -
402 mlx4_num_reserved_sqps(dev);
403 dev->quotas.cq = dev->caps.num_cqs - dev->caps.reserved_cqs;
404 dev->quotas.srq = dev->caps.num_srqs - dev->caps.reserved_srqs;
405 dev->quotas.mtt = dev->caps.num_mtts - dev->caps.reserved_mtts;
406 dev->quotas.mpt = dev->caps.num_mpts - dev->caps.reserved_mrws;
407 return;
408 }
409
410 pf = mlx4_master_func_num(dev);
411 dev->quotas.qp =
412 priv->mfunc.master.res_tracker.res_alloc[RES_QP].quota[pf];
413 dev->quotas.cq =
414 priv->mfunc.master.res_tracker.res_alloc[RES_CQ].quota[pf];
415 dev->quotas.srq =
416 priv->mfunc.master.res_tracker.res_alloc[RES_SRQ].quota[pf];
417 dev->quotas.mtt =
418 priv->mfunc.master.res_tracker.res_alloc[RES_MTT].quota[pf];
419 dev->quotas.mpt =
420 priv->mfunc.master.res_tracker.res_alloc[RES_MPT].quota[pf];
421}
277int mlx4_init_resource_tracker(struct mlx4_dev *dev) 422int mlx4_init_resource_tracker(struct mlx4_dev *dev)
278{ 423{
279 struct mlx4_priv *priv = mlx4_priv(dev); 424 struct mlx4_priv *priv = mlx4_priv(dev);
280 int i; 425 int i, j;
281 int t; 426 int t;
282 427
283 priv->mfunc.master.res_tracker.slave_list = 428 priv->mfunc.master.res_tracker.slave_list =
@@ -298,8 +443,105 @@ int mlx4_init_resource_tracker(struct mlx4_dev *dev)
298 for (i = 0 ; i < MLX4_NUM_OF_RESOURCE_TYPE; i++) 443 for (i = 0 ; i < MLX4_NUM_OF_RESOURCE_TYPE; i++)
299 priv->mfunc.master.res_tracker.res_tree[i] = RB_ROOT; 444 priv->mfunc.master.res_tracker.res_tree[i] = RB_ROOT;
300 445
446 for (i = 0; i < MLX4_NUM_OF_RESOURCE_TYPE; i++) {
447 struct resource_allocator *res_alloc =
448 &priv->mfunc.master.res_tracker.res_alloc[i];
449 res_alloc->quota = kmalloc((dev->num_vfs + 1) * sizeof(int), GFP_KERNEL);
450 res_alloc->guaranteed = kmalloc((dev->num_vfs + 1) * sizeof(int), GFP_KERNEL);
451 if (i == RES_MAC || i == RES_VLAN)
452 res_alloc->allocated = kzalloc(MLX4_MAX_PORTS *
453 (dev->num_vfs + 1) * sizeof(int),
454 GFP_KERNEL);
455 else
456 res_alloc->allocated = kzalloc((dev->num_vfs + 1) * sizeof(int), GFP_KERNEL);
457
458 if (!res_alloc->quota || !res_alloc->guaranteed ||
459 !res_alloc->allocated)
460 goto no_mem_err;
461
462 spin_lock_init(&res_alloc->alloc_lock);
463 for (t = 0; t < dev->num_vfs + 1; t++) {
464 switch (i) {
465 case RES_QP:
466 initialize_res_quotas(dev, res_alloc, RES_QP,
467 t, dev->caps.num_qps -
468 dev->caps.reserved_qps -
469 mlx4_num_reserved_sqps(dev));
470 break;
471 case RES_CQ:
472 initialize_res_quotas(dev, res_alloc, RES_CQ,
473 t, dev->caps.num_cqs -
474 dev->caps.reserved_cqs);
475 break;
476 case RES_SRQ:
477 initialize_res_quotas(dev, res_alloc, RES_SRQ,
478 t, dev->caps.num_srqs -
479 dev->caps.reserved_srqs);
480 break;
481 case RES_MPT:
482 initialize_res_quotas(dev, res_alloc, RES_MPT,
483 t, dev->caps.num_mpts -
484 dev->caps.reserved_mrws);
485 break;
486 case RES_MTT:
487 initialize_res_quotas(dev, res_alloc, RES_MTT,
488 t, dev->caps.num_mtts -
489 dev->caps.reserved_mtts);
490 break;
491 case RES_MAC:
492 if (t == mlx4_master_func_num(dev)) {
493 res_alloc->quota[t] = MLX4_MAX_MAC_NUM;
494 res_alloc->guaranteed[t] = 2;
495 for (j = 0; j < MLX4_MAX_PORTS; j++)
496 res_alloc->res_port_free[j] = MLX4_MAX_MAC_NUM;
497 } else {
498 res_alloc->quota[t] = MLX4_MAX_MAC_NUM;
499 res_alloc->guaranteed[t] = 2;
500 }
501 break;
502 case RES_VLAN:
503 if (t == mlx4_master_func_num(dev)) {
504 res_alloc->quota[t] = MLX4_MAX_VLAN_NUM;
505 res_alloc->guaranteed[t] = MLX4_MAX_VLAN_NUM / 2;
506 for (j = 0; j < MLX4_MAX_PORTS; j++)
507 res_alloc->res_port_free[j] =
508 res_alloc->quota[t];
509 } else {
510 res_alloc->quota[t] = MLX4_MAX_VLAN_NUM / 2;
511 res_alloc->guaranteed[t] = 0;
512 }
513 break;
514 case RES_COUNTER:
515 res_alloc->quota[t] = dev->caps.max_counters;
516 res_alloc->guaranteed[t] = 0;
517 if (t == mlx4_master_func_num(dev))
518 res_alloc->res_free = res_alloc->quota[t];
519 break;
520 default:
521 break;
522 }
523 if (i == RES_MAC || i == RES_VLAN) {
524 for (j = 0; j < MLX4_MAX_PORTS; j++)
525 res_alloc->res_port_rsvd[j] +=
526 res_alloc->guaranteed[t];
527 } else {
528 res_alloc->res_reserved += res_alloc->guaranteed[t];
529 }
530 }
531 }
301 spin_lock_init(&priv->mfunc.master.res_tracker.lock); 532 spin_lock_init(&priv->mfunc.master.res_tracker.lock);
302 return 0 ; 533 return 0;
534
535no_mem_err:
536 for (i = 0; i < MLX4_NUM_OF_RESOURCE_TYPE; i++) {
537 kfree(priv->mfunc.master.res_tracker.res_alloc[i].allocated);
538 priv->mfunc.master.res_tracker.res_alloc[i].allocated = NULL;
539 kfree(priv->mfunc.master.res_tracker.res_alloc[i].guaranteed);
540 priv->mfunc.master.res_tracker.res_alloc[i].guaranteed = NULL;
541 kfree(priv->mfunc.master.res_tracker.res_alloc[i].quota);
542 priv->mfunc.master.res_tracker.res_alloc[i].quota = NULL;
543 }
544 return -ENOMEM;
303} 545}
304 546
305void mlx4_free_resource_tracker(struct mlx4_dev *dev, 547void mlx4_free_resource_tracker(struct mlx4_dev *dev,
@@ -309,13 +551,28 @@ void mlx4_free_resource_tracker(struct mlx4_dev *dev,
309 int i; 551 int i;
310 552
311 if (priv->mfunc.master.res_tracker.slave_list) { 553 if (priv->mfunc.master.res_tracker.slave_list) {
312 if (type != RES_TR_FREE_STRUCTS_ONLY) 554 if (type != RES_TR_FREE_STRUCTS_ONLY) {
313 for (i = 0 ; i < dev->num_slaves; i++) 555 for (i = 0; i < dev->num_slaves; i++) {
314 if (type == RES_TR_FREE_ALL || 556 if (type == RES_TR_FREE_ALL ||
315 dev->caps.function != i) 557 dev->caps.function != i)
316 mlx4_delete_all_resources_for_slave(dev, i); 558 mlx4_delete_all_resources_for_slave(dev, i);
559 }
560 /* free master's vlans */
561 i = dev->caps.function;
562 mutex_lock(&priv->mfunc.master.res_tracker.slave_list[i].mutex);
563 rem_slave_vlans(dev, i);
564 mutex_unlock(&priv->mfunc.master.res_tracker.slave_list[i].mutex);
565 }
317 566
318 if (type != RES_TR_FREE_SLAVES_ONLY) { 567 if (type != RES_TR_FREE_SLAVES_ONLY) {
568 for (i = 0; i < MLX4_NUM_OF_RESOURCE_TYPE; i++) {
569 kfree(priv->mfunc.master.res_tracker.res_alloc[i].allocated);
570 priv->mfunc.master.res_tracker.res_alloc[i].allocated = NULL;
571 kfree(priv->mfunc.master.res_tracker.res_alloc[i].guaranteed);
572 priv->mfunc.master.res_tracker.res_alloc[i].guaranteed = NULL;
573 kfree(priv->mfunc.master.res_tracker.res_alloc[i].quota);
574 priv->mfunc.master.res_tracker.res_alloc[i].quota = NULL;
575 }
319 kfree(priv->mfunc.master.res_tracker.slave_list); 576 kfree(priv->mfunc.master.res_tracker.slave_list);
320 priv->mfunc.master.res_tracker.slave_list = NULL; 577 priv->mfunc.master.res_tracker.slave_list = NULL;
321 } 578 }
@@ -1229,12 +1486,19 @@ static int qp_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1229 case RES_OP_RESERVE: 1486 case RES_OP_RESERVE:
1230 count = get_param_l(&in_param); 1487 count = get_param_l(&in_param);
1231 align = get_param_h(&in_param); 1488 align = get_param_h(&in_param);
1232 err = __mlx4_qp_reserve_range(dev, count, align, &base); 1489 err = mlx4_grant_resource(dev, slave, RES_QP, count, 0);
1233 if (err) 1490 if (err)
1234 return err; 1491 return err;
1235 1492
1493 err = __mlx4_qp_reserve_range(dev, count, align, &base);
1494 if (err) {
1495 mlx4_release_resource(dev, slave, RES_QP, count, 0);
1496 return err;
1497 }
1498
1236 err = add_res_range(dev, slave, base, count, RES_QP, 0); 1499 err = add_res_range(dev, slave, base, count, RES_QP, 0);
1237 if (err) { 1500 if (err) {
1501 mlx4_release_resource(dev, slave, RES_QP, count, 0);
1238 __mlx4_qp_release_range(dev, base, count); 1502 __mlx4_qp_release_range(dev, base, count);
1239 return err; 1503 return err;
1240 } 1504 }
@@ -1282,15 +1546,24 @@ static int mtt_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1282 return err; 1546 return err;
1283 1547
1284 order = get_param_l(&in_param); 1548 order = get_param_l(&in_param);
1549
1550 err = mlx4_grant_resource(dev, slave, RES_MTT, 1 << order, 0);
1551 if (err)
1552 return err;
1553
1285 base = __mlx4_alloc_mtt_range(dev, order); 1554 base = __mlx4_alloc_mtt_range(dev, order);
1286 if (base == -1) 1555 if (base == -1) {
1556 mlx4_release_resource(dev, slave, RES_MTT, 1 << order, 0);
1287 return -ENOMEM; 1557 return -ENOMEM;
1558 }
1288 1559
1289 err = add_res_range(dev, slave, base, 1, RES_MTT, order); 1560 err = add_res_range(dev, slave, base, 1, RES_MTT, order);
1290 if (err) 1561 if (err) {
1562 mlx4_release_resource(dev, slave, RES_MTT, 1 << order, 0);
1291 __mlx4_free_mtt_range(dev, base, order); 1563 __mlx4_free_mtt_range(dev, base, order);
1292 else 1564 } else {
1293 set_param_l(out_param, base); 1565 set_param_l(out_param, base);
1566 }
1294 1567
1295 return err; 1568 return err;
1296} 1569}
@@ -1305,13 +1578,20 @@ static int mpt_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1305 1578
1306 switch (op) { 1579 switch (op) {
1307 case RES_OP_RESERVE: 1580 case RES_OP_RESERVE:
1581 err = mlx4_grant_resource(dev, slave, RES_MPT, 1, 0);
1582 if (err)
1583 break;
1584
1308 index = __mlx4_mpt_reserve(dev); 1585 index = __mlx4_mpt_reserve(dev);
1309 if (index == -1) 1586 if (index == -1) {
1587 mlx4_release_resource(dev, slave, RES_MPT, 1, 0);
1310 break; 1588 break;
1589 }
1311 id = index & mpt_mask(dev); 1590 id = index & mpt_mask(dev);
1312 1591
1313 err = add_res_range(dev, slave, id, 1, RES_MPT, index); 1592 err = add_res_range(dev, slave, id, 1, RES_MPT, index);
1314 if (err) { 1593 if (err) {
1594 mlx4_release_resource(dev, slave, RES_MPT, 1, 0);
1315 __mlx4_mpt_release(dev, index); 1595 __mlx4_mpt_release(dev, index);
1316 break; 1596 break;
1317 } 1597 }
@@ -1345,12 +1625,19 @@ static int cq_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1345 1625
1346 switch (op) { 1626 switch (op) {
1347 case RES_OP_RESERVE_AND_MAP: 1627 case RES_OP_RESERVE_AND_MAP:
1348 err = __mlx4_cq_alloc_icm(dev, &cqn); 1628 err = mlx4_grant_resource(dev, slave, RES_CQ, 1, 0);
1349 if (err) 1629 if (err)
1350 break; 1630 break;
1351 1631
1632 err = __mlx4_cq_alloc_icm(dev, &cqn);
1633 if (err) {
1634 mlx4_release_resource(dev, slave, RES_CQ, 1, 0);
1635 break;
1636 }
1637
1352 err = add_res_range(dev, slave, cqn, 1, RES_CQ, 0); 1638 err = add_res_range(dev, slave, cqn, 1, RES_CQ, 0);
1353 if (err) { 1639 if (err) {
1640 mlx4_release_resource(dev, slave, RES_CQ, 1, 0);
1354 __mlx4_cq_free_icm(dev, cqn); 1641 __mlx4_cq_free_icm(dev, cqn);
1355 break; 1642 break;
1356 } 1643 }
@@ -1373,12 +1660,19 @@ static int srq_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1373 1660
1374 switch (op) { 1661 switch (op) {
1375 case RES_OP_RESERVE_AND_MAP: 1662 case RES_OP_RESERVE_AND_MAP:
1376 err = __mlx4_srq_alloc_icm(dev, &srqn); 1663 err = mlx4_grant_resource(dev, slave, RES_SRQ, 1, 0);
1377 if (err) 1664 if (err)
1378 break; 1665 break;
1379 1666
1667 err = __mlx4_srq_alloc_icm(dev, &srqn);
1668 if (err) {
1669 mlx4_release_resource(dev, slave, RES_SRQ, 1, 0);
1670 break;
1671 }
1672
1380 err = add_res_range(dev, slave, srqn, 1, RES_SRQ, 0); 1673 err = add_res_range(dev, slave, srqn, 1, RES_SRQ, 0);
1381 if (err) { 1674 if (err) {
1675 mlx4_release_resource(dev, slave, RES_SRQ, 1, 0);
1382 __mlx4_srq_free_icm(dev, srqn); 1676 __mlx4_srq_free_icm(dev, srqn);
1383 break; 1677 break;
1384 } 1678 }
@@ -1399,9 +1693,13 @@ static int mac_add_to_slave(struct mlx4_dev *dev, int slave, u64 mac, int port)
1399 struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker; 1693 struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
1400 struct mac_res *res; 1694 struct mac_res *res;
1401 1695
1696 if (mlx4_grant_resource(dev, slave, RES_MAC, 1, port))
1697 return -EINVAL;
1402 res = kzalloc(sizeof *res, GFP_KERNEL); 1698 res = kzalloc(sizeof *res, GFP_KERNEL);
1403 if (!res) 1699 if (!res) {
1700 mlx4_release_resource(dev, slave, RES_MAC, 1, port);
1404 return -ENOMEM; 1701 return -ENOMEM;
1702 }
1405 res->mac = mac; 1703 res->mac = mac;
1406 res->port = (u8) port; 1704 res->port = (u8) port;
1407 list_add_tail(&res->list, 1705 list_add_tail(&res->list,
@@ -1421,6 +1719,7 @@ static void mac_del_from_slave(struct mlx4_dev *dev, int slave, u64 mac,
1421 list_for_each_entry_safe(res, tmp, mac_list, list) { 1719 list_for_each_entry_safe(res, tmp, mac_list, list) {
1422 if (res->mac == mac && res->port == (u8) port) { 1720 if (res->mac == mac && res->port == (u8) port) {
1423 list_del(&res->list); 1721 list_del(&res->list);
1722 mlx4_release_resource(dev, slave, RES_MAC, 1, port);
1424 kfree(res); 1723 kfree(res);
1425 break; 1724 break;
1426 } 1725 }
@@ -1438,12 +1737,13 @@ static void rem_slave_macs(struct mlx4_dev *dev, int slave)
1438 list_for_each_entry_safe(res, tmp, mac_list, list) { 1737 list_for_each_entry_safe(res, tmp, mac_list, list) {
1439 list_del(&res->list); 1738 list_del(&res->list);
1440 __mlx4_unregister_mac(dev, res->port, res->mac); 1739 __mlx4_unregister_mac(dev, res->port, res->mac);
1740 mlx4_release_resource(dev, slave, RES_MAC, 1, res->port);
1441 kfree(res); 1741 kfree(res);
1442 } 1742 }
1443} 1743}
1444 1744
1445static int mac_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd, 1745static int mac_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1446 u64 in_param, u64 *out_param) 1746 u64 in_param, u64 *out_param, int in_port)
1447{ 1747{
1448 int err = -EINVAL; 1748 int err = -EINVAL;
1449 int port; 1749 int port;
@@ -1452,7 +1752,7 @@ static int mac_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1452 if (op != RES_OP_RESERVE_AND_MAP) 1752 if (op != RES_OP_RESERVE_AND_MAP)
1453 return err; 1753 return err;
1454 1754
1455 port = get_param_l(out_param); 1755 port = !in_port ? get_param_l(out_param) : in_port;
1456 mac = in_param; 1756 mac = in_param;
1457 1757
1458 err = __mlx4_register_mac(dev, port, mac); 1758 err = __mlx4_register_mac(dev, port, mac);
@@ -1469,12 +1769,114 @@ static int mac_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1469 return err; 1769 return err;
1470} 1770}
1471 1771
1472static int vlan_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd, 1772static int vlan_add_to_slave(struct mlx4_dev *dev, int slave, u16 vlan,
1473 u64 in_param, u64 *out_param) 1773 int port, int vlan_index)
1474{ 1774{
1775 struct mlx4_priv *priv = mlx4_priv(dev);
1776 struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
1777 struct list_head *vlan_list =
1778 &tracker->slave_list[slave].res_list[RES_VLAN];
1779 struct vlan_res *res, *tmp;
1780
1781 list_for_each_entry_safe(res, tmp, vlan_list, list) {
1782 if (res->vlan == vlan && res->port == (u8) port) {
1783 /* vlan found. update ref count */
1784 ++res->ref_count;
1785 return 0;
1786 }
1787 }
1788
1789 if (mlx4_grant_resource(dev, slave, RES_VLAN, 1, port))
1790 return -EINVAL;
1791 res = kzalloc(sizeof(*res), GFP_KERNEL);
1792 if (!res) {
1793 mlx4_release_resource(dev, slave, RES_VLAN, 1, port);
1794 return -ENOMEM;
1795 }
1796 res->vlan = vlan;
1797 res->port = (u8) port;
1798 res->vlan_index = vlan_index;
1799 res->ref_count = 1;
1800 list_add_tail(&res->list,
1801 &tracker->slave_list[slave].res_list[RES_VLAN]);
1475 return 0; 1802 return 0;
1476} 1803}
1477 1804
1805
1806static void vlan_del_from_slave(struct mlx4_dev *dev, int slave, u16 vlan,
1807 int port)
1808{
1809 struct mlx4_priv *priv = mlx4_priv(dev);
1810 struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
1811 struct list_head *vlan_list =
1812 &tracker->slave_list[slave].res_list[RES_VLAN];
1813 struct vlan_res *res, *tmp;
1814
1815 list_for_each_entry_safe(res, tmp, vlan_list, list) {
1816 if (res->vlan == vlan && res->port == (u8) port) {
1817 if (!--res->ref_count) {
1818 list_del(&res->list);
1819 mlx4_release_resource(dev, slave, RES_VLAN,
1820 1, port);
1821 kfree(res);
1822 }
1823 break;
1824 }
1825 }
1826}
1827
1828static void rem_slave_vlans(struct mlx4_dev *dev, int slave)
1829{
1830 struct mlx4_priv *priv = mlx4_priv(dev);
1831 struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
1832 struct list_head *vlan_list =
1833 &tracker->slave_list[slave].res_list[RES_VLAN];
1834 struct vlan_res *res, *tmp;
1835 int i;
1836
1837 list_for_each_entry_safe(res, tmp, vlan_list, list) {
1838 list_del(&res->list);
1839 /* dereference the vlan the num times the slave referenced it */
1840 for (i = 0; i < res->ref_count; i++)
1841 __mlx4_unregister_vlan(dev, res->port, res->vlan);
1842 mlx4_release_resource(dev, slave, RES_VLAN, 1, res->port);
1843 kfree(res);
1844 }
1845}
1846
1847static int vlan_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1848 u64 in_param, u64 *out_param, int in_port)
1849{
1850 struct mlx4_priv *priv = mlx4_priv(dev);
1851 struct mlx4_slave_state *slave_state = priv->mfunc.master.slave_state;
1852 int err;
1853 u16 vlan;
1854 int vlan_index;
1855 int port;
1856
1857 port = !in_port ? get_param_l(out_param) : in_port;
1858
1859 if (!port || op != RES_OP_RESERVE_AND_MAP)
1860 return -EINVAL;
1861
1862 /* upstream kernels had NOP for reg/unreg vlan. Continue this. */
1863 if (!in_port && port > 0 && port <= dev->caps.num_ports) {
1864 slave_state[slave].old_vlan_api = true;
1865 return 0;
1866 }
1867
1868 vlan = (u16) in_param;
1869
1870 err = __mlx4_register_vlan(dev, port, vlan, &vlan_index);
1871 if (!err) {
1872 set_param_l(out_param, (u32) vlan_index);
1873 err = vlan_add_to_slave(dev, slave, vlan, port, vlan_index);
1874 if (err)
1875 __mlx4_unregister_vlan(dev, port, vlan);
1876 }
1877 return err;
1878}
1879
1478static int counter_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd, 1880static int counter_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1479 u64 in_param, u64 *out_param) 1881 u64 in_param, u64 *out_param)
1480{ 1882{
@@ -1484,15 +1886,23 @@ static int counter_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1484 if (op != RES_OP_RESERVE) 1886 if (op != RES_OP_RESERVE)
1485 return -EINVAL; 1887 return -EINVAL;
1486 1888
1487 err = __mlx4_counter_alloc(dev, &index); 1889 err = mlx4_grant_resource(dev, slave, RES_COUNTER, 1, 0);
1488 if (err) 1890 if (err)
1489 return err; 1891 return err;
1490 1892
1893 err = __mlx4_counter_alloc(dev, &index);
1894 if (err) {
1895 mlx4_release_resource(dev, slave, RES_COUNTER, 1, 0);
1896 return err;
1897 }
1898
1491 err = add_res_range(dev, slave, index, 1, RES_COUNTER, 0); 1899 err = add_res_range(dev, slave, index, 1, RES_COUNTER, 0);
1492 if (err) 1900 if (err) {
1493 __mlx4_counter_free(dev, index); 1901 __mlx4_counter_free(dev, index);
1494 else 1902 mlx4_release_resource(dev, slave, RES_COUNTER, 1, 0);
1903 } else {
1495 set_param_l(out_param, index); 1904 set_param_l(out_param, index);
1905 }
1496 1906
1497 return err; 1907 return err;
1498} 1908}
@@ -1528,7 +1938,7 @@ int mlx4_ALLOC_RES_wrapper(struct mlx4_dev *dev, int slave,
1528 int err; 1938 int err;
1529 int alop = vhcr->op_modifier; 1939 int alop = vhcr->op_modifier;
1530 1940
1531 switch (vhcr->in_modifier) { 1941 switch (vhcr->in_modifier & 0xFF) {
1532 case RES_QP: 1942 case RES_QP:
1533 err = qp_alloc_res(dev, slave, vhcr->op_modifier, alop, 1943 err = qp_alloc_res(dev, slave, vhcr->op_modifier, alop,
1534 vhcr->in_param, &vhcr->out_param); 1944 vhcr->in_param, &vhcr->out_param);
@@ -1556,12 +1966,14 @@ int mlx4_ALLOC_RES_wrapper(struct mlx4_dev *dev, int slave,
1556 1966
1557 case RES_MAC: 1967 case RES_MAC:
1558 err = mac_alloc_res(dev, slave, vhcr->op_modifier, alop, 1968 err = mac_alloc_res(dev, slave, vhcr->op_modifier, alop,
1559 vhcr->in_param, &vhcr->out_param); 1969 vhcr->in_param, &vhcr->out_param,
1970 (vhcr->in_modifier >> 8) & 0xFF);
1560 break; 1971 break;
1561 1972
1562 case RES_VLAN: 1973 case RES_VLAN:
1563 err = vlan_alloc_res(dev, slave, vhcr->op_modifier, alop, 1974 err = vlan_alloc_res(dev, slave, vhcr->op_modifier, alop,
1564 vhcr->in_param, &vhcr->out_param); 1975 vhcr->in_param, &vhcr->out_param,
1976 (vhcr->in_modifier >> 8) & 0xFF);
1565 break; 1977 break;
1566 1978
1567 case RES_COUNTER: 1979 case RES_COUNTER:
@@ -1597,6 +2009,7 @@ static int qp_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1597 err = rem_res_range(dev, slave, base, count, RES_QP, 0); 2009 err = rem_res_range(dev, slave, base, count, RES_QP, 0);
1598 if (err) 2010 if (err)
1599 break; 2011 break;
2012 mlx4_release_resource(dev, slave, RES_QP, count, 0);
1600 __mlx4_qp_release_range(dev, base, count); 2013 __mlx4_qp_release_range(dev, base, count);
1601 break; 2014 break;
1602 case RES_OP_MAP_ICM: 2015 case RES_OP_MAP_ICM:
@@ -1634,8 +2047,10 @@ static int mtt_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1634 base = get_param_l(&in_param); 2047 base = get_param_l(&in_param);
1635 order = get_param_h(&in_param); 2048 order = get_param_h(&in_param);
1636 err = rem_res_range(dev, slave, base, 1, RES_MTT, order); 2049 err = rem_res_range(dev, slave, base, 1, RES_MTT, order);
1637 if (!err) 2050 if (!err) {
2051 mlx4_release_resource(dev, slave, RES_MTT, 1 << order, 0);
1638 __mlx4_free_mtt_range(dev, base, order); 2052 __mlx4_free_mtt_range(dev, base, order);
2053 }
1639 return err; 2054 return err;
1640} 2055}
1641 2056
@@ -1660,6 +2075,7 @@ static int mpt_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1660 err = rem_res_range(dev, slave, id, 1, RES_MPT, 0); 2075 err = rem_res_range(dev, slave, id, 1, RES_MPT, 0);
1661 if (err) 2076 if (err)
1662 break; 2077 break;
2078 mlx4_release_resource(dev, slave, RES_MPT, 1, 0);
1663 __mlx4_mpt_release(dev, index); 2079 __mlx4_mpt_release(dev, index);
1664 break; 2080 break;
1665 case RES_OP_MAP_ICM: 2081 case RES_OP_MAP_ICM:
@@ -1694,6 +2110,7 @@ static int cq_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1694 if (err) 2110 if (err)
1695 break; 2111 break;
1696 2112
2113 mlx4_release_resource(dev, slave, RES_CQ, 1, 0);
1697 __mlx4_cq_free_icm(dev, cqn); 2114 __mlx4_cq_free_icm(dev, cqn);
1698 break; 2115 break;
1699 2116
@@ -1718,6 +2135,7 @@ static int srq_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1718 if (err) 2135 if (err)
1719 break; 2136 break;
1720 2137
2138 mlx4_release_resource(dev, slave, RES_SRQ, 1, 0);
1721 __mlx4_srq_free_icm(dev, srqn); 2139 __mlx4_srq_free_icm(dev, srqn);
1722 break; 2140 break;
1723 2141
@@ -1730,14 +2148,14 @@ static int srq_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1730} 2148}
1731 2149
1732static int mac_free_res(struct mlx4_dev *dev, int slave, int op, int cmd, 2150static int mac_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1733 u64 in_param, u64 *out_param) 2151 u64 in_param, u64 *out_param, int in_port)
1734{ 2152{
1735 int port; 2153 int port;
1736 int err = 0; 2154 int err = 0;
1737 2155
1738 switch (op) { 2156 switch (op) {
1739 case RES_OP_RESERVE_AND_MAP: 2157 case RES_OP_RESERVE_AND_MAP:
1740 port = get_param_l(out_param); 2158 port = !in_port ? get_param_l(out_param) : in_port;
1741 mac_del_from_slave(dev, slave, in_param, port); 2159 mac_del_from_slave(dev, slave, in_param, port);
1742 __mlx4_unregister_mac(dev, port, in_param); 2160 __mlx4_unregister_mac(dev, port, in_param);
1743 break; 2161 break;
@@ -1751,9 +2169,27 @@ static int mac_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1751} 2169}
1752 2170
1753static int vlan_free_res(struct mlx4_dev *dev, int slave, int op, int cmd, 2171static int vlan_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1754 u64 in_param, u64 *out_param) 2172 u64 in_param, u64 *out_param, int port)
1755{ 2173{
1756 return 0; 2174 struct mlx4_priv *priv = mlx4_priv(dev);
2175 struct mlx4_slave_state *slave_state = priv->mfunc.master.slave_state;
2176 int err = 0;
2177
2178 switch (op) {
2179 case RES_OP_RESERVE_AND_MAP:
2180 if (slave_state[slave].old_vlan_api)
2181 return 0;
2182 if (!port)
2183 return -EINVAL;
2184 vlan_del_from_slave(dev, slave, in_param, port);
2185 __mlx4_unregister_vlan(dev, port, in_param);
2186 break;
2187 default:
2188 err = -EINVAL;
2189 break;
2190 }
2191
2192 return err;
1757} 2193}
1758 2194
1759static int counter_free_res(struct mlx4_dev *dev, int slave, int op, int cmd, 2195static int counter_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
@@ -1771,6 +2207,7 @@ static int counter_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1771 return err; 2207 return err;
1772 2208
1773 __mlx4_counter_free(dev, index); 2209 __mlx4_counter_free(dev, index);
2210 mlx4_release_resource(dev, slave, RES_COUNTER, 1, 0);
1774 2211
1775 return err; 2212 return err;
1776} 2213}
@@ -1803,7 +2240,7 @@ int mlx4_FREE_RES_wrapper(struct mlx4_dev *dev, int slave,
1803 int err = -EINVAL; 2240 int err = -EINVAL;
1804 int alop = vhcr->op_modifier; 2241 int alop = vhcr->op_modifier;
1805 2242
1806 switch (vhcr->in_modifier) { 2243 switch (vhcr->in_modifier & 0xFF) {
1807 case RES_QP: 2244 case RES_QP:
1808 err = qp_free_res(dev, slave, vhcr->op_modifier, alop, 2245 err = qp_free_res(dev, slave, vhcr->op_modifier, alop,
1809 vhcr->in_param); 2246 vhcr->in_param);
@@ -1831,12 +2268,14 @@ int mlx4_FREE_RES_wrapper(struct mlx4_dev *dev, int slave,
1831 2268
1832 case RES_MAC: 2269 case RES_MAC:
1833 err = mac_free_res(dev, slave, vhcr->op_modifier, alop, 2270 err = mac_free_res(dev, slave, vhcr->op_modifier, alop,
1834 vhcr->in_param, &vhcr->out_param); 2271 vhcr->in_param, &vhcr->out_param,
2272 (vhcr->in_modifier >> 8) & 0xFF);
1835 break; 2273 break;
1836 2274
1837 case RES_VLAN: 2275 case RES_VLAN:
1838 err = vlan_free_res(dev, slave, vhcr->op_modifier, alop, 2276 err = vlan_free_res(dev, slave, vhcr->op_modifier, alop,
1839 vhcr->in_param, &vhcr->out_param); 2277 vhcr->in_param, &vhcr->out_param,
2278 (vhcr->in_modifier >> 8) & 0xFF);
1840 break; 2279 break;
1841 2280
1842 case RES_COUNTER: 2281 case RES_COUNTER:
@@ -2136,6 +2575,12 @@ int mlx4_RST2INIT_QP_wrapper(struct mlx4_dev *dev, int slave,
2136 return err; 2575 return err;
2137 qp->local_qpn = local_qpn; 2576 qp->local_qpn = local_qpn;
2138 qp->sched_queue = 0; 2577 qp->sched_queue = 0;
2578 qp->param3 = 0;
2579 qp->vlan_control = 0;
2580 qp->fvl_rx = 0;
2581 qp->pri_path_fl = 0;
2582 qp->vlan_index = 0;
2583 qp->feup = 0;
2139 qp->qpc_flags = be32_to_cpu(qpc->flags); 2584 qp->qpc_flags = be32_to_cpu(qpc->flags);
2140 2585
2141 err = get_res(dev, slave, mtt_base, RES_MTT, &mtt); 2586 err = get_res(dev, slave, mtt_base, RES_MTT, &mtt);
@@ -2862,6 +3307,12 @@ int mlx4_INIT2RTR_QP_wrapper(struct mlx4_dev *dev, int slave,
2862 int qpn = vhcr->in_modifier & 0x7fffff; 3307 int qpn = vhcr->in_modifier & 0x7fffff;
2863 struct res_qp *qp; 3308 struct res_qp *qp;
2864 u8 orig_sched_queue; 3309 u8 orig_sched_queue;
3310 __be32 orig_param3 = qpc->param3;
3311 u8 orig_vlan_control = qpc->pri_path.vlan_control;
3312 u8 orig_fvl_rx = qpc->pri_path.fvl_rx;
3313 u8 orig_pri_path_fl = qpc->pri_path.fl;
3314 u8 orig_vlan_index = qpc->pri_path.vlan_index;
3315 u8 orig_feup = qpc->pri_path.feup;
2865 3316
2866 err = verify_qp_parameters(dev, inbox, QP_TRANS_INIT2RTR, slave); 3317 err = verify_qp_parameters(dev, inbox, QP_TRANS_INIT2RTR, slave);
2867 if (err) 3318 if (err)
@@ -2889,9 +3340,15 @@ out:
2889 * essentially the QOS value provided by the VF. This will be useful 3340 * essentially the QOS value provided by the VF. This will be useful
2890 * if we allow dynamic changes from VST back to VGT 3341 * if we allow dynamic changes from VST back to VGT
2891 */ 3342 */
2892 if (!err) 3343 if (!err) {
2893 qp->sched_queue = orig_sched_queue; 3344 qp->sched_queue = orig_sched_queue;
2894 3345 qp->param3 = orig_param3;
3346 qp->vlan_control = orig_vlan_control;
3347 qp->fvl_rx = orig_fvl_rx;
3348 qp->pri_path_fl = orig_pri_path_fl;
3349 qp->vlan_index = orig_vlan_index;
3350 qp->feup = orig_feup;
3351 }
2895 put_res(dev, slave, qpn, RES_QP); 3352 put_res(dev, slave, qpn, RES_QP);
2896 return err; 3353 return err;
2897} 3354}
@@ -3498,6 +3955,11 @@ static void rem_slave_qps(struct mlx4_dev *dev, int slave)
3498 &tracker->res_tree[RES_QP]); 3955 &tracker->res_tree[RES_QP]);
3499 list_del(&qp->com.list); 3956 list_del(&qp->com.list);
3500 spin_unlock_irq(mlx4_tlock(dev)); 3957 spin_unlock_irq(mlx4_tlock(dev));
3958 if (!valid_reserved(dev, slave, qpn)) {
3959 __mlx4_qp_release_range(dev, qpn, 1);
3960 mlx4_release_resource(dev, slave,
3961 RES_QP, 1, 0);
3962 }
3501 kfree(qp); 3963 kfree(qp);
3502 state = 0; 3964 state = 0;
3503 break; 3965 break;
@@ -3569,6 +4031,8 @@ static void rem_slave_srqs(struct mlx4_dev *dev, int slave)
3569 &tracker->res_tree[RES_SRQ]); 4031 &tracker->res_tree[RES_SRQ]);
3570 list_del(&srq->com.list); 4032 list_del(&srq->com.list);
3571 spin_unlock_irq(mlx4_tlock(dev)); 4033 spin_unlock_irq(mlx4_tlock(dev));
4034 mlx4_release_resource(dev, slave,
4035 RES_SRQ, 1, 0);
3572 kfree(srq); 4036 kfree(srq);
3573 state = 0; 4037 state = 0;
3574 break; 4038 break;
@@ -3635,6 +4099,8 @@ static void rem_slave_cqs(struct mlx4_dev *dev, int slave)
3635 &tracker->res_tree[RES_CQ]); 4099 &tracker->res_tree[RES_CQ]);
3636 list_del(&cq->com.list); 4100 list_del(&cq->com.list);
3637 spin_unlock_irq(mlx4_tlock(dev)); 4101 spin_unlock_irq(mlx4_tlock(dev));
4102 mlx4_release_resource(dev, slave,
4103 RES_CQ, 1, 0);
3638 kfree(cq); 4104 kfree(cq);
3639 state = 0; 4105 state = 0;
3640 break; 4106 break;
@@ -3698,6 +4164,8 @@ static void rem_slave_mrs(struct mlx4_dev *dev, int slave)
3698 &tracker->res_tree[RES_MPT]); 4164 &tracker->res_tree[RES_MPT]);
3699 list_del(&mpt->com.list); 4165 list_del(&mpt->com.list);
3700 spin_unlock_irq(mlx4_tlock(dev)); 4166 spin_unlock_irq(mlx4_tlock(dev));
4167 mlx4_release_resource(dev, slave,
4168 RES_MPT, 1, 0);
3701 kfree(mpt); 4169 kfree(mpt);
3702 state = 0; 4170 state = 0;
3703 break; 4171 break;
@@ -3767,6 +4235,8 @@ static void rem_slave_mtts(struct mlx4_dev *dev, int slave)
3767 &tracker->res_tree[RES_MTT]); 4235 &tracker->res_tree[RES_MTT]);
3768 list_del(&mtt->com.list); 4236 list_del(&mtt->com.list);
3769 spin_unlock_irq(mlx4_tlock(dev)); 4237 spin_unlock_irq(mlx4_tlock(dev));
4238 mlx4_release_resource(dev, slave, RES_MTT,
4239 1 << mtt->order, 0);
3770 kfree(mtt); 4240 kfree(mtt);
3771 state = 0; 4241 state = 0;
3772 break; 4242 break;
@@ -3925,6 +4395,7 @@ static void rem_slave_counters(struct mlx4_dev *dev, int slave)
3925 list_del(&counter->com.list); 4395 list_del(&counter->com.list);
3926 kfree(counter); 4396 kfree(counter);
3927 __mlx4_counter_free(dev, index); 4397 __mlx4_counter_free(dev, index);
4398 mlx4_release_resource(dev, slave, RES_COUNTER, 1, 0);
3928 } 4399 }
3929 } 4400 }
3930 spin_unlock_irq(mlx4_tlock(dev)); 4401 spin_unlock_irq(mlx4_tlock(dev));
@@ -3964,7 +4435,7 @@ void mlx4_delete_all_resources_for_slave(struct mlx4_dev *dev, int slave)
3964 struct mlx4_priv *priv = mlx4_priv(dev); 4435 struct mlx4_priv *priv = mlx4_priv(dev);
3965 4436
3966 mutex_lock(&priv->mfunc.master.res_tracker.slave_list[slave].mutex); 4437 mutex_lock(&priv->mfunc.master.res_tracker.slave_list[slave].mutex);
3967 /*VLAN*/ 4438 rem_slave_vlans(dev, slave);
3968 rem_slave_macs(dev, slave); 4439 rem_slave_macs(dev, slave);
3969 rem_slave_fs_rule(dev, slave); 4440 rem_slave_fs_rule(dev, slave);
3970 rem_slave_qps(dev, slave); 4441 rem_slave_qps(dev, slave);
@@ -3991,13 +4462,20 @@ void mlx4_vf_immed_vlan_work_handler(struct work_struct *_work)
3991 &tracker->slave_list[work->slave].res_list[RES_QP]; 4462 &tracker->slave_list[work->slave].res_list[RES_QP];
3992 struct res_qp *qp; 4463 struct res_qp *qp;
3993 struct res_qp *tmp; 4464 struct res_qp *tmp;
3994 u64 qp_mask = ((1ULL << MLX4_UPD_QP_PATH_MASK_ETH_TX_BLOCK_UNTAGGED) | 4465 u64 qp_path_mask_vlan_ctrl =
4466 ((1ULL << MLX4_UPD_QP_PATH_MASK_ETH_TX_BLOCK_UNTAGGED) |
3995 (1ULL << MLX4_UPD_QP_PATH_MASK_ETH_TX_BLOCK_1P) | 4467 (1ULL << MLX4_UPD_QP_PATH_MASK_ETH_TX_BLOCK_1P) |
3996 (1ULL << MLX4_UPD_QP_PATH_MASK_ETH_TX_BLOCK_TAGGED) | 4468 (1ULL << MLX4_UPD_QP_PATH_MASK_ETH_TX_BLOCK_TAGGED) |
3997 (1ULL << MLX4_UPD_QP_PATH_MASK_ETH_RX_BLOCK_UNTAGGED) | 4469 (1ULL << MLX4_UPD_QP_PATH_MASK_ETH_RX_BLOCK_UNTAGGED) |
3998 (1ULL << MLX4_UPD_QP_PATH_MASK_ETH_RX_BLOCK_1P) | 4470 (1ULL << MLX4_UPD_QP_PATH_MASK_ETH_RX_BLOCK_1P) |
3999 (1ULL << MLX4_UPD_QP_PATH_MASK_ETH_RX_BLOCK_TAGGED) | 4471 (1ULL << MLX4_UPD_QP_PATH_MASK_ETH_RX_BLOCK_TAGGED));
4000 (1ULL << MLX4_UPD_QP_PATH_MASK_VLAN_INDEX) | 4472
4473 u64 qp_path_mask = ((1ULL << MLX4_UPD_QP_PATH_MASK_VLAN_INDEX) |
4474 (1ULL << MLX4_UPD_QP_PATH_MASK_FVL) |
4475 (1ULL << MLX4_UPD_QP_PATH_MASK_CV) |
4476 (1ULL << MLX4_UPD_QP_PATH_MASK_ETH_HIDE_CQE_VLAN) |
4477 (1ULL << MLX4_UPD_QP_PATH_MASK_FEUP) |
4478 (1ULL << MLX4_UPD_QP_PATH_MASK_FVL_RX) |
4001 (1ULL << MLX4_UPD_QP_PATH_MASK_SCHED_QUEUE)); 4479 (1ULL << MLX4_UPD_QP_PATH_MASK_SCHED_QUEUE));
4002 4480
4003 int err; 4481 int err;
@@ -4029,9 +4507,7 @@ void mlx4_vf_immed_vlan_work_handler(struct work_struct *_work)
4029 MLX4_VLAN_CTRL_ETH_RX_BLOCK_UNTAGGED; 4507 MLX4_VLAN_CTRL_ETH_RX_BLOCK_UNTAGGED;
4030 4508
4031 upd_context = mailbox->buf; 4509 upd_context = mailbox->buf;
4032 upd_context->primary_addr_path_mask = cpu_to_be64(qp_mask); 4510 upd_context->qp_mask = cpu_to_be64(MLX4_UPD_QP_MASK_VSD);
4033 upd_context->qp_context.pri_path.vlan_control = vlan_control;
4034 upd_context->qp_context.pri_path.vlan_index = work->vlan_ix;
4035 4511
4036 spin_lock_irq(mlx4_tlock(dev)); 4512 spin_lock_irq(mlx4_tlock(dev));
4037 list_for_each_entry_safe(qp, tmp, qp_list, com.list) { 4513 list_for_each_entry_safe(qp, tmp, qp_list, com.list) {
@@ -4049,10 +4525,35 @@ void mlx4_vf_immed_vlan_work_handler(struct work_struct *_work)
4049 spin_lock_irq(mlx4_tlock(dev)); 4525 spin_lock_irq(mlx4_tlock(dev));
4050 continue; 4526 continue;
4051 } 4527 }
4052 upd_context->qp_context.pri_path.sched_queue = 4528 if (MLX4_QP_ST_RC == ((qp->qpc_flags >> 16) & 0xff))
4053 qp->sched_queue & 0xC7; 4529 upd_context->primary_addr_path_mask = cpu_to_be64(qp_path_mask);
4054 upd_context->qp_context.pri_path.sched_queue |= 4530 else
4055 ((work->qos & 0x7) << 3); 4531 upd_context->primary_addr_path_mask =
4532 cpu_to_be64(qp_path_mask | qp_path_mask_vlan_ctrl);
4533 if (work->vlan_id == MLX4_VGT) {
4534 upd_context->qp_context.param3 = qp->param3;
4535 upd_context->qp_context.pri_path.vlan_control = qp->vlan_control;
4536 upd_context->qp_context.pri_path.fvl_rx = qp->fvl_rx;
4537 upd_context->qp_context.pri_path.vlan_index = qp->vlan_index;
4538 upd_context->qp_context.pri_path.fl = qp->pri_path_fl;
4539 upd_context->qp_context.pri_path.feup = qp->feup;
4540 upd_context->qp_context.pri_path.sched_queue =
4541 qp->sched_queue;
4542 } else {
4543 upd_context->qp_context.param3 = qp->param3 & ~cpu_to_be32(MLX4_STRIP_VLAN);
4544 upd_context->qp_context.pri_path.vlan_control = vlan_control;
4545 upd_context->qp_context.pri_path.vlan_index = work->vlan_ix;
4546 upd_context->qp_context.pri_path.fvl_rx =
4547 qp->fvl_rx | MLX4_FVL_RX_FORCE_ETH_VLAN;
4548 upd_context->qp_context.pri_path.fl =
4549 qp->pri_path_fl | MLX4_FL_CV | MLX4_FL_ETH_HIDE_CQE_VLAN;
4550 upd_context->qp_context.pri_path.feup =
4551 qp->feup | MLX4_FEUP_FORCE_ETH_UP | MLX4_FVL_FORCE_ETH_VLAN;
4552 upd_context->qp_context.pri_path.sched_queue =
4553 qp->sched_queue & 0xC7;
4554 upd_context->qp_context.pri_path.sched_queue |=
4555 ((work->qos & 0x7) << 3);
4556 }
4056 4557
4057 err = mlx4_cmd(dev, mailbox->dma, 4558 err = mlx4_cmd(dev, mailbox->dma,
4058 qp->local_qpn & 0xffffff, 4559 qp->local_qpn & 0xffffff,
@@ -4081,7 +4582,7 @@ void mlx4_vf_immed_vlan_work_handler(struct work_struct *_work)
4081 if (work->flags & MLX4_VF_IMMED_VLAN_FLAG_VLAN && !errors && 4582 if (work->flags & MLX4_VF_IMMED_VLAN_FLAG_VLAN && !errors &&
4082 NO_INDX != work->orig_vlan_ix) 4583 NO_INDX != work->orig_vlan_ix)
4083 __mlx4_unregister_vlan(&work->priv->dev, work->port, 4584 __mlx4_unregister_vlan(&work->priv->dev, work->port,
4084 work->orig_vlan_ix); 4585 work->orig_vlan_id);
4085out: 4586out:
4086 kfree(work); 4587 kfree(work);
4087 return; 4588 return;
diff --git a/drivers/net/ethernet/mellanox/mlx4/srq.c b/drivers/net/ethernet/mellanox/mlx4/srq.c
index 79fd269e2c54..8fdf23753779 100644
--- a/drivers/net/ethernet/mellanox/mlx4/srq.c
+++ b/drivers/net/ethernet/mellanox/mlx4/srq.c
@@ -34,6 +34,7 @@
34#include <linux/init.h> 34#include <linux/init.h>
35 35
36#include <linux/mlx4/cmd.h> 36#include <linux/mlx4/cmd.h>
37#include <linux/mlx4/srq.h>
37#include <linux/export.h> 38#include <linux/export.h>
38#include <linux/gfp.h> 39#include <linux/gfp.h>
39 40
@@ -188,8 +189,6 @@ int mlx4_srq_alloc(struct mlx4_dev *dev, u32 pdn, u32 cqn, u16 xrcd,
188 } 189 }
189 190
190 srq_context = mailbox->buf; 191 srq_context = mailbox->buf;
191 memset(srq_context, 0, sizeof *srq_context);
192
193 srq_context->state_logsize_srqn = cpu_to_be32((ilog2(srq->max) << 24) | 192 srq_context->state_logsize_srqn = cpu_to_be32((ilog2(srq->max) << 24) |
194 srq->srqn); 193 srq->srqn);
195 srq_context->logstride = srq->wqe_shift - 4; 194 srq_context->logstride = srq->wqe_shift - 4;
diff --git a/drivers/net/ethernet/micrel/ks8851_mll.c b/drivers/net/ethernet/micrel/ks8851_mll.c
index 075f4e21d33d..c83d16dc7cd5 100644
--- a/drivers/net/ethernet/micrel/ks8851_mll.c
+++ b/drivers/net/ethernet/micrel/ks8851_mll.c
@@ -1248,7 +1248,7 @@ static void ks_set_mac(struct ks_net *ks, u8 *data)
1248 w = ((u & 0xFF) << 8) | ((u >> 8) & 0xFF); 1248 w = ((u & 0xFF) << 8) | ((u >> 8) & 0xFF);
1249 ks_wrreg16(ks, KS_MARL, w); 1249 ks_wrreg16(ks, KS_MARL, w);
1250 1250
1251 memcpy(ks->mac_addr, data, 6); 1251 memcpy(ks->mac_addr, data, ETH_ALEN);
1252 1252
1253 if (ks->enabled) 1253 if (ks->enabled)
1254 ks_start_rx(ks); 1254 ks_start_rx(ks);
@@ -1651,7 +1651,7 @@ static int ks8851_probe(struct platform_device *pdev)
1651 } 1651 }
1652 netdev_info(netdev, "Mac address is: %pM\n", ks->mac_addr); 1652 netdev_info(netdev, "Mac address is: %pM\n", ks->mac_addr);
1653 1653
1654 memcpy(netdev->dev_addr, ks->mac_addr, 6); 1654 memcpy(netdev->dev_addr, ks->mac_addr, ETH_ALEN);
1655 1655
1656 ks_set_mac(ks, netdev->dev_addr); 1656 ks_set_mac(ks, netdev->dev_addr);
1657 1657
diff --git a/drivers/net/ethernet/micrel/ksz884x.c b/drivers/net/ethernet/micrel/ksz884x.c
index 8ebc352bcbe6..ddd252a3da9c 100644
--- a/drivers/net/ethernet/micrel/ksz884x.c
+++ b/drivers/net/ethernet/micrel/ksz884x.c
@@ -7150,8 +7150,6 @@ static void pcidev_exit(struct pci_dev *pdev)
7150 struct platform_info *info = pci_get_drvdata(pdev); 7150 struct platform_info *info = pci_get_drvdata(pdev);
7151 struct dev_info *hw_priv = &info->dev_info; 7151 struct dev_info *hw_priv = &info->dev_info;
7152 7152
7153 pci_set_drvdata(pdev, NULL);
7154
7155 release_mem_region(pci_resource_start(pdev, 0), 7153 release_mem_region(pci_resource_start(pdev, 0),
7156 pci_resource_len(pdev, 0)); 7154 pci_resource_len(pdev, 0));
7157 for (i = 0; i < hw_priv->hw.dev_count; i++) { 7155 for (i = 0; i < hw_priv->hw.dev_count; i++) {
@@ -7227,7 +7225,7 @@ static int pcidev_suspend(struct pci_dev *pdev, pm_message_t state)
7227 7225
7228static char pcidev_name[] = "ksz884xp"; 7226static char pcidev_name[] = "ksz884xp";
7229 7227
7230static struct pci_device_id pcidev_table[] = { 7228static DEFINE_PCI_DEVICE_TABLE(pcidev_table) = {
7231 { PCI_VENDOR_ID_MICREL_KS, 0x8841, 7229 { PCI_VENDOR_ID_MICREL_KS, 0x8841,
7232 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0 }, 7230 PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0 },
7233 { PCI_VENDOR_ID_MICREL_KS, 0x8842, 7231 { PCI_VENDOR_ID_MICREL_KS, 0x8842,
diff --git a/drivers/net/ethernet/moxa/moxart_ether.c b/drivers/net/ethernet/moxa/moxart_ether.c
index ea54d95e5b9f..cbd013379252 100644
--- a/drivers/net/ethernet/moxa/moxart_ether.c
+++ b/drivers/net/ethernet/moxa/moxart_ether.c
@@ -26,7 +26,6 @@
26#include <linux/of_irq.h> 26#include <linux/of_irq.h>
27#include <linux/crc32.h> 27#include <linux/crc32.h>
28#include <linux/crc32c.h> 28#include <linux/crc32c.h>
29#include <linux/dma-mapping.h>
30 29
31#include "moxart_ether.h" 30#include "moxart_ether.h"
32 31
diff --git a/drivers/net/ethernet/myricom/myri10ge/myri10ge.c b/drivers/net/ethernet/myricom/myri10ge/myri10ge.c
index 149355b52ad0..68026f7e8ba3 100644
--- a/drivers/net/ethernet/myricom/myri10ge/myri10ge.c
+++ b/drivers/net/ethernet/myricom/myri10ge/myri10ge.c
@@ -934,7 +934,7 @@ static inline void myri10ge_ss_init_lock(struct myri10ge_slice_state *ss)
934 934
935static inline bool myri10ge_ss_lock_napi(struct myri10ge_slice_state *ss) 935static inline bool myri10ge_ss_lock_napi(struct myri10ge_slice_state *ss)
936{ 936{
937 int rc = true; 937 bool rc = true;
938 spin_lock(&ss->lock); 938 spin_lock(&ss->lock);
939 if ((ss->state & SLICE_LOCKED)) { 939 if ((ss->state & SLICE_LOCKED)) {
940 WARN_ON((ss->state & SLICE_STATE_NAPI)); 940 WARN_ON((ss->state & SLICE_STATE_NAPI));
@@ -957,7 +957,7 @@ static inline void myri10ge_ss_unlock_napi(struct myri10ge_slice_state *ss)
957 957
958static inline bool myri10ge_ss_lock_poll(struct myri10ge_slice_state *ss) 958static inline bool myri10ge_ss_lock_poll(struct myri10ge_slice_state *ss)
959{ 959{
960 int rc = true; 960 bool rc = true;
961 spin_lock_bh(&ss->lock); 961 spin_lock_bh(&ss->lock);
962 if ((ss->state & SLICE_LOCKED)) { 962 if ((ss->state & SLICE_LOCKED)) {
963 ss->state |= SLICE_STATE_POLL_YIELD; 963 ss->state |= SLICE_STATE_POLL_YIELD;
@@ -3164,7 +3164,7 @@ static void myri10ge_set_multicast_list(struct net_device *dev)
3164 3164
3165 /* Walk the multicast list, and add each address */ 3165 /* Walk the multicast list, and add each address */
3166 netdev_for_each_mc_addr(ha, dev) { 3166 netdev_for_each_mc_addr(ha, dev) {
3167 memcpy(data, &ha->addr, 6); 3167 memcpy(data, &ha->addr, ETH_ALEN);
3168 cmd.data0 = ntohl(data[0]); 3168 cmd.data0 = ntohl(data[0]);
3169 cmd.data1 = ntohl(data[1]); 3169 cmd.data1 = ntohl(data[1]);
3170 err = myri10ge_send_cmd(mgp, MXGEFW_JOIN_MULTICAST_GROUP, 3170 err = myri10ge_send_cmd(mgp, MXGEFW_JOIN_MULTICAST_GROUP,
@@ -3207,7 +3207,7 @@ static int myri10ge_set_mac_address(struct net_device *dev, void *addr)
3207 } 3207 }
3208 3208
3209 /* change the dev structure */ 3209 /* change the dev structure */
3210 memcpy(dev->dev_addr, sa->sa_data, 6); 3210 memcpy(dev->dev_addr, sa->sa_data, ETH_ALEN);
3211 return 0; 3211 return 0;
3212} 3212}
3213 3213
@@ -4208,7 +4208,6 @@ static void myri10ge_remove(struct pci_dev *pdev)
4208 set_fw_name(mgp, NULL, false); 4208 set_fw_name(mgp, NULL, false);
4209 free_netdev(netdev); 4209 free_netdev(netdev);
4210 pci_disable_device(pdev); 4210 pci_disable_device(pdev);
4211 pci_set_drvdata(pdev, NULL);
4212} 4211}
4213 4212
4214#define PCI_DEVICE_ID_MYRICOM_MYRI10GE_Z8E 0x0008 4213#define PCI_DEVICE_ID_MYRICOM_MYRI10GE_Z8E 0x0008
diff --git a/drivers/net/ethernet/natsemi/natsemi.c b/drivers/net/ethernet/natsemi/natsemi.c
index 7a5e295588b0..64ec2a437f46 100644
--- a/drivers/net/ethernet/natsemi/natsemi.c
+++ b/drivers/net/ethernet/natsemi/natsemi.c
@@ -970,7 +970,6 @@ static int natsemi_probe1(struct pci_dev *pdev, const struct pci_device_id *ent)
970 970
971 err_ioremap: 971 err_ioremap:
972 pci_release_regions(pdev); 972 pci_release_regions(pdev);
973 pci_set_drvdata(pdev, NULL);
974 973
975 err_pci_request_regions: 974 err_pci_request_regions:
976 free_netdev(dev); 975 free_netdev(dev);
@@ -3220,7 +3219,6 @@ static void natsemi_remove1(struct pci_dev *pdev)
3220 pci_release_regions (pdev); 3219 pci_release_regions (pdev);
3221 iounmap(ioaddr); 3220 iounmap(ioaddr);
3222 free_netdev (dev); 3221 free_netdev (dev);
3223 pci_set_drvdata(pdev, NULL);
3224} 3222}
3225 3223
3226#ifdef CONFIG_PM 3224#ifdef CONFIG_PM
diff --git a/drivers/net/ethernet/natsemi/xtsonic.c b/drivers/net/ethernet/natsemi/xtsonic.c
index 4da172ac5599..7007d212f3e4 100644
--- a/drivers/net/ethernet/natsemi/xtsonic.c
+++ b/drivers/net/ethernet/natsemi/xtsonic.c
@@ -264,6 +264,7 @@ int xtsonic_probe(struct platform_device *pdev)
264 264
265 lp = netdev_priv(dev); 265 lp = netdev_priv(dev);
266 lp->device = &pdev->dev; 266 lp->device = &pdev->dev;
267 platform_set_drvdata(pdev, dev);
267 SET_NETDEV_DEV(dev, &pdev->dev); 268 SET_NETDEV_DEV(dev, &pdev->dev);
268 netdev_boot_setup_check(dev); 269 netdev_boot_setup_check(dev);
269 270
diff --git a/drivers/net/ethernet/neterion/s2io.c b/drivers/net/ethernet/neterion/s2io.c
index 51b00941302c..9eeddbd0b2c7 100644
--- a/drivers/net/ethernet/neterion/s2io.c
+++ b/drivers/net/ethernet/neterion/s2io.c
@@ -8185,7 +8185,6 @@ mem_alloc_failed:
8185 free_shared_mem(sp); 8185 free_shared_mem(sp);
8186 pci_disable_device(pdev); 8186 pci_disable_device(pdev);
8187 pci_release_regions(pdev); 8187 pci_release_regions(pdev);
8188 pci_set_drvdata(pdev, NULL);
8189 free_netdev(dev); 8188 free_netdev(dev);
8190 8189
8191 return ret; 8190 return ret;
@@ -8221,7 +8220,6 @@ static void s2io_rem_nic(struct pci_dev *pdev)
8221 iounmap(sp->bar0); 8220 iounmap(sp->bar0);
8222 iounmap(sp->bar1); 8221 iounmap(sp->bar1);
8223 pci_release_regions(pdev); 8222 pci_release_regions(pdev);
8224 pci_set_drvdata(pdev, NULL);
8225 free_netdev(dev); 8223 free_netdev(dev);
8226 pci_disable_device(pdev); 8224 pci_disable_device(pdev);
8227} 8225}
diff --git a/drivers/net/ethernet/neterion/vxge/vxge-main.c b/drivers/net/ethernet/neterion/vxge/vxge-main.c
index 5a20eaf903dd..8614eeb7de81 100644
--- a/drivers/net/ethernet/neterion/vxge/vxge-main.c
+++ b/drivers/net/ethernet/neterion/vxge/vxge-main.c
@@ -4739,7 +4739,6 @@ _exit6:
4739_exit5: 4739_exit5:
4740 vxge_device_unregister(hldev); 4740 vxge_device_unregister(hldev);
4741_exit4: 4741_exit4:
4742 pci_set_drvdata(pdev, NULL);
4743 vxge_hw_device_terminate(hldev); 4742 vxge_hw_device_terminate(hldev);
4744 pci_disable_sriov(pdev); 4743 pci_disable_sriov(pdev);
4745_exit3: 4744_exit3:
@@ -4782,7 +4781,6 @@ static void vxge_remove(struct pci_dev *pdev)
4782 vxge_free_mac_add_list(&vdev->vpaths[i]); 4781 vxge_free_mac_add_list(&vdev->vpaths[i]);
4783 4782
4784 vxge_device_unregister(hldev); 4783 vxge_device_unregister(hldev);
4785 pci_set_drvdata(pdev, NULL);
4786 /* Do not call pci_disable_sriov here, as it will break child devices */ 4784 /* Do not call pci_disable_sriov here, as it will break child devices */
4787 vxge_hw_device_terminate(hldev); 4785 vxge_hw_device_terminate(hldev);
4788 iounmap(vdev->bar0); 4786 iounmap(vdev->bar0);
diff --git a/drivers/net/ethernet/octeon/octeon_mgmt.c b/drivers/net/ethernet/octeon/octeon_mgmt.c
index 622aa75904c4..1b326cbcd34b 100644
--- a/drivers/net/ethernet/octeon/octeon_mgmt.c
+++ b/drivers/net/ethernet/octeon/octeon_mgmt.c
@@ -1545,7 +1545,7 @@ static int octeon_mgmt_probe(struct platform_device *pdev)
1545 1545
1546 mac = of_get_mac_address(pdev->dev.of_node); 1546 mac = of_get_mac_address(pdev->dev.of_node);
1547 1547
1548 if (mac && is_valid_ether_addr(mac)) 1548 if (mac)
1549 memcpy(netdev->dev_addr, mac, ETH_ALEN); 1549 memcpy(netdev->dev_addr, mac, ETH_ALEN);
1550 else 1550 else
1551 eth_hw_addr_random(netdev); 1551 eth_hw_addr_random(netdev);
diff --git a/drivers/net/ethernet/oki-semi/pch_gbe/pch_gbe.h b/drivers/net/ethernet/oki-semi/pch_gbe/pch_gbe.h
index 6797b1075874..2a9003071d51 100644
--- a/drivers/net/ethernet/oki-semi/pch_gbe/pch_gbe.h
+++ b/drivers/net/ethernet/oki-semi/pch_gbe/pch_gbe.h
@@ -653,38 +653,38 @@ struct pch_gbe_adapter {
653extern const char pch_driver_version[]; 653extern const char pch_driver_version[];
654 654
655/* pch_gbe_main.c */ 655/* pch_gbe_main.c */
656extern int pch_gbe_up(struct pch_gbe_adapter *adapter); 656int pch_gbe_up(struct pch_gbe_adapter *adapter);
657extern void pch_gbe_down(struct pch_gbe_adapter *adapter); 657void pch_gbe_down(struct pch_gbe_adapter *adapter);
658extern void pch_gbe_reinit_locked(struct pch_gbe_adapter *adapter); 658void pch_gbe_reinit_locked(struct pch_gbe_adapter *adapter);
659extern void pch_gbe_reset(struct pch_gbe_adapter *adapter); 659void pch_gbe_reset(struct pch_gbe_adapter *adapter);
660extern int pch_gbe_setup_tx_resources(struct pch_gbe_adapter *adapter, 660int pch_gbe_setup_tx_resources(struct pch_gbe_adapter *adapter,
661 struct pch_gbe_tx_ring *txdr); 661 struct pch_gbe_tx_ring *txdr);
662extern int pch_gbe_setup_rx_resources(struct pch_gbe_adapter *adapter, 662int pch_gbe_setup_rx_resources(struct pch_gbe_adapter *adapter,
663 struct pch_gbe_rx_ring *rxdr); 663 struct pch_gbe_rx_ring *rxdr);
664extern void pch_gbe_free_tx_resources(struct pch_gbe_adapter *adapter, 664void pch_gbe_free_tx_resources(struct pch_gbe_adapter *adapter,
665 struct pch_gbe_tx_ring *tx_ring); 665 struct pch_gbe_tx_ring *tx_ring);
666extern void pch_gbe_free_rx_resources(struct pch_gbe_adapter *adapter, 666void pch_gbe_free_rx_resources(struct pch_gbe_adapter *adapter,
667 struct pch_gbe_rx_ring *rx_ring); 667 struct pch_gbe_rx_ring *rx_ring);
668extern void pch_gbe_update_stats(struct pch_gbe_adapter *adapter); 668void pch_gbe_update_stats(struct pch_gbe_adapter *adapter);
669extern u32 pch_ch_control_read(struct pci_dev *pdev); 669u32 pch_ch_control_read(struct pci_dev *pdev);
670extern void pch_ch_control_write(struct pci_dev *pdev, u32 val); 670void pch_ch_control_write(struct pci_dev *pdev, u32 val);
671extern u32 pch_ch_event_read(struct pci_dev *pdev); 671u32 pch_ch_event_read(struct pci_dev *pdev);
672extern void pch_ch_event_write(struct pci_dev *pdev, u32 val); 672void pch_ch_event_write(struct pci_dev *pdev, u32 val);
673extern u32 pch_src_uuid_lo_read(struct pci_dev *pdev); 673u32 pch_src_uuid_lo_read(struct pci_dev *pdev);
674extern u32 pch_src_uuid_hi_read(struct pci_dev *pdev); 674u32 pch_src_uuid_hi_read(struct pci_dev *pdev);
675extern u64 pch_rx_snap_read(struct pci_dev *pdev); 675u64 pch_rx_snap_read(struct pci_dev *pdev);
676extern u64 pch_tx_snap_read(struct pci_dev *pdev); 676u64 pch_tx_snap_read(struct pci_dev *pdev);
677extern int pch_set_station_address(u8 *addr, struct pci_dev *pdev); 677int pch_set_station_address(u8 *addr, struct pci_dev *pdev);
678 678
679/* pch_gbe_param.c */ 679/* pch_gbe_param.c */
680extern void pch_gbe_check_options(struct pch_gbe_adapter *adapter); 680void pch_gbe_check_options(struct pch_gbe_adapter *adapter);
681 681
682/* pch_gbe_ethtool.c */ 682/* pch_gbe_ethtool.c */
683extern void pch_gbe_set_ethtool_ops(struct net_device *netdev); 683void pch_gbe_set_ethtool_ops(struct net_device *netdev);
684 684
685/* pch_gbe_mac.c */ 685/* pch_gbe_mac.c */
686extern s32 pch_gbe_mac_force_mac_fc(struct pch_gbe_hw *hw); 686s32 pch_gbe_mac_force_mac_fc(struct pch_gbe_hw *hw);
687extern s32 pch_gbe_mac_read_mac_addr(struct pch_gbe_hw *hw); 687s32 pch_gbe_mac_read_mac_addr(struct pch_gbe_hw *hw);
688extern u16 pch_gbe_mac_ctrl_miim(struct pch_gbe_hw *hw, 688u16 pch_gbe_mac_ctrl_miim(struct pch_gbe_hw *hw, u32 addr, u32 dir, u32 reg,
689 u32 addr, u32 dir, u32 reg, u16 data); 689 u16 data);
690#endif /* _PCH_GBE_H_ */ 690#endif /* _PCH_GBE_H_ */
diff --git a/drivers/net/ethernet/packetengines/hamachi.c b/drivers/net/ethernet/packetengines/hamachi.c
index cac33e5f9bc2..b6bdeb3c1971 100644
--- a/drivers/net/ethernet/packetengines/hamachi.c
+++ b/drivers/net/ethernet/packetengines/hamachi.c
@@ -1910,7 +1910,6 @@ static void hamachi_remove_one(struct pci_dev *pdev)
1910 iounmap(hmp->base); 1910 iounmap(hmp->base);
1911 free_netdev(dev); 1911 free_netdev(dev);
1912 pci_release_regions(pdev); 1912 pci_release_regions(pdev);
1913 pci_set_drvdata(pdev, NULL);
1914 } 1913 }
1915} 1914}
1916 1915
diff --git a/drivers/net/ethernet/packetengines/yellowfin.c b/drivers/net/ethernet/packetengines/yellowfin.c
index d28593b1fc3e..07a890eb72ad 100644
--- a/drivers/net/ethernet/packetengines/yellowfin.c
+++ b/drivers/net/ethernet/packetengines/yellowfin.c
@@ -513,7 +513,6 @@ err_out_unmap_rx:
513err_out_unmap_tx: 513err_out_unmap_tx:
514 pci_free_consistent(pdev, TX_TOTAL_SIZE, np->tx_ring, np->tx_ring_dma); 514 pci_free_consistent(pdev, TX_TOTAL_SIZE, np->tx_ring, np->tx_ring_dma);
515err_out_cleardev: 515err_out_cleardev:
516 pci_set_drvdata(pdev, NULL);
517 pci_iounmap(pdev, ioaddr); 516 pci_iounmap(pdev, ioaddr);
518err_out_free_res: 517err_out_free_res:
519 pci_release_regions(pdev); 518 pci_release_regions(pdev);
@@ -1392,7 +1391,6 @@ static void yellowfin_remove_one(struct pci_dev *pdev)
1392 pci_release_regions (pdev); 1391 pci_release_regions (pdev);
1393 1392
1394 free_netdev (dev); 1393 free_netdev (dev);
1395 pci_set_drvdata(pdev, NULL);
1396} 1394}
1397 1395
1398 1396
diff --git a/drivers/net/ethernet/pasemi/pasemi_mac.c b/drivers/net/ethernet/pasemi/pasemi_mac.c
index 5b65356e7568..dbaa49e58b0c 100644
--- a/drivers/net/ethernet/pasemi/pasemi_mac.c
+++ b/drivers/net/ethernet/pasemi/pasemi_mac.c
@@ -1870,7 +1870,6 @@ static void pasemi_mac_remove(struct pci_dev *pdev)
1870 pasemi_dma_free_chan(&mac->tx->chan); 1870 pasemi_dma_free_chan(&mac->tx->chan);
1871 pasemi_dma_free_chan(&mac->rx->chan); 1871 pasemi_dma_free_chan(&mac->rx->chan);
1872 1872
1873 pci_set_drvdata(pdev, NULL);
1874 free_netdev(netdev); 1873 free_netdev(netdev);
1875} 1874}
1876 1875
diff --git a/drivers/net/ethernet/qlogic/netxen/netxen_nic.h b/drivers/net/ethernet/qlogic/netxen/netxen_nic.h
index 32675e16021e..9adcdbb49476 100644
--- a/drivers/net/ethernet/qlogic/netxen/netxen_nic.h
+++ b/drivers/net/ethernet/qlogic/netxen/netxen_nic.h
@@ -53,8 +53,8 @@
53 53
54#define _NETXEN_NIC_LINUX_MAJOR 4 54#define _NETXEN_NIC_LINUX_MAJOR 4
55#define _NETXEN_NIC_LINUX_MINOR 0 55#define _NETXEN_NIC_LINUX_MINOR 0
56#define _NETXEN_NIC_LINUX_SUBVERSION 81 56#define _NETXEN_NIC_LINUX_SUBVERSION 82
57#define NETXEN_NIC_LINUX_VERSIONID "4.0.81" 57#define NETXEN_NIC_LINUX_VERSIONID "4.0.82"
58 58
59#define NETXEN_VERSION_CODE(a, b, c) (((a) << 24) + ((b) << 16) + (c)) 59#define NETXEN_VERSION_CODE(a, b, c) (((a) << 24) + ((b) << 16) + (c))
60#define _major(v) (((v) >> 24) & 0xff) 60#define _major(v) (((v) >> 24) & 0xff)
@@ -1883,9 +1883,8 @@ static inline u32 netxen_tx_avail(struct nx_host_tx_ring *tx_ring)
1883 1883
1884int netxen_get_flash_mac_addr(struct netxen_adapter *adapter, u64 *mac); 1884int netxen_get_flash_mac_addr(struct netxen_adapter *adapter, u64 *mac);
1885int netxen_p3_get_mac_addr(struct netxen_adapter *adapter, u64 *mac); 1885int netxen_p3_get_mac_addr(struct netxen_adapter *adapter, u64 *mac);
1886extern void netxen_change_ringparam(struct netxen_adapter *adapter); 1886void netxen_change_ringparam(struct netxen_adapter *adapter);
1887extern int netxen_rom_fast_read(struct netxen_adapter *adapter, int addr, 1887int netxen_rom_fast_read(struct netxen_adapter *adapter, int addr, int *valp);
1888 int *valp);
1889 1888
1890extern const struct ethtool_ops netxen_nic_ethtool_ops; 1889extern const struct ethtool_ops netxen_nic_ethtool_ops;
1891 1890
diff --git a/drivers/net/ethernet/qlogic/netxen/netxen_nic_hdr.h b/drivers/net/ethernet/qlogic/netxen/netxen_nic_hdr.h
index 32c790659f9c..0c64c82b9acf 100644
--- a/drivers/net/ethernet/qlogic/netxen/netxen_nic_hdr.h
+++ b/drivers/net/ethernet/qlogic/netxen/netxen_nic_hdr.h
@@ -958,6 +958,7 @@ enum {
958#define NETXEN_PEG_HALT_STATUS2 (NETXEN_CAM_RAM(0xac)) 958#define NETXEN_PEG_HALT_STATUS2 (NETXEN_CAM_RAM(0xac))
959#define NX_CRB_DEV_REF_COUNT (NETXEN_CAM_RAM(0x138)) 959#define NX_CRB_DEV_REF_COUNT (NETXEN_CAM_RAM(0x138))
960#define NX_CRB_DEV_STATE (NETXEN_CAM_RAM(0x140)) 960#define NX_CRB_DEV_STATE (NETXEN_CAM_RAM(0x140))
961#define NETXEN_ULA_KEY (NETXEN_CAM_RAM(0x178))
961 962
962/* MiniDIMM related macros */ 963/* MiniDIMM related macros */
963#define NETXEN_DIMM_CAPABILITY (NETXEN_CAM_RAM(0x258)) 964#define NETXEN_DIMM_CAPABILITY (NETXEN_CAM_RAM(0x258))
diff --git a/drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c b/drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
index 8375cbde9969..67efe754367d 100644
--- a/drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
+++ b/drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c
@@ -648,7 +648,7 @@ nx_p3_sre_macaddr_change(struct netxen_adapter *adapter, u8 *addr, unsigned op)
648 648
649 mac_req = (nx_mac_req_t *)&req.words[0]; 649 mac_req = (nx_mac_req_t *)&req.words[0];
650 mac_req->op = op; 650 mac_req->op = op;
651 memcpy(mac_req->mac_addr, addr, 6); 651 memcpy(mac_req->mac_addr, addr, ETH_ALEN);
652 652
653 return netxen_send_cmd_descs(adapter, (struct cmd_desc_type0 *)&req, 1); 653 return netxen_send_cmd_descs(adapter, (struct cmd_desc_type0 *)&req, 1);
654} 654}
diff --git a/drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c b/drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c
index cbd75f97ffb3..3bec8cfebf99 100644
--- a/drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c
+++ b/drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c
@@ -1415,6 +1415,32 @@ netxen_setup_netdev(struct netxen_adapter *adapter,
1415 return 0; 1415 return 0;
1416} 1416}
1417 1417
1418#define NETXEN_ULA_ADAPTER_KEY (0xdaddad01)
1419#define NETXEN_NON_ULA_ADAPTER_KEY (0xdaddad00)
1420
1421static void netxen_read_ula_info(struct netxen_adapter *adapter)
1422{
1423 u32 temp;
1424
1425 /* Print ULA info only once for an adapter */
1426 if (adapter->portnum != 0)
1427 return;
1428
1429 temp = NXRD32(adapter, NETXEN_ULA_KEY);
1430 switch (temp) {
1431 case NETXEN_ULA_ADAPTER_KEY:
1432 dev_info(&adapter->pdev->dev, "ULA adapter");
1433 break;
1434 case NETXEN_NON_ULA_ADAPTER_KEY:
1435 dev_info(&adapter->pdev->dev, "non ULA adapter");
1436 break;
1437 default:
1438 break;
1439 }
1440
1441 return;
1442}
1443
1418#ifdef CONFIG_PCIEAER 1444#ifdef CONFIG_PCIEAER
1419static void netxen_mask_aer_correctable(struct netxen_adapter *adapter) 1445static void netxen_mask_aer_correctable(struct netxen_adapter *adapter)
1420{ 1446{
@@ -1561,6 +1587,8 @@ netxen_nic_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
1561 goto err_out_disable_msi; 1587 goto err_out_disable_msi;
1562 } 1588 }
1563 1589
1590 netxen_read_ula_info(adapter);
1591
1564 err = netxen_setup_netdev(adapter, netdev); 1592 err = netxen_setup_netdev(adapter, netdev);
1565 if (err) 1593 if (err)
1566 goto err_out_disable_msi; 1594 goto err_out_disable_msi;
@@ -1602,7 +1630,6 @@ err_out_free_res:
1602 pci_release_regions(pdev); 1630 pci_release_regions(pdev);
1603 1631
1604err_out_disable_pdev: 1632err_out_disable_pdev:
1605 pci_set_drvdata(pdev, NULL);
1606 pci_disable_device(pdev); 1633 pci_disable_device(pdev);
1607 return err; 1634 return err;
1608} 1635}
@@ -1661,7 +1688,6 @@ static void netxen_nic_remove(struct pci_dev *pdev)
1661 1688
1662 pci_release_regions(pdev); 1689 pci_release_regions(pdev);
1663 pci_disable_device(pdev); 1690 pci_disable_device(pdev);
1664 pci_set_drvdata(pdev, NULL);
1665 1691
1666 free_netdev(netdev); 1692 free_netdev(netdev);
1667} 1693}
diff --git a/drivers/net/ethernet/qlogic/qla3xxx.c b/drivers/net/ethernet/qlogic/qla3xxx.c
index 91a8fcd6c246..0758b9435358 100644
--- a/drivers/net/ethernet/qlogic/qla3xxx.c
+++ b/drivers/net/ethernet/qlogic/qla3xxx.c
@@ -3916,7 +3916,6 @@ err_out_free_regions:
3916 pci_release_regions(pdev); 3916 pci_release_regions(pdev);
3917err_out_disable_pdev: 3917err_out_disable_pdev:
3918 pci_disable_device(pdev); 3918 pci_disable_device(pdev);
3919 pci_set_drvdata(pdev, NULL);
3920err_out: 3919err_out:
3921 return err; 3920 return err;
3922} 3921}
@@ -3939,7 +3938,6 @@ static void ql3xxx_remove(struct pci_dev *pdev)
3939 3938
3940 iounmap(qdev->mem_map_registers); 3939 iounmap(qdev->mem_map_registers);
3941 pci_release_regions(pdev); 3940 pci_release_regions(pdev);
3942 pci_set_drvdata(pdev, NULL);
3943 free_netdev(ndev); 3941 free_netdev(ndev);
3944} 3942}
3945 3943
diff --git a/drivers/net/ethernet/qlogic/qlcnic/qlcnic.h b/drivers/net/ethernet/qlogic/qlcnic/qlcnic.h
index 81bf83604c4f..631ea0ac1cd8 100644
--- a/drivers/net/ethernet/qlogic/qlcnic/qlcnic.h
+++ b/drivers/net/ethernet/qlogic/qlcnic/qlcnic.h
@@ -38,8 +38,8 @@
38 38
39#define _QLCNIC_LINUX_MAJOR 5 39#define _QLCNIC_LINUX_MAJOR 5
40#define _QLCNIC_LINUX_MINOR 3 40#define _QLCNIC_LINUX_MINOR 3
41#define _QLCNIC_LINUX_SUBVERSION 50 41#define _QLCNIC_LINUX_SUBVERSION 52
42#define QLCNIC_LINUX_VERSIONID "5.3.50" 42#define QLCNIC_LINUX_VERSIONID "5.3.52"
43#define QLCNIC_DRV_IDC_VER 0x01 43#define QLCNIC_DRV_IDC_VER 0x01
44#define QLCNIC_DRIVER_VERSION ((_QLCNIC_LINUX_MAJOR << 16) |\ 44#define QLCNIC_DRIVER_VERSION ((_QLCNIC_LINUX_MAJOR << 16) |\
45 (_QLCNIC_LINUX_MINOR << 8) | (_QLCNIC_LINUX_SUBVERSION)) 45 (_QLCNIC_LINUX_MINOR << 8) | (_QLCNIC_LINUX_SUBVERSION))
@@ -98,8 +98,22 @@
98#define TX_STOP_THRESH ((MAX_SKB_FRAGS >> 2) + MAX_TSO_HEADER_DESC \ 98#define TX_STOP_THRESH ((MAX_SKB_FRAGS >> 2) + MAX_TSO_HEADER_DESC \
99 + MGMT_CMD_DESC_RESV) 99 + MGMT_CMD_DESC_RESV)
100#define QLCNIC_MAX_TX_TIMEOUTS 2 100#define QLCNIC_MAX_TX_TIMEOUTS 2
101#define QLCNIC_MAX_TX_RINGS 8 101
102#define QLCNIC_MAX_SDS_RINGS 8 102/* Driver will use 1 Tx ring in INT-x/MSI/SRIOV mode. */
103#define QLCNIC_SINGLE_RING 1
104#define QLCNIC_DEF_SDS_RINGS 4
105#define QLCNIC_DEF_TX_RINGS 4
106#define QLCNIC_MAX_VNIC_TX_RINGS 4
107#define QLCNIC_MAX_VNIC_SDS_RINGS 4
108
109enum qlcnic_queue_type {
110 QLCNIC_TX_QUEUE = 1,
111 QLCNIC_RX_QUEUE,
112};
113
114/* Operational mode for driver */
115#define QLCNIC_VNIC_MODE 0xFF
116#define QLCNIC_DEFAULT_MODE 0x0
103 117
104/* 118/*
105 * Following are the states of the Phantom. Phantom will set them and 119 * Following are the states of the Phantom. Phantom will set them and
@@ -533,6 +547,14 @@ struct qlcnic_host_sds_ring {
533 char name[IFNAMSIZ + 12]; 547 char name[IFNAMSIZ + 12];
534} ____cacheline_internodealigned_in_smp; 548} ____cacheline_internodealigned_in_smp;
535 549
550struct qlcnic_tx_queue_stats {
551 u64 xmit_on;
552 u64 xmit_off;
553 u64 xmit_called;
554 u64 xmit_finished;
555 u64 tx_bytes;
556};
557
536struct qlcnic_host_tx_ring { 558struct qlcnic_host_tx_ring {
537 int irq; 559 int irq;
538 void __iomem *crb_intr_mask; 560 void __iomem *crb_intr_mask;
@@ -544,10 +566,7 @@ struct qlcnic_host_tx_ring {
544 u32 sw_consumer; 566 u32 sw_consumer;
545 u32 num_desc; 567 u32 num_desc;
546 568
547 u64 xmit_on; 569 struct qlcnic_tx_queue_stats tx_stats;
548 u64 xmit_off;
549 u64 xmit_called;
550 u64 xmit_finished;
551 570
552 void __iomem *crb_cmd_producer; 571 void __iomem *crb_cmd_producer;
553 struct cmd_desc_type0 *desc_head; 572 struct cmd_desc_type0 *desc_head;
@@ -940,8 +959,6 @@ struct qlcnic_ipaddr {
940#define QLCNIC_BEACON_EANBLE 0xC 959#define QLCNIC_BEACON_EANBLE 0xC
941#define QLCNIC_BEACON_DISABLE 0xD 960#define QLCNIC_BEACON_DISABLE 0xD
942 961
943#define QLCNIC_DEF_NUM_STS_DESC_RINGS 4
944#define QLCNIC_DEF_NUM_TX_RINGS 4
945#define QLCNIC_MSIX_TBL_SPACE 8192 962#define QLCNIC_MSIX_TBL_SPACE 8192
946#define QLCNIC_PCI_REG_MSIX_TBL 0x44 963#define QLCNIC_PCI_REG_MSIX_TBL 0x44
947#define QLCNIC_MSIX_TBL_PGSIZE 4096 964#define QLCNIC_MSIX_TBL_PGSIZE 4096
@@ -961,8 +978,7 @@ struct qlcnic_ipaddr {
961#define __QLCNIC_SRIOV_CAPABLE 11 978#define __QLCNIC_SRIOV_CAPABLE 11
962#define __QLCNIC_MBX_POLL_ENABLE 12 979#define __QLCNIC_MBX_POLL_ENABLE 12
963#define __QLCNIC_DIAG_MODE 13 980#define __QLCNIC_DIAG_MODE 13
964#define __QLCNIC_DCB_STATE 14 981#define __QLCNIC_MAINTENANCE_MODE 16
965#define __QLCNIC_DCB_IN_AEN 15
966 982
967#define QLCNIC_INTERRUPT_TEST 1 983#define QLCNIC_INTERRUPT_TEST 1
968#define QLCNIC_LOOPBACK_TEST 2 984#define QLCNIC_LOOPBACK_TEST 2
@@ -1013,7 +1029,6 @@ struct qlcnic_adapter {
1013 unsigned long state; 1029 unsigned long state;
1014 u32 flags; 1030 u32 flags;
1015 1031
1016 int max_drv_tx_rings;
1017 u16 num_txd; 1032 u16 num_txd;
1018 u16 num_rxd; 1033 u16 num_rxd;
1019 u16 num_jumbo_rxd; 1034 u16 num_jumbo_rxd;
@@ -1021,7 +1036,13 @@ struct qlcnic_adapter {
1021 u16 max_jumbo_rxd; 1036 u16 max_jumbo_rxd;
1022 1037
1023 u8 max_rds_rings; 1038 u8 max_rds_rings;
1024 u8 max_sds_rings; 1039
1040 u8 max_sds_rings; /* max sds rings supported by adapter */
1041 u8 max_tx_rings; /* max tx rings supported by adapter */
1042
1043 u8 drv_tx_rings; /* max tx rings supported by driver */
1044 u8 drv_sds_rings; /* max sds rings supported by driver */
1045
1025 u8 rx_csum; 1046 u8 rx_csum;
1026 u8 portnum; 1047 u8 portnum;
1027 1048
@@ -1199,6 +1220,7 @@ struct qlcnic_npar_info {
1199 u8 promisc_mode; 1220 u8 promisc_mode;
1200 u8 offload_flags; 1221 u8 offload_flags;
1201 u8 pci_func; 1222 u8 pci_func;
1223 u8 mac[ETH_ALEN];
1202}; 1224};
1203 1225
1204struct qlcnic_eswitch { 1226struct qlcnic_eswitch {
@@ -1543,12 +1565,13 @@ int qlcnic_loopback_test(struct net_device *, u8);
1543 1565
1544/* Functions from qlcnic_main.c */ 1566/* Functions from qlcnic_main.c */
1545int qlcnic_reset_context(struct qlcnic_adapter *); 1567int qlcnic_reset_context(struct qlcnic_adapter *);
1546void qlcnic_diag_free_res(struct net_device *netdev, int max_sds_rings); 1568void qlcnic_diag_free_res(struct net_device *netdev, int);
1547int qlcnic_diag_alloc_res(struct net_device *netdev, int test); 1569int qlcnic_diag_alloc_res(struct net_device *netdev, int);
1548netdev_tx_t qlcnic_xmit_frame(struct sk_buff *skb, struct net_device *netdev); 1570netdev_tx_t qlcnic_xmit_frame(struct sk_buff *, struct net_device *);
1549int qlcnic_set_max_rss(struct qlcnic_adapter *, u8, int); 1571void qlcnic_set_tx_ring_count(struct qlcnic_adapter *, u8);
1550int qlcnic_validate_max_rss(struct qlcnic_adapter *, __u32); 1572void qlcnic_set_sds_ring_count(struct qlcnic_adapter *, u8);
1551int qlcnic_validate_max_tx_rings(struct qlcnic_adapter *, u32 txq); 1573int qlcnic_setup_rings(struct qlcnic_adapter *, u8, u8);
1574int qlcnic_validate_rings(struct qlcnic_adapter *, __u32, int);
1552void qlcnic_alloc_lb_filters_mem(struct qlcnic_adapter *adapter); 1575void qlcnic_alloc_lb_filters_mem(struct qlcnic_adapter *adapter);
1553void qlcnic_82xx_set_mac_filter_count(struct qlcnic_adapter *); 1576void qlcnic_82xx_set_mac_filter_count(struct qlcnic_adapter *);
1554int qlcnic_enable_msix(struct qlcnic_adapter *, u32); 1577int qlcnic_enable_msix(struct qlcnic_adapter *, u32);
@@ -1641,19 +1664,18 @@ static inline u32 qlcnic_tx_avail(struct qlcnic_host_tx_ring *tx_ring)
1641static inline int qlcnic_set_real_num_queues(struct qlcnic_adapter *adapter, 1664static inline int qlcnic_set_real_num_queues(struct qlcnic_adapter *adapter,
1642 struct net_device *netdev) 1665 struct net_device *netdev)
1643{ 1666{
1644 int err, tx_q; 1667 int err;
1645
1646 tx_q = adapter->max_drv_tx_rings;
1647 1668
1648 netdev->num_tx_queues = tx_q; 1669 netdev->num_tx_queues = adapter->drv_tx_rings;
1649 netdev->real_num_tx_queues = tx_q; 1670 netdev->real_num_tx_queues = adapter->drv_tx_rings;
1650 1671
1651 err = netif_set_real_num_tx_queues(netdev, tx_q); 1672 err = netif_set_real_num_tx_queues(netdev, adapter->drv_tx_rings);
1652 if (err) 1673 if (err)
1653 dev_err(&adapter->pdev->dev, "failed to set %d Tx queues\n", 1674 dev_err(&adapter->pdev->dev, "failed to set %d Tx queues\n",
1654 tx_q); 1675 adapter->drv_tx_rings);
1655 else 1676 else
1656 dev_info(&adapter->pdev->dev, "set %d Tx queues\n", tx_q); 1677 dev_info(&adapter->pdev->dev, "Set %d Tx queues\n",
1678 adapter->drv_tx_rings);
1657 1679
1658 return err; 1680 return err;
1659} 1681}
@@ -1695,7 +1717,7 @@ struct qlcnic_hardware_ops {
1695 int (*write_reg) (struct qlcnic_adapter *, ulong, u32); 1717 int (*write_reg) (struct qlcnic_adapter *, ulong, u32);
1696 void (*get_ocm_win) (struct qlcnic_hardware_context *); 1718 void (*get_ocm_win) (struct qlcnic_hardware_context *);
1697 int (*get_mac_address) (struct qlcnic_adapter *, u8 *, u8); 1719 int (*get_mac_address) (struct qlcnic_adapter *, u8 *, u8);
1698 int (*setup_intr) (struct qlcnic_adapter *, u8, int); 1720 int (*setup_intr) (struct qlcnic_adapter *);
1699 int (*alloc_mbx_args)(struct qlcnic_cmd_args *, 1721 int (*alloc_mbx_args)(struct qlcnic_cmd_args *,
1700 struct qlcnic_adapter *, u32); 1722 struct qlcnic_adapter *, u32);
1701 int (*mbx_cmd) (struct qlcnic_adapter *, struct qlcnic_cmd_args *); 1723 int (*mbx_cmd) (struct qlcnic_adapter *, struct qlcnic_cmd_args *);
@@ -1766,10 +1788,9 @@ static inline int qlcnic_get_mac_address(struct qlcnic_adapter *adapter,
1766 return adapter->ahw->hw_ops->get_mac_address(adapter, mac, function); 1788 return adapter->ahw->hw_ops->get_mac_address(adapter, mac, function);
1767} 1789}
1768 1790
1769static inline int qlcnic_setup_intr(struct qlcnic_adapter *adapter, 1791static inline int qlcnic_setup_intr(struct qlcnic_adapter *adapter)
1770 u8 num_intr, int txq)
1771{ 1792{
1772 return adapter->ahw->hw_ops->setup_intr(adapter, num_intr, txq); 1793 return adapter->ahw->hw_ops->setup_intr(adapter);
1773} 1794}
1774 1795
1775static inline int qlcnic_alloc_mbx_args(struct qlcnic_cmd_args *mbx, 1796static inline int qlcnic_alloc_mbx_args(struct qlcnic_cmd_args *mbx,
@@ -2005,7 +2026,7 @@ static inline bool qlcnic_check_multi_tx(struct qlcnic_adapter *adapter)
2005static inline void qlcnic_disable_multi_tx(struct qlcnic_adapter *adapter) 2026static inline void qlcnic_disable_multi_tx(struct qlcnic_adapter *adapter)
2006{ 2027{
2007 test_and_clear_bit(__QLCNIC_MULTI_TX_UNIQUE, &adapter->state); 2028 test_and_clear_bit(__QLCNIC_MULTI_TX_UNIQUE, &adapter->state);
2008 adapter->max_drv_tx_rings = 1; 2029 adapter->drv_tx_rings = QLCNIC_SINGLE_RING;
2009} 2030}
2010 2031
2011/* When operating in a muti tx mode, driver needs to write 0x1 2032/* When operating in a muti tx mode, driver needs to write 0x1
@@ -2115,98 +2136,4 @@ static inline bool qlcnic_sriov_vf_check(struct qlcnic_adapter *adapter)
2115 2136
2116 return status; 2137 return status;
2117} 2138}
2118
2119static inline int qlcnic_dcb_get_hw_capability(struct qlcnic_adapter *adapter)
2120{
2121 struct qlcnic_dcb *dcb = adapter->dcb;
2122
2123 if (dcb && dcb->ops->get_hw_capability)
2124 return dcb->ops->get_hw_capability(adapter);
2125
2126 return 0;
2127}
2128
2129static inline void qlcnic_dcb_free(struct qlcnic_adapter *adapter)
2130{
2131 struct qlcnic_dcb *dcb = adapter->dcb;
2132
2133 if (dcb && dcb->ops->free)
2134 dcb->ops->free(adapter);
2135}
2136
2137static inline int qlcnic_dcb_attach(struct qlcnic_adapter *adapter)
2138{
2139 struct qlcnic_dcb *dcb = adapter->dcb;
2140
2141 if (dcb && dcb->ops->attach)
2142 return dcb->ops->attach(adapter);
2143
2144 return 0;
2145}
2146
2147static inline int
2148qlcnic_dcb_query_hw_capability(struct qlcnic_adapter *adapter, char *buf)
2149{
2150 struct qlcnic_dcb *dcb = adapter->dcb;
2151
2152 if (dcb && dcb->ops->query_hw_capability)
2153 return dcb->ops->query_hw_capability(adapter, buf);
2154
2155 return 0;
2156}
2157
2158static inline void qlcnic_dcb_get_info(struct qlcnic_adapter *adapter)
2159{
2160 struct qlcnic_dcb *dcb = adapter->dcb;
2161
2162 if (dcb && dcb->ops->get_info)
2163 dcb->ops->get_info(adapter);
2164}
2165
2166static inline int
2167qlcnic_dcb_query_cee_param(struct qlcnic_adapter *adapter, char *buf, u8 type)
2168{
2169 struct qlcnic_dcb *dcb = adapter->dcb;
2170
2171 if (dcb && dcb->ops->query_cee_param)
2172 return dcb->ops->query_cee_param(adapter, buf, type);
2173
2174 return 0;
2175}
2176
2177static inline int qlcnic_dcb_get_cee_cfg(struct qlcnic_adapter *adapter)
2178{
2179 struct qlcnic_dcb *dcb = adapter->dcb;
2180
2181 if (dcb && dcb->ops->get_cee_cfg)
2182 return dcb->ops->get_cee_cfg(adapter);
2183
2184 return 0;
2185}
2186
2187static inline void
2188qlcnic_dcb_register_aen(struct qlcnic_adapter *adapter, u8 flag)
2189{
2190 struct qlcnic_dcb *dcb = adapter->dcb;
2191
2192 if (dcb && dcb->ops->register_aen)
2193 dcb->ops->register_aen(adapter, flag);
2194}
2195
2196static inline void qlcnic_dcb_handle_aen(struct qlcnic_adapter *adapter,
2197 void *msg)
2198{
2199 struct qlcnic_dcb *dcb = adapter->dcb;
2200
2201 if (dcb && dcb->ops->handle_aen)
2202 dcb->ops->handle_aen(adapter, msg);
2203}
2204
2205static inline void qlcnic_dcb_init_dcbnl_ops(struct qlcnic_adapter *adapter)
2206{
2207 struct qlcnic_dcb *dcb = adapter->dcb;
2208
2209 if (dcb && dcb->ops->init_dcbnl_ops)
2210 dcb->ops->init_dcbnl_ops(adapter);
2211}
2212#endif /* __QLCNIC_H_ */ 2139#endif /* __QLCNIC_H_ */
diff --git a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
index ace217c447dd..09810ddd11ec 100644
--- a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
+++ b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
@@ -13,7 +13,6 @@
13#include <linux/interrupt.h> 13#include <linux/interrupt.h>
14#include <linux/aer.h> 14#include <linux/aer.h>
15 15
16#define QLCNIC_MAX_TX_QUEUES 1
17#define RSS_HASHTYPE_IP_TCP 0x3 16#define RSS_HASHTYPE_IP_TCP 0x3
18#define QLC_83XX_FW_MBX_CMD 0 17#define QLC_83XX_FW_MBX_CMD 0
19 18
@@ -268,20 +267,18 @@ int qlcnic_83xx_wrt_reg_indirect(struct qlcnic_adapter *adapter, ulong addr,
268 } 267 }
269} 268}
270 269
271int qlcnic_83xx_setup_intr(struct qlcnic_adapter *adapter, u8 num_intr, int txq) 270int qlcnic_83xx_setup_intr(struct qlcnic_adapter *adapter)
272{ 271{
273 int err, i, num_msix; 272 int err, i, num_msix;
274 struct qlcnic_hardware_context *ahw = adapter->ahw; 273 struct qlcnic_hardware_context *ahw = adapter->ahw;
275 274
276 if (!num_intr) 275 num_msix = adapter->drv_sds_rings;
277 num_intr = QLCNIC_DEF_NUM_STS_DESC_RINGS; 276
278 num_msix = rounddown_pow_of_two(min_t(int, num_online_cpus(),
279 num_intr));
280 /* account for AEN interrupt MSI-X based interrupts */ 277 /* account for AEN interrupt MSI-X based interrupts */
281 num_msix += 1; 278 num_msix += 1;
282 279
283 if (!(adapter->flags & QLCNIC_TX_INTR_SHARED)) 280 if (!(adapter->flags & QLCNIC_TX_INTR_SHARED))
284 num_msix += adapter->max_drv_tx_rings; 281 num_msix += adapter->drv_tx_rings;
285 282
286 err = qlcnic_enable_msix(adapter, num_msix); 283 err = qlcnic_enable_msix(adapter, num_msix);
287 if (err == -ENOMEM) 284 if (err == -ENOMEM)
@@ -325,7 +322,8 @@ inline void qlcnic_83xx_clear_legacy_intr_mask(struct qlcnic_adapter *adapter)
325 322
326inline void qlcnic_83xx_set_legacy_intr_mask(struct qlcnic_adapter *adapter) 323inline void qlcnic_83xx_set_legacy_intr_mask(struct qlcnic_adapter *adapter)
327{ 324{
328 writel(1, adapter->tgt_mask_reg); 325 if (adapter->tgt_mask_reg)
326 writel(1, adapter->tgt_mask_reg);
329} 327}
330 328
331/* Enable MSI-x and INT-x interrupts */ 329/* Enable MSI-x and INT-x interrupts */
@@ -498,8 +496,11 @@ void qlcnic_83xx_free_mbx_intr(struct qlcnic_adapter *adapter)
498 num_msix = 0; 496 num_msix = 0;
499 497
500 msleep(20); 498 msleep(20);
501 synchronize_irq(adapter->msix_entries[num_msix].vector); 499
502 free_irq(adapter->msix_entries[num_msix].vector, adapter); 500 if (adapter->msix_entries) {
501 synchronize_irq(adapter->msix_entries[num_msix].vector);
502 free_irq(adapter->msix_entries[num_msix].vector, adapter);
503 }
503} 504}
504 505
505int qlcnic_83xx_setup_mbx_intr(struct qlcnic_adapter *adapter) 506int qlcnic_83xx_setup_mbx_intr(struct qlcnic_adapter *adapter)
@@ -760,6 +761,9 @@ int qlcnic_83xx_issue_cmd(struct qlcnic_adapter *adapter,
760 int cmd_type, err, opcode; 761 int cmd_type, err, opcode;
761 unsigned long timeout; 762 unsigned long timeout;
762 763
764 if (!mbx)
765 return -EIO;
766
763 opcode = LSW(cmd->req.arg[0]); 767 opcode = LSW(cmd->req.arg[0]);
764 cmd_type = cmd->type; 768 cmd_type = cmd->type;
765 err = mbx->ops->enqueue_cmd(adapter, cmd, &timeout); 769 err = mbx->ops->enqueue_cmd(adapter, cmd, &timeout);
@@ -902,7 +906,7 @@ void __qlcnic_83xx_process_aen(struct qlcnic_adapter *adapter)
902 QLCNIC_MBX_RSP(event[0])); 906 QLCNIC_MBX_RSP(event[0]));
903 break; 907 break;
904 case QLCNIC_MBX_DCBX_CONFIG_CHANGE_EVENT: 908 case QLCNIC_MBX_DCBX_CONFIG_CHANGE_EVENT:
905 qlcnic_dcb_handle_aen(adapter, (void *)&event[1]); 909 qlcnic_dcb_aen_handler(adapter->dcb, (void *)&event[1]);
906 break; 910 break;
907 default: 911 default:
908 dev_dbg(&adapter->pdev->dev, "Unsupported AEN:0x%x.\n", 912 dev_dbg(&adapter->pdev->dev, "Unsupported AEN:0x%x.\n",
@@ -979,14 +983,14 @@ static int qlcnic_83xx_add_rings(struct qlcnic_adapter *adapter)
979 983
980 sds_mbx_size = sizeof(struct qlcnic_sds_mbx); 984 sds_mbx_size = sizeof(struct qlcnic_sds_mbx);
981 context_id = recv_ctx->context_id; 985 context_id = recv_ctx->context_id;
982 num_sds = (adapter->max_sds_rings - QLCNIC_MAX_RING_SETS); 986 num_sds = adapter->drv_sds_rings - QLCNIC_MAX_SDS_RINGS;
983 ahw->hw_ops->alloc_mbx_args(&cmd, adapter, 987 ahw->hw_ops->alloc_mbx_args(&cmd, adapter,
984 QLCNIC_CMD_ADD_RCV_RINGS); 988 QLCNIC_CMD_ADD_RCV_RINGS);
985 cmd.req.arg[1] = 0 | (num_sds << 8) | (context_id << 16); 989 cmd.req.arg[1] = 0 | (num_sds << 8) | (context_id << 16);
986 990
987 /* set up status rings, mbx 2-81 */ 991 /* set up status rings, mbx 2-81 */
988 index = 2; 992 index = 2;
989 for (i = 8; i < adapter->max_sds_rings; i++) { 993 for (i = 8; i < adapter->drv_sds_rings; i++) {
990 memset(&sds_mbx, 0, sds_mbx_size); 994 memset(&sds_mbx, 0, sds_mbx_size);
991 sds = &recv_ctx->sds_rings[i]; 995 sds = &recv_ctx->sds_rings[i];
992 sds->consumer = 0; 996 sds->consumer = 0;
@@ -1021,7 +1025,7 @@ static int qlcnic_83xx_add_rings(struct qlcnic_adapter *adapter)
1021 mbx_out = (struct qlcnic_add_rings_mbx_out *)&cmd.rsp.arg[1]; 1025 mbx_out = (struct qlcnic_add_rings_mbx_out *)&cmd.rsp.arg[1];
1022 index = 0; 1026 index = 0;
1023 /* status descriptor ring */ 1027 /* status descriptor ring */
1024 for (i = 8; i < adapter->max_sds_rings; i++) { 1028 for (i = 8; i < adapter->drv_sds_rings; i++) {
1025 sds = &recv_ctx->sds_rings[i]; 1029 sds = &recv_ctx->sds_rings[i];
1026 sds->crb_sts_consumer = ahw->pci_base0 + 1030 sds->crb_sts_consumer = ahw->pci_base0 +
1027 mbx_out->host_csmr[index]; 1031 mbx_out->host_csmr[index];
@@ -1079,10 +1083,10 @@ int qlcnic_83xx_create_rx_ctx(struct qlcnic_adapter *adapter)
1079 struct qlcnic_hardware_context *ahw = adapter->ahw; 1083 struct qlcnic_hardware_context *ahw = adapter->ahw;
1080 num_rds = adapter->max_rds_rings; 1084 num_rds = adapter->max_rds_rings;
1081 1085
1082 if (adapter->max_sds_rings <= QLCNIC_MAX_RING_SETS) 1086 if (adapter->drv_sds_rings <= QLCNIC_MAX_SDS_RINGS)
1083 num_sds = adapter->max_sds_rings; 1087 num_sds = adapter->drv_sds_rings;
1084 else 1088 else
1085 num_sds = QLCNIC_MAX_RING_SETS; 1089 num_sds = QLCNIC_MAX_SDS_RINGS;
1086 1090
1087 sds_mbx_size = sizeof(struct qlcnic_sds_mbx); 1091 sds_mbx_size = sizeof(struct qlcnic_sds_mbx);
1088 rds_mbx_size = sizeof(struct qlcnic_rds_mbx); 1092 rds_mbx_size = sizeof(struct qlcnic_rds_mbx);
@@ -1183,7 +1187,7 @@ int qlcnic_83xx_create_rx_ctx(struct qlcnic_adapter *adapter)
1183 sds->crb_intr_mask = ahw->pci_base0 + intr_mask; 1187 sds->crb_intr_mask = ahw->pci_base0 + intr_mask;
1184 } 1188 }
1185 1189
1186 if (adapter->max_sds_rings > QLCNIC_MAX_RING_SETS) 1190 if (adapter->drv_sds_rings > QLCNIC_MAX_SDS_RINGS)
1187 err = qlcnic_83xx_add_rings(adapter); 1191 err = qlcnic_83xx_add_rings(adapter);
1188out: 1192out:
1189 qlcnic_free_mbx_args(&cmd); 1193 qlcnic_free_mbx_args(&cmd);
@@ -1239,9 +1243,9 @@ int qlcnic_83xx_create_tx_ctx(struct qlcnic_adapter *adapter,
1239 mbx.size = tx->num_desc; 1243 mbx.size = tx->num_desc;
1240 if (adapter->flags & QLCNIC_MSIX_ENABLED) { 1244 if (adapter->flags & QLCNIC_MSIX_ENABLED) {
1241 if (!(adapter->flags & QLCNIC_TX_INTR_SHARED)) 1245 if (!(adapter->flags & QLCNIC_TX_INTR_SHARED))
1242 msix_vector = adapter->max_sds_rings + ring; 1246 msix_vector = adapter->drv_sds_rings + ring;
1243 else 1247 else
1244 msix_vector = adapter->max_sds_rings - 1; 1248 msix_vector = adapter->drv_sds_rings - 1;
1245 msix_id = ahw->intr_tbl[msix_vector].id; 1249 msix_id = ahw->intr_tbl[msix_vector].id;
1246 } else { 1250 } else {
1247 msix_id = QLCRDX(ahw, QLCNIC_DEF_INT_ID); 1251 msix_id = QLCRDX(ahw, QLCNIC_DEF_INT_ID);
@@ -1264,7 +1268,8 @@ int qlcnic_83xx_create_tx_ctx(struct qlcnic_adapter *adapter,
1264 qlcnic_pf_set_interface_id_create_tx_ctx(adapter, &temp); 1268 qlcnic_pf_set_interface_id_create_tx_ctx(adapter, &temp);
1265 1269
1266 cmd.req.arg[1] = QLCNIC_CAP0_LEGACY_CONTEXT; 1270 cmd.req.arg[1] = QLCNIC_CAP0_LEGACY_CONTEXT;
1267 cmd.req.arg[5] = QLCNIC_MAX_TX_QUEUES | temp; 1271 cmd.req.arg[5] = QLCNIC_SINGLE_RING | temp;
1272
1268 buf = &cmd.req.arg[6]; 1273 buf = &cmd.req.arg[6];
1269 memcpy(buf, &mbx, sizeof(struct qlcnic_tx_mbx)); 1274 memcpy(buf, &mbx, sizeof(struct qlcnic_tx_mbx));
1270 /* send the mailbox command*/ 1275 /* send the mailbox command*/
@@ -1279,7 +1284,7 @@ int qlcnic_83xx_create_tx_ctx(struct qlcnic_adapter *adapter,
1279 tx->ctx_id = mbx_out->ctx_id; 1284 tx->ctx_id = mbx_out->ctx_id;
1280 if ((adapter->flags & QLCNIC_MSIX_ENABLED) && 1285 if ((adapter->flags & QLCNIC_MSIX_ENABLED) &&
1281 !(adapter->flags & QLCNIC_TX_INTR_SHARED)) { 1286 !(adapter->flags & QLCNIC_TX_INTR_SHARED)) {
1282 intr_mask = ahw->intr_tbl[adapter->max_sds_rings + ring].src; 1287 intr_mask = ahw->intr_tbl[adapter->drv_sds_rings + ring].src;
1283 tx->crb_intr_mask = ahw->pci_base0 + intr_mask; 1288 tx->crb_intr_mask = ahw->pci_base0 + intr_mask;
1284 } 1289 }
1285 dev_info(&adapter->pdev->dev, "Tx Context[0x%x] Created, state:0x%x\n", 1290 dev_info(&adapter->pdev->dev, "Tx Context[0x%x] Created, state:0x%x\n",
@@ -1290,7 +1295,7 @@ out:
1290} 1295}
1291 1296
1292static int qlcnic_83xx_diag_alloc_res(struct net_device *netdev, int test, 1297static int qlcnic_83xx_diag_alloc_res(struct net_device *netdev, int test,
1293 int num_sds_ring) 1298 u8 num_sds_ring)
1294{ 1299{
1295 struct qlcnic_adapter *adapter = netdev_priv(netdev); 1300 struct qlcnic_adapter *adapter = netdev_priv(netdev);
1296 struct qlcnic_host_sds_ring *sds_ring; 1301 struct qlcnic_host_sds_ring *sds_ring;
@@ -1306,7 +1311,7 @@ static int qlcnic_83xx_diag_alloc_res(struct net_device *netdev, int test,
1306 1311
1307 qlcnic_detach(adapter); 1312 qlcnic_detach(adapter);
1308 1313
1309 adapter->max_sds_rings = 1; 1314 adapter->drv_sds_rings = QLCNIC_SINGLE_RING;
1310 adapter->ahw->diag_test = test; 1315 adapter->ahw->diag_test = test;
1311 adapter->ahw->linkup = 0; 1316 adapter->ahw->linkup = 0;
1312 1317
@@ -1320,7 +1325,7 @@ static int qlcnic_83xx_diag_alloc_res(struct net_device *netdev, int test,
1320 if (ret) { 1325 if (ret) {
1321 qlcnic_detach(adapter); 1326 qlcnic_detach(adapter);
1322 if (adapter_state == QLCNIC_ADAPTER_UP_MAGIC) { 1327 if (adapter_state == QLCNIC_ADAPTER_UP_MAGIC) {
1323 adapter->max_sds_rings = num_sds_ring; 1328 adapter->drv_sds_rings = num_sds_ring;
1324 qlcnic_attach(adapter); 1329 qlcnic_attach(adapter);
1325 } 1330 }
1326 netif_device_attach(netdev); 1331 netif_device_attach(netdev);
@@ -1333,7 +1338,7 @@ static int qlcnic_83xx_diag_alloc_res(struct net_device *netdev, int test,
1333 } 1338 }
1334 1339
1335 if (adapter->ahw->diag_test == QLCNIC_INTERRUPT_TEST) { 1340 if (adapter->ahw->diag_test == QLCNIC_INTERRUPT_TEST) {
1336 for (ring = 0; ring < adapter->max_sds_rings; ring++) { 1341 for (ring = 0; ring < adapter->drv_sds_rings; ring++) {
1337 sds_ring = &adapter->recv_ctx->sds_rings[ring]; 1342 sds_ring = &adapter->recv_ctx->sds_rings[ring];
1338 qlcnic_83xx_enable_intr(adapter, sds_ring); 1343 qlcnic_83xx_enable_intr(adapter, sds_ring);
1339 } 1344 }
@@ -1354,7 +1359,7 @@ static int qlcnic_83xx_diag_alloc_res(struct net_device *netdev, int test,
1354} 1359}
1355 1360
1356static void qlcnic_83xx_diag_free_res(struct net_device *netdev, 1361static void qlcnic_83xx_diag_free_res(struct net_device *netdev,
1357 int max_sds_rings) 1362 u8 drv_sds_rings)
1358{ 1363{
1359 struct qlcnic_adapter *adapter = netdev_priv(netdev); 1364 struct qlcnic_adapter *adapter = netdev_priv(netdev);
1360 struct qlcnic_host_sds_ring *sds_ring; 1365 struct qlcnic_host_sds_ring *sds_ring;
@@ -1362,7 +1367,7 @@ static void qlcnic_83xx_diag_free_res(struct net_device *netdev,
1362 1367
1363 clear_bit(__QLCNIC_DEV_UP, &adapter->state); 1368 clear_bit(__QLCNIC_DEV_UP, &adapter->state);
1364 if (adapter->ahw->diag_test == QLCNIC_INTERRUPT_TEST) { 1369 if (adapter->ahw->diag_test == QLCNIC_INTERRUPT_TEST) {
1365 for (ring = 0; ring < adapter->max_sds_rings; ring++) { 1370 for (ring = 0; ring < adapter->drv_sds_rings; ring++) {
1366 sds_ring = &adapter->recv_ctx->sds_rings[ring]; 1371 sds_ring = &adapter->recv_ctx->sds_rings[ring];
1367 qlcnic_83xx_disable_intr(adapter, sds_ring); 1372 qlcnic_83xx_disable_intr(adapter, sds_ring);
1368 if (!(adapter->flags & QLCNIC_MSIX_ENABLED)) 1373 if (!(adapter->flags & QLCNIC_MSIX_ENABLED))
@@ -1386,7 +1391,7 @@ static void qlcnic_83xx_diag_free_res(struct net_device *netdev,
1386 } 1391 }
1387 } 1392 }
1388 adapter->ahw->diag_test = 0; 1393 adapter->ahw->diag_test = 0;
1389 adapter->max_sds_rings = max_sds_rings; 1394 adapter->drv_sds_rings = drv_sds_rings;
1390 1395
1391 if (qlcnic_attach(adapter)) 1396 if (qlcnic_attach(adapter))
1392 goto out; 1397 goto out;
@@ -1648,7 +1653,9 @@ int qlcnic_83xx_loopback_test(struct net_device *netdev, u8 mode)
1648{ 1653{
1649 struct qlcnic_adapter *adapter = netdev_priv(netdev); 1654 struct qlcnic_adapter *adapter = netdev_priv(netdev);
1650 struct qlcnic_hardware_context *ahw = adapter->ahw; 1655 struct qlcnic_hardware_context *ahw = adapter->ahw;
1651 int ret = 0, loop = 0, max_sds_rings = adapter->max_sds_rings; 1656 u8 drv_sds_rings = adapter->drv_sds_rings;
1657 u8 drv_tx_rings = adapter->drv_tx_rings;
1658 int ret = 0, loop = 0;
1652 1659
1653 if (ahw->op_mode == QLCNIC_NON_PRIV_FUNC) { 1660 if (ahw->op_mode == QLCNIC_NON_PRIV_FUNC) {
1654 netdev_warn(netdev, 1661 netdev_warn(netdev,
@@ -1670,7 +1677,7 @@ int qlcnic_83xx_loopback_test(struct net_device *netdev, u8 mode)
1670 mode == QLCNIC_ILB_MODE ? "internal" : "external"); 1677 mode == QLCNIC_ILB_MODE ? "internal" : "external");
1671 1678
1672 ret = qlcnic_83xx_diag_alloc_res(netdev, QLCNIC_LOOPBACK_TEST, 1679 ret = qlcnic_83xx_diag_alloc_res(netdev, QLCNIC_LOOPBACK_TEST,
1673 max_sds_rings); 1680 drv_sds_rings);
1674 if (ret) 1681 if (ret)
1675 goto fail_diag_alloc; 1682 goto fail_diag_alloc;
1676 1683
@@ -1708,10 +1715,11 @@ int qlcnic_83xx_loopback_test(struct net_device *netdev, u8 mode)
1708 qlcnic_83xx_clear_lb_mode(adapter, mode); 1715 qlcnic_83xx_clear_lb_mode(adapter, mode);
1709 1716
1710free_diag_res: 1717free_diag_res:
1711 qlcnic_83xx_diag_free_res(netdev, max_sds_rings); 1718 qlcnic_83xx_diag_free_res(netdev, drv_sds_rings);
1712 1719
1713fail_diag_alloc: 1720fail_diag_alloc:
1714 adapter->max_sds_rings = max_sds_rings; 1721 adapter->drv_sds_rings = drv_sds_rings;
1722 adapter->drv_tx_rings = drv_tx_rings;
1715 qlcnic_release_diag_lock(adapter); 1723 qlcnic_release_diag_lock(adapter);
1716 return ret; 1724 return ret;
1717} 1725}
@@ -2321,19 +2329,7 @@ int qlcnic_83xx_get_pci_info(struct qlcnic_adapter *adapter,
2321 i++; 2329 i++;
2322 memcpy(pci_info->mac + sizeof(u32), &cmd.rsp.arg[i], 2); 2330 memcpy(pci_info->mac + sizeof(u32), &cmd.rsp.arg[i], 2);
2323 i = i + 3; 2331 i = i + 3;
2324 if (ahw->op_mode == QLCNIC_MGMT_FUNC)
2325 dev_info(dev, "id = %d active = %d type = %d\n"
2326 "\tport = %d min bw = %d max bw = %d\n"
2327 "\tmac_addr = %pM\n", pci_info->id,
2328 pci_info->active, pci_info->type,
2329 pci_info->default_port,
2330 pci_info->tx_min_bw,
2331 pci_info->tx_max_bw, pci_info->mac);
2332 } 2332 }
2333 if (ahw->op_mode == QLCNIC_MGMT_FUNC)
2334 dev_info(dev, "Max functions = %d, active functions = %d\n",
2335 ahw->max_pci_func, ahw->act_pci_func);
2336
2337 } else { 2333 } else {
2338 dev_err(dev, "Failed to get PCI Info, error = %d\n", err); 2334 dev_err(dev, "Failed to get PCI Info, error = %d\n", err);
2339 err = -EIO; 2335 err = -EIO;
@@ -3061,11 +3057,14 @@ int qlcnic_83xx_get_settings(struct qlcnic_adapter *adapter,
3061 int status = 0; 3057 int status = 0;
3062 struct qlcnic_hardware_context *ahw = adapter->ahw; 3058 struct qlcnic_hardware_context *ahw = adapter->ahw;
3063 3059
3064 /* Get port configuration info */ 3060 if (!test_bit(__QLCNIC_MAINTENANCE_MODE, &adapter->state)) {
3065 status = qlcnic_83xx_get_port_info(adapter); 3061 /* Get port configuration info */
3066 /* Get Link Status related info */ 3062 status = qlcnic_83xx_get_port_info(adapter);
3067 config = qlcnic_83xx_test_link(adapter); 3063 /* Get Link Status related info */
3068 ahw->module_type = QLC_83XX_SFP_MODULE_TYPE(config); 3064 config = qlcnic_83xx_test_link(adapter);
3065 ahw->module_type = QLC_83XX_SFP_MODULE_TYPE(config);
3066 }
3067
3069 /* hard code until there is a way to get it from flash */ 3068 /* hard code until there is a way to get it from flash */
3070 ahw->board_type = QLCNIC_BRDTYPE_83XX_10G; 3069 ahw->board_type = QLCNIC_BRDTYPE_83XX_10G;
3071 3070
@@ -3279,12 +3278,12 @@ int qlcnic_83xx_reg_test(struct qlcnic_adapter *adapter)
3279 return 0; 3278 return 0;
3280} 3279}
3281 3280
3282int qlcnic_83xx_get_regs_len(struct qlcnic_adapter *adapter) 3281inline int qlcnic_83xx_get_regs_len(struct qlcnic_adapter *adapter)
3283{ 3282{
3284 return (ARRAY_SIZE(qlcnic_83xx_ext_reg_tbl) * 3283 return (ARRAY_SIZE(qlcnic_83xx_ext_reg_tbl) *
3285 sizeof(adapter->ahw->ext_reg_tbl)) + 3284 sizeof(*adapter->ahw->ext_reg_tbl)) +
3286 (ARRAY_SIZE(qlcnic_83xx_reg_tbl) + 3285 (ARRAY_SIZE(qlcnic_83xx_reg_tbl) *
3287 sizeof(adapter->ahw->reg_tbl)); 3286 sizeof(*adapter->ahw->reg_tbl));
3288} 3287}
3289 3288
3290int qlcnic_83xx_get_registers(struct qlcnic_adapter *adapter, u32 *regs_buff) 3289int qlcnic_83xx_get_registers(struct qlcnic_adapter *adapter, u32 *regs_buff)
@@ -3305,10 +3304,11 @@ int qlcnic_83xx_interrupt_test(struct net_device *netdev)
3305 struct qlcnic_adapter *adapter = netdev_priv(netdev); 3304 struct qlcnic_adapter *adapter = netdev_priv(netdev);
3306 struct qlcnic_hardware_context *ahw = adapter->ahw; 3305 struct qlcnic_hardware_context *ahw = adapter->ahw;
3307 struct qlcnic_cmd_args cmd; 3306 struct qlcnic_cmd_args cmd;
3307 u8 val, drv_sds_rings = adapter->drv_sds_rings;
3308 u8 drv_tx_rings = adapter->drv_tx_rings;
3308 u32 data; 3309 u32 data;
3309 u16 intrpt_id, id; 3310 u16 intrpt_id, id;
3310 u8 val; 3311 int ret;
3311 int ret, max_sds_rings = adapter->max_sds_rings;
3312 3312
3313 if (test_bit(__QLCNIC_RESETTING, &adapter->state)) { 3313 if (test_bit(__QLCNIC_RESETTING, &adapter->state)) {
3314 netdev_info(netdev, "Device is resetting\n"); 3314 netdev_info(netdev, "Device is resetting\n");
@@ -3321,7 +3321,7 @@ int qlcnic_83xx_interrupt_test(struct net_device *netdev)
3321 } 3321 }
3322 3322
3323 ret = qlcnic_83xx_diag_alloc_res(netdev, QLCNIC_INTERRUPT_TEST, 3323 ret = qlcnic_83xx_diag_alloc_res(netdev, QLCNIC_INTERRUPT_TEST,
3324 max_sds_rings); 3324 drv_sds_rings);
3325 if (ret) 3325 if (ret)
3326 goto fail_diag_irq; 3326 goto fail_diag_irq;
3327 3327
@@ -3358,10 +3358,11 @@ int qlcnic_83xx_interrupt_test(struct net_device *netdev)
3358 3358
3359done: 3359done:
3360 qlcnic_free_mbx_args(&cmd); 3360 qlcnic_free_mbx_args(&cmd);
3361 qlcnic_83xx_diag_free_res(netdev, max_sds_rings); 3361 qlcnic_83xx_diag_free_res(netdev, drv_sds_rings);
3362 3362
3363fail_diag_irq: 3363fail_diag_irq:
3364 adapter->max_sds_rings = max_sds_rings; 3364 adapter->drv_sds_rings = drv_sds_rings;
3365 adapter->drv_tx_rings = drv_tx_rings;
3365 qlcnic_release_diag_lock(adapter); 3366 qlcnic_release_diag_lock(adapter);
3366 return ret; 3367 return ret;
3367} 3368}
@@ -3381,10 +3382,21 @@ void qlcnic_83xx_get_pauseparam(struct qlcnic_adapter *adapter,
3381 } 3382 }
3382 config = ahw->port_config; 3383 config = ahw->port_config;
3383 if (config & QLC_83XX_CFG_STD_PAUSE) { 3384 if (config & QLC_83XX_CFG_STD_PAUSE) {
3384 if (config & QLC_83XX_CFG_STD_TX_PAUSE) 3385 switch (MSW(config)) {
3386 case QLC_83XX_TX_PAUSE:
3387 pause->tx_pause = 1;
3388 break;
3389 case QLC_83XX_RX_PAUSE:
3390 pause->rx_pause = 1;
3391 break;
3392 case QLC_83XX_TX_RX_PAUSE:
3393 default:
3394 /* Backward compatibility for existing
3395 * flash definitions
3396 */
3385 pause->tx_pause = 1; 3397 pause->tx_pause = 1;
3386 if (config & QLC_83XX_CFG_STD_RX_PAUSE)
3387 pause->rx_pause = 1; 3398 pause->rx_pause = 1;
3399 }
3388 } 3400 }
3389 3401
3390 if (QLC_83XX_AUTONEG(config)) 3402 if (QLC_83XX_AUTONEG(config))
@@ -3427,7 +3439,8 @@ int qlcnic_83xx_set_pauseparam(struct qlcnic_adapter *adapter,
3427 ahw->port_config &= ~QLC_83XX_CFG_STD_RX_PAUSE; 3439 ahw->port_config &= ~QLC_83XX_CFG_STD_RX_PAUSE;
3428 ahw->port_config |= QLC_83XX_CFG_STD_TX_PAUSE; 3440 ahw->port_config |= QLC_83XX_CFG_STD_TX_PAUSE;
3429 } else if (!pause->rx_pause && !pause->tx_pause) { 3441 } else if (!pause->rx_pause && !pause->tx_pause) {
3430 ahw->port_config &= ~QLC_83XX_CFG_STD_TX_RX_PAUSE; 3442 ahw->port_config &= ~(QLC_83XX_CFG_STD_TX_RX_PAUSE |
3443 QLC_83XX_CFG_STD_PAUSE);
3431 } 3444 }
3432 status = qlcnic_83xx_set_port_config(adapter); 3445 status = qlcnic_83xx_set_port_config(adapter);
3433 if (status) { 3446 if (status) {
@@ -3503,7 +3516,7 @@ int qlcnic_83xx_resume(struct qlcnic_adapter *adapter)
3503 if (err) 3516 if (err)
3504 return err; 3517 return err;
3505 3518
3506 if (ahw->nic_mode == QLC_83XX_VIRTUAL_NIC_MODE) { 3519 if (ahw->nic_mode == QLCNIC_VNIC_MODE) {
3507 if (ahw->op_mode == QLCNIC_MGMT_FUNC) { 3520 if (ahw->op_mode == QLCNIC_MGMT_FUNC) {
3508 qlcnic_83xx_set_vnic_opmode(adapter); 3521 qlcnic_83xx_set_vnic_opmode(adapter);
3509 } else { 3522 } else {
@@ -3530,6 +3543,9 @@ void qlcnic_83xx_reinit_mbx_work(struct qlcnic_mailbox *mbx)
3530 3543
3531void qlcnic_83xx_free_mailbox(struct qlcnic_mailbox *mbx) 3544void qlcnic_83xx_free_mailbox(struct qlcnic_mailbox *mbx)
3532{ 3545{
3546 if (!mbx)
3547 return;
3548
3533 destroy_workqueue(mbx->work_q); 3549 destroy_workqueue(mbx->work_q);
3534 kfree(mbx); 3550 kfree(mbx);
3535} 3551}
@@ -3650,6 +3666,9 @@ void qlcnic_83xx_detach_mailbox_work(struct qlcnic_adapter *adapter)
3650{ 3666{
3651 struct qlcnic_mailbox *mbx = adapter->ahw->mailbox; 3667 struct qlcnic_mailbox *mbx = adapter->ahw->mailbox;
3652 3668
3669 if (!mbx)
3670 return;
3671
3653 clear_bit(QLC_83XX_MBX_READY, &mbx->status); 3672 clear_bit(QLC_83XX_MBX_READY, &mbx->status);
3654 complete(&mbx->completion); 3673 complete(&mbx->completion);
3655 cancel_work_sync(&mbx->work); 3674 cancel_work_sync(&mbx->work);
diff --git a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.h b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.h
index 533e150503af..4cae6caa6bfa 100644
--- a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.h
+++ b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.h
@@ -61,7 +61,6 @@
61#define QLC_83XX_HOST_SDS_MBX_IDX 8 61#define QLC_83XX_HOST_SDS_MBX_IDX 8
62 62
63#define QLCNIC_HOST_RDS_MBX_IDX 88 63#define QLCNIC_HOST_RDS_MBX_IDX 88
64#define QLCNIC_MAX_RING_SETS 8
65 64
66/* Pause control registers */ 65/* Pause control registers */
67#define QLC_83XX_SRE_SHIM_REG 0x0D200284 66#define QLC_83XX_SRE_SHIM_REG 0x0D200284
@@ -183,8 +182,8 @@ struct qlcnic_rcv_mbx_out {
183 u8 num_pci_func; 182 u8 num_pci_func;
184 u8 state; 183 u8 state;
185#endif 184#endif
186 u32 host_csmr[QLCNIC_MAX_RING_SETS]; 185 u32 host_csmr[QLCNIC_MAX_SDS_RINGS];
187 struct __host_producer_mbx host_prod[QLCNIC_MAX_RING_SETS]; 186 struct __host_producer_mbx host_prod[QLCNIC_MAX_SDS_RINGS];
188} __packed; 187} __packed;
189 188
190struct qlcnic_add_rings_mbx_out { 189struct qlcnic_add_rings_mbx_out {
@@ -197,8 +196,8 @@ struct qlcnic_add_rings_mbx_out {
197 u8 sts_num; 196 u8 sts_num;
198 u8 rcv_num; 197 u8 rcv_num;
199#endif 198#endif
200 u32 host_csmr[QLCNIC_MAX_RING_SETS]; 199 u32 host_csmr[QLCNIC_MAX_SDS_RINGS];
201 struct __host_producer_mbx host_prod[QLCNIC_MAX_RING_SETS]; 200 struct __host_producer_mbx host_prod[QLCNIC_MAX_SDS_RINGS];
202} __packed; 201} __packed;
203 202
204/* Transmit context mailbox inbox registers 203/* Transmit context mailbox inbox registers
@@ -363,6 +362,9 @@ enum qlcnic_83xx_states {
363#define QLC_83XX_LINK_EEE(data) ((data) & BIT_13) 362#define QLC_83XX_LINK_EEE(data) ((data) & BIT_13)
364#define QLC_83XX_DCBX(data) (((data) >> 28) & 7) 363#define QLC_83XX_DCBX(data) (((data) >> 28) & 7)
365#define QLC_83XX_AUTONEG(data) ((data) & BIT_15) 364#define QLC_83XX_AUTONEG(data) ((data) & BIT_15)
365#define QLC_83XX_TX_PAUSE 0x10
366#define QLC_83XX_RX_PAUSE 0x20
367#define QLC_83XX_TX_RX_PAUSE 0x30
366#define QLC_83XX_CFG_STD_PAUSE (1 << 5) 368#define QLC_83XX_CFG_STD_PAUSE (1 << 5)
367#define QLC_83XX_CFG_STD_TX_PAUSE (1 << 20) 369#define QLC_83XX_CFG_STD_TX_PAUSE (1 << 20)
368#define QLC_83XX_CFG_STD_RX_PAUSE (2 << 20) 370#define QLC_83XX_CFG_STD_RX_PAUSE (2 << 20)
@@ -412,8 +414,6 @@ enum qlcnic_83xx_states {
412#define QLC_83XX_GET_VLAN_ALIGN_CAPABILITY(val) (val & 0x4000) 414#define QLC_83XX_GET_VLAN_ALIGN_CAPABILITY(val) (val & 0x4000)
413#define QLC_83XX_GET_FW_LRO_MSS_CAPABILITY(val) (val & 0x20000) 415#define QLC_83XX_GET_FW_LRO_MSS_CAPABILITY(val) (val & 0x20000)
414#define QLC_83XX_ESWITCH_CAPABILITY BIT_23 416#define QLC_83XX_ESWITCH_CAPABILITY BIT_23
415#define QLC_83XX_VIRTUAL_NIC_MODE 0xFF
416#define QLC_83XX_DEFAULT_MODE 0x0
417#define QLC_83XX_SRIOV_MODE 0x1 417#define QLC_83XX_SRIOV_MODE 0x1
418#define QLCNIC_BRDTYPE_83XX_10G 0x0083 418#define QLCNIC_BRDTYPE_83XX_10G 0x0083
419 419
@@ -521,7 +521,7 @@ enum qlc_83xx_ext_regs {
521/* 83xx funcitons */ 521/* 83xx funcitons */
522int qlcnic_83xx_get_fw_version(struct qlcnic_adapter *); 522int qlcnic_83xx_get_fw_version(struct qlcnic_adapter *);
523int qlcnic_83xx_issue_cmd(struct qlcnic_adapter *, struct qlcnic_cmd_args *); 523int qlcnic_83xx_issue_cmd(struct qlcnic_adapter *, struct qlcnic_cmd_args *);
524int qlcnic_83xx_setup_intr(struct qlcnic_adapter *, u8, int); 524int qlcnic_83xx_setup_intr(struct qlcnic_adapter *);
525void qlcnic_83xx_get_func_no(struct qlcnic_adapter *); 525void qlcnic_83xx_get_func_no(struct qlcnic_adapter *);
526int qlcnic_83xx_cam_lock(struct qlcnic_adapter *); 526int qlcnic_83xx_cam_lock(struct qlcnic_adapter *);
527void qlcnic_83xx_cam_unlock(struct qlcnic_adapter *); 527void qlcnic_83xx_cam_unlock(struct qlcnic_adapter *);
@@ -626,7 +626,7 @@ int qlcnic_83xx_config_vnic_opmode(struct qlcnic_adapter *);
626int qlcnic_83xx_get_vnic_vport_info(struct qlcnic_adapter *, 626int qlcnic_83xx_get_vnic_vport_info(struct qlcnic_adapter *,
627 struct qlcnic_info *, u8); 627 struct qlcnic_info *, u8);
628int qlcnic_83xx_get_vnic_pf_info(struct qlcnic_adapter *, struct qlcnic_info *); 628int qlcnic_83xx_get_vnic_pf_info(struct qlcnic_adapter *, struct qlcnic_info *);
629int qlcnic_83xx_enable_port_eswitch(struct qlcnic_adapter *, int); 629int qlcnic_83xx_set_port_eswitch_status(struct qlcnic_adapter *, int, int *);
630 630
631void qlcnic_83xx_get_minidump_template(struct qlcnic_adapter *); 631void qlcnic_83xx_get_minidump_template(struct qlcnic_adapter *);
632void qlcnic_83xx_get_stats(struct qlcnic_adapter *adapter, u64 *data); 632void qlcnic_83xx_get_stats(struct qlcnic_adapter *adapter, u64 *data);
diff --git a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c
index f09e787af0b2..89208e5b25d6 100644
--- a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c
+++ b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c
@@ -636,7 +636,7 @@ int qlcnic_83xx_idc_reattach_driver(struct qlcnic_adapter *adapter)
636 if (adapter->portnum == 0) 636 if (adapter->portnum == 0)
637 qlcnic_set_drv_version(adapter); 637 qlcnic_set_drv_version(adapter);
638 638
639 qlcnic_dcb_get_info(adapter); 639 qlcnic_dcb_get_info(adapter->dcb);
640 qlcnic_83xx_idc_attach_driver(adapter); 640 qlcnic_83xx_idc_attach_driver(adapter);
641 641
642 return 0; 642 return 0;
@@ -818,6 +818,7 @@ static int qlcnic_83xx_idc_ready_state(struct qlcnic_adapter *adapter)
818 struct qlcnic_hardware_context *ahw = adapter->ahw; 818 struct qlcnic_hardware_context *ahw = adapter->ahw;
819 struct qlcnic_mailbox *mbx = ahw->mailbox; 819 struct qlcnic_mailbox *mbx = ahw->mailbox;
820 int ret = 0; 820 int ret = 0;
821 u32 owner;
821 u32 val; 822 u32 val;
822 823
823 /* Perform NIC configuration based ready state entry actions */ 824 /* Perform NIC configuration based ready state entry actions */
@@ -846,6 +847,10 @@ static int qlcnic_83xx_idc_ready_state(struct qlcnic_adapter *adapter)
846 clear_bit(QLC_83XX_MBX_READY, &mbx->status); 847 clear_bit(QLC_83XX_MBX_READY, &mbx->status);
847 set_bit(__QLCNIC_RESETTING, &adapter->state); 848 set_bit(__QLCNIC_RESETTING, &adapter->state);
848 qlcnic_83xx_idc_enter_need_reset_state(adapter, 1); 849 qlcnic_83xx_idc_enter_need_reset_state(adapter, 1);
850 } else {
851 owner = qlcnic_83xx_idc_find_reset_owner_id(adapter);
852 if (ahw->pci_func == owner)
853 qlcnic_dump_fw(adapter);
849 } 854 }
850 return -EIO; 855 return -EIO;
851 } 856 }
@@ -897,7 +902,7 @@ static int qlcnic_83xx_idc_need_reset_state(struct qlcnic_adapter *adapter)
897 qlcnic_83xx_idc_update_audit_reg(adapter, 0, 1); 902 qlcnic_83xx_idc_update_audit_reg(adapter, 0, 1);
898 set_bit(__QLCNIC_RESETTING, &adapter->state); 903 set_bit(__QLCNIC_RESETTING, &adapter->state);
899 clear_bit(QLC_83XX_MBX_READY, &mbx->status); 904 clear_bit(QLC_83XX_MBX_READY, &mbx->status);
900 if (adapter->ahw->nic_mode == QLC_83XX_VIRTUAL_NIC_MODE) 905 if (adapter->ahw->nic_mode == QLCNIC_VNIC_MODE)
901 qlcnic_83xx_disable_vnic_mode(adapter, 1); 906 qlcnic_83xx_disable_vnic_mode(adapter, 1);
902 907
903 if (qlcnic_check_diag_status(adapter)) { 908 if (qlcnic_check_diag_status(adapter)) {
@@ -1058,6 +1063,12 @@ void qlcnic_83xx_idc_poll_dev_state(struct work_struct *work)
1058 adapter->ahw->idc.prev_state = adapter->ahw->idc.curr_state; 1063 adapter->ahw->idc.prev_state = adapter->ahw->idc.curr_state;
1059 qlcnic_83xx_periodic_tasks(adapter); 1064 qlcnic_83xx_periodic_tasks(adapter);
1060 1065
1066 /* Do not reschedule if firmaware is in hanged state and auto
1067 * recovery is disabled
1068 */
1069 if ((adapter->flags & QLCNIC_FW_HANG) && !qlcnic_auto_fw_reset)
1070 return;
1071
1061 /* Re-schedule the function */ 1072 /* Re-schedule the function */
1062 if (test_bit(QLC_83XX_MODULE_LOADED, &adapter->ahw->idc.status)) 1073 if (test_bit(QLC_83XX_MODULE_LOADED, &adapter->ahw->idc.status))
1063 qlcnic_schedule_work(adapter, qlcnic_83xx_idc_poll_dev_state, 1074 qlcnic_schedule_work(adapter, qlcnic_83xx_idc_poll_dev_state,
@@ -2022,6 +2033,8 @@ int qlcnic_83xx_get_nic_configuration(struct qlcnic_adapter *adapter)
2022 ahw->max_mac_filters = nic_info.max_mac_filters; 2033 ahw->max_mac_filters = nic_info.max_mac_filters;
2023 ahw->max_mtu = nic_info.max_mtu; 2034 ahw->max_mtu = nic_info.max_mtu;
2024 2035
2036 adapter->max_tx_rings = ahw->max_tx_ques;
2037 adapter->max_sds_rings = ahw->max_rx_ques;
2025 /* eSwitch capability indicates vNIC mode. 2038 /* eSwitch capability indicates vNIC mode.
2026 * vNIC and SRIOV are mutually exclusive operational modes. 2039 * vNIC and SRIOV are mutually exclusive operational modes.
2027 * If SR-IOV capability is detected, SR-IOV physical function 2040 * If SR-IOV capability is detected, SR-IOV physical function
@@ -2034,7 +2047,7 @@ int qlcnic_83xx_get_nic_configuration(struct qlcnic_adapter *adapter)
2034 return QLC_83XX_DEFAULT_OPMODE; 2047 return QLC_83XX_DEFAULT_OPMODE;
2035 2048
2036 if (ahw->capabilities & QLC_83XX_ESWITCH_CAPABILITY) 2049 if (ahw->capabilities & QLC_83XX_ESWITCH_CAPABILITY)
2037 return QLC_83XX_VIRTUAL_NIC_MODE; 2050 return QLCNIC_VNIC_MODE;
2038 2051
2039 return QLC_83XX_DEFAULT_OPMODE; 2052 return QLC_83XX_DEFAULT_OPMODE;
2040} 2053}
@@ -2048,15 +2061,20 @@ int qlcnic_83xx_configure_opmode(struct qlcnic_adapter *adapter)
2048 if (ret == -EIO) 2061 if (ret == -EIO)
2049 return -EIO; 2062 return -EIO;
2050 2063
2051 if (ret == QLC_83XX_VIRTUAL_NIC_MODE) { 2064 if (ret == QLCNIC_VNIC_MODE) {
2052 ahw->nic_mode = QLC_83XX_VIRTUAL_NIC_MODE; 2065 ahw->nic_mode = QLCNIC_VNIC_MODE;
2066
2053 if (qlcnic_83xx_config_vnic_opmode(adapter)) 2067 if (qlcnic_83xx_config_vnic_opmode(adapter))
2054 return -EIO; 2068 return -EIO;
2055 2069
2070 adapter->max_sds_rings = QLCNIC_MAX_VNIC_SDS_RINGS;
2071 adapter->max_tx_rings = QLCNIC_MAX_VNIC_TX_RINGS;
2056 } else if (ret == QLC_83XX_DEFAULT_OPMODE) { 2072 } else if (ret == QLC_83XX_DEFAULT_OPMODE) {
2057 ahw->nic_mode = QLC_83XX_DEFAULT_MODE; 2073 ahw->nic_mode = QLCNIC_DEFAULT_MODE;
2058 adapter->nic_ops->init_driver = qlcnic_83xx_init_default_driver; 2074 adapter->nic_ops->init_driver = qlcnic_83xx_init_default_driver;
2059 ahw->idc.state_entry = qlcnic_83xx_idc_ready_state_entry; 2075 ahw->idc.state_entry = qlcnic_83xx_idc_ready_state_entry;
2076 adapter->max_sds_rings = ahw->max_rx_ques;
2077 adapter->max_tx_rings = ahw->max_tx_ques;
2060 } else { 2078 } else {
2061 return -EIO; 2079 return -EIO;
2062 } 2080 }
@@ -2159,13 +2177,34 @@ static int qlcnic_83xx_get_fw_info(struct qlcnic_adapter *adapter)
2159 return err; 2177 return err;
2160} 2178}
2161 2179
2180static void qlcnic_83xx_init_rings(struct qlcnic_adapter *adapter)
2181{
2182 u8 rx_cnt = QLCNIC_DEF_SDS_RINGS;
2183 u8 tx_cnt = QLCNIC_DEF_TX_RINGS;
2184
2185 adapter->max_tx_rings = QLCNIC_MAX_TX_RINGS;
2186 adapter->max_sds_rings = QLCNIC_MAX_SDS_RINGS;
2187
2188 if (!adapter->ahw->msix_supported) {
2189 rx_cnt = QLCNIC_SINGLE_RING;
2190 tx_cnt = QLCNIC_SINGLE_RING;
2191 }
2192
2193 /* compute and set drv sds rings */
2194 qlcnic_set_tx_ring_count(adapter, tx_cnt);
2195 qlcnic_set_sds_ring_count(adapter, rx_cnt);
2196}
2162 2197
2163int qlcnic_83xx_init(struct qlcnic_adapter *adapter, int pci_using_dac) 2198int qlcnic_83xx_init(struct qlcnic_adapter *adapter, int pci_using_dac)
2164{ 2199{
2165 struct qlcnic_hardware_context *ahw = adapter->ahw; 2200 struct qlcnic_hardware_context *ahw = adapter->ahw;
2201 struct qlcnic_dcb *dcb;
2166 int err = 0; 2202 int err = 0;
2167 2203
2168 ahw->msix_supported = !!qlcnic_use_msi_x; 2204 ahw->msix_supported = !!qlcnic_use_msi_x;
2205
2206 qlcnic_83xx_init_rings(adapter);
2207
2169 err = qlcnic_83xx_init_mailbox_work(adapter); 2208 err = qlcnic_83xx_init_mailbox_work(adapter);
2170 if (err) 2209 if (err)
2171 goto exit; 2210 goto exit;
@@ -2178,22 +2217,26 @@ int qlcnic_83xx_init(struct qlcnic_adapter *adapter, int pci_using_dac)
2178 return err; 2217 return err;
2179 } 2218 }
2180 2219
2220 if (qlcnic_83xx_read_flash_descriptor_table(adapter) ||
2221 qlcnic_83xx_read_flash_mfg_id(adapter)) {
2222 dev_err(&adapter->pdev->dev, "Failed reading flash mfg id\n");
2223 err = -ENOTRECOVERABLE;
2224 goto detach_mbx;
2225 }
2226
2181 err = qlcnic_83xx_check_hw_status(adapter); 2227 err = qlcnic_83xx_check_hw_status(adapter);
2182 if (err) 2228 if (err)
2183 goto detach_mbx; 2229 goto detach_mbx;
2184 2230
2185 if (!qlcnic_83xx_read_flash_descriptor_table(adapter))
2186 qlcnic_83xx_read_flash_mfg_id(adapter);
2187
2188 err = qlcnic_83xx_get_fw_info(adapter); 2231 err = qlcnic_83xx_get_fw_info(adapter);
2189 if (err) 2232 if (err)
2190 goto detach_mbx; 2233 goto detach_mbx;
2191 2234
2192 err = qlcnic_83xx_idc_init(adapter); 2235 err = qlcnic_83xx_idc_init(adapter);
2193 if (err) 2236 if (err)
2194 goto clear_fw_info; 2237 goto detach_mbx;
2195 2238
2196 err = qlcnic_setup_intr(adapter, 0, 0); 2239 err = qlcnic_setup_intr(adapter);
2197 if (err) { 2240 if (err) {
2198 dev_err(&adapter->pdev->dev, "Failed to setup interrupt\n"); 2241 dev_err(&adapter->pdev->dev, "Failed to setup interrupt\n");
2199 goto disable_intr; 2242 goto disable_intr;
@@ -2215,13 +2258,16 @@ int qlcnic_83xx_init(struct qlcnic_adapter *adapter, int pci_using_dac)
2215 if (err) 2258 if (err)
2216 goto disable_mbx_intr; 2259 goto disable_mbx_intr;
2217 2260
2261
2218 /* Perform operating mode specific initialization */ 2262 /* Perform operating mode specific initialization */
2219 err = adapter->nic_ops->init_driver(adapter); 2263 err = adapter->nic_ops->init_driver(adapter);
2220 if (err) 2264 if (err)
2221 goto disable_mbx_intr; 2265 goto disable_mbx_intr;
2222 2266
2223 if (adapter->dcb && qlcnic_dcb_attach(adapter)) 2267 dcb = adapter->dcb;
2224 qlcnic_clear_dcb_ops(adapter); 2268
2269 if (dcb && qlcnic_dcb_attach(dcb))
2270 qlcnic_clear_dcb_ops(dcb);
2225 2271
2226 /* Periodically monitor device status */ 2272 /* Periodically monitor device status */
2227 qlcnic_83xx_idc_poll_dev_state(&adapter->fw_work.work); 2273 qlcnic_83xx_idc_poll_dev_state(&adapter->fw_work.work);
@@ -2233,12 +2279,10 @@ disable_mbx_intr:
2233disable_intr: 2279disable_intr:
2234 qlcnic_teardown_intr(adapter); 2280 qlcnic_teardown_intr(adapter);
2235 2281
2236clear_fw_info:
2237 kfree(ahw->fw_info);
2238
2239detach_mbx: 2282detach_mbx:
2240 qlcnic_83xx_detach_mailbox_work(adapter); 2283 qlcnic_83xx_detach_mailbox_work(adapter);
2241 qlcnic_83xx_free_mailbox(ahw->mailbox); 2284 qlcnic_83xx_free_mailbox(ahw->mailbox);
2285 ahw->mailbox = NULL;
2242exit: 2286exit:
2243 return err; 2287 return err;
2244} 2288}
@@ -2251,7 +2295,7 @@ void qlcnic_83xx_aer_stop_poll_work(struct qlcnic_adapter *adapter)
2251 clear_bit(QLC_83XX_MBX_READY, &idc->status); 2295 clear_bit(QLC_83XX_MBX_READY, &idc->status);
2252 cancel_delayed_work_sync(&adapter->fw_work); 2296 cancel_delayed_work_sync(&adapter->fw_work);
2253 2297
2254 if (ahw->nic_mode == QLC_83XX_VIRTUAL_NIC_MODE) 2298 if (ahw->nic_mode == QLCNIC_VNIC_MODE)
2255 qlcnic_83xx_disable_vnic_mode(adapter, 1); 2299 qlcnic_83xx_disable_vnic_mode(adapter, 1);
2256 2300
2257 qlcnic_83xx_idc_detach_driver(adapter); 2301 qlcnic_83xx_idc_detach_driver(adapter);
diff --git a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_vnic.c b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_vnic.c
index 0248a4c2f5dd..734d28602ac3 100644
--- a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_vnic.c
+++ b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_vnic.c
@@ -94,13 +94,29 @@ qlcnic_83xx_config_vnic_buff_descriptors(struct qlcnic_adapter *adapter)
94 **/ 94 **/
95static int qlcnic_83xx_init_mgmt_vnic(struct qlcnic_adapter *adapter) 95static int qlcnic_83xx_init_mgmt_vnic(struct qlcnic_adapter *adapter)
96{ 96{
97 int err = -EIO; 97 struct qlcnic_hardware_context *ahw = adapter->ahw;
98 struct device *dev = &adapter->pdev->dev;
99 struct qlcnic_npar_info *npar;
100 int i, err = -EIO;
98 101
99 qlcnic_83xx_get_minidump_template(adapter); 102 qlcnic_83xx_get_minidump_template(adapter);
103
100 if (!(adapter->flags & QLCNIC_ADAPTER_INITIALIZED)) { 104 if (!(adapter->flags & QLCNIC_ADAPTER_INITIALIZED)) {
101 if (qlcnic_init_pci_info(adapter)) 105 if (qlcnic_init_pci_info(adapter))
102 return err; 106 return err;
103 107
108 npar = adapter->npars;
109
110 for (i = 0; i < ahw->act_pci_func; i++, npar++) {
111 dev_info(dev, "id:%d active:%d type:%d port:%d min_bw:%d max_bw:%d mac_addr:%pM\n",
112 npar->pci_func, npar->active, npar->type,
113 npar->phy_port, npar->min_bw, npar->max_bw,
114 npar->mac);
115 }
116
117 dev_info(dev, "Max functions = %d, active functions = %d\n",
118 ahw->max_pci_func, ahw->act_pci_func);
119
104 if (qlcnic_83xx_set_vnic_opmode(adapter)) 120 if (qlcnic_83xx_set_vnic_opmode(adapter))
105 return err; 121 return err;
106 122
@@ -115,12 +131,12 @@ static int qlcnic_83xx_init_mgmt_vnic(struct qlcnic_adapter *adapter)
115 return err; 131 return err;
116 132
117 qlcnic_83xx_config_vnic_buff_descriptors(adapter); 133 qlcnic_83xx_config_vnic_buff_descriptors(adapter);
118 adapter->ahw->msix_supported = !!qlcnic_use_msi_x; 134 ahw->msix_supported = qlcnic_use_msi_x ? 1 : 0;
119 adapter->flags |= QLCNIC_ADAPTER_INITIALIZED; 135 adapter->flags |= QLCNIC_ADAPTER_INITIALIZED;
120 qlcnic_83xx_enable_vnic_mode(adapter, 1); 136 qlcnic_83xx_enable_vnic_mode(adapter, 1);
121 137
122 dev_info(&adapter->pdev->dev, "HAL Version: %d, Management function\n", 138 dev_info(dev, "HAL Version: %d, Management function\n",
123 adapter->ahw->fw_hal_version); 139 ahw->fw_hal_version);
124 140
125 return 0; 141 return 0;
126} 142}
@@ -240,8 +256,8 @@ int qlcnic_83xx_check_vnic_state(struct qlcnic_adapter *adapter)
240 return 0; 256 return 0;
241} 257}
242 258
243static int qlcnic_83xx_get_eswitch_port_info(struct qlcnic_adapter *adapter, 259int qlcnic_83xx_set_port_eswitch_status(struct qlcnic_adapter *adapter,
244 int func, int *port_id) 260 int func, int *port_id)
245{ 261{
246 struct qlcnic_info nic_info; 262 struct qlcnic_info nic_info;
247 int err = 0; 263 int err = 0;
@@ -257,23 +273,8 @@ static int qlcnic_83xx_get_eswitch_port_info(struct qlcnic_adapter *adapter,
257 else 273 else
258 err = -EIO; 274 err = -EIO;
259 275
260 return err; 276 if (!err)
261} 277 adapter->eswitch[*port_id].flags |= QLCNIC_SWITCH_ENABLE;
262
263int qlcnic_83xx_enable_port_eswitch(struct qlcnic_adapter *adapter, int func)
264{
265 int id, err = 0;
266
267 err = qlcnic_83xx_get_eswitch_port_info(adapter, func, &id);
268 if (err)
269 return err;
270
271 if (!(adapter->eswitch[id].flags & QLCNIC_SWITCH_ENABLE)) {
272 if (!qlcnic_enable_eswitch(adapter, id, 1))
273 adapter->eswitch[id].flags |= QLCNIC_SWITCH_ENABLE;
274 else
275 err = -EIO;
276 }
277 278
278 return err; 279 return err;
279} 280}
diff --git a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_ctx.c b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_ctx.c
index 86850dd633a1..859cb161fc63 100644
--- a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_ctx.c
+++ b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_ctx.c
@@ -270,7 +270,7 @@ int qlcnic_82xx_fw_cmd_create_rx_ctx(struct qlcnic_adapter *adapter)
270 int err; 270 int err;
271 271
272 nrds_rings = adapter->max_rds_rings; 272 nrds_rings = adapter->max_rds_rings;
273 nsds_rings = adapter->max_sds_rings; 273 nsds_rings = adapter->drv_sds_rings;
274 274
275 rq_size = SIZEOF_HOSTRQ_RX(struct qlcnic_hostrq_rx_ctx, nrds_rings, 275 rq_size = SIZEOF_HOSTRQ_RX(struct qlcnic_hostrq_rx_ctx, nrds_rings,
276 nsds_rings); 276 nsds_rings);
@@ -475,7 +475,7 @@ int qlcnic_82xx_fw_cmd_create_tx_ctx(struct qlcnic_adapter *adapter,
475 475
476 if (qlcnic_check_multi_tx(adapter) && 476 if (qlcnic_check_multi_tx(adapter) &&
477 !adapter->ahw->diag_test) { 477 !adapter->ahw->diag_test) {
478 temp_nsds_rings = adapter->max_sds_rings; 478 temp_nsds_rings = adapter->drv_sds_rings;
479 index = temp_nsds_rings + ring; 479 index = temp_nsds_rings + ring;
480 msix_id = ahw->intr_tbl[index].id; 480 msix_id = ahw->intr_tbl[index].id;
481 prq->msi_index = cpu_to_le16(msix_id); 481 prq->msi_index = cpu_to_le16(msix_id);
@@ -512,7 +512,7 @@ int qlcnic_82xx_fw_cmd_create_tx_ctx(struct qlcnic_adapter *adapter,
512 if (qlcnic_check_multi_tx(adapter) && 512 if (qlcnic_check_multi_tx(adapter) &&
513 !adapter->ahw->diag_test && 513 !adapter->ahw->diag_test &&
514 (adapter->flags & QLCNIC_MSIX_ENABLED)) { 514 (adapter->flags & QLCNIC_MSIX_ENABLED)) {
515 index = adapter->max_sds_rings + ring; 515 index = adapter->drv_sds_rings + ring;
516 intr_mask = ahw->intr_tbl[index].src; 516 intr_mask = ahw->intr_tbl[index].src;
517 tx_ring->crb_intr_mask = ahw->pci_base0 + intr_mask; 517 tx_ring->crb_intr_mask = ahw->pci_base0 + intr_mask;
518 } 518 }
@@ -582,7 +582,7 @@ int qlcnic_alloc_hw_resources(struct qlcnic_adapter *adapter)
582 582
583 recv_ctx = adapter->recv_ctx; 583 recv_ctx = adapter->recv_ctx;
584 584
585 for (ring = 0; ring < adapter->max_drv_tx_rings; ring++) { 585 for (ring = 0; ring < adapter->drv_tx_rings; ring++) {
586 tx_ring = &adapter->tx_ring[ring]; 586 tx_ring = &adapter->tx_ring[ring];
587 ptr = (__le32 *)dma_alloc_coherent(&pdev->dev, sizeof(u32), 587 ptr = (__le32 *)dma_alloc_coherent(&pdev->dev, sizeof(u32),
588 &tx_ring->hw_cons_phys_addr, 588 &tx_ring->hw_cons_phys_addr,
@@ -616,7 +616,7 @@ int qlcnic_alloc_hw_resources(struct qlcnic_adapter *adapter)
616 616
617 } 617 }
618 618
619 for (ring = 0; ring < adapter->max_sds_rings; ring++) { 619 for (ring = 0; ring < adapter->drv_sds_rings; ring++) {
620 sds_ring = &recv_ctx->sds_rings[ring]; 620 sds_ring = &recv_ctx->sds_rings[ring];
621 621
622 addr = dma_alloc_coherent(&adapter->pdev->dev, 622 addr = dma_alloc_coherent(&adapter->pdev->dev,
@@ -664,7 +664,7 @@ int qlcnic_fw_create_ctx(struct qlcnic_adapter *dev)
664 if (err) 664 if (err)
665 goto err_out; 665 goto err_out;
666 666
667 for (ring = 0; ring < dev->max_drv_tx_rings; ring++) { 667 for (ring = 0; ring < dev->drv_tx_rings; ring++) {
668 err = qlcnic_fw_cmd_create_tx_ctx(dev, 668 err = qlcnic_fw_cmd_create_tx_ctx(dev,
669 &dev->tx_ring[ring], 669 &dev->tx_ring[ring],
670 ring); 670 ring);
@@ -703,7 +703,7 @@ void qlcnic_fw_destroy_ctx(struct qlcnic_adapter *adapter)
703 703
704 if (test_and_clear_bit(__QLCNIC_FW_ATTACHED, &adapter->state)) { 704 if (test_and_clear_bit(__QLCNIC_FW_ATTACHED, &adapter->state)) {
705 qlcnic_fw_cmd_del_rx_ctx(adapter); 705 qlcnic_fw_cmd_del_rx_ctx(adapter);
706 for (ring = 0; ring < adapter->max_drv_tx_rings; ring++) 706 for (ring = 0; ring < adapter->drv_tx_rings; ring++)
707 qlcnic_fw_cmd_del_tx_ctx(adapter, 707 qlcnic_fw_cmd_del_tx_ctx(adapter,
708 &adapter->tx_ring[ring]); 708 &adapter->tx_ring[ring]);
709 709
@@ -733,7 +733,7 @@ void qlcnic_free_hw_resources(struct qlcnic_adapter *adapter)
733 733
734 recv_ctx = adapter->recv_ctx; 734 recv_ctx = adapter->recv_ctx;
735 735
736 for (ring = 0; ring < adapter->max_drv_tx_rings; ring++) { 736 for (ring = 0; ring < adapter->drv_tx_rings; ring++) {
737 tx_ring = &adapter->tx_ring[ring]; 737 tx_ring = &adapter->tx_ring[ring];
738 if (tx_ring->hw_consumer != NULL) { 738 if (tx_ring->hw_consumer != NULL) {
739 dma_free_coherent(&adapter->pdev->dev, sizeof(u32), 739 dma_free_coherent(&adapter->pdev->dev, sizeof(u32),
@@ -764,7 +764,7 @@ void qlcnic_free_hw_resources(struct qlcnic_adapter *adapter)
764 } 764 }
765 } 765 }
766 766
767 for (ring = 0; ring < adapter->max_sds_rings; ring++) { 767 for (ring = 0; ring < adapter->drv_sds_rings; ring++) {
768 sds_ring = &recv_ctx->sds_rings[ring]; 768 sds_ring = &recv_ctx->sds_rings[ring];
769 769
770 if (sds_ring->desc_head != NULL) { 770 if (sds_ring->desc_head != NULL) {
@@ -895,6 +895,8 @@ int qlcnic_82xx_get_nic_info(struct qlcnic_adapter *adapter,
895 npar_info->max_rx_ques = le16_to_cpu(nic_info->max_rx_ques); 895 npar_info->max_rx_ques = le16_to_cpu(nic_info->max_rx_ques);
896 npar_info->capabilities = le32_to_cpu(nic_info->capabilities); 896 npar_info->capabilities = le32_to_cpu(nic_info->capabilities);
897 npar_info->max_mtu = le16_to_cpu(nic_info->max_mtu); 897 npar_info->max_mtu = le16_to_cpu(nic_info->max_mtu);
898 adapter->max_tx_rings = npar_info->max_tx_ques;
899 adapter->max_sds_rings = npar_info->max_rx_ques;
898 } 900 }
899 901
900 qlcnic_free_mbx_args(&cmd); 902 qlcnic_free_mbx_args(&cmd);
diff --git a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c
index d62d5ce432ec..86bca7c14f99 100644
--- a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c
+++ b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.c
@@ -57,22 +57,22 @@ static const struct dcbnl_rtnl_ops qlcnic_dcbnl_ops;
57static void qlcnic_dcb_aen_work(struct work_struct *); 57static void qlcnic_dcb_aen_work(struct work_struct *);
58static void qlcnic_dcb_data_cee_param_map(struct qlcnic_adapter *); 58static void qlcnic_dcb_data_cee_param_map(struct qlcnic_adapter *);
59 59
60static inline void __qlcnic_init_dcbnl_ops(struct qlcnic_adapter *); 60static inline void __qlcnic_init_dcbnl_ops(struct qlcnic_dcb *);
61static void __qlcnic_dcb_free(struct qlcnic_adapter *); 61static void __qlcnic_dcb_free(struct qlcnic_dcb *);
62static int __qlcnic_dcb_attach(struct qlcnic_adapter *); 62static int __qlcnic_dcb_attach(struct qlcnic_dcb *);
63static int __qlcnic_dcb_query_hw_capability(struct qlcnic_adapter *, char *); 63static int __qlcnic_dcb_query_hw_capability(struct qlcnic_dcb *, char *);
64static void __qlcnic_dcb_get_info(struct qlcnic_adapter *); 64static void __qlcnic_dcb_get_info(struct qlcnic_dcb *);
65 65
66static int qlcnic_82xx_dcb_get_hw_capability(struct qlcnic_adapter *); 66static int qlcnic_82xx_dcb_get_hw_capability(struct qlcnic_dcb *);
67static int qlcnic_82xx_dcb_query_cee_param(struct qlcnic_adapter *, char *, u8); 67static int qlcnic_82xx_dcb_query_cee_param(struct qlcnic_dcb *, char *, u8);
68static int qlcnic_82xx_dcb_get_cee_cfg(struct qlcnic_adapter *); 68static int qlcnic_82xx_dcb_get_cee_cfg(struct qlcnic_dcb *);
69static void qlcnic_82xx_dcb_handle_aen(struct qlcnic_adapter *, void *); 69static void qlcnic_82xx_dcb_aen_handler(struct qlcnic_dcb *, void *);
70 70
71static int qlcnic_83xx_dcb_get_hw_capability(struct qlcnic_adapter *); 71static int qlcnic_83xx_dcb_get_hw_capability(struct qlcnic_dcb *);
72static int qlcnic_83xx_dcb_query_cee_param(struct qlcnic_adapter *, char *, u8); 72static int qlcnic_83xx_dcb_query_cee_param(struct qlcnic_dcb *, char *, u8);
73static int qlcnic_83xx_dcb_get_cee_cfg(struct qlcnic_adapter *); 73static int qlcnic_83xx_dcb_get_cee_cfg(struct qlcnic_dcb *);
74static int qlcnic_83xx_dcb_register_aen(struct qlcnic_adapter *, bool); 74static int qlcnic_83xx_dcb_register_aen(struct qlcnic_dcb *, bool);
75static void qlcnic_83xx_dcb_handle_aen(struct qlcnic_adapter *, void *); 75static void qlcnic_83xx_dcb_aen_handler(struct qlcnic_dcb *, void *);
76 76
77struct qlcnic_dcb_capability { 77struct qlcnic_dcb_capability {
78 bool tsa_capability; 78 bool tsa_capability;
@@ -180,7 +180,7 @@ static struct qlcnic_dcb_ops qlcnic_83xx_dcb_ops = {
180 .query_cee_param = qlcnic_83xx_dcb_query_cee_param, 180 .query_cee_param = qlcnic_83xx_dcb_query_cee_param,
181 .get_cee_cfg = qlcnic_83xx_dcb_get_cee_cfg, 181 .get_cee_cfg = qlcnic_83xx_dcb_get_cee_cfg,
182 .register_aen = qlcnic_83xx_dcb_register_aen, 182 .register_aen = qlcnic_83xx_dcb_register_aen,
183 .handle_aen = qlcnic_83xx_dcb_handle_aen, 183 .aen_handler = qlcnic_83xx_dcb_aen_handler,
184}; 184};
185 185
186static struct qlcnic_dcb_ops qlcnic_82xx_dcb_ops = { 186static struct qlcnic_dcb_ops qlcnic_82xx_dcb_ops = {
@@ -193,7 +193,7 @@ static struct qlcnic_dcb_ops qlcnic_82xx_dcb_ops = {
193 .get_hw_capability = qlcnic_82xx_dcb_get_hw_capability, 193 .get_hw_capability = qlcnic_82xx_dcb_get_hw_capability,
194 .query_cee_param = qlcnic_82xx_dcb_query_cee_param, 194 .query_cee_param = qlcnic_82xx_dcb_query_cee_param,
195 .get_cee_cfg = qlcnic_82xx_dcb_get_cee_cfg, 195 .get_cee_cfg = qlcnic_82xx_dcb_get_cee_cfg,
196 .handle_aen = qlcnic_82xx_dcb_handle_aen, 196 .aen_handler = qlcnic_82xx_dcb_aen_handler,
197}; 197};
198 198
199static u8 qlcnic_dcb_get_num_app(struct qlcnic_adapter *adapter, u32 val) 199static u8 qlcnic_dcb_get_num_app(struct qlcnic_adapter *adapter, u32 val)
@@ -242,10 +242,10 @@ static int qlcnic_dcb_prio_count(u8 up_tc_map)
242 return j; 242 return j;
243} 243}
244 244
245static inline void __qlcnic_init_dcbnl_ops(struct qlcnic_adapter *adapter) 245static inline void __qlcnic_init_dcbnl_ops(struct qlcnic_dcb *dcb)
246{ 246{
247 if (test_bit(__QLCNIC_DCB_STATE, &adapter->state)) 247 if (test_bit(QLCNIC_DCB_STATE, &dcb->state))
248 adapter->netdev->dcbnl_ops = &qlcnic_dcbnl_ops; 248 dcb->adapter->netdev->dcbnl_ops = &qlcnic_dcbnl_ops;
249} 249}
250 250
251static void qlcnic_set_dcb_ops(struct qlcnic_adapter *adapter) 251static void qlcnic_set_dcb_ops(struct qlcnic_adapter *adapter)
@@ -256,7 +256,7 @@ static void qlcnic_set_dcb_ops(struct qlcnic_adapter *adapter)
256 adapter->dcb->ops = &qlcnic_83xx_dcb_ops; 256 adapter->dcb->ops = &qlcnic_83xx_dcb_ops;
257} 257}
258 258
259int __qlcnic_register_dcb(struct qlcnic_adapter *adapter) 259int qlcnic_register_dcb(struct qlcnic_adapter *adapter)
260{ 260{
261 struct qlcnic_dcb *dcb; 261 struct qlcnic_dcb *dcb;
262 262
@@ -267,20 +267,22 @@ int __qlcnic_register_dcb(struct qlcnic_adapter *adapter)
267 adapter->dcb = dcb; 267 adapter->dcb = dcb;
268 dcb->adapter = adapter; 268 dcb->adapter = adapter;
269 qlcnic_set_dcb_ops(adapter); 269 qlcnic_set_dcb_ops(adapter);
270 dcb->state = 0;
270 271
271 return 0; 272 return 0;
272} 273}
273 274
274static void __qlcnic_dcb_free(struct qlcnic_adapter *adapter) 275static void __qlcnic_dcb_free(struct qlcnic_dcb *dcb)
275{ 276{
276 struct qlcnic_dcb *dcb = adapter->dcb; 277 struct qlcnic_adapter *adapter;
277 278
278 if (!dcb) 279 if (!dcb)
279 return; 280 return;
280 281
281 qlcnic_dcb_register_aen(adapter, 0); 282 adapter = dcb->adapter;
283 qlcnic_dcb_register_aen(dcb, 0);
282 284
283 while (test_bit(__QLCNIC_DCB_IN_AEN, &adapter->state)) 285 while (test_bit(QLCNIC_DCB_AEN_MODE, &dcb->state))
284 usleep_range(10000, 11000); 286 usleep_range(10000, 11000);
285 287
286 cancel_delayed_work_sync(&dcb->aen_work); 288 cancel_delayed_work_sync(&dcb->aen_work);
@@ -298,23 +300,22 @@ static void __qlcnic_dcb_free(struct qlcnic_adapter *adapter)
298 adapter->dcb = NULL; 300 adapter->dcb = NULL;
299} 301}
300 302
301static void __qlcnic_dcb_get_info(struct qlcnic_adapter *adapter) 303static void __qlcnic_dcb_get_info(struct qlcnic_dcb *dcb)
302{ 304{
303 qlcnic_dcb_get_hw_capability(adapter); 305 qlcnic_dcb_get_hw_capability(dcb);
304 qlcnic_dcb_get_cee_cfg(adapter); 306 qlcnic_dcb_get_cee_cfg(dcb);
305 qlcnic_dcb_register_aen(adapter, 1); 307 qlcnic_dcb_register_aen(dcb, 1);
306} 308}
307 309
308static int __qlcnic_dcb_attach(struct qlcnic_adapter *adapter) 310static int __qlcnic_dcb_attach(struct qlcnic_dcb *dcb)
309{ 311{
310 struct qlcnic_dcb *dcb = adapter->dcb;
311 int err = 0; 312 int err = 0;
312 313
313 INIT_DELAYED_WORK(&dcb->aen_work, qlcnic_dcb_aen_work); 314 INIT_DELAYED_WORK(&dcb->aen_work, qlcnic_dcb_aen_work);
314 315
315 dcb->wq = create_singlethread_workqueue("qlcnic-dcb"); 316 dcb->wq = create_singlethread_workqueue("qlcnic-dcb");
316 if (!dcb->wq) { 317 if (!dcb->wq) {
317 dev_err(&adapter->pdev->dev, 318 dev_err(&dcb->adapter->pdev->dev,
318 "DCB workqueue allocation failed. DCB will be disabled\n"); 319 "DCB workqueue allocation failed. DCB will be disabled\n");
319 return -1; 320 return -1;
320 } 321 }
@@ -331,7 +332,7 @@ static int __qlcnic_dcb_attach(struct qlcnic_adapter *adapter)
331 goto out_free_cfg; 332 goto out_free_cfg;
332 } 333 }
333 334
334 qlcnic_dcb_get_info(adapter); 335 qlcnic_dcb_get_info(dcb);
335 336
336 return 0; 337 return 0;
337out_free_cfg: 338out_free_cfg:
@@ -345,9 +346,9 @@ out_free_wq:
345 return err; 346 return err;
346} 347}
347 348
348static int __qlcnic_dcb_query_hw_capability(struct qlcnic_adapter *adapter, 349static int __qlcnic_dcb_query_hw_capability(struct qlcnic_dcb *dcb, char *buf)
349 char *buf)
350{ 350{
351 struct qlcnic_adapter *adapter = dcb->adapter;
351 struct qlcnic_cmd_args cmd; 352 struct qlcnic_cmd_args cmd;
352 u32 mbx_out; 353 u32 mbx_out;
353 int err; 354 int err;
@@ -371,15 +372,15 @@ static int __qlcnic_dcb_query_hw_capability(struct qlcnic_adapter *adapter,
371 return err; 372 return err;
372} 373}
373 374
374static int __qlcnic_dcb_get_capability(struct qlcnic_adapter *adapter, u32 *val) 375static int __qlcnic_dcb_get_capability(struct qlcnic_dcb *dcb, u32 *val)
375{ 376{
376 struct qlcnic_dcb_capability *cap = &adapter->dcb->cfg->capability; 377 struct qlcnic_dcb_capability *cap = &dcb->cfg->capability;
377 u32 mbx_out; 378 u32 mbx_out;
378 int err; 379 int err;
379 380
380 memset(cap, 0, sizeof(struct qlcnic_dcb_capability)); 381 memset(cap, 0, sizeof(struct qlcnic_dcb_capability));
381 382
382 err = qlcnic_dcb_query_hw_capability(adapter, (char *)val); 383 err = qlcnic_dcb_query_hw_capability(dcb, (char *)val);
383 if (err) 384 if (err)
384 return err; 385 return err;
385 386
@@ -397,21 +398,21 @@ static int __qlcnic_dcb_get_capability(struct qlcnic_adapter *adapter, u32 *val)
397 if (cap->max_num_tc > QLC_DCB_MAX_TC || 398 if (cap->max_num_tc > QLC_DCB_MAX_TC ||
398 cap->max_ets_tc > cap->max_num_tc || 399 cap->max_ets_tc > cap->max_num_tc ||
399 cap->max_pfc_tc > cap->max_num_tc) { 400 cap->max_pfc_tc > cap->max_num_tc) {
400 dev_err(&adapter->pdev->dev, "Invalid DCB configuration\n"); 401 dev_err(&dcb->adapter->pdev->dev, "Invalid DCB configuration\n");
401 return -EINVAL; 402 return -EINVAL;
402 } 403 }
403 404
404 return err; 405 return err;
405} 406}
406 407
407static int qlcnic_82xx_dcb_get_hw_capability(struct qlcnic_adapter *adapter) 408static int qlcnic_82xx_dcb_get_hw_capability(struct qlcnic_dcb *dcb)
408{ 409{
409 struct qlcnic_dcb_cfg *cfg = adapter->dcb->cfg; 410 struct qlcnic_dcb_cfg *cfg = dcb->cfg;
410 struct qlcnic_dcb_capability *cap; 411 struct qlcnic_dcb_capability *cap;
411 u32 mbx_out; 412 u32 mbx_out;
412 int err; 413 int err;
413 414
414 err = __qlcnic_dcb_get_capability(adapter, &mbx_out); 415 err = __qlcnic_dcb_get_capability(dcb, &mbx_out);
415 if (err) 416 if (err)
416 return err; 417 return err;
417 418
@@ -419,15 +420,16 @@ static int qlcnic_82xx_dcb_get_hw_capability(struct qlcnic_adapter *adapter)
419 cap->dcb_capability = DCB_CAP_DCBX_VER_CEE | DCB_CAP_DCBX_LLD_MANAGED; 420 cap->dcb_capability = DCB_CAP_DCBX_VER_CEE | DCB_CAP_DCBX_LLD_MANAGED;
420 421
421 if (cap->dcb_capability && cap->tsa_capability && cap->ets_capability) 422 if (cap->dcb_capability && cap->tsa_capability && cap->ets_capability)
422 set_bit(__QLCNIC_DCB_STATE, &adapter->state); 423 set_bit(QLCNIC_DCB_STATE, &dcb->state);
423 424
424 return err; 425 return err;
425} 426}
426 427
427static int qlcnic_82xx_dcb_query_cee_param(struct qlcnic_adapter *adapter, 428static int qlcnic_82xx_dcb_query_cee_param(struct qlcnic_dcb *dcb,
428 char *buf, u8 type) 429 char *buf, u8 type)
429{ 430{
430 u16 size = sizeof(struct qlcnic_82xx_dcb_param_mbx_le); 431 u16 size = sizeof(struct qlcnic_82xx_dcb_param_mbx_le);
432 struct qlcnic_adapter *adapter = dcb->adapter;
431 struct qlcnic_82xx_dcb_param_mbx_le *prsp_le; 433 struct qlcnic_82xx_dcb_param_mbx_le *prsp_le;
432 struct device *dev = &adapter->pdev->dev; 434 struct device *dev = &adapter->pdev->dev;
433 dma_addr_t cardrsp_phys_addr; 435 dma_addr_t cardrsp_phys_addr;
@@ -447,8 +449,7 @@ static int qlcnic_82xx_dcb_query_cee_param(struct qlcnic_adapter *adapter,
447 return -EINVAL; 449 return -EINVAL;
448 } 450 }
449 451
450 addr = dma_alloc_coherent(&adapter->pdev->dev, size, &cardrsp_phys_addr, 452 addr = dma_alloc_coherent(dev, size, &cardrsp_phys_addr, GFP_KERNEL);
451 GFP_KERNEL);
452 if (addr == NULL) 453 if (addr == NULL)
453 return -ENOMEM; 454 return -ENOMEM;
454 455
@@ -488,72 +489,67 @@ out:
488 qlcnic_free_mbx_args(&cmd); 489 qlcnic_free_mbx_args(&cmd);
489 490
490out_free_rsp: 491out_free_rsp:
491 dma_free_coherent(&adapter->pdev->dev, size, addr, cardrsp_phys_addr); 492 dma_free_coherent(dev, size, addr, cardrsp_phys_addr);
492 493
493 return err; 494 return err;
494} 495}
495 496
496static int qlcnic_82xx_dcb_get_cee_cfg(struct qlcnic_adapter *adapter) 497static int qlcnic_82xx_dcb_get_cee_cfg(struct qlcnic_dcb *dcb)
497{ 498{
498 struct qlcnic_dcb_mbx_params *mbx; 499 struct qlcnic_dcb_mbx_params *mbx;
499 int err; 500 int err;
500 501
501 mbx = adapter->dcb->param; 502 mbx = dcb->param;
502 if (!mbx) 503 if (!mbx)
503 return 0; 504 return 0;
504 505
505 err = qlcnic_dcb_query_cee_param(adapter, (char *)&mbx->type[0], 506 err = qlcnic_dcb_query_cee_param(dcb, (char *)&mbx->type[0],
506 QLC_DCB_LOCAL_PARAM_FWID); 507 QLC_DCB_LOCAL_PARAM_FWID);
507 if (err) 508 if (err)
508 return err; 509 return err;
509 510
510 err = qlcnic_dcb_query_cee_param(adapter, (char *)&mbx->type[1], 511 err = qlcnic_dcb_query_cee_param(dcb, (char *)&mbx->type[1],
511 QLC_DCB_OPER_PARAM_FWID); 512 QLC_DCB_OPER_PARAM_FWID);
512 if (err) 513 if (err)
513 return err; 514 return err;
514 515
515 err = qlcnic_dcb_query_cee_param(adapter, (char *)&mbx->type[2], 516 err = qlcnic_dcb_query_cee_param(dcb, (char *)&mbx->type[2],
516 QLC_DCB_PEER_PARAM_FWID); 517 QLC_DCB_PEER_PARAM_FWID);
517 if (err) 518 if (err)
518 return err; 519 return err;
519 520
520 mbx->prio_tc_map = QLC_82XX_DCB_PRIO_TC_MAP; 521 mbx->prio_tc_map = QLC_82XX_DCB_PRIO_TC_MAP;
521 522
522 qlcnic_dcb_data_cee_param_map(adapter); 523 qlcnic_dcb_data_cee_param_map(dcb->adapter);
523 524
524 return err; 525 return err;
525} 526}
526 527
527static void qlcnic_dcb_aen_work(struct work_struct *work) 528static void qlcnic_dcb_aen_work(struct work_struct *work)
528{ 529{
529 struct qlcnic_adapter *adapter;
530 struct qlcnic_dcb *dcb; 530 struct qlcnic_dcb *dcb;
531 531
532 dcb = container_of(work, struct qlcnic_dcb, aen_work.work); 532 dcb = container_of(work, struct qlcnic_dcb, aen_work.work);
533 adapter = dcb->adapter;
534 533
535 qlcnic_dcb_get_cee_cfg(adapter); 534 qlcnic_dcb_get_cee_cfg(dcb);
536 clear_bit(__QLCNIC_DCB_IN_AEN, &adapter->state); 535 clear_bit(QLCNIC_DCB_AEN_MODE, &dcb->state);
537} 536}
538 537
539static void qlcnic_82xx_dcb_handle_aen(struct qlcnic_adapter *adapter, 538static void qlcnic_82xx_dcb_aen_handler(struct qlcnic_dcb *dcb, void *data)
540 void *data)
541{ 539{
542 struct qlcnic_dcb *dcb = adapter->dcb; 540 if (test_and_set_bit(QLCNIC_DCB_AEN_MODE, &dcb->state))
543
544 if (test_and_set_bit(__QLCNIC_DCB_IN_AEN, &adapter->state))
545 return; 541 return;
546 542
547 queue_delayed_work(dcb->wq, &dcb->aen_work, 0); 543 queue_delayed_work(dcb->wq, &dcb->aen_work, 0);
548} 544}
549 545
550static int qlcnic_83xx_dcb_get_hw_capability(struct qlcnic_adapter *adapter) 546static int qlcnic_83xx_dcb_get_hw_capability(struct qlcnic_dcb *dcb)
551{ 547{
552 struct qlcnic_dcb_capability *cap = &adapter->dcb->cfg->capability; 548 struct qlcnic_dcb_capability *cap = &dcb->cfg->capability;
553 u32 mbx_out; 549 u32 mbx_out;
554 int err; 550 int err;
555 551
556 err = __qlcnic_dcb_get_capability(adapter, &mbx_out); 552 err = __qlcnic_dcb_get_capability(dcb, &mbx_out);
557 if (err) 553 if (err)
558 return err; 554 return err;
559 555
@@ -565,14 +561,15 @@ static int qlcnic_83xx_dcb_get_hw_capability(struct qlcnic_adapter *adapter)
565 cap->dcb_capability |= DCB_CAP_DCBX_LLD_MANAGED; 561 cap->dcb_capability |= DCB_CAP_DCBX_LLD_MANAGED;
566 562
567 if (cap->dcb_capability && cap->tsa_capability && cap->ets_capability) 563 if (cap->dcb_capability && cap->tsa_capability && cap->ets_capability)
568 set_bit(__QLCNIC_DCB_STATE, &adapter->state); 564 set_bit(QLCNIC_DCB_STATE, &dcb->state);
569 565
570 return err; 566 return err;
571} 567}
572 568
573static int qlcnic_83xx_dcb_query_cee_param(struct qlcnic_adapter *adapter, 569static int qlcnic_83xx_dcb_query_cee_param(struct qlcnic_dcb *dcb,
574 char *buf, u8 idx) 570 char *buf, u8 idx)
575{ 571{
572 struct qlcnic_adapter *adapter = dcb->adapter;
576 struct qlcnic_dcb_mbx_params mbx_out; 573 struct qlcnic_dcb_mbx_params mbx_out;
577 int err, i, j, k, max_app, size; 574 int err, i, j, k, max_app, size;
578 struct qlcnic_dcb_param *each; 575 struct qlcnic_dcb_param *each;
@@ -632,24 +629,23 @@ out:
632 return err; 629 return err;
633} 630}
634 631
635static int qlcnic_83xx_dcb_get_cee_cfg(struct qlcnic_adapter *adapter) 632static int qlcnic_83xx_dcb_get_cee_cfg(struct qlcnic_dcb *dcb)
636{ 633{
637 struct qlcnic_dcb *dcb = adapter->dcb;
638 int err; 634 int err;
639 635
640 err = qlcnic_dcb_query_cee_param(adapter, (char *)dcb->param, 0); 636 err = qlcnic_dcb_query_cee_param(dcb, (char *)dcb->param, 0);
641 if (err) 637 if (err)
642 return err; 638 return err;
643 639
644 qlcnic_dcb_data_cee_param_map(adapter); 640 qlcnic_dcb_data_cee_param_map(dcb->adapter);
645 641
646 return err; 642 return err;
647} 643}
648 644
649static int qlcnic_83xx_dcb_register_aen(struct qlcnic_adapter *adapter, 645static int qlcnic_83xx_dcb_register_aen(struct qlcnic_dcb *dcb, bool flag)
650 bool flag)
651{ 646{
652 u8 val = (flag ? QLCNIC_CMD_INIT_NIC_FUNC : QLCNIC_CMD_STOP_NIC_FUNC); 647 u8 val = (flag ? QLCNIC_CMD_INIT_NIC_FUNC : QLCNIC_CMD_STOP_NIC_FUNC);
648 struct qlcnic_adapter *adapter = dcb->adapter;
653 struct qlcnic_cmd_args cmd; 649 struct qlcnic_cmd_args cmd;
654 int err; 650 int err;
655 651
@@ -669,19 +665,17 @@ static int qlcnic_83xx_dcb_register_aen(struct qlcnic_adapter *adapter,
669 return err; 665 return err;
670} 666}
671 667
672static void qlcnic_83xx_dcb_handle_aen(struct qlcnic_adapter *adapter, 668static void qlcnic_83xx_dcb_aen_handler(struct qlcnic_dcb *dcb, void *data)
673 void *data)
674{ 669{
675 struct qlcnic_dcb *dcb = adapter->dcb;
676 u32 *val = data; 670 u32 *val = data;
677 671
678 if (test_and_set_bit(__QLCNIC_DCB_IN_AEN, &adapter->state)) 672 if (test_and_set_bit(QLCNIC_DCB_AEN_MODE, &dcb->state))
679 return; 673 return;
680 674
681 if (*val & BIT_8) 675 if (*val & BIT_8)
682 set_bit(__QLCNIC_DCB_STATE, &adapter->state); 676 set_bit(QLCNIC_DCB_STATE, &dcb->state);
683 else 677 else
684 clear_bit(__QLCNIC_DCB_STATE, &adapter->state); 678 clear_bit(QLCNIC_DCB_STATE, &dcb->state);
685 679
686 queue_delayed_work(dcb->wq, &dcb->aen_work, 0); 680 queue_delayed_work(dcb->wq, &dcb->aen_work, 0);
687} 681}
@@ -814,12 +808,12 @@ static u8 qlcnic_dcb_get_state(struct net_device *netdev)
814{ 808{
815 struct qlcnic_adapter *adapter = netdev_priv(netdev); 809 struct qlcnic_adapter *adapter = netdev_priv(netdev);
816 810
817 return test_bit(__QLCNIC_DCB_STATE, &adapter->state); 811 return test_bit(QLCNIC_DCB_STATE, &adapter->dcb->state);
818} 812}
819 813
820static void qlcnic_dcb_get_perm_hw_addr(struct net_device *netdev, u8 *addr) 814static void qlcnic_dcb_get_perm_hw_addr(struct net_device *netdev, u8 *addr)
821{ 815{
822 memcpy(addr, netdev->dev_addr, netdev->addr_len); 816 memcpy(addr, netdev->perm_addr, netdev->addr_len);
823} 817}
824 818
825static void 819static void
@@ -834,7 +828,7 @@ qlcnic_dcb_get_pg_tc_cfg_tx(struct net_device *netdev, int tc, u8 *prio,
834 type = &adapter->dcb->cfg->type[QLC_DCB_OPER_IDX]; 828 type = &adapter->dcb->cfg->type[QLC_DCB_OPER_IDX];
835 *prio = *pgid = *bw_per = *up_tc_map = 0; 829 *prio = *pgid = *bw_per = *up_tc_map = 0;
836 830
837 if (!test_bit(__QLCNIC_DCB_STATE, &adapter->state) || 831 if (!test_bit(QLCNIC_DCB_STATE, &adapter->dcb->state) ||
838 !type->tc_param_valid) 832 !type->tc_param_valid)
839 return; 833 return;
840 834
@@ -870,7 +864,7 @@ static void qlcnic_dcb_get_pg_bwg_cfg_tx(struct net_device *netdev, int pgid,
870 *bw_pct = 0; 864 *bw_pct = 0;
871 type = &adapter->dcb->cfg->type[QLC_DCB_OPER_IDX]; 865 type = &adapter->dcb->cfg->type[QLC_DCB_OPER_IDX];
872 866
873 if (!test_bit(__QLCNIC_DCB_STATE, &adapter->state) || 867 if (!test_bit(QLCNIC_DCB_STATE, &adapter->dcb->state) ||
874 !type->tc_param_valid) 868 !type->tc_param_valid)
875 return; 869 return;
876 870
@@ -896,7 +890,7 @@ static void qlcnic_dcb_get_pfc_cfg(struct net_device *netdev, int prio,
896 *setting = 0; 890 *setting = 0;
897 type = &adapter->dcb->cfg->type[QLC_DCB_OPER_IDX]; 891 type = &adapter->dcb->cfg->type[QLC_DCB_OPER_IDX];
898 892
899 if (!test_bit(__QLCNIC_DCB_STATE, &adapter->state) || 893 if (!test_bit(QLCNIC_DCB_STATE, &adapter->dcb->state) ||
900 !type->pfc_mode_enable) 894 !type->pfc_mode_enable)
901 return; 895 return;
902 896
@@ -915,7 +909,7 @@ static u8 qlcnic_dcb_get_capability(struct net_device *netdev, int capid,
915{ 909{
916 struct qlcnic_adapter *adapter = netdev_priv(netdev); 910 struct qlcnic_adapter *adapter = netdev_priv(netdev);
917 911
918 if (!test_bit(__QLCNIC_DCB_STATE, &adapter->state)) 912 if (!test_bit(QLCNIC_DCB_STATE, &adapter->dcb->state))
919 return 0; 913 return 0;
920 914
921 switch (capid) { 915 switch (capid) {
@@ -944,7 +938,7 @@ static int qlcnic_dcb_get_num_tcs(struct net_device *netdev, int attr, u8 *num)
944 struct qlcnic_adapter *adapter = netdev_priv(netdev); 938 struct qlcnic_adapter *adapter = netdev_priv(netdev);
945 struct qlcnic_dcb_cfg *cfg = adapter->dcb->cfg; 939 struct qlcnic_dcb_cfg *cfg = adapter->dcb->cfg;
946 940
947 if (!test_bit(__QLCNIC_DCB_STATE, &adapter->state)) 941 if (!test_bit(QLCNIC_DCB_STATE, &adapter->dcb->state))
948 return -EINVAL; 942 return -EINVAL;
949 943
950 switch (attr) { 944 switch (attr) {
@@ -967,7 +961,7 @@ static u8 qlcnic_dcb_get_app(struct net_device *netdev, u8 idtype, u16 id)
967 .protocol = id, 961 .protocol = id,
968 }; 962 };
969 963
970 if (!test_bit(__QLCNIC_DCB_STATE, &adapter->state)) 964 if (!test_bit(QLCNIC_DCB_STATE, &adapter->dcb->state))
971 return 0; 965 return 0;
972 966
973 return dcb_getapp(netdev, &app); 967 return dcb_getapp(netdev, &app);
@@ -978,7 +972,7 @@ static u8 qlcnic_dcb_get_pfc_state(struct net_device *netdev)
978 struct qlcnic_adapter *adapter = netdev_priv(netdev); 972 struct qlcnic_adapter *adapter = netdev_priv(netdev);
979 struct qlcnic_dcb *dcb = adapter->dcb; 973 struct qlcnic_dcb *dcb = adapter->dcb;
980 974
981 if (!test_bit(__QLCNIC_DCB_STATE, &adapter->state)) 975 if (!test_bit(QLCNIC_DCB_STATE, &dcb->state))
982 return 0; 976 return 0;
983 977
984 return dcb->cfg->type[QLC_DCB_OPER_IDX].pfc_mode_enable; 978 return dcb->cfg->type[QLC_DCB_OPER_IDX].pfc_mode_enable;
@@ -989,7 +983,7 @@ static u8 qlcnic_dcb_get_dcbx(struct net_device *netdev)
989 struct qlcnic_adapter *adapter = netdev_priv(netdev); 983 struct qlcnic_adapter *adapter = netdev_priv(netdev);
990 struct qlcnic_dcb_cfg *cfg = adapter->dcb->cfg; 984 struct qlcnic_dcb_cfg *cfg = adapter->dcb->cfg;
991 985
992 if (!test_bit(__QLCNIC_DCB_STATE, &adapter->state)) 986 if (!test_bit(QLCNIC_DCB_STATE, &adapter->dcb->state))
993 return 0; 987 return 0;
994 988
995 return cfg->capability.dcb_capability; 989 return cfg->capability.dcb_capability;
@@ -1000,7 +994,7 @@ static u8 qlcnic_dcb_get_feat_cfg(struct net_device *netdev, int fid, u8 *flag)
1000 struct qlcnic_adapter *adapter = netdev_priv(netdev); 994 struct qlcnic_adapter *adapter = netdev_priv(netdev);
1001 struct qlcnic_dcb_cee *type; 995 struct qlcnic_dcb_cee *type;
1002 996
1003 if (!test_bit(__QLCNIC_DCB_STATE, &adapter->state)) 997 if (!test_bit(QLCNIC_DCB_STATE, &adapter->dcb->state))
1004 return 1; 998 return 1;
1005 999
1006 type = &adapter->dcb->cfg->type[QLC_DCB_OPER_IDX]; 1000 type = &adapter->dcb->cfg->type[QLC_DCB_OPER_IDX];
@@ -1055,7 +1049,7 @@ static int qlcnic_dcb_peer_app_info(struct net_device *netdev,
1055 1049
1056 *app_count = 0; 1050 *app_count = 0;
1057 1051
1058 if (!test_bit(__QLCNIC_DCB_STATE, &adapter->state)) 1052 if (!test_bit(QLCNIC_DCB_STATE, &adapter->dcb->state))
1059 return 0; 1053 return 0;
1060 1054
1061 peer = &adapter->dcb->cfg->type[QLC_DCB_PEER_IDX]; 1055 peer = &adapter->dcb->cfg->type[QLC_DCB_PEER_IDX];
@@ -1076,7 +1070,7 @@ static int qlcnic_dcb_peer_app_table(struct net_device *netdev,
1076 struct qlcnic_dcb_app *app; 1070 struct qlcnic_dcb_app *app;
1077 int i, j; 1071 int i, j;
1078 1072
1079 if (!test_bit(__QLCNIC_DCB_STATE, &adapter->state)) 1073 if (!test_bit(QLCNIC_DCB_STATE, &adapter->dcb->state))
1080 return 0; 1074 return 0;
1081 1075
1082 peer = &adapter->dcb->cfg->type[QLC_DCB_PEER_IDX]; 1076 peer = &adapter->dcb->cfg->type[QLC_DCB_PEER_IDX];
@@ -1101,7 +1095,7 @@ static int qlcnic_dcb_cee_peer_get_pg(struct net_device *netdev,
1101 struct qlcnic_dcb_cee *peer; 1095 struct qlcnic_dcb_cee *peer;
1102 u8 i, j, k, map; 1096 u8 i, j, k, map;
1103 1097
1104 if (!test_bit(__QLCNIC_DCB_STATE, &adapter->state)) 1098 if (!test_bit(QLCNIC_DCB_STATE, &adapter->dcb->state))
1105 return 0; 1099 return 0;
1106 1100
1107 peer = &adapter->dcb->cfg->type[QLC_DCB_PEER_IDX]; 1101 peer = &adapter->dcb->cfg->type[QLC_DCB_PEER_IDX];
@@ -1136,7 +1130,7 @@ static int qlcnic_dcb_cee_peer_get_pfc(struct net_device *netdev,
1136 1130
1137 pfc->pfc_en = 0; 1131 pfc->pfc_en = 0;
1138 1132
1139 if (!test_bit(__QLCNIC_DCB_STATE, &adapter->state)) 1133 if (!test_bit(QLCNIC_DCB_STATE, &adapter->dcb->state))
1140 return 0; 1134 return 0;
1141 1135
1142 peer = &cfg->type[QLC_DCB_PEER_IDX]; 1136 peer = &cfg->type[QLC_DCB_PEER_IDX];
diff --git a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.h b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.h
index b87ce9fb503e..c04ae0cdc108 100644
--- a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.h
+++ b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_dcb.h
@@ -8,26 +8,29 @@
8#ifndef __QLCNIC_DCBX_H 8#ifndef __QLCNIC_DCBX_H
9#define __QLCNIC_DCBX_H 9#define __QLCNIC_DCBX_H
10 10
11void qlcnic_clear_dcb_ops(struct qlcnic_adapter *); 11#define QLCNIC_DCB_STATE 0
12#define QLCNIC_DCB_AEN_MODE 1
12 13
13#ifdef CONFIG_QLCNIC_DCB 14#ifdef CONFIG_QLCNIC_DCB
14int __qlcnic_register_dcb(struct qlcnic_adapter *); 15int qlcnic_register_dcb(struct qlcnic_adapter *);
15#else 16#else
16static inline int __qlcnic_register_dcb(struct qlcnic_adapter *adapter) 17static inline int qlcnic_register_dcb(struct qlcnic_adapter *adapter)
17{ return 0; } 18{ return 0; }
18#endif 19#endif
19 20
21struct qlcnic_dcb;
22
20struct qlcnic_dcb_ops { 23struct qlcnic_dcb_ops {
21 void (*init_dcbnl_ops) (struct qlcnic_adapter *); 24 int (*query_hw_capability) (struct qlcnic_dcb *, char *);
22 void (*free) (struct qlcnic_adapter *); 25 int (*get_hw_capability) (struct qlcnic_dcb *);
23 int (*attach) (struct qlcnic_adapter *); 26 int (*query_cee_param) (struct qlcnic_dcb *, char *, u8);
24 int (*query_hw_capability) (struct qlcnic_adapter *, char *); 27 void (*init_dcbnl_ops) (struct qlcnic_dcb *);
25 int (*get_hw_capability) (struct qlcnic_adapter *); 28 int (*register_aen) (struct qlcnic_dcb *, bool);
26 void (*get_info) (struct qlcnic_adapter *); 29 void (*aen_handler) (struct qlcnic_dcb *, void *);
27 int (*query_cee_param) (struct qlcnic_adapter *, char *, u8); 30 int (*get_cee_cfg) (struct qlcnic_dcb *);
28 int (*get_cee_cfg) (struct qlcnic_adapter *); 31 void (*get_info) (struct qlcnic_dcb *);
29 int (*register_aen) (struct qlcnic_adapter *, bool); 32 int (*attach) (struct qlcnic_dcb *);
30 void (*handle_aen) (struct qlcnic_adapter *, void *); 33 void (*free) (struct qlcnic_dcb *);
31}; 34};
32 35
33struct qlcnic_dcb { 36struct qlcnic_dcb {
@@ -37,5 +40,85 @@ struct qlcnic_dcb {
37 struct workqueue_struct *wq; 40 struct workqueue_struct *wq;
38 struct qlcnic_dcb_ops *ops; 41 struct qlcnic_dcb_ops *ops;
39 struct qlcnic_dcb_cfg *cfg; 42 struct qlcnic_dcb_cfg *cfg;
43 unsigned long state;
40}; 44};
45
46static inline void qlcnic_clear_dcb_ops(struct qlcnic_dcb *dcb)
47{
48 kfree(dcb);
49 dcb = NULL;
50}
51
52static inline int qlcnic_dcb_get_hw_capability(struct qlcnic_dcb *dcb)
53{
54 if (dcb && dcb->ops->get_hw_capability)
55 return dcb->ops->get_hw_capability(dcb);
56
57 return 0;
58}
59
60static inline void qlcnic_dcb_free(struct qlcnic_dcb *dcb)
61{
62 if (dcb && dcb->ops->free)
63 dcb->ops->free(dcb);
64}
65
66static inline int qlcnic_dcb_attach(struct qlcnic_dcb *dcb)
67{
68 if (dcb && dcb->ops->attach)
69 return dcb->ops->attach(dcb);
70
71 return 0;
72}
73
74static inline int
75qlcnic_dcb_query_hw_capability(struct qlcnic_dcb *dcb, char *buf)
76{
77 if (dcb && dcb->ops->query_hw_capability)
78 return dcb->ops->query_hw_capability(dcb, buf);
79
80 return 0;
81}
82
83static inline void qlcnic_dcb_get_info(struct qlcnic_dcb *dcb)
84{
85 if (dcb && dcb->ops->get_info)
86 dcb->ops->get_info(dcb);
87}
88
89static inline int
90qlcnic_dcb_query_cee_param(struct qlcnic_dcb *dcb, char *buf, u8 type)
91{
92 if (dcb && dcb->ops->query_cee_param)
93 return dcb->ops->query_cee_param(dcb, buf, type);
94
95 return 0;
96}
97
98static inline int qlcnic_dcb_get_cee_cfg(struct qlcnic_dcb *dcb)
99{
100 if (dcb && dcb->ops->get_cee_cfg)
101 return dcb->ops->get_cee_cfg(dcb);
102
103 return 0;
104}
105
106static inline void
107qlcnic_dcb_register_aen(struct qlcnic_dcb *dcb, u8 flag)
108{
109 if (dcb && dcb->ops->register_aen)
110 dcb->ops->register_aen(dcb, flag);
111}
112
113static inline void qlcnic_dcb_aen_handler(struct qlcnic_dcb *dcb, void *msg)
114{
115 if (dcb && dcb->ops->aen_handler)
116 dcb->ops->aen_handler(dcb, msg);
117}
118
119static inline void qlcnic_dcb_init_dcbnl_ops(struct qlcnic_dcb *dcb)
120{
121 if (dcb && dcb->ops->init_dcbnl_ops)
122 dcb->ops->init_dcbnl_ops(dcb);
123}
41#endif 124#endif
diff --git a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_ethtool.c b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_ethtool.c
index ff83a9fcd4c5..b36c02fafcfd 100644
--- a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_ethtool.c
+++ b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_ethtool.c
@@ -27,43 +27,36 @@ static const u32 qlcnic_fw_dump_level[] = {
27}; 27};
28 28
29static const struct qlcnic_stats qlcnic_gstrings_stats[] = { 29static const struct qlcnic_stats qlcnic_gstrings_stats[] = {
30 {"xmit_on", QLC_SIZEOF(stats.xmit_on), QLC_OFF(stats.xmit_on)},
31 {"xmit_off", QLC_SIZEOF(stats.xmit_off), QLC_OFF(stats.xmit_off)},
30 {"xmit_called", QLC_SIZEOF(stats.xmitcalled), 32 {"xmit_called", QLC_SIZEOF(stats.xmitcalled),
31 QLC_OFF(stats.xmitcalled)}, 33 QLC_OFF(stats.xmitcalled)},
32 {"xmit_finished", QLC_SIZEOF(stats.xmitfinished), 34 {"xmit_finished", QLC_SIZEOF(stats.xmitfinished),
33 QLC_OFF(stats.xmitfinished)}, 35 QLC_OFF(stats.xmitfinished)},
34 {"rx_dropped", QLC_SIZEOF(stats.rxdropped), QLC_OFF(stats.rxdropped)}, 36 {"tx dma map error", QLC_SIZEOF(stats.tx_dma_map_error),
37 QLC_OFF(stats.tx_dma_map_error)},
38 {"tx_bytes", QLC_SIZEOF(stats.txbytes), QLC_OFF(stats.txbytes)},
35 {"tx_dropped", QLC_SIZEOF(stats.txdropped), QLC_OFF(stats.txdropped)}, 39 {"tx_dropped", QLC_SIZEOF(stats.txdropped), QLC_OFF(stats.txdropped)},
36 {"csummed", QLC_SIZEOF(stats.csummed), QLC_OFF(stats.csummed)}, 40 {"rx dma map error", QLC_SIZEOF(stats.rx_dma_map_error),
41 QLC_OFF(stats.rx_dma_map_error)},
37 {"rx_pkts", QLC_SIZEOF(stats.rx_pkts), QLC_OFF(stats.rx_pkts)}, 42 {"rx_pkts", QLC_SIZEOF(stats.rx_pkts), QLC_OFF(stats.rx_pkts)},
38 {"lro_pkts", QLC_SIZEOF(stats.lro_pkts), QLC_OFF(stats.lro_pkts)},
39 {"rx_bytes", QLC_SIZEOF(stats.rxbytes), QLC_OFF(stats.rxbytes)}, 43 {"rx_bytes", QLC_SIZEOF(stats.rxbytes), QLC_OFF(stats.rxbytes)},
40 {"tx_bytes", QLC_SIZEOF(stats.txbytes), QLC_OFF(stats.txbytes)}, 44 {"rx_dropped", QLC_SIZEOF(stats.rxdropped), QLC_OFF(stats.rxdropped)},
45 {"null rxbuf", QLC_SIZEOF(stats.null_rxbuf), QLC_OFF(stats.null_rxbuf)},
46 {"csummed", QLC_SIZEOF(stats.csummed), QLC_OFF(stats.csummed)},
47 {"lro_pkts", QLC_SIZEOF(stats.lro_pkts), QLC_OFF(stats.lro_pkts)},
41 {"lrobytes", QLC_SIZEOF(stats.lrobytes), QLC_OFF(stats.lrobytes)}, 48 {"lrobytes", QLC_SIZEOF(stats.lrobytes), QLC_OFF(stats.lrobytes)},
42 {"lso_frames", QLC_SIZEOF(stats.lso_frames), QLC_OFF(stats.lso_frames)}, 49 {"lso_frames", QLC_SIZEOF(stats.lso_frames), QLC_OFF(stats.lso_frames)},
43 {"xmit_on", QLC_SIZEOF(stats.xmit_on), QLC_OFF(stats.xmit_on)},
44 {"xmit_off", QLC_SIZEOF(stats.xmit_off), QLC_OFF(stats.xmit_off)},
45 {"skb_alloc_failure", QLC_SIZEOF(stats.skb_alloc_failure), 50 {"skb_alloc_failure", QLC_SIZEOF(stats.skb_alloc_failure),
46 QLC_OFF(stats.skb_alloc_failure)}, 51 QLC_OFF(stats.skb_alloc_failure)},
47 {"null rxbuf", QLC_SIZEOF(stats.null_rxbuf), QLC_OFF(stats.null_rxbuf)},
48 {"rx dma map error", QLC_SIZEOF(stats.rx_dma_map_error),
49 QLC_OFF(stats.rx_dma_map_error)},
50 {"tx dma map error", QLC_SIZEOF(stats.tx_dma_map_error),
51 QLC_OFF(stats.tx_dma_map_error)},
52 {"mac_filter_limit_overrun", QLC_SIZEOF(stats.mac_filter_limit_overrun), 52 {"mac_filter_limit_overrun", QLC_SIZEOF(stats.mac_filter_limit_overrun),
53 QLC_OFF(stats.mac_filter_limit_overrun)}, 53 QLC_OFF(stats.mac_filter_limit_overrun)},
54 {"spurious intr", QLC_SIZEOF(stats.spurious_intr), 54 {"spurious intr", QLC_SIZEOF(stats.spurious_intr),
55 QLC_OFF(stats.spurious_intr)}, 55 QLC_OFF(stats.spurious_intr)},
56 56
57}; 57};
58 58
59static const char qlcnic_device_gstrings_stats[][ETH_GSTRING_LEN] = { 59static const char qlcnic_device_gstrings_stats[][ETH_GSTRING_LEN] = {
60 "rx unicast frames",
61 "rx multicast frames",
62 "rx broadcast frames",
63 "rx dropped frames",
64 "rx errors",
65 "rx local frames",
66 "rx numbytes",
67 "tx unicast frames", 60 "tx unicast frames",
68 "tx multicast frames", 61 "tx multicast frames",
69 "tx broadcast frames", 62 "tx broadcast frames",
@@ -71,6 +64,13 @@ static const char qlcnic_device_gstrings_stats[][ETH_GSTRING_LEN] = {
71 "tx errors", 64 "tx errors",
72 "tx local frames", 65 "tx local frames",
73 "tx numbytes", 66 "tx numbytes",
67 "rx unicast frames",
68 "rx multicast frames",
69 "rx broadcast frames",
70 "rx dropped frames",
71 "rx errors",
72 "rx local frames",
73 "rx numbytes",
74}; 74};
75 75
76static const char qlcnic_83xx_tx_stats_strings[][ETH_GSTRING_LEN] = { 76static const char qlcnic_83xx_tx_stats_strings[][ETH_GSTRING_LEN] = {
@@ -126,13 +126,16 @@ static const char qlcnic_83xx_mac_stats_strings[][ETH_GSTRING_LEN] = {
126 126
127#define QLCNIC_STATS_LEN ARRAY_SIZE(qlcnic_gstrings_stats) 127#define QLCNIC_STATS_LEN ARRAY_SIZE(qlcnic_gstrings_stats)
128 128
129static const char qlcnic_tx_ring_stats_strings[][ETH_GSTRING_LEN] = { 129static const char qlcnic_tx_queue_stats_strings[][ETH_GSTRING_LEN] = {
130 "xmit_on", 130 "xmit_on",
131 "xmit_off", 131 "xmit_off",
132 "xmit_called", 132 "xmit_called",
133 "xmit_finished", 133 "xmit_finished",
134 "tx_bytes",
134}; 135};
135 136
137#define QLCNIC_TX_STATS_LEN ARRAY_SIZE(qlcnic_tx_queue_stats_strings)
138
136static const char qlcnic_83xx_rx_stats_strings[][ETH_GSTRING_LEN] = { 139static const char qlcnic_83xx_rx_stats_strings[][ETH_GSTRING_LEN] = {
137 "ctx_rx_bytes", 140 "ctx_rx_bytes",
138 "ctx_rx_pkts", 141 "ctx_rx_pkts",
@@ -187,8 +190,8 @@ static int qlcnic_dev_statistics_len(struct qlcnic_adapter *adapter)
187 return -1; 190 return -1;
188} 191}
189 192
190#define QLCNIC_RING_REGS_COUNT 20 193#define QLCNIC_TX_INTR_NOT_CONFIGURED 0X78563412
191#define QLCNIC_RING_REGS_LEN (QLCNIC_RING_REGS_COUNT * sizeof(u32)) 194
192#define QLCNIC_MAX_EEPROM_LEN 1024 195#define QLCNIC_MAX_EEPROM_LEN 1024
193 196
194static const u32 diag_registers[] = { 197static const u32 diag_registers[] = {
@@ -219,7 +222,15 @@ static const u32 ext_diag_registers[] = {
219}; 222};
220 223
221#define QLCNIC_MGMT_API_VERSION 2 224#define QLCNIC_MGMT_API_VERSION 2
222#define QLCNIC_ETHTOOL_REGS_VER 3 225#define QLCNIC_ETHTOOL_REGS_VER 4
226
227static inline int qlcnic_get_ring_regs_len(struct qlcnic_adapter *adapter)
228{
229 int ring_regs_cnt = (adapter->drv_tx_rings * 5) +
230 (adapter->max_rds_rings * 2) +
231 (adapter->drv_sds_rings * 3) + 5;
232 return ring_regs_cnt * sizeof(u32);
233}
223 234
224static int qlcnic_get_regs_len(struct net_device *dev) 235static int qlcnic_get_regs_len(struct net_device *dev)
225{ 236{
@@ -231,7 +242,9 @@ static int qlcnic_get_regs_len(struct net_device *dev)
231 else 242 else
232 len = sizeof(ext_diag_registers) + sizeof(diag_registers); 243 len = sizeof(ext_diag_registers) + sizeof(diag_registers);
233 244
234 return QLCNIC_RING_REGS_LEN + len + QLCNIC_DEV_INFO_SIZE + 1; 245 len += ((QLCNIC_DEV_INFO_SIZE + 2) * sizeof(u32));
246 len += qlcnic_get_ring_regs_len(adapter);
247 return len;
235} 248}
236 249
237static int qlcnic_get_eeprom_len(struct net_device *dev) 250static int qlcnic_get_eeprom_len(struct net_device *dev)
@@ -493,6 +506,8 @@ qlcnic_get_regs(struct net_device *dev, struct ethtool_regs *regs, void *p)
493 struct qlcnic_adapter *adapter = netdev_priv(dev); 506 struct qlcnic_adapter *adapter = netdev_priv(dev);
494 struct qlcnic_recv_context *recv_ctx = adapter->recv_ctx; 507 struct qlcnic_recv_context *recv_ctx = adapter->recv_ctx;
495 struct qlcnic_host_sds_ring *sds_ring; 508 struct qlcnic_host_sds_ring *sds_ring;
509 struct qlcnic_host_rds_ring *rds_rings;
510 struct qlcnic_host_tx_ring *tx_ring;
496 u32 *regs_buff = p; 511 u32 *regs_buff = p;
497 int ring, i = 0; 512 int ring, i = 0;
498 513
@@ -512,21 +527,35 @@ qlcnic_get_regs(struct net_device *dev, struct ethtool_regs *regs, void *p)
512 if (!test_bit(__QLCNIC_DEV_UP, &adapter->state)) 527 if (!test_bit(__QLCNIC_DEV_UP, &adapter->state))
513 return; 528 return;
514 529
515 regs_buff[i++] = 0xFFEFCDAB; /* Marker btw regs and ring count*/ 530 /* Marker btw regs and TX ring count */
516 531 regs_buff[i++] = 0xFFEFCDAB;
517 regs_buff[i++] = 1; /* No. of tx ring */ 532
518 regs_buff[i++] = le32_to_cpu(*(adapter->tx_ring->hw_consumer)); 533 regs_buff[i++] = adapter->drv_tx_rings; /* No. of TX ring */
519 regs_buff[i++] = readl(adapter->tx_ring->crb_cmd_producer); 534 for (ring = 0; ring < adapter->drv_tx_rings; ring++) {
520 535 tx_ring = &adapter->tx_ring[ring];
521 regs_buff[i++] = 2; /* No. of rx ring */ 536 regs_buff[i++] = le32_to_cpu(*(tx_ring->hw_consumer));
522 regs_buff[i++] = readl(recv_ctx->rds_rings[0].crb_rcv_producer); 537 regs_buff[i++] = tx_ring->sw_consumer;
523 regs_buff[i++] = readl(recv_ctx->rds_rings[1].crb_rcv_producer); 538 regs_buff[i++] = readl(tx_ring->crb_cmd_producer);
539 regs_buff[i++] = tx_ring->producer;
540 if (tx_ring->crb_intr_mask)
541 regs_buff[i++] = readl(tx_ring->crb_intr_mask);
542 else
543 regs_buff[i++] = QLCNIC_TX_INTR_NOT_CONFIGURED;
544 }
524 545
525 regs_buff[i++] = adapter->max_sds_rings; 546 regs_buff[i++] = adapter->max_rds_rings; /* No. of RX ring */
547 for (ring = 0; ring < adapter->max_rds_rings; ring++) {
548 rds_rings = &recv_ctx->rds_rings[ring];
549 regs_buff[i++] = readl(rds_rings->crb_rcv_producer);
550 regs_buff[i++] = rds_rings->producer;
551 }
526 552
527 for (ring = 0; ring < adapter->max_sds_rings; ring++) { 553 regs_buff[i++] = adapter->drv_sds_rings; /* No. of SDS ring */
554 for (ring = 0; ring < adapter->drv_sds_rings; ring++) {
528 sds_ring = &(recv_ctx->sds_rings[ring]); 555 sds_ring = &(recv_ctx->sds_rings[ring]);
529 regs_buff[i++] = readl(sds_ring->crb_sts_consumer); 556 regs_buff[i++] = readl(sds_ring->crb_sts_consumer);
557 regs_buff[i++] = sds_ring->consumer;
558 regs_buff[i++] = readl(sds_ring->crb_intr_mask);
530 } 559 }
531} 560}
532 561
@@ -635,46 +664,88 @@ qlcnic_set_ringparam(struct net_device *dev,
635 return qlcnic_reset_context(adapter); 664 return qlcnic_reset_context(adapter);
636} 665}
637 666
667static int qlcnic_validate_ring_count(struct qlcnic_adapter *adapter,
668 u8 rx_ring, u8 tx_ring)
669{
670 if (rx_ring != 0) {
671 if (rx_ring > adapter->max_sds_rings) {
672 netdev_err(adapter->netdev, "Invalid ring count, SDS ring count %d should not be greater than max %d driver sds rings.\n",
673 rx_ring, adapter->max_sds_rings);
674 return -EINVAL;
675 }
676 }
677
678 if (tx_ring != 0) {
679 if (qlcnic_82xx_check(adapter) &&
680 (tx_ring > adapter->max_tx_rings)) {
681 netdev_err(adapter->netdev,
682 "Invalid ring count, Tx ring count %d should not be greater than max %d driver Tx rings.\n",
683 tx_ring, adapter->max_tx_rings);
684 return -EINVAL;
685 }
686
687 if (qlcnic_83xx_check(adapter) &&
688 (tx_ring > QLCNIC_SINGLE_RING)) {
689 netdev_err(adapter->netdev,
690 "Invalid ring count, Tx ring count %d should not be greater than %d driver Tx rings.\n",
691 tx_ring, QLCNIC_SINGLE_RING);
692 return -EINVAL;
693 }
694 }
695
696 return 0;
697}
698
638static void qlcnic_get_channels(struct net_device *dev, 699static void qlcnic_get_channels(struct net_device *dev,
639 struct ethtool_channels *channel) 700 struct ethtool_channels *channel)
640{ 701{
641 struct qlcnic_adapter *adapter = netdev_priv(dev); 702 struct qlcnic_adapter *adapter = netdev_priv(dev);
642 int min;
643
644 min = min_t(int, adapter->ahw->max_rx_ques, num_online_cpus());
645 channel->max_rx = rounddown_pow_of_two(min);
646 channel->max_tx = min_t(int, QLCNIC_MAX_TX_RINGS, num_online_cpus());
647 703
648 channel->rx_count = adapter->max_sds_rings; 704 channel->max_rx = adapter->max_sds_rings;
649 channel->tx_count = adapter->max_drv_tx_rings; 705 channel->max_tx = adapter->max_tx_rings;
706 channel->rx_count = adapter->drv_sds_rings;
707 channel->tx_count = adapter->drv_tx_rings;
650} 708}
651 709
652static int qlcnic_set_channels(struct net_device *dev, 710static int qlcnic_set_channels(struct net_device *dev,
653 struct ethtool_channels *channel) 711 struct ethtool_channels *channel)
654{ 712{
655 struct qlcnic_adapter *adapter = netdev_priv(dev); 713 struct qlcnic_adapter *adapter = netdev_priv(dev);
656 int err; 714 int err;
657 int txq = 0;
658 715
659 if (channel->other_count || channel->combined_count) 716 if (channel->other_count || channel->combined_count)
660 return -EINVAL; 717 return -EINVAL;
661 718
719 err = qlcnic_validate_ring_count(adapter, channel->rx_count,
720 channel->tx_count);
721 if (err)
722 return err;
723
662 if (channel->rx_count) { 724 if (channel->rx_count) {
663 err = qlcnic_validate_max_rss(adapter, channel->rx_count); 725 err = qlcnic_validate_rings(adapter, channel->rx_count,
664 if (err) 726 QLCNIC_RX_QUEUE);
727 if (err) {
728 netdev_err(dev, "Unable to configure %u SDS rings\n",
729 channel->rx_count);
665 return err; 730 return err;
731 }
666 } 732 }
667 733
668 if (qlcnic_82xx_check(adapter) && channel->tx_count) { 734 if (channel->tx_count) {
669 err = qlcnic_validate_max_tx_rings(adapter, channel->tx_count); 735 err = qlcnic_validate_rings(adapter, channel->tx_count,
670 if (err) 736 QLCNIC_TX_QUEUE);
737 if (err) {
738 netdev_err(dev, "Unable to configure %u Tx rings\n",
739 channel->tx_count);
671 return err; 740 return err;
672 txq = channel->tx_count; 741 }
673 } 742 }
674 743
675 err = qlcnic_set_max_rss(adapter, channel->rx_count, txq); 744 err = qlcnic_setup_rings(adapter, channel->rx_count,
676 netdev_info(dev, "allocated 0x%x sds rings and 0x%x tx rings\n", 745 channel->tx_count);
677 adapter->max_sds_rings, adapter->max_drv_tx_rings); 746 netdev_info(dev, "Allocated %d SDS rings and %d Tx rings\n",
747 adapter->drv_sds_rings, adapter->drv_tx_rings);
748
678 return err; 749 return err;
679} 750}
680 751
@@ -876,7 +947,7 @@ static int qlcnic_irq_test(struct net_device *netdev)
876 struct qlcnic_adapter *adapter = netdev_priv(netdev); 947 struct qlcnic_adapter *adapter = netdev_priv(netdev);
877 struct qlcnic_hardware_context *ahw = adapter->ahw; 948 struct qlcnic_hardware_context *ahw = adapter->ahw;
878 struct qlcnic_cmd_args cmd; 949 struct qlcnic_cmd_args cmd;
879 int ret, max_sds_rings = adapter->max_sds_rings; 950 int ret, drv_sds_rings = adapter->drv_sds_rings;
880 951
881 if (qlcnic_83xx_check(adapter)) 952 if (qlcnic_83xx_check(adapter))
882 return qlcnic_83xx_interrupt_test(netdev); 953 return qlcnic_83xx_interrupt_test(netdev);
@@ -905,10 +976,10 @@ done:
905 qlcnic_free_mbx_args(&cmd); 976 qlcnic_free_mbx_args(&cmd);
906 977
907free_diag_res: 978free_diag_res:
908 qlcnic_diag_free_res(netdev, max_sds_rings); 979 qlcnic_diag_free_res(netdev, drv_sds_rings);
909 980
910clear_diag_irq: 981clear_diag_irq:
911 adapter->max_sds_rings = max_sds_rings; 982 adapter->drv_sds_rings = drv_sds_rings;
912 clear_bit(__QLCNIC_RESETTING, &adapter->state); 983 clear_bit(__QLCNIC_RESETTING, &adapter->state);
913 984
914 return ret; 985 return ret;
@@ -984,8 +1055,8 @@ int qlcnic_do_lb_test(struct qlcnic_adapter *adapter, u8 mode)
984int qlcnic_loopback_test(struct net_device *netdev, u8 mode) 1055int qlcnic_loopback_test(struct net_device *netdev, u8 mode)
985{ 1056{
986 struct qlcnic_adapter *adapter = netdev_priv(netdev); 1057 struct qlcnic_adapter *adapter = netdev_priv(netdev);
987 int max_drv_tx_rings = adapter->max_drv_tx_rings; 1058 int drv_tx_rings = adapter->drv_tx_rings;
988 int max_sds_rings = adapter->max_sds_rings; 1059 int drv_sds_rings = adapter->drv_sds_rings;
989 struct qlcnic_host_sds_ring *sds_ring; 1060 struct qlcnic_host_sds_ring *sds_ring;
990 struct qlcnic_hardware_context *ahw = adapter->ahw; 1061 struct qlcnic_hardware_context *ahw = adapter->ahw;
991 int loop = 0; 1062 int loop = 0;
@@ -1040,11 +1111,11 @@ int qlcnic_loopback_test(struct net_device *netdev, u8 mode)
1040 qlcnic_clear_lb_mode(adapter, mode); 1111 qlcnic_clear_lb_mode(adapter, mode);
1041 1112
1042 free_res: 1113 free_res:
1043 qlcnic_diag_free_res(netdev, max_sds_rings); 1114 qlcnic_diag_free_res(netdev, drv_sds_rings);
1044 1115
1045 clear_it: 1116 clear_it:
1046 adapter->max_sds_rings = max_sds_rings; 1117 adapter->drv_sds_rings = drv_sds_rings;
1047 adapter->max_drv_tx_rings = max_drv_tx_rings; 1118 adapter->drv_tx_rings = drv_tx_rings;
1048 clear_bit(__QLCNIC_RESETTING, &adapter->state); 1119 clear_bit(__QLCNIC_RESETTING, &adapter->state);
1049 return ret; 1120 return ret;
1050} 1121}
@@ -1097,11 +1168,11 @@ qlcnic_get_strings(struct net_device *dev, u32 stringset, u8 *data)
1097 QLCNIC_TEST_LEN * ETH_GSTRING_LEN); 1168 QLCNIC_TEST_LEN * ETH_GSTRING_LEN);
1098 break; 1169 break;
1099 case ETH_SS_STATS: 1170 case ETH_SS_STATS:
1100 num_stats = ARRAY_SIZE(qlcnic_tx_ring_stats_strings); 1171 num_stats = ARRAY_SIZE(qlcnic_tx_queue_stats_strings);
1101 for (i = 0; i < adapter->max_drv_tx_rings; i++) { 1172 for (i = 0; i < adapter->drv_tx_rings; i++) {
1102 for (index = 0; index < num_stats; index++) { 1173 for (index = 0; index < num_stats; index++) {
1103 sprintf(data, "tx_ring_%d %s", i, 1174 sprintf(data, "tx_queue_%d %s", i,
1104 qlcnic_tx_ring_stats_strings[index]); 1175 qlcnic_tx_queue_stats_strings[index]);
1105 data += ETH_GSTRING_LEN; 1176 data += ETH_GSTRING_LEN;
1106 } 1177 }
1107 } 1178 }
@@ -1199,6 +1270,36 @@ static u64 *qlcnic_fill_stats(u64 *data, void *stats, int type)
1199 return data; 1270 return data;
1200} 1271}
1201 1272
1273static void qlcnic_update_stats(struct qlcnic_adapter *adapter)
1274{
1275 struct qlcnic_host_tx_ring *tx_ring;
1276 int ring;
1277
1278 for (ring = 0; ring < adapter->drv_tx_rings; ring++) {
1279 tx_ring = &adapter->tx_ring[ring];
1280 adapter->stats.xmit_on += tx_ring->tx_stats.xmit_on;
1281 adapter->stats.xmit_off += tx_ring->tx_stats.xmit_off;
1282 adapter->stats.xmitcalled += tx_ring->tx_stats.xmit_called;
1283 adapter->stats.xmitfinished += tx_ring->tx_stats.xmit_finished;
1284 adapter->stats.txbytes += tx_ring->tx_stats.tx_bytes;
1285 }
1286}
1287
1288static u64 *qlcnic_fill_tx_queue_stats(u64 *data, void *stats)
1289{
1290 struct qlcnic_host_tx_ring *tx_ring;
1291
1292 tx_ring = (struct qlcnic_host_tx_ring *)stats;
1293
1294 *data++ = QLCNIC_FILL_STATS(tx_ring->tx_stats.xmit_on);
1295 *data++ = QLCNIC_FILL_STATS(tx_ring->tx_stats.xmit_off);
1296 *data++ = QLCNIC_FILL_STATS(tx_ring->tx_stats.xmit_called);
1297 *data++ = QLCNIC_FILL_STATS(tx_ring->tx_stats.xmit_finished);
1298 *data++ = QLCNIC_FILL_STATS(tx_ring->tx_stats.tx_bytes);
1299
1300 return data;
1301}
1302
1202static void qlcnic_get_ethtool_stats(struct net_device *dev, 1303static void qlcnic_get_ethtool_stats(struct net_device *dev,
1203 struct ethtool_stats *stats, u64 *data) 1304 struct ethtool_stats *stats, u64 *data)
1204{ 1305{
@@ -1206,19 +1307,20 @@ static void qlcnic_get_ethtool_stats(struct net_device *dev,
1206 struct qlcnic_host_tx_ring *tx_ring; 1307 struct qlcnic_host_tx_ring *tx_ring;
1207 struct qlcnic_esw_statistics port_stats; 1308 struct qlcnic_esw_statistics port_stats;
1208 struct qlcnic_mac_statistics mac_stats; 1309 struct qlcnic_mac_statistics mac_stats;
1209 int index, ret, length, size, ring; 1310 int index, ret, length, size, tx_size, ring;
1210 char *p; 1311 char *p;
1211 1312
1212 memset(data, 0, adapter->max_drv_tx_rings * 4 * sizeof(u64)); 1313 tx_size = adapter->drv_tx_rings * QLCNIC_TX_STATS_LEN;
1213 for (ring = 0, index = 0; ring < adapter->max_drv_tx_rings; ring++) { 1314
1315 memset(data, 0, tx_size * sizeof(u64));
1316 for (ring = 0, index = 0; ring < adapter->drv_tx_rings; ring++) {
1214 if (test_bit(__QLCNIC_DEV_UP, &adapter->state)) { 1317 if (test_bit(__QLCNIC_DEV_UP, &adapter->state)) {
1215 tx_ring = &adapter->tx_ring[ring]; 1318 tx_ring = &adapter->tx_ring[ring];
1216 *data++ = tx_ring->xmit_on; 1319 data = qlcnic_fill_tx_queue_stats(data, tx_ring);
1217 *data++ = tx_ring->xmit_off; 1320 qlcnic_update_stats(adapter);
1218 *data++ = tx_ring->xmit_called;
1219 *data++ = tx_ring->xmit_finished;
1220 } 1321 }
1221 } 1322 }
1323
1222 memset(data, 0, stats->n_stats * sizeof(u64)); 1324 memset(data, 0, stats->n_stats * sizeof(u64));
1223 length = QLCNIC_STATS_LEN; 1325 length = QLCNIC_STATS_LEN;
1224 for (index = 0; index < length; index++) { 1326 for (index = 0; index < length; index++) {
@@ -1260,7 +1362,7 @@ static int qlcnic_set_led(struct net_device *dev,
1260 enum ethtool_phys_id_state state) 1362 enum ethtool_phys_id_state state)
1261{ 1363{
1262 struct qlcnic_adapter *adapter = netdev_priv(dev); 1364 struct qlcnic_adapter *adapter = netdev_priv(dev);
1263 int max_sds_rings = adapter->max_sds_rings; 1365 int drv_sds_rings = adapter->drv_sds_rings;
1264 int err = -EIO, active = 1; 1366 int err = -EIO, active = 1;
1265 1367
1266 if (qlcnic_83xx_check(adapter)) 1368 if (qlcnic_83xx_check(adapter))
@@ -1318,7 +1420,7 @@ static int qlcnic_set_led(struct net_device *dev,
1318 } 1420 }
1319 1421
1320 if (test_and_clear_bit(__QLCNIC_DIAG_RES_ALLOC, &adapter->state)) 1422 if (test_and_clear_bit(__QLCNIC_DIAG_RES_ALLOC, &adapter->state))
1321 qlcnic_diag_free_res(dev, max_sds_rings); 1423 qlcnic_diag_free_res(dev, drv_sds_rings);
1322 1424
1323 if (!active || err) 1425 if (!active || err)
1324 clear_bit(__QLCNIC_LED_ENABLE, &adapter->state); 1426 clear_bit(__QLCNIC_LED_ENABLE, &adapter->state);
@@ -1659,7 +1761,6 @@ qlcnic_set_dump(struct net_device *netdev, struct ethtool_dump *val)
1659 struct qlcnic_fw_dump *fw_dump = &adapter->ahw->fw_dump; 1761 struct qlcnic_fw_dump *fw_dump = &adapter->ahw->fw_dump;
1660 bool valid_mask = false; 1762 bool valid_mask = false;
1661 int i, ret = 0; 1763 int i, ret = 0;
1662 u32 state;
1663 1764
1664 switch (val->flag) { 1765 switch (val->flag) {
1665 case QLCNIC_FORCE_FW_DUMP_KEY: 1766 case QLCNIC_FORCE_FW_DUMP_KEY:
@@ -1712,9 +1813,8 @@ qlcnic_set_dump(struct net_device *netdev, struct ethtool_dump *val)
1712 1813
1713 case QLCNIC_SET_QUIESCENT: 1814 case QLCNIC_SET_QUIESCENT:
1714 case QLCNIC_RESET_QUIESCENT: 1815 case QLCNIC_RESET_QUIESCENT:
1715 state = QLC_SHARED_REG_RD32(adapter, QLCNIC_CRB_DEV_STATE); 1816 if (test_bit(__QLCNIC_MAINTENANCE_MODE, &adapter->state))
1716 if (state == QLCNIC_DEV_FAILED || (state == QLCNIC_DEV_BADBAD)) 1817 netdev_info(netdev, "Device is in non-operational state\n");
1717 netdev_info(netdev, "Device in FAILED state\n");
1718 break; 1818 break;
1719 1819
1720 default: 1820 default:
diff --git a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
index b64e2bef9428..6f7f60c09f07 100644
--- a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
+++ b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c
@@ -445,7 +445,7 @@ int qlcnic_82xx_sre_macaddr_change(struct qlcnic_adapter *adapter, u8 *addr,
445 445
446 mac_req = (struct qlcnic_mac_req *)&req.words[0]; 446 mac_req = (struct qlcnic_mac_req *)&req.words[0];
447 mac_req->op = op; 447 mac_req->op = op;
448 memcpy(mac_req->mac_addr, addr, 6); 448 memcpy(mac_req->mac_addr, addr, ETH_ALEN);
449 449
450 vlan_req = (struct qlcnic_vlan_req *)&req.words[1]; 450 vlan_req = (struct qlcnic_vlan_req *)&req.words[1];
451 vlan_req->vlan_id = cpu_to_le16(vlan_id); 451 vlan_req->vlan_id = cpu_to_le16(vlan_id);
diff --git a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.h b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.h
index 272c356cf9b2..13303e7d1ed7 100644
--- a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.h
+++ b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.h
@@ -146,6 +146,12 @@ struct qlcnic_mailbox_metadata {
146#define QLCNIC_MBX_PORT_RSP_OK 0x1a 146#define QLCNIC_MBX_PORT_RSP_OK 0x1a
147#define QLCNIC_MBX_ASYNC_EVENT BIT_15 147#define QLCNIC_MBX_ASYNC_EVENT BIT_15
148 148
149/* Set HW Tx ring limit for 82xx adapter. */
150#define QLCNIC_MAX_HW_TX_RINGS 8
151#define QLCNIC_MAX_HW_VNIC_TX_RINGS 4
152#define QLCNIC_MAX_TX_RINGS 8
153#define QLCNIC_MAX_SDS_RINGS 8
154
149struct qlcnic_pci_info; 155struct qlcnic_pci_info;
150struct qlcnic_info; 156struct qlcnic_info;
151struct qlcnic_cmd_args; 157struct qlcnic_cmd_args;
@@ -176,7 +182,7 @@ int qlcnic_82xx_set_lb_mode(struct qlcnic_adapter *, u8);
176void qlcnic_82xx_write_crb(struct qlcnic_adapter *, char *, loff_t, size_t); 182void qlcnic_82xx_write_crb(struct qlcnic_adapter *, char *, loff_t, size_t);
177void qlcnic_82xx_read_crb(struct qlcnic_adapter *, char *, loff_t, size_t); 183void qlcnic_82xx_read_crb(struct qlcnic_adapter *, char *, loff_t, size_t);
178void qlcnic_82xx_dev_request_reset(struct qlcnic_adapter *, u32); 184void qlcnic_82xx_dev_request_reset(struct qlcnic_adapter *, u32);
179int qlcnic_82xx_setup_intr(struct qlcnic_adapter *, u8, int); 185int qlcnic_82xx_setup_intr(struct qlcnic_adapter *);
180irqreturn_t qlcnic_82xx_clear_legacy_intr(struct qlcnic_adapter *); 186irqreturn_t qlcnic_82xx_clear_legacy_intr(struct qlcnic_adapter *);
181int qlcnic_82xx_issue_cmd(struct qlcnic_adapter *adapter, 187int qlcnic_82xx_issue_cmd(struct qlcnic_adapter *adapter,
182 struct qlcnic_cmd_args *); 188 struct qlcnic_cmd_args *);
diff --git a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c
index 66c26cf7a2b8..e9c21e5d0ca9 100644
--- a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c
+++ b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c
@@ -236,7 +236,7 @@ int qlcnic_alloc_sw_resources(struct qlcnic_adapter *adapter)
236 spin_lock_init(&rds_ring->lock); 236 spin_lock_init(&rds_ring->lock);
237 } 237 }
238 238
239 for (ring = 0; ring < adapter->max_sds_rings; ring++) { 239 for (ring = 0; ring < adapter->drv_sds_rings; ring++) {
240 sds_ring = &recv_ctx->sds_rings[ring]; 240 sds_ring = &recv_ctx->sds_rings[ring];
241 sds_ring->irq = adapter->msix_entries[ring].vector; 241 sds_ring->irq = adapter->msix_entries[ring].vector;
242 sds_ring->adapter = adapter; 242 sds_ring->adapter = adapter;
diff --git a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
index 11b4bb83b930..0149c9495347 100644
--- a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
+++ b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
@@ -581,10 +581,7 @@ netdev_tx_t qlcnic_xmit_frame(struct sk_buff *skb, struct net_device *netdev)
581 goto drop_packet; 581 goto drop_packet;
582 } 582 }
583 583
584 if (qlcnic_check_multi_tx(adapter)) 584 tx_ring = &adapter->tx_ring[skb_get_queue_mapping(skb)];
585 tx_ring = &adapter->tx_ring[skb_get_queue_mapping(skb)];
586 else
587 tx_ring = &adapter->tx_ring[0];
588 num_txd = tx_ring->num_desc; 585 num_txd = tx_ring->num_desc;
589 586
590 frag_count = skb_shinfo(skb)->nr_frags + 1; 587 frag_count = skb_shinfo(skb)->nr_frags + 1;
@@ -607,8 +604,7 @@ netdev_tx_t qlcnic_xmit_frame(struct sk_buff *skb, struct net_device *netdev)
607 if (qlcnic_tx_avail(tx_ring) > TX_STOP_THRESH) { 604 if (qlcnic_tx_avail(tx_ring) > TX_STOP_THRESH) {
608 netif_tx_start_queue(tx_ring->txq); 605 netif_tx_start_queue(tx_ring->txq);
609 } else { 606 } else {
610 adapter->stats.xmit_off++; 607 tx_ring->tx_stats.xmit_off++;
611 tx_ring->xmit_off++;
612 return NETDEV_TX_BUSY; 608 return NETDEV_TX_BUSY;
613 } 609 }
614 } 610 }
@@ -669,9 +665,8 @@ netdev_tx_t qlcnic_xmit_frame(struct sk_buff *skb, struct net_device *netdev)
669 if (adapter->drv_mac_learn) 665 if (adapter->drv_mac_learn)
670 qlcnic_send_filter(adapter, first_desc, skb); 666 qlcnic_send_filter(adapter, first_desc, skb);
671 667
672 adapter->stats.txbytes += skb->len; 668 tx_ring->tx_stats.tx_bytes += skb->len;
673 adapter->stats.xmitcalled++; 669 tx_ring->tx_stats.xmit_called++;
674 tx_ring->xmit_called++;
675 670
676 qlcnic_update_cmd_producer(tx_ring); 671 qlcnic_update_cmd_producer(tx_ring);
677 672
@@ -789,6 +784,9 @@ static int qlcnic_process_cmd_ring(struct qlcnic_adapter *adapter,
789 struct net_device *netdev = adapter->netdev; 784 struct net_device *netdev = adapter->netdev;
790 struct qlcnic_skb_frag *frag; 785 struct qlcnic_skb_frag *frag;
791 786
787 if (!spin_trylock(&adapter->tx_clean_lock))
788 return 1;
789
792 sw_consumer = tx_ring->sw_consumer; 790 sw_consumer = tx_ring->sw_consumer;
793 hw_consumer = le32_to_cpu(*(tx_ring->hw_consumer)); 791 hw_consumer = le32_to_cpu(*(tx_ring->hw_consumer));
794 792
@@ -805,8 +803,7 @@ static int qlcnic_process_cmd_ring(struct qlcnic_adapter *adapter,
805 PCI_DMA_TODEVICE); 803 PCI_DMA_TODEVICE);
806 frag->dma = 0ULL; 804 frag->dma = 0ULL;
807 } 805 }
808 adapter->stats.xmitfinished++; 806 tx_ring->tx_stats.xmit_finished++;
809 tx_ring->xmit_finished++;
810 dev_kfree_skb_any(buffer->skb); 807 dev_kfree_skb_any(buffer->skb);
811 buffer->skb = NULL; 808 buffer->skb = NULL;
812 } 809 }
@@ -823,8 +820,7 @@ static int qlcnic_process_cmd_ring(struct qlcnic_adapter *adapter,
823 netif_carrier_ok(netdev)) { 820 netif_carrier_ok(netdev)) {
824 if (qlcnic_tx_avail(tx_ring) > TX_STOP_THRESH) { 821 if (qlcnic_tx_avail(tx_ring) > TX_STOP_THRESH) {
825 netif_tx_wake_queue(tx_ring->txq); 822 netif_tx_wake_queue(tx_ring->txq);
826 adapter->stats.xmit_on++; 823 tx_ring->tx_stats.xmit_on++;
827 tx_ring->xmit_on++;
828 } 824 }
829 } 825 }
830 adapter->tx_timeo_cnt = 0; 826 adapter->tx_timeo_cnt = 0;
@@ -844,6 +840,7 @@ static int qlcnic_process_cmd_ring(struct qlcnic_adapter *adapter,
844 */ 840 */
845 hw_consumer = le32_to_cpu(*(tx_ring->hw_consumer)); 841 hw_consumer = le32_to_cpu(*(tx_ring->hw_consumer));
846 done = (sw_consumer == hw_consumer); 842 done = (sw_consumer == hw_consumer);
843 spin_unlock(&adapter->tx_clean_lock);
847 844
848 return done; 845 return done;
849} 846}
@@ -1011,7 +1008,7 @@ static void qlcnic_handle_fw_message(int desc_cnt, int index,
1011 } 1008 }
1012 break; 1009 break;
1013 case QLCNIC_C2H_OPCODE_GET_DCB_AEN: 1010 case QLCNIC_C2H_OPCODE_GET_DCB_AEN:
1014 qlcnic_dcb_handle_aen(adapter, (void *)&msg); 1011 qlcnic_dcb_aen_handler(adapter->dcb, (void *)&msg);
1015 break; 1012 break;
1016 default: 1013 default:
1017 break; 1014 break;
@@ -1463,18 +1460,18 @@ int qlcnic_82xx_napi_add(struct qlcnic_adapter *adapter,
1463 struct qlcnic_recv_context *recv_ctx = adapter->recv_ctx; 1460 struct qlcnic_recv_context *recv_ctx = adapter->recv_ctx;
1464 struct qlcnic_host_tx_ring *tx_ring; 1461 struct qlcnic_host_tx_ring *tx_ring;
1465 1462
1466 if (qlcnic_alloc_sds_rings(recv_ctx, adapter->max_sds_rings)) 1463 if (qlcnic_alloc_sds_rings(recv_ctx, adapter->drv_sds_rings))
1467 return -ENOMEM; 1464 return -ENOMEM;
1468 1465
1469 for (ring = 0; ring < adapter->max_sds_rings; ring++) { 1466 for (ring = 0; ring < adapter->drv_sds_rings; ring++) {
1470 sds_ring = &recv_ctx->sds_rings[ring]; 1467 sds_ring = &recv_ctx->sds_rings[ring];
1471 if (qlcnic_check_multi_tx(adapter) && 1468 if (qlcnic_check_multi_tx(adapter) &&
1472 !adapter->ahw->diag_test && 1469 !adapter->ahw->diag_test &&
1473 (adapter->max_drv_tx_rings > 1)) { 1470 (adapter->drv_tx_rings > QLCNIC_SINGLE_RING)) {
1474 netif_napi_add(netdev, &sds_ring->napi, qlcnic_rx_poll, 1471 netif_napi_add(netdev, &sds_ring->napi, qlcnic_rx_poll,
1475 NAPI_POLL_WEIGHT); 1472 NAPI_POLL_WEIGHT);
1476 } else { 1473 } else {
1477 if (ring == (adapter->max_sds_rings - 1)) 1474 if (ring == (adapter->drv_sds_rings - 1))
1478 netif_napi_add(netdev, &sds_ring->napi, 1475 netif_napi_add(netdev, &sds_ring->napi,
1479 qlcnic_poll, 1476 qlcnic_poll,
1480 NAPI_POLL_WEIGHT); 1477 NAPI_POLL_WEIGHT);
@@ -1491,7 +1488,7 @@ int qlcnic_82xx_napi_add(struct qlcnic_adapter *adapter,
1491 } 1488 }
1492 1489
1493 if (qlcnic_check_multi_tx(adapter) && !adapter->ahw->diag_test) { 1490 if (qlcnic_check_multi_tx(adapter) && !adapter->ahw->diag_test) {
1494 for (ring = 0; ring < adapter->max_drv_tx_rings; ring++) { 1491 for (ring = 0; ring < adapter->drv_tx_rings; ring++) {
1495 tx_ring = &adapter->tx_ring[ring]; 1492 tx_ring = &adapter->tx_ring[ring];
1496 netif_napi_add(netdev, &tx_ring->napi, qlcnic_tx_poll, 1493 netif_napi_add(netdev, &tx_ring->napi, qlcnic_tx_poll,
1497 NAPI_POLL_WEIGHT); 1494 NAPI_POLL_WEIGHT);
@@ -1508,7 +1505,7 @@ void qlcnic_82xx_napi_del(struct qlcnic_adapter *adapter)
1508 struct qlcnic_recv_context *recv_ctx = adapter->recv_ctx; 1505 struct qlcnic_recv_context *recv_ctx = adapter->recv_ctx;
1509 struct qlcnic_host_tx_ring *tx_ring; 1506 struct qlcnic_host_tx_ring *tx_ring;
1510 1507
1511 for (ring = 0; ring < adapter->max_sds_rings; ring++) { 1508 for (ring = 0; ring < adapter->drv_sds_rings; ring++) {
1512 sds_ring = &recv_ctx->sds_rings[ring]; 1509 sds_ring = &recv_ctx->sds_rings[ring];
1513 netif_napi_del(&sds_ring->napi); 1510 netif_napi_del(&sds_ring->napi);
1514 } 1511 }
@@ -1516,7 +1513,7 @@ void qlcnic_82xx_napi_del(struct qlcnic_adapter *adapter)
1516 qlcnic_free_sds_rings(adapter->recv_ctx); 1513 qlcnic_free_sds_rings(adapter->recv_ctx);
1517 1514
1518 if (qlcnic_check_multi_tx(adapter) && !adapter->ahw->diag_test) { 1515 if (qlcnic_check_multi_tx(adapter) && !adapter->ahw->diag_test) {
1519 for (ring = 0; ring < adapter->max_drv_tx_rings; ring++) { 1516 for (ring = 0; ring < adapter->drv_tx_rings; ring++) {
1520 tx_ring = &adapter->tx_ring[ring]; 1517 tx_ring = &adapter->tx_ring[ring];
1521 netif_napi_del(&tx_ring->napi); 1518 netif_napi_del(&tx_ring->napi);
1522 } 1519 }
@@ -1535,7 +1532,7 @@ void qlcnic_82xx_napi_enable(struct qlcnic_adapter *adapter)
1535 if (adapter->is_up != QLCNIC_ADAPTER_UP_MAGIC) 1532 if (adapter->is_up != QLCNIC_ADAPTER_UP_MAGIC)
1536 return; 1533 return;
1537 1534
1538 for (ring = 0; ring < adapter->max_sds_rings; ring++) { 1535 for (ring = 0; ring < adapter->drv_sds_rings; ring++) {
1539 sds_ring = &recv_ctx->sds_rings[ring]; 1536 sds_ring = &recv_ctx->sds_rings[ring];
1540 napi_enable(&sds_ring->napi); 1537 napi_enable(&sds_ring->napi);
1541 qlcnic_enable_int(sds_ring); 1538 qlcnic_enable_int(sds_ring);
@@ -1544,8 +1541,8 @@ void qlcnic_82xx_napi_enable(struct qlcnic_adapter *adapter)
1544 if (qlcnic_check_multi_tx(adapter) && 1541 if (qlcnic_check_multi_tx(adapter) &&
1545 (adapter->flags & QLCNIC_MSIX_ENABLED) && 1542 (adapter->flags & QLCNIC_MSIX_ENABLED) &&
1546 !adapter->ahw->diag_test && 1543 !adapter->ahw->diag_test &&
1547 (adapter->max_drv_tx_rings > 1)) { 1544 (adapter->drv_tx_rings > QLCNIC_SINGLE_RING)) {
1548 for (ring = 0; ring < adapter->max_drv_tx_rings; ring++) { 1545 for (ring = 0; ring < adapter->drv_tx_rings; ring++) {
1549 tx_ring = &adapter->tx_ring[ring]; 1546 tx_ring = &adapter->tx_ring[ring];
1550 napi_enable(&tx_ring->napi); 1547 napi_enable(&tx_ring->napi);
1551 qlcnic_enable_tx_intr(adapter, tx_ring); 1548 qlcnic_enable_tx_intr(adapter, tx_ring);
@@ -1563,7 +1560,7 @@ void qlcnic_82xx_napi_disable(struct qlcnic_adapter *adapter)
1563 if (adapter->is_up != QLCNIC_ADAPTER_UP_MAGIC) 1560 if (adapter->is_up != QLCNIC_ADAPTER_UP_MAGIC)
1564 return; 1561 return;
1565 1562
1566 for (ring = 0; ring < adapter->max_sds_rings; ring++) { 1563 for (ring = 0; ring < adapter->drv_sds_rings; ring++) {
1567 sds_ring = &recv_ctx->sds_rings[ring]; 1564 sds_ring = &recv_ctx->sds_rings[ring];
1568 qlcnic_disable_int(sds_ring); 1565 qlcnic_disable_int(sds_ring);
1569 napi_synchronize(&sds_ring->napi); 1566 napi_synchronize(&sds_ring->napi);
@@ -1573,7 +1570,7 @@ void qlcnic_82xx_napi_disable(struct qlcnic_adapter *adapter)
1573 if ((adapter->flags & QLCNIC_MSIX_ENABLED) && 1570 if ((adapter->flags & QLCNIC_MSIX_ENABLED) &&
1574 !adapter->ahw->diag_test && 1571 !adapter->ahw->diag_test &&
1575 qlcnic_check_multi_tx(adapter)) { 1572 qlcnic_check_multi_tx(adapter)) {
1576 for (ring = 0; ring < adapter->max_drv_tx_rings; ring++) { 1573 for (ring = 0; ring < adapter->drv_tx_rings; ring++) {
1577 tx_ring = &adapter->tx_ring[ring]; 1574 tx_ring = &adapter->tx_ring[ring];
1578 qlcnic_disable_tx_int(adapter, tx_ring); 1575 qlcnic_disable_tx_int(adapter, tx_ring);
1579 napi_synchronize(&tx_ring->napi); 1576 napi_synchronize(&tx_ring->napi);
@@ -1911,7 +1908,7 @@ void qlcnic_83xx_napi_enable(struct qlcnic_adapter *adapter)
1911 if (adapter->is_up != QLCNIC_ADAPTER_UP_MAGIC) 1908 if (adapter->is_up != QLCNIC_ADAPTER_UP_MAGIC)
1912 return; 1909 return;
1913 1910
1914 for (ring = 0; ring < adapter->max_sds_rings; ring++) { 1911 for (ring = 0; ring < adapter->drv_sds_rings; ring++) {
1915 sds_ring = &recv_ctx->sds_rings[ring]; 1912 sds_ring = &recv_ctx->sds_rings[ring];
1916 napi_enable(&sds_ring->napi); 1913 napi_enable(&sds_ring->napi);
1917 if (adapter->flags & QLCNIC_MSIX_ENABLED) 1914 if (adapter->flags & QLCNIC_MSIX_ENABLED)
@@ -1920,7 +1917,7 @@ void qlcnic_83xx_napi_enable(struct qlcnic_adapter *adapter)
1920 1917
1921 if ((adapter->flags & QLCNIC_MSIX_ENABLED) && 1918 if ((adapter->flags & QLCNIC_MSIX_ENABLED) &&
1922 !(adapter->flags & QLCNIC_TX_INTR_SHARED)) { 1919 !(adapter->flags & QLCNIC_TX_INTR_SHARED)) {
1923 for (ring = 0; ring < adapter->max_drv_tx_rings; ring++) { 1920 for (ring = 0; ring < adapter->drv_tx_rings; ring++) {
1924 tx_ring = &adapter->tx_ring[ring]; 1921 tx_ring = &adapter->tx_ring[ring];
1925 napi_enable(&tx_ring->napi); 1922 napi_enable(&tx_ring->napi);
1926 qlcnic_83xx_enable_tx_intr(adapter, tx_ring); 1923 qlcnic_83xx_enable_tx_intr(adapter, tx_ring);
@@ -1938,7 +1935,7 @@ void qlcnic_83xx_napi_disable(struct qlcnic_adapter *adapter)
1938 if (adapter->is_up != QLCNIC_ADAPTER_UP_MAGIC) 1935 if (adapter->is_up != QLCNIC_ADAPTER_UP_MAGIC)
1939 return; 1936 return;
1940 1937
1941 for (ring = 0; ring < adapter->max_sds_rings; ring++) { 1938 for (ring = 0; ring < adapter->drv_sds_rings; ring++) {
1942 sds_ring = &recv_ctx->sds_rings[ring]; 1939 sds_ring = &recv_ctx->sds_rings[ring];
1943 if (adapter->flags & QLCNIC_MSIX_ENABLED) 1940 if (adapter->flags & QLCNIC_MSIX_ENABLED)
1944 qlcnic_83xx_disable_intr(adapter, sds_ring); 1941 qlcnic_83xx_disable_intr(adapter, sds_ring);
@@ -1948,7 +1945,7 @@ void qlcnic_83xx_napi_disable(struct qlcnic_adapter *adapter)
1948 1945
1949 if ((adapter->flags & QLCNIC_MSIX_ENABLED) && 1946 if ((adapter->flags & QLCNIC_MSIX_ENABLED) &&
1950 !(adapter->flags & QLCNIC_TX_INTR_SHARED)) { 1947 !(adapter->flags & QLCNIC_TX_INTR_SHARED)) {
1951 for (ring = 0; ring < adapter->max_drv_tx_rings; ring++) { 1948 for (ring = 0; ring < adapter->drv_tx_rings; ring++) {
1952 tx_ring = &adapter->tx_ring[ring]; 1949 tx_ring = &adapter->tx_ring[ring];
1953 qlcnic_83xx_disable_tx_intr(adapter, tx_ring); 1950 qlcnic_83xx_disable_tx_intr(adapter, tx_ring);
1954 napi_synchronize(&tx_ring->napi); 1951 napi_synchronize(&tx_ring->napi);
@@ -1965,10 +1962,10 @@ int qlcnic_83xx_napi_add(struct qlcnic_adapter *adapter,
1965 struct qlcnic_host_tx_ring *tx_ring; 1962 struct qlcnic_host_tx_ring *tx_ring;
1966 struct qlcnic_recv_context *recv_ctx = adapter->recv_ctx; 1963 struct qlcnic_recv_context *recv_ctx = adapter->recv_ctx;
1967 1964
1968 if (qlcnic_alloc_sds_rings(recv_ctx, adapter->max_sds_rings)) 1965 if (qlcnic_alloc_sds_rings(recv_ctx, adapter->drv_sds_rings))
1969 return -ENOMEM; 1966 return -ENOMEM;
1970 1967
1971 for (ring = 0; ring < adapter->max_sds_rings; ring++) { 1968 for (ring = 0; ring < adapter->drv_sds_rings; ring++) {
1972 sds_ring = &recv_ctx->sds_rings[ring]; 1969 sds_ring = &recv_ctx->sds_rings[ring];
1973 if (adapter->flags & QLCNIC_MSIX_ENABLED) { 1970 if (adapter->flags & QLCNIC_MSIX_ENABLED) {
1974 if (!(adapter->flags & QLCNIC_TX_INTR_SHARED)) 1971 if (!(adapter->flags & QLCNIC_TX_INTR_SHARED))
@@ -1994,7 +1991,7 @@ int qlcnic_83xx_napi_add(struct qlcnic_adapter *adapter,
1994 1991
1995 if ((adapter->flags & QLCNIC_MSIX_ENABLED) && 1992 if ((adapter->flags & QLCNIC_MSIX_ENABLED) &&
1996 !(adapter->flags & QLCNIC_TX_INTR_SHARED)) { 1993 !(adapter->flags & QLCNIC_TX_INTR_SHARED)) {
1997 for (ring = 0; ring < adapter->max_drv_tx_rings; ring++) { 1994 for (ring = 0; ring < adapter->drv_tx_rings; ring++) {
1998 tx_ring = &adapter->tx_ring[ring]; 1995 tx_ring = &adapter->tx_ring[ring];
1999 netif_napi_add(netdev, &tx_ring->napi, 1996 netif_napi_add(netdev, &tx_ring->napi,
2000 qlcnic_83xx_msix_tx_poll, 1997 qlcnic_83xx_msix_tx_poll,
@@ -2012,7 +2009,7 @@ void qlcnic_83xx_napi_del(struct qlcnic_adapter *adapter)
2012 struct qlcnic_recv_context *recv_ctx = adapter->recv_ctx; 2009 struct qlcnic_recv_context *recv_ctx = adapter->recv_ctx;
2013 struct qlcnic_host_tx_ring *tx_ring; 2010 struct qlcnic_host_tx_ring *tx_ring;
2014 2011
2015 for (ring = 0; ring < adapter->max_sds_rings; ring++) { 2012 for (ring = 0; ring < adapter->drv_sds_rings; ring++) {
2016 sds_ring = &recv_ctx->sds_rings[ring]; 2013 sds_ring = &recv_ctx->sds_rings[ring];
2017 netif_napi_del(&sds_ring->napi); 2014 netif_napi_del(&sds_ring->napi);
2018 } 2015 }
@@ -2021,7 +2018,7 @@ void qlcnic_83xx_napi_del(struct qlcnic_adapter *adapter)
2021 2018
2022 if ((adapter->flags & QLCNIC_MSIX_ENABLED) && 2019 if ((adapter->flags & QLCNIC_MSIX_ENABLED) &&
2023 !(adapter->flags & QLCNIC_TX_INTR_SHARED)) { 2020 !(adapter->flags & QLCNIC_TX_INTR_SHARED)) {
2024 for (ring = 0; ring < adapter->max_drv_tx_rings; ring++) { 2021 for (ring = 0; ring < adapter->drv_tx_rings; ring++) {
2025 tx_ring = &adapter->tx_ring[ring]; 2022 tx_ring = &adapter->tx_ring[ring];
2026 netif_napi_del(&tx_ring->napi); 2023 netif_napi_del(&tx_ring->napi);
2027 } 2024 }
diff --git a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c
index d8f4897e9e82..05c1eef8df13 100644
--- a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c
+++ b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c
@@ -548,36 +548,75 @@ static struct qlcnic_hardware_ops qlcnic_hw_ops = {
548 .io_resume = qlcnic_82xx_io_resume, 548 .io_resume = qlcnic_82xx_io_resume,
549}; 549};
550 550
551static void qlcnic_get_multiq_capability(struct qlcnic_adapter *adapter) 551static int qlcnic_check_multi_tx_capability(struct qlcnic_adapter *adapter)
552{ 552{
553 struct qlcnic_hardware_context *ahw = adapter->ahw; 553 struct qlcnic_hardware_context *ahw = adapter->ahw;
554 int num_tx_q;
555 554
556 if (ahw->msix_supported && 555 if (qlcnic_82xx_check(adapter) &&
557 (ahw->extra_capability[0] & QLCNIC_FW_CAPABILITY_2_MULTI_TX)) { 556 (ahw->extra_capability[0] & QLCNIC_FW_CAPABILITY_2_MULTI_TX)) {
558 num_tx_q = min_t(int, QLCNIC_DEF_NUM_TX_RINGS, 557 test_and_set_bit(__QLCNIC_MULTI_TX_UNIQUE, &adapter->state);
559 num_online_cpus()); 558 return 0;
560 if (num_tx_q > 1) {
561 test_and_set_bit(__QLCNIC_MULTI_TX_UNIQUE,
562 &adapter->state);
563 adapter->max_drv_tx_rings = num_tx_q;
564 }
565 } else { 559 } else {
566 adapter->max_drv_tx_rings = 1; 560 return 1;
567 } 561 }
568} 562}
569 563
564static int qlcnic_max_rings(struct qlcnic_adapter *adapter, u8 ring_cnt,
565 int queue_type)
566{
567 int num_rings, max_rings = QLCNIC_MAX_SDS_RINGS;
568
569 if (queue_type == QLCNIC_RX_QUEUE)
570 max_rings = adapter->max_sds_rings;
571 else if (queue_type == QLCNIC_TX_QUEUE)
572 max_rings = adapter->max_tx_rings;
573
574 num_rings = rounddown_pow_of_two(min_t(int, num_online_cpus(),
575 max_rings));
576
577 if (ring_cnt > num_rings)
578 return num_rings;
579 else
580 return ring_cnt;
581}
582
583void qlcnic_set_tx_ring_count(struct qlcnic_adapter *adapter, u8 tx_cnt)
584{
585 /* 83xx adapter does not have max_tx_rings intialized in probe */
586 if (adapter->max_tx_rings)
587 adapter->drv_tx_rings = qlcnic_max_rings(adapter, tx_cnt,
588 QLCNIC_TX_QUEUE);
589 else
590 adapter->drv_tx_rings = tx_cnt;
591
592 dev_info(&adapter->pdev->dev, "Set %d Tx rings\n",
593 adapter->drv_tx_rings);
594}
595
596void qlcnic_set_sds_ring_count(struct qlcnic_adapter *adapter, u8 rx_cnt)
597{
598 /* 83xx adapter does not have max_sds_rings intialized in probe */
599 if (adapter->max_sds_rings)
600 adapter->drv_sds_rings = qlcnic_max_rings(adapter, rx_cnt,
601 QLCNIC_RX_QUEUE);
602 else
603 adapter->drv_sds_rings = rx_cnt;
604
605 dev_info(&adapter->pdev->dev, "Set %d SDS rings\n",
606 adapter->drv_sds_rings);
607}
608
570int qlcnic_enable_msix(struct qlcnic_adapter *adapter, u32 num_msix) 609int qlcnic_enable_msix(struct qlcnic_adapter *adapter, u32 num_msix)
571{ 610{
572 struct pci_dev *pdev = adapter->pdev; 611 struct pci_dev *pdev = adapter->pdev;
573 int max_tx_rings, max_sds_rings, tx_vector; 612 int drv_tx_rings, drv_sds_rings, tx_vector;
574 int err = -1, i; 613 int err = -1, i;
575 614
576 if (adapter->flags & QLCNIC_TX_INTR_SHARED) { 615 if (adapter->flags & QLCNIC_TX_INTR_SHARED) {
577 max_tx_rings = 0; 616 drv_tx_rings = 0;
578 tx_vector = 0; 617 tx_vector = 0;
579 } else { 618 } else {
580 max_tx_rings = adapter->max_drv_tx_rings; 619 drv_tx_rings = adapter->drv_tx_rings;
581 tx_vector = 1; 620 tx_vector = 1;
582 } 621 }
583 622
@@ -589,7 +628,7 @@ int qlcnic_enable_msix(struct qlcnic_adapter *adapter, u32 num_msix)
589 return -ENOMEM; 628 return -ENOMEM;
590 } 629 }
591 630
592 adapter->max_sds_rings = 1; 631 adapter->drv_sds_rings = QLCNIC_SINGLE_RING;
593 adapter->flags &= ~(QLCNIC_MSI_ENABLED | QLCNIC_MSIX_ENABLED); 632 adapter->flags &= ~(QLCNIC_MSI_ENABLED | QLCNIC_MSIX_ENABLED);
594 633
595 if (adapter->ahw->msix_supported) { 634 if (adapter->ahw->msix_supported) {
@@ -602,18 +641,18 @@ int qlcnic_enable_msix(struct qlcnic_adapter *adapter, u32 num_msix)
602 if (qlcnic_83xx_check(adapter)) { 641 if (qlcnic_83xx_check(adapter)) {
603 adapter->ahw->num_msix = num_msix; 642 adapter->ahw->num_msix = num_msix;
604 /* subtract mail box and tx ring vectors */ 643 /* subtract mail box and tx ring vectors */
605 adapter->max_sds_rings = num_msix - 644 adapter->drv_sds_rings = num_msix -
606 max_tx_rings - 1; 645 drv_tx_rings - 1;
607 } else { 646 } else {
608 adapter->ahw->num_msix = num_msix; 647 adapter->ahw->num_msix = num_msix;
609 if (qlcnic_check_multi_tx(adapter) && 648 if (qlcnic_check_multi_tx(adapter) &&
610 !adapter->ahw->diag_test && 649 !adapter->ahw->diag_test &&
611 (adapter->max_drv_tx_rings > 1)) 650 (adapter->drv_tx_rings > 1))
612 max_sds_rings = num_msix - max_tx_rings; 651 drv_sds_rings = num_msix - drv_tx_rings;
613 else 652 else
614 max_sds_rings = num_msix; 653 drv_sds_rings = num_msix;
615 654
616 adapter->max_sds_rings = max_sds_rings; 655 adapter->drv_sds_rings = drv_sds_rings;
617 } 656 }
618 dev_info(&pdev->dev, "using msi-x interrupts\n"); 657 dev_info(&pdev->dev, "using msi-x interrupts\n");
619 return err; 658 return err;
@@ -624,13 +663,13 @@ int qlcnic_enable_msix(struct qlcnic_adapter *adapter, u32 num_msix)
624 if (qlcnic_83xx_check(adapter)) { 663 if (qlcnic_83xx_check(adapter)) {
625 if (err < (QLC_83XX_MINIMUM_VECTOR - tx_vector)) 664 if (err < (QLC_83XX_MINIMUM_VECTOR - tx_vector))
626 return err; 665 return err;
627 err -= (max_tx_rings + 1); 666 err -= drv_tx_rings + 1;
628 num_msix = rounddown_pow_of_two(err); 667 num_msix = rounddown_pow_of_two(err);
629 num_msix += (max_tx_rings + 1); 668 num_msix += drv_tx_rings + 1;
630 } else { 669 } else {
631 num_msix = rounddown_pow_of_two(err); 670 num_msix = rounddown_pow_of_two(err);
632 if (qlcnic_check_multi_tx(adapter)) 671 if (qlcnic_check_multi_tx(adapter))
633 num_msix += max_tx_rings; 672 num_msix += drv_tx_rings;
634 } 673 }
635 674
636 if (num_msix) { 675 if (num_msix) {
@@ -683,25 +722,14 @@ static int qlcnic_enable_msi_legacy(struct qlcnic_adapter *adapter)
683 return err; 722 return err;
684} 723}
685 724
686int qlcnic_82xx_setup_intr(struct qlcnic_adapter *adapter, u8 num_intr, int txq) 725int qlcnic_82xx_setup_intr(struct qlcnic_adapter *adapter)
687{ 726{
688 struct qlcnic_hardware_context *ahw = adapter->ahw;
689 int num_msix, err = 0; 727 int num_msix, err = 0;
690 728
691 if (!num_intr) 729 num_msix = adapter->drv_sds_rings;
692 num_intr = QLCNIC_DEF_NUM_STS_DESC_RINGS;
693 730
694 if (ahw->msix_supported) { 731 if (qlcnic_check_multi_tx(adapter))
695 num_msix = rounddown_pow_of_two(min_t(int, num_online_cpus(), 732 num_msix += adapter->drv_tx_rings;
696 num_intr));
697 if (qlcnic_check_multi_tx(adapter)) {
698 if (txq)
699 adapter->max_drv_tx_rings = txq;
700 num_msix += adapter->max_drv_tx_rings;
701 }
702 } else {
703 num_msix = 1;
704 }
705 733
706 err = qlcnic_enable_msix(adapter, num_msix); 734 err = qlcnic_enable_msix(adapter, num_msix);
707 if (err == -ENOMEM) 735 if (err == -ENOMEM)
@@ -819,7 +847,7 @@ static bool qlcnic_port_eswitch_cfg_capability(struct qlcnic_adapter *adapter)
819int qlcnic_init_pci_info(struct qlcnic_adapter *adapter) 847int qlcnic_init_pci_info(struct qlcnic_adapter *adapter)
820{ 848{
821 struct qlcnic_pci_info *pci_info; 849 struct qlcnic_pci_info *pci_info;
822 int i, ret = 0, j = 0; 850 int i, id = 0, ret = 0, j = 0;
823 u16 act_pci_func; 851 u16 act_pci_func;
824 u8 pfn; 852 u8 pfn;
825 853
@@ -860,7 +888,8 @@ int qlcnic_init_pci_info(struct qlcnic_adapter *adapter)
860 continue; 888 continue;
861 889
862 if (qlcnic_port_eswitch_cfg_capability(adapter)) { 890 if (qlcnic_port_eswitch_cfg_capability(adapter)) {
863 if (!qlcnic_83xx_enable_port_eswitch(adapter, pfn)) 891 if (!qlcnic_83xx_set_port_eswitch_status(adapter, pfn,
892 &id))
864 adapter->npars[j].eswitch_status = true; 893 adapter->npars[j].eswitch_status = true;
865 else 894 else
866 continue; 895 continue;
@@ -875,15 +904,16 @@ int qlcnic_init_pci_info(struct qlcnic_adapter *adapter)
875 adapter->npars[j].min_bw = pci_info[i].tx_min_bw; 904 adapter->npars[j].min_bw = pci_info[i].tx_min_bw;
876 adapter->npars[j].max_bw = pci_info[i].tx_max_bw; 905 adapter->npars[j].max_bw = pci_info[i].tx_max_bw;
877 906
907 memcpy(&adapter->npars[j].mac, &pci_info[i].mac, ETH_ALEN);
878 j++; 908 j++;
879 } 909 }
880 910
881 if (qlcnic_82xx_check(adapter)) { 911 /* Update eSwitch status for adapters without per port eSwitch
912 * configuration capability
913 */
914 if (!qlcnic_port_eswitch_cfg_capability(adapter)) {
882 for (i = 0; i < QLCNIC_NIU_MAX_XG_PORTS; i++) 915 for (i = 0; i < QLCNIC_NIU_MAX_XG_PORTS; i++)
883 adapter->eswitch[i].flags |= QLCNIC_SWITCH_ENABLE; 916 adapter->eswitch[i].flags |= QLCNIC_SWITCH_ENABLE;
884 } else if (!qlcnic_port_eswitch_cfg_capability(adapter)) {
885 for (i = 0; i < QLCNIC_NIU_MAX_XG_PORTS; i++)
886 qlcnic_enable_eswitch(adapter, i, 1);
887 } 917 }
888 918
889 kfree(pci_info); 919 kfree(pci_info);
@@ -1138,14 +1168,18 @@ qlcnic_initialize_nic(struct qlcnic_adapter *adapter)
1138 adapter->ahw->max_mac_filters = nic_info.max_mac_filters; 1168 adapter->ahw->max_mac_filters = nic_info.max_mac_filters;
1139 adapter->ahw->max_mtu = nic_info.max_mtu; 1169 adapter->ahw->max_mtu = nic_info.max_mtu;
1140 1170
1141 /* Disable NPAR for 83XX */ 1171 if (adapter->ahw->capabilities & BIT_6) {
1142 if (qlcnic_83xx_check(adapter))
1143 return err;
1144
1145 if (adapter->ahw->capabilities & BIT_6)
1146 adapter->flags |= QLCNIC_ESWITCH_ENABLED; 1172 adapter->flags |= QLCNIC_ESWITCH_ENABLED;
1147 else 1173 adapter->ahw->nic_mode = QLCNIC_VNIC_MODE;
1174 adapter->max_tx_rings = QLCNIC_MAX_HW_VNIC_TX_RINGS;
1175 adapter->max_sds_rings = QLCNIC_MAX_VNIC_SDS_RINGS;
1176
1177 dev_info(&adapter->pdev->dev, "vNIC mode enabled.\n");
1178 } else {
1179 adapter->ahw->nic_mode = QLCNIC_DEFAULT_MODE;
1180 adapter->max_tx_rings = QLCNIC_MAX_HW_TX_RINGS;
1148 adapter->flags &= ~QLCNIC_ESWITCH_ENABLED; 1181 adapter->flags &= ~QLCNIC_ESWITCH_ENABLED;
1182 }
1149 1183
1150 return err; 1184 return err;
1151} 1185}
@@ -1293,6 +1327,8 @@ qlcnic_check_eswitch_mode(struct qlcnic_adapter *adapter)
1293 "HAL Version: %d, Privileged function\n", 1327 "HAL Version: %d, Privileged function\n",
1294 adapter->ahw->fw_hal_version); 1328 adapter->ahw->fw_hal_version);
1295 } 1329 }
1330 } else {
1331 adapter->ahw->nic_mode = QLCNIC_DEFAULT_MODE;
1296 } 1332 }
1297 1333
1298 adapter->flags |= QLCNIC_ADAPTER_INITIALIZED; 1334 adapter->flags |= QLCNIC_ADAPTER_INITIALIZED;
@@ -1552,7 +1588,7 @@ qlcnic_request_irq(struct qlcnic_adapter *adapter)
1552 if (qlcnic_82xx_check(adapter) || 1588 if (qlcnic_82xx_check(adapter) ||
1553 (qlcnic_83xx_check(adapter) && 1589 (qlcnic_83xx_check(adapter) &&
1554 (adapter->flags & QLCNIC_MSIX_ENABLED))) { 1590 (adapter->flags & QLCNIC_MSIX_ENABLED))) {
1555 num_sds_rings = adapter->max_sds_rings; 1591 num_sds_rings = adapter->drv_sds_rings;
1556 for (ring = 0; ring < num_sds_rings; ring++) { 1592 for (ring = 0; ring < num_sds_rings; ring++) {
1557 sds_ring = &recv_ctx->sds_rings[ring]; 1593 sds_ring = &recv_ctx->sds_rings[ring];
1558 if (qlcnic_82xx_check(adapter) && 1594 if (qlcnic_82xx_check(adapter) &&
@@ -1586,7 +1622,7 @@ qlcnic_request_irq(struct qlcnic_adapter *adapter)
1586 (adapter->flags & QLCNIC_MSIX_ENABLED) && 1622 (adapter->flags & QLCNIC_MSIX_ENABLED) &&
1587 !(adapter->flags & QLCNIC_TX_INTR_SHARED))) { 1623 !(adapter->flags & QLCNIC_TX_INTR_SHARED))) {
1588 handler = qlcnic_msix_tx_intr; 1624 handler = qlcnic_msix_tx_intr;
1589 for (ring = 0; ring < adapter->max_drv_tx_rings; 1625 for (ring = 0; ring < adapter->drv_tx_rings;
1590 ring++) { 1626 ring++) {
1591 tx_ring = &adapter->tx_ring[ring]; 1627 tx_ring = &adapter->tx_ring[ring];
1592 snprintf(tx_ring->name, sizeof(tx_ring->name), 1628 snprintf(tx_ring->name, sizeof(tx_ring->name),
@@ -1614,7 +1650,7 @@ qlcnic_free_irq(struct qlcnic_adapter *adapter)
1614 if (qlcnic_82xx_check(adapter) || 1650 if (qlcnic_82xx_check(adapter) ||
1615 (qlcnic_83xx_check(adapter) && 1651 (qlcnic_83xx_check(adapter) &&
1616 (adapter->flags & QLCNIC_MSIX_ENABLED))) { 1652 (adapter->flags & QLCNIC_MSIX_ENABLED))) {
1617 for (ring = 0; ring < adapter->max_sds_rings; ring++) { 1653 for (ring = 0; ring < adapter->drv_sds_rings; ring++) {
1618 sds_ring = &recv_ctx->sds_rings[ring]; 1654 sds_ring = &recv_ctx->sds_rings[ring];
1619 free_irq(sds_ring->irq, sds_ring); 1655 free_irq(sds_ring->irq, sds_ring);
1620 } 1656 }
@@ -1623,7 +1659,7 @@ qlcnic_free_irq(struct qlcnic_adapter *adapter)
1623 !(adapter->flags & QLCNIC_TX_INTR_SHARED)) || 1659 !(adapter->flags & QLCNIC_TX_INTR_SHARED)) ||
1624 (qlcnic_82xx_check(adapter) && 1660 (qlcnic_82xx_check(adapter) &&
1625 qlcnic_check_multi_tx(adapter))) { 1661 qlcnic_check_multi_tx(adapter))) {
1626 for (ring = 0; ring < adapter->max_drv_tx_rings; 1662 for (ring = 0; ring < adapter->drv_tx_rings;
1627 ring++) { 1663 ring++) {
1628 tx_ring = &adapter->tx_ring[ring]; 1664 tx_ring = &adapter->tx_ring[ring];
1629 if (tx_ring->irq) 1665 if (tx_ring->irq)
@@ -1677,7 +1713,7 @@ int __qlcnic_up(struct qlcnic_adapter *adapter, struct net_device *netdev)
1677 1713
1678 adapter->ahw->linkup = 0; 1714 adapter->ahw->linkup = 0;
1679 1715
1680 if (adapter->max_sds_rings > 1) 1716 if (adapter->drv_sds_rings > 1)
1681 qlcnic_config_rss(adapter, 1); 1717 qlcnic_config_rss(adapter, 1);
1682 1718
1683 qlcnic_config_intr_coalesce(adapter); 1719 qlcnic_config_intr_coalesce(adapter);
@@ -1719,6 +1755,7 @@ void __qlcnic_down(struct qlcnic_adapter *adapter, struct net_device *netdev)
1719 if (qlcnic_sriov_vf_check(adapter)) 1755 if (qlcnic_sriov_vf_check(adapter))
1720 qlcnic_sriov_cleanup_async_list(&adapter->ahw->sriov->bc); 1756 qlcnic_sriov_cleanup_async_list(&adapter->ahw->sriov->bc);
1721 smp_mb(); 1757 smp_mb();
1758 spin_lock(&adapter->tx_clean_lock);
1722 netif_carrier_off(netdev); 1759 netif_carrier_off(netdev);
1723 adapter->ahw->linkup = 0; 1760 adapter->ahw->linkup = 0;
1724 netif_tx_disable(netdev); 1761 netif_tx_disable(netdev);
@@ -1737,8 +1774,9 @@ void __qlcnic_down(struct qlcnic_adapter *adapter, struct net_device *netdev)
1737 1774
1738 qlcnic_reset_rx_buffers_list(adapter); 1775 qlcnic_reset_rx_buffers_list(adapter);
1739 1776
1740 for (ring = 0; ring < adapter->max_drv_tx_rings; ring++) 1777 for (ring = 0; ring < adapter->drv_tx_rings; ring++)
1741 qlcnic_release_tx_buffers(adapter, &adapter->tx_ring[ring]); 1778 qlcnic_release_tx_buffers(adapter, &adapter->tx_ring[ring]);
1779 spin_unlock(&adapter->tx_clean_lock);
1742} 1780}
1743 1781
1744/* Usage: During suspend and firmware recovery module */ 1782/* Usage: During suspend and firmware recovery module */
@@ -1814,16 +1852,16 @@ void qlcnic_detach(struct qlcnic_adapter *adapter)
1814 adapter->is_up = 0; 1852 adapter->is_up = 0;
1815} 1853}
1816 1854
1817void qlcnic_diag_free_res(struct net_device *netdev, int max_sds_rings) 1855void qlcnic_diag_free_res(struct net_device *netdev, int drv_sds_rings)
1818{ 1856{
1819 struct qlcnic_adapter *adapter = netdev_priv(netdev); 1857 struct qlcnic_adapter *adapter = netdev_priv(netdev);
1820 struct qlcnic_host_sds_ring *sds_ring; 1858 struct qlcnic_host_sds_ring *sds_ring;
1821 int max_tx_rings = adapter->max_drv_tx_rings; 1859 int drv_tx_rings = adapter->drv_tx_rings;
1822 int ring; 1860 int ring;
1823 1861
1824 clear_bit(__QLCNIC_DEV_UP, &adapter->state); 1862 clear_bit(__QLCNIC_DEV_UP, &adapter->state);
1825 if (adapter->ahw->diag_test == QLCNIC_INTERRUPT_TEST) { 1863 if (adapter->ahw->diag_test == QLCNIC_INTERRUPT_TEST) {
1826 for (ring = 0; ring < adapter->max_sds_rings; ring++) { 1864 for (ring = 0; ring < adapter->drv_sds_rings; ring++) {
1827 sds_ring = &adapter->recv_ctx->sds_rings[ring]; 1865 sds_ring = &adapter->recv_ctx->sds_rings[ring];
1828 qlcnic_disable_int(sds_ring); 1866 qlcnic_disable_int(sds_ring);
1829 } 1867 }
@@ -1834,8 +1872,8 @@ void qlcnic_diag_free_res(struct net_device *netdev, int max_sds_rings)
1834 qlcnic_detach(adapter); 1872 qlcnic_detach(adapter);
1835 1873
1836 adapter->ahw->diag_test = 0; 1874 adapter->ahw->diag_test = 0;
1837 adapter->max_sds_rings = max_sds_rings; 1875 adapter->drv_sds_rings = drv_sds_rings;
1838 adapter->max_drv_tx_rings = max_tx_rings; 1876 adapter->drv_tx_rings = drv_tx_rings;
1839 1877
1840 if (qlcnic_attach(adapter)) 1878 if (qlcnic_attach(adapter))
1841 goto out; 1879 goto out;
@@ -1901,10 +1939,10 @@ int qlcnic_diag_alloc_res(struct net_device *netdev, int test)
1901 1939
1902 qlcnic_detach(adapter); 1940 qlcnic_detach(adapter);
1903 1941
1904 adapter->max_sds_rings = 1; 1942 adapter->drv_sds_rings = QLCNIC_SINGLE_RING;
1943 adapter->drv_tx_rings = QLCNIC_SINGLE_RING;
1905 adapter->ahw->diag_test = test; 1944 adapter->ahw->diag_test = test;
1906 adapter->ahw->linkup = 0; 1945 adapter->ahw->linkup = 0;
1907 adapter->max_drv_tx_rings = 1;
1908 1946
1909 ret = qlcnic_attach(adapter); 1947 ret = qlcnic_attach(adapter);
1910 if (ret) { 1948 if (ret) {
@@ -1925,7 +1963,7 @@ int qlcnic_diag_alloc_res(struct net_device *netdev, int test)
1925 } 1963 }
1926 1964
1927 if (adapter->ahw->diag_test == QLCNIC_INTERRUPT_TEST) { 1965 if (adapter->ahw->diag_test == QLCNIC_INTERRUPT_TEST) {
1928 for (ring = 0; ring < adapter->max_sds_rings; ring++) { 1966 for (ring = 0; ring < adapter->drv_sds_rings; ring++) {
1929 sds_ring = &adapter->recv_ctx->sds_rings[ring]; 1967 sds_ring = &adapter->recv_ctx->sds_rings[ring];
1930 qlcnic_enable_int(sds_ring); 1968 qlcnic_enable_int(sds_ring);
1931 } 1969 }
@@ -2072,7 +2110,7 @@ qlcnic_setup_netdev(struct qlcnic_adapter *adapter, struct net_device *netdev,
2072 return err; 2110 return err;
2073 } 2111 }
2074 2112
2075 qlcnic_dcb_init_dcbnl_ops(adapter); 2113 qlcnic_dcb_init_dcbnl_ops(adapter->dcb);
2076 2114
2077 return 0; 2115 return 0;
2078} 2116}
@@ -2098,7 +2136,7 @@ void qlcnic_free_tx_rings(struct qlcnic_adapter *adapter)
2098 int ring; 2136 int ring;
2099 struct qlcnic_host_tx_ring *tx_ring; 2137 struct qlcnic_host_tx_ring *tx_ring;
2100 2138
2101 for (ring = 0; ring < adapter->max_drv_tx_rings; ring++) { 2139 for (ring = 0; ring < adapter->drv_tx_rings; ring++) {
2102 tx_ring = &adapter->tx_ring[ring]; 2140 tx_ring = &adapter->tx_ring[ring];
2103 if (tx_ring && tx_ring->cmd_buf_arr != NULL) { 2141 if (tx_ring && tx_ring->cmd_buf_arr != NULL) {
2104 vfree(tx_ring->cmd_buf_arr); 2142 vfree(tx_ring->cmd_buf_arr);
@@ -2116,14 +2154,14 @@ int qlcnic_alloc_tx_rings(struct qlcnic_adapter *adapter,
2116 struct qlcnic_host_tx_ring *tx_ring; 2154 struct qlcnic_host_tx_ring *tx_ring;
2117 struct qlcnic_cmd_buffer *cmd_buf_arr; 2155 struct qlcnic_cmd_buffer *cmd_buf_arr;
2118 2156
2119 tx_ring = kcalloc(adapter->max_drv_tx_rings, 2157 tx_ring = kcalloc(adapter->drv_tx_rings,
2120 sizeof(struct qlcnic_host_tx_ring), GFP_KERNEL); 2158 sizeof(struct qlcnic_host_tx_ring), GFP_KERNEL);
2121 if (tx_ring == NULL) 2159 if (tx_ring == NULL)
2122 return -ENOMEM; 2160 return -ENOMEM;
2123 2161
2124 adapter->tx_ring = tx_ring; 2162 adapter->tx_ring = tx_ring;
2125 2163
2126 for (ring = 0; ring < adapter->max_drv_tx_rings; ring++) { 2164 for (ring = 0; ring < adapter->drv_tx_rings; ring++) {
2127 tx_ring = &adapter->tx_ring[ring]; 2165 tx_ring = &adapter->tx_ring[ring];
2128 tx_ring->num_desc = adapter->num_txd; 2166 tx_ring->num_desc = adapter->num_txd;
2129 tx_ring->txq = netdev_get_tx_queue(netdev, ring); 2167 tx_ring->txq = netdev_get_tx_queue(netdev, ring);
@@ -2138,11 +2176,11 @@ int qlcnic_alloc_tx_rings(struct qlcnic_adapter *adapter,
2138 2176
2139 if (qlcnic_83xx_check(adapter) || 2177 if (qlcnic_83xx_check(adapter) ||
2140 (qlcnic_82xx_check(adapter) && qlcnic_check_multi_tx(adapter))) { 2178 (qlcnic_82xx_check(adapter) && qlcnic_check_multi_tx(adapter))) {
2141 for (ring = 0; ring < adapter->max_drv_tx_rings; ring++) { 2179 for (ring = 0; ring < adapter->drv_tx_rings; ring++) {
2142 tx_ring = &adapter->tx_ring[ring]; 2180 tx_ring = &adapter->tx_ring[ring];
2143 tx_ring->adapter = adapter; 2181 tx_ring->adapter = adapter;
2144 if (adapter->flags & QLCNIC_MSIX_ENABLED) { 2182 if (adapter->flags & QLCNIC_MSIX_ENABLED) {
2145 index = adapter->max_sds_rings + ring; 2183 index = adapter->drv_sds_rings + ring;
2146 vector = adapter->msix_entries[index].vector; 2184 vector = adapter->msix_entries[index].vector;
2147 tx_ring->irq = vector; 2185 tx_ring->irq = vector;
2148 } 2186 }
@@ -2166,17 +2204,6 @@ void qlcnic_set_drv_version(struct qlcnic_adapter *adapter)
2166 qlcnic_fw_cmd_set_drv_version(adapter, fw_cmd); 2204 qlcnic_fw_cmd_set_drv_version(adapter, fw_cmd);
2167} 2205}
2168 2206
2169static int qlcnic_register_dcb(struct qlcnic_adapter *adapter)
2170{
2171 return __qlcnic_register_dcb(adapter);
2172}
2173
2174void qlcnic_clear_dcb_ops(struct qlcnic_adapter *adapter)
2175{
2176 kfree(adapter->dcb);
2177 adapter->dcb = NULL;
2178}
2179
2180static int 2207static int
2181qlcnic_probe(struct pci_dev *pdev, const struct pci_device_id *ent) 2208qlcnic_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
2182{ 2209{
@@ -2185,6 +2212,7 @@ qlcnic_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
2185 struct qlcnic_hardware_context *ahw; 2212 struct qlcnic_hardware_context *ahw;
2186 int err, pci_using_dac = -1; 2213 int err, pci_using_dac = -1;
2187 char board_name[QLCNIC_MAX_BOARD_NAME_LEN + 19]; /* MAC + ": " + name */ 2214 char board_name[QLCNIC_MAX_BOARD_NAME_LEN + 19]; /* MAC + ": " + name */
2215 struct qlcnic_dcb *dcb;
2188 2216
2189 if (pdev->is_virtfn) 2217 if (pdev->is_virtfn)
2190 return -ENODEV; 2218 return -ENODEV;
@@ -2271,6 +2299,7 @@ qlcnic_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
2271 rwlock_init(&adapter->ahw->crb_lock); 2299 rwlock_init(&adapter->ahw->crb_lock);
2272 mutex_init(&adapter->ahw->mem_lock); 2300 mutex_init(&adapter->ahw->mem_lock);
2273 2301
2302 spin_lock_init(&adapter->tx_clean_lock);
2274 INIT_LIST_HEAD(&adapter->mac_list); 2303 INIT_LIST_HEAD(&adapter->mac_list);
2275 2304
2276 qlcnic_register_dcb(adapter); 2305 qlcnic_register_dcb(adapter);
@@ -2285,38 +2314,51 @@ qlcnic_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
2285 goto err_out_maintenance_mode; 2314 goto err_out_maintenance_mode;
2286 } 2315 }
2287 2316
2288 qlcnic_get_multiq_capability(adapter); 2317 /* compute and set default and max tx/sds rings */
2289 2318 if (adapter->ahw->msix_supported) {
2290 if ((adapter->ahw->act_pci_func > 2) && 2319 if (qlcnic_check_multi_tx_capability(adapter) == 1)
2291 qlcnic_check_multi_tx(adapter)) { 2320 qlcnic_set_tx_ring_count(adapter,
2292 adapter->max_drv_tx_rings = QLCNIC_DEF_NUM_TX_RINGS; 2321 QLCNIC_SINGLE_RING);
2293 dev_info(&adapter->pdev->dev, 2322 else
2294 "vNIC mode enabled, Set max TX rings = %d\n", 2323 qlcnic_set_tx_ring_count(adapter,
2295 adapter->max_drv_tx_rings); 2324 QLCNIC_DEF_TX_RINGS);
2325 qlcnic_set_sds_ring_count(adapter,
2326 QLCNIC_DEF_SDS_RINGS);
2327 } else {
2328 qlcnic_set_tx_ring_count(adapter, QLCNIC_SINGLE_RING);
2329 qlcnic_set_sds_ring_count(adapter, QLCNIC_SINGLE_RING);
2296 } 2330 }
2297 2331
2298 if (!qlcnic_check_multi_tx(adapter)) {
2299 clear_bit(__QLCNIC_MULTI_TX_UNIQUE, &adapter->state);
2300 adapter->max_drv_tx_rings = 1;
2301 }
2302 err = qlcnic_setup_idc_param(adapter); 2332 err = qlcnic_setup_idc_param(adapter);
2303 if (err) 2333 if (err)
2304 goto err_out_free_hw; 2334 goto err_out_free_hw;
2305 2335
2306 adapter->flags |= QLCNIC_NEED_FLR; 2336 adapter->flags |= QLCNIC_NEED_FLR;
2307 2337
2308 if (adapter->dcb && qlcnic_dcb_attach(adapter)) 2338 dcb = adapter->dcb;
2309 qlcnic_clear_dcb_ops(adapter);
2310 2339
2340 if (dcb && qlcnic_dcb_attach(dcb))
2341 qlcnic_clear_dcb_ops(dcb);
2311 } else if (qlcnic_83xx_check(adapter)) { 2342 } else if (qlcnic_83xx_check(adapter)) {
2312 adapter->max_drv_tx_rings = 1;
2313 qlcnic_83xx_check_vf(adapter, ent); 2343 qlcnic_83xx_check_vf(adapter, ent);
2314 adapter->portnum = adapter->ahw->pci_func; 2344 adapter->portnum = adapter->ahw->pci_func;
2315 err = qlcnic_83xx_init(adapter, pci_using_dac); 2345 err = qlcnic_83xx_init(adapter, pci_using_dac);
2316 if (err) { 2346 if (err) {
2317 dev_err(&pdev->dev, "%s: failed\n", __func__); 2347 switch (err) {
2318 goto err_out_free_hw; 2348 case -ENOTRECOVERABLE:
2349 dev_err(&pdev->dev, "Adapter initialization failed due to a faulty hardware. Please reboot\n");
2350 dev_err(&pdev->dev, "If reboot doesn't help, please replace the adapter with new one and return the faulty adapter for repair\n");
2351 goto err_out_free_hw;
2352 case -ENOMEM:
2353 dev_err(&pdev->dev, "Adapter initialization failed. Please reboot\n");
2354 goto err_out_free_hw;
2355 default:
2356 dev_err(&pdev->dev, "Adapter initialization failed. A reboot may be required to recover from this failure\n");
2357 dev_err(&pdev->dev, "If reboot does not help to recover from this failure, try a flash update of the adapter\n");
2358 goto err_out_maintenance_mode;
2359 }
2319 } 2360 }
2361
2320 if (qlcnic_sriov_vf_check(adapter)) 2362 if (qlcnic_sriov_vf_check(adapter))
2321 return 0; 2363 return 0;
2322 } else { 2364 } else {
@@ -2344,7 +2386,7 @@ qlcnic_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
2344 "Device does not support MSI interrupts\n"); 2386 "Device does not support MSI interrupts\n");
2345 2387
2346 if (qlcnic_82xx_check(adapter)) { 2388 if (qlcnic_82xx_check(adapter)) {
2347 err = qlcnic_setup_intr(adapter, 0, 0); 2389 err = qlcnic_setup_intr(adapter);
2348 if (err) { 2390 if (err) {
2349 dev_err(&pdev->dev, "Failed to setup interrupt\n"); 2391 dev_err(&pdev->dev, "Failed to setup interrupt\n");
2350 goto err_out_disable_msi; 2392 goto err_out_disable_msi;
@@ -2414,13 +2456,20 @@ err_out_free_res:
2414 pci_release_regions(pdev); 2456 pci_release_regions(pdev);
2415 2457
2416err_out_disable_pdev: 2458err_out_disable_pdev:
2417 pci_set_drvdata(pdev, NULL);
2418 pci_disable_device(pdev); 2459 pci_disable_device(pdev);
2419 return err; 2460 return err;
2420 2461
2421err_out_maintenance_mode: 2462err_out_maintenance_mode:
2463 set_bit(__QLCNIC_MAINTENANCE_MODE, &adapter->state);
2422 netdev->netdev_ops = &qlcnic_netdev_failed_ops; 2464 netdev->netdev_ops = &qlcnic_netdev_failed_ops;
2423 SET_ETHTOOL_OPS(netdev, &qlcnic_ethtool_failed_ops); 2465 SET_ETHTOOL_OPS(netdev, &qlcnic_ethtool_failed_ops);
2466 ahw->port_type = QLCNIC_XGBE;
2467
2468 if (qlcnic_83xx_check(adapter))
2469 adapter->tgt_status_reg = NULL;
2470 else
2471 ahw->board_type = QLCNIC_BRDTYPE_P3P_10G_SFP_PLUS;
2472
2424 err = register_netdev(netdev); 2473 err = register_netdev(netdev);
2425 2474
2426 if (err) { 2475 if (err) {
@@ -2451,7 +2500,7 @@ static void qlcnic_remove(struct pci_dev *pdev)
2451 qlcnic_cancel_idc_work(adapter); 2500 qlcnic_cancel_idc_work(adapter);
2452 ahw = adapter->ahw; 2501 ahw = adapter->ahw;
2453 2502
2454 qlcnic_dcb_free(adapter); 2503 qlcnic_dcb_free(adapter->dcb);
2455 2504
2456 unregister_netdev(netdev); 2505 unregister_netdev(netdev);
2457 qlcnic_sriov_cleanup(adapter); 2506 qlcnic_sriov_cleanup(adapter);
@@ -2490,7 +2539,6 @@ static void qlcnic_remove(struct pci_dev *pdev)
2490 pci_disable_pcie_error_reporting(pdev); 2539 pci_disable_pcie_error_reporting(pdev);
2491 pci_release_regions(pdev); 2540 pci_release_regions(pdev);
2492 pci_disable_device(pdev); 2541 pci_disable_device(pdev);
2493 pci_set_drvdata(pdev, NULL);
2494 2542
2495 if (adapter->qlcnic_wq) { 2543 if (adapter->qlcnic_wq) {
2496 destroy_workqueue(adapter->qlcnic_wq); 2544 destroy_workqueue(adapter->qlcnic_wq);
@@ -2543,12 +2591,11 @@ static int qlcnic_resume(struct pci_dev *pdev)
2543static int qlcnic_open(struct net_device *netdev) 2591static int qlcnic_open(struct net_device *netdev)
2544{ 2592{
2545 struct qlcnic_adapter *adapter = netdev_priv(netdev); 2593 struct qlcnic_adapter *adapter = netdev_priv(netdev);
2546 u32 state;
2547 int err; 2594 int err;
2548 2595
2549 state = QLC_SHARED_REG_RD32(adapter, QLCNIC_CRB_DEV_STATE); 2596 if (test_bit(__QLCNIC_MAINTENANCE_MODE, &adapter->state)) {
2550 if (state == QLCNIC_DEV_FAILED || state == QLCNIC_DEV_BADBAD) { 2597 netdev_err(netdev, "%s: Device is in non-operational state\n",
2551 netdev_err(netdev, "%s: Device is in FAILED state\n", __func__); 2598 __func__);
2552 2599
2553 return -EIO; 2600 return -EIO;
2554 } 2601 }
@@ -2710,24 +2757,21 @@ static void qlcnic_tx_timeout(struct net_device *netdev)
2710 QLCNIC_FORCE_FW_DUMP_KEY); 2757 QLCNIC_FORCE_FW_DUMP_KEY);
2711 } else { 2758 } else {
2712 netdev_info(netdev, "Tx timeout, reset adapter context.\n"); 2759 netdev_info(netdev, "Tx timeout, reset adapter context.\n");
2713 if (qlcnic_82xx_check(adapter)) { 2760 for (ring = 0; ring < adapter->drv_tx_rings; ring++) {
2714 for (ring = 0; ring < adapter->max_drv_tx_rings; 2761 tx_ring = &adapter->tx_ring[ring];
2715 ring++) { 2762 netdev_info(netdev, "Tx ring=%d\n", ring);
2716 tx_ring = &adapter->tx_ring[ring]; 2763 netdev_info(netdev,
2717 dev_info(&netdev->dev, "ring=%d\n", ring); 2764 "crb_intr_mask=%d, producer=%d, sw_consumer=%d, hw_consumer=%d\n",
2718 dev_info(&netdev->dev, "crb_intr_mask=%d\n", 2765 readl(tx_ring->crb_intr_mask),
2719 readl(tx_ring->crb_intr_mask)); 2766 readl(tx_ring->crb_cmd_producer),
2720 dev_info(&netdev->dev, "producer=%d\n", 2767 tx_ring->sw_consumer,
2721 readl(tx_ring->crb_cmd_producer)); 2768 le32_to_cpu(*(tx_ring->hw_consumer)));
2722 dev_info(&netdev->dev, "sw_consumer = %d\n", 2769 netdev_info(netdev,
2723 tx_ring->sw_consumer); 2770 "xmit_finished=%llu, xmit_called=%llu, xmit_on=%llu, xmit_off=%llu\n",
2724 dev_info(&netdev->dev, "hw_consumer = %d\n", 2771 tx_ring->tx_stats.xmit_finished,
2725 le32_to_cpu(*(tx_ring->hw_consumer))); 2772 tx_ring->tx_stats.xmit_called,
2726 dev_info(&netdev->dev, "xmit-on=%llu\n", 2773 tx_ring->tx_stats.xmit_on,
2727 tx_ring->xmit_on); 2774 tx_ring->tx_stats.xmit_off);
2728 dev_info(&netdev->dev, "xmit-off=%llu\n",
2729 tx_ring->xmit_off);
2730 }
2731 } 2775 }
2732 adapter->ahw->reset_context = 1; 2776 adapter->ahw->reset_context = 1;
2733 } 2777 }
@@ -2841,7 +2885,7 @@ static void qlcnic_poll_controller(struct net_device *netdev)
2841 struct qlcnic_recv_context *recv_ctx = adapter->recv_ctx; 2885 struct qlcnic_recv_context *recv_ctx = adapter->recv_ctx;
2842 2886
2843 disable_irq(adapter->irq); 2887 disable_irq(adapter->irq);
2844 for (ring = 0; ring < adapter->max_sds_rings; ring++) { 2888 for (ring = 0; ring < adapter->drv_sds_rings; ring++) {
2845 sds_ring = &recv_ctx->sds_rings[ring]; 2889 sds_ring = &recv_ctx->sds_rings[ring];
2846 qlcnic_intr(adapter->irq, sds_ring); 2890 qlcnic_intr(adapter->irq, sds_ring);
2847 } 2891 }
@@ -3261,8 +3305,9 @@ void qlcnic_82xx_dev_request_reset(struct qlcnic_adapter *adapter, u32 key)
3261 return; 3305 return;
3262 3306
3263 state = QLC_SHARED_REG_RD32(adapter, QLCNIC_CRB_DEV_STATE); 3307 state = QLC_SHARED_REG_RD32(adapter, QLCNIC_CRB_DEV_STATE);
3264 if (state == QLCNIC_DEV_FAILED || state == QLCNIC_DEV_BADBAD) { 3308
3265 netdev_err(adapter->netdev, "%s: Device is in FAILED state\n", 3309 if (test_bit(__QLCNIC_MAINTENANCE_MODE, &adapter->state)) {
3310 netdev_err(adapter->netdev, "%s: Device is in non-operational state\n",
3266 __func__); 3311 __func__);
3267 qlcnic_api_unlock(adapter); 3312 qlcnic_api_unlock(adapter);
3268 3313
@@ -3329,7 +3374,7 @@ qlcnic_attach_work(struct work_struct *work)
3329 return; 3374 return;
3330 } 3375 }
3331attach: 3376attach:
3332 qlcnic_dcb_get_info(adapter); 3377 qlcnic_dcb_get_info(adapter->dcb);
3333 3378
3334 if (netif_running(netdev)) { 3379 if (netif_running(netdev)) {
3335 if (qlcnic_up(adapter, netdev)) 3380 if (qlcnic_up(adapter, netdev))
@@ -3354,6 +3399,8 @@ done:
3354static int 3399static int
3355qlcnic_check_health(struct qlcnic_adapter *adapter) 3400qlcnic_check_health(struct qlcnic_adapter *adapter)
3356{ 3401{
3402 struct qlcnic_hardware_context *ahw = adapter->ahw;
3403 struct qlcnic_fw_dump *fw_dump = &ahw->fw_dump;
3357 u32 state = 0, heartbeat; 3404 u32 state = 0, heartbeat;
3358 u32 peg_status; 3405 u32 peg_status;
3359 int err = 0; 3406 int err = 0;
@@ -3378,7 +3425,7 @@ qlcnic_check_health(struct qlcnic_adapter *adapter)
3378 if (adapter->need_fw_reset) 3425 if (adapter->need_fw_reset)
3379 goto detach; 3426 goto detach;
3380 3427
3381 if (adapter->ahw->reset_context && qlcnic_auto_fw_reset) 3428 if (ahw->reset_context && qlcnic_auto_fw_reset)
3382 qlcnic_reset_hw_context(adapter); 3429 qlcnic_reset_hw_context(adapter);
3383 3430
3384 return 0; 3431 return 0;
@@ -3421,6 +3468,9 @@ detach:
3421 3468
3422 qlcnic_schedule_work(adapter, qlcnic_detach_work, 0); 3469 qlcnic_schedule_work(adapter, qlcnic_detach_work, 0);
3423 QLCDB(adapter, DRV, "fw recovery scheduled.\n"); 3470 QLCDB(adapter, DRV, "fw recovery scheduled.\n");
3471 } else if (!qlcnic_auto_fw_reset && fw_dump->enable &&
3472 adapter->flags & QLCNIC_FW_RESET_OWNER) {
3473 qlcnic_dump_fw(adapter);
3424 } 3474 }
3425 3475
3426 return 1; 3476 return 1;
@@ -3502,7 +3552,7 @@ static int qlcnic_attach_func(struct pci_dev *pdev)
3502 qlcnic_clr_drv_state(adapter); 3552 qlcnic_clr_drv_state(adapter);
3503 kfree(adapter->msix_entries); 3553 kfree(adapter->msix_entries);
3504 adapter->msix_entries = NULL; 3554 adapter->msix_entries = NULL;
3505 err = qlcnic_setup_intr(adapter, 0, 0); 3555 err = qlcnic_setup_intr(adapter);
3506 3556
3507 if (err) { 3557 if (err) {
3508 kfree(adapter->msix_entries); 3558 kfree(adapter->msix_entries);
@@ -3647,130 +3697,90 @@ qlcnicvf_start_firmware(struct qlcnic_adapter *adapter)
3647 return err; 3697 return err;
3648} 3698}
3649 3699
3650int qlcnic_validate_max_tx_rings(struct qlcnic_adapter *adapter, u32 txq) 3700int qlcnic_validate_rings(struct qlcnic_adapter *adapter, __u32 ring_cnt,
3701 int queue_type)
3651{ 3702{
3652 struct net_device *netdev = adapter->netdev; 3703 struct net_device *netdev = adapter->netdev;
3653 u8 max_hw = QLCNIC_MAX_TX_RINGS; 3704 u8 max_hw_rings = 0;
3654 u32 max_allowed; 3705 char buf[8];
3706 int cur_rings;
3655 3707
3656 if (!qlcnic_use_msi_x && !qlcnic_use_msi) { 3708 if (queue_type == QLCNIC_RX_QUEUE) {
3657 netdev_err(netdev, "No Multi TX-Q support in INT-x mode\n"); 3709 max_hw_rings = adapter->max_sds_rings;
3658 return -EINVAL; 3710 cur_rings = adapter->drv_sds_rings;
3711 strcpy(buf, "SDS");
3712 } else if (queue_type == QLCNIC_TX_QUEUE) {
3713 max_hw_rings = adapter->max_tx_rings;
3714 cur_rings = adapter->drv_tx_rings;
3715 strcpy(buf, "Tx");
3659 } 3716 }
3660 3717
3661 if (!qlcnic_check_multi_tx(adapter)) { 3718 if (!qlcnic_use_msi_x && !qlcnic_use_msi) {
3662 netdev_err(netdev, "No Multi TX-Q support\n"); 3719 netdev_err(netdev, "No RSS/TSS support in INT-x mode\n");
3663 return -EINVAL; 3720 return -EINVAL;
3664 } 3721 }
3665 3722
3666 if (txq > QLCNIC_MAX_TX_RINGS) { 3723 if (adapter->flags & QLCNIC_MSI_ENABLED) {
3667 netdev_err(netdev, "Invalid ring count\n"); 3724 netdev_err(netdev, "No RSS/TSS support in MSI mode\n");
3668 return -EINVAL; 3725 return -EINVAL;
3669 } 3726 }
3670 3727
3671 max_allowed = rounddown_pow_of_two(min_t(int, max_hw, 3728 if (ring_cnt < 2) {
3672 num_online_cpus())); 3729 netdev_err(netdev,
3673 if ((txq > max_allowed) || !is_power_of_2(txq)) { 3730 "%s rings value should not be lower than 2\n", buf);
3674 if (!is_power_of_2(txq))
3675 netdev_err(netdev,
3676 "TX queue should be a power of 2\n");
3677 if (txq > num_online_cpus())
3678 netdev_err(netdev,
3679 "Tx queue should not be higher than [%u], number of online CPUs in the system\n",
3680 num_online_cpus());
3681 netdev_err(netdev, "Unable to configure %u Tx rings\n", txq);
3682 return -EINVAL; 3731 return -EINVAL;
3683 } 3732 }
3684 3733
3685 return 0; 3734 if (!is_power_of_2(ring_cnt)) {
3686} 3735 netdev_err(netdev, "%s rings value should be a power of 2\n",
3687 3736 buf);
3688int qlcnic_validate_max_rss(struct qlcnic_adapter *adapter,
3689 __u32 val)
3690{
3691 struct net_device *netdev = adapter->netdev;
3692 u8 max_hw = adapter->ahw->max_rx_ques;
3693 u32 max_allowed;
3694
3695 if (!qlcnic_use_msi_x && !qlcnic_use_msi) {
3696 netdev_err(netdev, "No RSS support in INT-x mode\n");
3697 return -EINVAL; 3737 return -EINVAL;
3698 } 3738 }
3699 3739
3700 if (val > QLCNIC_MAX_SDS_RINGS) { 3740 if (qlcnic_82xx_check(adapter) && (queue_type == QLCNIC_TX_QUEUE) &&
3701 netdev_err(netdev, "RSS value should not be higher than %u\n", 3741 !qlcnic_check_multi_tx(adapter)) {
3702 QLCNIC_MAX_SDS_RINGS); 3742 netdev_err(netdev, "No Multi Tx queue support\n");
3703 return -EINVAL; 3743 return -EINVAL;
3704 } 3744 }
3705 3745
3706 max_allowed = rounddown_pow_of_two(min_t(int, max_hw, 3746 if (ring_cnt > num_online_cpus()) {
3707 num_online_cpus())); 3747 netdev_err(netdev,
3708 if ((val > max_allowed) || (val < 2) || !is_power_of_2(val)) { 3748 "%s value[%u] should not be higher than, number of online CPUs\n",
3709 if (!is_power_of_2(val)) 3749 buf, num_online_cpus());
3710 netdev_err(netdev, "RSS value should be a power of 2\n");
3711
3712 if (val < 2)
3713 netdev_err(netdev, "RSS value should not be lower than 2\n");
3714
3715 if (val > max_hw)
3716 netdev_err(netdev,
3717 "RSS value should not be higher than[%u], the max RSS rings supported by the adapter\n",
3718 max_hw);
3719
3720 if (val > num_online_cpus())
3721 netdev_err(netdev,
3722 "RSS value should not be higher than[%u], number of online CPUs in the system\n",
3723 num_online_cpus());
3724
3725 netdev_err(netdev, "Unable to configure %u RSS rings\n", val);
3726
3727 return -EINVAL; 3750 return -EINVAL;
3728 } 3751 }
3752
3729 return 0; 3753 return 0;
3730} 3754}
3731 3755
3732int qlcnic_set_max_rss(struct qlcnic_adapter *adapter, u8 data, int txq) 3756int qlcnic_setup_rings(struct qlcnic_adapter *adapter, u8 rx_cnt, u8 tx_cnt)
3733{ 3757{
3734 int err;
3735 struct net_device *netdev = adapter->netdev; 3758 struct net_device *netdev = adapter->netdev;
3736 int num_msix; 3759 int err;
3737 3760
3738 if (test_bit(__QLCNIC_RESETTING, &adapter->state)) 3761 if (test_bit(__QLCNIC_RESETTING, &adapter->state))
3739 return -EBUSY; 3762 return -EBUSY;
3740 3763
3741 if (qlcnic_82xx_check(adapter) && !qlcnic_use_msi_x &&
3742 !qlcnic_use_msi) {
3743 netdev_err(netdev, "No RSS support in INT-x mode\n");
3744 return -EINVAL;
3745 }
3746
3747 netif_device_detach(netdev); 3764 netif_device_detach(netdev);
3748 if (netif_running(netdev)) 3765 if (netif_running(netdev))
3749 __qlcnic_down(adapter, netdev); 3766 __qlcnic_down(adapter, netdev);
3750 3767
3751 qlcnic_detach(adapter); 3768 qlcnic_detach(adapter);
3752 3769
3753 if (qlcnic_82xx_check(adapter)) {
3754 if (txq != 0)
3755 adapter->max_drv_tx_rings = txq;
3756
3757 if (qlcnic_check_multi_tx(adapter) &&
3758 (txq > adapter->max_drv_tx_rings))
3759 num_msix = adapter->max_drv_tx_rings;
3760 else
3761 num_msix = data;
3762 }
3763
3764 if (qlcnic_83xx_check(adapter)) { 3770 if (qlcnic_83xx_check(adapter)) {
3765 qlcnic_83xx_free_mbx_intr(adapter); 3771 qlcnic_83xx_free_mbx_intr(adapter);
3766 qlcnic_83xx_enable_mbx_poll(adapter); 3772 qlcnic_83xx_enable_mbx_poll(adapter);
3767 } 3773 }
3768 3774
3769 netif_set_real_num_tx_queues(netdev, adapter->max_drv_tx_rings);
3770
3771 qlcnic_teardown_intr(adapter); 3775 qlcnic_teardown_intr(adapter);
3772 3776
3773 err = qlcnic_setup_intr(adapter, data, txq); 3777 /* compute and set default and max tx/sds rings */
3778 qlcnic_set_tx_ring_count(adapter, tx_cnt);
3779 qlcnic_set_sds_ring_count(adapter, rx_cnt);
3780
3781 netif_set_real_num_tx_queues(netdev, adapter->drv_tx_rings);
3782
3783 err = qlcnic_setup_intr(adapter);
3774 if (err) { 3784 if (err) {
3775 kfree(adapter->msix_entries); 3785 kfree(adapter->msix_entries);
3776 netdev_err(netdev, "failed to setup interrupt\n"); 3786 netdev_err(netdev, "failed to setup interrupt\n");
diff --git a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
index 15513608d480..7763962e2ec4 100644
--- a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
+++ b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
@@ -1187,41 +1187,38 @@ int qlcnic_dump_fw(struct qlcnic_adapter *adapter)
1187 } 1187 }
1188 1188
1189 if (ops_index == ops_cnt) { 1189 if (ops_index == ops_cnt) {
1190 dev_info(&adapter->pdev->dev, 1190 dev_info(dev, "Skipping unknown entry opcode %d\n",
1191 "Invalid entry type %d, exiting dump\n",
1192 entry->hdr.type); 1191 entry->hdr.type);
1193 goto error; 1192 entry->hdr.flags |= QLCNIC_DUMP_SKIP;
1193 entry_offset += entry->hdr.offset;
1194 continue;
1194 } 1195 }
1195 1196
1196 /* Collect dump for this entry */ 1197 /* Collect dump for this entry */
1197 dump = fw_dump_ops[ops_index].handler(adapter, entry, buffer); 1198 dump = fw_dump_ops[ops_index].handler(adapter, entry, buffer);
1198 if (!qlcnic_valid_dump_entry(&adapter->pdev->dev, entry, dump)) 1199 if (!qlcnic_valid_dump_entry(dev, entry, dump)) {
1199 entry->hdr.flags |= QLCNIC_DUMP_SKIP; 1200 entry->hdr.flags |= QLCNIC_DUMP_SKIP;
1201 entry_offset += entry->hdr.offset;
1202 continue;
1203 }
1204
1200 buf_offset += entry->hdr.cap_size; 1205 buf_offset += entry->hdr.cap_size;
1201 entry_offset += entry->hdr.offset; 1206 entry_offset += entry->hdr.offset;
1202 buffer = fw_dump->data + buf_offset; 1207 buffer = fw_dump->data + buf_offset;
1203 } 1208 }
1204 if (dump_size != buf_offset) { 1209
1205 dev_info(&adapter->pdev->dev, 1210 fw_dump->clr = 1;
1206 "Captured(%d) and expected size(%d) do not match\n", 1211 snprintf(mesg, sizeof(mesg), "FW_DUMP=%s", adapter->netdev->name);
1207 buf_offset, dump_size); 1212 dev_info(dev, "%s: Dump data %d bytes captured, template header size %d bytes\n",
1208 goto error; 1213 adapter->netdev->name, fw_dump->size, tmpl_hdr->size);
1209 } else { 1214 /* Send a udev event to notify availability of FW dump */
1210 fw_dump->clr = 1; 1215 kobject_uevent_env(&dev->kobj, KOBJ_CHANGE, msg);
1211 snprintf(mesg, sizeof(mesg), "FW_DUMP=%s", 1216
1212 adapter->netdev->name);
1213 dev_info(&adapter->pdev->dev, "%s: Dump data, %d bytes captured\n",
1214 adapter->netdev->name, fw_dump->size);
1215 /* Send a udev event to notify availability of FW dump */
1216 kobject_uevent_env(&adapter->pdev->dev.kobj, KOBJ_CHANGE, msg);
1217 return 0;
1218 }
1219error:
1220 if (fw_dump->use_pex_dma) 1217 if (fw_dump->use_pex_dma)
1221 dma_free_coherent(dev, QLC_PEX_DMA_READ_SIZE, 1218 dma_free_coherent(dev, QLC_PEX_DMA_READ_SIZE,
1222 fw_dump->dma_buffer, fw_dump->phys_addr); 1219 fw_dump->dma_buffer, fw_dump->phys_addr);
1223 vfree(fw_dump->data); 1220
1224 return -EINVAL; 1221 return 0;
1225} 1222}
1226 1223
1227void qlcnic_83xx_get_minidump_template(struct qlcnic_adapter *adapter) 1224void qlcnic_83xx_get_minidump_template(struct qlcnic_adapter *adapter)
diff --git a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c
index 392b9bd12b4f..21a4b274d2e4 100644
--- a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c
+++ b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c
@@ -500,6 +500,7 @@ static int qlcnic_sriov_vf_init_driver(struct qlcnic_adapter *adapter)
500static int qlcnic_sriov_setup_vf(struct qlcnic_adapter *adapter, 500static int qlcnic_sriov_setup_vf(struct qlcnic_adapter *adapter,
501 int pci_using_dac) 501 int pci_using_dac)
502{ 502{
503 struct qlcnic_dcb *dcb;
503 int err; 504 int err;
504 505
505 INIT_LIST_HEAD(&adapter->vf_mc_list); 506 INIT_LIST_HEAD(&adapter->vf_mc_list);
@@ -507,7 +508,11 @@ static int qlcnic_sriov_setup_vf(struct qlcnic_adapter *adapter,
507 dev_warn(&adapter->pdev->dev, 508 dev_warn(&adapter->pdev->dev,
508 "Device does not support MSI interrupts\n"); 509 "Device does not support MSI interrupts\n");
509 510
510 err = qlcnic_setup_intr(adapter, 1, 0); 511 /* compute and set default and max tx/sds rings */
512 qlcnic_set_tx_ring_count(adapter, QLCNIC_SINGLE_RING);
513 qlcnic_set_sds_ring_count(adapter, QLCNIC_SINGLE_RING);
514
515 err = qlcnic_setup_intr(adapter);
511 if (err) { 516 if (err) {
512 dev_err(&adapter->pdev->dev, "Failed to setup interrupt\n"); 517 dev_err(&adapter->pdev->dev, "Failed to setup interrupt\n");
513 goto err_out_disable_msi; 518 goto err_out_disable_msi;
@@ -533,8 +538,10 @@ static int qlcnic_sriov_setup_vf(struct qlcnic_adapter *adapter,
533 if (err) 538 if (err)
534 goto err_out_send_channel_term; 539 goto err_out_send_channel_term;
535 540
536 if (adapter->dcb && qlcnic_dcb_attach(adapter)) 541 dcb = adapter->dcb;
537 qlcnic_clear_dcb_ops(adapter); 542
543 if (dcb && qlcnic_dcb_attach(dcb))
544 qlcnic_clear_dcb_ops(dcb);
538 545
539 err = qlcnic_setup_netdev(adapter, adapter->netdev, pci_using_dac); 546 err = qlcnic_setup_netdev(adapter, adapter->netdev, pci_using_dac);
540 if (err) 547 if (err)
@@ -1577,7 +1584,7 @@ static int qlcnic_sriov_vf_reinit_driver(struct qlcnic_adapter *adapter)
1577 if (err) 1584 if (err)
1578 goto err_out_term_channel; 1585 goto err_out_term_channel;
1579 1586
1580 qlcnic_dcb_get_info(adapter); 1587 qlcnic_dcb_get_info(adapter->dcb);
1581 1588
1582 return 0; 1589 return 0;
1583 1590
diff --git a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_sysfs.c b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_sysfs.c
index 019f4377307f..1a9f8a400e50 100644
--- a/drivers/net/ethernet/qlogic/qlcnic/qlcnic_sysfs.c
+++ b/drivers/net/ethernet/qlogic/qlcnic/qlcnic_sysfs.c
@@ -156,7 +156,7 @@ static int qlcnic_82xx_store_beacon(struct qlcnic_adapter *adapter,
156 const char *buf, size_t len) 156 const char *buf, size_t len)
157{ 157{
158 struct qlcnic_hardware_context *ahw = adapter->ahw; 158 struct qlcnic_hardware_context *ahw = adapter->ahw;
159 int err, max_sds_rings = adapter->max_sds_rings; 159 int err, drv_sds_rings = adapter->drv_sds_rings;
160 u16 beacon; 160 u16 beacon;
161 u8 h_beacon_state, b_state, b_rate; 161 u8 h_beacon_state, b_state, b_rate;
162 162
@@ -211,7 +211,7 @@ static int qlcnic_82xx_store_beacon(struct qlcnic_adapter *adapter,
211 } 211 }
212 212
213 if (test_and_clear_bit(__QLCNIC_DIAG_RES_ALLOC, &adapter->state)) 213 if (test_and_clear_bit(__QLCNIC_DIAG_RES_ALLOC, &adapter->state))
214 qlcnic_diag_free_res(adapter->netdev, max_sds_rings); 214 qlcnic_diag_free_res(adapter->netdev, drv_sds_rings);
215 215
216out: 216out:
217 if (!ahw->beacon_state) 217 if (!ahw->beacon_state)
@@ -1272,7 +1272,6 @@ void qlcnic_remove_sysfs_entries(struct qlcnic_adapter *adapter)
1272void qlcnic_create_diag_entries(struct qlcnic_adapter *adapter) 1272void qlcnic_create_diag_entries(struct qlcnic_adapter *adapter)
1273{ 1273{
1274 struct device *dev = &adapter->pdev->dev; 1274 struct device *dev = &adapter->pdev->dev;
1275 u32 state;
1276 1275
1277 if (device_create_bin_file(dev, &bin_attr_port_stats)) 1276 if (device_create_bin_file(dev, &bin_attr_port_stats))
1278 dev_info(dev, "failed to create port stats sysfs entry"); 1277 dev_info(dev, "failed to create port stats sysfs entry");
@@ -1286,8 +1285,7 @@ void qlcnic_create_diag_entries(struct qlcnic_adapter *adapter)
1286 if (device_create_bin_file(dev, &bin_attr_mem)) 1285 if (device_create_bin_file(dev, &bin_attr_mem))
1287 dev_info(dev, "failed to create mem sysfs entry\n"); 1286 dev_info(dev, "failed to create mem sysfs entry\n");
1288 1287
1289 state = QLC_SHARED_REG_RD32(adapter, QLCNIC_CRB_DEV_STATE); 1288 if (test_bit(__QLCNIC_MAINTENANCE_MODE, &adapter->state))
1290 if (state == QLCNIC_DEV_FAILED || state == QLCNIC_DEV_BADBAD)
1291 return; 1289 return;
1292 1290
1293 if (device_create_bin_file(dev, &bin_attr_pci_config)) 1291 if (device_create_bin_file(dev, &bin_attr_pci_config))
@@ -1313,7 +1311,6 @@ void qlcnic_create_diag_entries(struct qlcnic_adapter *adapter)
1313void qlcnic_remove_diag_entries(struct qlcnic_adapter *adapter) 1311void qlcnic_remove_diag_entries(struct qlcnic_adapter *adapter)
1314{ 1312{
1315 struct device *dev = &adapter->pdev->dev; 1313 struct device *dev = &adapter->pdev->dev;
1316 u32 state;
1317 1314
1318 device_remove_bin_file(dev, &bin_attr_port_stats); 1315 device_remove_bin_file(dev, &bin_attr_port_stats);
1319 1316
@@ -1323,8 +1320,7 @@ void qlcnic_remove_diag_entries(struct qlcnic_adapter *adapter)
1323 device_remove_bin_file(dev, &bin_attr_crb); 1320 device_remove_bin_file(dev, &bin_attr_crb);
1324 device_remove_bin_file(dev, &bin_attr_mem); 1321 device_remove_bin_file(dev, &bin_attr_mem);
1325 1322
1326 state = QLC_SHARED_REG_RD32(adapter, QLCNIC_CRB_DEV_STATE); 1323 if (test_bit(__QLCNIC_MAINTENANCE_MODE, &adapter->state))
1327 if (state == QLCNIC_DEV_FAILED || state == QLCNIC_DEV_BADBAD)
1328 return; 1324 return;
1329 1325
1330 device_remove_bin_file(dev, &bin_attr_pci_config); 1326 device_remove_bin_file(dev, &bin_attr_pci_config);
diff --git a/drivers/net/ethernet/qlogic/qlge/qlge.h b/drivers/net/ethernet/qlogic/qlge/qlge.h
index 899433778466..0c9c4e895595 100644
--- a/drivers/net/ethernet/qlogic/qlge/qlge.h
+++ b/drivers/net/ethernet/qlogic/qlge/qlge.h
@@ -18,7 +18,7 @@
18 */ 18 */
19#define DRV_NAME "qlge" 19#define DRV_NAME "qlge"
20#define DRV_STRING "QLogic 10 Gigabit PCI-E Ethernet Driver " 20#define DRV_STRING "QLogic 10 Gigabit PCI-E Ethernet Driver "
21#define DRV_VERSION "v1.00.00.32" 21#define DRV_VERSION "1.00.00.33"
22 22
23#define WQ_ADDR_ALIGN 0x3 /* 4 byte alignment */ 23#define WQ_ADDR_ALIGN 0x3 /* 4 byte alignment */
24 24
@@ -2206,14 +2206,14 @@ extern char qlge_driver_name[];
2206extern const char qlge_driver_version[]; 2206extern const char qlge_driver_version[];
2207extern const struct ethtool_ops qlge_ethtool_ops; 2207extern const struct ethtool_ops qlge_ethtool_ops;
2208 2208
2209extern int ql_sem_spinlock(struct ql_adapter *qdev, u32 sem_mask); 2209int ql_sem_spinlock(struct ql_adapter *qdev, u32 sem_mask);
2210extern void ql_sem_unlock(struct ql_adapter *qdev, u32 sem_mask); 2210void ql_sem_unlock(struct ql_adapter *qdev, u32 sem_mask);
2211extern int ql_read_xgmac_reg(struct ql_adapter *qdev, u32 reg, u32 *data); 2211int ql_read_xgmac_reg(struct ql_adapter *qdev, u32 reg, u32 *data);
2212extern int ql_get_mac_addr_reg(struct ql_adapter *qdev, u32 type, u16 index, 2212int ql_get_mac_addr_reg(struct ql_adapter *qdev, u32 type, u16 index,
2213 u32 *value); 2213 u32 *value);
2214extern int ql_get_routing_reg(struct ql_adapter *qdev, u32 index, u32 *value); 2214int ql_get_routing_reg(struct ql_adapter *qdev, u32 index, u32 *value);
2215extern int ql_write_cfg(struct ql_adapter *qdev, void *ptr, int size, u32 bit, 2215int ql_write_cfg(struct ql_adapter *qdev, void *ptr, int size, u32 bit,
2216 u16 q_id); 2216 u16 q_id);
2217void ql_queue_fw_error(struct ql_adapter *qdev); 2217void ql_queue_fw_error(struct ql_adapter *qdev);
2218void ql_mpi_work(struct work_struct *work); 2218void ql_mpi_work(struct work_struct *work);
2219void ql_mpi_reset_work(struct work_struct *work); 2219void ql_mpi_reset_work(struct work_struct *work);
@@ -2233,10 +2233,9 @@ int ql_unpause_mpi_risc(struct ql_adapter *qdev);
2233int ql_pause_mpi_risc(struct ql_adapter *qdev); 2233int ql_pause_mpi_risc(struct ql_adapter *qdev);
2234int ql_hard_reset_mpi_risc(struct ql_adapter *qdev); 2234int ql_hard_reset_mpi_risc(struct ql_adapter *qdev);
2235int ql_soft_reset_mpi_risc(struct ql_adapter *qdev); 2235int ql_soft_reset_mpi_risc(struct ql_adapter *qdev);
2236int ql_dump_risc_ram_area(struct ql_adapter *qdev, void *buf, 2236int ql_dump_risc_ram_area(struct ql_adapter *qdev, void *buf, u32 ram_addr,
2237 u32 ram_addr, int word_count); 2237 int word_count);
2238int ql_core_dump(struct ql_adapter *qdev, 2238int ql_core_dump(struct ql_adapter *qdev, struct ql_mpi_coredump *mpi_coredump);
2239 struct ql_mpi_coredump *mpi_coredump);
2240int ql_mb_about_fw(struct ql_adapter *qdev); 2239int ql_mb_about_fw(struct ql_adapter *qdev);
2241int ql_mb_wol_set_magic(struct ql_adapter *qdev, u32 enable_wol); 2240int ql_mb_wol_set_magic(struct ql_adapter *qdev, u32 enable_wol);
2242int ql_mb_wol_mode(struct ql_adapter *qdev, u32 wol); 2241int ql_mb_wol_mode(struct ql_adapter *qdev, u32 wol);
@@ -2249,8 +2248,7 @@ int ql_mb_get_port_cfg(struct ql_adapter *qdev);
2249int ql_mb_set_port_cfg(struct ql_adapter *qdev); 2248int ql_mb_set_port_cfg(struct ql_adapter *qdev);
2250int ql_wait_fifo_empty(struct ql_adapter *qdev); 2249int ql_wait_fifo_empty(struct ql_adapter *qdev);
2251void ql_get_dump(struct ql_adapter *qdev, void *buff); 2250void ql_get_dump(struct ql_adapter *qdev, void *buff);
2252void ql_gen_reg_dump(struct ql_adapter *qdev, 2251void ql_gen_reg_dump(struct ql_adapter *qdev, struct ql_reg_dump *mpi_coredump);
2253 struct ql_reg_dump *mpi_coredump);
2254netdev_tx_t ql_lb_send(struct sk_buff *skb, struct net_device *ndev); 2252netdev_tx_t ql_lb_send(struct sk_buff *skb, struct net_device *ndev);
2255void ql_check_lb_frame(struct ql_adapter *, struct sk_buff *); 2253void ql_check_lb_frame(struct ql_adapter *, struct sk_buff *);
2256int ql_own_firmware(struct ql_adapter *qdev); 2254int ql_own_firmware(struct ql_adapter *qdev);
@@ -2264,9 +2262,9 @@ int ql_clean_lb_rx_ring(struct rx_ring *rx_ring, int budget);
2264/* #define QL_OB_DUMP */ 2262/* #define QL_OB_DUMP */
2265 2263
2266#ifdef QL_REG_DUMP 2264#ifdef QL_REG_DUMP
2267extern void ql_dump_xgmac_control_regs(struct ql_adapter *qdev); 2265void ql_dump_xgmac_control_regs(struct ql_adapter *qdev);
2268extern void ql_dump_routing_entries(struct ql_adapter *qdev); 2266void ql_dump_routing_entries(struct ql_adapter *qdev);
2269extern void ql_dump_regs(struct ql_adapter *qdev); 2267void ql_dump_regs(struct ql_adapter *qdev);
2270#define QL_DUMP_REGS(qdev) ql_dump_regs(qdev) 2268#define QL_DUMP_REGS(qdev) ql_dump_regs(qdev)
2271#define QL_DUMP_ROUTE(qdev) ql_dump_routing_entries(qdev) 2269#define QL_DUMP_ROUTE(qdev) ql_dump_routing_entries(qdev)
2272#define QL_DUMP_XGMAC_CONTROL_REGS(qdev) ql_dump_xgmac_control_regs(qdev) 2270#define QL_DUMP_XGMAC_CONTROL_REGS(qdev) ql_dump_xgmac_control_regs(qdev)
@@ -2277,26 +2275,26 @@ extern void ql_dump_regs(struct ql_adapter *qdev);
2277#endif 2275#endif
2278 2276
2279#ifdef QL_STAT_DUMP 2277#ifdef QL_STAT_DUMP
2280extern void ql_dump_stat(struct ql_adapter *qdev); 2278void ql_dump_stat(struct ql_adapter *qdev);
2281#define QL_DUMP_STAT(qdev) ql_dump_stat(qdev) 2279#define QL_DUMP_STAT(qdev) ql_dump_stat(qdev)
2282#else 2280#else
2283#define QL_DUMP_STAT(qdev) 2281#define QL_DUMP_STAT(qdev)
2284#endif 2282#endif
2285 2283
2286#ifdef QL_DEV_DUMP 2284#ifdef QL_DEV_DUMP
2287extern void ql_dump_qdev(struct ql_adapter *qdev); 2285void ql_dump_qdev(struct ql_adapter *qdev);
2288#define QL_DUMP_QDEV(qdev) ql_dump_qdev(qdev) 2286#define QL_DUMP_QDEV(qdev) ql_dump_qdev(qdev)
2289#else 2287#else
2290#define QL_DUMP_QDEV(qdev) 2288#define QL_DUMP_QDEV(qdev)
2291#endif 2289#endif
2292 2290
2293#ifdef QL_CB_DUMP 2291#ifdef QL_CB_DUMP
2294extern void ql_dump_wqicb(struct wqicb *wqicb); 2292void ql_dump_wqicb(struct wqicb *wqicb);
2295extern void ql_dump_tx_ring(struct tx_ring *tx_ring); 2293void ql_dump_tx_ring(struct tx_ring *tx_ring);
2296extern void ql_dump_ricb(struct ricb *ricb); 2294void ql_dump_ricb(struct ricb *ricb);
2297extern void ql_dump_cqicb(struct cqicb *cqicb); 2295void ql_dump_cqicb(struct cqicb *cqicb);
2298extern void ql_dump_rx_ring(struct rx_ring *rx_ring); 2296void ql_dump_rx_ring(struct rx_ring *rx_ring);
2299extern void ql_dump_hw_cb(struct ql_adapter *qdev, int size, u32 bit, u16 q_id); 2297void ql_dump_hw_cb(struct ql_adapter *qdev, int size, u32 bit, u16 q_id);
2300#define QL_DUMP_RICB(ricb) ql_dump_ricb(ricb) 2298#define QL_DUMP_RICB(ricb) ql_dump_ricb(ricb)
2301#define QL_DUMP_WQICB(wqicb) ql_dump_wqicb(wqicb) 2299#define QL_DUMP_WQICB(wqicb) ql_dump_wqicb(wqicb)
2302#define QL_DUMP_TX_RING(tx_ring) ql_dump_tx_ring(tx_ring) 2300#define QL_DUMP_TX_RING(tx_ring) ql_dump_tx_ring(tx_ring)
@@ -2314,9 +2312,9 @@ extern void ql_dump_hw_cb(struct ql_adapter *qdev, int size, u32 bit, u16 q_id);
2314#endif 2312#endif
2315 2313
2316#ifdef QL_OB_DUMP 2314#ifdef QL_OB_DUMP
2317extern void ql_dump_tx_desc(struct tx_buf_desc *tbd); 2315void ql_dump_tx_desc(struct tx_buf_desc *tbd);
2318extern void ql_dump_ob_mac_iocb(struct ob_mac_iocb_req *ob_mac_iocb); 2316void ql_dump_ob_mac_iocb(struct ob_mac_iocb_req *ob_mac_iocb);
2319extern void ql_dump_ob_mac_rsp(struct ob_mac_iocb_rsp *ob_mac_rsp); 2317void ql_dump_ob_mac_rsp(struct ob_mac_iocb_rsp *ob_mac_rsp);
2320#define QL_DUMP_OB_MAC_IOCB(ob_mac_iocb) ql_dump_ob_mac_iocb(ob_mac_iocb) 2318#define QL_DUMP_OB_MAC_IOCB(ob_mac_iocb) ql_dump_ob_mac_iocb(ob_mac_iocb)
2321#define QL_DUMP_OB_MAC_RSP(ob_mac_rsp) ql_dump_ob_mac_rsp(ob_mac_rsp) 2319#define QL_DUMP_OB_MAC_RSP(ob_mac_rsp) ql_dump_ob_mac_rsp(ob_mac_rsp)
2322#else 2320#else
@@ -2325,14 +2323,14 @@ extern void ql_dump_ob_mac_rsp(struct ob_mac_iocb_rsp *ob_mac_rsp);
2325#endif 2323#endif
2326 2324
2327#ifdef QL_IB_DUMP 2325#ifdef QL_IB_DUMP
2328extern void ql_dump_ib_mac_rsp(struct ib_mac_iocb_rsp *ib_mac_rsp); 2326void ql_dump_ib_mac_rsp(struct ib_mac_iocb_rsp *ib_mac_rsp);
2329#define QL_DUMP_IB_MAC_RSP(ib_mac_rsp) ql_dump_ib_mac_rsp(ib_mac_rsp) 2327#define QL_DUMP_IB_MAC_RSP(ib_mac_rsp) ql_dump_ib_mac_rsp(ib_mac_rsp)
2330#else 2328#else
2331#define QL_DUMP_IB_MAC_RSP(ib_mac_rsp) 2329#define QL_DUMP_IB_MAC_RSP(ib_mac_rsp)
2332#endif 2330#endif
2333 2331
2334#ifdef QL_ALL_DUMP 2332#ifdef QL_ALL_DUMP
2335extern void ql_dump_all(struct ql_adapter *qdev); 2333void ql_dump_all(struct ql_adapter *qdev);
2336#define QL_DUMP_ALL(qdev) ql_dump_all(qdev) 2334#define QL_DUMP_ALL(qdev) ql_dump_all(qdev)
2337#else 2335#else
2338#define QL_DUMP_ALL(qdev) 2336#define QL_DUMP_ALL(qdev)
diff --git a/drivers/net/ethernet/qlogic/qlge/qlge_main.c b/drivers/net/ethernet/qlogic/qlge/qlge_main.c
index 2553cf4503b9..a245dc18d769 100644
--- a/drivers/net/ethernet/qlogic/qlge/qlge_main.c
+++ b/drivers/net/ethernet/qlogic/qlge/qlge_main.c
@@ -96,8 +96,10 @@ static DEFINE_PCI_DEVICE_TABLE(qlge_pci_tbl) = {
96 96
97MODULE_DEVICE_TABLE(pci, qlge_pci_tbl); 97MODULE_DEVICE_TABLE(pci, qlge_pci_tbl);
98 98
99static int ql_wol(struct ql_adapter *qdev); 99static int ql_wol(struct ql_adapter *);
100static void qlge_set_multicast_list(struct net_device *ndev); 100static void qlge_set_multicast_list(struct net_device *);
101static int ql_adapter_down(struct ql_adapter *);
102static int ql_adapter_up(struct ql_adapter *);
101 103
102/* This hardware semaphore causes exclusive access to 104/* This hardware semaphore causes exclusive access to
103 * resources shared between the NIC driver, MPI firmware, 105 * resources shared between the NIC driver, MPI firmware,
@@ -1464,6 +1466,29 @@ static void ql_categorize_rx_err(struct ql_adapter *qdev, u8 rx_err,
1464 } 1466 }
1465} 1467}
1466 1468
1469/**
1470 * ql_update_mac_hdr_len - helper routine to update the mac header length
1471 * based on vlan tags if present
1472 */
1473static void ql_update_mac_hdr_len(struct ql_adapter *qdev,
1474 struct ib_mac_iocb_rsp *ib_mac_rsp,
1475 void *page, size_t *len)
1476{
1477 u16 *tags;
1478
1479 if (qdev->ndev->features & NETIF_F_HW_VLAN_CTAG_RX)
1480 return;
1481 if (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_V) {
1482 tags = (u16 *)page;
1483 /* Look for stacked vlan tags in ethertype field */
1484 if (tags[6] == ETH_P_8021Q &&
1485 tags[8] == ETH_P_8021Q)
1486 *len += 2 * VLAN_HLEN;
1487 else
1488 *len += VLAN_HLEN;
1489 }
1490}
1491
1467/* Process an inbound completion from an rx ring. */ 1492/* Process an inbound completion from an rx ring. */
1468static void ql_process_mac_rx_gro_page(struct ql_adapter *qdev, 1493static void ql_process_mac_rx_gro_page(struct ql_adapter *qdev,
1469 struct rx_ring *rx_ring, 1494 struct rx_ring *rx_ring,
@@ -1523,6 +1548,7 @@ static void ql_process_mac_rx_page(struct ql_adapter *qdev,
1523 void *addr; 1548 void *addr;
1524 struct bq_desc *lbq_desc = ql_get_curr_lchunk(qdev, rx_ring); 1549 struct bq_desc *lbq_desc = ql_get_curr_lchunk(qdev, rx_ring);
1525 struct napi_struct *napi = &rx_ring->napi; 1550 struct napi_struct *napi = &rx_ring->napi;
1551 size_t hlen = ETH_HLEN;
1526 1552
1527 skb = netdev_alloc_skb(ndev, length); 1553 skb = netdev_alloc_skb(ndev, length);
1528 if (!skb) { 1554 if (!skb) {
@@ -1540,25 +1566,28 @@ static void ql_process_mac_rx_page(struct ql_adapter *qdev,
1540 goto err_out; 1566 goto err_out;
1541 } 1567 }
1542 1568
1569 /* Update the MAC header length*/
1570 ql_update_mac_hdr_len(qdev, ib_mac_rsp, addr, &hlen);
1571
1543 /* The max framesize filter on this chip is set higher than 1572 /* The max framesize filter on this chip is set higher than
1544 * MTU since FCoE uses 2k frames. 1573 * MTU since FCoE uses 2k frames.
1545 */ 1574 */
1546 if (skb->len > ndev->mtu + ETH_HLEN) { 1575 if (skb->len > ndev->mtu + hlen) {
1547 netif_err(qdev, drv, qdev->ndev, 1576 netif_err(qdev, drv, qdev->ndev,
1548 "Segment too small, dropping.\n"); 1577 "Segment too small, dropping.\n");
1549 rx_ring->rx_dropped++; 1578 rx_ring->rx_dropped++;
1550 goto err_out; 1579 goto err_out;
1551 } 1580 }
1552 memcpy(skb_put(skb, ETH_HLEN), addr, ETH_HLEN); 1581 memcpy(skb_put(skb, hlen), addr, hlen);
1553 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev, 1582 netif_printk(qdev, rx_status, KERN_DEBUG, qdev->ndev,
1554 "%d bytes of headers and data in large. Chain page to new skb and pull tail.\n", 1583 "%d bytes of headers and data in large. Chain page to new skb and pull tail.\n",
1555 length); 1584 length);
1556 skb_fill_page_desc(skb, 0, lbq_desc->p.pg_chunk.page, 1585 skb_fill_page_desc(skb, 0, lbq_desc->p.pg_chunk.page,
1557 lbq_desc->p.pg_chunk.offset+ETH_HLEN, 1586 lbq_desc->p.pg_chunk.offset + hlen,
1558 length-ETH_HLEN); 1587 length - hlen);
1559 skb->len += length-ETH_HLEN; 1588 skb->len += length - hlen;
1560 skb->data_len += length-ETH_HLEN; 1589 skb->data_len += length - hlen;
1561 skb->truesize += length-ETH_HLEN; 1590 skb->truesize += length - hlen;
1562 1591
1563 rx_ring->rx_packets++; 1592 rx_ring->rx_packets++;
1564 rx_ring->rx_bytes += skb->len; 1593 rx_ring->rx_bytes += skb->len;
@@ -1576,7 +1605,7 @@ static void ql_process_mac_rx_page(struct ql_adapter *qdev,
1576 (ib_mac_rsp->flags3 & IB_MAC_IOCB_RSP_V4)) { 1605 (ib_mac_rsp->flags3 & IB_MAC_IOCB_RSP_V4)) {
1577 /* Unfragmented ipv4 UDP frame. */ 1606 /* Unfragmented ipv4 UDP frame. */
1578 struct iphdr *iph = 1607 struct iphdr *iph =
1579 (struct iphdr *) ((u8 *)addr + ETH_HLEN); 1608 (struct iphdr *)((u8 *)addr + hlen);
1580 if (!(iph->frag_off & 1609 if (!(iph->frag_off &
1581 htons(IP_MF|IP_OFFSET))) { 1610 htons(IP_MF|IP_OFFSET))) {
1582 skb->ip_summed = CHECKSUM_UNNECESSARY; 1611 skb->ip_summed = CHECKSUM_UNNECESSARY;
@@ -1726,7 +1755,8 @@ static struct sk_buff *ql_build_rx_skb(struct ql_adapter *qdev,
1726 struct bq_desc *sbq_desc; 1755 struct bq_desc *sbq_desc;
1727 struct sk_buff *skb = NULL; 1756 struct sk_buff *skb = NULL;
1728 u32 length = le32_to_cpu(ib_mac_rsp->data_len); 1757 u32 length = le32_to_cpu(ib_mac_rsp->data_len);
1729 u32 hdr_len = le32_to_cpu(ib_mac_rsp->hdr_len); 1758 u32 hdr_len = le32_to_cpu(ib_mac_rsp->hdr_len);
1759 size_t hlen = ETH_HLEN;
1730 1760
1731 /* 1761 /*
1732 * Handle the header buffer if present. 1762 * Handle the header buffer if present.
@@ -1853,9 +1883,10 @@ static struct sk_buff *ql_build_rx_skb(struct ql_adapter *qdev,
1853 skb->data_len += length; 1883 skb->data_len += length;
1854 skb->truesize += length; 1884 skb->truesize += length;
1855 length -= length; 1885 length -= length;
1856 __pskb_pull_tail(skb, 1886 ql_update_mac_hdr_len(qdev, ib_mac_rsp,
1857 (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_V) ? 1887 lbq_desc->p.pg_chunk.va,
1858 VLAN_ETH_HLEN : ETH_HLEN); 1888 &hlen);
1889 __pskb_pull_tail(skb, hlen);
1859 } 1890 }
1860 } else { 1891 } else {
1861 /* 1892 /*
@@ -1910,8 +1941,9 @@ static struct sk_buff *ql_build_rx_skb(struct ql_adapter *qdev,
1910 length -= size; 1941 length -= size;
1911 i++; 1942 i++;
1912 } 1943 }
1913 __pskb_pull_tail(skb, (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_V) ? 1944 ql_update_mac_hdr_len(qdev, ib_mac_rsp, lbq_desc->p.pg_chunk.va,
1914 VLAN_ETH_HLEN : ETH_HLEN); 1945 &hlen);
1946 __pskb_pull_tail(skb, hlen);
1915 } 1947 }
1916 return skb; 1948 return skb;
1917} 1949}
@@ -2003,7 +2035,7 @@ static void ql_process_mac_split_rx_intr(struct ql_adapter *qdev,
2003 rx_ring->rx_packets++; 2035 rx_ring->rx_packets++;
2004 rx_ring->rx_bytes += skb->len; 2036 rx_ring->rx_bytes += skb->len;
2005 skb_record_rx_queue(skb, rx_ring->cq_id); 2037 skb_record_rx_queue(skb, rx_ring->cq_id);
2006 if ((ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_V) && (vlan_id != 0)) 2038 if (vlan_id != 0xffff)
2007 __vlan_hwaccel_put_tag(skb, htons(ETH_P_8021Q), vlan_id); 2039 __vlan_hwaccel_put_tag(skb, htons(ETH_P_8021Q), vlan_id);
2008 if (skb->ip_summed == CHECKSUM_UNNECESSARY) 2040 if (skb->ip_summed == CHECKSUM_UNNECESSARY)
2009 napi_gro_receive(&rx_ring->napi, skb); 2041 napi_gro_receive(&rx_ring->napi, skb);
@@ -2017,7 +2049,8 @@ static unsigned long ql_process_mac_rx_intr(struct ql_adapter *qdev,
2017 struct ib_mac_iocb_rsp *ib_mac_rsp) 2049 struct ib_mac_iocb_rsp *ib_mac_rsp)
2018{ 2050{
2019 u32 length = le32_to_cpu(ib_mac_rsp->data_len); 2051 u32 length = le32_to_cpu(ib_mac_rsp->data_len);
2020 u16 vlan_id = (ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_V) ? 2052 u16 vlan_id = ((ib_mac_rsp->flags2 & IB_MAC_IOCB_RSP_V) &&
2053 (qdev->ndev->features & NETIF_F_HW_VLAN_CTAG_RX)) ?
2021 ((le16_to_cpu(ib_mac_rsp->vlan_id) & 2054 ((le16_to_cpu(ib_mac_rsp->vlan_id) &
2022 IB_MAC_IOCB_RSP_VLAN_MASK)) : 0xffff; 2055 IB_MAC_IOCB_RSP_VLAN_MASK)) : 0xffff;
2023 2056
@@ -2310,9 +2343,39 @@ static void qlge_vlan_mode(struct net_device *ndev, netdev_features_t features)
2310 } 2343 }
2311} 2344}
2312 2345
2346/**
2347 * qlge_update_hw_vlan_features - helper routine to reinitialize the adapter
2348 * based on the features to enable/disable hardware vlan accel
2349 */
2350static int qlge_update_hw_vlan_features(struct net_device *ndev,
2351 netdev_features_t features)
2352{
2353 struct ql_adapter *qdev = netdev_priv(ndev);
2354 int status = 0;
2355
2356 status = ql_adapter_down(qdev);
2357 if (status) {
2358 netif_err(qdev, link, qdev->ndev,
2359 "Failed to bring down the adapter\n");
2360 return status;
2361 }
2362
2363 /* update the features with resent change */
2364 ndev->features = features;
2365
2366 status = ql_adapter_up(qdev);
2367 if (status) {
2368 netif_err(qdev, link, qdev->ndev,
2369 "Failed to bring up the adapter\n");
2370 return status;
2371 }
2372 return status;
2373}
2374
2313static netdev_features_t qlge_fix_features(struct net_device *ndev, 2375static netdev_features_t qlge_fix_features(struct net_device *ndev,
2314 netdev_features_t features) 2376 netdev_features_t features)
2315{ 2377{
2378 int err;
2316 /* 2379 /*
2317 * Since there is no support for separate rx/tx vlan accel 2380 * Since there is no support for separate rx/tx vlan accel
2318 * enable/disable make sure tx flag is always in same state as rx. 2381 * enable/disable make sure tx flag is always in same state as rx.
@@ -2322,6 +2385,11 @@ static netdev_features_t qlge_fix_features(struct net_device *ndev,
2322 else 2385 else
2323 features &= ~NETIF_F_HW_VLAN_CTAG_TX; 2386 features &= ~NETIF_F_HW_VLAN_CTAG_TX;
2324 2387
2388 /* Update the behavior of vlan accel in the adapter */
2389 err = qlge_update_hw_vlan_features(ndev, features);
2390 if (err)
2391 return err;
2392
2325 return features; 2393 return features;
2326} 2394}
2327 2395
@@ -3704,8 +3772,12 @@ static int ql_adapter_initialize(struct ql_adapter *qdev)
3704 ql_write32(qdev, SYS, mask | value); 3772 ql_write32(qdev, SYS, mask | value);
3705 3773
3706 /* Set the default queue, and VLAN behavior. */ 3774 /* Set the default queue, and VLAN behavior. */
3707 value = NIC_RCV_CFG_DFQ | NIC_RCV_CFG_RV; 3775 value = NIC_RCV_CFG_DFQ;
3708 mask = NIC_RCV_CFG_DFQ_MASK | (NIC_RCV_CFG_RV << 16); 3776 mask = NIC_RCV_CFG_DFQ_MASK;
3777 if (qdev->ndev->features & NETIF_F_HW_VLAN_CTAG_RX) {
3778 value |= NIC_RCV_CFG_RV;
3779 mask |= (NIC_RCV_CFG_RV << 16);
3780 }
3709 ql_write32(qdev, NIC_RCV_CFG, (mask | value)); 3781 ql_write32(qdev, NIC_RCV_CFG, (mask | value));
3710 3782
3711 /* Set the MPI interrupt to enabled. */ 3783 /* Set the MPI interrupt to enabled. */
@@ -4505,7 +4577,6 @@ static void ql_release_all(struct pci_dev *pdev)
4505 iounmap(qdev->doorbell_area); 4577 iounmap(qdev->doorbell_area);
4506 vfree(qdev->mpi_coredump); 4578 vfree(qdev->mpi_coredump);
4507 pci_release_regions(pdev); 4579 pci_release_regions(pdev);
4508 pci_set_drvdata(pdev, NULL);
4509} 4580}
4510 4581
4511static int ql_init_device(struct pci_dev *pdev, struct net_device *ndev, 4582static int ql_init_device(struct pci_dev *pdev, struct net_device *ndev,
@@ -4692,11 +4763,15 @@ static int qlge_probe(struct pci_dev *pdev,
4692 4763
4693 qdev = netdev_priv(ndev); 4764 qdev = netdev_priv(ndev);
4694 SET_NETDEV_DEV(ndev, &pdev->dev); 4765 SET_NETDEV_DEV(ndev, &pdev->dev);
4695 ndev->hw_features = NETIF_F_SG | NETIF_F_IP_CSUM | 4766 ndev->hw_features = NETIF_F_SG |
4696 NETIF_F_TSO | NETIF_F_TSO_ECN | 4767 NETIF_F_IP_CSUM |
4697 NETIF_F_HW_VLAN_CTAG_TX | NETIF_F_RXCSUM; 4768 NETIF_F_TSO |
4698 ndev->features = ndev->hw_features | 4769 NETIF_F_TSO_ECN |
4699 NETIF_F_HW_VLAN_CTAG_RX | NETIF_F_HW_VLAN_CTAG_FILTER; 4770 NETIF_F_HW_VLAN_CTAG_TX |
4771 NETIF_F_HW_VLAN_CTAG_RX |
4772 NETIF_F_HW_VLAN_CTAG_FILTER |
4773 NETIF_F_RXCSUM;
4774 ndev->features = ndev->hw_features;
4700 ndev->vlan_features = ndev->hw_features; 4775 ndev->vlan_features = ndev->hw_features;
4701 4776
4702 if (test_bit(QL_DMA64, &qdev->flags)) 4777 if (test_bit(QL_DMA64, &qdev->flags))
diff --git a/drivers/net/ethernet/rdc/r6040.c b/drivers/net/ethernet/rdc/r6040.c
index e9dc84943cfc..1e49ec5b2232 100644
--- a/drivers/net/ethernet/rdc/r6040.c
+++ b/drivers/net/ethernet/rdc/r6040.c
@@ -1231,7 +1231,6 @@ err_out_mdio:
1231 mdiobus_free(lp->mii_bus); 1231 mdiobus_free(lp->mii_bus);
1232err_out_unmap: 1232err_out_unmap:
1233 netif_napi_del(&lp->napi); 1233 netif_napi_del(&lp->napi);
1234 pci_set_drvdata(pdev, NULL);
1235 pci_iounmap(pdev, ioaddr); 1234 pci_iounmap(pdev, ioaddr);
1236err_out_free_res: 1235err_out_free_res:
1237 pci_release_regions(pdev); 1236 pci_release_regions(pdev);
@@ -1257,7 +1256,6 @@ static void r6040_remove_one(struct pci_dev *pdev)
1257 pci_release_regions(pdev); 1256 pci_release_regions(pdev);
1258 free_netdev(dev); 1257 free_netdev(dev);
1259 pci_disable_device(pdev); 1258 pci_disable_device(pdev);
1260 pci_set_drvdata(pdev, NULL);
1261} 1259}
1262 1260
1263 1261
diff --git a/drivers/net/ethernet/realtek/8139cp.c b/drivers/net/ethernet/realtek/8139cp.c
index d2e591955bdd..f2a2128165dd 100644
--- a/drivers/net/ethernet/realtek/8139cp.c
+++ b/drivers/net/ethernet/realtek/8139cp.c
@@ -2052,7 +2052,6 @@ static void cp_remove_one (struct pci_dev *pdev)
2052 pci_release_regions(pdev); 2052 pci_release_regions(pdev);
2053 pci_clear_mwi(pdev); 2053 pci_clear_mwi(pdev);
2054 pci_disable_device(pdev); 2054 pci_disable_device(pdev);
2055 pci_set_drvdata(pdev, NULL);
2056 free_netdev(dev); 2055 free_netdev(dev);
2057} 2056}
2058 2057
diff --git a/drivers/net/ethernet/realtek/8139too.c b/drivers/net/ethernet/realtek/8139too.c
index 3ccedeb8aba0..50a92104dd0a 100644
--- a/drivers/net/ethernet/realtek/8139too.c
+++ b/drivers/net/ethernet/realtek/8139too.c
@@ -727,7 +727,6 @@ static void __rtl8139_cleanup_dev (struct net_device *dev)
727 pci_release_regions (pdev); 727 pci_release_regions (pdev);
728 728
729 free_netdev(dev); 729 free_netdev(dev);
730 pci_set_drvdata (pdev, NULL);
731} 730}
732 731
733 732
diff --git a/drivers/net/ethernet/realtek/r8169.c b/drivers/net/ethernet/realtek/r8169.c
index 3397cee89777..799387570766 100644
--- a/drivers/net/ethernet/realtek/r8169.c
+++ b/drivers/net/ethernet/realtek/r8169.c
@@ -6811,7 +6811,6 @@ static void rtl_remove_one(struct pci_dev *pdev)
6811 6811
6812 rtl_disable_msi(pdev, tp); 6812 rtl_disable_msi(pdev, tp);
6813 rtl8169_release_board(pdev, dev, tp->mmio_addr); 6813 rtl8169_release_board(pdev, dev, tp->mmio_addr);
6814 pci_set_drvdata(pdev, NULL);
6815} 6814}
6816 6815
6817static const struct net_device_ops rtl_netdev_ops = { 6816static const struct net_device_ops rtl_netdev_ops = {
diff --git a/drivers/net/ethernet/renesas/sh_eth.c b/drivers/net/ethernet/renesas/sh_eth.c
index b57c278d3b46..d256ce19d4de 100644
--- a/drivers/net/ethernet/renesas/sh_eth.c
+++ b/drivers/net/ethernet/renesas/sh_eth.c
@@ -483,7 +483,7 @@ static struct sh_eth_cpu_data sh7757_data = {
483 .register_type = SH_ETH_REG_FAST_SH4, 483 .register_type = SH_ETH_REG_FAST_SH4,
484 484
485 .eesipr_value = DMAC_M_RFRMER | DMAC_M_ECI | 0x003fffff, 485 .eesipr_value = DMAC_M_RFRMER | DMAC_M_ECI | 0x003fffff,
486 .rmcr_value = 0x00000001, 486 .rmcr_value = RMCR_RNC,
487 487
488 .tx_check = EESR_FTC | EESR_CND | EESR_DLC | EESR_CD | EESR_RTO, 488 .tx_check = EESR_FTC | EESR_CND | EESR_DLC | EESR_CD | EESR_RTO,
489 .eesr_err_check = EESR_TWB | EESR_TABT | EESR_RABT | EESR_RFE | 489 .eesr_err_check = EESR_TWB | EESR_TABT | EESR_RABT | EESR_RFE |
@@ -561,7 +561,7 @@ static struct sh_eth_cpu_data sh7757_data_giga = {
561 EESR_RFE | EESR_RDE | EESR_RFRMER | EESR_TFE | 561 EESR_RFE | EESR_RDE | EESR_RFRMER | EESR_TFE |
562 EESR_TDE | EESR_ECI, 562 EESR_TDE | EESR_ECI,
563 .fdr_value = 0x0000072f, 563 .fdr_value = 0x0000072f,
564 .rmcr_value = 0x00000001, 564 .rmcr_value = RMCR_RNC,
565 565
566 .irq_flags = IRQF_SHARED, 566 .irq_flags = IRQF_SHARED,
567 .apr = 1, 567 .apr = 1,
@@ -689,7 +689,7 @@ static struct sh_eth_cpu_data r8a7740_data = {
689 EESR_RFE | EESR_RDE | EESR_RFRMER | EESR_TFE | 689 EESR_RFE | EESR_RDE | EESR_RFRMER | EESR_TFE |
690 EESR_TDE | EESR_ECI, 690 EESR_TDE | EESR_ECI,
691 .fdr_value = 0x0000070f, 691 .fdr_value = 0x0000070f,
692 .rmcr_value = 0x00000001, 692 .rmcr_value = RMCR_RNC,
693 693
694 .apr = 1, 694 .apr = 1,
695 .mpr = 1, 695 .mpr = 1,
@@ -872,7 +872,7 @@ static void update_mac_address(struct net_device *ndev)
872static void read_mac_address(struct net_device *ndev, unsigned char *mac) 872static void read_mac_address(struct net_device *ndev, unsigned char *mac)
873{ 873{
874 if (mac[0] || mac[1] || mac[2] || mac[3] || mac[4] || mac[5]) { 874 if (mac[0] || mac[1] || mac[2] || mac[3] || mac[4] || mac[5]) {
875 memcpy(ndev->dev_addr, mac, 6); 875 memcpy(ndev->dev_addr, mac, ETH_ALEN);
876 } else { 876 } else {
877 ndev->dev_addr[0] = (sh_eth_read(ndev, MAHR) >> 24); 877 ndev->dev_addr[0] = (sh_eth_read(ndev, MAHR) >> 24);
878 ndev->dev_addr[1] = (sh_eth_read(ndev, MAHR) >> 16) & 0xFF; 878 ndev->dev_addr[1] = (sh_eth_read(ndev, MAHR) >> 16) & 0xFF;
@@ -2663,6 +2663,12 @@ static int sh_eth_drv_probe(struct platform_device *pdev)
2663 pm_runtime_enable(&pdev->dev); 2663 pm_runtime_enable(&pdev->dev);
2664 pm_runtime_resume(&pdev->dev); 2664 pm_runtime_resume(&pdev->dev);
2665 2665
2666 if (!pd) {
2667 dev_err(&pdev->dev, "no platform data\n");
2668 ret = -EINVAL;
2669 goto out_release;
2670 }
2671
2666 /* get PHY ID */ 2672 /* get PHY ID */
2667 mdp->phy_id = pd->phy; 2673 mdp->phy_id = pd->phy;
2668 mdp->phy_interface = pd->phy_interface; 2674 mdp->phy_interface = pd->phy_interface;
diff --git a/drivers/net/ethernet/renesas/sh_eth.h b/drivers/net/ethernet/renesas/sh_eth.h
index a0db02c63b11..f32c1692d310 100644
--- a/drivers/net/ethernet/renesas/sh_eth.h
+++ b/drivers/net/ethernet/renesas/sh_eth.h
@@ -321,6 +321,9 @@ enum TD_STS_BIT {
321#define TD_TFP (TD_TFP1|TD_TFP0) 321#define TD_TFP (TD_TFP1|TD_TFP0)
322 322
323/* RMCR */ 323/* RMCR */
324enum RMCR_BIT {
325 RMCR_RNC = 0x00000001,
326};
324#define DEFAULT_RMCR_VALUE 0x00000000 327#define DEFAULT_RMCR_VALUE 0x00000000
325 328
326/* ECMR */ 329/* ECMR */
diff --git a/drivers/net/ethernet/sfc/ef10.c b/drivers/net/ethernet/sfc/ef10.c
index 21f9ad6392e9..676c3c057bfb 100644
--- a/drivers/net/ethernet/sfc/ef10.c
+++ b/drivers/net/ethernet/sfc/ef10.c
@@ -285,6 +285,181 @@ static int efx_ef10_free_vis(struct efx_nic *efx)
285 return rc; 285 return rc;
286} 286}
287 287
288#ifdef EFX_USE_PIO
289
290static void efx_ef10_free_piobufs(struct efx_nic *efx)
291{
292 struct efx_ef10_nic_data *nic_data = efx->nic_data;
293 MCDI_DECLARE_BUF(inbuf, MC_CMD_FREE_PIOBUF_IN_LEN);
294 unsigned int i;
295 int rc;
296
297 BUILD_BUG_ON(MC_CMD_FREE_PIOBUF_OUT_LEN != 0);
298
299 for (i = 0; i < nic_data->n_piobufs; i++) {
300 MCDI_SET_DWORD(inbuf, FREE_PIOBUF_IN_PIOBUF_HANDLE,
301 nic_data->piobuf_handle[i]);
302 rc = efx_mcdi_rpc(efx, MC_CMD_FREE_PIOBUF, inbuf, sizeof(inbuf),
303 NULL, 0, NULL);
304 WARN_ON(rc);
305 }
306
307 nic_data->n_piobufs = 0;
308}
309
310static int efx_ef10_alloc_piobufs(struct efx_nic *efx, unsigned int n)
311{
312 struct efx_ef10_nic_data *nic_data = efx->nic_data;
313 MCDI_DECLARE_BUF(outbuf, MC_CMD_ALLOC_PIOBUF_OUT_LEN);
314 unsigned int i;
315 size_t outlen;
316 int rc = 0;
317
318 BUILD_BUG_ON(MC_CMD_ALLOC_PIOBUF_IN_LEN != 0);
319
320 for (i = 0; i < n; i++) {
321 rc = efx_mcdi_rpc(efx, MC_CMD_ALLOC_PIOBUF, NULL, 0,
322 outbuf, sizeof(outbuf), &outlen);
323 if (rc)
324 break;
325 if (outlen < MC_CMD_ALLOC_PIOBUF_OUT_LEN) {
326 rc = -EIO;
327 break;
328 }
329 nic_data->piobuf_handle[i] =
330 MCDI_DWORD(outbuf, ALLOC_PIOBUF_OUT_PIOBUF_HANDLE);
331 netif_dbg(efx, probe, efx->net_dev,
332 "allocated PIO buffer %u handle %x\n", i,
333 nic_data->piobuf_handle[i]);
334 }
335
336 nic_data->n_piobufs = i;
337 if (rc)
338 efx_ef10_free_piobufs(efx);
339 return rc;
340}
341
342static int efx_ef10_link_piobufs(struct efx_nic *efx)
343{
344 struct efx_ef10_nic_data *nic_data = efx->nic_data;
345 MCDI_DECLARE_BUF(inbuf,
346 max(MC_CMD_LINK_PIOBUF_IN_LEN,
347 MC_CMD_UNLINK_PIOBUF_IN_LEN));
348 struct efx_channel *channel;
349 struct efx_tx_queue *tx_queue;
350 unsigned int offset, index;
351 int rc;
352
353 BUILD_BUG_ON(MC_CMD_LINK_PIOBUF_OUT_LEN != 0);
354 BUILD_BUG_ON(MC_CMD_UNLINK_PIOBUF_OUT_LEN != 0);
355
356 /* Link a buffer to each VI in the write-combining mapping */
357 for (index = 0; index < nic_data->n_piobufs; ++index) {
358 MCDI_SET_DWORD(inbuf, LINK_PIOBUF_IN_PIOBUF_HANDLE,
359 nic_data->piobuf_handle[index]);
360 MCDI_SET_DWORD(inbuf, LINK_PIOBUF_IN_TXQ_INSTANCE,
361 nic_data->pio_write_vi_base + index);
362 rc = efx_mcdi_rpc(efx, MC_CMD_LINK_PIOBUF,
363 inbuf, MC_CMD_LINK_PIOBUF_IN_LEN,
364 NULL, 0, NULL);
365 if (rc) {
366 netif_err(efx, drv, efx->net_dev,
367 "failed to link VI %u to PIO buffer %u (%d)\n",
368 nic_data->pio_write_vi_base + index, index,
369 rc);
370 goto fail;
371 }
372 netif_dbg(efx, probe, efx->net_dev,
373 "linked VI %u to PIO buffer %u\n",
374 nic_data->pio_write_vi_base + index, index);
375 }
376
377 /* Link a buffer to each TX queue */
378 efx_for_each_channel(channel, efx) {
379 efx_for_each_channel_tx_queue(tx_queue, channel) {
380 /* We assign the PIO buffers to queues in
381 * reverse order to allow for the following
382 * special case.
383 */
384 offset = ((efx->tx_channel_offset + efx->n_tx_channels -
385 tx_queue->channel->channel - 1) *
386 efx_piobuf_size);
387 index = offset / ER_DZ_TX_PIOBUF_SIZE;
388 offset = offset % ER_DZ_TX_PIOBUF_SIZE;
389
390 /* When the host page size is 4K, the first
391 * host page in the WC mapping may be within
392 * the same VI page as the last TX queue. We
393 * can only link one buffer to each VI.
394 */
395 if (tx_queue->queue == nic_data->pio_write_vi_base) {
396 BUG_ON(index != 0);
397 rc = 0;
398 } else {
399 MCDI_SET_DWORD(inbuf,
400 LINK_PIOBUF_IN_PIOBUF_HANDLE,
401 nic_data->piobuf_handle[index]);
402 MCDI_SET_DWORD(inbuf,
403 LINK_PIOBUF_IN_TXQ_INSTANCE,
404 tx_queue->queue);
405 rc = efx_mcdi_rpc(efx, MC_CMD_LINK_PIOBUF,
406 inbuf, MC_CMD_LINK_PIOBUF_IN_LEN,
407 NULL, 0, NULL);
408 }
409
410 if (rc) {
411 /* This is non-fatal; the TX path just
412 * won't use PIO for this queue
413 */
414 netif_err(efx, drv, efx->net_dev,
415 "failed to link VI %u to PIO buffer %u (%d)\n",
416 tx_queue->queue, index, rc);
417 tx_queue->piobuf = NULL;
418 } else {
419 tx_queue->piobuf =
420 nic_data->pio_write_base +
421 index * EFX_VI_PAGE_SIZE + offset;
422 tx_queue->piobuf_offset = offset;
423 netif_dbg(efx, probe, efx->net_dev,
424 "linked VI %u to PIO buffer %u offset %x addr %p\n",
425 tx_queue->queue, index,
426 tx_queue->piobuf_offset,
427 tx_queue->piobuf);
428 }
429 }
430 }
431
432 return 0;
433
434fail:
435 while (index--) {
436 MCDI_SET_DWORD(inbuf, UNLINK_PIOBUF_IN_TXQ_INSTANCE,
437 nic_data->pio_write_vi_base + index);
438 efx_mcdi_rpc(efx, MC_CMD_UNLINK_PIOBUF,
439 inbuf, MC_CMD_UNLINK_PIOBUF_IN_LEN,
440 NULL, 0, NULL);
441 }
442 return rc;
443}
444
445#else /* !EFX_USE_PIO */
446
447static int efx_ef10_alloc_piobufs(struct efx_nic *efx, unsigned int n)
448{
449 return n == 0 ? 0 : -ENOBUFS;
450}
451
452static int efx_ef10_link_piobufs(struct efx_nic *efx)
453{
454 return 0;
455}
456
457static void efx_ef10_free_piobufs(struct efx_nic *efx)
458{
459}
460
461#endif /* EFX_USE_PIO */
462
288static void efx_ef10_remove(struct efx_nic *efx) 463static void efx_ef10_remove(struct efx_nic *efx)
289{ 464{
290 struct efx_ef10_nic_data *nic_data = efx->nic_data; 465 struct efx_ef10_nic_data *nic_data = efx->nic_data;
@@ -295,9 +470,15 @@ static void efx_ef10_remove(struct efx_nic *efx)
295 /* This needs to be after efx_ptp_remove_channel() with no filters */ 470 /* This needs to be after efx_ptp_remove_channel() with no filters */
296 efx_ef10_rx_free_indir_table(efx); 471 efx_ef10_rx_free_indir_table(efx);
297 472
473 if (nic_data->wc_membase)
474 iounmap(nic_data->wc_membase);
475
298 rc = efx_ef10_free_vis(efx); 476 rc = efx_ef10_free_vis(efx);
299 WARN_ON(rc != 0); 477 WARN_ON(rc != 0);
300 478
479 if (!nic_data->must_restore_piobufs)
480 efx_ef10_free_piobufs(efx);
481
301 efx_mcdi_fini(efx); 482 efx_mcdi_fini(efx);
302 efx_nic_free_buffer(efx, &nic_data->mcdi_buf); 483 efx_nic_free_buffer(efx, &nic_data->mcdi_buf);
303 kfree(nic_data); 484 kfree(nic_data);
@@ -330,12 +511,126 @@ static int efx_ef10_alloc_vis(struct efx_nic *efx,
330 return 0; 511 return 0;
331} 512}
332 513
514/* Note that the failure path of this function does not free
515 * resources, as this will be done by efx_ef10_remove().
516 */
333static int efx_ef10_dimension_resources(struct efx_nic *efx) 517static int efx_ef10_dimension_resources(struct efx_nic *efx)
334{ 518{
335 unsigned int n_vis = 519 struct efx_ef10_nic_data *nic_data = efx->nic_data;
336 max(efx->n_channels, efx->n_tx_channels * EFX_TXQ_TYPES); 520 unsigned int uc_mem_map_size, wc_mem_map_size;
521 unsigned int min_vis, pio_write_vi_base, max_vis;
522 void __iomem *membase;
523 int rc;
524
525 min_vis = max(efx->n_channels, efx->n_tx_channels * EFX_TXQ_TYPES);
526
527#ifdef EFX_USE_PIO
528 /* Try to allocate PIO buffers if wanted and if the full
529 * number of PIO buffers would be sufficient to allocate one
530 * copy-buffer per TX channel. Failure is non-fatal, as there
531 * are only a small number of PIO buffers shared between all
532 * functions of the controller.
533 */
534 if (efx_piobuf_size != 0 &&
535 ER_DZ_TX_PIOBUF_SIZE / efx_piobuf_size * EF10_TX_PIOBUF_COUNT >=
536 efx->n_tx_channels) {
537 unsigned int n_piobufs =
538 DIV_ROUND_UP(efx->n_tx_channels,
539 ER_DZ_TX_PIOBUF_SIZE / efx_piobuf_size);
540
541 rc = efx_ef10_alloc_piobufs(efx, n_piobufs);
542 if (rc)
543 netif_err(efx, probe, efx->net_dev,
544 "failed to allocate PIO buffers (%d)\n", rc);
545 else
546 netif_dbg(efx, probe, efx->net_dev,
547 "allocated %u PIO buffers\n", n_piobufs);
548 }
549#else
550 nic_data->n_piobufs = 0;
551#endif
337 552
338 return efx_ef10_alloc_vis(efx, n_vis, n_vis); 553 /* PIO buffers should be mapped with write-combining enabled,
554 * and we want to make single UC and WC mappings rather than
555 * several of each (in fact that's the only option if host
556 * page size is >4K). So we may allocate some extra VIs just
557 * for writing PIO buffers through.
558 */
559 uc_mem_map_size = PAGE_ALIGN((min_vis - 1) * EFX_VI_PAGE_SIZE +
560 ER_DZ_TX_PIOBUF);
561 if (nic_data->n_piobufs) {
562 pio_write_vi_base = uc_mem_map_size / EFX_VI_PAGE_SIZE;
563 wc_mem_map_size = (PAGE_ALIGN((pio_write_vi_base +
564 nic_data->n_piobufs) *
565 EFX_VI_PAGE_SIZE) -
566 uc_mem_map_size);
567 max_vis = pio_write_vi_base + nic_data->n_piobufs;
568 } else {
569 pio_write_vi_base = 0;
570 wc_mem_map_size = 0;
571 max_vis = min_vis;
572 }
573
574 /* In case the last attached driver failed to free VIs, do it now */
575 rc = efx_ef10_free_vis(efx);
576 if (rc != 0)
577 return rc;
578
579 rc = efx_ef10_alloc_vis(efx, min_vis, max_vis);
580 if (rc != 0)
581 return rc;
582
583 /* If we didn't get enough VIs to map all the PIO buffers, free the
584 * PIO buffers
585 */
586 if (nic_data->n_piobufs &&
587 nic_data->n_allocated_vis <
588 pio_write_vi_base + nic_data->n_piobufs) {
589 netif_dbg(efx, probe, efx->net_dev,
590 "%u VIs are not sufficient to map %u PIO buffers\n",
591 nic_data->n_allocated_vis, nic_data->n_piobufs);
592 efx_ef10_free_piobufs(efx);
593 }
594
595 /* Shrink the original UC mapping of the memory BAR */
596 membase = ioremap_nocache(efx->membase_phys, uc_mem_map_size);
597 if (!membase) {
598 netif_err(efx, probe, efx->net_dev,
599 "could not shrink memory BAR to %x\n",
600 uc_mem_map_size);
601 return -ENOMEM;
602 }
603 iounmap(efx->membase);
604 efx->membase = membase;
605
606 /* Set up the WC mapping if needed */
607 if (wc_mem_map_size) {
608 nic_data->wc_membase = ioremap_wc(efx->membase_phys +
609 uc_mem_map_size,
610 wc_mem_map_size);
611 if (!nic_data->wc_membase) {
612 netif_err(efx, probe, efx->net_dev,
613 "could not allocate WC mapping of size %x\n",
614 wc_mem_map_size);
615 return -ENOMEM;
616 }
617 nic_data->pio_write_vi_base = pio_write_vi_base;
618 nic_data->pio_write_base =
619 nic_data->wc_membase +
620 (pio_write_vi_base * EFX_VI_PAGE_SIZE + ER_DZ_TX_PIOBUF -
621 uc_mem_map_size);
622
623 rc = efx_ef10_link_piobufs(efx);
624 if (rc)
625 efx_ef10_free_piobufs(efx);
626 }
627
628 netif_dbg(efx, probe, efx->net_dev,
629 "memory BAR at %pa (virtual %p+%x UC, %p+%x WC)\n",
630 &efx->membase_phys, efx->membase, uc_mem_map_size,
631 nic_data->wc_membase, wc_mem_map_size);
632
633 return 0;
339} 634}
340 635
341static int efx_ef10_init_nic(struct efx_nic *efx) 636static int efx_ef10_init_nic(struct efx_nic *efx)
@@ -359,6 +654,21 @@ static int efx_ef10_init_nic(struct efx_nic *efx)
359 nic_data->must_realloc_vis = false; 654 nic_data->must_realloc_vis = false;
360 } 655 }
361 656
657 if (nic_data->must_restore_piobufs && nic_data->n_piobufs) {
658 rc = efx_ef10_alloc_piobufs(efx, nic_data->n_piobufs);
659 if (rc == 0) {
660 rc = efx_ef10_link_piobufs(efx);
661 if (rc)
662 efx_ef10_free_piobufs(efx);
663 }
664
665 /* Log an error on failure, but this is non-fatal */
666 if (rc)
667 netif_err(efx, drv, efx->net_dev,
668 "failed to restore PIO buffers (%d)\n", rc);
669 nic_data->must_restore_piobufs = false;
670 }
671
362 efx_ef10_rx_push_indir_table(efx); 672 efx_ef10_rx_push_indir_table(efx);
363 return 0; 673 return 0;
364} 674}
@@ -759,6 +1069,7 @@ static int efx_ef10_mcdi_poll_reboot(struct efx_nic *efx)
759 /* All our allocations have been reset */ 1069 /* All our allocations have been reset */
760 nic_data->must_realloc_vis = true; 1070 nic_data->must_realloc_vis = true;
761 nic_data->must_restore_filters = true; 1071 nic_data->must_restore_filters = true;
1072 nic_data->must_restore_piobufs = true;
762 nic_data->rx_rss_context = EFX_EF10_RSS_CONTEXT_INVALID; 1073 nic_data->rx_rss_context = EFX_EF10_RSS_CONTEXT_INVALID;
763 1074
764 /* The datapath firmware might have been changed */ 1075 /* The datapath firmware might have been changed */
@@ -2180,7 +2491,7 @@ out_unlock:
2180 return rc; 2491 return rc;
2181} 2492}
2182 2493
2183void efx_ef10_filter_update_rx_scatter(struct efx_nic *efx) 2494static void efx_ef10_filter_update_rx_scatter(struct efx_nic *efx)
2184{ 2495{
2185 /* no need to do anything here on EF10 */ 2496 /* no need to do anything here on EF10 */
2186} 2497}
diff --git a/drivers/net/ethernet/sfc/ef10_regs.h b/drivers/net/ethernet/sfc/ef10_regs.h
index b3f4e3755fd9..207ac9a1e3de 100644
--- a/drivers/net/ethernet/sfc/ef10_regs.h
+++ b/drivers/net/ethernet/sfc/ef10_regs.h
@@ -315,6 +315,7 @@
315#define ESF_DZ_TX_PIO_TYPE_WIDTH 1 315#define ESF_DZ_TX_PIO_TYPE_WIDTH 1
316#define ESF_DZ_TX_PIO_OPT_LBN 60 316#define ESF_DZ_TX_PIO_OPT_LBN 60
317#define ESF_DZ_TX_PIO_OPT_WIDTH 3 317#define ESF_DZ_TX_PIO_OPT_WIDTH 3
318#define ESE_DZ_TX_OPTION_DESC_PIO 1
318#define ESF_DZ_TX_PIO_CONT_LBN 59 319#define ESF_DZ_TX_PIO_CONT_LBN 59
319#define ESF_DZ_TX_PIO_CONT_WIDTH 1 320#define ESF_DZ_TX_PIO_CONT_WIDTH 1
320#define ESF_DZ_TX_PIO_BYTE_CNT_LBN 32 321#define ESF_DZ_TX_PIO_BYTE_CNT_LBN 32
diff --git a/drivers/net/ethernet/sfc/efx.h b/drivers/net/ethernet/sfc/efx.h
index 34d00f5771fe..b8235ee5d7d7 100644
--- a/drivers/net/ethernet/sfc/efx.h
+++ b/drivers/net/ethernet/sfc/efx.h
@@ -18,37 +18,36 @@
18#define EFX_MEM_BAR 2 18#define EFX_MEM_BAR 2
19 19
20/* TX */ 20/* TX */
21extern int efx_probe_tx_queue(struct efx_tx_queue *tx_queue); 21int efx_probe_tx_queue(struct efx_tx_queue *tx_queue);
22extern void efx_remove_tx_queue(struct efx_tx_queue *tx_queue); 22void efx_remove_tx_queue(struct efx_tx_queue *tx_queue);
23extern void efx_init_tx_queue(struct efx_tx_queue *tx_queue); 23void efx_init_tx_queue(struct efx_tx_queue *tx_queue);
24extern void efx_init_tx_queue_core_txq(struct efx_tx_queue *tx_queue); 24void efx_init_tx_queue_core_txq(struct efx_tx_queue *tx_queue);
25extern void efx_fini_tx_queue(struct efx_tx_queue *tx_queue); 25void efx_fini_tx_queue(struct efx_tx_queue *tx_queue);
26extern netdev_tx_t 26netdev_tx_t efx_hard_start_xmit(struct sk_buff *skb,
27efx_hard_start_xmit(struct sk_buff *skb, struct net_device *net_dev); 27 struct net_device *net_dev);
28extern netdev_tx_t 28netdev_tx_t efx_enqueue_skb(struct efx_tx_queue *tx_queue, struct sk_buff *skb);
29efx_enqueue_skb(struct efx_tx_queue *tx_queue, struct sk_buff *skb); 29void efx_xmit_done(struct efx_tx_queue *tx_queue, unsigned int index);
30extern void efx_xmit_done(struct efx_tx_queue *tx_queue, unsigned int index); 30int efx_setup_tc(struct net_device *net_dev, u8 num_tc);
31extern int efx_setup_tc(struct net_device *net_dev, u8 num_tc); 31unsigned int efx_tx_max_skb_descs(struct efx_nic *efx);
32extern unsigned int efx_tx_max_skb_descs(struct efx_nic *efx); 32extern unsigned int efx_piobuf_size;
33 33
34/* RX */ 34/* RX */
35extern void efx_rx_config_page_split(struct efx_nic *efx); 35void efx_rx_config_page_split(struct efx_nic *efx);
36extern int efx_probe_rx_queue(struct efx_rx_queue *rx_queue); 36int efx_probe_rx_queue(struct efx_rx_queue *rx_queue);
37extern void efx_remove_rx_queue(struct efx_rx_queue *rx_queue); 37void efx_remove_rx_queue(struct efx_rx_queue *rx_queue);
38extern void efx_init_rx_queue(struct efx_rx_queue *rx_queue); 38void efx_init_rx_queue(struct efx_rx_queue *rx_queue);
39extern void efx_fini_rx_queue(struct efx_rx_queue *rx_queue); 39void efx_fini_rx_queue(struct efx_rx_queue *rx_queue);
40extern void efx_fast_push_rx_descriptors(struct efx_rx_queue *rx_queue); 40void efx_fast_push_rx_descriptors(struct efx_rx_queue *rx_queue);
41extern void efx_rx_slow_fill(unsigned long context); 41void efx_rx_slow_fill(unsigned long context);
42extern void __efx_rx_packet(struct efx_channel *channel); 42void __efx_rx_packet(struct efx_channel *channel);
43extern void efx_rx_packet(struct efx_rx_queue *rx_queue, 43void efx_rx_packet(struct efx_rx_queue *rx_queue, unsigned int index,
44 unsigned int index, unsigned int n_frags, 44 unsigned int n_frags, unsigned int len, u16 flags);
45 unsigned int len, u16 flags);
46static inline void efx_rx_flush_packet(struct efx_channel *channel) 45static inline void efx_rx_flush_packet(struct efx_channel *channel)
47{ 46{
48 if (channel->rx_pkt_n_frags) 47 if (channel->rx_pkt_n_frags)
49 __efx_rx_packet(channel); 48 __efx_rx_packet(channel);
50} 49}
51extern void efx_schedule_slow_fill(struct efx_rx_queue *rx_queue); 50void efx_schedule_slow_fill(struct efx_rx_queue *rx_queue);
52 51
53#define EFX_MAX_DMAQ_SIZE 4096UL 52#define EFX_MAX_DMAQ_SIZE 4096UL
54#define EFX_DEFAULT_DMAQ_SIZE 1024UL 53#define EFX_DEFAULT_DMAQ_SIZE 1024UL
@@ -162,9 +161,9 @@ static inline s32 efx_filter_get_rx_ids(struct efx_nic *efx,
162 return efx->type->filter_get_rx_ids(efx, priority, buf, size); 161 return efx->type->filter_get_rx_ids(efx, priority, buf, size);
163} 162}
164#ifdef CONFIG_RFS_ACCEL 163#ifdef CONFIG_RFS_ACCEL
165extern int efx_filter_rfs(struct net_device *net_dev, const struct sk_buff *skb, 164int efx_filter_rfs(struct net_device *net_dev, const struct sk_buff *skb,
166 u16 rxq_index, u32 flow_id); 165 u16 rxq_index, u32 flow_id);
167extern bool __efx_filter_rfs_expire(struct efx_nic *efx, unsigned quota); 166bool __efx_filter_rfs_expire(struct efx_nic *efx, unsigned quota);
168static inline void efx_filter_rfs_expire(struct efx_channel *channel) 167static inline void efx_filter_rfs_expire(struct efx_channel *channel)
169{ 168{
170 if (channel->rfs_filters_added >= 60 && 169 if (channel->rfs_filters_added >= 60 &&
@@ -176,50 +175,48 @@ static inline void efx_filter_rfs_expire(struct efx_channel *channel)
176static inline void efx_filter_rfs_expire(struct efx_channel *channel) {} 175static inline void efx_filter_rfs_expire(struct efx_channel *channel) {}
177#define efx_filter_rfs_enabled() 0 176#define efx_filter_rfs_enabled() 0
178#endif 177#endif
179extern bool efx_filter_is_mc_recipient(const struct efx_filter_spec *spec); 178bool efx_filter_is_mc_recipient(const struct efx_filter_spec *spec);
180 179
181/* Channels */ 180/* Channels */
182extern int efx_channel_dummy_op_int(struct efx_channel *channel); 181int efx_channel_dummy_op_int(struct efx_channel *channel);
183extern void efx_channel_dummy_op_void(struct efx_channel *channel); 182void efx_channel_dummy_op_void(struct efx_channel *channel);
184extern int 183int efx_realloc_channels(struct efx_nic *efx, u32 rxq_entries, u32 txq_entries);
185efx_realloc_channels(struct efx_nic *efx, u32 rxq_entries, u32 txq_entries);
186 184
187/* Ports */ 185/* Ports */
188extern int efx_reconfigure_port(struct efx_nic *efx); 186int efx_reconfigure_port(struct efx_nic *efx);
189extern int __efx_reconfigure_port(struct efx_nic *efx); 187int __efx_reconfigure_port(struct efx_nic *efx);
190 188
191/* Ethtool support */ 189/* Ethtool support */
192extern const struct ethtool_ops efx_ethtool_ops; 190extern const struct ethtool_ops efx_ethtool_ops;
193 191
194/* Reset handling */ 192/* Reset handling */
195extern int efx_reset(struct efx_nic *efx, enum reset_type method); 193int efx_reset(struct efx_nic *efx, enum reset_type method);
196extern void efx_reset_down(struct efx_nic *efx, enum reset_type method); 194void efx_reset_down(struct efx_nic *efx, enum reset_type method);
197extern int efx_reset_up(struct efx_nic *efx, enum reset_type method, bool ok); 195int efx_reset_up(struct efx_nic *efx, enum reset_type method, bool ok);
198extern int efx_try_recovery(struct efx_nic *efx); 196int efx_try_recovery(struct efx_nic *efx);
199 197
200/* Global */ 198/* Global */
201extern void efx_schedule_reset(struct efx_nic *efx, enum reset_type type); 199void efx_schedule_reset(struct efx_nic *efx, enum reset_type type);
202extern int efx_init_irq_moderation(struct efx_nic *efx, unsigned int tx_usecs, 200int efx_init_irq_moderation(struct efx_nic *efx, unsigned int tx_usecs,
203 unsigned int rx_usecs, bool rx_adaptive, 201 unsigned int rx_usecs, bool rx_adaptive,
204 bool rx_may_override_tx); 202 bool rx_may_override_tx);
205extern void efx_get_irq_moderation(struct efx_nic *efx, unsigned int *tx_usecs, 203void efx_get_irq_moderation(struct efx_nic *efx, unsigned int *tx_usecs,
206 unsigned int *rx_usecs, bool *rx_adaptive); 204 unsigned int *rx_usecs, bool *rx_adaptive);
207 205
208/* Dummy PHY ops for PHY drivers */ 206/* Dummy PHY ops for PHY drivers */
209extern int efx_port_dummy_op_int(struct efx_nic *efx); 207int efx_port_dummy_op_int(struct efx_nic *efx);
210extern void efx_port_dummy_op_void(struct efx_nic *efx); 208void efx_port_dummy_op_void(struct efx_nic *efx);
211
212 209
213/* MTD */ 210/* MTD */
214#ifdef CONFIG_SFC_MTD 211#ifdef CONFIG_SFC_MTD
215extern int efx_mtd_add(struct efx_nic *efx, struct efx_mtd_partition *parts, 212int efx_mtd_add(struct efx_nic *efx, struct efx_mtd_partition *parts,
216 size_t n_parts, size_t sizeof_part); 213 size_t n_parts, size_t sizeof_part);
217static inline int efx_mtd_probe(struct efx_nic *efx) 214static inline int efx_mtd_probe(struct efx_nic *efx)
218{ 215{
219 return efx->type->mtd_probe(efx); 216 return efx->type->mtd_probe(efx);
220} 217}
221extern void efx_mtd_rename(struct efx_nic *efx); 218void efx_mtd_rename(struct efx_nic *efx);
222extern void efx_mtd_remove(struct efx_nic *efx); 219void efx_mtd_remove(struct efx_nic *efx);
223#else 220#else
224static inline int efx_mtd_probe(struct efx_nic *efx) { return 0; } 221static inline int efx_mtd_probe(struct efx_nic *efx) { return 0; }
225static inline void efx_mtd_rename(struct efx_nic *efx) {} 222static inline void efx_mtd_rename(struct efx_nic *efx) {}
@@ -241,9 +238,9 @@ static inline void efx_schedule_channel_irq(struct efx_channel *channel)
241 efx_schedule_channel(channel); 238 efx_schedule_channel(channel);
242} 239}
243 240
244extern void efx_link_status_changed(struct efx_nic *efx); 241void efx_link_status_changed(struct efx_nic *efx);
245extern void efx_link_set_advertising(struct efx_nic *efx, u32); 242void efx_link_set_advertising(struct efx_nic *efx, u32);
246extern void efx_link_set_wanted_fc(struct efx_nic *efx, u8); 243void efx_link_set_wanted_fc(struct efx_nic *efx, u8);
247 244
248static inline void efx_device_detach_sync(struct efx_nic *efx) 245static inline void efx_device_detach_sync(struct efx_nic *efx)
249{ 246{
diff --git a/drivers/net/ethernet/sfc/ethtool.c b/drivers/net/ethernet/sfc/ethtool.c
index 5b471cf5c323..1f529fa2edb1 100644
--- a/drivers/net/ethernet/sfc/ethtool.c
+++ b/drivers/net/ethernet/sfc/ethtool.c
@@ -70,6 +70,7 @@ static const struct efx_sw_stat_desc efx_sw_stat_desc[] = {
70 EFX_ETHTOOL_UINT_TXQ_STAT(tso_long_headers), 70 EFX_ETHTOOL_UINT_TXQ_STAT(tso_long_headers),
71 EFX_ETHTOOL_UINT_TXQ_STAT(tso_packets), 71 EFX_ETHTOOL_UINT_TXQ_STAT(tso_packets),
72 EFX_ETHTOOL_UINT_TXQ_STAT(pushes), 72 EFX_ETHTOOL_UINT_TXQ_STAT(pushes),
73 EFX_ETHTOOL_UINT_TXQ_STAT(pio_packets),
73 EFX_ETHTOOL_ATOMIC_NIC_ERROR_STAT(rx_reset), 74 EFX_ETHTOOL_ATOMIC_NIC_ERROR_STAT(rx_reset),
74 EFX_ETHTOOL_UINT_CHANNEL_STAT(rx_tobe_disc), 75 EFX_ETHTOOL_UINT_CHANNEL_STAT(rx_tobe_disc),
75 EFX_ETHTOOL_UINT_CHANNEL_STAT(rx_ip_hdr_chksum_err), 76 EFX_ETHTOOL_UINT_CHANNEL_STAT(rx_ip_hdr_chksum_err),
@@ -1035,8 +1036,8 @@ static int efx_ethtool_set_rxfh_indir(struct net_device *net_dev,
1035 return 0; 1036 return 0;
1036} 1037}
1037 1038
1038int efx_ethtool_get_ts_info(struct net_device *net_dev, 1039static int efx_ethtool_get_ts_info(struct net_device *net_dev,
1039 struct ethtool_ts_info *ts_info) 1040 struct ethtool_ts_info *ts_info)
1040{ 1041{
1041 struct efx_nic *efx = netdev_priv(net_dev); 1042 struct efx_nic *efx = netdev_priv(net_dev);
1042 1043
diff --git a/drivers/net/ethernet/sfc/io.h b/drivers/net/ethernet/sfc/io.h
index 96ce507d8602..4d3f119b67b3 100644
--- a/drivers/net/ethernet/sfc/io.h
+++ b/drivers/net/ethernet/sfc/io.h
@@ -66,6 +66,11 @@
66#define EFX_USE_QWORD_IO 1 66#define EFX_USE_QWORD_IO 1
67#endif 67#endif
68 68
69/* PIO is a win only if write-combining is possible */
70#ifdef ARCH_HAS_IOREMAP_WC
71#define EFX_USE_PIO 1
72#endif
73
69#ifdef EFX_USE_QWORD_IO 74#ifdef EFX_USE_QWORD_IO
70static inline void _efx_writeq(struct efx_nic *efx, __le64 value, 75static inline void _efx_writeq(struct efx_nic *efx, __le64 value,
71 unsigned int reg) 76 unsigned int reg)
diff --git a/drivers/net/ethernet/sfc/mcdi.h b/drivers/net/ethernet/sfc/mcdi.h
index c34d0d4e10ee..656a3277c2b2 100644
--- a/drivers/net/ethernet/sfc/mcdi.h
+++ b/drivers/net/ethernet/sfc/mcdi.h
@@ -108,38 +108,35 @@ static inline struct efx_mcdi_mon *efx_mcdi_mon(struct efx_nic *efx)
108} 108}
109#endif 109#endif
110 110
111extern int efx_mcdi_init(struct efx_nic *efx); 111int efx_mcdi_init(struct efx_nic *efx);
112extern void efx_mcdi_fini(struct efx_nic *efx); 112void efx_mcdi_fini(struct efx_nic *efx);
113 113
114extern int efx_mcdi_rpc(struct efx_nic *efx, unsigned cmd, 114int efx_mcdi_rpc(struct efx_nic *efx, unsigned cmd, const efx_dword_t *inbuf,
115 const efx_dword_t *inbuf, size_t inlen, 115 size_t inlen, efx_dword_t *outbuf, size_t outlen,
116 size_t *outlen_actual);
117
118int efx_mcdi_rpc_start(struct efx_nic *efx, unsigned cmd,
119 const efx_dword_t *inbuf, size_t inlen);
120int efx_mcdi_rpc_finish(struct efx_nic *efx, unsigned cmd, size_t inlen,
116 efx_dword_t *outbuf, size_t outlen, 121 efx_dword_t *outbuf, size_t outlen,
117 size_t *outlen_actual); 122 size_t *outlen_actual);
118 123
119extern int efx_mcdi_rpc_start(struct efx_nic *efx, unsigned cmd,
120 const efx_dword_t *inbuf, size_t inlen);
121extern int efx_mcdi_rpc_finish(struct efx_nic *efx, unsigned cmd, size_t inlen,
122 efx_dword_t *outbuf, size_t outlen,
123 size_t *outlen_actual);
124
125typedef void efx_mcdi_async_completer(struct efx_nic *efx, 124typedef void efx_mcdi_async_completer(struct efx_nic *efx,
126 unsigned long cookie, int rc, 125 unsigned long cookie, int rc,
127 efx_dword_t *outbuf, 126 efx_dword_t *outbuf,
128 size_t outlen_actual); 127 size_t outlen_actual);
129extern int efx_mcdi_rpc_async(struct efx_nic *efx, unsigned int cmd, 128int efx_mcdi_rpc_async(struct efx_nic *efx, unsigned int cmd,
130 const efx_dword_t *inbuf, size_t inlen, 129 const efx_dword_t *inbuf, size_t inlen, size_t outlen,
131 size_t outlen, 130 efx_mcdi_async_completer *complete,
132 efx_mcdi_async_completer *complete, 131 unsigned long cookie);
133 unsigned long cookie);
134 132
135extern int efx_mcdi_poll_reboot(struct efx_nic *efx); 133int efx_mcdi_poll_reboot(struct efx_nic *efx);
136extern void efx_mcdi_mode_poll(struct efx_nic *efx); 134void efx_mcdi_mode_poll(struct efx_nic *efx);
137extern void efx_mcdi_mode_event(struct efx_nic *efx); 135void efx_mcdi_mode_event(struct efx_nic *efx);
138extern void efx_mcdi_flush_async(struct efx_nic *efx); 136void efx_mcdi_flush_async(struct efx_nic *efx);
139 137
140extern void efx_mcdi_process_event(struct efx_channel *channel, 138void efx_mcdi_process_event(struct efx_channel *channel, efx_qword_t *event);
141 efx_qword_t *event); 139void efx_mcdi_sensor_event(struct efx_nic *efx, efx_qword_t *ev);
142extern void efx_mcdi_sensor_event(struct efx_nic *efx, efx_qword_t *ev);
143 140
144/* We expect that 16- and 32-bit fields in MCDI requests and responses 141/* We expect that 16- and 32-bit fields in MCDI requests and responses
145 * are appropriately aligned, but 64-bit fields are only 142 * are appropriately aligned, but 64-bit fields are only
@@ -275,55 +272,54 @@ extern void efx_mcdi_sensor_event(struct efx_nic *efx, efx_qword_t *ev);
275#define MCDI_EVENT_FIELD(_ev, _field) \ 272#define MCDI_EVENT_FIELD(_ev, _field) \
276 EFX_QWORD_FIELD(_ev, MCDI_EVENT_ ## _field) 273 EFX_QWORD_FIELD(_ev, MCDI_EVENT_ ## _field)
277 274
278extern void efx_mcdi_print_fwver(struct efx_nic *efx, char *buf, size_t len); 275void efx_mcdi_print_fwver(struct efx_nic *efx, char *buf, size_t len);
279extern int efx_mcdi_get_board_cfg(struct efx_nic *efx, u8 *mac_address, 276int efx_mcdi_get_board_cfg(struct efx_nic *efx, u8 *mac_address,
280 u16 *fw_subtype_list, u32 *capabilities); 277 u16 *fw_subtype_list, u32 *capabilities);
281extern int efx_mcdi_log_ctrl(struct efx_nic *efx, bool evq, bool uart, 278int efx_mcdi_log_ctrl(struct efx_nic *efx, bool evq, bool uart, u32 dest_evq);
282 u32 dest_evq); 279int efx_mcdi_nvram_types(struct efx_nic *efx, u32 *nvram_types_out);
283extern int efx_mcdi_nvram_types(struct efx_nic *efx, u32 *nvram_types_out); 280int efx_mcdi_nvram_info(struct efx_nic *efx, unsigned int type,
284extern int efx_mcdi_nvram_info(struct efx_nic *efx, unsigned int type, 281 size_t *size_out, size_t *erase_size_out,
285 size_t *size_out, size_t *erase_size_out, 282 bool *protected_out);
286 bool *protected_out); 283int efx_mcdi_nvram_test_all(struct efx_nic *efx);
287extern int efx_mcdi_nvram_test_all(struct efx_nic *efx); 284int efx_mcdi_handle_assertion(struct efx_nic *efx);
288extern int efx_mcdi_handle_assertion(struct efx_nic *efx); 285void efx_mcdi_set_id_led(struct efx_nic *efx, enum efx_led_mode mode);
289extern void efx_mcdi_set_id_led(struct efx_nic *efx, enum efx_led_mode mode); 286int efx_mcdi_wol_filter_set_magic(struct efx_nic *efx, const u8 *mac,
290extern int efx_mcdi_wol_filter_set_magic(struct efx_nic *efx, 287 int *id_out);
291 const u8 *mac, int *id_out); 288int efx_mcdi_wol_filter_get_magic(struct efx_nic *efx, int *id_out);
292extern int efx_mcdi_wol_filter_get_magic(struct efx_nic *efx, int *id_out); 289int efx_mcdi_wol_filter_remove(struct efx_nic *efx, int id);
293extern int efx_mcdi_wol_filter_remove(struct efx_nic *efx, int id); 290int efx_mcdi_wol_filter_reset(struct efx_nic *efx);
294extern int efx_mcdi_wol_filter_reset(struct efx_nic *efx); 291int efx_mcdi_flush_rxqs(struct efx_nic *efx);
295extern int efx_mcdi_flush_rxqs(struct efx_nic *efx); 292int efx_mcdi_port_probe(struct efx_nic *efx);
296extern int efx_mcdi_port_probe(struct efx_nic *efx); 293void efx_mcdi_port_remove(struct efx_nic *efx);
297extern void efx_mcdi_port_remove(struct efx_nic *efx); 294int efx_mcdi_port_reconfigure(struct efx_nic *efx);
298extern int efx_mcdi_port_reconfigure(struct efx_nic *efx); 295int efx_mcdi_port_get_number(struct efx_nic *efx);
299extern int efx_mcdi_port_get_number(struct efx_nic *efx); 296u32 efx_mcdi_phy_get_caps(struct efx_nic *efx);
300extern u32 efx_mcdi_phy_get_caps(struct efx_nic *efx); 297void efx_mcdi_process_link_change(struct efx_nic *efx, efx_qword_t *ev);
301extern void efx_mcdi_process_link_change(struct efx_nic *efx, efx_qword_t *ev); 298int efx_mcdi_set_mac(struct efx_nic *efx);
302extern int efx_mcdi_set_mac(struct efx_nic *efx);
303#define EFX_MC_STATS_GENERATION_INVALID ((__force __le64)(-1)) 299#define EFX_MC_STATS_GENERATION_INVALID ((__force __le64)(-1))
304extern void efx_mcdi_mac_start_stats(struct efx_nic *efx); 300void efx_mcdi_mac_start_stats(struct efx_nic *efx);
305extern void efx_mcdi_mac_stop_stats(struct efx_nic *efx); 301void efx_mcdi_mac_stop_stats(struct efx_nic *efx);
306extern bool efx_mcdi_mac_check_fault(struct efx_nic *efx); 302bool efx_mcdi_mac_check_fault(struct efx_nic *efx);
307extern enum reset_type efx_mcdi_map_reset_reason(enum reset_type reason); 303enum reset_type efx_mcdi_map_reset_reason(enum reset_type reason);
308extern int efx_mcdi_reset(struct efx_nic *efx, enum reset_type method); 304int efx_mcdi_reset(struct efx_nic *efx, enum reset_type method);
309extern int efx_mcdi_set_workaround(struct efx_nic *efx, u32 type, bool enabled); 305int efx_mcdi_set_workaround(struct efx_nic *efx, u32 type, bool enabled);
310 306
311#ifdef CONFIG_SFC_MCDI_MON 307#ifdef CONFIG_SFC_MCDI_MON
312extern int efx_mcdi_mon_probe(struct efx_nic *efx); 308int efx_mcdi_mon_probe(struct efx_nic *efx);
313extern void efx_mcdi_mon_remove(struct efx_nic *efx); 309void efx_mcdi_mon_remove(struct efx_nic *efx);
314#else 310#else
315static inline int efx_mcdi_mon_probe(struct efx_nic *efx) { return 0; } 311static inline int efx_mcdi_mon_probe(struct efx_nic *efx) { return 0; }
316static inline void efx_mcdi_mon_remove(struct efx_nic *efx) {} 312static inline void efx_mcdi_mon_remove(struct efx_nic *efx) {}
317#endif 313#endif
318 314
319#ifdef CONFIG_SFC_MTD 315#ifdef CONFIG_SFC_MTD
320extern int efx_mcdi_mtd_read(struct mtd_info *mtd, loff_t start, 316int efx_mcdi_mtd_read(struct mtd_info *mtd, loff_t start, size_t len,
321 size_t len, size_t *retlen, u8 *buffer); 317 size_t *retlen, u8 *buffer);
322extern int efx_mcdi_mtd_erase(struct mtd_info *mtd, loff_t start, size_t len); 318int efx_mcdi_mtd_erase(struct mtd_info *mtd, loff_t start, size_t len);
323extern int efx_mcdi_mtd_write(struct mtd_info *mtd, loff_t start, 319int efx_mcdi_mtd_write(struct mtd_info *mtd, loff_t start, size_t len,
324 size_t len, size_t *retlen, const u8 *buffer); 320 size_t *retlen, const u8 *buffer);
325extern int efx_mcdi_mtd_sync(struct mtd_info *mtd); 321int efx_mcdi_mtd_sync(struct mtd_info *mtd);
326extern void efx_mcdi_mtd_rename(struct efx_mtd_partition *part); 322void efx_mcdi_mtd_rename(struct efx_mtd_partition *part);
327#endif 323#endif
328 324
329#endif /* EFX_MCDI_H */ 325#endif /* EFX_MCDI_H */
diff --git a/drivers/net/ethernet/sfc/mdio_10g.h b/drivers/net/ethernet/sfc/mdio_10g.h
index 16824fecc5ee..4a2dc4c281b7 100644
--- a/drivers/net/ethernet/sfc/mdio_10g.h
+++ b/drivers/net/ethernet/sfc/mdio_10g.h
@@ -20,7 +20,7 @@
20 20
21static inline unsigned efx_mdio_id_rev(u32 id) { return id & 0xf; } 21static inline unsigned efx_mdio_id_rev(u32 id) { return id & 0xf; }
22static inline unsigned efx_mdio_id_model(u32 id) { return (id >> 4) & 0x3f; } 22static inline unsigned efx_mdio_id_model(u32 id) { return (id >> 4) & 0x3f; }
23extern unsigned efx_mdio_id_oui(u32 id); 23unsigned efx_mdio_id_oui(u32 id);
24 24
25static inline int efx_mdio_read(struct efx_nic *efx, int devad, int addr) 25static inline int efx_mdio_read(struct efx_nic *efx, int devad, int addr)
26{ 26{
@@ -56,7 +56,7 @@ static inline bool efx_mdio_phyxgxs_lane_sync(struct efx_nic *efx)
56 return sync; 56 return sync;
57} 57}
58 58
59extern const char *efx_mdio_mmd_name(int mmd); 59const char *efx_mdio_mmd_name(int mmd);
60 60
61/* 61/*
62 * Reset a specific MMD and wait for reset to clear. 62 * Reset a specific MMD and wait for reset to clear.
@@ -64,30 +64,29 @@ extern const char *efx_mdio_mmd_name(int mmd);
64 * 64 *
65 * This function will sleep 65 * This function will sleep
66 */ 66 */
67extern int efx_mdio_reset_mmd(struct efx_nic *efx, int mmd, 67int efx_mdio_reset_mmd(struct efx_nic *efx, int mmd, int spins, int spintime);
68 int spins, int spintime);
69 68
70/* As efx_mdio_check_mmd but for multiple MMDs */ 69/* As efx_mdio_check_mmd but for multiple MMDs */
71int efx_mdio_check_mmds(struct efx_nic *efx, unsigned int mmd_mask); 70int efx_mdio_check_mmds(struct efx_nic *efx, unsigned int mmd_mask);
72 71
73/* Check the link status of specified mmds in bit mask */ 72/* Check the link status of specified mmds in bit mask */
74extern bool efx_mdio_links_ok(struct efx_nic *efx, unsigned int mmd_mask); 73bool efx_mdio_links_ok(struct efx_nic *efx, unsigned int mmd_mask);
75 74
76/* Generic transmit disable support though PMAPMD */ 75/* Generic transmit disable support though PMAPMD */
77extern void efx_mdio_transmit_disable(struct efx_nic *efx); 76void efx_mdio_transmit_disable(struct efx_nic *efx);
78 77
79/* Generic part of reconfigure: set/clear loopback bits */ 78/* Generic part of reconfigure: set/clear loopback bits */
80extern void efx_mdio_phy_reconfigure(struct efx_nic *efx); 79void efx_mdio_phy_reconfigure(struct efx_nic *efx);
81 80
82/* Set the power state of the specified MMDs */ 81/* Set the power state of the specified MMDs */
83extern void efx_mdio_set_mmds_lpower(struct efx_nic *efx, 82void efx_mdio_set_mmds_lpower(struct efx_nic *efx, int low_power,
84 int low_power, unsigned int mmd_mask); 83 unsigned int mmd_mask);
85 84
86/* Set (some of) the PHY settings over MDIO */ 85/* Set (some of) the PHY settings over MDIO */
87extern int efx_mdio_set_settings(struct efx_nic *efx, struct ethtool_cmd *ecmd); 86int efx_mdio_set_settings(struct efx_nic *efx, struct ethtool_cmd *ecmd);
88 87
89/* Push advertising flags and restart autonegotiation */ 88/* Push advertising flags and restart autonegotiation */
90extern void efx_mdio_an_reconfigure(struct efx_nic *efx); 89void efx_mdio_an_reconfigure(struct efx_nic *efx);
91 90
92/* Get pause parameters from AN if available (otherwise return 91/* Get pause parameters from AN if available (otherwise return
93 * requested pause parameters) 92 * requested pause parameters)
@@ -95,8 +94,7 @@ extern void efx_mdio_an_reconfigure(struct efx_nic *efx);
95u8 efx_mdio_get_pause(struct efx_nic *efx); 94u8 efx_mdio_get_pause(struct efx_nic *efx);
96 95
97/* Wait for specified MMDs to exit reset within a timeout */ 96/* Wait for specified MMDs to exit reset within a timeout */
98extern int efx_mdio_wait_reset_mmds(struct efx_nic *efx, 97int efx_mdio_wait_reset_mmds(struct efx_nic *efx, unsigned int mmd_mask);
99 unsigned int mmd_mask);
100 98
101/* Set or clear flag, debouncing */ 99/* Set or clear flag, debouncing */
102static inline void 100static inline void
@@ -107,6 +105,6 @@ efx_mdio_set_flag(struct efx_nic *efx, int devad, int addr,
107} 105}
108 106
109/* Liveness self-test for MDIO PHYs */ 107/* Liveness self-test for MDIO PHYs */
110extern int efx_mdio_test_alive(struct efx_nic *efx); 108int efx_mdio_test_alive(struct efx_nic *efx);
111 109
112#endif /* EFX_MDIO_10G_H */ 110#endif /* EFX_MDIO_10G_H */
diff --git a/drivers/net/ethernet/sfc/net_driver.h b/drivers/net/ethernet/sfc/net_driver.h
index b172ed133055..b14a717ac3e8 100644
--- a/drivers/net/ethernet/sfc/net_driver.h
+++ b/drivers/net/ethernet/sfc/net_driver.h
@@ -141,6 +141,8 @@ struct efx_special_buffer {
141 * @len: Length of this fragment. 141 * @len: Length of this fragment.
142 * This field is zero when the queue slot is empty. 142 * This field is zero when the queue slot is empty.
143 * @unmap_len: Length of this fragment to unmap 143 * @unmap_len: Length of this fragment to unmap
144 * @dma_offset: Offset of @dma_addr from the address of the backing DMA mapping.
145 * Only valid if @unmap_len != 0.
144 */ 146 */
145struct efx_tx_buffer { 147struct efx_tx_buffer {
146 union { 148 union {
@@ -154,6 +156,7 @@ struct efx_tx_buffer {
154 unsigned short flags; 156 unsigned short flags;
155 unsigned short len; 157 unsigned short len;
156 unsigned short unmap_len; 158 unsigned short unmap_len;
159 unsigned short dma_offset;
157}; 160};
158#define EFX_TX_BUF_CONT 1 /* not last descriptor of packet */ 161#define EFX_TX_BUF_CONT 1 /* not last descriptor of packet */
159#define EFX_TX_BUF_SKB 2 /* buffer is last part of skb */ 162#define EFX_TX_BUF_SKB 2 /* buffer is last part of skb */
@@ -182,6 +185,9 @@ struct efx_tx_buffer {
182 * @tsoh_page: Array of pages of TSO header buffers 185 * @tsoh_page: Array of pages of TSO header buffers
183 * @txd: The hardware descriptor ring 186 * @txd: The hardware descriptor ring
184 * @ptr_mask: The size of the ring minus 1. 187 * @ptr_mask: The size of the ring minus 1.
188 * @piobuf: PIO buffer region for this TX queue (shared with its partner).
189 * Size of the region is efx_piobuf_size.
190 * @piobuf_offset: Buffer offset to be specified in PIO descriptors
185 * @initialised: Has hardware queue been initialised? 191 * @initialised: Has hardware queue been initialised?
186 * @read_count: Current read pointer. 192 * @read_count: Current read pointer.
187 * This is the number of buffers that have been removed from both rings. 193 * This is the number of buffers that have been removed from both rings.
@@ -209,6 +215,7 @@ struct efx_tx_buffer {
209 * blocks 215 * blocks
210 * @tso_packets: Number of packets via the TSO xmit path 216 * @tso_packets: Number of packets via the TSO xmit path
211 * @pushes: Number of times the TX push feature has been used 217 * @pushes: Number of times the TX push feature has been used
218 * @pio_packets: Number of times the TX PIO feature has been used
212 * @empty_read_count: If the completion path has seen the queue as empty 219 * @empty_read_count: If the completion path has seen the queue as empty
213 * and the transmission path has not yet checked this, the value of 220 * and the transmission path has not yet checked this, the value of
214 * @read_count bitwise-added to %EFX_EMPTY_COUNT_VALID; otherwise 0. 221 * @read_count bitwise-added to %EFX_EMPTY_COUNT_VALID; otherwise 0.
@@ -223,6 +230,8 @@ struct efx_tx_queue {
223 struct efx_buffer *tsoh_page; 230 struct efx_buffer *tsoh_page;
224 struct efx_special_buffer txd; 231 struct efx_special_buffer txd;
225 unsigned int ptr_mask; 232 unsigned int ptr_mask;
233 void __iomem *piobuf;
234 unsigned int piobuf_offset;
226 bool initialised; 235 bool initialised;
227 236
228 /* Members used mainly on the completion path */ 237 /* Members used mainly on the completion path */
@@ -238,6 +247,7 @@ struct efx_tx_queue {
238 unsigned int tso_long_headers; 247 unsigned int tso_long_headers;
239 unsigned int tso_packets; 248 unsigned int tso_packets;
240 unsigned int pushes; 249 unsigned int pushes;
250 unsigned int pio_packets;
241 251
242 /* Members shared between paths and sometimes updated */ 252 /* Members shared between paths and sometimes updated */
243 unsigned int empty_read_count ____cacheline_aligned_in_smp; 253 unsigned int empty_read_count ____cacheline_aligned_in_smp;
diff --git a/drivers/net/ethernet/sfc/nic.c b/drivers/net/ethernet/sfc/nic.c
index 9826594c8a48..9c90bf56090f 100644
--- a/drivers/net/ethernet/sfc/nic.c
+++ b/drivers/net/ethernet/sfc/nic.c
@@ -19,6 +19,7 @@
19#include "bitfield.h" 19#include "bitfield.h"
20#include "efx.h" 20#include "efx.h"
21#include "nic.h" 21#include "nic.h"
22#include "ef10_regs.h"
22#include "farch_regs.h" 23#include "farch_regs.h"
23#include "io.h" 24#include "io.h"
24#include "workarounds.h" 25#include "workarounds.h"
@@ -166,26 +167,30 @@ void efx_nic_fini_interrupt(struct efx_nic *efx)
166 167
167/* Register dump */ 168/* Register dump */
168 169
169#define REGISTER_REVISION_A 1 170#define REGISTER_REVISION_FA 1
170#define REGISTER_REVISION_B 2 171#define REGISTER_REVISION_FB 2
171#define REGISTER_REVISION_C 3 172#define REGISTER_REVISION_FC 3
172#define REGISTER_REVISION_Z 3 /* latest revision */ 173#define REGISTER_REVISION_FZ 3 /* last Falcon arch revision */
174#define REGISTER_REVISION_ED 4
175#define REGISTER_REVISION_EZ 4 /* latest EF10 revision */
173 176
174struct efx_nic_reg { 177struct efx_nic_reg {
175 u32 offset:24; 178 u32 offset:24;
176 u32 min_revision:2, max_revision:2; 179 u32 min_revision:3, max_revision:3;
177}; 180};
178 181
179#define REGISTER(name, min_rev, max_rev) { \ 182#define REGISTER(name, arch, min_rev, max_rev) { \
180 FR_ ## min_rev ## max_rev ## _ ## name, \ 183 arch ## R_ ## min_rev ## max_rev ## _ ## name, \
181 REGISTER_REVISION_ ## min_rev, REGISTER_REVISION_ ## max_rev \ 184 REGISTER_REVISION_ ## arch ## min_rev, \
185 REGISTER_REVISION_ ## arch ## max_rev \
182} 186}
183#define REGISTER_AA(name) REGISTER(name, A, A) 187#define REGISTER_AA(name) REGISTER(name, F, A, A)
184#define REGISTER_AB(name) REGISTER(name, A, B) 188#define REGISTER_AB(name) REGISTER(name, F, A, B)
185#define REGISTER_AZ(name) REGISTER(name, A, Z) 189#define REGISTER_AZ(name) REGISTER(name, F, A, Z)
186#define REGISTER_BB(name) REGISTER(name, B, B) 190#define REGISTER_BB(name) REGISTER(name, F, B, B)
187#define REGISTER_BZ(name) REGISTER(name, B, Z) 191#define REGISTER_BZ(name) REGISTER(name, F, B, Z)
188#define REGISTER_CZ(name) REGISTER(name, C, Z) 192#define REGISTER_CZ(name) REGISTER(name, F, C, Z)
193#define REGISTER_DZ(name) REGISTER(name, E, D, Z)
189 194
190static const struct efx_nic_reg efx_nic_regs[] = { 195static const struct efx_nic_reg efx_nic_regs[] = {
191 REGISTER_AZ(ADR_REGION), 196 REGISTER_AZ(ADR_REGION),
@@ -292,37 +297,42 @@ static const struct efx_nic_reg efx_nic_regs[] = {
292 REGISTER_AB(XX_TXDRV_CTL), 297 REGISTER_AB(XX_TXDRV_CTL),
293 /* XX_PRBS_CTL, XX_PRBS_CHK and XX_PRBS_ERR are not used */ 298 /* XX_PRBS_CTL, XX_PRBS_CHK and XX_PRBS_ERR are not used */
294 /* XX_CORE_STAT is partly RC */ 299 /* XX_CORE_STAT is partly RC */
300 REGISTER_DZ(BIU_HW_REV_ID),
301 REGISTER_DZ(MC_DB_LWRD),
302 REGISTER_DZ(MC_DB_HWRD),
295}; 303};
296 304
297struct efx_nic_reg_table { 305struct efx_nic_reg_table {
298 u32 offset:24; 306 u32 offset:24;
299 u32 min_revision:2, max_revision:2; 307 u32 min_revision:3, max_revision:3;
300 u32 step:6, rows:21; 308 u32 step:6, rows:21;
301}; 309};
302 310
303#define REGISTER_TABLE_DIMENSIONS(_, offset, min_rev, max_rev, step, rows) { \ 311#define REGISTER_TABLE_DIMENSIONS(_, offset, arch, min_rev, max_rev, step, rows) { \
304 offset, \ 312 offset, \
305 REGISTER_REVISION_ ## min_rev, REGISTER_REVISION_ ## max_rev, \ 313 REGISTER_REVISION_ ## arch ## min_rev, \
314 REGISTER_REVISION_ ## arch ## max_rev, \
306 step, rows \ 315 step, rows \
307} 316}
308#define REGISTER_TABLE(name, min_rev, max_rev) \ 317#define REGISTER_TABLE(name, arch, min_rev, max_rev) \
309 REGISTER_TABLE_DIMENSIONS( \ 318 REGISTER_TABLE_DIMENSIONS( \
310 name, FR_ ## min_rev ## max_rev ## _ ## name, \ 319 name, arch ## R_ ## min_rev ## max_rev ## _ ## name, \
311 min_rev, max_rev, \ 320 arch, min_rev, max_rev, \
312 FR_ ## min_rev ## max_rev ## _ ## name ## _STEP, \ 321 arch ## R_ ## min_rev ## max_rev ## _ ## name ## _STEP, \
313 FR_ ## min_rev ## max_rev ## _ ## name ## _ROWS) 322 arch ## R_ ## min_rev ## max_rev ## _ ## name ## _ROWS)
314#define REGISTER_TABLE_AA(name) REGISTER_TABLE(name, A, A) 323#define REGISTER_TABLE_AA(name) REGISTER_TABLE(name, F, A, A)
315#define REGISTER_TABLE_AZ(name) REGISTER_TABLE(name, A, Z) 324#define REGISTER_TABLE_AZ(name) REGISTER_TABLE(name, F, A, Z)
316#define REGISTER_TABLE_BB(name) REGISTER_TABLE(name, B, B) 325#define REGISTER_TABLE_BB(name) REGISTER_TABLE(name, F, B, B)
317#define REGISTER_TABLE_BZ(name) REGISTER_TABLE(name, B, Z) 326#define REGISTER_TABLE_BZ(name) REGISTER_TABLE(name, F, B, Z)
318#define REGISTER_TABLE_BB_CZ(name) \ 327#define REGISTER_TABLE_BB_CZ(name) \
319 REGISTER_TABLE_DIMENSIONS(name, FR_BZ_ ## name, B, B, \ 328 REGISTER_TABLE_DIMENSIONS(name, FR_BZ_ ## name, F, B, B, \
320 FR_BZ_ ## name ## _STEP, \ 329 FR_BZ_ ## name ## _STEP, \
321 FR_BB_ ## name ## _ROWS), \ 330 FR_BB_ ## name ## _ROWS), \
322 REGISTER_TABLE_DIMENSIONS(name, FR_BZ_ ## name, C, Z, \ 331 REGISTER_TABLE_DIMENSIONS(name, FR_BZ_ ## name, F, C, Z, \
323 FR_BZ_ ## name ## _STEP, \ 332 FR_BZ_ ## name ## _STEP, \
324 FR_CZ_ ## name ## _ROWS) 333 FR_CZ_ ## name ## _ROWS)
325#define REGISTER_TABLE_CZ(name) REGISTER_TABLE(name, C, Z) 334#define REGISTER_TABLE_CZ(name) REGISTER_TABLE(name, F, C, Z)
335#define REGISTER_TABLE_DZ(name) REGISTER_TABLE(name, E, D, Z)
326 336
327static const struct efx_nic_reg_table efx_nic_reg_tables[] = { 337static const struct efx_nic_reg_table efx_nic_reg_tables[] = {
328 /* DRIVER is not used */ 338 /* DRIVER is not used */
@@ -340,9 +350,9 @@ static const struct efx_nic_reg_table efx_nic_reg_tables[] = {
340 * 1K entries allows for some expansion of queue count and 350 * 1K entries allows for some expansion of queue count and
341 * size before we need to change the version. */ 351 * size before we need to change the version. */
342 REGISTER_TABLE_DIMENSIONS(BUF_FULL_TBL_KER, FR_AA_BUF_FULL_TBL_KER, 352 REGISTER_TABLE_DIMENSIONS(BUF_FULL_TBL_KER, FR_AA_BUF_FULL_TBL_KER,
343 A, A, 8, 1024), 353 F, A, A, 8, 1024),
344 REGISTER_TABLE_DIMENSIONS(BUF_FULL_TBL, FR_BZ_BUF_FULL_TBL, 354 REGISTER_TABLE_DIMENSIONS(BUF_FULL_TBL, FR_BZ_BUF_FULL_TBL,
345 B, Z, 8, 1024), 355 F, B, Z, 8, 1024),
346 REGISTER_TABLE_CZ(RX_MAC_FILTER_TBL0), 356 REGISTER_TABLE_CZ(RX_MAC_FILTER_TBL0),
347 REGISTER_TABLE_BB_CZ(TIMER_TBL), 357 REGISTER_TABLE_BB_CZ(TIMER_TBL),
348 REGISTER_TABLE_BB_CZ(TX_PACE_TBL), 358 REGISTER_TABLE_BB_CZ(TX_PACE_TBL),
@@ -353,6 +363,7 @@ static const struct efx_nic_reg_table efx_nic_reg_tables[] = {
353 /* MSIX_PBA_TABLE is not mapped */ 363 /* MSIX_PBA_TABLE is not mapped */
354 /* SRM_DBG is not mapped (and is redundant with BUF_FLL_TBL) */ 364 /* SRM_DBG is not mapped (and is redundant with BUF_FLL_TBL) */
355 REGISTER_TABLE_BZ(RX_FILTER_TBL0), 365 REGISTER_TABLE_BZ(RX_FILTER_TBL0),
366 REGISTER_TABLE_DZ(BIU_MC_SFT_STATUS),
356}; 367};
357 368
358size_t efx_nic_get_regs_len(struct efx_nic *efx) 369size_t efx_nic_get_regs_len(struct efx_nic *efx)
diff --git a/drivers/net/ethernet/sfc/nic.h b/drivers/net/ethernet/sfc/nic.h
index 890bbbe8320e..11b6112d9249 100644
--- a/drivers/net/ethernet/sfc/nic.h
+++ b/drivers/net/ethernet/sfc/nic.h
@@ -30,7 +30,7 @@ static inline int efx_nic_rev(struct efx_nic *efx)
30 return efx->type->revision; 30 return efx->type->revision;
31} 31}
32 32
33extern u32 efx_farch_fpga_ver(struct efx_nic *efx); 33u32 efx_farch_fpga_ver(struct efx_nic *efx);
34 34
35/* NIC has two interlinked PCI functions for the same port. */ 35/* NIC has two interlinked PCI functions for the same port. */
36static inline bool efx_nic_is_dual_func(struct efx_nic *efx) 36static inline bool efx_nic_is_dual_func(struct efx_nic *efx)
@@ -71,6 +71,26 @@ efx_tx_desc(struct efx_tx_queue *tx_queue, unsigned int index)
71 return ((efx_qword_t *) (tx_queue->txd.buf.addr)) + index; 71 return ((efx_qword_t *) (tx_queue->txd.buf.addr)) + index;
72} 72}
73 73
74/* Report whether the NIC considers this TX queue empty, given the
75 * write_count used for the last doorbell push. May return false
76 * negative.
77 */
78static inline bool __efx_nic_tx_is_empty(struct efx_tx_queue *tx_queue,
79 unsigned int write_count)
80{
81 unsigned int empty_read_count = ACCESS_ONCE(tx_queue->empty_read_count);
82
83 if (empty_read_count == 0)
84 return false;
85
86 return ((empty_read_count ^ write_count) & ~EFX_EMPTY_COUNT_VALID) == 0;
87}
88
89static inline bool efx_nic_tx_is_empty(struct efx_tx_queue *tx_queue)
90{
91 return __efx_nic_tx_is_empty(tx_queue, tx_queue->write_count);
92}
93
74/* Decide whether to push a TX descriptor to the NIC vs merely writing 94/* Decide whether to push a TX descriptor to the NIC vs merely writing
75 * the doorbell. This can reduce latency when we are adding a single 95 * the doorbell. This can reduce latency when we are adding a single
76 * descriptor to an empty queue, but is otherwise pointless. Further, 96 * descriptor to an empty queue, but is otherwise pointless. Further,
@@ -80,14 +100,10 @@ efx_tx_desc(struct efx_tx_queue *tx_queue, unsigned int index)
80static inline bool efx_nic_may_push_tx_desc(struct efx_tx_queue *tx_queue, 100static inline bool efx_nic_may_push_tx_desc(struct efx_tx_queue *tx_queue,
81 unsigned int write_count) 101 unsigned int write_count)
82{ 102{
83 unsigned empty_read_count = ACCESS_ONCE(tx_queue->empty_read_count); 103 bool was_empty = __efx_nic_tx_is_empty(tx_queue, write_count);
84
85 if (empty_read_count == 0)
86 return false;
87 104
88 tx_queue->empty_read_count = 0; 105 tx_queue->empty_read_count = 0;
89 return ((empty_read_count ^ write_count) & ~EFX_EMPTY_COUNT_VALID) == 0 106 return was_empty && tx_queue->write_count - write_count == 1;
90 && tx_queue->write_count - write_count == 1;
91} 107}
92 108
93/* Returns a pointer to the specified descriptor in the RX descriptor queue */ 109/* Returns a pointer to the specified descriptor in the RX descriptor queue */
@@ -401,6 +417,12 @@ enum {
401 EF10_STAT_COUNT 417 EF10_STAT_COUNT
402}; 418};
403 419
420/* Maximum number of TX PIO buffers we may allocate to a function.
421 * This matches the total number of buffers on each SFC9100-family
422 * controller.
423 */
424#define EF10_TX_PIOBUF_COUNT 16
425
404/** 426/**
405 * struct efx_ef10_nic_data - EF10 architecture NIC state 427 * struct efx_ef10_nic_data - EF10 architecture NIC state
406 * @mcdi_buf: DMA buffer for MCDI 428 * @mcdi_buf: DMA buffer for MCDI
@@ -409,6 +431,13 @@ enum {
409 * @n_allocated_vis: Number of VIs allocated to this function 431 * @n_allocated_vis: Number of VIs allocated to this function
410 * @must_realloc_vis: Flag: VIs have yet to be reallocated after MC reboot 432 * @must_realloc_vis: Flag: VIs have yet to be reallocated after MC reboot
411 * @must_restore_filters: Flag: filters have yet to be restored after MC reboot 433 * @must_restore_filters: Flag: filters have yet to be restored after MC reboot
434 * @n_piobufs: Number of PIO buffers allocated to this function
435 * @wc_membase: Base address of write-combining mapping of the memory BAR
436 * @pio_write_base: Base address for writing PIO buffers
437 * @pio_write_vi_base: Relative VI number for @pio_write_base
438 * @piobuf_handle: Handle of each PIO buffer allocated
439 * @must_restore_piobufs: Flag: PIO buffers have yet to be restored after MC
440 * reboot
412 * @rx_rss_context: Firmware handle for our RSS context 441 * @rx_rss_context: Firmware handle for our RSS context
413 * @stats: Hardware statistics 442 * @stats: Hardware statistics
414 * @workaround_35388: Flag: firmware supports workaround for bug 35388 443 * @workaround_35388: Flag: firmware supports workaround for bug 35388
@@ -424,6 +453,11 @@ struct efx_ef10_nic_data {
424 unsigned int n_allocated_vis; 453 unsigned int n_allocated_vis;
425 bool must_realloc_vis; 454 bool must_realloc_vis;
426 bool must_restore_filters; 455 bool must_restore_filters;
456 unsigned int n_piobufs;
457 void __iomem *wc_membase, *pio_write_base;
458 unsigned int pio_write_vi_base;
459 unsigned int piobuf_handle[EF10_TX_PIOBUF_COUNT];
460 bool must_restore_piobufs;
427 u32 rx_rss_context; 461 u32 rx_rss_context;
428 u64 stats[EF10_STAT_COUNT]; 462 u64 stats[EF10_STAT_COUNT];
429 bool workaround_35388; 463 bool workaround_35388;
@@ -475,18 +509,18 @@ static inline unsigned int efx_vf_size(struct efx_nic *efx)
475 return 1 << efx->vi_scale; 509 return 1 << efx->vi_scale;
476} 510}
477 511
478extern int efx_init_sriov(void); 512int efx_init_sriov(void);
479extern void efx_sriov_probe(struct efx_nic *efx); 513void efx_sriov_probe(struct efx_nic *efx);
480extern int efx_sriov_init(struct efx_nic *efx); 514int efx_sriov_init(struct efx_nic *efx);
481extern void efx_sriov_mac_address_changed(struct efx_nic *efx); 515void efx_sriov_mac_address_changed(struct efx_nic *efx);
482extern void efx_sriov_tx_flush_done(struct efx_nic *efx, efx_qword_t *event); 516void efx_sriov_tx_flush_done(struct efx_nic *efx, efx_qword_t *event);
483extern void efx_sriov_rx_flush_done(struct efx_nic *efx, efx_qword_t *event); 517void efx_sriov_rx_flush_done(struct efx_nic *efx, efx_qword_t *event);
484extern void efx_sriov_event(struct efx_channel *channel, efx_qword_t *event); 518void efx_sriov_event(struct efx_channel *channel, efx_qword_t *event);
485extern void efx_sriov_desc_fetch_err(struct efx_nic *efx, unsigned dmaq); 519void efx_sriov_desc_fetch_err(struct efx_nic *efx, unsigned dmaq);
486extern void efx_sriov_flr(struct efx_nic *efx, unsigned flr); 520void efx_sriov_flr(struct efx_nic *efx, unsigned flr);
487extern void efx_sriov_reset(struct efx_nic *efx); 521void efx_sriov_reset(struct efx_nic *efx);
488extern void efx_sriov_fini(struct efx_nic *efx); 522void efx_sriov_fini(struct efx_nic *efx);
489extern void efx_fini_sriov(void); 523void efx_fini_sriov(void);
490 524
491#else 525#else
492 526
@@ -512,22 +546,20 @@ static inline void efx_fini_sriov(void) {}
512 546
513#endif 547#endif
514 548
515extern int efx_sriov_set_vf_mac(struct net_device *dev, int vf, u8 *mac); 549int efx_sriov_set_vf_mac(struct net_device *dev, int vf, u8 *mac);
516extern int efx_sriov_set_vf_vlan(struct net_device *dev, int vf, 550int efx_sriov_set_vf_vlan(struct net_device *dev, int vf, u16 vlan, u8 qos);
517 u16 vlan, u8 qos); 551int efx_sriov_get_vf_config(struct net_device *dev, int vf,
518extern int efx_sriov_get_vf_config(struct net_device *dev, int vf, 552 struct ifla_vf_info *ivf);
519 struct ifla_vf_info *ivf); 553int efx_sriov_set_vf_spoofchk(struct net_device *net_dev, int vf,
520extern int efx_sriov_set_vf_spoofchk(struct net_device *net_dev, int vf, 554 bool spoofchk);
521 bool spoofchk);
522 555
523struct ethtool_ts_info; 556struct ethtool_ts_info;
524extern void efx_ptp_probe(struct efx_nic *efx); 557void efx_ptp_probe(struct efx_nic *efx);
525extern int efx_ptp_ioctl(struct efx_nic *efx, struct ifreq *ifr, int cmd); 558int efx_ptp_ioctl(struct efx_nic *efx, struct ifreq *ifr, int cmd);
526extern void efx_ptp_get_ts_info(struct efx_nic *efx, 559void efx_ptp_get_ts_info(struct efx_nic *efx, struct ethtool_ts_info *ts_info);
527 struct ethtool_ts_info *ts_info); 560bool efx_ptp_is_ptp_tx(struct efx_nic *efx, struct sk_buff *skb);
528extern bool efx_ptp_is_ptp_tx(struct efx_nic *efx, struct sk_buff *skb); 561int efx_ptp_tx(struct efx_nic *efx, struct sk_buff *skb);
529extern int efx_ptp_tx(struct efx_nic *efx, struct sk_buff *skb); 562void efx_ptp_event(struct efx_nic *efx, efx_qword_t *ev);
530extern void efx_ptp_event(struct efx_nic *efx, efx_qword_t *ev);
531 563
532extern const struct efx_nic_type falcon_a1_nic_type; 564extern const struct efx_nic_type falcon_a1_nic_type;
533extern const struct efx_nic_type falcon_b0_nic_type; 565extern const struct efx_nic_type falcon_b0_nic_type;
@@ -541,7 +573,7 @@ extern const struct efx_nic_type efx_hunt_a0_nic_type;
541 ************************************************************************** 573 **************************************************************************
542 */ 574 */
543 575
544extern int falcon_probe_board(struct efx_nic *efx, u16 revision_info); 576int falcon_probe_board(struct efx_nic *efx, u16 revision_info);
545 577
546/* TX data path */ 578/* TX data path */
547static inline int efx_nic_probe_tx(struct efx_tx_queue *tx_queue) 579static inline int efx_nic_probe_tx(struct efx_tx_queue *tx_queue)
@@ -609,58 +641,58 @@ static inline void efx_nic_eventq_read_ack(struct efx_channel *channel)
609{ 641{
610 channel->efx->type->ev_read_ack(channel); 642 channel->efx->type->ev_read_ack(channel);
611} 643}
612extern void efx_nic_event_test_start(struct efx_channel *channel); 644void efx_nic_event_test_start(struct efx_channel *channel);
613 645
614/* Falcon/Siena queue operations */ 646/* Falcon/Siena queue operations */
615extern int efx_farch_tx_probe(struct efx_tx_queue *tx_queue); 647int efx_farch_tx_probe(struct efx_tx_queue *tx_queue);
616extern void efx_farch_tx_init(struct efx_tx_queue *tx_queue); 648void efx_farch_tx_init(struct efx_tx_queue *tx_queue);
617extern void efx_farch_tx_fini(struct efx_tx_queue *tx_queue); 649void efx_farch_tx_fini(struct efx_tx_queue *tx_queue);
618extern void efx_farch_tx_remove(struct efx_tx_queue *tx_queue); 650void efx_farch_tx_remove(struct efx_tx_queue *tx_queue);
619extern void efx_farch_tx_write(struct efx_tx_queue *tx_queue); 651void efx_farch_tx_write(struct efx_tx_queue *tx_queue);
620extern int efx_farch_rx_probe(struct efx_rx_queue *rx_queue); 652int efx_farch_rx_probe(struct efx_rx_queue *rx_queue);
621extern void efx_farch_rx_init(struct efx_rx_queue *rx_queue); 653void efx_farch_rx_init(struct efx_rx_queue *rx_queue);
622extern void efx_farch_rx_fini(struct efx_rx_queue *rx_queue); 654void efx_farch_rx_fini(struct efx_rx_queue *rx_queue);
623extern void efx_farch_rx_remove(struct efx_rx_queue *rx_queue); 655void efx_farch_rx_remove(struct efx_rx_queue *rx_queue);
624extern void efx_farch_rx_write(struct efx_rx_queue *rx_queue); 656void efx_farch_rx_write(struct efx_rx_queue *rx_queue);
625extern void efx_farch_rx_defer_refill(struct efx_rx_queue *rx_queue); 657void efx_farch_rx_defer_refill(struct efx_rx_queue *rx_queue);
626extern int efx_farch_ev_probe(struct efx_channel *channel); 658int efx_farch_ev_probe(struct efx_channel *channel);
627extern int efx_farch_ev_init(struct efx_channel *channel); 659int efx_farch_ev_init(struct efx_channel *channel);
628extern void efx_farch_ev_fini(struct efx_channel *channel); 660void efx_farch_ev_fini(struct efx_channel *channel);
629extern void efx_farch_ev_remove(struct efx_channel *channel); 661void efx_farch_ev_remove(struct efx_channel *channel);
630extern int efx_farch_ev_process(struct efx_channel *channel, int quota); 662int efx_farch_ev_process(struct efx_channel *channel, int quota);
631extern void efx_farch_ev_read_ack(struct efx_channel *channel); 663void efx_farch_ev_read_ack(struct efx_channel *channel);
632extern void efx_farch_ev_test_generate(struct efx_channel *channel); 664void efx_farch_ev_test_generate(struct efx_channel *channel);
633 665
634/* Falcon/Siena filter operations */ 666/* Falcon/Siena filter operations */
635extern int efx_farch_filter_table_probe(struct efx_nic *efx); 667int efx_farch_filter_table_probe(struct efx_nic *efx);
636extern void efx_farch_filter_table_restore(struct efx_nic *efx); 668void efx_farch_filter_table_restore(struct efx_nic *efx);
637extern void efx_farch_filter_table_remove(struct efx_nic *efx); 669void efx_farch_filter_table_remove(struct efx_nic *efx);
638extern void efx_farch_filter_update_rx_scatter(struct efx_nic *efx); 670void efx_farch_filter_update_rx_scatter(struct efx_nic *efx);
639extern s32 efx_farch_filter_insert(struct efx_nic *efx, 671s32 efx_farch_filter_insert(struct efx_nic *efx, struct efx_filter_spec *spec,
640 struct efx_filter_spec *spec, bool replace); 672 bool replace);
641extern int efx_farch_filter_remove_safe(struct efx_nic *efx, 673int efx_farch_filter_remove_safe(struct efx_nic *efx,
642 enum efx_filter_priority priority, 674 enum efx_filter_priority priority,
643 u32 filter_id); 675 u32 filter_id);
644extern int efx_farch_filter_get_safe(struct efx_nic *efx, 676int efx_farch_filter_get_safe(struct efx_nic *efx,
645 enum efx_filter_priority priority, 677 enum efx_filter_priority priority, u32 filter_id,
646 u32 filter_id, struct efx_filter_spec *); 678 struct efx_filter_spec *);
647extern void efx_farch_filter_clear_rx(struct efx_nic *efx, 679void efx_farch_filter_clear_rx(struct efx_nic *efx,
648 enum efx_filter_priority priority); 680 enum efx_filter_priority priority);
649extern u32 efx_farch_filter_count_rx_used(struct efx_nic *efx, 681u32 efx_farch_filter_count_rx_used(struct efx_nic *efx,
650 enum efx_filter_priority priority); 682 enum efx_filter_priority priority);
651extern u32 efx_farch_filter_get_rx_id_limit(struct efx_nic *efx); 683u32 efx_farch_filter_get_rx_id_limit(struct efx_nic *efx);
652extern s32 efx_farch_filter_get_rx_ids(struct efx_nic *efx, 684s32 efx_farch_filter_get_rx_ids(struct efx_nic *efx,
653 enum efx_filter_priority priority, 685 enum efx_filter_priority priority, u32 *buf,
654 u32 *buf, u32 size); 686 u32 size);
655#ifdef CONFIG_RFS_ACCEL 687#ifdef CONFIG_RFS_ACCEL
656extern s32 efx_farch_filter_rfs_insert(struct efx_nic *efx, 688s32 efx_farch_filter_rfs_insert(struct efx_nic *efx,
657 struct efx_filter_spec *spec); 689 struct efx_filter_spec *spec);
658extern bool efx_farch_filter_rfs_expire_one(struct efx_nic *efx, u32 flow_id, 690bool efx_farch_filter_rfs_expire_one(struct efx_nic *efx, u32 flow_id,
659 unsigned int index); 691 unsigned int index);
660#endif 692#endif
661extern void efx_farch_filter_sync_rx_mode(struct efx_nic *efx); 693void efx_farch_filter_sync_rx_mode(struct efx_nic *efx);
662 694
663extern bool efx_nic_event_present(struct efx_channel *channel); 695bool efx_nic_event_present(struct efx_channel *channel);
664 696
665/* Some statistics are computed as A - B where A and B each increase 697/* Some statistics are computed as A - B where A and B each increase
666 * linearly with some hardware counter(s) and the counters are read 698 * linearly with some hardware counter(s) and the counters are read
@@ -681,17 +713,17 @@ static inline void efx_update_diff_stat(u64 *stat, u64 diff)
681} 713}
682 714
683/* Interrupts */ 715/* Interrupts */
684extern int efx_nic_init_interrupt(struct efx_nic *efx); 716int efx_nic_init_interrupt(struct efx_nic *efx);
685extern void efx_nic_irq_test_start(struct efx_nic *efx); 717void efx_nic_irq_test_start(struct efx_nic *efx);
686extern void efx_nic_fini_interrupt(struct efx_nic *efx); 718void efx_nic_fini_interrupt(struct efx_nic *efx);
687 719
688/* Falcon/Siena interrupts */ 720/* Falcon/Siena interrupts */
689extern void efx_farch_irq_enable_master(struct efx_nic *efx); 721void efx_farch_irq_enable_master(struct efx_nic *efx);
690extern void efx_farch_irq_test_generate(struct efx_nic *efx); 722void efx_farch_irq_test_generate(struct efx_nic *efx);
691extern void efx_farch_irq_disable_master(struct efx_nic *efx); 723void efx_farch_irq_disable_master(struct efx_nic *efx);
692extern irqreturn_t efx_farch_msi_interrupt(int irq, void *dev_id); 724irqreturn_t efx_farch_msi_interrupt(int irq, void *dev_id);
693extern irqreturn_t efx_farch_legacy_interrupt(int irq, void *dev_id); 725irqreturn_t efx_farch_legacy_interrupt(int irq, void *dev_id);
694extern irqreturn_t efx_farch_fatal_interrupt(struct efx_nic *efx); 726irqreturn_t efx_farch_fatal_interrupt(struct efx_nic *efx);
695 727
696static inline int efx_nic_event_test_irq_cpu(struct efx_channel *channel) 728static inline int efx_nic_event_test_irq_cpu(struct efx_channel *channel)
697{ 729{
@@ -703,21 +735,21 @@ static inline int efx_nic_irq_test_irq_cpu(struct efx_nic *efx)
703} 735}
704 736
705/* Global Resources */ 737/* Global Resources */
706extern int efx_nic_flush_queues(struct efx_nic *efx); 738int efx_nic_flush_queues(struct efx_nic *efx);
707extern void siena_prepare_flush(struct efx_nic *efx); 739void siena_prepare_flush(struct efx_nic *efx);
708extern int efx_farch_fini_dmaq(struct efx_nic *efx); 740int efx_farch_fini_dmaq(struct efx_nic *efx);
709extern void siena_finish_flush(struct efx_nic *efx); 741void siena_finish_flush(struct efx_nic *efx);
710extern void falcon_start_nic_stats(struct efx_nic *efx); 742void falcon_start_nic_stats(struct efx_nic *efx);
711extern void falcon_stop_nic_stats(struct efx_nic *efx); 743void falcon_stop_nic_stats(struct efx_nic *efx);
712extern int falcon_reset_xaui(struct efx_nic *efx); 744int falcon_reset_xaui(struct efx_nic *efx);
713extern void efx_farch_dimension_resources(struct efx_nic *efx, unsigned sram_lim_qw); 745void efx_farch_dimension_resources(struct efx_nic *efx, unsigned sram_lim_qw);
714extern void efx_farch_init_common(struct efx_nic *efx); 746void efx_farch_init_common(struct efx_nic *efx);
715extern void efx_ef10_handle_drain_event(struct efx_nic *efx); 747void efx_ef10_handle_drain_event(struct efx_nic *efx);
716static inline void efx_nic_push_rx_indir_table(struct efx_nic *efx) 748static inline void efx_nic_push_rx_indir_table(struct efx_nic *efx)
717{ 749{
718 efx->type->rx_push_indir_table(efx); 750 efx->type->rx_push_indir_table(efx);
719} 751}
720extern void efx_farch_rx_push_indir_table(struct efx_nic *efx); 752void efx_farch_rx_push_indir_table(struct efx_nic *efx);
721 753
722int efx_nic_alloc_buffer(struct efx_nic *efx, struct efx_buffer *buffer, 754int efx_nic_alloc_buffer(struct efx_nic *efx, struct efx_buffer *buffer,
723 unsigned int len, gfp_t gfp_flags); 755 unsigned int len, gfp_t gfp_flags);
@@ -728,24 +760,22 @@ struct efx_farch_register_test {
728 unsigned address; 760 unsigned address;
729 efx_oword_t mask; 761 efx_oword_t mask;
730}; 762};
731extern int efx_farch_test_registers(struct efx_nic *efx, 763int efx_farch_test_registers(struct efx_nic *efx,
732 const struct efx_farch_register_test *regs, 764 const struct efx_farch_register_test *regs,
733 size_t n_regs); 765 size_t n_regs);
734 766
735extern size_t efx_nic_get_regs_len(struct efx_nic *efx); 767size_t efx_nic_get_regs_len(struct efx_nic *efx);
736extern void efx_nic_get_regs(struct efx_nic *efx, void *buf); 768void efx_nic_get_regs(struct efx_nic *efx, void *buf);
737 769
738extern size_t 770size_t efx_nic_describe_stats(const struct efx_hw_stat_desc *desc, size_t count,
739efx_nic_describe_stats(const struct efx_hw_stat_desc *desc, size_t count, 771 const unsigned long *mask, u8 *names);
740 const unsigned long *mask, u8 *names); 772void efx_nic_update_stats(const struct efx_hw_stat_desc *desc, size_t count,
741extern void 773 const unsigned long *mask, u64 *stats,
742efx_nic_update_stats(const struct efx_hw_stat_desc *desc, size_t count, 774 const void *dma_buf, bool accumulate);
743 const unsigned long *mask,
744 u64 *stats, const void *dma_buf, bool accumulate);
745 775
746#define EFX_MAX_FLUSH_TIME 5000 776#define EFX_MAX_FLUSH_TIME 5000
747 777
748extern void efx_farch_generate_event(struct efx_nic *efx, unsigned int evq, 778void efx_farch_generate_event(struct efx_nic *efx, unsigned int evq,
749 efx_qword_t *event); 779 efx_qword_t *event);
750 780
751#endif /* EFX_NIC_H */ 781#endif /* EFX_NIC_H */
diff --git a/drivers/net/ethernet/sfc/phy.h b/drivers/net/ethernet/sfc/phy.h
index 45eeb7075156..803bf445c08e 100644
--- a/drivers/net/ethernet/sfc/phy.h
+++ b/drivers/net/ethernet/sfc/phy.h
@@ -15,7 +15,7 @@
15 */ 15 */
16extern const struct efx_phy_operations falcon_sfx7101_phy_ops; 16extern const struct efx_phy_operations falcon_sfx7101_phy_ops;
17 17
18extern void tenxpress_set_id_led(struct efx_nic *efx, enum efx_led_mode mode); 18void tenxpress_set_id_led(struct efx_nic *efx, enum efx_led_mode mode);
19 19
20/**************************************************************************** 20/****************************************************************************
21 * AMCC/Quake QT202x PHYs 21 * AMCC/Quake QT202x PHYs
@@ -34,7 +34,7 @@ extern const struct efx_phy_operations falcon_qt202x_phy_ops;
34#define QUAKE_LED_TXLINK (0) 34#define QUAKE_LED_TXLINK (0)
35#define QUAKE_LED_RXLINK (8) 35#define QUAKE_LED_RXLINK (8)
36 36
37extern void falcon_qt202x_set_led(struct efx_nic *p, int led, int state); 37void falcon_qt202x_set_led(struct efx_nic *p, int led, int state);
38 38
39/**************************************************************************** 39/****************************************************************************
40* Transwitch CX4 retimer 40* Transwitch CX4 retimer
@@ -44,7 +44,7 @@ extern const struct efx_phy_operations falcon_txc_phy_ops;
44#define TXC_GPIO_DIR_INPUT 0 44#define TXC_GPIO_DIR_INPUT 0
45#define TXC_GPIO_DIR_OUTPUT 1 45#define TXC_GPIO_DIR_OUTPUT 1
46 46
47extern void falcon_txc_set_gpio_dir(struct efx_nic *efx, int pin, int dir); 47void falcon_txc_set_gpio_dir(struct efx_nic *efx, int pin, int dir);
48extern void falcon_txc_set_gpio_val(struct efx_nic *efx, int pin, int val); 48void falcon_txc_set_gpio_val(struct efx_nic *efx, int pin, int val);
49 49
50#endif 50#endif
diff --git a/drivers/net/ethernet/sfc/rx.c b/drivers/net/ethernet/sfc/rx.c
index 4a596725023f..8f09e686fc23 100644
--- a/drivers/net/ethernet/sfc/rx.c
+++ b/drivers/net/ethernet/sfc/rx.c
@@ -12,6 +12,7 @@
12#include <linux/in.h> 12#include <linux/in.h>
13#include <linux/slab.h> 13#include <linux/slab.h>
14#include <linux/ip.h> 14#include <linux/ip.h>
15#include <linux/ipv6.h>
15#include <linux/tcp.h> 16#include <linux/tcp.h>
16#include <linux/udp.h> 17#include <linux/udp.h>
17#include <linux/prefetch.h> 18#include <linux/prefetch.h>
@@ -818,44 +819,70 @@ int efx_filter_rfs(struct net_device *net_dev, const struct sk_buff *skb,
818 struct efx_nic *efx = netdev_priv(net_dev); 819 struct efx_nic *efx = netdev_priv(net_dev);
819 struct efx_channel *channel; 820 struct efx_channel *channel;
820 struct efx_filter_spec spec; 821 struct efx_filter_spec spec;
821 const struct iphdr *ip;
822 const __be16 *ports; 822 const __be16 *ports;
823 __be16 ether_type;
823 int nhoff; 824 int nhoff;
824 int rc; 825 int rc;
825 826
826 nhoff = skb_network_offset(skb); 827 /* The core RPS/RFS code has already parsed and validated
828 * VLAN, IP and transport headers. We assume they are in the
829 * header area.
830 */
827 831
828 if (skb->protocol == htons(ETH_P_8021Q)) { 832 if (skb->protocol == htons(ETH_P_8021Q)) {
829 EFX_BUG_ON_PARANOID(skb_headlen(skb) < 833 const struct vlan_hdr *vh =
830 nhoff + sizeof(struct vlan_hdr)); 834 (const struct vlan_hdr *)skb->data;
831 if (((const struct vlan_hdr *)skb->data + nhoff)->
832 h_vlan_encapsulated_proto != htons(ETH_P_IP))
833 return -EPROTONOSUPPORT;
834 835
835 /* This is IP over 802.1q VLAN. We can't filter on the 836 /* We can't filter on the IP 5-tuple and the vlan
836 * IP 5-tuple and the vlan together, so just strip the 837 * together, so just strip the vlan header and filter
837 * vlan header and filter on the IP part. 838 * on the IP part.
838 */ 839 */
839 nhoff += sizeof(struct vlan_hdr); 840 EFX_BUG_ON_PARANOID(skb_headlen(skb) < sizeof(*vh));
840 } else if (skb->protocol != htons(ETH_P_IP)) { 841 ether_type = vh->h_vlan_encapsulated_proto;
841 return -EPROTONOSUPPORT; 842 nhoff = sizeof(struct vlan_hdr);
843 } else {
844 ether_type = skb->protocol;
845 nhoff = 0;
842 } 846 }
843 847
844 /* RFS must validate the IP header length before calling us */ 848 if (ether_type != htons(ETH_P_IP) && ether_type != htons(ETH_P_IPV6))
845 EFX_BUG_ON_PARANOID(skb_headlen(skb) < nhoff + sizeof(*ip));
846 ip = (const struct iphdr *)(skb->data + nhoff);
847 if (ip_is_fragment(ip))
848 return -EPROTONOSUPPORT; 849 return -EPROTONOSUPPORT;
849 EFX_BUG_ON_PARANOID(skb_headlen(skb) < nhoff + 4 * ip->ihl + 4);
850 ports = (const __be16 *)(skb->data + nhoff + 4 * ip->ihl);
851 850
852 efx_filter_init_rx(&spec, EFX_FILTER_PRI_HINT, 851 efx_filter_init_rx(&spec, EFX_FILTER_PRI_HINT,
853 efx->rx_scatter ? EFX_FILTER_FLAG_RX_SCATTER : 0, 852 efx->rx_scatter ? EFX_FILTER_FLAG_RX_SCATTER : 0,
854 rxq_index); 853 rxq_index);
855 rc = efx_filter_set_ipv4_full(&spec, ip->protocol, 854 spec.match_flags =
856 ip->daddr, ports[1], ip->saddr, ports[0]); 855 EFX_FILTER_MATCH_ETHER_TYPE | EFX_FILTER_MATCH_IP_PROTO |
857 if (rc) 856 EFX_FILTER_MATCH_LOC_HOST | EFX_FILTER_MATCH_LOC_PORT |
858 return rc; 857 EFX_FILTER_MATCH_REM_HOST | EFX_FILTER_MATCH_REM_PORT;
858 spec.ether_type = ether_type;
859
860 if (ether_type == htons(ETH_P_IP)) {
861 const struct iphdr *ip =
862 (const struct iphdr *)(skb->data + nhoff);
863
864 EFX_BUG_ON_PARANOID(skb_headlen(skb) < nhoff + sizeof(*ip));
865 if (ip_is_fragment(ip))
866 return -EPROTONOSUPPORT;
867 spec.ip_proto = ip->protocol;
868 spec.rem_host[0] = ip->saddr;
869 spec.loc_host[0] = ip->daddr;
870 EFX_BUG_ON_PARANOID(skb_headlen(skb) < nhoff + 4 * ip->ihl + 4);
871 ports = (const __be16 *)(skb->data + nhoff + 4 * ip->ihl);
872 } else {
873 const struct ipv6hdr *ip6 =
874 (const struct ipv6hdr *)(skb->data + nhoff);
875
876 EFX_BUG_ON_PARANOID(skb_headlen(skb) <
877 nhoff + sizeof(*ip6) + 4);
878 spec.ip_proto = ip6->nexthdr;
879 memcpy(spec.rem_host, &ip6->saddr, sizeof(ip6->saddr));
880 memcpy(spec.loc_host, &ip6->daddr, sizeof(ip6->daddr));
881 ports = (const __be16 *)(ip6 + 1);
882 }
883
884 spec.rem_port = ports[0];
885 spec.loc_port = ports[1];
859 886
860 rc = efx->type->filter_rfs_insert(efx, &spec); 887 rc = efx->type->filter_rfs_insert(efx, &spec);
861 if (rc < 0) 888 if (rc < 0)
@@ -866,11 +893,18 @@ int efx_filter_rfs(struct net_device *net_dev, const struct sk_buff *skb,
866 channel = efx_get_channel(efx, skb_get_rx_queue(skb)); 893 channel = efx_get_channel(efx, skb_get_rx_queue(skb));
867 ++channel->rfs_filters_added; 894 ++channel->rfs_filters_added;
868 895
869 netif_info(efx, rx_status, efx->net_dev, 896 if (ether_type == htons(ETH_P_IP))
870 "steering %s %pI4:%u:%pI4:%u to queue %u [flow %u filter %d]\n", 897 netif_info(efx, rx_status, efx->net_dev,
871 (ip->protocol == IPPROTO_TCP) ? "TCP" : "UDP", 898 "steering %s %pI4:%u:%pI4:%u to queue %u [flow %u filter %d]\n",
872 &ip->saddr, ntohs(ports[0]), &ip->daddr, ntohs(ports[1]), 899 (spec.ip_proto == IPPROTO_TCP) ? "TCP" : "UDP",
873 rxq_index, flow_id, rc); 900 spec.rem_host, ntohs(ports[0]), spec.loc_host,
901 ntohs(ports[1]), rxq_index, flow_id, rc);
902 else
903 netif_info(efx, rx_status, efx->net_dev,
904 "steering %s [%pI6]:%u:[%pI6]:%u to queue %u [flow %u filter %d]\n",
905 (spec.ip_proto == IPPROTO_TCP) ? "TCP" : "UDP",
906 spec.rem_host, ntohs(ports[0]), spec.loc_host,
907 ntohs(ports[1]), rxq_index, flow_id, rc);
874 908
875 return rc; 909 return rc;
876} 910}
diff --git a/drivers/net/ethernet/sfc/selftest.h b/drivers/net/ethernet/sfc/selftest.h
index 87698ae0bf75..a2f4a06ffa4e 100644
--- a/drivers/net/ethernet/sfc/selftest.h
+++ b/drivers/net/ethernet/sfc/selftest.h
@@ -43,13 +43,12 @@ struct efx_self_tests {
43 struct efx_loopback_self_tests loopback[LOOPBACK_TEST_MAX + 1]; 43 struct efx_loopback_self_tests loopback[LOOPBACK_TEST_MAX + 1];
44}; 44};
45 45
46extern void efx_loopback_rx_packet(struct efx_nic *efx, 46void efx_loopback_rx_packet(struct efx_nic *efx, const char *buf_ptr,
47 const char *buf_ptr, int pkt_len); 47 int pkt_len);
48extern int efx_selftest(struct efx_nic *efx, 48int efx_selftest(struct efx_nic *efx, struct efx_self_tests *tests,
49 struct efx_self_tests *tests, 49 unsigned flags);
50 unsigned flags); 50void efx_selftest_async_start(struct efx_nic *efx);
51extern void efx_selftest_async_start(struct efx_nic *efx); 51void efx_selftest_async_cancel(struct efx_nic *efx);
52extern void efx_selftest_async_cancel(struct efx_nic *efx); 52void efx_selftest_async_work(struct work_struct *data);
53extern void efx_selftest_async_work(struct work_struct *data);
54 53
55#endif /* EFX_SELFTEST_H */ 54#endif /* EFX_SELFTEST_H */
diff --git a/drivers/net/ethernet/sfc/tx.c b/drivers/net/ethernet/sfc/tx.c
index 2ac91c5b5eea..c49d1fb16965 100644
--- a/drivers/net/ethernet/sfc/tx.c
+++ b/drivers/net/ethernet/sfc/tx.c
@@ -17,10 +17,46 @@
17#include <net/ipv6.h> 17#include <net/ipv6.h>
18#include <linux/if_ether.h> 18#include <linux/if_ether.h>
19#include <linux/highmem.h> 19#include <linux/highmem.h>
20#include <linux/cache.h>
20#include "net_driver.h" 21#include "net_driver.h"
21#include "efx.h" 22#include "efx.h"
23#include "io.h"
22#include "nic.h" 24#include "nic.h"
23#include "workarounds.h" 25#include "workarounds.h"
26#include "ef10_regs.h"
27
28#ifdef EFX_USE_PIO
29
30#define EFX_PIOBUF_SIZE_MAX ER_DZ_TX_PIOBUF_SIZE
31#define EFX_PIOBUF_SIZE_DEF ALIGN(256, L1_CACHE_BYTES)
32unsigned int efx_piobuf_size __read_mostly = EFX_PIOBUF_SIZE_DEF;
33
34#endif /* EFX_USE_PIO */
35
36static inline unsigned int
37efx_tx_queue_get_insert_index(const struct efx_tx_queue *tx_queue)
38{
39 return tx_queue->insert_count & tx_queue->ptr_mask;
40}
41
42static inline struct efx_tx_buffer *
43__efx_tx_queue_get_insert_buffer(const struct efx_tx_queue *tx_queue)
44{
45 return &tx_queue->buffer[efx_tx_queue_get_insert_index(tx_queue)];
46}
47
48static inline struct efx_tx_buffer *
49efx_tx_queue_get_insert_buffer(const struct efx_tx_queue *tx_queue)
50{
51 struct efx_tx_buffer *buffer =
52 __efx_tx_queue_get_insert_buffer(tx_queue);
53
54 EFX_BUG_ON_PARANOID(buffer->len);
55 EFX_BUG_ON_PARANOID(buffer->flags);
56 EFX_BUG_ON_PARANOID(buffer->unmap_len);
57
58 return buffer;
59}
24 60
25static void efx_dequeue_buffer(struct efx_tx_queue *tx_queue, 61static void efx_dequeue_buffer(struct efx_tx_queue *tx_queue,
26 struct efx_tx_buffer *buffer, 62 struct efx_tx_buffer *buffer,
@@ -29,8 +65,7 @@ static void efx_dequeue_buffer(struct efx_tx_queue *tx_queue,
29{ 65{
30 if (buffer->unmap_len) { 66 if (buffer->unmap_len) {
31 struct device *dma_dev = &tx_queue->efx->pci_dev->dev; 67 struct device *dma_dev = &tx_queue->efx->pci_dev->dev;
32 dma_addr_t unmap_addr = (buffer->dma_addr + buffer->len - 68 dma_addr_t unmap_addr = buffer->dma_addr - buffer->dma_offset;
33 buffer->unmap_len);
34 if (buffer->flags & EFX_TX_BUF_MAP_SINGLE) 69 if (buffer->flags & EFX_TX_BUF_MAP_SINGLE)
35 dma_unmap_single(dma_dev, unmap_addr, buffer->unmap_len, 70 dma_unmap_single(dma_dev, unmap_addr, buffer->unmap_len,
36 DMA_TO_DEVICE); 71 DMA_TO_DEVICE);
@@ -83,8 +118,10 @@ unsigned int efx_tx_max_skb_descs(struct efx_nic *efx)
83 */ 118 */
84 unsigned int max_descs = EFX_TSO_MAX_SEGS * 2 + MAX_SKB_FRAGS; 119 unsigned int max_descs = EFX_TSO_MAX_SEGS * 2 + MAX_SKB_FRAGS;
85 120
86 /* Possibly one more per segment for the alignment workaround */ 121 /* Possibly one more per segment for the alignment workaround,
87 if (EFX_WORKAROUND_5391(efx)) 122 * or for option descriptors
123 */
124 if (EFX_WORKAROUND_5391(efx) || efx_nic_rev(efx) >= EFX_REV_HUNT_A0)
88 max_descs += EFX_TSO_MAX_SEGS; 125 max_descs += EFX_TSO_MAX_SEGS;
89 126
90 /* Possibly more for PCIe page boundaries within input fragments */ 127 /* Possibly more for PCIe page boundaries within input fragments */
@@ -145,6 +182,145 @@ static void efx_tx_maybe_stop_queue(struct efx_tx_queue *txq1)
145 } 182 }
146} 183}
147 184
185#ifdef EFX_USE_PIO
186
187struct efx_short_copy_buffer {
188 int used;
189 u8 buf[L1_CACHE_BYTES];
190};
191
192/* Copy to PIO, respecting that writes to PIO buffers must be dword aligned.
193 * Advances piobuf pointer. Leaves additional data in the copy buffer.
194 */
195static void efx_memcpy_toio_aligned(struct efx_nic *efx, u8 __iomem **piobuf,
196 u8 *data, int len,
197 struct efx_short_copy_buffer *copy_buf)
198{
199 int block_len = len & ~(sizeof(copy_buf->buf) - 1);
200
201 memcpy_toio(*piobuf, data, block_len);
202 *piobuf += block_len;
203 len -= block_len;
204
205 if (len) {
206 data += block_len;
207 BUG_ON(copy_buf->used);
208 BUG_ON(len > sizeof(copy_buf->buf));
209 memcpy(copy_buf->buf, data, len);
210 copy_buf->used = len;
211 }
212}
213
214/* Copy to PIO, respecting dword alignment, popping data from copy buffer first.
215 * Advances piobuf pointer. Leaves additional data in the copy buffer.
216 */
217static void efx_memcpy_toio_aligned_cb(struct efx_nic *efx, u8 __iomem **piobuf,
218 u8 *data, int len,
219 struct efx_short_copy_buffer *copy_buf)
220{
221 if (copy_buf->used) {
222 /* if the copy buffer is partially full, fill it up and write */
223 int copy_to_buf =
224 min_t(int, sizeof(copy_buf->buf) - copy_buf->used, len);
225
226 memcpy(copy_buf->buf + copy_buf->used, data, copy_to_buf);
227 copy_buf->used += copy_to_buf;
228
229 /* if we didn't fill it up then we're done for now */
230 if (copy_buf->used < sizeof(copy_buf->buf))
231 return;
232
233 memcpy_toio(*piobuf, copy_buf->buf, sizeof(copy_buf->buf));
234 *piobuf += sizeof(copy_buf->buf);
235 data += copy_to_buf;
236 len -= copy_to_buf;
237 copy_buf->used = 0;
238 }
239
240 efx_memcpy_toio_aligned(efx, piobuf, data, len, copy_buf);
241}
242
243static void efx_flush_copy_buffer(struct efx_nic *efx, u8 __iomem *piobuf,
244 struct efx_short_copy_buffer *copy_buf)
245{
246 /* if there's anything in it, write the whole buffer, including junk */
247 if (copy_buf->used)
248 memcpy_toio(piobuf, copy_buf->buf, sizeof(copy_buf->buf));
249}
250
251/* Traverse skb structure and copy fragments in to PIO buffer.
252 * Advances piobuf pointer.
253 */
254static void efx_skb_copy_bits_to_pio(struct efx_nic *efx, struct sk_buff *skb,
255 u8 __iomem **piobuf,
256 struct efx_short_copy_buffer *copy_buf)
257{
258 int i;
259
260 efx_memcpy_toio_aligned(efx, piobuf, skb->data, skb_headlen(skb),
261 copy_buf);
262
263 for (i = 0; i < skb_shinfo(skb)->nr_frags; ++i) {
264 skb_frag_t *f = &skb_shinfo(skb)->frags[i];
265 u8 *vaddr;
266
267 vaddr = kmap_atomic(skb_frag_page(f));
268
269 efx_memcpy_toio_aligned_cb(efx, piobuf, vaddr + f->page_offset,
270 skb_frag_size(f), copy_buf);
271 kunmap_atomic(vaddr);
272 }
273
274 EFX_BUG_ON_PARANOID(skb_shinfo(skb)->frag_list);
275}
276
277static struct efx_tx_buffer *
278efx_enqueue_skb_pio(struct efx_tx_queue *tx_queue, struct sk_buff *skb)
279{
280 struct efx_tx_buffer *buffer =
281 efx_tx_queue_get_insert_buffer(tx_queue);
282 u8 __iomem *piobuf = tx_queue->piobuf;
283
284 /* Copy to PIO buffer. Ensure the writes are padded to the end
285 * of a cache line, as this is required for write-combining to be
286 * effective on at least x86.
287 */
288
289 if (skb_shinfo(skb)->nr_frags) {
290 /* The size of the copy buffer will ensure all writes
291 * are the size of a cache line.
292 */
293 struct efx_short_copy_buffer copy_buf;
294
295 copy_buf.used = 0;
296
297 efx_skb_copy_bits_to_pio(tx_queue->efx, skb,
298 &piobuf, &copy_buf);
299 efx_flush_copy_buffer(tx_queue->efx, piobuf, &copy_buf);
300 } else {
301 /* Pad the write to the size of a cache line.
302 * We can do this because we know the skb_shared_info sruct is
303 * after the source, and the destination buffer is big enough.
304 */
305 BUILD_BUG_ON(L1_CACHE_BYTES >
306 SKB_DATA_ALIGN(sizeof(struct skb_shared_info)));
307 memcpy_toio(tx_queue->piobuf, skb->data,
308 ALIGN(skb->len, L1_CACHE_BYTES));
309 }
310
311 EFX_POPULATE_QWORD_5(buffer->option,
312 ESF_DZ_TX_DESC_IS_OPT, 1,
313 ESF_DZ_TX_OPTION_TYPE, ESE_DZ_TX_OPTION_DESC_PIO,
314 ESF_DZ_TX_PIO_CONT, 0,
315 ESF_DZ_TX_PIO_BYTE_CNT, skb->len,
316 ESF_DZ_TX_PIO_BUF_ADDR,
317 tx_queue->piobuf_offset);
318 ++tx_queue->pio_packets;
319 ++tx_queue->insert_count;
320 return buffer;
321}
322#endif /* EFX_USE_PIO */
323
148/* 324/*
149 * Add a socket buffer to a TX queue 325 * Add a socket buffer to a TX queue
150 * 326 *
@@ -167,7 +343,7 @@ netdev_tx_t efx_enqueue_skb(struct efx_tx_queue *tx_queue, struct sk_buff *skb)
167 struct device *dma_dev = &efx->pci_dev->dev; 343 struct device *dma_dev = &efx->pci_dev->dev;
168 struct efx_tx_buffer *buffer; 344 struct efx_tx_buffer *buffer;
169 skb_frag_t *fragment; 345 skb_frag_t *fragment;
170 unsigned int len, unmap_len = 0, insert_ptr; 346 unsigned int len, unmap_len = 0;
171 dma_addr_t dma_addr, unmap_addr = 0; 347 dma_addr_t dma_addr, unmap_addr = 0;
172 unsigned int dma_len; 348 unsigned int dma_len;
173 unsigned short dma_flags; 349 unsigned short dma_flags;
@@ -189,6 +365,17 @@ netdev_tx_t efx_enqueue_skb(struct efx_tx_queue *tx_queue, struct sk_buff *skb)
189 return NETDEV_TX_OK; 365 return NETDEV_TX_OK;
190 } 366 }
191 367
368 /* Consider using PIO for short packets */
369#ifdef EFX_USE_PIO
370 if (skb->len <= efx_piobuf_size && tx_queue->piobuf &&
371 efx_nic_tx_is_empty(tx_queue) &&
372 efx_nic_tx_is_empty(efx_tx_queue_partner(tx_queue))) {
373 buffer = efx_enqueue_skb_pio(tx_queue, skb);
374 dma_flags = EFX_TX_BUF_OPTION;
375 goto finish_packet;
376 }
377#endif
378
192 /* Map for DMA. Use dma_map_single rather than dma_map_page 379 /* Map for DMA. Use dma_map_single rather than dma_map_page
193 * since this is more efficient on machines with sparse 380 * since this is more efficient on machines with sparse
194 * memory. 381 * memory.
@@ -208,11 +395,7 @@ netdev_tx_t efx_enqueue_skb(struct efx_tx_queue *tx_queue, struct sk_buff *skb)
208 395
209 /* Add to TX queue, splitting across DMA boundaries */ 396 /* Add to TX queue, splitting across DMA boundaries */
210 do { 397 do {
211 insert_ptr = tx_queue->insert_count & tx_queue->ptr_mask; 398 buffer = efx_tx_queue_get_insert_buffer(tx_queue);
212 buffer = &tx_queue->buffer[insert_ptr];
213 EFX_BUG_ON_PARANOID(buffer->flags);
214 EFX_BUG_ON_PARANOID(buffer->len);
215 EFX_BUG_ON_PARANOID(buffer->unmap_len);
216 399
217 dma_len = efx_max_tx_len(efx, dma_addr); 400 dma_len = efx_max_tx_len(efx, dma_addr);
218 if (likely(dma_len >= len)) 401 if (likely(dma_len >= len))
@@ -230,6 +413,7 @@ netdev_tx_t efx_enqueue_skb(struct efx_tx_queue *tx_queue, struct sk_buff *skb)
230 /* Transfer ownership of the unmapping to the final buffer */ 413 /* Transfer ownership of the unmapping to the final buffer */
231 buffer->flags = EFX_TX_BUF_CONT | dma_flags; 414 buffer->flags = EFX_TX_BUF_CONT | dma_flags;
232 buffer->unmap_len = unmap_len; 415 buffer->unmap_len = unmap_len;
416 buffer->dma_offset = buffer->dma_addr - unmap_addr;
233 unmap_len = 0; 417 unmap_len = 0;
234 418
235 /* Get address and size of next fragment */ 419 /* Get address and size of next fragment */
@@ -245,6 +429,7 @@ netdev_tx_t efx_enqueue_skb(struct efx_tx_queue *tx_queue, struct sk_buff *skb)
245 } 429 }
246 430
247 /* Transfer ownership of the skb to the final buffer */ 431 /* Transfer ownership of the skb to the final buffer */
432finish_packet:
248 buffer->skb = skb; 433 buffer->skb = skb;
249 buffer->flags = EFX_TX_BUF_SKB | dma_flags; 434 buffer->flags = EFX_TX_BUF_SKB | dma_flags;
250 435
@@ -270,8 +455,7 @@ netdev_tx_t efx_enqueue_skb(struct efx_tx_queue *tx_queue, struct sk_buff *skb)
270 while (tx_queue->insert_count != tx_queue->write_count) { 455 while (tx_queue->insert_count != tx_queue->write_count) {
271 unsigned int pkts_compl = 0, bytes_compl = 0; 456 unsigned int pkts_compl = 0, bytes_compl = 0;
272 --tx_queue->insert_count; 457 --tx_queue->insert_count;
273 insert_ptr = tx_queue->insert_count & tx_queue->ptr_mask; 458 buffer = __efx_tx_queue_get_insert_buffer(tx_queue);
274 buffer = &tx_queue->buffer[insert_ptr];
275 efx_dequeue_buffer(tx_queue, buffer, &pkts_compl, &bytes_compl); 459 efx_dequeue_buffer(tx_queue, buffer, &pkts_compl, &bytes_compl);
276 } 460 }
277 461
@@ -628,6 +812,9 @@ void efx_remove_tx_queue(struct efx_tx_queue *tx_queue)
628 * @tcp_off: Offset of TCP header 812 * @tcp_off: Offset of TCP header
629 * @header_len: Number of bytes of header 813 * @header_len: Number of bytes of header
630 * @ip_base_len: IPv4 tot_len or IPv6 payload_len, before TCP payload 814 * @ip_base_len: IPv4 tot_len or IPv6 payload_len, before TCP payload
815 * @header_dma_addr: Header DMA address, when using option descriptors
816 * @header_unmap_len: Header DMA mapped length, or 0 if not using option
817 * descriptors
631 * 818 *
632 * The state used during segmentation. It is put into this data structure 819 * The state used during segmentation. It is put into this data structure
633 * just to make it easy to pass into inline functions. 820 * just to make it easy to pass into inline functions.
@@ -636,7 +823,7 @@ struct tso_state {
636 /* Output position */ 823 /* Output position */
637 unsigned out_len; 824 unsigned out_len;
638 unsigned seqnum; 825 unsigned seqnum;
639 unsigned ipv4_id; 826 u16 ipv4_id;
640 unsigned packet_space; 827 unsigned packet_space;
641 828
642 /* Input position */ 829 /* Input position */
@@ -651,6 +838,8 @@ struct tso_state {
651 unsigned int tcp_off; 838 unsigned int tcp_off;
652 unsigned header_len; 839 unsigned header_len;
653 unsigned int ip_base_len; 840 unsigned int ip_base_len;
841 dma_addr_t header_dma_addr;
842 unsigned int header_unmap_len;
654}; 843};
655 844
656 845
@@ -737,23 +926,18 @@ static void efx_tx_queue_insert(struct efx_tx_queue *tx_queue,
737{ 926{
738 struct efx_tx_buffer *buffer; 927 struct efx_tx_buffer *buffer;
739 struct efx_nic *efx = tx_queue->efx; 928 struct efx_nic *efx = tx_queue->efx;
740 unsigned dma_len, insert_ptr; 929 unsigned dma_len;
741 930
742 EFX_BUG_ON_PARANOID(len <= 0); 931 EFX_BUG_ON_PARANOID(len <= 0);
743 932
744 while (1) { 933 while (1) {
745 insert_ptr = tx_queue->insert_count & tx_queue->ptr_mask; 934 buffer = efx_tx_queue_get_insert_buffer(tx_queue);
746 buffer = &tx_queue->buffer[insert_ptr];
747 ++tx_queue->insert_count; 935 ++tx_queue->insert_count;
748 936
749 EFX_BUG_ON_PARANOID(tx_queue->insert_count - 937 EFX_BUG_ON_PARANOID(tx_queue->insert_count -
750 tx_queue->read_count >= 938 tx_queue->read_count >=
751 efx->txq_entries); 939 efx->txq_entries);
752 940
753 EFX_BUG_ON_PARANOID(buffer->len);
754 EFX_BUG_ON_PARANOID(buffer->unmap_len);
755 EFX_BUG_ON_PARANOID(buffer->flags);
756
757 buffer->dma_addr = dma_addr; 941 buffer->dma_addr = dma_addr;
758 942
759 dma_len = efx_max_tx_len(efx, dma_addr); 943 dma_len = efx_max_tx_len(efx, dma_addr);
@@ -796,6 +980,7 @@ static int efx_tso_put_header(struct efx_tx_queue *tx_queue,
796 return -ENOMEM; 980 return -ENOMEM;
797 } 981 }
798 buffer->unmap_len = buffer->len; 982 buffer->unmap_len = buffer->len;
983 buffer->dma_offset = 0;
799 buffer->flags |= EFX_TX_BUF_MAP_SINGLE; 984 buffer->flags |= EFX_TX_BUF_MAP_SINGLE;
800 } 985 }
801 986
@@ -814,19 +999,27 @@ static void efx_enqueue_unwind(struct efx_tx_queue *tx_queue)
814 /* Work backwards until we hit the original insert pointer value */ 999 /* Work backwards until we hit the original insert pointer value */
815 while (tx_queue->insert_count != tx_queue->write_count) { 1000 while (tx_queue->insert_count != tx_queue->write_count) {
816 --tx_queue->insert_count; 1001 --tx_queue->insert_count;
817 buffer = &tx_queue->buffer[tx_queue->insert_count & 1002 buffer = __efx_tx_queue_get_insert_buffer(tx_queue);
818 tx_queue->ptr_mask];
819 efx_dequeue_buffer(tx_queue, buffer, NULL, NULL); 1003 efx_dequeue_buffer(tx_queue, buffer, NULL, NULL);
820 } 1004 }
821} 1005}
822 1006
823 1007
824/* Parse the SKB header and initialise state. */ 1008/* Parse the SKB header and initialise state. */
825static void tso_start(struct tso_state *st, const struct sk_buff *skb) 1009static int tso_start(struct tso_state *st, struct efx_nic *efx,
1010 const struct sk_buff *skb)
826{ 1011{
1012 bool use_options = efx_nic_rev(efx) >= EFX_REV_HUNT_A0;
1013 struct device *dma_dev = &efx->pci_dev->dev;
1014 unsigned int header_len, in_len;
1015 dma_addr_t dma_addr;
1016
827 st->ip_off = skb_network_header(skb) - skb->data; 1017 st->ip_off = skb_network_header(skb) - skb->data;
828 st->tcp_off = skb_transport_header(skb) - skb->data; 1018 st->tcp_off = skb_transport_header(skb) - skb->data;
829 st->header_len = st->tcp_off + (tcp_hdr(skb)->doff << 2u); 1019 header_len = st->tcp_off + (tcp_hdr(skb)->doff << 2u);
1020 in_len = skb_headlen(skb) - header_len;
1021 st->header_len = header_len;
1022 st->in_len = in_len;
830 if (st->protocol == htons(ETH_P_IP)) { 1023 if (st->protocol == htons(ETH_P_IP)) {
831 st->ip_base_len = st->header_len - st->ip_off; 1024 st->ip_base_len = st->header_len - st->ip_off;
832 st->ipv4_id = ntohs(ip_hdr(skb)->id); 1025 st->ipv4_id = ntohs(ip_hdr(skb)->id);
@@ -840,9 +1033,34 @@ static void tso_start(struct tso_state *st, const struct sk_buff *skb)
840 EFX_BUG_ON_PARANOID(tcp_hdr(skb)->syn); 1033 EFX_BUG_ON_PARANOID(tcp_hdr(skb)->syn);
841 EFX_BUG_ON_PARANOID(tcp_hdr(skb)->rst); 1034 EFX_BUG_ON_PARANOID(tcp_hdr(skb)->rst);
842 1035
843 st->out_len = skb->len - st->header_len; 1036 st->out_len = skb->len - header_len;
844 st->unmap_len = 0; 1037
845 st->dma_flags = 0; 1038 if (!use_options) {
1039 st->header_unmap_len = 0;
1040
1041 if (likely(in_len == 0)) {
1042 st->dma_flags = 0;
1043 st->unmap_len = 0;
1044 return 0;
1045 }
1046
1047 dma_addr = dma_map_single(dma_dev, skb->data + header_len,
1048 in_len, DMA_TO_DEVICE);
1049 st->dma_flags = EFX_TX_BUF_MAP_SINGLE;
1050 st->dma_addr = dma_addr;
1051 st->unmap_addr = dma_addr;
1052 st->unmap_len = in_len;
1053 } else {
1054 dma_addr = dma_map_single(dma_dev, skb->data,
1055 skb_headlen(skb), DMA_TO_DEVICE);
1056 st->header_dma_addr = dma_addr;
1057 st->header_unmap_len = skb_headlen(skb);
1058 st->dma_flags = 0;
1059 st->dma_addr = dma_addr + header_len;
1060 st->unmap_len = 0;
1061 }
1062
1063 return unlikely(dma_mapping_error(dma_dev, dma_addr)) ? -ENOMEM : 0;
846} 1064}
847 1065
848static int tso_get_fragment(struct tso_state *st, struct efx_nic *efx, 1066static int tso_get_fragment(struct tso_state *st, struct efx_nic *efx,
@@ -860,24 +1078,6 @@ static int tso_get_fragment(struct tso_state *st, struct efx_nic *efx,
860 return -ENOMEM; 1078 return -ENOMEM;
861} 1079}
862 1080
863static int tso_get_head_fragment(struct tso_state *st, struct efx_nic *efx,
864 const struct sk_buff *skb)
865{
866 int hl = st->header_len;
867 int len = skb_headlen(skb) - hl;
868
869 st->unmap_addr = dma_map_single(&efx->pci_dev->dev, skb->data + hl,
870 len, DMA_TO_DEVICE);
871 if (likely(!dma_mapping_error(&efx->pci_dev->dev, st->unmap_addr))) {
872 st->dma_flags = EFX_TX_BUF_MAP_SINGLE;
873 st->unmap_len = len;
874 st->in_len = len;
875 st->dma_addr = st->unmap_addr;
876 return 0;
877 }
878 return -ENOMEM;
879}
880
881 1081
882/** 1082/**
883 * tso_fill_packet_with_fragment - form descriptors for the current fragment 1083 * tso_fill_packet_with_fragment - form descriptors for the current fragment
@@ -922,6 +1122,7 @@ static void tso_fill_packet_with_fragment(struct efx_tx_queue *tx_queue,
922 if (st->in_len == 0) { 1122 if (st->in_len == 0) {
923 /* Transfer ownership of the DMA mapping */ 1123 /* Transfer ownership of the DMA mapping */
924 buffer->unmap_len = st->unmap_len; 1124 buffer->unmap_len = st->unmap_len;
1125 buffer->dma_offset = buffer->unmap_len - buffer->len;
925 buffer->flags |= st->dma_flags; 1126 buffer->flags |= st->dma_flags;
926 st->unmap_len = 0; 1127 st->unmap_len = 0;
927 } 1128 }
@@ -944,55 +1145,98 @@ static int tso_start_new_packet(struct efx_tx_queue *tx_queue,
944 struct tso_state *st) 1145 struct tso_state *st)
945{ 1146{
946 struct efx_tx_buffer *buffer = 1147 struct efx_tx_buffer *buffer =
947 &tx_queue->buffer[tx_queue->insert_count & tx_queue->ptr_mask]; 1148 efx_tx_queue_get_insert_buffer(tx_queue);
948 struct tcphdr *tsoh_th; 1149 bool is_last = st->out_len <= skb_shinfo(skb)->gso_size;
949 unsigned ip_length; 1150 u8 tcp_flags_clear;
950 u8 *header;
951 int rc;
952 1151
953 /* Allocate and insert a DMA-mapped header buffer. */ 1152 if (!is_last) {
954 header = efx_tsoh_get_buffer(tx_queue, buffer, st->header_len);
955 if (!header)
956 return -ENOMEM;
957
958 tsoh_th = (struct tcphdr *)(header + st->tcp_off);
959
960 /* Copy and update the headers. */
961 memcpy(header, skb->data, st->header_len);
962
963 tsoh_th->seq = htonl(st->seqnum);
964 st->seqnum += skb_shinfo(skb)->gso_size;
965 if (st->out_len > skb_shinfo(skb)->gso_size) {
966 /* This packet will not finish the TSO burst. */
967 st->packet_space = skb_shinfo(skb)->gso_size; 1153 st->packet_space = skb_shinfo(skb)->gso_size;
968 tsoh_th->fin = 0; 1154 tcp_flags_clear = 0x09; /* mask out FIN and PSH */
969 tsoh_th->psh = 0;
970 } else { 1155 } else {
971 /* This packet will be the last in the TSO burst. */
972 st->packet_space = st->out_len; 1156 st->packet_space = st->out_len;
973 tsoh_th->fin = tcp_hdr(skb)->fin; 1157 tcp_flags_clear = 0x00;
974 tsoh_th->psh = tcp_hdr(skb)->psh;
975 } 1158 }
976 ip_length = st->ip_base_len + st->packet_space;
977 1159
978 if (st->protocol == htons(ETH_P_IP)) { 1160 if (!st->header_unmap_len) {
979 struct iphdr *tsoh_iph = (struct iphdr *)(header + st->ip_off); 1161 /* Allocate and insert a DMA-mapped header buffer. */
1162 struct tcphdr *tsoh_th;
1163 unsigned ip_length;
1164 u8 *header;
1165 int rc;
1166
1167 header = efx_tsoh_get_buffer(tx_queue, buffer, st->header_len);
1168 if (!header)
1169 return -ENOMEM;
980 1170
981 tsoh_iph->tot_len = htons(ip_length); 1171 tsoh_th = (struct tcphdr *)(header + st->tcp_off);
1172
1173 /* Copy and update the headers. */
1174 memcpy(header, skb->data, st->header_len);
1175
1176 tsoh_th->seq = htonl(st->seqnum);
1177 ((u8 *)tsoh_th)[13] &= ~tcp_flags_clear;
1178
1179 ip_length = st->ip_base_len + st->packet_space;
1180
1181 if (st->protocol == htons(ETH_P_IP)) {
1182 struct iphdr *tsoh_iph =
1183 (struct iphdr *)(header + st->ip_off);
1184
1185 tsoh_iph->tot_len = htons(ip_length);
1186 tsoh_iph->id = htons(st->ipv4_id);
1187 } else {
1188 struct ipv6hdr *tsoh_iph =
1189 (struct ipv6hdr *)(header + st->ip_off);
1190
1191 tsoh_iph->payload_len = htons(ip_length);
1192 }
982 1193
983 /* Linux leaves suitable gaps in the IP ID space for us to fill. */ 1194 rc = efx_tso_put_header(tx_queue, buffer, header);
984 tsoh_iph->id = htons(st->ipv4_id); 1195 if (unlikely(rc))
985 st->ipv4_id++; 1196 return rc;
986 } else { 1197 } else {
987 struct ipv6hdr *tsoh_iph = 1198 /* Send the original headers with a TSO option descriptor
988 (struct ipv6hdr *)(header + st->ip_off); 1199 * in front
1200 */
1201 u8 tcp_flags = ((u8 *)tcp_hdr(skb))[13] & ~tcp_flags_clear;
989 1202
990 tsoh_iph->payload_len = htons(ip_length); 1203 buffer->flags = EFX_TX_BUF_OPTION;
1204 buffer->len = 0;
1205 buffer->unmap_len = 0;
1206 EFX_POPULATE_QWORD_5(buffer->option,
1207 ESF_DZ_TX_DESC_IS_OPT, 1,
1208 ESF_DZ_TX_OPTION_TYPE,
1209 ESE_DZ_TX_OPTION_DESC_TSO,
1210 ESF_DZ_TX_TSO_TCP_FLAGS, tcp_flags,
1211 ESF_DZ_TX_TSO_IP_ID, st->ipv4_id,
1212 ESF_DZ_TX_TSO_TCP_SEQNO, st->seqnum);
1213 ++tx_queue->insert_count;
1214
1215 /* We mapped the headers in tso_start(). Unmap them
1216 * when the last segment is completed.
1217 */
1218 buffer = efx_tx_queue_get_insert_buffer(tx_queue);
1219 buffer->dma_addr = st->header_dma_addr;
1220 buffer->len = st->header_len;
1221 if (is_last) {
1222 buffer->flags = EFX_TX_BUF_CONT | EFX_TX_BUF_MAP_SINGLE;
1223 buffer->unmap_len = st->header_unmap_len;
1224 buffer->dma_offset = 0;
1225 /* Ensure we only unmap them once in case of a
1226 * later DMA mapping error and rollback
1227 */
1228 st->header_unmap_len = 0;
1229 } else {
1230 buffer->flags = EFX_TX_BUF_CONT;
1231 buffer->unmap_len = 0;
1232 }
1233 ++tx_queue->insert_count;
991 } 1234 }
992 1235
993 rc = efx_tso_put_header(tx_queue, buffer, header); 1236 st->seqnum += skb_shinfo(skb)->gso_size;
994 if (unlikely(rc)) 1237
995 return rc; 1238 /* Linux leaves suitable gaps in the IP ID space for us to fill. */
1239 ++st->ipv4_id;
996 1240
997 ++tx_queue->tso_packets; 1241 ++tx_queue->tso_packets;
998 1242
@@ -1023,12 +1267,11 @@ static int efx_enqueue_skb_tso(struct efx_tx_queue *tx_queue,
1023 1267
1024 EFX_BUG_ON_PARANOID(tx_queue->write_count != tx_queue->insert_count); 1268 EFX_BUG_ON_PARANOID(tx_queue->write_count != tx_queue->insert_count);
1025 1269
1026 tso_start(&state, skb); 1270 rc = tso_start(&state, efx, skb);
1271 if (rc)
1272 goto mem_err;
1027 1273
1028 /* Assume that skb header area contains exactly the headers, and 1274 if (likely(state.in_len == 0)) {
1029 * all payload is in the frag list.
1030 */
1031 if (skb_headlen(skb) == state.header_len) {
1032 /* Grab the first payload fragment. */ 1275 /* Grab the first payload fragment. */
1033 EFX_BUG_ON_PARANOID(skb_shinfo(skb)->nr_frags < 1); 1276 EFX_BUG_ON_PARANOID(skb_shinfo(skb)->nr_frags < 1);
1034 frag_i = 0; 1277 frag_i = 0;
@@ -1037,9 +1280,7 @@ static int efx_enqueue_skb_tso(struct efx_tx_queue *tx_queue,
1037 if (rc) 1280 if (rc)
1038 goto mem_err; 1281 goto mem_err;
1039 } else { 1282 } else {
1040 rc = tso_get_head_fragment(&state, efx, skb); 1283 /* Payload starts in the header area. */
1041 if (rc)
1042 goto mem_err;
1043 frag_i = -1; 1284 frag_i = -1;
1044 } 1285 }
1045 1286
@@ -1091,6 +1332,11 @@ static int efx_enqueue_skb_tso(struct efx_tx_queue *tx_queue,
1091 state.unmap_len, DMA_TO_DEVICE); 1332 state.unmap_len, DMA_TO_DEVICE);
1092 } 1333 }
1093 1334
1335 /* Free the header DMA mapping, if using option descriptors */
1336 if (state.header_unmap_len)
1337 dma_unmap_single(&efx->pci_dev->dev, state.header_dma_addr,
1338 state.header_unmap_len, DMA_TO_DEVICE);
1339
1094 efx_enqueue_unwind(tx_queue); 1340 efx_enqueue_unwind(tx_queue);
1095 return NETDEV_TX_OK; 1341 return NETDEV_TX_OK;
1096} 1342}
diff --git a/drivers/net/ethernet/sgi/meth.c b/drivers/net/ethernet/sgi/meth.c
index 770036bc2d87..513ed8b1ba58 100644
--- a/drivers/net/ethernet/sgi/meth.c
+++ b/drivers/net/ethernet/sgi/meth.c
@@ -839,7 +839,7 @@ static int meth_probe(struct platform_device *pdev)
839 dev->watchdog_timeo = timeout; 839 dev->watchdog_timeo = timeout;
840 dev->irq = MACE_ETHERNET_IRQ; 840 dev->irq = MACE_ETHERNET_IRQ;
841 dev->base_addr = (unsigned long)&mace->eth; 841 dev->base_addr = (unsigned long)&mace->eth;
842 memcpy(dev->dev_addr, o2meth_eaddr, 6); 842 memcpy(dev->dev_addr, o2meth_eaddr, ETH_ALEN);
843 843
844 priv = netdev_priv(dev); 844 priv = netdev_priv(dev);
845 spin_lock_init(&priv->meth_lock); 845 spin_lock_init(&priv->meth_lock);
diff --git a/drivers/net/ethernet/sis/sis190.c b/drivers/net/ethernet/sis/sis190.c
index ee18e6f7b4fe..acbbe48a519c 100644
--- a/drivers/net/ethernet/sis/sis190.c
+++ b/drivers/net/ethernet/sis/sis190.c
@@ -1921,7 +1921,6 @@ static void sis190_remove_one(struct pci_dev *pdev)
1921 cancel_work_sync(&tp->phy_task); 1921 cancel_work_sync(&tp->phy_task);
1922 unregister_netdev(dev); 1922 unregister_netdev(dev);
1923 sis190_release_board(pdev); 1923 sis190_release_board(pdev);
1924 pci_set_drvdata(pdev, NULL);
1925} 1924}
1926 1925
1927static struct pci_driver sis190_pci_driver = { 1926static struct pci_driver sis190_pci_driver = {
diff --git a/drivers/net/ethernet/smsc/Kconfig b/drivers/net/ethernet/smsc/Kconfig
index 068fc44d37e1..753630f5d3d3 100644
--- a/drivers/net/ethernet/smsc/Kconfig
+++ b/drivers/net/ethernet/smsc/Kconfig
@@ -6,7 +6,7 @@ config NET_VENDOR_SMSC
6 bool "SMC (SMSC)/Western Digital devices" 6 bool "SMC (SMSC)/Western Digital devices"
7 default y 7 default y
8 depends on ARM || ISA || MAC || ARM64 || MIPS || M32R || SUPERH || \ 8 depends on ARM || ISA || MAC || ARM64 || MIPS || M32R || SUPERH || \
9 BLACKFIN || MN10300 || COLDFIRE || PCI || PCMCIA 9 BLACKFIN || MN10300 || COLDFIRE || XTENSA || PCI || PCMCIA
10 ---help--- 10 ---help---
11 If you have a network (Ethernet) card belonging to this class, say Y 11 If you have a network (Ethernet) card belonging to this class, say Y
12 and read the Ethernet-HOWTO, available from 12 and read the Ethernet-HOWTO, available from
@@ -39,7 +39,7 @@ config SMC91X
39 select CRC32 39 select CRC32
40 select MII 40 select MII
41 depends on (ARM || M32R || SUPERH || MIPS || BLACKFIN || \ 41 depends on (ARM || M32R || SUPERH || MIPS || BLACKFIN || \
42 MN10300 || COLDFIRE || ARM64) 42 MN10300 || COLDFIRE || ARM64 || XTENSA)
43 ---help--- 43 ---help---
44 This is a driver for SMC's 91x series of Ethernet chipsets, 44 This is a driver for SMC's 91x series of Ethernet chipsets,
45 including the SMC91C94 and the SMC91C111. Say Y if you want it 45 including the SMC91C94 and the SMC91C111. Say Y if you want it
diff --git a/drivers/net/ethernet/smsc/epic100.c b/drivers/net/ethernet/smsc/epic100.c
index 03b256af7ed5..8ae1f8a7bf38 100644
--- a/drivers/net/ethernet/smsc/epic100.c
+++ b/drivers/net/ethernet/smsc/epic100.c
@@ -91,9 +91,9 @@ static int rx_copybreak;
91 91
92/* These identify the driver base version and may not be removed. */ 92/* These identify the driver base version and may not be removed. */
93static char version[] = 93static char version[] =
94DRV_NAME ".c:v1.11 1/7/2001 Written by Donald Becker <becker@scyld.com>\n"; 94DRV_NAME ".c:v1.11 1/7/2001 Written by Donald Becker <becker@scyld.com>";
95static char version2[] = 95static char version2[] =
96" (unofficial 2.4.x kernel port, version " DRV_VERSION ", " DRV_RELDATE ")\n"; 96" (unofficial 2.4.x kernel port, version " DRV_VERSION ", " DRV_RELDATE ")";
97 97
98MODULE_AUTHOR("Donald Becker <becker@scyld.com>"); 98MODULE_AUTHOR("Donald Becker <becker@scyld.com>");
99MODULE_DESCRIPTION("SMC 83c170 EPIC series Ethernet driver"); 99MODULE_DESCRIPTION("SMC 83c170 EPIC series Ethernet driver");
@@ -332,9 +332,7 @@ static int epic_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
332 332
333/* when built into the kernel, we only print version if device is found */ 333/* when built into the kernel, we only print version if device is found */
334#ifndef MODULE 334#ifndef MODULE
335 static int printed_version; 335 pr_info_once("%s%s\n", version, version2);
336 if (!printed_version++)
337 printk(KERN_INFO "%s%s", version, version2);
338#endif 336#endif
339 337
340 card_idx++; 338 card_idx++;
@@ -423,9 +421,9 @@ static int epic_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
423 ((__le16 *)dev->dev_addr)[i] = cpu_to_le16(er16(LAN0 + i*4)); 421 ((__le16 *)dev->dev_addr)[i] = cpu_to_le16(er16(LAN0 + i*4));
424 422
425 if (debug > 2) { 423 if (debug > 2) {
426 dev_printk(KERN_DEBUG, &pdev->dev, "EEPROM contents:\n"); 424 dev_dbg(&pdev->dev, "EEPROM contents:\n");
427 for (i = 0; i < 64; i++) 425 for (i = 0; i < 64; i++)
428 printk(" %4.4x%s", read_eeprom(ep, i), 426 pr_cont(" %4.4x%s", read_eeprom(ep, i),
429 i % 16 == 15 ? "\n" : ""); 427 i % 16 == 15 ? "\n" : "");
430 } 428 }
431 429
@@ -490,10 +488,10 @@ static int epic_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
490 if (ret < 0) 488 if (ret < 0)
491 goto err_out_unmap_rx; 489 goto err_out_unmap_rx;
492 490
493 printk(KERN_INFO "%s: %s at %lx, IRQ %d, %pM\n", 491 netdev_info(dev, "%s at %lx, IRQ %d, %pM\n",
494 dev->name, pci_id_tbl[chip_idx].name, 492 pci_id_tbl[chip_idx].name,
495 (long)pci_resource_start(pdev, EPIC_BAR), pdev->irq, 493 (long)pci_resource_start(pdev, EPIC_BAR), pdev->irq,
496 dev->dev_addr); 494 dev->dev_addr);
497 495
498out: 496out:
499 return ret; 497 return ret;
@@ -703,9 +701,8 @@ static int epic_open(struct net_device *dev)
703 mdio_write(dev, ep->phys[0], MII_BMCR, media2miictl[dev->if_port&15]); 701 mdio_write(dev, ep->phys[0], MII_BMCR, media2miictl[dev->if_port&15]);
704 if (dev->if_port == 1) { 702 if (dev->if_port == 1) {
705 if (debug > 1) 703 if (debug > 1)
706 printk(KERN_INFO "%s: Using the 10base2 transceiver, MII " 704 netdev_info(dev, "Using the 10base2 transceiver, MII status %4.4x.\n",
707 "status %4.4x.\n", 705 mdio_read(dev, ep->phys[0], MII_BMSR));
708 dev->name, mdio_read(dev, ep->phys[0], MII_BMSR));
709 } 706 }
710 } else { 707 } else {
711 int mii_lpa = mdio_read(dev, ep->phys[0], MII_LPA); 708 int mii_lpa = mdio_read(dev, ep->phys[0], MII_LPA);
@@ -715,10 +712,10 @@ static int epic_open(struct net_device *dev)
715 else if (! (mii_lpa & LPA_LPACK)) 712 else if (! (mii_lpa & LPA_LPACK))
716 mdio_write(dev, ep->phys[0], MII_BMCR, BMCR_ANENABLE|BMCR_ANRESTART); 713 mdio_write(dev, ep->phys[0], MII_BMCR, BMCR_ANENABLE|BMCR_ANRESTART);
717 if (debug > 1) 714 if (debug > 1)
718 printk(KERN_INFO "%s: Setting %s-duplex based on MII xcvr %d" 715 netdev_info(dev, "Setting %s-duplex based on MII xcvr %d register read of %4.4x.\n",
719 " register read of %4.4x.\n", dev->name, 716 ep->mii.full_duplex ? "full"
720 ep->mii.full_duplex ? "full" : "half", 717 : "half",
721 ep->phys[0], mii_lpa); 718 ep->phys[0], mii_lpa);
722 } 719 }
723 } 720 }
724 721
@@ -738,10 +735,9 @@ static int epic_open(struct net_device *dev)
738 TxUnderrun); 735 TxUnderrun);
739 736
740 if (debug > 1) { 737 if (debug > 1) {
741 printk(KERN_DEBUG "%s: epic_open() ioaddr %p IRQ %d " 738 netdev_dbg(dev, "epic_open() ioaddr %p IRQ %d status %4.4x %s-duplex.\n",
742 "status %4.4x %s-duplex.\n", 739 ioaddr, irq, er32(GENCTL),
743 dev->name, ioaddr, irq, er32(GENCTL), 740 ep->mii.full_duplex ? "full" : "half");
744 ep->mii.full_duplex ? "full" : "half");
745 } 741 }
746 742
747 /* Set the timer to switch to check for link beat and perhaps switch 743 /* Set the timer to switch to check for link beat and perhaps switch
@@ -790,8 +786,8 @@ static void epic_restart(struct net_device *dev)
790 /* Soft reset the chip. */ 786 /* Soft reset the chip. */
791 ew32(GENCTL, 0x4001); 787 ew32(GENCTL, 0x4001);
792 788
793 printk(KERN_DEBUG "%s: Restarting the EPIC chip, Rx %d/%d Tx %d/%d.\n", 789 netdev_dbg(dev, "Restarting the EPIC chip, Rx %d/%d Tx %d/%d.\n",
794 dev->name, ep->cur_rx, ep->dirty_rx, ep->dirty_tx, ep->cur_tx); 790 ep->cur_rx, ep->dirty_rx, ep->dirty_tx, ep->cur_tx);
795 udelay(1); 791 udelay(1);
796 792
797 /* This magic is documented in SMSC app note 7.15 */ 793 /* This magic is documented in SMSC app note 7.15 */
@@ -827,9 +823,8 @@ static void epic_restart(struct net_device *dev)
827 ((ep->chip_flags & TYPE2_INTR) ? PCIBusErr175 : PCIBusErr170) | 823 ((ep->chip_flags & TYPE2_INTR) ? PCIBusErr175 : PCIBusErr170) |
828 TxUnderrun); 824 TxUnderrun);
829 825
830 printk(KERN_DEBUG "%s: epic_restart() done, cmd status %4.4x, ctl %4.4x" 826 netdev_dbg(dev, "epic_restart() done, cmd status %4.4x, ctl %4.4x interrupt %4.4x.\n",
831 " interrupt %4.4x.\n", 827 er32(COMMAND), er32(GENCTL), er32(INTSTAT));
832 dev->name, er32(COMMAND), er32(GENCTL), er32(INTSTAT));
833} 828}
834 829
835static void check_media(struct net_device *dev) 830static void check_media(struct net_device *dev)
@@ -846,9 +841,9 @@ static void check_media(struct net_device *dev)
846 return; 841 return;
847 if (ep->mii.full_duplex != duplex) { 842 if (ep->mii.full_duplex != duplex) {
848 ep->mii.full_duplex = duplex; 843 ep->mii.full_duplex = duplex;
849 printk(KERN_INFO "%s: Setting %s-duplex based on MII #%d link" 844 netdev_info(dev, "Setting %s-duplex based on MII #%d link partner capability of %4.4x.\n",
850 " partner capability of %4.4x.\n", dev->name, 845 ep->mii.full_duplex ? "full" : "half",
851 ep->mii.full_duplex ? "full" : "half", ep->phys[0], mii_lpa); 846 ep->phys[0], mii_lpa);
852 ew32(TxCtrl, ep->mii.full_duplex ? 0x7F : 0x79); 847 ew32(TxCtrl, ep->mii.full_duplex ? 0x7F : 0x79);
853 } 848 }
854} 849}
@@ -861,11 +856,10 @@ static void epic_timer(unsigned long data)
861 int next_tick = 5*HZ; 856 int next_tick = 5*HZ;
862 857
863 if (debug > 3) { 858 if (debug > 3) {
864 printk(KERN_DEBUG "%s: Media monitor tick, Tx status %8.8x.\n", 859 netdev_dbg(dev, "Media monitor tick, Tx status %8.8x.\n",
865 dev->name, er32(TxSTAT)); 860 er32(TxSTAT));
866 printk(KERN_DEBUG "%s: Other registers are IntMask %4.4x " 861 netdev_dbg(dev, "Other registers are IntMask %4.4x IntStatus %4.4x RxStatus %4.4x.\n",
867 "IntStatus %4.4x RxStatus %4.4x.\n", dev->name, 862 er32(INTMASK), er32(INTSTAT), er32(RxSTAT));
868 er32(INTMASK), er32(INTSTAT), er32(RxSTAT));
869 } 863 }
870 864
871 check_media(dev); 865 check_media(dev);
@@ -880,11 +874,11 @@ static void epic_tx_timeout(struct net_device *dev)
880 void __iomem *ioaddr = ep->ioaddr; 874 void __iomem *ioaddr = ep->ioaddr;
881 875
882 if (debug > 0) { 876 if (debug > 0) {
883 printk(KERN_WARNING "%s: Transmit timeout using MII device, " 877 netdev_warn(dev, "Transmit timeout using MII device, Tx status %4.4x.\n",
884 "Tx status %4.4x.\n", dev->name, er16(TxSTAT)); 878 er16(TxSTAT));
885 if (debug > 1) { 879 if (debug > 1) {
886 printk(KERN_DEBUG "%s: Tx indices: dirty_tx %d, cur_tx %d.\n", 880 netdev_dbg(dev, "Tx indices: dirty_tx %d, cur_tx %d.\n",
887 dev->name, ep->dirty_tx, ep->cur_tx); 881 ep->dirty_tx, ep->cur_tx);
888 } 882 }
889 } 883 }
890 if (er16(TxSTAT) & 0x10) { /* Tx FIFO underflow. */ 884 if (er16(TxSTAT) & 0x10) { /* Tx FIFO underflow. */
@@ -994,9 +988,8 @@ static netdev_tx_t epic_start_xmit(struct sk_buff *skb, struct net_device *dev)
994 ew32(COMMAND, TxQueued); 988 ew32(COMMAND, TxQueued);
995 989
996 if (debug > 4) 990 if (debug > 4)
997 printk(KERN_DEBUG "%s: Queued Tx packet size %d to slot %d, " 991 netdev_dbg(dev, "Queued Tx packet size %d to slot %d, flag %2.2x Tx status %8.8x.\n",
998 "flag %2.2x Tx status %8.8x.\n", dev->name, skb->len, 992 skb->len, entry, ctrl_word, er32(TxSTAT));
999 entry, ctrl_word, er32(TxSTAT));
1000 993
1001 return NETDEV_TX_OK; 994 return NETDEV_TX_OK;
1002} 995}
@@ -1009,8 +1002,8 @@ static void epic_tx_error(struct net_device *dev, struct epic_private *ep,
1009#ifndef final_version 1002#ifndef final_version
1010 /* There was an major error, log it. */ 1003 /* There was an major error, log it. */
1011 if (debug > 1) 1004 if (debug > 1)
1012 printk(KERN_DEBUG "%s: Transmit error, Tx status %8.8x.\n", 1005 netdev_dbg(dev, "Transmit error, Tx status %8.8x.\n",
1013 dev->name, status); 1006 status);
1014#endif 1007#endif
1015 stats->tx_errors++; 1008 stats->tx_errors++;
1016 if (status & 0x1050) 1009 if (status & 0x1050)
@@ -1057,9 +1050,8 @@ static void epic_tx(struct net_device *dev, struct epic_private *ep)
1057 1050
1058#ifndef final_version 1051#ifndef final_version
1059 if (cur_tx - dirty_tx > TX_RING_SIZE) { 1052 if (cur_tx - dirty_tx > TX_RING_SIZE) {
1060 printk(KERN_WARNING 1053 netdev_warn(dev, "Out-of-sync dirty pointer, %d vs. %d, full=%d.\n",
1061 "%s: Out-of-sync dirty pointer, %d vs. %d, full=%d.\n", 1054 dirty_tx, cur_tx, ep->tx_full);
1062 dev->name, dirty_tx, cur_tx, ep->tx_full);
1063 dirty_tx += TX_RING_SIZE; 1055 dirty_tx += TX_RING_SIZE;
1064 } 1056 }
1065#endif 1057#endif
@@ -1086,8 +1078,8 @@ static irqreturn_t epic_interrupt(int irq, void *dev_instance)
1086 ew32(INTSTAT, status & EpicNormalEvent); 1078 ew32(INTSTAT, status & EpicNormalEvent);
1087 1079
1088 if (debug > 4) { 1080 if (debug > 4) {
1089 printk(KERN_DEBUG "%s: Interrupt, status=%#8.8x new " 1081 netdev_dbg(dev, "Interrupt, status=%#8.8x new intstat=%#8.8x.\n",
1090 "intstat=%#8.8x.\n", dev->name, status, er32(INTSTAT)); 1082 status, er32(INTSTAT));
1091 } 1083 }
1092 1084
1093 if ((status & IntrSummary) == 0) 1085 if ((status & IntrSummary) == 0)
@@ -1125,8 +1117,8 @@ static irqreturn_t epic_interrupt(int irq, void *dev_instance)
1125 ew32(COMMAND, RestartTx); 1117 ew32(COMMAND, RestartTx);
1126 } 1118 }
1127 if (status & PCIBusErr170) { 1119 if (status & PCIBusErr170) {
1128 printk(KERN_ERR "%s: PCI Bus Error! status %4.4x.\n", 1120 netdev_err(dev, "PCI Bus Error! status %4.4x.\n",
1129 dev->name, status); 1121 status);
1130 epic_pause(dev); 1122 epic_pause(dev);
1131 epic_restart(dev); 1123 epic_restart(dev);
1132 } 1124 }
@@ -1136,8 +1128,8 @@ static irqreturn_t epic_interrupt(int irq, void *dev_instance)
1136 1128
1137out: 1129out:
1138 if (debug > 3) { 1130 if (debug > 3) {
1139 printk(KERN_DEBUG "%s: exit interrupt, intr_status=%#4.4x.\n", 1131 netdev_dbg(dev, "exit interrupt, intr_status=%#4.4x.\n",
1140 dev->name, status); 1132 status);
1141 } 1133 }
1142 1134
1143 return IRQ_RETVAL(handled); 1135 return IRQ_RETVAL(handled);
@@ -1151,7 +1143,7 @@ static int epic_rx(struct net_device *dev, int budget)
1151 int work_done = 0; 1143 int work_done = 0;
1152 1144
1153 if (debug > 4) 1145 if (debug > 4)
1154 printk(KERN_DEBUG " In epic_rx(), entry %d %8.8x.\n", entry, 1146 netdev_dbg(dev, " In epic_rx(), entry %d %8.8x.\n", entry,
1155 ep->rx_ring[entry].rxstatus); 1147 ep->rx_ring[entry].rxstatus);
1156 1148
1157 if (rx_work_limit > budget) 1149 if (rx_work_limit > budget)
@@ -1162,16 +1154,17 @@ static int epic_rx(struct net_device *dev, int budget)
1162 int status = ep->rx_ring[entry].rxstatus; 1154 int status = ep->rx_ring[entry].rxstatus;
1163 1155
1164 if (debug > 4) 1156 if (debug > 4)
1165 printk(KERN_DEBUG " epic_rx() status was %8.8x.\n", status); 1157 netdev_dbg(dev, " epic_rx() status was %8.8x.\n",
1158 status);
1166 if (--rx_work_limit < 0) 1159 if (--rx_work_limit < 0)
1167 break; 1160 break;
1168 if (status & 0x2006) { 1161 if (status & 0x2006) {
1169 if (debug > 2) 1162 if (debug > 2)
1170 printk(KERN_DEBUG "%s: epic_rx() error status was %8.8x.\n", 1163 netdev_dbg(dev, "epic_rx() error status was %8.8x.\n",
1171 dev->name, status); 1164 status);
1172 if (status & 0x2000) { 1165 if (status & 0x2000) {
1173 printk(KERN_WARNING "%s: Oversized Ethernet frame spanned " 1166 netdev_warn(dev, "Oversized Ethernet frame spanned multiple buffers, status %4.4x!\n",
1174 "multiple buffers, status %4.4x!\n", dev->name, status); 1167 status);
1175 dev->stats.rx_length_errors++; 1168 dev->stats.rx_length_errors++;
1176 } else if (status & 0x0006) 1169 } else if (status & 0x0006)
1177 /* Rx Frame errors are counted in hardware. */ 1170 /* Rx Frame errors are counted in hardware. */
@@ -1183,9 +1176,8 @@ static int epic_rx(struct net_device *dev, int budget)
1183 struct sk_buff *skb; 1176 struct sk_buff *skb;
1184 1177
1185 if (pkt_len > PKT_BUF_SZ - 4) { 1178 if (pkt_len > PKT_BUF_SZ - 4) {
1186 printk(KERN_ERR "%s: Oversized Ethernet frame, status %x " 1179 netdev_err(dev, "Oversized Ethernet frame, status %x %d bytes.\n",
1187 "%d bytes.\n", 1180 status, pkt_len);
1188 dev->name, status, pkt_len);
1189 pkt_len = 1514; 1181 pkt_len = 1514;
1190 } 1182 }
1191 /* Check if the packet is long enough to accept without copying 1183 /* Check if the packet is long enough to accept without copying
@@ -1305,8 +1297,8 @@ static int epic_close(struct net_device *dev)
1305 napi_disable(&ep->napi); 1297 napi_disable(&ep->napi);
1306 1298
1307 if (debug > 1) 1299 if (debug > 1)
1308 printk(KERN_DEBUG "%s: Shutting down ethercard, status was %2.2x.\n", 1300 netdev_dbg(dev, "Shutting down ethercard, status was %2.2x.\n",
1309 dev->name, er32(INTSTAT)); 1301 er32(INTSTAT));
1310 1302
1311 del_timer_sync(&ep->timer); 1303 del_timer_sync(&ep->timer);
1312 1304
@@ -1324,7 +1316,7 @@ static int epic_close(struct net_device *dev)
1324 ep->rx_ring[i].buflength = 0; 1316 ep->rx_ring[i].buflength = 0;
1325 if (skb) { 1317 if (skb) {
1326 pci_unmap_single(pdev, ep->rx_ring[i].bufaddr, 1318 pci_unmap_single(pdev, ep->rx_ring[i].bufaddr,
1327 ep->rx_buf_sz, PCI_DMA_FROMDEVICE); 1319 ep->rx_buf_sz, PCI_DMA_FROMDEVICE);
1328 dev_kfree_skb(skb); 1320 dev_kfree_skb(skb);
1329 } 1321 }
1330 ep->rx_ring[i].bufaddr = 0xBADF00D0; /* An invalid address. */ 1322 ep->rx_ring[i].bufaddr = 0xBADF00D0; /* An invalid address. */
@@ -1535,7 +1527,6 @@ static void epic_remove_one(struct pci_dev *pdev)
1535 pci_release_regions(pdev); 1527 pci_release_regions(pdev);
1536 free_netdev(dev); 1528 free_netdev(dev);
1537 pci_disable_device(pdev); 1529 pci_disable_device(pdev);
1538 pci_set_drvdata(pdev, NULL);
1539 /* pci_power_off(pdev, -1); */ 1530 /* pci_power_off(pdev, -1); */
1540} 1531}
1541 1532
@@ -1588,8 +1579,7 @@ static int __init epic_init (void)
1588{ 1579{
1589/* when a module, this is printed whether or not devices are found in probe */ 1580/* when a module, this is printed whether or not devices are found in probe */
1590#ifdef MODULE 1581#ifdef MODULE
1591 printk (KERN_INFO "%s%s", 1582 pr_info("%s%s\n", version, version2);
1592 version, version2);
1593#endif 1583#endif
1594 1584
1595 return pci_register_driver(&epic_driver); 1585 return pci_register_driver(&epic_driver);
diff --git a/drivers/net/ethernet/smsc/smc911x.c b/drivers/net/ethernet/smsc/smc911x.c
index afe01c4088a3..0f096a890059 100644
--- a/drivers/net/ethernet/smsc/smc911x.c
+++ b/drivers/net/ethernet/smsc/smc911x.c
@@ -106,16 +106,16 @@ MODULE_ALIAS("platform:smc911x");
106#define POWER_DOWN 1 106#define POWER_DOWN 1
107 107
108#if SMC_DEBUG > 0 108#if SMC_DEBUG > 0
109#define DBG(n, args...) \ 109#define DBG(n, dev, args...) \
110 do { \ 110 do { \
111 if (SMC_DEBUG & (n)) \ 111 if (SMC_DEBUG & (n)) \
112 printk(args); \ 112 netdev_dbg(dev, args); \
113 } while (0) 113 } while (0)
114 114
115#define PRINTK(args...) printk(args) 115#define PRINTK(dev, args...) netdev_info(dev, args)
116#else 116#else
117#define DBG(n, args...) do { } while (0) 117#define DBG(n, dev, args...) do { } while (0)
118#define PRINTK(args...) printk(KERN_DEBUG args) 118#define PRINTK(dev, args...) netdev_dbg(dev, args)
119#endif 119#endif
120 120
121#if SMC_DEBUG_PKTS > 0 121#if SMC_DEBUG_PKTS > 0
@@ -130,21 +130,23 @@ static void PRINT_PKT(u_char *buf, int length)
130 130
131 for (i = 0; i < lines ; i ++) { 131 for (i = 0; i < lines ; i ++) {
132 int cur; 132 int cur;
133 printk(KERN_DEBUG);
133 for (cur = 0; cur < 8; cur++) { 134 for (cur = 0; cur < 8; cur++) {
134 u_char a, b; 135 u_char a, b;
135 a = *buf++; 136 a = *buf++;
136 b = *buf++; 137 b = *buf++;
137 printk("%02x%02x ", a, b); 138 pr_cont("%02x%02x ", a, b);
138 } 139 }
139 printk("\n"); 140 pr_cont("\n");
140 } 141 }
142 printk(KERN_DEBUG);
141 for (i = 0; i < remainder/2 ; i++) { 143 for (i = 0; i < remainder/2 ; i++) {
142 u_char a, b; 144 u_char a, b;
143 a = *buf++; 145 a = *buf++;
144 b = *buf++; 146 b = *buf++;
145 printk("%02x%02x ", a, b); 147 pr_cont("%02x%02x ", a, b);
146 } 148 }
147 printk("\n"); 149 pr_cont("\n");
148} 150}
149#else 151#else
150#define PRINT_PKT(x...) do { } while (0) 152#define PRINT_PKT(x...) do { } while (0)
@@ -176,7 +178,7 @@ static void smc911x_reset(struct net_device *dev)
176 unsigned int reg, timeout=0, resets=1, irq_cfg; 178 unsigned int reg, timeout=0, resets=1, irq_cfg;
177 unsigned long flags; 179 unsigned long flags;
178 180
179 DBG(SMC_DEBUG_FUNC, "%s: --> %s\n", dev->name, __func__); 181 DBG(SMC_DEBUG_FUNC, dev, "--> %s\n", __func__);
180 182
181 /* Take out of PM setting first */ 183 /* Take out of PM setting first */
182 if ((SMC_GET_PMT_CTRL(lp) & PMT_CTRL_READY_) == 0) { 184 if ((SMC_GET_PMT_CTRL(lp) & PMT_CTRL_READY_) == 0) {
@@ -188,7 +190,7 @@ static void smc911x_reset(struct net_device *dev)
188 reg = SMC_GET_PMT_CTRL(lp) & PMT_CTRL_READY_; 190 reg = SMC_GET_PMT_CTRL(lp) & PMT_CTRL_READY_;
189 } while (--timeout && !reg); 191 } while (--timeout && !reg);
190 if (timeout == 0) { 192 if (timeout == 0) {
191 PRINTK("%s: smc911x_reset timeout waiting for PM restore\n", dev->name); 193 PRINTK(dev, "smc911x_reset timeout waiting for PM restore\n");
192 return; 194 return;
193 } 195 }
194 } 196 }
@@ -206,14 +208,14 @@ static void smc911x_reset(struct net_device *dev)
206 reg = SMC_GET_HW_CFG(lp); 208 reg = SMC_GET_HW_CFG(lp);
207 /* If chip indicates reset timeout then try again */ 209 /* If chip indicates reset timeout then try again */
208 if (reg & HW_CFG_SRST_TO_) { 210 if (reg & HW_CFG_SRST_TO_) {
209 PRINTK("%s: chip reset timeout, retrying...\n", dev->name); 211 PRINTK(dev, "chip reset timeout, retrying...\n");
210 resets++; 212 resets++;
211 break; 213 break;
212 } 214 }
213 } while (--timeout && (reg & HW_CFG_SRST_)); 215 } while (--timeout && (reg & HW_CFG_SRST_));
214 } 216 }
215 if (timeout == 0) { 217 if (timeout == 0) {
216 PRINTK("%s: smc911x_reset timeout waiting for reset\n", dev->name); 218 PRINTK(dev, "smc911x_reset timeout waiting for reset\n");
217 return; 219 return;
218 } 220 }
219 221
@@ -223,7 +225,7 @@ static void smc911x_reset(struct net_device *dev)
223 udelay(10); 225 udelay(10);
224 226
225 if (timeout == 0){ 227 if (timeout == 0){
226 PRINTK("%s: smc911x_reset timeout waiting for EEPROM busy\n", dev->name); 228 PRINTK(dev, "smc911x_reset timeout waiting for EEPROM busy\n");
227 return; 229 return;
228 } 230 }
229 231
@@ -270,7 +272,7 @@ static void smc911x_enable(struct net_device *dev)
270 unsigned mask, cfg, cr; 272 unsigned mask, cfg, cr;
271 unsigned long flags; 273 unsigned long flags;
272 274
273 DBG(SMC_DEBUG_FUNC, "%s: --> %s\n", dev->name, __func__); 275 DBG(SMC_DEBUG_FUNC, dev, "--> %s\n", __func__);
274 276
275 spin_lock_irqsave(&lp->lock, flags); 277 spin_lock_irqsave(&lp->lock, flags);
276 278
@@ -296,7 +298,7 @@ static void smc911x_enable(struct net_device *dev)
296 298
297 /* Turn on receiver and enable RX */ 299 /* Turn on receiver and enable RX */
298 if (cr & MAC_CR_RXEN_) 300 if (cr & MAC_CR_RXEN_)
299 DBG(SMC_DEBUG_RX, "%s: Receiver already enabled\n", dev->name); 301 DBG(SMC_DEBUG_RX, dev, "Receiver already enabled\n");
300 302
301 SMC_SET_MAC_CR(lp, cr | MAC_CR_RXEN_); 303 SMC_SET_MAC_CR(lp, cr | MAC_CR_RXEN_);
302 304
@@ -327,7 +329,7 @@ static void smc911x_shutdown(struct net_device *dev)
327 unsigned cr; 329 unsigned cr;
328 unsigned long flags; 330 unsigned long flags;
329 331
330 DBG(SMC_DEBUG_FUNC, "%s: --> %s\n", CARDNAME, __func__); 332 DBG(SMC_DEBUG_FUNC, dev, "%s: --> %s\n", CARDNAME, __func__);
331 333
332 /* Disable IRQ's */ 334 /* Disable IRQ's */
333 SMC_SET_INT_EN(lp, 0); 335 SMC_SET_INT_EN(lp, 0);
@@ -346,7 +348,8 @@ static inline void smc911x_drop_pkt(struct net_device *dev)
346 struct smc911x_local *lp = netdev_priv(dev); 348 struct smc911x_local *lp = netdev_priv(dev);
347 unsigned int fifo_count, timeout, reg; 349 unsigned int fifo_count, timeout, reg;
348 350
349 DBG(SMC_DEBUG_FUNC | SMC_DEBUG_RX, "%s: --> %s\n", CARDNAME, __func__); 351 DBG(SMC_DEBUG_FUNC | SMC_DEBUG_RX, dev, "%s: --> %s\n",
352 CARDNAME, __func__);
350 fifo_count = SMC_GET_RX_FIFO_INF(lp) & 0xFFFF; 353 fifo_count = SMC_GET_RX_FIFO_INF(lp) & 0xFFFF;
351 if (fifo_count <= 4) { 354 if (fifo_count <= 4) {
352 /* Manually dump the packet data */ 355 /* Manually dump the packet data */
@@ -361,7 +364,7 @@ static inline void smc911x_drop_pkt(struct net_device *dev)
361 reg = SMC_GET_RX_DP_CTRL(lp) & RX_DP_CTRL_FFWD_BUSY_; 364 reg = SMC_GET_RX_DP_CTRL(lp) & RX_DP_CTRL_FFWD_BUSY_;
362 } while (--timeout && reg); 365 } while (--timeout && reg);
363 if (timeout == 0) { 366 if (timeout == 0) {
364 PRINTK("%s: timeout waiting for RX fast forward\n", dev->name); 367 PRINTK(dev, "timeout waiting for RX fast forward\n");
365 } 368 }
366 } 369 }
367} 370}
@@ -379,11 +382,11 @@ static inline void smc911x_rcv(struct net_device *dev)
379 struct sk_buff *skb; 382 struct sk_buff *skb;
380 unsigned char *data; 383 unsigned char *data;
381 384
382 DBG(SMC_DEBUG_FUNC | SMC_DEBUG_RX, "%s: --> %s\n", 385 DBG(SMC_DEBUG_FUNC | SMC_DEBUG_RX, dev, "--> %s\n",
383 dev->name, __func__); 386 __func__);
384 status = SMC_GET_RX_STS_FIFO(lp); 387 status = SMC_GET_RX_STS_FIFO(lp);
385 DBG(SMC_DEBUG_RX, "%s: Rx pkt len %d status 0x%08x\n", 388 DBG(SMC_DEBUG_RX, dev, "Rx pkt len %d status 0x%08x\n",
386 dev->name, (status & 0x3fff0000) >> 16, status & 0xc000ffff); 389 (status & 0x3fff0000) >> 16, status & 0xc000ffff);
387 pkt_len = (status & RX_STS_PKT_LEN_) >> 16; 390 pkt_len = (status & RX_STS_PKT_LEN_) >> 16;
388 if (status & RX_STS_ES_) { 391 if (status & RX_STS_ES_) {
389 /* Deal with a bad packet */ 392 /* Deal with a bad packet */
@@ -403,8 +406,7 @@ static inline void smc911x_rcv(struct net_device *dev)
403 /* Alloc a buffer with extra room for DMA alignment */ 406 /* Alloc a buffer with extra room for DMA alignment */
404 skb = netdev_alloc_skb(dev, pkt_len+32); 407 skb = netdev_alloc_skb(dev, pkt_len+32);
405 if (unlikely(skb == NULL)) { 408 if (unlikely(skb == NULL)) {
406 PRINTK( "%s: Low memory, rcvd packet dropped.\n", 409 PRINTK(dev, "Low memory, rcvd packet dropped.\n");
407 dev->name);
408 dev->stats.rx_dropped++; 410 dev->stats.rx_dropped++;
409 smc911x_drop_pkt(dev); 411 smc911x_drop_pkt(dev);
410 return; 412 return;
@@ -422,8 +424,8 @@ static inline void smc911x_rcv(struct net_device *dev)
422 /* Lower the FIFO threshold if possible */ 424 /* Lower the FIFO threshold if possible */
423 fifo = SMC_GET_FIFO_INT(lp); 425 fifo = SMC_GET_FIFO_INT(lp);
424 if (fifo & 0xFF) fifo--; 426 if (fifo & 0xFF) fifo--;
425 DBG(SMC_DEBUG_RX, "%s: Setting RX stat FIFO threshold to %d\n", 427 DBG(SMC_DEBUG_RX, dev, "Setting RX stat FIFO threshold to %d\n",
426 dev->name, fifo & 0xff); 428 fifo & 0xff);
427 SMC_SET_FIFO_INT(lp, fifo); 429 SMC_SET_FIFO_INT(lp, fifo);
428 /* Setup RX DMA */ 430 /* Setup RX DMA */
429 SMC_SET_RX_CFG(lp, RX_CFG_RX_END_ALGN16_ | ((2<<8) & RX_CFG_RXDOFF_)); 431 SMC_SET_RX_CFG(lp, RX_CFG_RX_END_ALGN16_ | ((2<<8) & RX_CFG_RXDOFF_));
@@ -436,7 +438,7 @@ static inline void smc911x_rcv(struct net_device *dev)
436 SMC_SET_RX_CFG(lp, RX_CFG_RX_END_ALGN4_ | ((2<<8) & RX_CFG_RXDOFF_)); 438 SMC_SET_RX_CFG(lp, RX_CFG_RX_END_ALGN4_ | ((2<<8) & RX_CFG_RXDOFF_));
437 SMC_PULL_DATA(lp, data, pkt_len+2+3); 439 SMC_PULL_DATA(lp, data, pkt_len+2+3);
438 440
439 DBG(SMC_DEBUG_PKTS, "%s: Received packet\n", dev->name); 441 DBG(SMC_DEBUG_PKTS, dev, "Received packet\n");
440 PRINT_PKT(data, ((pkt_len - 4) <= 64) ? pkt_len - 4 : 64); 442 PRINT_PKT(data, ((pkt_len - 4) <= 64) ? pkt_len - 4 : 64);
441 skb->protocol = eth_type_trans(skb, dev); 443 skb->protocol = eth_type_trans(skb, dev);
442 netif_rx(skb); 444 netif_rx(skb);
@@ -456,7 +458,7 @@ static void smc911x_hardware_send_pkt(struct net_device *dev)
456 unsigned int cmdA, cmdB, len; 458 unsigned int cmdA, cmdB, len;
457 unsigned char *buf; 459 unsigned char *buf;
458 460
459 DBG(SMC_DEBUG_FUNC | SMC_DEBUG_TX, "%s: --> %s\n", dev->name, __func__); 461 DBG(SMC_DEBUG_FUNC | SMC_DEBUG_TX, dev, "--> %s\n", __func__);
460 BUG_ON(lp->pending_tx_skb == NULL); 462 BUG_ON(lp->pending_tx_skb == NULL);
461 463
462 skb = lp->pending_tx_skb; 464 skb = lp->pending_tx_skb;
@@ -481,12 +483,12 @@ static void smc911x_hardware_send_pkt(struct net_device *dev)
481 /* tag is packet length so we can use this in stats update later */ 483 /* tag is packet length so we can use this in stats update later */
482 cmdB = (skb->len << 16) | (skb->len & 0x7FF); 484 cmdB = (skb->len << 16) | (skb->len & 0x7FF);
483 485
484 DBG(SMC_DEBUG_TX, "%s: TX PKT LENGTH 0x%04x (%d) BUF 0x%p CMDA 0x%08x CMDB 0x%08x\n", 486 DBG(SMC_DEBUG_TX, dev, "TX PKT LENGTH 0x%04x (%d) BUF 0x%p CMDA 0x%08x CMDB 0x%08x\n",
485 dev->name, len, len, buf, cmdA, cmdB); 487 len, len, buf, cmdA, cmdB);
486 SMC_SET_TX_FIFO(lp, cmdA); 488 SMC_SET_TX_FIFO(lp, cmdA);
487 SMC_SET_TX_FIFO(lp, cmdB); 489 SMC_SET_TX_FIFO(lp, cmdB);
488 490
489 DBG(SMC_DEBUG_PKTS, "%s: Transmitted packet\n", dev->name); 491 DBG(SMC_DEBUG_PKTS, dev, "Transmitted packet\n");
490 PRINT_PKT(buf, len <= 64 ? len : 64); 492 PRINT_PKT(buf, len <= 64 ? len : 64);
491 493
492 /* Send pkt via PIO or DMA */ 494 /* Send pkt via PIO or DMA */
@@ -517,20 +519,20 @@ static int smc911x_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
517 unsigned int free; 519 unsigned int free;
518 unsigned long flags; 520 unsigned long flags;
519 521
520 DBG(SMC_DEBUG_FUNC | SMC_DEBUG_TX, "%s: --> %s\n", 522 DBG(SMC_DEBUG_FUNC | SMC_DEBUG_TX, dev, "--> %s\n",
521 dev->name, __func__); 523 __func__);
522 524
523 spin_lock_irqsave(&lp->lock, flags); 525 spin_lock_irqsave(&lp->lock, flags);
524 526
525 BUG_ON(lp->pending_tx_skb != NULL); 527 BUG_ON(lp->pending_tx_skb != NULL);
526 528
527 free = SMC_GET_TX_FIFO_INF(lp) & TX_FIFO_INF_TDFREE_; 529 free = SMC_GET_TX_FIFO_INF(lp) & TX_FIFO_INF_TDFREE_;
528 DBG(SMC_DEBUG_TX, "%s: TX free space %d\n", dev->name, free); 530 DBG(SMC_DEBUG_TX, dev, "TX free space %d\n", free);
529 531
530 /* Turn off the flow when running out of space in FIFO */ 532 /* Turn off the flow when running out of space in FIFO */
531 if (free <= SMC911X_TX_FIFO_LOW_THRESHOLD) { 533 if (free <= SMC911X_TX_FIFO_LOW_THRESHOLD) {
532 DBG(SMC_DEBUG_TX, "%s: Disabling data flow due to low FIFO space (%d)\n", 534 DBG(SMC_DEBUG_TX, dev, "Disabling data flow due to low FIFO space (%d)\n",
533 dev->name, free); 535 free);
534 /* Reenable when at least 1 packet of size MTU present */ 536 /* Reenable when at least 1 packet of size MTU present */
535 SMC_SET_FIFO_TDA(lp, (SMC911X_TX_FIFO_LOW_THRESHOLD)/64); 537 SMC_SET_FIFO_TDA(lp, (SMC911X_TX_FIFO_LOW_THRESHOLD)/64);
536 lp->tx_throttle = 1; 538 lp->tx_throttle = 1;
@@ -545,8 +547,8 @@ static int smc911x_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
545 * End padding 15 bytes 547 * End padding 15 bytes
546 */ 548 */
547 if (unlikely(free < (skb->len + 8 + 15 + 15))) { 549 if (unlikely(free < (skb->len + 8 + 15 + 15))) {
548 printk("%s: No Tx free space %d < %d\n", 550 netdev_warn(dev, "No Tx free space %d < %d\n",
549 dev->name, free, skb->len); 551 free, skb->len);
550 lp->pending_tx_skb = NULL; 552 lp->pending_tx_skb = NULL;
551 dev->stats.tx_errors++; 553 dev->stats.tx_errors++;
552 dev->stats.tx_dropped++; 554 dev->stats.tx_dropped++;
@@ -561,13 +563,13 @@ static int smc911x_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
561 * the DMA IRQ starts it 563 * the DMA IRQ starts it
562 */ 564 */
563 if (lp->txdma_active) { 565 if (lp->txdma_active) {
564 DBG(SMC_DEBUG_TX | SMC_DEBUG_DMA, "%s: Tx DMA running, deferring packet\n", dev->name); 566 DBG(SMC_DEBUG_TX | SMC_DEBUG_DMA, dev, "Tx DMA running, deferring packet\n");
565 lp->pending_tx_skb = skb; 567 lp->pending_tx_skb = skb;
566 netif_stop_queue(dev); 568 netif_stop_queue(dev);
567 spin_unlock_irqrestore(&lp->lock, flags); 569 spin_unlock_irqrestore(&lp->lock, flags);
568 return NETDEV_TX_OK; 570 return NETDEV_TX_OK;
569 } else { 571 } else {
570 DBG(SMC_DEBUG_TX | SMC_DEBUG_DMA, "%s: Activating Tx DMA\n", dev->name); 572 DBG(SMC_DEBUG_TX | SMC_DEBUG_DMA, dev, "Activating Tx DMA\n");
571 lp->txdma_active = 1; 573 lp->txdma_active = 1;
572 } 574 }
573 } 575 }
@@ -589,20 +591,19 @@ static void smc911x_tx(struct net_device *dev)
589 struct smc911x_local *lp = netdev_priv(dev); 591 struct smc911x_local *lp = netdev_priv(dev);
590 unsigned int tx_status; 592 unsigned int tx_status;
591 593
592 DBG(SMC_DEBUG_FUNC | SMC_DEBUG_TX, "%s: --> %s\n", 594 DBG(SMC_DEBUG_FUNC | SMC_DEBUG_TX, dev, "--> %s\n",
593 dev->name, __func__); 595 __func__);
594 596
595 /* Collect the TX status */ 597 /* Collect the TX status */
596 while (((SMC_GET_TX_FIFO_INF(lp) & TX_FIFO_INF_TSUSED_) >> 16) != 0) { 598 while (((SMC_GET_TX_FIFO_INF(lp) & TX_FIFO_INF_TSUSED_) >> 16) != 0) {
597 DBG(SMC_DEBUG_TX, "%s: Tx stat FIFO used 0x%04x\n", 599 DBG(SMC_DEBUG_TX, dev, "Tx stat FIFO used 0x%04x\n",
598 dev->name, 600 (SMC_GET_TX_FIFO_INF(lp) & TX_FIFO_INF_TSUSED_) >> 16);
599 (SMC_GET_TX_FIFO_INF(lp) & TX_FIFO_INF_TSUSED_) >> 16);
600 tx_status = SMC_GET_TX_STS_FIFO(lp); 601 tx_status = SMC_GET_TX_STS_FIFO(lp);
601 dev->stats.tx_packets++; 602 dev->stats.tx_packets++;
602 dev->stats.tx_bytes+=tx_status>>16; 603 dev->stats.tx_bytes+=tx_status>>16;
603 DBG(SMC_DEBUG_TX, "%s: Tx FIFO tag 0x%04x status 0x%04x\n", 604 DBG(SMC_DEBUG_TX, dev, "Tx FIFO tag 0x%04x status 0x%04x\n",
604 dev->name, (tx_status & 0xffff0000) >> 16, 605 (tx_status & 0xffff0000) >> 16,
605 tx_status & 0x0000ffff); 606 tx_status & 0x0000ffff);
606 /* count Tx errors, but ignore lost carrier errors when in 607 /* count Tx errors, but ignore lost carrier errors when in
607 * full-duplex mode */ 608 * full-duplex mode */
608 if ((tx_status & TX_STS_ES_) && !(lp->ctl_rfduplx && 609 if ((tx_status & TX_STS_ES_) && !(lp->ctl_rfduplx &&
@@ -640,8 +641,8 @@ static int smc911x_phy_read(struct net_device *dev, int phyaddr, int phyreg)
640 641
641 SMC_GET_MII(lp, phyreg, phyaddr, phydata); 642 SMC_GET_MII(lp, phyreg, phyaddr, phydata);
642 643
643 DBG(SMC_DEBUG_MISC, "%s: phyaddr=0x%x, phyreg=0x%02x, phydata=0x%04x\n", 644 DBG(SMC_DEBUG_MISC, dev, "%s: phyaddr=0x%x, phyreg=0x%02x, phydata=0x%04x\n",
644 __func__, phyaddr, phyreg, phydata); 645 __func__, phyaddr, phyreg, phydata);
645 return phydata; 646 return phydata;
646} 647}
647 648
@@ -654,8 +655,8 @@ static void smc911x_phy_write(struct net_device *dev, int phyaddr, int phyreg,
654{ 655{
655 struct smc911x_local *lp = netdev_priv(dev); 656 struct smc911x_local *lp = netdev_priv(dev);
656 657
657 DBG(SMC_DEBUG_MISC, "%s: phyaddr=0x%x, phyreg=0x%x, phydata=0x%x\n", 658 DBG(SMC_DEBUG_MISC, dev, "%s: phyaddr=0x%x, phyreg=0x%x, phydata=0x%x\n",
658 __func__, phyaddr, phyreg, phydata); 659 __func__, phyaddr, phyreg, phydata);
659 660
660 SMC_SET_MII(lp, phyreg, phyaddr, phydata); 661 SMC_SET_MII(lp, phyreg, phyaddr, phydata);
661} 662}
@@ -670,7 +671,7 @@ static void smc911x_phy_detect(struct net_device *dev)
670 int phyaddr; 671 int phyaddr;
671 unsigned int cfg, id1, id2; 672 unsigned int cfg, id1, id2;
672 673
673 DBG(SMC_DEBUG_FUNC, "%s: --> %s\n", dev->name, __func__); 674 DBG(SMC_DEBUG_FUNC, dev, "--> %s\n", __func__);
674 675
675 lp->phy_type = 0; 676 lp->phy_type = 0;
676 677
@@ -731,8 +732,8 @@ static void smc911x_phy_detect(struct net_device *dev)
731 lp->phy_type = id1 << 16 | id2; 732 lp->phy_type = id1 << 16 | id2;
732 } 733 }
733 734
734 DBG(SMC_DEBUG_MISC, "%s: phy_id1=0x%x, phy_id2=0x%x phyaddr=0x%d\n", 735 DBG(SMC_DEBUG_MISC, dev, "phy_id1=0x%x, phy_id2=0x%x phyaddr=0x%d\n",
735 dev->name, id1, id2, lp->mii.phy_id); 736 id1, id2, lp->mii.phy_id);
736} 737}
737 738
738/* 739/*
@@ -745,7 +746,7 @@ static int smc911x_phy_fixed(struct net_device *dev)
745 int phyaddr = lp->mii.phy_id; 746 int phyaddr = lp->mii.phy_id;
746 int bmcr; 747 int bmcr;
747 748
748 DBG(SMC_DEBUG_FUNC, "%s: --> %s\n", dev->name, __func__); 749 DBG(SMC_DEBUG_FUNC, dev, "--> %s\n", __func__);
749 750
750 /* Enter Link Disable state */ 751 /* Enter Link Disable state */
751 SMC_GET_PHY_BMCR(lp, phyaddr, bmcr); 752 SMC_GET_PHY_BMCR(lp, phyaddr, bmcr);
@@ -792,7 +793,7 @@ static int smc911x_phy_reset(struct net_device *dev, int phy)
792 unsigned long flags; 793 unsigned long flags;
793 unsigned int reg; 794 unsigned int reg;
794 795
795 DBG(SMC_DEBUG_FUNC, "%s: --> %s()\n", dev->name, __func__); 796 DBG(SMC_DEBUG_FUNC, dev, "--> %s()\n", __func__);
796 797
797 spin_lock_irqsave(&lp->lock, flags); 798 spin_lock_irqsave(&lp->lock, flags);
798 reg = SMC_GET_PMT_CTRL(lp); 799 reg = SMC_GET_PMT_CTRL(lp);
@@ -851,18 +852,18 @@ static void smc911x_phy_check_media(struct net_device *dev, int init)
851 int phyaddr = lp->mii.phy_id; 852 int phyaddr = lp->mii.phy_id;
852 unsigned int bmcr, cr; 853 unsigned int bmcr, cr;
853 854
854 DBG(SMC_DEBUG_FUNC, "%s: --> %s\n", dev->name, __func__); 855 DBG(SMC_DEBUG_FUNC, dev, "--> %s\n", __func__);
855 856
856 if (mii_check_media(&lp->mii, netif_msg_link(lp), init)) { 857 if (mii_check_media(&lp->mii, netif_msg_link(lp), init)) {
857 /* duplex state has changed */ 858 /* duplex state has changed */
858 SMC_GET_PHY_BMCR(lp, phyaddr, bmcr); 859 SMC_GET_PHY_BMCR(lp, phyaddr, bmcr);
859 SMC_GET_MAC_CR(lp, cr); 860 SMC_GET_MAC_CR(lp, cr);
860 if (lp->mii.full_duplex) { 861 if (lp->mii.full_duplex) {
861 DBG(SMC_DEBUG_MISC, "%s: Configuring for full-duplex mode\n", dev->name); 862 DBG(SMC_DEBUG_MISC, dev, "Configuring for full-duplex mode\n");
862 bmcr |= BMCR_FULLDPLX; 863 bmcr |= BMCR_FULLDPLX;
863 cr |= MAC_CR_RCVOWN_; 864 cr |= MAC_CR_RCVOWN_;
864 } else { 865 } else {
865 DBG(SMC_DEBUG_MISC, "%s: Configuring for half-duplex mode\n", dev->name); 866 DBG(SMC_DEBUG_MISC, dev, "Configuring for half-duplex mode\n");
866 bmcr &= ~BMCR_FULLDPLX; 867 bmcr &= ~BMCR_FULLDPLX;
867 cr &= ~MAC_CR_RCVOWN_; 868 cr &= ~MAC_CR_RCVOWN_;
868 } 869 }
@@ -891,7 +892,7 @@ static void smc911x_phy_configure(struct work_struct *work)
891 int status; 892 int status;
892 unsigned long flags; 893 unsigned long flags;
893 894
894 DBG(SMC_DEBUG_FUNC, "%s: --> %s()\n", dev->name, __func__); 895 DBG(SMC_DEBUG_FUNC, dev, "--> %s()\n", __func__);
895 896
896 /* 897 /*
897 * We should not be called if phy_type is zero. 898 * We should not be called if phy_type is zero.
@@ -900,7 +901,7 @@ static void smc911x_phy_configure(struct work_struct *work)
900 return; 901 return;
901 902
902 if (smc911x_phy_reset(dev, phyaddr)) { 903 if (smc911x_phy_reset(dev, phyaddr)) {
903 printk("%s: PHY reset timed out\n", dev->name); 904 netdev_info(dev, "PHY reset timed out\n");
904 return; 905 return;
905 } 906 }
906 spin_lock_irqsave(&lp->lock, flags); 907 spin_lock_irqsave(&lp->lock, flags);
@@ -922,7 +923,7 @@ static void smc911x_phy_configure(struct work_struct *work)
922 /* Copy our capabilities from MII_BMSR to MII_ADVERTISE */ 923 /* Copy our capabilities from MII_BMSR to MII_ADVERTISE */
923 SMC_GET_PHY_BMSR(lp, phyaddr, my_phy_caps); 924 SMC_GET_PHY_BMSR(lp, phyaddr, my_phy_caps);
924 if (!(my_phy_caps & BMSR_ANEGCAPABLE)) { 925 if (!(my_phy_caps & BMSR_ANEGCAPABLE)) {
925 printk(KERN_INFO "Auto negotiation NOT supported\n"); 926 netdev_info(dev, "Auto negotiation NOT supported\n");
926 smc911x_phy_fixed(dev); 927 smc911x_phy_fixed(dev);
927 goto smc911x_phy_configure_exit; 928 goto smc911x_phy_configure_exit;
928 } 929 }
@@ -960,8 +961,8 @@ static void smc911x_phy_configure(struct work_struct *work)
960 udelay(10); 961 udelay(10);
961 SMC_GET_PHY_MII_ADV(lp, phyaddr, status); 962 SMC_GET_PHY_MII_ADV(lp, phyaddr, status);
962 963
963 DBG(SMC_DEBUG_MISC, "%s: phy caps=0x%04x\n", dev->name, my_phy_caps); 964 DBG(SMC_DEBUG_MISC, dev, "phy caps=0x%04x\n", my_phy_caps);
964 DBG(SMC_DEBUG_MISC, "%s: phy advertised caps=0x%04x\n", dev->name, my_ad_caps); 965 DBG(SMC_DEBUG_MISC, dev, "phy advertised caps=0x%04x\n", my_ad_caps);
965 966
966 /* Restart auto-negotiation process in order to advertise my caps */ 967 /* Restart auto-negotiation process in order to advertise my caps */
967 SMC_SET_PHY_BMCR(lp, phyaddr, BMCR_ANENABLE | BMCR_ANRESTART); 968 SMC_SET_PHY_BMCR(lp, phyaddr, BMCR_ANENABLE | BMCR_ANRESTART);
@@ -984,7 +985,7 @@ static void smc911x_phy_interrupt(struct net_device *dev)
984 int phyaddr = lp->mii.phy_id; 985 int phyaddr = lp->mii.phy_id;
985 int status; 986 int status;
986 987
987 DBG(SMC_DEBUG_FUNC, "%s: --> %s\n", dev->name, __func__); 988 DBG(SMC_DEBUG_FUNC, dev, "--> %s\n", __func__);
988 989
989 if (lp->phy_type == 0) 990 if (lp->phy_type == 0)
990 return; 991 return;
@@ -992,10 +993,10 @@ static void smc911x_phy_interrupt(struct net_device *dev)
992 smc911x_phy_check_media(dev, 0); 993 smc911x_phy_check_media(dev, 0);
993 /* read to clear status bits */ 994 /* read to clear status bits */
994 SMC_GET_PHY_INT_SRC(lp, phyaddr,status); 995 SMC_GET_PHY_INT_SRC(lp, phyaddr,status);
995 DBG(SMC_DEBUG_MISC, "%s: PHY interrupt status 0x%04x\n", 996 DBG(SMC_DEBUG_MISC, dev, "PHY interrupt status 0x%04x\n",
996 dev->name, status & 0xffff); 997 status & 0xffff);
997 DBG(SMC_DEBUG_MISC, "%s: AFC_CFG 0x%08x\n", 998 DBG(SMC_DEBUG_MISC, dev, "AFC_CFG 0x%08x\n",
998 dev->name, SMC_GET_AFC_CFG(lp)); 999 SMC_GET_AFC_CFG(lp));
999} 1000}
1000 1001
1001/*--- END PHY CONTROL AND CONFIGURATION-------------------------------------*/ 1002/*--- END PHY CONTROL AND CONFIGURATION-------------------------------------*/
@@ -1012,7 +1013,7 @@ static irqreturn_t smc911x_interrupt(int irq, void *dev_id)
1012 unsigned int rx_overrun=0, cr, pkts; 1013 unsigned int rx_overrun=0, cr, pkts;
1013 unsigned long flags; 1014 unsigned long flags;
1014 1015
1015 DBG(SMC_DEBUG_FUNC, "%s: --> %s\n", dev->name, __func__); 1016 DBG(SMC_DEBUG_FUNC, dev, "--> %s\n", __func__);
1016 1017
1017 spin_lock_irqsave(&lp->lock, flags); 1018 spin_lock_irqsave(&lp->lock, flags);
1018 1019
@@ -1033,8 +1034,8 @@ static irqreturn_t smc911x_interrupt(int irq, void *dev_id)
1033 do { 1034 do {
1034 status = SMC_GET_INT(lp); 1035 status = SMC_GET_INT(lp);
1035 1036
1036 DBG(SMC_DEBUG_MISC, "%s: INT 0x%08x MASK 0x%08x OUTSIDE MASK 0x%08x\n", 1037 DBG(SMC_DEBUG_MISC, dev, "INT 0x%08x MASK 0x%08x OUTSIDE MASK 0x%08x\n",
1037 dev->name, status, mask, status & ~mask); 1038 status, mask, status & ~mask);
1038 1039
1039 status &= mask; 1040 status &= mask;
1040 if (!status) 1041 if (!status)
@@ -1066,7 +1067,7 @@ static irqreturn_t smc911x_interrupt(int irq, void *dev_id)
1066 SMC_GET_MAC_CR(lp, cr); 1067 SMC_GET_MAC_CR(lp, cr);
1067 cr &= ~MAC_CR_RXEN_; 1068 cr &= ~MAC_CR_RXEN_;
1068 SMC_SET_MAC_CR(lp, cr); 1069 SMC_SET_MAC_CR(lp, cr);
1069 DBG(SMC_DEBUG_RX, "%s: RX overrun\n", dev->name); 1070 DBG(SMC_DEBUG_RX, dev, "RX overrun\n");
1070 dev->stats.rx_errors++; 1071 dev->stats.rx_errors++;
1071 dev->stats.rx_fifo_errors++; 1072 dev->stats.rx_fifo_errors++;
1072 } 1073 }
@@ -1078,7 +1079,7 @@ static irqreturn_t smc911x_interrupt(int irq, void *dev_id)
1078 cr &= ~MAC_CR_RXEN_; 1079 cr &= ~MAC_CR_RXEN_;
1079 SMC_SET_MAC_CR(lp, cr); 1080 SMC_SET_MAC_CR(lp, cr);
1080 rx_overrun=1; 1081 rx_overrun=1;
1081 DBG(SMC_DEBUG_RX, "%s: RX overrun\n", dev->name); 1082 DBG(SMC_DEBUG_RX, dev, "RX overrun\n");
1082 dev->stats.rx_errors++; 1083 dev->stats.rx_errors++;
1083 dev->stats.rx_fifo_errors++; 1084 dev->stats.rx_fifo_errors++;
1084 } 1085 }
@@ -1087,23 +1088,23 @@ static irqreturn_t smc911x_interrupt(int irq, void *dev_id)
1087 /* Handle receive condition */ 1088 /* Handle receive condition */
1088 if ((status & INT_STS_RSFL_) || rx_overrun) { 1089 if ((status & INT_STS_RSFL_) || rx_overrun) {
1089 unsigned int fifo; 1090 unsigned int fifo;
1090 DBG(SMC_DEBUG_RX, "%s: RX irq\n", dev->name); 1091 DBG(SMC_DEBUG_RX, dev, "RX irq\n");
1091 fifo = SMC_GET_RX_FIFO_INF(lp); 1092 fifo = SMC_GET_RX_FIFO_INF(lp);
1092 pkts = (fifo & RX_FIFO_INF_RXSUSED_) >> 16; 1093 pkts = (fifo & RX_FIFO_INF_RXSUSED_) >> 16;
1093 DBG(SMC_DEBUG_RX, "%s: Rx FIFO pkts %d, bytes %d\n", 1094 DBG(SMC_DEBUG_RX, dev, "Rx FIFO pkts %d, bytes %d\n",
1094 dev->name, pkts, fifo & 0xFFFF ); 1095 pkts, fifo & 0xFFFF);
1095 if (pkts != 0) { 1096 if (pkts != 0) {
1096#ifdef SMC_USE_DMA 1097#ifdef SMC_USE_DMA
1097 unsigned int fifo; 1098 unsigned int fifo;
1098 if (lp->rxdma_active){ 1099 if (lp->rxdma_active){
1099 DBG(SMC_DEBUG_RX | SMC_DEBUG_DMA, 1100 DBG(SMC_DEBUG_RX | SMC_DEBUG_DMA, dev,
1100 "%s: RX DMA active\n", dev->name); 1101 "RX DMA active\n");
1101 /* The DMA is already running so up the IRQ threshold */ 1102 /* The DMA is already running so up the IRQ threshold */
1102 fifo = SMC_GET_FIFO_INT(lp) & ~0xFF; 1103 fifo = SMC_GET_FIFO_INT(lp) & ~0xFF;
1103 fifo |= pkts & 0xFF; 1104 fifo |= pkts & 0xFF;
1104 DBG(SMC_DEBUG_RX, 1105 DBG(SMC_DEBUG_RX, dev,
1105 "%s: Setting RX stat FIFO threshold to %d\n", 1106 "Setting RX stat FIFO threshold to %d\n",
1106 dev->name, fifo & 0xff); 1107 fifo & 0xff);
1107 SMC_SET_FIFO_INT(lp, fifo); 1108 SMC_SET_FIFO_INT(lp, fifo);
1108 } else 1109 } else
1109#endif 1110#endif
@@ -1113,7 +1114,7 @@ static irqreturn_t smc911x_interrupt(int irq, void *dev_id)
1113 } 1114 }
1114 /* Handle transmit FIFO available */ 1115 /* Handle transmit FIFO available */
1115 if (status & INT_STS_TDFA_) { 1116 if (status & INT_STS_TDFA_) {
1116 DBG(SMC_DEBUG_TX, "%s: TX data FIFO space available irq\n", dev->name); 1117 DBG(SMC_DEBUG_TX, dev, "TX data FIFO space available irq\n");
1117 SMC_SET_FIFO_TDA(lp, 0xFF); 1118 SMC_SET_FIFO_TDA(lp, 0xFF);
1118 lp->tx_throttle = 0; 1119 lp->tx_throttle = 0;
1119#ifdef SMC_USE_DMA 1120#ifdef SMC_USE_DMA
@@ -1125,9 +1126,9 @@ static irqreturn_t smc911x_interrupt(int irq, void *dev_id)
1125 /* Handle transmit done condition */ 1126 /* Handle transmit done condition */
1126#if 1 1127#if 1
1127 if (status & (INT_STS_TSFL_ | INT_STS_GPT_INT_)) { 1128 if (status & (INT_STS_TSFL_ | INT_STS_GPT_INT_)) {
1128 DBG(SMC_DEBUG_TX | SMC_DEBUG_MISC, 1129 DBG(SMC_DEBUG_TX | SMC_DEBUG_MISC, dev,
1129 "%s: Tx stat FIFO limit (%d) /GPT irq\n", 1130 "Tx stat FIFO limit (%d) /GPT irq\n",
1130 dev->name, (SMC_GET_FIFO_INT(lp) & 0x00ff0000) >> 16); 1131 (SMC_GET_FIFO_INT(lp) & 0x00ff0000) >> 16);
1131 smc911x_tx(dev); 1132 smc911x_tx(dev);
1132 SMC_SET_GPT_CFG(lp, GPT_CFG_TIMER_EN_ | 10000); 1133 SMC_SET_GPT_CFG(lp, GPT_CFG_TIMER_EN_ | 10000);
1133 SMC_ACK_INT(lp, INT_STS_TSFL_); 1134 SMC_ACK_INT(lp, INT_STS_TSFL_);
@@ -1135,23 +1136,20 @@ static irqreturn_t smc911x_interrupt(int irq, void *dev_id)
1135 } 1136 }
1136#else 1137#else
1137 if (status & INT_STS_TSFL_) { 1138 if (status & INT_STS_TSFL_) {
1138 DBG(SMC_DEBUG_TX, "%s: TX status FIFO limit (%d) irq\n", dev->name, ); 1139 DBG(SMC_DEBUG_TX, dev, "TX status FIFO limit (%d) irq\n", ?);
1139 smc911x_tx(dev); 1140 smc911x_tx(dev);
1140 SMC_ACK_INT(lp, INT_STS_TSFL_); 1141 SMC_ACK_INT(lp, INT_STS_TSFL_);
1141 } 1142 }
1142 1143
1143 if (status & INT_STS_GPT_INT_) { 1144 if (status & INT_STS_GPT_INT_) {
1144 DBG(SMC_DEBUG_RX, "%s: IRQ_CFG 0x%08x FIFO_INT 0x%08x RX_CFG 0x%08x\n", 1145 DBG(SMC_DEBUG_RX, dev, "IRQ_CFG 0x%08x FIFO_INT 0x%08x RX_CFG 0x%08x\n",
1145 dev->name, 1146 SMC_GET_IRQ_CFG(lp),
1146 SMC_GET_IRQ_CFG(lp), 1147 SMC_GET_FIFO_INT(lp),
1147 SMC_GET_FIFO_INT(lp), 1148 SMC_GET_RX_CFG(lp));
1148 SMC_GET_RX_CFG(lp)); 1149 DBG(SMC_DEBUG_RX, dev, "Rx Stat FIFO Used 0x%02x Data FIFO Used 0x%04x Stat FIFO 0x%08x\n",
1149 DBG(SMC_DEBUG_RX, "%s: Rx Stat FIFO Used 0x%02x " 1150 (SMC_GET_RX_FIFO_INF(lp) & 0x00ff0000) >> 16,
1150 "Data FIFO Used 0x%04x Stat FIFO 0x%08x\n", 1151 SMC_GET_RX_FIFO_INF(lp) & 0xffff,
1151 dev->name, 1152 SMC_GET_RX_STS_FIFO_PEEK(lp));
1152 (SMC_GET_RX_FIFO_INF(lp) & 0x00ff0000) >> 16,
1153 SMC_GET_RX_FIFO_INF(lp) & 0xffff,
1154 SMC_GET_RX_STS_FIFO_PEEK(lp));
1155 SMC_SET_GPT_CFG(lp, GPT_CFG_TIMER_EN_ | 10000); 1153 SMC_SET_GPT_CFG(lp, GPT_CFG_TIMER_EN_ | 10000);
1156 SMC_ACK_INT(lp, INT_STS_GPT_INT_); 1154 SMC_ACK_INT(lp, INT_STS_GPT_INT_);
1157 } 1155 }
@@ -1159,7 +1157,7 @@ static irqreturn_t smc911x_interrupt(int irq, void *dev_id)
1159 1157
1160 /* Handle PHY interrupt condition */ 1158 /* Handle PHY interrupt condition */
1161 if (status & INT_STS_PHY_INT_) { 1159 if (status & INT_STS_PHY_INT_) {
1162 DBG(SMC_DEBUG_MISC, "%s: PHY irq\n", dev->name); 1160 DBG(SMC_DEBUG_MISC, dev, "PHY irq\n");
1163 smc911x_phy_interrupt(dev); 1161 smc911x_phy_interrupt(dev);
1164 SMC_ACK_INT(lp, INT_STS_PHY_INT_); 1162 SMC_ACK_INT(lp, INT_STS_PHY_INT_);
1165 } 1163 }
@@ -1168,8 +1166,8 @@ static irqreturn_t smc911x_interrupt(int irq, void *dev_id)
1168 /* restore mask state */ 1166 /* restore mask state */
1169 SMC_SET_INT_EN(lp, mask); 1167 SMC_SET_INT_EN(lp, mask);
1170 1168
1171 DBG(SMC_DEBUG_MISC, "%s: Interrupt done (%d loops)\n", 1169 DBG(SMC_DEBUG_MISC, dev, "Interrupt done (%d loops)\n",
1172 dev->name, 8-timeout); 1170 8-timeout);
1173 1171
1174 spin_unlock_irqrestore(&lp->lock, flags); 1172 spin_unlock_irqrestore(&lp->lock, flags);
1175 1173
@@ -1185,9 +1183,9 @@ smc911x_tx_dma_irq(int dma, void *data)
1185 struct sk_buff *skb = lp->current_tx_skb; 1183 struct sk_buff *skb = lp->current_tx_skb;
1186 unsigned long flags; 1184 unsigned long flags;
1187 1185
1188 DBG(SMC_DEBUG_FUNC, "%s: --> %s\n", dev->name, __func__); 1186 DBG(SMC_DEBUG_FUNC, dev, "--> %s\n", __func__);
1189 1187
1190 DBG(SMC_DEBUG_TX | SMC_DEBUG_DMA, "%s: TX DMA irq handler\n", dev->name); 1188 DBG(SMC_DEBUG_TX | SMC_DEBUG_DMA, dev, "TX DMA irq handler\n");
1191 /* Clear the DMA interrupt sources */ 1189 /* Clear the DMA interrupt sources */
1192 SMC_DMA_ACK_IRQ(dev, dma); 1190 SMC_DMA_ACK_IRQ(dev, dma);
1193 BUG_ON(skb == NULL); 1191 BUG_ON(skb == NULL);
@@ -1198,8 +1196,8 @@ smc911x_tx_dma_irq(int dma, void *data)
1198 if (lp->pending_tx_skb != NULL) 1196 if (lp->pending_tx_skb != NULL)
1199 smc911x_hardware_send_pkt(dev); 1197 smc911x_hardware_send_pkt(dev);
1200 else { 1198 else {
1201 DBG(SMC_DEBUG_TX | SMC_DEBUG_DMA, 1199 DBG(SMC_DEBUG_TX | SMC_DEBUG_DMA, dev,
1202 "%s: No pending Tx packets. DMA disabled\n", dev->name); 1200 "No pending Tx packets. DMA disabled\n");
1203 spin_lock_irqsave(&lp->lock, flags); 1201 spin_lock_irqsave(&lp->lock, flags);
1204 lp->txdma_active = 0; 1202 lp->txdma_active = 0;
1205 if (!lp->tx_throttle) { 1203 if (!lp->tx_throttle) {
@@ -1208,8 +1206,8 @@ smc911x_tx_dma_irq(int dma, void *data)
1208 spin_unlock_irqrestore(&lp->lock, flags); 1206 spin_unlock_irqrestore(&lp->lock, flags);
1209 } 1207 }
1210 1208
1211 DBG(SMC_DEBUG_TX | SMC_DEBUG_DMA, 1209 DBG(SMC_DEBUG_TX | SMC_DEBUG_DMA, dev,
1212 "%s: TX DMA irq completed\n", dev->name); 1210 "TX DMA irq completed\n");
1213} 1211}
1214static void 1212static void
1215smc911x_rx_dma_irq(int dma, void *data) 1213smc911x_rx_dma_irq(int dma, void *data)
@@ -1221,8 +1219,8 @@ smc911x_rx_dma_irq(int dma, void *data)
1221 unsigned long flags; 1219 unsigned long flags;
1222 unsigned int pkts; 1220 unsigned int pkts;
1223 1221
1224 DBG(SMC_DEBUG_FUNC, "%s: --> %s\n", dev->name, __func__); 1222 DBG(SMC_DEBUG_FUNC, dev, "--> %s\n", __func__);
1225 DBG(SMC_DEBUG_RX | SMC_DEBUG_DMA, "%s: RX DMA irq handler\n", dev->name); 1223 DBG(SMC_DEBUG_RX | SMC_DEBUG_DMA, dev, "RX DMA irq handler\n");
1226 /* Clear the DMA interrupt sources */ 1224 /* Clear the DMA interrupt sources */
1227 SMC_DMA_ACK_IRQ(dev, dma); 1225 SMC_DMA_ACK_IRQ(dev, dma);
1228 dma_unmap_single(NULL, rx_dmabuf, rx_dmalen, DMA_FROM_DEVICE); 1226 dma_unmap_single(NULL, rx_dmabuf, rx_dmalen, DMA_FROM_DEVICE);
@@ -1242,9 +1240,9 @@ smc911x_rx_dma_irq(int dma, void *data)
1242 lp->rxdma_active = 0; 1240 lp->rxdma_active = 0;
1243 } 1241 }
1244 spin_unlock_irqrestore(&lp->lock, flags); 1242 spin_unlock_irqrestore(&lp->lock, flags);
1245 DBG(SMC_DEBUG_RX | SMC_DEBUG_DMA, 1243 DBG(SMC_DEBUG_RX | SMC_DEBUG_DMA, dev,
1246 "%s: RX DMA irq completed. DMA RX FIFO PKTS %d\n", 1244 "RX DMA irq completed. DMA RX FIFO PKTS %d\n",
1247 dev->name, pkts); 1245 pkts);
1248} 1246}
1249#endif /* SMC_USE_DMA */ 1247#endif /* SMC_USE_DMA */
1250 1248
@@ -1268,14 +1266,14 @@ static void smc911x_timeout(struct net_device *dev)
1268 int status, mask; 1266 int status, mask;
1269 unsigned long flags; 1267 unsigned long flags;
1270 1268
1271 DBG(SMC_DEBUG_FUNC, "%s: --> %s\n", dev->name, __func__); 1269 DBG(SMC_DEBUG_FUNC, dev, "--> %s\n", __func__);
1272 1270
1273 spin_lock_irqsave(&lp->lock, flags); 1271 spin_lock_irqsave(&lp->lock, flags);
1274 status = SMC_GET_INT(lp); 1272 status = SMC_GET_INT(lp);
1275 mask = SMC_GET_INT_EN(lp); 1273 mask = SMC_GET_INT_EN(lp);
1276 spin_unlock_irqrestore(&lp->lock, flags); 1274 spin_unlock_irqrestore(&lp->lock, flags);
1277 DBG(SMC_DEBUG_MISC, "%s: INT 0x%02x MASK 0x%02x\n", 1275 DBG(SMC_DEBUG_MISC, dev, "INT 0x%02x MASK 0x%02x\n",
1278 dev->name, status, mask); 1276 status, mask);
1279 1277
1280 /* Dump the current TX FIFO contents and restart */ 1278 /* Dump the current TX FIFO contents and restart */
1281 mask = SMC_GET_TX_CFG(lp); 1279 mask = SMC_GET_TX_CFG(lp);
@@ -1306,7 +1304,7 @@ static void smc911x_set_multicast_list(struct net_device *dev)
1306 unsigned int mcr, update_multicast = 0; 1304 unsigned int mcr, update_multicast = 0;
1307 unsigned long flags; 1305 unsigned long flags;
1308 1306
1309 DBG(SMC_DEBUG_FUNC, "%s: --> %s\n", dev->name, __func__); 1307 DBG(SMC_DEBUG_FUNC, dev, "--> %s\n", __func__);
1310 1308
1311 spin_lock_irqsave(&lp->lock, flags); 1309 spin_lock_irqsave(&lp->lock, flags);
1312 SMC_GET_MAC_CR(lp, mcr); 1310 SMC_GET_MAC_CR(lp, mcr);
@@ -1314,7 +1312,7 @@ static void smc911x_set_multicast_list(struct net_device *dev)
1314 1312
1315 if (dev->flags & IFF_PROMISC) { 1313 if (dev->flags & IFF_PROMISC) {
1316 1314
1317 DBG(SMC_DEBUG_MISC, "%s: RCR_PRMS\n", dev->name); 1315 DBG(SMC_DEBUG_MISC, dev, "RCR_PRMS\n");
1318 mcr |= MAC_CR_PRMS_; 1316 mcr |= MAC_CR_PRMS_;
1319 } 1317 }
1320 /* 1318 /*
@@ -1323,7 +1321,7 @@ static void smc911x_set_multicast_list(struct net_device *dev)
1323 * checked before the table is 1321 * checked before the table is
1324 */ 1322 */
1325 else if (dev->flags & IFF_ALLMULTI || netdev_mc_count(dev) > 16) { 1323 else if (dev->flags & IFF_ALLMULTI || netdev_mc_count(dev) > 16) {
1326 DBG(SMC_DEBUG_MISC, "%s: RCR_ALMUL\n", dev->name); 1324 DBG(SMC_DEBUG_MISC, dev, "RCR_ALMUL\n");
1327 mcr |= MAC_CR_MCPAS_; 1325 mcr |= MAC_CR_MCPAS_;
1328 } 1326 }
1329 1327
@@ -1363,8 +1361,7 @@ static void smc911x_set_multicast_list(struct net_device *dev)
1363 /* now, the table can be loaded into the chipset */ 1361 /* now, the table can be loaded into the chipset */
1364 update_multicast = 1; 1362 update_multicast = 1;
1365 } else { 1363 } else {
1366 DBG(SMC_DEBUG_MISC, "%s: ~(MAC_CR_PRMS_|MAC_CR_MCPAS_)\n", 1364 DBG(SMC_DEBUG_MISC, dev, "~(MAC_CR_PRMS_|MAC_CR_MCPAS_)\n");
1367 dev->name);
1368 mcr &= ~(MAC_CR_PRMS_ | MAC_CR_MCPAS_); 1365 mcr &= ~(MAC_CR_PRMS_ | MAC_CR_MCPAS_);
1369 1366
1370 /* 1367 /*
@@ -1378,9 +1375,9 @@ static void smc911x_set_multicast_list(struct net_device *dev)
1378 spin_lock_irqsave(&lp->lock, flags); 1375 spin_lock_irqsave(&lp->lock, flags);
1379 SMC_SET_MAC_CR(lp, mcr); 1376 SMC_SET_MAC_CR(lp, mcr);
1380 if (update_multicast) { 1377 if (update_multicast) {
1381 DBG(SMC_DEBUG_MISC, 1378 DBG(SMC_DEBUG_MISC, dev,
1382 "%s: update mcast hash table 0x%08x 0x%08x\n", 1379 "update mcast hash table 0x%08x 0x%08x\n",
1383 dev->name, multicast_table[0], multicast_table[1]); 1380 multicast_table[0], multicast_table[1]);
1384 SMC_SET_HASHL(lp, multicast_table[0]); 1381 SMC_SET_HASHL(lp, multicast_table[0]);
1385 SMC_SET_HASHH(lp, multicast_table[1]); 1382 SMC_SET_HASHH(lp, multicast_table[1]);
1386 } 1383 }
@@ -1398,7 +1395,7 @@ smc911x_open(struct net_device *dev)
1398{ 1395{
1399 struct smc911x_local *lp = netdev_priv(dev); 1396 struct smc911x_local *lp = netdev_priv(dev);
1400 1397
1401 DBG(SMC_DEBUG_FUNC, "%s: --> %s\n", dev->name, __func__); 1398 DBG(SMC_DEBUG_FUNC, dev, "--> %s\n", __func__);
1402 1399
1403 /* reset the hardware */ 1400 /* reset the hardware */
1404 smc911x_reset(dev); 1401 smc911x_reset(dev);
@@ -1425,7 +1422,7 @@ static int smc911x_close(struct net_device *dev)
1425{ 1422{
1426 struct smc911x_local *lp = netdev_priv(dev); 1423 struct smc911x_local *lp = netdev_priv(dev);
1427 1424
1428 DBG(SMC_DEBUG_FUNC, "%s: --> %s\n", dev->name, __func__); 1425 DBG(SMC_DEBUG_FUNC, dev, "--> %s\n", __func__);
1429 1426
1430 netif_stop_queue(dev); 1427 netif_stop_queue(dev);
1431 netif_carrier_off(dev); 1428 netif_carrier_off(dev);
@@ -1459,7 +1456,7 @@ smc911x_ethtool_getsettings(struct net_device *dev, struct ethtool_cmd *cmd)
1459 int ret, status; 1456 int ret, status;
1460 unsigned long flags; 1457 unsigned long flags;
1461 1458
1462 DBG(SMC_DEBUG_FUNC, "%s: --> %s\n", dev->name, __func__); 1459 DBG(SMC_DEBUG_FUNC, dev, "--> %s\n", __func__);
1463 cmd->maxtxpkt = 1; 1460 cmd->maxtxpkt = 1;
1464 cmd->maxrxpkt = 1; 1461 cmd->maxrxpkt = 1;
1465 1462
@@ -1597,16 +1594,16 @@ static int smc911x_ethtool_wait_eeprom_ready(struct net_device *dev)
1597 e2p_cmd = SMC_GET_E2P_CMD(lp); 1594 e2p_cmd = SMC_GET_E2P_CMD(lp);
1598 for(timeout=10;(e2p_cmd & E2P_CMD_EPC_BUSY_) && timeout; timeout--) { 1595 for(timeout=10;(e2p_cmd & E2P_CMD_EPC_BUSY_) && timeout; timeout--) {
1599 if (e2p_cmd & E2P_CMD_EPC_TIMEOUT_) { 1596 if (e2p_cmd & E2P_CMD_EPC_TIMEOUT_) {
1600 PRINTK("%s: %s timeout waiting for EEPROM to respond\n", 1597 PRINTK(dev, "%s timeout waiting for EEPROM to respond\n",
1601 dev->name, __func__); 1598 __func__);
1602 return -EFAULT; 1599 return -EFAULT;
1603 } 1600 }
1604 mdelay(1); 1601 mdelay(1);
1605 e2p_cmd = SMC_GET_E2P_CMD(lp); 1602 e2p_cmd = SMC_GET_E2P_CMD(lp);
1606 } 1603 }
1607 if (timeout == 0) { 1604 if (timeout == 0) {
1608 PRINTK("%s: %s timeout waiting for EEPROM CMD not busy\n", 1605 PRINTK(dev, "%s timeout waiting for EEPROM CMD not busy\n",
1609 dev->name, __func__); 1606 __func__);
1610 return -ETIMEDOUT; 1607 return -ETIMEDOUT;
1611 } 1608 }
1612 return 0; 1609 return 0;
@@ -1719,7 +1716,7 @@ static int smc911x_findirq(struct net_device *dev)
1719 int timeout = 20; 1716 int timeout = 20;
1720 unsigned long cookie; 1717 unsigned long cookie;
1721 1718
1722 DBG(SMC_DEBUG_FUNC, "--> %s\n", __func__); 1719 DBG(SMC_DEBUG_FUNC, dev, "--> %s\n", __func__);
1723 1720
1724 cookie = probe_irq_on(); 1721 cookie = probe_irq_on();
1725 1722
@@ -1799,13 +1796,14 @@ static int smc911x_probe(struct net_device *dev)
1799 const char *version_string; 1796 const char *version_string;
1800 unsigned long irq_flags; 1797 unsigned long irq_flags;
1801 1798
1802 DBG(SMC_DEBUG_FUNC, "%s: --> %s\n", dev->name, __func__); 1799 DBG(SMC_DEBUG_FUNC, dev, "--> %s\n", __func__);
1803 1800
1804 /* First, see if the endian word is recognized */ 1801 /* First, see if the endian word is recognized */
1805 val = SMC_GET_BYTE_TEST(lp); 1802 val = SMC_GET_BYTE_TEST(lp);
1806 DBG(SMC_DEBUG_MISC, "%s: endian probe returned 0x%04x\n", CARDNAME, val); 1803 DBG(SMC_DEBUG_MISC, dev, "%s: endian probe returned 0x%04x\n",
1804 CARDNAME, val);
1807 if (val != 0x87654321) { 1805 if (val != 0x87654321) {
1808 printk(KERN_ERR "Invalid chip endian 0x%08x\n",val); 1806 netdev_err(dev, "Invalid chip endian 0x%08x\n", val);
1809 retval = -ENODEV; 1807 retval = -ENODEV;
1810 goto err_out; 1808 goto err_out;
1811 } 1809 }
@@ -1816,26 +1814,29 @@ static int smc911x_probe(struct net_device *dev)
1816 * as future revisions could be added. 1814 * as future revisions could be added.
1817 */ 1815 */
1818 chip_id = SMC_GET_PN(lp); 1816 chip_id = SMC_GET_PN(lp);
1819 DBG(SMC_DEBUG_MISC, "%s: id probe returned 0x%04x\n", CARDNAME, chip_id); 1817 DBG(SMC_DEBUG_MISC, dev, "%s: id probe returned 0x%04x\n",
1818 CARDNAME, chip_id);
1820 for(i=0;chip_ids[i].id != 0; i++) { 1819 for(i=0;chip_ids[i].id != 0; i++) {
1821 if (chip_ids[i].id == chip_id) break; 1820 if (chip_ids[i].id == chip_id) break;
1822 } 1821 }
1823 if (!chip_ids[i].id) { 1822 if (!chip_ids[i].id) {
1824 printk(KERN_ERR "Unknown chip ID %04x\n", chip_id); 1823 netdev_err(dev, "Unknown chip ID %04x\n", chip_id);
1825 retval = -ENODEV; 1824 retval = -ENODEV;
1826 goto err_out; 1825 goto err_out;
1827 } 1826 }
1828 version_string = chip_ids[i].name; 1827 version_string = chip_ids[i].name;
1829 1828
1830 revision = SMC_GET_REV(lp); 1829 revision = SMC_GET_REV(lp);
1831 DBG(SMC_DEBUG_MISC, "%s: revision = 0x%04x\n", CARDNAME, revision); 1830 DBG(SMC_DEBUG_MISC, dev, "%s: revision = 0x%04x\n", CARDNAME, revision);
1832 1831
1833 /* At this point I'll assume that the chip is an SMC911x. */ 1832 /* At this point I'll assume that the chip is an SMC911x. */
1834 DBG(SMC_DEBUG_MISC, "%s: Found a %s\n", CARDNAME, chip_ids[i].name); 1833 DBG(SMC_DEBUG_MISC, dev, "%s: Found a %s\n",
1834 CARDNAME, chip_ids[i].name);
1835 1835
1836 /* Validate the TX FIFO size requested */ 1836 /* Validate the TX FIFO size requested */
1837 if ((tx_fifo_kb < 2) || (tx_fifo_kb > 14)) { 1837 if ((tx_fifo_kb < 2) || (tx_fifo_kb > 14)) {
1838 printk(KERN_ERR "Invalid TX FIFO size requested %d\n", tx_fifo_kb); 1838 netdev_err(dev, "Invalid TX FIFO size requested %d\n",
1839 tx_fifo_kb);
1839 retval = -EINVAL; 1840 retval = -EINVAL;
1840 goto err_out; 1841 goto err_out;
1841 } 1842 }
@@ -1887,14 +1888,13 @@ static int smc911x_probe(struct net_device *dev)
1887 case 14:/* 1920 Rx Data Fifo Size */ 1888 case 14:/* 1920 Rx Data Fifo Size */
1888 lp->afc_cfg=0x0006032F;break; 1889 lp->afc_cfg=0x0006032F;break;
1889 default: 1890 default:
1890 PRINTK("%s: ERROR -- no AFC_CFG setting found", 1891 PRINTK(dev, "ERROR -- no AFC_CFG setting found");
1891 dev->name);
1892 break; 1892 break;
1893 } 1893 }
1894 1894
1895 DBG(SMC_DEBUG_MISC | SMC_DEBUG_TX | SMC_DEBUG_RX, 1895 DBG(SMC_DEBUG_MISC | SMC_DEBUG_TX | SMC_DEBUG_RX, dev,
1896 "%s: tx_fifo %d rx_fifo %d afc_cfg 0x%08x\n", CARDNAME, 1896 "%s: tx_fifo %d rx_fifo %d afc_cfg 0x%08x\n", CARDNAME,
1897 lp->tx_fifo_size, lp->rx_fifo_size, lp->afc_cfg); 1897 lp->tx_fifo_size, lp->rx_fifo_size, lp->afc_cfg);
1898 1898
1899 spin_lock_init(&lp->lock); 1899 spin_lock_init(&lp->lock);
1900 1900
@@ -1924,8 +1924,7 @@ static int smc911x_probe(struct net_device *dev)
1924 } 1924 }
1925 } 1925 }
1926 if (dev->irq == 0) { 1926 if (dev->irq == 0) {
1927 printk("%s: Couldn't autodetect your IRQ. Use irq=xx.\n", 1927 netdev_warn(dev, "Couldn't autodetect your IRQ. Use irq=xx.\n");
1928 dev->name);
1929 retval = -ENODEV; 1928 retval = -ENODEV;
1930 goto err_out; 1929 goto err_out;
1931 } 1930 }
@@ -1980,33 +1979,32 @@ static int smc911x_probe(struct net_device *dev)
1980 retval = register_netdev(dev); 1979 retval = register_netdev(dev);
1981 if (retval == 0) { 1980 if (retval == 0) {
1982 /* now, print out the card info, in a short format.. */ 1981 /* now, print out the card info, in a short format.. */
1983 printk("%s: %s (rev %d) at %#lx IRQ %d", 1982 netdev_info(dev, "%s (rev %d) at %#lx IRQ %d",
1984 dev->name, version_string, lp->revision, 1983 version_string, lp->revision,
1985 dev->base_addr, dev->irq); 1984 dev->base_addr, dev->irq);
1986 1985
1987#ifdef SMC_USE_DMA 1986#ifdef SMC_USE_DMA
1988 if (lp->rxdma != -1) 1987 if (lp->rxdma != -1)
1989 printk(" RXDMA %d ", lp->rxdma); 1988 pr_cont(" RXDMA %d", lp->rxdma);
1990 1989
1991 if (lp->txdma != -1) 1990 if (lp->txdma != -1)
1992 printk("TXDMA %d", lp->txdma); 1991 pr_cont(" TXDMA %d", lp->txdma);
1993#endif 1992#endif
1994 printk("\n"); 1993 pr_cont("\n");
1995 if (!is_valid_ether_addr(dev->dev_addr)) { 1994 if (!is_valid_ether_addr(dev->dev_addr)) {
1996 printk("%s: Invalid ethernet MAC address. Please " 1995 netdev_warn(dev, "Invalid ethernet MAC address. Please set using ifconfig\n");
1997 "set using ifconfig\n", dev->name);
1998 } else { 1996 } else {
1999 /* Print the Ethernet address */ 1997 /* Print the Ethernet address */
2000 printk("%s: Ethernet addr: %pM\n", 1998 netdev_info(dev, "Ethernet addr: %pM\n",
2001 dev->name, dev->dev_addr); 1999 dev->dev_addr);
2002 } 2000 }
2003 2001
2004 if (lp->phy_type == 0) { 2002 if (lp->phy_type == 0) {
2005 PRINTK("%s: No PHY found\n", dev->name); 2003 PRINTK(dev, "No PHY found\n");
2006 } else if ((lp->phy_type & ~0xff) == LAN911X_INTERNAL_PHY_ID) { 2004 } else if ((lp->phy_type & ~0xff) == LAN911X_INTERNAL_PHY_ID) {
2007 PRINTK("%s: LAN911x Internal PHY\n", dev->name); 2005 PRINTK(dev, "LAN911x Internal PHY\n");
2008 } else { 2006 } else {
2009 PRINTK("%s: External PHY 0x%08x\n", dev->name, lp->phy_type); 2007 PRINTK(dev, "External PHY 0x%08x\n", lp->phy_type);
2010 } 2008 }
2011 } 2009 }
2012 2010
@@ -2025,7 +2023,7 @@ err_out:
2025} 2023}
2026 2024
2027/* 2025/*
2028 * smc911x_init(void) 2026 * smc911x_drv_probe(void)
2029 * 2027 *
2030 * Output: 2028 * Output:
2031 * 0 --> there is a device 2029 * 0 --> there is a device
@@ -2039,6 +2037,7 @@ static int smc911x_drv_probe(struct platform_device *pdev)
2039 void __iomem *addr; 2037 void __iomem *addr;
2040 int ret; 2038 int ret;
2041 2039
2040 /* ndev is not valid yet, so avoid passing it in. */
2042 DBG(SMC_DEBUG_FUNC, "--> %s\n", __func__); 2041 DBG(SMC_DEBUG_FUNC, "--> %s\n", __func__);
2043 res = platform_get_resource(pdev, IORESOURCE_MEM, 0); 2042 res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
2044 if (!res) { 2043 if (!res) {
@@ -2093,7 +2092,7 @@ release_both:
2093release_1: 2092release_1:
2094 release_mem_region(res->start, SMC911X_IO_EXTENT); 2093 release_mem_region(res->start, SMC911X_IO_EXTENT);
2095out: 2094out:
2096 printk("%s: not found (%d).\n", CARDNAME, ret); 2095 pr_info("%s: not found (%d).\n", CARDNAME, ret);
2097 } 2096 }
2098#ifdef SMC_USE_DMA 2097#ifdef SMC_USE_DMA
2099 else { 2098 else {
@@ -2111,7 +2110,7 @@ static int smc911x_drv_remove(struct platform_device *pdev)
2111 struct smc911x_local *lp = netdev_priv(ndev); 2110 struct smc911x_local *lp = netdev_priv(ndev);
2112 struct resource *res; 2111 struct resource *res;
2113 2112
2114 DBG(SMC_DEBUG_FUNC, "--> %s\n", __func__); 2113 DBG(SMC_DEBUG_FUNC, ndev, "--> %s\n", __func__);
2115 2114
2116 unregister_netdev(ndev); 2115 unregister_netdev(ndev);
2117 2116
@@ -2140,7 +2139,7 @@ static int smc911x_drv_suspend(struct platform_device *dev, pm_message_t state)
2140 struct net_device *ndev = platform_get_drvdata(dev); 2139 struct net_device *ndev = platform_get_drvdata(dev);
2141 struct smc911x_local *lp = netdev_priv(ndev); 2140 struct smc911x_local *lp = netdev_priv(ndev);
2142 2141
2143 DBG(SMC_DEBUG_FUNC, "--> %s\n", __func__); 2142 DBG(SMC_DEBUG_FUNC, ndev, "--> %s\n", __func__);
2144 if (ndev) { 2143 if (ndev) {
2145 if (netif_running(ndev)) { 2144 if (netif_running(ndev)) {
2146 netif_device_detach(ndev); 2145 netif_device_detach(ndev);
@@ -2158,7 +2157,7 @@ static int smc911x_drv_resume(struct platform_device *dev)
2158{ 2157{
2159 struct net_device *ndev = platform_get_drvdata(dev); 2158 struct net_device *ndev = platform_get_drvdata(dev);
2160 2159
2161 DBG(SMC_DEBUG_FUNC, "--> %s\n", __func__); 2160 DBG(SMC_DEBUG_FUNC, ndev, "--> %s\n", __func__);
2162 if (ndev) { 2161 if (ndev) {
2163 struct smc911x_local *lp = netdev_priv(ndev); 2162 struct smc911x_local *lp = netdev_priv(ndev);
2164 2163
diff --git a/drivers/net/ethernet/smsc/smc911x.h b/drivers/net/ethernet/smsc/smc911x.h
index d51261ba4642..9965da39281b 100644
--- a/drivers/net/ethernet/smsc/smc911x.h
+++ b/drivers/net/ethernet/smsc/smc911x.h
@@ -227,7 +227,7 @@ static inline void SMC_outsl(struct smc911x_local *lp, int reg,
227#define SMC_DMA_ACK_IRQ(dev, dma) \ 227#define SMC_DMA_ACK_IRQ(dev, dma) \
228{ \ 228{ \
229 if (DCSR(dma) & DCSR_BUSERR) { \ 229 if (DCSR(dma) & DCSR_BUSERR) { \
230 printk("%s: DMA %d bus error!\n", dev->name, dma); \ 230 netdev_err(dev, "DMA %d bus error!\n", dma); \
231 } \ 231 } \
232 DCSR(dma) = DCSR_STARTINTR|DCSR_ENDINTR|DCSR_BUSERR; \ 232 DCSR(dma) = DCSR_STARTINTR|DCSR_ENDINTR|DCSR_BUSERR; \
233} 233}
diff --git a/drivers/net/ethernet/smsc/smc9194.c b/drivers/net/ethernet/smsc/smc9194.c
index afd9873e9bdb..67d9fdeedd86 100644
--- a/drivers/net/ethernet/smsc/smc9194.c
+++ b/drivers/net/ethernet/smsc/smc9194.c
@@ -55,7 +55,7 @@
55 ----------------------------------------------------------------------------*/ 55 ----------------------------------------------------------------------------*/
56 56
57static const char version[] = 57static const char version[] =
58 "smc9194.c:v0.14 12/15/00 by Erik Stahlman (erik@vt.edu)\n"; 58 "smc9194.c:v0.14 12/15/00 by Erik Stahlman (erik@vt.edu)";
59 59
60#include <linux/module.h> 60#include <linux/module.h>
61#include <linux/kernel.h> 61#include <linux/kernel.h>
@@ -597,7 +597,7 @@ static void smc_hardware_send_packet( struct net_device * dev )
597 packet_no = inb( ioaddr + PNR_ARR + 1 ); 597 packet_no = inb( ioaddr + PNR_ARR + 1 );
598 if ( packet_no & 0x80 ) { 598 if ( packet_no & 0x80 ) {
599 /* or isn't there? BAD CHIP! */ 599 /* or isn't there? BAD CHIP! */
600 printk(KERN_DEBUG CARDNAME": Memory allocation failed.\n"); 600 netdev_dbg(dev, CARDNAME": Memory allocation failed.\n");
601 dev_kfree_skb_any(skb); 601 dev_kfree_skb_any(skb);
602 lp->saved_skb = NULL; 602 lp->saved_skb = NULL;
603 netif_wake_queue(dev); 603 netif_wake_queue(dev);
@@ -610,7 +610,7 @@ static void smc_hardware_send_packet( struct net_device * dev )
610 /* point to the beginning of the packet */ 610 /* point to the beginning of the packet */
611 outw( PTR_AUTOINC , ioaddr + POINTER ); 611 outw( PTR_AUTOINC , ioaddr + POINTER );
612 612
613 PRINTK3((CARDNAME": Trying to xmit packet of length %x\n", length )); 613 PRINTK3((CARDNAME": Trying to xmit packet of length %x\n", length));
614#if SMC_DEBUG > 2 614#if SMC_DEBUG > 2
615 print_packet( buf, length ); 615 print_packet( buf, length );
616#endif 616#endif
@@ -846,7 +846,6 @@ static const struct net_device_ops smc_netdev_ops = {
846static int __init smc_probe(struct net_device *dev, int ioaddr) 846static int __init smc_probe(struct net_device *dev, int ioaddr)
847{ 847{
848 int i, memory, retval; 848 int i, memory, retval;
849 static unsigned version_printed;
850 unsigned int bank; 849 unsigned int bank;
851 850
852 const char *version_string; 851 const char *version_string;
@@ -913,8 +912,7 @@ static int __init smc_probe(struct net_device *dev, int ioaddr)
913 It might be prudent to check a listing of MAC addresses 912 It might be prudent to check a listing of MAC addresses
914 against the hardware address, or do some other tests. */ 913 against the hardware address, or do some other tests. */
915 914
916 if (version_printed++ == 0) 915 pr_info_once("%s\n", version);
917 printk("%s", version);
918 916
919 /* fill in some of the fields */ 917 /* fill in some of the fields */
920 dev->base_addr = ioaddr; 918 dev->base_addr = ioaddr;
@@ -1003,21 +1001,21 @@ static int __init smc_probe(struct net_device *dev, int ioaddr)
1003 1001
1004 /* now, print out the card info, in a short format.. */ 1002 /* now, print out the card info, in a short format.. */
1005 1003
1006 printk("%s: %s(r:%d) at %#3x IRQ:%d INTF:%s MEM:%db ", dev->name, 1004 netdev_info(dev, "%s(r:%d) at %#3x IRQ:%d INTF:%s MEM:%db ",
1007 version_string, revision_register & 0xF, ioaddr, dev->irq, 1005 version_string, revision_register & 0xF, ioaddr, dev->irq,
1008 if_string, memory ); 1006 if_string, memory);
1009 /* 1007 /*
1010 . Print the Ethernet address 1008 . Print the Ethernet address
1011 */ 1009 */
1012 printk("ADDR: %pM\n", dev->dev_addr); 1010 netdev_info(dev, "ADDR: %pM\n", dev->dev_addr);
1013 1011
1014 /* Grab the IRQ */ 1012 /* Grab the IRQ */
1015 retval = request_irq(dev->irq, smc_interrupt, 0, DRV_NAME, dev); 1013 retval = request_irq(dev->irq, smc_interrupt, 0, DRV_NAME, dev);
1016 if (retval) { 1014 if (retval) {
1017 printk("%s: unable to get IRQ %d (irqval=%d).\n", DRV_NAME, 1015 netdev_warn(dev, "%s: unable to get IRQ %d (irqval=%d).\n",
1018 dev->irq, retval); 1016 DRV_NAME, dev->irq, retval);
1019 goto err_out; 1017 goto err_out;
1020 } 1018 }
1021 1019
1022 dev->netdev_ops = &smc_netdev_ops; 1020 dev->netdev_ops = &smc_netdev_ops;
1023 dev->watchdog_timeo = HZ/20; 1021 dev->watchdog_timeo = HZ/20;
@@ -1037,30 +1035,32 @@ static void print_packet( byte * buf, int length )
1037 int remainder; 1035 int remainder;
1038 int lines; 1036 int lines;
1039 1037
1040 printk("Packet of length %d\n", length); 1038 pr_dbg("Packet of length %d\n", length);
1041 lines = length / 16; 1039 lines = length / 16;
1042 remainder = length % 16; 1040 remainder = length % 16;
1043 1041
1044 for ( i = 0; i < lines ; i ++ ) { 1042 for ( i = 0; i < lines ; i ++ ) {
1045 int cur; 1043 int cur;
1046 1044
1045 printk(KERN_DEBUG);
1047 for ( cur = 0; cur < 8; cur ++ ) { 1046 for ( cur = 0; cur < 8; cur ++ ) {
1048 byte a, b; 1047 byte a, b;
1049 1048
1050 a = *(buf ++ ); 1049 a = *(buf ++ );
1051 b = *(buf ++ ); 1050 b = *(buf ++ );
1052 printk("%02x%02x ", a, b ); 1051 pr_cont("%02x%02x ", a, b);
1053 } 1052 }
1054 printk("\n"); 1053 pr_cont("\n");
1055 } 1054 }
1055 printk(KERN_DEBUG);
1056 for ( i = 0; i < remainder/2 ; i++ ) { 1056 for ( i = 0; i < remainder/2 ; i++ ) {
1057 byte a, b; 1057 byte a, b;
1058 1058
1059 a = *(buf ++ ); 1059 a = *(buf ++ );
1060 b = *(buf ++ ); 1060 b = *(buf ++ );
1061 printk("%02x%02x ", a, b ); 1061 pr_cont("%02x%02x ", a, b);
1062 } 1062 }
1063 printk("\n"); 1063 pr_cont("\n");
1064#endif 1064#endif
1065} 1065}
1066#endif 1066#endif
@@ -1127,9 +1127,8 @@ static void smc_timeout(struct net_device *dev)
1127{ 1127{
1128 /* If we get here, some higher level has decided we are broken. 1128 /* If we get here, some higher level has decided we are broken.
1129 There should really be a "kick me" function call instead. */ 1129 There should really be a "kick me" function call instead. */
1130 printk(KERN_WARNING CARDNAME": transmit timed out, %s?\n", 1130 netdev_warn(dev, CARDNAME": transmit timed out, %s?\n",
1131 tx_done(dev) ? "IRQ conflict" : 1131 tx_done(dev) ? "IRQ conflict" : "network cable problem");
1132 "network cable problem");
1133 /* "kick" the adaptor */ 1132 /* "kick" the adaptor */
1134 smc_reset( dev->base_addr ); 1133 smc_reset( dev->base_addr );
1135 smc_enable( dev->base_addr ); 1134 smc_enable( dev->base_addr );
@@ -1299,8 +1298,7 @@ static void smc_tx( struct net_device * dev )
1299 dev->stats.tx_errors++; 1298 dev->stats.tx_errors++;
1300 if ( tx_status & TS_LOSTCAR ) dev->stats.tx_carrier_errors++; 1299 if ( tx_status & TS_LOSTCAR ) dev->stats.tx_carrier_errors++;
1301 if ( tx_status & TS_LATCOL ) { 1300 if ( tx_status & TS_LATCOL ) {
1302 printk(KERN_DEBUG CARDNAME 1301 netdev_dbg(dev, CARDNAME": Late collision occurred on last xmit.\n");
1303 ": Late collision occurred on last xmit.\n");
1304 dev->stats.tx_window_errors++; 1302 dev->stats.tx_window_errors++;
1305 } 1303 }
1306#if 0 1304#if 0
@@ -1308,7 +1306,7 @@ static void smc_tx( struct net_device * dev )
1308#endif 1306#endif
1309 1307
1310 if ( tx_status & TS_SUCCESS ) { 1308 if ( tx_status & TS_SUCCESS ) {
1311 printk(CARDNAME": Successful packet caused interrupt\n"); 1309 netdev_info(dev, CARDNAME": Successful packet caused interrupt\n");
1312 } 1310 }
1313 /* re-enable transmit */ 1311 /* re-enable transmit */
1314 SMC_SELECT_BANK( 0 ); 1312 SMC_SELECT_BANK( 0 );
@@ -1547,9 +1545,7 @@ int __init init_module(void)
1547 1545
1548 /* copy the parameters from insmod into the device structure */ 1546 /* copy the parameters from insmod into the device structure */
1549 devSMC9194 = smc_init(-1); 1547 devSMC9194 = smc_init(-1);
1550 if (IS_ERR(devSMC9194)) 1548 return PTR_ERR_OR_ZERO(devSMC9194);
1551 return PTR_ERR(devSMC9194);
1552 return 0;
1553} 1549}
1554 1550
1555void __exit cleanup_module(void) 1551void __exit cleanup_module(void)
diff --git a/drivers/net/ethernet/smsc/smc91c92_cs.c b/drivers/net/ethernet/smsc/smc91c92_cs.c
index 656d2e2ebfc9..8ef70d9c20c1 100644
--- a/drivers/net/ethernet/smsc/smc91c92_cs.c
+++ b/drivers/net/ethernet/smsc/smc91c92_cs.c
@@ -740,7 +740,7 @@ static int smc91c92_resume(struct pcmcia_device *link)
740 (smc->cardid == PRODID_PSION_NET100))) { 740 (smc->cardid == PRODID_PSION_NET100))) {
741 i = osi_load_firmware(link); 741 i = osi_load_firmware(link);
742 if (i) { 742 if (i) {
743 pr_err("smc91c92_cs: Failed to load firmware\n"); 743 netdev_err(dev, "Failed to load firmware\n");
744 return i; 744 return i;
745 } 745 }
746 } 746 }
@@ -793,7 +793,7 @@ static int check_sig(struct pcmcia_device *link)
793 } 793 }
794 794
795 if (width) { 795 if (width) {
796 pr_info("using 8-bit IO window\n"); 796 netdev_info(dev, "using 8-bit IO window\n");
797 797
798 smc91c92_suspend(link); 798 smc91c92_suspend(link);
799 pcmcia_fixup_iowidth(link); 799 pcmcia_fixup_iowidth(link);
@@ -1036,7 +1036,7 @@ static void smc_dump(struct net_device *dev)
1036 save = inw(ioaddr + BANK_SELECT); 1036 save = inw(ioaddr + BANK_SELECT);
1037 for (w = 0; w < 4; w++) { 1037 for (w = 0; w < 4; w++) {
1038 SMC_SELECT_BANK(w); 1038 SMC_SELECT_BANK(w);
1039 netdev_printk(KERN_DEBUG, dev, "bank %d: ", w); 1039 netdev_dbg(dev, "bank %d: ", w);
1040 for (i = 0; i < 14; i += 2) 1040 for (i = 0; i < 14; i += 2)
1041 pr_cont(" %04x", inw(ioaddr + i)); 1041 pr_cont(" %04x", inw(ioaddr + i));
1042 pr_cont("\n"); 1042 pr_cont("\n");
@@ -1213,8 +1213,7 @@ static netdev_tx_t smc_start_xmit(struct sk_buff *skb,
1213 if (smc->saved_skb) { 1213 if (smc->saved_skb) {
1214 /* THIS SHOULD NEVER HAPPEN. */ 1214 /* THIS SHOULD NEVER HAPPEN. */
1215 dev->stats.tx_aborted_errors++; 1215 dev->stats.tx_aborted_errors++;
1216 netdev_printk(KERN_DEBUG, dev, 1216 netdev_dbg(dev, "Internal error -- sent packet while busy\n");
1217 "Internal error -- sent packet while busy\n");
1218 return NETDEV_TX_BUSY; 1217 return NETDEV_TX_BUSY;
1219 } 1218 }
1220 smc->saved_skb = skb; 1219 smc->saved_skb = skb;
@@ -1254,7 +1253,7 @@ static netdev_tx_t smc_start_xmit(struct sk_buff *skb,
1254 } 1253 }
1255 1254
1256 /* Otherwise defer until the Tx-space-allocated interrupt. */ 1255 /* Otherwise defer until the Tx-space-allocated interrupt. */
1257 pr_debug("%s: memory allocation deferred.\n", dev->name); 1256 netdev_dbg(dev, "memory allocation deferred.\n");
1258 outw((IM_ALLOC_INT << 8) | (ir & 0xff00), ioaddr + INTERRUPT); 1257 outw((IM_ALLOC_INT << 8) | (ir & 0xff00), ioaddr + INTERRUPT);
1259 spin_unlock_irqrestore(&smc->lock, flags); 1258 spin_unlock_irqrestore(&smc->lock, flags);
1260 1259
@@ -1317,8 +1316,8 @@ static void smc_eph_irq(struct net_device *dev)
1317 1316
1318 SMC_SELECT_BANK(0); 1317 SMC_SELECT_BANK(0);
1319 ephs = inw(ioaddr + EPH); 1318 ephs = inw(ioaddr + EPH);
1320 pr_debug("%s: Ethernet protocol handler interrupt, status" 1319 netdev_dbg(dev, "Ethernet protocol handler interrupt, status %4.4x.\n",
1321 " %4.4x.\n", dev->name, ephs); 1320 ephs);
1322 /* Could be a counter roll-over warning: update stats. */ 1321 /* Could be a counter roll-over warning: update stats. */
1323 card_stats = inw(ioaddr + COUNTER); 1322 card_stats = inw(ioaddr + COUNTER);
1324 /* single collisions */ 1323 /* single collisions */
@@ -1357,8 +1356,8 @@ static irqreturn_t smc_interrupt(int irq, void *dev_id)
1357 1356
1358 ioaddr = dev->base_addr; 1357 ioaddr = dev->base_addr;
1359 1358
1360 pr_debug("%s: SMC91c92 interrupt %d at %#x.\n", dev->name, 1359 netdev_dbg(dev, "SMC91c92 interrupt %d at %#x.\n",
1361 irq, ioaddr); 1360 irq, ioaddr);
1362 1361
1363 spin_lock(&smc->lock); 1362 spin_lock(&smc->lock);
1364 smc->watchdog = 0; 1363 smc->watchdog = 0;
@@ -1366,8 +1365,8 @@ static irqreturn_t smc_interrupt(int irq, void *dev_id)
1366 if ((saved_bank & 0xff00) != 0x3300) { 1365 if ((saved_bank & 0xff00) != 0x3300) {
1367 /* The device does not exist -- the card could be off-line, or 1366 /* The device does not exist -- the card could be off-line, or
1368 maybe it has been ejected. */ 1367 maybe it has been ejected. */
1369 pr_debug("%s: SMC91c92 interrupt %d for non-existent" 1368 netdev_dbg(dev, "SMC91c92 interrupt %d for non-existent/ejected device.\n",
1370 "/ejected device.\n", dev->name, irq); 1369 irq);
1371 handled = 0; 1370 handled = 0;
1372 goto irq_done; 1371 goto irq_done;
1373 } 1372 }
@@ -1380,8 +1379,8 @@ static irqreturn_t smc_interrupt(int irq, void *dev_id)
1380 1379
1381 do { /* read the status flag, and mask it */ 1380 do { /* read the status flag, and mask it */
1382 status = inw(ioaddr + INTERRUPT) & 0xff; 1381 status = inw(ioaddr + INTERRUPT) & 0xff;
1383 pr_debug("%s: Status is %#2.2x (mask %#2.2x).\n", dev->name, 1382 netdev_dbg(dev, "Status is %#2.2x (mask %#2.2x).\n",
1384 status, mask); 1383 status, mask);
1385 if ((status & mask) == 0) { 1384 if ((status & mask) == 0) {
1386 if (bogus_cnt == INTR_WORK) 1385 if (bogus_cnt == INTR_WORK)
1387 handled = 0; 1386 handled = 0;
@@ -1425,15 +1424,15 @@ static irqreturn_t smc_interrupt(int irq, void *dev_id)
1425 smc_eph_irq(dev); 1424 smc_eph_irq(dev);
1426 } while (--bogus_cnt); 1425 } while (--bogus_cnt);
1427 1426
1428 pr_debug(" Restoring saved registers mask %2.2x bank %4.4x" 1427 netdev_dbg(dev, " Restoring saved registers mask %2.2x bank %4.4x pointer %4.4x.\n",
1429 " pointer %4.4x.\n", mask, saved_bank, saved_pointer); 1428 mask, saved_bank, saved_pointer);
1430 1429
1431 /* restore state register */ 1430 /* restore state register */
1432 outw((mask<<8), ioaddr + INTERRUPT); 1431 outw((mask<<8), ioaddr + INTERRUPT);
1433 outw(saved_pointer, ioaddr + POINTER); 1432 outw(saved_pointer, ioaddr + POINTER);
1434 SMC_SELECT_BANK(saved_bank); 1433 SMC_SELECT_BANK(saved_bank);
1435 1434
1436 pr_debug("%s: Exiting interrupt IRQ%d.\n", dev->name, irq); 1435 netdev_dbg(dev, "Exiting interrupt IRQ%d.\n", irq);
1437 1436
1438irq_done: 1437irq_done:
1439 1438
@@ -1491,10 +1490,10 @@ static void smc_rx(struct net_device *dev)
1491 rx_status = inw(ioaddr + DATA_1); 1490 rx_status = inw(ioaddr + DATA_1);
1492 packet_length = inw(ioaddr + DATA_1) & 0x07ff; 1491 packet_length = inw(ioaddr + DATA_1) & 0x07ff;
1493 1492
1494 pr_debug("%s: Receive status %4.4x length %d.\n", 1493 netdev_dbg(dev, "Receive status %4.4x length %d.\n",
1495 dev->name, rx_status, packet_length); 1494 rx_status, packet_length);
1496 1495
1497 if (!(rx_status & RS_ERRORS)) { 1496 if (!(rx_status & RS_ERRORS)) {
1498 /* do stuff to make a new packet */ 1497 /* do stuff to make a new packet */
1499 struct sk_buff *skb; 1498 struct sk_buff *skb;
1500 1499
@@ -1502,7 +1501,7 @@ static void smc_rx(struct net_device *dev)
1502 skb = netdev_alloc_skb(dev, packet_length+2); 1501 skb = netdev_alloc_skb(dev, packet_length+2);
1503 1502
1504 if (skb == NULL) { 1503 if (skb == NULL) {
1505 pr_debug("%s: Low memory, packet dropped.\n", dev->name); 1504 netdev_dbg(dev, "Low memory, packet dropped.\n");
1506 dev->stats.rx_dropped++; 1505 dev->stats.rx_dropped++;
1507 outw(MC_RELEASE, ioaddr + MMU_CMD); 1506 outw(MC_RELEASE, ioaddr + MMU_CMD);
1508 return; 1507 return;
@@ -1643,7 +1642,7 @@ static void smc_reset(struct net_device *dev)
1643 struct smc_private *smc = netdev_priv(dev); 1642 struct smc_private *smc = netdev_priv(dev);
1644 int i; 1643 int i;
1645 1644
1646 pr_debug("%s: smc91c92 reset called.\n", dev->name); 1645 netdev_dbg(dev, "smc91c92 reset called.\n");
1647 1646
1648 /* The first interaction must be a write to bring the chip out 1647 /* The first interaction must be a write to bring the chip out
1649 of sleep mode. */ 1648 of sleep mode. */
diff --git a/drivers/net/ethernet/smsc/smc91x.c b/drivers/net/ethernet/smsc/smc91x.c
index 73be7f3982e6..0c9b5d94154f 100644
--- a/drivers/net/ethernet/smsc/smc91x.c
+++ b/drivers/net/ethernet/smsc/smc91x.c
@@ -58,7 +58,7 @@
58 * 22/09/04 Nicolas Pitre big update (see commit log for details) 58 * 22/09/04 Nicolas Pitre big update (see commit log for details)
59 */ 59 */
60static const char version[] = 60static const char version[] =
61 "smc91x.c: v1.1, sep 22 2004 by Nicolas Pitre <nico@fluxnic.net>\n"; 61 "smc91x.c: v1.1, sep 22 2004 by Nicolas Pitre <nico@fluxnic.net>";
62 62
63/* Debugging level */ 63/* Debugging level */
64#ifndef SMC_DEBUG 64#ifndef SMC_DEBUG
@@ -149,16 +149,16 @@ MODULE_ALIAS("platform:smc91x");
149#define MII_DELAY 1 149#define MII_DELAY 1
150 150
151#if SMC_DEBUG > 0 151#if SMC_DEBUG > 0
152#define DBG(n, args...) \ 152#define DBG(n, dev, args...) \
153 do { \ 153 do { \
154 if (SMC_DEBUG >= (n)) \ 154 if (SMC_DEBUG >= (n)) \
155 printk(args); \ 155 netdev_dbg(dev, args); \
156 } while (0) 156 } while (0)
157 157
158#define PRINTK(args...) printk(args) 158#define PRINTK(dev, args...) netdev_info(dev, args)
159#else 159#else
160#define DBG(n, args...) do { } while(0) 160#define DBG(n, dev, args...) do { } while (0)
161#define PRINTK(args...) printk(KERN_DEBUG args) 161#define PRINTK(dev, args...) netdev_dbg(dev, args)
162#endif 162#endif
163 163
164#if SMC_DEBUG > 3 164#if SMC_DEBUG > 3
@@ -173,24 +173,26 @@ static void PRINT_PKT(u_char *buf, int length)
173 173
174 for (i = 0; i < lines ; i ++) { 174 for (i = 0; i < lines ; i ++) {
175 int cur; 175 int cur;
176 printk(KERN_DEBUG);
176 for (cur = 0; cur < 8; cur++) { 177 for (cur = 0; cur < 8; cur++) {
177 u_char a, b; 178 u_char a, b;
178 a = *buf++; 179 a = *buf++;
179 b = *buf++; 180 b = *buf++;
180 printk("%02x%02x ", a, b); 181 pr_cont("%02x%02x ", a, b);
181 } 182 }
182 printk("\n"); 183 pr_cont("\n");
183 } 184 }
185 printk(KERN_DEBUG);
184 for (i = 0; i < remainder/2 ; i++) { 186 for (i = 0; i < remainder/2 ; i++) {
185 u_char a, b; 187 u_char a, b;
186 a = *buf++; 188 a = *buf++;
187 b = *buf++; 189 b = *buf++;
188 printk("%02x%02x ", a, b); 190 pr_cont("%02x%02x ", a, b);
189 } 191 }
190 printk("\n"); 192 pr_cont("\n");
191} 193}
192#else 194#else
193#define PRINT_PKT(x...) do { } while(0) 195#define PRINT_PKT(x...) do { } while (0)
194#endif 196#endif
195 197
196 198
@@ -226,8 +228,8 @@ static void PRINT_PKT(u_char *buf, int length)
226 unsigned long timeout = jiffies + 2; \ 228 unsigned long timeout = jiffies + 2; \
227 while (SMC_GET_MMU_CMD(lp) & MC_BUSY) { \ 229 while (SMC_GET_MMU_CMD(lp) & MC_BUSY) { \
228 if (time_after(jiffies, timeout)) { \ 230 if (time_after(jiffies, timeout)) { \
229 printk("%s: timeout %s line %d\n", \ 231 netdev_dbg(dev, "timeout %s line %d\n", \
230 dev->name, __FILE__, __LINE__); \ 232 __FILE__, __LINE__); \
231 break; \ 233 break; \
232 } \ 234 } \
233 cpu_relax(); \ 235 cpu_relax(); \
@@ -246,7 +248,7 @@ static void smc_reset(struct net_device *dev)
246 unsigned int ctl, cfg; 248 unsigned int ctl, cfg;
247 struct sk_buff *pending_skb; 249 struct sk_buff *pending_skb;
248 250
249 DBG(2, "%s: %s\n", dev->name, __func__); 251 DBG(2, dev, "%s\n", __func__);
250 252
251 /* Disable all interrupts, block TX tasklet */ 253 /* Disable all interrupts, block TX tasklet */
252 spin_lock_irq(&lp->lock); 254 spin_lock_irq(&lp->lock);
@@ -339,7 +341,7 @@ static void smc_enable(struct net_device *dev)
339 void __iomem *ioaddr = lp->base; 341 void __iomem *ioaddr = lp->base;
340 int mask; 342 int mask;
341 343
342 DBG(2, "%s: %s\n", dev->name, __func__); 344 DBG(2, dev, "%s\n", __func__);
343 345
344 /* see the header file for options in TCR/RCR DEFAULT */ 346 /* see the header file for options in TCR/RCR DEFAULT */
345 SMC_SELECT_BANK(lp, 0); 347 SMC_SELECT_BANK(lp, 0);
@@ -373,7 +375,7 @@ static void smc_shutdown(struct net_device *dev)
373 void __iomem *ioaddr = lp->base; 375 void __iomem *ioaddr = lp->base;
374 struct sk_buff *pending_skb; 376 struct sk_buff *pending_skb;
375 377
376 DBG(2, "%s: %s\n", CARDNAME, __func__); 378 DBG(2, dev, "%s: %s\n", CARDNAME, __func__);
377 379
378 /* no more interrupts for me */ 380 /* no more interrupts for me */
379 spin_lock_irq(&lp->lock); 381 spin_lock_irq(&lp->lock);
@@ -406,11 +408,11 @@ static inline void smc_rcv(struct net_device *dev)
406 void __iomem *ioaddr = lp->base; 408 void __iomem *ioaddr = lp->base;
407 unsigned int packet_number, status, packet_len; 409 unsigned int packet_number, status, packet_len;
408 410
409 DBG(3, "%s: %s\n", dev->name, __func__); 411 DBG(3, dev, "%s\n", __func__);
410 412
411 packet_number = SMC_GET_RXFIFO(lp); 413 packet_number = SMC_GET_RXFIFO(lp);
412 if (unlikely(packet_number & RXFIFO_REMPTY)) { 414 if (unlikely(packet_number & RXFIFO_REMPTY)) {
413 PRINTK("%s: smc_rcv with nothing on FIFO.\n", dev->name); 415 PRINTK(dev, "smc_rcv with nothing on FIFO.\n");
414 return; 416 return;
415 } 417 }
416 418
@@ -420,9 +422,8 @@ static inline void smc_rcv(struct net_device *dev)
420 /* First two words are status and packet length */ 422 /* First two words are status and packet length */
421 SMC_GET_PKT_HDR(lp, status, packet_len); 423 SMC_GET_PKT_HDR(lp, status, packet_len);
422 packet_len &= 0x07ff; /* mask off top bits */ 424 packet_len &= 0x07ff; /* mask off top bits */
423 DBG(2, "%s: RX PNR 0x%x STATUS 0x%04x LENGTH 0x%04x (%d)\n", 425 DBG(2, dev, "RX PNR 0x%x STATUS 0x%04x LENGTH 0x%04x (%d)\n",
424 dev->name, packet_number, status, 426 packet_number, status, packet_len, packet_len);
425 packet_len, packet_len);
426 427
427 back: 428 back:
428 if (unlikely(packet_len < 6 || status & RS_ERRORS)) { 429 if (unlikely(packet_len < 6 || status & RS_ERRORS)) {
@@ -433,8 +434,8 @@ static inline void smc_rcv(struct net_device *dev)
433 } 434 }
434 if (packet_len < 6) { 435 if (packet_len < 6) {
435 /* bloody hardware */ 436 /* bloody hardware */
436 printk(KERN_ERR "%s: fubar (rxlen %u status %x\n", 437 netdev_err(dev, "fubar (rxlen %u status %x\n",
437 dev->name, packet_len, status); 438 packet_len, status);
438 status |= RS_TOOSHORT; 439 status |= RS_TOOSHORT;
439 } 440 }
440 SMC_WAIT_MMU_BUSY(lp); 441 SMC_WAIT_MMU_BUSY(lp);
@@ -551,7 +552,7 @@ static void smc_hardware_send_pkt(unsigned long data)
551 unsigned char *buf; 552 unsigned char *buf;
552 unsigned long flags; 553 unsigned long flags;
553 554
554 DBG(3, "%s: %s\n", dev->name, __func__); 555 DBG(3, dev, "%s\n", __func__);
555 556
556 if (!smc_special_trylock(&lp->lock, flags)) { 557 if (!smc_special_trylock(&lp->lock, flags)) {
557 netif_stop_queue(dev); 558 netif_stop_queue(dev);
@@ -568,7 +569,7 @@ static void smc_hardware_send_pkt(unsigned long data)
568 569
569 packet_no = SMC_GET_AR(lp); 570 packet_no = SMC_GET_AR(lp);
570 if (unlikely(packet_no & AR_FAILED)) { 571 if (unlikely(packet_no & AR_FAILED)) {
571 printk("%s: Memory allocation failed.\n", dev->name); 572 netdev_err(dev, "Memory allocation failed.\n");
572 dev->stats.tx_errors++; 573 dev->stats.tx_errors++;
573 dev->stats.tx_fifo_errors++; 574 dev->stats.tx_fifo_errors++;
574 smc_special_unlock(&lp->lock, flags); 575 smc_special_unlock(&lp->lock, flags);
@@ -581,8 +582,8 @@ static void smc_hardware_send_pkt(unsigned long data)
581 582
582 buf = skb->data; 583 buf = skb->data;
583 len = skb->len; 584 len = skb->len;
584 DBG(2, "%s: TX PNR 0x%x LENGTH 0x%04x (%d) BUF 0x%p\n", 585 DBG(2, dev, "TX PNR 0x%x LENGTH 0x%04x (%d) BUF 0x%p\n",
585 dev->name, packet_no, len, len, buf); 586 packet_no, len, len, buf);
586 PRINT_PKT(buf, len); 587 PRINT_PKT(buf, len);
587 588
588 /* 589 /*
@@ -637,7 +638,7 @@ static int smc_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
637 unsigned int numPages, poll_count, status; 638 unsigned int numPages, poll_count, status;
638 unsigned long flags; 639 unsigned long flags;
639 640
640 DBG(3, "%s: %s\n", dev->name, __func__); 641 DBG(3, dev, "%s\n", __func__);
641 642
642 BUG_ON(lp->pending_tx_skb != NULL); 643 BUG_ON(lp->pending_tx_skb != NULL);
643 644
@@ -654,7 +655,7 @@ static int smc_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
654 */ 655 */
655 numPages = ((skb->len & ~1) + (6 - 1)) >> 8; 656 numPages = ((skb->len & ~1) + (6 - 1)) >> 8;
656 if (unlikely(numPages > 7)) { 657 if (unlikely(numPages > 7)) {
657 printk("%s: Far too big packet error.\n", dev->name); 658 netdev_warn(dev, "Far too big packet error.\n");
658 dev->stats.tx_errors++; 659 dev->stats.tx_errors++;
659 dev->stats.tx_dropped++; 660 dev->stats.tx_dropped++;
660 dev_kfree_skb(skb); 661 dev_kfree_skb(skb);
@@ -685,7 +686,7 @@ static int smc_hard_start_xmit(struct sk_buff *skb, struct net_device *dev)
685 if (!poll_count) { 686 if (!poll_count) {
686 /* oh well, wait until the chip finds memory later */ 687 /* oh well, wait until the chip finds memory later */
687 netif_stop_queue(dev); 688 netif_stop_queue(dev);
688 DBG(2, "%s: TX memory allocation deferred.\n", dev->name); 689 DBG(2, dev, "TX memory allocation deferred.\n");
689 SMC_ENABLE_INT(lp, IM_ALLOC_INT); 690 SMC_ENABLE_INT(lp, IM_ALLOC_INT);
690 } else { 691 } else {
691 /* 692 /*
@@ -709,12 +710,12 @@ static void smc_tx(struct net_device *dev)
709 void __iomem *ioaddr = lp->base; 710 void __iomem *ioaddr = lp->base;
710 unsigned int saved_packet, packet_no, tx_status, pkt_len; 711 unsigned int saved_packet, packet_no, tx_status, pkt_len;
711 712
712 DBG(3, "%s: %s\n", dev->name, __func__); 713 DBG(3, dev, "%s\n", __func__);
713 714
714 /* If the TX FIFO is empty then nothing to do */ 715 /* If the TX FIFO is empty then nothing to do */
715 packet_no = SMC_GET_TXFIFO(lp); 716 packet_no = SMC_GET_TXFIFO(lp);
716 if (unlikely(packet_no & TXFIFO_TEMPTY)) { 717 if (unlikely(packet_no & TXFIFO_TEMPTY)) {
717 PRINTK("%s: smc_tx with nothing on FIFO.\n", dev->name); 718 PRINTK(dev, "smc_tx with nothing on FIFO.\n");
718 return; 719 return;
719 } 720 }
720 721
@@ -725,8 +726,8 @@ static void smc_tx(struct net_device *dev)
725 /* read the first word (status word) from this packet */ 726 /* read the first word (status word) from this packet */
726 SMC_SET_PTR(lp, PTR_AUTOINC | PTR_READ); 727 SMC_SET_PTR(lp, PTR_AUTOINC | PTR_READ);
727 SMC_GET_PKT_HDR(lp, tx_status, pkt_len); 728 SMC_GET_PKT_HDR(lp, tx_status, pkt_len);
728 DBG(2, "%s: TX STATUS 0x%04x PNR 0x%02x\n", 729 DBG(2, dev, "TX STATUS 0x%04x PNR 0x%02x\n",
729 dev->name, tx_status, packet_no); 730 tx_status, packet_no);
730 731
731 if (!(tx_status & ES_TX_SUC)) 732 if (!(tx_status & ES_TX_SUC))
732 dev->stats.tx_errors++; 733 dev->stats.tx_errors++;
@@ -735,14 +736,12 @@ static void smc_tx(struct net_device *dev)
735 dev->stats.tx_carrier_errors++; 736 dev->stats.tx_carrier_errors++;
736 737
737 if (tx_status & (ES_LATCOL | ES_16COL)) { 738 if (tx_status & (ES_LATCOL | ES_16COL)) {
738 PRINTK("%s: %s occurred on last xmit\n", dev->name, 739 PRINTK(dev, "%s occurred on last xmit\n",
739 (tx_status & ES_LATCOL) ? 740 (tx_status & ES_LATCOL) ?
740 "late collision" : "too many collisions"); 741 "late collision" : "too many collisions");
741 dev->stats.tx_window_errors++; 742 dev->stats.tx_window_errors++;
742 if (!(dev->stats.tx_window_errors & 63) && net_ratelimit()) { 743 if (!(dev->stats.tx_window_errors & 63) && net_ratelimit()) {
743 printk(KERN_INFO "%s: unexpectedly large number of " 744 netdev_info(dev, "unexpectedly large number of bad collisions. Please check duplex setting.\n");
744 "bad collisions. Please check duplex "
745 "setting.\n", dev->name);
746 } 745 }
747 } 746 }
748 747
@@ -830,8 +829,8 @@ static int smc_phy_read(struct net_device *dev, int phyaddr, int phyreg)
830 /* Return to idle state */ 829 /* Return to idle state */
831 SMC_SET_MII(lp, SMC_GET_MII(lp) & ~(MII_MCLK|MII_MDOE|MII_MDO)); 830 SMC_SET_MII(lp, SMC_GET_MII(lp) & ~(MII_MCLK|MII_MDOE|MII_MDO));
832 831
833 DBG(3, "%s: phyaddr=0x%x, phyreg=0x%x, phydata=0x%x\n", 832 DBG(3, dev, "%s: phyaddr=0x%x, phyreg=0x%x, phydata=0x%x\n",
834 __func__, phyaddr, phyreg, phydata); 833 __func__, phyaddr, phyreg, phydata);
835 834
836 SMC_SELECT_BANK(lp, 2); 835 SMC_SELECT_BANK(lp, 2);
837 return phydata; 836 return phydata;
@@ -857,8 +856,8 @@ static void smc_phy_write(struct net_device *dev, int phyaddr, int phyreg,
857 /* Return to idle state */ 856 /* Return to idle state */
858 SMC_SET_MII(lp, SMC_GET_MII(lp) & ~(MII_MCLK|MII_MDOE|MII_MDO)); 857 SMC_SET_MII(lp, SMC_GET_MII(lp) & ~(MII_MCLK|MII_MDOE|MII_MDO));
859 858
860 DBG(3, "%s: phyaddr=0x%x, phyreg=0x%x, phydata=0x%x\n", 859 DBG(3, dev, "%s: phyaddr=0x%x, phyreg=0x%x, phydata=0x%x\n",
861 __func__, phyaddr, phyreg, phydata); 860 __func__, phyaddr, phyreg, phydata);
862 861
863 SMC_SELECT_BANK(lp, 2); 862 SMC_SELECT_BANK(lp, 2);
864} 863}
@@ -871,7 +870,7 @@ static void smc_phy_detect(struct net_device *dev)
871 struct smc_local *lp = netdev_priv(dev); 870 struct smc_local *lp = netdev_priv(dev);
872 int phyaddr; 871 int phyaddr;
873 872
874 DBG(2, "%s: %s\n", dev->name, __func__); 873 DBG(2, dev, "%s\n", __func__);
875 874
876 lp->phy_type = 0; 875 lp->phy_type = 0;
877 876
@@ -886,8 +885,8 @@ static void smc_phy_detect(struct net_device *dev)
886 id1 = smc_phy_read(dev, phyaddr & 31, MII_PHYSID1); 885 id1 = smc_phy_read(dev, phyaddr & 31, MII_PHYSID1);
887 id2 = smc_phy_read(dev, phyaddr & 31, MII_PHYSID2); 886 id2 = smc_phy_read(dev, phyaddr & 31, MII_PHYSID2);
888 887
889 DBG(3, "%s: phy_id1=0x%x, phy_id2=0x%x\n", 888 DBG(3, dev, "phy_id1=0x%x, phy_id2=0x%x\n",
890 dev->name, id1, id2); 889 id1, id2);
891 890
892 /* Make sure it is a valid identifier */ 891 /* Make sure it is a valid identifier */
893 if (id1 != 0x0000 && id1 != 0xffff && id1 != 0x8000 && 892 if (id1 != 0x0000 && id1 != 0xffff && id1 != 0x8000 &&
@@ -910,7 +909,7 @@ static int smc_phy_fixed(struct net_device *dev)
910 int phyaddr = lp->mii.phy_id; 909 int phyaddr = lp->mii.phy_id;
911 int bmcr, cfg1; 910 int bmcr, cfg1;
912 911
913 DBG(3, "%s: %s\n", dev->name, __func__); 912 DBG(3, dev, "%s\n", __func__);
914 913
915 /* Enter Link Disable state */ 914 /* Enter Link Disable state */
916 cfg1 = smc_phy_read(dev, phyaddr, PHY_CFG1_REG); 915 cfg1 = smc_phy_read(dev, phyaddr, PHY_CFG1_REG);
@@ -1044,7 +1043,7 @@ static void smc_phy_configure(struct work_struct *work)
1044 int my_ad_caps; /* My Advertised capabilities */ 1043 int my_ad_caps; /* My Advertised capabilities */
1045 int status; 1044 int status;
1046 1045
1047 DBG(3, "%s:smc_program_phy()\n", dev->name); 1046 DBG(3, dev, "smc_program_phy()\n");
1048 1047
1049 spin_lock_irq(&lp->lock); 1048 spin_lock_irq(&lp->lock);
1050 1049
@@ -1055,7 +1054,7 @@ static void smc_phy_configure(struct work_struct *work)
1055 goto smc_phy_configure_exit; 1054 goto smc_phy_configure_exit;
1056 1055
1057 if (smc_phy_reset(dev, phyaddr)) { 1056 if (smc_phy_reset(dev, phyaddr)) {
1058 printk("%s: PHY reset timed out\n", dev->name); 1057 netdev_info(dev, "PHY reset timed out\n");
1059 goto smc_phy_configure_exit; 1058 goto smc_phy_configure_exit;
1060 } 1059 }
1061 1060
@@ -1082,7 +1081,7 @@ static void smc_phy_configure(struct work_struct *work)
1082 my_phy_caps = smc_phy_read(dev, phyaddr, MII_BMSR); 1081 my_phy_caps = smc_phy_read(dev, phyaddr, MII_BMSR);
1083 1082
1084 if (!(my_phy_caps & BMSR_ANEGCAPABLE)) { 1083 if (!(my_phy_caps & BMSR_ANEGCAPABLE)) {
1085 printk(KERN_INFO "Auto negotiation NOT supported\n"); 1084 netdev_info(dev, "Auto negotiation NOT supported\n");
1086 smc_phy_fixed(dev); 1085 smc_phy_fixed(dev);
1087 goto smc_phy_configure_exit; 1086 goto smc_phy_configure_exit;
1088 } 1087 }
@@ -1118,8 +1117,8 @@ static void smc_phy_configure(struct work_struct *work)
1118 */ 1117 */
1119 status = smc_phy_read(dev, phyaddr, MII_ADVERTISE); 1118 status = smc_phy_read(dev, phyaddr, MII_ADVERTISE);
1120 1119
1121 DBG(2, "%s: phy caps=%x\n", dev->name, my_phy_caps); 1120 DBG(2, dev, "phy caps=%x\n", my_phy_caps);
1122 DBG(2, "%s: phy advertised caps=%x\n", dev->name, my_ad_caps); 1121 DBG(2, dev, "phy advertised caps=%x\n", my_ad_caps);
1123 1122
1124 /* Restart auto-negotiation process in order to advertise my caps */ 1123 /* Restart auto-negotiation process in order to advertise my caps */
1125 smc_phy_write(dev, phyaddr, MII_BMCR, BMCR_ANENABLE | BMCR_ANRESTART); 1124 smc_phy_write(dev, phyaddr, MII_BMCR, BMCR_ANENABLE | BMCR_ANRESTART);
@@ -1143,7 +1142,7 @@ static void smc_phy_interrupt(struct net_device *dev)
1143 int phyaddr = lp->mii.phy_id; 1142 int phyaddr = lp->mii.phy_id;
1144 int phy18; 1143 int phy18;
1145 1144
1146 DBG(2, "%s: %s\n", dev->name, __func__); 1145 DBG(2, dev, "%s\n", __func__);
1147 1146
1148 if (lp->phy_type == 0) 1147 if (lp->phy_type == 0)
1149 return; 1148 return;
@@ -1179,8 +1178,8 @@ static void smc_10bt_check_media(struct net_device *dev, int init)
1179 netif_carrier_on(dev); 1178 netif_carrier_on(dev);
1180 } 1179 }
1181 if (netif_msg_link(lp)) 1180 if (netif_msg_link(lp))
1182 printk(KERN_INFO "%s: link %s\n", dev->name, 1181 netdev_info(dev, "link %s\n",
1183 new_carrier ? "up" : "down"); 1182 new_carrier ? "up" : "down");
1184 } 1183 }
1185} 1184}
1186 1185
@@ -1211,7 +1210,7 @@ static irqreturn_t smc_interrupt(int irq, void *dev_id)
1211 int status, mask, timeout, card_stats; 1210 int status, mask, timeout, card_stats;
1212 int saved_pointer; 1211 int saved_pointer;
1213 1212
1214 DBG(3, "%s: %s\n", dev->name, __func__); 1213 DBG(3, dev, "%s\n", __func__);
1215 1214
1216 spin_lock(&lp->lock); 1215 spin_lock(&lp->lock);
1217 1216
@@ -1230,12 +1229,12 @@ static irqreturn_t smc_interrupt(int irq, void *dev_id)
1230 do { 1229 do {
1231 status = SMC_GET_INT(lp); 1230 status = SMC_GET_INT(lp);
1232 1231
1233 DBG(2, "%s: INT 0x%02x MASK 0x%02x MEM 0x%04x FIFO 0x%04x\n", 1232 DBG(2, dev, "INT 0x%02x MASK 0x%02x MEM 0x%04x FIFO 0x%04x\n",
1234 dev->name, status, mask, 1233 status, mask,
1235 ({ int meminfo; SMC_SELECT_BANK(lp, 0); 1234 ({ int meminfo; SMC_SELECT_BANK(lp, 0);
1236 meminfo = SMC_GET_MIR(lp); 1235 meminfo = SMC_GET_MIR(lp);
1237 SMC_SELECT_BANK(lp, 2); meminfo; }), 1236 SMC_SELECT_BANK(lp, 2); meminfo; }),
1238 SMC_GET_FIFO(lp)); 1237 SMC_GET_FIFO(lp));
1239 1238
1240 status &= mask; 1239 status &= mask;
1241 if (!status) 1240 if (!status)
@@ -1243,20 +1242,20 @@ static irqreturn_t smc_interrupt(int irq, void *dev_id)
1243 1242
1244 if (status & IM_TX_INT) { 1243 if (status & IM_TX_INT) {
1245 /* do this before RX as it will free memory quickly */ 1244 /* do this before RX as it will free memory quickly */
1246 DBG(3, "%s: TX int\n", dev->name); 1245 DBG(3, dev, "TX int\n");
1247 smc_tx(dev); 1246 smc_tx(dev);
1248 SMC_ACK_INT(lp, IM_TX_INT); 1247 SMC_ACK_INT(lp, IM_TX_INT);
1249 if (THROTTLE_TX_PKTS) 1248 if (THROTTLE_TX_PKTS)
1250 netif_wake_queue(dev); 1249 netif_wake_queue(dev);
1251 } else if (status & IM_RCV_INT) { 1250 } else if (status & IM_RCV_INT) {
1252 DBG(3, "%s: RX irq\n", dev->name); 1251 DBG(3, dev, "RX irq\n");
1253 smc_rcv(dev); 1252 smc_rcv(dev);
1254 } else if (status & IM_ALLOC_INT) { 1253 } else if (status & IM_ALLOC_INT) {
1255 DBG(3, "%s: Allocation irq\n", dev->name); 1254 DBG(3, dev, "Allocation irq\n");
1256 tasklet_hi_schedule(&lp->tx_task); 1255 tasklet_hi_schedule(&lp->tx_task);
1257 mask &= ~IM_ALLOC_INT; 1256 mask &= ~IM_ALLOC_INT;
1258 } else if (status & IM_TX_EMPTY_INT) { 1257 } else if (status & IM_TX_EMPTY_INT) {
1259 DBG(3, "%s: TX empty\n", dev->name); 1258 DBG(3, dev, "TX empty\n");
1260 mask &= ~IM_TX_EMPTY_INT; 1259 mask &= ~IM_TX_EMPTY_INT;
1261 1260
1262 /* update stats */ 1261 /* update stats */
@@ -1271,10 +1270,10 @@ static irqreturn_t smc_interrupt(int irq, void *dev_id)
1271 /* multiple collisions */ 1270 /* multiple collisions */
1272 dev->stats.collisions += card_stats & 0xF; 1271 dev->stats.collisions += card_stats & 0xF;
1273 } else if (status & IM_RX_OVRN_INT) { 1272 } else if (status & IM_RX_OVRN_INT) {
1274 DBG(1, "%s: RX overrun (EPH_ST 0x%04x)\n", dev->name, 1273 DBG(1, dev, "RX overrun (EPH_ST 0x%04x)\n",
1275 ({ int eph_st; SMC_SELECT_BANK(lp, 0); 1274 ({ int eph_st; SMC_SELECT_BANK(lp, 0);
1276 eph_st = SMC_GET_EPH_STATUS(lp); 1275 eph_st = SMC_GET_EPH_STATUS(lp);
1277 SMC_SELECT_BANK(lp, 2); eph_st; })); 1276 SMC_SELECT_BANK(lp, 2); eph_st; }));
1278 SMC_ACK_INT(lp, IM_RX_OVRN_INT); 1277 SMC_ACK_INT(lp, IM_RX_OVRN_INT);
1279 dev->stats.rx_errors++; 1278 dev->stats.rx_errors++;
1280 dev->stats.rx_fifo_errors++; 1279 dev->stats.rx_fifo_errors++;
@@ -1285,7 +1284,7 @@ static irqreturn_t smc_interrupt(int irq, void *dev_id)
1285 smc_phy_interrupt(dev); 1284 smc_phy_interrupt(dev);
1286 } else if (status & IM_ERCV_INT) { 1285 } else if (status & IM_ERCV_INT) {
1287 SMC_ACK_INT(lp, IM_ERCV_INT); 1286 SMC_ACK_INT(lp, IM_ERCV_INT);
1288 PRINTK("%s: UNSUPPORTED: ERCV INTERRUPT\n", dev->name); 1287 PRINTK(dev, "UNSUPPORTED: ERCV INTERRUPT\n");
1289 } 1288 }
1290 } while (--timeout); 1289 } while (--timeout);
1291 1290
@@ -1296,11 +1295,11 @@ static irqreturn_t smc_interrupt(int irq, void *dev_id)
1296 1295
1297#ifndef CONFIG_NET_POLL_CONTROLLER 1296#ifndef CONFIG_NET_POLL_CONTROLLER
1298 if (timeout == MAX_IRQ_LOOPS) 1297 if (timeout == MAX_IRQ_LOOPS)
1299 PRINTK("%s: spurious interrupt (mask = 0x%02x)\n", 1298 PRINTK(dev, "spurious interrupt (mask = 0x%02x)\n",
1300 dev->name, mask); 1299 mask);
1301#endif 1300#endif
1302 DBG(3, "%s: Interrupt done (%d loops)\n", 1301 DBG(3, dev, "Interrupt done (%d loops)\n",
1303 dev->name, MAX_IRQ_LOOPS - timeout); 1302 MAX_IRQ_LOOPS - timeout);
1304 1303
1305 /* 1304 /*
1306 * We return IRQ_HANDLED unconditionally here even if there was 1305 * We return IRQ_HANDLED unconditionally here even if there was
@@ -1333,7 +1332,7 @@ static void smc_timeout(struct net_device *dev)
1333 void __iomem *ioaddr = lp->base; 1332 void __iomem *ioaddr = lp->base;
1334 int status, mask, eph_st, meminfo, fifo; 1333 int status, mask, eph_st, meminfo, fifo;
1335 1334
1336 DBG(2, "%s: %s\n", dev->name, __func__); 1335 DBG(2, dev, "%s\n", __func__);
1337 1336
1338 spin_lock_irq(&lp->lock); 1337 spin_lock_irq(&lp->lock);
1339 status = SMC_GET_INT(lp); 1338 status = SMC_GET_INT(lp);
@@ -1344,9 +1343,8 @@ static void smc_timeout(struct net_device *dev)
1344 meminfo = SMC_GET_MIR(lp); 1343 meminfo = SMC_GET_MIR(lp);
1345 SMC_SELECT_BANK(lp, 2); 1344 SMC_SELECT_BANK(lp, 2);
1346 spin_unlock_irq(&lp->lock); 1345 spin_unlock_irq(&lp->lock);
1347 PRINTK( "%s: TX timeout (INT 0x%02x INTMASK 0x%02x " 1346 PRINTK(dev, "TX timeout (INT 0x%02x INTMASK 0x%02x MEM 0x%04x FIFO 0x%04x EPH_ST 0x%04x)\n",
1348 "MEM 0x%04x FIFO 0x%04x EPH_ST 0x%04x)\n", 1347 status, mask, meminfo, fifo, eph_st);
1349 dev->name, status, mask, meminfo, fifo, eph_st );
1350 1348
1351 smc_reset(dev); 1349 smc_reset(dev);
1352 smc_enable(dev); 1350 smc_enable(dev);
@@ -1377,10 +1375,10 @@ static void smc_set_multicast_list(struct net_device *dev)
1377 unsigned char multicast_table[8]; 1375 unsigned char multicast_table[8];
1378 int update_multicast = 0; 1376 int update_multicast = 0;
1379 1377
1380 DBG(2, "%s: %s\n", dev->name, __func__); 1378 DBG(2, dev, "%s\n", __func__);
1381 1379
1382 if (dev->flags & IFF_PROMISC) { 1380 if (dev->flags & IFF_PROMISC) {
1383 DBG(2, "%s: RCR_PRMS\n", dev->name); 1381 DBG(2, dev, "RCR_PRMS\n");
1384 lp->rcr_cur_mode |= RCR_PRMS; 1382 lp->rcr_cur_mode |= RCR_PRMS;
1385 } 1383 }
1386 1384
@@ -1395,7 +1393,7 @@ static void smc_set_multicast_list(struct net_device *dev)
1395 * checked before the table is 1393 * checked before the table is
1396 */ 1394 */
1397 else if (dev->flags & IFF_ALLMULTI || netdev_mc_count(dev) > 16) { 1395 else if (dev->flags & IFF_ALLMULTI || netdev_mc_count(dev) > 16) {
1398 DBG(2, "%s: RCR_ALMUL\n", dev->name); 1396 DBG(2, dev, "RCR_ALMUL\n");
1399 lp->rcr_cur_mode |= RCR_ALMUL; 1397 lp->rcr_cur_mode |= RCR_ALMUL;
1400 } 1398 }
1401 1399
@@ -1437,7 +1435,7 @@ static void smc_set_multicast_list(struct net_device *dev)
1437 /* now, the table can be loaded into the chipset */ 1435 /* now, the table can be loaded into the chipset */
1438 update_multicast = 1; 1436 update_multicast = 1;
1439 } else { 1437 } else {
1440 DBG(2, "%s: ~(RCR_PRMS|RCR_ALMUL)\n", dev->name); 1438 DBG(2, dev, "~(RCR_PRMS|RCR_ALMUL)\n");
1441 lp->rcr_cur_mode &= ~(RCR_PRMS | RCR_ALMUL); 1439 lp->rcr_cur_mode &= ~(RCR_PRMS | RCR_ALMUL);
1442 1440
1443 /* 1441 /*
@@ -1470,7 +1468,7 @@ smc_open(struct net_device *dev)
1470{ 1468{
1471 struct smc_local *lp = netdev_priv(dev); 1469 struct smc_local *lp = netdev_priv(dev);
1472 1470
1473 DBG(2, "%s: %s\n", dev->name, __func__); 1471 DBG(2, dev, "%s\n", __func__);
1474 1472
1475 /* Setup the default Register Modes */ 1473 /* Setup the default Register Modes */
1476 lp->tcr_cur_mode = TCR_DEFAULT; 1474 lp->tcr_cur_mode = TCR_DEFAULT;
@@ -1514,7 +1512,7 @@ static int smc_close(struct net_device *dev)
1514{ 1512{
1515 struct smc_local *lp = netdev_priv(dev); 1513 struct smc_local *lp = netdev_priv(dev);
1516 1514
1517 DBG(2, "%s: %s\n", dev->name, __func__); 1515 DBG(2, dev, "%s\n", __func__);
1518 1516
1519 netif_stop_queue(dev); 1517 netif_stop_queue(dev);
1520 netif_carrier_off(dev); 1518 netif_carrier_off(dev);
@@ -1694,7 +1692,7 @@ static int smc_ethtool_geteeprom(struct net_device *dev,
1694 int i; 1692 int i;
1695 int imax; 1693 int imax;
1696 1694
1697 DBG(1, "Reading %d bytes at %d(0x%x)\n", 1695 DBG(1, dev, "Reading %d bytes at %d(0x%x)\n",
1698 eeprom->len, eeprom->offset, eeprom->offset); 1696 eeprom->len, eeprom->offset, eeprom->offset);
1699 imax = smc_ethtool_geteeprom_len(dev); 1697 imax = smc_ethtool_geteeprom_len(dev);
1700 for (i = 0; i < eeprom->len; i += 2) { 1698 for (i = 0; i < eeprom->len; i += 2) {
@@ -1706,7 +1704,7 @@ static int smc_ethtool_geteeprom(struct net_device *dev,
1706 ret = smc_read_eeprom_word(dev, offset >> 1, &wbuf); 1704 ret = smc_read_eeprom_word(dev, offset >> 1, &wbuf);
1707 if (ret != 0) 1705 if (ret != 0)
1708 return ret; 1706 return ret;
1709 DBG(2, "Read 0x%x from 0x%x\n", wbuf, offset >> 1); 1707 DBG(2, dev, "Read 0x%x from 0x%x\n", wbuf, offset >> 1);
1710 data[i] = (wbuf >> 8) & 0xff; 1708 data[i] = (wbuf >> 8) & 0xff;
1711 data[i+1] = wbuf & 0xff; 1709 data[i+1] = wbuf & 0xff;
1712 } 1710 }
@@ -1719,8 +1717,8 @@ static int smc_ethtool_seteeprom(struct net_device *dev,
1719 int i; 1717 int i;
1720 int imax; 1718 int imax;
1721 1719
1722 DBG(1, "Writing %d bytes to %d(0x%x)\n", 1720 DBG(1, dev, "Writing %d bytes to %d(0x%x)\n",
1723 eeprom->len, eeprom->offset, eeprom->offset); 1721 eeprom->len, eeprom->offset, eeprom->offset);
1724 imax = smc_ethtool_geteeprom_len(dev); 1722 imax = smc_ethtool_geteeprom_len(dev);
1725 for (i = 0; i < eeprom->len; i += 2) { 1723 for (i = 0; i < eeprom->len; i += 2) {
1726 int ret; 1724 int ret;
@@ -1729,7 +1727,7 @@ static int smc_ethtool_seteeprom(struct net_device *dev,
1729 if (offset > imax) 1727 if (offset > imax)
1730 break; 1728 break;
1731 wbuf = (data[i] << 8) | data[i + 1]; 1729 wbuf = (data[i] << 8) | data[i + 1];
1732 DBG(2, "Writing 0x%x to 0x%x\n", wbuf, offset >> 1); 1730 DBG(2, dev, "Writing 0x%x to 0x%x\n", wbuf, offset >> 1);
1733 ret = smc_write_eeprom_word(dev, offset >> 1, wbuf); 1731 ret = smc_write_eeprom_word(dev, offset >> 1, wbuf);
1734 if (ret != 0) 1732 if (ret != 0)
1735 return ret; 1733 return ret;
@@ -1784,7 +1782,7 @@ static int smc_findirq(struct smc_local *lp)
1784 int timeout = 20; 1782 int timeout = 20;
1785 unsigned long cookie; 1783 unsigned long cookie;
1786 1784
1787 DBG(2, "%s: %s\n", CARDNAME, __func__); 1785 DBG(2, dev, "%s: %s\n", CARDNAME, __func__);
1788 1786
1789 cookie = probe_irq_on(); 1787 cookie = probe_irq_on();
1790 1788
@@ -1856,21 +1854,21 @@ static int smc_probe(struct net_device *dev, void __iomem *ioaddr,
1856 unsigned long irq_flags) 1854 unsigned long irq_flags)
1857{ 1855{
1858 struct smc_local *lp = netdev_priv(dev); 1856 struct smc_local *lp = netdev_priv(dev);
1859 static int version_printed = 0;
1860 int retval; 1857 int retval;
1861 unsigned int val, revision_register; 1858 unsigned int val, revision_register;
1862 const char *version_string; 1859 const char *version_string;
1863 1860
1864 DBG(2, "%s: %s\n", CARDNAME, __func__); 1861 DBG(2, dev, "%s: %s\n", CARDNAME, __func__);
1865 1862
1866 /* First, see if the high byte is 0x33 */ 1863 /* First, see if the high byte is 0x33 */
1867 val = SMC_CURRENT_BANK(lp); 1864 val = SMC_CURRENT_BANK(lp);
1868 DBG(2, "%s: bank signature probe returned 0x%04x\n", CARDNAME, val); 1865 DBG(2, dev, "%s: bank signature probe returned 0x%04x\n",
1866 CARDNAME, val);
1869 if ((val & 0xFF00) != 0x3300) { 1867 if ((val & 0xFF00) != 0x3300) {
1870 if ((val & 0xFF) == 0x33) { 1868 if ((val & 0xFF) == 0x33) {
1871 printk(KERN_WARNING 1869 netdev_warn(dev,
1872 "%s: Detected possible byte-swapped interface" 1870 "%s: Detected possible byte-swapped interface at IOADDR %p\n",
1873 " at IOADDR %p\n", CARDNAME, ioaddr); 1871 CARDNAME, ioaddr);
1874 } 1872 }
1875 retval = -ENODEV; 1873 retval = -ENODEV;
1876 goto err_out; 1874 goto err_out;
@@ -1897,8 +1895,8 @@ static int smc_probe(struct net_device *dev, void __iomem *ioaddr,
1897 val = SMC_GET_BASE(lp); 1895 val = SMC_GET_BASE(lp);
1898 val = ((val & 0x1F00) >> 3) << SMC_IO_SHIFT; 1896 val = ((val & 0x1F00) >> 3) << SMC_IO_SHIFT;
1899 if (((unsigned int)ioaddr & (0x3e0 << SMC_IO_SHIFT)) != val) { 1897 if (((unsigned int)ioaddr & (0x3e0 << SMC_IO_SHIFT)) != val) {
1900 printk("%s: IOADDR %p doesn't match configuration (%x).\n", 1898 netdev_warn(dev, "%s: IOADDR %p doesn't match configuration (%x).\n",
1901 CARDNAME, ioaddr, val); 1899 CARDNAME, ioaddr, val);
1902 } 1900 }
1903 1901
1904 /* 1902 /*
@@ -1908,21 +1906,19 @@ static int smc_probe(struct net_device *dev, void __iomem *ioaddr,
1908 */ 1906 */
1909 SMC_SELECT_BANK(lp, 3); 1907 SMC_SELECT_BANK(lp, 3);
1910 revision_register = SMC_GET_REV(lp); 1908 revision_register = SMC_GET_REV(lp);
1911 DBG(2, "%s: revision = 0x%04x\n", CARDNAME, revision_register); 1909 DBG(2, dev, "%s: revision = 0x%04x\n", CARDNAME, revision_register);
1912 version_string = chip_ids[ (revision_register >> 4) & 0xF]; 1910 version_string = chip_ids[ (revision_register >> 4) & 0xF];
1913 if (!version_string || (revision_register & 0xff00) != 0x3300) { 1911 if (!version_string || (revision_register & 0xff00) != 0x3300) {
1914 /* I don't recognize this chip, so... */ 1912 /* I don't recognize this chip, so... */
1915 printk("%s: IO %p: Unrecognized revision register 0x%04x" 1913 netdev_warn(dev, "%s: IO %p: Unrecognized revision register 0x%04x, Contact author.\n",
1916 ", Contact author.\n", CARDNAME, 1914 CARDNAME, ioaddr, revision_register);
1917 ioaddr, revision_register);
1918 1915
1919 retval = -ENODEV; 1916 retval = -ENODEV;
1920 goto err_out; 1917 goto err_out;
1921 } 1918 }
1922 1919
1923 /* At this point I'll assume that the chip is an SMC91x. */ 1920 /* At this point I'll assume that the chip is an SMC91x. */
1924 if (version_printed++ == 0) 1921 pr_info_once("%s\n", version);
1925 printk("%s", version);
1926 1922
1927 /* fill in some of the fields */ 1923 /* fill in some of the fields */
1928 dev->base_addr = (unsigned long)ioaddr; 1924 dev->base_addr = (unsigned long)ioaddr;
@@ -1940,7 +1936,7 @@ static int smc_probe(struct net_device *dev, void __iomem *ioaddr,
1940 /* 1936 /*
1941 * If dev->irq is 0, then the device has to be banged on to see 1937 * If dev->irq is 0, then the device has to be banged on to see
1942 * what the IRQ is. 1938 * what the IRQ is.
1943 * 1939 *
1944 * This banging doesn't always detect the IRQ, for unknown reasons. 1940 * This banging doesn't always detect the IRQ, for unknown reasons.
1945 * a workaround is to reset the chip and try again. 1941 * a workaround is to reset the chip and try again.
1946 * 1942 *
@@ -1965,8 +1961,7 @@ static int smc_probe(struct net_device *dev, void __iomem *ioaddr,
1965 } 1961 }
1966 } 1962 }
1967 if (dev->irq == 0) { 1963 if (dev->irq == 0) {
1968 printk("%s: Couldn't autodetect your IRQ. Use irq=xx.\n", 1964 netdev_warn(dev, "Couldn't autodetect your IRQ. Use irq=xx.\n");
1969 dev->name);
1970 retval = -ENODEV; 1965 retval = -ENODEV;
1971 goto err_out; 1966 goto err_out;
1972 } 1967 }
@@ -2030,32 +2025,31 @@ static int smc_probe(struct net_device *dev, void __iomem *ioaddr,
2030 retval = register_netdev(dev); 2025 retval = register_netdev(dev);
2031 if (retval == 0) { 2026 if (retval == 0) {
2032 /* now, print out the card info, in a short format.. */ 2027 /* now, print out the card info, in a short format.. */
2033 printk("%s: %s (rev %d) at %p IRQ %d", 2028 netdev_info(dev, "%s (rev %d) at %p IRQ %d",
2034 dev->name, version_string, revision_register & 0x0f, 2029 version_string, revision_register & 0x0f,
2035 lp->base, dev->irq); 2030 lp->base, dev->irq);
2036 2031
2037 if (dev->dma != (unsigned char)-1) 2032 if (dev->dma != (unsigned char)-1)
2038 printk(" DMA %d", dev->dma); 2033 pr_cont(" DMA %d", dev->dma);
2039 2034
2040 printk("%s%s\n", 2035 pr_cont("%s%s\n",
2041 lp->cfg.flags & SMC91X_NOWAIT ? " [nowait]" : "", 2036 lp->cfg.flags & SMC91X_NOWAIT ? " [nowait]" : "",
2042 THROTTLE_TX_PKTS ? " [throttle_tx]" : ""); 2037 THROTTLE_TX_PKTS ? " [throttle_tx]" : "");
2043 2038
2044 if (!is_valid_ether_addr(dev->dev_addr)) { 2039 if (!is_valid_ether_addr(dev->dev_addr)) {
2045 printk("%s: Invalid ethernet MAC address. Please " 2040 netdev_warn(dev, "Invalid ethernet MAC address. Please set using ifconfig\n");
2046 "set using ifconfig\n", dev->name);
2047 } else { 2041 } else {
2048 /* Print the Ethernet address */ 2042 /* Print the Ethernet address */
2049 printk("%s: Ethernet addr: %pM\n", 2043 netdev_info(dev, "Ethernet addr: %pM\n",
2050 dev->name, dev->dev_addr); 2044 dev->dev_addr);
2051 } 2045 }
2052 2046
2053 if (lp->phy_type == 0) { 2047 if (lp->phy_type == 0) {
2054 PRINTK("%s: No PHY found\n", dev->name); 2048 PRINTK(dev, "No PHY found\n");
2055 } else if ((lp->phy_type & 0xfffffff0) == 0x0016f840) { 2049 } else if ((lp->phy_type & 0xfffffff0) == 0x0016f840) {
2056 PRINTK("%s: PHY LAN83C183 (LAN91C111 Internal)\n", dev->name); 2050 PRINTK(dev, "PHY LAN83C183 (LAN91C111 Internal)\n");
2057 } else if ((lp->phy_type & 0xfffffff0) == 0x02821c50) { 2051 } else if ((lp->phy_type & 0xfffffff0) == 0x02821c50) {
2058 PRINTK("%s: PHY LAN83C180\n", dev->name); 2052 PRINTK(dev, "PHY LAN83C180\n");
2059 } 2053 }
2060 } 2054 }
2061 2055
@@ -2165,7 +2159,8 @@ static inline void smc_request_datacs(struct platform_device *pdev, struct net_d
2165 return; 2159 return;
2166 2160
2167 if(!request_mem_region(res->start, SMC_DATA_EXTENT, CARDNAME)) { 2161 if(!request_mem_region(res->start, SMC_DATA_EXTENT, CARDNAME)) {
2168 printk(KERN_INFO "%s: failed to request datacs memory region.\n", CARDNAME); 2162 netdev_info(ndev, "%s: failed to request datacs memory region.\n",
2163 CARDNAME);
2169 return; 2164 return;
2170 } 2165 }
2171 2166
@@ -2307,7 +2302,7 @@ static int smc_drv_probe(struct platform_device *pdev)
2307 out_free_netdev: 2302 out_free_netdev:
2308 free_netdev(ndev); 2303 free_netdev(ndev);
2309 out: 2304 out:
2310 printk("%s: not found (%d).\n", CARDNAME, ret); 2305 pr_info("%s: not found (%d).\n", CARDNAME, ret);
2311 2306
2312 return ret; 2307 return ret;
2313} 2308}
diff --git a/drivers/net/ethernet/smsc/smc91x.h b/drivers/net/ethernet/smsc/smc91x.h
index 98eedb90cdc3..c9d4c872e81d 100644
--- a/drivers/net/ethernet/smsc/smc91x.h
+++ b/drivers/net/ethernet/smsc/smc91x.h
@@ -907,8 +907,8 @@ static const char * chip_ids[ 16 ] = {
907 ({ \ 907 ({ \
908 int __b = SMC_CURRENT_BANK(lp); \ 908 int __b = SMC_CURRENT_BANK(lp); \
909 if (unlikely((__b & ~0xf0) != (0x3300 | bank))) { \ 909 if (unlikely((__b & ~0xf0) != (0x3300 | bank))) { \
910 printk( "%s: bank reg screwed (0x%04x)\n", \ 910 pr_err("%s: bank reg screwed (0x%04x)\n", \
911 CARDNAME, __b ); \ 911 CARDNAME, __b); \
912 BUG(); \ 912 BUG(); \
913 } \ 913 } \
914 reg<<SMC_IO_SHIFT; \ 914 reg<<SMC_IO_SHIFT; \
diff --git a/drivers/net/ethernet/smsc/smsc911x.c b/drivers/net/ethernet/smsc/smsc911x.c
index 5fdbc2686eb3..8564f23a6796 100644
--- a/drivers/net/ethernet/smsc/smsc911x.c
+++ b/drivers/net/ethernet/smsc/smsc911x.c
@@ -2167,7 +2167,7 @@ static int smsc911x_init(struct net_device *dev)
2167 udelay(1000); 2167 udelay(1000);
2168 2168
2169 if (to == 0) { 2169 if (to == 0) {
2170 pr_err("Device not READY in 100ms aborting\n"); 2170 netdev_err(dev, "Device not READY in 100ms aborting\n");
2171 return -ENODEV; 2171 return -ENODEV;
2172 } 2172 }
2173 2173
@@ -2502,7 +2502,7 @@ static int smsc911x_drv_probe(struct platform_device *pdev)
2502 SMSC_TRACE(pdata, probe, 2502 SMSC_TRACE(pdata, probe,
2503 "MAC Address is specified by configuration"); 2503 "MAC Address is specified by configuration");
2504 } else if (is_valid_ether_addr(pdata->config.mac)) { 2504 } else if (is_valid_ether_addr(pdata->config.mac)) {
2505 memcpy(dev->dev_addr, pdata->config.mac, 6); 2505 memcpy(dev->dev_addr, pdata->config.mac, ETH_ALEN);
2506 SMSC_TRACE(pdata, probe, 2506 SMSC_TRACE(pdata, probe,
2507 "MAC Address specified by platform data"); 2507 "MAC Address specified by platform data");
2508 } else { 2508 } else {
diff --git a/drivers/net/ethernet/smsc/smsc9420.c b/drivers/net/ethernet/smsc/smsc9420.c
index 5f9e79f7f2df..f433d97aa097 100644
--- a/drivers/net/ethernet/smsc/smsc9420.c
+++ b/drivers/net/ethernet/smsc/smsc9420.c
@@ -19,6 +19,8 @@
19 *************************************************************************** 19 ***************************************************************************
20 */ 20 */
21 21
22#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
23
22#include <linux/interrupt.h> 24#include <linux/interrupt.h>
23#include <linux/kernel.h> 25#include <linux/kernel.h>
24#include <linux/netdevice.h> 26#include <linux/netdevice.h>
@@ -33,7 +35,6 @@
33#include "smsc9420.h" 35#include "smsc9420.h"
34 36
35#define DRV_NAME "smsc9420" 37#define DRV_NAME "smsc9420"
36#define PFX DRV_NAME ": "
37#define DRV_MDIONAME "smsc9420-mdio" 38#define DRV_MDIONAME "smsc9420-mdio"
38#define DRV_DESCRIPTION "SMSC LAN9420 driver" 39#define DRV_DESCRIPTION "SMSC LAN9420 driver"
39#define DRV_VERSION "1.01" 40#define DRV_VERSION "1.01"
@@ -97,21 +98,6 @@ static uint debug = -1;
97module_param(debug, uint, 0); 98module_param(debug, uint, 0);
98MODULE_PARM_DESC(debug, "debug level"); 99MODULE_PARM_DESC(debug, "debug level");
99 100
100#define smsc_dbg(TYPE, f, a...) \
101do { if ((pd)->msg_enable & NETIF_MSG_##TYPE) \
102 printk(KERN_DEBUG PFX f "\n", ## a); \
103} while (0)
104
105#define smsc_info(TYPE, f, a...) \
106do { if ((pd)->msg_enable & NETIF_MSG_##TYPE) \
107 printk(KERN_INFO PFX f "\n", ## a); \
108} while (0)
109
110#define smsc_warn(TYPE, f, a...) \
111do { if ((pd)->msg_enable & NETIF_MSG_##TYPE) \
112 printk(KERN_WARNING PFX f "\n", ## a); \
113} while (0)
114
115static inline u32 smsc9420_reg_read(struct smsc9420_pdata *pd, u32 offset) 101static inline u32 smsc9420_reg_read(struct smsc9420_pdata *pd, u32 offset)
116{ 102{
117 return ioread32(pd->ioaddr + offset); 103 return ioread32(pd->ioaddr + offset);
@@ -140,7 +126,7 @@ static int smsc9420_mii_read(struct mii_bus *bus, int phyaddr, int regidx)
140 126
141 /* confirm MII not busy */ 127 /* confirm MII not busy */
142 if ((smsc9420_reg_read(pd, MII_ACCESS) & MII_ACCESS_MII_BUSY_)) { 128 if ((smsc9420_reg_read(pd, MII_ACCESS) & MII_ACCESS_MII_BUSY_)) {
143 smsc_warn(DRV, "MII is busy???"); 129 netif_warn(pd, drv, pd->dev, "MII is busy???\n");
144 goto out; 130 goto out;
145 } 131 }
146 132
@@ -159,7 +145,7 @@ static int smsc9420_mii_read(struct mii_bus *bus, int phyaddr, int regidx)
159 udelay(10); 145 udelay(10);
160 } 146 }
161 147
162 smsc_warn(DRV, "MII busy timeout!"); 148 netif_warn(pd, drv, pd->dev, "MII busy timeout!\n");
163 149
164out: 150out:
165 spin_unlock_irqrestore(&pd->phy_lock, flags); 151 spin_unlock_irqrestore(&pd->phy_lock, flags);
@@ -178,7 +164,7 @@ static int smsc9420_mii_write(struct mii_bus *bus, int phyaddr, int regidx,
178 164
179 /* confirm MII not busy */ 165 /* confirm MII not busy */
180 if ((smsc9420_reg_read(pd, MII_ACCESS) & MII_ACCESS_MII_BUSY_)) { 166 if ((smsc9420_reg_read(pd, MII_ACCESS) & MII_ACCESS_MII_BUSY_)) {
181 smsc_warn(DRV, "MII is busy???"); 167 netif_warn(pd, drv, pd->dev, "MII is busy???\n");
182 goto out; 168 goto out;
183 } 169 }
184 170
@@ -200,7 +186,7 @@ static int smsc9420_mii_write(struct mii_bus *bus, int phyaddr, int regidx,
200 udelay(10); 186 udelay(10);
201 } 187 }
202 188
203 smsc_warn(DRV, "MII busy timeout!"); 189 netif_warn(pd, drv, pd->dev, "MII busy timeout!\n");
204 190
205out: 191out:
206 spin_unlock_irqrestore(&pd->phy_lock, flags); 192 spin_unlock_irqrestore(&pd->phy_lock, flags);
@@ -222,7 +208,7 @@ static int smsc9420_eeprom_reload(struct smsc9420_pdata *pd)
222 BUG_ON(!pd); 208 BUG_ON(!pd);
223 209
224 if (smsc9420_reg_read(pd, E2P_CMD) & E2P_CMD_EPC_BUSY_) { 210 if (smsc9420_reg_read(pd, E2P_CMD) & E2P_CMD_EPC_BUSY_) {
225 smsc_dbg(DRV, "smsc9420_eeprom_reload: Eeprom busy"); 211 netif_dbg(pd, drv, pd->dev, "%s: Eeprom busy\n", __func__);
226 return -EIO; 212 return -EIO;
227 } 213 }
228 214
@@ -235,7 +221,7 @@ static int smsc9420_eeprom_reload(struct smsc9420_pdata *pd)
235 return 0; 221 return 0;
236 } while (timeout--); 222 } while (timeout--);
237 223
238 smsc_warn(DRV, "smsc9420_eeprom_reload: Eeprom timed out"); 224 netif_warn(pd, drv, pd->dev, "%s: Eeprom timed out\n", __func__);
239 return -EIO; 225 return -EIO;
240} 226}
241 227
@@ -347,9 +333,9 @@ static int smsc9420_eeprom_send_cmd(struct smsc9420_pdata *pd, u32 op)
347 int timeout = 100; 333 int timeout = 100;
348 u32 e2cmd; 334 u32 e2cmd;
349 335
350 smsc_dbg(HW, "op 0x%08x", op); 336 netif_dbg(pd, hw, pd->dev, "op 0x%08x\n", op);
351 if (smsc9420_reg_read(pd, E2P_CMD) & E2P_CMD_EPC_BUSY_) { 337 if (smsc9420_reg_read(pd, E2P_CMD) & E2P_CMD_EPC_BUSY_) {
352 smsc_warn(HW, "Busy at start"); 338 netif_warn(pd, hw, pd->dev, "Busy at start\n");
353 return -EBUSY; 339 return -EBUSY;
354 } 340 }
355 341
@@ -362,12 +348,13 @@ static int smsc9420_eeprom_send_cmd(struct smsc9420_pdata *pd, u32 op)
362 } while ((e2cmd & E2P_CMD_EPC_BUSY_) && (--timeout)); 348 } while ((e2cmd & E2P_CMD_EPC_BUSY_) && (--timeout));
363 349
364 if (!timeout) { 350 if (!timeout) {
365 smsc_info(HW, "TIMED OUT"); 351 netif_info(pd, hw, pd->dev, "TIMED OUT\n");
366 return -EAGAIN; 352 return -EAGAIN;
367 } 353 }
368 354
369 if (e2cmd & E2P_CMD_EPC_TIMEOUT_) { 355 if (e2cmd & E2P_CMD_EPC_TIMEOUT_) {
370 smsc_info(HW, "Error occurred during eeprom operation"); 356 netif_info(pd, hw, pd->dev,
357 "Error occurred during eeprom operation\n");
371 return -EINVAL; 358 return -EINVAL;
372 } 359 }
373 360
@@ -380,7 +367,7 @@ static int smsc9420_eeprom_read_location(struct smsc9420_pdata *pd,
380 u32 op = E2P_CMD_EPC_CMD_READ_ | address; 367 u32 op = E2P_CMD_EPC_CMD_READ_ | address;
381 int ret; 368 int ret;
382 369
383 smsc_dbg(HW, "address 0x%x", address); 370 netif_dbg(pd, hw, pd->dev, "address 0x%x\n", address);
384 ret = smsc9420_eeprom_send_cmd(pd, op); 371 ret = smsc9420_eeprom_send_cmd(pd, op);
385 372
386 if (!ret) 373 if (!ret)
@@ -395,7 +382,7 @@ static int smsc9420_eeprom_write_location(struct smsc9420_pdata *pd,
395 u32 op = E2P_CMD_EPC_CMD_ERASE_ | address; 382 u32 op = E2P_CMD_EPC_CMD_ERASE_ | address;
396 int ret; 383 int ret;
397 384
398 smsc_dbg(HW, "address 0x%x, data 0x%x", address, data); 385 netif_dbg(pd, hw, pd->dev, "address 0x%x, data 0x%x\n", address, data);
399 ret = smsc9420_eeprom_send_cmd(pd, op); 386 ret = smsc9420_eeprom_send_cmd(pd, op);
400 387
401 if (!ret) { 388 if (!ret) {
@@ -492,7 +479,8 @@ static void smsc9420_check_mac_address(struct net_device *dev)
492 /* Check if mac address has been specified when bringing interface up */ 479 /* Check if mac address has been specified when bringing interface up */
493 if (is_valid_ether_addr(dev->dev_addr)) { 480 if (is_valid_ether_addr(dev->dev_addr)) {
494 smsc9420_set_mac_address(dev); 481 smsc9420_set_mac_address(dev);
495 smsc_dbg(PROBE, "MAC Address is specified by configuration"); 482 netif_dbg(pd, probe, pd->dev,
483 "MAC Address is specified by configuration\n");
496 } else { 484 } else {
497 /* Try reading mac address from device. if EEPROM is present 485 /* Try reading mac address from device. if EEPROM is present
498 * it will already have been set */ 486 * it will already have been set */
@@ -507,12 +495,14 @@ static void smsc9420_check_mac_address(struct net_device *dev)
507 495
508 if (is_valid_ether_addr(dev->dev_addr)) { 496 if (is_valid_ether_addr(dev->dev_addr)) {
509 /* eeprom values are valid so use them */ 497 /* eeprom values are valid so use them */
510 smsc_dbg(PROBE, "Mac Address is read from EEPROM"); 498 netif_dbg(pd, probe, pd->dev,
499 "Mac Address is read from EEPROM\n");
511 } else { 500 } else {
512 /* eeprom values are invalid, generate random MAC */ 501 /* eeprom values are invalid, generate random MAC */
513 eth_hw_addr_random(dev); 502 eth_hw_addr_random(dev);
514 smsc9420_set_mac_address(dev); 503 smsc9420_set_mac_address(dev);
515 smsc_dbg(PROBE, "MAC Address is set to random"); 504 netif_dbg(pd, probe, pd->dev,
505 "MAC Address is set to random\n");
516 } 506 }
517 } 507 }
518} 508}
@@ -535,7 +525,7 @@ static void smsc9420_stop_tx(struct smsc9420_pdata *pd)
535 } 525 }
536 526
537 if (!timeout) 527 if (!timeout)
538 smsc_warn(IFDOWN, "TX DMAC failed to stop"); 528 netif_warn(pd, ifdown, pd->dev, "TX DMAC failed to stop\n");
539 529
540 /* ACK Tx DMAC stop bit */ 530 /* ACK Tx DMAC stop bit */
541 smsc9420_reg_write(pd, DMAC_STATUS, DMAC_STS_TXPS_); 531 smsc9420_reg_write(pd, DMAC_STATUS, DMAC_STS_TXPS_);
@@ -646,7 +636,8 @@ static void smsc9420_stop_rx(struct smsc9420_pdata *pd)
646 } 636 }
647 637
648 if (!timeout) 638 if (!timeout)
649 smsc_warn(IFDOWN, "RX DMAC did not stop! timeout."); 639 netif_warn(pd, ifdown, pd->dev,
640 "RX DMAC did not stop! timeout\n");
650 641
651 /* ACK the Rx DMAC stop bit */ 642 /* ACK the Rx DMAC stop bit */
652 smsc9420_reg_write(pd, DMAC_STATUS, DMAC_STS_RXPS_); 643 smsc9420_reg_write(pd, DMAC_STATUS, DMAC_STS_RXPS_);
@@ -736,7 +727,7 @@ static void smsc9420_dmac_soft_reset(struct smsc9420_pdata *pd)
736 smsc9420_reg_read(pd, BUS_MODE); 727 smsc9420_reg_read(pd, BUS_MODE);
737 udelay(2); 728 udelay(2);
738 if (smsc9420_reg_read(pd, BUS_MODE) & BUS_MODE_SWR_) 729 if (smsc9420_reg_read(pd, BUS_MODE) & BUS_MODE_SWR_)
739 smsc_warn(DRV, "Software reset not cleared"); 730 netif_warn(pd, drv, pd->dev, "Software reset not cleared\n");
740} 731}
741 732
742static int smsc9420_stop(struct net_device *dev) 733static int smsc9420_stop(struct net_device *dev)
@@ -855,7 +846,7 @@ static int smsc9420_alloc_rx_buffer(struct smsc9420_pdata *pd, int index)
855 PKT_BUF_SZ, PCI_DMA_FROMDEVICE); 846 PKT_BUF_SZ, PCI_DMA_FROMDEVICE);
856 if (pci_dma_mapping_error(pd->pdev, mapping)) { 847 if (pci_dma_mapping_error(pd->pdev, mapping)) {
857 dev_kfree_skb_any(skb); 848 dev_kfree_skb_any(skb);
858 smsc_warn(RX_ERR, "pci_map_single failed!"); 849 netif_warn(pd, rx_err, pd->dev, "pci_map_single failed!\n");
859 return -ENOMEM; 850 return -ENOMEM;
860 } 851 }
861 852
@@ -1004,7 +995,8 @@ static netdev_tx_t smsc9420_hard_start_xmit(struct sk_buff *skb,
1004 mapping = pci_map_single(pd->pdev, skb->data, 995 mapping = pci_map_single(pd->pdev, skb->data,
1005 skb->len, PCI_DMA_TODEVICE); 996 skb->len, PCI_DMA_TODEVICE);
1006 if (pci_dma_mapping_error(pd->pdev, mapping)) { 997 if (pci_dma_mapping_error(pd->pdev, mapping)) {
1007 smsc_warn(TX_ERR, "pci_map_single failed, dropping packet"); 998 netif_warn(pd, tx_err, pd->dev,
999 "pci_map_single failed, dropping packet\n");
1008 return NETDEV_TX_BUSY; 1000 return NETDEV_TX_BUSY;
1009 } 1001 }
1010 1002
@@ -1056,12 +1048,12 @@ static void smsc9420_set_multicast_list(struct net_device *dev)
1056 u32 mac_cr = smsc9420_reg_read(pd, MAC_CR); 1048 u32 mac_cr = smsc9420_reg_read(pd, MAC_CR);
1057 1049
1058 if (dev->flags & IFF_PROMISC) { 1050 if (dev->flags & IFF_PROMISC) {
1059 smsc_dbg(HW, "Promiscuous Mode Enabled"); 1051 netif_dbg(pd, hw, pd->dev, "Promiscuous Mode Enabled\n");
1060 mac_cr |= MAC_CR_PRMS_; 1052 mac_cr |= MAC_CR_PRMS_;
1061 mac_cr &= (~MAC_CR_MCPAS_); 1053 mac_cr &= (~MAC_CR_MCPAS_);
1062 mac_cr &= (~MAC_CR_HPFILT_); 1054 mac_cr &= (~MAC_CR_HPFILT_);
1063 } else if (dev->flags & IFF_ALLMULTI) { 1055 } else if (dev->flags & IFF_ALLMULTI) {
1064 smsc_dbg(HW, "Receive all Multicast Enabled"); 1056 netif_dbg(pd, hw, pd->dev, "Receive all Multicast Enabled\n");
1065 mac_cr &= (~MAC_CR_PRMS_); 1057 mac_cr &= (~MAC_CR_PRMS_);
1066 mac_cr |= MAC_CR_MCPAS_; 1058 mac_cr |= MAC_CR_MCPAS_;
1067 mac_cr &= (~MAC_CR_HPFILT_); 1059 mac_cr &= (~MAC_CR_HPFILT_);
@@ -1069,7 +1061,7 @@ static void smsc9420_set_multicast_list(struct net_device *dev)
1069 struct netdev_hw_addr *ha; 1061 struct netdev_hw_addr *ha;
1070 u32 hash_lo = 0, hash_hi = 0; 1062 u32 hash_lo = 0, hash_hi = 0;
1071 1063
1072 smsc_dbg(HW, "Multicast filter enabled"); 1064 netif_dbg(pd, hw, pd->dev, "Multicast filter enabled\n");
1073 netdev_for_each_mc_addr(ha, dev) { 1065 netdev_for_each_mc_addr(ha, dev) {
1074 u32 bit_num = smsc9420_hash(ha->addr); 1066 u32 bit_num = smsc9420_hash(ha->addr);
1075 u32 mask = 1 << (bit_num & 0x1F); 1067 u32 mask = 1 << (bit_num & 0x1F);
@@ -1087,7 +1079,7 @@ static void smsc9420_set_multicast_list(struct net_device *dev)
1087 mac_cr &= (~MAC_CR_MCPAS_); 1079 mac_cr &= (~MAC_CR_MCPAS_);
1088 mac_cr |= MAC_CR_HPFILT_; 1080 mac_cr |= MAC_CR_HPFILT_;
1089 } else { 1081 } else {
1090 smsc_dbg(HW, "Receive own packets only."); 1082 netif_dbg(pd, hw, pd->dev, "Receive own packets only\n");
1091 smsc9420_reg_write(pd, HASHH, 0); 1083 smsc9420_reg_write(pd, HASHH, 0);
1092 smsc9420_reg_write(pd, HASHL, 0); 1084 smsc9420_reg_write(pd, HASHL, 0);
1093 1085
@@ -1115,11 +1107,11 @@ static void smsc9420_phy_update_flowcontrol(struct smsc9420_pdata *pd)
1115 else 1107 else
1116 flow = 0; 1108 flow = 0;
1117 1109
1118 smsc_info(LINK, "rx pause %s, tx pause %s", 1110 netif_info(pd, link, pd->dev, "rx pause %s, tx pause %s\n",
1119 (cap & FLOW_CTRL_RX ? "enabled" : "disabled"), 1111 cap & FLOW_CTRL_RX ? "enabled" : "disabled",
1120 (cap & FLOW_CTRL_TX ? "enabled" : "disabled")); 1112 cap & FLOW_CTRL_TX ? "enabled" : "disabled");
1121 } else { 1113 } else {
1122 smsc_info(LINK, "half duplex"); 1114 netif_info(pd, link, pd->dev, "half duplex\n");
1123 flow = 0; 1115 flow = 0;
1124 } 1116 }
1125 1117
@@ -1137,10 +1129,10 @@ static void smsc9420_phy_adjust_link(struct net_device *dev)
1137 if (phy_dev->duplex != pd->last_duplex) { 1129 if (phy_dev->duplex != pd->last_duplex) {
1138 u32 mac_cr = smsc9420_reg_read(pd, MAC_CR); 1130 u32 mac_cr = smsc9420_reg_read(pd, MAC_CR);
1139 if (phy_dev->duplex) { 1131 if (phy_dev->duplex) {
1140 smsc_dbg(LINK, "full duplex mode"); 1132 netif_dbg(pd, link, pd->dev, "full duplex mode\n");
1141 mac_cr |= MAC_CR_FDPX_; 1133 mac_cr |= MAC_CR_FDPX_;
1142 } else { 1134 } else {
1143 smsc_dbg(LINK, "half duplex mode"); 1135 netif_dbg(pd, link, pd->dev, "half duplex mode\n");
1144 mac_cr &= ~MAC_CR_FDPX_; 1136 mac_cr &= ~MAC_CR_FDPX_;
1145 } 1137 }
1146 smsc9420_reg_write(pd, MAC_CR, mac_cr); 1138 smsc9420_reg_write(pd, MAC_CR, mac_cr);
@@ -1152,9 +1144,9 @@ static void smsc9420_phy_adjust_link(struct net_device *dev)
1152 carrier = netif_carrier_ok(dev); 1144 carrier = netif_carrier_ok(dev);
1153 if (carrier != pd->last_carrier) { 1145 if (carrier != pd->last_carrier) {
1154 if (carrier) 1146 if (carrier)
1155 smsc_dbg(LINK, "carrier OK"); 1147 netif_dbg(pd, link, pd->dev, "carrier OK\n");
1156 else 1148 else
1157 smsc_dbg(LINK, "no carrier"); 1149 netif_dbg(pd, link, pd->dev, "no carrier\n");
1158 pd->last_carrier = carrier; 1150 pd->last_carrier = carrier;
1159 } 1151 }
1160} 1152}
@@ -1168,24 +1160,24 @@ static int smsc9420_mii_probe(struct net_device *dev)
1168 1160
1169 /* Device only supports internal PHY at address 1 */ 1161 /* Device only supports internal PHY at address 1 */
1170 if (!pd->mii_bus->phy_map[1]) { 1162 if (!pd->mii_bus->phy_map[1]) {
1171 pr_err("%s: no PHY found at address 1\n", dev->name); 1163 netdev_err(dev, "no PHY found at address 1\n");
1172 return -ENODEV; 1164 return -ENODEV;
1173 } 1165 }
1174 1166
1175 phydev = pd->mii_bus->phy_map[1]; 1167 phydev = pd->mii_bus->phy_map[1];
1176 smsc_info(PROBE, "PHY addr %d, phy_id 0x%08X", phydev->addr, 1168 netif_info(pd, probe, pd->dev, "PHY addr %d, phy_id 0x%08X\n",
1177 phydev->phy_id); 1169 phydev->addr, phydev->phy_id);
1178 1170
1179 phydev = phy_connect(dev, dev_name(&phydev->dev), 1171 phydev = phy_connect(dev, dev_name(&phydev->dev),
1180 smsc9420_phy_adjust_link, PHY_INTERFACE_MODE_MII); 1172 smsc9420_phy_adjust_link, PHY_INTERFACE_MODE_MII);
1181 1173
1182 if (IS_ERR(phydev)) { 1174 if (IS_ERR(phydev)) {
1183 pr_err("%s: Could not attach to PHY\n", dev->name); 1175 netdev_err(dev, "Could not attach to PHY\n");
1184 return PTR_ERR(phydev); 1176 return PTR_ERR(phydev);
1185 } 1177 }
1186 1178
1187 pr_info("%s: attached PHY driver [%s] (mii_bus:phy_addr=%s, irq=%d)\n", 1179 netdev_info(dev, "attached PHY driver [%s] (mii_bus:phy_addr=%s, irq=%d)\n",
1188 dev->name, phydev->drv->name, dev_name(&phydev->dev), phydev->irq); 1180 phydev->drv->name, dev_name(&phydev->dev), phydev->irq);
1189 1181
1190 /* mask with MAC supported features */ 1182 /* mask with MAC supported features */
1191 phydev->supported &= (PHY_BASIC_FEATURES | SUPPORTED_Pause | 1183 phydev->supported &= (PHY_BASIC_FEATURES | SUPPORTED_Pause |
@@ -1223,12 +1215,12 @@ static int smsc9420_mii_init(struct net_device *dev)
1223 pd->mii_bus->phy_mask = ~(1 << 1); 1215 pd->mii_bus->phy_mask = ~(1 << 1);
1224 1216
1225 if (mdiobus_register(pd->mii_bus)) { 1217 if (mdiobus_register(pd->mii_bus)) {
1226 smsc_warn(PROBE, "Error registering mii bus"); 1218 netif_warn(pd, probe, pd->dev, "Error registering mii bus\n");
1227 goto err_out_free_bus_2; 1219 goto err_out_free_bus_2;
1228 } 1220 }
1229 1221
1230 if (smsc9420_mii_probe(dev) < 0) { 1222 if (smsc9420_mii_probe(dev) < 0) {
1231 smsc_warn(PROBE, "Error probing mii bus"); 1223 netif_warn(pd, probe, pd->dev, "Error probing mii bus\n");
1232 goto err_out_unregister_bus_3; 1224 goto err_out_unregister_bus_3;
1233 } 1225 }
1234 1226
@@ -1281,12 +1273,11 @@ static int smsc9420_alloc_rx_ring(struct smsc9420_pdata *pd)
1281 1273
1282 BUG_ON(!pd->rx_ring); 1274 BUG_ON(!pd->rx_ring);
1283 1275
1284 pd->rx_buffers = kmalloc((sizeof(struct smsc9420_ring_info) * 1276 pd->rx_buffers = kmalloc_array(RX_RING_SIZE,
1285 RX_RING_SIZE), GFP_KERNEL); 1277 sizeof(struct smsc9420_ring_info),
1286 if (pd->rx_buffers == NULL) { 1278 GFP_KERNEL);
1287 smsc_warn(IFUP, "Failed to allocated rx_buffers"); 1279 if (pd->rx_buffers == NULL)
1288 goto out; 1280 goto out;
1289 }
1290 1281
1291 /* initialize the rx ring */ 1282 /* initialize the rx ring */
1292 for (i = 0; i < RX_RING_SIZE; i++) { 1283 for (i = 0; i < RX_RING_SIZE; i++) {
@@ -1301,7 +1292,8 @@ static int smsc9420_alloc_rx_ring(struct smsc9420_pdata *pd)
1301 /* now allocate the entire ring of skbs */ 1292 /* now allocate the entire ring of skbs */
1302 for (i = 0; i < RX_RING_SIZE; i++) { 1293 for (i = 0; i < RX_RING_SIZE; i++) {
1303 if (smsc9420_alloc_rx_buffer(pd, i)) { 1294 if (smsc9420_alloc_rx_buffer(pd, i)) {
1304 smsc_warn(IFUP, "failed to allocate rx skb %d", i); 1295 netif_warn(pd, ifup, pd->dev,
1296 "failed to allocate rx skb %d\n", i);
1305 goto out_free_rx_skbs; 1297 goto out_free_rx_skbs;
1306 } 1298 }
1307 } 1299 }
@@ -1310,13 +1302,14 @@ static int smsc9420_alloc_rx_ring(struct smsc9420_pdata *pd)
1310 pd->rx_ring_tail = 0; 1302 pd->rx_ring_tail = 0;
1311 1303
1312 smsc9420_reg_write(pd, VLAN1, ETH_P_8021Q); 1304 smsc9420_reg_write(pd, VLAN1, ETH_P_8021Q);
1313 smsc_dbg(IFUP, "VLAN1 = 0x%08x", smsc9420_reg_read(pd, VLAN1)); 1305 netif_dbg(pd, ifup, pd->dev, "VLAN1 = 0x%08x\n",
1306 smsc9420_reg_read(pd, VLAN1));
1314 1307
1315 if (pd->rx_csum) { 1308 if (pd->rx_csum) {
1316 /* Enable RX COE */ 1309 /* Enable RX COE */
1317 u32 coe = smsc9420_reg_read(pd, COE_CR) | RX_COE_EN; 1310 u32 coe = smsc9420_reg_read(pd, COE_CR) | RX_COE_EN;
1318 smsc9420_reg_write(pd, COE_CR, coe); 1311 smsc9420_reg_write(pd, COE_CR, coe);
1319 smsc_dbg(IFUP, "COE_CR = 0x%08x", coe); 1312 netif_dbg(pd, ifup, pd->dev, "COE_CR = 0x%08x\n", coe);
1320 } 1313 }
1321 1314
1322 smsc9420_reg_write(pd, RX_BASE_ADDR, pd->rx_dma_addr); 1315 smsc9420_reg_write(pd, RX_BASE_ADDR, pd->rx_dma_addr);
@@ -1339,7 +1332,8 @@ static int smsc9420_open(struct net_device *dev)
1339 int result = 0, timeout; 1332 int result = 0, timeout;
1340 1333
1341 if (!is_valid_ether_addr(dev->dev_addr)) { 1334 if (!is_valid_ether_addr(dev->dev_addr)) {
1342 smsc_warn(IFUP, "dev_addr is not a valid MAC address"); 1335 netif_warn(pd, ifup, pd->dev,
1336 "dev_addr is not a valid MAC address\n");
1343 result = -EADDRNOTAVAIL; 1337 result = -EADDRNOTAVAIL;
1344 goto out_0; 1338 goto out_0;
1345 } 1339 }
@@ -1358,7 +1352,7 @@ static int smsc9420_open(struct net_device *dev)
1358 1352
1359 result = request_irq(irq, smsc9420_isr, IRQF_SHARED, DRV_NAME, pd); 1353 result = request_irq(irq, smsc9420_isr, IRQF_SHARED, DRV_NAME, pd);
1360 if (result) { 1354 if (result) {
1361 smsc_warn(IFUP, "Unable to use IRQ = %d", irq); 1355 netif_warn(pd, ifup, pd->dev, "Unable to use IRQ = %d\n", irq);
1362 result = -ENODEV; 1356 result = -ENODEV;
1363 goto out_0; 1357 goto out_0;
1364 } 1358 }
@@ -1393,7 +1387,7 @@ static int smsc9420_open(struct net_device *dev)
1393 smsc9420_pci_flush_write(pd); 1387 smsc9420_pci_flush_write(pd);
1394 1388
1395 /* test the IRQ connection to the ISR */ 1389 /* test the IRQ connection to the ISR */
1396 smsc_dbg(IFUP, "Testing ISR using IRQ %d", irq); 1390 netif_dbg(pd, ifup, pd->dev, "Testing ISR using IRQ %d\n", irq);
1397 pd->software_irq_signal = false; 1391 pd->software_irq_signal = false;
1398 1392
1399 spin_lock_irqsave(&pd->int_lock, flags); 1393 spin_lock_irqsave(&pd->int_lock, flags);
@@ -1423,30 +1417,32 @@ static int smsc9420_open(struct net_device *dev)
1423 spin_unlock_irqrestore(&pd->int_lock, flags); 1417 spin_unlock_irqrestore(&pd->int_lock, flags);
1424 1418
1425 if (!pd->software_irq_signal) { 1419 if (!pd->software_irq_signal) {
1426 smsc_warn(IFUP, "ISR failed signaling test"); 1420 netif_warn(pd, ifup, pd->dev, "ISR failed signaling test\n");
1427 result = -ENODEV; 1421 result = -ENODEV;
1428 goto out_free_irq_1; 1422 goto out_free_irq_1;
1429 } 1423 }
1430 1424
1431 smsc_dbg(IFUP, "ISR passed test using IRQ %d", irq); 1425 netif_dbg(pd, ifup, pd->dev, "ISR passed test using IRQ %d\n", irq);
1432 1426
1433 result = smsc9420_alloc_tx_ring(pd); 1427 result = smsc9420_alloc_tx_ring(pd);
1434 if (result) { 1428 if (result) {
1435 smsc_warn(IFUP, "Failed to Initialize tx dma ring"); 1429 netif_warn(pd, ifup, pd->dev,
1430 "Failed to Initialize tx dma ring\n");
1436 result = -ENOMEM; 1431 result = -ENOMEM;
1437 goto out_free_irq_1; 1432 goto out_free_irq_1;
1438 } 1433 }
1439 1434
1440 result = smsc9420_alloc_rx_ring(pd); 1435 result = smsc9420_alloc_rx_ring(pd);
1441 if (result) { 1436 if (result) {
1442 smsc_warn(IFUP, "Failed to Initialize rx dma ring"); 1437 netif_warn(pd, ifup, pd->dev,
1438 "Failed to Initialize rx dma ring\n");
1443 result = -ENOMEM; 1439 result = -ENOMEM;
1444 goto out_free_tx_ring_2; 1440 goto out_free_tx_ring_2;
1445 } 1441 }
1446 1442
1447 result = smsc9420_mii_init(dev); 1443 result = smsc9420_mii_init(dev);
1448 if (result) { 1444 if (result) {
1449 smsc_warn(IFUP, "Failed to initialize Phy"); 1445 netif_warn(pd, ifup, pd->dev, "Failed to initialize Phy\n");
1450 result = -ENODEV; 1446 result = -ENODEV;
1451 goto out_free_rx_ring_3; 1447 goto out_free_rx_ring_3;
1452 } 1448 }
@@ -1547,7 +1543,8 @@ static int smsc9420_resume(struct pci_dev *pdev)
1547 1543
1548 err = pci_enable_wake(pdev, 0, 0); 1544 err = pci_enable_wake(pdev, 0, 0);
1549 if (err) 1545 if (err)
1550 smsc_warn(IFUP, "pci_enable_wake failed: %d", err); 1546 netif_warn(pd, ifup, pd->dev, "pci_enable_wake failed: %d\n",
1547 err);
1551 1548
1552 if (netif_running(dev)) { 1549 if (netif_running(dev)) {
1553 /* FIXME: gross. It looks like ancient PM relic.*/ 1550 /* FIXME: gross. It looks like ancient PM relic.*/
@@ -1582,12 +1579,12 @@ smsc9420_probe(struct pci_dev *pdev, const struct pci_device_id *id)
1582 int result = 0; 1579 int result = 0;
1583 u32 id_rev; 1580 u32 id_rev;
1584 1581
1585 printk(KERN_INFO DRV_DESCRIPTION " version " DRV_VERSION "\n"); 1582 pr_info("%s version %s\n", DRV_DESCRIPTION, DRV_VERSION);
1586 1583
1587 /* First do the PCI initialisation */ 1584 /* First do the PCI initialisation */
1588 result = pci_enable_device(pdev); 1585 result = pci_enable_device(pdev);
1589 if (unlikely(result)) { 1586 if (unlikely(result)) {
1590 printk(KERN_ERR "Cannot enable smsc9420\n"); 1587 pr_err("Cannot enable smsc9420\n");
1591 goto out_0; 1588 goto out_0;
1592 } 1589 }
1593 1590
@@ -1600,24 +1597,24 @@ smsc9420_probe(struct pci_dev *pdev, const struct pci_device_id *id)
1600 SET_NETDEV_DEV(dev, &pdev->dev); 1597 SET_NETDEV_DEV(dev, &pdev->dev);
1601 1598
1602 if (!(pci_resource_flags(pdev, SMSC_BAR) & IORESOURCE_MEM)) { 1599 if (!(pci_resource_flags(pdev, SMSC_BAR) & IORESOURCE_MEM)) {
1603 printk(KERN_ERR "Cannot find PCI device base address\n"); 1600 netdev_err(dev, "Cannot find PCI device base address\n");
1604 goto out_free_netdev_2; 1601 goto out_free_netdev_2;
1605 } 1602 }
1606 1603
1607 if ((pci_request_regions(pdev, DRV_NAME))) { 1604 if ((pci_request_regions(pdev, DRV_NAME))) {
1608 printk(KERN_ERR "Cannot obtain PCI resources, aborting.\n"); 1605 netdev_err(dev, "Cannot obtain PCI resources, aborting\n");
1609 goto out_free_netdev_2; 1606 goto out_free_netdev_2;
1610 } 1607 }
1611 1608
1612 if (pci_set_dma_mask(pdev, DMA_BIT_MASK(32))) { 1609 if (pci_set_dma_mask(pdev, DMA_BIT_MASK(32))) {
1613 printk(KERN_ERR "No usable DMA configuration, aborting.\n"); 1610 netdev_err(dev, "No usable DMA configuration, aborting\n");
1614 goto out_free_regions_3; 1611 goto out_free_regions_3;
1615 } 1612 }
1616 1613
1617 virt_addr = ioremap(pci_resource_start(pdev, SMSC_BAR), 1614 virt_addr = ioremap(pci_resource_start(pdev, SMSC_BAR),
1618 pci_resource_len(pdev, SMSC_BAR)); 1615 pci_resource_len(pdev, SMSC_BAR));
1619 if (!virt_addr) { 1616 if (!virt_addr) {
1620 printk(KERN_ERR "Cannot map device registers, aborting.\n"); 1617 netdev_err(dev, "Cannot map device registers, aborting\n");
1621 goto out_free_regions_3; 1618 goto out_free_regions_3;
1622 } 1619 }
1623 1620
@@ -1646,16 +1643,17 @@ smsc9420_probe(struct pci_dev *pdev, const struct pci_device_id *id)
1646 pd->msg_enable = smsc_debug; 1643 pd->msg_enable = smsc_debug;
1647 pd->rx_csum = true; 1644 pd->rx_csum = true;
1648 1645
1649 smsc_dbg(PROBE, "lan_base=0x%08lx", (ulong)virt_addr); 1646 netif_dbg(pd, probe, pd->dev, "lan_base=0x%08lx\n", (ulong)virt_addr);
1650 1647
1651 id_rev = smsc9420_reg_read(pd, ID_REV); 1648 id_rev = smsc9420_reg_read(pd, ID_REV);
1652 switch (id_rev & 0xFFFF0000) { 1649 switch (id_rev & 0xFFFF0000) {
1653 case 0x94200000: 1650 case 0x94200000:
1654 smsc_info(PROBE, "LAN9420 identified, ID_REV=0x%08X", id_rev); 1651 netif_info(pd, probe, pd->dev,
1652 "LAN9420 identified, ID_REV=0x%08X\n", id_rev);
1655 break; 1653 break;
1656 default: 1654 default:
1657 smsc_warn(PROBE, "LAN9420 NOT identified"); 1655 netif_warn(pd, probe, pd->dev, "LAN9420 NOT identified\n");
1658 smsc_warn(PROBE, "ID_REV=0x%08X", id_rev); 1656 netif_warn(pd, probe, pd->dev, "ID_REV=0x%08X\n", id_rev);
1659 goto out_free_dmadesc_5; 1657 goto out_free_dmadesc_5;
1660 } 1658 }
1661 1659
@@ -1670,7 +1668,8 @@ smsc9420_probe(struct pci_dev *pdev, const struct pci_device_id *id)
1670 1668
1671 result = register_netdev(dev); 1669 result = register_netdev(dev);
1672 if (result) { 1670 if (result) {
1673 smsc_warn(PROBE, "error %i registering device", result); 1671 netif_warn(pd, probe, pd->dev, "error %i registering device\n",
1672 result);
1674 goto out_free_dmadesc_5; 1673 goto out_free_dmadesc_5;
1675 } 1674 }
1676 1675
@@ -1707,8 +1706,6 @@ static void smsc9420_remove(struct pci_dev *pdev)
1707 if (!dev) 1706 if (!dev)
1708 return; 1707 return;
1709 1708
1710 pci_set_drvdata(pdev, NULL);
1711
1712 pd = netdev_priv(dev); 1709 pd = netdev_priv(dev);
1713 unregister_netdev(dev); 1710 unregister_netdev(dev);
1714 1711
diff --git a/drivers/net/ethernet/stmicro/stmmac/common.h b/drivers/net/ethernet/stmicro/stmmac/common.h
index 7eb8babed2cb..fc94f202a43e 100644
--- a/drivers/net/ethernet/stmicro/stmmac/common.h
+++ b/drivers/net/ethernet/stmicro/stmmac/common.h
@@ -451,14 +451,14 @@ struct mac_device_info {
451struct mac_device_info *dwmac1000_setup(void __iomem *ioaddr); 451struct mac_device_info *dwmac1000_setup(void __iomem *ioaddr);
452struct mac_device_info *dwmac100_setup(void __iomem *ioaddr); 452struct mac_device_info *dwmac100_setup(void __iomem *ioaddr);
453 453
454extern void stmmac_set_mac_addr(void __iomem *ioaddr, u8 addr[6], 454void stmmac_set_mac_addr(void __iomem *ioaddr, u8 addr[6],
455 unsigned int high, unsigned int low); 455 unsigned int high, unsigned int low);
456extern void stmmac_get_mac_addr(void __iomem *ioaddr, unsigned char *addr, 456void stmmac_get_mac_addr(void __iomem *ioaddr, unsigned char *addr,
457 unsigned int high, unsigned int low); 457 unsigned int high, unsigned int low);
458 458
459extern void stmmac_set_mac(void __iomem *ioaddr, bool enable); 459void stmmac_set_mac(void __iomem *ioaddr, bool enable);
460 460
461extern void dwmac_dma_flush_tx_fifo(void __iomem *ioaddr); 461void dwmac_dma_flush_tx_fifo(void __iomem *ioaddr);
462extern const struct stmmac_ring_mode_ops ring_mode_ops; 462extern const struct stmmac_ring_mode_ops ring_mode_ops;
463extern const struct stmmac_chain_mode_ops chain_mode_ops; 463extern const struct stmmac_chain_mode_ops chain_mode_ops;
464 464
diff --git a/drivers/net/ethernet/stmicro/stmmac/dwmac_dma.h b/drivers/net/ethernet/stmicro/stmmac/dwmac_dma.h
index 8e5662ce488b..def266da55db 100644
--- a/drivers/net/ethernet/stmicro/stmmac/dwmac_dma.h
+++ b/drivers/net/ethernet/stmicro/stmmac/dwmac_dma.h
@@ -104,14 +104,13 @@
104#define DMA_STATUS_TI 0x00000001 /* Transmit Interrupt */ 104#define DMA_STATUS_TI 0x00000001 /* Transmit Interrupt */
105#define DMA_CONTROL_FTF 0x00100000 /* Flush transmit FIFO */ 105#define DMA_CONTROL_FTF 0x00100000 /* Flush transmit FIFO */
106 106
107extern void dwmac_enable_dma_transmission(void __iomem *ioaddr); 107void dwmac_enable_dma_transmission(void __iomem *ioaddr);
108extern void dwmac_enable_dma_irq(void __iomem *ioaddr); 108void dwmac_enable_dma_irq(void __iomem *ioaddr);
109extern void dwmac_disable_dma_irq(void __iomem *ioaddr); 109void dwmac_disable_dma_irq(void __iomem *ioaddr);
110extern void dwmac_dma_start_tx(void __iomem *ioaddr); 110void dwmac_dma_start_tx(void __iomem *ioaddr);
111extern void dwmac_dma_stop_tx(void __iomem *ioaddr); 111void dwmac_dma_stop_tx(void __iomem *ioaddr);
112extern void dwmac_dma_start_rx(void __iomem *ioaddr); 112void dwmac_dma_start_rx(void __iomem *ioaddr);
113extern void dwmac_dma_stop_rx(void __iomem *ioaddr); 113void dwmac_dma_stop_rx(void __iomem *ioaddr);
114extern int dwmac_dma_interrupt(void __iomem *ioaddr, 114int dwmac_dma_interrupt(void __iomem *ioaddr, struct stmmac_extra_stats *x);
115 struct stmmac_extra_stats *x);
116 115
117#endif /* __DWMAC_DMA_H__ */ 116#endif /* __DWMAC_DMA_H__ */
diff --git a/drivers/net/ethernet/stmicro/stmmac/mmc.h b/drivers/net/ethernet/stmicro/stmmac/mmc.h
index 48ec001566b5..8607488cbcfc 100644
--- a/drivers/net/ethernet/stmicro/stmmac/mmc.h
+++ b/drivers/net/ethernet/stmicro/stmmac/mmc.h
@@ -128,8 +128,8 @@ struct stmmac_counters {
128 unsigned int mmc_rx_icmp_err_octets; 128 unsigned int mmc_rx_icmp_err_octets;
129}; 129};
130 130
131extern void dwmac_mmc_ctrl(void __iomem *ioaddr, unsigned int mode); 131void dwmac_mmc_ctrl(void __iomem *ioaddr, unsigned int mode);
132extern void dwmac_mmc_intr_all_mask(void __iomem *ioaddr); 132void dwmac_mmc_intr_all_mask(void __iomem *ioaddr);
133extern void dwmac_mmc_read(void __iomem *ioaddr, struct stmmac_counters *mmc); 133void dwmac_mmc_read(void __iomem *ioaddr, struct stmmac_counters *mmc);
134 134
135#endif /* __MMC_H__ */ 135#endif /* __MMC_H__ */
diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac.h b/drivers/net/ethernet/stmicro/stmmac/stmmac.h
index f16a9bdf45bb..22f89ffdfd95 100644
--- a/drivers/net/ethernet/stmicro/stmmac/stmmac.h
+++ b/drivers/net/ethernet/stmicro/stmmac/stmmac.h
@@ -110,14 +110,14 @@ struct stmmac_priv {
110 110
111extern int phyaddr; 111extern int phyaddr;
112 112
113extern int stmmac_mdio_unregister(struct net_device *ndev); 113int stmmac_mdio_unregister(struct net_device *ndev);
114extern int stmmac_mdio_register(struct net_device *ndev); 114int stmmac_mdio_register(struct net_device *ndev);
115extern void stmmac_set_ethtool_ops(struct net_device *netdev); 115void stmmac_set_ethtool_ops(struct net_device *netdev);
116extern const struct stmmac_desc_ops enh_desc_ops; 116extern const struct stmmac_desc_ops enh_desc_ops;
117extern const struct stmmac_desc_ops ndesc_ops; 117extern const struct stmmac_desc_ops ndesc_ops;
118extern const struct stmmac_hwtimestamp stmmac_ptp; 118extern const struct stmmac_hwtimestamp stmmac_ptp;
119extern int stmmac_ptp_register(struct stmmac_priv *priv); 119int stmmac_ptp_register(struct stmmac_priv *priv);
120extern void stmmac_ptp_unregister(struct stmmac_priv *priv); 120void stmmac_ptp_unregister(struct stmmac_priv *priv);
121int stmmac_freeze(struct net_device *ndev); 121int stmmac_freeze(struct net_device *ndev);
122int stmmac_restore(struct net_device *ndev); 122int stmmac_restore(struct net_device *ndev);
123int stmmac_resume(struct net_device *ndev); 123int stmmac_resume(struct net_device *ndev);
diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_pci.c b/drivers/net/ethernet/stmicro/stmmac/stmmac_pci.c
index 023b7c29cb2f..644d80ece067 100644
--- a/drivers/net/ethernet/stmicro/stmmac/stmmac_pci.c
+++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_pci.c
@@ -138,7 +138,6 @@ static void stmmac_pci_remove(struct pci_dev *pdev)
138 138
139 stmmac_dvr_remove(ndev); 139 stmmac_dvr_remove(ndev);
140 140
141 pci_set_drvdata(pdev, NULL);
142 pci_iounmap(pdev, priv->ioaddr); 141 pci_iounmap(pdev, priv->ioaddr);
143 pci_release_regions(pdev); 142 pci_release_regions(pdev);
144 pci_disable_device(pdev); 143 pci_disable_device(pdev);
diff --git a/drivers/net/ethernet/sun/cassini.c b/drivers/net/ethernet/sun/cassini.c
index 759441b29e53..b4d50d74ba18 100644
--- a/drivers/net/ethernet/sun/cassini.c
+++ b/drivers/net/ethernet/sun/cassini.c
@@ -3354,7 +3354,7 @@ use_random_mac_addr:
3354#if defined(CONFIG_SPARC) 3354#if defined(CONFIG_SPARC)
3355 addr = of_get_property(cp->of_node, "local-mac-address", NULL); 3355 addr = of_get_property(cp->of_node, "local-mac-address", NULL);
3356 if (addr != NULL) { 3356 if (addr != NULL) {
3357 memcpy(dev_addr, addr, 6); 3357 memcpy(dev_addr, addr, ETH_ALEN);
3358 goto done; 3358 goto done;
3359 } 3359 }
3360#endif 3360#endif
@@ -5168,7 +5168,6 @@ err_out_free_netdev:
5168 5168
5169err_out_disable_pdev: 5169err_out_disable_pdev:
5170 pci_disable_device(pdev); 5170 pci_disable_device(pdev);
5171 pci_set_drvdata(pdev, NULL);
5172 return -ENODEV; 5171 return -ENODEV;
5173} 5172}
5174 5173
@@ -5206,7 +5205,6 @@ static void cas_remove_one(struct pci_dev *pdev)
5206 free_netdev(dev); 5205 free_netdev(dev);
5207 pci_release_regions(pdev); 5206 pci_release_regions(pdev);
5208 pci_disable_device(pdev); 5207 pci_disable_device(pdev);
5209 pci_set_drvdata(pdev, NULL);
5210} 5208}
5211 5209
5212#ifdef CONFIG_PM 5210#ifdef CONFIG_PM
diff --git a/drivers/net/ethernet/sun/niu.c b/drivers/net/ethernet/sun/niu.c
index f28460ce24a7..388540fcb977 100644
--- a/drivers/net/ethernet/sun/niu.c
+++ b/drivers/net/ethernet/sun/niu.c
@@ -9875,7 +9875,6 @@ err_out_free_res:
9875 9875
9876err_out_disable_pdev: 9876err_out_disable_pdev:
9877 pci_disable_device(pdev); 9877 pci_disable_device(pdev);
9878 pci_set_drvdata(pdev, NULL);
9879 9878
9880 return err; 9879 return err;
9881} 9880}
@@ -9900,7 +9899,6 @@ static void niu_pci_remove_one(struct pci_dev *pdev)
9900 free_netdev(dev); 9899 free_netdev(dev);
9901 pci_release_regions(pdev); 9900 pci_release_regions(pdev);
9902 pci_disable_device(pdev); 9901 pci_disable_device(pdev);
9903 pci_set_drvdata(pdev, NULL);
9904 } 9902 }
9905} 9903}
9906 9904
diff --git a/drivers/net/ethernet/sun/sungem.c b/drivers/net/ethernet/sun/sungem.c
index e62df2b81302..b5655b79bd3b 100644
--- a/drivers/net/ethernet/sun/sungem.c
+++ b/drivers/net/ethernet/sun/sungem.c
@@ -2779,7 +2779,7 @@ static int gem_get_device_address(struct gem *gp)
2779 return -1; 2779 return -1;
2780#endif 2780#endif
2781 } 2781 }
2782 memcpy(dev->dev_addr, addr, 6); 2782 memcpy(dev->dev_addr, addr, ETH_ALEN);
2783#else 2783#else
2784 get_gem_mac_nonobp(gp->pdev, gp->dev->dev_addr); 2784 get_gem_mac_nonobp(gp->pdev, gp->dev->dev_addr);
2785#endif 2785#endif
@@ -2806,8 +2806,6 @@ static void gem_remove_one(struct pci_dev *pdev)
2806 iounmap(gp->regs); 2806 iounmap(gp->regs);
2807 pci_release_regions(pdev); 2807 pci_release_regions(pdev);
2808 free_netdev(dev); 2808 free_netdev(dev);
2809
2810 pci_set_drvdata(pdev, NULL);
2811 } 2809 }
2812} 2810}
2813 2811
diff --git a/drivers/net/ethernet/sun/sunhme.c b/drivers/net/ethernet/sun/sunhme.c
index e37b587b3860..0dbf46f08ed5 100644
--- a/drivers/net/ethernet/sun/sunhme.c
+++ b/drivers/net/ethernet/sun/sunhme.c
@@ -2675,10 +2675,10 @@ static int happy_meal_sbus_probe_one(struct platform_device *op, int is_qfe)
2675 2675
2676 addr = of_get_property(dp, "local-mac-address", &len); 2676 addr = of_get_property(dp, "local-mac-address", &len);
2677 2677
2678 if (qfe_slot != -1 && addr && len == 6) 2678 if (qfe_slot != -1 && addr && len == ETH_ALEN)
2679 memcpy(dev->dev_addr, addr, 6); 2679 memcpy(dev->dev_addr, addr, ETH_ALEN);
2680 else 2680 else
2681 memcpy(dev->dev_addr, idprom->id_ethaddr, 6); 2681 memcpy(dev->dev_addr, idprom->id_ethaddr, ETH_ALEN);
2682 } 2682 }
2683 2683
2684 hp = netdev_priv(dev); 2684 hp = netdev_priv(dev);
@@ -3024,9 +3024,9 @@ static int happy_meal_pci_probe(struct pci_dev *pdev,
3024 (addr = of_get_property(dp, "local-mac-address", &len)) 3024 (addr = of_get_property(dp, "local-mac-address", &len))
3025 != NULL && 3025 != NULL &&
3026 len == 6) { 3026 len == 6) {
3027 memcpy(dev->dev_addr, addr, 6); 3027 memcpy(dev->dev_addr, addr, ETH_ALEN);
3028 } else { 3028 } else {
3029 memcpy(dev->dev_addr, idprom->id_ethaddr, 6); 3029 memcpy(dev->dev_addr, idprom->id_ethaddr, ETH_ALEN);
3030 } 3030 }
3031#else 3031#else
3032 get_hme_mac_nonsparc(pdev, &dev->dev_addr[0]); 3032 get_hme_mac_nonsparc(pdev, &dev->dev_addr[0]);
@@ -3170,8 +3170,6 @@ static void happy_meal_pci_remove(struct pci_dev *pdev)
3170 pci_release_regions(hp->happy_dev); 3170 pci_release_regions(hp->happy_dev);
3171 3171
3172 free_netdev(net_dev); 3172 free_netdev(net_dev);
3173
3174 pci_set_drvdata(pdev, NULL);
3175} 3173}
3176 3174
3177static DEFINE_PCI_DEVICE_TABLE(happymeal_pci_ids) = { 3175static DEFINE_PCI_DEVICE_TABLE(happymeal_pci_ids) = {
diff --git a/drivers/net/ethernet/sun/sunqe.c b/drivers/net/ethernet/sun/sunqe.c
index b072f4dba033..5695ae2411de 100644
--- a/drivers/net/ethernet/sun/sunqe.c
+++ b/drivers/net/ethernet/sun/sunqe.c
@@ -843,7 +843,7 @@ static int qec_ether_init(struct platform_device *op)
843 if (!dev) 843 if (!dev)
844 return -ENOMEM; 844 return -ENOMEM;
845 845
846 memcpy(dev->dev_addr, idprom->id_ethaddr, 6); 846 memcpy(dev->dev_addr, idprom->id_ethaddr, ETH_ALEN);
847 847
848 qe = netdev_priv(dev); 848 qe = netdev_priv(dev);
849 849
diff --git a/drivers/net/ethernet/tehuti/tehuti.c b/drivers/net/ethernet/tehuti/tehuti.c
index 571452e786d5..dd0dd6279b4e 100644
--- a/drivers/net/ethernet/tehuti/tehuti.c
+++ b/drivers/net/ethernet/tehuti/tehuti.c
@@ -2447,7 +2447,6 @@ static void bdx_remove(struct pci_dev *pdev)
2447 iounmap(nic->regs); 2447 iounmap(nic->regs);
2448 pci_release_regions(pdev); 2448 pci_release_regions(pdev);
2449 pci_disable_device(pdev); 2449 pci_disable_device(pdev);
2450 pci_set_drvdata(pdev, NULL);
2451 vfree(nic); 2450 vfree(nic);
2452 2451
2453 RET(); 2452 RET();
diff --git a/drivers/net/ethernet/ti/Kconfig b/drivers/net/ethernet/ti/Kconfig
index de71b1ec4625..53150c25a96b 100644
--- a/drivers/net/ethernet/ti/Kconfig
+++ b/drivers/net/ethernet/ti/Kconfig
@@ -49,11 +49,19 @@ config TI_DAVINCI_CPDMA
49 To compile this driver as a module, choose M here: the module 49 To compile this driver as a module, choose M here: the module
50 will be called davinci_cpdma. This is recommended. 50 will be called davinci_cpdma. This is recommended.
51 51
52config TI_CPSW_PHY_SEL
53 boolean "TI CPSW Switch Phy sel Support"
54 depends on TI_CPSW
55 ---help---
56 This driver supports configuring of the phy mode connected to
57 the CPSW.
58
52config TI_CPSW 59config TI_CPSW
53 tristate "TI CPSW Switch Support" 60 tristate "TI CPSW Switch Support"
54 depends on ARM && (ARCH_DAVINCI || SOC_AM33XX) 61 depends on ARM && (ARCH_DAVINCI || SOC_AM33XX)
55 select TI_DAVINCI_CPDMA 62 select TI_DAVINCI_CPDMA
56 select TI_DAVINCI_MDIO 63 select TI_DAVINCI_MDIO
64 select TI_CPSW_PHY_SEL
57 ---help--- 65 ---help---
58 This driver supports TI's CPSW Ethernet Switch. 66 This driver supports TI's CPSW Ethernet Switch.
59 67
diff --git a/drivers/net/ethernet/ti/Makefile b/drivers/net/ethernet/ti/Makefile
index c65148e8aa1d..9cfaab8152be 100644
--- a/drivers/net/ethernet/ti/Makefile
+++ b/drivers/net/ethernet/ti/Makefile
@@ -7,5 +7,6 @@ obj-$(CONFIG_CPMAC) += cpmac.o
7obj-$(CONFIG_TI_DAVINCI_EMAC) += davinci_emac.o 7obj-$(CONFIG_TI_DAVINCI_EMAC) += davinci_emac.o
8obj-$(CONFIG_TI_DAVINCI_MDIO) += davinci_mdio.o 8obj-$(CONFIG_TI_DAVINCI_MDIO) += davinci_mdio.o
9obj-$(CONFIG_TI_DAVINCI_CPDMA) += davinci_cpdma.o 9obj-$(CONFIG_TI_DAVINCI_CPDMA) += davinci_cpdma.o
10obj-$(CONFIG_TI_CPSW_PHY_SEL) += cpsw-phy-sel.o
10obj-$(CONFIG_TI_CPSW) += ti_cpsw.o 11obj-$(CONFIG_TI_CPSW) += ti_cpsw.o
11ti_cpsw-y := cpsw_ale.o cpsw.o cpts.o 12ti_cpsw-y := cpsw_ale.o cpsw.o cpts.o
diff --git a/drivers/net/ethernet/ti/cpsw-phy-sel.c b/drivers/net/ethernet/ti/cpsw-phy-sel.c
new file mode 100644
index 000000000000..148da9ae8366
--- /dev/null
+++ b/drivers/net/ethernet/ti/cpsw-phy-sel.c
@@ -0,0 +1,161 @@
1/* Texas Instruments Ethernet Switch Driver
2 *
3 * Copyright (C) 2013 Texas Instruments
4 *
5 * This program is free software; you can redistribute it and/or
6 * modify it under the terms of the GNU General Public License
7 * version 2 as published by the Free Software Foundation.
8 *
9 * This program is distributed "as is" WITHOUT ANY WARRANTY of any
10 * kind, whether express or implied; without even the implied warranty
11 * of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12 * GNU General Public License for more details.
13 */
14
15#include <linux/platform_device.h>
16#include <linux/module.h>
17#include <linux/netdevice.h>
18#include <linux/phy.h>
19#include <linux/of.h>
20#include <linux/of_device.h>
21
22#include "cpsw.h"
23
24/* AM33xx SoC specific definitions for the CONTROL port */
25#define AM33XX_GMII_SEL_MODE_MII 0
26#define AM33XX_GMII_SEL_MODE_RMII 1
27#define AM33XX_GMII_SEL_MODE_RGMII 2
28
29#define AM33XX_GMII_SEL_RMII2_IO_CLK_EN BIT(7)
30#define AM33XX_GMII_SEL_RMII1_IO_CLK_EN BIT(6)
31
32struct cpsw_phy_sel_priv {
33 struct device *dev;
34 u32 __iomem *gmii_sel;
35 bool rmii_clock_external;
36 void (*cpsw_phy_sel)(struct cpsw_phy_sel_priv *priv,
37 phy_interface_t phy_mode, int slave);
38};
39
40
41static void cpsw_gmii_sel_am3352(struct cpsw_phy_sel_priv *priv,
42 phy_interface_t phy_mode, int slave)
43{
44 u32 reg;
45 u32 mask;
46 u32 mode = 0;
47
48 reg = readl(priv->gmii_sel);
49
50 switch (phy_mode) {
51 case PHY_INTERFACE_MODE_RMII:
52 mode = AM33XX_GMII_SEL_MODE_RMII;
53 break;
54
55 case PHY_INTERFACE_MODE_RGMII:
56 case PHY_INTERFACE_MODE_RGMII_ID:
57 case PHY_INTERFACE_MODE_RGMII_RXID:
58 case PHY_INTERFACE_MODE_RGMII_TXID:
59 mode = AM33XX_GMII_SEL_MODE_RGMII;
60 break;
61
62 case PHY_INTERFACE_MODE_MII:
63 default:
64 mode = AM33XX_GMII_SEL_MODE_MII;
65 break;
66 };
67
68 mask = 0x3 << (slave * 2) | BIT(slave + 6);
69 mode <<= slave * 2;
70
71 if (priv->rmii_clock_external) {
72 if (slave == 0)
73 mode |= AM33XX_GMII_SEL_RMII1_IO_CLK_EN;
74 else
75 mode |= AM33XX_GMII_SEL_RMII2_IO_CLK_EN;
76 }
77
78 reg &= ~mask;
79 reg |= mode;
80
81 writel(reg, priv->gmii_sel);
82}
83
84static struct platform_driver cpsw_phy_sel_driver;
85static int match(struct device *dev, void *data)
86{
87 struct device_node *node = (struct device_node *)data;
88 return dev->of_node == node &&
89 dev->driver == &cpsw_phy_sel_driver.driver;
90}
91
92void cpsw_phy_sel(struct device *dev, phy_interface_t phy_mode, int slave)
93{
94 struct device_node *node;
95 struct cpsw_phy_sel_priv *priv;
96
97 node = of_get_child_by_name(dev->of_node, "cpsw-phy-sel");
98 if (!node) {
99 dev_err(dev, "Phy mode driver DT not found\n");
100 return;
101 }
102
103 dev = bus_find_device(&platform_bus_type, NULL, node, match);
104 priv = dev_get_drvdata(dev);
105
106 priv->cpsw_phy_sel(priv, phy_mode, slave);
107}
108EXPORT_SYMBOL_GPL(cpsw_phy_sel);
109
110static const struct of_device_id cpsw_phy_sel_id_table[] = {
111 {
112 .compatible = "ti,am3352-cpsw-phy-sel",
113 .data = &cpsw_gmii_sel_am3352,
114 },
115 {}
116};
117MODULE_DEVICE_TABLE(of, cpsw_phy_sel_id_table);
118
119static int cpsw_phy_sel_probe(struct platform_device *pdev)
120{
121 struct resource *res;
122 const struct of_device_id *of_id;
123 struct cpsw_phy_sel_priv *priv;
124
125 of_id = of_match_node(cpsw_phy_sel_id_table, pdev->dev.of_node);
126 if (!of_id)
127 return -EINVAL;
128
129 priv = devm_kzalloc(&pdev->dev, sizeof(*priv), GFP_KERNEL);
130 if (!priv) {
131 dev_err(&pdev->dev, "unable to alloc memory for cpsw phy sel\n");
132 return -ENOMEM;
133 }
134
135 priv->cpsw_phy_sel = of_id->data;
136
137 res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "gmii-sel");
138 priv->gmii_sel = devm_ioremap_resource(&pdev->dev, res);
139 if (IS_ERR(priv->gmii_sel))
140 return PTR_ERR(priv->gmii_sel);
141
142 if (of_find_property(pdev->dev.of_node, "rmii-clock-ext", NULL))
143 priv->rmii_clock_external = true;
144
145 dev_set_drvdata(&pdev->dev, priv);
146
147 return 0;
148}
149
150static struct platform_driver cpsw_phy_sel_driver = {
151 .probe = cpsw_phy_sel_probe,
152 .driver = {
153 .name = "cpsw-phy-sel",
154 .owner = THIS_MODULE,
155 .of_match_table = cpsw_phy_sel_id_table,
156 },
157};
158
159module_platform_driver(cpsw_phy_sel_driver);
160MODULE_AUTHOR("Mugunthan V N <mugunthanvnm@ti.com>");
161MODULE_LICENSE("GPL v2");
diff --git a/drivers/net/ethernet/ti/cpsw.c b/drivers/net/ethernet/ti/cpsw.c
index cc3ce557e4aa..90d41d26ec6d 100644
--- a/drivers/net/ethernet/ti/cpsw.c
+++ b/drivers/net/ethernet/ti/cpsw.c
@@ -367,8 +367,6 @@ struct cpsw_priv {
367 spinlock_t lock; 367 spinlock_t lock;
368 struct platform_device *pdev; 368 struct platform_device *pdev;
369 struct net_device *ndev; 369 struct net_device *ndev;
370 struct resource *cpsw_res;
371 struct resource *cpsw_wr_res;
372 struct napi_struct napi; 370 struct napi_struct napi;
373 struct device *dev; 371 struct device *dev;
374 struct cpsw_platform_data data; 372 struct cpsw_platform_data data;
@@ -1016,6 +1014,10 @@ static void cpsw_slave_open(struct cpsw_slave *slave, struct cpsw_priv *priv)
1016 dev_info(priv->dev, "phy found : id is : 0x%x\n", 1014 dev_info(priv->dev, "phy found : id is : 0x%x\n",
1017 slave->phy->phy_id); 1015 slave->phy->phy_id);
1018 phy_start(slave->phy); 1016 phy_start(slave->phy);
1017
1018 /* Configure GMII_SEL register */
1019 cpsw_phy_sel(&priv->pdev->dev, slave->phy->interface,
1020 slave->slave_num);
1019 } 1021 }
1020} 1022}
1021 1023
@@ -1705,62 +1707,55 @@ static int cpsw_probe_dt(struct cpsw_platform_data *data,
1705 1707
1706 if (of_property_read_u32(node, "active_slave", &prop)) { 1708 if (of_property_read_u32(node, "active_slave", &prop)) {
1707 pr_err("Missing active_slave property in the DT.\n"); 1709 pr_err("Missing active_slave property in the DT.\n");
1708 ret = -EINVAL; 1710 return -EINVAL;
1709 goto error_ret;
1710 } 1711 }
1711 data->active_slave = prop; 1712 data->active_slave = prop;
1712 1713
1713 if (of_property_read_u32(node, "cpts_clock_mult", &prop)) { 1714 if (of_property_read_u32(node, "cpts_clock_mult", &prop)) {
1714 pr_err("Missing cpts_clock_mult property in the DT.\n"); 1715 pr_err("Missing cpts_clock_mult property in the DT.\n");
1715 ret = -EINVAL; 1716 return -EINVAL;
1716 goto error_ret;
1717 } 1717 }
1718 data->cpts_clock_mult = prop; 1718 data->cpts_clock_mult = prop;
1719 1719
1720 if (of_property_read_u32(node, "cpts_clock_shift", &prop)) { 1720 if (of_property_read_u32(node, "cpts_clock_shift", &prop)) {
1721 pr_err("Missing cpts_clock_shift property in the DT.\n"); 1721 pr_err("Missing cpts_clock_shift property in the DT.\n");
1722 ret = -EINVAL; 1722 return -EINVAL;
1723 goto error_ret;
1724 } 1723 }
1725 data->cpts_clock_shift = prop; 1724 data->cpts_clock_shift = prop;
1726 1725
1727 data->slave_data = kcalloc(data->slaves, sizeof(struct cpsw_slave_data), 1726 data->slave_data = devm_kzalloc(&pdev->dev, data->slaves
1728 GFP_KERNEL); 1727 * sizeof(struct cpsw_slave_data),
1728 GFP_KERNEL);
1729 if (!data->slave_data) 1729 if (!data->slave_data)
1730 return -EINVAL; 1730 return -ENOMEM;
1731 1731
1732 if (of_property_read_u32(node, "cpdma_channels", &prop)) { 1732 if (of_property_read_u32(node, "cpdma_channels", &prop)) {
1733 pr_err("Missing cpdma_channels property in the DT.\n"); 1733 pr_err("Missing cpdma_channels property in the DT.\n");
1734 ret = -EINVAL; 1734 return -EINVAL;
1735 goto error_ret;
1736 } 1735 }
1737 data->channels = prop; 1736 data->channels = prop;
1738 1737
1739 if (of_property_read_u32(node, "ale_entries", &prop)) { 1738 if (of_property_read_u32(node, "ale_entries", &prop)) {
1740 pr_err("Missing ale_entries property in the DT.\n"); 1739 pr_err("Missing ale_entries property in the DT.\n");
1741 ret = -EINVAL; 1740 return -EINVAL;
1742 goto error_ret;
1743 } 1741 }
1744 data->ale_entries = prop; 1742 data->ale_entries = prop;
1745 1743
1746 if (of_property_read_u32(node, "bd_ram_size", &prop)) { 1744 if (of_property_read_u32(node, "bd_ram_size", &prop)) {
1747 pr_err("Missing bd_ram_size property in the DT.\n"); 1745 pr_err("Missing bd_ram_size property in the DT.\n");
1748 ret = -EINVAL; 1746 return -EINVAL;
1749 goto error_ret;
1750 } 1747 }
1751 data->bd_ram_size = prop; 1748 data->bd_ram_size = prop;
1752 1749
1753 if (of_property_read_u32(node, "rx_descs", &prop)) { 1750 if (of_property_read_u32(node, "rx_descs", &prop)) {
1754 pr_err("Missing rx_descs property in the DT.\n"); 1751 pr_err("Missing rx_descs property in the DT.\n");
1755 ret = -EINVAL; 1752 return -EINVAL;
1756 goto error_ret;
1757 } 1753 }
1758 data->rx_descs = prop; 1754 data->rx_descs = prop;
1759 1755
1760 if (of_property_read_u32(node, "mac_control", &prop)) { 1756 if (of_property_read_u32(node, "mac_control", &prop)) {
1761 pr_err("Missing mac_control property in the DT.\n"); 1757 pr_err("Missing mac_control property in the DT.\n");
1762 ret = -EINVAL; 1758 return -EINVAL;
1763 goto error_ret;
1764 } 1759 }
1765 data->mac_control = prop; 1760 data->mac_control = prop;
1766 1761
@@ -1791,8 +1786,7 @@ static int cpsw_probe_dt(struct cpsw_platform_data *data,
1791 parp = of_get_property(slave_node, "phy_id", &lenp); 1786 parp = of_get_property(slave_node, "phy_id", &lenp);
1792 if ((parp == NULL) || (lenp != (sizeof(void *) * 2))) { 1787 if ((parp == NULL) || (lenp != (sizeof(void *) * 2))) {
1793 pr_err("Missing slave[%d] phy_id property\n", i); 1788 pr_err("Missing slave[%d] phy_id property\n", i);
1794 ret = -EINVAL; 1789 return -EINVAL;
1795 goto error_ret;
1796 } 1790 }
1797 mdio_node = of_find_node_by_phandle(be32_to_cpup(parp)); 1791 mdio_node = of_find_node_by_phandle(be32_to_cpup(parp));
1798 phyid = be32_to_cpup(parp+1); 1792 phyid = be32_to_cpup(parp+1);
@@ -1822,10 +1816,6 @@ static int cpsw_probe_dt(struct cpsw_platform_data *data,
1822 } 1816 }
1823 1817
1824 return 0; 1818 return 0;
1825
1826error_ret:
1827 kfree(data->slave_data);
1828 return ret;
1829} 1819}
1830 1820
1831static int cpsw_probe_dual_emac(struct platform_device *pdev, 1821static int cpsw_probe_dual_emac(struct platform_device *pdev,
@@ -1867,7 +1857,6 @@ static int cpsw_probe_dual_emac(struct platform_device *pdev,
1867 priv_sl2->coal_intvl = 0; 1857 priv_sl2->coal_intvl = 0;
1868 priv_sl2->bus_freq_mhz = priv->bus_freq_mhz; 1858 priv_sl2->bus_freq_mhz = priv->bus_freq_mhz;
1869 1859
1870 priv_sl2->cpsw_res = priv->cpsw_res;
1871 priv_sl2->regs = priv->regs; 1860 priv_sl2->regs = priv->regs;
1872 priv_sl2->host_port = priv->host_port; 1861 priv_sl2->host_port = priv->host_port;
1873 priv_sl2->host_port_regs = priv->host_port_regs; 1862 priv_sl2->host_port_regs = priv->host_port_regs;
@@ -1911,8 +1900,8 @@ static int cpsw_probe(struct platform_device *pdev)
1911 struct cpsw_priv *priv; 1900 struct cpsw_priv *priv;
1912 struct cpdma_params dma_params; 1901 struct cpdma_params dma_params;
1913 struct cpsw_ale_params ale_params; 1902 struct cpsw_ale_params ale_params;
1914 void __iomem *ss_regs, *wr_regs; 1903 void __iomem *ss_regs;
1915 struct resource *res; 1904 struct resource *res, *ss_res;
1916 u32 slave_offset, sliver_offset, slave_size; 1905 u32 slave_offset, sliver_offset, slave_size;
1917 int ret = 0, i, k = 0; 1906 int ret = 0, i, k = 0;
1918 1907
@@ -1948,7 +1937,7 @@ static int cpsw_probe(struct platform_device *pdev)
1948 if (cpsw_probe_dt(&priv->data, pdev)) { 1937 if (cpsw_probe_dt(&priv->data, pdev)) {
1949 pr_err("cpsw: platform data missing\n"); 1938 pr_err("cpsw: platform data missing\n");
1950 ret = -ENODEV; 1939 ret = -ENODEV;
1951 goto clean_ndev_ret; 1940 goto clean_runtime_disable_ret;
1952 } 1941 }
1953 data = &priv->data; 1942 data = &priv->data;
1954 1943
@@ -1962,11 +1951,12 @@ static int cpsw_probe(struct platform_device *pdev)
1962 1951
1963 memcpy(ndev->dev_addr, priv->mac_addr, ETH_ALEN); 1952 memcpy(ndev->dev_addr, priv->mac_addr, ETH_ALEN);
1964 1953
1965 priv->slaves = kzalloc(sizeof(struct cpsw_slave) * data->slaves, 1954 priv->slaves = devm_kzalloc(&pdev->dev,
1966 GFP_KERNEL); 1955 sizeof(struct cpsw_slave) * data->slaves,
1956 GFP_KERNEL);
1967 if (!priv->slaves) { 1957 if (!priv->slaves) {
1968 ret = -EBUSY; 1958 ret = -ENOMEM;
1969 goto clean_ndev_ret; 1959 goto clean_runtime_disable_ret;
1970 } 1960 }
1971 for (i = 0; i < data->slaves; i++) 1961 for (i = 0; i < data->slaves; i++)
1972 priv->slaves[i].slave_num = i; 1962 priv->slaves[i].slave_num = i;
@@ -1974,55 +1964,31 @@ static int cpsw_probe(struct platform_device *pdev)
1974 priv->slaves[0].ndev = ndev; 1964 priv->slaves[0].ndev = ndev;
1975 priv->emac_port = 0; 1965 priv->emac_port = 0;
1976 1966
1977 priv->clk = clk_get(&pdev->dev, "fck"); 1967 priv->clk = devm_clk_get(&pdev->dev, "fck");
1978 if (IS_ERR(priv->clk)) { 1968 if (IS_ERR(priv->clk)) {
1979 dev_err(&pdev->dev, "fck is not found\n"); 1969 dev_err(priv->dev, "fck is not found\n");
1980 ret = -ENODEV; 1970 ret = -ENODEV;
1981 goto clean_slave_ret; 1971 goto clean_runtime_disable_ret;
1982 } 1972 }
1983 priv->coal_intvl = 0; 1973 priv->coal_intvl = 0;
1984 priv->bus_freq_mhz = clk_get_rate(priv->clk) / 1000000; 1974 priv->bus_freq_mhz = clk_get_rate(priv->clk) / 1000000;
1985 1975
1986 priv->cpsw_res = platform_get_resource(pdev, IORESOURCE_MEM, 0); 1976 ss_res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
1987 if (!priv->cpsw_res) { 1977 ss_regs = devm_ioremap_resource(&pdev->dev, ss_res);
1988 dev_err(priv->dev, "error getting i/o resource\n"); 1978 if (IS_ERR(ss_regs)) {
1989 ret = -ENOENT; 1979 ret = PTR_ERR(ss_regs);
1990 goto clean_clk_ret; 1980 goto clean_runtime_disable_ret;
1991 }
1992 if (!request_mem_region(priv->cpsw_res->start,
1993 resource_size(priv->cpsw_res), ndev->name)) {
1994 dev_err(priv->dev, "failed request i/o region\n");
1995 ret = -ENXIO;
1996 goto clean_clk_ret;
1997 }
1998 ss_regs = ioremap(priv->cpsw_res->start, resource_size(priv->cpsw_res));
1999 if (!ss_regs) {
2000 dev_err(priv->dev, "unable to map i/o region\n");
2001 goto clean_cpsw_iores_ret;
2002 } 1981 }
2003 priv->regs = ss_regs; 1982 priv->regs = ss_regs;
2004 priv->version = __raw_readl(&priv->regs->id_ver); 1983 priv->version = __raw_readl(&priv->regs->id_ver);
2005 priv->host_port = HOST_PORT_NUM; 1984 priv->host_port = HOST_PORT_NUM;
2006 1985
2007 priv->cpsw_wr_res = platform_get_resource(pdev, IORESOURCE_MEM, 1); 1986 res = platform_get_resource(pdev, IORESOURCE_MEM, 1);
2008 if (!priv->cpsw_wr_res) { 1987 priv->wr_regs = devm_ioremap_resource(&pdev->dev, res);
2009 dev_err(priv->dev, "error getting i/o resource\n"); 1988 if (IS_ERR(priv->wr_regs)) {
2010 ret = -ENOENT; 1989 ret = PTR_ERR(priv->wr_regs);
2011 goto clean_iomap_ret; 1990 goto clean_runtime_disable_ret;
2012 }
2013 if (!request_mem_region(priv->cpsw_wr_res->start,
2014 resource_size(priv->cpsw_wr_res), ndev->name)) {
2015 dev_err(priv->dev, "failed request i/o region\n");
2016 ret = -ENXIO;
2017 goto clean_iomap_ret;
2018 } 1991 }
2019 wr_regs = ioremap(priv->cpsw_wr_res->start,
2020 resource_size(priv->cpsw_wr_res));
2021 if (!wr_regs) {
2022 dev_err(priv->dev, "unable to map i/o region\n");
2023 goto clean_cpsw_wr_iores_ret;
2024 }
2025 priv->wr_regs = wr_regs;
2026 1992
2027 memset(&dma_params, 0, sizeof(dma_params)); 1993 memset(&dma_params, 0, sizeof(dma_params));
2028 memset(&ale_params, 0, sizeof(ale_params)); 1994 memset(&ale_params, 0, sizeof(ale_params));
@@ -2053,12 +2019,12 @@ static int cpsw_probe(struct platform_device *pdev)
2053 slave_size = CPSW2_SLAVE_SIZE; 2019 slave_size = CPSW2_SLAVE_SIZE;
2054 sliver_offset = CPSW2_SLIVER_OFFSET; 2020 sliver_offset = CPSW2_SLIVER_OFFSET;
2055 dma_params.desc_mem_phys = 2021 dma_params.desc_mem_phys =
2056 (u32 __force) priv->cpsw_res->start + CPSW2_BD_OFFSET; 2022 (u32 __force) ss_res->start + CPSW2_BD_OFFSET;
2057 break; 2023 break;
2058 default: 2024 default:
2059 dev_err(priv->dev, "unknown version 0x%08x\n", priv->version); 2025 dev_err(priv->dev, "unknown version 0x%08x\n", priv->version);
2060 ret = -ENODEV; 2026 ret = -ENODEV;
2061 goto clean_cpsw_wr_iores_ret; 2027 goto clean_runtime_disable_ret;
2062 } 2028 }
2063 for (i = 0; i < priv->data.slaves; i++) { 2029 for (i = 0; i < priv->data.slaves; i++) {
2064 struct cpsw_slave *slave = &priv->slaves[i]; 2030 struct cpsw_slave *slave = &priv->slaves[i];
@@ -2086,7 +2052,7 @@ static int cpsw_probe(struct platform_device *pdev)
2086 if (!priv->dma) { 2052 if (!priv->dma) {
2087 dev_err(priv->dev, "error initializing dma\n"); 2053 dev_err(priv->dev, "error initializing dma\n");
2088 ret = -ENOMEM; 2054 ret = -ENOMEM;
2089 goto clean_wr_iomap_ret; 2055 goto clean_runtime_disable_ret;
2090 } 2056 }
2091 2057
2092 priv->txch = cpdma_chan_create(priv->dma, tx_chan_num(0), 2058 priv->txch = cpdma_chan_create(priv->dma, tx_chan_num(0),
@@ -2121,8 +2087,8 @@ static int cpsw_probe(struct platform_device *pdev)
2121 2087
2122 while ((res = platform_get_resource(priv->pdev, IORESOURCE_IRQ, k))) { 2088 while ((res = platform_get_resource(priv->pdev, IORESOURCE_IRQ, k))) {
2123 for (i = res->start; i <= res->end; i++) { 2089 for (i = res->start; i <= res->end; i++) {
2124 if (request_irq(i, cpsw_interrupt, 0, 2090 if (devm_request_irq(&pdev->dev, i, cpsw_interrupt, 0,
2125 dev_name(&pdev->dev), priv)) { 2091 dev_name(priv->dev), priv)) {
2126 dev_err(priv->dev, "error attaching irq\n"); 2092 dev_err(priv->dev, "error attaching irq\n");
2127 goto clean_ale_ret; 2093 goto clean_ale_ret;
2128 } 2094 }
@@ -2144,7 +2110,7 @@ static int cpsw_probe(struct platform_device *pdev)
2144 if (ret) { 2110 if (ret) {
2145 dev_err(priv->dev, "error registering net device\n"); 2111 dev_err(priv->dev, "error registering net device\n");
2146 ret = -ENODEV; 2112 ret = -ENODEV;
2147 goto clean_irq_ret; 2113 goto clean_ale_ret;
2148 } 2114 }
2149 2115
2150 if (cpts_register(&pdev->dev, priv->cpts, 2116 if (cpts_register(&pdev->dev, priv->cpts,
@@ -2152,44 +2118,27 @@ static int cpsw_probe(struct platform_device *pdev)
2152 dev_err(priv->dev, "error registering cpts device\n"); 2118 dev_err(priv->dev, "error registering cpts device\n");
2153 2119
2154 cpsw_notice(priv, probe, "initialized device (regs %x, irq %d)\n", 2120 cpsw_notice(priv, probe, "initialized device (regs %x, irq %d)\n",
2155 priv->cpsw_res->start, ndev->irq); 2121 ss_res->start, ndev->irq);
2156 2122
2157 if (priv->data.dual_emac) { 2123 if (priv->data.dual_emac) {
2158 ret = cpsw_probe_dual_emac(pdev, priv); 2124 ret = cpsw_probe_dual_emac(pdev, priv);
2159 if (ret) { 2125 if (ret) {
2160 cpsw_err(priv, probe, "error probe slave 2 emac interface\n"); 2126 cpsw_err(priv, probe, "error probe slave 2 emac interface\n");
2161 goto clean_irq_ret; 2127 goto clean_ale_ret;
2162 } 2128 }
2163 } 2129 }
2164 2130
2165 return 0; 2131 return 0;
2166 2132
2167clean_irq_ret:
2168 for (i = 0; i < priv->num_irqs; i++)
2169 free_irq(priv->irqs_table[i], priv);
2170clean_ale_ret: 2133clean_ale_ret:
2171 cpsw_ale_destroy(priv->ale); 2134 cpsw_ale_destroy(priv->ale);
2172clean_dma_ret: 2135clean_dma_ret:
2173 cpdma_chan_destroy(priv->txch); 2136 cpdma_chan_destroy(priv->txch);
2174 cpdma_chan_destroy(priv->rxch); 2137 cpdma_chan_destroy(priv->rxch);
2175 cpdma_ctlr_destroy(priv->dma); 2138 cpdma_ctlr_destroy(priv->dma);
2176clean_wr_iomap_ret: 2139clean_runtime_disable_ret:
2177 iounmap(priv->wr_regs);
2178clean_cpsw_wr_iores_ret:
2179 release_mem_region(priv->cpsw_wr_res->start,
2180 resource_size(priv->cpsw_wr_res));
2181clean_iomap_ret:
2182 iounmap(priv->regs);
2183clean_cpsw_iores_ret:
2184 release_mem_region(priv->cpsw_res->start,
2185 resource_size(priv->cpsw_res));
2186clean_clk_ret:
2187 clk_put(priv->clk);
2188clean_slave_ret:
2189 pm_runtime_disable(&pdev->dev); 2140 pm_runtime_disable(&pdev->dev);
2190 kfree(priv->slaves);
2191clean_ndev_ret: 2141clean_ndev_ret:
2192 kfree(priv->data.slave_data);
2193 free_netdev(priv->ndev); 2142 free_netdev(priv->ndev);
2194 return ret; 2143 return ret;
2195} 2144}
@@ -2198,30 +2147,18 @@ static int cpsw_remove(struct platform_device *pdev)
2198{ 2147{
2199 struct net_device *ndev = platform_get_drvdata(pdev); 2148 struct net_device *ndev = platform_get_drvdata(pdev);
2200 struct cpsw_priv *priv = netdev_priv(ndev); 2149 struct cpsw_priv *priv = netdev_priv(ndev);
2201 int i;
2202 2150
2203 if (priv->data.dual_emac) 2151 if (priv->data.dual_emac)
2204 unregister_netdev(cpsw_get_slave_ndev(priv, 1)); 2152 unregister_netdev(cpsw_get_slave_ndev(priv, 1));
2205 unregister_netdev(ndev); 2153 unregister_netdev(ndev);
2206 2154
2207 cpts_unregister(priv->cpts); 2155 cpts_unregister(priv->cpts);
2208 for (i = 0; i < priv->num_irqs; i++)
2209 free_irq(priv->irqs_table[i], priv);
2210 2156
2211 cpsw_ale_destroy(priv->ale); 2157 cpsw_ale_destroy(priv->ale);
2212 cpdma_chan_destroy(priv->txch); 2158 cpdma_chan_destroy(priv->txch);
2213 cpdma_chan_destroy(priv->rxch); 2159 cpdma_chan_destroy(priv->rxch);
2214 cpdma_ctlr_destroy(priv->dma); 2160 cpdma_ctlr_destroy(priv->dma);
2215 iounmap(priv->regs);
2216 release_mem_region(priv->cpsw_res->start,
2217 resource_size(priv->cpsw_res));
2218 iounmap(priv->wr_regs);
2219 release_mem_region(priv->cpsw_wr_res->start,
2220 resource_size(priv->cpsw_wr_res));
2221 pm_runtime_disable(&pdev->dev); 2161 pm_runtime_disable(&pdev->dev);
2222 clk_put(priv->clk);
2223 kfree(priv->slaves);
2224 kfree(priv->data.slave_data);
2225 if (priv->data.dual_emac) 2162 if (priv->data.dual_emac)
2226 free_netdev(cpsw_get_slave_ndev(priv, 1)); 2163 free_netdev(cpsw_get_slave_ndev(priv, 1));
2227 free_netdev(ndev); 2164 free_netdev(ndev);
@@ -2277,7 +2214,7 @@ static struct platform_driver cpsw_driver = {
2277 .name = "cpsw", 2214 .name = "cpsw",
2278 .owner = THIS_MODULE, 2215 .owner = THIS_MODULE,
2279 .pm = &cpsw_pm_ops, 2216 .pm = &cpsw_pm_ops,
2280 .of_match_table = of_match_ptr(cpsw_of_mtable), 2217 .of_match_table = cpsw_of_mtable,
2281 }, 2218 },
2282 .probe = cpsw_probe, 2219 .probe = cpsw_probe,
2283 .remove = cpsw_remove, 2220 .remove = cpsw_remove,
diff --git a/drivers/net/ethernet/ti/cpsw.h b/drivers/net/ethernet/ti/cpsw.h
index eb3e101ec048..574f49da693f 100644
--- a/drivers/net/ethernet/ti/cpsw.h
+++ b/drivers/net/ethernet/ti/cpsw.h
@@ -39,4 +39,6 @@ struct cpsw_platform_data {
39 bool dual_emac; /* Enable Dual EMAC mode */ 39 bool dual_emac; /* Enable Dual EMAC mode */
40}; 40};
41 41
42void cpsw_phy_sel(struct device *dev, phy_interface_t phy_mode, int slave);
43
42#endif /* __CPSW_H__ */ 44#endif /* __CPSW_H__ */
diff --git a/drivers/net/ethernet/ti/cpts.h b/drivers/net/ethernet/ti/cpts.h
index fe993cdd7e23..1a581ef7eee8 100644
--- a/drivers/net/ethernet/ti/cpts.h
+++ b/drivers/net/ethernet/ti/cpts.h
@@ -127,8 +127,8 @@ struct cpts {
127}; 127};
128 128
129#ifdef CONFIG_TI_CPTS 129#ifdef CONFIG_TI_CPTS
130extern void cpts_rx_timestamp(struct cpts *cpts, struct sk_buff *skb); 130void cpts_rx_timestamp(struct cpts *cpts, struct sk_buff *skb);
131extern void cpts_tx_timestamp(struct cpts *cpts, struct sk_buff *skb); 131void cpts_tx_timestamp(struct cpts *cpts, struct sk_buff *skb);
132#else 132#else
133static inline void cpts_rx_timestamp(struct cpts *cpts, struct sk_buff *skb) 133static inline void cpts_rx_timestamp(struct cpts *cpts, struct sk_buff *skb)
134{ 134{
@@ -138,8 +138,7 @@ static inline void cpts_tx_timestamp(struct cpts *cpts, struct sk_buff *skb)
138} 138}
139#endif 139#endif
140 140
141extern int cpts_register(struct device *dev, struct cpts *cpts, 141int cpts_register(struct device *dev, struct cpts *cpts, u32 mult, u32 shift);
142 u32 mult, u32 shift); 142void cpts_unregister(struct cpts *cpts);
143extern void cpts_unregister(struct cpts *cpts);
144 143
145#endif 144#endif
diff --git a/drivers/net/ethernet/ti/davinci_emac.c b/drivers/net/ethernet/ti/davinci_emac.c
index 6a32ef9d63ae..41ba974bf37c 100644
--- a/drivers/net/ethernet/ti/davinci_emac.c
+++ b/drivers/net/ethernet/ti/davinci_emac.c
@@ -1852,7 +1852,7 @@ static int davinci_emac_probe(struct platform_device *pdev)
1852 } 1852 }
1853 1853
1854 /* MAC addr and PHY mask , RMII enable info from platform_data */ 1854 /* MAC addr and PHY mask , RMII enable info from platform_data */
1855 memcpy(priv->mac_addr, pdata->mac_addr, 6); 1855 memcpy(priv->mac_addr, pdata->mac_addr, ETH_ALEN);
1856 priv->phy_id = pdata->phy_id; 1856 priv->phy_id = pdata->phy_id;
1857 priv->rmii_en = pdata->rmii_en; 1857 priv->rmii_en = pdata->rmii_en;
1858 priv->version = pdata->version; 1858 priv->version = pdata->version;
diff --git a/drivers/net/ethernet/ti/tlan.c b/drivers/net/ethernet/ti/tlan.c
index 591437e59b90..62b19be5183d 100644
--- a/drivers/net/ethernet/ti/tlan.c
+++ b/drivers/net/ethernet/ti/tlan.c
@@ -319,7 +319,6 @@ static void tlan_remove_one(struct pci_dev *pdev)
319 319
320 free_netdev(dev); 320 free_netdev(dev);
321 321
322 pci_set_drvdata(pdev, NULL);
323 cancel_work_sync(&priv->tlan_tqueue); 322 cancel_work_sync(&priv->tlan_tqueue);
324} 323}
325 324
diff --git a/drivers/net/ethernet/tile/tilegx.c b/drivers/net/ethernet/tile/tilegx.c
index 13e6fff8ca23..628b736e5ae7 100644
--- a/drivers/net/ethernet/tile/tilegx.c
+++ b/drivers/net/ethernet/tile/tilegx.c
@@ -2230,7 +2230,7 @@ static void tile_net_dev_init(const char *name, const uint8_t *mac)
2230 nz_addr |= mac[i]; 2230 nz_addr |= mac[i];
2231 2231
2232 if (nz_addr) { 2232 if (nz_addr) {
2233 memcpy(dev->dev_addr, mac, 6); 2233 memcpy(dev->dev_addr, mac, ETH_ALEN);
2234 dev->addr_len = 6; 2234 dev->addr_len = 6;
2235 } else { 2235 } else {
2236 eth_hw_addr_random(dev); 2236 eth_hw_addr_random(dev);
diff --git a/drivers/net/ethernet/toshiba/ps3_gelic_net.h b/drivers/net/ethernet/toshiba/ps3_gelic_net.h
index 309abb472aa2..8505196be9f5 100644
--- a/drivers/net/ethernet/toshiba/ps3_gelic_net.h
+++ b/drivers/net/ethernet/toshiba/ps3_gelic_net.h
@@ -359,27 +359,26 @@ static inline void *port_priv(struct gelic_port *port)
359} 359}
360 360
361#ifdef CONFIG_PPC_EARLY_DEBUG_PS3GELIC 361#ifdef CONFIG_PPC_EARLY_DEBUG_PS3GELIC
362extern void udbg_shutdown_ps3gelic(void); 362void udbg_shutdown_ps3gelic(void);
363#else 363#else
364static inline void udbg_shutdown_ps3gelic(void) {} 364static inline void udbg_shutdown_ps3gelic(void) {}
365#endif 365#endif
366 366
367extern int gelic_card_set_irq_mask(struct gelic_card *card, u64 mask); 367int gelic_card_set_irq_mask(struct gelic_card *card, u64 mask);
368/* shared netdev ops */ 368/* shared netdev ops */
369extern void gelic_card_up(struct gelic_card *card); 369void gelic_card_up(struct gelic_card *card);
370extern void gelic_card_down(struct gelic_card *card); 370void gelic_card_down(struct gelic_card *card);
371extern int gelic_net_open(struct net_device *netdev); 371int gelic_net_open(struct net_device *netdev);
372extern int gelic_net_stop(struct net_device *netdev); 372int gelic_net_stop(struct net_device *netdev);
373extern int gelic_net_xmit(struct sk_buff *skb, struct net_device *netdev); 373int gelic_net_xmit(struct sk_buff *skb, struct net_device *netdev);
374extern void gelic_net_set_multi(struct net_device *netdev); 374void gelic_net_set_multi(struct net_device *netdev);
375extern void gelic_net_tx_timeout(struct net_device *netdev); 375void gelic_net_tx_timeout(struct net_device *netdev);
376extern int gelic_net_change_mtu(struct net_device *netdev, int new_mtu); 376int gelic_net_change_mtu(struct net_device *netdev, int new_mtu);
377extern int gelic_net_setup_netdev(struct net_device *netdev, 377int gelic_net_setup_netdev(struct net_device *netdev, struct gelic_card *card);
378 struct gelic_card *card);
379 378
380/* shared ethtool ops */ 379/* shared ethtool ops */
381extern void gelic_net_get_drvinfo(struct net_device *netdev, 380void gelic_net_get_drvinfo(struct net_device *netdev,
382 struct ethtool_drvinfo *info); 381 struct ethtool_drvinfo *info);
383extern void gelic_net_poll_controller(struct net_device *netdev); 382void gelic_net_poll_controller(struct net_device *netdev);
384 383
385#endif /* _GELIC_NET_H */ 384#endif /* _GELIC_NET_H */
diff --git a/drivers/net/ethernet/toshiba/ps3_gelic_wireless.h b/drivers/net/ethernet/toshiba/ps3_gelic_wireless.h
index f7e51b7d7049..11f443d8e4ea 100644
--- a/drivers/net/ethernet/toshiba/ps3_gelic_wireless.h
+++ b/drivers/net/ethernet/toshiba/ps3_gelic_wireless.h
@@ -320,7 +320,7 @@ struct gelic_eurus_cmd {
320#define GELIC_WL_PRIV_SET_PSK (SIOCIWFIRSTPRIV + 0) 320#define GELIC_WL_PRIV_SET_PSK (SIOCIWFIRSTPRIV + 0)
321#define GELIC_WL_PRIV_GET_PSK (SIOCIWFIRSTPRIV + 1) 321#define GELIC_WL_PRIV_GET_PSK (SIOCIWFIRSTPRIV + 1)
322 322
323extern int gelic_wl_driver_probe(struct gelic_card *card); 323int gelic_wl_driver_probe(struct gelic_card *card);
324extern int gelic_wl_driver_remove(struct gelic_card *card); 324int gelic_wl_driver_remove(struct gelic_card *card);
325extern void gelic_wl_interrupt(struct net_device *netdev, u64 status); 325void gelic_wl_interrupt(struct net_device *netdev, u64 status);
326#endif /* _GELIC_WIRELESS_H */ 326#endif /* _GELIC_WIRELESS_H */
diff --git a/drivers/net/ethernet/toshiba/spider_net.c b/drivers/net/ethernet/toshiba/spider_net.c
index 5734480c1ecf..3f4a32e39d27 100644
--- a/drivers/net/ethernet/toshiba/spider_net.c
+++ b/drivers/net/ethernet/toshiba/spider_net.c
@@ -2478,7 +2478,6 @@ out_release_regions:
2478 pci_release_regions(pdev); 2478 pci_release_regions(pdev);
2479out_disable_dev: 2479out_disable_dev:
2480 pci_disable_device(pdev); 2480 pci_disable_device(pdev);
2481 pci_set_drvdata(pdev, NULL);
2482 return NULL; 2481 return NULL;
2483} 2482}
2484 2483
diff --git a/drivers/net/ethernet/toshiba/spider_net.h b/drivers/net/ethernet/toshiba/spider_net.h
index 4ba2135474d1..9b6af0845a11 100644
--- a/drivers/net/ethernet/toshiba/spider_net.h
+++ b/drivers/net/ethernet/toshiba/spider_net.h
@@ -29,8 +29,8 @@
29 29
30#include <linux/sungem_phy.h> 30#include <linux/sungem_phy.h>
31 31
32extern int spider_net_stop(struct net_device *netdev); 32int spider_net_stop(struct net_device *netdev);
33extern int spider_net_open(struct net_device *netdev); 33int spider_net_open(struct net_device *netdev);
34 34
35extern const struct ethtool_ops spider_net_ethtool_ops; 35extern const struct ethtool_ops spider_net_ethtool_ops;
36 36
diff --git a/drivers/net/ethernet/toshiba/tc35815.c b/drivers/net/ethernet/toshiba/tc35815.c
index a971b9cca564..1322546d92ac 100644
--- a/drivers/net/ethernet/toshiba/tc35815.c
+++ b/drivers/net/ethernet/toshiba/tc35815.c
@@ -887,7 +887,6 @@ static void tc35815_remove_one(struct pci_dev *pdev)
887 mdiobus_free(lp->mii_bus); 887 mdiobus_free(lp->mii_bus);
888 unregister_netdev(dev); 888 unregister_netdev(dev);
889 free_netdev(dev); 889 free_netdev(dev);
890 pci_set_drvdata(pdev, NULL);
891} 890}
892 891
893static int 892static int
diff --git a/drivers/net/ethernet/via/via-rhine.c b/drivers/net/ethernet/via/via-rhine.c
index bdf697b184ae..4a7293ed95e9 100644
--- a/drivers/net/ethernet/via/via-rhine.c
+++ b/drivers/net/ethernet/via/via-rhine.c
@@ -2292,7 +2292,6 @@ static void rhine_remove_one(struct pci_dev *pdev)
2292 2292
2293 free_netdev(dev); 2293 free_netdev(dev);
2294 pci_disable_device(pdev); 2294 pci_disable_device(pdev);
2295 pci_set_drvdata(pdev, NULL);
2296} 2295}
2297 2296
2298static void rhine_shutdown (struct pci_dev *pdev) 2297static void rhine_shutdown (struct pci_dev *pdev)
diff --git a/drivers/net/ethernet/xilinx/xilinx_emaclite.c b/drivers/net/ethernet/xilinx/xilinx_emaclite.c
index 4c619ea5189f..74234a51c851 100644
--- a/drivers/net/ethernet/xilinx/xilinx_emaclite.c
+++ b/drivers/net/ethernet/xilinx/xilinx_emaclite.c
@@ -31,7 +31,7 @@
31#define DRIVER_NAME "xilinx_emaclite" 31#define DRIVER_NAME "xilinx_emaclite"
32 32
33/* Register offsets for the EmacLite Core */ 33/* Register offsets for the EmacLite Core */
34#define XEL_TXBUFF_OFFSET 0x0 /* Transmit Buffer */ 34#define XEL_TXBUFF_OFFSET 0x0 /* Transmit Buffer */
35#define XEL_MDIOADDR_OFFSET 0x07E4 /* MDIO Address Register */ 35#define XEL_MDIOADDR_OFFSET 0x07E4 /* MDIO Address Register */
36#define XEL_MDIOWR_OFFSET 0x07E8 /* MDIO Write Data Register */ 36#define XEL_MDIOWR_OFFSET 0x07E8 /* MDIO Write Data Register */
37#define XEL_MDIORD_OFFSET 0x07EC /* MDIO Read Data Register */ 37#define XEL_MDIORD_OFFSET 0x07EC /* MDIO Read Data Register */
@@ -63,13 +63,13 @@
63#define XEL_MDIOCTRL_MDIOEN_MASK 0x00000008 /* MDIO Enable */ 63#define XEL_MDIOCTRL_MDIOEN_MASK 0x00000008 /* MDIO Enable */
64 64
65/* Global Interrupt Enable Register (GIER) Bit Masks */ 65/* Global Interrupt Enable Register (GIER) Bit Masks */
66#define XEL_GIER_GIE_MASK 0x80000000 /* Global Enable */ 66#define XEL_GIER_GIE_MASK 0x80000000 /* Global Enable */
67 67
68/* Transmit Status Register (TSR) Bit Masks */ 68/* Transmit Status Register (TSR) Bit Masks */
69#define XEL_TSR_XMIT_BUSY_MASK 0x00000001 /* Tx complete */ 69#define XEL_TSR_XMIT_BUSY_MASK 0x00000001 /* Tx complete */
70#define XEL_TSR_PROGRAM_MASK 0x00000002 /* Program the MAC address */ 70#define XEL_TSR_PROGRAM_MASK 0x00000002 /* Program the MAC address */
71#define XEL_TSR_XMIT_IE_MASK 0x00000008 /* Tx interrupt enable bit */ 71#define XEL_TSR_XMIT_IE_MASK 0x00000008 /* Tx interrupt enable bit */
72#define XEL_TSR_XMIT_ACTIVE_MASK 0x80000000 /* Buffer is active, SW bit 72#define XEL_TSR_XMIT_ACTIVE_MASK 0x80000000 /* Buffer is active, SW bit
73 * only. This is not documented 73 * only. This is not documented
74 * in the HW spec */ 74 * in the HW spec */
75 75
@@ -77,21 +77,21 @@
77#define XEL_TSR_PROG_MAC_ADDR (XEL_TSR_XMIT_BUSY_MASK | XEL_TSR_PROGRAM_MASK) 77#define XEL_TSR_PROG_MAC_ADDR (XEL_TSR_XMIT_BUSY_MASK | XEL_TSR_PROGRAM_MASK)
78 78
79/* Receive Status Register (RSR) */ 79/* Receive Status Register (RSR) */
80#define XEL_RSR_RECV_DONE_MASK 0x00000001 /* Rx complete */ 80#define XEL_RSR_RECV_DONE_MASK 0x00000001 /* Rx complete */
81#define XEL_RSR_RECV_IE_MASK 0x00000008 /* Rx interrupt enable bit */ 81#define XEL_RSR_RECV_IE_MASK 0x00000008 /* Rx interrupt enable bit */
82 82
83/* Transmit Packet Length Register (TPLR) */ 83/* Transmit Packet Length Register (TPLR) */
84#define XEL_TPLR_LENGTH_MASK 0x0000FFFF /* Tx packet length */ 84#define XEL_TPLR_LENGTH_MASK 0x0000FFFF /* Tx packet length */
85 85
86/* Receive Packet Length Register (RPLR) */ 86/* Receive Packet Length Register (RPLR) */
87#define XEL_RPLR_LENGTH_MASK 0x0000FFFF /* Rx packet length */ 87#define XEL_RPLR_LENGTH_MASK 0x0000FFFF /* Rx packet length */
88 88
89#define XEL_HEADER_OFFSET 12 /* Offset to length field */ 89#define XEL_HEADER_OFFSET 12 /* Offset to length field */
90#define XEL_HEADER_SHIFT 16 /* Shift value for length */ 90#define XEL_HEADER_SHIFT 16 /* Shift value for length */
91 91
92/* General Ethernet Definitions */ 92/* General Ethernet Definitions */
93#define XEL_ARP_PACKET_SIZE 28 /* Max ARP packet size */ 93#define XEL_ARP_PACKET_SIZE 28 /* Max ARP packet size */
94#define XEL_HEADER_IP_LENGTH_OFFSET 16 /* IP Length Offset */ 94#define XEL_HEADER_IP_LENGTH_OFFSET 16 /* IP Length Offset */
95 95
96 96
97 97
@@ -1075,14 +1075,9 @@ static int xemaclite_send(struct sk_buff *orig_skb, struct net_device *dev)
1075 * This function un maps the IO region of the Emaclite device and frees the net 1075 * This function un maps the IO region of the Emaclite device and frees the net
1076 * device. 1076 * device.
1077 */ 1077 */
1078static void xemaclite_remove_ndev(struct net_device *ndev, 1078static void xemaclite_remove_ndev(struct net_device *ndev)
1079 struct platform_device *pdev)
1080{ 1079{
1081 if (ndev) { 1080 if (ndev) {
1082 struct net_local *lp = netdev_priv(ndev);
1083
1084 if (lp->base_addr)
1085 devm_iounmap(&pdev->dev, lp->base_addr);
1086 free_netdev(ndev); 1081 free_netdev(ndev);
1087 } 1082 }
1088} 1083}
@@ -1177,7 +1172,7 @@ static int xemaclite_of_probe(struct platform_device *ofdev)
1177 1172
1178 if (mac_address) 1173 if (mac_address)
1179 /* Set the MAC address. */ 1174 /* Set the MAC address. */
1180 memcpy(ndev->dev_addr, mac_address, 6); 1175 memcpy(ndev->dev_addr, mac_address, ETH_ALEN);
1181 else 1176 else
1182 dev_warn(dev, "No MAC address found\n"); 1177 dev_warn(dev, "No MAC address found\n");
1183 1178
@@ -1214,7 +1209,7 @@ static int xemaclite_of_probe(struct platform_device *ofdev)
1214 return 0; 1209 return 0;
1215 1210
1216error: 1211error:
1217 xemaclite_remove_ndev(ndev, ofdev); 1212 xemaclite_remove_ndev(ndev);
1218 return rc; 1213 return rc;
1219} 1214}
1220 1215
@@ -1248,7 +1243,7 @@ static int xemaclite_of_remove(struct platform_device *of_dev)
1248 of_node_put(lp->phy_node); 1243 of_node_put(lp->phy_node);
1249 lp->phy_node = NULL; 1244 lp->phy_node = NULL;
1250 1245
1251 xemaclite_remove_ndev(ndev, of_dev); 1246 xemaclite_remove_ndev(ndev);
1252 1247
1253 return 0; 1248 return 0;
1254} 1249}
diff --git a/drivers/net/fddi/skfp/fplustm.c b/drivers/net/fddi/skfp/fplustm.c
index a20ed1a98099..f83993590174 100644
--- a/drivers/net/fddi/skfp/fplustm.c
+++ b/drivers/net/fddi/skfp/fplustm.c
@@ -453,7 +453,7 @@ static void directed_beacon(struct s_smc *smc)
453 */ 453 */
454 * (char *) a = (char) ((long)DBEACON_INFO<<24L) ; 454 * (char *) a = (char) ((long)DBEACON_INFO<<24L) ;
455 a[1] = 0 ; 455 a[1] = 0 ;
456 memcpy((char *)a+1,(char *) &smc->mib.m[MAC0].fddiMACUpstreamNbr,6) ; 456 memcpy((char *)a+1, (char *) &smc->mib.m[MAC0].fddiMACUpstreamNbr, ETH_ALEN);
457 457
458 CHECK_NPP() ; 458 CHECK_NPP() ;
459 /* set memory address reg for writes */ 459 /* set memory address reg for writes */
diff --git a/drivers/net/fddi/skfp/h/smc.h b/drivers/net/fddi/skfp/h/smc.h
index 3ca308b28214..bd1166bf8f61 100644
--- a/drivers/net/fddi/skfp/h/smc.h
+++ b/drivers/net/fddi/skfp/h/smc.h
@@ -469,20 +469,20 @@ struct s_smc {
469 469
470extern const struct fddi_addr fddi_broadcast; 470extern const struct fddi_addr fddi_broadcast;
471 471
472extern void all_selection_criteria(struct s_smc *smc); 472void all_selection_criteria(struct s_smc *smc);
473extern void card_stop(struct s_smc *smc); 473void card_stop(struct s_smc *smc);
474extern void init_board(struct s_smc *smc, u_char *mac_addr); 474void init_board(struct s_smc *smc, u_char *mac_addr);
475extern int init_fplus(struct s_smc *smc); 475int init_fplus(struct s_smc *smc);
476extern void init_plc(struct s_smc *smc); 476void init_plc(struct s_smc *smc);
477extern int init_smt(struct s_smc *smc, u_char * mac_addr); 477int init_smt(struct s_smc *smc, u_char *mac_addr);
478extern void mac1_irq(struct s_smc *smc, u_short stu, u_short stl); 478void mac1_irq(struct s_smc *smc, u_short stu, u_short stl);
479extern void mac2_irq(struct s_smc *smc, u_short code_s2u, u_short code_s2l); 479void mac2_irq(struct s_smc *smc, u_short code_s2u, u_short code_s2l);
480extern void mac3_irq(struct s_smc *smc, u_short code_s3u, u_short code_s3l); 480void mac3_irq(struct s_smc *smc, u_short code_s3u, u_short code_s3l);
481extern int pcm_status_twisted(struct s_smc *smc); 481int pcm_status_twisted(struct s_smc *smc);
482extern void plc1_irq(struct s_smc *smc); 482void plc1_irq(struct s_smc *smc);
483extern void plc2_irq(struct s_smc *smc); 483void plc2_irq(struct s_smc *smc);
484extern void read_address(struct s_smc *smc, u_char * mac_addr); 484void read_address(struct s_smc *smc, u_char *mac_addr);
485extern void timer_irq(struct s_smc *smc); 485void timer_irq(struct s_smc *smc);
486 486
487#endif /* _SCMECM_ */ 487#endif /* _SCMECM_ */
488 488
diff --git a/drivers/net/fddi/skfp/skfddi.c b/drivers/net/fddi/skfp/skfddi.c
index f5d7305a5784..713d303a06a9 100644
--- a/drivers/net/fddi/skfp/skfddi.c
+++ b/drivers/net/fddi/skfp/skfddi.c
@@ -436,7 +436,7 @@ static int skfp_driver_init(struct net_device *dev)
436 } 436 }
437 read_address(smc, NULL); 437 read_address(smc, NULL);
438 pr_debug("HW-Addr: %pMF\n", smc->hw.fddi_canon_addr.a); 438 pr_debug("HW-Addr: %pMF\n", smc->hw.fddi_canon_addr.a);
439 memcpy(dev->dev_addr, smc->hw.fddi_canon_addr.a, 6); 439 memcpy(dev->dev_addr, smc->hw.fddi_canon_addr.a, ETH_ALEN);
440 440
441 smt_reset_defaults(smc, 0); 441 smt_reset_defaults(smc, 0);
442 442
@@ -503,7 +503,7 @@ static int skfp_open(struct net_device *dev)
503 * address. 503 * address.
504 */ 504 */
505 read_address(smc, NULL); 505 read_address(smc, NULL);
506 memcpy(dev->dev_addr, smc->hw.fddi_canon_addr.a, 6); 506 memcpy(dev->dev_addr, smc->hw.fddi_canon_addr.a, ETH_ALEN);
507 507
508 init_smt(smc, NULL); 508 init_smt(smc, NULL);
509 smt_online(smc, 1); 509 smt_online(smc, 1);
@@ -1213,7 +1213,7 @@ static void CheckSourceAddress(unsigned char *frame, unsigned char *hw_addr)
1213 if ((unsigned short) frame[1 + 10] != 0) 1213 if ((unsigned short) frame[1 + 10] != 0)
1214 return; 1214 return;
1215 SRBit = frame[1 + 6] & 0x01; 1215 SRBit = frame[1 + 6] & 0x01;
1216 memcpy(&frame[1 + 6], hw_addr, 6); 1216 memcpy(&frame[1 + 6], hw_addr, ETH_ALEN);
1217 frame[8] |= SRBit; 1217 frame[8] |= SRBit;
1218} // CheckSourceAddress 1218} // CheckSourceAddress
1219 1219
diff --git a/drivers/net/hamradio/baycom_ser_fdx.c b/drivers/net/hamradio/baycom_ser_fdx.c
index a974727dd9a2..636b65c66d49 100644
--- a/drivers/net/hamradio/baycom_ser_fdx.c
+++ b/drivers/net/hamradio/baycom_ser_fdx.c
@@ -445,7 +445,7 @@ static int ser12_open(struct net_device *dev)
445 outb(0, FCR(dev->base_addr)); /* disable FIFOs */ 445 outb(0, FCR(dev->base_addr)); /* disable FIFOs */
446 outb(0x0d, MCR(dev->base_addr)); 446 outb(0x0d, MCR(dev->base_addr));
447 outb(0, IER(dev->base_addr)); 447 outb(0, IER(dev->base_addr));
448 if (request_irq(dev->irq, ser12_interrupt, IRQF_DISABLED | IRQF_SHARED, 448 if (request_irq(dev->irq, ser12_interrupt, IRQF_SHARED,
449 "baycom_ser_fdx", dev)) { 449 "baycom_ser_fdx", dev)) {
450 release_region(dev->base_addr, SER12_EXTENT); 450 release_region(dev->base_addr, SER12_EXTENT);
451 return -EBUSY; 451 return -EBUSY;
diff --git a/drivers/net/hamradio/baycom_ser_hdx.c b/drivers/net/hamradio/baycom_ser_hdx.c
index e349d867449b..f9a8976195ba 100644
--- a/drivers/net/hamradio/baycom_ser_hdx.c
+++ b/drivers/net/hamradio/baycom_ser_hdx.c
@@ -490,7 +490,7 @@ static int ser12_open(struct net_device *dev)
490 outb(0, FCR(dev->base_addr)); /* disable FIFOs */ 490 outb(0, FCR(dev->base_addr)); /* disable FIFOs */
491 outb(0x0d, MCR(dev->base_addr)); 491 outb(0x0d, MCR(dev->base_addr));
492 outb(0, IER(dev->base_addr)); 492 outb(0, IER(dev->base_addr));
493 if (request_irq(dev->irq, ser12_interrupt, IRQF_DISABLED | IRQF_SHARED, 493 if (request_irq(dev->irq, ser12_interrupt, IRQF_SHARED,
494 "baycom_ser12", dev)) { 494 "baycom_ser12", dev)) {
495 release_region(dev->base_addr, SER12_EXTENT); 495 release_region(dev->base_addr, SER12_EXTENT);
496 return -EBUSY; 496 return -EBUSY;
diff --git a/drivers/net/hamradio/scc.c b/drivers/net/hamradio/scc.c
index bc1d52170389..4bc6ee8e7987 100644
--- a/drivers/net/hamradio/scc.c
+++ b/drivers/net/hamradio/scc.c
@@ -1734,7 +1734,7 @@ static int scc_net_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd)
1734 if (!Ivec[hwcfg.irq].used && hwcfg.irq) 1734 if (!Ivec[hwcfg.irq].used && hwcfg.irq)
1735 { 1735 {
1736 if (request_irq(hwcfg.irq, scc_isr, 1736 if (request_irq(hwcfg.irq, scc_isr,
1737 IRQF_DISABLED, "AX.25 SCC", 1737 0, "AX.25 SCC",
1738 (void *)(long) hwcfg.irq)) 1738 (void *)(long) hwcfg.irq))
1739 printk(KERN_WARNING "z8530drv: warning, cannot get IRQ %d\n", hwcfg.irq); 1739 printk(KERN_WARNING "z8530drv: warning, cannot get IRQ %d\n", hwcfg.irq);
1740 else 1740 else
diff --git a/drivers/net/hamradio/yam.c b/drivers/net/hamradio/yam.c
index 5af1c3e5032a..1971411574db 100644
--- a/drivers/net/hamradio/yam.c
+++ b/drivers/net/hamradio/yam.c
@@ -888,7 +888,7 @@ static int yam_open(struct net_device *dev)
888 goto out_release_base; 888 goto out_release_base;
889 } 889 }
890 outb(0, IER(dev->base_addr)); 890 outb(0, IER(dev->base_addr));
891 if (request_irq(dev->irq, yam_interrupt, IRQF_DISABLED | IRQF_SHARED, dev->name, dev)) { 891 if (request_irq(dev->irq, yam_interrupt, IRQF_SHARED, dev->name, dev)) {
892 printk(KERN_ERR "%s: irq %d busy\n", dev->name, dev->irq); 892 printk(KERN_ERR "%s: irq %d busy\n", dev->name, dev->irq);
893 ret = -EBUSY; 893 ret = -EBUSY;
894 goto out_release_base; 894 goto out_release_base;
diff --git a/drivers/net/irda/bfin_sir.c b/drivers/net/irda/bfin_sir.c
index c74f384c87d5..303c4bd26e17 100644
--- a/drivers/net/irda/bfin_sir.c
+++ b/drivers/net/irda/bfin_sir.c
@@ -411,12 +411,12 @@ static int bfin_sir_startup(struct bfin_sir_port *port, struct net_device *dev)
411 411
412#else 412#else
413 413
414 if (request_irq(port->irq, bfin_sir_rx_int, IRQF_DISABLED, "BFIN_SIR_RX", dev)) { 414 if (request_irq(port->irq, bfin_sir_rx_int, 0, "BFIN_SIR_RX", dev)) {
415 dev_warn(&dev->dev, "Unable to attach SIR RX interrupt\n"); 415 dev_warn(&dev->dev, "Unable to attach SIR RX interrupt\n");
416 return -EBUSY; 416 return -EBUSY;
417 } 417 }
418 418
419 if (request_irq(port->irq+1, bfin_sir_tx_int, IRQF_DISABLED, "BFIN_SIR_TX", dev)) { 419 if (request_irq(port->irq+1, bfin_sir_tx_int, 0, "BFIN_SIR_TX", dev)) {
420 dev_warn(&dev->dev, "Unable to attach SIR TX interrupt\n"); 420 dev_warn(&dev->dev, "Unable to attach SIR TX interrupt\n");
421 free_irq(port->irq, dev); 421 free_irq(port->irq, dev);
422 return -EBUSY; 422 return -EBUSY;
diff --git a/drivers/net/irda/donauboe.c b/drivers/net/irda/donauboe.c
index 31bcb98ef356..768dfe9a9315 100644
--- a/drivers/net/irda/donauboe.c
+++ b/drivers/net/irda/donauboe.c
@@ -1352,7 +1352,7 @@ toshoboe_net_open (struct net_device *dev)
1352 return 0; 1352 return 0;
1353 1353
1354 rc = request_irq (self->io.irq, toshoboe_interrupt, 1354 rc = request_irq (self->io.irq, toshoboe_interrupt,
1355 IRQF_SHARED | IRQF_DISABLED, dev->name, self); 1355 IRQF_SHARED, dev->name, self);
1356 if (rc) 1356 if (rc)
1357 return rc; 1357 return rc;
1358 1358
@@ -1559,7 +1559,7 @@ toshoboe_open (struct pci_dev *pci_dev, const struct pci_device_id *pdid)
1559 self->io.fir_base = self->base; 1559 self->io.fir_base = self->base;
1560 self->io.fir_ext = OBOE_IO_EXTENT; 1560 self->io.fir_ext = OBOE_IO_EXTENT;
1561 self->io.irq = pci_dev->irq; 1561 self->io.irq = pci_dev->irq;
1562 self->io.irqflags = IRQF_SHARED | IRQF_DISABLED; 1562 self->io.irqflags = IRQF_SHARED;
1563 1563
1564 self->speed = self->io.speed = 9600; 1564 self->speed = self->io.speed = 9600;
1565 self->async = 0; 1565 self->async = 0;
diff --git a/drivers/net/irda/sh_irda.c b/drivers/net/irda/sh_irda.c
index 4455425f1c77..ff45cd0d60e8 100644
--- a/drivers/net/irda/sh_irda.c
+++ b/drivers/net/irda/sh_irda.c
@@ -804,7 +804,7 @@ static int sh_irda_probe(struct platform_device *pdev)
804 goto err_mem_4; 804 goto err_mem_4;
805 805
806 platform_set_drvdata(pdev, ndev); 806 platform_set_drvdata(pdev, ndev);
807 err = request_irq(irq, sh_irda_irq, IRQF_DISABLED, "sh_irda", self); 807 err = request_irq(irq, sh_irda_irq, 0, "sh_irda", self);
808 if (err) { 808 if (err) {
809 dev_warn(&pdev->dev, "Unable to attach sh_irda interrupt\n"); 809 dev_warn(&pdev->dev, "Unable to attach sh_irda interrupt\n");
810 goto err_mem_4; 810 goto err_mem_4;
diff --git a/drivers/net/irda/sh_sir.c b/drivers/net/irda/sh_sir.c
index 89682b49900f..8d9ae5a086d5 100644
--- a/drivers/net/irda/sh_sir.c
+++ b/drivers/net/irda/sh_sir.c
@@ -761,7 +761,7 @@ static int sh_sir_probe(struct platform_device *pdev)
761 goto err_mem_4; 761 goto err_mem_4;
762 762
763 platform_set_drvdata(pdev, ndev); 763 platform_set_drvdata(pdev, ndev);
764 err = request_irq(irq, sh_sir_irq, IRQF_DISABLED, "sh_sir", self); 764 err = request_irq(irq, sh_sir_irq, 0, "sh_sir", self);
765 if (err) { 765 if (err) {
766 dev_warn(&pdev->dev, "Unable to attach sh_sir interrupt\n"); 766 dev_warn(&pdev->dev, "Unable to attach sh_sir interrupt\n");
767 goto err_mem_4; 767 goto err_mem_4;
diff --git a/drivers/net/irda/sir-dev.h b/drivers/net/irda/sir-dev.h
index 6d5b1e2b1289..f50b9c1c0639 100644
--- a/drivers/net/irda/sir-dev.h
+++ b/drivers/net/irda/sir-dev.h
@@ -102,28 +102,29 @@ struct sir_driver {
102 102
103/* exported */ 103/* exported */
104 104
105extern int irda_register_dongle(struct dongle_driver *new); 105int irda_register_dongle(struct dongle_driver *new);
106extern int irda_unregister_dongle(struct dongle_driver *drv); 106int irda_unregister_dongle(struct dongle_driver *drv);
107 107
108extern struct sir_dev * sirdev_get_instance(const struct sir_driver *drv, const char *name); 108struct sir_dev *sirdev_get_instance(const struct sir_driver *drv,
109extern int sirdev_put_instance(struct sir_dev *self); 109 const char *name);
110int sirdev_put_instance(struct sir_dev *self);
110 111
111extern int sirdev_set_dongle(struct sir_dev *dev, IRDA_DONGLE type); 112int sirdev_set_dongle(struct sir_dev *dev, IRDA_DONGLE type);
112extern void sirdev_write_complete(struct sir_dev *dev); 113void sirdev_write_complete(struct sir_dev *dev);
113extern int sirdev_receive(struct sir_dev *dev, const unsigned char *cp, size_t count); 114int sirdev_receive(struct sir_dev *dev, const unsigned char *cp, size_t count);
114 115
115/* low level helpers for SIR device/dongle setup */ 116/* low level helpers for SIR device/dongle setup */
116extern int sirdev_raw_write(struct sir_dev *dev, const char *buf, int len); 117int sirdev_raw_write(struct sir_dev *dev, const char *buf, int len);
117extern int sirdev_raw_read(struct sir_dev *dev, char *buf, int len); 118int sirdev_raw_read(struct sir_dev *dev, char *buf, int len);
118extern int sirdev_set_dtr_rts(struct sir_dev *dev, int dtr, int rts); 119int sirdev_set_dtr_rts(struct sir_dev *dev, int dtr, int rts);
119 120
120/* not exported */ 121/* not exported */
121 122
122extern int sirdev_get_dongle(struct sir_dev *self, IRDA_DONGLE type); 123int sirdev_get_dongle(struct sir_dev *self, IRDA_DONGLE type);
123extern int sirdev_put_dongle(struct sir_dev *self); 124int sirdev_put_dongle(struct sir_dev *self);
124 125
125extern void sirdev_enable_rx(struct sir_dev *dev); 126void sirdev_enable_rx(struct sir_dev *dev);
126extern int sirdev_schedule_request(struct sir_dev *dev, int state, unsigned param); 127int sirdev_schedule_request(struct sir_dev *dev, int state, unsigned param);
127 128
128/* inline helpers */ 129/* inline helpers */
129 130
diff --git a/drivers/net/macvlan.c b/drivers/net/macvlan.c
index 9bf46bd19b87..af4aaa5893ff 100644
--- a/drivers/net/macvlan.c
+++ b/drivers/net/macvlan.c
@@ -297,7 +297,13 @@ netdev_tx_t macvlan_start_xmit(struct sk_buff *skb,
297 int ret; 297 int ret;
298 const struct macvlan_dev *vlan = netdev_priv(dev); 298 const struct macvlan_dev *vlan = netdev_priv(dev);
299 299
300 ret = macvlan_queue_xmit(skb, dev); 300 if (vlan->fwd_priv) {
301 skb->dev = vlan->lowerdev;
302 ret = dev_hard_start_xmit(skb, skb->dev, NULL, vlan->fwd_priv);
303 } else {
304 ret = macvlan_queue_xmit(skb, dev);
305 }
306
301 if (likely(ret == NET_XMIT_SUCCESS || ret == NET_XMIT_CN)) { 307 if (likely(ret == NET_XMIT_SUCCESS || ret == NET_XMIT_CN)) {
302 struct macvlan_pcpu_stats *pcpu_stats; 308 struct macvlan_pcpu_stats *pcpu_stats;
303 309
@@ -347,6 +353,21 @@ static int macvlan_open(struct net_device *dev)
347 goto hash_add; 353 goto hash_add;
348 } 354 }
349 355
356 if (lowerdev->features & NETIF_F_HW_L2FW_DOFFLOAD) {
357 vlan->fwd_priv =
358 lowerdev->netdev_ops->ndo_dfwd_add_station(lowerdev, dev);
359
360 /* If we get a NULL pointer back, or if we get an error
361 * then we should just fall through to the non accelerated path
362 */
363 if (IS_ERR_OR_NULL(vlan->fwd_priv)) {
364 vlan->fwd_priv = NULL;
365 } else {
366 dev->features &= ~NETIF_F_LLTX;
367 return 0;
368 }
369 }
370
350 err = -EBUSY; 371 err = -EBUSY;
351 if (macvlan_addr_busy(vlan->port, dev->dev_addr)) 372 if (macvlan_addr_busy(vlan->port, dev->dev_addr))
352 goto out; 373 goto out;
@@ -367,6 +388,11 @@ hash_add:
367del_unicast: 388del_unicast:
368 dev_uc_del(lowerdev, dev->dev_addr); 389 dev_uc_del(lowerdev, dev->dev_addr);
369out: 390out:
391 if (vlan->fwd_priv) {
392 lowerdev->netdev_ops->ndo_dfwd_del_station(lowerdev,
393 vlan->fwd_priv);
394 vlan->fwd_priv = NULL;
395 }
370 return err; 396 return err;
371} 397}
372 398
@@ -375,6 +401,13 @@ static int macvlan_stop(struct net_device *dev)
375 struct macvlan_dev *vlan = netdev_priv(dev); 401 struct macvlan_dev *vlan = netdev_priv(dev);
376 struct net_device *lowerdev = vlan->lowerdev; 402 struct net_device *lowerdev = vlan->lowerdev;
377 403
404 if (vlan->fwd_priv) {
405 lowerdev->netdev_ops->ndo_dfwd_del_station(lowerdev,
406 vlan->fwd_priv);
407 vlan->fwd_priv = NULL;
408 return 0;
409 }
410
378 dev_uc_unsync(lowerdev, dev); 411 dev_uc_unsync(lowerdev, dev);
379 dev_mc_unsync(lowerdev, dev); 412 dev_mc_unsync(lowerdev, dev);
380 413
@@ -828,22 +861,22 @@ int macvlan_common_newlink(struct net *src_net, struct net_device *dev,
828 eth_hw_addr_inherit(dev, lowerdev); 861 eth_hw_addr_inherit(dev, lowerdev);
829 } 862 }
830 863
864 port->count += 1;
865 err = register_netdevice(dev);
866 if (err < 0)
867 goto destroy_port;
868
869 dev->priv_flags |= IFF_MACVLAN;
831 err = netdev_upper_dev_link(lowerdev, dev); 870 err = netdev_upper_dev_link(lowerdev, dev);
832 if (err) 871 if (err)
833 goto destroy_port; 872 goto destroy_port;
834 873
835 port->count += 1;
836 err = register_netdevice(dev);
837 if (err < 0)
838 goto upper_dev_unlink;
839 874
840 list_add_tail_rcu(&vlan->list, &port->vlans); 875 list_add_tail_rcu(&vlan->list, &port->vlans);
841 netif_stacked_transfer_operstate(lowerdev, dev); 876 netif_stacked_transfer_operstate(lowerdev, dev);
842 877
843 return 0; 878 return 0;
844 879
845upper_dev_unlink:
846 netdev_upper_dev_unlink(lowerdev, dev);
847destroy_port: 880destroy_port:
848 port->count -= 1; 881 port->count -= 1;
849 if (!port->count) 882 if (!port->count)
diff --git a/drivers/net/netconsole.c b/drivers/net/netconsole.c
index c9a15925a1f7..ba2f5e710af1 100644
--- a/drivers/net/netconsole.c
+++ b/drivers/net/netconsole.c
@@ -34,6 +34,8 @@
34 * 34 *
35 ****************************************************************/ 35 ****************************************************************/
36 36
37#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
38
37#include <linux/mm.h> 39#include <linux/mm.h>
38#include <linux/init.h> 40#include <linux/init.h>
39#include <linux/module.h> 41#include <linux/module.h>
@@ -320,8 +322,8 @@ static ssize_t store_enabled(struct netconsole_target *nt,
320 if (enabled < 0 || enabled > 1) 322 if (enabled < 0 || enabled > 1)
321 return -EINVAL; 323 return -EINVAL;
322 if (enabled == nt->enabled) { 324 if (enabled == nt->enabled) {
323 printk(KERN_INFO "netconsole: network logging has already %s\n", 325 pr_info("network logging has already %s\n",
324 nt->enabled ? "started" : "stopped"); 326 nt->enabled ? "started" : "stopped");
325 return -EINVAL; 327 return -EINVAL;
326 } 328 }
327 329
@@ -336,7 +338,7 @@ static ssize_t store_enabled(struct netconsole_target *nt,
336 if (err) 338 if (err)
337 return err; 339 return err;
338 340
339 printk(KERN_INFO "netconsole: network logging started\n"); 341 pr_info("netconsole: network logging started\n");
340 } else { /* 0 */ 342 } else { /* 0 */
341 /* We need to disable the netconsole before cleaning it up 343 /* We need to disable the netconsole before cleaning it up
342 * otherwise we might end up in write_msg() with 344 * otherwise we might end up in write_msg() with
@@ -360,9 +362,8 @@ static ssize_t store_dev_name(struct netconsole_target *nt,
360 size_t len; 362 size_t len;
361 363
362 if (nt->enabled) { 364 if (nt->enabled) {
363 printk(KERN_ERR "netconsole: target (%s) is enabled, " 365 pr_err("target (%s) is enabled, disable to update parameters\n",
364 "disable to update parameters\n", 366 config_item_name(&nt->item));
365 config_item_name(&nt->item));
366 return -EINVAL; 367 return -EINVAL;
367 } 368 }
368 369
@@ -383,9 +384,8 @@ static ssize_t store_local_port(struct netconsole_target *nt,
383 int rv; 384 int rv;
384 385
385 if (nt->enabled) { 386 if (nt->enabled) {
386 printk(KERN_ERR "netconsole: target (%s) is enabled, " 387 pr_err("target (%s) is enabled, disable to update parameters\n",
387 "disable to update parameters\n", 388 config_item_name(&nt->item));
388 config_item_name(&nt->item));
389 return -EINVAL; 389 return -EINVAL;
390 } 390 }
391 391
@@ -402,9 +402,8 @@ static ssize_t store_remote_port(struct netconsole_target *nt,
402 int rv; 402 int rv;
403 403
404 if (nt->enabled) { 404 if (nt->enabled) {
405 printk(KERN_ERR "netconsole: target (%s) is enabled, " 405 pr_err("target (%s) is enabled, disable to update parameters\n",
406 "disable to update parameters\n", 406 config_item_name(&nt->item));
407 config_item_name(&nt->item));
408 return -EINVAL; 407 return -EINVAL;
409 } 408 }
410 409
@@ -419,9 +418,8 @@ static ssize_t store_local_ip(struct netconsole_target *nt,
419 size_t count) 418 size_t count)
420{ 419{
421 if (nt->enabled) { 420 if (nt->enabled) {
422 printk(KERN_ERR "netconsole: target (%s) is enabled, " 421 pr_err("target (%s) is enabled, disable to update parameters\n",
423 "disable to update parameters\n", 422 config_item_name(&nt->item));
424 config_item_name(&nt->item));
425 return -EINVAL; 423 return -EINVAL;
426 } 424 }
427 425
@@ -429,7 +427,7 @@ static ssize_t store_local_ip(struct netconsole_target *nt,
429 const char *end; 427 const char *end;
430 if (in6_pton(buf, count, nt->np.local_ip.in6.s6_addr, -1, &end) > 0) { 428 if (in6_pton(buf, count, nt->np.local_ip.in6.s6_addr, -1, &end) > 0) {
431 if (*end && *end != '\n') { 429 if (*end && *end != '\n') {
432 printk(KERN_ERR "netconsole: invalid IPv6 address at: <%c>\n", *end); 430 pr_err("invalid IPv6 address at: <%c>\n", *end);
433 return -EINVAL; 431 return -EINVAL;
434 } 432 }
435 nt->np.ipv6 = true; 433 nt->np.ipv6 = true;
@@ -450,9 +448,8 @@ static ssize_t store_remote_ip(struct netconsole_target *nt,
450 size_t count) 448 size_t count)
451{ 449{
452 if (nt->enabled) { 450 if (nt->enabled) {
453 printk(KERN_ERR "netconsole: target (%s) is enabled, " 451 pr_err("target (%s) is enabled, disable to update parameters\n",
454 "disable to update parameters\n", 452 config_item_name(&nt->item));
455 config_item_name(&nt->item));
456 return -EINVAL; 453 return -EINVAL;
457 } 454 }
458 455
@@ -460,7 +457,7 @@ static ssize_t store_remote_ip(struct netconsole_target *nt,
460 const char *end; 457 const char *end;
461 if (in6_pton(buf, count, nt->np.remote_ip.in6.s6_addr, -1, &end) > 0) { 458 if (in6_pton(buf, count, nt->np.remote_ip.in6.s6_addr, -1, &end) > 0) {
462 if (*end && *end != '\n') { 459 if (*end && *end != '\n') {
463 printk(KERN_ERR "netconsole: invalid IPv6 address at: <%c>\n", *end); 460 pr_err("invalid IPv6 address at: <%c>\n", *end);
464 return -EINVAL; 461 return -EINVAL;
465 } 462 }
466 nt->np.ipv6 = true; 463 nt->np.ipv6 = true;
@@ -483,9 +480,8 @@ static ssize_t store_remote_mac(struct netconsole_target *nt,
483 u8 remote_mac[ETH_ALEN]; 480 u8 remote_mac[ETH_ALEN];
484 481
485 if (nt->enabled) { 482 if (nt->enabled) {
486 printk(KERN_ERR "netconsole: target (%s) is enabled, " 483 pr_err("target (%s) is enabled, disable to update parameters\n",
487 "disable to update parameters\n", 484 config_item_name(&nt->item));
488 config_item_name(&nt->item));
489 return -EINVAL; 485 return -EINVAL;
490 } 486 }
491 487
@@ -708,19 +704,20 @@ restart:
708 } 704 }
709 spin_unlock_irqrestore(&target_list_lock, flags); 705 spin_unlock_irqrestore(&target_list_lock, flags);
710 if (stopped) { 706 if (stopped) {
711 printk(KERN_INFO "netconsole: network logging stopped on " 707 const char *msg = "had an event";
712 "interface %s as it ", dev->name);
713 switch (event) { 708 switch (event) {
714 case NETDEV_UNREGISTER: 709 case NETDEV_UNREGISTER:
715 printk(KERN_CONT "unregistered\n"); 710 msg = "unregistered";
716 break; 711 break;
717 case NETDEV_RELEASE: 712 case NETDEV_RELEASE:
718 printk(KERN_CONT "released slaves\n"); 713 msg = "released slaves";
719 break; 714 break;
720 case NETDEV_JOIN: 715 case NETDEV_JOIN:
721 printk(KERN_CONT "is joining a master device\n"); 716 msg = "is joining a master device";
722 break; 717 break;
723 } 718 }
719 pr_info("network logging stopped on interface %s as it %s\n",
720 dev->name, msg);
724 } 721 }
725 722
726done: 723done:
@@ -806,7 +803,7 @@ static int __init init_netconsole(void)
806 goto undonotifier; 803 goto undonotifier;
807 804
808 register_console(&netconsole); 805 register_console(&netconsole);
809 printk(KERN_INFO "netconsole: network logging started\n"); 806 pr_info("network logging started\n");
810 807
811 return err; 808 return err;
812 809
@@ -814,7 +811,7 @@ undonotifier:
814 unregister_netdevice_notifier(&netconsole_netdev_notifier); 811 unregister_netdevice_notifier(&netconsole_netdev_notifier);
815 812
816fail: 813fail:
817 printk(KERN_ERR "netconsole: cleaning up\n"); 814 pr_err("cleaning up\n");
818 815
819 /* 816 /*
820 * Remove all targets and destroy them (only targets created 817 * Remove all targets and destroy them (only targets created
diff --git a/drivers/net/phy/Kconfig b/drivers/net/phy/Kconfig
index 342561ad3158..9b5d46c03eed 100644
--- a/drivers/net/phy/Kconfig
+++ b/drivers/net/phy/Kconfig
@@ -154,6 +154,13 @@ config MDIO_SUN4I
154 interface units of the Allwinner SoC that have an EMAC (A10, 154 interface units of the Allwinner SoC that have an EMAC (A10,
155 A12, A10s, etc.) 155 A12, A10s, etc.)
156 156
157config MDIO_MOXART
158 tristate "MOXA ART MDIO interface support"
159 depends on ARCH_MOXART
160 help
161 This driver supports the MDIO interface found in the network
162 interface units of the MOXA ART SoC
163
157config MDIO_BUS_MUX 164config MDIO_BUS_MUX
158 tristate 165 tristate
159 depends on OF_MDIO 166 depends on OF_MDIO
diff --git a/drivers/net/phy/Makefile b/drivers/net/phy/Makefile
index 23a2ab2e847e..9013dfa12aa3 100644
--- a/drivers/net/phy/Makefile
+++ b/drivers/net/phy/Makefile
@@ -31,3 +31,4 @@ obj-$(CONFIG_MDIO_BUS_MUX) += mdio-mux.o
31obj-$(CONFIG_MDIO_BUS_MUX_GPIO) += mdio-mux-gpio.o 31obj-$(CONFIG_MDIO_BUS_MUX_GPIO) += mdio-mux-gpio.o
32obj-$(CONFIG_MDIO_BUS_MUX_MMIOREG) += mdio-mux-mmioreg.o 32obj-$(CONFIG_MDIO_BUS_MUX_MMIOREG) += mdio-mux-mmioreg.o
33obj-$(CONFIG_MDIO_SUN4I) += mdio-sun4i.o 33obj-$(CONFIG_MDIO_SUN4I) += mdio-sun4i.o
34obj-$(CONFIG_MDIO_MOXART) += mdio-moxart.o
diff --git a/drivers/net/phy/at803x.c b/drivers/net/phy/at803x.c
index ac22283aaf23..bc71947b1ec3 100644
--- a/drivers/net/phy/at803x.c
+++ b/drivers/net/phy/at803x.c
@@ -100,6 +100,45 @@ static void at803x_get_wol(struct phy_device *phydev,
100 wol->wolopts |= WAKE_MAGIC; 100 wol->wolopts |= WAKE_MAGIC;
101} 101}
102 102
103static int at803x_suspend(struct phy_device *phydev)
104{
105 int value;
106 int wol_enabled;
107
108 mutex_lock(&phydev->lock);
109
110 value = phy_read(phydev, AT803X_INTR_ENABLE);
111 wol_enabled = value & AT803X_WOL_ENABLE;
112
113 value = phy_read(phydev, MII_BMCR);
114
115 if (wol_enabled)
116 value |= BMCR_ISOLATE;
117 else
118 value |= BMCR_PDOWN;
119
120 phy_write(phydev, MII_BMCR, value);
121
122 mutex_unlock(&phydev->lock);
123
124 return 0;
125}
126
127static int at803x_resume(struct phy_device *phydev)
128{
129 int value;
130
131 mutex_lock(&phydev->lock);
132
133 value = phy_read(phydev, MII_BMCR);
134 value &= ~(BMCR_PDOWN | BMCR_ISOLATE);
135 phy_write(phydev, MII_BMCR, value);
136
137 mutex_unlock(&phydev->lock);
138
139 return 0;
140}
141
103static int at803x_config_init(struct phy_device *phydev) 142static int at803x_config_init(struct phy_device *phydev)
104{ 143{
105 int val; 144 int val;
@@ -161,10 +200,12 @@ static struct phy_driver at803x_driver[] = {
161 .config_init = at803x_config_init, 200 .config_init = at803x_config_init,
162 .set_wol = at803x_set_wol, 201 .set_wol = at803x_set_wol,
163 .get_wol = at803x_get_wol, 202 .get_wol = at803x_get_wol,
203 .suspend = at803x_suspend,
204 .resume = at803x_resume,
164 .features = PHY_GBIT_FEATURES, 205 .features = PHY_GBIT_FEATURES,
165 .flags = PHY_HAS_INTERRUPT, 206 .flags = PHY_HAS_INTERRUPT,
166 .config_aneg = &genphy_config_aneg, 207 .config_aneg = genphy_config_aneg,
167 .read_status = &genphy_read_status, 208 .read_status = genphy_read_status,
168 .driver = { 209 .driver = {
169 .owner = THIS_MODULE, 210 .owner = THIS_MODULE,
170 }, 211 },
@@ -176,10 +217,12 @@ static struct phy_driver at803x_driver[] = {
176 .config_init = at803x_config_init, 217 .config_init = at803x_config_init,
177 .set_wol = at803x_set_wol, 218 .set_wol = at803x_set_wol,
178 .get_wol = at803x_get_wol, 219 .get_wol = at803x_get_wol,
220 .suspend = at803x_suspend,
221 .resume = at803x_resume,
179 .features = PHY_GBIT_FEATURES, 222 .features = PHY_GBIT_FEATURES,
180 .flags = PHY_HAS_INTERRUPT, 223 .flags = PHY_HAS_INTERRUPT,
181 .config_aneg = &genphy_config_aneg, 224 .config_aneg = genphy_config_aneg,
182 .read_status = &genphy_read_status, 225 .read_status = genphy_read_status,
183 .driver = { 226 .driver = {
184 .owner = THIS_MODULE, 227 .owner = THIS_MODULE,
185 }, 228 },
@@ -191,10 +234,12 @@ static struct phy_driver at803x_driver[] = {
191 .config_init = at803x_config_init, 234 .config_init = at803x_config_init,
192 .set_wol = at803x_set_wol, 235 .set_wol = at803x_set_wol,
193 .get_wol = at803x_get_wol, 236 .get_wol = at803x_get_wol,
237 .suspend = at803x_suspend,
238 .resume = at803x_resume,
194 .features = PHY_GBIT_FEATURES, 239 .features = PHY_GBIT_FEATURES,
195 .flags = PHY_HAS_INTERRUPT, 240 .flags = PHY_HAS_INTERRUPT,
196 .config_aneg = &genphy_config_aneg, 241 .config_aneg = genphy_config_aneg,
197 .read_status = &genphy_read_status, 242 .read_status = genphy_read_status,
198 .driver = { 243 .driver = {
199 .owner = THIS_MODULE, 244 .owner = THIS_MODULE,
200 }, 245 },
diff --git a/drivers/net/phy/marvell.c b/drivers/net/phy/marvell.c
index 2e91477362d4..2e3c778ea9bf 100644
--- a/drivers/net/phy/marvell.c
+++ b/drivers/net/phy/marvell.c
@@ -34,9 +34,9 @@
34#include <linux/marvell_phy.h> 34#include <linux/marvell_phy.h>
35#include <linux/of.h> 35#include <linux/of.h>
36 36
37#include <asm/io.h> 37#include <linux/io.h>
38#include <asm/irq.h> 38#include <asm/irq.h>
39#include <asm/uaccess.h> 39#include <linux/uaccess.h>
40 40
41#define MII_MARVELL_PHY_PAGE 22 41#define MII_MARVELL_PHY_PAGE 22
42 42
diff --git a/drivers/net/phy/mdio-moxart.c b/drivers/net/phy/mdio-moxart.c
new file mode 100644
index 000000000000..a5741cb0304e
--- /dev/null
+++ b/drivers/net/phy/mdio-moxart.c
@@ -0,0 +1,201 @@
1/* MOXA ART Ethernet (RTL8201CP) MDIO interface driver
2 *
3 * Copyright (C) 2013 Jonas Jensen <jonas.jensen@gmail.com>
4 *
5 * This file is licensed under the terms of the GNU General Public
6 * License version 2. This program is licensed "as is" without any
7 * warranty of any kind, whether express or implied.
8 */
9
10#include <linux/delay.h>
11#include <linux/init.h>
12#include <linux/kernel.h>
13#include <linux/module.h>
14#include <linux/mutex.h>
15#include <linux/of_address.h>
16#include <linux/of_mdio.h>
17#include <linux/phy.h>
18#include <linux/platform_device.h>
19#include <linux/regulator/consumer.h>
20
21#define REG_PHY_CTRL 0
22#define REG_PHY_WRITE_DATA 4
23
24/* REG_PHY_CTRL */
25#define MIIWR BIT(27) /* init write sequence (auto cleared)*/
26#define MIIRD BIT(26)
27#define REGAD_MASK 0x3e00000
28#define PHYAD_MASK 0x1f0000
29#define MIIRDATA_MASK 0xffff
30
31/* REG_PHY_WRITE_DATA */
32#define MIIWDATA_MASK 0xffff
33
34struct moxart_mdio_data {
35 void __iomem *base;
36};
37
38static int moxart_mdio_read(struct mii_bus *bus, int mii_id, int regnum)
39{
40 struct moxart_mdio_data *data = bus->priv;
41 u32 ctrl = 0;
42 unsigned int count = 5;
43
44 dev_dbg(&bus->dev, "%s\n", __func__);
45
46 ctrl |= MIIRD | ((mii_id << 16) & PHYAD_MASK) |
47 ((regnum << 21) & REGAD_MASK);
48
49 writel(ctrl, data->base + REG_PHY_CTRL);
50
51 do {
52 ctrl = readl(data->base + REG_PHY_CTRL);
53
54 if (!(ctrl & MIIRD))
55 return ctrl & MIIRDATA_MASK;
56
57 mdelay(10);
58 count--;
59 } while (count > 0);
60
61 dev_dbg(&bus->dev, "%s timed out\n", __func__);
62
63 return -ETIMEDOUT;
64}
65
66static int moxart_mdio_write(struct mii_bus *bus, int mii_id,
67 int regnum, u16 value)
68{
69 struct moxart_mdio_data *data = bus->priv;
70 u32 ctrl = 0;
71 unsigned int count = 5;
72
73 dev_dbg(&bus->dev, "%s\n", __func__);
74
75 ctrl |= MIIWR | ((mii_id << 16) & PHYAD_MASK) |
76 ((regnum << 21) & REGAD_MASK);
77
78 value &= MIIWDATA_MASK;
79
80 writel(value, data->base + REG_PHY_WRITE_DATA);
81 writel(ctrl, data->base + REG_PHY_CTRL);
82
83 do {
84 ctrl = readl(data->base + REG_PHY_CTRL);
85
86 if (!(ctrl & MIIWR))
87 return 0;
88
89 mdelay(10);
90 count--;
91 } while (count > 0);
92
93 dev_dbg(&bus->dev, "%s timed out\n", __func__);
94
95 return -ETIMEDOUT;
96}
97
98static int moxart_mdio_reset(struct mii_bus *bus)
99{
100 int data, i;
101
102 for (i = 0; i < PHY_MAX_ADDR; i++) {
103 data = moxart_mdio_read(bus, i, MII_BMCR);
104 if (data < 0)
105 continue;
106
107 data |= BMCR_RESET;
108 if (moxart_mdio_write(bus, i, MII_BMCR, data) < 0)
109 continue;
110 }
111
112 return 0;
113}
114
115static int moxart_mdio_probe(struct platform_device *pdev)
116{
117 struct device_node *np = pdev->dev.of_node;
118 struct mii_bus *bus;
119 struct moxart_mdio_data *data;
120 struct resource *res;
121 int ret, i;
122
123 bus = mdiobus_alloc_size(sizeof(*data));
124 if (!bus)
125 return -ENOMEM;
126
127 bus->name = "MOXA ART Ethernet MII";
128 bus->read = &moxart_mdio_read;
129 bus->write = &moxart_mdio_write;
130 bus->reset = &moxart_mdio_reset;
131 snprintf(bus->id, MII_BUS_ID_SIZE, "%s-%d-mii", pdev->name, pdev->id);
132 bus->parent = &pdev->dev;
133
134 bus->irq = devm_kzalloc(&pdev->dev, sizeof(int) * PHY_MAX_ADDR,
135 GFP_KERNEL);
136 if (!bus->irq) {
137 ret = -ENOMEM;
138 goto err_out_free_mdiobus;
139 }
140
141 /* Setting PHY_IGNORE_INTERRUPT here even if it has no effect,
142 * of_mdiobus_register() sets these PHY_POLL.
143 * Ideally, the interrupt from MAC controller could be used to
144 * detect link state changes, not polling, i.e. if there was
145 * a way phy_driver could set PHY_HAS_INTERRUPT but have that
146 * interrupt handled in ethernet drivercode.
147 */
148 for (i = 0; i < PHY_MAX_ADDR; i++)
149 bus->irq[i] = PHY_IGNORE_INTERRUPT;
150
151 data = bus->priv;
152 res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
153 data->base = devm_ioremap_resource(&pdev->dev, res);
154 if (IS_ERR(data->base)) {
155 ret = PTR_ERR(data->base);
156 goto err_out_free_mdiobus;
157 }
158
159 ret = of_mdiobus_register(bus, np);
160 if (ret < 0)
161 goto err_out_free_mdiobus;
162
163 platform_set_drvdata(pdev, bus);
164
165 return 0;
166
167err_out_free_mdiobus:
168 mdiobus_free(bus);
169 return ret;
170}
171
172static int moxart_mdio_remove(struct platform_device *pdev)
173{
174 struct mii_bus *bus = platform_get_drvdata(pdev);
175
176 mdiobus_unregister(bus);
177 mdiobus_free(bus);
178
179 return 0;
180}
181
182static const struct of_device_id moxart_mdio_dt_ids[] = {
183 { .compatible = "moxa,moxart-mdio" },
184 { }
185};
186MODULE_DEVICE_TABLE(of, moxart_mdio_dt_ids);
187
188static struct platform_driver moxart_mdio_driver = {
189 .probe = moxart_mdio_probe,
190 .remove = moxart_mdio_remove,
191 .driver = {
192 .name = "moxart-mdio",
193 .of_match_table = moxart_mdio_dt_ids,
194 },
195};
196
197module_platform_driver(moxart_mdio_driver);
198
199MODULE_DESCRIPTION("MOXA ART MDIO interface driver");
200MODULE_AUTHOR("Jonas Jensen <jonas.jensen@gmail.com>");
201MODULE_LICENSE("GPL");
diff --git a/drivers/net/phy/micrel.c b/drivers/net/phy/micrel.c
index c31aad0004cb..3ae28f420868 100644
--- a/drivers/net/phy/micrel.c
+++ b/drivers/net/phy/micrel.c
@@ -287,6 +287,8 @@ static struct phy_driver ksphy_driver[] = {
287 .read_status = genphy_read_status, 287 .read_status = genphy_read_status,
288 .ack_interrupt = kszphy_ack_interrupt, 288 .ack_interrupt = kszphy_ack_interrupt,
289 .config_intr = ks8737_config_intr, 289 .config_intr = ks8737_config_intr,
290 .suspend = genphy_suspend,
291 .resume = genphy_resume,
290 .driver = { .owner = THIS_MODULE,}, 292 .driver = { .owner = THIS_MODULE,},
291}, { 293}, {
292 .phy_id = PHY_ID_KSZ8021, 294 .phy_id = PHY_ID_KSZ8021,
@@ -300,6 +302,8 @@ static struct phy_driver ksphy_driver[] = {
300 .read_status = genphy_read_status, 302 .read_status = genphy_read_status,
301 .ack_interrupt = kszphy_ack_interrupt, 303 .ack_interrupt = kszphy_ack_interrupt,
302 .config_intr = kszphy_config_intr, 304 .config_intr = kszphy_config_intr,
305 .suspend = genphy_suspend,
306 .resume = genphy_resume,
303 .driver = { .owner = THIS_MODULE,}, 307 .driver = { .owner = THIS_MODULE,},
304}, { 308}, {
305 .phy_id = PHY_ID_KSZ8031, 309 .phy_id = PHY_ID_KSZ8031,
@@ -313,6 +317,8 @@ static struct phy_driver ksphy_driver[] = {
313 .read_status = genphy_read_status, 317 .read_status = genphy_read_status,
314 .ack_interrupt = kszphy_ack_interrupt, 318 .ack_interrupt = kszphy_ack_interrupt,
315 .config_intr = kszphy_config_intr, 319 .config_intr = kszphy_config_intr,
320 .suspend = genphy_suspend,
321 .resume = genphy_resume,
316 .driver = { .owner = THIS_MODULE,}, 322 .driver = { .owner = THIS_MODULE,},
317}, { 323}, {
318 .phy_id = PHY_ID_KSZ8041, 324 .phy_id = PHY_ID_KSZ8041,
@@ -326,6 +332,8 @@ static struct phy_driver ksphy_driver[] = {
326 .read_status = genphy_read_status, 332 .read_status = genphy_read_status,
327 .ack_interrupt = kszphy_ack_interrupt, 333 .ack_interrupt = kszphy_ack_interrupt,
328 .config_intr = kszphy_config_intr, 334 .config_intr = kszphy_config_intr,
335 .suspend = genphy_suspend,
336 .resume = genphy_resume,
329 .driver = { .owner = THIS_MODULE,}, 337 .driver = { .owner = THIS_MODULE,},
330}, { 338}, {
331 .phy_id = PHY_ID_KSZ8051, 339 .phy_id = PHY_ID_KSZ8051,
@@ -339,6 +347,8 @@ static struct phy_driver ksphy_driver[] = {
339 .read_status = genphy_read_status, 347 .read_status = genphy_read_status,
340 .ack_interrupt = kszphy_ack_interrupt, 348 .ack_interrupt = kszphy_ack_interrupt,
341 .config_intr = kszphy_config_intr, 349 .config_intr = kszphy_config_intr,
350 .suspend = genphy_suspend,
351 .resume = genphy_resume,
342 .driver = { .owner = THIS_MODULE,}, 352 .driver = { .owner = THIS_MODULE,},
343}, { 353}, {
344 .phy_id = PHY_ID_KSZ8001, 354 .phy_id = PHY_ID_KSZ8001,
@@ -351,6 +361,8 @@ static struct phy_driver ksphy_driver[] = {
351 .read_status = genphy_read_status, 361 .read_status = genphy_read_status,
352 .ack_interrupt = kszphy_ack_interrupt, 362 .ack_interrupt = kszphy_ack_interrupt,
353 .config_intr = kszphy_config_intr, 363 .config_intr = kszphy_config_intr,
364 .suspend = genphy_suspend,
365 .resume = genphy_resume,
354 .driver = { .owner = THIS_MODULE,}, 366 .driver = { .owner = THIS_MODULE,},
355}, { 367}, {
356 .phy_id = PHY_ID_KSZ8081, 368 .phy_id = PHY_ID_KSZ8081,
@@ -363,6 +375,8 @@ static struct phy_driver ksphy_driver[] = {
363 .read_status = genphy_read_status, 375 .read_status = genphy_read_status,
364 .ack_interrupt = kszphy_ack_interrupt, 376 .ack_interrupt = kszphy_ack_interrupt,
365 .config_intr = kszphy_config_intr, 377 .config_intr = kszphy_config_intr,
378 .suspend = genphy_suspend,
379 .resume = genphy_resume,
366 .driver = { .owner = THIS_MODULE,}, 380 .driver = { .owner = THIS_MODULE,},
367}, { 381}, {
368 .phy_id = PHY_ID_KSZ8061, 382 .phy_id = PHY_ID_KSZ8061,
@@ -375,6 +389,8 @@ static struct phy_driver ksphy_driver[] = {
375 .read_status = genphy_read_status, 389 .read_status = genphy_read_status,
376 .ack_interrupt = kszphy_ack_interrupt, 390 .ack_interrupt = kszphy_ack_interrupt,
377 .config_intr = kszphy_config_intr, 391 .config_intr = kszphy_config_intr,
392 .suspend = genphy_suspend,
393 .resume = genphy_resume,
378 .driver = { .owner = THIS_MODULE,}, 394 .driver = { .owner = THIS_MODULE,},
379}, { 395}, {
380 .phy_id = PHY_ID_KSZ9021, 396 .phy_id = PHY_ID_KSZ9021,
@@ -387,6 +403,8 @@ static struct phy_driver ksphy_driver[] = {
387 .read_status = genphy_read_status, 403 .read_status = genphy_read_status,
388 .ack_interrupt = kszphy_ack_interrupt, 404 .ack_interrupt = kszphy_ack_interrupt,
389 .config_intr = ksz9021_config_intr, 405 .config_intr = ksz9021_config_intr,
406 .suspend = genphy_suspend,
407 .resume = genphy_resume,
390 .driver = { .owner = THIS_MODULE, }, 408 .driver = { .owner = THIS_MODULE, },
391}, { 409}, {
392 .phy_id = PHY_ID_KSZ9031, 410 .phy_id = PHY_ID_KSZ9031,
@@ -400,6 +418,8 @@ static struct phy_driver ksphy_driver[] = {
400 .read_status = genphy_read_status, 418 .read_status = genphy_read_status,
401 .ack_interrupt = kszphy_ack_interrupt, 419 .ack_interrupt = kszphy_ack_interrupt,
402 .config_intr = ksz9021_config_intr, 420 .config_intr = ksz9021_config_intr,
421 .suspend = genphy_suspend,
422 .resume = genphy_resume,
403 .driver = { .owner = THIS_MODULE, }, 423 .driver = { .owner = THIS_MODULE, },
404}, { 424}, {
405 .phy_id = PHY_ID_KSZ8873MLL, 425 .phy_id = PHY_ID_KSZ8873MLL,
@@ -410,6 +430,8 @@ static struct phy_driver ksphy_driver[] = {
410 .config_init = kszphy_config_init, 430 .config_init = kszphy_config_init,
411 .config_aneg = ksz8873mll_config_aneg, 431 .config_aneg = ksz8873mll_config_aneg,
412 .read_status = ksz8873mll_read_status, 432 .read_status = ksz8873mll_read_status,
433 .suspend = genphy_suspend,
434 .resume = genphy_resume,
413 .driver = { .owner = THIS_MODULE, }, 435 .driver = { .owner = THIS_MODULE, },
414}, { 436}, {
415 .phy_id = PHY_ID_KSZ886X, 437 .phy_id = PHY_ID_KSZ886X,
@@ -420,6 +442,8 @@ static struct phy_driver ksphy_driver[] = {
420 .config_init = kszphy_config_init, 442 .config_init = kszphy_config_init,
421 .config_aneg = genphy_config_aneg, 443 .config_aneg = genphy_config_aneg,
422 .read_status = genphy_read_status, 444 .read_status = genphy_read_status,
445 .suspend = genphy_suspend,
446 .resume = genphy_resume,
423 .driver = { .owner = THIS_MODULE, }, 447 .driver = { .owner = THIS_MODULE, },
424} }; 448} };
425 449
diff --git a/drivers/net/phy/realtek.c b/drivers/net/phy/realtek.c
index 138de837977f..fa1d69a38ccf 100644
--- a/drivers/net/phy/realtek.c
+++ b/drivers/net/phy/realtek.c
@@ -64,6 +64,18 @@ static int rtl8211e_config_intr(struct phy_device *phydev)
64 return err; 64 return err;
65} 65}
66 66
67/* RTL8201CP */
68static struct phy_driver rtl8201cp_driver = {
69 .phy_id = 0x00008201,
70 .name = "RTL8201CP Ethernet",
71 .phy_id_mask = 0x0000ffff,
72 .features = PHY_BASIC_FEATURES,
73 .flags = PHY_HAS_INTERRUPT,
74 .config_aneg = &genphy_config_aneg,
75 .read_status = &genphy_read_status,
76 .driver = { .owner = THIS_MODULE,},
77};
78
67/* RTL8211B */ 79/* RTL8211B */
68static struct phy_driver rtl8211b_driver = { 80static struct phy_driver rtl8211b_driver = {
69 .phy_id = 0x001cc912, 81 .phy_id = 0x001cc912,
@@ -98,6 +110,9 @@ static int __init realtek_init(void)
98{ 110{
99 int ret; 111 int ret;
100 112
113 ret = phy_driver_register(&rtl8201cp_driver);
114 if (ret < 0)
115 return -ENODEV;
101 ret = phy_driver_register(&rtl8211b_driver); 116 ret = phy_driver_register(&rtl8211b_driver);
102 if (ret < 0) 117 if (ret < 0)
103 return -ENODEV; 118 return -ENODEV;
diff --git a/drivers/net/plip/plip.c b/drivers/net/plip/plip.c
index 1f7bef90b467..7b4ff35c8bf7 100644
--- a/drivers/net/plip/plip.c
+++ b/drivers/net/plip/plip.c
@@ -1002,7 +1002,7 @@ plip_rewrite_address(const struct net_device *dev, struct ethhdr *eth)
1002 /* Any address will do - we take the first */ 1002 /* Any address will do - we take the first */
1003 const struct in_ifaddr *ifa = in_dev->ifa_list; 1003 const struct in_ifaddr *ifa = in_dev->ifa_list;
1004 if (ifa) { 1004 if (ifa) {
1005 memcpy(eth->h_source, dev->dev_addr, 6); 1005 memcpy(eth->h_source, dev->dev_addr, ETH_ALEN);
1006 memset(eth->h_dest, 0xfc, 2); 1006 memset(eth->h_dest, 0xfc, 2);
1007 memcpy(eth->h_dest+2, &ifa->ifa_address, 4); 1007 memcpy(eth->h_dest+2, &ifa->ifa_address, 4);
1008 } 1008 }
diff --git a/drivers/net/usb/Kconfig b/drivers/net/usb/Kconfig
index 40db31233313..85e4a01670f0 100644
--- a/drivers/net/usb/Kconfig
+++ b/drivers/net/usb/Kconfig
@@ -242,6 +242,21 @@ config USB_NET_CDC_NCM
242 * ST-Ericsson M343 HSPA Mobile Broadband Modem (reference design) 242 * ST-Ericsson M343 HSPA Mobile Broadband Modem (reference design)
243 * Ericsson F5521gw Mobile Broadband Module 243 * Ericsson F5521gw Mobile Broadband Module
244 244
245config USB_NET_HUAWEI_CDC_NCM
246 tristate "Huawei NCM embedded AT channel support"
247 depends on USB_USBNET
248 select USB_WDM
249 select USB_NET_CDC_NCM
250 help
251 This driver supports huawei-style NCM devices, that use NCM as a
252 transport for other protocols, usually an embedded AT channel.
253 Good examples are:
254 * Huawei E3131
255 * Huawei E3251
256
257 To compile this driver as a module, choose M here: the module will be
258 called huawei_cdc_ncm.ko.
259
245config USB_NET_CDC_MBIM 260config USB_NET_CDC_MBIM
246 tristate "CDC MBIM support" 261 tristate "CDC MBIM support"
247 depends on USB_USBNET 262 depends on USB_USBNET
diff --git a/drivers/net/usb/Makefile b/drivers/net/usb/Makefile
index 8b342cf992fd..b17b5e88bbaf 100644
--- a/drivers/net/usb/Makefile
+++ b/drivers/net/usb/Makefile
@@ -32,6 +32,7 @@ obj-$(CONFIG_USB_IPHETH) += ipheth.o
32obj-$(CONFIG_USB_SIERRA_NET) += sierra_net.o 32obj-$(CONFIG_USB_SIERRA_NET) += sierra_net.o
33obj-$(CONFIG_USB_NET_CX82310_ETH) += cx82310_eth.o 33obj-$(CONFIG_USB_NET_CX82310_ETH) += cx82310_eth.o
34obj-$(CONFIG_USB_NET_CDC_NCM) += cdc_ncm.o 34obj-$(CONFIG_USB_NET_CDC_NCM) += cdc_ncm.o
35obj-$(CONFIG_USB_NET_HUAWEI_CDC_NCM) += huawei_cdc_ncm.o
35obj-$(CONFIG_USB_VL600) += lg-vl600.o 36obj-$(CONFIG_USB_VL600) += lg-vl600.o
36obj-$(CONFIG_USB_NET_QMI_WWAN) += qmi_wwan.o 37obj-$(CONFIG_USB_NET_QMI_WWAN) += qmi_wwan.o
37obj-$(CONFIG_USB_NET_CDC_MBIM) += cdc_mbim.o 38obj-$(CONFIG_USB_NET_CDC_MBIM) += cdc_mbim.o
diff --git a/drivers/net/usb/catc.c b/drivers/net/usb/catc.c
index 8d5cac2d8e33..df507e6dbb9c 100644
--- a/drivers/net/usb/catc.c
+++ b/drivers/net/usb/catc.c
@@ -640,10 +640,10 @@ static void catc_set_multicast_list(struct net_device *netdev)
640{ 640{
641 struct catc *catc = netdev_priv(netdev); 641 struct catc *catc = netdev_priv(netdev);
642 struct netdev_hw_addr *ha; 642 struct netdev_hw_addr *ha;
643 u8 broadcast[6]; 643 u8 broadcast[ETH_ALEN];
644 u8 rx = RxEnable | RxPolarity | RxMultiCast; 644 u8 rx = RxEnable | RxPolarity | RxMultiCast;
645 645
646 memset(broadcast, 0xff, 6); 646 memset(broadcast, 0xff, ETH_ALEN);
647 memset(catc->multicast, 0, 64); 647 memset(catc->multicast, 0, 64);
648 648
649 catc_multicast(broadcast, catc->multicast); 649 catc_multicast(broadcast, catc->multicast);
@@ -778,7 +778,7 @@ static int catc_probe(struct usb_interface *intf, const struct usb_device_id *id
778 struct usb_device *usbdev = interface_to_usbdev(intf); 778 struct usb_device *usbdev = interface_to_usbdev(intf);
779 struct net_device *netdev; 779 struct net_device *netdev;
780 struct catc *catc; 780 struct catc *catc;
781 u8 broadcast[6]; 781 u8 broadcast[ETH_ALEN];
782 int i, pktsz; 782 int i, pktsz;
783 783
784 if (usb_set_interface(usbdev, 784 if (usb_set_interface(usbdev,
@@ -882,7 +882,7 @@ static int catc_probe(struct usb_interface *intf, const struct usb_device_id *id
882 882
883 dev_dbg(dev, "Filling the multicast list.\n"); 883 dev_dbg(dev, "Filling the multicast list.\n");
884 884
885 memset(broadcast, 0xff, 6); 885 memset(broadcast, 0xff, ETH_ALEN);
886 catc_multicast(broadcast, catc->multicast); 886 catc_multicast(broadcast, catc->multicast);
887 catc_multicast(netdev->dev_addr, catc->multicast); 887 catc_multicast(netdev->dev_addr, catc->multicast);
888 catc_write_mem(catc, 0xfa80, catc->multicast, 64); 888 catc_write_mem(catc, 0xfa80, catc->multicast, 64);
diff --git a/drivers/net/usb/cdc-phonet.c b/drivers/net/usb/cdc-phonet.c
index 7d78669000d7..6358d420e185 100644
--- a/drivers/net/usb/cdc-phonet.c
+++ b/drivers/net/usb/cdc-phonet.c
@@ -328,7 +328,7 @@ MODULE_DEVICE_TABLE(usb, usbpn_ids);
328 328
329static struct usb_driver usbpn_driver; 329static struct usb_driver usbpn_driver;
330 330
331int usbpn_probe(struct usb_interface *intf, const struct usb_device_id *id) 331static int usbpn_probe(struct usb_interface *intf, const struct usb_device_id *id)
332{ 332{
333 static const char ifname[] = "usbpn%d"; 333 static const char ifname[] = "usbpn%d";
334 const struct usb_cdc_union_desc *union_header = NULL; 334 const struct usb_cdc_union_desc *union_header = NULL;
diff --git a/drivers/net/usb/cdc_mbim.c b/drivers/net/usb/cdc_mbim.c
index 25ba7eca9a13..c9f3281506af 100644
--- a/drivers/net/usb/cdc_mbim.c
+++ b/drivers/net/usb/cdc_mbim.c
@@ -21,6 +21,8 @@
21#include <linux/usb/usbnet.h> 21#include <linux/usb/usbnet.h>
22#include <linux/usb/cdc-wdm.h> 22#include <linux/usb/cdc-wdm.h>
23#include <linux/usb/cdc_ncm.h> 23#include <linux/usb/cdc_ncm.h>
24#include <net/ipv6.h>
25#include <net/addrconf.h>
24 26
25/* driver specific data - must match cdc_ncm usage */ 27/* driver specific data - must match cdc_ncm usage */
26struct cdc_mbim_state { 28struct cdc_mbim_state {
@@ -42,13 +44,11 @@ static int cdc_mbim_manage_power(struct usbnet *dev, int on)
42 if ((on && atomic_add_return(1, &info->pmcount) == 1) || (!on && atomic_dec_and_test(&info->pmcount))) { 44 if ((on && atomic_add_return(1, &info->pmcount) == 1) || (!on && atomic_dec_and_test(&info->pmcount))) {
43 /* need autopm_get/put here to ensure the usbcore sees the new value */ 45 /* need autopm_get/put here to ensure the usbcore sees the new value */
44 rv = usb_autopm_get_interface(dev->intf); 46 rv = usb_autopm_get_interface(dev->intf);
45 if (rv < 0)
46 goto err;
47 dev->intf->needs_remote_wakeup = on; 47 dev->intf->needs_remote_wakeup = on;
48 usb_autopm_put_interface(dev->intf); 48 if (!rv)
49 usb_autopm_put_interface(dev->intf);
49 } 50 }
50err: 51 return 0;
51 return rv;
52} 52}
53 53
54static int cdc_mbim_wdm_manage_power(struct usb_interface *intf, int status) 54static int cdc_mbim_wdm_manage_power(struct usb_interface *intf, int status)
@@ -173,7 +173,7 @@ static struct sk_buff *cdc_mbim_tx_fixup(struct usbnet *dev, struct sk_buff *skb
173 } 173 }
174 174
175 spin_lock_bh(&ctx->mtx); 175 spin_lock_bh(&ctx->mtx);
176 skb_out = cdc_ncm_fill_tx_frame(ctx, skb, sign); 176 skb_out = cdc_ncm_fill_tx_frame(dev, skb, sign);
177 spin_unlock_bh(&ctx->mtx); 177 spin_unlock_bh(&ctx->mtx);
178 return skb_out; 178 return skb_out;
179 179
@@ -184,6 +184,60 @@ error:
184 return NULL; 184 return NULL;
185} 185}
186 186
187/* Some devices are known to send Neigbor Solicitation messages and
188 * require Neigbor Advertisement replies. The IPv6 core will not
189 * respond since IFF_NOARP is set, so we must handle them ourselves.
190 */
191static void do_neigh_solicit(struct usbnet *dev, u8 *buf, u16 tci)
192{
193 struct ipv6hdr *iph = (void *)buf;
194 struct nd_msg *msg = (void *)(iph + 1);
195 struct net_device *netdev;
196 struct inet6_dev *in6_dev;
197 bool is_router;
198
199 /* we'll only respond to requests from unicast addresses to
200 * our solicited node addresses.
201 */
202 if (!ipv6_addr_is_solict_mult(&iph->daddr) ||
203 !(ipv6_addr_type(&iph->saddr) & IPV6_ADDR_UNICAST))
204 return;
205
206 /* need to send the NA on the VLAN dev, if any */
207 if (tci)
208 netdev = __vlan_find_dev_deep(dev->net, htons(ETH_P_8021Q),
209 tci);
210 else
211 netdev = dev->net;
212 if (!netdev)
213 return;
214
215 in6_dev = in6_dev_get(netdev);
216 if (!in6_dev)
217 return;
218 is_router = !!in6_dev->cnf.forwarding;
219 in6_dev_put(in6_dev);
220
221 /* ipv6_stub != NULL if in6_dev_get returned an inet6_dev */
222 ipv6_stub->ndisc_send_na(netdev, NULL, &iph->saddr, &msg->target,
223 is_router /* router */,
224 true /* solicited */,
225 false /* override */,
226 true /* inc_opt */);
227}
228
229static bool is_neigh_solicit(u8 *buf, size_t len)
230{
231 struct ipv6hdr *iph = (void *)buf;
232 struct nd_msg *msg = (void *)(iph + 1);
233
234 return (len >= sizeof(struct ipv6hdr) + sizeof(struct nd_msg) &&
235 iph->nexthdr == IPPROTO_ICMPV6 &&
236 msg->icmph.icmp6_code == 0 &&
237 msg->icmph.icmp6_type == NDISC_NEIGHBOUR_SOLICITATION);
238}
239
240
187static struct sk_buff *cdc_mbim_process_dgram(struct usbnet *dev, u8 *buf, size_t len, u16 tci) 241static struct sk_buff *cdc_mbim_process_dgram(struct usbnet *dev, u8 *buf, size_t len, u16 tci)
188{ 242{
189 __be16 proto = htons(ETH_P_802_3); 243 __be16 proto = htons(ETH_P_802_3);
@@ -198,6 +252,8 @@ static struct sk_buff *cdc_mbim_process_dgram(struct usbnet *dev, u8 *buf, size_
198 proto = htons(ETH_P_IP); 252 proto = htons(ETH_P_IP);
199 break; 253 break;
200 case 0x60: 254 case 0x60:
255 if (is_neigh_solicit(buf, len))
256 do_neigh_solicit(dev, buf, tci);
201 proto = htons(ETH_P_IPV6); 257 proto = htons(ETH_P_IPV6);
202 break; 258 break;
203 default: 259 default:
@@ -313,15 +369,13 @@ error:
313 369
314static int cdc_mbim_suspend(struct usb_interface *intf, pm_message_t message) 370static int cdc_mbim_suspend(struct usb_interface *intf, pm_message_t message)
315{ 371{
316 int ret = 0; 372 int ret = -ENODEV;
317 struct usbnet *dev = usb_get_intfdata(intf); 373 struct usbnet *dev = usb_get_intfdata(intf);
318 struct cdc_mbim_state *info = (void *)&dev->data; 374 struct cdc_mbim_state *info = (void *)&dev->data;
319 struct cdc_ncm_ctx *ctx = info->ctx; 375 struct cdc_ncm_ctx *ctx = info->ctx;
320 376
321 if (ctx == NULL) { 377 if (!ctx)
322 ret = -1;
323 goto error; 378 goto error;
324 }
325 379
326 /* 380 /*
327 * Both usbnet_suspend() and subdriver->suspend() MUST return 0 381 * Both usbnet_suspend() and subdriver->suspend() MUST return 0
@@ -354,7 +408,7 @@ static int cdc_mbim_resume(struct usb_interface *intf)
354 if (ret < 0) 408 if (ret < 0)
355 goto err; 409 goto err;
356 ret = usbnet_resume(intf); 410 ret = usbnet_resume(intf);
357 if (ret < 0 && callsub && info->subdriver->suspend) 411 if (ret < 0 && callsub)
358 info->subdriver->suspend(intf, PMSG_SUSPEND); 412 info->subdriver->suspend(intf, PMSG_SUSPEND);
359err: 413err:
360 return ret; 414 return ret;
@@ -371,9 +425,18 @@ static const struct driver_info cdc_mbim_info = {
371}; 425};
372 426
373/* MBIM and NCM devices should not need a ZLP after NTBs with 427/* MBIM and NCM devices should not need a ZLP after NTBs with
374 * dwNtbOutMaxSize length. This driver_info is for the exceptional 428 * dwNtbOutMaxSize length. Nevertheless, a number of devices from
375 * devices requiring it anyway, allowing them to be supported without 429 * different vendor IDs will fail unless we send ZLPs, forcing us
376 * forcing the performance penalty on all the sane devices. 430 * to make this the default.
431 *
432 * This default may cause a performance penalty for spec conforming
433 * devices wanting to take advantage of optimizations possible without
434 * ZLPs. A whitelist is added in an attempt to avoid this for devices
435 * known to conform to the MBIM specification.
436 *
437 * All known devices supporting NCM compatibility mode are also
438 * conforming to the NCM and MBIM specifications. For this reason, the
439 * NCM subclass entry is also in the ZLP whitelist.
377 */ 440 */
378static const struct driver_info cdc_mbim_info_zlp = { 441static const struct driver_info cdc_mbim_info_zlp = {
379 .description = "CDC MBIM", 442 .description = "CDC MBIM",
@@ -396,16 +459,13 @@ static const struct usb_device_id mbim_devs[] = {
396 { USB_INTERFACE_INFO(USB_CLASS_COMM, USB_CDC_SUBCLASS_NCM, USB_CDC_PROTO_NONE), 459 { USB_INTERFACE_INFO(USB_CLASS_COMM, USB_CDC_SUBCLASS_NCM, USB_CDC_PROTO_NONE),
397 .driver_info = (unsigned long)&cdc_mbim_info, 460 .driver_info = (unsigned long)&cdc_mbim_info,
398 }, 461 },
399 /* Sierra Wireless MC7710 need ZLPs */ 462 /* ZLP conformance whitelist: All Ericsson MBIM devices */
400 { USB_DEVICE_AND_INTERFACE_INFO(0x1199, 0x68a2, USB_CLASS_COMM, USB_CDC_SUBCLASS_MBIM, USB_CDC_PROTO_NONE), 463 { USB_VENDOR_AND_INTERFACE_INFO(0x0bdb, USB_CLASS_COMM, USB_CDC_SUBCLASS_MBIM, USB_CDC_PROTO_NONE),
401 .driver_info = (unsigned long)&cdc_mbim_info_zlp, 464 .driver_info = (unsigned long)&cdc_mbim_info,
402 },
403 /* HP hs2434 Mobile Broadband Module needs ZLPs */
404 { USB_DEVICE_AND_INTERFACE_INFO(0x3f0, 0x4b1d, USB_CLASS_COMM, USB_CDC_SUBCLASS_MBIM, USB_CDC_PROTO_NONE),
405 .driver_info = (unsigned long)&cdc_mbim_info_zlp,
406 }, 465 },
466 /* default entry */
407 { USB_INTERFACE_INFO(USB_CLASS_COMM, USB_CDC_SUBCLASS_MBIM, USB_CDC_PROTO_NONE), 467 { USB_INTERFACE_INFO(USB_CLASS_COMM, USB_CDC_SUBCLASS_MBIM, USB_CDC_PROTO_NONE),
408 .driver_info = (unsigned long)&cdc_mbim_info, 468 .driver_info = (unsigned long)&cdc_mbim_info_zlp,
409 }, 469 },
410 { 470 {
411 }, 471 },
diff --git a/drivers/net/usb/cdc_ncm.c b/drivers/net/usb/cdc_ncm.c
index 43afde8f48d2..f74786aa37be 100644
--- a/drivers/net/usb/cdc_ncm.c
+++ b/drivers/net/usb/cdc_ncm.c
@@ -53,8 +53,6 @@
53#include <linux/usb/cdc.h> 53#include <linux/usb/cdc.h>
54#include <linux/usb/cdc_ncm.h> 54#include <linux/usb/cdc_ncm.h>
55 55
56#define DRIVER_VERSION "14-Mar-2012"
57
58#if IS_ENABLED(CONFIG_USB_NET_CDC_MBIM) 56#if IS_ENABLED(CONFIG_USB_NET_CDC_MBIM)
59static bool prefer_mbim = true; 57static bool prefer_mbim = true;
60#else 58#else
@@ -68,71 +66,67 @@ static void cdc_ncm_tx_timeout_start(struct cdc_ncm_ctx *ctx);
68static enum hrtimer_restart cdc_ncm_tx_timer_cb(struct hrtimer *hr_timer); 66static enum hrtimer_restart cdc_ncm_tx_timer_cb(struct hrtimer *hr_timer);
69static struct usb_driver cdc_ncm_driver; 67static struct usb_driver cdc_ncm_driver;
70 68
71static void 69static u8 cdc_ncm_setup(struct usbnet *dev)
72cdc_ncm_get_drvinfo(struct net_device *net, struct ethtool_drvinfo *info)
73{
74 struct usbnet *dev = netdev_priv(net);
75
76 strlcpy(info->driver, dev->driver_name, sizeof(info->driver));
77 strlcpy(info->version, DRIVER_VERSION, sizeof(info->version));
78 strlcpy(info->fw_version, dev->driver_info->description,
79 sizeof(info->fw_version));
80 usb_make_path(dev->udev, info->bus_info, sizeof(info->bus_info));
81}
82
83static u8 cdc_ncm_setup(struct cdc_ncm_ctx *ctx)
84{ 70{
71 struct cdc_ncm_ctx *ctx = (struct cdc_ncm_ctx *)dev->data[0];
72 struct usb_cdc_ncm_ntb_parameters ncm_parm;
85 u32 val; 73 u32 val;
86 u8 flags; 74 u8 flags;
87 u8 iface_no; 75 u8 iface_no;
88 int err; 76 int err;
89 int eth_hlen; 77 int eth_hlen;
90 u16 ntb_fmt_supported; 78 u16 ntb_fmt_supported;
91 u32 min_dgram_size; 79 __le16 max_datagram_size;
92 u32 min_hdr_size;
93 struct usbnet *dev = netdev_priv(ctx->netdev);
94 80
95 iface_no = ctx->control->cur_altsetting->desc.bInterfaceNumber; 81 iface_no = ctx->control->cur_altsetting->desc.bInterfaceNumber;
96 82
97 err = usbnet_read_cmd(dev, USB_CDC_GET_NTB_PARAMETERS, 83 err = usbnet_read_cmd(dev, USB_CDC_GET_NTB_PARAMETERS,
98 USB_TYPE_CLASS | USB_DIR_IN 84 USB_TYPE_CLASS | USB_DIR_IN
99 |USB_RECIP_INTERFACE, 85 |USB_RECIP_INTERFACE,
100 0, iface_no, &ctx->ncm_parm, 86 0, iface_no, &ncm_parm,
101 sizeof(ctx->ncm_parm)); 87 sizeof(ncm_parm));
102 if (err < 0) { 88 if (err < 0) {
103 pr_debug("failed GET_NTB_PARAMETERS\n"); 89 dev_err(&dev->intf->dev, "failed GET_NTB_PARAMETERS\n");
104 return 1; 90 return err; /* GET_NTB_PARAMETERS is required */
105 } 91 }
106 92
107 /* read correct set of parameters according to device mode */ 93 /* read correct set of parameters according to device mode */
108 ctx->rx_max = le32_to_cpu(ctx->ncm_parm.dwNtbInMaxSize); 94 ctx->rx_max = le32_to_cpu(ncm_parm.dwNtbInMaxSize);
109 ctx->tx_max = le32_to_cpu(ctx->ncm_parm.dwNtbOutMaxSize); 95 ctx->tx_max = le32_to_cpu(ncm_parm.dwNtbOutMaxSize);
110 ctx->tx_remainder = le16_to_cpu(ctx->ncm_parm.wNdpOutPayloadRemainder); 96 ctx->tx_remainder = le16_to_cpu(ncm_parm.wNdpOutPayloadRemainder);
111 ctx->tx_modulus = le16_to_cpu(ctx->ncm_parm.wNdpOutDivisor); 97 ctx->tx_modulus = le16_to_cpu(ncm_parm.wNdpOutDivisor);
112 ctx->tx_ndp_modulus = le16_to_cpu(ctx->ncm_parm.wNdpOutAlignment); 98 ctx->tx_ndp_modulus = le16_to_cpu(ncm_parm.wNdpOutAlignment);
113 /* devices prior to NCM Errata shall set this field to zero */ 99 /* devices prior to NCM Errata shall set this field to zero */
114 ctx->tx_max_datagrams = le16_to_cpu(ctx->ncm_parm.wNtbOutMaxDatagrams); 100 ctx->tx_max_datagrams = le16_to_cpu(ncm_parm.wNtbOutMaxDatagrams);
115 ntb_fmt_supported = le16_to_cpu(ctx->ncm_parm.bmNtbFormatsSupported); 101 ntb_fmt_supported = le16_to_cpu(ncm_parm.bmNtbFormatsSupported);
116 102
117 eth_hlen = ETH_HLEN; 103 /* there are some minor differences in NCM and MBIM defaults */
118 min_dgram_size = CDC_NCM_MIN_DATAGRAM_SIZE; 104 if (cdc_ncm_comm_intf_is_mbim(ctx->control->cur_altsetting)) {
119 min_hdr_size = CDC_NCM_MIN_HDR_SIZE; 105 if (!ctx->mbim_desc)
120 if (ctx->mbim_desc != NULL) { 106 return -EINVAL;
121 flags = ctx->mbim_desc->bmNetworkCapabilities;
122 eth_hlen = 0; 107 eth_hlen = 0;
123 min_dgram_size = CDC_MBIM_MIN_DATAGRAM_SIZE; 108 flags = ctx->mbim_desc->bmNetworkCapabilities;
124 min_hdr_size = 0; 109 ctx->max_datagram_size = le16_to_cpu(ctx->mbim_desc->wMaxSegmentSize);
125 } else if (ctx->func_desc != NULL) { 110 if (ctx->max_datagram_size < CDC_MBIM_MIN_DATAGRAM_SIZE)
126 flags = ctx->func_desc->bmNetworkCapabilities; 111 ctx->max_datagram_size = CDC_MBIM_MIN_DATAGRAM_SIZE;
127 } else { 112 } else {
128 flags = 0; 113 if (!ctx->func_desc)
114 return -EINVAL;
115 eth_hlen = ETH_HLEN;
116 flags = ctx->func_desc->bmNetworkCapabilities;
117 ctx->max_datagram_size = le16_to_cpu(ctx->ether_desc->wMaxSegmentSize);
118 if (ctx->max_datagram_size < CDC_NCM_MIN_DATAGRAM_SIZE)
119 ctx->max_datagram_size = CDC_NCM_MIN_DATAGRAM_SIZE;
129 } 120 }
130 121
131 pr_debug("dwNtbInMaxSize=%u dwNtbOutMaxSize=%u " 122 /* common absolute max for NCM and MBIM */
132 "wNdpOutPayloadRemainder=%u wNdpOutDivisor=%u " 123 if (ctx->max_datagram_size > CDC_NCM_MAX_DATAGRAM_SIZE)
133 "wNdpOutAlignment=%u wNtbOutMaxDatagrams=%u flags=0x%x\n", 124 ctx->max_datagram_size = CDC_NCM_MAX_DATAGRAM_SIZE;
134 ctx->rx_max, ctx->tx_max, ctx->tx_remainder, ctx->tx_modulus, 125
135 ctx->tx_ndp_modulus, ctx->tx_max_datagrams, flags); 126 dev_dbg(&dev->intf->dev,
127 "dwNtbInMaxSize=%u dwNtbOutMaxSize=%u wNdpOutPayloadRemainder=%u wNdpOutDivisor=%u wNdpOutAlignment=%u wNtbOutMaxDatagrams=%u flags=0x%x\n",
128 ctx->rx_max, ctx->tx_max, ctx->tx_remainder, ctx->tx_modulus,
129 ctx->tx_ndp_modulus, ctx->tx_max_datagrams, flags);
136 130
137 /* max count of tx datagrams */ 131 /* max count of tx datagrams */
138 if ((ctx->tx_max_datagrams == 0) || 132 if ((ctx->tx_max_datagrams == 0) ||
@@ -141,19 +135,19 @@ static u8 cdc_ncm_setup(struct cdc_ncm_ctx *ctx)
141 135
142 /* verify maximum size of received NTB in bytes */ 136 /* verify maximum size of received NTB in bytes */
143 if (ctx->rx_max < USB_CDC_NCM_NTB_MIN_IN_SIZE) { 137 if (ctx->rx_max < USB_CDC_NCM_NTB_MIN_IN_SIZE) {
144 pr_debug("Using min receive length=%d\n", 138 dev_dbg(&dev->intf->dev, "Using min receive length=%d\n",
145 USB_CDC_NCM_NTB_MIN_IN_SIZE); 139 USB_CDC_NCM_NTB_MIN_IN_SIZE);
146 ctx->rx_max = USB_CDC_NCM_NTB_MIN_IN_SIZE; 140 ctx->rx_max = USB_CDC_NCM_NTB_MIN_IN_SIZE;
147 } 141 }
148 142
149 if (ctx->rx_max > CDC_NCM_NTB_MAX_SIZE_RX) { 143 if (ctx->rx_max > CDC_NCM_NTB_MAX_SIZE_RX) {
150 pr_debug("Using default maximum receive length=%d\n", 144 dev_dbg(&dev->intf->dev, "Using default maximum receive length=%d\n",
151 CDC_NCM_NTB_MAX_SIZE_RX); 145 CDC_NCM_NTB_MAX_SIZE_RX);
152 ctx->rx_max = CDC_NCM_NTB_MAX_SIZE_RX; 146 ctx->rx_max = CDC_NCM_NTB_MAX_SIZE_RX;
153 } 147 }
154 148
155 /* inform device about NTB input size changes */ 149 /* inform device about NTB input size changes */
156 if (ctx->rx_max != le32_to_cpu(ctx->ncm_parm.dwNtbInMaxSize)) { 150 if (ctx->rx_max != le32_to_cpu(ncm_parm.dwNtbInMaxSize)) {
157 __le32 dwNtbInMaxSize = cpu_to_le32(ctx->rx_max); 151 __le32 dwNtbInMaxSize = cpu_to_le32(ctx->rx_max);
158 152
159 err = usbnet_write_cmd(dev, USB_CDC_SET_NTB_INPUT_SIZE, 153 err = usbnet_write_cmd(dev, USB_CDC_SET_NTB_INPUT_SIZE,
@@ -161,16 +155,22 @@ static u8 cdc_ncm_setup(struct cdc_ncm_ctx *ctx)
161 | USB_RECIP_INTERFACE, 155 | USB_RECIP_INTERFACE,
162 0, iface_no, &dwNtbInMaxSize, 4); 156 0, iface_no, &dwNtbInMaxSize, 4);
163 if (err < 0) 157 if (err < 0)
164 pr_debug("Setting NTB Input Size failed\n"); 158 dev_dbg(&dev->intf->dev, "Setting NTB Input Size failed\n");
165 } 159 }
166 160
167 /* verify maximum size of transmitted NTB in bytes */ 161 /* verify maximum size of transmitted NTB in bytes */
168 if ((ctx->tx_max < 162 if (ctx->tx_max > CDC_NCM_NTB_MAX_SIZE_TX) {
169 (min_hdr_size + min_dgram_size)) || 163 dev_dbg(&dev->intf->dev, "Using default maximum transmit length=%d\n",
170 (ctx->tx_max > CDC_NCM_NTB_MAX_SIZE_TX)) { 164 CDC_NCM_NTB_MAX_SIZE_TX);
171 pr_debug("Using default maximum transmit length=%d\n",
172 CDC_NCM_NTB_MAX_SIZE_TX);
173 ctx->tx_max = CDC_NCM_NTB_MAX_SIZE_TX; 165 ctx->tx_max = CDC_NCM_NTB_MAX_SIZE_TX;
166
167 /* Adding a pad byte here simplifies the handling in
168 * cdc_ncm_fill_tx_frame, by making tx_max always
169 * represent the real skb max size.
170 */
171 if (ctx->tx_max % usb_maxpacket(dev->udev, dev->out, 1) == 0)
172 ctx->tx_max++;
173
174 } 174 }
175 175
176 /* 176 /*
@@ -183,7 +183,7 @@ static u8 cdc_ncm_setup(struct cdc_ncm_ctx *ctx)
183 183
184 if ((val < USB_CDC_NCM_NDP_ALIGN_MIN_SIZE) || 184 if ((val < USB_CDC_NCM_NDP_ALIGN_MIN_SIZE) ||
185 (val != ((-val) & val)) || (val >= ctx->tx_max)) { 185 (val != ((-val) & val)) || (val >= ctx->tx_max)) {
186 pr_debug("Using default alignment: 4 bytes\n"); 186 dev_dbg(&dev->intf->dev, "Using default alignment: 4 bytes\n");
187 ctx->tx_ndp_modulus = USB_CDC_NCM_NDP_ALIGN_MIN_SIZE; 187 ctx->tx_ndp_modulus = USB_CDC_NCM_NDP_ALIGN_MIN_SIZE;
188 } 188 }
189 189
@@ -197,13 +197,13 @@ static u8 cdc_ncm_setup(struct cdc_ncm_ctx *ctx)
197 197
198 if ((val < USB_CDC_NCM_NDP_ALIGN_MIN_SIZE) || 198 if ((val < USB_CDC_NCM_NDP_ALIGN_MIN_SIZE) ||
199 (val != ((-val) & val)) || (val >= ctx->tx_max)) { 199 (val != ((-val) & val)) || (val >= ctx->tx_max)) {
200 pr_debug("Using default transmit modulus: 4 bytes\n"); 200 dev_dbg(&dev->intf->dev, "Using default transmit modulus: 4 bytes\n");
201 ctx->tx_modulus = USB_CDC_NCM_NDP_ALIGN_MIN_SIZE; 201 ctx->tx_modulus = USB_CDC_NCM_NDP_ALIGN_MIN_SIZE;
202 } 202 }
203 203
204 /* verify the payload remainder */ 204 /* verify the payload remainder */
205 if (ctx->tx_remainder >= ctx->tx_modulus) { 205 if (ctx->tx_remainder >= ctx->tx_modulus) {
206 pr_debug("Using default transmit remainder: 0 bytes\n"); 206 dev_dbg(&dev->intf->dev, "Using default transmit remainder: 0 bytes\n");
207 ctx->tx_remainder = 0; 207 ctx->tx_remainder = 0;
208 } 208 }
209 209
@@ -221,7 +221,7 @@ static u8 cdc_ncm_setup(struct cdc_ncm_ctx *ctx)
221 USB_CDC_NCM_CRC_NOT_APPENDED, 221 USB_CDC_NCM_CRC_NOT_APPENDED,
222 iface_no, NULL, 0); 222 iface_no, NULL, 0);
223 if (err < 0) 223 if (err < 0)
224 pr_debug("Setting CRC mode off failed\n"); 224 dev_dbg(&dev->intf->dev, "Setting CRC mode off failed\n");
225 } 225 }
226 226
227 /* set NTB format, if both formats are supported */ 227 /* set NTB format, if both formats are supported */
@@ -232,69 +232,43 @@ static u8 cdc_ncm_setup(struct cdc_ncm_ctx *ctx)
232 USB_CDC_NCM_NTB16_FORMAT, 232 USB_CDC_NCM_NTB16_FORMAT,
233 iface_no, NULL, 0); 233 iface_no, NULL, 0);
234 if (err < 0) 234 if (err < 0)
235 pr_debug("Setting NTB format to 16-bit failed\n"); 235 dev_dbg(&dev->intf->dev, "Setting NTB format to 16-bit failed\n");
236 } 236 }
237 237
238 ctx->max_datagram_size = min_dgram_size; 238 /* inform the device about the selected Max Datagram Size */
239 if (!(flags & USB_CDC_NCM_NCAP_MAX_DATAGRAM_SIZE))
240 goto out;
239 241
240 /* set Max Datagram Size (MTU) */ 242 /* read current mtu value from device */
241 if (flags & USB_CDC_NCM_NCAP_MAX_DATAGRAM_SIZE) { 243 err = usbnet_read_cmd(dev, USB_CDC_GET_MAX_DATAGRAM_SIZE,
242 __le16 max_datagram_size; 244 USB_TYPE_CLASS | USB_DIR_IN | USB_RECIP_INTERFACE,
243 u16 eth_max_sz; 245 0, iface_no, &max_datagram_size, 2);
244 if (ctx->ether_desc != NULL) 246 if (err < 0) {
245 eth_max_sz = le16_to_cpu(ctx->ether_desc->wMaxSegmentSize); 247 dev_dbg(&dev->intf->dev, "GET_MAX_DATAGRAM_SIZE failed\n");
246 else if (ctx->mbim_desc != NULL) 248 goto out;
247 eth_max_sz = le16_to_cpu(ctx->mbim_desc->wMaxSegmentSize);
248 else
249 goto max_dgram_err;
250
251 err = usbnet_read_cmd(dev, USB_CDC_GET_MAX_DATAGRAM_SIZE,
252 USB_TYPE_CLASS | USB_DIR_IN
253 | USB_RECIP_INTERFACE,
254 0, iface_no, &max_datagram_size, 2);
255 if (err < 0) {
256 pr_debug("GET_MAX_DATAGRAM_SIZE failed, use size=%u\n",
257 min_dgram_size);
258 } else {
259 ctx->max_datagram_size =
260 le16_to_cpu(max_datagram_size);
261 /* Check Eth descriptor value */
262 if (ctx->max_datagram_size > eth_max_sz)
263 ctx->max_datagram_size = eth_max_sz;
264
265 if (ctx->max_datagram_size > CDC_NCM_MAX_DATAGRAM_SIZE)
266 ctx->max_datagram_size = CDC_NCM_MAX_DATAGRAM_SIZE;
267
268 if (ctx->max_datagram_size < min_dgram_size)
269 ctx->max_datagram_size = min_dgram_size;
270
271 /* if value changed, update device */
272 if (ctx->max_datagram_size !=
273 le16_to_cpu(max_datagram_size)) {
274 err = usbnet_write_cmd(dev,
275 USB_CDC_SET_MAX_DATAGRAM_SIZE,
276 USB_TYPE_CLASS | USB_DIR_OUT
277 | USB_RECIP_INTERFACE,
278 0,
279 iface_no, &max_datagram_size,
280 2);
281 if (err < 0)
282 pr_debug("SET_MAX_DGRAM_SIZE failed\n");
283 }
284 }
285 } 249 }
286 250
287max_dgram_err: 251 if (le16_to_cpu(max_datagram_size) == ctx->max_datagram_size)
288 if (ctx->netdev->mtu != (ctx->max_datagram_size - eth_hlen)) 252 goto out;
289 ctx->netdev->mtu = ctx->max_datagram_size - eth_hlen; 253
254 max_datagram_size = cpu_to_le16(ctx->max_datagram_size);
255 err = usbnet_write_cmd(dev, USB_CDC_SET_MAX_DATAGRAM_SIZE,
256 USB_TYPE_CLASS | USB_DIR_OUT | USB_RECIP_INTERFACE,
257 0, iface_no, &max_datagram_size, 2);
258 if (err < 0)
259 dev_dbg(&dev->intf->dev, "SET_MAX_DATAGRAM_SIZE failed\n");
290 260
261out:
262 /* set MTU to max supported by the device if necessary */
263 if (dev->net->mtu > ctx->max_datagram_size - eth_hlen)
264 dev->net->mtu = ctx->max_datagram_size - eth_hlen;
291 return 0; 265 return 0;
292} 266}
293 267
294static void 268static void
295cdc_ncm_find_endpoints(struct cdc_ncm_ctx *ctx, struct usb_interface *intf) 269cdc_ncm_find_endpoints(struct usbnet *dev, struct usb_interface *intf)
296{ 270{
297 struct usb_host_endpoint *e; 271 struct usb_host_endpoint *e, *in = NULL, *out = NULL;
298 u8 ep; 272 u8 ep;
299 273
300 for (ep = 0; ep < intf->cur_altsetting->desc.bNumEndpoints; ep++) { 274 for (ep = 0; ep < intf->cur_altsetting->desc.bNumEndpoints; ep++) {
@@ -303,18 +277,18 @@ cdc_ncm_find_endpoints(struct cdc_ncm_ctx *ctx, struct usb_interface *intf)
303 switch (e->desc.bmAttributes & USB_ENDPOINT_XFERTYPE_MASK) { 277 switch (e->desc.bmAttributes & USB_ENDPOINT_XFERTYPE_MASK) {
304 case USB_ENDPOINT_XFER_INT: 278 case USB_ENDPOINT_XFER_INT:
305 if (usb_endpoint_dir_in(&e->desc)) { 279 if (usb_endpoint_dir_in(&e->desc)) {
306 if (ctx->status_ep == NULL) 280 if (!dev->status)
307 ctx->status_ep = e; 281 dev->status = e;
308 } 282 }
309 break; 283 break;
310 284
311 case USB_ENDPOINT_XFER_BULK: 285 case USB_ENDPOINT_XFER_BULK:
312 if (usb_endpoint_dir_in(&e->desc)) { 286 if (usb_endpoint_dir_in(&e->desc)) {
313 if (ctx->in_ep == NULL) 287 if (!in)
314 ctx->in_ep = e; 288 in = e;
315 } else { 289 } else {
316 if (ctx->out_ep == NULL) 290 if (!out)
317 ctx->out_ep = e; 291 out = e;
318 } 292 }
319 break; 293 break;
320 294
@@ -322,6 +296,14 @@ cdc_ncm_find_endpoints(struct cdc_ncm_ctx *ctx, struct usb_interface *intf)
322 break; 296 break;
323 } 297 }
324 } 298 }
299 if (in && !dev->in)
300 dev->in = usb_rcvbulkpipe(dev->udev,
301 in->desc.bEndpointAddress &
302 USB_ENDPOINT_NUMBER_MASK);
303 if (out && !dev->out)
304 dev->out = usb_sndbulkpipe(dev->udev,
305 out->desc.bEndpointAddress &
306 USB_ENDPOINT_NUMBER_MASK);
325} 307}
326 308
327static void cdc_ncm_free(struct cdc_ncm_ctx *ctx) 309static void cdc_ncm_free(struct cdc_ncm_ctx *ctx)
@@ -342,18 +324,9 @@ static void cdc_ncm_free(struct cdc_ncm_ctx *ctx)
342 kfree(ctx); 324 kfree(ctx);
343} 325}
344 326
345static const struct ethtool_ops cdc_ncm_ethtool_ops = {
346 .get_drvinfo = cdc_ncm_get_drvinfo,
347 .get_link = usbnet_get_link,
348 .get_msglevel = usbnet_get_msglevel,
349 .set_msglevel = usbnet_set_msglevel,
350 .get_settings = usbnet_get_settings,
351 .set_settings = usbnet_set_settings,
352 .nway_reset = usbnet_nway_reset,
353};
354
355int cdc_ncm_bind_common(struct usbnet *dev, struct usb_interface *intf, u8 data_altsetting) 327int cdc_ncm_bind_common(struct usbnet *dev, struct usb_interface *intf, u8 data_altsetting)
356{ 328{
329 const struct usb_cdc_union_desc *union_desc = NULL;
357 struct cdc_ncm_ctx *ctx; 330 struct cdc_ncm_ctx *ctx;
358 struct usb_driver *driver; 331 struct usb_driver *driver;
359 u8 *buf; 332 u8 *buf;
@@ -367,23 +340,22 @@ int cdc_ncm_bind_common(struct usbnet *dev, struct usb_interface *intf, u8 data_
367 340
368 hrtimer_init(&ctx->tx_timer, CLOCK_MONOTONIC, HRTIMER_MODE_REL); 341 hrtimer_init(&ctx->tx_timer, CLOCK_MONOTONIC, HRTIMER_MODE_REL);
369 ctx->tx_timer.function = &cdc_ncm_tx_timer_cb; 342 ctx->tx_timer.function = &cdc_ncm_tx_timer_cb;
370 ctx->bh.data = (unsigned long)ctx; 343 ctx->bh.data = (unsigned long)dev;
371 ctx->bh.func = cdc_ncm_txpath_bh; 344 ctx->bh.func = cdc_ncm_txpath_bh;
372 atomic_set(&ctx->stop, 0); 345 atomic_set(&ctx->stop, 0);
373 spin_lock_init(&ctx->mtx); 346 spin_lock_init(&ctx->mtx);
374 ctx->netdev = dev->net;
375 347
376 /* store ctx pointer in device data field */ 348 /* store ctx pointer in device data field */
377 dev->data[0] = (unsigned long)ctx; 349 dev->data[0] = (unsigned long)ctx;
378 350
351 /* only the control interface can be successfully probed */
352 ctx->control = intf;
353
379 /* get some pointers */ 354 /* get some pointers */
380 driver = driver_of(intf); 355 driver = driver_of(intf);
381 buf = intf->cur_altsetting->extra; 356 buf = intf->cur_altsetting->extra;
382 len = intf->cur_altsetting->extralen; 357 len = intf->cur_altsetting->extralen;
383 358
384 ctx->udev = dev->udev;
385 ctx->intf = intf;
386
387 /* parse through descriptors associated with control interface */ 359 /* parse through descriptors associated with control interface */
388 while ((len > 0) && (buf[0] > 2) && (buf[0] <= len)) { 360 while ((len > 0) && (buf[0] > 2) && (buf[0] <= len)) {
389 361
@@ -392,16 +364,18 @@ int cdc_ncm_bind_common(struct usbnet *dev, struct usb_interface *intf, u8 data_
392 364
393 switch (buf[2]) { 365 switch (buf[2]) {
394 case USB_CDC_UNION_TYPE: 366 case USB_CDC_UNION_TYPE:
395 if (buf[0] < sizeof(*(ctx->union_desc))) 367 if (buf[0] < sizeof(*union_desc))
396 break; 368 break;
397 369
398 ctx->union_desc = 370 union_desc = (const struct usb_cdc_union_desc *)buf;
399 (const struct usb_cdc_union_desc *)buf; 371 /* the master must be the interface we are probing */
400 372 if (intf->cur_altsetting->desc.bInterfaceNumber !=
401 ctx->control = usb_ifnum_to_if(dev->udev, 373 union_desc->bMasterInterface0) {
402 ctx->union_desc->bMasterInterface0); 374 dev_dbg(&intf->dev, "bogus CDC Union\n");
375 goto error;
376 }
403 ctx->data = usb_ifnum_to_if(dev->udev, 377 ctx->data = usb_ifnum_to_if(dev->udev,
404 ctx->union_desc->bSlaveInterface0); 378 union_desc->bSlaveInterface0);
405 break; 379 break;
406 380
407 case USB_CDC_ETHERNET_TYPE: 381 case USB_CDC_ETHERNET_TYPE:
@@ -410,13 +384,6 @@ int cdc_ncm_bind_common(struct usbnet *dev, struct usb_interface *intf, u8 data_
410 384
411 ctx->ether_desc = 385 ctx->ether_desc =
412 (const struct usb_cdc_ether_desc *)buf; 386 (const struct usb_cdc_ether_desc *)buf;
413 dev->hard_mtu =
414 le16_to_cpu(ctx->ether_desc->wMaxSegmentSize);
415
416 if (dev->hard_mtu < CDC_NCM_MIN_DATAGRAM_SIZE)
417 dev->hard_mtu = CDC_NCM_MIN_DATAGRAM_SIZE;
418 else if (dev->hard_mtu > CDC_NCM_MAX_DATAGRAM_SIZE)
419 dev->hard_mtu = CDC_NCM_MAX_DATAGRAM_SIZE;
420 break; 387 break;
421 388
422 case USB_CDC_NCM_TYPE: 389 case USB_CDC_NCM_TYPE:
@@ -444,69 +411,71 @@ advance:
444 } 411 }
445 412
446 /* some buggy devices have an IAD but no CDC Union */ 413 /* some buggy devices have an IAD but no CDC Union */
447 if (!ctx->union_desc && intf->intf_assoc && intf->intf_assoc->bInterfaceCount == 2) { 414 if (!union_desc && intf->intf_assoc && intf->intf_assoc->bInterfaceCount == 2) {
448 ctx->control = intf;
449 ctx->data = usb_ifnum_to_if(dev->udev, intf->cur_altsetting->desc.bInterfaceNumber + 1); 415 ctx->data = usb_ifnum_to_if(dev->udev, intf->cur_altsetting->desc.bInterfaceNumber + 1);
450 dev_dbg(&intf->dev, "CDC Union missing - got slave from IAD\n"); 416 dev_dbg(&intf->dev, "CDC Union missing - got slave from IAD\n");
451 } 417 }
452 418
453 /* check if we got everything */ 419 /* check if we got everything */
454 if ((ctx->control == NULL) || (ctx->data == NULL) || 420 if (!ctx->data || (!ctx->mbim_desc && !ctx->ether_desc)) {
455 ((!ctx->mbim_desc) && ((ctx->ether_desc == NULL) || (ctx->control != intf)))) 421 dev_dbg(&intf->dev, "CDC descriptors missing\n");
456 goto error; 422 goto error;
423 }
457 424
458 /* claim data interface, if different from control */ 425 /* claim data interface, if different from control */
459 if (ctx->data != ctx->control) { 426 if (ctx->data != ctx->control) {
460 temp = usb_driver_claim_interface(driver, ctx->data, dev); 427 temp = usb_driver_claim_interface(driver, ctx->data, dev);
461 if (temp) 428 if (temp) {
429 dev_dbg(&intf->dev, "failed to claim data intf\n");
462 goto error; 430 goto error;
431 }
463 } 432 }
464 433
465 iface_no = ctx->data->cur_altsetting->desc.bInterfaceNumber; 434 iface_no = ctx->data->cur_altsetting->desc.bInterfaceNumber;
466 435
467 /* reset data interface */ 436 /* reset data interface */
468 temp = usb_set_interface(dev->udev, iface_no, 0); 437 temp = usb_set_interface(dev->udev, iface_no, 0);
469 if (temp) 438 if (temp) {
470 goto error2; 439 dev_dbg(&intf->dev, "set interface failed\n");
471
472 /* initialize data interface */
473 if (cdc_ncm_setup(ctx))
474 goto error2; 440 goto error2;
441 }
475 442
476 /* configure data interface */ 443 /* configure data interface */
477 temp = usb_set_interface(dev->udev, iface_no, data_altsetting); 444 temp = usb_set_interface(dev->udev, iface_no, data_altsetting);
478 if (temp) 445 if (temp) {
446 dev_dbg(&intf->dev, "set interface failed\n");
479 goto error2; 447 goto error2;
448 }
480 449
481 cdc_ncm_find_endpoints(ctx, ctx->data); 450 cdc_ncm_find_endpoints(dev, ctx->data);
482 cdc_ncm_find_endpoints(ctx, ctx->control); 451 cdc_ncm_find_endpoints(dev, ctx->control);
483 452 if (!dev->in || !dev->out || !dev->status) {
484 if ((ctx->in_ep == NULL) || (ctx->out_ep == NULL) || 453 dev_dbg(&intf->dev, "failed to collect endpoints\n");
485 (ctx->status_ep == NULL))
486 goto error2; 454 goto error2;
455 }
487 456
488 dev->net->ethtool_ops = &cdc_ncm_ethtool_ops; 457 /* initialize data interface */
458 if (cdc_ncm_setup(dev)) {
459 dev_dbg(&intf->dev, "cdc_ncm_setup() failed\n");
460 goto error2;
461 }
489 462
490 usb_set_intfdata(ctx->data, dev); 463 usb_set_intfdata(ctx->data, dev);
491 usb_set_intfdata(ctx->control, dev); 464 usb_set_intfdata(ctx->control, dev);
492 usb_set_intfdata(ctx->intf, dev);
493 465
494 if (ctx->ether_desc) { 466 if (ctx->ether_desc) {
495 temp = usbnet_get_ethernet_addr(dev, ctx->ether_desc->iMACAddress); 467 temp = usbnet_get_ethernet_addr(dev, ctx->ether_desc->iMACAddress);
496 if (temp) 468 if (temp) {
469 dev_dbg(&intf->dev, "failed to get mac address\n");
497 goto error2; 470 goto error2;
498 dev_info(&dev->udev->dev, "MAC-Address: %pM\n", dev->net->dev_addr); 471 }
472 dev_info(&intf->dev, "MAC-Address: %pM\n", dev->net->dev_addr);
499 } 473 }
500 474
501 475 /* usbnet use these values for sizing tx/rx queues */
502 dev->in = usb_rcvbulkpipe(dev->udev, 476 dev->hard_mtu = ctx->tx_max;
503 ctx->in_ep->desc.bEndpointAddress & USB_ENDPOINT_NUMBER_MASK);
504 dev->out = usb_sndbulkpipe(dev->udev,
505 ctx->out_ep->desc.bEndpointAddress & USB_ENDPOINT_NUMBER_MASK);
506 dev->status = ctx->status_ep;
507 dev->rx_urb_size = ctx->rx_max; 477 dev->rx_urb_size = ctx->rx_max;
508 478
509 ctx->tx_speed = ctx->rx_speed = 0;
510 return 0; 479 return 0;
511 480
512error2: 481error2:
@@ -517,7 +486,7 @@ error2:
517error: 486error:
518 cdc_ncm_free((struct cdc_ncm_ctx *)dev->data[0]); 487 cdc_ncm_free((struct cdc_ncm_ctx *)dev->data[0]);
519 dev->data[0] = 0; 488 dev->data[0] = 0;
520 dev_info(&dev->udev->dev, "bind() failure\n"); 489 dev_info(&intf->dev, "bind() failure\n");
521 return -ENODEV; 490 return -ENODEV;
522} 491}
523EXPORT_SYMBOL_GPL(cdc_ncm_bind_common); 492EXPORT_SYMBOL_GPL(cdc_ncm_bind_common);
@@ -553,7 +522,7 @@ void cdc_ncm_unbind(struct usbnet *dev, struct usb_interface *intf)
553 ctx->control = NULL; 522 ctx->control = NULL;
554 } 523 }
555 524
556 usb_set_intfdata(ctx->intf, NULL); 525 usb_set_intfdata(intf, NULL);
557 cdc_ncm_free(ctx); 526 cdc_ncm_free(ctx);
558} 527}
559EXPORT_SYMBOL_GPL(cdc_ncm_unbind); 528EXPORT_SYMBOL_GPL(cdc_ncm_unbind);
@@ -662,8 +631,9 @@ static struct usb_cdc_ncm_ndp16 *cdc_ncm_ndp(struct cdc_ncm_ctx *ctx, struct sk_
662} 631}
663 632
664struct sk_buff * 633struct sk_buff *
665cdc_ncm_fill_tx_frame(struct cdc_ncm_ctx *ctx, struct sk_buff *skb, __le32 sign) 634cdc_ncm_fill_tx_frame(struct usbnet *dev, struct sk_buff *skb, __le32 sign)
666{ 635{
636 struct cdc_ncm_ctx *ctx = (struct cdc_ncm_ctx *)dev->data[0];
667 struct usb_cdc_ncm_nth16 *nth16; 637 struct usb_cdc_ncm_nth16 *nth16;
668 struct usb_cdc_ncm_ndp16 *ndp16; 638 struct usb_cdc_ncm_ndp16 *ndp16;
669 struct sk_buff *skb_out; 639 struct sk_buff *skb_out;
@@ -683,11 +653,11 @@ cdc_ncm_fill_tx_frame(struct cdc_ncm_ctx *ctx, struct sk_buff *skb, __le32 sign)
683 653
684 /* allocate a new OUT skb */ 654 /* allocate a new OUT skb */
685 if (!skb_out) { 655 if (!skb_out) {
686 skb_out = alloc_skb((ctx->tx_max + 1), GFP_ATOMIC); 656 skb_out = alloc_skb(ctx->tx_max, GFP_ATOMIC);
687 if (skb_out == NULL) { 657 if (skb_out == NULL) {
688 if (skb != NULL) { 658 if (skb != NULL) {
689 dev_kfree_skb_any(skb); 659 dev_kfree_skb_any(skb);
690 ctx->netdev->stats.tx_dropped++; 660 dev->net->stats.tx_dropped++;
691 } 661 }
692 goto exit_no_skb; 662 goto exit_no_skb;
693 } 663 }
@@ -725,12 +695,12 @@ cdc_ncm_fill_tx_frame(struct cdc_ncm_ctx *ctx, struct sk_buff *skb, __le32 sign)
725 /* won't fit, MTU problem? */ 695 /* won't fit, MTU problem? */
726 dev_kfree_skb_any(skb); 696 dev_kfree_skb_any(skb);
727 skb = NULL; 697 skb = NULL;
728 ctx->netdev->stats.tx_dropped++; 698 dev->net->stats.tx_dropped++;
729 } else { 699 } else {
730 /* no room for skb - store for later */ 700 /* no room for skb - store for later */
731 if (ctx->tx_rem_skb != NULL) { 701 if (ctx->tx_rem_skb != NULL) {
732 dev_kfree_skb_any(ctx->tx_rem_skb); 702 dev_kfree_skb_any(ctx->tx_rem_skb);
733 ctx->netdev->stats.tx_dropped++; 703 dev->net->stats.tx_dropped++;
734 } 704 }
735 ctx->tx_rem_skb = skb; 705 ctx->tx_rem_skb = skb;
736 ctx->tx_rem_sign = sign; 706 ctx->tx_rem_sign = sign;
@@ -763,7 +733,7 @@ cdc_ncm_fill_tx_frame(struct cdc_ncm_ctx *ctx, struct sk_buff *skb, __le32 sign)
763 if (skb != NULL) { 733 if (skb != NULL) {
764 dev_kfree_skb_any(skb); 734 dev_kfree_skb_any(skb);
765 skb = NULL; 735 skb = NULL;
766 ctx->netdev->stats.tx_dropped++; 736 dev->net->stats.tx_dropped++;
767 } 737 }
768 738
769 ctx->tx_curr_frame_num = n; 739 ctx->tx_curr_frame_num = n;
@@ -788,19 +758,20 @@ cdc_ncm_fill_tx_frame(struct cdc_ncm_ctx *ctx, struct sk_buff *skb, __le32 sign)
788 /* variables will be reset at next call */ 758 /* variables will be reset at next call */
789 } 759 }
790 760
791 /* 761 /* If collected data size is less or equal CDC_NCM_MIN_TX_PKT
792 * If collected data size is less or equal CDC_NCM_MIN_TX_PKT bytes, 762 * bytes, we send buffers as it is. If we get more data, it
793 * we send buffers as it is. If we get more data, it would be more 763 * would be more efficient for USB HS mobile device with DMA
794 * efficient for USB HS mobile device with DMA engine to receive a full 764 * engine to receive a full size NTB, than canceling DMA
795 * size NTB, than canceling DMA transfer and receiving a short packet. 765 * transfer and receiving a short packet.
766 *
767 * This optimization support is pointless if we end up sending
768 * a ZLP after full sized NTBs.
796 */ 769 */
797 if (skb_out->len > CDC_NCM_MIN_TX_PKT) 770 if (!(dev->driver_info->flags & FLAG_SEND_ZLP) &&
798 /* final zero padding */ 771 skb_out->len > CDC_NCM_MIN_TX_PKT)
799 memset(skb_put(skb_out, ctx->tx_max - skb_out->len), 0, ctx->tx_max - skb_out->len); 772 memset(skb_put(skb_out, ctx->tx_max - skb_out->len), 0,
800 773 ctx->tx_max - skb_out->len);
801 /* do we need to prevent a ZLP? */ 774 else if ((skb_out->len % dev->maxpacket) == 0)
802 if (((skb_out->len % le16_to_cpu(ctx->out_ep->desc.wMaxPacketSize)) == 0) &&
803 (skb_out->len < le32_to_cpu(ctx->ncm_parm.dwNtbOutMaxSize)) && skb_tailroom(skb_out))
804 *skb_put(skb_out, 1) = 0; /* force short packet */ 775 *skb_put(skb_out, 1) = 0; /* force short packet */
805 776
806 /* set final frame length */ 777 /* set final frame length */
@@ -809,7 +780,7 @@ cdc_ncm_fill_tx_frame(struct cdc_ncm_ctx *ctx, struct sk_buff *skb, __le32 sign)
809 780
810 /* return skb */ 781 /* return skb */
811 ctx->tx_curr_skb = NULL; 782 ctx->tx_curr_skb = NULL;
812 ctx->netdev->stats.tx_packets += ctx->tx_curr_frame_num; 783 dev->net->stats.tx_packets += ctx->tx_curr_frame_num;
813 return skb_out; 784 return skb_out;
814 785
815exit_no_skb: 786exit_no_skb:
@@ -841,24 +812,25 @@ static enum hrtimer_restart cdc_ncm_tx_timer_cb(struct hrtimer *timer)
841 812
842static void cdc_ncm_txpath_bh(unsigned long param) 813static void cdc_ncm_txpath_bh(unsigned long param)
843{ 814{
844 struct cdc_ncm_ctx *ctx = (struct cdc_ncm_ctx *)param; 815 struct usbnet *dev = (struct usbnet *)param;
816 struct cdc_ncm_ctx *ctx = (struct cdc_ncm_ctx *)dev->data[0];
845 817
846 spin_lock_bh(&ctx->mtx); 818 spin_lock_bh(&ctx->mtx);
847 if (ctx->tx_timer_pending != 0) { 819 if (ctx->tx_timer_pending != 0) {
848 ctx->tx_timer_pending--; 820 ctx->tx_timer_pending--;
849 cdc_ncm_tx_timeout_start(ctx); 821 cdc_ncm_tx_timeout_start(ctx);
850 spin_unlock_bh(&ctx->mtx); 822 spin_unlock_bh(&ctx->mtx);
851 } else if (ctx->netdev != NULL) { 823 } else if (dev->net != NULL) {
852 spin_unlock_bh(&ctx->mtx); 824 spin_unlock_bh(&ctx->mtx);
853 netif_tx_lock_bh(ctx->netdev); 825 netif_tx_lock_bh(dev->net);
854 usbnet_start_xmit(NULL, ctx->netdev); 826 usbnet_start_xmit(NULL, dev->net);
855 netif_tx_unlock_bh(ctx->netdev); 827 netif_tx_unlock_bh(dev->net);
856 } else { 828 } else {
857 spin_unlock_bh(&ctx->mtx); 829 spin_unlock_bh(&ctx->mtx);
858 } 830 }
859} 831}
860 832
861static struct sk_buff * 833struct sk_buff *
862cdc_ncm_tx_fixup(struct usbnet *dev, struct sk_buff *skb, gfp_t flags) 834cdc_ncm_tx_fixup(struct usbnet *dev, struct sk_buff *skb, gfp_t flags)
863{ 835{
864 struct sk_buff *skb_out; 836 struct sk_buff *skb_out;
@@ -875,7 +847,7 @@ cdc_ncm_tx_fixup(struct usbnet *dev, struct sk_buff *skb, gfp_t flags)
875 goto error; 847 goto error;
876 848
877 spin_lock_bh(&ctx->mtx); 849 spin_lock_bh(&ctx->mtx);
878 skb_out = cdc_ncm_fill_tx_frame(ctx, skb, cpu_to_le32(USB_CDC_NCM_NDP16_NOCRC_SIGN)); 850 skb_out = cdc_ncm_fill_tx_frame(dev, skb, cpu_to_le32(USB_CDC_NCM_NDP16_NOCRC_SIGN));
879 spin_unlock_bh(&ctx->mtx); 851 spin_unlock_bh(&ctx->mtx);
880 return skb_out; 852 return skb_out;
881 853
@@ -885,10 +857,12 @@ error:
885 857
886 return NULL; 858 return NULL;
887} 859}
860EXPORT_SYMBOL_GPL(cdc_ncm_tx_fixup);
888 861
889/* verify NTB header and return offset of first NDP, or negative error */ 862/* verify NTB header and return offset of first NDP, or negative error */
890int cdc_ncm_rx_verify_nth16(struct cdc_ncm_ctx *ctx, struct sk_buff *skb_in) 863int cdc_ncm_rx_verify_nth16(struct cdc_ncm_ctx *ctx, struct sk_buff *skb_in)
891{ 864{
865 struct usbnet *dev = netdev_priv(skb_in->dev);
892 struct usb_cdc_ncm_nth16 *nth16; 866 struct usb_cdc_ncm_nth16 *nth16;
893 int len; 867 int len;
894 int ret = -EINVAL; 868 int ret = -EINVAL;
@@ -898,30 +872,33 @@ int cdc_ncm_rx_verify_nth16(struct cdc_ncm_ctx *ctx, struct sk_buff *skb_in)
898 872
899 if (skb_in->len < (sizeof(struct usb_cdc_ncm_nth16) + 873 if (skb_in->len < (sizeof(struct usb_cdc_ncm_nth16) +
900 sizeof(struct usb_cdc_ncm_ndp16))) { 874 sizeof(struct usb_cdc_ncm_ndp16))) {
901 pr_debug("frame too short\n"); 875 netif_dbg(dev, rx_err, dev->net, "frame too short\n");
902 goto error; 876 goto error;
903 } 877 }
904 878
905 nth16 = (struct usb_cdc_ncm_nth16 *)skb_in->data; 879 nth16 = (struct usb_cdc_ncm_nth16 *)skb_in->data;
906 880
907 if (le32_to_cpu(nth16->dwSignature) != USB_CDC_NCM_NTH16_SIGN) { 881 if (nth16->dwSignature != cpu_to_le32(USB_CDC_NCM_NTH16_SIGN)) {
908 pr_debug("invalid NTH16 signature <%u>\n", 882 netif_dbg(dev, rx_err, dev->net,
909 le32_to_cpu(nth16->dwSignature)); 883 "invalid NTH16 signature <%#010x>\n",
884 le32_to_cpu(nth16->dwSignature));
910 goto error; 885 goto error;
911 } 886 }
912 887
913 len = le16_to_cpu(nth16->wBlockLength); 888 len = le16_to_cpu(nth16->wBlockLength);
914 if (len > ctx->rx_max) { 889 if (len > ctx->rx_max) {
915 pr_debug("unsupported NTB block length %u/%u\n", len, 890 netif_dbg(dev, rx_err, dev->net,
916 ctx->rx_max); 891 "unsupported NTB block length %u/%u\n", len,
892 ctx->rx_max);
917 goto error; 893 goto error;
918 } 894 }
919 895
920 if ((ctx->rx_seq + 1) != le16_to_cpu(nth16->wSequence) && 896 if ((ctx->rx_seq + 1) != le16_to_cpu(nth16->wSequence) &&
921 (ctx->rx_seq || le16_to_cpu(nth16->wSequence)) && 897 (ctx->rx_seq || le16_to_cpu(nth16->wSequence)) &&
922 !((ctx->rx_seq == 0xffff) && !le16_to_cpu(nth16->wSequence))) { 898 !((ctx->rx_seq == 0xffff) && !le16_to_cpu(nth16->wSequence))) {
923 pr_debug("sequence number glitch prev=%d curr=%d\n", 899 netif_dbg(dev, rx_err, dev->net,
924 ctx->rx_seq, le16_to_cpu(nth16->wSequence)); 900 "sequence number glitch prev=%d curr=%d\n",
901 ctx->rx_seq, le16_to_cpu(nth16->wSequence));
925 } 902 }
926 ctx->rx_seq = le16_to_cpu(nth16->wSequence); 903 ctx->rx_seq = le16_to_cpu(nth16->wSequence);
927 904
@@ -934,18 +911,20 @@ EXPORT_SYMBOL_GPL(cdc_ncm_rx_verify_nth16);
934/* verify NDP header and return number of datagrams, or negative error */ 911/* verify NDP header and return number of datagrams, or negative error */
935int cdc_ncm_rx_verify_ndp16(struct sk_buff *skb_in, int ndpoffset) 912int cdc_ncm_rx_verify_ndp16(struct sk_buff *skb_in, int ndpoffset)
936{ 913{
914 struct usbnet *dev = netdev_priv(skb_in->dev);
937 struct usb_cdc_ncm_ndp16 *ndp16; 915 struct usb_cdc_ncm_ndp16 *ndp16;
938 int ret = -EINVAL; 916 int ret = -EINVAL;
939 917
940 if ((ndpoffset + sizeof(struct usb_cdc_ncm_ndp16)) > skb_in->len) { 918 if ((ndpoffset + sizeof(struct usb_cdc_ncm_ndp16)) > skb_in->len) {
941 pr_debug("invalid NDP offset <%u>\n", ndpoffset); 919 netif_dbg(dev, rx_err, dev->net, "invalid NDP offset <%u>\n",
920 ndpoffset);
942 goto error; 921 goto error;
943 } 922 }
944 ndp16 = (struct usb_cdc_ncm_ndp16 *)(skb_in->data + ndpoffset); 923 ndp16 = (struct usb_cdc_ncm_ndp16 *)(skb_in->data + ndpoffset);
945 924
946 if (le16_to_cpu(ndp16->wLength) < USB_CDC_NCM_NDP16_LENGTH_MIN) { 925 if (le16_to_cpu(ndp16->wLength) < USB_CDC_NCM_NDP16_LENGTH_MIN) {
947 pr_debug("invalid DPT16 length <%u>\n", 926 netif_dbg(dev, rx_err, dev->net, "invalid DPT16 length <%u>\n",
948 le32_to_cpu(ndp16->dwSignature)); 927 le16_to_cpu(ndp16->wLength));
949 goto error; 928 goto error;
950 } 929 }
951 930
@@ -954,9 +933,9 @@ int cdc_ncm_rx_verify_ndp16(struct sk_buff *skb_in, int ndpoffset)
954 sizeof(struct usb_cdc_ncm_dpe16)); 933 sizeof(struct usb_cdc_ncm_dpe16));
955 ret--; /* we process NDP entries except for the last one */ 934 ret--; /* we process NDP entries except for the last one */
956 935
957 if ((sizeof(struct usb_cdc_ncm_ndp16) + ret * (sizeof(struct usb_cdc_ncm_dpe16))) > 936 if ((sizeof(struct usb_cdc_ncm_ndp16) +
958 skb_in->len) { 937 ret * (sizeof(struct usb_cdc_ncm_dpe16))) > skb_in->len) {
959 pr_debug("Invalid nframes = %d\n", ret); 938 netif_dbg(dev, rx_err, dev->net, "Invalid nframes = %d\n", ret);
960 ret = -EINVAL; 939 ret = -EINVAL;
961 } 940 }
962 941
@@ -965,7 +944,7 @@ error:
965} 944}
966EXPORT_SYMBOL_GPL(cdc_ncm_rx_verify_ndp16); 945EXPORT_SYMBOL_GPL(cdc_ncm_rx_verify_ndp16);
967 946
968static int cdc_ncm_rx_fixup(struct usbnet *dev, struct sk_buff *skb_in) 947int cdc_ncm_rx_fixup(struct usbnet *dev, struct sk_buff *skb_in)
969{ 948{
970 struct sk_buff *skb; 949 struct sk_buff *skb;
971 struct cdc_ncm_ctx *ctx = (struct cdc_ncm_ctx *)dev->data[0]; 950 struct cdc_ncm_ctx *ctx = (struct cdc_ncm_ctx *)dev->data[0];
@@ -989,9 +968,10 @@ next_ndp:
989 968
990 ndp16 = (struct usb_cdc_ncm_ndp16 *)(skb_in->data + ndpoffset); 969 ndp16 = (struct usb_cdc_ncm_ndp16 *)(skb_in->data + ndpoffset);
991 970
992 if (le32_to_cpu(ndp16->dwSignature) != USB_CDC_NCM_NDP16_NOCRC_SIGN) { 971 if (ndp16->dwSignature != cpu_to_le32(USB_CDC_NCM_NDP16_NOCRC_SIGN)) {
993 pr_debug("invalid DPT16 signature <%u>\n", 972 netif_dbg(dev, rx_err, dev->net,
994 le32_to_cpu(ndp16->dwSignature)); 973 "invalid DPT16 signature <%#010x>\n",
974 le32_to_cpu(ndp16->dwSignature));
995 goto err_ndp; 975 goto err_ndp;
996 } 976 }
997 dpe16 = ndp16->dpe16; 977 dpe16 = ndp16->dpe16;
@@ -1013,9 +993,9 @@ next_ndp:
1013 /* sanity checking */ 993 /* sanity checking */
1014 if (((offset + len) > skb_in->len) || 994 if (((offset + len) > skb_in->len) ||
1015 (len > ctx->rx_max) || (len < ETH_HLEN)) { 995 (len > ctx->rx_max) || (len < ETH_HLEN)) {
1016 pr_debug("invalid frame detected (ignored)" 996 netif_dbg(dev, rx_err, dev->net,
1017 "offset[%u]=%u, length=%u, skb=%p\n", 997 "invalid frame detected (ignored) offset[%u]=%u, length=%u, skb=%p\n",
1018 x, offset, len, skb_in); 998 x, offset, len, skb_in);
1019 if (!x) 999 if (!x)
1020 goto err_ndp; 1000 goto err_ndp;
1021 break; 1001 break;
@@ -1040,9 +1020,10 @@ err_ndp:
1040error: 1020error:
1041 return 0; 1021 return 0;
1042} 1022}
1023EXPORT_SYMBOL_GPL(cdc_ncm_rx_fixup);
1043 1024
1044static void 1025static void
1045cdc_ncm_speed_change(struct cdc_ncm_ctx *ctx, 1026cdc_ncm_speed_change(struct usbnet *dev,
1046 struct usb_cdc_speed_change *data) 1027 struct usb_cdc_speed_change *data)
1047{ 1028{
1048 uint32_t rx_speed = le32_to_cpu(data->DLBitRRate); 1029 uint32_t rx_speed = le32_to_cpu(data->DLBitRRate);
@@ -1052,25 +1033,16 @@ cdc_ncm_speed_change(struct cdc_ncm_ctx *ctx,
1052 * Currently the USB-NET API does not support reporting the actual 1033 * Currently the USB-NET API does not support reporting the actual
1053 * device speed. Do print it instead. 1034 * device speed. Do print it instead.
1054 */ 1035 */
1055 if ((tx_speed != ctx->tx_speed) || (rx_speed != ctx->rx_speed)) { 1036 if ((tx_speed > 1000000) && (rx_speed > 1000000)) {
1056 ctx->tx_speed = tx_speed; 1037 netif_info(dev, link, dev->net,
1057 ctx->rx_speed = rx_speed; 1038 "%u mbit/s downlink %u mbit/s uplink\n",
1058 1039 (unsigned int)(rx_speed / 1000000U),
1059 if ((tx_speed > 1000000) && (rx_speed > 1000000)) { 1040 (unsigned int)(tx_speed / 1000000U));
1060 printk(KERN_INFO KBUILD_MODNAME 1041 } else {
1061 ": %s: %u mbit/s downlink " 1042 netif_info(dev, link, dev->net,
1062 "%u mbit/s uplink\n", 1043 "%u kbit/s downlink %u kbit/s uplink\n",
1063 ctx->netdev->name, 1044 (unsigned int)(rx_speed / 1000U),
1064 (unsigned int)(rx_speed / 1000000U), 1045 (unsigned int)(tx_speed / 1000U));
1065 (unsigned int)(tx_speed / 1000000U));
1066 } else {
1067 printk(KERN_INFO KBUILD_MODNAME
1068 ": %s: %u kbit/s downlink "
1069 "%u kbit/s uplink\n",
1070 ctx->netdev->name,
1071 (unsigned int)(rx_speed / 1000U),
1072 (unsigned int)(tx_speed / 1000U));
1073 }
1074 } 1046 }
1075} 1047}
1076 1048
@@ -1086,7 +1058,7 @@ static void cdc_ncm_status(struct usbnet *dev, struct urb *urb)
1086 1058
1087 /* test for split data in 8-byte chunks */ 1059 /* test for split data in 8-byte chunks */
1088 if (test_and_clear_bit(EVENT_STS_SPLIT, &dev->flags)) { 1060 if (test_and_clear_bit(EVENT_STS_SPLIT, &dev->flags)) {
1089 cdc_ncm_speed_change(ctx, 1061 cdc_ncm_speed_change(dev,
1090 (struct usb_cdc_speed_change *)urb->transfer_buffer); 1062 (struct usb_cdc_speed_change *)urb->transfer_buffer);
1091 return; 1063 return;
1092 } 1064 }
@@ -1101,14 +1073,10 @@ static void cdc_ncm_status(struct usbnet *dev, struct urb *urb)
1101 * sent by device after USB_CDC_NOTIFY_SPEED_CHANGE. 1073 * sent by device after USB_CDC_NOTIFY_SPEED_CHANGE.
1102 */ 1074 */
1103 ctx->connected = le16_to_cpu(event->wValue); 1075 ctx->connected = le16_to_cpu(event->wValue);
1104 1076 netif_info(dev, link, dev->net,
1105 printk(KERN_INFO KBUILD_MODNAME ": %s: network connection:" 1077 "network connection: %sconnected\n",
1106 " %sconnected\n", 1078 ctx->connected ? "" : "dis");
1107 ctx->netdev->name, ctx->connected ? "" : "dis");
1108
1109 usbnet_link_change(dev, ctx->connected, 0); 1079 usbnet_link_change(dev, ctx->connected, 0);
1110 if (!ctx->connected)
1111 ctx->tx_speed = ctx->rx_speed = 0;
1112 break; 1080 break;
1113 1081
1114 case USB_CDC_NOTIFY_SPEED_CHANGE: 1082 case USB_CDC_NOTIFY_SPEED_CHANGE:
@@ -1116,8 +1084,8 @@ static void cdc_ncm_status(struct usbnet *dev, struct urb *urb)
1116 sizeof(struct usb_cdc_speed_change))) 1084 sizeof(struct usb_cdc_speed_change)))
1117 set_bit(EVENT_STS_SPLIT, &dev->flags); 1085 set_bit(EVENT_STS_SPLIT, &dev->flags);
1118 else 1086 else
1119 cdc_ncm_speed_change(ctx, 1087 cdc_ncm_speed_change(dev,
1120 (struct usb_cdc_speed_change *) &event[1]); 1088 (struct usb_cdc_speed_change *)&event[1]);
1121 break; 1089 break;
1122 1090
1123 default: 1091 default:
@@ -1139,22 +1107,6 @@ static int cdc_ncm_check_connect(struct usbnet *dev)
1139 return !ctx->connected; 1107 return !ctx->connected;
1140} 1108}
1141 1109
1142static int
1143cdc_ncm_probe(struct usb_interface *udev, const struct usb_device_id *prod)
1144{
1145 return usbnet_probe(udev, prod);
1146}
1147
1148static void cdc_ncm_disconnect(struct usb_interface *intf)
1149{
1150 struct usbnet *dev = usb_get_intfdata(intf);
1151
1152 if (dev == NULL)
1153 return; /* already disconnected */
1154
1155 usbnet_disconnect(intf);
1156}
1157
1158static const struct driver_info cdc_ncm_info = { 1110static const struct driver_info cdc_ncm_info = {
1159 .description = "CDC NCM", 1111 .description = "CDC NCM",
1160 .flags = FLAG_POINTTOPOINT | FLAG_NO_SETINT | FLAG_MULTI_PACKET, 1112 .flags = FLAG_POINTTOPOINT | FLAG_NO_SETINT | FLAG_MULTI_PACKET,
@@ -1234,17 +1186,6 @@ static const struct usb_device_id cdc_devs[] = {
1234 .driver_info = (unsigned long)&wwan_info, 1186 .driver_info = (unsigned long)&wwan_info,
1235 }, 1187 },
1236 1188
1237 /* Huawei NCM devices disguised as vendor specific */
1238 { USB_VENDOR_AND_INTERFACE_INFO(0x12d1, 0xff, 0x02, 0x16),
1239 .driver_info = (unsigned long)&wwan_info,
1240 },
1241 { USB_VENDOR_AND_INTERFACE_INFO(0x12d1, 0xff, 0x02, 0x46),
1242 .driver_info = (unsigned long)&wwan_info,
1243 },
1244 { USB_VENDOR_AND_INTERFACE_INFO(0x12d1, 0xff, 0x02, 0x76),
1245 .driver_info = (unsigned long)&wwan_info,
1246 },
1247
1248 /* Infineon(now Intel) HSPA Modem platform */ 1189 /* Infineon(now Intel) HSPA Modem platform */
1249 { USB_DEVICE_AND_INTERFACE_INFO(0x1519, 0x0443, 1190 { USB_DEVICE_AND_INTERFACE_INFO(0x1519, 0x0443,
1250 USB_CLASS_COMM, 1191 USB_CLASS_COMM,
@@ -1265,8 +1206,8 @@ MODULE_DEVICE_TABLE(usb, cdc_devs);
1265static struct usb_driver cdc_ncm_driver = { 1206static struct usb_driver cdc_ncm_driver = {
1266 .name = "cdc_ncm", 1207 .name = "cdc_ncm",
1267 .id_table = cdc_devs, 1208 .id_table = cdc_devs,
1268 .probe = cdc_ncm_probe, 1209 .probe = usbnet_probe,
1269 .disconnect = cdc_ncm_disconnect, 1210 .disconnect = usbnet_disconnect,
1270 .suspend = usbnet_suspend, 1211 .suspend = usbnet_suspend,
1271 .resume = usbnet_resume, 1212 .resume = usbnet_resume,
1272 .reset_resume = usbnet_resume, 1213 .reset_resume = usbnet_resume,
diff --git a/drivers/net/usb/huawei_cdc_ncm.c b/drivers/net/usb/huawei_cdc_ncm.c
new file mode 100644
index 000000000000..312178d7b698
--- /dev/null
+++ b/drivers/net/usb/huawei_cdc_ncm.c
@@ -0,0 +1,230 @@
1/* huawei_cdc_ncm.c - handles Huawei devices using the CDC NCM protocol as
2 * transport layer.
3 * Copyright (C) 2013 Enrico Mioso <mrkiko.rs@gmail.com>
4 *
5 *
6 * ABSTRACT:
7 * This driver handles devices resembling the CDC NCM standard, but
8 * encapsulating another protocol inside it. An example are some Huawei 3G
9 * devices, exposing an embedded AT channel where you can set up the NCM
10 * connection.
11 * This code has been heavily inspired by the cdc_mbim.c driver, which is
12 * Copyright (c) 2012 Smith Micro Software, Inc.
13 * Copyright (c) 2012 Bjørn Mork <bjorn@mork.no>
14 *
15 * This program is free software; you can redistribute it and/or
16 * modify it under the terms of the GNU General Public License
17 * version 2 as published by the Free Software Foundation.
18 */
19
20#include <linux/module.h>
21#include <linux/netdevice.h>
22#include <linux/ethtool.h>
23#include <linux/if_vlan.h>
24#include <linux/ip.h>
25#include <linux/mii.h>
26#include <linux/usb.h>
27#include <linux/usb/cdc.h>
28#include <linux/usb/usbnet.h>
29#include <linux/usb/cdc-wdm.h>
30#include <linux/usb/cdc_ncm.h>
31
32/* Driver data */
33struct huawei_cdc_ncm_state {
34 struct cdc_ncm_ctx *ctx;
35 atomic_t pmcount;
36 struct usb_driver *subdriver;
37 struct usb_interface *control;
38 struct usb_interface *data;
39};
40
41static int huawei_cdc_ncm_manage_power(struct usbnet *usbnet_dev, int on)
42{
43 struct huawei_cdc_ncm_state *drvstate = (void *)&usbnet_dev->data;
44 int rv;
45
46 if ((on && atomic_add_return(1, &drvstate->pmcount) == 1) ||
47 (!on && atomic_dec_and_test(&drvstate->pmcount))) {
48 rv = usb_autopm_get_interface(usbnet_dev->intf);
49 usbnet_dev->intf->needs_remote_wakeup = on;
50 if (!rv)
51 usb_autopm_put_interface(usbnet_dev->intf);
52 }
53 return 0;
54}
55
56static int huawei_cdc_ncm_wdm_manage_power(struct usb_interface *intf,
57 int status)
58{
59 struct usbnet *usbnet_dev = usb_get_intfdata(intf);
60
61 /* can be called while disconnecting */
62 if (!usbnet_dev)
63 return 0;
64
65 return huawei_cdc_ncm_manage_power(usbnet_dev, status);
66}
67
68
69static int huawei_cdc_ncm_bind(struct usbnet *usbnet_dev,
70 struct usb_interface *intf)
71{
72 struct cdc_ncm_ctx *ctx;
73 struct usb_driver *subdriver = ERR_PTR(-ENODEV);
74 int ret = -ENODEV;
75 struct huawei_cdc_ncm_state *drvstate = (void *)&usbnet_dev->data;
76
77 /* altsetting should always be 1 for NCM devices - so we hard-coded
78 * it here
79 */
80 ret = cdc_ncm_bind_common(usbnet_dev, intf, 1);
81 if (ret)
82 goto err;
83
84 ctx = drvstate->ctx;
85
86 if (usbnet_dev->status)
87 /* CDC-WMC r1.1 requires wMaxCommand to be "at least 256
88 * decimal (0x100)"
89 */
90 subdriver = usb_cdc_wdm_register(ctx->control,
91 &usbnet_dev->status->desc,
92 256, /* wMaxCommand */
93 huawei_cdc_ncm_wdm_manage_power);
94 if (IS_ERR(subdriver)) {
95 ret = PTR_ERR(subdriver);
96 cdc_ncm_unbind(usbnet_dev, intf);
97 goto err;
98 }
99
100 /* Prevent usbnet from using the status descriptor */
101 usbnet_dev->status = NULL;
102
103 drvstate->subdriver = subdriver;
104
105err:
106 return ret;
107}
108
109static void huawei_cdc_ncm_unbind(struct usbnet *usbnet_dev,
110 struct usb_interface *intf)
111{
112 struct huawei_cdc_ncm_state *drvstate = (void *)&usbnet_dev->data;
113 struct cdc_ncm_ctx *ctx = drvstate->ctx;
114
115 if (drvstate->subdriver && drvstate->subdriver->disconnect)
116 drvstate->subdriver->disconnect(ctx->control);
117 drvstate->subdriver = NULL;
118
119 cdc_ncm_unbind(usbnet_dev, intf);
120}
121
122static int huawei_cdc_ncm_suspend(struct usb_interface *intf,
123 pm_message_t message)
124{
125 int ret = 0;
126 struct usbnet *usbnet_dev = usb_get_intfdata(intf);
127 struct huawei_cdc_ncm_state *drvstate = (void *)&usbnet_dev->data;
128 struct cdc_ncm_ctx *ctx = drvstate->ctx;
129
130 if (ctx == NULL) {
131 ret = -ENODEV;
132 goto error;
133 }
134
135 ret = usbnet_suspend(intf, message);
136 if (ret < 0)
137 goto error;
138
139 if (intf == ctx->control &&
140 drvstate->subdriver &&
141 drvstate->subdriver->suspend)
142 ret = drvstate->subdriver->suspend(intf, message);
143 if (ret < 0)
144 usbnet_resume(intf);
145
146error:
147 return ret;
148}
149
150static int huawei_cdc_ncm_resume(struct usb_interface *intf)
151{
152 int ret = 0;
153 struct usbnet *usbnet_dev = usb_get_intfdata(intf);
154 struct huawei_cdc_ncm_state *drvstate = (void *)&usbnet_dev->data;
155 bool callsub;
156 struct cdc_ncm_ctx *ctx = drvstate->ctx;
157
158 /* should we call subdriver's resume function? */
159 callsub =
160 (intf == ctx->control &&
161 drvstate->subdriver &&
162 drvstate->subdriver->resume);
163
164 if (callsub)
165 ret = drvstate->subdriver->resume(intf);
166 if (ret < 0)
167 goto err;
168 ret = usbnet_resume(intf);
169 if (ret < 0 && callsub)
170 drvstate->subdriver->suspend(intf, PMSG_SUSPEND);
171err:
172 return ret;
173}
174
175static int huawei_cdc_ncm_check_connect(struct usbnet *usbnet_dev)
176{
177 struct cdc_ncm_ctx *ctx;
178
179 ctx = (struct cdc_ncm_ctx *)usbnet_dev->data[0];
180
181 if (ctx == NULL)
182 return 1; /* disconnected */
183
184 return !ctx->connected;
185}
186
187static const struct driver_info huawei_cdc_ncm_info = {
188 .description = "Huawei CDC NCM device",
189 .flags = FLAG_NO_SETINT | FLAG_MULTI_PACKET | FLAG_WWAN,
190 .bind = huawei_cdc_ncm_bind,
191 .unbind = huawei_cdc_ncm_unbind,
192 .check_connect = huawei_cdc_ncm_check_connect,
193 .manage_power = huawei_cdc_ncm_manage_power,
194 .rx_fixup = cdc_ncm_rx_fixup,
195 .tx_fixup = cdc_ncm_tx_fixup,
196};
197
198static const struct usb_device_id huawei_cdc_ncm_devs[] = {
199 /* Huawei NCM devices disguised as vendor specific */
200 { USB_VENDOR_AND_INTERFACE_INFO(0x12d1, 0xff, 0x02, 0x16),
201 .driver_info = (unsigned long)&huawei_cdc_ncm_info,
202 },
203 { USB_VENDOR_AND_INTERFACE_INFO(0x12d1, 0xff, 0x02, 0x46),
204 .driver_info = (unsigned long)&huawei_cdc_ncm_info,
205 },
206 { USB_VENDOR_AND_INTERFACE_INFO(0x12d1, 0xff, 0x02, 0x76),
207 .driver_info = (unsigned long)&huawei_cdc_ncm_info,
208 },
209
210 /* Terminating entry */
211 {
212 },
213};
214MODULE_DEVICE_TABLE(usb, huawei_cdc_ncm_devs);
215
216static struct usb_driver huawei_cdc_ncm_driver = {
217 .name = "huawei_cdc_ncm",
218 .id_table = huawei_cdc_ncm_devs,
219 .probe = usbnet_probe,
220 .disconnect = usbnet_disconnect,
221 .suspend = huawei_cdc_ncm_suspend,
222 .resume = huawei_cdc_ncm_resume,
223 .reset_resume = huawei_cdc_ncm_resume,
224 .supports_autosuspend = 1,
225 .disable_hub_initiated_lpm = 1,
226};
227module_usb_driver(huawei_cdc_ncm_driver);
228MODULE_AUTHOR("Enrico Mioso <mrkiko.rs@gmail.com>");
229MODULE_DESCRIPTION("USB CDC NCM host driver with encapsulated protocol support");
230MODULE_LICENSE("GPL");
diff --git a/drivers/net/usb/qmi_wwan.c b/drivers/net/usb/qmi_wwan.c
index 818ce90185b5..23bdd5b9274d 100644
--- a/drivers/net/usb/qmi_wwan.c
+++ b/drivers/net/usb/qmi_wwan.c
@@ -143,24 +143,28 @@ static const struct net_device_ops qmi_wwan_netdev_ops = {
143 .ndo_validate_addr = eth_validate_addr, 143 .ndo_validate_addr = eth_validate_addr,
144}; 144};
145 145
146/* using a counter to merge subdriver requests with our own into a combined state */ 146/* using a counter to merge subdriver requests with our own into a
147 * combined state
148 */
147static int qmi_wwan_manage_power(struct usbnet *dev, int on) 149static int qmi_wwan_manage_power(struct usbnet *dev, int on)
148{ 150{
149 struct qmi_wwan_state *info = (void *)&dev->data; 151 struct qmi_wwan_state *info = (void *)&dev->data;
150 int rv = 0; 152 int rv;
151 153
152 dev_dbg(&dev->intf->dev, "%s() pmcount=%d, on=%d\n", __func__, atomic_read(&info->pmcount), on); 154 dev_dbg(&dev->intf->dev, "%s() pmcount=%d, on=%d\n", __func__,
155 atomic_read(&info->pmcount), on);
153 156
154 if ((on && atomic_add_return(1, &info->pmcount) == 1) || (!on && atomic_dec_and_test(&info->pmcount))) { 157 if ((on && atomic_add_return(1, &info->pmcount) == 1) ||
155 /* need autopm_get/put here to ensure the usbcore sees the new value */ 158 (!on && atomic_dec_and_test(&info->pmcount))) {
159 /* need autopm_get/put here to ensure the usbcore sees
160 * the new value
161 */
156 rv = usb_autopm_get_interface(dev->intf); 162 rv = usb_autopm_get_interface(dev->intf);
157 if (rv < 0)
158 goto err;
159 dev->intf->needs_remote_wakeup = on; 163 dev->intf->needs_remote_wakeup = on;
160 usb_autopm_put_interface(dev->intf); 164 if (!rv)
165 usb_autopm_put_interface(dev->intf);
161 } 166 }
162err: 167 return 0;
163 return rv;
164} 168}
165 169
166static int qmi_wwan_cdc_wdm_manage_power(struct usb_interface *intf, int on) 170static int qmi_wwan_cdc_wdm_manage_power(struct usb_interface *intf, int on)
@@ -199,7 +203,8 @@ static int qmi_wwan_register_subdriver(struct usbnet *dev)
199 atomic_set(&info->pmcount, 0); 203 atomic_set(&info->pmcount, 0);
200 204
201 /* register subdriver */ 205 /* register subdriver */
202 subdriver = usb_cdc_wdm_register(info->control, &dev->status->desc, 4096, &qmi_wwan_cdc_wdm_manage_power); 206 subdriver = usb_cdc_wdm_register(info->control, &dev->status->desc,
207 4096, &qmi_wwan_cdc_wdm_manage_power);
203 if (IS_ERR(subdriver)) { 208 if (IS_ERR(subdriver)) {
204 dev_err(&info->control->dev, "subdriver registration failed\n"); 209 dev_err(&info->control->dev, "subdriver registration failed\n");
205 rv = PTR_ERR(subdriver); 210 rv = PTR_ERR(subdriver);
@@ -228,7 +233,8 @@ static int qmi_wwan_bind(struct usbnet *dev, struct usb_interface *intf)
228 struct usb_driver *driver = driver_of(intf); 233 struct usb_driver *driver = driver_of(intf);
229 struct qmi_wwan_state *info = (void *)&dev->data; 234 struct qmi_wwan_state *info = (void *)&dev->data;
230 235
231 BUILD_BUG_ON((sizeof(((struct usbnet *)0)->data) < sizeof(struct qmi_wwan_state))); 236 BUILD_BUG_ON((sizeof(((struct usbnet *)0)->data) <
237 sizeof(struct qmi_wwan_state)));
232 238
233 /* set up initial state */ 239 /* set up initial state */
234 info->control = intf; 240 info->control = intf;
@@ -250,7 +256,8 @@ static int qmi_wwan_bind(struct usbnet *dev, struct usb_interface *intf)
250 goto err; 256 goto err;
251 } 257 }
252 if (h->bLength != sizeof(struct usb_cdc_header_desc)) { 258 if (h->bLength != sizeof(struct usb_cdc_header_desc)) {
253 dev_dbg(&intf->dev, "CDC header len %u\n", h->bLength); 259 dev_dbg(&intf->dev, "CDC header len %u\n",
260 h->bLength);
254 goto err; 261 goto err;
255 } 262 }
256 break; 263 break;
@@ -260,7 +267,8 @@ static int qmi_wwan_bind(struct usbnet *dev, struct usb_interface *intf)
260 goto err; 267 goto err;
261 } 268 }
262 if (h->bLength != sizeof(struct usb_cdc_union_desc)) { 269 if (h->bLength != sizeof(struct usb_cdc_union_desc)) {
263 dev_dbg(&intf->dev, "CDC union len %u\n", h->bLength); 270 dev_dbg(&intf->dev, "CDC union len %u\n",
271 h->bLength);
264 goto err; 272 goto err;
265 } 273 }
266 cdc_union = (struct usb_cdc_union_desc *)buf; 274 cdc_union = (struct usb_cdc_union_desc *)buf;
@@ -271,15 +279,15 @@ static int qmi_wwan_bind(struct usbnet *dev, struct usb_interface *intf)
271 goto err; 279 goto err;
272 } 280 }
273 if (h->bLength != sizeof(struct usb_cdc_ether_desc)) { 281 if (h->bLength != sizeof(struct usb_cdc_ether_desc)) {
274 dev_dbg(&intf->dev, "CDC ether len %u\n", h->bLength); 282 dev_dbg(&intf->dev, "CDC ether len %u\n",
283 h->bLength);
275 goto err; 284 goto err;
276 } 285 }
277 cdc_ether = (struct usb_cdc_ether_desc *)buf; 286 cdc_ether = (struct usb_cdc_ether_desc *)buf;
278 break; 287 break;
279 } 288 }
280 289
281 /* 290 /* Remember which CDC functional descriptors we've seen. Works
282 * Remember which CDC functional descriptors we've seen. Works
283 * for all types we care about, of which USB_CDC_ETHERNET_TYPE 291 * for all types we care about, of which USB_CDC_ETHERNET_TYPE
284 * (0x0f) is the highest numbered 292 * (0x0f) is the highest numbered
285 */ 293 */
@@ -293,10 +301,14 @@ next_desc:
293 301
294 /* Use separate control and data interfaces if we found a CDC Union */ 302 /* Use separate control and data interfaces if we found a CDC Union */
295 if (cdc_union) { 303 if (cdc_union) {
296 info->data = usb_ifnum_to_if(dev->udev, cdc_union->bSlaveInterface0); 304 info->data = usb_ifnum_to_if(dev->udev,
297 if (desc->bInterfaceNumber != cdc_union->bMasterInterface0 || !info->data) { 305 cdc_union->bSlaveInterface0);
298 dev_err(&intf->dev, "bogus CDC Union: master=%u, slave=%u\n", 306 if (desc->bInterfaceNumber != cdc_union->bMasterInterface0 ||
299 cdc_union->bMasterInterface0, cdc_union->bSlaveInterface0); 307 !info->data) {
308 dev_err(&intf->dev,
309 "bogus CDC Union: master=%u, slave=%u\n",
310 cdc_union->bMasterInterface0,
311 cdc_union->bSlaveInterface0);
300 goto err; 312 goto err;
301 } 313 }
302 } 314 }
@@ -374,8 +386,7 @@ static int qmi_wwan_suspend(struct usb_interface *intf, pm_message_t message)
374 struct qmi_wwan_state *info = (void *)&dev->data; 386 struct qmi_wwan_state *info = (void *)&dev->data;
375 int ret; 387 int ret;
376 388
377 /* 389 /* Both usbnet_suspend() and subdriver->suspend() MUST return 0
378 * Both usbnet_suspend() and subdriver->suspend() MUST return 0
379 * in system sleep context, otherwise, the resume callback has 390 * in system sleep context, otherwise, the resume callback has
380 * to recover device from previous suspend failure. 391 * to recover device from previous suspend failure.
381 */ 392 */
@@ -383,7 +394,8 @@ static int qmi_wwan_suspend(struct usb_interface *intf, pm_message_t message)
383 if (ret < 0) 394 if (ret < 0)
384 goto err; 395 goto err;
385 396
386 if (intf == info->control && info->subdriver && info->subdriver->suspend) 397 if (intf == info->control && info->subdriver &&
398 info->subdriver->suspend)
387 ret = info->subdriver->suspend(intf, message); 399 ret = info->subdriver->suspend(intf, message);
388 if (ret < 0) 400 if (ret < 0)
389 usbnet_resume(intf); 401 usbnet_resume(intf);
@@ -396,14 +408,15 @@ static int qmi_wwan_resume(struct usb_interface *intf)
396 struct usbnet *dev = usb_get_intfdata(intf); 408 struct usbnet *dev = usb_get_intfdata(intf);
397 struct qmi_wwan_state *info = (void *)&dev->data; 409 struct qmi_wwan_state *info = (void *)&dev->data;
398 int ret = 0; 410 int ret = 0;
399 bool callsub = (intf == info->control && info->subdriver && info->subdriver->resume); 411 bool callsub = (intf == info->control && info->subdriver &&
412 info->subdriver->resume);
400 413
401 if (callsub) 414 if (callsub)
402 ret = info->subdriver->resume(intf); 415 ret = info->subdriver->resume(intf);
403 if (ret < 0) 416 if (ret < 0)
404 goto err; 417 goto err;
405 ret = usbnet_resume(intf); 418 ret = usbnet_resume(intf);
406 if (ret < 0 && callsub && info->subdriver->suspend) 419 if (ret < 0 && callsub)
407 info->subdriver->suspend(intf, PMSG_SUSPEND); 420 info->subdriver->suspend(intf, PMSG_SUSPEND);
408err: 421err:
409 return ret; 422 return ret;
@@ -714,6 +727,7 @@ static const struct usb_device_id products[] = {
714 {QMI_FIXED_INTF(0x2357, 0x0201, 4)}, /* TP-LINK HSUPA Modem MA180 */ 727 {QMI_FIXED_INTF(0x2357, 0x0201, 4)}, /* TP-LINK HSUPA Modem MA180 */
715 {QMI_FIXED_INTF(0x2357, 0x9000, 4)}, /* TP-LINK MA260 */ 728 {QMI_FIXED_INTF(0x2357, 0x9000, 4)}, /* TP-LINK MA260 */
716 {QMI_FIXED_INTF(0x1bc7, 0x1200, 5)}, /* Telit LE920 */ 729 {QMI_FIXED_INTF(0x1bc7, 0x1200, 5)}, /* Telit LE920 */
730 {QMI_FIXED_INTF(0x1bc7, 0x1201, 2)}, /* Telit LE920 */
717 {QMI_FIXED_INTF(0x0b3c, 0xc005, 6)}, /* Olivetti Olicard 200 */ 731 {QMI_FIXED_INTF(0x0b3c, 0xc005, 6)}, /* Olivetti Olicard 200 */
718 {QMI_FIXED_INTF(0x1e2d, 0x0060, 4)}, /* Cinterion PLxx */ 732 {QMI_FIXED_INTF(0x1e2d, 0x0060, 4)}, /* Cinterion PLxx */
719 733
@@ -777,7 +791,8 @@ static const struct usb_device_id products[] = {
777}; 791};
778MODULE_DEVICE_TABLE(usb, products); 792MODULE_DEVICE_TABLE(usb, products);
779 793
780static int qmi_wwan_probe(struct usb_interface *intf, const struct usb_device_id *prod) 794static int qmi_wwan_probe(struct usb_interface *intf,
795 const struct usb_device_id *prod)
781{ 796{
782 struct usb_device_id *id = (struct usb_device_id *)prod; 797 struct usb_device_id *id = (struct usb_device_id *)prod;
783 798
diff --git a/drivers/net/veth.c b/drivers/net/veth.c
index eee1f19ef1e9..b24db7acbf12 100644
--- a/drivers/net/veth.c
+++ b/drivers/net/veth.c
@@ -188,6 +188,11 @@ static struct rtnl_link_stats64 *veth_get_stats64(struct net_device *dev,
188 return tot; 188 return tot;
189} 189}
190 190
191/* fake multicast ability */
192static void veth_set_multicast_list(struct net_device *dev)
193{
194}
195
191static int veth_open(struct net_device *dev) 196static int veth_open(struct net_device *dev)
192{ 197{
193 struct veth_priv *priv = netdev_priv(dev); 198 struct veth_priv *priv = netdev_priv(dev);
@@ -250,11 +255,14 @@ static const struct net_device_ops veth_netdev_ops = {
250 .ndo_start_xmit = veth_xmit, 255 .ndo_start_xmit = veth_xmit,
251 .ndo_change_mtu = veth_change_mtu, 256 .ndo_change_mtu = veth_change_mtu,
252 .ndo_get_stats64 = veth_get_stats64, 257 .ndo_get_stats64 = veth_get_stats64,
258 .ndo_set_rx_mode = veth_set_multicast_list,
253 .ndo_set_mac_address = eth_mac_addr, 259 .ndo_set_mac_address = eth_mac_addr,
254}; 260};
255 261
256#define VETH_FEATURES (NETIF_F_SG | NETIF_F_FRAGLIST | NETIF_F_ALL_TSO | \ 262#define VETH_FEATURES (NETIF_F_SG | NETIF_F_FRAGLIST | NETIF_F_ALL_TSO | \
257 NETIF_F_HW_CSUM | NETIF_F_RXCSUM | NETIF_F_HIGHDMA | \ 263 NETIF_F_HW_CSUM | NETIF_F_RXCSUM | NETIF_F_HIGHDMA | \
264 NETIF_F_GSO_GRE | NETIF_F_GSO_UDP_TUNNEL | \
265 NETIF_F_GSO_IPIP | NETIF_F_GSO_SIT | NETIF_F_UFO | \
258 NETIF_F_HW_VLAN_CTAG_TX | NETIF_F_HW_VLAN_CTAG_RX | \ 266 NETIF_F_HW_VLAN_CTAG_TX | NETIF_F_HW_VLAN_CTAG_RX | \
259 NETIF_F_HW_VLAN_STAG_TX | NETIF_F_HW_VLAN_STAG_RX ) 267 NETIF_F_HW_VLAN_STAG_TX | NETIF_F_HW_VLAN_STAG_RX )
260 268
@@ -273,6 +281,7 @@ static void veth_setup(struct net_device *dev)
273 dev->destructor = veth_dev_free; 281 dev->destructor = veth_dev_free;
274 282
275 dev->hw_features = VETH_FEATURES; 283 dev->hw_features = VETH_FEATURES;
284 dev->hw_enc_features = VETH_FEATURES;
276} 285}
277 286
278/* 287/*
diff --git a/drivers/net/virtio_net.c b/drivers/net/virtio_net.c
index bbc9cb84ec1f..01f4eb5c8b78 100644
--- a/drivers/net/virtio_net.c
+++ b/drivers/net/virtio_net.c
@@ -124,12 +124,14 @@ struct virtnet_info {
124 /* Lock for config space updates */ 124 /* Lock for config space updates */
125 struct mutex config_lock; 125 struct mutex config_lock;
126 126
127 /* Page_frag for GFP_KERNEL packet buffer allocation when we run
128 * low on memory.
129 */
130 struct page_frag alloc_frag;
131
127 /* Does the affinity hint is set for virtqueues? */ 132 /* Does the affinity hint is set for virtqueues? */
128 bool affinity_hint_set; 133 bool affinity_hint_set;
129 134
130 /* Per-cpu variable to show the mapping from CPU to virtqueue */
131 int __percpu *vq_index;
132
133 /* CPU hot plug notifier */ 135 /* CPU hot plug notifier */
134 struct notifier_block nb; 136 struct notifier_block nb;
135}; 137};
@@ -217,33 +219,18 @@ static void skb_xmit_done(struct virtqueue *vq)
217 netif_wake_subqueue(vi->dev, vq2txq(vq)); 219 netif_wake_subqueue(vi->dev, vq2txq(vq));
218} 220}
219 221
220static void set_skb_frag(struct sk_buff *skb, struct page *page,
221 unsigned int offset, unsigned int *len)
222{
223 int size = min((unsigned)PAGE_SIZE - offset, *len);
224 int i = skb_shinfo(skb)->nr_frags;
225
226 __skb_fill_page_desc(skb, i, page, offset, size);
227
228 skb->data_len += size;
229 skb->len += size;
230 skb->truesize += PAGE_SIZE;
231 skb_shinfo(skb)->nr_frags++;
232 skb_shinfo(skb)->tx_flags |= SKBTX_SHARED_FRAG;
233 *len -= size;
234}
235
236/* Called from bottom half context */ 222/* Called from bottom half context */
237static struct sk_buff *page_to_skb(struct receive_queue *rq, 223static struct sk_buff *page_to_skb(struct receive_queue *rq,
238 struct page *page, unsigned int len) 224 struct page *page, unsigned int offset,
225 unsigned int len, unsigned int truesize)
239{ 226{
240 struct virtnet_info *vi = rq->vq->vdev->priv; 227 struct virtnet_info *vi = rq->vq->vdev->priv;
241 struct sk_buff *skb; 228 struct sk_buff *skb;
242 struct skb_vnet_hdr *hdr; 229 struct skb_vnet_hdr *hdr;
243 unsigned int copy, hdr_len, offset; 230 unsigned int copy, hdr_len, hdr_padded_len;
244 char *p; 231 char *p;
245 232
246 p = page_address(page); 233 p = page_address(page) + offset;
247 234
248 /* copy small packet so we can reuse these pages for small data */ 235 /* copy small packet so we can reuse these pages for small data */
249 skb = netdev_alloc_skb_ip_align(vi->dev, GOOD_COPY_LEN); 236 skb = netdev_alloc_skb_ip_align(vi->dev, GOOD_COPY_LEN);
@@ -254,16 +241,17 @@ static struct sk_buff *page_to_skb(struct receive_queue *rq,
254 241
255 if (vi->mergeable_rx_bufs) { 242 if (vi->mergeable_rx_bufs) {
256 hdr_len = sizeof hdr->mhdr; 243 hdr_len = sizeof hdr->mhdr;
257 offset = hdr_len; 244 hdr_padded_len = sizeof hdr->mhdr;
258 } else { 245 } else {
259 hdr_len = sizeof hdr->hdr; 246 hdr_len = sizeof hdr->hdr;
260 offset = sizeof(struct padded_vnet_hdr); 247 hdr_padded_len = sizeof(struct padded_vnet_hdr);
261 } 248 }
262 249
263 memcpy(hdr, p, hdr_len); 250 memcpy(hdr, p, hdr_len);
264 251
265 len -= hdr_len; 252 len -= hdr_len;
266 p += offset; 253 offset += hdr_padded_len;
254 p += hdr_padded_len;
267 255
268 copy = len; 256 copy = len;
269 if (copy > skb_tailroom(skb)) 257 if (copy > skb_tailroom(skb))
@@ -273,6 +261,14 @@ static struct sk_buff *page_to_skb(struct receive_queue *rq,
273 len -= copy; 261 len -= copy;
274 offset += copy; 262 offset += copy;
275 263
264 if (vi->mergeable_rx_bufs) {
265 if (len)
266 skb_add_rx_frag(skb, 0, page, offset, len, truesize);
267 else
268 put_page(page);
269 return skb;
270 }
271
276 /* 272 /*
277 * Verify that we can indeed put this data into a skb. 273 * Verify that we can indeed put this data into a skb.
278 * This is here to handle cases when the device erroneously 274 * This is here to handle cases when the device erroneously
@@ -284,9 +280,12 @@ static struct sk_buff *page_to_skb(struct receive_queue *rq,
284 dev_kfree_skb(skb); 280 dev_kfree_skb(skb);
285 return NULL; 281 return NULL;
286 } 282 }
287 283 BUG_ON(offset >= PAGE_SIZE);
288 while (len) { 284 while (len) {
289 set_skb_frag(skb, page, offset, &len); 285 unsigned int frag_size = min((unsigned)PAGE_SIZE - offset, len);
286 skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, page, offset,
287 frag_size, truesize);
288 len -= frag_size;
290 page = (struct page *)page->private; 289 page = (struct page *)page->private;
291 offset = 0; 290 offset = 0;
292 } 291 }
@@ -297,33 +296,59 @@ static struct sk_buff *page_to_skb(struct receive_queue *rq,
297 return skb; 296 return skb;
298} 297}
299 298
300static int receive_mergeable(struct receive_queue *rq, struct sk_buff *skb) 299static int receive_mergeable(struct receive_queue *rq, struct sk_buff *head_skb)
301{ 300{
302 struct skb_vnet_hdr *hdr = skb_vnet_hdr(skb); 301 struct skb_vnet_hdr *hdr = skb_vnet_hdr(head_skb);
302 struct sk_buff *curr_skb = head_skb;
303 char *buf;
303 struct page *page; 304 struct page *page;
304 int num_buf, i, len; 305 int num_buf, len, offset;
305 306
306 num_buf = hdr->mhdr.num_buffers; 307 num_buf = hdr->mhdr.num_buffers;
307 while (--num_buf) { 308 while (--num_buf) {
308 i = skb_shinfo(skb)->nr_frags; 309 int num_skb_frags = skb_shinfo(curr_skb)->nr_frags;
309 if (i >= MAX_SKB_FRAGS) { 310 buf = virtqueue_get_buf(rq->vq, &len);
310 pr_debug("%s: packet too long\n", skb->dev->name); 311 if (unlikely(!buf)) {
311 skb->dev->stats.rx_length_errors++;
312 return -EINVAL;
313 }
314 page = virtqueue_get_buf(rq->vq, &len);
315 if (!page) {
316 pr_debug("%s: rx error: %d buffers missing\n", 312 pr_debug("%s: rx error: %d buffers missing\n",
317 skb->dev->name, hdr->mhdr.num_buffers); 313 head_skb->dev->name, hdr->mhdr.num_buffers);
318 skb->dev->stats.rx_length_errors++; 314 head_skb->dev->stats.rx_length_errors++;
319 return -EINVAL; 315 return -EINVAL;
320 } 316 }
321 317 if (unlikely(len > MAX_PACKET_LEN)) {
322 if (len > PAGE_SIZE) 318 pr_debug("%s: rx error: merge buffer too long\n",
323 len = PAGE_SIZE; 319 head_skb->dev->name);
324 320 len = MAX_PACKET_LEN;
325 set_skb_frag(skb, page, 0, &len); 321 }
326 322 if (unlikely(num_skb_frags == MAX_SKB_FRAGS)) {
323 struct sk_buff *nskb = alloc_skb(0, GFP_ATOMIC);
324 if (unlikely(!nskb)) {
325 head_skb->dev->stats.rx_dropped++;
326 return -ENOMEM;
327 }
328 if (curr_skb == head_skb)
329 skb_shinfo(curr_skb)->frag_list = nskb;
330 else
331 curr_skb->next = nskb;
332 curr_skb = nskb;
333 head_skb->truesize += nskb->truesize;
334 num_skb_frags = 0;
335 }
336 if (curr_skb != head_skb) {
337 head_skb->data_len += len;
338 head_skb->len += len;
339 head_skb->truesize += MAX_PACKET_LEN;
340 }
341 page = virt_to_head_page(buf);
342 offset = buf - (char *)page_address(page);
343 if (skb_can_coalesce(curr_skb, num_skb_frags, page, offset)) {
344 put_page(page);
345 skb_coalesce_rx_frag(curr_skb, num_skb_frags - 1,
346 len, MAX_PACKET_LEN);
347 } else {
348 skb_add_rx_frag(curr_skb, num_skb_frags, page,
349 offset, len,
350 MAX_PACKET_LEN);
351 }
327 --rq->num; 352 --rq->num;
328 } 353 }
329 return 0; 354 return 0;
@@ -341,8 +366,10 @@ static void receive_buf(struct receive_queue *rq, void *buf, unsigned int len)
341 if (unlikely(len < sizeof(struct virtio_net_hdr) + ETH_HLEN)) { 366 if (unlikely(len < sizeof(struct virtio_net_hdr) + ETH_HLEN)) {
342 pr_debug("%s: short packet %i\n", dev->name, len); 367 pr_debug("%s: short packet %i\n", dev->name, len);
343 dev->stats.rx_length_errors++; 368 dev->stats.rx_length_errors++;
344 if (vi->mergeable_rx_bufs || vi->big_packets) 369 if (vi->big_packets)
345 give_pages(rq, buf); 370 give_pages(rq, buf);
371 else if (vi->mergeable_rx_bufs)
372 put_page(virt_to_head_page(buf));
346 else 373 else
347 dev_kfree_skb(buf); 374 dev_kfree_skb(buf);
348 return; 375 return;
@@ -352,19 +379,28 @@ static void receive_buf(struct receive_queue *rq, void *buf, unsigned int len)
352 skb = buf; 379 skb = buf;
353 len -= sizeof(struct virtio_net_hdr); 380 len -= sizeof(struct virtio_net_hdr);
354 skb_trim(skb, len); 381 skb_trim(skb, len);
382 } else if (vi->mergeable_rx_bufs) {
383 struct page *page = virt_to_head_page(buf);
384 skb = page_to_skb(rq, page,
385 (char *)buf - (char *)page_address(page),
386 len, MAX_PACKET_LEN);
387 if (unlikely(!skb)) {
388 dev->stats.rx_dropped++;
389 put_page(page);
390 return;
391 }
392 if (receive_mergeable(rq, skb)) {
393 dev_kfree_skb(skb);
394 return;
395 }
355 } else { 396 } else {
356 page = buf; 397 page = buf;
357 skb = page_to_skb(rq, page, len); 398 skb = page_to_skb(rq, page, 0, len, PAGE_SIZE);
358 if (unlikely(!skb)) { 399 if (unlikely(!skb)) {
359 dev->stats.rx_dropped++; 400 dev->stats.rx_dropped++;
360 give_pages(rq, page); 401 give_pages(rq, page);
361 return; 402 return;
362 } 403 }
363 if (vi->mergeable_rx_bufs)
364 if (receive_mergeable(rq, skb)) {
365 dev_kfree_skb(skb);
366 return;
367 }
368 } 404 }
369 405
370 hdr = skb_vnet_hdr(skb); 406 hdr = skb_vnet_hdr(skb);
@@ -501,18 +537,28 @@ static int add_recvbuf_big(struct receive_queue *rq, gfp_t gfp)
501 537
502static int add_recvbuf_mergeable(struct receive_queue *rq, gfp_t gfp) 538static int add_recvbuf_mergeable(struct receive_queue *rq, gfp_t gfp)
503{ 539{
504 struct page *page; 540 struct virtnet_info *vi = rq->vq->vdev->priv;
541 char *buf = NULL;
505 int err; 542 int err;
506 543
507 page = get_a_page(rq, gfp); 544 if (gfp & __GFP_WAIT) {
508 if (!page) 545 if (skb_page_frag_refill(MAX_PACKET_LEN, &vi->alloc_frag,
546 gfp)) {
547 buf = (char *)page_address(vi->alloc_frag.page) +
548 vi->alloc_frag.offset;
549 get_page(vi->alloc_frag.page);
550 vi->alloc_frag.offset += MAX_PACKET_LEN;
551 }
552 } else {
553 buf = netdev_alloc_frag(MAX_PACKET_LEN);
554 }
555 if (!buf)
509 return -ENOMEM; 556 return -ENOMEM;
510 557
511 sg_init_one(rq->sg, page_address(page), PAGE_SIZE); 558 sg_init_one(rq->sg, buf, MAX_PACKET_LEN);
512 559 err = virtqueue_add_inbuf(rq->vq, rq->sg, 1, buf, gfp);
513 err = virtqueue_add_inbuf(rq->vq, rq->sg, 1, page, gfp);
514 if (err < 0) 560 if (err < 0)
515 give_pages(rq, page); 561 put_page(virt_to_head_page(buf));
516 562
517 return err; 563 return err;
518} 564}
@@ -1065,7 +1111,6 @@ static int virtnet_vlan_rx_kill_vid(struct net_device *dev,
1065static void virtnet_clean_affinity(struct virtnet_info *vi, long hcpu) 1111static void virtnet_clean_affinity(struct virtnet_info *vi, long hcpu)
1066{ 1112{
1067 int i; 1113 int i;
1068 int cpu;
1069 1114
1070 if (vi->affinity_hint_set) { 1115 if (vi->affinity_hint_set) {
1071 for (i = 0; i < vi->max_queue_pairs; i++) { 1116 for (i = 0; i < vi->max_queue_pairs; i++) {
@@ -1075,16 +1120,6 @@ static void virtnet_clean_affinity(struct virtnet_info *vi, long hcpu)
1075 1120
1076 vi->affinity_hint_set = false; 1121 vi->affinity_hint_set = false;
1077 } 1122 }
1078
1079 i = 0;
1080 for_each_online_cpu(cpu) {
1081 if (cpu == hcpu) {
1082 *per_cpu_ptr(vi->vq_index, cpu) = -1;
1083 } else {
1084 *per_cpu_ptr(vi->vq_index, cpu) =
1085 ++i % vi->curr_queue_pairs;
1086 }
1087 }
1088} 1123}
1089 1124
1090static void virtnet_set_affinity(struct virtnet_info *vi) 1125static void virtnet_set_affinity(struct virtnet_info *vi)
@@ -1106,7 +1141,7 @@ static void virtnet_set_affinity(struct virtnet_info *vi)
1106 for_each_online_cpu(cpu) { 1141 for_each_online_cpu(cpu) {
1107 virtqueue_set_affinity(vi->rq[i].vq, cpu); 1142 virtqueue_set_affinity(vi->rq[i].vq, cpu);
1108 virtqueue_set_affinity(vi->sq[i].vq, cpu); 1143 virtqueue_set_affinity(vi->sq[i].vq, cpu);
1109 *per_cpu_ptr(vi->vq_index, cpu) = i; 1144 netif_set_xps_queue(vi->dev, cpumask_of(cpu), i);
1110 i++; 1145 i++;
1111 } 1146 }
1112 1147
@@ -1220,28 +1255,6 @@ static int virtnet_change_mtu(struct net_device *dev, int new_mtu)
1220 return 0; 1255 return 0;
1221} 1256}
1222 1257
1223/* To avoid contending a lock hold by a vcpu who would exit to host, select the
1224 * txq based on the processor id.
1225 */
1226static u16 virtnet_select_queue(struct net_device *dev, struct sk_buff *skb)
1227{
1228 int txq;
1229 struct virtnet_info *vi = netdev_priv(dev);
1230
1231 if (skb_rx_queue_recorded(skb)) {
1232 txq = skb_get_rx_queue(skb);
1233 } else {
1234 txq = *__this_cpu_ptr(vi->vq_index);
1235 if (txq == -1)
1236 txq = 0;
1237 }
1238
1239 while (unlikely(txq >= dev->real_num_tx_queues))
1240 txq -= dev->real_num_tx_queues;
1241
1242 return txq;
1243}
1244
1245static const struct net_device_ops virtnet_netdev = { 1258static const struct net_device_ops virtnet_netdev = {
1246 .ndo_open = virtnet_open, 1259 .ndo_open = virtnet_open,
1247 .ndo_stop = virtnet_close, 1260 .ndo_stop = virtnet_close,
@@ -1253,7 +1266,6 @@ static const struct net_device_ops virtnet_netdev = {
1253 .ndo_get_stats64 = virtnet_stats, 1266 .ndo_get_stats64 = virtnet_stats,
1254 .ndo_vlan_rx_add_vid = virtnet_vlan_rx_add_vid, 1267 .ndo_vlan_rx_add_vid = virtnet_vlan_rx_add_vid,
1255 .ndo_vlan_rx_kill_vid = virtnet_vlan_rx_kill_vid, 1268 .ndo_vlan_rx_kill_vid = virtnet_vlan_rx_kill_vid,
1256 .ndo_select_queue = virtnet_select_queue,
1257#ifdef CONFIG_NET_POLL_CONTROLLER 1269#ifdef CONFIG_NET_POLL_CONTROLLER
1258 .ndo_poll_controller = virtnet_netpoll, 1270 .ndo_poll_controller = virtnet_netpoll,
1259#endif 1271#endif
@@ -1336,8 +1348,10 @@ static void free_unused_bufs(struct virtnet_info *vi)
1336 struct virtqueue *vq = vi->rq[i].vq; 1348 struct virtqueue *vq = vi->rq[i].vq;
1337 1349
1338 while ((buf = virtqueue_detach_unused_buf(vq)) != NULL) { 1350 while ((buf = virtqueue_detach_unused_buf(vq)) != NULL) {
1339 if (vi->mergeable_rx_bufs || vi->big_packets) 1351 if (vi->big_packets)
1340 give_pages(&vi->rq[i], buf); 1352 give_pages(&vi->rq[i], buf);
1353 else if (vi->mergeable_rx_bufs)
1354 put_page(virt_to_head_page(buf));
1341 else 1355 else
1342 dev_kfree_skb(buf); 1356 dev_kfree_skb(buf);
1343 --vi->rq[i].num; 1357 --vi->rq[i].num;
@@ -1562,10 +1576,6 @@ static int virtnet_probe(struct virtio_device *vdev)
1562 if (vi->stats == NULL) 1576 if (vi->stats == NULL)
1563 goto free; 1577 goto free;
1564 1578
1565 vi->vq_index = alloc_percpu(int);
1566 if (vi->vq_index == NULL)
1567 goto free_stats;
1568
1569 mutex_init(&vi->config_lock); 1579 mutex_init(&vi->config_lock);
1570 vi->config_enable = true; 1580 vi->config_enable = true;
1571 INIT_WORK(&vi->config_work, virtnet_config_changed_work); 1581 INIT_WORK(&vi->config_work, virtnet_config_changed_work);
@@ -1592,7 +1602,7 @@ static int virtnet_probe(struct virtio_device *vdev)
1592 /* Allocate/initialize the rx/tx queues, and invoke find_vqs */ 1602 /* Allocate/initialize the rx/tx queues, and invoke find_vqs */
1593 err = init_vqs(vi); 1603 err = init_vqs(vi);
1594 if (err) 1604 if (err)
1595 goto free_index; 1605 goto free_stats;
1596 1606
1597 netif_set_real_num_tx_queues(dev, 1); 1607 netif_set_real_num_tx_queues(dev, 1);
1598 netif_set_real_num_rx_queues(dev, 1); 1608 netif_set_real_num_rx_queues(dev, 1);
@@ -1643,8 +1653,8 @@ free_recv_bufs:
1643free_vqs: 1653free_vqs:
1644 cancel_delayed_work_sync(&vi->refill); 1654 cancel_delayed_work_sync(&vi->refill);
1645 virtnet_del_vqs(vi); 1655 virtnet_del_vqs(vi);
1646free_index: 1656 if (vi->alloc_frag.page)
1647 free_percpu(vi->vq_index); 1657 put_page(vi->alloc_frag.page);
1648free_stats: 1658free_stats:
1649 free_percpu(vi->stats); 1659 free_percpu(vi->stats);
1650free: 1660free:
@@ -1678,10 +1688,11 @@ static void virtnet_remove(struct virtio_device *vdev)
1678 unregister_netdev(vi->dev); 1688 unregister_netdev(vi->dev);
1679 1689
1680 remove_vq_common(vi); 1690 remove_vq_common(vi);
1691 if (vi->alloc_frag.page)
1692 put_page(vi->alloc_frag.page);
1681 1693
1682 flush_work(&vi->config_work); 1694 flush_work(&vi->config_work);
1683 1695
1684 free_percpu(vi->vq_index);
1685 free_percpu(vi->stats); 1696 free_percpu(vi->stats);
1686 free_netdev(vi->dev); 1697 free_netdev(vi->dev);
1687} 1698}
diff --git a/drivers/net/vmxnet3/vmxnet3_int.h b/drivers/net/vmxnet3/vmxnet3_int.h
index a03f358fd58b..12040a35d95d 100644
--- a/drivers/net/vmxnet3/vmxnet3_int.h
+++ b/drivers/net/vmxnet3/vmxnet3_int.h
@@ -410,9 +410,9 @@ int
410vmxnet3_create_queues(struct vmxnet3_adapter *adapter, 410vmxnet3_create_queues(struct vmxnet3_adapter *adapter,
411 u32 tx_ring_size, u32 rx_ring_size, u32 rx_ring2_size); 411 u32 tx_ring_size, u32 rx_ring_size, u32 rx_ring2_size);
412 412
413extern void vmxnet3_set_ethtool_ops(struct net_device *netdev); 413void vmxnet3_set_ethtool_ops(struct net_device *netdev);
414 414
415extern struct rtnl_link_stats64 * 415struct rtnl_link_stats64 *
416vmxnet3_get_stats64(struct net_device *dev, struct rtnl_link_stats64 *stats); 416vmxnet3_get_stats64(struct net_device *dev, struct rtnl_link_stats64 *stats);
417 417
418extern char vmxnet3_driver_name[]; 418extern char vmxnet3_driver_name[];
diff --git a/drivers/net/vxlan.c b/drivers/net/vxlan.c
index 2ef5b6219f3f..78df8f39e57c 100644
--- a/drivers/net/vxlan.c
+++ b/drivers/net/vxlan.c
@@ -60,10 +60,6 @@
60 60
61#define VXLAN_N_VID (1u << 24) 61#define VXLAN_N_VID (1u << 24)
62#define VXLAN_VID_MASK (VXLAN_N_VID - 1) 62#define VXLAN_VID_MASK (VXLAN_N_VID - 1)
63/* IP header + UDP + VXLAN + Ethernet header */
64#define VXLAN_HEADROOM (20 + 8 + 8 + 14)
65/* IPv6 header + UDP + VXLAN + Ethernet header */
66#define VXLAN6_HEADROOM (40 + 8 + 8 + 14)
67#define VXLAN_HLEN (sizeof(struct udphdr) + sizeof(struct vxlanhdr)) 63#define VXLAN_HLEN (sizeof(struct udphdr) + sizeof(struct vxlanhdr))
68 64
69#define VXLAN_FLAGS 0x08000000 /* struct vxlanhdr.vx_flags required value. */ 65#define VXLAN_FLAGS 0x08000000 /* struct vxlanhdr.vx_flags required value. */
@@ -2087,7 +2083,7 @@ static void vxlan_setup(struct net_device *dev)
2087 vxlan->age_timer.function = vxlan_cleanup; 2083 vxlan->age_timer.function = vxlan_cleanup;
2088 vxlan->age_timer.data = (unsigned long) vxlan; 2084 vxlan->age_timer.data = (unsigned long) vxlan;
2089 2085
2090 inet_get_local_port_range(&low, &high); 2086 inet_get_local_port_range(dev_net(dev), &low, &high);
2091 vxlan->port_min = low; 2087 vxlan->port_min = low;
2092 vxlan->port_max = high; 2088 vxlan->port_max = high;
2093 vxlan->dst_port = htons(vxlan_port); 2089 vxlan->dst_port = htons(vxlan_port);
@@ -2180,7 +2176,7 @@ static void vxlan_del_work(struct work_struct *work)
2180 * could be used for both IPv4 and IPv6 communications, but 2176 * could be used for both IPv4 and IPv6 communications, but
2181 * users may set bindv6only=1. 2177 * users may set bindv6only=1.
2182 */ 2178 */
2183static int create_v6_sock(struct net *net, __be16 port, struct socket **psock) 2179static struct socket *create_v6_sock(struct net *net, __be16 port)
2184{ 2180{
2185 struct sock *sk; 2181 struct sock *sk;
2186 struct socket *sock; 2182 struct socket *sock;
@@ -2193,7 +2189,7 @@ static int create_v6_sock(struct net *net, __be16 port, struct socket **psock)
2193 rc = sock_create_kern(AF_INET6, SOCK_DGRAM, IPPROTO_UDP, &sock); 2189 rc = sock_create_kern(AF_INET6, SOCK_DGRAM, IPPROTO_UDP, &sock);
2194 if (rc < 0) { 2190 if (rc < 0) {
2195 pr_debug("UDPv6 socket create failed\n"); 2191 pr_debug("UDPv6 socket create failed\n");
2196 return rc; 2192 return ERR_PTR(rc);
2197 } 2193 }
2198 2194
2199 /* Put in proper namespace */ 2195 /* Put in proper namespace */
@@ -2208,28 +2204,27 @@ static int create_v6_sock(struct net *net, __be16 port, struct socket **psock)
2208 pr_debug("bind for UDPv6 socket %pI6:%u (%d)\n", 2204 pr_debug("bind for UDPv6 socket %pI6:%u (%d)\n",
2209 &vxlan_addr.sin6_addr, ntohs(vxlan_addr.sin6_port), rc); 2205 &vxlan_addr.sin6_addr, ntohs(vxlan_addr.sin6_port), rc);
2210 sk_release_kernel(sk); 2206 sk_release_kernel(sk);
2211 return rc; 2207 return ERR_PTR(rc);
2212 } 2208 }
2213 /* At this point, IPv6 module should have been loaded in 2209 /* At this point, IPv6 module should have been loaded in
2214 * sock_create_kern(). 2210 * sock_create_kern().
2215 */ 2211 */
2216 BUG_ON(!ipv6_stub); 2212 BUG_ON(!ipv6_stub);
2217 2213
2218 *psock = sock;
2219 /* Disable multicast loopback */ 2214 /* Disable multicast loopback */
2220 inet_sk(sk)->mc_loop = 0; 2215 inet_sk(sk)->mc_loop = 0;
2221 return 0; 2216 return sock;
2222} 2217}
2223 2218
2224#else 2219#else
2225 2220
2226static int create_v6_sock(struct net *net, __be16 port, struct socket **psock) 2221static struct socket *create_v6_sock(struct net *net, __be16 port)
2227{ 2222{
2228 return -EPFNOSUPPORT; 2223 return ERR_PTR(-EPFNOSUPPORT);
2229} 2224}
2230#endif 2225#endif
2231 2226
2232static int create_v4_sock(struct net *net, __be16 port, struct socket **psock) 2227static struct socket *create_v4_sock(struct net *net, __be16 port)
2233{ 2228{
2234 struct sock *sk; 2229 struct sock *sk;
2235 struct socket *sock; 2230 struct socket *sock;
@@ -2244,7 +2239,7 @@ static int create_v4_sock(struct net *net, __be16 port, struct socket **psock)
2244 rc = sock_create_kern(AF_INET, SOCK_DGRAM, IPPROTO_UDP, &sock); 2239 rc = sock_create_kern(AF_INET, SOCK_DGRAM, IPPROTO_UDP, &sock);
2245 if (rc < 0) { 2240 if (rc < 0) {
2246 pr_debug("UDP socket create failed\n"); 2241 pr_debug("UDP socket create failed\n");
2247 return rc; 2242 return ERR_PTR(rc);
2248 } 2243 }
2249 2244
2250 /* Put in proper namespace */ 2245 /* Put in proper namespace */
@@ -2257,13 +2252,12 @@ static int create_v4_sock(struct net *net, __be16 port, struct socket **psock)
2257 pr_debug("bind for UDP socket %pI4:%u (%d)\n", 2252 pr_debug("bind for UDP socket %pI4:%u (%d)\n",
2258 &vxlan_addr.sin_addr, ntohs(vxlan_addr.sin_port), rc); 2253 &vxlan_addr.sin_addr, ntohs(vxlan_addr.sin_port), rc);
2259 sk_release_kernel(sk); 2254 sk_release_kernel(sk);
2260 return rc; 2255 return ERR_PTR(rc);
2261 } 2256 }
2262 2257
2263 *psock = sock;
2264 /* Disable multicast loopback */ 2258 /* Disable multicast loopback */
2265 inet_sk(sk)->mc_loop = 0; 2259 inet_sk(sk)->mc_loop = 0;
2266 return 0; 2260 return sock;
2267} 2261}
2268 2262
2269/* Create new listen socket if needed */ 2263/* Create new listen socket if needed */
@@ -2274,7 +2268,6 @@ static struct vxlan_sock *vxlan_socket_create(struct net *net, __be16 port,
2274 struct vxlan_sock *vs; 2268 struct vxlan_sock *vs;
2275 struct socket *sock; 2269 struct socket *sock;
2276 struct sock *sk; 2270 struct sock *sk;
2277 int rc = 0;
2278 unsigned int h; 2271 unsigned int h;
2279 2272
2280 vs = kmalloc(sizeof(*vs), GFP_KERNEL); 2273 vs = kmalloc(sizeof(*vs), GFP_KERNEL);
@@ -2287,12 +2280,12 @@ static struct vxlan_sock *vxlan_socket_create(struct net *net, __be16 port,
2287 INIT_WORK(&vs->del_work, vxlan_del_work); 2280 INIT_WORK(&vs->del_work, vxlan_del_work);
2288 2281
2289 if (ipv6) 2282 if (ipv6)
2290 rc = create_v6_sock(net, port, &sock); 2283 sock = create_v6_sock(net, port);
2291 else 2284 else
2292 rc = create_v4_sock(net, port, &sock); 2285 sock = create_v4_sock(net, port);
2293 if (rc < 0) { 2286 if (IS_ERR(sock)) {
2294 kfree(vs); 2287 kfree(vs);
2295 return ERR_PTR(rc); 2288 return ERR_CAST(sock);
2296 } 2289 }
2297 2290
2298 vs->sock = sock; 2291 vs->sock = sock;
diff --git a/drivers/net/wan/hostess_sv11.c b/drivers/net/wan/hostess_sv11.c
index 3d80e4267de8..3d741663fd67 100644
--- a/drivers/net/wan/hostess_sv11.c
+++ b/drivers/net/wan/hostess_sv11.c
@@ -220,7 +220,7 @@ static struct z8530_dev *sv11_init(int iobase, int irq)
220 /* We want a fast IRQ for this device. Actually we'd like an even faster 220 /* We want a fast IRQ for this device. Actually we'd like an even faster
221 IRQ ;) - This is one driver RtLinux is made for */ 221 IRQ ;) - This is one driver RtLinux is made for */
222 222
223 if (request_irq(irq, z8530_interrupt, IRQF_DISABLED, 223 if (request_irq(irq, z8530_interrupt, 0,
224 "Hostess SV11", sv) < 0) { 224 "Hostess SV11", sv) < 0) {
225 pr_warn("IRQ %d already in use\n", irq); 225 pr_warn("IRQ %d already in use\n", irq);
226 goto err_irq; 226 goto err_irq;
diff --git a/drivers/net/wan/sealevel.c b/drivers/net/wan/sealevel.c
index 4f7748478984..27860b4f5908 100644
--- a/drivers/net/wan/sealevel.c
+++ b/drivers/net/wan/sealevel.c
@@ -266,7 +266,7 @@ static __init struct slvl_board *slvl_init(int iobase, int irq,
266 /* We want a fast IRQ for this device. Actually we'd like an even faster 266 /* We want a fast IRQ for this device. Actually we'd like an even faster
267 IRQ ;) - This is one driver RtLinux is made for */ 267 IRQ ;) - This is one driver RtLinux is made for */
268 268
269 if (request_irq(irq, z8530_interrupt, IRQF_DISABLED, 269 if (request_irq(irq, z8530_interrupt, 0,
270 "SeaLevel", dev) < 0) { 270 "SeaLevel", dev) < 0) {
271 pr_warn("IRQ %d already in use\n", irq); 271 pr_warn("IRQ %d already in use\n", irq);
272 goto err_request_irq; 272 goto err_request_irq;
diff --git a/drivers/net/wan/x25_asy.h b/drivers/net/wan/x25_asy.h
index 8f0fc2e57e2b..f57ee67836ae 100644
--- a/drivers/net/wan/x25_asy.h
+++ b/drivers/net/wan/x25_asy.h
@@ -41,6 +41,6 @@ struct x25_asy {
41 41
42#define X25_ASY_MAGIC 0x5303 42#define X25_ASY_MAGIC 0x5303
43 43
44extern int x25_asy_init(struct net_device *dev); 44int x25_asy_init(struct net_device *dev);
45 45
46#endif /* _LINUX_X25_ASY.H */ 46#endif /* _LINUX_X25_ASY.H */
diff --git a/drivers/net/wan/z85230.h b/drivers/net/wan/z85230.h
index f29d554fc07d..2416a9d60bd6 100644
--- a/drivers/net/wan/z85230.h
+++ b/drivers/net/wan/z85230.h
@@ -395,20 +395,19 @@ struct z8530_dev
395extern u8 z8530_dead_port[]; 395extern u8 z8530_dead_port[];
396extern u8 z8530_hdlc_kilostream_85230[]; 396extern u8 z8530_hdlc_kilostream_85230[];
397extern u8 z8530_hdlc_kilostream[]; 397extern u8 z8530_hdlc_kilostream[];
398extern irqreturn_t z8530_interrupt(int, void *); 398irqreturn_t z8530_interrupt(int, void *);
399extern void z8530_describe(struct z8530_dev *, char *mapping, unsigned long io); 399void z8530_describe(struct z8530_dev *, char *mapping, unsigned long io);
400extern int z8530_init(struct z8530_dev *); 400int z8530_init(struct z8530_dev *);
401extern int z8530_shutdown(struct z8530_dev *); 401int z8530_shutdown(struct z8530_dev *);
402extern int z8530_sync_open(struct net_device *, struct z8530_channel *); 402int z8530_sync_open(struct net_device *, struct z8530_channel *);
403extern int z8530_sync_close(struct net_device *, struct z8530_channel *); 403int z8530_sync_close(struct net_device *, struct z8530_channel *);
404extern int z8530_sync_dma_open(struct net_device *, struct z8530_channel *); 404int z8530_sync_dma_open(struct net_device *, struct z8530_channel *);
405extern int z8530_sync_dma_close(struct net_device *, struct z8530_channel *); 405int z8530_sync_dma_close(struct net_device *, struct z8530_channel *);
406extern int z8530_sync_txdma_open(struct net_device *, struct z8530_channel *); 406int z8530_sync_txdma_open(struct net_device *, struct z8530_channel *);
407extern int z8530_sync_txdma_close(struct net_device *, struct z8530_channel *); 407int z8530_sync_txdma_close(struct net_device *, struct z8530_channel *);
408extern int z8530_channel_load(struct z8530_channel *, u8 *); 408int z8530_channel_load(struct z8530_channel *, u8 *);
409extern netdev_tx_t z8530_queue_xmit(struct z8530_channel *c, 409netdev_tx_t z8530_queue_xmit(struct z8530_channel *c, struct sk_buff *skb);
410 struct sk_buff *skb); 410void z8530_null_rx(struct z8530_channel *c, struct sk_buff *skb);
411extern void z8530_null_rx(struct z8530_channel *c, struct sk_buff *skb);
412 411
413 412
414/* 413/*
diff --git a/drivers/net/wimax/i2400m/i2400m-usb.h b/drivers/net/wimax/i2400m/i2400m-usb.h
index 9f1e947f3557..649ecad6844c 100644
--- a/drivers/net/wimax/i2400m/i2400m-usb.h
+++ b/drivers/net/wimax/i2400m/i2400m-usb.h
@@ -256,21 +256,20 @@ void i2400mu_init(struct i2400mu *i2400mu)
256 i2400mu->rx_size_auto_shrink = 1; 256 i2400mu->rx_size_auto_shrink = 1;
257} 257}
258 258
259extern int i2400mu_notification_setup(struct i2400mu *); 259int i2400mu_notification_setup(struct i2400mu *);
260extern void i2400mu_notification_release(struct i2400mu *); 260void i2400mu_notification_release(struct i2400mu *);
261 261
262extern int i2400mu_rx_setup(struct i2400mu *); 262int i2400mu_rx_setup(struct i2400mu *);
263extern void i2400mu_rx_release(struct i2400mu *); 263void i2400mu_rx_release(struct i2400mu *);
264extern void i2400mu_rx_kick(struct i2400mu *); 264void i2400mu_rx_kick(struct i2400mu *);
265 265
266extern int i2400mu_tx_setup(struct i2400mu *); 266int i2400mu_tx_setup(struct i2400mu *);
267extern void i2400mu_tx_release(struct i2400mu *); 267void i2400mu_tx_release(struct i2400mu *);
268extern void i2400mu_bus_tx_kick(struct i2400m *); 268void i2400mu_bus_tx_kick(struct i2400m *);
269 269
270extern ssize_t i2400mu_bus_bm_cmd_send(struct i2400m *, 270ssize_t i2400mu_bus_bm_cmd_send(struct i2400m *,
271 const struct i2400m_bootrom_header *, 271 const struct i2400m_bootrom_header *, size_t,
272 size_t, int); 272 int);
273extern ssize_t i2400mu_bus_bm_wait_for_ack(struct i2400m *, 273ssize_t i2400mu_bus_bm_wait_for_ack(struct i2400m *,
274 struct i2400m_bootrom_header *, 274 struct i2400m_bootrom_header *, size_t);
275 size_t);
276#endif /* #ifndef __I2400M_USB_H__ */ 275#endif /* #ifndef __I2400M_USB_H__ */
diff --git a/drivers/net/wimax/i2400m/i2400m.h b/drivers/net/wimax/i2400m/i2400m.h
index 79c6505b5c20..5a34e72bab9a 100644
--- a/drivers/net/wimax/i2400m/i2400m.h
+++ b/drivers/net/wimax/i2400m/i2400m.h
@@ -710,18 +710,18 @@ enum i2400m_bri {
710 I2400M_BRI_MAC_REINIT = 1 << 3, 710 I2400M_BRI_MAC_REINIT = 1 << 3,
711}; 711};
712 712
713extern void i2400m_bm_cmd_prepare(struct i2400m_bootrom_header *); 713void i2400m_bm_cmd_prepare(struct i2400m_bootrom_header *);
714extern int i2400m_dev_bootstrap(struct i2400m *, enum i2400m_bri); 714int i2400m_dev_bootstrap(struct i2400m *, enum i2400m_bri);
715extern int i2400m_read_mac_addr(struct i2400m *); 715int i2400m_read_mac_addr(struct i2400m *);
716extern int i2400m_bootrom_init(struct i2400m *, enum i2400m_bri); 716int i2400m_bootrom_init(struct i2400m *, enum i2400m_bri);
717extern int i2400m_is_boot_barker(struct i2400m *, const void *, size_t); 717int i2400m_is_boot_barker(struct i2400m *, const void *, size_t);
718static inline 718static inline
719int i2400m_is_d2h_barker(const void *buf) 719int i2400m_is_d2h_barker(const void *buf)
720{ 720{
721 const __le32 *barker = buf; 721 const __le32 *barker = buf;
722 return le32_to_cpu(*barker) == I2400M_D2H_MSG_BARKER; 722 return le32_to_cpu(*barker) == I2400M_D2H_MSG_BARKER;
723} 723}
724extern void i2400m_unknown_barker(struct i2400m *, const void *, size_t); 724void i2400m_unknown_barker(struct i2400m *, const void *, size_t);
725 725
726/* Make/grok boot-rom header commands */ 726/* Make/grok boot-rom header commands */
727 727
@@ -789,32 +789,31 @@ unsigned i2400m_brh_get_signature(const struct i2400m_bootrom_header *hdr)
789/* 789/*
790 * Driver / device setup and internal functions 790 * Driver / device setup and internal functions
791 */ 791 */
792extern void i2400m_init(struct i2400m *); 792void i2400m_init(struct i2400m *);
793extern int i2400m_reset(struct i2400m *, enum i2400m_reset_type); 793int i2400m_reset(struct i2400m *, enum i2400m_reset_type);
794extern void i2400m_netdev_setup(struct net_device *net_dev); 794void i2400m_netdev_setup(struct net_device *net_dev);
795extern int i2400m_sysfs_setup(struct device_driver *); 795int i2400m_sysfs_setup(struct device_driver *);
796extern void i2400m_sysfs_release(struct device_driver *); 796void i2400m_sysfs_release(struct device_driver *);
797extern int i2400m_tx_setup(struct i2400m *); 797int i2400m_tx_setup(struct i2400m *);
798extern void i2400m_wake_tx_work(struct work_struct *); 798void i2400m_wake_tx_work(struct work_struct *);
799extern void i2400m_tx_release(struct i2400m *); 799void i2400m_tx_release(struct i2400m *);
800 800
801extern int i2400m_rx_setup(struct i2400m *); 801int i2400m_rx_setup(struct i2400m *);
802extern void i2400m_rx_release(struct i2400m *); 802void i2400m_rx_release(struct i2400m *);
803 803
804extern void i2400m_fw_cache(struct i2400m *); 804void i2400m_fw_cache(struct i2400m *);
805extern void i2400m_fw_uncache(struct i2400m *); 805void i2400m_fw_uncache(struct i2400m *);
806 806
807extern void i2400m_net_rx(struct i2400m *, struct sk_buff *, unsigned, 807void i2400m_net_rx(struct i2400m *, struct sk_buff *, unsigned, const void *,
808 const void *, int); 808 int);
809extern void i2400m_net_erx(struct i2400m *, struct sk_buff *, 809void i2400m_net_erx(struct i2400m *, struct sk_buff *, enum i2400m_cs);
810 enum i2400m_cs); 810void i2400m_net_wake_stop(struct i2400m *);
811extern void i2400m_net_wake_stop(struct i2400m *);
812enum i2400m_pt; 811enum i2400m_pt;
813extern int i2400m_tx(struct i2400m *, const void *, size_t, enum i2400m_pt); 812int i2400m_tx(struct i2400m *, const void *, size_t, enum i2400m_pt);
814 813
815#ifdef CONFIG_DEBUG_FS 814#ifdef CONFIG_DEBUG_FS
816extern int i2400m_debugfs_add(struct i2400m *); 815int i2400m_debugfs_add(struct i2400m *);
817extern void i2400m_debugfs_rm(struct i2400m *); 816void i2400m_debugfs_rm(struct i2400m *);
818#else 817#else
819static inline int i2400m_debugfs_add(struct i2400m *i2400m) 818static inline int i2400m_debugfs_add(struct i2400m *i2400m)
820{ 819{
@@ -824,8 +823,8 @@ static inline void i2400m_debugfs_rm(struct i2400m *i2400m) {}
824#endif 823#endif
825 824
826/* Initialize/shutdown the device */ 825/* Initialize/shutdown the device */
827extern int i2400m_dev_initialize(struct i2400m *); 826int i2400m_dev_initialize(struct i2400m *);
828extern void i2400m_dev_shutdown(struct i2400m *); 827void i2400m_dev_shutdown(struct i2400m *);
829 828
830extern struct attribute_group i2400m_dev_attr_group; 829extern struct attribute_group i2400m_dev_attr_group;
831 830
@@ -873,21 +872,21 @@ void i2400m_put(struct i2400m *i2400m)
873 dev_put(i2400m->wimax_dev.net_dev); 872 dev_put(i2400m->wimax_dev.net_dev);
874} 873}
875 874
876extern int i2400m_dev_reset_handle(struct i2400m *, const char *); 875int i2400m_dev_reset_handle(struct i2400m *, const char *);
877extern int i2400m_pre_reset(struct i2400m *); 876int i2400m_pre_reset(struct i2400m *);
878extern int i2400m_post_reset(struct i2400m *); 877int i2400m_post_reset(struct i2400m *);
879extern void i2400m_error_recovery(struct i2400m *); 878void i2400m_error_recovery(struct i2400m *);
880 879
881/* 880/*
882 * _setup()/_release() are called by the probe/disconnect functions of 881 * _setup()/_release() are called by the probe/disconnect functions of
883 * the bus-specific drivers. 882 * the bus-specific drivers.
884 */ 883 */
885extern int i2400m_setup(struct i2400m *, enum i2400m_bri bm_flags); 884int i2400m_setup(struct i2400m *, enum i2400m_bri bm_flags);
886extern void i2400m_release(struct i2400m *); 885void i2400m_release(struct i2400m *);
887 886
888extern int i2400m_rx(struct i2400m *, struct sk_buff *); 887int i2400m_rx(struct i2400m *, struct sk_buff *);
889extern struct i2400m_msg_hdr *i2400m_tx_msg_get(struct i2400m *, size_t *); 888struct i2400m_msg_hdr *i2400m_tx_msg_get(struct i2400m *, size_t *);
890extern void i2400m_tx_msg_sent(struct i2400m *); 889void i2400m_tx_msg_sent(struct i2400m *);
891 890
892 891
893/* 892/*
@@ -900,20 +899,19 @@ struct device *i2400m_dev(struct i2400m *i2400m)
900 return i2400m->wimax_dev.net_dev->dev.parent; 899 return i2400m->wimax_dev.net_dev->dev.parent;
901} 900}
902 901
903extern int i2400m_msg_check_status(const struct i2400m_l3l4_hdr *, 902int i2400m_msg_check_status(const struct i2400m_l3l4_hdr *, char *, size_t);
904 char *, size_t); 903int i2400m_msg_size_check(struct i2400m *, const struct i2400m_l3l4_hdr *,
905extern int i2400m_msg_size_check(struct i2400m *, 904 size_t);
906 const struct i2400m_l3l4_hdr *, size_t); 905struct sk_buff *i2400m_msg_to_dev(struct i2400m *, const void *, size_t);
907extern struct sk_buff *i2400m_msg_to_dev(struct i2400m *, const void *, size_t); 906void i2400m_msg_to_dev_cancel_wait(struct i2400m *, int);
908extern void i2400m_msg_to_dev_cancel_wait(struct i2400m *, int); 907void i2400m_report_hook(struct i2400m *, const struct i2400m_l3l4_hdr *,
909extern void i2400m_report_hook(struct i2400m *, 908 size_t);
910 const struct i2400m_l3l4_hdr *, size_t); 909void i2400m_report_hook_work(struct work_struct *);
911extern void i2400m_report_hook_work(struct work_struct *); 910int i2400m_cmd_enter_powersave(struct i2400m *);
912extern int i2400m_cmd_enter_powersave(struct i2400m *); 911int i2400m_cmd_exit_idle(struct i2400m *);
913extern int i2400m_cmd_exit_idle(struct i2400m *); 912struct sk_buff *i2400m_get_device_info(struct i2400m *);
914extern struct sk_buff *i2400m_get_device_info(struct i2400m *); 913int i2400m_firmware_check(struct i2400m *);
915extern int i2400m_firmware_check(struct i2400m *); 914int i2400m_set_idle_timeout(struct i2400m *, unsigned);
916extern int i2400m_set_idle_timeout(struct i2400m *, unsigned);
917 915
918static inline 916static inline
919struct usb_endpoint_descriptor *usb_get_epd(struct usb_interface *iface, int ep) 917struct usb_endpoint_descriptor *usb_get_epd(struct usb_interface *iface, int ep)
@@ -921,10 +919,9 @@ struct usb_endpoint_descriptor *usb_get_epd(struct usb_interface *iface, int ep)
921 return &iface->cur_altsetting->endpoint[ep].desc; 919 return &iface->cur_altsetting->endpoint[ep].desc;
922} 920}
923 921
924extern int i2400m_op_rfkill_sw_toggle(struct wimax_dev *, 922int i2400m_op_rfkill_sw_toggle(struct wimax_dev *, enum wimax_rf_state);
925 enum wimax_rf_state); 923void i2400m_report_tlv_rf_switches_status(struct i2400m *,
926extern void i2400m_report_tlv_rf_switches_status( 924 const struct i2400m_tlv_rf_switches_status *);
927 struct i2400m *, const struct i2400m_tlv_rf_switches_status *);
928 925
929/* 926/*
930 * Helpers for firmware backwards compatibility 927 * Helpers for firmware backwards compatibility
@@ -968,8 +965,8 @@ void __i2400m_msleep(unsigned ms)
968 965
969 966
970/* module initialization helpers */ 967/* module initialization helpers */
971extern int i2400m_barker_db_init(const char *); 968int i2400m_barker_db_init(const char *);
972extern void i2400m_barker_db_exit(void); 969void i2400m_barker_db_exit(void);
973 970
974 971
975 972
diff --git a/drivers/net/wireless/adm8211.c b/drivers/net/wireless/adm8211.c
index f9a24e599dee..cfce83e1f273 100644
--- a/drivers/net/wireless/adm8211.c
+++ b/drivers/net/wireless/adm8211.c
@@ -1924,7 +1924,6 @@ static int adm8211_probe(struct pci_dev *pdev,
1924 pci_iounmap(pdev, priv->map); 1924 pci_iounmap(pdev, priv->map);
1925 1925
1926 err_free_dev: 1926 err_free_dev:
1927 pci_set_drvdata(pdev, NULL);
1928 ieee80211_free_hw(dev); 1927 ieee80211_free_hw(dev);
1929 1928
1930 err_free_reg: 1929 err_free_reg:
diff --git a/drivers/net/wireless/airo.c b/drivers/net/wireless/airo.c
index 7fe19648f10e..edf4b57c4aaa 100644
--- a/drivers/net/wireless/airo.c
+++ b/drivers/net/wireless/airo.c
@@ -5570,7 +5570,6 @@ static void airo_pci_remove(struct pci_dev *pdev)
5570 airo_print_info(dev->name, "Unregistering..."); 5570 airo_print_info(dev->name, "Unregistering...");
5571 stop_airo_card(dev, 1); 5571 stop_airo_card(dev, 1);
5572 pci_disable_device(pdev); 5572 pci_disable_device(pdev);
5573 pci_set_drvdata(pdev, NULL);
5574} 5573}
5575 5574
5576static int airo_pci_suspend(struct pci_dev *pdev, pm_message_t state) 5575static int airo_pci_suspend(struct pci_dev *pdev, pm_message_t state)
diff --git a/drivers/net/wireless/ath/Kconfig b/drivers/net/wireless/ath/Kconfig
index 1abf1d421173..c63d1159db5c 100644
--- a/drivers/net/wireless/ath/Kconfig
+++ b/drivers/net/wireless/ath/Kconfig
@@ -25,6 +25,23 @@ config ATH_DEBUG
25 Say Y, if you want to debug atheros wireless drivers. 25 Say Y, if you want to debug atheros wireless drivers.
26 Right now only ath9k makes use of this. 26 Right now only ath9k makes use of this.
27 27
28config ATH_REG_DYNAMIC_USER_REG_HINTS
29 bool "Atheros dynamic user regulatory hints"
30 depends on CFG80211_CERTIFICATION_ONUS
31 default n
32 ---help---
33 Say N. This should only be enabled in countries where
34 this feature is explicitly allowed and only on cards that
35 specifically have been tested for this.
36
37config ATH_REG_DYNAMIC_USER_CERT_TESTING
38 bool "Atheros dynamic user regulatory testing"
39 depends on ATH_REG_DYNAMIC_USER_REG_HINTS && CFG80211_CERTIFICATION_ONUS
40 default n
41 ---help---
42 Say N. This should only be enabled on systems
43 undergoing certification testing.
44
28source "drivers/net/wireless/ath/ath5k/Kconfig" 45source "drivers/net/wireless/ath/ath5k/Kconfig"
29source "drivers/net/wireless/ath/ath9k/Kconfig" 46source "drivers/net/wireless/ath/ath9k/Kconfig"
30source "drivers/net/wireless/ath/carl9170/Kconfig" 47source "drivers/net/wireless/ath/carl9170/Kconfig"
@@ -32,5 +49,6 @@ source "drivers/net/wireless/ath/ath6kl/Kconfig"
32source "drivers/net/wireless/ath/ar5523/Kconfig" 49source "drivers/net/wireless/ath/ar5523/Kconfig"
33source "drivers/net/wireless/ath/wil6210/Kconfig" 50source "drivers/net/wireless/ath/wil6210/Kconfig"
34source "drivers/net/wireless/ath/ath10k/Kconfig" 51source "drivers/net/wireless/ath/ath10k/Kconfig"
52source "drivers/net/wireless/ath/wcn36xx/Kconfig"
35 53
36endif 54endif
diff --git a/drivers/net/wireless/ath/Makefile b/drivers/net/wireless/ath/Makefile
index fb05cfd19361..7d023b0f13b4 100644
--- a/drivers/net/wireless/ath/Makefile
+++ b/drivers/net/wireless/ath/Makefile
@@ -5,13 +5,16 @@ obj-$(CONFIG_ATH6KL) += ath6kl/
5obj-$(CONFIG_AR5523) += ar5523/ 5obj-$(CONFIG_AR5523) += ar5523/
6obj-$(CONFIG_WIL6210) += wil6210/ 6obj-$(CONFIG_WIL6210) += wil6210/
7obj-$(CONFIG_ATH10K) += ath10k/ 7obj-$(CONFIG_ATH10K) += ath10k/
8obj-$(CONFIG_WCN36XX) += wcn36xx/
8 9
9obj-$(CONFIG_ATH_COMMON) += ath.o 10obj-$(CONFIG_ATH_COMMON) += ath.o
10 11
11ath-objs := main.o \ 12ath-objs := main.o \
12 regd.o \ 13 regd.o \
13 hw.o \ 14 hw.o \
14 key.o 15 key.o \
16 dfs_pattern_detector.o \
17 dfs_pri_detector.o
15 18
16ath-$(CONFIG_ATH_DEBUG) += debug.o 19ath-$(CONFIG_ATH_DEBUG) += debug.o
17ccflags-y += -D__CHECK_ENDIAN__ 20ccflags-y += -D__CHECK_ENDIAN__
diff --git a/drivers/net/wireless/ath/ar5523/ar5523.c b/drivers/net/wireless/ath/ar5523/ar5523.c
index 17d7fece35d2..280fc3d53a36 100644
--- a/drivers/net/wireless/ath/ar5523/ar5523.c
+++ b/drivers/net/wireless/ath/ar5523/ar5523.c
@@ -1762,6 +1762,7 @@ static struct usb_device_id ar5523_id_table[] = {
1762 AR5523_DEVICE_UX(0x2001, 0x3a00), /* Dlink / DWLAG132 */ 1762 AR5523_DEVICE_UX(0x2001, 0x3a00), /* Dlink / DWLAG132 */
1763 AR5523_DEVICE_UG(0x2001, 0x3a02), /* Dlink / DWLG132 */ 1763 AR5523_DEVICE_UG(0x2001, 0x3a02), /* Dlink / DWLG132 */
1764 AR5523_DEVICE_UX(0x2001, 0x3a04), /* Dlink / DWLAG122 */ 1764 AR5523_DEVICE_UX(0x2001, 0x3a04), /* Dlink / DWLAG122 */
1765 AR5523_DEVICE_UG(0x07d1, 0x3a07), /* D-Link / WUA-2340 rev A1 */
1765 AR5523_DEVICE_UG(0x1690, 0x0712), /* Gigaset / AR5523 */ 1766 AR5523_DEVICE_UG(0x1690, 0x0712), /* Gigaset / AR5523 */
1766 AR5523_DEVICE_UG(0x1690, 0x0710), /* Gigaset / SMCWUSBTG */ 1767 AR5523_DEVICE_UG(0x1690, 0x0710), /* Gigaset / SMCWUSBTG */
1767 AR5523_DEVICE_UG(0x129b, 0x160c), /* Gigaset / USB stick 108 1768 AR5523_DEVICE_UG(0x129b, 0x160c), /* Gigaset / USB stick 108
diff --git a/drivers/net/wireless/ath/ath10k/bmi.c b/drivers/net/wireless/ath/ath10k/bmi.c
index 744da6d1c405..a1f099628850 100644
--- a/drivers/net/wireless/ath/ath10k/bmi.c
+++ b/drivers/net/wireless/ath/ath10k/bmi.c
@@ -22,7 +22,8 @@
22 22
23void ath10k_bmi_start(struct ath10k *ar) 23void ath10k_bmi_start(struct ath10k *ar)
24{ 24{
25 ath10k_dbg(ATH10K_DBG_CORE, "BMI started\n"); 25 ath10k_dbg(ATH10K_DBG_BMI, "bmi start\n");
26
26 ar->bmi.done_sent = false; 27 ar->bmi.done_sent = false;
27} 28}
28 29
@@ -32,8 +33,10 @@ int ath10k_bmi_done(struct ath10k *ar)
32 u32 cmdlen = sizeof(cmd.id) + sizeof(cmd.done); 33 u32 cmdlen = sizeof(cmd.id) + sizeof(cmd.done);
33 int ret; 34 int ret;
34 35
36 ath10k_dbg(ATH10K_DBG_BMI, "bmi done\n");
37
35 if (ar->bmi.done_sent) { 38 if (ar->bmi.done_sent) {
36 ath10k_dbg(ATH10K_DBG_CORE, "%s skipped\n", __func__); 39 ath10k_dbg(ATH10K_DBG_BMI, "bmi skipped\n");
37 return 0; 40 return 0;
38 } 41 }
39 42
@@ -46,7 +49,6 @@ int ath10k_bmi_done(struct ath10k *ar)
46 return ret; 49 return ret;
47 } 50 }
48 51
49 ath10k_dbg(ATH10K_DBG_CORE, "BMI done\n");
50 return 0; 52 return 0;
51} 53}
52 54
@@ -59,6 +61,8 @@ int ath10k_bmi_get_target_info(struct ath10k *ar,
59 u32 resplen = sizeof(resp.get_target_info); 61 u32 resplen = sizeof(resp.get_target_info);
60 int ret; 62 int ret;
61 63
64 ath10k_dbg(ATH10K_DBG_BMI, "bmi get target info\n");
65
62 if (ar->bmi.done_sent) { 66 if (ar->bmi.done_sent) {
63 ath10k_warn("BMI Get Target Info Command disallowed\n"); 67 ath10k_warn("BMI Get Target Info Command disallowed\n");
64 return -EBUSY; 68 return -EBUSY;
@@ -80,6 +84,7 @@ int ath10k_bmi_get_target_info(struct ath10k *ar,
80 84
81 target_info->version = __le32_to_cpu(resp.get_target_info.version); 85 target_info->version = __le32_to_cpu(resp.get_target_info.version);
82 target_info->type = __le32_to_cpu(resp.get_target_info.type); 86 target_info->type = __le32_to_cpu(resp.get_target_info.type);
87
83 return 0; 88 return 0;
84} 89}
85 90
@@ -92,15 +97,14 @@ int ath10k_bmi_read_memory(struct ath10k *ar,
92 u32 rxlen; 97 u32 rxlen;
93 int ret; 98 int ret;
94 99
100 ath10k_dbg(ATH10K_DBG_BMI, "bmi read address 0x%x length %d\n",
101 address, length);
102
95 if (ar->bmi.done_sent) { 103 if (ar->bmi.done_sent) {
96 ath10k_warn("command disallowed\n"); 104 ath10k_warn("command disallowed\n");
97 return -EBUSY; 105 return -EBUSY;
98 } 106 }
99 107
100 ath10k_dbg(ATH10K_DBG_CORE,
101 "%s: (device: 0x%p, address: 0x%x, length: %d)\n",
102 __func__, ar, address, length);
103
104 while (length) { 108 while (length) {
105 rxlen = min_t(u32, length, BMI_MAX_DATA_SIZE); 109 rxlen = min_t(u32, length, BMI_MAX_DATA_SIZE);
106 110
@@ -133,15 +137,14 @@ int ath10k_bmi_write_memory(struct ath10k *ar,
133 u32 txlen; 137 u32 txlen;
134 int ret; 138 int ret;
135 139
140 ath10k_dbg(ATH10K_DBG_BMI, "bmi write address 0x%x length %d\n",
141 address, length);
142
136 if (ar->bmi.done_sent) { 143 if (ar->bmi.done_sent) {
137 ath10k_warn("command disallowed\n"); 144 ath10k_warn("command disallowed\n");
138 return -EBUSY; 145 return -EBUSY;
139 } 146 }
140 147
141 ath10k_dbg(ATH10K_DBG_CORE,
142 "%s: (device: 0x%p, address: 0x%x, length: %d)\n",
143 __func__, ar, address, length);
144
145 while (length) { 148 while (length) {
146 txlen = min(length, BMI_MAX_DATA_SIZE - hdrlen); 149 txlen = min(length, BMI_MAX_DATA_SIZE - hdrlen);
147 150
@@ -180,15 +183,14 @@ int ath10k_bmi_execute(struct ath10k *ar, u32 address, u32 *param)
180 u32 resplen = sizeof(resp.execute); 183 u32 resplen = sizeof(resp.execute);
181 int ret; 184 int ret;
182 185
186 ath10k_dbg(ATH10K_DBG_BMI, "bmi execute address 0x%x param 0x%x\n",
187 address, *param);
188
183 if (ar->bmi.done_sent) { 189 if (ar->bmi.done_sent) {
184 ath10k_warn("command disallowed\n"); 190 ath10k_warn("command disallowed\n");
185 return -EBUSY; 191 return -EBUSY;
186 } 192 }
187 193
188 ath10k_dbg(ATH10K_DBG_CORE,
189 "%s: (device: 0x%p, address: 0x%x, param: %d)\n",
190 __func__, ar, address, *param);
191
192 cmd.id = __cpu_to_le32(BMI_EXECUTE); 194 cmd.id = __cpu_to_le32(BMI_EXECUTE);
193 cmd.execute.addr = __cpu_to_le32(address); 195 cmd.execute.addr = __cpu_to_le32(address);
194 cmd.execute.param = __cpu_to_le32(*param); 196 cmd.execute.param = __cpu_to_le32(*param);
@@ -216,6 +218,9 @@ int ath10k_bmi_lz_data(struct ath10k *ar, const void *buffer, u32 length)
216 u32 txlen; 218 u32 txlen;
217 int ret; 219 int ret;
218 220
221 ath10k_dbg(ATH10K_DBG_BMI, "bmi lz data buffer 0x%p length %d\n",
222 buffer, length);
223
219 if (ar->bmi.done_sent) { 224 if (ar->bmi.done_sent) {
220 ath10k_warn("command disallowed\n"); 225 ath10k_warn("command disallowed\n");
221 return -EBUSY; 226 return -EBUSY;
@@ -250,6 +255,9 @@ int ath10k_bmi_lz_stream_start(struct ath10k *ar, u32 address)
250 u32 cmdlen = sizeof(cmd.id) + sizeof(cmd.lz_start); 255 u32 cmdlen = sizeof(cmd.id) + sizeof(cmd.lz_start);
251 int ret; 256 int ret;
252 257
258 ath10k_dbg(ATH10K_DBG_BMI, "bmi lz stream start address 0x%x\n",
259 address);
260
253 if (ar->bmi.done_sent) { 261 if (ar->bmi.done_sent) {
254 ath10k_warn("command disallowed\n"); 262 ath10k_warn("command disallowed\n");
255 return -EBUSY; 263 return -EBUSY;
@@ -275,6 +283,10 @@ int ath10k_bmi_fast_download(struct ath10k *ar,
275 u32 trailer_len = length - head_len; 283 u32 trailer_len = length - head_len;
276 int ret; 284 int ret;
277 285
286 ath10k_dbg(ATH10K_DBG_BMI,
287 "bmi fast download address 0x%x buffer 0x%p length %d\n",
288 address, buffer, length);
289
278 ret = ath10k_bmi_lz_stream_start(ar, address); 290 ret = ath10k_bmi_lz_stream_start(ar, address);
279 if (ret) 291 if (ret)
280 return ret; 292 return ret;
diff --git a/drivers/net/wireless/ath/ath10k/ce.c b/drivers/net/wireless/ath/ath10k/ce.c
index f8b969f518f8..e46951b8fb92 100644
--- a/drivers/net/wireless/ath/ath10k/ce.c
+++ b/drivers/net/wireless/ath/ath10k/ce.c
@@ -76,36 +76,7 @@ static inline void ath10k_ce_src_ring_write_index_set(struct ath10k *ar,
76 u32 ce_ctrl_addr, 76 u32 ce_ctrl_addr,
77 unsigned int n) 77 unsigned int n)
78{ 78{
79 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 79 ath10k_pci_write32(ar, ce_ctrl_addr + SR_WR_INDEX_ADDRESS, n);
80 void __iomem *indicator_addr;
81
82 if (!test_bit(ATH10K_PCI_FEATURE_HW_1_0_WORKAROUND, ar_pci->features)) {
83 ath10k_pci_write32(ar, ce_ctrl_addr + SR_WR_INDEX_ADDRESS, n);
84 return;
85 }
86
87 /* workaround for QCA988x_1.0 HW CE */
88 indicator_addr = ar_pci->mem + ce_ctrl_addr + DST_WATERMARK_ADDRESS;
89
90 if (ce_ctrl_addr == ath10k_ce_base_address(CDC_WAR_DATA_CE)) {
91 iowrite32((CDC_WAR_MAGIC_STR | n), indicator_addr);
92 } else {
93 unsigned long irq_flags;
94 local_irq_save(irq_flags);
95 iowrite32(1, indicator_addr);
96
97 /*
98 * PCIE write waits for ACK in IPQ8K, there is no
99 * need to read back value.
100 */
101 (void)ioread32(indicator_addr);
102 (void)ioread32(indicator_addr); /* conservative */
103
104 ath10k_pci_write32(ar, ce_ctrl_addr + SR_WR_INDEX_ADDRESS, n);
105
106 iowrite32(0, indicator_addr);
107 local_irq_restore(irq_flags);
108 }
109} 80}
110 81
111static inline u32 ath10k_ce_src_ring_write_index_get(struct ath10k *ar, 82static inline u32 ath10k_ce_src_ring_write_index_get(struct ath10k *ar,
@@ -285,7 +256,7 @@ static inline void ath10k_ce_engine_int_status_clear(struct ath10k *ar,
285 * ath10k_ce_sendlist_send. 256 * ath10k_ce_sendlist_send.
286 * The caller takes responsibility for any needed locking. 257 * The caller takes responsibility for any needed locking.
287 */ 258 */
288static int ath10k_ce_send_nolock(struct ce_state *ce_state, 259static int ath10k_ce_send_nolock(struct ath10k_ce_pipe *ce_state,
289 void *per_transfer_context, 260 void *per_transfer_context,
290 u32 buffer, 261 u32 buffer,
291 unsigned int nbytes, 262 unsigned int nbytes,
@@ -293,7 +264,7 @@ static int ath10k_ce_send_nolock(struct ce_state *ce_state,
293 unsigned int flags) 264 unsigned int flags)
294{ 265{
295 struct ath10k *ar = ce_state->ar; 266 struct ath10k *ar = ce_state->ar;
296 struct ce_ring_state *src_ring = ce_state->src_ring; 267 struct ath10k_ce_ring *src_ring = ce_state->src_ring;
297 struct ce_desc *desc, *sdesc; 268 struct ce_desc *desc, *sdesc;
298 unsigned int nentries_mask = src_ring->nentries_mask; 269 unsigned int nentries_mask = src_ring->nentries_mask;
299 unsigned int sw_index = src_ring->sw_index; 270 unsigned int sw_index = src_ring->sw_index;
@@ -306,11 +277,13 @@ static int ath10k_ce_send_nolock(struct ce_state *ce_state,
306 ath10k_warn("%s: send more we can (nbytes: %d, max: %d)\n", 277 ath10k_warn("%s: send more we can (nbytes: %d, max: %d)\n",
307 __func__, nbytes, ce_state->src_sz_max); 278 __func__, nbytes, ce_state->src_sz_max);
308 279
309 ath10k_pci_wake(ar); 280 ret = ath10k_pci_wake(ar);
281 if (ret)
282 return ret;
310 283
311 if (unlikely(CE_RING_DELTA(nentries_mask, 284 if (unlikely(CE_RING_DELTA(nentries_mask,
312 write_index, sw_index - 1) <= 0)) { 285 write_index, sw_index - 1) <= 0)) {
313 ret = -EIO; 286 ret = -ENOSR;
314 goto exit; 287 goto exit;
315 } 288 }
316 289
@@ -346,7 +319,7 @@ exit:
346 return ret; 319 return ret;
347} 320}
348 321
349int ath10k_ce_send(struct ce_state *ce_state, 322int ath10k_ce_send(struct ath10k_ce_pipe *ce_state,
350 void *per_transfer_context, 323 void *per_transfer_context,
351 u32 buffer, 324 u32 buffer,
352 unsigned int nbytes, 325 unsigned int nbytes,
@@ -365,77 +338,26 @@ int ath10k_ce_send(struct ce_state *ce_state,
365 return ret; 338 return ret;
366} 339}
367 340
368void ath10k_ce_sendlist_buf_add(struct ce_sendlist *sendlist, u32 buffer, 341int ath10k_ce_num_free_src_entries(struct ath10k_ce_pipe *pipe)
369 unsigned int nbytes, u32 flags)
370{ 342{
371 unsigned int num_items = sendlist->num_items; 343 struct ath10k *ar = pipe->ar;
372 struct ce_sendlist_item *item;
373
374 item = &sendlist->item[num_items];
375 item->data = buffer;
376 item->u.nbytes = nbytes;
377 item->flags = flags;
378 sendlist->num_items++;
379}
380
381int ath10k_ce_sendlist_send(struct ce_state *ce_state,
382 void *per_transfer_context,
383 struct ce_sendlist *sendlist,
384 unsigned int transfer_id)
385{
386 struct ce_ring_state *src_ring = ce_state->src_ring;
387 struct ce_sendlist_item *item;
388 struct ath10k *ar = ce_state->ar;
389 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 344 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
390 unsigned int nentries_mask = src_ring->nentries_mask; 345 int delta;
391 unsigned int num_items = sendlist->num_items;
392 unsigned int sw_index;
393 unsigned int write_index;
394 int i, delta, ret = -ENOMEM;
395 346
396 spin_lock_bh(&ar_pci->ce_lock); 347 spin_lock_bh(&ar_pci->ce_lock);
397 348 delta = CE_RING_DELTA(pipe->src_ring->nentries_mask,
398 sw_index = src_ring->sw_index; 349 pipe->src_ring->write_index,
399 write_index = src_ring->write_index; 350 pipe->src_ring->sw_index - 1);
400
401 delta = CE_RING_DELTA(nentries_mask, write_index, sw_index - 1);
402
403 if (delta >= num_items) {
404 /*
405 * Handle all but the last item uniformly.
406 */
407 for (i = 0; i < num_items - 1; i++) {
408 item = &sendlist->item[i];
409 ret = ath10k_ce_send_nolock(ce_state,
410 CE_SENDLIST_ITEM_CTXT,
411 (u32) item->data,
412 item->u.nbytes, transfer_id,
413 item->flags |
414 CE_SEND_FLAG_GATHER);
415 if (ret)
416 ath10k_warn("CE send failed for item: %d\n", i);
417 }
418 /*
419 * Provide valid context pointer for final item.
420 */
421 item = &sendlist->item[i];
422 ret = ath10k_ce_send_nolock(ce_state, per_transfer_context,
423 (u32) item->data, item->u.nbytes,
424 transfer_id, item->flags);
425 if (ret)
426 ath10k_warn("CE send failed for last item: %d\n", i);
427 }
428
429 spin_unlock_bh(&ar_pci->ce_lock); 351 spin_unlock_bh(&ar_pci->ce_lock);
430 352
431 return ret; 353 return delta;
432} 354}
433 355
434int ath10k_ce_recv_buf_enqueue(struct ce_state *ce_state, 356int ath10k_ce_recv_buf_enqueue(struct ath10k_ce_pipe *ce_state,
435 void *per_recv_context, 357 void *per_recv_context,
436 u32 buffer) 358 u32 buffer)
437{ 359{
438 struct ce_ring_state *dest_ring = ce_state->dest_ring; 360 struct ath10k_ce_ring *dest_ring = ce_state->dest_ring;
439 u32 ctrl_addr = ce_state->ctrl_addr; 361 u32 ctrl_addr = ce_state->ctrl_addr;
440 struct ath10k *ar = ce_state->ar; 362 struct ath10k *ar = ce_state->ar;
441 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 363 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
@@ -448,7 +370,9 @@ int ath10k_ce_recv_buf_enqueue(struct ce_state *ce_state,
448 write_index = dest_ring->write_index; 370 write_index = dest_ring->write_index;
449 sw_index = dest_ring->sw_index; 371 sw_index = dest_ring->sw_index;
450 372
451 ath10k_pci_wake(ar); 373 ret = ath10k_pci_wake(ar);
374 if (ret)
375 goto out;
452 376
453 if (CE_RING_DELTA(nentries_mask, write_index, sw_index - 1) > 0) { 377 if (CE_RING_DELTA(nentries_mask, write_index, sw_index - 1) > 0) {
454 struct ce_desc *base = dest_ring->base_addr_owner_space; 378 struct ce_desc *base = dest_ring->base_addr_owner_space;
@@ -470,6 +394,8 @@ int ath10k_ce_recv_buf_enqueue(struct ce_state *ce_state,
470 ret = -EIO; 394 ret = -EIO;
471 } 395 }
472 ath10k_pci_sleep(ar); 396 ath10k_pci_sleep(ar);
397
398out:
473 spin_unlock_bh(&ar_pci->ce_lock); 399 spin_unlock_bh(&ar_pci->ce_lock);
474 400
475 return ret; 401 return ret;
@@ -479,14 +405,14 @@ int ath10k_ce_recv_buf_enqueue(struct ce_state *ce_state,
479 * Guts of ath10k_ce_completed_recv_next. 405 * Guts of ath10k_ce_completed_recv_next.
480 * The caller takes responsibility for any necessary locking. 406 * The caller takes responsibility for any necessary locking.
481 */ 407 */
482static int ath10k_ce_completed_recv_next_nolock(struct ce_state *ce_state, 408static int ath10k_ce_completed_recv_next_nolock(struct ath10k_ce_pipe *ce_state,
483 void **per_transfer_contextp, 409 void **per_transfer_contextp,
484 u32 *bufferp, 410 u32 *bufferp,
485 unsigned int *nbytesp, 411 unsigned int *nbytesp,
486 unsigned int *transfer_idp, 412 unsigned int *transfer_idp,
487 unsigned int *flagsp) 413 unsigned int *flagsp)
488{ 414{
489 struct ce_ring_state *dest_ring = ce_state->dest_ring; 415 struct ath10k_ce_ring *dest_ring = ce_state->dest_ring;
490 unsigned int nentries_mask = dest_ring->nentries_mask; 416 unsigned int nentries_mask = dest_ring->nentries_mask;
491 unsigned int sw_index = dest_ring->sw_index; 417 unsigned int sw_index = dest_ring->sw_index;
492 418
@@ -535,7 +461,7 @@ static int ath10k_ce_completed_recv_next_nolock(struct ce_state *ce_state,
535 return 0; 461 return 0;
536} 462}
537 463
538int ath10k_ce_completed_recv_next(struct ce_state *ce_state, 464int ath10k_ce_completed_recv_next(struct ath10k_ce_pipe *ce_state,
539 void **per_transfer_contextp, 465 void **per_transfer_contextp,
540 u32 *bufferp, 466 u32 *bufferp,
541 unsigned int *nbytesp, 467 unsigned int *nbytesp,
@@ -556,11 +482,11 @@ int ath10k_ce_completed_recv_next(struct ce_state *ce_state,
556 return ret; 482 return ret;
557} 483}
558 484
559int ath10k_ce_revoke_recv_next(struct ce_state *ce_state, 485int ath10k_ce_revoke_recv_next(struct ath10k_ce_pipe *ce_state,
560 void **per_transfer_contextp, 486 void **per_transfer_contextp,
561 u32 *bufferp) 487 u32 *bufferp)
562{ 488{
563 struct ce_ring_state *dest_ring; 489 struct ath10k_ce_ring *dest_ring;
564 unsigned int nentries_mask; 490 unsigned int nentries_mask;
565 unsigned int sw_index; 491 unsigned int sw_index;
566 unsigned int write_index; 492 unsigned int write_index;
@@ -612,19 +538,20 @@ int ath10k_ce_revoke_recv_next(struct ce_state *ce_state,
612 * Guts of ath10k_ce_completed_send_next. 538 * Guts of ath10k_ce_completed_send_next.
613 * The caller takes responsibility for any necessary locking. 539 * The caller takes responsibility for any necessary locking.
614 */ 540 */
615static int ath10k_ce_completed_send_next_nolock(struct ce_state *ce_state, 541static int ath10k_ce_completed_send_next_nolock(struct ath10k_ce_pipe *ce_state,
616 void **per_transfer_contextp, 542 void **per_transfer_contextp,
617 u32 *bufferp, 543 u32 *bufferp,
618 unsigned int *nbytesp, 544 unsigned int *nbytesp,
619 unsigned int *transfer_idp) 545 unsigned int *transfer_idp)
620{ 546{
621 struct ce_ring_state *src_ring = ce_state->src_ring; 547 struct ath10k_ce_ring *src_ring = ce_state->src_ring;
622 u32 ctrl_addr = ce_state->ctrl_addr; 548 u32 ctrl_addr = ce_state->ctrl_addr;
623 struct ath10k *ar = ce_state->ar; 549 struct ath10k *ar = ce_state->ar;
624 unsigned int nentries_mask = src_ring->nentries_mask; 550 unsigned int nentries_mask = src_ring->nentries_mask;
625 unsigned int sw_index = src_ring->sw_index; 551 unsigned int sw_index = src_ring->sw_index;
552 struct ce_desc *sdesc, *sbase;
626 unsigned int read_index; 553 unsigned int read_index;
627 int ret = -EIO; 554 int ret;
628 555
629 if (src_ring->hw_index == sw_index) { 556 if (src_ring->hw_index == sw_index) {
630 /* 557 /*
@@ -634,48 +561,54 @@ static int ath10k_ce_completed_send_next_nolock(struct ce_state *ce_state,
634 * the SW has really caught up to the HW, or if the cached 561 * the SW has really caught up to the HW, or if the cached
635 * value of the HW index has become stale. 562 * value of the HW index has become stale.
636 */ 563 */
637 ath10k_pci_wake(ar); 564
565 ret = ath10k_pci_wake(ar);
566 if (ret)
567 return ret;
568
638 src_ring->hw_index = 569 src_ring->hw_index =
639 ath10k_ce_src_ring_read_index_get(ar, ctrl_addr); 570 ath10k_ce_src_ring_read_index_get(ar, ctrl_addr);
640 src_ring->hw_index &= nentries_mask; 571 src_ring->hw_index &= nentries_mask;
572
641 ath10k_pci_sleep(ar); 573 ath10k_pci_sleep(ar);
642 } 574 }
575
643 read_index = src_ring->hw_index; 576 read_index = src_ring->hw_index;
644 577
645 if ((read_index != sw_index) && (read_index != 0xffffffff)) { 578 if ((read_index == sw_index) || (read_index == 0xffffffff))
646 struct ce_desc *sbase = src_ring->shadow_base; 579 return -EIO;
647 struct ce_desc *sdesc = CE_SRC_RING_TO_DESC(sbase, sw_index);
648 580
649 /* Return data from completed source descriptor */ 581 sbase = src_ring->shadow_base;
650 *bufferp = __le32_to_cpu(sdesc->addr); 582 sdesc = CE_SRC_RING_TO_DESC(sbase, sw_index);
651 *nbytesp = __le16_to_cpu(sdesc->nbytes);
652 *transfer_idp = MS(__le16_to_cpu(sdesc->flags),
653 CE_DESC_FLAGS_META_DATA);
654 583
655 if (per_transfer_contextp) 584 /* Return data from completed source descriptor */
656 *per_transfer_contextp = 585 *bufferp = __le32_to_cpu(sdesc->addr);
657 src_ring->per_transfer_context[sw_index]; 586 *nbytesp = __le16_to_cpu(sdesc->nbytes);
587 *transfer_idp = MS(__le16_to_cpu(sdesc->flags),
588 CE_DESC_FLAGS_META_DATA);
658 589
659 /* sanity */ 590 if (per_transfer_contextp)
660 src_ring->per_transfer_context[sw_index] = NULL; 591 *per_transfer_contextp =
592 src_ring->per_transfer_context[sw_index];
661 593
662 /* Update sw_index */ 594 /* sanity */
663 sw_index = CE_RING_IDX_INCR(nentries_mask, sw_index); 595 src_ring->per_transfer_context[sw_index] = NULL;
664 src_ring->sw_index = sw_index;
665 ret = 0;
666 }
667 596
668 return ret; 597 /* Update sw_index */
598 sw_index = CE_RING_IDX_INCR(nentries_mask, sw_index);
599 src_ring->sw_index = sw_index;
600
601 return 0;
669} 602}
670 603
671/* NB: Modeled after ath10k_ce_completed_send_next */ 604/* NB: Modeled after ath10k_ce_completed_send_next */
672int ath10k_ce_cancel_send_next(struct ce_state *ce_state, 605int ath10k_ce_cancel_send_next(struct ath10k_ce_pipe *ce_state,
673 void **per_transfer_contextp, 606 void **per_transfer_contextp,
674 u32 *bufferp, 607 u32 *bufferp,
675 unsigned int *nbytesp, 608 unsigned int *nbytesp,
676 unsigned int *transfer_idp) 609 unsigned int *transfer_idp)
677{ 610{
678 struct ce_ring_state *src_ring; 611 struct ath10k_ce_ring *src_ring;
679 unsigned int nentries_mask; 612 unsigned int nentries_mask;
680 unsigned int sw_index; 613 unsigned int sw_index;
681 unsigned int write_index; 614 unsigned int write_index;
@@ -727,7 +660,7 @@ int ath10k_ce_cancel_send_next(struct ce_state *ce_state,
727 return ret; 660 return ret;
728} 661}
729 662
730int ath10k_ce_completed_send_next(struct ce_state *ce_state, 663int ath10k_ce_completed_send_next(struct ath10k_ce_pipe *ce_state,
731 void **per_transfer_contextp, 664 void **per_transfer_contextp,
732 u32 *bufferp, 665 u32 *bufferp,
733 unsigned int *nbytesp, 666 unsigned int *nbytesp,
@@ -756,53 +689,29 @@ int ath10k_ce_completed_send_next(struct ce_state *ce_state,
756void ath10k_ce_per_engine_service(struct ath10k *ar, unsigned int ce_id) 689void ath10k_ce_per_engine_service(struct ath10k *ar, unsigned int ce_id)
757{ 690{
758 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 691 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
759 struct ce_state *ce_state = ar_pci->ce_id_to_state[ce_id]; 692 struct ath10k_ce_pipe *ce_state = &ar_pci->ce_states[ce_id];
760 u32 ctrl_addr = ce_state->ctrl_addr; 693 u32 ctrl_addr = ce_state->ctrl_addr;
761 void *transfer_context; 694 int ret;
762 u32 buf; 695
763 unsigned int nbytes; 696 ret = ath10k_pci_wake(ar);
764 unsigned int id; 697 if (ret)
765 unsigned int flags; 698 return;
766 699
767 ath10k_pci_wake(ar);
768 spin_lock_bh(&ar_pci->ce_lock); 700 spin_lock_bh(&ar_pci->ce_lock);
769 701
770 /* Clear the copy-complete interrupts that will be handled here. */ 702 /* Clear the copy-complete interrupts that will be handled here. */
771 ath10k_ce_engine_int_status_clear(ar, ctrl_addr, 703 ath10k_ce_engine_int_status_clear(ar, ctrl_addr,
772 HOST_IS_COPY_COMPLETE_MASK); 704 HOST_IS_COPY_COMPLETE_MASK);
773 705
774 if (ce_state->recv_cb) { 706 spin_unlock_bh(&ar_pci->ce_lock);
775 /*
776 * Pop completed recv buffers and call the registered
777 * recv callback for each
778 */
779 while (ath10k_ce_completed_recv_next_nolock(ce_state,
780 &transfer_context,
781 &buf, &nbytes,
782 &id, &flags) == 0) {
783 spin_unlock_bh(&ar_pci->ce_lock);
784 ce_state->recv_cb(ce_state, transfer_context, buf,
785 nbytes, id, flags);
786 spin_lock_bh(&ar_pci->ce_lock);
787 }
788 }
789 707
790 if (ce_state->send_cb) { 708 if (ce_state->recv_cb)
791 /* 709 ce_state->recv_cb(ce_state);
792 * Pop completed send buffers and call the registered 710
793 * send callback for each 711 if (ce_state->send_cb)
794 */ 712 ce_state->send_cb(ce_state);
795 while (ath10k_ce_completed_send_next_nolock(ce_state, 713
796 &transfer_context, 714 spin_lock_bh(&ar_pci->ce_lock);
797 &buf,
798 &nbytes,
799 &id) == 0) {
800 spin_unlock_bh(&ar_pci->ce_lock);
801 ce_state->send_cb(ce_state, transfer_context,
802 buf, nbytes, id);
803 spin_lock_bh(&ar_pci->ce_lock);
804 }
805 }
806 715
807 /* 716 /*
808 * Misc CE interrupts are not being handled, but still need 717 * Misc CE interrupts are not being handled, but still need
@@ -823,10 +732,13 @@ void ath10k_ce_per_engine_service(struct ath10k *ar, unsigned int ce_id)
823void ath10k_ce_per_engine_service_any(struct ath10k *ar) 732void ath10k_ce_per_engine_service_any(struct ath10k *ar)
824{ 733{
825 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 734 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
826 int ce_id; 735 int ce_id, ret;
827 u32 intr_summary; 736 u32 intr_summary;
828 737
829 ath10k_pci_wake(ar); 738 ret = ath10k_pci_wake(ar);
739 if (ret)
740 return;
741
830 intr_summary = CE_INTERRUPT_SUMMARY(ar); 742 intr_summary = CE_INTERRUPT_SUMMARY(ar);
831 743
832 for (ce_id = 0; intr_summary && (ce_id < ar_pci->ce_count); ce_id++) { 744 for (ce_id = 0; intr_summary && (ce_id < ar_pci->ce_count); ce_id++) {
@@ -849,13 +761,16 @@ void ath10k_ce_per_engine_service_any(struct ath10k *ar)
849 * 761 *
850 * Called with ce_lock held. 762 * Called with ce_lock held.
851 */ 763 */
852static void ath10k_ce_per_engine_handler_adjust(struct ce_state *ce_state, 764static void ath10k_ce_per_engine_handler_adjust(struct ath10k_ce_pipe *ce_state,
853 int disable_copy_compl_intr) 765 int disable_copy_compl_intr)
854{ 766{
855 u32 ctrl_addr = ce_state->ctrl_addr; 767 u32 ctrl_addr = ce_state->ctrl_addr;
856 struct ath10k *ar = ce_state->ar; 768 struct ath10k *ar = ce_state->ar;
769 int ret;
857 770
858 ath10k_pci_wake(ar); 771 ret = ath10k_pci_wake(ar);
772 if (ret)
773 return;
859 774
860 if ((!disable_copy_compl_intr) && 775 if ((!disable_copy_compl_intr) &&
861 (ce_state->send_cb || ce_state->recv_cb)) 776 (ce_state->send_cb || ce_state->recv_cb))
@@ -871,11 +786,14 @@ static void ath10k_ce_per_engine_handler_adjust(struct ce_state *ce_state,
871void ath10k_ce_disable_interrupts(struct ath10k *ar) 786void ath10k_ce_disable_interrupts(struct ath10k *ar)
872{ 787{
873 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 788 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
874 int ce_id; 789 int ce_id, ret;
790
791 ret = ath10k_pci_wake(ar);
792 if (ret)
793 return;
875 794
876 ath10k_pci_wake(ar);
877 for (ce_id = 0; ce_id < ar_pci->ce_count; ce_id++) { 795 for (ce_id = 0; ce_id < ar_pci->ce_count; ce_id++) {
878 struct ce_state *ce_state = ar_pci->ce_id_to_state[ce_id]; 796 struct ath10k_ce_pipe *ce_state = &ar_pci->ce_states[ce_id];
879 u32 ctrl_addr = ce_state->ctrl_addr; 797 u32 ctrl_addr = ce_state->ctrl_addr;
880 798
881 ath10k_ce_copy_complete_intr_disable(ar, ctrl_addr); 799 ath10k_ce_copy_complete_intr_disable(ar, ctrl_addr);
@@ -883,12 +801,8 @@ void ath10k_ce_disable_interrupts(struct ath10k *ar)
883 ath10k_pci_sleep(ar); 801 ath10k_pci_sleep(ar);
884} 802}
885 803
886void ath10k_ce_send_cb_register(struct ce_state *ce_state, 804void ath10k_ce_send_cb_register(struct ath10k_ce_pipe *ce_state,
887 void (*send_cb) (struct ce_state *ce_state, 805 void (*send_cb)(struct ath10k_ce_pipe *),
888 void *transfer_context,
889 u32 buffer,
890 unsigned int nbytes,
891 unsigned int transfer_id),
892 int disable_interrupts) 806 int disable_interrupts)
893{ 807{
894 struct ath10k *ar = ce_state->ar; 808 struct ath10k *ar = ce_state->ar;
@@ -900,13 +814,8 @@ void ath10k_ce_send_cb_register(struct ce_state *ce_state,
900 spin_unlock_bh(&ar_pci->ce_lock); 814 spin_unlock_bh(&ar_pci->ce_lock);
901} 815}
902 816
903void ath10k_ce_recv_cb_register(struct ce_state *ce_state, 817void ath10k_ce_recv_cb_register(struct ath10k_ce_pipe *ce_state,
904 void (*recv_cb) (struct ce_state *ce_state, 818 void (*recv_cb)(struct ath10k_ce_pipe *))
905 void *transfer_context,
906 u32 buffer,
907 unsigned int nbytes,
908 unsigned int transfer_id,
909 unsigned int flags))
910{ 819{
911 struct ath10k *ar = ce_state->ar; 820 struct ath10k *ar = ce_state->ar;
912 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 821 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
@@ -919,11 +828,11 @@ void ath10k_ce_recv_cb_register(struct ce_state *ce_state,
919 828
920static int ath10k_ce_init_src_ring(struct ath10k *ar, 829static int ath10k_ce_init_src_ring(struct ath10k *ar,
921 unsigned int ce_id, 830 unsigned int ce_id,
922 struct ce_state *ce_state, 831 struct ath10k_ce_pipe *ce_state,
923 const struct ce_attr *attr) 832 const struct ce_attr *attr)
924{ 833{
925 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 834 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
926 struct ce_ring_state *src_ring; 835 struct ath10k_ce_ring *src_ring;
927 unsigned int nentries = attr->src_nentries; 836 unsigned int nentries = attr->src_nentries;
928 unsigned int ce_nbytes; 837 unsigned int ce_nbytes;
929 u32 ctrl_addr = ath10k_ce_base_address(ce_id); 838 u32 ctrl_addr = ath10k_ce_base_address(ce_id);
@@ -937,19 +846,18 @@ static int ath10k_ce_init_src_ring(struct ath10k *ar,
937 return 0; 846 return 0;
938 } 847 }
939 848
940 ce_nbytes = sizeof(struct ce_ring_state) + (nentries * sizeof(void *)); 849 ce_nbytes = sizeof(struct ath10k_ce_ring) + (nentries * sizeof(void *));
941 ptr = kzalloc(ce_nbytes, GFP_KERNEL); 850 ptr = kzalloc(ce_nbytes, GFP_KERNEL);
942 if (ptr == NULL) 851 if (ptr == NULL)
943 return -ENOMEM; 852 return -ENOMEM;
944 853
945 ce_state->src_ring = (struct ce_ring_state *)ptr; 854 ce_state->src_ring = (struct ath10k_ce_ring *)ptr;
946 src_ring = ce_state->src_ring; 855 src_ring = ce_state->src_ring;
947 856
948 ptr += sizeof(struct ce_ring_state); 857 ptr += sizeof(struct ath10k_ce_ring);
949 src_ring->nentries = nentries; 858 src_ring->nentries = nentries;
950 src_ring->nentries_mask = nentries - 1; 859 src_ring->nentries_mask = nentries - 1;
951 860
952 ath10k_pci_wake(ar);
953 src_ring->sw_index = ath10k_ce_src_ring_read_index_get(ar, ctrl_addr); 861 src_ring->sw_index = ath10k_ce_src_ring_read_index_get(ar, ctrl_addr);
954 src_ring->sw_index &= src_ring->nentries_mask; 862 src_ring->sw_index &= src_ring->nentries_mask;
955 src_ring->hw_index = src_ring->sw_index; 863 src_ring->hw_index = src_ring->sw_index;
@@ -957,7 +865,6 @@ static int ath10k_ce_init_src_ring(struct ath10k *ar,
957 src_ring->write_index = 865 src_ring->write_index =
958 ath10k_ce_src_ring_write_index_get(ar, ctrl_addr); 866 ath10k_ce_src_ring_write_index_get(ar, ctrl_addr);
959 src_ring->write_index &= src_ring->nentries_mask; 867 src_ring->write_index &= src_ring->nentries_mask;
960 ath10k_pci_sleep(ar);
961 868
962 src_ring->per_transfer_context = (void **)ptr; 869 src_ring->per_transfer_context = (void **)ptr;
963 870
@@ -970,6 +877,12 @@ static int ath10k_ce_init_src_ring(struct ath10k *ar,
970 (nentries * sizeof(struct ce_desc) + 877 (nentries * sizeof(struct ce_desc) +
971 CE_DESC_RING_ALIGN), 878 CE_DESC_RING_ALIGN),
972 &base_addr); 879 &base_addr);
880 if (!src_ring->base_addr_owner_space_unaligned) {
881 kfree(ce_state->src_ring);
882 ce_state->src_ring = NULL;
883 return -ENOMEM;
884 }
885
973 src_ring->base_addr_ce_space_unaligned = base_addr; 886 src_ring->base_addr_ce_space_unaligned = base_addr;
974 887
975 src_ring->base_addr_owner_space = PTR_ALIGN( 888 src_ring->base_addr_owner_space = PTR_ALIGN(
@@ -986,12 +899,21 @@ static int ath10k_ce_init_src_ring(struct ath10k *ar,
986 src_ring->shadow_base_unaligned = 899 src_ring->shadow_base_unaligned =
987 kmalloc((nentries * sizeof(struct ce_desc) + 900 kmalloc((nentries * sizeof(struct ce_desc) +
988 CE_DESC_RING_ALIGN), GFP_KERNEL); 901 CE_DESC_RING_ALIGN), GFP_KERNEL);
902 if (!src_ring->shadow_base_unaligned) {
903 pci_free_consistent(ar_pci->pdev,
904 (nentries * sizeof(struct ce_desc) +
905 CE_DESC_RING_ALIGN),
906 src_ring->base_addr_owner_space,
907 src_ring->base_addr_ce_space);
908 kfree(ce_state->src_ring);
909 ce_state->src_ring = NULL;
910 return -ENOMEM;
911 }
989 912
990 src_ring->shadow_base = PTR_ALIGN( 913 src_ring->shadow_base = PTR_ALIGN(
991 src_ring->shadow_base_unaligned, 914 src_ring->shadow_base_unaligned,
992 CE_DESC_RING_ALIGN); 915 CE_DESC_RING_ALIGN);
993 916
994 ath10k_pci_wake(ar);
995 ath10k_ce_src_ring_base_addr_set(ar, ctrl_addr, 917 ath10k_ce_src_ring_base_addr_set(ar, ctrl_addr,
996 src_ring->base_addr_ce_space); 918 src_ring->base_addr_ce_space);
997 ath10k_ce_src_ring_size_set(ar, ctrl_addr, nentries); 919 ath10k_ce_src_ring_size_set(ar, ctrl_addr, nentries);
@@ -999,18 +921,21 @@ static int ath10k_ce_init_src_ring(struct ath10k *ar,
999 ath10k_ce_src_ring_byte_swap_set(ar, ctrl_addr, 0); 921 ath10k_ce_src_ring_byte_swap_set(ar, ctrl_addr, 0);
1000 ath10k_ce_src_ring_lowmark_set(ar, ctrl_addr, 0); 922 ath10k_ce_src_ring_lowmark_set(ar, ctrl_addr, 0);
1001 ath10k_ce_src_ring_highmark_set(ar, ctrl_addr, nentries); 923 ath10k_ce_src_ring_highmark_set(ar, ctrl_addr, nentries);
1002 ath10k_pci_sleep(ar); 924
925 ath10k_dbg(ATH10K_DBG_BOOT,
926 "boot ce src ring id %d entries %d base_addr %p\n",
927 ce_id, nentries, src_ring->base_addr_owner_space);
1003 928
1004 return 0; 929 return 0;
1005} 930}
1006 931
1007static int ath10k_ce_init_dest_ring(struct ath10k *ar, 932static int ath10k_ce_init_dest_ring(struct ath10k *ar,
1008 unsigned int ce_id, 933 unsigned int ce_id,
1009 struct ce_state *ce_state, 934 struct ath10k_ce_pipe *ce_state,
1010 const struct ce_attr *attr) 935 const struct ce_attr *attr)
1011{ 936{
1012 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 937 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
1013 struct ce_ring_state *dest_ring; 938 struct ath10k_ce_ring *dest_ring;
1014 unsigned int nentries = attr->dest_nentries; 939 unsigned int nentries = attr->dest_nentries;
1015 unsigned int ce_nbytes; 940 unsigned int ce_nbytes;
1016 u32 ctrl_addr = ath10k_ce_base_address(ce_id); 941 u32 ctrl_addr = ath10k_ce_base_address(ce_id);
@@ -1024,25 +949,23 @@ static int ath10k_ce_init_dest_ring(struct ath10k *ar,
1024 return 0; 949 return 0;
1025 } 950 }
1026 951
1027 ce_nbytes = sizeof(struct ce_ring_state) + (nentries * sizeof(void *)); 952 ce_nbytes = sizeof(struct ath10k_ce_ring) + (nentries * sizeof(void *));
1028 ptr = kzalloc(ce_nbytes, GFP_KERNEL); 953 ptr = kzalloc(ce_nbytes, GFP_KERNEL);
1029 if (ptr == NULL) 954 if (ptr == NULL)
1030 return -ENOMEM; 955 return -ENOMEM;
1031 956
1032 ce_state->dest_ring = (struct ce_ring_state *)ptr; 957 ce_state->dest_ring = (struct ath10k_ce_ring *)ptr;
1033 dest_ring = ce_state->dest_ring; 958 dest_ring = ce_state->dest_ring;
1034 959
1035 ptr += sizeof(struct ce_ring_state); 960 ptr += sizeof(struct ath10k_ce_ring);
1036 dest_ring->nentries = nentries; 961 dest_ring->nentries = nentries;
1037 dest_ring->nentries_mask = nentries - 1; 962 dest_ring->nentries_mask = nentries - 1;
1038 963
1039 ath10k_pci_wake(ar);
1040 dest_ring->sw_index = ath10k_ce_dest_ring_read_index_get(ar, ctrl_addr); 964 dest_ring->sw_index = ath10k_ce_dest_ring_read_index_get(ar, ctrl_addr);
1041 dest_ring->sw_index &= dest_ring->nentries_mask; 965 dest_ring->sw_index &= dest_ring->nentries_mask;
1042 dest_ring->write_index = 966 dest_ring->write_index =
1043 ath10k_ce_dest_ring_write_index_get(ar, ctrl_addr); 967 ath10k_ce_dest_ring_write_index_get(ar, ctrl_addr);
1044 dest_ring->write_index &= dest_ring->nentries_mask; 968 dest_ring->write_index &= dest_ring->nentries_mask;
1045 ath10k_pci_sleep(ar);
1046 969
1047 dest_ring->per_transfer_context = (void **)ptr; 970 dest_ring->per_transfer_context = (void **)ptr;
1048 971
@@ -1055,6 +978,12 @@ static int ath10k_ce_init_dest_ring(struct ath10k *ar,
1055 (nentries * sizeof(struct ce_desc) + 978 (nentries * sizeof(struct ce_desc) +
1056 CE_DESC_RING_ALIGN), 979 CE_DESC_RING_ALIGN),
1057 &base_addr); 980 &base_addr);
981 if (!dest_ring->base_addr_owner_space_unaligned) {
982 kfree(ce_state->dest_ring);
983 ce_state->dest_ring = NULL;
984 return -ENOMEM;
985 }
986
1058 dest_ring->base_addr_ce_space_unaligned = base_addr; 987 dest_ring->base_addr_ce_space_unaligned = base_addr;
1059 988
1060 /* 989 /*
@@ -1071,44 +1000,35 @@ static int ath10k_ce_init_dest_ring(struct ath10k *ar,
1071 dest_ring->base_addr_ce_space_unaligned, 1000 dest_ring->base_addr_ce_space_unaligned,
1072 CE_DESC_RING_ALIGN); 1001 CE_DESC_RING_ALIGN);
1073 1002
1074 ath10k_pci_wake(ar);
1075 ath10k_ce_dest_ring_base_addr_set(ar, ctrl_addr, 1003 ath10k_ce_dest_ring_base_addr_set(ar, ctrl_addr,
1076 dest_ring->base_addr_ce_space); 1004 dest_ring->base_addr_ce_space);
1077 ath10k_ce_dest_ring_size_set(ar, ctrl_addr, nentries); 1005 ath10k_ce_dest_ring_size_set(ar, ctrl_addr, nentries);
1078 ath10k_ce_dest_ring_byte_swap_set(ar, ctrl_addr, 0); 1006 ath10k_ce_dest_ring_byte_swap_set(ar, ctrl_addr, 0);
1079 ath10k_ce_dest_ring_lowmark_set(ar, ctrl_addr, 0); 1007 ath10k_ce_dest_ring_lowmark_set(ar, ctrl_addr, 0);
1080 ath10k_ce_dest_ring_highmark_set(ar, ctrl_addr, nentries); 1008 ath10k_ce_dest_ring_highmark_set(ar, ctrl_addr, nentries);
1081 ath10k_pci_sleep(ar); 1009
1010 ath10k_dbg(ATH10K_DBG_BOOT,
1011 "boot ce dest ring id %d entries %d base_addr %p\n",
1012 ce_id, nentries, dest_ring->base_addr_owner_space);
1082 1013
1083 return 0; 1014 return 0;
1084} 1015}
1085 1016
1086static struct ce_state *ath10k_ce_init_state(struct ath10k *ar, 1017static struct ath10k_ce_pipe *ath10k_ce_init_state(struct ath10k *ar,
1087 unsigned int ce_id, 1018 unsigned int ce_id,
1088 const struct ce_attr *attr) 1019 const struct ce_attr *attr)
1089{ 1020{
1090 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 1021 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
1091 struct ce_state *ce_state = NULL; 1022 struct ath10k_ce_pipe *ce_state = &ar_pci->ce_states[ce_id];
1092 u32 ctrl_addr = ath10k_ce_base_address(ce_id); 1023 u32 ctrl_addr = ath10k_ce_base_address(ce_id);
1093 1024
1094 spin_lock_bh(&ar_pci->ce_lock); 1025 spin_lock_bh(&ar_pci->ce_lock);
1095 1026
1096 if (!ar_pci->ce_id_to_state[ce_id]) { 1027 ce_state->ar = ar;
1097 ce_state = kzalloc(sizeof(*ce_state), GFP_ATOMIC); 1028 ce_state->id = ce_id;
1098 if (ce_state == NULL) { 1029 ce_state->ctrl_addr = ctrl_addr;
1099 spin_unlock_bh(&ar_pci->ce_lock); 1030 ce_state->attr_flags = attr->flags;
1100 return NULL; 1031 ce_state->src_sz_max = attr->src_sz_max;
1101 }
1102
1103 ar_pci->ce_id_to_state[ce_id] = ce_state;
1104 ce_state->ar = ar;
1105 ce_state->id = ce_id;
1106 ce_state->ctrl_addr = ctrl_addr;
1107 ce_state->state = CE_RUNNING;
1108 /* Save attribute flags */
1109 ce_state->attr_flags = attr->flags;
1110 ce_state->src_sz_max = attr->src_sz_max;
1111 }
1112 1032
1113 spin_unlock_bh(&ar_pci->ce_lock); 1033 spin_unlock_bh(&ar_pci->ce_lock);
1114 1034
@@ -1122,12 +1042,17 @@ static struct ce_state *ath10k_ce_init_state(struct ath10k *ar,
1122 * initialization. It may be that only one side or the other is 1042 * initialization. It may be that only one side or the other is
1123 * initialized by software/firmware. 1043 * initialized by software/firmware.
1124 */ 1044 */
1125struct ce_state *ath10k_ce_init(struct ath10k *ar, 1045struct ath10k_ce_pipe *ath10k_ce_init(struct ath10k *ar,
1126 unsigned int ce_id, 1046 unsigned int ce_id,
1127 const struct ce_attr *attr) 1047 const struct ce_attr *attr)
1128{ 1048{
1129 struct ce_state *ce_state; 1049 struct ath10k_ce_pipe *ce_state;
1130 u32 ctrl_addr = ath10k_ce_base_address(ce_id); 1050 u32 ctrl_addr = ath10k_ce_base_address(ce_id);
1051 int ret;
1052
1053 ret = ath10k_pci_wake(ar);
1054 if (ret)
1055 return NULL;
1131 1056
1132 ce_state = ath10k_ce_init_state(ar, ce_id, attr); 1057 ce_state = ath10k_ce_init_state(ar, ce_id, attr);
1133 if (!ce_state) { 1058 if (!ce_state) {
@@ -1136,40 +1061,38 @@ struct ce_state *ath10k_ce_init(struct ath10k *ar,
1136 } 1061 }
1137 1062
1138 if (attr->src_nentries) { 1063 if (attr->src_nentries) {
1139 if (ath10k_ce_init_src_ring(ar, ce_id, ce_state, attr)) { 1064 ret = ath10k_ce_init_src_ring(ar, ce_id, ce_state, attr);
1140 ath10k_err("Failed to initialize CE src ring for ID: %d\n", 1065 if (ret) {
1141 ce_id); 1066 ath10k_err("Failed to initialize CE src ring for ID: %d (%d)\n",
1067 ce_id, ret);
1142 ath10k_ce_deinit(ce_state); 1068 ath10k_ce_deinit(ce_state);
1143 return NULL; 1069 return NULL;
1144 } 1070 }
1145 } 1071 }
1146 1072
1147 if (attr->dest_nentries) { 1073 if (attr->dest_nentries) {
1148 if (ath10k_ce_init_dest_ring(ar, ce_id, ce_state, attr)) { 1074 ret = ath10k_ce_init_dest_ring(ar, ce_id, ce_state, attr);
1149 ath10k_err("Failed to initialize CE dest ring for ID: %d\n", 1075 if (ret) {
1150 ce_id); 1076 ath10k_err("Failed to initialize CE dest ring for ID: %d (%d)\n",
1077 ce_id, ret);
1151 ath10k_ce_deinit(ce_state); 1078 ath10k_ce_deinit(ce_state);
1152 return NULL; 1079 return NULL;
1153 } 1080 }
1154 } 1081 }
1155 1082
1156 /* Enable CE error interrupts */ 1083 /* Enable CE error interrupts */
1157 ath10k_pci_wake(ar);
1158 ath10k_ce_error_intr_enable(ar, ctrl_addr); 1084 ath10k_ce_error_intr_enable(ar, ctrl_addr);
1085
1159 ath10k_pci_sleep(ar); 1086 ath10k_pci_sleep(ar);
1160 1087
1161 return ce_state; 1088 return ce_state;
1162} 1089}
1163 1090
1164void ath10k_ce_deinit(struct ce_state *ce_state) 1091void ath10k_ce_deinit(struct ath10k_ce_pipe *ce_state)
1165{ 1092{
1166 unsigned int ce_id = ce_state->id;
1167 struct ath10k *ar = ce_state->ar; 1093 struct ath10k *ar = ce_state->ar;
1168 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 1094 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
1169 1095
1170 ce_state->state = CE_UNUSED;
1171 ar_pci->ce_id_to_state[ce_id] = NULL;
1172
1173 if (ce_state->src_ring) { 1096 if (ce_state->src_ring) {
1174 kfree(ce_state->src_ring->shadow_base_unaligned); 1097 kfree(ce_state->src_ring->shadow_base_unaligned);
1175 pci_free_consistent(ar_pci->pdev, 1098 pci_free_consistent(ar_pci->pdev,
@@ -1190,5 +1113,7 @@ void ath10k_ce_deinit(struct ce_state *ce_state)
1190 ce_state->dest_ring->base_addr_ce_space); 1113 ce_state->dest_ring->base_addr_ce_space);
1191 kfree(ce_state->dest_ring); 1114 kfree(ce_state->dest_ring);
1192 } 1115 }
1193 kfree(ce_state); 1116
1117 ce_state->src_ring = NULL;
1118 ce_state->dest_ring = NULL;
1194} 1119}
diff --git a/drivers/net/wireless/ath/ath10k/ce.h b/drivers/net/wireless/ath/ath10k/ce.h
index c17f07c026f4..15d45b5b7615 100644
--- a/drivers/net/wireless/ath/ath10k/ce.h
+++ b/drivers/net/wireless/ath/ath10k/ce.h
@@ -27,7 +27,6 @@
27 27
28/* Descriptor rings must be aligned to this boundary */ 28/* Descriptor rings must be aligned to this boundary */
29#define CE_DESC_RING_ALIGN 8 29#define CE_DESC_RING_ALIGN 8
30#define CE_SENDLIST_ITEMS_MAX 12
31#define CE_SEND_FLAG_GATHER 0x00010000 30#define CE_SEND_FLAG_GATHER 0x00010000
32 31
33/* 32/*
@@ -36,16 +35,9 @@
36 * how to use copy engines. 35 * how to use copy engines.
37 */ 36 */
38 37
39struct ce_state; 38struct ath10k_ce_pipe;
40 39
41 40
42/* Copy Engine operational state */
43enum ce_op_state {
44 CE_UNUSED,
45 CE_PAUSED,
46 CE_RUNNING,
47};
48
49#define CE_DESC_FLAGS_GATHER (1 << 0) 41#define CE_DESC_FLAGS_GATHER (1 << 0)
50#define CE_DESC_FLAGS_BYTE_SWAP (1 << 1) 42#define CE_DESC_FLAGS_BYTE_SWAP (1 << 1)
51#define CE_DESC_FLAGS_META_DATA_MASK 0xFFFC 43#define CE_DESC_FLAGS_META_DATA_MASK 0xFFFC
@@ -57,8 +49,7 @@ struct ce_desc {
57 __le16 flags; /* %CE_DESC_FLAGS_ */ 49 __le16 flags; /* %CE_DESC_FLAGS_ */
58}; 50};
59 51
60/* Copy Engine Ring internal state */ 52struct ath10k_ce_ring {
61struct ce_ring_state {
62 /* Number of entries in this ring; must be power of 2 */ 53 /* Number of entries in this ring; must be power of 2 */
63 unsigned int nentries; 54 unsigned int nentries;
64 unsigned int nentries_mask; 55 unsigned int nentries_mask;
@@ -116,49 +107,20 @@ struct ce_ring_state {
116 void **per_transfer_context; 107 void **per_transfer_context;
117}; 108};
118 109
119/* Copy Engine internal state */ 110struct ath10k_ce_pipe {
120struct ce_state {
121 struct ath10k *ar; 111 struct ath10k *ar;
122 unsigned int id; 112 unsigned int id;
123 113
124 unsigned int attr_flags; 114 unsigned int attr_flags;
125 115
126 u32 ctrl_addr; 116 u32 ctrl_addr;
127 enum ce_op_state state;
128
129 void (*send_cb) (struct ce_state *ce_state,
130 void *per_transfer_send_context,
131 u32 buffer,
132 unsigned int nbytes,
133 unsigned int transfer_id);
134 void (*recv_cb) (struct ce_state *ce_state,
135 void *per_transfer_recv_context,
136 u32 buffer,
137 unsigned int nbytes,
138 unsigned int transfer_id,
139 unsigned int flags);
140
141 unsigned int src_sz_max;
142 struct ce_ring_state *src_ring;
143 struct ce_ring_state *dest_ring;
144};
145 117
146struct ce_sendlist_item { 118 void (*send_cb)(struct ath10k_ce_pipe *);
147 /* e.g. buffer or desc list */ 119 void (*recv_cb)(struct ath10k_ce_pipe *);
148 dma_addr_t data;
149 union {
150 /* simple buffer */
151 unsigned int nbytes;
152 /* Rx descriptor list */
153 unsigned int ndesc;
154 } u;
155 /* externally-specified flags; OR-ed with internal flags */
156 u32 flags;
157};
158 120
159struct ce_sendlist { 121 unsigned int src_sz_max;
160 unsigned int num_items; 122 struct ath10k_ce_ring *src_ring;
161 struct ce_sendlist_item item[CE_SENDLIST_ITEMS_MAX]; 123 struct ath10k_ce_ring *dest_ring;
162}; 124};
163 125
164/* Copy Engine settable attributes */ 126/* Copy Engine settable attributes */
@@ -182,7 +144,7 @@ struct ce_attr;
182 * 144 *
183 * Implementation note: pushes 1 buffer to Source ring 145 * Implementation note: pushes 1 buffer to Source ring
184 */ 146 */
185int ath10k_ce_send(struct ce_state *ce_state, 147int ath10k_ce_send(struct ath10k_ce_pipe *ce_state,
186 void *per_transfer_send_context, 148 void *per_transfer_send_context,
187 u32 buffer, 149 u32 buffer,
188 unsigned int nbytes, 150 unsigned int nbytes,
@@ -190,36 +152,11 @@ int ath10k_ce_send(struct ce_state *ce_state,
190 unsigned int transfer_id, 152 unsigned int transfer_id,
191 unsigned int flags); 153 unsigned int flags);
192 154
193void ath10k_ce_send_cb_register(struct ce_state *ce_state, 155void ath10k_ce_send_cb_register(struct ath10k_ce_pipe *ce_state,
194 void (*send_cb) (struct ce_state *ce_state, 156 void (*send_cb)(struct ath10k_ce_pipe *),
195 void *transfer_context,
196 u32 buffer,
197 unsigned int nbytes,
198 unsigned int transfer_id),
199 int disable_interrupts); 157 int disable_interrupts);
200 158
201/* Append a simple buffer (address/length) to a sendlist. */ 159int ath10k_ce_num_free_src_entries(struct ath10k_ce_pipe *pipe);
202void ath10k_ce_sendlist_buf_add(struct ce_sendlist *sendlist,
203 u32 buffer,
204 unsigned int nbytes,
205 /* OR-ed with internal flags */
206 u32 flags);
207
208/*
209 * Queue a "sendlist" of buffers to be sent using gather to a single
210 * anonymous destination buffer
211 * ce - which copy engine to use
212 * sendlist - list of simple buffers to send using gather
213 * transfer_id - arbitrary ID; reflected to destination
214 * Returns 0 on success; otherwise an error status.
215 *
216 * Implemenation note: Pushes multiple buffers with Gather to Source ring.
217 */
218int ath10k_ce_sendlist_send(struct ce_state *ce_state,
219 void *per_transfer_send_context,
220 struct ce_sendlist *sendlist,
221 /* 14 bits */
222 unsigned int transfer_id);
223 160
224/*==================Recv=======================*/ 161/*==================Recv=======================*/
225 162
@@ -233,17 +170,12 @@ int ath10k_ce_sendlist_send(struct ce_state *ce_state,
233 * 170 *
234 * Implemenation note: Pushes a buffer to Dest ring. 171 * Implemenation note: Pushes a buffer to Dest ring.
235 */ 172 */
236int ath10k_ce_recv_buf_enqueue(struct ce_state *ce_state, 173int ath10k_ce_recv_buf_enqueue(struct ath10k_ce_pipe *ce_state,
237 void *per_transfer_recv_context, 174 void *per_transfer_recv_context,
238 u32 buffer); 175 u32 buffer);
239 176
240void ath10k_ce_recv_cb_register(struct ce_state *ce_state, 177void ath10k_ce_recv_cb_register(struct ath10k_ce_pipe *ce_state,
241 void (*recv_cb) (struct ce_state *ce_state, 178 void (*recv_cb)(struct ath10k_ce_pipe *));
242 void *transfer_context,
243 u32 buffer,
244 unsigned int nbytes,
245 unsigned int transfer_id,
246 unsigned int flags));
247 179
248/* recv flags */ 180/* recv flags */
249/* Data is byte-swapped */ 181/* Data is byte-swapped */
@@ -253,7 +185,7 @@ void ath10k_ce_recv_cb_register(struct ce_state *ce_state,
253 * Supply data for the next completed unprocessed receive descriptor. 185 * Supply data for the next completed unprocessed receive descriptor.
254 * Pops buffer from Dest ring. 186 * Pops buffer from Dest ring.
255 */ 187 */
256int ath10k_ce_completed_recv_next(struct ce_state *ce_state, 188int ath10k_ce_completed_recv_next(struct ath10k_ce_pipe *ce_state,
257 void **per_transfer_contextp, 189 void **per_transfer_contextp,
258 u32 *bufferp, 190 u32 *bufferp,
259 unsigned int *nbytesp, 191 unsigned int *nbytesp,
@@ -263,7 +195,7 @@ int ath10k_ce_completed_recv_next(struct ce_state *ce_state,
263 * Supply data for the next completed unprocessed send descriptor. 195 * Supply data for the next completed unprocessed send descriptor.
264 * Pops 1 completed send buffer from Source ring. 196 * Pops 1 completed send buffer from Source ring.
265 */ 197 */
266int ath10k_ce_completed_send_next(struct ce_state *ce_state, 198int ath10k_ce_completed_send_next(struct ath10k_ce_pipe *ce_state,
267 void **per_transfer_contextp, 199 void **per_transfer_contextp,
268 u32 *bufferp, 200 u32 *bufferp,
269 unsigned int *nbytesp, 201 unsigned int *nbytesp,
@@ -272,7 +204,7 @@ int ath10k_ce_completed_send_next(struct ce_state *ce_state,
272/*==================CE Engine Initialization=======================*/ 204/*==================CE Engine Initialization=======================*/
273 205
274/* Initialize an instance of a CE */ 206/* Initialize an instance of a CE */
275struct ce_state *ath10k_ce_init(struct ath10k *ar, 207struct ath10k_ce_pipe *ath10k_ce_init(struct ath10k *ar,
276 unsigned int ce_id, 208 unsigned int ce_id,
277 const struct ce_attr *attr); 209 const struct ce_attr *attr);
278 210
@@ -282,7 +214,7 @@ struct ce_state *ath10k_ce_init(struct ath10k *ar,
282 * receive buffers. Target DMA must be stopped before using 214 * receive buffers. Target DMA must be stopped before using
283 * this API. 215 * this API.
284 */ 216 */
285int ath10k_ce_revoke_recv_next(struct ce_state *ce_state, 217int ath10k_ce_revoke_recv_next(struct ath10k_ce_pipe *ce_state,
286 void **per_transfer_contextp, 218 void **per_transfer_contextp,
287 u32 *bufferp); 219 u32 *bufferp);
288 220
@@ -291,13 +223,13 @@ int ath10k_ce_revoke_recv_next(struct ce_state *ce_state,
291 * pending sends. Target DMA must be stopped before using 223 * pending sends. Target DMA must be stopped before using
292 * this API. 224 * this API.
293 */ 225 */
294int ath10k_ce_cancel_send_next(struct ce_state *ce_state, 226int ath10k_ce_cancel_send_next(struct ath10k_ce_pipe *ce_state,
295 void **per_transfer_contextp, 227 void **per_transfer_contextp,
296 u32 *bufferp, 228 u32 *bufferp,
297 unsigned int *nbytesp, 229 unsigned int *nbytesp,
298 unsigned int *transfer_idp); 230 unsigned int *transfer_idp);
299 231
300void ath10k_ce_deinit(struct ce_state *ce_state); 232void ath10k_ce_deinit(struct ath10k_ce_pipe *ce_state);
301 233
302/*==================CE Interrupt Handlers====================*/ 234/*==================CE Interrupt Handlers====================*/
303void ath10k_ce_per_engine_service_any(struct ath10k *ar); 235void ath10k_ce_per_engine_service_any(struct ath10k *ar);
@@ -322,9 +254,6 @@ struct ce_attr {
322 /* CE_ATTR_* values */ 254 /* CE_ATTR_* values */
323 unsigned int flags; 255 unsigned int flags;
324 256
325 /* currently not in use */
326 unsigned int priority;
327
328 /* #entries in source ring - Must be a power of 2 */ 257 /* #entries in source ring - Must be a power of 2 */
329 unsigned int src_nentries; 258 unsigned int src_nentries;
330 259
@@ -336,21 +265,8 @@ struct ce_attr {
336 265
337 /* #entries in destination ring - Must be a power of 2 */ 266 /* #entries in destination ring - Must be a power of 2 */
338 unsigned int dest_nentries; 267 unsigned int dest_nentries;
339
340 /* Future use */
341 void *reserved;
342}; 268};
343 269
344/*
345 * When using sendlist_send to transfer multiple buffer fragments, the
346 * transfer context of each fragment, except last one, will be filled
347 * with CE_SENDLIST_ITEM_CTXT. ce_completed_send will return success for
348 * each fragment done with send and the transfer context would be
349 * CE_SENDLIST_ITEM_CTXT. Upper layer could use this to identify the
350 * status of a send completion.
351 */
352#define CE_SENDLIST_ITEM_CTXT ((void *)0xcecebeef)
353
354#define SR_BA_ADDRESS 0x0000 270#define SR_BA_ADDRESS 0x0000
355#define SR_SIZE_ADDRESS 0x0004 271#define SR_SIZE_ADDRESS 0x0004
356#define DR_BA_ADDRESS 0x0008 272#define DR_BA_ADDRESS 0x0008
diff --git a/drivers/net/wireless/ath/ath10k/core.c b/drivers/net/wireless/ath/ath10k/core.c
index 7226c23b9569..1129994fb105 100644
--- a/drivers/net/wireless/ath/ath10k/core.c
+++ b/drivers/net/wireless/ath/ath10k/core.c
@@ -39,17 +39,6 @@ MODULE_PARM_DESC(p2p, "Enable ath10k P2P support");
39 39
40static const struct ath10k_hw_params ath10k_hw_params_list[] = { 40static const struct ath10k_hw_params ath10k_hw_params_list[] = {
41 { 41 {
42 .id = QCA988X_HW_1_0_VERSION,
43 .name = "qca988x hw1.0",
44 .patch_load_addr = QCA988X_HW_1_0_PATCH_LOAD_ADDR,
45 .fw = {
46 .dir = QCA988X_HW_1_0_FW_DIR,
47 .fw = QCA988X_HW_1_0_FW_FILE,
48 .otp = QCA988X_HW_1_0_OTP_FILE,
49 .board = QCA988X_HW_1_0_BOARD_DATA_FILE,
50 },
51 },
52 {
53 .id = QCA988X_HW_2_0_VERSION, 42 .id = QCA988X_HW_2_0_VERSION,
54 .name = "qca988x hw2.0", 43 .name = "qca988x hw2.0",
55 .patch_load_addr = QCA988X_HW_2_0_PATCH_LOAD_ADDR, 44 .patch_load_addr = QCA988X_HW_2_0_PATCH_LOAD_ADDR,
@@ -64,33 +53,12 @@ static const struct ath10k_hw_params ath10k_hw_params_list[] = {
64 53
65static void ath10k_send_suspend_complete(struct ath10k *ar) 54static void ath10k_send_suspend_complete(struct ath10k *ar)
66{ 55{
67 ath10k_dbg(ATH10K_DBG_CORE, "%s\n", __func__); 56 ath10k_dbg(ATH10K_DBG_BOOT, "boot suspend complete\n");
68 57
69 ar->is_target_paused = true; 58 ar->is_target_paused = true;
70 wake_up(&ar->event_queue); 59 wake_up(&ar->event_queue);
71} 60}
72 61
73static int ath10k_check_fw_version(struct ath10k *ar)
74{
75 char version[32];
76
77 if (ar->fw_version_major >= SUPPORTED_FW_MAJOR &&
78 ar->fw_version_minor >= SUPPORTED_FW_MINOR &&
79 ar->fw_version_release >= SUPPORTED_FW_RELEASE &&
80 ar->fw_version_build >= SUPPORTED_FW_BUILD)
81 return 0;
82
83 snprintf(version, sizeof(version), "%u.%u.%u.%u",
84 SUPPORTED_FW_MAJOR, SUPPORTED_FW_MINOR,
85 SUPPORTED_FW_RELEASE, SUPPORTED_FW_BUILD);
86
87 ath10k_warn("WARNING: Firmware version %s is not officially supported.\n",
88 ar->hw->wiphy->fw_version);
89 ath10k_warn("Please upgrade to version %s (or newer)\n", version);
90
91 return 0;
92}
93
94static int ath10k_init_connect_htc(struct ath10k *ar) 62static int ath10k_init_connect_htc(struct ath10k *ar)
95{ 63{
96 int status; 64 int status;
@@ -112,7 +80,7 @@ static int ath10k_init_connect_htc(struct ath10k *ar)
112 goto timeout; 80 goto timeout;
113 } 81 }
114 82
115 ath10k_dbg(ATH10K_DBG_CORE, "core wmi ready\n"); 83 ath10k_dbg(ATH10K_DBG_BOOT, "boot wmi ready\n");
116 return 0; 84 return 0;
117 85
118timeout: 86timeout:
@@ -200,8 +168,7 @@ static const struct firmware *ath10k_fetch_fw_file(struct ath10k *ar,
200 return fw; 168 return fw;
201} 169}
202 170
203static int ath10k_push_board_ext_data(struct ath10k *ar, 171static int ath10k_push_board_ext_data(struct ath10k *ar)
204 const struct firmware *fw)
205{ 172{
206 u32 board_data_size = QCA988X_BOARD_DATA_SZ; 173 u32 board_data_size = QCA988X_BOARD_DATA_SZ;
207 u32 board_ext_data_size = QCA988X_BOARD_EXT_DATA_SZ; 174 u32 board_ext_data_size = QCA988X_BOARD_EXT_DATA_SZ;
@@ -214,21 +181,21 @@ static int ath10k_push_board_ext_data(struct ath10k *ar,
214 return ret; 181 return ret;
215 } 182 }
216 183
217 ath10k_dbg(ATH10K_DBG_CORE, 184 ath10k_dbg(ATH10K_DBG_BOOT,
218 "ath10k: Board extended Data download addr: 0x%x\n", 185 "boot push board extended data addr 0x%x\n",
219 board_ext_data_addr); 186 board_ext_data_addr);
220 187
221 if (board_ext_data_addr == 0) 188 if (board_ext_data_addr == 0)
222 return 0; 189 return 0;
223 190
224 if (fw->size != (board_data_size + board_ext_data_size)) { 191 if (ar->board_len != (board_data_size + board_ext_data_size)) {
225 ath10k_err("invalid board (ext) data sizes %zu != %d+%d\n", 192 ath10k_err("invalid board (ext) data sizes %zu != %d+%d\n",
226 fw->size, board_data_size, board_ext_data_size); 193 ar->board_len, board_data_size, board_ext_data_size);
227 return -EINVAL; 194 return -EINVAL;
228 } 195 }
229 196
230 ret = ath10k_bmi_write_memory(ar, board_ext_data_addr, 197 ret = ath10k_bmi_write_memory(ar, board_ext_data_addr,
231 fw->data + board_data_size, 198 ar->board_data + board_data_size,
232 board_ext_data_size); 199 board_ext_data_size);
233 if (ret) { 200 if (ret) {
234 ath10k_err("could not write board ext data (%d)\n", ret); 201 ath10k_err("could not write board ext data (%d)\n", ret);
@@ -247,12 +214,11 @@ static int ath10k_push_board_ext_data(struct ath10k *ar,
247 214
248static int ath10k_download_board_data(struct ath10k *ar) 215static int ath10k_download_board_data(struct ath10k *ar)
249{ 216{
250 const struct firmware *fw = ar->board_data;
251 u32 board_data_size = QCA988X_BOARD_DATA_SZ; 217 u32 board_data_size = QCA988X_BOARD_DATA_SZ;
252 u32 address; 218 u32 address;
253 int ret; 219 int ret;
254 220
255 ret = ath10k_push_board_ext_data(ar, fw); 221 ret = ath10k_push_board_ext_data(ar);
256 if (ret) { 222 if (ret) {
257 ath10k_err("could not push board ext data (%d)\n", ret); 223 ath10k_err("could not push board ext data (%d)\n", ret);
258 goto exit; 224 goto exit;
@@ -264,8 +230,9 @@ static int ath10k_download_board_data(struct ath10k *ar)
264 goto exit; 230 goto exit;
265 } 231 }
266 232
267 ret = ath10k_bmi_write_memory(ar, address, fw->data, 233 ret = ath10k_bmi_write_memory(ar, address, ar->board_data,
268 min_t(u32, board_data_size, fw->size)); 234 min_t(u32, board_data_size,
235 ar->board_len));
269 if (ret) { 236 if (ret) {
270 ath10k_err("could not write board data (%d)\n", ret); 237 ath10k_err("could not write board data (%d)\n", ret);
271 goto exit; 238 goto exit;
@@ -283,17 +250,16 @@ exit:
283 250
284static int ath10k_download_and_run_otp(struct ath10k *ar) 251static int ath10k_download_and_run_otp(struct ath10k *ar)
285{ 252{
286 const struct firmware *fw = ar->otp;
287 u32 address = ar->hw_params.patch_load_addr; 253 u32 address = ar->hw_params.patch_load_addr;
288 u32 exec_param; 254 u32 exec_param;
289 int ret; 255 int ret;
290 256
291 /* OTP is optional */ 257 /* OTP is optional */
292 258
293 if (!ar->otp) 259 if (!ar->otp_data || !ar->otp_len)
294 return 0; 260 return 0;
295 261
296 ret = ath10k_bmi_fast_download(ar, address, fw->data, fw->size); 262 ret = ath10k_bmi_fast_download(ar, address, ar->otp_data, ar->otp_len);
297 if (ret) { 263 if (ret) {
298 ath10k_err("could not write otp (%d)\n", ret); 264 ath10k_err("could not write otp (%d)\n", ret);
299 goto exit; 265 goto exit;
@@ -312,13 +278,13 @@ exit:
312 278
313static int ath10k_download_fw(struct ath10k *ar) 279static int ath10k_download_fw(struct ath10k *ar)
314{ 280{
315 const struct firmware *fw = ar->firmware;
316 u32 address; 281 u32 address;
317 int ret; 282 int ret;
318 283
319 address = ar->hw_params.patch_load_addr; 284 address = ar->hw_params.patch_load_addr;
320 285
321 ret = ath10k_bmi_fast_download(ar, address, fw->data, fw->size); 286 ret = ath10k_bmi_fast_download(ar, address, ar->firmware_data,
287 ar->firmware_len);
322 if (ret) { 288 if (ret) {
323 ath10k_err("could not write fw (%d)\n", ret); 289 ath10k_err("could not write fw (%d)\n", ret);
324 goto exit; 290 goto exit;
@@ -330,8 +296,8 @@ exit:
330 296
331static void ath10k_core_free_firmware_files(struct ath10k *ar) 297static void ath10k_core_free_firmware_files(struct ath10k *ar)
332{ 298{
333 if (ar->board_data && !IS_ERR(ar->board_data)) 299 if (ar->board && !IS_ERR(ar->board))
334 release_firmware(ar->board_data); 300 release_firmware(ar->board);
335 301
336 if (ar->otp && !IS_ERR(ar->otp)) 302 if (ar->otp && !IS_ERR(ar->otp))
337 release_firmware(ar->otp); 303 release_firmware(ar->otp);
@@ -339,12 +305,20 @@ static void ath10k_core_free_firmware_files(struct ath10k *ar)
339 if (ar->firmware && !IS_ERR(ar->firmware)) 305 if (ar->firmware && !IS_ERR(ar->firmware))
340 release_firmware(ar->firmware); 306 release_firmware(ar->firmware);
341 307
308 ar->board = NULL;
342 ar->board_data = NULL; 309 ar->board_data = NULL;
310 ar->board_len = 0;
311
343 ar->otp = NULL; 312 ar->otp = NULL;
313 ar->otp_data = NULL;
314 ar->otp_len = 0;
315
344 ar->firmware = NULL; 316 ar->firmware = NULL;
317 ar->firmware_data = NULL;
318 ar->firmware_len = 0;
345} 319}
346 320
347static int ath10k_core_fetch_firmware_files(struct ath10k *ar) 321static int ath10k_core_fetch_firmware_api_1(struct ath10k *ar)
348{ 322{
349 int ret = 0; 323 int ret = 0;
350 324
@@ -358,15 +332,18 @@ static int ath10k_core_fetch_firmware_files(struct ath10k *ar)
358 return -EINVAL; 332 return -EINVAL;
359 } 333 }
360 334
361 ar->board_data = ath10k_fetch_fw_file(ar, 335 ar->board = ath10k_fetch_fw_file(ar,
362 ar->hw_params.fw.dir, 336 ar->hw_params.fw.dir,
363 ar->hw_params.fw.board); 337 ar->hw_params.fw.board);
364 if (IS_ERR(ar->board_data)) { 338 if (IS_ERR(ar->board)) {
365 ret = PTR_ERR(ar->board_data); 339 ret = PTR_ERR(ar->board);
366 ath10k_err("could not fetch board data (%d)\n", ret); 340 ath10k_err("could not fetch board data (%d)\n", ret);
367 goto err; 341 goto err;
368 } 342 }
369 343
344 ar->board_data = ar->board->data;
345 ar->board_len = ar->board->size;
346
370 ar->firmware = ath10k_fetch_fw_file(ar, 347 ar->firmware = ath10k_fetch_fw_file(ar,
371 ar->hw_params.fw.dir, 348 ar->hw_params.fw.dir,
372 ar->hw_params.fw.fw); 349 ar->hw_params.fw.fw);
@@ -376,6 +353,9 @@ static int ath10k_core_fetch_firmware_files(struct ath10k *ar)
376 goto err; 353 goto err;
377 } 354 }
378 355
356 ar->firmware_data = ar->firmware->data;
357 ar->firmware_len = ar->firmware->size;
358
379 /* OTP may be undefined. If so, don't fetch it at all */ 359 /* OTP may be undefined. If so, don't fetch it at all */
380 if (ar->hw_params.fw.otp == NULL) 360 if (ar->hw_params.fw.otp == NULL)
381 return 0; 361 return 0;
@@ -389,6 +369,172 @@ static int ath10k_core_fetch_firmware_files(struct ath10k *ar)
389 goto err; 369 goto err;
390 } 370 }
391 371
372 ar->otp_data = ar->otp->data;
373 ar->otp_len = ar->otp->size;
374
375 return 0;
376
377err:
378 ath10k_core_free_firmware_files(ar);
379 return ret;
380}
381
382static int ath10k_core_fetch_firmware_api_n(struct ath10k *ar, const char *name)
383{
384 size_t magic_len, len, ie_len;
385 int ie_id, i, index, bit, ret;
386 struct ath10k_fw_ie *hdr;
387 const u8 *data;
388 __le32 *timestamp;
389
390 /* first fetch the firmware file (firmware-*.bin) */
391 ar->firmware = ath10k_fetch_fw_file(ar, ar->hw_params.fw.dir, name);
392 if (IS_ERR(ar->firmware)) {
393 ath10k_err("Could not fetch firmware file '%s': %ld\n",
394 name, PTR_ERR(ar->firmware));
395 return PTR_ERR(ar->firmware);
396 }
397
398 data = ar->firmware->data;
399 len = ar->firmware->size;
400
401 /* magic also includes the null byte, check that as well */
402 magic_len = strlen(ATH10K_FIRMWARE_MAGIC) + 1;
403
404 if (len < magic_len) {
405 ath10k_err("firmware image too small to contain magic: %zu\n",
406 len);
407 ret = -EINVAL;
408 goto err;
409 }
410
411 if (memcmp(data, ATH10K_FIRMWARE_MAGIC, magic_len) != 0) {
412 ath10k_err("Invalid firmware magic\n");
413 ret = -EINVAL;
414 goto err;
415 }
416
417 /* jump over the padding */
418 magic_len = ALIGN(magic_len, 4);
419
420 len -= magic_len;
421 data += magic_len;
422
423 /* loop elements */
424 while (len > sizeof(struct ath10k_fw_ie)) {
425 hdr = (struct ath10k_fw_ie *)data;
426
427 ie_id = le32_to_cpu(hdr->id);
428 ie_len = le32_to_cpu(hdr->len);
429
430 len -= sizeof(*hdr);
431 data += sizeof(*hdr);
432
433 if (len < ie_len) {
434 ath10k_err("Invalid length for FW IE %d (%zu < %zu)\n",
435 ie_id, len, ie_len);
436 ret = -EINVAL;
437 goto err;
438 }
439
440 switch (ie_id) {
441 case ATH10K_FW_IE_FW_VERSION:
442 if (ie_len > sizeof(ar->hw->wiphy->fw_version) - 1)
443 break;
444
445 memcpy(ar->hw->wiphy->fw_version, data, ie_len);
446 ar->hw->wiphy->fw_version[ie_len] = '\0';
447
448 ath10k_dbg(ATH10K_DBG_BOOT,
449 "found fw version %s\n",
450 ar->hw->wiphy->fw_version);
451 break;
452 case ATH10K_FW_IE_TIMESTAMP:
453 if (ie_len != sizeof(u32))
454 break;
455
456 timestamp = (__le32 *)data;
457
458 ath10k_dbg(ATH10K_DBG_BOOT, "found fw timestamp %d\n",
459 le32_to_cpup(timestamp));
460 break;
461 case ATH10K_FW_IE_FEATURES:
462 ath10k_dbg(ATH10K_DBG_BOOT,
463 "found firmware features ie (%zd B)\n",
464 ie_len);
465
466 for (i = 0; i < ATH10K_FW_FEATURE_COUNT; i++) {
467 index = i / 8;
468 bit = i % 8;
469
470 if (index == ie_len)
471 break;
472
473 if (data[index] & (1 << bit))
474 __set_bit(i, ar->fw_features);
475 }
476
477 ath10k_dbg_dump(ATH10K_DBG_BOOT, "features", "",
478 ar->fw_features,
479 sizeof(ar->fw_features));
480 break;
481 case ATH10K_FW_IE_FW_IMAGE:
482 ath10k_dbg(ATH10K_DBG_BOOT,
483 "found fw image ie (%zd B)\n",
484 ie_len);
485
486 ar->firmware_data = data;
487 ar->firmware_len = ie_len;
488
489 break;
490 case ATH10K_FW_IE_OTP_IMAGE:
491 ath10k_dbg(ATH10K_DBG_BOOT,
492 "found otp image ie (%zd B)\n",
493 ie_len);
494
495 ar->otp_data = data;
496 ar->otp_len = ie_len;
497
498 break;
499 default:
500 ath10k_warn("Unknown FW IE: %u\n",
501 le32_to_cpu(hdr->id));
502 break;
503 }
504
505 /* jump over the padding */
506 ie_len = ALIGN(ie_len, 4);
507
508 len -= ie_len;
509 data += ie_len;
510 }
511
512 if (!ar->firmware_data || !ar->firmware_len) {
513 ath10k_warn("No ATH10K_FW_IE_FW_IMAGE found from %s, skipping\n",
514 name);
515 ret = -ENOMEDIUM;
516 goto err;
517 }
518
519 /* now fetch the board file */
520 if (ar->hw_params.fw.board == NULL) {
521 ath10k_err("board data file not defined");
522 ret = -EINVAL;
523 goto err;
524 }
525
526 ar->board = ath10k_fetch_fw_file(ar,
527 ar->hw_params.fw.dir,
528 ar->hw_params.fw.board);
529 if (IS_ERR(ar->board)) {
530 ret = PTR_ERR(ar->board);
531 ath10k_err("could not fetch board data (%d)\n", ret);
532 goto err;
533 }
534
535 ar->board_data = ar->board->data;
536 ar->board_len = ar->board->size;
537
392 return 0; 538 return 0;
393 539
394err: 540err:
@@ -396,6 +542,28 @@ err:
396 return ret; 542 return ret;
397} 543}
398 544
545static int ath10k_core_fetch_firmware_files(struct ath10k *ar)
546{
547 int ret;
548
549 ret = ath10k_core_fetch_firmware_api_n(ar, ATH10K_FW_API2_FILE);
550 if (ret == 0) {
551 ar->fw_api = 2;
552 goto out;
553 }
554
555 ret = ath10k_core_fetch_firmware_api_1(ar);
556 if (ret)
557 return ret;
558
559 ar->fw_api = 1;
560
561out:
562 ath10k_dbg(ATH10K_DBG_BOOT, "using fw api %d\n", ar->fw_api);
563
564 return 0;
565}
566
399static int ath10k_init_download_firmware(struct ath10k *ar) 567static int ath10k_init_download_firmware(struct ath10k *ar)
400{ 568{
401 int ret; 569 int ret;
@@ -446,6 +614,13 @@ static int ath10k_init_uart(struct ath10k *ar)
446 return ret; 614 return ret;
447 } 615 }
448 616
617 /* Set the UART baud rate to 19200. */
618 ret = ath10k_bmi_write32(ar, hi_desired_baud_rate, 19200);
619 if (ret) {
620 ath10k_warn("could not set the baud rate (%d)\n", ret);
621 return ret;
622 }
623
449 ath10k_info("UART prints enabled\n"); 624 ath10k_info("UART prints enabled\n");
450 return 0; 625 return 0;
451} 626}
@@ -545,6 +720,9 @@ struct ath10k *ath10k_core_create(void *hif_priv, struct device *dev,
545 INIT_WORK(&ar->offchan_tx_work, ath10k_offchan_tx_work); 720 INIT_WORK(&ar->offchan_tx_work, ath10k_offchan_tx_work);
546 skb_queue_head_init(&ar->offchan_tx_queue); 721 skb_queue_head_init(&ar->offchan_tx_queue);
547 722
723 INIT_WORK(&ar->wmi_mgmt_tx_work, ath10k_mgmt_over_wmi_tx_work);
724 skb_queue_head_init(&ar->wmi_mgmt_tx_queue);
725
548 init_waitqueue_head(&ar->event_queue); 726 init_waitqueue_head(&ar->event_queue);
549 727
550 INIT_WORK(&ar->restart_work, ath10k_core_restart); 728 INIT_WORK(&ar->restart_work, ath10k_core_restart);
@@ -559,6 +737,8 @@ EXPORT_SYMBOL(ath10k_core_create);
559 737
560void ath10k_core_destroy(struct ath10k *ar) 738void ath10k_core_destroy(struct ath10k *ar)
561{ 739{
740 ath10k_debug_destroy(ar);
741
562 flush_workqueue(ar->workqueue); 742 flush_workqueue(ar->workqueue);
563 destroy_workqueue(ar->workqueue); 743 destroy_workqueue(ar->workqueue);
564 744
@@ -570,6 +750,8 @@ int ath10k_core_start(struct ath10k *ar)
570{ 750{
571 int status; 751 int status;
572 752
753 lockdep_assert_held(&ar->conf_mutex);
754
573 ath10k_bmi_start(ar); 755 ath10k_bmi_start(ar);
574 756
575 if (ath10k_init_configure_target(ar)) { 757 if (ath10k_init_configure_target(ar)) {
@@ -620,10 +802,6 @@ int ath10k_core_start(struct ath10k *ar)
620 802
621 ath10k_info("firmware %s booted\n", ar->hw->wiphy->fw_version); 803 ath10k_info("firmware %s booted\n", ar->hw->wiphy->fw_version);
622 804
623 status = ath10k_check_fw_version(ar);
624 if (status)
625 goto err_disconnect_htc;
626
627 status = ath10k_wmi_cmd_init(ar); 805 status = ath10k_wmi_cmd_init(ar);
628 if (status) { 806 if (status) {
629 ath10k_err("could not send WMI init command (%d)\n", status); 807 ath10k_err("could not send WMI init command (%d)\n", status);
@@ -641,7 +819,12 @@ int ath10k_core_start(struct ath10k *ar)
641 if (status) 819 if (status)
642 goto err_disconnect_htc; 820 goto err_disconnect_htc;
643 821
822 status = ath10k_debug_start(ar);
823 if (status)
824 goto err_disconnect_htc;
825
644 ar->free_vdev_map = (1 << TARGET_NUM_VDEVS) - 1; 826 ar->free_vdev_map = (1 << TARGET_NUM_VDEVS) - 1;
827 INIT_LIST_HEAD(&ar->arvifs);
645 828
646 return 0; 829 return 0;
647 830
@@ -658,6 +841,9 @@ EXPORT_SYMBOL(ath10k_core_start);
658 841
659void ath10k_core_stop(struct ath10k *ar) 842void ath10k_core_stop(struct ath10k *ar)
660{ 843{
844 lockdep_assert_held(&ar->conf_mutex);
845
846 ath10k_debug_stop(ar);
661 ath10k_htc_stop(&ar->htc); 847 ath10k_htc_stop(&ar->htc);
662 ath10k_htt_detach(&ar->htt); 848 ath10k_htt_detach(&ar->htt);
663 ath10k_wmi_detach(ar); 849 ath10k_wmi_detach(ar);
@@ -704,23 +890,65 @@ static int ath10k_core_probe_fw(struct ath10k *ar)
704 return ret; 890 return ret;
705 } 891 }
706 892
893 mutex_lock(&ar->conf_mutex);
894
707 ret = ath10k_core_start(ar); 895 ret = ath10k_core_start(ar);
708 if (ret) { 896 if (ret) {
709 ath10k_err("could not init core (%d)\n", ret); 897 ath10k_err("could not init core (%d)\n", ret);
710 ath10k_core_free_firmware_files(ar); 898 ath10k_core_free_firmware_files(ar);
711 ath10k_hif_power_down(ar); 899 ath10k_hif_power_down(ar);
900 mutex_unlock(&ar->conf_mutex);
712 return ret; 901 return ret;
713 } 902 }
714 903
715 ath10k_core_stop(ar); 904 ath10k_core_stop(ar);
905
906 mutex_unlock(&ar->conf_mutex);
907
716 ath10k_hif_power_down(ar); 908 ath10k_hif_power_down(ar);
717 return 0; 909 return 0;
718} 910}
719 911
720int ath10k_core_register(struct ath10k *ar) 912static int ath10k_core_check_chip_id(struct ath10k *ar)
913{
914 u32 hw_revision = MS(ar->chip_id, SOC_CHIP_ID_REV);
915
916 ath10k_dbg(ATH10K_DBG_BOOT, "boot chip_id 0x%08x hw_revision 0x%x\n",
917 ar->chip_id, hw_revision);
918
919 /* Check that we are not using hw1.0 (some of them have same pci id
920 * as hw2.0) before doing anything else as ath10k crashes horribly
921 * due to missing hw1.0 workarounds. */
922 switch (hw_revision) {
923 case QCA988X_HW_1_0_CHIP_ID_REV:
924 ath10k_err("ERROR: qca988x hw1.0 is not supported\n");
925 return -EOPNOTSUPP;
926
927 case QCA988X_HW_2_0_CHIP_ID_REV:
928 /* known hardware revision, continue normally */
929 return 0;
930
931 default:
932 ath10k_warn("Warning: hardware revision unknown (0x%x), expect problems\n",
933 ar->chip_id);
934 return 0;
935 }
936
937 return 0;
938}
939
940int ath10k_core_register(struct ath10k *ar, u32 chip_id)
721{ 941{
722 int status; 942 int status;
723 943
944 ar->chip_id = chip_id;
945
946 status = ath10k_core_check_chip_id(ar);
947 if (status) {
948 ath10k_err("Unsupported chip id 0x%08x\n", ar->chip_id);
949 return status;
950 }
951
724 status = ath10k_core_probe_fw(ar); 952 status = ath10k_core_probe_fw(ar);
725 if (status) { 953 if (status) {
726 ath10k_err("could not probe fw (%d)\n", status); 954 ath10k_err("could not probe fw (%d)\n", status);
@@ -755,6 +983,7 @@ void ath10k_core_unregister(struct ath10k *ar)
755 * Otherwise we will fail to submit commands to FW and mac80211 will be 983 * Otherwise we will fail to submit commands to FW and mac80211 will be
756 * unhappy about callback failures. */ 984 * unhappy about callback failures. */
757 ath10k_mac_unregister(ar); 985 ath10k_mac_unregister(ar);
986
758 ath10k_core_free_firmware_files(ar); 987 ath10k_core_free_firmware_files(ar);
759} 988}
760EXPORT_SYMBOL(ath10k_core_unregister); 989EXPORT_SYMBOL(ath10k_core_unregister);
diff --git a/drivers/net/wireless/ath/ath10k/core.h b/drivers/net/wireless/ath/ath10k/core.h
index e4bba563ed42..0934f7633de3 100644
--- a/drivers/net/wireless/ath/ath10k/core.h
+++ b/drivers/net/wireless/ath/ath10k/core.h
@@ -43,27 +43,23 @@
43/* Antenna noise floor */ 43/* Antenna noise floor */
44#define ATH10K_DEFAULT_NOISE_FLOOR -95 44#define ATH10K_DEFAULT_NOISE_FLOOR -95
45 45
46#define ATH10K_MAX_NUM_MGMT_PENDING 16
47
46struct ath10k; 48struct ath10k;
47 49
48struct ath10k_skb_cb { 50struct ath10k_skb_cb {
49 dma_addr_t paddr; 51 dma_addr_t paddr;
50 bool is_mapped; 52 bool is_mapped;
51 bool is_aborted; 53 bool is_aborted;
54 u8 vdev_id;
52 55
53 struct { 56 struct {
54 u8 vdev_id;
55 u16 msdu_id;
56 u8 tid; 57 u8 tid;
57 bool is_offchan; 58 bool is_offchan;
58 bool is_conf;
59 bool discard;
60 bool no_ack;
61 u8 refcount;
62 struct sk_buff *txfrag;
63 struct sk_buff *msdu;
64 } __packed htt;
65 59
66 /* 4 bytes left on 64bit arch */ 60 u8 frag_len;
61 u8 pad_len;
62 } __packed htt;
67} __packed; 63} __packed;
68 64
69static inline struct ath10k_skb_cb *ATH10K_SKB_CB(struct sk_buff *skb) 65static inline struct ath10k_skb_cb *ATH10K_SKB_CB(struct sk_buff *skb)
@@ -108,15 +104,26 @@ struct ath10k_bmi {
108 bool done_sent; 104 bool done_sent;
109}; 105};
110 106
107#define ATH10K_MAX_MEM_REQS 16
108
109struct ath10k_mem_chunk {
110 void *vaddr;
111 dma_addr_t paddr;
112 u32 len;
113 u32 req_id;
114};
115
111struct ath10k_wmi { 116struct ath10k_wmi {
112 enum ath10k_htc_ep_id eid; 117 enum ath10k_htc_ep_id eid;
113 struct completion service_ready; 118 struct completion service_ready;
114 struct completion unified_ready; 119 struct completion unified_ready;
115 atomic_t pending_tx_count; 120 wait_queue_head_t tx_credits_wq;
116 wait_queue_head_t wq; 121 struct wmi_cmd_map *cmd;
122 struct wmi_vdev_param_map *vdev_param;
123 struct wmi_pdev_param_map *pdev_param;
117 124
118 struct sk_buff_head wmi_event_list; 125 u32 num_mem_chunks;
119 struct work_struct wmi_event_work; 126 struct ath10k_mem_chunk mem_chunks[ATH10K_MAX_MEM_REQS];
120}; 127};
121 128
122struct ath10k_peer_stat { 129struct ath10k_peer_stat {
@@ -198,17 +205,22 @@ struct ath10k_peer {
198#define ATH10K_VDEV_SETUP_TIMEOUT_HZ (5*HZ) 205#define ATH10K_VDEV_SETUP_TIMEOUT_HZ (5*HZ)
199 206
200struct ath10k_vif { 207struct ath10k_vif {
208 struct list_head list;
209
201 u32 vdev_id; 210 u32 vdev_id;
202 enum wmi_vdev_type vdev_type; 211 enum wmi_vdev_type vdev_type;
203 enum wmi_vdev_subtype vdev_subtype; 212 enum wmi_vdev_subtype vdev_subtype;
204 u32 beacon_interval; 213 u32 beacon_interval;
205 u32 dtim_period; 214 u32 dtim_period;
215 struct sk_buff *beacon;
206 216
207 struct ath10k *ar; 217 struct ath10k *ar;
208 struct ieee80211_vif *vif; 218 struct ieee80211_vif *vif;
209 219
220 struct work_struct wep_key_work;
210 struct ieee80211_key_conf *wep_keys[WMI_MAX_KEY_INDEX + 1]; 221 struct ieee80211_key_conf *wep_keys[WMI_MAX_KEY_INDEX + 1];
211 u8 def_wep_key_index; 222 u8 def_wep_key_idx;
223 u8 def_wep_key_newidx;
212 224
213 u16 tx_seq_no; 225 u16 tx_seq_no;
214 226
@@ -246,6 +258,9 @@ struct ath10k_debug {
246 u32 wmi_service_bitmap[WMI_SERVICE_BM_SIZE]; 258 u32 wmi_service_bitmap[WMI_SERVICE_BM_SIZE];
247 259
248 struct completion event_stats_compl; 260 struct completion event_stats_compl;
261
262 unsigned long htt_stats_mask;
263 struct delayed_work htt_stats_dwork;
249}; 264};
250 265
251enum ath10k_state { 266enum ath10k_state {
@@ -270,12 +285,27 @@ enum ath10k_state {
270 ATH10K_STATE_WEDGED, 285 ATH10K_STATE_WEDGED,
271}; 286};
272 287
288enum ath10k_fw_features {
289 /* wmi_mgmt_rx_hdr contains extra RSSI information */
290 ATH10K_FW_FEATURE_EXT_WMI_MGMT_RX = 0,
291
292 /* firmware from 10X branch */
293 ATH10K_FW_FEATURE_WMI_10X = 1,
294
295 /* firmware support tx frame management over WMI, otherwise it's HTT */
296 ATH10K_FW_FEATURE_HAS_WMI_MGMT_TX = 2,
297
298 /* keep last */
299 ATH10K_FW_FEATURE_COUNT,
300};
301
273struct ath10k { 302struct ath10k {
274 struct ath_common ath_common; 303 struct ath_common ath_common;
275 struct ieee80211_hw *hw; 304 struct ieee80211_hw *hw;
276 struct device *dev; 305 struct device *dev;
277 u8 mac_addr[ETH_ALEN]; 306 u8 mac_addr[ETH_ALEN];
278 307
308 u32 chip_id;
279 u32 target_version; 309 u32 target_version;
280 u8 fw_version_major; 310 u8 fw_version_major;
281 u32 fw_version_minor; 311 u32 fw_version_minor;
@@ -288,6 +318,8 @@ struct ath10k {
288 u32 vht_cap_info; 318 u32 vht_cap_info;
289 u32 num_rf_chains; 319 u32 num_rf_chains;
290 320
321 DECLARE_BITMAP(fw_features, ATH10K_FW_FEATURE_COUNT);
322
291 struct targetdef *targetdef; 323 struct targetdef *targetdef;
292 struct hostdef *hostdef; 324 struct hostdef *hostdef;
293 325
@@ -319,9 +351,19 @@ struct ath10k {
319 } fw; 351 } fw;
320 } hw_params; 352 } hw_params;
321 353
322 const struct firmware *board_data; 354 const struct firmware *board;
355 const void *board_data;
356 size_t board_len;
357
323 const struct firmware *otp; 358 const struct firmware *otp;
359 const void *otp_data;
360 size_t otp_len;
361
324 const struct firmware *firmware; 362 const struct firmware *firmware;
363 const void *firmware_data;
364 size_t firmware_len;
365
366 int fw_api;
325 367
326 struct { 368 struct {
327 struct completion started; 369 struct completion started;
@@ -364,6 +406,7 @@ struct ath10k {
364 /* protects shared structure data */ 406 /* protects shared structure data */
365 spinlock_t data_lock; 407 spinlock_t data_lock;
366 408
409 struct list_head arvifs;
367 struct list_head peers; 410 struct list_head peers;
368 wait_queue_head_t peer_mapping_wq; 411 wait_queue_head_t peer_mapping_wq;
369 412
@@ -372,6 +415,9 @@ struct ath10k {
372 struct completion offchan_tx_completed; 415 struct completion offchan_tx_completed;
373 struct sk_buff *offchan_tx_skb; 416 struct sk_buff *offchan_tx_skb;
374 417
418 struct work_struct wmi_mgmt_tx_work;
419 struct sk_buff_head wmi_mgmt_tx_queue;
420
375 enum ath10k_state state; 421 enum ath10k_state state;
376 422
377 struct work_struct restart_work; 423 struct work_struct restart_work;
@@ -393,7 +439,7 @@ void ath10k_core_destroy(struct ath10k *ar);
393 439
394int ath10k_core_start(struct ath10k *ar); 440int ath10k_core_start(struct ath10k *ar);
395void ath10k_core_stop(struct ath10k *ar); 441void ath10k_core_stop(struct ath10k *ar);
396int ath10k_core_register(struct ath10k *ar); 442int ath10k_core_register(struct ath10k *ar, u32 chip_id);
397void ath10k_core_unregister(struct ath10k *ar); 443void ath10k_core_unregister(struct ath10k *ar);
398 444
399#endif /* _CORE_H_ */ 445#endif /* _CORE_H_ */
diff --git a/drivers/net/wireless/ath/ath10k/debug.c b/drivers/net/wireless/ath/ath10k/debug.c
index 3d65594fa098..760ff2289e3c 100644
--- a/drivers/net/wireless/ath/ath10k/debug.c
+++ b/drivers/net/wireless/ath/ath10k/debug.c
@@ -21,6 +21,9 @@
21#include "core.h" 21#include "core.h"
22#include "debug.h" 22#include "debug.h"
23 23
24/* ms */
25#define ATH10K_DEBUG_HTT_STATS_INTERVAL 1000
26
24static int ath10k_printk(const char *level, const char *fmt, ...) 27static int ath10k_printk(const char *level, const char *fmt, ...)
25{ 28{
26 struct va_format vaf; 29 struct va_format vaf;
@@ -260,7 +263,6 @@ void ath10k_debug_read_target_stats(struct ath10k *ar,
260 } 263 }
261 264
262 spin_unlock_bh(&ar->data_lock); 265 spin_unlock_bh(&ar->data_lock);
263 mutex_unlock(&ar->conf_mutex);
264 complete(&ar->debug.event_stats_compl); 266 complete(&ar->debug.event_stats_compl);
265} 267}
266 268
@@ -499,6 +501,144 @@ static const struct file_operations fops_simulate_fw_crash = {
499 .llseek = default_llseek, 501 .llseek = default_llseek,
500}; 502};
501 503
504static ssize_t ath10k_read_chip_id(struct file *file, char __user *user_buf,
505 size_t count, loff_t *ppos)
506{
507 struct ath10k *ar = file->private_data;
508 unsigned int len;
509 char buf[50];
510
511 len = scnprintf(buf, sizeof(buf), "0x%08x\n", ar->chip_id);
512
513 return simple_read_from_buffer(user_buf, count, ppos, buf, len);
514}
515
516static const struct file_operations fops_chip_id = {
517 .read = ath10k_read_chip_id,
518 .open = simple_open,
519 .owner = THIS_MODULE,
520 .llseek = default_llseek,
521};
522
523static int ath10k_debug_htt_stats_req(struct ath10k *ar)
524{
525 u64 cookie;
526 int ret;
527
528 lockdep_assert_held(&ar->conf_mutex);
529
530 if (ar->debug.htt_stats_mask == 0)
531 /* htt stats are disabled */
532 return 0;
533
534 if (ar->state != ATH10K_STATE_ON)
535 return 0;
536
537 cookie = get_jiffies_64();
538
539 ret = ath10k_htt_h2t_stats_req(&ar->htt, ar->debug.htt_stats_mask,
540 cookie);
541 if (ret) {
542 ath10k_warn("failed to send htt stats request: %d\n", ret);
543 return ret;
544 }
545
546 queue_delayed_work(ar->workqueue, &ar->debug.htt_stats_dwork,
547 msecs_to_jiffies(ATH10K_DEBUG_HTT_STATS_INTERVAL));
548
549 return 0;
550}
551
552static void ath10k_debug_htt_stats_dwork(struct work_struct *work)
553{
554 struct ath10k *ar = container_of(work, struct ath10k,
555 debug.htt_stats_dwork.work);
556
557 mutex_lock(&ar->conf_mutex);
558
559 ath10k_debug_htt_stats_req(ar);
560
561 mutex_unlock(&ar->conf_mutex);
562}
563
564static ssize_t ath10k_read_htt_stats_mask(struct file *file,
565 char __user *user_buf,
566 size_t count, loff_t *ppos)
567{
568 struct ath10k *ar = file->private_data;
569 char buf[32];
570 unsigned int len;
571
572 len = scnprintf(buf, sizeof(buf), "%lu\n", ar->debug.htt_stats_mask);
573
574 return simple_read_from_buffer(user_buf, count, ppos, buf, len);
575}
576
577static ssize_t ath10k_write_htt_stats_mask(struct file *file,
578 const char __user *user_buf,
579 size_t count, loff_t *ppos)
580{
581 struct ath10k *ar = file->private_data;
582 unsigned long mask;
583 int ret;
584
585 ret = kstrtoul_from_user(user_buf, count, 0, &mask);
586 if (ret)
587 return ret;
588
589 /* max 8 bit masks (for now) */
590 if (mask > 0xff)
591 return -E2BIG;
592
593 mutex_lock(&ar->conf_mutex);
594
595 ar->debug.htt_stats_mask = mask;
596
597 ret = ath10k_debug_htt_stats_req(ar);
598 if (ret)
599 goto out;
600
601 ret = count;
602
603out:
604 mutex_unlock(&ar->conf_mutex);
605
606 return ret;
607}
608
609static const struct file_operations fops_htt_stats_mask = {
610 .read = ath10k_read_htt_stats_mask,
611 .write = ath10k_write_htt_stats_mask,
612 .open = simple_open,
613 .owner = THIS_MODULE,
614 .llseek = default_llseek,
615};
616
617int ath10k_debug_start(struct ath10k *ar)
618{
619 int ret;
620
621 lockdep_assert_held(&ar->conf_mutex);
622
623 ret = ath10k_debug_htt_stats_req(ar);
624 if (ret)
625 /* continue normally anyway, this isn't serious */
626 ath10k_warn("failed to start htt stats workqueue: %d\n", ret);
627
628 return 0;
629}
630
631void ath10k_debug_stop(struct ath10k *ar)
632{
633 lockdep_assert_held(&ar->conf_mutex);
634
635 /* Must not use _sync to avoid deadlock, we do that in
636 * ath10k_debug_destroy(). The check for htt_stats_mask is to avoid
637 * warning from del_timer(). */
638 if (ar->debug.htt_stats_mask != 0)
639 cancel_delayed_work(&ar->debug.htt_stats_dwork);
640}
641
502int ath10k_debug_create(struct ath10k *ar) 642int ath10k_debug_create(struct ath10k *ar)
503{ 643{
504 ar->debug.debugfs_phy = debugfs_create_dir("ath10k", 644 ar->debug.debugfs_phy = debugfs_create_dir("ath10k",
@@ -507,6 +647,9 @@ int ath10k_debug_create(struct ath10k *ar)
507 if (!ar->debug.debugfs_phy) 647 if (!ar->debug.debugfs_phy)
508 return -ENOMEM; 648 return -ENOMEM;
509 649
650 INIT_DELAYED_WORK(&ar->debug.htt_stats_dwork,
651 ath10k_debug_htt_stats_dwork);
652
510 init_completion(&ar->debug.event_stats_compl); 653 init_completion(&ar->debug.event_stats_compl);
511 654
512 debugfs_create_file("fw_stats", S_IRUSR, ar->debug.debugfs_phy, ar, 655 debugfs_create_file("fw_stats", S_IRUSR, ar->debug.debugfs_phy, ar,
@@ -518,8 +661,20 @@ int ath10k_debug_create(struct ath10k *ar)
518 debugfs_create_file("simulate_fw_crash", S_IRUSR, ar->debug.debugfs_phy, 661 debugfs_create_file("simulate_fw_crash", S_IRUSR, ar->debug.debugfs_phy,
519 ar, &fops_simulate_fw_crash); 662 ar, &fops_simulate_fw_crash);
520 663
664 debugfs_create_file("chip_id", S_IRUSR, ar->debug.debugfs_phy,
665 ar, &fops_chip_id);
666
667 debugfs_create_file("htt_stats_mask", S_IRUSR, ar->debug.debugfs_phy,
668 ar, &fops_htt_stats_mask);
669
521 return 0; 670 return 0;
522} 671}
672
673void ath10k_debug_destroy(struct ath10k *ar)
674{
675 cancel_delayed_work_sync(&ar->debug.htt_stats_dwork);
676}
677
523#endif /* CONFIG_ATH10K_DEBUGFS */ 678#endif /* CONFIG_ATH10K_DEBUGFS */
524 679
525#ifdef CONFIG_ATH10K_DEBUG 680#ifdef CONFIG_ATH10K_DEBUG
diff --git a/drivers/net/wireless/ath/ath10k/debug.h b/drivers/net/wireless/ath/ath10k/debug.h
index 168140c54028..3cfe3ee90dbe 100644
--- a/drivers/net/wireless/ath/ath10k/debug.h
+++ b/drivers/net/wireless/ath/ath10k/debug.h
@@ -27,22 +27,26 @@ enum ath10k_debug_mask {
27 ATH10K_DBG_HTC = 0x00000004, 27 ATH10K_DBG_HTC = 0x00000004,
28 ATH10K_DBG_HTT = 0x00000008, 28 ATH10K_DBG_HTT = 0x00000008,
29 ATH10K_DBG_MAC = 0x00000010, 29 ATH10K_DBG_MAC = 0x00000010,
30 ATH10K_DBG_CORE = 0x00000020, 30 ATH10K_DBG_BOOT = 0x00000020,
31 ATH10K_DBG_PCI_DUMP = 0x00000040, 31 ATH10K_DBG_PCI_DUMP = 0x00000040,
32 ATH10K_DBG_HTT_DUMP = 0x00000080, 32 ATH10K_DBG_HTT_DUMP = 0x00000080,
33 ATH10K_DBG_MGMT = 0x00000100, 33 ATH10K_DBG_MGMT = 0x00000100,
34 ATH10K_DBG_DATA = 0x00000200, 34 ATH10K_DBG_DATA = 0x00000200,
35 ATH10K_DBG_BMI = 0x00000400,
35 ATH10K_DBG_ANY = 0xffffffff, 36 ATH10K_DBG_ANY = 0xffffffff,
36}; 37};
37 38
38extern unsigned int ath10k_debug_mask; 39extern unsigned int ath10k_debug_mask;
39 40
40extern __printf(1, 2) int ath10k_info(const char *fmt, ...); 41__printf(1, 2) int ath10k_info(const char *fmt, ...);
41extern __printf(1, 2) int ath10k_err(const char *fmt, ...); 42__printf(1, 2) int ath10k_err(const char *fmt, ...);
42extern __printf(1, 2) int ath10k_warn(const char *fmt, ...); 43__printf(1, 2) int ath10k_warn(const char *fmt, ...);
43 44
44#ifdef CONFIG_ATH10K_DEBUGFS 45#ifdef CONFIG_ATH10K_DEBUGFS
46int ath10k_debug_start(struct ath10k *ar);
47void ath10k_debug_stop(struct ath10k *ar);
45int ath10k_debug_create(struct ath10k *ar); 48int ath10k_debug_create(struct ath10k *ar);
49void ath10k_debug_destroy(struct ath10k *ar);
46void ath10k_debug_read_service_map(struct ath10k *ar, 50void ath10k_debug_read_service_map(struct ath10k *ar,
47 void *service_map, 51 void *service_map,
48 size_t map_size); 52 size_t map_size);
@@ -50,11 +54,24 @@ void ath10k_debug_read_target_stats(struct ath10k *ar,
50 struct wmi_stats_event *ev); 54 struct wmi_stats_event *ev);
51 55
52#else 56#else
57static inline int ath10k_debug_start(struct ath10k *ar)
58{
59 return 0;
60}
61
62static inline void ath10k_debug_stop(struct ath10k *ar)
63{
64}
65
53static inline int ath10k_debug_create(struct ath10k *ar) 66static inline int ath10k_debug_create(struct ath10k *ar)
54{ 67{
55 return 0; 68 return 0;
56} 69}
57 70
71static inline void ath10k_debug_destroy(struct ath10k *ar)
72{
73}
74
58static inline void ath10k_debug_read_service_map(struct ath10k *ar, 75static inline void ath10k_debug_read_service_map(struct ath10k *ar,
59 void *service_map, 76 void *service_map,
60 size_t map_size) 77 size_t map_size)
@@ -68,7 +85,7 @@ static inline void ath10k_debug_read_target_stats(struct ath10k *ar,
68#endif /* CONFIG_ATH10K_DEBUGFS */ 85#endif /* CONFIG_ATH10K_DEBUGFS */
69 86
70#ifdef CONFIG_ATH10K_DEBUG 87#ifdef CONFIG_ATH10K_DEBUG
71extern __printf(2, 3) void ath10k_dbg(enum ath10k_debug_mask mask, 88__printf(2, 3) void ath10k_dbg(enum ath10k_debug_mask mask,
72 const char *fmt, ...); 89 const char *fmt, ...);
73void ath10k_dbg_dump(enum ath10k_debug_mask mask, 90void ath10k_dbg_dump(enum ath10k_debug_mask mask,
74 const char *msg, const char *prefix, 91 const char *msg, const char *prefix,
diff --git a/drivers/net/wireless/ath/ath10k/htc.c b/drivers/net/wireless/ath/ath10k/htc.c
index ef3329ef52f3..3118d7506734 100644
--- a/drivers/net/wireless/ath/ath10k/htc.c
+++ b/drivers/net/wireless/ath/ath10k/htc.c
@@ -103,10 +103,10 @@ static void ath10k_htc_prepare_tx_skb(struct ath10k_htc_ep *ep,
103 struct ath10k_htc_hdr *hdr; 103 struct ath10k_htc_hdr *hdr;
104 104
105 hdr = (struct ath10k_htc_hdr *)skb->data; 105 hdr = (struct ath10k_htc_hdr *)skb->data;
106 memset(hdr, 0, sizeof(*hdr));
107 106
108 hdr->eid = ep->eid; 107 hdr->eid = ep->eid;
109 hdr->len = __cpu_to_le16(skb->len - sizeof(*hdr)); 108 hdr->len = __cpu_to_le16(skb->len - sizeof(*hdr));
109 hdr->flags = 0;
110 110
111 spin_lock_bh(&ep->htc->tx_lock); 111 spin_lock_bh(&ep->htc->tx_lock);
112 hdr->seq_no = ep->seq_no++; 112 hdr->seq_no = ep->seq_no++;
@@ -117,134 +117,13 @@ static void ath10k_htc_prepare_tx_skb(struct ath10k_htc_ep *ep,
117 spin_unlock_bh(&ep->htc->tx_lock); 117 spin_unlock_bh(&ep->htc->tx_lock);
118} 118}
119 119
120static int ath10k_htc_issue_skb(struct ath10k_htc *htc,
121 struct ath10k_htc_ep *ep,
122 struct sk_buff *skb,
123 u8 credits)
124{
125 struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(skb);
126 int ret;
127
128 ath10k_dbg(ATH10K_DBG_HTC, "%s: ep %d skb %p\n", __func__,
129 ep->eid, skb);
130
131 ath10k_htc_prepare_tx_skb(ep, skb);
132
133 ret = ath10k_skb_map(htc->ar->dev, skb);
134 if (ret)
135 goto err;
136
137 ret = ath10k_hif_send_head(htc->ar,
138 ep->ul_pipe_id,
139 ep->eid,
140 skb->len,
141 skb);
142 if (unlikely(ret))
143 goto err;
144
145 return 0;
146err:
147 ath10k_warn("HTC issue failed: %d\n", ret);
148
149 spin_lock_bh(&htc->tx_lock);
150 ep->tx_credits += credits;
151 spin_unlock_bh(&htc->tx_lock);
152
153 /* this is the simplest way to handle out-of-resources for non-credit
154 * based endpoints. credit based endpoints can still get -ENOSR, but
155 * this is highly unlikely as credit reservation should prevent that */
156 if (ret == -ENOSR) {
157 spin_lock_bh(&htc->tx_lock);
158 __skb_queue_head(&ep->tx_queue, skb);
159 spin_unlock_bh(&htc->tx_lock);
160
161 return ret;
162 }
163
164 skb_cb->is_aborted = true;
165 ath10k_htc_notify_tx_completion(ep, skb);
166
167 return ret;
168}
169
170static struct sk_buff *ath10k_htc_get_skb_credit_based(struct ath10k_htc *htc,
171 struct ath10k_htc_ep *ep,
172 u8 *credits)
173{
174 struct sk_buff *skb;
175 struct ath10k_skb_cb *skb_cb;
176 int credits_required;
177 int remainder;
178 unsigned int transfer_len;
179
180 lockdep_assert_held(&htc->tx_lock);
181
182 skb = __skb_dequeue(&ep->tx_queue);
183 if (!skb)
184 return NULL;
185
186 skb_cb = ATH10K_SKB_CB(skb);
187 transfer_len = skb->len;
188
189 if (likely(transfer_len <= htc->target_credit_size)) {
190 credits_required = 1;
191 } else {
192 /* figure out how many credits this message requires */
193 credits_required = transfer_len / htc->target_credit_size;
194 remainder = transfer_len % htc->target_credit_size;
195
196 if (remainder)
197 credits_required++;
198 }
199
200 ath10k_dbg(ATH10K_DBG_HTC, "Credits required %d got %d\n",
201 credits_required, ep->tx_credits);
202
203 if (ep->tx_credits < credits_required) {
204 __skb_queue_head(&ep->tx_queue, skb);
205 return NULL;
206 }
207
208 ep->tx_credits -= credits_required;
209 *credits = credits_required;
210 return skb;
211}
212
213static void ath10k_htc_send_work(struct work_struct *work)
214{
215 struct ath10k_htc_ep *ep = container_of(work,
216 struct ath10k_htc_ep, send_work);
217 struct ath10k_htc *htc = ep->htc;
218 struct sk_buff *skb;
219 u8 credits = 0;
220 int ret;
221
222 while (true) {
223 if (ep->ul_is_polled)
224 ath10k_htc_send_complete_check(ep, 0);
225
226 spin_lock_bh(&htc->tx_lock);
227 if (ep->tx_credit_flow_enabled)
228 skb = ath10k_htc_get_skb_credit_based(htc, ep,
229 &credits);
230 else
231 skb = __skb_dequeue(&ep->tx_queue);
232 spin_unlock_bh(&htc->tx_lock);
233
234 if (!skb)
235 break;
236
237 ret = ath10k_htc_issue_skb(htc, ep, skb, credits);
238 if (ret == -ENOSR)
239 break;
240 }
241}
242
243int ath10k_htc_send(struct ath10k_htc *htc, 120int ath10k_htc_send(struct ath10k_htc *htc,
244 enum ath10k_htc_ep_id eid, 121 enum ath10k_htc_ep_id eid,
245 struct sk_buff *skb) 122 struct sk_buff *skb)
246{ 123{
247 struct ath10k_htc_ep *ep = &htc->endpoint[eid]; 124 struct ath10k_htc_ep *ep = &htc->endpoint[eid];
125 int credits = 0;
126 int ret;
248 127
249 if (htc->ar->state == ATH10K_STATE_WEDGED) 128 if (htc->ar->state == ATH10K_STATE_WEDGED)
250 return -ECOMM; 129 return -ECOMM;
@@ -254,18 +133,55 @@ int ath10k_htc_send(struct ath10k_htc *htc,
254 return -ENOENT; 133 return -ENOENT;
255 } 134 }
256 135
136 /* FIXME: This looks ugly, can we fix it? */
257 spin_lock_bh(&htc->tx_lock); 137 spin_lock_bh(&htc->tx_lock);
258 if (htc->stopped) { 138 if (htc->stopped) {
259 spin_unlock_bh(&htc->tx_lock); 139 spin_unlock_bh(&htc->tx_lock);
260 return -ESHUTDOWN; 140 return -ESHUTDOWN;
261 } 141 }
142 spin_unlock_bh(&htc->tx_lock);
262 143
263 __skb_queue_tail(&ep->tx_queue, skb);
264 skb_push(skb, sizeof(struct ath10k_htc_hdr)); 144 skb_push(skb, sizeof(struct ath10k_htc_hdr));
265 spin_unlock_bh(&htc->tx_lock);
266 145
267 queue_work(htc->ar->workqueue, &ep->send_work); 146 if (ep->tx_credit_flow_enabled) {
147 credits = DIV_ROUND_UP(skb->len, htc->target_credit_size);
148 spin_lock_bh(&htc->tx_lock);
149 if (ep->tx_credits < credits) {
150 spin_unlock_bh(&htc->tx_lock);
151 ret = -EAGAIN;
152 goto err_pull;
153 }
154 ep->tx_credits -= credits;
155 spin_unlock_bh(&htc->tx_lock);
156 }
157
158 ath10k_htc_prepare_tx_skb(ep, skb);
159
160 ret = ath10k_skb_map(htc->ar->dev, skb);
161 if (ret)
162 goto err_credits;
163
164 ret = ath10k_hif_send_head(htc->ar, ep->ul_pipe_id, ep->eid,
165 skb->len, skb);
166 if (ret)
167 goto err_unmap;
168
268 return 0; 169 return 0;
170
171err_unmap:
172 ath10k_skb_unmap(htc->ar->dev, skb);
173err_credits:
174 if (ep->tx_credit_flow_enabled) {
175 spin_lock_bh(&htc->tx_lock);
176 ep->tx_credits += credits;
177 spin_unlock_bh(&htc->tx_lock);
178
179 if (ep->ep_ops.ep_tx_credits)
180 ep->ep_ops.ep_tx_credits(htc->ar);
181 }
182err_pull:
183 skb_pull(skb, sizeof(struct ath10k_htc_hdr));
184 return ret;
269} 185}
270 186
271static int ath10k_htc_tx_completion_handler(struct ath10k *ar, 187static int ath10k_htc_tx_completion_handler(struct ath10k *ar,
@@ -278,39 +194,9 @@ static int ath10k_htc_tx_completion_handler(struct ath10k *ar,
278 ath10k_htc_notify_tx_completion(ep, skb); 194 ath10k_htc_notify_tx_completion(ep, skb);
279 /* the skb now belongs to the completion handler */ 195 /* the skb now belongs to the completion handler */
280 196
281 /* note: when using TX credit flow, the re-checking of queues happens
282 * when credits flow back from the target. in the non-TX credit case,
283 * we recheck after the packet completes */
284 spin_lock_bh(&htc->tx_lock);
285 if (!ep->tx_credit_flow_enabled && !htc->stopped)
286 queue_work(ar->workqueue, &ep->send_work);
287 spin_unlock_bh(&htc->tx_lock);
288
289 return 0; 197 return 0;
290} 198}
291 199
292/* flush endpoint TX queue */
293static void ath10k_htc_flush_endpoint_tx(struct ath10k_htc *htc,
294 struct ath10k_htc_ep *ep)
295{
296 struct sk_buff *skb;
297 struct ath10k_skb_cb *skb_cb;
298
299 spin_lock_bh(&htc->tx_lock);
300 for (;;) {
301 skb = __skb_dequeue(&ep->tx_queue);
302 if (!skb)
303 break;
304
305 skb_cb = ATH10K_SKB_CB(skb);
306 skb_cb->is_aborted = true;
307 ath10k_htc_notify_tx_completion(ep, skb);
308 }
309 spin_unlock_bh(&htc->tx_lock);
310
311 cancel_work_sync(&ep->send_work);
312}
313
314/***********/ 200/***********/
315/* Receive */ 201/* Receive */
316/***********/ 202/***********/
@@ -340,8 +226,11 @@ ath10k_htc_process_credit_report(struct ath10k_htc *htc,
340 ep = &htc->endpoint[report->eid]; 226 ep = &htc->endpoint[report->eid];
341 ep->tx_credits += report->credits; 227 ep->tx_credits += report->credits;
342 228
343 if (ep->tx_credits && !skb_queue_empty(&ep->tx_queue)) 229 if (ep->ep_ops.ep_tx_credits) {
344 queue_work(htc->ar->workqueue, &ep->send_work); 230 spin_unlock_bh(&htc->tx_lock);
231 ep->ep_ops.ep_tx_credits(htc->ar);
232 spin_lock_bh(&htc->tx_lock);
233 }
345 } 234 }
346 spin_unlock_bh(&htc->tx_lock); 235 spin_unlock_bh(&htc->tx_lock);
347} 236}
@@ -599,10 +488,8 @@ static void ath10k_htc_reset_endpoint_states(struct ath10k_htc *htc)
599 ep->max_ep_message_len = 0; 488 ep->max_ep_message_len = 0;
600 ep->max_tx_queue_depth = 0; 489 ep->max_tx_queue_depth = 0;
601 ep->eid = i; 490 ep->eid = i;
602 skb_queue_head_init(&ep->tx_queue);
603 ep->htc = htc; 491 ep->htc = htc;
604 ep->tx_credit_flow_enabled = true; 492 ep->tx_credit_flow_enabled = true;
605 INIT_WORK(&ep->send_work, ath10k_htc_send_work);
606 } 493 }
607} 494}
608 495
@@ -752,8 +639,8 @@ int ath10k_htc_connect_service(struct ath10k_htc *htc,
752 tx_alloc = ath10k_htc_get_credit_allocation(htc, 639 tx_alloc = ath10k_htc_get_credit_allocation(htc,
753 conn_req->service_id); 640 conn_req->service_id);
754 if (!tx_alloc) 641 if (!tx_alloc)
755 ath10k_dbg(ATH10K_DBG_HTC, 642 ath10k_dbg(ATH10K_DBG_BOOT,
756 "HTC Service %s does not allocate target credits\n", 643 "boot htc service %s does not allocate target credits\n",
757 htc_service_name(conn_req->service_id)); 644 htc_service_name(conn_req->service_id));
758 645
759 skb = ath10k_htc_build_tx_ctrl_skb(htc->ar); 646 skb = ath10k_htc_build_tx_ctrl_skb(htc->ar);
@@ -772,16 +659,16 @@ int ath10k_htc_connect_service(struct ath10k_htc *htc,
772 659
773 flags |= SM(tx_alloc, ATH10K_HTC_CONN_FLAGS_RECV_ALLOC); 660 flags |= SM(tx_alloc, ATH10K_HTC_CONN_FLAGS_RECV_ALLOC);
774 661
775 req_msg = &msg->connect_service;
776 req_msg->flags = __cpu_to_le16(flags);
777 req_msg->service_id = __cpu_to_le16(conn_req->service_id);
778
779 /* Only enable credit flow control for WMI ctrl service */ 662 /* Only enable credit flow control for WMI ctrl service */
780 if (conn_req->service_id != ATH10K_HTC_SVC_ID_WMI_CONTROL) { 663 if (conn_req->service_id != ATH10K_HTC_SVC_ID_WMI_CONTROL) {
781 flags |= ATH10K_HTC_CONN_FLAGS_DISABLE_CREDIT_FLOW_CTRL; 664 flags |= ATH10K_HTC_CONN_FLAGS_DISABLE_CREDIT_FLOW_CTRL;
782 disable_credit_flow_ctrl = true; 665 disable_credit_flow_ctrl = true;
783 } 666 }
784 667
668 req_msg = &msg->connect_service;
669 req_msg->flags = __cpu_to_le16(flags);
670 req_msg->service_id = __cpu_to_le16(conn_req->service_id);
671
785 INIT_COMPLETION(htc->ctl_resp); 672 INIT_COMPLETION(htc->ctl_resp);
786 673
787 status = ath10k_htc_send(htc, ATH10K_HTC_EP_0, skb); 674 status = ath10k_htc_send(htc, ATH10K_HTC_EP_0, skb);
@@ -873,19 +760,19 @@ setup:
873 if (status) 760 if (status)
874 return status; 761 return status;
875 762
876 ath10k_dbg(ATH10K_DBG_HTC, 763 ath10k_dbg(ATH10K_DBG_BOOT,
877 "HTC service: %s UL pipe: %d DL pipe: %d eid: %d ready\n", 764 "boot htc service '%s' ul pipe %d dl pipe %d eid %d ready\n",
878 htc_service_name(ep->service_id), ep->ul_pipe_id, 765 htc_service_name(ep->service_id), ep->ul_pipe_id,
879 ep->dl_pipe_id, ep->eid); 766 ep->dl_pipe_id, ep->eid);
880 767
881 ath10k_dbg(ATH10K_DBG_HTC, 768 ath10k_dbg(ATH10K_DBG_BOOT,
882 "EP %d UL polled: %d, DL polled: %d\n", 769 "boot htc ep %d ul polled %d dl polled %d\n",
883 ep->eid, ep->ul_is_polled, ep->dl_is_polled); 770 ep->eid, ep->ul_is_polled, ep->dl_is_polled);
884 771
885 if (disable_credit_flow_ctrl && ep->tx_credit_flow_enabled) { 772 if (disable_credit_flow_ctrl && ep->tx_credit_flow_enabled) {
886 ep->tx_credit_flow_enabled = false; 773 ep->tx_credit_flow_enabled = false;
887 ath10k_dbg(ATH10K_DBG_HTC, 774 ath10k_dbg(ATH10K_DBG_BOOT,
888 "HTC service: %s eid: %d TX flow control disabled\n", 775 "boot htc service '%s' eid %d TX flow control disabled\n",
889 htc_service_name(ep->service_id), assigned_eid); 776 htc_service_name(ep->service_id), assigned_eid);
890 } 777 }
891 778
@@ -945,18 +832,10 @@ int ath10k_htc_start(struct ath10k_htc *htc)
945 */ 832 */
946void ath10k_htc_stop(struct ath10k_htc *htc) 833void ath10k_htc_stop(struct ath10k_htc *htc)
947{ 834{
948 int i;
949 struct ath10k_htc_ep *ep;
950
951 spin_lock_bh(&htc->tx_lock); 835 spin_lock_bh(&htc->tx_lock);
952 htc->stopped = true; 836 htc->stopped = true;
953 spin_unlock_bh(&htc->tx_lock); 837 spin_unlock_bh(&htc->tx_lock);
954 838
955 for (i = ATH10K_HTC_EP_0; i < ATH10K_HTC_EP_COUNT; i++) {
956 ep = &htc->endpoint[i];
957 ath10k_htc_flush_endpoint_tx(htc, ep);
958 }
959
960 ath10k_hif_stop(htc->ar); 839 ath10k_hif_stop(htc->ar);
961} 840}
962 841
diff --git a/drivers/net/wireless/ath/ath10k/htc.h b/drivers/net/wireless/ath/ath10k/htc.h
index e1dd8c761853..4716d331e6b6 100644
--- a/drivers/net/wireless/ath/ath10k/htc.h
+++ b/drivers/net/wireless/ath/ath10k/htc.h
@@ -276,6 +276,7 @@ struct ath10k_htc_ops {
276struct ath10k_htc_ep_ops { 276struct ath10k_htc_ep_ops {
277 void (*ep_tx_complete)(struct ath10k *, struct sk_buff *); 277 void (*ep_tx_complete)(struct ath10k *, struct sk_buff *);
278 void (*ep_rx_complete)(struct ath10k *, struct sk_buff *); 278 void (*ep_rx_complete)(struct ath10k *, struct sk_buff *);
279 void (*ep_tx_credits)(struct ath10k *);
279}; 280};
280 281
281/* service connection information */ 282/* service connection information */
@@ -315,15 +316,11 @@ struct ath10k_htc_ep {
315 int ul_is_polled; /* call HIF to get tx completions */ 316 int ul_is_polled; /* call HIF to get tx completions */
316 int dl_is_polled; /* call HIF to fetch rx (not implemented) */ 317 int dl_is_polled; /* call HIF to fetch rx (not implemented) */
317 318
318 struct sk_buff_head tx_queue;
319
320 u8 seq_no; /* for debugging */ 319 u8 seq_no; /* for debugging */
321 int tx_credits; 320 int tx_credits;
322 int tx_credit_size; 321 int tx_credit_size;
323 int tx_credits_per_max_message; 322 int tx_credits_per_max_message;
324 bool tx_credit_flow_enabled; 323 bool tx_credit_flow_enabled;
325
326 struct work_struct send_work;
327}; 324};
328 325
329struct ath10k_htc_svc_tx_credits { 326struct ath10k_htc_svc_tx_credits {
diff --git a/drivers/net/wireless/ath/ath10k/htt.c b/drivers/net/wireless/ath/ath10k/htt.c
index 39342c5cfcb2..5f7eeebc5432 100644
--- a/drivers/net/wireless/ath/ath10k/htt.c
+++ b/drivers/net/wireless/ath/ath10k/htt.c
@@ -104,21 +104,16 @@ err_htc_attach:
104 104
105static int ath10k_htt_verify_version(struct ath10k_htt *htt) 105static int ath10k_htt_verify_version(struct ath10k_htt *htt)
106{ 106{
107 ath10k_dbg(ATH10K_DBG_HTT, 107 ath10k_info("htt target version %d.%d\n",
108 "htt target version %d.%d; host version %d.%d\n", 108 htt->target_version_major, htt->target_version_minor);
109 htt->target_version_major, 109
110 htt->target_version_minor, 110 if (htt->target_version_major != 2 &&
111 HTT_CURRENT_VERSION_MAJOR, 111 htt->target_version_major != 3) {
112 HTT_CURRENT_VERSION_MINOR); 112 ath10k_err("unsupported htt major version %d. supported versions are 2 and 3\n",
113 113 htt->target_version_major);
114 if (htt->target_version_major != HTT_CURRENT_VERSION_MAJOR) {
115 ath10k_err("htt major versions are incompatible!\n");
116 return -ENOTSUPP; 114 return -ENOTSUPP;
117 } 115 }
118 116
119 if (htt->target_version_minor != HTT_CURRENT_VERSION_MINOR)
120 ath10k_warn("htt minor version differ but still compatible\n");
121
122 return 0; 117 return 0;
123} 118}
124 119
diff --git a/drivers/net/wireless/ath/ath10k/htt.h b/drivers/net/wireless/ath/ath10k/htt.h
index 318be4629cde..1a337e93b7e9 100644
--- a/drivers/net/wireless/ath/ath10k/htt.h
+++ b/drivers/net/wireless/ath/ath10k/htt.h
@@ -19,13 +19,11 @@
19#define _HTT_H_ 19#define _HTT_H_
20 20
21#include <linux/bug.h> 21#include <linux/bug.h>
22#include <linux/interrupt.h>
22 23
23#include "htc.h" 24#include "htc.h"
24#include "rx_desc.h" 25#include "rx_desc.h"
25 26
26#define HTT_CURRENT_VERSION_MAJOR 2
27#define HTT_CURRENT_VERSION_MINOR 1
28
29enum htt_dbg_stats_type { 27enum htt_dbg_stats_type {
30 HTT_DBG_STATS_WAL_PDEV_TXRX = 1 << 0, 28 HTT_DBG_STATS_WAL_PDEV_TXRX = 1 << 0,
31 HTT_DBG_STATS_RX_REORDER = 1 << 1, 29 HTT_DBG_STATS_RX_REORDER = 1 << 1,
@@ -45,6 +43,9 @@ enum htt_h2t_msg_type { /* host-to-target */
45 HTT_H2T_MSG_TYPE_SYNC = 4, 43 HTT_H2T_MSG_TYPE_SYNC = 4,
46 HTT_H2T_MSG_TYPE_AGGR_CFG = 5, 44 HTT_H2T_MSG_TYPE_AGGR_CFG = 5,
47 HTT_H2T_MSG_TYPE_FRAG_DESC_BANK_CFG = 6, 45 HTT_H2T_MSG_TYPE_FRAG_DESC_BANK_CFG = 6,
46
47 /* This command is used for sending management frames in HTT < 3.0.
48 * HTT >= 3.0 uses TX_FRM for everything. */
48 HTT_H2T_MSG_TYPE_MGMT_TX = 7, 49 HTT_H2T_MSG_TYPE_MGMT_TX = 7,
49 50
50 HTT_H2T_NUM_MSGS /* keep this last */ 51 HTT_H2T_NUM_MSGS /* keep this last */
@@ -1268,6 +1269,7 @@ struct ath10k_htt {
1268 /* set if host-fw communication goes haywire 1269 /* set if host-fw communication goes haywire
1269 * used to avoid further failures */ 1270 * used to avoid further failures */
1270 bool rx_confused; 1271 bool rx_confused;
1272 struct tasklet_struct rx_replenish_task;
1271}; 1273};
1272 1274
1273#define RX_HTT_HDR_STATUS_LEN 64 1275#define RX_HTT_HDR_STATUS_LEN 64
@@ -1308,6 +1310,10 @@ struct htt_rx_desc {
1308#define HTT_RX_BUF_SIZE 1920 1310#define HTT_RX_BUF_SIZE 1920
1309#define HTT_RX_MSDU_SIZE (HTT_RX_BUF_SIZE - (int)sizeof(struct htt_rx_desc)) 1311#define HTT_RX_MSDU_SIZE (HTT_RX_BUF_SIZE - (int)sizeof(struct htt_rx_desc))
1310 1312
1313/* Refill a bunch of RX buffers for each refill round so that FW/HW can handle
1314 * aggregated traffic more nicely. */
1315#define ATH10K_HTT_MAX_NUM_REFILL 16
1316
1311/* 1317/*
1312 * DMA_MAP expects the buffer to be an integral number of cache lines. 1318 * DMA_MAP expects the buffer to be an integral number of cache lines.
1313 * Rather than checking the actual cache line size, this code makes a 1319 * Rather than checking the actual cache line size, this code makes a
@@ -1327,6 +1333,7 @@ void ath10k_htt_rx_detach(struct ath10k_htt *htt);
1327void ath10k_htt_htc_tx_complete(struct ath10k *ar, struct sk_buff *skb); 1333void ath10k_htt_htc_tx_complete(struct ath10k *ar, struct sk_buff *skb);
1328void ath10k_htt_t2h_msg_handler(struct ath10k *ar, struct sk_buff *skb); 1334void ath10k_htt_t2h_msg_handler(struct ath10k *ar, struct sk_buff *skb);
1329int ath10k_htt_h2t_ver_req_msg(struct ath10k_htt *htt); 1335int ath10k_htt_h2t_ver_req_msg(struct ath10k_htt *htt);
1336int ath10k_htt_h2t_stats_req(struct ath10k_htt *htt, u8 mask, u64 cookie);
1330int ath10k_htt_send_rx_ring_cfg_ll(struct ath10k_htt *htt); 1337int ath10k_htt_send_rx_ring_cfg_ll(struct ath10k_htt *htt);
1331 1338
1332void __ath10k_htt_tx_dec_pending(struct ath10k_htt *htt); 1339void __ath10k_htt_tx_dec_pending(struct ath10k_htt *htt);
diff --git a/drivers/net/wireless/ath/ath10k/htt_rx.c b/drivers/net/wireless/ath/ath10k/htt_rx.c
index e784c40b904b..90d4f74c28d7 100644
--- a/drivers/net/wireless/ath/ath10k/htt_rx.c
+++ b/drivers/net/wireless/ath/ath10k/htt_rx.c
@@ -20,6 +20,7 @@
20#include "htt.h" 20#include "htt.h"
21#include "txrx.h" 21#include "txrx.h"
22#include "debug.h" 22#include "debug.h"
23#include "trace.h"
23 24
24#include <linux/log2.h> 25#include <linux/log2.h>
25 26
@@ -40,6 +41,10 @@
40/* when under memory pressure rx ring refill may fail and needs a retry */ 41/* when under memory pressure rx ring refill may fail and needs a retry */
41#define HTT_RX_RING_REFILL_RETRY_MS 50 42#define HTT_RX_RING_REFILL_RETRY_MS 50
42 43
44
45static int ath10k_htt_rx_get_csum_state(struct sk_buff *skb);
46
47
43static int ath10k_htt_rx_ring_size(struct ath10k_htt *htt) 48static int ath10k_htt_rx_ring_size(struct ath10k_htt *htt)
44{ 49{
45 int size; 50 int size;
@@ -177,10 +182,27 @@ static int ath10k_htt_rx_ring_fill_n(struct ath10k_htt *htt, int num)
177 182
178static void ath10k_htt_rx_msdu_buff_replenish(struct ath10k_htt *htt) 183static void ath10k_htt_rx_msdu_buff_replenish(struct ath10k_htt *htt)
179{ 184{
180 int ret, num_to_fill; 185 int ret, num_deficit, num_to_fill;
181 186
187 /* Refilling the whole RX ring buffer proves to be a bad idea. The
188 * reason is RX may take up significant amount of CPU cycles and starve
189 * other tasks, e.g. TX on an ethernet device while acting as a bridge
190 * with ath10k wlan interface. This ended up with very poor performance
191 * once CPU the host system was overwhelmed with RX on ath10k.
192 *
193 * By limiting the number of refills the replenishing occurs
194 * progressively. This in turns makes use of the fact tasklets are
195 * processed in FIFO order. This means actual RX processing can starve
196 * out refilling. If there's not enough buffers on RX ring FW will not
197 * report RX until it is refilled with enough buffers. This
198 * automatically balances load wrt to CPU power.
199 *
200 * This probably comes at a cost of lower maximum throughput but
201 * improves the avarage and stability. */
182 spin_lock_bh(&htt->rx_ring.lock); 202 spin_lock_bh(&htt->rx_ring.lock);
183 num_to_fill = htt->rx_ring.fill_level - htt->rx_ring.fill_cnt; 203 num_deficit = htt->rx_ring.fill_level - htt->rx_ring.fill_cnt;
204 num_to_fill = min(ATH10K_HTT_MAX_NUM_REFILL, num_deficit);
205 num_deficit -= num_to_fill;
184 ret = ath10k_htt_rx_ring_fill_n(htt, num_to_fill); 206 ret = ath10k_htt_rx_ring_fill_n(htt, num_to_fill);
185 if (ret == -ENOMEM) { 207 if (ret == -ENOMEM) {
186 /* 208 /*
@@ -191,6 +213,8 @@ static void ath10k_htt_rx_msdu_buff_replenish(struct ath10k_htt *htt)
191 */ 213 */
192 mod_timer(&htt->rx_ring.refill_retry_timer, jiffies + 214 mod_timer(&htt->rx_ring.refill_retry_timer, jiffies +
193 msecs_to_jiffies(HTT_RX_RING_REFILL_RETRY_MS)); 215 msecs_to_jiffies(HTT_RX_RING_REFILL_RETRY_MS));
216 } else if (num_deficit > 0) {
217 tasklet_schedule(&htt->rx_replenish_task);
194 } 218 }
195 spin_unlock_bh(&htt->rx_ring.lock); 219 spin_unlock_bh(&htt->rx_ring.lock);
196} 220}
@@ -212,6 +236,7 @@ void ath10k_htt_rx_detach(struct ath10k_htt *htt)
212 int sw_rd_idx = htt->rx_ring.sw_rd_idx.msdu_payld; 236 int sw_rd_idx = htt->rx_ring.sw_rd_idx.msdu_payld;
213 237
214 del_timer_sync(&htt->rx_ring.refill_retry_timer); 238 del_timer_sync(&htt->rx_ring.refill_retry_timer);
239 tasklet_kill(&htt->rx_replenish_task);
215 240
216 while (sw_rd_idx != __le32_to_cpu(*(htt->rx_ring.alloc_idx.vaddr))) { 241 while (sw_rd_idx != __le32_to_cpu(*(htt->rx_ring.alloc_idx.vaddr))) {
217 struct sk_buff *skb = 242 struct sk_buff *skb =
@@ -441,6 +466,12 @@ static int ath10k_htt_rx_amsdu_pop(struct ath10k_htt *htt,
441 return msdu_chaining; 466 return msdu_chaining;
442} 467}
443 468
469static void ath10k_htt_rx_replenish_task(unsigned long ptr)
470{
471 struct ath10k_htt *htt = (struct ath10k_htt *)ptr;
472 ath10k_htt_rx_msdu_buff_replenish(htt);
473}
474
444int ath10k_htt_rx_attach(struct ath10k_htt *htt) 475int ath10k_htt_rx_attach(struct ath10k_htt *htt)
445{ 476{
446 dma_addr_t paddr; 477 dma_addr_t paddr;
@@ -501,7 +532,10 @@ int ath10k_htt_rx_attach(struct ath10k_htt *htt)
501 if (__ath10k_htt_rx_ring_fill_n(htt, htt->rx_ring.fill_level)) 532 if (__ath10k_htt_rx_ring_fill_n(htt, htt->rx_ring.fill_level))
502 goto err_fill_ring; 533 goto err_fill_ring;
503 534
504 ath10k_dbg(ATH10K_DBG_HTT, "HTT RX ring size: %d, fill_level: %d\n", 535 tasklet_init(&htt->rx_replenish_task, ath10k_htt_rx_replenish_task,
536 (unsigned long)htt);
537
538 ath10k_dbg(ATH10K_DBG_BOOT, "htt rx ring size %d fill_level %d\n",
505 htt->rx_ring.size, htt->rx_ring.fill_level); 539 htt->rx_ring.size, htt->rx_ring.fill_level);
506 return 0; 540 return 0;
507 541
@@ -590,134 +624,144 @@ static bool ath10k_htt_rx_hdr_is_amsdu(struct ieee80211_hdr *hdr)
590 return false; 624 return false;
591} 625}
592 626
593static int ath10k_htt_rx_amsdu(struct ath10k_htt *htt, 627struct rfc1042_hdr {
594 struct htt_rx_info *info) 628 u8 llc_dsap;
629 u8 llc_ssap;
630 u8 llc_ctrl;
631 u8 snap_oui[3];
632 __be16 snap_type;
633} __packed;
634
635struct amsdu_subframe_hdr {
636 u8 dst[ETH_ALEN];
637 u8 src[ETH_ALEN];
638 __be16 len;
639} __packed;
640
641static void ath10k_htt_rx_amsdu(struct ath10k_htt *htt,
642 struct htt_rx_info *info)
595{ 643{
596 struct htt_rx_desc *rxd; 644 struct htt_rx_desc *rxd;
597 struct sk_buff *amsdu;
598 struct sk_buff *first; 645 struct sk_buff *first;
599 struct ieee80211_hdr *hdr;
600 struct sk_buff *skb = info->skb; 646 struct sk_buff *skb = info->skb;
601 enum rx_msdu_decap_format fmt; 647 enum rx_msdu_decap_format fmt;
602 enum htt_rx_mpdu_encrypt_type enctype; 648 enum htt_rx_mpdu_encrypt_type enctype;
649 struct ieee80211_hdr *hdr;
650 u8 hdr_buf[64], addr[ETH_ALEN], *qos;
603 unsigned int hdr_len; 651 unsigned int hdr_len;
604 int crypto_len;
605 652
606 rxd = (void *)skb->data - sizeof(*rxd); 653 rxd = (void *)skb->data - sizeof(*rxd);
607 fmt = MS(__le32_to_cpu(rxd->msdu_start.info1),
608 RX_MSDU_START_INFO1_DECAP_FORMAT);
609 enctype = MS(__le32_to_cpu(rxd->mpdu_start.info0), 654 enctype = MS(__le32_to_cpu(rxd->mpdu_start.info0),
610 RX_MPDU_START_INFO0_ENCRYPT_TYPE); 655 RX_MPDU_START_INFO0_ENCRYPT_TYPE);
611 656
612 /* FIXME: No idea what assumptions are safe here. Need logs */ 657 hdr = (struct ieee80211_hdr *)rxd->rx_hdr_status;
613 if ((fmt == RX_MSDU_DECAP_RAW && skb->next) || 658 hdr_len = ieee80211_hdrlen(hdr->frame_control);
614 (fmt == RX_MSDU_DECAP_8023_SNAP_LLC)) { 659 memcpy(hdr_buf, hdr, hdr_len);
615 ath10k_htt_rx_free_msdu_chain(skb->next); 660 hdr = (struct ieee80211_hdr *)hdr_buf;
616 skb->next = NULL;
617 return -ENOTSUPP;
618 }
619 661
620 /* A-MSDU max is a little less than 8K */ 662 /* FIXME: Hopefully this is a temporary measure.
621 amsdu = dev_alloc_skb(8*1024); 663 *
622 if (!amsdu) { 664 * Reporting individual A-MSDU subframes means each reported frame
623 ath10k_warn("A-MSDU allocation failed\n"); 665 * shares the same sequence number.
624 ath10k_htt_rx_free_msdu_chain(skb->next); 666 *
625 skb->next = NULL; 667 * mac80211 drops frames it recognizes as duplicates, i.e.
626 return -ENOMEM; 668 * retransmission flag is set and sequence number matches sequence
627 } 669 * number from a previous frame (as per IEEE 802.11-2012: 9.3.2.10
628 670 * "Duplicate detection and recovery")
629 if (fmt >= RX_MSDU_DECAP_NATIVE_WIFI) { 671 *
630 int hdrlen; 672 * To avoid frames being dropped clear retransmission flag for all
631 673 * received A-MSDUs.
632 hdr = (void *)rxd->rx_hdr_status; 674 *
633 hdrlen = ieee80211_hdrlen(hdr->frame_control); 675 * Worst case: actual duplicate frames will be reported but this should
634 memcpy(skb_put(amsdu, hdrlen), hdr, hdrlen); 676 * still be handled gracefully by other OSI/ISO layers. */
635 } 677 hdr->frame_control &= cpu_to_le16(~IEEE80211_FCTL_RETRY);
636 678
637 first = skb; 679 first = skb;
638 while (skb) { 680 while (skb) {
639 void *decap_hdr; 681 void *decap_hdr;
640 int decap_len = 0; 682 int len;
641 683
642 rxd = (void *)skb->data - sizeof(*rxd); 684 rxd = (void *)skb->data - sizeof(*rxd);
643 fmt = MS(__le32_to_cpu(rxd->msdu_start.info1), 685 fmt = MS(__le32_to_cpu(rxd->msdu_start.info1),
644 RX_MSDU_START_INFO1_DECAP_FORMAT); 686 RX_MSDU_START_INFO1_DECAP_FORMAT);
645 decap_hdr = (void *)rxd->rx_hdr_status; 687 decap_hdr = (void *)rxd->rx_hdr_status;
646 688
647 if (skb == first) { 689 skb->ip_summed = ath10k_htt_rx_get_csum_state(skb);
648 /* We receive linked A-MSDU subframe skbuffs. The
649 * first one contains the original 802.11 header (and
650 * possible crypto param) in the RX descriptor. The
651 * A-MSDU subframe header follows that. Each part is
652 * aligned to 4 byte boundary. */
653
654 hdr = (void *)amsdu->data;
655 hdr_len = ieee80211_hdrlen(hdr->frame_control);
656 crypto_len = ath10k_htt_rx_crypto_param_len(enctype);
657
658 decap_hdr += roundup(hdr_len, 4);
659 decap_hdr += roundup(crypto_len, 4);
660 }
661 690
662 if (fmt == RX_MSDU_DECAP_ETHERNET2_DIX) { 691 /* First frame in an A-MSDU chain has more decapped data. */
663 /* Ethernet2 decap inserts ethernet header in place of 692 if (skb == first) {
664 * A-MSDU subframe header. */ 693 len = round_up(ieee80211_hdrlen(hdr->frame_control), 4);
665 skb_pull(skb, 6 + 6 + 2); 694 len += round_up(ath10k_htt_rx_crypto_param_len(enctype),
666 695 4);
667 /* A-MSDU subframe header length */ 696 decap_hdr += len;
668 decap_len += 6 + 6 + 2;
669
670 /* Ethernet2 decap also strips the LLC/SNAP so we need
671 * to re-insert it. The LLC/SNAP follows A-MSDU
672 * subframe header. */
673 /* FIXME: Not all LLCs are 8 bytes long */
674 decap_len += 8;
675
676 memcpy(skb_put(amsdu, decap_len), decap_hdr, decap_len);
677 } 697 }
678 698
679 if (fmt == RX_MSDU_DECAP_NATIVE_WIFI) { 699 switch (fmt) {
680 /* Native Wifi decap inserts regular 802.11 header 700 case RX_MSDU_DECAP_RAW:
681 * in place of A-MSDU subframe header. */ 701 /* remove trailing FCS */
702 skb_trim(skb, skb->len - FCS_LEN);
703 break;
704 case RX_MSDU_DECAP_NATIVE_WIFI:
705 /* pull decapped header and copy DA */
682 hdr = (struct ieee80211_hdr *)skb->data; 706 hdr = (struct ieee80211_hdr *)skb->data;
683 skb_pull(skb, ieee80211_hdrlen(hdr->frame_control)); 707 hdr_len = ieee80211_hdrlen(hdr->frame_control);
708 memcpy(addr, ieee80211_get_DA(hdr), ETH_ALEN);
709 skb_pull(skb, hdr_len);
684 710
685 /* A-MSDU subframe header length */ 711 /* push original 802.11 header */
686 decap_len += 6 + 6 + 2; 712 hdr = (struct ieee80211_hdr *)hdr_buf;
713 hdr_len = ieee80211_hdrlen(hdr->frame_control);
714 memcpy(skb_push(skb, hdr_len), hdr, hdr_len);
687 715
688 memcpy(skb_put(amsdu, decap_len), decap_hdr, decap_len); 716 /* original A-MSDU header has the bit set but we're
689 } 717 * not including A-MSDU subframe header */
718 hdr = (struct ieee80211_hdr *)skb->data;
719 qos = ieee80211_get_qos_ctl(hdr);
720 qos[0] &= ~IEEE80211_QOS_CTL_A_MSDU_PRESENT;
690 721
691 if (fmt == RX_MSDU_DECAP_RAW) 722 /* original 802.11 header has a different DA */
692 skb_trim(skb, skb->len - 4); /* remove FCS */ 723 memcpy(ieee80211_get_DA(hdr), addr, ETH_ALEN);
724 break;
725 case RX_MSDU_DECAP_ETHERNET2_DIX:
726 /* strip ethernet header and insert decapped 802.11
727 * header, amsdu subframe header and rfc1042 header */
693 728
694 memcpy(skb_put(amsdu, skb->len), skb->data, skb->len); 729 len = 0;
730 len += sizeof(struct rfc1042_hdr);
731 len += sizeof(struct amsdu_subframe_hdr);
695 732
696 /* A-MSDU subframes are padded to 4bytes 733 skb_pull(skb, sizeof(struct ethhdr));
697 * but relative to first subframe, not the whole MPDU */ 734 memcpy(skb_push(skb, len), decap_hdr, len);
698 if (skb->next && ((decap_len + skb->len) & 3)) { 735 memcpy(skb_push(skb, hdr_len), hdr, hdr_len);
699 int padlen = 4 - ((decap_len + skb->len) & 3); 736 break;
700 memset(skb_put(amsdu, padlen), 0, padlen); 737 case RX_MSDU_DECAP_8023_SNAP_LLC:
738 /* insert decapped 802.11 header making a singly
739 * A-MSDU */
740 memcpy(skb_push(skb, hdr_len), hdr, hdr_len);
741 break;
701 } 742 }
702 743
744 info->skb = skb;
745 info->encrypt_type = enctype;
703 skb = skb->next; 746 skb = skb->next;
704 } 747 info->skb->next = NULL;
705 748
706 info->skb = amsdu; 749 ath10k_process_rx(htt->ar, info);
707 info->encrypt_type = enctype; 750 }
708
709 ath10k_htt_rx_free_msdu_chain(first);
710 751
711 return 0; 752 /* FIXME: It might be nice to re-assemble the A-MSDU when there's a
753 * monitor interface active for sniffing purposes. */
712} 754}
713 755
714static int ath10k_htt_rx_msdu(struct ath10k_htt *htt, struct htt_rx_info *info) 756static void ath10k_htt_rx_msdu(struct ath10k_htt *htt, struct htt_rx_info *info)
715{ 757{
716 struct sk_buff *skb = info->skb; 758 struct sk_buff *skb = info->skb;
717 struct htt_rx_desc *rxd; 759 struct htt_rx_desc *rxd;
718 struct ieee80211_hdr *hdr; 760 struct ieee80211_hdr *hdr;
719 enum rx_msdu_decap_format fmt; 761 enum rx_msdu_decap_format fmt;
720 enum htt_rx_mpdu_encrypt_type enctype; 762 enum htt_rx_mpdu_encrypt_type enctype;
763 int hdr_len;
764 void *rfc1042;
721 765
722 /* This shouldn't happen. If it does than it may be a FW bug. */ 766 /* This shouldn't happen. If it does than it may be a FW bug. */
723 if (skb->next) { 767 if (skb->next) {
@@ -731,49 +775,53 @@ static int ath10k_htt_rx_msdu(struct ath10k_htt *htt, struct htt_rx_info *info)
731 RX_MSDU_START_INFO1_DECAP_FORMAT); 775 RX_MSDU_START_INFO1_DECAP_FORMAT);
732 enctype = MS(__le32_to_cpu(rxd->mpdu_start.info0), 776 enctype = MS(__le32_to_cpu(rxd->mpdu_start.info0),
733 RX_MPDU_START_INFO0_ENCRYPT_TYPE); 777 RX_MPDU_START_INFO0_ENCRYPT_TYPE);
734 hdr = (void *)skb->data - RX_HTT_HDR_STATUS_LEN; 778 hdr = (struct ieee80211_hdr *)rxd->rx_hdr_status;
779 hdr_len = ieee80211_hdrlen(hdr->frame_control);
780
781 skb->ip_summed = ath10k_htt_rx_get_csum_state(skb);
735 782
736 switch (fmt) { 783 switch (fmt) {
737 case RX_MSDU_DECAP_RAW: 784 case RX_MSDU_DECAP_RAW:
738 /* remove trailing FCS */ 785 /* remove trailing FCS */
739 skb_trim(skb, skb->len - 4); 786 skb_trim(skb, skb->len - FCS_LEN);
740 break; 787 break;
741 case RX_MSDU_DECAP_NATIVE_WIFI: 788 case RX_MSDU_DECAP_NATIVE_WIFI:
742 /* nothing to do here */ 789 /* Pull decapped header */
790 hdr = (struct ieee80211_hdr *)skb->data;
791 hdr_len = ieee80211_hdrlen(hdr->frame_control);
792 skb_pull(skb, hdr_len);
793
794 /* Push original header */
795 hdr = (struct ieee80211_hdr *)rxd->rx_hdr_status;
796 hdr_len = ieee80211_hdrlen(hdr->frame_control);
797 memcpy(skb_push(skb, hdr_len), hdr, hdr_len);
743 break; 798 break;
744 case RX_MSDU_DECAP_ETHERNET2_DIX: 799 case RX_MSDU_DECAP_ETHERNET2_DIX:
745 /* macaddr[6] + macaddr[6] + ethertype[2] */ 800 /* strip ethernet header and insert decapped 802.11 header and
746 skb_pull(skb, 6 + 6 + 2); 801 * rfc1042 header */
747 break;
748 case RX_MSDU_DECAP_8023_SNAP_LLC:
749 /* macaddr[6] + macaddr[6] + len[2] */
750 /* we don't need this for non-A-MSDU */
751 skb_pull(skb, 6 + 6 + 2);
752 break;
753 }
754 802
755 if (fmt == RX_MSDU_DECAP_ETHERNET2_DIX) { 803 rfc1042 = hdr;
756 void *llc; 804 rfc1042 += roundup(hdr_len, 4);
757 int llclen; 805 rfc1042 += roundup(ath10k_htt_rx_crypto_param_len(enctype), 4);
758 806
759 llclen = 8; 807 skb_pull(skb, sizeof(struct ethhdr));
760 llc = hdr; 808 memcpy(skb_push(skb, sizeof(struct rfc1042_hdr)),
761 llc += roundup(ieee80211_hdrlen(hdr->frame_control), 4); 809 rfc1042, sizeof(struct rfc1042_hdr));
762 llc += roundup(ath10k_htt_rx_crypto_param_len(enctype), 4); 810 memcpy(skb_push(skb, hdr_len), hdr, hdr_len);
763 811 break;
764 skb_push(skb, llclen); 812 case RX_MSDU_DECAP_8023_SNAP_LLC:
765 memcpy(skb->data, llc, llclen); 813 /* remove A-MSDU subframe header and insert
766 } 814 * decapped 802.11 header. rfc1042 header is already there */
767 815
768 if (fmt >= RX_MSDU_DECAP_ETHERNET2_DIX) { 816 skb_pull(skb, sizeof(struct amsdu_subframe_hdr));
769 int len = ieee80211_hdrlen(hdr->frame_control); 817 memcpy(skb_push(skb, hdr_len), hdr, hdr_len);
770 skb_push(skb, len); 818 break;
771 memcpy(skb->data, hdr, len);
772 } 819 }
773 820
774 info->skb = skb; 821 info->skb = skb;
775 info->encrypt_type = enctype; 822 info->encrypt_type = enctype;
776 return 0; 823
824 ath10k_process_rx(htt->ar, info);
777} 825}
778 826
779static bool ath10k_htt_rx_has_decrypt_err(struct sk_buff *skb) 827static bool ath10k_htt_rx_has_decrypt_err(struct sk_buff *skb)
@@ -845,8 +893,6 @@ static void ath10k_htt_rx_handler(struct ath10k_htt *htt,
845 int fw_desc_len; 893 int fw_desc_len;
846 u8 *fw_desc; 894 u8 *fw_desc;
847 int i, j; 895 int i, j;
848 int ret;
849 int ip_summed;
850 896
851 memset(&info, 0, sizeof(info)); 897 memset(&info, 0, sizeof(info));
852 898
@@ -921,11 +967,6 @@ static void ath10k_htt_rx_handler(struct ath10k_htt *htt,
921 continue; 967 continue;
922 } 968 }
923 969
924 /* The skb is not yet processed and it may be
925 * reallocated. Since the offload is in the original
926 * skb extract the checksum now and assign it later */
927 ip_summed = ath10k_htt_rx_get_csum_state(msdu_head);
928
929 info.skb = msdu_head; 970 info.skb = msdu_head;
930 info.fcs_err = ath10k_htt_rx_has_fcs_err(msdu_head); 971 info.fcs_err = ath10k_htt_rx_has_fcs_err(msdu_head);
931 info.signal = ATH10K_DEFAULT_NOISE_FLOOR; 972 info.signal = ATH10K_DEFAULT_NOISE_FLOOR;
@@ -938,28 +979,13 @@ static void ath10k_htt_rx_handler(struct ath10k_htt *htt,
938 hdr = ath10k_htt_rx_skb_get_hdr(msdu_head); 979 hdr = ath10k_htt_rx_skb_get_hdr(msdu_head);
939 980
940 if (ath10k_htt_rx_hdr_is_amsdu(hdr)) 981 if (ath10k_htt_rx_hdr_is_amsdu(hdr))
941 ret = ath10k_htt_rx_amsdu(htt, &info); 982 ath10k_htt_rx_amsdu(htt, &info);
942 else 983 else
943 ret = ath10k_htt_rx_msdu(htt, &info); 984 ath10k_htt_rx_msdu(htt, &info);
944
945 if (ret && !info.fcs_err) {
946 ath10k_warn("error processing msdus %d\n", ret);
947 dev_kfree_skb_any(info.skb);
948 continue;
949 }
950
951 if (ath10k_htt_rx_hdr_is_amsdu((void *)info.skb->data))
952 ath10k_dbg(ATH10K_DBG_HTT, "htt mpdu is amsdu\n");
953
954 info.skb->ip_summed = ip_summed;
955
956 ath10k_dbg_dump(ATH10K_DBG_HTT_DUMP, NULL, "htt mpdu: ",
957 info.skb->data, info.skb->len);
958 ath10k_process_rx(htt->ar, &info);
959 } 985 }
960 } 986 }
961 987
962 ath10k_htt_rx_msdu_buff_replenish(htt); 988 tasklet_schedule(&htt->rx_replenish_task);
963} 989}
964 990
965static void ath10k_htt_rx_frag_handler(struct ath10k_htt *htt, 991static void ath10k_htt_rx_frag_handler(struct ath10k_htt *htt,
@@ -1131,7 +1157,7 @@ void ath10k_htt_t2h_msg_handler(struct ath10k *ar, struct sk_buff *skb)
1131 break; 1157 break;
1132 } 1158 }
1133 1159
1134 ath10k_txrx_tx_completed(htt, &tx_done); 1160 ath10k_txrx_tx_unref(htt, &tx_done);
1135 break; 1161 break;
1136 } 1162 }
1137 case HTT_T2H_MSG_TYPE_TX_COMPL_IND: { 1163 case HTT_T2H_MSG_TYPE_TX_COMPL_IND: {
@@ -1165,7 +1191,7 @@ void ath10k_htt_t2h_msg_handler(struct ath10k *ar, struct sk_buff *skb)
1165 for (i = 0; i < resp->data_tx_completion.num_msdus; i++) { 1191 for (i = 0; i < resp->data_tx_completion.num_msdus; i++) {
1166 msdu_id = resp->data_tx_completion.msdus[i]; 1192 msdu_id = resp->data_tx_completion.msdus[i];
1167 tx_done.msdu_id = __le16_to_cpu(msdu_id); 1193 tx_done.msdu_id = __le16_to_cpu(msdu_id);
1168 ath10k_txrx_tx_completed(htt, &tx_done); 1194 ath10k_txrx_tx_unref(htt, &tx_done);
1169 } 1195 }
1170 break; 1196 break;
1171 } 1197 }
@@ -1190,8 +1216,10 @@ void ath10k_htt_t2h_msg_handler(struct ath10k *ar, struct sk_buff *skb)
1190 case HTT_T2H_MSG_TYPE_TEST: 1216 case HTT_T2H_MSG_TYPE_TEST:
1191 /* FIX THIS */ 1217 /* FIX THIS */
1192 break; 1218 break;
1193 case HTT_T2H_MSG_TYPE_TX_INSPECT_IND:
1194 case HTT_T2H_MSG_TYPE_STATS_CONF: 1219 case HTT_T2H_MSG_TYPE_STATS_CONF:
1220 trace_ath10k_htt_stats(skb->data, skb->len);
1221 break;
1222 case HTT_T2H_MSG_TYPE_TX_INSPECT_IND:
1195 case HTT_T2H_MSG_TYPE_RX_ADDBA: 1223 case HTT_T2H_MSG_TYPE_RX_ADDBA:
1196 case HTT_T2H_MSG_TYPE_RX_DELBA: 1224 case HTT_T2H_MSG_TYPE_RX_DELBA:
1197 case HTT_T2H_MSG_TYPE_RX_FLUSH: 1225 case HTT_T2H_MSG_TYPE_RX_FLUSH:
diff --git a/drivers/net/wireless/ath/ath10k/htt_tx.c b/drivers/net/wireless/ath/ath10k/htt_tx.c
index 656c2546b294..d9335e9d0d04 100644
--- a/drivers/net/wireless/ath/ath10k/htt_tx.c
+++ b/drivers/net/wireless/ath/ath10k/htt_tx.c
@@ -96,7 +96,7 @@ int ath10k_htt_tx_attach(struct ath10k_htt *htt)
96 htt->max_num_pending_tx = ath10k_hif_get_free_queue_number(htt->ar, 96 htt->max_num_pending_tx = ath10k_hif_get_free_queue_number(htt->ar,
97 pipe); 97 pipe);
98 98
99 ath10k_dbg(ATH10K_DBG_HTT, "htt tx max num pending tx %d\n", 99 ath10k_dbg(ATH10K_DBG_BOOT, "htt tx max num pending tx %d\n",
100 htt->max_num_pending_tx); 100 htt->max_num_pending_tx);
101 101
102 htt->pending_tx = kzalloc(sizeof(*htt->pending_tx) * 102 htt->pending_tx = kzalloc(sizeof(*htt->pending_tx) *
@@ -117,7 +117,7 @@ int ath10k_htt_tx_attach(struct ath10k_htt *htt)
117 117
118static void ath10k_htt_tx_cleanup_pending(struct ath10k_htt *htt) 118static void ath10k_htt_tx_cleanup_pending(struct ath10k_htt *htt)
119{ 119{
120 struct sk_buff *txdesc; 120 struct htt_tx_done tx_done = {0};
121 int msdu_id; 121 int msdu_id;
122 122
123 /* No locks needed. Called after communication with the device has 123 /* No locks needed. Called after communication with the device has
@@ -127,18 +127,13 @@ static void ath10k_htt_tx_cleanup_pending(struct ath10k_htt *htt)
127 if (!test_bit(msdu_id, htt->used_msdu_ids)) 127 if (!test_bit(msdu_id, htt->used_msdu_ids))
128 continue; 128 continue;
129 129
130 txdesc = htt->pending_tx[msdu_id];
131 if (!txdesc)
132 continue;
133
134 ath10k_dbg(ATH10K_DBG_HTT, "force cleanup msdu_id %hu\n", 130 ath10k_dbg(ATH10K_DBG_HTT, "force cleanup msdu_id %hu\n",
135 msdu_id); 131 msdu_id);
136 132
137 if (ATH10K_SKB_CB(txdesc)->htt.refcount > 0) 133 tx_done.discard = 1;
138 ATH10K_SKB_CB(txdesc)->htt.refcount = 1; 134 tx_done.msdu_id = msdu_id;
139 135
140 ATH10K_SKB_CB(txdesc)->htt.discard = true; 136 ath10k_txrx_tx_unref(htt, &tx_done);
141 ath10k_txrx_tx_unref(htt, txdesc);
142 } 137 }
143} 138}
144 139
@@ -152,26 +147,7 @@ void ath10k_htt_tx_detach(struct ath10k_htt *htt)
152 147
153void ath10k_htt_htc_tx_complete(struct ath10k *ar, struct sk_buff *skb) 148void ath10k_htt_htc_tx_complete(struct ath10k *ar, struct sk_buff *skb)
154{ 149{
155 struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(skb); 150 dev_kfree_skb_any(skb);
156 struct ath10k_htt *htt = &ar->htt;
157
158 if (skb_cb->htt.is_conf) {
159 dev_kfree_skb_any(skb);
160 return;
161 }
162
163 if (skb_cb->is_aborted) {
164 skb_cb->htt.discard = true;
165
166 /* if the skbuff is aborted we need to make sure we'll free up
167 * the tx resources, we can't simply run tx_unref() 2 times
168 * because if htt tx completion came in earlier we'd access
169 * unallocated memory */
170 if (skb_cb->htt.refcount > 1)
171 skb_cb->htt.refcount = 1;
172 }
173
174 ath10k_txrx_tx_unref(htt, skb);
175} 151}
176 152
177int ath10k_htt_h2t_ver_req_msg(struct ath10k_htt *htt) 153int ath10k_htt_h2t_ver_req_msg(struct ath10k_htt *htt)
@@ -192,10 +168,48 @@ int ath10k_htt_h2t_ver_req_msg(struct ath10k_htt *htt)
192 cmd = (struct htt_cmd *)skb->data; 168 cmd = (struct htt_cmd *)skb->data;
193 cmd->hdr.msg_type = HTT_H2T_MSG_TYPE_VERSION_REQ; 169 cmd->hdr.msg_type = HTT_H2T_MSG_TYPE_VERSION_REQ;
194 170
195 ATH10K_SKB_CB(skb)->htt.is_conf = true; 171 ret = ath10k_htc_send(&htt->ar->htc, htt->eid, skb);
172 if (ret) {
173 dev_kfree_skb_any(skb);
174 return ret;
175 }
176
177 return 0;
178}
179
180int ath10k_htt_h2t_stats_req(struct ath10k_htt *htt, u8 mask, u64 cookie)
181{
182 struct htt_stats_req *req;
183 struct sk_buff *skb;
184 struct htt_cmd *cmd;
185 int len = 0, ret;
186
187 len += sizeof(cmd->hdr);
188 len += sizeof(cmd->stats_req);
189
190 skb = ath10k_htc_alloc_skb(len);
191 if (!skb)
192 return -ENOMEM;
193
194 skb_put(skb, len);
195 cmd = (struct htt_cmd *)skb->data;
196 cmd->hdr.msg_type = HTT_H2T_MSG_TYPE_STATS_REQ;
197
198 req = &cmd->stats_req;
199
200 memset(req, 0, sizeof(*req));
201
202 /* currently we support only max 8 bit masks so no need to worry
203 * about endian support */
204 req->upload_types[0] = mask;
205 req->reset_types[0] = mask;
206 req->stat_type = HTT_STATS_REQ_CFG_STAT_TYPE_INVALID;
207 req->cookie_lsb = cpu_to_le32(cookie & 0xffffffff);
208 req->cookie_msb = cpu_to_le32((cookie & 0xffffffff00000000ULL) >> 32);
196 209
197 ret = ath10k_htc_send(&htt->ar->htc, htt->eid, skb); 210 ret = ath10k_htc_send(&htt->ar->htc, htt->eid, skb);
198 if (ret) { 211 if (ret) {
212 ath10k_warn("failed to send htt type stats request: %d", ret);
199 dev_kfree_skb_any(skb); 213 dev_kfree_skb_any(skb);
200 return ret; 214 return ret;
201 } 215 }
@@ -279,8 +293,6 @@ int ath10k_htt_send_rx_ring_cfg_ll(struct ath10k_htt *htt)
279 293
280#undef desc_offset 294#undef desc_offset
281 295
282 ATH10K_SKB_CB(skb)->htt.is_conf = true;
283
284 ret = ath10k_htc_send(&htt->ar->htc, htt->eid, skb); 296 ret = ath10k_htc_send(&htt->ar->htc, htt->eid, skb);
285 if (ret) { 297 if (ret) {
286 dev_kfree_skb_any(skb); 298 dev_kfree_skb_any(skb);
@@ -293,10 +305,10 @@ int ath10k_htt_send_rx_ring_cfg_ll(struct ath10k_htt *htt)
293int ath10k_htt_mgmt_tx(struct ath10k_htt *htt, struct sk_buff *msdu) 305int ath10k_htt_mgmt_tx(struct ath10k_htt *htt, struct sk_buff *msdu)
294{ 306{
295 struct device *dev = htt->ar->dev; 307 struct device *dev = htt->ar->dev;
296 struct ath10k_skb_cb *skb_cb;
297 struct sk_buff *txdesc = NULL; 308 struct sk_buff *txdesc = NULL;
298 struct htt_cmd *cmd; 309 struct htt_cmd *cmd;
299 u8 vdev_id = ATH10K_SKB_CB(msdu)->htt.vdev_id; 310 struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(msdu);
311 u8 vdev_id = skb_cb->vdev_id;
300 int len = 0; 312 int len = 0;
301 int msdu_id = -1; 313 int msdu_id = -1;
302 int res; 314 int res;
@@ -304,30 +316,30 @@ int ath10k_htt_mgmt_tx(struct ath10k_htt *htt, struct sk_buff *msdu)
304 316
305 res = ath10k_htt_tx_inc_pending(htt); 317 res = ath10k_htt_tx_inc_pending(htt);
306 if (res) 318 if (res)
307 return res; 319 goto err;
308 320
309 len += sizeof(cmd->hdr); 321 len += sizeof(cmd->hdr);
310 len += sizeof(cmd->mgmt_tx); 322 len += sizeof(cmd->mgmt_tx);
311 323
312 txdesc = ath10k_htc_alloc_skb(len);
313 if (!txdesc) {
314 res = -ENOMEM;
315 goto err;
316 }
317
318 spin_lock_bh(&htt->tx_lock); 324 spin_lock_bh(&htt->tx_lock);
319 msdu_id = ath10k_htt_tx_alloc_msdu_id(htt); 325 res = ath10k_htt_tx_alloc_msdu_id(htt);
320 if (msdu_id < 0) { 326 if (res < 0) {
321 spin_unlock_bh(&htt->tx_lock); 327 spin_unlock_bh(&htt->tx_lock);
322 res = msdu_id; 328 goto err_tx_dec;
323 goto err;
324 } 329 }
325 htt->pending_tx[msdu_id] = txdesc; 330 msdu_id = res;
331 htt->pending_tx[msdu_id] = msdu;
326 spin_unlock_bh(&htt->tx_lock); 332 spin_unlock_bh(&htt->tx_lock);
327 333
334 txdesc = ath10k_htc_alloc_skb(len);
335 if (!txdesc) {
336 res = -ENOMEM;
337 goto err_free_msdu_id;
338 }
339
328 res = ath10k_skb_map(dev, msdu); 340 res = ath10k_skb_map(dev, msdu);
329 if (res) 341 if (res)
330 goto err; 342 goto err_free_txdesc;
331 343
332 skb_put(txdesc, len); 344 skb_put(txdesc, len);
333 cmd = (struct htt_cmd *)txdesc->data; 345 cmd = (struct htt_cmd *)txdesc->data;
@@ -339,31 +351,27 @@ int ath10k_htt_mgmt_tx(struct ath10k_htt *htt, struct sk_buff *msdu)
339 memcpy(cmd->mgmt_tx.hdr, msdu->data, 351 memcpy(cmd->mgmt_tx.hdr, msdu->data,
340 min_t(int, msdu->len, HTT_MGMT_FRM_HDR_DOWNLOAD_LEN)); 352 min_t(int, msdu->len, HTT_MGMT_FRM_HDR_DOWNLOAD_LEN));
341 353
342 /* refcount is decremented by HTC and HTT completions until it reaches 354 skb_cb->htt.frag_len = 0;
343 * zero and is freed */ 355 skb_cb->htt.pad_len = 0;
344 skb_cb = ATH10K_SKB_CB(txdesc);
345 skb_cb->htt.msdu_id = msdu_id;
346 skb_cb->htt.refcount = 2;
347 skb_cb->htt.msdu = msdu;
348 356
349 res = ath10k_htc_send(&htt->ar->htc, htt->eid, txdesc); 357 res = ath10k_htc_send(&htt->ar->htc, htt->eid, txdesc);
350 if (res) 358 if (res)
351 goto err; 359 goto err_unmap_msdu;
352 360
353 return 0; 361 return 0;
354 362
355err: 363err_unmap_msdu:
356 ath10k_skb_unmap(dev, msdu); 364 ath10k_skb_unmap(dev, msdu);
357 365err_free_txdesc:
358 if (txdesc) 366 dev_kfree_skb_any(txdesc);
359 dev_kfree_skb_any(txdesc); 367err_free_msdu_id:
360 if (msdu_id >= 0) { 368 spin_lock_bh(&htt->tx_lock);
361 spin_lock_bh(&htt->tx_lock); 369 htt->pending_tx[msdu_id] = NULL;
362 htt->pending_tx[msdu_id] = NULL; 370 ath10k_htt_tx_free_msdu_id(htt, msdu_id);
363 ath10k_htt_tx_free_msdu_id(htt, msdu_id); 371 spin_unlock_bh(&htt->tx_lock);
364 spin_unlock_bh(&htt->tx_lock); 372err_tx_dec:
365 }
366 ath10k_htt_tx_dec_pending(htt); 373 ath10k_htt_tx_dec_pending(htt);
374err:
367 return res; 375 return res;
368} 376}
369 377
@@ -373,13 +381,12 @@ int ath10k_htt_tx(struct ath10k_htt *htt, struct sk_buff *msdu)
373 struct htt_cmd *cmd; 381 struct htt_cmd *cmd;
374 struct htt_data_tx_desc_frag *tx_frags; 382 struct htt_data_tx_desc_frag *tx_frags;
375 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)msdu->data; 383 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)msdu->data;
376 struct ath10k_skb_cb *skb_cb; 384 struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(msdu);
377 struct sk_buff *txdesc = NULL; 385 struct sk_buff *txdesc = NULL;
378 struct sk_buff *txfrag = NULL; 386 bool use_frags;
379 u8 vdev_id = ATH10K_SKB_CB(msdu)->htt.vdev_id; 387 u8 vdev_id = ATH10K_SKB_CB(msdu)->vdev_id;
380 u8 tid; 388 u8 tid;
381 int prefetch_len, desc_len, frag_len; 389 int prefetch_len, desc_len;
382 dma_addr_t frags_paddr;
383 int msdu_id = -1; 390 int msdu_id = -1;
384 int res; 391 int res;
385 u8 flags0; 392 u8 flags0;
@@ -387,69 +394,82 @@ int ath10k_htt_tx(struct ath10k_htt *htt, struct sk_buff *msdu)
387 394
388 res = ath10k_htt_tx_inc_pending(htt); 395 res = ath10k_htt_tx_inc_pending(htt);
389 if (res) 396 if (res)
390 return res; 397 goto err;
398
399 spin_lock_bh(&htt->tx_lock);
400 res = ath10k_htt_tx_alloc_msdu_id(htt);
401 if (res < 0) {
402 spin_unlock_bh(&htt->tx_lock);
403 goto err_tx_dec;
404 }
405 msdu_id = res;
406 htt->pending_tx[msdu_id] = msdu;
407 spin_unlock_bh(&htt->tx_lock);
391 408
392 prefetch_len = min(htt->prefetch_len, msdu->len); 409 prefetch_len = min(htt->prefetch_len, msdu->len);
393 prefetch_len = roundup(prefetch_len, 4); 410 prefetch_len = roundup(prefetch_len, 4);
394 411
395 desc_len = sizeof(cmd->hdr) + sizeof(cmd->data_tx) + prefetch_len; 412 desc_len = sizeof(cmd->hdr) + sizeof(cmd->data_tx) + prefetch_len;
396 frag_len = sizeof(*tx_frags) * 2;
397 413
398 txdesc = ath10k_htc_alloc_skb(desc_len); 414 txdesc = ath10k_htc_alloc_skb(desc_len);
399 if (!txdesc) { 415 if (!txdesc) {
400 res = -ENOMEM; 416 res = -ENOMEM;
401 goto err; 417 goto err_free_msdu_id;
402 } 418 }
403 419
404 txfrag = dev_alloc_skb(frag_len); 420 /* Since HTT 3.0 there is no separate mgmt tx command. However in case
405 if (!txfrag) { 421 * of mgmt tx using TX_FRM there is not tx fragment list. Instead of tx
406 res = -ENOMEM; 422 * fragment list host driver specifies directly frame pointer. */
407 goto err; 423 use_frags = htt->target_version_major < 3 ||
408 } 424 !ieee80211_is_mgmt(hdr->frame_control);
409 425
410 if (!IS_ALIGNED((unsigned long)txdesc->data, 4)) { 426 if (!IS_ALIGNED((unsigned long)txdesc->data, 4)) {
411 ath10k_warn("htt alignment check failed. dropping packet.\n"); 427 ath10k_warn("htt alignment check failed. dropping packet.\n");
412 res = -EIO; 428 res = -EIO;
413 goto err; 429 goto err_free_txdesc;
414 } 430 }
415 431
416 spin_lock_bh(&htt->tx_lock); 432 if (use_frags) {
417 msdu_id = ath10k_htt_tx_alloc_msdu_id(htt); 433 skb_cb->htt.frag_len = sizeof(*tx_frags) * 2;
418 if (msdu_id < 0) { 434 skb_cb->htt.pad_len = (unsigned long)msdu->data -
419 spin_unlock_bh(&htt->tx_lock); 435 round_down((unsigned long)msdu->data, 4);
420 res = msdu_id; 436
421 goto err; 437 skb_push(msdu, skb_cb->htt.frag_len + skb_cb->htt.pad_len);
438 } else {
439 skb_cb->htt.frag_len = 0;
440 skb_cb->htt.pad_len = 0;
422 } 441 }
423 htt->pending_tx[msdu_id] = txdesc;
424 spin_unlock_bh(&htt->tx_lock);
425 442
426 res = ath10k_skb_map(dev, msdu); 443 res = ath10k_skb_map(dev, msdu);
427 if (res) 444 if (res)
428 goto err; 445 goto err_pull_txfrag;
429 446
430 /* tx fragment list must be terminated with zero-entry */ 447 if (use_frags) {
431 skb_put(txfrag, frag_len); 448 dma_sync_single_for_cpu(dev, skb_cb->paddr, msdu->len,
432 tx_frags = (struct htt_data_tx_desc_frag *)txfrag->data; 449 DMA_TO_DEVICE);
433 tx_frags[0].paddr = __cpu_to_le32(ATH10K_SKB_CB(msdu)->paddr); 450
434 tx_frags[0].len = __cpu_to_le32(msdu->len); 451 /* tx fragment list must be terminated with zero-entry */
435 tx_frags[1].paddr = __cpu_to_le32(0); 452 tx_frags = (struct htt_data_tx_desc_frag *)msdu->data;
436 tx_frags[1].len = __cpu_to_le32(0); 453 tx_frags[0].paddr = __cpu_to_le32(skb_cb->paddr +
437 454 skb_cb->htt.frag_len +
438 res = ath10k_skb_map(dev, txfrag); 455 skb_cb->htt.pad_len);
439 if (res) 456 tx_frags[0].len = __cpu_to_le32(msdu->len -
440 goto err; 457 skb_cb->htt.frag_len -
458 skb_cb->htt.pad_len);
459 tx_frags[1].paddr = __cpu_to_le32(0);
460 tx_frags[1].len = __cpu_to_le32(0);
461
462 dma_sync_single_for_device(dev, skb_cb->paddr, msdu->len,
463 DMA_TO_DEVICE);
464 }
441 465
442 ath10k_dbg(ATH10K_DBG_HTT, "txfrag 0x%llx msdu 0x%llx\n", 466 ath10k_dbg(ATH10K_DBG_HTT, "msdu 0x%llx\n",
443 (unsigned long long) ATH10K_SKB_CB(txfrag)->paddr,
444 (unsigned long long) ATH10K_SKB_CB(msdu)->paddr); 467 (unsigned long long) ATH10K_SKB_CB(msdu)->paddr);
445 ath10k_dbg_dump(ATH10K_DBG_HTT_DUMP, NULL, "txfrag: ",
446 txfrag->data, frag_len);
447 ath10k_dbg_dump(ATH10K_DBG_HTT_DUMP, NULL, "msdu: ", 468 ath10k_dbg_dump(ATH10K_DBG_HTT_DUMP, NULL, "msdu: ",
448 msdu->data, msdu->len); 469 msdu->data, msdu->len);
449 470
450 skb_put(txdesc, desc_len); 471 skb_put(txdesc, desc_len);
451 cmd = (struct htt_cmd *)txdesc->data; 472 cmd = (struct htt_cmd *)txdesc->data;
452 memset(cmd, 0, desc_len);
453 473
454 tid = ATH10K_SKB_CB(msdu)->htt.tid; 474 tid = ATH10K_SKB_CB(msdu)->htt.tid;
455 475
@@ -459,8 +479,13 @@ int ath10k_htt_tx(struct ath10k_htt *htt, struct sk_buff *msdu)
459 if (!ieee80211_has_protected(hdr->frame_control)) 479 if (!ieee80211_has_protected(hdr->frame_control))
460 flags0 |= HTT_DATA_TX_DESC_FLAGS0_NO_ENCRYPT; 480 flags0 |= HTT_DATA_TX_DESC_FLAGS0_NO_ENCRYPT;
461 flags0 |= HTT_DATA_TX_DESC_FLAGS0_MAC_HDR_PRESENT; 481 flags0 |= HTT_DATA_TX_DESC_FLAGS0_MAC_HDR_PRESENT;
462 flags0 |= SM(ATH10K_HW_TXRX_NATIVE_WIFI, 482
463 HTT_DATA_TX_DESC_FLAGS0_PKT_TYPE); 483 if (use_frags)
484 flags0 |= SM(ATH10K_HW_TXRX_NATIVE_WIFI,
485 HTT_DATA_TX_DESC_FLAGS0_PKT_TYPE);
486 else
487 flags0 |= SM(ATH10K_HW_TXRX_MGMT,
488 HTT_DATA_TX_DESC_FLAGS0_PKT_TYPE);
464 489
465 flags1 = 0; 490 flags1 = 0;
466 flags1 |= SM((u16)vdev_id, HTT_DATA_TX_DESC_FLAGS1_VDEV_ID); 491 flags1 |= SM((u16)vdev_id, HTT_DATA_TX_DESC_FLAGS1_VDEV_ID);
@@ -468,45 +493,37 @@ int ath10k_htt_tx(struct ath10k_htt *htt, struct sk_buff *msdu)
468 flags1 |= HTT_DATA_TX_DESC_FLAGS1_CKSUM_L3_OFFLOAD; 493 flags1 |= HTT_DATA_TX_DESC_FLAGS1_CKSUM_L3_OFFLOAD;
469 flags1 |= HTT_DATA_TX_DESC_FLAGS1_CKSUM_L4_OFFLOAD; 494 flags1 |= HTT_DATA_TX_DESC_FLAGS1_CKSUM_L4_OFFLOAD;
470 495
471 frags_paddr = ATH10K_SKB_CB(txfrag)->paddr;
472
473 cmd->hdr.msg_type = HTT_H2T_MSG_TYPE_TX_FRM; 496 cmd->hdr.msg_type = HTT_H2T_MSG_TYPE_TX_FRM;
474 cmd->data_tx.flags0 = flags0; 497 cmd->data_tx.flags0 = flags0;
475 cmd->data_tx.flags1 = __cpu_to_le16(flags1); 498 cmd->data_tx.flags1 = __cpu_to_le16(flags1);
476 cmd->data_tx.len = __cpu_to_le16(msdu->len); 499 cmd->data_tx.len = __cpu_to_le16(msdu->len -
500 skb_cb->htt.frag_len -
501 skb_cb->htt.pad_len);
477 cmd->data_tx.id = __cpu_to_le16(msdu_id); 502 cmd->data_tx.id = __cpu_to_le16(msdu_id);
478 cmd->data_tx.frags_paddr = __cpu_to_le32(frags_paddr); 503 cmd->data_tx.frags_paddr = __cpu_to_le32(skb_cb->paddr);
479 cmd->data_tx.peerid = __cpu_to_le32(HTT_INVALID_PEERID); 504 cmd->data_tx.peerid = __cpu_to_le32(HTT_INVALID_PEERID);
480 505
481 memcpy(cmd->data_tx.prefetch, msdu->data, prefetch_len); 506 memcpy(cmd->data_tx.prefetch, hdr, prefetch_len);
482
483 /* refcount is decremented by HTC and HTT completions until it reaches
484 * zero and is freed */
485 skb_cb = ATH10K_SKB_CB(txdesc);
486 skb_cb->htt.msdu_id = msdu_id;
487 skb_cb->htt.refcount = 2;
488 skb_cb->htt.txfrag = txfrag;
489 skb_cb->htt.msdu = msdu;
490 507
491 res = ath10k_htc_send(&htt->ar->htc, htt->eid, txdesc); 508 res = ath10k_htc_send(&htt->ar->htc, htt->eid, txdesc);
492 if (res) 509 if (res)
493 goto err; 510 goto err_unmap_msdu;
494 511
495 return 0; 512 return 0;
496err: 513
497 if (txfrag) 514err_unmap_msdu:
498 ath10k_skb_unmap(dev, txfrag);
499 if (txdesc)
500 dev_kfree_skb_any(txdesc);
501 if (txfrag)
502 dev_kfree_skb_any(txfrag);
503 if (msdu_id >= 0) {
504 spin_lock_bh(&htt->tx_lock);
505 htt->pending_tx[msdu_id] = NULL;
506 ath10k_htt_tx_free_msdu_id(htt, msdu_id);
507 spin_unlock_bh(&htt->tx_lock);
508 }
509 ath10k_htt_tx_dec_pending(htt);
510 ath10k_skb_unmap(dev, msdu); 515 ath10k_skb_unmap(dev, msdu);
516err_pull_txfrag:
517 skb_pull(msdu, skb_cb->htt.frag_len + skb_cb->htt.pad_len);
518err_free_txdesc:
519 dev_kfree_skb_any(txdesc);
520err_free_msdu_id:
521 spin_lock_bh(&htt->tx_lock);
522 htt->pending_tx[msdu_id] = NULL;
523 ath10k_htt_tx_free_msdu_id(htt, msdu_id);
524 spin_unlock_bh(&htt->tx_lock);
525err_tx_dec:
526 ath10k_htt_tx_dec_pending(htt);
527err:
511 return res; 528 return res;
512} 529}
diff --git a/drivers/net/wireless/ath/ath10k/hw.h b/drivers/net/wireless/ath/ath10k/hw.h
index 44ed5af0a204..8aeb46d9b534 100644
--- a/drivers/net/wireless/ath/ath10k/hw.h
+++ b/drivers/net/wireless/ath/ath10k/hw.h
@@ -20,28 +20,37 @@
20 20
21#include "targaddrs.h" 21#include "targaddrs.h"
22 22
23/* Supported FW version */ 23/* QCA988X 1.0 definitions (unsupported) */
24#define SUPPORTED_FW_MAJOR 1 24#define QCA988X_HW_1_0_CHIP_ID_REV 0x0
25#define SUPPORTED_FW_MINOR 0
26#define SUPPORTED_FW_RELEASE 0
27#define SUPPORTED_FW_BUILD 629
28
29/* QCA988X 1.0 definitions */
30#define QCA988X_HW_1_0_VERSION 0x4000002c
31#define QCA988X_HW_1_0_FW_DIR "ath10k/QCA988X/hw1.0"
32#define QCA988X_HW_1_0_FW_FILE "firmware.bin"
33#define QCA988X_HW_1_0_OTP_FILE "otp.bin"
34#define QCA988X_HW_1_0_BOARD_DATA_FILE "board.bin"
35#define QCA988X_HW_1_0_PATCH_LOAD_ADDR 0x1234
36 25
37/* QCA988X 2.0 definitions */ 26/* QCA988X 2.0 definitions */
38#define QCA988X_HW_2_0_VERSION 0x4100016c 27#define QCA988X_HW_2_0_VERSION 0x4100016c
28#define QCA988X_HW_2_0_CHIP_ID_REV 0x2
39#define QCA988X_HW_2_0_FW_DIR "ath10k/QCA988X/hw2.0" 29#define QCA988X_HW_2_0_FW_DIR "ath10k/QCA988X/hw2.0"
40#define QCA988X_HW_2_0_FW_FILE "firmware.bin" 30#define QCA988X_HW_2_0_FW_FILE "firmware.bin"
41#define QCA988X_HW_2_0_OTP_FILE "otp.bin" 31#define QCA988X_HW_2_0_OTP_FILE "otp.bin"
42#define QCA988X_HW_2_0_BOARD_DATA_FILE "board.bin" 32#define QCA988X_HW_2_0_BOARD_DATA_FILE "board.bin"
43#define QCA988X_HW_2_0_PATCH_LOAD_ADDR 0x1234 33#define QCA988X_HW_2_0_PATCH_LOAD_ADDR 0x1234
44 34
35#define ATH10K_FW_API2_FILE "firmware-2.bin"
36
37/* includes also the null byte */
38#define ATH10K_FIRMWARE_MAGIC "QCA-ATH10K"
39
40struct ath10k_fw_ie {
41 __le32 id;
42 __le32 len;
43 u8 data[0];
44};
45
46enum ath10k_fw_ie_type {
47 ATH10K_FW_IE_FW_VERSION = 0,
48 ATH10K_FW_IE_TIMESTAMP = 1,
49 ATH10K_FW_IE_FEATURES = 2,
50 ATH10K_FW_IE_FW_IMAGE = 3,
51 ATH10K_FW_IE_OTP_IMAGE = 4,
52};
53
45/* Known pecularities: 54/* Known pecularities:
46 * - current FW doesn't support raw rx mode (last tested v599) 55 * - current FW doesn't support raw rx mode (last tested v599)
47 * - current FW dumps upon raw tx mode (last tested v599) 56 * - current FW dumps upon raw tx mode (last tested v599)
@@ -53,6 +62,9 @@ enum ath10k_hw_txrx_mode {
53 ATH10K_HW_TXRX_RAW = 0, 62 ATH10K_HW_TXRX_RAW = 0,
54 ATH10K_HW_TXRX_NATIVE_WIFI = 1, 63 ATH10K_HW_TXRX_NATIVE_WIFI = 1,
55 ATH10K_HW_TXRX_ETHERNET = 2, 64 ATH10K_HW_TXRX_ETHERNET = 2,
65
66 /* Valid for HTT >= 3.0. Used for management frames in TX_FRM. */
67 ATH10K_HW_TXRX_MGMT = 3,
56}; 68};
57 69
58enum ath10k_mcast2ucast_mode { 70enum ath10k_mcast2ucast_mode {
@@ -60,6 +72,7 @@ enum ath10k_mcast2ucast_mode {
60 ATH10K_MCAST2UCAST_ENABLED = 1, 72 ATH10K_MCAST2UCAST_ENABLED = 1,
61}; 73};
62 74
75/* Target specific defines for MAIN firmware */
63#define TARGET_NUM_VDEVS 8 76#define TARGET_NUM_VDEVS 8
64#define TARGET_NUM_PEER_AST 2 77#define TARGET_NUM_PEER_AST 2
65#define TARGET_NUM_WDS_ENTRIES 32 78#define TARGET_NUM_WDS_ENTRIES 32
@@ -75,7 +88,11 @@ enum ath10k_mcast2ucast_mode {
75#define TARGET_RX_CHAIN_MASK (BIT(0) | BIT(1) | BIT(2)) 88#define TARGET_RX_CHAIN_MASK (BIT(0) | BIT(1) | BIT(2))
76#define TARGET_RX_TIMEOUT_LO_PRI 100 89#define TARGET_RX_TIMEOUT_LO_PRI 100
77#define TARGET_RX_TIMEOUT_HI_PRI 40 90#define TARGET_RX_TIMEOUT_HI_PRI 40
78#define TARGET_RX_DECAP_MODE ATH10K_HW_TXRX_ETHERNET 91
92/* Native Wifi decap mode is used to align IP frames to 4-byte boundaries and
93 * avoid a very expensive re-alignment in mac80211. */
94#define TARGET_RX_DECAP_MODE ATH10K_HW_TXRX_NATIVE_WIFI
95
79#define TARGET_SCAN_MAX_PENDING_REQS 4 96#define TARGET_SCAN_MAX_PENDING_REQS 4
80#define TARGET_BMISS_OFFLOAD_MAX_VDEV 3 97#define TARGET_BMISS_OFFLOAD_MAX_VDEV 3
81#define TARGET_ROAM_OFFLOAD_MAX_VDEV 3 98#define TARGET_ROAM_OFFLOAD_MAX_VDEV 3
@@ -90,6 +107,36 @@ enum ath10k_mcast2ucast_mode {
90#define TARGET_NUM_MSDU_DESC (1024 + 400) 107#define TARGET_NUM_MSDU_DESC (1024 + 400)
91#define TARGET_MAX_FRAG_ENTRIES 0 108#define TARGET_MAX_FRAG_ENTRIES 0
92 109
110/* Target specific defines for 10.X firmware */
111#define TARGET_10X_NUM_VDEVS 16
112#define TARGET_10X_NUM_PEER_AST 2
113#define TARGET_10X_NUM_WDS_ENTRIES 32
114#define TARGET_10X_DMA_BURST_SIZE 0
115#define TARGET_10X_MAC_AGGR_DELIM 0
116#define TARGET_10X_AST_SKID_LIMIT 16
117#define TARGET_10X_NUM_PEERS (128 + (TARGET_10X_NUM_VDEVS))
118#define TARGET_10X_NUM_OFFLOAD_PEERS 0
119#define TARGET_10X_NUM_OFFLOAD_REORDER_BUFS 0
120#define TARGET_10X_NUM_PEER_KEYS 2
121#define TARGET_10X_NUM_TIDS 256
122#define TARGET_10X_TX_CHAIN_MASK (BIT(0) | BIT(1) | BIT(2))
123#define TARGET_10X_RX_CHAIN_MASK (BIT(0) | BIT(1) | BIT(2))
124#define TARGET_10X_RX_TIMEOUT_LO_PRI 100
125#define TARGET_10X_RX_TIMEOUT_HI_PRI 40
126#define TARGET_10X_RX_DECAP_MODE ATH10K_HW_TXRX_NATIVE_WIFI
127#define TARGET_10X_SCAN_MAX_PENDING_REQS 4
128#define TARGET_10X_BMISS_OFFLOAD_MAX_VDEV 2
129#define TARGET_10X_ROAM_OFFLOAD_MAX_VDEV 2
130#define TARGET_10X_ROAM_OFFLOAD_MAX_AP_PROFILES 8
131#define TARGET_10X_GTK_OFFLOAD_MAX_VDEV 3
132#define TARGET_10X_NUM_MCAST_GROUPS 0
133#define TARGET_10X_NUM_MCAST_TABLE_ELEMS 0
134#define TARGET_10X_MCAST2UCAST_MODE ATH10K_MCAST2UCAST_DISABLED
135#define TARGET_10X_TX_DBG_LOG_SIZE 1024
136#define TARGET_10X_RX_SKIP_DEFRAG_TIMEOUT_DUP_DETECTION_CHECK 1
137#define TARGET_10X_VOW_CONFIG 0
138#define TARGET_10X_NUM_MSDU_DESC (1024 + 400)
139#define TARGET_10X_MAX_FRAG_ENTRIES 0
93 140
94/* Number of Copy Engines supported */ 141/* Number of Copy Engines supported */
95#define CE_COUNT 8 142#define CE_COUNT 8
@@ -169,6 +216,10 @@ enum ath10k_mcast2ucast_mode {
169#define SOC_LPO_CAL_ENABLE_LSB 20 216#define SOC_LPO_CAL_ENABLE_LSB 20
170#define SOC_LPO_CAL_ENABLE_MASK 0x00100000 217#define SOC_LPO_CAL_ENABLE_MASK 0x00100000
171 218
219#define SOC_CHIP_ID_ADDRESS 0x000000ec
220#define SOC_CHIP_ID_REV_LSB 8
221#define SOC_CHIP_ID_REV_MASK 0x00000f00
222
172#define WLAN_RESET_CONTROL_COLD_RST_MASK 0x00000008 223#define WLAN_RESET_CONTROL_COLD_RST_MASK 0x00000008
173#define WLAN_RESET_CONTROL_WARM_RST_MASK 0x00000004 224#define WLAN_RESET_CONTROL_WARM_RST_MASK 0x00000004
174#define WLAN_SYSTEM_SLEEP_DISABLE_LSB 0 225#define WLAN_SYSTEM_SLEEP_DISABLE_LSB 0
diff --git a/drivers/net/wireless/ath/ath10k/mac.c b/drivers/net/wireless/ath/ath10k/mac.c
index cf2ba4d850c9..0b1cc516e778 100644
--- a/drivers/net/wireless/ath/ath10k/mac.c
+++ b/drivers/net/wireless/ath/ath10k/mac.c
@@ -334,25 +334,29 @@ static int ath10k_peer_create(struct ath10k *ar, u32 vdev_id, const u8 *addr)
334 334
335static int ath10k_mac_set_rts(struct ath10k_vif *arvif, u32 value) 335static int ath10k_mac_set_rts(struct ath10k_vif *arvif, u32 value)
336{ 336{
337 struct ath10k *ar = arvif->ar;
338 u32 vdev_param;
339
337 if (value != 0xFFFFFFFF) 340 if (value != 0xFFFFFFFF)
338 value = min_t(u32, arvif->ar->hw->wiphy->rts_threshold, 341 value = min_t(u32, arvif->ar->hw->wiphy->rts_threshold,
339 ATH10K_RTS_MAX); 342 ATH10K_RTS_MAX);
340 343
341 return ath10k_wmi_vdev_set_param(arvif->ar, arvif->vdev_id, 344 vdev_param = ar->wmi.vdev_param->rts_threshold;
342 WMI_VDEV_PARAM_RTS_THRESHOLD, 345 return ath10k_wmi_vdev_set_param(ar, arvif->vdev_id, vdev_param, value);
343 value);
344} 346}
345 347
346static int ath10k_mac_set_frag(struct ath10k_vif *arvif, u32 value) 348static int ath10k_mac_set_frag(struct ath10k_vif *arvif, u32 value)
347{ 349{
350 struct ath10k *ar = arvif->ar;
351 u32 vdev_param;
352
348 if (value != 0xFFFFFFFF) 353 if (value != 0xFFFFFFFF)
349 value = clamp_t(u32, arvif->ar->hw->wiphy->frag_threshold, 354 value = clamp_t(u32, arvif->ar->hw->wiphy->frag_threshold,
350 ATH10K_FRAGMT_THRESHOLD_MIN, 355 ATH10K_FRAGMT_THRESHOLD_MIN,
351 ATH10K_FRAGMT_THRESHOLD_MAX); 356 ATH10K_FRAGMT_THRESHOLD_MAX);
352 357
353 return ath10k_wmi_vdev_set_param(arvif->ar, arvif->vdev_id, 358 vdev_param = ar->wmi.vdev_param->fragmentation_threshold;
354 WMI_VDEV_PARAM_FRAGMENTATION_THRESHOLD, 359 return ath10k_wmi_vdev_set_param(ar, arvif->vdev_id, vdev_param, value);
355 value);
356} 360}
357 361
358static int ath10k_peer_delete(struct ath10k *ar, u32 vdev_id, const u8 *addr) 362static int ath10k_peer_delete(struct ath10k *ar, u32 vdev_id, const u8 *addr)
@@ -460,6 +464,11 @@ static int ath10k_vdev_start(struct ath10k_vif *arvif)
460 arg.ssid_len = arvif->vif->bss_conf.ssid_len; 464 arg.ssid_len = arvif->vif->bss_conf.ssid_len;
461 } 465 }
462 466
467 ath10k_dbg(ATH10K_DBG_MAC,
468 "mac vdev %d start center_freq %d phymode %s\n",
469 arg.vdev_id, arg.channel.freq,
470 ath10k_wmi_phymode_str(arg.channel.mode));
471
463 ret = ath10k_wmi_vdev_start(ar, &arg); 472 ret = ath10k_wmi_vdev_start(ar, &arg);
464 if (ret) { 473 if (ret) {
465 ath10k_warn("WMI vdev start failed: ret %d\n", ret); 474 ath10k_warn("WMI vdev start failed: ret %d\n", ret);
@@ -503,13 +512,10 @@ static int ath10k_monitor_start(struct ath10k *ar, int vdev_id)
503{ 512{
504 struct ieee80211_channel *channel = ar->hw->conf.chandef.chan; 513 struct ieee80211_channel *channel = ar->hw->conf.chandef.chan;
505 struct wmi_vdev_start_request_arg arg = {}; 514 struct wmi_vdev_start_request_arg arg = {};
506 enum nl80211_channel_type type;
507 int ret = 0; 515 int ret = 0;
508 516
509 lockdep_assert_held(&ar->conf_mutex); 517 lockdep_assert_held(&ar->conf_mutex);
510 518
511 type = cfg80211_get_chandef_type(&ar->hw->conf.chandef);
512
513 arg.vdev_id = vdev_id; 519 arg.vdev_id = vdev_id;
514 arg.channel.freq = channel->center_freq; 520 arg.channel.freq = channel->center_freq;
515 arg.channel.band_center_freq1 = ar->hw->conf.chandef.center_freq1; 521 arg.channel.band_center_freq1 = ar->hw->conf.chandef.center_freq1;
@@ -560,12 +566,9 @@ static int ath10k_monitor_stop(struct ath10k *ar)
560 566
561 lockdep_assert_held(&ar->conf_mutex); 567 lockdep_assert_held(&ar->conf_mutex);
562 568
563 /* For some reasons, ath10k_wmi_vdev_down() here couse 569 ret = ath10k_wmi_vdev_down(ar, ar->monitor_vdev_id);
564 * often ath10k_wmi_vdev_stop() to fail. Next we could 570 if (ret)
565 * not run monitor vdev and driver reload 571 ath10k_warn("Monitor vdev down failed: %d\n", ret);
566 * required. Don't see such problems we skip
567 * ath10k_wmi_vdev_down() here.
568 */
569 572
570 ret = ath10k_wmi_vdev_stop(ar, ar->monitor_vdev_id); 573 ret = ath10k_wmi_vdev_stop(ar, ar->monitor_vdev_id);
571 if (ret) 574 if (ret)
@@ -607,7 +610,7 @@ static int ath10k_monitor_create(struct ath10k *ar)
607 goto vdev_fail; 610 goto vdev_fail;
608 } 611 }
609 612
610 ath10k_dbg(ATH10K_DBG_MAC, "Monitor interface created, vdev id: %d\n", 613 ath10k_dbg(ATH10K_DBG_MAC, "mac monitor vdev %d created\n",
611 ar->monitor_vdev_id); 614 ar->monitor_vdev_id);
612 615
613 ar->monitor_present = true; 616 ar->monitor_present = true;
@@ -639,7 +642,7 @@ static int ath10k_monitor_destroy(struct ath10k *ar)
639 ar->free_vdev_map |= 1 << (ar->monitor_vdev_id); 642 ar->free_vdev_map |= 1 << (ar->monitor_vdev_id);
640 ar->monitor_present = false; 643 ar->monitor_present = false;
641 644
642 ath10k_dbg(ATH10K_DBG_MAC, "Monitor interface destroyed, vdev id: %d\n", 645 ath10k_dbg(ATH10K_DBG_MAC, "mac monitor vdev %d deleted\n",
643 ar->monitor_vdev_id); 646 ar->monitor_vdev_id);
644 return ret; 647 return ret;
645} 648}
@@ -668,13 +671,14 @@ static void ath10k_control_beaconing(struct ath10k_vif *arvif,
668 arvif->vdev_id); 671 arvif->vdev_id);
669 return; 672 return;
670 } 673 }
671 ath10k_dbg(ATH10K_DBG_MAC, "VDEV: %d up\n", arvif->vdev_id); 674 ath10k_dbg(ATH10K_DBG_MAC, "mac vdev %d up\n", arvif->vdev_id);
672} 675}
673 676
674static void ath10k_control_ibss(struct ath10k_vif *arvif, 677static void ath10k_control_ibss(struct ath10k_vif *arvif,
675 struct ieee80211_bss_conf *info, 678 struct ieee80211_bss_conf *info,
676 const u8 self_peer[ETH_ALEN]) 679 const u8 self_peer[ETH_ALEN])
677{ 680{
681 u32 vdev_param;
678 int ret = 0; 682 int ret = 0;
679 683
680 lockdep_assert_held(&arvif->ar->conf_mutex); 684 lockdep_assert_held(&arvif->ar->conf_mutex);
@@ -708,8 +712,8 @@ static void ath10k_control_ibss(struct ath10k_vif *arvif,
708 return; 712 return;
709 } 713 }
710 714
711 ret = ath10k_wmi_vdev_set_param(arvif->ar, arvif->vdev_id, 715 vdev_param = arvif->ar->wmi.vdev_param->atim_window;
712 WMI_VDEV_PARAM_ATIM_WINDOW, 716 ret = ath10k_wmi_vdev_set_param(arvif->ar, arvif->vdev_id, vdev_param,
713 ATH10K_DEFAULT_ATIM); 717 ATH10K_DEFAULT_ATIM);
714 if (ret) 718 if (ret)
715 ath10k_warn("Failed to set IBSS ATIM for VDEV:%d ret:%d\n", 719 ath10k_warn("Failed to set IBSS ATIM for VDEV:%d ret:%d\n",
@@ -719,47 +723,45 @@ static void ath10k_control_ibss(struct ath10k_vif *arvif,
719/* 723/*
720 * Review this when mac80211 gains per-interface powersave support. 724 * Review this when mac80211 gains per-interface powersave support.
721 */ 725 */
722static void ath10k_ps_iter(void *data, u8 *mac, struct ieee80211_vif *vif) 726static int ath10k_mac_vif_setup_ps(struct ath10k_vif *arvif)
723{ 727{
724 struct ath10k_generic_iter *ar_iter = data; 728 struct ath10k *ar = arvif->ar;
725 struct ieee80211_conf *conf = &ar_iter->ar->hw->conf; 729 struct ieee80211_conf *conf = &ar->hw->conf;
726 struct ath10k_vif *arvif = ath10k_vif_to_arvif(vif);
727 enum wmi_sta_powersave_param param; 730 enum wmi_sta_powersave_param param;
728 enum wmi_sta_ps_mode psmode; 731 enum wmi_sta_ps_mode psmode;
729 int ret; 732 int ret;
730 733
731 lockdep_assert_held(&arvif->ar->conf_mutex); 734 lockdep_assert_held(&arvif->ar->conf_mutex);
732 735
733 if (vif->type != NL80211_IFTYPE_STATION) 736 if (arvif->vif->type != NL80211_IFTYPE_STATION)
734 return; 737 return 0;
735 738
736 if (conf->flags & IEEE80211_CONF_PS) { 739 if (conf->flags & IEEE80211_CONF_PS) {
737 psmode = WMI_STA_PS_MODE_ENABLED; 740 psmode = WMI_STA_PS_MODE_ENABLED;
738 param = WMI_STA_PS_PARAM_INACTIVITY_TIME; 741 param = WMI_STA_PS_PARAM_INACTIVITY_TIME;
739 742
740 ret = ath10k_wmi_set_sta_ps_param(ar_iter->ar, 743 ret = ath10k_wmi_set_sta_ps_param(ar, arvif->vdev_id, param,
741 arvif->vdev_id,
742 param,
743 conf->dynamic_ps_timeout); 744 conf->dynamic_ps_timeout);
744 if (ret) { 745 if (ret) {
745 ath10k_warn("Failed to set inactivity time for VDEV: %d\n", 746 ath10k_warn("Failed to set inactivity time for VDEV: %d\n",
746 arvif->vdev_id); 747 arvif->vdev_id);
747 return; 748 return ret;
748 } 749 }
749
750 ar_iter->ret = ret;
751 } else { 750 } else {
752 psmode = WMI_STA_PS_MODE_DISABLED; 751 psmode = WMI_STA_PS_MODE_DISABLED;
753 } 752 }
754 753
755 ar_iter->ret = ath10k_wmi_set_psmode(ar_iter->ar, arvif->vdev_id, 754 ath10k_dbg(ATH10K_DBG_MAC, "mac vdev %d psmode %s\n",
756 psmode); 755 arvif->vdev_id, psmode ? "enable" : "disable");
757 if (ar_iter->ret) 756
757 ret = ath10k_wmi_set_psmode(ar, arvif->vdev_id, psmode);
758 if (ret) {
758 ath10k_warn("Failed to set PS Mode: %d for VDEV: %d\n", 759 ath10k_warn("Failed to set PS Mode: %d for VDEV: %d\n",
759 psmode, arvif->vdev_id); 760 psmode, arvif->vdev_id);
760 else 761 return ret;
761 ath10k_dbg(ATH10K_DBG_MAC, "Set PS Mode: %d for VDEV: %d\n", 762 }
762 psmode, arvif->vdev_id); 763
764 return 0;
763} 765}
764 766
765/**********************/ 767/**********************/
@@ -949,7 +951,8 @@ static void ath10k_peer_assoc_h_ht(struct ath10k *ar,
949 arg->peer_ht_rates.num_rates = n; 951 arg->peer_ht_rates.num_rates = n;
950 arg->peer_num_spatial_streams = max((n+7) / 8, 1); 952 arg->peer_num_spatial_streams = max((n+7) / 8, 1);
951 953
952 ath10k_dbg(ATH10K_DBG_MAC, "mcs cnt %d nss %d\n", 954 ath10k_dbg(ATH10K_DBG_MAC, "mac ht peer %pM mcs cnt %d nss %d\n",
955 arg->addr,
953 arg->peer_ht_rates.num_rates, 956 arg->peer_ht_rates.num_rates,
954 arg->peer_num_spatial_streams); 957 arg->peer_num_spatial_streams);
955} 958}
@@ -969,11 +972,11 @@ static void ath10k_peer_assoc_h_qos_ap(struct ath10k *ar,
969 arg->peer_flags |= WMI_PEER_QOS; 972 arg->peer_flags |= WMI_PEER_QOS;
970 973
971 if (sta->wme && sta->uapsd_queues) { 974 if (sta->wme && sta->uapsd_queues) {
972 ath10k_dbg(ATH10K_DBG_MAC, "uapsd_queues: 0x%X, max_sp: %d\n", 975 ath10k_dbg(ATH10K_DBG_MAC, "mac uapsd_queues 0x%x max_sp %d\n",
973 sta->uapsd_queues, sta->max_sp); 976 sta->uapsd_queues, sta->max_sp);
974 977
975 arg->peer_flags |= WMI_PEER_APSD; 978 arg->peer_flags |= WMI_PEER_APSD;
976 arg->peer_flags |= WMI_RC_UAPSD_FLAG; 979 arg->peer_rate_caps |= WMI_RC_UAPSD_FLAG;
977 980
978 if (sta->uapsd_queues & IEEE80211_WMM_IE_STA_QOSINFO_AC_VO) 981 if (sta->uapsd_queues & IEEE80211_WMM_IE_STA_QOSINFO_AC_VO)
979 uapsd |= WMI_AP_PS_UAPSD_AC3_DELIVERY_EN | 982 uapsd |= WMI_AP_PS_UAPSD_AC3_DELIVERY_EN |
@@ -1028,14 +1031,27 @@ static void ath10k_peer_assoc_h_vht(struct ath10k *ar,
1028 struct wmi_peer_assoc_complete_arg *arg) 1031 struct wmi_peer_assoc_complete_arg *arg)
1029{ 1032{
1030 const struct ieee80211_sta_vht_cap *vht_cap = &sta->vht_cap; 1033 const struct ieee80211_sta_vht_cap *vht_cap = &sta->vht_cap;
1034 u8 ampdu_factor;
1031 1035
1032 if (!vht_cap->vht_supported) 1036 if (!vht_cap->vht_supported)
1033 return; 1037 return;
1034 1038
1035 arg->peer_flags |= WMI_PEER_VHT; 1039 arg->peer_flags |= WMI_PEER_VHT;
1036
1037 arg->peer_vht_caps = vht_cap->cap; 1040 arg->peer_vht_caps = vht_cap->cap;
1038 1041
1042
1043 ampdu_factor = (vht_cap->cap &
1044 IEEE80211_VHT_CAP_MAX_A_MPDU_LENGTH_EXPONENT_MASK) >>
1045 IEEE80211_VHT_CAP_MAX_A_MPDU_LENGTH_EXPONENT_SHIFT;
1046
1047 /* Workaround: Some Netgear/Linksys 11ac APs set Rx A-MPDU factor to
1048 * zero in VHT IE. Using it would result in degraded throughput.
1049 * arg->peer_max_mpdu at this point contains HT max_mpdu so keep
1050 * it if VHT max_mpdu is smaller. */
1051 arg->peer_max_mpdu = max(arg->peer_max_mpdu,
1052 (1U << (IEEE80211_HT_MAX_AMPDU_FACTOR +
1053 ampdu_factor)) - 1);
1054
1039 if (sta->bandwidth == IEEE80211_STA_RX_BW_80) 1055 if (sta->bandwidth == IEEE80211_STA_RX_BW_80)
1040 arg->peer_flags |= WMI_PEER_80MHZ; 1056 arg->peer_flags |= WMI_PEER_80MHZ;
1041 1057
@@ -1048,7 +1064,8 @@ static void ath10k_peer_assoc_h_vht(struct ath10k *ar,
1048 arg->peer_vht_rates.tx_mcs_set = 1064 arg->peer_vht_rates.tx_mcs_set =
1049 __le16_to_cpu(vht_cap->vht_mcs.tx_mcs_map); 1065 __le16_to_cpu(vht_cap->vht_mcs.tx_mcs_map);
1050 1066
1051 ath10k_dbg(ATH10K_DBG_MAC, "mac vht peer\n"); 1067 ath10k_dbg(ATH10K_DBG_MAC, "mac vht peer %pM max_mpdu %d flags 0x%x\n",
1068 sta->addr, arg->peer_max_mpdu, arg->peer_flags);
1052} 1069}
1053 1070
1054static void ath10k_peer_assoc_h_qos(struct ath10k *ar, 1071static void ath10k_peer_assoc_h_qos(struct ath10k *ar,
@@ -1076,8 +1093,6 @@ static void ath10k_peer_assoc_h_phymode(struct ath10k *ar,
1076{ 1093{
1077 enum wmi_phy_mode phymode = MODE_UNKNOWN; 1094 enum wmi_phy_mode phymode = MODE_UNKNOWN;
1078 1095
1079 /* FIXME: add VHT */
1080
1081 switch (ar->hw->conf.chandef.chan->band) { 1096 switch (ar->hw->conf.chandef.chan->band) {
1082 case IEEE80211_BAND_2GHZ: 1097 case IEEE80211_BAND_2GHZ:
1083 if (sta->ht_cap.ht_supported) { 1098 if (sta->ht_cap.ht_supported) {
@@ -1091,7 +1106,17 @@ static void ath10k_peer_assoc_h_phymode(struct ath10k *ar,
1091 1106
1092 break; 1107 break;
1093 case IEEE80211_BAND_5GHZ: 1108 case IEEE80211_BAND_5GHZ:
1094 if (sta->ht_cap.ht_supported) { 1109 /*
1110 * Check VHT first.
1111 */
1112 if (sta->vht_cap.vht_supported) {
1113 if (sta->bandwidth == IEEE80211_STA_RX_BW_80)
1114 phymode = MODE_11AC_VHT80;
1115 else if (sta->bandwidth == IEEE80211_STA_RX_BW_40)
1116 phymode = MODE_11AC_VHT40;
1117 else if (sta->bandwidth == IEEE80211_STA_RX_BW_20)
1118 phymode = MODE_11AC_VHT20;
1119 } else if (sta->ht_cap.ht_supported) {
1095 if (sta->bandwidth == IEEE80211_STA_RX_BW_40) 1120 if (sta->bandwidth == IEEE80211_STA_RX_BW_40)
1096 phymode = MODE_11NA_HT40; 1121 phymode = MODE_11NA_HT40;
1097 else 1122 else
@@ -1105,30 +1130,32 @@ static void ath10k_peer_assoc_h_phymode(struct ath10k *ar,
1105 break; 1130 break;
1106 } 1131 }
1107 1132
1133 ath10k_dbg(ATH10K_DBG_MAC, "mac peer %pM phymode %s\n",
1134 sta->addr, ath10k_wmi_phymode_str(phymode));
1135
1108 arg->peer_phymode = phymode; 1136 arg->peer_phymode = phymode;
1109 WARN_ON(phymode == MODE_UNKNOWN); 1137 WARN_ON(phymode == MODE_UNKNOWN);
1110} 1138}
1111 1139
1112static int ath10k_peer_assoc(struct ath10k *ar, 1140static int ath10k_peer_assoc_prepare(struct ath10k *ar,
1113 struct ath10k_vif *arvif, 1141 struct ath10k_vif *arvif,
1114 struct ieee80211_sta *sta, 1142 struct ieee80211_sta *sta,
1115 struct ieee80211_bss_conf *bss_conf) 1143 struct ieee80211_bss_conf *bss_conf,
1144 struct wmi_peer_assoc_complete_arg *arg)
1116{ 1145{
1117 struct wmi_peer_assoc_complete_arg arg;
1118
1119 lockdep_assert_held(&ar->conf_mutex); 1146 lockdep_assert_held(&ar->conf_mutex);
1120 1147
1121 memset(&arg, 0, sizeof(struct wmi_peer_assoc_complete_arg)); 1148 memset(arg, 0, sizeof(*arg));
1122 1149
1123 ath10k_peer_assoc_h_basic(ar, arvif, sta, bss_conf, &arg); 1150 ath10k_peer_assoc_h_basic(ar, arvif, sta, bss_conf, arg);
1124 ath10k_peer_assoc_h_crypto(ar, arvif, &arg); 1151 ath10k_peer_assoc_h_crypto(ar, arvif, arg);
1125 ath10k_peer_assoc_h_rates(ar, sta, &arg); 1152 ath10k_peer_assoc_h_rates(ar, sta, arg);
1126 ath10k_peer_assoc_h_ht(ar, sta, &arg); 1153 ath10k_peer_assoc_h_ht(ar, sta, arg);
1127 ath10k_peer_assoc_h_vht(ar, sta, &arg); 1154 ath10k_peer_assoc_h_vht(ar, sta, arg);
1128 ath10k_peer_assoc_h_qos(ar, arvif, sta, bss_conf, &arg); 1155 ath10k_peer_assoc_h_qos(ar, arvif, sta, bss_conf, arg);
1129 ath10k_peer_assoc_h_phymode(ar, arvif, sta, &arg); 1156 ath10k_peer_assoc_h_phymode(ar, arvif, sta, arg);
1130 1157
1131 return ath10k_wmi_peer_assoc(ar, &arg); 1158 return 0;
1132} 1159}
1133 1160
1134/* can be called only in mac80211 callbacks due to `key_count` usage */ 1161/* can be called only in mac80211 callbacks due to `key_count` usage */
@@ -1138,6 +1165,7 @@ static void ath10k_bss_assoc(struct ieee80211_hw *hw,
1138{ 1165{
1139 struct ath10k *ar = hw->priv; 1166 struct ath10k *ar = hw->priv;
1140 struct ath10k_vif *arvif = ath10k_vif_to_arvif(vif); 1167 struct ath10k_vif *arvif = ath10k_vif_to_arvif(vif);
1168 struct wmi_peer_assoc_complete_arg peer_arg;
1141 struct ieee80211_sta *ap_sta; 1169 struct ieee80211_sta *ap_sta;
1142 int ret; 1170 int ret;
1143 1171
@@ -1153,24 +1181,33 @@ static void ath10k_bss_assoc(struct ieee80211_hw *hw,
1153 return; 1181 return;
1154 } 1182 }
1155 1183
1156 ret = ath10k_peer_assoc(ar, arvif, ap_sta, bss_conf); 1184 ret = ath10k_peer_assoc_prepare(ar, arvif, ap_sta,
1185 bss_conf, &peer_arg);
1157 if (ret) { 1186 if (ret) {
1158 ath10k_warn("Peer assoc failed for %pM\n", bss_conf->bssid); 1187 ath10k_warn("Peer assoc prepare failed for %pM\n: %d",
1188 bss_conf->bssid, ret);
1159 rcu_read_unlock(); 1189 rcu_read_unlock();
1160 return; 1190 return;
1161 } 1191 }
1162 1192
1163 rcu_read_unlock(); 1193 rcu_read_unlock();
1164 1194
1195 ret = ath10k_wmi_peer_assoc(ar, &peer_arg);
1196 if (ret) {
1197 ath10k_warn("Peer assoc failed for %pM\n: %d",
1198 bss_conf->bssid, ret);
1199 return;
1200 }
1201
1202 ath10k_dbg(ATH10K_DBG_MAC,
1203 "mac vdev %d up (associated) bssid %pM aid %d\n",
1204 arvif->vdev_id, bss_conf->bssid, bss_conf->aid);
1205
1165 ret = ath10k_wmi_vdev_up(ar, arvif->vdev_id, bss_conf->aid, 1206 ret = ath10k_wmi_vdev_up(ar, arvif->vdev_id, bss_conf->aid,
1166 bss_conf->bssid); 1207 bss_conf->bssid);
1167 if (ret) 1208 if (ret)
1168 ath10k_warn("VDEV: %d up failed: ret %d\n", 1209 ath10k_warn("VDEV: %d up failed: ret %d\n",
1169 arvif->vdev_id, ret); 1210 arvif->vdev_id, ret);
1170 else
1171 ath10k_dbg(ATH10K_DBG_MAC,
1172 "VDEV: %d associated, BSSID: %pM, AID: %d\n",
1173 arvif->vdev_id, bss_conf->bssid, bss_conf->aid);
1174} 1211}
1175 1212
1176/* 1213/*
@@ -1191,10 +1228,11 @@ static void ath10k_bss_disassoc(struct ieee80211_hw *hw,
1191 * No idea why this happens, even though VDEV-DOWN is supposed 1228 * No idea why this happens, even though VDEV-DOWN is supposed
1192 * to be analogous to link down, so just stop the VDEV. 1229 * to be analogous to link down, so just stop the VDEV.
1193 */ 1230 */
1231 ath10k_dbg(ATH10K_DBG_MAC, "mac vdev %d stop (disassociated\n",
1232 arvif->vdev_id);
1233
1234 /* FIXME: check return value */
1194 ret = ath10k_vdev_stop(arvif); 1235 ret = ath10k_vdev_stop(arvif);
1195 if (!ret)
1196 ath10k_dbg(ATH10K_DBG_MAC, "VDEV: %d stopped\n",
1197 arvif->vdev_id);
1198 1236
1199 /* 1237 /*
1200 * If we don't call VDEV-DOWN after VDEV-STOP FW will remain active and 1238 * If we don't call VDEV-DOWN after VDEV-STOP FW will remain active and
@@ -1203,26 +1241,33 @@ static void ath10k_bss_disassoc(struct ieee80211_hw *hw,
1203 * interfaces as it expects there is no rx when no interface is 1241 * interfaces as it expects there is no rx when no interface is
1204 * running. 1242 * running.
1205 */ 1243 */
1206 ret = ath10k_wmi_vdev_down(ar, arvif->vdev_id); 1244 ath10k_dbg(ATH10K_DBG_MAC, "mac vdev %d down\n", arvif->vdev_id);
1207 if (ret)
1208 ath10k_dbg(ATH10K_DBG_MAC, "VDEV: %d ath10k_wmi_vdev_down failed (%d)\n",
1209 arvif->vdev_id, ret);
1210 1245
1211 ath10k_wmi_flush_tx(ar); 1246 /* FIXME: why don't we print error if wmi call fails? */
1247 ret = ath10k_wmi_vdev_down(ar, arvif->vdev_id);
1212 1248
1213 arvif->def_wep_key_index = 0; 1249 arvif->def_wep_key_idx = 0;
1214} 1250}
1215 1251
1216static int ath10k_station_assoc(struct ath10k *ar, struct ath10k_vif *arvif, 1252static int ath10k_station_assoc(struct ath10k *ar, struct ath10k_vif *arvif,
1217 struct ieee80211_sta *sta) 1253 struct ieee80211_sta *sta)
1218{ 1254{
1255 struct wmi_peer_assoc_complete_arg peer_arg;
1219 int ret = 0; 1256 int ret = 0;
1220 1257
1221 lockdep_assert_held(&ar->conf_mutex); 1258 lockdep_assert_held(&ar->conf_mutex);
1222 1259
1223 ret = ath10k_peer_assoc(ar, arvif, sta, NULL); 1260 ret = ath10k_peer_assoc_prepare(ar, arvif, sta, NULL, &peer_arg);
1261 if (ret) {
1262 ath10k_warn("WMI peer assoc prepare failed for %pM\n",
1263 sta->addr);
1264 return ret;
1265 }
1266
1267 ret = ath10k_wmi_peer_assoc(ar, &peer_arg);
1224 if (ret) { 1268 if (ret) {
1225 ath10k_warn("WMI peer assoc failed for %pM\n", sta->addr); 1269 ath10k_warn("Peer assoc failed for STA %pM\n: %d",
1270 sta->addr, ret);
1226 return ret; 1271 return ret;
1227 } 1272 }
1228 1273
@@ -1333,8 +1378,8 @@ static int ath10k_update_channel_list(struct ath10k *ar)
1333 continue; 1378 continue;
1334 1379
1335 ath10k_dbg(ATH10K_DBG_WMI, 1380 ath10k_dbg(ATH10K_DBG_WMI,
1336 "%s: [%zd/%d] freq %d maxpower %d regpower %d antenna %d mode %d\n", 1381 "mac channel [%zd/%d] freq %d maxpower %d regpower %d antenna %d mode %d\n",
1337 __func__, ch - arg.channels, arg.n_channels, 1382 ch - arg.channels, arg.n_channels,
1338 ch->freq, ch->max_power, ch->max_reg_power, 1383 ch->freq, ch->max_power, ch->max_reg_power,
1339 ch->max_antenna_gain, ch->mode); 1384 ch->max_antenna_gain, ch->mode);
1340 1385
@@ -1391,6 +1436,33 @@ static void ath10k_reg_notifier(struct wiphy *wiphy,
1391/* TX handlers */ 1436/* TX handlers */
1392/***************/ 1437/***************/
1393 1438
1439static u8 ath10k_tx_h_get_tid(struct ieee80211_hdr *hdr)
1440{
1441 if (ieee80211_is_mgmt(hdr->frame_control))
1442 return HTT_DATA_TX_EXT_TID_MGMT;
1443
1444 if (!ieee80211_is_data_qos(hdr->frame_control))
1445 return HTT_DATA_TX_EXT_TID_NON_QOS_MCAST_BCAST;
1446
1447 if (!is_unicast_ether_addr(ieee80211_get_DA(hdr)))
1448 return HTT_DATA_TX_EXT_TID_NON_QOS_MCAST_BCAST;
1449
1450 return ieee80211_get_qos_ctl(hdr)[0] & IEEE80211_QOS_CTL_TID_MASK;
1451}
1452
1453static u8 ath10k_tx_h_get_vdev_id(struct ath10k *ar,
1454 struct ieee80211_tx_info *info)
1455{
1456 if (info->control.vif)
1457 return ath10k_vif_to_arvif(info->control.vif)->vdev_id;
1458
1459 if (ar->monitor_enabled)
1460 return ar->monitor_vdev_id;
1461
1462 ath10k_warn("could not resolve vdev id\n");
1463 return 0;
1464}
1465
1394/* 1466/*
1395 * Frames sent to the FW have to be in "Native Wifi" format. 1467 * Frames sent to the FW have to be in "Native Wifi" format.
1396 * Strip the QoS field from the 802.11 header. 1468 * Strip the QoS field from the 802.11 header.
@@ -1411,6 +1483,30 @@ static void ath10k_tx_h_qos_workaround(struct ieee80211_hw *hw,
1411 skb_pull(skb, IEEE80211_QOS_CTL_LEN); 1483 skb_pull(skb, IEEE80211_QOS_CTL_LEN);
1412} 1484}
1413 1485
1486static void ath10k_tx_wep_key_work(struct work_struct *work)
1487{
1488 struct ath10k_vif *arvif = container_of(work, struct ath10k_vif,
1489 wep_key_work);
1490 int ret, keyidx = arvif->def_wep_key_newidx;
1491
1492 if (arvif->def_wep_key_idx == keyidx)
1493 return;
1494
1495 ath10k_dbg(ATH10K_DBG_MAC, "mac vdev %d set keyidx %d\n",
1496 arvif->vdev_id, keyidx);
1497
1498 ret = ath10k_wmi_vdev_set_param(arvif->ar,
1499 arvif->vdev_id,
1500 arvif->ar->wmi.vdev_param->def_keyid,
1501 keyidx);
1502 if (ret) {
1503 ath10k_warn("could not update wep keyidx (%d)\n", ret);
1504 return;
1505 }
1506
1507 arvif->def_wep_key_idx = keyidx;
1508}
1509
1414static void ath10k_tx_h_update_wep_key(struct sk_buff *skb) 1510static void ath10k_tx_h_update_wep_key(struct sk_buff *skb)
1415{ 1511{
1416 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); 1512 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb);
@@ -1419,11 +1515,6 @@ static void ath10k_tx_h_update_wep_key(struct sk_buff *skb)
1419 struct ath10k *ar = arvif->ar; 1515 struct ath10k *ar = arvif->ar;
1420 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)skb->data; 1516 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)skb->data;
1421 struct ieee80211_key_conf *key = info->control.hw_key; 1517 struct ieee80211_key_conf *key = info->control.hw_key;
1422 int ret;
1423
1424 /* TODO AP mode should be implemented */
1425 if (vif->type != NL80211_IFTYPE_STATION)
1426 return;
1427 1518
1428 if (!ieee80211_has_protected(hdr->frame_control)) 1519 if (!ieee80211_has_protected(hdr->frame_control))
1429 return; 1520 return;
@@ -1435,20 +1526,14 @@ static void ath10k_tx_h_update_wep_key(struct sk_buff *skb)
1435 key->cipher != WLAN_CIPHER_SUITE_WEP104) 1526 key->cipher != WLAN_CIPHER_SUITE_WEP104)
1436 return; 1527 return;
1437 1528
1438 if (key->keyidx == arvif->def_wep_key_index) 1529 if (key->keyidx == arvif->def_wep_key_idx)
1439 return; 1530 return;
1440 1531
1441 ath10k_dbg(ATH10K_DBG_MAC, "new wep keyidx will be %d\n", key->keyidx); 1532 /* FIXME: Most likely a few frames will be TXed with an old key. Simply
1442 1533 * queueing frames until key index is updated is not an option because
1443 ret = ath10k_wmi_vdev_set_param(ar, arvif->vdev_id, 1534 * sk_buff may need more processing to be done, e.g. offchannel */
1444 WMI_VDEV_PARAM_DEF_KEYID, 1535 arvif->def_wep_key_newidx = key->keyidx;
1445 key->keyidx); 1536 ieee80211_queue_work(ar->hw, &arvif->wep_key_work);
1446 if (ret) {
1447 ath10k_warn("could not update wep keyidx (%d)\n", ret);
1448 return;
1449 }
1450
1451 arvif->def_wep_key_index = key->keyidx;
1452} 1537}
1453 1538
1454static void ath10k_tx_h_add_p2p_noa_ie(struct ath10k *ar, struct sk_buff *skb) 1539static void ath10k_tx_h_add_p2p_noa_ie(struct ath10k *ar, struct sk_buff *skb)
@@ -1478,19 +1563,42 @@ static void ath10k_tx_h_add_p2p_noa_ie(struct ath10k *ar, struct sk_buff *skb)
1478static void ath10k_tx_htt(struct ath10k *ar, struct sk_buff *skb) 1563static void ath10k_tx_htt(struct ath10k *ar, struct sk_buff *skb)
1479{ 1564{
1480 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)skb->data; 1565 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)skb->data;
1481 int ret; 1566 int ret = 0;
1482 1567
1483 if (ieee80211_is_mgmt(hdr->frame_control)) 1568 if (ar->htt.target_version_major >= 3) {
1484 ret = ath10k_htt_mgmt_tx(&ar->htt, skb); 1569 /* Since HTT 3.0 there is no separate mgmt tx command */
1485 else if (ieee80211_is_nullfunc(hdr->frame_control)) 1570 ret = ath10k_htt_tx(&ar->htt, skb);
1571 goto exit;
1572 }
1573
1574 if (ieee80211_is_mgmt(hdr->frame_control)) {
1575 if (test_bit(ATH10K_FW_FEATURE_HAS_WMI_MGMT_TX,
1576 ar->fw_features)) {
1577 if (skb_queue_len(&ar->wmi_mgmt_tx_queue) >=
1578 ATH10K_MAX_NUM_MGMT_PENDING) {
1579 ath10k_warn("wmi mgmt_tx queue limit reached\n");
1580 ret = -EBUSY;
1581 goto exit;
1582 }
1583
1584 skb_queue_tail(&ar->wmi_mgmt_tx_queue, skb);
1585 ieee80211_queue_work(ar->hw, &ar->wmi_mgmt_tx_work);
1586 } else {
1587 ret = ath10k_htt_mgmt_tx(&ar->htt, skb);
1588 }
1589 } else if (!test_bit(ATH10K_FW_FEATURE_HAS_WMI_MGMT_TX,
1590 ar->fw_features) &&
1591 ieee80211_is_nullfunc(hdr->frame_control)) {
1486 /* FW does not report tx status properly for NullFunc frames 1592 /* FW does not report tx status properly for NullFunc frames
1487 * unless they are sent through mgmt tx path. mac80211 sends 1593 * unless they are sent through mgmt tx path. mac80211 sends
1488 * those frames when it detects link/beacon loss and depends on 1594 * those frames when it detects link/beacon loss and depends
1489 * the tx status to be correct. */ 1595 * on the tx status to be correct. */
1490 ret = ath10k_htt_mgmt_tx(&ar->htt, skb); 1596 ret = ath10k_htt_mgmt_tx(&ar->htt, skb);
1491 else 1597 } else {
1492 ret = ath10k_htt_tx(&ar->htt, skb); 1598 ret = ath10k_htt_tx(&ar->htt, skb);
1599 }
1493 1600
1601exit:
1494 if (ret) { 1602 if (ret) {
1495 ath10k_warn("tx failed (%d). dropping packet.\n", ret); 1603 ath10k_warn("tx failed (%d). dropping packet.\n", ret);
1496 ieee80211_free_txskb(ar->hw, skb); 1604 ieee80211_free_txskb(ar->hw, skb);
@@ -1534,18 +1642,19 @@ void ath10k_offchan_tx_work(struct work_struct *work)
1534 1642
1535 mutex_lock(&ar->conf_mutex); 1643 mutex_lock(&ar->conf_mutex);
1536 1644
1537 ath10k_dbg(ATH10K_DBG_MAC, "processing offchannel skb %p\n", 1645 ath10k_dbg(ATH10K_DBG_MAC, "mac offchannel skb %p\n",
1538 skb); 1646 skb);
1539 1647
1540 hdr = (struct ieee80211_hdr *)skb->data; 1648 hdr = (struct ieee80211_hdr *)skb->data;
1541 peer_addr = ieee80211_get_DA(hdr); 1649 peer_addr = ieee80211_get_DA(hdr);
1542 vdev_id = ATH10K_SKB_CB(skb)->htt.vdev_id; 1650 vdev_id = ATH10K_SKB_CB(skb)->vdev_id;
1543 1651
1544 spin_lock_bh(&ar->data_lock); 1652 spin_lock_bh(&ar->data_lock);
1545 peer = ath10k_peer_find(ar, vdev_id, peer_addr); 1653 peer = ath10k_peer_find(ar, vdev_id, peer_addr);
1546 spin_unlock_bh(&ar->data_lock); 1654 spin_unlock_bh(&ar->data_lock);
1547 1655
1548 if (peer) 1656 if (peer)
1657 /* FIXME: should this use ath10k_warn()? */
1549 ath10k_dbg(ATH10K_DBG_MAC, "peer %pM on vdev %d already present\n", 1658 ath10k_dbg(ATH10K_DBG_MAC, "peer %pM on vdev %d already present\n",
1550 peer_addr, vdev_id); 1659 peer_addr, vdev_id);
1551 1660
@@ -1580,6 +1689,36 @@ void ath10k_offchan_tx_work(struct work_struct *work)
1580 } 1689 }
1581} 1690}
1582 1691
1692void ath10k_mgmt_over_wmi_tx_purge(struct ath10k *ar)
1693{
1694 struct sk_buff *skb;
1695
1696 for (;;) {
1697 skb = skb_dequeue(&ar->wmi_mgmt_tx_queue);
1698 if (!skb)
1699 break;
1700
1701 ieee80211_free_txskb(ar->hw, skb);
1702 }
1703}
1704
1705void ath10k_mgmt_over_wmi_tx_work(struct work_struct *work)
1706{
1707 struct ath10k *ar = container_of(work, struct ath10k, wmi_mgmt_tx_work);
1708 struct sk_buff *skb;
1709 int ret;
1710
1711 for (;;) {
1712 skb = skb_dequeue(&ar->wmi_mgmt_tx_queue);
1713 if (!skb)
1714 break;
1715
1716 ret = ath10k_wmi_mgmt_tx(ar, skb);
1717 if (ret)
1718 ath10k_warn("wmi mgmt_tx failed (%d)\n", ret);
1719 }
1720}
1721
1583/************/ 1722/************/
1584/* Scanning */ 1723/* Scanning */
1585/************/ 1724/************/
@@ -1643,8 +1782,6 @@ static int ath10k_abort_scan(struct ath10k *ar)
1643 return -EIO; 1782 return -EIO;
1644 } 1783 }
1645 1784
1646 ath10k_wmi_flush_tx(ar);
1647
1648 ret = wait_for_completion_timeout(&ar->scan.completed, 3*HZ); 1785 ret = wait_for_completion_timeout(&ar->scan.completed, 3*HZ);
1649 if (ret == 0) 1786 if (ret == 0)
1650 ath10k_warn("timed out while waiting for scan to stop\n"); 1787 ath10k_warn("timed out while waiting for scan to stop\n");
@@ -1678,10 +1815,6 @@ static int ath10k_start_scan(struct ath10k *ar,
1678 if (ret) 1815 if (ret)
1679 return ret; 1816 return ret;
1680 1817
1681 /* make sure we submit the command so the completion
1682 * timeout makes sense */
1683 ath10k_wmi_flush_tx(ar);
1684
1685 ret = wait_for_completion_timeout(&ar->scan.started, 1*HZ); 1818 ret = wait_for_completion_timeout(&ar->scan.started, 1*HZ);
1686 if (ret == 0) { 1819 if (ret == 0) {
1687 ath10k_abort_scan(ar); 1820 ath10k_abort_scan(ar);
@@ -1709,16 +1842,7 @@ static void ath10k_tx(struct ieee80211_hw *hw,
1709 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); 1842 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb);
1710 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)skb->data; 1843 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)skb->data;
1711 struct ath10k *ar = hw->priv; 1844 struct ath10k *ar = hw->priv;
1712 struct ath10k_vif *arvif = NULL; 1845 u8 tid, vdev_id;
1713 u32 vdev_id = 0;
1714 u8 tid;
1715
1716 if (info->control.vif) {
1717 arvif = ath10k_vif_to_arvif(info->control.vif);
1718 vdev_id = arvif->vdev_id;
1719 } else if (ar->monitor_enabled) {
1720 vdev_id = ar->monitor_vdev_id;
1721 }
1722 1846
1723 /* We should disable CCK RATE due to P2P */ 1847 /* We should disable CCK RATE due to P2P */
1724 if (info->flags & IEEE80211_TX_CTL_NO_CCK_RATE) 1848 if (info->flags & IEEE80211_TX_CTL_NO_CCK_RATE)
@@ -1726,12 +1850,8 @@ static void ath10k_tx(struct ieee80211_hw *hw,
1726 1850
1727 /* we must calculate tid before we apply qos workaround 1851 /* we must calculate tid before we apply qos workaround
1728 * as we'd lose the qos control field */ 1852 * as we'd lose the qos control field */
1729 tid = HTT_DATA_TX_EXT_TID_NON_QOS_MCAST_BCAST; 1853 tid = ath10k_tx_h_get_tid(hdr);
1730 if (ieee80211_is_data_qos(hdr->frame_control) && 1854 vdev_id = ath10k_tx_h_get_vdev_id(ar, info);
1731 is_unicast_ether_addr(ieee80211_get_DA(hdr))) {
1732 u8 *qc = ieee80211_get_qos_ctl(hdr);
1733 tid = qc[0] & IEEE80211_QOS_CTL_TID_MASK;
1734 }
1735 1855
1736 /* it makes no sense to process injected frames like that */ 1856 /* it makes no sense to process injected frames like that */
1737 if (info->control.vif && 1857 if (info->control.vif &&
@@ -1742,14 +1862,14 @@ static void ath10k_tx(struct ieee80211_hw *hw,
1742 ath10k_tx_h_seq_no(skb); 1862 ath10k_tx_h_seq_no(skb);
1743 } 1863 }
1744 1864
1745 memset(ATH10K_SKB_CB(skb), 0, sizeof(*ATH10K_SKB_CB(skb))); 1865 ATH10K_SKB_CB(skb)->vdev_id = vdev_id;
1746 ATH10K_SKB_CB(skb)->htt.vdev_id = vdev_id; 1866 ATH10K_SKB_CB(skb)->htt.is_offchan = false;
1747 ATH10K_SKB_CB(skb)->htt.tid = tid; 1867 ATH10K_SKB_CB(skb)->htt.tid = tid;
1748 1868
1749 if (info->flags & IEEE80211_TX_CTL_TX_OFFCHAN) { 1869 if (info->flags & IEEE80211_TX_CTL_TX_OFFCHAN) {
1750 spin_lock_bh(&ar->data_lock); 1870 spin_lock_bh(&ar->data_lock);
1751 ATH10K_SKB_CB(skb)->htt.is_offchan = true; 1871 ATH10K_SKB_CB(skb)->htt.is_offchan = true;
1752 ATH10K_SKB_CB(skb)->htt.vdev_id = ar->scan.vdev_id; 1872 ATH10K_SKB_CB(skb)->vdev_id = ar->scan.vdev_id;
1753 spin_unlock_bh(&ar->data_lock); 1873 spin_unlock_bh(&ar->data_lock);
1754 1874
1755 ath10k_dbg(ATH10K_DBG_MAC, "queued offchannel skb %p\n", skb); 1875 ath10k_dbg(ATH10K_DBG_MAC, "queued offchannel skb %p\n", skb);
@@ -1771,6 +1891,7 @@ void ath10k_halt(struct ath10k *ar)
1771 1891
1772 del_timer_sync(&ar->scan.timeout); 1892 del_timer_sync(&ar->scan.timeout);
1773 ath10k_offchan_tx_purge(ar); 1893 ath10k_offchan_tx_purge(ar);
1894 ath10k_mgmt_over_wmi_tx_purge(ar);
1774 ath10k_peer_cleanup_all(ar); 1895 ath10k_peer_cleanup_all(ar);
1775 ath10k_core_stop(ar); 1896 ath10k_core_stop(ar);
1776 ath10k_hif_power_down(ar); 1897 ath10k_hif_power_down(ar);
@@ -1817,12 +1938,12 @@ static int ath10k_start(struct ieee80211_hw *hw)
1817 else if (ar->state == ATH10K_STATE_RESTARTING) 1938 else if (ar->state == ATH10K_STATE_RESTARTING)
1818 ar->state = ATH10K_STATE_RESTARTED; 1939 ar->state = ATH10K_STATE_RESTARTED;
1819 1940
1820 ret = ath10k_wmi_pdev_set_param(ar, WMI_PDEV_PARAM_PMF_QOS, 1); 1941 ret = ath10k_wmi_pdev_set_param(ar, ar->wmi.pdev_param->pmf_qos, 1);
1821 if (ret) 1942 if (ret)
1822 ath10k_warn("could not enable WMI_PDEV_PARAM_PMF_QOS (%d)\n", 1943 ath10k_warn("could not enable WMI_PDEV_PARAM_PMF_QOS (%d)\n",
1823 ret); 1944 ret);
1824 1945
1825 ret = ath10k_wmi_pdev_set_param(ar, WMI_PDEV_PARAM_DYNAMIC_BW, 0); 1946 ret = ath10k_wmi_pdev_set_param(ar, ar->wmi.pdev_param->dynamic_bw, 0);
1826 if (ret) 1947 if (ret)
1827 ath10k_warn("could not init WMI_PDEV_PARAM_DYNAMIC_BW (%d)\n", 1948 ath10k_warn("could not init WMI_PDEV_PARAM_DYNAMIC_BW (%d)\n",
1828 ret); 1949 ret);
@@ -1847,32 +1968,29 @@ static void ath10k_stop(struct ieee80211_hw *hw)
1847 ar->state = ATH10K_STATE_OFF; 1968 ar->state = ATH10K_STATE_OFF;
1848 mutex_unlock(&ar->conf_mutex); 1969 mutex_unlock(&ar->conf_mutex);
1849 1970
1971 ath10k_mgmt_over_wmi_tx_purge(ar);
1972
1850 cancel_work_sync(&ar->offchan_tx_work); 1973 cancel_work_sync(&ar->offchan_tx_work);
1974 cancel_work_sync(&ar->wmi_mgmt_tx_work);
1851 cancel_work_sync(&ar->restart_work); 1975 cancel_work_sync(&ar->restart_work);
1852} 1976}
1853 1977
1854static void ath10k_config_ps(struct ath10k *ar) 1978static int ath10k_config_ps(struct ath10k *ar)
1855{ 1979{
1856 struct ath10k_generic_iter ar_iter; 1980 struct ath10k_vif *arvif;
1981 int ret = 0;
1857 1982
1858 lockdep_assert_held(&ar->conf_mutex); 1983 lockdep_assert_held(&ar->conf_mutex);
1859 1984
1860 /* During HW reconfiguration mac80211 reports all interfaces that were 1985 list_for_each_entry(arvif, &ar->arvifs, list) {
1861 * running until reconfiguration was started. Since FW doesn't have any 1986 ret = ath10k_mac_vif_setup_ps(arvif);
1862 * vdevs at this point we must not iterate over this interface list. 1987 if (ret) {
1863 * This setting will be updated upon add_interface(). */ 1988 ath10k_warn("could not setup powersave (%d)\n", ret);
1864 if (ar->state == ATH10K_STATE_RESTARTED) 1989 break;
1865 return; 1990 }
1866 1991 }
1867 memset(&ar_iter, 0, sizeof(struct ath10k_generic_iter));
1868 ar_iter.ar = ar;
1869
1870 ieee80211_iterate_active_interfaces_atomic(
1871 ar->hw, IEEE80211_IFACE_ITER_NORMAL,
1872 ath10k_ps_iter, &ar_iter);
1873 1992
1874 if (ar_iter.ret) 1993 return ret;
1875 ath10k_warn("failed to set ps config (%d)\n", ar_iter.ret);
1876} 1994}
1877 1995
1878static int ath10k_config(struct ieee80211_hw *hw, u32 changed) 1996static int ath10k_config(struct ieee80211_hw *hw, u32 changed)
@@ -1884,7 +2002,7 @@ static int ath10k_config(struct ieee80211_hw *hw, u32 changed)
1884 mutex_lock(&ar->conf_mutex); 2002 mutex_lock(&ar->conf_mutex);
1885 2003
1886 if (changed & IEEE80211_CONF_CHANGE_CHANNEL) { 2004 if (changed & IEEE80211_CONF_CHANGE_CHANNEL) {
1887 ath10k_dbg(ATH10K_DBG_MAC, "Config channel %d mhz\n", 2005 ath10k_dbg(ATH10K_DBG_MAC, "mac config channel %d mhz\n",
1888 conf->chandef.chan->center_freq); 2006 conf->chandef.chan->center_freq);
1889 spin_lock_bh(&ar->data_lock); 2007 spin_lock_bh(&ar->data_lock);
1890 ar->rx_channel = conf->chandef.chan; 2008 ar->rx_channel = conf->chandef.chan;
@@ -1901,7 +2019,6 @@ static int ath10k_config(struct ieee80211_hw *hw, u32 changed)
1901 ret = ath10k_monitor_destroy(ar); 2019 ret = ath10k_monitor_destroy(ar);
1902 } 2020 }
1903 2021
1904 ath10k_wmi_flush_tx(ar);
1905 mutex_unlock(&ar->conf_mutex); 2022 mutex_unlock(&ar->conf_mutex);
1906 return ret; 2023 return ret;
1907} 2024}
@@ -1922,6 +2039,7 @@ static int ath10k_add_interface(struct ieee80211_hw *hw,
1922 int ret = 0; 2039 int ret = 0;
1923 u32 value; 2040 u32 value;
1924 int bit; 2041 int bit;
2042 u32 vdev_param;
1925 2043
1926 mutex_lock(&ar->conf_mutex); 2044 mutex_lock(&ar->conf_mutex);
1927 2045
@@ -1930,21 +2048,22 @@ static int ath10k_add_interface(struct ieee80211_hw *hw,
1930 arvif->ar = ar; 2048 arvif->ar = ar;
1931 arvif->vif = vif; 2049 arvif->vif = vif;
1932 2050
2051 INIT_WORK(&arvif->wep_key_work, ath10k_tx_wep_key_work);
2052
1933 if ((vif->type == NL80211_IFTYPE_MONITOR) && ar->monitor_present) { 2053 if ((vif->type == NL80211_IFTYPE_MONITOR) && ar->monitor_present) {
1934 ath10k_warn("Only one monitor interface allowed\n"); 2054 ath10k_warn("Only one monitor interface allowed\n");
1935 ret = -EBUSY; 2055 ret = -EBUSY;
1936 goto exit; 2056 goto err;
1937 } 2057 }
1938 2058
1939 bit = ffs(ar->free_vdev_map); 2059 bit = ffs(ar->free_vdev_map);
1940 if (bit == 0) { 2060 if (bit == 0) {
1941 ret = -EBUSY; 2061 ret = -EBUSY;
1942 goto exit; 2062 goto err;
1943 } 2063 }
1944 2064
1945 arvif->vdev_id = bit - 1; 2065 arvif->vdev_id = bit - 1;
1946 arvif->vdev_subtype = WMI_VDEV_SUBTYPE_NONE; 2066 arvif->vdev_subtype = WMI_VDEV_SUBTYPE_NONE;
1947 ar->free_vdev_map &= ~(1 << arvif->vdev_id);
1948 2067
1949 if (ar->p2p) 2068 if (ar->p2p)
1950 arvif->vdev_subtype = WMI_VDEV_SUBTYPE_P2P_DEVICE; 2069 arvif->vdev_subtype = WMI_VDEV_SUBTYPE_P2P_DEVICE;
@@ -1973,32 +2092,41 @@ static int ath10k_add_interface(struct ieee80211_hw *hw,
1973 break; 2092 break;
1974 } 2093 }
1975 2094
1976 ath10k_dbg(ATH10K_DBG_MAC, "Add interface: id %d type %d subtype %d\n", 2095 ath10k_dbg(ATH10K_DBG_MAC, "mac vdev create %d (add interface) type %d subtype %d\n",
1977 arvif->vdev_id, arvif->vdev_type, arvif->vdev_subtype); 2096 arvif->vdev_id, arvif->vdev_type, arvif->vdev_subtype);
1978 2097
1979 ret = ath10k_wmi_vdev_create(ar, arvif->vdev_id, arvif->vdev_type, 2098 ret = ath10k_wmi_vdev_create(ar, arvif->vdev_id, arvif->vdev_type,
1980 arvif->vdev_subtype, vif->addr); 2099 arvif->vdev_subtype, vif->addr);
1981 if (ret) { 2100 if (ret) {
1982 ath10k_warn("WMI vdev create failed: ret %d\n", ret); 2101 ath10k_warn("WMI vdev create failed: ret %d\n", ret);
1983 goto exit; 2102 goto err;
1984 } 2103 }
1985 2104
1986 ret = ath10k_wmi_vdev_set_param(ar, 0, WMI_VDEV_PARAM_DEF_KEYID, 2105 ar->free_vdev_map &= ~BIT(arvif->vdev_id);
1987 arvif->def_wep_key_index); 2106 list_add(&arvif->list, &ar->arvifs);
1988 if (ret) 2107
2108 vdev_param = ar->wmi.vdev_param->def_keyid;
2109 ret = ath10k_wmi_vdev_set_param(ar, 0, vdev_param,
2110 arvif->def_wep_key_idx);
2111 if (ret) {
1989 ath10k_warn("Failed to set default keyid: %d\n", ret); 2112 ath10k_warn("Failed to set default keyid: %d\n", ret);
2113 goto err_vdev_delete;
2114 }
1990 2115
1991 ret = ath10k_wmi_vdev_set_param(ar, arvif->vdev_id, 2116 vdev_param = ar->wmi.vdev_param->tx_encap_type;
1992 WMI_VDEV_PARAM_TX_ENCAP_TYPE, 2117 ret = ath10k_wmi_vdev_set_param(ar, arvif->vdev_id, vdev_param,
1993 ATH10K_HW_TXRX_NATIVE_WIFI); 2118 ATH10K_HW_TXRX_NATIVE_WIFI);
1994 if (ret) 2119 /* 10.X firmware does not support this VDEV parameter. Do not warn */
2120 if (ret && ret != -EOPNOTSUPP) {
1995 ath10k_warn("Failed to set TX encap: %d\n", ret); 2121 ath10k_warn("Failed to set TX encap: %d\n", ret);
2122 goto err_vdev_delete;
2123 }
1996 2124
1997 if (arvif->vdev_type == WMI_VDEV_TYPE_AP) { 2125 if (arvif->vdev_type == WMI_VDEV_TYPE_AP) {
1998 ret = ath10k_peer_create(ar, arvif->vdev_id, vif->addr); 2126 ret = ath10k_peer_create(ar, arvif->vdev_id, vif->addr);
1999 if (ret) { 2127 if (ret) {
2000 ath10k_warn("Failed to create peer for AP: %d\n", ret); 2128 ath10k_warn("Failed to create peer for AP: %d\n", ret);
2001 goto exit; 2129 goto err_vdev_delete;
2002 } 2130 }
2003 } 2131 }
2004 2132
@@ -2007,39 +2135,62 @@ static int ath10k_add_interface(struct ieee80211_hw *hw,
2007 value = WMI_STA_PS_RX_WAKE_POLICY_WAKE; 2135 value = WMI_STA_PS_RX_WAKE_POLICY_WAKE;
2008 ret = ath10k_wmi_set_sta_ps_param(ar, arvif->vdev_id, 2136 ret = ath10k_wmi_set_sta_ps_param(ar, arvif->vdev_id,
2009 param, value); 2137 param, value);
2010 if (ret) 2138 if (ret) {
2011 ath10k_warn("Failed to set RX wake policy: %d\n", ret); 2139 ath10k_warn("Failed to set RX wake policy: %d\n", ret);
2140 goto err_peer_delete;
2141 }
2012 2142
2013 param = WMI_STA_PS_PARAM_TX_WAKE_THRESHOLD; 2143 param = WMI_STA_PS_PARAM_TX_WAKE_THRESHOLD;
2014 value = WMI_STA_PS_TX_WAKE_THRESHOLD_ALWAYS; 2144 value = WMI_STA_PS_TX_WAKE_THRESHOLD_ALWAYS;
2015 ret = ath10k_wmi_set_sta_ps_param(ar, arvif->vdev_id, 2145 ret = ath10k_wmi_set_sta_ps_param(ar, arvif->vdev_id,
2016 param, value); 2146 param, value);
2017 if (ret) 2147 if (ret) {
2018 ath10k_warn("Failed to set TX wake thresh: %d\n", ret); 2148 ath10k_warn("Failed to set TX wake thresh: %d\n", ret);
2149 goto err_peer_delete;
2150 }
2019 2151
2020 param = WMI_STA_PS_PARAM_PSPOLL_COUNT; 2152 param = WMI_STA_PS_PARAM_PSPOLL_COUNT;
2021 value = WMI_STA_PS_PSPOLL_COUNT_NO_MAX; 2153 value = WMI_STA_PS_PSPOLL_COUNT_NO_MAX;
2022 ret = ath10k_wmi_set_sta_ps_param(ar, arvif->vdev_id, 2154 ret = ath10k_wmi_set_sta_ps_param(ar, arvif->vdev_id,
2023 param, value); 2155 param, value);
2024 if (ret) 2156 if (ret) {
2025 ath10k_warn("Failed to set PSPOLL count: %d\n", ret); 2157 ath10k_warn("Failed to set PSPOLL count: %d\n", ret);
2158 goto err_peer_delete;
2159 }
2026 } 2160 }
2027 2161
2028 ret = ath10k_mac_set_rts(arvif, ar->hw->wiphy->rts_threshold); 2162 ret = ath10k_mac_set_rts(arvif, ar->hw->wiphy->rts_threshold);
2029 if (ret) 2163 if (ret) {
2030 ath10k_warn("failed to set rts threshold for vdev %d (%d)\n", 2164 ath10k_warn("failed to set rts threshold for vdev %d (%d)\n",
2031 arvif->vdev_id, ret); 2165 arvif->vdev_id, ret);
2166 goto err_peer_delete;
2167 }
2032 2168
2033 ret = ath10k_mac_set_frag(arvif, ar->hw->wiphy->frag_threshold); 2169 ret = ath10k_mac_set_frag(arvif, ar->hw->wiphy->frag_threshold);
2034 if (ret) 2170 if (ret) {
2035 ath10k_warn("failed to set frag threshold for vdev %d (%d)\n", 2171 ath10k_warn("failed to set frag threshold for vdev %d (%d)\n",
2036 arvif->vdev_id, ret); 2172 arvif->vdev_id, ret);
2173 goto err_peer_delete;
2174 }
2037 2175
2038 if (arvif->vdev_type == WMI_VDEV_TYPE_MONITOR) 2176 if (arvif->vdev_type == WMI_VDEV_TYPE_MONITOR)
2039 ar->monitor_present = true; 2177 ar->monitor_present = true;
2040 2178
2041exit:
2042 mutex_unlock(&ar->conf_mutex); 2179 mutex_unlock(&ar->conf_mutex);
2180 return 0;
2181
2182err_peer_delete:
2183 if (arvif->vdev_type == WMI_VDEV_TYPE_AP)
2184 ath10k_wmi_peer_delete(ar, arvif->vdev_id, vif->addr);
2185
2186err_vdev_delete:
2187 ath10k_wmi_vdev_delete(ar, arvif->vdev_id);
2188 ar->free_vdev_map &= ~BIT(arvif->vdev_id);
2189 list_del(&arvif->list);
2190
2191err:
2192 mutex_unlock(&ar->conf_mutex);
2193
2043 return ret; 2194 return ret;
2044} 2195}
2045 2196
@@ -2052,9 +2203,17 @@ static void ath10k_remove_interface(struct ieee80211_hw *hw,
2052 2203
2053 mutex_lock(&ar->conf_mutex); 2204 mutex_lock(&ar->conf_mutex);
2054 2205
2055 ath10k_dbg(ATH10K_DBG_MAC, "Remove interface: id %d\n", arvif->vdev_id); 2206 cancel_work_sync(&arvif->wep_key_work);
2207
2208 spin_lock_bh(&ar->data_lock);
2209 if (arvif->beacon) {
2210 dev_kfree_skb_any(arvif->beacon);
2211 arvif->beacon = NULL;
2212 }
2213 spin_unlock_bh(&ar->data_lock);
2056 2214
2057 ar->free_vdev_map |= 1 << (arvif->vdev_id); 2215 ar->free_vdev_map |= 1 << (arvif->vdev_id);
2216 list_del(&arvif->list);
2058 2217
2059 if (arvif->vdev_type == WMI_VDEV_TYPE_AP) { 2218 if (arvif->vdev_type == WMI_VDEV_TYPE_AP) {
2060 ret = ath10k_peer_delete(arvif->ar, arvif->vdev_id, vif->addr); 2219 ret = ath10k_peer_delete(arvif->ar, arvif->vdev_id, vif->addr);
@@ -2064,6 +2223,9 @@ static void ath10k_remove_interface(struct ieee80211_hw *hw,
2064 kfree(arvif->u.ap.noa_data); 2223 kfree(arvif->u.ap.noa_data);
2065 } 2224 }
2066 2225
2226 ath10k_dbg(ATH10K_DBG_MAC, "mac vdev delete %d (remove interface)\n",
2227 arvif->vdev_id);
2228
2067 ret = ath10k_wmi_vdev_delete(ar, arvif->vdev_id); 2229 ret = ath10k_wmi_vdev_delete(ar, arvif->vdev_id);
2068 if (ret) 2230 if (ret)
2069 ath10k_warn("WMI vdev delete failed: %d\n", ret); 2231 ath10k_warn("WMI vdev delete failed: %d\n", ret);
@@ -2105,18 +2267,20 @@ static void ath10k_configure_filter(struct ieee80211_hw *hw,
2105 2267
2106 if ((ar->filter_flags & FIF_PROMISC_IN_BSS) && 2268 if ((ar->filter_flags & FIF_PROMISC_IN_BSS) &&
2107 !ar->monitor_enabled) { 2269 !ar->monitor_enabled) {
2270 ath10k_dbg(ATH10K_DBG_MAC, "mac monitor %d start\n",
2271 ar->monitor_vdev_id);
2272
2108 ret = ath10k_monitor_start(ar, ar->monitor_vdev_id); 2273 ret = ath10k_monitor_start(ar, ar->monitor_vdev_id);
2109 if (ret) 2274 if (ret)
2110 ath10k_warn("Unable to start monitor mode\n"); 2275 ath10k_warn("Unable to start monitor mode\n");
2111 else
2112 ath10k_dbg(ATH10K_DBG_MAC, "Monitor mode started\n");
2113 } else if (!(ar->filter_flags & FIF_PROMISC_IN_BSS) && 2276 } else if (!(ar->filter_flags & FIF_PROMISC_IN_BSS) &&
2114 ar->monitor_enabled) { 2277 ar->monitor_enabled) {
2278 ath10k_dbg(ATH10K_DBG_MAC, "mac monitor %d stop\n",
2279 ar->monitor_vdev_id);
2280
2115 ret = ath10k_monitor_stop(ar); 2281 ret = ath10k_monitor_stop(ar);
2116 if (ret) 2282 if (ret)
2117 ath10k_warn("Unable to stop monitor mode\n"); 2283 ath10k_warn("Unable to stop monitor mode\n");
2118 else
2119 ath10k_dbg(ATH10K_DBG_MAC, "Monitor mode stopped\n");
2120 } 2284 }
2121 2285
2122 mutex_unlock(&ar->conf_mutex); 2286 mutex_unlock(&ar->conf_mutex);
@@ -2130,6 +2294,7 @@ static void ath10k_bss_info_changed(struct ieee80211_hw *hw,
2130 struct ath10k *ar = hw->priv; 2294 struct ath10k *ar = hw->priv;
2131 struct ath10k_vif *arvif = ath10k_vif_to_arvif(vif); 2295 struct ath10k_vif *arvif = ath10k_vif_to_arvif(vif);
2132 int ret = 0; 2296 int ret = 0;
2297 u32 vdev_param, pdev_param;
2133 2298
2134 mutex_lock(&ar->conf_mutex); 2299 mutex_lock(&ar->conf_mutex);
2135 2300
@@ -2138,44 +2303,44 @@ static void ath10k_bss_info_changed(struct ieee80211_hw *hw,
2138 2303
2139 if (changed & BSS_CHANGED_BEACON_INT) { 2304 if (changed & BSS_CHANGED_BEACON_INT) {
2140 arvif->beacon_interval = info->beacon_int; 2305 arvif->beacon_interval = info->beacon_int;
2141 ret = ath10k_wmi_vdev_set_param(ar, arvif->vdev_id, 2306 vdev_param = ar->wmi.vdev_param->beacon_interval;
2142 WMI_VDEV_PARAM_BEACON_INTERVAL, 2307 ret = ath10k_wmi_vdev_set_param(ar, arvif->vdev_id, vdev_param,
2143 arvif->beacon_interval); 2308 arvif->beacon_interval);
2309 ath10k_dbg(ATH10K_DBG_MAC,
2310 "mac vdev %d beacon_interval %d\n",
2311 arvif->vdev_id, arvif->beacon_interval);
2312
2144 if (ret) 2313 if (ret)
2145 ath10k_warn("Failed to set beacon interval for VDEV: %d\n", 2314 ath10k_warn("Failed to set beacon interval for VDEV: %d\n",
2146 arvif->vdev_id); 2315 arvif->vdev_id);
2147 else
2148 ath10k_dbg(ATH10K_DBG_MAC,
2149 "Beacon interval: %d set for VDEV: %d\n",
2150 arvif->beacon_interval, arvif->vdev_id);
2151 } 2316 }
2152 2317
2153 if (changed & BSS_CHANGED_BEACON) { 2318 if (changed & BSS_CHANGED_BEACON) {
2154 ret = ath10k_wmi_pdev_set_param(ar, 2319 ath10k_dbg(ATH10K_DBG_MAC,
2155 WMI_PDEV_PARAM_BEACON_TX_MODE, 2320 "vdev %d set beacon tx mode to staggered\n",
2321 arvif->vdev_id);
2322
2323 pdev_param = ar->wmi.pdev_param->beacon_tx_mode;
2324 ret = ath10k_wmi_pdev_set_param(ar, pdev_param,
2156 WMI_BEACON_STAGGERED_MODE); 2325 WMI_BEACON_STAGGERED_MODE);
2157 if (ret) 2326 if (ret)
2158 ath10k_warn("Failed to set beacon mode for VDEV: %d\n", 2327 ath10k_warn("Failed to set beacon mode for VDEV: %d\n",
2159 arvif->vdev_id); 2328 arvif->vdev_id);
2160 else
2161 ath10k_dbg(ATH10K_DBG_MAC,
2162 "Set staggered beacon mode for VDEV: %d\n",
2163 arvif->vdev_id);
2164 } 2329 }
2165 2330
2166 if (changed & BSS_CHANGED_BEACON_INFO) { 2331 if (changed & BSS_CHANGED_BEACON_INFO) {
2167 arvif->dtim_period = info->dtim_period; 2332 arvif->dtim_period = info->dtim_period;
2168 2333
2169 ret = ath10k_wmi_vdev_set_param(ar, arvif->vdev_id, 2334 ath10k_dbg(ATH10K_DBG_MAC,
2170 WMI_VDEV_PARAM_DTIM_PERIOD, 2335 "mac vdev %d dtim_period %d\n",
2336 arvif->vdev_id, arvif->dtim_period);
2337
2338 vdev_param = ar->wmi.vdev_param->dtim_period;
2339 ret = ath10k_wmi_vdev_set_param(ar, arvif->vdev_id, vdev_param,
2171 arvif->dtim_period); 2340 arvif->dtim_period);
2172 if (ret) 2341 if (ret)
2173 ath10k_warn("Failed to set dtim period for VDEV: %d\n", 2342 ath10k_warn("Failed to set dtim period for VDEV: %d\n",
2174 arvif->vdev_id); 2343 arvif->vdev_id);
2175 else
2176 ath10k_dbg(ATH10K_DBG_MAC,
2177 "Set dtim period: %d for VDEV: %d\n",
2178 arvif->dtim_period, arvif->vdev_id);
2179 } 2344 }
2180 2345
2181 if (changed & BSS_CHANGED_SSID && 2346 if (changed & BSS_CHANGED_SSID &&
@@ -2188,16 +2353,15 @@ static void ath10k_bss_info_changed(struct ieee80211_hw *hw,
2188 2353
2189 if (changed & BSS_CHANGED_BSSID) { 2354 if (changed & BSS_CHANGED_BSSID) {
2190 if (!is_zero_ether_addr(info->bssid)) { 2355 if (!is_zero_ether_addr(info->bssid)) {
2356 ath10k_dbg(ATH10K_DBG_MAC,
2357 "mac vdev %d create peer %pM\n",
2358 arvif->vdev_id, info->bssid);
2359
2191 ret = ath10k_peer_create(ar, arvif->vdev_id, 2360 ret = ath10k_peer_create(ar, arvif->vdev_id,
2192 info->bssid); 2361 info->bssid);
2193 if (ret) 2362 if (ret)
2194 ath10k_warn("Failed to add peer: %pM for VDEV: %d\n", 2363 ath10k_warn("Failed to add peer: %pM for VDEV: %d\n",
2195 info->bssid, arvif->vdev_id); 2364 info->bssid, arvif->vdev_id);
2196 else
2197 ath10k_dbg(ATH10K_DBG_MAC,
2198 "Added peer: %pM for VDEV: %d\n",
2199 info->bssid, arvif->vdev_id);
2200
2201 2365
2202 if (vif->type == NL80211_IFTYPE_STATION) { 2366 if (vif->type == NL80211_IFTYPE_STATION) {
2203 /* 2367 /*
@@ -2207,11 +2371,12 @@ static void ath10k_bss_info_changed(struct ieee80211_hw *hw,
2207 memcpy(arvif->u.sta.bssid, info->bssid, 2371 memcpy(arvif->u.sta.bssid, info->bssid,
2208 ETH_ALEN); 2372 ETH_ALEN);
2209 2373
2374 ath10k_dbg(ATH10K_DBG_MAC,
2375 "mac vdev %d start %pM\n",
2376 arvif->vdev_id, info->bssid);
2377
2378 /* FIXME: check return value */
2210 ret = ath10k_vdev_start(arvif); 2379 ret = ath10k_vdev_start(arvif);
2211 if (!ret)
2212 ath10k_dbg(ATH10K_DBG_MAC,
2213 "VDEV: %d started with BSSID: %pM\n",
2214 arvif->vdev_id, info->bssid);
2215 } 2380 }
2216 2381
2217 /* 2382 /*
@@ -2235,16 +2400,15 @@ static void ath10k_bss_info_changed(struct ieee80211_hw *hw,
2235 else 2400 else
2236 cts_prot = 0; 2401 cts_prot = 0;
2237 2402
2238 ret = ath10k_wmi_vdev_set_param(ar, arvif->vdev_id, 2403 ath10k_dbg(ATH10K_DBG_MAC, "mac vdev %d cts_prot %d\n",
2239 WMI_VDEV_PARAM_ENABLE_RTSCTS, 2404 arvif->vdev_id, cts_prot);
2405
2406 vdev_param = ar->wmi.vdev_param->enable_rtscts;
2407 ret = ath10k_wmi_vdev_set_param(ar, arvif->vdev_id, vdev_param,
2240 cts_prot); 2408 cts_prot);
2241 if (ret) 2409 if (ret)
2242 ath10k_warn("Failed to set CTS prot for VDEV: %d\n", 2410 ath10k_warn("Failed to set CTS prot for VDEV: %d\n",
2243 arvif->vdev_id); 2411 arvif->vdev_id);
2244 else
2245 ath10k_dbg(ATH10K_DBG_MAC,
2246 "Set CTS prot: %d for VDEV: %d\n",
2247 cts_prot, arvif->vdev_id);
2248 } 2412 }
2249 2413
2250 if (changed & BSS_CHANGED_ERP_SLOT) { 2414 if (changed & BSS_CHANGED_ERP_SLOT) {
@@ -2255,16 +2419,15 @@ static void ath10k_bss_info_changed(struct ieee80211_hw *hw,
2255 else 2419 else
2256 slottime = WMI_VDEV_SLOT_TIME_LONG; /* 20us */ 2420 slottime = WMI_VDEV_SLOT_TIME_LONG; /* 20us */
2257 2421
2258 ret = ath10k_wmi_vdev_set_param(ar, arvif->vdev_id, 2422 ath10k_dbg(ATH10K_DBG_MAC, "mac vdev %d slot_time %d\n",
2259 WMI_VDEV_PARAM_SLOT_TIME, 2423 arvif->vdev_id, slottime);
2424
2425 vdev_param = ar->wmi.vdev_param->slot_time;
2426 ret = ath10k_wmi_vdev_set_param(ar, arvif->vdev_id, vdev_param,
2260 slottime); 2427 slottime);
2261 if (ret) 2428 if (ret)
2262 ath10k_warn("Failed to set erp slot for VDEV: %d\n", 2429 ath10k_warn("Failed to set erp slot for VDEV: %d\n",
2263 arvif->vdev_id); 2430 arvif->vdev_id);
2264 else
2265 ath10k_dbg(ATH10K_DBG_MAC,
2266 "Set slottime: %d for VDEV: %d\n",
2267 slottime, arvif->vdev_id);
2268 } 2431 }
2269 2432
2270 if (changed & BSS_CHANGED_ERP_PREAMBLE) { 2433 if (changed & BSS_CHANGED_ERP_PREAMBLE) {
@@ -2274,16 +2437,16 @@ static void ath10k_bss_info_changed(struct ieee80211_hw *hw,
2274 else 2437 else
2275 preamble = WMI_VDEV_PREAMBLE_LONG; 2438 preamble = WMI_VDEV_PREAMBLE_LONG;
2276 2439
2277 ret = ath10k_wmi_vdev_set_param(ar, arvif->vdev_id, 2440 ath10k_dbg(ATH10K_DBG_MAC,
2278 WMI_VDEV_PARAM_PREAMBLE, 2441 "mac vdev %d preamble %dn",
2442 arvif->vdev_id, preamble);
2443
2444 vdev_param = ar->wmi.vdev_param->preamble;
2445 ret = ath10k_wmi_vdev_set_param(ar, arvif->vdev_id, vdev_param,
2279 preamble); 2446 preamble);
2280 if (ret) 2447 if (ret)
2281 ath10k_warn("Failed to set preamble for VDEV: %d\n", 2448 ath10k_warn("Failed to set preamble for VDEV: %d\n",
2282 arvif->vdev_id); 2449 arvif->vdev_id);
2283 else
2284 ath10k_dbg(ATH10K_DBG_MAC,
2285 "Set preamble: %d for VDEV: %d\n",
2286 preamble, arvif->vdev_id);
2287 } 2450 }
2288 2451
2289 if (changed & BSS_CHANGED_ASSOC) { 2452 if (changed & BSS_CHANGED_ASSOC) {
@@ -2474,27 +2637,26 @@ static int ath10k_sta_state(struct ieee80211_hw *hw,
2474 /* 2637 /*
2475 * New station addition. 2638 * New station addition.
2476 */ 2639 */
2640 ath10k_dbg(ATH10K_DBG_MAC,
2641 "mac vdev %d peer create %pM (new sta)\n",
2642 arvif->vdev_id, sta->addr);
2643
2477 ret = ath10k_peer_create(ar, arvif->vdev_id, sta->addr); 2644 ret = ath10k_peer_create(ar, arvif->vdev_id, sta->addr);
2478 if (ret) 2645 if (ret)
2479 ath10k_warn("Failed to add peer: %pM for VDEV: %d\n", 2646 ath10k_warn("Failed to add peer: %pM for VDEV: %d\n",
2480 sta->addr, arvif->vdev_id); 2647 sta->addr, arvif->vdev_id);
2481 else
2482 ath10k_dbg(ATH10K_DBG_MAC,
2483 "Added peer: %pM for VDEV: %d\n",
2484 sta->addr, arvif->vdev_id);
2485 } else if ((old_state == IEEE80211_STA_NONE && 2648 } else if ((old_state == IEEE80211_STA_NONE &&
2486 new_state == IEEE80211_STA_NOTEXIST)) { 2649 new_state == IEEE80211_STA_NOTEXIST)) {
2487 /* 2650 /*
2488 * Existing station deletion. 2651 * Existing station deletion.
2489 */ 2652 */
2653 ath10k_dbg(ATH10K_DBG_MAC,
2654 "mac vdev %d peer delete %pM (sta gone)\n",
2655 arvif->vdev_id, sta->addr);
2490 ret = ath10k_peer_delete(ar, arvif->vdev_id, sta->addr); 2656 ret = ath10k_peer_delete(ar, arvif->vdev_id, sta->addr);
2491 if (ret) 2657 if (ret)
2492 ath10k_warn("Failed to delete peer: %pM for VDEV: %d\n", 2658 ath10k_warn("Failed to delete peer: %pM for VDEV: %d\n",
2493 sta->addr, arvif->vdev_id); 2659 sta->addr, arvif->vdev_id);
2494 else
2495 ath10k_dbg(ATH10K_DBG_MAC,
2496 "Removed peer: %pM for VDEV: %d\n",
2497 sta->addr, arvif->vdev_id);
2498 2660
2499 if (vif->type == NL80211_IFTYPE_STATION) 2661 if (vif->type == NL80211_IFTYPE_STATION)
2500 ath10k_bss_disassoc(hw, vif); 2662 ath10k_bss_disassoc(hw, vif);
@@ -2505,14 +2667,13 @@ static int ath10k_sta_state(struct ieee80211_hw *hw,
2505 /* 2667 /*
2506 * New association. 2668 * New association.
2507 */ 2669 */
2670 ath10k_dbg(ATH10K_DBG_MAC, "mac sta %pM associated\n",
2671 sta->addr);
2672
2508 ret = ath10k_station_assoc(ar, arvif, sta); 2673 ret = ath10k_station_assoc(ar, arvif, sta);
2509 if (ret) 2674 if (ret)
2510 ath10k_warn("Failed to associate station: %pM\n", 2675 ath10k_warn("Failed to associate station: %pM\n",
2511 sta->addr); 2676 sta->addr);
2512 else
2513 ath10k_dbg(ATH10K_DBG_MAC,
2514 "Station %pM moved to assoc state\n",
2515 sta->addr);
2516 } else if (old_state == IEEE80211_STA_ASSOC && 2677 } else if (old_state == IEEE80211_STA_ASSOC &&
2517 new_state == IEEE80211_STA_AUTH && 2678 new_state == IEEE80211_STA_AUTH &&
2518 (vif->type == NL80211_IFTYPE_AP || 2679 (vif->type == NL80211_IFTYPE_AP ||
@@ -2520,14 +2681,13 @@ static int ath10k_sta_state(struct ieee80211_hw *hw,
2520 /* 2681 /*
2521 * Disassociation. 2682 * Disassociation.
2522 */ 2683 */
2684 ath10k_dbg(ATH10K_DBG_MAC, "mac sta %pM disassociated\n",
2685 sta->addr);
2686
2523 ret = ath10k_station_disassoc(ar, arvif, sta); 2687 ret = ath10k_station_disassoc(ar, arvif, sta);
2524 if (ret) 2688 if (ret)
2525 ath10k_warn("Failed to disassociate station: %pM\n", 2689 ath10k_warn("Failed to disassociate station: %pM\n",
2526 sta->addr); 2690 sta->addr);
2527 else
2528 ath10k_dbg(ATH10K_DBG_MAC,
2529 "Station %pM moved to disassociated state\n",
2530 sta->addr);
2531 } 2691 }
2532 2692
2533 mutex_unlock(&ar->conf_mutex); 2693 mutex_unlock(&ar->conf_mutex);
@@ -2732,88 +2892,51 @@ static int ath10k_cancel_remain_on_channel(struct ieee80211_hw *hw)
2732 * Both RTS and Fragmentation threshold are interface-specific 2892 * Both RTS and Fragmentation threshold are interface-specific
2733 * in ath10k, but device-specific in mac80211. 2893 * in ath10k, but device-specific in mac80211.
2734 */ 2894 */
2735static void ath10k_set_rts_iter(void *data, u8 *mac, struct ieee80211_vif *vif)
2736{
2737 struct ath10k_generic_iter *ar_iter = data;
2738 struct ath10k_vif *arvif = ath10k_vif_to_arvif(vif);
2739 u32 rts = ar_iter->ar->hw->wiphy->rts_threshold;
2740
2741 lockdep_assert_held(&arvif->ar->conf_mutex);
2742
2743 /* During HW reconfiguration mac80211 reports all interfaces that were
2744 * running until reconfiguration was started. Since FW doesn't have any
2745 * vdevs at this point we must not iterate over this interface list.
2746 * This setting will be updated upon add_interface(). */
2747 if (ar_iter->ar->state == ATH10K_STATE_RESTARTED)
2748 return;
2749
2750 ar_iter->ret = ath10k_mac_set_rts(arvif, rts);
2751 if (ar_iter->ret)
2752 ath10k_warn("Failed to set RTS threshold for VDEV: %d\n",
2753 arvif->vdev_id);
2754 else
2755 ath10k_dbg(ATH10K_DBG_MAC,
2756 "Set RTS threshold: %d for VDEV: %d\n",
2757 rts, arvif->vdev_id);
2758}
2759 2895
2760static int ath10k_set_rts_threshold(struct ieee80211_hw *hw, u32 value) 2896static int ath10k_set_rts_threshold(struct ieee80211_hw *hw, u32 value)
2761{ 2897{
2762 struct ath10k_generic_iter ar_iter;
2763 struct ath10k *ar = hw->priv; 2898 struct ath10k *ar = hw->priv;
2764 2899 struct ath10k_vif *arvif;
2765 memset(&ar_iter, 0, sizeof(struct ath10k_generic_iter)); 2900 int ret = 0;
2766 ar_iter.ar = ar;
2767 2901
2768 mutex_lock(&ar->conf_mutex); 2902 mutex_lock(&ar->conf_mutex);
2769 ieee80211_iterate_active_interfaces_atomic( 2903 list_for_each_entry(arvif, &ar->arvifs, list) {
2770 hw, IEEE80211_IFACE_ITER_NORMAL, 2904 ath10k_dbg(ATH10K_DBG_MAC, "mac vdev %d rts threshold %d\n",
2771 ath10k_set_rts_iter, &ar_iter); 2905 arvif->vdev_id, value);
2772 mutex_unlock(&ar->conf_mutex);
2773
2774 return ar_iter.ret;
2775}
2776 2906
2777static void ath10k_set_frag_iter(void *data, u8 *mac, struct ieee80211_vif *vif) 2907 ret = ath10k_mac_set_rts(arvif, value);
2778{ 2908 if (ret) {
2779 struct ath10k_generic_iter *ar_iter = data; 2909 ath10k_warn("could not set rts threshold for vdev %d (%d)\n",
2780 struct ath10k_vif *arvif = ath10k_vif_to_arvif(vif); 2910 arvif->vdev_id, ret);
2781 u32 frag = ar_iter->ar->hw->wiphy->frag_threshold; 2911 break;
2782 2912 }
2783 lockdep_assert_held(&arvif->ar->conf_mutex); 2913 }
2784 2914 mutex_unlock(&ar->conf_mutex);
2785 /* During HW reconfiguration mac80211 reports all interfaces that were
2786 * running until reconfiguration was started. Since FW doesn't have any
2787 * vdevs at this point we must not iterate over this interface list.
2788 * This setting will be updated upon add_interface(). */
2789 if (ar_iter->ar->state == ATH10K_STATE_RESTARTED)
2790 return;
2791 2915
2792 ar_iter->ret = ath10k_mac_set_frag(arvif, frag); 2916 return ret;
2793 if (ar_iter->ret)
2794 ath10k_warn("Failed to set frag threshold for VDEV: %d\n",
2795 arvif->vdev_id);
2796 else
2797 ath10k_dbg(ATH10K_DBG_MAC,
2798 "Set frag threshold: %d for VDEV: %d\n",
2799 frag, arvif->vdev_id);
2800} 2917}
2801 2918
2802static int ath10k_set_frag_threshold(struct ieee80211_hw *hw, u32 value) 2919static int ath10k_set_frag_threshold(struct ieee80211_hw *hw, u32 value)
2803{ 2920{
2804 struct ath10k_generic_iter ar_iter;
2805 struct ath10k *ar = hw->priv; 2921 struct ath10k *ar = hw->priv;
2806 2922 struct ath10k_vif *arvif;
2807 memset(&ar_iter, 0, sizeof(struct ath10k_generic_iter)); 2923 int ret = 0;
2808 ar_iter.ar = ar;
2809 2924
2810 mutex_lock(&ar->conf_mutex); 2925 mutex_lock(&ar->conf_mutex);
2811 ieee80211_iterate_active_interfaces_atomic( 2926 list_for_each_entry(arvif, &ar->arvifs, list) {
2812 hw, IEEE80211_IFACE_ITER_NORMAL, 2927 ath10k_dbg(ATH10K_DBG_MAC, "mac vdev %d fragmentation threshold %d\n",
2813 ath10k_set_frag_iter, &ar_iter); 2928 arvif->vdev_id, value);
2929
2930 ret = ath10k_mac_set_rts(arvif, value);
2931 if (ret) {
2932 ath10k_warn("could not set fragmentation threshold for vdev %d (%d)\n",
2933 arvif->vdev_id, ret);
2934 break;
2935 }
2936 }
2814 mutex_unlock(&ar->conf_mutex); 2937 mutex_unlock(&ar->conf_mutex);
2815 2938
2816 return ar_iter.ret; 2939 return ret;
2817} 2940}
2818 2941
2819static void ath10k_flush(struct ieee80211_hw *hw, u32 queues, bool drop) 2942static void ath10k_flush(struct ieee80211_hw *hw, u32 queues, bool drop)
@@ -2836,8 +2959,7 @@ static void ath10k_flush(struct ieee80211_hw *hw, u32 queues, bool drop)
2836 bool empty; 2959 bool empty;
2837 2960
2838 spin_lock_bh(&ar->htt.tx_lock); 2961 spin_lock_bh(&ar->htt.tx_lock);
2839 empty = bitmap_empty(ar->htt.used_msdu_ids, 2962 empty = (ar->htt.num_pending_tx == 0);
2840 ar->htt.max_num_pending_tx);
2841 spin_unlock_bh(&ar->htt.tx_lock); 2963 spin_unlock_bh(&ar->htt.tx_lock);
2842 2964
2843 skip = (ar->state == ATH10K_STATE_WEDGED); 2965 skip = (ar->state == ATH10K_STATE_WEDGED);
@@ -3326,6 +3448,10 @@ int ath10k_mac_register(struct ath10k *ar)
3326 IEEE80211_HW_WANT_MONITOR_VIF | 3448 IEEE80211_HW_WANT_MONITOR_VIF |
3327 IEEE80211_HW_AP_LINK_PS; 3449 IEEE80211_HW_AP_LINK_PS;
3328 3450
3451 /* MSDU can have HTT TX fragment pushed in front. The additional 4
3452 * bytes is used for padding/alignment if necessary. */
3453 ar->hw->extra_tx_headroom += sizeof(struct htt_data_tx_desc_frag)*2 + 4;
3454
3329 if (ar->ht_cap_info & WMI_HT_CAP_DYNAMIC_SMPS) 3455 if (ar->ht_cap_info & WMI_HT_CAP_DYNAMIC_SMPS)
3330 ar->hw->flags |= IEEE80211_HW_SUPPORTS_DYNAMIC_SMPS; 3456 ar->hw->flags |= IEEE80211_HW_SUPPORTS_DYNAMIC_SMPS;
3331 3457
diff --git a/drivers/net/wireless/ath/ath10k/mac.h b/drivers/net/wireless/ath/ath10k/mac.h
index 6fce9bfb19a5..ba1021997b8f 100644
--- a/drivers/net/wireless/ath/ath10k/mac.h
+++ b/drivers/net/wireless/ath/ath10k/mac.h
@@ -34,6 +34,8 @@ struct ath10k_vif *ath10k_get_arvif(struct ath10k *ar, u32 vdev_id);
34void ath10k_reset_scan(unsigned long ptr); 34void ath10k_reset_scan(unsigned long ptr);
35void ath10k_offchan_tx_purge(struct ath10k *ar); 35void ath10k_offchan_tx_purge(struct ath10k *ar);
36void ath10k_offchan_tx_work(struct work_struct *work); 36void ath10k_offchan_tx_work(struct work_struct *work);
37void ath10k_mgmt_over_wmi_tx_purge(struct ath10k *ar);
38void ath10k_mgmt_over_wmi_tx_work(struct work_struct *work);
37void ath10k_halt(struct ath10k *ar); 39void ath10k_halt(struct ath10k *ar);
38 40
39static inline struct ath10k_vif *ath10k_vif_to_arvif(struct ieee80211_vif *vif) 41static inline struct ath10k_vif *ath10k_vif_to_arvif(struct ieee80211_vif *vif)
diff --git a/drivers/net/wireless/ath/ath10k/pci.c b/drivers/net/wireless/ath/ath10k/pci.c
index e2f9ef50b1bd..f8d59c7b9082 100644
--- a/drivers/net/wireless/ath/ath10k/pci.c
+++ b/drivers/net/wireless/ath/ath10k/pci.c
@@ -36,11 +36,9 @@ static unsigned int ath10k_target_ps;
36module_param(ath10k_target_ps, uint, 0644); 36module_param(ath10k_target_ps, uint, 0644);
37MODULE_PARM_DESC(ath10k_target_ps, "Enable ath10k Target (SoC) PS option"); 37MODULE_PARM_DESC(ath10k_target_ps, "Enable ath10k Target (SoC) PS option");
38 38
39#define QCA988X_1_0_DEVICE_ID (0xabcd)
40#define QCA988X_2_0_DEVICE_ID (0x003c) 39#define QCA988X_2_0_DEVICE_ID (0x003c)
41 40
42static DEFINE_PCI_DEVICE_TABLE(ath10k_pci_id_table) = { 41static DEFINE_PCI_DEVICE_TABLE(ath10k_pci_id_table) = {
43 { PCI_VDEVICE(ATHEROS, QCA988X_1_0_DEVICE_ID) }, /* PCI-E QCA988X V1 */
44 { PCI_VDEVICE(ATHEROS, QCA988X_2_0_DEVICE_ID) }, /* PCI-E QCA988X V2 */ 42 { PCI_VDEVICE(ATHEROS, QCA988X_2_0_DEVICE_ID) }, /* PCI-E QCA988X V2 */
45 {0} 43 {0}
46}; 44};
@@ -50,9 +48,9 @@ static int ath10k_pci_diag_read_access(struct ath10k *ar, u32 address,
50 48
51static void ath10k_pci_process_ce(struct ath10k *ar); 49static void ath10k_pci_process_ce(struct ath10k *ar);
52static int ath10k_pci_post_rx(struct ath10k *ar); 50static int ath10k_pci_post_rx(struct ath10k *ar);
53static int ath10k_pci_post_rx_pipe(struct hif_ce_pipe_info *pipe_info, 51static int ath10k_pci_post_rx_pipe(struct ath10k_pci_pipe *pipe_info,
54 int num); 52 int num);
55static void ath10k_pci_rx_pipe_cleanup(struct hif_ce_pipe_info *pipe_info); 53static void ath10k_pci_rx_pipe_cleanup(struct ath10k_pci_pipe *pipe_info);
56static void ath10k_pci_stop_ce(struct ath10k *ar); 54static void ath10k_pci_stop_ce(struct ath10k *ar);
57static void ath10k_pci_device_reset(struct ath10k *ar); 55static void ath10k_pci_device_reset(struct ath10k *ar);
58static int ath10k_pci_reset_target(struct ath10k *ar); 56static int ath10k_pci_reset_target(struct ath10k *ar);
@@ -60,43 +58,145 @@ static int ath10k_pci_start_intr(struct ath10k *ar);
60static void ath10k_pci_stop_intr(struct ath10k *ar); 58static void ath10k_pci_stop_intr(struct ath10k *ar);
61 59
62static const struct ce_attr host_ce_config_wlan[] = { 60static const struct ce_attr host_ce_config_wlan[] = {
63 /* host->target HTC control and raw streams */ 61 /* CE0: host->target HTC control and raw streams */
64 { /* CE0 */ CE_ATTR_FLAGS, 0, 16, 256, 0, NULL,}, 62 {
65 /* could be moved to share CE3 */ 63 .flags = CE_ATTR_FLAGS,
66 /* target->host HTT + HTC control */ 64 .src_nentries = 16,
67 { /* CE1 */ CE_ATTR_FLAGS, 0, 0, 512, 512, NULL,}, 65 .src_sz_max = 256,
68 /* target->host WMI */ 66 .dest_nentries = 0,
69 { /* CE2 */ CE_ATTR_FLAGS, 0, 0, 2048, 32, NULL,}, 67 },
70 /* host->target WMI */ 68
71 { /* CE3 */ CE_ATTR_FLAGS, 0, 32, 2048, 0, NULL,}, 69 /* CE1: target->host HTT + HTC control */
72 /* host->target HTT */ 70 {
73 { /* CE4 */ CE_ATTR_FLAGS | CE_ATTR_DIS_INTR, 0, 71 .flags = CE_ATTR_FLAGS,
74 CE_HTT_H2T_MSG_SRC_NENTRIES, 256, 0, NULL,}, 72 .src_nentries = 0,
75 /* unused */ 73 .src_sz_max = 512,
76 { /* CE5 */ CE_ATTR_FLAGS, 0, 0, 0, 0, NULL,}, 74 .dest_nentries = 512,
77 /* Target autonomous hif_memcpy */ 75 },
78 { /* CE6 */ CE_ATTR_FLAGS, 0, 0, 0, 0, NULL,}, 76
79 /* ce_diag, the Diagnostic Window */ 77 /* CE2: target->host WMI */
80 { /* CE7 */ CE_ATTR_FLAGS, 0, 2, DIAG_TRANSFER_LIMIT, 2, NULL,}, 78 {
79 .flags = CE_ATTR_FLAGS,
80 .src_nentries = 0,
81 .src_sz_max = 2048,
82 .dest_nentries = 32,
83 },
84
85 /* CE3: host->target WMI */
86 {
87 .flags = CE_ATTR_FLAGS,
88 .src_nentries = 32,
89 .src_sz_max = 2048,
90 .dest_nentries = 0,
91 },
92
93 /* CE4: host->target HTT */
94 {
95 .flags = CE_ATTR_FLAGS | CE_ATTR_DIS_INTR,
96 .src_nentries = CE_HTT_H2T_MSG_SRC_NENTRIES,
97 .src_sz_max = 256,
98 .dest_nentries = 0,
99 },
100
101 /* CE5: unused */
102 {
103 .flags = CE_ATTR_FLAGS,
104 .src_nentries = 0,
105 .src_sz_max = 0,
106 .dest_nentries = 0,
107 },
108
109 /* CE6: target autonomous hif_memcpy */
110 {
111 .flags = CE_ATTR_FLAGS,
112 .src_nentries = 0,
113 .src_sz_max = 0,
114 .dest_nentries = 0,
115 },
116
117 /* CE7: ce_diag, the Diagnostic Window */
118 {
119 .flags = CE_ATTR_FLAGS,
120 .src_nentries = 2,
121 .src_sz_max = DIAG_TRANSFER_LIMIT,
122 .dest_nentries = 2,
123 },
81}; 124};
82 125
83/* Target firmware's Copy Engine configuration. */ 126/* Target firmware's Copy Engine configuration. */
84static const struct ce_pipe_config target_ce_config_wlan[] = { 127static const struct ce_pipe_config target_ce_config_wlan[] = {
85 /* host->target HTC control and raw streams */ 128 /* CE0: host->target HTC control and raw streams */
86 { /* CE0 */ 0, PIPEDIR_OUT, 32, 256, CE_ATTR_FLAGS, 0,}, 129 {
87 /* target->host HTT + HTC control */ 130 .pipenum = 0,
88 { /* CE1 */ 1, PIPEDIR_IN, 32, 512, CE_ATTR_FLAGS, 0,}, 131 .pipedir = PIPEDIR_OUT,
89 /* target->host WMI */ 132 .nentries = 32,
90 { /* CE2 */ 2, PIPEDIR_IN, 32, 2048, CE_ATTR_FLAGS, 0,}, 133 .nbytes_max = 256,
91 /* host->target WMI */ 134 .flags = CE_ATTR_FLAGS,
92 { /* CE3 */ 3, PIPEDIR_OUT, 32, 2048, CE_ATTR_FLAGS, 0,}, 135 .reserved = 0,
93 /* host->target HTT */ 136 },
94 { /* CE4 */ 4, PIPEDIR_OUT, 256, 256, CE_ATTR_FLAGS, 0,}, 137
138 /* CE1: target->host HTT + HTC control */
139 {
140 .pipenum = 1,
141 .pipedir = PIPEDIR_IN,
142 .nentries = 32,
143 .nbytes_max = 512,
144 .flags = CE_ATTR_FLAGS,
145 .reserved = 0,
146 },
147
148 /* CE2: target->host WMI */
149 {
150 .pipenum = 2,
151 .pipedir = PIPEDIR_IN,
152 .nentries = 32,
153 .nbytes_max = 2048,
154 .flags = CE_ATTR_FLAGS,
155 .reserved = 0,
156 },
157
158 /* CE3: host->target WMI */
159 {
160 .pipenum = 3,
161 .pipedir = PIPEDIR_OUT,
162 .nentries = 32,
163 .nbytes_max = 2048,
164 .flags = CE_ATTR_FLAGS,
165 .reserved = 0,
166 },
167
168 /* CE4: host->target HTT */
169 {
170 .pipenum = 4,
171 .pipedir = PIPEDIR_OUT,
172 .nentries = 256,
173 .nbytes_max = 256,
174 .flags = CE_ATTR_FLAGS,
175 .reserved = 0,
176 },
177
95 /* NB: 50% of src nentries, since tx has 2 frags */ 178 /* NB: 50% of src nentries, since tx has 2 frags */
96 /* unused */ 179
97 { /* CE5 */ 5, PIPEDIR_OUT, 32, 2048, CE_ATTR_FLAGS, 0,}, 180 /* CE5: unused */
98 /* Reserved for target autonomous hif_memcpy */ 181 {
99 { /* CE6 */ 6, PIPEDIR_INOUT, 32, 4096, CE_ATTR_FLAGS, 0,}, 182 .pipenum = 5,
183 .pipedir = PIPEDIR_OUT,
184 .nentries = 32,
185 .nbytes_max = 2048,
186 .flags = CE_ATTR_FLAGS,
187 .reserved = 0,
188 },
189
190 /* CE6: Reserved for target autonomous hif_memcpy */
191 {
192 .pipenum = 6,
193 .pipedir = PIPEDIR_INOUT,
194 .nentries = 32,
195 .nbytes_max = 4096,
196 .flags = CE_ATTR_FLAGS,
197 .reserved = 0,
198 },
199
100 /* CE7 used only by Host */ 200 /* CE7 used only by Host */
101}; 201};
102 202
@@ -114,7 +214,7 @@ static int ath10k_pci_diag_read_mem(struct ath10k *ar, u32 address, void *data,
114 unsigned int completed_nbytes, orig_nbytes, remaining_bytes; 214 unsigned int completed_nbytes, orig_nbytes, remaining_bytes;
115 unsigned int id; 215 unsigned int id;
116 unsigned int flags; 216 unsigned int flags;
117 struct ce_state *ce_diag; 217 struct ath10k_ce_pipe *ce_diag;
118 /* Host buffer address in CE space */ 218 /* Host buffer address in CE space */
119 u32 ce_data; 219 u32 ce_data;
120 dma_addr_t ce_data_base = 0; 220 dma_addr_t ce_data_base = 0;
@@ -278,7 +378,7 @@ static int ath10k_pci_diag_write_mem(struct ath10k *ar, u32 address,
278 unsigned int completed_nbytes, orig_nbytes, remaining_bytes; 378 unsigned int completed_nbytes, orig_nbytes, remaining_bytes;
279 unsigned int id; 379 unsigned int id;
280 unsigned int flags; 380 unsigned int flags;
281 struct ce_state *ce_diag; 381 struct ath10k_ce_pipe *ce_diag;
282 void *data_buf = NULL; 382 void *data_buf = NULL;
283 u32 ce_data; /* Host buffer address in CE space */ 383 u32 ce_data; /* Host buffer address in CE space */
284 dma_addr_t ce_data_base = 0; 384 dma_addr_t ce_data_base = 0;
@@ -437,7 +537,7 @@ static void ath10k_pci_wait(struct ath10k *ar)
437 ath10k_warn("Unable to wakeup target\n"); 537 ath10k_warn("Unable to wakeup target\n");
438} 538}
439 539
440void ath10k_do_pci_wake(struct ath10k *ar) 540int ath10k_do_pci_wake(struct ath10k *ar)
441{ 541{
442 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 542 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
443 void __iomem *pci_addr = ar_pci->mem; 543 void __iomem *pci_addr = ar_pci->mem;
@@ -453,18 +553,19 @@ void ath10k_do_pci_wake(struct ath10k *ar)
453 atomic_inc(&ar_pci->keep_awake_count); 553 atomic_inc(&ar_pci->keep_awake_count);
454 554
455 if (ar_pci->verified_awake) 555 if (ar_pci->verified_awake)
456 return; 556 return 0;
457 557
458 for (;;) { 558 for (;;) {
459 if (ath10k_pci_target_is_awake(ar)) { 559 if (ath10k_pci_target_is_awake(ar)) {
460 ar_pci->verified_awake = true; 560 ar_pci->verified_awake = true;
461 break; 561 return 0;
462 } 562 }
463 563
464 if (tot_delay > PCIE_WAKE_TIMEOUT) { 564 if (tot_delay > PCIE_WAKE_TIMEOUT) {
465 ath10k_warn("target takes too long to wake up (awake count %d)\n", 565 ath10k_warn("target took longer %d us to wake up (awake count %d)\n",
566 PCIE_WAKE_TIMEOUT,
466 atomic_read(&ar_pci->keep_awake_count)); 567 atomic_read(&ar_pci->keep_awake_count));
467 break; 568 return -ETIMEDOUT;
468 } 569 }
469 570
470 udelay(curr_delay); 571 udelay(curr_delay);
@@ -493,7 +594,7 @@ void ath10k_do_pci_sleep(struct ath10k *ar)
493 * FIXME: Handle OOM properly. 594 * FIXME: Handle OOM properly.
494 */ 595 */
495static inline 596static inline
496struct ath10k_pci_compl *get_free_compl(struct hif_ce_pipe_info *pipe_info) 597struct ath10k_pci_compl *get_free_compl(struct ath10k_pci_pipe *pipe_info)
497{ 598{
498 struct ath10k_pci_compl *compl = NULL; 599 struct ath10k_pci_compl *compl = NULL;
499 600
@@ -511,39 +612,28 @@ exit:
511} 612}
512 613
513/* Called by lower (CE) layer when a send to Target completes. */ 614/* Called by lower (CE) layer when a send to Target completes. */
514static void ath10k_pci_ce_send_done(struct ce_state *ce_state, 615static void ath10k_pci_ce_send_done(struct ath10k_ce_pipe *ce_state)
515 void *transfer_context,
516 u32 ce_data,
517 unsigned int nbytes,
518 unsigned int transfer_id)
519{ 616{
520 struct ath10k *ar = ce_state->ar; 617 struct ath10k *ar = ce_state->ar;
521 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 618 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
522 struct hif_ce_pipe_info *pipe_info = &ar_pci->pipe_info[ce_state->id]; 619 struct ath10k_pci_pipe *pipe_info = &ar_pci->pipe_info[ce_state->id];
523 struct ath10k_pci_compl *compl; 620 struct ath10k_pci_compl *compl;
524 bool process = false; 621 void *transfer_context;
525 622 u32 ce_data;
526 do { 623 unsigned int nbytes;
527 /* 624 unsigned int transfer_id;
528 * For the send completion of an item in sendlist, just
529 * increment num_sends_allowed. The upper layer callback will
530 * be triggered when last fragment is done with send.
531 */
532 if (transfer_context == CE_SENDLIST_ITEM_CTXT) {
533 spin_lock_bh(&pipe_info->pipe_lock);
534 pipe_info->num_sends_allowed++;
535 spin_unlock_bh(&pipe_info->pipe_lock);
536 continue;
537 }
538 625
626 while (ath10k_ce_completed_send_next(ce_state, &transfer_context,
627 &ce_data, &nbytes,
628 &transfer_id) == 0) {
539 compl = get_free_compl(pipe_info); 629 compl = get_free_compl(pipe_info);
540 if (!compl) 630 if (!compl)
541 break; 631 break;
542 632
543 compl->send_or_recv = HIF_CE_COMPLETE_SEND; 633 compl->state = ATH10K_PCI_COMPL_SEND;
544 compl->ce_state = ce_state; 634 compl->ce_state = ce_state;
545 compl->pipe_info = pipe_info; 635 compl->pipe_info = pipe_info;
546 compl->transfer_context = transfer_context; 636 compl->skb = transfer_context;
547 compl->nbytes = nbytes; 637 compl->nbytes = nbytes;
548 compl->transfer_id = transfer_id; 638 compl->transfer_id = transfer_id;
549 compl->flags = 0; 639 compl->flags = 0;
@@ -554,46 +644,36 @@ static void ath10k_pci_ce_send_done(struct ce_state *ce_state,
554 spin_lock_bh(&ar_pci->compl_lock); 644 spin_lock_bh(&ar_pci->compl_lock);
555 list_add_tail(&compl->list, &ar_pci->compl_process); 645 list_add_tail(&compl->list, &ar_pci->compl_process);
556 spin_unlock_bh(&ar_pci->compl_lock); 646 spin_unlock_bh(&ar_pci->compl_lock);
557 647 }
558 process = true;
559 } while (ath10k_ce_completed_send_next(ce_state,
560 &transfer_context,
561 &ce_data, &nbytes,
562 &transfer_id) == 0);
563
564 /*
565 * If only some of the items within a sendlist have completed,
566 * don't invoke completion processing until the entire sendlist
567 * has been sent.
568 */
569 if (!process)
570 return;
571 648
572 ath10k_pci_process_ce(ar); 649 ath10k_pci_process_ce(ar);
573} 650}
574 651
575/* Called by lower (CE) layer when data is received from the Target. */ 652/* Called by lower (CE) layer when data is received from the Target. */
576static void ath10k_pci_ce_recv_data(struct ce_state *ce_state, 653static void ath10k_pci_ce_recv_data(struct ath10k_ce_pipe *ce_state)
577 void *transfer_context, u32 ce_data,
578 unsigned int nbytes,
579 unsigned int transfer_id,
580 unsigned int flags)
581{ 654{
582 struct ath10k *ar = ce_state->ar; 655 struct ath10k *ar = ce_state->ar;
583 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 656 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
584 struct hif_ce_pipe_info *pipe_info = &ar_pci->pipe_info[ce_state->id]; 657 struct ath10k_pci_pipe *pipe_info = &ar_pci->pipe_info[ce_state->id];
585 struct ath10k_pci_compl *compl; 658 struct ath10k_pci_compl *compl;
586 struct sk_buff *skb; 659 struct sk_buff *skb;
660 void *transfer_context;
661 u32 ce_data;
662 unsigned int nbytes;
663 unsigned int transfer_id;
664 unsigned int flags;
587 665
588 do { 666 while (ath10k_ce_completed_recv_next(ce_state, &transfer_context,
667 &ce_data, &nbytes, &transfer_id,
668 &flags) == 0) {
589 compl = get_free_compl(pipe_info); 669 compl = get_free_compl(pipe_info);
590 if (!compl) 670 if (!compl)
591 break; 671 break;
592 672
593 compl->send_or_recv = HIF_CE_COMPLETE_RECV; 673 compl->state = ATH10K_PCI_COMPL_RECV;
594 compl->ce_state = ce_state; 674 compl->ce_state = ce_state;
595 compl->pipe_info = pipe_info; 675 compl->pipe_info = pipe_info;
596 compl->transfer_context = transfer_context; 676 compl->skb = transfer_context;
597 compl->nbytes = nbytes; 677 compl->nbytes = nbytes;
598 compl->transfer_id = transfer_id; 678 compl->transfer_id = transfer_id;
599 compl->flags = flags; 679 compl->flags = flags;
@@ -608,12 +688,7 @@ static void ath10k_pci_ce_recv_data(struct ce_state *ce_state,
608 spin_lock_bh(&ar_pci->compl_lock); 688 spin_lock_bh(&ar_pci->compl_lock);
609 list_add_tail(&compl->list, &ar_pci->compl_process); 689 list_add_tail(&compl->list, &ar_pci->compl_process);
610 spin_unlock_bh(&ar_pci->compl_lock); 690 spin_unlock_bh(&ar_pci->compl_lock);
611 691 }
612 } while (ath10k_ce_completed_recv_next(ce_state,
613 &transfer_context,
614 &ce_data, &nbytes,
615 &transfer_id,
616 &flags) == 0);
617 692
618 ath10k_pci_process_ce(ar); 693 ath10k_pci_process_ce(ar);
619} 694}
@@ -625,15 +700,12 @@ static int ath10k_pci_hif_send_head(struct ath10k *ar, u8 pipe_id,
625{ 700{
626 struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(nbuf); 701 struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(nbuf);
627 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 702 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
628 struct hif_ce_pipe_info *pipe_info = &(ar_pci->pipe_info[pipe_id]); 703 struct ath10k_pci_pipe *pipe_info = &(ar_pci->pipe_info[pipe_id]);
629 struct ce_state *ce_hdl = pipe_info->ce_hdl; 704 struct ath10k_ce_pipe *ce_hdl = pipe_info->ce_hdl;
630 struct ce_sendlist sendlist;
631 unsigned int len; 705 unsigned int len;
632 u32 flags = 0; 706 u32 flags = 0;
633 int ret; 707 int ret;
634 708
635 memset(&sendlist, 0, sizeof(struct ce_sendlist));
636
637 len = min(bytes, nbuf->len); 709 len = min(bytes, nbuf->len);
638 bytes -= len; 710 bytes -= len;
639 711
@@ -648,19 +720,8 @@ static int ath10k_pci_hif_send_head(struct ath10k *ar, u8 pipe_id,
648 "ath10k tx: data: ", 720 "ath10k tx: data: ",
649 nbuf->data, nbuf->len); 721 nbuf->data, nbuf->len);
650 722
651 ath10k_ce_sendlist_buf_add(&sendlist, skb_cb->paddr, len, flags); 723 ret = ath10k_ce_send(ce_hdl, nbuf, skb_cb->paddr, len, transfer_id,
652 724 flags);
653 /* Make sure we have resources to handle this request */
654 spin_lock_bh(&pipe_info->pipe_lock);
655 if (!pipe_info->num_sends_allowed) {
656 ath10k_warn("Pipe: %d is full\n", pipe_id);
657 spin_unlock_bh(&pipe_info->pipe_lock);
658 return -ENOSR;
659 }
660 pipe_info->num_sends_allowed--;
661 spin_unlock_bh(&pipe_info->pipe_lock);
662
663 ret = ath10k_ce_sendlist_send(ce_hdl, nbuf, &sendlist, transfer_id);
664 if (ret) 725 if (ret)
665 ath10k_warn("CE send failed: %p\n", nbuf); 726 ath10k_warn("CE send failed: %p\n", nbuf);
666 727
@@ -670,14 +731,7 @@ static int ath10k_pci_hif_send_head(struct ath10k *ar, u8 pipe_id,
670static u16 ath10k_pci_hif_get_free_queue_number(struct ath10k *ar, u8 pipe) 731static u16 ath10k_pci_hif_get_free_queue_number(struct ath10k *ar, u8 pipe)
671{ 732{
672 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 733 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
673 struct hif_ce_pipe_info *pipe_info = &(ar_pci->pipe_info[pipe]); 734 return ath10k_ce_num_free_src_entries(ar_pci->pipe_info[pipe].ce_hdl);
674 int ret;
675
676 spin_lock_bh(&pipe_info->pipe_lock);
677 ret = pipe_info->num_sends_allowed;
678 spin_unlock_bh(&pipe_info->pipe_lock);
679
680 return ret;
681} 735}
682 736
683static void ath10k_pci_hif_dump_area(struct ath10k *ar) 737static void ath10k_pci_hif_dump_area(struct ath10k *ar)
@@ -764,9 +818,9 @@ static void ath10k_pci_hif_set_callbacks(struct ath10k *ar,
764static int ath10k_pci_start_ce(struct ath10k *ar) 818static int ath10k_pci_start_ce(struct ath10k *ar)
765{ 819{
766 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 820 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
767 struct ce_state *ce_diag = ar_pci->ce_diag; 821 struct ath10k_ce_pipe *ce_diag = ar_pci->ce_diag;
768 const struct ce_attr *attr; 822 const struct ce_attr *attr;
769 struct hif_ce_pipe_info *pipe_info; 823 struct ath10k_pci_pipe *pipe_info;
770 struct ath10k_pci_compl *compl; 824 struct ath10k_pci_compl *compl;
771 int i, pipe_num, completions, disable_interrupts; 825 int i, pipe_num, completions, disable_interrupts;
772 826
@@ -792,7 +846,6 @@ static int ath10k_pci_start_ce(struct ath10k *ar)
792 ath10k_pci_ce_send_done, 846 ath10k_pci_ce_send_done,
793 disable_interrupts); 847 disable_interrupts);
794 completions += attr->src_nentries; 848 completions += attr->src_nentries;
795 pipe_info->num_sends_allowed = attr->src_nentries - 1;
796 } 849 }
797 850
798 if (attr->dest_nentries) { 851 if (attr->dest_nentries) {
@@ -805,15 +858,14 @@ static int ath10k_pci_start_ce(struct ath10k *ar)
805 continue; 858 continue;
806 859
807 for (i = 0; i < completions; i++) { 860 for (i = 0; i < completions; i++) {
808 compl = kmalloc(sizeof(struct ath10k_pci_compl), 861 compl = kmalloc(sizeof(*compl), GFP_KERNEL);
809 GFP_KERNEL);
810 if (!compl) { 862 if (!compl) {
811 ath10k_warn("No memory for completion state\n"); 863 ath10k_warn("No memory for completion state\n");
812 ath10k_pci_stop_ce(ar); 864 ath10k_pci_stop_ce(ar);
813 return -ENOMEM; 865 return -ENOMEM;
814 } 866 }
815 867
816 compl->send_or_recv = HIF_CE_COMPLETE_FREE; 868 compl->state = ATH10K_PCI_COMPL_FREE;
817 list_add_tail(&compl->list, &pipe_info->compl_free); 869 list_add_tail(&compl->list, &pipe_info->compl_free);
818 } 870 }
819 } 871 }
@@ -840,7 +892,7 @@ static void ath10k_pci_stop_ce(struct ath10k *ar)
840 * their associated resources */ 892 * their associated resources */
841 spin_lock_bh(&ar_pci->compl_lock); 893 spin_lock_bh(&ar_pci->compl_lock);
842 list_for_each_entry(compl, &ar_pci->compl_process, list) { 894 list_for_each_entry(compl, &ar_pci->compl_process, list) {
843 skb = (struct sk_buff *)compl->transfer_context; 895 skb = compl->skb;
844 ATH10K_SKB_CB(skb)->is_aborted = true; 896 ATH10K_SKB_CB(skb)->is_aborted = true;
845 } 897 }
846 spin_unlock_bh(&ar_pci->compl_lock); 898 spin_unlock_bh(&ar_pci->compl_lock);
@@ -850,7 +902,7 @@ static void ath10k_pci_cleanup_ce(struct ath10k *ar)
850{ 902{
851 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 903 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
852 struct ath10k_pci_compl *compl, *tmp; 904 struct ath10k_pci_compl *compl, *tmp;
853 struct hif_ce_pipe_info *pipe_info; 905 struct ath10k_pci_pipe *pipe_info;
854 struct sk_buff *netbuf; 906 struct sk_buff *netbuf;
855 int pipe_num; 907 int pipe_num;
856 908
@@ -861,7 +913,7 @@ static void ath10k_pci_cleanup_ce(struct ath10k *ar)
861 913
862 list_for_each_entry_safe(compl, tmp, &ar_pci->compl_process, list) { 914 list_for_each_entry_safe(compl, tmp, &ar_pci->compl_process, list) {
863 list_del(&compl->list); 915 list_del(&compl->list);
864 netbuf = (struct sk_buff *)compl->transfer_context; 916 netbuf = compl->skb;
865 dev_kfree_skb_any(netbuf); 917 dev_kfree_skb_any(netbuf);
866 kfree(compl); 918 kfree(compl);
867 } 919 }
@@ -912,12 +964,14 @@ static void ath10k_pci_process_ce(struct ath10k *ar)
912 list_del(&compl->list); 964 list_del(&compl->list);
913 spin_unlock_bh(&ar_pci->compl_lock); 965 spin_unlock_bh(&ar_pci->compl_lock);
914 966
915 if (compl->send_or_recv == HIF_CE_COMPLETE_SEND) { 967 switch (compl->state) {
968 case ATH10K_PCI_COMPL_SEND:
916 cb->tx_completion(ar, 969 cb->tx_completion(ar,
917 compl->transfer_context, 970 compl->skb,
918 compl->transfer_id); 971 compl->transfer_id);
919 send_done = 1; 972 send_done = 1;
920 } else { 973 break;
974 case ATH10K_PCI_COMPL_RECV:
921 ret = ath10k_pci_post_rx_pipe(compl->pipe_info, 1); 975 ret = ath10k_pci_post_rx_pipe(compl->pipe_info, 1);
922 if (ret) { 976 if (ret) {
923 ath10k_warn("Unable to post recv buffer for pipe: %d\n", 977 ath10k_warn("Unable to post recv buffer for pipe: %d\n",
@@ -925,7 +979,7 @@ static void ath10k_pci_process_ce(struct ath10k *ar)
925 break; 979 break;
926 } 980 }
927 981
928 skb = (struct sk_buff *)compl->transfer_context; 982 skb = compl->skb;
929 nbytes = compl->nbytes; 983 nbytes = compl->nbytes;
930 984
931 ath10k_dbg(ATH10K_DBG_PCI, 985 ath10k_dbg(ATH10K_DBG_PCI,
@@ -944,16 +998,23 @@ static void ath10k_pci_process_ce(struct ath10k *ar)
944 nbytes, 998 nbytes,
945 skb->len + skb_tailroom(skb)); 999 skb->len + skb_tailroom(skb));
946 } 1000 }
1001 break;
1002 case ATH10K_PCI_COMPL_FREE:
1003 ath10k_warn("free completion cannot be processed\n");
1004 break;
1005 default:
1006 ath10k_warn("invalid completion state (%d)\n",
1007 compl->state);
1008 break;
947 } 1009 }
948 1010
949 compl->send_or_recv = HIF_CE_COMPLETE_FREE; 1011 compl->state = ATH10K_PCI_COMPL_FREE;
950 1012
951 /* 1013 /*
952 * Add completion back to the pipe's free list. 1014 * Add completion back to the pipe's free list.
953 */ 1015 */
954 spin_lock_bh(&compl->pipe_info->pipe_lock); 1016 spin_lock_bh(&compl->pipe_info->pipe_lock);
955 list_add_tail(&compl->list, &compl->pipe_info->compl_free); 1017 list_add_tail(&compl->list, &compl->pipe_info->compl_free);
956 compl->pipe_info->num_sends_allowed += send_done;
957 spin_unlock_bh(&compl->pipe_info->pipe_lock); 1018 spin_unlock_bh(&compl->pipe_info->pipe_lock);
958 } 1019 }
959 1020
@@ -1037,12 +1098,12 @@ static void ath10k_pci_hif_get_default_pipe(struct ath10k *ar,
1037 &dl_is_polled); 1098 &dl_is_polled);
1038} 1099}
1039 1100
1040static int ath10k_pci_post_rx_pipe(struct hif_ce_pipe_info *pipe_info, 1101static int ath10k_pci_post_rx_pipe(struct ath10k_pci_pipe *pipe_info,
1041 int num) 1102 int num)
1042{ 1103{
1043 struct ath10k *ar = pipe_info->hif_ce_state; 1104 struct ath10k *ar = pipe_info->hif_ce_state;
1044 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 1105 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
1045 struct ce_state *ce_state = pipe_info->ce_hdl; 1106 struct ath10k_ce_pipe *ce_state = pipe_info->ce_hdl;
1046 struct sk_buff *skb; 1107 struct sk_buff *skb;
1047 dma_addr_t ce_data; 1108 dma_addr_t ce_data;
1048 int i, ret = 0; 1109 int i, ret = 0;
@@ -1097,7 +1158,7 @@ err:
1097static int ath10k_pci_post_rx(struct ath10k *ar) 1158static int ath10k_pci_post_rx(struct ath10k *ar)
1098{ 1159{
1099 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 1160 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
1100 struct hif_ce_pipe_info *pipe_info; 1161 struct ath10k_pci_pipe *pipe_info;
1101 const struct ce_attr *attr; 1162 const struct ce_attr *attr;
1102 int pipe_num, ret = 0; 1163 int pipe_num, ret = 0;
1103 1164
@@ -1147,11 +1208,11 @@ static int ath10k_pci_hif_start(struct ath10k *ar)
1147 return 0; 1208 return 0;
1148} 1209}
1149 1210
1150static void ath10k_pci_rx_pipe_cleanup(struct hif_ce_pipe_info *pipe_info) 1211static void ath10k_pci_rx_pipe_cleanup(struct ath10k_pci_pipe *pipe_info)
1151{ 1212{
1152 struct ath10k *ar; 1213 struct ath10k *ar;
1153 struct ath10k_pci *ar_pci; 1214 struct ath10k_pci *ar_pci;
1154 struct ce_state *ce_hdl; 1215 struct ath10k_ce_pipe *ce_hdl;
1155 u32 buf_sz; 1216 u32 buf_sz;
1156 struct sk_buff *netbuf; 1217 struct sk_buff *netbuf;
1157 u32 ce_data; 1218 u32 ce_data;
@@ -1179,11 +1240,11 @@ static void ath10k_pci_rx_pipe_cleanup(struct hif_ce_pipe_info *pipe_info)
1179 } 1240 }
1180} 1241}
1181 1242
1182static void ath10k_pci_tx_pipe_cleanup(struct hif_ce_pipe_info *pipe_info) 1243static void ath10k_pci_tx_pipe_cleanup(struct ath10k_pci_pipe *pipe_info)
1183{ 1244{
1184 struct ath10k *ar; 1245 struct ath10k *ar;
1185 struct ath10k_pci *ar_pci; 1246 struct ath10k_pci *ar_pci;
1186 struct ce_state *ce_hdl; 1247 struct ath10k_ce_pipe *ce_hdl;
1187 struct sk_buff *netbuf; 1248 struct sk_buff *netbuf;
1188 u32 ce_data; 1249 u32 ce_data;
1189 unsigned int nbytes; 1250 unsigned int nbytes;
@@ -1206,15 +1267,14 @@ static void ath10k_pci_tx_pipe_cleanup(struct hif_ce_pipe_info *pipe_info)
1206 1267
1207 while (ath10k_ce_cancel_send_next(ce_hdl, (void **)&netbuf, 1268 while (ath10k_ce_cancel_send_next(ce_hdl, (void **)&netbuf,
1208 &ce_data, &nbytes, &id) == 0) { 1269 &ce_data, &nbytes, &id) == 0) {
1209 if (netbuf != CE_SENDLIST_ITEM_CTXT) 1270 /*
1210 /* 1271 * Indicate the completion to higer layer to free
1211 * Indicate the completion to higer layer to free 1272 * the buffer
1212 * the buffer 1273 */
1213 */ 1274 ATH10K_SKB_CB(netbuf)->is_aborted = true;
1214 ATH10K_SKB_CB(netbuf)->is_aborted = true; 1275 ar_pci->msg_callbacks_current.tx_completion(ar,
1215 ar_pci->msg_callbacks_current.tx_completion(ar, 1276 netbuf,
1216 netbuf, 1277 id);
1217 id);
1218 } 1278 }
1219} 1279}
1220 1280
@@ -1232,7 +1292,7 @@ static void ath10k_pci_buffer_cleanup(struct ath10k *ar)
1232 int pipe_num; 1292 int pipe_num;
1233 1293
1234 for (pipe_num = 0; pipe_num < ar_pci->ce_count; pipe_num++) { 1294 for (pipe_num = 0; pipe_num < ar_pci->ce_count; pipe_num++) {
1235 struct hif_ce_pipe_info *pipe_info; 1295 struct ath10k_pci_pipe *pipe_info;
1236 1296
1237 pipe_info = &ar_pci->pipe_info[pipe_num]; 1297 pipe_info = &ar_pci->pipe_info[pipe_num];
1238 ath10k_pci_rx_pipe_cleanup(pipe_info); 1298 ath10k_pci_rx_pipe_cleanup(pipe_info);
@@ -1243,7 +1303,7 @@ static void ath10k_pci_buffer_cleanup(struct ath10k *ar)
1243static void ath10k_pci_ce_deinit(struct ath10k *ar) 1303static void ath10k_pci_ce_deinit(struct ath10k *ar)
1244{ 1304{
1245 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 1305 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
1246 struct hif_ce_pipe_info *pipe_info; 1306 struct ath10k_pci_pipe *pipe_info;
1247 int pipe_num; 1307 int pipe_num;
1248 1308
1249 for (pipe_num = 0; pipe_num < ar_pci->ce_count; pipe_num++) { 1309 for (pipe_num = 0; pipe_num < ar_pci->ce_count; pipe_num++) {
@@ -1293,8 +1353,10 @@ static int ath10k_pci_hif_exchange_bmi_msg(struct ath10k *ar,
1293 void *resp, u32 *resp_len) 1353 void *resp, u32 *resp_len)
1294{ 1354{
1295 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 1355 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
1296 struct ce_state *ce_tx = ar_pci->pipe_info[BMI_CE_NUM_TO_TARG].ce_hdl; 1356 struct ath10k_pci_pipe *pci_tx = &ar_pci->pipe_info[BMI_CE_NUM_TO_TARG];
1297 struct ce_state *ce_rx = ar_pci->pipe_info[BMI_CE_NUM_TO_HOST].ce_hdl; 1357 struct ath10k_pci_pipe *pci_rx = &ar_pci->pipe_info[BMI_CE_NUM_TO_HOST];
1358 struct ath10k_ce_pipe *ce_tx = pci_tx->ce_hdl;
1359 struct ath10k_ce_pipe *ce_rx = pci_rx->ce_hdl;
1298 dma_addr_t req_paddr = 0; 1360 dma_addr_t req_paddr = 0;
1299 dma_addr_t resp_paddr = 0; 1361 dma_addr_t resp_paddr = 0;
1300 struct bmi_xfer xfer = {}; 1362 struct bmi_xfer xfer = {};
@@ -1378,13 +1440,16 @@ err_dma:
1378 return ret; 1440 return ret;
1379} 1441}
1380 1442
1381static void ath10k_pci_bmi_send_done(struct ce_state *ce_state, 1443static void ath10k_pci_bmi_send_done(struct ath10k_ce_pipe *ce_state)
1382 void *transfer_context,
1383 u32 data,
1384 unsigned int nbytes,
1385 unsigned int transfer_id)
1386{ 1444{
1387 struct bmi_xfer *xfer = transfer_context; 1445 struct bmi_xfer *xfer;
1446 u32 ce_data;
1447 unsigned int nbytes;
1448 unsigned int transfer_id;
1449
1450 if (ath10k_ce_completed_send_next(ce_state, (void **)&xfer, &ce_data,
1451 &nbytes, &transfer_id))
1452 return;
1388 1453
1389 if (xfer->wait_for_resp) 1454 if (xfer->wait_for_resp)
1390 return; 1455 return;
@@ -1392,14 +1457,17 @@ static void ath10k_pci_bmi_send_done(struct ce_state *ce_state,
1392 complete(&xfer->done); 1457 complete(&xfer->done);
1393} 1458}
1394 1459
1395static void ath10k_pci_bmi_recv_data(struct ce_state *ce_state, 1460static void ath10k_pci_bmi_recv_data(struct ath10k_ce_pipe *ce_state)
1396 void *transfer_context,
1397 u32 data,
1398 unsigned int nbytes,
1399 unsigned int transfer_id,
1400 unsigned int flags)
1401{ 1461{
1402 struct bmi_xfer *xfer = transfer_context; 1462 struct bmi_xfer *xfer;
1463 u32 ce_data;
1464 unsigned int nbytes;
1465 unsigned int transfer_id;
1466 unsigned int flags;
1467
1468 if (ath10k_ce_completed_recv_next(ce_state, (void **)&xfer, &ce_data,
1469 &nbytes, &transfer_id, &flags))
1470 return;
1403 1471
1404 if (!xfer->wait_for_resp) { 1472 if (!xfer->wait_for_resp) {
1405 ath10k_warn("unexpected: BMI data received; ignoring\n"); 1473 ath10k_warn("unexpected: BMI data received; ignoring\n");
@@ -1679,7 +1747,7 @@ static int ath10k_pci_init_config(struct ath10k *ar)
1679static int ath10k_pci_ce_init(struct ath10k *ar) 1747static int ath10k_pci_ce_init(struct ath10k *ar)
1680{ 1748{
1681 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 1749 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
1682 struct hif_ce_pipe_info *pipe_info; 1750 struct ath10k_pci_pipe *pipe_info;
1683 const struct ce_attr *attr; 1751 const struct ce_attr *attr;
1684 int pipe_num; 1752 int pipe_num;
1685 1753
@@ -1895,7 +1963,7 @@ static const struct ath10k_hif_ops ath10k_pci_hif_ops = {
1895 1963
1896static void ath10k_pci_ce_tasklet(unsigned long ptr) 1964static void ath10k_pci_ce_tasklet(unsigned long ptr)
1897{ 1965{
1898 struct hif_ce_pipe_info *pipe = (struct hif_ce_pipe_info *)ptr; 1966 struct ath10k_pci_pipe *pipe = (struct ath10k_pci_pipe *)ptr;
1899 struct ath10k_pci *ar_pci = pipe->ar_pci; 1967 struct ath10k_pci *ar_pci = pipe->ar_pci;
1900 1968
1901 ath10k_ce_per_engine_service(ar_pci->ar, pipe->pipe_num); 1969 ath10k_ce_per_engine_service(ar_pci->ar, pipe->pipe_num);
@@ -2212,18 +2280,13 @@ static int ath10k_pci_reset_target(struct ath10k *ar)
2212 2280
2213static void ath10k_pci_device_reset(struct ath10k *ar) 2281static void ath10k_pci_device_reset(struct ath10k *ar)
2214{ 2282{
2215 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
2216 void __iomem *mem = ar_pci->mem;
2217 int i; 2283 int i;
2218 u32 val; 2284 u32 val;
2219 2285
2220 if (!SOC_GLOBAL_RESET_ADDRESS) 2286 if (!SOC_GLOBAL_RESET_ADDRESS)
2221 return; 2287 return;
2222 2288
2223 if (!mem) 2289 ath10k_pci_reg_write32(ar, PCIE_SOC_WAKE_ADDRESS,
2224 return;
2225
2226 ath10k_pci_reg_write32(mem, PCIE_SOC_WAKE_ADDRESS,
2227 PCIE_SOC_WAKE_V_MASK); 2290 PCIE_SOC_WAKE_V_MASK);
2228 for (i = 0; i < ATH_PCI_RESET_WAIT_MAX; i++) { 2291 for (i = 0; i < ATH_PCI_RESET_WAIT_MAX; i++) {
2229 if (ath10k_pci_target_is_awake(ar)) 2292 if (ath10k_pci_target_is_awake(ar))
@@ -2232,12 +2295,12 @@ static void ath10k_pci_device_reset(struct ath10k *ar)
2232 } 2295 }
2233 2296
2234 /* Put Target, including PCIe, into RESET. */ 2297 /* Put Target, including PCIe, into RESET. */
2235 val = ath10k_pci_reg_read32(mem, SOC_GLOBAL_RESET_ADDRESS); 2298 val = ath10k_pci_reg_read32(ar, SOC_GLOBAL_RESET_ADDRESS);
2236 val |= 1; 2299 val |= 1;
2237 ath10k_pci_reg_write32(mem, SOC_GLOBAL_RESET_ADDRESS, val); 2300 ath10k_pci_reg_write32(ar, SOC_GLOBAL_RESET_ADDRESS, val);
2238 2301
2239 for (i = 0; i < ATH_PCI_RESET_WAIT_MAX; i++) { 2302 for (i = 0; i < ATH_PCI_RESET_WAIT_MAX; i++) {
2240 if (ath10k_pci_reg_read32(mem, RTC_STATE_ADDRESS) & 2303 if (ath10k_pci_reg_read32(ar, RTC_STATE_ADDRESS) &
2241 RTC_STATE_COLD_RESET_MASK) 2304 RTC_STATE_COLD_RESET_MASK)
2242 break; 2305 break;
2243 msleep(1); 2306 msleep(1);
@@ -2245,16 +2308,16 @@ static void ath10k_pci_device_reset(struct ath10k *ar)
2245 2308
2246 /* Pull Target, including PCIe, out of RESET. */ 2309 /* Pull Target, including PCIe, out of RESET. */
2247 val &= ~1; 2310 val &= ~1;
2248 ath10k_pci_reg_write32(mem, SOC_GLOBAL_RESET_ADDRESS, val); 2311 ath10k_pci_reg_write32(ar, SOC_GLOBAL_RESET_ADDRESS, val);
2249 2312
2250 for (i = 0; i < ATH_PCI_RESET_WAIT_MAX; i++) { 2313 for (i = 0; i < ATH_PCI_RESET_WAIT_MAX; i++) {
2251 if (!(ath10k_pci_reg_read32(mem, RTC_STATE_ADDRESS) & 2314 if (!(ath10k_pci_reg_read32(ar, RTC_STATE_ADDRESS) &
2252 RTC_STATE_COLD_RESET_MASK)) 2315 RTC_STATE_COLD_RESET_MASK))
2253 break; 2316 break;
2254 msleep(1); 2317 msleep(1);
2255 } 2318 }
2256 2319
2257 ath10k_pci_reg_write32(mem, PCIE_SOC_WAKE_ADDRESS, PCIE_SOC_WAKE_RESET); 2320 ath10k_pci_reg_write32(ar, PCIE_SOC_WAKE_ADDRESS, PCIE_SOC_WAKE_RESET);
2258} 2321}
2259 2322
2260static void ath10k_pci_dump_features(struct ath10k_pci *ar_pci) 2323static void ath10k_pci_dump_features(struct ath10k_pci *ar_pci)
@@ -2267,13 +2330,10 @@ static void ath10k_pci_dump_features(struct ath10k_pci *ar_pci)
2267 2330
2268 switch (i) { 2331 switch (i) {
2269 case ATH10K_PCI_FEATURE_MSI_X: 2332 case ATH10K_PCI_FEATURE_MSI_X:
2270 ath10k_dbg(ATH10K_DBG_PCI, "device supports MSI-X\n"); 2333 ath10k_dbg(ATH10K_DBG_BOOT, "device supports MSI-X\n");
2271 break;
2272 case ATH10K_PCI_FEATURE_HW_1_0_WORKAROUND:
2273 ath10k_dbg(ATH10K_DBG_PCI, "QCA988X_1.0 workaround enabled\n");
2274 break; 2334 break;
2275 case ATH10K_PCI_FEATURE_SOC_POWER_SAVE: 2335 case ATH10K_PCI_FEATURE_SOC_POWER_SAVE:
2276 ath10k_dbg(ATH10K_DBG_PCI, "QCA98XX SoC power save enabled\n"); 2336 ath10k_dbg(ATH10K_DBG_BOOT, "QCA98XX SoC power save enabled\n");
2277 break; 2337 break;
2278 } 2338 }
2279 } 2339 }
@@ -2286,7 +2346,7 @@ static int ath10k_pci_probe(struct pci_dev *pdev,
2286 int ret = 0; 2346 int ret = 0;
2287 struct ath10k *ar; 2347 struct ath10k *ar;
2288 struct ath10k_pci *ar_pci; 2348 struct ath10k_pci *ar_pci;
2289 u32 lcr_val; 2349 u32 lcr_val, chip_id;
2290 2350
2291 ath10k_dbg(ATH10K_DBG_PCI, "%s\n", __func__); 2351 ath10k_dbg(ATH10K_DBG_PCI, "%s\n", __func__);
2292 2352
@@ -2298,9 +2358,6 @@ static int ath10k_pci_probe(struct pci_dev *pdev,
2298 ar_pci->dev = &pdev->dev; 2358 ar_pci->dev = &pdev->dev;
2299 2359
2300 switch (pci_dev->device) { 2360 switch (pci_dev->device) {
2301 case QCA988X_1_0_DEVICE_ID:
2302 set_bit(ATH10K_PCI_FEATURE_HW_1_0_WORKAROUND, ar_pci->features);
2303 break;
2304 case QCA988X_2_0_DEVICE_ID: 2361 case QCA988X_2_0_DEVICE_ID:
2305 set_bit(ATH10K_PCI_FEATURE_MSI_X, ar_pci->features); 2362 set_bit(ATH10K_PCI_FEATURE_MSI_X, ar_pci->features);
2306 break; 2363 break;
@@ -2322,10 +2379,6 @@ static int ath10k_pci_probe(struct pci_dev *pdev,
2322 goto err_ar_pci; 2379 goto err_ar_pci;
2323 } 2380 }
2324 2381
2325 /* Enable QCA988X_1.0 HW workarounds */
2326 if (test_bit(ATH10K_PCI_FEATURE_HW_1_0_WORKAROUND, ar_pci->features))
2327 spin_lock_init(&ar_pci->hw_v1_workaround_lock);
2328
2329 ar_pci->ar = ar; 2382 ar_pci->ar = ar;
2330 ar_pci->fw_indicator_address = FW_INDICATOR_ADDRESS; 2383 ar_pci->fw_indicator_address = FW_INDICATOR_ADDRESS;
2331 atomic_set(&ar_pci->keep_awake_count, 0); 2384 atomic_set(&ar_pci->keep_awake_count, 0);
@@ -2395,9 +2448,20 @@ static int ath10k_pci_probe(struct pci_dev *pdev,
2395 2448
2396 spin_lock_init(&ar_pci->ce_lock); 2449 spin_lock_init(&ar_pci->ce_lock);
2397 2450
2398 ar_pci->cacheline_sz = dma_get_cache_alignment(); 2451 ret = ath10k_do_pci_wake(ar);
2452 if (ret) {
2453 ath10k_err("Failed to get chip id: %d\n", ret);
2454 return ret;
2455 }
2456
2457 chip_id = ath10k_pci_read32(ar,
2458 RTC_SOC_BASE_ADDRESS + SOC_CHIP_ID_ADDRESS);
2459
2460 ath10k_do_pci_sleep(ar);
2461
2462 ath10k_dbg(ATH10K_DBG_BOOT, "boot pci_mem 0x%p\n", ar_pci->mem);
2399 2463
2400 ret = ath10k_core_register(ar); 2464 ret = ath10k_core_register(ar, chip_id);
2401 if (ret) { 2465 if (ret) {
2402 ath10k_err("could not register driver core (%d)\n", ret); 2466 ath10k_err("could not register driver core (%d)\n", ret);
2403 goto err_iomap; 2467 goto err_iomap;
@@ -2414,7 +2478,6 @@ err_region:
2414err_device: 2478err_device:
2415 pci_disable_device(pdev); 2479 pci_disable_device(pdev);
2416err_ar: 2480err_ar:
2417 pci_set_drvdata(pdev, NULL);
2418 ath10k_core_destroy(ar); 2481 ath10k_core_destroy(ar);
2419err_ar_pci: 2482err_ar_pci:
2420 /* call HIF PCI free here */ 2483 /* call HIF PCI free here */
@@ -2442,7 +2505,6 @@ static void ath10k_pci_remove(struct pci_dev *pdev)
2442 2505
2443 ath10k_core_unregister(ar); 2506 ath10k_core_unregister(ar);
2444 2507
2445 pci_set_drvdata(pdev, NULL);
2446 pci_iounmap(pdev, ar_pci->mem); 2508 pci_iounmap(pdev, ar_pci->mem);
2447 pci_release_region(pdev, BAR_NUM); 2509 pci_release_region(pdev, BAR_NUM);
2448 pci_clear_master(pdev); 2510 pci_clear_master(pdev);
@@ -2483,9 +2545,6 @@ module_exit(ath10k_pci_exit);
2483MODULE_AUTHOR("Qualcomm Atheros"); 2545MODULE_AUTHOR("Qualcomm Atheros");
2484MODULE_DESCRIPTION("Driver support for Atheros QCA988X PCIe devices"); 2546MODULE_DESCRIPTION("Driver support for Atheros QCA988X PCIe devices");
2485MODULE_LICENSE("Dual BSD/GPL"); 2547MODULE_LICENSE("Dual BSD/GPL");
2486MODULE_FIRMWARE(QCA988X_HW_1_0_FW_DIR "/" QCA988X_HW_1_0_FW_FILE);
2487MODULE_FIRMWARE(QCA988X_HW_1_0_FW_DIR "/" QCA988X_HW_1_0_OTP_FILE);
2488MODULE_FIRMWARE(QCA988X_HW_1_0_FW_DIR "/" QCA988X_HW_1_0_BOARD_DATA_FILE);
2489MODULE_FIRMWARE(QCA988X_HW_2_0_FW_DIR "/" QCA988X_HW_2_0_FW_FILE); 2548MODULE_FIRMWARE(QCA988X_HW_2_0_FW_DIR "/" QCA988X_HW_2_0_FW_FILE);
2490MODULE_FIRMWARE(QCA988X_HW_2_0_FW_DIR "/" QCA988X_HW_2_0_OTP_FILE); 2549MODULE_FIRMWARE(QCA988X_HW_2_0_FW_DIR "/" QCA988X_HW_2_0_OTP_FILE);
2491MODULE_FIRMWARE(QCA988X_HW_2_0_FW_DIR "/" QCA988X_HW_2_0_BOARD_DATA_FILE); 2550MODULE_FIRMWARE(QCA988X_HW_2_0_FW_DIR "/" QCA988X_HW_2_0_BOARD_DATA_FILE);
diff --git a/drivers/net/wireless/ath/ath10k/pci.h b/drivers/net/wireless/ath/ath10k/pci.h
index 871bb339d56d..52fb7b973571 100644
--- a/drivers/net/wireless/ath/ath10k/pci.h
+++ b/drivers/net/wireless/ath/ath10k/pci.h
@@ -43,22 +43,23 @@ struct bmi_xfer {
43 u32 resp_len; 43 u32 resp_len;
44}; 44};
45 45
46enum ath10k_pci_compl_state {
47 ATH10K_PCI_COMPL_FREE = 0,
48 ATH10K_PCI_COMPL_SEND,
49 ATH10K_PCI_COMPL_RECV,
50};
51
46struct ath10k_pci_compl { 52struct ath10k_pci_compl {
47 struct list_head list; 53 struct list_head list;
48 int send_or_recv; 54 enum ath10k_pci_compl_state state;
49 struct ce_state *ce_state; 55 struct ath10k_ce_pipe *ce_state;
50 struct hif_ce_pipe_info *pipe_info; 56 struct ath10k_pci_pipe *pipe_info;
51 void *transfer_context; 57 struct sk_buff *skb;
52 unsigned int nbytes; 58 unsigned int nbytes;
53 unsigned int transfer_id; 59 unsigned int transfer_id;
54 unsigned int flags; 60 unsigned int flags;
55}; 61};
56 62
57/* compl_state.send_or_recv */
58#define HIF_CE_COMPLETE_FREE 0
59#define HIF_CE_COMPLETE_SEND 1
60#define HIF_CE_COMPLETE_RECV 2
61
62/* 63/*
63 * PCI-specific Target state 64 * PCI-specific Target state
64 * 65 *
@@ -152,17 +153,16 @@ struct service_to_pipe {
152 153
153enum ath10k_pci_features { 154enum ath10k_pci_features {
154 ATH10K_PCI_FEATURE_MSI_X = 0, 155 ATH10K_PCI_FEATURE_MSI_X = 0,
155 ATH10K_PCI_FEATURE_HW_1_0_WORKAROUND = 1, 156 ATH10K_PCI_FEATURE_SOC_POWER_SAVE = 1,
156 ATH10K_PCI_FEATURE_SOC_POWER_SAVE = 2,
157 157
158 /* keep last */ 158 /* keep last */
159 ATH10K_PCI_FEATURE_COUNT 159 ATH10K_PCI_FEATURE_COUNT
160}; 160};
161 161
162/* Per-pipe state. */ 162/* Per-pipe state. */
163struct hif_ce_pipe_info { 163struct ath10k_pci_pipe {
164 /* Handle of underlying Copy Engine */ 164 /* Handle of underlying Copy Engine */
165 struct ce_state *ce_hdl; 165 struct ath10k_ce_pipe *ce_hdl;
166 166
167 /* Our pipe number; facilitiates use of pipe_info ptrs. */ 167 /* Our pipe number; facilitiates use of pipe_info ptrs. */
168 u8 pipe_num; 168 u8 pipe_num;
@@ -178,9 +178,6 @@ struct hif_ce_pipe_info {
178 /* List of free CE completion slots */ 178 /* List of free CE completion slots */
179 struct list_head compl_free; 179 struct list_head compl_free;
180 180
181 /* Limit the number of outstanding send requests. */
182 int num_sends_allowed;
183
184 struct ath10k_pci *ar_pci; 181 struct ath10k_pci *ar_pci;
185 struct tasklet_struct intr; 182 struct tasklet_struct intr;
186}; 183};
@@ -190,7 +187,6 @@ struct ath10k_pci {
190 struct device *dev; 187 struct device *dev;
191 struct ath10k *ar; 188 struct ath10k *ar;
192 void __iomem *mem; 189 void __iomem *mem;
193 int cacheline_sz;
194 190
195 DECLARE_BITMAP(features, ATH10K_PCI_FEATURE_COUNT); 191 DECLARE_BITMAP(features, ATH10K_PCI_FEATURE_COUNT);
196 192
@@ -219,7 +215,7 @@ struct ath10k_pci {
219 215
220 bool compl_processing; 216 bool compl_processing;
221 217
222 struct hif_ce_pipe_info pipe_info[CE_COUNT_MAX]; 218 struct ath10k_pci_pipe pipe_info[CE_COUNT_MAX];
223 219
224 struct ath10k_hif_cb msg_callbacks_current; 220 struct ath10k_hif_cb msg_callbacks_current;
225 221
@@ -227,16 +223,13 @@ struct ath10k_pci {
227 u32 fw_indicator_address; 223 u32 fw_indicator_address;
228 224
229 /* Copy Engine used for Diagnostic Accesses */ 225 /* Copy Engine used for Diagnostic Accesses */
230 struct ce_state *ce_diag; 226 struct ath10k_ce_pipe *ce_diag;
231 227
232 /* FIXME: document what this really protects */ 228 /* FIXME: document what this really protects */
233 spinlock_t ce_lock; 229 spinlock_t ce_lock;
234 230
235 /* Map CE id to ce_state */ 231 /* Map CE id to ce_state */
236 struct ce_state *ce_id_to_state[CE_COUNT_MAX]; 232 struct ath10k_ce_pipe ce_states[CE_COUNT_MAX];
237
238 /* makes sure that dummy reads are atomic */
239 spinlock_t hw_v1_workaround_lock;
240}; 233};
241 234
242static inline struct ath10k_pci *ath10k_pci_priv(struct ath10k *ar) 235static inline struct ath10k_pci *ath10k_pci_priv(struct ath10k *ar)
@@ -244,14 +237,18 @@ static inline struct ath10k_pci *ath10k_pci_priv(struct ath10k *ar)
244 return ar->hif.priv; 237 return ar->hif.priv;
245} 238}
246 239
247static inline u32 ath10k_pci_reg_read32(void __iomem *mem, u32 addr) 240static inline u32 ath10k_pci_reg_read32(struct ath10k *ar, u32 addr)
248{ 241{
249 return ioread32(mem + PCIE_LOCAL_BASE_ADDRESS + addr); 242 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
243
244 return ioread32(ar_pci->mem + PCIE_LOCAL_BASE_ADDRESS + addr);
250} 245}
251 246
252static inline void ath10k_pci_reg_write32(void __iomem *mem, u32 addr, u32 val) 247static inline void ath10k_pci_reg_write32(struct ath10k *ar, u32 addr, u32 val)
253{ 248{
254 iowrite32(val, mem + PCIE_LOCAL_BASE_ADDRESS + addr); 249 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
250
251 iowrite32(val, ar_pci->mem + PCIE_LOCAL_BASE_ADDRESS + addr);
255} 252}
256 253
257#define ATH_PCI_RESET_WAIT_MAX 10 /* ms */ 254#define ATH_PCI_RESET_WAIT_MAX 10 /* ms */
@@ -310,23 +307,8 @@ static inline void ath10k_pci_write32(struct ath10k *ar, u32 offset,
310 u32 value) 307 u32 value)
311{ 308{
312 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 309 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
313 void __iomem *addr = ar_pci->mem;
314
315 if (test_bit(ATH10K_PCI_FEATURE_HW_1_0_WORKAROUND, ar_pci->features)) {
316 unsigned long irq_flags;
317 310
318 spin_lock_irqsave(&ar_pci->hw_v1_workaround_lock, irq_flags); 311 iowrite32(value, ar_pci->mem + offset);
319
320 ioread32(addr+offset+4); /* 3rd read prior to write */
321 ioread32(addr+offset+4); /* 2nd read prior to write */
322 ioread32(addr+offset+4); /* 1st read prior to write */
323 iowrite32(value, addr+offset);
324
325 spin_unlock_irqrestore(&ar_pci->hw_v1_workaround_lock,
326 irq_flags);
327 } else {
328 iowrite32(value, addr+offset);
329 }
330} 312}
331 313
332static inline u32 ath10k_pci_read32(struct ath10k *ar, u32 offset) 314static inline u32 ath10k_pci_read32(struct ath10k *ar, u32 offset)
@@ -336,15 +318,17 @@ static inline u32 ath10k_pci_read32(struct ath10k *ar, u32 offset)
336 return ioread32(ar_pci->mem + offset); 318 return ioread32(ar_pci->mem + offset);
337} 319}
338 320
339void ath10k_do_pci_wake(struct ath10k *ar); 321int ath10k_do_pci_wake(struct ath10k *ar);
340void ath10k_do_pci_sleep(struct ath10k *ar); 322void ath10k_do_pci_sleep(struct ath10k *ar);
341 323
342static inline void ath10k_pci_wake(struct ath10k *ar) 324static inline int ath10k_pci_wake(struct ath10k *ar)
343{ 325{
344 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar); 326 struct ath10k_pci *ar_pci = ath10k_pci_priv(ar);
345 327
346 if (test_bit(ATH10K_PCI_FEATURE_SOC_POWER_SAVE, ar_pci->features)) 328 if (test_bit(ATH10K_PCI_FEATURE_SOC_POWER_SAVE, ar_pci->features))
347 ath10k_do_pci_wake(ar); 329 return ath10k_do_pci_wake(ar);
330
331 return 0;
348} 332}
349 333
350static inline void ath10k_pci_sleep(struct ath10k *ar) 334static inline void ath10k_pci_sleep(struct ath10k *ar)
diff --git a/drivers/net/wireless/ath/ath10k/rx_desc.h b/drivers/net/wireless/ath/ath10k/rx_desc.h
index bfec6c8f2ecb..1c584c4b019c 100644
--- a/drivers/net/wireless/ath/ath10k/rx_desc.h
+++ b/drivers/net/wireless/ath/ath10k/rx_desc.h
@@ -422,10 +422,30 @@ struct rx_mpdu_end {
422#define RX_MSDU_START_INFO1_IP_FRAG (1 << 14) 422#define RX_MSDU_START_INFO1_IP_FRAG (1 << 14)
423#define RX_MSDU_START_INFO1_TCP_ONLY_ACK (1 << 15) 423#define RX_MSDU_START_INFO1_TCP_ONLY_ACK (1 << 15)
424 424
425/* The decapped header (rx_hdr_status) contains the following:
426 * a) 802.11 header
427 * [padding to 4 bytes]
428 * b) HW crypto parameter
429 * - 0 bytes for no security
430 * - 4 bytes for WEP
431 * - 8 bytes for TKIP, AES
432 * [padding to 4 bytes]
433 * c) A-MSDU subframe header (14 bytes) if appliable
434 * d) LLC/SNAP (RFC1042, 8 bytes)
435 *
436 * In case of A-MSDU only first frame in sequence contains (a) and (b). */
425enum rx_msdu_decap_format { 437enum rx_msdu_decap_format {
426 RX_MSDU_DECAP_RAW = 0, 438 RX_MSDU_DECAP_RAW = 0,
427 RX_MSDU_DECAP_NATIVE_WIFI = 1, 439
440 /* Note: QoS frames are reported as non-QoS. The rx_hdr_status in
441 * htt_rx_desc contains the original decapped 802.11 header. */
442 RX_MSDU_DECAP_NATIVE_WIFI = 1,
443
444 /* Payload contains an ethernet header (struct ethhdr). */
428 RX_MSDU_DECAP_ETHERNET2_DIX = 2, 445 RX_MSDU_DECAP_ETHERNET2_DIX = 2,
446
447 /* Payload contains two 48-bit addresses and 2-byte length (14 bytes
448 * total), followed by an RFC1042 header (8 bytes). */
429 RX_MSDU_DECAP_8023_SNAP_LLC = 3 449 RX_MSDU_DECAP_8023_SNAP_LLC = 3
430}; 450};
431 451
diff --git a/drivers/net/wireless/ath/ath10k/trace.h b/drivers/net/wireless/ath/ath10k/trace.h
index 85e806bf7257..90817ddc92ba 100644
--- a/drivers/net/wireless/ath/ath10k/trace.h
+++ b/drivers/net/wireless/ath/ath10k/trace.h
@@ -111,26 +111,29 @@ TRACE_EVENT(ath10k_log_dbg_dump,
111); 111);
112 112
113TRACE_EVENT(ath10k_wmi_cmd, 113TRACE_EVENT(ath10k_wmi_cmd,
114 TP_PROTO(int id, void *buf, size_t buf_len), 114 TP_PROTO(int id, void *buf, size_t buf_len, int ret),
115 115
116 TP_ARGS(id, buf, buf_len), 116 TP_ARGS(id, buf, buf_len, ret),
117 117
118 TP_STRUCT__entry( 118 TP_STRUCT__entry(
119 __field(unsigned int, id) 119 __field(unsigned int, id)
120 __field(size_t, buf_len) 120 __field(size_t, buf_len)
121 __dynamic_array(u8, buf, buf_len) 121 __dynamic_array(u8, buf, buf_len)
122 __field(int, ret)
122 ), 123 ),
123 124
124 TP_fast_assign( 125 TP_fast_assign(
125 __entry->id = id; 126 __entry->id = id;
126 __entry->buf_len = buf_len; 127 __entry->buf_len = buf_len;
128 __entry->ret = ret;
127 memcpy(__get_dynamic_array(buf), buf, buf_len); 129 memcpy(__get_dynamic_array(buf), buf, buf_len);
128 ), 130 ),
129 131
130 TP_printk( 132 TP_printk(
131 "id %d len %zu", 133 "id %d len %zu ret %d",
132 __entry->id, 134 __entry->id,
133 __entry->buf_len 135 __entry->buf_len,
136 __entry->ret
134 ) 137 )
135); 138);
136 139
@@ -158,6 +161,27 @@ TRACE_EVENT(ath10k_wmi_event,
158 ) 161 )
159); 162);
160 163
164TRACE_EVENT(ath10k_htt_stats,
165 TP_PROTO(void *buf, size_t buf_len),
166
167 TP_ARGS(buf, buf_len),
168
169 TP_STRUCT__entry(
170 __field(size_t, buf_len)
171 __dynamic_array(u8, buf, buf_len)
172 ),
173
174 TP_fast_assign(
175 __entry->buf_len = buf_len;
176 memcpy(__get_dynamic_array(buf), buf, buf_len);
177 ),
178
179 TP_printk(
180 "len %zu",
181 __entry->buf_len
182 )
183);
184
161#endif /* _TRACE_H_ || TRACE_HEADER_MULTI_READ*/ 185#endif /* _TRACE_H_ || TRACE_HEADER_MULTI_READ*/
162 186
163/* we don't want to use include/trace/events */ 187/* we don't want to use include/trace/events */
diff --git a/drivers/net/wireless/ath/ath10k/txrx.c b/drivers/net/wireless/ath/ath10k/txrx.c
index 68b6faefd1d8..5ae373a1e294 100644
--- a/drivers/net/wireless/ath/ath10k/txrx.c
+++ b/drivers/net/wireless/ath/ath10k/txrx.c
@@ -44,40 +44,39 @@ out:
44 spin_unlock_bh(&ar->data_lock); 44 spin_unlock_bh(&ar->data_lock);
45} 45}
46 46
47void ath10k_txrx_tx_unref(struct ath10k_htt *htt, struct sk_buff *txdesc) 47void ath10k_txrx_tx_unref(struct ath10k_htt *htt,
48 const struct htt_tx_done *tx_done)
48{ 49{
49 struct device *dev = htt->ar->dev; 50 struct device *dev = htt->ar->dev;
50 struct ieee80211_tx_info *info; 51 struct ieee80211_tx_info *info;
51 struct sk_buff *txfrag = ATH10K_SKB_CB(txdesc)->htt.txfrag; 52 struct ath10k_skb_cb *skb_cb;
52 struct sk_buff *msdu = ATH10K_SKB_CB(txdesc)->htt.msdu; 53 struct sk_buff *msdu;
53 int ret; 54 int ret;
54 55
55 if (ATH10K_SKB_CB(txdesc)->htt.refcount == 0) 56 ath10k_dbg(ATH10K_DBG_HTT, "htt tx completion msdu_id %u discard %d no_ack %d\n",
56 return; 57 tx_done->msdu_id, !!tx_done->discard, !!tx_done->no_ack);
57
58 ATH10K_SKB_CB(txdesc)->htt.refcount--;
59 58
60 if (ATH10K_SKB_CB(txdesc)->htt.refcount > 0) 59 if (tx_done->msdu_id >= htt->max_num_pending_tx) {
60 ath10k_warn("warning: msdu_id %d too big, ignoring\n",
61 tx_done->msdu_id);
61 return; 62 return;
62
63 if (txfrag) {
64 ret = ath10k_skb_unmap(dev, txfrag);
65 if (ret)
66 ath10k_warn("txfrag unmap failed (%d)\n", ret);
67
68 dev_kfree_skb_any(txfrag);
69 } 63 }
70 64
65 msdu = htt->pending_tx[tx_done->msdu_id];
66 skb_cb = ATH10K_SKB_CB(msdu);
67
71 ret = ath10k_skb_unmap(dev, msdu); 68 ret = ath10k_skb_unmap(dev, msdu);
72 if (ret) 69 if (ret)
73 ath10k_warn("data skb unmap failed (%d)\n", ret); 70 ath10k_warn("data skb unmap failed (%d)\n", ret);
74 71
72 if (skb_cb->htt.frag_len)
73 skb_pull(msdu, skb_cb->htt.frag_len + skb_cb->htt.pad_len);
74
75 ath10k_report_offchan_tx(htt->ar, msdu); 75 ath10k_report_offchan_tx(htt->ar, msdu);
76 76
77 info = IEEE80211_SKB_CB(msdu); 77 info = IEEE80211_SKB_CB(msdu);
78 memset(&info->status, 0, sizeof(info->status));
79 78
80 if (ATH10K_SKB_CB(txdesc)->htt.discard) { 79 if (tx_done->discard) {
81 ieee80211_free_txskb(htt->ar->hw, msdu); 80 ieee80211_free_txskb(htt->ar->hw, msdu);
82 goto exit; 81 goto exit;
83 } 82 }
@@ -85,7 +84,7 @@ void ath10k_txrx_tx_unref(struct ath10k_htt *htt, struct sk_buff *txdesc)
85 if (!(info->flags & IEEE80211_TX_CTL_NO_ACK)) 84 if (!(info->flags & IEEE80211_TX_CTL_NO_ACK))
86 info->flags |= IEEE80211_TX_STAT_ACK; 85 info->flags |= IEEE80211_TX_STAT_ACK;
87 86
88 if (ATH10K_SKB_CB(txdesc)->htt.no_ack) 87 if (tx_done->no_ack)
89 info->flags &= ~IEEE80211_TX_STAT_ACK; 88 info->flags &= ~IEEE80211_TX_STAT_ACK;
90 89
91 ieee80211_tx_status(htt->ar->hw, msdu); 90 ieee80211_tx_status(htt->ar->hw, msdu);
@@ -93,36 +92,12 @@ void ath10k_txrx_tx_unref(struct ath10k_htt *htt, struct sk_buff *txdesc)
93 92
94exit: 93exit:
95 spin_lock_bh(&htt->tx_lock); 94 spin_lock_bh(&htt->tx_lock);
96 htt->pending_tx[ATH10K_SKB_CB(txdesc)->htt.msdu_id] = NULL; 95 htt->pending_tx[tx_done->msdu_id] = NULL;
97 ath10k_htt_tx_free_msdu_id(htt, ATH10K_SKB_CB(txdesc)->htt.msdu_id); 96 ath10k_htt_tx_free_msdu_id(htt, tx_done->msdu_id);
98 __ath10k_htt_tx_dec_pending(htt); 97 __ath10k_htt_tx_dec_pending(htt);
99 if (bitmap_empty(htt->used_msdu_ids, htt->max_num_pending_tx)) 98 if (htt->num_pending_tx == 0)
100 wake_up(&htt->empty_tx_wq); 99 wake_up(&htt->empty_tx_wq);
101 spin_unlock_bh(&htt->tx_lock); 100 spin_unlock_bh(&htt->tx_lock);
102
103 dev_kfree_skb_any(txdesc);
104}
105
106void ath10k_txrx_tx_completed(struct ath10k_htt *htt,
107 const struct htt_tx_done *tx_done)
108{
109 struct sk_buff *txdesc;
110
111 ath10k_dbg(ATH10K_DBG_HTT, "htt tx completion msdu_id %u discard %d no_ack %d\n",
112 tx_done->msdu_id, !!tx_done->discard, !!tx_done->no_ack);
113
114 if (tx_done->msdu_id >= htt->max_num_pending_tx) {
115 ath10k_warn("warning: msdu_id %d too big, ignoring\n",
116 tx_done->msdu_id);
117 return;
118 }
119
120 txdesc = htt->pending_tx[tx_done->msdu_id];
121
122 ATH10K_SKB_CB(txdesc)->htt.discard = tx_done->discard;
123 ATH10K_SKB_CB(txdesc)->htt.no_ack = tx_done->no_ack;
124
125 ath10k_txrx_tx_unref(htt, txdesc);
126} 101}
127 102
128static const u8 rx_legacy_rate_idx[] = { 103static const u8 rx_legacy_rate_idx[] = {
@@ -293,6 +268,8 @@ void ath10k_process_rx(struct ath10k *ar, struct htt_rx_info *info)
293 status->vht_nss, 268 status->vht_nss,
294 status->freq, 269 status->freq,
295 status->band); 270 status->band);
271 ath10k_dbg_dump(ATH10K_DBG_HTT_DUMP, NULL, "rx skb: ",
272 info->skb->data, info->skb->len);
296 273
297 ieee80211_rx(ar->hw, info->skb); 274 ieee80211_rx(ar->hw, info->skb);
298} 275}
diff --git a/drivers/net/wireless/ath/ath10k/txrx.h b/drivers/net/wireless/ath/ath10k/txrx.h
index e78632a76df7..356dc9c04c9e 100644
--- a/drivers/net/wireless/ath/ath10k/txrx.h
+++ b/drivers/net/wireless/ath/ath10k/txrx.h
@@ -19,9 +19,8 @@
19 19
20#include "htt.h" 20#include "htt.h"
21 21
22void ath10k_txrx_tx_unref(struct ath10k_htt *htt, struct sk_buff *txdesc); 22void ath10k_txrx_tx_unref(struct ath10k_htt *htt,
23void ath10k_txrx_tx_completed(struct ath10k_htt *htt, 23 const struct htt_tx_done *tx_done);
24 const struct htt_tx_done *tx_done);
25void ath10k_process_rx(struct ath10k *ar, struct htt_rx_info *info); 24void ath10k_process_rx(struct ath10k *ar, struct htt_rx_info *info);
26 25
27struct ath10k_peer *ath10k_peer_find(struct ath10k *ar, int vdev_id, 26struct ath10k_peer *ath10k_peer_find(struct ath10k *ar, int vdev_id,
diff --git a/drivers/net/wireless/ath/ath10k/wmi.c b/drivers/net/wireless/ath/ath10k/wmi.c
index 55f90c761868..ccf3597fd9e2 100644
--- a/drivers/net/wireless/ath/ath10k/wmi.c
+++ b/drivers/net/wireless/ath/ath10k/wmi.c
@@ -23,29 +23,470 @@
23#include "wmi.h" 23#include "wmi.h"
24#include "mac.h" 24#include "mac.h"
25 25
26void ath10k_wmi_flush_tx(struct ath10k *ar) 26/* MAIN WMI cmd track */
27{ 27static struct wmi_cmd_map wmi_cmd_map = {
28 int ret; 28 .init_cmdid = WMI_INIT_CMDID,
29 29 .start_scan_cmdid = WMI_START_SCAN_CMDID,
30 lockdep_assert_held(&ar->conf_mutex); 30 .stop_scan_cmdid = WMI_STOP_SCAN_CMDID,
31 31 .scan_chan_list_cmdid = WMI_SCAN_CHAN_LIST_CMDID,
32 if (ar->state == ATH10K_STATE_WEDGED) { 32 .scan_sch_prio_tbl_cmdid = WMI_SCAN_SCH_PRIO_TBL_CMDID,
33 ath10k_warn("wmi flush skipped - device is wedged anyway\n"); 33 .pdev_set_regdomain_cmdid = WMI_PDEV_SET_REGDOMAIN_CMDID,
34 return; 34 .pdev_set_channel_cmdid = WMI_PDEV_SET_CHANNEL_CMDID,
35 } 35 .pdev_set_param_cmdid = WMI_PDEV_SET_PARAM_CMDID,
36 36 .pdev_pktlog_enable_cmdid = WMI_PDEV_PKTLOG_ENABLE_CMDID,
37 ret = wait_event_timeout(ar->wmi.wq, 37 .pdev_pktlog_disable_cmdid = WMI_PDEV_PKTLOG_DISABLE_CMDID,
38 atomic_read(&ar->wmi.pending_tx_count) == 0, 38 .pdev_set_wmm_params_cmdid = WMI_PDEV_SET_WMM_PARAMS_CMDID,
39 5*HZ); 39 .pdev_set_ht_cap_ie_cmdid = WMI_PDEV_SET_HT_CAP_IE_CMDID,
40 if (atomic_read(&ar->wmi.pending_tx_count) == 0) 40 .pdev_set_vht_cap_ie_cmdid = WMI_PDEV_SET_VHT_CAP_IE_CMDID,
41 return; 41 .pdev_set_dscp_tid_map_cmdid = WMI_PDEV_SET_DSCP_TID_MAP_CMDID,
42 42 .pdev_set_quiet_mode_cmdid = WMI_PDEV_SET_QUIET_MODE_CMDID,
43 if (ret == 0) 43 .pdev_green_ap_ps_enable_cmdid = WMI_PDEV_GREEN_AP_PS_ENABLE_CMDID,
44 ret = -ETIMEDOUT; 44 .pdev_get_tpc_config_cmdid = WMI_PDEV_GET_TPC_CONFIG_CMDID,
45 45 .pdev_set_base_macaddr_cmdid = WMI_PDEV_SET_BASE_MACADDR_CMDID,
46 if (ret < 0) 46 .vdev_create_cmdid = WMI_VDEV_CREATE_CMDID,
47 ath10k_warn("wmi flush failed (%d)\n", ret); 47 .vdev_delete_cmdid = WMI_VDEV_DELETE_CMDID,
48} 48 .vdev_start_request_cmdid = WMI_VDEV_START_REQUEST_CMDID,
49 .vdev_restart_request_cmdid = WMI_VDEV_RESTART_REQUEST_CMDID,
50 .vdev_up_cmdid = WMI_VDEV_UP_CMDID,
51 .vdev_stop_cmdid = WMI_VDEV_STOP_CMDID,
52 .vdev_down_cmdid = WMI_VDEV_DOWN_CMDID,
53 .vdev_set_param_cmdid = WMI_VDEV_SET_PARAM_CMDID,
54 .vdev_install_key_cmdid = WMI_VDEV_INSTALL_KEY_CMDID,
55 .peer_create_cmdid = WMI_PEER_CREATE_CMDID,
56 .peer_delete_cmdid = WMI_PEER_DELETE_CMDID,
57 .peer_flush_tids_cmdid = WMI_PEER_FLUSH_TIDS_CMDID,
58 .peer_set_param_cmdid = WMI_PEER_SET_PARAM_CMDID,
59 .peer_assoc_cmdid = WMI_PEER_ASSOC_CMDID,
60 .peer_add_wds_entry_cmdid = WMI_PEER_ADD_WDS_ENTRY_CMDID,
61 .peer_remove_wds_entry_cmdid = WMI_PEER_REMOVE_WDS_ENTRY_CMDID,
62 .peer_mcast_group_cmdid = WMI_PEER_MCAST_GROUP_CMDID,
63 .bcn_tx_cmdid = WMI_BCN_TX_CMDID,
64 .pdev_send_bcn_cmdid = WMI_PDEV_SEND_BCN_CMDID,
65 .bcn_tmpl_cmdid = WMI_BCN_TMPL_CMDID,
66 .bcn_filter_rx_cmdid = WMI_BCN_FILTER_RX_CMDID,
67 .prb_req_filter_rx_cmdid = WMI_PRB_REQ_FILTER_RX_CMDID,
68 .mgmt_tx_cmdid = WMI_MGMT_TX_CMDID,
69 .prb_tmpl_cmdid = WMI_PRB_TMPL_CMDID,
70 .addba_clear_resp_cmdid = WMI_ADDBA_CLEAR_RESP_CMDID,
71 .addba_send_cmdid = WMI_ADDBA_SEND_CMDID,
72 .addba_status_cmdid = WMI_ADDBA_STATUS_CMDID,
73 .delba_send_cmdid = WMI_DELBA_SEND_CMDID,
74 .addba_set_resp_cmdid = WMI_ADDBA_SET_RESP_CMDID,
75 .send_singleamsdu_cmdid = WMI_SEND_SINGLEAMSDU_CMDID,
76 .sta_powersave_mode_cmdid = WMI_STA_POWERSAVE_MODE_CMDID,
77 .sta_powersave_param_cmdid = WMI_STA_POWERSAVE_PARAM_CMDID,
78 .sta_mimo_ps_mode_cmdid = WMI_STA_MIMO_PS_MODE_CMDID,
79 .pdev_dfs_enable_cmdid = WMI_PDEV_DFS_ENABLE_CMDID,
80 .pdev_dfs_disable_cmdid = WMI_PDEV_DFS_DISABLE_CMDID,
81 .roam_scan_mode = WMI_ROAM_SCAN_MODE,
82 .roam_scan_rssi_threshold = WMI_ROAM_SCAN_RSSI_THRESHOLD,
83 .roam_scan_period = WMI_ROAM_SCAN_PERIOD,
84 .roam_scan_rssi_change_threshold = WMI_ROAM_SCAN_RSSI_CHANGE_THRESHOLD,
85 .roam_ap_profile = WMI_ROAM_AP_PROFILE,
86 .ofl_scan_add_ap_profile = WMI_ROAM_AP_PROFILE,
87 .ofl_scan_remove_ap_profile = WMI_OFL_SCAN_REMOVE_AP_PROFILE,
88 .ofl_scan_period = WMI_OFL_SCAN_PERIOD,
89 .p2p_dev_set_device_info = WMI_P2P_DEV_SET_DEVICE_INFO,
90 .p2p_dev_set_discoverability = WMI_P2P_DEV_SET_DISCOVERABILITY,
91 .p2p_go_set_beacon_ie = WMI_P2P_GO_SET_BEACON_IE,
92 .p2p_go_set_probe_resp_ie = WMI_P2P_GO_SET_PROBE_RESP_IE,
93 .p2p_set_vendor_ie_data_cmdid = WMI_P2P_SET_VENDOR_IE_DATA_CMDID,
94 .ap_ps_peer_param_cmdid = WMI_AP_PS_PEER_PARAM_CMDID,
95 .ap_ps_peer_uapsd_coex_cmdid = WMI_AP_PS_PEER_UAPSD_COEX_CMDID,
96 .peer_rate_retry_sched_cmdid = WMI_PEER_RATE_RETRY_SCHED_CMDID,
97 .wlan_profile_trigger_cmdid = WMI_WLAN_PROFILE_TRIGGER_CMDID,
98 .wlan_profile_set_hist_intvl_cmdid =
99 WMI_WLAN_PROFILE_SET_HIST_INTVL_CMDID,
100 .wlan_profile_get_profile_data_cmdid =
101 WMI_WLAN_PROFILE_GET_PROFILE_DATA_CMDID,
102 .wlan_profile_enable_profile_id_cmdid =
103 WMI_WLAN_PROFILE_ENABLE_PROFILE_ID_CMDID,
104 .wlan_profile_list_profile_id_cmdid =
105 WMI_WLAN_PROFILE_LIST_PROFILE_ID_CMDID,
106 .pdev_suspend_cmdid = WMI_PDEV_SUSPEND_CMDID,
107 .pdev_resume_cmdid = WMI_PDEV_RESUME_CMDID,
108 .add_bcn_filter_cmdid = WMI_ADD_BCN_FILTER_CMDID,
109 .rmv_bcn_filter_cmdid = WMI_RMV_BCN_FILTER_CMDID,
110 .wow_add_wake_pattern_cmdid = WMI_WOW_ADD_WAKE_PATTERN_CMDID,
111 .wow_del_wake_pattern_cmdid = WMI_WOW_DEL_WAKE_PATTERN_CMDID,
112 .wow_enable_disable_wake_event_cmdid =
113 WMI_WOW_ENABLE_DISABLE_WAKE_EVENT_CMDID,
114 .wow_enable_cmdid = WMI_WOW_ENABLE_CMDID,
115 .wow_hostwakeup_from_sleep_cmdid = WMI_WOW_HOSTWAKEUP_FROM_SLEEP_CMDID,
116 .rtt_measreq_cmdid = WMI_RTT_MEASREQ_CMDID,
117 .rtt_tsf_cmdid = WMI_RTT_TSF_CMDID,
118 .vdev_spectral_scan_configure_cmdid =
119 WMI_VDEV_SPECTRAL_SCAN_CONFIGURE_CMDID,
120 .vdev_spectral_scan_enable_cmdid = WMI_VDEV_SPECTRAL_SCAN_ENABLE_CMDID,
121 .request_stats_cmdid = WMI_REQUEST_STATS_CMDID,
122 .set_arp_ns_offload_cmdid = WMI_SET_ARP_NS_OFFLOAD_CMDID,
123 .network_list_offload_config_cmdid =
124 WMI_NETWORK_LIST_OFFLOAD_CONFIG_CMDID,
125 .gtk_offload_cmdid = WMI_GTK_OFFLOAD_CMDID,
126 .csa_offload_enable_cmdid = WMI_CSA_OFFLOAD_ENABLE_CMDID,
127 .csa_offload_chanswitch_cmdid = WMI_CSA_OFFLOAD_CHANSWITCH_CMDID,
128 .chatter_set_mode_cmdid = WMI_CHATTER_SET_MODE_CMDID,
129 .peer_tid_addba_cmdid = WMI_PEER_TID_ADDBA_CMDID,
130 .peer_tid_delba_cmdid = WMI_PEER_TID_DELBA_CMDID,
131 .sta_dtim_ps_method_cmdid = WMI_STA_DTIM_PS_METHOD_CMDID,
132 .sta_uapsd_auto_trig_cmdid = WMI_STA_UAPSD_AUTO_TRIG_CMDID,
133 .sta_keepalive_cmd = WMI_STA_KEEPALIVE_CMD,
134 .echo_cmdid = WMI_ECHO_CMDID,
135 .pdev_utf_cmdid = WMI_PDEV_UTF_CMDID,
136 .dbglog_cfg_cmdid = WMI_DBGLOG_CFG_CMDID,
137 .pdev_qvit_cmdid = WMI_PDEV_QVIT_CMDID,
138 .pdev_ftm_intg_cmdid = WMI_PDEV_FTM_INTG_CMDID,
139 .vdev_set_keepalive_cmdid = WMI_VDEV_SET_KEEPALIVE_CMDID,
140 .vdev_get_keepalive_cmdid = WMI_VDEV_GET_KEEPALIVE_CMDID,
141 .force_fw_hang_cmdid = WMI_FORCE_FW_HANG_CMDID,
142 .gpio_config_cmdid = WMI_GPIO_CONFIG_CMDID,
143 .gpio_output_cmdid = WMI_GPIO_OUTPUT_CMDID,
144};
145
146/* 10.X WMI cmd track */
147static struct wmi_cmd_map wmi_10x_cmd_map = {
148 .init_cmdid = WMI_10X_INIT_CMDID,
149 .start_scan_cmdid = WMI_10X_START_SCAN_CMDID,
150 .stop_scan_cmdid = WMI_10X_STOP_SCAN_CMDID,
151 .scan_chan_list_cmdid = WMI_10X_SCAN_CHAN_LIST_CMDID,
152 .scan_sch_prio_tbl_cmdid = WMI_CMD_UNSUPPORTED,
153 .pdev_set_regdomain_cmdid = WMI_10X_PDEV_SET_REGDOMAIN_CMDID,
154 .pdev_set_channel_cmdid = WMI_10X_PDEV_SET_CHANNEL_CMDID,
155 .pdev_set_param_cmdid = WMI_10X_PDEV_SET_PARAM_CMDID,
156 .pdev_pktlog_enable_cmdid = WMI_10X_PDEV_PKTLOG_ENABLE_CMDID,
157 .pdev_pktlog_disable_cmdid = WMI_10X_PDEV_PKTLOG_DISABLE_CMDID,
158 .pdev_set_wmm_params_cmdid = WMI_10X_PDEV_SET_WMM_PARAMS_CMDID,
159 .pdev_set_ht_cap_ie_cmdid = WMI_10X_PDEV_SET_HT_CAP_IE_CMDID,
160 .pdev_set_vht_cap_ie_cmdid = WMI_10X_PDEV_SET_VHT_CAP_IE_CMDID,
161 .pdev_set_dscp_tid_map_cmdid = WMI_10X_PDEV_SET_DSCP_TID_MAP_CMDID,
162 .pdev_set_quiet_mode_cmdid = WMI_10X_PDEV_SET_QUIET_MODE_CMDID,
163 .pdev_green_ap_ps_enable_cmdid = WMI_10X_PDEV_GREEN_AP_PS_ENABLE_CMDID,
164 .pdev_get_tpc_config_cmdid = WMI_10X_PDEV_GET_TPC_CONFIG_CMDID,
165 .pdev_set_base_macaddr_cmdid = WMI_10X_PDEV_SET_BASE_MACADDR_CMDID,
166 .vdev_create_cmdid = WMI_10X_VDEV_CREATE_CMDID,
167 .vdev_delete_cmdid = WMI_10X_VDEV_DELETE_CMDID,
168 .vdev_start_request_cmdid = WMI_10X_VDEV_START_REQUEST_CMDID,
169 .vdev_restart_request_cmdid = WMI_10X_VDEV_RESTART_REQUEST_CMDID,
170 .vdev_up_cmdid = WMI_10X_VDEV_UP_CMDID,
171 .vdev_stop_cmdid = WMI_10X_VDEV_STOP_CMDID,
172 .vdev_down_cmdid = WMI_10X_VDEV_DOWN_CMDID,
173 .vdev_set_param_cmdid = WMI_10X_VDEV_SET_PARAM_CMDID,
174 .vdev_install_key_cmdid = WMI_10X_VDEV_INSTALL_KEY_CMDID,
175 .peer_create_cmdid = WMI_10X_PEER_CREATE_CMDID,
176 .peer_delete_cmdid = WMI_10X_PEER_DELETE_CMDID,
177 .peer_flush_tids_cmdid = WMI_10X_PEER_FLUSH_TIDS_CMDID,
178 .peer_set_param_cmdid = WMI_10X_PEER_SET_PARAM_CMDID,
179 .peer_assoc_cmdid = WMI_10X_PEER_ASSOC_CMDID,
180 .peer_add_wds_entry_cmdid = WMI_10X_PEER_ADD_WDS_ENTRY_CMDID,
181 .peer_remove_wds_entry_cmdid = WMI_10X_PEER_REMOVE_WDS_ENTRY_CMDID,
182 .peer_mcast_group_cmdid = WMI_10X_PEER_MCAST_GROUP_CMDID,
183 .bcn_tx_cmdid = WMI_10X_BCN_TX_CMDID,
184 .pdev_send_bcn_cmdid = WMI_10X_PDEV_SEND_BCN_CMDID,
185 .bcn_tmpl_cmdid = WMI_CMD_UNSUPPORTED,
186 .bcn_filter_rx_cmdid = WMI_10X_BCN_FILTER_RX_CMDID,
187 .prb_req_filter_rx_cmdid = WMI_10X_PRB_REQ_FILTER_RX_CMDID,
188 .mgmt_tx_cmdid = WMI_10X_MGMT_TX_CMDID,
189 .prb_tmpl_cmdid = WMI_CMD_UNSUPPORTED,
190 .addba_clear_resp_cmdid = WMI_10X_ADDBA_CLEAR_RESP_CMDID,
191 .addba_send_cmdid = WMI_10X_ADDBA_SEND_CMDID,
192 .addba_status_cmdid = WMI_10X_ADDBA_STATUS_CMDID,
193 .delba_send_cmdid = WMI_10X_DELBA_SEND_CMDID,
194 .addba_set_resp_cmdid = WMI_10X_ADDBA_SET_RESP_CMDID,
195 .send_singleamsdu_cmdid = WMI_10X_SEND_SINGLEAMSDU_CMDID,
196 .sta_powersave_mode_cmdid = WMI_10X_STA_POWERSAVE_MODE_CMDID,
197 .sta_powersave_param_cmdid = WMI_10X_STA_POWERSAVE_PARAM_CMDID,
198 .sta_mimo_ps_mode_cmdid = WMI_10X_STA_MIMO_PS_MODE_CMDID,
199 .pdev_dfs_enable_cmdid = WMI_10X_PDEV_DFS_ENABLE_CMDID,
200 .pdev_dfs_disable_cmdid = WMI_10X_PDEV_DFS_DISABLE_CMDID,
201 .roam_scan_mode = WMI_10X_ROAM_SCAN_MODE,
202 .roam_scan_rssi_threshold = WMI_10X_ROAM_SCAN_RSSI_THRESHOLD,
203 .roam_scan_period = WMI_10X_ROAM_SCAN_PERIOD,
204 .roam_scan_rssi_change_threshold =
205 WMI_10X_ROAM_SCAN_RSSI_CHANGE_THRESHOLD,
206 .roam_ap_profile = WMI_10X_ROAM_AP_PROFILE,
207 .ofl_scan_add_ap_profile = WMI_10X_OFL_SCAN_ADD_AP_PROFILE,
208 .ofl_scan_remove_ap_profile = WMI_10X_OFL_SCAN_REMOVE_AP_PROFILE,
209 .ofl_scan_period = WMI_10X_OFL_SCAN_PERIOD,
210 .p2p_dev_set_device_info = WMI_10X_P2P_DEV_SET_DEVICE_INFO,
211 .p2p_dev_set_discoverability = WMI_10X_P2P_DEV_SET_DISCOVERABILITY,
212 .p2p_go_set_beacon_ie = WMI_10X_P2P_GO_SET_BEACON_IE,
213 .p2p_go_set_probe_resp_ie = WMI_10X_P2P_GO_SET_PROBE_RESP_IE,
214 .p2p_set_vendor_ie_data_cmdid = WMI_CMD_UNSUPPORTED,
215 .ap_ps_peer_param_cmdid = WMI_CMD_UNSUPPORTED,
216 .ap_ps_peer_uapsd_coex_cmdid = WMI_CMD_UNSUPPORTED,
217 .peer_rate_retry_sched_cmdid = WMI_10X_PEER_RATE_RETRY_SCHED_CMDID,
218 .wlan_profile_trigger_cmdid = WMI_10X_WLAN_PROFILE_TRIGGER_CMDID,
219 .wlan_profile_set_hist_intvl_cmdid =
220 WMI_10X_WLAN_PROFILE_SET_HIST_INTVL_CMDID,
221 .wlan_profile_get_profile_data_cmdid =
222 WMI_10X_WLAN_PROFILE_GET_PROFILE_DATA_CMDID,
223 .wlan_profile_enable_profile_id_cmdid =
224 WMI_10X_WLAN_PROFILE_ENABLE_PROFILE_ID_CMDID,
225 .wlan_profile_list_profile_id_cmdid =
226 WMI_10X_WLAN_PROFILE_LIST_PROFILE_ID_CMDID,
227 .pdev_suspend_cmdid = WMI_10X_PDEV_SUSPEND_CMDID,
228 .pdev_resume_cmdid = WMI_10X_PDEV_RESUME_CMDID,
229 .add_bcn_filter_cmdid = WMI_10X_ADD_BCN_FILTER_CMDID,
230 .rmv_bcn_filter_cmdid = WMI_10X_RMV_BCN_FILTER_CMDID,
231 .wow_add_wake_pattern_cmdid = WMI_10X_WOW_ADD_WAKE_PATTERN_CMDID,
232 .wow_del_wake_pattern_cmdid = WMI_10X_WOW_DEL_WAKE_PATTERN_CMDID,
233 .wow_enable_disable_wake_event_cmdid =
234 WMI_10X_WOW_ENABLE_DISABLE_WAKE_EVENT_CMDID,
235 .wow_enable_cmdid = WMI_10X_WOW_ENABLE_CMDID,
236 .wow_hostwakeup_from_sleep_cmdid =
237 WMI_10X_WOW_HOSTWAKEUP_FROM_SLEEP_CMDID,
238 .rtt_measreq_cmdid = WMI_10X_RTT_MEASREQ_CMDID,
239 .rtt_tsf_cmdid = WMI_10X_RTT_TSF_CMDID,
240 .vdev_spectral_scan_configure_cmdid =
241 WMI_10X_VDEV_SPECTRAL_SCAN_CONFIGURE_CMDID,
242 .vdev_spectral_scan_enable_cmdid =
243 WMI_10X_VDEV_SPECTRAL_SCAN_ENABLE_CMDID,
244 .request_stats_cmdid = WMI_10X_REQUEST_STATS_CMDID,
245 .set_arp_ns_offload_cmdid = WMI_CMD_UNSUPPORTED,
246 .network_list_offload_config_cmdid = WMI_CMD_UNSUPPORTED,
247 .gtk_offload_cmdid = WMI_CMD_UNSUPPORTED,
248 .csa_offload_enable_cmdid = WMI_CMD_UNSUPPORTED,
249 .csa_offload_chanswitch_cmdid = WMI_CMD_UNSUPPORTED,
250 .chatter_set_mode_cmdid = WMI_CMD_UNSUPPORTED,
251 .peer_tid_addba_cmdid = WMI_CMD_UNSUPPORTED,
252 .peer_tid_delba_cmdid = WMI_CMD_UNSUPPORTED,
253 .sta_dtim_ps_method_cmdid = WMI_CMD_UNSUPPORTED,
254 .sta_uapsd_auto_trig_cmdid = WMI_CMD_UNSUPPORTED,
255 .sta_keepalive_cmd = WMI_CMD_UNSUPPORTED,
256 .echo_cmdid = WMI_10X_ECHO_CMDID,
257 .pdev_utf_cmdid = WMI_10X_PDEV_UTF_CMDID,
258 .dbglog_cfg_cmdid = WMI_10X_DBGLOG_CFG_CMDID,
259 .pdev_qvit_cmdid = WMI_10X_PDEV_QVIT_CMDID,
260 .pdev_ftm_intg_cmdid = WMI_CMD_UNSUPPORTED,
261 .vdev_set_keepalive_cmdid = WMI_CMD_UNSUPPORTED,
262 .vdev_get_keepalive_cmdid = WMI_CMD_UNSUPPORTED,
263 .force_fw_hang_cmdid = WMI_CMD_UNSUPPORTED,
264 .gpio_config_cmdid = WMI_10X_GPIO_CONFIG_CMDID,
265 .gpio_output_cmdid = WMI_10X_GPIO_OUTPUT_CMDID,
266};
267
268/* MAIN WMI VDEV param map */
269static struct wmi_vdev_param_map wmi_vdev_param_map = {
270 .rts_threshold = WMI_VDEV_PARAM_RTS_THRESHOLD,
271 .fragmentation_threshold = WMI_VDEV_PARAM_FRAGMENTATION_THRESHOLD,
272 .beacon_interval = WMI_VDEV_PARAM_BEACON_INTERVAL,
273 .listen_interval = WMI_VDEV_PARAM_LISTEN_INTERVAL,
274 .multicast_rate = WMI_VDEV_PARAM_MULTICAST_RATE,
275 .mgmt_tx_rate = WMI_VDEV_PARAM_MGMT_TX_RATE,
276 .slot_time = WMI_VDEV_PARAM_SLOT_TIME,
277 .preamble = WMI_VDEV_PARAM_PREAMBLE,
278 .swba_time = WMI_VDEV_PARAM_SWBA_TIME,
279 .wmi_vdev_stats_update_period = WMI_VDEV_STATS_UPDATE_PERIOD,
280 .wmi_vdev_pwrsave_ageout_time = WMI_VDEV_PWRSAVE_AGEOUT_TIME,
281 .wmi_vdev_host_swba_interval = WMI_VDEV_HOST_SWBA_INTERVAL,
282 .dtim_period = WMI_VDEV_PARAM_DTIM_PERIOD,
283 .wmi_vdev_oc_scheduler_air_time_limit =
284 WMI_VDEV_OC_SCHEDULER_AIR_TIME_LIMIT,
285 .wds = WMI_VDEV_PARAM_WDS,
286 .atim_window = WMI_VDEV_PARAM_ATIM_WINDOW,
287 .bmiss_count_max = WMI_VDEV_PARAM_BMISS_COUNT_MAX,
288 .bmiss_first_bcnt = WMI_VDEV_PARAM_BMISS_FIRST_BCNT,
289 .bmiss_final_bcnt = WMI_VDEV_PARAM_BMISS_FINAL_BCNT,
290 .feature_wmm = WMI_VDEV_PARAM_FEATURE_WMM,
291 .chwidth = WMI_VDEV_PARAM_CHWIDTH,
292 .chextoffset = WMI_VDEV_PARAM_CHEXTOFFSET,
293 .disable_htprotection = WMI_VDEV_PARAM_DISABLE_HTPROTECTION,
294 .sta_quickkickout = WMI_VDEV_PARAM_STA_QUICKKICKOUT,
295 .mgmt_rate = WMI_VDEV_PARAM_MGMT_RATE,
296 .protection_mode = WMI_VDEV_PARAM_PROTECTION_MODE,
297 .fixed_rate = WMI_VDEV_PARAM_FIXED_RATE,
298 .sgi = WMI_VDEV_PARAM_SGI,
299 .ldpc = WMI_VDEV_PARAM_LDPC,
300 .tx_stbc = WMI_VDEV_PARAM_TX_STBC,
301 .rx_stbc = WMI_VDEV_PARAM_RX_STBC,
302 .intra_bss_fwd = WMI_VDEV_PARAM_INTRA_BSS_FWD,
303 .def_keyid = WMI_VDEV_PARAM_DEF_KEYID,
304 .nss = WMI_VDEV_PARAM_NSS,
305 .bcast_data_rate = WMI_VDEV_PARAM_BCAST_DATA_RATE,
306 .mcast_data_rate = WMI_VDEV_PARAM_MCAST_DATA_RATE,
307 .mcast_indicate = WMI_VDEV_PARAM_MCAST_INDICATE,
308 .dhcp_indicate = WMI_VDEV_PARAM_DHCP_INDICATE,
309 .unknown_dest_indicate = WMI_VDEV_PARAM_UNKNOWN_DEST_INDICATE,
310 .ap_keepalive_min_idle_inactive_time_secs =
311 WMI_VDEV_PARAM_AP_KEEPALIVE_MIN_IDLE_INACTIVE_TIME_SECS,
312 .ap_keepalive_max_idle_inactive_time_secs =
313 WMI_VDEV_PARAM_AP_KEEPALIVE_MAX_IDLE_INACTIVE_TIME_SECS,
314 .ap_keepalive_max_unresponsive_time_secs =
315 WMI_VDEV_PARAM_AP_KEEPALIVE_MAX_UNRESPONSIVE_TIME_SECS,
316 .ap_enable_nawds = WMI_VDEV_PARAM_AP_ENABLE_NAWDS,
317 .mcast2ucast_set = WMI_VDEV_PARAM_UNSUPPORTED,
318 .enable_rtscts = WMI_VDEV_PARAM_ENABLE_RTSCTS,
319 .txbf = WMI_VDEV_PARAM_TXBF,
320 .packet_powersave = WMI_VDEV_PARAM_PACKET_POWERSAVE,
321 .drop_unencry = WMI_VDEV_PARAM_DROP_UNENCRY,
322 .tx_encap_type = WMI_VDEV_PARAM_TX_ENCAP_TYPE,
323 .ap_detect_out_of_sync_sleeping_sta_time_secs =
324 WMI_VDEV_PARAM_UNSUPPORTED,
325};
326
327/* 10.X WMI VDEV param map */
328static struct wmi_vdev_param_map wmi_10x_vdev_param_map = {
329 .rts_threshold = WMI_10X_VDEV_PARAM_RTS_THRESHOLD,
330 .fragmentation_threshold = WMI_10X_VDEV_PARAM_FRAGMENTATION_THRESHOLD,
331 .beacon_interval = WMI_10X_VDEV_PARAM_BEACON_INTERVAL,
332 .listen_interval = WMI_10X_VDEV_PARAM_LISTEN_INTERVAL,
333 .multicast_rate = WMI_10X_VDEV_PARAM_MULTICAST_RATE,
334 .mgmt_tx_rate = WMI_10X_VDEV_PARAM_MGMT_TX_RATE,
335 .slot_time = WMI_10X_VDEV_PARAM_SLOT_TIME,
336 .preamble = WMI_10X_VDEV_PARAM_PREAMBLE,
337 .swba_time = WMI_10X_VDEV_PARAM_SWBA_TIME,
338 .wmi_vdev_stats_update_period = WMI_10X_VDEV_STATS_UPDATE_PERIOD,
339 .wmi_vdev_pwrsave_ageout_time = WMI_10X_VDEV_PWRSAVE_AGEOUT_TIME,
340 .wmi_vdev_host_swba_interval = WMI_10X_VDEV_HOST_SWBA_INTERVAL,
341 .dtim_period = WMI_10X_VDEV_PARAM_DTIM_PERIOD,
342 .wmi_vdev_oc_scheduler_air_time_limit =
343 WMI_10X_VDEV_OC_SCHEDULER_AIR_TIME_LIMIT,
344 .wds = WMI_10X_VDEV_PARAM_WDS,
345 .atim_window = WMI_10X_VDEV_PARAM_ATIM_WINDOW,
346 .bmiss_count_max = WMI_10X_VDEV_PARAM_BMISS_COUNT_MAX,
347 .bmiss_first_bcnt = WMI_VDEV_PARAM_UNSUPPORTED,
348 .bmiss_final_bcnt = WMI_VDEV_PARAM_UNSUPPORTED,
349 .feature_wmm = WMI_10X_VDEV_PARAM_FEATURE_WMM,
350 .chwidth = WMI_10X_VDEV_PARAM_CHWIDTH,
351 .chextoffset = WMI_10X_VDEV_PARAM_CHEXTOFFSET,
352 .disable_htprotection = WMI_10X_VDEV_PARAM_DISABLE_HTPROTECTION,
353 .sta_quickkickout = WMI_10X_VDEV_PARAM_STA_QUICKKICKOUT,
354 .mgmt_rate = WMI_10X_VDEV_PARAM_MGMT_RATE,
355 .protection_mode = WMI_10X_VDEV_PARAM_PROTECTION_MODE,
356 .fixed_rate = WMI_10X_VDEV_PARAM_FIXED_RATE,
357 .sgi = WMI_10X_VDEV_PARAM_SGI,
358 .ldpc = WMI_10X_VDEV_PARAM_LDPC,
359 .tx_stbc = WMI_10X_VDEV_PARAM_TX_STBC,
360 .rx_stbc = WMI_10X_VDEV_PARAM_RX_STBC,
361 .intra_bss_fwd = WMI_10X_VDEV_PARAM_INTRA_BSS_FWD,
362 .def_keyid = WMI_10X_VDEV_PARAM_DEF_KEYID,
363 .nss = WMI_10X_VDEV_PARAM_NSS,
364 .bcast_data_rate = WMI_10X_VDEV_PARAM_BCAST_DATA_RATE,
365 .mcast_data_rate = WMI_10X_VDEV_PARAM_MCAST_DATA_RATE,
366 .mcast_indicate = WMI_10X_VDEV_PARAM_MCAST_INDICATE,
367 .dhcp_indicate = WMI_10X_VDEV_PARAM_DHCP_INDICATE,
368 .unknown_dest_indicate = WMI_10X_VDEV_PARAM_UNKNOWN_DEST_INDICATE,
369 .ap_keepalive_min_idle_inactive_time_secs =
370 WMI_10X_VDEV_PARAM_AP_KEEPALIVE_MIN_IDLE_INACTIVE_TIME_SECS,
371 .ap_keepalive_max_idle_inactive_time_secs =
372 WMI_10X_VDEV_PARAM_AP_KEEPALIVE_MAX_IDLE_INACTIVE_TIME_SECS,
373 .ap_keepalive_max_unresponsive_time_secs =
374 WMI_10X_VDEV_PARAM_AP_KEEPALIVE_MAX_UNRESPONSIVE_TIME_SECS,
375 .ap_enable_nawds = WMI_10X_VDEV_PARAM_AP_ENABLE_NAWDS,
376 .mcast2ucast_set = WMI_10X_VDEV_PARAM_MCAST2UCAST_SET,
377 .enable_rtscts = WMI_10X_VDEV_PARAM_ENABLE_RTSCTS,
378 .txbf = WMI_VDEV_PARAM_UNSUPPORTED,
379 .packet_powersave = WMI_VDEV_PARAM_UNSUPPORTED,
380 .drop_unencry = WMI_VDEV_PARAM_UNSUPPORTED,
381 .tx_encap_type = WMI_VDEV_PARAM_UNSUPPORTED,
382 .ap_detect_out_of_sync_sleeping_sta_time_secs =
383 WMI_10X_VDEV_PARAM_AP_DETECT_OUT_OF_SYNC_SLEEPING_STA_TIME_SECS,
384};
385
386static struct wmi_pdev_param_map wmi_pdev_param_map = {
387 .tx_chain_mask = WMI_PDEV_PARAM_TX_CHAIN_MASK,
388 .rx_chain_mask = WMI_PDEV_PARAM_RX_CHAIN_MASK,
389 .txpower_limit2g = WMI_PDEV_PARAM_TXPOWER_LIMIT2G,
390 .txpower_limit5g = WMI_PDEV_PARAM_TXPOWER_LIMIT5G,
391 .txpower_scale = WMI_PDEV_PARAM_TXPOWER_SCALE,
392 .beacon_gen_mode = WMI_PDEV_PARAM_BEACON_GEN_MODE,
393 .beacon_tx_mode = WMI_PDEV_PARAM_BEACON_TX_MODE,
394 .resmgr_offchan_mode = WMI_PDEV_PARAM_RESMGR_OFFCHAN_MODE,
395 .protection_mode = WMI_PDEV_PARAM_PROTECTION_MODE,
396 .dynamic_bw = WMI_PDEV_PARAM_DYNAMIC_BW,
397 .non_agg_sw_retry_th = WMI_PDEV_PARAM_NON_AGG_SW_RETRY_TH,
398 .agg_sw_retry_th = WMI_PDEV_PARAM_AGG_SW_RETRY_TH,
399 .sta_kickout_th = WMI_PDEV_PARAM_STA_KICKOUT_TH,
400 .ac_aggrsize_scaling = WMI_PDEV_PARAM_AC_AGGRSIZE_SCALING,
401 .ltr_enable = WMI_PDEV_PARAM_LTR_ENABLE,
402 .ltr_ac_latency_be = WMI_PDEV_PARAM_LTR_AC_LATENCY_BE,
403 .ltr_ac_latency_bk = WMI_PDEV_PARAM_LTR_AC_LATENCY_BK,
404 .ltr_ac_latency_vi = WMI_PDEV_PARAM_LTR_AC_LATENCY_VI,
405 .ltr_ac_latency_vo = WMI_PDEV_PARAM_LTR_AC_LATENCY_VO,
406 .ltr_ac_latency_timeout = WMI_PDEV_PARAM_LTR_AC_LATENCY_TIMEOUT,
407 .ltr_sleep_override = WMI_PDEV_PARAM_LTR_SLEEP_OVERRIDE,
408 .ltr_rx_override = WMI_PDEV_PARAM_LTR_RX_OVERRIDE,
409 .ltr_tx_activity_timeout = WMI_PDEV_PARAM_LTR_TX_ACTIVITY_TIMEOUT,
410 .l1ss_enable = WMI_PDEV_PARAM_L1SS_ENABLE,
411 .dsleep_enable = WMI_PDEV_PARAM_DSLEEP_ENABLE,
412 .pcielp_txbuf_flush = WMI_PDEV_PARAM_PCIELP_TXBUF_FLUSH,
413 .pcielp_txbuf_watermark = WMI_PDEV_PARAM_PCIELP_TXBUF_TMO_EN,
414 .pcielp_txbuf_tmo_en = WMI_PDEV_PARAM_PCIELP_TXBUF_TMO_EN,
415 .pcielp_txbuf_tmo_value = WMI_PDEV_PARAM_PCIELP_TXBUF_TMO_VALUE,
416 .pdev_stats_update_period = WMI_PDEV_PARAM_PDEV_STATS_UPDATE_PERIOD,
417 .vdev_stats_update_period = WMI_PDEV_PARAM_VDEV_STATS_UPDATE_PERIOD,
418 .peer_stats_update_period = WMI_PDEV_PARAM_PEER_STATS_UPDATE_PERIOD,
419 .bcnflt_stats_update_period = WMI_PDEV_PARAM_BCNFLT_STATS_UPDATE_PERIOD,
420 .pmf_qos = WMI_PDEV_PARAM_PMF_QOS,
421 .arp_ac_override = WMI_PDEV_PARAM_ARP_AC_OVERRIDE,
422 .arpdhcp_ac_override = WMI_PDEV_PARAM_UNSUPPORTED,
423 .dcs = WMI_PDEV_PARAM_DCS,
424 .ani_enable = WMI_PDEV_PARAM_ANI_ENABLE,
425 .ani_poll_period = WMI_PDEV_PARAM_ANI_POLL_PERIOD,
426 .ani_listen_period = WMI_PDEV_PARAM_ANI_LISTEN_PERIOD,
427 .ani_ofdm_level = WMI_PDEV_PARAM_ANI_OFDM_LEVEL,
428 .ani_cck_level = WMI_PDEV_PARAM_ANI_CCK_LEVEL,
429 .dyntxchain = WMI_PDEV_PARAM_DYNTXCHAIN,
430 .proxy_sta = WMI_PDEV_PARAM_PROXY_STA,
431 .idle_ps_config = WMI_PDEV_PARAM_IDLE_PS_CONFIG,
432 .power_gating_sleep = WMI_PDEV_PARAM_POWER_GATING_SLEEP,
433 .fast_channel_reset = WMI_PDEV_PARAM_UNSUPPORTED,
434 .burst_dur = WMI_PDEV_PARAM_UNSUPPORTED,
435 .burst_enable = WMI_PDEV_PARAM_UNSUPPORTED,
436};
437
438static struct wmi_pdev_param_map wmi_10x_pdev_param_map = {
439 .tx_chain_mask = WMI_10X_PDEV_PARAM_TX_CHAIN_MASK,
440 .rx_chain_mask = WMI_10X_PDEV_PARAM_RX_CHAIN_MASK,
441 .txpower_limit2g = WMI_10X_PDEV_PARAM_TXPOWER_LIMIT2G,
442 .txpower_limit5g = WMI_10X_PDEV_PARAM_TXPOWER_LIMIT5G,
443 .txpower_scale = WMI_10X_PDEV_PARAM_TXPOWER_SCALE,
444 .beacon_gen_mode = WMI_10X_PDEV_PARAM_BEACON_GEN_MODE,
445 .beacon_tx_mode = WMI_10X_PDEV_PARAM_BEACON_TX_MODE,
446 .resmgr_offchan_mode = WMI_10X_PDEV_PARAM_RESMGR_OFFCHAN_MODE,
447 .protection_mode = WMI_10X_PDEV_PARAM_PROTECTION_MODE,
448 .dynamic_bw = WMI_10X_PDEV_PARAM_DYNAMIC_BW,
449 .non_agg_sw_retry_th = WMI_10X_PDEV_PARAM_NON_AGG_SW_RETRY_TH,
450 .agg_sw_retry_th = WMI_10X_PDEV_PARAM_AGG_SW_RETRY_TH,
451 .sta_kickout_th = WMI_10X_PDEV_PARAM_STA_KICKOUT_TH,
452 .ac_aggrsize_scaling = WMI_10X_PDEV_PARAM_AC_AGGRSIZE_SCALING,
453 .ltr_enable = WMI_10X_PDEV_PARAM_LTR_ENABLE,
454 .ltr_ac_latency_be = WMI_10X_PDEV_PARAM_LTR_AC_LATENCY_BE,
455 .ltr_ac_latency_bk = WMI_10X_PDEV_PARAM_LTR_AC_LATENCY_BK,
456 .ltr_ac_latency_vi = WMI_10X_PDEV_PARAM_LTR_AC_LATENCY_VI,
457 .ltr_ac_latency_vo = WMI_10X_PDEV_PARAM_LTR_AC_LATENCY_VO,
458 .ltr_ac_latency_timeout = WMI_10X_PDEV_PARAM_LTR_AC_LATENCY_TIMEOUT,
459 .ltr_sleep_override = WMI_10X_PDEV_PARAM_LTR_SLEEP_OVERRIDE,
460 .ltr_rx_override = WMI_10X_PDEV_PARAM_LTR_RX_OVERRIDE,
461 .ltr_tx_activity_timeout = WMI_10X_PDEV_PARAM_LTR_TX_ACTIVITY_TIMEOUT,
462 .l1ss_enable = WMI_10X_PDEV_PARAM_L1SS_ENABLE,
463 .dsleep_enable = WMI_10X_PDEV_PARAM_DSLEEP_ENABLE,
464 .pcielp_txbuf_flush = WMI_PDEV_PARAM_UNSUPPORTED,
465 .pcielp_txbuf_watermark = WMI_PDEV_PARAM_UNSUPPORTED,
466 .pcielp_txbuf_tmo_en = WMI_PDEV_PARAM_UNSUPPORTED,
467 .pcielp_txbuf_tmo_value = WMI_PDEV_PARAM_UNSUPPORTED,
468 .pdev_stats_update_period = WMI_10X_PDEV_PARAM_PDEV_STATS_UPDATE_PERIOD,
469 .vdev_stats_update_period = WMI_10X_PDEV_PARAM_VDEV_STATS_UPDATE_PERIOD,
470 .peer_stats_update_period = WMI_10X_PDEV_PARAM_PEER_STATS_UPDATE_PERIOD,
471 .bcnflt_stats_update_period =
472 WMI_10X_PDEV_PARAM_BCNFLT_STATS_UPDATE_PERIOD,
473 .pmf_qos = WMI_10X_PDEV_PARAM_PMF_QOS,
474 .arp_ac_override = WMI_PDEV_PARAM_UNSUPPORTED,
475 .arpdhcp_ac_override = WMI_10X_PDEV_PARAM_ARPDHCP_AC_OVERRIDE,
476 .dcs = WMI_10X_PDEV_PARAM_DCS,
477 .ani_enable = WMI_10X_PDEV_PARAM_ANI_ENABLE,
478 .ani_poll_period = WMI_10X_PDEV_PARAM_ANI_POLL_PERIOD,
479 .ani_listen_period = WMI_10X_PDEV_PARAM_ANI_LISTEN_PERIOD,
480 .ani_ofdm_level = WMI_10X_PDEV_PARAM_ANI_OFDM_LEVEL,
481 .ani_cck_level = WMI_10X_PDEV_PARAM_ANI_CCK_LEVEL,
482 .dyntxchain = WMI_10X_PDEV_PARAM_DYNTXCHAIN,
483 .proxy_sta = WMI_PDEV_PARAM_UNSUPPORTED,
484 .idle_ps_config = WMI_PDEV_PARAM_UNSUPPORTED,
485 .power_gating_sleep = WMI_PDEV_PARAM_UNSUPPORTED,
486 .fast_channel_reset = WMI_10X_PDEV_PARAM_FAST_CHANNEL_RESET,
487 .burst_dur = WMI_10X_PDEV_PARAM_BURST_DUR,
488 .burst_enable = WMI_10X_PDEV_PARAM_BURST_ENABLE,
489};
49 490
50int ath10k_wmi_wait_for_service_ready(struct ath10k *ar) 491int ath10k_wmi_wait_for_service_ready(struct ath10k *ar)
51{ 492{
@@ -85,18 +526,14 @@ static struct sk_buff *ath10k_wmi_alloc_skb(u32 len)
85static void ath10k_wmi_htc_tx_complete(struct ath10k *ar, struct sk_buff *skb) 526static void ath10k_wmi_htc_tx_complete(struct ath10k *ar, struct sk_buff *skb)
86{ 527{
87 dev_kfree_skb(skb); 528 dev_kfree_skb(skb);
88
89 if (atomic_sub_return(1, &ar->wmi.pending_tx_count) == 0)
90 wake_up(&ar->wmi.wq);
91} 529}
92 530
93/* WMI command API */ 531static int ath10k_wmi_cmd_send_nowait(struct ath10k *ar, struct sk_buff *skb,
94static int ath10k_wmi_cmd_send(struct ath10k *ar, struct sk_buff *skb, 532 u32 cmd_id)
95 enum wmi_cmd_id cmd_id)
96{ 533{
97 struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(skb); 534 struct ath10k_skb_cb *skb_cb = ATH10K_SKB_CB(skb);
98 struct wmi_cmd_hdr *cmd_hdr; 535 struct wmi_cmd_hdr *cmd_hdr;
99 int status; 536 int ret;
100 u32 cmd = 0; 537 u32 cmd = 0;
101 538
102 if (skb_push(skb, sizeof(struct wmi_cmd_hdr)) == NULL) 539 if (skb_push(skb, sizeof(struct wmi_cmd_hdr)) == NULL)
@@ -107,25 +544,146 @@ static int ath10k_wmi_cmd_send(struct ath10k *ar, struct sk_buff *skb,
107 cmd_hdr = (struct wmi_cmd_hdr *)skb->data; 544 cmd_hdr = (struct wmi_cmd_hdr *)skb->data;
108 cmd_hdr->cmd_id = __cpu_to_le32(cmd); 545 cmd_hdr->cmd_id = __cpu_to_le32(cmd);
109 546
110 if (atomic_add_return(1, &ar->wmi.pending_tx_count) > 547 memset(skb_cb, 0, sizeof(*skb_cb));
111 WMI_MAX_PENDING_TX_COUNT) { 548 ret = ath10k_htc_send(&ar->htc, ar->wmi.eid, skb);
112 /* avoid using up memory when FW hangs */ 549 trace_ath10k_wmi_cmd(cmd_id, skb->data, skb->len, ret);
113 atomic_dec(&ar->wmi.pending_tx_count); 550
114 return -EBUSY; 551 if (ret)
552 goto err_pull;
553
554 return 0;
555
556err_pull:
557 skb_pull(skb, sizeof(struct wmi_cmd_hdr));
558 return ret;
559}
560
561static void ath10k_wmi_tx_beacon_nowait(struct ath10k_vif *arvif)
562{
563 struct wmi_bcn_tx_arg arg = {0};
564 int ret;
565
566 lockdep_assert_held(&arvif->ar->data_lock);
567
568 if (arvif->beacon == NULL)
569 return;
570
571 arg.vdev_id = arvif->vdev_id;
572 arg.tx_rate = 0;
573 arg.tx_power = 0;
574 arg.bcn = arvif->beacon->data;
575 arg.bcn_len = arvif->beacon->len;
576
577 ret = ath10k_wmi_beacon_send_nowait(arvif->ar, &arg);
578 if (ret)
579 return;
580
581 dev_kfree_skb_any(arvif->beacon);
582 arvif->beacon = NULL;
583}
584
585static void ath10k_wmi_tx_beacons_iter(void *data, u8 *mac,
586 struct ieee80211_vif *vif)
587{
588 struct ath10k_vif *arvif = ath10k_vif_to_arvif(vif);
589
590 ath10k_wmi_tx_beacon_nowait(arvif);
591}
592
593static void ath10k_wmi_tx_beacons_nowait(struct ath10k *ar)
594{
595 spin_lock_bh(&ar->data_lock);
596 ieee80211_iterate_active_interfaces_atomic(ar->hw,
597 IEEE80211_IFACE_ITER_NORMAL,
598 ath10k_wmi_tx_beacons_iter,
599 NULL);
600 spin_unlock_bh(&ar->data_lock);
601}
602
603static void ath10k_wmi_op_ep_tx_credits(struct ath10k *ar)
604{
605 /* try to send pending beacons first. they take priority */
606 ath10k_wmi_tx_beacons_nowait(ar);
607
608 wake_up(&ar->wmi.tx_credits_wq);
609}
610
611static int ath10k_wmi_cmd_send(struct ath10k *ar, struct sk_buff *skb,
612 u32 cmd_id)
613{
614 int ret = -EOPNOTSUPP;
615
616 might_sleep();
617
618 if (cmd_id == WMI_CMD_UNSUPPORTED) {
619 ath10k_warn("wmi command %d is not supported by firmware\n",
620 cmd_id);
621 return ret;
115 } 622 }
116 623
117 memset(skb_cb, 0, sizeof(*skb_cb)); 624 wait_event_timeout(ar->wmi.tx_credits_wq, ({
625 /* try to send pending beacons first. they take priority */
626 ath10k_wmi_tx_beacons_nowait(ar);
118 627
119 trace_ath10k_wmi_cmd(cmd_id, skb->data, skb->len); 628 ret = ath10k_wmi_cmd_send_nowait(ar, skb, cmd_id);
629 (ret != -EAGAIN);
630 }), 3*HZ);
120 631
121 status = ath10k_htc_send(&ar->htc, ar->wmi.eid, skb); 632 if (ret)
122 if (status) {
123 dev_kfree_skb_any(skb); 633 dev_kfree_skb_any(skb);
124 atomic_dec(&ar->wmi.pending_tx_count); 634
125 return status; 635 return ret;
636}
637
638int ath10k_wmi_mgmt_tx(struct ath10k *ar, struct sk_buff *skb)
639{
640 int ret = 0;
641 struct wmi_mgmt_tx_cmd *cmd;
642 struct ieee80211_hdr *hdr;
643 struct sk_buff *wmi_skb;
644 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb);
645 int len;
646 u16 fc;
647
648 hdr = (struct ieee80211_hdr *)skb->data;
649 fc = le16_to_cpu(hdr->frame_control);
650
651 if (WARN_ON_ONCE(!ieee80211_is_mgmt(hdr->frame_control)))
652 return -EINVAL;
653
654 len = sizeof(cmd->hdr) + skb->len;
655 len = round_up(len, 4);
656
657 wmi_skb = ath10k_wmi_alloc_skb(len);
658 if (!wmi_skb)
659 return -ENOMEM;
660
661 cmd = (struct wmi_mgmt_tx_cmd *)wmi_skb->data;
662
663 cmd->hdr.vdev_id = __cpu_to_le32(ATH10K_SKB_CB(skb)->vdev_id);
664 cmd->hdr.tx_rate = 0;
665 cmd->hdr.tx_power = 0;
666 cmd->hdr.buf_len = __cpu_to_le32((u32)(skb->len));
667
668 memcpy(cmd->hdr.peer_macaddr.addr, ieee80211_get_DA(hdr), ETH_ALEN);
669 memcpy(cmd->buf, skb->data, skb->len);
670
671 ath10k_dbg(ATH10K_DBG_WMI, "wmi mgmt tx skb %p len %d ftype %02x stype %02x\n",
672 wmi_skb, wmi_skb->len, fc & IEEE80211_FCTL_FTYPE,
673 fc & IEEE80211_FCTL_STYPE);
674
675 /* Send the management frame buffer to the target */
676 ret = ath10k_wmi_cmd_send(ar, wmi_skb, ar->wmi.cmd->mgmt_tx_cmdid);
677 if (ret) {
678 dev_kfree_skb_any(skb);
679 return ret;
126 } 680 }
127 681
128 return 0; 682 /* TODO: report tx status to mac80211 - temporary just ACK */
683 info->flags |= IEEE80211_TX_STAT_ACK;
684 ieee80211_tx_status_irqsafe(ar->hw, skb);
685
686 return ret;
129} 687}
130 688
131static int ath10k_wmi_event_scan(struct ath10k *ar, struct sk_buff *skb) 689static int ath10k_wmi_event_scan(struct ath10k *ar, struct sk_buff *skb)
@@ -315,7 +873,9 @@ static inline u8 get_rate_idx(u32 rate, enum ieee80211_band band)
315 873
316static int ath10k_wmi_event_mgmt_rx(struct ath10k *ar, struct sk_buff *skb) 874static int ath10k_wmi_event_mgmt_rx(struct ath10k *ar, struct sk_buff *skb)
317{ 875{
318 struct wmi_mgmt_rx_event *event = (struct wmi_mgmt_rx_event *)skb->data; 876 struct wmi_mgmt_rx_event_v1 *ev_v1;
877 struct wmi_mgmt_rx_event_v2 *ev_v2;
878 struct wmi_mgmt_rx_hdr_v1 *ev_hdr;
319 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(skb); 879 struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(skb);
320 struct ieee80211_hdr *hdr; 880 struct ieee80211_hdr *hdr;
321 u32 rx_status; 881 u32 rx_status;
@@ -325,13 +885,24 @@ static int ath10k_wmi_event_mgmt_rx(struct ath10k *ar, struct sk_buff *skb)
325 u32 rate; 885 u32 rate;
326 u32 buf_len; 886 u32 buf_len;
327 u16 fc; 887 u16 fc;
888 int pull_len;
889
890 if (test_bit(ATH10K_FW_FEATURE_EXT_WMI_MGMT_RX, ar->fw_features)) {
891 ev_v2 = (struct wmi_mgmt_rx_event_v2 *)skb->data;
892 ev_hdr = &ev_v2->hdr.v1;
893 pull_len = sizeof(*ev_v2);
894 } else {
895 ev_v1 = (struct wmi_mgmt_rx_event_v1 *)skb->data;
896 ev_hdr = &ev_v1->hdr;
897 pull_len = sizeof(*ev_v1);
898 }
328 899
329 channel = __le32_to_cpu(event->hdr.channel); 900 channel = __le32_to_cpu(ev_hdr->channel);
330 buf_len = __le32_to_cpu(event->hdr.buf_len); 901 buf_len = __le32_to_cpu(ev_hdr->buf_len);
331 rx_status = __le32_to_cpu(event->hdr.status); 902 rx_status = __le32_to_cpu(ev_hdr->status);
332 snr = __le32_to_cpu(event->hdr.snr); 903 snr = __le32_to_cpu(ev_hdr->snr);
333 phy_mode = __le32_to_cpu(event->hdr.phy_mode); 904 phy_mode = __le32_to_cpu(ev_hdr->phy_mode);
334 rate = __le32_to_cpu(event->hdr.rate); 905 rate = __le32_to_cpu(ev_hdr->rate);
335 906
336 memset(status, 0, sizeof(*status)); 907 memset(status, 0, sizeof(*status));
337 908
@@ -358,7 +929,7 @@ static int ath10k_wmi_event_mgmt_rx(struct ath10k *ar, struct sk_buff *skb)
358 status->signal = snr + ATH10K_DEFAULT_NOISE_FLOOR; 929 status->signal = snr + ATH10K_DEFAULT_NOISE_FLOOR;
359 status->rate_idx = get_rate_idx(rate, status->band); 930 status->rate_idx = get_rate_idx(rate, status->band);
360 931
361 skb_pull(skb, sizeof(event->hdr)); 932 skb_pull(skb, pull_len);
362 933
363 hdr = (struct ieee80211_hdr *)skb->data; 934 hdr = (struct ieee80211_hdr *)skb->data;
364 fc = le16_to_cpu(hdr->frame_control); 935 fc = le16_to_cpu(hdr->frame_control);
@@ -734,10 +1305,8 @@ static void ath10k_wmi_event_host_swba(struct ath10k *ar, struct sk_buff *skb)
734 int i = -1; 1305 int i = -1;
735 struct wmi_bcn_info *bcn_info; 1306 struct wmi_bcn_info *bcn_info;
736 struct ath10k_vif *arvif; 1307 struct ath10k_vif *arvif;
737 struct wmi_bcn_tx_arg arg;
738 struct sk_buff *bcn; 1308 struct sk_buff *bcn;
739 int vdev_id = 0; 1309 int vdev_id = 0;
740 int ret;
741 1310
742 ath10k_dbg(ATH10K_DBG_MGMT, "WMI_HOST_SWBA_EVENTID\n"); 1311 ath10k_dbg(ATH10K_DBG_MGMT, "WMI_HOST_SWBA_EVENTID\n");
743 1312
@@ -794,17 +1363,17 @@ static void ath10k_wmi_event_host_swba(struct ath10k *ar, struct sk_buff *skb)
794 ath10k_wmi_update_tim(ar, arvif, bcn, bcn_info); 1363 ath10k_wmi_update_tim(ar, arvif, bcn, bcn_info);
795 ath10k_wmi_update_noa(ar, arvif, bcn, bcn_info); 1364 ath10k_wmi_update_noa(ar, arvif, bcn, bcn_info);
796 1365
797 arg.vdev_id = arvif->vdev_id; 1366 spin_lock_bh(&ar->data_lock);
798 arg.tx_rate = 0; 1367 if (arvif->beacon) {
799 arg.tx_power = 0; 1368 ath10k_warn("SWBA overrun on vdev %d\n",
800 arg.bcn = bcn->data; 1369 arvif->vdev_id);
801 arg.bcn_len = bcn->len; 1370 dev_kfree_skb_any(arvif->beacon);
1371 }
802 1372
803 ret = ath10k_wmi_beacon_send(ar, &arg); 1373 arvif->beacon = bcn;
804 if (ret)
805 ath10k_warn("could not send beacon (%d)\n", ret);
806 1374
807 dev_kfree_skb_any(bcn); 1375 ath10k_wmi_tx_beacon_nowait(arvif);
1376 spin_unlock_bh(&ar->data_lock);
808 } 1377 }
809} 1378}
810 1379
@@ -919,6 +1488,55 @@ static void ath10k_wmi_event_vdev_install_key_complete(struct ath10k *ar,
919 ath10k_dbg(ATH10K_DBG_WMI, "WMI_VDEV_INSTALL_KEY_COMPLETE_EVENTID\n"); 1488 ath10k_dbg(ATH10K_DBG_WMI, "WMI_VDEV_INSTALL_KEY_COMPLETE_EVENTID\n");
920} 1489}
921 1490
1491static void ath10k_wmi_event_inst_rssi_stats(struct ath10k *ar,
1492 struct sk_buff *skb)
1493{
1494 ath10k_dbg(ATH10K_DBG_WMI, "WMI_INST_RSSI_STATS_EVENTID\n");
1495}
1496
1497static void ath10k_wmi_event_vdev_standby_req(struct ath10k *ar,
1498 struct sk_buff *skb)
1499{
1500 ath10k_dbg(ATH10K_DBG_WMI, "WMI_VDEV_STANDBY_REQ_EVENTID\n");
1501}
1502
1503static void ath10k_wmi_event_vdev_resume_req(struct ath10k *ar,
1504 struct sk_buff *skb)
1505{
1506 ath10k_dbg(ATH10K_DBG_WMI, "WMI_VDEV_RESUME_REQ_EVENTID\n");
1507}
1508
1509static int ath10k_wmi_alloc_host_mem(struct ath10k *ar, u32 req_id,
1510 u32 num_units, u32 unit_len)
1511{
1512 dma_addr_t paddr;
1513 u32 pool_size;
1514 int idx = ar->wmi.num_mem_chunks;
1515
1516 pool_size = num_units * round_up(unit_len, 4);
1517
1518 if (!pool_size)
1519 return -EINVAL;
1520
1521 ar->wmi.mem_chunks[idx].vaddr = dma_alloc_coherent(ar->dev,
1522 pool_size,
1523 &paddr,
1524 GFP_ATOMIC);
1525 if (!ar->wmi.mem_chunks[idx].vaddr) {
1526 ath10k_warn("failed to allocate memory chunk\n");
1527 return -ENOMEM;
1528 }
1529
1530 memset(ar->wmi.mem_chunks[idx].vaddr, 0, pool_size);
1531
1532 ar->wmi.mem_chunks[idx].paddr = paddr;
1533 ar->wmi.mem_chunks[idx].len = pool_size;
1534 ar->wmi.mem_chunks[idx].req_id = req_id;
1535 ar->wmi.num_mem_chunks++;
1536
1537 return 0;
1538}
1539
922static void ath10k_wmi_service_ready_event_rx(struct ath10k *ar, 1540static void ath10k_wmi_service_ready_event_rx(struct ath10k *ar,
923 struct sk_buff *skb) 1541 struct sk_buff *skb)
924{ 1542{
@@ -943,6 +1561,10 @@ static void ath10k_wmi_service_ready_event_rx(struct ath10k *ar,
943 ar->phy_capability = __le32_to_cpu(ev->phy_capability); 1561 ar->phy_capability = __le32_to_cpu(ev->phy_capability);
944 ar->num_rf_chains = __le32_to_cpu(ev->num_rf_chains); 1562 ar->num_rf_chains = __le32_to_cpu(ev->num_rf_chains);
945 1563
1564 /* only manually set fw features when not using FW IE format */
1565 if (ar->fw_api == 1 && ar->fw_version_build > 636)
1566 set_bit(ATH10K_FW_FEATURE_EXT_WMI_MGMT_RX, ar->fw_features);
1567
946 if (ar->num_rf_chains > WMI_MAX_SPATIAL_STREAM) { 1568 if (ar->num_rf_chains > WMI_MAX_SPATIAL_STREAM) {
947 ath10k_warn("hardware advertises support for more spatial streams than it should (%d > %d)\n", 1569 ath10k_warn("hardware advertises support for more spatial streams than it should (%d > %d)\n",
948 ar->num_rf_chains, WMI_MAX_SPATIAL_STREAM); 1570 ar->num_rf_chains, WMI_MAX_SPATIAL_STREAM);
@@ -987,6 +1609,108 @@ static void ath10k_wmi_service_ready_event_rx(struct ath10k *ar,
987 complete(&ar->wmi.service_ready); 1609 complete(&ar->wmi.service_ready);
988} 1610}
989 1611
1612static void ath10k_wmi_10x_service_ready_event_rx(struct ath10k *ar,
1613 struct sk_buff *skb)
1614{
1615 u32 num_units, req_id, unit_size, num_mem_reqs, num_unit_info, i;
1616 int ret;
1617 struct wmi_service_ready_event_10x *ev = (void *)skb->data;
1618
1619 if (skb->len < sizeof(*ev)) {
1620 ath10k_warn("Service ready event was %d B but expected %zu B. Wrong firmware version?\n",
1621 skb->len, sizeof(*ev));
1622 return;
1623 }
1624
1625 ar->hw_min_tx_power = __le32_to_cpu(ev->hw_min_tx_power);
1626 ar->hw_max_tx_power = __le32_to_cpu(ev->hw_max_tx_power);
1627 ar->ht_cap_info = __le32_to_cpu(ev->ht_cap_info);
1628 ar->vht_cap_info = __le32_to_cpu(ev->vht_cap_info);
1629 ar->fw_version_major =
1630 (__le32_to_cpu(ev->sw_version) & 0xff000000) >> 24;
1631 ar->fw_version_minor = (__le32_to_cpu(ev->sw_version) & 0x00ffffff);
1632 ar->phy_capability = __le32_to_cpu(ev->phy_capability);
1633 ar->num_rf_chains = __le32_to_cpu(ev->num_rf_chains);
1634
1635 if (ar->num_rf_chains > WMI_MAX_SPATIAL_STREAM) {
1636 ath10k_warn("hardware advertises support for more spatial streams than it should (%d > %d)\n",
1637 ar->num_rf_chains, WMI_MAX_SPATIAL_STREAM);
1638 ar->num_rf_chains = WMI_MAX_SPATIAL_STREAM;
1639 }
1640
1641 ar->ath_common.regulatory.current_rd =
1642 __le32_to_cpu(ev->hal_reg_capabilities.eeprom_rd);
1643
1644 ath10k_debug_read_service_map(ar, ev->wmi_service_bitmap,
1645 sizeof(ev->wmi_service_bitmap));
1646
1647 if (strlen(ar->hw->wiphy->fw_version) == 0) {
1648 snprintf(ar->hw->wiphy->fw_version,
1649 sizeof(ar->hw->wiphy->fw_version),
1650 "%u.%u",
1651 ar->fw_version_major,
1652 ar->fw_version_minor);
1653 }
1654
1655 num_mem_reqs = __le32_to_cpu(ev->num_mem_reqs);
1656
1657 if (num_mem_reqs > ATH10K_MAX_MEM_REQS) {
1658 ath10k_warn("requested memory chunks number (%d) exceeds the limit\n",
1659 num_mem_reqs);
1660 return;
1661 }
1662
1663 if (!num_mem_reqs)
1664 goto exit;
1665
1666 ath10k_dbg(ATH10K_DBG_WMI, "firmware has requested %d memory chunks\n",
1667 num_mem_reqs);
1668
1669 for (i = 0; i < num_mem_reqs; ++i) {
1670 req_id = __le32_to_cpu(ev->mem_reqs[i].req_id);
1671 num_units = __le32_to_cpu(ev->mem_reqs[i].num_units);
1672 unit_size = __le32_to_cpu(ev->mem_reqs[i].unit_size);
1673 num_unit_info = __le32_to_cpu(ev->mem_reqs[i].num_unit_info);
1674
1675 if (num_unit_info & NUM_UNITS_IS_NUM_PEERS)
1676 /* number of units to allocate is number of
1677 * peers, 1 extra for self peer on target */
1678 /* this needs to be tied, host and target
1679 * can get out of sync */
1680 num_units = TARGET_10X_NUM_PEERS + 1;
1681 else if (num_unit_info & NUM_UNITS_IS_NUM_VDEVS)
1682 num_units = TARGET_10X_NUM_VDEVS + 1;
1683
1684 ath10k_dbg(ATH10K_DBG_WMI,
1685 "wmi mem_req_id %d num_units %d num_unit_info %d unit size %d actual units %d\n",
1686 req_id,
1687 __le32_to_cpu(ev->mem_reqs[i].num_units),
1688 num_unit_info,
1689 unit_size,
1690 num_units);
1691
1692 ret = ath10k_wmi_alloc_host_mem(ar, req_id, num_units,
1693 unit_size);
1694 if (ret)
1695 return;
1696 }
1697
1698exit:
1699 ath10k_dbg(ATH10K_DBG_WMI,
1700 "wmi event service ready sw_ver 0x%08x abi_ver %u phy_cap 0x%08x ht_cap 0x%08x vht_cap 0x%08x vht_supp_msc 0x%08x sys_cap_info 0x%08x mem_reqs %u num_rf_chains %u\n",
1701 __le32_to_cpu(ev->sw_version),
1702 __le32_to_cpu(ev->abi_version),
1703 __le32_to_cpu(ev->phy_capability),
1704 __le32_to_cpu(ev->ht_cap_info),
1705 __le32_to_cpu(ev->vht_cap_info),
1706 __le32_to_cpu(ev->vht_supp_mcs),
1707 __le32_to_cpu(ev->sys_cap_info),
1708 __le32_to_cpu(ev->num_mem_reqs),
1709 __le32_to_cpu(ev->num_rf_chains));
1710
1711 complete(&ar->wmi.service_ready);
1712}
1713
990static int ath10k_wmi_ready_event_rx(struct ath10k *ar, struct sk_buff *skb) 1714static int ath10k_wmi_ready_event_rx(struct ath10k *ar, struct sk_buff *skb)
991{ 1715{
992 struct wmi_ready_event *ev = (struct wmi_ready_event *)skb->data; 1716 struct wmi_ready_event *ev = (struct wmi_ready_event *)skb->data;
@@ -1007,7 +1731,7 @@ static int ath10k_wmi_ready_event_rx(struct ath10k *ar, struct sk_buff *skb)
1007 return 0; 1731 return 0;
1008} 1732}
1009 1733
1010static void ath10k_wmi_event_process(struct ath10k *ar, struct sk_buff *skb) 1734static void ath10k_wmi_main_process_rx(struct ath10k *ar, struct sk_buff *skb)
1011{ 1735{
1012 struct wmi_cmd_hdr *cmd_hdr; 1736 struct wmi_cmd_hdr *cmd_hdr;
1013 enum wmi_event_id id; 1737 enum wmi_event_id id;
@@ -1126,64 +1850,158 @@ static void ath10k_wmi_event_process(struct ath10k *ar, struct sk_buff *skb)
1126 dev_kfree_skb(skb); 1850 dev_kfree_skb(skb);
1127} 1851}
1128 1852
1129static void ath10k_wmi_event_work(struct work_struct *work) 1853static void ath10k_wmi_10x_process_rx(struct ath10k *ar, struct sk_buff *skb)
1130{ 1854{
1131 struct ath10k *ar = container_of(work, struct ath10k, 1855 struct wmi_cmd_hdr *cmd_hdr;
1132 wmi.wmi_event_work); 1856 enum wmi_10x_event_id id;
1133 struct sk_buff *skb; 1857 u16 len;
1134 1858
1135 for (;;) { 1859 cmd_hdr = (struct wmi_cmd_hdr *)skb->data;
1136 skb = skb_dequeue(&ar->wmi.wmi_event_list); 1860 id = MS(__le32_to_cpu(cmd_hdr->cmd_id), WMI_CMD_HDR_CMD_ID);
1137 if (!skb)
1138 break;
1139 1861
1140 ath10k_wmi_event_process(ar, skb); 1862 if (skb_pull(skb, sizeof(struct wmi_cmd_hdr)) == NULL)
1141 } 1863 return;
1142}
1143 1864
1144static void ath10k_wmi_process_rx(struct ath10k *ar, struct sk_buff *skb) 1865 len = skb->len;
1145{
1146 struct wmi_cmd_hdr *cmd_hdr = (struct wmi_cmd_hdr *)skb->data;
1147 enum wmi_event_id event_id;
1148 1866
1149 event_id = MS(__le32_to_cpu(cmd_hdr->cmd_id), WMI_CMD_HDR_CMD_ID); 1867 trace_ath10k_wmi_event(id, skb->data, skb->len);
1150 1868
1151 /* some events require to be handled ASAP 1869 switch (id) {
1152 * thus can't be defered to a worker thread */ 1870 case WMI_10X_MGMT_RX_EVENTID:
1153 switch (event_id) { 1871 ath10k_wmi_event_mgmt_rx(ar, skb);
1154 case WMI_HOST_SWBA_EVENTID: 1872 /* mgmt_rx() owns the skb now! */
1155 case WMI_MGMT_RX_EVENTID:
1156 ath10k_wmi_event_process(ar, skb);
1157 return; 1873 return;
1874 case WMI_10X_SCAN_EVENTID:
1875 ath10k_wmi_event_scan(ar, skb);
1876 break;
1877 case WMI_10X_CHAN_INFO_EVENTID:
1878 ath10k_wmi_event_chan_info(ar, skb);
1879 break;
1880 case WMI_10X_ECHO_EVENTID:
1881 ath10k_wmi_event_echo(ar, skb);
1882 break;
1883 case WMI_10X_DEBUG_MESG_EVENTID:
1884 ath10k_wmi_event_debug_mesg(ar, skb);
1885 break;
1886 case WMI_10X_UPDATE_STATS_EVENTID:
1887 ath10k_wmi_event_update_stats(ar, skb);
1888 break;
1889 case WMI_10X_VDEV_START_RESP_EVENTID:
1890 ath10k_wmi_event_vdev_start_resp(ar, skb);
1891 break;
1892 case WMI_10X_VDEV_STOPPED_EVENTID:
1893 ath10k_wmi_event_vdev_stopped(ar, skb);
1894 break;
1895 case WMI_10X_PEER_STA_KICKOUT_EVENTID:
1896 ath10k_wmi_event_peer_sta_kickout(ar, skb);
1897 break;
1898 case WMI_10X_HOST_SWBA_EVENTID:
1899 ath10k_wmi_event_host_swba(ar, skb);
1900 break;
1901 case WMI_10X_TBTTOFFSET_UPDATE_EVENTID:
1902 ath10k_wmi_event_tbttoffset_update(ar, skb);
1903 break;
1904 case WMI_10X_PHYERR_EVENTID:
1905 ath10k_wmi_event_phyerr(ar, skb);
1906 break;
1907 case WMI_10X_ROAM_EVENTID:
1908 ath10k_wmi_event_roam(ar, skb);
1909 break;
1910 case WMI_10X_PROFILE_MATCH:
1911 ath10k_wmi_event_profile_match(ar, skb);
1912 break;
1913 case WMI_10X_DEBUG_PRINT_EVENTID:
1914 ath10k_wmi_event_debug_print(ar, skb);
1915 break;
1916 case WMI_10X_PDEV_QVIT_EVENTID:
1917 ath10k_wmi_event_pdev_qvit(ar, skb);
1918 break;
1919 case WMI_10X_WLAN_PROFILE_DATA_EVENTID:
1920 ath10k_wmi_event_wlan_profile_data(ar, skb);
1921 break;
1922 case WMI_10X_RTT_MEASUREMENT_REPORT_EVENTID:
1923 ath10k_wmi_event_rtt_measurement_report(ar, skb);
1924 break;
1925 case WMI_10X_TSF_MEASUREMENT_REPORT_EVENTID:
1926 ath10k_wmi_event_tsf_measurement_report(ar, skb);
1927 break;
1928 case WMI_10X_RTT_ERROR_REPORT_EVENTID:
1929 ath10k_wmi_event_rtt_error_report(ar, skb);
1930 break;
1931 case WMI_10X_WOW_WAKEUP_HOST_EVENTID:
1932 ath10k_wmi_event_wow_wakeup_host(ar, skb);
1933 break;
1934 case WMI_10X_DCS_INTERFERENCE_EVENTID:
1935 ath10k_wmi_event_dcs_interference(ar, skb);
1936 break;
1937 case WMI_10X_PDEV_TPC_CONFIG_EVENTID:
1938 ath10k_wmi_event_pdev_tpc_config(ar, skb);
1939 break;
1940 case WMI_10X_INST_RSSI_STATS_EVENTID:
1941 ath10k_wmi_event_inst_rssi_stats(ar, skb);
1942 break;
1943 case WMI_10X_VDEV_STANDBY_REQ_EVENTID:
1944 ath10k_wmi_event_vdev_standby_req(ar, skb);
1945 break;
1946 case WMI_10X_VDEV_RESUME_REQ_EVENTID:
1947 ath10k_wmi_event_vdev_resume_req(ar, skb);
1948 break;
1949 case WMI_10X_SERVICE_READY_EVENTID:
1950 ath10k_wmi_10x_service_ready_event_rx(ar, skb);
1951 break;
1952 case WMI_10X_READY_EVENTID:
1953 ath10k_wmi_ready_event_rx(ar, skb);
1954 break;
1158 default: 1955 default:
1956 ath10k_warn("Unknown eventid: %d\n", id);
1159 break; 1957 break;
1160 } 1958 }
1161 1959
1162 skb_queue_tail(&ar->wmi.wmi_event_list, skb); 1960 dev_kfree_skb(skb);
1163 queue_work(ar->workqueue, &ar->wmi.wmi_event_work); 1961}
1962
1963
1964static void ath10k_wmi_process_rx(struct ath10k *ar, struct sk_buff *skb)
1965{
1966 if (test_bit(ATH10K_FW_FEATURE_WMI_10X, ar->fw_features))
1967 ath10k_wmi_10x_process_rx(ar, skb);
1968 else
1969 ath10k_wmi_main_process_rx(ar, skb);
1164} 1970}
1165 1971
1166/* WMI Initialization functions */ 1972/* WMI Initialization functions */
1167int ath10k_wmi_attach(struct ath10k *ar) 1973int ath10k_wmi_attach(struct ath10k *ar)
1168{ 1974{
1975 if (test_bit(ATH10K_FW_FEATURE_WMI_10X, ar->fw_features)) {
1976 ar->wmi.cmd = &wmi_10x_cmd_map;
1977 ar->wmi.vdev_param = &wmi_10x_vdev_param_map;
1978 ar->wmi.pdev_param = &wmi_10x_pdev_param_map;
1979 } else {
1980 ar->wmi.cmd = &wmi_cmd_map;
1981 ar->wmi.vdev_param = &wmi_vdev_param_map;
1982 ar->wmi.pdev_param = &wmi_pdev_param_map;
1983 }
1984
1169 init_completion(&ar->wmi.service_ready); 1985 init_completion(&ar->wmi.service_ready);
1170 init_completion(&ar->wmi.unified_ready); 1986 init_completion(&ar->wmi.unified_ready);
1171 init_waitqueue_head(&ar->wmi.wq); 1987 init_waitqueue_head(&ar->wmi.tx_credits_wq);
1172
1173 skb_queue_head_init(&ar->wmi.wmi_event_list);
1174 INIT_WORK(&ar->wmi.wmi_event_work, ath10k_wmi_event_work);
1175 1988
1176 return 0; 1989 return 0;
1177} 1990}
1178 1991
1179void ath10k_wmi_detach(struct ath10k *ar) 1992void ath10k_wmi_detach(struct ath10k *ar)
1180{ 1993{
1181 /* HTC should've drained the packets already */ 1994 int i;
1182 if (WARN_ON(atomic_read(&ar->wmi.pending_tx_count) > 0)) 1995
1183 ath10k_warn("there are still pending packets\n"); 1996 /* free the host memory chunks requested by firmware */
1997 for (i = 0; i < ar->wmi.num_mem_chunks; i++) {
1998 dma_free_coherent(ar->dev,
1999 ar->wmi.mem_chunks[i].len,
2000 ar->wmi.mem_chunks[i].vaddr,
2001 ar->wmi.mem_chunks[i].paddr);
2002 }
1184 2003
1185 cancel_work_sync(&ar->wmi.wmi_event_work); 2004 ar->wmi.num_mem_chunks = 0;
1186 skb_queue_purge(&ar->wmi.wmi_event_list);
1187} 2005}
1188 2006
1189int ath10k_wmi_connect_htc_service(struct ath10k *ar) 2007int ath10k_wmi_connect_htc_service(struct ath10k *ar)
@@ -1198,6 +2016,7 @@ int ath10k_wmi_connect_htc_service(struct ath10k *ar)
1198 /* these fields are the same for all service endpoints */ 2016 /* these fields are the same for all service endpoints */
1199 conn_req.ep_ops.ep_tx_complete = ath10k_wmi_htc_tx_complete; 2017 conn_req.ep_ops.ep_tx_complete = ath10k_wmi_htc_tx_complete;
1200 conn_req.ep_ops.ep_rx_complete = ath10k_wmi_process_rx; 2018 conn_req.ep_ops.ep_rx_complete = ath10k_wmi_process_rx;
2019 conn_req.ep_ops.ep_tx_credits = ath10k_wmi_op_ep_tx_credits;
1201 2020
1202 /* connect to control service */ 2021 /* connect to control service */
1203 conn_req.service_id = ATH10K_HTC_SVC_ID_WMI_CONTROL; 2022 conn_req.service_id = ATH10K_HTC_SVC_ID_WMI_CONTROL;
@@ -1234,7 +2053,8 @@ int ath10k_wmi_pdev_set_regdomain(struct ath10k *ar, u16 rd, u16 rd2g,
1234 "wmi pdev regdomain rd %x rd2g %x rd5g %x ctl2g %x ctl5g %x\n", 2053 "wmi pdev regdomain rd %x rd2g %x rd5g %x ctl2g %x ctl5g %x\n",
1235 rd, rd2g, rd5g, ctl2g, ctl5g); 2054 rd, rd2g, rd5g, ctl2g, ctl5g);
1236 2055
1237 return ath10k_wmi_cmd_send(ar, skb, WMI_PDEV_SET_REGDOMAIN_CMDID); 2056 return ath10k_wmi_cmd_send(ar, skb,
2057 ar->wmi.cmd->pdev_set_regdomain_cmdid);
1238} 2058}
1239 2059
1240int ath10k_wmi_pdev_set_channel(struct ath10k *ar, 2060int ath10k_wmi_pdev_set_channel(struct ath10k *ar,
@@ -1264,7 +2084,8 @@ int ath10k_wmi_pdev_set_channel(struct ath10k *ar,
1264 "wmi set channel mode %d freq %d\n", 2084 "wmi set channel mode %d freq %d\n",
1265 arg->mode, arg->freq); 2085 arg->mode, arg->freq);
1266 2086
1267 return ath10k_wmi_cmd_send(ar, skb, WMI_PDEV_SET_CHANNEL_CMDID); 2087 return ath10k_wmi_cmd_send(ar, skb,
2088 ar->wmi.cmd->pdev_set_channel_cmdid);
1268} 2089}
1269 2090
1270int ath10k_wmi_pdev_suspend_target(struct ath10k *ar) 2091int ath10k_wmi_pdev_suspend_target(struct ath10k *ar)
@@ -1279,7 +2100,7 @@ int ath10k_wmi_pdev_suspend_target(struct ath10k *ar)
1279 cmd = (struct wmi_pdev_suspend_cmd *)skb->data; 2100 cmd = (struct wmi_pdev_suspend_cmd *)skb->data;
1280 cmd->suspend_opt = WMI_PDEV_SUSPEND; 2101 cmd->suspend_opt = WMI_PDEV_SUSPEND;
1281 2102
1282 return ath10k_wmi_cmd_send(ar, skb, WMI_PDEV_SUSPEND_CMDID); 2103 return ath10k_wmi_cmd_send(ar, skb, ar->wmi.cmd->pdev_suspend_cmdid);
1283} 2104}
1284 2105
1285int ath10k_wmi_pdev_resume_target(struct ath10k *ar) 2106int ath10k_wmi_pdev_resume_target(struct ath10k *ar)
@@ -1290,15 +2111,19 @@ int ath10k_wmi_pdev_resume_target(struct ath10k *ar)
1290 if (skb == NULL) 2111 if (skb == NULL)
1291 return -ENOMEM; 2112 return -ENOMEM;
1292 2113
1293 return ath10k_wmi_cmd_send(ar, skb, WMI_PDEV_RESUME_CMDID); 2114 return ath10k_wmi_cmd_send(ar, skb, ar->wmi.cmd->pdev_resume_cmdid);
1294} 2115}
1295 2116
1296int ath10k_wmi_pdev_set_param(struct ath10k *ar, enum wmi_pdev_param id, 2117int ath10k_wmi_pdev_set_param(struct ath10k *ar, u32 id, u32 value)
1297 u32 value)
1298{ 2118{
1299 struct wmi_pdev_set_param_cmd *cmd; 2119 struct wmi_pdev_set_param_cmd *cmd;
1300 struct sk_buff *skb; 2120 struct sk_buff *skb;
1301 2121
2122 if (id == WMI_PDEV_PARAM_UNSUPPORTED) {
2123 ath10k_warn("pdev param %d not supported by firmware\n", id);
2124 return -EOPNOTSUPP;
2125 }
2126
1302 skb = ath10k_wmi_alloc_skb(sizeof(*cmd)); 2127 skb = ath10k_wmi_alloc_skb(sizeof(*cmd));
1303 if (!skb) 2128 if (!skb)
1304 return -ENOMEM; 2129 return -ENOMEM;
@@ -1309,15 +2134,16 @@ int ath10k_wmi_pdev_set_param(struct ath10k *ar, enum wmi_pdev_param id,
1309 2134
1310 ath10k_dbg(ATH10K_DBG_WMI, "wmi pdev set param %d value %d\n", 2135 ath10k_dbg(ATH10K_DBG_WMI, "wmi pdev set param %d value %d\n",
1311 id, value); 2136 id, value);
1312 return ath10k_wmi_cmd_send(ar, skb, WMI_PDEV_SET_PARAM_CMDID); 2137 return ath10k_wmi_cmd_send(ar, skb, ar->wmi.cmd->pdev_set_param_cmdid);
1313} 2138}
1314 2139
1315int ath10k_wmi_cmd_init(struct ath10k *ar) 2140static int ath10k_wmi_main_cmd_init(struct ath10k *ar)
1316{ 2141{
1317 struct wmi_init_cmd *cmd; 2142 struct wmi_init_cmd *cmd;
1318 struct sk_buff *buf; 2143 struct sk_buff *buf;
1319 struct wmi_resource_config config = {}; 2144 struct wmi_resource_config config = {};
1320 u32 val; 2145 u32 len, val;
2146 int i;
1321 2147
1322 config.num_vdevs = __cpu_to_le32(TARGET_NUM_VDEVS); 2148 config.num_vdevs = __cpu_to_le32(TARGET_NUM_VDEVS);
1323 config.num_peers = __cpu_to_le32(TARGET_NUM_PEERS + TARGET_NUM_VDEVS); 2149 config.num_peers = __cpu_to_le32(TARGET_NUM_PEERS + TARGET_NUM_VDEVS);
@@ -1370,23 +2196,158 @@ int ath10k_wmi_cmd_init(struct ath10k *ar)
1370 config.num_msdu_desc = __cpu_to_le32(TARGET_NUM_MSDU_DESC); 2196 config.num_msdu_desc = __cpu_to_le32(TARGET_NUM_MSDU_DESC);
1371 config.max_frag_entries = __cpu_to_le32(TARGET_MAX_FRAG_ENTRIES); 2197 config.max_frag_entries = __cpu_to_le32(TARGET_MAX_FRAG_ENTRIES);
1372 2198
1373 buf = ath10k_wmi_alloc_skb(sizeof(*cmd)); 2199 len = sizeof(*cmd) +
2200 (sizeof(struct host_memory_chunk) * ar->wmi.num_mem_chunks);
2201
2202 buf = ath10k_wmi_alloc_skb(len);
1374 if (!buf) 2203 if (!buf)
1375 return -ENOMEM; 2204 return -ENOMEM;
1376 2205
1377 cmd = (struct wmi_init_cmd *)buf->data; 2206 cmd = (struct wmi_init_cmd *)buf->data;
1378 cmd->num_host_mem_chunks = 0; 2207
2208 if (ar->wmi.num_mem_chunks == 0) {
2209 cmd->num_host_mem_chunks = 0;
2210 goto out;
2211 }
2212
2213 ath10k_dbg(ATH10K_DBG_WMI, "wmi sending %d memory chunks info.\n",
2214 __cpu_to_le32(ar->wmi.num_mem_chunks));
2215
2216 cmd->num_host_mem_chunks = __cpu_to_le32(ar->wmi.num_mem_chunks);
2217
2218 for (i = 0; i < ar->wmi.num_mem_chunks; i++) {
2219 cmd->host_mem_chunks[i].ptr =
2220 __cpu_to_le32(ar->wmi.mem_chunks[i].paddr);
2221 cmd->host_mem_chunks[i].size =
2222 __cpu_to_le32(ar->wmi.mem_chunks[i].len);
2223 cmd->host_mem_chunks[i].req_id =
2224 __cpu_to_le32(ar->wmi.mem_chunks[i].req_id);
2225
2226 ath10k_dbg(ATH10K_DBG_WMI,
2227 "wmi chunk %d len %d requested, addr 0x%x\n",
2228 i,
2229 cmd->host_mem_chunks[i].size,
2230 cmd->host_mem_chunks[i].ptr);
2231 }
2232out:
1379 memcpy(&cmd->resource_config, &config, sizeof(config)); 2233 memcpy(&cmd->resource_config, &config, sizeof(config));
1380 2234
1381 ath10k_dbg(ATH10K_DBG_WMI, "wmi init\n"); 2235 ath10k_dbg(ATH10K_DBG_WMI, "wmi init\n");
1382 return ath10k_wmi_cmd_send(ar, buf, WMI_INIT_CMDID); 2236 return ath10k_wmi_cmd_send(ar, buf, ar->wmi.cmd->init_cmdid);
1383} 2237}
1384 2238
1385static int ath10k_wmi_start_scan_calc_len(const struct wmi_start_scan_arg *arg) 2239static int ath10k_wmi_10x_cmd_init(struct ath10k *ar)
2240{
2241 struct wmi_init_cmd_10x *cmd;
2242 struct sk_buff *buf;
2243 struct wmi_resource_config_10x config = {};
2244 u32 len, val;
2245 int i;
2246
2247 config.num_vdevs = __cpu_to_le32(TARGET_10X_NUM_VDEVS);
2248 config.num_peers = __cpu_to_le32(TARGET_10X_NUM_PEERS);
2249 config.num_peer_keys = __cpu_to_le32(TARGET_10X_NUM_PEER_KEYS);
2250 config.num_tids = __cpu_to_le32(TARGET_10X_NUM_TIDS);
2251 config.ast_skid_limit = __cpu_to_le32(TARGET_10X_AST_SKID_LIMIT);
2252 config.tx_chain_mask = __cpu_to_le32(TARGET_10X_TX_CHAIN_MASK);
2253 config.rx_chain_mask = __cpu_to_le32(TARGET_10X_RX_CHAIN_MASK);
2254 config.rx_timeout_pri_vo = __cpu_to_le32(TARGET_10X_RX_TIMEOUT_LO_PRI);
2255 config.rx_timeout_pri_vi = __cpu_to_le32(TARGET_10X_RX_TIMEOUT_LO_PRI);
2256 config.rx_timeout_pri_be = __cpu_to_le32(TARGET_10X_RX_TIMEOUT_LO_PRI);
2257 config.rx_timeout_pri_bk = __cpu_to_le32(TARGET_10X_RX_TIMEOUT_HI_PRI);
2258 config.rx_decap_mode = __cpu_to_le32(TARGET_10X_RX_DECAP_MODE);
2259
2260 config.scan_max_pending_reqs =
2261 __cpu_to_le32(TARGET_10X_SCAN_MAX_PENDING_REQS);
2262
2263 config.bmiss_offload_max_vdev =
2264 __cpu_to_le32(TARGET_10X_BMISS_OFFLOAD_MAX_VDEV);
2265
2266 config.roam_offload_max_vdev =
2267 __cpu_to_le32(TARGET_10X_ROAM_OFFLOAD_MAX_VDEV);
2268
2269 config.roam_offload_max_ap_profiles =
2270 __cpu_to_le32(TARGET_10X_ROAM_OFFLOAD_MAX_AP_PROFILES);
2271
2272 config.num_mcast_groups = __cpu_to_le32(TARGET_10X_NUM_MCAST_GROUPS);
2273 config.num_mcast_table_elems =
2274 __cpu_to_le32(TARGET_10X_NUM_MCAST_TABLE_ELEMS);
2275
2276 config.mcast2ucast_mode = __cpu_to_le32(TARGET_10X_MCAST2UCAST_MODE);
2277 config.tx_dbg_log_size = __cpu_to_le32(TARGET_10X_TX_DBG_LOG_SIZE);
2278 config.num_wds_entries = __cpu_to_le32(TARGET_10X_NUM_WDS_ENTRIES);
2279 config.dma_burst_size = __cpu_to_le32(TARGET_10X_DMA_BURST_SIZE);
2280 config.mac_aggr_delim = __cpu_to_le32(TARGET_10X_MAC_AGGR_DELIM);
2281
2282 val = TARGET_10X_RX_SKIP_DEFRAG_TIMEOUT_DUP_DETECTION_CHECK;
2283 config.rx_skip_defrag_timeout_dup_detection_check = __cpu_to_le32(val);
2284
2285 config.vow_config = __cpu_to_le32(TARGET_10X_VOW_CONFIG);
2286
2287 config.num_msdu_desc = __cpu_to_le32(TARGET_10X_NUM_MSDU_DESC);
2288 config.max_frag_entries = __cpu_to_le32(TARGET_10X_MAX_FRAG_ENTRIES);
2289
2290 len = sizeof(*cmd) +
2291 (sizeof(struct host_memory_chunk) * ar->wmi.num_mem_chunks);
2292
2293 buf = ath10k_wmi_alloc_skb(len);
2294 if (!buf)
2295 return -ENOMEM;
2296
2297 cmd = (struct wmi_init_cmd_10x *)buf->data;
2298
2299 if (ar->wmi.num_mem_chunks == 0) {
2300 cmd->num_host_mem_chunks = 0;
2301 goto out;
2302 }
2303
2304 ath10k_dbg(ATH10K_DBG_WMI, "wmi sending %d memory chunks info.\n",
2305 __cpu_to_le32(ar->wmi.num_mem_chunks));
2306
2307 cmd->num_host_mem_chunks = __cpu_to_le32(ar->wmi.num_mem_chunks);
2308
2309 for (i = 0; i < ar->wmi.num_mem_chunks; i++) {
2310 cmd->host_mem_chunks[i].ptr =
2311 __cpu_to_le32(ar->wmi.mem_chunks[i].paddr);
2312 cmd->host_mem_chunks[i].size =
2313 __cpu_to_le32(ar->wmi.mem_chunks[i].len);
2314 cmd->host_mem_chunks[i].req_id =
2315 __cpu_to_le32(ar->wmi.mem_chunks[i].req_id);
2316
2317 ath10k_dbg(ATH10K_DBG_WMI,
2318 "wmi chunk %d len %d requested, addr 0x%x\n",
2319 i,
2320 cmd->host_mem_chunks[i].size,
2321 cmd->host_mem_chunks[i].ptr);
2322 }
2323out:
2324 memcpy(&cmd->resource_config, &config, sizeof(config));
2325
2326 ath10k_dbg(ATH10K_DBG_WMI, "wmi init 10x\n");
2327 return ath10k_wmi_cmd_send(ar, buf, ar->wmi.cmd->init_cmdid);
2328}
2329
2330int ath10k_wmi_cmd_init(struct ath10k *ar)
2331{
2332 int ret;
2333
2334 if (test_bit(ATH10K_FW_FEATURE_WMI_10X, ar->fw_features))
2335 ret = ath10k_wmi_10x_cmd_init(ar);
2336 else
2337 ret = ath10k_wmi_main_cmd_init(ar);
2338
2339 return ret;
2340}
2341
2342static int ath10k_wmi_start_scan_calc_len(struct ath10k *ar,
2343 const struct wmi_start_scan_arg *arg)
1386{ 2344{
1387 int len; 2345 int len;
1388 2346
1389 len = sizeof(struct wmi_start_scan_cmd); 2347 if (test_bit(ATH10K_FW_FEATURE_WMI_10X, ar->fw_features))
2348 len = sizeof(struct wmi_start_scan_cmd_10x);
2349 else
2350 len = sizeof(struct wmi_start_scan_cmd);
1390 2351
1391 if (arg->ie_len) { 2352 if (arg->ie_len) {
1392 if (!arg->ie) 2353 if (!arg->ie)
@@ -1446,7 +2407,7 @@ int ath10k_wmi_start_scan(struct ath10k *ar,
1446 int len = 0; 2407 int len = 0;
1447 int i; 2408 int i;
1448 2409
1449 len = ath10k_wmi_start_scan_calc_len(arg); 2410 len = ath10k_wmi_start_scan_calc_len(ar, arg);
1450 if (len < 0) 2411 if (len < 0)
1451 return len; /* len contains error code here */ 2412 return len; /* len contains error code here */
1452 2413
@@ -1478,7 +2439,14 @@ int ath10k_wmi_start_scan(struct ath10k *ar,
1478 cmd->scan_ctrl_flags = __cpu_to_le32(arg->scan_ctrl_flags); 2439 cmd->scan_ctrl_flags = __cpu_to_le32(arg->scan_ctrl_flags);
1479 2440
1480 /* TLV list starts after fields included in the struct */ 2441 /* TLV list starts after fields included in the struct */
1481 off = sizeof(*cmd); 2442 /* There's just one filed that differes the two start_scan
2443 * structures - burst_duration, which we are not using btw,
2444 no point to make the split here, just shift the buffer to fit with
2445 given FW */
2446 if (test_bit(ATH10K_FW_FEATURE_WMI_10X, ar->fw_features))
2447 off = sizeof(struct wmi_start_scan_cmd_10x);
2448 else
2449 off = sizeof(struct wmi_start_scan_cmd);
1482 2450
1483 if (arg->n_channels) { 2451 if (arg->n_channels) {
1484 channels = (void *)skb->data + off; 2452 channels = (void *)skb->data + off;
@@ -1540,7 +2508,7 @@ int ath10k_wmi_start_scan(struct ath10k *ar,
1540 } 2508 }
1541 2509
1542 ath10k_dbg(ATH10K_DBG_WMI, "wmi start scan\n"); 2510 ath10k_dbg(ATH10K_DBG_WMI, "wmi start scan\n");
1543 return ath10k_wmi_cmd_send(ar, skb, WMI_START_SCAN_CMDID); 2511 return ath10k_wmi_cmd_send(ar, skb, ar->wmi.cmd->start_scan_cmdid);
1544} 2512}
1545 2513
1546void ath10k_wmi_start_scan_init(struct ath10k *ar, 2514void ath10k_wmi_start_scan_init(struct ath10k *ar,
@@ -1556,7 +2524,7 @@ void ath10k_wmi_start_scan_init(struct ath10k *ar,
1556 arg->repeat_probe_time = 0; 2524 arg->repeat_probe_time = 0;
1557 arg->probe_spacing_time = 0; 2525 arg->probe_spacing_time = 0;
1558 arg->idle_time = 0; 2526 arg->idle_time = 0;
1559 arg->max_scan_time = 5000; 2527 arg->max_scan_time = 20000;
1560 arg->probe_delay = 5; 2528 arg->probe_delay = 5;
1561 arg->notify_scan_events = WMI_SCAN_EVENT_STARTED 2529 arg->notify_scan_events = WMI_SCAN_EVENT_STARTED
1562 | WMI_SCAN_EVENT_COMPLETED 2530 | WMI_SCAN_EVENT_COMPLETED
@@ -1600,7 +2568,7 @@ int ath10k_wmi_stop_scan(struct ath10k *ar, const struct wmi_stop_scan_arg *arg)
1600 ath10k_dbg(ATH10K_DBG_WMI, 2568 ath10k_dbg(ATH10K_DBG_WMI,
1601 "wmi stop scan reqid %d req_type %d vdev/scan_id %d\n", 2569 "wmi stop scan reqid %d req_type %d vdev/scan_id %d\n",
1602 arg->req_id, arg->req_type, arg->u.scan_id); 2570 arg->req_id, arg->req_type, arg->u.scan_id);
1603 return ath10k_wmi_cmd_send(ar, skb, WMI_STOP_SCAN_CMDID); 2571 return ath10k_wmi_cmd_send(ar, skb, ar->wmi.cmd->stop_scan_cmdid);
1604} 2572}
1605 2573
1606int ath10k_wmi_vdev_create(struct ath10k *ar, u32 vdev_id, 2574int ath10k_wmi_vdev_create(struct ath10k *ar, u32 vdev_id,
@@ -1625,7 +2593,7 @@ int ath10k_wmi_vdev_create(struct ath10k *ar, u32 vdev_id,
1625 "WMI vdev create: id %d type %d subtype %d macaddr %pM\n", 2593 "WMI vdev create: id %d type %d subtype %d macaddr %pM\n",
1626 vdev_id, type, subtype, macaddr); 2594 vdev_id, type, subtype, macaddr);
1627 2595
1628 return ath10k_wmi_cmd_send(ar, skb, WMI_VDEV_CREATE_CMDID); 2596 return ath10k_wmi_cmd_send(ar, skb, ar->wmi.cmd->vdev_create_cmdid);
1629} 2597}
1630 2598
1631int ath10k_wmi_vdev_delete(struct ath10k *ar, u32 vdev_id) 2599int ath10k_wmi_vdev_delete(struct ath10k *ar, u32 vdev_id)
@@ -1643,20 +2611,20 @@ int ath10k_wmi_vdev_delete(struct ath10k *ar, u32 vdev_id)
1643 ath10k_dbg(ATH10K_DBG_WMI, 2611 ath10k_dbg(ATH10K_DBG_WMI,
1644 "WMI vdev delete id %d\n", vdev_id); 2612 "WMI vdev delete id %d\n", vdev_id);
1645 2613
1646 return ath10k_wmi_cmd_send(ar, skb, WMI_VDEV_DELETE_CMDID); 2614 return ath10k_wmi_cmd_send(ar, skb, ar->wmi.cmd->vdev_delete_cmdid);
1647} 2615}
1648 2616
1649static int ath10k_wmi_vdev_start_restart(struct ath10k *ar, 2617static int ath10k_wmi_vdev_start_restart(struct ath10k *ar,
1650 const struct wmi_vdev_start_request_arg *arg, 2618 const struct wmi_vdev_start_request_arg *arg,
1651 enum wmi_cmd_id cmd_id) 2619 u32 cmd_id)
1652{ 2620{
1653 struct wmi_vdev_start_request_cmd *cmd; 2621 struct wmi_vdev_start_request_cmd *cmd;
1654 struct sk_buff *skb; 2622 struct sk_buff *skb;
1655 const char *cmdname; 2623 const char *cmdname;
1656 u32 flags = 0; 2624 u32 flags = 0;
1657 2625
1658 if (cmd_id != WMI_VDEV_START_REQUEST_CMDID && 2626 if (cmd_id != ar->wmi.cmd->vdev_start_request_cmdid &&
1659 cmd_id != WMI_VDEV_RESTART_REQUEST_CMDID) 2627 cmd_id != ar->wmi.cmd->vdev_restart_request_cmdid)
1660 return -EINVAL; 2628 return -EINVAL;
1661 if (WARN_ON(arg->ssid && arg->ssid_len == 0)) 2629 if (WARN_ON(arg->ssid && arg->ssid_len == 0))
1662 return -EINVAL; 2630 return -EINVAL;
@@ -1665,9 +2633,9 @@ static int ath10k_wmi_vdev_start_restart(struct ath10k *ar,
1665 if (WARN_ON(arg->ssid_len > sizeof(cmd->ssid.ssid))) 2633 if (WARN_ON(arg->ssid_len > sizeof(cmd->ssid.ssid)))
1666 return -EINVAL; 2634 return -EINVAL;
1667 2635
1668 if (cmd_id == WMI_VDEV_START_REQUEST_CMDID) 2636 if (cmd_id == ar->wmi.cmd->vdev_start_request_cmdid)
1669 cmdname = "start"; 2637 cmdname = "start";
1670 else if (cmd_id == WMI_VDEV_RESTART_REQUEST_CMDID) 2638 else if (cmd_id == ar->wmi.cmd->vdev_restart_request_cmdid)
1671 cmdname = "restart"; 2639 cmdname = "restart";
1672 else 2640 else
1673 return -EINVAL; /* should not happen, we already check cmd_id */ 2641 return -EINVAL; /* should not happen, we already check cmd_id */
@@ -1718,15 +2686,17 @@ static int ath10k_wmi_vdev_start_restart(struct ath10k *ar,
1718int ath10k_wmi_vdev_start(struct ath10k *ar, 2686int ath10k_wmi_vdev_start(struct ath10k *ar,
1719 const struct wmi_vdev_start_request_arg *arg) 2687 const struct wmi_vdev_start_request_arg *arg)
1720{ 2688{
1721 return ath10k_wmi_vdev_start_restart(ar, arg, 2689 u32 cmd_id = ar->wmi.cmd->vdev_start_request_cmdid;
1722 WMI_VDEV_START_REQUEST_CMDID); 2690
2691 return ath10k_wmi_vdev_start_restart(ar, arg, cmd_id);
1723} 2692}
1724 2693
1725int ath10k_wmi_vdev_restart(struct ath10k *ar, 2694int ath10k_wmi_vdev_restart(struct ath10k *ar,
1726 const struct wmi_vdev_start_request_arg *arg) 2695 const struct wmi_vdev_start_request_arg *arg)
1727{ 2696{
1728 return ath10k_wmi_vdev_start_restart(ar, arg, 2697 u32 cmd_id = ar->wmi.cmd->vdev_restart_request_cmdid;
1729 WMI_VDEV_RESTART_REQUEST_CMDID); 2698
2699 return ath10k_wmi_vdev_start_restart(ar, arg, cmd_id);
1730} 2700}
1731 2701
1732int ath10k_wmi_vdev_stop(struct ath10k *ar, u32 vdev_id) 2702int ath10k_wmi_vdev_stop(struct ath10k *ar, u32 vdev_id)
@@ -1743,7 +2713,7 @@ int ath10k_wmi_vdev_stop(struct ath10k *ar, u32 vdev_id)
1743 2713
1744 ath10k_dbg(ATH10K_DBG_WMI, "wmi vdev stop id 0x%x\n", vdev_id); 2714 ath10k_dbg(ATH10K_DBG_WMI, "wmi vdev stop id 0x%x\n", vdev_id);
1745 2715
1746 return ath10k_wmi_cmd_send(ar, skb, WMI_VDEV_STOP_CMDID); 2716 return ath10k_wmi_cmd_send(ar, skb, ar->wmi.cmd->vdev_stop_cmdid);
1747} 2717}
1748 2718
1749int ath10k_wmi_vdev_up(struct ath10k *ar, u32 vdev_id, u32 aid, const u8 *bssid) 2719int ath10k_wmi_vdev_up(struct ath10k *ar, u32 vdev_id, u32 aid, const u8 *bssid)
@@ -1758,13 +2728,13 @@ int ath10k_wmi_vdev_up(struct ath10k *ar, u32 vdev_id, u32 aid, const u8 *bssid)
1758 cmd = (struct wmi_vdev_up_cmd *)skb->data; 2728 cmd = (struct wmi_vdev_up_cmd *)skb->data;
1759 cmd->vdev_id = __cpu_to_le32(vdev_id); 2729 cmd->vdev_id = __cpu_to_le32(vdev_id);
1760 cmd->vdev_assoc_id = __cpu_to_le32(aid); 2730 cmd->vdev_assoc_id = __cpu_to_le32(aid);
1761 memcpy(&cmd->vdev_bssid.addr, bssid, 6); 2731 memcpy(&cmd->vdev_bssid.addr, bssid, ETH_ALEN);
1762 2732
1763 ath10k_dbg(ATH10K_DBG_WMI, 2733 ath10k_dbg(ATH10K_DBG_WMI,
1764 "wmi mgmt vdev up id 0x%x assoc id %d bssid %pM\n", 2734 "wmi mgmt vdev up id 0x%x assoc id %d bssid %pM\n",
1765 vdev_id, aid, bssid); 2735 vdev_id, aid, bssid);
1766 2736
1767 return ath10k_wmi_cmd_send(ar, skb, WMI_VDEV_UP_CMDID); 2737 return ath10k_wmi_cmd_send(ar, skb, ar->wmi.cmd->vdev_up_cmdid);
1768} 2738}
1769 2739
1770int ath10k_wmi_vdev_down(struct ath10k *ar, u32 vdev_id) 2740int ath10k_wmi_vdev_down(struct ath10k *ar, u32 vdev_id)
@@ -1782,15 +2752,22 @@ int ath10k_wmi_vdev_down(struct ath10k *ar, u32 vdev_id)
1782 ath10k_dbg(ATH10K_DBG_WMI, 2752 ath10k_dbg(ATH10K_DBG_WMI,
1783 "wmi mgmt vdev down id 0x%x\n", vdev_id); 2753 "wmi mgmt vdev down id 0x%x\n", vdev_id);
1784 2754
1785 return ath10k_wmi_cmd_send(ar, skb, WMI_VDEV_DOWN_CMDID); 2755 return ath10k_wmi_cmd_send(ar, skb, ar->wmi.cmd->vdev_down_cmdid);
1786} 2756}
1787 2757
1788int ath10k_wmi_vdev_set_param(struct ath10k *ar, u32 vdev_id, 2758int ath10k_wmi_vdev_set_param(struct ath10k *ar, u32 vdev_id,
1789 enum wmi_vdev_param param_id, u32 param_value) 2759 u32 param_id, u32 param_value)
1790{ 2760{
1791 struct wmi_vdev_set_param_cmd *cmd; 2761 struct wmi_vdev_set_param_cmd *cmd;
1792 struct sk_buff *skb; 2762 struct sk_buff *skb;
1793 2763
2764 if (param_id == WMI_VDEV_PARAM_UNSUPPORTED) {
2765 ath10k_dbg(ATH10K_DBG_WMI,
2766 "vdev param %d not supported by firmware\n",
2767 param_id);
2768 return -EOPNOTSUPP;
2769 }
2770
1794 skb = ath10k_wmi_alloc_skb(sizeof(*cmd)); 2771 skb = ath10k_wmi_alloc_skb(sizeof(*cmd));
1795 if (!skb) 2772 if (!skb)
1796 return -ENOMEM; 2773 return -ENOMEM;
@@ -1804,7 +2781,7 @@ int ath10k_wmi_vdev_set_param(struct ath10k *ar, u32 vdev_id,
1804 "wmi vdev id 0x%x set param %d value %d\n", 2781 "wmi vdev id 0x%x set param %d value %d\n",
1805 vdev_id, param_id, param_value); 2782 vdev_id, param_id, param_value);
1806 2783
1807 return ath10k_wmi_cmd_send(ar, skb, WMI_VDEV_SET_PARAM_CMDID); 2784 return ath10k_wmi_cmd_send(ar, skb, ar->wmi.cmd->vdev_set_param_cmdid);
1808} 2785}
1809 2786
1810int ath10k_wmi_vdev_install_key(struct ath10k *ar, 2787int ath10k_wmi_vdev_install_key(struct ath10k *ar,
@@ -1839,7 +2816,8 @@ int ath10k_wmi_vdev_install_key(struct ath10k *ar,
1839 ath10k_dbg(ATH10K_DBG_WMI, 2816 ath10k_dbg(ATH10K_DBG_WMI,
1840 "wmi vdev install key idx %d cipher %d len %d\n", 2817 "wmi vdev install key idx %d cipher %d len %d\n",
1841 arg->key_idx, arg->key_cipher, arg->key_len); 2818 arg->key_idx, arg->key_cipher, arg->key_len);
1842 return ath10k_wmi_cmd_send(ar, skb, WMI_VDEV_INSTALL_KEY_CMDID); 2819 return ath10k_wmi_cmd_send(ar, skb,
2820 ar->wmi.cmd->vdev_install_key_cmdid);
1843} 2821}
1844 2822
1845int ath10k_wmi_peer_create(struct ath10k *ar, u32 vdev_id, 2823int ath10k_wmi_peer_create(struct ath10k *ar, u32 vdev_id,
@@ -1859,7 +2837,7 @@ int ath10k_wmi_peer_create(struct ath10k *ar, u32 vdev_id,
1859 ath10k_dbg(ATH10K_DBG_WMI, 2837 ath10k_dbg(ATH10K_DBG_WMI,
1860 "wmi peer create vdev_id %d peer_addr %pM\n", 2838 "wmi peer create vdev_id %d peer_addr %pM\n",
1861 vdev_id, peer_addr); 2839 vdev_id, peer_addr);
1862 return ath10k_wmi_cmd_send(ar, skb, WMI_PEER_CREATE_CMDID); 2840 return ath10k_wmi_cmd_send(ar, skb, ar->wmi.cmd->peer_create_cmdid);
1863} 2841}
1864 2842
1865int ath10k_wmi_peer_delete(struct ath10k *ar, u32 vdev_id, 2843int ath10k_wmi_peer_delete(struct ath10k *ar, u32 vdev_id,
@@ -1879,7 +2857,7 @@ int ath10k_wmi_peer_delete(struct ath10k *ar, u32 vdev_id,
1879 ath10k_dbg(ATH10K_DBG_WMI, 2857 ath10k_dbg(ATH10K_DBG_WMI,
1880 "wmi peer delete vdev_id %d peer_addr %pM\n", 2858 "wmi peer delete vdev_id %d peer_addr %pM\n",
1881 vdev_id, peer_addr); 2859 vdev_id, peer_addr);
1882 return ath10k_wmi_cmd_send(ar, skb, WMI_PEER_DELETE_CMDID); 2860 return ath10k_wmi_cmd_send(ar, skb, ar->wmi.cmd->peer_delete_cmdid);
1883} 2861}
1884 2862
1885int ath10k_wmi_peer_flush(struct ath10k *ar, u32 vdev_id, 2863int ath10k_wmi_peer_flush(struct ath10k *ar, u32 vdev_id,
@@ -1900,7 +2878,7 @@ int ath10k_wmi_peer_flush(struct ath10k *ar, u32 vdev_id,
1900 ath10k_dbg(ATH10K_DBG_WMI, 2878 ath10k_dbg(ATH10K_DBG_WMI,
1901 "wmi peer flush vdev_id %d peer_addr %pM tids %08x\n", 2879 "wmi peer flush vdev_id %d peer_addr %pM tids %08x\n",
1902 vdev_id, peer_addr, tid_bitmap); 2880 vdev_id, peer_addr, tid_bitmap);
1903 return ath10k_wmi_cmd_send(ar, skb, WMI_PEER_FLUSH_TIDS_CMDID); 2881 return ath10k_wmi_cmd_send(ar, skb, ar->wmi.cmd->peer_flush_tids_cmdid);
1904} 2882}
1905 2883
1906int ath10k_wmi_peer_set_param(struct ath10k *ar, u32 vdev_id, 2884int ath10k_wmi_peer_set_param(struct ath10k *ar, u32 vdev_id,
@@ -1918,13 +2896,13 @@ int ath10k_wmi_peer_set_param(struct ath10k *ar, u32 vdev_id,
1918 cmd->vdev_id = __cpu_to_le32(vdev_id); 2896 cmd->vdev_id = __cpu_to_le32(vdev_id);
1919 cmd->param_id = __cpu_to_le32(param_id); 2897 cmd->param_id = __cpu_to_le32(param_id);
1920 cmd->param_value = __cpu_to_le32(param_value); 2898 cmd->param_value = __cpu_to_le32(param_value);
1921 memcpy(&cmd->peer_macaddr.addr, peer_addr, 6); 2899 memcpy(&cmd->peer_macaddr.addr, peer_addr, ETH_ALEN);
1922 2900
1923 ath10k_dbg(ATH10K_DBG_WMI, 2901 ath10k_dbg(ATH10K_DBG_WMI,
1924 "wmi vdev %d peer 0x%pM set param %d value %d\n", 2902 "wmi vdev %d peer 0x%pM set param %d value %d\n",
1925 vdev_id, peer_addr, param_id, param_value); 2903 vdev_id, peer_addr, param_id, param_value);
1926 2904
1927 return ath10k_wmi_cmd_send(ar, skb, WMI_PEER_SET_PARAM_CMDID); 2905 return ath10k_wmi_cmd_send(ar, skb, ar->wmi.cmd->peer_set_param_cmdid);
1928} 2906}
1929 2907
1930int ath10k_wmi_set_psmode(struct ath10k *ar, u32 vdev_id, 2908int ath10k_wmi_set_psmode(struct ath10k *ar, u32 vdev_id,
@@ -1945,7 +2923,8 @@ int ath10k_wmi_set_psmode(struct ath10k *ar, u32 vdev_id,
1945 "wmi set powersave id 0x%x mode %d\n", 2923 "wmi set powersave id 0x%x mode %d\n",
1946 vdev_id, psmode); 2924 vdev_id, psmode);
1947 2925
1948 return ath10k_wmi_cmd_send(ar, skb, WMI_STA_POWERSAVE_MODE_CMDID); 2926 return ath10k_wmi_cmd_send(ar, skb,
2927 ar->wmi.cmd->sta_powersave_mode_cmdid);
1949} 2928}
1950 2929
1951int ath10k_wmi_set_sta_ps_param(struct ath10k *ar, u32 vdev_id, 2930int ath10k_wmi_set_sta_ps_param(struct ath10k *ar, u32 vdev_id,
@@ -1967,7 +2946,8 @@ int ath10k_wmi_set_sta_ps_param(struct ath10k *ar, u32 vdev_id,
1967 ath10k_dbg(ATH10K_DBG_WMI, 2946 ath10k_dbg(ATH10K_DBG_WMI,
1968 "wmi sta ps param vdev_id 0x%x param %d value %d\n", 2947 "wmi sta ps param vdev_id 0x%x param %d value %d\n",
1969 vdev_id, param_id, value); 2948 vdev_id, param_id, value);
1970 return ath10k_wmi_cmd_send(ar, skb, WMI_STA_POWERSAVE_PARAM_CMDID); 2949 return ath10k_wmi_cmd_send(ar, skb,
2950 ar->wmi.cmd->sta_powersave_param_cmdid);
1971} 2951}
1972 2952
1973int ath10k_wmi_set_ap_ps_param(struct ath10k *ar, u32 vdev_id, const u8 *mac, 2953int ath10k_wmi_set_ap_ps_param(struct ath10k *ar, u32 vdev_id, const u8 *mac,
@@ -1993,7 +2973,8 @@ int ath10k_wmi_set_ap_ps_param(struct ath10k *ar, u32 vdev_id, const u8 *mac,
1993 "wmi ap ps param vdev_id 0x%X param %d value %d mac_addr %pM\n", 2973 "wmi ap ps param vdev_id 0x%X param %d value %d mac_addr %pM\n",
1994 vdev_id, param_id, value, mac); 2974 vdev_id, param_id, value, mac);
1995 2975
1996 return ath10k_wmi_cmd_send(ar, skb, WMI_AP_PS_PEER_PARAM_CMDID); 2976 return ath10k_wmi_cmd_send(ar, skb,
2977 ar->wmi.cmd->ap_ps_peer_param_cmdid);
1997} 2978}
1998 2979
1999int ath10k_wmi_scan_chan_list(struct ath10k *ar, 2980int ath10k_wmi_scan_chan_list(struct ath10k *ar,
@@ -2046,7 +3027,7 @@ int ath10k_wmi_scan_chan_list(struct ath10k *ar,
2046 ci->flags |= __cpu_to_le32(flags); 3027 ci->flags |= __cpu_to_le32(flags);
2047 } 3028 }
2048 3029
2049 return ath10k_wmi_cmd_send(ar, skb, WMI_SCAN_CHAN_LIST_CMDID); 3030 return ath10k_wmi_cmd_send(ar, skb, ar->wmi.cmd->scan_chan_list_cmdid);
2050} 3031}
2051 3032
2052int ath10k_wmi_peer_assoc(struct ath10k *ar, 3033int ath10k_wmi_peer_assoc(struct ath10k *ar,
@@ -2105,10 +3086,11 @@ int ath10k_wmi_peer_assoc(struct ath10k *ar,
2105 ath10k_dbg(ATH10K_DBG_WMI, 3086 ath10k_dbg(ATH10K_DBG_WMI,
2106 "wmi peer assoc vdev %d addr %pM\n", 3087 "wmi peer assoc vdev %d addr %pM\n",
2107 arg->vdev_id, arg->addr); 3088 arg->vdev_id, arg->addr);
2108 return ath10k_wmi_cmd_send(ar, skb, WMI_PEER_ASSOC_CMDID); 3089 return ath10k_wmi_cmd_send(ar, skb, ar->wmi.cmd->peer_assoc_cmdid);
2109} 3090}
2110 3091
2111int ath10k_wmi_beacon_send(struct ath10k *ar, const struct wmi_bcn_tx_arg *arg) 3092int ath10k_wmi_beacon_send_nowait(struct ath10k *ar,
3093 const struct wmi_bcn_tx_arg *arg)
2112{ 3094{
2113 struct wmi_bcn_tx_cmd *cmd; 3095 struct wmi_bcn_tx_cmd *cmd;
2114 struct sk_buff *skb; 3096 struct sk_buff *skb;
@@ -2124,7 +3106,7 @@ int ath10k_wmi_beacon_send(struct ath10k *ar, const struct wmi_bcn_tx_arg *arg)
2124 cmd->hdr.bcn_len = __cpu_to_le32(arg->bcn_len); 3106 cmd->hdr.bcn_len = __cpu_to_le32(arg->bcn_len);
2125 memcpy(cmd->bcn, arg->bcn, arg->bcn_len); 3107 memcpy(cmd->bcn, arg->bcn, arg->bcn_len);
2126 3108
2127 return ath10k_wmi_cmd_send(ar, skb, WMI_BCN_TX_CMDID); 3109 return ath10k_wmi_cmd_send_nowait(ar, skb, ar->wmi.cmd->bcn_tx_cmdid);
2128} 3110}
2129 3111
2130static void ath10k_wmi_pdev_set_wmm_param(struct wmi_wmm_params *params, 3112static void ath10k_wmi_pdev_set_wmm_param(struct wmi_wmm_params *params,
@@ -2155,7 +3137,8 @@ int ath10k_wmi_pdev_set_wmm_params(struct ath10k *ar,
2155 ath10k_wmi_pdev_set_wmm_param(&cmd->ac_vo, &arg->ac_vo); 3137 ath10k_wmi_pdev_set_wmm_param(&cmd->ac_vo, &arg->ac_vo);
2156 3138
2157 ath10k_dbg(ATH10K_DBG_WMI, "wmi pdev set wmm params\n"); 3139 ath10k_dbg(ATH10K_DBG_WMI, "wmi pdev set wmm params\n");
2158 return ath10k_wmi_cmd_send(ar, skb, WMI_PDEV_SET_WMM_PARAMS_CMDID); 3140 return ath10k_wmi_cmd_send(ar, skb,
3141 ar->wmi.cmd->pdev_set_wmm_params_cmdid);
2159} 3142}
2160 3143
2161int ath10k_wmi_request_stats(struct ath10k *ar, enum wmi_stats_id stats_id) 3144int ath10k_wmi_request_stats(struct ath10k *ar, enum wmi_stats_id stats_id)
@@ -2171,7 +3154,7 @@ int ath10k_wmi_request_stats(struct ath10k *ar, enum wmi_stats_id stats_id)
2171 cmd->stats_id = __cpu_to_le32(stats_id); 3154 cmd->stats_id = __cpu_to_le32(stats_id);
2172 3155
2173 ath10k_dbg(ATH10K_DBG_WMI, "wmi request stats %d\n", (int)stats_id); 3156 ath10k_dbg(ATH10K_DBG_WMI, "wmi request stats %d\n", (int)stats_id);
2174 return ath10k_wmi_cmd_send(ar, skb, WMI_REQUEST_STATS_CMDID); 3157 return ath10k_wmi_cmd_send(ar, skb, ar->wmi.cmd->request_stats_cmdid);
2175} 3158}
2176 3159
2177int ath10k_wmi_force_fw_hang(struct ath10k *ar, 3160int ath10k_wmi_force_fw_hang(struct ath10k *ar,
@@ -2190,5 +3173,5 @@ int ath10k_wmi_force_fw_hang(struct ath10k *ar,
2190 3173
2191 ath10k_dbg(ATH10K_DBG_WMI, "wmi force fw hang %d delay %d\n", 3174 ath10k_dbg(ATH10K_DBG_WMI, "wmi force fw hang %d delay %d\n",
2192 type, delay_ms); 3175 type, delay_ms);
2193 return ath10k_wmi_cmd_send(ar, skb, WMI_FORCE_FW_HANG_CMDID); 3176 return ath10k_wmi_cmd_send(ar, skb, ar->wmi.cmd->force_fw_hang_cmdid);
2194} 3177}
diff --git a/drivers/net/wireless/ath/ath10k/wmi.h b/drivers/net/wireless/ath/ath10k/wmi.h
index 2c5a4f8daf2e..78c991aec7f9 100644
--- a/drivers/net/wireless/ath/ath10k/wmi.h
+++ b/drivers/net/wireless/ath/ath10k/wmi.h
@@ -208,6 +208,118 @@ struct wmi_mac_addr {
208 (c_macaddr)[5] = (((pwmi_mac_addr)->word1) >> 8) & 0xff; \ 208 (c_macaddr)[5] = (((pwmi_mac_addr)->word1) >> 8) & 0xff; \
209 } while (0) 209 } while (0)
210 210
211struct wmi_cmd_map {
212 u32 init_cmdid;
213 u32 start_scan_cmdid;
214 u32 stop_scan_cmdid;
215 u32 scan_chan_list_cmdid;
216 u32 scan_sch_prio_tbl_cmdid;
217 u32 pdev_set_regdomain_cmdid;
218 u32 pdev_set_channel_cmdid;
219 u32 pdev_set_param_cmdid;
220 u32 pdev_pktlog_enable_cmdid;
221 u32 pdev_pktlog_disable_cmdid;
222 u32 pdev_set_wmm_params_cmdid;
223 u32 pdev_set_ht_cap_ie_cmdid;
224 u32 pdev_set_vht_cap_ie_cmdid;
225 u32 pdev_set_dscp_tid_map_cmdid;
226 u32 pdev_set_quiet_mode_cmdid;
227 u32 pdev_green_ap_ps_enable_cmdid;
228 u32 pdev_get_tpc_config_cmdid;
229 u32 pdev_set_base_macaddr_cmdid;
230 u32 vdev_create_cmdid;
231 u32 vdev_delete_cmdid;
232 u32 vdev_start_request_cmdid;
233 u32 vdev_restart_request_cmdid;
234 u32 vdev_up_cmdid;
235 u32 vdev_stop_cmdid;
236 u32 vdev_down_cmdid;
237 u32 vdev_set_param_cmdid;
238 u32 vdev_install_key_cmdid;
239 u32 peer_create_cmdid;
240 u32 peer_delete_cmdid;
241 u32 peer_flush_tids_cmdid;
242 u32 peer_set_param_cmdid;
243 u32 peer_assoc_cmdid;
244 u32 peer_add_wds_entry_cmdid;
245 u32 peer_remove_wds_entry_cmdid;
246 u32 peer_mcast_group_cmdid;
247 u32 bcn_tx_cmdid;
248 u32 pdev_send_bcn_cmdid;
249 u32 bcn_tmpl_cmdid;
250 u32 bcn_filter_rx_cmdid;
251 u32 prb_req_filter_rx_cmdid;
252 u32 mgmt_tx_cmdid;
253 u32 prb_tmpl_cmdid;
254 u32 addba_clear_resp_cmdid;
255 u32 addba_send_cmdid;
256 u32 addba_status_cmdid;
257 u32 delba_send_cmdid;
258 u32 addba_set_resp_cmdid;
259 u32 send_singleamsdu_cmdid;
260 u32 sta_powersave_mode_cmdid;
261 u32 sta_powersave_param_cmdid;
262 u32 sta_mimo_ps_mode_cmdid;
263 u32 pdev_dfs_enable_cmdid;
264 u32 pdev_dfs_disable_cmdid;
265 u32 roam_scan_mode;
266 u32 roam_scan_rssi_threshold;
267 u32 roam_scan_period;
268 u32 roam_scan_rssi_change_threshold;
269 u32 roam_ap_profile;
270 u32 ofl_scan_add_ap_profile;
271 u32 ofl_scan_remove_ap_profile;
272 u32 ofl_scan_period;
273 u32 p2p_dev_set_device_info;
274 u32 p2p_dev_set_discoverability;
275 u32 p2p_go_set_beacon_ie;
276 u32 p2p_go_set_probe_resp_ie;
277 u32 p2p_set_vendor_ie_data_cmdid;
278 u32 ap_ps_peer_param_cmdid;
279 u32 ap_ps_peer_uapsd_coex_cmdid;
280 u32 peer_rate_retry_sched_cmdid;
281 u32 wlan_profile_trigger_cmdid;
282 u32 wlan_profile_set_hist_intvl_cmdid;
283 u32 wlan_profile_get_profile_data_cmdid;
284 u32 wlan_profile_enable_profile_id_cmdid;
285 u32 wlan_profile_list_profile_id_cmdid;
286 u32 pdev_suspend_cmdid;
287 u32 pdev_resume_cmdid;
288 u32 add_bcn_filter_cmdid;
289 u32 rmv_bcn_filter_cmdid;
290 u32 wow_add_wake_pattern_cmdid;
291 u32 wow_del_wake_pattern_cmdid;
292 u32 wow_enable_disable_wake_event_cmdid;
293 u32 wow_enable_cmdid;
294 u32 wow_hostwakeup_from_sleep_cmdid;
295 u32 rtt_measreq_cmdid;
296 u32 rtt_tsf_cmdid;
297 u32 vdev_spectral_scan_configure_cmdid;
298 u32 vdev_spectral_scan_enable_cmdid;
299 u32 request_stats_cmdid;
300 u32 set_arp_ns_offload_cmdid;
301 u32 network_list_offload_config_cmdid;
302 u32 gtk_offload_cmdid;
303 u32 csa_offload_enable_cmdid;
304 u32 csa_offload_chanswitch_cmdid;
305 u32 chatter_set_mode_cmdid;
306 u32 peer_tid_addba_cmdid;
307 u32 peer_tid_delba_cmdid;
308 u32 sta_dtim_ps_method_cmdid;
309 u32 sta_uapsd_auto_trig_cmdid;
310 u32 sta_keepalive_cmd;
311 u32 echo_cmdid;
312 u32 pdev_utf_cmdid;
313 u32 dbglog_cfg_cmdid;
314 u32 pdev_qvit_cmdid;
315 u32 pdev_ftm_intg_cmdid;
316 u32 vdev_set_keepalive_cmdid;
317 u32 vdev_get_keepalive_cmdid;
318 u32 force_fw_hang_cmdid;
319 u32 gpio_config_cmdid;
320 u32 gpio_output_cmdid;
321};
322
211/* 323/*
212 * wmi command groups. 324 * wmi command groups.
213 */ 325 */
@@ -247,7 +359,9 @@ enum wmi_cmd_group {
247#define WMI_CMD_GRP(grp_id) (((grp_id) << 12) | 0x1) 359#define WMI_CMD_GRP(grp_id) (((grp_id) << 12) | 0x1)
248#define WMI_EVT_GRP_START_ID(grp_id) (((grp_id) << 12) | 0x1) 360#define WMI_EVT_GRP_START_ID(grp_id) (((grp_id) << 12) | 0x1)
249 361
250/* Command IDs and commande events. */ 362#define WMI_CMD_UNSUPPORTED 0
363
364/* Command IDs and command events for MAIN FW. */
251enum wmi_cmd_id { 365enum wmi_cmd_id {
252 WMI_INIT_CMDID = 0x1, 366 WMI_INIT_CMDID = 0x1,
253 367
@@ -488,6 +602,217 @@ enum wmi_event_id {
488 WMI_GPIO_INPUT_EVENTID = WMI_EVT_GRP_START_ID(WMI_GRP_GPIO), 602 WMI_GPIO_INPUT_EVENTID = WMI_EVT_GRP_START_ID(WMI_GRP_GPIO),
489}; 603};
490 604
605/* Command IDs and command events for 10.X firmware */
606enum wmi_10x_cmd_id {
607 WMI_10X_START_CMDID = 0x9000,
608 WMI_10X_END_CMDID = 0x9FFF,
609
610 /* initialize the wlan sub system */
611 WMI_10X_INIT_CMDID,
612
613 /* Scan specific commands */
614
615 WMI_10X_START_SCAN_CMDID = WMI_10X_START_CMDID,
616 WMI_10X_STOP_SCAN_CMDID,
617 WMI_10X_SCAN_CHAN_LIST_CMDID,
618 WMI_10X_ECHO_CMDID,
619
620 /* PDEV(physical device) specific commands */
621 WMI_10X_PDEV_SET_REGDOMAIN_CMDID,
622 WMI_10X_PDEV_SET_CHANNEL_CMDID,
623 WMI_10X_PDEV_SET_PARAM_CMDID,
624 WMI_10X_PDEV_PKTLOG_ENABLE_CMDID,
625 WMI_10X_PDEV_PKTLOG_DISABLE_CMDID,
626 WMI_10X_PDEV_SET_WMM_PARAMS_CMDID,
627 WMI_10X_PDEV_SET_HT_CAP_IE_CMDID,
628 WMI_10X_PDEV_SET_VHT_CAP_IE_CMDID,
629 WMI_10X_PDEV_SET_BASE_MACADDR_CMDID,
630 WMI_10X_PDEV_SET_DSCP_TID_MAP_CMDID,
631 WMI_10X_PDEV_SET_QUIET_MODE_CMDID,
632 WMI_10X_PDEV_GREEN_AP_PS_ENABLE_CMDID,
633 WMI_10X_PDEV_GET_TPC_CONFIG_CMDID,
634
635 /* VDEV(virtual device) specific commands */
636 WMI_10X_VDEV_CREATE_CMDID,
637 WMI_10X_VDEV_DELETE_CMDID,
638 WMI_10X_VDEV_START_REQUEST_CMDID,
639 WMI_10X_VDEV_RESTART_REQUEST_CMDID,
640 WMI_10X_VDEV_UP_CMDID,
641 WMI_10X_VDEV_STOP_CMDID,
642 WMI_10X_VDEV_DOWN_CMDID,
643 WMI_10X_VDEV_STANDBY_RESPONSE_CMDID,
644 WMI_10X_VDEV_RESUME_RESPONSE_CMDID,
645 WMI_10X_VDEV_SET_PARAM_CMDID,
646 WMI_10X_VDEV_INSTALL_KEY_CMDID,
647
648 /* peer specific commands */
649 WMI_10X_PEER_CREATE_CMDID,
650 WMI_10X_PEER_DELETE_CMDID,
651 WMI_10X_PEER_FLUSH_TIDS_CMDID,
652 WMI_10X_PEER_SET_PARAM_CMDID,
653 WMI_10X_PEER_ASSOC_CMDID,
654 WMI_10X_PEER_ADD_WDS_ENTRY_CMDID,
655 WMI_10X_PEER_REMOVE_WDS_ENTRY_CMDID,
656 WMI_10X_PEER_MCAST_GROUP_CMDID,
657
658 /* beacon/management specific commands */
659
660 WMI_10X_BCN_TX_CMDID,
661 WMI_10X_BCN_PRB_TMPL_CMDID,
662 WMI_10X_BCN_FILTER_RX_CMDID,
663 WMI_10X_PRB_REQ_FILTER_RX_CMDID,
664 WMI_10X_MGMT_TX_CMDID,
665
666 /* commands to directly control ba negotiation directly from host. */
667 WMI_10X_ADDBA_CLEAR_RESP_CMDID,
668 WMI_10X_ADDBA_SEND_CMDID,
669 WMI_10X_ADDBA_STATUS_CMDID,
670 WMI_10X_DELBA_SEND_CMDID,
671 WMI_10X_ADDBA_SET_RESP_CMDID,
672 WMI_10X_SEND_SINGLEAMSDU_CMDID,
673
674 /* Station power save specific config */
675 WMI_10X_STA_POWERSAVE_MODE_CMDID,
676 WMI_10X_STA_POWERSAVE_PARAM_CMDID,
677 WMI_10X_STA_MIMO_PS_MODE_CMDID,
678
679 /* set debug log config */
680 WMI_10X_DBGLOG_CFG_CMDID,
681
682 /* DFS-specific commands */
683 WMI_10X_PDEV_DFS_ENABLE_CMDID,
684 WMI_10X_PDEV_DFS_DISABLE_CMDID,
685
686 /* QVIT specific command id */
687 WMI_10X_PDEV_QVIT_CMDID,
688
689 /* Offload Scan and Roaming related commands */
690 WMI_10X_ROAM_SCAN_MODE,
691 WMI_10X_ROAM_SCAN_RSSI_THRESHOLD,
692 WMI_10X_ROAM_SCAN_PERIOD,
693 WMI_10X_ROAM_SCAN_RSSI_CHANGE_THRESHOLD,
694 WMI_10X_ROAM_AP_PROFILE,
695 WMI_10X_OFL_SCAN_ADD_AP_PROFILE,
696 WMI_10X_OFL_SCAN_REMOVE_AP_PROFILE,
697 WMI_10X_OFL_SCAN_PERIOD,
698
699 /* P2P specific commands */
700 WMI_10X_P2P_DEV_SET_DEVICE_INFO,
701 WMI_10X_P2P_DEV_SET_DISCOVERABILITY,
702 WMI_10X_P2P_GO_SET_BEACON_IE,
703 WMI_10X_P2P_GO_SET_PROBE_RESP_IE,
704
705 /* AP power save specific config */
706 WMI_10X_AP_PS_PEER_PARAM_CMDID,
707 WMI_10X_AP_PS_PEER_UAPSD_COEX_CMDID,
708
709 /* Rate-control specific commands */
710 WMI_10X_PEER_RATE_RETRY_SCHED_CMDID,
711
712 /* WLAN Profiling commands. */
713 WMI_10X_WLAN_PROFILE_TRIGGER_CMDID,
714 WMI_10X_WLAN_PROFILE_SET_HIST_INTVL_CMDID,
715 WMI_10X_WLAN_PROFILE_GET_PROFILE_DATA_CMDID,
716 WMI_10X_WLAN_PROFILE_ENABLE_PROFILE_ID_CMDID,
717 WMI_10X_WLAN_PROFILE_LIST_PROFILE_ID_CMDID,
718
719 /* Suspend resume command Ids */
720 WMI_10X_PDEV_SUSPEND_CMDID,
721 WMI_10X_PDEV_RESUME_CMDID,
722
723 /* Beacon filter commands */
724 WMI_10X_ADD_BCN_FILTER_CMDID,
725 WMI_10X_RMV_BCN_FILTER_CMDID,
726
727 /* WOW Specific WMI commands*/
728 WMI_10X_WOW_ADD_WAKE_PATTERN_CMDID,
729 WMI_10X_WOW_DEL_WAKE_PATTERN_CMDID,
730 WMI_10X_WOW_ENABLE_DISABLE_WAKE_EVENT_CMDID,
731 WMI_10X_WOW_ENABLE_CMDID,
732 WMI_10X_WOW_HOSTWAKEUP_FROM_SLEEP_CMDID,
733
734 /* RTT measurement related cmd */
735 WMI_10X_RTT_MEASREQ_CMDID,
736 WMI_10X_RTT_TSF_CMDID,
737
738 /* transmit beacon by value */
739 WMI_10X_PDEV_SEND_BCN_CMDID,
740
741 /* F/W stats */
742 WMI_10X_VDEV_SPECTRAL_SCAN_CONFIGURE_CMDID,
743 WMI_10X_VDEV_SPECTRAL_SCAN_ENABLE_CMDID,
744 WMI_10X_REQUEST_STATS_CMDID,
745
746 /* GPIO Configuration */
747 WMI_10X_GPIO_CONFIG_CMDID,
748 WMI_10X_GPIO_OUTPUT_CMDID,
749
750 WMI_10X_PDEV_UTF_CMDID = WMI_10X_END_CMDID - 1,
751};
752
753enum wmi_10x_event_id {
754 WMI_10X_SERVICE_READY_EVENTID = 0x8000,
755 WMI_10X_READY_EVENTID,
756 WMI_10X_START_EVENTID = 0x9000,
757 WMI_10X_END_EVENTID = 0x9FFF,
758
759 /* Scan specific events */
760 WMI_10X_SCAN_EVENTID = WMI_10X_START_EVENTID,
761 WMI_10X_ECHO_EVENTID,
762 WMI_10X_DEBUG_MESG_EVENTID,
763 WMI_10X_UPDATE_STATS_EVENTID,
764
765 /* Instantaneous RSSI event */
766 WMI_10X_INST_RSSI_STATS_EVENTID,
767
768 /* VDEV specific events */
769 WMI_10X_VDEV_START_RESP_EVENTID,
770 WMI_10X_VDEV_STANDBY_REQ_EVENTID,
771 WMI_10X_VDEV_RESUME_REQ_EVENTID,
772 WMI_10X_VDEV_STOPPED_EVENTID,
773
774 /* peer specific events */
775 WMI_10X_PEER_STA_KICKOUT_EVENTID,
776
777 /* beacon/mgmt specific events */
778 WMI_10X_HOST_SWBA_EVENTID,
779 WMI_10X_TBTTOFFSET_UPDATE_EVENTID,
780 WMI_10X_MGMT_RX_EVENTID,
781
782 /* Channel stats event */
783 WMI_10X_CHAN_INFO_EVENTID,
784
785 /* PHY Error specific WMI event */
786 WMI_10X_PHYERR_EVENTID,
787
788 /* Roam event to trigger roaming on host */
789 WMI_10X_ROAM_EVENTID,
790
791 /* matching AP found from list of profiles */
792 WMI_10X_PROFILE_MATCH,
793
794 /* debug print message used for tracing FW code while debugging */
795 WMI_10X_DEBUG_PRINT_EVENTID,
796 /* VI spoecific event */
797 WMI_10X_PDEV_QVIT_EVENTID,
798 /* FW code profile data in response to profile request */
799 WMI_10X_WLAN_PROFILE_DATA_EVENTID,
800
801 /*RTT related event ID*/
802 WMI_10X_RTT_MEASUREMENT_REPORT_EVENTID,
803 WMI_10X_TSF_MEASUREMENT_REPORT_EVENTID,
804 WMI_10X_RTT_ERROR_REPORT_EVENTID,
805
806 WMI_10X_WOW_WAKEUP_HOST_EVENTID,
807 WMI_10X_DCS_INTERFERENCE_EVENTID,
808
809 /* TPC config for the current operating channel */
810 WMI_10X_PDEV_TPC_CONFIG_EVENTID,
811
812 WMI_10X_GPIO_INPUT_EVENTID,
813 WMI_10X_PDEV_UTF_EVENTID = WMI_10X_END_EVENTID-1,
814};
815
491enum wmi_phy_mode { 816enum wmi_phy_mode {
492 MODE_11A = 0, /* 11a Mode */ 817 MODE_11A = 0, /* 11a Mode */
493 MODE_11G = 1, /* 11b/g Mode */ 818 MODE_11G = 1, /* 11b/g Mode */
@@ -508,6 +833,48 @@ enum wmi_phy_mode {
508 MODE_MAX = 14 833 MODE_MAX = 14
509}; 834};
510 835
836static inline const char *ath10k_wmi_phymode_str(enum wmi_phy_mode mode)
837{
838 switch (mode) {
839 case MODE_11A:
840 return "11a";
841 case MODE_11G:
842 return "11g";
843 case MODE_11B:
844 return "11b";
845 case MODE_11GONLY:
846 return "11gonly";
847 case MODE_11NA_HT20:
848 return "11na-ht20";
849 case MODE_11NG_HT20:
850 return "11ng-ht20";
851 case MODE_11NA_HT40:
852 return "11na-ht40";
853 case MODE_11NG_HT40:
854 return "11ng-ht40";
855 case MODE_11AC_VHT20:
856 return "11ac-vht20";
857 case MODE_11AC_VHT40:
858 return "11ac-vht40";
859 case MODE_11AC_VHT80:
860 return "11ac-vht80";
861 case MODE_11AC_VHT20_2G:
862 return "11ac-vht20-2g";
863 case MODE_11AC_VHT40_2G:
864 return "11ac-vht40-2g";
865 case MODE_11AC_VHT80_2G:
866 return "11ac-vht80-2g";
867 case MODE_UNKNOWN:
868 /* skip */
869 break;
870
871 /* no default handler to allow compiler to check that the
872 * enum is fully handled */
873 };
874
875 return "<unknown>";
876}
877
511#define WMI_CHAN_LIST_TAG 0x1 878#define WMI_CHAN_LIST_TAG 0x1
512#define WMI_SSID_LIST_TAG 0x2 879#define WMI_SSID_LIST_TAG 0x2
513#define WMI_BSSID_LIST_TAG 0x3 880#define WMI_BSSID_LIST_TAG 0x3
@@ -763,13 +1130,45 @@ struct wmi_service_ready_event {
763 struct wlan_host_mem_req mem_reqs[1]; 1130 struct wlan_host_mem_req mem_reqs[1];
764} __packed; 1131} __packed;
765 1132
766/* 1133/* This is the definition from 10.X firmware branch */
767 * status consists of upper 16 bits fo int status and lower 16 bits of 1134struct wmi_service_ready_event_10x {
768 * module ID that retuned status 1135 __le32 sw_version;
769 */ 1136 __le32 abi_version;
770#define WLAN_INIT_STATUS_SUCCESS 0x0 1137
771#define WLAN_GET_INIT_STATUS_REASON(status) ((status) & 0xffff) 1138 /* WMI_PHY_CAPABILITY */
772#define WLAN_GET_INIT_STATUS_MODULE_ID(status) (((status) >> 16) & 0xffff) 1139 __le32 phy_capability;
1140
1141 /* Maximum number of frag table entries that SW will populate less 1 */
1142 __le32 max_frag_entry;
1143 __le32 wmi_service_bitmap[WMI_SERVICE_BM_SIZE];
1144 __le32 num_rf_chains;
1145
1146 /*
1147 * The following field is only valid for service type
1148 * WMI_SERVICE_11AC
1149 */
1150 __le32 ht_cap_info; /* WMI HT Capability */
1151 __le32 vht_cap_info; /* VHT capability info field of 802.11ac */
1152 __le32 vht_supp_mcs; /* VHT Supported MCS Set field Rx/Tx same */
1153 __le32 hw_min_tx_power;
1154 __le32 hw_max_tx_power;
1155
1156 struct hal_reg_capabilities hal_reg_capabilities;
1157
1158 __le32 sys_cap_info;
1159 __le32 min_pkt_size_enable; /* Enterprise mode short pkt enable */
1160
1161 /*
1162 * request to host to allocate a chuck of memory and pss it down to FW
1163 * via WM_INIT. FW uses this as FW extesnsion memory for saving its
1164 * data structures. Only valid for low latency interfaces like PCIE
1165 * where FW can access this memory directly (or) by DMA.
1166 */
1167 __le32 num_mem_reqs;
1168
1169 struct wlan_host_mem_req mem_reqs[1];
1170} __packed;
1171
773 1172
774#define WMI_SERVICE_READY_TIMEOUT_HZ (5*HZ) 1173#define WMI_SERVICE_READY_TIMEOUT_HZ (5*HZ)
775#define WMI_UNIFIED_READY_TIMEOUT_HZ (5*HZ) 1174#define WMI_UNIFIED_READY_TIMEOUT_HZ (5*HZ)
@@ -978,6 +1377,192 @@ struct wmi_resource_config {
978 __le32 max_frag_entries; 1377 __le32 max_frag_entries;
979} __packed; 1378} __packed;
980 1379
1380struct wmi_resource_config_10x {
1381 /* number of virtual devices (VAPs) to support */
1382 __le32 num_vdevs;
1383
1384 /* number of peer nodes to support */
1385 __le32 num_peers;
1386
1387 /* number of keys per peer */
1388 __le32 num_peer_keys;
1389
1390 /* total number of TX/RX data TIDs */
1391 __le32 num_tids;
1392
1393 /*
1394 * max skid for resolving hash collisions
1395 *
1396 * The address search table is sparse, so that if two MAC addresses
1397 * result in the same hash value, the second of these conflicting
1398 * entries can slide to the next index in the address search table,
1399 * and use it, if it is unoccupied. This ast_skid_limit parameter
1400 * specifies the upper bound on how many subsequent indices to search
1401 * over to find an unoccupied space.
1402 */
1403 __le32 ast_skid_limit;
1404
1405 /*
1406 * the nominal chain mask for transmit
1407 *
1408 * The chain mask may be modified dynamically, e.g. to operate AP
1409 * tx with a reduced number of chains if no clients are associated.
1410 * This configuration parameter specifies the nominal chain-mask that
1411 * should be used when not operating with a reduced set of tx chains.
1412 */
1413 __le32 tx_chain_mask;
1414
1415 /*
1416 * the nominal chain mask for receive
1417 *
1418 * The chain mask may be modified dynamically, e.g. for a client
1419 * to use a reduced number of chains for receive if the traffic to
1420 * the client is low enough that it doesn't require downlink MIMO
1421 * or antenna diversity.
1422 * This configuration parameter specifies the nominal chain-mask that
1423 * should be used when not operating with a reduced set of rx chains.
1424 */
1425 __le32 rx_chain_mask;
1426
1427 /*
1428 * what rx reorder timeout (ms) to use for the AC
1429 *
1430 * Each WMM access class (voice, video, best-effort, background) will
1431 * have its own timeout value to dictate how long to wait for missing
1432 * rx MPDUs to arrive before flushing subsequent MPDUs that have
1433 * already been received.
1434 * This parameter specifies the timeout in milliseconds for each
1435 * class.
1436 */
1437 __le32 rx_timeout_pri_vi;
1438 __le32 rx_timeout_pri_vo;
1439 __le32 rx_timeout_pri_be;
1440 __le32 rx_timeout_pri_bk;
1441
1442 /*
1443 * what mode the rx should decap packets to
1444 *
1445 * MAC can decap to RAW (no decap), native wifi or Ethernet types
1446 * THis setting also determines the default TX behavior, however TX
1447 * behavior can be modified on a per VAP basis during VAP init
1448 */
1449 __le32 rx_decap_mode;
1450
1451 /* what is the maximum scan requests than can be queued */
1452 __le32 scan_max_pending_reqs;
1453
1454 /* maximum VDEV that could use BMISS offload */
1455 __le32 bmiss_offload_max_vdev;
1456
1457 /* maximum VDEV that could use offload roaming */
1458 __le32 roam_offload_max_vdev;
1459
1460 /* maximum AP profiles that would push to offload roaming */
1461 __le32 roam_offload_max_ap_profiles;
1462
1463 /*
1464 * how many groups to use for mcast->ucast conversion
1465 *
1466 * The target's WAL maintains a table to hold information regarding
1467 * which peers belong to a given multicast group, so that if
1468 * multicast->unicast conversion is enabled, the target can convert
1469 * multicast tx frames to a series of unicast tx frames, to each
1470 * peer within the multicast group.
1471 This num_mcast_groups configuration parameter tells the target how
1472 * many multicast groups to provide storage for within its multicast
1473 * group membership table.
1474 */
1475 __le32 num_mcast_groups;
1476
1477 /*
1478 * size to alloc for the mcast membership table
1479 *
1480 * This num_mcast_table_elems configuration parameter tells the
1481 * target how many peer elements it needs to provide storage for in
1482 * its multicast group membership table.
1483 * These multicast group membership table elements are shared by the
1484 * multicast groups stored within the table.
1485 */
1486 __le32 num_mcast_table_elems;
1487
1488 /*
1489 * whether/how to do multicast->unicast conversion
1490 *
1491 * This configuration parameter specifies whether the target should
1492 * perform multicast --> unicast conversion on transmit, and if so,
1493 * what to do if it finds no entries in its multicast group
1494 * membership table for the multicast IP address in the tx frame.
1495 * Configuration value:
1496 * 0 -> Do not perform multicast to unicast conversion.
1497 * 1 -> Convert multicast frames to unicast, if the IP multicast
1498 * address from the tx frame is found in the multicast group
1499 * membership table. If the IP multicast address is not found,
1500 * drop the frame.
1501 * 2 -> Convert multicast frames to unicast, if the IP multicast
1502 * address from the tx frame is found in the multicast group
1503 * membership table. If the IP multicast address is not found,
1504 * transmit the frame as multicast.
1505 */
1506 __le32 mcast2ucast_mode;
1507
1508 /*
1509 * how much memory to allocate for a tx PPDU dbg log
1510 *
1511 * This parameter controls how much memory the target will allocate
1512 * to store a log of tx PPDU meta-information (how large the PPDU
1513 * was, when it was sent, whether it was successful, etc.)
1514 */
1515 __le32 tx_dbg_log_size;
1516
1517 /* how many AST entries to be allocated for WDS */
1518 __le32 num_wds_entries;
1519
1520 /*
1521 * MAC DMA burst size, e.g., For target PCI limit can be
1522 * 0 -default, 1 256B
1523 */
1524 __le32 dma_burst_size;
1525
1526 /*
1527 * Fixed delimiters to be inserted after every MPDU to
1528 * account for interface latency to avoid underrun.
1529 */
1530 __le32 mac_aggr_delim;
1531
1532 /*
1533 * determine whether target is responsible for detecting duplicate
1534 * non-aggregate MPDU and timing out stale fragments.
1535 *
1536 * A-MPDU reordering is always performed on the target.
1537 *
1538 * 0: target responsible for frag timeout and dup checking
1539 * 1: host responsible for frag timeout and dup checking
1540 */
1541 __le32 rx_skip_defrag_timeout_dup_detection_check;
1542
1543 /*
1544 * Configuration for VoW :
1545 * No of Video Nodes to be supported
1546 * and Max no of descriptors for each Video link (node).
1547 */
1548 __le32 vow_config;
1549
1550 /* Number of msdu descriptors target should use */
1551 __le32 num_msdu_desc;
1552
1553 /*
1554 * Max. number of Tx fragments per MSDU
1555 * This parameter controls the max number of Tx fragments per MSDU.
1556 * This is sent by the target as part of the WMI_SERVICE_READY event
1557 * and is overriden by the OS shim as required.
1558 */
1559 __le32 max_frag_entries;
1560} __packed;
1561
1562
1563#define NUM_UNITS_IS_NUM_VDEVS 0x1
1564#define NUM_UNITS_IS_NUM_PEERS 0x2
1565
981/* strucutre describing host memory chunk. */ 1566/* strucutre describing host memory chunk. */
982struct host_memory_chunk { 1567struct host_memory_chunk {
983 /* id of the request that is passed up in service ready */ 1568 /* id of the request that is passed up in service ready */
@@ -999,6 +1584,18 @@ struct wmi_init_cmd {
999 struct host_memory_chunk host_mem_chunks[1]; 1584 struct host_memory_chunk host_mem_chunks[1];
1000} __packed; 1585} __packed;
1001 1586
1587/* _10x stucture is from 10.X FW API */
1588struct wmi_init_cmd_10x {
1589 struct wmi_resource_config_10x resource_config;
1590 __le32 num_host_mem_chunks;
1591
1592 /*
1593 * variable number of host memory chunks.
1594 * This should be the last element in the structure
1595 */
1596 struct host_memory_chunk host_mem_chunks[1];
1597} __packed;
1598
1002/* TLV for channel list */ 1599/* TLV for channel list */
1003struct wmi_chan_list { 1600struct wmi_chan_list {
1004 __le32 tag; /* WMI_CHAN_LIST_TAG */ 1601 __le32 tag; /* WMI_CHAN_LIST_TAG */
@@ -1118,6 +1715,88 @@ struct wmi_start_scan_cmd {
1118 */ 1715 */
1119} __packed; 1716} __packed;
1120 1717
1718/* This is the definition from 10.X firmware branch */
1719struct wmi_start_scan_cmd_10x {
1720 /* Scan ID */
1721 __le32 scan_id;
1722
1723 /* Scan requestor ID */
1724 __le32 scan_req_id;
1725
1726 /* VDEV id(interface) that is requesting scan */
1727 __le32 vdev_id;
1728
1729 /* Scan Priority, input to scan scheduler */
1730 __le32 scan_priority;
1731
1732 /* Scan events subscription */
1733 __le32 notify_scan_events;
1734
1735 /* dwell time in msec on active channels */
1736 __le32 dwell_time_active;
1737
1738 /* dwell time in msec on passive channels */
1739 __le32 dwell_time_passive;
1740
1741 /*
1742 * min time in msec on the BSS channel,only valid if atleast one
1743 * VDEV is active
1744 */
1745 __le32 min_rest_time;
1746
1747 /*
1748 * max rest time in msec on the BSS channel,only valid if at least
1749 * one VDEV is active
1750 */
1751 /*
1752 * the scanner will rest on the bss channel at least min_rest_time
1753 * after min_rest_time the scanner will start checking for tx/rx
1754 * activity on all VDEVs. if there is no activity the scanner will
1755 * switch to off channel. if there is activity the scanner will let
1756 * the radio on the bss channel until max_rest_time expires.at
1757 * max_rest_time scanner will switch to off channel irrespective of
1758 * activity. activity is determined by the idle_time parameter.
1759 */
1760 __le32 max_rest_time;
1761
1762 /*
1763 * time before sending next set of probe requests.
1764 * The scanner keeps repeating probe requests transmission with
1765 * period specified by repeat_probe_time.
1766 * The number of probe requests specified depends on the ssid_list
1767 * and bssid_list
1768 */
1769 __le32 repeat_probe_time;
1770
1771 /* time in msec between 2 consequetive probe requests with in a set. */
1772 __le32 probe_spacing_time;
1773
1774 /*
1775 * data inactivity time in msec on bss channel that will be used by
1776 * scanner for measuring the inactivity.
1777 */
1778 __le32 idle_time;
1779
1780 /* maximum time in msec allowed for scan */
1781 __le32 max_scan_time;
1782
1783 /*
1784 * delay in msec before sending first probe request after switching
1785 * to a channel
1786 */
1787 __le32 probe_delay;
1788
1789 /* Scan control flags */
1790 __le32 scan_ctrl_flags;
1791
1792 /*
1793 * TLV (tag length value ) paramerters follow the scan_cmd structure.
1794 * TLV can contain channel list, bssid list, ssid list and
1795 * ie. the TLV tags are defined above;
1796 */
1797} __packed;
1798
1799
1121struct wmi_ssid_arg { 1800struct wmi_ssid_arg {
1122 int len; 1801 int len;
1123 const u8 *ssid; 1802 const u8 *ssid;
@@ -1268,7 +1947,7 @@ struct wmi_scan_event {
1268 * good idea to pass all the fields in the RX status 1947 * good idea to pass all the fields in the RX status
1269 * descriptor up to the host. 1948 * descriptor up to the host.
1270 */ 1949 */
1271struct wmi_mgmt_rx_hdr { 1950struct wmi_mgmt_rx_hdr_v1 {
1272 __le32 channel; 1951 __le32 channel;
1273 __le32 snr; 1952 __le32 snr;
1274 __le32 rate; 1953 __le32 rate;
@@ -1277,8 +1956,18 @@ struct wmi_mgmt_rx_hdr {
1277 __le32 status; /* %WMI_RX_STATUS_ */ 1956 __le32 status; /* %WMI_RX_STATUS_ */
1278} __packed; 1957} __packed;
1279 1958
1280struct wmi_mgmt_rx_event { 1959struct wmi_mgmt_rx_hdr_v2 {
1281 struct wmi_mgmt_rx_hdr hdr; 1960 struct wmi_mgmt_rx_hdr_v1 v1;
1961 __le32 rssi_ctl[4];
1962} __packed;
1963
1964struct wmi_mgmt_rx_event_v1 {
1965 struct wmi_mgmt_rx_hdr_v1 hdr;
1966 u8 buf[0];
1967} __packed;
1968
1969struct wmi_mgmt_rx_event_v2 {
1970 struct wmi_mgmt_rx_hdr_v2 hdr;
1282 u8 buf[0]; 1971 u8 buf[0];
1283} __packed; 1972} __packed;
1284 1973
@@ -1465,6 +2154,60 @@ struct wmi_csa_event {
1465#define VDEV_DEFAULT_STATS_UPDATE_PERIOD 500 2154#define VDEV_DEFAULT_STATS_UPDATE_PERIOD 500
1466#define PEER_DEFAULT_STATS_UPDATE_PERIOD 500 2155#define PEER_DEFAULT_STATS_UPDATE_PERIOD 500
1467 2156
2157struct wmi_pdev_param_map {
2158 u32 tx_chain_mask;
2159 u32 rx_chain_mask;
2160 u32 txpower_limit2g;
2161 u32 txpower_limit5g;
2162 u32 txpower_scale;
2163 u32 beacon_gen_mode;
2164 u32 beacon_tx_mode;
2165 u32 resmgr_offchan_mode;
2166 u32 protection_mode;
2167 u32 dynamic_bw;
2168 u32 non_agg_sw_retry_th;
2169 u32 agg_sw_retry_th;
2170 u32 sta_kickout_th;
2171 u32 ac_aggrsize_scaling;
2172 u32 ltr_enable;
2173 u32 ltr_ac_latency_be;
2174 u32 ltr_ac_latency_bk;
2175 u32 ltr_ac_latency_vi;
2176 u32 ltr_ac_latency_vo;
2177 u32 ltr_ac_latency_timeout;
2178 u32 ltr_sleep_override;
2179 u32 ltr_rx_override;
2180 u32 ltr_tx_activity_timeout;
2181 u32 l1ss_enable;
2182 u32 dsleep_enable;
2183 u32 pcielp_txbuf_flush;
2184 u32 pcielp_txbuf_watermark;
2185 u32 pcielp_txbuf_tmo_en;
2186 u32 pcielp_txbuf_tmo_value;
2187 u32 pdev_stats_update_period;
2188 u32 vdev_stats_update_period;
2189 u32 peer_stats_update_period;
2190 u32 bcnflt_stats_update_period;
2191 u32 pmf_qos;
2192 u32 arp_ac_override;
2193 u32 arpdhcp_ac_override;
2194 u32 dcs;
2195 u32 ani_enable;
2196 u32 ani_poll_period;
2197 u32 ani_listen_period;
2198 u32 ani_ofdm_level;
2199 u32 ani_cck_level;
2200 u32 dyntxchain;
2201 u32 proxy_sta;
2202 u32 idle_ps_config;
2203 u32 power_gating_sleep;
2204 u32 fast_channel_reset;
2205 u32 burst_dur;
2206 u32 burst_enable;
2207};
2208
2209#define WMI_PDEV_PARAM_UNSUPPORTED 0
2210
1468enum wmi_pdev_param { 2211enum wmi_pdev_param {
1469 /* TX chian mask */ 2212 /* TX chian mask */
1470 WMI_PDEV_PARAM_TX_CHAIN_MASK = 0x1, 2213 WMI_PDEV_PARAM_TX_CHAIN_MASK = 0x1,
@@ -1564,6 +2307,97 @@ enum wmi_pdev_param {
1564 WMI_PDEV_PARAM_POWER_GATING_SLEEP, 2307 WMI_PDEV_PARAM_POWER_GATING_SLEEP,
1565}; 2308};
1566 2309
2310enum wmi_10x_pdev_param {
2311 /* TX chian mask */
2312 WMI_10X_PDEV_PARAM_TX_CHAIN_MASK = 0x1,
2313 /* RX chian mask */
2314 WMI_10X_PDEV_PARAM_RX_CHAIN_MASK,
2315 /* TX power limit for 2G Radio */
2316 WMI_10X_PDEV_PARAM_TXPOWER_LIMIT2G,
2317 /* TX power limit for 5G Radio */
2318 WMI_10X_PDEV_PARAM_TXPOWER_LIMIT5G,
2319 /* TX power scale */
2320 WMI_10X_PDEV_PARAM_TXPOWER_SCALE,
2321 /* Beacon generation mode . 0: host, 1: target */
2322 WMI_10X_PDEV_PARAM_BEACON_GEN_MODE,
2323 /* Beacon generation mode . 0: staggered 1: bursted */
2324 WMI_10X_PDEV_PARAM_BEACON_TX_MODE,
2325 /*
2326 * Resource manager off chan mode .
2327 * 0: turn off off chan mode. 1: turn on offchan mode
2328 */
2329 WMI_10X_PDEV_PARAM_RESMGR_OFFCHAN_MODE,
2330 /*
2331 * Protection mode:
2332 * 0: no protection 1:use CTS-to-self 2: use RTS/CTS
2333 */
2334 WMI_10X_PDEV_PARAM_PROTECTION_MODE,
2335 /* Dynamic bandwidth 0: disable 1: enable */
2336 WMI_10X_PDEV_PARAM_DYNAMIC_BW,
2337 /* Non aggregrate/ 11g sw retry threshold.0-disable */
2338 WMI_10X_PDEV_PARAM_NON_AGG_SW_RETRY_TH,
2339 /* aggregrate sw retry threshold. 0-disable*/
2340 WMI_10X_PDEV_PARAM_AGG_SW_RETRY_TH,
2341 /* Station kickout threshold (non of consecutive failures).0-disable */
2342 WMI_10X_PDEV_PARAM_STA_KICKOUT_TH,
2343 /* Aggerate size scaling configuration per AC */
2344 WMI_10X_PDEV_PARAM_AC_AGGRSIZE_SCALING,
2345 /* LTR enable */
2346 WMI_10X_PDEV_PARAM_LTR_ENABLE,
2347 /* LTR latency for BE, in us */
2348 WMI_10X_PDEV_PARAM_LTR_AC_LATENCY_BE,
2349 /* LTR latency for BK, in us */
2350 WMI_10X_PDEV_PARAM_LTR_AC_LATENCY_BK,
2351 /* LTR latency for VI, in us */
2352 WMI_10X_PDEV_PARAM_LTR_AC_LATENCY_VI,
2353 /* LTR latency for VO, in us */
2354 WMI_10X_PDEV_PARAM_LTR_AC_LATENCY_VO,
2355 /* LTR AC latency timeout, in ms */
2356 WMI_10X_PDEV_PARAM_LTR_AC_LATENCY_TIMEOUT,
2357 /* LTR platform latency override, in us */
2358 WMI_10X_PDEV_PARAM_LTR_SLEEP_OVERRIDE,
2359 /* LTR-RX override, in us */
2360 WMI_10X_PDEV_PARAM_LTR_RX_OVERRIDE,
2361 /* Tx activity timeout for LTR, in us */
2362 WMI_10X_PDEV_PARAM_LTR_TX_ACTIVITY_TIMEOUT,
2363 /* L1SS state machine enable */
2364 WMI_10X_PDEV_PARAM_L1SS_ENABLE,
2365 /* Deep sleep state machine enable */
2366 WMI_10X_PDEV_PARAM_DSLEEP_ENABLE,
2367 /* pdev level stats update period in ms */
2368 WMI_10X_PDEV_PARAM_PDEV_STATS_UPDATE_PERIOD,
2369 /* vdev level stats update period in ms */
2370 WMI_10X_PDEV_PARAM_VDEV_STATS_UPDATE_PERIOD,
2371 /* peer level stats update period in ms */
2372 WMI_10X_PDEV_PARAM_PEER_STATS_UPDATE_PERIOD,
2373 /* beacon filter status update period */
2374 WMI_10X_PDEV_PARAM_BCNFLT_STATS_UPDATE_PERIOD,
2375 /* QOS Mgmt frame protection MFP/PMF 0: disable, 1: enable */
2376 WMI_10X_PDEV_PARAM_PMF_QOS,
2377 /* Access category on which ARP and DHCP frames are sent */
2378 WMI_10X_PDEV_PARAM_ARPDHCP_AC_OVERRIDE,
2379 /* DCS configuration */
2380 WMI_10X_PDEV_PARAM_DCS,
2381 /* Enable/Disable ANI on target */
2382 WMI_10X_PDEV_PARAM_ANI_ENABLE,
2383 /* configure the ANI polling period */
2384 WMI_10X_PDEV_PARAM_ANI_POLL_PERIOD,
2385 /* configure the ANI listening period */
2386 WMI_10X_PDEV_PARAM_ANI_LISTEN_PERIOD,
2387 /* configure OFDM immunity level */
2388 WMI_10X_PDEV_PARAM_ANI_OFDM_LEVEL,
2389 /* configure CCK immunity level */
2390 WMI_10X_PDEV_PARAM_ANI_CCK_LEVEL,
2391 /* Enable/Disable CDD for 1x1 STAs in rate control module */
2392 WMI_10X_PDEV_PARAM_DYNTXCHAIN,
2393 /* Enable/Disable Fast channel reset*/
2394 WMI_10X_PDEV_PARAM_FAST_CHANNEL_RESET,
2395 /* Set Bursting DUR */
2396 WMI_10X_PDEV_PARAM_BURST_DUR,
2397 /* Set Bursting Enable*/
2398 WMI_10X_PDEV_PARAM_BURST_ENABLE,
2399};
2400
1567struct wmi_pdev_set_param_cmd { 2401struct wmi_pdev_set_param_cmd {
1568 __le32 param_id; 2402 __le32 param_id;
1569 __le32 param_value; 2403 __le32 param_value;
@@ -2088,6 +2922,61 @@ enum wmi_rate_preamble {
2088/* Value to disable fixed rate setting */ 2922/* Value to disable fixed rate setting */
2089#define WMI_FIXED_RATE_NONE (0xff) 2923#define WMI_FIXED_RATE_NONE (0xff)
2090 2924
2925struct wmi_vdev_param_map {
2926 u32 rts_threshold;
2927 u32 fragmentation_threshold;
2928 u32 beacon_interval;
2929 u32 listen_interval;
2930 u32 multicast_rate;
2931 u32 mgmt_tx_rate;
2932 u32 slot_time;
2933 u32 preamble;
2934 u32 swba_time;
2935 u32 wmi_vdev_stats_update_period;
2936 u32 wmi_vdev_pwrsave_ageout_time;
2937 u32 wmi_vdev_host_swba_interval;
2938 u32 dtim_period;
2939 u32 wmi_vdev_oc_scheduler_air_time_limit;
2940 u32 wds;
2941 u32 atim_window;
2942 u32 bmiss_count_max;
2943 u32 bmiss_first_bcnt;
2944 u32 bmiss_final_bcnt;
2945 u32 feature_wmm;
2946 u32 chwidth;
2947 u32 chextoffset;
2948 u32 disable_htprotection;
2949 u32 sta_quickkickout;
2950 u32 mgmt_rate;
2951 u32 protection_mode;
2952 u32 fixed_rate;
2953 u32 sgi;
2954 u32 ldpc;
2955 u32 tx_stbc;
2956 u32 rx_stbc;
2957 u32 intra_bss_fwd;
2958 u32 def_keyid;
2959 u32 nss;
2960 u32 bcast_data_rate;
2961 u32 mcast_data_rate;
2962 u32 mcast_indicate;
2963 u32 dhcp_indicate;
2964 u32 unknown_dest_indicate;
2965 u32 ap_keepalive_min_idle_inactive_time_secs;
2966 u32 ap_keepalive_max_idle_inactive_time_secs;
2967 u32 ap_keepalive_max_unresponsive_time_secs;
2968 u32 ap_enable_nawds;
2969 u32 mcast2ucast_set;
2970 u32 enable_rtscts;
2971 u32 txbf;
2972 u32 packet_powersave;
2973 u32 drop_unencry;
2974 u32 tx_encap_type;
2975 u32 ap_detect_out_of_sync_sleeping_sta_time_secs;
2976};
2977
2978#define WMI_VDEV_PARAM_UNSUPPORTED 0
2979
2091/* the definition of different VDEV parameters */ 2980/* the definition of different VDEV parameters */
2092enum wmi_vdev_param { 2981enum wmi_vdev_param {
2093 /* RTS Threshold */ 2982 /* RTS Threshold */
@@ -2219,6 +3108,121 @@ enum wmi_vdev_param {
2219 WMI_VDEV_PARAM_TX_ENCAP_TYPE, 3108 WMI_VDEV_PARAM_TX_ENCAP_TYPE,
2220}; 3109};
2221 3110
3111/* the definition of different VDEV parameters */
3112enum wmi_10x_vdev_param {
3113 /* RTS Threshold */
3114 WMI_10X_VDEV_PARAM_RTS_THRESHOLD = 0x1,
3115 /* Fragmentation threshold */
3116 WMI_10X_VDEV_PARAM_FRAGMENTATION_THRESHOLD,
3117 /* beacon interval in TUs */
3118 WMI_10X_VDEV_PARAM_BEACON_INTERVAL,
3119 /* Listen interval in TUs */
3120 WMI_10X_VDEV_PARAM_LISTEN_INTERVAL,
3121 /* muticast rate in Mbps */
3122 WMI_10X_VDEV_PARAM_MULTICAST_RATE,
3123 /* management frame rate in Mbps */
3124 WMI_10X_VDEV_PARAM_MGMT_TX_RATE,
3125 /* slot time (long vs short) */
3126 WMI_10X_VDEV_PARAM_SLOT_TIME,
3127 /* preamble (long vs short) */
3128 WMI_10X_VDEV_PARAM_PREAMBLE,
3129 /* SWBA time (time before tbtt in msec) */
3130 WMI_10X_VDEV_PARAM_SWBA_TIME,
3131 /* time period for updating VDEV stats */
3132 WMI_10X_VDEV_STATS_UPDATE_PERIOD,
3133 /* age out time in msec for frames queued for station in power save */
3134 WMI_10X_VDEV_PWRSAVE_AGEOUT_TIME,
3135 /*
3136 * Host SWBA interval (time in msec before tbtt for SWBA event
3137 * generation).
3138 */
3139 WMI_10X_VDEV_HOST_SWBA_INTERVAL,
3140 /* DTIM period (specified in units of num beacon intervals) */
3141 WMI_10X_VDEV_PARAM_DTIM_PERIOD,
3142 /*
3143 * scheduler air time limit for this VDEV. used by off chan
3144 * scheduler.
3145 */
3146 WMI_10X_VDEV_OC_SCHEDULER_AIR_TIME_LIMIT,
3147 /* enable/dsiable WDS for this VDEV */
3148 WMI_10X_VDEV_PARAM_WDS,
3149 /* ATIM Window */
3150 WMI_10X_VDEV_PARAM_ATIM_WINDOW,
3151 /* BMISS max */
3152 WMI_10X_VDEV_PARAM_BMISS_COUNT_MAX,
3153 /* WMM enables/disabled */
3154 WMI_10X_VDEV_PARAM_FEATURE_WMM,
3155 /* Channel width */
3156 WMI_10X_VDEV_PARAM_CHWIDTH,
3157 /* Channel Offset */
3158 WMI_10X_VDEV_PARAM_CHEXTOFFSET,
3159 /* Disable HT Protection */
3160 WMI_10X_VDEV_PARAM_DISABLE_HTPROTECTION,
3161 /* Quick STA Kickout */
3162 WMI_10X_VDEV_PARAM_STA_QUICKKICKOUT,
3163 /* Rate to be used with Management frames */
3164 WMI_10X_VDEV_PARAM_MGMT_RATE,
3165 /* Protection Mode */
3166 WMI_10X_VDEV_PARAM_PROTECTION_MODE,
3167 /* Fixed rate setting */
3168 WMI_10X_VDEV_PARAM_FIXED_RATE,
3169 /* Short GI Enable/Disable */
3170 WMI_10X_VDEV_PARAM_SGI,
3171 /* Enable LDPC */
3172 WMI_10X_VDEV_PARAM_LDPC,
3173 /* Enable Tx STBC */
3174 WMI_10X_VDEV_PARAM_TX_STBC,
3175 /* Enable Rx STBC */
3176 WMI_10X_VDEV_PARAM_RX_STBC,
3177 /* Intra BSS forwarding */
3178 WMI_10X_VDEV_PARAM_INTRA_BSS_FWD,
3179 /* Setting Default xmit key for Vdev */
3180 WMI_10X_VDEV_PARAM_DEF_KEYID,
3181 /* NSS width */
3182 WMI_10X_VDEV_PARAM_NSS,
3183 /* Set the custom rate for the broadcast data frames */
3184 WMI_10X_VDEV_PARAM_BCAST_DATA_RATE,
3185 /* Set the custom rate (rate-code) for multicast data frames */
3186 WMI_10X_VDEV_PARAM_MCAST_DATA_RATE,
3187 /* Tx multicast packet indicate Enable/Disable */
3188 WMI_10X_VDEV_PARAM_MCAST_INDICATE,
3189 /* Tx DHCP packet indicate Enable/Disable */
3190 WMI_10X_VDEV_PARAM_DHCP_INDICATE,
3191 /* Enable host inspection of Tx unicast packet to unknown destination */
3192 WMI_10X_VDEV_PARAM_UNKNOWN_DEST_INDICATE,
3193
3194 /* The minimum amount of time AP begins to consider STA inactive */
3195 WMI_10X_VDEV_PARAM_AP_KEEPALIVE_MIN_IDLE_INACTIVE_TIME_SECS,
3196
3197 /*
3198 * An associated STA is considered inactive when there is no recent
3199 * TX/RX activity and no downlink frames are buffered for it. Once a
3200 * STA exceeds the maximum idle inactive time, the AP will send an
3201 * 802.11 data-null as a keep alive to verify the STA is still
3202 * associated. If the STA does ACK the data-null, or if the data-null
3203 * is buffered and the STA does not retrieve it, the STA will be
3204 * considered unresponsive
3205 * (see WMI_10X_VDEV_AP_KEEPALIVE_MAX_UNRESPONSIVE_TIME_SECS).
3206 */
3207 WMI_10X_VDEV_PARAM_AP_KEEPALIVE_MAX_IDLE_INACTIVE_TIME_SECS,
3208
3209 /*
3210 * An associated STA is considered unresponsive if there is no recent
3211 * TX/RX activity and downlink frames are buffered for it. Once a STA
3212 * exceeds the maximum unresponsive time, the AP will send a
3213 * WMI_10X_STA_KICKOUT event to the host so the STA can be deleted. */
3214 WMI_10X_VDEV_PARAM_AP_KEEPALIVE_MAX_UNRESPONSIVE_TIME_SECS,
3215
3216 /* Enable NAWDS : MCAST INSPECT Enable, NAWDS Flag set */
3217 WMI_10X_VDEV_PARAM_AP_ENABLE_NAWDS,
3218
3219 WMI_10X_VDEV_PARAM_MCAST2UCAST_SET,
3220 /* Enable/Disable RTS-CTS */
3221 WMI_10X_VDEV_PARAM_ENABLE_RTSCTS,
3222
3223 WMI_10X_VDEV_PARAM_AP_DETECT_OUT_OF_SYNC_SLEEPING_STA_TIME_SECS,
3224};
3225
2222/* slot time long */ 3226/* slot time long */
2223#define WMI_VDEV_SLOT_TIME_LONG 0x1 3227#define WMI_VDEV_SLOT_TIME_LONG 0x1
2224/* slot time short */ 3228/* slot time short */
@@ -3000,7 +4004,6 @@ struct wmi_force_fw_hang_cmd {
3000 4004
3001#define WMI_MAX_EVENT 0x1000 4005#define WMI_MAX_EVENT 0x1000
3002/* Maximum number of pending TXed WMI packets */ 4006/* Maximum number of pending TXed WMI packets */
3003#define WMI_MAX_PENDING_TX_COUNT 128
3004#define WMI_SKB_HEADROOM sizeof(struct wmi_cmd_hdr) 4007#define WMI_SKB_HEADROOM sizeof(struct wmi_cmd_hdr)
3005 4008
3006/* By default disable power save for IBSS */ 4009/* By default disable power save for IBSS */
@@ -3013,7 +4016,6 @@ int ath10k_wmi_attach(struct ath10k *ar);
3013void ath10k_wmi_detach(struct ath10k *ar); 4016void ath10k_wmi_detach(struct ath10k *ar);
3014int ath10k_wmi_wait_for_service_ready(struct ath10k *ar); 4017int ath10k_wmi_wait_for_service_ready(struct ath10k *ar);
3015int ath10k_wmi_wait_for_unified_ready(struct ath10k *ar); 4018int ath10k_wmi_wait_for_unified_ready(struct ath10k *ar);
3016void ath10k_wmi_flush_tx(struct ath10k *ar);
3017 4019
3018int ath10k_wmi_connect_htc_service(struct ath10k *ar); 4020int ath10k_wmi_connect_htc_service(struct ath10k *ar);
3019int ath10k_wmi_pdev_set_channel(struct ath10k *ar, 4021int ath10k_wmi_pdev_set_channel(struct ath10k *ar,
@@ -3022,8 +4024,7 @@ int ath10k_wmi_pdev_suspend_target(struct ath10k *ar);
3022int ath10k_wmi_pdev_resume_target(struct ath10k *ar); 4024int ath10k_wmi_pdev_resume_target(struct ath10k *ar);
3023int ath10k_wmi_pdev_set_regdomain(struct ath10k *ar, u16 rd, u16 rd2g, 4025int ath10k_wmi_pdev_set_regdomain(struct ath10k *ar, u16 rd, u16 rd2g,
3024 u16 rd5g, u16 ctl2g, u16 ctl5g); 4026 u16 rd5g, u16 ctl2g, u16 ctl5g);
3025int ath10k_wmi_pdev_set_param(struct ath10k *ar, enum wmi_pdev_param id, 4027int ath10k_wmi_pdev_set_param(struct ath10k *ar, u32 id, u32 value);
3026 u32 value);
3027int ath10k_wmi_cmd_init(struct ath10k *ar); 4028int ath10k_wmi_cmd_init(struct ath10k *ar);
3028int ath10k_wmi_start_scan(struct ath10k *ar, const struct wmi_start_scan_arg *); 4029int ath10k_wmi_start_scan(struct ath10k *ar, const struct wmi_start_scan_arg *);
3029void ath10k_wmi_start_scan_init(struct ath10k *ar, struct wmi_start_scan_arg *); 4030void ath10k_wmi_start_scan_init(struct ath10k *ar, struct wmi_start_scan_arg *);
@@ -3043,7 +4044,7 @@ int ath10k_wmi_vdev_up(struct ath10k *ar, u32 vdev_id, u32 aid,
3043 const u8 *bssid); 4044 const u8 *bssid);
3044int ath10k_wmi_vdev_down(struct ath10k *ar, u32 vdev_id); 4045int ath10k_wmi_vdev_down(struct ath10k *ar, u32 vdev_id);
3045int ath10k_wmi_vdev_set_param(struct ath10k *ar, u32 vdev_id, 4046int ath10k_wmi_vdev_set_param(struct ath10k *ar, u32 vdev_id,
3046 enum wmi_vdev_param param_id, u32 param_value); 4047 u32 param_id, u32 param_value);
3047int ath10k_wmi_vdev_install_key(struct ath10k *ar, 4048int ath10k_wmi_vdev_install_key(struct ath10k *ar,
3048 const struct wmi_vdev_install_key_arg *arg); 4049 const struct wmi_vdev_install_key_arg *arg);
3049int ath10k_wmi_peer_create(struct ath10k *ar, u32 vdev_id, 4050int ath10k_wmi_peer_create(struct ath10k *ar, u32 vdev_id,
@@ -3066,11 +4067,13 @@ int ath10k_wmi_set_ap_ps_param(struct ath10k *ar, u32 vdev_id, const u8 *mac,
3066 enum wmi_ap_ps_peer_param param_id, u32 value); 4067 enum wmi_ap_ps_peer_param param_id, u32 value);
3067int ath10k_wmi_scan_chan_list(struct ath10k *ar, 4068int ath10k_wmi_scan_chan_list(struct ath10k *ar,
3068 const struct wmi_scan_chan_list_arg *arg); 4069 const struct wmi_scan_chan_list_arg *arg);
3069int ath10k_wmi_beacon_send(struct ath10k *ar, const struct wmi_bcn_tx_arg *arg); 4070int ath10k_wmi_beacon_send_nowait(struct ath10k *ar,
4071 const struct wmi_bcn_tx_arg *arg);
3070int ath10k_wmi_pdev_set_wmm_params(struct ath10k *ar, 4072int ath10k_wmi_pdev_set_wmm_params(struct ath10k *ar,
3071 const struct wmi_pdev_set_wmm_params_arg *arg); 4073 const struct wmi_pdev_set_wmm_params_arg *arg);
3072int ath10k_wmi_request_stats(struct ath10k *ar, enum wmi_stats_id stats_id); 4074int ath10k_wmi_request_stats(struct ath10k *ar, enum wmi_stats_id stats_id);
3073int ath10k_wmi_force_fw_hang(struct ath10k *ar, 4075int ath10k_wmi_force_fw_hang(struct ath10k *ar,
3074 enum wmi_force_fw_hang_type type, u32 delay_ms); 4076 enum wmi_force_fw_hang_type type, u32 delay_ms);
4077int ath10k_wmi_mgmt_tx(struct ath10k *ar, struct sk_buff *skb);
3075 4078
3076#endif /* _WMI_H_ */ 4079#endif /* _WMI_H_ */
diff --git a/drivers/net/wireless/ath/ath5k/ahb.c b/drivers/net/wireless/ath/ath5k/ahb.c
index e9bc9e616b69..79bffe165cab 100644
--- a/drivers/net/wireless/ath/ath5k/ahb.c
+++ b/drivers/net/wireless/ath/ath5k/ahb.c
@@ -37,12 +37,9 @@ ath5k_ahb_eeprom_read(struct ath_common *common, u32 off, u16 *data)
37{ 37{
38 struct ath5k_hw *ah = common->priv; 38 struct ath5k_hw *ah = common->priv;
39 struct platform_device *pdev = to_platform_device(ah->dev); 39 struct platform_device *pdev = to_platform_device(ah->dev);
40 struct ar231x_board_config *bcfg = pdev->dev.platform_data; 40 struct ar231x_board_config *bcfg = dev_get_platdata(&pdev->dev);
41 u16 *eeprom, *eeprom_end; 41 u16 *eeprom, *eeprom_end;
42 42
43
44
45 bcfg = pdev->dev.platform_data;
46 eeprom = (u16 *) bcfg->radio; 43 eeprom = (u16 *) bcfg->radio;
47 eeprom_end = ((void *) bcfg->config) + BOARD_CONFIG_BUFSZ; 44 eeprom_end = ((void *) bcfg->config) + BOARD_CONFIG_BUFSZ;
48 45
@@ -57,7 +54,7 @@ ath5k_ahb_eeprom_read(struct ath_common *common, u32 off, u16 *data)
57int ath5k_hw_read_srev(struct ath5k_hw *ah) 54int ath5k_hw_read_srev(struct ath5k_hw *ah)
58{ 55{
59 struct platform_device *pdev = to_platform_device(ah->dev); 56 struct platform_device *pdev = to_platform_device(ah->dev);
60 struct ar231x_board_config *bcfg = pdev->dev.platform_data; 57 struct ar231x_board_config *bcfg = dev_get_platdata(&pdev->dev);
61 ah->ah_mac_srev = bcfg->devid; 58 ah->ah_mac_srev = bcfg->devid;
62 return 0; 59 return 0;
63} 60}
@@ -65,7 +62,7 @@ int ath5k_hw_read_srev(struct ath5k_hw *ah)
65static int ath5k_ahb_eeprom_read_mac(struct ath5k_hw *ah, u8 *mac) 62static int ath5k_ahb_eeprom_read_mac(struct ath5k_hw *ah, u8 *mac)
66{ 63{
67 struct platform_device *pdev = to_platform_device(ah->dev); 64 struct platform_device *pdev = to_platform_device(ah->dev);
68 struct ar231x_board_config *bcfg = pdev->dev.platform_data; 65 struct ar231x_board_config *bcfg = dev_get_platdata(&pdev->dev);
69 u8 *cfg_mac; 66 u8 *cfg_mac;
70 67
71 if (to_platform_device(ah->dev)->id == 0) 68 if (to_platform_device(ah->dev)->id == 0)
@@ -87,7 +84,7 @@ static const struct ath_bus_ops ath_ahb_bus_ops = {
87/*Initialization*/ 84/*Initialization*/
88static int ath_ahb_probe(struct platform_device *pdev) 85static int ath_ahb_probe(struct platform_device *pdev)
89{ 86{
90 struct ar231x_board_config *bcfg = pdev->dev.platform_data; 87 struct ar231x_board_config *bcfg = dev_get_platdata(&pdev->dev);
91 struct ath5k_hw *ah; 88 struct ath5k_hw *ah;
92 struct ieee80211_hw *hw; 89 struct ieee80211_hw *hw;
93 struct resource *res; 90 struct resource *res;
@@ -96,7 +93,7 @@ static int ath_ahb_probe(struct platform_device *pdev)
96 int ret = 0; 93 int ret = 0;
97 u32 reg; 94 u32 reg;
98 95
99 if (!pdev->dev.platform_data) { 96 if (!dev_get_platdata(&pdev->dev)) {
100 dev_err(&pdev->dev, "no platform data specified\n"); 97 dev_err(&pdev->dev, "no platform data specified\n");
101 ret = -EINVAL; 98 ret = -EINVAL;
102 goto err_out; 99 goto err_out;
@@ -193,7 +190,7 @@ static int ath_ahb_probe(struct platform_device *pdev)
193 190
194static int ath_ahb_remove(struct platform_device *pdev) 191static int ath_ahb_remove(struct platform_device *pdev)
195{ 192{
196 struct ar231x_board_config *bcfg = pdev->dev.platform_data; 193 struct ar231x_board_config *bcfg = dev_get_platdata(&pdev->dev);
197 struct ieee80211_hw *hw = platform_get_drvdata(pdev); 194 struct ieee80211_hw *hw = platform_get_drvdata(pdev);
198 struct ath5k_hw *ah; 195 struct ath5k_hw *ah;
199 u32 reg; 196 u32 reg;
diff --git a/drivers/net/wireless/ath/ath5k/base.c b/drivers/net/wireless/ath/ath5k/base.c
index 48161edec8de..69f58b073e85 100644
--- a/drivers/net/wireless/ath/ath5k/base.c
+++ b/drivers/net/wireless/ath/ath5k/base.c
@@ -1663,15 +1663,15 @@ ath5k_tx_frame_completed(struct ath5k_hw *ah, struct sk_buff *skb,
1663 ah->stats.tx_bytes_count += skb->len; 1663 ah->stats.tx_bytes_count += skb->len;
1664 info = IEEE80211_SKB_CB(skb); 1664 info = IEEE80211_SKB_CB(skb);
1665 1665
1666 size = min_t(int, sizeof(info->status.rates), sizeof(bf->rates));
1667 memcpy(info->status.rates, bf->rates, size);
1668
1666 tries[0] = info->status.rates[0].count; 1669 tries[0] = info->status.rates[0].count;
1667 tries[1] = info->status.rates[1].count; 1670 tries[1] = info->status.rates[1].count;
1668 tries[2] = info->status.rates[2].count; 1671 tries[2] = info->status.rates[2].count;
1669 1672
1670 ieee80211_tx_info_clear_status(info); 1673 ieee80211_tx_info_clear_status(info);
1671 1674
1672 size = min_t(int, sizeof(info->status.rates), sizeof(bf->rates));
1673 memcpy(info->status.rates, bf->rates, size);
1674
1675 for (i = 0; i < ts->ts_final_idx; i++) { 1675 for (i = 0; i < ts->ts_final_idx; i++) {
1676 struct ieee80211_tx_rate *r = 1676 struct ieee80211_tx_rate *r =
1677 &info->status.rates[i]; 1677 &info->status.rates[i];
diff --git a/drivers/net/wireless/ath/ath6kl/common.h b/drivers/net/wireless/ath/ath6kl/common.h
index 98a886154d9c..05debf700a84 100644
--- a/drivers/net/wireless/ath/ath6kl/common.h
+++ b/drivers/net/wireless/ath/ath6kl/common.h
@@ -22,8 +22,7 @@
22 22
23#define ATH6KL_MAX_IE 256 23#define ATH6KL_MAX_IE 256
24 24
25extern __printf(2, 3) 25__printf(2, 3) int ath6kl_printk(const char *level, const char *fmt, ...);
26int ath6kl_printk(const char *level, const char *fmt, ...);
27 26
28/* 27/*
29 * Reflects the version of binary interface exposed by ATH6KL target 28 * Reflects the version of binary interface exposed by ATH6KL target
diff --git a/drivers/net/wireless/ath/ath6kl/debug.h b/drivers/net/wireless/ath/ath6kl/debug.h
index 74369de00fb5..ca9ba005f287 100644
--- a/drivers/net/wireless/ath/ath6kl/debug.h
+++ b/drivers/net/wireless/ath/ath6kl/debug.h
@@ -50,11 +50,10 @@ enum ATH6K_DEBUG_MASK {
50}; 50};
51 51
52extern unsigned int debug_mask; 52extern unsigned int debug_mask;
53extern __printf(2, 3) 53__printf(2, 3) int ath6kl_printk(const char *level, const char *fmt, ...);
54int ath6kl_printk(const char *level, const char *fmt, ...); 54__printf(1, 2) int ath6kl_info(const char *fmt, ...);
55extern __printf(1, 2) int ath6kl_info(const char *fmt, ...); 55__printf(1, 2) int ath6kl_err(const char *fmt, ...);
56extern __printf(1, 2) int ath6kl_err(const char *fmt, ...); 56__printf(1, 2) int ath6kl_warn(const char *fmt, ...);
57extern __printf(1, 2) int ath6kl_warn(const char *fmt, ...);
58 57
59enum ath6kl_war { 58enum ath6kl_war {
60 ATH6KL_WAR_INVALID_RATE, 59 ATH6KL_WAR_INVALID_RATE,
diff --git a/drivers/net/wireless/ath/ath6kl/htc.h b/drivers/net/wireless/ath/ath6kl/htc.h
index a2c8ff809793..14cab1403dd6 100644
--- a/drivers/net/wireless/ath/ath6kl/htc.h
+++ b/drivers/net/wireless/ath/ath6kl/htc.h
@@ -60,7 +60,7 @@
60/* disable credit flow control on a specific service */ 60/* disable credit flow control on a specific service */
61#define HTC_CONN_FLGS_DISABLE_CRED_FLOW_CTRL (1 << 3) 61#define HTC_CONN_FLGS_DISABLE_CRED_FLOW_CTRL (1 << 3)
62#define HTC_CONN_FLGS_SET_RECV_ALLOC_SHIFT 8 62#define HTC_CONN_FLGS_SET_RECV_ALLOC_SHIFT 8
63#define HTC_CONN_FLGS_SET_RECV_ALLOC_MASK 0xFF00 63#define HTC_CONN_FLGS_SET_RECV_ALLOC_MASK 0xFF00U
64 64
65/* connect response status codes */ 65/* connect response status codes */
66#define HTC_SERVICE_SUCCESS 0 66#define HTC_SERVICE_SUCCESS 0
diff --git a/drivers/net/wireless/ath/ath9k/Kconfig b/drivers/net/wireless/ath/ath9k/Kconfig
index 7944c25c9a43..32f139e2e897 100644
--- a/drivers/net/wireless/ath/ath9k/Kconfig
+++ b/drivers/net/wireless/ath/ath9k/Kconfig
@@ -84,6 +84,26 @@ config ATH9K_DFS_CERTIFIED
84 developed. At this point enabling this option won't do anything 84 developed. At this point enabling this option won't do anything
85 except increase code size. 85 except increase code size.
86 86
87config ATH9K_TX99
88 bool "Atheros ath9k TX99 testing support"
89 depends on CFG80211_CERTIFICATION_ONUS
90 default n
91 ---help---
92 Say N. This should only be enabled on systems undergoing
93 certification testing and evaluation in a controlled environment.
94 Enabling this will only enable TX99 support, all other modes of
95 operation will be disabled.
96
97 TX99 support enables Specific Absorption Rate (SAR) testing.
98 SAR is the unit of measurement for the amount of radio frequency(RF)
99 absorbed by the body when using a wireless device. The RF exposure
100 limits used are expressed in the terms of SAR, which is a measure
101 of the electric and magnetic field strength and power density for
102 transmitters operating at frequencies from 300 kHz to 100 GHz.
103 Regulatory bodies around the world require that wireless device
104 be evaluated to meet the RF exposure limits set forth in the
105 governmental SAR regulations.
106
87config ATH9K_LEGACY_RATE_CONTROL 107config ATH9K_LEGACY_RATE_CONTROL
88 bool "Atheros ath9k rate control" 108 bool "Atheros ath9k rate control"
89 depends on ATH9K 109 depends on ATH9K
diff --git a/drivers/net/wireless/ath/ath9k/Makefile b/drivers/net/wireless/ath/ath9k/Makefile
index 75ee9e7704ce..6205ef5a9321 100644
--- a/drivers/net/wireless/ath/ath9k/Makefile
+++ b/drivers/net/wireless/ath/ath9k/Makefile
@@ -14,9 +14,7 @@ ath9k-$(CONFIG_ATH9K_AHB) += ahb.o
14ath9k-$(CONFIG_ATH9K_DEBUGFS) += debug.o 14ath9k-$(CONFIG_ATH9K_DEBUGFS) += debug.o
15ath9k-$(CONFIG_ATH9K_DFS_DEBUGFS) += dfs_debug.o 15ath9k-$(CONFIG_ATH9K_DFS_DEBUGFS) += dfs_debug.o
16ath9k-$(CONFIG_ATH9K_DFS_CERTIFIED) += \ 16ath9k-$(CONFIG_ATH9K_DFS_CERTIFIED) += \
17 dfs.o \ 17 dfs.o
18 dfs_pattern_detector.o \
19 dfs_pri_detector.o
20ath9k-$(CONFIG_PM_SLEEP) += wow.o 18ath9k-$(CONFIG_PM_SLEEP) += wow.o
21 19
22obj-$(CONFIG_ATH9K) += ath9k.o 20obj-$(CONFIG_ATH9K) += ath9k.o
diff --git a/drivers/net/wireless/ath/ath9k/ahb.c b/drivers/net/wireless/ath/ath9k/ahb.c
index 072e4b531067..2dff2765769b 100644
--- a/drivers/net/wireless/ath/ath9k/ahb.c
+++ b/drivers/net/wireless/ath/ath9k/ahb.c
@@ -54,7 +54,7 @@ static bool ath_ahb_eeprom_read(struct ath_common *common, u32 off, u16 *data)
54 struct platform_device *pdev = to_platform_device(sc->dev); 54 struct platform_device *pdev = to_platform_device(sc->dev);
55 struct ath9k_platform_data *pdata; 55 struct ath9k_platform_data *pdata;
56 56
57 pdata = (struct ath9k_platform_data *) pdev->dev.platform_data; 57 pdata = dev_get_platdata(&pdev->dev);
58 if (off >= (ARRAY_SIZE(pdata->eeprom_data))) { 58 if (off >= (ARRAY_SIZE(pdata->eeprom_data))) {
59 ath_err(common, 59 ath_err(common,
60 "%s: flash read failed, offset %08x is out of range\n", 60 "%s: flash read failed, offset %08x is out of range\n",
@@ -84,7 +84,7 @@ static int ath_ahb_probe(struct platform_device *pdev)
84 struct ath_hw *ah; 84 struct ath_hw *ah;
85 char hw_name[64]; 85 char hw_name[64];
86 86
87 if (!pdev->dev.platform_data) { 87 if (!dev_get_platdata(&pdev->dev)) {
88 dev_err(&pdev->dev, "no platform data specified\n"); 88 dev_err(&pdev->dev, "no platform data specified\n");
89 return -EINVAL; 89 return -EINVAL;
90 } 90 }
diff --git a/drivers/net/wireless/ath/ath9k/ani.c b/drivers/net/wireless/ath/ath9k/ani.c
index be466b0ef7a7..d28923b7435b 100644
--- a/drivers/net/wireless/ath/ath9k/ani.c
+++ b/drivers/net/wireless/ath/ath9k/ani.c
@@ -338,10 +338,9 @@ void ath9k_ani_reset(struct ath_hw *ah, bool is_scanning)
338 aniState->cckNoiseImmunityLevel != 338 aniState->cckNoiseImmunityLevel !=
339 ATH9K_ANI_CCK_DEF_LEVEL) { 339 ATH9K_ANI_CCK_DEF_LEVEL) {
340 ath_dbg(common, ANI, 340 ath_dbg(common, ANI,
341 "Restore defaults: opmode %u chan %d Mhz/0x%x is_scanning=%d ofdm:%d cck:%d\n", 341 "Restore defaults: opmode %u chan %d Mhz is_scanning=%d ofdm:%d cck:%d\n",
342 ah->opmode, 342 ah->opmode,
343 chan->channel, 343 chan->channel,
344 chan->channelFlags,
345 is_scanning, 344 is_scanning,
346 aniState->ofdmNoiseImmunityLevel, 345 aniState->ofdmNoiseImmunityLevel,
347 aniState->cckNoiseImmunityLevel); 346 aniState->cckNoiseImmunityLevel);
@@ -354,10 +353,9 @@ void ath9k_ani_reset(struct ath_hw *ah, bool is_scanning)
354 * restore historical levels for this channel 353 * restore historical levels for this channel
355 */ 354 */
356 ath_dbg(common, ANI, 355 ath_dbg(common, ANI,
357 "Restore history: opmode %u chan %d Mhz/0x%x is_scanning=%d ofdm:%d cck:%d\n", 356 "Restore history: opmode %u chan %d Mhz is_scanning=%d ofdm:%d cck:%d\n",
358 ah->opmode, 357 ah->opmode,
359 chan->channel, 358 chan->channel,
360 chan->channelFlags,
361 is_scanning, 359 is_scanning,
362 aniState->ofdmNoiseImmunityLevel, 360 aniState->ofdmNoiseImmunityLevel,
363 aniState->cckNoiseImmunityLevel); 361 aniState->cckNoiseImmunityLevel);
diff --git a/drivers/net/wireless/ath/ath9k/antenna.c b/drivers/net/wireless/ath/ath9k/antenna.c
index dd1cc73d7946..bd048cc69a33 100644
--- a/drivers/net/wireless/ath/ath9k/antenna.c
+++ b/drivers/net/wireless/ath/ath9k/antenna.c
@@ -332,7 +332,7 @@ static void ath_select_ant_div_from_quick_scan(struct ath_ant_comb *antcomb,
332 } 332 }
333 333
334 if (antcomb->rssi_lna2 > antcomb->rssi_lna1 + 334 if (antcomb->rssi_lna2 > antcomb->rssi_lna1 +
335 ATH_ANT_DIV_COMB_LNA1_LNA2_SWITCH_DELTA) 335 div_ant_conf->lna1_lna2_switch_delta)
336 div_ant_conf->main_lna_conf = ATH_ANT_DIV_COMB_LNA2; 336 div_ant_conf->main_lna_conf = ATH_ANT_DIV_COMB_LNA2;
337 else 337 else
338 div_ant_conf->main_lna_conf = ATH_ANT_DIV_COMB_LNA1; 338 div_ant_conf->main_lna_conf = ATH_ANT_DIV_COMB_LNA1;
@@ -554,42 +554,22 @@ static void ath_ant_div_conf_fast_divbias(struct ath_hw_antcomb_conf *ant_conf,
554 ant_conf->fast_div_bias = 0x1; 554 ant_conf->fast_div_bias = 0x1;
555 break; 555 break;
556 case 0x10: /* LNA2 A-B */ 556 case 0x10: /* LNA2 A-B */
557 if ((antcomb->scan == 0) && 557 ant_conf->fast_div_bias = 0x2;
558 (alt_ratio > ATH_ANT_DIV_COMB_ALT_ANT_RATIO)) {
559 ant_conf->fast_div_bias = 0x3f;
560 } else {
561 ant_conf->fast_div_bias = 0x1;
562 }
563 break; 558 break;
564 case 0x12: /* LNA2 LNA1 */ 559 case 0x12: /* LNA2 LNA1 */
565 ant_conf->fast_div_bias = 0x39; 560 ant_conf->fast_div_bias = 0x3f;
566 break; 561 break;
567 case 0x13: /* LNA2 A+B */ 562 case 0x13: /* LNA2 A+B */
568 if ((antcomb->scan == 0) && 563 ant_conf->fast_div_bias = 0x2;
569 (alt_ratio > ATH_ANT_DIV_COMB_ALT_ANT_RATIO)) {
570 ant_conf->fast_div_bias = 0x3f;
571 } else {
572 ant_conf->fast_div_bias = 0x1;
573 }
574 break; 564 break;
575 case 0x20: /* LNA1 A-B */ 565 case 0x20: /* LNA1 A-B */
576 if ((antcomb->scan == 0) && 566 ant_conf->fast_div_bias = 0x3;
577 (alt_ratio > ATH_ANT_DIV_COMB_ALT_ANT_RATIO)) {
578 ant_conf->fast_div_bias = 0x3f;
579 } else {
580 ant_conf->fast_div_bias = 0x4;
581 }
582 break; 567 break;
583 case 0x21: /* LNA1 LNA2 */ 568 case 0x21: /* LNA1 LNA2 */
584 ant_conf->fast_div_bias = 0x6; 569 ant_conf->fast_div_bias = 0x3;
585 break; 570 break;
586 case 0x23: /* LNA1 A+B */ 571 case 0x23: /* LNA1 A+B */
587 if ((antcomb->scan == 0) && 572 ant_conf->fast_div_bias = 0x3;
588 (alt_ratio > ATH_ANT_DIV_COMB_ALT_ANT_RATIO)) {
589 ant_conf->fast_div_bias = 0x3f;
590 } else {
591 ant_conf->fast_div_bias = 0x6;
592 }
593 break; 573 break;
594 case 0x30: /* A+B A-B */ 574 case 0x30: /* A+B A-B */
595 ant_conf->fast_div_bias = 0x1; 575 ant_conf->fast_div_bias = 0x1;
@@ -638,7 +618,7 @@ static void ath_ant_try_scan(struct ath_ant_comb *antcomb,
638 antcomb->rssi_sub = alt_rssi_avg; 618 antcomb->rssi_sub = alt_rssi_avg;
639 antcomb->scan = false; 619 antcomb->scan = false;
640 if (antcomb->rssi_lna2 > 620 if (antcomb->rssi_lna2 >
641 (antcomb->rssi_lna1 + ATH_ANT_DIV_COMB_LNA1_LNA2_SWITCH_DELTA)) { 621 (antcomb->rssi_lna1 + conf->lna1_lna2_switch_delta)) {
642 /* use LNA2 as main LNA */ 622 /* use LNA2 as main LNA */
643 if ((antcomb->rssi_add > antcomb->rssi_lna1) && 623 if ((antcomb->rssi_add > antcomb->rssi_lna1) &&
644 (antcomb->rssi_add > antcomb->rssi_sub)) { 624 (antcomb->rssi_add > antcomb->rssi_sub)) {
diff --git a/drivers/net/wireless/ath/ath9k/ar5008_phy.c b/drivers/net/wireless/ath/ath9k/ar5008_phy.c
index 08656473c63e..ff415e863ee9 100644
--- a/drivers/net/wireless/ath/ath9k/ar5008_phy.c
+++ b/drivers/net/wireless/ath/ath9k/ar5008_phy.c
@@ -626,12 +626,11 @@ static void ar5008_hw_override_ini(struct ath_hw *ah,
626 if (AR_SREV_9287_11_OR_LATER(ah)) 626 if (AR_SREV_9287_11_OR_LATER(ah))
627 val = val & (~AR_PCU_MISC_MODE2_HWWAR2); 627 val = val & (~AR_PCU_MISC_MODE2_HWWAR2);
628 628
629 val |= AR_PCU_MISC_MODE2_CFP_IGNORE;
630
629 REG_WRITE(ah, AR_PCU_MISC_MODE2, val); 631 REG_WRITE(ah, AR_PCU_MISC_MODE2, val);
630 } 632 }
631 633
632 REG_SET_BIT(ah, AR_PHY_CCK_DETECT,
633 AR_PHY_CCK_DETECT_BB_ENABLE_ANT_FAST_DIV);
634
635 if (AR_SREV_9280_20_OR_LATER(ah)) 634 if (AR_SREV_9280_20_OR_LATER(ah))
636 return; 635 return;
637 /* 636 /*
@@ -667,14 +666,13 @@ static void ar5008_hw_set_channel_regs(struct ath_hw *ah,
667 if (IS_CHAN_HT40(chan)) { 666 if (IS_CHAN_HT40(chan)) {
668 phymode |= AR_PHY_FC_DYN2040_EN; 667 phymode |= AR_PHY_FC_DYN2040_EN;
669 668
670 if ((chan->chanmode == CHANNEL_A_HT40PLUS) || 669 if (IS_CHAN_HT40PLUS(chan))
671 (chan->chanmode == CHANNEL_G_HT40PLUS))
672 phymode |= AR_PHY_FC_DYN2040_PRI_CH; 670 phymode |= AR_PHY_FC_DYN2040_PRI_CH;
673 671
674 } 672 }
675 REG_WRITE(ah, AR_PHY_TURBO, phymode); 673 REG_WRITE(ah, AR_PHY_TURBO, phymode);
676 674
677 ath9k_hw_set11nmac2040(ah); 675 ath9k_hw_set11nmac2040(ah, chan);
678 676
679 ENABLE_REGWRITE_BUFFER(ah); 677 ENABLE_REGWRITE_BUFFER(ah);
680 678
@@ -692,31 +690,12 @@ static int ar5008_hw_process_ini(struct ath_hw *ah,
692 int i, regWrites = 0; 690 int i, regWrites = 0;
693 u32 modesIndex, freqIndex; 691 u32 modesIndex, freqIndex;
694 692
695 switch (chan->chanmode) { 693 if (IS_CHAN_5GHZ(chan)) {
696 case CHANNEL_A:
697 case CHANNEL_A_HT20:
698 modesIndex = 1;
699 freqIndex = 1;
700 break;
701 case CHANNEL_A_HT40PLUS:
702 case CHANNEL_A_HT40MINUS:
703 modesIndex = 2;
704 freqIndex = 1; 694 freqIndex = 1;
705 break; 695 modesIndex = IS_CHAN_HT40(chan) ? 2 : 1;
706 case CHANNEL_G: 696 } else {
707 case CHANNEL_G_HT20:
708 case CHANNEL_B:
709 modesIndex = 4;
710 freqIndex = 2;
711 break;
712 case CHANNEL_G_HT40PLUS:
713 case CHANNEL_G_HT40MINUS:
714 modesIndex = 3;
715 freqIndex = 2; 697 freqIndex = 2;
716 break; 698 modesIndex = IS_CHAN_HT40(chan) ? 3 : 4;
717
718 default:
719 return -EINVAL;
720 } 699 }
721 700
722 /* 701 /*
@@ -815,8 +794,10 @@ static void ar5008_hw_set_rfmode(struct ath_hw *ah, struct ath9k_channel *chan)
815 if (chan == NULL) 794 if (chan == NULL)
816 return; 795 return;
817 796
818 rfMode |= (IS_CHAN_B(chan) || IS_CHAN_G(chan)) 797 if (IS_CHAN_2GHZ(chan))
819 ? AR_PHY_MODE_DYNAMIC : AR_PHY_MODE_OFDM; 798 rfMode |= AR_PHY_MODE_DYNAMIC;
799 else
800 rfMode |= AR_PHY_MODE_OFDM;
820 801
821 if (!AR_SREV_9280_20_OR_LATER(ah)) 802 if (!AR_SREV_9280_20_OR_LATER(ah))
822 rfMode |= (IS_CHAN_5GHZ(chan)) ? 803 rfMode |= (IS_CHAN_5GHZ(chan)) ?
@@ -1219,12 +1200,11 @@ static void ar5008_hw_ani_cache_ini_regs(struct ath_hw *ah)
1219 1200
1220 iniDef = &aniState->iniDef; 1201 iniDef = &aniState->iniDef;
1221 1202
1222 ath_dbg(common, ANI, "ver %d.%d opmode %u chan %d Mhz/0x%x\n", 1203 ath_dbg(common, ANI, "ver %d.%d opmode %u chan %d Mhz\n",
1223 ah->hw_version.macVersion, 1204 ah->hw_version.macVersion,
1224 ah->hw_version.macRev, 1205 ah->hw_version.macRev,
1225 ah->opmode, 1206 ah->opmode,
1226 chan->channel, 1207 chan->channel);
1227 chan->channelFlags);
1228 1208
1229 val = REG_READ(ah, AR_PHY_SFCORR); 1209 val = REG_READ(ah, AR_PHY_SFCORR);
1230 iniDef->m1Thresh = MS(val, AR_PHY_SFCORR_M1_THRESH); 1210 iniDef->m1Thresh = MS(val, AR_PHY_SFCORR_M1_THRESH);
diff --git a/drivers/net/wireless/ath/ath9k/ar9002_calib.c b/drivers/net/wireless/ath/ath9k/ar9002_calib.c
index 9f589744a9f9..cdc74005650c 100644
--- a/drivers/net/wireless/ath/ath9k/ar9002_calib.c
+++ b/drivers/net/wireless/ath/ath9k/ar9002_calib.c
@@ -33,15 +33,12 @@ static bool ar9002_hw_is_cal_supported(struct ath_hw *ah,
33 bool supported = false; 33 bool supported = false;
34 switch (ah->supp_cals & cal_type) { 34 switch (ah->supp_cals & cal_type) {
35 case IQ_MISMATCH_CAL: 35 case IQ_MISMATCH_CAL:
36 /* Run IQ Mismatch for non-CCK only */ 36 supported = true;
37 if (!IS_CHAN_B(chan))
38 supported = true;
39 break; 37 break;
40 case ADC_GAIN_CAL: 38 case ADC_GAIN_CAL:
41 case ADC_DC_CAL: 39 case ADC_DC_CAL:
42 /* Run ADC Gain Cal for non-CCK & non 2GHz-HT20 only */ 40 /* Run ADC Gain Cal for non-CCK & non 2GHz-HT20 only */
43 if (!IS_CHAN_B(chan) && 41 if (!((IS_CHAN_2GHZ(chan) || IS_CHAN_A_FAST_CLOCK(ah, chan)) &&
44 !((IS_CHAN_2GHZ(chan) || IS_CHAN_A_FAST_CLOCK(ah, chan)) &&
45 IS_CHAN_HT20(chan))) 42 IS_CHAN_HT20(chan)))
46 supported = true; 43 supported = true;
47 break; 44 break;
@@ -671,7 +668,7 @@ static bool ar9002_hw_calibrate(struct ath_hw *ah,
671 668
672 nfcal = !!(REG_READ(ah, AR_PHY_AGC_CONTROL) & AR_PHY_AGC_CONTROL_NF); 669 nfcal = !!(REG_READ(ah, AR_PHY_AGC_CONTROL) & AR_PHY_AGC_CONTROL_NF);
673 if (ah->caldata) 670 if (ah->caldata)
674 nfcal_pending = ah->caldata->nfcal_pending; 671 nfcal_pending = test_bit(NFCAL_PENDING, &ah->caldata->cal_flags);
675 672
676 if (currCal && !nfcal && 673 if (currCal && !nfcal &&
677 (currCal->calState == CAL_RUNNING || 674 (currCal->calState == CAL_RUNNING ||
@@ -861,7 +858,7 @@ static bool ar9002_hw_init_cal(struct ath_hw *ah, struct ath9k_channel *chan)
861 ar9002_hw_pa_cal(ah, true); 858 ar9002_hw_pa_cal(ah, true);
862 859
863 if (ah->caldata) 860 if (ah->caldata)
864 ah->caldata->nfcal_pending = true; 861 set_bit(NFCAL_PENDING, &ah->caldata->cal_flags);
865 862
866 ah->cal_list = ah->cal_list_last = ah->cal_list_curr = NULL; 863 ah->cal_list = ah->cal_list_last = ah->cal_list_curr = NULL;
867 864
diff --git a/drivers/net/wireless/ath/ath9k/ar9002_hw.c b/drivers/net/wireless/ath/ath9k/ar9002_hw.c
index fb61b081d172..5c95fd9e9c9e 100644
--- a/drivers/net/wireless/ath/ath9k/ar9002_hw.c
+++ b/drivers/net/wireless/ath/ath9k/ar9002_hw.c
@@ -419,28 +419,10 @@ void ar9002_hw_load_ani_reg(struct ath_hw *ah, struct ath9k_channel *chan)
419 u32 modesIndex; 419 u32 modesIndex;
420 int i; 420 int i;
421 421
422 switch (chan->chanmode) { 422 if (IS_CHAN_5GHZ(chan))
423 case CHANNEL_A: 423 modesIndex = IS_CHAN_HT40(chan) ? 2 : 1;
424 case CHANNEL_A_HT20: 424 else
425 modesIndex = 1; 425 modesIndex = IS_CHAN_HT40(chan) ? 3 : 4;
426 break;
427 case CHANNEL_A_HT40PLUS:
428 case CHANNEL_A_HT40MINUS:
429 modesIndex = 2;
430 break;
431 case CHANNEL_G:
432 case CHANNEL_G_HT20:
433 case CHANNEL_B:
434 modesIndex = 4;
435 break;
436 case CHANNEL_G_HT40PLUS:
437 case CHANNEL_G_HT40MINUS:
438 modesIndex = 3;
439 break;
440
441 default:
442 return;
443 }
444 426
445 ENABLE_REGWRITE_BUFFER(ah); 427 ENABLE_REGWRITE_BUFFER(ah);
446 428
diff --git a/drivers/net/wireless/ath/ath9k/ar9002_phy.c b/drivers/net/wireless/ath/ath9k/ar9002_phy.c
index 1fc1fa955d44..f087117b2e6b 100644
--- a/drivers/net/wireless/ath/ath9k/ar9002_phy.c
+++ b/drivers/net/wireless/ath/ath9k/ar9002_phy.c
@@ -485,7 +485,7 @@ static void ar9002_hw_do_getnf(struct ath_hw *ah,
485 if (IS_CHAN_HT40(ah->curchan)) 485 if (IS_CHAN_HT40(ah->curchan))
486 nfarray[3] = sign_extend32(nf, 8); 486 nfarray[3] = sign_extend32(nf, 8);
487 487
488 if (AR_SREV_9285(ah) || AR_SREV_9271(ah)) 488 if (!(ah->rxchainmask & BIT(1)))
489 return; 489 return;
490 490
491 nf = MS(REG_READ(ah, AR_PHY_CH1_CCA), AR9280_PHY_CH1_MINCCA_PWR); 491 nf = MS(REG_READ(ah, AR_PHY_CH1_CCA), AR9280_PHY_CH1_MINCCA_PWR);
@@ -532,6 +532,7 @@ static void ar9002_hw_antdiv_comb_conf_get(struct ath_hw *ah,
532 AR_PHY_9285_ANT_DIV_ALT_LNACONF_S; 532 AR_PHY_9285_ANT_DIV_ALT_LNACONF_S;
533 antconf->fast_div_bias = (regval & AR_PHY_9285_FAST_DIV_BIAS) >> 533 antconf->fast_div_bias = (regval & AR_PHY_9285_FAST_DIV_BIAS) >>
534 AR_PHY_9285_FAST_DIV_BIAS_S; 534 AR_PHY_9285_FAST_DIV_BIAS_S;
535 antconf->lna1_lna2_switch_delta = -1;
535 antconf->lna1_lna2_delta = -3; 536 antconf->lna1_lna2_delta = -3;
536 antconf->div_group = 0; 537 antconf->div_group = 0;
537} 538}
@@ -679,6 +680,26 @@ static void ar9002_hw_spectral_scan_wait(struct ath_hw *ah)
679 } 680 }
680} 681}
681 682
683static void ar9002_hw_tx99_start(struct ath_hw *ah, u32 qnum)
684{
685 REG_SET_BIT(ah, 0x9864, 0x7f000);
686 REG_SET_BIT(ah, 0x9924, 0x7f00fe);
687 REG_CLR_BIT(ah, AR_DIAG_SW, AR_DIAG_RX_DIS);
688 REG_WRITE(ah, AR_CR, AR_CR_RXD);
689 REG_WRITE(ah, AR_DLCL_IFS(qnum), 0);
690 REG_WRITE(ah, AR_D_GBL_IFS_SIFS, 20);
691 REG_WRITE(ah, AR_D_GBL_IFS_EIFS, 20);
692 REG_WRITE(ah, AR_D_FPCTL, 0x10|qnum);
693 REG_WRITE(ah, AR_TIME_OUT, 0x00000400);
694 REG_WRITE(ah, AR_DRETRY_LIMIT(qnum), 0xffffffff);
695 REG_SET_BIT(ah, AR_QMISC(qnum), AR_Q_MISC_DCU_EARLY_TERM_REQ);
696}
697
698static void ar9002_hw_tx99_stop(struct ath_hw *ah)
699{
700 REG_SET_BIT(ah, AR_DIAG_SW, AR_DIAG_RX_DIS);
701}
702
682void ar9002_hw_attach_phy_ops(struct ath_hw *ah) 703void ar9002_hw_attach_phy_ops(struct ath_hw *ah)
683{ 704{
684 struct ath_hw_private_ops *priv_ops = ath9k_hw_private_ops(ah); 705 struct ath_hw_private_ops *priv_ops = ath9k_hw_private_ops(ah);
@@ -700,6 +721,8 @@ void ar9002_hw_attach_phy_ops(struct ath_hw *ah)
700#ifdef CONFIG_ATH9K_BTCOEX_SUPPORT 721#ifdef CONFIG_ATH9K_BTCOEX_SUPPORT
701 ops->set_bt_ant_diversity = ar9002_hw_set_bt_ant_diversity; 722 ops->set_bt_ant_diversity = ar9002_hw_set_bt_ant_diversity;
702#endif 723#endif
724 ops->tx99_start = ar9002_hw_tx99_start;
725 ops->tx99_stop = ar9002_hw_tx99_stop;
703 726
704 ar9002_hw_set_nf_limits(ah); 727 ar9002_hw_set_nf_limits(ah);
705} 728}
diff --git a/drivers/net/wireless/ath/ath9k/ar9003_calib.c b/drivers/net/wireless/ath/ath9k/ar9003_calib.c
index 6988e1d081f2..22934d3ca544 100644
--- a/drivers/net/wireless/ath/ath9k/ar9003_calib.c
+++ b/drivers/net/wireless/ath/ath9k/ar9003_calib.c
@@ -727,8 +727,12 @@ static void ar9003_hw_tx_iqcal_load_avg_2_passes(struct ath_hw *ah,
727 REG_RMW_FIELD(ah, AR_PHY_RX_IQCAL_CORR_B0, 727 REG_RMW_FIELD(ah, AR_PHY_RX_IQCAL_CORR_B0,
728 AR_PHY_RX_IQCAL_CORR_B0_LOOPBACK_IQCORR_EN, 0x1); 728 AR_PHY_RX_IQCAL_CORR_B0_LOOPBACK_IQCORR_EN, 0x1);
729 729
730 if (caldata) 730 if (caldata) {
731 caldata->done_txiqcal_once = is_reusable; 731 if (is_reusable)
732 set_bit(TXIQCAL_DONE, &caldata->cal_flags);
733 else
734 clear_bit(TXIQCAL_DONE, &caldata->cal_flags);
735 }
732 736
733 return; 737 return;
734} 738}
@@ -961,18 +965,44 @@ static void ar9003_hw_manual_peak_cal(struct ath_hw *ah, u8 chain, bool is_2g)
961} 965}
962 966
963static void ar9003_hw_do_manual_peak_cal(struct ath_hw *ah, 967static void ar9003_hw_do_manual_peak_cal(struct ath_hw *ah,
964 struct ath9k_channel *chan) 968 struct ath9k_channel *chan,
969 bool run_rtt_cal)
965{ 970{
971 struct ath9k_hw_cal_data *caldata = ah->caldata;
966 int i; 972 int i;
967 973
968 if (!AR_SREV_9462(ah) && !AR_SREV_9565(ah) && !AR_SREV_9485(ah)) 974 if (!AR_SREV_9462(ah) && !AR_SREV_9565(ah) && !AR_SREV_9485(ah))
969 return; 975 return;
970 976
977 if ((ah->caps.hw_caps & ATH9K_HW_CAP_RTT) && !run_rtt_cal)
978 return;
979
971 for (i = 0; i < AR9300_MAX_CHAINS; i++) { 980 for (i = 0; i < AR9300_MAX_CHAINS; i++) {
972 if (!(ah->rxchainmask & (1 << i))) 981 if (!(ah->rxchainmask & (1 << i)))
973 continue; 982 continue;
974 ar9003_hw_manual_peak_cal(ah, i, IS_CHAN_2GHZ(chan)); 983 ar9003_hw_manual_peak_cal(ah, i, IS_CHAN_2GHZ(chan));
975 } 984 }
985
986 if (caldata)
987 set_bit(SW_PKDET_DONE, &caldata->cal_flags);
988
989 if ((ah->caps.hw_caps & ATH9K_HW_CAP_RTT) && caldata) {
990 if (IS_CHAN_2GHZ(chan)){
991 caldata->caldac[0] = REG_READ_FIELD(ah,
992 AR_PHY_65NM_RXRF_AGC(0),
993 AR_PHY_65NM_RXRF_AGC_AGC2G_CALDAC_OVR);
994 caldata->caldac[1] = REG_READ_FIELD(ah,
995 AR_PHY_65NM_RXRF_AGC(1),
996 AR_PHY_65NM_RXRF_AGC_AGC2G_CALDAC_OVR);
997 } else {
998 caldata->caldac[0] = REG_READ_FIELD(ah,
999 AR_PHY_65NM_RXRF_AGC(0),
1000 AR_PHY_65NM_RXRF_AGC_AGC5G_CALDAC_OVR);
1001 caldata->caldac[1] = REG_READ_FIELD(ah,
1002 AR_PHY_65NM_RXRF_AGC(1),
1003 AR_PHY_65NM_RXRF_AGC_AGC5G_CALDAC_OVR);
1004 }
1005 }
976} 1006}
977 1007
978static void ar9003_hw_cl_cal_post_proc(struct ath_hw *ah, bool is_reusable) 1008static void ar9003_hw_cl_cal_post_proc(struct ath_hw *ah, bool is_reusable)
@@ -990,7 +1020,7 @@ static void ar9003_hw_cl_cal_post_proc(struct ath_hw *ah, bool is_reusable)
990 txclcal_done = !!(REG_READ(ah, AR_PHY_AGC_CONTROL) & 1020 txclcal_done = !!(REG_READ(ah, AR_PHY_AGC_CONTROL) &
991 AR_PHY_AGC_CONTROL_CLC_SUCCESS); 1021 AR_PHY_AGC_CONTROL_CLC_SUCCESS);
992 1022
993 if (caldata->done_txclcal_once) { 1023 if (test_bit(TXCLCAL_DONE, &caldata->cal_flags)) {
994 for (i = 0; i < AR9300_MAX_CHAINS; i++) { 1024 for (i = 0; i < AR9300_MAX_CHAINS; i++) {
995 if (!(ah->txchainmask & (1 << i))) 1025 if (!(ah->txchainmask & (1 << i)))
996 continue; 1026 continue;
@@ -1006,7 +1036,7 @@ static void ar9003_hw_cl_cal_post_proc(struct ath_hw *ah, bool is_reusable)
1006 caldata->tx_clcal[i][j] = 1036 caldata->tx_clcal[i][j] =
1007 REG_READ(ah, CL_TAB_ENTRY(cl_idx[i])); 1037 REG_READ(ah, CL_TAB_ENTRY(cl_idx[i]));
1008 } 1038 }
1009 caldata->done_txclcal_once = true; 1039 set_bit(TXCLCAL_DONE, &caldata->cal_flags);
1010 } 1040 }
1011} 1041}
1012 1042
@@ -1019,6 +1049,7 @@ static bool ar9003_hw_init_cal(struct ath_hw *ah,
1019 bool is_reusable = true, status = true; 1049 bool is_reusable = true, status = true;
1020 bool run_rtt_cal = false, run_agc_cal, sep_iq_cal = false; 1050 bool run_rtt_cal = false, run_agc_cal, sep_iq_cal = false;
1021 bool rtt = !!(ah->caps.hw_caps & ATH9K_HW_CAP_RTT); 1051 bool rtt = !!(ah->caps.hw_caps & ATH9K_HW_CAP_RTT);
1052 u32 rx_delay = 0;
1022 u32 agc_ctrl = 0, agc_supp_cals = AR_PHY_AGC_CONTROL_OFFSET_CAL | 1053 u32 agc_ctrl = 0, agc_supp_cals = AR_PHY_AGC_CONTROL_OFFSET_CAL |
1023 AR_PHY_AGC_CONTROL_FLTR_CAL | 1054 AR_PHY_AGC_CONTROL_FLTR_CAL |
1024 AR_PHY_AGC_CONTROL_PKDET_CAL; 1055 AR_PHY_AGC_CONTROL_PKDET_CAL;
@@ -1042,17 +1073,22 @@ static bool ar9003_hw_init_cal(struct ath_hw *ah,
1042 ar9003_hw_rtt_clear_hist(ah); 1073 ar9003_hw_rtt_clear_hist(ah);
1043 } 1074 }
1044 1075
1045 if (rtt && !run_rtt_cal) { 1076 if (rtt) {
1046 agc_ctrl = REG_READ(ah, AR_PHY_AGC_CONTROL); 1077 if (!run_rtt_cal) {
1047 agc_supp_cals &= agc_ctrl; 1078 agc_ctrl = REG_READ(ah, AR_PHY_AGC_CONTROL);
1048 agc_ctrl &= ~(AR_PHY_AGC_CONTROL_OFFSET_CAL | 1079 agc_supp_cals &= agc_ctrl;
1049 AR_PHY_AGC_CONTROL_FLTR_CAL | 1080 agc_ctrl &= ~(AR_PHY_AGC_CONTROL_OFFSET_CAL |
1050 AR_PHY_AGC_CONTROL_PKDET_CAL); 1081 AR_PHY_AGC_CONTROL_FLTR_CAL |
1051 REG_WRITE(ah, AR_PHY_AGC_CONTROL, agc_ctrl); 1082 AR_PHY_AGC_CONTROL_PKDET_CAL);
1083 REG_WRITE(ah, AR_PHY_AGC_CONTROL, agc_ctrl);
1084 } else {
1085 if (ah->ah_flags & AH_FASTCC)
1086 run_agc_cal = true;
1087 }
1052 } 1088 }
1053 1089
1054 if (ah->enabled_cals & TX_CL_CAL) { 1090 if (ah->enabled_cals & TX_CL_CAL) {
1055 if (caldata && caldata->done_txclcal_once) 1091 if (caldata && test_bit(TXCLCAL_DONE, &caldata->cal_flags))
1056 REG_CLR_BIT(ah, AR_PHY_CL_CAL_CTL, 1092 REG_CLR_BIT(ah, AR_PHY_CL_CAL_CTL,
1057 AR_PHY_CL_CAL_ENABLE); 1093 AR_PHY_CL_CAL_ENABLE);
1058 else { 1094 else {
@@ -1076,14 +1112,14 @@ static bool ar9003_hw_init_cal(struct ath_hw *ah,
1076 * AGC calibration 1112 * AGC calibration
1077 */ 1113 */
1078 if (ah->enabled_cals & TX_IQ_ON_AGC_CAL) { 1114 if (ah->enabled_cals & TX_IQ_ON_AGC_CAL) {
1079 if (caldata && !caldata->done_txiqcal_once) 1115 if (caldata && !test_bit(TXIQCAL_DONE, &caldata->cal_flags))
1080 REG_SET_BIT(ah, AR_PHY_TX_IQCAL_CONTROL_0, 1116 REG_SET_BIT(ah, AR_PHY_TX_IQCAL_CONTROL_0,
1081 AR_PHY_TX_IQCAL_CONTROL_0_ENABLE_TXIQ_CAL); 1117 AR_PHY_TX_IQCAL_CONTROL_0_ENABLE_TXIQ_CAL);
1082 else 1118 else
1083 REG_CLR_BIT(ah, AR_PHY_TX_IQCAL_CONTROL_0, 1119 REG_CLR_BIT(ah, AR_PHY_TX_IQCAL_CONTROL_0,
1084 AR_PHY_TX_IQCAL_CONTROL_0_ENABLE_TXIQ_CAL); 1120 AR_PHY_TX_IQCAL_CONTROL_0_ENABLE_TXIQ_CAL);
1085 txiqcal_done = run_agc_cal = true; 1121 txiqcal_done = run_agc_cal = true;
1086 } else if (caldata && !caldata->done_txiqcal_once) { 1122 } else if (caldata && !test_bit(TXIQCAL_DONE, &caldata->cal_flags)) {
1087 run_agc_cal = true; 1123 run_agc_cal = true;
1088 sep_iq_cal = true; 1124 sep_iq_cal = true;
1089 } 1125 }
@@ -1099,6 +1135,15 @@ skip_tx_iqcal:
1099 REG_WRITE(ah, AR_PHY_ACTIVE, AR_PHY_ACTIVE_EN); 1135 REG_WRITE(ah, AR_PHY_ACTIVE, AR_PHY_ACTIVE_EN);
1100 } 1136 }
1101 1137
1138 if (REG_READ(ah, AR_PHY_CL_CAL_CTL) & AR_PHY_CL_CAL_ENABLE) {
1139 rx_delay = REG_READ(ah, AR_PHY_RX_DELAY);
1140 /* Disable BB_active */
1141 REG_WRITE(ah, AR_PHY_ACTIVE, AR_PHY_ACTIVE_DIS);
1142 udelay(5);
1143 REG_WRITE(ah, AR_PHY_RX_DELAY, AR_PHY_RX_DELAY_DELAY);
1144 REG_WRITE(ah, AR_PHY_ACTIVE, AR_PHY_ACTIVE_EN);
1145 }
1146
1102 if (run_agc_cal || !(ah->ah_flags & AH_FASTCC)) { 1147 if (run_agc_cal || !(ah->ah_flags & AH_FASTCC)) {
1103 /* Calibrate the AGC */ 1148 /* Calibrate the AGC */
1104 REG_WRITE(ah, AR_PHY_AGC_CONTROL, 1149 REG_WRITE(ah, AR_PHY_AGC_CONTROL,
@@ -1110,7 +1155,12 @@ skip_tx_iqcal:
1110 AR_PHY_AGC_CONTROL_CAL, 1155 AR_PHY_AGC_CONTROL_CAL,
1111 0, AH_WAIT_TIMEOUT); 1156 0, AH_WAIT_TIMEOUT);
1112 1157
1113 ar9003_hw_do_manual_peak_cal(ah, chan); 1158 ar9003_hw_do_manual_peak_cal(ah, chan, run_rtt_cal);
1159 }
1160
1161 if (REG_READ(ah, AR_PHY_CL_CAL_CTL) & AR_PHY_CL_CAL_ENABLE) {
1162 REG_WRITE(ah, AR_PHY_RX_DELAY, rx_delay);
1163 udelay(5);
1114 } 1164 }
1115 1165
1116 if (ath9k_hw_mci_is_enabled(ah) && IS_CHAN_2GHZ(chan) && run_agc_cal) 1166 if (ath9k_hw_mci_is_enabled(ah) && IS_CHAN_2GHZ(chan) && run_agc_cal)
@@ -1133,19 +1183,23 @@ skip_tx_iqcal:
1133 1183
1134 if (txiqcal_done) 1184 if (txiqcal_done)
1135 ar9003_hw_tx_iq_cal_post_proc(ah, is_reusable); 1185 ar9003_hw_tx_iq_cal_post_proc(ah, is_reusable);
1136 else if (caldata && caldata->done_txiqcal_once) 1186 else if (caldata && test_bit(TXIQCAL_DONE, &caldata->cal_flags))
1137 ar9003_hw_tx_iq_cal_reload(ah); 1187 ar9003_hw_tx_iq_cal_reload(ah);
1138 1188
1139 ar9003_hw_cl_cal_post_proc(ah, is_reusable); 1189 ar9003_hw_cl_cal_post_proc(ah, is_reusable);
1140 1190
1141 if (run_rtt_cal && caldata) { 1191 if (run_rtt_cal && caldata) {
1142 if (is_reusable) { 1192 if (is_reusable) {
1143 if (!ath9k_hw_rfbus_req(ah)) 1193 if (!ath9k_hw_rfbus_req(ah)) {
1144 ath_err(ath9k_hw_common(ah), 1194 ath_err(ath9k_hw_common(ah),
1145 "Could not stop baseband\n"); 1195 "Could not stop baseband\n");
1146 else 1196 } else {
1147 ar9003_hw_rtt_fill_hist(ah); 1197 ar9003_hw_rtt_fill_hist(ah);
1148 1198
1199 if (test_bit(SW_PKDET_DONE, &caldata->cal_flags))
1200 ar9003_hw_rtt_load_hist(ah);
1201 }
1202
1149 ath9k_hw_rfbus_done(ah); 1203 ath9k_hw_rfbus_done(ah);
1150 } 1204 }
1151 1205
diff --git a/drivers/net/wireless/ath/ath9k/ar9003_eeprom.c b/drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
index f4864807e15b..1ec52356b5a1 100644
--- a/drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
+++ b/drivers/net/wireless/ath/ath9k/ar9003_eeprom.c
@@ -2991,7 +2991,10 @@ static u32 ath9k_hw_ar9300_get_eeprom(struct ath_hw *ah,
2991 case EEP_CHAIN_MASK_REDUCE: 2991 case EEP_CHAIN_MASK_REDUCE:
2992 return (pBase->miscConfiguration >> 0x3) & 0x1; 2992 return (pBase->miscConfiguration >> 0x3) & 0x1;
2993 case EEP_ANT_DIV_CTL1: 2993 case EEP_ANT_DIV_CTL1:
2994 return eep->base_ext1.ant_div_control; 2994 if (AR_SREV_9565(ah))
2995 return AR9300_EEP_ANTDIV_CONTROL_DEFAULT_VALUE;
2996 else
2997 return eep->base_ext1.ant_div_control;
2995 case EEP_ANTENNA_GAIN_5G: 2998 case EEP_ANTENNA_GAIN_5G:
2996 return eep->modalHeader5G.antennaGain; 2999 return eep->modalHeader5G.antennaGain;
2997 case EEP_ANTENNA_GAIN_2G: 3000 case EEP_ANTENNA_GAIN_2G:
@@ -3424,12 +3427,12 @@ static u32 ath9k_hw_ar9003_dump_eeprom(struct ath_hw *ah, bool dump_base_hdr,
3424 struct ar9300_base_eep_hdr *pBase; 3427 struct ar9300_base_eep_hdr *pBase;
3425 3428
3426 if (!dump_base_hdr) { 3429 if (!dump_base_hdr) {
3427 len += snprintf(buf + len, size - len, 3430 len += scnprintf(buf + len, size - len,
3428 "%20s :\n", "2GHz modal Header"); 3431 "%20s :\n", "2GHz modal Header");
3429 len = ar9003_dump_modal_eeprom(buf, len, size, 3432 len = ar9003_dump_modal_eeprom(buf, len, size,
3430 &eep->modalHeader2G); 3433 &eep->modalHeader2G);
3431 len += snprintf(buf + len, size - len, 3434 len += scnprintf(buf + len, size - len,
3432 "%20s :\n", "5GHz modal Header"); 3435 "%20s :\n", "5GHz modal Header");
3433 len = ar9003_dump_modal_eeprom(buf, len, size, 3436 len = ar9003_dump_modal_eeprom(buf, len, size,
3434 &eep->modalHeader5G); 3437 &eep->modalHeader5G);
3435 goto out; 3438 goto out;
@@ -3479,8 +3482,8 @@ static u32 ath9k_hw_ar9003_dump_eeprom(struct ath_hw *ah, bool dump_base_hdr,
3479 PR_EEP("Rx Gain", pBase->txrxgain & 0xf); 3482 PR_EEP("Rx Gain", pBase->txrxgain & 0xf);
3480 PR_EEP("SW Reg", le32_to_cpu(pBase->swreg)); 3483 PR_EEP("SW Reg", le32_to_cpu(pBase->swreg));
3481 3484
3482 len += snprintf(buf + len, size - len, "%20s : %pM\n", "MacAddress", 3485 len += scnprintf(buf + len, size - len, "%20s : %pM\n", "MacAddress",
3483 ah->eeprom.ar9300_eep.macAddr); 3486 ah->eeprom.ar9300_eep.macAddr);
3484out: 3487out:
3485 if (len > size) 3488 if (len > size)
3486 len = size; 3489 len = size;
@@ -3656,9 +3659,23 @@ static void ar9003_hw_ant_ctrl_apply(struct ath_hw *ah, bool is2ghz)
3656 if (AR_SREV_9565(ah)) { 3659 if (AR_SREV_9565(ah)) {
3657 if (common->bt_ant_diversity) { 3660 if (common->bt_ant_diversity) {
3658 regval |= (1 << AR_PHY_ANT_SW_RX_PROT_S); 3661 regval |= (1 << AR_PHY_ANT_SW_RX_PROT_S);
3662
3663 REG_SET_BIT(ah, AR_PHY_RESTART,
3664 AR_PHY_RESTART_ENABLE_DIV_M2FLAG);
3665
3666 /* Force WLAN LNA diversity ON */
3667 REG_SET_BIT(ah, AR_BTCOEX_WL_LNADIV,
3668 AR_BTCOEX_WL_LNADIV_FORCE_ON);
3659 } else { 3669 } else {
3660 regval &= ~(1 << AR_PHY_ANT_DIV_LNADIV_S); 3670 regval &= ~(1 << AR_PHY_ANT_DIV_LNADIV_S);
3661 regval &= ~(1 << AR_PHY_ANT_SW_RX_PROT_S); 3671 regval &= ~(1 << AR_PHY_ANT_SW_RX_PROT_S);
3672
3673 REG_CLR_BIT(ah, AR_PHY_MC_GAIN_CTRL,
3674 (1 << AR_PHY_ANT_SW_RX_PROT_S));
3675
3676 /* Force WLAN LNA diversity OFF */
3677 REG_CLR_BIT(ah, AR_BTCOEX_WL_LNADIV,
3678 AR_BTCOEX_WL_LNADIV_FORCE_ON);
3662 } 3679 }
3663 } 3680 }
3664 3681
@@ -3669,7 +3686,8 @@ static void ar9003_hw_ant_ctrl_apply(struct ath_hw *ah, bool is2ghz)
3669 regval &= (~AR_FAST_DIV_ENABLE); 3686 regval &= (~AR_FAST_DIV_ENABLE);
3670 regval |= ((value >> 7) & 0x1) << AR_FAST_DIV_ENABLE_S; 3687 regval |= ((value >> 7) & 0x1) << AR_FAST_DIV_ENABLE_S;
3671 3688
3672 if (AR_SREV_9485(ah) && common->bt_ant_diversity) 3689 if ((AR_SREV_9485(ah) || AR_SREV_9565(ah))
3690 && common->bt_ant_diversity)
3673 regval |= AR_FAST_DIV_ENABLE; 3691 regval |= AR_FAST_DIV_ENABLE;
3674 3692
3675 REG_WRITE(ah, AR_PHY_CCK_DETECT, regval); 3693 REG_WRITE(ah, AR_PHY_CCK_DETECT, regval);
diff --git a/drivers/net/wireless/ath/ath9k/ar9003_eeprom.h b/drivers/net/wireless/ath/ath9k/ar9003_eeprom.h
index 75d4fb41962f..0e5daa58a4fc 100644
--- a/drivers/net/wireless/ath/ath9k/ar9003_eeprom.h
+++ b/drivers/net/wireless/ath/ath9k/ar9003_eeprom.h
@@ -52,6 +52,8 @@
52#define AR9300_PAPRD_SCALE_2 0x70000000 52#define AR9300_PAPRD_SCALE_2 0x70000000
53#define AR9300_PAPRD_SCALE_2_S 28 53#define AR9300_PAPRD_SCALE_2_S 28
54 54
55#define AR9300_EEP_ANTDIV_CONTROL_DEFAULT_VALUE 0xc9
56
55/* Delta from which to start power to pdadc table */ 57/* Delta from which to start power to pdadc table */
56/* This offset is used in both open loop and closed loop power control 58/* This offset is used in both open loop and closed loop power control
57 * schemes. In open loop power control, it is not really needed, but for 59 * schemes. In open loop power control, it is not really needed, but for
diff --git a/drivers/net/wireless/ath/ath9k/ar9003_hw.c b/drivers/net/wireless/ath/ath9k/ar9003_hw.c
index 608bb4824e2a..b07f164d65cf 100644
--- a/drivers/net/wireless/ath/ath9k/ar9003_hw.c
+++ b/drivers/net/wireless/ath/ath9k/ar9003_hw.c
@@ -364,6 +364,8 @@ static void ar9003_hw_init_mode_regs(struct ath_hw *ah)
364 364
365 INIT_INI_ARRAY(&ah->iniModesFastClock, 365 INIT_INI_ARRAY(&ah->iniModesFastClock,
366 ar9565_1p0_modes_fast_clock); 366 ar9565_1p0_modes_fast_clock);
367 INIT_INI_ARRAY(&ah->iniCckfirJapan2484,
368 ar9565_1p0_baseband_core_txfir_coeff_japan_2484);
367 } else { 369 } else {
368 /* mac */ 370 /* mac */
369 INIT_INI_ARRAY(&ah->iniMac[ATH_INI_CORE], 371 INIT_INI_ARRAY(&ah->iniMac[ATH_INI_CORE],
@@ -628,6 +630,9 @@ static void ar9003_rx_gain_table_mode0(struct ath_hw *ah)
628 else if (AR_SREV_9462_20(ah)) 630 else if (AR_SREV_9462_20(ah))
629 INIT_INI_ARRAY(&ah->iniModesRxGain, 631 INIT_INI_ARRAY(&ah->iniModesRxGain,
630 ar9462_common_rx_gain_table_2p0); 632 ar9462_common_rx_gain_table_2p0);
633 else if (AR_SREV_9565(ah))
634 INIT_INI_ARRAY(&ah->iniModesRxGain,
635 ar9565_1p0_Common_rx_gain_table);
631 else 636 else
632 INIT_INI_ARRAY(&ah->iniModesRxGain, 637 INIT_INI_ARRAY(&ah->iniModesRxGain,
633 ar9300Common_rx_gain_table_2p2); 638 ar9300Common_rx_gain_table_2p2);
diff --git a/drivers/net/wireless/ath/ath9k/ar9003_mci.c b/drivers/net/wireless/ath/ath9k/ar9003_mci.c
index 8dd069259e7b..7b94a6c7db3d 100644
--- a/drivers/net/wireless/ath/ath9k/ar9003_mci.c
+++ b/drivers/net/wireless/ath/ath9k/ar9003_mci.c
@@ -753,9 +753,9 @@ int ar9003_mci_end_reset(struct ath_hw *ah, struct ath9k_channel *chan,
753 1 << AR_PHY_TIMING_CONTROL4_DO_GAIN_DC_IQ_CAL_SHIFT); 753 1 << AR_PHY_TIMING_CONTROL4_DO_GAIN_DC_IQ_CAL_SHIFT);
754 754
755 if (caldata) { 755 if (caldata) {
756 caldata->done_txiqcal_once = false; 756 clear_bit(TXIQCAL_DONE, &caldata->cal_flags);
757 caldata->done_txclcal_once = false; 757 clear_bit(TXCLCAL_DONE, &caldata->cal_flags);
758 caldata->rtt_done = false; 758 clear_bit(RTT_DONE, &caldata->cal_flags);
759 } 759 }
760 760
761 if (!ath9k_hw_init_cal(ah, chan)) 761 if (!ath9k_hw_init_cal(ah, chan))
diff --git a/drivers/net/wireless/ath/ath9k/ar9003_phy.c b/drivers/net/wireless/ath/ath9k/ar9003_phy.c
index e897648d3233..11f53589a3f3 100644
--- a/drivers/net/wireless/ath/ath9k/ar9003_phy.c
+++ b/drivers/net/wireless/ath/ath9k/ar9003_phy.c
@@ -551,8 +551,7 @@ static void ar9003_hw_set_channel_regs(struct ath_hw *ah,
551 if (IS_CHAN_HT40(chan)) { 551 if (IS_CHAN_HT40(chan)) {
552 phymode |= AR_PHY_GC_DYN2040_EN; 552 phymode |= AR_PHY_GC_DYN2040_EN;
553 /* Configure control (primary) channel at +-10MHz */ 553 /* Configure control (primary) channel at +-10MHz */
554 if ((chan->chanmode == CHANNEL_A_HT40PLUS) || 554 if (IS_CHAN_HT40PLUS(chan))
555 (chan->chanmode == CHANNEL_G_HT40PLUS))
556 phymode |= AR_PHY_GC_DYN2040_PRI_CH; 555 phymode |= AR_PHY_GC_DYN2040_PRI_CH;
557 556
558 } 557 }
@@ -565,7 +564,7 @@ static void ar9003_hw_set_channel_regs(struct ath_hw *ah,
565 REG_WRITE(ah, AR_PHY_GEN_CTRL, phymode); 564 REG_WRITE(ah, AR_PHY_GEN_CTRL, phymode);
566 565
567 /* Configure MAC for 20/40 operation */ 566 /* Configure MAC for 20/40 operation */
568 ath9k_hw_set11nmac2040(ah); 567 ath9k_hw_set11nmac2040(ah, chan);
569 568
570 /* global transmit timeout (25 TUs default)*/ 569 /* global transmit timeout (25 TUs default)*/
571 REG_WRITE(ah, AR_GTXTO, 25 << AR_GTXTO_TIMEOUT_LIMIT_S); 570 REG_WRITE(ah, AR_GTXTO, 25 << AR_GTXTO_TIMEOUT_LIMIT_S);
@@ -627,11 +626,10 @@ static void ar9003_hw_override_ini(struct ath_hw *ah)
627 * MAC addr only will fail. 626 * MAC addr only will fail.
628 */ 627 */
629 val = REG_READ(ah, AR_PCU_MISC_MODE2) & (~AR_ADHOC_MCAST_KEYID_ENABLE); 628 val = REG_READ(ah, AR_PCU_MISC_MODE2) & (~AR_ADHOC_MCAST_KEYID_ENABLE);
630 REG_WRITE(ah, AR_PCU_MISC_MODE2, 629 val |= AR_AGG_WEP_ENABLE_FIX |
631 val | AR_AGG_WEP_ENABLE_FIX | AR_AGG_WEP_ENABLE); 630 AR_AGG_WEP_ENABLE |
632 631 AR_PCU_MISC_MODE2_CFP_IGNORE;
633 REG_SET_BIT(ah, AR_PHY_CCK_DETECT, 632 REG_WRITE(ah, AR_PCU_MISC_MODE2, val);
634 AR_PHY_CCK_DETECT_BB_ENABLE_ANT_FAST_DIV);
635 633
636 if (AR_SREV_9462(ah) || AR_SREV_9565(ah)) { 634 if (AR_SREV_9462(ah) || AR_SREV_9565(ah)) {
637 REG_WRITE(ah, AR_GLB_SWREG_DISCONT_MODE, 635 REG_WRITE(ah, AR_GLB_SWREG_DISCONT_MODE,
@@ -683,41 +681,22 @@ static int ar9550_hw_get_modes_txgain_index(struct ath_hw *ah,
683{ 681{
684 int ret; 682 int ret;
685 683
686 switch (chan->chanmode) { 684 if (IS_CHAN_2GHZ(chan)) {
687 case CHANNEL_A: 685 if (IS_CHAN_HT40(chan))
688 case CHANNEL_A_HT20: 686 return 7;
689 if (chan->channel <= 5350)
690 ret = 1;
691 else if ((chan->channel > 5350) && (chan->channel <= 5600))
692 ret = 3;
693 else
694 ret = 5;
695 break;
696
697 case CHANNEL_A_HT40PLUS:
698 case CHANNEL_A_HT40MINUS:
699 if (chan->channel <= 5350)
700 ret = 2;
701 else if ((chan->channel > 5350) && (chan->channel <= 5600))
702 ret = 4;
703 else 687 else
704 ret = 6; 688 return 8;
705 break; 689 }
706
707 case CHANNEL_G:
708 case CHANNEL_G_HT20:
709 case CHANNEL_B:
710 ret = 8;
711 break;
712 690
713 case CHANNEL_G_HT40PLUS: 691 if (chan->channel <= 5350)
714 case CHANNEL_G_HT40MINUS: 692 ret = 1;
715 ret = 7; 693 else if ((chan->channel > 5350) && (chan->channel <= 5600))
716 break; 694 ret = 3;
695 else
696 ret = 5;
717 697
718 default: 698 if (IS_CHAN_HT40(chan))
719 ret = -EINVAL; 699 ret++;
720 }
721 700
722 return ret; 701 return ret;
723} 702}
@@ -728,28 +707,10 @@ static int ar9003_hw_process_ini(struct ath_hw *ah,
728 unsigned int regWrites = 0, i; 707 unsigned int regWrites = 0, i;
729 u32 modesIndex; 708 u32 modesIndex;
730 709
731 switch (chan->chanmode) { 710 if (IS_CHAN_5GHZ(chan))
732 case CHANNEL_A: 711 modesIndex = IS_CHAN_HT40(chan) ? 2 : 1;
733 case CHANNEL_A_HT20: 712 else
734 modesIndex = 1; 713 modesIndex = IS_CHAN_HT40(chan) ? 3 : 4;
735 break;
736 case CHANNEL_A_HT40PLUS:
737 case CHANNEL_A_HT40MINUS:
738 modesIndex = 2;
739 break;
740 case CHANNEL_G:
741 case CHANNEL_G_HT20:
742 case CHANNEL_B:
743 modesIndex = 4;
744 break;
745 case CHANNEL_G_HT40PLUS:
746 case CHANNEL_G_HT40MINUS:
747 modesIndex = 3;
748 break;
749
750 default:
751 return -EINVAL;
752 }
753 714
754 /* 715 /*
755 * SOC, MAC, BB, RADIO initvals. 716 * SOC, MAC, BB, RADIO initvals.
@@ -847,8 +808,10 @@ static void ar9003_hw_set_rfmode(struct ath_hw *ah,
847 if (chan == NULL) 808 if (chan == NULL)
848 return; 809 return;
849 810
850 rfMode |= (IS_CHAN_B(chan) || IS_CHAN_G(chan)) 811 if (IS_CHAN_2GHZ(chan))
851 ? AR_PHY_MODE_DYNAMIC : AR_PHY_MODE_OFDM; 812 rfMode |= AR_PHY_MODE_DYNAMIC;
813 else
814 rfMode |= AR_PHY_MODE_OFDM;
852 815
853 if (IS_CHAN_A_FAST_CLOCK(ah, chan)) 816 if (IS_CHAN_A_FAST_CLOCK(ah, chan))
854 rfMode |= (AR_PHY_MODE_DYNAMIC | AR_PHY_MODE_DYN_CCK_DISABLE); 817 rfMode |= (AR_PHY_MODE_DYNAMIC | AR_PHY_MODE_DYN_CCK_DISABLE);
@@ -1274,12 +1237,11 @@ static void ar9003_hw_ani_cache_ini_regs(struct ath_hw *ah)
1274 aniState = &ah->ani; 1237 aniState = &ah->ani;
1275 iniDef = &aniState->iniDef; 1238 iniDef = &aniState->iniDef;
1276 1239
1277 ath_dbg(common, ANI, "ver %d.%d opmode %u chan %d Mhz/0x%x\n", 1240 ath_dbg(common, ANI, "ver %d.%d opmode %u chan %d Mhz\n",
1278 ah->hw_version.macVersion, 1241 ah->hw_version.macVersion,
1279 ah->hw_version.macRev, 1242 ah->hw_version.macRev,
1280 ah->opmode, 1243 ah->opmode,
1281 chan->channel, 1244 chan->channel);
1282 chan->channelFlags);
1283 1245
1284 val = REG_READ(ah, AR_PHY_SFCORR); 1246 val = REG_READ(ah, AR_PHY_SFCORR);
1285 iniDef->m1Thresh = MS(val, AR_PHY_SFCORR_M1_THRESH); 1247 iniDef->m1Thresh = MS(val, AR_PHY_SFCORR_M1_THRESH);
@@ -1375,15 +1337,19 @@ static void ar9003_hw_antdiv_comb_conf_get(struct ath_hw *ah,
1375 AR_PHY_ANT_FAST_DIV_BIAS_S; 1337 AR_PHY_ANT_FAST_DIV_BIAS_S;
1376 1338
1377 if (AR_SREV_9330_11(ah)) { 1339 if (AR_SREV_9330_11(ah)) {
1340 antconf->lna1_lna2_switch_delta = -1;
1378 antconf->lna1_lna2_delta = -9; 1341 antconf->lna1_lna2_delta = -9;
1379 antconf->div_group = 1; 1342 antconf->div_group = 1;
1380 } else if (AR_SREV_9485(ah)) { 1343 } else if (AR_SREV_9485(ah)) {
1344 antconf->lna1_lna2_switch_delta = -1;
1381 antconf->lna1_lna2_delta = -9; 1345 antconf->lna1_lna2_delta = -9;
1382 antconf->div_group = 2; 1346 antconf->div_group = 2;
1383 } else if (AR_SREV_9565(ah)) { 1347 } else if (AR_SREV_9565(ah)) {
1384 antconf->lna1_lna2_delta = -3; 1348 antconf->lna1_lna2_switch_delta = 3;
1349 antconf->lna1_lna2_delta = -9;
1385 antconf->div_group = 3; 1350 antconf->div_group = 3;
1386 } else { 1351 } else {
1352 antconf->lna1_lna2_switch_delta = -1;
1387 antconf->lna1_lna2_delta = -3; 1353 antconf->lna1_lna2_delta = -3;
1388 antconf->div_group = 0; 1354 antconf->div_group = 0;
1389 } 1355 }
@@ -1489,17 +1455,24 @@ static void ar9003_hw_set_bt_ant_diversity(struct ath_hw *ah, bool enable)
1489 } else if (AR_SREV_9565(ah)) { 1455 } else if (AR_SREV_9565(ah)) {
1490 if (enable) { 1456 if (enable) {
1491 REG_SET_BIT(ah, AR_PHY_MC_GAIN_CTRL, 1457 REG_SET_BIT(ah, AR_PHY_MC_GAIN_CTRL,
1458 AR_ANT_DIV_ENABLE);
1459 REG_SET_BIT(ah, AR_PHY_MC_GAIN_CTRL,
1492 (1 << AR_PHY_ANT_SW_RX_PROT_S)); 1460 (1 << AR_PHY_ANT_SW_RX_PROT_S));
1493 if (ah->curchan && IS_CHAN_2GHZ(ah->curchan)) 1461 REG_SET_BIT(ah, AR_PHY_CCK_DETECT,
1494 REG_SET_BIT(ah, AR_PHY_RESTART, 1462 AR_FAST_DIV_ENABLE);
1495 AR_PHY_RESTART_ENABLE_DIV_M2FLAG); 1463 REG_SET_BIT(ah, AR_PHY_RESTART,
1464 AR_PHY_RESTART_ENABLE_DIV_M2FLAG);
1496 REG_SET_BIT(ah, AR_BTCOEX_WL_LNADIV, 1465 REG_SET_BIT(ah, AR_BTCOEX_WL_LNADIV,
1497 AR_BTCOEX_WL_LNADIV_FORCE_ON); 1466 AR_BTCOEX_WL_LNADIV_FORCE_ON);
1498 } else { 1467 } else {
1499 REG_CLR_BIT(ah, AR_PHY_MC_GAIN_CTRL, AR_ANT_DIV_ENABLE); 1468 REG_CLR_BIT(ah, AR_PHY_MC_GAIN_CTRL,
1469 AR_ANT_DIV_ENABLE);
1500 REG_CLR_BIT(ah, AR_PHY_MC_GAIN_CTRL, 1470 REG_CLR_BIT(ah, AR_PHY_MC_GAIN_CTRL,
1501 (1 << AR_PHY_ANT_SW_RX_PROT_S)); 1471 (1 << AR_PHY_ANT_SW_RX_PROT_S));
1502 REG_CLR_BIT(ah, AR_PHY_CCK_DETECT, AR_FAST_DIV_ENABLE); 1472 REG_CLR_BIT(ah, AR_PHY_CCK_DETECT,
1473 AR_FAST_DIV_ENABLE);
1474 REG_CLR_BIT(ah, AR_PHY_RESTART,
1475 AR_PHY_RESTART_ENABLE_DIV_M2FLAG);
1503 REG_CLR_BIT(ah, AR_BTCOEX_WL_LNADIV, 1476 REG_CLR_BIT(ah, AR_BTCOEX_WL_LNADIV,
1504 AR_BTCOEX_WL_LNADIV_FORCE_ON); 1477 AR_BTCOEX_WL_LNADIV_FORCE_ON);
1505 1478
@@ -1526,28 +1499,10 @@ static int ar9003_hw_fast_chan_change(struct ath_hw *ah,
1526 unsigned int regWrites = 0; 1499 unsigned int regWrites = 0;
1527 u32 modesIndex; 1500 u32 modesIndex;
1528 1501
1529 switch (chan->chanmode) { 1502 if (IS_CHAN_5GHZ(chan))
1530 case CHANNEL_A: 1503 modesIndex = IS_CHAN_HT40(chan) ? 2 : 1;
1531 case CHANNEL_A_HT20: 1504 else
1532 modesIndex = 1; 1505 modesIndex = IS_CHAN_HT40(chan) ? 3 : 4;
1533 break;
1534 case CHANNEL_A_HT40PLUS:
1535 case CHANNEL_A_HT40MINUS:
1536 modesIndex = 2;
1537 break;
1538 case CHANNEL_G:
1539 case CHANNEL_G_HT20:
1540 case CHANNEL_B:
1541 modesIndex = 4;
1542 break;
1543 case CHANNEL_G_HT40PLUS:
1544 case CHANNEL_G_HT40MINUS:
1545 modesIndex = 3;
1546 break;
1547
1548 default:
1549 return -EINVAL;
1550 }
1551 1506
1552 if (modesIndex == ah->modes_index) { 1507 if (modesIndex == ah->modes_index) {
1553 *ini_reloaded = false; 1508 *ini_reloaded = false;
@@ -1662,6 +1617,98 @@ static void ar9003_hw_spectral_scan_wait(struct ath_hw *ah)
1662 } 1617 }
1663} 1618}
1664 1619
1620static void ar9003_hw_tx99_start(struct ath_hw *ah, u32 qnum)
1621{
1622 REG_SET_BIT(ah, AR_PHY_TEST, PHY_AGC_CLR);
1623 REG_SET_BIT(ah, 0x9864, 0x7f000);
1624 REG_SET_BIT(ah, 0x9924, 0x7f00fe);
1625 REG_CLR_BIT(ah, AR_DIAG_SW, AR_DIAG_RX_DIS);
1626 REG_WRITE(ah, AR_CR, AR_CR_RXD);
1627 REG_WRITE(ah, AR_DLCL_IFS(qnum), 0);
1628 REG_WRITE(ah, AR_D_GBL_IFS_SIFS, 20); /* 50 OK */
1629 REG_WRITE(ah, AR_D_GBL_IFS_EIFS, 20);
1630 REG_WRITE(ah, AR_TIME_OUT, 0x00000400);
1631 REG_WRITE(ah, AR_DRETRY_LIMIT(qnum), 0xffffffff);
1632 REG_SET_BIT(ah, AR_QMISC(qnum), AR_Q_MISC_DCU_EARLY_TERM_REQ);
1633}
1634
1635static void ar9003_hw_tx99_stop(struct ath_hw *ah)
1636{
1637 REG_CLR_BIT(ah, AR_PHY_TEST, PHY_AGC_CLR);
1638 REG_SET_BIT(ah, AR_DIAG_SW, AR_DIAG_RX_DIS);
1639}
1640
1641static void ar9003_hw_tx99_set_txpower(struct ath_hw *ah, u8 txpower)
1642{
1643 static s16 p_pwr_array[ar9300RateSize] = { 0 };
1644 unsigned int i;
1645
1646 if (txpower <= MAX_RATE_POWER) {
1647 for (i = 0; i < ar9300RateSize; i++)
1648 p_pwr_array[i] = txpower;
1649 } else {
1650 for (i = 0; i < ar9300RateSize; i++)
1651 p_pwr_array[i] = MAX_RATE_POWER;
1652 }
1653
1654 REG_WRITE(ah, 0xa458, 0);
1655
1656 REG_WRITE(ah, 0xa3c0,
1657 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_LEGACY_6_24], 24) |
1658 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_LEGACY_6_24], 16) |
1659 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_LEGACY_6_24], 8) |
1660 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_LEGACY_6_24], 0));
1661 REG_WRITE(ah, 0xa3c4,
1662 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_LEGACY_54], 24) |
1663 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_LEGACY_48], 16) |
1664 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_LEGACY_36], 8) |
1665 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_LEGACY_6_24], 0));
1666 REG_WRITE(ah, 0xa3c8,
1667 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_LEGACY_1L_5L], 24) |
1668 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_LEGACY_1L_5L], 16) |
1669 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_LEGACY_1L_5L], 0));
1670 REG_WRITE(ah, 0xa3cc,
1671 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_LEGACY_11S], 24) |
1672 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_LEGACY_11L], 16) |
1673 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_LEGACY_5S], 8) |
1674 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_LEGACY_1L_5L], 0));
1675 REG_WRITE(ah, 0xa3d0,
1676 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT20_5], 24) |
1677 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT20_4], 16) |
1678 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT20_1_3_9_11_17_19], 8)|
1679 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT20_0_8_16], 0));
1680 REG_WRITE(ah, 0xa3d4,
1681 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT20_13], 24) |
1682 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT20_12], 16) |
1683 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT20_7], 8) |
1684 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT20_6], 0));
1685 REG_WRITE(ah, 0xa3e4,
1686 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT20_21], 24) |
1687 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT20_20], 16) |
1688 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT20_15], 8) |
1689 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT20_14], 0));
1690 REG_WRITE(ah, 0xa3e8,
1691 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT40_23], 24) |
1692 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT40_22], 16) |
1693 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT20_23], 8) |
1694 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT20_22], 0));
1695 REG_WRITE(ah, 0xa3d8,
1696 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT40_5], 24) |
1697 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT40_4], 16) |
1698 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT40_1_3_9_11_17_19], 8) |
1699 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT40_0_8_16], 0));
1700 REG_WRITE(ah, 0xa3dc,
1701 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT40_13], 24) |
1702 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT40_12], 16) |
1703 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT40_7], 8) |
1704 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT40_6], 0));
1705 REG_WRITE(ah, 0xa3ec,
1706 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT40_21], 24) |
1707 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT40_20], 16) |
1708 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT40_15], 8) |
1709 ATH9K_POW_SM(p_pwr_array[ALL_TARGET_HT40_14], 0));
1710}
1711
1665void ar9003_hw_attach_phy_ops(struct ath_hw *ah) 1712void ar9003_hw_attach_phy_ops(struct ath_hw *ah)
1666{ 1713{
1667 struct ath_hw_private_ops *priv_ops = ath9k_hw_private_ops(ah); 1714 struct ath_hw_private_ops *priv_ops = ath9k_hw_private_ops(ah);
@@ -1701,6 +1748,9 @@ void ar9003_hw_attach_phy_ops(struct ath_hw *ah)
1701#ifdef CONFIG_ATH9K_BTCOEX_SUPPORT 1748#ifdef CONFIG_ATH9K_BTCOEX_SUPPORT
1702 ops->set_bt_ant_diversity = ar9003_hw_set_bt_ant_diversity; 1749 ops->set_bt_ant_diversity = ar9003_hw_set_bt_ant_diversity;
1703#endif 1750#endif
1751 ops->tx99_start = ar9003_hw_tx99_start;
1752 ops->tx99_stop = ar9003_hw_tx99_stop;
1753 ops->tx99_set_txpower = ar9003_hw_tx99_set_txpower;
1704 1754
1705 ar9003_hw_set_nf_limits(ah); 1755 ar9003_hw_set_nf_limits(ah);
1706 ar9003_hw_set_radar_conf(ah); 1756 ar9003_hw_set_radar_conf(ah);
diff --git a/drivers/net/wireless/ath/ath9k/ar9003_phy.h b/drivers/net/wireless/ath/ath9k/ar9003_phy.h
index 6fd752321e36..fca624322dc8 100644
--- a/drivers/net/wireless/ath/ath9k/ar9003_phy.h
+++ b/drivers/net/wireless/ath/ath9k/ar9003_phy.h
@@ -343,8 +343,12 @@
343 343
344#define AR_PHY_CCA_NOM_VAL_9462_2GHZ -127 344#define AR_PHY_CCA_NOM_VAL_9462_2GHZ -127
345#define AR_PHY_CCA_MIN_GOOD_VAL_9462_2GHZ -127 345#define AR_PHY_CCA_MIN_GOOD_VAL_9462_2GHZ -127
346#define AR_PHY_CCA_MAX_GOOD_VAL_9462_2GHZ -60
347#define AR_PHY_CCA_MAX_GOOD_VAL_9462_FCC_2GHZ -95
346#define AR_PHY_CCA_NOM_VAL_9462_5GHZ -127 348#define AR_PHY_CCA_NOM_VAL_9462_5GHZ -127
347#define AR_PHY_CCA_MIN_GOOD_VAL_9462_5GHZ -127 349#define AR_PHY_CCA_MIN_GOOD_VAL_9462_5GHZ -127
350#define AR_PHY_CCA_MAX_GOOD_VAL_9462_5GHZ -60
351#define AR_PHY_CCA_MAX_GOOD_VAL_9462_FCC_5GHZ -100
348 352
349#define AR_PHY_CCA_NOM_VAL_9330_2GHZ -118 353#define AR_PHY_CCA_NOM_VAL_9330_2GHZ -118
350 354
diff --git a/drivers/net/wireless/ath/ath9k/ar9003_rtt.c b/drivers/net/wireless/ath/ath9k/ar9003_rtt.c
index 74de3539c2c8..934418872e8e 100644
--- a/drivers/net/wireless/ath/ath9k/ar9003_rtt.c
+++ b/drivers/net/wireless/ath/ath9k/ar9003_rtt.c
@@ -118,6 +118,27 @@ void ar9003_hw_rtt_load_hist(struct ath_hw *ah)
118 } 118 }
119} 119}
120 120
121static void ar9003_hw_patch_rtt(struct ath_hw *ah, int index, int chain)
122{
123 int agc, caldac;
124
125 if (!test_bit(SW_PKDET_DONE, &ah->caldata->cal_flags))
126 return;
127
128 if ((index != 5) || (chain >= 2))
129 return;
130
131 agc = REG_READ_FIELD(ah, AR_PHY_65NM_RXRF_AGC(chain),
132 AR_PHY_65NM_RXRF_AGC_AGC_OVERRIDE);
133 if (!agc)
134 return;
135
136 caldac = ah->caldata->caldac[chain];
137 ah->caldata->rtt_table[chain][index] &= 0xFFFF05FF;
138 caldac = (caldac & 0x20) | ((caldac & 0x1F) << 7);
139 ah->caldata->rtt_table[chain][index] |= (caldac << 4);
140}
141
121static int ar9003_hw_rtt_fill_hist_entry(struct ath_hw *ah, u8 chain, u32 index) 142static int ar9003_hw_rtt_fill_hist_entry(struct ath_hw *ah, u8 chain, u32 index)
122{ 143{
123 u32 val; 144 u32 val;
@@ -155,13 +176,16 @@ void ar9003_hw_rtt_fill_hist(struct ath_hw *ah)
155 for (i = 0; i < MAX_RTT_TABLE_ENTRY; i++) { 176 for (i = 0; i < MAX_RTT_TABLE_ENTRY; i++) {
156 ah->caldata->rtt_table[chain][i] = 177 ah->caldata->rtt_table[chain][i] =
157 ar9003_hw_rtt_fill_hist_entry(ah, chain, i); 178 ar9003_hw_rtt_fill_hist_entry(ah, chain, i);
179
180 ar9003_hw_patch_rtt(ah, i, chain);
181
158 ath_dbg(ath9k_hw_common(ah), CALIBRATE, 182 ath_dbg(ath9k_hw_common(ah), CALIBRATE,
159 "RTT value at idx %d, chain %d is: 0x%x\n", 183 "RTT value at idx %d, chain %d is: 0x%x\n",
160 i, chain, ah->caldata->rtt_table[chain][i]); 184 i, chain, ah->caldata->rtt_table[chain][i]);
161 } 185 }
162 } 186 }
163 187
164 ah->caldata->rtt_done = true; 188 set_bit(RTT_DONE, &ah->caldata->cal_flags);
165} 189}
166 190
167void ar9003_hw_rtt_clear_hist(struct ath_hw *ah) 191void ar9003_hw_rtt_clear_hist(struct ath_hw *ah)
@@ -176,7 +200,7 @@ void ar9003_hw_rtt_clear_hist(struct ath_hw *ah)
176 } 200 }
177 201
178 if (ah->caldata) 202 if (ah->caldata)
179 ah->caldata->rtt_done = false; 203 clear_bit(RTT_DONE, &ah->caldata->cal_flags);
180} 204}
181 205
182bool ar9003_hw_rtt_restore(struct ath_hw *ah, struct ath9k_channel *chan) 206bool ar9003_hw_rtt_restore(struct ath_hw *ah, struct ath9k_channel *chan)
@@ -186,11 +210,37 @@ bool ar9003_hw_rtt_restore(struct ath_hw *ah, struct ath9k_channel *chan)
186 if (!ah->caldata) 210 if (!ah->caldata)
187 return false; 211 return false;
188 212
189 if (!ah->caldata->rtt_done) 213 if (test_bit(SW_PKDET_DONE, &ah->caldata->cal_flags)) {
214 if (IS_CHAN_2GHZ(chan)){
215 REG_RMW_FIELD(ah, AR_PHY_65NM_RXRF_AGC(0),
216 AR_PHY_65NM_RXRF_AGC_AGC2G_CALDAC_OVR,
217 ah->caldata->caldac[0]);
218 REG_RMW_FIELD(ah, AR_PHY_65NM_RXRF_AGC(1),
219 AR_PHY_65NM_RXRF_AGC_AGC2G_CALDAC_OVR,
220 ah->caldata->caldac[1]);
221 } else {
222 REG_RMW_FIELD(ah, AR_PHY_65NM_RXRF_AGC(0),
223 AR_PHY_65NM_RXRF_AGC_AGC5G_CALDAC_OVR,
224 ah->caldata->caldac[0]);
225 REG_RMW_FIELD(ah, AR_PHY_65NM_RXRF_AGC(1),
226 AR_PHY_65NM_RXRF_AGC_AGC5G_CALDAC_OVR,
227 ah->caldata->caldac[1]);
228 }
229 REG_RMW_FIELD(ah, AR_PHY_65NM_RXRF_AGC(1),
230 AR_PHY_65NM_RXRF_AGC_AGC_OVERRIDE, 0x1);
231 REG_RMW_FIELD(ah, AR_PHY_65NM_RXRF_AGC(0),
232 AR_PHY_65NM_RXRF_AGC_AGC_OVERRIDE, 0x1);
233 }
234
235 if (!test_bit(RTT_DONE, &ah->caldata->cal_flags))
190 return false; 236 return false;
191 237
192 ar9003_hw_rtt_enable(ah); 238 ar9003_hw_rtt_enable(ah);
193 ar9003_hw_rtt_set_mask(ah, 0x10); 239
240 if (test_bit(SW_PKDET_DONE, &ah->caldata->cal_flags))
241 ar9003_hw_rtt_set_mask(ah, 0x30);
242 else
243 ar9003_hw_rtt_set_mask(ah, 0x10);
194 244
195 if (!ath9k_hw_rfbus_req(ah)) { 245 if (!ath9k_hw_rfbus_req(ah)) {
196 ath_err(ath9k_hw_common(ah), "Could not stop baseband\n"); 246 ath_err(ath9k_hw_common(ah), "Could not stop baseband\n");
diff --git a/drivers/net/wireless/ath/ath9k/ar9485_initvals.h b/drivers/net/wireless/ath/ath9k/ar9485_initvals.h
index 88ff1d7b53ab..6f899c692647 100644
--- a/drivers/net/wireless/ath/ath9k/ar9485_initvals.h
+++ b/drivers/net/wireless/ath/ath9k/ar9485_initvals.h
@@ -20,7 +20,17 @@
20 20
21/* AR9485 1.1 */ 21/* AR9485 1.1 */
22 22
23#define ar9485_1_1_mac_postamble ar9300_2p2_mac_postamble 23static const u32 ar9485_1_1_mac_postamble[][5] = {
24 /* Addr 5G_HT20 5G_HT40 2G_HT40 2G_HT20 */
25 {0x00001030, 0x00000230, 0x00000460, 0x000002c0, 0x00000160},
26 {0x00001070, 0x00000168, 0x000002d0, 0x00000318, 0x0000018c},
27 {0x000010b0, 0x00000e60, 0x00001cc0, 0x00007c70, 0x00003e38},
28 {0x00008014, 0x03e803e8, 0x07d007d0, 0x10801600, 0x08400b00},
29 {0x0000801c, 0x128d8027, 0x128d804f, 0x12e00057, 0x12e0002b},
30 {0x00008120, 0x08f04800, 0x08f04800, 0x08f04810, 0x08f04810},
31 {0x000081d0, 0x00003210, 0x00003210, 0x0000320a, 0x0000320a},
32 {0x00008318, 0x00003e80, 0x00007d00, 0x00006880, 0x00003440},
33};
24 34
25static const u32 ar9485_1_1_pcie_phy_pll_on_clkreq_disable_L1[][2] = { 35static const u32 ar9485_1_1_pcie_phy_pll_on_clkreq_disable_L1[][2] = {
26 /* Addr allmodes */ 36 /* Addr allmodes */
@@ -34,6 +44,7 @@ static const u32 ar9485Common_wo_xlna_rx_gain_1_1[][2] = {
34 {0x00009e00, 0x037216a0}, 44 {0x00009e00, 0x037216a0},
35 {0x00009e04, 0x00182020}, 45 {0x00009e04, 0x00182020},
36 {0x00009e18, 0x00000000}, 46 {0x00009e18, 0x00000000},
47 {0x00009e20, 0x000003a8},
37 {0x00009e2c, 0x00004121}, 48 {0x00009e2c, 0x00004121},
38 {0x00009e44, 0x02282324}, 49 {0x00009e44, 0x02282324},
39 {0x0000a000, 0x00060005}, 50 {0x0000a000, 0x00060005},
@@ -174,7 +185,7 @@ static const u32 ar9485Modes_high_power_tx_gain_1_1[][5] = {
174 {0x0000a2e0, 0x00000000, 0x00000000, 0xfe2d3552, 0xfe2d3552}, 185 {0x0000a2e0, 0x00000000, 0x00000000, 0xfe2d3552, 0xfe2d3552},
175 {0x0000a2e4, 0x00000000, 0x00000000, 0xfe2d3552, 0xfe2d3552}, 186 {0x0000a2e4, 0x00000000, 0x00000000, 0xfe2d3552, 0xfe2d3552},
176 {0x0000a2e8, 0x00000000, 0x00000000, 0xfe2d3552, 0xfe2d3552}, 187 {0x0000a2e8, 0x00000000, 0x00000000, 0xfe2d3552, 0xfe2d3552},
177 {0x0000a410, 0x000050d9, 0x000050d9, 0x000050d8, 0x000050d8}, 188 {0x0000a410, 0x000050d9, 0x000050d9, 0x000050da, 0x000050da},
178 {0x0000a458, 0x00000000, 0x00000000, 0x00000000, 0x00000000}, 189 {0x0000a458, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
179 {0x0000a500, 0x00022200, 0x00022200, 0x00000000, 0x00000000}, 190 {0x0000a500, 0x00022200, 0x00022200, 0x00000000, 0x00000000},
180 {0x0000a504, 0x05062002, 0x05062002, 0x04000002, 0x04000002}, 191 {0x0000a504, 0x05062002, 0x05062002, 0x04000002, 0x04000002},
@@ -200,14 +211,14 @@ static const u32 ar9485Modes_high_power_tx_gain_1_1[][5] = {
200 {0x0000a554, 0x7d06ffcb, 0x7d06ffcb, 0x57001ce9, 0x57001ce9}, 211 {0x0000a554, 0x7d06ffcb, 0x7d06ffcb, 0x57001ce9, 0x57001ce9},
201 {0x0000a558, 0x8407ff0b, 0x8407ff0b, 0x5a001ceb, 0x5a001ceb}, 212 {0x0000a558, 0x8407ff0b, 0x8407ff0b, 0x5a001ceb, 0x5a001ceb},
202 {0x0000a55c, 0x8907ffcb, 0x8907ffcb, 0x5e001eeb, 0x5e001eeb}, 213 {0x0000a55c, 0x8907ffcb, 0x8907ffcb, 0x5e001eeb, 0x5e001eeb},
203 {0x0000a560, 0x900fff0b, 0x900fff0b, 0x5e001eeb, 0x5e001eeb}, 214 {0x0000a560, 0x900fff0b, 0x900fff0b, 0x62001eee, 0x62001eee},
204 {0x0000a564, 0x960fffcb, 0x960fffcb, 0x5e001eeb, 0x5e001eeb}, 215 {0x0000a564, 0x960fffcb, 0x960fffcb, 0x66001ff6, 0x66001ff6},
205 {0x0000a568, 0x9c1fff0b, 0x9c1fff0b, 0x5e001eeb, 0x5e001eeb}, 216 {0x0000a568, 0x9c1fff0b, 0x9c1fff0b, 0x66001ff6, 0x66001ff6},
206 {0x0000a56c, 0x9c1fff0b, 0x9c1fff0b, 0x5e001eeb, 0x5e001eeb}, 217 {0x0000a56c, 0x9c1fff0b, 0x9c1fff0b, 0x66001ff6, 0x66001ff6},
207 {0x0000a570, 0x9c1fff0b, 0x9c1fff0b, 0x5e001eeb, 0x5e001eeb}, 218 {0x0000a570, 0x9c1fff0b, 0x9c1fff0b, 0x66001ff6, 0x66001ff6},
208 {0x0000a574, 0x9c1fff0b, 0x9c1fff0b, 0x5e001eeb, 0x5e001eeb}, 219 {0x0000a574, 0x9c1fff0b, 0x9c1fff0b, 0x66001ff6, 0x66001ff6},
209 {0x0000a578, 0x9c1fff0b, 0x9c1fff0b, 0x5e001eeb, 0x5e001eeb}, 220 {0x0000a578, 0x9c1fff0b, 0x9c1fff0b, 0x66001ff6, 0x66001ff6},
210 {0x0000a57c, 0x9c1fff0b, 0x9c1fff0b, 0x5e001eeb, 0x5e001eeb}, 221 {0x0000a57c, 0x9c1fff0b, 0x9c1fff0b, 0x66001ff6, 0x66001ff6},
211 {0x0000a580, 0x00000000, 0x00000000, 0x00000000, 0x00000000}, 222 {0x0000a580, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
212 {0x0000a584, 0x00000000, 0x00000000, 0x00000000, 0x00000000}, 223 {0x0000a584, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
213 {0x0000a588, 0x00000000, 0x00000000, 0x00000000, 0x00000000}, 224 {0x0000a588, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
@@ -263,6 +274,11 @@ static const u32 ar9485Modes_high_power_tx_gain_1_1[][5] = {
263static const u32 ar9485Modes_green_ob_db_tx_gain_1_1[][5] = { 274static const u32 ar9485Modes_green_ob_db_tx_gain_1_1[][5] = {
264 /* Addr 5G_HT20 5G_HT40 2G_HT40 2G_HT20 */ 275 /* Addr 5G_HT20 5G_HT40 2G_HT40 2G_HT20 */
265 {0x000098bc, 0x00000003, 0x00000003, 0x00000003, 0x00000003}, 276 {0x000098bc, 0x00000003, 0x00000003, 0x00000003, 0x00000003},
277 {0x0000a2d8, 0xf999a83a, 0xf999a83a, 0x7999a83a, 0x7999a83a},
278 {0x0000a2dc, 0x00000000, 0x00000000, 0xfe2d3552, 0xfe2d3552},
279 {0x0000a2e0, 0x00000000, 0x00000000, 0xfe2d3552, 0xfe2d3552},
280 {0x0000a2e4, 0x00000000, 0x00000000, 0xfe2d3552, 0xfe2d3552},
281 {0x0000a2e8, 0x00000000, 0x00000000, 0xfe2d3552, 0xfe2d3552},
266 {0x0000a410, 0x000050d9, 0x000050d9, 0x000050d8, 0x000050d8}, 282 {0x0000a410, 0x000050d9, 0x000050d9, 0x000050d8, 0x000050d8},
267 {0x0000a458, 0x80000000, 0x80000000, 0x80000000, 0x80000000}, 283 {0x0000a458, 0x80000000, 0x80000000, 0x80000000, 0x80000000},
268 {0x0000a500, 0x00022200, 0x00022200, 0x00000006, 0x00000006}, 284 {0x0000a500, 0x00022200, 0x00022200, 0x00000006, 0x00000006},
@@ -297,6 +313,22 @@ static const u32 ar9485Modes_green_ob_db_tx_gain_1_1[][5] = {
297 {0x0000a574, 0x9c1fff0b, 0x9c1fff0b, 0x5e001eeb, 0x5e001eeb}, 313 {0x0000a574, 0x9c1fff0b, 0x9c1fff0b, 0x5e001eeb, 0x5e001eeb},
298 {0x0000a578, 0x9c1fff0b, 0x9c1fff0b, 0x5e001eeb, 0x5e001eeb}, 314 {0x0000a578, 0x9c1fff0b, 0x9c1fff0b, 0x5e001eeb, 0x5e001eeb},
299 {0x0000a57c, 0x9c1fff0b, 0x9c1fff0b, 0x5e001eeb, 0x5e001eeb}, 315 {0x0000a57c, 0x9c1fff0b, 0x9c1fff0b, 0x5e001eeb, 0x5e001eeb},
316 {0x0000a580, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
317 {0x0000a584, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
318 {0x0000a588, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
319 {0x0000a58c, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
320 {0x0000a590, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
321 {0x0000a594, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
322 {0x0000a598, 0x00000000, 0x00000000, 0x01404501, 0x01404501},
323 {0x0000a59c, 0x00000000, 0x00000000, 0x02808a02, 0x02808a02},
324 {0x0000a5a0, 0x00000000, 0x00000000, 0x02808a02, 0x02808a02},
325 {0x0000a5a4, 0x00000000, 0x00000000, 0x02808803, 0x02808803},
326 {0x0000a5a8, 0x00000000, 0x00000000, 0x04c14b04, 0x04c14b04},
327 {0x0000a5ac, 0x00000000, 0x00000000, 0x04c15305, 0x04c15305},
328 {0x0000a5b0, 0x00000000, 0x00000000, 0x04c15305, 0x04c15305},
329 {0x0000a5b4, 0x00000000, 0x00000000, 0x04c15305, 0x04c15305},
330 {0x0000a5b8, 0x00000000, 0x00000000, 0x04c15305, 0x04c15305},
331 {0x0000a5bc, 0x00000000, 0x00000000, 0x04c15305, 0x04c15305},
300 {0x0000b500, 0x0000001a, 0x0000001a, 0x0000001a, 0x0000001a}, 332 {0x0000b500, 0x0000001a, 0x0000001a, 0x0000001a, 0x0000001a},
301 {0x0000b504, 0x0000001a, 0x0000001a, 0x0000001a, 0x0000001a}, 333 {0x0000b504, 0x0000001a, 0x0000001a, 0x0000001a, 0x0000001a},
302 {0x0000b508, 0x0000001a, 0x0000001a, 0x0000001a, 0x0000001a}, 334 {0x0000b508, 0x0000001a, 0x0000001a, 0x0000001a, 0x0000001a},
@@ -341,6 +373,100 @@ static const u32 ar9485Modes_high_ob_db_tx_gain_1_1[][5] = {
341 {0x0000a2e0, 0x00000000, 0x00000000, 0xffc63a84, 0xffc63a84}, 373 {0x0000a2e0, 0x00000000, 0x00000000, 0xffc63a84, 0xffc63a84},
342 {0x0000a2e4, 0x00000000, 0x00000000, 0xfe0fc000, 0xfe0fc000}, 374 {0x0000a2e4, 0x00000000, 0x00000000, 0xfe0fc000, 0xfe0fc000},
343 {0x0000a2e8, 0x00000000, 0x00000000, 0xfff00000, 0xfff00000}, 375 {0x0000a2e8, 0x00000000, 0x00000000, 0xfff00000, 0xfff00000},
376 {0x0000a410, 0x000050d9, 0x000050d9, 0x000050da, 0x000050da},
377 {0x0000a458, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
378 {0x0000a500, 0x00022200, 0x00022200, 0x00000000, 0x00000000},
379 {0x0000a504, 0x05062002, 0x05062002, 0x04000002, 0x04000002},
380 {0x0000a508, 0x0c002e00, 0x0c002e00, 0x08000004, 0x08000004},
381 {0x0000a50c, 0x11062202, 0x11062202, 0x0d000200, 0x0d000200},
382 {0x0000a510, 0x17022e00, 0x17022e00, 0x11000202, 0x11000202},
383 {0x0000a514, 0x1d000ec2, 0x1d000ec2, 0x15000400, 0x15000400},
384 {0x0000a518, 0x25020ec0, 0x25020ec0, 0x19000402, 0x19000402},
385 {0x0000a51c, 0x2b020ec3, 0x2b020ec3, 0x1d000404, 0x1d000404},
386 {0x0000a520, 0x2f001f04, 0x2f001f04, 0x21000603, 0x21000603},
387 {0x0000a524, 0x35001fc4, 0x35001fc4, 0x25000605, 0x25000605},
388 {0x0000a528, 0x3c022f04, 0x3c022f04, 0x2a000a03, 0x2a000a03},
389 {0x0000a52c, 0x41023e85, 0x41023e85, 0x2c000a04, 0x2c000a04},
390 {0x0000a530, 0x48023ec6, 0x48023ec6, 0x34000e20, 0x34000e20},
391 {0x0000a534, 0x4d023f01, 0x4d023f01, 0x35000e21, 0x35000e21},
392 {0x0000a538, 0x53023f4b, 0x53023f4b, 0x43000e62, 0x43000e62},
393 {0x0000a53c, 0x5a027f09, 0x5a027f09, 0x45000e63, 0x45000e63},
394 {0x0000a540, 0x5f027fc9, 0x5f027fc9, 0x49000e65, 0x49000e65},
395 {0x0000a544, 0x6502feca, 0x6502feca, 0x4b000e66, 0x4b000e66},
396 {0x0000a548, 0x6b02ff4a, 0x6b02ff4a, 0x4d001645, 0x4d001645},
397 {0x0000a54c, 0x7203feca, 0x7203feca, 0x51001865, 0x51001865},
398 {0x0000a550, 0x7703ff0b, 0x7703ff0b, 0x55001a86, 0x55001a86},
399 {0x0000a554, 0x7d06ffcb, 0x7d06ffcb, 0x57001ce9, 0x57001ce9},
400 {0x0000a558, 0x8407ff0b, 0x8407ff0b, 0x5a001ceb, 0x5a001ceb},
401 {0x0000a55c, 0x8907ffcb, 0x8907ffcb, 0x5e001eeb, 0x5e001eeb},
402 {0x0000a560, 0x900fff0b, 0x900fff0b, 0x62001eee, 0x62001eee},
403 {0x0000a564, 0x960fffcb, 0x960fffcb, 0x66001ff6, 0x66001ff6},
404 {0x0000a568, 0x9c1fff0b, 0x9c1fff0b, 0x66001ff6, 0x66001ff6},
405 {0x0000a56c, 0x9c1fff0b, 0x9c1fff0b, 0x66001ff6, 0x66001ff6},
406 {0x0000a570, 0x9c1fff0b, 0x9c1fff0b, 0x66001ff6, 0x66001ff6},
407 {0x0000a574, 0x9c1fff0b, 0x9c1fff0b, 0x66001ff6, 0x66001ff6},
408 {0x0000a578, 0x9c1fff0b, 0x9c1fff0b, 0x66001ff6, 0x66001ff6},
409 {0x0000a57c, 0x9c1fff0b, 0x9c1fff0b, 0x66001ff6, 0x66001ff6},
410 {0x0000a580, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
411 {0x0000a584, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
412 {0x0000a588, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
413 {0x0000a58c, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
414 {0x0000a590, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
415 {0x0000a594, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
416 {0x0000a598, 0x00000000, 0x00000000, 0x01404501, 0x01404501},
417 {0x0000a59c, 0x00000000, 0x00000000, 0x02808a02, 0x02808a02},
418 {0x0000a5a0, 0x00000000, 0x00000000, 0x02808a02, 0x02808a02},
419 {0x0000a5a4, 0x00000000, 0x00000000, 0x02808803, 0x02808803},
420 {0x0000a5a8, 0x00000000, 0x00000000, 0x04c14b04, 0x04c14b04},
421 {0x0000a5ac, 0x00000000, 0x00000000, 0x04c15305, 0x04c15305},
422 {0x0000a5b0, 0x00000000, 0x00000000, 0x04c15305, 0x04c15305},
423 {0x0000a5b4, 0x00000000, 0x00000000, 0x04c15305, 0x04c15305},
424 {0x0000a5b8, 0x00000000, 0x00000000, 0x04c15305, 0x04c15305},
425 {0x0000a5bc, 0x00000000, 0x00000000, 0x04c15305, 0x04c15305},
426 {0x0000b500, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
427 {0x0000b504, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
428 {0x0000b508, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
429 {0x0000b50c, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
430 {0x0000b510, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
431 {0x0000b514, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
432 {0x0000b518, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
433 {0x0000b51c, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
434 {0x0000b520, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
435 {0x0000b524, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
436 {0x0000b528, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
437 {0x0000b52c, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
438 {0x0000b530, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
439 {0x0000b534, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
440 {0x0000b538, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
441 {0x0000b53c, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
442 {0x0000b540, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
443 {0x0000b544, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
444 {0x0000b548, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
445 {0x0000b54c, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
446 {0x0000b550, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
447 {0x0000b554, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
448 {0x0000b558, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
449 {0x0000b55c, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
450 {0x0000b560, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
451 {0x0000b564, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
452 {0x0000b568, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
453 {0x0000b56c, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
454 {0x0000b570, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
455 {0x0000b574, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
456 {0x0000b578, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
457 {0x0000b57c, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
458 {0x00016044, 0x05d6b2db, 0x05d6b2db, 0x05d6b2db, 0x05d6b2db},
459 {0x00016048, 0x6c924260, 0x6c924260, 0x6c924260, 0x6c924260},
460};
461
462static const u32 ar9485Modes_low_ob_db_tx_gain_1_1[][5] = {
463 /* Addr 5G_HT20 5G_HT40 2G_HT40 2G_HT20 */
464 {0x000098bc, 0x00000002, 0x00000002, 0x00000002, 0x00000002},
465 {0x0000a2d8, 0xf999a83a, 0xf999a83a, 0x7999a83a, 0x7999a83a},
466 {0x0000a2dc, 0x00000000, 0x00000000, 0xfe2d3552, 0xfe2d3552},
467 {0x0000a2e0, 0x00000000, 0x00000000, 0xfe2d3552, 0xfe2d3552},
468 {0x0000a2e4, 0x00000000, 0x00000000, 0xfe2d3552, 0xfe2d3552},
469 {0x0000a2e8, 0x00000000, 0x00000000, 0xfe2d3552, 0xfe2d3552},
344 {0x0000a410, 0x000050d9, 0x000050d9, 0x000050d8, 0x000050d8}, 470 {0x0000a410, 0x000050d9, 0x000050d9, 0x000050d8, 0x000050d8},
345 {0x0000a458, 0x00000000, 0x00000000, 0x00000000, 0x00000000}, 471 {0x0000a458, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
346 {0x0000a500, 0x00022200, 0x00022200, 0x00000000, 0x00000000}, 472 {0x0000a500, 0x00022200, 0x00022200, 0x00000000, 0x00000000},
@@ -427,7 +553,7 @@ static const u32 ar9485Modes_high_ob_db_tx_gain_1_1[][5] = {
427 {0x00016048, 0x6c924260, 0x6c924260, 0x6c924260, 0x6c924260}, 553 {0x00016048, 0x6c924260, 0x6c924260, 0x6c924260, 0x6c924260},
428}; 554};
429 555
430static const u32 ar9485Modes_low_ob_db_tx_gain_1_1[][5] = { 556static const u32 ar9485_modes_lowest_ob_db_tx_gain_1_1[][5] = {
431 /* Addr 5G_HT20 5G_HT40 2G_HT40 2G_HT20 */ 557 /* Addr 5G_HT20 5G_HT40 2G_HT40 2G_HT20 */
432 {0x000098bc, 0x00000002, 0x00000002, 0x00000002, 0x00000002}, 558 {0x000098bc, 0x00000002, 0x00000002, 0x00000002, 0x00000002},
433 {0x0000a2d8, 0xf999a83a, 0xf999a83a, 0x7999a83a, 0x7999a83a}, 559 {0x0000a2d8, 0xf999a83a, 0xf999a83a, 0x7999a83a, 0x7999a83a},
@@ -521,12 +647,15 @@ static const u32 ar9485Modes_low_ob_db_tx_gain_1_1[][5] = {
521 {0x00016048, 0x6c924260, 0x6c924260, 0x6c924260, 0x6c924260}, 647 {0x00016048, 0x6c924260, 0x6c924260, 0x6c924260, 0x6c924260},
522}; 648};
523 649
524#define ar9485_modes_lowest_ob_db_tx_gain_1_1 ar9485Modes_low_ob_db_tx_gain_1_1
525
526static const u32 ar9485Modes_green_spur_ob_db_tx_gain_1_1[][5] = { 650static const u32 ar9485Modes_green_spur_ob_db_tx_gain_1_1[][5] = {
527 /* Addr 5G_HT20 5G_HT40 2G_HT40 2G_HT20 */ 651 /* Addr 5G_HT20 5G_HT40 2G_HT40 2G_HT20 */
528 {0x000098bc, 0x00000003, 0x00000003, 0x00000003, 0x00000003}, 652 {0x000098bc, 0x00000003, 0x00000003, 0x00000003, 0x00000003},
529 {0x0000a410, 0x000050d9, 0x000050d9, 0x000050d8, 0x000050d8}, 653 {0x0000a2d8, 0xf999a83a, 0xf999a83a, 0x7999a83a, 0x7999a83a},
654 {0x0000a2dc, 0x00000000, 0x00000000, 0xffad452a, 0xffad452a},
655 {0x0000a2e0, 0x00000000, 0x00000000, 0xffc98634, 0xffc98634},
656 {0x0000a2e4, 0x00000000, 0x00000000, 0xfff60780, 0xfff60780},
657 {0x0000a2e8, 0x00000000, 0x00000000, 0xfffff800, 0xfffff800},
658 {0x0000a410, 0x000050d9, 0x000050d9, 0x000050d9, 0x000050d9},
530 {0x0000a458, 0x80000000, 0x80000000, 0x80000000, 0x80000000}, 659 {0x0000a458, 0x80000000, 0x80000000, 0x80000000, 0x80000000},
531 {0x0000a500, 0x00022200, 0x00022200, 0x00000006, 0x00000006}, 660 {0x0000a500, 0x00022200, 0x00022200, 0x00000006, 0x00000006},
532 {0x0000a504, 0x05062002, 0x05062002, 0x03000201, 0x03000201}, 661 {0x0000a504, 0x05062002, 0x05062002, 0x03000201, 0x03000201},
@@ -543,23 +672,39 @@ static const u32 ar9485Modes_green_spur_ob_db_tx_gain_1_1[][5] = {
543 {0x0000a530, 0x48023ec6, 0x48023ec6, 0x310006e0, 0x310006e0}, 672 {0x0000a530, 0x48023ec6, 0x48023ec6, 0x310006e0, 0x310006e0},
544 {0x0000a534, 0x4d023f01, 0x4d023f01, 0x330006e0, 0x330006e0}, 673 {0x0000a534, 0x4d023f01, 0x4d023f01, 0x330006e0, 0x330006e0},
545 {0x0000a538, 0x53023f4b, 0x53023f4b, 0x3e0008e3, 0x3e0008e3}, 674 {0x0000a538, 0x53023f4b, 0x53023f4b, 0x3e0008e3, 0x3e0008e3},
546 {0x0000a53c, 0x5a027f09, 0x5a027f09, 0x410008e5, 0x410008e5}, 675 {0x0000a53c, 0x5a027f09, 0x5a027f09, 0x430008e6, 0x430008e6},
547 {0x0000a540, 0x5f027fc9, 0x5f027fc9, 0x430008e6, 0x430008e6}, 676 {0x0000a540, 0x5f027fc9, 0x5f027fc9, 0x4a0008ec, 0x4a0008ec},
548 {0x0000a544, 0x6502feca, 0x6502feca, 0x4a0008ec, 0x4a0008ec}, 677 {0x0000a544, 0x6502feca, 0x6502feca, 0x4e0008f1, 0x4e0008f1},
549 {0x0000a548, 0x6b02ff4a, 0x6b02ff4a, 0x4e0008f1, 0x4e0008f1}, 678 {0x0000a548, 0x6b02ff4a, 0x6b02ff4a, 0x520008f3, 0x520008f3},
550 {0x0000a54c, 0x7203feca, 0x7203feca, 0x520008f3, 0x520008f3}, 679 {0x0000a54c, 0x7203feca, 0x7203feca, 0x54000eed, 0x54000eed},
551 {0x0000a550, 0x7703ff0b, 0x7703ff0b, 0x54000eed, 0x54000eed}, 680 {0x0000a550, 0x7703ff0b, 0x7703ff0b, 0x58000ef1, 0x58000ef1},
552 {0x0000a554, 0x7d06ffcb, 0x7d06ffcb, 0x58000ef1, 0x58000ef1}, 681 {0x0000a554, 0x7d06ffcb, 0x7d06ffcb, 0x5c000ef3, 0x5c000ef3},
553 {0x0000a558, 0x8407ff0b, 0x8407ff0b, 0x5c000ef3, 0x5c000ef3}, 682 {0x0000a558, 0x8407ff0b, 0x8407ff0b, 0x62000ef6, 0x62000ef6},
554 {0x0000a55c, 0x8907ffcb, 0x8907ffcb, 0x60000ef5, 0x60000ef5}, 683 {0x0000a55c, 0x8907ffcb, 0x8907ffcb, 0x66001ff0, 0x66001ff0},
555 {0x0000a560, 0x900fff0b, 0x900fff0b, 0x62000ef6, 0x62000ef6}, 684 {0x0000a560, 0x900fff0b, 0x900fff0b, 0x68001ff6, 0x68001ff6},
556 {0x0000a564, 0x960fffcb, 0x960fffcb, 0x62000ef6, 0x62000ef6}, 685 {0x0000a564, 0x960fffcb, 0x960fffcb, 0x68001ff6, 0x68001ff6},
557 {0x0000a568, 0x9c1fff0b, 0x9c1fff0b, 0x62000ef6, 0x62000ef6}, 686 {0x0000a568, 0x9c1fff0b, 0x9c1fff0b, 0x68001ff6, 0x68001ff6},
558 {0x0000a56c, 0x9c1fff0b, 0x9c1fff0b, 0x62000ef6, 0x62000ef6}, 687 {0x0000a56c, 0x9c1fff0b, 0x9c1fff0b, 0x68001ff6, 0x68001ff6},
559 {0x0000a570, 0x9c1fff0b, 0x9c1fff0b, 0x62000ef6, 0x62000ef6}, 688 {0x0000a570, 0x9c1fff0b, 0x9c1fff0b, 0x68001ff6, 0x68001ff6},
560 {0x0000a574, 0x9c1fff0b, 0x9c1fff0b, 0x62000ef6, 0x62000ef6}, 689 {0x0000a574, 0x9c1fff0b, 0x9c1fff0b, 0x68001ff6, 0x68001ff6},
561 {0x0000a578, 0x9c1fff0b, 0x9c1fff0b, 0x62000ef6, 0x62000ef6}, 690 {0x0000a578, 0x9c1fff0b, 0x9c1fff0b, 0x68001ff6, 0x68001ff6},
562 {0x0000a57c, 0x9c1fff0b, 0x9c1fff0b, 0x62000ef6, 0x62000ef6}, 691 {0x0000a57c, 0x9c1fff0b, 0x9c1fff0b, 0x68001ff6, 0x68001ff6},
692 {0x0000a580, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
693 {0x0000a584, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
694 {0x0000a588, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
695 {0x0000a58c, 0x00000000, 0x00000000, 0x01804000, 0x01804000},
696 {0x0000a590, 0x00000000, 0x00000000, 0x02808a02, 0x02808a02},
697 {0x0000a594, 0x00000000, 0x00000000, 0x0340ca02, 0x0340ca02},
698 {0x0000a598, 0x00000000, 0x00000000, 0x0340cd03, 0x0340cd03},
699 {0x0000a59c, 0x00000000, 0x00000000, 0x0340cd03, 0x0340cd03},
700 {0x0000a5a0, 0x00000000, 0x00000000, 0x06415304, 0x06415304},
701 {0x0000a5a4, 0x00000000, 0x00000000, 0x04c11905, 0x04c11905},
702 {0x0000a5a8, 0x00000000, 0x00000000, 0x06415905, 0x06415905},
703 {0x0000a5ac, 0x00000000, 0x00000000, 0x06415905, 0x06415905},
704 {0x0000a5b0, 0x00000000, 0x00000000, 0x06415905, 0x06415905},
705 {0x0000a5b4, 0x00000000, 0x00000000, 0x06415905, 0x06415905},
706 {0x0000a5b8, 0x00000000, 0x00000000, 0x06415905, 0x06415905},
707 {0x0000a5bc, 0x00000000, 0x00000000, 0x06415905, 0x06415905},
563 {0x0000b500, 0x0000001a, 0x0000001a, 0x0000001a, 0x0000001a}, 708 {0x0000b500, 0x0000001a, 0x0000001a, 0x0000001a, 0x0000001a},
564 {0x0000b504, 0x0000001a, 0x0000001a, 0x0000001a, 0x0000001a}, 709 {0x0000b504, 0x0000001a, 0x0000001a, 0x0000001a, 0x0000001a},
565 {0x0000b508, 0x0000001a, 0x0000001a, 0x0000001a, 0x0000001a}, 710 {0x0000b508, 0x0000001a, 0x0000001a, 0x0000001a, 0x0000001a},
@@ -823,6 +968,7 @@ static const u32 ar9485_common_rx_gain_1_1[][2] = {
823 {0x00009e00, 0x03721b20}, 968 {0x00009e00, 0x03721b20},
824 {0x00009e04, 0x00082020}, 969 {0x00009e04, 0x00082020},
825 {0x00009e18, 0x0300501e}, 970 {0x00009e18, 0x0300501e},
971 {0x00009e20, 0x000003ba},
826 {0x00009e2c, 0x00002e21}, 972 {0x00009e2c, 0x00002e21},
827 {0x00009e44, 0x02182324}, 973 {0x00009e44, 0x02182324},
828 {0x0000a000, 0x00060005}, 974 {0x0000a000, 0x00060005},
@@ -1001,7 +1147,6 @@ static const u32 ar9485_1_1_baseband_postamble[][5] = {
1001 {0x00009e10, 0x7ec88d2e, 0x7ec88d2e, 0x7ec80d2e, 0x7ec80d2e}, 1147 {0x00009e10, 0x7ec88d2e, 0x7ec88d2e, 0x7ec80d2e, 0x7ec80d2e},
1002 {0x00009e14, 0x31395d53, 0x31396053, 0x312e6053, 0x312e5d53}, 1148 {0x00009e14, 0x31395d53, 0x31396053, 0x312e6053, 0x312e5d53},
1003 {0x00009e1c, 0x0001cf9c, 0x0001cf9c, 0x00021f9c, 0x00021f9c}, 1149 {0x00009e1c, 0x0001cf9c, 0x0001cf9c, 0x00021f9c, 0x00021f9c},
1004 {0x00009e20, 0x000003b5, 0x000003b5, 0x000003ce, 0x000003ce},
1005 {0x00009e3c, 0xcf946220, 0xcf946220, 0xcf946222, 0xcf946222}, 1150 {0x00009e3c, 0xcf946220, 0xcf946220, 0xcf946222, 0xcf946222},
1006 {0x00009e48, 0x5030201a, 0x5030201a, 0x50302010, 0x50302010}, 1151 {0x00009e48, 0x5030201a, 0x5030201a, 0x50302010, 0x50302010},
1007 {0x00009fc8, 0x0003f000, 0x0003f000, 0x0001a000, 0x0001a000}, 1152 {0x00009fc8, 0x0003f000, 0x0003f000, 0x0001a000, 0x0001a000},
@@ -1020,7 +1165,7 @@ static const u32 ar9485_1_1_baseband_postamble[][5] = {
1020 {0x0000a284, 0x00000000, 0x00000000, 0x000002a0, 0x000002a0}, 1165 {0x0000a284, 0x00000000, 0x00000000, 0x000002a0, 0x000002a0},
1021 {0x0000a288, 0x00000000, 0x00000000, 0x00000000, 0x00000000}, 1166 {0x0000a288, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
1022 {0x0000a28c, 0x00000000, 0x00000000, 0x00000000, 0x00000000}, 1167 {0x0000a28c, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
1023 {0x0000a2c4, 0x00158d18, 0x00158d18, 0x00058d18, 0x00058d18}, 1168 {0x0000a2c4, 0x00158d18, 0x00158d18, 0x00158d18, 0x00158d18},
1024 {0x0000a2d0, 0x00071981, 0x00071981, 0x00071982, 0x00071982}, 1169 {0x0000a2d0, 0x00071981, 0x00071981, 0x00071982, 0x00071982},
1025 {0x0000a2d8, 0xf999a83a, 0xf999a83a, 0xf999a83a, 0xf999a83a}, 1170 {0x0000a2d8, 0xf999a83a, 0xf999a83a, 0xf999a83a, 0xf999a83a},
1026 {0x0000a358, 0x00000000, 0x00000000, 0x00000000, 0x00000000}, 1171 {0x0000a358, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
@@ -1206,6 +1351,11 @@ static const u32 ar9485_1_1_mac_core[][2] = {
1206 {0x000083d0, 0x000301ff}, 1351 {0x000083d0, 0x000301ff},
1207}; 1352};
1208 1353
1209#define ar9485_1_1_baseband_core_txfir_coeff_japan_2484 ar9462_2p0_baseband_core_txfir_coeff_japan_2484 1354static const u32 ar9485_1_1_baseband_core_txfir_coeff_japan_2484[][2] = {
1355 /* Addr allmodes */
1356 {0x0000a398, 0x00000000},
1357 {0x0000a39c, 0x6f7f0301},
1358 {0x0000a3a0, 0xca9228ee},
1359};
1210 1360
1211#endif /* INITVALS_9485_H */ 1361#endif /* INITVALS_9485_H */
diff --git a/drivers/net/wireless/ath/ath9k/ar9565_1p0_initvals.h b/drivers/net/wireless/ath/ath9k/ar9565_1p0_initvals.h
index e85a8b076c22..a8c757b6124f 100644
--- a/drivers/net/wireless/ath/ath9k/ar9565_1p0_initvals.h
+++ b/drivers/net/wireless/ath/ath9k/ar9565_1p0_initvals.h
@@ -272,9 +272,9 @@ static const u32 ar9565_1p0_baseband_core[][2] = {
272 {0x0000a398, 0x001f0e0f}, 272 {0x0000a398, 0x001f0e0f},
273 {0x0000a39c, 0x0075393f}, 273 {0x0000a39c, 0x0075393f},
274 {0x0000a3a0, 0xb79f6427}, 274 {0x0000a3a0, 0xb79f6427},
275 {0x0000a3a4, 0x00000000}, 275 {0x0000a3a4, 0x00000011},
276 {0x0000a3a8, 0xaaaaaaaa}, 276 {0x0000a3a8, 0xaaaaaa6e},
277 {0x0000a3ac, 0x3c466478}, 277 {0x0000a3ac, 0x3c466455},
278 {0x0000a3c0, 0x20202020}, 278 {0x0000a3c0, 0x20202020},
279 {0x0000a3c4, 0x22222220}, 279 {0x0000a3c4, 0x22222220},
280 {0x0000a3c8, 0x20200020}, 280 {0x0000a3c8, 0x20200020},
@@ -295,11 +295,11 @@ static const u32 ar9565_1p0_baseband_core[][2] = {
295 {0x0000a404, 0x00000000}, 295 {0x0000a404, 0x00000000},
296 {0x0000a408, 0x0e79e5c6}, 296 {0x0000a408, 0x0e79e5c6},
297 {0x0000a40c, 0x00820820}, 297 {0x0000a40c, 0x00820820},
298 {0x0000a414, 0x1ce739ce}, 298 {0x0000a414, 0x1ce739c5},
299 {0x0000a418, 0x2d001dce}, 299 {0x0000a418, 0x2d001dce},
300 {0x0000a41c, 0x1ce739ce}, 300 {0x0000a41c, 0x1ce739c5},
301 {0x0000a420, 0x000001ce}, 301 {0x0000a420, 0x000001ce},
302 {0x0000a424, 0x1ce739ce}, 302 {0x0000a424, 0x1ce739c5},
303 {0x0000a428, 0x000001ce}, 303 {0x0000a428, 0x000001ce},
304 {0x0000a42c, 0x1ce739ce}, 304 {0x0000a42c, 0x1ce739ce},
305 {0x0000a430, 0x1ce739ce}, 305 {0x0000a430, 0x1ce739ce},
@@ -351,9 +351,9 @@ static const u32 ar9565_1p0_baseband_postamble[][5] = {
351 {0x00009e14, 0x37b95d5e, 0x37b9605e, 0x3379605e, 0x33795d5e}, 351 {0x00009e14, 0x37b95d5e, 0x37b9605e, 0x3379605e, 0x33795d5e},
352 {0x00009e18, 0x00000000, 0x00000000, 0x00000000, 0x00000000}, 352 {0x00009e18, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
353 {0x00009e1c, 0x0001cf9c, 0x0001cf9c, 0x00021f9c, 0x00021f9c}, 353 {0x00009e1c, 0x0001cf9c, 0x0001cf9c, 0x00021f9c, 0x00021f9c},
354 {0x00009e20, 0x000003b5, 0x000003b5, 0x000003ce, 0x000003ce}, 354 {0x00009e20, 0x000003b5, 0x000003b5, 0x000003a4, 0x000003a4},
355 {0x00009e2c, 0x0000001c, 0x0000001c, 0x00000021, 0x00000021}, 355 {0x00009e2c, 0x0000001c, 0x0000001c, 0x00000021, 0x00000021},
356 {0x00009e3c, 0xcf946222, 0xcf946222, 0xcf946222, 0xcf946222}, 356 {0x00009e3c, 0xcf946222, 0xcf946222, 0xcf946220, 0xcf946220},
357 {0x00009e44, 0xfe321e27, 0xfe321e27, 0xfe291e27, 0xfe291e27}, 357 {0x00009e44, 0xfe321e27, 0xfe321e27, 0xfe291e27, 0xfe291e27},
358 {0x00009e48, 0x5030201a, 0x5030201a, 0x50302012, 0x50302012}, 358 {0x00009e48, 0x5030201a, 0x5030201a, 0x50302012, 0x50302012},
359 {0x00009fc8, 0x0003f000, 0x0003f000, 0x0001a000, 0x0001a000}, 359 {0x00009fc8, 0x0003f000, 0x0003f000, 0x0001a000, 0x0001a000},
@@ -452,6 +452,7 @@ static const u32 ar9565_1p0_Common_rx_gain_table[][2] = {
452 /* Addr allmodes */ 452 /* Addr allmodes */
453 {0x00004050, 0x00300300}, 453 {0x00004050, 0x00300300},
454 {0x0000406c, 0x00100000}, 454 {0x0000406c, 0x00100000},
455 {0x00009e20, 0x000003b6},
455 {0x0000a000, 0x00010000}, 456 {0x0000a000, 0x00010000},
456 {0x0000a004, 0x00030002}, 457 {0x0000a004, 0x00030002},
457 {0x0000a008, 0x00050004}, 458 {0x0000a008, 0x00050004},
@@ -1230,4 +1231,11 @@ static const u32 ar9565_1p0_modes_high_power_tx_gain_table[][5] = {
1230 {0x00016054, 0x00000000, 0x00000000, 0x00000000, 0x00000000}, 1231 {0x00016054, 0x00000000, 0x00000000, 0x00000000, 0x00000000},
1231}; 1232};
1232 1233
1234static const u32 ar9565_1p0_baseband_core_txfir_coeff_japan_2484[][2] = {
1235 /* Addr allmodes */
1236 {0x0000a398, 0x00000000},
1237 {0x0000a39c, 0x6f7f0301},
1238 {0x0000a3a0, 0xca9228ee},
1239};
1240
1233#endif /* INITVALS_9565_1P0_H */ 1241#endif /* INITVALS_9565_1P0_H */
diff --git a/drivers/net/wireless/ath/ath9k/ath9k.h b/drivers/net/wireless/ath/ath9k/ath9k.h
index 2ee35f677c0e..e7a38d844a6a 100644
--- a/drivers/net/wireless/ath/ath9k/ath9k.h
+++ b/drivers/net/wireless/ath/ath9k/ath9k.h
@@ -64,7 +64,6 @@ struct ath_node;
64 64
65struct ath_config { 65struct ath_config {
66 u16 txpowlimit; 66 u16 txpowlimit;
67 u8 cabqReadytime;
68}; 67};
69 68
70/*************************/ 69/*************************/
@@ -207,6 +206,14 @@ struct ath_frame_info {
207 u8 baw_tracked : 1; 206 u8 baw_tracked : 1;
208}; 207};
209 208
209struct ath_rxbuf {
210 struct list_head list;
211 struct sk_buff *bf_mpdu;
212 void *bf_desc;
213 dma_addr_t bf_daddr;
214 dma_addr_t bf_buf_addr;
215};
216
210struct ath_buf_state { 217struct ath_buf_state {
211 u8 bf_type; 218 u8 bf_type;
212 u8 bfs_paprd; 219 u8 bfs_paprd;
@@ -307,7 +314,7 @@ struct ath_rx {
307 struct ath_descdma rxdma; 314 struct ath_descdma rxdma;
308 struct ath_rx_edma rx_edma[ATH9K_RX_QUEUE_MAX]; 315 struct ath_rx_edma rx_edma[ATH9K_RX_QUEUE_MAX];
309 316
310 struct ath_buf *buf_hold; 317 struct ath_rxbuf *buf_hold;
311 struct sk_buff *frag; 318 struct sk_buff *frag;
312 319
313 u32 ampdu_ref; 320 u32 ampdu_ref;
@@ -459,8 +466,8 @@ void ath9k_queue_reset(struct ath_softc *sc, enum ath_reset_type type);
459 466
460#define ATH_DUMP_BTCOEX(_s, _val) \ 467#define ATH_DUMP_BTCOEX(_s, _val) \
461 do { \ 468 do { \
462 len += snprintf(buf + len, size - len, \ 469 len += scnprintf(buf + len, size - len, \
463 "%20s : %10d\n", _s, (_val)); \ 470 "%20s : %10d\n", _s, (_val)); \
464 } while (0) 471 } while (0)
465 472
466enum bt_op_flags { 473enum bt_op_flags {
@@ -581,7 +588,6 @@ static inline void ath_fill_led_pin(struct ath_softc *sc)
581#define ATH_ANT_DIV_COMB_ALT_ANT_RATIO_LOW_RSSI 50 588#define ATH_ANT_DIV_COMB_ALT_ANT_RATIO_LOW_RSSI 50
582#define ATH_ANT_DIV_COMB_ALT_ANT_RATIO2_LOW_RSSI 50 589#define ATH_ANT_DIV_COMB_ALT_ANT_RATIO2_LOW_RSSI 50
583 590
584#define ATH_ANT_DIV_COMB_LNA1_LNA2_SWITCH_DELTA -1
585#define ATH_ANT_DIV_COMB_LNA1_DELTA_HI -4 591#define ATH_ANT_DIV_COMB_LNA1_DELTA_HI -4
586#define ATH_ANT_DIV_COMB_LNA1_DELTA_MID -2 592#define ATH_ANT_DIV_COMB_LNA1_DELTA_MID -2
587#define ATH_ANT_DIV_COMB_LNA1_DELTA_LOW 2 593#define ATH_ANT_DIV_COMB_LNA1_DELTA_LOW 2
@@ -626,12 +632,15 @@ void ath_ant_comb_scan(struct ath_softc *sc, struct ath_rx_status *rs);
626/* Main driver core */ 632/* Main driver core */
627/********************/ 633/********************/
628 634
629#define ATH9K_PCI_CUS198 0x0001 635#define ATH9K_PCI_CUS198 0x0001
630#define ATH9K_PCI_CUS230 0x0002 636#define ATH9K_PCI_CUS230 0x0002
631#define ATH9K_PCI_CUS217 0x0004 637#define ATH9K_PCI_CUS217 0x0004
632#define ATH9K_PCI_WOW 0x0008 638#define ATH9K_PCI_CUS252 0x0008
633#define ATH9K_PCI_BT_ANT_DIV 0x0010 639#define ATH9K_PCI_WOW 0x0010
634#define ATH9K_PCI_D3_L1_WAR 0x0020 640#define ATH9K_PCI_BT_ANT_DIV 0x0020
641#define ATH9K_PCI_D3_L1_WAR 0x0040
642#define ATH9K_PCI_AR9565_1ANT 0x0080
643#define ATH9K_PCI_AR9565_2ANT 0x0100
635 644
636/* 645/*
637 * Default cache line size, in bytes. 646 * Default cache line size, in bytes.
@@ -769,6 +778,11 @@ struct ath_softc {
769 enum spectral_mode spectral_mode; 778 enum spectral_mode spectral_mode;
770 struct ath_spec_scan spec_config; 779 struct ath_spec_scan spec_config;
771 780
781 struct ieee80211_vif *tx99_vif;
782 struct sk_buff *tx99_skb;
783 bool tx99_state;
784 s16 tx99_power;
785
772#ifdef CONFIG_PM_SLEEP 786#ifdef CONFIG_PM_SLEEP
773 atomic_t wow_got_bmiss_intr; 787 atomic_t wow_got_bmiss_intr;
774 atomic_t wow_sleep_proc_intr; /* in the middle of WoW sleep ? */ 788 atomic_t wow_sleep_proc_intr; /* in the middle of WoW sleep ? */
@@ -877,6 +891,7 @@ static inline u8 spectral_bitmap_weight(u8 *bins)
877 */ 891 */
878enum ath_fft_sample_type { 892enum ath_fft_sample_type {
879 ATH_FFT_SAMPLE_HT20 = 1, 893 ATH_FFT_SAMPLE_HT20 = 1,
894 ATH_FFT_SAMPLE_HT20_40,
880}; 895};
881 896
882struct fft_sample_tlv { 897struct fft_sample_tlv {
@@ -903,6 +918,39 @@ struct fft_sample_ht20 {
903 u8 data[SPECTRAL_HT20_NUM_BINS]; 918 u8 data[SPECTRAL_HT20_NUM_BINS];
904} __packed; 919} __packed;
905 920
921struct fft_sample_ht20_40 {
922 struct fft_sample_tlv tlv;
923
924 u8 channel_type;
925 __be16 freq;
926
927 s8 lower_rssi;
928 s8 upper_rssi;
929
930 __be64 tsf;
931
932 s8 lower_noise;
933 s8 upper_noise;
934
935 __be16 lower_max_magnitude;
936 __be16 upper_max_magnitude;
937
938 u8 lower_max_index;
939 u8 upper_max_index;
940
941 u8 lower_bitmap_weight;
942 u8 upper_bitmap_weight;
943
944 u8 max_exp;
945
946 u8 data[SPECTRAL_HT20_40_NUM_BINS];
947} __packed;
948
949int ath9k_tx99_init(struct ath_softc *sc);
950void ath9k_tx99_deinit(struct ath_softc *sc);
951int ath9k_tx99_send(struct ath_softc *sc, struct sk_buff *skb,
952 struct ath_tx_control *txctl);
953
906void ath9k_tasklet(unsigned long data); 954void ath9k_tasklet(unsigned long data);
907int ath_cabq_update(struct ath_softc *); 955int ath_cabq_update(struct ath_softc *);
908 956
@@ -924,7 +972,6 @@ void ath9k_deinit_device(struct ath_softc *sc);
924void ath9k_set_hw_capab(struct ath_softc *sc, struct ieee80211_hw *hw); 972void ath9k_set_hw_capab(struct ath_softc *sc, struct ieee80211_hw *hw);
925void ath9k_reload_chainmask_settings(struct ath_softc *sc); 973void ath9k_reload_chainmask_settings(struct ath_softc *sc);
926 974
927bool ath9k_uses_beacons(int type);
928void ath9k_spectral_scan_trigger(struct ieee80211_hw *hw); 975void ath9k_spectral_scan_trigger(struct ieee80211_hw *hw);
929int ath9k_spectral_scan_config(struct ieee80211_hw *hw, 976int ath9k_spectral_scan_config(struct ieee80211_hw *hw,
930 enum spectral_mode spectral_mode); 977 enum spectral_mode spectral_mode);
@@ -952,7 +999,7 @@ void ath9k_ps_restore(struct ath_softc *sc);
952u8 ath_txchainmask_reduction(struct ath_softc *sc, u8 chainmask, u32 rate); 999u8 ath_txchainmask_reduction(struct ath_softc *sc, u8 chainmask, u32 rate);
953 1000
954void ath_start_rfkill_poll(struct ath_softc *sc); 1001void ath_start_rfkill_poll(struct ath_softc *sc);
955extern void ath9k_rfkill_poll_state(struct ieee80211_hw *hw); 1002void ath9k_rfkill_poll_state(struct ieee80211_hw *hw);
956void ath9k_calculate_iter_data(struct ieee80211_hw *hw, 1003void ath9k_calculate_iter_data(struct ieee80211_hw *hw,
957 struct ieee80211_vif *vif, 1004 struct ieee80211_vif *vif,
958 struct ath9k_vif_iter_data *iter_data); 1005 struct ath9k_vif_iter_data *iter_data);
diff --git a/drivers/net/wireless/ath/ath9k/beacon.c b/drivers/net/wireless/ath/ath9k/beacon.c
index b5c16b3a37b9..17be35392bb4 100644
--- a/drivers/net/wireless/ath/ath9k/beacon.c
+++ b/drivers/net/wireless/ath/ath9k/beacon.c
@@ -334,6 +334,8 @@ void ath9k_beacon_tasklet(unsigned long data)
334 if (ath9k_hw_numtxpending(ah, sc->beacon.beaconq) != 0) { 334 if (ath9k_hw_numtxpending(ah, sc->beacon.beaconq) != 0) {
335 sc->beacon.bmisscnt++; 335 sc->beacon.bmisscnt++;
336 336
337 ath9k_hw_check_nav(ah);
338
337 if (!ath9k_hw_check_alive(ah)) 339 if (!ath9k_hw_check_alive(ah))
338 ieee80211_queue_work(sc->hw, &sc->hw_check_work); 340 ieee80211_queue_work(sc->hw, &sc->hw_check_work);
339 341
diff --git a/drivers/net/wireless/ath/ath9k/calib.c b/drivers/net/wireless/ath/ath9k/calib.c
index 5e8219a91e25..278365b8a895 100644
--- a/drivers/net/wireless/ath/ath9k/calib.c
+++ b/drivers/net/wireless/ath/ath9k/calib.c
@@ -63,13 +63,13 @@ static s16 ath9k_hw_get_default_nf(struct ath_hw *ah,
63 return ath9k_hw_get_nf_limits(ah, chan)->nominal; 63 return ath9k_hw_get_nf_limits(ah, chan)->nominal;
64} 64}
65 65
66s16 ath9k_hw_getchan_noise(struct ath_hw *ah, struct ath9k_channel *chan) 66s16 ath9k_hw_getchan_noise(struct ath_hw *ah, struct ath9k_channel *chan,
67 s16 nf)
67{ 68{
68 s8 noise = ATH_DEFAULT_NOISE_FLOOR; 69 s8 noise = ATH_DEFAULT_NOISE_FLOOR;
69 70
70 if (chan && chan->noisefloor) { 71 if (nf) {
71 s8 delta = chan->noisefloor - 72 s8 delta = nf - ATH9K_NF_CAL_NOISE_THRESH -
72 ATH9K_NF_CAL_NOISE_THRESH -
73 ath9k_hw_get_default_nf(ah, chan); 73 ath9k_hw_get_default_nf(ah, chan);
74 if (delta > 0) 74 if (delta > 0)
75 noise += delta; 75 noise += delta;
@@ -119,7 +119,7 @@ static void ath9k_hw_update_nfcal_hist_buffer(struct ath_hw *ah,
119 ath_dbg(common, CALIBRATE, 119 ath_dbg(common, CALIBRATE,
120 "NFmid[%d] (%d) > MAX (%d), %s\n", 120 "NFmid[%d] (%d) > MAX (%d), %s\n",
121 i, h[i].privNF, limit->max, 121 i, h[i].privNF, limit->max,
122 (cal->nfcal_interference ? 122 (test_bit(NFCAL_INTF, &cal->cal_flags) ?
123 "not corrected (due to interference)" : 123 "not corrected (due to interference)" :
124 "correcting to MAX")); 124 "correcting to MAX"));
125 125
@@ -130,7 +130,7 @@ static void ath9k_hw_update_nfcal_hist_buffer(struct ath_hw *ah,
130 * we bypass this limit here in order to better deal 130 * we bypass this limit here in order to better deal
131 * with our environment. 131 * with our environment.
132 */ 132 */
133 if (!cal->nfcal_interference) 133 if (!test_bit(NFCAL_INTF, &cal->cal_flags))
134 h[i].privNF = limit->max; 134 h[i].privNF = limit->max;
135 } 135 }
136 } 136 }
@@ -141,7 +141,7 @@ static void ath9k_hw_update_nfcal_hist_buffer(struct ath_hw *ah,
141 * Re-enable the enforcement of the NF maximum again. 141 * Re-enable the enforcement of the NF maximum again.
142 */ 142 */
143 if (!high_nf_mid) 143 if (!high_nf_mid)
144 cal->nfcal_interference = false; 144 clear_bit(NFCAL_INTF, &cal->cal_flags);
145} 145}
146 146
147static bool ath9k_hw_get_nf_thresh(struct ath_hw *ah, 147static bool ath9k_hw_get_nf_thresh(struct ath_hw *ah,
@@ -186,7 +186,6 @@ void ath9k_hw_reset_calibration(struct ath_hw *ah,
186bool ath9k_hw_reset_calvalid(struct ath_hw *ah) 186bool ath9k_hw_reset_calvalid(struct ath_hw *ah)
187{ 187{
188 struct ath_common *common = ath9k_hw_common(ah); 188 struct ath_common *common = ath9k_hw_common(ah);
189 struct ieee80211_conf *conf = &common->hw->conf;
190 struct ath9k_cal_list *currCal = ah->cal_list_curr; 189 struct ath9k_cal_list *currCal = ah->cal_list_curr;
191 190
192 if (!ah->caldata) 191 if (!ah->caldata)
@@ -208,7 +207,7 @@ bool ath9k_hw_reset_calvalid(struct ath_hw *ah)
208 return true; 207 return true;
209 208
210 ath_dbg(common, CALIBRATE, "Resetting Cal %d state for channel %u\n", 209 ath_dbg(common, CALIBRATE, "Resetting Cal %d state for channel %u\n",
211 currCal->calData->calType, conf->chandef.chan->center_freq); 210 currCal->calData->calType, ah->curchan->chan->center_freq);
212 211
213 ah->caldata->CalValid &= ~currCal->calData->calType; 212 ah->caldata->CalValid &= ~currCal->calData->calType;
214 currCal->calState = CAL_WAITING; 213 currCal->calState = CAL_WAITING;
@@ -220,7 +219,7 @@ EXPORT_SYMBOL(ath9k_hw_reset_calvalid);
220void ath9k_hw_start_nfcal(struct ath_hw *ah, bool update) 219void ath9k_hw_start_nfcal(struct ath_hw *ah, bool update)
221{ 220{
222 if (ah->caldata) 221 if (ah->caldata)
223 ah->caldata->nfcal_pending = true; 222 set_bit(NFCAL_PENDING, &ah->caldata->cal_flags);
224 223
225 REG_SET_BIT(ah, AR_PHY_AGC_CONTROL, 224 REG_SET_BIT(ah, AR_PHY_AGC_CONTROL,
226 AR_PHY_AGC_CONTROL_ENABLE_NF); 225 AR_PHY_AGC_CONTROL_ENABLE_NF);
@@ -242,7 +241,6 @@ void ath9k_hw_loadnf(struct ath_hw *ah, struct ath9k_channel *chan)
242 int32_t val; 241 int32_t val;
243 u8 chainmask = (ah->rxchainmask << 3) | ah->rxchainmask; 242 u8 chainmask = (ah->rxchainmask << 3) | ah->rxchainmask;
244 struct ath_common *common = ath9k_hw_common(ah); 243 struct ath_common *common = ath9k_hw_common(ah);
245 struct ieee80211_conf *conf = &common->hw->conf;
246 s16 default_nf = ath9k_hw_get_default_nf(ah, chan); 244 s16 default_nf = ath9k_hw_get_default_nf(ah, chan);
247 245
248 if (ah->caldata) 246 if (ah->caldata)
@@ -252,7 +250,7 @@ void ath9k_hw_loadnf(struct ath_hw *ah, struct ath9k_channel *chan)
252 if (chainmask & (1 << i)) { 250 if (chainmask & (1 << i)) {
253 s16 nfval; 251 s16 nfval;
254 252
255 if ((i >= AR5416_MAX_CHAINS) && !conf_is_ht40(conf)) 253 if ((i >= AR5416_MAX_CHAINS) && !IS_CHAN_HT40(chan))
256 continue; 254 continue;
257 255
258 if (h) 256 if (h)
@@ -314,7 +312,7 @@ void ath9k_hw_loadnf(struct ath_hw *ah, struct ath9k_channel *chan)
314 ENABLE_REGWRITE_BUFFER(ah); 312 ENABLE_REGWRITE_BUFFER(ah);
315 for (i = 0; i < NUM_NF_READINGS; i++) { 313 for (i = 0; i < NUM_NF_READINGS; i++) {
316 if (chainmask & (1 << i)) { 314 if (chainmask & (1 << i)) {
317 if ((i >= AR5416_MAX_CHAINS) && !conf_is_ht40(conf)) 315 if ((i >= AR5416_MAX_CHAINS) && !IS_CHAN_HT40(chan))
318 continue; 316 continue;
319 317
320 val = REG_READ(ah, ah->nf_regs[i]); 318 val = REG_READ(ah, ah->nf_regs[i]);
@@ -391,10 +389,10 @@ bool ath9k_hw_getnf(struct ath_hw *ah, struct ath9k_channel *chan)
391 } 389 }
392 390
393 h = caldata->nfCalHist; 391 h = caldata->nfCalHist;
394 caldata->nfcal_pending = false; 392 clear_bit(NFCAL_PENDING, &caldata->cal_flags);
395 ath9k_hw_update_nfcal_hist_buffer(ah, caldata, nfarray); 393 ath9k_hw_update_nfcal_hist_buffer(ah, caldata, nfarray);
396 chan->noisefloor = h[0].privNF; 394 chan->noisefloor = h[0].privNF;
397 ah->noise = ath9k_hw_getchan_noise(ah, chan); 395 ah->noise = ath9k_hw_getchan_noise(ah, chan, chan->noisefloor);
398 return true; 396 return true;
399} 397}
400EXPORT_SYMBOL(ath9k_hw_getnf); 398EXPORT_SYMBOL(ath9k_hw_getnf);
@@ -408,7 +406,6 @@ void ath9k_init_nfcal_hist_buffer(struct ath_hw *ah,
408 406
409 ah->caldata->channel = chan->channel; 407 ah->caldata->channel = chan->channel;
410 ah->caldata->channelFlags = chan->channelFlags; 408 ah->caldata->channelFlags = chan->channelFlags;
411 ah->caldata->chanmode = chan->chanmode;
412 h = ah->caldata->nfCalHist; 409 h = ah->caldata->nfCalHist;
413 default_nf = ath9k_hw_get_default_nf(ah, chan); 410 default_nf = ath9k_hw_get_default_nf(ah, chan);
414 for (i = 0; i < NUM_NF_READINGS; i++) { 411 for (i = 0; i < NUM_NF_READINGS; i++) {
@@ -437,12 +434,12 @@ void ath9k_hw_bstuck_nfcal(struct ath_hw *ah)
437 * the baseband update the internal NF value itself, similar to 434 * the baseband update the internal NF value itself, similar to
438 * what is being done after a full reset. 435 * what is being done after a full reset.
439 */ 436 */
440 if (!caldata->nfcal_pending) 437 if (!test_bit(NFCAL_PENDING, &caldata->cal_flags))
441 ath9k_hw_start_nfcal(ah, true); 438 ath9k_hw_start_nfcal(ah, true);
442 else if (!(REG_READ(ah, AR_PHY_AGC_CONTROL) & AR_PHY_AGC_CONTROL_NF)) 439 else if (!(REG_READ(ah, AR_PHY_AGC_CONTROL) & AR_PHY_AGC_CONTROL_NF))
443 ath9k_hw_getnf(ah, ah->curchan); 440 ath9k_hw_getnf(ah, ah->curchan);
444 441
445 caldata->nfcal_interference = true; 442 set_bit(NFCAL_INTF, &caldata->cal_flags);
446} 443}
447EXPORT_SYMBOL(ath9k_hw_bstuck_nfcal); 444EXPORT_SYMBOL(ath9k_hw_bstuck_nfcal);
448 445
diff --git a/drivers/net/wireless/ath/ath9k/calib.h b/drivers/net/wireless/ath/ath9k/calib.h
index 3d70b8c2bcdd..b8ed95e9a335 100644
--- a/drivers/net/wireless/ath/ath9k/calib.h
+++ b/drivers/net/wireless/ath/ath9k/calib.h
@@ -116,7 +116,8 @@ void ath9k_init_nfcal_hist_buffer(struct ath_hw *ah,
116void ath9k_hw_bstuck_nfcal(struct ath_hw *ah); 116void ath9k_hw_bstuck_nfcal(struct ath_hw *ah);
117void ath9k_hw_reset_calibration(struct ath_hw *ah, 117void ath9k_hw_reset_calibration(struct ath_hw *ah,
118 struct ath9k_cal_list *currCal); 118 struct ath9k_cal_list *currCal);
119s16 ath9k_hw_getchan_noise(struct ath_hw *ah, struct ath9k_channel *chan); 119s16 ath9k_hw_getchan_noise(struct ath_hw *ah, struct ath9k_channel *chan,
120 s16 nf);
120 121
121 122
122#endif /* CALIB_H */ 123#endif /* CALIB_H */
diff --git a/drivers/net/wireless/ath/ath9k/common.c b/drivers/net/wireless/ath/ath9k/common.c
index d3063c21e16c..a7e5a05b2eff 100644
--- a/drivers/net/wireless/ath/ath9k/common.c
+++ b/drivers/net/wireless/ath/ath9k/common.c
@@ -49,103 +49,64 @@ int ath9k_cmn_get_hw_crypto_keytype(struct sk_buff *skb)
49} 49}
50EXPORT_SYMBOL(ath9k_cmn_get_hw_crypto_keytype); 50EXPORT_SYMBOL(ath9k_cmn_get_hw_crypto_keytype);
51 51
52static u32 ath9k_get_extchanmode(struct cfg80211_chan_def *chandef)
53{
54 u32 chanmode = 0;
55
56 switch (chandef->chan->band) {
57 case IEEE80211_BAND_2GHZ:
58 switch (chandef->width) {
59 case NL80211_CHAN_WIDTH_20_NOHT:
60 case NL80211_CHAN_WIDTH_20:
61 chanmode = CHANNEL_G_HT20;
62 break;
63 case NL80211_CHAN_WIDTH_40:
64 if (chandef->center_freq1 > chandef->chan->center_freq)
65 chanmode = CHANNEL_G_HT40PLUS;
66 else
67 chanmode = CHANNEL_G_HT40MINUS;
68 break;
69 default:
70 break;
71 }
72 break;
73 case IEEE80211_BAND_5GHZ:
74 switch (chandef->width) {
75 case NL80211_CHAN_WIDTH_20_NOHT:
76 case NL80211_CHAN_WIDTH_20:
77 chanmode = CHANNEL_A_HT20;
78 break;
79 case NL80211_CHAN_WIDTH_40:
80 if (chandef->center_freq1 > chandef->chan->center_freq)
81 chanmode = CHANNEL_A_HT40PLUS;
82 else
83 chanmode = CHANNEL_A_HT40MINUS;
84 break;
85 default:
86 break;
87 }
88 break;
89 default:
90 break;
91 }
92
93 return chanmode;
94}
95
96/* 52/*
97 * Update internal channel flags. 53 * Update internal channel flags.
98 */ 54 */
99void ath9k_cmn_update_ichannel(struct ath9k_channel *ichan, 55static void ath9k_cmn_update_ichannel(struct ath9k_channel *ichan,
100 struct cfg80211_chan_def *chandef) 56 struct cfg80211_chan_def *chandef)
101{ 57{
102 ichan->channel = chandef->chan->center_freq; 58 struct ieee80211_channel *chan = chandef->chan;
103 ichan->chan = chandef->chan; 59 u16 flags = 0;
104 60
105 if (chandef->chan->band == IEEE80211_BAND_2GHZ) { 61 ichan->channel = chan->center_freq;
106 ichan->chanmode = CHANNEL_G; 62 ichan->chan = chan;
107 ichan->channelFlags = CHANNEL_2GHZ | CHANNEL_OFDM; 63
108 } else { 64 if (chan->band == IEEE80211_BAND_5GHZ)
109 ichan->chanmode = CHANNEL_A; 65 flags |= CHANNEL_5GHZ;
110 ichan->channelFlags = CHANNEL_5GHZ | CHANNEL_OFDM;
111 }
112 66
113 switch (chandef->width) { 67 switch (chandef->width) {
114 case NL80211_CHAN_WIDTH_5: 68 case NL80211_CHAN_WIDTH_5:
115 ichan->channelFlags |= CHANNEL_QUARTER; 69 flags |= CHANNEL_QUARTER;
116 break; 70 break;
117 case NL80211_CHAN_WIDTH_10: 71 case NL80211_CHAN_WIDTH_10:
118 ichan->channelFlags |= CHANNEL_HALF; 72 flags |= CHANNEL_HALF;
119 break; 73 break;
120 case NL80211_CHAN_WIDTH_20_NOHT: 74 case NL80211_CHAN_WIDTH_20_NOHT:
121 break; 75 break;
122 case NL80211_CHAN_WIDTH_20: 76 case NL80211_CHAN_WIDTH_20:
77 flags |= CHANNEL_HT;
78 break;
123 case NL80211_CHAN_WIDTH_40: 79 case NL80211_CHAN_WIDTH_40:
124 ichan->chanmode = ath9k_get_extchanmode(chandef); 80 if (chandef->center_freq1 > chandef->chan->center_freq)
81 flags |= CHANNEL_HT40PLUS | CHANNEL_HT;
82 else
83 flags |= CHANNEL_HT40MINUS | CHANNEL_HT;
125 break; 84 break;
126 default: 85 default:
127 WARN_ON(1); 86 WARN_ON(1);
128 } 87 }
88
89 ichan->channelFlags = flags;
129} 90}
130EXPORT_SYMBOL(ath9k_cmn_update_ichannel);
131 91
132/* 92/*
133 * Get the internal channel reference. 93 * Get the internal channel reference.
134 */ 94 */
135struct ath9k_channel *ath9k_cmn_get_curchannel(struct ieee80211_hw *hw, 95struct ath9k_channel *ath9k_cmn_get_channel(struct ieee80211_hw *hw,
136 struct ath_hw *ah) 96 struct ath_hw *ah,
97 struct cfg80211_chan_def *chandef)
137{ 98{
138 struct ieee80211_channel *curchan = hw->conf.chandef.chan; 99 struct ieee80211_channel *curchan = chandef->chan;
139 struct ath9k_channel *channel; 100 struct ath9k_channel *channel;
140 u8 chan_idx; 101 u8 chan_idx;
141 102
142 chan_idx = curchan->hw_value; 103 chan_idx = curchan->hw_value;
143 channel = &ah->channels[chan_idx]; 104 channel = &ah->channels[chan_idx];
144 ath9k_cmn_update_ichannel(channel, &hw->conf.chandef); 105 ath9k_cmn_update_ichannel(channel, chandef);
145 106
146 return channel; 107 return channel;
147} 108}
148EXPORT_SYMBOL(ath9k_cmn_get_curchannel); 109EXPORT_SYMBOL(ath9k_cmn_get_channel);
149 110
150int ath9k_cmn_count_streams(unsigned int chainmask, int max) 111int ath9k_cmn_count_streams(unsigned int chainmask, int max)
151{ 112{
diff --git a/drivers/net/wireless/ath/ath9k/common.h b/drivers/net/wireless/ath/ath9k/common.h
index e039bcbfbd79..eb85e1bdca88 100644
--- a/drivers/net/wireless/ath/ath9k/common.h
+++ b/drivers/net/wireless/ath/ath9k/common.h
@@ -43,10 +43,9 @@
43 (((x) + ((mul)/2)) / (mul)) 43 (((x) + ((mul)/2)) / (mul))
44 44
45int ath9k_cmn_get_hw_crypto_keytype(struct sk_buff *skb); 45int ath9k_cmn_get_hw_crypto_keytype(struct sk_buff *skb);
46void ath9k_cmn_update_ichannel(struct ath9k_channel *ichan, 46struct ath9k_channel *ath9k_cmn_get_channel(struct ieee80211_hw *hw,
47 struct cfg80211_chan_def *chandef); 47 struct ath_hw *ah,
48struct ath9k_channel *ath9k_cmn_get_curchannel(struct ieee80211_hw *hw, 48 struct cfg80211_chan_def *chandef);
49 struct ath_hw *ah);
50int ath9k_cmn_count_streams(unsigned int chainmask, int max); 49int ath9k_cmn_count_streams(unsigned int chainmask, int max);
51void ath9k_cmn_btcoex_bt_stomp(struct ath_common *common, 50void ath9k_cmn_btcoex_bt_stomp(struct ath_common *common,
52 enum ath_stomp_type stomp_type); 51 enum ath_stomp_type stomp_type);
diff --git a/drivers/net/wireless/ath/ath9k/debug.c b/drivers/net/wireless/ath/ath9k/debug.c
index c088744a6bfb..83a2c59f680b 100644
--- a/drivers/net/wireless/ath/ath9k/debug.c
+++ b/drivers/net/wireless/ath/ath9k/debug.c
@@ -104,37 +104,37 @@ static ssize_t read_file_ani(struct file *file, char __user *user_buf,
104 return -ENOMEM; 104 return -ENOMEM;
105 105
106 if (common->disable_ani) { 106 if (common->disable_ani) {
107 len += snprintf(buf + len, size - len, "%s: %s\n", 107 len += scnprintf(buf + len, size - len, "%s: %s\n",
108 "ANI", "DISABLED"); 108 "ANI", "DISABLED");
109 goto exit; 109 goto exit;
110 } 110 }
111 111
112 len += snprintf(buf + len, size - len, "%15s: %s\n", 112 len += scnprintf(buf + len, size - len, "%15s: %s\n",
113 "ANI", "ENABLED"); 113 "ANI", "ENABLED");
114 len += snprintf(buf + len, size - len, "%15s: %u\n", 114 len += scnprintf(buf + len, size - len, "%15s: %u\n",
115 "ANI RESET", ah->stats.ast_ani_reset); 115 "ANI RESET", ah->stats.ast_ani_reset);
116 len += snprintf(buf + len, size - len, "%15s: %u\n", 116 len += scnprintf(buf + len, size - len, "%15s: %u\n",
117 "SPUR UP", ah->stats.ast_ani_spurup); 117 "SPUR UP", ah->stats.ast_ani_spurup);
118 len += snprintf(buf + len, size - len, "%15s: %u\n", 118 len += scnprintf(buf + len, size - len, "%15s: %u\n",
119 "SPUR DOWN", ah->stats.ast_ani_spurup); 119 "SPUR DOWN", ah->stats.ast_ani_spurup);
120 len += snprintf(buf + len, size - len, "%15s: %u\n", 120 len += scnprintf(buf + len, size - len, "%15s: %u\n",
121 "OFDM WS-DET ON", ah->stats.ast_ani_ofdmon); 121 "OFDM WS-DET ON", ah->stats.ast_ani_ofdmon);
122 len += snprintf(buf + len, size - len, "%15s: %u\n", 122 len += scnprintf(buf + len, size - len, "%15s: %u\n",
123 "OFDM WS-DET OFF", ah->stats.ast_ani_ofdmoff); 123 "OFDM WS-DET OFF", ah->stats.ast_ani_ofdmoff);
124 len += snprintf(buf + len, size - len, "%15s: %u\n", 124 len += scnprintf(buf + len, size - len, "%15s: %u\n",
125 "MRC-CCK ON", ah->stats.ast_ani_ccklow); 125 "MRC-CCK ON", ah->stats.ast_ani_ccklow);
126 len += snprintf(buf + len, size - len, "%15s: %u\n", 126 len += scnprintf(buf + len, size - len, "%15s: %u\n",
127 "MRC-CCK OFF", ah->stats.ast_ani_cckhigh); 127 "MRC-CCK OFF", ah->stats.ast_ani_cckhigh);
128 len += snprintf(buf + len, size - len, "%15s: %u\n", 128 len += scnprintf(buf + len, size - len, "%15s: %u\n",
129 "FIR-STEP UP", ah->stats.ast_ani_stepup); 129 "FIR-STEP UP", ah->stats.ast_ani_stepup);
130 len += snprintf(buf + len, size - len, "%15s: %u\n", 130 len += scnprintf(buf + len, size - len, "%15s: %u\n",
131 "FIR-STEP DOWN", ah->stats.ast_ani_stepdown); 131 "FIR-STEP DOWN", ah->stats.ast_ani_stepdown);
132 len += snprintf(buf + len, size - len, "%15s: %u\n", 132 len += scnprintf(buf + len, size - len, "%15s: %u\n",
133 "INV LISTENTIME", ah->stats.ast_ani_lneg_or_lzero); 133 "INV LISTENTIME", ah->stats.ast_ani_lneg_or_lzero);
134 len += snprintf(buf + len, size - len, "%15s: %u\n", 134 len += scnprintf(buf + len, size - len, "%15s: %u\n",
135 "OFDM ERRORS", ah->stats.ast_ani_ofdmerrs); 135 "OFDM ERRORS", ah->stats.ast_ani_ofdmerrs);
136 len += snprintf(buf + len, size - len, "%15s: %u\n", 136 len += scnprintf(buf + len, size - len, "%15s: %u\n",
137 "CCK ERRORS", ah->stats.ast_ani_cckerrs); 137 "CCK ERRORS", ah->stats.ast_ani_cckerrs);
138exit: 138exit:
139 if (len > size) 139 if (len > size)
140 len = size; 140 len = size;
@@ -280,70 +280,70 @@ static ssize_t read_file_antenna_diversity(struct file *file,
280 return -ENOMEM; 280 return -ENOMEM;
281 281
282 if (!(pCap->hw_caps & ATH9K_HW_CAP_ANT_DIV_COMB)) { 282 if (!(pCap->hw_caps & ATH9K_HW_CAP_ANT_DIV_COMB)) {
283 len += snprintf(buf + len, size - len, "%s\n", 283 len += scnprintf(buf + len, size - len, "%s\n",
284 "Antenna Diversity Combining is disabled"); 284 "Antenna Diversity Combining is disabled");
285 goto exit; 285 goto exit;
286 } 286 }
287 287
288 ath9k_ps_wakeup(sc); 288 ath9k_ps_wakeup(sc);
289 ath9k_hw_antdiv_comb_conf_get(ah, &div_ant_conf); 289 ath9k_hw_antdiv_comb_conf_get(ah, &div_ant_conf);
290 len += snprintf(buf + len, size - len, "Current MAIN config : %s\n", 290 len += scnprintf(buf + len, size - len, "Current MAIN config : %s\n",
291 lna_conf_str[div_ant_conf.main_lna_conf]); 291 lna_conf_str[div_ant_conf.main_lna_conf]);
292 len += snprintf(buf + len, size - len, "Current ALT config : %s\n", 292 len += scnprintf(buf + len, size - len, "Current ALT config : %s\n",
293 lna_conf_str[div_ant_conf.alt_lna_conf]); 293 lna_conf_str[div_ant_conf.alt_lna_conf]);
294 len += snprintf(buf + len, size - len, "Average MAIN RSSI : %d\n", 294 len += scnprintf(buf + len, size - len, "Average MAIN RSSI : %d\n",
295 as_main->rssi_avg); 295 as_main->rssi_avg);
296 len += snprintf(buf + len, size - len, "Average ALT RSSI : %d\n\n", 296 len += scnprintf(buf + len, size - len, "Average ALT RSSI : %d\n\n",
297 as_alt->rssi_avg); 297 as_alt->rssi_avg);
298 ath9k_ps_restore(sc); 298 ath9k_ps_restore(sc);
299 299
300 len += snprintf(buf + len, size - len, "Packet Receive Cnt:\n"); 300 len += scnprintf(buf + len, size - len, "Packet Receive Cnt:\n");
301 len += snprintf(buf + len, size - len, "-------------------\n"); 301 len += scnprintf(buf + len, size - len, "-------------------\n");
302 302
303 len += snprintf(buf + len, size - len, "%30s%15s\n", 303 len += scnprintf(buf + len, size - len, "%30s%15s\n",
304 "MAIN", "ALT"); 304 "MAIN", "ALT");
305 len += snprintf(buf + len, size - len, "%-14s:%15d%15d\n", 305 len += scnprintf(buf + len, size - len, "%-14s:%15d%15d\n",
306 "TOTAL COUNT", 306 "TOTAL COUNT",
307 as_main->recv_cnt, 307 as_main->recv_cnt,
308 as_alt->recv_cnt); 308 as_alt->recv_cnt);
309 len += snprintf(buf + len, size - len, "%-14s:%15d%15d\n", 309 len += scnprintf(buf + len, size - len, "%-14s:%15d%15d\n",
310 "LNA1", 310 "LNA1",
311 as_main->lna_recv_cnt[ATH_ANT_DIV_COMB_LNA1], 311 as_main->lna_recv_cnt[ATH_ANT_DIV_COMB_LNA1],
312 as_alt->lna_recv_cnt[ATH_ANT_DIV_COMB_LNA1]); 312 as_alt->lna_recv_cnt[ATH_ANT_DIV_COMB_LNA1]);
313 len += snprintf(buf + len, size - len, "%-14s:%15d%15d\n", 313 len += scnprintf(buf + len, size - len, "%-14s:%15d%15d\n",
314 "LNA2", 314 "LNA2",
315 as_main->lna_recv_cnt[ATH_ANT_DIV_COMB_LNA2], 315 as_main->lna_recv_cnt[ATH_ANT_DIV_COMB_LNA2],
316 as_alt->lna_recv_cnt[ATH_ANT_DIV_COMB_LNA2]); 316 as_alt->lna_recv_cnt[ATH_ANT_DIV_COMB_LNA2]);
317 len += snprintf(buf + len, size - len, "%-14s:%15d%15d\n", 317 len += scnprintf(buf + len, size - len, "%-14s:%15d%15d\n",
318 "LNA1 + LNA2", 318 "LNA1 + LNA2",
319 as_main->lna_recv_cnt[ATH_ANT_DIV_COMB_LNA1_PLUS_LNA2], 319 as_main->lna_recv_cnt[ATH_ANT_DIV_COMB_LNA1_PLUS_LNA2],
320 as_alt->lna_recv_cnt[ATH_ANT_DIV_COMB_LNA1_PLUS_LNA2]); 320 as_alt->lna_recv_cnt[ATH_ANT_DIV_COMB_LNA1_PLUS_LNA2]);
321 len += snprintf(buf + len, size - len, "%-14s:%15d%15d\n", 321 len += scnprintf(buf + len, size - len, "%-14s:%15d%15d\n",
322 "LNA1 - LNA2", 322 "LNA1 - LNA2",
323 as_main->lna_recv_cnt[ATH_ANT_DIV_COMB_LNA1_MINUS_LNA2], 323 as_main->lna_recv_cnt[ATH_ANT_DIV_COMB_LNA1_MINUS_LNA2],
324 as_alt->lna_recv_cnt[ATH_ANT_DIV_COMB_LNA1_MINUS_LNA2]); 324 as_alt->lna_recv_cnt[ATH_ANT_DIV_COMB_LNA1_MINUS_LNA2]);
325 325
326 len += snprintf(buf + len, size - len, "\nLNA Config Attempts:\n"); 326 len += scnprintf(buf + len, size - len, "\nLNA Config Attempts:\n");
327 len += snprintf(buf + len, size - len, "--------------------\n"); 327 len += scnprintf(buf + len, size - len, "--------------------\n");
328 328
329 len += snprintf(buf + len, size - len, "%30s%15s\n", 329 len += scnprintf(buf + len, size - len, "%30s%15s\n",
330 "MAIN", "ALT"); 330 "MAIN", "ALT");
331 len += snprintf(buf + len, size - len, "%-14s:%15d%15d\n", 331 len += scnprintf(buf + len, size - len, "%-14s:%15d%15d\n",
332 "LNA1", 332 "LNA1",
333 as_main->lna_attempt_cnt[ATH_ANT_DIV_COMB_LNA1], 333 as_main->lna_attempt_cnt[ATH_ANT_DIV_COMB_LNA1],
334 as_alt->lna_attempt_cnt[ATH_ANT_DIV_COMB_LNA1]); 334 as_alt->lna_attempt_cnt[ATH_ANT_DIV_COMB_LNA1]);
335 len += snprintf(buf + len, size - len, "%-14s:%15d%15d\n", 335 len += scnprintf(buf + len, size - len, "%-14s:%15d%15d\n",
336 "LNA2", 336 "LNA2",
337 as_main->lna_attempt_cnt[ATH_ANT_DIV_COMB_LNA2], 337 as_main->lna_attempt_cnt[ATH_ANT_DIV_COMB_LNA2],
338 as_alt->lna_attempt_cnt[ATH_ANT_DIV_COMB_LNA2]); 338 as_alt->lna_attempt_cnt[ATH_ANT_DIV_COMB_LNA2]);
339 len += snprintf(buf + len, size - len, "%-14s:%15d%15d\n", 339 len += scnprintf(buf + len, size - len, "%-14s:%15d%15d\n",
340 "LNA1 + LNA2", 340 "LNA1 + LNA2",
341 as_main->lna_attempt_cnt[ATH_ANT_DIV_COMB_LNA1_PLUS_LNA2], 341 as_main->lna_attempt_cnt[ATH_ANT_DIV_COMB_LNA1_PLUS_LNA2],
342 as_alt->lna_attempt_cnt[ATH_ANT_DIV_COMB_LNA1_PLUS_LNA2]); 342 as_alt->lna_attempt_cnt[ATH_ANT_DIV_COMB_LNA1_PLUS_LNA2]);
343 len += snprintf(buf + len, size - len, "%-14s:%15d%15d\n", 343 len += scnprintf(buf + len, size - len, "%-14s:%15d%15d\n",
344 "LNA1 - LNA2", 344 "LNA1 - LNA2",
345 as_main->lna_attempt_cnt[ATH_ANT_DIV_COMB_LNA1_MINUS_LNA2], 345 as_main->lna_attempt_cnt[ATH_ANT_DIV_COMB_LNA1_MINUS_LNA2],
346 as_alt->lna_attempt_cnt[ATH_ANT_DIV_COMB_LNA1_MINUS_LNA2]); 346 as_alt->lna_attempt_cnt[ATH_ANT_DIV_COMB_LNA1_MINUS_LNA2]);
347 347
348exit: 348exit:
349 if (len > size) 349 if (len > size)
@@ -385,21 +385,21 @@ static ssize_t read_file_dma(struct file *file, char __user *user_buf,
385 (AR_MACMISC_MISC_OBS_BUS_1 << 385 (AR_MACMISC_MISC_OBS_BUS_1 <<
386 AR_MACMISC_MISC_OBS_BUS_MSB_S))); 386 AR_MACMISC_MISC_OBS_BUS_MSB_S)));
387 387
388 len += snprintf(buf + len, DMA_BUF_LEN - len, 388 len += scnprintf(buf + len, DMA_BUF_LEN - len,
389 "Raw DMA Debug values:\n"); 389 "Raw DMA Debug values:\n");
390 390
391 for (i = 0; i < ATH9K_NUM_DMA_DEBUG_REGS; i++) { 391 for (i = 0; i < ATH9K_NUM_DMA_DEBUG_REGS; i++) {
392 if (i % 4 == 0) 392 if (i % 4 == 0)
393 len += snprintf(buf + len, DMA_BUF_LEN - len, "\n"); 393 len += scnprintf(buf + len, DMA_BUF_LEN - len, "\n");
394 394
395 val[i] = REG_READ_D(ah, AR_DMADBG_0 + (i * sizeof(u32))); 395 val[i] = REG_READ_D(ah, AR_DMADBG_0 + (i * sizeof(u32)));
396 len += snprintf(buf + len, DMA_BUF_LEN - len, "%d: %08x ", 396 len += scnprintf(buf + len, DMA_BUF_LEN - len, "%d: %08x ",
397 i, val[i]); 397 i, val[i]);
398 } 398 }
399 399
400 len += snprintf(buf + len, DMA_BUF_LEN - len, "\n\n"); 400 len += scnprintf(buf + len, DMA_BUF_LEN - len, "\n\n");
401 len += snprintf(buf + len, DMA_BUF_LEN - len, 401 len += scnprintf(buf + len, DMA_BUF_LEN - len,
402 "Num QCU: chain_st fsp_ok fsp_st DCU: chain_st\n"); 402 "Num QCU: chain_st fsp_ok fsp_st DCU: chain_st\n");
403 403
404 for (i = 0; i < ATH9K_NUM_QUEUES; i++, qcuOffset += 4, dcuOffset += 5) { 404 for (i = 0; i < ATH9K_NUM_QUEUES; i++, qcuOffset += 4, dcuOffset += 5) {
405 if (i == 8) { 405 if (i == 8) {
@@ -412,39 +412,39 @@ static ssize_t read_file_dma(struct file *file, char __user *user_buf,
412 dcuBase++; 412 dcuBase++;
413 } 413 }
414 414
415 len += snprintf(buf + len, DMA_BUF_LEN - len, 415 len += scnprintf(buf + len, DMA_BUF_LEN - len,
416 "%2d %2x %1x %2x %2x\n", 416 "%2d %2x %1x %2x %2x\n",
417 i, (*qcuBase & (0x7 << qcuOffset)) >> qcuOffset, 417 i, (*qcuBase & (0x7 << qcuOffset)) >> qcuOffset,
418 (*qcuBase & (0x8 << qcuOffset)) >> (qcuOffset + 3), 418 (*qcuBase & (0x8 << qcuOffset)) >> (qcuOffset + 3),
419 val[2] & (0x7 << (i * 3)) >> (i * 3), 419 val[2] & (0x7 << (i * 3)) >> (i * 3),
420 (*dcuBase & (0x1f << dcuOffset)) >> dcuOffset); 420 (*dcuBase & (0x1f << dcuOffset)) >> dcuOffset);
421 } 421 }
422 422
423 len += snprintf(buf + len, DMA_BUF_LEN - len, "\n"); 423 len += scnprintf(buf + len, DMA_BUF_LEN - len, "\n");
424 424
425 len += snprintf(buf + len, DMA_BUF_LEN - len, 425 len += scnprintf(buf + len, DMA_BUF_LEN - len,
426 "qcu_stitch state: %2x qcu_fetch state: %2x\n", 426 "qcu_stitch state: %2x qcu_fetch state: %2x\n",
427 (val[3] & 0x003c0000) >> 18, (val[3] & 0x03c00000) >> 22); 427 (val[3] & 0x003c0000) >> 18, (val[3] & 0x03c00000) >> 22);
428 len += snprintf(buf + len, DMA_BUF_LEN - len, 428 len += scnprintf(buf + len, DMA_BUF_LEN - len,
429 "qcu_complete state: %2x dcu_complete state: %2x\n", 429 "qcu_complete state: %2x dcu_complete state: %2x\n",
430 (val[3] & 0x1c000000) >> 26, (val[6] & 0x3)); 430 (val[3] & 0x1c000000) >> 26, (val[6] & 0x3));
431 len += snprintf(buf + len, DMA_BUF_LEN - len, 431 len += scnprintf(buf + len, DMA_BUF_LEN - len,
432 "dcu_arb state: %2x dcu_fp state: %2x\n", 432 "dcu_arb state: %2x dcu_fp state: %2x\n",
433 (val[5] & 0x06000000) >> 25, (val[5] & 0x38000000) >> 27); 433 (val[5] & 0x06000000) >> 25, (val[5] & 0x38000000) >> 27);
434 len += snprintf(buf + len, DMA_BUF_LEN - len, 434 len += scnprintf(buf + len, DMA_BUF_LEN - len,
435 "chan_idle_dur: %3d chan_idle_dur_valid: %1d\n", 435 "chan_idle_dur: %3d chan_idle_dur_valid: %1d\n",
436 (val[6] & 0x000003fc) >> 2, (val[6] & 0x00000400) >> 10); 436 (val[6] & 0x000003fc) >> 2, (val[6] & 0x00000400) >> 10);
437 len += snprintf(buf + len, DMA_BUF_LEN - len, 437 len += scnprintf(buf + len, DMA_BUF_LEN - len,
438 "txfifo_valid_0: %1d txfifo_valid_1: %1d\n", 438 "txfifo_valid_0: %1d txfifo_valid_1: %1d\n",
439 (val[6] & 0x00000800) >> 11, (val[6] & 0x00001000) >> 12); 439 (val[6] & 0x00000800) >> 11, (val[6] & 0x00001000) >> 12);
440 len += snprintf(buf + len, DMA_BUF_LEN - len, 440 len += scnprintf(buf + len, DMA_BUF_LEN - len,
441 "txfifo_dcu_num_0: %2d txfifo_dcu_num_1: %2d\n", 441 "txfifo_dcu_num_0: %2d txfifo_dcu_num_1: %2d\n",
442 (val[6] & 0x0001e000) >> 13, (val[6] & 0x001e0000) >> 17); 442 (val[6] & 0x0001e000) >> 13, (val[6] & 0x001e0000) >> 17);
443 443
444 len += snprintf(buf + len, DMA_BUF_LEN - len, "pcu observe: 0x%x\n", 444 len += scnprintf(buf + len, DMA_BUF_LEN - len, "pcu observe: 0x%x\n",
445 REG_READ_D(ah, AR_OBS_BUS_1)); 445 REG_READ_D(ah, AR_OBS_BUS_1));
446 len += snprintf(buf + len, DMA_BUF_LEN - len, 446 len += scnprintf(buf + len, DMA_BUF_LEN - len,
447 "AR_CR: 0x%x\n", REG_READ_D(ah, AR_CR)); 447 "AR_CR: 0x%x\n", REG_READ_D(ah, AR_CR));
448 448
449 ath9k_ps_restore(sc); 449 ath9k_ps_restore(sc);
450 450
@@ -530,9 +530,9 @@ static ssize_t read_file_interrupt(struct file *file, char __user *user_buf,
530 530
531#define PR_IS(a, s) \ 531#define PR_IS(a, s) \
532 do { \ 532 do { \
533 len += snprintf(buf + len, mxlen - len, \ 533 len += scnprintf(buf + len, mxlen - len, \
534 "%21s: %10u\n", a, \ 534 "%21s: %10u\n", a, \
535 sc->debug.stats.istats.s); \ 535 sc->debug.stats.istats.s); \
536 } while (0) 536 } while (0)
537 537
538 if (sc->sc_ah->caps.hw_caps & ATH9K_HW_CAP_EDMA) { 538 if (sc->sc_ah->caps.hw_caps & ATH9K_HW_CAP_EDMA) {
@@ -563,8 +563,8 @@ static ssize_t read_file_interrupt(struct file *file, char __user *user_buf,
563 PR_IS("GENTIMER", gen_timer); 563 PR_IS("GENTIMER", gen_timer);
564 PR_IS("TOTAL", total); 564 PR_IS("TOTAL", total);
565 565
566 len += snprintf(buf + len, mxlen - len, 566 len += scnprintf(buf + len, mxlen - len,
567 "SYNC_CAUSE stats:\n"); 567 "SYNC_CAUSE stats:\n");
568 568
569 PR_IS("Sync-All", sync_cause_all); 569 PR_IS("Sync-All", sync_cause_all);
570 PR_IS("RTC-IRQ", sync_rtc_irq); 570 PR_IS("RTC-IRQ", sync_rtc_irq);
@@ -655,16 +655,16 @@ static ssize_t print_queue(struct ath_softc *sc, struct ath_txq *txq,
655 655
656 ath_txq_lock(sc, txq); 656 ath_txq_lock(sc, txq);
657 657
658 len += snprintf(buf + len, size - len, "%s: %d ", 658 len += scnprintf(buf + len, size - len, "%s: %d ",
659 "qnum", txq->axq_qnum); 659 "qnum", txq->axq_qnum);
660 len += snprintf(buf + len, size - len, "%s: %2d ", 660 len += scnprintf(buf + len, size - len, "%s: %2d ",
661 "qdepth", txq->axq_depth); 661 "qdepth", txq->axq_depth);
662 len += snprintf(buf + len, size - len, "%s: %2d ", 662 len += scnprintf(buf + len, size - len, "%s: %2d ",
663 "ampdu-depth", txq->axq_ampdu_depth); 663 "ampdu-depth", txq->axq_ampdu_depth);
664 len += snprintf(buf + len, size - len, "%s: %3d ", 664 len += scnprintf(buf + len, size - len, "%s: %3d ",
665 "pending", txq->pending_frames); 665 "pending", txq->pending_frames);
666 len += snprintf(buf + len, size - len, "%s: %d\n", 666 len += scnprintf(buf + len, size - len, "%s: %d\n",
667 "stopped", txq->stopped); 667 "stopped", txq->stopped);
668 668
669 ath_txq_unlock(sc, txq); 669 ath_txq_unlock(sc, txq);
670 return len; 670 return len;
@@ -687,11 +687,11 @@ static ssize_t read_file_queues(struct file *file, char __user *user_buf,
687 687
688 for (i = 0; i < IEEE80211_NUM_ACS; i++) { 688 for (i = 0; i < IEEE80211_NUM_ACS; i++) {
689 txq = sc->tx.txq_map[i]; 689 txq = sc->tx.txq_map[i];
690 len += snprintf(buf + len, size - len, "(%s): ", qname[i]); 690 len += scnprintf(buf + len, size - len, "(%s): ", qname[i]);
691 len += print_queue(sc, txq, buf + len, size - len); 691 len += print_queue(sc, txq, buf + len, size - len);
692 } 692 }
693 693
694 len += snprintf(buf + len, size - len, "(CAB): "); 694 len += scnprintf(buf + len, size - len, "(CAB): ");
695 len += print_queue(sc, sc->beacon.cabq, buf + len, size - len); 695 len += print_queue(sc, sc->beacon.cabq, buf + len, size - len);
696 696
697 if (len > size) 697 if (len > size)
@@ -716,80 +716,82 @@ static ssize_t read_file_misc(struct file *file, char __user *user_buf,
716 unsigned int reg; 716 unsigned int reg;
717 u32 rxfilter; 717 u32 rxfilter;
718 718
719 len += snprintf(buf + len, sizeof(buf) - len, 719 len += scnprintf(buf + len, sizeof(buf) - len,
720 "BSSID: %pM\n", common->curbssid); 720 "BSSID: %pM\n", common->curbssid);
721 len += snprintf(buf + len, sizeof(buf) - len, 721 len += scnprintf(buf + len, sizeof(buf) - len,
722 "BSSID-MASK: %pM\n", common->bssidmask); 722 "BSSID-MASK: %pM\n", common->bssidmask);
723 len += snprintf(buf + len, sizeof(buf) - len, 723 len += scnprintf(buf + len, sizeof(buf) - len,
724 "OPMODE: %s\n", ath_opmode_to_string(sc->sc_ah->opmode)); 724 "OPMODE: %s\n",
725 ath_opmode_to_string(sc->sc_ah->opmode));
725 726
726 ath9k_ps_wakeup(sc); 727 ath9k_ps_wakeup(sc);
727 rxfilter = ath9k_hw_getrxfilter(sc->sc_ah); 728 rxfilter = ath9k_hw_getrxfilter(sc->sc_ah);
728 ath9k_ps_restore(sc); 729 ath9k_ps_restore(sc);
729 730
730 len += snprintf(buf + len, sizeof(buf) - len, 731 len += scnprintf(buf + len, sizeof(buf) - len,
731 "RXFILTER: 0x%x", rxfilter); 732 "RXFILTER: 0x%x", rxfilter);
732 733
733 if (rxfilter & ATH9K_RX_FILTER_UCAST) 734 if (rxfilter & ATH9K_RX_FILTER_UCAST)
734 len += snprintf(buf + len, sizeof(buf) - len, " UCAST"); 735 len += scnprintf(buf + len, sizeof(buf) - len, " UCAST");
735 if (rxfilter & ATH9K_RX_FILTER_MCAST) 736 if (rxfilter & ATH9K_RX_FILTER_MCAST)
736 len += snprintf(buf + len, sizeof(buf) - len, " MCAST"); 737 len += scnprintf(buf + len, sizeof(buf) - len, " MCAST");
737 if (rxfilter & ATH9K_RX_FILTER_BCAST) 738 if (rxfilter & ATH9K_RX_FILTER_BCAST)
738 len += snprintf(buf + len, sizeof(buf) - len, " BCAST"); 739 len += scnprintf(buf + len, sizeof(buf) - len, " BCAST");
739 if (rxfilter & ATH9K_RX_FILTER_CONTROL) 740 if (rxfilter & ATH9K_RX_FILTER_CONTROL)
740 len += snprintf(buf + len, sizeof(buf) - len, " CONTROL"); 741 len += scnprintf(buf + len, sizeof(buf) - len, " CONTROL");
741 if (rxfilter & ATH9K_RX_FILTER_BEACON) 742 if (rxfilter & ATH9K_RX_FILTER_BEACON)
742 len += snprintf(buf + len, sizeof(buf) - len, " BEACON"); 743 len += scnprintf(buf + len, sizeof(buf) - len, " BEACON");
743 if (rxfilter & ATH9K_RX_FILTER_PROM) 744 if (rxfilter & ATH9K_RX_FILTER_PROM)
744 len += snprintf(buf + len, sizeof(buf) - len, " PROM"); 745 len += scnprintf(buf + len, sizeof(buf) - len, " PROM");
745 if (rxfilter & ATH9K_RX_FILTER_PROBEREQ) 746 if (rxfilter & ATH9K_RX_FILTER_PROBEREQ)
746 len += snprintf(buf + len, sizeof(buf) - len, " PROBEREQ"); 747 len += scnprintf(buf + len, sizeof(buf) - len, " PROBEREQ");
747 if (rxfilter & ATH9K_RX_FILTER_PHYERR) 748 if (rxfilter & ATH9K_RX_FILTER_PHYERR)
748 len += snprintf(buf + len, sizeof(buf) - len, " PHYERR"); 749 len += scnprintf(buf + len, sizeof(buf) - len, " PHYERR");
749 if (rxfilter & ATH9K_RX_FILTER_MYBEACON) 750 if (rxfilter & ATH9K_RX_FILTER_MYBEACON)
750 len += snprintf(buf + len, sizeof(buf) - len, " MYBEACON"); 751 len += scnprintf(buf + len, sizeof(buf) - len, " MYBEACON");
751 if (rxfilter & ATH9K_RX_FILTER_COMP_BAR) 752 if (rxfilter & ATH9K_RX_FILTER_COMP_BAR)
752 len += snprintf(buf + len, sizeof(buf) - len, " COMP_BAR"); 753 len += scnprintf(buf + len, sizeof(buf) - len, " COMP_BAR");
753 if (rxfilter & ATH9K_RX_FILTER_PSPOLL) 754 if (rxfilter & ATH9K_RX_FILTER_PSPOLL)
754 len += snprintf(buf + len, sizeof(buf) - len, " PSPOLL"); 755 len += scnprintf(buf + len, sizeof(buf) - len, " PSPOLL");
755 if (rxfilter & ATH9K_RX_FILTER_PHYRADAR) 756 if (rxfilter & ATH9K_RX_FILTER_PHYRADAR)
756 len += snprintf(buf + len, sizeof(buf) - len, " PHYRADAR"); 757 len += scnprintf(buf + len, sizeof(buf) - len, " PHYRADAR");
757 if (rxfilter & ATH9K_RX_FILTER_MCAST_BCAST_ALL) 758 if (rxfilter & ATH9K_RX_FILTER_MCAST_BCAST_ALL)
758 len += snprintf(buf + len, sizeof(buf) - len, " MCAST_BCAST_ALL"); 759 len += scnprintf(buf + len, sizeof(buf) - len, " MCAST_BCAST_ALL");
759 if (rxfilter & ATH9K_RX_FILTER_CONTROL_WRAPPER) 760 if (rxfilter & ATH9K_RX_FILTER_CONTROL_WRAPPER)
760 len += snprintf(buf + len, sizeof(buf) - len, " CONTROL_WRAPPER"); 761 len += scnprintf(buf + len, sizeof(buf) - len, " CONTROL_WRAPPER");
761 762
762 len += snprintf(buf + len, sizeof(buf) - len, "\n"); 763 len += scnprintf(buf + len, sizeof(buf) - len, "\n");
763 764
764 reg = sc->sc_ah->imask; 765 reg = sc->sc_ah->imask;
765 766
766 len += snprintf(buf + len, sizeof(buf) - len, "INTERRUPT-MASK: 0x%x", reg); 767 len += scnprintf(buf + len, sizeof(buf) - len,
768 "INTERRUPT-MASK: 0x%x", reg);
767 769
768 if (reg & ATH9K_INT_SWBA) 770 if (reg & ATH9K_INT_SWBA)
769 len += snprintf(buf + len, sizeof(buf) - len, " SWBA"); 771 len += scnprintf(buf + len, sizeof(buf) - len, " SWBA");
770 if (reg & ATH9K_INT_BMISS) 772 if (reg & ATH9K_INT_BMISS)
771 len += snprintf(buf + len, sizeof(buf) - len, " BMISS"); 773 len += scnprintf(buf + len, sizeof(buf) - len, " BMISS");
772 if (reg & ATH9K_INT_CST) 774 if (reg & ATH9K_INT_CST)
773 len += snprintf(buf + len, sizeof(buf) - len, " CST"); 775 len += scnprintf(buf + len, sizeof(buf) - len, " CST");
774 if (reg & ATH9K_INT_RX) 776 if (reg & ATH9K_INT_RX)
775 len += snprintf(buf + len, sizeof(buf) - len, " RX"); 777 len += scnprintf(buf + len, sizeof(buf) - len, " RX");
776 if (reg & ATH9K_INT_RXHP) 778 if (reg & ATH9K_INT_RXHP)
777 len += snprintf(buf + len, sizeof(buf) - len, " RXHP"); 779 len += scnprintf(buf + len, sizeof(buf) - len, " RXHP");
778 if (reg & ATH9K_INT_RXLP) 780 if (reg & ATH9K_INT_RXLP)
779 len += snprintf(buf + len, sizeof(buf) - len, " RXLP"); 781 len += scnprintf(buf + len, sizeof(buf) - len, " RXLP");
780 if (reg & ATH9K_INT_BB_WATCHDOG) 782 if (reg & ATH9K_INT_BB_WATCHDOG)
781 len += snprintf(buf + len, sizeof(buf) - len, " BB_WATCHDOG"); 783 len += scnprintf(buf + len, sizeof(buf) - len, " BB_WATCHDOG");
782 784
783 len += snprintf(buf + len, sizeof(buf) - len, "\n"); 785 len += scnprintf(buf + len, sizeof(buf) - len, "\n");
784 786
785 ath9k_calculate_iter_data(hw, NULL, &iter_data); 787 ath9k_calculate_iter_data(hw, NULL, &iter_data);
786 788
787 len += snprintf(buf + len, sizeof(buf) - len, 789 len += scnprintf(buf + len, sizeof(buf) - len,
788 "VIF-COUNTS: AP: %i STA: %i MESH: %i WDS: %i" 790 "VIF-COUNTS: AP: %i STA: %i MESH: %i WDS: %i"
789 " ADHOC: %i TOTAL: %hi BEACON-VIF: %hi\n", 791 " ADHOC: %i TOTAL: %hi BEACON-VIF: %hi\n",
790 iter_data.naps, iter_data.nstations, iter_data.nmeshes, 792 iter_data.naps, iter_data.nstations, iter_data.nmeshes,
791 iter_data.nwds, iter_data.nadhocs, 793 iter_data.nwds, iter_data.nadhocs,
792 sc->nvifs, sc->nbcnvifs); 794 sc->nvifs, sc->nbcnvifs);
793 795
794 if (len > sizeof(buf)) 796 if (len > sizeof(buf))
795 len = sizeof(buf); 797 len = sizeof(buf);
@@ -805,27 +807,27 @@ static ssize_t read_file_reset(struct file *file, char __user *user_buf,
805 char buf[512]; 807 char buf[512];
806 unsigned int len = 0; 808 unsigned int len = 0;
807 809
808 len += snprintf(buf + len, sizeof(buf) - len, 810 len += scnprintf(buf + len, sizeof(buf) - len,
809 "%17s: %2d\n", "Baseband Hang", 811 "%17s: %2d\n", "Baseband Hang",
810 sc->debug.stats.reset[RESET_TYPE_BB_HANG]); 812 sc->debug.stats.reset[RESET_TYPE_BB_HANG]);
811 len += snprintf(buf + len, sizeof(buf) - len, 813 len += scnprintf(buf + len, sizeof(buf) - len,
812 "%17s: %2d\n", "Baseband Watchdog", 814 "%17s: %2d\n", "Baseband Watchdog",
813 sc->debug.stats.reset[RESET_TYPE_BB_WATCHDOG]); 815 sc->debug.stats.reset[RESET_TYPE_BB_WATCHDOG]);
814 len += snprintf(buf + len, sizeof(buf) - len, 816 len += scnprintf(buf + len, sizeof(buf) - len,
815 "%17s: %2d\n", "Fatal HW Error", 817 "%17s: %2d\n", "Fatal HW Error",
816 sc->debug.stats.reset[RESET_TYPE_FATAL_INT]); 818 sc->debug.stats.reset[RESET_TYPE_FATAL_INT]);
817 len += snprintf(buf + len, sizeof(buf) - len, 819 len += scnprintf(buf + len, sizeof(buf) - len,
818 "%17s: %2d\n", "TX HW error", 820 "%17s: %2d\n", "TX HW error",
819 sc->debug.stats.reset[RESET_TYPE_TX_ERROR]); 821 sc->debug.stats.reset[RESET_TYPE_TX_ERROR]);
820 len += snprintf(buf + len, sizeof(buf) - len, 822 len += scnprintf(buf + len, sizeof(buf) - len,
821 "%17s: %2d\n", "TX Path Hang", 823 "%17s: %2d\n", "TX Path Hang",
822 sc->debug.stats.reset[RESET_TYPE_TX_HANG]); 824 sc->debug.stats.reset[RESET_TYPE_TX_HANG]);
823 len += snprintf(buf + len, sizeof(buf) - len, 825 len += scnprintf(buf + len, sizeof(buf) - len,
824 "%17s: %2d\n", "PLL RX Hang", 826 "%17s: %2d\n", "PLL RX Hang",
825 sc->debug.stats.reset[RESET_TYPE_PLL_HANG]); 827 sc->debug.stats.reset[RESET_TYPE_PLL_HANG]);
826 len += snprintf(buf + len, sizeof(buf) - len, 828 len += scnprintf(buf + len, sizeof(buf) - len,
827 "%17s: %2d\n", "MCI Reset", 829 "%17s: %2d\n", "MCI Reset",
828 sc->debug.stats.reset[RESET_TYPE_MCI]); 830 sc->debug.stats.reset[RESET_TYPE_MCI]);
829 831
830 if (len > sizeof(buf)) 832 if (len > sizeof(buf))
831 len = sizeof(buf); 833 len = sizeof(buf);
@@ -902,14 +904,14 @@ static ssize_t read_file_recv(struct file *file, char __user *user_buf,
902 size_t count, loff_t *ppos) 904 size_t count, loff_t *ppos)
903{ 905{
904#define PHY_ERR(s, p) \ 906#define PHY_ERR(s, p) \
905 len += snprintf(buf + len, size - len, "%22s : %10u\n", s, \ 907 len += scnprintf(buf + len, size - len, "%22s : %10u\n", s, \
906 sc->debug.stats.rxstats.phy_err_stats[p]); 908 sc->debug.stats.rxstats.phy_err_stats[p]);
907 909
908#define RXS_ERR(s, e) \ 910#define RXS_ERR(s, e) \
909 do { \ 911 do { \
910 len += snprintf(buf + len, size - len, \ 912 len += scnprintf(buf + len, size - len, \
911 "%22s : %10u\n", s, \ 913 "%22s : %10u\n", s, \
912 sc->debug.stats.rxstats.e); \ 914 sc->debug.stats.rxstats.e);\
913 } while (0) 915 } while (0)
914 916
915 struct ath_softc *sc = file->private_data; 917 struct ath_softc *sc = file->private_data;
@@ -1048,6 +1050,9 @@ static ssize_t write_file_spec_scan_ctl(struct file *file,
1048 char buf[32]; 1050 char buf[32];
1049 ssize_t len; 1051 ssize_t len;
1050 1052
1053 if (config_enabled(CONFIG_ATH9K_TX99))
1054 return -EOPNOTSUPP;
1055
1051 len = min(count, sizeof(buf) - 1); 1056 len = min(count, sizeof(buf) - 1);
1052 if (copy_from_user(buf, user_buf, len)) 1057 if (copy_from_user(buf, user_buf, len))
1053 return -EFAULT; 1058 return -EFAULT;
@@ -1439,22 +1444,22 @@ static ssize_t read_file_dump_nfcal(struct file *file, char __user *user_buf,
1439 if (!buf) 1444 if (!buf)
1440 return -ENOMEM; 1445 return -ENOMEM;
1441 1446
1442 len += snprintf(buf + len, size - len, 1447 len += scnprintf(buf + len, size - len,
1443 "Channel Noise Floor : %d\n", ah->noise); 1448 "Channel Noise Floor : %d\n", ah->noise);
1444 len += snprintf(buf + len, size - len, 1449 len += scnprintf(buf + len, size - len,
1445 "Chain | privNF | # Readings | NF Readings\n"); 1450 "Chain | privNF | # Readings | NF Readings\n");
1446 for (i = 0; i < NUM_NF_READINGS; i++) { 1451 for (i = 0; i < NUM_NF_READINGS; i++) {
1447 if (!(chainmask & (1 << i)) || 1452 if (!(chainmask & (1 << i)) ||
1448 ((i >= AR5416_MAX_CHAINS) && !conf_is_ht40(conf))) 1453 ((i >= AR5416_MAX_CHAINS) && !conf_is_ht40(conf)))
1449 continue; 1454 continue;
1450 1455
1451 nread = AR_PHY_CCA_FILTERWINDOW_LENGTH - h[i].invalidNFcount; 1456 nread = AR_PHY_CCA_FILTERWINDOW_LENGTH - h[i].invalidNFcount;
1452 len += snprintf(buf + len, size - len, " %d\t %d\t %d\t\t", 1457 len += scnprintf(buf + len, size - len, " %d\t %d\t %d\t\t",
1453 i, h[i].privNF, nread); 1458 i, h[i].privNF, nread);
1454 for (j = 0; j < nread; j++) 1459 for (j = 0; j < nread; j++)
1455 len += snprintf(buf + len, size - len, 1460 len += scnprintf(buf + len, size - len,
1456 " %d", h[i].nfCalBuffer[j]); 1461 " %d", h[i].nfCalBuffer[j]);
1457 len += snprintf(buf + len, size - len, "\n"); 1462 len += scnprintf(buf + len, size - len, "\n");
1458 } 1463 }
1459 1464
1460 if (len > size) 1465 if (len > size)
@@ -1543,8 +1548,8 @@ static ssize_t read_file_btcoex(struct file *file, char __user *user_buf,
1543 return -ENOMEM; 1548 return -ENOMEM;
1544 1549
1545 if (!sc->sc_ah->common.btcoex_enabled) { 1550 if (!sc->sc_ah->common.btcoex_enabled) {
1546 len = snprintf(buf, size, "%s\n", 1551 len = scnprintf(buf, size, "%s\n",
1547 "BTCOEX is disabled"); 1552 "BTCOEX is disabled");
1548 goto exit; 1553 goto exit;
1549 } 1554 }
1550 1555
@@ -1582,43 +1587,43 @@ static ssize_t read_file_node_stat(struct file *file, char __user *user_buf,
1582 return -ENOMEM; 1587 return -ENOMEM;
1583 1588
1584 if (!an->sta->ht_cap.ht_supported) { 1589 if (!an->sta->ht_cap.ht_supported) {
1585 len = snprintf(buf, size, "%s\n", 1590 len = scnprintf(buf, size, "%s\n",
1586 "HT not supported"); 1591 "HT not supported");
1587 goto exit; 1592 goto exit;
1588 } 1593 }
1589 1594
1590 len = snprintf(buf, size, "Max-AMPDU: %d\n", 1595 len = scnprintf(buf, size, "Max-AMPDU: %d\n",
1591 an->maxampdu); 1596 an->maxampdu);
1592 len += snprintf(buf + len, size - len, "MPDU Density: %d\n\n", 1597 len += scnprintf(buf + len, size - len, "MPDU Density: %d\n\n",
1593 an->mpdudensity); 1598 an->mpdudensity);
1594 1599
1595 len += snprintf(buf + len, size - len, 1600 len += scnprintf(buf + len, size - len,
1596 "%2s%7s\n", "AC", "SCHED"); 1601 "%2s%7s\n", "AC", "SCHED");
1597 1602
1598 for (acno = 0, ac = &an->ac[acno]; 1603 for (acno = 0, ac = &an->ac[acno];
1599 acno < IEEE80211_NUM_ACS; acno++, ac++) { 1604 acno < IEEE80211_NUM_ACS; acno++, ac++) {
1600 txq = ac->txq; 1605 txq = ac->txq;
1601 ath_txq_lock(sc, txq); 1606 ath_txq_lock(sc, txq);
1602 len += snprintf(buf + len, size - len, 1607 len += scnprintf(buf + len, size - len,
1603 "%2d%7d\n", 1608 "%2d%7d\n",
1604 acno, ac->sched); 1609 acno, ac->sched);
1605 ath_txq_unlock(sc, txq); 1610 ath_txq_unlock(sc, txq);
1606 } 1611 }
1607 1612
1608 len += snprintf(buf + len, size - len, 1613 len += scnprintf(buf + len, size - len,
1609 "\n%3s%11s%10s%10s%10s%10s%9s%6s%8s\n", 1614 "\n%3s%11s%10s%10s%10s%10s%9s%6s%8s\n",
1610 "TID", "SEQ_START", "SEQ_NEXT", "BAW_SIZE", 1615 "TID", "SEQ_START", "SEQ_NEXT", "BAW_SIZE",
1611 "BAW_HEAD", "BAW_TAIL", "BAR_IDX", "SCHED", "PAUSED"); 1616 "BAW_HEAD", "BAW_TAIL", "BAR_IDX", "SCHED", "PAUSED");
1612 1617
1613 for (tidno = 0, tid = &an->tid[tidno]; 1618 for (tidno = 0, tid = &an->tid[tidno];
1614 tidno < IEEE80211_NUM_TIDS; tidno++, tid++) { 1619 tidno < IEEE80211_NUM_TIDS; tidno++, tid++) {
1615 txq = tid->ac->txq; 1620 txq = tid->ac->txq;
1616 ath_txq_lock(sc, txq); 1621 ath_txq_lock(sc, txq);
1617 len += snprintf(buf + len, size - len, 1622 len += scnprintf(buf + len, size - len,
1618 "%3d%11d%10d%10d%10d%10d%9d%6d%8d\n", 1623 "%3d%11d%10d%10d%10d%10d%9d%6d%8d\n",
1619 tid->tidno, tid->seq_start, tid->seq_next, 1624 tid->tidno, tid->seq_start, tid->seq_next,
1620 tid->baw_size, tid->baw_head, tid->baw_tail, 1625 tid->baw_size, tid->baw_head, tid->baw_tail,
1621 tid->bar_index, tid->sched, tid->paused); 1626 tid->bar_index, tid->sched, tid->paused);
1622 ath_txq_unlock(sc, txq); 1627 ath_txq_unlock(sc, txq);
1623 } 1628 }
1624exit: 1629exit:
@@ -1773,6 +1778,111 @@ void ath9k_deinit_debug(struct ath_softc *sc)
1773 } 1778 }
1774} 1779}
1775 1780
1781static ssize_t read_file_tx99(struct file *file, char __user *user_buf,
1782 size_t count, loff_t *ppos)
1783{
1784 struct ath_softc *sc = file->private_data;
1785 char buf[3];
1786 unsigned int len;
1787
1788 len = sprintf(buf, "%d\n", sc->tx99_state);
1789 return simple_read_from_buffer(user_buf, count, ppos, buf, len);
1790}
1791
1792static ssize_t write_file_tx99(struct file *file, const char __user *user_buf,
1793 size_t count, loff_t *ppos)
1794{
1795 struct ath_softc *sc = file->private_data;
1796 struct ath_common *common = ath9k_hw_common(sc->sc_ah);
1797 char buf[32];
1798 bool start;
1799 ssize_t len;
1800 int r;
1801
1802 if (sc->nvifs > 1)
1803 return -EOPNOTSUPP;
1804
1805 len = min(count, sizeof(buf) - 1);
1806 if (copy_from_user(buf, user_buf, len))
1807 return -EFAULT;
1808
1809 if (strtobool(buf, &start))
1810 return -EINVAL;
1811
1812 if (start == sc->tx99_state) {
1813 if (!start)
1814 return count;
1815 ath_dbg(common, XMIT, "Resetting TX99\n");
1816 ath9k_tx99_deinit(sc);
1817 }
1818
1819 if (!start) {
1820 ath9k_tx99_deinit(sc);
1821 return count;
1822 }
1823
1824 r = ath9k_tx99_init(sc);
1825 if (r)
1826 return r;
1827
1828 return count;
1829}
1830
1831static const struct file_operations fops_tx99 = {
1832 .read = read_file_tx99,
1833 .write = write_file_tx99,
1834 .open = simple_open,
1835 .owner = THIS_MODULE,
1836 .llseek = default_llseek,
1837};
1838
1839static ssize_t read_file_tx99_power(struct file *file,
1840 char __user *user_buf,
1841 size_t count, loff_t *ppos)
1842{
1843 struct ath_softc *sc = file->private_data;
1844 char buf[32];
1845 unsigned int len;
1846
1847 len = sprintf(buf, "%d (%d dBm)\n",
1848 sc->tx99_power,
1849 sc->tx99_power / 2);
1850
1851 return simple_read_from_buffer(user_buf, count, ppos, buf, len);
1852}
1853
1854static ssize_t write_file_tx99_power(struct file *file,
1855 const char __user *user_buf,
1856 size_t count, loff_t *ppos)
1857{
1858 struct ath_softc *sc = file->private_data;
1859 int r;
1860 u8 tx_power;
1861
1862 r = kstrtou8_from_user(user_buf, count, 0, &tx_power);
1863 if (r)
1864 return r;
1865
1866 if (tx_power > MAX_RATE_POWER)
1867 return -EINVAL;
1868
1869 sc->tx99_power = tx_power;
1870
1871 ath9k_ps_wakeup(sc);
1872 ath9k_hw_tx99_set_txpower(sc->sc_ah, sc->tx99_power);
1873 ath9k_ps_restore(sc);
1874
1875 return count;
1876}
1877
1878static const struct file_operations fops_tx99_power = {
1879 .read = read_file_tx99_power,
1880 .write = write_file_tx99_power,
1881 .open = simple_open,
1882 .owner = THIS_MODULE,
1883 .llseek = default_llseek,
1884};
1885
1776int ath9k_init_debug(struct ath_hw *ah) 1886int ath9k_init_debug(struct ath_hw *ah)
1777{ 1887{
1778 struct ath_common *common = ath9k_hw_common(ah); 1888 struct ath_common *common = ath9k_hw_common(ah);
@@ -1864,5 +1974,15 @@ int ath9k_init_debug(struct ath_hw *ah)
1864 debugfs_create_file("btcoex", S_IRUSR, sc->debug.debugfs_phy, sc, 1974 debugfs_create_file("btcoex", S_IRUSR, sc->debug.debugfs_phy, sc,
1865 &fops_btcoex); 1975 &fops_btcoex);
1866#endif 1976#endif
1977 if (config_enabled(CONFIG_ATH9K_TX99) &&
1978 AR_SREV_9300_20_OR_LATER(ah)) {
1979 debugfs_create_file("tx99", S_IRUSR | S_IWUSR,
1980 sc->debug.debugfs_phy, sc,
1981 &fops_tx99);
1982 debugfs_create_file("tx99_power", S_IRUSR | S_IWUSR,
1983 sc->debug.debugfs_phy, sc,
1984 &fops_tx99_power);
1985 }
1986
1867 return 0; 1987 return 0;
1868} 1988}
diff --git a/drivers/net/wireless/ath/ath9k/debug.h b/drivers/net/wireless/ath/ath9k/debug.h
index 6e1556fa2f3e..d6e3fa4299a4 100644
--- a/drivers/net/wireless/ath/ath9k/debug.h
+++ b/drivers/net/wireless/ath/ath9k/debug.h
@@ -193,12 +193,12 @@ struct ath_tx_stats {
193#define TXSTATS sc->debug.stats.txstats 193#define TXSTATS sc->debug.stats.txstats
194#define PR(str, elem) \ 194#define PR(str, elem) \
195 do { \ 195 do { \
196 len += snprintf(buf + len, size - len, \ 196 len += scnprintf(buf + len, size - len, \
197 "%s%13u%11u%10u%10u\n", str, \ 197 "%s%13u%11u%10u%10u\n", str, \
198 TXSTATS[PR_QNUM(IEEE80211_AC_BE)].elem, \ 198 TXSTATS[PR_QNUM(IEEE80211_AC_BE)].elem,\
199 TXSTATS[PR_QNUM(IEEE80211_AC_BK)].elem, \ 199 TXSTATS[PR_QNUM(IEEE80211_AC_BK)].elem,\
200 TXSTATS[PR_QNUM(IEEE80211_AC_VI)].elem, \ 200 TXSTATS[PR_QNUM(IEEE80211_AC_VI)].elem,\
201 TXSTATS[PR_QNUM(IEEE80211_AC_VO)].elem); \ 201 TXSTATS[PR_QNUM(IEEE80211_AC_VO)].elem); \
202 } while(0) 202 } while(0)
203 203
204#define RX_STAT_INC(c) (sc->debug.stats.rxstats.c++) 204#define RX_STAT_INC(c) (sc->debug.stats.rxstats.c++)
diff --git a/drivers/net/wireless/ath/ath9k/dfs.h b/drivers/net/wireless/ath/ath9k/dfs.h
index 3c839f06a06a..c6fa3d5b5d74 100644
--- a/drivers/net/wireless/ath/ath9k/dfs.h
+++ b/drivers/net/wireless/ath/ath9k/dfs.h
@@ -17,7 +17,7 @@
17 17
18#ifndef ATH9K_DFS_H 18#ifndef ATH9K_DFS_H
19#define ATH9K_DFS_H 19#define ATH9K_DFS_H
20#include "dfs_pattern_detector.h" 20#include "../dfs_pattern_detector.h"
21 21
22#if defined(CONFIG_ATH9K_DFS_CERTIFIED) 22#if defined(CONFIG_ATH9K_DFS_CERTIFIED)
23/** 23/**
diff --git a/drivers/net/wireless/ath/ath9k/dfs_debug.c b/drivers/net/wireless/ath/ath9k/dfs_debug.c
index 3c6e4138a95d..90b8342d1ed4 100644
--- a/drivers/net/wireless/ath/ath9k/dfs_debug.c
+++ b/drivers/net/wireless/ath/ath9k/dfs_debug.c
@@ -20,16 +20,16 @@
20 20
21#include "ath9k.h" 21#include "ath9k.h"
22#include "dfs_debug.h" 22#include "dfs_debug.h"
23#include "../dfs_pattern_detector.h"
23 24
24 25static struct ath_dfs_pool_stats dfs_pool_stats = { 0 };
25struct ath_dfs_pool_stats global_dfs_pool_stats = { 0 };
26 26
27#define ATH9K_DFS_STAT(s, p) \ 27#define ATH9K_DFS_STAT(s, p) \
28 len += snprintf(buf + len, size - len, "%28s : %10u\n", s, \ 28 len += scnprintf(buf + len, size - len, "%28s : %10u\n", s, \
29 sc->debug.stats.dfs_stats.p); 29 sc->debug.stats.dfs_stats.p);
30#define ATH9K_DFS_POOL_STAT(s, p) \ 30#define ATH9K_DFS_POOL_STAT(s, p) \
31 len += snprintf(buf + len, size - len, "%28s : %10u\n", s, \ 31 len += scnprintf(buf + len, size - len, "%28s : %10u\n", s, \
32 global_dfs_pool_stats.p); 32 dfs_pool_stats.p);
33 33
34static ssize_t read_file_dfs(struct file *file, char __user *user_buf, 34static ssize_t read_file_dfs(struct file *file, char __user *user_buf,
35 size_t count, loff_t *ppos) 35 size_t count, loff_t *ppos)
@@ -44,12 +44,15 @@ static ssize_t read_file_dfs(struct file *file, char __user *user_buf,
44 if (buf == NULL) 44 if (buf == NULL)
45 return -ENOMEM; 45 return -ENOMEM;
46 46
47 len += snprintf(buf + len, size - len, "DFS support for " 47 if (sc->dfs_detector)
48 "macVersion = 0x%x, macRev = 0x%x: %s\n", 48 dfs_pool_stats = sc->dfs_detector->get_stats(sc->dfs_detector);
49 hw_ver->macVersion, hw_ver->macRev, 49
50 (sc->sc_ah->caps.hw_caps & ATH9K_HW_CAP_DFS) ? 50 len += scnprintf(buf + len, size - len, "DFS support for "
51 "macVersion = 0x%x, macRev = 0x%x: %s\n",
52 hw_ver->macVersion, hw_ver->macRev,
53 (sc->sc_ah->caps.hw_caps & ATH9K_HW_CAP_DFS) ?
51 "enabled" : "disabled"); 54 "enabled" : "disabled");
52 len += snprintf(buf + len, size - len, "Pulse detector statistics:\n"); 55 len += scnprintf(buf + len, size - len, "Pulse detector statistics:\n");
53 ATH9K_DFS_STAT("pulse events reported ", pulses_total); 56 ATH9K_DFS_STAT("pulse events reported ", pulses_total);
54 ATH9K_DFS_STAT("invalid pulse events ", pulses_no_dfs); 57 ATH9K_DFS_STAT("invalid pulse events ", pulses_no_dfs);
55 ATH9K_DFS_STAT("DFS pulses detected ", pulses_detected); 58 ATH9K_DFS_STAT("DFS pulses detected ", pulses_detected);
@@ -59,11 +62,12 @@ static ssize_t read_file_dfs(struct file *file, char __user *user_buf,
59 ATH9K_DFS_STAT("Primary channel pulses ", pri_phy_errors); 62 ATH9K_DFS_STAT("Primary channel pulses ", pri_phy_errors);
60 ATH9K_DFS_STAT("Secondary channel pulses", ext_phy_errors); 63 ATH9K_DFS_STAT("Secondary channel pulses", ext_phy_errors);
61 ATH9K_DFS_STAT("Dual channel pulses ", dc_phy_errors); 64 ATH9K_DFS_STAT("Dual channel pulses ", dc_phy_errors);
62 len += snprintf(buf + len, size - len, "Radar detector statistics " 65 len += scnprintf(buf + len, size - len, "Radar detector statistics "
63 "(current DFS region: %d)\n", sc->dfs_detector->region); 66 "(current DFS region: %d)\n",
67 sc->dfs_detector->region);
64 ATH9K_DFS_STAT("Pulse events processed ", pulses_processed); 68 ATH9K_DFS_STAT("Pulse events processed ", pulses_processed);
65 ATH9K_DFS_STAT("Radars detected ", radar_detected); 69 ATH9K_DFS_STAT("Radars detected ", radar_detected);
66 len += snprintf(buf + len, size - len, "Global Pool statistics:\n"); 70 len += scnprintf(buf + len, size - len, "Global Pool statistics:\n");
67 ATH9K_DFS_POOL_STAT("Pool references ", pool_reference); 71 ATH9K_DFS_POOL_STAT("Pool references ", pool_reference);
68 ATH9K_DFS_POOL_STAT("Pulses allocated ", pulse_allocated); 72 ATH9K_DFS_POOL_STAT("Pulses allocated ", pulse_allocated);
69 ATH9K_DFS_POOL_STAT("Pulses alloc error ", pulse_alloc_error); 73 ATH9K_DFS_POOL_STAT("Pulses alloc error ", pulse_alloc_error);
diff --git a/drivers/net/wireless/ath/ath9k/dfs_debug.h b/drivers/net/wireless/ath/ath9k/dfs_debug.h
index e36810a4b585..0a7ddf4c88c9 100644
--- a/drivers/net/wireless/ath/ath9k/dfs_debug.h
+++ b/drivers/net/wireless/ath/ath9k/dfs_debug.h
@@ -51,25 +51,11 @@ struct ath_dfs_stats {
51 u32 radar_detected; 51 u32 radar_detected;
52}; 52};
53 53
54/**
55 * struct ath_dfs_pool_stats - DFS Statistics for global pools
56 */
57struct ath_dfs_pool_stats {
58 u32 pool_reference;
59 u32 pulse_allocated;
60 u32 pulse_alloc_error;
61 u32 pulse_used;
62 u32 pseq_allocated;
63 u32 pseq_alloc_error;
64 u32 pseq_used;
65};
66#if defined(CONFIG_ATH9K_DFS_DEBUGFS) 54#if defined(CONFIG_ATH9K_DFS_DEBUGFS)
67 55
68#define DFS_STAT_INC(sc, c) (sc->debug.stats.dfs_stats.c++) 56#define DFS_STAT_INC(sc, c) (sc->debug.stats.dfs_stats.c++)
69void ath9k_dfs_init_debug(struct ath_softc *sc); 57void ath9k_dfs_init_debug(struct ath_softc *sc);
70 58
71#define DFS_POOL_STAT_INC(c) (global_dfs_pool_stats.c++)
72#define DFS_POOL_STAT_DEC(c) (global_dfs_pool_stats.c--)
73extern struct ath_dfs_pool_stats global_dfs_pool_stats; 59extern struct ath_dfs_pool_stats global_dfs_pool_stats;
74 60
75#else 61#else
@@ -77,8 +63,6 @@ extern struct ath_dfs_pool_stats global_dfs_pool_stats;
77#define DFS_STAT_INC(sc, c) do { } while (0) 63#define DFS_STAT_INC(sc, c) do { } while (0)
78static inline void ath9k_dfs_init_debug(struct ath_softc *sc) { } 64static inline void ath9k_dfs_init_debug(struct ath_softc *sc) { }
79 65
80#define DFS_POOL_STAT_INC(c) do { } while (0)
81#define DFS_POOL_STAT_DEC(c) do { } while (0)
82#endif /* CONFIG_ATH9K_DFS_DEBUGFS */ 66#endif /* CONFIG_ATH9K_DFS_DEBUGFS */
83 67
84#endif /* ATH9K_DFS_DEBUG_H */ 68#endif /* ATH9K_DFS_DEBUG_H */
diff --git a/drivers/net/wireless/ath/ath9k/eeprom_4k.c b/drivers/net/wireless/ath/ath9k/eeprom_4k.c
index 9ea8e4b779c9..b4091716e9b3 100644
--- a/drivers/net/wireless/ath/ath9k/eeprom_4k.c
+++ b/drivers/net/wireless/ath/ath9k/eeprom_4k.c
@@ -129,10 +129,10 @@ static u32 ath9k_hw_4k_dump_eeprom(struct ath_hw *ah, bool dump_base_hdr,
129 struct base_eep_header_4k *pBase = &eep->baseEepHeader; 129 struct base_eep_header_4k *pBase = &eep->baseEepHeader;
130 130
131 if (!dump_base_hdr) { 131 if (!dump_base_hdr) {
132 len += snprintf(buf + len, size - len, 132 len += scnprintf(buf + len, size - len,
133 "%20s :\n", "2GHz modal Header"); 133 "%20s :\n", "2GHz modal Header");
134 len = ath9k_dump_4k_modal_eeprom(buf, len, size, 134 len = ath9k_dump_4k_modal_eeprom(buf, len, size,
135 &eep->modalHeader); 135 &eep->modalHeader);
136 goto out; 136 goto out;
137 } 137 }
138 138
@@ -160,8 +160,8 @@ static u32 ath9k_hw_4k_dump_eeprom(struct ath_hw *ah, bool dump_base_hdr,
160 PR_EEP("Cal Bin Build", (pBase->binBuildNumber >> 8) & 0xFF); 160 PR_EEP("Cal Bin Build", (pBase->binBuildNumber >> 8) & 0xFF);
161 PR_EEP("TX Gain type", pBase->txGainType); 161 PR_EEP("TX Gain type", pBase->txGainType);
162 162
163 len += snprintf(buf + len, size - len, "%20s : %pM\n", "MacAddress", 163 len += scnprintf(buf + len, size - len, "%20s : %pM\n", "MacAddress",
164 pBase->macAddr); 164 pBase->macAddr);
165 165
166out: 166out:
167 if (len > size) 167 if (len > size)
diff --git a/drivers/net/wireless/ath/ath9k/eeprom_9287.c b/drivers/net/wireless/ath/ath9k/eeprom_9287.c
index 3ae1f3df0637..e1d0c217c104 100644
--- a/drivers/net/wireless/ath/ath9k/eeprom_9287.c
+++ b/drivers/net/wireless/ath/ath9k/eeprom_9287.c
@@ -125,8 +125,8 @@ static u32 ath9k_hw_ar9287_dump_eeprom(struct ath_hw *ah, bool dump_base_hdr,
125 struct base_eep_ar9287_header *pBase = &eep->baseEepHeader; 125 struct base_eep_ar9287_header *pBase = &eep->baseEepHeader;
126 126
127 if (!dump_base_hdr) { 127 if (!dump_base_hdr) {
128 len += snprintf(buf + len, size - len, 128 len += scnprintf(buf + len, size - len,
129 "%20s :\n", "2GHz modal Header"); 129 "%20s :\n", "2GHz modal Header");
130 len = ar9287_dump_modal_eeprom(buf, len, size, 130 len = ar9287_dump_modal_eeprom(buf, len, size,
131 &eep->modalHeader); 131 &eep->modalHeader);
132 goto out; 132 goto out;
@@ -157,8 +157,8 @@ static u32 ath9k_hw_ar9287_dump_eeprom(struct ath_hw *ah, bool dump_base_hdr,
157 PR_EEP("Power Table Offset", pBase->pwrTableOffset); 157 PR_EEP("Power Table Offset", pBase->pwrTableOffset);
158 PR_EEP("OpenLoop Power Ctrl", pBase->openLoopPwrCntl); 158 PR_EEP("OpenLoop Power Ctrl", pBase->openLoopPwrCntl);
159 159
160 len += snprintf(buf + len, size - len, "%20s : %pM\n", "MacAddress", 160 len += scnprintf(buf + len, size - len, "%20s : %pM\n", "MacAddress",
161 pBase->macAddr); 161 pBase->macAddr);
162 162
163out: 163out:
164 if (len > size) 164 if (len > size)
diff --git a/drivers/net/wireless/ath/ath9k/eeprom_def.c b/drivers/net/wireless/ath/ath9k/eeprom_def.c
index 1c25368b3836..39107e31e79a 100644
--- a/drivers/net/wireless/ath/ath9k/eeprom_def.c
+++ b/drivers/net/wireless/ath/ath9k/eeprom_def.c
@@ -205,12 +205,12 @@ static u32 ath9k_hw_def_dump_eeprom(struct ath_hw *ah, bool dump_base_hdr,
205 struct base_eep_header *pBase = &eep->baseEepHeader; 205 struct base_eep_header *pBase = &eep->baseEepHeader;
206 206
207 if (!dump_base_hdr) { 207 if (!dump_base_hdr) {
208 len += snprintf(buf + len, size - len, 208 len += scnprintf(buf + len, size - len,
209 "%20s :\n", "2GHz modal Header"); 209 "%20s :\n", "2GHz modal Header");
210 len = ath9k_def_dump_modal_eeprom(buf, len, size, 210 len = ath9k_def_dump_modal_eeprom(buf, len, size,
211 &eep->modalHeader[0]); 211 &eep->modalHeader[0]);
212 len += snprintf(buf + len, size - len, 212 len += scnprintf(buf + len, size - len,
213 "%20s :\n", "5GHz modal Header"); 213 "%20s :\n", "5GHz modal Header");
214 len = ath9k_def_dump_modal_eeprom(buf, len, size, 214 len = ath9k_def_dump_modal_eeprom(buf, len, size,
215 &eep->modalHeader[1]); 215 &eep->modalHeader[1]);
216 goto out; 216 goto out;
@@ -240,8 +240,8 @@ static u32 ath9k_hw_def_dump_eeprom(struct ath_hw *ah, bool dump_base_hdr,
240 PR_EEP("Cal Bin Build", (pBase->binBuildNumber >> 8) & 0xFF); 240 PR_EEP("Cal Bin Build", (pBase->binBuildNumber >> 8) & 0xFF);
241 PR_EEP("OpenLoop Power Ctrl", pBase->openLoopPwrCntl); 241 PR_EEP("OpenLoop Power Ctrl", pBase->openLoopPwrCntl);
242 242
243 len += snprintf(buf + len, size - len, "%20s : %pM\n", "MacAddress", 243 len += scnprintf(buf + len, size - len, "%20s : %pM\n", "MacAddress",
244 pBase->macAddr); 244 pBase->macAddr);
245 245
246out: 246out:
247 if (len > size) 247 if (len > size)
diff --git a/drivers/net/wireless/ath/ath9k/gpio.c b/drivers/net/wireless/ath/ath9k/gpio.c
index 4b412aaf4f36..c34f21241da9 100644
--- a/drivers/net/wireless/ath/ath9k/gpio.c
+++ b/drivers/net/wireless/ath/ath9k/gpio.c
@@ -522,22 +522,22 @@ static int ath9k_dump_mci_btcoex(struct ath_softc *sc, u8 *buf, u32 size)
522 ATH_DUMP_BTCOEX("Concurrent Tx", btcoex_hw->mci.concur_tx); 522 ATH_DUMP_BTCOEX("Concurrent Tx", btcoex_hw->mci.concur_tx);
523 ATH_DUMP_BTCOEX("Concurrent RSSI cnt", btcoex->rssi_count); 523 ATH_DUMP_BTCOEX("Concurrent RSSI cnt", btcoex->rssi_count);
524 524
525 len += snprintf(buf + len, size - len, "BT Weights: "); 525 len += scnprintf(buf + len, size - len, "BT Weights: ");
526 for (i = 0; i < AR9300_NUM_BT_WEIGHTS; i++) 526 for (i = 0; i < AR9300_NUM_BT_WEIGHTS; i++)
527 len += snprintf(buf + len, size - len, "%08x ", 527 len += scnprintf(buf + len, size - len, "%08x ",
528 btcoex_hw->bt_weight[i]); 528 btcoex_hw->bt_weight[i]);
529 len += snprintf(buf + len, size - len, "\n"); 529 len += scnprintf(buf + len, size - len, "\n");
530 len += snprintf(buf + len, size - len, "WLAN Weights: "); 530 len += scnprintf(buf + len, size - len, "WLAN Weights: ");
531 for (i = 0; i < AR9300_NUM_BT_WEIGHTS; i++) 531 for (i = 0; i < AR9300_NUM_BT_WEIGHTS; i++)
532 len += snprintf(buf + len, size - len, "%08x ", 532 len += scnprintf(buf + len, size - len, "%08x ",
533 btcoex_hw->wlan_weight[i]); 533 btcoex_hw->wlan_weight[i]);
534 len += snprintf(buf + len, size - len, "\n"); 534 len += scnprintf(buf + len, size - len, "\n");
535 len += snprintf(buf + len, size - len, "Tx Priorities: "); 535 len += scnprintf(buf + len, size - len, "Tx Priorities: ");
536 for (i = 0; i < ATH_BTCOEX_STOMP_MAX; i++) 536 for (i = 0; i < ATH_BTCOEX_STOMP_MAX; i++)
537 len += snprintf(buf + len, size - len, "%08x ", 537 len += scnprintf(buf + len, size - len, "%08x ",
538 btcoex_hw->tx_prio[i]); 538 btcoex_hw->tx_prio[i]);
539 539
540 len += snprintf(buf + len, size - len, "\n"); 540 len += scnprintf(buf + len, size - len, "\n");
541 541
542 return len; 542 return len;
543} 543}
diff --git a/drivers/net/wireless/ath/ath9k/htc_drv_debug.c b/drivers/net/wireless/ath/ath9k/htc_drv_debug.c
index c1b45e2f8481..fb071ee4fcfb 100644
--- a/drivers/net/wireless/ath/ath9k/htc_drv_debug.c
+++ b/drivers/net/wireless/ath/ath9k/htc_drv_debug.c
@@ -37,29 +37,29 @@ static ssize_t read_file_tgt_int_stats(struct file *file, char __user *user_buf,
37 37
38 ath9k_htc_ps_restore(priv); 38 ath9k_htc_ps_restore(priv);
39 39
40 len += snprintf(buf + len, sizeof(buf) - len, 40 len += scnprintf(buf + len, sizeof(buf) - len,
41 "%20s : %10u\n", "RX", 41 "%20s : %10u\n", "RX",
42 be32_to_cpu(cmd_rsp.rx)); 42 be32_to_cpu(cmd_rsp.rx));
43 43
44 len += snprintf(buf + len, sizeof(buf) - len, 44 len += scnprintf(buf + len, sizeof(buf) - len,
45 "%20s : %10u\n", "RXORN", 45 "%20s : %10u\n", "RXORN",
46 be32_to_cpu(cmd_rsp.rxorn)); 46 be32_to_cpu(cmd_rsp.rxorn));
47 47
48 len += snprintf(buf + len, sizeof(buf) - len, 48 len += scnprintf(buf + len, sizeof(buf) - len,
49 "%20s : %10u\n", "RXEOL", 49 "%20s : %10u\n", "RXEOL",
50 be32_to_cpu(cmd_rsp.rxeol)); 50 be32_to_cpu(cmd_rsp.rxeol));
51 51
52 len += snprintf(buf + len, sizeof(buf) - len, 52 len += scnprintf(buf + len, sizeof(buf) - len,
53 "%20s : %10u\n", "TXURN", 53 "%20s : %10u\n", "TXURN",
54 be32_to_cpu(cmd_rsp.txurn)); 54 be32_to_cpu(cmd_rsp.txurn));
55 55
56 len += snprintf(buf + len, sizeof(buf) - len, 56 len += scnprintf(buf + len, sizeof(buf) - len,
57 "%20s : %10u\n", "TXTO", 57 "%20s : %10u\n", "TXTO",
58 be32_to_cpu(cmd_rsp.txto)); 58 be32_to_cpu(cmd_rsp.txto));
59 59
60 len += snprintf(buf + len, sizeof(buf) - len, 60 len += scnprintf(buf + len, sizeof(buf) - len,
61 "%20s : %10u\n", "CST", 61 "%20s : %10u\n", "CST",
62 be32_to_cpu(cmd_rsp.cst)); 62 be32_to_cpu(cmd_rsp.cst));
63 63
64 if (len > sizeof(buf)) 64 if (len > sizeof(buf))
65 len = sizeof(buf); 65 len = sizeof(buf);
@@ -95,41 +95,41 @@ static ssize_t read_file_tgt_tx_stats(struct file *file, char __user *user_buf,
95 95
96 ath9k_htc_ps_restore(priv); 96 ath9k_htc_ps_restore(priv);
97 97
98 len += snprintf(buf + len, sizeof(buf) - len, 98 len += scnprintf(buf + len, sizeof(buf) - len,
99 "%20s : %10u\n", "Xretries", 99 "%20s : %10u\n", "Xretries",
100 be32_to_cpu(cmd_rsp.xretries)); 100 be32_to_cpu(cmd_rsp.xretries));
101 101
102 len += snprintf(buf + len, sizeof(buf) - len, 102 len += scnprintf(buf + len, sizeof(buf) - len,
103 "%20s : %10u\n", "FifoErr", 103 "%20s : %10u\n", "FifoErr",
104 be32_to_cpu(cmd_rsp.fifoerr)); 104 be32_to_cpu(cmd_rsp.fifoerr));
105 105
106 len += snprintf(buf + len, sizeof(buf) - len, 106 len += scnprintf(buf + len, sizeof(buf) - len,
107 "%20s : %10u\n", "Filtered", 107 "%20s : %10u\n", "Filtered",
108 be32_to_cpu(cmd_rsp.filtered)); 108 be32_to_cpu(cmd_rsp.filtered));
109 109
110 len += snprintf(buf + len, sizeof(buf) - len, 110 len += scnprintf(buf + len, sizeof(buf) - len,
111 "%20s : %10u\n", "TimerExp", 111 "%20s : %10u\n", "TimerExp",
112 be32_to_cpu(cmd_rsp.timer_exp)); 112 be32_to_cpu(cmd_rsp.timer_exp));
113 113
114 len += snprintf(buf + len, sizeof(buf) - len, 114 len += scnprintf(buf + len, sizeof(buf) - len,
115 "%20s : %10u\n", "ShortRetries", 115 "%20s : %10u\n", "ShortRetries",
116 be32_to_cpu(cmd_rsp.shortretries)); 116 be32_to_cpu(cmd_rsp.shortretries));
117 117
118 len += snprintf(buf + len, sizeof(buf) - len, 118 len += scnprintf(buf + len, sizeof(buf) - len,
119 "%20s : %10u\n", "LongRetries", 119 "%20s : %10u\n", "LongRetries",
120 be32_to_cpu(cmd_rsp.longretries)); 120 be32_to_cpu(cmd_rsp.longretries));
121 121
122 len += snprintf(buf + len, sizeof(buf) - len, 122 len += scnprintf(buf + len, sizeof(buf) - len,
123 "%20s : %10u\n", "QueueNull", 123 "%20s : %10u\n", "QueueNull",
124 be32_to_cpu(cmd_rsp.qnull)); 124 be32_to_cpu(cmd_rsp.qnull));
125 125
126 len += snprintf(buf + len, sizeof(buf) - len, 126 len += scnprintf(buf + len, sizeof(buf) - len,
127 "%20s : %10u\n", "EncapFail", 127 "%20s : %10u\n", "EncapFail",
128 be32_to_cpu(cmd_rsp.encap_fail)); 128 be32_to_cpu(cmd_rsp.encap_fail));
129 129
130 len += snprintf(buf + len, sizeof(buf) - len, 130 len += scnprintf(buf + len, sizeof(buf) - len,
131 "%20s : %10u\n", "NoBuf", 131 "%20s : %10u\n", "NoBuf",
132 be32_to_cpu(cmd_rsp.nobuf)); 132 be32_to_cpu(cmd_rsp.nobuf));
133 133
134 if (len > sizeof(buf)) 134 if (len > sizeof(buf))
135 len = sizeof(buf); 135 len = sizeof(buf);
@@ -165,17 +165,17 @@ static ssize_t read_file_tgt_rx_stats(struct file *file, char __user *user_buf,
165 165
166 ath9k_htc_ps_restore(priv); 166 ath9k_htc_ps_restore(priv);
167 167
168 len += snprintf(buf + len, sizeof(buf) - len, 168 len += scnprintf(buf + len, sizeof(buf) - len,
169 "%20s : %10u\n", "NoBuf", 169 "%20s : %10u\n", "NoBuf",
170 be32_to_cpu(cmd_rsp.nobuf)); 170 be32_to_cpu(cmd_rsp.nobuf));
171 171
172 len += snprintf(buf + len, sizeof(buf) - len, 172 len += scnprintf(buf + len, sizeof(buf) - len,
173 "%20s : %10u\n", "HostSend", 173 "%20s : %10u\n", "HostSend",
174 be32_to_cpu(cmd_rsp.host_send)); 174 be32_to_cpu(cmd_rsp.host_send));
175 175
176 len += snprintf(buf + len, sizeof(buf) - len, 176 len += scnprintf(buf + len, sizeof(buf) - len,
177 "%20s : %10u\n", "HostDone", 177 "%20s : %10u\n", "HostDone",
178 be32_to_cpu(cmd_rsp.host_done)); 178 be32_to_cpu(cmd_rsp.host_done));
179 179
180 if (len > sizeof(buf)) 180 if (len > sizeof(buf))
181 len = sizeof(buf); 181 len = sizeof(buf);
@@ -197,37 +197,37 @@ static ssize_t read_file_xmit(struct file *file, char __user *user_buf,
197 char buf[512]; 197 char buf[512];
198 unsigned int len = 0; 198 unsigned int len = 0;
199 199
200 len += snprintf(buf + len, sizeof(buf) - len, 200 len += scnprintf(buf + len, sizeof(buf) - len,
201 "%20s : %10u\n", "Buffers queued", 201 "%20s : %10u\n", "Buffers queued",
202 priv->debug.tx_stats.buf_queued); 202 priv->debug.tx_stats.buf_queued);
203 len += snprintf(buf + len, sizeof(buf) - len, 203 len += scnprintf(buf + len, sizeof(buf) - len,
204 "%20s : %10u\n", "Buffers completed", 204 "%20s : %10u\n", "Buffers completed",
205 priv->debug.tx_stats.buf_completed); 205 priv->debug.tx_stats.buf_completed);
206 len += snprintf(buf + len, sizeof(buf) - len, 206 len += scnprintf(buf + len, sizeof(buf) - len,
207 "%20s : %10u\n", "SKBs queued", 207 "%20s : %10u\n", "SKBs queued",
208 priv->debug.tx_stats.skb_queued); 208 priv->debug.tx_stats.skb_queued);
209 len += snprintf(buf + len, sizeof(buf) - len, 209 len += scnprintf(buf + len, sizeof(buf) - len,
210 "%20s : %10u\n", "SKBs success", 210 "%20s : %10u\n", "SKBs success",
211 priv->debug.tx_stats.skb_success); 211 priv->debug.tx_stats.skb_success);
212 len += snprintf(buf + len, sizeof(buf) - len, 212 len += scnprintf(buf + len, sizeof(buf) - len,
213 "%20s : %10u\n", "SKBs failed", 213 "%20s : %10u\n", "SKBs failed",
214 priv->debug.tx_stats.skb_failed); 214 priv->debug.tx_stats.skb_failed);
215 len += snprintf(buf + len, sizeof(buf) - len, 215 len += scnprintf(buf + len, sizeof(buf) - len,
216 "%20s : %10u\n", "CAB queued", 216 "%20s : %10u\n", "CAB queued",
217 priv->debug.tx_stats.cab_queued); 217 priv->debug.tx_stats.cab_queued);
218 218
219 len += snprintf(buf + len, sizeof(buf) - len, 219 len += scnprintf(buf + len, sizeof(buf) - len,
220 "%20s : %10u\n", "BE queued", 220 "%20s : %10u\n", "BE queued",
221 priv->debug.tx_stats.queue_stats[IEEE80211_AC_BE]); 221 priv->debug.tx_stats.queue_stats[IEEE80211_AC_BE]);
222 len += snprintf(buf + len, sizeof(buf) - len, 222 len += scnprintf(buf + len, sizeof(buf) - len,
223 "%20s : %10u\n", "BK queued", 223 "%20s : %10u\n", "BK queued",
224 priv->debug.tx_stats.queue_stats[IEEE80211_AC_BK]); 224 priv->debug.tx_stats.queue_stats[IEEE80211_AC_BK]);
225 len += snprintf(buf + len, sizeof(buf) - len, 225 len += scnprintf(buf + len, sizeof(buf) - len,
226 "%20s : %10u\n", "VI queued", 226 "%20s : %10u\n", "VI queued",
227 priv->debug.tx_stats.queue_stats[IEEE80211_AC_VI]); 227 priv->debug.tx_stats.queue_stats[IEEE80211_AC_VI]);
228 len += snprintf(buf + len, sizeof(buf) - len, 228 len += scnprintf(buf + len, sizeof(buf) - len,
229 "%20s : %10u\n", "VO queued", 229 "%20s : %10u\n", "VO queued",
230 priv->debug.tx_stats.queue_stats[IEEE80211_AC_VO]); 230 priv->debug.tx_stats.queue_stats[IEEE80211_AC_VO]);
231 231
232 if (len > sizeof(buf)) 232 if (len > sizeof(buf))
233 len = sizeof(buf); 233 len = sizeof(buf);
@@ -273,8 +273,8 @@ static ssize_t read_file_recv(struct file *file, char __user *user_buf,
273 size_t count, loff_t *ppos) 273 size_t count, loff_t *ppos)
274{ 274{
275#define PHY_ERR(s, p) \ 275#define PHY_ERR(s, p) \
276 len += snprintf(buf + len, size - len, "%20s : %10u\n", s, \ 276 len += scnprintf(buf + len, size - len, "%20s : %10u\n", s, \
277 priv->debug.rx_stats.err_phy_stats[p]); 277 priv->debug.rx_stats.err_phy_stats[p]);
278 278
279 struct ath9k_htc_priv *priv = file->private_data; 279 struct ath9k_htc_priv *priv = file->private_data;
280 char *buf; 280 char *buf;
@@ -285,37 +285,37 @@ static ssize_t read_file_recv(struct file *file, char __user *user_buf,
285 if (buf == NULL) 285 if (buf == NULL)
286 return -ENOMEM; 286 return -ENOMEM;
287 287
288 len += snprintf(buf + len, size - len, 288 len += scnprintf(buf + len, size - len,
289 "%20s : %10u\n", "SKBs allocated", 289 "%20s : %10u\n", "SKBs allocated",
290 priv->debug.rx_stats.skb_allocated); 290 priv->debug.rx_stats.skb_allocated);
291 len += snprintf(buf + len, size - len, 291 len += scnprintf(buf + len, size - len,
292 "%20s : %10u\n", "SKBs completed", 292 "%20s : %10u\n", "SKBs completed",
293 priv->debug.rx_stats.skb_completed); 293 priv->debug.rx_stats.skb_completed);
294 len += snprintf(buf + len, size - len, 294 len += scnprintf(buf + len, size - len,
295 "%20s : %10u\n", "SKBs Dropped", 295 "%20s : %10u\n", "SKBs Dropped",
296 priv->debug.rx_stats.skb_dropped); 296 priv->debug.rx_stats.skb_dropped);
297 297
298 len += snprintf(buf + len, size - len, 298 len += scnprintf(buf + len, size - len,
299 "%20s : %10u\n", "CRC ERR", 299 "%20s : %10u\n", "CRC ERR",
300 priv->debug.rx_stats.err_crc); 300 priv->debug.rx_stats.err_crc);
301 len += snprintf(buf + len, size - len, 301 len += scnprintf(buf + len, size - len,
302 "%20s : %10u\n", "DECRYPT CRC ERR", 302 "%20s : %10u\n", "DECRYPT CRC ERR",
303 priv->debug.rx_stats.err_decrypt_crc); 303 priv->debug.rx_stats.err_decrypt_crc);
304 len += snprintf(buf + len, size - len, 304 len += scnprintf(buf + len, size - len,
305 "%20s : %10u\n", "MIC ERR", 305 "%20s : %10u\n", "MIC ERR",
306 priv->debug.rx_stats.err_mic); 306 priv->debug.rx_stats.err_mic);
307 len += snprintf(buf + len, size - len, 307 len += scnprintf(buf + len, size - len,
308 "%20s : %10u\n", "PRE-DELIM CRC ERR", 308 "%20s : %10u\n", "PRE-DELIM CRC ERR",
309 priv->debug.rx_stats.err_pre_delim); 309 priv->debug.rx_stats.err_pre_delim);
310 len += snprintf(buf + len, size - len, 310 len += scnprintf(buf + len, size - len,
311 "%20s : %10u\n", "POST-DELIM CRC ERR", 311 "%20s : %10u\n", "POST-DELIM CRC ERR",
312 priv->debug.rx_stats.err_post_delim); 312 priv->debug.rx_stats.err_post_delim);
313 len += snprintf(buf + len, size - len, 313 len += scnprintf(buf + len, size - len,
314 "%20s : %10u\n", "DECRYPT BUSY ERR", 314 "%20s : %10u\n", "DECRYPT BUSY ERR",
315 priv->debug.rx_stats.err_decrypt_busy); 315 priv->debug.rx_stats.err_decrypt_busy);
316 len += snprintf(buf + len, size - len, 316 len += scnprintf(buf + len, size - len,
317 "%20s : %10u\n", "TOTAL PHY ERR", 317 "%20s : %10u\n", "TOTAL PHY ERR",
318 priv->debug.rx_stats.err_phy); 318 priv->debug.rx_stats.err_phy);
319 319
320 320
321 PHY_ERR("UNDERRUN", ATH9K_PHYERR_UNDERRUN); 321 PHY_ERR("UNDERRUN", ATH9K_PHYERR_UNDERRUN);
@@ -372,16 +372,16 @@ static ssize_t read_file_slot(struct file *file, char __user *user_buf,
372 372
373 spin_lock_bh(&priv->tx.tx_lock); 373 spin_lock_bh(&priv->tx.tx_lock);
374 374
375 len += snprintf(buf + len, sizeof(buf) - len, "TX slot bitmap : "); 375 len += scnprintf(buf + len, sizeof(buf) - len, "TX slot bitmap : ");
376 376
377 len += bitmap_scnprintf(buf + len, sizeof(buf) - len, 377 len += bitmap_scnprintf(buf + len, sizeof(buf) - len,
378 priv->tx.tx_slot, MAX_TX_BUF_NUM); 378 priv->tx.tx_slot, MAX_TX_BUF_NUM);
379 379
380 len += snprintf(buf + len, sizeof(buf) - len, "\n"); 380 len += scnprintf(buf + len, sizeof(buf) - len, "\n");
381 381
382 len += snprintf(buf + len, sizeof(buf) - len, 382 len += scnprintf(buf + len, sizeof(buf) - len,
383 "Used slots : %d\n", 383 "Used slots : %d\n",
384 bitmap_weight(priv->tx.tx_slot, MAX_TX_BUF_NUM)); 384 bitmap_weight(priv->tx.tx_slot, MAX_TX_BUF_NUM));
385 385
386 spin_unlock_bh(&priv->tx.tx_lock); 386 spin_unlock_bh(&priv->tx.tx_lock);
387 387
@@ -405,30 +405,30 @@ static ssize_t read_file_queue(struct file *file, char __user *user_buf,
405 char buf[512]; 405 char buf[512];
406 unsigned int len = 0; 406 unsigned int len = 0;
407 407
408 len += snprintf(buf + len, sizeof(buf) - len, "%20s : %10u\n", 408 len += scnprintf(buf + len, sizeof(buf) - len, "%20s : %10u\n",
409 "Mgmt endpoint", skb_queue_len(&priv->tx.mgmt_ep_queue)); 409 "Mgmt endpoint", skb_queue_len(&priv->tx.mgmt_ep_queue));
410 410
411 len += snprintf(buf + len, sizeof(buf) - len, "%20s : %10u\n", 411 len += scnprintf(buf + len, sizeof(buf) - len, "%20s : %10u\n",
412 "Cab endpoint", skb_queue_len(&priv->tx.cab_ep_queue)); 412 "Cab endpoint", skb_queue_len(&priv->tx.cab_ep_queue));
413 413
414 len += snprintf(buf + len, sizeof(buf) - len, "%20s : %10u\n", 414 len += scnprintf(buf + len, sizeof(buf) - len, "%20s : %10u\n",
415 "Data BE endpoint", skb_queue_len(&priv->tx.data_be_queue)); 415 "Data BE endpoint", skb_queue_len(&priv->tx.data_be_queue));
416 416
417 len += snprintf(buf + len, sizeof(buf) - len, "%20s : %10u\n", 417 len += scnprintf(buf + len, sizeof(buf) - len, "%20s : %10u\n",
418 "Data BK endpoint", skb_queue_len(&priv->tx.data_bk_queue)); 418 "Data BK endpoint", skb_queue_len(&priv->tx.data_bk_queue));
419 419
420 len += snprintf(buf + len, sizeof(buf) - len, "%20s : %10u\n", 420 len += scnprintf(buf + len, sizeof(buf) - len, "%20s : %10u\n",
421 "Data VI endpoint", skb_queue_len(&priv->tx.data_vi_queue)); 421 "Data VI endpoint", skb_queue_len(&priv->tx.data_vi_queue));
422 422
423 len += snprintf(buf + len, sizeof(buf) - len, "%20s : %10u\n", 423 len += scnprintf(buf + len, sizeof(buf) - len, "%20s : %10u\n",
424 "Data VO endpoint", skb_queue_len(&priv->tx.data_vo_queue)); 424 "Data VO endpoint", skb_queue_len(&priv->tx.data_vo_queue));
425 425
426 len += snprintf(buf + len, sizeof(buf) - len, "%20s : %10u\n", 426 len += scnprintf(buf + len, sizeof(buf) - len, "%20s : %10u\n",
427 "Failed queue", skb_queue_len(&priv->tx.tx_failed)); 427 "Failed queue", skb_queue_len(&priv->tx.tx_failed));
428 428
429 spin_lock_bh(&priv->tx.tx_lock); 429 spin_lock_bh(&priv->tx.tx_lock);
430 len += snprintf(buf + len, sizeof(buf) - len, "%20s : %10u\n", 430 len += scnprintf(buf + len, sizeof(buf) - len, "%20s : %10u\n",
431 "Queued count", priv->tx.queued_cnt); 431 "Queued count", priv->tx.queued_cnt);
432 spin_unlock_bh(&priv->tx.tx_lock); 432 spin_unlock_bh(&priv->tx.tx_lock);
433 433
434 if (len > sizeof(buf)) 434 if (len > sizeof(buf))
@@ -507,70 +507,70 @@ static ssize_t read_file_base_eeprom(struct file *file, char __user *user_buf,
507 if (buf == NULL) 507 if (buf == NULL)
508 return -ENOMEM; 508 return -ENOMEM;
509 509
510 len += snprintf(buf + len, size - len, 510 len += scnprintf(buf + len, size - len,
511 "%20s : %10d\n", "Major Version", 511 "%20s : %10d\n", "Major Version",
512 pBase->version >> 12); 512 pBase->version >> 12);
513 len += snprintf(buf + len, size - len, 513 len += scnprintf(buf + len, size - len,
514 "%20s : %10d\n", "Minor Version", 514 "%20s : %10d\n", "Minor Version",
515 pBase->version & 0xFFF); 515 pBase->version & 0xFFF);
516 len += snprintf(buf + len, size - len, 516 len += scnprintf(buf + len, size - len,
517 "%20s : %10d\n", "Checksum", 517 "%20s : %10d\n", "Checksum",
518 pBase->checksum); 518 pBase->checksum);
519 len += snprintf(buf + len, size - len, 519 len += scnprintf(buf + len, size - len,
520 "%20s : %10d\n", "Length", 520 "%20s : %10d\n", "Length",
521 pBase->length); 521 pBase->length);
522 len += snprintf(buf + len, size - len, 522 len += scnprintf(buf + len, size - len,
523 "%20s : %10d\n", "RegDomain1", 523 "%20s : %10d\n", "RegDomain1",
524 pBase->regDmn[0]); 524 pBase->regDmn[0]);
525 len += snprintf(buf + len, size - len, 525 len += scnprintf(buf + len, size - len,
526 "%20s : %10d\n", "RegDomain2", 526 "%20s : %10d\n", "RegDomain2",
527 pBase->regDmn[1]); 527 pBase->regDmn[1]);
528 len += snprintf(buf + len, size - len, 528 len += scnprintf(buf + len, size - len,
529 "%20s : %10d\n", 529 "%20s : %10d\n",
530 "TX Mask", pBase->txMask); 530 "TX Mask", pBase->txMask);
531 len += snprintf(buf + len, size - len, 531 len += scnprintf(buf + len, size - len,
532 "%20s : %10d\n", 532 "%20s : %10d\n",
533 "RX Mask", pBase->rxMask); 533 "RX Mask", pBase->rxMask);
534 len += snprintf(buf + len, size - len, 534 len += scnprintf(buf + len, size - len,
535 "%20s : %10d\n", 535 "%20s : %10d\n",
536 "Allow 5GHz", 536 "Allow 5GHz",
537 !!(pBase->opCapFlags & AR5416_OPFLAGS_11A)); 537 !!(pBase->opCapFlags & AR5416_OPFLAGS_11A));
538 len += snprintf(buf + len, size - len, 538 len += scnprintf(buf + len, size - len,
539 "%20s : %10d\n", 539 "%20s : %10d\n",
540 "Allow 2GHz", 540 "Allow 2GHz",
541 !!(pBase->opCapFlags & AR5416_OPFLAGS_11G)); 541 !!(pBase->opCapFlags & AR5416_OPFLAGS_11G));
542 len += snprintf(buf + len, size - len, 542 len += scnprintf(buf + len, size - len,
543 "%20s : %10d\n", 543 "%20s : %10d\n",
544 "Disable 2GHz HT20", 544 "Disable 2GHz HT20",
545 !!(pBase->opCapFlags & AR5416_OPFLAGS_N_2G_HT20)); 545 !!(pBase->opCapFlags & AR5416_OPFLAGS_N_2G_HT20));
546 len += snprintf(buf + len, size - len, 546 len += scnprintf(buf + len, size - len,
547 "%20s : %10d\n", 547 "%20s : %10d\n",
548 "Disable 2GHz HT40", 548 "Disable 2GHz HT40",
549 !!(pBase->opCapFlags & AR5416_OPFLAGS_N_2G_HT40)); 549 !!(pBase->opCapFlags & AR5416_OPFLAGS_N_2G_HT40));
550 len += snprintf(buf + len, size - len, 550 len += scnprintf(buf + len, size - len,
551 "%20s : %10d\n", 551 "%20s : %10d\n",
552 "Disable 5Ghz HT20", 552 "Disable 5Ghz HT20",
553 !!(pBase->opCapFlags & AR5416_OPFLAGS_N_5G_HT20)); 553 !!(pBase->opCapFlags & AR5416_OPFLAGS_N_5G_HT20));
554 len += snprintf(buf + len, size - len, 554 len += scnprintf(buf + len, size - len,
555 "%20s : %10d\n", 555 "%20s : %10d\n",
556 "Disable 5Ghz HT40", 556 "Disable 5Ghz HT40",
557 !!(pBase->opCapFlags & AR5416_OPFLAGS_N_5G_HT40)); 557 !!(pBase->opCapFlags & AR5416_OPFLAGS_N_5G_HT40));
558 len += snprintf(buf + len, size - len, 558 len += scnprintf(buf + len, size - len,
559 "%20s : %10d\n", 559 "%20s : %10d\n",
560 "Big Endian", 560 "Big Endian",
561 !!(pBase->eepMisc & 0x01)); 561 !!(pBase->eepMisc & 0x01));
562 len += snprintf(buf + len, size - len, 562 len += scnprintf(buf + len, size - len,
563 "%20s : %10d\n", 563 "%20s : %10d\n",
564 "Cal Bin Major Ver", 564 "Cal Bin Major Ver",
565 (pBase->binBuildNumber >> 24) & 0xFF); 565 (pBase->binBuildNumber >> 24) & 0xFF);
566 len += snprintf(buf + len, size - len, 566 len += scnprintf(buf + len, size - len,
567 "%20s : %10d\n", 567 "%20s : %10d\n",
568 "Cal Bin Minor Ver", 568 "Cal Bin Minor Ver",
569 (pBase->binBuildNumber >> 16) & 0xFF); 569 (pBase->binBuildNumber >> 16) & 0xFF);
570 len += snprintf(buf + len, size - len, 570 len += scnprintf(buf + len, size - len,
571 "%20s : %10d\n", 571 "%20s : %10d\n",
572 "Cal Bin Build", 572 "Cal Bin Build",
573 (pBase->binBuildNumber >> 8) & 0xFF); 573 (pBase->binBuildNumber >> 8) & 0xFF);
574 574
575 /* 575 /*
576 * UB91 specific data. 576 * UB91 specific data.
@@ -579,10 +579,10 @@ static ssize_t read_file_base_eeprom(struct file *file, char __user *user_buf,
579 struct base_eep_header_4k *pBase4k = 579 struct base_eep_header_4k *pBase4k =
580 &priv->ah->eeprom.map4k.baseEepHeader; 580 &priv->ah->eeprom.map4k.baseEepHeader;
581 581
582 len += snprintf(buf + len, size - len, 582 len += scnprintf(buf + len, size - len,
583 "%20s : %10d\n", 583 "%20s : %10d\n",
584 "TX Gain type", 584 "TX Gain type",
585 pBase4k->txGainType); 585 pBase4k->txGainType);
586 } 586 }
587 587
588 /* 588 /*
@@ -592,19 +592,19 @@ static ssize_t read_file_base_eeprom(struct file *file, char __user *user_buf,
592 struct base_eep_ar9287_header *pBase9287 = 592 struct base_eep_ar9287_header *pBase9287 =
593 &priv->ah->eeprom.map9287.baseEepHeader; 593 &priv->ah->eeprom.map9287.baseEepHeader;
594 594
595 len += snprintf(buf + len, size - len, 595 len += scnprintf(buf + len, size - len,
596 "%20s : %10ddB\n", 596 "%20s : %10ddB\n",
597 "Power Table Offset", 597 "Power Table Offset",
598 pBase9287->pwrTableOffset); 598 pBase9287->pwrTableOffset);
599 599
600 len += snprintf(buf + len, size - len, 600 len += scnprintf(buf + len, size - len,
601 "%20s : %10d\n", 601 "%20s : %10d\n",
602 "OpenLoop Power Ctrl", 602 "OpenLoop Power Ctrl",
603 pBase9287->openLoopPwrCntl); 603 pBase9287->openLoopPwrCntl);
604 } 604 }
605 605
606 len += snprintf(buf + len, size - len, "%20s : %pM\n", "MacAddress", 606 len += scnprintf(buf + len, size - len, "%20s : %pM\n", "MacAddress",
607 pBase->macAddr); 607 pBase->macAddr);
608 if (len > size) 608 if (len > size)
609 len = size; 609 len = size;
610 610
@@ -627,8 +627,8 @@ static ssize_t read_4k_modal_eeprom(struct file *file,
627{ 627{
628#define PR_EEP(_s, _val) \ 628#define PR_EEP(_s, _val) \
629 do { \ 629 do { \
630 len += snprintf(buf + len, size - len, "%20s : %10d\n", \ 630 len += scnprintf(buf + len, size - len, "%20s : %10d\n",\
631 _s, (_val)); \ 631 _s, (_val)); \
632 } while (0) 632 } while (0)
633 633
634 struct ath9k_htc_priv *priv = file->private_data; 634 struct ath9k_htc_priv *priv = file->private_data;
@@ -708,12 +708,12 @@ static ssize_t read_def_modal_eeprom(struct file *file,
708 do { \ 708 do { \
709 if (pBase->opCapFlags & AR5416_OPFLAGS_11G) { \ 709 if (pBase->opCapFlags & AR5416_OPFLAGS_11G) { \
710 pModal = &priv->ah->eeprom.def.modalHeader[1]; \ 710 pModal = &priv->ah->eeprom.def.modalHeader[1]; \
711 len += snprintf(buf + len, size - len, "%20s : %8d%7s", \ 711 len += scnprintf(buf + len, size - len, "%20s : %8d%7s", \
712 _s, (_val), "|"); \ 712 _s, (_val), "|"); \
713 } \ 713 } \
714 if (pBase->opCapFlags & AR5416_OPFLAGS_11A) { \ 714 if (pBase->opCapFlags & AR5416_OPFLAGS_11A) { \
715 pModal = &priv->ah->eeprom.def.modalHeader[0]; \ 715 pModal = &priv->ah->eeprom.def.modalHeader[0]; \
716 len += snprintf(buf + len, size - len, "%9d\n", \ 716 len += scnprintf(buf + len, size - len, "%9d\n",\
717 (_val)); \ 717 (_val)); \
718 } \ 718 } \
719 } while (0) 719 } while (0)
@@ -729,10 +729,10 @@ static ssize_t read_def_modal_eeprom(struct file *file,
729 if (buf == NULL) 729 if (buf == NULL)
730 return -ENOMEM; 730 return -ENOMEM;
731 731
732 len += snprintf(buf + len, size - len, 732 len += scnprintf(buf + len, size - len,
733 "%31s %15s\n", "2G", "5G"); 733 "%31s %15s\n", "2G", "5G");
734 len += snprintf(buf + len, size - len, 734 len += scnprintf(buf + len, size - len,
735 "%32s %16s\n", "====", "====\n"); 735 "%32s %16s\n", "====", "====\n");
736 736
737 PR_EEP("Chain0 Ant. Control", pModal->antCtrlChain[0]); 737 PR_EEP("Chain0 Ant. Control", pModal->antCtrlChain[0]);
738 PR_EEP("Chain1 Ant. Control", pModal->antCtrlChain[1]); 738 PR_EEP("Chain1 Ant. Control", pModal->antCtrlChain[1]);
@@ -814,8 +814,8 @@ static ssize_t read_9287_modal_eeprom(struct file *file,
814{ 814{
815#define PR_EEP(_s, _val) \ 815#define PR_EEP(_s, _val) \
816 do { \ 816 do { \
817 len += snprintf(buf + len, size - len, "%20s : %10d\n", \ 817 len += scnprintf(buf + len, size - len, "%20s : %10d\n",\
818 _s, (_val)); \ 818 _s, (_val)); \
819 } while (0) 819 } while (0)
820 820
821 struct ath9k_htc_priv *priv = file->private_data; 821 struct ath9k_htc_priv *priv = file->private_data;
diff --git a/drivers/net/wireless/ath/ath9k/htc_drv_main.c b/drivers/net/wireless/ath/ath9k/htc_drv_main.c
index d44258172c0f..9a2657fdd9cc 100644
--- a/drivers/net/wireless/ath/ath9k/htc_drv_main.c
+++ b/drivers/net/wireless/ath/ath9k/htc_drv_main.c
@@ -24,30 +24,10 @@
24static enum htc_phymode ath9k_htc_get_curmode(struct ath9k_htc_priv *priv, 24static enum htc_phymode ath9k_htc_get_curmode(struct ath9k_htc_priv *priv,
25 struct ath9k_channel *ichan) 25 struct ath9k_channel *ichan)
26{ 26{
27 enum htc_phymode mode; 27 if (IS_CHAN_5GHZ(ichan))
28 28 return HTC_MODE_11NA;
29 mode = -EINVAL;
30
31 switch (ichan->chanmode) {
32 case CHANNEL_G:
33 case CHANNEL_G_HT20:
34 case CHANNEL_G_HT40PLUS:
35 case CHANNEL_G_HT40MINUS:
36 mode = HTC_MODE_11NG;
37 break;
38 case CHANNEL_A:
39 case CHANNEL_A_HT20:
40 case CHANNEL_A_HT40PLUS:
41 case CHANNEL_A_HT40MINUS:
42 mode = HTC_MODE_11NA;
43 break;
44 default:
45 break;
46 }
47 29
48 WARN_ON(mode < 0); 30 return HTC_MODE_11NG;
49
50 return mode;
51} 31}
52 32
53bool ath9k_htc_setpower(struct ath9k_htc_priv *priv, 33bool ath9k_htc_setpower(struct ath9k_htc_priv *priv,
@@ -926,7 +906,7 @@ static int ath9k_htc_start(struct ieee80211_hw *hw)
926 WMI_CMD(WMI_FLUSH_RECV_CMDID); 906 WMI_CMD(WMI_FLUSH_RECV_CMDID);
927 907
928 /* setup initial channel */ 908 /* setup initial channel */
929 init_channel = ath9k_cmn_get_curchannel(hw, ah); 909 init_channel = ath9k_cmn_get_channel(hw, ah, &hw->conf.chandef);
930 910
931 ret = ath9k_hw_reset(ah, init_channel, ah->caldata, false); 911 ret = ath9k_hw_reset(ah, init_channel, ah->caldata, false);
932 if (ret) { 912 if (ret) {
@@ -1208,9 +1188,7 @@ static int ath9k_htc_config(struct ieee80211_hw *hw, u32 changed)
1208 ath_dbg(common, CONFIG, "Set channel: %d MHz\n", 1188 ath_dbg(common, CONFIG, "Set channel: %d MHz\n",
1209 curchan->center_freq); 1189 curchan->center_freq);
1210 1190
1211 ath9k_cmn_update_ichannel(&priv->ah->channels[pos], 1191 ath9k_cmn_get_channel(hw, priv->ah, &hw->conf.chandef);
1212 &hw->conf.chandef);
1213
1214 if (ath9k_htc_set_channel(priv, hw, &priv->ah->channels[pos]) < 0) { 1192 if (ath9k_htc_set_channel(priv, hw, &priv->ah->channels[pos]) < 0) {
1215 ath_err(common, "Unable to set channel\n"); 1193 ath_err(common, "Unable to set channel\n");
1216 ret = -EINVAL; 1194 ret = -EINVAL;
diff --git a/drivers/net/wireless/ath/ath9k/hw-ops.h b/drivers/net/wireless/ath/ath9k/hw-ops.h
index 83f4927aeaca..4f9378ddf07f 100644
--- a/drivers/net/wireless/ath/ath9k/hw-ops.h
+++ b/drivers/net/wireless/ath/ath9k/hw-ops.h
@@ -78,6 +78,22 @@ static inline void ath9k_hw_antdiv_comb_conf_set(struct ath_hw *ah,
78 ath9k_hw_ops(ah)->antdiv_comb_conf_set(ah, antconf); 78 ath9k_hw_ops(ah)->antdiv_comb_conf_set(ah, antconf);
79} 79}
80 80
81static inline void ath9k_hw_tx99_start(struct ath_hw *ah, u32 qnum)
82{
83 ath9k_hw_ops(ah)->tx99_start(ah, qnum);
84}
85
86static inline void ath9k_hw_tx99_stop(struct ath_hw *ah)
87{
88 ath9k_hw_ops(ah)->tx99_stop(ah);
89}
90
91static inline void ath9k_hw_tx99_set_txpower(struct ath_hw *ah, u8 power)
92{
93 if (ath9k_hw_ops(ah)->tx99_set_txpower)
94 ath9k_hw_ops(ah)->tx99_set_txpower(ah, power);
95}
96
81#ifdef CONFIG_ATH9K_BTCOEX_SUPPORT 97#ifdef CONFIG_ATH9K_BTCOEX_SUPPORT
82 98
83static inline void ath9k_hw_set_bt_ant_diversity(struct ath_hw *ah, bool enable) 99static inline void ath9k_hw_set_bt_ant_diversity(struct ath_hw *ah, bool enable)
diff --git a/drivers/net/wireless/ath/ath9k/hw.c b/drivers/net/wireless/ath/ath9k/hw.c
index ecc6ec4a1edb..54b04155e43b 100644
--- a/drivers/net/wireless/ath/ath9k/hw.c
+++ b/drivers/net/wireless/ath/ath9k/hw.c
@@ -130,29 +130,29 @@ void ath9k_debug_sync_cause(struct ath_common *common, u32 sync_cause)
130 130
131static void ath9k_hw_set_clockrate(struct ath_hw *ah) 131static void ath9k_hw_set_clockrate(struct ath_hw *ah)
132{ 132{
133 struct ieee80211_conf *conf = &ath9k_hw_common(ah)->hw->conf;
134 struct ath_common *common = ath9k_hw_common(ah); 133 struct ath_common *common = ath9k_hw_common(ah);
134 struct ath9k_channel *chan = ah->curchan;
135 unsigned int clockrate; 135 unsigned int clockrate;
136 136
137 /* AR9287 v1.3+ uses async FIFO and runs the MAC at 117 MHz */ 137 /* AR9287 v1.3+ uses async FIFO and runs the MAC at 117 MHz */
138 if (AR_SREV_9287(ah) && AR_SREV_9287_13_OR_LATER(ah)) 138 if (AR_SREV_9287(ah) && AR_SREV_9287_13_OR_LATER(ah))
139 clockrate = 117; 139 clockrate = 117;
140 else if (!ah->curchan) /* should really check for CCK instead */ 140 else if (!chan) /* should really check for CCK instead */
141 clockrate = ATH9K_CLOCK_RATE_CCK; 141 clockrate = ATH9K_CLOCK_RATE_CCK;
142 else if (conf->chandef.chan->band == IEEE80211_BAND_2GHZ) 142 else if (IS_CHAN_2GHZ(chan))
143 clockrate = ATH9K_CLOCK_RATE_2GHZ_OFDM; 143 clockrate = ATH9K_CLOCK_RATE_2GHZ_OFDM;
144 else if (ah->caps.hw_caps & ATH9K_HW_CAP_FASTCLOCK) 144 else if (ah->caps.hw_caps & ATH9K_HW_CAP_FASTCLOCK)
145 clockrate = ATH9K_CLOCK_FAST_RATE_5GHZ_OFDM; 145 clockrate = ATH9K_CLOCK_FAST_RATE_5GHZ_OFDM;
146 else 146 else
147 clockrate = ATH9K_CLOCK_RATE_5GHZ_OFDM; 147 clockrate = ATH9K_CLOCK_RATE_5GHZ_OFDM;
148 148
149 if (conf_is_ht40(conf)) 149 if (IS_CHAN_HT40(chan))
150 clockrate *= 2; 150 clockrate *= 2;
151 151
152 if (ah->curchan) { 152 if (ah->curchan) {
153 if (IS_CHAN_HALF_RATE(ah->curchan)) 153 if (IS_CHAN_HALF_RATE(chan))
154 clockrate /= 2; 154 clockrate /= 2;
155 if (IS_CHAN_QUARTER_RATE(ah->curchan)) 155 if (IS_CHAN_QUARTER_RATE(chan))
156 clockrate /= 4; 156 clockrate /= 4;
157 } 157 }
158 158
@@ -190,10 +190,7 @@ EXPORT_SYMBOL(ath9k_hw_wait);
190void ath9k_hw_synth_delay(struct ath_hw *ah, struct ath9k_channel *chan, 190void ath9k_hw_synth_delay(struct ath_hw *ah, struct ath9k_channel *chan,
191 int hw_delay) 191 int hw_delay)
192{ 192{
193 if (IS_CHAN_B(chan)) 193 hw_delay /= 10;
194 hw_delay = (4 * hw_delay) / 22;
195 else
196 hw_delay /= 10;
197 194
198 if (IS_CHAN_HALF_RATE(chan)) 195 if (IS_CHAN_HALF_RATE(chan))
199 hw_delay *= 2; 196 hw_delay *= 2;
@@ -294,8 +291,7 @@ void ath9k_hw_get_channel_centers(struct ath_hw *ah,
294 return; 291 return;
295 } 292 }
296 293
297 if ((chan->chanmode == CHANNEL_A_HT40PLUS) || 294 if (IS_CHAN_HT40PLUS(chan)) {
298 (chan->chanmode == CHANNEL_G_HT40PLUS)) {
299 centers->synth_center = 295 centers->synth_center =
300 chan->channel + HT40_CHANNEL_CENTER_SHIFT; 296 chan->channel + HT40_CHANNEL_CENTER_SHIFT;
301 extoff = 1; 297 extoff = 1;
@@ -549,6 +545,18 @@ static int ath9k_hw_post_init(struct ath_hw *ah)
549 545
550 ath9k_hw_ani_init(ah); 546 ath9k_hw_ani_init(ah);
551 547
548 /*
549 * EEPROM needs to be initialized before we do this.
550 * This is required for regulatory compliance.
551 */
552 if (AR_SREV_9462(ah) || AR_SREV_9565(ah)) {
553 u16 regdmn = ah->eep_ops->get_eeprom(ah, EEP_REG_0);
554 if ((regdmn & 0xF0) == CTL_FCC) {
555 ah->nf_2g.max = AR_PHY_CCA_MAX_GOOD_VAL_9462_FCC_2GHZ;
556 ah->nf_5g.max = AR_PHY_CCA_MAX_GOOD_VAL_9462_FCC_5GHZ;
557 }
558 }
559
552 return 0; 560 return 0;
553} 561}
554 562
@@ -1030,7 +1038,6 @@ static bool ath9k_hw_set_global_txtimeout(struct ath_hw *ah, u32 tu)
1030void ath9k_hw_init_global_settings(struct ath_hw *ah) 1038void ath9k_hw_init_global_settings(struct ath_hw *ah)
1031{ 1039{
1032 struct ath_common *common = ath9k_hw_common(ah); 1040 struct ath_common *common = ath9k_hw_common(ah);
1033 struct ieee80211_conf *conf = &common->hw->conf;
1034 const struct ath9k_channel *chan = ah->curchan; 1041 const struct ath9k_channel *chan = ah->curchan;
1035 int acktimeout, ctstimeout, ack_offset = 0; 1042 int acktimeout, ctstimeout, ack_offset = 0;
1036 int slottime; 1043 int slottime;
@@ -1105,8 +1112,7 @@ void ath9k_hw_init_global_settings(struct ath_hw *ah)
1105 * BA frames in some implementations, but it has been found to fix ACK 1112 * BA frames in some implementations, but it has been found to fix ACK
1106 * timeout issues in other cases as well. 1113 * timeout issues in other cases as well.
1107 */ 1114 */
1108 if (conf->chandef.chan && 1115 if (IS_CHAN_2GHZ(chan) &&
1109 conf->chandef.chan->band == IEEE80211_BAND_2GHZ &&
1110 !IS_CHAN_HALF_RATE(chan) && !IS_CHAN_QUARTER_RATE(chan)) { 1116 !IS_CHAN_HALF_RATE(chan) && !IS_CHAN_QUARTER_RATE(chan)) {
1111 acktimeout += 64 - sifstime - ah->slottime; 1117 acktimeout += 64 - sifstime - ah->slottime;
1112 ctstimeout += 48 - sifstime - ah->slottime; 1118 ctstimeout += 48 - sifstime - ah->slottime;
@@ -1148,9 +1154,7 @@ u32 ath9k_regd_get_ctl(struct ath_regulatory *reg, struct ath9k_channel *chan)
1148{ 1154{
1149 u32 ctl = ath_regd_get_band_ctl(reg, chan->chan->band); 1155 u32 ctl = ath_regd_get_band_ctl(reg, chan->chan->band);
1150 1156
1151 if (IS_CHAN_B(chan)) 1157 if (IS_CHAN_2GHZ(chan))
1152 ctl |= CTL_11B;
1153 else if (IS_CHAN_G(chan))
1154 ctl |= CTL_11G; 1158 ctl |= CTL_11G;
1155 else 1159 else
1156 ctl |= CTL_11A; 1160 ctl |= CTL_11A;
@@ -1498,10 +1502,8 @@ static bool ath9k_hw_channel_change(struct ath_hw *ah,
1498 int r; 1502 int r;
1499 1503
1500 if (pCap->hw_caps & ATH9K_HW_CAP_FCC_BAND_SWITCH) { 1504 if (pCap->hw_caps & ATH9K_HW_CAP_FCC_BAND_SWITCH) {
1501 u32 cur = ah->curchan->channelFlags & (CHANNEL_2GHZ | CHANNEL_5GHZ); 1505 band_switch = IS_CHAN_5GHZ(ah->curchan) != IS_CHAN_5GHZ(chan);
1502 u32 new = chan->channelFlags & (CHANNEL_2GHZ | CHANNEL_5GHZ); 1506 mode_diff = (chan->channelFlags != ah->curchan->channelFlags);
1503 band_switch = (cur != new);
1504 mode_diff = (chan->chanmode != ah->curchan->chanmode);
1505 } 1507 }
1506 1508
1507 for (qnum = 0; qnum < AR_NUM_QCU; qnum++) { 1509 for (qnum = 0; qnum < AR_NUM_QCU; qnum++) {
@@ -1540,9 +1542,7 @@ static bool ath9k_hw_channel_change(struct ath_hw *ah,
1540 ath9k_hw_set_clockrate(ah); 1542 ath9k_hw_set_clockrate(ah);
1541 ath9k_hw_apply_txpower(ah, chan, false); 1543 ath9k_hw_apply_txpower(ah, chan, false);
1542 1544
1543 if (IS_CHAN_OFDM(chan) || IS_CHAN_HT(chan)) 1545 ath9k_hw_set_delta_slope(ah, chan);
1544 ath9k_hw_set_delta_slope(ah, chan);
1545
1546 ath9k_hw_spur_mitigate_freq(ah, chan); 1546 ath9k_hw_spur_mitigate_freq(ah, chan);
1547 1547
1548 if (band_switch || ini_reloaded) 1548 if (band_switch || ini_reloaded)
@@ -1644,6 +1644,19 @@ hang_check_iter:
1644 return true; 1644 return true;
1645} 1645}
1646 1646
1647void ath9k_hw_check_nav(struct ath_hw *ah)
1648{
1649 struct ath_common *common = ath9k_hw_common(ah);
1650 u32 val;
1651
1652 val = REG_READ(ah, AR_NAV);
1653 if (val != 0xdeadbeef && val > 0x7fff) {
1654 ath_dbg(common, BSTUCK, "Abnormal NAV: 0x%x\n", val);
1655 REG_WRITE(ah, AR_NAV, 0);
1656 }
1657}
1658EXPORT_SYMBOL(ath9k_hw_check_nav);
1659
1647bool ath9k_hw_check_alive(struct ath_hw *ah) 1660bool ath9k_hw_check_alive(struct ath_hw *ah)
1648{ 1661{
1649 int count = 50; 1662 int count = 50;
@@ -1799,20 +1812,11 @@ static int ath9k_hw_do_fastcc(struct ath_hw *ah, struct ath9k_channel *chan)
1799 goto fail; 1812 goto fail;
1800 1813
1801 /* 1814 /*
1802 * If cross-band fcc is not supoprted, bail out if 1815 * If cross-band fcc is not supoprted, bail out if channelFlags differ.
1803 * either channelFlags or chanmode differ.
1804 *
1805 * chanmode will be different if the HT operating mode
1806 * changes because of CSA.
1807 */ 1816 */
1808 if (!(pCap->hw_caps & ATH9K_HW_CAP_FCC_BAND_SWITCH)) { 1817 if (!(pCap->hw_caps & ATH9K_HW_CAP_FCC_BAND_SWITCH) &&
1809 if ((chan->channelFlags & CHANNEL_ALL) != 1818 chan->channelFlags != ah->curchan->channelFlags)
1810 (ah->curchan->channelFlags & CHANNEL_ALL)) 1819 goto fail;
1811 goto fail;
1812
1813 if (chan->chanmode != ah->curchan->chanmode)
1814 goto fail;
1815 }
1816 1820
1817 if (!ath9k_hw_check_alive(ah)) 1821 if (!ath9k_hw_check_alive(ah))
1818 goto fail; 1822 goto fail;
@@ -1822,9 +1826,9 @@ static int ath9k_hw_do_fastcc(struct ath_hw *ah, struct ath9k_channel *chan)
1822 * re-using are present. 1826 * re-using are present.
1823 */ 1827 */
1824 if (AR_SREV_9462(ah) && (ah->caldata && 1828 if (AR_SREV_9462(ah) && (ah->caldata &&
1825 (!ah->caldata->done_txiqcal_once || 1829 (!test_bit(TXIQCAL_DONE, &ah->caldata->cal_flags) ||
1826 !ah->caldata->done_txclcal_once || 1830 !test_bit(TXCLCAL_DONE, &ah->caldata->cal_flags) ||
1827 !ah->caldata->rtt_done))) 1831 !test_bit(RTT_DONE, &ah->caldata->cal_flags))))
1828 goto fail; 1832 goto fail;
1829 1833
1830 ath_dbg(common, RESET, "FastChannelChange for %d -> %d\n", 1834 ath_dbg(common, RESET, "FastChannelChange for %d -> %d\n",
@@ -1874,15 +1878,14 @@ int ath9k_hw_reset(struct ath_hw *ah, struct ath9k_channel *chan,
1874 1878
1875 ah->caldata = caldata; 1879 ah->caldata = caldata;
1876 if (caldata && (chan->channel != caldata->channel || 1880 if (caldata && (chan->channel != caldata->channel ||
1877 chan->channelFlags != caldata->channelFlags || 1881 chan->channelFlags != caldata->channelFlags)) {
1878 chan->chanmode != caldata->chanmode)) {
1879 /* Operating channel changed, reset channel calibration data */ 1882 /* Operating channel changed, reset channel calibration data */
1880 memset(caldata, 0, sizeof(*caldata)); 1883 memset(caldata, 0, sizeof(*caldata));
1881 ath9k_init_nfcal_hist_buffer(ah, chan); 1884 ath9k_init_nfcal_hist_buffer(ah, chan);
1882 } else if (caldata) { 1885 } else if (caldata) {
1883 caldata->paprd_packet_sent = false; 1886 clear_bit(PAPRD_PACKET_SENT, &caldata->cal_flags);
1884 } 1887 }
1885 ah->noise = ath9k_hw_getchan_noise(ah, chan); 1888 ah->noise = ath9k_hw_getchan_noise(ah, chan, chan->noisefloor);
1886 1889
1887 if (fastcc) { 1890 if (fastcc) {
1888 r = ath9k_hw_do_fastcc(ah, chan); 1891 r = ath9k_hw_do_fastcc(ah, chan);
@@ -1964,9 +1967,7 @@ int ath9k_hw_reset(struct ath_hw *ah, struct ath9k_channel *chan,
1964 1967
1965 ath9k_hw_init_mfp(ah); 1968 ath9k_hw_init_mfp(ah);
1966 1969
1967 if (IS_CHAN_OFDM(chan) || IS_CHAN_HT(chan)) 1970 ath9k_hw_set_delta_slope(ah, chan);
1968 ath9k_hw_set_delta_slope(ah, chan);
1969
1970 ath9k_hw_spur_mitigate_freq(ah, chan); 1971 ath9k_hw_spur_mitigate_freq(ah, chan);
1971 ah->eep_ops->set_board_values(ah, chan); 1972 ah->eep_ops->set_board_values(ah, chan);
1972 1973
@@ -2017,8 +2018,8 @@ int ath9k_hw_reset(struct ath_hw *ah, struct ath9k_channel *chan,
2017 ath9k_hw_init_bb(ah, chan); 2018 ath9k_hw_init_bb(ah, chan);
2018 2019
2019 if (caldata) { 2020 if (caldata) {
2020 caldata->done_txiqcal_once = false; 2021 clear_bit(TXIQCAL_DONE, &caldata->cal_flags);
2021 caldata->done_txclcal_once = false; 2022 clear_bit(TXCLCAL_DONE, &caldata->cal_flags);
2022 } 2023 }
2023 if (!ath9k_hw_init_cal(ah, chan)) 2024 if (!ath9k_hw_init_cal(ah, chan))
2024 return -EIO; 2025 return -EIO;
@@ -2943,12 +2944,11 @@ void ath9k_hw_set_tsfadjust(struct ath_hw *ah, bool set)
2943} 2944}
2944EXPORT_SYMBOL(ath9k_hw_set_tsfadjust); 2945EXPORT_SYMBOL(ath9k_hw_set_tsfadjust);
2945 2946
2946void ath9k_hw_set11nmac2040(struct ath_hw *ah) 2947void ath9k_hw_set11nmac2040(struct ath_hw *ah, struct ath9k_channel *chan)
2947{ 2948{
2948 struct ieee80211_conf *conf = &ath9k_hw_common(ah)->hw->conf;
2949 u32 macmode; 2949 u32 macmode;
2950 2950
2951 if (conf_is_ht40(conf) && !ah->config.cwm_ignore_extcca) 2951 if (IS_CHAN_HT40(chan) && !ah->config.cwm_ignore_extcca)
2952 macmode = AR_2040_JOINED_RX_CLEAR; 2952 macmode = AR_2040_JOINED_RX_CLEAR;
2953 else 2953 else
2954 macmode = 0; 2954 macmode = 0;
@@ -3240,19 +3240,19 @@ void ath9k_hw_name(struct ath_hw *ah, char *hw_name, size_t len)
3240 3240
3241 /* chipsets >= AR9280 are single-chip */ 3241 /* chipsets >= AR9280 are single-chip */
3242 if (AR_SREV_9280_20_OR_LATER(ah)) { 3242 if (AR_SREV_9280_20_OR_LATER(ah)) {
3243 used = snprintf(hw_name, len, 3243 used = scnprintf(hw_name, len,
3244 "Atheros AR%s Rev:%x", 3244 "Atheros AR%s Rev:%x",
3245 ath9k_hw_mac_bb_name(ah->hw_version.macVersion), 3245 ath9k_hw_mac_bb_name(ah->hw_version.macVersion),
3246 ah->hw_version.macRev); 3246 ah->hw_version.macRev);
3247 } 3247 }
3248 else { 3248 else {
3249 used = snprintf(hw_name, len, 3249 used = scnprintf(hw_name, len,
3250 "Atheros AR%s MAC/BB Rev:%x AR%s RF Rev:%x", 3250 "Atheros AR%s MAC/BB Rev:%x AR%s RF Rev:%x",
3251 ath9k_hw_mac_bb_name(ah->hw_version.macVersion), 3251 ath9k_hw_mac_bb_name(ah->hw_version.macVersion),
3252 ah->hw_version.macRev, 3252 ah->hw_version.macRev,
3253 ath9k_hw_rf_name((ah->hw_version.analog5GhzRev & 3253 ath9k_hw_rf_name((ah->hw_version.analog5GhzRev
3254 AR_RADIO_SREV_MAJOR)), 3254 & AR_RADIO_SREV_MAJOR)),
3255 ah->hw_version.phyRev); 3255 ah->hw_version.phyRev);
3256 } 3256 }
3257 3257
3258 hw_name[used] = '\0'; 3258 hw_name[used] = '\0';
diff --git a/drivers/net/wireless/ath/ath9k/hw.h b/drivers/net/wireless/ath/ath9k/hw.h
index 69a907b55a73..9ea24f1cba73 100644
--- a/drivers/net/wireless/ath/ath9k/hw.h
+++ b/drivers/net/wireless/ath/ath9k/hw.h
@@ -98,8 +98,8 @@
98 98
99#define PR_EEP(_s, _val) \ 99#define PR_EEP(_s, _val) \
100 do { \ 100 do { \
101 len += snprintf(buf + len, size - len, "%20s : %10d\n", \ 101 len += scnprintf(buf + len, size - len, "%20s : %10d\n",\
102 _s, (_val)); \ 102 _s, (_val)); \
103 } while (0) 103 } while (0)
104 104
105#define SM(_v, _f) (((_v) << _f##_S) & _f) 105#define SM(_v, _f) (((_v) << _f##_S) & _f)
@@ -369,55 +369,30 @@ enum ath9k_int {
369 ATH9K_INT_NOCARD = 0xffffffff 369 ATH9K_INT_NOCARD = 0xffffffff
370}; 370};
371 371
372#define CHANNEL_CCK 0x00020
373#define CHANNEL_OFDM 0x00040
374#define CHANNEL_2GHZ 0x00080
375#define CHANNEL_5GHZ 0x00100
376#define CHANNEL_PASSIVE 0x00200
377#define CHANNEL_DYN 0x00400
378#define CHANNEL_HALF 0x04000
379#define CHANNEL_QUARTER 0x08000
380#define CHANNEL_HT20 0x10000
381#define CHANNEL_HT40PLUS 0x20000
382#define CHANNEL_HT40MINUS 0x40000
383
384#define CHANNEL_A (CHANNEL_5GHZ|CHANNEL_OFDM)
385#define CHANNEL_B (CHANNEL_2GHZ|CHANNEL_CCK)
386#define CHANNEL_G (CHANNEL_2GHZ|CHANNEL_OFDM)
387#define CHANNEL_G_HT20 (CHANNEL_2GHZ|CHANNEL_HT20)
388#define CHANNEL_A_HT20 (CHANNEL_5GHZ|CHANNEL_HT20)
389#define CHANNEL_G_HT40PLUS (CHANNEL_2GHZ|CHANNEL_HT40PLUS)
390#define CHANNEL_G_HT40MINUS (CHANNEL_2GHZ|CHANNEL_HT40MINUS)
391#define CHANNEL_A_HT40PLUS (CHANNEL_5GHZ|CHANNEL_HT40PLUS)
392#define CHANNEL_A_HT40MINUS (CHANNEL_5GHZ|CHANNEL_HT40MINUS)
393#define CHANNEL_ALL \
394 (CHANNEL_OFDM| \
395 CHANNEL_CCK| \
396 CHANNEL_2GHZ | \
397 CHANNEL_5GHZ | \
398 CHANNEL_HT20 | \
399 CHANNEL_HT40PLUS | \
400 CHANNEL_HT40MINUS)
401
402#define MAX_RTT_TABLE_ENTRY 6 372#define MAX_RTT_TABLE_ENTRY 6
403#define MAX_IQCAL_MEASUREMENT 8 373#define MAX_IQCAL_MEASUREMENT 8
404#define MAX_CL_TAB_ENTRY 16 374#define MAX_CL_TAB_ENTRY 16
405#define CL_TAB_ENTRY(reg_base) (reg_base + (4 * j)) 375#define CL_TAB_ENTRY(reg_base) (reg_base + (4 * j))
406 376
377enum ath9k_cal_flags {
378 RTT_DONE,
379 PAPRD_PACKET_SENT,
380 PAPRD_DONE,
381 NFCAL_PENDING,
382 NFCAL_INTF,
383 TXIQCAL_DONE,
384 TXCLCAL_DONE,
385 SW_PKDET_DONE,
386};
387
407struct ath9k_hw_cal_data { 388struct ath9k_hw_cal_data {
408 u16 channel; 389 u16 channel;
409 u32 channelFlags; 390 u16 channelFlags;
410 u32 chanmode; 391 unsigned long cal_flags;
411 int32_t CalValid; 392 int32_t CalValid;
412 int8_t iCoff; 393 int8_t iCoff;
413 int8_t qCoff; 394 int8_t qCoff;
414 bool rtt_done; 395 u8 caldac[2];
415 bool paprd_packet_sent;
416 bool paprd_done;
417 bool nfcal_pending;
418 bool nfcal_interference;
419 bool done_txiqcal_once;
420 bool done_txclcal_once;
421 u16 small_signal_gain[AR9300_MAX_CHAINS]; 396 u16 small_signal_gain[AR9300_MAX_CHAINS];
422 u32 pa_table[AR9300_MAX_CHAINS][PAPRD_TABLE_SZ]; 397 u32 pa_table[AR9300_MAX_CHAINS][PAPRD_TABLE_SZ];
423 u32 num_measures[AR9300_MAX_CHAINS]; 398 u32 num_measures[AR9300_MAX_CHAINS];
@@ -430,33 +405,34 @@ struct ath9k_hw_cal_data {
430struct ath9k_channel { 405struct ath9k_channel {
431 struct ieee80211_channel *chan; 406 struct ieee80211_channel *chan;
432 u16 channel; 407 u16 channel;
433 u32 channelFlags; 408 u16 channelFlags;
434 u32 chanmode;
435 s16 noisefloor; 409 s16 noisefloor;
436}; 410};
437 411
438#define IS_CHAN_G(_c) ((((_c)->channelFlags & (CHANNEL_G)) == CHANNEL_G) || \ 412#define CHANNEL_5GHZ BIT(0)
439 (((_c)->channelFlags & CHANNEL_G_HT20) == CHANNEL_G_HT20) || \ 413#define CHANNEL_HALF BIT(1)
440 (((_c)->channelFlags & CHANNEL_G_HT40PLUS) == CHANNEL_G_HT40PLUS) || \ 414#define CHANNEL_QUARTER BIT(2)
441 (((_c)->channelFlags & CHANNEL_G_HT40MINUS) == CHANNEL_G_HT40MINUS)) 415#define CHANNEL_HT BIT(3)
442#define IS_CHAN_OFDM(_c) (((_c)->channelFlags & CHANNEL_OFDM) != 0) 416#define CHANNEL_HT40PLUS BIT(4)
443#define IS_CHAN_5GHZ(_c) (((_c)->channelFlags & CHANNEL_5GHZ) != 0) 417#define CHANNEL_HT40MINUS BIT(5)
444#define IS_CHAN_2GHZ(_c) (((_c)->channelFlags & CHANNEL_2GHZ) != 0) 418
445#define IS_CHAN_HALF_RATE(_c) (((_c)->channelFlags & CHANNEL_HALF) != 0) 419#define IS_CHAN_5GHZ(_c) (!!((_c)->channelFlags & CHANNEL_5GHZ))
446#define IS_CHAN_QUARTER_RATE(_c) (((_c)->channelFlags & CHANNEL_QUARTER) != 0) 420#define IS_CHAN_2GHZ(_c) (!IS_CHAN_5GHZ(_c))
421
422#define IS_CHAN_HALF_RATE(_c) (!!((_c)->channelFlags & CHANNEL_HALF))
423#define IS_CHAN_QUARTER_RATE(_c) (!!((_c)->channelFlags & CHANNEL_QUARTER))
447#define IS_CHAN_A_FAST_CLOCK(_ah, _c) \ 424#define IS_CHAN_A_FAST_CLOCK(_ah, _c) \
448 ((((_c)->channelFlags & CHANNEL_5GHZ) != 0) && \ 425 (IS_CHAN_5GHZ(_c) && ((_ah)->caps.hw_caps & ATH9K_HW_CAP_FASTCLOCK))
449 ((_ah)->caps.hw_caps & ATH9K_HW_CAP_FASTCLOCK)) 426
450 427#define IS_CHAN_HT(_c) ((_c)->channelFlags & CHANNEL_HT)
451/* These macros check chanmode and not channelFlags */ 428
452#define IS_CHAN_B(_c) ((_c)->chanmode == CHANNEL_B) 429#define IS_CHAN_HT20(_c) (IS_CHAN_HT(_c) && !IS_CHAN_HT40(_c))
453#define IS_CHAN_HT20(_c) (((_c)->chanmode == CHANNEL_A_HT20) || \ 430
454 ((_c)->chanmode == CHANNEL_G_HT20)) 431#define IS_CHAN_HT40(_c) \
455#define IS_CHAN_HT40(_c) (((_c)->chanmode == CHANNEL_A_HT40PLUS) || \ 432 (!!((_c)->channelFlags & (CHANNEL_HT40PLUS | CHANNEL_HT40MINUS)))
456 ((_c)->chanmode == CHANNEL_A_HT40MINUS) || \ 433
457 ((_c)->chanmode == CHANNEL_G_HT40PLUS) || \ 434#define IS_CHAN_HT40PLUS(_c) ((_c)->channelFlags & CHANNEL_HT40PLUS)
458 ((_c)->chanmode == CHANNEL_G_HT40MINUS)) 435#define IS_CHAN_HT40MINUS(_c) ((_c)->channelFlags & CHANNEL_HT40MINUS)
459#define IS_CHAN_HT(_c) (IS_CHAN_HT20((_c)) || IS_CHAN_HT40((_c)))
460 436
461enum ath9k_power_mode { 437enum ath9k_power_mode {
462 ATH9K_PM_AWAKE = 0, 438 ATH9K_PM_AWAKE = 0,
@@ -558,6 +534,7 @@ struct ath_hw_antcomb_conf {
558 u8 main_gaintb; 534 u8 main_gaintb;
559 u8 alt_gaintb; 535 u8 alt_gaintb;
560 int lna1_lna2_delta; 536 int lna1_lna2_delta;
537 int lna1_lna2_switch_delta;
561 u8 div_group; 538 u8 div_group;
562}; 539};
563 540
@@ -726,6 +703,10 @@ struct ath_hw_ops {
726 void (*spectral_scan_trigger)(struct ath_hw *ah); 703 void (*spectral_scan_trigger)(struct ath_hw *ah);
727 void (*spectral_scan_wait)(struct ath_hw *ah); 704 void (*spectral_scan_wait)(struct ath_hw *ah);
728 705
706 void (*tx99_start)(struct ath_hw *ah, u32 qnum);
707 void (*tx99_stop)(struct ath_hw *ah);
708 void (*tx99_set_txpower)(struct ath_hw *ah, u8 power);
709
729#ifdef CONFIG_ATH9K_BTCOEX_SUPPORT 710#ifdef CONFIG_ATH9K_BTCOEX_SUPPORT
730 void (*set_bt_ant_diversity)(struct ath_hw *hw, bool enable); 711 void (*set_bt_ant_diversity)(struct ath_hw *hw, bool enable);
731#endif 712#endif
@@ -1026,10 +1007,11 @@ void ath9k_hw_reset_tsf(struct ath_hw *ah);
1026void ath9k_hw_set_tsfadjust(struct ath_hw *ah, bool set); 1007void ath9k_hw_set_tsfadjust(struct ath_hw *ah, bool set);
1027void ath9k_hw_init_global_settings(struct ath_hw *ah); 1008void ath9k_hw_init_global_settings(struct ath_hw *ah);
1028u32 ar9003_get_pll_sqsum_dvc(struct ath_hw *ah); 1009u32 ar9003_get_pll_sqsum_dvc(struct ath_hw *ah);
1029void ath9k_hw_set11nmac2040(struct ath_hw *ah); 1010void ath9k_hw_set11nmac2040(struct ath_hw *ah, struct ath9k_channel *chan);
1030void ath9k_hw_beaconinit(struct ath_hw *ah, u32 next_beacon, u32 beacon_period); 1011void ath9k_hw_beaconinit(struct ath_hw *ah, u32 next_beacon, u32 beacon_period);
1031void ath9k_hw_set_sta_beacon_timers(struct ath_hw *ah, 1012void ath9k_hw_set_sta_beacon_timers(struct ath_hw *ah,
1032 const struct ath9k_beacon_state *bs); 1013 const struct ath9k_beacon_state *bs);
1014void ath9k_hw_check_nav(struct ath_hw *ah);
1033bool ath9k_hw_check_alive(struct ath_hw *ah); 1015bool ath9k_hw_check_alive(struct ath_hw *ah);
1034 1016
1035bool ath9k_hw_setpower(struct ath_hw *ah, enum ath9k_power_mode mode); 1017bool ath9k_hw_setpower(struct ath_hw *ah, enum ath9k_power_mode mode);
diff --git a/drivers/net/wireless/ath/ath9k/init.c b/drivers/net/wireless/ath/ath9k/init.c
index 9a1f349f9260..d8643ebabd30 100644
--- a/drivers/net/wireless/ath/ath9k/init.c
+++ b/drivers/net/wireless/ath/ath9k/init.c
@@ -347,7 +347,6 @@ int ath_descdma_setup(struct ath_softc *sc, struct ath_descdma *dd,
347{ 347{
348 struct ath_common *common = ath9k_hw_common(sc->sc_ah); 348 struct ath_common *common = ath9k_hw_common(sc->sc_ah);
349 u8 *ds; 349 u8 *ds;
350 struct ath_buf *bf;
351 int i, bsize, desc_len; 350 int i, bsize, desc_len;
352 351
353 ath_dbg(common, CONFIG, "%s DMA: %u buffers %u desc/buf\n", 352 ath_dbg(common, CONFIG, "%s DMA: %u buffers %u desc/buf\n",
@@ -399,33 +398,68 @@ int ath_descdma_setup(struct ath_softc *sc, struct ath_descdma *dd,
399 ito64(dd->dd_desc_paddr), /*XXX*/(u32) dd->dd_desc_len); 398 ito64(dd->dd_desc_paddr), /*XXX*/(u32) dd->dd_desc_len);
400 399
401 /* allocate buffers */ 400 /* allocate buffers */
402 bsize = sizeof(struct ath_buf) * nbuf; 401 if (is_tx) {
403 bf = devm_kzalloc(sc->dev, bsize, GFP_KERNEL); 402 struct ath_buf *bf;
404 if (!bf) 403
405 return -ENOMEM; 404 bsize = sizeof(struct ath_buf) * nbuf;
405 bf = devm_kzalloc(sc->dev, bsize, GFP_KERNEL);
406 if (!bf)
407 return -ENOMEM;
408
409 for (i = 0; i < nbuf; i++, bf++, ds += (desc_len * ndesc)) {
410 bf->bf_desc = ds;
411 bf->bf_daddr = DS2PHYS(dd, ds);
412
413 if (!(sc->sc_ah->caps.hw_caps &
414 ATH9K_HW_CAP_4KB_SPLITTRANS)) {
415 /*
416 * Skip descriptor addresses which can cause 4KB
417 * boundary crossing (addr + length) with a 32 dword
418 * descriptor fetch.
419 */
420 while (ATH_DESC_4KB_BOUND_CHECK(bf->bf_daddr)) {
421 BUG_ON((caddr_t) bf->bf_desc >=
422 ((caddr_t) dd->dd_desc +
423 dd->dd_desc_len));
424
425 ds += (desc_len * ndesc);
426 bf->bf_desc = ds;
427 bf->bf_daddr = DS2PHYS(dd, ds);
428 }
429 }
430 list_add_tail(&bf->list, head);
431 }
432 } else {
433 struct ath_rxbuf *bf;
406 434
407 for (i = 0; i < nbuf; i++, bf++, ds += (desc_len * ndesc)) { 435 bsize = sizeof(struct ath_rxbuf) * nbuf;
408 bf->bf_desc = ds; 436 bf = devm_kzalloc(sc->dev, bsize, GFP_KERNEL);
409 bf->bf_daddr = DS2PHYS(dd, ds); 437 if (!bf)
410 438 return -ENOMEM;
411 if (!(sc->sc_ah->caps.hw_caps & 439
412 ATH9K_HW_CAP_4KB_SPLITTRANS)) { 440 for (i = 0; i < nbuf; i++, bf++, ds += (desc_len * ndesc)) {
413 /* 441 bf->bf_desc = ds;
414 * Skip descriptor addresses which can cause 4KB 442 bf->bf_daddr = DS2PHYS(dd, ds);
415 * boundary crossing (addr + length) with a 32 dword 443
416 * descriptor fetch. 444 if (!(sc->sc_ah->caps.hw_caps &
417 */ 445 ATH9K_HW_CAP_4KB_SPLITTRANS)) {
418 while (ATH_DESC_4KB_BOUND_CHECK(bf->bf_daddr)) { 446 /*
419 BUG_ON((caddr_t) bf->bf_desc >= 447 * Skip descriptor addresses which can cause 4KB
420 ((caddr_t) dd->dd_desc + 448 * boundary crossing (addr + length) with a 32 dword
421 dd->dd_desc_len)); 449 * descriptor fetch.
422 450 */
423 ds += (desc_len * ndesc); 451 while (ATH_DESC_4KB_BOUND_CHECK(bf->bf_daddr)) {
424 bf->bf_desc = ds; 452 BUG_ON((caddr_t) bf->bf_desc >=
425 bf->bf_daddr = DS2PHYS(dd, ds); 453 ((caddr_t) dd->dd_desc +
454 dd->dd_desc_len));
455
456 ds += (desc_len * ndesc);
457 bf->bf_desc = ds;
458 bf->bf_daddr = DS2PHYS(dd, ds);
459 }
426 } 460 }
461 list_add_tail(&bf->list, head);
427 } 462 }
428 list_add_tail(&bf->list, head);
429 } 463 }
430 return 0; 464 return 0;
431} 465}
@@ -437,7 +471,6 @@ static int ath9k_init_queues(struct ath_softc *sc)
437 sc->beacon.beaconq = ath9k_hw_beaconq_setup(sc->sc_ah); 471 sc->beacon.beaconq = ath9k_hw_beaconq_setup(sc->sc_ah);
438 sc->beacon.cabq = ath_txq_setup(sc, ATH9K_TX_QUEUE_CAB, 0); 472 sc->beacon.cabq = ath_txq_setup(sc, ATH9K_TX_QUEUE_CAB, 0);
439 473
440 sc->config.cabqReadytime = ATH_CABQ_READY_TIME;
441 ath_cabq_update(sc); 474 ath_cabq_update(sc);
442 475
443 sc->tx.uapsdq = ath_txq_setup(sc, ATH9K_TX_QUEUE_UAPSD, 0); 476 sc->tx.uapsdq = ath_txq_setup(sc, ATH9K_TX_QUEUE_UAPSD, 0);
@@ -547,6 +580,26 @@ static void ath9k_init_platform(struct ath_softc *sc)
547 if (sc->driver_data & ATH9K_PCI_CUS217) 580 if (sc->driver_data & ATH9K_PCI_CUS217)
548 ath_info(common, "CUS217 card detected\n"); 581 ath_info(common, "CUS217 card detected\n");
549 582
583 if (sc->driver_data & ATH9K_PCI_CUS252)
584 ath_info(common, "CUS252 card detected\n");
585
586 if (sc->driver_data & ATH9K_PCI_AR9565_1ANT)
587 ath_info(common, "WB335 1-ANT card detected\n");
588
589 if (sc->driver_data & ATH9K_PCI_AR9565_2ANT)
590 ath_info(common, "WB335 2-ANT card detected\n");
591
592 /*
593 * Some WB335 cards do not support antenna diversity. Since
594 * we use a hardcoded value for AR9565 instead of using the
595 * EEPROM/OTP data, remove the combining feature from
596 * the HW capabilities bitmap.
597 */
598 if (sc->driver_data & (ATH9K_PCI_AR9565_1ANT | ATH9K_PCI_AR9565_2ANT)) {
599 if (!(sc->driver_data & ATH9K_PCI_BT_ANT_DIV))
600 pCap->hw_caps &= ~ATH9K_HW_CAP_ANT_DIV_COMB;
601 }
602
550 if (sc->driver_data & ATH9K_PCI_BT_ANT_DIV) { 603 if (sc->driver_data & ATH9K_PCI_BT_ANT_DIV) {
551 pCap->hw_caps |= ATH9K_HW_CAP_BT_ANT_DIV; 604 pCap->hw_caps |= ATH9K_HW_CAP_BT_ANT_DIV;
552 ath_info(common, "Set BT/WLAN RX diversity capability\n"); 605 ath_info(common, "Set BT/WLAN RX diversity capability\n");
@@ -627,7 +680,9 @@ static int ath9k_init_softc(u16 devid, struct ath_softc *sc,
627 sc->sc_ah = ah; 680 sc->sc_ah = ah;
628 pCap = &ah->caps; 681 pCap = &ah->caps;
629 682
630 sc->dfs_detector = dfs_pattern_detector_init(ah, NL80211_DFS_UNSET); 683 common = ath9k_hw_common(ah);
684 sc->dfs_detector = dfs_pattern_detector_init(common, NL80211_DFS_UNSET);
685 sc->tx99_power = MAX_RATE_POWER + 1;
631 686
632 if (!pdata) { 687 if (!pdata) {
633 ah->ah_flags |= AH_USE_EEPROM; 688 ah->ah_flags |= AH_USE_EEPROM;
@@ -641,7 +696,6 @@ static int ath9k_init_softc(u16 devid, struct ath_softc *sc,
641 ah->external_reset = pdata->external_reset; 696 ah->external_reset = pdata->external_reset;
642 } 697 }
643 698
644 common = ath9k_hw_common(ah);
645 common->ops = &ah->reg_ops; 699 common->ops = &ah->reg_ops;
646 common->bus_ops = bus_ops; 700 common->bus_ops = bus_ops;
647 common->ah = ah; 701 common->ah = ah;
@@ -732,6 +786,7 @@ err_queues:
732 ath9k_hw_deinit(ah); 786 ath9k_hw_deinit(ah);
733err_hw: 787err_hw:
734 ath9k_eeprom_release(sc); 788 ath9k_eeprom_release(sc);
789 dev_kfree_skb_any(sc->tx99_skb);
735 return ret; 790 return ret;
736} 791}
737 792
@@ -748,7 +803,7 @@ static void ath9k_init_band_txpower(struct ath_softc *sc, int band)
748 chan = &sband->channels[i]; 803 chan = &sband->channels[i];
749 ah->curchan = &ah->channels[chan->hw_value]; 804 ah->curchan = &ah->channels[chan->hw_value];
750 cfg80211_chandef_create(&chandef, chan, NL80211_CHAN_HT20); 805 cfg80211_chandef_create(&chandef, chan, NL80211_CHAN_HT20);
751 ath9k_cmn_update_ichannel(ah->curchan, &chandef); 806 ath9k_cmn_get_channel(sc->hw, ah, &chandef);
752 ath9k_hw_set_txpowerlimit(ah, MAX_RATE_POWER, true); 807 ath9k_hw_set_txpowerlimit(ah, MAX_RATE_POWER, true);
753 } 808 }
754} 809}
@@ -789,9 +844,9 @@ static const struct ieee80211_iface_limit if_limits[] = {
789 BIT(NL80211_IFTYPE_P2P_GO) }, 844 BIT(NL80211_IFTYPE_P2P_GO) },
790}; 845};
791 846
792
793static const struct ieee80211_iface_limit if_dfs_limits[] = { 847static const struct ieee80211_iface_limit if_dfs_limits[] = {
794 { .max = 1, .types = BIT(NL80211_IFTYPE_AP) }, 848 { .max = 1, .types = BIT(NL80211_IFTYPE_AP) |
849 BIT(NL80211_IFTYPE_ADHOC) },
795}; 850};
796 851
797static const struct ieee80211_iface_combination if_comb[] = { 852static const struct ieee80211_iface_combination if_comb[] = {
@@ -850,17 +905,18 @@ void ath9k_set_hw_capab(struct ath_softc *sc, struct ieee80211_hw *hw)
850 905
851 hw->wiphy->features |= NL80211_FEATURE_ACTIVE_MONITOR; 906 hw->wiphy->features |= NL80211_FEATURE_ACTIVE_MONITOR;
852 907
853 hw->wiphy->interface_modes = 908 if (!config_enabled(CONFIG_ATH9K_TX99)) {
854 BIT(NL80211_IFTYPE_P2P_GO) | 909 hw->wiphy->interface_modes =
855 BIT(NL80211_IFTYPE_P2P_CLIENT) | 910 BIT(NL80211_IFTYPE_P2P_GO) |
856 BIT(NL80211_IFTYPE_AP) | 911 BIT(NL80211_IFTYPE_P2P_CLIENT) |
857 BIT(NL80211_IFTYPE_WDS) | 912 BIT(NL80211_IFTYPE_AP) |
858 BIT(NL80211_IFTYPE_STATION) | 913 BIT(NL80211_IFTYPE_WDS) |
859 BIT(NL80211_IFTYPE_ADHOC) | 914 BIT(NL80211_IFTYPE_STATION) |
860 BIT(NL80211_IFTYPE_MESH_POINT); 915 BIT(NL80211_IFTYPE_ADHOC) |
861 916 BIT(NL80211_IFTYPE_MESH_POINT);
862 hw->wiphy->iface_combinations = if_comb; 917 hw->wiphy->iface_combinations = if_comb;
863 hw->wiphy->n_iface_combinations = ARRAY_SIZE(if_comb); 918 hw->wiphy->n_iface_combinations = ARRAY_SIZE(if_comb);
919 }
864 920
865 hw->wiphy->flags &= ~WIPHY_FLAG_PS_ON_BY_DEFAULT; 921 hw->wiphy->flags &= ~WIPHY_FLAG_PS_ON_BY_DEFAULT;
866 922
diff --git a/drivers/net/wireless/ath/ath9k/link.c b/drivers/net/wireless/ath/ath9k/link.c
index 2f831db396ac..aed7e29dc50f 100644
--- a/drivers/net/wireless/ath/ath9k/link.c
+++ b/drivers/net/wireless/ath/ath9k/link.c
@@ -28,6 +28,13 @@ void ath_tx_complete_poll_work(struct work_struct *work)
28 int i; 28 int i;
29 bool needreset = false; 29 bool needreset = false;
30 30
31
32 if (sc->tx99_state) {
33 ath_dbg(ath9k_hw_common(sc->sc_ah), RESET,
34 "skip tx hung detection on tx99\n");
35 return;
36 }
37
31 for (i = 0; i < IEEE80211_NUM_ACS; i++) { 38 for (i = 0; i < IEEE80211_NUM_ACS; i++) {
32 txq = sc->tx.txq_map[i]; 39 txq = sc->tx.txq_map[i];
33 40
@@ -70,7 +77,7 @@ void ath_hw_check(struct work_struct *work)
70 ath9k_ps_wakeup(sc); 77 ath9k_ps_wakeup(sc);
71 is_alive = ath9k_hw_check_alive(sc->sc_ah); 78 is_alive = ath9k_hw_check_alive(sc->sc_ah);
72 79
73 if (is_alive && !AR_SREV_9300(sc->sc_ah)) 80 if ((is_alive && !AR_SREV_9300(sc->sc_ah)) || sc->tx99_state)
74 goto out; 81 goto out;
75 else if (!is_alive && AR_SREV_9300(sc->sc_ah)) { 82 else if (!is_alive && AR_SREV_9300(sc->sc_ah)) {
76 ath_dbg(common, RESET, 83 ath_dbg(common, RESET,
@@ -141,6 +148,9 @@ void ath_hw_pll_work(struct work_struct *work)
141 if (!test_bit(SC_OP_BEACONS, &sc->sc_flags)) 148 if (!test_bit(SC_OP_BEACONS, &sc->sc_flags))
142 return; 149 return;
143 150
151 if (sc->tx99_state)
152 return;
153
144 ath9k_ps_wakeup(sc); 154 ath9k_ps_wakeup(sc);
145 pll_sqsum = ar9003_get_pll_sqsum_dvc(sc->sc_ah); 155 pll_sqsum = ar9003_get_pll_sqsum_dvc(sc->sc_ah);
146 ath9k_ps_restore(sc); 156 ath9k_ps_restore(sc);
@@ -184,7 +194,7 @@ static void ath_paprd_activate(struct ath_softc *sc)
184 struct ath9k_hw_cal_data *caldata = ah->caldata; 194 struct ath9k_hw_cal_data *caldata = ah->caldata;
185 int chain; 195 int chain;
186 196
187 if (!caldata || !caldata->paprd_done) { 197 if (!caldata || !test_bit(PAPRD_DONE, &caldata->cal_flags)) {
188 ath_dbg(common, CALIBRATE, "Failed to activate PAPRD\n"); 198 ath_dbg(common, CALIBRATE, "Failed to activate PAPRD\n");
189 return; 199 return;
190 } 200 }
@@ -256,7 +266,9 @@ void ath_paprd_calibrate(struct work_struct *work)
256 int len = 1800; 266 int len = 1800;
257 int ret; 267 int ret;
258 268
259 if (!caldata || !caldata->paprd_packet_sent || caldata->paprd_done) { 269 if (!caldata ||
270 !test_bit(PAPRD_PACKET_SENT, &caldata->cal_flags) ||
271 test_bit(PAPRD_DONE, &caldata->cal_flags)) {
260 ath_dbg(common, CALIBRATE, "Skipping PAPRD calibration\n"); 272 ath_dbg(common, CALIBRATE, "Skipping PAPRD calibration\n");
261 return; 273 return;
262 } 274 }
@@ -316,7 +328,7 @@ void ath_paprd_calibrate(struct work_struct *work)
316 kfree_skb(skb); 328 kfree_skb(skb);
317 329
318 if (chain_ok) { 330 if (chain_ok) {
319 caldata->paprd_done = true; 331 set_bit(PAPRD_DONE, &caldata->cal_flags);
320 ath_paprd_activate(sc); 332 ath_paprd_activate(sc);
321 } 333 }
322 334
@@ -343,7 +355,7 @@ void ath_ani_calibrate(unsigned long data)
343 u32 cal_interval, short_cal_interval, long_cal_interval; 355 u32 cal_interval, short_cal_interval, long_cal_interval;
344 unsigned long flags; 356 unsigned long flags;
345 357
346 if (ah->caldata && ah->caldata->nfcal_interference) 358 if (ah->caldata && test_bit(NFCAL_INTF, &ah->caldata->cal_flags))
347 long_cal_interval = ATH_LONG_CALINTERVAL_INT; 359 long_cal_interval = ATH_LONG_CALINTERVAL_INT;
348 else 360 else
349 long_cal_interval = ATH_LONG_CALINTERVAL; 361 long_cal_interval = ATH_LONG_CALINTERVAL;
@@ -432,7 +444,7 @@ set_timer:
432 mod_timer(&common->ani.timer, jiffies + msecs_to_jiffies(cal_interval)); 444 mod_timer(&common->ani.timer, jiffies + msecs_to_jiffies(cal_interval));
433 445
434 if (ar9003_is_paprd_enabled(ah) && ah->caldata) { 446 if (ar9003_is_paprd_enabled(ah) && ah->caldata) {
435 if (!ah->caldata->paprd_done) { 447 if (!test_bit(PAPRD_DONE, &ah->caldata->cal_flags)) {
436 ieee80211_queue_work(sc->hw, &sc->paprd_work); 448 ieee80211_queue_work(sc->hw, &sc->paprd_work);
437 } else if (!ah->paprd_table_write_done) { 449 } else if (!ah->paprd_table_write_done) {
438 ath9k_ps_wakeup(sc); 450 ath9k_ps_wakeup(sc);
@@ -516,7 +528,8 @@ void ath_update_survey_nf(struct ath_softc *sc, int channel)
516 528
517 if (chan->noisefloor) { 529 if (chan->noisefloor) {
518 survey->filled |= SURVEY_INFO_NOISE_DBM; 530 survey->filled |= SURVEY_INFO_NOISE_DBM;
519 survey->noise = ath9k_hw_getchan_noise(ah, chan); 531 survey->noise = ath9k_hw_getchan_noise(ah, chan,
532 chan->noisefloor);
520 } 533 }
521} 534}
522 535
diff --git a/drivers/net/wireless/ath/ath9k/mac.c b/drivers/net/wireless/ath/ath9k/mac.c
index a3eff0986a3f..6a18f9d3e9cc 100644
--- a/drivers/net/wireless/ath/ath9k/mac.c
+++ b/drivers/net/wireless/ath/ath9k/mac.c
@@ -374,7 +374,6 @@ EXPORT_SYMBOL(ath9k_hw_releasetxqueue);
374bool ath9k_hw_resettxqueue(struct ath_hw *ah, u32 q) 374bool ath9k_hw_resettxqueue(struct ath_hw *ah, u32 q)
375{ 375{
376 struct ath_common *common = ath9k_hw_common(ah); 376 struct ath_common *common = ath9k_hw_common(ah);
377 struct ath9k_channel *chan = ah->curchan;
378 struct ath9k_tx_queue_info *qi; 377 struct ath9k_tx_queue_info *qi;
379 u32 cwMin, chanCwMin, value; 378 u32 cwMin, chanCwMin, value;
380 379
@@ -387,10 +386,7 @@ bool ath9k_hw_resettxqueue(struct ath_hw *ah, u32 q)
387 ath_dbg(common, QUEUE, "Reset TX queue: %u\n", q); 386 ath_dbg(common, QUEUE, "Reset TX queue: %u\n", q);
388 387
389 if (qi->tqi_cwmin == ATH9K_TXQ_USEDEFAULT) { 388 if (qi->tqi_cwmin == ATH9K_TXQ_USEDEFAULT) {
390 if (chan && IS_CHAN_B(chan)) 389 chanCwMin = INIT_CWMIN;
391 chanCwMin = INIT_CWMIN_11B;
392 else
393 chanCwMin = INIT_CWMIN;
394 390
395 for (cwMin = 1; cwMin < chanCwMin; cwMin = (cwMin << 1) | 1); 391 for (cwMin = 1; cwMin < chanCwMin; cwMin = (cwMin << 1) | 1);
396 } else 392 } else
diff --git a/drivers/net/wireless/ath/ath9k/mac.h b/drivers/net/wireless/ath/ath9k/mac.h
index bfccaceed44e..e3eed81f2439 100644
--- a/drivers/net/wireless/ath/ath9k/mac.h
+++ b/drivers/net/wireless/ath/ath9k/mac.h
@@ -603,8 +603,6 @@ enum ath9k_tx_queue_flags {
603#define ATH9K_TXQ_USE_LOCKOUT_BKOFF_DIS 0x00000001 603#define ATH9K_TXQ_USE_LOCKOUT_BKOFF_DIS 0x00000001
604 604
605#define ATH9K_DECOMP_MASK_SIZE 128 605#define ATH9K_DECOMP_MASK_SIZE 128
606#define ATH9K_READY_TIME_LO_BOUND 50
607#define ATH9K_READY_TIME_HI_BOUND 96
608 606
609enum ath9k_pkt_type { 607enum ath9k_pkt_type {
610 ATH9K_PKT_TYPE_NORMAL = 0, 608 ATH9K_PKT_TYPE_NORMAL = 0,
diff --git a/drivers/net/wireless/ath/ath9k/main.c b/drivers/net/wireless/ath/ath9k/main.c
index 709301f88dcd..74f452c7b166 100644
--- a/drivers/net/wireless/ath/ath9k/main.c
+++ b/drivers/net/wireless/ath/ath9k/main.c
@@ -312,17 +312,91 @@ out:
312 * by reseting the chip. To accomplish this we must first cleanup any pending 312 * by reseting the chip. To accomplish this we must first cleanup any pending
313 * DMA, then restart stuff. 313 * DMA, then restart stuff.
314*/ 314*/
315static int ath_set_channel(struct ath_softc *sc, struct ieee80211_hw *hw, 315static int ath_set_channel(struct ath_softc *sc, struct cfg80211_chan_def *chandef)
316 struct ath9k_channel *hchan)
317{ 316{
317 struct ath_hw *ah = sc->sc_ah;
318 struct ath_common *common = ath9k_hw_common(ah);
319 struct ieee80211_hw *hw = sc->hw;
320 struct ath9k_channel *hchan;
321 struct ieee80211_channel *chan = chandef->chan;
322 unsigned long flags;
323 bool offchannel;
324 int pos = chan->hw_value;
325 int old_pos = -1;
318 int r; 326 int r;
319 327
320 if (test_bit(SC_OP_INVALID, &sc->sc_flags)) 328 if (test_bit(SC_OP_INVALID, &sc->sc_flags))
321 return -EIO; 329 return -EIO;
322 330
331 offchannel = !!(hw->conf.flags & IEEE80211_CONF_OFFCHANNEL);
332
333 if (ah->curchan)
334 old_pos = ah->curchan - &ah->channels[0];
335
336 ath_dbg(common, CONFIG, "Set channel: %d MHz width: %d\n",
337 chan->center_freq, chandef->width);
338
339 /* update survey stats for the old channel before switching */
340 spin_lock_irqsave(&common->cc_lock, flags);
341 ath_update_survey_stats(sc);
342 spin_unlock_irqrestore(&common->cc_lock, flags);
343
344 ath9k_cmn_get_channel(hw, ah, chandef);
345
346 /*
347 * If the operating channel changes, change the survey in-use flags
348 * along with it.
349 * Reset the survey data for the new channel, unless we're switching
350 * back to the operating channel from an off-channel operation.
351 */
352 if (!offchannel && sc->cur_survey != &sc->survey[pos]) {
353 if (sc->cur_survey)
354 sc->cur_survey->filled &= ~SURVEY_INFO_IN_USE;
355
356 sc->cur_survey = &sc->survey[pos];
357
358 memset(sc->cur_survey, 0, sizeof(struct survey_info));
359 sc->cur_survey->filled |= SURVEY_INFO_IN_USE;
360 } else if (!(sc->survey[pos].filled & SURVEY_INFO_IN_USE)) {
361 memset(&sc->survey[pos], 0, sizeof(struct survey_info));
362 }
363
364 hchan = &sc->sc_ah->channels[pos];
323 r = ath_reset_internal(sc, hchan); 365 r = ath_reset_internal(sc, hchan);
366 if (r)
367 return r;
324 368
325 return r; 369 /*
370 * The most recent snapshot of channel->noisefloor for the old
371 * channel is only available after the hardware reset. Copy it to
372 * the survey stats now.
373 */
374 if (old_pos >= 0)
375 ath_update_survey_nf(sc, old_pos);
376
377 /*
378 * Enable radar pulse detection if on a DFS channel. Spectral
379 * scanning and radar detection can not be used concurrently.
380 */
381 if (hw->conf.radar_enabled) {
382 u32 rxfilter;
383
384 /* set HW specific DFS configuration */
385 ath9k_hw_set_radar_params(ah);
386 rxfilter = ath9k_hw_getrxfilter(ah);
387 rxfilter |= ATH9K_RX_FILTER_PHYRADAR |
388 ATH9K_RX_FILTER_PHYERR;
389 ath9k_hw_setrxfilter(ah, rxfilter);
390 ath_dbg(common, DFS, "DFS enabled at freq %d\n",
391 chan->center_freq);
392 } else {
393 /* perform spectral scan if requested. */
394 if (test_bit(SC_OP_SCANNING, &sc->sc_flags) &&
395 sc->spectral_mode == SPECTRAL_CHANSCAN)
396 ath9k_spectral_scan_trigger(hw);
397 }
398
399 return 0;
326} 400}
327 401
328static void ath_node_attach(struct ath_softc *sc, struct ieee80211_sta *sta, 402static void ath_node_attach(struct ath_softc *sc, struct ieee80211_sta *sta,
@@ -372,6 +446,13 @@ void ath9k_tasklet(unsigned long data)
372 type = RESET_TYPE_BB_WATCHDOG; 446 type = RESET_TYPE_BB_WATCHDOG;
373 447
374 ath9k_queue_reset(sc, type); 448 ath9k_queue_reset(sc, type);
449
450 /*
451 * Increment the ref. counter here so that
452 * interrupts are enabled in the reset routine.
453 */
454 atomic_inc(&ah->intr_ref_cnt);
455 ath_dbg(common, ANY, "FATAL: Skipping interrupts\n");
375 goto out; 456 goto out;
376 } 457 }
377 458
@@ -410,10 +491,9 @@ void ath9k_tasklet(unsigned long data)
410 491
411 ath9k_btcoex_handle_interrupt(sc, status); 492 ath9k_btcoex_handle_interrupt(sc, status);
412 493
413out:
414 /* re-enable hardware interrupt */ 494 /* re-enable hardware interrupt */
415 ath9k_hw_enable_interrupts(ah); 495 ath9k_hw_enable_interrupts(ah);
416 496out:
417 spin_unlock(&sc->sc_pcu_lock); 497 spin_unlock(&sc->sc_pcu_lock);
418 ath9k_ps_restore(sc); 498 ath9k_ps_restore(sc);
419} 499}
@@ -594,7 +674,7 @@ static int ath9k_start(struct ieee80211_hw *hw)
594 ath9k_ps_wakeup(sc); 674 ath9k_ps_wakeup(sc);
595 mutex_lock(&sc->mutex); 675 mutex_lock(&sc->mutex);
596 676
597 init_channel = ath9k_cmn_get_curchannel(hw, ah); 677 init_channel = ath9k_cmn_get_channel(hw, ah, &hw->conf.chandef);
598 678
599 /* Reset SERDES registers */ 679 /* Reset SERDES registers */
600 ath9k_hw_configpcipowersave(ah, false); 680 ath9k_hw_configpcipowersave(ah, false);
@@ -797,7 +877,7 @@ static void ath9k_stop(struct ieee80211_hw *hw)
797 } 877 }
798 878
799 if (!ah->curchan) 879 if (!ah->curchan)
800 ah->curchan = ath9k_cmn_get_curchannel(hw, ah); 880 ah->curchan = ath9k_cmn_get_channel(hw, ah, &hw->conf.chandef);
801 881
802 ath9k_hw_reset(ah, ah->curchan, ah->caldata, false); 882 ath9k_hw_reset(ah, ah->curchan, ah->caldata, false);
803 ath9k_hw_phy_disable(ah); 883 ath9k_hw_phy_disable(ah);
@@ -816,7 +896,7 @@ static void ath9k_stop(struct ieee80211_hw *hw)
816 ath_dbg(common, CONFIG, "Driver halt\n"); 896 ath_dbg(common, CONFIG, "Driver halt\n");
817} 897}
818 898
819bool ath9k_uses_beacons(int type) 899static bool ath9k_uses_beacons(int type)
820{ 900{
821 switch (type) { 901 switch (type) {
822 case NL80211_IFTYPE_AP: 902 case NL80211_IFTYPE_AP:
@@ -966,6 +1046,14 @@ static int ath9k_add_interface(struct ieee80211_hw *hw,
966 1046
967 mutex_lock(&sc->mutex); 1047 mutex_lock(&sc->mutex);
968 1048
1049 if (config_enabled(CONFIG_ATH9K_TX99)) {
1050 if (sc->nvifs >= 1) {
1051 mutex_unlock(&sc->mutex);
1052 return -EOPNOTSUPP;
1053 }
1054 sc->tx99_vif = vif;
1055 }
1056
969 ath_dbg(common, CONFIG, "Attach a VIF of type: %d\n", vif->type); 1057 ath_dbg(common, CONFIG, "Attach a VIF of type: %d\n", vif->type);
970 sc->nvifs++; 1058 sc->nvifs++;
971 1059
@@ -994,9 +1082,15 @@ static int ath9k_change_interface(struct ieee80211_hw *hw,
994 struct ath_softc *sc = hw->priv; 1082 struct ath_softc *sc = hw->priv;
995 struct ath_common *common = ath9k_hw_common(sc->sc_ah); 1083 struct ath_common *common = ath9k_hw_common(sc->sc_ah);
996 1084
997 ath_dbg(common, CONFIG, "Change Interface\n");
998 mutex_lock(&sc->mutex); 1085 mutex_lock(&sc->mutex);
999 1086
1087 if (config_enabled(CONFIG_ATH9K_TX99)) {
1088 mutex_unlock(&sc->mutex);
1089 return -EOPNOTSUPP;
1090 }
1091
1092 ath_dbg(common, CONFIG, "Change Interface\n");
1093
1000 if (ath9k_uses_beacons(vif->type)) 1094 if (ath9k_uses_beacons(vif->type))
1001 ath9k_beacon_remove_slot(sc, vif); 1095 ath9k_beacon_remove_slot(sc, vif);
1002 1096
@@ -1026,6 +1120,7 @@ static void ath9k_remove_interface(struct ieee80211_hw *hw,
1026 mutex_lock(&sc->mutex); 1120 mutex_lock(&sc->mutex);
1027 1121
1028 sc->nvifs--; 1122 sc->nvifs--;
1123 sc->tx99_vif = NULL;
1029 1124
1030 if (ath9k_uses_beacons(vif->type)) 1125 if (ath9k_uses_beacons(vif->type))
1031 ath9k_beacon_remove_slot(sc, vif); 1126 ath9k_beacon_remove_slot(sc, vif);
@@ -1047,6 +1142,9 @@ static void ath9k_enable_ps(struct ath_softc *sc)
1047 struct ath_hw *ah = sc->sc_ah; 1142 struct ath_hw *ah = sc->sc_ah;
1048 struct ath_common *common = ath9k_hw_common(ah); 1143 struct ath_common *common = ath9k_hw_common(ah);
1049 1144
1145 if (config_enabled(CONFIG_ATH9K_TX99))
1146 return;
1147
1050 sc->ps_enabled = true; 1148 sc->ps_enabled = true;
1051 if (!(ah->caps.hw_caps & ATH9K_HW_CAP_AUTOSLEEP)) { 1149 if (!(ah->caps.hw_caps & ATH9K_HW_CAP_AUTOSLEEP)) {
1052 if ((ah->imask & ATH9K_INT_TIM_TIMER) == 0) { 1150 if ((ah->imask & ATH9K_INT_TIM_TIMER) == 0) {
@@ -1063,6 +1161,9 @@ static void ath9k_disable_ps(struct ath_softc *sc)
1063 struct ath_hw *ah = sc->sc_ah; 1161 struct ath_hw *ah = sc->sc_ah;
1064 struct ath_common *common = ath9k_hw_common(ah); 1162 struct ath_common *common = ath9k_hw_common(ah);
1065 1163
1164 if (config_enabled(CONFIG_ATH9K_TX99))
1165 return;
1166
1066 sc->ps_enabled = false; 1167 sc->ps_enabled = false;
1067 ath9k_hw_setpower(ah, ATH9K_PM_AWAKE); 1168 ath9k_hw_setpower(ah, ATH9K_PM_AWAKE);
1068 if (!(ah->caps.hw_caps & ATH9K_HW_CAP_AUTOSLEEP)) { 1169 if (!(ah->caps.hw_caps & ATH9K_HW_CAP_AUTOSLEEP)) {
@@ -1086,6 +1187,9 @@ void ath9k_spectral_scan_trigger(struct ieee80211_hw *hw)
1086 struct ath_common *common = ath9k_hw_common(ah); 1187 struct ath_common *common = ath9k_hw_common(ah);
1087 u32 rxfilter; 1188 u32 rxfilter;
1088 1189
1190 if (config_enabled(CONFIG_ATH9K_TX99))
1191 return;
1192
1089 if (!ath9k_hw_ops(ah)->spectral_scan_trigger) { 1193 if (!ath9k_hw_ops(ah)->spectral_scan_trigger) {
1090 ath_err(common, "spectrum analyzer not implemented on this hardware\n"); 1194 ath_err(common, "spectrum analyzer not implemented on this hardware\n");
1091 return; 1195 return;
@@ -1201,81 +1305,12 @@ static int ath9k_config(struct ieee80211_hw *hw, u32 changed)
1201 } 1305 }
1202 1306
1203 if ((changed & IEEE80211_CONF_CHANGE_CHANNEL) || reset_channel) { 1307 if ((changed & IEEE80211_CONF_CHANGE_CHANNEL) || reset_channel) {
1204 struct ieee80211_channel *curchan = hw->conf.chandef.chan; 1308 if (ath_set_channel(sc, &hw->conf.chandef) < 0) {
1205 int pos = curchan->hw_value;
1206 int old_pos = -1;
1207 unsigned long flags;
1208
1209 if (ah->curchan)
1210 old_pos = ah->curchan - &ah->channels[0];
1211
1212 ath_dbg(common, CONFIG, "Set channel: %d MHz width: %d\n",
1213 curchan->center_freq, hw->conf.chandef.width);
1214
1215 /* update survey stats for the old channel before switching */
1216 spin_lock_irqsave(&common->cc_lock, flags);
1217 ath_update_survey_stats(sc);
1218 spin_unlock_irqrestore(&common->cc_lock, flags);
1219
1220 ath9k_cmn_update_ichannel(&sc->sc_ah->channels[pos],
1221 &conf->chandef);
1222
1223 /*
1224 * If the operating channel changes, change the survey in-use flags
1225 * along with it.
1226 * Reset the survey data for the new channel, unless we're switching
1227 * back to the operating channel from an off-channel operation.
1228 */
1229 if (!(hw->conf.flags & IEEE80211_CONF_OFFCHANNEL) &&
1230 sc->cur_survey != &sc->survey[pos]) {
1231
1232 if (sc->cur_survey)
1233 sc->cur_survey->filled &= ~SURVEY_INFO_IN_USE;
1234
1235 sc->cur_survey = &sc->survey[pos];
1236
1237 memset(sc->cur_survey, 0, sizeof(struct survey_info));
1238 sc->cur_survey->filled |= SURVEY_INFO_IN_USE;
1239 } else if (!(sc->survey[pos].filled & SURVEY_INFO_IN_USE)) {
1240 memset(&sc->survey[pos], 0, sizeof(struct survey_info));
1241 }
1242
1243 if (ath_set_channel(sc, hw, &sc->sc_ah->channels[pos]) < 0) {
1244 ath_err(common, "Unable to set channel\n"); 1309 ath_err(common, "Unable to set channel\n");
1245 mutex_unlock(&sc->mutex); 1310 mutex_unlock(&sc->mutex);
1246 ath9k_ps_restore(sc); 1311 ath9k_ps_restore(sc);
1247 return -EINVAL; 1312 return -EINVAL;
1248 } 1313 }
1249
1250 /*
1251 * The most recent snapshot of channel->noisefloor for the old
1252 * channel is only available after the hardware reset. Copy it to
1253 * the survey stats now.
1254 */
1255 if (old_pos >= 0)
1256 ath_update_survey_nf(sc, old_pos);
1257
1258 /*
1259 * Enable radar pulse detection if on a DFS channel. Spectral
1260 * scanning and radar detection can not be used concurrently.
1261 */
1262 if (hw->conf.radar_enabled) {
1263 u32 rxfilter;
1264
1265 /* set HW specific DFS configuration */
1266 ath9k_hw_set_radar_params(ah);
1267 rxfilter = ath9k_hw_getrxfilter(ah);
1268 rxfilter |= ATH9K_RX_FILTER_PHYRADAR |
1269 ATH9K_RX_FILTER_PHYERR;
1270 ath9k_hw_setrxfilter(ah, rxfilter);
1271 ath_dbg(common, DFS, "DFS enabled at freq %d\n",
1272 curchan->center_freq);
1273 } else {
1274 /* perform spectral scan if requested. */
1275 if (test_bit(SC_OP_SCANNING, &sc->sc_flags) &&
1276 sc->spectral_mode == SPECTRAL_CHANSCAN)
1277 ath9k_spectral_scan_trigger(hw);
1278 }
1279 } 1314 }
1280 1315
1281 if (changed & IEEE80211_CONF_CHANGE_POWER) { 1316 if (changed & IEEE80211_CONF_CHANGE_POWER) {
@@ -1734,6 +1769,9 @@ static int ath9k_get_survey(struct ieee80211_hw *hw, int idx,
1734 unsigned long flags; 1769 unsigned long flags;
1735 int pos; 1770 int pos;
1736 1771
1772 if (config_enabled(CONFIG_ATH9K_TX99))
1773 return -EOPNOTSUPP;
1774
1737 spin_lock_irqsave(&common->cc_lock, flags); 1775 spin_lock_irqsave(&common->cc_lock, flags);
1738 if (idx == 0) 1776 if (idx == 0)
1739 ath_update_survey_stats(sc); 1777 ath_update_survey_stats(sc);
@@ -1766,6 +1804,9 @@ static void ath9k_set_coverage_class(struct ieee80211_hw *hw, u8 coverage_class)
1766 struct ath_softc *sc = hw->priv; 1804 struct ath_softc *sc = hw->priv;
1767 struct ath_hw *ah = sc->sc_ah; 1805 struct ath_hw *ah = sc->sc_ah;
1768 1806
1807 if (config_enabled(CONFIG_ATH9K_TX99))
1808 return;
1809
1769 mutex_lock(&sc->mutex); 1810 mutex_lock(&sc->mutex);
1770 ah->coverage_class = coverage_class; 1811 ah->coverage_class = coverage_class;
1771 1812
@@ -2332,6 +2373,134 @@ static void ath9k_channel_switch_beacon(struct ieee80211_hw *hw,
2332 sc->csa_vif = vif; 2373 sc->csa_vif = vif;
2333} 2374}
2334 2375
2376static void ath9k_tx99_stop(struct ath_softc *sc)
2377{
2378 struct ath_hw *ah = sc->sc_ah;
2379 struct ath_common *common = ath9k_hw_common(ah);
2380
2381 ath_drain_all_txq(sc);
2382 ath_startrecv(sc);
2383
2384 ath9k_hw_set_interrupts(ah);
2385 ath9k_hw_enable_interrupts(ah);
2386
2387 ieee80211_wake_queues(sc->hw);
2388
2389 kfree_skb(sc->tx99_skb);
2390 sc->tx99_skb = NULL;
2391 sc->tx99_state = false;
2392
2393 ath9k_hw_tx99_stop(sc->sc_ah);
2394 ath_dbg(common, XMIT, "TX99 stopped\n");
2395}
2396
2397static struct sk_buff *ath9k_build_tx99_skb(struct ath_softc *sc)
2398{
2399 static u8 PN9Data[] = {0xff, 0x87, 0xb8, 0x59, 0xb7, 0xa1, 0xcc, 0x24,
2400 0x57, 0x5e, 0x4b, 0x9c, 0x0e, 0xe9, 0xea, 0x50,
2401 0x2a, 0xbe, 0xb4, 0x1b, 0xb6, 0xb0, 0x5d, 0xf1,
2402 0xe6, 0x9a, 0xe3, 0x45, 0xfd, 0x2c, 0x53, 0x18,
2403 0x0c, 0xca, 0xc9, 0xfb, 0x49, 0x37, 0xe5, 0xa8,
2404 0x51, 0x3b, 0x2f, 0x61, 0xaa, 0x72, 0x18, 0x84,
2405 0x02, 0x23, 0x23, 0xab, 0x63, 0x89, 0x51, 0xb3,
2406 0xe7, 0x8b, 0x72, 0x90, 0x4c, 0xe8, 0xfb, 0xc0};
2407 u32 len = 1200;
2408 struct ieee80211_hw *hw = sc->hw;
2409 struct ieee80211_hdr *hdr;
2410 struct ieee80211_tx_info *tx_info;
2411 struct sk_buff *skb;
2412
2413 skb = alloc_skb(len, GFP_KERNEL);
2414 if (!skb)
2415 return NULL;
2416
2417 skb_put(skb, len);
2418
2419 memset(skb->data, 0, len);
2420
2421 hdr = (struct ieee80211_hdr *)skb->data;
2422 hdr->frame_control = cpu_to_le16(IEEE80211_FTYPE_DATA);
2423 hdr->duration_id = 0;
2424
2425 memcpy(hdr->addr1, hw->wiphy->perm_addr, ETH_ALEN);
2426 memcpy(hdr->addr2, hw->wiphy->perm_addr, ETH_ALEN);
2427 memcpy(hdr->addr3, hw->wiphy->perm_addr, ETH_ALEN);
2428
2429 hdr->seq_ctrl |= cpu_to_le16(sc->tx.seq_no);
2430
2431 tx_info = IEEE80211_SKB_CB(skb);
2432 memset(tx_info, 0, sizeof(*tx_info));
2433 tx_info->band = hw->conf.chandef.chan->band;
2434 tx_info->flags = IEEE80211_TX_CTL_NO_ACK;
2435 tx_info->control.vif = sc->tx99_vif;
2436
2437 memcpy(skb->data + sizeof(*hdr), PN9Data, sizeof(PN9Data));
2438
2439 return skb;
2440}
2441
2442void ath9k_tx99_deinit(struct ath_softc *sc)
2443{
2444 ath_reset(sc);
2445
2446 ath9k_ps_wakeup(sc);
2447 ath9k_tx99_stop(sc);
2448 ath9k_ps_restore(sc);
2449}
2450
2451int ath9k_tx99_init(struct ath_softc *sc)
2452{
2453 struct ieee80211_hw *hw = sc->hw;
2454 struct ath_hw *ah = sc->sc_ah;
2455 struct ath_common *common = ath9k_hw_common(ah);
2456 struct ath_tx_control txctl;
2457 int r;
2458
2459 if (sc->sc_flags & SC_OP_INVALID) {
2460 ath_err(common,
2461 "driver is in invalid state unable to use TX99");
2462 return -EINVAL;
2463 }
2464
2465 sc->tx99_skb = ath9k_build_tx99_skb(sc);
2466 if (!sc->tx99_skb)
2467 return -ENOMEM;
2468
2469 memset(&txctl, 0, sizeof(txctl));
2470 txctl.txq = sc->tx.txq_map[IEEE80211_AC_VO];
2471
2472 ath_reset(sc);
2473
2474 ath9k_ps_wakeup(sc);
2475
2476 ath9k_hw_disable_interrupts(ah);
2477 atomic_set(&ah->intr_ref_cnt, -1);
2478 ath_drain_all_txq(sc);
2479 ath_stoprecv(sc);
2480
2481 sc->tx99_state = true;
2482
2483 ieee80211_stop_queues(hw);
2484
2485 if (sc->tx99_power == MAX_RATE_POWER + 1)
2486 sc->tx99_power = MAX_RATE_POWER;
2487
2488 ath9k_hw_tx99_set_txpower(ah, sc->tx99_power);
2489 r = ath9k_tx99_send(sc, sc->tx99_skb, &txctl);
2490 if (r) {
2491 ath_dbg(common, XMIT, "Failed to xmit TX99 skb\n");
2492 return r;
2493 }
2494
2495 ath_dbg(common, XMIT, "TX99 xmit started using %d ( %ddBm)\n",
2496 sc->tx99_power,
2497 sc->tx99_power / 2);
2498
2499 /* We leave the harware awake as it will be chugging on */
2500
2501 return 0;
2502}
2503
2335struct ieee80211_ops ath9k_ops = { 2504struct ieee80211_ops ath9k_ops = {
2336 .tx = ath9k_tx, 2505 .tx = ath9k_tx,
2337 .start = ath9k_start, 2506 .start = ath9k_start,
diff --git a/drivers/net/wireless/ath/ath9k/mci.c b/drivers/net/wireless/ath/ath9k/mci.c
index 815bee21c19a..0ac1b5f04256 100644
--- a/drivers/net/wireless/ath/ath9k/mci.c
+++ b/drivers/net/wireless/ath/ath9k/mci.c
@@ -661,9 +661,9 @@ void ath9k_mci_update_wlan_channels(struct ath_softc *sc, bool allow_all)
661 chan_start = wlan_chan - 10; 661 chan_start = wlan_chan - 10;
662 chan_end = wlan_chan + 10; 662 chan_end = wlan_chan + 10;
663 663
664 if (chan->chanmode == CHANNEL_G_HT40PLUS) 664 if (IS_CHAN_HT40PLUS(chan))
665 chan_end += 20; 665 chan_end += 20;
666 else if (chan->chanmode == CHANNEL_G_HT40MINUS) 666 else if (IS_CHAN_HT40MINUS(chan))
667 chan_start -= 20; 667 chan_start -= 20;
668 668
669 /* adjust side band */ 669 /* adjust side band */
@@ -707,11 +707,11 @@ void ath9k_mci_set_txpower(struct ath_softc *sc, bool setchannel,
707 707
708 if (setchannel) { 708 if (setchannel) {
709 struct ath9k_hw_cal_data *caldata = &sc->caldata; 709 struct ath9k_hw_cal_data *caldata = &sc->caldata;
710 if ((caldata->chanmode == CHANNEL_G_HT40PLUS) && 710 if (IS_CHAN_HT40PLUS(ah->curchan) &&
711 (ah->curchan->channel > caldata->channel) && 711 (ah->curchan->channel > caldata->channel) &&
712 (ah->curchan->channel <= caldata->channel + 20)) 712 (ah->curchan->channel <= caldata->channel + 20))
713 return; 713 return;
714 if ((caldata->chanmode == CHANNEL_G_HT40MINUS) && 714 if (IS_CHAN_HT40MINUS(ah->curchan) &&
715 (ah->curchan->channel < caldata->channel) && 715 (ah->curchan->channel < caldata->channel) &&
716 (ah->curchan->channel >= caldata->channel - 20)) 716 (ah->curchan->channel >= caldata->channel - 20))
717 return; 717 return;
diff --git a/drivers/net/wireless/ath/ath9k/pci.c b/drivers/net/wireless/ath/ath9k/pci.c
index d089a7cf01c4..7e4c2524b630 100644
--- a/drivers/net/wireless/ath/ath9k/pci.c
+++ b/drivers/net/wireless/ath/ath9k/pci.c
@@ -269,7 +269,200 @@ static DEFINE_PCI_DEVICE_TABLE(ath_pci_id_table) = {
269 269
270 { PCI_VDEVICE(ATHEROS, 0x0034) }, /* PCI-E AR9462 */ 270 { PCI_VDEVICE(ATHEROS, 0x0034) }, /* PCI-E AR9462 */
271 { PCI_VDEVICE(ATHEROS, 0x0037) }, /* PCI-E AR1111/AR9485 */ 271 { PCI_VDEVICE(ATHEROS, 0x0037) }, /* PCI-E AR1111/AR9485 */
272 { PCI_VDEVICE(ATHEROS, 0x0036) }, /* PCI-E AR9565 */ 272
273 /* CUS252 */
274 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
275 0x0036,
276 PCI_VENDOR_ID_ATHEROS,
277 0x3028),
278 .driver_data = ATH9K_PCI_CUS252 |
279 ATH9K_PCI_AR9565_2ANT |
280 ATH9K_PCI_BT_ANT_DIV },
281 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
282 0x0036,
283 PCI_VENDOR_ID_AZWAVE,
284 0x2176),
285 .driver_data = ATH9K_PCI_CUS252 |
286 ATH9K_PCI_AR9565_2ANT |
287 ATH9K_PCI_BT_ANT_DIV },
288
289 /* WB335 1-ANT */
290 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
291 0x0036,
292 PCI_VENDOR_ID_FOXCONN,
293 0xE068),
294 .driver_data = ATH9K_PCI_AR9565_1ANT },
295 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
296 0x0036,
297 0x185F, /* WNC */
298 0xA119),
299 .driver_data = ATH9K_PCI_AR9565_1ANT },
300 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
301 0x0036,
302 0x11AD, /* LITEON */
303 0x0632),
304 .driver_data = ATH9K_PCI_AR9565_1ANT },
305 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
306 0x0036,
307 0x11AD, /* LITEON */
308 0x6671),
309 .driver_data = ATH9K_PCI_AR9565_1ANT },
310 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
311 0x0036,
312 0x1B9A, /* XAVI */
313 0x2811),
314 .driver_data = ATH9K_PCI_AR9565_1ANT },
315 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
316 0x0036,
317 0x1B9A, /* XAVI */
318 0x2812),
319 .driver_data = ATH9K_PCI_AR9565_1ANT },
320
321 /* WB335 1-ANT / Antenna Diversity */
322 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
323 0x0036,
324 PCI_VENDOR_ID_ATHEROS,
325 0x3025),
326 .driver_data = ATH9K_PCI_AR9565_1ANT | ATH9K_PCI_BT_ANT_DIV },
327 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
328 0x0036,
329 PCI_VENDOR_ID_ATHEROS,
330 0x3026),
331 .driver_data = ATH9K_PCI_AR9565_1ANT | ATH9K_PCI_BT_ANT_DIV },
332 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
333 0x0036,
334 PCI_VENDOR_ID_ATHEROS,
335 0x302B),
336 .driver_data = ATH9K_PCI_AR9565_1ANT | ATH9K_PCI_BT_ANT_DIV },
337 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
338 0x0036,
339 PCI_VENDOR_ID_FOXCONN,
340 0xE069),
341 .driver_data = ATH9K_PCI_AR9565_1ANT | ATH9K_PCI_BT_ANT_DIV },
342 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
343 0x0036,
344 0x185F, /* WNC */
345 0x3028),
346 .driver_data = ATH9K_PCI_AR9565_1ANT | ATH9K_PCI_BT_ANT_DIV },
347 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
348 0x0036,
349 0x11AD, /* LITEON */
350 0x0622),
351 .driver_data = ATH9K_PCI_AR9565_1ANT | ATH9K_PCI_BT_ANT_DIV },
352 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
353 0x0036,
354 0x11AD, /* LITEON */
355 0x0672),
356 .driver_data = ATH9K_PCI_AR9565_1ANT | ATH9K_PCI_BT_ANT_DIV },
357 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
358 0x0036,
359 0x11AD, /* LITEON */
360 0x0662),
361 .driver_data = ATH9K_PCI_AR9565_1ANT | ATH9K_PCI_BT_ANT_DIV },
362 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
363 0x0036,
364 PCI_VENDOR_ID_AZWAVE,
365 0x213A),
366 .driver_data = ATH9K_PCI_AR9565_1ANT | ATH9K_PCI_BT_ANT_DIV },
367 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
368 0x0036,
369 PCI_VENDOR_ID_LENOVO,
370 0x3026),
371 .driver_data = ATH9K_PCI_AR9565_1ANT | ATH9K_PCI_BT_ANT_DIV },
372 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
373 0x0036,
374 PCI_VENDOR_ID_HP,
375 0x18E3),
376 .driver_data = ATH9K_PCI_AR9565_1ANT | ATH9K_PCI_BT_ANT_DIV },
377 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
378 0x0036,
379 PCI_VENDOR_ID_HP,
380 0x217F),
381 .driver_data = ATH9K_PCI_AR9565_1ANT | ATH9K_PCI_BT_ANT_DIV },
382 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
383 0x0036,
384 PCI_VENDOR_ID_DELL,
385 0x020E),
386 .driver_data = ATH9K_PCI_AR9565_1ANT | ATH9K_PCI_BT_ANT_DIV },
387
388 /* WB335 2-ANT */
389 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
390 0x0036,
391 PCI_VENDOR_ID_SAMSUNG,
392 0x411A),
393 .driver_data = ATH9K_PCI_AR9565_2ANT },
394 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
395 0x0036,
396 PCI_VENDOR_ID_SAMSUNG,
397 0x411B),
398 .driver_data = ATH9K_PCI_AR9565_2ANT },
399 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
400 0x0036,
401 PCI_VENDOR_ID_SAMSUNG,
402 0x411C),
403 .driver_data = ATH9K_PCI_AR9565_2ANT },
404 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
405 0x0036,
406 PCI_VENDOR_ID_SAMSUNG,
407 0x411D),
408 .driver_data = ATH9K_PCI_AR9565_2ANT },
409 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
410 0x0036,
411 PCI_VENDOR_ID_SAMSUNG,
412 0x411E),
413 .driver_data = ATH9K_PCI_AR9565_2ANT },
414
415 /* WB335 2-ANT / Antenna-Diversity */
416 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
417 0x0036,
418 PCI_VENDOR_ID_ATHEROS,
419 0x3027),
420 .driver_data = ATH9K_PCI_AR9565_2ANT | ATH9K_PCI_BT_ANT_DIV },
421 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
422 0x0036,
423 PCI_VENDOR_ID_ATHEROS,
424 0x302C),
425 .driver_data = ATH9K_PCI_AR9565_2ANT | ATH9K_PCI_BT_ANT_DIV },
426 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
427 0x0036,
428 0x11AD, /* LITEON */
429 0x0642),
430 .driver_data = ATH9K_PCI_AR9565_2ANT | ATH9K_PCI_BT_ANT_DIV },
431 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
432 0x0036,
433 0x11AD, /* LITEON */
434 0x0652),
435 .driver_data = ATH9K_PCI_AR9565_2ANT | ATH9K_PCI_BT_ANT_DIV },
436 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
437 0x0036,
438 0x11AD, /* LITEON */
439 0x0612),
440 .driver_data = ATH9K_PCI_AR9565_2ANT | ATH9K_PCI_BT_ANT_DIV },
441 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
442 0x0036,
443 PCI_VENDOR_ID_AZWAVE,
444 0x2130),
445 .driver_data = ATH9K_PCI_AR9565_2ANT | ATH9K_PCI_BT_ANT_DIV },
446 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
447 0x0036,
448 0x144F, /* ASKEY */
449 0x7202),
450 .driver_data = ATH9K_PCI_AR9565_2ANT | ATH9K_PCI_BT_ANT_DIV },
451 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
452 0x0036,
453 0x1B9A, /* XAVI */
454 0x2810),
455 .driver_data = ATH9K_PCI_AR9565_2ANT | ATH9K_PCI_BT_ANT_DIV },
456 { PCI_DEVICE_SUB(PCI_VENDOR_ID_ATHEROS,
457 0x0036,
458 0x185F, /* WNC */
459 0x3027),
460 .driver_data = ATH9K_PCI_AR9565_2ANT | ATH9K_PCI_BT_ANT_DIV },
461
462 /* PCI-E AR9565 (WB335) */
463 { PCI_VDEVICE(ATHEROS, 0x0036),
464 .driver_data = ATH9K_PCI_BT_ANT_DIV },
465
273 { 0 } 466 { 0 }
274}; 467};
275 468
diff --git a/drivers/net/wireless/ath/ath9k/rc.c b/drivers/net/wireless/ath/ath9k/rc.c
index d3d7c51fa6c8..d829bb62a3fc 100644
--- a/drivers/net/wireless/ath/ath9k/rc.c
+++ b/drivers/net/wireless/ath/ath9k/rc.c
@@ -1387,31 +1387,31 @@ static ssize_t read_file_rcstat(struct file *file, char __user *user_buf,
1387 int used_mcs = 0, used_htmode = 0; 1387 int used_mcs = 0, used_htmode = 0;
1388 1388
1389 if (WLAN_RC_PHY_HT(rc->rate_table->info[i].phy)) { 1389 if (WLAN_RC_PHY_HT(rc->rate_table->info[i].phy)) {
1390 used_mcs = snprintf(mcs, 5, "%d", 1390 used_mcs = scnprintf(mcs, 5, "%d",
1391 rc->rate_table->info[i].ratecode); 1391 rc->rate_table->info[i].ratecode);
1392 1392
1393 if (WLAN_RC_PHY_40(rc->rate_table->info[i].phy)) 1393 if (WLAN_RC_PHY_40(rc->rate_table->info[i].phy))
1394 used_htmode = snprintf(htmode, 5, "HT40"); 1394 used_htmode = scnprintf(htmode, 5, "HT40");
1395 else if (WLAN_RC_PHY_20(rc->rate_table->info[i].phy)) 1395 else if (WLAN_RC_PHY_20(rc->rate_table->info[i].phy))
1396 used_htmode = snprintf(htmode, 5, "HT20"); 1396 used_htmode = scnprintf(htmode, 5, "HT20");
1397 else 1397 else
1398 used_htmode = snprintf(htmode, 5, "????"); 1398 used_htmode = scnprintf(htmode, 5, "????");
1399 } 1399 }
1400 1400
1401 mcs[used_mcs] = '\0'; 1401 mcs[used_mcs] = '\0';
1402 htmode[used_htmode] = '\0'; 1402 htmode[used_htmode] = '\0';
1403 1403
1404 len += snprintf(buf + len, max - len, 1404 len += scnprintf(buf + len, max - len,
1405 "%6s %6s %3u.%d: " 1405 "%6s %6s %3u.%d: "
1406 "%10u %10u %10u %10u\n", 1406 "%10u %10u %10u %10u\n",
1407 htmode, 1407 htmode,
1408 mcs, 1408 mcs,
1409 ratekbps / 1000, 1409 ratekbps / 1000,
1410 (ratekbps % 1000) / 100, 1410 (ratekbps % 1000) / 100,
1411 stats->success, 1411 stats->success,
1412 stats->retries, 1412 stats->retries,
1413 stats->xretries, 1413 stats->xretries,
1414 stats->per); 1414 stats->per);
1415 } 1415 }
1416 1416
1417 if (len > max) 1417 if (len > max)
diff --git a/drivers/net/wireless/ath/ath9k/recv.c b/drivers/net/wireless/ath/ath9k/recv.c
index ab9e3a8410bc..95ddca5495d4 100644
--- a/drivers/net/wireless/ath/ath9k/recv.c
+++ b/drivers/net/wireless/ath/ath9k/recv.c
@@ -19,7 +19,7 @@
19#include "ath9k.h" 19#include "ath9k.h"
20#include "ar9003_mac.h" 20#include "ar9003_mac.h"
21 21
22#define SKB_CB_ATHBUF(__skb) (*((struct ath_buf **)__skb->cb)) 22#define SKB_CB_ATHBUF(__skb) (*((struct ath_rxbuf **)__skb->cb))
23 23
24static inline bool ath9k_check_auto_sleep(struct ath_softc *sc) 24static inline bool ath9k_check_auto_sleep(struct ath_softc *sc)
25{ 25{
@@ -35,7 +35,7 @@ static inline bool ath9k_check_auto_sleep(struct ath_softc *sc)
35 * buffer (or rx fifo). This can incorrectly acknowledge packets 35 * buffer (or rx fifo). This can incorrectly acknowledge packets
36 * to a sender if last desc is self-linked. 36 * to a sender if last desc is self-linked.
37 */ 37 */
38static void ath_rx_buf_link(struct ath_softc *sc, struct ath_buf *bf) 38static void ath_rx_buf_link(struct ath_softc *sc, struct ath_rxbuf *bf)
39{ 39{
40 struct ath_hw *ah = sc->sc_ah; 40 struct ath_hw *ah = sc->sc_ah;
41 struct ath_common *common = ath9k_hw_common(ah); 41 struct ath_common *common = ath9k_hw_common(ah);
@@ -68,7 +68,7 @@ static void ath_rx_buf_link(struct ath_softc *sc, struct ath_buf *bf)
68 sc->rx.rxlink = &ds->ds_link; 68 sc->rx.rxlink = &ds->ds_link;
69} 69}
70 70
71static void ath_rx_buf_relink(struct ath_softc *sc, struct ath_buf *bf) 71static void ath_rx_buf_relink(struct ath_softc *sc, struct ath_rxbuf *bf)
72{ 72{
73 if (sc->rx.buf_hold) 73 if (sc->rx.buf_hold)
74 ath_rx_buf_link(sc, sc->rx.buf_hold); 74 ath_rx_buf_link(sc, sc->rx.buf_hold);
@@ -112,13 +112,13 @@ static bool ath_rx_edma_buf_link(struct ath_softc *sc,
112 struct ath_hw *ah = sc->sc_ah; 112 struct ath_hw *ah = sc->sc_ah;
113 struct ath_rx_edma *rx_edma; 113 struct ath_rx_edma *rx_edma;
114 struct sk_buff *skb; 114 struct sk_buff *skb;
115 struct ath_buf *bf; 115 struct ath_rxbuf *bf;
116 116
117 rx_edma = &sc->rx.rx_edma[qtype]; 117 rx_edma = &sc->rx.rx_edma[qtype];
118 if (skb_queue_len(&rx_edma->rx_fifo) >= rx_edma->rx_fifo_hwsize) 118 if (skb_queue_len(&rx_edma->rx_fifo) >= rx_edma->rx_fifo_hwsize)
119 return false; 119 return false;
120 120
121 bf = list_first_entry(&sc->rx.rxbuf, struct ath_buf, list); 121 bf = list_first_entry(&sc->rx.rxbuf, struct ath_rxbuf, list);
122 list_del_init(&bf->list); 122 list_del_init(&bf->list);
123 123
124 skb = bf->bf_mpdu; 124 skb = bf->bf_mpdu;
@@ -138,7 +138,7 @@ static void ath_rx_addbuffer_edma(struct ath_softc *sc,
138 enum ath9k_rx_qtype qtype) 138 enum ath9k_rx_qtype qtype)
139{ 139{
140 struct ath_common *common = ath9k_hw_common(sc->sc_ah); 140 struct ath_common *common = ath9k_hw_common(sc->sc_ah);
141 struct ath_buf *bf, *tbf; 141 struct ath_rxbuf *bf, *tbf;
142 142
143 if (list_empty(&sc->rx.rxbuf)) { 143 if (list_empty(&sc->rx.rxbuf)) {
144 ath_dbg(common, QUEUE, "No free rx buf available\n"); 144 ath_dbg(common, QUEUE, "No free rx buf available\n");
@@ -154,7 +154,7 @@ static void ath_rx_addbuffer_edma(struct ath_softc *sc,
154static void ath_rx_remove_buffer(struct ath_softc *sc, 154static void ath_rx_remove_buffer(struct ath_softc *sc,
155 enum ath9k_rx_qtype qtype) 155 enum ath9k_rx_qtype qtype)
156{ 156{
157 struct ath_buf *bf; 157 struct ath_rxbuf *bf;
158 struct ath_rx_edma *rx_edma; 158 struct ath_rx_edma *rx_edma;
159 struct sk_buff *skb; 159 struct sk_buff *skb;
160 160
@@ -171,7 +171,7 @@ static void ath_rx_edma_cleanup(struct ath_softc *sc)
171{ 171{
172 struct ath_hw *ah = sc->sc_ah; 172 struct ath_hw *ah = sc->sc_ah;
173 struct ath_common *common = ath9k_hw_common(ah); 173 struct ath_common *common = ath9k_hw_common(ah);
174 struct ath_buf *bf; 174 struct ath_rxbuf *bf;
175 175
176 ath_rx_remove_buffer(sc, ATH9K_RX_QUEUE_LP); 176 ath_rx_remove_buffer(sc, ATH9K_RX_QUEUE_LP);
177 ath_rx_remove_buffer(sc, ATH9K_RX_QUEUE_HP); 177 ath_rx_remove_buffer(sc, ATH9K_RX_QUEUE_HP);
@@ -199,7 +199,7 @@ static int ath_rx_edma_init(struct ath_softc *sc, int nbufs)
199 struct ath_common *common = ath9k_hw_common(sc->sc_ah); 199 struct ath_common *common = ath9k_hw_common(sc->sc_ah);
200 struct ath_hw *ah = sc->sc_ah; 200 struct ath_hw *ah = sc->sc_ah;
201 struct sk_buff *skb; 201 struct sk_buff *skb;
202 struct ath_buf *bf; 202 struct ath_rxbuf *bf;
203 int error = 0, i; 203 int error = 0, i;
204 u32 size; 204 u32 size;
205 205
@@ -211,7 +211,7 @@ static int ath_rx_edma_init(struct ath_softc *sc, int nbufs)
211 ath_rx_edma_init_queue(&sc->rx.rx_edma[ATH9K_RX_QUEUE_HP], 211 ath_rx_edma_init_queue(&sc->rx.rx_edma[ATH9K_RX_QUEUE_HP],
212 ah->caps.rx_hp_qdepth); 212 ah->caps.rx_hp_qdepth);
213 213
214 size = sizeof(struct ath_buf) * nbufs; 214 size = sizeof(struct ath_rxbuf) * nbufs;
215 bf = devm_kzalloc(sc->dev, size, GFP_KERNEL); 215 bf = devm_kzalloc(sc->dev, size, GFP_KERNEL);
216 if (!bf) 216 if (!bf)
217 return -ENOMEM; 217 return -ENOMEM;
@@ -271,7 +271,7 @@ int ath_rx_init(struct ath_softc *sc, int nbufs)
271{ 271{
272 struct ath_common *common = ath9k_hw_common(sc->sc_ah); 272 struct ath_common *common = ath9k_hw_common(sc->sc_ah);
273 struct sk_buff *skb; 273 struct sk_buff *skb;
274 struct ath_buf *bf; 274 struct ath_rxbuf *bf;
275 int error = 0; 275 int error = 0;
276 276
277 spin_lock_init(&sc->sc_pcu_lock); 277 spin_lock_init(&sc->sc_pcu_lock);
@@ -332,7 +332,7 @@ void ath_rx_cleanup(struct ath_softc *sc)
332 struct ath_hw *ah = sc->sc_ah; 332 struct ath_hw *ah = sc->sc_ah;
333 struct ath_common *common = ath9k_hw_common(ah); 333 struct ath_common *common = ath9k_hw_common(ah);
334 struct sk_buff *skb; 334 struct sk_buff *skb;
335 struct ath_buf *bf; 335 struct ath_rxbuf *bf;
336 336
337 if (sc->sc_ah->caps.hw_caps & ATH9K_HW_CAP_EDMA) { 337 if (sc->sc_ah->caps.hw_caps & ATH9K_HW_CAP_EDMA) {
338 ath_rx_edma_cleanup(sc); 338 ath_rx_edma_cleanup(sc);
@@ -375,6 +375,9 @@ u32 ath_calcrxfilter(struct ath_softc *sc)
375{ 375{
376 u32 rfilt; 376 u32 rfilt;
377 377
378 if (config_enabled(CONFIG_ATH9K_TX99))
379 return 0;
380
378 rfilt = ATH9K_RX_FILTER_UCAST | ATH9K_RX_FILTER_BCAST 381 rfilt = ATH9K_RX_FILTER_UCAST | ATH9K_RX_FILTER_BCAST
379 | ATH9K_RX_FILTER_MCAST; 382 | ATH9K_RX_FILTER_MCAST;
380 383
@@ -427,7 +430,7 @@ u32 ath_calcrxfilter(struct ath_softc *sc)
427int ath_startrecv(struct ath_softc *sc) 430int ath_startrecv(struct ath_softc *sc)
428{ 431{
429 struct ath_hw *ah = sc->sc_ah; 432 struct ath_hw *ah = sc->sc_ah;
430 struct ath_buf *bf, *tbf; 433 struct ath_rxbuf *bf, *tbf;
431 434
432 if (ah->caps.hw_caps & ATH9K_HW_CAP_EDMA) { 435 if (ah->caps.hw_caps & ATH9K_HW_CAP_EDMA) {
433 ath_edma_start_recv(sc); 436 ath_edma_start_recv(sc);
@@ -447,7 +450,7 @@ int ath_startrecv(struct ath_softc *sc)
447 if (list_empty(&sc->rx.rxbuf)) 450 if (list_empty(&sc->rx.rxbuf))
448 goto start_recv; 451 goto start_recv;
449 452
450 bf = list_first_entry(&sc->rx.rxbuf, struct ath_buf, list); 453 bf = list_first_entry(&sc->rx.rxbuf, struct ath_rxbuf, list);
451 ath9k_hw_putrxbuf(ah, bf->bf_daddr); 454 ath9k_hw_putrxbuf(ah, bf->bf_daddr);
452 ath9k_hw_rxena(ah); 455 ath9k_hw_rxena(ah);
453 456
@@ -603,13 +606,13 @@ static void ath_rx_ps(struct ath_softc *sc, struct sk_buff *skb, bool mybeacon)
603static bool ath_edma_get_buffers(struct ath_softc *sc, 606static bool ath_edma_get_buffers(struct ath_softc *sc,
604 enum ath9k_rx_qtype qtype, 607 enum ath9k_rx_qtype qtype,
605 struct ath_rx_status *rs, 608 struct ath_rx_status *rs,
606 struct ath_buf **dest) 609 struct ath_rxbuf **dest)
607{ 610{
608 struct ath_rx_edma *rx_edma = &sc->rx.rx_edma[qtype]; 611 struct ath_rx_edma *rx_edma = &sc->rx.rx_edma[qtype];
609 struct ath_hw *ah = sc->sc_ah; 612 struct ath_hw *ah = sc->sc_ah;
610 struct ath_common *common = ath9k_hw_common(ah); 613 struct ath_common *common = ath9k_hw_common(ah);
611 struct sk_buff *skb; 614 struct sk_buff *skb;
612 struct ath_buf *bf; 615 struct ath_rxbuf *bf;
613 int ret; 616 int ret;
614 617
615 skb = skb_peek(&rx_edma->rx_fifo); 618 skb = skb_peek(&rx_edma->rx_fifo);
@@ -653,11 +656,11 @@ static bool ath_edma_get_buffers(struct ath_softc *sc,
653 return true; 656 return true;
654} 657}
655 658
656static struct ath_buf *ath_edma_get_next_rx_buf(struct ath_softc *sc, 659static struct ath_rxbuf *ath_edma_get_next_rx_buf(struct ath_softc *sc,
657 struct ath_rx_status *rs, 660 struct ath_rx_status *rs,
658 enum ath9k_rx_qtype qtype) 661 enum ath9k_rx_qtype qtype)
659{ 662{
660 struct ath_buf *bf = NULL; 663 struct ath_rxbuf *bf = NULL;
661 664
662 while (ath_edma_get_buffers(sc, qtype, rs, &bf)) { 665 while (ath_edma_get_buffers(sc, qtype, rs, &bf)) {
663 if (!bf) 666 if (!bf)
@@ -668,13 +671,13 @@ static struct ath_buf *ath_edma_get_next_rx_buf(struct ath_softc *sc,
668 return NULL; 671 return NULL;
669} 672}
670 673
671static struct ath_buf *ath_get_next_rx_buf(struct ath_softc *sc, 674static struct ath_rxbuf *ath_get_next_rx_buf(struct ath_softc *sc,
672 struct ath_rx_status *rs) 675 struct ath_rx_status *rs)
673{ 676{
674 struct ath_hw *ah = sc->sc_ah; 677 struct ath_hw *ah = sc->sc_ah;
675 struct ath_common *common = ath9k_hw_common(ah); 678 struct ath_common *common = ath9k_hw_common(ah);
676 struct ath_desc *ds; 679 struct ath_desc *ds;
677 struct ath_buf *bf; 680 struct ath_rxbuf *bf;
678 int ret; 681 int ret;
679 682
680 if (list_empty(&sc->rx.rxbuf)) { 683 if (list_empty(&sc->rx.rxbuf)) {
@@ -682,7 +685,7 @@ static struct ath_buf *ath_get_next_rx_buf(struct ath_softc *sc,
682 return NULL; 685 return NULL;
683 } 686 }
684 687
685 bf = list_first_entry(&sc->rx.rxbuf, struct ath_buf, list); 688 bf = list_first_entry(&sc->rx.rxbuf, struct ath_rxbuf, list);
686 if (bf == sc->rx.buf_hold) 689 if (bf == sc->rx.buf_hold)
687 return NULL; 690 return NULL;
688 691
@@ -702,7 +705,7 @@ static struct ath_buf *ath_get_next_rx_buf(struct ath_softc *sc,
702 ret = ath9k_hw_rxprocdesc(ah, ds, rs); 705 ret = ath9k_hw_rxprocdesc(ah, ds, rs);
703 if (ret == -EINPROGRESS) { 706 if (ret == -EINPROGRESS) {
704 struct ath_rx_status trs; 707 struct ath_rx_status trs;
705 struct ath_buf *tbf; 708 struct ath_rxbuf *tbf;
706 struct ath_desc *tds; 709 struct ath_desc *tds;
707 710
708 memset(&trs, 0, sizeof(trs)); 711 memset(&trs, 0, sizeof(trs));
@@ -711,7 +714,7 @@ static struct ath_buf *ath_get_next_rx_buf(struct ath_softc *sc,
711 return NULL; 714 return NULL;
712 } 715 }
713 716
714 tbf = list_entry(bf->list.next, struct ath_buf, list); 717 tbf = list_entry(bf->list.next, struct ath_rxbuf, list);
715 718
716 /* 719 /*
717 * On some hardware the descriptor status words could 720 * On some hardware the descriptor status words could
@@ -972,14 +975,15 @@ static int ath_process_fft(struct ath_softc *sc, struct ieee80211_hdr *hdr,
972{ 975{
973#ifdef CONFIG_ATH9K_DEBUGFS 976#ifdef CONFIG_ATH9K_DEBUGFS
974 struct ath_hw *ah = sc->sc_ah; 977 struct ath_hw *ah = sc->sc_ah;
975 u8 bins[SPECTRAL_HT20_NUM_BINS]; 978 u8 num_bins, *bins, *vdata = (u8 *)hdr;
976 u8 *vdata = (u8 *)hdr; 979 struct fft_sample_ht20 fft_sample_20;
977 struct fft_sample_ht20 fft_sample; 980 struct fft_sample_ht20_40 fft_sample_40;
981 struct fft_sample_tlv *tlv;
978 struct ath_radar_info *radar_info; 982 struct ath_radar_info *radar_info;
979 struct ath_ht20_mag_info *mag_info;
980 int len = rs->rs_datalen; 983 int len = rs->rs_datalen;
981 int dc_pos; 984 int dc_pos;
982 u16 length, max_magnitude; 985 u16 fft_len, length, freq = ah->curchan->chan->center_freq;
986 enum nl80211_channel_type chan_type;
983 987
984 /* AR9280 and before report via ATH9K_PHYERR_RADAR, AR93xx and newer 988 /* AR9280 and before report via ATH9K_PHYERR_RADAR, AR93xx and newer
985 * via ATH9K_PHYERR_SPECTRAL. Haven't seen ATH9K_PHYERR_FALSE_RADAR_EXT 989 * via ATH9K_PHYERR_SPECTRAL. Haven't seen ATH9K_PHYERR_FALSE_RADAR_EXT
@@ -997,45 +1001,44 @@ static int ath_process_fft(struct ath_softc *sc, struct ieee80211_hdr *hdr,
997 if (!(radar_info->pulse_bw_info & SPECTRAL_SCAN_BITMASK)) 1001 if (!(radar_info->pulse_bw_info & SPECTRAL_SCAN_BITMASK))
998 return 0; 1002 return 0;
999 1003
1000 /* Variation in the data length is possible and will be fixed later. 1004 chan_type = cfg80211_get_chandef_type(&sc->hw->conf.chandef);
1001 * Note that we only support HT20 for now. 1005 if ((chan_type == NL80211_CHAN_HT40MINUS) ||
1002 * 1006 (chan_type == NL80211_CHAN_HT40PLUS)) {
1003 * TODO: add HT20_40 support as well. 1007 fft_len = SPECTRAL_HT20_40_TOTAL_DATA_LEN;
1004 */ 1008 num_bins = SPECTRAL_HT20_40_NUM_BINS;
1005 if ((len > SPECTRAL_HT20_TOTAL_DATA_LEN + 2) || 1009 bins = (u8 *)fft_sample_40.data;
1006 (len < SPECTRAL_HT20_TOTAL_DATA_LEN - 1)) 1010 } else {
1007 return 1; 1011 fft_len = SPECTRAL_HT20_TOTAL_DATA_LEN;
1008 1012 num_bins = SPECTRAL_HT20_NUM_BINS;
1009 fft_sample.tlv.type = ATH_FFT_SAMPLE_HT20; 1013 bins = (u8 *)fft_sample_20.data;
1010 length = sizeof(fft_sample) - sizeof(fft_sample.tlv); 1014 }
1011 fft_sample.tlv.length = __cpu_to_be16(length);
1012 1015
1013 fft_sample.freq = __cpu_to_be16(ah->curchan->chan->center_freq); 1016 /* Variation in the data length is possible and will be fixed later */
1014 fft_sample.rssi = fix_rssi_inv_only(rs->rs_rssi_ctl0); 1017 if ((len > fft_len + 2) || (len < fft_len - 1))
1015 fft_sample.noise = ah->noise; 1018 return 1;
1016 1019
1017 switch (len - SPECTRAL_HT20_TOTAL_DATA_LEN) { 1020 switch (len - fft_len) {
1018 case 0: 1021 case 0:
1019 /* length correct, nothing to do. */ 1022 /* length correct, nothing to do. */
1020 memcpy(bins, vdata, SPECTRAL_HT20_NUM_BINS); 1023 memcpy(bins, vdata, num_bins);
1021 break; 1024 break;
1022 case -1: 1025 case -1:
1023 /* first byte missing, duplicate it. */ 1026 /* first byte missing, duplicate it. */
1024 memcpy(&bins[1], vdata, SPECTRAL_HT20_NUM_BINS - 1); 1027 memcpy(&bins[1], vdata, num_bins - 1);
1025 bins[0] = vdata[0]; 1028 bins[0] = vdata[0];
1026 break; 1029 break;
1027 case 2: 1030 case 2:
1028 /* MAC added 2 extra bytes at bin 30 and 32, remove them. */ 1031 /* MAC added 2 extra bytes at bin 30 and 32, remove them. */
1029 memcpy(bins, vdata, 30); 1032 memcpy(bins, vdata, 30);
1030 bins[30] = vdata[31]; 1033 bins[30] = vdata[31];
1031 memcpy(&bins[31], &vdata[33], SPECTRAL_HT20_NUM_BINS - 31); 1034 memcpy(&bins[31], &vdata[33], num_bins - 31);
1032 break; 1035 break;
1033 case 1: 1036 case 1:
1034 /* MAC added 2 extra bytes AND first byte is missing. */ 1037 /* MAC added 2 extra bytes AND first byte is missing. */
1035 bins[0] = vdata[0]; 1038 bins[0] = vdata[0];
1036 memcpy(&bins[0], vdata, 30); 1039 memcpy(&bins[1], vdata, 30);
1037 bins[31] = vdata[31]; 1040 bins[31] = vdata[31];
1038 memcpy(&bins[32], &vdata[33], SPECTRAL_HT20_NUM_BINS - 32); 1041 memcpy(&bins[32], &vdata[33], num_bins - 32);
1039 break; 1042 break;
1040 default: 1043 default:
1041 return 1; 1044 return 1;
@@ -1044,23 +1047,93 @@ static int ath_process_fft(struct ath_softc *sc, struct ieee80211_hdr *hdr,
1044 /* DC value (value in the middle) is the blind spot of the spectral 1047 /* DC value (value in the middle) is the blind spot of the spectral
1045 * sample and invalid, interpolate it. 1048 * sample and invalid, interpolate it.
1046 */ 1049 */
1047 dc_pos = SPECTRAL_HT20_NUM_BINS / 2; 1050 dc_pos = num_bins / 2;
1048 bins[dc_pos] = (bins[dc_pos + 1] + bins[dc_pos - 1]) / 2; 1051 bins[dc_pos] = (bins[dc_pos + 1] + bins[dc_pos - 1]) / 2;
1049 1052
1050 /* mag data is at the end of the frame, in front of radar_info */ 1053 if ((chan_type == NL80211_CHAN_HT40MINUS) ||
1051 mag_info = ((struct ath_ht20_mag_info *)radar_info) - 1; 1054 (chan_type == NL80211_CHAN_HT40PLUS)) {
1055 s8 lower_rssi, upper_rssi;
1056 s16 ext_nf;
1057 u8 lower_max_index, upper_max_index;
1058 u8 lower_bitmap_w, upper_bitmap_w;
1059 u16 lower_mag, upper_mag;
1060 struct ath9k_hw_cal_data *caldata = ah->caldata;
1061 struct ath_ht20_40_mag_info *mag_info;
1062
1063 if (caldata)
1064 ext_nf = ath9k_hw_getchan_noise(ah, ah->curchan,
1065 caldata->nfCalHist[3].privNF);
1066 else
1067 ext_nf = ATH_DEFAULT_NOISE_FLOOR;
1068
1069 length = sizeof(fft_sample_40) - sizeof(struct fft_sample_tlv);
1070 fft_sample_40.tlv.type = ATH_FFT_SAMPLE_HT20_40;
1071 fft_sample_40.tlv.length = __cpu_to_be16(length);
1072 fft_sample_40.freq = __cpu_to_be16(freq);
1073 fft_sample_40.channel_type = chan_type;
1074
1075 if (chan_type == NL80211_CHAN_HT40PLUS) {
1076 lower_rssi = fix_rssi_inv_only(rs->rs_rssi_ctl0);
1077 upper_rssi = fix_rssi_inv_only(rs->rs_rssi_ext0);
1052 1078
1053 /* copy raw bins without scaling them */ 1079 fft_sample_40.lower_noise = ah->noise;
1054 memcpy(fft_sample.data, bins, SPECTRAL_HT20_NUM_BINS); 1080 fft_sample_40.upper_noise = ext_nf;
1055 fft_sample.max_exp = mag_info->max_exp & 0xf; 1081 } else {
1082 lower_rssi = fix_rssi_inv_only(rs->rs_rssi_ext0);
1083 upper_rssi = fix_rssi_inv_only(rs->rs_rssi_ctl0);
1056 1084
1057 max_magnitude = spectral_max_magnitude(mag_info->all_bins); 1085 fft_sample_40.lower_noise = ext_nf;
1058 fft_sample.max_magnitude = __cpu_to_be16(max_magnitude); 1086 fft_sample_40.upper_noise = ah->noise;
1059 fft_sample.max_index = spectral_max_index(mag_info->all_bins); 1087 }
1060 fft_sample.bitmap_weight = spectral_bitmap_weight(mag_info->all_bins); 1088 fft_sample_40.lower_rssi = lower_rssi;
1061 fft_sample.tsf = __cpu_to_be64(tsf); 1089 fft_sample_40.upper_rssi = upper_rssi;
1090
1091 mag_info = ((struct ath_ht20_40_mag_info *)radar_info) - 1;
1092 lower_mag = spectral_max_magnitude(mag_info->lower_bins);
1093 upper_mag = spectral_max_magnitude(mag_info->upper_bins);
1094 fft_sample_40.lower_max_magnitude = __cpu_to_be16(lower_mag);
1095 fft_sample_40.upper_max_magnitude = __cpu_to_be16(upper_mag);
1096 lower_max_index = spectral_max_index(mag_info->lower_bins);
1097 upper_max_index = spectral_max_index(mag_info->upper_bins);
1098 fft_sample_40.lower_max_index = lower_max_index;
1099 fft_sample_40.upper_max_index = upper_max_index;
1100 lower_bitmap_w = spectral_bitmap_weight(mag_info->lower_bins);
1101 upper_bitmap_w = spectral_bitmap_weight(mag_info->upper_bins);
1102 fft_sample_40.lower_bitmap_weight = lower_bitmap_w;
1103 fft_sample_40.upper_bitmap_weight = upper_bitmap_w;
1104 fft_sample_40.max_exp = mag_info->max_exp & 0xf;
1105
1106 fft_sample_40.tsf = __cpu_to_be64(tsf);
1107
1108 tlv = (struct fft_sample_tlv *)&fft_sample_40;
1109 } else {
1110 u8 max_index, bitmap_w;
1111 u16 magnitude;
1112 struct ath_ht20_mag_info *mag_info;
1113
1114 length = sizeof(fft_sample_20) - sizeof(struct fft_sample_tlv);
1115 fft_sample_20.tlv.type = ATH_FFT_SAMPLE_HT20;
1116 fft_sample_20.tlv.length = __cpu_to_be16(length);
1117 fft_sample_20.freq = __cpu_to_be16(freq);
1118
1119 fft_sample_20.rssi = fix_rssi_inv_only(rs->rs_rssi_ctl0);
1120 fft_sample_20.noise = ah->noise;
1121
1122 mag_info = ((struct ath_ht20_mag_info *)radar_info) - 1;
1123 magnitude = spectral_max_magnitude(mag_info->all_bins);
1124 fft_sample_20.max_magnitude = __cpu_to_be16(magnitude);
1125 max_index = spectral_max_index(mag_info->all_bins);
1126 fft_sample_20.max_index = max_index;
1127 bitmap_w = spectral_bitmap_weight(mag_info->all_bins);
1128 fft_sample_20.bitmap_weight = bitmap_w;
1129 fft_sample_20.max_exp = mag_info->max_exp & 0xf;
1130
1131 fft_sample_20.tsf = __cpu_to_be64(tsf);
1132
1133 tlv = (struct fft_sample_tlv *)&fft_sample_20;
1134 }
1062 1135
1063 ath_debug_send_fft_sample(sc, &fft_sample.tlv); 1136 ath_debug_send_fft_sample(sc, tlv);
1064 return 1; 1137 return 1;
1065#else 1138#else
1066 return 0; 1139 return 0;
@@ -1308,7 +1381,7 @@ static void ath9k_apply_ampdu_details(struct ath_softc *sc,
1308 1381
1309int ath_rx_tasklet(struct ath_softc *sc, int flush, bool hp) 1382int ath_rx_tasklet(struct ath_softc *sc, int flush, bool hp)
1310{ 1383{
1311 struct ath_buf *bf; 1384 struct ath_rxbuf *bf;
1312 struct sk_buff *skb = NULL, *requeue_skb, *hdr_skb; 1385 struct sk_buff *skb = NULL, *requeue_skb, *hdr_skb;
1313 struct ieee80211_rx_status *rxs; 1386 struct ieee80211_rx_status *rxs;
1314 struct ath_hw *ah = sc->sc_ah; 1387 struct ath_hw *ah = sc->sc_ah;
diff --git a/drivers/net/wireless/ath/ath9k/wmi.h b/drivers/net/wireless/ath/ath9k/wmi.h
index fde6da619f30..0db37f230018 100644
--- a/drivers/net/wireless/ath/ath9k/wmi.h
+++ b/drivers/net/wireless/ath/ath9k/wmi.h
@@ -39,7 +39,7 @@ struct wmi_fw_version {
39struct wmi_event_swba { 39struct wmi_event_swba {
40 __be64 tsf; 40 __be64 tsf;
41 u8 beacon_pending; 41 u8 beacon_pending;
42}; 42} __packed;
43 43
44/* 44/*
45 * 64 - HTC header - WMI header - 1 / txstatus 45 * 64 - HTC header - WMI header - 1 / txstatus
diff --git a/drivers/net/wireless/ath/ath9k/xmit.c b/drivers/net/wireless/ath/ath9k/xmit.c
index dd30452df966..09cdbcd09739 100644
--- a/drivers/net/wireless/ath/ath9k/xmit.c
+++ b/drivers/net/wireless/ath/ath9k/xmit.c
@@ -1241,12 +1241,13 @@ static void ath_tx_fill_desc(struct ath_softc *sc, struct ath_buf *bf,
1241 if (bf->bf_next) 1241 if (bf->bf_next)
1242 info.link = bf->bf_next->bf_daddr; 1242 info.link = bf->bf_next->bf_daddr;
1243 else 1243 else
1244 info.link = 0; 1244 info.link = (sc->tx99_state) ? bf->bf_daddr : 0;
1245 1245
1246 if (!bf_first) { 1246 if (!bf_first) {
1247 bf_first = bf; 1247 bf_first = bf;
1248 1248
1249 info.flags = ATH9K_TXDESC_INTREQ; 1249 if (!sc->tx99_state)
1250 info.flags = ATH9K_TXDESC_INTREQ;
1250 if ((tx_info->flags & IEEE80211_TX_CTL_CLEAR_PS_FILT) || 1251 if ((tx_info->flags & IEEE80211_TX_CTL_CLEAR_PS_FILT) ||
1251 txq == sc->tx.uapsdq) 1252 txq == sc->tx.uapsdq)
1252 info.flags |= ATH9K_TXDESC_CLRDMASK; 1253 info.flags |= ATH9K_TXDESC_CLRDMASK;
@@ -1704,16 +1705,9 @@ int ath_cabq_update(struct ath_softc *sc)
1704 int qnum = sc->beacon.cabq->axq_qnum; 1705 int qnum = sc->beacon.cabq->axq_qnum;
1705 1706
1706 ath9k_hw_get_txq_props(sc->sc_ah, qnum, &qi); 1707 ath9k_hw_get_txq_props(sc->sc_ah, qnum, &qi);
1707 /*
1708 * Ensure the readytime % is within the bounds.
1709 */
1710 if (sc->config.cabqReadytime < ATH9K_READY_TIME_LO_BOUND)
1711 sc->config.cabqReadytime = ATH9K_READY_TIME_LO_BOUND;
1712 else if (sc->config.cabqReadytime > ATH9K_READY_TIME_HI_BOUND)
1713 sc->config.cabqReadytime = ATH9K_READY_TIME_HI_BOUND;
1714 1708
1715 qi.tqi_readyTime = (cur_conf->beacon_interval * 1709 qi.tqi_readyTime = (cur_conf->beacon_interval *
1716 sc->config.cabqReadytime) / 100; 1710 ATH_CABQ_READY_TIME) / 100;
1717 ath_txq_update(sc, qnum, &qi); 1711 ath_txq_update(sc, qnum, &qi);
1718 1712
1719 return 0; 1713 return 0;
@@ -1948,7 +1942,7 @@ static void ath_tx_txqaddbuf(struct ath_softc *sc, struct ath_txq *txq,
1948 txq->axq_qnum, ito64(bf->bf_daddr), bf->bf_desc); 1942 txq->axq_qnum, ito64(bf->bf_daddr), bf->bf_desc);
1949 } 1943 }
1950 1944
1951 if (!edma) { 1945 if (!edma || sc->tx99_state) {
1952 TX_STAT_INC(txq->axq_qnum, txstart); 1946 TX_STAT_INC(txq->axq_qnum, txstart);
1953 ath9k_hw_txstart(ah, txq->axq_qnum); 1947 ath9k_hw_txstart(ah, txq->axq_qnum);
1954 } 1948 }
@@ -2027,6 +2021,9 @@ static void setup_frame_info(struct ieee80211_hw *hw,
2027 fi->keyix = ATH9K_TXKEYIX_INVALID; 2021 fi->keyix = ATH9K_TXKEYIX_INVALID;
2028 fi->keytype = keytype; 2022 fi->keytype = keytype;
2029 fi->framelen = framelen; 2023 fi->framelen = framelen;
2024
2025 if (!rate)
2026 return;
2030 fi->rtscts_rate = rate->hw_value; 2027 fi->rtscts_rate = rate->hw_value;
2031 if (short_preamble) 2028 if (short_preamble)
2032 fi->rtscts_rate |= rate->hw_value_short; 2029 fi->rtscts_rate |= rate->hw_value_short;
@@ -2037,8 +2034,7 @@ u8 ath_txchainmask_reduction(struct ath_softc *sc, u8 chainmask, u32 rate)
2037 struct ath_hw *ah = sc->sc_ah; 2034 struct ath_hw *ah = sc->sc_ah;
2038 struct ath9k_channel *curchan = ah->curchan; 2035 struct ath9k_channel *curchan = ah->curchan;
2039 2036
2040 if ((ah->caps.hw_caps & ATH9K_HW_CAP_APM) && 2037 if ((ah->caps.hw_caps & ATH9K_HW_CAP_APM) && IS_CHAN_5GHZ(curchan) &&
2041 (curchan->channelFlags & CHANNEL_5GHZ) &&
2042 (chainmask == 0x7) && (rate < 0x90)) 2038 (chainmask == 0x7) && (rate < 0x90))
2043 return 0x3; 2039 return 0x3;
2044 else if (AR_SREV_9462(ah) && ath9k_hw_btcoex_is_enabled(ah) && 2040 else if (AR_SREV_9462(ah) && ath9k_hw_btcoex_is_enabled(ah) &&
@@ -2329,7 +2325,7 @@ static void ath_tx_complete(struct ath_softc *sc, struct sk_buff *skb,
2329 ath_dbg(common, XMIT, "TX complete: skb: %p\n", skb); 2325 ath_dbg(common, XMIT, "TX complete: skb: %p\n", skb);
2330 2326
2331 if (sc->sc_ah->caldata) 2327 if (sc->sc_ah->caldata)
2332 sc->sc_ah->caldata->paprd_packet_sent = true; 2328 set_bit(PAPRD_PACKET_SENT, &sc->sc_ah->caldata->cal_flags);
2333 2329
2334 if (!(tx_flags & ATH_TX_ERROR)) 2330 if (!(tx_flags & ATH_TX_ERROR))
2335 /* Frame was ACKed */ 2331 /* Frame was ACKed */
@@ -2379,6 +2375,8 @@ static void ath_tx_complete_buf(struct ath_softc *sc, struct ath_buf *bf,
2379 2375
2380 dma_unmap_single(sc->dev, bf->bf_buf_addr, skb->len, DMA_TO_DEVICE); 2376 dma_unmap_single(sc->dev, bf->bf_buf_addr, skb->len, DMA_TO_DEVICE);
2381 bf->bf_buf_addr = 0; 2377 bf->bf_buf_addr = 0;
2378 if (sc->tx99_state)
2379 goto skip_tx_complete;
2382 2380
2383 if (bf->bf_state.bfs_paprd) { 2381 if (bf->bf_state.bfs_paprd) {
2384 if (time_after(jiffies, 2382 if (time_after(jiffies,
@@ -2391,6 +2389,7 @@ static void ath_tx_complete_buf(struct ath_softc *sc, struct ath_buf *bf,
2391 ath_debug_stat_tx(sc, bf, ts, txq, tx_flags); 2389 ath_debug_stat_tx(sc, bf, ts, txq, tx_flags);
2392 ath_tx_complete(sc, skb, tx_flags, txq); 2390 ath_tx_complete(sc, skb, tx_flags, txq);
2393 } 2391 }
2392skip_tx_complete:
2394 /* At this point, skb (bf->bf_mpdu) is consumed...make sure we don't 2393 /* At this point, skb (bf->bf_mpdu) is consumed...make sure we don't
2395 * accidentally reference it later. 2394 * accidentally reference it later.
2396 */ 2395 */
@@ -2749,3 +2748,46 @@ void ath_tx_node_cleanup(struct ath_softc *sc, struct ath_node *an)
2749 ath_txq_unlock(sc, txq); 2748 ath_txq_unlock(sc, txq);
2750 } 2749 }
2751} 2750}
2751
2752int ath9k_tx99_send(struct ath_softc *sc, struct sk_buff *skb,
2753 struct ath_tx_control *txctl)
2754{
2755 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *) skb->data;
2756 struct ath_frame_info *fi = get_frame_info(skb);
2757 struct ath_common *common = ath9k_hw_common(sc->sc_ah);
2758 struct ath_buf *bf;
2759 int padpos, padsize;
2760
2761 padpos = ieee80211_hdrlen(hdr->frame_control);
2762 padsize = padpos & 3;
2763
2764 if (padsize && skb->len > padpos) {
2765 if (skb_headroom(skb) < padsize) {
2766 ath_dbg(common, XMIT,
2767 "tx99 padding failed\n");
2768 return -EINVAL;
2769 }
2770
2771 skb_push(skb, padsize);
2772 memmove(skb->data, skb->data + padsize, padpos);
2773 }
2774
2775 fi->keyix = ATH9K_TXKEYIX_INVALID;
2776 fi->framelen = skb->len + FCS_LEN;
2777 fi->keytype = ATH9K_KEY_TYPE_CLEAR;
2778
2779 bf = ath_tx_setup_buffer(sc, txctl->txq, NULL, skb);
2780 if (!bf) {
2781 ath_dbg(common, XMIT, "tx99 buffer setup failed\n");
2782 return -EINVAL;
2783 }
2784
2785 ath_set_rates(sc->tx99_vif, NULL, bf);
2786
2787 ath9k_hw_set_desc_link(sc->sc_ah, bf->bf_desc, bf->bf_daddr);
2788 ath9k_hw_tx99_start(sc->sc_ah, txctl->txq->axq_qnum);
2789
2790 ath_tx_send_normal(sc, txctl->txq, NULL, skb);
2791
2792 return 0;
2793}
diff --git a/drivers/net/wireless/ath/ath9k/dfs_pattern_detector.c b/drivers/net/wireless/ath/dfs_pattern_detector.c
index 491305c81fce..a1a69c5db409 100644
--- a/drivers/net/wireless/ath/ath9k/dfs_pattern_detector.c
+++ b/drivers/net/wireless/ath/dfs_pattern_detector.c
@@ -19,7 +19,7 @@
19 19
20#include "dfs_pattern_detector.h" 20#include "dfs_pattern_detector.h"
21#include "dfs_pri_detector.h" 21#include "dfs_pri_detector.h"
22#include "ath9k.h" 22#include "ath.h"
23 23
24/* 24/*
25 * tolerated deviation of radar time stamp in usecs on both sides 25 * tolerated deviation of radar time stamp in usecs on both sides
@@ -143,7 +143,6 @@ channel_detector_create(struct dfs_pattern_detector *dpd, u16 freq)
143{ 143{
144 u32 sz, i; 144 u32 sz, i;
145 struct channel_detector *cd; 145 struct channel_detector *cd;
146 struct ath_common *common = ath9k_hw_common(dpd->ah);
147 146
148 cd = kmalloc(sizeof(*cd), GFP_ATOMIC); 147 cd = kmalloc(sizeof(*cd), GFP_ATOMIC);
149 if (cd == NULL) 148 if (cd == NULL)
@@ -167,7 +166,7 @@ channel_detector_create(struct dfs_pattern_detector *dpd, u16 freq)
167 return cd; 166 return cd;
168 167
169fail: 168fail:
170 ath_dbg(common, DFS, 169 ath_dbg(dpd->common, DFS,
171 "failed to allocate channel_detector for freq=%d\n", freq); 170 "failed to allocate channel_detector for freq=%d\n", freq);
172 channel_detector_exit(dpd, cd); 171 channel_detector_exit(dpd, cd);
173 return NULL; 172 return NULL;
@@ -242,7 +241,7 @@ dpd_add_pulse(struct dfs_pattern_detector *dpd, struct pulse_event *event)
242 struct pri_detector *pd = cd->detectors[i]; 241 struct pri_detector *pd = cd->detectors[i];
243 struct pri_sequence *ps = pd->add_pulse(pd, event); 242 struct pri_sequence *ps = pd->add_pulse(pd, event);
244 if (ps != NULL) { 243 if (ps != NULL) {
245 ath_dbg(ath9k_hw_common(dpd->ah), DFS, 244 ath_dbg(dpd->common, DFS,
246 "DFS: radar found on freq=%d: id=%d, pri=%d, " 245 "DFS: radar found on freq=%d: id=%d, pri=%d, "
247 "count=%d, count_false=%d\n", 246 "count=%d, count_false=%d\n",
248 event->freq, pd->rs->type_id, 247 event->freq, pd->rs->type_id,
@@ -254,6 +253,12 @@ dpd_add_pulse(struct dfs_pattern_detector *dpd, struct pulse_event *event)
254 return false; 253 return false;
255} 254}
256 255
256static struct ath_dfs_pool_stats
257dpd_get_stats(struct dfs_pattern_detector *dpd)
258{
259 return global_dfs_pool_stats;
260}
261
257static bool dpd_set_domain(struct dfs_pattern_detector *dpd, 262static bool dpd_set_domain(struct dfs_pattern_detector *dpd,
258 enum nl80211_dfs_regions region) 263 enum nl80211_dfs_regions region)
259{ 264{
@@ -284,14 +289,18 @@ static struct dfs_pattern_detector default_dpd = {
284 .exit = dpd_exit, 289 .exit = dpd_exit,
285 .set_dfs_domain = dpd_set_domain, 290 .set_dfs_domain = dpd_set_domain,
286 .add_pulse = dpd_add_pulse, 291 .add_pulse = dpd_add_pulse,
292 .get_stats = dpd_get_stats,
287 .region = NL80211_DFS_UNSET, 293 .region = NL80211_DFS_UNSET,
288}; 294};
289 295
290struct dfs_pattern_detector * 296struct dfs_pattern_detector *
291dfs_pattern_detector_init(struct ath_hw *ah, enum nl80211_dfs_regions region) 297dfs_pattern_detector_init(struct ath_common *common,
298 enum nl80211_dfs_regions region)
292{ 299{
293 struct dfs_pattern_detector *dpd; 300 struct dfs_pattern_detector *dpd;
294 struct ath_common *common = ath9k_hw_common(ah); 301
302 if (!config_enabled(CONFIG_CFG80211_CERTIFICATION_ONUS))
303 return NULL;
295 304
296 dpd = kmalloc(sizeof(*dpd), GFP_KERNEL); 305 dpd = kmalloc(sizeof(*dpd), GFP_KERNEL);
297 if (dpd == NULL) 306 if (dpd == NULL)
@@ -300,7 +309,7 @@ dfs_pattern_detector_init(struct ath_hw *ah, enum nl80211_dfs_regions region)
300 *dpd = default_dpd; 309 *dpd = default_dpd;
301 INIT_LIST_HEAD(&dpd->channel_detectors); 310 INIT_LIST_HEAD(&dpd->channel_detectors);
302 311
303 dpd->ah = ah; 312 dpd->common = common;
304 if (dpd->set_dfs_domain(dpd, region)) 313 if (dpd->set_dfs_domain(dpd, region))
305 return dpd; 314 return dpd;
306 315
diff --git a/drivers/net/wireless/ath/ath9k/dfs_pattern_detector.h b/drivers/net/wireless/ath/dfs_pattern_detector.h
index 90a5abcc4265..dde2652b787c 100644
--- a/drivers/net/wireless/ath/ath9k/dfs_pattern_detector.h
+++ b/drivers/net/wireless/ath/dfs_pattern_detector.h
@@ -22,6 +22,19 @@
22#include <linux/nl80211.h> 22#include <linux/nl80211.h>
23 23
24/** 24/**
25 * struct ath_dfs_pool_stats - DFS Statistics for global pools
26 */
27struct ath_dfs_pool_stats {
28 u32 pool_reference;
29 u32 pulse_allocated;
30 u32 pulse_alloc_error;
31 u32 pulse_used;
32 u32 pseq_allocated;
33 u32 pseq_alloc_error;
34 u32 pseq_used;
35};
36
37/**
25 * struct pulse_event - describing pulses reported by PHY 38 * struct pulse_event - describing pulses reported by PHY
26 * @ts: pulse time stamp in us 39 * @ts: pulse time stamp in us
27 * @freq: channel frequency in MHz 40 * @freq: channel frequency in MHz
@@ -77,11 +90,12 @@ struct dfs_pattern_detector {
77 bool (*add_pulse)(struct dfs_pattern_detector *dpd, 90 bool (*add_pulse)(struct dfs_pattern_detector *dpd,
78 struct pulse_event *pe); 91 struct pulse_event *pe);
79 92
93 struct ath_dfs_pool_stats (*get_stats)(struct dfs_pattern_detector *dpd);
80 enum nl80211_dfs_regions region; 94 enum nl80211_dfs_regions region;
81 u8 num_radar_types; 95 u8 num_radar_types;
82 u64 last_pulse_ts; 96 u64 last_pulse_ts;
83 /* needed for ath_dbg() */ 97 /* needed for ath_dbg() */
84 struct ath_hw *ah; 98 struct ath_common *common;
85 99
86 const struct radar_detector_specs *radar_spec; 100 const struct radar_detector_specs *radar_spec;
87 struct list_head channel_detectors; 101 struct list_head channel_detectors;
@@ -92,15 +106,7 @@ struct dfs_pattern_detector {
92 * @param region: DFS domain to be used, can be NL80211_DFS_UNSET at creation 106 * @param region: DFS domain to be used, can be NL80211_DFS_UNSET at creation
93 * @return instance pointer on success, NULL otherwise 107 * @return instance pointer on success, NULL otherwise
94 */ 108 */
95#if defined(CONFIG_ATH9K_DFS_CERTIFIED)
96extern struct dfs_pattern_detector * 109extern struct dfs_pattern_detector *
97dfs_pattern_detector_init(struct ath_hw *ah, enum nl80211_dfs_regions region); 110dfs_pattern_detector_init(struct ath_common *common,
98#else 111 enum nl80211_dfs_regions region);
99static inline struct dfs_pattern_detector *
100dfs_pattern_detector_init(struct ath_hw *ah, enum nl80211_dfs_regions region)
101{
102 return NULL;
103}
104#endif /* CONFIG_ATH9K_DFS_CERTIFIED */
105
106#endif /* DFS_PATTERN_DETECTOR_H */ 112#endif /* DFS_PATTERN_DETECTOR_H */
diff --git a/drivers/net/wireless/ath/ath9k/dfs_pri_detector.c b/drivers/net/wireless/ath/dfs_pri_detector.c
index 5ba4b6fe37c0..43b608178884 100644
--- a/drivers/net/wireless/ath/ath9k/dfs_pri_detector.c
+++ b/drivers/net/wireless/ath/dfs_pri_detector.c
@@ -17,10 +17,14 @@
17#include <linux/slab.h> 17#include <linux/slab.h>
18#include <linux/spinlock.h> 18#include <linux/spinlock.h>
19 19
20#include "ath9k.h" 20#include "ath.h"
21#include "dfs_pattern_detector.h" 21#include "dfs_pattern_detector.h"
22#include "dfs_pri_detector.h" 22#include "dfs_pri_detector.h"
23#include "dfs_debug.h" 23
24struct ath_dfs_pool_stats global_dfs_pool_stats = {};
25
26#define DFS_POOL_STAT_INC(c) (global_dfs_pool_stats.c++)
27#define DFS_POOL_STAT_DEC(c) (global_dfs_pool_stats.c--)
24 28
25/** 29/**
26 * struct pulse_elem - elements in pulse queue 30 * struct pulse_elem - elements in pulse queue
@@ -392,7 +396,7 @@ static struct pri_sequence *pri_detector_add_pulse(struct pri_detector *de,
392 396
393 if (!pseq_handler_create_sequences(de, ts, max_updated_seq)) { 397 if (!pseq_handler_create_sequences(de, ts, max_updated_seq)) {
394 pri_detector_reset(de, ts); 398 pri_detector_reset(de, ts);
395 return false; 399 return NULL;
396 } 400 }
397 401
398 ps = pseq_handler_check_detection(de); 402 ps = pseq_handler_check_detection(de);
diff --git a/drivers/net/wireless/ath/ath9k/dfs_pri_detector.h b/drivers/net/wireless/ath/dfs_pri_detector.h
index 723962d1abc6..79f0fff4d1e6 100644
--- a/drivers/net/wireless/ath/ath9k/dfs_pri_detector.h
+++ b/drivers/net/wireless/ath/dfs_pri_detector.h
@@ -19,6 +19,8 @@
19 19
20#include <linux/list.h> 20#include <linux/list.h>
21 21
22extern struct ath_dfs_pool_stats global_dfs_pool_stats;
23
22/** 24/**
23 * struct pri_sequence - sequence of pulses matching one PRI 25 * struct pri_sequence - sequence of pulses matching one PRI
24 * @head: list_head 26 * @head: list_head
diff --git a/drivers/net/wireless/ath/regd.c b/drivers/net/wireless/ath/regd.c
index 7d077c752dd5..c00687e05688 100644
--- a/drivers/net/wireless/ath/regd.c
+++ b/drivers/net/wireless/ath/regd.c
@@ -356,14 +356,131 @@ static u16 ath_regd_find_country_by_name(char *alpha2)
356 return -1; 356 return -1;
357} 357}
358 358
359static int __ath_reg_dyn_country(struct wiphy *wiphy,
360 struct ath_regulatory *reg,
361 struct regulatory_request *request)
362{
363 u16 country_code;
364
365 if (!ath_is_world_regd(reg))
366 return -EINVAL;
367
368 country_code = ath_regd_find_country_by_name(request->alpha2);
369 if (country_code == (u16) -1)
370 return -EINVAL;
371
372 reg->current_rd = COUNTRY_ERD_FLAG;
373 reg->current_rd |= country_code;
374
375 __ath_regd_init(reg);
376
377 ath_reg_apply_world_flags(wiphy, request->initiator, reg);
378
379 return 0;
380}
381
382static void ath_reg_dyn_country(struct wiphy *wiphy,
383 struct ath_regulatory *reg,
384 struct regulatory_request *request)
385{
386 if (__ath_reg_dyn_country(wiphy, reg, request))
387 return;
388
389 printk(KERN_DEBUG "ath: regdomain 0x%0x "
390 "dynamically updated by %s\n",
391 reg->current_rd,
392 reg_initiator_name(request->initiator));
393}
394
395static bool dynamic_country_user_possible(struct ath_regulatory *reg)
396{
397 if (config_enabled(CONFIG_ATH_REG_DYNAMIC_USER_CERT_TESTING))
398 return true;
399
400 switch (reg->country_code) {
401 case CTRY_UNITED_STATES:
402 case CTRY_JAPAN1:
403 case CTRY_JAPAN2:
404 case CTRY_JAPAN3:
405 case CTRY_JAPAN4:
406 case CTRY_JAPAN5:
407 case CTRY_JAPAN6:
408 case CTRY_JAPAN7:
409 case CTRY_JAPAN8:
410 case CTRY_JAPAN9:
411 case CTRY_JAPAN10:
412 case CTRY_JAPAN11:
413 case CTRY_JAPAN12:
414 case CTRY_JAPAN13:
415 case CTRY_JAPAN14:
416 case CTRY_JAPAN15:
417 case CTRY_JAPAN16:
418 case CTRY_JAPAN17:
419 case CTRY_JAPAN18:
420 case CTRY_JAPAN19:
421 case CTRY_JAPAN20:
422 case CTRY_JAPAN21:
423 case CTRY_JAPAN22:
424 case CTRY_JAPAN23:
425 case CTRY_JAPAN24:
426 case CTRY_JAPAN25:
427 case CTRY_JAPAN26:
428 case CTRY_JAPAN27:
429 case CTRY_JAPAN28:
430 case CTRY_JAPAN29:
431 case CTRY_JAPAN30:
432 case CTRY_JAPAN31:
433 case CTRY_JAPAN32:
434 case CTRY_JAPAN33:
435 case CTRY_JAPAN34:
436 case CTRY_JAPAN35:
437 case CTRY_JAPAN36:
438 case CTRY_JAPAN37:
439 case CTRY_JAPAN38:
440 case CTRY_JAPAN39:
441 case CTRY_JAPAN40:
442 case CTRY_JAPAN41:
443 case CTRY_JAPAN42:
444 case CTRY_JAPAN43:
445 case CTRY_JAPAN44:
446 case CTRY_JAPAN45:
447 case CTRY_JAPAN46:
448 case CTRY_JAPAN47:
449 case CTRY_JAPAN48:
450 case CTRY_JAPAN49:
451 case CTRY_JAPAN50:
452 case CTRY_JAPAN51:
453 case CTRY_JAPAN52:
454 case CTRY_JAPAN53:
455 case CTRY_JAPAN54:
456 case CTRY_JAPAN55:
457 case CTRY_JAPAN56:
458 case CTRY_JAPAN57:
459 case CTRY_JAPAN58:
460 case CTRY_JAPAN59:
461 return false;
462 }
463
464 return true;
465}
466
467static void ath_reg_dyn_country_user(struct wiphy *wiphy,
468 struct ath_regulatory *reg,
469 struct regulatory_request *request)
470{
471 if (!config_enabled(CONFIG_ATH_REG_DYNAMIC_USER_REG_HINTS))
472 return;
473 if (!dynamic_country_user_possible(reg))
474 return;
475 ath_reg_dyn_country(wiphy, reg, request);
476}
477
359void ath_reg_notifier_apply(struct wiphy *wiphy, 478void ath_reg_notifier_apply(struct wiphy *wiphy,
360 struct regulatory_request *request, 479 struct regulatory_request *request,
361 struct ath_regulatory *reg) 480 struct ath_regulatory *reg)
362{ 481{
363 struct ath_common *common = container_of(reg, struct ath_common, 482 struct ath_common *common = container_of(reg, struct ath_common,
364 regulatory); 483 regulatory);
365 u16 country_code;
366
367 /* We always apply this */ 484 /* We always apply this */
368 ath_reg_apply_radar_flags(wiphy); 485 ath_reg_apply_radar_flags(wiphy);
369 486
@@ -388,25 +505,12 @@ void ath_reg_notifier_apply(struct wiphy *wiphy,
388 sizeof(struct ath_regulatory)); 505 sizeof(struct ath_regulatory));
389 break; 506 break;
390 case NL80211_REGDOM_SET_BY_DRIVER: 507 case NL80211_REGDOM_SET_BY_DRIVER:
508 break;
391 case NL80211_REGDOM_SET_BY_USER: 509 case NL80211_REGDOM_SET_BY_USER:
510 ath_reg_dyn_country_user(wiphy, reg, request);
392 break; 511 break;
393 case NL80211_REGDOM_SET_BY_COUNTRY_IE: 512 case NL80211_REGDOM_SET_BY_COUNTRY_IE:
394 if (!ath_is_world_regd(reg)) 513 ath_reg_dyn_country(wiphy, reg, request);
395 break;
396
397 country_code = ath_regd_find_country_by_name(request->alpha2);
398 if (country_code == (u16) -1)
399 break;
400
401 reg->current_rd = COUNTRY_ERD_FLAG;
402 reg->current_rd |= country_code;
403
404 printk(KERN_DEBUG "ath: regdomain 0x%0x updated by CountryIE\n",
405 reg->current_rd);
406 __ath_regd_init(reg);
407
408 ath_reg_apply_world_flags(wiphy, request->initiator, reg);
409
410 break; 514 break;
411 } 515 }
412} 516}
diff --git a/drivers/net/wireless/ath/wcn36xx/Kconfig b/drivers/net/wireless/ath/wcn36xx/Kconfig
new file mode 100644
index 000000000000..591ebaea8265
--- /dev/null
+++ b/drivers/net/wireless/ath/wcn36xx/Kconfig
@@ -0,0 +1,16 @@
1config WCN36XX
2 tristate "Qualcomm Atheros WCN3660/3680 support"
3 depends on MAC80211 && HAS_DMA
4 ---help---
5 This module adds support for wireless adapters based on
6 Qualcomm Atheros WCN3660 and WCN3680 mobile chipsets.
7
8 If you choose to build a module, it'll be called wcn36xx.
9
10config WCN36XX_DEBUGFS
11 bool "WCN36XX debugfs support"
12 depends on WCN36XX
13 ---help---
14 Enabled debugfs support
15
16 If unsure, say Y to make it easier to debug problems.
diff --git a/drivers/net/wireless/ath/wcn36xx/Makefile b/drivers/net/wireless/ath/wcn36xx/Makefile
new file mode 100644
index 000000000000..50c43b4382ba
--- /dev/null
+++ b/drivers/net/wireless/ath/wcn36xx/Makefile
@@ -0,0 +1,7 @@
1obj-$(CONFIG_WCN36XX) := wcn36xx.o
2wcn36xx-y += main.o \
3 dxe.o \
4 txrx.o \
5 smd.o \
6 pmc.o \
7 debug.o
diff --git a/drivers/net/wireless/ath/wcn36xx/debug.c b/drivers/net/wireless/ath/wcn36xx/debug.c
new file mode 100644
index 000000000000..5b84f7ae0b1e
--- /dev/null
+++ b/drivers/net/wireless/ath/wcn36xx/debug.c
@@ -0,0 +1,181 @@
1/*
2 * Copyright (c) 2013 Eugene Krasnikov <k.eugene.e@gmail.com>
3 *
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
7 *
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
11 * SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION
13 * OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN
14 * CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
15 */
16
17#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
18
19#include <linux/debugfs.h>
20#include <linux/uaccess.h>
21#include "wcn36xx.h"
22#include "debug.h"
23#include "pmc.h"
24
25#ifdef CONFIG_WCN36XX_DEBUGFS
26
27static ssize_t read_file_bool_bmps(struct file *file, char __user *user_buf,
28 size_t count, loff_t *ppos)
29{
30 struct wcn36xx *wcn = file->private_data;
31 struct wcn36xx_vif *vif_priv = NULL;
32 struct ieee80211_vif *vif = NULL;
33 char buf[3];
34
35 list_for_each_entry(vif_priv, &wcn->vif_list, list) {
36 vif = container_of((void *)vif_priv,
37 struct ieee80211_vif,
38 drv_priv);
39 if (NL80211_IFTYPE_STATION == vif->type) {
40 if (vif_priv->pw_state == WCN36XX_BMPS)
41 buf[0] = '1';
42 else
43 buf[0] = '0';
44 break;
45 }
46 }
47 buf[1] = '\n';
48 buf[2] = 0x00;
49
50 return simple_read_from_buffer(user_buf, count, ppos, buf, 2);
51}
52
53static ssize_t write_file_bool_bmps(struct file *file,
54 const char __user *user_buf,
55 size_t count, loff_t *ppos)
56{
57 struct wcn36xx *wcn = file->private_data;
58 struct wcn36xx_vif *vif_priv = NULL;
59 struct ieee80211_vif *vif = NULL;
60
61 char buf[32];
62 int buf_size;
63
64 buf_size = min(count, (sizeof(buf)-1));
65 if (copy_from_user(buf, user_buf, buf_size))
66 return -EFAULT;
67
68 switch (buf[0]) {
69 case 'y':
70 case 'Y':
71 case '1':
72 list_for_each_entry(vif_priv, &wcn->vif_list, list) {
73 vif = container_of((void *)vif_priv,
74 struct ieee80211_vif,
75 drv_priv);
76 if (NL80211_IFTYPE_STATION == vif->type) {
77 wcn36xx_enable_keep_alive_null_packet(wcn, vif);
78 wcn36xx_pmc_enter_bmps_state(wcn, vif);
79 }
80 }
81 break;
82 case 'n':
83 case 'N':
84 case '0':
85 list_for_each_entry(vif_priv, &wcn->vif_list, list) {
86 vif = container_of((void *)vif_priv,
87 struct ieee80211_vif,
88 drv_priv);
89 if (NL80211_IFTYPE_STATION == vif->type)
90 wcn36xx_pmc_exit_bmps_state(wcn, vif);
91 }
92 break;
93 }
94
95 return count;
96}
97
98static const struct file_operations fops_wcn36xx_bmps = {
99 .open = simple_open,
100 .read = read_file_bool_bmps,
101 .write = write_file_bool_bmps,
102};
103
104static ssize_t write_file_dump(struct file *file,
105 const char __user *user_buf,
106 size_t count, loff_t *ppos)
107{
108 struct wcn36xx *wcn = file->private_data;
109 char buf[255], *tmp;
110 int buf_size;
111 u32 arg[WCN36xx_MAX_DUMP_ARGS];
112 int i;
113
114 memset(buf, 0, sizeof(buf));
115 memset(arg, 0, sizeof(arg));
116
117 buf_size = min(count, (sizeof(buf) - 1));
118 if (copy_from_user(buf, user_buf, buf_size))
119 return -EFAULT;
120
121 tmp = buf;
122
123 for (i = 0; i < WCN36xx_MAX_DUMP_ARGS; i++) {
124 char *begin;
125 begin = strsep(&tmp, " ");
126 if (begin == NULL)
127 break;
128
129 if (kstrtoul(begin, 0, (unsigned long *)(arg + i)) != 0)
130 break;
131 }
132
133 wcn36xx_info("DUMP args is %d %d %d %d %d\n", arg[0], arg[1], arg[2],
134 arg[3], arg[4]);
135 wcn36xx_smd_dump_cmd_req(wcn, arg[0], arg[1], arg[2], arg[3], arg[4]);
136
137 return count;
138}
139
140static const struct file_operations fops_wcn36xx_dump = {
141 .open = simple_open,
142 .write = write_file_dump,
143};
144
145#define ADD_FILE(name, mode, fop, priv_data) \
146 do { \
147 struct dentry *d; \
148 d = debugfs_create_file(__stringify(name), \
149 mode, dfs->rootdir, \
150 priv_data, fop); \
151 dfs->file_##name.dentry = d; \
152 if (IS_ERR(d)) { \
153 wcn36xx_warn("Create the debugfs entry failed");\
154 dfs->file_##name.dentry = NULL; \
155 } \
156 } while (0)
157
158
159void wcn36xx_debugfs_init(struct wcn36xx *wcn)
160{
161 struct wcn36xx_dfs_entry *dfs = &wcn->dfs;
162
163 dfs->rootdir = debugfs_create_dir(KBUILD_MODNAME,
164 wcn->hw->wiphy->debugfsdir);
165 if (IS_ERR(dfs->rootdir)) {
166 wcn36xx_warn("Create the debugfs failed\n");
167 dfs->rootdir = NULL;
168 }
169
170 ADD_FILE(bmps_switcher, S_IRUSR | S_IWUSR,
171 &fops_wcn36xx_bmps, wcn);
172 ADD_FILE(dump, S_IWUSR, &fops_wcn36xx_dump, wcn);
173}
174
175void wcn36xx_debugfs_exit(struct wcn36xx *wcn)
176{
177 struct wcn36xx_dfs_entry *dfs = &wcn->dfs;
178 debugfs_remove_recursive(dfs->rootdir);
179}
180
181#endif /* CONFIG_WCN36XX_DEBUGFS */
diff --git a/drivers/net/wireless/ath/wcn36xx/debug.h b/drivers/net/wireless/ath/wcn36xx/debug.h
new file mode 100644
index 000000000000..46307aa562d3
--- /dev/null
+++ b/drivers/net/wireless/ath/wcn36xx/debug.h
@@ -0,0 +1,49 @@
1/*
2 * Copyright (c) 2013 Eugene Krasnikov <k.eugene.e@gmail.com>
3 *
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
7 *
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
11 * SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION
13 * OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN
14 * CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
15 */
16
17#ifndef _WCN36XX_DEBUG_H_
18#define _WCN36XX_DEBUG_H_
19
20#include <linux/kernel.h>
21
22#define WCN36xx_MAX_DUMP_ARGS 5
23
24#ifdef CONFIG_WCN36XX_DEBUGFS
25struct wcn36xx_dfs_file {
26 struct dentry *dentry;
27 u32 value;
28};
29
30struct wcn36xx_dfs_entry {
31 struct dentry *rootdir;
32 struct wcn36xx_dfs_file file_bmps_switcher;
33 struct wcn36xx_dfs_file file_dump;
34};
35
36void wcn36xx_debugfs_init(struct wcn36xx *wcn);
37void wcn36xx_debugfs_exit(struct wcn36xx *wcn);
38
39#else
40static inline void wcn36xx_debugfs_init(struct wcn36xx *wcn)
41{
42}
43static inline void wcn36xx_debugfs_exit(struct wcn36xx *wcn)
44{
45}
46
47#endif /* CONFIG_WCN36XX_DEBUGFS */
48
49#endif /* _WCN36XX_DEBUG_H_ */
diff --git a/drivers/net/wireless/ath/wcn36xx/dxe.c b/drivers/net/wireless/ath/wcn36xx/dxe.c
new file mode 100644
index 000000000000..ee25786b4447
--- /dev/null
+++ b/drivers/net/wireless/ath/wcn36xx/dxe.c
@@ -0,0 +1,805 @@
1/*
2 * Copyright (c) 2013 Eugene Krasnikov <k.eugene.e@gmail.com>
3 *
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
7 *
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
11 * SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION
13 * OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN
14 * CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
15 */
16
17/* DXE - DMA transfer engine
18 * we have 2 channels(High prio and Low prio) for TX and 2 channels for RX.
19 * through low channels data packets are transfered
20 * through high channels managment packets are transfered
21 */
22
23#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
24
25#include <linux/interrupt.h>
26#include "wcn36xx.h"
27#include "txrx.h"
28
29void *wcn36xx_dxe_get_next_bd(struct wcn36xx *wcn, bool is_low)
30{
31 struct wcn36xx_dxe_ch *ch = is_low ?
32 &wcn->dxe_tx_l_ch :
33 &wcn->dxe_tx_h_ch;
34
35 return ch->head_blk_ctl->bd_cpu_addr;
36}
37
38static void wcn36xx_dxe_write_register(struct wcn36xx *wcn, int addr, int data)
39{
40 wcn36xx_dbg(WCN36XX_DBG_DXE,
41 "wcn36xx_dxe_write_register: addr=%x, data=%x\n",
42 addr, data);
43
44 writel(data, wcn->mmio + addr);
45}
46
47static void wcn36xx_dxe_read_register(struct wcn36xx *wcn, int addr, int *data)
48{
49 *data = readl(wcn->mmio + addr);
50
51 wcn36xx_dbg(WCN36XX_DBG_DXE,
52 "wcn36xx_dxe_read_register: addr=%x, data=%x\n",
53 addr, *data);
54}
55
56static void wcn36xx_dxe_free_ctl_block(struct wcn36xx_dxe_ch *ch)
57{
58 struct wcn36xx_dxe_ctl *ctl = ch->head_blk_ctl, *next;
59 int i;
60
61 for (i = 0; i < ch->desc_num && ctl; i++) {
62 next = ctl->next;
63 kfree(ctl);
64 ctl = next;
65 }
66}
67
68static int wcn36xx_dxe_allocate_ctl_block(struct wcn36xx_dxe_ch *ch)
69{
70 struct wcn36xx_dxe_ctl *prev_ctl = NULL;
71 struct wcn36xx_dxe_ctl *cur_ctl = NULL;
72 int i;
73
74 for (i = 0; i < ch->desc_num; i++) {
75 cur_ctl = kzalloc(sizeof(*cur_ctl), GFP_KERNEL);
76 if (!cur_ctl)
77 goto out_fail;
78
79 cur_ctl->ctl_blk_order = i;
80 if (i == 0) {
81 ch->head_blk_ctl = cur_ctl;
82 ch->tail_blk_ctl = cur_ctl;
83 } else if (ch->desc_num - 1 == i) {
84 prev_ctl->next = cur_ctl;
85 cur_ctl->next = ch->head_blk_ctl;
86 } else {
87 prev_ctl->next = cur_ctl;
88 }
89 prev_ctl = cur_ctl;
90 }
91
92 return 0;
93
94out_fail:
95 wcn36xx_dxe_free_ctl_block(ch);
96 return -ENOMEM;
97}
98
99int wcn36xx_dxe_alloc_ctl_blks(struct wcn36xx *wcn)
100{
101 int ret;
102
103 wcn->dxe_tx_l_ch.ch_type = WCN36XX_DXE_CH_TX_L;
104 wcn->dxe_tx_h_ch.ch_type = WCN36XX_DXE_CH_TX_H;
105 wcn->dxe_rx_l_ch.ch_type = WCN36XX_DXE_CH_RX_L;
106 wcn->dxe_rx_h_ch.ch_type = WCN36XX_DXE_CH_RX_H;
107
108 wcn->dxe_tx_l_ch.desc_num = WCN36XX_DXE_CH_DESC_NUMB_TX_L;
109 wcn->dxe_tx_h_ch.desc_num = WCN36XX_DXE_CH_DESC_NUMB_TX_H;
110 wcn->dxe_rx_l_ch.desc_num = WCN36XX_DXE_CH_DESC_NUMB_RX_L;
111 wcn->dxe_rx_h_ch.desc_num = WCN36XX_DXE_CH_DESC_NUMB_RX_H;
112
113 wcn->dxe_tx_l_ch.dxe_wq = WCN36XX_DXE_WQ_TX_L;
114 wcn->dxe_tx_h_ch.dxe_wq = WCN36XX_DXE_WQ_TX_H;
115
116 wcn->dxe_tx_l_ch.ctrl_bd = WCN36XX_DXE_CTRL_TX_L_BD;
117 wcn->dxe_tx_h_ch.ctrl_bd = WCN36XX_DXE_CTRL_TX_H_BD;
118
119 wcn->dxe_tx_l_ch.ctrl_skb = WCN36XX_DXE_CTRL_TX_L_SKB;
120 wcn->dxe_tx_h_ch.ctrl_skb = WCN36XX_DXE_CTRL_TX_H_SKB;
121
122 wcn->dxe_tx_l_ch.reg_ctrl = WCN36XX_DXE_REG_CTL_TX_L;
123 wcn->dxe_tx_h_ch.reg_ctrl = WCN36XX_DXE_REG_CTL_TX_H;
124
125 wcn->dxe_tx_l_ch.def_ctrl = WCN36XX_DXE_CH_DEFAULT_CTL_TX_L;
126 wcn->dxe_tx_h_ch.def_ctrl = WCN36XX_DXE_CH_DEFAULT_CTL_TX_H;
127
128 /* DXE control block allocation */
129 ret = wcn36xx_dxe_allocate_ctl_block(&wcn->dxe_tx_l_ch);
130 if (ret)
131 goto out_err;
132 ret = wcn36xx_dxe_allocate_ctl_block(&wcn->dxe_tx_h_ch);
133 if (ret)
134 goto out_err;
135 ret = wcn36xx_dxe_allocate_ctl_block(&wcn->dxe_rx_l_ch);
136 if (ret)
137 goto out_err;
138 ret = wcn36xx_dxe_allocate_ctl_block(&wcn->dxe_rx_h_ch);
139 if (ret)
140 goto out_err;
141
142 /* Initialize SMSM state Clear TX Enable RING EMPTY STATE */
143 ret = wcn->ctrl_ops->smsm_change_state(
144 WCN36XX_SMSM_WLAN_TX_ENABLE,
145 WCN36XX_SMSM_WLAN_TX_RINGS_EMPTY);
146
147 return 0;
148
149out_err:
150 wcn36xx_err("Failed to allocate DXE control blocks\n");
151 wcn36xx_dxe_free_ctl_blks(wcn);
152 return -ENOMEM;
153}
154
155void wcn36xx_dxe_free_ctl_blks(struct wcn36xx *wcn)
156{
157 wcn36xx_dxe_free_ctl_block(&wcn->dxe_tx_l_ch);
158 wcn36xx_dxe_free_ctl_block(&wcn->dxe_tx_h_ch);
159 wcn36xx_dxe_free_ctl_block(&wcn->dxe_rx_l_ch);
160 wcn36xx_dxe_free_ctl_block(&wcn->dxe_rx_h_ch);
161}
162
163static int wcn36xx_dxe_init_descs(struct wcn36xx_dxe_ch *wcn_ch)
164{
165 struct wcn36xx_dxe_desc *cur_dxe = NULL;
166 struct wcn36xx_dxe_desc *prev_dxe = NULL;
167 struct wcn36xx_dxe_ctl *cur_ctl = NULL;
168 size_t size;
169 int i;
170
171 size = wcn_ch->desc_num * sizeof(struct wcn36xx_dxe_desc);
172 wcn_ch->cpu_addr = dma_alloc_coherent(NULL, size, &wcn_ch->dma_addr,
173 GFP_KERNEL);
174 if (!wcn_ch->cpu_addr)
175 return -ENOMEM;
176
177 memset(wcn_ch->cpu_addr, 0, size);
178
179 cur_dxe = (struct wcn36xx_dxe_desc *)wcn_ch->cpu_addr;
180 cur_ctl = wcn_ch->head_blk_ctl;
181
182 for (i = 0; i < wcn_ch->desc_num; i++) {
183 cur_ctl->desc = cur_dxe;
184 cur_ctl->desc_phy_addr = wcn_ch->dma_addr +
185 i * sizeof(struct wcn36xx_dxe_desc);
186
187 switch (wcn_ch->ch_type) {
188 case WCN36XX_DXE_CH_TX_L:
189 cur_dxe->ctrl = WCN36XX_DXE_CTRL_TX_L;
190 cur_dxe->dst_addr_l = WCN36XX_DXE_WQ_TX_L;
191 break;
192 case WCN36XX_DXE_CH_TX_H:
193 cur_dxe->ctrl = WCN36XX_DXE_CTRL_TX_H;
194 cur_dxe->dst_addr_l = WCN36XX_DXE_WQ_TX_H;
195 break;
196 case WCN36XX_DXE_CH_RX_L:
197 cur_dxe->ctrl = WCN36XX_DXE_CTRL_RX_L;
198 cur_dxe->src_addr_l = WCN36XX_DXE_WQ_RX_L;
199 break;
200 case WCN36XX_DXE_CH_RX_H:
201 cur_dxe->ctrl = WCN36XX_DXE_CTRL_RX_H;
202 cur_dxe->src_addr_l = WCN36XX_DXE_WQ_RX_H;
203 break;
204 }
205 if (0 == i) {
206 cur_dxe->phy_next_l = 0;
207 } else if ((0 < i) && (i < wcn_ch->desc_num - 1)) {
208 prev_dxe->phy_next_l =
209 cur_ctl->desc_phy_addr;
210 } else if (i == (wcn_ch->desc_num - 1)) {
211 prev_dxe->phy_next_l =
212 cur_ctl->desc_phy_addr;
213 cur_dxe->phy_next_l =
214 wcn_ch->head_blk_ctl->desc_phy_addr;
215 }
216 cur_ctl = cur_ctl->next;
217 prev_dxe = cur_dxe;
218 cur_dxe++;
219 }
220
221 return 0;
222}
223
224static void wcn36xx_dxe_init_tx_bd(struct wcn36xx_dxe_ch *ch,
225 struct wcn36xx_dxe_mem_pool *pool)
226{
227 int i, chunk_size = pool->chunk_size;
228 dma_addr_t bd_phy_addr = pool->phy_addr;
229 void *bd_cpu_addr = pool->virt_addr;
230 struct wcn36xx_dxe_ctl *cur = ch->head_blk_ctl;
231
232 for (i = 0; i < ch->desc_num; i++) {
233 /* Only every second dxe needs a bd pointer,
234 the other will point to the skb data */
235 if (!(i & 1)) {
236 cur->bd_phy_addr = bd_phy_addr;
237 cur->bd_cpu_addr = bd_cpu_addr;
238 bd_phy_addr += chunk_size;
239 bd_cpu_addr += chunk_size;
240 } else {
241 cur->bd_phy_addr = 0;
242 cur->bd_cpu_addr = NULL;
243 }
244 cur = cur->next;
245 }
246}
247
248static int wcn36xx_dxe_enable_ch_int(struct wcn36xx *wcn, u16 wcn_ch)
249{
250 int reg_data = 0;
251
252 wcn36xx_dxe_read_register(wcn,
253 WCN36XX_DXE_INT_MASK_REG,
254 &reg_data);
255
256 reg_data |= wcn_ch;
257
258 wcn36xx_dxe_write_register(wcn,
259 WCN36XX_DXE_INT_MASK_REG,
260 (int)reg_data);
261 return 0;
262}
263
264static int wcn36xx_dxe_fill_skb(struct wcn36xx_dxe_ctl *ctl)
265{
266 struct wcn36xx_dxe_desc *dxe = ctl->desc;
267 struct sk_buff *skb;
268
269 skb = alloc_skb(WCN36XX_PKT_SIZE, GFP_ATOMIC);
270 if (skb == NULL)
271 return -ENOMEM;
272
273 dxe->dst_addr_l = dma_map_single(NULL,
274 skb_tail_pointer(skb),
275 WCN36XX_PKT_SIZE,
276 DMA_FROM_DEVICE);
277 ctl->skb = skb;
278
279 return 0;
280}
281
282static int wcn36xx_dxe_ch_alloc_skb(struct wcn36xx *wcn,
283 struct wcn36xx_dxe_ch *wcn_ch)
284{
285 int i;
286 struct wcn36xx_dxe_ctl *cur_ctl = NULL;
287
288 cur_ctl = wcn_ch->head_blk_ctl;
289
290 for (i = 0; i < wcn_ch->desc_num; i++) {
291 wcn36xx_dxe_fill_skb(cur_ctl);
292 cur_ctl = cur_ctl->next;
293 }
294
295 return 0;
296}
297
298static void wcn36xx_dxe_ch_free_skbs(struct wcn36xx *wcn,
299 struct wcn36xx_dxe_ch *wcn_ch)
300{
301 struct wcn36xx_dxe_ctl *cur = wcn_ch->head_blk_ctl;
302 int i;
303
304 for (i = 0; i < wcn_ch->desc_num; i++) {
305 kfree_skb(cur->skb);
306 cur = cur->next;
307 }
308}
309
310void wcn36xx_dxe_tx_ack_ind(struct wcn36xx *wcn, u32 status)
311{
312 struct ieee80211_tx_info *info;
313 struct sk_buff *skb;
314 unsigned long flags;
315
316 spin_lock_irqsave(&wcn->dxe_lock, flags);
317 skb = wcn->tx_ack_skb;
318 wcn->tx_ack_skb = NULL;
319 spin_unlock_irqrestore(&wcn->dxe_lock, flags);
320
321 if (!skb) {
322 wcn36xx_warn("Spurious TX complete indication\n");
323 return;
324 }
325
326 info = IEEE80211_SKB_CB(skb);
327
328 if (status == 1)
329 info->flags |= IEEE80211_TX_STAT_ACK;
330
331 wcn36xx_dbg(WCN36XX_DBG_DXE, "dxe tx ack status: %d\n", status);
332
333 ieee80211_tx_status_irqsafe(wcn->hw, skb);
334 ieee80211_wake_queues(wcn->hw);
335}
336
337static void reap_tx_dxes(struct wcn36xx *wcn, struct wcn36xx_dxe_ch *ch)
338{
339 struct wcn36xx_dxe_ctl *ctl = ch->tail_blk_ctl;
340 struct ieee80211_tx_info *info;
341 unsigned long flags;
342
343 /*
344 * Make at least one loop of do-while because in case ring is
345 * completely full head and tail are pointing to the same element
346 * and while-do will not make any cycles.
347 */
348 do {
349 if (ctl->skb) {
350 dma_unmap_single(NULL, ctl->desc->src_addr_l,
351 ctl->skb->len, DMA_TO_DEVICE);
352 info = IEEE80211_SKB_CB(ctl->skb);
353 if (!(info->flags & IEEE80211_TX_CTL_REQ_TX_STATUS)) {
354 /* Keep frame until TX status comes */
355 ieee80211_free_txskb(wcn->hw, ctl->skb);
356 }
357 spin_lock_irqsave(&ctl->skb_lock, flags);
358 if (wcn->queues_stopped) {
359 wcn->queues_stopped = false;
360 ieee80211_wake_queues(wcn->hw);
361 }
362 spin_unlock_irqrestore(&ctl->skb_lock, flags);
363
364 ctl->skb = NULL;
365 }
366 ctl = ctl->next;
367 } while (ctl != ch->head_blk_ctl &&
368 !(ctl->desc->ctrl & WCN36XX_DXE_CTRL_VALID_MASK));
369
370 ch->tail_blk_ctl = ctl;
371}
372
373static irqreturn_t wcn36xx_irq_tx_complete(int irq, void *dev)
374{
375 struct wcn36xx *wcn = (struct wcn36xx *)dev;
376 int int_src, int_reason;
377
378 wcn36xx_dxe_read_register(wcn, WCN36XX_DXE_INT_SRC_RAW_REG, &int_src);
379
380 if (int_src & WCN36XX_INT_MASK_CHAN_TX_H) {
381 wcn36xx_dxe_read_register(wcn,
382 WCN36XX_DXE_CH_STATUS_REG_ADDR_TX_H,
383 &int_reason);
384
385 /* TODO: Check int_reason */
386
387 wcn36xx_dxe_write_register(wcn,
388 WCN36XX_DXE_0_INT_CLR,
389 WCN36XX_INT_MASK_CHAN_TX_H);
390
391 wcn36xx_dxe_write_register(wcn, WCN36XX_DXE_0_INT_ED_CLR,
392 WCN36XX_INT_MASK_CHAN_TX_H);
393 wcn36xx_dbg(WCN36XX_DBG_DXE, "dxe tx ready high\n");
394 reap_tx_dxes(wcn, &wcn->dxe_tx_h_ch);
395 }
396
397 if (int_src & WCN36XX_INT_MASK_CHAN_TX_L) {
398 wcn36xx_dxe_read_register(wcn,
399 WCN36XX_DXE_CH_STATUS_REG_ADDR_TX_L,
400 &int_reason);
401 /* TODO: Check int_reason */
402
403 wcn36xx_dxe_write_register(wcn,
404 WCN36XX_DXE_0_INT_CLR,
405 WCN36XX_INT_MASK_CHAN_TX_L);
406
407 wcn36xx_dxe_write_register(wcn, WCN36XX_DXE_0_INT_ED_CLR,
408 WCN36XX_INT_MASK_CHAN_TX_L);
409 wcn36xx_dbg(WCN36XX_DBG_DXE, "dxe tx ready low\n");
410 reap_tx_dxes(wcn, &wcn->dxe_tx_l_ch);
411 }
412
413 return IRQ_HANDLED;
414}
415
416static irqreturn_t wcn36xx_irq_rx_ready(int irq, void *dev)
417{
418 struct wcn36xx *wcn = (struct wcn36xx *)dev;
419
420 disable_irq_nosync(wcn->rx_irq);
421 wcn36xx_dxe_rx_frame(wcn);
422 enable_irq(wcn->rx_irq);
423 return IRQ_HANDLED;
424}
425
426static int wcn36xx_dxe_request_irqs(struct wcn36xx *wcn)
427{
428 int ret;
429
430 ret = request_irq(wcn->tx_irq, wcn36xx_irq_tx_complete,
431 IRQF_TRIGGER_HIGH, "wcn36xx_tx", wcn);
432 if (ret) {
433 wcn36xx_err("failed to alloc tx irq\n");
434 goto out_err;
435 }
436
437 ret = request_irq(wcn->rx_irq, wcn36xx_irq_rx_ready, IRQF_TRIGGER_HIGH,
438 "wcn36xx_rx", wcn);
439 if (ret) {
440 wcn36xx_err("failed to alloc rx irq\n");
441 goto out_txirq;
442 }
443
444 enable_irq_wake(wcn->rx_irq);
445
446 return 0;
447
448out_txirq:
449 free_irq(wcn->tx_irq, wcn);
450out_err:
451 return ret;
452
453}
454
455static int wcn36xx_rx_handle_packets(struct wcn36xx *wcn,
456 struct wcn36xx_dxe_ch *ch)
457{
458 struct wcn36xx_dxe_ctl *ctl = ch->head_blk_ctl;
459 struct wcn36xx_dxe_desc *dxe = ctl->desc;
460 dma_addr_t dma_addr;
461 struct sk_buff *skb;
462
463 while (!(dxe->ctrl & WCN36XX_DXE_CTRL_VALID_MASK)) {
464 skb = ctl->skb;
465 dma_addr = dxe->dst_addr_l;
466 wcn36xx_dxe_fill_skb(ctl);
467
468 switch (ch->ch_type) {
469 case WCN36XX_DXE_CH_RX_L:
470 dxe->ctrl = WCN36XX_DXE_CTRL_RX_L;
471 wcn36xx_dxe_write_register(wcn, WCN36XX_DXE_ENCH_ADDR,
472 WCN36XX_DXE_INT_CH1_MASK);
473 break;
474 case WCN36XX_DXE_CH_RX_H:
475 dxe->ctrl = WCN36XX_DXE_CTRL_RX_H;
476 wcn36xx_dxe_write_register(wcn, WCN36XX_DXE_ENCH_ADDR,
477 WCN36XX_DXE_INT_CH3_MASK);
478 break;
479 default:
480 wcn36xx_warn("Unknown channel\n");
481 }
482
483 dma_unmap_single(NULL, dma_addr, WCN36XX_PKT_SIZE,
484 DMA_FROM_DEVICE);
485 wcn36xx_rx_skb(wcn, skb);
486 ctl = ctl->next;
487 dxe = ctl->desc;
488 }
489
490 ch->head_blk_ctl = ctl;
491
492 return 0;
493}
494
495void wcn36xx_dxe_rx_frame(struct wcn36xx *wcn)
496{
497 int int_src;
498
499 wcn36xx_dxe_read_register(wcn, WCN36XX_DXE_INT_SRC_RAW_REG, &int_src);
500
501 /* RX_LOW_PRI */
502 if (int_src & WCN36XX_DXE_INT_CH1_MASK) {
503 wcn36xx_dxe_write_register(wcn, WCN36XX_DXE_0_INT_CLR,
504 WCN36XX_DXE_INT_CH1_MASK);
505 wcn36xx_rx_handle_packets(wcn, &(wcn->dxe_rx_l_ch));
506 }
507
508 /* RX_HIGH_PRI */
509 if (int_src & WCN36XX_DXE_INT_CH3_MASK) {
510 /* Clean up all the INT within this channel */
511 wcn36xx_dxe_write_register(wcn, WCN36XX_DXE_0_INT_CLR,
512 WCN36XX_DXE_INT_CH3_MASK);
513 wcn36xx_rx_handle_packets(wcn, &(wcn->dxe_rx_h_ch));
514 }
515
516 if (!int_src)
517 wcn36xx_warn("No DXE interrupt pending\n");
518}
519
520int wcn36xx_dxe_allocate_mem_pools(struct wcn36xx *wcn)
521{
522 size_t s;
523 void *cpu_addr;
524
525 /* Allocate BD headers for MGMT frames */
526
527 /* Where this come from ask QC */
528 wcn->mgmt_mem_pool.chunk_size = WCN36XX_BD_CHUNK_SIZE +
529 16 - (WCN36XX_BD_CHUNK_SIZE % 8);
530
531 s = wcn->mgmt_mem_pool.chunk_size * WCN36XX_DXE_CH_DESC_NUMB_TX_H;
532 cpu_addr = dma_alloc_coherent(NULL, s, &wcn->mgmt_mem_pool.phy_addr,
533 GFP_KERNEL);
534 if (!cpu_addr)
535 goto out_err;
536
537 wcn->mgmt_mem_pool.virt_addr = cpu_addr;
538 memset(cpu_addr, 0, s);
539
540 /* Allocate BD headers for DATA frames */
541
542 /* Where this come from ask QC */
543 wcn->data_mem_pool.chunk_size = WCN36XX_BD_CHUNK_SIZE +
544 16 - (WCN36XX_BD_CHUNK_SIZE % 8);
545
546 s = wcn->data_mem_pool.chunk_size * WCN36XX_DXE_CH_DESC_NUMB_TX_L;
547 cpu_addr = dma_alloc_coherent(NULL, s, &wcn->data_mem_pool.phy_addr,
548 GFP_KERNEL);
549 if (!cpu_addr)
550 goto out_err;
551
552 wcn->data_mem_pool.virt_addr = cpu_addr;
553 memset(cpu_addr, 0, s);
554
555 return 0;
556
557out_err:
558 wcn36xx_dxe_free_mem_pools(wcn);
559 wcn36xx_err("Failed to allocate BD mempool\n");
560 return -ENOMEM;
561}
562
563void wcn36xx_dxe_free_mem_pools(struct wcn36xx *wcn)
564{
565 if (wcn->mgmt_mem_pool.virt_addr)
566 dma_free_coherent(NULL, wcn->mgmt_mem_pool.chunk_size *
567 WCN36XX_DXE_CH_DESC_NUMB_TX_H,
568 wcn->mgmt_mem_pool.virt_addr,
569 wcn->mgmt_mem_pool.phy_addr);
570
571 if (wcn->data_mem_pool.virt_addr) {
572 dma_free_coherent(NULL, wcn->data_mem_pool.chunk_size *
573 WCN36XX_DXE_CH_DESC_NUMB_TX_L,
574 wcn->data_mem_pool.virt_addr,
575 wcn->data_mem_pool.phy_addr);
576 }
577}
578
579int wcn36xx_dxe_tx_frame(struct wcn36xx *wcn,
580 struct wcn36xx_vif *vif_priv,
581 struct sk_buff *skb,
582 bool is_low)
583{
584 struct wcn36xx_dxe_ctl *ctl = NULL;
585 struct wcn36xx_dxe_desc *desc = NULL;
586 struct wcn36xx_dxe_ch *ch = NULL;
587 unsigned long flags;
588
589 ch = is_low ? &wcn->dxe_tx_l_ch : &wcn->dxe_tx_h_ch;
590
591 ctl = ch->head_blk_ctl;
592
593 spin_lock_irqsave(&ctl->next->skb_lock, flags);
594
595 /*
596 * If skb is not null that means that we reached the tail of the ring
597 * hence ring is full. Stop queues to let mac80211 back off until ring
598 * has an empty slot again.
599 */
600 if (NULL != ctl->next->skb) {
601 ieee80211_stop_queues(wcn->hw);
602 wcn->queues_stopped = true;
603 spin_unlock_irqrestore(&ctl->next->skb_lock, flags);
604 return -EBUSY;
605 }
606 spin_unlock_irqrestore(&ctl->next->skb_lock, flags);
607
608 ctl->skb = NULL;
609 desc = ctl->desc;
610
611 /* Set source address of the BD we send */
612 desc->src_addr_l = ctl->bd_phy_addr;
613
614 desc->dst_addr_l = ch->dxe_wq;
615 desc->fr_len = sizeof(struct wcn36xx_tx_bd);
616 desc->ctrl = ch->ctrl_bd;
617
618 wcn36xx_dbg(WCN36XX_DBG_DXE, "DXE TX\n");
619
620 wcn36xx_dbg_dump(WCN36XX_DBG_DXE_DUMP, "DESC1 >>> ",
621 (char *)desc, sizeof(*desc));
622 wcn36xx_dbg_dump(WCN36XX_DBG_DXE_DUMP,
623 "BD >>> ", (char *)ctl->bd_cpu_addr,
624 sizeof(struct wcn36xx_tx_bd));
625
626 /* Set source address of the SKB we send */
627 ctl = ctl->next;
628 ctl->skb = skb;
629 desc = ctl->desc;
630 if (ctl->bd_cpu_addr) {
631 wcn36xx_err("bd_cpu_addr cannot be NULL for skb DXE\n");
632 return -EINVAL;
633 }
634
635 desc->src_addr_l = dma_map_single(NULL,
636 ctl->skb->data,
637 ctl->skb->len,
638 DMA_TO_DEVICE);
639
640 desc->dst_addr_l = ch->dxe_wq;
641 desc->fr_len = ctl->skb->len;
642
643 /* set dxe descriptor to VALID */
644 desc->ctrl = ch->ctrl_skb;
645
646 wcn36xx_dbg_dump(WCN36XX_DBG_DXE_DUMP, "DESC2 >>> ",
647 (char *)desc, sizeof(*desc));
648 wcn36xx_dbg_dump(WCN36XX_DBG_DXE_DUMP, "SKB >>> ",
649 (char *)ctl->skb->data, ctl->skb->len);
650
651 /* Move the head of the ring to the next empty descriptor */
652 ch->head_blk_ctl = ctl->next;
653
654 /*
655 * When connected and trying to send data frame chip can be in sleep
656 * mode and writing to the register will not wake up the chip. Instead
657 * notify chip about new frame through SMSM bus.
658 */
659 if (is_low && vif_priv->pw_state == WCN36XX_BMPS) {
660 wcn->ctrl_ops->smsm_change_state(
661 0,
662 WCN36XX_SMSM_WLAN_TX_ENABLE);
663 } else {
664 /* indicate End Of Packet and generate interrupt on descriptor
665 * done.
666 */
667 wcn36xx_dxe_write_register(wcn,
668 ch->reg_ctrl, ch->def_ctrl);
669 }
670
671 return 0;
672}
673
674int wcn36xx_dxe_init(struct wcn36xx *wcn)
675{
676 int reg_data = 0, ret;
677
678 reg_data = WCN36XX_DXE_REG_RESET;
679 wcn36xx_dxe_write_register(wcn, WCN36XX_DXE_REG_CSR_RESET, reg_data);
680
681 /* Setting interrupt path */
682 reg_data = WCN36XX_DXE_CCU_INT;
683 wcn36xx_dxe_write_register(wcn, WCN36XX_DXE_REG_CCU_INT, reg_data);
684
685 /***************************************/
686 /* Init descriptors for TX LOW channel */
687 /***************************************/
688 wcn36xx_dxe_init_descs(&wcn->dxe_tx_l_ch);
689 wcn36xx_dxe_init_tx_bd(&wcn->dxe_tx_l_ch, &wcn->data_mem_pool);
690
691 /* Write channel head to a NEXT register */
692 wcn36xx_dxe_write_register(wcn, WCN36XX_DXE_CH_NEXT_DESC_ADDR_TX_L,
693 wcn->dxe_tx_l_ch.head_blk_ctl->desc_phy_addr);
694
695 /* Program DMA destination addr for TX LOW */
696 wcn36xx_dxe_write_register(wcn,
697 WCN36XX_DXE_CH_DEST_ADDR_TX_L,
698 WCN36XX_DXE_WQ_TX_L);
699
700 wcn36xx_dxe_read_register(wcn, WCN36XX_DXE_REG_CH_EN, &reg_data);
701 wcn36xx_dxe_enable_ch_int(wcn, WCN36XX_INT_MASK_CHAN_TX_L);
702
703 /***************************************/
704 /* Init descriptors for TX HIGH channel */
705 /***************************************/
706 wcn36xx_dxe_init_descs(&wcn->dxe_tx_h_ch);
707 wcn36xx_dxe_init_tx_bd(&wcn->dxe_tx_h_ch, &wcn->mgmt_mem_pool);
708
709 /* Write channel head to a NEXT register */
710 wcn36xx_dxe_write_register(wcn, WCN36XX_DXE_CH_NEXT_DESC_ADDR_TX_H,
711 wcn->dxe_tx_h_ch.head_blk_ctl->desc_phy_addr);
712
713 /* Program DMA destination addr for TX HIGH */
714 wcn36xx_dxe_write_register(wcn,
715 WCN36XX_DXE_CH_DEST_ADDR_TX_H,
716 WCN36XX_DXE_WQ_TX_H);
717
718 wcn36xx_dxe_read_register(wcn, WCN36XX_DXE_REG_CH_EN, &reg_data);
719
720 /* Enable channel interrupts */
721 wcn36xx_dxe_enable_ch_int(wcn, WCN36XX_INT_MASK_CHAN_TX_H);
722
723 /***************************************/
724 /* Init descriptors for RX LOW channel */
725 /***************************************/
726 wcn36xx_dxe_init_descs(&wcn->dxe_rx_l_ch);
727
728 /* For RX we need to preallocated buffers */
729 wcn36xx_dxe_ch_alloc_skb(wcn, &wcn->dxe_rx_l_ch);
730
731 /* Write channel head to a NEXT register */
732 wcn36xx_dxe_write_register(wcn, WCN36XX_DXE_CH_NEXT_DESC_ADDR_RX_L,
733 wcn->dxe_rx_l_ch.head_blk_ctl->desc_phy_addr);
734
735 /* Write DMA source address */
736 wcn36xx_dxe_write_register(wcn,
737 WCN36XX_DXE_CH_SRC_ADDR_RX_L,
738 WCN36XX_DXE_WQ_RX_L);
739
740 /* Program preallocated destination address */
741 wcn36xx_dxe_write_register(wcn,
742 WCN36XX_DXE_CH_DEST_ADDR_RX_L,
743 wcn->dxe_rx_l_ch.head_blk_ctl->desc->phy_next_l);
744
745 /* Enable default control registers */
746 wcn36xx_dxe_write_register(wcn,
747 WCN36XX_DXE_REG_CTL_RX_L,
748 WCN36XX_DXE_CH_DEFAULT_CTL_RX_L);
749
750 /* Enable channel interrupts */
751 wcn36xx_dxe_enable_ch_int(wcn, WCN36XX_INT_MASK_CHAN_RX_L);
752
753 /***************************************/
754 /* Init descriptors for RX HIGH channel */
755 /***************************************/
756 wcn36xx_dxe_init_descs(&wcn->dxe_rx_h_ch);
757
758 /* For RX we need to prealocat buffers */
759 wcn36xx_dxe_ch_alloc_skb(wcn, &wcn->dxe_rx_h_ch);
760
761 /* Write chanel head to a NEXT register */
762 wcn36xx_dxe_write_register(wcn, WCN36XX_DXE_CH_NEXT_DESC_ADDR_RX_H,
763 wcn->dxe_rx_h_ch.head_blk_ctl->desc_phy_addr);
764
765 /* Write DMA source address */
766 wcn36xx_dxe_write_register(wcn,
767 WCN36XX_DXE_CH_SRC_ADDR_RX_H,
768 WCN36XX_DXE_WQ_RX_H);
769
770 /* Program preallocated destination address */
771 wcn36xx_dxe_write_register(wcn,
772 WCN36XX_DXE_CH_DEST_ADDR_RX_H,
773 wcn->dxe_rx_h_ch.head_blk_ctl->desc->phy_next_l);
774
775 /* Enable default control registers */
776 wcn36xx_dxe_write_register(wcn,
777 WCN36XX_DXE_REG_CTL_RX_H,
778 WCN36XX_DXE_CH_DEFAULT_CTL_RX_H);
779
780 /* Enable channel interrupts */
781 wcn36xx_dxe_enable_ch_int(wcn, WCN36XX_INT_MASK_CHAN_RX_H);
782
783 ret = wcn36xx_dxe_request_irqs(wcn);
784 if (ret < 0)
785 goto out_err;
786
787 return 0;
788
789out_err:
790 return ret;
791}
792
793void wcn36xx_dxe_deinit(struct wcn36xx *wcn)
794{
795 free_irq(wcn->tx_irq, wcn);
796 free_irq(wcn->rx_irq, wcn);
797
798 if (wcn->tx_ack_skb) {
799 ieee80211_tx_status_irqsafe(wcn->hw, wcn->tx_ack_skb);
800 wcn->tx_ack_skb = NULL;
801 }
802
803 wcn36xx_dxe_ch_free_skbs(wcn, &wcn->dxe_rx_l_ch);
804 wcn36xx_dxe_ch_free_skbs(wcn, &wcn->dxe_rx_h_ch);
805}
diff --git a/drivers/net/wireless/ath/wcn36xx/dxe.h b/drivers/net/wireless/ath/wcn36xx/dxe.h
new file mode 100644
index 000000000000..c88562f85de1
--- /dev/null
+++ b/drivers/net/wireless/ath/wcn36xx/dxe.h
@@ -0,0 +1,284 @@
1/*
2 * Copyright (c) 2013 Eugene Krasnikov <k.eugene.e@gmail.com>
3 *
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
7 *
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
11 * SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION
13 * OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN
14 * CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
15 */
16
17#ifndef _DXE_H_
18#define _DXE_H_
19
20#include "wcn36xx.h"
21
22/*
23TX_LOW = DMA0
24TX_HIGH = DMA4
25RX_LOW = DMA1
26RX_HIGH = DMA3
27H2H_TEST_RX_TX = DMA2
28*/
29
30/* DXE registers */
31#define WCN36XX_DXE_MEM_BASE 0x03000000
32#define WCN36XX_DXE_MEM_REG 0x202000
33
34#define WCN36XX_DXE_CCU_INT 0xA0011
35#define WCN36XX_DXE_REG_CCU_INT 0x200b10
36
37/* TODO This must calculated properly but not hardcoded */
38#define WCN36XX_DXE_CTRL_TX_L 0x328a44
39#define WCN36XX_DXE_CTRL_TX_H 0x32ce44
40#define WCN36XX_DXE_CTRL_RX_L 0x12ad2f
41#define WCN36XX_DXE_CTRL_RX_H 0x12d12f
42#define WCN36XX_DXE_CTRL_TX_H_BD 0x30ce45
43#define WCN36XX_DXE_CTRL_TX_H_SKB 0x32ce4d
44#define WCN36XX_DXE_CTRL_TX_L_BD 0x308a45
45#define WCN36XX_DXE_CTRL_TX_L_SKB 0x328a4d
46
47/* TODO This must calculated properly but not hardcoded */
48#define WCN36XX_DXE_WQ_TX_L 0x17
49#define WCN36XX_DXE_WQ_TX_H 0x17
50#define WCN36XX_DXE_WQ_RX_L 0xB
51#define WCN36XX_DXE_WQ_RX_H 0x4
52
53/* DXE descriptor control filed */
54#define WCN36XX_DXE_CTRL_VALID_MASK (0x00000001)
55
56/* TODO This must calculated properly but not hardcoded */
57/* DXE default control register values */
58#define WCN36XX_DXE_CH_DEFAULT_CTL_RX_L 0x847EAD2F
59#define WCN36XX_DXE_CH_DEFAULT_CTL_RX_H 0x84FED12F
60#define WCN36XX_DXE_CH_DEFAULT_CTL_TX_H 0x853ECF4D
61#define WCN36XX_DXE_CH_DEFAULT_CTL_TX_L 0x843e8b4d
62
63/* Common DXE registers */
64#define WCN36XX_DXE_MEM_CSR (WCN36XX_DXE_MEM_REG + 0x00)
65#define WCN36XX_DXE_REG_CSR_RESET (WCN36XX_DXE_MEM_REG + 0x00)
66#define WCN36XX_DXE_ENCH_ADDR (WCN36XX_DXE_MEM_REG + 0x04)
67#define WCN36XX_DXE_REG_CH_EN (WCN36XX_DXE_MEM_REG + 0x08)
68#define WCN36XX_DXE_REG_CH_DONE (WCN36XX_DXE_MEM_REG + 0x0C)
69#define WCN36XX_DXE_REG_CH_ERR (WCN36XX_DXE_MEM_REG + 0x10)
70#define WCN36XX_DXE_INT_MASK_REG (WCN36XX_DXE_MEM_REG + 0x18)
71#define WCN36XX_DXE_INT_SRC_RAW_REG (WCN36XX_DXE_MEM_REG + 0x20)
72 /* #define WCN36XX_DXE_INT_CH6_MASK 0x00000040 */
73 /* #define WCN36XX_DXE_INT_CH5_MASK 0x00000020 */
74 #define WCN36XX_DXE_INT_CH4_MASK 0x00000010
75 #define WCN36XX_DXE_INT_CH3_MASK 0x00000008
76 /* #define WCN36XX_DXE_INT_CH2_MASK 0x00000004 */
77 #define WCN36XX_DXE_INT_CH1_MASK 0x00000002
78 #define WCN36XX_DXE_INT_CH0_MASK 0x00000001
79#define WCN36XX_DXE_0_INT_CLR (WCN36XX_DXE_MEM_REG + 0x30)
80#define WCN36XX_DXE_0_INT_ED_CLR (WCN36XX_DXE_MEM_REG + 0x34)
81#define WCN36XX_DXE_0_INT_DONE_CLR (WCN36XX_DXE_MEM_REG + 0x38)
82#define WCN36XX_DXE_0_INT_ERR_CLR (WCN36XX_DXE_MEM_REG + 0x3C)
83
84#define WCN36XX_DXE_0_CH0_STATUS (WCN36XX_DXE_MEM_REG + 0x404)
85#define WCN36XX_DXE_0_CH1_STATUS (WCN36XX_DXE_MEM_REG + 0x444)
86#define WCN36XX_DXE_0_CH2_STATUS (WCN36XX_DXE_MEM_REG + 0x484)
87#define WCN36XX_DXE_0_CH3_STATUS (WCN36XX_DXE_MEM_REG + 0x4C4)
88#define WCN36XX_DXE_0_CH4_STATUS (WCN36XX_DXE_MEM_REG + 0x504)
89
90#define WCN36XX_DXE_REG_RESET 0x5c89
91
92/* Temporary BMU Workqueue 4 */
93#define WCN36XX_DXE_BMU_WQ_RX_LOW 0xB
94#define WCN36XX_DXE_BMU_WQ_RX_HIGH 0x4
95/* DMA channel offset */
96#define WCN36XX_DXE_TX_LOW_OFFSET 0x400
97#define WCN36XX_DXE_TX_HIGH_OFFSET 0x500
98#define WCN36XX_DXE_RX_LOW_OFFSET 0x440
99#define WCN36XX_DXE_RX_HIGH_OFFSET 0x4C0
100
101/* Address of the next DXE descriptor */
102#define WCN36XX_DXE_CH_NEXT_DESC_ADDR 0x001C
103#define WCN36XX_DXE_CH_NEXT_DESC_ADDR_TX_L (WCN36XX_DXE_MEM_REG + \
104 WCN36XX_DXE_TX_LOW_OFFSET + \
105 WCN36XX_DXE_CH_NEXT_DESC_ADDR)
106#define WCN36XX_DXE_CH_NEXT_DESC_ADDR_TX_H (WCN36XX_DXE_MEM_REG + \
107 WCN36XX_DXE_TX_HIGH_OFFSET + \
108 WCN36XX_DXE_CH_NEXT_DESC_ADDR)
109#define WCN36XX_DXE_CH_NEXT_DESC_ADDR_RX_L (WCN36XX_DXE_MEM_REG + \
110 WCN36XX_DXE_RX_LOW_OFFSET + \
111 WCN36XX_DXE_CH_NEXT_DESC_ADDR)
112#define WCN36XX_DXE_CH_NEXT_DESC_ADDR_RX_H (WCN36XX_DXE_MEM_REG + \
113 WCN36XX_DXE_RX_HIGH_OFFSET + \
114 WCN36XX_DXE_CH_NEXT_DESC_ADDR)
115
116/* DXE Descriptor source address */
117#define WCN36XX_DXE_CH_SRC_ADDR 0x000C
118#define WCN36XX_DXE_CH_SRC_ADDR_RX_L (WCN36XX_DXE_MEM_REG + \
119 WCN36XX_DXE_RX_LOW_OFFSET + \
120 WCN36XX_DXE_CH_SRC_ADDR)
121#define WCN36XX_DXE_CH_SRC_ADDR_RX_H (WCN36XX_DXE_MEM_REG + \
122 WCN36XX_DXE_RX_HIGH_OFFSET + \
123 WCN36XX_DXE_CH_SRC_ADDR)
124
125/* DXE Descriptor address destination address */
126#define WCN36XX_DXE_CH_DEST_ADDR 0x0014
127#define WCN36XX_DXE_CH_DEST_ADDR_TX_L (WCN36XX_DXE_MEM_REG + \
128 WCN36XX_DXE_TX_LOW_OFFSET + \
129 WCN36XX_DXE_CH_DEST_ADDR)
130#define WCN36XX_DXE_CH_DEST_ADDR_TX_H (WCN36XX_DXE_MEM_REG + \
131 WCN36XX_DXE_TX_HIGH_OFFSET + \
132 WCN36XX_DXE_CH_DEST_ADDR)
133#define WCN36XX_DXE_CH_DEST_ADDR_RX_L (WCN36XX_DXE_MEM_REG + \
134 WCN36XX_DXE_RX_LOW_OFFSET + \
135 WCN36XX_DXE_CH_DEST_ADDR)
136#define WCN36XX_DXE_CH_DEST_ADDR_RX_H (WCN36XX_DXE_MEM_REG + \
137 WCN36XX_DXE_RX_HIGH_OFFSET + \
138 WCN36XX_DXE_CH_DEST_ADDR)
139
140/* Interrupt status */
141#define WCN36XX_DXE_CH_STATUS_REG_ADDR 0x0004
142#define WCN36XX_DXE_CH_STATUS_REG_ADDR_TX_L (WCN36XX_DXE_MEM_REG + \
143 WCN36XX_DXE_TX_LOW_OFFSET + \
144 WCN36XX_DXE_CH_STATUS_REG_ADDR)
145#define WCN36XX_DXE_CH_STATUS_REG_ADDR_TX_H (WCN36XX_DXE_MEM_REG + \
146 WCN36XX_DXE_TX_HIGH_OFFSET + \
147 WCN36XX_DXE_CH_STATUS_REG_ADDR)
148#define WCN36XX_DXE_CH_STATUS_REG_ADDR_RX_L (WCN36XX_DXE_MEM_REG + \
149 WCN36XX_DXE_RX_LOW_OFFSET + \
150 WCN36XX_DXE_CH_STATUS_REG_ADDR)
151#define WCN36XX_DXE_CH_STATUS_REG_ADDR_RX_H (WCN36XX_DXE_MEM_REG + \
152 WCN36XX_DXE_RX_HIGH_OFFSET + \
153 WCN36XX_DXE_CH_STATUS_REG_ADDR)
154
155
156/* DXE default control register */
157#define WCN36XX_DXE_REG_CTL_RX_L (WCN36XX_DXE_MEM_REG + \
158 WCN36XX_DXE_RX_LOW_OFFSET)
159#define WCN36XX_DXE_REG_CTL_RX_H (WCN36XX_DXE_MEM_REG + \
160 WCN36XX_DXE_RX_HIGH_OFFSET)
161#define WCN36XX_DXE_REG_CTL_TX_H (WCN36XX_DXE_MEM_REG + \
162 WCN36XX_DXE_TX_HIGH_OFFSET)
163#define WCN36XX_DXE_REG_CTL_TX_L (WCN36XX_DXE_MEM_REG + \
164 WCN36XX_DXE_TX_LOW_OFFSET)
165
166#define WCN36XX_SMSM_WLAN_TX_ENABLE 0x00000400
167#define WCN36XX_SMSM_WLAN_TX_RINGS_EMPTY 0x00000200
168
169
170/* Interrupt control channel mask */
171#define WCN36XX_INT_MASK_CHAN_TX_L 0x00000001
172#define WCN36XX_INT_MASK_CHAN_RX_L 0x00000002
173#define WCN36XX_INT_MASK_CHAN_RX_H 0x00000008
174#define WCN36XX_INT_MASK_CHAN_TX_H 0x00000010
175
176#define WCN36XX_BD_CHUNK_SIZE 128
177
178#define WCN36XX_PKT_SIZE 0xF20
179enum wcn36xx_dxe_ch_type {
180 WCN36XX_DXE_CH_TX_L,
181 WCN36XX_DXE_CH_TX_H,
182 WCN36XX_DXE_CH_RX_L,
183 WCN36XX_DXE_CH_RX_H
184};
185
186/* amount of descriptors per channel */
187enum wcn36xx_dxe_ch_desc_num {
188 WCN36XX_DXE_CH_DESC_NUMB_TX_L = 128,
189 WCN36XX_DXE_CH_DESC_NUMB_TX_H = 10,
190 WCN36XX_DXE_CH_DESC_NUMB_RX_L = 512,
191 WCN36XX_DXE_CH_DESC_NUMB_RX_H = 40
192};
193
194/**
195 * struct wcn36xx_dxe_desc - describes descriptor of one DXE buffer
196 *
197 * @ctrl: is a union that consists of following bits:
198 * union {
199 * u32 valid :1; //0 = DMA stop, 1 = DMA continue with this
200 * //descriptor
201 * u32 transfer_type :2; //0 = Host to Host space
202 * u32 eop :1; //End of Packet
203 * u32 bd_handling :1; //if transferType = Host to BMU, then 0
204 * // means first 128 bytes contain BD, and 1
205 * // means create new empty BD
206 * u32 siq :1; // SIQ
207 * u32 diq :1; // DIQ
208 * u32 pdu_rel :1; //0 = don't release BD and PDUs when done,
209 * // 1 = release them
210 * u32 bthld_sel :4; //BMU Threshold Select
211 * u32 prio :3; //Specifies the priority level to use for
212 * // the transfer
213 * u32 stop_channel :1; //1 = DMA stops processing further, channel
214 * //requires re-enabling after this
215 * u32 intr :1; //Interrupt on Descriptor Done
216 * u32 rsvd :1; //reserved
217 * u32 size :14;//14 bits used - ignored for BMU transfers,
218 * //only used for host to host transfers?
219 * } ctrl;
220 */
221struct wcn36xx_dxe_desc {
222 u32 ctrl;
223 u32 fr_len;
224
225 u32 src_addr_l;
226 u32 dst_addr_l;
227 u32 phy_next_l;
228 u32 src_addr_h;
229 u32 dst_addr_h;
230 u32 phy_next_h;
231} __packed;
232
233/* DXE Control block */
234struct wcn36xx_dxe_ctl {
235 struct wcn36xx_dxe_ctl *next;
236 struct wcn36xx_dxe_desc *desc;
237 unsigned int desc_phy_addr;
238 int ctl_blk_order;
239 struct sk_buff *skb;
240 spinlock_t skb_lock;
241 void *bd_cpu_addr;
242 dma_addr_t bd_phy_addr;
243};
244
245struct wcn36xx_dxe_ch {
246 enum wcn36xx_dxe_ch_type ch_type;
247 void *cpu_addr;
248 dma_addr_t dma_addr;
249 enum wcn36xx_dxe_ch_desc_num desc_num;
250 /* DXE control block ring */
251 struct wcn36xx_dxe_ctl *head_blk_ctl;
252 struct wcn36xx_dxe_ctl *tail_blk_ctl;
253
254 /* DXE channel specific configs */
255 u32 dxe_wq;
256 u32 ctrl_bd;
257 u32 ctrl_skb;
258 u32 reg_ctrl;
259 u32 def_ctrl;
260};
261
262/* Memory Pool for BD headers */
263struct wcn36xx_dxe_mem_pool {
264 int chunk_size;
265 void *virt_addr;
266 dma_addr_t phy_addr;
267};
268
269struct wcn36xx_vif;
270int wcn36xx_dxe_allocate_mem_pools(struct wcn36xx *wcn);
271void wcn36xx_dxe_free_mem_pools(struct wcn36xx *wcn);
272void wcn36xx_dxe_rx_frame(struct wcn36xx *wcn);
273int wcn36xx_dxe_alloc_ctl_blks(struct wcn36xx *wcn);
274void wcn36xx_dxe_free_ctl_blks(struct wcn36xx *wcn);
275int wcn36xx_dxe_init(struct wcn36xx *wcn);
276void wcn36xx_dxe_deinit(struct wcn36xx *wcn);
277int wcn36xx_dxe_init_channels(struct wcn36xx *wcn);
278int wcn36xx_dxe_tx_frame(struct wcn36xx *wcn,
279 struct wcn36xx_vif *vif_priv,
280 struct sk_buff *skb,
281 bool is_low);
282void wcn36xx_dxe_tx_ack_ind(struct wcn36xx *wcn, u32 status);
283void *wcn36xx_dxe_get_next_bd(struct wcn36xx *wcn, bool is_low);
284#endif /* _DXE_H_ */
diff --git a/drivers/net/wireless/ath/wcn36xx/hal.h b/drivers/net/wireless/ath/wcn36xx/hal.h
new file mode 100644
index 000000000000..c02dbc618724
--- /dev/null
+++ b/drivers/net/wireless/ath/wcn36xx/hal.h
@@ -0,0 +1,4657 @@
1/*
2 * Copyright (c) 2013 Eugene Krasnikov <k.eugene.e@gmail.com>
3 *
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
7 *
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
11 * SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION
13 * OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN
14 * CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
15 */
16
17#ifndef _HAL_H_
18#define _HAL_H_
19
20/*---------------------------------------------------------------------------
21 API VERSIONING INFORMATION
22
23 The RIVA API is versioned as MAJOR.MINOR.VERSION.REVISION
24 The MAJOR is incremented for major product/architecture changes
25 (and then MINOR/VERSION/REVISION are zeroed)
26 The MINOR is incremented for minor product/architecture changes
27 (and then VERSION/REVISION are zeroed)
28 The VERSION is incremented if a significant API change occurs
29 (and then REVISION is zeroed)
30 The REVISION is incremented if an insignificant API change occurs
31 or if a new API is added
32 All values are in the range 0..255 (ie they are 8-bit values)
33 ---------------------------------------------------------------------------*/
34#define WCN36XX_HAL_VER_MAJOR 1
35#define WCN36XX_HAL_VER_MINOR 4
36#define WCN36XX_HAL_VER_VERSION 1
37#define WCN36XX_HAL_VER_REVISION 2
38
39/* This is to force compiler to use the maximum of an int ( 4 bytes ) */
40#define WCN36XX_HAL_MAX_ENUM_SIZE 0x7FFFFFFF
41#define WCN36XX_HAL_MSG_TYPE_MAX_ENUM_SIZE 0x7FFF
42
43/* Max no. of transmit categories */
44#define STACFG_MAX_TC 8
45
46/* The maximum value of access category */
47#define WCN36XX_HAL_MAX_AC 4
48
49#define WCN36XX_HAL_IPV4_ADDR_LEN 4
50
51#define WALN_HAL_STA_INVALID_IDX 0xFF
52#define WCN36XX_HAL_BSS_INVALID_IDX 0xFF
53
54/* Default Beacon template size */
55#define BEACON_TEMPLATE_SIZE 0x180
56
57/* Param Change Bitmap sent to HAL */
58#define PARAM_BCN_INTERVAL_CHANGED (1 << 0)
59#define PARAM_SHORT_PREAMBLE_CHANGED (1 << 1)
60#define PARAM_SHORT_SLOT_TIME_CHANGED (1 << 2)
61#define PARAM_llACOEXIST_CHANGED (1 << 3)
62#define PARAM_llBCOEXIST_CHANGED (1 << 4)
63#define PARAM_llGCOEXIST_CHANGED (1 << 5)
64#define PARAM_HT20MHZCOEXIST_CHANGED (1<<6)
65#define PARAM_NON_GF_DEVICES_PRESENT_CHANGED (1<<7)
66#define PARAM_RIFS_MODE_CHANGED (1<<8)
67#define PARAM_LSIG_TXOP_FULL_SUPPORT_CHANGED (1<<9)
68#define PARAM_OBSS_MODE_CHANGED (1<<10)
69#define PARAM_BEACON_UPDATE_MASK \
70 (PARAM_BCN_INTERVAL_CHANGED | \
71 PARAM_SHORT_PREAMBLE_CHANGED | \
72 PARAM_SHORT_SLOT_TIME_CHANGED | \
73 PARAM_llACOEXIST_CHANGED | \
74 PARAM_llBCOEXIST_CHANGED | \
75 PARAM_llGCOEXIST_CHANGED | \
76 PARAM_HT20MHZCOEXIST_CHANGED | \
77 PARAM_NON_GF_DEVICES_PRESENT_CHANGED | \
78 PARAM_RIFS_MODE_CHANGED | \
79 PARAM_LSIG_TXOP_FULL_SUPPORT_CHANGED | \
80 PARAM_OBSS_MODE_CHANGED)
81
82/* dump command response Buffer size */
83#define DUMPCMD_RSP_BUFFER 100
84
85/* version string max length (including NULL) */
86#define WCN36XX_HAL_VERSION_LENGTH 64
87
88/* message types for messages exchanged between WDI and HAL */
89enum wcn36xx_hal_host_msg_type {
90 /* Init/De-Init */
91 WCN36XX_HAL_START_REQ = 0,
92 WCN36XX_HAL_START_RSP = 1,
93 WCN36XX_HAL_STOP_REQ = 2,
94 WCN36XX_HAL_STOP_RSP = 3,
95
96 /* Scan */
97 WCN36XX_HAL_INIT_SCAN_REQ = 4,
98 WCN36XX_HAL_INIT_SCAN_RSP = 5,
99 WCN36XX_HAL_START_SCAN_REQ = 6,
100 WCN36XX_HAL_START_SCAN_RSP = 7,
101 WCN36XX_HAL_END_SCAN_REQ = 8,
102 WCN36XX_HAL_END_SCAN_RSP = 9,
103 WCN36XX_HAL_FINISH_SCAN_REQ = 10,
104 WCN36XX_HAL_FINISH_SCAN_RSP = 11,
105
106 /* HW STA configuration/deconfiguration */
107 WCN36XX_HAL_CONFIG_STA_REQ = 12,
108 WCN36XX_HAL_CONFIG_STA_RSP = 13,
109 WCN36XX_HAL_DELETE_STA_REQ = 14,
110 WCN36XX_HAL_DELETE_STA_RSP = 15,
111 WCN36XX_HAL_CONFIG_BSS_REQ = 16,
112 WCN36XX_HAL_CONFIG_BSS_RSP = 17,
113 WCN36XX_HAL_DELETE_BSS_REQ = 18,
114 WCN36XX_HAL_DELETE_BSS_RSP = 19,
115
116 /* Infra STA asscoiation */
117 WCN36XX_HAL_JOIN_REQ = 20,
118 WCN36XX_HAL_JOIN_RSP = 21,
119 WCN36XX_HAL_POST_ASSOC_REQ = 22,
120 WCN36XX_HAL_POST_ASSOC_RSP = 23,
121
122 /* Security */
123 WCN36XX_HAL_SET_BSSKEY_REQ = 24,
124 WCN36XX_HAL_SET_BSSKEY_RSP = 25,
125 WCN36XX_HAL_SET_STAKEY_REQ = 26,
126 WCN36XX_HAL_SET_STAKEY_RSP = 27,
127 WCN36XX_HAL_RMV_BSSKEY_REQ = 28,
128 WCN36XX_HAL_RMV_BSSKEY_RSP = 29,
129 WCN36XX_HAL_RMV_STAKEY_REQ = 30,
130 WCN36XX_HAL_RMV_STAKEY_RSP = 31,
131
132 /* Qos Related */
133 WCN36XX_HAL_ADD_TS_REQ = 32,
134 WCN36XX_HAL_ADD_TS_RSP = 33,
135 WCN36XX_HAL_DEL_TS_REQ = 34,
136 WCN36XX_HAL_DEL_TS_RSP = 35,
137 WCN36XX_HAL_UPD_EDCA_PARAMS_REQ = 36,
138 WCN36XX_HAL_UPD_EDCA_PARAMS_RSP = 37,
139 WCN36XX_HAL_ADD_BA_REQ = 38,
140 WCN36XX_HAL_ADD_BA_RSP = 39,
141 WCN36XX_HAL_DEL_BA_REQ = 40,
142 WCN36XX_HAL_DEL_BA_RSP = 41,
143
144 WCN36XX_HAL_CH_SWITCH_REQ = 42,
145 WCN36XX_HAL_CH_SWITCH_RSP = 43,
146 WCN36XX_HAL_SET_LINK_ST_REQ = 44,
147 WCN36XX_HAL_SET_LINK_ST_RSP = 45,
148 WCN36XX_HAL_GET_STATS_REQ = 46,
149 WCN36XX_HAL_GET_STATS_RSP = 47,
150 WCN36XX_HAL_UPDATE_CFG_REQ = 48,
151 WCN36XX_HAL_UPDATE_CFG_RSP = 49,
152
153 WCN36XX_HAL_MISSED_BEACON_IND = 50,
154 WCN36XX_HAL_UNKNOWN_ADDR2_FRAME_RX_IND = 51,
155 WCN36XX_HAL_MIC_FAILURE_IND = 52,
156 WCN36XX_HAL_FATAL_ERROR_IND = 53,
157 WCN36XX_HAL_SET_KEYDONE_MSG = 54,
158
159 /* NV Interface */
160 WCN36XX_HAL_DOWNLOAD_NV_REQ = 55,
161 WCN36XX_HAL_DOWNLOAD_NV_RSP = 56,
162
163 WCN36XX_HAL_ADD_BA_SESSION_REQ = 57,
164 WCN36XX_HAL_ADD_BA_SESSION_RSP = 58,
165 WCN36XX_HAL_TRIGGER_BA_REQ = 59,
166 WCN36XX_HAL_TRIGGER_BA_RSP = 60,
167 WCN36XX_HAL_UPDATE_BEACON_REQ = 61,
168 WCN36XX_HAL_UPDATE_BEACON_RSP = 62,
169 WCN36XX_HAL_SEND_BEACON_REQ = 63,
170 WCN36XX_HAL_SEND_BEACON_RSP = 64,
171
172 WCN36XX_HAL_SET_BCASTKEY_REQ = 65,
173 WCN36XX_HAL_SET_BCASTKEY_RSP = 66,
174 WCN36XX_HAL_DELETE_STA_CONTEXT_IND = 67,
175 WCN36XX_HAL_UPDATE_PROBE_RSP_TEMPLATE_REQ = 68,
176 WCN36XX_HAL_UPDATE_PROBE_RSP_TEMPLATE_RSP = 69,
177
178 /* PTT interface support */
179 WCN36XX_HAL_PROCESS_PTT_REQ = 70,
180 WCN36XX_HAL_PROCESS_PTT_RSP = 71,
181
182 /* BTAMP related events */
183 WCN36XX_HAL_SIGNAL_BTAMP_EVENT_REQ = 72,
184 WCN36XX_HAL_SIGNAL_BTAMP_EVENT_RSP = 73,
185 WCN36XX_HAL_TL_HAL_FLUSH_AC_REQ = 74,
186 WCN36XX_HAL_TL_HAL_FLUSH_AC_RSP = 75,
187
188 WCN36XX_HAL_ENTER_IMPS_REQ = 76,
189 WCN36XX_HAL_EXIT_IMPS_REQ = 77,
190 WCN36XX_HAL_ENTER_BMPS_REQ = 78,
191 WCN36XX_HAL_EXIT_BMPS_REQ = 79,
192 WCN36XX_HAL_ENTER_UAPSD_REQ = 80,
193 WCN36XX_HAL_EXIT_UAPSD_REQ = 81,
194 WCN36XX_HAL_UPDATE_UAPSD_PARAM_REQ = 82,
195 WCN36XX_HAL_CONFIGURE_RXP_FILTER_REQ = 83,
196 WCN36XX_HAL_ADD_BCN_FILTER_REQ = 84,
197 WCN36XX_HAL_REM_BCN_FILTER_REQ = 85,
198 WCN36XX_HAL_ADD_WOWL_BCAST_PTRN = 86,
199 WCN36XX_HAL_DEL_WOWL_BCAST_PTRN = 87,
200 WCN36XX_HAL_ENTER_WOWL_REQ = 88,
201 WCN36XX_HAL_EXIT_WOWL_REQ = 89,
202 WCN36XX_HAL_HOST_OFFLOAD_REQ = 90,
203 WCN36XX_HAL_SET_RSSI_THRESH_REQ = 91,
204 WCN36XX_HAL_GET_RSSI_REQ = 92,
205 WCN36XX_HAL_SET_UAPSD_AC_PARAMS_REQ = 93,
206 WCN36XX_HAL_CONFIGURE_APPS_CPU_WAKEUP_STATE_REQ = 94,
207
208 WCN36XX_HAL_ENTER_IMPS_RSP = 95,
209 WCN36XX_HAL_EXIT_IMPS_RSP = 96,
210 WCN36XX_HAL_ENTER_BMPS_RSP = 97,
211 WCN36XX_HAL_EXIT_BMPS_RSP = 98,
212 WCN36XX_HAL_ENTER_UAPSD_RSP = 99,
213 WCN36XX_HAL_EXIT_UAPSD_RSP = 100,
214 WCN36XX_HAL_SET_UAPSD_AC_PARAMS_RSP = 101,
215 WCN36XX_HAL_UPDATE_UAPSD_PARAM_RSP = 102,
216 WCN36XX_HAL_CONFIGURE_RXP_FILTER_RSP = 103,
217 WCN36XX_HAL_ADD_BCN_FILTER_RSP = 104,
218 WCN36XX_HAL_REM_BCN_FILTER_RSP = 105,
219 WCN36XX_HAL_SET_RSSI_THRESH_RSP = 106,
220 WCN36XX_HAL_HOST_OFFLOAD_RSP = 107,
221 WCN36XX_HAL_ADD_WOWL_BCAST_PTRN_RSP = 108,
222 WCN36XX_HAL_DEL_WOWL_BCAST_PTRN_RSP = 109,
223 WCN36XX_HAL_ENTER_WOWL_RSP = 110,
224 WCN36XX_HAL_EXIT_WOWL_RSP = 111,
225 WCN36XX_HAL_RSSI_NOTIFICATION_IND = 112,
226 WCN36XX_HAL_GET_RSSI_RSP = 113,
227 WCN36XX_HAL_CONFIGURE_APPS_CPU_WAKEUP_STATE_RSP = 114,
228
229 /* 11k related events */
230 WCN36XX_HAL_SET_MAX_TX_POWER_REQ = 115,
231 WCN36XX_HAL_SET_MAX_TX_POWER_RSP = 116,
232
233 /* 11R related msgs */
234 WCN36XX_HAL_AGGR_ADD_TS_REQ = 117,
235 WCN36XX_HAL_AGGR_ADD_TS_RSP = 118,
236
237 /* P2P WLAN_FEATURE_P2P */
238 WCN36XX_HAL_SET_P2P_GONOA_REQ = 119,
239 WCN36XX_HAL_SET_P2P_GONOA_RSP = 120,
240
241 /* WLAN Dump commands */
242 WCN36XX_HAL_DUMP_COMMAND_REQ = 121,
243 WCN36XX_HAL_DUMP_COMMAND_RSP = 122,
244
245 /* OEM_DATA FEATURE SUPPORT */
246 WCN36XX_HAL_START_OEM_DATA_REQ = 123,
247 WCN36XX_HAL_START_OEM_DATA_RSP = 124,
248
249 /* ADD SELF STA REQ and RSP */
250 WCN36XX_HAL_ADD_STA_SELF_REQ = 125,
251 WCN36XX_HAL_ADD_STA_SELF_RSP = 126,
252
253 /* DEL SELF STA SUPPORT */
254 WCN36XX_HAL_DEL_STA_SELF_REQ = 127,
255 WCN36XX_HAL_DEL_STA_SELF_RSP = 128,
256
257 /* Coex Indication */
258 WCN36XX_HAL_COEX_IND = 129,
259
260 /* Tx Complete Indication */
261 WCN36XX_HAL_OTA_TX_COMPL_IND = 130,
262
263 /* Host Suspend/resume messages */
264 WCN36XX_HAL_HOST_SUSPEND_IND = 131,
265 WCN36XX_HAL_HOST_RESUME_REQ = 132,
266 WCN36XX_HAL_HOST_RESUME_RSP = 133,
267
268 WCN36XX_HAL_SET_TX_POWER_REQ = 134,
269 WCN36XX_HAL_SET_TX_POWER_RSP = 135,
270 WCN36XX_HAL_GET_TX_POWER_REQ = 136,
271 WCN36XX_HAL_GET_TX_POWER_RSP = 137,
272
273 WCN36XX_HAL_P2P_NOA_ATTR_IND = 138,
274
275 WCN36XX_HAL_ENABLE_RADAR_DETECT_REQ = 139,
276 WCN36XX_HAL_ENABLE_RADAR_DETECT_RSP = 140,
277 WCN36XX_HAL_GET_TPC_REPORT_REQ = 141,
278 WCN36XX_HAL_GET_TPC_REPORT_RSP = 142,
279 WCN36XX_HAL_RADAR_DETECT_IND = 143,
280 WCN36XX_HAL_RADAR_DETECT_INTR_IND = 144,
281 WCN36XX_HAL_KEEP_ALIVE_REQ = 145,
282 WCN36XX_HAL_KEEP_ALIVE_RSP = 146,
283
284 /* PNO messages */
285 WCN36XX_HAL_SET_PREF_NETWORK_REQ = 147,
286 WCN36XX_HAL_SET_PREF_NETWORK_RSP = 148,
287 WCN36XX_HAL_SET_RSSI_FILTER_REQ = 149,
288 WCN36XX_HAL_SET_RSSI_FILTER_RSP = 150,
289 WCN36XX_HAL_UPDATE_SCAN_PARAM_REQ = 151,
290 WCN36XX_HAL_UPDATE_SCAN_PARAM_RSP = 152,
291 WCN36XX_HAL_PREF_NETW_FOUND_IND = 153,
292
293 WCN36XX_HAL_SET_TX_PER_TRACKING_REQ = 154,
294 WCN36XX_HAL_SET_TX_PER_TRACKING_RSP = 155,
295 WCN36XX_HAL_TX_PER_HIT_IND = 156,
296
297 WCN36XX_HAL_8023_MULTICAST_LIST_REQ = 157,
298 WCN36XX_HAL_8023_MULTICAST_LIST_RSP = 158,
299
300 WCN36XX_HAL_SET_PACKET_FILTER_REQ = 159,
301 WCN36XX_HAL_SET_PACKET_FILTER_RSP = 160,
302 WCN36XX_HAL_PACKET_FILTER_MATCH_COUNT_REQ = 161,
303 WCN36XX_HAL_PACKET_FILTER_MATCH_COUNT_RSP = 162,
304 WCN36XX_HAL_CLEAR_PACKET_FILTER_REQ = 163,
305 WCN36XX_HAL_CLEAR_PACKET_FILTER_RSP = 164,
306
307 /*
308 * This is temp fix. Should be removed once Host and Riva code is
309 * in sync.
310 */
311 WCN36XX_HAL_INIT_SCAN_CON_REQ = 165,
312
313 WCN36XX_HAL_SET_POWER_PARAMS_REQ = 166,
314 WCN36XX_HAL_SET_POWER_PARAMS_RSP = 167,
315
316 WCN36XX_HAL_TSM_STATS_REQ = 168,
317 WCN36XX_HAL_TSM_STATS_RSP = 169,
318
319 /* wake reason indication (WOW) */
320 WCN36XX_HAL_WAKE_REASON_IND = 170,
321
322 /* GTK offload support */
323 WCN36XX_HAL_GTK_OFFLOAD_REQ = 171,
324 WCN36XX_HAL_GTK_OFFLOAD_RSP = 172,
325 WCN36XX_HAL_GTK_OFFLOAD_GETINFO_REQ = 173,
326 WCN36XX_HAL_GTK_OFFLOAD_GETINFO_RSP = 174,
327
328 WCN36XX_HAL_FEATURE_CAPS_EXCHANGE_REQ = 175,
329 WCN36XX_HAL_FEATURE_CAPS_EXCHANGE_RSP = 176,
330 WCN36XX_HAL_EXCLUDE_UNENCRYPTED_IND = 177,
331
332 WCN36XX_HAL_SET_THERMAL_MITIGATION_REQ = 178,
333 WCN36XX_HAL_SET_THERMAL_MITIGATION_RSP = 179,
334
335 WCN36XX_HAL_UPDATE_VHT_OP_MODE_REQ = 182,
336 WCN36XX_HAL_UPDATE_VHT_OP_MODE_RSP = 183,
337
338 WCN36XX_HAL_P2P_NOA_START_IND = 184,
339
340 WCN36XX_HAL_GET_ROAM_RSSI_REQ = 185,
341 WCN36XX_HAL_GET_ROAM_RSSI_RSP = 186,
342
343 WCN36XX_HAL_CLASS_B_STATS_IND = 187,
344 WCN36XX_HAL_DEL_BA_IND = 188,
345 WCN36XX_HAL_DHCP_START_IND = 189,
346 WCN36XX_HAL_DHCP_STOP_IND = 190,
347
348 WCN36XX_HAL_MSG_MAX = WCN36XX_HAL_MSG_TYPE_MAX_ENUM_SIZE
349};
350
351/* Enumeration for Version */
352enum wcn36xx_hal_host_msg_version {
353 WCN36XX_HAL_MSG_VERSION0 = 0,
354 WCN36XX_HAL_MSG_VERSION1 = 1,
355 /* define as 2 bytes data */
356 WCN36XX_HAL_MSG_WCNSS_CTRL_VERSION = 0x7FFF,
357 WCN36XX_HAL_MSG_VERSION_MAX_FIELD = WCN36XX_HAL_MSG_WCNSS_CTRL_VERSION
358};
359
360enum driver_type {
361 DRIVER_TYPE_PRODUCTION = 0,
362 DRIVER_TYPE_MFG = 1,
363 DRIVER_TYPE_DVT = 2,
364 DRIVER_TYPE_MAX = WCN36XX_HAL_MAX_ENUM_SIZE
365};
366
367enum wcn36xx_hal_stop_type {
368 HAL_STOP_TYPE_SYS_RESET,
369 HAL_STOP_TYPE_SYS_DEEP_SLEEP,
370 HAL_STOP_TYPE_RF_KILL,
371 HAL_STOP_TYPE_MAX = WCN36XX_HAL_MAX_ENUM_SIZE
372};
373
374enum wcn36xx_hal_sys_mode {
375 HAL_SYS_MODE_NORMAL,
376 HAL_SYS_MODE_LEARN,
377 HAL_SYS_MODE_SCAN,
378 HAL_SYS_MODE_PROMISC,
379 HAL_SYS_MODE_SUSPEND_LINK,
380 HAL_SYS_MODE_ROAM_SCAN,
381 HAL_SYS_MODE_ROAM_SUSPEND_LINK,
382 HAL_SYS_MODE_MAX = WCN36XX_HAL_MAX_ENUM_SIZE
383};
384
385enum phy_chan_bond_state {
386 /* 20MHz IF bandwidth centered on IF carrier */
387 PHY_SINGLE_CHANNEL_CENTERED = 0,
388
389 /* 40MHz IF bandwidth with lower 20MHz supporting the primary channel */
390 PHY_DOUBLE_CHANNEL_LOW_PRIMARY = 1,
391
392 /* 40MHz IF bandwidth centered on IF carrier */
393 PHY_DOUBLE_CHANNEL_CENTERED = 2,
394
395 /* 40MHz IF bandwidth with higher 20MHz supporting the primary ch */
396 PHY_DOUBLE_CHANNEL_HIGH_PRIMARY = 3,
397
398 /* 20/40MHZ offset LOW 40/80MHZ offset CENTERED */
399 PHY_QUADRUPLE_CHANNEL_20MHZ_LOW_40MHZ_CENTERED = 4,
400
401 /* 20/40MHZ offset CENTERED 40/80MHZ offset CENTERED */
402 PHY_QUADRUPLE_CHANNEL_20MHZ_CENTERED_40MHZ_CENTERED = 5,
403
404 /* 20/40MHZ offset HIGH 40/80MHZ offset CENTERED */
405 PHY_QUADRUPLE_CHANNEL_20MHZ_HIGH_40MHZ_CENTERED = 6,
406
407 /* 20/40MHZ offset LOW 40/80MHZ offset LOW */
408 PHY_QUADRUPLE_CHANNEL_20MHZ_LOW_40MHZ_LOW = 7,
409
410 /* 20/40MHZ offset HIGH 40/80MHZ offset LOW */
411 PHY_QUADRUPLE_CHANNEL_20MHZ_HIGH_40MHZ_LOW = 8,
412
413 /* 20/40MHZ offset LOW 40/80MHZ offset HIGH */
414 PHY_QUADRUPLE_CHANNEL_20MHZ_LOW_40MHZ_HIGH = 9,
415
416 /* 20/40MHZ offset-HIGH 40/80MHZ offset HIGH */
417 PHY_QUADRUPLE_CHANNEL_20MHZ_HIGH_40MHZ_HIGH = 10,
418
419 PHY_CHANNEL_BONDING_STATE_MAX = WCN36XX_HAL_MAX_ENUM_SIZE
420};
421
422/* Spatial Multiplexing(SM) Power Save mode */
423enum wcn36xx_hal_ht_mimo_state {
424 /* Static SM Power Save mode */
425 WCN36XX_HAL_HT_MIMO_PS_STATIC = 0,
426
427 /* Dynamic SM Power Save mode */
428 WCN36XX_HAL_HT_MIMO_PS_DYNAMIC = 1,
429
430 /* reserved */
431 WCN36XX_HAL_HT_MIMO_PS_NA = 2,
432
433 /* SM Power Save disabled */
434 WCN36XX_HAL_HT_MIMO_PS_NO_LIMIT = 3,
435
436 WCN36XX_HAL_HT_MIMO_PS_MAX = WCN36XX_HAL_MAX_ENUM_SIZE
437};
438
439/* each station added has a rate mode which specifies the sta attributes */
440enum sta_rate_mode {
441 STA_TAURUS = 0,
442 STA_TITAN,
443 STA_POLARIS,
444 STA_11b,
445 STA_11bg,
446 STA_11a,
447 STA_11n,
448 STA_11ac,
449 STA_INVALID_RATE_MODE = WCN36XX_HAL_MAX_ENUM_SIZE
450};
451
452/* 1,2,5.5,11 */
453#define WCN36XX_HAL_NUM_DSSS_RATES 4
454
455/* 6,9,12,18,24,36,48,54 */
456#define WCN36XX_HAL_NUM_OFDM_RATES 8
457
458/* 72,96,108 */
459#define WCN36XX_HAL_NUM_POLARIS_RATES 3
460
461#define WCN36XX_HAL_MAC_MAX_SUPPORTED_MCS_SET 16
462
463enum wcn36xx_hal_bss_type {
464 WCN36XX_HAL_INFRASTRUCTURE_MODE,
465
466 /* Added for softAP support */
467 WCN36XX_HAL_INFRA_AP_MODE,
468
469 WCN36XX_HAL_IBSS_MODE,
470
471 /* Added for BT-AMP support */
472 WCN36XX_HAL_BTAMP_STA_MODE,
473
474 /* Added for BT-AMP support */
475 WCN36XX_HAL_BTAMP_AP_MODE,
476
477 WCN36XX_HAL_AUTO_MODE,
478
479 WCN36XX_HAL_DONOT_USE_BSS_TYPE = WCN36XX_HAL_MAX_ENUM_SIZE
480};
481
482enum wcn36xx_hal_nw_type {
483 WCN36XX_HAL_11A_NW_TYPE,
484 WCN36XX_HAL_11B_NW_TYPE,
485 WCN36XX_HAL_11G_NW_TYPE,
486 WCN36XX_HAL_11N_NW_TYPE,
487 WCN36XX_HAL_DONOT_USE_NW_TYPE = WCN36XX_HAL_MAX_ENUM_SIZE
488};
489
490#define WCN36XX_HAL_MAC_RATESET_EID_MAX 12
491
492enum wcn36xx_hal_ht_operating_mode {
493 /* No Protection */
494 WCN36XX_HAL_HT_OP_MODE_PURE,
495
496 /* Overlap Legacy device present, protection is optional */
497 WCN36XX_HAL_HT_OP_MODE_OVERLAP_LEGACY,
498
499 /* No legacy device, but 20 MHz HT present */
500 WCN36XX_HAL_HT_OP_MODE_NO_LEGACY_20MHZ_HT,
501
502 /* Protection is required */
503 WCN36XX_HAL_HT_OP_MODE_MIXED,
504
505 WCN36XX_HAL_HT_OP_MODE_MAX = WCN36XX_HAL_MAX_ENUM_SIZE
506};
507
508/* Encryption type enum used with peer */
509enum ani_ed_type {
510 WCN36XX_HAL_ED_NONE,
511 WCN36XX_HAL_ED_WEP40,
512 WCN36XX_HAL_ED_WEP104,
513 WCN36XX_HAL_ED_TKIP,
514 WCN36XX_HAL_ED_CCMP,
515 WCN36XX_HAL_ED_WPI,
516 WCN36XX_HAL_ED_AES_128_CMAC,
517 WCN36XX_HAL_ED_NOT_IMPLEMENTED = WCN36XX_HAL_MAX_ENUM_SIZE
518};
519
520#define WLAN_MAX_KEY_RSC_LEN 16
521#define WLAN_WAPI_KEY_RSC_LEN 16
522
523/* MAX key length when ULA is used */
524#define WCN36XX_HAL_MAC_MAX_KEY_LENGTH 32
525#define WCN36XX_HAL_MAC_MAX_NUM_OF_DEFAULT_KEYS 4
526
527/*
528 * Enum to specify whether key is used for TX only, RX only or both.
529 */
530enum ani_key_direction {
531 WCN36XX_HAL_TX_ONLY,
532 WCN36XX_HAL_RX_ONLY,
533 WCN36XX_HAL_TX_RX,
534 WCN36XX_HAL_TX_DEFAULT,
535 WCN36XX_HAL_DONOT_USE_KEY_DIRECTION = WCN36XX_HAL_MAX_ENUM_SIZE
536};
537
538enum ani_wep_type {
539 WCN36XX_HAL_WEP_STATIC,
540 WCN36XX_HAL_WEP_DYNAMIC,
541 WCN36XX_HAL_WEP_MAX = WCN36XX_HAL_MAX_ENUM_SIZE
542};
543
544enum wcn36xx_hal_link_state {
545
546 WCN36XX_HAL_LINK_IDLE_STATE = 0,
547 WCN36XX_HAL_LINK_PREASSOC_STATE = 1,
548 WCN36XX_HAL_LINK_POSTASSOC_STATE = 2,
549 WCN36XX_HAL_LINK_AP_STATE = 3,
550 WCN36XX_HAL_LINK_IBSS_STATE = 4,
551
552 /* BT-AMP Case */
553 WCN36XX_HAL_LINK_BTAMP_PREASSOC_STATE = 5,
554 WCN36XX_HAL_LINK_BTAMP_POSTASSOC_STATE = 6,
555 WCN36XX_HAL_LINK_BTAMP_AP_STATE = 7,
556 WCN36XX_HAL_LINK_BTAMP_STA_STATE = 8,
557
558 /* Reserved for HAL Internal Use */
559 WCN36XX_HAL_LINK_LEARN_STATE = 9,
560 WCN36XX_HAL_LINK_SCAN_STATE = 10,
561 WCN36XX_HAL_LINK_FINISH_SCAN_STATE = 11,
562 WCN36XX_HAL_LINK_INIT_CAL_STATE = 12,
563 WCN36XX_HAL_LINK_FINISH_CAL_STATE = 13,
564 WCN36XX_HAL_LINK_LISTEN_STATE = 14,
565
566 WCN36XX_HAL_LINK_MAX = WCN36XX_HAL_MAX_ENUM_SIZE
567};
568
569enum wcn36xx_hal_stats_mask {
570 HAL_SUMMARY_STATS_INFO = 0x00000001,
571 HAL_GLOBAL_CLASS_A_STATS_INFO = 0x00000002,
572 HAL_GLOBAL_CLASS_B_STATS_INFO = 0x00000004,
573 HAL_GLOBAL_CLASS_C_STATS_INFO = 0x00000008,
574 HAL_GLOBAL_CLASS_D_STATS_INFO = 0x00000010,
575 HAL_PER_STA_STATS_INFO = 0x00000020
576};
577
578/* BT-AMP events type */
579enum bt_amp_event_type {
580 BTAMP_EVENT_CONNECTION_START,
581 BTAMP_EVENT_CONNECTION_STOP,
582 BTAMP_EVENT_CONNECTION_TERMINATED,
583
584 /* This and beyond are invalid values */
585 BTAMP_EVENT_TYPE_MAX = WCN36XX_HAL_MAX_ENUM_SIZE,
586};
587
588/* PE Statistics */
589enum pe_stats_mask {
590 PE_SUMMARY_STATS_INFO = 0x00000001,
591 PE_GLOBAL_CLASS_A_STATS_INFO = 0x00000002,
592 PE_GLOBAL_CLASS_B_STATS_INFO = 0x00000004,
593 PE_GLOBAL_CLASS_C_STATS_INFO = 0x00000008,
594 PE_GLOBAL_CLASS_D_STATS_INFO = 0x00000010,
595 PE_PER_STA_STATS_INFO = 0x00000020,
596
597 /* This and beyond are invalid values */
598 PE_STATS_TYPE_MAX = WCN36XX_HAL_MAX_ENUM_SIZE
599};
600
601/*
602 * Configuration Parameter IDs
603 */
604#define WCN36XX_HAL_CFG_STA_ID 0
605#define WCN36XX_HAL_CFG_CURRENT_TX_ANTENNA 1
606#define WCN36XX_HAL_CFG_CURRENT_RX_ANTENNA 2
607#define WCN36XX_HAL_CFG_LOW_GAIN_OVERRIDE 3
608#define WCN36XX_HAL_CFG_POWER_STATE_PER_CHAIN 4
609#define WCN36XX_HAL_CFG_CAL_PERIOD 5
610#define WCN36XX_HAL_CFG_CAL_CONTROL 6
611#define WCN36XX_HAL_CFG_PROXIMITY 7
612#define WCN36XX_HAL_CFG_NETWORK_DENSITY 8
613#define WCN36XX_HAL_CFG_MAX_MEDIUM_TIME 9
614#define WCN36XX_HAL_CFG_MAX_MPDUS_IN_AMPDU 10
615#define WCN36XX_HAL_CFG_RTS_THRESHOLD 11
616#define WCN36XX_HAL_CFG_SHORT_RETRY_LIMIT 12
617#define WCN36XX_HAL_CFG_LONG_RETRY_LIMIT 13
618#define WCN36XX_HAL_CFG_FRAGMENTATION_THRESHOLD 14
619#define WCN36XX_HAL_CFG_DYNAMIC_THRESHOLD_ZERO 15
620#define WCN36XX_HAL_CFG_DYNAMIC_THRESHOLD_ONE 16
621#define WCN36XX_HAL_CFG_DYNAMIC_THRESHOLD_TWO 17
622#define WCN36XX_HAL_CFG_FIXED_RATE 18
623#define WCN36XX_HAL_CFG_RETRYRATE_POLICY 19
624#define WCN36XX_HAL_CFG_RETRYRATE_SECONDARY 20
625#define WCN36XX_HAL_CFG_RETRYRATE_TERTIARY 21
626#define WCN36XX_HAL_CFG_FORCE_POLICY_PROTECTION 22
627#define WCN36XX_HAL_CFG_FIXED_RATE_MULTICAST_24GHZ 23
628#define WCN36XX_HAL_CFG_FIXED_RATE_MULTICAST_5GHZ 24
629#define WCN36XX_HAL_CFG_DEFAULT_RATE_INDEX_24GHZ 25
630#define WCN36XX_HAL_CFG_DEFAULT_RATE_INDEX_5GHZ 26
631#define WCN36XX_HAL_CFG_MAX_BA_SESSIONS 27
632#define WCN36XX_HAL_CFG_PS_DATA_INACTIVITY_TIMEOUT 28
633#define WCN36XX_HAL_CFG_PS_ENABLE_BCN_FILTER 29
634#define WCN36XX_HAL_CFG_PS_ENABLE_RSSI_MONITOR 30
635#define WCN36XX_HAL_CFG_NUM_BEACON_PER_RSSI_AVERAGE 31
636#define WCN36XX_HAL_CFG_STATS_PERIOD 32
637#define WCN36XX_HAL_CFG_CFP_MAX_DURATION 33
638#define WCN36XX_HAL_CFG_FRAME_TRANS_ENABLED 34
639#define WCN36XX_HAL_CFG_DTIM_PERIOD 35
640#define WCN36XX_HAL_CFG_EDCA_WMM_ACBK 36
641#define WCN36XX_HAL_CFG_EDCA_WMM_ACBE 37
642#define WCN36XX_HAL_CFG_EDCA_WMM_ACVO 38
643#define WCN36XX_HAL_CFG_EDCA_WMM_ACVI 39
644#define WCN36XX_HAL_CFG_BA_THRESHOLD_HIGH 40
645#define WCN36XX_HAL_CFG_MAX_BA_BUFFERS 41
646#define WCN36XX_HAL_CFG_RPE_POLLING_THRESHOLD 42
647#define WCN36XX_HAL_CFG_RPE_AGING_THRESHOLD_FOR_AC0_REG 43
648#define WCN36XX_HAL_CFG_RPE_AGING_THRESHOLD_FOR_AC1_REG 44
649#define WCN36XX_HAL_CFG_RPE_AGING_THRESHOLD_FOR_AC2_REG 45
650#define WCN36XX_HAL_CFG_RPE_AGING_THRESHOLD_FOR_AC3_REG 46
651#define WCN36XX_HAL_CFG_NO_OF_ONCHIP_REORDER_SESSIONS 47
652#define WCN36XX_HAL_CFG_PS_LISTEN_INTERVAL 48
653#define WCN36XX_HAL_CFG_PS_HEART_BEAT_THRESHOLD 49
654#define WCN36XX_HAL_CFG_PS_NTH_BEACON_FILTER 50
655#define WCN36XX_HAL_CFG_PS_MAX_PS_POLL 51
656#define WCN36XX_HAL_CFG_PS_MIN_RSSI_THRESHOLD 52
657#define WCN36XX_HAL_CFG_PS_RSSI_FILTER_PERIOD 53
658#define WCN36XX_HAL_CFG_PS_BROADCAST_FRAME_FILTER_ENABLE 54
659#define WCN36XX_HAL_CFG_PS_IGNORE_DTIM 55
660#define WCN36XX_HAL_CFG_PS_ENABLE_BCN_EARLY_TERM 56
661#define WCN36XX_HAL_CFG_DYNAMIC_PS_POLL_VALUE 57
662#define WCN36XX_HAL_CFG_PS_NULLDATA_AP_RESP_TIMEOUT 58
663#define WCN36XX_HAL_CFG_TELE_BCN_WAKEUP_EN 59
664#define WCN36XX_HAL_CFG_TELE_BCN_TRANS_LI 60
665#define WCN36XX_HAL_CFG_TELE_BCN_TRANS_LI_IDLE_BCNS 61
666#define WCN36XX_HAL_CFG_TELE_BCN_MAX_LI 62
667#define WCN36XX_HAL_CFG_TELE_BCN_MAX_LI_IDLE_BCNS 63
668#define WCN36XX_HAL_CFG_TX_PWR_CTRL_ENABLE 64
669#define WCN36XX_HAL_CFG_VALID_RADAR_CHANNEL_LIST 65
670#define WCN36XX_HAL_CFG_TX_POWER_24_20 66
671#define WCN36XX_HAL_CFG_TX_POWER_24_40 67
672#define WCN36XX_HAL_CFG_TX_POWER_50_20 68
673#define WCN36XX_HAL_CFG_TX_POWER_50_40 69
674#define WCN36XX_HAL_CFG_MCAST_BCAST_FILTER_SETTING 70
675#define WCN36XX_HAL_CFG_BCN_EARLY_TERM_WAKEUP_INTERVAL 71
676#define WCN36XX_HAL_CFG_MAX_TX_POWER_2_4 72
677#define WCN36XX_HAL_CFG_MAX_TX_POWER_5 73
678#define WCN36XX_HAL_CFG_INFRA_STA_KEEP_ALIVE_PERIOD 74
679#define WCN36XX_HAL_CFG_ENABLE_CLOSE_LOOP 75
680#define WCN36XX_HAL_CFG_BTC_EXECUTION_MODE 76
681#define WCN36XX_HAL_CFG_BTC_DHCP_BT_SLOTS_TO_BLOCK 77
682#define WCN36XX_HAL_CFG_BTC_A2DP_DHCP_BT_SUB_INTERVALS 78
683#define WCN36XX_HAL_CFG_PS_TX_INACTIVITY_TIMEOUT 79
684#define WCN36XX_HAL_CFG_WCNSS_API_VERSION 80
685#define WCN36XX_HAL_CFG_AP_KEEPALIVE_TIMEOUT 81
686#define WCN36XX_HAL_CFG_GO_KEEPALIVE_TIMEOUT 82
687#define WCN36XX_HAL_CFG_ENABLE_MC_ADDR_LIST 83
688#define WCN36XX_HAL_CFG_BTC_STATIC_LEN_INQ_BT 84
689#define WCN36XX_HAL_CFG_BTC_STATIC_LEN_PAGE_BT 85
690#define WCN36XX_HAL_CFG_BTC_STATIC_LEN_CONN_BT 86
691#define WCN36XX_HAL_CFG_BTC_STATIC_LEN_LE_BT 87
692#define WCN36XX_HAL_CFG_BTC_STATIC_LEN_INQ_WLAN 88
693#define WCN36XX_HAL_CFG_BTC_STATIC_LEN_PAGE_WLAN 89
694#define WCN36XX_HAL_CFG_BTC_STATIC_LEN_CONN_WLAN 90
695#define WCN36XX_HAL_CFG_BTC_STATIC_LEN_LE_WLAN 91
696#define WCN36XX_HAL_CFG_BTC_DYN_MAX_LEN_BT 92
697#define WCN36XX_HAL_CFG_BTC_DYN_MAX_LEN_WLAN 93
698#define WCN36XX_HAL_CFG_BTC_MAX_SCO_BLOCK_PERC 94
699#define WCN36XX_HAL_CFG_BTC_DHCP_PROT_ON_A2DP 95
700#define WCN36XX_HAL_CFG_BTC_DHCP_PROT_ON_SCO 96
701#define WCN36XX_HAL_CFG_ENABLE_UNICAST_FILTER 97
702#define WCN36XX_HAL_CFG_MAX_ASSOC_LIMIT 98
703#define WCN36XX_HAL_CFG_ENABLE_LPWR_IMG_TRANSITION 99
704#define WCN36XX_HAL_CFG_ENABLE_MCC_ADAPTIVE_SCHEDULER 100
705#define WCN36XX_HAL_CFG_ENABLE_DETECT_PS_SUPPORT 101
706#define WCN36XX_HAL_CFG_AP_LINK_MONITOR_TIMEOUT 102
707#define WCN36XX_HAL_CFG_BTC_DWELL_TIME_MULTIPLIER 103
708#define WCN36XX_HAL_CFG_ENABLE_TDLS_OXYGEN_MODE 104
709#define WCN36XX_HAL_CFG_MAX_PARAMS 105
710
711/* Message definitons - All the messages below need to be packed */
712
713/* Definition for HAL API Version. */
714struct wcnss_wlan_version {
715 u8 revision;
716 u8 version;
717 u8 minor;
718 u8 major;
719} __packed;
720
721/* Definition for Encryption Keys */
722struct wcn36xx_hal_keys {
723 u8 id;
724
725 /* 0 for multicast */
726 u8 unicast;
727
728 enum ani_key_direction direction;
729
730 /* Usage is unknown */
731 u8 rsc[WLAN_MAX_KEY_RSC_LEN];
732
733 /* =1 for authenticator,=0 for supplicant */
734 u8 pae_role;
735
736 u16 length;
737 u8 key[WCN36XX_HAL_MAC_MAX_KEY_LENGTH];
738} __packed;
739
740/*
741 * set_sta_key_params Moving here since it is shared by
742 * configbss/setstakey msgs
743 */
744struct wcn36xx_hal_set_sta_key_params {
745 /* STA Index */
746 u16 sta_index;
747
748 /* Encryption Type used with peer */
749 enum ani_ed_type enc_type;
750
751 /* STATIC/DYNAMIC - valid only for WEP */
752 enum ani_wep_type wep_type;
753
754 /* Default WEP key, valid only for static WEP, must between 0 and 3. */
755 u8 def_wep_idx;
756
757 /* valid only for non-static WEP encyrptions */
758 struct wcn36xx_hal_keys key[WCN36XX_HAL_MAC_MAX_NUM_OF_DEFAULT_KEYS];
759
760 /*
761 * Control for Replay Count, 1= Single TID based replay count on Tx
762 * 0 = Per TID based replay count on TX
763 */
764 u8 single_tid_rc;
765
766} __packed;
767
768/* 4-byte control message header used by HAL*/
769struct wcn36xx_hal_msg_header {
770 enum wcn36xx_hal_host_msg_type msg_type:16;
771 enum wcn36xx_hal_host_msg_version msg_version:16;
772 u32 len;
773} __packed;
774
775/* Config format required by HAL for each CFG item*/
776struct wcn36xx_hal_cfg {
777 /* Cfg Id. The Id required by HAL is exported by HAL
778 * in shared header file between UMAC and HAL.*/
779 u16 id;
780
781 /* Length of the Cfg. This parameter is used to go to next cfg
782 * in the TLV format.*/
783 u16 len;
784
785 /* Padding bytes for unaligned address's */
786 u16 pad_bytes;
787
788 /* Reserve bytes for making cfgVal to align address */
789 u16 reserve;
790
791 /* Following the uCfgLen field there should be a 'uCfgLen' bytes
792 * containing the uCfgValue ; u8 uCfgValue[uCfgLen] */
793} __packed;
794
795struct wcn36xx_hal_mac_start_parameters {
796 /* Drive Type - Production or FTM etc */
797 enum driver_type type;
798
799 /* Length of the config buffer */
800 u32 len;
801
802 /* Following this there is a TLV formatted buffer of length
803 * "len" bytes containing all config values.
804 * The TLV is expected to be formatted like this:
805 * 0 15 31 31+CFG_LEN-1 length-1
806 * | CFG_ID | CFG_LEN | CFG_BODY | CFG_ID |......|
807 */
808} __packed;
809
810struct wcn36xx_hal_mac_start_req_msg {
811 /* config buffer must start in TLV format just here */
812 struct wcn36xx_hal_msg_header header;
813 struct wcn36xx_hal_mac_start_parameters params;
814} __packed;
815
816struct wcn36xx_hal_mac_start_rsp_params {
817 /* success or failure */
818 u16 status;
819
820 /* Max number of STA supported by the device */
821 u8 stations;
822
823 /* Max number of BSS supported by the device */
824 u8 bssids;
825
826 /* API Version */
827 struct wcnss_wlan_version version;
828
829 /* CRM build information */
830 u8 crm_version[WCN36XX_HAL_VERSION_LENGTH];
831
832 /* hardware/chipset/misc version information */
833 u8 wlan_version[WCN36XX_HAL_VERSION_LENGTH];
834
835} __packed;
836
837struct wcn36xx_hal_mac_start_rsp_msg {
838 struct wcn36xx_hal_msg_header header;
839 struct wcn36xx_hal_mac_start_rsp_params start_rsp_params;
840} __packed;
841
842struct wcn36xx_hal_mac_stop_req_params {
843 /* The reason for which the device is being stopped */
844 enum wcn36xx_hal_stop_type reason;
845
846} __packed;
847
848struct wcn36xx_hal_mac_stop_req_msg {
849 struct wcn36xx_hal_msg_header header;
850 struct wcn36xx_hal_mac_stop_req_params stop_req_params;
851} __packed;
852
853struct wcn36xx_hal_mac_stop_rsp_msg {
854 struct wcn36xx_hal_msg_header header;
855
856 /* success or failure */
857 u32 status;
858} __packed;
859
860struct wcn36xx_hal_update_cfg_req_msg {
861 /*
862 * Note: The length specified in tHalUpdateCfgReqMsg messages should be
863 * header.msgLen = sizeof(tHalUpdateCfgReqMsg) + uConfigBufferLen
864 */
865 struct wcn36xx_hal_msg_header header;
866
867 /* Length of the config buffer. Allows UMAC to update multiple CFGs */
868 u32 len;
869
870 /*
871 * Following this there is a TLV formatted buffer of length
872 * "uConfigBufferLen" bytes containing all config values.
873 * The TLV is expected to be formatted like this:
874 * 0 15 31 31+CFG_LEN-1 length-1
875 * | CFG_ID | CFG_LEN | CFG_BODY | CFG_ID |......|
876 */
877
878} __packed;
879
880struct wcn36xx_hal_update_cfg_rsp_msg {
881 struct wcn36xx_hal_msg_header header;
882
883 /* success or failure */
884 u32 status;
885
886} __packed;
887
888/* Frame control field format (2 bytes) */
889struct wcn36xx_hal_mac_frame_ctl {
890
891#ifndef ANI_LITTLE_BIT_ENDIAN
892
893 u8 subType:4;
894 u8 type:2;
895 u8 protVer:2;
896
897 u8 order:1;
898 u8 wep:1;
899 u8 moreData:1;
900 u8 powerMgmt:1;
901 u8 retry:1;
902 u8 moreFrag:1;
903 u8 fromDS:1;
904 u8 toDS:1;
905
906#else
907
908 u8 protVer:2;
909 u8 type:2;
910 u8 subType:4;
911
912 u8 toDS:1;
913 u8 fromDS:1;
914 u8 moreFrag:1;
915 u8 retry:1;
916 u8 powerMgmt:1;
917 u8 moreData:1;
918 u8 wep:1;
919 u8 order:1;
920
921#endif
922
923};
924
925/* Sequence control field */
926struct wcn36xx_hal_mac_seq_ctl {
927 u8 fragNum:4;
928 u8 seqNumLo:4;
929 u8 seqNumHi:8;
930};
931
932/* Management header format */
933struct wcn36xx_hal_mac_mgmt_hdr {
934 struct wcn36xx_hal_mac_frame_ctl fc;
935 u8 durationLo;
936 u8 durationHi;
937 u8 da[6];
938 u8 sa[6];
939 u8 bssId[6];
940 struct wcn36xx_hal_mac_seq_ctl seqControl;
941};
942
943/* FIXME: pronto v1 apparently has 4 */
944#define WCN36XX_HAL_NUM_BSSID 2
945
946/* Scan Entry to hold active BSS idx's */
947struct wcn36xx_hal_scan_entry {
948 u8 bss_index[WCN36XX_HAL_NUM_BSSID];
949 u8 active_bss_count;
950};
951
952struct wcn36xx_hal_init_scan_req_msg {
953 struct wcn36xx_hal_msg_header header;
954
955 /* LEARN - AP Role
956 SCAN - STA Role */
957 enum wcn36xx_hal_sys_mode mode;
958
959 /* BSSID of the BSS */
960 u8 bssid[ETH_ALEN];
961
962 /* Whether BSS needs to be notified */
963 u8 notify;
964
965 /* Kind of frame to be used for notifying the BSS (Data Null, QoS
966 * Null, or CTS to Self). Must always be a valid frame type. */
967 u8 frame_type;
968
969 /* UMAC has the option of passing the MAC frame to be used for
970 * notifying the BSS. If non-zero, HAL will use the MAC frame
971 * buffer pointed to by macMgmtHdr. If zero, HAL will generate the
972 * appropriate MAC frame based on frameType. */
973 u8 frame_len;
974
975 /* Following the framelength there is a MAC frame buffer if
976 * frameLength is non-zero. */
977 struct wcn36xx_hal_mac_mgmt_hdr mac_mgmt_hdr;
978
979 /* Entry to hold number of active BSS idx's */
980 struct wcn36xx_hal_scan_entry scan_entry;
981};
982
983struct wcn36xx_hal_init_scan_con_req_msg {
984 struct wcn36xx_hal_msg_header header;
985
986 /* LEARN - AP Role
987 SCAN - STA Role */
988 enum wcn36xx_hal_sys_mode mode;
989
990 /* BSSID of the BSS */
991 u8 bssid[ETH_ALEN];
992
993 /* Whether BSS needs to be notified */
994 u8 notify;
995
996 /* Kind of frame to be used for notifying the BSS (Data Null, QoS
997 * Null, or CTS to Self). Must always be a valid frame type. */
998 u8 frame_type;
999
1000 /* UMAC has the option of passing the MAC frame to be used for
1001 * notifying the BSS. If non-zero, HAL will use the MAC frame
1002 * buffer pointed to by macMgmtHdr. If zero, HAL will generate the
1003 * appropriate MAC frame based on frameType. */
1004 u8 frame_length;
1005
1006 /* Following the framelength there is a MAC frame buffer if
1007 * frameLength is non-zero. */
1008 struct wcn36xx_hal_mac_mgmt_hdr mac_mgmt_hdr;
1009
1010 /* Entry to hold number of active BSS idx's */
1011 struct wcn36xx_hal_scan_entry scan_entry;
1012
1013 /* Single NoA usage in Scanning */
1014 u8 use_noa;
1015
1016 /* Indicates the scan duration (in ms) */
1017 u16 scan_duration;
1018
1019};
1020
1021struct wcn36xx_hal_init_scan_rsp_msg {
1022 struct wcn36xx_hal_msg_header header;
1023
1024 /* success or failure */
1025 u32 status;
1026
1027} __packed;
1028
1029struct wcn36xx_hal_start_scan_req_msg {
1030 struct wcn36xx_hal_msg_header header;
1031
1032 /* Indicates the channel to scan */
1033 u8 scan_channel;
1034} __packed;
1035
1036struct wcn36xx_hal_start_rsp_msg {
1037 struct wcn36xx_hal_msg_header header;
1038
1039 /* success or failure */
1040 u32 status;
1041
1042 u32 start_tsf[2];
1043 u8 tx_mgmt_power;
1044
1045} __packed;
1046
1047struct wcn36xx_hal_end_scan_req_msg {
1048 struct wcn36xx_hal_msg_header header;
1049
1050 /* Indicates the channel to stop scanning. Not used really. But
1051 * retained for symmetry with "start Scan" message. It can also
1052 * help in error check if needed. */
1053 u8 scan_channel;
1054} __packed;
1055
1056struct wcn36xx_hal_end_scan_rsp_msg {
1057 struct wcn36xx_hal_msg_header header;
1058
1059 /* success or failure */
1060 u32 status;
1061} __packed;
1062
1063struct wcn36xx_hal_finish_scan_req_msg {
1064 struct wcn36xx_hal_msg_header header;
1065
1066 /* Identifies the operational state of the AP/STA
1067 * LEARN - AP Role SCAN - STA Role */
1068 enum wcn36xx_hal_sys_mode mode;
1069
1070 /* Operating channel to tune to. */
1071 u8 oper_channel;
1072
1073 /* Channel Bonding state If 20/40 MHz is operational, this will
1074 * indicate the 40 MHz extension channel in combination with the
1075 * control channel */
1076 enum phy_chan_bond_state cb_state;
1077
1078 /* BSSID of the BSS */
1079 u8 bssid[ETH_ALEN];
1080
1081 /* Whether BSS needs to be notified */
1082 u8 notify;
1083
1084 /* Kind of frame to be used for notifying the BSS (Data Null, QoS
1085 * Null, or CTS to Self). Must always be a valid frame type. */
1086 u8 frame_type;
1087
1088 /* UMAC has the option of passing the MAC frame to be used for
1089 * notifying the BSS. If non-zero, HAL will use the MAC frame
1090 * buffer pointed to by macMgmtHdr. If zero, HAL will generate the
1091 * appropriate MAC frame based on frameType. */
1092 u8 frame_length;
1093
1094 /* Following the framelength there is a MAC frame buffer if
1095 * frameLength is non-zero. */
1096 struct wcn36xx_hal_mac_mgmt_hdr mac_mgmt_hdr;
1097
1098 /* Entry to hold number of active BSS idx's */
1099 struct wcn36xx_hal_scan_entry scan_entry;
1100
1101} __packed;
1102
1103struct wcn36xx_hal_finish_scan_rsp_msg {
1104 struct wcn36xx_hal_msg_header header;
1105
1106 /* success or failure */
1107 u32 status;
1108
1109} __packed;
1110
1111enum wcn36xx_hal_rate_index {
1112 HW_RATE_INDEX_1MBPS = 0x82,
1113 HW_RATE_INDEX_2MBPS = 0x84,
1114 HW_RATE_INDEX_5_5MBPS = 0x8B,
1115 HW_RATE_INDEX_6MBPS = 0x0C,
1116 HW_RATE_INDEX_9MBPS = 0x12,
1117 HW_RATE_INDEX_11MBPS = 0x96,
1118 HW_RATE_INDEX_12MBPS = 0x18,
1119 HW_RATE_INDEX_18MBPS = 0x24,
1120 HW_RATE_INDEX_24MBPS = 0x30,
1121 HW_RATE_INDEX_36MBPS = 0x48,
1122 HW_RATE_INDEX_48MBPS = 0x60,
1123 HW_RATE_INDEX_54MBPS = 0x6C
1124};
1125
1126struct wcn36xx_hal_supported_rates {
1127 /*
1128 * For Self STA Entry: this represents Self Mode.
1129 * For Peer Stations, this represents the mode of the peer.
1130 * On Station:
1131 *
1132 * --this mode is updated when PE adds the Self Entry.
1133 *
1134 * -- OR when PE sends 'ADD_BSS' message and station context in BSS
1135 * is used to indicate the mode of the AP.
1136 *
1137 * ON AP:
1138 *
1139 * -- this mode is updated when PE sends 'ADD_BSS' and Sta entry
1140 * for that BSS is used to indicate the self mode of the AP.
1141 *
1142 * -- OR when a station is associated, PE sends 'ADD_STA' message
1143 * with this mode updated.
1144 */
1145
1146 enum sta_rate_mode op_rate_mode;
1147
1148 /* 11b, 11a and aniLegacyRates are IE rates which gives rate in
1149 * unit of 500Kbps */
1150 u16 dsss_rates[WCN36XX_HAL_NUM_DSSS_RATES];
1151 u16 ofdm_rates[WCN36XX_HAL_NUM_OFDM_RATES];
1152 u16 legacy_rates[WCN36XX_HAL_NUM_POLARIS_RATES];
1153 u16 reserved;
1154
1155 /* Taurus only supports 26 Titan Rates(no ESF/concat Rates will be
1156 * supported) First 26 bits are reserved for those Titan rates and
1157 * the last 4 bits(bit28-31) for Taurus, 2(bit26-27) bits are
1158 * reserved. */
1159 /* Titan and Taurus Rates */
1160 u32 enhanced_rate_bitmap;
1161
1162 /*
1163 * 0-76 bits used, remaining reserved
1164 * bits 0-15 and 32 should be set.
1165 */
1166 u8 supported_mcs_set[WCN36XX_HAL_MAC_MAX_SUPPORTED_MCS_SET];
1167
1168 /*
1169 * RX Highest Supported Data Rate defines the highest data
1170 * rate that the STA is able to receive, in unites of 1Mbps.
1171 * This value is derived from "Supported MCS Set field" inside
1172 * the HT capability element.
1173 */
1174 u16 rx_highest_data_rate;
1175
1176} __packed;
1177
1178struct wcn36xx_hal_config_sta_params {
1179 /* BSSID of STA */
1180 u8 bssid[ETH_ALEN];
1181
1182 /* ASSOC ID, as assigned by UMAC */
1183 u16 aid;
1184
1185 /* STA entry Type: 0 - Self, 1 - Other/Peer, 2 - BSSID, 3 - BCAST */
1186 u8 type;
1187
1188 /* Short Preamble Supported. */
1189 u8 short_preamble_supported;
1190
1191 /* MAC Address of STA */
1192 u8 mac[ETH_ALEN];
1193
1194 /* Listen interval of the STA */
1195 u16 listen_interval;
1196
1197 /* Support for 11e/WMM */
1198 u8 wmm_enabled;
1199
1200 /* 11n HT capable STA */
1201 u8 ht_capable;
1202
1203 /* TX Width Set: 0 - 20 MHz only, 1 - 20/40 MHz */
1204 u8 tx_channel_width_set;
1205
1206 /* RIFS mode 0 - NA, 1 - Allowed */
1207 u8 rifs_mode;
1208
1209 /* L-SIG TXOP Protection mechanism
1210 0 - No Support, 1 - Supported
1211 SG - there is global field */
1212 u8 lsig_txop_protection;
1213
1214 /* Max Ampdu Size supported by STA. TPE programming.
1215 0 : 8k , 1 : 16k, 2 : 32k, 3 : 64k */
1216 u8 max_ampdu_size;
1217
1218 /* Max Ampdu density. Used by RA. 3 : 0~7 : 2^(11nAMPDUdensity -4) */
1219 u8 max_ampdu_density;
1220
1221 /* Max AMSDU size 1 : 3839 bytes, 0 : 7935 bytes */
1222 u8 max_amsdu_size;
1223
1224 /* Short GI support for 40Mhz packets */
1225 u8 sgi_40mhz;
1226
1227 /* Short GI support for 20Mhz packets */
1228 u8 sgi_20Mhz;
1229
1230 /* TODO move this parameter to the end for 3680 */
1231 /* These rates are the intersection of peer and self capabilities. */
1232 struct wcn36xx_hal_supported_rates supported_rates;
1233
1234 /* Robust Management Frame (RMF) enabled/disabled */
1235 u8 rmf;
1236
1237 /* The unicast encryption type in the association */
1238 u32 encrypt_type;
1239
1240 /* HAL should update the existing STA entry, if this flag is set. UMAC
1241 will set this flag in case of RE-ASSOC, where we want to reuse the
1242 old STA ID. 0 = Add, 1 = Update */
1243 u8 action;
1244
1245 /* U-APSD Flags: 1b per AC. Encoded as follows:
1246 b7 b6 b5 b4 b3 b2 b1 b0 =
1247 X X X X BE BK VI VO */
1248 u8 uapsd;
1249
1250 /* Max SP Length */
1251 u8 max_sp_len;
1252
1253 /* 11n Green Field preamble support
1254 0 - Not supported, 1 - Supported */
1255 u8 green_field_capable;
1256
1257 /* MIMO Power Save mode */
1258 enum wcn36xx_hal_ht_mimo_state mimo_ps;
1259
1260 /* Delayed BA Support */
1261 u8 delayed_ba_support;
1262
1263 /* Max AMPDU duration in 32us */
1264 u8 max_ampdu_duration;
1265
1266 /* HT STA should set it to 1 if it is enabled in BSS. HT STA should
1267 * set it to 0 if AP does not support it. This indication is sent
1268 * to HAL and HAL uses this flag to pickup up appropriate 40Mhz
1269 * rates. */
1270 u8 dsss_cck_mode_40mhz;
1271
1272 /* Valid STA Idx when action=Update. Set to 0xFF when invalid!
1273 * Retained for backward compalibity with existing HAL code */
1274 u8 sta_index;
1275
1276 /* BSSID of BSS to which station is associated. Set to 0xFF when
1277 * invalid. Retained for backward compalibity with existing HAL
1278 * code */
1279 u8 bssid_index;
1280
1281 u8 p2p;
1282
1283 /* TODO add this parameter for 3680. */
1284 /* Reserved to align next field on a dword boundary */
1285 /* u8 reserved; */
1286} __packed;
1287
1288struct wcn36xx_hal_config_sta_req_msg {
1289 struct wcn36xx_hal_msg_header header;
1290 struct wcn36xx_hal_config_sta_params sta_params;
1291} __packed;
1292
1293struct wcn36xx_hal_config_sta_params_v1 {
1294 /* BSSID of STA */
1295 u8 bssid[ETH_ALEN];
1296
1297 /* ASSOC ID, as assigned by UMAC */
1298 u16 aid;
1299
1300 /* STA entry Type: 0 - Self, 1 - Other/Peer, 2 - BSSID, 3 - BCAST */
1301 u8 type;
1302
1303 /* Short Preamble Supported. */
1304 u8 short_preamble_supported;
1305
1306 /* MAC Address of STA */
1307 u8 mac[ETH_ALEN];
1308
1309 /* Listen interval of the STA */
1310 u16 listen_interval;
1311
1312 /* Support for 11e/WMM */
1313 u8 wmm_enabled;
1314
1315 /* 11n HT capable STA */
1316 u8 ht_capable;
1317
1318 /* TX Width Set: 0 - 20 MHz only, 1 - 20/40 MHz */
1319 u8 tx_channel_width_set;
1320
1321 /* RIFS mode 0 - NA, 1 - Allowed */
1322 u8 rifs_mode;
1323
1324 /* L-SIG TXOP Protection mechanism
1325 0 - No Support, 1 - Supported
1326 SG - there is global field */
1327 u8 lsig_txop_protection;
1328
1329 /* Max Ampdu Size supported by STA. TPE programming.
1330 0 : 8k , 1 : 16k, 2 : 32k, 3 : 64k */
1331 u8 max_ampdu_size;
1332
1333 /* Max Ampdu density. Used by RA. 3 : 0~7 : 2^(11nAMPDUdensity -4) */
1334 u8 max_ampdu_density;
1335
1336 /* Max AMSDU size 1 : 3839 bytes, 0 : 7935 bytes */
1337 u8 max_amsdu_size;
1338
1339 /* Short GI support for 40Mhz packets */
1340 u8 sgi_40mhz;
1341
1342 /* Short GI support for 20Mhz packets */
1343 u8 sgi_20Mhz;
1344
1345 /* Robust Management Frame (RMF) enabled/disabled */
1346 u8 rmf;
1347
1348 /* The unicast encryption type in the association */
1349 u32 encrypt_type;
1350
1351 /* HAL should update the existing STA entry, if this flag is set. UMAC
1352 will set this flag in case of RE-ASSOC, where we want to reuse the
1353 old STA ID. 0 = Add, 1 = Update */
1354 u8 action;
1355
1356 /* U-APSD Flags: 1b per AC. Encoded as follows:
1357 b7 b6 b5 b4 b3 b2 b1 b0 =
1358 X X X X BE BK VI VO */
1359 u8 uapsd;
1360
1361 /* Max SP Length */
1362 u8 max_sp_len;
1363
1364 /* 11n Green Field preamble support
1365 0 - Not supported, 1 - Supported */
1366 u8 green_field_capable;
1367
1368 /* MIMO Power Save mode */
1369 enum wcn36xx_hal_ht_mimo_state mimo_ps;
1370
1371 /* Delayed BA Support */
1372 u8 delayed_ba_support;
1373
1374 /* Max AMPDU duration in 32us */
1375 u8 max_ampdu_duration;
1376
1377 /* HT STA should set it to 1 if it is enabled in BSS. HT STA should
1378 * set it to 0 if AP does not support it. This indication is sent
1379 * to HAL and HAL uses this flag to pickup up appropriate 40Mhz
1380 * rates. */
1381 u8 dsss_cck_mode_40mhz;
1382
1383 /* Valid STA Idx when action=Update. Set to 0xFF when invalid!
1384 * Retained for backward compalibity with existing HAL code */
1385 u8 sta_index;
1386
1387 /* BSSID of BSS to which station is associated. Set to 0xFF when
1388 * invalid. Retained for backward compalibity with existing HAL
1389 * code */
1390 u8 bssid_index;
1391
1392 u8 p2p;
1393
1394 /* Reserved to align next field on a dword boundary */
1395 u8 reserved;
1396
1397 /* These rates are the intersection of peer and self capabilities. */
1398 struct wcn36xx_hal_supported_rates supported_rates;
1399} __packed;
1400
1401struct wcn36xx_hal_config_sta_req_msg_v1 {
1402 struct wcn36xx_hal_msg_header header;
1403 struct wcn36xx_hal_config_sta_params_v1 sta_params;
1404} __packed;
1405
1406struct config_sta_rsp_params {
1407 /* success or failure */
1408 u32 status;
1409
1410 /* Station index; valid only when 'status' field value SUCCESS */
1411 u8 sta_index;
1412
1413 /* BSSID Index of BSS to which the station is associated */
1414 u8 bssid_index;
1415
1416 /* DPU Index for PTK */
1417 u8 dpu_index;
1418
1419 /* DPU Index for GTK */
1420 u8 bcast_dpu_index;
1421
1422 /* DPU Index for IGTK */
1423 u8 bcast_mgmt_dpu_idx;
1424
1425 /* PTK DPU signature */
1426 u8 uc_ucast_sig;
1427
1428 /* GTK DPU isignature */
1429 u8 uc_bcast_sig;
1430
1431 /* IGTK DPU signature */
1432 u8 uc_mgmt_sig;
1433
1434 u8 p2p;
1435
1436} __packed;
1437
1438struct wcn36xx_hal_config_sta_rsp_msg {
1439 struct wcn36xx_hal_msg_header header;
1440
1441 struct config_sta_rsp_params params;
1442} __packed;
1443
1444/* Delete STA Request message */
1445struct wcn36xx_hal_delete_sta_req_msg {
1446 struct wcn36xx_hal_msg_header header;
1447
1448 /* Index of STA to delete */
1449 u8 sta_index;
1450
1451} __packed;
1452
1453/* Delete STA Response message */
1454struct wcn36xx_hal_delete_sta_rsp_msg {
1455 struct wcn36xx_hal_msg_header header;
1456
1457 /* success or failure */
1458 u32 status;
1459
1460 /* Index of STA deleted */
1461 u8 sta_id;
1462} __packed;
1463
1464/* 12 Bytes long because this structure can be used to represent rate and
1465 * extended rate set IEs. The parser assume this to be at least 12 */
1466struct wcn36xx_hal_rate_set {
1467 u8 num_rates;
1468 u8 rate[WCN36XX_HAL_MAC_RATESET_EID_MAX];
1469} __packed;
1470
1471/* access category record */
1472struct wcn36xx_hal_aci_aifsn {
1473#ifndef ANI_LITTLE_BIT_ENDIAN
1474 u8 rsvd:1;
1475 u8 aci:2;
1476 u8 acm:1;
1477 u8 aifsn:4;
1478#else
1479 u8 aifsn:4;
1480 u8 acm:1;
1481 u8 aci:2;
1482 u8 rsvd:1;
1483#endif
1484} __packed;
1485
1486/* contention window size */
1487struct wcn36xx_hal_mac_cw {
1488#ifndef ANI_LITTLE_BIT_ENDIAN
1489 u8 max:4;
1490 u8 min:4;
1491#else
1492 u8 min:4;
1493 u8 max:4;
1494#endif
1495} __packed;
1496
1497struct wcn36xx_hal_edca_param_record {
1498 struct wcn36xx_hal_aci_aifsn aci;
1499 struct wcn36xx_hal_mac_cw cw;
1500 u16 txop_limit;
1501} __packed;
1502
1503struct wcn36xx_hal_mac_ssid {
1504 u8 length;
1505 u8 ssid[32];
1506} __packed;
1507
1508/* Concurrency role. These are generic IDs that identify the various roles
1509 * in the software system. */
1510enum wcn36xx_hal_con_mode {
1511 WCN36XX_HAL_STA_MODE = 0,
1512
1513 /* to support softAp mode . This is misleading.
1514 It means AP MODE only. */
1515 WCN36XX_HAL_STA_SAP_MODE = 1,
1516
1517 WCN36XX_HAL_P2P_CLIENT_MODE,
1518 WCN36XX_HAL_P2P_GO_MODE,
1519 WCN36XX_HAL_MONITOR_MODE,
1520};
1521
1522/* This is a bit pattern to be set for each mode
1523 * bit 0 - sta mode
1524 * bit 1 - ap mode
1525 * bit 2 - p2p client mode
1526 * bit 3 - p2p go mode */
1527enum wcn36xx_hal_concurrency_mode {
1528 HAL_STA = 1,
1529 HAL_SAP = 2,
1530
1531 /* to support sta, softAp mode . This means STA+AP mode */
1532 HAL_STA_SAP = 3,
1533
1534 HAL_P2P_CLIENT = 4,
1535 HAL_P2P_GO = 8,
1536 HAL_MAX_CONCURRENCY_PERSONA = 4
1537};
1538
1539struct wcn36xx_hal_config_bss_params {
1540 /* BSSID */
1541 u8 bssid[ETH_ALEN];
1542
1543 /* Self Mac Address */
1544 u8 self_mac_addr[ETH_ALEN];
1545
1546 /* BSS type */
1547 enum wcn36xx_hal_bss_type bss_type;
1548
1549 /* Operational Mode: AP =0, STA = 1 */
1550 u8 oper_mode;
1551
1552 /* Network Type */
1553 enum wcn36xx_hal_nw_type nw_type;
1554
1555 /* Used to classify PURE_11G/11G_MIXED to program MTU */
1556 u8 short_slot_time_supported;
1557
1558 /* Co-exist with 11a STA */
1559 u8 lla_coexist;
1560
1561 /* Co-exist with 11b STA */
1562 u8 llb_coexist;
1563
1564 /* Co-exist with 11g STA */
1565 u8 llg_coexist;
1566
1567 /* Coexistence with 11n STA */
1568 u8 ht20_coexist;
1569
1570 /* Non GF coexist flag */
1571 u8 lln_non_gf_coexist;
1572
1573 /* TXOP protection support */
1574 u8 lsig_tx_op_protection_full_support;
1575
1576 /* RIFS mode */
1577 u8 rifs_mode;
1578
1579 /* Beacon Interval in TU */
1580 u16 beacon_interval;
1581
1582 /* DTIM period */
1583 u8 dtim_period;
1584
1585 /* TX Width Set: 0 - 20 MHz only, 1 - 20/40 MHz */
1586 u8 tx_channel_width_set;
1587
1588 /* Operating channel */
1589 u8 oper_channel;
1590
1591 /* Extension channel for channel bonding */
1592 u8 ext_channel;
1593
1594 /* Reserved to align next field on a dword boundary */
1595 u8 reserved;
1596
1597 /* TODO move sta to the end for 3680 */
1598 /* Context of the station being added in HW
1599 * Add a STA entry for "itself" -
1600 *
1601 * On AP - Add the AP itself in an "STA context"
1602 *
1603 * On STA - Add the AP to which this STA is joining in an
1604 * "STA context"
1605 */
1606 struct wcn36xx_hal_config_sta_params sta;
1607 /* SSID of the BSS */
1608 struct wcn36xx_hal_mac_ssid ssid;
1609
1610 /* HAL should update the existing BSS entry, if this flag is set.
1611 * UMAC will set this flag in case of reassoc, where we want to
1612 * resue the the old BSSID and still return success 0 = Add, 1 =
1613 * Update */
1614 u8 action;
1615
1616 /* MAC Rate Set */
1617 struct wcn36xx_hal_rate_set rateset;
1618
1619 /* Enable/Disable HT capabilities of the BSS */
1620 u8 ht;
1621
1622 /* Enable/Disable OBSS protection */
1623 u8 obss_prot_enabled;
1624
1625 /* RMF enabled/disabled */
1626 u8 rmf;
1627
1628 /* HT Operating Mode operating mode of the 802.11n STA */
1629 enum wcn36xx_hal_ht_operating_mode ht_oper_mode;
1630
1631 /* Dual CTS Protection: 0 - Unused, 1 - Used */
1632 u8 dual_cts_protection;
1633
1634 /* Probe Response Max retries */
1635 u8 max_probe_resp_retry_limit;
1636
1637 /* To Enable Hidden ssid */
1638 u8 hidden_ssid;
1639
1640 /* To Enable Disable FW Proxy Probe Resp */
1641 u8 proxy_probe_resp;
1642
1643 /* Boolean to indicate if EDCA params are valid. UMAC might not
1644 * have valid EDCA params or might not desire to apply EDCA params
1645 * during config BSS. 0 implies Not Valid ; Non-Zero implies
1646 * valid */
1647 u8 edca_params_valid;
1648
1649 /* EDCA Parameters for Best Effort Access Category */
1650 struct wcn36xx_hal_edca_param_record acbe;
1651
1652 /* EDCA Parameters forBackground Access Category */
1653 struct wcn36xx_hal_edca_param_record acbk;
1654
1655 /* EDCA Parameters for Video Access Category */
1656 struct wcn36xx_hal_edca_param_record acvi;
1657
1658 /* EDCA Parameters for Voice Access Category */
1659 struct wcn36xx_hal_edca_param_record acvo;
1660
1661 /* Ext Bss Config Msg if set */
1662 u8 ext_set_sta_key_param_valid;
1663
1664 /* SetStaKeyParams for ext bss msg */
1665 struct wcn36xx_hal_set_sta_key_params ext_set_sta_key_param;
1666
1667 /* Persona for the BSS can be STA,AP,GO,CLIENT value same as enum
1668 * wcn36xx_hal_con_mode */
1669 u8 wcn36xx_hal_persona;
1670
1671 u8 spectrum_mgt_enable;
1672
1673 /* HAL fills in the tx power used for mgmt frames in txMgmtPower */
1674 s8 tx_mgmt_power;
1675
1676 /* maxTxPower has max power to be used after applying the power
1677 * constraint if any */
1678 s8 max_tx_power;
1679} __packed;
1680
1681struct wcn36xx_hal_config_bss_req_msg {
1682 struct wcn36xx_hal_msg_header header;
1683 struct wcn36xx_hal_config_bss_params bss_params;
1684} __packed;
1685
1686struct wcn36xx_hal_config_bss_params_v1 {
1687 /* BSSID */
1688 u8 bssid[ETH_ALEN];
1689
1690 /* Self Mac Address */
1691 u8 self_mac_addr[ETH_ALEN];
1692
1693 /* BSS type */
1694 enum wcn36xx_hal_bss_type bss_type;
1695
1696 /* Operational Mode: AP =0, STA = 1 */
1697 u8 oper_mode;
1698
1699 /* Network Type */
1700 enum wcn36xx_hal_nw_type nw_type;
1701
1702 /* Used to classify PURE_11G/11G_MIXED to program MTU */
1703 u8 short_slot_time_supported;
1704
1705 /* Co-exist with 11a STA */
1706 u8 lla_coexist;
1707
1708 /* Co-exist with 11b STA */
1709 u8 llb_coexist;
1710
1711 /* Co-exist with 11g STA */
1712 u8 llg_coexist;
1713
1714 /* Coexistence with 11n STA */
1715 u8 ht20_coexist;
1716
1717 /* Non GF coexist flag */
1718 u8 lln_non_gf_coexist;
1719
1720 /* TXOP protection support */
1721 u8 lsig_tx_op_protection_full_support;
1722
1723 /* RIFS mode */
1724 u8 rifs_mode;
1725
1726 /* Beacon Interval in TU */
1727 u16 beacon_interval;
1728
1729 /* DTIM period */
1730 u8 dtim_period;
1731
1732 /* TX Width Set: 0 - 20 MHz only, 1 - 20/40 MHz */
1733 u8 tx_channel_width_set;
1734
1735 /* Operating channel */
1736 u8 oper_channel;
1737
1738 /* Extension channel for channel bonding */
1739 u8 ext_channel;
1740
1741 /* Reserved to align next field on a dword boundary */
1742 u8 reserved;
1743
1744 /* SSID of the BSS */
1745 struct wcn36xx_hal_mac_ssid ssid;
1746
1747 /* HAL should update the existing BSS entry, if this flag is set.
1748 * UMAC will set this flag in case of reassoc, where we want to
1749 * resue the the old BSSID and still return success 0 = Add, 1 =
1750 * Update */
1751 u8 action;
1752
1753 /* MAC Rate Set */
1754 struct wcn36xx_hal_rate_set rateset;
1755
1756 /* Enable/Disable HT capabilities of the BSS */
1757 u8 ht;
1758
1759 /* Enable/Disable OBSS protection */
1760 u8 obss_prot_enabled;
1761
1762 /* RMF enabled/disabled */
1763 u8 rmf;
1764
1765 /* HT Operating Mode operating mode of the 802.11n STA */
1766 enum wcn36xx_hal_ht_operating_mode ht_oper_mode;
1767
1768 /* Dual CTS Protection: 0 - Unused, 1 - Used */
1769 u8 dual_cts_protection;
1770
1771 /* Probe Response Max retries */
1772 u8 max_probe_resp_retry_limit;
1773
1774 /* To Enable Hidden ssid */
1775 u8 hidden_ssid;
1776
1777 /* To Enable Disable FW Proxy Probe Resp */
1778 u8 proxy_probe_resp;
1779
1780 /* Boolean to indicate if EDCA params are valid. UMAC might not
1781 * have valid EDCA params or might not desire to apply EDCA params
1782 * during config BSS. 0 implies Not Valid ; Non-Zero implies
1783 * valid */
1784 u8 edca_params_valid;
1785
1786 /* EDCA Parameters for Best Effort Access Category */
1787 struct wcn36xx_hal_edca_param_record acbe;
1788
1789 /* EDCA Parameters forBackground Access Category */
1790 struct wcn36xx_hal_edca_param_record acbk;
1791
1792 /* EDCA Parameters for Video Access Category */
1793 struct wcn36xx_hal_edca_param_record acvi;
1794
1795 /* EDCA Parameters for Voice Access Category */
1796 struct wcn36xx_hal_edca_param_record acvo;
1797
1798 /* Ext Bss Config Msg if set */
1799 u8 ext_set_sta_key_param_valid;
1800
1801 /* SetStaKeyParams for ext bss msg */
1802 struct wcn36xx_hal_set_sta_key_params ext_set_sta_key_param;
1803
1804 /* Persona for the BSS can be STA,AP,GO,CLIENT value same as enum
1805 * wcn36xx_hal_con_mode */
1806 u8 wcn36xx_hal_persona;
1807
1808 u8 spectrum_mgt_enable;
1809
1810 /* HAL fills in the tx power used for mgmt frames in txMgmtPower */
1811 s8 tx_mgmt_power;
1812
1813 /* maxTxPower has max power to be used after applying the power
1814 * constraint if any */
1815 s8 max_tx_power;
1816
1817 /* Context of the station being added in HW
1818 * Add a STA entry for "itself" -
1819 *
1820 * On AP - Add the AP itself in an "STA context"
1821 *
1822 * On STA - Add the AP to which this STA is joining in an
1823 * "STA context"
1824 */
1825 struct wcn36xx_hal_config_sta_params_v1 sta;
1826} __packed;
1827
1828struct wcn36xx_hal_config_bss_req_msg_v1 {
1829 struct wcn36xx_hal_msg_header header;
1830 struct wcn36xx_hal_config_bss_params_v1 bss_params;
1831} __packed;
1832
1833struct wcn36xx_hal_config_bss_rsp_params {
1834 /* Success or Failure */
1835 u32 status;
1836
1837 /* BSS index allocated by HAL */
1838 u8 bss_index;
1839
1840 /* DPU descriptor index for PTK */
1841 u8 dpu_desc_index;
1842
1843 /* PTK DPU signature */
1844 u8 ucast_dpu_signature;
1845
1846 /* DPU descriptor index for GTK */
1847 u8 bcast_dpu_desc_indx;
1848
1849 /* GTK DPU signature */
1850 u8 bcast_dpu_signature;
1851
1852 /* DPU descriptor for IGTK */
1853 u8 mgmt_dpu_desc_index;
1854
1855 /* IGTK DPU signature */
1856 u8 mgmt_dpu_signature;
1857
1858 /* Station Index for BSS entry */
1859 u8 bss_sta_index;
1860
1861 /* Self station index for this BSS */
1862 u8 bss_self_sta_index;
1863
1864 /* Bcast station for buffering bcast frames in AP role */
1865 u8 bss_bcast_sta_idx;
1866
1867 /* MAC Address of STA(PEER/SELF) in staContext of configBSSReq */
1868 u8 mac[ETH_ALEN];
1869
1870 /* HAL fills in the tx power used for mgmt frames in this field. */
1871 s8 tx_mgmt_power;
1872
1873} __packed;
1874
1875struct wcn36xx_hal_config_bss_rsp_msg {
1876 struct wcn36xx_hal_msg_header header;
1877 struct wcn36xx_hal_config_bss_rsp_params bss_rsp_params;
1878} __packed;
1879
1880struct wcn36xx_hal_delete_bss_req_msg {
1881 struct wcn36xx_hal_msg_header header;
1882
1883 /* BSS index to be deleted */
1884 u8 bss_index;
1885
1886} __packed;
1887
1888struct wcn36xx_hal_delete_bss_rsp_msg {
1889 struct wcn36xx_hal_msg_header header;
1890
1891 /* Success or Failure */
1892 u32 status;
1893
1894 /* BSS index that has been deleted */
1895 u8 bss_index;
1896
1897} __packed;
1898
1899struct wcn36xx_hal_join_req_msg {
1900 struct wcn36xx_hal_msg_header header;
1901
1902 /* Indicates the BSSID to which STA is going to associate */
1903 u8 bssid[ETH_ALEN];
1904
1905 /* Indicates the channel to switch to. */
1906 u8 channel;
1907
1908 /* Self STA MAC */
1909 u8 self_sta_mac_addr[ETH_ALEN];
1910
1911 /* Local power constraint */
1912 u8 local_power_constraint;
1913
1914 /* Secondary channel offset */
1915 enum phy_chan_bond_state secondary_channel_offset;
1916
1917 /* link State */
1918 enum wcn36xx_hal_link_state link_state;
1919
1920 /* Max TX power */
1921 s8 max_tx_power;
1922} __packed;
1923
1924struct wcn36xx_hal_join_rsp_msg {
1925 struct wcn36xx_hal_msg_header header;
1926
1927 /* success or failure */
1928 u32 status;
1929
1930 /* HAL fills in the tx power used for mgmt frames in this field */
1931 u8 tx_mgmt_power;
1932} __packed;
1933
1934struct post_assoc_req_msg {
1935 struct wcn36xx_hal_msg_header header;
1936
1937 struct wcn36xx_hal_config_sta_params sta_params;
1938 struct wcn36xx_hal_config_bss_params bss_params;
1939};
1940
1941struct post_assoc_rsp_msg {
1942 struct wcn36xx_hal_msg_header header;
1943 struct config_sta_rsp_params sta_rsp_params;
1944 struct wcn36xx_hal_config_bss_rsp_params bss_rsp_params;
1945};
1946
1947/* This is used to create a set of WEP keys for a given BSS. */
1948struct wcn36xx_hal_set_bss_key_req_msg {
1949 struct wcn36xx_hal_msg_header header;
1950
1951 /* BSS Index of the BSS */
1952 u8 bss_idx;
1953
1954 /* Encryption Type used with peer */
1955 enum ani_ed_type enc_type;
1956
1957 /* Number of keys */
1958 u8 num_keys;
1959
1960 /* Array of keys. */
1961 struct wcn36xx_hal_keys keys[WCN36XX_HAL_MAC_MAX_NUM_OF_DEFAULT_KEYS];
1962
1963 /* Control for Replay Count, 1= Single TID based replay count on Tx
1964 * 0 = Per TID based replay count on TX */
1965 u8 single_tid_rc;
1966} __packed;
1967
1968/* tagged version of set bss key */
1969struct wcn36xx_hal_set_bss_key_req_msg_tagged {
1970 struct wcn36xx_hal_set_bss_key_req_msg Msg;
1971 u32 tag;
1972} __packed;
1973
1974struct wcn36xx_hal_set_bss_key_rsp_msg {
1975 struct wcn36xx_hal_msg_header header;
1976
1977 /* success or failure */
1978 u32 status;
1979} __packed;
1980
1981/*
1982 * This is used configure the key information on a given station.
1983 * When the sec_type is WEP40 or WEP104, the def_wep_idx is used to locate
1984 * a preconfigured key from a BSS the station assoicated with; otherwise
1985 * a new key descriptor is created based on the key field.
1986 */
1987struct wcn36xx_hal_set_sta_key_req_msg {
1988 struct wcn36xx_hal_msg_header header;
1989 struct wcn36xx_hal_set_sta_key_params set_sta_key_params;
1990} __packed;
1991
1992struct wcn36xx_hal_set_sta_key_rsp_msg {
1993 struct wcn36xx_hal_msg_header header;
1994
1995 /* success or failure */
1996 u32 status;
1997} __packed;
1998
1999struct wcn36xx_hal_remove_bss_key_req_msg {
2000 struct wcn36xx_hal_msg_header header;
2001
2002 /* BSS Index of the BSS */
2003 u8 bss_idx;
2004
2005 /* Encryption Type used with peer */
2006 enum ani_ed_type enc_type;
2007
2008 /* Key Id */
2009 u8 key_id;
2010
2011 /* STATIC/DYNAMIC. Used in Nullifying in Key Descriptors for
2012 * Static/Dynamic keys */
2013 enum ani_wep_type wep_type;
2014} __packed;
2015
2016struct wcn36xx_hal_remove_bss_key_rsp_msg {
2017 struct wcn36xx_hal_msg_header header;
2018
2019 /* success or failure */
2020 u32 status;
2021} __packed;
2022
2023/*
2024 * This is used by PE to Remove the key information on a given station.
2025 */
2026struct wcn36xx_hal_remove_sta_key_req_msg {
2027 struct wcn36xx_hal_msg_header header;
2028
2029 /* STA Index */
2030 u16 sta_idx;
2031
2032 /* Encryption Type used with peer */
2033 enum ani_ed_type enc_type;
2034
2035 /* Key Id */
2036 u8 key_id;
2037
2038 /* Whether to invalidate the Broadcast key or Unicast key. In case
2039 * of WEP, the same key is used for both broadcast and unicast. */
2040 u8 unicast;
2041
2042} __packed;
2043
2044struct wcn36xx_hal_remove_sta_key_rsp_msg {
2045 struct wcn36xx_hal_msg_header header;
2046
2047 /*success or failure */
2048 u32 status;
2049
2050} __packed;
2051
2052#ifdef FEATURE_OEM_DATA_SUPPORT
2053
2054#ifndef OEM_DATA_REQ_SIZE
2055#define OEM_DATA_REQ_SIZE 134
2056#endif
2057
2058#ifndef OEM_DATA_RSP_SIZE
2059#define OEM_DATA_RSP_SIZE 1968
2060#endif
2061
2062struct start_oem_data_req_msg {
2063 struct wcn36xx_hal_msg_header header;
2064
2065 u32 status;
2066 tSirMacAddr self_mac_addr;
2067 u8 oem_data_req[OEM_DATA_REQ_SIZE];
2068
2069};
2070
2071struct start_oem_data_rsp_msg {
2072 struct wcn36xx_hal_msg_header header;
2073
2074 u8 oem_data_rsp[OEM_DATA_RSP_SIZE];
2075};
2076
2077#endif
2078
2079struct wcn36xx_hal_switch_channel_req_msg {
2080 struct wcn36xx_hal_msg_header header;
2081
2082 /* Channel number */
2083 u8 channel_number;
2084
2085 /* Local power constraint */
2086 u8 local_power_constraint;
2087
2088 /* Secondary channel offset */
2089 enum phy_chan_bond_state secondary_channel_offset;
2090
2091 /* HAL fills in the tx power used for mgmt frames in this field. */
2092 u8 tx_mgmt_power;
2093
2094 /* Max TX power */
2095 u8 max_tx_power;
2096
2097 /* Self STA MAC */
2098 u8 self_sta_mac_addr[ETH_ALEN];
2099
2100 /* VO WIFI comment: BSSID needed to identify session. As the
2101 * request has power constraints, this should be applied only to
2102 * that session Since MTU timing and EDCA are sessionized, this
2103 * struct needs to be sessionized and bssid needs to be out of the
2104 * VOWifi feature flag V IMP: Keep bssId field at the end of this
2105 * msg. It is used to mantain backward compatbility by way of
2106 * ignoring if using new host/old FW or old host/new FW since it is
2107 * at the end of this struct
2108 */
2109 u8 bssid[ETH_ALEN];
2110} __packed;
2111
2112struct wcn36xx_hal_switch_channel_rsp_msg {
2113 struct wcn36xx_hal_msg_header header;
2114
2115 /* Status */
2116 u32 status;
2117
2118 /* Channel number - same as in request */
2119 u8 channel_number;
2120
2121 /* HAL fills in the tx power used for mgmt frames in this field */
2122 u8 tx_mgmt_power;
2123
2124 /* BSSID needed to identify session - same as in request */
2125 u8 bssid[ETH_ALEN];
2126
2127} __packed;
2128
2129struct update_edca_params_req_msg {
2130 struct wcn36xx_hal_msg_header header;
2131
2132 /*BSS Index */
2133 u16 bss_index;
2134
2135 /* Best Effort */
2136 struct wcn36xx_hal_edca_param_record acbe;
2137
2138 /* Background */
2139 struct wcn36xx_hal_edca_param_record acbk;
2140
2141 /* Video */
2142 struct wcn36xx_hal_edca_param_record acvi;
2143
2144 /* Voice */
2145 struct wcn36xx_hal_edca_param_record acvo;
2146};
2147
2148struct update_edca_params_rsp_msg {
2149 struct wcn36xx_hal_msg_header header;
2150
2151 /* success or failure */
2152 u32 status;
2153};
2154
2155struct dpu_stats_params {
2156 /* Index of STA to which the statistics */
2157 u16 sta_index;
2158
2159 /* Encryption mode */
2160 u8 enc_mode;
2161
2162 /* status */
2163 u32 status;
2164
2165 /* Statistics */
2166 u32 send_blocks;
2167 u32 recv_blocks;
2168 u32 replays;
2169 u8 mic_error_cnt;
2170 u32 prot_excl_cnt;
2171 u16 format_err_cnt;
2172 u16 un_decryptable_cnt;
2173 u32 decrypt_err_cnt;
2174 u32 decrypt_ok_cnt;
2175};
2176
2177struct wcn36xx_hal_stats_req_msg {
2178 struct wcn36xx_hal_msg_header header;
2179
2180 /* Valid STA Idx for per STA stats request */
2181 u32 sta_id;
2182
2183 /* Categories of stats requested as specified in eHalStatsMask */
2184 u32 stats_mask;
2185};
2186
2187struct ani_summary_stats_info {
2188 /* Total number of packets(per AC) that were successfully
2189 * transmitted with retries */
2190 u32 retry_cnt[4];
2191
2192 /* The number of MSDU packets and MMPDU frames per AC that the
2193 * 802.11 station successfully transmitted after more than one
2194 * retransmission attempt */
2195 u32 multiple_retry_cnt[4];
2196
2197 /* Total number of packets(per AC) that were successfully
2198 * transmitted (with and without retries, including multi-cast,
2199 * broadcast) */
2200 u32 tx_frm_cnt[4];
2201
2202 /* Total number of packets that were successfully received (after
2203 * appropriate filter rules including multi-cast, broadcast) */
2204 u32 rx_frm_cnt;
2205
2206 /* Total number of duplicate frames received successfully */
2207 u32 frm_dup_cnt;
2208
2209 /* Total number packets(per AC) failed to transmit */
2210 u32 fail_cnt[4];
2211
2212 /* Total number of RTS/CTS sequence failures for transmission of a
2213 * packet */
2214 u32 rts_fail_cnt;
2215
2216 /* Total number packets failed transmit because of no ACK from the
2217 * remote entity */
2218 u32 ack_fail_cnt;
2219
2220 /* Total number of RTS/CTS sequence success for transmission of a
2221 * packet */
2222 u32 rts_succ_cnt;
2223
2224 /* The sum of the receive error count and dropped-receive-buffer
2225 * error count. HAL will provide this as a sum of (FCS error) +
2226 * (Fail get BD/PDU in HW) */
2227 u32 rx_discard_cnt;
2228
2229 /*
2230 * The receive error count. HAL will provide the RxP FCS error
2231 * global counter. */
2232 u32 rx_error_cnt;
2233
2234 /* The sum of the transmit-directed byte count, transmit-multicast
2235 * byte count and transmit-broadcast byte count. HAL will sum TPE
2236 * UC/MC/BCAST global counters to provide this. */
2237 u32 tx_byte_cnt;
2238};
2239
2240/* defines tx_rate_flags */
2241enum tx_rate_info {
2242 /* Legacy rates */
2243 HAL_TX_RATE_LEGACY = 0x1,
2244
2245 /* HT20 rates */
2246 HAL_TX_RATE_HT20 = 0x2,
2247
2248 /* HT40 rates */
2249 HAL_TX_RATE_HT40 = 0x4,
2250
2251 /* Rate with Short guard interval */
2252 HAL_TX_RATE_SGI = 0x8,
2253
2254 /* Rate with Long guard interval */
2255 HAL_TX_RATE_LGI = 0x10
2256};
2257
2258struct ani_global_class_a_stats_info {
2259 /* The number of MPDU frames received by the 802.11 station for
2260 * MSDU packets or MMPDU frames */
2261 u32 rx_frag_cnt;
2262
2263 /* The number of MPDU frames received by the 802.11 station for
2264 * MSDU packets or MMPDU frames when a promiscuous packet filter
2265 * was enabled */
2266 u32 promiscuous_rx_frag_cnt;
2267
2268 /* The receiver input sensitivity referenced to a FER of 8% at an
2269 * MPDU length of 1024 bytes at the antenna connector. Each element
2270 * of the array shall correspond to a supported rate and the order
2271 * shall be the same as the supporteRates parameter. */
2272 u32 rx_input_sensitivity;
2273
2274 /* The maximum transmit power in dBm upto one decimal. for eg: if
2275 * it is 10.5dBm, the value would be 105 */
2276 u32 max_pwr;
2277
2278 /* Number of times the receiver failed to synchronize with the
2279 * incoming signal after detecting the sync in the preamble of the
2280 * transmitted PLCP protocol data unit. */
2281 u32 sync_fail_cnt;
2282
2283 /* Legacy transmit rate, in units of 500 kbit/sec, for the most
2284 * recently transmitted frame */
2285 u32 tx_rate;
2286
2287 /* mcs index for HT20 and HT40 rates */
2288 u32 mcs_index;
2289
2290 /* to differentiate between HT20 and HT40 rates; short and long
2291 * guard interval */
2292 u32 tx_rate_flags;
2293};
2294
2295struct ani_global_security_stats {
2296 /* The number of unencrypted received MPDU frames that the MAC
2297 * layer discarded when the IEEE 802.11 dot11ExcludeUnencrypted
2298 * management information base (MIB) object is enabled */
2299 u32 rx_wep_unencrypted_frm_cnt;
2300
2301 /* The number of received MSDU packets that that the 802.11 station
2302 * discarded because of MIC failures */
2303 u32 rx_mic_fail_cnt;
2304
2305 /* The number of encrypted MPDU frames that the 802.11 station
2306 * failed to decrypt because of a TKIP ICV error */
2307 u32 tkip_icv_err;
2308
2309 /* The number of received MPDU frames that the 802.11 discarded
2310 * because of an invalid AES-CCMP format */
2311 u32 aes_ccmp_format_err;
2312
2313 /* The number of received MPDU frames that the 802.11 station
2314 * discarded because of the AES-CCMP replay protection procedure */
2315 u32 aes_ccmp_replay_cnt;
2316
2317 /* The number of received MPDU frames that the 802.11 station
2318 * discarded because of errors detected by the AES-CCMP decryption
2319 * algorithm */
2320 u32 aes_ccmp_decrpt_err;
2321
2322 /* The number of encrypted MPDU frames received for which a WEP
2323 * decryption key was not available on the 802.11 station */
2324 u32 wep_undecryptable_cnt;
2325
2326 /* The number of encrypted MPDU frames that the 802.11 station
2327 * failed to decrypt because of a WEP ICV error */
2328 u32 wep_icv_err;
2329
2330 /* The number of received encrypted packets that the 802.11 station
2331 * successfully decrypted */
2332 u32 rx_decrypt_succ_cnt;
2333
2334 /* The number of encrypted packets that the 802.11 station failed
2335 * to decrypt */
2336 u32 rx_decrypt_fail_cnt;
2337};
2338
2339struct ani_global_class_b_stats_info {
2340 struct ani_global_security_stats uc_stats;
2341 struct ani_global_security_stats mc_bc_stats;
2342};
2343
2344struct ani_global_class_c_stats_info {
2345 /* This counter shall be incremented for a received A-MSDU frame
2346 * with the stations MAC address in the address 1 field or an
2347 * A-MSDU frame with a group address in the address 1 field */
2348 u32 rx_amsdu_cnt;
2349
2350 /* This counter shall be incremented when the MAC receives an AMPDU
2351 * from the PHY */
2352 u32 rx_ampdu_cnt;
2353
2354 /* This counter shall be incremented when a Frame is transmitted
2355 * only on the primary channel */
2356 u32 tx_20_frm_cnt;
2357
2358 /* This counter shall be incremented when a Frame is received only
2359 * on the primary channel */
2360 u32 rx_20_frm_cnt;
2361
2362 /* This counter shall be incremented by the number of MPDUs
2363 * received in the A-MPDU when an A-MPDU is received */
2364 u32 rx_mpdu_in_ampdu_cnt;
2365
2366 /* This counter shall be incremented when an MPDU delimiter has a
2367 * CRC error when this is the first CRC error in the received AMPDU
2368 * or when the previous delimiter has been decoded correctly */
2369 u32 ampdu_delimiter_crc_err;
2370};
2371
2372struct ani_per_sta_stats_info {
2373 /* The number of MPDU frames that the 802.11 station transmitted
2374 * and acknowledged through a received 802.11 ACK frame */
2375 u32 tx_frag_cnt[4];
2376
2377 /* This counter shall be incremented when an A-MPDU is transmitted */
2378 u32 tx_ampdu_cnt;
2379
2380 /* This counter shall increment by the number of MPDUs in the AMPDU
2381 * when an A-MPDU is transmitted */
2382 u32 tx_mpdu_in_ampdu_cnt;
2383};
2384
2385struct wcn36xx_hal_stats_rsp_msg {
2386 struct wcn36xx_hal_msg_header header;
2387
2388 /* Success or Failure */
2389 u32 status;
2390
2391 /* STA Idx */
2392 u32 sta_index;
2393
2394 /* Categories of STATS being returned as per eHalStatsMask */
2395 u32 stats_mask;
2396
2397 /* message type is same as the request type */
2398 u16 msg_type;
2399
2400 /* length of the entire request, includes the pStatsBuf length too */
2401 u16 msg_len;
2402};
2403
2404struct wcn36xx_hal_set_link_state_req_msg {
2405 struct wcn36xx_hal_msg_header header;
2406
2407 u8 bssid[ETH_ALEN];
2408 enum wcn36xx_hal_link_state state;
2409 u8 self_mac_addr[ETH_ALEN];
2410
2411} __packed;
2412
2413struct set_link_state_rsp_msg {
2414 struct wcn36xx_hal_msg_header header;
2415
2416 /* success or failure */
2417 u32 status;
2418};
2419
2420/* TSPEC Params */
2421struct wcn36xx_hal_ts_info_tfc {
2422#ifndef ANI_LITTLE_BIT_ENDIAN
2423 u16 ackPolicy:2;
2424 u16 userPrio:3;
2425 u16 psb:1;
2426 u16 aggregation:1;
2427 u16 accessPolicy:2;
2428 u16 direction:2;
2429 u16 tsid:4;
2430 u16 trafficType:1;
2431#else
2432 u16 trafficType:1;
2433 u16 tsid:4;
2434 u16 direction:2;
2435 u16 accessPolicy:2;
2436 u16 aggregation:1;
2437 u16 psb:1;
2438 u16 userPrio:3;
2439 u16 ackPolicy:2;
2440#endif
2441};
2442
2443/* Flag to schedule the traffic type */
2444struct wcn36xx_hal_ts_info_sch {
2445#ifndef ANI_LITTLE_BIT_ENDIAN
2446 u8 rsvd:7;
2447 u8 schedule:1;
2448#else
2449 u8 schedule:1;
2450 u8 rsvd:7;
2451#endif
2452};
2453
2454/* Traffic and scheduling info */
2455struct wcn36xx_hal_ts_info {
2456 struct wcn36xx_hal_ts_info_tfc traffic;
2457 struct wcn36xx_hal_ts_info_sch schedule;
2458};
2459
2460/* Information elements */
2461struct wcn36xx_hal_tspec_ie {
2462 u8 type;
2463 u8 length;
2464 struct wcn36xx_hal_ts_info ts_info;
2465 u16 nom_msdu_size;
2466 u16 max_msdu_size;
2467 u32 min_svc_interval;
2468 u32 max_svc_interval;
2469 u32 inact_interval;
2470 u32 suspend_interval;
2471 u32 svc_start_time;
2472 u32 min_data_rate;
2473 u32 mean_data_rate;
2474 u32 peak_data_rate;
2475 u32 max_burst_sz;
2476 u32 delay_bound;
2477 u32 min_phy_rate;
2478 u16 surplus_bw;
2479 u16 medium_time;
2480};
2481
2482struct add_ts_req_msg {
2483 struct wcn36xx_hal_msg_header header;
2484
2485 /* Station Index */
2486 u16 sta_index;
2487
2488 /* TSPEC handler uniquely identifying a TSPEC for a STA in a BSS */
2489 u16 tspec_index;
2490
2491 /* To program TPE with required parameters */
2492 struct wcn36xx_hal_tspec_ie tspec;
2493
2494 /* U-APSD Flags: 1b per AC. Encoded as follows:
2495 b7 b6 b5 b4 b3 b2 b1 b0 =
2496 X X X X BE BK VI VO */
2497 u8 uapsd;
2498
2499 /* These parameters are for all the access categories */
2500
2501 /* Service Interval */
2502 u32 service_interval[WCN36XX_HAL_MAX_AC];
2503
2504 /* Suspend Interval */
2505 u32 suspend_interval[WCN36XX_HAL_MAX_AC];
2506
2507 /* Delay Interval */
2508 u32 delay_interval[WCN36XX_HAL_MAX_AC];
2509};
2510
2511struct add_rs_rsp_msg {
2512 struct wcn36xx_hal_msg_header header;
2513
2514 /* success or failure */
2515 u32 status;
2516};
2517
2518struct del_ts_req_msg {
2519 struct wcn36xx_hal_msg_header header;
2520
2521 /* Station Index */
2522 u16 sta_index;
2523
2524 /* TSPEC identifier uniquely identifying a TSPEC for a STA in a BSS */
2525 u16 tspec_index;
2526
2527 /* To lookup station id using the mac address */
2528 u8 bssid[ETH_ALEN];
2529};
2530
2531struct del_ts_rsp_msg {
2532 struct wcn36xx_hal_msg_header header;
2533
2534 /* success or failure */
2535 u32 status;
2536};
2537
2538/* End of TSpec Parameters */
2539
2540/* Start of BLOCK ACK related Parameters */
2541
2542struct wcn36xx_hal_add_ba_session_req_msg {
2543 struct wcn36xx_hal_msg_header header;
2544
2545 /* Station Index */
2546 u16 sta_index;
2547
2548 /* Peer MAC Address */
2549 u8 mac_addr[ETH_ALEN];
2550
2551 /* ADDBA Action Frame dialog token
2552 HAL will not interpret this object */
2553 u8 dialog_token;
2554
2555 /* TID for which the BA is being setup
2556 This identifies the TC or TS of interest */
2557 u8 tid;
2558
2559 /* 0 - Delayed BA (Not supported)
2560 1 - Immediate BA */
2561 u8 policy;
2562
2563 /* Indicates the number of buffers for this TID (baTID)
2564 NOTE - This is the requested buffer size. When this
2565 is processed by HAL and subsequently by HDD, it is
2566 possible that HDD may change this buffer size. Any
2567 change in the buffer size should be noted by PE and
2568 advertized appropriately in the ADDBA response */
2569 u16 buffer_size;
2570
2571 /* BA timeout in TU's 0 means no timeout will occur */
2572 u16 timeout;
2573
2574 /* b0..b3 - Fragment Number - Always set to 0
2575 b4..b15 - Starting Sequence Number of first MSDU
2576 for which this BA is setup */
2577 u16 ssn;
2578
2579 /* ADDBA direction
2580 1 - Originator
2581 0 - Recipient */
2582 u8 direction;
2583} __packed;
2584
2585struct wcn36xx_hal_add_ba_session_rsp_msg {
2586 struct wcn36xx_hal_msg_header header;
2587
2588 /* success or failure */
2589 u32 status;
2590
2591 /* Dialog token */
2592 u8 dialog_token;
2593
2594 /* TID for which the BA session has been setup */
2595 u8 ba_tid;
2596
2597 /* BA Buffer Size allocated for the current BA session */
2598 u8 ba_buffer_size;
2599
2600 u8 ba_session_id;
2601
2602 /* Reordering Window buffer */
2603 u8 win_size;
2604
2605 /* Station Index to id the sta */
2606 u8 sta_index;
2607
2608 /* Starting Sequence Number */
2609 u16 ssn;
2610} __packed;
2611
2612struct wcn36xx_hal_add_ba_req_msg {
2613 struct wcn36xx_hal_msg_header header;
2614
2615 /* Session Id */
2616 u8 session_id;
2617
2618 /* Reorder Window Size */
2619 u8 win_size;
2620/* Old FW 1.2.2.4 does not support this*/
2621#ifdef FEATURE_ON_CHIP_REORDERING
2622 u8 reordering_done_on_chip;
2623#endif
2624} __packed;
2625
2626struct wcn36xx_hal_add_ba_rsp_msg {
2627 struct wcn36xx_hal_msg_header header;
2628
2629 /* success or failure */
2630 u32 status;
2631
2632 /* Dialog token */
2633 u8 dialog_token;
2634} __packed;
2635
2636struct add_ba_info {
2637 u16 ba_enable:1;
2638 u16 starting_seq_num:12;
2639 u16 reserved:3;
2640};
2641
2642struct wcn36xx_hal_trigger_ba_rsp_candidate {
2643 u8 sta_addr[ETH_ALEN];
2644 struct add_ba_info ba_info[STACFG_MAX_TC];
2645} __packed;
2646
2647struct wcn36xx_hal_trigget_ba_req_candidate {
2648 u8 sta_index;
2649 u8 tid_bitmap;
2650} __packed;
2651
2652struct wcn36xx_hal_trigger_ba_req_msg {
2653 struct wcn36xx_hal_msg_header header;
2654
2655 /* Session Id */
2656 u8 session_id;
2657
2658 /* baCandidateCnt is followed by trigger BA
2659 * Candidate List(tTriggerBaCandidate)
2660 */
2661 u16 candidate_cnt;
2662
2663} __packed;
2664
2665struct wcn36xx_hal_trigger_ba_rsp_msg {
2666 struct wcn36xx_hal_msg_header header;
2667
2668 /* TO SUPPORT BT-AMP */
2669 u8 bssid[ETH_ALEN];
2670
2671 /* success or failure */
2672 u32 status;
2673
2674 /* baCandidateCnt is followed by trigger BA
2675 * Rsp Candidate List(tTriggerRspBaCandidate)
2676 */
2677 u16 candidate_cnt;
2678} __packed;
2679
2680struct wcn36xx_hal_del_ba_req_msg {
2681 struct wcn36xx_hal_msg_header header;
2682
2683 /* Station Index */
2684 u16 sta_index;
2685
2686 /* TID for which the BA session is being deleted */
2687 u8 tid;
2688
2689 /* DELBA direction
2690 1 - Originator
2691 0 - Recipient */
2692 u8 direction;
2693} __packed;
2694
2695struct wcn36xx_hal_del_ba_rsp_msg {
2696 struct wcn36xx_hal_msg_header header;
2697
2698 /* success or failure */
2699 u32 status;
2700} __packed;
2701
2702struct tsm_stats_req_msg {
2703 struct wcn36xx_hal_msg_header header;
2704
2705 /* Traffic Id */
2706 u8 tid;
2707
2708 u8 bssid[ETH_ALEN];
2709};
2710
2711struct tsm_stats_rsp_msg {
2712 struct wcn36xx_hal_msg_header header;
2713
2714 /*success or failure */
2715 u32 status;
2716
2717 /* Uplink Packet Queue delay */
2718 u16 uplink_pkt_queue_delay;
2719
2720 /* Uplink Packet Queue delay histogram */
2721 u16 uplink_pkt_queue_delay_hist[4];
2722
2723 /* Uplink Packet Transmit delay */
2724 u32 uplink_pkt_tx_delay;
2725
2726 /* Uplink Packet loss */
2727 u16 uplink_pkt_loss;
2728
2729 /* Uplink Packet count */
2730 u16 uplink_pkt_count;
2731
2732 /* Roaming count */
2733 u8 roaming_count;
2734
2735 /* Roaming Delay */
2736 u16 roaming_delay;
2737};
2738
2739struct set_key_done_msg {
2740 struct wcn36xx_hal_msg_header header;
2741
2742 /*bssid of the keys */
2743 u8 bssidx;
2744 u8 enc_type;
2745};
2746
2747struct wcn36xx_hal_nv_img_download_req_msg {
2748 /* Note: The length specified in wcn36xx_hal_nv_img_download_req_msg
2749 * messages should be
2750 * header.len = sizeof(wcn36xx_hal_nv_img_download_req_msg) +
2751 * nv_img_buffer_size */
2752 struct wcn36xx_hal_msg_header header;
2753
2754 /* Fragment sequence number of the NV Image. Note that NV Image
2755 * might not fit into one message due to size limitation of the SMD
2756 * channel FIFO. UMAC can hence choose to chop the NV blob into
2757 * multiple fragments starting with seqeunce number 0, 1, 2 etc.
2758 * The last fragment MUST be indicated by marking the
2759 * isLastFragment field to 1. Note that all the NV blobs would be
2760 * concatenated together by HAL without any padding bytes in
2761 * between.*/
2762 u16 frag_number;
2763
2764 /* Is this the last fragment? When set to 1 it indicates that no
2765 * more fragments will be sent by UMAC and HAL can concatenate all
2766 * the NV blobs rcvd & proceed with the parsing. HAL would generate
2767 * a WCN36XX_HAL_DOWNLOAD_NV_RSP to the WCN36XX_HAL_DOWNLOAD_NV_REQ
2768 * after it receives each fragment */
2769 u16 last_fragment;
2770
2771 /* NV Image size (number of bytes) */
2772 u32 nv_img_buffer_size;
2773
2774 /* Following the 'nv_img_buffer_size', there should be
2775 * nv_img_buffer_size bytes of NV Image i.e.
2776 * u8[nv_img_buffer_size] */
2777} __packed;
2778
2779struct wcn36xx_hal_nv_img_download_rsp_msg {
2780 struct wcn36xx_hal_msg_header header;
2781
2782 /* Success or Failure. HAL would generate a
2783 * WCN36XX_HAL_DOWNLOAD_NV_RSP after each fragment */
2784 u32 status;
2785} __packed;
2786
2787struct wcn36xx_hal_nv_store_ind {
2788 /* Note: The length specified in tHalNvStoreInd messages should be
2789 * header.msgLen = sizeof(tHalNvStoreInd) + nvBlobSize */
2790 struct wcn36xx_hal_msg_header header;
2791
2792 /* NV Item */
2793 u32 table_id;
2794
2795 /* Size of NV Blob */
2796 u32 nv_blob_size;
2797
2798 /* Following the 'nvBlobSize', there should be nvBlobSize bytes of
2799 * NV blob i.e. u8[nvBlobSize] */
2800};
2801
2802/* End of Block Ack Related Parameters */
2803
2804#define WCN36XX_HAL_CIPHER_SEQ_CTR_SIZE 6
2805
2806/* Definition for MIC failure indication MAC reports this each time a MIC
2807 * failure occures on Rx TKIP packet
2808 */
2809struct mic_failure_ind_msg {
2810 struct wcn36xx_hal_msg_header header;
2811
2812 u8 bssid[ETH_ALEN];
2813
2814 /* address used to compute MIC */
2815 u8 src_addr[ETH_ALEN];
2816
2817 /* transmitter address */
2818 u8 ta_addr[ETH_ALEN];
2819
2820 u8 dst_addr[ETH_ALEN];
2821
2822 u8 multicast;
2823
2824 /* first byte of IV */
2825 u8 iv1;
2826
2827 /* second byte of IV */
2828 u8 key_id;
2829
2830 /* sequence number */
2831 u8 tsc[WCN36XX_HAL_CIPHER_SEQ_CTR_SIZE];
2832
2833 /* receive address */
2834 u8 rx_addr[ETH_ALEN];
2835};
2836
2837struct update_vht_op_mode_req_msg {
2838 struct wcn36xx_hal_msg_header header;
2839
2840 u16 op_mode;
2841 u16 sta_id;
2842};
2843
2844struct update_vht_op_mode_params_rsp_msg {
2845 struct wcn36xx_hal_msg_header header;
2846
2847 u32 status;
2848};
2849
2850struct update_beacon_req_msg {
2851 struct wcn36xx_hal_msg_header header;
2852
2853 u8 bss_index;
2854
2855 /* shortPreamble mode. HAL should update all the STA rates when it
2856 * receives this message */
2857 u8 short_preamble;
2858
2859 /* short Slot time. */
2860 u8 short_slot_time;
2861
2862 /* Beacon Interval */
2863 u16 beacon_interval;
2864
2865 /* Protection related */
2866 u8 lla_coexist;
2867 u8 llb_coexist;
2868 u8 llg_coexist;
2869 u8 ht20_coexist;
2870 u8 lln_non_gf_coexist;
2871 u8 lsig_tx_op_protection_full_support;
2872 u8 rifs_mode;
2873
2874 u16 param_change_bitmap;
2875};
2876
2877struct update_beacon_rsp_msg {
2878 struct wcn36xx_hal_msg_header header;
2879 u32 status;
2880};
2881
2882struct wcn36xx_hal_send_beacon_req_msg {
2883 struct wcn36xx_hal_msg_header header;
2884
2885 /* length of the template. */
2886 u32 beacon_length;
2887
2888 /* Beacon data. */
2889 u8 beacon[BEACON_TEMPLATE_SIZE];
2890
2891 u8 bssid[ETH_ALEN];
2892
2893 /* TIM IE offset from the beginning of the template. */
2894 u32 tim_ie_offset;
2895
2896 /* P2P IE offset from the begining of the template */
2897 u16 p2p_ie_offset;
2898} __packed;
2899
2900struct send_beacon_rsp_msg {
2901 struct wcn36xx_hal_msg_header header;
2902 u32 status;
2903} __packed;
2904
2905struct enable_radar_req_msg {
2906 struct wcn36xx_hal_msg_header header;
2907
2908 u8 bssid[ETH_ALEN];
2909 u8 channel;
2910};
2911
2912struct enable_radar_rsp_msg {
2913 struct wcn36xx_hal_msg_header header;
2914
2915 /* Link Parameters */
2916 u8 bssid[ETH_ALEN];
2917
2918 /* success or failure */
2919 u32 status;
2920};
2921
2922struct radar_detect_intr_ind_msg {
2923 struct wcn36xx_hal_msg_header header;
2924
2925 u8 radar_det_channel;
2926};
2927
2928struct radar_detect_ind_msg {
2929 struct wcn36xx_hal_msg_header header;
2930
2931 /* channel number in which the RADAR detected */
2932 u8 channel_number;
2933
2934 /* RADAR pulse width in usecond */
2935 u16 radar_pulse_width;
2936
2937 /* Number of RADAR pulses */
2938 u16 num_radar_pulse;
2939};
2940
2941struct wcn36xx_hal_get_tpc_report_req_msg {
2942 struct wcn36xx_hal_msg_header header;
2943
2944 u8 sta[ETH_ALEN];
2945 u8 dialog_token;
2946 u8 txpower;
2947};
2948
2949struct wcn36xx_hal_get_tpc_report_rsp_msg {
2950 struct wcn36xx_hal_msg_header header;
2951
2952 /* success or failure */
2953 u32 status;
2954};
2955
2956struct wcn36xx_hal_send_probe_resp_req_msg {
2957 struct wcn36xx_hal_msg_header header;
2958
2959 u8 probe_resp_template[BEACON_TEMPLATE_SIZE];
2960 u32 probe_resp_template_len;
2961 u32 proxy_probe_req_valid_ie_bmap[8];
2962 u8 bssid[ETH_ALEN];
2963};
2964
2965struct send_probe_resp_rsp_msg {
2966 struct wcn36xx_hal_msg_header header;
2967
2968 /* success or failure */
2969 u32 status;
2970};
2971
2972struct send_unknown_frame_rx_ind_msg {
2973 struct wcn36xx_hal_msg_header header;
2974
2975 /* success or failure */
2976 u32 status;
2977};
2978
2979struct wcn36xx_hal_delete_sta_context_ind_msg {
2980 struct wcn36xx_hal_msg_header header;
2981
2982 u16 aid;
2983 u16 sta_id;
2984
2985 /* TO SUPPORT BT-AMP */
2986 u8 bssid[ETH_ALEN];
2987
2988 /* HAL copies bssid from the sta table. */
2989 u8 addr2[ETH_ALEN];
2990
2991 /* To unify the keepalive / unknown A2 / tim-based disa */
2992 u16 reason_code;
2993} __packed;
2994
2995struct indicate_del_sta {
2996 struct wcn36xx_hal_msg_header header;
2997 u8 aid;
2998 u8 sta_index;
2999 u8 bss_index;
3000 u8 reason_code;
3001 u32 status;
3002};
3003
3004struct bt_amp_event_msg {
3005 struct wcn36xx_hal_msg_header header;
3006
3007 enum bt_amp_event_type btAmpEventType;
3008};
3009
3010struct bt_amp_event_rsp {
3011 struct wcn36xx_hal_msg_header header;
3012
3013 /* success or failure */
3014 u32 status;
3015};
3016
3017struct tl_hal_flush_ac_req_msg {
3018 struct wcn36xx_hal_msg_header header;
3019
3020 /* Station Index. originates from HAL */
3021 u8 sta_id;
3022
3023 /* TID for which the transmit queue is being flushed */
3024 u8 tid;
3025};
3026
3027struct tl_hal_flush_ac_rsp_msg {
3028 struct wcn36xx_hal_msg_header header;
3029
3030 /* Station Index. originates from HAL */
3031 u8 sta_id;
3032
3033 /* TID for which the transmit queue is being flushed */
3034 u8 tid;
3035
3036 /* success or failure */
3037 u32 status;
3038};
3039
3040struct wcn36xx_hal_enter_imps_req_msg {
3041 struct wcn36xx_hal_msg_header header;
3042};
3043
3044struct wcn36xx_hal_exit_imps_req {
3045 struct wcn36xx_hal_msg_header header;
3046};
3047
3048struct wcn36xx_hal_enter_bmps_req_msg {
3049 struct wcn36xx_hal_msg_header header;
3050
3051 u8 bss_index;
3052
3053 /* TBTT value derived from the last beacon */
3054#ifndef BUILD_QWPTTSTATIC
3055 u64 tbtt;
3056#endif
3057 u8 dtim_count;
3058
3059 /* DTIM period given to HAL during association may not be valid, if
3060 * association is based on ProbeRsp instead of beacon. */
3061 u8 dtim_period;
3062
3063 /* For CCX and 11R Roaming */
3064 u32 rssi_filter_period;
3065
3066 u32 num_beacon_per_rssi_average;
3067 u8 rssi_filter_enable;
3068} __packed;
3069
3070struct wcn36xx_hal_exit_bmps_req_msg {
3071 struct wcn36xx_hal_msg_header header;
3072
3073 u8 send_data_null;
3074 u8 bss_index;
3075} __packed;
3076
3077struct wcn36xx_hal_missed_beacon_ind_msg {
3078 struct wcn36xx_hal_msg_header header;
3079
3080 u8 bss_index;
3081} __packed;
3082
3083/* Beacon Filtering data structures */
3084
3085/* The above structure would be followed by multiple of below mentioned
3086 * structure
3087 */
3088struct beacon_filter_ie {
3089 u8 element_id;
3090 u8 check_ie_presence;
3091 u8 offset;
3092 u8 value;
3093 u8 bitmask;
3094 u8 ref;
3095};
3096
3097struct wcn36xx_hal_add_bcn_filter_req_msg {
3098 struct wcn36xx_hal_msg_header header;
3099
3100 u16 capability_info;
3101 u16 capability_mask;
3102 u16 beacon_interval;
3103 u16 ie_num;
3104 u8 bss_index;
3105 u8 reserved;
3106};
3107
3108struct wcn36xx_hal_rem_bcn_filter_req {
3109 struct wcn36xx_hal_msg_header header;
3110
3111 u8 ie_Count;
3112 u8 rem_ie_id[1];
3113};
3114
3115#define WCN36XX_HAL_IPV4_ARP_REPLY_OFFLOAD 0
3116#define WCN36XX_HAL_IPV6_NEIGHBOR_DISCOVERY_OFFLOAD 1
3117#define WCN36XX_HAL_IPV6_NS_OFFLOAD 2
3118#define WCN36XX_HAL_IPV6_ADDR_LEN 16
3119#define WCN36XX_HAL_OFFLOAD_DISABLE 0
3120#define WCN36XX_HAL_OFFLOAD_ENABLE 1
3121#define WCN36XX_HAL_OFFLOAD_BCAST_FILTER_ENABLE 0x2
3122#define WCN36XX_HAL_OFFLOAD_ARP_AND_BCAST_FILTER_ENABLE \
3123 (HAL_OFFLOAD_ENABLE|HAL_OFFLOAD_BCAST_FILTER_ENABLE)
3124
3125struct wcn36xx_hal_ns_offload_params {
3126 u8 src_ipv6_addr[WCN36XX_HAL_IPV6_ADDR_LEN];
3127 u8 self_ipv6_addr[WCN36XX_HAL_IPV6_ADDR_LEN];
3128
3129 /* Only support 2 possible Network Advertisement IPv6 address */
3130 u8 target_ipv6_addr1[WCN36XX_HAL_IPV6_ADDR_LEN];
3131 u8 target_ipv6_addr2[WCN36XX_HAL_IPV6_ADDR_LEN];
3132
3133 u8 self_addr[ETH_ALEN];
3134 u8 src_ipv6_addr_valid:1;
3135 u8 target_ipv6_addr1_valid:1;
3136 u8 target_ipv6_addr2_valid:1;
3137 u8 reserved1:5;
3138
3139 /* make it DWORD aligned */
3140 u8 reserved2;
3141
3142 /* slot index for this offload */
3143 u32 slot_index;
3144 u8 bss_index;
3145};
3146
3147struct wcn36xx_hal_host_offload_req {
3148 u8 offload_Type;
3149
3150 /* enable or disable */
3151 u8 enable;
3152
3153 union {
3154 u8 host_ipv4_addr[4];
3155 u8 host_ipv6_addr[WCN36XX_HAL_IPV6_ADDR_LEN];
3156 } u;
3157};
3158
3159struct wcn36xx_hal_host_offload_req_msg {
3160 struct wcn36xx_hal_msg_header header;
3161 struct wcn36xx_hal_host_offload_req host_offload_params;
3162 struct wcn36xx_hal_ns_offload_params ns_offload_params;
3163};
3164
3165/* Packet Types. */
3166#define WCN36XX_HAL_KEEP_ALIVE_NULL_PKT 1
3167#define WCN36XX_HAL_KEEP_ALIVE_UNSOLICIT_ARP_RSP 2
3168
3169/* Enable or disable keep alive */
3170#define WCN36XX_HAL_KEEP_ALIVE_DISABLE 0
3171#define WCN36XX_HAL_KEEP_ALIVE_ENABLE 1
3172#define WCN36XX_KEEP_ALIVE_TIME_PERIOD 30 /* unit: s */
3173
3174/* Keep Alive request. */
3175struct wcn36xx_hal_keep_alive_req_msg {
3176 struct wcn36xx_hal_msg_header header;
3177
3178 u8 packet_type;
3179 u32 time_period;
3180 u8 host_ipv4_addr[WCN36XX_HAL_IPV4_ADDR_LEN];
3181 u8 dest_ipv4_addr[WCN36XX_HAL_IPV4_ADDR_LEN];
3182 u8 dest_addr[ETH_ALEN];
3183 u8 bss_index;
3184} __packed;
3185
3186struct wcn36xx_hal_rssi_threshold_req_msg {
3187 struct wcn36xx_hal_msg_header header;
3188
3189 s8 threshold1:8;
3190 s8 threshold2:8;
3191 s8 threshold3:8;
3192 u8 thres1_pos_notify:1;
3193 u8 thres1_neg_notify:1;
3194 u8 thres2_pos_notify:1;
3195 u8 thres2_neg_notify:1;
3196 u8 thres3_pos_notify:1;
3197 u8 thres3_neg_notify:1;
3198 u8 reserved10:2;
3199};
3200
3201struct wcn36xx_hal_enter_uapsd_req_msg {
3202 struct wcn36xx_hal_msg_header header;
3203
3204 u8 bk_delivery:1;
3205 u8 be_delivery:1;
3206 u8 vi_delivery:1;
3207 u8 vo_delivery:1;
3208 u8 bk_trigger:1;
3209 u8 be_trigger:1;
3210 u8 vi_trigger:1;
3211 u8 vo_trigger:1;
3212 u8 bss_index;
3213};
3214
3215struct wcn36xx_hal_exit_uapsd_req_msg {
3216 struct wcn36xx_hal_msg_header header;
3217 u8 bss_index;
3218};
3219
3220#define WCN36XX_HAL_WOWL_BCAST_PATTERN_MAX_SIZE 128
3221#define WCN36XX_HAL_WOWL_BCAST_MAX_NUM_PATTERNS 16
3222
3223struct wcn36xx_hal_wowl_add_bcast_ptrn_req_msg {
3224 struct wcn36xx_hal_msg_header header;
3225
3226 /* Pattern ID */
3227 u8 id;
3228
3229 /* Pattern byte offset from beginning of the 802.11 packet to start
3230 * of the wake-up pattern */
3231 u8 byte_Offset;
3232
3233 /* Non-Zero Pattern size */
3234 u8 size;
3235
3236 /* Pattern */
3237 u8 pattern[WCN36XX_HAL_WOWL_BCAST_PATTERN_MAX_SIZE];
3238
3239 /* Non-zero pattern mask size */
3240 u8 mask_size;
3241
3242 /* Pattern mask */
3243 u8 mask[WCN36XX_HAL_WOWL_BCAST_PATTERN_MAX_SIZE];
3244
3245 /* Extra pattern */
3246 u8 extra[WCN36XX_HAL_WOWL_BCAST_PATTERN_MAX_SIZE];
3247
3248 /* Extra pattern mask */
3249 u8 mask_extra[WCN36XX_HAL_WOWL_BCAST_PATTERN_MAX_SIZE];
3250
3251 u8 bss_index;
3252};
3253
3254struct wcn36xx_hal_wow_del_bcast_ptrn_req_msg {
3255 struct wcn36xx_hal_msg_header header;
3256
3257 /* Pattern ID of the wakeup pattern to be deleted */
3258 u8 id;
3259 u8 bss_index;
3260};
3261
3262struct wcn36xx_hal_wowl_enter_req_msg {
3263 struct wcn36xx_hal_msg_header header;
3264
3265 /* Enables/disables magic packet filtering */
3266 u8 magic_packet_enable;
3267
3268 /* Magic pattern */
3269 u8 magic_pattern[ETH_ALEN];
3270
3271 /* Enables/disables packet pattern filtering in firmware. Enabling
3272 * this flag enables broadcast pattern matching in Firmware. If
3273 * unicast pattern matching is also desired,
3274 * ucUcastPatternFilteringEnable flag must be set tot true as well
3275 */
3276 u8 pattern_filtering_enable;
3277
3278 /* Enables/disables unicast packet pattern filtering. This flag
3279 * specifies whether we want to do pattern match on unicast packets
3280 * as well and not just broadcast packets. This flag has no effect
3281 * if the ucPatternFilteringEnable (main controlling flag) is set
3282 * to false
3283 */
3284 u8 ucast_pattern_filtering_enable;
3285
3286 /* This configuration is valid only when magicPktEnable=1. It
3287 * requests hardware to wake up when it receives the Channel Switch
3288 * Action Frame.
3289 */
3290 u8 wow_channel_switch_receive;
3291
3292 /* This configuration is valid only when magicPktEnable=1. It
3293 * requests hardware to wake up when it receives the
3294 * Deauthentication Frame.
3295 */
3296 u8 wow_deauth_receive;
3297
3298 /* This configuration is valid only when magicPktEnable=1. It
3299 * requests hardware to wake up when it receives the Disassociation
3300 * Frame.
3301 */
3302 u8 wow_disassoc_receive;
3303
3304 /* This configuration is valid only when magicPktEnable=1. It
3305 * requests hardware to wake up when it has missed consecutive
3306 * beacons. This is a hardware register configuration (NOT a
3307 * firmware configuration).
3308 */
3309 u8 wow_max_missed_beacons;
3310
3311 /* This configuration is valid only when magicPktEnable=1. This is
3312 * a timeout value in units of microsec. It requests hardware to
3313 * unconditionally wake up after it has stayed in WoWLAN mode for
3314 * some time. Set 0 to disable this feature.
3315 */
3316 u8 wow_max_sleep;
3317
3318 /* This configuration directs the WoW packet filtering to look for
3319 * EAP-ID requests embedded in EAPOL frames and use this as a wake
3320 * source.
3321 */
3322 u8 wow_eap_id_request_enable;
3323
3324 /* This configuration directs the WoW packet filtering to look for
3325 * EAPOL-4WAY requests and use this as a wake source.
3326 */
3327 u8 wow_eapol_4way_enable;
3328
3329 /* This configuration allows a host wakeup on an network scan
3330 * offload match.
3331 */
3332 u8 wow_net_scan_offload_match;
3333
3334 /* This configuration allows a host wakeup on any GTK rekeying
3335 * error.
3336 */
3337 u8 wow_gtk_rekey_error;
3338
3339 /* This configuration allows a host wakeup on BSS connection loss.
3340 */
3341 u8 wow_bss_connection_loss;
3342
3343 u8 bss_index;
3344};
3345
3346struct wcn36xx_hal_wowl_exit_req_msg {
3347 struct wcn36xx_hal_msg_header header;
3348
3349 u8 bss_index;
3350};
3351
3352struct wcn36xx_hal_get_rssi_req_msg {
3353 struct wcn36xx_hal_msg_header header;
3354};
3355
3356struct wcn36xx_hal_get_roam_rssi_req_msg {
3357 struct wcn36xx_hal_msg_header header;
3358
3359 /* Valid STA Idx for per STA stats request */
3360 u32 sta_id;
3361};
3362
3363struct wcn36xx_hal_set_uapsd_ac_params_req_msg {
3364 struct wcn36xx_hal_msg_header header;
3365
3366 /* STA index */
3367 u8 sta_idx;
3368
3369 /* Access Category */
3370 u8 ac;
3371
3372 /* User Priority */
3373 u8 up;
3374
3375 /* Service Interval */
3376 u32 service_interval;
3377
3378 /* Suspend Interval */
3379 u32 suspend_interval;
3380
3381 /* Delay Interval */
3382 u32 delay_interval;
3383};
3384
3385struct wcn36xx_hal_configure_rxp_filter_req_msg {
3386 struct wcn36xx_hal_msg_header header;
3387
3388 u8 set_mcst_bcst_filter_setting;
3389 u8 set_mcst_bcst_filter;
3390};
3391
3392struct wcn36xx_hal_enter_imps_rsp_msg {
3393 struct wcn36xx_hal_msg_header header;
3394
3395 /* success or failure */
3396 u32 status;
3397};
3398
3399struct wcn36xx_hal_exit_imps_rsp_msg {
3400 struct wcn36xx_hal_msg_header header;
3401
3402 /* success or failure */
3403 u32 status;
3404};
3405
3406struct wcn36xx_hal_enter_bmps_rsp_msg {
3407 struct wcn36xx_hal_msg_header header;
3408
3409 /* success or failure */
3410 u32 status;
3411
3412 u8 bss_index;
3413} __packed;
3414
3415struct wcn36xx_hal_exit_bmps_rsp_msg {
3416 struct wcn36xx_hal_msg_header header;
3417
3418 /* success or failure */
3419 u32 status;
3420
3421 u8 bss_index;
3422} __packed;
3423
3424struct wcn36xx_hal_enter_uapsd_rsp_msg {
3425 struct wcn36xx_hal_msg_header header;
3426
3427 /* success or failure */
3428 u32 status;
3429
3430 u8 bss_index;
3431};
3432
3433struct wcn36xx_hal_exit_uapsd_rsp_msg {
3434 struct wcn36xx_hal_msg_header header;
3435
3436 /* success or failure */
3437 u32 status;
3438
3439 u8 bss_index;
3440};
3441
3442struct wcn36xx_hal_rssi_notification_ind_msg {
3443 struct wcn36xx_hal_msg_header header;
3444
3445 u32 rssi_thres1_pos_cross:1;
3446 u32 rssi_thres1_neg_cross:1;
3447 u32 rssi_thres2_pos_cross:1;
3448 u32 rssi_thres2_neg_cross:1;
3449 u32 rssi_thres3_pos_cross:1;
3450 u32 rssi_thres3_neg_cross:1;
3451 u32 avg_rssi:8;
3452 u32 reserved:18;
3453
3454};
3455
3456struct wcn36xx_hal_get_rssio_rsp_msg {
3457 struct wcn36xx_hal_msg_header header;
3458
3459 /* success or failure */
3460 u32 status;
3461 s8 rssi;
3462
3463};
3464
3465struct wcn36xx_hal_get_roam_rssi_rsp_msg {
3466 struct wcn36xx_hal_msg_header header;
3467
3468 /* success or failure */
3469 u32 status;
3470
3471 u8 sta_id;
3472 s8 rssi;
3473};
3474
3475struct wcn36xx_hal_wowl_enter_rsp_msg {
3476 struct wcn36xx_hal_msg_header header;
3477
3478 /* success or failure */
3479 u32 status;
3480 u8 bss_index;
3481};
3482
3483struct wcn36xx_hal_wowl_exit_rsp_msg {
3484 struct wcn36xx_hal_msg_header header;
3485
3486 /* success or failure */
3487 u32 status;
3488 u8 bss_index;
3489};
3490
3491struct wcn36xx_hal_add_bcn_filter_rsp_msg {
3492 struct wcn36xx_hal_msg_header header;
3493
3494 /* success or failure */
3495 u32 status;
3496};
3497
3498struct wcn36xx_hal_rem_bcn_filter_rsp_msg {
3499 struct wcn36xx_hal_msg_header header;
3500
3501 /* success or failure */
3502 u32 status;
3503};
3504
3505struct wcn36xx_hal_add_wowl_bcast_ptrn_rsp_msg {
3506 struct wcn36xx_hal_msg_header header;
3507
3508 /* success or failure */
3509 u32 status;
3510 u8 bss_index;
3511};
3512
3513struct wcn36xx_hal_del_wowl_bcast_ptrn_rsp_msg {
3514 struct wcn36xx_hal_msg_header header;
3515
3516 /* success or failure */
3517 u32 status;
3518 u8 bss_index;
3519};
3520
3521struct wcn36xx_hal_host_offload_rsp_msg {
3522 struct wcn36xx_hal_msg_header header;
3523
3524 /* success or failure */
3525 u32 status;
3526};
3527
3528struct wcn36xx_hal_keep_alive_rsp_msg {
3529 struct wcn36xx_hal_msg_header header;
3530
3531 /* success or failure */
3532 u32 status;
3533};
3534
3535struct wcn36xx_hal_set_rssi_thresh_rsp_msg {
3536 struct wcn36xx_hal_msg_header header;
3537
3538 /* success or failure */
3539 u32 status;
3540};
3541
3542struct wcn36xx_hal_set_uapsd_ac_params_rsp_msg {
3543 struct wcn36xx_hal_msg_header header;
3544
3545 /* success or failure */
3546 u32 status;
3547};
3548
3549struct wcn36xx_hal_configure_rxp_filter_rsp_msg {
3550 struct wcn36xx_hal_msg_header header;
3551
3552 /* success or failure */
3553 u32 status;
3554};
3555
3556struct set_max_tx_pwr_req {
3557 struct wcn36xx_hal_msg_header header;
3558
3559 /* BSSID is needed to identify which session issued this request.
3560 * As the request has power constraints, this should be applied
3561 * only to that session */
3562 u8 bssid[ETH_ALEN];
3563
3564 u8 self_addr[ETH_ALEN];
3565
3566 /* In request, power == MaxTx power to be used. */
3567 u8 power;
3568};
3569
3570struct set_max_tx_pwr_rsp_msg {
3571 struct wcn36xx_hal_msg_header header;
3572
3573 /* power == tx power used for management frames */
3574 u8 power;
3575
3576 /* success or failure */
3577 u32 status;
3578};
3579
3580struct set_tx_pwr_req_msg {
3581 struct wcn36xx_hal_msg_header header;
3582
3583 /* TX Power in milli watts */
3584 u32 tx_power;
3585
3586 u8 bss_index;
3587};
3588
3589struct set_tx_pwr_rsp_msg {
3590 struct wcn36xx_hal_msg_header header;
3591
3592 /* success or failure */
3593 u32 status;
3594};
3595
3596struct get_tx_pwr_req_msg {
3597 struct wcn36xx_hal_msg_header header;
3598
3599 u8 sta_id;
3600};
3601
3602struct get_tx_pwr_rsp_msg {
3603 struct wcn36xx_hal_msg_header header;
3604
3605 /* success or failure */
3606 u32 status;
3607
3608 /* TX Power in milli watts */
3609 u32 tx_power;
3610};
3611
3612struct set_p2p_gonoa_req_msg {
3613 struct wcn36xx_hal_msg_header header;
3614
3615 u8 opp_ps;
3616 u32 ct_window;
3617 u8 count;
3618 u32 duration;
3619 u32 interval;
3620 u32 single_noa_duration;
3621 u8 ps_selection;
3622};
3623
3624struct set_p2p_gonoa_rsp_msg {
3625 struct wcn36xx_hal_msg_header header;
3626
3627 /* success or failure */
3628 u32 status;
3629};
3630
3631struct wcn36xx_hal_add_sta_self_req {
3632 struct wcn36xx_hal_msg_header header;
3633
3634 u8 self_addr[ETH_ALEN];
3635 u32 status;
3636} __packed;
3637
3638struct wcn36xx_hal_add_sta_self_rsp_msg {
3639 struct wcn36xx_hal_msg_header header;
3640
3641 /* success or failure */
3642 u32 status;
3643
3644 /* Self STA Index */
3645 u8 self_sta_index;
3646
3647 /* DPU Index (IGTK, PTK, GTK all same) */
3648 u8 dpu_index;
3649
3650 /* DPU Signature */
3651 u8 dpu_signature;
3652} __packed;
3653
3654struct wcn36xx_hal_del_sta_self_req_msg {
3655 struct wcn36xx_hal_msg_header header;
3656
3657 u8 self_addr[ETH_ALEN];
3658} __packed;
3659
3660struct wcn36xx_hal_del_sta_self_rsp_msg {
3661 struct wcn36xx_hal_msg_header header;
3662
3663 /*success or failure */
3664 u32 status;
3665
3666 u8 self_addr[ETH_ALEN];
3667} __packed;
3668
3669struct aggr_add_ts_req {
3670 struct wcn36xx_hal_msg_header header;
3671
3672 /* Station Index */
3673 u16 sta_idx;
3674
3675 /* TSPEC handler uniquely identifying a TSPEC for a STA in a BSS.
3676 * This will carry the bitmap with the bit positions representing
3677 * different AC.s */
3678 u16 tspec_index;
3679
3680 /* Tspec info per AC To program TPE with required parameters */
3681 struct wcn36xx_hal_tspec_ie tspec[WCN36XX_HAL_MAX_AC];
3682
3683 /* U-APSD Flags: 1b per AC. Encoded as follows:
3684 b7 b6 b5 b4 b3 b2 b1 b0 =
3685 X X X X BE BK VI VO */
3686 u8 uapsd;
3687
3688 /* These parameters are for all the access categories */
3689
3690 /* Service Interval */
3691 u32 service_interval[WCN36XX_HAL_MAX_AC];
3692
3693 /* Suspend Interval */
3694 u32 suspend_interval[WCN36XX_HAL_MAX_AC];
3695
3696 /* Delay Interval */
3697 u32 delay_interval[WCN36XX_HAL_MAX_AC];
3698};
3699
3700struct aggr_add_ts_rsp_msg {
3701 struct wcn36xx_hal_msg_header header;
3702
3703 /* success or failure */
3704 u32 status0;
3705
3706 /* FIXME PRIMA for future use for 11R */
3707 u32 status1;
3708};
3709
3710struct wcn36xx_hal_configure_apps_cpu_wakeup_state_req_msg {
3711 struct wcn36xx_hal_msg_header header;
3712
3713 u8 is_apps_cpu_awake;
3714};
3715
3716struct wcn36xx_hal_configure_apps_cpu_wakeup_state_rsp_msg {
3717 struct wcn36xx_hal_msg_header header;
3718
3719 /* success or failure */
3720 u32 status;
3721};
3722
3723struct wcn36xx_hal_dump_cmd_req_msg {
3724 struct wcn36xx_hal_msg_header header;
3725
3726 u32 arg1;
3727 u32 arg2;
3728 u32 arg3;
3729 u32 arg4;
3730 u32 arg5;
3731} __packed;
3732
3733struct wcn36xx_hal_dump_cmd_rsp_msg {
3734 struct wcn36xx_hal_msg_header header;
3735
3736 /* success or failure */
3737 u32 status;
3738
3739 /* Length of the responce message */
3740 u32 rsp_length;
3741
3742 /* FIXME: Currently considering the the responce will be less than
3743 * 100bytes */
3744 u8 rsp_buffer[DUMPCMD_RSP_BUFFER];
3745} __packed;
3746
3747#define WLAN_COEX_IND_DATA_SIZE (4)
3748#define WLAN_COEX_IND_TYPE_DISABLE_HB_MONITOR (0)
3749#define WLAN_COEX_IND_TYPE_ENABLE_HB_MONITOR (1)
3750
3751struct coex_ind_msg {
3752 struct wcn36xx_hal_msg_header header;
3753
3754 /* Coex Indication Type */
3755 u32 type;
3756
3757 /* Coex Indication Data */
3758 u32 data[WLAN_COEX_IND_DATA_SIZE];
3759};
3760
3761struct wcn36xx_hal_tx_compl_ind_msg {
3762 struct wcn36xx_hal_msg_header header;
3763
3764 /* Tx Complete Indication Success or Failure */
3765 u32 status;
3766};
3767
3768struct wcn36xx_hal_wlan_host_suspend_ind_msg {
3769 struct wcn36xx_hal_msg_header header;
3770
3771 u32 configured_mcst_bcst_filter_setting;
3772 u32 active_session_count;
3773};
3774
3775struct wcn36xx_hal_wlan_exclude_unencrpted_ind_msg {
3776 struct wcn36xx_hal_msg_header header;
3777
3778 u8 dot11_exclude_unencrypted;
3779 u8 bssid[ETH_ALEN];
3780};
3781
3782struct noa_attr_ind_msg {
3783 struct wcn36xx_hal_msg_header header;
3784
3785 u8 index;
3786 u8 opp_ps_flag;
3787 u16 ctwin;
3788
3789 u16 noa1_interval_count;
3790 u16 bss_index;
3791 u32 noa1_duration;
3792 u32 noa1_interval;
3793 u32 noa1_starttime;
3794
3795 u16 noa2_interval_count;
3796 u16 reserved2;
3797 u32 noa2_duration;
3798 u32 noa2_interval;
3799 u32 noa2_start_time;
3800
3801 u32 status;
3802};
3803
3804struct noa_start_ind_msg {
3805 struct wcn36xx_hal_msg_header header;
3806
3807 u32 status;
3808 u32 bss_index;
3809};
3810
3811struct wcn36xx_hal_wlan_host_resume_req_msg {
3812 struct wcn36xx_hal_msg_header header;
3813
3814 u8 configured_mcst_bcst_filter_setting;
3815};
3816
3817struct wcn36xx_hal_host_resume_rsp_msg {
3818 struct wcn36xx_hal_msg_header header;
3819
3820 /* success or failure */
3821 u32 status;
3822};
3823
3824struct wcn36xx_hal_del_ba_ind_msg {
3825 struct wcn36xx_hal_msg_header header;
3826
3827 u16 sta_idx;
3828
3829 /* Peer MAC Address, whose BA session has timed out */
3830 u8 peer_addr[ETH_ALEN];
3831
3832 /* TID for which a BA session timeout is being triggered */
3833 u8 ba_tid;
3834
3835 /* DELBA direction
3836 * 1 - Originator
3837 * 0 - Recipient
3838 */
3839 u8 direction;
3840
3841 u32 reason_code;
3842
3843 /* TO SUPPORT BT-AMP */
3844 u8 bssid[ETH_ALEN];
3845};
3846
3847/* PNO Messages */
3848
3849/* Max number of channels that a network can be found on */
3850#define WCN36XX_HAL_PNO_MAX_NETW_CHANNELS 26
3851
3852/* Max number of channels that a network can be found on */
3853#define WCN36XX_HAL_PNO_MAX_NETW_CHANNELS_EX 60
3854
3855/* Maximum numbers of networks supported by PNO */
3856#define WCN36XX_HAL_PNO_MAX_SUPP_NETWORKS 16
3857
3858/* The number of scan time intervals that can be programmed into PNO */
3859#define WCN36XX_HAL_PNO_MAX_SCAN_TIMERS 10
3860
3861/* Maximum size of the probe template */
3862#define WCN36XX_HAL_PNO_MAX_PROBE_SIZE 450
3863
3864/* Type of PNO enabling:
3865 *
3866 * Immediate - scanning will start immediately and PNO procedure will be
3867 * repeated based on timer
3868 *
3869 * Suspend - scanning will start at suspend
3870 *
3871 * Resume - scanning will start on system resume
3872 */
3873enum pno_mode {
3874 PNO_MODE_IMMEDIATE,
3875 PNO_MODE_ON_SUSPEND,
3876 PNO_MODE_ON_RESUME,
3877 PNO_MODE_MAX = WCN36XX_HAL_MAX_ENUM_SIZE
3878};
3879
3880/* Authentication type */
3881enum auth_type {
3882 AUTH_TYPE_ANY = 0,
3883 AUTH_TYPE_OPEN_SYSTEM = 1,
3884
3885 /* Upper layer authentication types */
3886 AUTH_TYPE_WPA = 2,
3887 AUTH_TYPE_WPA_PSK = 3,
3888
3889 AUTH_TYPE_RSN = 4,
3890 AUTH_TYPE_RSN_PSK = 5,
3891 AUTH_TYPE_FT_RSN = 6,
3892 AUTH_TYPE_FT_RSN_PSK = 7,
3893 AUTH_TYPE_WAPI_WAI_CERTIFICATE = 8,
3894 AUTH_TYPE_WAPI_WAI_PSK = 9,
3895
3896 AUTH_TYPE_MAX = WCN36XX_HAL_MAX_ENUM_SIZE
3897};
3898
3899/* Encryption type */
3900enum ed_type {
3901 ED_ANY = 0,
3902 ED_NONE = 1,
3903 ED_WEP = 2,
3904 ED_TKIP = 3,
3905 ED_CCMP = 4,
3906 ED_WPI = 5,
3907
3908 ED_TYPE_MAX = WCN36XX_HAL_MAX_ENUM_SIZE
3909};
3910
3911/* SSID broadcast type */
3912enum ssid_bcast_type {
3913 BCAST_UNKNOWN = 0,
3914 BCAST_NORMAL = 1,
3915 BCAST_HIDDEN = 2,
3916
3917 BCAST_TYPE_MAX = WCN36XX_HAL_MAX_ENUM_SIZE
3918};
3919
3920/* The network description for which PNO will have to look for */
3921struct network_type {
3922 /* SSID of the BSS */
3923 struct wcn36xx_hal_mac_ssid ssid;
3924
3925 /* Authentication type for the network */
3926 enum auth_type authentication;
3927
3928 /* Encryption type for the network */
3929 enum ed_type encryption;
3930
3931 /* Indicate the channel on which the Network can be found 0 - if
3932 * all channels */
3933 u8 channel_count;
3934 u8 channels[WCN36XX_HAL_PNO_MAX_NETW_CHANNELS];
3935
3936 /* Indicates the RSSI threshold for the network to be considered */
3937 u8 rssi_threshold;
3938};
3939
3940struct scan_timer {
3941 /* How much it should wait */
3942 u32 value;
3943
3944 /* How many times it should repeat that wait value 0 - keep using
3945 * this timer until PNO is disabled */
3946 u32 repeat;
3947
3948 /* e.g: 2 3 4 0 - it will wait 2s between consecutive scans for 3
3949 * times - after that it will wait 4s between consecutive scans
3950 * until disabled */
3951};
3952
3953/* The network parameters to be sent to the PNO algorithm */
3954struct scan_timers_type {
3955 /* set to 0 if you wish for PNO to use its default telescopic timer */
3956 u8 count;
3957
3958 /* A set value represents the amount of time that PNO will wait
3959 * between two consecutive scan procedures If the desired is for a
3960 * uniform timer that fires always at the exact same interval - one
3961 * single value is to be set If there is a desire for a more
3962 * complex - telescopic like timer multiple values can be set -
3963 * once PNO reaches the end of the array it will continue scanning
3964 * at intervals presented by the last value */
3965 struct scan_timer values[WCN36XX_HAL_PNO_MAX_SCAN_TIMERS];
3966};
3967
3968/* Preferred network list request */
3969struct set_pref_netw_list_req {
3970 struct wcn36xx_hal_msg_header header;
3971
3972 /* Enable PNO */
3973 u32 enable;
3974
3975 /* Immediate, On Suspend, On Resume */
3976 enum pno_mode mode;
3977
3978 /* Number of networks sent for PNO */
3979 u32 networks_count;
3980
3981 /* The networks that PNO needs to look for */
3982 struct network_type networks[WCN36XX_HAL_PNO_MAX_SUPP_NETWORKS];
3983
3984 /* The scan timers required for PNO */
3985 struct scan_timers_type scan_timers;
3986
3987 /* Probe template for 2.4GHz band */
3988 u16 band_24g_probe_size;
3989 u8 band_24g_probe_template[WCN36XX_HAL_PNO_MAX_PROBE_SIZE];
3990
3991 /* Probe template for 5GHz band */
3992 u16 band_5g_probe_size;
3993 u8 band_5g_probe_template[WCN36XX_HAL_PNO_MAX_PROBE_SIZE];
3994};
3995
3996/* The network description for which PNO will have to look for */
3997struct network_type_new {
3998 /* SSID of the BSS */
3999 struct wcn36xx_hal_mac_ssid ssid;
4000
4001 /* Authentication type for the network */
4002 enum auth_type authentication;
4003
4004 /* Encryption type for the network */
4005 enum ed_type encryption;
4006
4007 /* SSID broadcast type, normal, hidden or unknown */
4008 enum ssid_bcast_type bcast_network_type;
4009
4010 /* Indicate the channel on which the Network can be found 0 - if
4011 * all channels */
4012 u8 channel_count;
4013 u8 channels[WCN36XX_HAL_PNO_MAX_NETW_CHANNELS];
4014
4015 /* Indicates the RSSI threshold for the network to be considered */
4016 u8 rssi_threshold;
4017};
4018
4019/* Preferred network list request new */
4020struct set_pref_netw_list_req_new {
4021 struct wcn36xx_hal_msg_header header;
4022
4023 /* Enable PNO */
4024 u32 enable;
4025
4026 /* Immediate, On Suspend, On Resume */
4027 enum pno_mode mode;
4028
4029 /* Number of networks sent for PNO */
4030 u32 networks_count;
4031
4032 /* The networks that PNO needs to look for */
4033 struct network_type_new networks[WCN36XX_HAL_PNO_MAX_SUPP_NETWORKS];
4034
4035 /* The scan timers required for PNO */
4036 struct scan_timers_type scan_timers;
4037
4038 /* Probe template for 2.4GHz band */
4039 u16 band_24g_probe_size;
4040 u8 band_24g_probe_template[WCN36XX_HAL_PNO_MAX_PROBE_SIZE];
4041
4042 /* Probe template for 5GHz band */
4043 u16 band_5g_probe_size;
4044 u8 band_5g_probe_template[WCN36XX_HAL_PNO_MAX_PROBE_SIZE];
4045};
4046
4047/* Preferred network list response */
4048struct set_pref_netw_list_resp {
4049 struct wcn36xx_hal_msg_header header;
4050
4051 /* status of the request - just to indicate that PNO has
4052 * acknowledged the request and will start scanning */
4053 u32 status;
4054};
4055
4056/* Preferred network found indication */
4057struct pref_netw_found_ind {
4058
4059 struct wcn36xx_hal_msg_header header;
4060
4061 /* Network that was found with the highest RSSI */
4062 struct wcn36xx_hal_mac_ssid ssid;
4063
4064 /* Indicates the RSSI */
4065 u8 rssi;
4066};
4067
4068/* RSSI Filter request */
4069struct set_rssi_filter_req {
4070 struct wcn36xx_hal_msg_header header;
4071
4072 /* RSSI Threshold */
4073 u8 rssi_threshold;
4074};
4075
4076/* Set RSSI filter resp */
4077struct set_rssi_filter_resp {
4078 struct wcn36xx_hal_msg_header header;
4079
4080 /* status of the request */
4081 u32 status;
4082};
4083
4084/* Update scan params - sent from host to PNO to be used during PNO
4085 * scanningx */
4086struct wcn36xx_hal_update_scan_params_req {
4087
4088 struct wcn36xx_hal_msg_header header;
4089
4090 /* Host setting for 11d */
4091 u8 dot11d_enabled;
4092
4093 /* Lets PNO know that host has determined the regulatory domain */
4094 u8 dot11d_resolved;
4095
4096 /* Channels on which PNO is allowed to scan */
4097 u8 channel_count;
4098 u8 channels[WCN36XX_HAL_PNO_MAX_NETW_CHANNELS];
4099
4100 /* Minimum channel time */
4101 u16 active_min_ch_time;
4102
4103 /* Maximum channel time */
4104 u16 active_max_ch_time;
4105
4106 /* Minimum channel time */
4107 u16 passive_min_ch_time;
4108
4109 /* Maximum channel time */
4110 u16 passive_max_ch_time;
4111
4112 /* Cb State */
4113 enum phy_chan_bond_state state;
4114} __packed;
4115
4116/* Update scan params - sent from host to PNO to be used during PNO
4117 * scanningx */
4118struct update_scan_params_req_ex {
4119
4120 struct wcn36xx_hal_msg_header header;
4121
4122 /* Host setting for 11d */
4123 u8 dot11d_enabled;
4124
4125 /* Lets PNO know that host has determined the regulatory domain */
4126 u8 dot11d_resolved;
4127
4128 /* Channels on which PNO is allowed to scan */
4129 u8 channel_count;
4130 u8 channels[WCN36XX_HAL_PNO_MAX_NETW_CHANNELS_EX];
4131
4132 /* Minimum channel time */
4133 u16 active_min_ch_time;
4134
4135 /* Maximum channel time */
4136 u16 active_max_ch_time;
4137
4138 /* Minimum channel time */
4139 u16 passive_min_ch_time;
4140
4141 /* Maximum channel time */
4142 u16 passive_max_ch_time;
4143
4144 /* Cb State */
4145 enum phy_chan_bond_state state;
4146};
4147
4148/* Update scan params - sent from host to PNO to be used during PNO
4149 * scanningx */
4150struct wcn36xx_hal_update_scan_params_resp {
4151
4152 struct wcn36xx_hal_msg_header header;
4153
4154 /* status of the request */
4155 u32 status;
4156} __packed;
4157
4158struct wcn36xx_hal_set_tx_per_tracking_req_msg {
4159 struct wcn36xx_hal_msg_header header;
4160
4161 /* 0: disable, 1:enable */
4162 u8 tx_per_tracking_enable;
4163
4164 /* Check period, unit is sec. */
4165 u8 tx_per_tracking_period;
4166
4167 /* (Fail TX packet)/(Total TX packet) ratio, the unit is 10%. */
4168 u8 tx_per_tracking_ratio;
4169
4170 /* A watermark of check number, once the tx packet exceed this
4171 * number, we do the check, default is 5 */
4172 u32 tx_per_tracking_watermark;
4173};
4174
4175struct wcn36xx_hal_set_tx_per_tracking_rsp_msg {
4176 struct wcn36xx_hal_msg_header header;
4177
4178 /* success or failure */
4179 u32 status;
4180
4181};
4182
4183struct tx_per_hit_ind_msg {
4184 struct wcn36xx_hal_msg_header header;
4185};
4186
4187/* Packet Filtering Definitions Begin */
4188#define WCN36XX_HAL_PROTOCOL_DATA_LEN 8
4189#define WCN36XX_HAL_MAX_NUM_MULTICAST_ADDRESS 240
4190#define WCN36XX_HAL_MAX_NUM_FILTERS 20
4191#define WCN36XX_HAL_MAX_CMP_PER_FILTER 10
4192
4193enum wcn36xx_hal_receive_packet_filter_type {
4194 HAL_RCV_FILTER_TYPE_INVALID,
4195 HAL_RCV_FILTER_TYPE_FILTER_PKT,
4196 HAL_RCV_FILTER_TYPE_BUFFER_PKT,
4197 HAL_RCV_FILTER_TYPE_MAX_ENUM_SIZE
4198};
4199
4200enum wcn36xx_hal_rcv_pkt_flt_protocol_type {
4201 HAL_FILTER_PROTO_TYPE_INVALID,
4202 HAL_FILTER_PROTO_TYPE_MAC,
4203 HAL_FILTER_PROTO_TYPE_ARP,
4204 HAL_FILTER_PROTO_TYPE_IPV4,
4205 HAL_FILTER_PROTO_TYPE_IPV6,
4206 HAL_FILTER_PROTO_TYPE_UDP,
4207 HAL_FILTER_PROTO_TYPE_MAX
4208};
4209
4210enum wcn36xx_hal_rcv_pkt_flt_cmp_flag_type {
4211 HAL_FILTER_CMP_TYPE_INVALID,
4212 HAL_FILTER_CMP_TYPE_EQUAL,
4213 HAL_FILTER_CMP_TYPE_MASK_EQUAL,
4214 HAL_FILTER_CMP_TYPE_NOT_EQUAL,
4215 HAL_FILTER_CMP_TYPE_MAX
4216};
4217
4218struct wcn36xx_hal_rcv_pkt_filter_params {
4219 u8 protocol_layer;
4220 u8 cmp_flag;
4221
4222 /* Length of the data to compare */
4223 u16 data_length;
4224
4225 /* from start of the respective frame header */
4226 u8 data_offset;
4227
4228 /* Reserved field */
4229 u8 reserved;
4230
4231 /* Data to compare */
4232 u8 compare_data[WCN36XX_HAL_PROTOCOL_DATA_LEN];
4233
4234 /* Mask to be applied on the received packet data before compare */
4235 u8 data_mask[WCN36XX_HAL_PROTOCOL_DATA_LEN];
4236};
4237
4238struct wcn36xx_hal_sessionized_rcv_pkt_filter_cfg_type {
4239 u8 id;
4240 u8 type;
4241 u8 params_count;
4242 u32 coleasce_time;
4243 u8 bss_index;
4244 struct wcn36xx_hal_rcv_pkt_filter_params params[1];
4245};
4246
4247struct wcn36xx_hal_set_rcv_pkt_filter_req_msg {
4248 struct wcn36xx_hal_msg_header header;
4249
4250 u8 id;
4251 u8 type;
4252 u8 params_count;
4253 u32 coalesce_time;
4254 struct wcn36xx_hal_rcv_pkt_filter_params params[1];
4255};
4256
4257struct wcn36xx_hal_rcv_flt_mc_addr_list_type {
4258 /* from start of the respective frame header */
4259 u8 data_offset;
4260
4261 u32 mc_addr_count;
4262 u8 mc_addr[ETH_ALEN][WCN36XX_HAL_MAX_NUM_MULTICAST_ADDRESS];
4263 u8 bss_index;
4264};
4265
4266struct wcn36xx_hal_set_pkt_filter_rsp_msg {
4267 struct wcn36xx_hal_msg_header header;
4268
4269 /* success or failure */
4270 u32 status;
4271
4272 u8 bss_index;
4273};
4274
4275struct wcn36xx_hal_rcv_flt_pkt_match_cnt_req_msg {
4276 struct wcn36xx_hal_msg_header header;
4277
4278 u8 bss_index;
4279};
4280
4281struct wcn36xx_hal_rcv_flt_pkt_match_cnt {
4282 u8 id;
4283 u32 match_cnt;
4284};
4285
4286struct wcn36xx_hal_rcv_flt_pkt_match_cnt_rsp_msg {
4287 struct wcn36xx_hal_msg_header header;
4288
4289 /* Success or Failure */
4290 u32 status;
4291
4292 u32 match_count;
4293 struct wcn36xx_hal_rcv_flt_pkt_match_cnt
4294 matches[WCN36XX_HAL_MAX_NUM_FILTERS];
4295 u8 bss_index;
4296};
4297
4298struct wcn36xx_hal_rcv_flt_pkt_clear_param {
4299 /* only valid for response message */
4300 u32 status;
4301 u8 id;
4302 u8 bss_index;
4303};
4304
4305struct wcn36xx_hal_rcv_flt_pkt_clear_req_msg {
4306 struct wcn36xx_hal_msg_header header;
4307 struct wcn36xx_hal_rcv_flt_pkt_clear_param param;
4308};
4309
4310struct wcn36xx_hal_rcv_flt_pkt_clear_rsp_msg {
4311 struct wcn36xx_hal_msg_header header;
4312 struct wcn36xx_hal_rcv_flt_pkt_clear_param param;
4313};
4314
4315struct wcn36xx_hal_rcv_flt_pkt_set_mc_list_req_msg {
4316 struct wcn36xx_hal_msg_header header;
4317 struct wcn36xx_hal_rcv_flt_mc_addr_list_type mc_addr_list;
4318};
4319
4320struct wcn36xx_hal_rcv_flt_pkt_set_mc_list_rsp_msg {
4321 struct wcn36xx_hal_msg_header header;
4322 u32 status;
4323 u8 bss_index;
4324};
4325
4326/* Packet Filtering Definitions End */
4327
4328struct wcn36xx_hal_set_power_params_req_msg {
4329 struct wcn36xx_hal_msg_header header;
4330
4331 /* Ignore DTIM */
4332 u32 ignore_dtim;
4333
4334 /* DTIM Period */
4335 u32 dtim_period;
4336
4337 /* Listen Interval */
4338 u32 listen_interval;
4339
4340 /* Broadcast Multicast Filter */
4341 u32 bcast_mcast_filter;
4342
4343 /* Beacon Early Termination */
4344 u32 enable_bet;
4345
4346 /* Beacon Early Termination Interval */
4347 u32 bet_interval;
4348} __packed;
4349
4350struct wcn36xx_hal_set_power_params_resp {
4351
4352 struct wcn36xx_hal_msg_header header;
4353
4354 /* status of the request */
4355 u32 status;
4356} __packed;
4357
4358/* Capability bitmap exchange definitions and macros starts */
4359
4360enum place_holder_in_cap_bitmap {
4361 MCC = 0,
4362 P2P = 1,
4363 DOT11AC = 2,
4364 SLM_SESSIONIZATION = 3,
4365 DOT11AC_OPMODE = 4,
4366 SAP32STA = 5,
4367 TDLS = 6,
4368 P2P_GO_NOA_DECOUPLE_INIT_SCAN = 7,
4369 WLANACTIVE_OFFLOAD = 8,
4370 BEACON_OFFLOAD = 9,
4371 SCAN_OFFLOAD = 10,
4372 ROAM_OFFLOAD = 11,
4373 BCN_MISS_OFFLOAD = 12,
4374 STA_POWERSAVE = 13,
4375 STA_ADVANCED_PWRSAVE = 14,
4376 AP_UAPSD = 15,
4377 AP_DFS = 16,
4378 BLOCKACK = 17,
4379 PHY_ERR = 18,
4380 BCN_FILTER = 19,
4381 RTT = 20,
4382 RATECTRL = 21,
4383 WOW = 22,
4384 MAX_FEATURE_SUPPORTED = 128,
4385};
4386
4387struct wcn36xx_hal_feat_caps_msg {
4388
4389 struct wcn36xx_hal_msg_header header;
4390
4391 u32 feat_caps[4];
4392} __packed;
4393
4394/* status codes to help debug rekey failures */
4395enum gtk_rekey_status {
4396 WCN36XX_HAL_GTK_REKEY_STATUS_SUCCESS = 0,
4397
4398 /* rekey detected, but not handled */
4399 WCN36XX_HAL_GTK_REKEY_STATUS_NOT_HANDLED = 1,
4400
4401 /* MIC check error on M1 */
4402 WCN36XX_HAL_GTK_REKEY_STATUS_MIC_ERROR = 2,
4403
4404 /* decryption error on M1 */
4405 WCN36XX_HAL_GTK_REKEY_STATUS_DECRYPT_ERROR = 3,
4406
4407 /* M1 replay detected */
4408 WCN36XX_HAL_GTK_REKEY_STATUS_REPLAY_ERROR = 4,
4409
4410 /* missing GTK key descriptor in M1 */
4411 WCN36XX_HAL_GTK_REKEY_STATUS_MISSING_KDE = 5,
4412
4413 /* missing iGTK key descriptor in M1 */
4414 WCN36XX_HAL_GTK_REKEY_STATUS_MISSING_IGTK_KDE = 6,
4415
4416 /* key installation error */
4417 WCN36XX_HAL_GTK_REKEY_STATUS_INSTALL_ERROR = 7,
4418
4419 /* iGTK key installation error */
4420 WCN36XX_HAL_GTK_REKEY_STATUS_IGTK_INSTALL_ERROR = 8,
4421
4422 /* GTK rekey M2 response TX error */
4423 WCN36XX_HAL_GTK_REKEY_STATUS_RESP_TX_ERROR = 9,
4424
4425 /* non-specific general error */
4426 WCN36XX_HAL_GTK_REKEY_STATUS_GEN_ERROR = 255
4427};
4428
4429/* wake reason types */
4430enum wake_reason_type {
4431 WCN36XX_HAL_WAKE_REASON_NONE = 0,
4432
4433 /* magic packet match */
4434 WCN36XX_HAL_WAKE_REASON_MAGIC_PACKET = 1,
4435
4436 /* host defined pattern match */
4437 WCN36XX_HAL_WAKE_REASON_PATTERN_MATCH = 2,
4438
4439 /* EAP-ID frame detected */
4440 WCN36XX_HAL_WAKE_REASON_EAPID_PACKET = 3,
4441
4442 /* start of EAPOL 4-way handshake detected */
4443 WCN36XX_HAL_WAKE_REASON_EAPOL4WAY_PACKET = 4,
4444
4445 /* network scan offload match */
4446 WCN36XX_HAL_WAKE_REASON_NETSCAN_OFFL_MATCH = 5,
4447
4448 /* GTK rekey status wakeup (see status) */
4449 WCN36XX_HAL_WAKE_REASON_GTK_REKEY_STATUS = 6,
4450
4451 /* BSS connection lost */
4452 WCN36XX_HAL_WAKE_REASON_BSS_CONN_LOST = 7,
4453};
4454
4455/*
4456 Wake Packet which is saved at tWakeReasonParams.DataStart
4457 This data is sent for any wake reasons that involve a packet-based wakeup :
4458
4459 WCN36XX_HAL_WAKE_REASON_TYPE_MAGIC_PACKET
4460 WCN36XX_HAL_WAKE_REASON_TYPE_PATTERN_MATCH
4461 WCN36XX_HAL_WAKE_REASON_TYPE_EAPID_PACKET
4462 WCN36XX_HAL_WAKE_REASON_TYPE_EAPOL4WAY_PACKET
4463 WCN36XX_HAL_WAKE_REASON_TYPE_GTK_REKEY_STATUS
4464
4465 The information is provided to the host for auditing and debug purposes
4466
4467*/
4468
4469/* Wake reason indication */
4470struct wcn36xx_hal_wake_reason_ind {
4471 struct wcn36xx_hal_msg_header header;
4472
4473 /* see tWakeReasonType */
4474 u32 reason;
4475
4476 /* argument specific to the reason type */
4477 u32 reason_arg;
4478
4479 /* length of optional data stored in this message, in case HAL
4480 * truncates the data (i.e. data packets) this length will be less
4481 * than the actual length */
4482 u32 stored_data_len;
4483
4484 /* actual length of data */
4485 u32 actual_data_len;
4486
4487 /* variable length start of data (length == storedDataLen) see
4488 * specific wake type */
4489 u8 data_start[1];
4490
4491 u32 bss_index:8;
4492 u32 reserved:24;
4493};
4494
4495#define WCN36XX_HAL_GTK_KEK_BYTES 16
4496#define WCN36XX_HAL_GTK_KCK_BYTES 16
4497
4498#define WCN36XX_HAL_GTK_OFFLOAD_FLAGS_DISABLE (1 << 0)
4499
4500#define GTK_SET_BSS_KEY_TAG 0x1234AA55
4501
4502struct wcn36xx_hal_gtk_offload_req_msg {
4503 struct wcn36xx_hal_msg_header header;
4504
4505 /* optional flags */
4506 u32 flags;
4507
4508 /* Key confirmation key */
4509 u8 kck[WCN36XX_HAL_GTK_KCK_BYTES];
4510
4511 /* key encryption key */
4512 u8 kek[WCN36XX_HAL_GTK_KEK_BYTES];
4513
4514 /* replay counter */
4515 u64 key_replay_counter;
4516
4517 u8 bss_index;
4518};
4519
4520struct wcn36xx_hal_gtk_offload_rsp_msg {
4521 struct wcn36xx_hal_msg_header header;
4522
4523 /* success or failure */
4524 u32 status;
4525
4526 u8 bss_index;
4527};
4528
4529struct wcn36xx_hal_gtk_offload_get_info_req_msg {
4530 struct wcn36xx_hal_msg_header header;
4531 u8 bss_index;
4532};
4533
4534struct wcn36xx_hal_gtk_offload_get_info_rsp_msg {
4535 struct wcn36xx_hal_msg_header header;
4536
4537 /* success or failure */
4538 u32 status;
4539
4540 /* last rekey status when the rekey was offloaded */
4541 u32 last_rekey_status;
4542
4543 /* current replay counter value */
4544 u64 key_replay_counter;
4545
4546 /* total rekey attempts */
4547 u32 total_rekey_count;
4548
4549 /* successful GTK rekeys */
4550 u32 gtk_rekey_count;
4551
4552 /* successful iGTK rekeys */
4553 u32 igtk_rekey_count;
4554
4555 u8 bss_index;
4556};
4557
4558struct dhcp_info {
4559 /* Indicates the device mode which indicates about the DHCP activity */
4560 u8 device_mode;
4561
4562 u8 addr[ETH_ALEN];
4563};
4564
4565struct dhcp_ind_status {
4566 struct wcn36xx_hal_msg_header header;
4567
4568 /* success or failure */
4569 u32 status;
4570};
4571
4572/*
4573 * Thermal Mitigation mode of operation.
4574 *
4575 * WCN36XX_HAL_THERMAL_MITIGATION_MODE_0 - Based on AMPDU disabling aggregation
4576 *
4577 * WCN36XX_HAL_THERMAL_MITIGATION_MODE_1 - Based on AMPDU disabling aggregation
4578 * and reducing transmit power
4579 *
4580 * WCN36XX_HAL_THERMAL_MITIGATION_MODE_2 - Not supported */
4581enum wcn36xx_hal_thermal_mitigation_mode_type {
4582 HAL_THERMAL_MITIGATION_MODE_INVALID = -1,
4583 HAL_THERMAL_MITIGATION_MODE_0,
4584 HAL_THERMAL_MITIGATION_MODE_1,
4585 HAL_THERMAL_MITIGATION_MODE_2,
4586 HAL_THERMAL_MITIGATION_MODE_MAX = WCN36XX_HAL_MAX_ENUM_SIZE,
4587};
4588
4589
4590/*
4591 * Thermal Mitigation level.
4592 * Note the levels are incremental i.e WCN36XX_HAL_THERMAL_MITIGATION_LEVEL_2 =
4593 * WCN36XX_HAL_THERMAL_MITIGATION_LEVEL_0 +
4594 * WCN36XX_HAL_THERMAL_MITIGATION_LEVEL_1
4595 *
4596 * WCN36XX_HAL_THERMAL_MITIGATION_LEVEL_0 - lowest level of thermal mitigation.
4597 * This level indicates normal mode of operation
4598 *
4599 * WCN36XX_HAL_THERMAL_MITIGATION_LEVEL_1 - 1st level of thermal mitigation
4600 *
4601 * WCN36XX_HAL_THERMAL_MITIGATION_LEVEL_2 - 2nd level of thermal mitigation
4602 *
4603 * WCN36XX_HAL_THERMAL_MITIGATION_LEVEL_3 - 3rd level of thermal mitigation
4604 *
4605 * WCN36XX_HAL_THERMAL_MITIGATION_LEVEL_4 - 4th level of thermal mitigation
4606 */
4607enum wcn36xx_hal_thermal_mitigation_level_type {
4608 HAL_THERMAL_MITIGATION_LEVEL_INVALID = -1,
4609 HAL_THERMAL_MITIGATION_LEVEL_0,
4610 HAL_THERMAL_MITIGATION_LEVEL_1,
4611 HAL_THERMAL_MITIGATION_LEVEL_2,
4612 HAL_THERMAL_MITIGATION_LEVEL_3,
4613 HAL_THERMAL_MITIGATION_LEVEL_4,
4614 HAL_THERMAL_MITIGATION_LEVEL_MAX = WCN36XX_HAL_MAX_ENUM_SIZE,
4615};
4616
4617
4618/* WCN36XX_HAL_SET_THERMAL_MITIGATION_REQ */
4619struct set_thermal_mitigation_req_msg {
4620 struct wcn36xx_hal_msg_header header;
4621
4622 /* Thermal Mitigation Operation Mode */
4623 enum wcn36xx_hal_thermal_mitigation_mode_type mode;
4624
4625 /* Thermal Mitigation Level */
4626 enum wcn36xx_hal_thermal_mitigation_level_type level;
4627};
4628
4629struct set_thermal_mitigation_resp {
4630
4631 struct wcn36xx_hal_msg_header header;
4632
4633 /* status of the request */
4634 u32 status;
4635};
4636
4637/* Per STA Class B Statistics. Class B statistics are STA TX/RX stats
4638 * provided to FW from Host via periodic messages */
4639struct stats_class_b_ind {
4640 struct wcn36xx_hal_msg_header header;
4641
4642 /* Duration over which this stats was collected */
4643 u32 duration;
4644
4645 /* Per STA Stats */
4646
4647 /* TX stats */
4648 u32 tx_bytes_pushed;
4649 u32 tx_packets_pushed;
4650
4651 /* RX stats */
4652 u32 rx_bytes_rcvd;
4653 u32 rx_packets_rcvd;
4654 u32 rx_time_total;
4655};
4656
4657#endif /* _HAL_H_ */
diff --git a/drivers/net/wireless/ath/wcn36xx/main.c b/drivers/net/wireless/ath/wcn36xx/main.c
new file mode 100644
index 000000000000..7839b31e4826
--- /dev/null
+++ b/drivers/net/wireless/ath/wcn36xx/main.c
@@ -0,0 +1,1036 @@
1/*
2 * Copyright (c) 2013 Eugene Krasnikov <k.eugene.e@gmail.com>
3 *
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
7 *
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
11 * SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION
13 * OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN
14 * CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
15 */
16
17#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
18
19#include <linux/module.h>
20#include <linux/platform_device.h>
21#include "wcn36xx.h"
22
23unsigned int wcn36xx_dbg_mask;
24module_param_named(debug_mask, wcn36xx_dbg_mask, uint, 0644);
25MODULE_PARM_DESC(debug_mask, "Debugging mask");
26
27#define CHAN2G(_freq, _idx) { \
28 .band = IEEE80211_BAND_2GHZ, \
29 .center_freq = (_freq), \
30 .hw_value = (_idx), \
31 .max_power = 25, \
32}
33
34#define CHAN5G(_freq, _idx) { \
35 .band = IEEE80211_BAND_5GHZ, \
36 .center_freq = (_freq), \
37 .hw_value = (_idx), \
38 .max_power = 25, \
39}
40
41/* The wcn firmware expects channel values to matching
42 * their mnemonic values. So use these for .hw_value. */
43static struct ieee80211_channel wcn_2ghz_channels[] = {
44 CHAN2G(2412, 1), /* Channel 1 */
45 CHAN2G(2417, 2), /* Channel 2 */
46 CHAN2G(2422, 3), /* Channel 3 */
47 CHAN2G(2427, 4), /* Channel 4 */
48 CHAN2G(2432, 5), /* Channel 5 */
49 CHAN2G(2437, 6), /* Channel 6 */
50 CHAN2G(2442, 7), /* Channel 7 */
51 CHAN2G(2447, 8), /* Channel 8 */
52 CHAN2G(2452, 9), /* Channel 9 */
53 CHAN2G(2457, 10), /* Channel 10 */
54 CHAN2G(2462, 11), /* Channel 11 */
55 CHAN2G(2467, 12), /* Channel 12 */
56 CHAN2G(2472, 13), /* Channel 13 */
57 CHAN2G(2484, 14) /* Channel 14 */
58
59};
60
61static struct ieee80211_channel wcn_5ghz_channels[] = {
62 CHAN5G(5180, 36),
63 CHAN5G(5200, 40),
64 CHAN5G(5220, 44),
65 CHAN5G(5240, 48),
66 CHAN5G(5260, 52),
67 CHAN5G(5280, 56),
68 CHAN5G(5300, 60),
69 CHAN5G(5320, 64),
70 CHAN5G(5500, 100),
71 CHAN5G(5520, 104),
72 CHAN5G(5540, 108),
73 CHAN5G(5560, 112),
74 CHAN5G(5580, 116),
75 CHAN5G(5600, 120),
76 CHAN5G(5620, 124),
77 CHAN5G(5640, 128),
78 CHAN5G(5660, 132),
79 CHAN5G(5700, 140),
80 CHAN5G(5745, 149),
81 CHAN5G(5765, 153),
82 CHAN5G(5785, 157),
83 CHAN5G(5805, 161),
84 CHAN5G(5825, 165)
85};
86
87#define RATE(_bitrate, _hw_rate, _flags) { \
88 .bitrate = (_bitrate), \
89 .flags = (_flags), \
90 .hw_value = (_hw_rate), \
91 .hw_value_short = (_hw_rate) \
92}
93
94static struct ieee80211_rate wcn_2ghz_rates[] = {
95 RATE(10, HW_RATE_INDEX_1MBPS, 0),
96 RATE(20, HW_RATE_INDEX_2MBPS, IEEE80211_RATE_SHORT_PREAMBLE),
97 RATE(55, HW_RATE_INDEX_5_5MBPS, IEEE80211_RATE_SHORT_PREAMBLE),
98 RATE(110, HW_RATE_INDEX_11MBPS, IEEE80211_RATE_SHORT_PREAMBLE),
99 RATE(60, HW_RATE_INDEX_6MBPS, 0),
100 RATE(90, HW_RATE_INDEX_9MBPS, 0),
101 RATE(120, HW_RATE_INDEX_12MBPS, 0),
102 RATE(180, HW_RATE_INDEX_18MBPS, 0),
103 RATE(240, HW_RATE_INDEX_24MBPS, 0),
104 RATE(360, HW_RATE_INDEX_36MBPS, 0),
105 RATE(480, HW_RATE_INDEX_48MBPS, 0),
106 RATE(540, HW_RATE_INDEX_54MBPS, 0)
107};
108
109static struct ieee80211_rate wcn_5ghz_rates[] = {
110 RATE(60, HW_RATE_INDEX_6MBPS, 0),
111 RATE(90, HW_RATE_INDEX_9MBPS, 0),
112 RATE(120, HW_RATE_INDEX_12MBPS, 0),
113 RATE(180, HW_RATE_INDEX_18MBPS, 0),
114 RATE(240, HW_RATE_INDEX_24MBPS, 0),
115 RATE(360, HW_RATE_INDEX_36MBPS, 0),
116 RATE(480, HW_RATE_INDEX_48MBPS, 0),
117 RATE(540, HW_RATE_INDEX_54MBPS, 0)
118};
119
120static struct ieee80211_supported_band wcn_band_2ghz = {
121 .channels = wcn_2ghz_channels,
122 .n_channels = ARRAY_SIZE(wcn_2ghz_channels),
123 .bitrates = wcn_2ghz_rates,
124 .n_bitrates = ARRAY_SIZE(wcn_2ghz_rates),
125 .ht_cap = {
126 .cap = IEEE80211_HT_CAP_GRN_FLD |
127 IEEE80211_HT_CAP_SGI_20 |
128 IEEE80211_HT_CAP_DSSSCCK40 |
129 IEEE80211_HT_CAP_LSIG_TXOP_PROT,
130 .ht_supported = true,
131 .ampdu_factor = IEEE80211_HT_MAX_AMPDU_64K,
132 .ampdu_density = IEEE80211_HT_MPDU_DENSITY_16,
133 .mcs = {
134 .rx_mask = { 0xff, 0, 0, 0, 0, 0, 0, 0, 0, 0, },
135 .rx_highest = cpu_to_le16(72),
136 .tx_params = IEEE80211_HT_MCS_TX_DEFINED,
137 }
138 }
139};
140
141static struct ieee80211_supported_band wcn_band_5ghz = {
142 .channels = wcn_5ghz_channels,
143 .n_channels = ARRAY_SIZE(wcn_5ghz_channels),
144 .bitrates = wcn_5ghz_rates,
145 .n_bitrates = ARRAY_SIZE(wcn_5ghz_rates),
146 .ht_cap = {
147 .cap = IEEE80211_HT_CAP_GRN_FLD |
148 IEEE80211_HT_CAP_SGI_20 |
149 IEEE80211_HT_CAP_DSSSCCK40 |
150 IEEE80211_HT_CAP_LSIG_TXOP_PROT |
151 IEEE80211_HT_CAP_SGI_40 |
152 IEEE80211_HT_CAP_SUP_WIDTH_20_40,
153 .ht_supported = true,
154 .ampdu_factor = IEEE80211_HT_MAX_AMPDU_64K,
155 .ampdu_density = IEEE80211_HT_MPDU_DENSITY_16,
156 .mcs = {
157 .rx_mask = { 0xff, 0, 0, 0, 0, 0, 0, 0, 0, 0, },
158 .rx_highest = cpu_to_le16(72),
159 .tx_params = IEEE80211_HT_MCS_TX_DEFINED,
160 }
161 }
162};
163
164#ifdef CONFIG_PM
165
166static const struct wiphy_wowlan_support wowlan_support = {
167 .flags = WIPHY_WOWLAN_ANY
168};
169
170#endif
171
172static inline u8 get_sta_index(struct ieee80211_vif *vif,
173 struct wcn36xx_sta *sta_priv)
174{
175 return NL80211_IFTYPE_STATION == vif->type ?
176 sta_priv->bss_sta_index :
177 sta_priv->sta_index;
178}
179
180static int wcn36xx_start(struct ieee80211_hw *hw)
181{
182 struct wcn36xx *wcn = hw->priv;
183 int ret;
184
185 wcn36xx_dbg(WCN36XX_DBG_MAC, "mac start\n");
186
187 /* SMD initialization */
188 ret = wcn36xx_smd_open(wcn);
189 if (ret) {
190 wcn36xx_err("Failed to open smd channel: %d\n", ret);
191 goto out_err;
192 }
193
194 /* Allocate memory pools for Mgmt BD headers and Data BD headers */
195 ret = wcn36xx_dxe_allocate_mem_pools(wcn);
196 if (ret) {
197 wcn36xx_err("Failed to alloc DXE mempool: %d\n", ret);
198 goto out_smd_close;
199 }
200
201 ret = wcn36xx_dxe_alloc_ctl_blks(wcn);
202 if (ret) {
203 wcn36xx_err("Failed to alloc DXE ctl blocks: %d\n", ret);
204 goto out_free_dxe_pool;
205 }
206
207 wcn->hal_buf = kmalloc(WCN36XX_HAL_BUF_SIZE, GFP_KERNEL);
208 if (!wcn->hal_buf) {
209 wcn36xx_err("Failed to allocate smd buf\n");
210 ret = -ENOMEM;
211 goto out_free_dxe_ctl;
212 }
213
214 ret = wcn36xx_smd_load_nv(wcn);
215 if (ret) {
216 wcn36xx_err("Failed to push NV to chip\n");
217 goto out_free_smd_buf;
218 }
219
220 ret = wcn36xx_smd_start(wcn);
221 if (ret) {
222 wcn36xx_err("Failed to start chip\n");
223 goto out_free_smd_buf;
224 }
225
226 /* DMA channel initialization */
227 ret = wcn36xx_dxe_init(wcn);
228 if (ret) {
229 wcn36xx_err("DXE init failed\n");
230 goto out_smd_stop;
231 }
232
233 wcn36xx_debugfs_init(wcn);
234
235 if (!wcn36xx_is_fw_version(wcn, 1, 2, 2, 24)) {
236 ret = wcn36xx_smd_feature_caps_exchange(wcn);
237 if (ret)
238 wcn36xx_warn("Exchange feature caps failed\n");
239 }
240 INIT_LIST_HEAD(&wcn->vif_list);
241 return 0;
242
243out_smd_stop:
244 wcn36xx_smd_stop(wcn);
245out_free_smd_buf:
246 kfree(wcn->hal_buf);
247out_free_dxe_pool:
248 wcn36xx_dxe_free_mem_pools(wcn);
249out_free_dxe_ctl:
250 wcn36xx_dxe_free_ctl_blks(wcn);
251out_smd_close:
252 wcn36xx_smd_close(wcn);
253out_err:
254 return ret;
255}
256
257static void wcn36xx_stop(struct ieee80211_hw *hw)
258{
259 struct wcn36xx *wcn = hw->priv;
260
261 wcn36xx_dbg(WCN36XX_DBG_MAC, "mac stop\n");
262
263 wcn36xx_debugfs_exit(wcn);
264 wcn36xx_smd_stop(wcn);
265 wcn36xx_dxe_deinit(wcn);
266 wcn36xx_smd_close(wcn);
267
268 wcn36xx_dxe_free_mem_pools(wcn);
269 wcn36xx_dxe_free_ctl_blks(wcn);
270
271 kfree(wcn->hal_buf);
272}
273
274static int wcn36xx_config(struct ieee80211_hw *hw, u32 changed)
275{
276 struct wcn36xx *wcn = hw->priv;
277 struct ieee80211_vif *vif = NULL;
278 struct wcn36xx_vif *tmp;
279
280 wcn36xx_dbg(WCN36XX_DBG_MAC, "mac config changed 0x%08x\n", changed);
281
282 if (changed & IEEE80211_CONF_CHANGE_CHANNEL) {
283 int ch = WCN36XX_HW_CHANNEL(wcn);
284 wcn36xx_dbg(WCN36XX_DBG_MAC, "wcn36xx_config channel switch=%d\n",
285 ch);
286 list_for_each_entry(tmp, &wcn->vif_list, list) {
287 vif = container_of((void *)tmp,
288 struct ieee80211_vif,
289 drv_priv);
290 wcn36xx_smd_switch_channel(wcn, vif, ch);
291 }
292 }
293
294 return 0;
295}
296
297#define WCN36XX_SUPPORTED_FILTERS (0)
298
299static void wcn36xx_configure_filter(struct ieee80211_hw *hw,
300 unsigned int changed,
301 unsigned int *total, u64 multicast)
302{
303 wcn36xx_dbg(WCN36XX_DBG_MAC, "mac configure filter\n");
304
305 *total &= WCN36XX_SUPPORTED_FILTERS;
306}
307
308static void wcn36xx_tx(struct ieee80211_hw *hw,
309 struct ieee80211_tx_control *control,
310 struct sk_buff *skb)
311{
312 struct wcn36xx *wcn = hw->priv;
313 struct wcn36xx_sta *sta_priv = NULL;
314
315 if (control->sta)
316 sta_priv = (struct wcn36xx_sta *)control->sta->drv_priv;
317
318 if (wcn36xx_start_tx(wcn, sta_priv, skb))
319 ieee80211_free_txskb(wcn->hw, skb);
320}
321
322static int wcn36xx_set_key(struct ieee80211_hw *hw, enum set_key_cmd cmd,
323 struct ieee80211_vif *vif,
324 struct ieee80211_sta *sta,
325 struct ieee80211_key_conf *key_conf)
326{
327 struct wcn36xx *wcn = hw->priv;
328 struct wcn36xx_vif *vif_priv = (struct wcn36xx_vif *)vif->drv_priv;
329 struct wcn36xx_sta *sta_priv = vif_priv->sta;
330 int ret = 0;
331 u8 key[WLAN_MAX_KEY_LEN];
332
333 wcn36xx_dbg(WCN36XX_DBG_MAC, "mac80211 set key\n");
334 wcn36xx_dbg(WCN36XX_DBG_MAC, "Key: cmd=0x%x algo:0x%x, id:%d, len:%d flags 0x%x\n",
335 cmd, key_conf->cipher, key_conf->keyidx,
336 key_conf->keylen, key_conf->flags);
337 wcn36xx_dbg_dump(WCN36XX_DBG_MAC, "KEY: ",
338 key_conf->key,
339 key_conf->keylen);
340
341 switch (key_conf->cipher) {
342 case WLAN_CIPHER_SUITE_WEP40:
343 vif_priv->encrypt_type = WCN36XX_HAL_ED_WEP40;
344 break;
345 case WLAN_CIPHER_SUITE_WEP104:
346 vif_priv->encrypt_type = WCN36XX_HAL_ED_WEP40;
347 break;
348 case WLAN_CIPHER_SUITE_CCMP:
349 vif_priv->encrypt_type = WCN36XX_HAL_ED_CCMP;
350 break;
351 case WLAN_CIPHER_SUITE_TKIP:
352 vif_priv->encrypt_type = WCN36XX_HAL_ED_TKIP;
353 break;
354 default:
355 wcn36xx_err("Unsupported key type 0x%x\n",
356 key_conf->cipher);
357 ret = -EOPNOTSUPP;
358 goto out;
359 }
360
361 switch (cmd) {
362 case SET_KEY:
363 if (WCN36XX_HAL_ED_TKIP == vif_priv->encrypt_type) {
364 /*
365 * Supplicant is sending key in the wrong order:
366 * Temporal Key (16 b) - TX MIC (8 b) - RX MIC (8 b)
367 * but HW expects it to be in the order as described in
368 * IEEE 802.11 spec (see chapter 11.7) like this:
369 * Temporal Key (16 b) - RX MIC (8 b) - TX MIC (8 b)
370 */
371 memcpy(key, key_conf->key, 16);
372 memcpy(key + 16, key_conf->key + 24, 8);
373 memcpy(key + 24, key_conf->key + 16, 8);
374 } else {
375 memcpy(key, key_conf->key, key_conf->keylen);
376 }
377
378 if (IEEE80211_KEY_FLAG_PAIRWISE & key_conf->flags) {
379 sta_priv->is_data_encrypted = true;
380 /* Reconfigure bss with encrypt_type */
381 if (NL80211_IFTYPE_STATION == vif->type)
382 wcn36xx_smd_config_bss(wcn,
383 vif,
384 sta,
385 sta->addr,
386 true);
387
388 wcn36xx_smd_set_stakey(wcn,
389 vif_priv->encrypt_type,
390 key_conf->keyidx,
391 key_conf->keylen,
392 key,
393 get_sta_index(vif, sta_priv));
394 } else {
395 wcn36xx_smd_set_bsskey(wcn,
396 vif_priv->encrypt_type,
397 key_conf->keyidx,
398 key_conf->keylen,
399 key);
400 if ((WLAN_CIPHER_SUITE_WEP40 == key_conf->cipher) ||
401 (WLAN_CIPHER_SUITE_WEP104 == key_conf->cipher)) {
402 sta_priv->is_data_encrypted = true;
403 wcn36xx_smd_set_stakey(wcn,
404 vif_priv->encrypt_type,
405 key_conf->keyidx,
406 key_conf->keylen,
407 key,
408 get_sta_index(vif, sta_priv));
409 }
410 }
411 break;
412 case DISABLE_KEY:
413 if (!(IEEE80211_KEY_FLAG_PAIRWISE & key_conf->flags)) {
414 wcn36xx_smd_remove_bsskey(wcn,
415 vif_priv->encrypt_type,
416 key_conf->keyidx);
417 } else {
418 sta_priv->is_data_encrypted = false;
419 /* do not remove key if disassociated */
420 if (sta_priv->aid)
421 wcn36xx_smd_remove_stakey(wcn,
422 vif_priv->encrypt_type,
423 key_conf->keyidx,
424 get_sta_index(vif, sta_priv));
425 }
426 break;
427 default:
428 wcn36xx_err("Unsupported key cmd 0x%x\n", cmd);
429 ret = -EOPNOTSUPP;
430 goto out;
431 break;
432 }
433
434out:
435 return ret;
436}
437
438static void wcn36xx_sw_scan_start(struct ieee80211_hw *hw)
439{
440 struct wcn36xx *wcn = hw->priv;
441
442 wcn36xx_smd_init_scan(wcn, HAL_SYS_MODE_SCAN);
443 wcn36xx_smd_start_scan(wcn);
444}
445
446static void wcn36xx_sw_scan_complete(struct ieee80211_hw *hw)
447{
448 struct wcn36xx *wcn = hw->priv;
449
450 wcn36xx_smd_end_scan(wcn);
451 wcn36xx_smd_finish_scan(wcn, HAL_SYS_MODE_SCAN);
452}
453
454static void wcn36xx_update_allowed_rates(struct ieee80211_sta *sta,
455 enum ieee80211_band band)
456{
457 int i, size;
458 u16 *rates_table;
459 struct wcn36xx_sta *sta_priv = (struct wcn36xx_sta *)sta->drv_priv;
460 u32 rates = sta->supp_rates[band];
461
462 memset(&sta_priv->supported_rates, 0,
463 sizeof(sta_priv->supported_rates));
464 sta_priv->supported_rates.op_rate_mode = STA_11n;
465
466 size = ARRAY_SIZE(sta_priv->supported_rates.dsss_rates);
467 rates_table = sta_priv->supported_rates.dsss_rates;
468 if (band == IEEE80211_BAND_2GHZ) {
469 for (i = 0; i < size; i++) {
470 if (rates & 0x01) {
471 rates_table[i] = wcn_2ghz_rates[i].hw_value;
472 rates = rates >> 1;
473 }
474 }
475 }
476
477 size = ARRAY_SIZE(sta_priv->supported_rates.ofdm_rates);
478 rates_table = sta_priv->supported_rates.ofdm_rates;
479 for (i = 0; i < size; i++) {
480 if (rates & 0x01) {
481 rates_table[i] = wcn_5ghz_rates[i].hw_value;
482 rates = rates >> 1;
483 }
484 }
485
486 if (sta->ht_cap.ht_supported) {
487 BUILD_BUG_ON(sizeof(sta->ht_cap.mcs.rx_mask) >
488 sizeof(sta_priv->supported_rates.supported_mcs_set));
489 memcpy(sta_priv->supported_rates.supported_mcs_set,
490 sta->ht_cap.mcs.rx_mask,
491 sizeof(sta->ht_cap.mcs.rx_mask));
492 }
493}
494void wcn36xx_set_default_rates(struct wcn36xx_hal_supported_rates *rates)
495{
496 u16 ofdm_rates[WCN36XX_HAL_NUM_OFDM_RATES] = {
497 HW_RATE_INDEX_6MBPS,
498 HW_RATE_INDEX_9MBPS,
499 HW_RATE_INDEX_12MBPS,
500 HW_RATE_INDEX_18MBPS,
501 HW_RATE_INDEX_24MBPS,
502 HW_RATE_INDEX_36MBPS,
503 HW_RATE_INDEX_48MBPS,
504 HW_RATE_INDEX_54MBPS
505 };
506 u16 dsss_rates[WCN36XX_HAL_NUM_DSSS_RATES] = {
507 HW_RATE_INDEX_1MBPS,
508 HW_RATE_INDEX_2MBPS,
509 HW_RATE_INDEX_5_5MBPS,
510 HW_RATE_INDEX_11MBPS
511 };
512
513 rates->op_rate_mode = STA_11n;
514 memcpy(rates->dsss_rates, dsss_rates,
515 sizeof(*dsss_rates) * WCN36XX_HAL_NUM_DSSS_RATES);
516 memcpy(rates->ofdm_rates, ofdm_rates,
517 sizeof(*ofdm_rates) * WCN36XX_HAL_NUM_OFDM_RATES);
518 rates->supported_mcs_set[0] = 0xFF;
519}
520static void wcn36xx_bss_info_changed(struct ieee80211_hw *hw,
521 struct ieee80211_vif *vif,
522 struct ieee80211_bss_conf *bss_conf,
523 u32 changed)
524{
525 struct wcn36xx *wcn = hw->priv;
526 struct sk_buff *skb = NULL;
527 u16 tim_off, tim_len;
528 enum wcn36xx_hal_link_state link_state;
529 struct wcn36xx_vif *vif_priv = (struct wcn36xx_vif *)vif->drv_priv;
530
531 wcn36xx_dbg(WCN36XX_DBG_MAC, "mac bss info changed vif %p changed 0x%08x\n",
532 vif, changed);
533
534 if (changed & BSS_CHANGED_BEACON_INFO) {
535 wcn36xx_dbg(WCN36XX_DBG_MAC,
536 "mac bss changed dtim period %d\n",
537 bss_conf->dtim_period);
538
539 vif_priv->dtim_period = bss_conf->dtim_period;
540 }
541
542 if (changed & BSS_CHANGED_PS) {
543 wcn36xx_dbg(WCN36XX_DBG_MAC,
544 "mac bss PS set %d\n",
545 bss_conf->ps);
546 if (bss_conf->ps) {
547 wcn36xx_pmc_enter_bmps_state(wcn, vif);
548 } else {
549 wcn36xx_pmc_exit_bmps_state(wcn, vif);
550 }
551 }
552
553 if (changed & BSS_CHANGED_BSSID) {
554 wcn36xx_dbg(WCN36XX_DBG_MAC, "mac bss changed_bssid %pM\n",
555 bss_conf->bssid);
556
557 if (!is_zero_ether_addr(bss_conf->bssid)) {
558 vif_priv->is_joining = true;
559 vif_priv->bss_index = 0xff;
560 wcn36xx_smd_join(wcn, bss_conf->bssid,
561 vif->addr, WCN36XX_HW_CHANNEL(wcn));
562 wcn36xx_smd_config_bss(wcn, vif, NULL,
563 bss_conf->bssid, false);
564 } else {
565 vif_priv->is_joining = false;
566 wcn36xx_smd_delete_bss(wcn, vif);
567 }
568 }
569
570 if (changed & BSS_CHANGED_SSID) {
571 wcn36xx_dbg(WCN36XX_DBG_MAC,
572 "mac bss changed ssid\n");
573 wcn36xx_dbg_dump(WCN36XX_DBG_MAC, "ssid ",
574 bss_conf->ssid, bss_conf->ssid_len);
575
576 vif_priv->ssid.length = bss_conf->ssid_len;
577 memcpy(&vif_priv->ssid.ssid,
578 bss_conf->ssid,
579 bss_conf->ssid_len);
580 }
581
582 if (changed & BSS_CHANGED_ASSOC) {
583 vif_priv->is_joining = false;
584 if (bss_conf->assoc) {
585 struct ieee80211_sta *sta;
586 struct wcn36xx_sta *sta_priv;
587
588 wcn36xx_dbg(WCN36XX_DBG_MAC,
589 "mac assoc bss %pM vif %pM AID=%d\n",
590 bss_conf->bssid,
591 vif->addr,
592 bss_conf->aid);
593
594 rcu_read_lock();
595 sta = ieee80211_find_sta(vif, bss_conf->bssid);
596 if (!sta) {
597 wcn36xx_err("sta %pM is not found\n",
598 bss_conf->bssid);
599 rcu_read_unlock();
600 goto out;
601 }
602 sta_priv = (struct wcn36xx_sta *)sta->drv_priv;
603
604 wcn36xx_update_allowed_rates(sta, WCN36XX_BAND(wcn));
605
606 wcn36xx_smd_set_link_st(wcn, bss_conf->bssid,
607 vif->addr,
608 WCN36XX_HAL_LINK_POSTASSOC_STATE);
609 wcn36xx_smd_config_bss(wcn, vif, sta,
610 bss_conf->bssid,
611 true);
612 sta_priv->aid = bss_conf->aid;
613 /*
614 * config_sta must be called from because this is the
615 * place where AID is available.
616 */
617 wcn36xx_smd_config_sta(wcn, vif, sta);
618 rcu_read_unlock();
619 } else {
620 wcn36xx_dbg(WCN36XX_DBG_MAC,
621 "disassociated bss %pM vif %pM AID=%d\n",
622 bss_conf->bssid,
623 vif->addr,
624 bss_conf->aid);
625 wcn36xx_smd_set_link_st(wcn,
626 bss_conf->bssid,
627 vif->addr,
628 WCN36XX_HAL_LINK_IDLE_STATE);
629 }
630 }
631
632 if (changed & BSS_CHANGED_AP_PROBE_RESP) {
633 wcn36xx_dbg(WCN36XX_DBG_MAC, "mac bss changed ap probe resp\n");
634 skb = ieee80211_proberesp_get(hw, vif);
635 if (!skb) {
636 wcn36xx_err("failed to alloc probereq skb\n");
637 goto out;
638 }
639
640 wcn36xx_smd_update_proberesp_tmpl(wcn, vif, skb);
641 dev_kfree_skb(skb);
642 }
643
644 if (changed & BSS_CHANGED_BEACON_ENABLED) {
645 wcn36xx_dbg(WCN36XX_DBG_MAC,
646 "mac bss changed beacon enabled %d\n",
647 bss_conf->enable_beacon);
648
649 if (bss_conf->enable_beacon) {
650 vif_priv->bss_index = 0xff;
651 wcn36xx_smd_config_bss(wcn, vif, NULL,
652 vif->addr, false);
653 skb = ieee80211_beacon_get_tim(hw, vif, &tim_off,
654 &tim_len);
655 if (!skb) {
656 wcn36xx_err("failed to alloc beacon skb\n");
657 goto out;
658 }
659 wcn36xx_smd_send_beacon(wcn, vif, skb, tim_off, 0);
660 dev_kfree_skb(skb);
661
662 if (vif->type == NL80211_IFTYPE_ADHOC ||
663 vif->type == NL80211_IFTYPE_MESH_POINT)
664 link_state = WCN36XX_HAL_LINK_IBSS_STATE;
665 else
666 link_state = WCN36XX_HAL_LINK_AP_STATE;
667
668 wcn36xx_smd_set_link_st(wcn, vif->addr, vif->addr,
669 link_state);
670 } else {
671 wcn36xx_smd_set_link_st(wcn, vif->addr, vif->addr,
672 WCN36XX_HAL_LINK_IDLE_STATE);
673 wcn36xx_smd_delete_bss(wcn, vif);
674 }
675 }
676out:
677 return;
678}
679
680/* this is required when using IEEE80211_HW_HAS_RATE_CONTROL */
681static int wcn36xx_set_rts_threshold(struct ieee80211_hw *hw, u32 value)
682{
683 struct wcn36xx *wcn = hw->priv;
684 wcn36xx_dbg(WCN36XX_DBG_MAC, "mac set RTS threshold %d\n", value);
685
686 wcn36xx_smd_update_cfg(wcn, WCN36XX_HAL_CFG_RTS_THRESHOLD, value);
687 return 0;
688}
689
690static void wcn36xx_remove_interface(struct ieee80211_hw *hw,
691 struct ieee80211_vif *vif)
692{
693 struct wcn36xx *wcn = hw->priv;
694 struct wcn36xx_vif *vif_priv = (struct wcn36xx_vif *)vif->drv_priv;
695 wcn36xx_dbg(WCN36XX_DBG_MAC, "mac remove interface vif %p\n", vif);
696
697 list_del(&vif_priv->list);
698 wcn36xx_smd_delete_sta_self(wcn, vif->addr);
699}
700
701static int wcn36xx_add_interface(struct ieee80211_hw *hw,
702 struct ieee80211_vif *vif)
703{
704 struct wcn36xx *wcn = hw->priv;
705 struct wcn36xx_vif *vif_priv = (struct wcn36xx_vif *)vif->drv_priv;
706
707 wcn36xx_dbg(WCN36XX_DBG_MAC, "mac add interface vif %p type %d\n",
708 vif, vif->type);
709
710 if (!(NL80211_IFTYPE_STATION == vif->type ||
711 NL80211_IFTYPE_AP == vif->type ||
712 NL80211_IFTYPE_ADHOC == vif->type ||
713 NL80211_IFTYPE_MESH_POINT == vif->type)) {
714 wcn36xx_warn("Unsupported interface type requested: %d\n",
715 vif->type);
716 return -EOPNOTSUPP;
717 }
718
719 list_add(&vif_priv->list, &wcn->vif_list);
720 wcn36xx_smd_add_sta_self(wcn, vif);
721
722 return 0;
723}
724
725static int wcn36xx_sta_add(struct ieee80211_hw *hw, struct ieee80211_vif *vif,
726 struct ieee80211_sta *sta)
727{
728 struct wcn36xx *wcn = hw->priv;
729 struct wcn36xx_vif *vif_priv = (struct wcn36xx_vif *)vif->drv_priv;
730 struct wcn36xx_sta *sta_priv = (struct wcn36xx_sta *)sta->drv_priv;
731 wcn36xx_dbg(WCN36XX_DBG_MAC, "mac sta add vif %p sta %pM\n",
732 vif, sta->addr);
733
734 vif_priv->sta = sta_priv;
735 sta_priv->vif = vif_priv;
736 /*
737 * For STA mode HW will be configured on BSS_CHANGED_ASSOC because
738 * at this stage AID is not available yet.
739 */
740 if (NL80211_IFTYPE_STATION != vif->type) {
741 wcn36xx_update_allowed_rates(sta, WCN36XX_BAND(wcn));
742 sta_priv->aid = sta->aid;
743 wcn36xx_smd_config_sta(wcn, vif, sta);
744 }
745 return 0;
746}
747
748static int wcn36xx_sta_remove(struct ieee80211_hw *hw,
749 struct ieee80211_vif *vif,
750 struct ieee80211_sta *sta)
751{
752 struct wcn36xx *wcn = hw->priv;
753 struct wcn36xx_vif *vif_priv = (struct wcn36xx_vif *)vif->drv_priv;
754 struct wcn36xx_sta *sta_priv = (struct wcn36xx_sta *)sta->drv_priv;
755
756 wcn36xx_dbg(WCN36XX_DBG_MAC, "mac sta remove vif %p sta %pM index %d\n",
757 vif, sta->addr, sta_priv->sta_index);
758
759 wcn36xx_smd_delete_sta(wcn, sta_priv->sta_index);
760 vif_priv->sta = NULL;
761 sta_priv->vif = NULL;
762 return 0;
763}
764
765#ifdef CONFIG_PM
766
767static int wcn36xx_suspend(struct ieee80211_hw *hw, struct cfg80211_wowlan *wow)
768{
769 struct wcn36xx *wcn = hw->priv;
770
771 wcn36xx_dbg(WCN36XX_DBG_MAC, "mac suspend\n");
772
773 flush_workqueue(wcn->hal_ind_wq);
774 wcn36xx_smd_set_power_params(wcn, true);
775 return 0;
776}
777
778static int wcn36xx_resume(struct ieee80211_hw *hw)
779{
780 struct wcn36xx *wcn = hw->priv;
781
782 wcn36xx_dbg(WCN36XX_DBG_MAC, "mac resume\n");
783
784 flush_workqueue(wcn->hal_ind_wq);
785 wcn36xx_smd_set_power_params(wcn, false);
786 return 0;
787}
788
789#endif
790
791static int wcn36xx_ampdu_action(struct ieee80211_hw *hw,
792 struct ieee80211_vif *vif,
793 enum ieee80211_ampdu_mlme_action action,
794 struct ieee80211_sta *sta, u16 tid, u16 *ssn,
795 u8 buf_size)
796{
797 struct wcn36xx *wcn = hw->priv;
798 struct wcn36xx_sta *sta_priv = NULL;
799
800 wcn36xx_dbg(WCN36XX_DBG_MAC, "mac ampdu action action %d tid %d\n",
801 action, tid);
802
803 sta_priv = (struct wcn36xx_sta *)sta->drv_priv;
804
805 switch (action) {
806 case IEEE80211_AMPDU_RX_START:
807 sta_priv->tid = tid;
808 wcn36xx_smd_add_ba_session(wcn, sta, tid, ssn, 0,
809 get_sta_index(vif, sta_priv));
810 wcn36xx_smd_add_ba(wcn);
811 wcn36xx_smd_trigger_ba(wcn, get_sta_index(vif, sta_priv));
812 ieee80211_start_tx_ba_session(sta, tid, 0);
813 break;
814 case IEEE80211_AMPDU_RX_STOP:
815 wcn36xx_smd_del_ba(wcn, tid, get_sta_index(vif, sta_priv));
816 break;
817 case IEEE80211_AMPDU_TX_START:
818 ieee80211_start_tx_ba_cb_irqsafe(vif, sta->addr, tid);
819 break;
820 case IEEE80211_AMPDU_TX_OPERATIONAL:
821 wcn36xx_smd_add_ba_session(wcn, sta, tid, ssn, 1,
822 get_sta_index(vif, sta_priv));
823 break;
824 case IEEE80211_AMPDU_TX_STOP_FLUSH:
825 case IEEE80211_AMPDU_TX_STOP_FLUSH_CONT:
826 case IEEE80211_AMPDU_TX_STOP_CONT:
827 ieee80211_stop_tx_ba_cb_irqsafe(vif, sta->addr, tid);
828 break;
829 default:
830 wcn36xx_err("Unknown AMPDU action\n");
831 }
832
833 return 0;
834}
835
836static const struct ieee80211_ops wcn36xx_ops = {
837 .start = wcn36xx_start,
838 .stop = wcn36xx_stop,
839 .add_interface = wcn36xx_add_interface,
840 .remove_interface = wcn36xx_remove_interface,
841#ifdef CONFIG_PM
842 .suspend = wcn36xx_suspend,
843 .resume = wcn36xx_resume,
844#endif
845 .config = wcn36xx_config,
846 .configure_filter = wcn36xx_configure_filter,
847 .tx = wcn36xx_tx,
848 .set_key = wcn36xx_set_key,
849 .sw_scan_start = wcn36xx_sw_scan_start,
850 .sw_scan_complete = wcn36xx_sw_scan_complete,
851 .bss_info_changed = wcn36xx_bss_info_changed,
852 .set_rts_threshold = wcn36xx_set_rts_threshold,
853 .sta_add = wcn36xx_sta_add,
854 .sta_remove = wcn36xx_sta_remove,
855 .ampdu_action = wcn36xx_ampdu_action,
856};
857
858static int wcn36xx_init_ieee80211(struct wcn36xx *wcn)
859{
860 int ret = 0;
861
862 static const u32 cipher_suites[] = {
863 WLAN_CIPHER_SUITE_WEP40,
864 WLAN_CIPHER_SUITE_WEP104,
865 WLAN_CIPHER_SUITE_TKIP,
866 WLAN_CIPHER_SUITE_CCMP,
867 };
868
869 wcn->hw->flags = IEEE80211_HW_SIGNAL_DBM |
870 IEEE80211_HW_HAS_RATE_CONTROL |
871 IEEE80211_HW_SUPPORTS_PS |
872 IEEE80211_HW_CONNECTION_MONITOR |
873 IEEE80211_HW_AMPDU_AGGREGATION |
874 IEEE80211_HW_TIMING_BEACON_ONLY;
875
876 wcn->hw->wiphy->interface_modes = BIT(NL80211_IFTYPE_STATION) |
877 BIT(NL80211_IFTYPE_AP) |
878 BIT(NL80211_IFTYPE_ADHOC) |
879 BIT(NL80211_IFTYPE_MESH_POINT);
880
881 wcn->hw->wiphy->bands[IEEE80211_BAND_2GHZ] = &wcn_band_2ghz;
882 wcn->hw->wiphy->bands[IEEE80211_BAND_5GHZ] = &wcn_band_5ghz;
883
884 wcn->hw->wiphy->cipher_suites = cipher_suites;
885 wcn->hw->wiphy->n_cipher_suites = ARRAY_SIZE(cipher_suites);
886
887 wcn->hw->wiphy->flags |= WIPHY_FLAG_AP_PROBE_RESP_OFFLOAD;
888
889#ifdef CONFIG_PM
890 wcn->hw->wiphy->wowlan = &wowlan_support;
891#endif
892
893 wcn->hw->max_listen_interval = 200;
894
895 wcn->hw->queues = 4;
896
897 SET_IEEE80211_DEV(wcn->hw, wcn->dev);
898
899 wcn->hw->sta_data_size = sizeof(struct wcn36xx_sta);
900 wcn->hw->vif_data_size = sizeof(struct wcn36xx_vif);
901
902 return ret;
903}
904
905static int wcn36xx_platform_get_resources(struct wcn36xx *wcn,
906 struct platform_device *pdev)
907{
908 struct resource *res;
909 /* Set TX IRQ */
910 res = platform_get_resource_byname(pdev, IORESOURCE_IRQ,
911 "wcnss_wlantx_irq");
912 if (!res) {
913 wcn36xx_err("failed to get tx_irq\n");
914 return -ENOENT;
915 }
916 wcn->tx_irq = res->start;
917
918 /* Set RX IRQ */
919 res = platform_get_resource_byname(pdev, IORESOURCE_IRQ,
920 "wcnss_wlanrx_irq");
921 if (!res) {
922 wcn36xx_err("failed to get rx_irq\n");
923 return -ENOENT;
924 }
925 wcn->rx_irq = res->start;
926
927 /* Map the memory */
928 res = platform_get_resource_byname(pdev, IORESOURCE_MEM,
929 "wcnss_mmio");
930 if (!res) {
931 wcn36xx_err("failed to get mmio\n");
932 return -ENOENT;
933 }
934 wcn->mmio = ioremap(res->start, resource_size(res));
935 if (!wcn->mmio) {
936 wcn36xx_err("failed to map io memory\n");
937 return -ENOMEM;
938 }
939 return 0;
940}
941
942static int wcn36xx_probe(struct platform_device *pdev)
943{
944 struct ieee80211_hw *hw;
945 struct wcn36xx *wcn;
946 int ret;
947 u8 addr[ETH_ALEN];
948
949 wcn36xx_dbg(WCN36XX_DBG_MAC, "platform probe\n");
950
951 hw = ieee80211_alloc_hw(sizeof(struct wcn36xx), &wcn36xx_ops);
952 if (!hw) {
953 wcn36xx_err("failed to alloc hw\n");
954 ret = -ENOMEM;
955 goto out_err;
956 }
957 platform_set_drvdata(pdev, hw);
958 wcn = hw->priv;
959 wcn->hw = hw;
960 wcn->dev = &pdev->dev;
961 wcn->ctrl_ops = pdev->dev.platform_data;
962
963 mutex_init(&wcn->hal_mutex);
964
965 if (!wcn->ctrl_ops->get_hw_mac(addr)) {
966 wcn36xx_info("mac address: %pM\n", addr);
967 SET_IEEE80211_PERM_ADDR(wcn->hw, addr);
968 }
969
970 ret = wcn36xx_platform_get_resources(wcn, pdev);
971 if (ret)
972 goto out_wq;
973
974 wcn36xx_init_ieee80211(wcn);
975 ret = ieee80211_register_hw(wcn->hw);
976 if (ret)
977 goto out_unmap;
978
979 return 0;
980
981out_unmap:
982 iounmap(wcn->mmio);
983out_wq:
984 ieee80211_free_hw(hw);
985out_err:
986 return ret;
987}
988static int wcn36xx_remove(struct platform_device *pdev)
989{
990 struct ieee80211_hw *hw = platform_get_drvdata(pdev);
991 struct wcn36xx *wcn = hw->priv;
992 wcn36xx_dbg(WCN36XX_DBG_MAC, "platform remove\n");
993
994 mutex_destroy(&wcn->hal_mutex);
995
996 ieee80211_unregister_hw(hw);
997 iounmap(wcn->mmio);
998 ieee80211_free_hw(hw);
999
1000 return 0;
1001}
1002static const struct platform_device_id wcn36xx_platform_id_table[] = {
1003 {
1004 .name = "wcn36xx",
1005 .driver_data = 0
1006 },
1007 {}
1008};
1009MODULE_DEVICE_TABLE(platform, wcn36xx_platform_id_table);
1010
1011static struct platform_driver wcn36xx_driver = {
1012 .probe = wcn36xx_probe,
1013 .remove = wcn36xx_remove,
1014 .driver = {
1015 .name = "wcn36xx",
1016 .owner = THIS_MODULE,
1017 },
1018 .id_table = wcn36xx_platform_id_table,
1019};
1020
1021static int __init wcn36xx_init(void)
1022{
1023 platform_driver_register(&wcn36xx_driver);
1024 return 0;
1025}
1026module_init(wcn36xx_init);
1027
1028static void __exit wcn36xx_exit(void)
1029{
1030 platform_driver_unregister(&wcn36xx_driver);
1031}
1032module_exit(wcn36xx_exit);
1033
1034MODULE_LICENSE("Dual BSD/GPL");
1035MODULE_AUTHOR("Eugene Krasnikov k.eugene.e@gmail.com");
1036MODULE_FIRMWARE(WLAN_NV_FILE);
diff --git a/drivers/net/wireless/ath/wcn36xx/pmc.c b/drivers/net/wireless/ath/wcn36xx/pmc.c
new file mode 100644
index 000000000000..28b515c81b0e
--- /dev/null
+++ b/drivers/net/wireless/ath/wcn36xx/pmc.c
@@ -0,0 +1,62 @@
1/*
2 * Copyright (c) 2013 Eugene Krasnikov <k.eugene.e@gmail.com>
3 *
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
7 *
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
11 * SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION
13 * OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN
14 * CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
15 */
16
17#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
18
19#include "wcn36xx.h"
20
21int wcn36xx_pmc_enter_bmps_state(struct wcn36xx *wcn,
22 struct ieee80211_vif *vif)
23{
24 int ret = 0;
25 struct wcn36xx_vif *vif_priv = (struct wcn36xx_vif *)vif->drv_priv;
26 /* TODO: Make sure the TX chain clean */
27 ret = wcn36xx_smd_enter_bmps(wcn, vif);
28 if (!ret) {
29 wcn36xx_dbg(WCN36XX_DBG_PMC, "Entered BMPS\n");
30 vif_priv->pw_state = WCN36XX_BMPS;
31 } else {
32 /*
33 * One of the reasons why HW will not enter BMPS is because
34 * driver is trying to enter bmps before first beacon was
35 * received just after auth complete
36 */
37 wcn36xx_err("Can not enter BMPS!\n");
38 }
39 return ret;
40}
41
42int wcn36xx_pmc_exit_bmps_state(struct wcn36xx *wcn,
43 struct ieee80211_vif *vif)
44{
45 struct wcn36xx_vif *vif_priv = (struct wcn36xx_vif *)vif->drv_priv;
46
47 if (WCN36XX_BMPS != vif_priv->pw_state) {
48 wcn36xx_err("Not in BMPS mode, no need to exit from BMPS mode!\n");
49 return -EINVAL;
50 }
51 wcn36xx_smd_exit_bmps(wcn, vif);
52 vif_priv->pw_state = WCN36XX_FULL_POWER;
53 return 0;
54}
55
56int wcn36xx_enable_keep_alive_null_packet(struct wcn36xx *wcn,
57 struct ieee80211_vif *vif)
58{
59 wcn36xx_dbg(WCN36XX_DBG_PMC, "%s\n", __func__);
60 return wcn36xx_smd_keep_alive_req(wcn, vif,
61 WCN36XX_HAL_KEEP_ALIVE_NULL_PKT);
62}
diff --git a/drivers/net/wireless/ath/wcn36xx/pmc.h b/drivers/net/wireless/ath/wcn36xx/pmc.h
new file mode 100644
index 000000000000..f72ed68b5a07
--- /dev/null
+++ b/drivers/net/wireless/ath/wcn36xx/pmc.h
@@ -0,0 +1,33 @@
1/*
2 * Copyright (c) 2013 Eugene Krasnikov <k.eugene.e@gmail.com>
3 *
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
7 *
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
11 * SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION
13 * OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN
14 * CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
15 */
16
17#ifndef _WCN36XX_PMC_H_
18#define _WCN36XX_PMC_H_
19
20struct wcn36xx;
21
22enum wcn36xx_power_state {
23 WCN36XX_FULL_POWER,
24 WCN36XX_BMPS
25};
26
27int wcn36xx_pmc_enter_bmps_state(struct wcn36xx *wcn,
28 struct ieee80211_vif *vif);
29int wcn36xx_pmc_exit_bmps_state(struct wcn36xx *wcn,
30 struct ieee80211_vif *vif);
31int wcn36xx_enable_keep_alive_null_packet(struct wcn36xx *wcn,
32 struct ieee80211_vif *vif);
33#endif /* _WCN36XX_PMC_H_ */
diff --git a/drivers/net/wireless/ath/wcn36xx/smd.c b/drivers/net/wireless/ath/wcn36xx/smd.c
new file mode 100644
index 000000000000..f8c3a10510c2
--- /dev/null
+++ b/drivers/net/wireless/ath/wcn36xx/smd.c
@@ -0,0 +1,2126 @@
1/*
2 * Copyright (c) 2013 Eugene Krasnikov <k.eugene.e@gmail.com>
3 *
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
7 *
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
11 * SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION
13 * OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN
14 * CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
15 */
16
17#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
18
19#include <linux/etherdevice.h>
20#include <linux/firmware.h>
21#include <linux/bitops.h>
22#include "smd.h"
23
24static int put_cfg_tlv_u32(struct wcn36xx *wcn, size_t *len, u32 id, u32 value)
25{
26 struct wcn36xx_hal_cfg *entry;
27 u32 *val;
28
29 if (*len + sizeof(*entry) + sizeof(u32) >= WCN36XX_HAL_BUF_SIZE) {
30 wcn36xx_err("Not enough room for TLV entry\n");
31 return -ENOMEM;
32 }
33
34 entry = (struct wcn36xx_hal_cfg *) (wcn->hal_buf + *len);
35 entry->id = id;
36 entry->len = sizeof(u32);
37 entry->pad_bytes = 0;
38 entry->reserve = 0;
39
40 val = (u32 *) (entry + 1);
41 *val = value;
42
43 *len += sizeof(*entry) + sizeof(u32);
44
45 return 0;
46}
47
48static void wcn36xx_smd_set_bss_nw_type(struct wcn36xx *wcn,
49 struct ieee80211_sta *sta,
50 struct wcn36xx_hal_config_bss_params *bss_params)
51{
52 if (IEEE80211_BAND_5GHZ == WCN36XX_BAND(wcn))
53 bss_params->nw_type = WCN36XX_HAL_11A_NW_TYPE;
54 else if (sta && sta->ht_cap.ht_supported)
55 bss_params->nw_type = WCN36XX_HAL_11N_NW_TYPE;
56 else if (sta && (sta->supp_rates[IEEE80211_BAND_2GHZ] & 0x7f))
57 bss_params->nw_type = WCN36XX_HAL_11G_NW_TYPE;
58 else
59 bss_params->nw_type = WCN36XX_HAL_11B_NW_TYPE;
60}
61
62static inline u8 is_cap_supported(unsigned long caps, unsigned long flag)
63{
64 return caps & flag ? 1 : 0;
65}
66static void wcn36xx_smd_set_bss_ht_params(struct ieee80211_vif *vif,
67 struct ieee80211_sta *sta,
68 struct wcn36xx_hal_config_bss_params *bss_params)
69{
70 if (sta && sta->ht_cap.ht_supported) {
71 unsigned long caps = sta->ht_cap.cap;
72 bss_params->ht = sta->ht_cap.ht_supported;
73 bss_params->tx_channel_width_set = is_cap_supported(caps,
74 IEEE80211_HT_CAP_SUP_WIDTH_20_40);
75 bss_params->lsig_tx_op_protection_full_support =
76 is_cap_supported(caps,
77 IEEE80211_HT_CAP_LSIG_TXOP_PROT);
78
79 bss_params->ht_oper_mode = vif->bss_conf.ht_operation_mode;
80 bss_params->lln_non_gf_coexist =
81 !!(vif->bss_conf.ht_operation_mode &
82 IEEE80211_HT_OP_MODE_NON_GF_STA_PRSNT);
83 /* IEEE80211_HT_STBC_PARAM_DUAL_CTS_PROT */
84 bss_params->dual_cts_protection = 0;
85 /* IEEE80211_HT_OP_MODE_PROTECTION_20MHZ */
86 bss_params->ht20_coexist = 0;
87 }
88}
89
90static void wcn36xx_smd_set_sta_ht_params(struct ieee80211_sta *sta,
91 struct wcn36xx_hal_config_sta_params *sta_params)
92{
93 if (sta->ht_cap.ht_supported) {
94 unsigned long caps = sta->ht_cap.cap;
95 sta_params->ht_capable = sta->ht_cap.ht_supported;
96 sta_params->tx_channel_width_set = is_cap_supported(caps,
97 IEEE80211_HT_CAP_SUP_WIDTH_20_40);
98 sta_params->lsig_txop_protection = is_cap_supported(caps,
99 IEEE80211_HT_CAP_LSIG_TXOP_PROT);
100
101 sta_params->max_ampdu_size = sta->ht_cap.ampdu_factor;
102 sta_params->max_ampdu_density = sta->ht_cap.ampdu_density;
103 sta_params->max_amsdu_size = is_cap_supported(caps,
104 IEEE80211_HT_CAP_MAX_AMSDU);
105 sta_params->sgi_20Mhz = is_cap_supported(caps,
106 IEEE80211_HT_CAP_SGI_20);
107 sta_params->sgi_40mhz = is_cap_supported(caps,
108 IEEE80211_HT_CAP_SGI_40);
109 sta_params->green_field_capable = is_cap_supported(caps,
110 IEEE80211_HT_CAP_GRN_FLD);
111 sta_params->delayed_ba_support = is_cap_supported(caps,
112 IEEE80211_HT_CAP_DELAY_BA);
113 sta_params->dsss_cck_mode_40mhz = is_cap_supported(caps,
114 IEEE80211_HT_CAP_DSSSCCK40);
115 }
116}
117
118static void wcn36xx_smd_set_sta_params(struct wcn36xx *wcn,
119 struct ieee80211_vif *vif,
120 struct ieee80211_sta *sta,
121 struct wcn36xx_hal_config_sta_params *sta_params)
122{
123 struct wcn36xx_vif *priv_vif = (struct wcn36xx_vif *)vif->drv_priv;
124 struct wcn36xx_sta *priv_sta = NULL;
125 if (vif->type == NL80211_IFTYPE_ADHOC ||
126 vif->type == NL80211_IFTYPE_AP ||
127 vif->type == NL80211_IFTYPE_MESH_POINT) {
128 sta_params->type = 1;
129 sta_params->sta_index = 0xFF;
130 } else {
131 sta_params->type = 0;
132 sta_params->sta_index = 1;
133 }
134
135 sta_params->listen_interval = WCN36XX_LISTEN_INTERVAL(wcn);
136
137 /*
138 * In STA mode ieee80211_sta contains bssid and ieee80211_vif
139 * contains our mac address. In AP mode we are bssid so vif
140 * contains bssid and ieee80211_sta contains mac.
141 */
142 if (NL80211_IFTYPE_STATION == vif->type)
143 memcpy(&sta_params->mac, vif->addr, ETH_ALEN);
144 else
145 memcpy(&sta_params->bssid, vif->addr, ETH_ALEN);
146
147 sta_params->encrypt_type = priv_vif->encrypt_type;
148 sta_params->short_preamble_supported =
149 !(WCN36XX_FLAGS(wcn) &
150 IEEE80211_HW_2GHZ_SHORT_PREAMBLE_INCAPABLE);
151
152 sta_params->rifs_mode = 0;
153 sta_params->rmf = 0;
154 sta_params->action = 0;
155 sta_params->uapsd = 0;
156 sta_params->mimo_ps = WCN36XX_HAL_HT_MIMO_PS_STATIC;
157 sta_params->max_ampdu_duration = 0;
158 sta_params->bssid_index = priv_vif->bss_index;
159 sta_params->p2p = 0;
160
161 if (sta) {
162 priv_sta = (struct wcn36xx_sta *)sta->drv_priv;
163 if (NL80211_IFTYPE_STATION == vif->type)
164 memcpy(&sta_params->bssid, sta->addr, ETH_ALEN);
165 else
166 memcpy(&sta_params->mac, sta->addr, ETH_ALEN);
167 sta_params->wmm_enabled = sta->wme;
168 sta_params->max_sp_len = sta->max_sp;
169 sta_params->aid = priv_sta->aid;
170 wcn36xx_smd_set_sta_ht_params(sta, sta_params);
171 memcpy(&sta_params->supported_rates, &priv_sta->supported_rates,
172 sizeof(priv_sta->supported_rates));
173 } else {
174 wcn36xx_set_default_rates(&sta_params->supported_rates);
175 }
176}
177
178static int wcn36xx_smd_send_and_wait(struct wcn36xx *wcn, size_t len)
179{
180 int ret = 0;
181 wcn36xx_dbg_dump(WCN36XX_DBG_SMD_DUMP, "HAL >>> ", wcn->hal_buf, len);
182
183 init_completion(&wcn->hal_rsp_compl);
184 ret = wcn->ctrl_ops->tx(wcn->hal_buf, len);
185 if (ret) {
186 wcn36xx_err("HAL TX failed\n");
187 goto out;
188 }
189 if (wait_for_completion_timeout(&wcn->hal_rsp_compl,
190 msecs_to_jiffies(HAL_MSG_TIMEOUT)) <= 0) {
191 wcn36xx_err("Timeout while waiting SMD response\n");
192 ret = -ETIME;
193 goto out;
194 }
195out:
196 return ret;
197}
198
199#define INIT_HAL_MSG(msg_body, type) \
200 do { \
201 memset(&msg_body, 0, sizeof(msg_body)); \
202 msg_body.header.msg_type = type; \
203 msg_body.header.msg_version = WCN36XX_HAL_MSG_VERSION0; \
204 msg_body.header.len = sizeof(msg_body); \
205 } while (0) \
206
207#define PREPARE_HAL_BUF(send_buf, msg_body) \
208 do { \
209 memset(send_buf, 0, msg_body.header.len); \
210 memcpy(send_buf, &msg_body, sizeof(msg_body)); \
211 } while (0) \
212
213static int wcn36xx_smd_rsp_status_check(void *buf, size_t len)
214{
215 struct wcn36xx_fw_msg_status_rsp *rsp;
216
217 if (len < sizeof(struct wcn36xx_hal_msg_header) +
218 sizeof(struct wcn36xx_fw_msg_status_rsp))
219 return -EIO;
220
221 rsp = (struct wcn36xx_fw_msg_status_rsp *)
222 (buf + sizeof(struct wcn36xx_hal_msg_header));
223
224 if (WCN36XX_FW_MSG_RESULT_SUCCESS != rsp->status)
225 return rsp->status;
226
227 return 0;
228}
229
230int wcn36xx_smd_load_nv(struct wcn36xx *wcn)
231{
232 const struct firmware *nv;
233 struct nv_data *nv_d;
234 struct wcn36xx_hal_nv_img_download_req_msg msg_body;
235 int fw_bytes_left;
236 int ret;
237 u16 fm_offset = 0;
238
239 ret = request_firmware(&nv, WLAN_NV_FILE, wcn->dev);
240 if (ret) {
241 wcn36xx_err("Failed to load nv file %s: %d\n",
242 WLAN_NV_FILE, ret);
243 goto out_free_nv;
244 }
245
246 nv_d = (struct nv_data *)nv->data;
247 INIT_HAL_MSG(msg_body, WCN36XX_HAL_DOWNLOAD_NV_REQ);
248
249 msg_body.header.len += WCN36XX_NV_FRAGMENT_SIZE;
250
251 msg_body.frag_number = 0;
252 /* hal_buf must be protected with mutex */
253 mutex_lock(&wcn->hal_mutex);
254
255 do {
256 fw_bytes_left = nv->size - fm_offset - 4;
257 if (fw_bytes_left > WCN36XX_NV_FRAGMENT_SIZE) {
258 msg_body.last_fragment = 0;
259 msg_body.nv_img_buffer_size = WCN36XX_NV_FRAGMENT_SIZE;
260 } else {
261 msg_body.last_fragment = 1;
262 msg_body.nv_img_buffer_size = fw_bytes_left;
263
264 /* Do not forget update general message len */
265 msg_body.header.len = sizeof(msg_body) + fw_bytes_left;
266
267 }
268
269 /* Add load NV request message header */
270 memcpy(wcn->hal_buf, &msg_body, sizeof(msg_body));
271
272 /* Add NV body itself */
273 memcpy(wcn->hal_buf + sizeof(msg_body),
274 &nv_d->table + fm_offset,
275 msg_body.nv_img_buffer_size);
276
277 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
278 if (ret)
279 goto out_unlock;
280 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf,
281 wcn->hal_rsp_len);
282 if (ret) {
283 wcn36xx_err("hal_load_nv response failed err=%d\n",
284 ret);
285 goto out_unlock;
286 }
287 msg_body.frag_number++;
288 fm_offset += WCN36XX_NV_FRAGMENT_SIZE;
289
290 } while (msg_body.last_fragment != 1);
291
292out_unlock:
293 mutex_unlock(&wcn->hal_mutex);
294out_free_nv:
295 release_firmware(nv);
296
297 return ret;
298}
299
300static int wcn36xx_smd_start_rsp(struct wcn36xx *wcn, void *buf, size_t len)
301{
302 struct wcn36xx_hal_mac_start_rsp_msg *rsp;
303
304 if (len < sizeof(*rsp))
305 return -EIO;
306
307 rsp = (struct wcn36xx_hal_mac_start_rsp_msg *)buf;
308
309 if (WCN36XX_FW_MSG_RESULT_SUCCESS != rsp->start_rsp_params.status)
310 return -EIO;
311
312 memcpy(wcn->crm_version, rsp->start_rsp_params.crm_version,
313 WCN36XX_HAL_VERSION_LENGTH);
314 memcpy(wcn->wlan_version, rsp->start_rsp_params.wlan_version,
315 WCN36XX_HAL_VERSION_LENGTH);
316
317 /* null terminate the strings, just in case */
318 wcn->crm_version[WCN36XX_HAL_VERSION_LENGTH] = '\0';
319 wcn->wlan_version[WCN36XX_HAL_VERSION_LENGTH] = '\0';
320
321 wcn->fw_revision = rsp->start_rsp_params.version.revision;
322 wcn->fw_version = rsp->start_rsp_params.version.version;
323 wcn->fw_minor = rsp->start_rsp_params.version.minor;
324 wcn->fw_major = rsp->start_rsp_params.version.major;
325
326 wcn36xx_info("firmware WLAN version '%s' and CRM version '%s'\n",
327 wcn->wlan_version, wcn->crm_version);
328
329 wcn36xx_info("firmware API %u.%u.%u.%u, %u stations, %u bssids\n",
330 wcn->fw_major, wcn->fw_minor,
331 wcn->fw_version, wcn->fw_revision,
332 rsp->start_rsp_params.stations,
333 rsp->start_rsp_params.bssids);
334
335 return 0;
336}
337
338int wcn36xx_smd_start(struct wcn36xx *wcn)
339{
340 struct wcn36xx_hal_mac_start_req_msg msg_body;
341 int ret = 0;
342
343 mutex_lock(&wcn->hal_mutex);
344 INIT_HAL_MSG(msg_body, WCN36XX_HAL_START_REQ);
345
346 msg_body.params.type = DRIVER_TYPE_PRODUCTION;
347 msg_body.params.len = 0;
348
349 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
350
351 wcn36xx_dbg(WCN36XX_DBG_HAL, "hal start type %d\n",
352 msg_body.params.type);
353
354 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
355 if (ret) {
356 wcn36xx_err("Sending hal_start failed\n");
357 goto out;
358 }
359
360 ret = wcn36xx_smd_start_rsp(wcn, wcn->hal_buf, wcn->hal_rsp_len);
361 if (ret) {
362 wcn36xx_err("hal_start response failed err=%d\n", ret);
363 goto out;
364 }
365
366out:
367 mutex_unlock(&wcn->hal_mutex);
368 return ret;
369}
370
371int wcn36xx_smd_stop(struct wcn36xx *wcn)
372{
373 struct wcn36xx_hal_mac_stop_req_msg msg_body;
374 int ret = 0;
375
376 mutex_lock(&wcn->hal_mutex);
377 INIT_HAL_MSG(msg_body, WCN36XX_HAL_STOP_REQ);
378
379 msg_body.stop_req_params.reason = HAL_STOP_TYPE_RF_KILL;
380
381 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
382
383 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
384 if (ret) {
385 wcn36xx_err("Sending hal_stop failed\n");
386 goto out;
387 }
388 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
389 if (ret) {
390 wcn36xx_err("hal_stop response failed err=%d\n", ret);
391 goto out;
392 }
393out:
394 mutex_unlock(&wcn->hal_mutex);
395 return ret;
396}
397
398int wcn36xx_smd_init_scan(struct wcn36xx *wcn, enum wcn36xx_hal_sys_mode mode)
399{
400 struct wcn36xx_hal_init_scan_req_msg msg_body;
401 int ret = 0;
402
403 mutex_lock(&wcn->hal_mutex);
404 INIT_HAL_MSG(msg_body, WCN36XX_HAL_INIT_SCAN_REQ);
405
406 msg_body.mode = mode;
407
408 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
409
410 wcn36xx_dbg(WCN36XX_DBG_HAL, "hal init scan mode %d\n", msg_body.mode);
411
412 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
413 if (ret) {
414 wcn36xx_err("Sending hal_init_scan failed\n");
415 goto out;
416 }
417 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
418 if (ret) {
419 wcn36xx_err("hal_init_scan response failed err=%d\n", ret);
420 goto out;
421 }
422out:
423 mutex_unlock(&wcn->hal_mutex);
424 return ret;
425}
426
427int wcn36xx_smd_start_scan(struct wcn36xx *wcn)
428{
429 struct wcn36xx_hal_start_scan_req_msg msg_body;
430 int ret = 0;
431
432 mutex_lock(&wcn->hal_mutex);
433 INIT_HAL_MSG(msg_body, WCN36XX_HAL_START_SCAN_REQ);
434
435 msg_body.scan_channel = WCN36XX_HW_CHANNEL(wcn);
436
437 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
438
439 wcn36xx_dbg(WCN36XX_DBG_HAL, "hal start scan channel %d\n",
440 msg_body.scan_channel);
441
442 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
443 if (ret) {
444 wcn36xx_err("Sending hal_start_scan failed\n");
445 goto out;
446 }
447 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
448 if (ret) {
449 wcn36xx_err("hal_start_scan response failed err=%d\n", ret);
450 goto out;
451 }
452out:
453 mutex_unlock(&wcn->hal_mutex);
454 return ret;
455}
456
457int wcn36xx_smd_end_scan(struct wcn36xx *wcn)
458{
459 struct wcn36xx_hal_end_scan_req_msg msg_body;
460 int ret = 0;
461
462 mutex_lock(&wcn->hal_mutex);
463 INIT_HAL_MSG(msg_body, WCN36XX_HAL_END_SCAN_REQ);
464
465 msg_body.scan_channel = WCN36XX_HW_CHANNEL(wcn);
466
467 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
468
469 wcn36xx_dbg(WCN36XX_DBG_HAL, "hal end scan channel %d\n",
470 msg_body.scan_channel);
471
472 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
473 if (ret) {
474 wcn36xx_err("Sending hal_end_scan failed\n");
475 goto out;
476 }
477 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
478 if (ret) {
479 wcn36xx_err("hal_end_scan response failed err=%d\n", ret);
480 goto out;
481 }
482out:
483 mutex_unlock(&wcn->hal_mutex);
484 return ret;
485}
486
487int wcn36xx_smd_finish_scan(struct wcn36xx *wcn,
488 enum wcn36xx_hal_sys_mode mode)
489{
490 struct wcn36xx_hal_finish_scan_req_msg msg_body;
491 int ret = 0;
492
493 mutex_lock(&wcn->hal_mutex);
494 INIT_HAL_MSG(msg_body, WCN36XX_HAL_FINISH_SCAN_REQ);
495
496 msg_body.mode = mode;
497
498 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
499
500 wcn36xx_dbg(WCN36XX_DBG_HAL, "hal finish scan mode %d\n",
501 msg_body.mode);
502
503 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
504 if (ret) {
505 wcn36xx_err("Sending hal_finish_scan failed\n");
506 goto out;
507 }
508 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
509 if (ret) {
510 wcn36xx_err("hal_finish_scan response failed err=%d\n", ret);
511 goto out;
512 }
513out:
514 mutex_unlock(&wcn->hal_mutex);
515 return ret;
516}
517
518static int wcn36xx_smd_switch_channel_rsp(void *buf, size_t len)
519{
520 struct wcn36xx_hal_switch_channel_rsp_msg *rsp;
521 int ret = 0;
522
523 ret = wcn36xx_smd_rsp_status_check(buf, len);
524 if (ret)
525 return ret;
526 rsp = (struct wcn36xx_hal_switch_channel_rsp_msg *)buf;
527 wcn36xx_dbg(WCN36XX_DBG_HAL, "channel switched to: %d, status: %d\n",
528 rsp->channel_number, rsp->status);
529 return ret;
530}
531
532int wcn36xx_smd_switch_channel(struct wcn36xx *wcn,
533 struct ieee80211_vif *vif, int ch)
534{
535 struct wcn36xx_hal_switch_channel_req_msg msg_body;
536 int ret = 0;
537
538 mutex_lock(&wcn->hal_mutex);
539 INIT_HAL_MSG(msg_body, WCN36XX_HAL_CH_SWITCH_REQ);
540
541 msg_body.channel_number = (u8)ch;
542 msg_body.tx_mgmt_power = 0xbf;
543 msg_body.max_tx_power = 0xbf;
544 memcpy(msg_body.self_sta_mac_addr, vif->addr, ETH_ALEN);
545
546 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
547
548 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
549 if (ret) {
550 wcn36xx_err("Sending hal_switch_channel failed\n");
551 goto out;
552 }
553 ret = wcn36xx_smd_switch_channel_rsp(wcn->hal_buf, wcn->hal_rsp_len);
554 if (ret) {
555 wcn36xx_err("hal_switch_channel response failed err=%d\n", ret);
556 goto out;
557 }
558out:
559 mutex_unlock(&wcn->hal_mutex);
560 return ret;
561}
562
563static int wcn36xx_smd_update_scan_params_rsp(void *buf, size_t len)
564{
565 struct wcn36xx_hal_update_scan_params_resp *rsp;
566
567 rsp = (struct wcn36xx_hal_update_scan_params_resp *)buf;
568
569 /* Remove the PNO version bit */
570 rsp->status &= (~(WCN36XX_FW_MSG_PNO_VERSION_MASK));
571
572 if (WCN36XX_FW_MSG_RESULT_SUCCESS != rsp->status) {
573 wcn36xx_warn("error response from update scan\n");
574 return rsp->status;
575 }
576
577 return 0;
578}
579
580int wcn36xx_smd_update_scan_params(struct wcn36xx *wcn)
581{
582 struct wcn36xx_hal_update_scan_params_req msg_body;
583 int ret = 0;
584
585 mutex_lock(&wcn->hal_mutex);
586 INIT_HAL_MSG(msg_body, WCN36XX_HAL_UPDATE_SCAN_PARAM_REQ);
587
588 msg_body.dot11d_enabled = 0;
589 msg_body.dot11d_resolved = 0;
590 msg_body.channel_count = 26;
591 msg_body.active_min_ch_time = 60;
592 msg_body.active_max_ch_time = 120;
593 msg_body.passive_min_ch_time = 60;
594 msg_body.passive_max_ch_time = 110;
595 msg_body.state = 0;
596
597 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
598
599 wcn36xx_dbg(WCN36XX_DBG_HAL,
600 "hal update scan params channel_count %d\n",
601 msg_body.channel_count);
602
603 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
604 if (ret) {
605 wcn36xx_err("Sending hal_update_scan_params failed\n");
606 goto out;
607 }
608 ret = wcn36xx_smd_update_scan_params_rsp(wcn->hal_buf,
609 wcn->hal_rsp_len);
610 if (ret) {
611 wcn36xx_err("hal_update_scan_params response failed err=%d\n",
612 ret);
613 goto out;
614 }
615out:
616 mutex_unlock(&wcn->hal_mutex);
617 return ret;
618}
619
620static int wcn36xx_smd_add_sta_self_rsp(struct wcn36xx *wcn,
621 struct ieee80211_vif *vif,
622 void *buf,
623 size_t len)
624{
625 struct wcn36xx_hal_add_sta_self_rsp_msg *rsp;
626 struct wcn36xx_vif *priv_vif = (struct wcn36xx_vif *)vif->drv_priv;
627
628 if (len < sizeof(*rsp))
629 return -EINVAL;
630
631 rsp = (struct wcn36xx_hal_add_sta_self_rsp_msg *)buf;
632
633 if (rsp->status != WCN36XX_FW_MSG_RESULT_SUCCESS) {
634 wcn36xx_warn("hal add sta self failure: %d\n",
635 rsp->status);
636 return rsp->status;
637 }
638
639 wcn36xx_dbg(WCN36XX_DBG_HAL,
640 "hal add sta self status %d self_sta_index %d dpu_index %d\n",
641 rsp->status, rsp->self_sta_index, rsp->dpu_index);
642
643 priv_vif->self_sta_index = rsp->self_sta_index;
644 priv_vif->self_dpu_desc_index = rsp->dpu_index;
645
646 return 0;
647}
648
649int wcn36xx_smd_add_sta_self(struct wcn36xx *wcn, struct ieee80211_vif *vif)
650{
651 struct wcn36xx_hal_add_sta_self_req msg_body;
652 int ret = 0;
653
654 mutex_lock(&wcn->hal_mutex);
655 INIT_HAL_MSG(msg_body, WCN36XX_HAL_ADD_STA_SELF_REQ);
656
657 memcpy(&msg_body.self_addr, vif->addr, ETH_ALEN);
658
659 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
660
661 wcn36xx_dbg(WCN36XX_DBG_HAL,
662 "hal add sta self self_addr %pM status %d\n",
663 msg_body.self_addr, msg_body.status);
664
665 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
666 if (ret) {
667 wcn36xx_err("Sending hal_add_sta_self failed\n");
668 goto out;
669 }
670 ret = wcn36xx_smd_add_sta_self_rsp(wcn,
671 vif,
672 wcn->hal_buf,
673 wcn->hal_rsp_len);
674 if (ret) {
675 wcn36xx_err("hal_add_sta_self response failed err=%d\n", ret);
676 goto out;
677 }
678out:
679 mutex_unlock(&wcn->hal_mutex);
680 return ret;
681}
682
683int wcn36xx_smd_delete_sta_self(struct wcn36xx *wcn, u8 *addr)
684{
685 struct wcn36xx_hal_del_sta_self_req_msg msg_body;
686 int ret = 0;
687
688 mutex_lock(&wcn->hal_mutex);
689 INIT_HAL_MSG(msg_body, WCN36XX_HAL_DEL_STA_SELF_REQ);
690
691 memcpy(&msg_body.self_addr, addr, ETH_ALEN);
692
693 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
694
695 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
696 if (ret) {
697 wcn36xx_err("Sending hal_delete_sta_self failed\n");
698 goto out;
699 }
700 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
701 if (ret) {
702 wcn36xx_err("hal_delete_sta_self response failed err=%d\n",
703 ret);
704 goto out;
705 }
706out:
707 mutex_unlock(&wcn->hal_mutex);
708 return ret;
709}
710
711int wcn36xx_smd_delete_sta(struct wcn36xx *wcn, u8 sta_index)
712{
713 struct wcn36xx_hal_delete_sta_req_msg msg_body;
714 int ret = 0;
715
716 mutex_lock(&wcn->hal_mutex);
717 INIT_HAL_MSG(msg_body, WCN36XX_HAL_DELETE_STA_REQ);
718
719 msg_body.sta_index = sta_index;
720
721 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
722
723 wcn36xx_dbg(WCN36XX_DBG_HAL,
724 "hal delete sta sta_index %d\n",
725 msg_body.sta_index);
726
727 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
728 if (ret) {
729 wcn36xx_err("Sending hal_delete_sta failed\n");
730 goto out;
731 }
732 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
733 if (ret) {
734 wcn36xx_err("hal_delete_sta response failed err=%d\n", ret);
735 goto out;
736 }
737out:
738 mutex_unlock(&wcn->hal_mutex);
739 return ret;
740}
741
742static int wcn36xx_smd_join_rsp(void *buf, size_t len)
743{
744 struct wcn36xx_hal_join_rsp_msg *rsp;
745
746 if (wcn36xx_smd_rsp_status_check(buf, len))
747 return -EIO;
748
749 rsp = (struct wcn36xx_hal_join_rsp_msg *)buf;
750
751 wcn36xx_dbg(WCN36XX_DBG_HAL,
752 "hal rsp join status %d tx_mgmt_power %d\n",
753 rsp->status, rsp->tx_mgmt_power);
754
755 return 0;
756}
757
758int wcn36xx_smd_join(struct wcn36xx *wcn, const u8 *bssid, u8 *vif, u8 ch)
759{
760 struct wcn36xx_hal_join_req_msg msg_body;
761 int ret = 0;
762
763 mutex_lock(&wcn->hal_mutex);
764 INIT_HAL_MSG(msg_body, WCN36XX_HAL_JOIN_REQ);
765
766 memcpy(&msg_body.bssid, bssid, ETH_ALEN);
767 memcpy(&msg_body.self_sta_mac_addr, vif, ETH_ALEN);
768 msg_body.channel = ch;
769
770 if (conf_is_ht40_minus(&wcn->hw->conf))
771 msg_body.secondary_channel_offset =
772 PHY_DOUBLE_CHANNEL_HIGH_PRIMARY;
773 else if (conf_is_ht40_plus(&wcn->hw->conf))
774 msg_body.secondary_channel_offset =
775 PHY_DOUBLE_CHANNEL_LOW_PRIMARY;
776 else
777 msg_body.secondary_channel_offset =
778 PHY_SINGLE_CHANNEL_CENTERED;
779
780 msg_body.link_state = WCN36XX_HAL_LINK_PREASSOC_STATE;
781
782 msg_body.max_tx_power = 0xbf;
783 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
784
785 wcn36xx_dbg(WCN36XX_DBG_HAL,
786 "hal join req bssid %pM self_sta_mac_addr %pM channel %d link_state %d\n",
787 msg_body.bssid, msg_body.self_sta_mac_addr,
788 msg_body.channel, msg_body.link_state);
789
790 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
791 if (ret) {
792 wcn36xx_err("Sending hal_join failed\n");
793 goto out;
794 }
795 ret = wcn36xx_smd_join_rsp(wcn->hal_buf, wcn->hal_rsp_len);
796 if (ret) {
797 wcn36xx_err("hal_join response failed err=%d\n", ret);
798 goto out;
799 }
800out:
801 mutex_unlock(&wcn->hal_mutex);
802 return ret;
803}
804
805int wcn36xx_smd_set_link_st(struct wcn36xx *wcn, const u8 *bssid,
806 const u8 *sta_mac,
807 enum wcn36xx_hal_link_state state)
808{
809 struct wcn36xx_hal_set_link_state_req_msg msg_body;
810 int ret = 0;
811
812 mutex_lock(&wcn->hal_mutex);
813 INIT_HAL_MSG(msg_body, WCN36XX_HAL_SET_LINK_ST_REQ);
814
815 memcpy(&msg_body.bssid, bssid, ETH_ALEN);
816 memcpy(&msg_body.self_mac_addr, sta_mac, ETH_ALEN);
817 msg_body.state = state;
818
819 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
820
821 wcn36xx_dbg(WCN36XX_DBG_HAL,
822 "hal set link state bssid %pM self_mac_addr %pM state %d\n",
823 msg_body.bssid, msg_body.self_mac_addr, msg_body.state);
824
825 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
826 if (ret) {
827 wcn36xx_err("Sending hal_set_link_st failed\n");
828 goto out;
829 }
830 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
831 if (ret) {
832 wcn36xx_err("hal_set_link_st response failed err=%d\n", ret);
833 goto out;
834 }
835out:
836 mutex_unlock(&wcn->hal_mutex);
837 return ret;
838}
839
840static void wcn36xx_smd_convert_sta_to_v1(struct wcn36xx *wcn,
841 const struct wcn36xx_hal_config_sta_params *orig,
842 struct wcn36xx_hal_config_sta_params_v1 *v1)
843{
844 /* convert orig to v1 format */
845 memcpy(&v1->bssid, orig->bssid, ETH_ALEN);
846 memcpy(&v1->mac, orig->mac, ETH_ALEN);
847 v1->aid = orig->aid;
848 v1->type = orig->type;
849 v1->listen_interval = orig->listen_interval;
850 v1->ht_capable = orig->ht_capable;
851
852 v1->max_ampdu_size = orig->max_ampdu_size;
853 v1->max_ampdu_density = orig->max_ampdu_density;
854 v1->sgi_40mhz = orig->sgi_40mhz;
855 v1->sgi_20Mhz = orig->sgi_20Mhz;
856
857 memcpy(&v1->supported_rates, &orig->supported_rates,
858 sizeof(orig->supported_rates));
859 v1->sta_index = orig->sta_index;
860}
861
862static int wcn36xx_smd_config_sta_rsp(struct wcn36xx *wcn,
863 struct ieee80211_sta *sta,
864 void *buf,
865 size_t len)
866{
867 struct wcn36xx_hal_config_sta_rsp_msg *rsp;
868 struct config_sta_rsp_params *params;
869 struct wcn36xx_sta *sta_priv = (struct wcn36xx_sta *)sta->drv_priv;
870
871 if (len < sizeof(*rsp))
872 return -EINVAL;
873
874 rsp = (struct wcn36xx_hal_config_sta_rsp_msg *)buf;
875 params = &rsp->params;
876
877 if (params->status != WCN36XX_FW_MSG_RESULT_SUCCESS) {
878 wcn36xx_warn("hal config sta response failure: %d\n",
879 params->status);
880 return -EIO;
881 }
882
883 sta_priv->sta_index = params->sta_index;
884 sta_priv->dpu_desc_index = params->dpu_index;
885
886 wcn36xx_dbg(WCN36XX_DBG_HAL,
887 "hal config sta rsp status %d sta_index %d bssid_index %d p2p %d\n",
888 params->status, params->sta_index, params->bssid_index,
889 params->p2p);
890
891 return 0;
892}
893
894static int wcn36xx_smd_config_sta_v1(struct wcn36xx *wcn,
895 const struct wcn36xx_hal_config_sta_req_msg *orig)
896{
897 struct wcn36xx_hal_config_sta_req_msg_v1 msg_body;
898 struct wcn36xx_hal_config_sta_params_v1 *sta = &msg_body.sta_params;
899
900 INIT_HAL_MSG(msg_body, WCN36XX_HAL_CONFIG_STA_REQ);
901
902 wcn36xx_smd_convert_sta_to_v1(wcn, &orig->sta_params,
903 &msg_body.sta_params);
904
905 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
906
907 wcn36xx_dbg(WCN36XX_DBG_HAL,
908 "hal config sta v1 action %d sta_index %d bssid_index %d bssid %pM type %d mac %pM aid %d\n",
909 sta->action, sta->sta_index, sta->bssid_index,
910 sta->bssid, sta->type, sta->mac, sta->aid);
911
912 return wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
913}
914
915int wcn36xx_smd_config_sta(struct wcn36xx *wcn, struct ieee80211_vif *vif,
916 struct ieee80211_sta *sta)
917{
918 struct wcn36xx_hal_config_sta_req_msg msg;
919 struct wcn36xx_hal_config_sta_params *sta_params;
920 int ret = 0;
921
922 mutex_lock(&wcn->hal_mutex);
923 INIT_HAL_MSG(msg, WCN36XX_HAL_CONFIG_STA_REQ);
924
925 sta_params = &msg.sta_params;
926
927 wcn36xx_smd_set_sta_params(wcn, vif, sta, sta_params);
928
929 if (!wcn36xx_is_fw_version(wcn, 1, 2, 2, 24)) {
930 ret = wcn36xx_smd_config_sta_v1(wcn, &msg);
931 } else {
932 PREPARE_HAL_BUF(wcn->hal_buf, msg);
933
934 wcn36xx_dbg(WCN36XX_DBG_HAL,
935 "hal config sta action %d sta_index %d bssid_index %d bssid %pM type %d mac %pM aid %d\n",
936 sta_params->action, sta_params->sta_index,
937 sta_params->bssid_index, sta_params->bssid,
938 sta_params->type, sta_params->mac, sta_params->aid);
939
940 ret = wcn36xx_smd_send_and_wait(wcn, msg.header.len);
941 }
942 if (ret) {
943 wcn36xx_err("Sending hal_config_sta failed\n");
944 goto out;
945 }
946 ret = wcn36xx_smd_config_sta_rsp(wcn,
947 sta,
948 wcn->hal_buf,
949 wcn->hal_rsp_len);
950 if (ret) {
951 wcn36xx_err("hal_config_sta response failed err=%d\n", ret);
952 goto out;
953 }
954out:
955 mutex_unlock(&wcn->hal_mutex);
956 return ret;
957}
958
959static int wcn36xx_smd_config_bss_v1(struct wcn36xx *wcn,
960 const struct wcn36xx_hal_config_bss_req_msg *orig)
961{
962 struct wcn36xx_hal_config_bss_req_msg_v1 msg_body;
963 struct wcn36xx_hal_config_bss_params_v1 *bss = &msg_body.bss_params;
964 struct wcn36xx_hal_config_sta_params_v1 *sta = &bss->sta;
965
966 INIT_HAL_MSG(msg_body, WCN36XX_HAL_CONFIG_BSS_REQ);
967
968 /* convert orig to v1 */
969 memcpy(&msg_body.bss_params.bssid,
970 &orig->bss_params.bssid, ETH_ALEN);
971 memcpy(&msg_body.bss_params.self_mac_addr,
972 &orig->bss_params.self_mac_addr, ETH_ALEN);
973
974 msg_body.bss_params.bss_type = orig->bss_params.bss_type;
975 msg_body.bss_params.oper_mode = orig->bss_params.oper_mode;
976 msg_body.bss_params.nw_type = orig->bss_params.nw_type;
977
978 msg_body.bss_params.short_slot_time_supported =
979 orig->bss_params.short_slot_time_supported;
980 msg_body.bss_params.lla_coexist = orig->bss_params.lla_coexist;
981 msg_body.bss_params.llb_coexist = orig->bss_params.llb_coexist;
982 msg_body.bss_params.llg_coexist = orig->bss_params.llg_coexist;
983 msg_body.bss_params.ht20_coexist = orig->bss_params.ht20_coexist;
984 msg_body.bss_params.lln_non_gf_coexist =
985 orig->bss_params.lln_non_gf_coexist;
986
987 msg_body.bss_params.lsig_tx_op_protection_full_support =
988 orig->bss_params.lsig_tx_op_protection_full_support;
989 msg_body.bss_params.rifs_mode = orig->bss_params.rifs_mode;
990 msg_body.bss_params.beacon_interval = orig->bss_params.beacon_interval;
991 msg_body.bss_params.dtim_period = orig->bss_params.dtim_period;
992 msg_body.bss_params.tx_channel_width_set =
993 orig->bss_params.tx_channel_width_set;
994 msg_body.bss_params.oper_channel = orig->bss_params.oper_channel;
995 msg_body.bss_params.ext_channel = orig->bss_params.ext_channel;
996
997 msg_body.bss_params.reserved = orig->bss_params.reserved;
998
999 memcpy(&msg_body.bss_params.ssid,
1000 &orig->bss_params.ssid,
1001 sizeof(orig->bss_params.ssid));
1002
1003 msg_body.bss_params.action = orig->bss_params.action;
1004 msg_body.bss_params.rateset = orig->bss_params.rateset;
1005 msg_body.bss_params.ht = orig->bss_params.ht;
1006 msg_body.bss_params.obss_prot_enabled =
1007 orig->bss_params.obss_prot_enabled;
1008 msg_body.bss_params.rmf = orig->bss_params.rmf;
1009 msg_body.bss_params.ht_oper_mode = orig->bss_params.ht_oper_mode;
1010 msg_body.bss_params.dual_cts_protection =
1011 orig->bss_params.dual_cts_protection;
1012
1013 msg_body.bss_params.max_probe_resp_retry_limit =
1014 orig->bss_params.max_probe_resp_retry_limit;
1015 msg_body.bss_params.hidden_ssid = orig->bss_params.hidden_ssid;
1016 msg_body.bss_params.proxy_probe_resp =
1017 orig->bss_params.proxy_probe_resp;
1018 msg_body.bss_params.edca_params_valid =
1019 orig->bss_params.edca_params_valid;
1020
1021 memcpy(&msg_body.bss_params.acbe,
1022 &orig->bss_params.acbe,
1023 sizeof(orig->bss_params.acbe));
1024 memcpy(&msg_body.bss_params.acbk,
1025 &orig->bss_params.acbk,
1026 sizeof(orig->bss_params.acbk));
1027 memcpy(&msg_body.bss_params.acvi,
1028 &orig->bss_params.acvi,
1029 sizeof(orig->bss_params.acvi));
1030 memcpy(&msg_body.bss_params.acvo,
1031 &orig->bss_params.acvo,
1032 sizeof(orig->bss_params.acvo));
1033
1034 msg_body.bss_params.ext_set_sta_key_param_valid =
1035 orig->bss_params.ext_set_sta_key_param_valid;
1036
1037 memcpy(&msg_body.bss_params.ext_set_sta_key_param,
1038 &orig->bss_params.ext_set_sta_key_param,
1039 sizeof(orig->bss_params.acvo));
1040
1041 msg_body.bss_params.wcn36xx_hal_persona =
1042 orig->bss_params.wcn36xx_hal_persona;
1043 msg_body.bss_params.spectrum_mgt_enable =
1044 orig->bss_params.spectrum_mgt_enable;
1045 msg_body.bss_params.tx_mgmt_power = orig->bss_params.tx_mgmt_power;
1046 msg_body.bss_params.max_tx_power = orig->bss_params.max_tx_power;
1047
1048 wcn36xx_smd_convert_sta_to_v1(wcn, &orig->bss_params.sta,
1049 &msg_body.bss_params.sta);
1050
1051 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
1052
1053 wcn36xx_dbg(WCN36XX_DBG_HAL,
1054 "hal config bss v1 bssid %pM self_mac_addr %pM bss_type %d oper_mode %d nw_type %d\n",
1055 bss->bssid, bss->self_mac_addr, bss->bss_type,
1056 bss->oper_mode, bss->nw_type);
1057
1058 wcn36xx_dbg(WCN36XX_DBG_HAL,
1059 "- sta bssid %pM action %d sta_index %d bssid_index %d aid %d type %d mac %pM\n",
1060 sta->bssid, sta->action, sta->sta_index,
1061 sta->bssid_index, sta->aid, sta->type, sta->mac);
1062
1063 return wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
1064}
1065
1066
1067static int wcn36xx_smd_config_bss_rsp(struct wcn36xx *wcn,
1068 struct ieee80211_vif *vif,
1069 void *buf,
1070 size_t len)
1071{
1072 struct wcn36xx_hal_config_bss_rsp_msg *rsp;
1073 struct wcn36xx_hal_config_bss_rsp_params *params;
1074 struct wcn36xx_vif *priv_vif = (struct wcn36xx_vif *)vif->drv_priv;
1075
1076 if (len < sizeof(*rsp))
1077 return -EINVAL;
1078
1079 rsp = (struct wcn36xx_hal_config_bss_rsp_msg *)buf;
1080 params = &rsp->bss_rsp_params;
1081
1082 if (params->status != WCN36XX_FW_MSG_RESULT_SUCCESS) {
1083 wcn36xx_warn("hal config bss response failure: %d\n",
1084 params->status);
1085 return -EIO;
1086 }
1087
1088 wcn36xx_dbg(WCN36XX_DBG_HAL,
1089 "hal config bss rsp status %d bss_idx %d dpu_desc_index %d"
1090 " sta_idx %d self_idx %d bcast_idx %d mac %pM"
1091 " power %d ucast_dpu_signature %d\n",
1092 params->status, params->bss_index, params->dpu_desc_index,
1093 params->bss_sta_index, params->bss_self_sta_index,
1094 params->bss_bcast_sta_idx, params->mac,
1095 params->tx_mgmt_power, params->ucast_dpu_signature);
1096
1097 priv_vif->bss_index = params->bss_index;
1098
1099 if (priv_vif->sta) {
1100 priv_vif->sta->bss_sta_index = params->bss_sta_index;
1101 priv_vif->sta->bss_dpu_desc_index = params->dpu_desc_index;
1102 }
1103
1104 priv_vif->ucast_dpu_signature = params->ucast_dpu_signature;
1105
1106 return 0;
1107}
1108
1109int wcn36xx_smd_config_bss(struct wcn36xx *wcn, struct ieee80211_vif *vif,
1110 struct ieee80211_sta *sta, const u8 *bssid,
1111 bool update)
1112{
1113 struct wcn36xx_hal_config_bss_req_msg msg;
1114 struct wcn36xx_hal_config_bss_params *bss;
1115 struct wcn36xx_hal_config_sta_params *sta_params;
1116 struct wcn36xx_vif *vif_priv = (struct wcn36xx_vif *)vif->drv_priv;
1117 int ret = 0;
1118
1119 mutex_lock(&wcn->hal_mutex);
1120 INIT_HAL_MSG(msg, WCN36XX_HAL_CONFIG_BSS_REQ);
1121
1122 bss = &msg.bss_params;
1123 sta_params = &bss->sta;
1124
1125 WARN_ON(is_zero_ether_addr(bssid));
1126
1127 memcpy(&bss->bssid, bssid, ETH_ALEN);
1128
1129 memcpy(bss->self_mac_addr, vif->addr, ETH_ALEN);
1130
1131 if (vif->type == NL80211_IFTYPE_STATION) {
1132 bss->bss_type = WCN36XX_HAL_INFRASTRUCTURE_MODE;
1133
1134 /* STA */
1135 bss->oper_mode = 1;
1136 bss->wcn36xx_hal_persona = WCN36XX_HAL_STA_MODE;
1137 } else if (vif->type == NL80211_IFTYPE_AP) {
1138 bss->bss_type = WCN36XX_HAL_INFRA_AP_MODE;
1139
1140 /* AP */
1141 bss->oper_mode = 0;
1142 bss->wcn36xx_hal_persona = WCN36XX_HAL_STA_SAP_MODE;
1143 } else if (vif->type == NL80211_IFTYPE_ADHOC ||
1144 vif->type == NL80211_IFTYPE_MESH_POINT) {
1145 bss->bss_type = WCN36XX_HAL_IBSS_MODE;
1146
1147 /* STA */
1148 bss->oper_mode = 1;
1149 } else {
1150 wcn36xx_warn("Unknown type for bss config: %d\n", vif->type);
1151 }
1152
1153 if (vif->type == NL80211_IFTYPE_STATION)
1154 wcn36xx_smd_set_bss_nw_type(wcn, sta, bss);
1155 else
1156 bss->nw_type = WCN36XX_HAL_11N_NW_TYPE;
1157
1158 bss->short_slot_time_supported = vif->bss_conf.use_short_slot;
1159 bss->lla_coexist = 0;
1160 bss->llb_coexist = 0;
1161 bss->llg_coexist = 0;
1162 bss->rifs_mode = 0;
1163 bss->beacon_interval = vif->bss_conf.beacon_int;
1164 bss->dtim_period = vif_priv->dtim_period;
1165
1166 wcn36xx_smd_set_bss_ht_params(vif, sta, bss);
1167
1168 bss->oper_channel = WCN36XX_HW_CHANNEL(wcn);
1169
1170 if (conf_is_ht40_minus(&wcn->hw->conf))
1171 bss->ext_channel = IEEE80211_HT_PARAM_CHA_SEC_BELOW;
1172 else if (conf_is_ht40_plus(&wcn->hw->conf))
1173 bss->ext_channel = IEEE80211_HT_PARAM_CHA_SEC_ABOVE;
1174 else
1175 bss->ext_channel = IEEE80211_HT_PARAM_CHA_SEC_NONE;
1176
1177 bss->reserved = 0;
1178 wcn36xx_smd_set_sta_params(wcn, vif, sta, sta_params);
1179
1180 /* wcn->ssid is only valid in AP and IBSS mode */
1181 bss->ssid.length = vif_priv->ssid.length;
1182 memcpy(bss->ssid.ssid, vif_priv->ssid.ssid, vif_priv->ssid.length);
1183
1184 bss->obss_prot_enabled = 0;
1185 bss->rmf = 0;
1186 bss->max_probe_resp_retry_limit = 0;
1187 bss->hidden_ssid = vif->bss_conf.hidden_ssid;
1188 bss->proxy_probe_resp = 0;
1189 bss->edca_params_valid = 0;
1190
1191 /* FIXME: set acbe, acbk, acvi and acvo */
1192
1193 bss->ext_set_sta_key_param_valid = 0;
1194
1195 /* FIXME: set ext_set_sta_key_param */
1196
1197 bss->spectrum_mgt_enable = 0;
1198 bss->tx_mgmt_power = 0;
1199 bss->max_tx_power = WCN36XX_MAX_POWER(wcn);
1200
1201 bss->action = update;
1202
1203 wcn36xx_dbg(WCN36XX_DBG_HAL,
1204 "hal config bss bssid %pM self_mac_addr %pM bss_type %d oper_mode %d nw_type %d\n",
1205 bss->bssid, bss->self_mac_addr, bss->bss_type,
1206 bss->oper_mode, bss->nw_type);
1207
1208 wcn36xx_dbg(WCN36XX_DBG_HAL,
1209 "- sta bssid %pM action %d sta_index %d bssid_index %d aid %d type %d mac %pM\n",
1210 sta_params->bssid, sta_params->action,
1211 sta_params->sta_index, sta_params->bssid_index,
1212 sta_params->aid, sta_params->type,
1213 sta_params->mac);
1214
1215 if (!wcn36xx_is_fw_version(wcn, 1, 2, 2, 24)) {
1216 ret = wcn36xx_smd_config_bss_v1(wcn, &msg);
1217 } else {
1218 PREPARE_HAL_BUF(wcn->hal_buf, msg);
1219
1220 ret = wcn36xx_smd_send_and_wait(wcn, msg.header.len);
1221 }
1222 if (ret) {
1223 wcn36xx_err("Sending hal_config_bss failed\n");
1224 goto out;
1225 }
1226 ret = wcn36xx_smd_config_bss_rsp(wcn,
1227 vif,
1228 wcn->hal_buf,
1229 wcn->hal_rsp_len);
1230 if (ret) {
1231 wcn36xx_err("hal_config_bss response failed err=%d\n", ret);
1232 goto out;
1233 }
1234out:
1235 mutex_unlock(&wcn->hal_mutex);
1236 return ret;
1237}
1238
1239int wcn36xx_smd_delete_bss(struct wcn36xx *wcn, struct ieee80211_vif *vif)
1240{
1241 struct wcn36xx_hal_delete_bss_req_msg msg_body;
1242 struct wcn36xx_vif *priv_vif = (struct wcn36xx_vif *)vif->drv_priv;
1243 int ret = 0;
1244
1245 mutex_lock(&wcn->hal_mutex);
1246 INIT_HAL_MSG(msg_body, WCN36XX_HAL_DELETE_BSS_REQ);
1247
1248 msg_body.bss_index = priv_vif->bss_index;
1249
1250 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
1251
1252 wcn36xx_dbg(WCN36XX_DBG_HAL, "hal delete bss %d\n", msg_body.bss_index);
1253
1254 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
1255 if (ret) {
1256 wcn36xx_err("Sending hal_delete_bss failed\n");
1257 goto out;
1258 }
1259 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
1260 if (ret) {
1261 wcn36xx_err("hal_delete_bss response failed err=%d\n", ret);
1262 goto out;
1263 }
1264out:
1265 mutex_unlock(&wcn->hal_mutex);
1266 return ret;
1267}
1268
1269int wcn36xx_smd_send_beacon(struct wcn36xx *wcn, struct ieee80211_vif *vif,
1270 struct sk_buff *skb_beacon, u16 tim_off,
1271 u16 p2p_off)
1272{
1273 struct wcn36xx_hal_send_beacon_req_msg msg_body;
1274 int ret = 0;
1275
1276 mutex_lock(&wcn->hal_mutex);
1277 INIT_HAL_MSG(msg_body, WCN36XX_HAL_SEND_BEACON_REQ);
1278
1279 /* TODO need to find out why this is needed? */
1280 msg_body.beacon_length = skb_beacon->len + 6;
1281
1282 if (BEACON_TEMPLATE_SIZE > msg_body.beacon_length) {
1283 memcpy(&msg_body.beacon, &skb_beacon->len, sizeof(u32));
1284 memcpy(&(msg_body.beacon[4]), skb_beacon->data,
1285 skb_beacon->len);
1286 } else {
1287 wcn36xx_err("Beacon is to big: beacon size=%d\n",
1288 msg_body.beacon_length);
1289 return -ENOMEM;
1290 }
1291 memcpy(msg_body.bssid, vif->addr, ETH_ALEN);
1292
1293 /* TODO need to find out why this is needed? */
1294 msg_body.tim_ie_offset = tim_off+4;
1295 msg_body.p2p_ie_offset = p2p_off;
1296 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
1297
1298 wcn36xx_dbg(WCN36XX_DBG_HAL,
1299 "hal send beacon beacon_length %d\n",
1300 msg_body.beacon_length);
1301
1302 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
1303 if (ret) {
1304 wcn36xx_err("Sending hal_send_beacon failed\n");
1305 goto out;
1306 }
1307 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
1308 if (ret) {
1309 wcn36xx_err("hal_send_beacon response failed err=%d\n", ret);
1310 goto out;
1311 }
1312out:
1313 mutex_unlock(&wcn->hal_mutex);
1314 return ret;
1315}
1316
1317int wcn36xx_smd_update_proberesp_tmpl(struct wcn36xx *wcn,
1318 struct ieee80211_vif *vif,
1319 struct sk_buff *skb)
1320{
1321 struct wcn36xx_hal_send_probe_resp_req_msg msg;
1322 int ret = 0;
1323
1324 mutex_lock(&wcn->hal_mutex);
1325 INIT_HAL_MSG(msg, WCN36XX_HAL_UPDATE_PROBE_RSP_TEMPLATE_REQ);
1326
1327 if (skb->len > BEACON_TEMPLATE_SIZE) {
1328 wcn36xx_warn("probe response template is too big: %d\n",
1329 skb->len);
1330 return -E2BIG;
1331 }
1332
1333 msg.probe_resp_template_len = skb->len;
1334 memcpy(&msg.probe_resp_template, skb->data, skb->len);
1335
1336 memcpy(msg.bssid, vif->addr, ETH_ALEN);
1337
1338 PREPARE_HAL_BUF(wcn->hal_buf, msg);
1339
1340 wcn36xx_dbg(WCN36XX_DBG_HAL,
1341 "hal update probe rsp len %d bssid %pM\n",
1342 msg.probe_resp_template_len, msg.bssid);
1343
1344 ret = wcn36xx_smd_send_and_wait(wcn, msg.header.len);
1345 if (ret) {
1346 wcn36xx_err("Sending hal_update_proberesp_tmpl failed\n");
1347 goto out;
1348 }
1349 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
1350 if (ret) {
1351 wcn36xx_err("hal_update_proberesp_tmpl response failed err=%d\n",
1352 ret);
1353 goto out;
1354 }
1355out:
1356 mutex_unlock(&wcn->hal_mutex);
1357 return ret;
1358}
1359
1360int wcn36xx_smd_set_stakey(struct wcn36xx *wcn,
1361 enum ani_ed_type enc_type,
1362 u8 keyidx,
1363 u8 keylen,
1364 u8 *key,
1365 u8 sta_index)
1366{
1367 struct wcn36xx_hal_set_sta_key_req_msg msg_body;
1368 int ret = 0;
1369
1370 mutex_lock(&wcn->hal_mutex);
1371 INIT_HAL_MSG(msg_body, WCN36XX_HAL_SET_STAKEY_REQ);
1372
1373 msg_body.set_sta_key_params.sta_index = sta_index;
1374 msg_body.set_sta_key_params.enc_type = enc_type;
1375
1376 msg_body.set_sta_key_params.key[0].id = keyidx;
1377 msg_body.set_sta_key_params.key[0].unicast = 1;
1378 msg_body.set_sta_key_params.key[0].direction = WCN36XX_HAL_TX_RX;
1379 msg_body.set_sta_key_params.key[0].pae_role = 0;
1380 msg_body.set_sta_key_params.key[0].length = keylen;
1381 memcpy(msg_body.set_sta_key_params.key[0].key, key, keylen);
1382 msg_body.set_sta_key_params.single_tid_rc = 1;
1383
1384 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
1385
1386 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
1387 if (ret) {
1388 wcn36xx_err("Sending hal_set_stakey failed\n");
1389 goto out;
1390 }
1391 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
1392 if (ret) {
1393 wcn36xx_err("hal_set_stakey response failed err=%d\n", ret);
1394 goto out;
1395 }
1396out:
1397 mutex_unlock(&wcn->hal_mutex);
1398 return ret;
1399}
1400
1401int wcn36xx_smd_set_bsskey(struct wcn36xx *wcn,
1402 enum ani_ed_type enc_type,
1403 u8 keyidx,
1404 u8 keylen,
1405 u8 *key)
1406{
1407 struct wcn36xx_hal_set_bss_key_req_msg msg_body;
1408 int ret = 0;
1409
1410 mutex_lock(&wcn->hal_mutex);
1411 INIT_HAL_MSG(msg_body, WCN36XX_HAL_SET_BSSKEY_REQ);
1412 msg_body.bss_idx = 0;
1413 msg_body.enc_type = enc_type;
1414 msg_body.num_keys = 1;
1415 msg_body.keys[0].id = keyidx;
1416 msg_body.keys[0].unicast = 0;
1417 msg_body.keys[0].direction = WCN36XX_HAL_RX_ONLY;
1418 msg_body.keys[0].pae_role = 0;
1419 msg_body.keys[0].length = keylen;
1420 memcpy(msg_body.keys[0].key, key, keylen);
1421
1422 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
1423
1424 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
1425 if (ret) {
1426 wcn36xx_err("Sending hal_set_bsskey failed\n");
1427 goto out;
1428 }
1429 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
1430 if (ret) {
1431 wcn36xx_err("hal_set_bsskey response failed err=%d\n", ret);
1432 goto out;
1433 }
1434out:
1435 mutex_unlock(&wcn->hal_mutex);
1436 return ret;
1437}
1438
1439int wcn36xx_smd_remove_stakey(struct wcn36xx *wcn,
1440 enum ani_ed_type enc_type,
1441 u8 keyidx,
1442 u8 sta_index)
1443{
1444 struct wcn36xx_hal_remove_sta_key_req_msg msg_body;
1445 int ret = 0;
1446
1447 mutex_lock(&wcn->hal_mutex);
1448 INIT_HAL_MSG(msg_body, WCN36XX_HAL_RMV_STAKEY_REQ);
1449
1450 msg_body.sta_idx = sta_index;
1451 msg_body.enc_type = enc_type;
1452 msg_body.key_id = keyidx;
1453
1454 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
1455
1456 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
1457 if (ret) {
1458 wcn36xx_err("Sending hal_remove_stakey failed\n");
1459 goto out;
1460 }
1461 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
1462 if (ret) {
1463 wcn36xx_err("hal_remove_stakey response failed err=%d\n", ret);
1464 goto out;
1465 }
1466out:
1467 mutex_unlock(&wcn->hal_mutex);
1468 return ret;
1469}
1470
1471int wcn36xx_smd_remove_bsskey(struct wcn36xx *wcn,
1472 enum ani_ed_type enc_type,
1473 u8 keyidx)
1474{
1475 struct wcn36xx_hal_remove_bss_key_req_msg msg_body;
1476 int ret = 0;
1477
1478 mutex_lock(&wcn->hal_mutex);
1479 INIT_HAL_MSG(msg_body, WCN36XX_HAL_RMV_BSSKEY_REQ);
1480 msg_body.bss_idx = 0;
1481 msg_body.enc_type = enc_type;
1482 msg_body.key_id = keyidx;
1483
1484 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
1485
1486 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
1487 if (ret) {
1488 wcn36xx_err("Sending hal_remove_bsskey failed\n");
1489 goto out;
1490 }
1491 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
1492 if (ret) {
1493 wcn36xx_err("hal_remove_bsskey response failed err=%d\n", ret);
1494 goto out;
1495 }
1496out:
1497 mutex_unlock(&wcn->hal_mutex);
1498 return ret;
1499}
1500
1501int wcn36xx_smd_enter_bmps(struct wcn36xx *wcn, struct ieee80211_vif *vif)
1502{
1503 struct wcn36xx_hal_enter_bmps_req_msg msg_body;
1504 struct wcn36xx_vif *vif_priv = (struct wcn36xx_vif *)vif->drv_priv;
1505 int ret = 0;
1506
1507 mutex_lock(&wcn->hal_mutex);
1508 INIT_HAL_MSG(msg_body, WCN36XX_HAL_ENTER_BMPS_REQ);
1509
1510 msg_body.bss_index = vif_priv->bss_index;
1511 msg_body.tbtt = vif->bss_conf.sync_tsf;
1512 msg_body.dtim_period = vif_priv->dtim_period;
1513
1514 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
1515
1516 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
1517 if (ret) {
1518 wcn36xx_err("Sending hal_enter_bmps failed\n");
1519 goto out;
1520 }
1521 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
1522 if (ret) {
1523 wcn36xx_err("hal_enter_bmps response failed err=%d\n", ret);
1524 goto out;
1525 }
1526out:
1527 mutex_unlock(&wcn->hal_mutex);
1528 return ret;
1529}
1530
1531int wcn36xx_smd_exit_bmps(struct wcn36xx *wcn, struct ieee80211_vif *vif)
1532{
1533 struct wcn36xx_hal_enter_bmps_req_msg msg_body;
1534 struct wcn36xx_vif *vif_priv = (struct wcn36xx_vif *)vif->drv_priv;
1535 int ret = 0;
1536
1537 mutex_lock(&wcn->hal_mutex);
1538 INIT_HAL_MSG(msg_body, WCN36XX_HAL_EXIT_BMPS_REQ);
1539
1540 msg_body.bss_index = vif_priv->bss_index;
1541
1542 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
1543
1544 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
1545 if (ret) {
1546 wcn36xx_err("Sending hal_exit_bmps failed\n");
1547 goto out;
1548 }
1549 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
1550 if (ret) {
1551 wcn36xx_err("hal_exit_bmps response failed err=%d\n", ret);
1552 goto out;
1553 }
1554out:
1555 mutex_unlock(&wcn->hal_mutex);
1556 return ret;
1557}
1558int wcn36xx_smd_set_power_params(struct wcn36xx *wcn, bool ignore_dtim)
1559{
1560 struct wcn36xx_hal_set_power_params_req_msg msg_body;
1561 int ret = 0;
1562
1563 mutex_lock(&wcn->hal_mutex);
1564 INIT_HAL_MSG(msg_body, WCN36XX_HAL_SET_POWER_PARAMS_REQ);
1565
1566 /*
1567 * When host is down ignore every second dtim
1568 */
1569 if (ignore_dtim) {
1570 msg_body.ignore_dtim = 1;
1571 msg_body.dtim_period = 2;
1572 }
1573 msg_body.listen_interval = WCN36XX_LISTEN_INTERVAL(wcn);
1574
1575 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
1576
1577 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
1578 if (ret) {
1579 wcn36xx_err("Sending hal_set_power_params failed\n");
1580 goto out;
1581 }
1582
1583out:
1584 mutex_unlock(&wcn->hal_mutex);
1585 return ret;
1586}
1587/* Notice: This function should be called after associated, or else it
1588 * will be invalid
1589 */
1590int wcn36xx_smd_keep_alive_req(struct wcn36xx *wcn,
1591 struct ieee80211_vif *vif,
1592 int packet_type)
1593{
1594 struct wcn36xx_hal_keep_alive_req_msg msg_body;
1595 struct wcn36xx_vif *vif_priv = (struct wcn36xx_vif *)vif->drv_priv;
1596 int ret = 0;
1597
1598 mutex_lock(&wcn->hal_mutex);
1599 INIT_HAL_MSG(msg_body, WCN36XX_HAL_KEEP_ALIVE_REQ);
1600
1601 if (packet_type == WCN36XX_HAL_KEEP_ALIVE_NULL_PKT) {
1602 msg_body.bss_index = vif_priv->bss_index;
1603 msg_body.packet_type = WCN36XX_HAL_KEEP_ALIVE_NULL_PKT;
1604 msg_body.time_period = WCN36XX_KEEP_ALIVE_TIME_PERIOD;
1605 } else if (packet_type == WCN36XX_HAL_KEEP_ALIVE_UNSOLICIT_ARP_RSP) {
1606 /* TODO: it also support ARP response type */
1607 } else {
1608 wcn36xx_warn("unknow keep alive packet type %d\n", packet_type);
1609 return -EINVAL;
1610 }
1611
1612 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
1613
1614 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
1615 if (ret) {
1616 wcn36xx_err("Sending hal_exit_bmps failed\n");
1617 goto out;
1618 }
1619 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
1620 if (ret) {
1621 wcn36xx_err("hal_exit_bmps response failed err=%d\n", ret);
1622 goto out;
1623 }
1624out:
1625 mutex_unlock(&wcn->hal_mutex);
1626 return ret;
1627}
1628
1629int wcn36xx_smd_dump_cmd_req(struct wcn36xx *wcn, u32 arg1, u32 arg2,
1630 u32 arg3, u32 arg4, u32 arg5)
1631{
1632 struct wcn36xx_hal_dump_cmd_req_msg msg_body;
1633 int ret = 0;
1634
1635 mutex_lock(&wcn->hal_mutex);
1636 INIT_HAL_MSG(msg_body, WCN36XX_HAL_DUMP_COMMAND_REQ);
1637
1638 msg_body.arg1 = arg1;
1639 msg_body.arg2 = arg2;
1640 msg_body.arg3 = arg3;
1641 msg_body.arg4 = arg4;
1642 msg_body.arg5 = arg5;
1643
1644 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
1645
1646 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
1647 if (ret) {
1648 wcn36xx_err("Sending hal_dump_cmd failed\n");
1649 goto out;
1650 }
1651 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
1652 if (ret) {
1653 wcn36xx_err("hal_dump_cmd response failed err=%d\n", ret);
1654 goto out;
1655 }
1656out:
1657 mutex_unlock(&wcn->hal_mutex);
1658 return ret;
1659}
1660
1661static inline void set_feat_caps(u32 *bitmap,
1662 enum place_holder_in_cap_bitmap cap)
1663{
1664 int arr_idx, bit_idx;
1665
1666 if (cap < 0 || cap > 127) {
1667 wcn36xx_warn("error cap idx %d\n", cap);
1668 return;
1669 }
1670
1671 arr_idx = cap / 32;
1672 bit_idx = cap % 32;
1673 bitmap[arr_idx] |= (1 << bit_idx);
1674}
1675
1676static inline int get_feat_caps(u32 *bitmap,
1677 enum place_holder_in_cap_bitmap cap)
1678{
1679 int arr_idx, bit_idx;
1680 int ret = 0;
1681
1682 if (cap < 0 || cap > 127) {
1683 wcn36xx_warn("error cap idx %d\n", cap);
1684 return -EINVAL;
1685 }
1686
1687 arr_idx = cap / 32;
1688 bit_idx = cap % 32;
1689 ret = (bitmap[arr_idx] & (1 << bit_idx)) ? 1 : 0;
1690 return ret;
1691}
1692
1693static inline void clear_feat_caps(u32 *bitmap,
1694 enum place_holder_in_cap_bitmap cap)
1695{
1696 int arr_idx, bit_idx;
1697
1698 if (cap < 0 || cap > 127) {
1699 wcn36xx_warn("error cap idx %d\n", cap);
1700 return;
1701 }
1702
1703 arr_idx = cap / 32;
1704 bit_idx = cap % 32;
1705 bitmap[arr_idx] &= ~(1 << bit_idx);
1706}
1707
1708int wcn36xx_smd_feature_caps_exchange(struct wcn36xx *wcn)
1709{
1710 struct wcn36xx_hal_feat_caps_msg msg_body;
1711 int ret = 0;
1712
1713 mutex_lock(&wcn->hal_mutex);
1714 INIT_HAL_MSG(msg_body, WCN36XX_HAL_FEATURE_CAPS_EXCHANGE_REQ);
1715
1716 set_feat_caps(msg_body.feat_caps, STA_POWERSAVE);
1717
1718 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
1719
1720 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
1721 if (ret) {
1722 wcn36xx_err("Sending hal_feature_caps_exchange failed\n");
1723 goto out;
1724 }
1725 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
1726 if (ret) {
1727 wcn36xx_err("hal_feature_caps_exchange response failed err=%d\n",
1728 ret);
1729 goto out;
1730 }
1731out:
1732 mutex_unlock(&wcn->hal_mutex);
1733 return ret;
1734}
1735
1736int wcn36xx_smd_add_ba_session(struct wcn36xx *wcn,
1737 struct ieee80211_sta *sta,
1738 u16 tid,
1739 u16 *ssn,
1740 u8 direction,
1741 u8 sta_index)
1742{
1743 struct wcn36xx_hal_add_ba_session_req_msg msg_body;
1744 int ret = 0;
1745
1746 mutex_lock(&wcn->hal_mutex);
1747 INIT_HAL_MSG(msg_body, WCN36XX_HAL_ADD_BA_SESSION_REQ);
1748
1749 msg_body.sta_index = sta_index;
1750 memcpy(&msg_body.mac_addr, sta->addr, ETH_ALEN);
1751 msg_body.dialog_token = 0x10;
1752 msg_body.tid = tid;
1753
1754 /* Immediate BA because Delayed BA is not supported */
1755 msg_body.policy = 1;
1756 msg_body.buffer_size = WCN36XX_AGGR_BUFFER_SIZE;
1757 msg_body.timeout = 0;
1758 if (ssn)
1759 msg_body.ssn = *ssn;
1760 msg_body.direction = direction;
1761
1762 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
1763
1764 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
1765 if (ret) {
1766 wcn36xx_err("Sending hal_add_ba_session failed\n");
1767 goto out;
1768 }
1769 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
1770 if (ret) {
1771 wcn36xx_err("hal_add_ba_session response failed err=%d\n", ret);
1772 goto out;
1773 }
1774out:
1775 mutex_unlock(&wcn->hal_mutex);
1776 return ret;
1777}
1778
1779int wcn36xx_smd_add_ba(struct wcn36xx *wcn)
1780{
1781 struct wcn36xx_hal_add_ba_req_msg msg_body;
1782 int ret = 0;
1783
1784 mutex_lock(&wcn->hal_mutex);
1785 INIT_HAL_MSG(msg_body, WCN36XX_HAL_ADD_BA_REQ);
1786
1787 msg_body.session_id = 0;
1788 msg_body.win_size = WCN36XX_AGGR_BUFFER_SIZE;
1789
1790 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
1791
1792 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
1793 if (ret) {
1794 wcn36xx_err("Sending hal_add_ba failed\n");
1795 goto out;
1796 }
1797 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
1798 if (ret) {
1799 wcn36xx_err("hal_add_ba response failed err=%d\n", ret);
1800 goto out;
1801 }
1802out:
1803 mutex_unlock(&wcn->hal_mutex);
1804 return ret;
1805}
1806
1807int wcn36xx_smd_del_ba(struct wcn36xx *wcn, u16 tid, u8 sta_index)
1808{
1809 struct wcn36xx_hal_del_ba_req_msg msg_body;
1810 int ret = 0;
1811
1812 mutex_lock(&wcn->hal_mutex);
1813 INIT_HAL_MSG(msg_body, WCN36XX_HAL_DEL_BA_REQ);
1814
1815 msg_body.sta_index = sta_index;
1816 msg_body.tid = tid;
1817 msg_body.direction = 0;
1818 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
1819
1820 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
1821 if (ret) {
1822 wcn36xx_err("Sending hal_del_ba failed\n");
1823 goto out;
1824 }
1825 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
1826 if (ret) {
1827 wcn36xx_err("hal_del_ba response failed err=%d\n", ret);
1828 goto out;
1829 }
1830out:
1831 mutex_unlock(&wcn->hal_mutex);
1832 return ret;
1833}
1834
1835int wcn36xx_smd_trigger_ba(struct wcn36xx *wcn, u8 sta_index)
1836{
1837 struct wcn36xx_hal_trigger_ba_req_msg msg_body;
1838 struct wcn36xx_hal_trigget_ba_req_candidate *candidate;
1839 int ret = 0;
1840
1841 mutex_lock(&wcn->hal_mutex);
1842 INIT_HAL_MSG(msg_body, WCN36XX_HAL_TRIGGER_BA_REQ);
1843
1844 msg_body.session_id = 0;
1845 msg_body.candidate_cnt = 1;
1846 msg_body.header.len += sizeof(*candidate);
1847 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
1848
1849 candidate = (struct wcn36xx_hal_trigget_ba_req_candidate *)
1850 (wcn->hal_buf + sizeof(msg_body));
1851 candidate->sta_index = sta_index;
1852 candidate->tid_bitmap = 1;
1853
1854 ret = wcn36xx_smd_send_and_wait(wcn, msg_body.header.len);
1855 if (ret) {
1856 wcn36xx_err("Sending hal_trigger_ba failed\n");
1857 goto out;
1858 }
1859 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
1860 if (ret) {
1861 wcn36xx_err("hal_trigger_ba response failed err=%d\n", ret);
1862 goto out;
1863 }
1864out:
1865 mutex_unlock(&wcn->hal_mutex);
1866 return ret;
1867}
1868
1869static int wcn36xx_smd_tx_compl_ind(struct wcn36xx *wcn, void *buf, size_t len)
1870{
1871 struct wcn36xx_hal_tx_compl_ind_msg *rsp = buf;
1872
1873 if (len != sizeof(*rsp)) {
1874 wcn36xx_warn("Bad TX complete indication\n");
1875 return -EIO;
1876 }
1877
1878 wcn36xx_dxe_tx_ack_ind(wcn, rsp->status);
1879
1880 return 0;
1881}
1882
1883static int wcn36xx_smd_missed_beacon_ind(struct wcn36xx *wcn,
1884 void *buf,
1885 size_t len)
1886{
1887 struct wcn36xx_hal_missed_beacon_ind_msg *rsp = buf;
1888 struct ieee80211_vif *vif = NULL;
1889 struct wcn36xx_vif *tmp;
1890
1891 /* Old FW does not have bss index */
1892 if (wcn36xx_is_fw_version(wcn, 1, 2, 2, 24)) {
1893 list_for_each_entry(tmp, &wcn->vif_list, list) {
1894 wcn36xx_dbg(WCN36XX_DBG_HAL, "beacon missed bss_index %d\n",
1895 tmp->bss_index);
1896 vif = container_of((void *)tmp,
1897 struct ieee80211_vif,
1898 drv_priv);
1899 ieee80211_connection_loss(vif);
1900 }
1901 return 0;
1902 }
1903
1904 if (len != sizeof(*rsp)) {
1905 wcn36xx_warn("Corrupted missed beacon indication\n");
1906 return -EIO;
1907 }
1908
1909 list_for_each_entry(tmp, &wcn->vif_list, list) {
1910 if (tmp->bss_index == rsp->bss_index) {
1911 wcn36xx_dbg(WCN36XX_DBG_HAL, "beacon missed bss_index %d\n",
1912 rsp->bss_index);
1913 vif = container_of((void *)tmp,
1914 struct ieee80211_vif,
1915 drv_priv);
1916 ieee80211_connection_loss(vif);
1917 return 0;
1918 }
1919 }
1920
1921 wcn36xx_warn("BSS index %d not found\n", rsp->bss_index);
1922 return -ENOENT;
1923}
1924
1925static int wcn36xx_smd_delete_sta_context_ind(struct wcn36xx *wcn,
1926 void *buf,
1927 size_t len)
1928{
1929 struct wcn36xx_hal_delete_sta_context_ind_msg *rsp = buf;
1930 struct wcn36xx_vif *tmp;
1931 struct ieee80211_sta *sta = NULL;
1932
1933 if (len != sizeof(*rsp)) {
1934 wcn36xx_warn("Corrupted delete sta indication\n");
1935 return -EIO;
1936 }
1937
1938 list_for_each_entry(tmp, &wcn->vif_list, list) {
1939 if (sta && (tmp->sta->sta_index == rsp->sta_id)) {
1940 sta = container_of((void *)tmp->sta,
1941 struct ieee80211_sta,
1942 drv_priv);
1943 wcn36xx_dbg(WCN36XX_DBG_HAL,
1944 "delete station indication %pM index %d\n",
1945 rsp->addr2,
1946 rsp->sta_id);
1947 ieee80211_report_low_ack(sta, 0);
1948 return 0;
1949 }
1950 }
1951
1952 wcn36xx_warn("STA with addr %pM and index %d not found\n",
1953 rsp->addr2,
1954 rsp->sta_id);
1955 return -ENOENT;
1956}
1957
1958int wcn36xx_smd_update_cfg(struct wcn36xx *wcn, u32 cfg_id, u32 value)
1959{
1960 struct wcn36xx_hal_update_cfg_req_msg msg_body, *body;
1961 size_t len;
1962 int ret = 0;
1963
1964 mutex_lock(&wcn->hal_mutex);
1965 INIT_HAL_MSG(msg_body, WCN36XX_HAL_UPDATE_CFG_REQ);
1966
1967 PREPARE_HAL_BUF(wcn->hal_buf, msg_body);
1968
1969 body = (struct wcn36xx_hal_update_cfg_req_msg *) wcn->hal_buf;
1970 len = msg_body.header.len;
1971
1972 put_cfg_tlv_u32(wcn, &len, cfg_id, value);
1973 body->header.len = len;
1974 body->len = len - sizeof(*body);
1975
1976 ret = wcn36xx_smd_send_and_wait(wcn, body->header.len);
1977 if (ret) {
1978 wcn36xx_err("Sending hal_update_cfg failed\n");
1979 goto out;
1980 }
1981 ret = wcn36xx_smd_rsp_status_check(wcn->hal_buf, wcn->hal_rsp_len);
1982 if (ret) {
1983 wcn36xx_err("hal_update_cfg response failed err=%d\n", ret);
1984 goto out;
1985 }
1986out:
1987 mutex_unlock(&wcn->hal_mutex);
1988 return ret;
1989}
1990static void wcn36xx_smd_rsp_process(struct wcn36xx *wcn, void *buf, size_t len)
1991{
1992 struct wcn36xx_hal_msg_header *msg_header = buf;
1993 struct wcn36xx_hal_ind_msg *msg_ind;
1994 wcn36xx_dbg_dump(WCN36XX_DBG_SMD_DUMP, "SMD <<< ", buf, len);
1995
1996 switch (msg_header->msg_type) {
1997 case WCN36XX_HAL_START_RSP:
1998 case WCN36XX_HAL_CONFIG_STA_RSP:
1999 case WCN36XX_HAL_CONFIG_BSS_RSP:
2000 case WCN36XX_HAL_ADD_STA_SELF_RSP:
2001 case WCN36XX_HAL_STOP_RSP:
2002 case WCN36XX_HAL_DEL_STA_SELF_RSP:
2003 case WCN36XX_HAL_DELETE_STA_RSP:
2004 case WCN36XX_HAL_INIT_SCAN_RSP:
2005 case WCN36XX_HAL_START_SCAN_RSP:
2006 case WCN36XX_HAL_END_SCAN_RSP:
2007 case WCN36XX_HAL_FINISH_SCAN_RSP:
2008 case WCN36XX_HAL_DOWNLOAD_NV_RSP:
2009 case WCN36XX_HAL_DELETE_BSS_RSP:
2010 case WCN36XX_HAL_SEND_BEACON_RSP:
2011 case WCN36XX_HAL_SET_LINK_ST_RSP:
2012 case WCN36XX_HAL_UPDATE_PROBE_RSP_TEMPLATE_RSP:
2013 case WCN36XX_HAL_SET_BSSKEY_RSP:
2014 case WCN36XX_HAL_SET_STAKEY_RSP:
2015 case WCN36XX_HAL_RMV_STAKEY_RSP:
2016 case WCN36XX_HAL_RMV_BSSKEY_RSP:
2017 case WCN36XX_HAL_ENTER_BMPS_RSP:
2018 case WCN36XX_HAL_SET_POWER_PARAMS_RSP:
2019 case WCN36XX_HAL_EXIT_BMPS_RSP:
2020 case WCN36XX_HAL_KEEP_ALIVE_RSP:
2021 case WCN36XX_HAL_DUMP_COMMAND_RSP:
2022 case WCN36XX_HAL_ADD_BA_SESSION_RSP:
2023 case WCN36XX_HAL_ADD_BA_RSP:
2024 case WCN36XX_HAL_DEL_BA_RSP:
2025 case WCN36XX_HAL_TRIGGER_BA_RSP:
2026 case WCN36XX_HAL_UPDATE_CFG_RSP:
2027 case WCN36XX_HAL_JOIN_RSP:
2028 case WCN36XX_HAL_UPDATE_SCAN_PARAM_RSP:
2029 case WCN36XX_HAL_CH_SWITCH_RSP:
2030 case WCN36XX_HAL_FEATURE_CAPS_EXCHANGE_RSP:
2031 memcpy(wcn->hal_buf, buf, len);
2032 wcn->hal_rsp_len = len;
2033 complete(&wcn->hal_rsp_compl);
2034 break;
2035
2036 case WCN36XX_HAL_OTA_TX_COMPL_IND:
2037 case WCN36XX_HAL_MISSED_BEACON_IND:
2038 case WCN36XX_HAL_DELETE_STA_CONTEXT_IND:
2039 mutex_lock(&wcn->hal_ind_mutex);
2040 msg_ind = kmalloc(sizeof(*msg_ind), GFP_KERNEL);
2041 msg_ind->msg_len = len;
2042 msg_ind->msg = kmalloc(len, GFP_KERNEL);
2043 memcpy(msg_ind->msg, buf, len);
2044 list_add_tail(&msg_ind->list, &wcn->hal_ind_queue);
2045 queue_work(wcn->hal_ind_wq, &wcn->hal_ind_work);
2046 wcn36xx_dbg(WCN36XX_DBG_HAL, "indication arrived\n");
2047 mutex_unlock(&wcn->hal_ind_mutex);
2048 break;
2049 default:
2050 wcn36xx_err("SMD_EVENT (%d) not supported\n",
2051 msg_header->msg_type);
2052 }
2053}
2054static void wcn36xx_ind_smd_work(struct work_struct *work)
2055{
2056 struct wcn36xx *wcn =
2057 container_of(work, struct wcn36xx, hal_ind_work);
2058 struct wcn36xx_hal_msg_header *msg_header;
2059 struct wcn36xx_hal_ind_msg *hal_ind_msg;
2060
2061 mutex_lock(&wcn->hal_ind_mutex);
2062
2063 hal_ind_msg = list_first_entry(&wcn->hal_ind_queue,
2064 struct wcn36xx_hal_ind_msg,
2065 list);
2066
2067 msg_header = (struct wcn36xx_hal_msg_header *)hal_ind_msg->msg;
2068
2069 switch (msg_header->msg_type) {
2070 case WCN36XX_HAL_OTA_TX_COMPL_IND:
2071 wcn36xx_smd_tx_compl_ind(wcn,
2072 hal_ind_msg->msg,
2073 hal_ind_msg->msg_len);
2074 break;
2075 case WCN36XX_HAL_MISSED_BEACON_IND:
2076 wcn36xx_smd_missed_beacon_ind(wcn,
2077 hal_ind_msg->msg,
2078 hal_ind_msg->msg_len);
2079 break;
2080 case WCN36XX_HAL_DELETE_STA_CONTEXT_IND:
2081 wcn36xx_smd_delete_sta_context_ind(wcn,
2082 hal_ind_msg->msg,
2083 hal_ind_msg->msg_len);
2084 break;
2085 default:
2086 wcn36xx_err("SMD_EVENT (%d) not supported\n",
2087 msg_header->msg_type);
2088 }
2089 list_del(wcn->hal_ind_queue.next);
2090 kfree(hal_ind_msg->msg);
2091 kfree(hal_ind_msg);
2092 mutex_unlock(&wcn->hal_ind_mutex);
2093}
2094int wcn36xx_smd_open(struct wcn36xx *wcn)
2095{
2096 int ret = 0;
2097 wcn->hal_ind_wq = create_freezable_workqueue("wcn36xx_smd_ind");
2098 if (!wcn->hal_ind_wq) {
2099 wcn36xx_err("failed to allocate wq\n");
2100 ret = -ENOMEM;
2101 goto out;
2102 }
2103 INIT_WORK(&wcn->hal_ind_work, wcn36xx_ind_smd_work);
2104 INIT_LIST_HEAD(&wcn->hal_ind_queue);
2105 mutex_init(&wcn->hal_ind_mutex);
2106
2107 ret = wcn->ctrl_ops->open(wcn, wcn36xx_smd_rsp_process);
2108 if (ret) {
2109 wcn36xx_err("failed to open control channel\n");
2110 goto free_wq;
2111 }
2112
2113 return ret;
2114
2115free_wq:
2116 destroy_workqueue(wcn->hal_ind_wq);
2117out:
2118 return ret;
2119}
2120
2121void wcn36xx_smd_close(struct wcn36xx *wcn)
2122{
2123 wcn->ctrl_ops->close();
2124 destroy_workqueue(wcn->hal_ind_wq);
2125 mutex_destroy(&wcn->hal_ind_mutex);
2126}
diff --git a/drivers/net/wireless/ath/wcn36xx/smd.h b/drivers/net/wireless/ath/wcn36xx/smd.h
new file mode 100644
index 000000000000..e7c39019c6f1
--- /dev/null
+++ b/drivers/net/wireless/ath/wcn36xx/smd.h
@@ -0,0 +1,127 @@
1/*
2 * Copyright (c) 2013 Eugene Krasnikov <k.eugene.e@gmail.com>
3 *
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
7 *
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
11 * SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION
13 * OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN
14 * CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
15 */
16
17#ifndef _SMD_H_
18#define _SMD_H_
19
20#include "wcn36xx.h"
21
22/* Max shared size is 4k but we take less.*/
23#define WCN36XX_NV_FRAGMENT_SIZE 3072
24
25#define WCN36XX_HAL_BUF_SIZE 4096
26
27#define HAL_MSG_TIMEOUT 200
28#define WCN36XX_SMSM_WLAN_TX_ENABLE 0x00000400
29#define WCN36XX_SMSM_WLAN_TX_RINGS_EMPTY 0x00000200
30/* The PNO version info be contained in the rsp msg */
31#define WCN36XX_FW_MSG_PNO_VERSION_MASK 0x8000
32
33enum wcn36xx_fw_msg_result {
34 WCN36XX_FW_MSG_RESULT_SUCCESS = 0,
35 WCN36XX_FW_MSG_RESULT_SUCCESS_SYNC = 1,
36
37 WCN36XX_FW_MSG_RESULT_MEM_FAIL = 5,
38};
39
40/******************************/
41/* SMD requests and responses */
42/******************************/
43struct wcn36xx_fw_msg_status_rsp {
44 u32 status;
45} __packed;
46
47struct wcn36xx_hal_ind_msg {
48 struct list_head list;
49 u8 *msg;
50 size_t msg_len;
51};
52
53struct wcn36xx;
54
55int wcn36xx_smd_open(struct wcn36xx *wcn);
56void wcn36xx_smd_close(struct wcn36xx *wcn);
57
58int wcn36xx_smd_load_nv(struct wcn36xx *wcn);
59int wcn36xx_smd_start(struct wcn36xx *wcn);
60int wcn36xx_smd_stop(struct wcn36xx *wcn);
61int wcn36xx_smd_init_scan(struct wcn36xx *wcn, enum wcn36xx_hal_sys_mode mode);
62int wcn36xx_smd_start_scan(struct wcn36xx *wcn);
63int wcn36xx_smd_end_scan(struct wcn36xx *wcn);
64int wcn36xx_smd_finish_scan(struct wcn36xx *wcn,
65 enum wcn36xx_hal_sys_mode mode);
66int wcn36xx_smd_update_scan_params(struct wcn36xx *wcn);
67int wcn36xx_smd_add_sta_self(struct wcn36xx *wcn, struct ieee80211_vif *vif);
68int wcn36xx_smd_delete_sta_self(struct wcn36xx *wcn, u8 *addr);
69int wcn36xx_smd_delete_sta(struct wcn36xx *wcn, u8 sta_index);
70int wcn36xx_smd_join(struct wcn36xx *wcn, const u8 *bssid, u8 *vif, u8 ch);
71int wcn36xx_smd_set_link_st(struct wcn36xx *wcn, const u8 *bssid,
72 const u8 *sta_mac,
73 enum wcn36xx_hal_link_state state);
74int wcn36xx_smd_config_bss(struct wcn36xx *wcn, struct ieee80211_vif *vif,
75 struct ieee80211_sta *sta, const u8 *bssid,
76 bool update);
77int wcn36xx_smd_delete_bss(struct wcn36xx *wcn, struct ieee80211_vif *vif);
78int wcn36xx_smd_config_sta(struct wcn36xx *wcn, struct ieee80211_vif *vif,
79 struct ieee80211_sta *sta);
80int wcn36xx_smd_send_beacon(struct wcn36xx *wcn, struct ieee80211_vif *vif,
81 struct sk_buff *skb_beacon, u16 tim_off,
82 u16 p2p_off);
83int wcn36xx_smd_switch_channel(struct wcn36xx *wcn,
84 struct ieee80211_vif *vif, int ch);
85int wcn36xx_smd_update_proberesp_tmpl(struct wcn36xx *wcn,
86 struct ieee80211_vif *vif,
87 struct sk_buff *skb);
88int wcn36xx_smd_set_stakey(struct wcn36xx *wcn,
89 enum ani_ed_type enc_type,
90 u8 keyidx,
91 u8 keylen,
92 u8 *key,
93 u8 sta_index);
94int wcn36xx_smd_set_bsskey(struct wcn36xx *wcn,
95 enum ani_ed_type enc_type,
96 u8 keyidx,
97 u8 keylen,
98 u8 *key);
99int wcn36xx_smd_remove_stakey(struct wcn36xx *wcn,
100 enum ani_ed_type enc_type,
101 u8 keyidx,
102 u8 sta_index);
103int wcn36xx_smd_remove_bsskey(struct wcn36xx *wcn,
104 enum ani_ed_type enc_type,
105 u8 keyidx);
106int wcn36xx_smd_enter_bmps(struct wcn36xx *wcn, struct ieee80211_vif *vif);
107int wcn36xx_smd_exit_bmps(struct wcn36xx *wcn, struct ieee80211_vif *vif);
108int wcn36xx_smd_set_power_params(struct wcn36xx *wcn, bool ignore_dtim);
109int wcn36xx_smd_keep_alive_req(struct wcn36xx *wcn,
110 struct ieee80211_vif *vif,
111 int packet_type);
112int wcn36xx_smd_dump_cmd_req(struct wcn36xx *wcn, u32 arg1, u32 arg2,
113 u32 arg3, u32 arg4, u32 arg5);
114int wcn36xx_smd_feature_caps_exchange(struct wcn36xx *wcn);
115
116int wcn36xx_smd_add_ba_session(struct wcn36xx *wcn,
117 struct ieee80211_sta *sta,
118 u16 tid,
119 u16 *ssn,
120 u8 direction,
121 u8 sta_index);
122int wcn36xx_smd_add_ba(struct wcn36xx *wcn);
123int wcn36xx_smd_del_ba(struct wcn36xx *wcn, u16 tid, u8 sta_index);
124int wcn36xx_smd_trigger_ba(struct wcn36xx *wcn, u8 sta_index);
125
126int wcn36xx_smd_update_cfg(struct wcn36xx *wcn, u32 cfg_id, u32 value);
127#endif /* _SMD_H_ */
diff --git a/drivers/net/wireless/ath/wcn36xx/txrx.c b/drivers/net/wireless/ath/wcn36xx/txrx.c
new file mode 100644
index 000000000000..b2b60e30caaf
--- /dev/null
+++ b/drivers/net/wireless/ath/wcn36xx/txrx.c
@@ -0,0 +1,284 @@
1/*
2 * Copyright (c) 2013 Eugene Krasnikov <k.eugene.e@gmail.com>
3 *
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
7 *
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
11 * SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION
13 * OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN
14 * CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
15 */
16
17#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
18
19#include "txrx.h"
20
21static inline int get_rssi0(struct wcn36xx_rx_bd *bd)
22{
23 return 100 - ((bd->phy_stat0 >> 24) & 0xff);
24}
25
26int wcn36xx_rx_skb(struct wcn36xx *wcn, struct sk_buff *skb)
27{
28 struct ieee80211_rx_status status;
29 struct ieee80211_hdr *hdr;
30 struct wcn36xx_rx_bd *bd;
31 u16 fc, sn;
32
33 /*
34 * All fields must be 0, otherwise it can lead to
35 * unexpected consequences.
36 */
37 memset(&status, 0, sizeof(status));
38
39 bd = (struct wcn36xx_rx_bd *)skb->data;
40 buff_to_be((u32 *)bd, sizeof(*bd)/sizeof(u32));
41 wcn36xx_dbg_dump(WCN36XX_DBG_RX_DUMP,
42 "BD <<< ", (char *)bd,
43 sizeof(struct wcn36xx_rx_bd));
44
45 skb_put(skb, bd->pdu.mpdu_header_off + bd->pdu.mpdu_len);
46 skb_pull(skb, bd->pdu.mpdu_header_off);
47
48 status.mactime = 10;
49 status.freq = WCN36XX_CENTER_FREQ(wcn);
50 status.band = WCN36XX_BAND(wcn);
51 status.signal = -get_rssi0(bd);
52 status.antenna = 1;
53 status.rate_idx = 1;
54 status.flag = 0;
55 status.rx_flags = 0;
56 status.flag |= RX_FLAG_IV_STRIPPED |
57 RX_FLAG_MMIC_STRIPPED |
58 RX_FLAG_DECRYPTED;
59
60 wcn36xx_dbg(WCN36XX_DBG_RX, "status.flags=%x status->vendor_radiotap_len=%x\n",
61 status.flag, status.vendor_radiotap_len);
62
63 memcpy(IEEE80211_SKB_RXCB(skb), &status, sizeof(status));
64
65 hdr = (struct ieee80211_hdr *) skb->data;
66 fc = __le16_to_cpu(hdr->frame_control);
67 sn = IEEE80211_SEQ_TO_SN(__le16_to_cpu(hdr->seq_ctrl));
68
69 if (ieee80211_is_beacon(hdr->frame_control)) {
70 wcn36xx_dbg(WCN36XX_DBG_BEACON, "beacon skb %p len %d fc %04x sn %d\n",
71 skb, skb->len, fc, sn);
72 wcn36xx_dbg_dump(WCN36XX_DBG_BEACON_DUMP, "SKB <<< ",
73 (char *)skb->data, skb->len);
74 } else {
75 wcn36xx_dbg(WCN36XX_DBG_RX, "rx skb %p len %d fc %04x sn %d\n",
76 skb, skb->len, fc, sn);
77 wcn36xx_dbg_dump(WCN36XX_DBG_RX_DUMP, "SKB <<< ",
78 (char *)skb->data, skb->len);
79 }
80
81 ieee80211_rx_irqsafe(wcn->hw, skb);
82
83 return 0;
84}
85
86static void wcn36xx_set_tx_pdu(struct wcn36xx_tx_bd *bd,
87 u32 mpdu_header_len,
88 u32 len,
89 u16 tid)
90{
91 bd->pdu.mpdu_header_len = mpdu_header_len;
92 bd->pdu.mpdu_header_off = sizeof(*bd);
93 bd->pdu.mpdu_data_off = bd->pdu.mpdu_header_len +
94 bd->pdu.mpdu_header_off;
95 bd->pdu.mpdu_len = len;
96 bd->pdu.tid = tid;
97}
98
99static inline struct wcn36xx_vif *get_vif_by_addr(struct wcn36xx *wcn,
100 u8 *addr)
101{
102 struct wcn36xx_vif *vif_priv = NULL;
103 struct ieee80211_vif *vif = NULL;
104 list_for_each_entry(vif_priv, &wcn->vif_list, list) {
105 vif = container_of((void *)vif_priv,
106 struct ieee80211_vif,
107 drv_priv);
108 if (memcmp(vif->addr, addr, ETH_ALEN) == 0)
109 return vif_priv;
110 }
111 wcn36xx_warn("vif %pM not found\n", addr);
112 return NULL;
113}
114static void wcn36xx_set_tx_data(struct wcn36xx_tx_bd *bd,
115 struct wcn36xx *wcn,
116 struct wcn36xx_vif **vif_priv,
117 struct wcn36xx_sta *sta_priv,
118 struct ieee80211_hdr *hdr,
119 bool bcast)
120{
121 struct ieee80211_vif *vif = NULL;
122 struct wcn36xx_vif *__vif_priv = NULL;
123 bd->bd_rate = WCN36XX_BD_RATE_DATA;
124
125 /*
126 * For not unicast frames mac80211 will not set sta pointer so use
127 * self_sta_index instead.
128 */
129 if (sta_priv) {
130 __vif_priv = sta_priv->vif;
131 vif = container_of((void *)__vif_priv,
132 struct ieee80211_vif,
133 drv_priv);
134
135 if (vif->type == NL80211_IFTYPE_STATION) {
136 bd->sta_index = sta_priv->bss_sta_index;
137 bd->dpu_desc_idx = sta_priv->bss_dpu_desc_index;
138 } else if (vif->type == NL80211_IFTYPE_AP ||
139 vif->type == NL80211_IFTYPE_ADHOC ||
140 vif->type == NL80211_IFTYPE_MESH_POINT) {
141 bd->sta_index = sta_priv->sta_index;
142 bd->dpu_desc_idx = sta_priv->dpu_desc_index;
143 }
144 } else {
145 __vif_priv = get_vif_by_addr(wcn, hdr->addr2);
146 bd->sta_index = __vif_priv->self_sta_index;
147 bd->dpu_desc_idx = __vif_priv->self_dpu_desc_index;
148 }
149
150 bd->dpu_sign = __vif_priv->ucast_dpu_signature;
151
152 if (ieee80211_is_nullfunc(hdr->frame_control) ||
153 (sta_priv && !sta_priv->is_data_encrypted))
154 bd->dpu_ne = 1;
155
156 if (bcast) {
157 bd->ub = 1;
158 bd->ack_policy = 1;
159 }
160 *vif_priv = __vif_priv;
161}
162
163static void wcn36xx_set_tx_mgmt(struct wcn36xx_tx_bd *bd,
164 struct wcn36xx *wcn,
165 struct wcn36xx_vif **vif_priv,
166 struct ieee80211_hdr *hdr,
167 bool bcast)
168{
169 struct wcn36xx_vif *__vif_priv =
170 get_vif_by_addr(wcn, hdr->addr2);
171 bd->sta_index = __vif_priv->self_sta_index;
172 bd->dpu_desc_idx = __vif_priv->self_dpu_desc_index;
173 bd->dpu_ne = 1;
174
175 /* default rate for unicast */
176 if (ieee80211_is_mgmt(hdr->frame_control))
177 bd->bd_rate = (WCN36XX_BAND(wcn) == IEEE80211_BAND_5GHZ) ?
178 WCN36XX_BD_RATE_CTRL :
179 WCN36XX_BD_RATE_MGMT;
180 else if (ieee80211_is_ctl(hdr->frame_control))
181 bd->bd_rate = WCN36XX_BD_RATE_CTRL;
182 else
183 wcn36xx_warn("frame control type unknown\n");
184
185 /*
186 * In joining state trick hardware that probe is sent as
187 * unicast even if address is broadcast.
188 */
189 if (__vif_priv->is_joining &&
190 ieee80211_is_probe_req(hdr->frame_control))
191 bcast = false;
192
193 if (bcast) {
194 /* broadcast */
195 bd->ub = 1;
196 /* No ack needed not unicast */
197 bd->ack_policy = 1;
198 bd->queue_id = WCN36XX_TX_B_WQ_ID;
199 } else
200 bd->queue_id = WCN36XX_TX_U_WQ_ID;
201 *vif_priv = __vif_priv;
202}
203
204int wcn36xx_start_tx(struct wcn36xx *wcn,
205 struct wcn36xx_sta *sta_priv,
206 struct sk_buff *skb)
207{
208 struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)skb->data;
209 struct wcn36xx_vif *vif_priv = NULL;
210 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb);
211 unsigned long flags;
212 bool is_low = ieee80211_is_data(hdr->frame_control);
213 bool bcast = is_broadcast_ether_addr(hdr->addr1) ||
214 is_multicast_ether_addr(hdr->addr1);
215 struct wcn36xx_tx_bd *bd = wcn36xx_dxe_get_next_bd(wcn, is_low);
216
217 if (!bd) {
218 /*
219 * TX DXE are used in pairs. One for the BD and one for the
220 * actual frame. The BD DXE's has a preallocated buffer while
221 * the skb ones does not. If this isn't true something is really
222 * wierd. TODO: Recover from this situation
223 */
224
225 wcn36xx_err("bd address may not be NULL for BD DXE\n");
226 return -EINVAL;
227 }
228
229 memset(bd, 0, sizeof(*bd));
230
231 wcn36xx_dbg(WCN36XX_DBG_TX,
232 "tx skb %p len %d fc %04x sn %d %s %s\n",
233 skb, skb->len, __le16_to_cpu(hdr->frame_control),
234 IEEE80211_SEQ_TO_SN(__le16_to_cpu(hdr->seq_ctrl)),
235 is_low ? "low" : "high", bcast ? "bcast" : "ucast");
236
237 wcn36xx_dbg_dump(WCN36XX_DBG_TX_DUMP, "", skb->data, skb->len);
238
239 bd->dpu_rf = WCN36XX_BMU_WQ_TX;
240
241 bd->tx_comp = info->flags & IEEE80211_TX_CTL_REQ_TX_STATUS;
242 if (bd->tx_comp) {
243 wcn36xx_dbg(WCN36XX_DBG_DXE, "TX_ACK status requested\n");
244 spin_lock_irqsave(&wcn->dxe_lock, flags);
245 if (wcn->tx_ack_skb) {
246 spin_unlock_irqrestore(&wcn->dxe_lock, flags);
247 wcn36xx_warn("tx_ack_skb already set\n");
248 return -EINVAL;
249 }
250
251 wcn->tx_ack_skb = skb;
252 spin_unlock_irqrestore(&wcn->dxe_lock, flags);
253
254 /* Only one at a time is supported by fw. Stop the TX queues
255 * until the ack status gets back.
256 *
257 * TODO: Add watchdog in case FW does not answer
258 */
259 ieee80211_stop_queues(wcn->hw);
260 }
261
262 /* Data frames served first*/
263 if (is_low) {
264 wcn36xx_set_tx_data(bd, wcn, &vif_priv, sta_priv, hdr, bcast);
265 wcn36xx_set_tx_pdu(bd,
266 ieee80211_is_data_qos(hdr->frame_control) ?
267 sizeof(struct ieee80211_qos_hdr) :
268 sizeof(struct ieee80211_hdr_3addr),
269 skb->len, sta_priv ? sta_priv->tid : 0);
270 } else {
271 /* MGMT and CTRL frames are handeld here*/
272 wcn36xx_set_tx_mgmt(bd, wcn, &vif_priv, hdr, bcast);
273 wcn36xx_set_tx_pdu(bd,
274 ieee80211_is_data_qos(hdr->frame_control) ?
275 sizeof(struct ieee80211_qos_hdr) :
276 sizeof(struct ieee80211_hdr_3addr),
277 skb->len, WCN36XX_TID);
278 }
279
280 buff_to_be((u32 *)bd, sizeof(*bd)/sizeof(u32));
281 bd->tx_bd_sign = 0xbdbdbdbd;
282
283 return wcn36xx_dxe_tx_frame(wcn, vif_priv, skb, is_low);
284}
diff --git a/drivers/net/wireless/ath/wcn36xx/txrx.h b/drivers/net/wireless/ath/wcn36xx/txrx.h
new file mode 100644
index 000000000000..bbfbcf808c77
--- /dev/null
+++ b/drivers/net/wireless/ath/wcn36xx/txrx.h
@@ -0,0 +1,160 @@
1/*
2 * Copyright (c) 2013 Eugene Krasnikov <k.eugene.e@gmail.com>
3 *
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
7 *
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
11 * SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION
13 * OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN
14 * CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
15 */
16
17#ifndef _TXRX_H_
18#define _TXRX_H_
19
20#include <linux/etherdevice.h>
21#include "wcn36xx.h"
22
23/* TODO describe all properties */
24#define WCN36XX_802_11_HEADER_LEN 24
25#define WCN36XX_BMU_WQ_TX 25
26#define WCN36XX_TID 7
27/* broadcast wq ID */
28#define WCN36XX_TX_B_WQ_ID 0xA
29#define WCN36XX_TX_U_WQ_ID 0x9
30/* bd_rate */
31#define WCN36XX_BD_RATE_DATA 0
32#define WCN36XX_BD_RATE_MGMT 2
33#define WCN36XX_BD_RATE_CTRL 3
34
35struct wcn36xx_pdu {
36 u32 dpu_fb:8;
37 u32 adu_fb:8;
38 u32 pdu_id:16;
39
40 /* 0x04*/
41 u32 tail_pdu_idx:16;
42 u32 head_pdu_idx:16;
43
44 /* 0x08*/
45 u32 pdu_count:7;
46 u32 mpdu_data_off:9;
47 u32 mpdu_header_off:8;
48 u32 mpdu_header_len:8;
49
50 /* 0x0c*/
51 u32 reserved4:8;
52 u32 tid:4;
53 u32 reserved3:4;
54 u32 mpdu_len:16;
55};
56
57struct wcn36xx_rx_bd {
58 u32 bdt:2;
59 u32 ft:1;
60 u32 dpu_ne:1;
61 u32 rx_key_id:3;
62 u32 ub:1;
63 u32 rmf:1;
64 u32 uma_bypass:1;
65 u32 csr11:1;
66 u32 reserved0:1;
67 u32 scan_learn:1;
68 u32 rx_ch:4;
69 u32 rtsf:1;
70 u32 bsf:1;
71 u32 a2hf:1;
72 u32 st_auf:1;
73 u32 dpu_sign:3;
74 u32 dpu_rf:8;
75
76 struct wcn36xx_pdu pdu;
77
78 /* 0x14*/
79 u32 addr3:8;
80 u32 addr2:8;
81 u32 addr1:8;
82 u32 dpu_desc_idx:8;
83
84 /* 0x18*/
85 u32 rxp_flags:23;
86 u32 rate_id:9;
87
88 u32 phy_stat0;
89 u32 phy_stat1;
90
91 /* 0x24 */
92 u32 rx_times;
93
94 u32 pmi_cmd[6];
95
96 /* 0x40 */
97 u32 reserved7:4;
98 u32 reorder_slot_id:6;
99 u32 reorder_fwd_id:6;
100 u32 reserved6:12;
101 u32 reorder_code:4;
102
103 /* 0x44 */
104 u32 exp_seq_num:12;
105 u32 cur_seq_num:12;
106 u32 fr_type_subtype:8;
107
108 /* 0x48 */
109 u32 msdu_size:16;
110 u32 sub_fr_id:4;
111 u32 proc_order:4;
112 u32 reserved9:4;
113 u32 aef:1;
114 u32 lsf:1;
115 u32 esf:1;
116 u32 asf:1;
117};
118
119struct wcn36xx_tx_bd {
120 u32 bdt:2;
121 u32 ft:1;
122 u32 dpu_ne:1;
123 u32 fw_tx_comp:1;
124 u32 tx_comp:1;
125 u32 reserved1:1;
126 u32 ub:1;
127 u32 rmf:1;
128 u32 reserved0:12;
129 u32 dpu_sign:3;
130 u32 dpu_rf:8;
131
132 struct wcn36xx_pdu pdu;
133
134 /* 0x14*/
135 u32 reserved5:7;
136 u32 queue_id:5;
137 u32 bd_rate:2;
138 u32 ack_policy:2;
139 u32 sta_index:8;
140 u32 dpu_desc_idx:8;
141
142 u32 tx_bd_sign;
143 u32 reserved6;
144 u32 dxe_start_time;
145 u32 dxe_end_time;
146
147 /*u32 tcp_udp_start_off:10;
148 u32 header_cks:16;
149 u32 reserved7:6;*/
150};
151
152struct wcn36xx_sta;
153struct wcn36xx;
154
155int wcn36xx_rx_skb(struct wcn36xx *wcn, struct sk_buff *skb);
156int wcn36xx_start_tx(struct wcn36xx *wcn,
157 struct wcn36xx_sta *sta_priv,
158 struct sk_buff *skb);
159
160#endif /* _TXRX_H_ */
diff --git a/drivers/net/wireless/ath/wcn36xx/wcn36xx.h b/drivers/net/wireless/ath/wcn36xx/wcn36xx.h
new file mode 100644
index 000000000000..58b63833e8e7
--- /dev/null
+++ b/drivers/net/wireless/ath/wcn36xx/wcn36xx.h
@@ -0,0 +1,238 @@
1/*
2 * Copyright (c) 2013 Eugene Krasnikov <k.eugene.e@gmail.com>
3 *
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
7 *
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
11 * SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION
13 * OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN
14 * CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
15 */
16
17#ifndef _WCN36XX_H_
18#define _WCN36XX_H_
19
20#include <linux/completion.h>
21#include <linux/printk.h>
22#include <linux/spinlock.h>
23#include <net/mac80211.h>
24
25#include "hal.h"
26#include "smd.h"
27#include "txrx.h"
28#include "dxe.h"
29#include "pmc.h"
30#include "debug.h"
31
32#define WLAN_NV_FILE "wlan/prima/WCNSS_qcom_wlan_nv.bin"
33#define WCN36XX_AGGR_BUFFER_SIZE 64
34
35extern unsigned int wcn36xx_dbg_mask;
36
37enum wcn36xx_debug_mask {
38 WCN36XX_DBG_DXE = 0x00000001,
39 WCN36XX_DBG_DXE_DUMP = 0x00000002,
40 WCN36XX_DBG_SMD = 0x00000004,
41 WCN36XX_DBG_SMD_DUMP = 0x00000008,
42 WCN36XX_DBG_RX = 0x00000010,
43 WCN36XX_DBG_RX_DUMP = 0x00000020,
44 WCN36XX_DBG_TX = 0x00000040,
45 WCN36XX_DBG_TX_DUMP = 0x00000080,
46 WCN36XX_DBG_HAL = 0x00000100,
47 WCN36XX_DBG_HAL_DUMP = 0x00000200,
48 WCN36XX_DBG_MAC = 0x00000400,
49 WCN36XX_DBG_BEACON = 0x00000800,
50 WCN36XX_DBG_BEACON_DUMP = 0x00001000,
51 WCN36XX_DBG_PMC = 0x00002000,
52 WCN36XX_DBG_PMC_DUMP = 0x00004000,
53 WCN36XX_DBG_ANY = 0xffffffff,
54};
55
56#define wcn36xx_err(fmt, arg...) \
57 printk(KERN_ERR pr_fmt("ERROR " fmt), ##arg);
58
59#define wcn36xx_warn(fmt, arg...) \
60 printk(KERN_WARNING pr_fmt("WARNING " fmt), ##arg)
61
62#define wcn36xx_info(fmt, arg...) \
63 printk(KERN_INFO pr_fmt(fmt), ##arg)
64
65#define wcn36xx_dbg(mask, fmt, arg...) do { \
66 if (wcn36xx_dbg_mask & mask) \
67 printk(KERN_DEBUG pr_fmt(fmt), ##arg); \
68} while (0)
69
70#define wcn36xx_dbg_dump(mask, prefix_str, buf, len) do { \
71 if (wcn36xx_dbg_mask & mask) \
72 print_hex_dump(KERN_DEBUG, pr_fmt(prefix_str), \
73 DUMP_PREFIX_OFFSET, 32, 1, \
74 buf, len, false); \
75} while (0)
76
77#define WCN36XX_HW_CHANNEL(__wcn) (__wcn->hw->conf.chandef.chan->hw_value)
78#define WCN36XX_BAND(__wcn) (__wcn->hw->conf.chandef.chan->band)
79#define WCN36XX_CENTER_FREQ(__wcn) (__wcn->hw->conf.chandef.chan->center_freq)
80#define WCN36XX_LISTEN_INTERVAL(__wcn) (__wcn->hw->conf.listen_interval)
81#define WCN36XX_FLAGS(__wcn) (__wcn->hw->flags)
82#define WCN36XX_MAX_POWER(__wcn) (__wcn->hw->conf.chandef.chan->max_power)
83
84static inline void buff_to_be(u32 *buf, size_t len)
85{
86 int i;
87 for (i = 0; i < len; i++)
88 buf[i] = cpu_to_be32(buf[i]);
89}
90
91struct nv_data {
92 int is_valid;
93 u8 table;
94};
95
96/* Interface for platform control path
97 *
98 * @open: hook must be called when wcn36xx wants to open control channel.
99 * @tx: sends a buffer.
100 */
101struct wcn36xx_platform_ctrl_ops {
102 int (*open)(void *drv_priv, void *rsp_cb);
103 void (*close)(void);
104 int (*tx)(char *buf, size_t len);
105 int (*get_hw_mac)(u8 *addr);
106 int (*smsm_change_state)(u32 clear_mask, u32 set_mask);
107};
108
109/**
110 * struct wcn36xx_vif - holds VIF related fields
111 *
112 * @bss_index: bss_index is initially set to 0xFF. bss_index is received from
113 * HW after first config_bss call and must be used in delete_bss and
114 * enter/exit_bmps.
115 */
116struct wcn36xx_vif {
117 struct list_head list;
118 struct wcn36xx_sta *sta;
119 u8 dtim_period;
120 enum ani_ed_type encrypt_type;
121 bool is_joining;
122 struct wcn36xx_hal_mac_ssid ssid;
123
124 /* Power management */
125 enum wcn36xx_power_state pw_state;
126
127 u8 bss_index;
128 u8 ucast_dpu_signature;
129 /* Returned from WCN36XX_HAL_ADD_STA_SELF_RSP */
130 u8 self_sta_index;
131 u8 self_dpu_desc_index;
132};
133
134/**
135 * struct wcn36xx_sta - holds STA related fields
136 *
137 * @tid: traffic ID that is used during AMPDU and in TX BD.
138 * @sta_index: STA index is returned from HW after config_sta call and is
139 * used in both SMD channel and TX BD.
140 * @dpu_desc_index: DPU descriptor index is returned from HW after config_sta
141 * call and is used in TX BD.
142 * @bss_sta_index: STA index is returned from HW after config_bss call and is
143 * used in both SMD channel and TX BD. See table bellow when it is used.
144 * @bss_dpu_desc_index: DPU descriptor index is returned from HW after
145 * config_bss call and is used in TX BD.
146 * ______________________________________________
147 * | | STA | AP |
148 * |______________|_____________|_______________|
149 * | TX BD |bss_sta_index| sta_index |
150 * |______________|_____________|_______________|
151 * |all SMD calls |bss_sta_index| sta_index |
152 * |______________|_____________|_______________|
153 * |smd_delete_sta| sta_index | sta_index |
154 * |______________|_____________|_______________|
155 */
156struct wcn36xx_sta {
157 struct wcn36xx_vif *vif;
158 u16 aid;
159 u16 tid;
160 u8 sta_index;
161 u8 dpu_desc_index;
162 u8 bss_sta_index;
163 u8 bss_dpu_desc_index;
164 bool is_data_encrypted;
165 /* Rates */
166 struct wcn36xx_hal_supported_rates supported_rates;
167};
168struct wcn36xx_dxe_ch;
169struct wcn36xx {
170 struct ieee80211_hw *hw;
171 struct device *dev;
172 struct list_head vif_list;
173
174 u8 fw_revision;
175 u8 fw_version;
176 u8 fw_minor;
177 u8 fw_major;
178
179 /* extra byte for the NULL termination */
180 u8 crm_version[WCN36XX_HAL_VERSION_LENGTH + 1];
181 u8 wlan_version[WCN36XX_HAL_VERSION_LENGTH + 1];
182
183 /* IRQs */
184 int tx_irq;
185 int rx_irq;
186 void __iomem *mmio;
187
188 struct wcn36xx_platform_ctrl_ops *ctrl_ops;
189 /*
190 * smd_buf must be protected with smd_mutex to garantee
191 * that all messages are sent one after another
192 */
193 u8 *hal_buf;
194 size_t hal_rsp_len;
195 struct mutex hal_mutex;
196 struct completion hal_rsp_compl;
197 struct workqueue_struct *hal_ind_wq;
198 struct work_struct hal_ind_work;
199 struct mutex hal_ind_mutex;
200 struct list_head hal_ind_queue;
201
202 /* DXE channels */
203 struct wcn36xx_dxe_ch dxe_tx_l_ch; /* TX low */
204 struct wcn36xx_dxe_ch dxe_tx_h_ch; /* TX high */
205 struct wcn36xx_dxe_ch dxe_rx_l_ch; /* RX low */
206 struct wcn36xx_dxe_ch dxe_rx_h_ch; /* RX high */
207
208 /* For synchronization of DXE resources from BH, IRQ and WQ contexts */
209 spinlock_t dxe_lock;
210 bool queues_stopped;
211
212 /* Memory pools */
213 struct wcn36xx_dxe_mem_pool mgmt_mem_pool;
214 struct wcn36xx_dxe_mem_pool data_mem_pool;
215
216 struct sk_buff *tx_ack_skb;
217
218#ifdef CONFIG_WCN36XX_DEBUGFS
219 /* Debug file system entry */
220 struct wcn36xx_dfs_entry dfs;
221#endif /* CONFIG_WCN36XX_DEBUGFS */
222
223};
224
225static inline bool wcn36xx_is_fw_version(struct wcn36xx *wcn,
226 u8 major,
227 u8 minor,
228 u8 version,
229 u8 revision)
230{
231 return (wcn->fw_major == major &&
232 wcn->fw_minor == minor &&
233 wcn->fw_version == version &&
234 wcn->fw_revision == revision);
235}
236void wcn36xx_set_default_rates(struct wcn36xx_hal_supported_rates *rates);
237
238#endif /* _WCN36XX_H_ */
diff --git a/drivers/net/wireless/ath/wil6210/cfg80211.c b/drivers/net/wireless/ath/wil6210/cfg80211.c
index 61c302a6bdea..5b340769d5bb 100644
--- a/drivers/net/wireless/ath/wil6210/cfg80211.c
+++ b/drivers/net/wireless/ath/wil6210/cfg80211.c
@@ -316,8 +316,8 @@ static int wil_cfg80211_connect(struct wiphy *wiphy,
316 } 316 }
317 conn.channel = ch - 1; 317 conn.channel = ch - 1;
318 318
319 memcpy(conn.bssid, bss->bssid, 6); 319 memcpy(conn.bssid, bss->bssid, ETH_ALEN);
320 memcpy(conn.dst_mac, bss->bssid, 6); 320 memcpy(conn.dst_mac, bss->bssid, ETH_ALEN);
321 /* 321 /*
322 * FW don't support scan after connection attempt 322 * FW don't support scan after connection attempt
323 */ 323 */
diff --git a/drivers/net/wireless/ath/wil6210/pcie_bus.c b/drivers/net/wireless/ath/wil6210/pcie_bus.c
index eb1dc7ad80fb..eeceab39cda2 100644
--- a/drivers/net/wireless/ath/wil6210/pcie_bus.c
+++ b/drivers/net/wireless/ath/wil6210/pcie_bus.c
@@ -197,7 +197,6 @@ static void wil_pcie_remove(struct pci_dev *pdev)
197 pci_iounmap(pdev, wil->csr); 197 pci_iounmap(pdev, wil->csr);
198 pci_release_region(pdev, 0); 198 pci_release_region(pdev, 0);
199 pci_disable_device(pdev); 199 pci_disable_device(pdev);
200 pci_set_drvdata(pdev, NULL);
201} 200}
202 201
203static DEFINE_PCI_DEVICE_TABLE(wil6210_pcie_ids) = { 202static DEFINE_PCI_DEVICE_TABLE(wil6210_pcie_ids) = {
diff --git a/drivers/net/wireless/atmel.c b/drivers/net/wireless/atmel.c
index b827d51c30a3..0d950f209dae 100644
--- a/drivers/net/wireless/atmel.c
+++ b/drivers/net/wireless/atmel.c
@@ -844,18 +844,18 @@ static netdev_tx_t start_tx(struct sk_buff *skb, struct net_device *dev)
844 if (priv->wep_is_on) 844 if (priv->wep_is_on)
845 frame_ctl |= IEEE80211_FCTL_PROTECTED; 845 frame_ctl |= IEEE80211_FCTL_PROTECTED;
846 if (priv->operating_mode == IW_MODE_ADHOC) { 846 if (priv->operating_mode == IW_MODE_ADHOC) {
847 skb_copy_from_linear_data(skb, &header.addr1, 6); 847 skb_copy_from_linear_data(skb, &header.addr1, ETH_ALEN);
848 memcpy(&header.addr2, dev->dev_addr, 6); 848 memcpy(&header.addr2, dev->dev_addr, ETH_ALEN);
849 memcpy(&header.addr3, priv->BSSID, 6); 849 memcpy(&header.addr3, priv->BSSID, ETH_ALEN);
850 } else { 850 } else {
851 frame_ctl |= IEEE80211_FCTL_TODS; 851 frame_ctl |= IEEE80211_FCTL_TODS;
852 memcpy(&header.addr1, priv->CurrentBSSID, 6); 852 memcpy(&header.addr1, priv->CurrentBSSID, ETH_ALEN);
853 memcpy(&header.addr2, dev->dev_addr, 6); 853 memcpy(&header.addr2, dev->dev_addr, ETH_ALEN);
854 skb_copy_from_linear_data(skb, &header.addr3, 6); 854 skb_copy_from_linear_data(skb, &header.addr3, ETH_ALEN);
855 } 855 }
856 856
857 if (priv->use_wpa) 857 if (priv->use_wpa)
858 memcpy(&header.addr4, SNAP_RFC1024, 6); 858 memcpy(&header.addr4, SNAP_RFC1024, ETH_ALEN);
859 859
860 header.frame_control = cpu_to_le16(frame_ctl); 860 header.frame_control = cpu_to_le16(frame_ctl);
861 /* Copy the wireless header into the card */ 861 /* Copy the wireless header into the card */
@@ -929,11 +929,11 @@ static void fast_rx_path(struct atmel_private *priv,
929 } 929 }
930 } 930 }
931 931
932 memcpy(skbp, header->addr1, 6); /* destination address */ 932 memcpy(skbp, header->addr1, ETH_ALEN); /* destination address */
933 if (le16_to_cpu(header->frame_control) & IEEE80211_FCTL_FROMDS) 933 if (le16_to_cpu(header->frame_control) & IEEE80211_FCTL_FROMDS)
934 memcpy(&skbp[6], header->addr3, 6); 934 memcpy(&skbp[ETH_ALEN], header->addr3, ETH_ALEN);
935 else 935 else
936 memcpy(&skbp[6], header->addr2, 6); /* source address */ 936 memcpy(&skbp[ETH_ALEN], header->addr2, ETH_ALEN); /* source address */
937 937
938 skb->protocol = eth_type_trans(skb, priv->dev); 938 skb->protocol = eth_type_trans(skb, priv->dev);
939 skb->ip_summed = CHECKSUM_NONE; 939 skb->ip_summed = CHECKSUM_NONE;
@@ -969,14 +969,14 @@ static void frag_rx_path(struct atmel_private *priv,
969 u16 msdu_size, u16 rx_packet_loc, u32 crc, u16 seq_no, 969 u16 msdu_size, u16 rx_packet_loc, u32 crc, u16 seq_no,
970 u8 frag_no, int more_frags) 970 u8 frag_no, int more_frags)
971{ 971{
972 u8 mac4[6]; 972 u8 mac4[ETH_ALEN];
973 u8 source[6]; 973 u8 source[ETH_ALEN];
974 struct sk_buff *skb; 974 struct sk_buff *skb;
975 975
976 if (le16_to_cpu(header->frame_control) & IEEE80211_FCTL_FROMDS) 976 if (le16_to_cpu(header->frame_control) & IEEE80211_FCTL_FROMDS)
977 memcpy(source, header->addr3, 6); 977 memcpy(source, header->addr3, ETH_ALEN);
978 else 978 else
979 memcpy(source, header->addr2, 6); 979 memcpy(source, header->addr2, ETH_ALEN);
980 980
981 rx_packet_loc += 24; /* skip header */ 981 rx_packet_loc += 24; /* skip header */
982 982
@@ -984,9 +984,9 @@ static void frag_rx_path(struct atmel_private *priv,
984 msdu_size -= 4; 984 msdu_size -= 4;
985 985
986 if (frag_no == 0) { /* first fragment */ 986 if (frag_no == 0) { /* first fragment */
987 atmel_copy_to_host(priv->dev, mac4, rx_packet_loc, 6); 987 atmel_copy_to_host(priv->dev, mac4, rx_packet_loc, ETH_ALEN);
988 msdu_size -= 6; 988 msdu_size -= ETH_ALEN;
989 rx_packet_loc += 6; 989 rx_packet_loc += ETH_ALEN;
990 990
991 if (priv->do_rx_crc) 991 if (priv->do_rx_crc)
992 crc = crc32_le(crc, mac4, 6); 992 crc = crc32_le(crc, mac4, 6);
@@ -994,9 +994,9 @@ static void frag_rx_path(struct atmel_private *priv,
994 priv->frag_seq = seq_no; 994 priv->frag_seq = seq_no;
995 priv->frag_no = 1; 995 priv->frag_no = 1;
996 priv->frag_len = msdu_size; 996 priv->frag_len = msdu_size;
997 memcpy(priv->frag_source, source, 6); 997 memcpy(priv->frag_source, source, ETH_ALEN);
998 memcpy(&priv->rx_buf[6], source, 6); 998 memcpy(&priv->rx_buf[ETH_ALEN], source, ETH_ALEN);
999 memcpy(priv->rx_buf, header->addr1, 6); 999 memcpy(priv->rx_buf, header->addr1, ETH_ALEN);
1000 1000
1001 atmel_copy_to_host(priv->dev, &priv->rx_buf[12], rx_packet_loc, msdu_size); 1001 atmel_copy_to_host(priv->dev, &priv->rx_buf[12], rx_packet_loc, msdu_size);
1002 1002
@@ -1006,13 +1006,13 @@ static void frag_rx_path(struct atmel_private *priv,
1006 atmel_copy_to_host(priv->dev, (void *)&netcrc, rx_packet_loc + msdu_size, 4); 1006 atmel_copy_to_host(priv->dev, (void *)&netcrc, rx_packet_loc + msdu_size, 4);
1007 if ((crc ^ 0xffffffff) != netcrc) { 1007 if ((crc ^ 0xffffffff) != netcrc) {
1008 priv->dev->stats.rx_crc_errors++; 1008 priv->dev->stats.rx_crc_errors++;
1009 memset(priv->frag_source, 0xff, 6); 1009 memset(priv->frag_source, 0xff, ETH_ALEN);
1010 } 1010 }
1011 } 1011 }
1012 1012
1013 } else if (priv->frag_no == frag_no && 1013 } else if (priv->frag_no == frag_no &&
1014 priv->frag_seq == seq_no && 1014 priv->frag_seq == seq_no &&
1015 memcmp(priv->frag_source, source, 6) == 0) { 1015 memcmp(priv->frag_source, source, ETH_ALEN) == 0) {
1016 1016
1017 atmel_copy_to_host(priv->dev, &priv->rx_buf[12 + priv->frag_len], 1017 atmel_copy_to_host(priv->dev, &priv->rx_buf[12 + priv->frag_len],
1018 rx_packet_loc, msdu_size); 1018 rx_packet_loc, msdu_size);
@@ -1024,7 +1024,7 @@ static void frag_rx_path(struct atmel_private *priv,
1024 atmel_copy_to_host(priv->dev, (void *)&netcrc, rx_packet_loc + msdu_size, 4); 1024 atmel_copy_to_host(priv->dev, (void *)&netcrc, rx_packet_loc + msdu_size, 4);
1025 if ((crc ^ 0xffffffff) != netcrc) { 1025 if ((crc ^ 0xffffffff) != netcrc) {
1026 priv->dev->stats.rx_crc_errors++; 1026 priv->dev->stats.rx_crc_errors++;
1027 memset(priv->frag_source, 0xff, 6); 1027 memset(priv->frag_source, 0xff, ETH_ALEN);
1028 more_frags = 1; /* don't send broken assembly */ 1028 more_frags = 1; /* don't send broken assembly */
1029 } 1029 }
1030 } 1030 }
@@ -1033,7 +1033,7 @@ static void frag_rx_path(struct atmel_private *priv,
1033 priv->frag_no++; 1033 priv->frag_no++;
1034 1034
1035 if (!more_frags) { /* last one */ 1035 if (!more_frags) { /* last one */
1036 memset(priv->frag_source, 0xff, 6); 1036 memset(priv->frag_source, 0xff, ETH_ALEN);
1037 if (!(skb = dev_alloc_skb(priv->frag_len + 14))) { 1037 if (!(skb = dev_alloc_skb(priv->frag_len + 14))) {
1038 priv->dev->stats.rx_dropped++; 1038 priv->dev->stats.rx_dropped++;
1039 } else { 1039 } else {
@@ -1129,7 +1129,7 @@ static void rx_done_irq(struct atmel_private *priv)
1129 atmel_copy_to_host(priv->dev, (unsigned char *)&priv->rx_buf, rx_packet_loc + 24, msdu_size); 1129 atmel_copy_to_host(priv->dev, (unsigned char *)&priv->rx_buf, rx_packet_loc + 24, msdu_size);
1130 1130
1131 /* we use the same buffer for frag reassembly and control packets */ 1131 /* we use the same buffer for frag reassembly and control packets */
1132 memset(priv->frag_source, 0xff, 6); 1132 memset(priv->frag_source, 0xff, ETH_ALEN);
1133 1133
1134 if (priv->do_rx_crc) { 1134 if (priv->do_rx_crc) {
1135 /* last 4 octets is crc */ 1135 /* last 4 octets is crc */
@@ -1557,7 +1557,7 @@ struct net_device *init_atmel_card(unsigned short irq, unsigned long port,
1557 priv->last_qual = jiffies; 1557 priv->last_qual = jiffies;
1558 priv->last_beacon_timestamp = 0; 1558 priv->last_beacon_timestamp = 0;
1559 memset(priv->frag_source, 0xff, sizeof(priv->frag_source)); 1559 memset(priv->frag_source, 0xff, sizeof(priv->frag_source));
1560 memset(priv->BSSID, 0, 6); 1560 memset(priv->BSSID, 0, ETH_ALEN);
1561 priv->CurrentBSSID[0] = 0xFF; /* Initialize to something invalid.... */ 1561 priv->CurrentBSSID[0] = 0xFF; /* Initialize to something invalid.... */
1562 priv->station_was_associated = 0; 1562 priv->station_was_associated = 0;
1563 1563
@@ -1718,7 +1718,7 @@ static int atmel_get_wap(struct net_device *dev,
1718 char *extra) 1718 char *extra)
1719{ 1719{
1720 struct atmel_private *priv = netdev_priv(dev); 1720 struct atmel_private *priv = netdev_priv(dev);
1721 memcpy(awrq->sa_data, priv->CurrentBSSID, 6); 1721 memcpy(awrq->sa_data, priv->CurrentBSSID, ETH_ALEN);
1722 awrq->sa_family = ARPHRD_ETHER; 1722 awrq->sa_family = ARPHRD_ETHER;
1723 1723
1724 return 0; 1724 return 0;
@@ -2356,7 +2356,7 @@ static int atmel_get_scan(struct net_device *dev,
2356 for (i = 0; i < priv->BSS_list_entries; i++) { 2356 for (i = 0; i < priv->BSS_list_entries; i++) {
2357 iwe.cmd = SIOCGIWAP; 2357 iwe.cmd = SIOCGIWAP;
2358 iwe.u.ap_addr.sa_family = ARPHRD_ETHER; 2358 iwe.u.ap_addr.sa_family = ARPHRD_ETHER;
2359 memcpy(iwe.u.ap_addr.sa_data, priv->BSSinfo[i].BSSID, 6); 2359 memcpy(iwe.u.ap_addr.sa_data, priv->BSSinfo[i].BSSID, ETH_ALEN);
2360 current_ev = iwe_stream_add_event(info, current_ev, 2360 current_ev = iwe_stream_add_event(info, current_ev,
2361 extra + IW_SCAN_MAX_DATA, 2361 extra + IW_SCAN_MAX_DATA,
2362 &iwe, IW_EV_ADDR_LEN); 2362 &iwe, IW_EV_ADDR_LEN);
@@ -2760,7 +2760,7 @@ static void atmel_enter_state(struct atmel_private *priv, int new_state)
2760static void atmel_scan(struct atmel_private *priv, int specific_ssid) 2760static void atmel_scan(struct atmel_private *priv, int specific_ssid)
2761{ 2761{
2762 struct { 2762 struct {
2763 u8 BSSID[6]; 2763 u8 BSSID[ETH_ALEN];
2764 u8 SSID[MAX_SSID_LENGTH]; 2764 u8 SSID[MAX_SSID_LENGTH];
2765 u8 scan_type; 2765 u8 scan_type;
2766 u8 channel; 2766 u8 channel;
@@ -2771,7 +2771,7 @@ static void atmel_scan(struct atmel_private *priv, int specific_ssid)
2771 u8 SSID_size; 2771 u8 SSID_size;
2772 } cmd; 2772 } cmd;
2773 2773
2774 memset(cmd.BSSID, 0xff, 6); 2774 memset(cmd.BSSID, 0xff, ETH_ALEN);
2775 2775
2776 if (priv->fast_scan) { 2776 if (priv->fast_scan) {
2777 cmd.SSID_size = priv->SSID_size; 2777 cmd.SSID_size = priv->SSID_size;
@@ -2816,7 +2816,7 @@ static void join(struct atmel_private *priv, int type)
2816 2816
2817 cmd.SSID_size = priv->SSID_size; 2817 cmd.SSID_size = priv->SSID_size;
2818 memcpy(cmd.SSID, priv->SSID, priv->SSID_size); 2818 memcpy(cmd.SSID, priv->SSID, priv->SSID_size);
2819 memcpy(cmd.BSSID, priv->CurrentBSSID, 6); 2819 memcpy(cmd.BSSID, priv->CurrentBSSID, ETH_ALEN);
2820 cmd.channel = (priv->channel & 0x7f); 2820 cmd.channel = (priv->channel & 0x7f);
2821 cmd.BSS_type = type; 2821 cmd.BSS_type = type;
2822 cmd.timeout = cpu_to_le16(2000); 2822 cmd.timeout = cpu_to_le16(2000);
@@ -2837,7 +2837,7 @@ static void start(struct atmel_private *priv, int type)
2837 2837
2838 cmd.SSID_size = priv->SSID_size; 2838 cmd.SSID_size = priv->SSID_size;
2839 memcpy(cmd.SSID, priv->SSID, priv->SSID_size); 2839 memcpy(cmd.SSID, priv->SSID, priv->SSID_size);
2840 memcpy(cmd.BSSID, priv->BSSID, 6); 2840 memcpy(cmd.BSSID, priv->BSSID, ETH_ALEN);
2841 cmd.BSS_type = type; 2841 cmd.BSS_type = type;
2842 cmd.channel = (priv->channel & 0x7f); 2842 cmd.channel = (priv->channel & 0x7f);
2843 2843
@@ -2883,9 +2883,9 @@ static void send_authentication_request(struct atmel_private *priv, u16 system,
2883 header.frame_control = cpu_to_le16(IEEE80211_FTYPE_MGMT | IEEE80211_STYPE_AUTH); 2883 header.frame_control = cpu_to_le16(IEEE80211_FTYPE_MGMT | IEEE80211_STYPE_AUTH);
2884 header.duration_id = cpu_to_le16(0x8000); 2884 header.duration_id = cpu_to_le16(0x8000);
2885 header.seq_ctrl = 0; 2885 header.seq_ctrl = 0;
2886 memcpy(header.addr1, priv->CurrentBSSID, 6); 2886 memcpy(header.addr1, priv->CurrentBSSID, ETH_ALEN);
2887 memcpy(header.addr2, priv->dev->dev_addr, 6); 2887 memcpy(header.addr2, priv->dev->dev_addr, ETH_ALEN);
2888 memcpy(header.addr3, priv->CurrentBSSID, 6); 2888 memcpy(header.addr3, priv->CurrentBSSID, ETH_ALEN);
2889 2889
2890 if (priv->wep_is_on && priv->CurrentAuthentTransactionSeqNum != 1) 2890 if (priv->wep_is_on && priv->CurrentAuthentTransactionSeqNum != 1)
2891 /* no WEP for authentication frames with TrSeqNo 1 */ 2891 /* no WEP for authentication frames with TrSeqNo 1 */
@@ -2916,7 +2916,7 @@ static void send_association_request(struct atmel_private *priv, int is_reassoc)
2916 struct ass_req_format { 2916 struct ass_req_format {
2917 __le16 capability; 2917 __le16 capability;
2918 __le16 listen_interval; 2918 __le16 listen_interval;
2919 u8 ap[6]; /* nothing after here directly accessible */ 2919 u8 ap[ETH_ALEN]; /* nothing after here directly accessible */
2920 u8 ssid_el_id; 2920 u8 ssid_el_id;
2921 u8 ssid_len; 2921 u8 ssid_len;
2922 u8 ssid[MAX_SSID_LENGTH]; 2922 u8 ssid[MAX_SSID_LENGTH];
@@ -2930,9 +2930,9 @@ static void send_association_request(struct atmel_private *priv, int is_reassoc)
2930 header.duration_id = cpu_to_le16(0x8000); 2930 header.duration_id = cpu_to_le16(0x8000);
2931 header.seq_ctrl = 0; 2931 header.seq_ctrl = 0;
2932 2932
2933 memcpy(header.addr1, priv->CurrentBSSID, 6); 2933 memcpy(header.addr1, priv->CurrentBSSID, ETH_ALEN);
2934 memcpy(header.addr2, priv->dev->dev_addr, 6); 2934 memcpy(header.addr2, priv->dev->dev_addr, ETH_ALEN);
2935 memcpy(header.addr3, priv->CurrentBSSID, 6); 2935 memcpy(header.addr3, priv->CurrentBSSID, ETH_ALEN);
2936 2936
2937 body.capability = cpu_to_le16(WLAN_CAPABILITY_ESS); 2937 body.capability = cpu_to_le16(WLAN_CAPABILITY_ESS);
2938 if (priv->wep_is_on) 2938 if (priv->wep_is_on)
@@ -2944,7 +2944,7 @@ static void send_association_request(struct atmel_private *priv, int is_reassoc)
2944 2944
2945 /* current AP address - only in reassoc frame */ 2945 /* current AP address - only in reassoc frame */
2946 if (is_reassoc) { 2946 if (is_reassoc) {
2947 memcpy(body.ap, priv->CurrentBSSID, 6); 2947 memcpy(body.ap, priv->CurrentBSSID, ETH_ALEN);
2948 ssid_el_p = &body.ssid_el_id; 2948 ssid_el_p = &body.ssid_el_id;
2949 bodysize = 18 + priv->SSID_size; 2949 bodysize = 18 + priv->SSID_size;
2950 } else { 2950 } else {
@@ -3021,7 +3021,7 @@ static void store_bss_info(struct atmel_private *priv,
3021 int i, index; 3021 int i, index;
3022 3022
3023 for (index = -1, i = 0; i < priv->BSS_list_entries; i++) 3023 for (index = -1, i = 0; i < priv->BSS_list_entries; i++)
3024 if (memcmp(bss, priv->BSSinfo[i].BSSID, 6) == 0) 3024 if (memcmp(bss, priv->BSSinfo[i].BSSID, ETH_ALEN) == 0)
3025 index = i; 3025 index = i;
3026 3026
3027 /* If we process a probe and an entry from this BSS exists 3027 /* If we process a probe and an entry from this BSS exists
@@ -3032,7 +3032,7 @@ static void store_bss_info(struct atmel_private *priv,
3032 if (priv->BSS_list_entries == MAX_BSS_ENTRIES) 3032 if (priv->BSS_list_entries == MAX_BSS_ENTRIES)
3033 return; 3033 return;
3034 index = priv->BSS_list_entries++; 3034 index = priv->BSS_list_entries++;
3035 memcpy(priv->BSSinfo[index].BSSID, bss, 6); 3035 memcpy(priv->BSSinfo[index].BSSID, bss, ETH_ALEN);
3036 priv->BSSinfo[index].RSSI = rssi; 3036 priv->BSSinfo[index].RSSI = rssi;
3037 } else { 3037 } else {
3038 if (rssi > priv->BSSinfo[index].RSSI) 3038 if (rssi > priv->BSSinfo[index].RSSI)
@@ -3212,7 +3212,7 @@ static void associate(struct atmel_private *priv, u16 frame_len, u16 subtype)
3212 if (subtype == IEEE80211_STYPE_REASSOC_RESP && 3212 if (subtype == IEEE80211_STYPE_REASSOC_RESP &&
3213 status != WLAN_STATUS_ASSOC_DENIED_RATES && 3213 status != WLAN_STATUS_ASSOC_DENIED_RATES &&
3214 status != WLAN_STATUS_CAPS_UNSUPPORTED && 3214 status != WLAN_STATUS_CAPS_UNSUPPORTED &&
3215 priv->AssociationRequestRetryCnt < MAX_ASSOCIATION_RETRIES) { 3215 priv->ReAssociationRequestRetryCnt < MAX_ASSOCIATION_RETRIES) {
3216 mod_timer(&priv->management_timer, jiffies + MGMT_JIFFIES); 3216 mod_timer(&priv->management_timer, jiffies + MGMT_JIFFIES);
3217 priv->ReAssociationRequestRetryCnt++; 3217 priv->ReAssociationRequestRetryCnt++;
3218 send_association_request(priv, 1); 3218 send_association_request(priv, 1);
@@ -3235,7 +3235,7 @@ static void atmel_join_bss(struct atmel_private *priv, int bss_index)
3235{ 3235{
3236 struct bss_info *bss = &priv->BSSinfo[bss_index]; 3236 struct bss_info *bss = &priv->BSSinfo[bss_index];
3237 3237
3238 memcpy(priv->CurrentBSSID, bss->BSSID, 6); 3238 memcpy(priv->CurrentBSSID, bss->BSSID, ETH_ALEN);
3239 memcpy(priv->SSID, bss->SSID, priv->SSID_size = bss->SSIDsize); 3239 memcpy(priv->SSID, bss->SSID, priv->SSID_size = bss->SSIDsize);
3240 3240
3241 /* The WPA stuff cares about the current AP address */ 3241 /* The WPA stuff cares about the current AP address */
@@ -3767,7 +3767,7 @@ static int probe_atmel_card(struct net_device *dev)
3767 0x00, 0x04, 0x25, 0x00, 0x00, 0x00 3767 0x00, 0x04, 0x25, 0x00, 0x00, 0x00
3768 }; 3768 };
3769 printk(KERN_ALERT "%s: *** Invalid MAC address. UPGRADE Firmware ****\n", dev->name); 3769 printk(KERN_ALERT "%s: *** Invalid MAC address. UPGRADE Firmware ****\n", dev->name);
3770 memcpy(dev->dev_addr, default_mac, 6); 3770 memcpy(dev->dev_addr, default_mac, ETH_ALEN);
3771 } 3771 }
3772 } 3772 }
3773 3773
@@ -3819,7 +3819,7 @@ static void build_wpa_mib(struct atmel_private *priv)
3819 3819
3820 struct { /* NB this is matched to the hardware, don't change. */ 3820 struct { /* NB this is matched to the hardware, don't change. */
3821 u8 cipher_default_key_value[MAX_ENCRYPTION_KEYS][MAX_ENCRYPTION_KEY_SIZE]; 3821 u8 cipher_default_key_value[MAX_ENCRYPTION_KEYS][MAX_ENCRYPTION_KEY_SIZE];
3822 u8 receiver_address[6]; 3822 u8 receiver_address[ETH_ALEN];
3823 u8 wep_is_on; 3823 u8 wep_is_on;
3824 u8 default_key; /* 0..3 */ 3824 u8 default_key; /* 0..3 */
3825 u8 group_key; 3825 u8 group_key;
@@ -3837,7 +3837,7 @@ static void build_wpa_mib(struct atmel_private *priv)
3837 3837
3838 mib.wep_is_on = priv->wep_is_on; 3838 mib.wep_is_on = priv->wep_is_on;
3839 mib.exclude_unencrypted = priv->exclude_unencrypted; 3839 mib.exclude_unencrypted = priv->exclude_unencrypted;
3840 memcpy(mib.receiver_address, priv->CurrentBSSID, 6); 3840 memcpy(mib.receiver_address, priv->CurrentBSSID, ETH_ALEN);
3841 3841
3842 /* zero all the keys before adding in valid ones. */ 3842 /* zero all the keys before adding in valid ones. */
3843 memset(mib.cipher_default_key_value, 0, sizeof(mib.cipher_default_key_value)); 3843 memset(mib.cipher_default_key_value, 0, sizeof(mib.cipher_default_key_value));
diff --git a/drivers/net/wireless/b43/phy_n.c b/drivers/net/wireless/b43/phy_n.c
index 7c970d3ae358..05ee7f10cc8f 100644
--- a/drivers/net/wireless/b43/phy_n.c
+++ b/drivers/net/wireless/b43/phy_n.c
@@ -164,7 +164,8 @@ static void b43_nphy_rf_ctl_override_rev7(struct b43_wldev *dev, u16 field,
164 } 164 }
165 en_addr = en_addrs[override][i]; 165 en_addr = en_addrs[override][i];
166 166
167 val_addr = (i == 0) ? e->val_addr_core0 : e->val_addr_core1; 167 if (e)
168 val_addr = (i == 0) ? e->val_addr_core0 : e->val_addr_core1;
168 169
169 if (off) { 170 if (off) {
170 b43_phy_mask(dev, en_addr, ~en_mask); 171 b43_phy_mask(dev, en_addr, ~en_mask);
diff --git a/drivers/net/wireless/b43/xmit.c b/drivers/net/wireless/b43/xmit.c
index 8cb206a89083..4ae63f4ddfb2 100644
--- a/drivers/net/wireless/b43/xmit.c
+++ b/drivers/net/wireless/b43/xmit.c
@@ -278,7 +278,7 @@ int b43_generate_txhdr(struct b43_wldev *dev,
278 else 278 else
279 txhdr->phy_rate = b43_plcp_get_ratecode_cck(rate); 279 txhdr->phy_rate = b43_plcp_get_ratecode_cck(rate);
280 txhdr->mac_frame_ctl = wlhdr->frame_control; 280 txhdr->mac_frame_ctl = wlhdr->frame_control;
281 memcpy(txhdr->tx_receiver, wlhdr->addr1, 6); 281 memcpy(txhdr->tx_receiver, wlhdr->addr1, ETH_ALEN);
282 282
283 /* Calculate duration for fallback rate */ 283 /* Calculate duration for fallback rate */
284 if ((rate_fb == rate) || 284 if ((rate_fb == rate) ||
diff --git a/drivers/net/wireless/b43legacy/xmit.c b/drivers/net/wireless/b43legacy/xmit.c
index 849a28c80302..86588c9ff0f2 100644
--- a/drivers/net/wireless/b43legacy/xmit.c
+++ b/drivers/net/wireless/b43legacy/xmit.c
@@ -215,7 +215,7 @@ static int generate_txhdr_fw3(struct b43legacy_wldev *dev,
215 rate_fb_ofdm = b43legacy_is_ofdm_rate(rate_fb->hw_value); 215 rate_fb_ofdm = b43legacy_is_ofdm_rate(rate_fb->hw_value);
216 216
217 txhdr->mac_frame_ctl = wlhdr->frame_control; 217 txhdr->mac_frame_ctl = wlhdr->frame_control;
218 memcpy(txhdr->tx_receiver, wlhdr->addr1, 6); 218 memcpy(txhdr->tx_receiver, wlhdr->addr1, ETH_ALEN);
219 219
220 /* Calculate duration for fallback rate */ 220 /* Calculate duration for fallback rate */
221 if ((rate_fb->hw_value == rate) || 221 if ((rate_fb->hw_value == rate) ||
diff --git a/drivers/net/wireless/brcm80211/brcmfmac/bcmsdh.c b/drivers/net/wireless/brcm80211/brcmfmac/bcmsdh.c
index e13b1a65c65f..3e10b801eee8 100644
--- a/drivers/net/wireless/brcm80211/brcmfmac/bcmsdh.c
+++ b/drivers/net/wireless/brcm80211/brcmfmac/bcmsdh.c
@@ -26,7 +26,6 @@
26#include <linux/mmc/sdio.h> 26#include <linux/mmc/sdio.h>
27#include <linux/mmc/sdio_func.h> 27#include <linux/mmc/sdio_func.h>
28#include <linux/mmc/card.h> 28#include <linux/mmc/card.h>
29#include <linux/mmc/host.h>
30#include <linux/platform_data/brcmfmac-sdio.h> 29#include <linux/platform_data/brcmfmac-sdio.h>
31 30
32#include <defs.h> 31#include <defs.h>
@@ -239,7 +238,9 @@ brcmf_sdio_regrw_helper(struct brcmf_sdio_dev *sdiodev, u32 addr,
239 func_num = SDIO_FUNC_1; 238 func_num = SDIO_FUNC_1;
240 reg_size = 4; 239 reg_size = 4;
241 240
242 brcmf_sdio_addrprep(sdiodev, reg_size, &addr); 241 ret = brcmf_sdio_addrprep(sdiodev, reg_size, &addr);
242 if (ret)
243 goto done;
243 } 244 }
244 245
245 do { 246 do {
@@ -255,6 +256,7 @@ brcmf_sdio_regrw_helper(struct brcmf_sdio_dev *sdiodev, u32 addr,
255 func_num, addr, data, 4); 256 func_num, addr, data, 4);
256 } while (ret != 0 && retry++ < SDIOH_API_ACCESS_RETRY_LIMIT); 257 } while (ret != 0 && retry++ < SDIOH_API_ACCESS_RETRY_LIMIT);
257 258
259done:
258 if (ret != 0) 260 if (ret != 0)
259 brcmf_err("failed with %d\n", ret); 261 brcmf_err("failed with %d\n", ret);
260 262
@@ -315,8 +317,36 @@ void brcmf_sdio_regwl(struct brcmf_sdio_dev *sdiodev, u32 addr,
315 *ret = retval; 317 *ret = retval;
316} 318}
317 319
320static int brcmf_sdio_buffrw(struct brcmf_sdio_dev *sdiodev, uint fn,
321 bool write, u32 addr, struct sk_buff *pkt)
322{
323 unsigned int req_sz;
324
325 brcmf_pm_resume_wait(sdiodev, &sdiodev->request_buffer_wait);
326 if (brcmf_pm_resume_error(sdiodev))
327 return -EIO;
328
329 /* Single skb use the standard mmc interface */
330 req_sz = pkt->len + 3;
331 req_sz &= (uint)~3;
332
333 if (write)
334 return sdio_memcpy_toio(sdiodev->func[fn], addr,
335 ((u8 *)(pkt->data)),
336 req_sz);
337 else if (fn == 1)
338 return sdio_memcpy_fromio(sdiodev->func[fn],
339 ((u8 *)(pkt->data)),
340 addr, req_sz);
341 else
342 /* function 2 read is FIFO operation */
343 return sdio_readsb(sdiodev->func[fn],
344 ((u8 *)(pkt->data)), addr,
345 req_sz);
346}
347
318/** 348/**
319 * brcmf_sdio_buffrw - SDIO interface function for block data access 349 * brcmf_sdio_sglist_rw - SDIO interface function for block data access
320 * @sdiodev: brcmfmac sdio device 350 * @sdiodev: brcmfmac sdio device
321 * @fn: SDIO function number 351 * @fn: SDIO function number
322 * @write: direction flag 352 * @write: direction flag
@@ -327,12 +357,13 @@ void brcmf_sdio_regwl(struct brcmf_sdio_dev *sdiodev, u32 addr,
327 * stack for block data access. It assumes that the skb passed down by the 357 * stack for block data access. It assumes that the skb passed down by the
328 * caller has already been padded and aligned. 358 * caller has already been padded and aligned.
329 */ 359 */
330static int brcmf_sdio_buffrw(struct brcmf_sdio_dev *sdiodev, uint fn, 360static int brcmf_sdio_sglist_rw(struct brcmf_sdio_dev *sdiodev, uint fn,
331 bool write, u32 addr, struct sk_buff_head *pktlist) 361 bool write, u32 addr,
362 struct sk_buff_head *pktlist)
332{ 363{
333 unsigned int req_sz, func_blk_sz, sg_cnt, sg_data_sz, pkt_offset; 364 unsigned int req_sz, func_blk_sz, sg_cnt, sg_data_sz, pkt_offset;
334 unsigned int max_blks, max_req_sz, orig_offset, dst_offset; 365 unsigned int max_req_sz, orig_offset, dst_offset;
335 unsigned short max_seg_sz, seg_sz; 366 unsigned short max_seg_cnt, seg_sz;
336 unsigned char *pkt_data, *orig_data, *dst_data; 367 unsigned char *pkt_data, *orig_data, *dst_data;
337 struct sk_buff *pkt_next = NULL, *local_pkt_next; 368 struct sk_buff *pkt_next = NULL, *local_pkt_next;
338 struct sk_buff_head local_list, *target_list; 369 struct sk_buff_head local_list, *target_list;
@@ -341,7 +372,6 @@ static int brcmf_sdio_buffrw(struct brcmf_sdio_dev *sdiodev, uint fn,
341 struct mmc_data mmc_dat; 372 struct mmc_data mmc_dat;
342 struct sg_table st; 373 struct sg_table st;
343 struct scatterlist *sgl; 374 struct scatterlist *sgl;
344 struct mmc_host *host;
345 int ret = 0; 375 int ret = 0;
346 376
347 if (!pktlist->qlen) 377 if (!pktlist->qlen)
@@ -351,27 +381,6 @@ static int brcmf_sdio_buffrw(struct brcmf_sdio_dev *sdiodev, uint fn,
351 if (brcmf_pm_resume_error(sdiodev)) 381 if (brcmf_pm_resume_error(sdiodev))
352 return -EIO; 382 return -EIO;
353 383
354 /* Single skb use the standard mmc interface */
355 if (pktlist->qlen == 1) {
356 pkt_next = pktlist->next;
357 req_sz = pkt_next->len + 3;
358 req_sz &= (uint)~3;
359
360 if (write)
361 return sdio_memcpy_toio(sdiodev->func[fn], addr,
362 ((u8 *)(pkt_next->data)),
363 req_sz);
364 else if (fn == 1)
365 return sdio_memcpy_fromio(sdiodev->func[fn],
366 ((u8 *)(pkt_next->data)),
367 addr, req_sz);
368 else
369 /* function 2 read is FIFO operation */
370 return sdio_readsb(sdiodev->func[fn],
371 ((u8 *)(pkt_next->data)), addr,
372 req_sz);
373 }
374
375 target_list = pktlist; 384 target_list = pktlist;
376 /* for host with broken sg support, prepare a page aligned list */ 385 /* for host with broken sg support, prepare a page aligned list */
377 __skb_queue_head_init(&local_list); 386 __skb_queue_head_init(&local_list);
@@ -398,38 +407,46 @@ static int brcmf_sdio_buffrw(struct brcmf_sdio_dev *sdiodev, uint fn,
398 target_list = &local_list; 407 target_list = &local_list;
399 } 408 }
400 409
401 host = sdiodev->func[fn]->card->host;
402 func_blk_sz = sdiodev->func[fn]->cur_blksize; 410 func_blk_sz = sdiodev->func[fn]->cur_blksize;
403 /* Blocks per command is limited by host count, host transfer 411 max_req_sz = sdiodev->max_request_size;
404 * size and the maximum for IO_RW_EXTENDED of 511 blocks. 412 max_seg_cnt = min_t(unsigned short, sdiodev->max_segment_count,
405 */ 413 target_list->qlen);
406 max_blks = min_t(unsigned int, host->max_blk_count, 511u);
407 max_req_sz = min_t(unsigned int, host->max_req_size,
408 max_blks * func_blk_sz);
409 max_seg_sz = min_t(unsigned short, host->max_segs, SG_MAX_SINGLE_ALLOC);
410 max_seg_sz = min_t(unsigned short, max_seg_sz, target_list->qlen);
411 seg_sz = target_list->qlen; 414 seg_sz = target_list->qlen;
412 pkt_offset = 0; 415 pkt_offset = 0;
413 pkt_next = target_list->next; 416 pkt_next = target_list->next;
414 417
415 if (sg_alloc_table(&st, max_seg_sz, GFP_KERNEL)) { 418 if (sg_alloc_table(&st, max_seg_cnt, GFP_KERNEL)) {
416 ret = -ENOMEM; 419 ret = -ENOMEM;
417 goto exit; 420 goto exit;
418 } 421 }
419 422
423 memset(&mmc_req, 0, sizeof(struct mmc_request));
424 memset(&mmc_cmd, 0, sizeof(struct mmc_command));
425 memset(&mmc_dat, 0, sizeof(struct mmc_data));
426
427 mmc_dat.sg = st.sgl;
428 mmc_dat.blksz = func_blk_sz;
429 mmc_dat.flags = write ? MMC_DATA_WRITE : MMC_DATA_READ;
430 mmc_cmd.opcode = SD_IO_RW_EXTENDED;
431 mmc_cmd.arg = write ? 1<<31 : 0; /* write flag */
432 mmc_cmd.arg |= (fn & 0x7) << 28; /* SDIO func num */
433 mmc_cmd.arg |= 1<<27; /* block mode */
434 /* for function 1 the addr will be incremented */
435 mmc_cmd.arg |= (fn == 1) ? 1<<26 : 0;
436 mmc_cmd.flags = MMC_RSP_SPI_R5 | MMC_RSP_R5 | MMC_CMD_ADTC;
437 mmc_req.cmd = &mmc_cmd;
438 mmc_req.data = &mmc_dat;
439
420 while (seg_sz) { 440 while (seg_sz) {
421 req_sz = 0; 441 req_sz = 0;
422 sg_cnt = 0; 442 sg_cnt = 0;
423 memset(&mmc_req, 0, sizeof(struct mmc_request));
424 memset(&mmc_cmd, 0, sizeof(struct mmc_command));
425 memset(&mmc_dat, 0, sizeof(struct mmc_data));
426 sgl = st.sgl; 443 sgl = st.sgl;
427 /* prep sg table */ 444 /* prep sg table */
428 while (pkt_next != (struct sk_buff *)target_list) { 445 while (pkt_next != (struct sk_buff *)target_list) {
429 pkt_data = pkt_next->data + pkt_offset; 446 pkt_data = pkt_next->data + pkt_offset;
430 sg_data_sz = pkt_next->len - pkt_offset; 447 sg_data_sz = pkt_next->len - pkt_offset;
431 if (sg_data_sz > host->max_seg_size) 448 if (sg_data_sz > sdiodev->max_segment_size)
432 sg_data_sz = host->max_seg_size; 449 sg_data_sz = sdiodev->max_segment_size;
433 if (sg_data_sz > max_req_sz - req_sz) 450 if (sg_data_sz > max_req_sz - req_sz)
434 sg_data_sz = max_req_sz - req_sz; 451 sg_data_sz = max_req_sz - req_sz;
435 452
@@ -444,7 +461,7 @@ static int brcmf_sdio_buffrw(struct brcmf_sdio_dev *sdiodev, uint fn,
444 pkt_next = pkt_next->next; 461 pkt_next = pkt_next->next;
445 } 462 }
446 463
447 if (req_sz >= max_req_sz || sg_cnt >= max_seg_sz) 464 if (req_sz >= max_req_sz || sg_cnt >= max_seg_cnt)
448 break; 465 break;
449 } 466 }
450 seg_sz -= sg_cnt; 467 seg_sz -= sg_cnt;
@@ -455,27 +472,17 @@ static int brcmf_sdio_buffrw(struct brcmf_sdio_dev *sdiodev, uint fn,
455 ret = -ENOTBLK; 472 ret = -ENOTBLK;
456 goto exit; 473 goto exit;
457 } 474 }
458 mmc_dat.sg = st.sgl; 475
459 mmc_dat.sg_len = sg_cnt; 476 mmc_dat.sg_len = sg_cnt;
460 mmc_dat.blksz = func_blk_sz;
461 mmc_dat.blocks = req_sz / func_blk_sz; 477 mmc_dat.blocks = req_sz / func_blk_sz;
462 mmc_dat.flags = write ? MMC_DATA_WRITE : MMC_DATA_READ;
463 mmc_cmd.opcode = SD_IO_RW_EXTENDED;
464 mmc_cmd.arg = write ? 1<<31 : 0; /* write flag */
465 mmc_cmd.arg |= (fn & 0x7) << 28; /* SDIO func num */
466 mmc_cmd.arg |= 1<<27; /* block mode */
467 /* incrementing addr for function 1 */
468 mmc_cmd.arg |= (fn == 1) ? 1<<26 : 0;
469 mmc_cmd.arg |= (addr & 0x1FFFF) << 9; /* address */ 478 mmc_cmd.arg |= (addr & 0x1FFFF) << 9; /* address */
470 mmc_cmd.arg |= mmc_dat.blocks & 0x1FF; /* block count */ 479 mmc_cmd.arg |= mmc_dat.blocks & 0x1FF; /* block count */
471 mmc_cmd.flags = MMC_RSP_SPI_R5 | MMC_RSP_R5 | MMC_CMD_ADTC; 480 /* incrementing addr for function 1 */
472 mmc_req.cmd = &mmc_cmd;
473 mmc_req.data = &mmc_dat;
474 if (fn == 1) 481 if (fn == 1)
475 addr += req_sz; 482 addr += req_sz;
476 483
477 mmc_set_data_timeout(&mmc_dat, sdiodev->func[fn]->card); 484 mmc_set_data_timeout(&mmc_dat, sdiodev->func[fn]->card);
478 mmc_wait_for_req(host, &mmc_req); 485 mmc_wait_for_req(sdiodev->func[fn]->card->host, &mmc_req);
479 486
480 ret = mmc_cmd.error ? mmc_cmd.error : mmc_dat.error; 487 ret = mmc_cmd.error ? mmc_cmd.error : mmc_dat.error;
481 if (ret != 0) { 488 if (ret != 0) {
@@ -546,7 +553,6 @@ brcmf_sdcard_recv_pkt(struct brcmf_sdio_dev *sdiodev, u32 addr, uint fn,
546{ 553{
547 uint width; 554 uint width;
548 int err = 0; 555 int err = 0;
549 struct sk_buff_head pkt_list;
550 556
551 brcmf_dbg(SDIO, "fun = %d, addr = 0x%x, size = %d\n", 557 brcmf_dbg(SDIO, "fun = %d, addr = 0x%x, size = %d\n",
552 fn, addr, pkt->len); 558 fn, addr, pkt->len);
@@ -556,19 +562,17 @@ brcmf_sdcard_recv_pkt(struct brcmf_sdio_dev *sdiodev, u32 addr, uint fn,
556 if (err) 562 if (err)
557 goto done; 563 goto done;
558 564
559 skb_queue_head_init(&pkt_list); 565 err = brcmf_sdio_buffrw(sdiodev, fn, false, addr, pkt);
560 skb_queue_tail(&pkt_list, pkt);
561 err = brcmf_sdio_buffrw(sdiodev, fn, false, addr, &pkt_list);
562 skb_dequeue_tail(&pkt_list);
563 566
564done: 567done:
565 return err; 568 return err;
566} 569}
567 570
568int brcmf_sdcard_recv_chain(struct brcmf_sdio_dev *sdiodev, u32 addr, uint fn, 571int brcmf_sdcard_recv_chain(struct brcmf_sdio_dev *sdiodev, u32 addr, uint fn,
569 uint flags, struct sk_buff_head *pktq) 572 uint flags, struct sk_buff_head *pktq, uint totlen)
570{ 573{
571 uint incr_fix; 574 struct sk_buff *glom_skb;
575 struct sk_buff *skb;
572 uint width; 576 uint width;
573 int err = 0; 577 int err = 0;
574 578
@@ -580,8 +584,22 @@ int brcmf_sdcard_recv_chain(struct brcmf_sdio_dev *sdiodev, u32 addr, uint fn,
580 if (err) 584 if (err)
581 goto done; 585 goto done;
582 586
583 incr_fix = (flags & SDIO_REQ_FIXED) ? SDIOH_DATA_FIX : SDIOH_DATA_INC; 587 if (pktq->qlen == 1)
584 err = brcmf_sdio_buffrw(sdiodev, fn, false, addr, pktq); 588 err = brcmf_sdio_buffrw(sdiodev, fn, false, addr, pktq->next);
589 else if (!sdiodev->sg_support) {
590 glom_skb = brcmu_pkt_buf_get_skb(totlen);
591 if (!glom_skb)
592 return -ENOMEM;
593 err = brcmf_sdio_buffrw(sdiodev, fn, false, addr, glom_skb);
594 if (err)
595 goto done;
596
597 skb_queue_walk(pktq, skb) {
598 memcpy(skb->data, glom_skb->data, skb->len);
599 skb_pull(glom_skb, skb->len);
600 }
601 } else
602 err = brcmf_sdio_sglist_rw(sdiodev, fn, false, addr, pktq);
585 603
586done: 604done:
587 return err; 605 return err;
@@ -592,7 +610,7 @@ brcmf_sdcard_send_buf(struct brcmf_sdio_dev *sdiodev, u32 addr, uint fn,
592 uint flags, u8 *buf, uint nbytes) 610 uint flags, u8 *buf, uint nbytes)
593{ 611{
594 struct sk_buff *mypkt; 612 struct sk_buff *mypkt;
595 struct sk_buff_head pktq; 613 uint width;
596 int err; 614 int err;
597 615
598 mypkt = brcmu_pkt_buf_get_skb(nbytes); 616 mypkt = brcmu_pkt_buf_get_skb(nbytes);
@@ -603,10 +621,12 @@ brcmf_sdcard_send_buf(struct brcmf_sdio_dev *sdiodev, u32 addr, uint fn,
603 } 621 }
604 622
605 memcpy(mypkt->data, buf, nbytes); 623 memcpy(mypkt->data, buf, nbytes);
606 __skb_queue_head_init(&pktq); 624
607 __skb_queue_tail(&pktq, mypkt); 625 width = (flags & SDIO_REQ_4BYTE) ? 4 : 2;
608 err = brcmf_sdcard_send_pkt(sdiodev, addr, fn, flags, &pktq); 626 err = brcmf_sdio_addrprep(sdiodev, width, &addr);
609 __skb_dequeue_tail(&pktq); 627
628 if (!err)
629 err = brcmf_sdio_buffrw(sdiodev, fn, true, addr, mypkt);
610 630
611 brcmu_pkt_buf_free_skb(mypkt); 631 brcmu_pkt_buf_free_skb(mypkt);
612 return err; 632 return err;
@@ -617,16 +637,26 @@ int
617brcmf_sdcard_send_pkt(struct brcmf_sdio_dev *sdiodev, u32 addr, uint fn, 637brcmf_sdcard_send_pkt(struct brcmf_sdio_dev *sdiodev, u32 addr, uint fn,
618 uint flags, struct sk_buff_head *pktq) 638 uint flags, struct sk_buff_head *pktq)
619{ 639{
640 struct sk_buff *skb;
620 uint width; 641 uint width;
621 int err = 0; 642 int err;
622 643
623 brcmf_dbg(SDIO, "fun = %d, addr = 0x%x, size = %d\n", 644 brcmf_dbg(SDIO, "fun = %d, addr = 0x%x, size = %d\n",
624 fn, addr, pktq->qlen); 645 fn, addr, pktq->qlen);
625 646
626 width = (flags & SDIO_REQ_4BYTE) ? 4 : 2; 647 width = (flags & SDIO_REQ_4BYTE) ? 4 : 2;
627 brcmf_sdio_addrprep(sdiodev, width, &addr); 648 err = brcmf_sdio_addrprep(sdiodev, width, &addr);
649 if (err)
650 return err;
628 651
629 err = brcmf_sdio_buffrw(sdiodev, fn, true, addr, pktq); 652 if (pktq->qlen == 1 || !sdiodev->sg_support)
653 skb_queue_walk(pktq, skb) {
654 err = brcmf_sdio_buffrw(sdiodev, fn, true, addr, skb);
655 if (err)
656 break;
657 }
658 else
659 err = brcmf_sdio_sglist_rw(sdiodev, fn, true, addr, pktq);
630 660
631 return err; 661 return err;
632} 662}
@@ -639,7 +669,6 @@ brcmf_sdio_ramrw(struct brcmf_sdio_dev *sdiodev, bool write, u32 address,
639 struct sk_buff *pkt; 669 struct sk_buff *pkt;
640 u32 sdaddr; 670 u32 sdaddr;
641 uint dsize; 671 uint dsize;
642 struct sk_buff_head pkt_list;
643 672
644 dsize = min_t(uint, SBSDIO_SB_OFT_ADDR_LIMIT, size); 673 dsize = min_t(uint, SBSDIO_SB_OFT_ADDR_LIMIT, size);
645 pkt = dev_alloc_skb(dsize); 674 pkt = dev_alloc_skb(dsize);
@@ -648,7 +677,6 @@ brcmf_sdio_ramrw(struct brcmf_sdio_dev *sdiodev, bool write, u32 address,
648 return -EIO; 677 return -EIO;
649 } 678 }
650 pkt->priority = 0; 679 pkt->priority = 0;
651 skb_queue_head_init(&pkt_list);
652 680
653 /* Determine initial transfer parameters */ 681 /* Determine initial transfer parameters */
654 sdaddr = address & SBSDIO_SB_OFT_ADDR_MASK; 682 sdaddr = address & SBSDIO_SB_OFT_ADDR_MASK;
@@ -676,10 +704,8 @@ brcmf_sdio_ramrw(struct brcmf_sdio_dev *sdiodev, bool write, u32 address,
676 skb_put(pkt, dsize); 704 skb_put(pkt, dsize);
677 if (write) 705 if (write)
678 memcpy(pkt->data, data, dsize); 706 memcpy(pkt->data, data, dsize);
679 skb_queue_tail(&pkt_list, pkt);
680 bcmerror = brcmf_sdio_buffrw(sdiodev, SDIO_FUNC_1, write, 707 bcmerror = brcmf_sdio_buffrw(sdiodev, SDIO_FUNC_1, write,
681 sdaddr, &pkt_list); 708 sdaddr, pkt);
682 skb_dequeue_tail(&pkt_list);
683 if (bcmerror) { 709 if (bcmerror) {
684 brcmf_err("membytes transfer failed\n"); 710 brcmf_err("membytes transfer failed\n");
685 break; 711 break;
diff --git a/drivers/net/wireless/brcm80211/brcmfmac/bcmsdh_sdmmc.c b/drivers/net/wireless/brcm80211/brcmfmac/bcmsdh_sdmmc.c
index c3462b75bd08..905704e335d7 100644
--- a/drivers/net/wireless/brcm80211/brcmfmac/bcmsdh_sdmmc.c
+++ b/drivers/net/wireless/brcm80211/brcmfmac/bcmsdh_sdmmc.c
@@ -21,6 +21,7 @@
21#include <linux/mmc/sdio_func.h> 21#include <linux/mmc/sdio_func.h>
22#include <linux/mmc/sdio_ids.h> 22#include <linux/mmc/sdio_ids.h>
23#include <linux/mmc/card.h> 23#include <linux/mmc/card.h>
24#include <linux/mmc/host.h>
24#include <linux/suspend.h> 25#include <linux/suspend.h>
25#include <linux/errno.h> 26#include <linux/errno.h>
26#include <linux/sched.h> /* request_irq() */ 27#include <linux/sched.h> /* request_irq() */
@@ -34,6 +35,7 @@
34#include <brcmu_utils.h> 35#include <brcmu_utils.h>
35#include <brcmu_wifi.h> 36#include <brcmu_wifi.h>
36#include "sdio_host.h" 37#include "sdio_host.h"
38#include "sdio_chip.h"
37#include "dhd_dbg.h" 39#include "dhd_dbg.h"
38#include "dhd_bus.h" 40#include "dhd_bus.h"
39 41
@@ -41,13 +43,6 @@
41 43
42#define DMA_ALIGN_MASK 0x03 44#define DMA_ALIGN_MASK 0x03
43 45
44#define SDIO_DEVICE_ID_BROADCOM_43143 43143
45#define SDIO_DEVICE_ID_BROADCOM_43241 0x4324
46#define SDIO_DEVICE_ID_BROADCOM_4329 0x4329
47#define SDIO_DEVICE_ID_BROADCOM_4330 0x4330
48#define SDIO_DEVICE_ID_BROADCOM_4334 0x4334
49#define SDIO_DEVICE_ID_BROADCOM_4335 0x4335
50
51#define SDIO_FUNC1_BLOCKSIZE 64 46#define SDIO_FUNC1_BLOCKSIZE 64
52#define SDIO_FUNC2_BLOCKSIZE 512 47#define SDIO_FUNC2_BLOCKSIZE 512
53 48
@@ -58,7 +53,8 @@ static const struct sdio_device_id brcmf_sdmmc_ids[] = {
58 {SDIO_DEVICE(SDIO_VENDOR_ID_BROADCOM, SDIO_DEVICE_ID_BROADCOM_4329)}, 53 {SDIO_DEVICE(SDIO_VENDOR_ID_BROADCOM, SDIO_DEVICE_ID_BROADCOM_4329)},
59 {SDIO_DEVICE(SDIO_VENDOR_ID_BROADCOM, SDIO_DEVICE_ID_BROADCOM_4330)}, 54 {SDIO_DEVICE(SDIO_VENDOR_ID_BROADCOM, SDIO_DEVICE_ID_BROADCOM_4330)},
60 {SDIO_DEVICE(SDIO_VENDOR_ID_BROADCOM, SDIO_DEVICE_ID_BROADCOM_4334)}, 55 {SDIO_DEVICE(SDIO_VENDOR_ID_BROADCOM, SDIO_DEVICE_ID_BROADCOM_4334)},
61 {SDIO_DEVICE(SDIO_VENDOR_ID_BROADCOM, SDIO_DEVICE_ID_BROADCOM_4335)}, 56 {SDIO_DEVICE(SDIO_VENDOR_ID_BROADCOM,
57 SDIO_DEVICE_ID_BROADCOM_4335_4339)},
62 { /* end: all zeroes */ }, 58 { /* end: all zeroes */ },
63}; 59};
64MODULE_DEVICE_TABLE(sdio, brcmf_sdmmc_ids); 60MODULE_DEVICE_TABLE(sdio, brcmf_sdmmc_ids);
@@ -320,6 +316,8 @@ static int brcmf_ops_sdio_probe(struct sdio_func *func,
320 int err; 316 int err;
321 struct brcmf_sdio_dev *sdiodev; 317 struct brcmf_sdio_dev *sdiodev;
322 struct brcmf_bus *bus_if; 318 struct brcmf_bus *bus_if;
319 struct mmc_host *host;
320 uint max_blocks;
323 321
324 brcmf_dbg(SDIO, "Enter\n"); 322 brcmf_dbg(SDIO, "Enter\n");
325 brcmf_dbg(SDIO, "Class=%x\n", func->class); 323 brcmf_dbg(SDIO, "Class=%x\n", func->class);
@@ -366,6 +364,20 @@ static int brcmf_ops_sdio_probe(struct sdio_func *func,
366 brcmf_err("F2 error, probe failed %d...\n", err); 364 brcmf_err("F2 error, probe failed %d...\n", err);
367 goto fail; 365 goto fail;
368 } 366 }
367
368 /*
369 * determine host related variables after brcmf_sdio_probe()
370 * as func->cur_blksize is properly set and F2 init has been
371 * completed successfully.
372 */
373 host = func->card->host;
374 sdiodev->sg_support = host->max_segs > 1;
375 max_blocks = min_t(uint, host->max_blk_count, 511u);
376 sdiodev->max_request_size = min_t(uint, host->max_req_size,
377 max_blocks * func->cur_blksize);
378 sdiodev->max_segment_count = min_t(uint, host->max_segs,
379 SG_MAX_SINGLE_ALLOC);
380 sdiodev->max_segment_size = host->max_seg_size;
369 brcmf_dbg(SDIO, "F2 init completed...\n"); 381 brcmf_dbg(SDIO, "F2 init completed...\n");
370 return 0; 382 return 0;
371 383
@@ -466,7 +478,7 @@ static int brcmf_sdio_pd_probe(struct platform_device *pdev)
466{ 478{
467 brcmf_dbg(SDIO, "Enter\n"); 479 brcmf_dbg(SDIO, "Enter\n");
468 480
469 brcmfmac_sdio_pdata = pdev->dev.platform_data; 481 brcmfmac_sdio_pdata = dev_get_platdata(&pdev->dev);
470 482
471 if (brcmfmac_sdio_pdata->power_on) 483 if (brcmfmac_sdio_pdata->power_on)
472 brcmfmac_sdio_pdata->power_on(); 484 brcmfmac_sdio_pdata->power_on();
diff --git a/drivers/net/wireless/brcm80211/brcmfmac/dhd.h b/drivers/net/wireless/brcm80211/brcmfmac/dhd.h
index 2eb9e642c9bf..899a2ada5b82 100644
--- a/drivers/net/wireless/brcm80211/brcmfmac/dhd.h
+++ b/drivers/net/wireless/brcm80211/brcmfmac/dhd.h
@@ -97,8 +97,6 @@
97#define WLC_PHY_TYPE_LCN 8 97#define WLC_PHY_TYPE_LCN 8
98#define WLC_PHY_TYPE_NULL 0xf 98#define WLC_PHY_TYPE_NULL 0xf
99 99
100#define BRCMF_EVENTING_MASK_LEN 16
101
102#define TOE_TX_CSUM_OL 0x00000001 100#define TOE_TX_CSUM_OL 0x00000001
103#define TOE_RX_CSUM_OL 0x00000002 101#define TOE_RX_CSUM_OL 0x00000002
104 102
@@ -632,29 +630,29 @@ struct brcmf_skb_reorder_data {
632 u8 *reorder; 630 u8 *reorder;
633}; 631};
634 632
635extern int brcmf_netdev_wait_pend8021x(struct net_device *ndev); 633int brcmf_netdev_wait_pend8021x(struct net_device *ndev);
636 634
637/* Return pointer to interface name */ 635/* Return pointer to interface name */
638extern char *brcmf_ifname(struct brcmf_pub *drvr, int idx); 636char *brcmf_ifname(struct brcmf_pub *drvr, int idx);
639 637
640/* Query dongle */ 638/* Query dongle */
641extern int brcmf_proto_cdc_query_dcmd(struct brcmf_pub *drvr, int ifidx, 639int brcmf_proto_cdc_query_dcmd(struct brcmf_pub *drvr, int ifidx, uint cmd,
642 uint cmd, void *buf, uint len); 640 void *buf, uint len);
643extern int brcmf_proto_cdc_set_dcmd(struct brcmf_pub *drvr, int ifidx, uint cmd, 641int brcmf_proto_cdc_set_dcmd(struct brcmf_pub *drvr, int ifidx, uint cmd,
644 void *buf, uint len); 642 void *buf, uint len);
645 643
646/* Remove any protocol-specific data header. */ 644/* Remove any protocol-specific data header. */
647extern int brcmf_proto_hdrpull(struct brcmf_pub *drvr, bool do_fws, u8 *ifidx, 645int brcmf_proto_hdrpull(struct brcmf_pub *drvr, bool do_fws, u8 *ifidx,
648 struct sk_buff *rxp); 646 struct sk_buff *rxp);
649 647
650extern int brcmf_net_attach(struct brcmf_if *ifp, bool rtnl_locked); 648int brcmf_net_attach(struct brcmf_if *ifp, bool rtnl_locked);
651extern struct brcmf_if *brcmf_add_if(struct brcmf_pub *drvr, s32 bssidx, 649struct brcmf_if *brcmf_add_if(struct brcmf_pub *drvr, s32 bssidx, s32 ifidx,
652 s32 ifidx, char *name, u8 *mac_addr); 650 char *name, u8 *mac_addr);
653extern void brcmf_del_if(struct brcmf_pub *drvr, s32 bssidx); 651void brcmf_del_if(struct brcmf_pub *drvr, s32 bssidx);
654void brcmf_txflowblock_if(struct brcmf_if *ifp, 652void brcmf_txflowblock_if(struct brcmf_if *ifp,
655 enum brcmf_netif_stop_reason reason, bool state); 653 enum brcmf_netif_stop_reason reason, bool state);
656extern u32 brcmf_get_chip_info(struct brcmf_if *ifp); 654u32 brcmf_get_chip_info(struct brcmf_if *ifp);
657extern void brcmf_txfinalize(struct brcmf_pub *drvr, struct sk_buff *txp, 655void brcmf_txfinalize(struct brcmf_pub *drvr, struct sk_buff *txp,
658 bool success); 656 bool success);
659 657
660#endif /* _BRCMF_H_ */ 658#endif /* _BRCMF_H_ */
diff --git a/drivers/net/wireless/brcm80211/brcmfmac/dhd_bus.h b/drivers/net/wireless/brcm80211/brcmfmac/dhd_bus.h
index 74156f84180c..a6eb09e5d46f 100644
--- a/drivers/net/wireless/brcm80211/brcmfmac/dhd_bus.h
+++ b/drivers/net/wireless/brcm80211/brcmfmac/dhd_bus.h
@@ -132,35 +132,34 @@ struct pktq *brcmf_bus_gettxq(struct brcmf_bus *bus)
132 * interface functions from common layer 132 * interface functions from common layer
133 */ 133 */
134 134
135extern bool brcmf_c_prec_enq(struct device *dev, struct pktq *q, 135bool brcmf_c_prec_enq(struct device *dev, struct pktq *q, struct sk_buff *pkt,
136 struct sk_buff *pkt, int prec); 136 int prec);
137 137
138/* Receive frame for delivery to OS. Callee disposes of rxp. */ 138/* Receive frame for delivery to OS. Callee disposes of rxp. */
139extern void brcmf_rx_frames(struct device *dev, struct sk_buff_head *rxlist); 139void brcmf_rx_frame(struct device *dev, struct sk_buff *rxp);
140 140
141/* Indication from bus module regarding presence/insertion of dongle. */ 141/* Indication from bus module regarding presence/insertion of dongle. */
142extern int brcmf_attach(uint bus_hdrlen, struct device *dev); 142int brcmf_attach(uint bus_hdrlen, struct device *dev);
143/* Indication from bus module regarding removal/absence of dongle */ 143/* Indication from bus module regarding removal/absence of dongle */
144extern void brcmf_detach(struct device *dev); 144void brcmf_detach(struct device *dev);
145/* Indication from bus module that dongle should be reset */ 145/* Indication from bus module that dongle should be reset */
146extern void brcmf_dev_reset(struct device *dev); 146void brcmf_dev_reset(struct device *dev);
147/* Indication from bus module to change flow-control state */ 147/* Indication from bus module to change flow-control state */
148extern void brcmf_txflowblock(struct device *dev, bool state); 148void brcmf_txflowblock(struct device *dev, bool state);
149 149
150/* Notify the bus has transferred the tx packet to firmware */ 150/* Notify the bus has transferred the tx packet to firmware */
151extern void brcmf_txcomplete(struct device *dev, struct sk_buff *txp, 151void brcmf_txcomplete(struct device *dev, struct sk_buff *txp, bool success);
152 bool success);
153 152
154extern int brcmf_bus_start(struct device *dev); 153int brcmf_bus_start(struct device *dev);
155 154
156#ifdef CONFIG_BRCMFMAC_SDIO 155#ifdef CONFIG_BRCMFMAC_SDIO
157extern void brcmf_sdio_exit(void); 156void brcmf_sdio_exit(void);
158extern void brcmf_sdio_init(void); 157void brcmf_sdio_init(void);
159extern void brcmf_sdio_register(void); 158void brcmf_sdio_register(void);
160#endif 159#endif
161#ifdef CONFIG_BRCMFMAC_USB 160#ifdef CONFIG_BRCMFMAC_USB
162extern void brcmf_usb_exit(void); 161void brcmf_usb_exit(void);
163extern void brcmf_usb_register(void); 162void brcmf_usb_register(void);
164#endif 163#endif
165 164
166#endif /* _BRCMF_BUS_H_ */ 165#endif /* _BRCMF_BUS_H_ */
diff --git a/drivers/net/wireless/brcm80211/brcmfmac/dhd_linux.c b/drivers/net/wireless/brcm80211/brcmfmac/dhd_linux.c
index 40e7f854e10f..64e9cff241b9 100644
--- a/drivers/net/wireless/brcm80211/brcmfmac/dhd_linux.c
+++ b/drivers/net/wireless/brcm80211/brcmfmac/dhd_linux.c
@@ -509,9 +509,8 @@ netif_rx:
509 } 509 }
510} 510}
511 511
512void brcmf_rx_frames(struct device *dev, struct sk_buff_head *skb_list) 512void brcmf_rx_frame(struct device *dev, struct sk_buff *skb)
513{ 513{
514 struct sk_buff *skb, *pnext;
515 struct brcmf_if *ifp; 514 struct brcmf_if *ifp;
516 struct brcmf_bus *bus_if = dev_get_drvdata(dev); 515 struct brcmf_bus *bus_if = dev_get_drvdata(dev);
517 struct brcmf_pub *drvr = bus_if->drvr; 516 struct brcmf_pub *drvr = bus_if->drvr;
@@ -519,29 +518,24 @@ void brcmf_rx_frames(struct device *dev, struct sk_buff_head *skb_list)
519 u8 ifidx; 518 u8 ifidx;
520 int ret; 519 int ret;
521 520
522 brcmf_dbg(DATA, "Enter: %s: count=%u\n", dev_name(dev), 521 brcmf_dbg(DATA, "Enter: %s: rxp=%p\n", dev_name(dev), skb);
523 skb_queue_len(skb_list));
524 522
525 skb_queue_walk_safe(skb_list, skb, pnext) { 523 /* process and remove protocol-specific header */
526 skb_unlink(skb, skb_list); 524 ret = brcmf_proto_hdrpull(drvr, true, &ifidx, skb);
527 525 ifp = drvr->iflist[ifidx];
528 /* process and remove protocol-specific header */
529 ret = brcmf_proto_hdrpull(drvr, true, &ifidx, skb);
530 ifp = drvr->iflist[ifidx];
531
532 if (ret || !ifp || !ifp->ndev) {
533 if ((ret != -ENODATA) && ifp)
534 ifp->stats.rx_errors++;
535 brcmu_pkt_buf_free_skb(skb);
536 continue;
537 }
538 526
539 rd = (struct brcmf_skb_reorder_data *)skb->cb; 527 if (ret || !ifp || !ifp->ndev) {
540 if (rd->reorder) 528 if ((ret != -ENODATA) && ifp)
541 brcmf_rxreorder_process_info(ifp, rd->reorder, skb); 529 ifp->stats.rx_errors++;
542 else 530 brcmu_pkt_buf_free_skb(skb);
543 brcmf_netif_rx(ifp, skb); 531 return;
544 } 532 }
533
534 rd = (struct brcmf_skb_reorder_data *)skb->cb;
535 if (rd->reorder)
536 brcmf_rxreorder_process_info(ifp, rd->reorder, skb);
537 else
538 brcmf_netif_rx(ifp, skb);
545} 539}
546 540
547void brcmf_txfinalize(struct brcmf_pub *drvr, struct sk_buff *txp, 541void brcmf_txfinalize(struct brcmf_pub *drvr, struct sk_buff *txp,
diff --git a/drivers/net/wireless/brcm80211/brcmfmac/dhd_proto.h b/drivers/net/wireless/brcm80211/brcmfmac/dhd_proto.h
index ef9179883748..53c6e710f2cb 100644
--- a/drivers/net/wireless/brcm80211/brcmfmac/dhd_proto.h
+++ b/drivers/net/wireless/brcm80211/brcmfmac/dhd_proto.h
@@ -22,21 +22,21 @@
22 */ 22 */
23 23
24/* Linkage, sets prot link and updates hdrlen in pub */ 24/* Linkage, sets prot link and updates hdrlen in pub */
25extern int brcmf_proto_attach(struct brcmf_pub *drvr); 25int brcmf_proto_attach(struct brcmf_pub *drvr);
26 26
27/* Unlink, frees allocated protocol memory (including brcmf_proto) */ 27/* Unlink, frees allocated protocol memory (including brcmf_proto) */
28extern void brcmf_proto_detach(struct brcmf_pub *drvr); 28void brcmf_proto_detach(struct brcmf_pub *drvr);
29 29
30/* Stop protocol: sync w/dongle state. */ 30/* Stop protocol: sync w/dongle state. */
31extern void brcmf_proto_stop(struct brcmf_pub *drvr); 31void brcmf_proto_stop(struct brcmf_pub *drvr);
32 32
33/* Add any protocol-specific data header. 33/* Add any protocol-specific data header.
34 * Caller must reserve prot_hdrlen prepend space. 34 * Caller must reserve prot_hdrlen prepend space.
35 */ 35 */
36extern void brcmf_proto_hdrpush(struct brcmf_pub *, int ifidx, u8 offset, 36void brcmf_proto_hdrpush(struct brcmf_pub *, int ifidx, u8 offset,
37 struct sk_buff *txp); 37 struct sk_buff *txp);
38 38
39/* Sets dongle media info (drv_version, mac address). */ 39/* Sets dongle media info (drv_version, mac address). */
40extern int brcmf_c_preinit_dcmds(struct brcmf_if *ifp); 40int brcmf_c_preinit_dcmds(struct brcmf_if *ifp);
41 41
42#endif /* _BRCMF_PROTO_H_ */ 42#endif /* _BRCMF_PROTO_H_ */
diff --git a/drivers/net/wireless/brcm80211/brcmfmac/dhd_sdio.c b/drivers/net/wireless/brcm80211/brcmfmac/dhd_sdio.c
index 1aa75d5951b8..b02953c4ade7 100644
--- a/drivers/net/wireless/brcm80211/brcmfmac/dhd_sdio.c
+++ b/drivers/net/wireless/brcm80211/brcmfmac/dhd_sdio.c
@@ -275,11 +275,6 @@ struct rte_console {
275/* Flags for SDH calls */ 275/* Flags for SDH calls */
276#define F2SYNC (SDIO_REQ_4BYTE | SDIO_REQ_FIXED) 276#define F2SYNC (SDIO_REQ_4BYTE | SDIO_REQ_FIXED)
277 277
278#define BRCMF_SDIO_FW_NAME "brcm/brcmfmac-sdio.bin"
279#define BRCMF_SDIO_NV_NAME "brcm/brcmfmac-sdio.txt"
280MODULE_FIRMWARE(BRCMF_SDIO_FW_NAME);
281MODULE_FIRMWARE(BRCMF_SDIO_NV_NAME);
282
283#define BRCMF_IDLE_IMMEDIATE (-1) /* Enter idle immediately */ 278#define BRCMF_IDLE_IMMEDIATE (-1) /* Enter idle immediately */
284#define BRCMF_IDLE_ACTIVE 0 /* Do not request any SD clock change 279#define BRCMF_IDLE_ACTIVE 0 /* Do not request any SD clock change
285 * when idle 280 * when idle
@@ -454,9 +449,6 @@ struct brcmf_sdio {
454 struct work_struct datawork; 449 struct work_struct datawork;
455 atomic_t dpc_tskcnt; 450 atomic_t dpc_tskcnt;
456 451
457 const struct firmware *firmware;
458 u32 fw_ptr;
459
460 bool txoff; /* Transmit flow-controlled */ 452 bool txoff; /* Transmit flow-controlled */
461 struct brcmf_sdio_count sdcnt; 453 struct brcmf_sdio_count sdcnt;
462 bool sr_enabled; /* SaveRestore enabled */ 454 bool sr_enabled; /* SaveRestore enabled */
@@ -493,6 +485,100 @@ enum brcmf_sdio_frmtype {
493 BRCMF_SDIO_FT_SUB, 485 BRCMF_SDIO_FT_SUB,
494}; 486};
495 487
488#define BCM43143_FIRMWARE_NAME "brcm/brcmfmac43143-sdio.bin"
489#define BCM43143_NVRAM_NAME "brcm/brcmfmac43143-sdio.txt"
490#define BCM43241B0_FIRMWARE_NAME "brcm/brcmfmac43241b0-sdio.bin"
491#define BCM43241B0_NVRAM_NAME "brcm/brcmfmac43241b0-sdio.txt"
492#define BCM43241B4_FIRMWARE_NAME "brcm/brcmfmac43241b4-sdio.bin"
493#define BCM43241B4_NVRAM_NAME "brcm/brcmfmac43241b4-sdio.txt"
494#define BCM4329_FIRMWARE_NAME "brcm/brcmfmac4329-sdio.bin"
495#define BCM4329_NVRAM_NAME "brcm/brcmfmac4329-sdio.txt"
496#define BCM4330_FIRMWARE_NAME "brcm/brcmfmac4330-sdio.bin"
497#define BCM4330_NVRAM_NAME "brcm/brcmfmac4330-sdio.txt"
498#define BCM4334_FIRMWARE_NAME "brcm/brcmfmac4334-sdio.bin"
499#define BCM4334_NVRAM_NAME "brcm/brcmfmac4334-sdio.txt"
500#define BCM4335_FIRMWARE_NAME "brcm/brcmfmac4335-sdio.bin"
501#define BCM4335_NVRAM_NAME "brcm/brcmfmac4335-sdio.txt"
502
503MODULE_FIRMWARE(BCM43143_FIRMWARE_NAME);
504MODULE_FIRMWARE(BCM43143_NVRAM_NAME);
505MODULE_FIRMWARE(BCM43241B0_FIRMWARE_NAME);
506MODULE_FIRMWARE(BCM43241B0_NVRAM_NAME);
507MODULE_FIRMWARE(BCM43241B4_FIRMWARE_NAME);
508MODULE_FIRMWARE(BCM43241B4_NVRAM_NAME);
509MODULE_FIRMWARE(BCM4329_FIRMWARE_NAME);
510MODULE_FIRMWARE(BCM4329_NVRAM_NAME);
511MODULE_FIRMWARE(BCM4330_FIRMWARE_NAME);
512MODULE_FIRMWARE(BCM4330_NVRAM_NAME);
513MODULE_FIRMWARE(BCM4334_FIRMWARE_NAME);
514MODULE_FIRMWARE(BCM4334_NVRAM_NAME);
515MODULE_FIRMWARE(BCM4335_FIRMWARE_NAME);
516MODULE_FIRMWARE(BCM4335_NVRAM_NAME);
517
518struct brcmf_firmware_names {
519 u32 chipid;
520 u32 revmsk;
521 const char *bin;
522 const char *nv;
523};
524
525enum brcmf_firmware_type {
526 BRCMF_FIRMWARE_BIN,
527 BRCMF_FIRMWARE_NVRAM
528};
529
530#define BRCMF_FIRMWARE_NVRAM(name) \
531 name ## _FIRMWARE_NAME, name ## _NVRAM_NAME
532
533static const struct brcmf_firmware_names brcmf_fwname_data[] = {
534 { BCM43143_CHIP_ID, 0xFFFFFFFF, BRCMF_FIRMWARE_NVRAM(BCM43143) },
535 { BCM43241_CHIP_ID, 0x0000001F, BRCMF_FIRMWARE_NVRAM(BCM43241B0) },
536 { BCM43241_CHIP_ID, 0xFFFFFFE0, BRCMF_FIRMWARE_NVRAM(BCM43241B4) },
537 { BCM4329_CHIP_ID, 0xFFFFFFFF, BRCMF_FIRMWARE_NVRAM(BCM4329) },
538 { BCM4330_CHIP_ID, 0xFFFFFFFF, BRCMF_FIRMWARE_NVRAM(BCM4330) },
539 { BCM4334_CHIP_ID, 0xFFFFFFFF, BRCMF_FIRMWARE_NVRAM(BCM4334) },
540 { BCM4335_CHIP_ID, 0xFFFFFFFF, BRCMF_FIRMWARE_NVRAM(BCM4335) }
541};
542
543
544static const struct firmware *brcmf_sdbrcm_get_fw(struct brcmf_sdio *bus,
545 enum brcmf_firmware_type type)
546{
547 const struct firmware *fw;
548 const char *name;
549 int err, i;
550
551 for (i = 0; i < ARRAY_SIZE(brcmf_fwname_data); i++) {
552 if (brcmf_fwname_data[i].chipid == bus->ci->chip &&
553 brcmf_fwname_data[i].revmsk & BIT(bus->ci->chiprev)) {
554 switch (type) {
555 case BRCMF_FIRMWARE_BIN:
556 name = brcmf_fwname_data[i].bin;
557 break;
558 case BRCMF_FIRMWARE_NVRAM:
559 name = brcmf_fwname_data[i].nv;
560 break;
561 default:
562 brcmf_err("invalid firmware type (%d)\n", type);
563 return NULL;
564 }
565 goto found;
566 }
567 }
568 brcmf_err("Unknown chipid %d [%d]\n",
569 bus->ci->chip, bus->ci->chiprev);
570 return NULL;
571
572found:
573 err = request_firmware(&fw, name, &bus->sdiodev->func[2]->dev);
574 if ((err) || (!fw)) {
575 brcmf_err("fail to request firmware %s (%d)\n", name, err);
576 return NULL;
577 }
578
579 return fw;
580}
581
496static void pkt_align(struct sk_buff *p, int len, int align) 582static void pkt_align(struct sk_buff *p, int len, int align)
497{ 583{
498 uint datalign; 584 uint datalign;
@@ -1061,6 +1147,8 @@ static int brcmf_sdio_hdparse(struct brcmf_sdio *bus, u8 *header,
1061 u8 rx_seq, fc, tx_seq_max; 1147 u8 rx_seq, fc, tx_seq_max;
1062 u32 swheader; 1148 u32 swheader;
1063 1149
1150 trace_brcmf_sdpcm_hdr(false, header);
1151
1064 /* hw header */ 1152 /* hw header */
1065 len = get_unaligned_le16(header); 1153 len = get_unaligned_le16(header);
1066 checksum = get_unaligned_le16(header + sizeof(u16)); 1154 checksum = get_unaligned_le16(header + sizeof(u16));
@@ -1183,6 +1271,7 @@ static void brcmf_sdio_hdpack(struct brcmf_sdio *bus, u8 *header,
1183 SDPCM_DOFFSET_MASK; 1271 SDPCM_DOFFSET_MASK;
1184 *(((__le32 *)header) + 1) = cpu_to_le32(sw_header); 1272 *(((__le32 *)header) + 1) = cpu_to_le32(sw_header);
1185 *(((__le32 *)header) + 2) = 0; 1273 *(((__le32 *)header) + 2) = 0;
1274 trace_brcmf_sdpcm_hdr(true, header);
1186} 1275}
1187 1276
1188static u8 brcmf_sdbrcm_rxglom(struct brcmf_sdio *bus, u8 rxseq) 1277static u8 brcmf_sdbrcm_rxglom(struct brcmf_sdio *bus, u8 rxseq)
@@ -1303,7 +1392,7 @@ static u8 brcmf_sdbrcm_rxglom(struct brcmf_sdio *bus, u8 rxseq)
1303 sdio_claim_host(bus->sdiodev->func[1]); 1392 sdio_claim_host(bus->sdiodev->func[1]);
1304 errcode = brcmf_sdcard_recv_chain(bus->sdiodev, 1393 errcode = brcmf_sdcard_recv_chain(bus->sdiodev,
1305 bus->sdiodev->sbwad, 1394 bus->sdiodev->sbwad,
1306 SDIO_FUNC_2, F2SYNC, &bus->glom); 1395 SDIO_FUNC_2, F2SYNC, &bus->glom, dlen);
1307 sdio_release_host(bus->sdiodev->func[1]); 1396 sdio_release_host(bus->sdiodev->func[1]);
1308 bus->sdcnt.f2rxdata++; 1397 bus->sdcnt.f2rxdata++;
1309 1398
@@ -1406,13 +1495,12 @@ static u8 brcmf_sdbrcm_rxglom(struct brcmf_sdio *bus, u8 rxseq)
1406 bus->glom.qlen, pfirst, pfirst->data, 1495 bus->glom.qlen, pfirst, pfirst->data,
1407 pfirst->len, pfirst->next, 1496 pfirst->len, pfirst->next,
1408 pfirst->prev); 1497 pfirst->prev);
1498 skb_unlink(pfirst, &bus->glom);
1499 brcmf_rx_frame(bus->sdiodev->dev, pfirst);
1500 bus->sdcnt.rxglompkts++;
1409 } 1501 }
1410 /* sent any remaining packets up */
1411 if (bus->glom.qlen)
1412 brcmf_rx_frames(bus->sdiodev->dev, &bus->glom);
1413 1502
1414 bus->sdcnt.rxglomframes++; 1503 bus->sdcnt.rxglomframes++;
1415 bus->sdcnt.rxglompkts += bus->glom.qlen;
1416 } 1504 }
1417 return num; 1505 return num;
1418} 1506}
@@ -1557,7 +1645,6 @@ static void brcmf_pad(struct brcmf_sdio *bus, u16 *pad, u16 *rdlen)
1557static uint brcmf_sdio_readframes(struct brcmf_sdio *bus, uint maxframes) 1645static uint brcmf_sdio_readframes(struct brcmf_sdio *bus, uint maxframes)
1558{ 1646{
1559 struct sk_buff *pkt; /* Packet for event or data frames */ 1647 struct sk_buff *pkt; /* Packet for event or data frames */
1560 struct sk_buff_head pktlist; /* needed for bus interface */
1561 u16 pad; /* Number of pad bytes to read */ 1648 u16 pad; /* Number of pad bytes to read */
1562 uint rxleft = 0; /* Remaining number of frames allowed */ 1649 uint rxleft = 0; /* Remaining number of frames allowed */
1563 int ret; /* Return code from calls */ 1650 int ret; /* Return code from calls */
@@ -1759,9 +1846,7 @@ static uint brcmf_sdio_readframes(struct brcmf_sdio *bus, uint maxframes)
1759 continue; 1846 continue;
1760 } 1847 }
1761 1848
1762 skb_queue_head_init(&pktlist); 1849 brcmf_rx_frame(bus->sdiodev->dev, pkt);
1763 skb_queue_tail(&pktlist, pkt);
1764 brcmf_rx_frames(bus->sdiodev->dev, &pktlist);
1765 } 1850 }
1766 1851
1767 rxcount = maxframes - rxleft; 1852 rxcount = maxframes - rxleft;
@@ -1786,10 +1871,65 @@ brcmf_sdbrcm_wait_event_wakeup(struct brcmf_sdio *bus)
1786 return; 1871 return;
1787} 1872}
1788 1873
1874/**
1875 * struct brcmf_skbuff_cb reserves first two bytes in sk_buff::cb for
1876 * bus layer usage.
1877 */
1789/* flag marking a dummy skb added for DMA alignment requirement */ 1878/* flag marking a dummy skb added for DMA alignment requirement */
1790#define DUMMY_SKB_FLAG 0x10000 1879#define ALIGN_SKB_FLAG 0x8000
1791/* bit mask of data length chopped from the previous packet */ 1880/* bit mask of data length chopped from the previous packet */
1792#define DUMMY_SKB_CHOP_LEN_MASK 0xffff 1881#define ALIGN_SKB_CHOP_LEN_MASK 0x7fff
1882
1883static int brcmf_sdio_txpkt_prep_sg(struct brcmf_sdio_dev *sdiodev,
1884 struct sk_buff_head *pktq,
1885 struct sk_buff *pkt, uint chan)
1886{
1887 struct sk_buff *pkt_pad;
1888 u16 tail_pad, tail_chop, sg_align;
1889 unsigned int blksize;
1890 u8 *dat_buf;
1891 int ntail;
1892
1893 blksize = sdiodev->func[SDIO_FUNC_2]->cur_blksize;
1894 sg_align = 4;
1895 if (sdiodev->pdata && sdiodev->pdata->sd_sgentry_align > 4)
1896 sg_align = sdiodev->pdata->sd_sgentry_align;
1897 /* sg entry alignment should be a divisor of block size */
1898 WARN_ON(blksize % sg_align);
1899
1900 /* Check tail padding */
1901 pkt_pad = NULL;
1902 tail_chop = pkt->len % sg_align;
1903 tail_pad = sg_align - tail_chop;
1904 tail_pad += blksize - (pkt->len + tail_pad) % blksize;
1905 if (skb_tailroom(pkt) < tail_pad && pkt->len > blksize) {
1906 pkt_pad = brcmu_pkt_buf_get_skb(tail_pad + tail_chop);
1907 if (pkt_pad == NULL)
1908 return -ENOMEM;
1909 memcpy(pkt_pad->data,
1910 pkt->data + pkt->len - tail_chop,
1911 tail_chop);
1912 *(u32 *)(pkt_pad->cb) = ALIGN_SKB_FLAG + tail_chop;
1913 skb_trim(pkt, pkt->len - tail_chop);
1914 __skb_queue_after(pktq, pkt, pkt_pad);
1915 } else {
1916 ntail = pkt->data_len + tail_pad -
1917 (pkt->end - pkt->tail);
1918 if (skb_cloned(pkt) || ntail > 0)
1919 if (pskb_expand_head(pkt, 0, ntail, GFP_ATOMIC))
1920 return -ENOMEM;
1921 if (skb_linearize(pkt))
1922 return -ENOMEM;
1923 dat_buf = (u8 *)(pkt->data);
1924 __skb_put(pkt, tail_pad);
1925 }
1926
1927 if (pkt_pad)
1928 return pkt->len + tail_chop;
1929 else
1930 return pkt->len - tail_pad;
1931}
1932
1793/** 1933/**
1794 * brcmf_sdio_txpkt_prep - packet preparation for transmit 1934 * brcmf_sdio_txpkt_prep - packet preparation for transmit
1795 * @bus: brcmf_sdio structure pointer 1935 * @bus: brcmf_sdio structure pointer
@@ -1806,24 +1946,16 @@ static int
1806brcmf_sdio_txpkt_prep(struct brcmf_sdio *bus, struct sk_buff_head *pktq, 1946brcmf_sdio_txpkt_prep(struct brcmf_sdio *bus, struct sk_buff_head *pktq,
1807 uint chan) 1947 uint chan)
1808{ 1948{
1809 u16 head_pad, tail_pad, tail_chop, head_align, sg_align; 1949 u16 head_pad, head_align;
1810 int ntail; 1950 struct sk_buff *pkt_next;
1811 struct sk_buff *pkt_next, *pkt_new;
1812 u8 *dat_buf; 1951 u8 *dat_buf;
1813 unsigned blksize = bus->sdiodev->func[SDIO_FUNC_2]->cur_blksize; 1952 int err;
1814 struct brcmf_sdio_hdrinfo hd_info = {0}; 1953 struct brcmf_sdio_hdrinfo hd_info = {0};
1815 1954
1816 /* SDIO ADMA requires at least 32 bit alignment */ 1955 /* SDIO ADMA requires at least 32 bit alignment */
1817 head_align = 4; 1956 head_align = 4;
1818 sg_align = 4; 1957 if (bus->sdiodev->pdata && bus->sdiodev->pdata->sd_head_align > 4)
1819 if (bus->sdiodev->pdata) { 1958 head_align = bus->sdiodev->pdata->sd_head_align;
1820 head_align = bus->sdiodev->pdata->sd_head_align > 4 ?
1821 bus->sdiodev->pdata->sd_head_align : 4;
1822 sg_align = bus->sdiodev->pdata->sd_sgentry_align > 4 ?
1823 bus->sdiodev->pdata->sd_sgentry_align : 4;
1824 }
1825 /* sg entry alignment should be a divisor of block size */
1826 WARN_ON(blksize % sg_align);
1827 1959
1828 pkt_next = pktq->next; 1960 pkt_next = pktq->next;
1829 dat_buf = (u8 *)(pkt_next->data); 1961 dat_buf = (u8 *)(pkt_next->data);
@@ -1842,40 +1974,20 @@ brcmf_sdio_txpkt_prep(struct brcmf_sdio *bus, struct sk_buff_head *pktq,
1842 memset(dat_buf, 0, head_pad + bus->tx_hdrlen); 1974 memset(dat_buf, 0, head_pad + bus->tx_hdrlen);
1843 } 1975 }
1844 1976
1845 /* Check tail padding */ 1977 if (bus->sdiodev->sg_support && pktq->qlen > 1) {
1846 pkt_new = NULL; 1978 err = brcmf_sdio_txpkt_prep_sg(bus->sdiodev, pktq,
1847 tail_chop = pkt_next->len % sg_align; 1979 pkt_next, chan);
1848 tail_pad = sg_align - tail_chop; 1980 if (err < 0)
1849 tail_pad += blksize - (pkt_next->len + tail_pad) % blksize; 1981 return err;
1850 if (skb_tailroom(pkt_next) < tail_pad && pkt_next->len > blksize) { 1982 hd_info.len = (u16)err;
1851 pkt_new = brcmu_pkt_buf_get_skb(tail_pad + tail_chop);
1852 if (pkt_new == NULL)
1853 return -ENOMEM;
1854 memcpy(pkt_new->data,
1855 pkt_next->data + pkt_next->len - tail_chop,
1856 tail_chop);
1857 *(u32 *)(pkt_new->cb) = DUMMY_SKB_FLAG + tail_chop;
1858 skb_trim(pkt_next, pkt_next->len - tail_chop);
1859 __skb_queue_after(pktq, pkt_next, pkt_new);
1860 } else { 1983 } else {
1861 ntail = pkt_next->data_len + tail_pad - 1984 hd_info.len = pkt_next->len;
1862 (pkt_next->end - pkt_next->tail);
1863 if (skb_cloned(pkt_next) || ntail > 0)
1864 if (pskb_expand_head(pkt_next, 0, ntail, GFP_ATOMIC))
1865 return -ENOMEM;
1866 if (skb_linearize(pkt_next))
1867 return -ENOMEM;
1868 dat_buf = (u8 *)(pkt_next->data);
1869 __skb_put(pkt_next, tail_pad);
1870 } 1985 }
1871 1986
1872 /* Now prep the header */
1873 if (pkt_new)
1874 hd_info.len = pkt_next->len + tail_chop;
1875 else
1876 hd_info.len = pkt_next->len - tail_pad;
1877 hd_info.channel = chan; 1987 hd_info.channel = chan;
1878 hd_info.dat_offset = head_pad + bus->tx_hdrlen; 1988 hd_info.dat_offset = head_pad + bus->tx_hdrlen;
1989
1990 /* Now fill the header */
1879 brcmf_sdio_hdpack(bus, dat_buf, &hd_info); 1991 brcmf_sdio_hdpack(bus, dat_buf, &hd_info);
1880 1992
1881 if (BRCMF_BYTES_ON() && 1993 if (BRCMF_BYTES_ON() &&
@@ -1908,8 +2020,8 @@ brcmf_sdio_txpkt_postp(struct brcmf_sdio *bus, struct sk_buff_head *pktq)
1908 2020
1909 skb_queue_walk_safe(pktq, pkt_next, tmp) { 2021 skb_queue_walk_safe(pktq, pkt_next, tmp) {
1910 dummy_flags = *(u32 *)(pkt_next->cb); 2022 dummy_flags = *(u32 *)(pkt_next->cb);
1911 if (dummy_flags & DUMMY_SKB_FLAG) { 2023 if (dummy_flags & ALIGN_SKB_FLAG) {
1912 chop_len = dummy_flags & DUMMY_SKB_CHOP_LEN_MASK; 2024 chop_len = dummy_flags & ALIGN_SKB_CHOP_LEN_MASK;
1913 if (chop_len) { 2025 if (chop_len) {
1914 pkt_prev = pkt_next->prev; 2026 pkt_prev = pkt_next->prev;
1915 memcpy(pkt_prev->data + pkt_prev->len, 2027 memcpy(pkt_prev->data + pkt_prev->len,
@@ -3037,69 +3149,43 @@ static bool brcmf_sdbrcm_download_state(struct brcmf_sdio *bus, bool enter)
3037 return true; 3149 return true;
3038} 3150}
3039 3151
3040static int brcmf_sdbrcm_get_image(char *buf, int len, struct brcmf_sdio *bus)
3041{
3042 if (bus->firmware->size < bus->fw_ptr + len)
3043 len = bus->firmware->size - bus->fw_ptr;
3044
3045 memcpy(buf, &bus->firmware->data[bus->fw_ptr], len);
3046 bus->fw_ptr += len;
3047 return len;
3048}
3049
3050static int brcmf_sdbrcm_download_code_file(struct brcmf_sdio *bus) 3152static int brcmf_sdbrcm_download_code_file(struct brcmf_sdio *bus)
3051{ 3153{
3154 const struct firmware *fw;
3155 int err;
3052 int offset; 3156 int offset;
3053 uint len; 3157 int address;
3054 u8 *memblock = NULL, *memptr; 3158 int len;
3055 int ret; 3159
3056 u8 idx; 3160 fw = brcmf_sdbrcm_get_fw(bus, BRCMF_FIRMWARE_BIN);
3057 3161 if (fw == NULL)
3058 brcmf_dbg(INFO, "Enter\n"); 3162 return -ENOENT;
3059 3163
3060 ret = request_firmware(&bus->firmware, BRCMF_SDIO_FW_NAME, 3164 if (brcmf_sdio_chip_getinfidx(bus->ci, BCMA_CORE_ARM_CR4) !=
3061 &bus->sdiodev->func[2]->dev); 3165 BRCMF_MAX_CORENUM)
3062 if (ret) { 3166 memcpy(&bus->ci->rst_vec, fw->data, sizeof(bus->ci->rst_vec));
3063 brcmf_err("Fail to request firmware %d\n", ret); 3167
3064 return ret; 3168 err = 0;
3065 } 3169 offset = 0;
3066 bus->fw_ptr = 0; 3170 address = bus->ci->rambase;
3067 3171 while (offset < fw->size) {
3068 memptr = memblock = kmalloc(MEMBLOCK + BRCMF_SDALIGN, GFP_ATOMIC); 3172 len = ((offset + MEMBLOCK) < fw->size) ? MEMBLOCK :
3069 if (memblock == NULL) { 3173 fw->size - offset;
3070 ret = -ENOMEM; 3174 err = brcmf_sdio_ramrw(bus->sdiodev, true, address,
3071 goto err; 3175 (u8 *)&fw->data[offset], len);
3072 } 3176 if (err) {
3073 if ((u32)(unsigned long)memblock % BRCMF_SDALIGN)
3074 memptr += (BRCMF_SDALIGN -
3075 ((u32)(unsigned long)memblock % BRCMF_SDALIGN));
3076
3077 offset = bus->ci->rambase;
3078
3079 /* Download image */
3080 len = brcmf_sdbrcm_get_image((char *)memptr, MEMBLOCK, bus);
3081 idx = brcmf_sdio_chip_getinfidx(bus->ci, BCMA_CORE_ARM_CR4);
3082 if (BRCMF_MAX_CORENUM != idx)
3083 memcpy(&bus->ci->rst_vec, memptr, sizeof(bus->ci->rst_vec));
3084 while (len) {
3085 ret = brcmf_sdio_ramrw(bus->sdiodev, true, offset, memptr, len);
3086 if (ret) {
3087 brcmf_err("error %d on writing %d membytes at 0x%08x\n", 3177 brcmf_err("error %d on writing %d membytes at 0x%08x\n",
3088 ret, MEMBLOCK, offset); 3178 err, len, address);
3089 goto err; 3179 goto failure;
3090 } 3180 }
3091 3181 offset += len;
3092 offset += MEMBLOCK; 3182 address += len;
3093 len = brcmf_sdbrcm_get_image((char *)memptr, MEMBLOCK, bus);
3094 } 3183 }
3095 3184
3096err: 3185failure:
3097 kfree(memblock); 3186 release_firmware(fw);
3098
3099 release_firmware(bus->firmware);
3100 bus->fw_ptr = 0;
3101 3187
3102 return ret; 3188 return err;
3103} 3189}
3104 3190
3105/* 3191/*
@@ -3111,7 +3197,8 @@ err:
3111 * by two NULs. 3197 * by two NULs.
3112*/ 3198*/
3113 3199
3114static int brcmf_process_nvram_vars(struct brcmf_sdio *bus) 3200static int brcmf_process_nvram_vars(struct brcmf_sdio *bus,
3201 const struct firmware *nv)
3115{ 3202{
3116 char *varbuf; 3203 char *varbuf;
3117 char *dp; 3204 char *dp;
@@ -3120,12 +3207,12 @@ static int brcmf_process_nvram_vars(struct brcmf_sdio *bus)
3120 int ret = 0; 3207 int ret = 0;
3121 uint buf_len, n, len; 3208 uint buf_len, n, len;
3122 3209
3123 len = bus->firmware->size; 3210 len = nv->size;
3124 varbuf = vmalloc(len); 3211 varbuf = vmalloc(len);
3125 if (!varbuf) 3212 if (!varbuf)
3126 return -ENOMEM; 3213 return -ENOMEM;
3127 3214
3128 memcpy(varbuf, bus->firmware->data, len); 3215 memcpy(varbuf, nv->data, len);
3129 dp = varbuf; 3216 dp = varbuf;
3130 3217
3131 findNewline = false; 3218 findNewline = false;
@@ -3177,18 +3264,16 @@ err:
3177 3264
3178static int brcmf_sdbrcm_download_nvram(struct brcmf_sdio *bus) 3265static int brcmf_sdbrcm_download_nvram(struct brcmf_sdio *bus)
3179{ 3266{
3267 const struct firmware *nv;
3180 int ret; 3268 int ret;
3181 3269
3182 ret = request_firmware(&bus->firmware, BRCMF_SDIO_NV_NAME, 3270 nv = brcmf_sdbrcm_get_fw(bus, BRCMF_FIRMWARE_NVRAM);
3183 &bus->sdiodev->func[2]->dev); 3271 if (nv == NULL)
3184 if (ret) { 3272 return -ENOENT;
3185 brcmf_err("Fail to request nvram %d\n", ret);
3186 return ret;
3187 }
3188 3273
3189 ret = brcmf_process_nvram_vars(bus); 3274 ret = brcmf_process_nvram_vars(bus, nv);
3190 3275
3191 release_firmware(bus->firmware); 3276 release_firmware(nv);
3192 3277
3193 return ret; 3278 return ret;
3194} 3279}
diff --git a/drivers/net/wireless/brcm80211/brcmfmac/fweh.h b/drivers/net/wireless/brcm80211/brcmfmac/fweh.h
index e679214b3c98..14bc24dc5bae 100644
--- a/drivers/net/wireless/brcm80211/brcmfmac/fweh.h
+++ b/drivers/net/wireless/brcm80211/brcmfmac/fweh.h
@@ -102,7 +102,8 @@ struct brcmf_event;
102 BRCMF_ENUM_DEF(DCS_REQUEST, 73) \ 102 BRCMF_ENUM_DEF(DCS_REQUEST, 73) \
103 BRCMF_ENUM_DEF(FIFO_CREDIT_MAP, 74) \ 103 BRCMF_ENUM_DEF(FIFO_CREDIT_MAP, 74) \
104 BRCMF_ENUM_DEF(ACTION_FRAME_RX, 75) \ 104 BRCMF_ENUM_DEF(ACTION_FRAME_RX, 75) \
105 BRCMF_ENUM_DEF(BCMC_CREDIT_SUPPORT, 127) 105 BRCMF_ENUM_DEF(BCMC_CREDIT_SUPPORT, 127) \
106 BRCMF_ENUM_DEF(PSTA_PRIMARY_INTF_IND, 128)
106 107
107#define BRCMF_ENUM_DEF(id, val) \ 108#define BRCMF_ENUM_DEF(id, val) \
108 BRCMF_E_##id = (val), 109 BRCMF_E_##id = (val),
@@ -114,6 +115,8 @@ enum brcmf_fweh_event_code {
114}; 115};
115#undef BRCMF_ENUM_DEF 116#undef BRCMF_ENUM_DEF
116 117
118#define BRCMF_EVENTING_MASK_LEN DIV_ROUND_UP(BRCMF_E_LAST, 8)
119
117/* flags field values in struct brcmf_event_msg */ 120/* flags field values in struct brcmf_event_msg */
118#define BRCMF_EVENT_MSG_LINK 0x01 121#define BRCMF_EVENT_MSG_LINK 0x01
119#define BRCMF_EVENT_MSG_FLUSHTXQ 0x02 122#define BRCMF_EVENT_MSG_FLUSHTXQ 0x02
diff --git a/drivers/net/wireless/brcm80211/brcmfmac/fwsignal.c b/drivers/net/wireless/brcm80211/brcmfmac/fwsignal.c
index 82f9140f3d35..d0cd0bf95c5a 100644
--- a/drivers/net/wireless/brcm80211/brcmfmac/fwsignal.c
+++ b/drivers/net/wireless/brcm80211/brcmfmac/fwsignal.c
@@ -168,6 +168,7 @@ enum brcmf_fws_skb_state {
168/** 168/**
169 * struct brcmf_skbuff_cb - control buffer associated with skbuff. 169 * struct brcmf_skbuff_cb - control buffer associated with skbuff.
170 * 170 *
171 * @bus_flags: 2 bytes reserved for bus specific parameters
171 * @if_flags: holds interface index and packet related flags. 172 * @if_flags: holds interface index and packet related flags.
172 * @htod: host to device packet identifier (used in PKTTAG tlv). 173 * @htod: host to device packet identifier (used in PKTTAG tlv).
173 * @state: transmit state of the packet. 174 * @state: transmit state of the packet.
@@ -177,6 +178,7 @@ enum brcmf_fws_skb_state {
177 * provides 48 bytes of storage so this structure should not exceed that. 178 * provides 48 bytes of storage so this structure should not exceed that.
178 */ 179 */
179struct brcmf_skbuff_cb { 180struct brcmf_skbuff_cb {
181 u16 bus_flags;
180 u16 if_flags; 182 u16 if_flags;
181 u32 htod; 183 u32 htod;
182 enum brcmf_fws_skb_state state; 184 enum brcmf_fws_skb_state state;
diff --git a/drivers/net/wireless/brcm80211/brcmfmac/sdio_chip.c b/drivers/net/wireless/brcm80211/brcmfmac/sdio_chip.c
index ca72177388b9..2096a14ef1fb 100644
--- a/drivers/net/wireless/brcm80211/brcmfmac/sdio_chip.c
+++ b/drivers/net/wireless/brcm80211/brcmfmac/sdio_chip.c
@@ -18,6 +18,7 @@
18#include <linux/types.h> 18#include <linux/types.h>
19#include <linux/netdevice.h> 19#include <linux/netdevice.h>
20#include <linux/mmc/card.h> 20#include <linux/mmc/card.h>
21#include <linux/mmc/sdio_func.h>
21#include <linux/ssb/ssb_regs.h> 22#include <linux/ssb/ssb_regs.h>
22#include <linux/bcma/bcma.h> 23#include <linux/bcma/bcma.h>
23 24
@@ -136,6 +137,8 @@ brcmf_sdio_sb_iscoreup(struct brcmf_sdio_dev *sdiodev,
136 u8 idx; 137 u8 idx;
137 138
138 idx = brcmf_sdio_chip_getinfidx(ci, coreid); 139 idx = brcmf_sdio_chip_getinfidx(ci, coreid);
140 if (idx == BRCMF_MAX_CORENUM)
141 return false;
139 142
140 regdata = brcmf_sdio_regrl(sdiodev, 143 regdata = brcmf_sdio_regrl(sdiodev,
141 CORE_SB(ci->c_inf[idx].base, sbtmstatelow), 144 CORE_SB(ci->c_inf[idx].base, sbtmstatelow),
@@ -154,6 +157,8 @@ brcmf_sdio_ai_iscoreup(struct brcmf_sdio_dev *sdiodev,
154 bool ret; 157 bool ret;
155 158
156 idx = brcmf_sdio_chip_getinfidx(ci, coreid); 159 idx = brcmf_sdio_chip_getinfidx(ci, coreid);
160 if (idx == BRCMF_MAX_CORENUM)
161 return false;
157 162
158 regdata = brcmf_sdio_regrl(sdiodev, ci->c_inf[idx].wrapbase+BCMA_IOCTL, 163 regdata = brcmf_sdio_regrl(sdiodev, ci->c_inf[idx].wrapbase+BCMA_IOCTL,
159 NULL); 164 NULL);
@@ -261,6 +266,8 @@ brcmf_sdio_ai_coredisable(struct brcmf_sdio_dev *sdiodev,
261 u32 regdata; 266 u32 regdata;
262 267
263 idx = brcmf_sdio_chip_getinfidx(ci, coreid); 268 idx = brcmf_sdio_chip_getinfidx(ci, coreid);
269 if (idx == BRCMF_MAX_CORENUM)
270 return;
264 271
265 /* if core is already in reset, just return */ 272 /* if core is already in reset, just return */
266 regdata = brcmf_sdio_regrl(sdiodev, 273 regdata = brcmf_sdio_regrl(sdiodev,
@@ -304,6 +311,8 @@ brcmf_sdio_sb_resetcore(struct brcmf_sdio_dev *sdiodev,
304 u8 idx; 311 u8 idx;
305 312
306 idx = brcmf_sdio_chip_getinfidx(ci, coreid); 313 idx = brcmf_sdio_chip_getinfidx(ci, coreid);
314 if (idx == BRCMF_MAX_CORENUM)
315 return;
307 316
308 /* 317 /*
309 * Must do the disable sequence first to work for 318 * Must do the disable sequence first to work for
@@ -368,6 +377,8 @@ brcmf_sdio_ai_resetcore(struct brcmf_sdio_dev *sdiodev,
368 u32 regdata; 377 u32 regdata;
369 378
370 idx = brcmf_sdio_chip_getinfidx(ci, coreid); 379 idx = brcmf_sdio_chip_getinfidx(ci, coreid);
380 if (idx == BRCMF_MAX_CORENUM)
381 return;
371 382
372 /* must disable first to work for arbitrary current core state */ 383 /* must disable first to work for arbitrary current core state */
373 brcmf_sdio_ai_coredisable(sdiodev, ci, coreid, core_bits); 384 brcmf_sdio_ai_coredisable(sdiodev, ci, coreid, core_bits);
@@ -444,6 +455,9 @@ static int brcmf_sdio_chip_recognition(struct brcmf_sdio_dev *sdiodev,
444 NULL); 455 NULL);
445 ci->chip = regdata & CID_ID_MASK; 456 ci->chip = regdata & CID_ID_MASK;
446 ci->chiprev = (regdata & CID_REV_MASK) >> CID_REV_SHIFT; 457 ci->chiprev = (regdata & CID_REV_MASK) >> CID_REV_SHIFT;
458 if (sdiodev->func[0]->device == SDIO_DEVICE_ID_BROADCOM_4335_4339 &&
459 ci->chiprev >= 2)
460 ci->chip = BCM4339_CHIP_ID;
447 ci->socitype = (regdata & CID_TYPE_MASK) >> CID_TYPE_SHIFT; 461 ci->socitype = (regdata & CID_TYPE_MASK) >> CID_TYPE_SHIFT;
448 462
449 brcmf_dbg(INFO, "chipid=0x%x chiprev=%d\n", ci->chip, ci->chiprev); 463 brcmf_dbg(INFO, "chipid=0x%x chiprev=%d\n", ci->chip, ci->chiprev);
@@ -541,6 +555,20 @@ static int brcmf_sdio_chip_recognition(struct brcmf_sdio_dev *sdiodev,
541 ci->ramsize = 0xc0000; 555 ci->ramsize = 0xc0000;
542 ci->rambase = 0x180000; 556 ci->rambase = 0x180000;
543 break; 557 break;
558 case BCM4339_CHIP_ID:
559 ci->c_inf[0].wrapbase = 0x18100000;
560 ci->c_inf[0].cib = 0x2e084411;
561 ci->c_inf[1].id = BCMA_CORE_SDIO_DEV;
562 ci->c_inf[1].base = 0x18005000;
563 ci->c_inf[1].wrapbase = 0x18105000;
564 ci->c_inf[1].cib = 0x15004211;
565 ci->c_inf[2].id = BCMA_CORE_ARM_CR4;
566 ci->c_inf[2].base = 0x18002000;
567 ci->c_inf[2].wrapbase = 0x18102000;
568 ci->c_inf[2].cib = 0x04084411;
569 ci->ramsize = 0xc0000;
570 ci->rambase = 0x180000;
571 break;
544 default: 572 default:
545 brcmf_err("chipid 0x%x is not supported\n", ci->chip); 573 brcmf_err("chipid 0x%x is not supported\n", ci->chip);
546 return -ENODEV; 574 return -ENODEV;
diff --git a/drivers/net/wireless/brcm80211/brcmfmac/sdio_chip.h b/drivers/net/wireless/brcm80211/brcmfmac/sdio_chip.h
index 83c041f1bf4a..507c61c991fa 100644
--- a/drivers/net/wireless/brcm80211/brcmfmac/sdio_chip.h
+++ b/drivers/net/wireless/brcm80211/brcmfmac/sdio_chip.h
@@ -54,6 +54,14 @@
54 54
55#define BRCMF_MAX_CORENUM 6 55#define BRCMF_MAX_CORENUM 6
56 56
57/* SDIO device ID */
58#define SDIO_DEVICE_ID_BROADCOM_43143 43143
59#define SDIO_DEVICE_ID_BROADCOM_43241 0x4324
60#define SDIO_DEVICE_ID_BROADCOM_4329 0x4329
61#define SDIO_DEVICE_ID_BROADCOM_4330 0x4330
62#define SDIO_DEVICE_ID_BROADCOM_4334 0x4334
63#define SDIO_DEVICE_ID_BROADCOM_4335_4339 0x4335
64
57struct chip_core_info { 65struct chip_core_info {
58 u16 id; 66 u16 id;
59 u16 rev; 67 u16 rev;
@@ -215,17 +223,16 @@ struct sdpcmd_regs {
215 u16 PAD[0x80]; 223 u16 PAD[0x80];
216}; 224};
217 225
218extern int brcmf_sdio_chip_attach(struct brcmf_sdio_dev *sdiodev, 226int brcmf_sdio_chip_attach(struct brcmf_sdio_dev *sdiodev,
219 struct chip_info **ci_ptr, u32 regs); 227 struct chip_info **ci_ptr, u32 regs);
220extern void brcmf_sdio_chip_detach(struct chip_info **ci_ptr); 228void brcmf_sdio_chip_detach(struct chip_info **ci_ptr);
221extern void brcmf_sdio_chip_drivestrengthinit(struct brcmf_sdio_dev *sdiodev, 229void brcmf_sdio_chip_drivestrengthinit(struct brcmf_sdio_dev *sdiodev,
222 struct chip_info *ci, 230 struct chip_info *ci, u32 drivestrength);
223 u32 drivestrength); 231u8 brcmf_sdio_chip_getinfidx(struct chip_info *ci, u16 coreid);
224extern u8 brcmf_sdio_chip_getinfidx(struct chip_info *ci, u16 coreid); 232void brcmf_sdio_chip_enter_download(struct brcmf_sdio_dev *sdiodev,
225extern void brcmf_sdio_chip_enter_download(struct brcmf_sdio_dev *sdiodev, 233 struct chip_info *ci);
226 struct chip_info *ci); 234bool brcmf_sdio_chip_exit_download(struct brcmf_sdio_dev *sdiodev,
227extern bool brcmf_sdio_chip_exit_download(struct brcmf_sdio_dev *sdiodev, 235 struct chip_info *ci, char *nvram_dat,
228 struct chip_info *ci, char *nvram_dat, 236 uint nvram_sz);
229 uint nvram_sz);
230 237
231#endif /* _BRCMFMAC_SDIO_CHIP_H_ */ 238#endif /* _BRCMFMAC_SDIO_CHIP_H_ */
diff --git a/drivers/net/wireless/brcm80211/brcmfmac/sdio_host.h b/drivers/net/wireless/brcm80211/brcmfmac/sdio_host.h
index 2b5407f002e5..fc0d4f0129db 100644
--- a/drivers/net/wireless/brcm80211/brcmfmac/sdio_host.h
+++ b/drivers/net/wireless/brcm80211/brcmfmac/sdio_host.h
@@ -178,21 +178,25 @@ struct brcmf_sdio_dev {
178 bool irq_en; /* irq enable flags */ 178 bool irq_en; /* irq enable flags */
179 spinlock_t irq_en_lock; 179 spinlock_t irq_en_lock;
180 bool irq_wake; /* irq wake enable flags */ 180 bool irq_wake; /* irq wake enable flags */
181 bool sg_support;
182 uint max_request_size;
183 ushort max_segment_count;
184 uint max_segment_size;
181}; 185};
182 186
183/* Register/deregister interrupt handler. */ 187/* Register/deregister interrupt handler. */
184extern int brcmf_sdio_intr_register(struct brcmf_sdio_dev *sdiodev); 188int brcmf_sdio_intr_register(struct brcmf_sdio_dev *sdiodev);
185extern int brcmf_sdio_intr_unregister(struct brcmf_sdio_dev *sdiodev); 189int brcmf_sdio_intr_unregister(struct brcmf_sdio_dev *sdiodev);
186 190
187/* sdio device register access interface */ 191/* sdio device register access interface */
188extern u8 brcmf_sdio_regrb(struct brcmf_sdio_dev *sdiodev, u32 addr, int *ret); 192u8 brcmf_sdio_regrb(struct brcmf_sdio_dev *sdiodev, u32 addr, int *ret);
189extern u32 brcmf_sdio_regrl(struct brcmf_sdio_dev *sdiodev, u32 addr, int *ret); 193u32 brcmf_sdio_regrl(struct brcmf_sdio_dev *sdiodev, u32 addr, int *ret);
190extern void brcmf_sdio_regwb(struct brcmf_sdio_dev *sdiodev, u32 addr, 194void brcmf_sdio_regwb(struct brcmf_sdio_dev *sdiodev, u32 addr, u8 data,
191 u8 data, int *ret); 195 int *ret);
192extern void brcmf_sdio_regwl(struct brcmf_sdio_dev *sdiodev, u32 addr, 196void brcmf_sdio_regwl(struct brcmf_sdio_dev *sdiodev, u32 addr, u32 data,
193 u32 data, int *ret); 197 int *ret);
194extern int brcmf_sdio_regrw_helper(struct brcmf_sdio_dev *sdiodev, u32 addr, 198int brcmf_sdio_regrw_helper(struct brcmf_sdio_dev *sdiodev, u32 addr,
195 void *data, bool write); 199 void *data, bool write);
196 200
197/* Buffer transfer to/from device (client) core via cmd53. 201/* Buffer transfer to/from device (client) core via cmd53.
198 * fn: function number 202 * fn: function number
@@ -206,22 +210,17 @@ extern int brcmf_sdio_regrw_helper(struct brcmf_sdio_dev *sdiodev, u32 addr,
206 * Returns 0 or error code. 210 * Returns 0 or error code.
207 * NOTE: Async operation is not currently supported. 211 * NOTE: Async operation is not currently supported.
208 */ 212 */
209extern int 213int brcmf_sdcard_send_pkt(struct brcmf_sdio_dev *sdiodev, u32 addr, uint fn,
210brcmf_sdcard_send_pkt(struct brcmf_sdio_dev *sdiodev, u32 addr, uint fn, 214 uint flags, struct sk_buff_head *pktq);
211 uint flags, struct sk_buff_head *pktq); 215int brcmf_sdcard_send_buf(struct brcmf_sdio_dev *sdiodev, u32 addr, uint fn,
212extern int 216 uint flags, u8 *buf, uint nbytes);
213brcmf_sdcard_send_buf(struct brcmf_sdio_dev *sdiodev, u32 addr, uint fn, 217
214 uint flags, u8 *buf, uint nbytes); 218int brcmf_sdcard_recv_pkt(struct brcmf_sdio_dev *sdiodev, u32 addr, uint fn,
215 219 uint flags, struct sk_buff *pkt);
216extern int 220int brcmf_sdcard_recv_buf(struct brcmf_sdio_dev *sdiodev, u32 addr, uint fn,
217brcmf_sdcard_recv_pkt(struct brcmf_sdio_dev *sdiodev, u32 addr, uint fn, 221 uint flags, u8 *buf, uint nbytes);
218 uint flags, struct sk_buff *pkt); 222int brcmf_sdcard_recv_chain(struct brcmf_sdio_dev *sdiodev, u32 addr, uint fn,
219extern int 223 uint flags, struct sk_buff_head *pktq, uint totlen);
220brcmf_sdcard_recv_buf(struct brcmf_sdio_dev *sdiodev, u32 addr, uint fn,
221 uint flags, u8 *buf, uint nbytes);
222extern int
223brcmf_sdcard_recv_chain(struct brcmf_sdio_dev *sdiodev, u32 addr, uint fn,
224 uint flags, struct sk_buff_head *pktq);
225 224
226/* Flags bits */ 225/* Flags bits */
227 226
@@ -237,46 +236,43 @@ brcmf_sdcard_recv_chain(struct brcmf_sdio_dev *sdiodev, u32 addr, uint fn,
237 * nbytes: number of bytes to transfer to/from buf 236 * nbytes: number of bytes to transfer to/from buf
238 * Returns 0 or error code. 237 * Returns 0 or error code.
239 */ 238 */
240extern int brcmf_sdcard_rwdata(struct brcmf_sdio_dev *sdiodev, uint rw, 239int brcmf_sdcard_rwdata(struct brcmf_sdio_dev *sdiodev, uint rw, u32 addr,
241 u32 addr, u8 *buf, uint nbytes); 240 u8 *buf, uint nbytes);
242extern int brcmf_sdio_ramrw(struct brcmf_sdio_dev *sdiodev, bool write, 241int brcmf_sdio_ramrw(struct brcmf_sdio_dev *sdiodev, bool write, u32 address,
243 u32 address, u8 *data, uint size); 242 u8 *data, uint size);
244 243
245/* Issue an abort to the specified function */ 244/* Issue an abort to the specified function */
246extern int brcmf_sdcard_abort(struct brcmf_sdio_dev *sdiodev, uint fn); 245int brcmf_sdcard_abort(struct brcmf_sdio_dev *sdiodev, uint fn);
247 246
248/* platform specific/high level functions */ 247/* platform specific/high level functions */
249extern int brcmf_sdio_probe(struct brcmf_sdio_dev *sdiodev); 248int brcmf_sdio_probe(struct brcmf_sdio_dev *sdiodev);
250extern int brcmf_sdio_remove(struct brcmf_sdio_dev *sdiodev); 249int brcmf_sdio_remove(struct brcmf_sdio_dev *sdiodev);
251 250
252/* attach, return handler on success, NULL if failed. 251/* attach, return handler on success, NULL if failed.
253 * The handler shall be provided by all subsequent calls. No local cache 252 * The handler shall be provided by all subsequent calls. No local cache
254 * cfghdl points to the starting address of pci device mapped memory 253 * cfghdl points to the starting address of pci device mapped memory
255 */ 254 */
256extern int brcmf_sdioh_attach(struct brcmf_sdio_dev *sdiodev); 255int brcmf_sdioh_attach(struct brcmf_sdio_dev *sdiodev);
257extern void brcmf_sdioh_detach(struct brcmf_sdio_dev *sdiodev); 256void brcmf_sdioh_detach(struct brcmf_sdio_dev *sdiodev);
258 257
259/* read or write one byte using cmd52 */ 258/* read or write one byte using cmd52 */
260extern int brcmf_sdioh_request_byte(struct brcmf_sdio_dev *sdiodev, uint rw, 259int brcmf_sdioh_request_byte(struct brcmf_sdio_dev *sdiodev, uint rw, uint fnc,
261 uint fnc, uint addr, u8 *byte); 260 uint addr, u8 *byte);
262 261
263/* read or write 2/4 bytes using cmd53 */ 262/* read or write 2/4 bytes using cmd53 */
264extern int 263int brcmf_sdioh_request_word(struct brcmf_sdio_dev *sdiodev, uint rw, uint fnc,
265brcmf_sdioh_request_word(struct brcmf_sdio_dev *sdiodev, 264 uint addr, u32 *word, uint nbyte);
266 uint rw, uint fnc, uint addr,
267 u32 *word, uint nbyte);
268 265
269/* Watchdog timer interface for pm ops */ 266/* Watchdog timer interface for pm ops */
270extern void brcmf_sdio_wdtmr_enable(struct brcmf_sdio_dev *sdiodev, 267void brcmf_sdio_wdtmr_enable(struct brcmf_sdio_dev *sdiodev, bool enable);
271 bool enable);
272 268
273extern void *brcmf_sdbrcm_probe(u32 regsva, struct brcmf_sdio_dev *sdiodev); 269void *brcmf_sdbrcm_probe(u32 regsva, struct brcmf_sdio_dev *sdiodev);
274extern void brcmf_sdbrcm_disconnect(void *ptr); 270void brcmf_sdbrcm_disconnect(void *ptr);
275extern void brcmf_sdbrcm_isr(void *arg); 271void brcmf_sdbrcm_isr(void *arg);
276 272
277extern void brcmf_sdbrcm_wd_timer(struct brcmf_sdio *bus, uint wdtick); 273void brcmf_sdbrcm_wd_timer(struct brcmf_sdio *bus, uint wdtick);
278 274
279extern void brcmf_pm_resume_wait(struct brcmf_sdio_dev *sdiodev, 275void brcmf_pm_resume_wait(struct brcmf_sdio_dev *sdiodev,
280 wait_queue_head_t *wq); 276 wait_queue_head_t *wq);
281extern bool brcmf_pm_resume_error(struct brcmf_sdio_dev *sdiodev); 277bool brcmf_pm_resume_error(struct brcmf_sdio_dev *sdiodev);
282#endif /* _BRCM_SDH_H_ */ 278#endif /* _BRCM_SDH_H_ */
diff --git a/drivers/net/wireless/brcm80211/brcmfmac/tracepoint.h b/drivers/net/wireless/brcm80211/brcmfmac/tracepoint.h
index bc2917112899..3c67529b9074 100644
--- a/drivers/net/wireless/brcm80211/brcmfmac/tracepoint.h
+++ b/drivers/net/wireless/brcm80211/brcmfmac/tracepoint.h
@@ -78,13 +78,15 @@ TRACE_EVENT(brcmf_hexdump,
78 TP_ARGS(data, len), 78 TP_ARGS(data, len),
79 TP_STRUCT__entry( 79 TP_STRUCT__entry(
80 __field(unsigned long, len) 80 __field(unsigned long, len)
81 __field(unsigned long, addr)
81 __dynamic_array(u8, hdata, len) 82 __dynamic_array(u8, hdata, len)
82 ), 83 ),
83 TP_fast_assign( 84 TP_fast_assign(
84 __entry->len = len; 85 __entry->len = len;
86 __entry->addr = (unsigned long)data;
85 memcpy(__get_dynamic_array(hdata), data, len); 87 memcpy(__get_dynamic_array(hdata), data, len);
86 ), 88 ),
87 TP_printk("hexdump [length=%lu]", __entry->len) 89 TP_printk("hexdump [addr=%lx, length=%lu]", __entry->addr, __entry->len)
88); 90);
89 91
90TRACE_EVENT(brcmf_bdchdr, 92TRACE_EVENT(brcmf_bdchdr,
@@ -108,6 +110,23 @@ TRACE_EVENT(brcmf_bdchdr,
108 TP_printk("bdc: prio=%d siglen=%d", __entry->prio, __entry->siglen) 110 TP_printk("bdc: prio=%d siglen=%d", __entry->prio, __entry->siglen)
109); 111);
110 112
113TRACE_EVENT(brcmf_sdpcm_hdr,
114 TP_PROTO(bool tx, void *data),
115 TP_ARGS(tx, data),
116 TP_STRUCT__entry(
117 __field(u8, tx)
118 __field(u16, len)
119 __array(u8, hdr, 12)
120 ),
121 TP_fast_assign(
122 memcpy(__entry->hdr, data, 12);
123 __entry->len = __entry->hdr[0] | (__entry->hdr[1] << 8);
124 __entry->tx = tx ? 1 : 0;
125 ),
126 TP_printk("sdpcm: %s len %u, seq %d", __entry->tx ? "TX" : "RX",
127 __entry->len, __entry->hdr[4])
128);
129
111#ifdef CONFIG_BRCM_TRACING 130#ifdef CONFIG_BRCM_TRACING
112 131
113#undef TRACE_INCLUDE_PATH 132#undef TRACE_INCLUDE_PATH
diff --git a/drivers/net/wireless/brcm80211/brcmfmac/usb.c b/drivers/net/wireless/brcm80211/brcmfmac/usb.c
index f4aea47e0730..422f44c63175 100644
--- a/drivers/net/wireless/brcm80211/brcmfmac/usb.c
+++ b/drivers/net/wireless/brcm80211/brcmfmac/usb.c
@@ -435,7 +435,6 @@ static void brcmf_usb_rx_complete(struct urb *urb)
435 struct brcmf_usbreq *req = (struct brcmf_usbreq *)urb->context; 435 struct brcmf_usbreq *req = (struct brcmf_usbreq *)urb->context;
436 struct brcmf_usbdev_info *devinfo = req->devinfo; 436 struct brcmf_usbdev_info *devinfo = req->devinfo;
437 struct sk_buff *skb; 437 struct sk_buff *skb;
438 struct sk_buff_head skbq;
439 438
440 brcmf_dbg(USB, "Enter, urb->status=%d\n", urb->status); 439 brcmf_dbg(USB, "Enter, urb->status=%d\n", urb->status);
441 brcmf_usb_del_fromq(devinfo, req); 440 brcmf_usb_del_fromq(devinfo, req);
@@ -450,10 +449,8 @@ static void brcmf_usb_rx_complete(struct urb *urb)
450 } 449 }
451 450
452 if (devinfo->bus_pub.state == BRCMFMAC_USB_STATE_UP) { 451 if (devinfo->bus_pub.state == BRCMFMAC_USB_STATE_UP) {
453 skb_queue_head_init(&skbq);
454 skb_queue_tail(&skbq, skb);
455 skb_put(skb, urb->actual_length); 452 skb_put(skb, urb->actual_length);
456 brcmf_rx_frames(devinfo->dev, &skbq); 453 brcmf_rx_frame(devinfo->dev, skb);
457 brcmf_usb_rx_refill(devinfo, req); 454 brcmf_usb_rx_refill(devinfo, req);
458 } else { 455 } else {
459 brcmu_pkt_buf_free_skb(skb); 456 brcmu_pkt_buf_free_skb(skb);
diff --git a/drivers/net/wireless/brcm80211/brcmsmac/aiutils.h b/drivers/net/wireless/brcm80211/brcmsmac/aiutils.h
index a8a267b5b87a..2d08c155c23b 100644
--- a/drivers/net/wireless/brcm80211/brcmsmac/aiutils.h
+++ b/drivers/net/wireless/brcm80211/brcmsmac/aiutils.h
@@ -172,19 +172,19 @@ struct si_info {
172 172
173 173
174/* AMBA Interconnect exported externs */ 174/* AMBA Interconnect exported externs */
175extern u32 ai_core_cflags(struct bcma_device *core, u32 mask, u32 val); 175u32 ai_core_cflags(struct bcma_device *core, u32 mask, u32 val);
176 176
177/* === exported functions === */ 177/* === exported functions === */
178extern struct si_pub *ai_attach(struct bcma_bus *pbus); 178struct si_pub *ai_attach(struct bcma_bus *pbus);
179extern void ai_detach(struct si_pub *sih); 179void ai_detach(struct si_pub *sih);
180extern uint ai_cc_reg(struct si_pub *sih, uint regoff, u32 mask, u32 val); 180uint ai_cc_reg(struct si_pub *sih, uint regoff, u32 mask, u32 val);
181extern void ai_clkctl_init(struct si_pub *sih); 181void ai_clkctl_init(struct si_pub *sih);
182extern u16 ai_clkctl_fast_pwrup_delay(struct si_pub *sih); 182u16 ai_clkctl_fast_pwrup_delay(struct si_pub *sih);
183extern bool ai_clkctl_cc(struct si_pub *sih, enum bcma_clkmode mode); 183bool ai_clkctl_cc(struct si_pub *sih, enum bcma_clkmode mode);
184extern bool ai_deviceremoved(struct si_pub *sih); 184bool ai_deviceremoved(struct si_pub *sih);
185 185
186/* Enable Ex-PA for 4313 */ 186/* Enable Ex-PA for 4313 */
187extern void ai_epa_4313war(struct si_pub *sih); 187void ai_epa_4313war(struct si_pub *sih);
188 188
189static inline u32 ai_get_cccaps(struct si_pub *sih) 189static inline u32 ai_get_cccaps(struct si_pub *sih)
190{ 190{
diff --git a/drivers/net/wireless/brcm80211/brcmsmac/ampdu.h b/drivers/net/wireless/brcm80211/brcmsmac/ampdu.h
index 73d01e586109..03bdcf29bd50 100644
--- a/drivers/net/wireless/brcm80211/brcmsmac/ampdu.h
+++ b/drivers/net/wireless/brcm80211/brcmsmac/ampdu.h
@@ -37,17 +37,17 @@ struct brcms_ampdu_session {
37 u16 dma_len; 37 u16 dma_len;
38}; 38};
39 39
40extern void brcms_c_ampdu_reset_session(struct brcms_ampdu_session *session, 40void brcms_c_ampdu_reset_session(struct brcms_ampdu_session *session,
41 struct brcms_c_info *wlc); 41 struct brcms_c_info *wlc);
42extern int brcms_c_ampdu_add_frame(struct brcms_ampdu_session *session, 42int brcms_c_ampdu_add_frame(struct brcms_ampdu_session *session,
43 struct sk_buff *p); 43 struct sk_buff *p);
44extern void brcms_c_ampdu_finalize(struct brcms_ampdu_session *session); 44void brcms_c_ampdu_finalize(struct brcms_ampdu_session *session);
45 45
46extern struct ampdu_info *brcms_c_ampdu_attach(struct brcms_c_info *wlc); 46struct ampdu_info *brcms_c_ampdu_attach(struct brcms_c_info *wlc);
47extern void brcms_c_ampdu_detach(struct ampdu_info *ampdu); 47void brcms_c_ampdu_detach(struct ampdu_info *ampdu);
48extern void brcms_c_ampdu_dotxstatus(struct ampdu_info *ampdu, struct scb *scb, 48void brcms_c_ampdu_dotxstatus(struct ampdu_info *ampdu, struct scb *scb,
49 struct sk_buff *p, struct tx_status *txs); 49 struct sk_buff *p, struct tx_status *txs);
50extern void brcms_c_ampdu_macaddr_upd(struct brcms_c_info *wlc); 50void brcms_c_ampdu_macaddr_upd(struct brcms_c_info *wlc);
51extern void brcms_c_ampdu_shm_upd(struct ampdu_info *ampdu); 51void brcms_c_ampdu_shm_upd(struct ampdu_info *ampdu);
52 52
53#endif /* _BRCM_AMPDU_H_ */ 53#endif /* _BRCM_AMPDU_H_ */
diff --git a/drivers/net/wireless/brcm80211/brcmsmac/antsel.h b/drivers/net/wireless/brcm80211/brcmsmac/antsel.h
index 97ea3881a8ec..a3d487ab1964 100644
--- a/drivers/net/wireless/brcm80211/brcmsmac/antsel.h
+++ b/drivers/net/wireless/brcm80211/brcmsmac/antsel.h
@@ -17,13 +17,11 @@
17#ifndef _BRCM_ANTSEL_H_ 17#ifndef _BRCM_ANTSEL_H_
18#define _BRCM_ANTSEL_H_ 18#define _BRCM_ANTSEL_H_
19 19
20extern struct antsel_info *brcms_c_antsel_attach(struct brcms_c_info *wlc); 20struct antsel_info *brcms_c_antsel_attach(struct brcms_c_info *wlc);
21extern void brcms_c_antsel_detach(struct antsel_info *asi); 21void brcms_c_antsel_detach(struct antsel_info *asi);
22extern void brcms_c_antsel_init(struct antsel_info *asi); 22void brcms_c_antsel_init(struct antsel_info *asi);
23extern void brcms_c_antsel_antcfg_get(struct antsel_info *asi, bool usedef, 23void brcms_c_antsel_antcfg_get(struct antsel_info *asi, bool usedef, bool sel,
24 bool sel, 24 u8 id, u8 fbid, u8 *antcfg, u8 *fbantcfg);
25 u8 id, u8 fbid, u8 *antcfg, 25u8 brcms_c_antsel_antsel2id(struct antsel_info *asi, u16 antsel);
26 u8 *fbantcfg);
27extern u8 brcms_c_antsel_antsel2id(struct antsel_info *asi, u16 antsel);
28 26
29#endif /* _BRCM_ANTSEL_H_ */ 27#endif /* _BRCM_ANTSEL_H_ */
diff --git a/drivers/net/wireless/brcm80211/brcmsmac/channel.h b/drivers/net/wireless/brcm80211/brcmsmac/channel.h
index 006483a0abe6..39dd3a5b2979 100644
--- a/drivers/net/wireless/brcm80211/brcmsmac/channel.h
+++ b/drivers/net/wireless/brcm80211/brcmsmac/channel.h
@@ -32,20 +32,16 @@
32 32
33#define BRCMS_DFS_EU (BRCMS_DFS_TPC | BRCMS_RADAR_TYPE_EU) /* Flag for DFS EU */ 33#define BRCMS_DFS_EU (BRCMS_DFS_TPC | BRCMS_RADAR_TYPE_EU) /* Flag for DFS EU */
34 34
35extern struct brcms_cm_info * 35struct brcms_cm_info *brcms_c_channel_mgr_attach(struct brcms_c_info *wlc);
36brcms_c_channel_mgr_attach(struct brcms_c_info *wlc);
37 36
38extern void brcms_c_channel_mgr_detach(struct brcms_cm_info *wlc_cm); 37void brcms_c_channel_mgr_detach(struct brcms_cm_info *wlc_cm);
39 38
40extern bool brcms_c_valid_chanspec_db(struct brcms_cm_info *wlc_cm, 39bool brcms_c_valid_chanspec_db(struct brcms_cm_info *wlc_cm, u16 chspec);
41 u16 chspec);
42 40
43extern void brcms_c_channel_reg_limits(struct brcms_cm_info *wlc_cm, 41void brcms_c_channel_reg_limits(struct brcms_cm_info *wlc_cm, u16 chanspec,
44 u16 chanspec, 42 struct txpwr_limits *txpwr);
45 struct txpwr_limits *txpwr); 43void brcms_c_channel_set_chanspec(struct brcms_cm_info *wlc_cm, u16 chanspec,
46extern void brcms_c_channel_set_chanspec(struct brcms_cm_info *wlc_cm, 44 u8 local_constraint_qdbm);
47 u16 chanspec, 45void brcms_c_regd_init(struct brcms_c_info *wlc);
48 u8 local_constraint_qdbm);
49extern void brcms_c_regd_init(struct brcms_c_info *wlc);
50 46
51#endif /* _WLC_CHANNEL_H */ 47#endif /* _WLC_CHANNEL_H */
diff --git a/drivers/net/wireless/brcm80211/brcmsmac/mac80211_if.h b/drivers/net/wireless/brcm80211/brcmsmac/mac80211_if.h
index 4090032e81a2..198053dfc310 100644
--- a/drivers/net/wireless/brcm80211/brcmsmac/mac80211_if.h
+++ b/drivers/net/wireless/brcm80211/brcmsmac/mac80211_if.h
@@ -88,26 +88,26 @@ struct brcms_info {
88}; 88};
89 89
90/* misc callbacks */ 90/* misc callbacks */
91extern void brcms_init(struct brcms_info *wl); 91void brcms_init(struct brcms_info *wl);
92extern uint brcms_reset(struct brcms_info *wl); 92uint brcms_reset(struct brcms_info *wl);
93extern void brcms_intrson(struct brcms_info *wl); 93void brcms_intrson(struct brcms_info *wl);
94extern u32 brcms_intrsoff(struct brcms_info *wl); 94u32 brcms_intrsoff(struct brcms_info *wl);
95extern void brcms_intrsrestore(struct brcms_info *wl, u32 macintmask); 95void brcms_intrsrestore(struct brcms_info *wl, u32 macintmask);
96extern int brcms_up(struct brcms_info *wl); 96int brcms_up(struct brcms_info *wl);
97extern void brcms_down(struct brcms_info *wl); 97void brcms_down(struct brcms_info *wl);
98extern void brcms_txflowcontrol(struct brcms_info *wl, struct brcms_if *wlif, 98void brcms_txflowcontrol(struct brcms_info *wl, struct brcms_if *wlif,
99 bool state, int prio); 99 bool state, int prio);
100extern bool brcms_rfkill_set_hw_state(struct brcms_info *wl); 100bool brcms_rfkill_set_hw_state(struct brcms_info *wl);
101 101
102/* timer functions */ 102/* timer functions */
103extern struct brcms_timer *brcms_init_timer(struct brcms_info *wl, 103struct brcms_timer *brcms_init_timer(struct brcms_info *wl,
104 void (*fn) (void *arg), void *arg, 104 void (*fn) (void *arg), void *arg,
105 const char *name); 105 const char *name);
106extern void brcms_free_timer(struct brcms_timer *timer); 106void brcms_free_timer(struct brcms_timer *timer);
107extern void brcms_add_timer(struct brcms_timer *timer, uint ms, int periodic); 107void brcms_add_timer(struct brcms_timer *timer, uint ms, int periodic);
108extern bool brcms_del_timer(struct brcms_timer *timer); 108bool brcms_del_timer(struct brcms_timer *timer);
109extern void brcms_dpc(unsigned long data); 109void brcms_dpc(unsigned long data);
110extern void brcms_timer(struct brcms_timer *t); 110void brcms_timer(struct brcms_timer *t);
111extern void brcms_fatal_error(struct brcms_info *wl); 111void brcms_fatal_error(struct brcms_info *wl);
112 112
113#endif /* _BRCM_MAC80211_IF_H_ */ 113#endif /* _BRCM_MAC80211_IF_H_ */
diff --git a/drivers/net/wireless/brcm80211/brcmsmac/main.c b/drivers/net/wireless/brcm80211/brcmsmac/main.c
index 4608e0eb1493..8138f1cff4e5 100644
--- a/drivers/net/wireless/brcm80211/brcmsmac/main.c
+++ b/drivers/net/wireless/brcm80211/brcmsmac/main.c
@@ -1906,14 +1906,14 @@ static void brcms_c_get_macaddr(struct brcms_hardware *wlc_hw, u8 etheraddr[ETH_
1906 1906
1907 /* If macaddr exists, use it (Sromrev4, CIS, ...). */ 1907 /* If macaddr exists, use it (Sromrev4, CIS, ...). */
1908 if (!is_zero_ether_addr(sprom->il0mac)) { 1908 if (!is_zero_ether_addr(sprom->il0mac)) {
1909 memcpy(etheraddr, sprom->il0mac, 6); 1909 memcpy(etheraddr, sprom->il0mac, ETH_ALEN);
1910 return; 1910 return;
1911 } 1911 }
1912 1912
1913 if (wlc_hw->_nbands > 1) 1913 if (wlc_hw->_nbands > 1)
1914 memcpy(etheraddr, sprom->et1mac, 6); 1914 memcpy(etheraddr, sprom->et1mac, ETH_ALEN);
1915 else 1915 else
1916 memcpy(etheraddr, sprom->il0mac, 6); 1916 memcpy(etheraddr, sprom->il0mac, ETH_ALEN);
1917} 1917}
1918 1918
1919/* power both the pll and external oscillator on/off */ 1919/* power both the pll and external oscillator on/off */
@@ -5695,7 +5695,7 @@ static bool brcms_c_chipmatch_pci(struct bcma_device *core)
5695 return true; 5695 return true;
5696 if ((device == BCM43224_D11N_ID) || (device == BCM43225_D11N2G_ID)) 5696 if ((device == BCM43224_D11N_ID) || (device == BCM43225_D11N2G_ID))
5697 return true; 5697 return true;
5698 if (device == BCM4313_D11N2G_ID) 5698 if (device == BCM4313_D11N2G_ID || device == BCM4313_CHIP_ID)
5699 return true; 5699 return true;
5700 if ((device == BCM43236_D11N_ID) || (device == BCM43236_D11N2G_ID)) 5700 if ((device == BCM43236_D11N_ID) || (device == BCM43236_D11N2G_ID))
5701 return true; 5701 return true;
diff --git a/drivers/net/wireless/brcm80211/brcmsmac/main.h b/drivers/net/wireless/brcm80211/brcmsmac/main.h
index b5d7a38b53fe..c4d135cff04a 100644
--- a/drivers/net/wireless/brcm80211/brcmsmac/main.h
+++ b/drivers/net/wireless/brcm80211/brcmsmac/main.h
@@ -616,66 +616,54 @@ struct brcms_bss_cfg {
616 struct brcms_bss_info *current_bss; 616 struct brcms_bss_info *current_bss;
617}; 617};
618 618
619extern int brcms_c_txfifo(struct brcms_c_info *wlc, uint fifo, 619int brcms_c_txfifo(struct brcms_c_info *wlc, uint fifo, struct sk_buff *p);
620 struct sk_buff *p); 620int brcms_b_xmtfifo_sz_get(struct brcms_hardware *wlc_hw, uint fifo,
621extern int brcms_b_xmtfifo_sz_get(struct brcms_hardware *wlc_hw, uint fifo, 621 uint *blocks);
622 uint *blocks); 622
623 623int brcms_c_set_gmode(struct brcms_c_info *wlc, u8 gmode, bool config);
624extern int brcms_c_set_gmode(struct brcms_c_info *wlc, u8 gmode, bool config); 624void brcms_c_mac_promisc(struct brcms_c_info *wlc, uint filter_flags);
625extern void brcms_c_mac_promisc(struct brcms_c_info *wlc, uint filter_flags); 625u16 brcms_c_calc_lsig_len(struct brcms_c_info *wlc, u32 ratespec, uint mac_len);
626extern u16 brcms_c_calc_lsig_len(struct brcms_c_info *wlc, u32 ratespec, 626u32 brcms_c_rspec_to_rts_rspec(struct brcms_c_info *wlc, u32 rspec,
627 uint mac_len); 627 bool use_rspec, u16 mimo_ctlchbw);
628extern u32 brcms_c_rspec_to_rts_rspec(struct brcms_c_info *wlc, 628u16 brcms_c_compute_rtscts_dur(struct brcms_c_info *wlc, bool cts_only,
629 u32 rspec, 629 u32 rts_rate, u32 frame_rate,
630 bool use_rspec, u16 mimo_ctlchbw); 630 u8 rts_preamble_type, u8 frame_preamble_type,
631extern u16 brcms_c_compute_rtscts_dur(struct brcms_c_info *wlc, bool cts_only, 631 uint frame_len, bool ba);
632 u32 rts_rate, 632void brcms_c_inval_dma_pkts(struct brcms_hardware *hw,
633 u32 frame_rate, 633 struct ieee80211_sta *sta, void (*dma_callback_fn));
634 u8 rts_preamble_type, 634void brcms_c_update_probe_resp(struct brcms_c_info *wlc, bool suspend);
635 u8 frame_preamble_type, uint frame_len, 635int brcms_c_set_nmode(struct brcms_c_info *wlc);
636 bool ba); 636void brcms_c_beacon_phytxctl_txant_upd(struct brcms_c_info *wlc, u32 bcn_rate);
637extern void brcms_c_inval_dma_pkts(struct brcms_hardware *hw, 637void brcms_b_antsel_type_set(struct brcms_hardware *wlc_hw, u8 antsel_type);
638 struct ieee80211_sta *sta, 638void brcms_b_set_chanspec(struct brcms_hardware *wlc_hw, u16 chanspec,
639 void (*dma_callback_fn)); 639 bool mute, struct txpwr_limits *txpwr);
640extern void brcms_c_update_probe_resp(struct brcms_c_info *wlc, bool suspend); 640void brcms_b_write_shm(struct brcms_hardware *wlc_hw, uint offset, u16 v);
641extern int brcms_c_set_nmode(struct brcms_c_info *wlc); 641u16 brcms_b_read_shm(struct brcms_hardware *wlc_hw, uint offset);
642extern void brcms_c_beacon_phytxctl_txant_upd(struct brcms_c_info *wlc, 642void brcms_b_mhf(struct brcms_hardware *wlc_hw, u8 idx, u16 mask, u16 val,
643 u32 bcn_rate); 643 int bands);
644extern void brcms_b_antsel_type_set(struct brcms_hardware *wlc_hw, 644void brcms_b_corereset(struct brcms_hardware *wlc_hw, u32 flags);
645 u8 antsel_type); 645void brcms_b_mctrl(struct brcms_hardware *wlc_hw, u32 mask, u32 val);
646extern void brcms_b_set_chanspec(struct brcms_hardware *wlc_hw, 646void brcms_b_phy_reset(struct brcms_hardware *wlc_hw);
647 u16 chanspec, 647void brcms_b_bw_set(struct brcms_hardware *wlc_hw, u16 bw);
648 bool mute, struct txpwr_limits *txpwr); 648void brcms_b_core_phypll_reset(struct brcms_hardware *wlc_hw);
649extern void brcms_b_write_shm(struct brcms_hardware *wlc_hw, uint offset, 649void brcms_c_ucode_wake_override_set(struct brcms_hardware *wlc_hw,
650 u16 v); 650 u32 override_bit);
651extern u16 brcms_b_read_shm(struct brcms_hardware *wlc_hw, uint offset); 651void brcms_c_ucode_wake_override_clear(struct brcms_hardware *wlc_hw,
652extern void brcms_b_mhf(struct brcms_hardware *wlc_hw, u8 idx, u16 mask, 652 u32 override_bit);
653 u16 val, int bands); 653void brcms_b_write_template_ram(struct brcms_hardware *wlc_hw, int offset,
654extern void brcms_b_corereset(struct brcms_hardware *wlc_hw, u32 flags); 654 int len, void *buf);
655extern void brcms_b_mctrl(struct brcms_hardware *wlc_hw, u32 mask, u32 val); 655u16 brcms_b_rate_shm_offset(struct brcms_hardware *wlc_hw, u8 rate);
656extern void brcms_b_phy_reset(struct brcms_hardware *wlc_hw); 656void brcms_b_copyto_objmem(struct brcms_hardware *wlc_hw, uint offset,
657extern void brcms_b_bw_set(struct brcms_hardware *wlc_hw, u16 bw); 657 const void *buf, int len, u32 sel);
658extern void brcms_b_core_phypll_reset(struct brcms_hardware *wlc_hw); 658void brcms_b_copyfrom_objmem(struct brcms_hardware *wlc_hw, uint offset,
659extern void brcms_c_ucode_wake_override_set(struct brcms_hardware *wlc_hw, 659 void *buf, int len, u32 sel);
660 u32 override_bit); 660void brcms_b_switch_macfreq(struct brcms_hardware *wlc_hw, u8 spurmode);
661extern void brcms_c_ucode_wake_override_clear(struct brcms_hardware *wlc_hw, 661u16 brcms_b_get_txant(struct brcms_hardware *wlc_hw);
662 u32 override_bit); 662void brcms_b_phyclk_fgc(struct brcms_hardware *wlc_hw, bool clk);
663extern void brcms_b_write_template_ram(struct brcms_hardware *wlc_hw, 663void brcms_b_macphyclk_set(struct brcms_hardware *wlc_hw, bool clk);
664 int offset, int len, void *buf); 664void brcms_b_core_phypll_ctl(struct brcms_hardware *wlc_hw, bool on);
665extern u16 brcms_b_rate_shm_offset(struct brcms_hardware *wlc_hw, u8 rate); 665void brcms_b_txant_set(struct brcms_hardware *wlc_hw, u16 phytxant);
666extern void brcms_b_copyto_objmem(struct brcms_hardware *wlc_hw, 666void brcms_b_band_stf_ss_set(struct brcms_hardware *wlc_hw, u8 stf_mode);
667 uint offset, const void *buf, int len, 667void brcms_c_init_scb(struct scb *scb);
668 u32 sel);
669extern void brcms_b_copyfrom_objmem(struct brcms_hardware *wlc_hw, uint offset,
670 void *buf, int len, u32 sel);
671extern void brcms_b_switch_macfreq(struct brcms_hardware *wlc_hw, u8 spurmode);
672extern u16 brcms_b_get_txant(struct brcms_hardware *wlc_hw);
673extern void brcms_b_phyclk_fgc(struct brcms_hardware *wlc_hw, bool clk);
674extern void brcms_b_macphyclk_set(struct brcms_hardware *wlc_hw, bool clk);
675extern void brcms_b_core_phypll_ctl(struct brcms_hardware *wlc_hw, bool on);
676extern void brcms_b_txant_set(struct brcms_hardware *wlc_hw, u16 phytxant);
677extern void brcms_b_band_stf_ss_set(struct brcms_hardware *wlc_hw,
678 u8 stf_mode);
679extern void brcms_c_init_scb(struct scb *scb);
680 668
681#endif /* _BRCM_MAIN_H_ */ 669#endif /* _BRCM_MAIN_H_ */
diff --git a/drivers/net/wireless/brcm80211/brcmsmac/phy/phy_hal.h b/drivers/net/wireless/brcm80211/brcmsmac/phy/phy_hal.h
index e34a71e7d242..4d3734f48d9c 100644
--- a/drivers/net/wireless/brcm80211/brcmsmac/phy/phy_hal.h
+++ b/drivers/net/wireless/brcm80211/brcmsmac/phy/phy_hal.h
@@ -179,121 +179,106 @@ struct shared_phy_params {
179}; 179};
180 180
181 181
182extern struct shared_phy *wlc_phy_shared_attach(struct shared_phy_params *shp); 182struct shared_phy *wlc_phy_shared_attach(struct shared_phy_params *shp);
183extern struct brcms_phy_pub *wlc_phy_attach(struct shared_phy *sh, 183struct brcms_phy_pub *wlc_phy_attach(struct shared_phy *sh,
184 struct bcma_device *d11core, 184 struct bcma_device *d11core, int bandtype,
185 int bandtype, struct wiphy *wiphy); 185 struct wiphy *wiphy);
186extern void wlc_phy_detach(struct brcms_phy_pub *ppi); 186void wlc_phy_detach(struct brcms_phy_pub *ppi);
187 187
188extern bool wlc_phy_get_phyversion(struct brcms_phy_pub *pih, u16 *phytype, 188bool wlc_phy_get_phyversion(struct brcms_phy_pub *pih, u16 *phytype,
189 u16 *phyrev, u16 *radioid, 189 u16 *phyrev, u16 *radioid, u16 *radiover);
190 u16 *radiover); 190bool wlc_phy_get_encore(struct brcms_phy_pub *pih);
191extern bool wlc_phy_get_encore(struct brcms_phy_pub *pih); 191u32 wlc_phy_get_coreflags(struct brcms_phy_pub *pih);
192extern u32 wlc_phy_get_coreflags(struct brcms_phy_pub *pih); 192
193 193void wlc_phy_hw_clk_state_upd(struct brcms_phy_pub *ppi, bool newstate);
194extern void wlc_phy_hw_clk_state_upd(struct brcms_phy_pub *ppi, bool newstate); 194void wlc_phy_hw_state_upd(struct brcms_phy_pub *ppi, bool newstate);
195extern void wlc_phy_hw_state_upd(struct brcms_phy_pub *ppi, bool newstate); 195void wlc_phy_init(struct brcms_phy_pub *ppi, u16 chanspec);
196extern void wlc_phy_init(struct brcms_phy_pub *ppi, u16 chanspec); 196void wlc_phy_watchdog(struct brcms_phy_pub *ppi);
197extern void wlc_phy_watchdog(struct brcms_phy_pub *ppi); 197int wlc_phy_down(struct brcms_phy_pub *ppi);
198extern int wlc_phy_down(struct brcms_phy_pub *ppi); 198u32 wlc_phy_clk_bwbits(struct brcms_phy_pub *pih);
199extern u32 wlc_phy_clk_bwbits(struct brcms_phy_pub *pih); 199void wlc_phy_cal_init(struct brcms_phy_pub *ppi);
200extern void wlc_phy_cal_init(struct brcms_phy_pub *ppi); 200void wlc_phy_antsel_init(struct brcms_phy_pub *ppi, bool lut_init);
201extern void wlc_phy_antsel_init(struct brcms_phy_pub *ppi, bool lut_init); 201
202 202void wlc_phy_chanspec_set(struct brcms_phy_pub *ppi, u16 chanspec);
203extern void wlc_phy_chanspec_set(struct brcms_phy_pub *ppi, 203u16 wlc_phy_chanspec_get(struct brcms_phy_pub *ppi);
204 u16 chanspec); 204void wlc_phy_chanspec_radio_set(struct brcms_phy_pub *ppi, u16 newch);
205extern u16 wlc_phy_chanspec_get(struct brcms_phy_pub *ppi); 205u16 wlc_phy_bw_state_get(struct brcms_phy_pub *ppi);
206extern void wlc_phy_chanspec_radio_set(struct brcms_phy_pub *ppi, 206void wlc_phy_bw_state_set(struct brcms_phy_pub *ppi, u16 bw);
207 u16 newch); 207
208extern u16 wlc_phy_bw_state_get(struct brcms_phy_pub *ppi); 208int wlc_phy_rssi_compute(struct brcms_phy_pub *pih, struct d11rxhdr *rxh);
209extern void wlc_phy_bw_state_set(struct brcms_phy_pub *ppi, u16 bw); 209void wlc_phy_por_inform(struct brcms_phy_pub *ppi);
210 210void wlc_phy_noise_sample_intr(struct brcms_phy_pub *ppi);
211extern int wlc_phy_rssi_compute(struct brcms_phy_pub *pih, 211bool wlc_phy_bist_check_phy(struct brcms_phy_pub *ppi);
212 struct d11rxhdr *rxh); 212
213extern void wlc_phy_por_inform(struct brcms_phy_pub *ppi); 213void wlc_phy_set_deaf(struct brcms_phy_pub *ppi, bool user_flag);
214extern void wlc_phy_noise_sample_intr(struct brcms_phy_pub *ppi); 214
215extern bool wlc_phy_bist_check_phy(struct brcms_phy_pub *ppi); 215void wlc_phy_switch_radio(struct brcms_phy_pub *ppi, bool on);
216 216void wlc_phy_anacore(struct brcms_phy_pub *ppi, bool on);
217extern void wlc_phy_set_deaf(struct brcms_phy_pub *ppi, bool user_flag); 217
218 218
219extern void wlc_phy_switch_radio(struct brcms_phy_pub *ppi, bool on); 219void wlc_phy_BSSinit(struct brcms_phy_pub *ppi, bool bonlyap, int rssi);
220extern void wlc_phy_anacore(struct brcms_phy_pub *ppi, bool on); 220
221 221void wlc_phy_chanspec_ch14_widefilter_set(struct brcms_phy_pub *ppi,
222 222 bool wide_filter);
223extern void wlc_phy_BSSinit(struct brcms_phy_pub *ppi, bool bonlyap, int rssi); 223void wlc_phy_chanspec_band_validch(struct brcms_phy_pub *ppi, uint band,
224 224 struct brcms_chanvec *channels);
225extern void wlc_phy_chanspec_ch14_widefilter_set(struct brcms_phy_pub *ppi, 225u16 wlc_phy_chanspec_band_firstch(struct brcms_phy_pub *ppi, uint band);
226 bool wide_filter); 226
227extern void wlc_phy_chanspec_band_validch(struct brcms_phy_pub *ppi, uint band, 227void wlc_phy_txpower_sromlimit(struct brcms_phy_pub *ppi, uint chan, u8 *_min_,
228 struct brcms_chanvec *channels); 228 u8 *_max_, int rate);
229extern u16 wlc_phy_chanspec_band_firstch(struct brcms_phy_pub *ppi, 229void wlc_phy_txpower_sromlimit_max_get(struct brcms_phy_pub *ppi, uint chan,
230 uint band); 230 u8 *_max_, u8 *_min_);
231 231void wlc_phy_txpower_boardlimit_band(struct brcms_phy_pub *ppi, uint band,
232extern void wlc_phy_txpower_sromlimit(struct brcms_phy_pub *ppi, uint chan, 232 s32 *, s32 *, u32 *);
233 u8 *_min_, u8 *_max_, int rate); 233void wlc_phy_txpower_limit_set(struct brcms_phy_pub *ppi, struct txpwr_limits *,
234extern void wlc_phy_txpower_sromlimit_max_get(struct brcms_phy_pub *ppi, 234 u16 chanspec);
235 uint chan, u8 *_max_, u8 *_min_); 235int wlc_phy_txpower_get(struct brcms_phy_pub *ppi, uint *qdbm, bool *override);
236extern void wlc_phy_txpower_boardlimit_band(struct brcms_phy_pub *ppi, 236int wlc_phy_txpower_set(struct brcms_phy_pub *ppi, uint qdbm, bool override);
237 uint band, s32 *, s32 *, u32 *); 237void wlc_phy_txpower_target_set(struct brcms_phy_pub *ppi,
238extern void wlc_phy_txpower_limit_set(struct brcms_phy_pub *ppi, 238 struct txpwr_limits *);
239 struct txpwr_limits *, 239bool wlc_phy_txpower_hw_ctrl_get(struct brcms_phy_pub *ppi);
240 u16 chanspec); 240void wlc_phy_txpower_hw_ctrl_set(struct brcms_phy_pub *ppi, bool hwpwrctrl);
241extern int wlc_phy_txpower_get(struct brcms_phy_pub *ppi, uint *qdbm, 241u8 wlc_phy_txpower_get_target_min(struct brcms_phy_pub *ppi);
242 bool *override); 242u8 wlc_phy_txpower_get_target_max(struct brcms_phy_pub *ppi);
243extern int wlc_phy_txpower_set(struct brcms_phy_pub *ppi, uint qdbm, 243bool wlc_phy_txpower_ipa_ison(struct brcms_phy_pub *pih);
244 bool override); 244
245extern void wlc_phy_txpower_target_set(struct brcms_phy_pub *ppi, 245void wlc_phy_stf_chain_init(struct brcms_phy_pub *pih, u8 txchain, u8 rxchain);
246 struct txpwr_limits *); 246void wlc_phy_stf_chain_set(struct brcms_phy_pub *pih, u8 txchain, u8 rxchain);
247extern bool wlc_phy_txpower_hw_ctrl_get(struct brcms_phy_pub *ppi); 247void wlc_phy_stf_chain_get(struct brcms_phy_pub *pih, u8 *txchain, u8 *rxchain);
248extern void wlc_phy_txpower_hw_ctrl_set(struct brcms_phy_pub *ppi, 248u8 wlc_phy_stf_chain_active_get(struct brcms_phy_pub *pih);
249 bool hwpwrctrl); 249s8 wlc_phy_stf_ssmode_get(struct brcms_phy_pub *pih, u16 chanspec);
250extern u8 wlc_phy_txpower_get_target_min(struct brcms_phy_pub *ppi); 250void wlc_phy_ldpc_override_set(struct brcms_phy_pub *ppi, bool val);
251extern u8 wlc_phy_txpower_get_target_max(struct brcms_phy_pub *ppi); 251
252extern bool wlc_phy_txpower_ipa_ison(struct brcms_phy_pub *pih); 252void wlc_phy_cal_perical(struct brcms_phy_pub *ppi, u8 reason);
253 253void wlc_phy_noise_sample_request_external(struct brcms_phy_pub *ppi);
254extern void wlc_phy_stf_chain_init(struct brcms_phy_pub *pih, u8 txchain, 254void wlc_phy_edcrs_lock(struct brcms_phy_pub *pih, bool lock);
255 u8 rxchain); 255void wlc_phy_cal_papd_recal(struct brcms_phy_pub *ppi);
256extern void wlc_phy_stf_chain_set(struct brcms_phy_pub *pih, u8 txchain, 256
257 u8 rxchain); 257void wlc_phy_ant_rxdiv_set(struct brcms_phy_pub *ppi, u8 val);
258extern void wlc_phy_stf_chain_get(struct brcms_phy_pub *pih, u8 *txchain, 258void wlc_phy_clear_tssi(struct brcms_phy_pub *ppi);
259 u8 *rxchain); 259void wlc_phy_hold_upd(struct brcms_phy_pub *ppi, u32 id, bool val);
260extern u8 wlc_phy_stf_chain_active_get(struct brcms_phy_pub *pih); 260void wlc_phy_mute_upd(struct brcms_phy_pub *ppi, bool val, u32 flags);
261extern s8 wlc_phy_stf_ssmode_get(struct brcms_phy_pub *pih, 261
262 u16 chanspec); 262void wlc_phy_antsel_type_set(struct brcms_phy_pub *ppi, u8 antsel_type);
263extern void wlc_phy_ldpc_override_set(struct brcms_phy_pub *ppi, bool val); 263
264 264void wlc_phy_txpower_get_current(struct brcms_phy_pub *ppi,
265extern void wlc_phy_cal_perical(struct brcms_phy_pub *ppi, u8 reason); 265 struct tx_power *power, uint channel);
266extern void wlc_phy_noise_sample_request_external(struct brcms_phy_pub *ppi); 266
267extern void wlc_phy_edcrs_lock(struct brcms_phy_pub *pih, bool lock); 267void wlc_phy_initcal_enable(struct brcms_phy_pub *pih, bool initcal);
268extern void wlc_phy_cal_papd_recal(struct brcms_phy_pub *ppi); 268bool wlc_phy_test_ison(struct brcms_phy_pub *ppi);
269 269void wlc_phy_txpwr_percent_set(struct brcms_phy_pub *ppi, u8 txpwr_percent);
270extern void wlc_phy_ant_rxdiv_set(struct brcms_phy_pub *ppi, u8 val); 270void wlc_phy_ofdm_rateset_war(struct brcms_phy_pub *pih, bool war);
271extern void wlc_phy_clear_tssi(struct brcms_phy_pub *ppi); 271void wlc_phy_bf_preempt_enable(struct brcms_phy_pub *pih, bool bf_preempt);
272extern void wlc_phy_hold_upd(struct brcms_phy_pub *ppi, u32 id, bool val); 272void wlc_phy_machwcap_set(struct brcms_phy_pub *ppi, u32 machwcap);
273extern void wlc_phy_mute_upd(struct brcms_phy_pub *ppi, bool val, u32 flags); 273
274 274void wlc_phy_runbist_config(struct brcms_phy_pub *ppi, bool start_end);
275extern void wlc_phy_antsel_type_set(struct brcms_phy_pub *ppi, u8 antsel_type); 275
276 276void wlc_phy_freqtrack_start(struct brcms_phy_pub *ppi);
277extern void wlc_phy_txpower_get_current(struct brcms_phy_pub *ppi, 277void wlc_phy_freqtrack_end(struct brcms_phy_pub *ppi);
278 struct tx_power *power, uint channel); 278
279 279const u8 *wlc_phy_get_ofdm_rate_lookup(void);
280extern void wlc_phy_initcal_enable(struct brcms_phy_pub *pih, bool initcal); 280
281extern bool wlc_phy_test_ison(struct brcms_phy_pub *ppi); 281s8 wlc_phy_get_tx_power_offset_by_mcs(struct brcms_phy_pub *ppi,
282extern void wlc_phy_txpwr_percent_set(struct brcms_phy_pub *ppi, 282 u8 mcs_offset);
283 u8 txpwr_percent); 283s8 wlc_phy_get_tx_power_offset(struct brcms_phy_pub *ppi, u8 tbl_offset);
284extern void wlc_phy_ofdm_rateset_war(struct brcms_phy_pub *pih, bool war);
285extern void wlc_phy_bf_preempt_enable(struct brcms_phy_pub *pih,
286 bool bf_preempt);
287extern void wlc_phy_machwcap_set(struct brcms_phy_pub *ppi, u32 machwcap);
288
289extern void wlc_phy_runbist_config(struct brcms_phy_pub *ppi, bool start_end);
290
291extern void wlc_phy_freqtrack_start(struct brcms_phy_pub *ppi);
292extern void wlc_phy_freqtrack_end(struct brcms_phy_pub *ppi);
293
294extern const u8 *wlc_phy_get_ofdm_rate_lookup(void);
295
296extern s8 wlc_phy_get_tx_power_offset_by_mcs(struct brcms_phy_pub *ppi,
297 u8 mcs_offset);
298extern s8 wlc_phy_get_tx_power_offset(struct brcms_phy_pub *ppi, u8 tbl_offset);
299#endif /* _BRCM_PHY_HAL_H_ */ 284#endif /* _BRCM_PHY_HAL_H_ */
diff --git a/drivers/net/wireless/brcm80211/brcmsmac/phy/phy_int.h b/drivers/net/wireless/brcm80211/brcmsmac/phy/phy_int.h
index 1dc767c31653..4960f7d26804 100644
--- a/drivers/net/wireless/brcm80211/brcmsmac/phy/phy_int.h
+++ b/drivers/net/wireless/brcm80211/brcmsmac/phy/phy_int.h
@@ -910,113 +910,103 @@ struct lcnphy_radio_regs {
910 u8 do_init_g; 910 u8 do_init_g;
911}; 911};
912 912
913extern u16 read_phy_reg(struct brcms_phy *pi, u16 addr); 913u16 read_phy_reg(struct brcms_phy *pi, u16 addr);
914extern void write_phy_reg(struct brcms_phy *pi, u16 addr, u16 val); 914void write_phy_reg(struct brcms_phy *pi, u16 addr, u16 val);
915extern void and_phy_reg(struct brcms_phy *pi, u16 addr, u16 val); 915void and_phy_reg(struct brcms_phy *pi, u16 addr, u16 val);
916extern void or_phy_reg(struct brcms_phy *pi, u16 addr, u16 val); 916void or_phy_reg(struct brcms_phy *pi, u16 addr, u16 val);
917extern void mod_phy_reg(struct brcms_phy *pi, u16 addr, u16 mask, u16 val); 917void mod_phy_reg(struct brcms_phy *pi, u16 addr, u16 mask, u16 val);
918 918
919extern u16 read_radio_reg(struct brcms_phy *pi, u16 addr); 919u16 read_radio_reg(struct brcms_phy *pi, u16 addr);
920extern void or_radio_reg(struct brcms_phy *pi, u16 addr, u16 val); 920void or_radio_reg(struct brcms_phy *pi, u16 addr, u16 val);
921extern void and_radio_reg(struct brcms_phy *pi, u16 addr, u16 val); 921void and_radio_reg(struct brcms_phy *pi, u16 addr, u16 val);
922extern void mod_radio_reg(struct brcms_phy *pi, u16 addr, u16 mask, 922void mod_radio_reg(struct brcms_phy *pi, u16 addr, u16 mask, u16 val);
923 u16 val); 923void xor_radio_reg(struct brcms_phy *pi, u16 addr, u16 mask);
924extern void xor_radio_reg(struct brcms_phy *pi, u16 addr, u16 mask); 924
925 925void write_radio_reg(struct brcms_phy *pi, u16 addr, u16 val);
926extern void write_radio_reg(struct brcms_phy *pi, u16 addr, u16 val); 926
927 927void wlc_phyreg_enter(struct brcms_phy_pub *pih);
928extern void wlc_phyreg_enter(struct brcms_phy_pub *pih); 928void wlc_phyreg_exit(struct brcms_phy_pub *pih);
929extern void wlc_phyreg_exit(struct brcms_phy_pub *pih); 929void wlc_radioreg_enter(struct brcms_phy_pub *pih);
930extern void wlc_radioreg_enter(struct brcms_phy_pub *pih); 930void wlc_radioreg_exit(struct brcms_phy_pub *pih);
931extern void wlc_radioreg_exit(struct brcms_phy_pub *pih); 931
932 932void wlc_phy_read_table(struct brcms_phy *pi,
933extern void wlc_phy_read_table(struct brcms_phy *pi, 933 const struct phytbl_info *ptbl_info,
934 const struct phytbl_info *ptbl_info, 934 u16 tblAddr, u16 tblDataHi, u16 tblDatalo);
935 u16 tblAddr, u16 tblDataHi, 935void wlc_phy_write_table(struct brcms_phy *pi,
936 u16 tblDatalo); 936 const struct phytbl_info *ptbl_info,
937extern void wlc_phy_write_table(struct brcms_phy *pi, 937 u16 tblAddr, u16 tblDataHi, u16 tblDatalo);
938 const struct phytbl_info *ptbl_info, 938void wlc_phy_table_addr(struct brcms_phy *pi, uint tbl_id, uint tbl_offset,
939 u16 tblAddr, u16 tblDataHi, u16 tblDatalo); 939 u16 tblAddr, u16 tblDataHi, u16 tblDataLo);
940extern void wlc_phy_table_addr(struct brcms_phy *pi, uint tbl_id, 940void wlc_phy_table_data_write(struct brcms_phy *pi, uint width, u32 val);
941 uint tbl_offset, u16 tblAddr, u16 tblDataHi, 941
942 u16 tblDataLo); 942void write_phy_channel_reg(struct brcms_phy *pi, uint val);
943extern void wlc_phy_table_data_write(struct brcms_phy *pi, uint width, u32 val); 943void wlc_phy_txpower_update_shm(struct brcms_phy *pi);
944 944
945extern void write_phy_channel_reg(struct brcms_phy *pi, uint val); 945u8 wlc_phy_nbits(s32 value);
946extern void wlc_phy_txpower_update_shm(struct brcms_phy *pi); 946void wlc_phy_compute_dB(u32 *cmplx_pwr, s8 *p_dB, u8 core);
947 947
948extern u8 wlc_phy_nbits(s32 value); 948uint wlc_phy_init_radio_regs_allbands(struct brcms_phy *pi,
949extern void wlc_phy_compute_dB(u32 *cmplx_pwr, s8 *p_dB, u8 core); 949 struct radio_20xx_regs *radioregs);
950 950uint wlc_phy_init_radio_regs(struct brcms_phy *pi,
951extern uint wlc_phy_init_radio_regs_allbands(struct brcms_phy *pi, 951 const struct radio_regs *radioregs,
952 struct radio_20xx_regs *radioregs); 952 u16 core_offset);
953extern uint wlc_phy_init_radio_regs(struct brcms_phy *pi, 953
954 const struct radio_regs *radioregs, 954void wlc_phy_txpower_ipa_upd(struct brcms_phy *pi);
955 u16 core_offset); 955
956 956void wlc_phy_do_dummy_tx(struct brcms_phy *pi, bool ofdm, bool pa_on);
957extern void wlc_phy_txpower_ipa_upd(struct brcms_phy *pi); 957void wlc_phy_papd_decode_epsilon(u32 epsilon, s32 *eps_real, s32 *eps_imag);
958 958
959extern void wlc_phy_do_dummy_tx(struct brcms_phy *pi, bool ofdm, bool pa_on); 959void wlc_phy_cal_perical_mphase_reset(struct brcms_phy *pi);
960extern void wlc_phy_papd_decode_epsilon(u32 epsilon, s32 *eps_real, 960void wlc_phy_cal_perical_mphase_restart(struct brcms_phy *pi);
961 s32 *eps_imag); 961
962 962bool wlc_phy_attach_nphy(struct brcms_phy *pi);
963extern void wlc_phy_cal_perical_mphase_reset(struct brcms_phy *pi); 963bool wlc_phy_attach_lcnphy(struct brcms_phy *pi);
964extern void wlc_phy_cal_perical_mphase_restart(struct brcms_phy *pi); 964
965 965void wlc_phy_detach_lcnphy(struct brcms_phy *pi);
966extern bool wlc_phy_attach_nphy(struct brcms_phy *pi); 966
967extern bool wlc_phy_attach_lcnphy(struct brcms_phy *pi); 967void wlc_phy_init_nphy(struct brcms_phy *pi);
968 968void wlc_phy_init_lcnphy(struct brcms_phy *pi);
969extern void wlc_phy_detach_lcnphy(struct brcms_phy *pi); 969
970 970void wlc_phy_cal_init_nphy(struct brcms_phy *pi);
971extern void wlc_phy_init_nphy(struct brcms_phy *pi); 971void wlc_phy_cal_init_lcnphy(struct brcms_phy *pi);
972extern void wlc_phy_init_lcnphy(struct brcms_phy *pi); 972
973 973void wlc_phy_chanspec_set_nphy(struct brcms_phy *pi, u16 chanspec);
974extern void wlc_phy_cal_init_nphy(struct brcms_phy *pi); 974void wlc_phy_chanspec_set_lcnphy(struct brcms_phy *pi, u16 chanspec);
975extern void wlc_phy_cal_init_lcnphy(struct brcms_phy *pi); 975void wlc_phy_chanspec_set_fixup_lcnphy(struct brcms_phy *pi, u16 chanspec);
976 976int wlc_phy_channel2freq(uint channel);
977extern void wlc_phy_chanspec_set_nphy(struct brcms_phy *pi, 977int wlc_phy_chanspec_freq2bandrange_lpssn(uint);
978 u16 chanspec); 978int wlc_phy_chanspec_bandrange_get(struct brcms_phy *, u16 chanspec);
979extern void wlc_phy_chanspec_set_lcnphy(struct brcms_phy *pi, 979
980 u16 chanspec); 980void wlc_lcnphy_set_tx_pwr_ctrl(struct brcms_phy *pi, u16 mode);
981extern void wlc_phy_chanspec_set_fixup_lcnphy(struct brcms_phy *pi, 981s8 wlc_lcnphy_get_current_tx_pwr_idx(struct brcms_phy *pi);
982 u16 chanspec); 982
983extern int wlc_phy_channel2freq(uint channel); 983void wlc_phy_txpower_recalc_target_nphy(struct brcms_phy *pi);
984extern int wlc_phy_chanspec_freq2bandrange_lpssn(uint); 984void wlc_lcnphy_txpower_recalc_target(struct brcms_phy *pi);
985extern int wlc_phy_chanspec_bandrange_get(struct brcms_phy *, u16 chanspec); 985void wlc_phy_txpower_recalc_target_lcnphy(struct brcms_phy *pi);
986 986
987extern void wlc_lcnphy_set_tx_pwr_ctrl(struct brcms_phy *pi, u16 mode); 987void wlc_lcnphy_set_tx_pwr_by_index(struct brcms_phy *pi, int index);
988extern s8 wlc_lcnphy_get_current_tx_pwr_idx(struct brcms_phy *pi); 988void wlc_lcnphy_tx_pu(struct brcms_phy *pi, bool bEnable);
989 989void wlc_lcnphy_stop_tx_tone(struct brcms_phy *pi);
990extern void wlc_phy_txpower_recalc_target_nphy(struct brcms_phy *pi); 990void wlc_lcnphy_start_tx_tone(struct brcms_phy *pi, s32 f_kHz, u16 max_val,
991extern void wlc_lcnphy_txpower_recalc_target(struct brcms_phy *pi); 991 bool iqcalmode);
992extern void wlc_phy_txpower_recalc_target_lcnphy(struct brcms_phy *pi); 992
993 993void wlc_phy_txpower_sromlimit_get_nphy(struct brcms_phy *pi, uint chan,
994extern void wlc_lcnphy_set_tx_pwr_by_index(struct brcms_phy *pi, int index); 994 u8 *max_pwr, u8 rate_id);
995extern void wlc_lcnphy_tx_pu(struct brcms_phy *pi, bool bEnable); 995void wlc_phy_ofdm_to_mcs_powers_nphy(u8 *power, u8 rate_mcs_start,
996extern void wlc_lcnphy_stop_tx_tone(struct brcms_phy *pi); 996 u8 rate_mcs_end, u8 rate_ofdm_start);
997extern void wlc_lcnphy_start_tx_tone(struct brcms_phy *pi, s32 f_kHz, 997void wlc_phy_mcs_to_ofdm_powers_nphy(u8 *power, u8 rate_ofdm_start,
998 u16 max_val, bool iqcalmode); 998 u8 rate_ofdm_end, u8 rate_mcs_start);
999 999
1000extern void wlc_phy_txpower_sromlimit_get_nphy(struct brcms_phy *pi, uint chan, 1000u16 wlc_lcnphy_tempsense(struct brcms_phy *pi, bool mode);
1001 u8 *max_pwr, u8 rate_id); 1001s16 wlc_lcnphy_tempsense_new(struct brcms_phy *pi, bool mode);
1002extern void wlc_phy_ofdm_to_mcs_powers_nphy(u8 *power, u8 rate_mcs_start, 1002s8 wlc_lcnphy_tempsense_degree(struct brcms_phy *pi, bool mode);
1003 u8 rate_mcs_end, 1003s8 wlc_lcnphy_vbatsense(struct brcms_phy *pi, bool mode);
1004 u8 rate_ofdm_start); 1004void wlc_phy_carrier_suppress_lcnphy(struct brcms_phy *pi);
1005extern void wlc_phy_mcs_to_ofdm_powers_nphy(u8 *power, 1005void wlc_lcnphy_crsuprs(struct brcms_phy *pi, int channel);
1006 u8 rate_ofdm_start, 1006void wlc_lcnphy_epa_switch(struct brcms_phy *pi, bool mode);
1007 u8 rate_ofdm_end, 1007void wlc_2064_vco_cal(struct brcms_phy *pi);
1008 u8 rate_mcs_start); 1008
1009 1009void wlc_phy_txpower_recalc_target(struct brcms_phy *pi);
1010extern u16 wlc_lcnphy_tempsense(struct brcms_phy *pi, bool mode);
1011extern s16 wlc_lcnphy_tempsense_new(struct brcms_phy *pi, bool mode);
1012extern s8 wlc_lcnphy_tempsense_degree(struct brcms_phy *pi, bool mode);
1013extern s8 wlc_lcnphy_vbatsense(struct brcms_phy *pi, bool mode);
1014extern void wlc_phy_carrier_suppress_lcnphy(struct brcms_phy *pi);
1015extern void wlc_lcnphy_crsuprs(struct brcms_phy *pi, int channel);
1016extern void wlc_lcnphy_epa_switch(struct brcms_phy *pi, bool mode);
1017extern void wlc_2064_vco_cal(struct brcms_phy *pi);
1018
1019extern void wlc_phy_txpower_recalc_target(struct brcms_phy *pi);
1020 1010
1021#define LCNPHY_TBL_ID_PAPDCOMPDELTATBL 0x18 1011#define LCNPHY_TBL_ID_PAPDCOMPDELTATBL 0x18
1022#define LCNPHY_TX_POWER_TABLE_SIZE 128 1012#define LCNPHY_TX_POWER_TABLE_SIZE 128
@@ -1030,26 +1020,24 @@ extern void wlc_phy_txpower_recalc_target(struct brcms_phy *pi);
1030 1020
1031#define LCNPHY_TX_PWR_CTRL_TEMPBASED 0xE001 1021#define LCNPHY_TX_PWR_CTRL_TEMPBASED 0xE001
1032 1022
1033extern void wlc_lcnphy_write_table(struct brcms_phy *pi, 1023void wlc_lcnphy_write_table(struct brcms_phy *pi,
1034 const struct phytbl_info *pti); 1024 const struct phytbl_info *pti);
1035extern void wlc_lcnphy_read_table(struct brcms_phy *pi, 1025void wlc_lcnphy_read_table(struct brcms_phy *pi, struct phytbl_info *pti);
1036 struct phytbl_info *pti); 1026void wlc_lcnphy_set_tx_iqcc(struct brcms_phy *pi, u16 a, u16 b);
1037extern void wlc_lcnphy_set_tx_iqcc(struct brcms_phy *pi, u16 a, u16 b); 1027void wlc_lcnphy_set_tx_locc(struct brcms_phy *pi, u16 didq);
1038extern void wlc_lcnphy_set_tx_locc(struct brcms_phy *pi, u16 didq); 1028void wlc_lcnphy_get_tx_iqcc(struct brcms_phy *pi, u16 *a, u16 *b);
1039extern void wlc_lcnphy_get_tx_iqcc(struct brcms_phy *pi, u16 *a, u16 *b); 1029u16 wlc_lcnphy_get_tx_locc(struct brcms_phy *pi);
1040extern u16 wlc_lcnphy_get_tx_locc(struct brcms_phy *pi); 1030void wlc_lcnphy_get_radio_loft(struct brcms_phy *pi, u8 *ei0, u8 *eq0, u8 *fi0,
1041extern void wlc_lcnphy_get_radio_loft(struct brcms_phy *pi, u8 *ei0, 1031 u8 *fq0);
1042 u8 *eq0, u8 *fi0, u8 *fq0); 1032void wlc_lcnphy_calib_modes(struct brcms_phy *pi, uint mode);
1043extern void wlc_lcnphy_calib_modes(struct brcms_phy *pi, uint mode); 1033void wlc_lcnphy_deaf_mode(struct brcms_phy *pi, bool mode);
1044extern void wlc_lcnphy_deaf_mode(struct brcms_phy *pi, bool mode); 1034bool wlc_phy_tpc_isenabled_lcnphy(struct brcms_phy *pi);
1045extern bool wlc_phy_tpc_isenabled_lcnphy(struct brcms_phy *pi); 1035void wlc_lcnphy_tx_pwr_update_npt(struct brcms_phy *pi);
1046extern void wlc_lcnphy_tx_pwr_update_npt(struct brcms_phy *pi); 1036s32 wlc_lcnphy_tssi2dbm(s32 tssi, s32 a1, s32 b0, s32 b1);
1047extern s32 wlc_lcnphy_tssi2dbm(s32 tssi, s32 a1, s32 b0, s32 b1); 1037void wlc_lcnphy_get_tssi(struct brcms_phy *pi, s8 *ofdm_pwr, s8 *cck_pwr);
1048extern void wlc_lcnphy_get_tssi(struct brcms_phy *pi, s8 *ofdm_pwr, 1038void wlc_lcnphy_tx_power_adjustment(struct brcms_phy_pub *ppi);
1049 s8 *cck_pwr); 1039
1050extern void wlc_lcnphy_tx_power_adjustment(struct brcms_phy_pub *ppi); 1040s32 wlc_lcnphy_rx_signal_power(struct brcms_phy *pi, s32 gain_index);
1051
1052extern s32 wlc_lcnphy_rx_signal_power(struct brcms_phy *pi, s32 gain_index);
1053 1041
1054#define NPHY_MAX_HPVGA1_INDEX 10 1042#define NPHY_MAX_HPVGA1_INDEX 10
1055#define NPHY_DEF_HPVGA1_INDEXLIMIT 7 1043#define NPHY_DEF_HPVGA1_INDEXLIMIT 7
@@ -1060,9 +1048,8 @@ struct phy_iq_est {
1060 u32 q_pwr; 1048 u32 q_pwr;
1061}; 1049};
1062 1050
1063extern void wlc_phy_stay_in_carriersearch_nphy(struct brcms_phy *pi, 1051void wlc_phy_stay_in_carriersearch_nphy(struct brcms_phy *pi, bool enable);
1064 bool enable); 1052void wlc_nphy_deaf_mode(struct brcms_phy *pi, bool mode);
1065extern void wlc_nphy_deaf_mode(struct brcms_phy *pi, bool mode);
1066 1053
1067#define wlc_phy_write_table_nphy(pi, pti) \ 1054#define wlc_phy_write_table_nphy(pi, pti) \
1068 wlc_phy_write_table(pi, pti, 0x72, 0x74, 0x73) 1055 wlc_phy_write_table(pi, pti, 0x72, 0x74, 0x73)
@@ -1076,10 +1063,10 @@ extern void wlc_nphy_deaf_mode(struct brcms_phy *pi, bool mode);
1076#define wlc_nphy_table_data_write(pi, w, v) \ 1063#define wlc_nphy_table_data_write(pi, w, v) \
1077 wlc_phy_table_data_write((pi), (w), (v)) 1064 wlc_phy_table_data_write((pi), (w), (v))
1078 1065
1079extern void wlc_phy_table_read_nphy(struct brcms_phy *pi, u32, u32 l, u32 o, 1066void wlc_phy_table_read_nphy(struct brcms_phy *pi, u32, u32 l, u32 o, u32 w,
1080 u32 w, void *d); 1067 void *d);
1081extern void wlc_phy_table_write_nphy(struct brcms_phy *pi, u32, u32, u32, 1068void wlc_phy_table_write_nphy(struct brcms_phy *pi, u32, u32, u32, u32,
1082 u32, const void *); 1069 const void *);
1083 1070
1084#define PHY_IPA(pi) \ 1071#define PHY_IPA(pi) \
1085 ((pi->ipa2g_on && CHSPEC_IS2G(pi->radio_chanspec)) || \ 1072 ((pi->ipa2g_on && CHSPEC_IS2G(pi->radio_chanspec)) || \
@@ -1089,73 +1076,67 @@ extern void wlc_phy_table_write_nphy(struct brcms_phy *pi, u32, u32, u32,
1089 if (NREV_LT((pi)->pubpi.phy_rev, 3)) \ 1076 if (NREV_LT((pi)->pubpi.phy_rev, 3)) \
1090 (void)bcma_read32(pi->d11core, D11REGOFFS(maccontrol)) 1077 (void)bcma_read32(pi->d11core, D11REGOFFS(maccontrol))
1091 1078
1092extern void wlc_phy_cal_perical_nphy_run(struct brcms_phy *pi, u8 caltype); 1079void wlc_phy_cal_perical_nphy_run(struct brcms_phy *pi, u8 caltype);
1093extern void wlc_phy_aci_reset_nphy(struct brcms_phy *pi); 1080void wlc_phy_aci_reset_nphy(struct brcms_phy *pi);
1094extern void wlc_phy_pa_override_nphy(struct brcms_phy *pi, bool en); 1081void wlc_phy_pa_override_nphy(struct brcms_phy *pi, bool en);
1095 1082
1096extern u8 wlc_phy_get_chan_freq_range_nphy(struct brcms_phy *pi, uint chan); 1083u8 wlc_phy_get_chan_freq_range_nphy(struct brcms_phy *pi, uint chan);
1097extern void wlc_phy_switch_radio_nphy(struct brcms_phy *pi, bool on); 1084void wlc_phy_switch_radio_nphy(struct brcms_phy *pi, bool on);
1098 1085
1099extern void wlc_phy_stf_chain_upd_nphy(struct brcms_phy *pi); 1086void wlc_phy_stf_chain_upd_nphy(struct brcms_phy *pi);
1100 1087
1101extern void wlc_phy_force_rfseq_nphy(struct brcms_phy *pi, u8 cmd); 1088void wlc_phy_force_rfseq_nphy(struct brcms_phy *pi, u8 cmd);
1102extern s16 wlc_phy_tempsense_nphy(struct brcms_phy *pi); 1089s16 wlc_phy_tempsense_nphy(struct brcms_phy *pi);
1103 1090
1104extern u16 wlc_phy_classifier_nphy(struct brcms_phy *pi, u16 mask, u16 val); 1091u16 wlc_phy_classifier_nphy(struct brcms_phy *pi, u16 mask, u16 val);
1105 1092
1106extern void wlc_phy_rx_iq_est_nphy(struct brcms_phy *pi, struct phy_iq_est *est, 1093void wlc_phy_rx_iq_est_nphy(struct brcms_phy *pi, struct phy_iq_est *est,
1107 u16 num_samps, u8 wait_time, 1094 u16 num_samps, u8 wait_time, u8 wait_for_crs);
1108 u8 wait_for_crs); 1095
1109 1096void wlc_phy_rx_iq_coeffs_nphy(struct brcms_phy *pi, u8 write,
1110extern void wlc_phy_rx_iq_coeffs_nphy(struct brcms_phy *pi, u8 write, 1097 struct nphy_iq_comp *comp);
1111 struct nphy_iq_comp *comp); 1098void wlc_phy_aci_and_noise_reduction_nphy(struct brcms_phy *pi);
1112extern void wlc_phy_aci_and_noise_reduction_nphy(struct brcms_phy *pi); 1099
1113 1100void wlc_phy_rxcore_setstate_nphy(struct brcms_phy_pub *pih, u8 rxcore_bitmask);
1114extern void wlc_phy_rxcore_setstate_nphy(struct brcms_phy_pub *pih, 1101u8 wlc_phy_rxcore_getstate_nphy(struct brcms_phy_pub *pih);
1115 u8 rxcore_bitmask); 1102
1116extern u8 wlc_phy_rxcore_getstate_nphy(struct brcms_phy_pub *pih); 1103void wlc_phy_txpwrctrl_enable_nphy(struct brcms_phy *pi, u8 ctrl_type);
1117 1104void wlc_phy_txpwr_fixpower_nphy(struct brcms_phy *pi);
1118extern void wlc_phy_txpwrctrl_enable_nphy(struct brcms_phy *pi, u8 ctrl_type); 1105void wlc_phy_txpwr_apply_nphy(struct brcms_phy *pi);
1119extern void wlc_phy_txpwr_fixpower_nphy(struct brcms_phy *pi); 1106void wlc_phy_txpwr_papd_cal_nphy(struct brcms_phy *pi);
1120extern void wlc_phy_txpwr_apply_nphy(struct brcms_phy *pi); 1107u16 wlc_phy_txpwr_idx_get_nphy(struct brcms_phy *pi);
1121extern void wlc_phy_txpwr_papd_cal_nphy(struct brcms_phy *pi); 1108
1122extern u16 wlc_phy_txpwr_idx_get_nphy(struct brcms_phy *pi); 1109struct nphy_txgains wlc_phy_get_tx_gain_nphy(struct brcms_phy *pi);
1123 1110int wlc_phy_cal_txiqlo_nphy(struct brcms_phy *pi,
1124extern struct nphy_txgains wlc_phy_get_tx_gain_nphy(struct brcms_phy *pi); 1111 struct nphy_txgains target_gain, bool full, bool m);
1125extern int wlc_phy_cal_txiqlo_nphy(struct brcms_phy *pi, 1112int wlc_phy_cal_rxiq_nphy(struct brcms_phy *pi, struct nphy_txgains target_gain,
1126 struct nphy_txgains target_gain, 1113 u8 type, bool d);
1127 bool full, bool m); 1114void wlc_phy_txpwr_index_nphy(struct brcms_phy *pi, u8 core_mask,
1128extern int wlc_phy_cal_rxiq_nphy(struct brcms_phy *pi, 1115 s8 txpwrindex, bool res);
1129 struct nphy_txgains target_gain, 1116void wlc_phy_rssisel_nphy(struct brcms_phy *pi, u8 core, u8 rssi_type);
1130 u8 type, bool d); 1117int wlc_phy_poll_rssi_nphy(struct brcms_phy *pi, u8 rssi_type,
1131extern void wlc_phy_txpwr_index_nphy(struct brcms_phy *pi, u8 core_mask, 1118 s32 *rssi_buf, u8 nsamps);
1132 s8 txpwrindex, bool res); 1119void wlc_phy_rssi_cal_nphy(struct brcms_phy *pi);
1133extern void wlc_phy_rssisel_nphy(struct brcms_phy *pi, u8 core, u8 rssi_type); 1120int wlc_phy_aci_scan_nphy(struct brcms_phy *pi);
1134extern int wlc_phy_poll_rssi_nphy(struct brcms_phy *pi, u8 rssi_type, 1121void wlc_phy_cal_txgainctrl_nphy(struct brcms_phy *pi, s32 dBm_targetpower,
1135 s32 *rssi_buf, u8 nsamps); 1122 bool debug);
1136extern void wlc_phy_rssi_cal_nphy(struct brcms_phy *pi); 1123int wlc_phy_tx_tone_nphy(struct brcms_phy *pi, u32 f_kHz, u16 max_val, u8 mode,
1137extern int wlc_phy_aci_scan_nphy(struct brcms_phy *pi); 1124 u8, bool);
1138extern void wlc_phy_cal_txgainctrl_nphy(struct brcms_phy *pi, 1125void wlc_phy_stopplayback_nphy(struct brcms_phy *pi);
1139 s32 dBm_targetpower, bool debug); 1126void wlc_phy_est_tonepwr_nphy(struct brcms_phy *pi, s32 *qdBm_pwrbuf,
1140extern int wlc_phy_tx_tone_nphy(struct brcms_phy *pi, u32 f_kHz, u16 max_val, 1127 u8 num_samps);
1141 u8 mode, u8, bool); 1128void wlc_phy_radio205x_vcocal_nphy(struct brcms_phy *pi);
1142extern void wlc_phy_stopplayback_nphy(struct brcms_phy *pi); 1129
1143extern void wlc_phy_est_tonepwr_nphy(struct brcms_phy *pi, s32 *qdBm_pwrbuf, 1130int wlc_phy_rssi_compute_nphy(struct brcms_phy *pi, struct d11rxhdr *rxh);
1144 u8 num_samps);
1145extern void wlc_phy_radio205x_vcocal_nphy(struct brcms_phy *pi);
1146
1147extern int wlc_phy_rssi_compute_nphy(struct brcms_phy *pi,
1148 struct d11rxhdr *rxh);
1149 1131
1150#define NPHY_TESTPATTERN_BPHY_EVM 0 1132#define NPHY_TESTPATTERN_BPHY_EVM 0
1151#define NPHY_TESTPATTERN_BPHY_RFCS 1 1133#define NPHY_TESTPATTERN_BPHY_RFCS 1
1152 1134
1153extern void wlc_phy_nphy_tkip_rifs_war(struct brcms_phy *pi, u8 rifs); 1135void wlc_phy_nphy_tkip_rifs_war(struct brcms_phy *pi, u8 rifs);
1154 1136
1155void wlc_phy_get_pwrdet_offsets(struct brcms_phy *pi, s8 *cckoffset, 1137void wlc_phy_get_pwrdet_offsets(struct brcms_phy *pi, s8 *cckoffset,
1156 s8 *ofdmoffset); 1138 s8 *ofdmoffset);
1157extern s8 wlc_phy_upd_rssi_offset(struct brcms_phy *pi, s8 rssi, 1139s8 wlc_phy_upd_rssi_offset(struct brcms_phy *pi, s8 rssi, u16 chanspec);
1158 u16 chanspec);
1159 1140
1160extern bool wlc_phy_n_txpower_ipa_ison(struct brcms_phy *pih); 1141bool wlc_phy_n_txpower_ipa_ison(struct brcms_phy *pih);
1161#endif /* _BRCM_PHY_INT_H_ */ 1142#endif /* _BRCM_PHY_INT_H_ */
diff --git a/drivers/net/wireless/brcm80211/brcmsmac/phy_shim.h b/drivers/net/wireless/brcm80211/brcmsmac/phy_shim.h
index 2c5b66b75970..dd8774717ade 100644
--- a/drivers/net/wireless/brcm80211/brcmsmac/phy_shim.h
+++ b/drivers/net/wireless/brcm80211/brcmsmac/phy_shim.h
@@ -124,56 +124,49 @@
124 124
125struct brcms_phy; 125struct brcms_phy;
126 126
127extern struct phy_shim_info *wlc_phy_shim_attach(struct brcms_hardware *wlc_hw, 127struct phy_shim_info *wlc_phy_shim_attach(struct brcms_hardware *wlc_hw,
128 struct brcms_info *wl, 128 struct brcms_info *wl,
129 struct brcms_c_info *wlc); 129 struct brcms_c_info *wlc);
130extern void wlc_phy_shim_detach(struct phy_shim_info *physhim); 130void wlc_phy_shim_detach(struct phy_shim_info *physhim);
131 131
132/* PHY to WL utility functions */ 132/* PHY to WL utility functions */
133extern struct wlapi_timer *wlapi_init_timer(struct phy_shim_info *physhim, 133struct wlapi_timer *wlapi_init_timer(struct phy_shim_info *physhim,
134 void (*fn) (struct brcms_phy *pi), 134 void (*fn)(struct brcms_phy *pi),
135 void *arg, const char *name); 135 void *arg, const char *name);
136extern void wlapi_free_timer(struct wlapi_timer *t); 136void wlapi_free_timer(struct wlapi_timer *t);
137extern void wlapi_add_timer(struct wlapi_timer *t, uint ms, int periodic); 137void wlapi_add_timer(struct wlapi_timer *t, uint ms, int periodic);
138extern bool wlapi_del_timer(struct wlapi_timer *t); 138bool wlapi_del_timer(struct wlapi_timer *t);
139extern void wlapi_intrson(struct phy_shim_info *physhim); 139void wlapi_intrson(struct phy_shim_info *physhim);
140extern u32 wlapi_intrsoff(struct phy_shim_info *physhim); 140u32 wlapi_intrsoff(struct phy_shim_info *physhim);
141extern void wlapi_intrsrestore(struct phy_shim_info *physhim, 141void wlapi_intrsrestore(struct phy_shim_info *physhim, u32 macintmask);
142 u32 macintmask); 142
143 143void wlapi_bmac_write_shm(struct phy_shim_info *physhim, uint offset, u16 v);
144extern void wlapi_bmac_write_shm(struct phy_shim_info *physhim, uint offset, 144u16 wlapi_bmac_read_shm(struct phy_shim_info *physhim, uint offset);
145 u16 v); 145void wlapi_bmac_mhf(struct phy_shim_info *physhim, u8 idx, u16 mask, u16 val,
146extern u16 wlapi_bmac_read_shm(struct phy_shim_info *physhim, uint offset); 146 int bands);
147extern void wlapi_bmac_mhf(struct phy_shim_info *physhim, u8 idx, 147void wlapi_bmac_corereset(struct phy_shim_info *physhim, u32 flags);
148 u16 mask, u16 val, int bands); 148void wlapi_suspend_mac_and_wait(struct phy_shim_info *physhim);
149extern void wlapi_bmac_corereset(struct phy_shim_info *physhim, u32 flags); 149void wlapi_switch_macfreq(struct phy_shim_info *physhim, u8 spurmode);
150extern void wlapi_suspend_mac_and_wait(struct phy_shim_info *physhim); 150void wlapi_enable_mac(struct phy_shim_info *physhim);
151extern void wlapi_switch_macfreq(struct phy_shim_info *physhim, u8 spurmode); 151void wlapi_bmac_mctrl(struct phy_shim_info *physhim, u32 mask, u32 val);
152extern void wlapi_enable_mac(struct phy_shim_info *physhim); 152void wlapi_bmac_phy_reset(struct phy_shim_info *physhim);
153extern void wlapi_bmac_mctrl(struct phy_shim_info *physhim, u32 mask, 153void wlapi_bmac_bw_set(struct phy_shim_info *physhim, u16 bw);
154 u32 val); 154void wlapi_bmac_phyclk_fgc(struct phy_shim_info *physhim, bool clk);
155extern void wlapi_bmac_phy_reset(struct phy_shim_info *physhim); 155void wlapi_bmac_macphyclk_set(struct phy_shim_info *physhim, bool clk);
156extern void wlapi_bmac_bw_set(struct phy_shim_info *physhim, u16 bw); 156void wlapi_bmac_core_phypll_ctl(struct phy_shim_info *physhim, bool on);
157extern void wlapi_bmac_phyclk_fgc(struct phy_shim_info *physhim, bool clk); 157void wlapi_bmac_core_phypll_reset(struct phy_shim_info *physhim);
158extern void wlapi_bmac_macphyclk_set(struct phy_shim_info *physhim, bool clk); 158void wlapi_bmac_ucode_wake_override_phyreg_set(struct phy_shim_info *physhim);
159extern void wlapi_bmac_core_phypll_ctl(struct phy_shim_info *physhim, bool on); 159void wlapi_bmac_ucode_wake_override_phyreg_clear(struct phy_shim_info *physhim);
160extern void wlapi_bmac_core_phypll_reset(struct phy_shim_info *physhim); 160void wlapi_bmac_write_template_ram(struct phy_shim_info *physhim, int o,
161extern void wlapi_bmac_ucode_wake_override_phyreg_set(struct phy_shim_info * 161 int len, void *buf);
162 physhim); 162u16 wlapi_bmac_rate_shm_offset(struct phy_shim_info *physhim, u8 rate);
163extern void wlapi_bmac_ucode_wake_override_phyreg_clear(struct phy_shim_info * 163void wlapi_ucode_sample_init(struct phy_shim_info *physhim);
164 physhim); 164void wlapi_copyfrom_objmem(struct phy_shim_info *physhim, uint, void *buf,
165extern void wlapi_bmac_write_template_ram(struct phy_shim_info *physhim, int o, 165 int, u32 sel);
166 int len, void *buf); 166void wlapi_copyto_objmem(struct phy_shim_info *physhim, uint, const void *buf,
167extern u16 wlapi_bmac_rate_shm_offset(struct phy_shim_info *physhim, 167 int, u32);
168 u8 rate); 168
169extern void wlapi_ucode_sample_init(struct phy_shim_info *physhim); 169void wlapi_high_update_phy_mode(struct phy_shim_info *physhim, u32 phy_mode);
170extern void wlapi_copyfrom_objmem(struct phy_shim_info *physhim, uint, 170u16 wlapi_bmac_get_txant(struct phy_shim_info *physhim);
171 void *buf, int, u32 sel);
172extern void wlapi_copyto_objmem(struct phy_shim_info *physhim, uint,
173 const void *buf, int, u32);
174
175extern void wlapi_high_update_phy_mode(struct phy_shim_info *physhim,
176 u32 phy_mode);
177extern u16 wlapi_bmac_get_txant(struct phy_shim_info *physhim);
178 171
179#endif /* _BRCM_PHY_SHIM_H_ */ 172#endif /* _BRCM_PHY_SHIM_H_ */
diff --git a/drivers/net/wireless/brcm80211/brcmsmac/pmu.h b/drivers/net/wireless/brcm80211/brcmsmac/pmu.h
index 20e2012d5a3a..a014bbc4f935 100644
--- a/drivers/net/wireless/brcm80211/brcmsmac/pmu.h
+++ b/drivers/net/wireless/brcm80211/brcmsmac/pmu.h
@@ -20,7 +20,7 @@
20 20
21#include "types.h" 21#include "types.h"
22 22
23extern u16 si_pmu_fast_pwrup_delay(struct si_pub *sih); 23u16 si_pmu_fast_pwrup_delay(struct si_pub *sih);
24extern u32 si_pmu_measure_alpclk(struct si_pub *sih); 24u32 si_pmu_measure_alpclk(struct si_pub *sih);
25 25
26#endif /* _BRCM_PMU_H_ */ 26#endif /* _BRCM_PMU_H_ */
diff --git a/drivers/net/wireless/brcm80211/brcmsmac/pub.h b/drivers/net/wireless/brcm80211/brcmsmac/pub.h
index d36ea5e1cc49..4da38cb4f318 100644
--- a/drivers/net/wireless/brcm80211/brcmsmac/pub.h
+++ b/drivers/net/wireless/brcm80211/brcmsmac/pub.h
@@ -266,83 +266,76 @@ struct brcms_antselcfg {
266}; 266};
267 267
268/* common functions for every port */ 268/* common functions for every port */
269extern struct brcms_c_info * 269struct brcms_c_info *brcms_c_attach(struct brcms_info *wl,
270brcms_c_attach(struct brcms_info *wl, struct bcma_device *core, uint unit, 270 struct bcma_device *core, uint unit,
271 bool piomode, uint *perr); 271 bool piomode, uint *perr);
272extern uint brcms_c_detach(struct brcms_c_info *wlc); 272uint brcms_c_detach(struct brcms_c_info *wlc);
273extern int brcms_c_up(struct brcms_c_info *wlc); 273int brcms_c_up(struct brcms_c_info *wlc);
274extern uint brcms_c_down(struct brcms_c_info *wlc); 274uint brcms_c_down(struct brcms_c_info *wlc);
275 275
276extern bool brcms_c_chipmatch(struct bcma_device *core); 276bool brcms_c_chipmatch(struct bcma_device *core);
277extern void brcms_c_init(struct brcms_c_info *wlc, bool mute_tx); 277void brcms_c_init(struct brcms_c_info *wlc, bool mute_tx);
278extern void brcms_c_reset(struct brcms_c_info *wlc); 278void brcms_c_reset(struct brcms_c_info *wlc);
279 279
280extern void brcms_c_intrson(struct brcms_c_info *wlc); 280void brcms_c_intrson(struct brcms_c_info *wlc);
281extern u32 brcms_c_intrsoff(struct brcms_c_info *wlc); 281u32 brcms_c_intrsoff(struct brcms_c_info *wlc);
282extern void brcms_c_intrsrestore(struct brcms_c_info *wlc, u32 macintmask); 282void brcms_c_intrsrestore(struct brcms_c_info *wlc, u32 macintmask);
283extern bool brcms_c_intrsupd(struct brcms_c_info *wlc); 283bool brcms_c_intrsupd(struct brcms_c_info *wlc);
284extern bool brcms_c_isr(struct brcms_c_info *wlc); 284bool brcms_c_isr(struct brcms_c_info *wlc);
285extern bool brcms_c_dpc(struct brcms_c_info *wlc, bool bounded); 285bool brcms_c_dpc(struct brcms_c_info *wlc, bool bounded);
286extern bool brcms_c_sendpkt_mac80211(struct brcms_c_info *wlc, 286bool brcms_c_sendpkt_mac80211(struct brcms_c_info *wlc, struct sk_buff *sdu,
287 struct sk_buff *sdu, 287 struct ieee80211_hw *hw);
288 struct ieee80211_hw *hw); 288bool brcms_c_aggregatable(struct brcms_c_info *wlc, u8 tid);
289extern bool brcms_c_aggregatable(struct brcms_c_info *wlc, u8 tid); 289void brcms_c_protection_upd(struct brcms_c_info *wlc, uint idx, int val);
290extern void brcms_c_protection_upd(struct brcms_c_info *wlc, uint idx, 290int brcms_c_get_header_len(void);
291 int val); 291void brcms_c_set_addrmatch(struct brcms_c_info *wlc, int match_reg_offset,
292extern int brcms_c_get_header_len(void); 292 const u8 *addr);
293extern void brcms_c_set_addrmatch(struct brcms_c_info *wlc, 293void brcms_c_wme_setparams(struct brcms_c_info *wlc, u16 aci,
294 int match_reg_offset, 294 const struct ieee80211_tx_queue_params *arg,
295 const u8 *addr); 295 bool suspend);
296extern void brcms_c_wme_setparams(struct brcms_c_info *wlc, u16 aci, 296struct brcms_pub *brcms_c_pub(struct brcms_c_info *wlc);
297 const struct ieee80211_tx_queue_params *arg, 297void brcms_c_ampdu_flush(struct brcms_c_info *wlc, struct ieee80211_sta *sta,
298 bool suspend); 298 u16 tid);
299extern struct brcms_pub *brcms_c_pub(struct brcms_c_info *wlc); 299void brcms_c_ampdu_tx_operational(struct brcms_c_info *wlc, u8 tid,
300extern void brcms_c_ampdu_flush(struct brcms_c_info *wlc, 300 u8 ba_wsize, uint max_rx_ampdu_bytes);
301 struct ieee80211_sta *sta, u16 tid); 301int brcms_c_module_register(struct brcms_pub *pub, const char *name,
302extern void brcms_c_ampdu_tx_operational(struct brcms_c_info *wlc, u8 tid, 302 struct brcms_info *hdl,
303 u8 ba_wsize, uint max_rx_ampdu_bytes); 303 int (*down_fn)(void *handle));
304extern int brcms_c_module_register(struct brcms_pub *pub, 304int brcms_c_module_unregister(struct brcms_pub *pub, const char *name,
305 const char *name, struct brcms_info *hdl, 305 struct brcms_info *hdl);
306 int (*down_fn)(void *handle)); 306void brcms_c_suspend_mac_and_wait(struct brcms_c_info *wlc);
307extern int brcms_c_module_unregister(struct brcms_pub *pub, const char *name, 307void brcms_c_enable_mac(struct brcms_c_info *wlc);
308 struct brcms_info *hdl); 308void brcms_c_associate_upd(struct brcms_c_info *wlc, bool state);
309extern void brcms_c_suspend_mac_and_wait(struct brcms_c_info *wlc); 309void brcms_c_scan_start(struct brcms_c_info *wlc);
310extern void brcms_c_enable_mac(struct brcms_c_info *wlc); 310void brcms_c_scan_stop(struct brcms_c_info *wlc);
311extern void brcms_c_associate_upd(struct brcms_c_info *wlc, bool state); 311int brcms_c_get_curband(struct brcms_c_info *wlc);
312extern void brcms_c_scan_start(struct brcms_c_info *wlc); 312int brcms_c_set_channel(struct brcms_c_info *wlc, u16 channel);
313extern void brcms_c_scan_stop(struct brcms_c_info *wlc); 313int brcms_c_set_rate_limit(struct brcms_c_info *wlc, u16 srl, u16 lrl);
314extern int brcms_c_get_curband(struct brcms_c_info *wlc); 314void brcms_c_get_current_rateset(struct brcms_c_info *wlc,
315extern int brcms_c_set_channel(struct brcms_c_info *wlc, u16 channel);
316extern int brcms_c_set_rate_limit(struct brcms_c_info *wlc, u16 srl, u16 lrl);
317extern void brcms_c_get_current_rateset(struct brcms_c_info *wlc,
318 struct brcm_rateset *currs); 315 struct brcm_rateset *currs);
319extern int brcms_c_set_rateset(struct brcms_c_info *wlc, 316int brcms_c_set_rateset(struct brcms_c_info *wlc, struct brcm_rateset *rs);
320 struct brcm_rateset *rs); 317int brcms_c_set_beacon_period(struct brcms_c_info *wlc, u16 period);
321extern int brcms_c_set_beacon_period(struct brcms_c_info *wlc, u16 period); 318u16 brcms_c_get_phy_type(struct brcms_c_info *wlc, int phyidx);
322extern u16 brcms_c_get_phy_type(struct brcms_c_info *wlc, int phyidx); 319void brcms_c_set_shortslot_override(struct brcms_c_info *wlc,
323extern void brcms_c_set_shortslot_override(struct brcms_c_info *wlc,
324 s8 sslot_override); 320 s8 sslot_override);
325extern void brcms_c_set_beacon_listen_interval(struct brcms_c_info *wlc, 321void brcms_c_set_beacon_listen_interval(struct brcms_c_info *wlc, u8 interval);
326 u8 interval); 322u64 brcms_c_tsf_get(struct brcms_c_info *wlc);
327extern u64 brcms_c_tsf_get(struct brcms_c_info *wlc); 323void brcms_c_tsf_set(struct brcms_c_info *wlc, u64 tsf);
328extern void brcms_c_tsf_set(struct brcms_c_info *wlc, u64 tsf); 324int brcms_c_set_tx_power(struct brcms_c_info *wlc, int txpwr);
329extern int brcms_c_set_tx_power(struct brcms_c_info *wlc, int txpwr); 325int brcms_c_get_tx_power(struct brcms_c_info *wlc);
330extern int brcms_c_get_tx_power(struct brcms_c_info *wlc); 326bool brcms_c_check_radio_disabled(struct brcms_c_info *wlc);
331extern bool brcms_c_check_radio_disabled(struct brcms_c_info *wlc); 327void brcms_c_mute(struct brcms_c_info *wlc, bool on);
332extern void brcms_c_mute(struct brcms_c_info *wlc, bool on); 328bool brcms_c_tx_flush_completed(struct brcms_c_info *wlc);
333extern bool brcms_c_tx_flush_completed(struct brcms_c_info *wlc); 329void brcms_c_start_station(struct brcms_c_info *wlc, u8 *addr);
334extern void brcms_c_start_station(struct brcms_c_info *wlc, u8 *addr); 330void brcms_c_start_ap(struct brcms_c_info *wlc, u8 *addr, const u8 *bssid,
335extern void brcms_c_start_ap(struct brcms_c_info *wlc, u8 *addr, 331 u8 *ssid, size_t ssid_len);
336 const u8 *bssid, u8 *ssid, size_t ssid_len); 332void brcms_c_start_adhoc(struct brcms_c_info *wlc, u8 *addr);
337extern void brcms_c_start_adhoc(struct brcms_c_info *wlc, u8 *addr); 333void brcms_c_update_beacon(struct brcms_c_info *wlc);
338extern void brcms_c_update_beacon(struct brcms_c_info *wlc); 334void brcms_c_set_new_beacon(struct brcms_c_info *wlc, struct sk_buff *beacon,
339extern void brcms_c_set_new_beacon(struct brcms_c_info *wlc, 335 u16 tim_offset, u16 dtim_period);
340 struct sk_buff *beacon, u16 tim_offset, 336void brcms_c_set_new_probe_resp(struct brcms_c_info *wlc,
341 u16 dtim_period); 337 struct sk_buff *probe_resp);
342extern void brcms_c_set_new_probe_resp(struct brcms_c_info *wlc, 338void brcms_c_enable_probe_resp(struct brcms_c_info *wlc, bool enable);
343 struct sk_buff *probe_resp); 339void brcms_c_set_ssid(struct brcms_c_info *wlc, u8 *ssid, size_t ssid_len);
344extern void brcms_c_enable_probe_resp(struct brcms_c_info *wlc, bool enable);
345extern void brcms_c_set_ssid(struct brcms_c_info *wlc, u8 *ssid,
346 size_t ssid_len);
347 340
348#endif /* _BRCM_PUB_H_ */ 341#endif /* _BRCM_PUB_H_ */
diff --git a/drivers/net/wireless/brcm80211/brcmsmac/rate.h b/drivers/net/wireless/brcm80211/brcmsmac/rate.h
index 980d578825cc..5bb88b78ed64 100644
--- a/drivers/net/wireless/brcm80211/brcmsmac/rate.h
+++ b/drivers/net/wireless/brcm80211/brcmsmac/rate.h
@@ -216,34 +216,30 @@ static inline u8 cck_phy2mac_rate(u8 signal)
216 216
217/* sanitize, and sort a rateset with the basic bit(s) preserved, validate 217/* sanitize, and sort a rateset with the basic bit(s) preserved, validate
218 * rateset */ 218 * rateset */
219extern bool 219bool brcms_c_rate_hwrs_filter_sort_validate(struct brcms_c_rateset *rs,
220brcms_c_rate_hwrs_filter_sort_validate(struct brcms_c_rateset *rs, 220 const struct brcms_c_rateset *hw_rs,
221 const struct brcms_c_rateset *hw_rs, 221 bool check_brate, u8 txstreams);
222 bool check_brate, u8 txstreams);
223/* copy rateset src to dst as-is (no masking or sorting) */ 222/* copy rateset src to dst as-is (no masking or sorting) */
224extern void brcms_c_rateset_copy(const struct brcms_c_rateset *src, 223void brcms_c_rateset_copy(const struct brcms_c_rateset *src,
225 struct brcms_c_rateset *dst); 224 struct brcms_c_rateset *dst);
226 225
227/* would be nice to have these documented ... */ 226/* would be nice to have these documented ... */
228extern u32 brcms_c_compute_rspec(struct d11rxhdr *rxh, u8 *plcp); 227u32 brcms_c_compute_rspec(struct d11rxhdr *rxh, u8 *plcp);
229 228
230extern void brcms_c_rateset_filter(struct brcms_c_rateset *src, 229void brcms_c_rateset_filter(struct brcms_c_rateset *src,
231 struct brcms_c_rateset *dst, bool basic_only, u8 rates, uint xmask, 230 struct brcms_c_rateset *dst, bool basic_only,
232 bool mcsallow); 231 u8 rates, uint xmask, bool mcsallow);
233 232
234extern void 233void brcms_c_rateset_default(struct brcms_c_rateset *rs_tgt,
235brcms_c_rateset_default(struct brcms_c_rateset *rs_tgt, 234 const struct brcms_c_rateset *rs_hw, uint phy_type,
236 const struct brcms_c_rateset *rs_hw, uint phy_type, 235 int bandtype, bool cck_only, uint rate_mask,
237 int bandtype, bool cck_only, uint rate_mask, 236 bool mcsallow, u8 bw, u8 txstreams);
238 bool mcsallow, u8 bw, u8 txstreams); 237
239 238s16 brcms_c_rate_legacy_phyctl(uint rate);
240extern s16 brcms_c_rate_legacy_phyctl(uint rate); 239
241 240void brcms_c_rateset_mcs_upd(struct brcms_c_rateset *rs, u8 txstreams);
242extern void brcms_c_rateset_mcs_upd(struct brcms_c_rateset *rs, u8 txstreams); 241void brcms_c_rateset_mcs_clear(struct brcms_c_rateset *rateset);
243extern void brcms_c_rateset_mcs_clear(struct brcms_c_rateset *rateset); 242void brcms_c_rateset_mcs_build(struct brcms_c_rateset *rateset, u8 txstreams);
244extern void brcms_c_rateset_mcs_build(struct brcms_c_rateset *rateset, 243void brcms_c_rateset_bw_mcs_filter(struct brcms_c_rateset *rateset, u8 bw);
245 u8 txstreams);
246extern void brcms_c_rateset_bw_mcs_filter(struct brcms_c_rateset *rateset,
247 u8 bw);
248 244
249#endif /* _BRCM_RATE_H_ */ 245#endif /* _BRCM_RATE_H_ */
diff --git a/drivers/net/wireless/brcm80211/brcmsmac/stf.h b/drivers/net/wireless/brcm80211/brcmsmac/stf.h
index 19f6580f69be..ba9493009a33 100644
--- a/drivers/net/wireless/brcm80211/brcmsmac/stf.h
+++ b/drivers/net/wireless/brcm80211/brcmsmac/stf.h
@@ -19,24 +19,19 @@
19 19
20#include "types.h" 20#include "types.h"
21 21
22extern int brcms_c_stf_attach(struct brcms_c_info *wlc); 22int brcms_c_stf_attach(struct brcms_c_info *wlc);
23extern void brcms_c_stf_detach(struct brcms_c_info *wlc); 23void brcms_c_stf_detach(struct brcms_c_info *wlc);
24 24
25extern void brcms_c_tempsense_upd(struct brcms_c_info *wlc); 25void brcms_c_tempsense_upd(struct brcms_c_info *wlc);
26extern void brcms_c_stf_ss_algo_channel_get(struct brcms_c_info *wlc, 26void brcms_c_stf_ss_algo_channel_get(struct brcms_c_info *wlc,
27 u16 *ss_algo_channel, 27 u16 *ss_algo_channel, u16 chanspec);
28 u16 chanspec); 28int brcms_c_stf_ss_update(struct brcms_c_info *wlc, struct brcms_band *band);
29extern int brcms_c_stf_ss_update(struct brcms_c_info *wlc, 29void brcms_c_stf_phy_txant_upd(struct brcms_c_info *wlc);
30 struct brcms_band *band); 30int brcms_c_stf_txchain_set(struct brcms_c_info *wlc, s32 int_val, bool force);
31extern void brcms_c_stf_phy_txant_upd(struct brcms_c_info *wlc); 31bool brcms_c_stf_stbc_rx_set(struct brcms_c_info *wlc, s32 int_val);
32extern int brcms_c_stf_txchain_set(struct brcms_c_info *wlc, s32 int_val, 32void brcms_c_stf_phy_txant_upd(struct brcms_c_info *wlc);
33 bool force); 33void brcms_c_stf_phy_chain_calc(struct brcms_c_info *wlc);
34extern bool brcms_c_stf_stbc_rx_set(struct brcms_c_info *wlc, s32 int_val); 34u16 brcms_c_stf_phytxchain_sel(struct brcms_c_info *wlc, u32 rspec);
35extern void brcms_c_stf_phy_txant_upd(struct brcms_c_info *wlc); 35u16 brcms_c_stf_d11hdrs_phyctl_txant(struct brcms_c_info *wlc, u32 rspec);
36extern void brcms_c_stf_phy_chain_calc(struct brcms_c_info *wlc);
37extern u16 brcms_c_stf_phytxchain_sel(struct brcms_c_info *wlc,
38 u32 rspec);
39extern u16 brcms_c_stf_d11hdrs_phyctl_txant(struct brcms_c_info *wlc,
40 u32 rspec);
41 36
42#endif /* _BRCM_STF_H_ */ 37#endif /* _BRCM_STF_H_ */
diff --git a/drivers/net/wireless/brcm80211/brcmsmac/ucode_loader.h b/drivers/net/wireless/brcm80211/brcmsmac/ucode_loader.h
index 18750a814b4f..c87dd89bcb78 100644
--- a/drivers/net/wireless/brcm80211/brcmsmac/ucode_loader.h
+++ b/drivers/net/wireless/brcm80211/brcmsmac/ucode_loader.h
@@ -43,16 +43,14 @@ struct brcms_ucode {
43 u32 *bcm43xx_bomminor; 43 u32 *bcm43xx_bomminor;
44}; 44};
45 45
46extern int 46int brcms_ucode_data_init(struct brcms_info *wl, struct brcms_ucode *ucode);
47brcms_ucode_data_init(struct brcms_info *wl, struct brcms_ucode *ucode);
48 47
49extern void brcms_ucode_data_free(struct brcms_ucode *ucode); 48void brcms_ucode_data_free(struct brcms_ucode *ucode);
50 49
51extern int brcms_ucode_init_buf(struct brcms_info *wl, void **pbuf, 50int brcms_ucode_init_buf(struct brcms_info *wl, void **pbuf, unsigned int idx);
52 unsigned int idx); 51int brcms_ucode_init_uint(struct brcms_info *wl, size_t *n_bytes,
53extern int brcms_ucode_init_uint(struct brcms_info *wl, size_t *n_bytes, 52 unsigned int idx);
54 unsigned int idx); 53void brcms_ucode_free_buf(void *);
55extern void brcms_ucode_free_buf(void *); 54int brcms_check_firmwares(struct brcms_info *wl);
56extern int brcms_check_firmwares(struct brcms_info *wl);
57 55
58#endif /* _BRCM_UCODE_H_ */ 56#endif /* _BRCM_UCODE_H_ */
diff --git a/drivers/net/wireless/brcm80211/include/brcm_hw_ids.h b/drivers/net/wireless/brcm80211/include/brcm_hw_ids.h
index c1fe245bb07e..84113ea16f84 100644
--- a/drivers/net/wireless/brcm80211/include/brcm_hw_ids.h
+++ b/drivers/net/wireless/brcm80211/include/brcm_hw_ids.h
@@ -41,5 +41,6 @@
41#define BCM4331_CHIP_ID 0x4331 41#define BCM4331_CHIP_ID 0x4331
42#define BCM4334_CHIP_ID 0x4334 42#define BCM4334_CHIP_ID 0x4334
43#define BCM4335_CHIP_ID 0x4335 43#define BCM4335_CHIP_ID 0x4335
44#define BCM4339_CHIP_ID 0x4339
44 45
45#endif /* _BRCM_HW_IDS_H_ */ 46#endif /* _BRCM_HW_IDS_H_ */
diff --git a/drivers/net/wireless/brcm80211/include/brcmu_d11.h b/drivers/net/wireless/brcm80211/include/brcmu_d11.h
index 92623f02b1c0..8660a2cba098 100644
--- a/drivers/net/wireless/brcm80211/include/brcmu_d11.h
+++ b/drivers/net/wireless/brcm80211/include/brcmu_d11.h
@@ -140,6 +140,6 @@ struct brcmu_d11inf {
140 void (*decchspec)(struct brcmu_chan *ch); 140 void (*decchspec)(struct brcmu_chan *ch);
141}; 141};
142 142
143extern void brcmu_d11_attach(struct brcmu_d11inf *d11inf); 143void brcmu_d11_attach(struct brcmu_d11inf *d11inf);
144 144
145#endif /* _BRCMU_CHANNELS_H_ */ 145#endif /* _BRCMU_CHANNELS_H_ */
diff --git a/drivers/net/wireless/brcm80211/include/brcmu_utils.h b/drivers/net/wireless/brcm80211/include/brcmu_utils.h
index 898cacb8d01d..8ba445b3fd72 100644
--- a/drivers/net/wireless/brcm80211/include/brcmu_utils.h
+++ b/drivers/net/wireless/brcm80211/include/brcmu_utils.h
@@ -114,31 +114,29 @@ static inline struct sk_buff *pktq_ppeek_tail(struct pktq *pq, int prec)
114 return skb_peek_tail(&pq->q[prec].skblist); 114 return skb_peek_tail(&pq->q[prec].skblist);
115} 115}
116 116
117extern struct sk_buff *brcmu_pktq_penq(struct pktq *pq, int prec, 117struct sk_buff *brcmu_pktq_penq(struct pktq *pq, int prec, struct sk_buff *p);
118 struct sk_buff *p); 118struct sk_buff *brcmu_pktq_penq_head(struct pktq *pq, int prec,
119extern struct sk_buff *brcmu_pktq_penq_head(struct pktq *pq, int prec, 119 struct sk_buff *p);
120 struct sk_buff *p); 120struct sk_buff *brcmu_pktq_pdeq(struct pktq *pq, int prec);
121extern struct sk_buff *brcmu_pktq_pdeq(struct pktq *pq, int prec); 121struct sk_buff *brcmu_pktq_pdeq_tail(struct pktq *pq, int prec);
122extern struct sk_buff *brcmu_pktq_pdeq_tail(struct pktq *pq, int prec); 122struct sk_buff *brcmu_pktq_pdeq_match(struct pktq *pq, int prec,
123extern struct sk_buff *brcmu_pktq_pdeq_match(struct pktq *pq, int prec, 123 bool (*match_fn)(struct sk_buff *p,
124 bool (*match_fn)(struct sk_buff *p, 124 void *arg),
125 void *arg), 125 void *arg);
126 void *arg);
127 126
128/* packet primitives */ 127/* packet primitives */
129extern struct sk_buff *brcmu_pkt_buf_get_skb(uint len); 128struct sk_buff *brcmu_pkt_buf_get_skb(uint len);
130extern void brcmu_pkt_buf_free_skb(struct sk_buff *skb); 129void brcmu_pkt_buf_free_skb(struct sk_buff *skb);
131 130
132/* Empty the queue at particular precedence level */ 131/* Empty the queue at particular precedence level */
133/* callback function fn(pkt, arg) returns true if pkt belongs to if */ 132/* callback function fn(pkt, arg) returns true if pkt belongs to if */
134extern void brcmu_pktq_pflush(struct pktq *pq, int prec, 133void brcmu_pktq_pflush(struct pktq *pq, int prec, bool dir,
135 bool dir, bool (*fn)(struct sk_buff *, void *), void *arg); 134 bool (*fn)(struct sk_buff *, void *), void *arg);
136 135
137/* operations on a set of precedences in packet queue */ 136/* operations on a set of precedences in packet queue */
138 137
139extern int brcmu_pktq_mlen(struct pktq *pq, uint prec_bmp); 138int brcmu_pktq_mlen(struct pktq *pq, uint prec_bmp);
140extern struct sk_buff *brcmu_pktq_mdeq(struct pktq *pq, uint prec_bmp, 139struct sk_buff *brcmu_pktq_mdeq(struct pktq *pq, uint prec_bmp, int *prec_out);
141 int *prec_out);
142 140
143/* operations on packet queue as a whole */ 141/* operations on packet queue as a whole */
144 142
@@ -167,11 +165,11 @@ static inline bool pktq_empty(struct pktq *pq)
167 return pq->len == 0; 165 return pq->len == 0;
168} 166}
169 167
170extern void brcmu_pktq_init(struct pktq *pq, int num_prec, int max_len); 168void brcmu_pktq_init(struct pktq *pq, int num_prec, int max_len);
171/* prec_out may be NULL if caller is not interested in return value */ 169/* prec_out may be NULL if caller is not interested in return value */
172extern struct sk_buff *brcmu_pktq_peek_tail(struct pktq *pq, int *prec_out); 170struct sk_buff *brcmu_pktq_peek_tail(struct pktq *pq, int *prec_out);
173extern void brcmu_pktq_flush(struct pktq *pq, bool dir, 171void brcmu_pktq_flush(struct pktq *pq, bool dir,
174 bool (*fn)(struct sk_buff *, void *), void *arg); 172 bool (*fn)(struct sk_buff *, void *), void *arg);
175 173
176/* externs */ 174/* externs */
177/* ip address */ 175/* ip address */
@@ -204,13 +202,13 @@ static inline u16 brcmu_maskget16(u16 var, u16 mask, u8 shift)
204/* externs */ 202/* externs */
205/* format/print */ 203/* format/print */
206#ifdef DEBUG 204#ifdef DEBUG
207extern void brcmu_prpkt(const char *msg, struct sk_buff *p0); 205void brcmu_prpkt(const char *msg, struct sk_buff *p0);
208#else 206#else
209#define brcmu_prpkt(a, b) 207#define brcmu_prpkt(a, b)
210#endif /* DEBUG */ 208#endif /* DEBUG */
211 209
212#ifdef DEBUG 210#ifdef DEBUG
213extern __printf(3, 4) 211__printf(3, 4)
214void brcmu_dbg_hex_dump(const void *data, size_t size, const char *fmt, ...); 212void brcmu_dbg_hex_dump(const void *data, size_t size, const char *fmt, ...);
215#else 213#else
216__printf(3, 4) 214__printf(3, 4)
diff --git a/drivers/net/wireless/cw1200/cw1200_spi.c b/drivers/net/wireless/cw1200/cw1200_spi.c
index 755a0c8edfe1..40078f5f932e 100644
--- a/drivers/net/wireless/cw1200/cw1200_spi.c
+++ b/drivers/net/wireless/cw1200/cw1200_spi.c
@@ -365,7 +365,7 @@ static struct hwbus_ops cw1200_spi_hwbus_ops = {
365static int cw1200_spi_probe(struct spi_device *func) 365static int cw1200_spi_probe(struct spi_device *func)
366{ 366{
367 const struct cw1200_platform_data_spi *plat_data = 367 const struct cw1200_platform_data_spi *plat_data =
368 func->dev.platform_data; 368 dev_get_platdata(&func->dev);
369 struct hwbus_priv *self; 369 struct hwbus_priv *self;
370 int status; 370 int status;
371 371
@@ -443,7 +443,7 @@ static int cw1200_spi_disconnect(struct spi_device *func)
443 } 443 }
444 kfree(self); 444 kfree(self);
445 } 445 }
446 cw1200_spi_off(func->dev.platform_data); 446 cw1200_spi_off(dev_get_platdata(&func->dev));
447 447
448 return 0; 448 return 0;
449} 449}
diff --git a/drivers/net/wireless/hostap/hostap_info.c b/drivers/net/wireless/hostap/hostap_info.c
index 970a48baaf80..de7c4ffec309 100644
--- a/drivers/net/wireless/hostap/hostap_info.c
+++ b/drivers/net/wireless/hostap/hostap_info.c
@@ -217,7 +217,7 @@ static void prism2_host_roaming(local_info_t *local)
217 } 217 }
218 } 218 }
219 219
220 memcpy(req.bssid, selected->bssid, 6); 220 memcpy(req.bssid, selected->bssid, ETH_ALEN);
221 req.channel = selected->chid; 221 req.channel = selected->chid;
222 spin_unlock_irqrestore(&local->lock, flags); 222 spin_unlock_irqrestore(&local->lock, flags);
223 223
diff --git a/drivers/net/wireless/ipw2x00/ipw2200.c b/drivers/net/wireless/ipw2x00/ipw2200.c
index 6b823a1ab789..81903e33d5b1 100644
--- a/drivers/net/wireless/ipw2x00/ipw2200.c
+++ b/drivers/net/wireless/ipw2x00/ipw2200.c
@@ -2698,7 +2698,7 @@ static u16 eeprom_read_u16(struct ipw_priv *priv, u8 addr)
2698/* data's copy of the eeprom data */ 2698/* data's copy of the eeprom data */
2699static void eeprom_parse_mac(struct ipw_priv *priv, u8 * mac) 2699static void eeprom_parse_mac(struct ipw_priv *priv, u8 * mac)
2700{ 2700{
2701 memcpy(mac, &priv->eeprom[EEPROM_MAC_ADDRESS], 6); 2701 memcpy(mac, &priv->eeprom[EEPROM_MAC_ADDRESS], ETH_ALEN);
2702} 2702}
2703 2703
2704static void ipw_read_eeprom(struct ipw_priv *priv) 2704static void ipw_read_eeprom(struct ipw_priv *priv)
@@ -11885,7 +11885,6 @@ static int ipw_pci_probe(struct pci_dev *pdev,
11885 pci_release_regions(pdev); 11885 pci_release_regions(pdev);
11886 out_pci_disable_device: 11886 out_pci_disable_device:
11887 pci_disable_device(pdev); 11887 pci_disable_device(pdev);
11888 pci_set_drvdata(pdev, NULL);
11889 out_free_libipw: 11888 out_free_libipw:
11890 free_libipw(priv->net_dev, 0); 11889 free_libipw(priv->net_dev, 0);
11891 out: 11890 out:
@@ -11966,7 +11965,6 @@ static void ipw_pci_remove(struct pci_dev *pdev)
11966 iounmap(priv->hw_base); 11965 iounmap(priv->hw_base);
11967 pci_release_regions(pdev); 11966 pci_release_regions(pdev);
11968 pci_disable_device(pdev); 11967 pci_disable_device(pdev);
11969 pci_set_drvdata(pdev, NULL);
11970 /* wiphy_unregister needs to be here, before free_libipw */ 11968 /* wiphy_unregister needs to be here, before free_libipw */
11971 wiphy_unregister(priv->ieee->wdev.wiphy); 11969 wiphy_unregister(priv->ieee->wdev.wiphy);
11972 kfree(priv->ieee->a_band.channels); 11970 kfree(priv->ieee->a_band.channels);
diff --git a/drivers/net/wireless/ipw2x00/libipw.h b/drivers/net/wireless/ipw2x00/libipw.h
index 6eede52ad8c0..5ce2f59d3378 100644
--- a/drivers/net/wireless/ipw2x00/libipw.h
+++ b/drivers/net/wireless/ipw2x00/libipw.h
@@ -950,66 +950,55 @@ static inline int libipw_is_cck_rate(u8 rate)
950} 950}
951 951
952/* libipw.c */ 952/* libipw.c */
953extern void free_libipw(struct net_device *dev, int monitor); 953void free_libipw(struct net_device *dev, int monitor);
954extern struct net_device *alloc_libipw(int sizeof_priv, int monitor); 954struct net_device *alloc_libipw(int sizeof_priv, int monitor);
955extern int libipw_change_mtu(struct net_device *dev, int new_mtu); 955int libipw_change_mtu(struct net_device *dev, int new_mtu);
956 956
957extern void libipw_networks_age(struct libipw_device *ieee, 957void libipw_networks_age(struct libipw_device *ieee, unsigned long age_secs);
958 unsigned long age_secs);
959 958
960extern int libipw_set_encryption(struct libipw_device *ieee); 959int libipw_set_encryption(struct libipw_device *ieee);
961 960
962/* libipw_tx.c */ 961/* libipw_tx.c */
963extern netdev_tx_t libipw_xmit(struct sk_buff *skb, 962netdev_tx_t libipw_xmit(struct sk_buff *skb, struct net_device *dev);
964 struct net_device *dev); 963void libipw_txb_free(struct libipw_txb *);
965extern void libipw_txb_free(struct libipw_txb *);
966 964
967/* libipw_rx.c */ 965/* libipw_rx.c */
968extern void libipw_rx_any(struct libipw_device *ieee, 966void libipw_rx_any(struct libipw_device *ieee, struct sk_buff *skb,
969 struct sk_buff *skb, struct libipw_rx_stats *stats); 967 struct libipw_rx_stats *stats);
970extern int libipw_rx(struct libipw_device *ieee, struct sk_buff *skb, 968int libipw_rx(struct libipw_device *ieee, struct sk_buff *skb,
971 struct libipw_rx_stats *rx_stats); 969 struct libipw_rx_stats *rx_stats);
972/* make sure to set stats->len */ 970/* make sure to set stats->len */
973extern void libipw_rx_mgt(struct libipw_device *ieee, 971void libipw_rx_mgt(struct libipw_device *ieee, struct libipw_hdr_4addr *header,
974 struct libipw_hdr_4addr *header, 972 struct libipw_rx_stats *stats);
975 struct libipw_rx_stats *stats); 973void libipw_network_reset(struct libipw_network *network);
976extern void libipw_network_reset(struct libipw_network *network);
977 974
978/* libipw_geo.c */ 975/* libipw_geo.c */
979extern const struct libipw_geo *libipw_get_geo(struct libipw_device 976const struct libipw_geo *libipw_get_geo(struct libipw_device *ieee);
980 *ieee); 977void libipw_set_geo(struct libipw_device *ieee, const struct libipw_geo *geo);
981extern void libipw_set_geo(struct libipw_device *ieee, 978
982 const struct libipw_geo *geo); 979int libipw_is_valid_channel(struct libipw_device *ieee, u8 channel);
983 980int libipw_channel_to_index(struct libipw_device *ieee, u8 channel);
984extern int libipw_is_valid_channel(struct libipw_device *ieee, 981u8 libipw_freq_to_channel(struct libipw_device *ieee, u32 freq);
985 u8 channel); 982u8 libipw_get_channel_flags(struct libipw_device *ieee, u8 channel);
986extern int libipw_channel_to_index(struct libipw_device *ieee, 983const struct libipw_channel *libipw_get_channel(struct libipw_device *ieee,
987 u8 channel); 984 u8 channel);
988extern u8 libipw_freq_to_channel(struct libipw_device *ieee, u32 freq); 985u32 libipw_channel_to_freq(struct libipw_device *ieee, u8 channel);
989extern u8 libipw_get_channel_flags(struct libipw_device *ieee,
990 u8 channel);
991extern const struct libipw_channel *libipw_get_channel(struct
992 libipw_device
993 *ieee, u8 channel);
994extern u32 libipw_channel_to_freq(struct libipw_device * ieee,
995 u8 channel);
996 986
997/* libipw_wx.c */ 987/* libipw_wx.c */
998extern int libipw_wx_get_scan(struct libipw_device *ieee, 988int libipw_wx_get_scan(struct libipw_device *ieee, struct iw_request_info *info,
999 struct iw_request_info *info, 989 union iwreq_data *wrqu, char *key);
1000 union iwreq_data *wrqu, char *key); 990int libipw_wx_set_encode(struct libipw_device *ieee,
1001extern int libipw_wx_set_encode(struct libipw_device *ieee, 991 struct iw_request_info *info, union iwreq_data *wrqu,
1002 struct iw_request_info *info, 992 char *key);
1003 union iwreq_data *wrqu, char *key); 993int libipw_wx_get_encode(struct libipw_device *ieee,
1004extern int libipw_wx_get_encode(struct libipw_device *ieee, 994 struct iw_request_info *info, union iwreq_data *wrqu,
1005 struct iw_request_info *info, 995 char *key);
1006 union iwreq_data *wrqu, char *key); 996int libipw_wx_set_encodeext(struct libipw_device *ieee,
1007extern int libipw_wx_set_encodeext(struct libipw_device *ieee, 997 struct iw_request_info *info,
1008 struct iw_request_info *info, 998 union iwreq_data *wrqu, char *extra);
1009 union iwreq_data *wrqu, char *extra); 999int libipw_wx_get_encodeext(struct libipw_device *ieee,
1010extern int libipw_wx_get_encodeext(struct libipw_device *ieee, 1000 struct iw_request_info *info,
1011 struct iw_request_info *info, 1001 union iwreq_data *wrqu, char *extra);
1012 union iwreq_data *wrqu, char *extra);
1013 1002
1014static inline void libipw_increment_scans(struct libipw_device *ieee) 1003static inline void libipw_increment_scans(struct libipw_device *ieee)
1015{ 1004{
diff --git a/drivers/net/wireless/iwlegacy/3945-mac.c b/drivers/net/wireless/iwlegacy/3945-mac.c
index 9581d07a4242..dea3b50d68b9 100644
--- a/drivers/net/wireless/iwlegacy/3945-mac.c
+++ b/drivers/net/wireless/iwlegacy/3945-mac.c
@@ -3811,7 +3811,6 @@ out_iounmap:
3811out_pci_release_regions: 3811out_pci_release_regions:
3812 pci_release_regions(pdev); 3812 pci_release_regions(pdev);
3813out_pci_disable_device: 3813out_pci_disable_device:
3814 pci_set_drvdata(pdev, NULL);
3815 pci_disable_device(pdev); 3814 pci_disable_device(pdev);
3816out_ieee80211_free_hw: 3815out_ieee80211_free_hw:
3817 ieee80211_free_hw(il->hw); 3816 ieee80211_free_hw(il->hw);
@@ -3888,7 +3887,6 @@ il3945_pci_remove(struct pci_dev *pdev)
3888 iounmap(il->hw_base); 3887 iounmap(il->hw_base);
3889 pci_release_regions(pdev); 3888 pci_release_regions(pdev);
3890 pci_disable_device(pdev); 3889 pci_disable_device(pdev);
3891 pci_set_drvdata(pdev, NULL);
3892 3890
3893 il_free_channel_map(il); 3891 il_free_channel_map(il);
3894 il_free_geos(il); 3892 il_free_geos(il);
diff --git a/drivers/net/wireless/iwlegacy/3945.h b/drivers/net/wireless/iwlegacy/3945.h
index 9a8703def0ba..00030d43a194 100644
--- a/drivers/net/wireless/iwlegacy/3945.h
+++ b/drivers/net/wireless/iwlegacy/3945.h
@@ -189,15 +189,14 @@ struct il3945_ibss_seq {
189 * for use by iwl-*.c 189 * for use by iwl-*.c
190 * 190 *
191 *****************************************************************************/ 191 *****************************************************************************/
192extern int il3945_calc_db_from_ratio(int sig_ratio); 192int il3945_calc_db_from_ratio(int sig_ratio);
193extern void il3945_rx_replenish(void *data); 193void il3945_rx_replenish(void *data);
194extern void il3945_rx_queue_reset(struct il_priv *il, struct il_rx_queue *rxq); 194void il3945_rx_queue_reset(struct il_priv *il, struct il_rx_queue *rxq);
195extern unsigned int il3945_fill_beacon_frame(struct il_priv *il, 195unsigned int il3945_fill_beacon_frame(struct il_priv *il,
196 struct ieee80211_hdr *hdr, 196 struct ieee80211_hdr *hdr, int left);
197 int left); 197int il3945_dump_nic_event_log(struct il_priv *il, bool full_log, char **buf,
198extern int il3945_dump_nic_event_log(struct il_priv *il, bool full_log, 198 bool display);
199 char **buf, bool display); 199void il3945_dump_nic_error_log(struct il_priv *il);
200extern void il3945_dump_nic_error_log(struct il_priv *il);
201 200
202/****************************************************************************** 201/******************************************************************************
203 * 202 *
@@ -215,39 +214,36 @@ extern void il3945_dump_nic_error_log(struct il_priv *il);
215 * il3945_mac_ <-- mac80211 callback 214 * il3945_mac_ <-- mac80211 callback
216 * 215 *
217 ****************************************************************************/ 216 ****************************************************************************/
218extern void il3945_hw_handler_setup(struct il_priv *il); 217void il3945_hw_handler_setup(struct il_priv *il);
219extern void il3945_hw_setup_deferred_work(struct il_priv *il); 218void il3945_hw_setup_deferred_work(struct il_priv *il);
220extern void il3945_hw_cancel_deferred_work(struct il_priv *il); 219void il3945_hw_cancel_deferred_work(struct il_priv *il);
221extern int il3945_hw_rxq_stop(struct il_priv *il); 220int il3945_hw_rxq_stop(struct il_priv *il);
222extern int il3945_hw_set_hw_params(struct il_priv *il); 221int il3945_hw_set_hw_params(struct il_priv *il);
223extern int il3945_hw_nic_init(struct il_priv *il); 222int il3945_hw_nic_init(struct il_priv *il);
224extern int il3945_hw_nic_stop_master(struct il_priv *il); 223int il3945_hw_nic_stop_master(struct il_priv *il);
225extern void il3945_hw_txq_ctx_free(struct il_priv *il); 224void il3945_hw_txq_ctx_free(struct il_priv *il);
226extern void il3945_hw_txq_ctx_stop(struct il_priv *il); 225void il3945_hw_txq_ctx_stop(struct il_priv *il);
227extern int il3945_hw_nic_reset(struct il_priv *il); 226int il3945_hw_nic_reset(struct il_priv *il);
228extern int il3945_hw_txq_attach_buf_to_tfd(struct il_priv *il, 227int il3945_hw_txq_attach_buf_to_tfd(struct il_priv *il, struct il_tx_queue *txq,
229 struct il_tx_queue *txq, 228 dma_addr_t addr, u16 len, u8 reset, u8 pad);
230 dma_addr_t addr, u16 len, u8 reset, 229void il3945_hw_txq_free_tfd(struct il_priv *il, struct il_tx_queue *txq);
231 u8 pad); 230int il3945_hw_get_temperature(struct il_priv *il);
232extern void il3945_hw_txq_free_tfd(struct il_priv *il, struct il_tx_queue *txq); 231int il3945_hw_tx_queue_init(struct il_priv *il, struct il_tx_queue *txq);
233extern int il3945_hw_get_temperature(struct il_priv *il); 232unsigned int il3945_hw_get_beacon_cmd(struct il_priv *il,
234extern int il3945_hw_tx_queue_init(struct il_priv *il, struct il_tx_queue *txq); 233 struct il3945_frame *frame, u8 rate);
235extern unsigned int il3945_hw_get_beacon_cmd(struct il_priv *il,
236 struct il3945_frame *frame,
237 u8 rate);
238void il3945_hw_build_tx_cmd_rate(struct il_priv *il, struct il_device_cmd *cmd, 234void il3945_hw_build_tx_cmd_rate(struct il_priv *il, struct il_device_cmd *cmd,
239 struct ieee80211_tx_info *info, 235 struct ieee80211_tx_info *info,
240 struct ieee80211_hdr *hdr, int sta_id); 236 struct ieee80211_hdr *hdr, int sta_id);
241extern int il3945_hw_reg_send_txpower(struct il_priv *il); 237int il3945_hw_reg_send_txpower(struct il_priv *il);
242extern int il3945_hw_reg_set_txpower(struct il_priv *il, s8 power); 238int il3945_hw_reg_set_txpower(struct il_priv *il, s8 power);
243extern void il3945_hdl_stats(struct il_priv *il, struct il_rx_buf *rxb); 239void il3945_hdl_stats(struct il_priv *il, struct il_rx_buf *rxb);
244void il3945_hdl_c_stats(struct il_priv *il, struct il_rx_buf *rxb); 240void il3945_hdl_c_stats(struct il_priv *il, struct il_rx_buf *rxb);
245extern void il3945_disable_events(struct il_priv *il); 241void il3945_disable_events(struct il_priv *il);
246extern int il4965_get_temperature(const struct il_priv *il); 242int il4965_get_temperature(const struct il_priv *il);
247extern void il3945_post_associate(struct il_priv *il); 243void il3945_post_associate(struct il_priv *il);
248extern void il3945_config_ap(struct il_priv *il); 244void il3945_config_ap(struct il_priv *il);
249 245
250extern int il3945_commit_rxon(struct il_priv *il); 246int il3945_commit_rxon(struct il_priv *il);
251 247
252/** 248/**
253 * il3945_hw_find_station - Find station id for a given BSSID 249 * il3945_hw_find_station - Find station id for a given BSSID
@@ -257,14 +253,14 @@ extern int il3945_commit_rxon(struct il_priv *il);
257 * not yet been merged into a single common layer for managing the 253 * not yet been merged into a single common layer for managing the
258 * station tables. 254 * station tables.
259 */ 255 */
260extern u8 il3945_hw_find_station(struct il_priv *il, const u8 * bssid); 256u8 il3945_hw_find_station(struct il_priv *il, const u8 *bssid);
261 257
262extern __le32 il3945_get_antenna_flags(const struct il_priv *il); 258__le32 il3945_get_antenna_flags(const struct il_priv *il);
263extern int il3945_init_hw_rate_table(struct il_priv *il); 259int il3945_init_hw_rate_table(struct il_priv *il);
264extern void il3945_reg_txpower_periodic(struct il_priv *il); 260void il3945_reg_txpower_periodic(struct il_priv *il);
265extern int il3945_txpower_set_from_eeprom(struct il_priv *il); 261int il3945_txpower_set_from_eeprom(struct il_priv *il);
266 262
267extern int il3945_rs_next_rate(struct il_priv *il, int rate); 263int il3945_rs_next_rate(struct il_priv *il, int rate);
268 264
269/* scanning */ 265/* scanning */
270int il3945_request_scan(struct il_priv *il, struct ieee80211_vif *vif); 266int il3945_request_scan(struct il_priv *il, struct ieee80211_vif *vif);
diff --git a/drivers/net/wireless/iwlegacy/4965-mac.c b/drivers/net/wireless/iwlegacy/4965-mac.c
index 5ab50a5b48b1..3982ab76f375 100644
--- a/drivers/net/wireless/iwlegacy/4965-mac.c
+++ b/drivers/net/wireless/iwlegacy/4965-mac.c
@@ -6706,7 +6706,6 @@ out_free_eeprom:
6706out_iounmap: 6706out_iounmap:
6707 iounmap(il->hw_base); 6707 iounmap(il->hw_base);
6708out_pci_release_regions: 6708out_pci_release_regions:
6709 pci_set_drvdata(pdev, NULL);
6710 pci_release_regions(pdev); 6709 pci_release_regions(pdev);
6711out_pci_disable_device: 6710out_pci_disable_device:
6712 pci_disable_device(pdev); 6711 pci_disable_device(pdev);
@@ -6787,7 +6786,6 @@ il4965_pci_remove(struct pci_dev *pdev)
6787 iounmap(il->hw_base); 6786 iounmap(il->hw_base);
6788 pci_release_regions(pdev); 6787 pci_release_regions(pdev);
6789 pci_disable_device(pdev); 6788 pci_disable_device(pdev);
6790 pci_set_drvdata(pdev, NULL);
6791 6789
6792 il4965_uninit_drv(il); 6790 il4965_uninit_drv(il);
6793 6791
diff --git a/drivers/net/wireless/iwlegacy/4965.h b/drivers/net/wireless/iwlegacy/4965.h
index 1b15b0b2292b..337dfcf3bbde 100644
--- a/drivers/net/wireless/iwlegacy/4965.h
+++ b/drivers/net/wireless/iwlegacy/4965.h
@@ -272,7 +272,7 @@ il4965_hw_valid_rtc_data_addr(u32 addr)
272 ((t) < IL_TX_POWER_TEMPERATURE_MIN || \ 272 ((t) < IL_TX_POWER_TEMPERATURE_MIN || \
273 (t) > IL_TX_POWER_TEMPERATURE_MAX) 273 (t) > IL_TX_POWER_TEMPERATURE_MAX)
274 274
275extern void il4965_temperature_calib(struct il_priv *il); 275void il4965_temperature_calib(struct il_priv *il);
276/********************* END TEMPERATURE ***************************************/ 276/********************* END TEMPERATURE ***************************************/
277 277
278/********************* START TXPOWER *****************************************/ 278/********************* START TXPOWER *****************************************/
diff --git a/drivers/net/wireless/iwlegacy/common.h b/drivers/net/wireless/iwlegacy/common.h
index 83f8ed8a5528..ad123d66ab6c 100644
--- a/drivers/net/wireless/iwlegacy/common.h
+++ b/drivers/net/wireless/iwlegacy/common.h
@@ -858,9 +858,9 @@ struct il_hw_params {
858 * il4965_mac_ <-- mac80211 callback 858 * il4965_mac_ <-- mac80211 callback
859 * 859 *
860 ****************************************************************************/ 860 ****************************************************************************/
861extern void il4965_update_chain_flags(struct il_priv *il); 861void il4965_update_chain_flags(struct il_priv *il);
862extern const u8 il_bcast_addr[ETH_ALEN]; 862extern const u8 il_bcast_addr[ETH_ALEN];
863extern int il_queue_space(const struct il_queue *q); 863int il_queue_space(const struct il_queue *q);
864static inline int 864static inline int
865il_queue_used(const struct il_queue *q, int i) 865il_queue_used(const struct il_queue *q, int i)
866{ 866{
@@ -1727,7 +1727,7 @@ int il_alloc_txq_mem(struct il_priv *il);
1727void il_free_txq_mem(struct il_priv *il); 1727void il_free_txq_mem(struct il_priv *il);
1728 1728
1729#ifdef CONFIG_IWLEGACY_DEBUGFS 1729#ifdef CONFIG_IWLEGACY_DEBUGFS
1730extern void il_update_stats(struct il_priv *il, bool is_tx, __le16 fc, u16 len); 1730void il_update_stats(struct il_priv *il, bool is_tx, __le16 fc, u16 len);
1731#else 1731#else
1732static inline void 1732static inline void
1733il_update_stats(struct il_priv *il, bool is_tx, __le16 fc, u16 len) 1733il_update_stats(struct il_priv *il, bool is_tx, __le16 fc, u16 len)
@@ -1760,12 +1760,12 @@ void il_chswitch_done(struct il_priv *il, bool is_success);
1760/***************************************************** 1760/*****************************************************
1761* TX 1761* TX
1762******************************************************/ 1762******************************************************/
1763extern void il_txq_update_write_ptr(struct il_priv *il, struct il_tx_queue *txq); 1763void il_txq_update_write_ptr(struct il_priv *il, struct il_tx_queue *txq);
1764extern int il_tx_queue_init(struct il_priv *il, u32 txq_id); 1764int il_tx_queue_init(struct il_priv *il, u32 txq_id);
1765extern void il_tx_queue_reset(struct il_priv *il, u32 txq_id); 1765void il_tx_queue_reset(struct il_priv *il, u32 txq_id);
1766extern void il_tx_queue_unmap(struct il_priv *il, int txq_id); 1766void il_tx_queue_unmap(struct il_priv *il, int txq_id);
1767extern void il_tx_queue_free(struct il_priv *il, int txq_id); 1767void il_tx_queue_free(struct il_priv *il, int txq_id);
1768extern void il_setup_watchdog(struct il_priv *il); 1768void il_setup_watchdog(struct il_priv *il);
1769/***************************************************** 1769/*****************************************************
1770 * TX power 1770 * TX power
1771 ****************************************************/ 1771 ****************************************************/
@@ -1931,10 +1931,10 @@ il_is_ready_rf(struct il_priv *il)
1931 return il_is_ready(il); 1931 return il_is_ready(il);
1932} 1932}
1933 1933
1934extern void il_send_bt_config(struct il_priv *il); 1934void il_send_bt_config(struct il_priv *il);
1935extern int il_send_stats_request(struct il_priv *il, u8 flags, bool clear); 1935int il_send_stats_request(struct il_priv *il, u8 flags, bool clear);
1936extern void il_apm_stop(struct il_priv *il); 1936void il_apm_stop(struct il_priv *il);
1937extern void _il_apm_stop(struct il_priv *il); 1937void _il_apm_stop(struct il_priv *il);
1938 1938
1939int il_apm_init(struct il_priv *il); 1939int il_apm_init(struct il_priv *il);
1940 1940
@@ -1968,15 +1968,15 @@ void il_tx_cmd_protection(struct il_priv *il, struct ieee80211_tx_info *info,
1968 1968
1969irqreturn_t il_isr(int irq, void *data); 1969irqreturn_t il_isr(int irq, void *data);
1970 1970
1971extern void il_set_bit(struct il_priv *p, u32 r, u32 m); 1971void il_set_bit(struct il_priv *p, u32 r, u32 m);
1972extern void il_clear_bit(struct il_priv *p, u32 r, u32 m); 1972void il_clear_bit(struct il_priv *p, u32 r, u32 m);
1973extern bool _il_grab_nic_access(struct il_priv *il); 1973bool _il_grab_nic_access(struct il_priv *il);
1974extern int _il_poll_bit(struct il_priv *il, u32 addr, u32 bits, u32 mask, int timeout); 1974int _il_poll_bit(struct il_priv *il, u32 addr, u32 bits, u32 mask, int timeout);
1975extern int il_poll_bit(struct il_priv *il, u32 addr, u32 mask, int timeout); 1975int il_poll_bit(struct il_priv *il, u32 addr, u32 mask, int timeout);
1976extern u32 il_rd_prph(struct il_priv *il, u32 reg); 1976u32 il_rd_prph(struct il_priv *il, u32 reg);
1977extern void il_wr_prph(struct il_priv *il, u32 addr, u32 val); 1977void il_wr_prph(struct il_priv *il, u32 addr, u32 val);
1978extern u32 il_read_targ_mem(struct il_priv *il, u32 addr); 1978u32 il_read_targ_mem(struct il_priv *il, u32 addr);
1979extern void il_write_targ_mem(struct il_priv *il, u32 addr, u32 val); 1979void il_write_targ_mem(struct il_priv *il, u32 addr, u32 val);
1980 1980
1981static inline void 1981static inline void
1982_il_write8(struct il_priv *il, u32 ofs, u8 val) 1982_il_write8(struct il_priv *il, u32 ofs, u8 val)
@@ -2868,13 +2868,13 @@ il4965_first_antenna(u8 mask)
2868 * The specific throughput table used is based on the type of network 2868 * The specific throughput table used is based on the type of network
2869 * the associated with, including A, B, G, and G w/ TGG protection 2869 * the associated with, including A, B, G, and G w/ TGG protection
2870 */ 2870 */
2871extern void il3945_rate_scale_init(struct ieee80211_hw *hw, s32 sta_id); 2871void il3945_rate_scale_init(struct ieee80211_hw *hw, s32 sta_id);
2872 2872
2873/* Initialize station's rate scaling information after adding station */ 2873/* Initialize station's rate scaling information after adding station */
2874extern void il4965_rs_rate_init(struct il_priv *il, struct ieee80211_sta *sta, 2874void il4965_rs_rate_init(struct il_priv *il, struct ieee80211_sta *sta,
2875 u8 sta_id); 2875 u8 sta_id);
2876extern void il3945_rs_rate_init(struct il_priv *il, struct ieee80211_sta *sta, 2876void il3945_rs_rate_init(struct il_priv *il, struct ieee80211_sta *sta,
2877 u8 sta_id); 2877 u8 sta_id);
2878 2878
2879/** 2879/**
2880 * il_rate_control_register - Register the rate control algorithm callbacks 2880 * il_rate_control_register - Register the rate control algorithm callbacks
@@ -2886,8 +2886,8 @@ extern void il3945_rs_rate_init(struct il_priv *il, struct ieee80211_sta *sta,
2886 * ieee80211_register_hw 2886 * ieee80211_register_hw
2887 * 2887 *
2888 */ 2888 */
2889extern int il4965_rate_control_register(void); 2889int il4965_rate_control_register(void);
2890extern int il3945_rate_control_register(void); 2890int il3945_rate_control_register(void);
2891 2891
2892/** 2892/**
2893 * il_rate_control_unregister - Unregister the rate control callbacks 2893 * il_rate_control_unregister - Unregister the rate control callbacks
@@ -2895,11 +2895,11 @@ extern int il3945_rate_control_register(void);
2895 * This should be called after calling ieee80211_unregister_hw, but before 2895 * This should be called after calling ieee80211_unregister_hw, but before
2896 * the driver is unloaded. 2896 * the driver is unloaded.
2897 */ 2897 */
2898extern void il4965_rate_control_unregister(void); 2898void il4965_rate_control_unregister(void);
2899extern void il3945_rate_control_unregister(void); 2899void il3945_rate_control_unregister(void);
2900 2900
2901extern int il_power_update_mode(struct il_priv *il, bool force); 2901int il_power_update_mode(struct il_priv *il, bool force);
2902extern void il_power_initialize(struct il_priv *il); 2902void il_power_initialize(struct il_priv *il);
2903 2903
2904extern u32 il_debug_level; 2904extern u32 il_debug_level;
2905 2905
diff --git a/drivers/net/wireless/iwlwifi/dvm/agn.h b/drivers/net/wireless/iwlwifi/dvm/agn.h
index f2a86ffc3b4c..23d5f0275ce9 100644
--- a/drivers/net/wireless/iwlwifi/dvm/agn.h
+++ b/drivers/net/wireless/iwlwifi/dvm/agn.h
@@ -397,7 +397,7 @@ static inline __le32 iwl_hw_set_rate_n_flags(u8 rate, u32 flags)
397 return cpu_to_le32(flags|(u32)rate); 397 return cpu_to_le32(flags|(u32)rate);
398} 398}
399 399
400extern int iwl_alive_start(struct iwl_priv *priv); 400int iwl_alive_start(struct iwl_priv *priv);
401 401
402#ifdef CONFIG_IWLWIFI_DEBUG 402#ifdef CONFIG_IWLWIFI_DEBUG
403void iwl_print_rx_config_cmd(struct iwl_priv *priv, 403void iwl_print_rx_config_cmd(struct iwl_priv *priv,
diff --git a/drivers/net/wireless/iwlwifi/dvm/dev.h b/drivers/net/wireless/iwlwifi/dvm/dev.h
index a79fdd137f95..7434d9edf3b7 100644
--- a/drivers/net/wireless/iwlwifi/dvm/dev.h
+++ b/drivers/net/wireless/iwlwifi/dvm/dev.h
@@ -270,7 +270,7 @@ struct iwl_sensitivity_ranges {
270 * iwlXXXX_ <-- Hardware specific (implemented in iwl-XXXX.c for XXXX) 270 * iwlXXXX_ <-- Hardware specific (implemented in iwl-XXXX.c for XXXX)
271 * 271 *
272 ****************************************************************************/ 272 ****************************************************************************/
273extern void iwl_update_chain_flags(struct iwl_priv *priv); 273void iwl_update_chain_flags(struct iwl_priv *priv);
274extern const u8 iwl_bcast_addr[ETH_ALEN]; 274extern const u8 iwl_bcast_addr[ETH_ALEN];
275 275
276#define IWL_OPERATION_MODE_AUTO 0 276#define IWL_OPERATION_MODE_AUTO 0
diff --git a/drivers/net/wireless/iwlwifi/dvm/rs.h b/drivers/net/wireless/iwlwifi/dvm/rs.h
index 5d83cab22d62..26fc550cd68c 100644
--- a/drivers/net/wireless/iwlwifi/dvm/rs.h
+++ b/drivers/net/wireless/iwlwifi/dvm/rs.h
@@ -407,8 +407,8 @@ static inline u8 first_antenna(u8 mask)
407 407
408 408
409/* Initialize station's rate scaling information after adding station */ 409/* Initialize station's rate scaling information after adding station */
410extern void iwl_rs_rate_init(struct iwl_priv *priv, 410void iwl_rs_rate_init(struct iwl_priv *priv, struct ieee80211_sta *sta,
411 struct ieee80211_sta *sta, u8 sta_id); 411 u8 sta_id);
412 412
413/** 413/**
414 * iwl_rate_control_register - Register the rate control algorithm callbacks 414 * iwl_rate_control_register - Register the rate control algorithm callbacks
@@ -420,7 +420,7 @@ extern void iwl_rs_rate_init(struct iwl_priv *priv,
420 * ieee80211_register_hw 420 * ieee80211_register_hw
421 * 421 *
422 */ 422 */
423extern int iwlagn_rate_control_register(void); 423int iwlagn_rate_control_register(void);
424 424
425/** 425/**
426 * iwl_rate_control_unregister - Unregister the rate control callbacks 426 * iwl_rate_control_unregister - Unregister the rate control callbacks
@@ -428,6 +428,6 @@ extern int iwlagn_rate_control_register(void);
428 * This should be called after calling ieee80211_unregister_hw, but before 428 * This should be called after calling ieee80211_unregister_hw, but before
429 * the driver is unloaded. 429 * the driver is unloaded.
430 */ 430 */
431extern void iwlagn_rate_control_unregister(void); 431void iwlagn_rate_control_unregister(void);
432 432
433#endif /* __iwl_agn__rs__ */ 433#endif /* __iwl_agn__rs__ */
diff --git a/drivers/net/wireless/iwlwifi/dvm/tx.c b/drivers/net/wireless/iwlwifi/dvm/tx.c
index da442b81370a..1fef5240e6ad 100644
--- a/drivers/net/wireless/iwlwifi/dvm/tx.c
+++ b/drivers/net/wireless/iwlwifi/dvm/tx.c
@@ -433,27 +433,19 @@ int iwlagn_tx_skb(struct iwl_priv *priv,
433 /* Copy MAC header from skb into command buffer */ 433 /* Copy MAC header from skb into command buffer */
434 memcpy(tx_cmd->hdr, hdr, hdr_len); 434 memcpy(tx_cmd->hdr, hdr, hdr_len);
435 435
436 txq_id = info->hw_queue;
437
436 if (is_agg) 438 if (is_agg)
437 txq_id = priv->tid_data[sta_id][tid].agg.txq_id; 439 txq_id = priv->tid_data[sta_id][tid].agg.txq_id;
438 else if (info->flags & IEEE80211_TX_CTL_SEND_AFTER_DTIM) { 440 else if (info->flags & IEEE80211_TX_CTL_SEND_AFTER_DTIM) {
439 /* 441 /*
440 * Send this frame after DTIM -- there's a special queue
441 * reserved for this for contexts that support AP mode.
442 */
443 txq_id = ctx->mcast_queue;
444
445 /*
446 * The microcode will clear the more data 442 * The microcode will clear the more data
447 * bit in the last frame it transmits. 443 * bit in the last frame it transmits.
448 */ 444 */
449 hdr->frame_control |= 445 hdr->frame_control |=
450 cpu_to_le16(IEEE80211_FCTL_MOREDATA); 446 cpu_to_le16(IEEE80211_FCTL_MOREDATA);
451 } else if (info->flags & IEEE80211_TX_CTL_TX_OFFCHAN) 447 }
452 txq_id = IWL_AUX_QUEUE;
453 else
454 txq_id = ctx->ac_to_queue[skb_get_queue_mapping(skb)];
455 448
456 WARN_ON_ONCE(!is_agg && txq_id != info->hw_queue);
457 WARN_ON_ONCE(is_agg && 449 WARN_ON_ONCE(is_agg &&
458 priv->queue_to_mac80211[txq_id] != info->hw_queue); 450 priv->queue_to_mac80211[txq_id] != info->hw_queue);
459 451
diff --git a/drivers/net/wireless/iwlwifi/dvm/ucode.c b/drivers/net/wireless/iwlwifi/dvm/ucode.c
index 86270b69cd02..63637949a146 100644
--- a/drivers/net/wireless/iwlwifi/dvm/ucode.c
+++ b/drivers/net/wireless/iwlwifi/dvm/ucode.c
@@ -330,15 +330,14 @@ int iwl_load_ucode_wait_alive(struct iwl_priv *priv,
330 enum iwl_ucode_type old_type; 330 enum iwl_ucode_type old_type;
331 static const u8 alive_cmd[] = { REPLY_ALIVE }; 331 static const u8 alive_cmd[] = { REPLY_ALIVE };
332 332
333 old_type = priv->cur_ucode;
334 priv->cur_ucode = ucode_type;
335 fw = iwl_get_ucode_image(priv, ucode_type); 333 fw = iwl_get_ucode_image(priv, ucode_type);
334 if (WARN_ON(!fw))
335 return -EINVAL;
336 336
337 old_type = priv->cur_ucode;
338 priv->cur_ucode = ucode_type;
337 priv->ucode_loaded = false; 339 priv->ucode_loaded = false;
338 340
339 if (!fw)
340 return -EINVAL;
341
342 iwl_init_notification_wait(&priv->notif_wait, &alive_wait, 341 iwl_init_notification_wait(&priv->notif_wait, &alive_wait,
343 alive_cmd, ARRAY_SIZE(alive_cmd), 342 alive_cmd, ARRAY_SIZE(alive_cmd),
344 iwl_alive_fn, &alive_data); 343 iwl_alive_fn, &alive_data);
diff --git a/drivers/net/wireless/iwlwifi/iwl-7000.c b/drivers/net/wireless/iwlwifi/iwl-7000.c
index 76e14c046d94..85879dbaa402 100644
--- a/drivers/net/wireless/iwlwifi/iwl-7000.c
+++ b/drivers/net/wireless/iwlwifi/iwl-7000.c
@@ -83,6 +83,8 @@
83#define IWL7260_TX_POWER_VERSION 0xffff /* meaningless */ 83#define IWL7260_TX_POWER_VERSION 0xffff /* meaningless */
84#define IWL3160_NVM_VERSION 0x709 84#define IWL3160_NVM_VERSION 0x709
85#define IWL3160_TX_POWER_VERSION 0xffff /* meaningless */ 85#define IWL3160_TX_POWER_VERSION 0xffff /* meaningless */
86#define IWL7265_NVM_VERSION 0x0a1d
87#define IWL7265_TX_POWER_VERSION 0xffff /* meaningless */
86 88
87#define IWL7260_FW_PRE "iwlwifi-7260-" 89#define IWL7260_FW_PRE "iwlwifi-7260-"
88#define IWL7260_MODULE_FIRMWARE(api) IWL7260_FW_PRE __stringify(api) ".ucode" 90#define IWL7260_MODULE_FIRMWARE(api) IWL7260_FW_PRE __stringify(api) ".ucode"
@@ -90,6 +92,9 @@
90#define IWL3160_FW_PRE "iwlwifi-3160-" 92#define IWL3160_FW_PRE "iwlwifi-3160-"
91#define IWL3160_MODULE_FIRMWARE(api) IWL3160_FW_PRE __stringify(api) ".ucode" 93#define IWL3160_MODULE_FIRMWARE(api) IWL3160_FW_PRE __stringify(api) ".ucode"
92 94
95#define IWL7265_FW_PRE "iwlwifi-7265-"
96#define IWL7265_MODULE_FIRMWARE(api) IWL7265_FW_PRE __stringify(api) ".ucode"
97
93static const struct iwl_base_params iwl7000_base_params = { 98static const struct iwl_base_params iwl7000_base_params = {
94 .eeprom_size = OTP_LOW_IMAGE_SIZE, 99 .eeprom_size = OTP_LOW_IMAGE_SIZE,
95 .num_of_queues = IWLAGN_NUM_QUEUES, 100 .num_of_queues = IWLAGN_NUM_QUEUES,
@@ -182,5 +187,14 @@ const struct iwl_cfg iwl3160_n_cfg = {
182 .nvm_calib_ver = IWL3160_TX_POWER_VERSION, 187 .nvm_calib_ver = IWL3160_TX_POWER_VERSION,
183}; 188};
184 189
190const struct iwl_cfg iwl7265_2ac_cfg = {
191 .name = "Intel(R) Dual Band Wireless AC 7265",
192 .fw_name_pre = IWL7265_FW_PRE,
193 IWL_DEVICE_7000,
194 .ht_params = &iwl7000_ht_params,
195 .nvm_ver = IWL7265_NVM_VERSION,
196 .nvm_calib_ver = IWL7265_TX_POWER_VERSION,
197};
198
185MODULE_FIRMWARE(IWL7260_MODULE_FIRMWARE(IWL7260_UCODE_API_OK)); 199MODULE_FIRMWARE(IWL7260_MODULE_FIRMWARE(IWL7260_UCODE_API_OK));
186MODULE_FIRMWARE(IWL3160_MODULE_FIRMWARE(IWL3160_UCODE_API_OK)); 200MODULE_FIRMWARE(IWL3160_MODULE_FIRMWARE(IWL3160_UCODE_API_OK));
diff --git a/drivers/net/wireless/iwlwifi/iwl-config.h b/drivers/net/wireless/iwlwifi/iwl-config.h
index b03c25e14903..18f232e8e812 100644
--- a/drivers/net/wireless/iwlwifi/iwl-config.h
+++ b/drivers/net/wireless/iwlwifi/iwl-config.h
@@ -293,6 +293,7 @@ extern const struct iwl_cfg iwl7260_n_cfg;
293extern const struct iwl_cfg iwl3160_2ac_cfg; 293extern const struct iwl_cfg iwl3160_2ac_cfg;
294extern const struct iwl_cfg iwl3160_2n_cfg; 294extern const struct iwl_cfg iwl3160_2n_cfg;
295extern const struct iwl_cfg iwl3160_n_cfg; 295extern const struct iwl_cfg iwl3160_n_cfg;
296extern const struct iwl_cfg iwl7265_2ac_cfg;
296#endif /* CONFIG_IWLMVM */ 297#endif /* CONFIG_IWLMVM */
297 298
298#endif /* __IWL_CONFIG_H__ */ 299#endif /* __IWL_CONFIG_H__ */
diff --git a/drivers/net/wireless/iwlwifi/iwl-csr.h b/drivers/net/wireless/iwlwifi/iwl-csr.h
index a276af476e2d..54a4fdc631b7 100644
--- a/drivers/net/wireless/iwlwifi/iwl-csr.h
+++ b/drivers/net/wireless/iwlwifi/iwl-csr.h
@@ -394,6 +394,38 @@
394#define CSR_DRAM_INT_TBL_ENABLE (1 << 31) 394#define CSR_DRAM_INT_TBL_ENABLE (1 << 31)
395#define CSR_DRAM_INIT_TBL_WRAP_CHECK (1 << 27) 395#define CSR_DRAM_INIT_TBL_WRAP_CHECK (1 << 27)
396 396
397/* SECURE boot registers */
398#define CSR_SECURE_BOOT_CONFIG_ADDR (0x100)
399enum secure_boot_config_reg {
400 CSR_SECURE_BOOT_CONFIG_INSPECTOR_BURNED_IN_OTP = 0x00000001,
401 CSR_SECURE_BOOT_CONFIG_INSPECTOR_NOT_REQ = 0x00000002,
402};
403
404#define CSR_SECURE_BOOT_CPU1_STATUS_ADDR (0x100)
405#define CSR_SECURE_BOOT_CPU2_STATUS_ADDR (0x100)
406enum secure_boot_status_reg {
407 CSR_SECURE_BOOT_CPU_STATUS_VERF_STATUS = 0x00000003,
408 CSR_SECURE_BOOT_CPU_STATUS_VERF_COMPLETED = 0x00000002,
409 CSR_SECURE_BOOT_CPU_STATUS_VERF_SUCCESS = 0x00000004,
410 CSR_SECURE_BOOT_CPU_STATUS_VERF_FAIL = 0x00000008,
411 CSR_SECURE_BOOT_CPU_STATUS_SIGN_VERF_FAIL = 0x00000010,
412};
413
414#define CSR_UCODE_LOAD_STATUS_ADDR (0x100)
415enum secure_load_status_reg {
416 CSR_CPU_STATUS_LOADING_STARTED = 0x00000001,
417 CSR_CPU_STATUS_LOADING_COMPLETED = 0x00000002,
418 CSR_CPU_STATUS_NUM_OF_LAST_COMPLETED = 0x000000F8,
419 CSR_CPU_STATUS_NUM_OF_LAST_LOADED_BLOCK = 0x0000FF00,
420};
421
422#define CSR_SECURE_INSPECTOR_CODE_ADDR (0x100)
423#define CSR_SECURE_INSPECTOR_DATA_ADDR (0x100)
424
425#define CSR_SECURE_TIME_OUT (100)
426
427#define FH_TCSR_0_REG0 (0x1D00)
428
397/* 429/*
398 * HBUS (Host-side Bus) 430 * HBUS (Host-side Bus)
399 * 431 *
diff --git a/drivers/net/wireless/iwlwifi/iwl-drv.c b/drivers/net/wireless/iwlwifi/iwl-drv.c
index 99e1da3123c9..ff570027e9dd 100644
--- a/drivers/net/wireless/iwlwifi/iwl-drv.c
+++ b/drivers/net/wireless/iwlwifi/iwl-drv.c
@@ -483,6 +483,7 @@ static int iwl_parse_tlv_firmware(struct iwl_drv *drv,
483 const u8 *tlv_data; 483 const u8 *tlv_data;
484 char buildstr[25]; 484 char buildstr[25];
485 u32 build; 485 u32 build;
486 int num_of_cpus;
486 487
487 if (len < sizeof(*ucode)) { 488 if (len < sizeof(*ucode)) {
488 IWL_ERR(drv, "uCode has invalid length: %zd\n", len); 489 IWL_ERR(drv, "uCode has invalid length: %zd\n", len);
@@ -692,6 +693,42 @@ static int iwl_parse_tlv_firmware(struct iwl_drv *drv,
692 goto invalid_tlv_len; 693 goto invalid_tlv_len;
693 drv->fw.phy_config = le32_to_cpup((__le32 *)tlv_data); 694 drv->fw.phy_config = le32_to_cpup((__le32 *)tlv_data);
694 break; 695 break;
696 case IWL_UCODE_TLV_SECURE_SEC_RT:
697 iwl_store_ucode_sec(pieces, tlv_data, IWL_UCODE_REGULAR,
698 tlv_len);
699 drv->fw.mvm_fw = true;
700 drv->fw.img[IWL_UCODE_REGULAR].is_secure = true;
701 break;
702 case IWL_UCODE_TLV_SECURE_SEC_INIT:
703 iwl_store_ucode_sec(pieces, tlv_data, IWL_UCODE_INIT,
704 tlv_len);
705 drv->fw.mvm_fw = true;
706 drv->fw.img[IWL_UCODE_INIT].is_secure = true;
707 break;
708 case IWL_UCODE_TLV_SECURE_SEC_WOWLAN:
709 iwl_store_ucode_sec(pieces, tlv_data, IWL_UCODE_WOWLAN,
710 tlv_len);
711 drv->fw.mvm_fw = true;
712 drv->fw.img[IWL_UCODE_WOWLAN].is_secure = true;
713 break;
714 case IWL_UCODE_TLV_NUM_OF_CPU:
715 if (tlv_len != sizeof(u32))
716 goto invalid_tlv_len;
717 num_of_cpus =
718 le32_to_cpup((__le32 *)tlv_data);
719
720 if (num_of_cpus == 2) {
721 drv->fw.img[IWL_UCODE_REGULAR].is_dual_cpus =
722 true;
723 drv->fw.img[IWL_UCODE_INIT].is_dual_cpus =
724 true;
725 drv->fw.img[IWL_UCODE_WOWLAN].is_dual_cpus =
726 true;
727 } else if ((num_of_cpus > 2) || (num_of_cpus < 1)) {
728 IWL_ERR(drv, "Driver support upto 2 CPUs\n");
729 return -EINVAL;
730 }
731 break;
695 default: 732 default:
696 IWL_DEBUG_INFO(drv, "unknown TLV: %d\n", tlv_type); 733 IWL_DEBUG_INFO(drv, "unknown TLV: %d\n", tlv_type);
697 break; 734 break;
diff --git a/drivers/net/wireless/iwlwifi/iwl-fw-file.h b/drivers/net/wireless/iwlwifi/iwl-fw-file.h
index 8b6c6fd95ed0..6c6c35c5228c 100644
--- a/drivers/net/wireless/iwlwifi/iwl-fw-file.h
+++ b/drivers/net/wireless/iwlwifi/iwl-fw-file.h
@@ -121,6 +121,10 @@ enum iwl_ucode_tlv_type {
121 IWL_UCODE_TLV_SEC_WOWLAN = 21, 121 IWL_UCODE_TLV_SEC_WOWLAN = 21,
122 IWL_UCODE_TLV_DEF_CALIB = 22, 122 IWL_UCODE_TLV_DEF_CALIB = 22,
123 IWL_UCODE_TLV_PHY_SKU = 23, 123 IWL_UCODE_TLV_PHY_SKU = 23,
124 IWL_UCODE_TLV_SECURE_SEC_RT = 24,
125 IWL_UCODE_TLV_SECURE_SEC_INIT = 25,
126 IWL_UCODE_TLV_SECURE_SEC_WOWLAN = 26,
127 IWL_UCODE_TLV_NUM_OF_CPU = 27,
124}; 128};
125 129
126struct iwl_ucode_tlv { 130struct iwl_ucode_tlv {
diff --git a/drivers/net/wireless/iwlwifi/iwl-fw.h b/drivers/net/wireless/iwlwifi/iwl-fw.h
index a1223680bc70..75db087120c3 100644
--- a/drivers/net/wireless/iwlwifi/iwl-fw.h
+++ b/drivers/net/wireless/iwlwifi/iwl-fw.h
@@ -75,11 +75,23 @@
75 * @IWL_UCODE_TLV_FLAGS_P2P: This uCode image supports P2P. 75 * @IWL_UCODE_TLV_FLAGS_P2P: This uCode image supports P2P.
76 * @IWL_UCODE_TLV_FLAGS_DW_BC_TABLE: The SCD byte count table is in DWORDS 76 * @IWL_UCODE_TLV_FLAGS_DW_BC_TABLE: The SCD byte count table is in DWORDS
77 * @IWL_UCODE_TLV_FLAGS_UAPSD: This uCode image supports uAPSD 77 * @IWL_UCODE_TLV_FLAGS_UAPSD: This uCode image supports uAPSD
78 * @IWL_UCODE_TLV_FLAGS_SHORT_BL: 16 entries of black list instead of 64 in scan
79 * offload profile config command.
78 * @IWL_UCODE_TLV_FLAGS_RX_ENERGY_API: supports rx signal strength api 80 * @IWL_UCODE_TLV_FLAGS_RX_ENERGY_API: supports rx signal strength api
79 * @IWL_UCODE_TLV_FLAGS_TIME_EVENT_API_V2: using the new time event API. 81 * @IWL_UCODE_TLV_FLAGS_TIME_EVENT_API_V2: using the new time event API.
80 * @IWL_UCODE_TLV_FLAGS_D3_6_IPV6_ADDRS: D3 image supports up to six 82 * @IWL_UCODE_TLV_FLAGS_D3_6_IPV6_ADDRS: D3 image supports up to six
81 * (rather than two) IPv6 addresses 83 * (rather than two) IPv6 addresses
82 * @IWL_UCODE_TLV_FLAGS_BF_UPDATED: new beacon filtering API 84 * @IWL_UCODE_TLV_FLAGS_BF_UPDATED: new beacon filtering API
85 * @IWL_UCODE_TLV_FLAGS_NO_BASIC_SSID: not sending a probe with the SSID element
86 * from the probe request template.
87 * @IWL_UCODE_TLV_FLAGS_D3_CONTINUITY_API: modified D3 API to allow keeping
88 * connection when going back to D0
89 * @IWL_UCODE_TLV_FLAGS_NEW_NSOFFL_SMALL: new NS offload (small version)
90 * @IWL_UCODE_TLV_FLAGS_NEW_NSOFFL_LARGE: new NS offload (large version)
91 * @IWL_UCODE_TLV_FLAGS_SCHED_SCAN: this uCode image supports scheduled scan.
92 * @IWL_UCODE_TLV_FLAGS_STA_KEY_CMD: new ADD_STA and ADD_STA_KEY command API
93 * @IWL_UCODE_TLV_FLAGS_DEVICE_PS_CMD: support device wide power command
94 * containing CAM (Continuous Active Mode) indication.
83 */ 95 */
84enum iwl_ucode_tlv_flag { 96enum iwl_ucode_tlv_flag {
85 IWL_UCODE_TLV_FLAGS_PAN = BIT(0), 97 IWL_UCODE_TLV_FLAGS_PAN = BIT(0),
@@ -87,11 +99,21 @@ enum iwl_ucode_tlv_flag {
87 IWL_UCODE_TLV_FLAGS_MFP = BIT(2), 99 IWL_UCODE_TLV_FLAGS_MFP = BIT(2),
88 IWL_UCODE_TLV_FLAGS_P2P = BIT(3), 100 IWL_UCODE_TLV_FLAGS_P2P = BIT(3),
89 IWL_UCODE_TLV_FLAGS_DW_BC_TABLE = BIT(4), 101 IWL_UCODE_TLV_FLAGS_DW_BC_TABLE = BIT(4),
90 IWL_UCODE_TLV_FLAGS_UAPSD = BIT(6), 102 IWL_UCODE_TLV_FLAGS_NEWBT_COEX = BIT(5),
103 IWL_UCODE_TLV_FLAGS_PM_CMD_SUPPORT = BIT(6),
104 IWL_UCODE_TLV_FLAGS_SHORT_BL = BIT(7),
91 IWL_UCODE_TLV_FLAGS_RX_ENERGY_API = BIT(8), 105 IWL_UCODE_TLV_FLAGS_RX_ENERGY_API = BIT(8),
92 IWL_UCODE_TLV_FLAGS_TIME_EVENT_API_V2 = BIT(9), 106 IWL_UCODE_TLV_FLAGS_TIME_EVENT_API_V2 = BIT(9),
93 IWL_UCODE_TLV_FLAGS_D3_6_IPV6_ADDRS = BIT(10), 107 IWL_UCODE_TLV_FLAGS_D3_6_IPV6_ADDRS = BIT(10),
94 IWL_UCODE_TLV_FLAGS_BF_UPDATED = BIT(11), 108 IWL_UCODE_TLV_FLAGS_BF_UPDATED = BIT(11),
109 IWL_UCODE_TLV_FLAGS_NO_BASIC_SSID = BIT(12),
110 IWL_UCODE_TLV_FLAGS_D3_CONTINUITY_API = BIT(14),
111 IWL_UCODE_TLV_FLAGS_NEW_NSOFFL_SMALL = BIT(15),
112 IWL_UCODE_TLV_FLAGS_NEW_NSOFFL_LARGE = BIT(16),
113 IWL_UCODE_TLV_FLAGS_SCHED_SCAN = BIT(17),
114 IWL_UCODE_TLV_FLAGS_STA_KEY_CMD = BIT(19),
115 IWL_UCODE_TLV_FLAGS_DEVICE_PS_CMD = BIT(20),
116 IWL_UCODE_TLV_FLAGS_UAPSD_SUPPORT = BIT(24),
95}; 117};
96 118
97/* The default calibrate table size if not specified by firmware file */ 119/* The default calibrate table size if not specified by firmware file */
@@ -133,7 +155,8 @@ enum iwl_ucode_sec {
133 * For 16.0 uCode and above, there is no differentiation between sections, 155 * For 16.0 uCode and above, there is no differentiation between sections,
134 * just an offset to the HW address. 156 * just an offset to the HW address.
135 */ 157 */
136#define IWL_UCODE_SECTION_MAX 4 158#define IWL_UCODE_SECTION_MAX 6
159#define IWL_UCODE_FIRST_SECTION_OF_SECOND_CPU (IWL_UCODE_SECTION_MAX/2)
137 160
138struct iwl_ucode_capabilities { 161struct iwl_ucode_capabilities {
139 u32 max_probe_length; 162 u32 max_probe_length;
@@ -150,6 +173,8 @@ struct fw_desc {
150 173
151struct fw_img { 174struct fw_img {
152 struct fw_desc sec[IWL_UCODE_SECTION_MAX]; 175 struct fw_desc sec[IWL_UCODE_SECTION_MAX];
176 bool is_secure;
177 bool is_dual_cpus;
153}; 178};
154 179
155/* uCode version contains 4 values: Major/Minor/API/Serial */ 180/* uCode version contains 4 values: Major/Minor/API/Serial */
diff --git a/drivers/net/wireless/iwlwifi/iwl-io.c b/drivers/net/wireless/iwlwifi/iwl-io.c
index dfa4d2e3aaa2..ad8e19a56eca 100644
--- a/drivers/net/wireless/iwlwifi/iwl-io.c
+++ b/drivers/net/wireless/iwlwifi/iwl-io.c
@@ -34,7 +34,6 @@
34#include "iwl-csr.h" 34#include "iwl-csr.h"
35#include "iwl-debug.h" 35#include "iwl-debug.h"
36#include "iwl-fh.h" 36#include "iwl-fh.h"
37#include "iwl-csr.h"
38 37
39#define IWL_POLL_INTERVAL 10 /* microseconds */ 38#define IWL_POLL_INTERVAL 10 /* microseconds */
40 39
diff --git a/drivers/net/wireless/iwlwifi/iwl-prph.h b/drivers/net/wireless/iwlwifi/iwl-prph.h
index ff8cc75c189d..a70c7b9d9bad 100644
--- a/drivers/net/wireless/iwlwifi/iwl-prph.h
+++ b/drivers/net/wireless/iwlwifi/iwl-prph.h
@@ -97,6 +97,8 @@
97 97
98#define APMG_PCIDEV_STT_VAL_L1_ACT_DIS (0x00000800) 98#define APMG_PCIDEV_STT_VAL_L1_ACT_DIS (0x00000800)
99 99
100#define APMG_RTC_INT_STT_RFKILL (0x10000000)
101
100/* Device system time */ 102/* Device system time */
101#define DEVICE_SYSTEM_TIME_REG 0xA0206C 103#define DEVICE_SYSTEM_TIME_REG 0xA0206C
102 104
diff --git a/drivers/net/wireless/iwlwifi/iwl-trans.h b/drivers/net/wireless/iwlwifi/iwl-trans.h
index 80b47508647c..143292b4dbbf 100644
--- a/drivers/net/wireless/iwlwifi/iwl-trans.h
+++ b/drivers/net/wireless/iwlwifi/iwl-trans.h
@@ -344,7 +344,7 @@ struct iwl_trans_config {
344 u8 cmd_queue; 344 u8 cmd_queue;
345 u8 cmd_fifo; 345 u8 cmd_fifo;
346 const u8 *no_reclaim_cmds; 346 const u8 *no_reclaim_cmds;
347 int n_no_reclaim_cmds; 347 unsigned int n_no_reclaim_cmds;
348 348
349 bool rx_buf_size_8k; 349 bool rx_buf_size_8k;
350 bool bc_table_dword; 350 bool bc_table_dword;
@@ -601,7 +601,7 @@ static inline int iwl_trans_send_cmd(struct iwl_trans *trans,
601{ 601{
602 int ret; 602 int ret;
603 603
604 if (trans->state != IWL_TRANS_FW_ALIVE) { 604 if (unlikely(trans->state != IWL_TRANS_FW_ALIVE)) {
605 IWL_ERR(trans, "%s bad state = %d", __func__, trans->state); 605 IWL_ERR(trans, "%s bad state = %d", __func__, trans->state);
606 return -EIO; 606 return -EIO;
607 } 607 }
@@ -640,8 +640,8 @@ static inline void iwl_trans_free_tx_cmd(struct iwl_trans *trans,
640static inline int iwl_trans_tx(struct iwl_trans *trans, struct sk_buff *skb, 640static inline int iwl_trans_tx(struct iwl_trans *trans, struct sk_buff *skb,
641 struct iwl_device_cmd *dev_cmd, int queue) 641 struct iwl_device_cmd *dev_cmd, int queue)
642{ 642{
643 WARN_ONCE(trans->state != IWL_TRANS_FW_ALIVE, 643 if (unlikely(trans->state != IWL_TRANS_FW_ALIVE))
644 "%s bad state = %d", __func__, trans->state); 644 IWL_ERR(trans, "%s bad state = %d", __func__, trans->state);
645 645
646 return trans->ops->tx(trans, skb, dev_cmd, queue); 646 return trans->ops->tx(trans, skb, dev_cmd, queue);
647} 647}
@@ -649,16 +649,16 @@ static inline int iwl_trans_tx(struct iwl_trans *trans, struct sk_buff *skb,
649static inline void iwl_trans_reclaim(struct iwl_trans *trans, int queue, 649static inline void iwl_trans_reclaim(struct iwl_trans *trans, int queue,
650 int ssn, struct sk_buff_head *skbs) 650 int ssn, struct sk_buff_head *skbs)
651{ 651{
652 WARN_ONCE(trans->state != IWL_TRANS_FW_ALIVE, 652 if (unlikely(trans->state != IWL_TRANS_FW_ALIVE))
653 "%s bad state = %d", __func__, trans->state); 653 IWL_ERR(trans, "%s bad state = %d", __func__, trans->state);
654 654
655 trans->ops->reclaim(trans, queue, ssn, skbs); 655 trans->ops->reclaim(trans, queue, ssn, skbs);
656} 656}
657 657
658static inline void iwl_trans_txq_disable(struct iwl_trans *trans, int queue) 658static inline void iwl_trans_txq_disable(struct iwl_trans *trans, int queue)
659{ 659{
660 WARN_ONCE(trans->state != IWL_TRANS_FW_ALIVE, 660 if (unlikely(trans->state != IWL_TRANS_FW_ALIVE))
661 "%s bad state = %d", __func__, trans->state); 661 IWL_ERR(trans, "%s bad state = %d", __func__, trans->state);
662 662
663 trans->ops->txq_disable(trans, queue); 663 trans->ops->txq_disable(trans, queue);
664} 664}
@@ -669,8 +669,8 @@ static inline void iwl_trans_txq_enable(struct iwl_trans *trans, int queue,
669{ 669{
670 might_sleep(); 670 might_sleep();
671 671
672 WARN_ONCE(trans->state != IWL_TRANS_FW_ALIVE, 672 if (unlikely((trans->state != IWL_TRANS_FW_ALIVE)))
673 "%s bad state = %d", __func__, trans->state); 673 IWL_ERR(trans, "%s bad state = %d", __func__, trans->state);
674 674
675 trans->ops->txq_enable(trans, queue, fifo, sta_id, tid, 675 trans->ops->txq_enable(trans, queue, fifo, sta_id, tid,
676 frame_limit, ssn); 676 frame_limit, ssn);
@@ -685,8 +685,8 @@ static inline void iwl_trans_ac_txq_enable(struct iwl_trans *trans, int queue,
685 685
686static inline int iwl_trans_wait_tx_queue_empty(struct iwl_trans *trans) 686static inline int iwl_trans_wait_tx_queue_empty(struct iwl_trans *trans)
687{ 687{
688 WARN_ONCE(trans->state != IWL_TRANS_FW_ALIVE, 688 if (unlikely(trans->state != IWL_TRANS_FW_ALIVE))
689 "%s bad state = %d", __func__, trans->state); 689 IWL_ERR(trans, "%s bad state = %d", __func__, trans->state);
690 690
691 return trans->ops->wait_tx_queue_empty(trans); 691 return trans->ops->wait_tx_queue_empty(trans);
692} 692}
diff --git a/drivers/net/wireless/iwlwifi/mvm/bt-coex.c b/drivers/net/wireless/iwlwifi/mvm/bt-coex.c
index 0fad98b85f60..5d066cbc5ac7 100644
--- a/drivers/net/wireless/iwlwifi/mvm/bt-coex.c
+++ b/drivers/net/wireless/iwlwifi/mvm/bt-coex.c
@@ -98,126 +98,258 @@ static const u8 iwl_bt_prio_tbl[BT_COEX_PRIO_TBL_EVT_MAX] = {
98 98
99#undef EVENT_PRIO_ANT 99#undef EVENT_PRIO_ANT
100 100
101/* BT Antenna Coupling Threshold (dB) */
102#define IWL_BT_ANTENNA_COUPLING_THRESHOLD (35)
103#define IWL_BT_LOAD_FORCE_SISO_THRESHOLD (3)
104
105#define BT_ENABLE_REDUCED_TXPOWER_THRESHOLD (-62) 101#define BT_ENABLE_REDUCED_TXPOWER_THRESHOLD (-62)
106#define BT_DISABLE_REDUCED_TXPOWER_THRESHOLD (-65) 102#define BT_DISABLE_REDUCED_TXPOWER_THRESHOLD (-65)
107#define BT_REDUCED_TX_POWER_BIT BIT(7) 103#define BT_ANTENNA_COUPLING_THRESHOLD (30)
108
109static inline bool is_loose_coex(void)
110{
111 return iwlwifi_mod_params.ant_coupling >
112 IWL_BT_ANTENNA_COUPLING_THRESHOLD;
113}
114 104
115int iwl_send_bt_prio_tbl(struct iwl_mvm *mvm) 105int iwl_send_bt_prio_tbl(struct iwl_mvm *mvm)
116{ 106{
107 if (!(mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_NEWBT_COEX))
108 return 0;
109
117 return iwl_mvm_send_cmd_pdu(mvm, BT_COEX_PRIO_TABLE, CMD_SYNC, 110 return iwl_mvm_send_cmd_pdu(mvm, BT_COEX_PRIO_TABLE, CMD_SYNC,
118 sizeof(struct iwl_bt_coex_prio_tbl_cmd), 111 sizeof(struct iwl_bt_coex_prio_tbl_cmd),
119 &iwl_bt_prio_tbl); 112 &iwl_bt_prio_tbl);
120} 113}
121 114
122static int iwl_send_bt_env(struct iwl_mvm *mvm, u8 action, u8 type) 115const u32 iwl_bt_ack_kill_msk[BT_KILL_MSK_MAX] = {
123{
124 struct iwl_bt_coex_prot_env_cmd env_cmd;
125 int ret;
126
127 env_cmd.action = action;
128 env_cmd.type = type;
129 ret = iwl_mvm_send_cmd_pdu(mvm, BT_COEX_PROT_ENV, CMD_SYNC,
130 sizeof(env_cmd), &env_cmd);
131 if (ret)
132 IWL_ERR(mvm, "failed to send BT env command\n");
133 return ret;
134}
135
136enum iwl_bt_kill_msk {
137 BT_KILL_MSK_DEFAULT,
138 BT_KILL_MSK_SCO_HID_A2DP,
139 BT_KILL_MSK_REDUCED_TXPOW,
140 BT_KILL_MSK_MAX,
141};
142
143static const u32 iwl_bt_ack_kill_msk[BT_KILL_MSK_MAX] = {
144 [BT_KILL_MSK_DEFAULT] = 0xffff0000, 116 [BT_KILL_MSK_DEFAULT] = 0xffff0000,
145 [BT_KILL_MSK_SCO_HID_A2DP] = 0xffffffff, 117 [BT_KILL_MSK_SCO_HID_A2DP] = 0xffffffff,
146 [BT_KILL_MSK_REDUCED_TXPOW] = 0, 118 [BT_KILL_MSK_REDUCED_TXPOW] = 0,
147}; 119};
148 120
149static const u32 iwl_bt_cts_kill_msk[BT_KILL_MSK_MAX] = { 121const u32 iwl_bt_cts_kill_msk[BT_KILL_MSK_MAX] = {
150 [BT_KILL_MSK_DEFAULT] = 0xffff0000, 122 [BT_KILL_MSK_DEFAULT] = 0xffff0000,
151 [BT_KILL_MSK_SCO_HID_A2DP] = 0xffffffff, 123 [BT_KILL_MSK_SCO_HID_A2DP] = 0xffffffff,
152 [BT_KILL_MSK_REDUCED_TXPOW] = 0, 124 [BT_KILL_MSK_REDUCED_TXPOW] = 0,
153}; 125};
154 126
155#define IWL_BT_DEFAULT_BOOST (0xf0f0f0f0) 127static const __le32 iwl_bt_prio_boost[BT_COEX_BOOST_SIZE] = {
156 128 cpu_to_le32(0xf0f0f0f0),
157/* Tight Coex */ 129 cpu_to_le32(0xc0c0c0c0),
158static const __le32 iwl_tight_lookup[BT_COEX_LUT_SIZE] = { 130 cpu_to_le32(0xfcfcfcfc),
159 cpu_to_le32(0xaaaaaaaa), 131 cpu_to_le32(0xff00ff00),
160 cpu_to_le32(0xaaaaaaaa),
161 cpu_to_le32(0xaeaaaaaa),
162 cpu_to_le32(0xaaaaaaaa),
163 cpu_to_le32(0xcc00ff28),
164 cpu_to_le32(0x0000aaaa),
165 cpu_to_le32(0xcc00aaaa),
166 cpu_to_le32(0x0000aaaa),
167 cpu_to_le32(0xc0004000),
168 cpu_to_le32(0x00000000),
169 cpu_to_le32(0xf0005000),
170 cpu_to_le32(0xf0005000),
171}; 132};
172 133
173/* Loose Coex */ 134static const __le32 iwl_single_shared_ant[BT_COEX_MAX_LUT][BT_COEX_LUT_SIZE] = {
174static const __le32 iwl_loose_lookup[BT_COEX_LUT_SIZE] = { 135 {
175 cpu_to_le32(0xaaaaaaaa), 136 cpu_to_le32(0x40000000),
176 cpu_to_le32(0xaaaaaaaa), 137 cpu_to_le32(0x00000000),
177 cpu_to_le32(0xaaaaaaaa), 138 cpu_to_le32(0x44000000),
178 cpu_to_le32(0xaaaaaaaa), 139 cpu_to_le32(0x00000000),
179 cpu_to_le32(0xcc00ff28), 140 cpu_to_le32(0x40000000),
180 cpu_to_le32(0x0000aaaa), 141 cpu_to_le32(0x00000000),
181 cpu_to_le32(0xcc00aaaa), 142 cpu_to_le32(0x44000000),
182 cpu_to_le32(0x0000aaaa), 143 cpu_to_le32(0x00000000),
183 cpu_to_le32(0x00000000), 144 cpu_to_le32(0xc0004000),
184 cpu_to_le32(0x00000000), 145 cpu_to_le32(0xf0005000),
185 cpu_to_le32(0xf0005000), 146 cpu_to_le32(0xc0004000),
186 cpu_to_le32(0xf0005000), 147 cpu_to_le32(0xf0005000),
148 },
149 {
150 cpu_to_le32(0x40000000),
151 cpu_to_le32(0x00000000),
152 cpu_to_le32(0x44000000),
153 cpu_to_le32(0x00000000),
154 cpu_to_le32(0x40000000),
155 cpu_to_le32(0x00000000),
156 cpu_to_le32(0x44000000),
157 cpu_to_le32(0x00000000),
158 cpu_to_le32(0xc0004000),
159 cpu_to_le32(0xf0005000),
160 cpu_to_le32(0xc0004000),
161 cpu_to_le32(0xf0005000),
162 },
163 {
164 cpu_to_le32(0x40000000),
165 cpu_to_le32(0x00000000),
166 cpu_to_le32(0x44000000),
167 cpu_to_le32(0x00000000),
168 cpu_to_le32(0x40000000),
169 cpu_to_le32(0x00000000),
170 cpu_to_le32(0x44000000),
171 cpu_to_le32(0x00000000),
172 cpu_to_le32(0xc0004000),
173 cpu_to_le32(0xf0005000),
174 cpu_to_le32(0xc0004000),
175 cpu_to_le32(0xf0005000),
176 },
187}; 177};
188 178
189/* Full concurrency */ 179static const __le32 iwl_combined_lookup[BT_COEX_MAX_LUT][BT_COEX_LUT_SIZE] = {
190static const __le32 iwl_concurrent_lookup[BT_COEX_LUT_SIZE] = { 180 {
191 cpu_to_le32(0xaaaaaaaa), 181 /* Tight */
192 cpu_to_le32(0xaaaaaaaa), 182 cpu_to_le32(0xaaaaaaaa),
193 cpu_to_le32(0xaaaaaaaa), 183 cpu_to_le32(0xaaaaaaaa),
194 cpu_to_le32(0xaaaaaaaa), 184 cpu_to_le32(0xaeaaaaaa),
195 cpu_to_le32(0xaaaaaaaa), 185 cpu_to_le32(0xaaaaaaaa),
196 cpu_to_le32(0xaaaaaaaa), 186 cpu_to_le32(0xcc00ff28),
197 cpu_to_le32(0xaaaaaaaa), 187 cpu_to_le32(0x0000aaaa),
198 cpu_to_le32(0xaaaaaaaa), 188 cpu_to_le32(0xcc00aaaa),
199 cpu_to_le32(0x00000000), 189 cpu_to_le32(0x0000aaaa),
200 cpu_to_le32(0x00000000), 190 cpu_to_le32(0xc0004000),
201 cpu_to_le32(0x00000000), 191 cpu_to_le32(0x00000000),
202 cpu_to_le32(0x00000000), 192 cpu_to_le32(0xf0005000),
193 cpu_to_le32(0xf0005000),
194 },
195 {
196 /* Loose */
197 cpu_to_le32(0xaaaaaaaa),
198 cpu_to_le32(0xaaaaaaaa),
199 cpu_to_le32(0xaaaaaaaa),
200 cpu_to_le32(0xaaaaaaaa),
201 cpu_to_le32(0xcc00ff28),
202 cpu_to_le32(0x0000aaaa),
203 cpu_to_le32(0xcc00aaaa),
204 cpu_to_le32(0x0000aaaa),
205 cpu_to_le32(0x00000000),
206 cpu_to_le32(0x00000000),
207 cpu_to_le32(0xf0005000),
208 cpu_to_le32(0xf0005000),
209 },
210 {
211 /* Tx Tx disabled */
212 cpu_to_le32(0xaaaaaaaa),
213 cpu_to_le32(0xaaaaaaaa),
214 cpu_to_le32(0xaaaaaaaa),
215 cpu_to_le32(0xaaaaaaaa),
216 cpu_to_le32(0xcc00ff28),
217 cpu_to_le32(0x0000aaaa),
218 cpu_to_le32(0xcc00aaaa),
219 cpu_to_le32(0x0000aaaa),
220 cpu_to_le32(0xC0004000),
221 cpu_to_le32(0xC0004000),
222 cpu_to_le32(0xF0005000),
223 cpu_to_le32(0xF0005000),
224 },
203}; 225};
204 226
205/* single shared antenna */ 227/* 20MHz / 40MHz below / 40Mhz above*/
206static const __le32 iwl_single_shared_ant_lookup[BT_COEX_LUT_SIZE] = { 228static const __le64 iwl_ci_mask[][3] = {
207 cpu_to_le32(0x40000000), 229 /* dummy entry for channel 0 */
208 cpu_to_le32(0x00000000), 230 {cpu_to_le64(0), cpu_to_le64(0), cpu_to_le64(0)},
209 cpu_to_le32(0x44000000), 231 {
210 cpu_to_le32(0x00000000), 232 cpu_to_le64(0x0000001FFFULL),
211 cpu_to_le32(0x40000000), 233 cpu_to_le64(0x0ULL),
212 cpu_to_le32(0x00000000), 234 cpu_to_le64(0x00007FFFFFULL),
213 cpu_to_le32(0x44000000), 235 },
214 cpu_to_le32(0x00000000), 236 {
215 cpu_to_le32(0xC0004000), 237 cpu_to_le64(0x000000FFFFULL),
216 cpu_to_le32(0xF0005000), 238 cpu_to_le64(0x0ULL),
217 cpu_to_le32(0xC0004000), 239 cpu_to_le64(0x0003FFFFFFULL),
218 cpu_to_le32(0xF0005000), 240 },
241 {
242 cpu_to_le64(0x000003FFFCULL),
243 cpu_to_le64(0x0ULL),
244 cpu_to_le64(0x000FFFFFFCULL),
245 },
246 {
247 cpu_to_le64(0x00001FFFE0ULL),
248 cpu_to_le64(0x0ULL),
249 cpu_to_le64(0x007FFFFFE0ULL),
250 },
251 {
252 cpu_to_le64(0x00007FFF80ULL),
253 cpu_to_le64(0x00007FFFFFULL),
254 cpu_to_le64(0x01FFFFFF80ULL),
255 },
256 {
257 cpu_to_le64(0x0003FFFC00ULL),
258 cpu_to_le64(0x0003FFFFFFULL),
259 cpu_to_le64(0x0FFFFFFC00ULL),
260 },
261 {
262 cpu_to_le64(0x000FFFF000ULL),
263 cpu_to_le64(0x000FFFFFFCULL),
264 cpu_to_le64(0x3FFFFFF000ULL),
265 },
266 {
267 cpu_to_le64(0x007FFF8000ULL),
268 cpu_to_le64(0x007FFFFFE0ULL),
269 cpu_to_le64(0xFFFFFF8000ULL),
270 },
271 {
272 cpu_to_le64(0x01FFFE0000ULL),
273 cpu_to_le64(0x01FFFFFF80ULL),
274 cpu_to_le64(0xFFFFFE0000ULL),
275 },
276 {
277 cpu_to_le64(0x0FFFF00000ULL),
278 cpu_to_le64(0x0FFFFFFC00ULL),
279 cpu_to_le64(0x0ULL),
280 },
281 {
282 cpu_to_le64(0x3FFFC00000ULL),
283 cpu_to_le64(0x3FFFFFF000ULL),
284 cpu_to_le64(0x0)
285 },
286 {
287 cpu_to_le64(0xFFFE000000ULL),
288 cpu_to_le64(0xFFFFFF8000ULL),
289 cpu_to_le64(0x0)
290 },
291 {
292 cpu_to_le64(0xFFF8000000ULL),
293 cpu_to_le64(0xFFFFFE0000ULL),
294 cpu_to_le64(0x0)
295 },
296 {
297 cpu_to_le64(0xFE00000000ULL),
298 cpu_to_le64(0x0ULL),
299 cpu_to_le64(0x0)
300 },
219}; 301};
220 302
303static const __le32 iwl_bt_mprio_lut[BT_COEX_MULTI_PRIO_LUT_SIZE] = {
304 cpu_to_le32(0x22002200),
305 cpu_to_le32(0x33113311),
306};
307
308static enum iwl_bt_coex_lut_type
309iwl_get_coex_type(struct iwl_mvm *mvm, const struct ieee80211_vif *vif)
310{
311 struct ieee80211_chanctx_conf *chanctx_conf;
312 enum iwl_bt_coex_lut_type ret;
313 u16 phy_ctx_id;
314
315 /*
316 * Checking that we hold mvm->mutex is a good idea, but the rate
317 * control can't acquire the mutex since it runs in Tx path.
318 * So this is racy in that case, but in the worst case, the AMPDU
319 * size limit will be wrong for a short time which is not a big
320 * issue.
321 */
322
323 rcu_read_lock();
324
325 chanctx_conf = rcu_dereference(vif->chanctx_conf);
326
327 if (!chanctx_conf ||
328 chanctx_conf->def.chan->band != IEEE80211_BAND_2GHZ) {
329 rcu_read_unlock();
330 return BT_COEX_LOOSE_LUT;
331 }
332
333 ret = BT_COEX_TX_DIS_LUT;
334
335 if (mvm->cfg->bt_shared_single_ant) {
336 rcu_read_unlock();
337 return ret;
338 }
339
340 phy_ctx_id = *((u16 *)chanctx_conf->drv_priv);
341
342 if (mvm->last_bt_ci_cmd.primary_ch_phy_id == phy_ctx_id)
343 ret = le32_to_cpu(mvm->last_bt_notif.primary_ch_lut);
344 else if (mvm->last_bt_ci_cmd.secondary_ch_phy_id == phy_ctx_id)
345 ret = le32_to_cpu(mvm->last_bt_notif.secondary_ch_lut);
346 /* else - default = TX TX disallowed */
347
348 rcu_read_unlock();
349
350 return ret;
351}
352
221int iwl_send_bt_init_conf(struct iwl_mvm *mvm) 353int iwl_send_bt_init_conf(struct iwl_mvm *mvm)
222{ 354{
223 struct iwl_bt_coex_cmd *bt_cmd; 355 struct iwl_bt_coex_cmd *bt_cmd;
@@ -228,17 +360,10 @@ int iwl_send_bt_init_conf(struct iwl_mvm *mvm)
228 .flags = CMD_SYNC, 360 .flags = CMD_SYNC,
229 }; 361 };
230 int ret; 362 int ret;
363 u32 flags;
231 364
232 /* go to CALIB state in internal BT-Coex state machine */ 365 if (!(mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_NEWBT_COEX))
233 ret = iwl_send_bt_env(mvm, BT_COEX_ENV_OPEN, 366 return 0;
234 BT_COEX_PRIO_TBL_EVT_INIT_CALIB2);
235 if (ret)
236 return ret;
237
238 ret = iwl_send_bt_env(mvm, BT_COEX_ENV_CLOSE,
239 BT_COEX_PRIO_TBL_EVT_INIT_CALIB2);
240 if (ret)
241 return ret;
242 367
243 bt_cmd = kzalloc(sizeof(*bt_cmd), GFP_KERNEL); 368 bt_cmd = kzalloc(sizeof(*bt_cmd), GFP_KERNEL);
244 if (!bt_cmd) 369 if (!bt_cmd)
@@ -246,40 +371,52 @@ int iwl_send_bt_init_conf(struct iwl_mvm *mvm)
246 cmd.data[0] = bt_cmd; 371 cmd.data[0] = bt_cmd;
247 372
248 bt_cmd->max_kill = 5; 373 bt_cmd->max_kill = 5;
249 bt_cmd->bt3_time_t7_value = 1; 374 bt_cmd->bt4_antenna_isolation_thr = BT_ANTENNA_COUPLING_THRESHOLD,
250 bt_cmd->bt3_prio_sample_time = 2; 375 bt_cmd->bt4_antenna_isolation = iwlwifi_mod_params.ant_coupling,
251 bt_cmd->bt3_timer_t2_value = 0xc; 376 bt_cmd->bt4_tx_tx_delta_freq_thr = 15,
377 bt_cmd->bt4_tx_rx_max_freq0 = 15,
252 378
253 bt_cmd->flags = iwlwifi_mod_params.bt_coex_active ? 379 flags = iwlwifi_mod_params.bt_coex_active ?
254 BT_COEX_NW : BT_COEX_DISABLE; 380 BT_COEX_NW : BT_COEX_DISABLE;
255 bt_cmd->flags |= BT_CH_PRIMARY_EN | BT_SYNC_2_BT_DISABLE; 381 flags |= BT_CH_PRIMARY_EN | BT_CH_SECONDARY_EN | BT_SYNC_2_BT_DISABLE;
382 bt_cmd->flags = cpu_to_le32(flags);
256 383
257 bt_cmd->valid_bit_msk = cpu_to_le16(BT_VALID_ENABLE | 384 bt_cmd->valid_bit_msk = cpu_to_le32(BT_VALID_ENABLE |
258 BT_VALID_BT_PRIO_BOOST | 385 BT_VALID_BT_PRIO_BOOST |
259 BT_VALID_MAX_KILL | 386 BT_VALID_MAX_KILL |
260 BT_VALID_3W_TMRS | 387 BT_VALID_3W_TMRS |
261 BT_VALID_KILL_ACK | 388 BT_VALID_KILL_ACK |
262 BT_VALID_KILL_CTS | 389 BT_VALID_KILL_CTS |
263 BT_VALID_REDUCED_TX_POWER | 390 BT_VALID_REDUCED_TX_POWER |
264 BT_VALID_LUT); 391 BT_VALID_LUT |
392 BT_VALID_WIFI_RX_SW_PRIO_BOOST |
393 BT_VALID_WIFI_TX_SW_PRIO_BOOST |
394 BT_VALID_MULTI_PRIO_LUT |
395 BT_VALID_CORUN_LUT_20 |
396 BT_VALID_CORUN_LUT_40 |
397 BT_VALID_ANT_ISOLATION |
398 BT_VALID_ANT_ISOLATION_THRS |
399 BT_VALID_TXTX_DELTA_FREQ_THRS |
400 BT_VALID_TXRX_MAX_FREQ_0);
265 401
266 if (mvm->cfg->bt_shared_single_ant) 402 if (mvm->cfg->bt_shared_single_ant)
267 memcpy(&bt_cmd->decision_lut, iwl_single_shared_ant_lookup, 403 memcpy(&bt_cmd->decision_lut, iwl_single_shared_ant,
268 sizeof(iwl_single_shared_ant_lookup)); 404 sizeof(iwl_single_shared_ant));
269 else if (is_loose_coex())
270 memcpy(&bt_cmd->decision_lut, iwl_loose_lookup,
271 sizeof(iwl_tight_lookup));
272 else 405 else
273 memcpy(&bt_cmd->decision_lut, iwl_tight_lookup, 406 memcpy(&bt_cmd->decision_lut, iwl_combined_lookup,
274 sizeof(iwl_tight_lookup)); 407 sizeof(iwl_combined_lookup));
275 408
276 bt_cmd->bt_prio_boost = cpu_to_le32(IWL_BT_DEFAULT_BOOST); 409 memcpy(&bt_cmd->bt_prio_boost, iwl_bt_prio_boost,
410 sizeof(iwl_bt_prio_boost));
411 memcpy(&bt_cmd->bt4_multiprio_lut, iwl_bt_mprio_lut,
412 sizeof(iwl_bt_mprio_lut));
277 bt_cmd->kill_ack_msk = 413 bt_cmd->kill_ack_msk =
278 cpu_to_le32(iwl_bt_ack_kill_msk[BT_KILL_MSK_DEFAULT]); 414 cpu_to_le32(iwl_bt_ack_kill_msk[BT_KILL_MSK_DEFAULT]);
279 bt_cmd->kill_cts_msk = 415 bt_cmd->kill_cts_msk =
280 cpu_to_le32(iwl_bt_cts_kill_msk[BT_KILL_MSK_DEFAULT]); 416 cpu_to_le32(iwl_bt_cts_kill_msk[BT_KILL_MSK_DEFAULT]);
281 417
282 memset(&mvm->last_bt_notif, 0, sizeof(mvm->last_bt_notif)); 418 memset(&mvm->last_bt_notif, 0, sizeof(mvm->last_bt_notif));
419 memset(&mvm->last_bt_ci_cmd, 0, sizeof(mvm->last_bt_ci_cmd));
283 420
284 ret = iwl_mvm_send_cmd(mvm, &cmd); 421 ret = iwl_mvm_send_cmd(mvm, &cmd);
285 422
@@ -334,13 +471,17 @@ static int iwl_mvm_bt_udpate_ctrl_kill_msk(struct iwl_mvm *mvm,
334 if (!bt_cmd) 471 if (!bt_cmd)
335 return -ENOMEM; 472 return -ENOMEM;
336 cmd.data[0] = bt_cmd; 473 cmd.data[0] = bt_cmd;
474 bt_cmd->flags = cpu_to_le32(BT_COEX_NW);
337 475
338 bt_cmd->kill_ack_msk = cpu_to_le32(iwl_bt_ack_kill_msk[bt_kill_msk]); 476 bt_cmd->kill_ack_msk = cpu_to_le32(iwl_bt_ack_kill_msk[bt_kill_msk]);
339 bt_cmd->kill_cts_msk = cpu_to_le32(iwl_bt_cts_kill_msk[bt_kill_msk]); 477 bt_cmd->kill_cts_msk = cpu_to_le32(iwl_bt_cts_kill_msk[bt_kill_msk]);
340 bt_cmd->valid_bit_msk = 478 bt_cmd->valid_bit_msk |= cpu_to_le32(BT_VALID_ENABLE |
341 cpu_to_le16(BT_VALID_KILL_ACK | BT_VALID_KILL_CTS); 479 BT_VALID_KILL_ACK |
480 BT_VALID_KILL_CTS);
342 481
343 IWL_DEBUG_COEX(mvm, "bt_kill_msk = %d\n", bt_kill_msk); 482 IWL_DEBUG_COEX(mvm, "ACK Kill msk = 0x%08x, CTS Kill msk = 0x%08x\n",
483 iwl_bt_ack_kill_msk[bt_kill_msk],
484 iwl_bt_cts_kill_msk[bt_kill_msk]);
344 485
345 ret = iwl_mvm_send_cmd(mvm, &cmd); 486 ret = iwl_mvm_send_cmd(mvm, &cmd);
346 487
@@ -364,12 +505,16 @@ static int iwl_mvm_bt_coex_reduced_txp(struct iwl_mvm *mvm, u8 sta_id,
364 struct iwl_mvm_sta *mvmsta; 505 struct iwl_mvm_sta *mvmsta;
365 int ret; 506 int ret;
366 507
367 /* This can happen if the station has been removed right now */
368 if (sta_id == IWL_MVM_STATION_COUNT) 508 if (sta_id == IWL_MVM_STATION_COUNT)
369 return 0; 509 return 0;
370 510
371 sta = rcu_dereference_protected(mvm->fw_id_to_mac_id[sta_id], 511 sta = rcu_dereference_protected(mvm->fw_id_to_mac_id[sta_id],
372 lockdep_is_held(&mvm->mutex)); 512 lockdep_is_held(&mvm->mutex));
513
514 /* This can happen if the station has been removed right now */
515 if (IS_ERR_OR_NULL(sta))
516 return 0;
517
373 mvmsta = (void *)sta->drv_priv; 518 mvmsta = (void *)sta->drv_priv;
374 519
375 /* nothing to do */ 520 /* nothing to do */
@@ -380,8 +525,10 @@ static int iwl_mvm_bt_coex_reduced_txp(struct iwl_mvm *mvm, u8 sta_id,
380 if (!bt_cmd) 525 if (!bt_cmd)
381 return -ENOMEM; 526 return -ENOMEM;
382 cmd.data[0] = bt_cmd; 527 cmd.data[0] = bt_cmd;
528 bt_cmd->flags = cpu_to_le32(BT_COEX_NW);
383 529
384 bt_cmd->valid_bit_msk = cpu_to_le16(BT_VALID_REDUCED_TX_POWER), 530 bt_cmd->valid_bit_msk =
531 cpu_to_le32(BT_VALID_ENABLE | BT_VALID_REDUCED_TX_POWER);
385 bt_cmd->bt_reduced_tx_power = sta_id; 532 bt_cmd->bt_reduced_tx_power = sta_id;
386 533
387 if (enable) 534 if (enable)
@@ -403,8 +550,25 @@ struct iwl_bt_iterator_data {
403 struct iwl_mvm *mvm; 550 struct iwl_mvm *mvm;
404 u32 num_bss_ifaces; 551 u32 num_bss_ifaces;
405 bool reduced_tx_power; 552 bool reduced_tx_power;
553 struct ieee80211_chanctx_conf *primary;
554 struct ieee80211_chanctx_conf *secondary;
406}; 555};
407 556
557static inline
558void iwl_mvm_bt_coex_enable_rssi_event(struct iwl_mvm *mvm,
559 struct ieee80211_vif *vif,
560 bool enable, int rssi)
561{
562 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif);
563
564 mvmvif->bf_data.last_bt_coex_event = rssi;
565 mvmvif->bf_data.bt_coex_max_thold =
566 enable ? BT_ENABLE_REDUCED_TXPOWER_THRESHOLD : 0;
567 mvmvif->bf_data.bt_coex_min_thold =
568 enable ? BT_DISABLE_REDUCED_TXPOWER_THRESHOLD : 0;
569}
570
571/* must be called under rcu_read_lock */
408static void iwl_mvm_bt_notif_iterator(void *_data, u8 *mac, 572static void iwl_mvm_bt_notif_iterator(void *_data, u8 *mac,
409 struct ieee80211_vif *vif) 573 struct ieee80211_vif *vif)
410{ 574{
@@ -413,65 +577,94 @@ static void iwl_mvm_bt_notif_iterator(void *_data, u8 *mac,
413 struct iwl_mvm *mvm = data->mvm; 577 struct iwl_mvm *mvm = data->mvm;
414 struct ieee80211_chanctx_conf *chanctx_conf; 578 struct ieee80211_chanctx_conf *chanctx_conf;
415 enum ieee80211_smps_mode smps_mode; 579 enum ieee80211_smps_mode smps_mode;
416 enum ieee80211_band band;
417 int ave_rssi; 580 int ave_rssi;
418 581
419 lockdep_assert_held(&mvm->mutex); 582 lockdep_assert_held(&mvm->mutex);
420 if (vif->type != NL80211_IFTYPE_STATION)
421 return;
422 583
423 rcu_read_lock(); 584 if (vif->type != NL80211_IFTYPE_STATION &&
424 chanctx_conf = rcu_dereference(vif->chanctx_conf); 585 vif->type != NL80211_IFTYPE_AP)
425 if (chanctx_conf && chanctx_conf->def.chan) 586 return;
426 band = chanctx_conf->def.chan->band;
427 else
428 band = -1;
429 rcu_read_unlock();
430 587
431 smps_mode = IEEE80211_SMPS_AUTOMATIC; 588 smps_mode = IEEE80211_SMPS_AUTOMATIC;
432 589
433 /* non associated BSSes aren't to be considered */ 590 chanctx_conf = rcu_dereference(vif->chanctx_conf);
434 if (!vif->bss_conf.assoc) 591
592 /* If channel context is invalid or not on 2.4GHz .. */
593 if ((!chanctx_conf ||
594 chanctx_conf->def.chan->band != IEEE80211_BAND_2GHZ)) {
595 /* ... and it is an associated STATION, relax constraints */
596 if (vif->type == NL80211_IFTYPE_STATION && vif->bss_conf.assoc)
597 iwl_mvm_update_smps(mvm, vif, IWL_MVM_SMPS_REQ_BT_COEX,
598 smps_mode);
599 iwl_mvm_bt_coex_enable_rssi_event(mvm, vif, false, 0);
435 return; 600 return;
601 }
602
603 /* SoftAP / GO will always be primary */
604 if (vif->type == NL80211_IFTYPE_AP) {
605 if (!mvmvif->ap_ibss_active)
606 return;
436 607
437 if (band != IEEE80211_BAND_2GHZ) { 608 /* the Ack / Cts kill mask must be default if AP / GO */
438 iwl_mvm_update_smps(mvm, vif, IWL_MVM_SMPS_REQ_BT_COEX, 609 data->reduced_tx_power = false;
439 smps_mode); 610
611 if (chanctx_conf == data->primary)
612 return;
613
614 /* downgrade the current primary no matter what its type is */
615 data->secondary = data->primary;
616 data->primary = chanctx_conf;
440 return; 617 return;
441 } 618 }
442 619
443 if (data->notif->bt_status) 620 data->num_bss_ifaces++;
444 smps_mode = IEEE80211_SMPS_DYNAMIC; 621
622 /* we are now a STA / P2P Client, and take associated ones only */
623 if (!vif->bss_conf.assoc)
624 return;
625
626 /* STA / P2P Client, try to be primary if first vif */
627 if (!data->primary || data->primary == chanctx_conf)
628 data->primary = chanctx_conf;
629 else if (!data->secondary)
630 /* if secondary is not NULL, it might be a GO */
631 data->secondary = chanctx_conf;
445 632
446 if (data->notif->bt_traffic_load >= IWL_BT_LOAD_FORCE_SISO_THRESHOLD) 633 if (le32_to_cpu(data->notif->bt_activity_grading) >= BT_HIGH_TRAFFIC)
447 smps_mode = IEEE80211_SMPS_STATIC; 634 smps_mode = IEEE80211_SMPS_STATIC;
635 else if (le32_to_cpu(data->notif->bt_activity_grading) >=
636 BT_LOW_TRAFFIC)
637 smps_mode = IEEE80211_SMPS_DYNAMIC;
448 638
449 IWL_DEBUG_COEX(data->mvm, 639 IWL_DEBUG_COEX(data->mvm,
450 "mac %d: bt_status %d traffic_load %d smps_req %d\n", 640 "mac %d: bt_status %d bt_activity_grading %d smps_req %d\n",
451 mvmvif->id, data->notif->bt_status, 641 mvmvif->id, data->notif->bt_status,
452 data->notif->bt_traffic_load, smps_mode); 642 data->notif->bt_activity_grading, smps_mode);
453 643
454 iwl_mvm_update_smps(mvm, vif, IWL_MVM_SMPS_REQ_BT_COEX, smps_mode); 644 iwl_mvm_update_smps(mvm, vif, IWL_MVM_SMPS_REQ_BT_COEX, smps_mode);
455 645
456 /* don't reduce the Tx power if in loose scheme */ 646 /* don't reduce the Tx power if in loose scheme */
457 if (is_loose_coex()) 647 if (iwl_get_coex_type(mvm, vif) == BT_COEX_LOOSE_LUT ||
648 mvm->cfg->bt_shared_single_ant) {
649 data->reduced_tx_power = false;
650 iwl_mvm_bt_coex_enable_rssi_event(mvm, vif, false, 0);
458 return; 651 return;
652 }
459 653
460 data->num_bss_ifaces++; 654 /* reduced Txpower only if BT is on, so ...*/
461 655 if (!data->notif->bt_status) {
462 /* reduced Txpower only if there are open BT connections, so ...*/
463 if (!BT_MBOX_MSG(data->notif, 3, OPEN_CON_2)) {
464 /* ... cancel reduced Tx power ... */ 656 /* ... cancel reduced Tx power ... */
465 if (iwl_mvm_bt_coex_reduced_txp(mvm, mvmvif->ap_sta_id, false)) 657 if (iwl_mvm_bt_coex_reduced_txp(mvm, mvmvif->ap_sta_id, false))
466 IWL_ERR(mvm, "Couldn't send BT_CONFIG cmd\n"); 658 IWL_ERR(mvm, "Couldn't send BT_CONFIG cmd\n");
467 data->reduced_tx_power = false; 659 data->reduced_tx_power = false;
468 660
469 /* ... and there is no need to get reports on RSSI any more. */ 661 /* ... and there is no need to get reports on RSSI any more. */
470 ieee80211_disable_rssi_reports(vif); 662 iwl_mvm_bt_coex_enable_rssi_event(mvm, vif, false, 0);
471 return; 663 return;
472 } 664 }
473 665
474 ave_rssi = ieee80211_ave_rssi(vif); 666 /* try to get the avg rssi from fw */
667 ave_rssi = mvmvif->bf_data.ave_beacon_signal;
475 668
476 /* if the RSSI isn't valid, fake it is very low */ 669 /* if the RSSI isn't valid, fake it is very low */
477 if (!ave_rssi) 670 if (!ave_rssi)
@@ -499,8 +692,7 @@ static void iwl_mvm_bt_notif_iterator(void *_data, u8 *mac,
499 } 692 }
500 693
501 /* Begin to monitor the RSSI: it may influence the reduced Tx power */ 694 /* Begin to monitor the RSSI: it may influence the reduced Tx power */
502 ieee80211_enable_rssi_reports(vif, BT_DISABLE_REDUCED_TXPOWER_THRESHOLD, 695 iwl_mvm_bt_coex_enable_rssi_event(mvm, vif, true, ave_rssi);
503 BT_ENABLE_REDUCED_TXPOWER_THRESHOLD);
504} 696}
505 697
506static void iwl_mvm_bt_coex_notif_handle(struct iwl_mvm *mvm) 698static void iwl_mvm_bt_coex_notif_handle(struct iwl_mvm *mvm)
@@ -510,11 +702,72 @@ static void iwl_mvm_bt_coex_notif_handle(struct iwl_mvm *mvm)
510 .notif = &mvm->last_bt_notif, 702 .notif = &mvm->last_bt_notif,
511 .reduced_tx_power = true, 703 .reduced_tx_power = true,
512 }; 704 };
705 struct iwl_bt_coex_ci_cmd cmd = {};
706 u8 ci_bw_idx;
513 707
708 rcu_read_lock();
514 ieee80211_iterate_active_interfaces_atomic( 709 ieee80211_iterate_active_interfaces_atomic(
515 mvm->hw, IEEE80211_IFACE_ITER_NORMAL, 710 mvm->hw, IEEE80211_IFACE_ITER_NORMAL,
516 iwl_mvm_bt_notif_iterator, &data); 711 iwl_mvm_bt_notif_iterator, &data);
517 712
713 if (data.primary) {
714 struct ieee80211_chanctx_conf *chan = data.primary;
715 if (WARN_ON(!chan->def.chan)) {
716 rcu_read_unlock();
717 return;
718 }
719
720 if (chan->def.width < NL80211_CHAN_WIDTH_40) {
721 ci_bw_idx = 0;
722 cmd.co_run_bw_primary = 0;
723 } else {
724 cmd.co_run_bw_primary = 1;
725 if (chan->def.center_freq1 >
726 chan->def.chan->center_freq)
727 ci_bw_idx = 2;
728 else
729 ci_bw_idx = 1;
730 }
731
732 cmd.bt_primary_ci =
733 iwl_ci_mask[chan->def.chan->hw_value][ci_bw_idx];
734 cmd.primary_ch_phy_id = *((u16 *)data.primary->drv_priv);
735 }
736
737 if (data.secondary) {
738 struct ieee80211_chanctx_conf *chan = data.secondary;
739 if (WARN_ON(!data.secondary->def.chan)) {
740 rcu_read_unlock();
741 return;
742 }
743
744 if (chan->def.width < NL80211_CHAN_WIDTH_40) {
745 ci_bw_idx = 0;
746 cmd.co_run_bw_secondary = 0;
747 } else {
748 cmd.co_run_bw_secondary = 1;
749 if (chan->def.center_freq1 >
750 chan->def.chan->center_freq)
751 ci_bw_idx = 2;
752 else
753 ci_bw_idx = 1;
754 }
755
756 cmd.bt_secondary_ci =
757 iwl_ci_mask[chan->def.chan->hw_value][ci_bw_idx];
758 cmd.secondary_ch_phy_id = *((u16 *)data.secondary->drv_priv);
759 }
760
761 rcu_read_unlock();
762
763 /* Don't spam the fw with the same command over and over */
764 if (memcmp(&cmd, &mvm->last_bt_ci_cmd, sizeof(cmd))) {
765 if (iwl_mvm_send_cmd_pdu(mvm, BT_COEX_CI, CMD_SYNC,
766 sizeof(cmd), &cmd))
767 IWL_ERR(mvm, "Failed to send BT_CI cmd");
768 memcpy(&mvm->last_bt_ci_cmd, &cmd, sizeof(cmd));
769 }
770
518 /* 771 /*
519 * If there are no BSS / P2P client interfaces, reduced Tx Power is 772 * If there are no BSS / P2P client interfaces, reduced Tx Power is
520 * irrelevant since it is based on the RSSI coming from the beacon. 773 * irrelevant since it is based on the RSSI coming from the beacon.
@@ -536,12 +789,18 @@ int iwl_mvm_rx_bt_coex_notif(struct iwl_mvm *mvm,
536 789
537 790
538 IWL_DEBUG_COEX(mvm, "BT Coex Notification received\n"); 791 IWL_DEBUG_COEX(mvm, "BT Coex Notification received\n");
539 IWL_DEBUG_COEX(mvm, "\tBT %salive\n", notif->bt_status ? "" : "not "); 792 IWL_DEBUG_COEX(mvm, "\tBT status: %s\n",
793 notif->bt_status ? "ON" : "OFF");
540 IWL_DEBUG_COEX(mvm, "\tBT open conn %d\n", notif->bt_open_conn); 794 IWL_DEBUG_COEX(mvm, "\tBT open conn %d\n", notif->bt_open_conn);
541 IWL_DEBUG_COEX(mvm, "\tBT traffic load %d\n", notif->bt_traffic_load); 795 IWL_DEBUG_COEX(mvm, "\tBT ci compliance %d\n", notif->bt_ci_compliance);
796 IWL_DEBUG_COEX(mvm, "\tBT primary_ch_lut %d\n",
797 le32_to_cpu(notif->primary_ch_lut));
798 IWL_DEBUG_COEX(mvm, "\tBT secondary_ch_lut %d\n",
799 le32_to_cpu(notif->secondary_ch_lut));
800 IWL_DEBUG_COEX(mvm, "\tBT activity grading %d\n",
801 le32_to_cpu(notif->bt_activity_grading));
542 IWL_DEBUG_COEX(mvm, "\tBT agg traffic load %d\n", 802 IWL_DEBUG_COEX(mvm, "\tBT agg traffic load %d\n",
543 notif->bt_agg_traffic_load); 803 notif->bt_agg_traffic_load);
544 IWL_DEBUG_COEX(mvm, "\tBT ci compliance %d\n", notif->bt_ci_compliance);
545 804
546 /* remember this notification for future use: rssi fluctuations */ 805 /* remember this notification for future use: rssi fluctuations */
547 memcpy(&mvm->last_bt_notif, notif, sizeof(mvm->last_bt_notif)); 806 memcpy(&mvm->last_bt_notif, notif, sizeof(mvm->last_bt_notif));
@@ -565,6 +824,18 @@ static void iwl_mvm_bt_rssi_iterator(void *_data, u8 *mac,
565 struct ieee80211_sta *sta; 824 struct ieee80211_sta *sta;
566 struct iwl_mvm_sta *mvmsta; 825 struct iwl_mvm_sta *mvmsta;
567 826
827 struct ieee80211_chanctx_conf *chanctx_conf;
828
829 rcu_read_lock();
830 chanctx_conf = rcu_dereference(vif->chanctx_conf);
831 /* If channel context is invalid or not on 2.4GHz - don't count it */
832 if (!chanctx_conf ||
833 chanctx_conf->def.chan->band != IEEE80211_BAND_2GHZ) {
834 rcu_read_unlock();
835 return;
836 }
837 rcu_read_unlock();
838
568 if (vif->type != NL80211_IFTYPE_STATION || 839 if (vif->type != NL80211_IFTYPE_STATION ||
569 mvmvif->ap_sta_id == IWL_MVM_STATION_COUNT) 840 mvmvif->ap_sta_id == IWL_MVM_STATION_COUNT)
570 return; 841 return;
@@ -594,15 +865,15 @@ void iwl_mvm_bt_rssi_event(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
594 }; 865 };
595 int ret; 866 int ret;
596 867
597 mutex_lock(&mvm->mutex); 868 lockdep_assert_held(&mvm->mutex);
598 869
599 /* Rssi update while not associated ?! */ 870 /* Rssi update while not associated ?! */
600 if (WARN_ON_ONCE(mvmvif->ap_sta_id == IWL_MVM_STATION_COUNT)) 871 if (WARN_ON_ONCE(mvmvif->ap_sta_id == IWL_MVM_STATION_COUNT))
601 goto out_unlock; 872 return;
602 873
603 /* No open connection - reports should be disabled */ 874 /* No BT - reports should be disabled */
604 if (!BT_MBOX_MSG(&mvm->last_bt_notif, 3, OPEN_CON_2)) 875 if (!mvm->last_bt_notif.bt_status)
605 goto out_unlock; 876 return;
606 877
607 IWL_DEBUG_COEX(mvm, "RSSI for %pM is now %s\n", vif->bss_conf.bssid, 878 IWL_DEBUG_COEX(mvm, "RSSI for %pM is now %s\n", vif->bss_conf.bssid,
608 rssi_event == RSSI_EVENT_HIGH ? "HIGH" : "LOW"); 879 rssi_event == RSSI_EVENT_HIGH ? "HIGH" : "LOW");
@@ -611,7 +882,8 @@ void iwl_mvm_bt_rssi_event(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
611 * Check if rssi is good enough for reduced Tx power, but not in loose 882 * Check if rssi is good enough for reduced Tx power, but not in loose
612 * scheme. 883 * scheme.
613 */ 884 */
614 if (rssi_event == RSSI_EVENT_LOW || is_loose_coex()) 885 if (rssi_event == RSSI_EVENT_LOW || mvm->cfg->bt_shared_single_ant ||
886 iwl_get_coex_type(mvm, vif) == BT_COEX_LOOSE_LUT)
615 ret = iwl_mvm_bt_coex_reduced_txp(mvm, mvmvif->ap_sta_id, 887 ret = iwl_mvm_bt_coex_reduced_txp(mvm, mvmvif->ap_sta_id,
616 false); 888 false);
617 else 889 else
@@ -633,12 +905,52 @@ void iwl_mvm_bt_rssi_event(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
633 905
634 if (iwl_mvm_bt_udpate_ctrl_kill_msk(mvm, data.reduced_tx_power)) 906 if (iwl_mvm_bt_udpate_ctrl_kill_msk(mvm, data.reduced_tx_power))
635 IWL_ERR(mvm, "Failed to update the ctrl_kill_msk\n"); 907 IWL_ERR(mvm, "Failed to update the ctrl_kill_msk\n");
908}
636 909
637 out_unlock: 910#define LINK_QUAL_AGG_TIME_LIMIT_DEF (4000)
638 mutex_unlock(&mvm->mutex); 911#define LINK_QUAL_AGG_TIME_LIMIT_BT_ACT (1200)
912
913u16 iwl_mvm_bt_coex_agg_time_limit(struct iwl_mvm *mvm,
914 struct ieee80211_sta *sta)
915{
916 struct iwl_mvm_sta *mvmsta = (void *)sta->drv_priv;
917 enum iwl_bt_coex_lut_type lut_type;
918
919 if (le32_to_cpu(mvm->last_bt_notif.bt_activity_grading) <
920 BT_LOW_TRAFFIC)
921 return LINK_QUAL_AGG_TIME_LIMIT_DEF;
922
923 lut_type = iwl_get_coex_type(mvm, mvmsta->vif);
924
925 if (lut_type == BT_COEX_LOOSE_LUT)
926 return LINK_QUAL_AGG_TIME_LIMIT_DEF;
927
928 /* tight coex, high bt traffic, reduce AGG time limit */
929 return LINK_QUAL_AGG_TIME_LIMIT_BT_ACT;
639} 930}
640 931
641void iwl_mvm_bt_coex_vif_assoc(struct iwl_mvm *mvm, struct ieee80211_vif *vif) 932bool iwl_mvm_bt_coex_is_mimo_allowed(struct iwl_mvm *mvm,
933 struct ieee80211_sta *sta)
642{ 934{
935 struct iwl_mvm_sta *mvmsta = (void *)sta->drv_priv;
936
937 if (le32_to_cpu(mvm->last_bt_notif.bt_activity_grading) <
938 BT_HIGH_TRAFFIC)
939 return true;
940
941 /*
942 * In Tight, BT can't Rx while we Tx, so use both antennas since BT is
943 * already killed.
944 * In Loose, BT can Rx while we Tx, so forbid MIMO to let BT Rx while we
945 * Tx.
946 */
947 return iwl_get_coex_type(mvm, mvmsta->vif) == BT_COEX_TIGHT_LUT;
948}
949
950void iwl_mvm_bt_coex_vif_change(struct iwl_mvm *mvm)
951{
952 if (!(mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_NEWBT_COEX))
953 return;
954
643 iwl_mvm_bt_coex_notif_handle(mvm); 955 iwl_mvm_bt_coex_notif_handle(mvm);
644} 956}
diff --git a/drivers/net/wireless/iwlwifi/mvm/constants.h b/drivers/net/wireless/iwlwifi/mvm/constants.h
index 2bf29f7992ee..4b6d670c3509 100644
--- a/drivers/net/wireless/iwlwifi/mvm/constants.h
+++ b/drivers/net/wireless/iwlwifi/mvm/constants.h
@@ -70,7 +70,9 @@
70#define IWL_MVM_UAPSD_RX_DATA_TIMEOUT (50 * USEC_PER_MSEC) 70#define IWL_MVM_UAPSD_RX_DATA_TIMEOUT (50 * USEC_PER_MSEC)
71#define IWL_MVM_UAPSD_TX_DATA_TIMEOUT (50 * USEC_PER_MSEC) 71#define IWL_MVM_UAPSD_TX_DATA_TIMEOUT (50 * USEC_PER_MSEC)
72#define IWL_MVM_PS_HEAVY_TX_THLD_PACKETS 20 72#define IWL_MVM_PS_HEAVY_TX_THLD_PACKETS 20
73#define IWL_MVM_PS_HEAVY_RX_THLD_PACKETS 20 73#define IWL_MVM_PS_HEAVY_RX_THLD_PACKETS 8
74#define IWL_MVM_PS_SNOOZE_HEAVY_TX_THLD_PACKETS 30
75#define IWL_MVM_PS_SNOOZE_HEAVY_RX_THLD_PACKETS 20
74#define IWL_MVM_PS_HEAVY_TX_THLD_PERCENT 50 76#define IWL_MVM_PS_HEAVY_TX_THLD_PERCENT 50
75#define IWL_MVM_PS_HEAVY_RX_THLD_PERCENT 50 77#define IWL_MVM_PS_HEAVY_RX_THLD_PERCENT 50
76#define IWL_MVM_PS_SNOOZE_INTERVAL 25 78#define IWL_MVM_PS_SNOOZE_INTERVAL 25
diff --git a/drivers/net/wireless/iwlwifi/mvm/d3.c b/drivers/net/wireless/iwlwifi/mvm/d3.c
index 417639f77b01..6f45966817bb 100644
--- a/drivers/net/wireless/iwlwifi/mvm/d3.c
+++ b/drivers/net/wireless/iwlwifi/mvm/d3.c
@@ -67,6 +67,7 @@
67#include <net/cfg80211.h> 67#include <net/cfg80211.h>
68#include <net/ipv6.h> 68#include <net/ipv6.h>
69#include <net/tcp.h> 69#include <net/tcp.h>
70#include <net/addrconf.h>
70#include "iwl-modparams.h" 71#include "iwl-modparams.h"
71#include "fw-api.h" 72#include "fw-api.h"
72#include "mvm.h" 73#include "mvm.h"
@@ -381,14 +382,74 @@ static int iwl_mvm_send_proto_offload(struct iwl_mvm *mvm,
381 union { 382 union {
382 struct iwl_proto_offload_cmd_v1 v1; 383 struct iwl_proto_offload_cmd_v1 v1;
383 struct iwl_proto_offload_cmd_v2 v2; 384 struct iwl_proto_offload_cmd_v2 v2;
385 struct iwl_proto_offload_cmd_v3_small v3s;
386 struct iwl_proto_offload_cmd_v3_large v3l;
384 } cmd = {}; 387 } cmd = {};
388 struct iwl_host_cmd hcmd = {
389 .id = PROT_OFFLOAD_CONFIG_CMD,
390 .flags = CMD_SYNC,
391 .data[0] = &cmd,
392 .dataflags[0] = IWL_HCMD_DFL_DUP,
393 };
385 struct iwl_proto_offload_cmd_common *common; 394 struct iwl_proto_offload_cmd_common *common;
386 u32 enabled = 0, size; 395 u32 enabled = 0, size;
396 u32 capa_flags = mvm->fw->ucode_capa.flags;
387#if IS_ENABLED(CONFIG_IPV6) 397#if IS_ENABLED(CONFIG_IPV6)
388 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 398 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif);
389 int i; 399 int i;
390 400
391 if (mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_D3_6_IPV6_ADDRS) { 401 if (capa_flags & IWL_UCODE_TLV_FLAGS_NEW_NSOFFL_SMALL ||
402 capa_flags & IWL_UCODE_TLV_FLAGS_NEW_NSOFFL_LARGE) {
403 struct iwl_ns_config *nsc;
404 struct iwl_targ_addr *addrs;
405 int n_nsc, n_addrs;
406 int c;
407
408 if (capa_flags & IWL_UCODE_TLV_FLAGS_NEW_NSOFFL_SMALL) {
409 nsc = cmd.v3s.ns_config;
410 n_nsc = IWL_PROTO_OFFLOAD_NUM_NS_CONFIG_V3S;
411 addrs = cmd.v3s.targ_addrs;
412 n_addrs = IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_V3S;
413 } else {
414 nsc = cmd.v3l.ns_config;
415 n_nsc = IWL_PROTO_OFFLOAD_NUM_NS_CONFIG_V3L;
416 addrs = cmd.v3l.targ_addrs;
417 n_addrs = IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_V3L;
418 }
419
420 if (mvmvif->num_target_ipv6_addrs)
421 enabled |= IWL_D3_PROTO_OFFLOAD_NS;
422
423 /*
424 * For each address we have (and that will fit) fill a target
425 * address struct and combine for NS offload structs with the
426 * solicited node addresses.
427 */
428 for (i = 0, c = 0;
429 i < mvmvif->num_target_ipv6_addrs &&
430 i < n_addrs && c < n_nsc; i++) {
431 struct in6_addr solicited_addr;
432 int j;
433
434 addrconf_addr_solict_mult(&mvmvif->target_ipv6_addrs[i],
435 &solicited_addr);
436 for (j = 0; j < c; j++)
437 if (ipv6_addr_cmp(&nsc[j].dest_ipv6_addr,
438 &solicited_addr) == 0)
439 break;
440 if (j == c)
441 c++;
442 addrs[i].addr = mvmvif->target_ipv6_addrs[i];
443 addrs[i].config_num = cpu_to_le32(j);
444 nsc[j].dest_ipv6_addr = solicited_addr;
445 memcpy(nsc[j].target_mac_addr, vif->addr, ETH_ALEN);
446 }
447
448 if (capa_flags & IWL_UCODE_TLV_FLAGS_NEW_NSOFFL_SMALL)
449 cmd.v3s.num_valid_ipv6_addrs = cpu_to_le32(i);
450 else
451 cmd.v3l.num_valid_ipv6_addrs = cpu_to_le32(i);
452 } else if (capa_flags & IWL_UCODE_TLV_FLAGS_D3_6_IPV6_ADDRS) {
392 if (mvmvif->num_target_ipv6_addrs) { 453 if (mvmvif->num_target_ipv6_addrs) {
393 enabled |= IWL_D3_PROTO_OFFLOAD_NS; 454 enabled |= IWL_D3_PROTO_OFFLOAD_NS;
394 memcpy(cmd.v2.ndp_mac_addr, vif->addr, ETH_ALEN); 455 memcpy(cmd.v2.ndp_mac_addr, vif->addr, ETH_ALEN);
@@ -419,7 +480,13 @@ static int iwl_mvm_send_proto_offload(struct iwl_mvm *mvm,
419 } 480 }
420#endif 481#endif
421 482
422 if (mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_D3_6_IPV6_ADDRS) { 483 if (capa_flags & IWL_UCODE_TLV_FLAGS_NEW_NSOFFL_SMALL) {
484 common = &cmd.v3s.common;
485 size = sizeof(cmd.v3s);
486 } else if (capa_flags & IWL_UCODE_TLV_FLAGS_NEW_NSOFFL_LARGE) {
487 common = &cmd.v3l.common;
488 size = sizeof(cmd.v3l);
489 } else if (capa_flags & IWL_UCODE_TLV_FLAGS_D3_6_IPV6_ADDRS) {
423 common = &cmd.v2.common; 490 common = &cmd.v2.common;
424 size = sizeof(cmd.v2); 491 size = sizeof(cmd.v2);
425 } else { 492 } else {
@@ -438,8 +505,8 @@ static int iwl_mvm_send_proto_offload(struct iwl_mvm *mvm,
438 505
439 common->enabled = cpu_to_le32(enabled); 506 common->enabled = cpu_to_le32(enabled);
440 507
441 return iwl_mvm_send_cmd_pdu(mvm, PROT_OFFLOAD_CONFIG_CMD, CMD_SYNC, 508 hcmd.len[0] = size;
442 size, &cmd); 509 return iwl_mvm_send_cmd(mvm, &hcmd);
443} 510}
444 511
445enum iwl_mvm_tcp_packet_type { 512enum iwl_mvm_tcp_packet_type {
@@ -793,6 +860,74 @@ static int iwl_mvm_d3_reprogram(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
793 return 0; 860 return 0;
794} 861}
795 862
863static int iwl_mvm_get_last_nonqos_seq(struct iwl_mvm *mvm,
864 struct ieee80211_vif *vif)
865{
866 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif);
867 struct iwl_nonqos_seq_query_cmd query_cmd = {
868 .get_set_flag = cpu_to_le32(IWL_NONQOS_SEQ_GET),
869 .mac_id_n_color =
870 cpu_to_le32(FW_CMD_ID_AND_COLOR(mvmvif->id,
871 mvmvif->color)),
872 };
873 struct iwl_host_cmd cmd = {
874 .id = NON_QOS_TX_COUNTER_CMD,
875 .flags = CMD_SYNC | CMD_WANT_SKB,
876 };
877 int err;
878 u32 size;
879
880 if (mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_D3_CONTINUITY_API) {
881 cmd.data[0] = &query_cmd;
882 cmd.len[0] = sizeof(query_cmd);
883 }
884
885 err = iwl_mvm_send_cmd(mvm, &cmd);
886 if (err)
887 return err;
888
889 size = le32_to_cpu(cmd.resp_pkt->len_n_flags) & FH_RSCSR_FRAME_SIZE_MSK;
890 size -= sizeof(cmd.resp_pkt->hdr);
891 if (size < sizeof(__le16)) {
892 err = -EINVAL;
893 } else {
894 err = le16_to_cpup((__le16 *)cmd.resp_pkt->data);
895 /* new API returns next, not last-used seqno */
896 if (mvm->fw->ucode_capa.flags &
897 IWL_UCODE_TLV_FLAGS_D3_CONTINUITY_API)
898 err -= 0x10;
899 }
900
901 iwl_free_resp(&cmd);
902 return err;
903}
904
905void iwl_mvm_set_last_nonqos_seq(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
906{
907 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif);
908 struct iwl_nonqos_seq_query_cmd query_cmd = {
909 .get_set_flag = cpu_to_le32(IWL_NONQOS_SEQ_SET),
910 .mac_id_n_color =
911 cpu_to_le32(FW_CMD_ID_AND_COLOR(mvmvif->id,
912 mvmvif->color)),
913 .value = cpu_to_le16(mvmvif->seqno),
914 };
915
916 /* return if called during restart, not resume from D3 */
917 if (!mvmvif->seqno_valid)
918 return;
919
920 mvmvif->seqno_valid = false;
921
922 if (!(mvm->fw->ucode_capa.flags &
923 IWL_UCODE_TLV_FLAGS_D3_CONTINUITY_API))
924 return;
925
926 if (iwl_mvm_send_cmd_pdu(mvm, NON_QOS_TX_COUNTER_CMD, CMD_SYNC,
927 sizeof(query_cmd), &query_cmd))
928 IWL_ERR(mvm, "failed to set non-QoS seqno\n");
929}
930
796static int __iwl_mvm_suspend(struct ieee80211_hw *hw, 931static int __iwl_mvm_suspend(struct ieee80211_hw *hw,
797 struct cfg80211_wowlan *wowlan, 932 struct cfg80211_wowlan *wowlan,
798 bool test) 933 bool test)
@@ -829,7 +964,6 @@ static int __iwl_mvm_suspend(struct ieee80211_hw *hw,
829 }; 964 };
830 int ret, i; 965 int ret, i;
831 int len __maybe_unused; 966 int len __maybe_unused;
832 u16 seq;
833 u8 old_aux_sta_id, old_ap_sta_id = IWL_MVM_STATION_COUNT; 967 u8 old_aux_sta_id, old_ap_sta_id = IWL_MVM_STATION_COUNT;
834 968
835 if (!wowlan) { 969 if (!wowlan) {
@@ -872,26 +1006,15 @@ static int __iwl_mvm_suspend(struct ieee80211_hw *hw,
872 1006
873 mvm_ap_sta = (struct iwl_mvm_sta *)ap_sta->drv_priv; 1007 mvm_ap_sta = (struct iwl_mvm_sta *)ap_sta->drv_priv;
874 1008
875 /*
876 * The D3 firmware still hardcodes the AP station ID for the
877 * BSS we're associated with as 0. Store the real STA ID here
878 * and assign 0. When we leave this function, we'll restore
879 * the original value for the resume code.
880 */
881 old_ap_sta_id = mvm_ap_sta->sta_id;
882 mvm_ap_sta->sta_id = 0;
883 mvmvif->ap_sta_id = 0;
884
885 /* TODO: wowlan_config_cmd.wowlan_ba_teardown_tids */ 1009 /* TODO: wowlan_config_cmd.wowlan_ba_teardown_tids */
886 1010
887 wowlan_config_cmd.is_11n_connection = ap_sta->ht_cap.ht_supported; 1011 wowlan_config_cmd.is_11n_connection = ap_sta->ht_cap.ht_supported;
888 1012
889 /* 1013 /* Query the last used seqno and set it */
890 * We know the last used seqno, and the uCode expects to know that 1014 ret = iwl_mvm_get_last_nonqos_seq(mvm, vif);
891 * one, it will increment before TX. 1015 if (ret < 0)
892 */ 1016 goto out_noreset;
893 seq = mvm_ap_sta->last_seq_ctl & IEEE80211_SCTL_SEQ; 1017 wowlan_config_cmd.non_qos_seq = cpu_to_le16(ret);
894 wowlan_config_cmd.non_qos_seq = cpu_to_le16(seq);
895 1018
896 /* 1019 /*
897 * For QoS counters, we store the one to use next, so subtract 0x10 1020 * For QoS counters, we store the one to use next, so subtract 0x10
@@ -899,7 +1022,7 @@ static int __iwl_mvm_suspend(struct ieee80211_hw *hw,
899 * increment after using the value (i.e. store the next value to use). 1022 * increment after using the value (i.e. store the next value to use).
900 */ 1023 */
901 for (i = 0; i < IWL_MAX_TID_COUNT; i++) { 1024 for (i = 0; i < IWL_MAX_TID_COUNT; i++) {
902 seq = mvm_ap_sta->tid_data[i].seq_number; 1025 u16 seq = mvm_ap_sta->tid_data[i].seq_number;
903 seq -= 0x10; 1026 seq -= 0x10;
904 wowlan_config_cmd.qos_seq[i] = cpu_to_le16(seq); 1027 wowlan_config_cmd.qos_seq[i] = cpu_to_le16(seq);
905 } 1028 }
@@ -945,6 +1068,16 @@ static int __iwl_mvm_suspend(struct ieee80211_hw *hw,
945 iwl_trans_stop_device(mvm->trans); 1068 iwl_trans_stop_device(mvm->trans);
946 1069
947 /* 1070 /*
1071 * The D3 firmware still hardcodes the AP station ID for the
1072 * BSS we're associated with as 0. Store the real STA ID here
1073 * and assign 0. When we leave this function, we'll restore
1074 * the original value for the resume code.
1075 */
1076 old_ap_sta_id = mvm_ap_sta->sta_id;
1077 mvm_ap_sta->sta_id = 0;
1078 mvmvif->ap_sta_id = 0;
1079
1080 /*
948 * Set the HW restart bit -- this is mostly true as we're 1081 * Set the HW restart bit -- this is mostly true as we're
949 * going to load new firmware and reprogram that, though 1082 * going to load new firmware and reprogram that, though
950 * the reprogramming is going to be manual to avoid adding 1083 * the reprogramming is going to be manual to avoid adding
@@ -1059,6 +1192,10 @@ static int __iwl_mvm_suspend(struct ieee80211_hw *hw,
1059 if (ret) 1192 if (ret)
1060 goto out; 1193 goto out;
1061 1194
1195 ret = iwl_mvm_power_update_device_mode(mvm);
1196 if (ret)
1197 goto out;
1198
1062 ret = iwl_mvm_power_update_mode(mvm, vif); 1199 ret = iwl_mvm_power_update_mode(mvm, vif);
1063 if (ret) 1200 if (ret)
1064 goto out; 1201 goto out;
@@ -1109,16 +1246,26 @@ int iwl_mvm_suspend(struct ieee80211_hw *hw, struct cfg80211_wowlan *wowlan)
1109 return __iwl_mvm_suspend(hw, wowlan, false); 1246 return __iwl_mvm_suspend(hw, wowlan, false);
1110} 1247}
1111 1248
1249/* converted data from the different status responses */
1250struct iwl_wowlan_status_data {
1251 u16 pattern_number;
1252 u16 qos_seq_ctr[8];
1253 u32 wakeup_reasons;
1254 u32 wake_packet_length;
1255 u32 wake_packet_bufsize;
1256 const u8 *wake_packet;
1257};
1258
1112static void iwl_mvm_report_wakeup_reasons(struct iwl_mvm *mvm, 1259static void iwl_mvm_report_wakeup_reasons(struct iwl_mvm *mvm,
1113 struct ieee80211_vif *vif, 1260 struct ieee80211_vif *vif,
1114 struct iwl_wowlan_status *status) 1261 struct iwl_wowlan_status_data *status)
1115{ 1262{
1116 struct sk_buff *pkt = NULL; 1263 struct sk_buff *pkt = NULL;
1117 struct cfg80211_wowlan_wakeup wakeup = { 1264 struct cfg80211_wowlan_wakeup wakeup = {
1118 .pattern_idx = -1, 1265 .pattern_idx = -1,
1119 }; 1266 };
1120 struct cfg80211_wowlan_wakeup *wakeup_report = &wakeup; 1267 struct cfg80211_wowlan_wakeup *wakeup_report = &wakeup;
1121 u32 reasons = le32_to_cpu(status->wakeup_reasons); 1268 u32 reasons = status->wakeup_reasons;
1122 1269
1123 if (reasons == IWL_WOWLAN_WAKEUP_BY_NON_WIRELESS) { 1270 if (reasons == IWL_WOWLAN_WAKEUP_BY_NON_WIRELESS) {
1124 wakeup_report = NULL; 1271 wakeup_report = NULL;
@@ -1130,7 +1277,7 @@ static void iwl_mvm_report_wakeup_reasons(struct iwl_mvm *mvm,
1130 1277
1131 if (reasons & IWL_WOWLAN_WAKEUP_BY_PATTERN) 1278 if (reasons & IWL_WOWLAN_WAKEUP_BY_PATTERN)
1132 wakeup.pattern_idx = 1279 wakeup.pattern_idx =
1133 le16_to_cpu(status->pattern_number); 1280 status->pattern_number;
1134 1281
1135 if (reasons & (IWL_WOWLAN_WAKEUP_BY_DISCONNECTION_ON_MISSED_BEACON | 1282 if (reasons & (IWL_WOWLAN_WAKEUP_BY_DISCONNECTION_ON_MISSED_BEACON |
1136 IWL_WOWLAN_WAKEUP_BY_DISCONNECTION_ON_DEAUTH)) 1283 IWL_WOWLAN_WAKEUP_BY_DISCONNECTION_ON_DEAUTH))
@@ -1158,8 +1305,8 @@ static void iwl_mvm_report_wakeup_reasons(struct iwl_mvm *mvm,
1158 wakeup.tcp_match = true; 1305 wakeup.tcp_match = true;
1159 1306
1160 if (status->wake_packet_bufsize) { 1307 if (status->wake_packet_bufsize) {
1161 int pktsize = le32_to_cpu(status->wake_packet_bufsize); 1308 int pktsize = status->wake_packet_bufsize;
1162 int pktlen = le32_to_cpu(status->wake_packet_length); 1309 int pktlen = status->wake_packet_length;
1163 const u8 *pktdata = status->wake_packet; 1310 const u8 *pktdata = status->wake_packet;
1164 struct ieee80211_hdr *hdr = (void *)pktdata; 1311 struct ieee80211_hdr *hdr = (void *)pktdata;
1165 int truncated = pktlen - pktsize; 1312 int truncated = pktlen - pktsize;
@@ -1239,8 +1386,229 @@ static void iwl_mvm_report_wakeup_reasons(struct iwl_mvm *mvm,
1239 kfree_skb(pkt); 1386 kfree_skb(pkt);
1240} 1387}
1241 1388
1389static void iwl_mvm_aes_sc_to_seq(struct aes_sc *sc,
1390 struct ieee80211_key_seq *seq)
1391{
1392 u64 pn;
1393
1394 pn = le64_to_cpu(sc->pn);
1395 seq->ccmp.pn[0] = pn >> 40;
1396 seq->ccmp.pn[1] = pn >> 32;
1397 seq->ccmp.pn[2] = pn >> 24;
1398 seq->ccmp.pn[3] = pn >> 16;
1399 seq->ccmp.pn[4] = pn >> 8;
1400 seq->ccmp.pn[5] = pn;
1401}
1402
1403static void iwl_mvm_tkip_sc_to_seq(struct tkip_sc *sc,
1404 struct ieee80211_key_seq *seq)
1405{
1406 seq->tkip.iv32 = le32_to_cpu(sc->iv32);
1407 seq->tkip.iv16 = le16_to_cpu(sc->iv16);
1408}
1409
1410static void iwl_mvm_set_aes_rx_seq(struct aes_sc *scs,
1411 struct ieee80211_key_conf *key)
1412{
1413 int tid;
1414
1415 BUILD_BUG_ON(IWL_NUM_RSC != IEEE80211_NUM_TIDS);
1416
1417 for (tid = 0; tid < IWL_NUM_RSC; tid++) {
1418 struct ieee80211_key_seq seq = {};
1419
1420 iwl_mvm_aes_sc_to_seq(&scs[tid], &seq);
1421 ieee80211_set_key_rx_seq(key, tid, &seq);
1422 }
1423}
1424
1425static void iwl_mvm_set_tkip_rx_seq(struct tkip_sc *scs,
1426 struct ieee80211_key_conf *key)
1427{
1428 int tid;
1429
1430 BUILD_BUG_ON(IWL_NUM_RSC != IEEE80211_NUM_TIDS);
1431
1432 for (tid = 0; tid < IWL_NUM_RSC; tid++) {
1433 struct ieee80211_key_seq seq = {};
1434
1435 iwl_mvm_tkip_sc_to_seq(&scs[tid], &seq);
1436 ieee80211_set_key_rx_seq(key, tid, &seq);
1437 }
1438}
1439
1440static void iwl_mvm_set_key_rx_seq(struct ieee80211_key_conf *key,
1441 struct iwl_wowlan_status_v6 *status)
1442{
1443 union iwl_all_tsc_rsc *rsc = &status->gtk.rsc.all_tsc_rsc;
1444
1445 switch (key->cipher) {
1446 case WLAN_CIPHER_SUITE_CCMP:
1447 iwl_mvm_set_aes_rx_seq(rsc->aes.multicast_rsc, key);
1448 break;
1449 case WLAN_CIPHER_SUITE_TKIP:
1450 iwl_mvm_set_tkip_rx_seq(rsc->tkip.multicast_rsc, key);
1451 break;
1452 default:
1453 WARN_ON(1);
1454 }
1455}
1456
1457struct iwl_mvm_d3_gtk_iter_data {
1458 struct iwl_wowlan_status_v6 *status;
1459 void *last_gtk;
1460 u32 cipher;
1461 bool find_phase, unhandled_cipher;
1462 int num_keys;
1463};
1464
1465static void iwl_mvm_d3_update_gtks(struct ieee80211_hw *hw,
1466 struct ieee80211_vif *vif,
1467 struct ieee80211_sta *sta,
1468 struct ieee80211_key_conf *key,
1469 void *_data)
1470{
1471 struct iwl_mvm_d3_gtk_iter_data *data = _data;
1472
1473 if (data->unhandled_cipher)
1474 return;
1475
1476 switch (key->cipher) {
1477 case WLAN_CIPHER_SUITE_WEP40:
1478 case WLAN_CIPHER_SUITE_WEP104:
1479 /* ignore WEP completely, nothing to do */
1480 return;
1481 case WLAN_CIPHER_SUITE_CCMP:
1482 case WLAN_CIPHER_SUITE_TKIP:
1483 /* we support these */
1484 break;
1485 default:
1486 /* everything else (even CMAC for MFP) - disconnect from AP */
1487 data->unhandled_cipher = true;
1488 return;
1489 }
1490
1491 data->num_keys++;
1492
1493 /*
1494 * pairwise key - update sequence counters only;
1495 * note that this assumes no TDLS sessions are active
1496 */
1497 if (sta) {
1498 struct ieee80211_key_seq seq = {};
1499 union iwl_all_tsc_rsc *sc = &data->status->gtk.rsc.all_tsc_rsc;
1500
1501 if (data->find_phase)
1502 return;
1503
1504 switch (key->cipher) {
1505 case WLAN_CIPHER_SUITE_CCMP:
1506 iwl_mvm_aes_sc_to_seq(&sc->aes.tsc, &seq);
1507 iwl_mvm_set_aes_rx_seq(sc->aes.unicast_rsc, key);
1508 break;
1509 case WLAN_CIPHER_SUITE_TKIP:
1510 iwl_mvm_tkip_sc_to_seq(&sc->tkip.tsc, &seq);
1511 iwl_mvm_set_tkip_rx_seq(sc->tkip.unicast_rsc, key);
1512 break;
1513 }
1514 ieee80211_set_key_tx_seq(key, &seq);
1515
1516 /* that's it for this key */
1517 return;
1518 }
1519
1520 if (data->find_phase) {
1521 data->last_gtk = key;
1522 data->cipher = key->cipher;
1523 return;
1524 }
1525
1526 if (data->status->num_of_gtk_rekeys)
1527 ieee80211_remove_key(key);
1528 else if (data->last_gtk == key)
1529 iwl_mvm_set_key_rx_seq(key, data->status);
1530}
1531
1532static bool iwl_mvm_setup_connection_keep(struct iwl_mvm *mvm,
1533 struct ieee80211_vif *vif,
1534 struct iwl_wowlan_status_v6 *status)
1535{
1536 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif);
1537 struct iwl_mvm_d3_gtk_iter_data gtkdata = {
1538 .status = status,
1539 };
1540
1541 if (!status || !vif->bss_conf.bssid)
1542 return false;
1543
1544 /* find last GTK that we used initially, if any */
1545 gtkdata.find_phase = true;
1546 ieee80211_iter_keys(mvm->hw, vif,
1547 iwl_mvm_d3_update_gtks, &gtkdata);
1548 /* not trying to keep connections with MFP/unhandled ciphers */
1549 if (gtkdata.unhandled_cipher)
1550 return false;
1551 if (!gtkdata.num_keys)
1552 return true;
1553 if (!gtkdata.last_gtk)
1554 return false;
1555
1556 /*
1557 * invalidate all other GTKs that might still exist and update
1558 * the one that we used
1559 */
1560 gtkdata.find_phase = false;
1561 ieee80211_iter_keys(mvm->hw, vif,
1562 iwl_mvm_d3_update_gtks, &gtkdata);
1563
1564 if (status->num_of_gtk_rekeys) {
1565 struct ieee80211_key_conf *key;
1566 struct {
1567 struct ieee80211_key_conf conf;
1568 u8 key[32];
1569 } conf = {
1570 .conf.cipher = gtkdata.cipher,
1571 .conf.keyidx = status->gtk.key_index,
1572 };
1573
1574 switch (gtkdata.cipher) {
1575 case WLAN_CIPHER_SUITE_CCMP:
1576 conf.conf.keylen = WLAN_KEY_LEN_CCMP;
1577 memcpy(conf.conf.key, status->gtk.decrypt_key,
1578 WLAN_KEY_LEN_CCMP);
1579 break;
1580 case WLAN_CIPHER_SUITE_TKIP:
1581 conf.conf.keylen = WLAN_KEY_LEN_TKIP;
1582 memcpy(conf.conf.key, status->gtk.decrypt_key, 16);
1583 /* leave TX MIC key zeroed, we don't use it anyway */
1584 memcpy(conf.conf.key +
1585 NL80211_TKIP_DATA_OFFSET_RX_MIC_KEY,
1586 status->gtk.tkip_mic_key, 8);
1587 break;
1588 }
1589
1590 key = ieee80211_gtk_rekey_add(vif, &conf.conf);
1591 if (IS_ERR(key))
1592 return false;
1593 iwl_mvm_set_key_rx_seq(key, status);
1594 }
1595
1596 if (status->num_of_gtk_rekeys) {
1597 __be64 replay_ctr =
1598 cpu_to_be64(le64_to_cpu(status->replay_ctr));
1599 ieee80211_gtk_rekey_notify(vif, vif->bss_conf.bssid,
1600 (void *)&replay_ctr, GFP_KERNEL);
1601 }
1602
1603 mvmvif->seqno_valid = true;
1604 /* +0x10 because the set API expects next-to-use, not last-used */
1605 mvmvif->seqno = le16_to_cpu(status->non_qos_seq_ctr) + 0x10;
1606
1607 return true;
1608}
1609
1242/* releases the MVM mutex */ 1610/* releases the MVM mutex */
1243static void iwl_mvm_query_wakeup_reasons(struct iwl_mvm *mvm, 1611static bool iwl_mvm_query_wakeup_reasons(struct iwl_mvm *mvm,
1244 struct ieee80211_vif *vif) 1612 struct ieee80211_vif *vif)
1245{ 1613{
1246 u32 base = mvm->error_event_table; 1614 u32 base = mvm->error_event_table;
@@ -1253,8 +1621,12 @@ static void iwl_mvm_query_wakeup_reasons(struct iwl_mvm *mvm,
1253 .id = WOWLAN_GET_STATUSES, 1621 .id = WOWLAN_GET_STATUSES,
1254 .flags = CMD_SYNC | CMD_WANT_SKB, 1622 .flags = CMD_SYNC | CMD_WANT_SKB,
1255 }; 1623 };
1256 struct iwl_wowlan_status *status; 1624 struct iwl_wowlan_status_data status;
1257 int ret, len; 1625 struct iwl_wowlan_status_v6 *status_v6;
1626 int ret, len, status_size, i;
1627 bool keep;
1628 struct ieee80211_sta *ap_sta;
1629 struct iwl_mvm_sta *mvm_ap_sta;
1258 1630
1259 iwl_trans_read_mem_bytes(mvm->trans, base, 1631 iwl_trans_read_mem_bytes(mvm->trans, base,
1260 &err_info, sizeof(err_info)); 1632 &err_info, sizeof(err_info));
@@ -1287,32 +1659,83 @@ static void iwl_mvm_query_wakeup_reasons(struct iwl_mvm *mvm,
1287 if (!cmd.resp_pkt) 1659 if (!cmd.resp_pkt)
1288 goto out_unlock; 1660 goto out_unlock;
1289 1661
1662 if (mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_D3_CONTINUITY_API)
1663 status_size = sizeof(struct iwl_wowlan_status_v6);
1664 else
1665 status_size = sizeof(struct iwl_wowlan_status_v4);
1666
1290 len = le32_to_cpu(cmd.resp_pkt->len_n_flags) & FH_RSCSR_FRAME_SIZE_MSK; 1667 len = le32_to_cpu(cmd.resp_pkt->len_n_flags) & FH_RSCSR_FRAME_SIZE_MSK;
1291 if (len - sizeof(struct iwl_cmd_header) < sizeof(*status)) { 1668 if (len - sizeof(struct iwl_cmd_header) < status_size) {
1292 IWL_ERR(mvm, "Invalid WoWLAN status response!\n"); 1669 IWL_ERR(mvm, "Invalid WoWLAN status response!\n");
1293 goto out_free_resp; 1670 goto out_free_resp;
1294 } 1671 }
1295 1672
1296 status = (void *)cmd.resp_pkt->data; 1673 if (mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_D3_CONTINUITY_API) {
1674 status_v6 = (void *)cmd.resp_pkt->data;
1675
1676 status.pattern_number = le16_to_cpu(status_v6->pattern_number);
1677 for (i = 0; i < 8; i++)
1678 status.qos_seq_ctr[i] =
1679 le16_to_cpu(status_v6->qos_seq_ctr[i]);
1680 status.wakeup_reasons = le32_to_cpu(status_v6->wakeup_reasons);
1681 status.wake_packet_length =
1682 le32_to_cpu(status_v6->wake_packet_length);
1683 status.wake_packet_bufsize =
1684 le32_to_cpu(status_v6->wake_packet_bufsize);
1685 status.wake_packet = status_v6->wake_packet;
1686 } else {
1687 struct iwl_wowlan_status_v4 *status_v4;
1688 status_v6 = NULL;
1689 status_v4 = (void *)cmd.resp_pkt->data;
1690
1691 status.pattern_number = le16_to_cpu(status_v4->pattern_number);
1692 for (i = 0; i < 8; i++)
1693 status.qos_seq_ctr[i] =
1694 le16_to_cpu(status_v4->qos_seq_ctr[i]);
1695 status.wakeup_reasons = le32_to_cpu(status_v4->wakeup_reasons);
1696 status.wake_packet_length =
1697 le32_to_cpu(status_v4->wake_packet_length);
1698 status.wake_packet_bufsize =
1699 le32_to_cpu(status_v4->wake_packet_bufsize);
1700 status.wake_packet = status_v4->wake_packet;
1701 }
1297 1702
1298 if (len - sizeof(struct iwl_cmd_header) != 1703 if (len - sizeof(struct iwl_cmd_header) !=
1299 sizeof(*status) + 1704 status_size + ALIGN(status.wake_packet_bufsize, 4)) {
1300 ALIGN(le32_to_cpu(status->wake_packet_bufsize), 4)) {
1301 IWL_ERR(mvm, "Invalid WoWLAN status response!\n"); 1705 IWL_ERR(mvm, "Invalid WoWLAN status response!\n");
1302 goto out_free_resp; 1706 goto out_free_resp;
1303 } 1707 }
1304 1708
1709 /* still at hard-coded place 0 for D3 image */
1710 ap_sta = rcu_dereference_protected(
1711 mvm->fw_id_to_mac_id[0],
1712 lockdep_is_held(&mvm->mutex));
1713 if (IS_ERR_OR_NULL(ap_sta))
1714 goto out_free_resp;
1715
1716 mvm_ap_sta = (struct iwl_mvm_sta *)ap_sta->drv_priv;
1717 for (i = 0; i < IWL_MAX_TID_COUNT; i++) {
1718 u16 seq = status.qos_seq_ctr[i];
1719 /* firmware stores last-used value, we store next value */
1720 seq += 0x10;
1721 mvm_ap_sta->tid_data[i].seq_number = seq;
1722 }
1723
1305 /* now we have all the data we need, unlock to avoid mac80211 issues */ 1724 /* now we have all the data we need, unlock to avoid mac80211 issues */
1306 mutex_unlock(&mvm->mutex); 1725 mutex_unlock(&mvm->mutex);
1307 1726
1308 iwl_mvm_report_wakeup_reasons(mvm, vif, status); 1727 iwl_mvm_report_wakeup_reasons(mvm, vif, &status);
1728
1729 keep = iwl_mvm_setup_connection_keep(mvm, vif, status_v6);
1730
1309 iwl_free_resp(&cmd); 1731 iwl_free_resp(&cmd);
1310 return; 1732 return keep;
1311 1733
1312 out_free_resp: 1734 out_free_resp:
1313 iwl_free_resp(&cmd); 1735 iwl_free_resp(&cmd);
1314 out_unlock: 1736 out_unlock:
1315 mutex_unlock(&mvm->mutex); 1737 mutex_unlock(&mvm->mutex);
1738 return false;
1316} 1739}
1317 1740
1318static void iwl_mvm_read_d3_sram(struct iwl_mvm *mvm) 1741static void iwl_mvm_read_d3_sram(struct iwl_mvm *mvm)
@@ -1335,6 +1758,17 @@ static void iwl_mvm_read_d3_sram(struct iwl_mvm *mvm)
1335#endif 1758#endif
1336} 1759}
1337 1760
1761static void iwl_mvm_d3_disconnect_iter(void *data, u8 *mac,
1762 struct ieee80211_vif *vif)
1763{
1764 /* skip the one we keep connection on */
1765 if (data == vif)
1766 return;
1767
1768 if (vif->type == NL80211_IFTYPE_STATION)
1769 ieee80211_resume_disconnect(vif);
1770}
1771
1338static int __iwl_mvm_resume(struct iwl_mvm *mvm, bool test) 1772static int __iwl_mvm_resume(struct iwl_mvm *mvm, bool test)
1339{ 1773{
1340 struct iwl_d3_iter_data resume_iter_data = { 1774 struct iwl_d3_iter_data resume_iter_data = {
@@ -1343,6 +1777,7 @@ static int __iwl_mvm_resume(struct iwl_mvm *mvm, bool test)
1343 struct ieee80211_vif *vif = NULL; 1777 struct ieee80211_vif *vif = NULL;
1344 int ret; 1778 int ret;
1345 enum iwl_d3_status d3_status; 1779 enum iwl_d3_status d3_status;
1780 bool keep = false;
1346 1781
1347 mutex_lock(&mvm->mutex); 1782 mutex_lock(&mvm->mutex);
1348 1783
@@ -1368,7 +1803,7 @@ static int __iwl_mvm_resume(struct iwl_mvm *mvm, bool test)
1368 /* query SRAM first in case we want event logging */ 1803 /* query SRAM first in case we want event logging */
1369 iwl_mvm_read_d3_sram(mvm); 1804 iwl_mvm_read_d3_sram(mvm);
1370 1805
1371 iwl_mvm_query_wakeup_reasons(mvm, vif); 1806 keep = iwl_mvm_query_wakeup_reasons(mvm, vif);
1372 /* has unlocked the mutex, so skip that */ 1807 /* has unlocked the mutex, so skip that */
1373 goto out; 1808 goto out;
1374 1809
@@ -1376,8 +1811,10 @@ static int __iwl_mvm_resume(struct iwl_mvm *mvm, bool test)
1376 mutex_unlock(&mvm->mutex); 1811 mutex_unlock(&mvm->mutex);
1377 1812
1378 out: 1813 out:
1379 if (!test && vif) 1814 if (!test)
1380 ieee80211_resume_disconnect(vif); 1815 ieee80211_iterate_active_interfaces_rtnl(mvm->hw,
1816 IEEE80211_IFACE_ITER_NORMAL,
1817 iwl_mvm_d3_disconnect_iter, keep ? vif : NULL);
1381 1818
1382 /* return 1 to reconfigure the device */ 1819 /* return 1 to reconfigure the device */
1383 set_bit(IWL_MVM_STATUS_IN_HW_RESTART, &mvm->status); 1820 set_bit(IWL_MVM_STATUS_IN_HW_RESTART, &mvm->status);
diff --git a/drivers/net/wireless/iwlwifi/mvm/debugfs.c b/drivers/net/wireless/iwlwifi/mvm/debugfs.c
index aac81b8984b0..9864d713eb2c 100644
--- a/drivers/net/wireless/iwlwifi/mvm/debugfs.c
+++ b/drivers/net/wireless/iwlwifi/mvm/debugfs.c
@@ -246,58 +246,56 @@ static ssize_t iwl_dbgfs_stations_read(struct file *file, char __user *user_buf,
246 return simple_read_from_buffer(user_buf, count, ppos, buf, pos); 246 return simple_read_from_buffer(user_buf, count, ppos, buf, pos);
247} 247}
248 248
249static ssize_t iwl_dbgfs_power_down_allow_write(struct file *file, 249static ssize_t iwl_dbgfs_disable_power_off_read(struct file *file,
250 const char __user *user_buf, 250 char __user *user_buf,
251 size_t count, loff_t *ppos) 251 size_t count, loff_t *ppos)
252{ 252{
253 struct iwl_mvm *mvm = file->private_data; 253 struct iwl_mvm *mvm = file->private_data;
254 char buf[8] = {}; 254 char buf[64];
255 int allow; 255 int bufsz = sizeof(buf);
256 256 int pos = 0;
257 if (!mvm->ucode_loaded)
258 return -EIO;
259
260 if (copy_from_user(buf, user_buf, sizeof(buf)))
261 return -EFAULT;
262
263 if (sscanf(buf, "%d", &allow) != 1)
264 return -EINVAL;
265
266 IWL_DEBUG_POWER(mvm, "%s device power down\n",
267 allow ? "allow" : "prevent");
268 257
269 /* 258 pos += scnprintf(buf+pos, bufsz-pos, "disable_power_off_d0=%d\n",
270 * TODO: Send REPLY_DEBUG_CMD (0xf0) when FW support it 259 mvm->disable_power_off);
271 */ 260 pos += scnprintf(buf+pos, bufsz-pos, "disable_power_off_d3=%d\n",
261 mvm->disable_power_off_d3);
272 262
273 return count; 263 return simple_read_from_buffer(user_buf, count, ppos, buf, pos);
274} 264}
275 265
276static ssize_t iwl_dbgfs_power_down_d3_allow_write(struct file *file, 266static ssize_t iwl_dbgfs_disable_power_off_write(struct file *file,
277 const char __user *user_buf, 267 const char __user *user_buf,
278 size_t count, loff_t *ppos) 268 size_t count, loff_t *ppos)
279{ 269{
280 struct iwl_mvm *mvm = file->private_data; 270 struct iwl_mvm *mvm = file->private_data;
281 char buf[8] = {}; 271 char buf[64] = {};
282 int allow; 272 int ret;
273 int val;
283 274
284 if (copy_from_user(buf, user_buf, sizeof(buf))) 275 if (!mvm->ucode_loaded)
276 return -EIO;
277
278 count = min_t(size_t, count, sizeof(buf) - 1);
279 if (copy_from_user(buf, user_buf, count))
285 return -EFAULT; 280 return -EFAULT;
286 281
287 if (sscanf(buf, "%d", &allow) != 1) 282 if (!strncmp("disable_power_off_d0=", buf, 21)) {
283 if (sscanf(buf + 21, "%d", &val) != 1)
284 return -EINVAL;
285 mvm->disable_power_off = val;
286 } else if (!strncmp("disable_power_off_d3=", buf, 21)) {
287 if (sscanf(buf + 21, "%d", &val) != 1)
288 return -EINVAL;
289 mvm->disable_power_off_d3 = val;
290 } else {
288 return -EINVAL; 291 return -EINVAL;
292 }
289 293
290 IWL_DEBUG_POWER(mvm, "%s device power down in d3\n", 294 mutex_lock(&mvm->mutex);
291 allow ? "allow" : "prevent"); 295 ret = iwl_mvm_power_update_device_mode(mvm);
292 296 mutex_unlock(&mvm->mutex);
293 /*
294 * TODO: When WoWLAN FW alive notification happens, driver will send
295 * REPLY_DEBUG_CMD setting power_down_allow flag according to
296 * mvm->prevent_power_down_d3
297 */
298 mvm->prevent_power_down_d3 = !allow;
299 297
300 return count; 298 return ret ?: count;
301} 299}
302 300
303static void iwl_dbgfs_update_pm(struct iwl_mvm *mvm, 301static void iwl_dbgfs_update_pm(struct iwl_mvm *mvm,
@@ -344,6 +342,7 @@ static void iwl_dbgfs_update_pm(struct iwl_mvm *mvm,
344 case MVM_DEBUGFS_PM_DISABLE_POWER_OFF: 342 case MVM_DEBUGFS_PM_DISABLE_POWER_OFF:
345 IWL_DEBUG_POWER(mvm, "disable_power_off=%d\n", val); 343 IWL_DEBUG_POWER(mvm, "disable_power_off=%d\n", val);
346 dbgfs_pm->disable_power_off = val; 344 dbgfs_pm->disable_power_off = val;
345 break;
347 case MVM_DEBUGFS_PM_LPRX_ENA: 346 case MVM_DEBUGFS_PM_LPRX_ENA:
348 IWL_DEBUG_POWER(mvm, "lprx %s\n", val ? "enabled" : "disabled"); 347 IWL_DEBUG_POWER(mvm, "lprx %s\n", val ? "enabled" : "disabled");
349 dbgfs_pm->lprx_ena = val; 348 dbgfs_pm->lprx_ena = val;
@@ -371,7 +370,8 @@ static ssize_t iwl_dbgfs_pm_params_write(struct file *file,
371 int val; 370 int val;
372 int ret; 371 int ret;
373 372
374 if (copy_from_user(buf, user_buf, sizeof(buf))) 373 count = min_t(size_t, count, sizeof(buf) - 1);
374 if (copy_from_user(buf, user_buf, count))
375 return -EFAULT; 375 return -EFAULT;
376 376
377 if (!strncmp("keep_alive=", buf, 11)) { 377 if (!strncmp("keep_alive=", buf, 11)) {
@@ -394,7 +394,9 @@ static ssize_t iwl_dbgfs_pm_params_write(struct file *file,
394 if (sscanf(buf + 16, "%d", &val) != 1) 394 if (sscanf(buf + 16, "%d", &val) != 1)
395 return -EINVAL; 395 return -EINVAL;
396 param = MVM_DEBUGFS_PM_TX_DATA_TIMEOUT; 396 param = MVM_DEBUGFS_PM_TX_DATA_TIMEOUT;
397 } else if (!strncmp("disable_power_off=", buf, 18)) { 397 } else if (!strncmp("disable_power_off=", buf, 18) &&
398 !(mvm->fw->ucode_capa.flags &
399 IWL_UCODE_TLV_FLAGS_DEVICE_PS_CMD)) {
398 if (sscanf(buf + 18, "%d", &val) != 1) 400 if (sscanf(buf + 18, "%d", &val) != 1)
399 return -EINVAL; 401 return -EINVAL;
400 param = MVM_DEBUGFS_PM_DISABLE_POWER_OFF; 402 param = MVM_DEBUGFS_PM_DISABLE_POWER_OFF;
@@ -581,15 +583,21 @@ static ssize_t iwl_dbgfs_bt_notif_read(struct file *file, char __user *user_buf,
581 BT_MBOX_PRINT(3, UPDATE_REQUEST, true); 583 BT_MBOX_PRINT(3, UPDATE_REQUEST, true);
582 584
583 pos += scnprintf(buf+pos, bufsz-pos, "bt_status = %d\n", 585 pos += scnprintf(buf+pos, bufsz-pos, "bt_status = %d\n",
584 notif->bt_status); 586 notif->bt_status);
585 pos += scnprintf(buf+pos, bufsz-pos, "bt_open_conn = %d\n", 587 pos += scnprintf(buf+pos, bufsz-pos, "bt_open_conn = %d\n",
586 notif->bt_open_conn); 588 notif->bt_open_conn);
587 pos += scnprintf(buf+pos, bufsz-pos, "bt_traffic_load = %d\n", 589 pos += scnprintf(buf+pos, bufsz-pos, "bt_traffic_load = %d\n",
588 notif->bt_traffic_load); 590 notif->bt_traffic_load);
589 pos += scnprintf(buf+pos, bufsz-pos, "bt_agg_traffic_load = %d\n", 591 pos += scnprintf(buf+pos, bufsz-pos, "bt_agg_traffic_load = %d\n",
590 notif->bt_agg_traffic_load); 592 notif->bt_agg_traffic_load);
591 pos += scnprintf(buf+pos, bufsz-pos, "bt_ci_compliance = %d\n", 593 pos += scnprintf(buf+pos, bufsz-pos, "bt_ci_compliance = %d\n",
592 notif->bt_ci_compliance); 594 notif->bt_ci_compliance);
595 pos += scnprintf(buf+pos, bufsz-pos, "primary_ch_lut = %d\n",
596 le32_to_cpu(notif->primary_ch_lut));
597 pos += scnprintf(buf+pos, bufsz-pos, "secondary_ch_lut = %d\n",
598 le32_to_cpu(notif->secondary_ch_lut));
599 pos += scnprintf(buf+pos, bufsz-pos, "bt_activity_grading = %d\n",
600 le32_to_cpu(notif->bt_activity_grading));
593 601
594 mutex_unlock(&mvm->mutex); 602 mutex_unlock(&mvm->mutex);
595 603
@@ -600,6 +608,38 @@ static ssize_t iwl_dbgfs_bt_notif_read(struct file *file, char __user *user_buf,
600} 608}
601#undef BT_MBOX_PRINT 609#undef BT_MBOX_PRINT
602 610
611static ssize_t iwl_dbgfs_bt_cmd_read(struct file *file, char __user *user_buf,
612 size_t count, loff_t *ppos)
613{
614 struct iwl_mvm *mvm = file->private_data;
615 struct iwl_bt_coex_ci_cmd *cmd = &mvm->last_bt_ci_cmd;
616 char buf[256];
617 int bufsz = sizeof(buf);
618 int pos = 0;
619
620 mutex_lock(&mvm->mutex);
621
622 pos += scnprintf(buf+pos, bufsz-pos, "Channel inhibition CMD\n");
623 pos += scnprintf(buf+pos, bufsz-pos,
624 "\tPrimary Channel Bitmap 0x%016llx Fat: %d\n",
625 le64_to_cpu(cmd->bt_primary_ci),
626 !!cmd->co_run_bw_primary);
627 pos += scnprintf(buf+pos, bufsz-pos,
628 "\tSecondary Channel Bitmap 0x%016llx Fat: %d\n",
629 le64_to_cpu(cmd->bt_secondary_ci),
630 !!cmd->co_run_bw_secondary);
631
632 pos += scnprintf(buf+pos, bufsz-pos, "BT Configuration CMD\n");
633 pos += scnprintf(buf+pos, bufsz-pos, "\tACK Kill Mask 0x%08x\n",
634 iwl_bt_ack_kill_msk[mvm->bt_kill_msk]);
635 pos += scnprintf(buf+pos, bufsz-pos, "\tCTS Kill Mask 0x%08x\n",
636 iwl_bt_cts_kill_msk[mvm->bt_kill_msk]);
637
638 mutex_unlock(&mvm->mutex);
639
640 return simple_read_from_buffer(user_buf, count, ppos, buf, pos);
641}
642
603#define PRINT_STATS_LE32(_str, _val) \ 643#define PRINT_STATS_LE32(_str, _val) \
604 pos += scnprintf(buf + pos, bufsz - pos, \ 644 pos += scnprintf(buf + pos, bufsz - pos, \
605 fmt_table, _str, \ 645 fmt_table, _str, \
@@ -615,9 +655,11 @@ static ssize_t iwl_dbgfs_fw_rx_stats_read(struct file *file,
615 int pos = 0; 655 int pos = 0;
616 char *buf; 656 char *buf;
617 int ret; 657 int ret;
618 int bufsz = sizeof(struct mvm_statistics_rx_phy) * 20 + 658 /* 43 is the size of each data line, 33 is the size of each header */
619 sizeof(struct mvm_statistics_rx_non_phy) * 10 + 659 size_t bufsz =
620 sizeof(struct mvm_statistics_rx_ht_phy) * 10 + 200; 660 ((sizeof(struct mvm_statistics_rx) / sizeof(__le32)) * 43) +
661 (4 * 33) + 1;
662
621 struct mvm_statistics_rx_phy *ofdm; 663 struct mvm_statistics_rx_phy *ofdm;
622 struct mvm_statistics_rx_phy *cck; 664 struct mvm_statistics_rx_phy *cck;
623 struct mvm_statistics_rx_non_phy *general; 665 struct mvm_statistics_rx_non_phy *general;
@@ -712,6 +754,7 @@ static ssize_t iwl_dbgfs_fw_rx_stats_read(struct file *file,
712 PRINT_STATS_LE32("beacon_energy_b", general->beacon_energy_b); 754 PRINT_STATS_LE32("beacon_energy_b", general->beacon_energy_b);
713 PRINT_STATS_LE32("beacon_energy_c", general->beacon_energy_c); 755 PRINT_STATS_LE32("beacon_energy_c", general->beacon_energy_c);
714 PRINT_STATS_LE32("num_bt_kills", general->num_bt_kills); 756 PRINT_STATS_LE32("num_bt_kills", general->num_bt_kills);
757 PRINT_STATS_LE32("mac_id", general->mac_id);
715 PRINT_STATS_LE32("directed_data_mpdu", general->directed_data_mpdu); 758 PRINT_STATS_LE32("directed_data_mpdu", general->directed_data_mpdu);
716 759
717 pos += scnprintf(buf + pos, bufsz - pos, fmt_header, 760 pos += scnprintf(buf + pos, bufsz - pos, fmt_header,
@@ -757,6 +800,59 @@ static ssize_t iwl_dbgfs_fw_restart_write(struct file *file,
757 return count; 800 return count;
758} 801}
759 802
803static ssize_t
804iwl_dbgfs_scan_ant_rxchain_read(struct file *file,
805 char __user *user_buf,
806 size_t count, loff_t *ppos)
807{
808 struct iwl_mvm *mvm = file->private_data;
809 int pos = 0;
810 char buf[32];
811 const size_t bufsz = sizeof(buf);
812
813 /* print which antennas were set for the scan command by the user */
814 pos += scnprintf(buf + pos, bufsz - pos, "Antennas for scan: ");
815 if (mvm->scan_rx_ant & ANT_A)
816 pos += scnprintf(buf + pos, bufsz - pos, "A");
817 if (mvm->scan_rx_ant & ANT_B)
818 pos += scnprintf(buf + pos, bufsz - pos, "B");
819 if (mvm->scan_rx_ant & ANT_C)
820 pos += scnprintf(buf + pos, bufsz - pos, "C");
821 pos += scnprintf(buf + pos, bufsz - pos, " (%hhx)\n", mvm->scan_rx_ant);
822
823 return simple_read_from_buffer(user_buf, count, ppos, buf, pos);
824}
825
826static ssize_t
827iwl_dbgfs_scan_ant_rxchain_write(struct file *file,
828 const char __user *user_buf,
829 size_t count, loff_t *ppos)
830{
831 struct iwl_mvm *mvm = file->private_data;
832 char buf[8];
833 int buf_size;
834 u8 scan_rx_ant;
835
836 memset(buf, 0, sizeof(buf));
837 buf_size = min(count, sizeof(buf) - 1);
838
839 /* get the argument from the user and check if it is valid */
840 if (copy_from_user(buf, user_buf, buf_size))
841 return -EFAULT;
842 if (sscanf(buf, "%hhx", &scan_rx_ant) != 1)
843 return -EINVAL;
844 if (scan_rx_ant > ANT_ABC)
845 return -EINVAL;
846 if (scan_rx_ant & ~iwl_fw_valid_rx_ant(mvm->fw))
847 return -EINVAL;
848
849 /* change the rx antennas for scan command */
850 mvm->scan_rx_ant = scan_rx_ant;
851
852 return count;
853}
854
855
760static void iwl_dbgfs_update_bf(struct ieee80211_vif *vif, 856static void iwl_dbgfs_update_bf(struct ieee80211_vif *vif,
761 enum iwl_dbgfs_bf_mask param, int value) 857 enum iwl_dbgfs_bf_mask param, int value)
762{ 858{
@@ -968,7 +1064,8 @@ static ssize_t iwl_dbgfs_d3_sram_write(struct file *file,
968 char buf[8] = {}; 1064 char buf[8] = {};
969 int store; 1065 int store;
970 1066
971 if (copy_from_user(buf, user_buf, sizeof(buf))) 1067 count = min_t(size_t, count, sizeof(buf) - 1);
1068 if (copy_from_user(buf, user_buf, count))
972 return -EFAULT; 1069 return -EFAULT;
973 1070
974 if (sscanf(buf, "%d", &store) != 1) 1071 if (sscanf(buf, "%d", &store) != 1)
@@ -1063,10 +1160,12 @@ MVM_DEBUGFS_WRITE_FILE_OPS(sta_drain);
1063MVM_DEBUGFS_READ_WRITE_FILE_OPS(sram); 1160MVM_DEBUGFS_READ_WRITE_FILE_OPS(sram);
1064MVM_DEBUGFS_READ_FILE_OPS(stations); 1161MVM_DEBUGFS_READ_FILE_OPS(stations);
1065MVM_DEBUGFS_READ_FILE_OPS(bt_notif); 1162MVM_DEBUGFS_READ_FILE_OPS(bt_notif);
1066MVM_DEBUGFS_WRITE_FILE_OPS(power_down_allow); 1163MVM_DEBUGFS_READ_FILE_OPS(bt_cmd);
1067MVM_DEBUGFS_WRITE_FILE_OPS(power_down_d3_allow); 1164MVM_DEBUGFS_READ_WRITE_FILE_OPS(disable_power_off);
1068MVM_DEBUGFS_READ_FILE_OPS(fw_rx_stats); 1165MVM_DEBUGFS_READ_FILE_OPS(fw_rx_stats);
1069MVM_DEBUGFS_WRITE_FILE_OPS(fw_restart); 1166MVM_DEBUGFS_WRITE_FILE_OPS(fw_restart);
1167MVM_DEBUGFS_READ_WRITE_FILE_OPS(scan_ant_rxchain);
1168
1070#ifdef CONFIG_PM_SLEEP 1169#ifdef CONFIG_PM_SLEEP
1071MVM_DEBUGFS_READ_WRITE_FILE_OPS(d3_sram); 1170MVM_DEBUGFS_READ_WRITE_FILE_OPS(d3_sram);
1072#endif 1171#endif
@@ -1087,10 +1186,14 @@ int iwl_mvm_dbgfs_register(struct iwl_mvm *mvm, struct dentry *dbgfs_dir)
1087 MVM_DEBUGFS_ADD_FILE(sram, mvm->debugfs_dir, S_IWUSR | S_IRUSR); 1186 MVM_DEBUGFS_ADD_FILE(sram, mvm->debugfs_dir, S_IWUSR | S_IRUSR);
1088 MVM_DEBUGFS_ADD_FILE(stations, dbgfs_dir, S_IRUSR); 1187 MVM_DEBUGFS_ADD_FILE(stations, dbgfs_dir, S_IRUSR);
1089 MVM_DEBUGFS_ADD_FILE(bt_notif, dbgfs_dir, S_IRUSR); 1188 MVM_DEBUGFS_ADD_FILE(bt_notif, dbgfs_dir, S_IRUSR);
1090 MVM_DEBUGFS_ADD_FILE(power_down_allow, mvm->debugfs_dir, S_IWUSR); 1189 MVM_DEBUGFS_ADD_FILE(bt_cmd, dbgfs_dir, S_IRUSR);
1091 MVM_DEBUGFS_ADD_FILE(power_down_d3_allow, mvm->debugfs_dir, S_IWUSR); 1190 if (mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_DEVICE_PS_CMD)
1191 MVM_DEBUGFS_ADD_FILE(disable_power_off, mvm->debugfs_dir,
1192 S_IRUSR | S_IWUSR);
1092 MVM_DEBUGFS_ADD_FILE(fw_rx_stats, mvm->debugfs_dir, S_IRUSR); 1193 MVM_DEBUGFS_ADD_FILE(fw_rx_stats, mvm->debugfs_dir, S_IRUSR);
1093 MVM_DEBUGFS_ADD_FILE(fw_restart, mvm->debugfs_dir, S_IWUSR); 1194 MVM_DEBUGFS_ADD_FILE(fw_restart, mvm->debugfs_dir, S_IWUSR);
1195 MVM_DEBUGFS_ADD_FILE(scan_ant_rxchain, mvm->debugfs_dir,
1196 S_IWUSR | S_IRUSR);
1094#ifdef CONFIG_PM_SLEEP 1197#ifdef CONFIG_PM_SLEEP
1095 MVM_DEBUGFS_ADD_FILE(d3_sram, mvm->debugfs_dir, S_IRUSR | S_IWUSR); 1198 MVM_DEBUGFS_ADD_FILE(d3_sram, mvm->debugfs_dir, S_IRUSR | S_IWUSR);
1096 MVM_DEBUGFS_ADD_FILE(d3_test, mvm->debugfs_dir, S_IRUSR); 1199 MVM_DEBUGFS_ADD_FILE(d3_test, mvm->debugfs_dir, S_IRUSR);
diff --git a/drivers/net/wireless/iwlwifi/mvm/fw-api-bt-coex.h b/drivers/net/wireless/iwlwifi/mvm/fw-api-bt-coex.h
index 05c61d6f384e..4ea5e24ca92d 100644
--- a/drivers/net/wireless/iwlwifi/mvm/fw-api-bt-coex.h
+++ b/drivers/net/wireless/iwlwifi/mvm/fw-api-bt-coex.h
@@ -82,6 +82,8 @@
82 * @BT_USE_DEFAULTS: 82 * @BT_USE_DEFAULTS:
83 * @BT_SYNC_2_BT_DISABLE: 83 * @BT_SYNC_2_BT_DISABLE:
84 * @BT_COEX_CORUNNING_TBL_EN: 84 * @BT_COEX_CORUNNING_TBL_EN:
85 *
86 * The COEX_MODE must be set for each command. Even if it is not changed.
85 */ 87 */
86enum iwl_bt_coex_flags { 88enum iwl_bt_coex_flags {
87 BT_CH_PRIMARY_EN = BIT(0), 89 BT_CH_PRIMARY_EN = BIT(0),
@@ -95,14 +97,16 @@ enum iwl_bt_coex_flags {
95 BT_COEX_NW = 0x3 << BT_COEX_MODE_POS, 97 BT_COEX_NW = 0x3 << BT_COEX_MODE_POS,
96 BT_USE_DEFAULTS = BIT(6), 98 BT_USE_DEFAULTS = BIT(6),
97 BT_SYNC_2_BT_DISABLE = BIT(7), 99 BT_SYNC_2_BT_DISABLE = BIT(7),
98 /* 100 BT_COEX_CORUNNING_TBL_EN = BIT(8),
99 * For future use - when the flags will be enlarged 101 BT_COEX_MPLUT_TBL_EN = BIT(9),
100 * BT_COEX_CORUNNING_TBL_EN = BIT(8), 102 /* Bit 10 is reserved */
101 */ 103 BT_COEX_WF_PRIO_BOOST_CHECK_EN = BIT(11),
102}; 104};
103 105
104/* 106/*
105 * indicates what has changed in the BT_COEX command. 107 * indicates what has changed in the BT_COEX command.
108 * BT_VALID_ENABLE must be set for each command. Commands without this bit will
109 * discarded by the firmware
106 */ 110 */
107enum iwl_bt_coex_valid_bit_msk { 111enum iwl_bt_coex_valid_bit_msk {
108 BT_VALID_ENABLE = BIT(0), 112 BT_VALID_ENABLE = BIT(0),
@@ -121,11 +125,8 @@ enum iwl_bt_coex_valid_bit_msk {
121 BT_VALID_CORUN_LUT_40 = BIT(13), 125 BT_VALID_CORUN_LUT_40 = BIT(13),
122 BT_VALID_ANT_ISOLATION = BIT(14), 126 BT_VALID_ANT_ISOLATION = BIT(14),
123 BT_VALID_ANT_ISOLATION_THRS = BIT(15), 127 BT_VALID_ANT_ISOLATION_THRS = BIT(15),
124 /* 128 BT_VALID_TXTX_DELTA_FREQ_THRS = BIT(16),
125 * For future use - when the valid flags will be enlarged 129 BT_VALID_TXRX_MAX_FREQ_0 = BIT(17),
126 * BT_VALID_TXTX_DELTA_FREQ_THRS = BIT(16),
127 * BT_VALID_TXRX_MAX_FREQ_0 = BIT(17),
128 */
129}; 130};
130 131
131/** 132/**
@@ -142,48 +143,88 @@ enum iwl_bt_reduced_tx_power {
142 BT_REDUCED_TX_POWER_DATA = BIT(1), 143 BT_REDUCED_TX_POWER_DATA = BIT(1),
143}; 144};
144 145
146enum iwl_bt_coex_lut_type {
147 BT_COEX_TIGHT_LUT = 0,
148 BT_COEX_LOOSE_LUT,
149 BT_COEX_TX_DIS_LUT,
150
151 BT_COEX_MAX_LUT,
152};
153
145#define BT_COEX_LUT_SIZE (12) 154#define BT_COEX_LUT_SIZE (12)
155#define BT_COEX_CORUN_LUT_SIZE (32)
156#define BT_COEX_MULTI_PRIO_LUT_SIZE (2)
157#define BT_COEX_BOOST_SIZE (4)
158#define BT_REDUCED_TX_POWER_BIT BIT(7)
146 159
147/** 160/**
148 * struct iwl_bt_coex_cmd - bt coex configuration command 161 * struct iwl_bt_coex_cmd - bt coex configuration command
149 * @flags:&enum iwl_bt_coex_flags 162 * @flags:&enum iwl_bt_coex_flags
150 * @lead_time:
151 * @max_kill: 163 * @max_kill:
152 * @bt3_time_t7_value:
153 * @kill_ack_msk:
154 * @kill_cts_msk:
155 * @bt3_prio_sample_time:
156 * @bt3_timer_t2_value:
157 * @bt4_reaction_time:
158 * @decision_lut[12]:
159 * @bt_reduced_tx_power: enum %iwl_bt_reduced_tx_power 164 * @bt_reduced_tx_power: enum %iwl_bt_reduced_tx_power
160 * @valid_bit_msk: enum %iwl_bt_coex_valid_bit_msk 165 * @bt4_antenna_isolation:
161 * @bt_prio_boost: values for PTA boost register 166 * @bt4_antenna_isolation_thr:
167 * @bt4_tx_tx_delta_freq_thr:
168 * @bt4_tx_rx_max_freq0:
169 * @bt_prio_boost:
162 * @wifi_tx_prio_boost: SW boost of wifi tx priority 170 * @wifi_tx_prio_boost: SW boost of wifi tx priority
163 * @wifi_rx_prio_boost: SW boost of wifi rx priority 171 * @wifi_rx_prio_boost: SW boost of wifi rx priority
172 * @kill_ack_msk:
173 * @kill_cts_msk:
174 * @decision_lut:
175 * @bt4_multiprio_lut:
176 * @bt4_corun_lut20:
177 * @bt4_corun_lut40:
178 * @valid_bit_msk: enum %iwl_bt_coex_valid_bit_msk
164 * 179 *
165 * The structure is used for the BT_COEX command. 180 * The structure is used for the BT_COEX command.
166 */ 181 */
167struct iwl_bt_coex_cmd { 182struct iwl_bt_coex_cmd {
168 u8 flags; 183 __le32 flags;
169 u8 lead_time;
170 u8 max_kill; 184 u8 max_kill;
171 u8 bt3_time_t7_value; 185 u8 bt_reduced_tx_power;
186 u8 reserved[2];
187
188 u8 bt4_antenna_isolation;
189 u8 bt4_antenna_isolation_thr;
190 u8 bt4_tx_tx_delta_freq_thr;
191 u8 bt4_tx_rx_max_freq0;
192
193 __le32 bt_prio_boost[BT_COEX_BOOST_SIZE];
194 __le32 wifi_tx_prio_boost;
195 __le32 wifi_rx_prio_boost;
172 __le32 kill_ack_msk; 196 __le32 kill_ack_msk;
173 __le32 kill_cts_msk; 197 __le32 kill_cts_msk;
174 u8 bt3_prio_sample_time; 198
175 u8 bt3_timer_t2_value; 199 __le32 decision_lut[BT_COEX_MAX_LUT][BT_COEX_LUT_SIZE];
176 __le16 bt4_reaction_time; 200 __le32 bt4_multiprio_lut[BT_COEX_MULTI_PRIO_LUT_SIZE];
177 __le32 decision_lut[BT_COEX_LUT_SIZE]; 201 __le32 bt4_corun_lut20[BT_COEX_CORUN_LUT_SIZE];
178 u8 bt_reduced_tx_power; 202 __le32 bt4_corun_lut40[BT_COEX_CORUN_LUT_SIZE];
179 u8 reserved; 203
180 __le16 valid_bit_msk; 204 __le32 valid_bit_msk;
181 __le32 bt_prio_boost;
182 u8 reserved2;
183 u8 wifi_tx_prio_boost;
184 __le16 wifi_rx_prio_boost;
185} __packed; /* BT_COEX_CMD_API_S_VER_3 */ 205} __packed; /* BT_COEX_CMD_API_S_VER_3 */
186 206
207/**
208 * struct iwl_bt_coex_ci_cmd - bt coex channel inhibition command
209 * @bt_primary_ci:
210 * @bt_secondary_ci:
211 * @co_run_bw_primary:
212 * @co_run_bw_secondary:
213 * @primary_ch_phy_id:
214 * @secondary_ch_phy_id:
215 *
216 * Used for BT_COEX_CI command
217 */
218struct iwl_bt_coex_ci_cmd {
219 __le64 bt_primary_ci;
220 __le64 bt_secondary_ci;
221
222 u8 co_run_bw_primary;
223 u8 co_run_bw_secondary;
224 u8 primary_ch_phy_id;
225 u8 secondary_ch_phy_id;
226} __packed; /* BT_CI_MSG_API_S_VER_1 */
227
187#define BT_MBOX(n_dw, _msg, _pos, _nbits) \ 228#define BT_MBOX(n_dw, _msg, _pos, _nbits) \
188 BT_MBOX##n_dw##_##_msg##_POS = (_pos), \ 229 BT_MBOX##n_dw##_##_msg##_POS = (_pos), \
189 BT_MBOX##n_dw##_##_msg = BITS(_nbits) << BT_MBOX##n_dw##_##_msg##_POS 230 BT_MBOX##n_dw##_##_msg = BITS(_nbits) << BT_MBOX##n_dw##_##_msg##_POS
@@ -244,23 +285,39 @@ enum iwl_bt_mxbox_dw3 {
244 ((le32_to_cpu((_notif)->mbox_msg[(_num)]) & BT_MBOX##_num##_##_field)\ 285 ((le32_to_cpu((_notif)->mbox_msg[(_num)]) & BT_MBOX##_num##_##_field)\
245 >> BT_MBOX##_num##_##_field##_POS) 286 >> BT_MBOX##_num##_##_field##_POS)
246 287
288enum iwl_bt_activity_grading {
289 BT_OFF = 0,
290 BT_ON_NO_CONNECTION = 1,
291 BT_LOW_TRAFFIC = 2,
292 BT_HIGH_TRAFFIC = 3,
293};
294
247/** 295/**
248 * struct iwl_bt_coex_profile_notif - notification about BT coex 296 * struct iwl_bt_coex_profile_notif - notification about BT coex
249 * @mbox_msg: message from BT to WiFi 297 * @mbox_msg: message from BT to WiFi
250 * @:bt_status: 0 - off, 1 - on 298 * @msg_idx: the index of the message
251 * @:bt_open_conn: number of BT connections open 299 * @bt_status: 0 - off, 1 - on
252 * @:bt_traffic_load: load of BT traffic 300 * @bt_open_conn: number of BT connections open
253 * @:bt_agg_traffic_load: aggregated load of BT traffic 301 * @bt_traffic_load: load of BT traffic
254 * @:bt_ci_compliance: 0 - no CI compliance, 1 - CI compliant 302 * @bt_agg_traffic_load: aggregated load of BT traffic
303 * @bt_ci_compliance: 0 - no CI compliance, 1 - CI compliant
304 * @primary_ch_lut: LUT used for primary channel
305 * @secondary_ch_lut: LUT used for secondary channel
306 * @bt_activity_grading: the activity of BT enum %iwl_bt_activity_grading
255 */ 307 */
256struct iwl_bt_coex_profile_notif { 308struct iwl_bt_coex_profile_notif {
257 __le32 mbox_msg[4]; 309 __le32 mbox_msg[4];
310 __le32 msg_idx;
258 u8 bt_status; 311 u8 bt_status;
259 u8 bt_open_conn; 312 u8 bt_open_conn;
260 u8 bt_traffic_load; 313 u8 bt_traffic_load;
261 u8 bt_agg_traffic_load; 314 u8 bt_agg_traffic_load;
262 u8 bt_ci_compliance; 315 u8 bt_ci_compliance;
263 u8 reserved[3]; 316 u8 reserved[3];
317
318 __le32 primary_ch_lut;
319 __le32 secondary_ch_lut;
320 __le32 bt_activity_grading;
264} __packed; /* BT_COEX_PROFILE_NTFY_API_S_VER_2 */ 321} __packed; /* BT_COEX_PROFILE_NTFY_API_S_VER_2 */
265 322
266enum iwl_bt_coex_prio_table_event { 323enum iwl_bt_coex_prio_table_event {
@@ -300,20 +357,4 @@ struct iwl_bt_coex_prio_tbl_cmd {
300 u8 prio_tbl[BT_COEX_PRIO_TBL_EVT_MAX]; 357 u8 prio_tbl[BT_COEX_PRIO_TBL_EVT_MAX];
301} __packed; 358} __packed;
302 359
303enum iwl_bt_coex_env_action {
304 BT_COEX_ENV_CLOSE = 0,
305 BT_COEX_ENV_OPEN = 1,
306}; /* BT_COEX_PROT_ENV_ACTION_API_E_VER_1 */
307
308/**
309 * struct iwl_bt_coex_prot_env_cmd - BT Protection Envelope
310 * @action: enum %iwl_bt_coex_env_action
311 * @type: enum %iwl_bt_coex_prio_table_event
312 */
313struct iwl_bt_coex_prot_env_cmd {
314 u8 action; /* 0 = closed, 1 = open */
315 u8 type; /* 0 .. 15 */
316 u8 reserved[2];
317} __packed;
318
319#endif /* __fw_api_bt_coex_h__ */ 360#endif /* __fw_api_bt_coex_h__ */
diff --git a/drivers/net/wireless/iwlwifi/mvm/fw-api-d3.h b/drivers/net/wireless/iwlwifi/mvm/fw-api-d3.h
index df72fcdf8170..4e7dd8cf87dc 100644
--- a/drivers/net/wireless/iwlwifi/mvm/fw-api-d3.h
+++ b/drivers/net/wireless/iwlwifi/mvm/fw-api-d3.h
@@ -100,7 +100,12 @@ enum iwl_proto_offloads {
100 100
101#define IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_V1 2 101#define IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_V1 2
102#define IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_V2 6 102#define IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_V2 6
103#define IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_MAX 6 103#define IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_V3L 12
104#define IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_V3S 4
105#define IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_MAX 12
106
107#define IWL_PROTO_OFFLOAD_NUM_NS_CONFIG_V3L 4
108#define IWL_PROTO_OFFLOAD_NUM_NS_CONFIG_V3S 2
104 109
105/** 110/**
106 * struct iwl_proto_offload_cmd_common - ARP/NS offload common part 111 * struct iwl_proto_offload_cmd_common - ARP/NS offload common part
@@ -155,6 +160,43 @@ struct iwl_proto_offload_cmd_v2 {
155 u8 reserved2[3]; 160 u8 reserved2[3];
156} __packed; /* PROT_OFFLOAD_CONFIG_CMD_DB_S_VER_2 */ 161} __packed; /* PROT_OFFLOAD_CONFIG_CMD_DB_S_VER_2 */
157 162
163struct iwl_ns_config {
164 struct in6_addr source_ipv6_addr;
165 struct in6_addr dest_ipv6_addr;
166 u8 target_mac_addr[ETH_ALEN];
167 __le16 reserved;
168} __packed; /* NS_OFFLOAD_CONFIG */
169
170struct iwl_targ_addr {
171 struct in6_addr addr;
172 __le32 config_num;
173} __packed; /* TARGET_IPV6_ADDRESS */
174
175/**
176 * struct iwl_proto_offload_cmd_v3_small - ARP/NS offload configuration
177 * @common: common/IPv4 configuration
178 * @target_ipv6_addr: target IPv6 addresses
179 * @ns_config: NS offload configurations
180 */
181struct iwl_proto_offload_cmd_v3_small {
182 struct iwl_proto_offload_cmd_common common;
183 __le32 num_valid_ipv6_addrs;
184 struct iwl_targ_addr targ_addrs[IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_V3S];
185 struct iwl_ns_config ns_config[IWL_PROTO_OFFLOAD_NUM_NS_CONFIG_V3S];
186} __packed; /* PROT_OFFLOAD_CONFIG_CMD_DB_S_VER_3 */
187
188/**
189 * struct iwl_proto_offload_cmd_v3_large - ARP/NS offload configuration
190 * @common: common/IPv4 configuration
191 * @target_ipv6_addr: target IPv6 addresses
192 * @ns_config: NS offload configurations
193 */
194struct iwl_proto_offload_cmd_v3_large {
195 struct iwl_proto_offload_cmd_common common;
196 __le32 num_valid_ipv6_addrs;
197 struct iwl_targ_addr targ_addrs[IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_V3L];
198 struct iwl_ns_config ns_config[IWL_PROTO_OFFLOAD_NUM_NS_CONFIG_V3L];
199} __packed; /* PROT_OFFLOAD_CONFIG_CMD_DB_S_VER_3 */
158 200
159/* 201/*
160 * WOWLAN_PATTERNS 202 * WOWLAN_PATTERNS
@@ -293,7 +335,7 @@ enum iwl_wowlan_wakeup_reason {
293 IWL_WOWLAN_WAKEUP_BY_REM_WAKE_WAKEUP_PACKET = BIT(12), 335 IWL_WOWLAN_WAKEUP_BY_REM_WAKE_WAKEUP_PACKET = BIT(12),
294}; /* WOWLAN_WAKE_UP_REASON_API_E_VER_2 */ 336}; /* WOWLAN_WAKE_UP_REASON_API_E_VER_2 */
295 337
296struct iwl_wowlan_status { 338struct iwl_wowlan_status_v4 {
297 __le64 replay_ctr; 339 __le64 replay_ctr;
298 __le16 pattern_number; 340 __le16 pattern_number;
299 __le16 non_qos_seq_ctr; 341 __le16 non_qos_seq_ctr;
@@ -308,6 +350,29 @@ struct iwl_wowlan_status {
308 u8 wake_packet[]; /* can be truncated from _length to _bufsize */ 350 u8 wake_packet[]; /* can be truncated from _length to _bufsize */
309} __packed; /* WOWLAN_STATUSES_API_S_VER_4 */ 351} __packed; /* WOWLAN_STATUSES_API_S_VER_4 */
310 352
353struct iwl_wowlan_gtk_status {
354 u8 key_index;
355 u8 reserved[3];
356 u8 decrypt_key[16];
357 u8 tkip_mic_key[8];
358 struct iwl_wowlan_rsc_tsc_params_cmd rsc;
359} __packed;
360
361struct iwl_wowlan_status_v6 {
362 struct iwl_wowlan_gtk_status gtk;
363 __le64 replay_ctr;
364 __le16 pattern_number;
365 __le16 non_qos_seq_ctr;
366 __le16 qos_seq_ctr[8];
367 __le32 wakeup_reasons;
368 __le32 num_of_gtk_rekeys;
369 __le32 transmitted_ndps;
370 __le32 received_beacons;
371 __le32 wake_packet_length;
372 __le32 wake_packet_bufsize;
373 u8 wake_packet[]; /* can be truncated from _length to _bufsize */
374} __packed; /* WOWLAN_STATUSES_API_S_VER_6 */
375
311#define IWL_WOWLAN_TCP_MAX_PACKET_LEN 64 376#define IWL_WOWLAN_TCP_MAX_PACKET_LEN 64
312#define IWL_WOWLAN_REMOTE_WAKE_MAX_PACKET_LEN 128 377#define IWL_WOWLAN_REMOTE_WAKE_MAX_PACKET_LEN 128
313#define IWL_WOWLAN_REMOTE_WAKE_MAX_TOKENS 2048 378#define IWL_WOWLAN_REMOTE_WAKE_MAX_TOKENS 2048
diff --git a/drivers/net/wireless/iwlwifi/mvm/fw-api-mac.h b/drivers/net/wireless/iwlwifi/mvm/fw-api-mac.h
index 98b1feb43d38..39c3148bdfa8 100644
--- a/drivers/net/wireless/iwlwifi/mvm/fw-api-mac.h
+++ b/drivers/net/wireless/iwlwifi/mvm/fw-api-mac.h
@@ -170,12 +170,14 @@ struct iwl_mac_data_ap {
170 * @beacon_tsf: beacon transmit time in TSF 170 * @beacon_tsf: beacon transmit time in TSF
171 * @bi: beacon interval in TU 171 * @bi: beacon interval in TU
172 * @bi_reciprocal: 2^32 / bi 172 * @bi_reciprocal: 2^32 / bi
173 * @beacon_template: beacon template ID
173 */ 174 */
174struct iwl_mac_data_ibss { 175struct iwl_mac_data_ibss {
175 __le32 beacon_time; 176 __le32 beacon_time;
176 __le64 beacon_tsf; 177 __le64 beacon_tsf;
177 __le32 bi; 178 __le32 bi;
178 __le32 bi_reciprocal; 179 __le32 bi_reciprocal;
180 __le32 beacon_template;
179} __packed; /* IBSS_MAC_DATA_API_S_VER_1 */ 181} __packed; /* IBSS_MAC_DATA_API_S_VER_1 */
180 182
181/** 183/**
@@ -372,4 +374,13 @@ static inline u32 iwl_mvm_reciprocal(u32 v)
372 return 0xFFFFFFFF / v; 374 return 0xFFFFFFFF / v;
373} 375}
374 376
377#define IWL_NONQOS_SEQ_GET 0x1
378#define IWL_NONQOS_SEQ_SET 0x2
379struct iwl_nonqos_seq_query_cmd {
380 __le32 get_set_flag;
381 __le32 mac_id_n_color;
382 __le16 value;
383 __le16 reserved;
384} __packed; /* NON_QOS_TX_COUNTER_GET_SET_API_S_VER_1 */
385
375#endif /* __fw_api_mac_h__ */ 386#endif /* __fw_api_mac_h__ */
diff --git a/drivers/net/wireless/iwlwifi/mvm/fw-api-power.h b/drivers/net/wireless/iwlwifi/mvm/fw-api-power.h
index 8e7ab41079ca..5cb93ae5cd2f 100644
--- a/drivers/net/wireless/iwlwifi/mvm/fw-api-power.h
+++ b/drivers/net/wireless/iwlwifi/mvm/fw-api-power.h
@@ -132,6 +132,33 @@ struct iwl_powertable_cmd {
132} __packed; 132} __packed;
133 133
134/** 134/**
135 * enum iwl_device_power_flags - masks for device power command flags
136 * @DEVIC_POWER_FLAGS_POWER_SAVE_ENA_MSK: '1' Allow to save power by turning off
137 * receiver and transmitter. '0' - does not allow. This flag should be
138 * always set to '1' unless one need to disable actual power down for debug
139 * purposes.
140 * @DEVICE_POWER_FLAGS_CAM_MSK: '1' CAM (Continuous Active Mode) is set, meaning
141 * that power management is disabled. '0' Power management is enabled, one
142 * of power schemes is applied.
143*/
144enum iwl_device_power_flags {
145 DEVICE_POWER_FLAGS_POWER_SAVE_ENA_MSK = BIT(0),
146 DEVICE_POWER_FLAGS_CAM_MSK = BIT(13),
147};
148
149/**
150 * struct iwl_device_power_cmd - device wide power command.
151 * DEVICE_POWER_CMD = 0x77 (command, has simple generic response)
152 *
153 * @flags: Power table command flags from DEVICE_POWER_FLAGS_*
154 */
155struct iwl_device_power_cmd {
156 /* PM_POWER_TABLE_CMD_API_S_VER_6 */
157 __le16 flags;
158 __le16 reserved;
159} __packed;
160
161/**
135 * struct iwl_mac_power_cmd - New power command containing uAPSD support 162 * struct iwl_mac_power_cmd - New power command containing uAPSD support
136 * MAC_PM_POWER_TABLE = 0xA9 (command, has simple generic response) 163 * MAC_PM_POWER_TABLE = 0xA9 (command, has simple generic response)
137 * @id_and_color: MAC contex identifier 164 * @id_and_color: MAC contex identifier
@@ -290,7 +317,7 @@ struct iwl_beacon_filter_cmd {
290#define IWL_BF_ESCAPE_TIMER_MIN 0 317#define IWL_BF_ESCAPE_TIMER_MIN 0
291 318
292#define IWL_BA_ESCAPE_TIMER_DEFAULT 6 319#define IWL_BA_ESCAPE_TIMER_DEFAULT 6
293#define IWL_BA_ESCAPE_TIMER_D3 6 320#define IWL_BA_ESCAPE_TIMER_D3 9
294#define IWL_BA_ESCAPE_TIMER_MAX 1024 321#define IWL_BA_ESCAPE_TIMER_MAX 1024
295#define IWL_BA_ESCAPE_TIMER_MIN 0 322#define IWL_BA_ESCAPE_TIMER_MIN 0
296 323
diff --git a/drivers/net/wireless/iwlwifi/mvm/fw-api-rs.h b/drivers/net/wireless/iwlwifi/mvm/fw-api-rs.h
index fdd33bc0a594..538f1c7a5966 100644
--- a/drivers/net/wireless/iwlwifi/mvm/fw-api-rs.h
+++ b/drivers/net/wireless/iwlwifi/mvm/fw-api-rs.h
@@ -68,6 +68,7 @@
68/* 68/*
69 * These serve as indexes into 69 * These serve as indexes into
70 * struct iwl_rate_info fw_rate_idx_to_plcp[IWL_RATE_COUNT]; 70 * struct iwl_rate_info fw_rate_idx_to_plcp[IWL_RATE_COUNT];
71 * TODO: avoid overlap between legacy and HT rates
71 */ 72 */
72enum { 73enum {
73 IWL_RATE_1M_INDEX = 0, 74 IWL_RATE_1M_INDEX = 0,
@@ -78,18 +79,31 @@ enum {
78 IWL_LAST_CCK_RATE = IWL_RATE_11M_INDEX, 79 IWL_LAST_CCK_RATE = IWL_RATE_11M_INDEX,
79 IWL_RATE_6M_INDEX, 80 IWL_RATE_6M_INDEX,
80 IWL_FIRST_OFDM_RATE = IWL_RATE_6M_INDEX, 81 IWL_FIRST_OFDM_RATE = IWL_RATE_6M_INDEX,
82 IWL_RATE_MCS_0_INDEX = IWL_RATE_6M_INDEX,
83 IWL_FIRST_HT_RATE = IWL_RATE_MCS_0_INDEX,
84 IWL_FIRST_VHT_RATE = IWL_RATE_MCS_0_INDEX,
81 IWL_RATE_9M_INDEX, 85 IWL_RATE_9M_INDEX,
82 IWL_RATE_12M_INDEX, 86 IWL_RATE_12M_INDEX,
87 IWL_RATE_MCS_1_INDEX = IWL_RATE_12M_INDEX,
83 IWL_RATE_18M_INDEX, 88 IWL_RATE_18M_INDEX,
89 IWL_RATE_MCS_2_INDEX = IWL_RATE_18M_INDEX,
84 IWL_RATE_24M_INDEX, 90 IWL_RATE_24M_INDEX,
91 IWL_RATE_MCS_3_INDEX = IWL_RATE_24M_INDEX,
85 IWL_RATE_36M_INDEX, 92 IWL_RATE_36M_INDEX,
93 IWL_RATE_MCS_4_INDEX = IWL_RATE_36M_INDEX,
86 IWL_RATE_48M_INDEX, 94 IWL_RATE_48M_INDEX,
95 IWL_RATE_MCS_5_INDEX = IWL_RATE_48M_INDEX,
87 IWL_RATE_54M_INDEX, 96 IWL_RATE_54M_INDEX,
97 IWL_RATE_MCS_6_INDEX = IWL_RATE_54M_INDEX,
88 IWL_LAST_NON_HT_RATE = IWL_RATE_54M_INDEX, 98 IWL_LAST_NON_HT_RATE = IWL_RATE_54M_INDEX,
89 IWL_RATE_60M_INDEX, 99 IWL_RATE_60M_INDEX,
90 IWL_LAST_OFDM_RATE = IWL_RATE_60M_INDEX, 100 IWL_RATE_MCS_7_INDEX = IWL_RATE_60M_INDEX,
101 IWL_LAST_HT_RATE = IWL_RATE_MCS_7_INDEX,
102 IWL_RATE_MCS_8_INDEX,
103 IWL_RATE_MCS_9_INDEX,
104 IWL_LAST_VHT_RATE = IWL_RATE_MCS_9_INDEX,
91 IWL_RATE_COUNT_LEGACY = IWL_LAST_NON_HT_RATE + 1, 105 IWL_RATE_COUNT_LEGACY = IWL_LAST_NON_HT_RATE + 1,
92 IWL_RATE_COUNT, 106 IWL_RATE_COUNT = IWL_LAST_VHT_RATE + 1,
93}; 107};
94 108
95#define IWL_RATE_BIT_MSK(r) BIT(IWL_RATE_##r##M_INDEX) 109#define IWL_RATE_BIT_MSK(r) BIT(IWL_RATE_##r##M_INDEX)
@@ -108,6 +122,7 @@ enum {
108 IWL_RATE_2M_PLCP = 20, 122 IWL_RATE_2M_PLCP = 20,
109 IWL_RATE_5M_PLCP = 55, 123 IWL_RATE_5M_PLCP = 55,
110 IWL_RATE_11M_PLCP = 110, 124 IWL_RATE_11M_PLCP = 110,
125 IWL_RATE_INVM_PLCP = -1,
111}; 126};
112 127
113/* 128/*
@@ -164,6 +179,8 @@ enum {
164 * which is the duplicate 20 MHz MCS (bit 5 set, all others zero.) 179 * which is the duplicate 20 MHz MCS (bit 5 set, all others zero.)
165 */ 180 */
166#define RATE_HT_MCS_RATE_CODE_MSK 0x7 181#define RATE_HT_MCS_RATE_CODE_MSK 0x7
182#define RATE_HT_MCS_NSS_POS 3
183#define RATE_HT_MCS_NSS_MSK (3 << RATE_HT_MCS_NSS_POS)
167 184
168/* Bit 10: (1) Use Green Field preamble */ 185/* Bit 10: (1) Use Green Field preamble */
169#define RATE_HT_MCS_GF_POS 10 186#define RATE_HT_MCS_GF_POS 10
diff --git a/drivers/net/wireless/iwlwifi/mvm/fw-api-scan.h b/drivers/net/wireless/iwlwifi/mvm/fw-api-scan.h
index 83cb9b992ea4..c3782b48ded1 100644
--- a/drivers/net/wireless/iwlwifi/mvm/fw-api-scan.h
+++ b/drivers/net/wireless/iwlwifi/mvm/fw-api-scan.h
@@ -356,6 +356,7 @@ struct iwl_scan_complete_notif {
356/* scan offload */ 356/* scan offload */
357#define IWL_MAX_SCAN_CHANNELS 40 357#define IWL_MAX_SCAN_CHANNELS 40
358#define IWL_SCAN_MAX_BLACKLIST_LEN 64 358#define IWL_SCAN_MAX_BLACKLIST_LEN 64
359#define IWL_SCAN_SHORT_BLACKLIST_LEN 16
359#define IWL_SCAN_MAX_PROFILES 11 360#define IWL_SCAN_MAX_PROFILES 11
360#define SCAN_OFFLOAD_PROBE_REQ_SIZE 512 361#define SCAN_OFFLOAD_PROBE_REQ_SIZE 512
361 362
@@ -368,6 +369,12 @@ struct iwl_scan_complete_notif {
368#define IWL_FULL_SCAN_MULTIPLIER 5 369#define IWL_FULL_SCAN_MULTIPLIER 5
369#define IWL_FAST_SCHED_SCAN_ITERATIONS 3 370#define IWL_FAST_SCHED_SCAN_ITERATIONS 3
370 371
372enum scan_framework_client {
373 SCAN_CLIENT_SCHED_SCAN = BIT(0),
374 SCAN_CLIENT_NETDETECT = BIT(1),
375 SCAN_CLIENT_ASSET_TRACKING = BIT(2),
376};
377
371/** 378/**
372 * struct iwl_scan_offload_cmd - SCAN_REQUEST_FIXED_PART_API_S_VER_6 379 * struct iwl_scan_offload_cmd - SCAN_REQUEST_FIXED_PART_API_S_VER_6
373 * @scan_flags: see enum iwl_scan_flags 380 * @scan_flags: see enum iwl_scan_flags
@@ -449,11 +456,12 @@ struct iwl_scan_offload_cfg {
449 * iwl_scan_offload_blacklist - SCAN_OFFLOAD_BLACKLIST_S 456 * iwl_scan_offload_blacklist - SCAN_OFFLOAD_BLACKLIST_S
450 * @ssid: MAC address to filter out 457 * @ssid: MAC address to filter out
451 * @reported_rssi: AP rssi reported to the host 458 * @reported_rssi: AP rssi reported to the host
459 * @client_bitmap: clients ignore this entry - enum scan_framework_client
452 */ 460 */
453struct iwl_scan_offload_blacklist { 461struct iwl_scan_offload_blacklist {
454 u8 ssid[ETH_ALEN]; 462 u8 ssid[ETH_ALEN];
455 u8 reported_rssi; 463 u8 reported_rssi;
456 u8 reserved; 464 u8 client_bitmap;
457} __packed; 465} __packed;
458 466
459enum iwl_scan_offload_network_type { 467enum iwl_scan_offload_network_type {
@@ -475,6 +483,7 @@ enum iwl_scan_offload_band_selection {
475 * @aut_alg: authentication olgorithm to match - bitmap 483 * @aut_alg: authentication olgorithm to match - bitmap
476 * @network_type: enum iwl_scan_offload_network_type 484 * @network_type: enum iwl_scan_offload_network_type
477 * @band_selection: enum iwl_scan_offload_band_selection 485 * @band_selection: enum iwl_scan_offload_band_selection
486 * @client_bitmap: clients waiting for match - enum scan_framework_client
478 */ 487 */
479struct iwl_scan_offload_profile { 488struct iwl_scan_offload_profile {
480 u8 ssid_index; 489 u8 ssid_index;
@@ -482,7 +491,8 @@ struct iwl_scan_offload_profile {
482 u8 auth_alg; 491 u8 auth_alg;
483 u8 network_type; 492 u8 network_type;
484 u8 band_selection; 493 u8 band_selection;
485 u8 reserved[3]; 494 u8 client_bitmap;
495 u8 reserved[2];
486} __packed; 496} __packed;
487 497
488/** 498/**
@@ -491,13 +501,18 @@ struct iwl_scan_offload_profile {
491 * @profiles: profiles to search for match 501 * @profiles: profiles to search for match
492 * @blacklist_len: length of blacklist 502 * @blacklist_len: length of blacklist
493 * @num_profiles: num of profiles in the list 503 * @num_profiles: num of profiles in the list
504 * @match_notify: clients waiting for match found notification
505 * @pass_match: clients waiting for the results
506 * @active_clients: active clients bitmap - enum scan_framework_client
494 */ 507 */
495struct iwl_scan_offload_profile_cfg { 508struct iwl_scan_offload_profile_cfg {
496 struct iwl_scan_offload_blacklist blacklist[IWL_SCAN_MAX_BLACKLIST_LEN];
497 struct iwl_scan_offload_profile profiles[IWL_SCAN_MAX_PROFILES]; 509 struct iwl_scan_offload_profile profiles[IWL_SCAN_MAX_PROFILES];
498 u8 blacklist_len; 510 u8 blacklist_len;
499 u8 num_profiles; 511 u8 num_profiles;
500 u8 reserved[2]; 512 u8 match_notify;
513 u8 pass_match;
514 u8 active_clients;
515 u8 reserved[3];
501} __packed; 516} __packed;
502 517
503/** 518/**
@@ -560,4 +575,15 @@ struct iwl_scan_offload_complete {
560 u8 reserved; 575 u8 reserved;
561} __packed; 576} __packed;
562 577
578/**
579 * iwl_sched_scan_results - SCAN_OFFLOAD_MATCH_FOUND_NTF_API_S_VER_1
580 * @ssid_bitmap: SSIDs indexes found in this iteration
581 * @client_bitmap: clients that are active and wait for this notification
582 */
583struct iwl_sched_scan_results {
584 __le16 ssid_bitmap;
585 u8 client_bitmap;
586 u8 reserved;
587};
588
563#endif 589#endif
diff --git a/drivers/net/wireless/iwlwifi/mvm/fw-api-sta.h b/drivers/net/wireless/iwlwifi/mvm/fw-api-sta.h
index a30691a8a85b..4aca5933a65d 100644
--- a/drivers/net/wireless/iwlwifi/mvm/fw-api-sta.h
+++ b/drivers/net/wireless/iwlwifi/mvm/fw-api-sta.h
@@ -247,7 +247,7 @@ struct iwl_mvm_keyinfo {
247} __packed; 247} __packed;
248 248
249/** 249/**
250 * struct iwl_mvm_add_sta_cmd - Add / modify a station in the fw's station table 250 * struct iwl_mvm_add_sta_cmd_v5 - Add/modify a station in the fw's sta table.
251 * ( REPLY_ADD_STA = 0x18 ) 251 * ( REPLY_ADD_STA = 0x18 )
252 * @add_modify: 1: modify existing, 0: add new station 252 * @add_modify: 1: modify existing, 0: add new station
253 * @unicast_tx_key_id: unicast tx key id. Relevant only when unicast key sent 253 * @unicast_tx_key_id: unicast tx key id. Relevant only when unicast key sent
@@ -286,7 +286,7 @@ struct iwl_mvm_keyinfo {
286 * ADD_STA sets up the table entry for one station, either creating a new 286 * ADD_STA sets up the table entry for one station, either creating a new
287 * entry, or modifying a pre-existing one. 287 * entry, or modifying a pre-existing one.
288 */ 288 */
289struct iwl_mvm_add_sta_cmd { 289struct iwl_mvm_add_sta_cmd_v5 {
290 u8 add_modify; 290 u8 add_modify;
291 u8 unicast_tx_key_id; 291 u8 unicast_tx_key_id;
292 u8 multicast_tx_key_id; 292 u8 multicast_tx_key_id;
@@ -313,6 +313,57 @@ struct iwl_mvm_add_sta_cmd {
313} __packed; /* ADD_STA_CMD_API_S_VER_5 */ 313} __packed; /* ADD_STA_CMD_API_S_VER_5 */
314 314
315/** 315/**
316 * struct iwl_mvm_add_sta_cmd_v6 - Add / modify a station
317 * VER_6 of this command is quite similar to VER_5 except
318 * exclusion of all fields related to the security key installation.
319 */
320struct iwl_mvm_add_sta_cmd_v6 {
321 u8 add_modify;
322 u8 reserved1;
323 __le16 tid_disable_tx;
324 __le32 mac_id_n_color;
325 u8 addr[ETH_ALEN]; /* _STA_ID_MODIFY_INFO_API_S_VER_1 */
326 __le16 reserved2;
327 u8 sta_id;
328 u8 modify_mask;
329 __le16 reserved3;
330 __le32 station_flags;
331 __le32 station_flags_msk;
332 u8 add_immediate_ba_tid;
333 u8 remove_immediate_ba_tid;
334 __le16 add_immediate_ba_ssn;
335 __le16 sleep_tx_count;
336 __le16 sleep_state_flags;
337 __le16 assoc_id;
338 __le16 beamform_flags;
339 __le32 tfd_queue_msk;
340} __packed; /* ADD_STA_CMD_API_S_VER_6 */
341
342/**
343 * struct iwl_mvm_add_sta_key_cmd - add/modify sta key
344 * ( REPLY_ADD_STA_KEY = 0x17 )
345 * @sta_id: index of station in uCode's station table
346 * @key_offset: key offset in key storage
347 * @key_flags: type %iwl_sta_key_flag
348 * @key: key material data
349 * @key2: key material data
350 * @rx_secur_seq_cnt: RX security sequence counter for the key
351 * @tkip_rx_tsc_byte2: TSC[2] for key mix ph1 detection
352 * @tkip_rx_ttak: 10-byte unicast TKIP TTAK for Rx
353 */
354struct iwl_mvm_add_sta_key_cmd {
355 u8 sta_id;
356 u8 key_offset;
357 __le16 key_flags;
358 u8 key[16];
359 u8 key2[16];
360 u8 rx_secur_seq_cnt[16];
361 u8 tkip_rx_tsc_byte2;
362 u8 reserved;
363 __le16 tkip_rx_ttak[5];
364} __packed; /* ADD_MODIFY_STA_KEY_API_S_VER_1 */
365
366/**
316 * enum iwl_mvm_add_sta_rsp_status - status in the response to ADD_STA command 367 * enum iwl_mvm_add_sta_rsp_status - status in the response to ADD_STA command
317 * @ADD_STA_SUCCESS: operation was executed successfully 368 * @ADD_STA_SUCCESS: operation was executed successfully
318 * @ADD_STA_STATIONS_OVERLOAD: no room left in the fw's station table 369 * @ADD_STA_STATIONS_OVERLOAD: no room left in the fw's station table
diff --git a/drivers/net/wireless/iwlwifi/mvm/fw-api.h b/drivers/net/wireless/iwlwifi/mvm/fw-api.h
index 66264cc5a016..bad5a552dd8d 100644
--- a/drivers/net/wireless/iwlwifi/mvm/fw-api.h
+++ b/drivers/net/wireless/iwlwifi/mvm/fw-api.h
@@ -72,17 +72,17 @@
72#include "fw-api-d3.h" 72#include "fw-api-d3.h"
73#include "fw-api-bt-coex.h" 73#include "fw-api-bt-coex.h"
74 74
75/* queue and FIFO numbers by usage */ 75/* maximal number of Tx queues in any platform */
76#define IWL_MVM_MAX_QUEUES 20
77
78/* Tx queue numbers */
76enum { 79enum {
77 IWL_MVM_OFFCHANNEL_QUEUE = 8, 80 IWL_MVM_OFFCHANNEL_QUEUE = 8,
78 IWL_MVM_CMD_QUEUE = 9, 81 IWL_MVM_CMD_QUEUE = 9,
79 IWL_MVM_AUX_QUEUE = 15,
80 IWL_MVM_FIRST_AGG_QUEUE = 16,
81 IWL_MVM_NUM_QUEUES = 20,
82 IWL_MVM_LAST_AGG_QUEUE = IWL_MVM_NUM_QUEUES - 1,
83 IWL_MVM_CMD_FIFO = 7
84}; 82};
85 83
84#define IWL_MVM_CMD_FIFO 7
85
86#define IWL_MVM_STATION_COUNT 16 86#define IWL_MVM_STATION_COUNT 16
87 87
88/* commands */ 88/* commands */
@@ -97,6 +97,7 @@ enum {
97 DBG_CFG = 0x9, 97 DBG_CFG = 0x9,
98 98
99 /* station table */ 99 /* station table */
100 ADD_STA_KEY = 0x17,
100 ADD_STA = 0x18, 101 ADD_STA = 0x18,
101 REMOVE_STA = 0x19, 102 REMOVE_STA = 0x19,
102 103
@@ -114,6 +115,7 @@ enum {
114 TIME_EVENT_NOTIFICATION = 0x2a, 115 TIME_EVENT_NOTIFICATION = 0x2a,
115 BINDING_CONTEXT_CMD = 0x2b, 116 BINDING_CONTEXT_CMD = 0x2b,
116 TIME_QUOTA_CMD = 0x2c, 117 TIME_QUOTA_CMD = 0x2c,
118 NON_QOS_TX_COUNTER_CMD = 0x2d,
117 119
118 LQ_CMD = 0x4e, 120 LQ_CMD = 0x4e,
119 121
@@ -130,6 +132,7 @@ enum {
130 SCAN_OFFLOAD_COMPLETE = 0x6D, 132 SCAN_OFFLOAD_COMPLETE = 0x6D,
131 SCAN_OFFLOAD_UPDATE_PROFILES_CMD = 0x6E, 133 SCAN_OFFLOAD_UPDATE_PROFILES_CMD = 0x6E,
132 SCAN_OFFLOAD_CONFIG_CMD = 0x6f, 134 SCAN_OFFLOAD_CONFIG_CMD = 0x6f,
135 MATCH_FOUND_NOTIFICATION = 0xd9,
133 136
134 /* Phy */ 137 /* Phy */
135 PHY_CONFIGURATION_CMD = 0x6a, 138 PHY_CONFIGURATION_CMD = 0x6a,
@@ -178,6 +181,7 @@ enum {
178 BT_COEX_PRIO_TABLE = 0xcc, 181 BT_COEX_PRIO_TABLE = 0xcc,
179 BT_COEX_PROT_ENV = 0xcd, 182 BT_COEX_PROT_ENV = 0xcd,
180 BT_PROFILE_NOTIFICATION = 0xce, 183 BT_PROFILE_NOTIFICATION = 0xce,
184 BT_COEX_CI = 0x5d,
181 185
182 REPLY_BEACON_FILTERING_CMD = 0xd2, 186 REPLY_BEACON_FILTERING_CMD = 0xd2,
183 187
diff --git a/drivers/net/wireless/iwlwifi/mvm/fw.c b/drivers/net/wireless/iwlwifi/mvm/fw.c
index c76299a3a1e0..70e5297646b2 100644
--- a/drivers/net/wireless/iwlwifi/mvm/fw.c
+++ b/drivers/net/wireless/iwlwifi/mvm/fw.c
@@ -151,13 +151,11 @@ static int iwl_mvm_load_ucode_wait_alive(struct iwl_mvm *mvm,
151 enum iwl_ucode_type old_type = mvm->cur_ucode; 151 enum iwl_ucode_type old_type = mvm->cur_ucode;
152 static const u8 alive_cmd[] = { MVM_ALIVE }; 152 static const u8 alive_cmd[] = { MVM_ALIVE };
153 153
154 mvm->cur_ucode = ucode_type;
155 fw = iwl_get_ucode_image(mvm, ucode_type); 154 fw = iwl_get_ucode_image(mvm, ucode_type);
156 155 if (WARN_ON(!fw))
157 mvm->ucode_loaded = false;
158
159 if (!fw)
160 return -EINVAL; 156 return -EINVAL;
157 mvm->cur_ucode = ucode_type;
158 mvm->ucode_loaded = false;
161 159
162 iwl_init_notification_wait(&mvm->notif_wait, &alive_wait, 160 iwl_init_notification_wait(&mvm->notif_wait, &alive_wait,
163 alive_cmd, ARRAY_SIZE(alive_cmd), 161 alive_cmd, ARRAY_SIZE(alive_cmd),
@@ -199,7 +197,7 @@ static int iwl_mvm_load_ucode_wait_alive(struct iwl_mvm *mvm,
199 */ 197 */
200 198
201 for (i = 0; i < IWL_MAX_HW_QUEUES; i++) { 199 for (i = 0; i < IWL_MAX_HW_QUEUES; i++) {
202 if (i < IWL_MVM_FIRST_AGG_QUEUE && i != IWL_MVM_CMD_QUEUE) 200 if (i < mvm->first_agg_queue && i != IWL_MVM_CMD_QUEUE)
203 mvm->queue_to_mac80211[i] = i; 201 mvm->queue_to_mac80211[i] = i;
204 else 202 else
205 mvm->queue_to_mac80211[i] = IWL_INVALID_MAC80211_QUEUE; 203 mvm->queue_to_mac80211[i] = IWL_INVALID_MAC80211_QUEUE;
@@ -243,7 +241,7 @@ int iwl_run_init_mvm_ucode(struct iwl_mvm *mvm, bool read_nvm)
243 241
244 lockdep_assert_held(&mvm->mutex); 242 lockdep_assert_held(&mvm->mutex);
245 243
246 if (mvm->init_ucode_run) 244 if (mvm->init_ucode_complete)
247 return 0; 245 return 0;
248 246
249 iwl_init_notification_wait(&mvm->notif_wait, 247 iwl_init_notification_wait(&mvm->notif_wait,
@@ -264,6 +262,7 @@ int iwl_run_init_mvm_ucode(struct iwl_mvm *mvm, bool read_nvm)
264 if (ret) 262 if (ret)
265 goto error; 263 goto error;
266 264
265 /* Read the NVM only at driver load time, no need to do this twice */
267 if (read_nvm) { 266 if (read_nvm) {
268 /* Read nvm */ 267 /* Read nvm */
269 ret = iwl_nvm_init(mvm); 268 ret = iwl_nvm_init(mvm);
@@ -273,6 +272,10 @@ int iwl_run_init_mvm_ucode(struct iwl_mvm *mvm, bool read_nvm)
273 } 272 }
274 } 273 }
275 274
275 /* In case we read the NVM from external file, load it to the NIC */
276 if (iwlwifi_mod_params.nvm_file)
277 iwl_mvm_load_nvm_to_nic(mvm);
278
276 ret = iwl_nvm_check_version(mvm->nvm_data, mvm->trans); 279 ret = iwl_nvm_check_version(mvm->nvm_data, mvm->trans);
277 WARN_ON(ret); 280 WARN_ON(ret);
278 281
@@ -310,7 +313,7 @@ int iwl_run_init_mvm_ucode(struct iwl_mvm *mvm, bool read_nvm)
310 ret = iwl_wait_notification(&mvm->notif_wait, &calib_wait, 313 ret = iwl_wait_notification(&mvm->notif_wait, &calib_wait,
311 MVM_UCODE_CALIB_TIMEOUT); 314 MVM_UCODE_CALIB_TIMEOUT);
312 if (!ret) 315 if (!ret)
313 mvm->init_ucode_run = true; 316 mvm->init_ucode_complete = true;
314 goto out; 317 goto out;
315 318
316error: 319error:
@@ -353,8 +356,12 @@ int iwl_mvm_up(struct iwl_mvm *mvm)
353 if (ret) 356 if (ret)
354 return ret; 357 return ret;
355 358
356 /* If we were in RFKILL during module loading, load init ucode now */ 359 /*
357 if (!mvm->init_ucode_run) { 360 * If we haven't completed the run of the init ucode during
361 * module loading, load init ucode now
362 * (for example, if we were in RFKILL)
363 */
364 if (!mvm->init_ucode_complete) {
358 ret = iwl_run_init_mvm_ucode(mvm, false); 365 ret = iwl_run_init_mvm_ucode(mvm, false);
359 if (ret && !iwlmvm_mod_params.init_dbg) { 366 if (ret && !iwlmvm_mod_params.init_dbg) {
360 IWL_ERR(mvm, "Failed to run INIT ucode: %d\n", ret); 367 IWL_ERR(mvm, "Failed to run INIT ucode: %d\n", ret);
@@ -424,6 +431,10 @@ int iwl_mvm_up(struct iwl_mvm *mvm)
424 goto error; 431 goto error;
425 } 432 }
426 433
434 ret = iwl_mvm_power_update_device_mode(mvm);
435 if (ret)
436 goto error;
437
427 IWL_DEBUG_INFO(mvm, "RT uCode started.\n"); 438 IWL_DEBUG_INFO(mvm, "RT uCode started.\n");
428 return 0; 439 return 0;
429 error: 440 error:
diff --git a/drivers/net/wireless/iwlwifi/mvm/mac-ctxt.c b/drivers/net/wireless/iwlwifi/mvm/mac-ctxt.c
index 5fe23a5ea9b6..f41f9b079831 100644
--- a/drivers/net/wireless/iwlwifi/mvm/mac-ctxt.c
+++ b/drivers/net/wireless/iwlwifi/mvm/mac-ctxt.c
@@ -80,7 +80,7 @@ struct iwl_mvm_mac_iface_iterator_data {
80 struct ieee80211_vif *vif; 80 struct ieee80211_vif *vif;
81 unsigned long available_mac_ids[BITS_TO_LONGS(NUM_MAC_INDEX_DRIVER)]; 81 unsigned long available_mac_ids[BITS_TO_LONGS(NUM_MAC_INDEX_DRIVER)];
82 unsigned long available_tsf_ids[BITS_TO_LONGS(NUM_TSF_IDS)]; 82 unsigned long available_tsf_ids[BITS_TO_LONGS(NUM_TSF_IDS)];
83 unsigned long used_hw_queues[BITS_TO_LONGS(IWL_MVM_FIRST_AGG_QUEUE)]; 83 unsigned long used_hw_queues[BITS_TO_LONGS(IWL_MVM_MAX_QUEUES)];
84 enum iwl_tsf_id preferred_tsf; 84 enum iwl_tsf_id preferred_tsf;
85 bool found_vif; 85 bool found_vif;
86}; 86};
@@ -218,7 +218,7 @@ static int iwl_mvm_mac_ctxt_allocate_resources(struct iwl_mvm *mvm,
218 .preferred_tsf = NUM_TSF_IDS, 218 .preferred_tsf = NUM_TSF_IDS,
219 .used_hw_queues = { 219 .used_hw_queues = {
220 BIT(IWL_MVM_OFFCHANNEL_QUEUE) | 220 BIT(IWL_MVM_OFFCHANNEL_QUEUE) |
221 BIT(IWL_MVM_AUX_QUEUE) | 221 BIT(mvm->aux_queue) |
222 BIT(IWL_MVM_CMD_QUEUE) 222 BIT(IWL_MVM_CMD_QUEUE)
223 }, 223 },
224 .found_vif = false, 224 .found_vif = false,
@@ -242,9 +242,17 @@ static int iwl_mvm_mac_ctxt_allocate_resources(struct iwl_mvm *mvm,
242 * that we should share it with another interface. 242 * that we should share it with another interface.
243 */ 243 */
244 244
245 /* Currently, MAC ID 0 should be used only for the managed vif */ 245 /* Currently, MAC ID 0 should be used only for the managed/IBSS vif */
246 if (vif->type != NL80211_IFTYPE_STATION || vif->p2p) 246 switch (vif->type) {
247 case NL80211_IFTYPE_ADHOC:
248 break;
249 case NL80211_IFTYPE_STATION:
250 if (!vif->p2p)
251 break;
252 /* fall through */
253 default:
247 __clear_bit(0, data.available_mac_ids); 254 __clear_bit(0, data.available_mac_ids);
255 }
248 256
249 ieee80211_iterate_active_interfaces_atomic( 257 ieee80211_iterate_active_interfaces_atomic(
250 mvm->hw, IEEE80211_IFACE_ITER_RESUME_ALL, 258 mvm->hw, IEEE80211_IFACE_ITER_RESUME_ALL,
@@ -302,9 +310,9 @@ static int iwl_mvm_mac_ctxt_allocate_resources(struct iwl_mvm *mvm,
302 /* Find available queues, and allocate them to the ACs */ 310 /* Find available queues, and allocate them to the ACs */
303 for (ac = 0; ac < IEEE80211_NUM_ACS; ac++) { 311 for (ac = 0; ac < IEEE80211_NUM_ACS; ac++) {
304 u8 queue = find_first_zero_bit(data.used_hw_queues, 312 u8 queue = find_first_zero_bit(data.used_hw_queues,
305 IWL_MVM_FIRST_AGG_QUEUE); 313 mvm->first_agg_queue);
306 314
307 if (queue >= IWL_MVM_FIRST_AGG_QUEUE) { 315 if (queue >= mvm->first_agg_queue) {
308 IWL_ERR(mvm, "Failed to allocate queue\n"); 316 IWL_ERR(mvm, "Failed to allocate queue\n");
309 ret = -EIO; 317 ret = -EIO;
310 goto exit_fail; 318 goto exit_fail;
@@ -317,9 +325,9 @@ static int iwl_mvm_mac_ctxt_allocate_resources(struct iwl_mvm *mvm,
317 /* Allocate the CAB queue for softAP and GO interfaces */ 325 /* Allocate the CAB queue for softAP and GO interfaces */
318 if (vif->type == NL80211_IFTYPE_AP) { 326 if (vif->type == NL80211_IFTYPE_AP) {
319 u8 queue = find_first_zero_bit(data.used_hw_queues, 327 u8 queue = find_first_zero_bit(data.used_hw_queues,
320 IWL_MVM_FIRST_AGG_QUEUE); 328 mvm->first_agg_queue);
321 329
322 if (queue >= IWL_MVM_FIRST_AGG_QUEUE) { 330 if (queue >= mvm->first_agg_queue) {
323 IWL_ERR(mvm, "Failed to allocate cab queue\n"); 331 IWL_ERR(mvm, "Failed to allocate cab queue\n");
324 ret = -EIO; 332 ret = -EIO;
325 goto exit_fail; 333 goto exit_fail;
@@ -559,8 +567,12 @@ static void iwl_mvm_mac_ctxt_cmd_common(struct iwl_mvm *mvm,
559 cmd->qos_flags |= cpu_to_le32(MAC_QOS_FLG_UPDATE_EDCA); 567 cmd->qos_flags |= cpu_to_le32(MAC_QOS_FLG_UPDATE_EDCA);
560 568
561 /* Don't use cts to self as the fw doesn't support it currently. */ 569 /* Don't use cts to self as the fw doesn't support it currently. */
562 if (vif->bss_conf.use_cts_prot) 570 if (vif->bss_conf.use_cts_prot) {
563 cmd->protection_flags |= cpu_to_le32(MAC_PROT_FLG_TGG_PROTECT); 571 cmd->protection_flags |= cpu_to_le32(MAC_PROT_FLG_TGG_PROTECT);
572 if (IWL_UCODE_API(mvm->fw->ucode_ver) >= 8)
573 cmd->protection_flags |=
574 cpu_to_le32(MAC_PROT_FLG_SELF_CTS_EN);
575 }
564 576
565 /* 577 /*
566 * I think that we should enable these 2 flags regardless the HT PROT 578 * I think that we should enable these 2 flags regardless the HT PROT
@@ -707,8 +719,35 @@ static int iwl_mvm_mac_ctxt_cmd_listener(struct iwl_mvm *mvm,
707 cmd.filter_flags = cpu_to_le32(MAC_FILTER_IN_PROMISC | 719 cmd.filter_flags = cpu_to_le32(MAC_FILTER_IN_PROMISC |
708 MAC_FILTER_IN_CONTROL_AND_MGMT | 720 MAC_FILTER_IN_CONTROL_AND_MGMT |
709 MAC_FILTER_IN_BEACON | 721 MAC_FILTER_IN_BEACON |
722 MAC_FILTER_IN_PROBE_REQUEST |
723 MAC_FILTER_IN_CRC32);
724 mvm->hw->flags |= IEEE80211_HW_RX_INCLUDES_FCS;
725
726 return iwl_mvm_mac_ctxt_send_cmd(mvm, &cmd);
727}
728
729static int iwl_mvm_mac_ctxt_cmd_ibss(struct iwl_mvm *mvm,
730 struct ieee80211_vif *vif,
731 u32 action)
732{
733 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif);
734 struct iwl_mac_ctx_cmd cmd = {};
735
736 WARN_ON(vif->type != NL80211_IFTYPE_ADHOC);
737
738 iwl_mvm_mac_ctxt_cmd_common(mvm, vif, &cmd, action);
739
740 cmd.filter_flags = cpu_to_le32(MAC_FILTER_IN_BEACON |
710 MAC_FILTER_IN_PROBE_REQUEST); 741 MAC_FILTER_IN_PROBE_REQUEST);
711 742
743 /* cmd.ibss.beacon_time/cmd.ibss.beacon_tsf are curently ignored */
744 cmd.ibss.bi = cpu_to_le32(vif->bss_conf.beacon_int);
745 cmd.ibss.bi_reciprocal =
746 cpu_to_le32(iwl_mvm_reciprocal(vif->bss_conf.beacon_int));
747
748 /* TODO: Assumes that the beacon id == mac context id */
749 cmd.ibss.beacon_template = cpu_to_le32(mvmvif->id);
750
712 return iwl_mvm_mac_ctxt_send_cmd(mvm, &cmd); 751 return iwl_mvm_mac_ctxt_send_cmd(mvm, &cmd);
713} 752}
714 753
@@ -721,7 +760,8 @@ static void iwl_mvm_go_iterator(void *_data, u8 *mac, struct ieee80211_vif *vif)
721 struct iwl_mvm_go_iterator_data *data = _data; 760 struct iwl_mvm_go_iterator_data *data = _data;
722 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 761 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif);
723 762
724 if (vif->type == NL80211_IFTYPE_AP && vif->p2p && mvmvif->ap_active) 763 if (vif->type == NL80211_IFTYPE_AP && vif->p2p &&
764 mvmvif->ap_ibss_active)
725 data->go_active = true; 765 data->go_active = true;
726} 766}
727 767
@@ -833,9 +873,10 @@ static int iwl_mvm_mac_ctxt_send_beacon(struct iwl_mvm *mvm,
833 cpu_to_le32(iwl_mvm_mac80211_idx_to_hwrate(rate)); 873 cpu_to_le32(iwl_mvm_mac80211_idx_to_hwrate(rate));
834 874
835 /* Set up TX beacon command fields */ 875 /* Set up TX beacon command fields */
836 iwl_mvm_mac_ctxt_set_tim(mvm, &beacon_cmd, 876 if (vif->type == NL80211_IFTYPE_AP)
837 beacon->data, 877 iwl_mvm_mac_ctxt_set_tim(mvm, &beacon_cmd,
838 beacon_skb_len); 878 beacon->data,
879 beacon_skb_len);
839 880
840 /* Submit command */ 881 /* Submit command */
841 cmd.len[0] = sizeof(beacon_cmd); 882 cmd.len[0] = sizeof(beacon_cmd);
@@ -848,14 +889,15 @@ static int iwl_mvm_mac_ctxt_send_beacon(struct iwl_mvm *mvm,
848 return iwl_mvm_send_cmd(mvm, &cmd); 889 return iwl_mvm_send_cmd(mvm, &cmd);
849} 890}
850 891
851/* The beacon template for the AP/GO context has changed and needs update */ 892/* The beacon template for the AP/GO/IBSS has changed and needs update */
852int iwl_mvm_mac_ctxt_beacon_changed(struct iwl_mvm *mvm, 893int iwl_mvm_mac_ctxt_beacon_changed(struct iwl_mvm *mvm,
853 struct ieee80211_vif *vif) 894 struct ieee80211_vif *vif)
854{ 895{
855 struct sk_buff *beacon; 896 struct sk_buff *beacon;
856 int ret; 897 int ret;
857 898
858 WARN_ON(vif->type != NL80211_IFTYPE_AP); 899 WARN_ON(vif->type != NL80211_IFTYPE_AP &&
900 vif->type != NL80211_IFTYPE_ADHOC);
859 901
860 beacon = ieee80211_beacon_get(mvm->hw, vif); 902 beacon = ieee80211_beacon_get(mvm->hw, vif);
861 if (!beacon) 903 if (!beacon)
@@ -1018,6 +1060,8 @@ static int iwl_mvm_mac_ctx_send(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
1018 return iwl_mvm_mac_ctxt_cmd_listener(mvm, vif, action); 1060 return iwl_mvm_mac_ctxt_cmd_listener(mvm, vif, action);
1019 case NL80211_IFTYPE_P2P_DEVICE: 1061 case NL80211_IFTYPE_P2P_DEVICE:
1020 return iwl_mvm_mac_ctxt_cmd_p2p_device(mvm, vif, action); 1062 return iwl_mvm_mac_ctxt_cmd_p2p_device(mvm, vif, action);
1063 case NL80211_IFTYPE_ADHOC:
1064 return iwl_mvm_mac_ctxt_cmd_ibss(mvm, vif, action);
1021 default: 1065 default:
1022 break; 1066 break;
1023 } 1067 }
@@ -1038,6 +1082,9 @@ int iwl_mvm_mac_ctxt_add(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
1038 if (ret) 1082 if (ret)
1039 return ret; 1083 return ret;
1040 1084
1085 /* will only do anything at resume from D3 time */
1086 iwl_mvm_set_last_nonqos_seq(mvm, vif);
1087
1041 mvmvif->uploaded = true; 1088 mvmvif->uploaded = true;
1042 return 0; 1089 return 0;
1043} 1090}
@@ -1077,6 +1124,10 @@ int iwl_mvm_mac_ctxt_remove(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
1077 } 1124 }
1078 1125
1079 mvmvif->uploaded = false; 1126 mvmvif->uploaded = false;
1127
1128 if (vif->type == NL80211_IFTYPE_MONITOR)
1129 mvm->hw->flags &= ~IEEE80211_HW_RX_INCLUDES_FCS;
1130
1080 return 0; 1131 return 0;
1081} 1132}
1082 1133
diff --git a/drivers/net/wireless/iwlwifi/mvm/mac80211.c b/drivers/net/wireless/iwlwifi/mvm/mac80211.c
index 9833cdf6177c..74bc2c8af06d 100644
--- a/drivers/net/wireless/iwlwifi/mvm/mac80211.c
+++ b/drivers/net/wireless/iwlwifi/mvm/mac80211.c
@@ -77,6 +77,7 @@
77#include "iwl-eeprom-parse.h" 77#include "iwl-eeprom-parse.h"
78#include "fw-api-scan.h" 78#include "fw-api-scan.h"
79#include "iwl-phy-db.h" 79#include "iwl-phy-db.h"
80#include "testmode.h"
80 81
81static const struct ieee80211_iface_limit iwl_mvm_limits[] = { 82static const struct ieee80211_iface_limit iwl_mvm_limits[] = {
82 { 83 {
@@ -138,6 +139,14 @@ static void iwl_mvm_reset_phy_ctxts(struct iwl_mvm *mvm)
138 } 139 }
139} 140}
140 141
142static int iwl_mvm_max_scan_ie_len(struct iwl_mvm *mvm)
143{
144 /* we create the 802.11 header and SSID element */
145 if (mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_NO_BASIC_SSID)
146 return mvm->fw->ucode_capa.max_probe_length - 24 - 2;
147 return mvm->fw->ucode_capa.max_probe_length - 24 - 34;
148}
149
141int iwl_mvm_mac_setup_register(struct iwl_mvm *mvm) 150int iwl_mvm_mac_setup_register(struct iwl_mvm *mvm)
142{ 151{
143 struct ieee80211_hw *hw = mvm->hw; 152 struct ieee80211_hw *hw = mvm->hw;
@@ -155,10 +164,9 @@ int iwl_mvm_mac_setup_register(struct iwl_mvm *mvm)
155 IEEE80211_HW_TIMING_BEACON_ONLY | 164 IEEE80211_HW_TIMING_BEACON_ONLY |
156 IEEE80211_HW_CONNECTION_MONITOR | 165 IEEE80211_HW_CONNECTION_MONITOR |
157 IEEE80211_HW_SUPPORTS_DYNAMIC_SMPS | 166 IEEE80211_HW_SUPPORTS_DYNAMIC_SMPS |
158 IEEE80211_HW_SUPPORTS_STATIC_SMPS | 167 IEEE80211_HW_SUPPORTS_STATIC_SMPS;
159 IEEE80211_HW_SUPPORTS_UAPSD;
160 168
161 hw->queues = IWL_MVM_FIRST_AGG_QUEUE; 169 hw->queues = mvm->first_agg_queue;
162 hw->offchannel_tx_hw_queue = IWL_MVM_OFFCHANNEL_QUEUE; 170 hw->offchannel_tx_hw_queue = IWL_MVM_OFFCHANNEL_QUEUE;
163 hw->rate_control_algorithm = "iwl-mvm-rs"; 171 hw->rate_control_algorithm = "iwl-mvm-rs";
164 172
@@ -171,6 +179,12 @@ int iwl_mvm_mac_setup_register(struct iwl_mvm *mvm)
171 !iwlwifi_mod_params.sw_crypto) 179 !iwlwifi_mod_params.sw_crypto)
172 hw->flags |= IEEE80211_HW_MFP_CAPABLE; 180 hw->flags |= IEEE80211_HW_MFP_CAPABLE;
173 181
182 if (mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_UAPSD_SUPPORT) {
183 hw->flags |= IEEE80211_HW_SUPPORTS_UAPSD;
184 hw->uapsd_queues = IWL_UAPSD_AC_INFO;
185 hw->uapsd_max_sp_len = IWL_UAPSD_MAX_SP;
186 }
187
174 hw->sta_data_size = sizeof(struct iwl_mvm_sta); 188 hw->sta_data_size = sizeof(struct iwl_mvm_sta);
175 hw->vif_data_size = sizeof(struct iwl_mvm_vif); 189 hw->vif_data_size = sizeof(struct iwl_mvm_vif);
176 hw->chanctx_data_size = sizeof(u16); 190 hw->chanctx_data_size = sizeof(u16);
@@ -181,6 +195,10 @@ int iwl_mvm_mac_setup_register(struct iwl_mvm *mvm)
181 BIT(NL80211_IFTYPE_P2P_GO) | 195 BIT(NL80211_IFTYPE_P2P_GO) |
182 BIT(NL80211_IFTYPE_P2P_DEVICE); 196 BIT(NL80211_IFTYPE_P2P_DEVICE);
183 197
198 /* IBSS has bugs in older versions */
199 if (IWL_UCODE_API(mvm->fw->ucode_ver) >= 8)
200 hw->wiphy->interface_modes |= BIT(NL80211_IFTYPE_ADHOC);
201
184 hw->wiphy->flags |= WIPHY_FLAG_CUSTOM_REGULATORY | 202 hw->wiphy->flags |= WIPHY_FLAG_CUSTOM_REGULATORY |
185 WIPHY_FLAG_DISABLE_BEACON_HINTS | 203 WIPHY_FLAG_DISABLE_BEACON_HINTS |
186 WIPHY_FLAG_IBSS_RSN; 204 WIPHY_FLAG_IBSS_RSN;
@@ -191,8 +209,6 @@ int iwl_mvm_mac_setup_register(struct iwl_mvm *mvm)
191 209
192 hw->wiphy->max_remain_on_channel_duration = 10000; 210 hw->wiphy->max_remain_on_channel_duration = 10000;
193 hw->max_listen_interval = IWL_CONN_MAX_LISTEN_INTERVAL; 211 hw->max_listen_interval = IWL_CONN_MAX_LISTEN_INTERVAL;
194 hw->uapsd_queues = IWL_UAPSD_AC_INFO;
195 hw->uapsd_max_sp_len = IWL_UAPSD_MAX_SP;
196 212
197 /* Extract MAC address */ 213 /* Extract MAC address */
198 memcpy(mvm->addresses[0].addr, mvm->nvm_data->hw_addr, ETH_ALEN); 214 memcpy(mvm->addresses[0].addr, mvm->nvm_data->hw_addr, ETH_ALEN);
@@ -212,9 +228,8 @@ int iwl_mvm_mac_setup_register(struct iwl_mvm *mvm)
212 228
213 iwl_mvm_reset_phy_ctxts(mvm); 229 iwl_mvm_reset_phy_ctxts(mvm);
214 230
215 /* we create the 802.11 header and a max-length SSID element */ 231 hw->wiphy->max_scan_ie_len = iwl_mvm_max_scan_ie_len(mvm);
216 hw->wiphy->max_scan_ie_len = 232
217 mvm->fw->ucode_capa.max_probe_length - 24 - 34;
218 hw->wiphy->max_scan_ssids = PROBE_OPTION_MAX; 233 hw->wiphy->max_scan_ssids = PROBE_OPTION_MAX;
219 234
220 if (mvm->nvm_data->bands[IEEE80211_BAND_2GHZ].n_channels) 235 if (mvm->nvm_data->bands[IEEE80211_BAND_2GHZ].n_channels)
@@ -231,6 +246,15 @@ int iwl_mvm_mac_setup_register(struct iwl_mvm *mvm)
231 else 246 else
232 hw->wiphy->flags &= ~WIPHY_FLAG_PS_ON_BY_DEFAULT; 247 hw->wiphy->flags &= ~WIPHY_FLAG_PS_ON_BY_DEFAULT;
233 248
249 if (mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_SCHED_SCAN) {
250 hw->wiphy->flags |= WIPHY_FLAG_SUPPORTS_SCHED_SCAN;
251 hw->wiphy->max_sched_scan_ssids = PROBE_OPTION_MAX;
252 hw->wiphy->max_match_sets = IWL_SCAN_MAX_PROFILES;
253 /* we create the 802.11 header and zero length SSID IE. */
254 hw->wiphy->max_sched_scan_ie_len =
255 SCAN_OFFLOAD_PROBE_REQ_SIZE - 24 - 2;
256 }
257
234 hw->wiphy->features |= NL80211_FEATURE_P2P_GO_CTWIN | 258 hw->wiphy->features |= NL80211_FEATURE_P2P_GO_CTWIN |
235 NL80211_FEATURE_P2P_GO_OPPPS; 259 NL80211_FEATURE_P2P_GO_OPPPS;
236 260
@@ -548,7 +572,8 @@ static int iwl_mvm_mac_add_interface(struct ieee80211_hw *hw,
548 * In short: there's not much we can do at this point, other than 572 * In short: there's not much we can do at this point, other than
549 * allocating resources :) 573 * allocating resources :)
550 */ 574 */
551 if (vif->type == NL80211_IFTYPE_AP) { 575 if (vif->type == NL80211_IFTYPE_AP ||
576 vif->type == NL80211_IFTYPE_ADHOC) {
552 u32 qmask = iwl_mvm_mac_get_queues_mask(mvm, vif); 577 u32 qmask = iwl_mvm_mac_get_queues_mask(mvm, vif);
553 ret = iwl_mvm_allocate_int_sta(mvm, &mvmvif->bcast_sta, 578 ret = iwl_mvm_allocate_int_sta(mvm, &mvmvif->bcast_sta,
554 qmask); 579 qmask);
@@ -698,7 +723,14 @@ static void iwl_mvm_mac_remove_interface(struct ieee80211_hw *hw,
698 * For AP/GO interface, the tear down of the resources allocated to the 723 * For AP/GO interface, the tear down of the resources allocated to the
699 * interface is be handled as part of the stop_ap flow. 724 * interface is be handled as part of the stop_ap flow.
700 */ 725 */
701 if (vif->type == NL80211_IFTYPE_AP) { 726 if (vif->type == NL80211_IFTYPE_AP ||
727 vif->type == NL80211_IFTYPE_ADHOC) {
728#ifdef CONFIG_NL80211_TESTMODE
729 if (vif == mvm->noa_vif) {
730 mvm->noa_vif = NULL;
731 mvm->noa_duration = 0;
732 }
733#endif
702 iwl_mvm_dealloc_int_sta(mvm, &mvmvif->bcast_sta); 734 iwl_mvm_dealloc_int_sta(mvm, &mvmvif->bcast_sta);
703 goto out_release; 735 goto out_release;
704 } 736 }
@@ -796,6 +828,27 @@ static void iwl_mvm_bss_info_changed_station(struct iwl_mvm *mvm,
796 return; 828 return;
797 } 829 }
798 iwl_mvm_configure_mcast_filter(mvm, vif); 830 iwl_mvm_configure_mcast_filter(mvm, vif);
831
832 if (test_bit(IWL_MVM_STATUS_IN_HW_RESTART,
833 &mvm->status)) {
834 /*
835 * If we're restarting then the firmware will
836 * obviously have lost synchronisation with
837 * the AP. It will attempt to synchronise by
838 * itself, but we can make it more reliable by
839 * scheduling a session protection time event.
840 *
841 * The firmware needs to receive a beacon to
842 * catch up with synchronisation, use 110% of
843 * the beacon interval.
844 *
845 * Set a large maximum delay to allow for more
846 * than a single interface.
847 */
848 u32 dur = (11 * vif->bss_conf.beacon_int) / 10;
849 iwl_mvm_protect_session(mvm, vif, dur, dur,
850 5 * dur);
851 }
799 } else if (mvmvif->ap_sta_id != IWL_MVM_STATION_COUNT) { 852 } else if (mvmvif->ap_sta_id != IWL_MVM_STATION_COUNT) {
800 /* remove AP station now that the MAC is unassoc */ 853 /* remove AP station now that the MAC is unassoc */
801 ret = iwl_mvm_rm_sta_id(mvm, vif, mvmvif->ap_sta_id); 854 ret = iwl_mvm_rm_sta_id(mvm, vif, mvmvif->ap_sta_id);
@@ -811,7 +864,8 @@ static void iwl_mvm_bss_info_changed_station(struct iwl_mvm *mvm,
811 /* reset rssi values */ 864 /* reset rssi values */
812 mvmvif->bf_data.ave_beacon_signal = 0; 865 mvmvif->bf_data.ave_beacon_signal = 0;
813 866
814 if (!(mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_UAPSD)) { 867 if (!(mvm->fw->ucode_capa.flags &
868 IWL_UCODE_TLV_FLAGS_PM_CMD_SUPPORT)) {
815 /* Workaround for FW bug, otherwise FW disables device 869 /* Workaround for FW bug, otherwise FW disables device
816 * power save upon disassociation 870 * power save upon disassociation
817 */ 871 */
@@ -819,7 +873,7 @@ static void iwl_mvm_bss_info_changed_station(struct iwl_mvm *mvm,
819 if (ret) 873 if (ret)
820 IWL_ERR(mvm, "failed to update power mode\n"); 874 IWL_ERR(mvm, "failed to update power mode\n");
821 } 875 }
822 iwl_mvm_bt_coex_vif_assoc(mvm, vif); 876 iwl_mvm_bt_coex_vif_change(mvm);
823 } else if (changes & BSS_CHANGED_BEACON_INFO) { 877 } else if (changes & BSS_CHANGED_BEACON_INFO) {
824 /* 878 /*
825 * We received a beacon _after_ association so 879 * We received a beacon _after_ association so
@@ -848,7 +902,8 @@ static void iwl_mvm_bss_info_changed_station(struct iwl_mvm *mvm,
848 } 902 }
849} 903}
850 904
851static int iwl_mvm_start_ap(struct ieee80211_hw *hw, struct ieee80211_vif *vif) 905static int iwl_mvm_start_ap_ibss(struct ieee80211_hw *hw,
906 struct ieee80211_vif *vif)
852{ 907{
853 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 908 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw);
854 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 909 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif);
@@ -871,7 +926,7 @@ static int iwl_mvm_start_ap(struct ieee80211_hw *hw, struct ieee80211_vif *vif)
871 if (ret) 926 if (ret)
872 goto out_remove; 927 goto out_remove;
873 928
874 mvmvif->ap_active = true; 929 mvmvif->ap_ibss_active = true;
875 930
876 /* Send the bcast station. At this stage the TBTT and DTIM time events 931 /* Send the bcast station. At this stage the TBTT and DTIM time events
877 * are added and applied to the scheduler */ 932 * are added and applied to the scheduler */
@@ -883,10 +938,12 @@ static int iwl_mvm_start_ap(struct ieee80211_hw *hw, struct ieee80211_vif *vif)
883 if (ret) 938 if (ret)
884 goto out_rm_bcast; 939 goto out_rm_bcast;
885 940
886 /* Need to update the P2P Device MAC */ 941 /* Need to update the P2P Device MAC (only GO, IBSS is single vif) */
887 if (vif->p2p && mvm->p2p_device_vif) 942 if (vif->p2p && mvm->p2p_device_vif)
888 iwl_mvm_mac_ctxt_changed(mvm, mvm->p2p_device_vif); 943 iwl_mvm_mac_ctxt_changed(mvm, mvm->p2p_device_vif);
889 944
945 iwl_mvm_bt_coex_vif_change(mvm);
946
890 mutex_unlock(&mvm->mutex); 947 mutex_unlock(&mvm->mutex);
891 return 0; 948 return 0;
892 949
@@ -901,7 +958,8 @@ out_unlock:
901 return ret; 958 return ret;
902} 959}
903 960
904static void iwl_mvm_stop_ap(struct ieee80211_hw *hw, struct ieee80211_vif *vif) 961static void iwl_mvm_stop_ap_ibss(struct ieee80211_hw *hw,
962 struct ieee80211_vif *vif)
905{ 963{
906 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 964 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw);
907 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 965 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif);
@@ -910,9 +968,11 @@ static void iwl_mvm_stop_ap(struct ieee80211_hw *hw, struct ieee80211_vif *vif)
910 968
911 mutex_lock(&mvm->mutex); 969 mutex_lock(&mvm->mutex);
912 970
913 mvmvif->ap_active = false; 971 mvmvif->ap_ibss_active = false;
972
973 iwl_mvm_bt_coex_vif_change(mvm);
914 974
915 /* Need to update the P2P Device MAC */ 975 /* Need to update the P2P Device MAC (only GO, IBSS is single vif) */
916 if (vif->p2p && mvm->p2p_device_vif) 976 if (vif->p2p && mvm->p2p_device_vif)
917 iwl_mvm_mac_ctxt_changed(mvm, mvm->p2p_device_vif); 977 iwl_mvm_mac_ctxt_changed(mvm, mvm->p2p_device_vif);
918 978
@@ -924,10 +984,11 @@ static void iwl_mvm_stop_ap(struct ieee80211_hw *hw, struct ieee80211_vif *vif)
924 mutex_unlock(&mvm->mutex); 984 mutex_unlock(&mvm->mutex);
925} 985}
926 986
927static void iwl_mvm_bss_info_changed_ap(struct iwl_mvm *mvm, 987static void
928 struct ieee80211_vif *vif, 988iwl_mvm_bss_info_changed_ap_ibss(struct iwl_mvm *mvm,
929 struct ieee80211_bss_conf *bss_conf, 989 struct ieee80211_vif *vif,
930 u32 changes) 990 struct ieee80211_bss_conf *bss_conf,
991 u32 changes)
931{ 992{
932 /* Need to send a new beacon template to the FW */ 993 /* Need to send a new beacon template to the FW */
933 if (changes & BSS_CHANGED_BEACON) { 994 if (changes & BSS_CHANGED_BEACON) {
@@ -950,7 +1011,8 @@ static void iwl_mvm_bss_info_changed(struct ieee80211_hw *hw,
950 iwl_mvm_bss_info_changed_station(mvm, vif, bss_conf, changes); 1011 iwl_mvm_bss_info_changed_station(mvm, vif, bss_conf, changes);
951 break; 1012 break;
952 case NL80211_IFTYPE_AP: 1013 case NL80211_IFTYPE_AP:
953 iwl_mvm_bss_info_changed_ap(mvm, vif, bss_conf, changes); 1014 case NL80211_IFTYPE_ADHOC:
1015 iwl_mvm_bss_info_changed_ap_ibss(mvm, vif, bss_conf, changes);
954 break; 1016 break;
955 default: 1017 default:
956 /* shouldn't happen */ 1018 /* shouldn't happen */
@@ -1163,7 +1225,54 @@ static void iwl_mvm_mac_mgd_prepare_tx(struct ieee80211_hw *hw,
1163 1225
1164 mutex_lock(&mvm->mutex); 1226 mutex_lock(&mvm->mutex);
1165 /* Try really hard to protect the session and hear a beacon */ 1227 /* Try really hard to protect the session and hear a beacon */
1166 iwl_mvm_protect_session(mvm, vif, duration, min_duration); 1228 iwl_mvm_protect_session(mvm, vif, duration, min_duration, 500);
1229 mutex_unlock(&mvm->mutex);
1230}
1231
1232static int iwl_mvm_mac_sched_scan_start(struct ieee80211_hw *hw,
1233 struct ieee80211_vif *vif,
1234 struct cfg80211_sched_scan_request *req,
1235 struct ieee80211_sched_scan_ies *ies)
1236{
1237 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw);
1238 int ret;
1239
1240 mutex_lock(&mvm->mutex);
1241
1242 if (mvm->scan_status != IWL_MVM_SCAN_NONE) {
1243 IWL_DEBUG_SCAN(mvm,
1244 "SCHED SCAN request during internal scan - abort\n");
1245 ret = -EBUSY;
1246 goto out;
1247 }
1248
1249 mvm->scan_status = IWL_MVM_SCAN_SCHED;
1250
1251 ret = iwl_mvm_config_sched_scan(mvm, vif, req, ies);
1252 if (ret)
1253 goto err;
1254
1255 ret = iwl_mvm_config_sched_scan_profiles(mvm, req);
1256 if (ret)
1257 goto err;
1258
1259 ret = iwl_mvm_sched_scan_start(mvm, req);
1260 if (!ret)
1261 goto out;
1262err:
1263 mvm->scan_status = IWL_MVM_SCAN_NONE;
1264out:
1265 mutex_unlock(&mvm->mutex);
1266 return ret;
1267}
1268
1269static void iwl_mvm_mac_sched_scan_stop(struct ieee80211_hw *hw,
1270 struct ieee80211_vif *vif)
1271{
1272 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw);
1273
1274 mutex_lock(&mvm->mutex);
1275 iwl_mvm_sched_scan_stop(mvm);
1167 mutex_unlock(&mvm->mutex); 1276 mutex_unlock(&mvm->mutex);
1168} 1277}
1169 1278
@@ -1207,8 +1316,13 @@ static int iwl_mvm_mac_set_key(struct ieee80211_hw *hw,
1207 1316
1208 switch (cmd) { 1317 switch (cmd) {
1209 case SET_KEY: 1318 case SET_KEY:
1210 if (vif->type == NL80211_IFTYPE_AP && !sta) { 1319 if ((vif->type == NL80211_IFTYPE_ADHOC ||
1211 /* GTK on AP interface is a TX-only key, return 0 */ 1320 vif->type == NL80211_IFTYPE_AP) && !sta) {
1321 /*
1322 * GTK on AP interface is a TX-only key, return 0;
1323 * on IBSS they're per-station and because we're lazy
1324 * we don't support them for RX, so do the same.
1325 */
1212 ret = 0; 1326 ret = 0;
1213 key->hw_key_idx = STA_KEY_IDX_INVALID; 1327 key->hw_key_idx = STA_KEY_IDX_INVALID;
1214 break; 1328 break;
@@ -1252,6 +1366,9 @@ static void iwl_mvm_mac_update_tkip_key(struct ieee80211_hw *hw,
1252{ 1366{
1253 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 1367 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw);
1254 1368
1369 if (keyconf->hw_key_idx == STA_KEY_IDX_INVALID)
1370 return;
1371
1255 iwl_mvm_update_tkip_key(mvm, vif, keyconf, sta, iv32, phase1key); 1372 iwl_mvm_update_tkip_key(mvm, vif, keyconf, sta, iv32, phase1key);
1256} 1373}
1257 1374
@@ -1445,6 +1562,7 @@ static void iwl_mvm_change_chanctx(struct ieee80211_hw *hw,
1445 iwl_mvm_phy_ctxt_changed(mvm, phy_ctxt, &ctx->def, 1562 iwl_mvm_phy_ctxt_changed(mvm, phy_ctxt, &ctx->def,
1446 ctx->rx_chains_static, 1563 ctx->rx_chains_static,
1447 ctx->rx_chains_dynamic); 1564 ctx->rx_chains_dynamic);
1565 iwl_mvm_bt_coex_vif_change(mvm);
1448 mutex_unlock(&mvm->mutex); 1566 mutex_unlock(&mvm->mutex);
1449} 1567}
1450 1568
@@ -1464,14 +1582,14 @@ static int iwl_mvm_assign_vif_chanctx(struct ieee80211_hw *hw,
1464 1582
1465 switch (vif->type) { 1583 switch (vif->type) {
1466 case NL80211_IFTYPE_AP: 1584 case NL80211_IFTYPE_AP:
1585 case NL80211_IFTYPE_ADHOC:
1467 /* 1586 /*
1468 * The AP binding flow is handled as part of the start_ap flow 1587 * The AP binding flow is handled as part of the start_ap flow
1469 * (in bss_info_changed). 1588 * (in bss_info_changed), similarly for IBSS.
1470 */ 1589 */
1471 ret = 0; 1590 ret = 0;
1472 goto out_unlock; 1591 goto out_unlock;
1473 case NL80211_IFTYPE_STATION: 1592 case NL80211_IFTYPE_STATION:
1474 case NL80211_IFTYPE_ADHOC:
1475 case NL80211_IFTYPE_MONITOR: 1593 case NL80211_IFTYPE_MONITOR:
1476 break; 1594 break;
1477 default: 1595 default:
@@ -1517,10 +1635,10 @@ static void iwl_mvm_unassign_vif_chanctx(struct ieee80211_hw *hw,
1517 1635
1518 iwl_mvm_remove_time_event(mvm, mvmvif, &mvmvif->time_event_data); 1636 iwl_mvm_remove_time_event(mvm, mvmvif, &mvmvif->time_event_data);
1519 1637
1520 if (vif->type == NL80211_IFTYPE_AP)
1521 goto out_unlock;
1522
1523 switch (vif->type) { 1638 switch (vif->type) {
1639 case NL80211_IFTYPE_AP:
1640 case NL80211_IFTYPE_ADHOC:
1641 goto out_unlock;
1524 case NL80211_IFTYPE_MONITOR: 1642 case NL80211_IFTYPE_MONITOR:
1525 mvmvif->monitor_active = false; 1643 mvmvif->monitor_active = false;
1526 iwl_mvm_update_quotas(mvm, NULL); 1644 iwl_mvm_update_quotas(mvm, NULL);
@@ -1550,14 +1668,72 @@ static int iwl_mvm_set_tim(struct ieee80211_hw *hw,
1550 return iwl_mvm_mac_ctxt_beacon_changed(mvm, mvm_sta->vif); 1668 return iwl_mvm_mac_ctxt_beacon_changed(mvm, mvm_sta->vif);
1551} 1669}
1552 1670
1553static void iwl_mvm_mac_rssi_callback(struct ieee80211_hw *hw, 1671#ifdef CONFIG_NL80211_TESTMODE
1672static const struct nla_policy iwl_mvm_tm_policy[IWL_MVM_TM_ATTR_MAX + 1] = {
1673 [IWL_MVM_TM_ATTR_CMD] = { .type = NLA_U32 },
1674 [IWL_MVM_TM_ATTR_NOA_DURATION] = { .type = NLA_U32 },
1675 [IWL_MVM_TM_ATTR_BEACON_FILTER_STATE] = { .type = NLA_U32 },
1676};
1677
1678static int __iwl_mvm_mac_testmode_cmd(struct iwl_mvm *mvm,
1554 struct ieee80211_vif *vif, 1679 struct ieee80211_vif *vif,
1555 enum ieee80211_rssi_event rssi_event) 1680 void *data, int len)
1681{
1682 struct nlattr *tb[IWL_MVM_TM_ATTR_MAX + 1];
1683 int err;
1684 u32 noa_duration;
1685
1686 err = nla_parse(tb, IWL_MVM_TM_ATTR_MAX, data, len, iwl_mvm_tm_policy);
1687 if (err)
1688 return err;
1689
1690 if (!tb[IWL_MVM_TM_ATTR_CMD])
1691 return -EINVAL;
1692
1693 switch (nla_get_u32(tb[IWL_MVM_TM_ATTR_CMD])) {
1694 case IWL_MVM_TM_CMD_SET_NOA:
1695 if (!vif || vif->type != NL80211_IFTYPE_AP || !vif->p2p ||
1696 !vif->bss_conf.enable_beacon ||
1697 !tb[IWL_MVM_TM_ATTR_NOA_DURATION])
1698 return -EINVAL;
1699
1700 noa_duration = nla_get_u32(tb[IWL_MVM_TM_ATTR_NOA_DURATION]);
1701 if (noa_duration >= vif->bss_conf.beacon_int)
1702 return -EINVAL;
1703
1704 mvm->noa_duration = noa_duration;
1705 mvm->noa_vif = vif;
1706
1707 return iwl_mvm_update_quotas(mvm, NULL);
1708 case IWL_MVM_TM_CMD_SET_BEACON_FILTER:
1709 /* must be associated client vif - ignore authorized */
1710 if (!vif || vif->type != NL80211_IFTYPE_STATION ||
1711 !vif->bss_conf.assoc || !vif->bss_conf.dtim_period ||
1712 !tb[IWL_MVM_TM_ATTR_BEACON_FILTER_STATE])
1713 return -EINVAL;
1714
1715 if (nla_get_u32(tb[IWL_MVM_TM_ATTR_BEACON_FILTER_STATE]))
1716 return iwl_mvm_enable_beacon_filter(mvm, vif);
1717 return iwl_mvm_disable_beacon_filter(mvm, vif);
1718 }
1719
1720 return -EOPNOTSUPP;
1721}
1722
1723static int iwl_mvm_mac_testmode_cmd(struct ieee80211_hw *hw,
1724 struct ieee80211_vif *vif,
1725 void *data, int len)
1556{ 1726{
1557 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw); 1727 struct iwl_mvm *mvm = IWL_MAC80211_GET_MVM(hw);
1728 int err;
1558 1729
1559 iwl_mvm_bt_rssi_event(mvm, vif, rssi_event); 1730 mutex_lock(&mvm->mutex);
1731 err = __iwl_mvm_mac_testmode_cmd(mvm, vif, data, len);
1732 mutex_unlock(&mvm->mutex);
1733
1734 return err;
1560} 1735}
1736#endif
1561 1737
1562struct ieee80211_ops iwl_mvm_hw_ops = { 1738struct ieee80211_ops iwl_mvm_hw_ops = {
1563 .tx = iwl_mvm_mac_tx, 1739 .tx = iwl_mvm_mac_tx,
@@ -1578,23 +1754,27 @@ struct ieee80211_ops iwl_mvm_hw_ops = {
1578 .set_rts_threshold = iwl_mvm_mac_set_rts_threshold, 1754 .set_rts_threshold = iwl_mvm_mac_set_rts_threshold,
1579 .conf_tx = iwl_mvm_mac_conf_tx, 1755 .conf_tx = iwl_mvm_mac_conf_tx,
1580 .mgd_prepare_tx = iwl_mvm_mac_mgd_prepare_tx, 1756 .mgd_prepare_tx = iwl_mvm_mac_mgd_prepare_tx,
1757 .sched_scan_start = iwl_mvm_mac_sched_scan_start,
1758 .sched_scan_stop = iwl_mvm_mac_sched_scan_stop,
1581 .set_key = iwl_mvm_mac_set_key, 1759 .set_key = iwl_mvm_mac_set_key,
1582 .update_tkip_key = iwl_mvm_mac_update_tkip_key, 1760 .update_tkip_key = iwl_mvm_mac_update_tkip_key,
1583 .remain_on_channel = iwl_mvm_roc, 1761 .remain_on_channel = iwl_mvm_roc,
1584 .cancel_remain_on_channel = iwl_mvm_cancel_roc, 1762 .cancel_remain_on_channel = iwl_mvm_cancel_roc,
1585 .rssi_callback = iwl_mvm_mac_rssi_callback,
1586
1587 .add_chanctx = iwl_mvm_add_chanctx, 1763 .add_chanctx = iwl_mvm_add_chanctx,
1588 .remove_chanctx = iwl_mvm_remove_chanctx, 1764 .remove_chanctx = iwl_mvm_remove_chanctx,
1589 .change_chanctx = iwl_mvm_change_chanctx, 1765 .change_chanctx = iwl_mvm_change_chanctx,
1590 .assign_vif_chanctx = iwl_mvm_assign_vif_chanctx, 1766 .assign_vif_chanctx = iwl_mvm_assign_vif_chanctx,
1591 .unassign_vif_chanctx = iwl_mvm_unassign_vif_chanctx, 1767 .unassign_vif_chanctx = iwl_mvm_unassign_vif_chanctx,
1592 1768
1593 .start_ap = iwl_mvm_start_ap, 1769 .start_ap = iwl_mvm_start_ap_ibss,
1594 .stop_ap = iwl_mvm_stop_ap, 1770 .stop_ap = iwl_mvm_stop_ap_ibss,
1771 .join_ibss = iwl_mvm_start_ap_ibss,
1772 .leave_ibss = iwl_mvm_stop_ap_ibss,
1595 1773
1596 .set_tim = iwl_mvm_set_tim, 1774 .set_tim = iwl_mvm_set_tim,
1597 1775
1776 CFG80211_TESTMODE_CMD(iwl_mvm_mac_testmode_cmd)
1777
1598#ifdef CONFIG_PM_SLEEP 1778#ifdef CONFIG_PM_SLEEP
1599 /* look at d3.c */ 1779 /* look at d3.c */
1600 .suspend = iwl_mvm_suspend, 1780 .suspend = iwl_mvm_suspend,
diff --git a/drivers/net/wireless/iwlwifi/mvm/mvm.h b/drivers/net/wireless/iwlwifi/mvm/mvm.h
index b0389279cc1e..fed21ef4162d 100644
--- a/drivers/net/wireless/iwlwifi/mvm/mvm.h
+++ b/drivers/net/wireless/iwlwifi/mvm/mvm.h
@@ -73,7 +73,6 @@
73#include "iwl-trans.h" 73#include "iwl-trans.h"
74#include "iwl-notif-wait.h" 74#include "iwl-notif-wait.h"
75#include "iwl-eeprom-parse.h" 75#include "iwl-eeprom-parse.h"
76#include "iwl-trans.h"
77#include "sta.h" 76#include "sta.h"
78#include "fw-api.h" 77#include "fw-api.h"
79#include "constants.h" 78#include "constants.h"
@@ -162,6 +161,7 @@ enum iwl_power_scheme {
162struct iwl_mvm_power_ops { 161struct iwl_mvm_power_ops {
163 int (*power_update_mode)(struct iwl_mvm *mvm, 162 int (*power_update_mode)(struct iwl_mvm *mvm,
164 struct ieee80211_vif *vif); 163 struct ieee80211_vif *vif);
164 int (*power_update_device_mode)(struct iwl_mvm *mvm);
165 int (*power_disable)(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 165 int (*power_disable)(struct iwl_mvm *mvm, struct ieee80211_vif *vif);
166#ifdef CONFIG_IWLWIFI_DEBUGFS 166#ifdef CONFIG_IWLWIFI_DEBUGFS
167 int (*power_dbgfs_read)(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 167 int (*power_dbgfs_read)(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
@@ -241,12 +241,18 @@ enum iwl_mvm_smps_type_request {
241* @last_beacon_signal: last beacon rssi signal in dbm 241* @last_beacon_signal: last beacon rssi signal in dbm
242* @ave_beacon_signal: average beacon signal 242* @ave_beacon_signal: average beacon signal
243* @last_cqm_event: rssi of the last cqm event 243* @last_cqm_event: rssi of the last cqm event
244* @bt_coex_min_thold: minimum threshold for BT coex
245* @bt_coex_max_thold: maximum threshold for BT coex
246* @last_bt_coex_event: rssi of the last BT coex event
244*/ 247*/
245struct iwl_mvm_vif_bf_data { 248struct iwl_mvm_vif_bf_data {
246 bool bf_enabled; 249 bool bf_enabled;
247 bool ba_enabled; 250 bool ba_enabled;
248 s8 ave_beacon_signal; 251 s8 ave_beacon_signal;
249 s8 last_cqm_event; 252 s8 last_cqm_event;
253 s8 bt_coex_min_thold;
254 s8 bt_coex_max_thold;
255 s8 last_bt_coex_event;
250}; 256};
251 257
252/** 258/**
@@ -255,8 +261,8 @@ struct iwl_mvm_vif_bf_data {
255 * @color: to solve races upon MAC addition and removal 261 * @color: to solve races upon MAC addition and removal
256 * @ap_sta_id: the sta_id of the AP - valid only if VIF type is STA 262 * @ap_sta_id: the sta_id of the AP - valid only if VIF type is STA
257 * @uploaded: indicates the MAC context has been added to the device 263 * @uploaded: indicates the MAC context has been added to the device
258 * @ap_active: indicates that ap context is configured, and that the interface 264 * @ap_ibss_active: indicates that AP/IBSS is configured and that the interface
259 * should get quota etc. 265 * should get quota etc.
260 * @monitor_active: indicates that monitor context is configured, and that the 266 * @monitor_active: indicates that monitor context is configured, and that the
261 * interface should get quota etc. 267 * interface should get quota etc.
262 * @queue_params: QoS params for this MAC 268 * @queue_params: QoS params for this MAC
@@ -272,7 +278,7 @@ struct iwl_mvm_vif {
272 u8 ap_sta_id; 278 u8 ap_sta_id;
273 279
274 bool uploaded; 280 bool uploaded;
275 bool ap_active; 281 bool ap_ibss_active;
276 bool monitor_active; 282 bool monitor_active;
277 struct iwl_mvm_vif_bf_data bf_data; 283 struct iwl_mvm_vif_bf_data bf_data;
278 284
@@ -306,6 +312,9 @@ struct iwl_mvm_vif {
306 312
307 int tx_key_idx; 313 int tx_key_idx;
308 314
315 bool seqno_valid;
316 u16 seqno;
317
309#if IS_ENABLED(CONFIG_IPV6) 318#if IS_ENABLED(CONFIG_IPV6)
310 /* IPv6 addresses for WoWLAN */ 319 /* IPv6 addresses for WoWLAN */
311 struct in6_addr target_ipv6_addrs[IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_MAX]; 320 struct in6_addr target_ipv6_addrs[IWL_PROTO_OFFLOAD_NUM_IPV6_ADDRS_MAX];
@@ -333,6 +342,7 @@ iwl_mvm_vif_from_mac80211(struct ieee80211_vif *vif)
333enum iwl_scan_status { 342enum iwl_scan_status {
334 IWL_MVM_SCAN_NONE, 343 IWL_MVM_SCAN_NONE,
335 IWL_MVM_SCAN_OS, 344 IWL_MVM_SCAN_OS,
345 IWL_MVM_SCAN_SCHED,
336}; 346};
337 347
338/** 348/**
@@ -434,7 +444,7 @@ struct iwl_mvm {
434 444
435 enum iwl_ucode_type cur_ucode; 445 enum iwl_ucode_type cur_ucode;
436 bool ucode_loaded; 446 bool ucode_loaded;
437 bool init_ucode_run; 447 bool init_ucode_complete;
438 u32 error_event_table; 448 u32 error_event_table;
439 u32 log_event_table; 449 u32 log_event_table;
440 450
@@ -470,6 +480,9 @@ struct iwl_mvm {
470 enum iwl_scan_status scan_status; 480 enum iwl_scan_status scan_status;
471 struct iwl_scan_cmd *scan_cmd; 481 struct iwl_scan_cmd *scan_cmd;
472 482
483 /* rx chain antennas set through debugfs for the scan command */
484 u8 scan_rx_ant;
485
473 /* Internal station */ 486 /* Internal station */
474 struct iwl_mvm_int_sta aux_sta; 487 struct iwl_mvm_int_sta aux_sta;
475 488
@@ -479,7 +492,8 @@ struct iwl_mvm {
479#ifdef CONFIG_IWLWIFI_DEBUGFS 492#ifdef CONFIG_IWLWIFI_DEBUGFS
480 struct dentry *debugfs_dir; 493 struct dentry *debugfs_dir;
481 u32 dbgfs_sram_offset, dbgfs_sram_len; 494 u32 dbgfs_sram_offset, dbgfs_sram_len;
482 bool prevent_power_down_d3; 495 bool disable_power_off;
496 bool disable_power_off_d3;
483#endif 497#endif
484 498
485 struct iwl_mvm_phy_ctxt phy_ctxts[NUM_PHY_CTX]; 499 struct iwl_mvm_phy_ctxt phy_ctxts[NUM_PHY_CTX];
@@ -523,12 +537,23 @@ struct iwl_mvm {
523 /* BT-Coex */ 537 /* BT-Coex */
524 u8 bt_kill_msk; 538 u8 bt_kill_msk;
525 struct iwl_bt_coex_profile_notif last_bt_notif; 539 struct iwl_bt_coex_profile_notif last_bt_notif;
540 struct iwl_bt_coex_ci_cmd last_bt_ci_cmd;
526 541
527 /* Thermal Throttling and CTkill */ 542 /* Thermal Throttling and CTkill */
528 struct iwl_mvm_tt_mgmt thermal_throttle; 543 struct iwl_mvm_tt_mgmt thermal_throttle;
529 s32 temperature; /* Celsius */ 544 s32 temperature; /* Celsius */
530 545
531 const struct iwl_mvm_power_ops *pm_ops; 546 const struct iwl_mvm_power_ops *pm_ops;
547
548#ifdef CONFIG_NL80211_TESTMODE
549 u32 noa_duration;
550 struct ieee80211_vif *noa_vif;
551#endif
552
553 /* Tx queues */
554 u8 aux_queue;
555 u8 first_agg_queue;
556 u8 last_agg_queue;
532}; 557};
533 558
534/* Extract MVM priv from op_mode and _hw */ 559/* Extract MVM priv from op_mode and _hw */
@@ -570,6 +595,9 @@ int iwl_run_init_mvm_ucode(struct iwl_mvm *mvm, bool read_nvm);
570/* Utils */ 595/* Utils */
571int iwl_mvm_legacy_rate_to_mac80211_idx(u32 rate_n_flags, 596int iwl_mvm_legacy_rate_to_mac80211_idx(u32 rate_n_flags,
572 enum ieee80211_band band); 597 enum ieee80211_band band);
598void iwl_mvm_hwrate_to_tx_rate(u32 rate_n_flags,
599 enum ieee80211_band band,
600 struct ieee80211_tx_rate *r);
573u8 iwl_mvm_mac80211_idx_to_hwrate(int rate_idx); 601u8 iwl_mvm_mac80211_idx_to_hwrate(int rate_idx);
574void iwl_mvm_dump_nic_error_log(struct iwl_mvm *mvm); 602void iwl_mvm_dump_nic_error_log(struct iwl_mvm *mvm);
575void iwl_mvm_dump_sram(struct iwl_mvm *mvm); 603void iwl_mvm_dump_sram(struct iwl_mvm *mvm);
@@ -608,6 +636,7 @@ int iwl_mvm_rx_statistics(struct iwl_mvm *mvm,
608 636
609/* NVM */ 637/* NVM */
610int iwl_nvm_init(struct iwl_mvm *mvm); 638int iwl_nvm_init(struct iwl_mvm *mvm);
639int iwl_mvm_load_nvm_to_nic(struct iwl_mvm *mvm);
611 640
612int iwl_mvm_up(struct iwl_mvm *mvm); 641int iwl_mvm_up(struct iwl_mvm *mvm);
613int iwl_mvm_load_d3_fw(struct iwl_mvm *mvm); 642int iwl_mvm_load_d3_fw(struct iwl_mvm *mvm);
@@ -682,6 +711,23 @@ int iwl_mvm_rx_scan_complete(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb,
682 struct iwl_device_cmd *cmd); 711 struct iwl_device_cmd *cmd);
683void iwl_mvm_cancel_scan(struct iwl_mvm *mvm); 712void iwl_mvm_cancel_scan(struct iwl_mvm *mvm);
684 713
714/* Scheduled scan */
715int iwl_mvm_rx_scan_offload_complete_notif(struct iwl_mvm *mvm,
716 struct iwl_rx_cmd_buffer *rxb,
717 struct iwl_device_cmd *cmd);
718int iwl_mvm_config_sched_scan(struct iwl_mvm *mvm,
719 struct ieee80211_vif *vif,
720 struct cfg80211_sched_scan_request *req,
721 struct ieee80211_sched_scan_ies *ies);
722int iwl_mvm_config_sched_scan_profiles(struct iwl_mvm *mvm,
723 struct cfg80211_sched_scan_request *req);
724int iwl_mvm_sched_scan_start(struct iwl_mvm *mvm,
725 struct cfg80211_sched_scan_request *req);
726void iwl_mvm_sched_scan_stop(struct iwl_mvm *mvm);
727int iwl_mvm_rx_sched_scan_results(struct iwl_mvm *mvm,
728 struct iwl_rx_cmd_buffer *rxb,
729 struct iwl_device_cmd *cmd);
730
685/* MVM debugfs */ 731/* MVM debugfs */
686#ifdef CONFIG_IWLWIFI_DEBUGFS 732#ifdef CONFIG_IWLWIFI_DEBUGFS
687int iwl_mvm_dbgfs_register(struct iwl_mvm *mvm, struct dentry *dbgfs_dir); 733int iwl_mvm_dbgfs_register(struct iwl_mvm *mvm, struct dentry *dbgfs_dir);
@@ -720,6 +766,13 @@ static inline int iwl_mvm_power_disable(struct iwl_mvm *mvm,
720 return mvm->pm_ops->power_disable(mvm, vif); 766 return mvm->pm_ops->power_disable(mvm, vif);
721} 767}
722 768
769static inline int iwl_mvm_power_update_device_mode(struct iwl_mvm *mvm)
770{
771 if (mvm->pm_ops->power_update_device_mode)
772 return mvm->pm_ops->power_update_device_mode(mvm);
773 return 0;
774}
775
723#ifdef CONFIG_IWLWIFI_DEBUGFS 776#ifdef CONFIG_IWLWIFI_DEBUGFS
724static inline int iwl_mvm_power_dbgfs_read(struct iwl_mvm *mvm, 777static inline int iwl_mvm_power_dbgfs_read(struct iwl_mvm *mvm,
725 struct ieee80211_vif *vif, 778 struct ieee80211_vif *vif,
@@ -745,6 +798,15 @@ void iwl_mvm_ipv6_addr_change(struct ieee80211_hw *hw,
745void iwl_mvm_set_default_unicast_key(struct ieee80211_hw *hw, 798void iwl_mvm_set_default_unicast_key(struct ieee80211_hw *hw,
746 struct ieee80211_vif *vif, int idx); 799 struct ieee80211_vif *vif, int idx);
747extern const struct file_operations iwl_dbgfs_d3_test_ops; 800extern const struct file_operations iwl_dbgfs_d3_test_ops;
801#ifdef CONFIG_PM_SLEEP
802void iwl_mvm_set_last_nonqos_seq(struct iwl_mvm *mvm,
803 struct ieee80211_vif *vif);
804#else
805static inline void
806iwl_mvm_set_last_nonqos_seq(struct iwl_mvm *mvm, struct ieee80211_vif *vif)
807{
808}
809#endif
748 810
749/* BT Coex */ 811/* BT Coex */
750int iwl_send_bt_prio_tbl(struct iwl_mvm *mvm); 812int iwl_send_bt_prio_tbl(struct iwl_mvm *mvm);
@@ -754,7 +816,20 @@ int iwl_mvm_rx_bt_coex_notif(struct iwl_mvm *mvm,
754 struct iwl_device_cmd *cmd); 816 struct iwl_device_cmd *cmd);
755void iwl_mvm_bt_rssi_event(struct iwl_mvm *mvm, struct ieee80211_vif *vif, 817void iwl_mvm_bt_rssi_event(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
756 enum ieee80211_rssi_event rssi_event); 818 enum ieee80211_rssi_event rssi_event);
757void iwl_mvm_bt_coex_vif_assoc(struct iwl_mvm *mvm, struct ieee80211_vif *vif); 819void iwl_mvm_bt_coex_vif_change(struct iwl_mvm *mvm);
820u16 iwl_mvm_bt_coex_agg_time_limit(struct iwl_mvm *mvm,
821 struct ieee80211_sta *sta);
822bool iwl_mvm_bt_coex_is_mimo_allowed(struct iwl_mvm *mvm,
823 struct ieee80211_sta *sta);
824
825enum iwl_bt_kill_msk {
826 BT_KILL_MSK_DEFAULT,
827 BT_KILL_MSK_SCO_HID_A2DP,
828 BT_KILL_MSK_REDUCED_TXPOW,
829 BT_KILL_MSK_MAX,
830};
831extern const u32 iwl_bt_ack_kill_msk[BT_KILL_MSK_MAX];
832extern const u32 iwl_bt_cts_kill_msk[BT_KILL_MSK_MAX];
758 833
759/* beacon filtering */ 834/* beacon filtering */
760#ifdef CONFIG_IWLWIFI_DEBUGFS 835#ifdef CONFIG_IWLWIFI_DEBUGFS
diff --git a/drivers/net/wireless/iwlwifi/mvm/nvm.c b/drivers/net/wireless/iwlwifi/mvm/nvm.c
index edb94ea31654..2beffd028b67 100644
--- a/drivers/net/wireless/iwlwifi/mvm/nvm.c
+++ b/drivers/net/wireless/iwlwifi/mvm/nvm.c
@@ -77,7 +77,7 @@ static const int nvm_to_read[] = {
77 77
78/* Default NVM size to read */ 78/* Default NVM size to read */
79#define IWL_NVM_DEFAULT_CHUNK_SIZE (2*1024) 79#define IWL_NVM_DEFAULT_CHUNK_SIZE (2*1024)
80#define IWL_MAX_NVM_SECTION_SIZE 6000 80#define IWL_MAX_NVM_SECTION_SIZE 7000
81 81
82#define NVM_WRITE_OPCODE 1 82#define NVM_WRITE_OPCODE 1
83#define NVM_READ_OPCODE 0 83#define NVM_READ_OPCODE 0
@@ -259,6 +259,8 @@ iwl_parse_nvm_sections(struct iwl_mvm *mvm)
259#define MAX_NVM_FILE_LEN 16384 259#define MAX_NVM_FILE_LEN 16384
260 260
261/* 261/*
262 * Reads external NVM from a file into mvm->nvm_sections
263 *
262 * HOW TO CREATE THE NVM FILE FORMAT: 264 * HOW TO CREATE THE NVM FILE FORMAT:
263 * ------------------------------ 265 * ------------------------------
264 * 1. create hex file, format: 266 * 1. create hex file, format:
@@ -277,20 +279,23 @@ iwl_parse_nvm_sections(struct iwl_mvm *mvm)
277 * 279 *
278 * 4. save as "iNVM_xxx.bin" under /lib/firmware 280 * 4. save as "iNVM_xxx.bin" under /lib/firmware
279 */ 281 */
280static int iwl_mvm_load_external_nvm(struct iwl_mvm *mvm) 282static int iwl_mvm_read_external_nvm(struct iwl_mvm *mvm)
281{ 283{
282 int ret, section_id, section_size; 284 int ret, section_size;
285 u16 section_id;
283 const struct firmware *fw_entry; 286 const struct firmware *fw_entry;
284 const struct { 287 const struct {
285 __le16 word1; 288 __le16 word1;
286 __le16 word2; 289 __le16 word2;
287 u8 data[]; 290 u8 data[];
288 } *file_sec; 291 } *file_sec;
289 const u8 *eof; 292 const u8 *eof, *temp;
290 293
291#define NVM_WORD1_LEN(x) (8 * (x & 0x03FF)) 294#define NVM_WORD1_LEN(x) (8 * (x & 0x03FF))
292#define NVM_WORD2_ID(x) (x >> 12) 295#define NVM_WORD2_ID(x) (x >> 12)
293 296
297 IWL_DEBUG_EEPROM(mvm->trans->dev, "Read from external NVM\n");
298
294 /* 299 /*
295 * Obtain NVM image via request_firmware. Since we already used 300 * Obtain NVM image via request_firmware. Since we already used
296 * request_firmware_nowait() for the firmware binary load and only 301 * request_firmware_nowait() for the firmware binary load and only
@@ -362,12 +367,18 @@ static int iwl_mvm_load_external_nvm(struct iwl_mvm *mvm)
362 break; 367 break;
363 } 368 }
364 369
365 ret = iwl_nvm_write_section(mvm, section_id, file_sec->data, 370 temp = kmemdup(file_sec->data, section_size, GFP_KERNEL);
366 section_size); 371 if (!temp) {
367 if (ret < 0) { 372 ret = -ENOMEM;
368 IWL_ERR(mvm, "iwl_mvm_send_cmd failed: %d\n", ret); 373 break;
374 }
375 if (WARN_ON(section_id >= NVM_NUM_OF_SECTIONS)) {
376 IWL_ERR(mvm, "Invalid NVM section ID\n");
377 ret = -EINVAL;
369 break; 378 break;
370 } 379 }
380 mvm->nvm_sections[section_id].data = temp;
381 mvm->nvm_sections[section_id].length = section_size;
371 382
372 /* advance to the next section */ 383 /* advance to the next section */
373 file_sec = (void *)(file_sec->data + section_size); 384 file_sec = (void *)(file_sec->data + section_size);
@@ -377,6 +388,28 @@ out:
377 return ret; 388 return ret;
378} 389}
379 390
391/* Loads the NVM data stored in mvm->nvm_sections into the NIC */
392int iwl_mvm_load_nvm_to_nic(struct iwl_mvm *mvm)
393{
394 int i, ret;
395 u16 section_id;
396 struct iwl_nvm_section *sections = mvm->nvm_sections;
397
398 IWL_DEBUG_EEPROM(mvm->trans->dev, "'Write to NVM\n");
399
400 for (i = 0; i < ARRAY_SIZE(nvm_to_read); i++) {
401 section_id = nvm_to_read[i];
402 ret = iwl_nvm_write_section(mvm, section_id,
403 sections[section_id].data,
404 sections[section_id].length);
405 if (ret < 0) {
406 IWL_ERR(mvm, "iwl_mvm_send_cmd failed: %d\n", ret);
407 break;
408 }
409 }
410 return ret;
411}
412
380int iwl_nvm_init(struct iwl_mvm *mvm) 413int iwl_nvm_init(struct iwl_mvm *mvm)
381{ 414{
382 int ret, i, section; 415 int ret, i, section;
@@ -385,36 +418,36 @@ int iwl_nvm_init(struct iwl_mvm *mvm)
385 /* load external NVM if configured */ 418 /* load external NVM if configured */
386 if (iwlwifi_mod_params.nvm_file) { 419 if (iwlwifi_mod_params.nvm_file) {
387 /* move to External NVM flow */ 420 /* move to External NVM flow */
388 ret = iwl_mvm_load_external_nvm(mvm); 421 ret = iwl_mvm_read_external_nvm(mvm);
389 if (ret) 422 if (ret)
390 return ret; 423 return ret;
391 } 424 } else {
392 425 /* Read From FW NVM */
393 /* Read From FW NVM */ 426 IWL_DEBUG_EEPROM(mvm->trans->dev, "Read from NVM\n");
394 IWL_DEBUG_EEPROM(mvm->trans->dev, "Read from NVM\n"); 427
395 428 /* TODO: find correct NVM max size for a section */
396 /* TODO: find correct NVM max size for a section */ 429 nvm_buffer = kmalloc(mvm->cfg->base_params->eeprom_size,
397 nvm_buffer = kmalloc(mvm->cfg->base_params->eeprom_size, 430 GFP_KERNEL);
398 GFP_KERNEL); 431 if (!nvm_buffer)
399 if (!nvm_buffer) 432 return -ENOMEM;
400 return -ENOMEM; 433 for (i = 0; i < ARRAY_SIZE(nvm_to_read); i++) {
401 for (i = 0; i < ARRAY_SIZE(nvm_to_read); i++) { 434 section = nvm_to_read[i];
402 section = nvm_to_read[i]; 435 /* we override the constness for initial read */
403 /* we override the constness for initial read */ 436 ret = iwl_nvm_read_section(mvm, section, nvm_buffer);
404 ret = iwl_nvm_read_section(mvm, section, nvm_buffer); 437 if (ret < 0)
405 if (ret < 0) 438 break;
406 break; 439 temp = kmemdup(nvm_buffer, ret, GFP_KERNEL);
407 temp = kmemdup(nvm_buffer, ret, GFP_KERNEL); 440 if (!temp) {
408 if (!temp) { 441 ret = -ENOMEM;
409 ret = -ENOMEM; 442 break;
410 break; 443 }
444 mvm->nvm_sections[section].data = temp;
445 mvm->nvm_sections[section].length = ret;
411 } 446 }
412 mvm->nvm_sections[section].data = temp; 447 kfree(nvm_buffer);
413 mvm->nvm_sections[section].length = ret; 448 if (ret < 0)
449 return ret;
414 } 450 }
415 kfree(nvm_buffer);
416 if (ret < 0)
417 return ret;
418 451
419 mvm->nvm_data = iwl_parse_nvm_sections(mvm); 452 mvm->nvm_data = iwl_parse_nvm_sections(mvm);
420 if (!mvm->nvm_data) 453 if (!mvm->nvm_data)
diff --git a/drivers/net/wireless/iwlwifi/mvm/ops.c b/drivers/net/wireless/iwlwifi/mvm/ops.c
index 2fcc8ef88a68..d86083c6f445 100644
--- a/drivers/net/wireless/iwlwifi/mvm/ops.c
+++ b/drivers/net/wireless/iwlwifi/mvm/ops.c
@@ -224,6 +224,10 @@ static const struct iwl_rx_handlers iwl_mvm_rx_handlers[] = {
224 224
225 RX_HANDLER(SCAN_REQUEST_CMD, iwl_mvm_rx_scan_response, false), 225 RX_HANDLER(SCAN_REQUEST_CMD, iwl_mvm_rx_scan_response, false),
226 RX_HANDLER(SCAN_COMPLETE_NOTIFICATION, iwl_mvm_rx_scan_complete, false), 226 RX_HANDLER(SCAN_COMPLETE_NOTIFICATION, iwl_mvm_rx_scan_complete, false),
227 RX_HANDLER(SCAN_OFFLOAD_COMPLETE,
228 iwl_mvm_rx_scan_offload_complete_notif, false),
229 RX_HANDLER(MATCH_FOUND_NOTIFICATION, iwl_mvm_rx_sched_scan_results,
230 false),
227 231
228 RX_HANDLER(RADIO_VERSION_NOTIFICATION, iwl_mvm_rx_radio_ver, false), 232 RX_HANDLER(RADIO_VERSION_NOTIFICATION, iwl_mvm_rx_radio_ver, false),
229 RX_HANDLER(CARD_STATE_NOTIFICATION, iwl_mvm_rx_card_state_notif, false), 233 RX_HANDLER(CARD_STATE_NOTIFICATION, iwl_mvm_rx_card_state_notif, false),
@@ -249,6 +253,7 @@ static const char *iwl_mvm_cmd_strings[REPLY_MAX] = {
249 CMD(TIME_EVENT_NOTIFICATION), 253 CMD(TIME_EVENT_NOTIFICATION),
250 CMD(BINDING_CONTEXT_CMD), 254 CMD(BINDING_CONTEXT_CMD),
251 CMD(TIME_QUOTA_CMD), 255 CMD(TIME_QUOTA_CMD),
256 CMD(NON_QOS_TX_COUNTER_CMD),
252 CMD(RADIO_VERSION_NOTIFICATION), 257 CMD(RADIO_VERSION_NOTIFICATION),
253 CMD(SCAN_REQUEST_CMD), 258 CMD(SCAN_REQUEST_CMD),
254 CMD(SCAN_ABORT_CMD), 259 CMD(SCAN_ABORT_CMD),
@@ -260,10 +265,12 @@ static const char *iwl_mvm_cmd_strings[REPLY_MAX] = {
260 CMD(CALIB_RES_NOTIF_PHY_DB), 265 CMD(CALIB_RES_NOTIF_PHY_DB),
261 CMD(SET_CALIB_DEFAULT_CMD), 266 CMD(SET_CALIB_DEFAULT_CMD),
262 CMD(CALIBRATION_COMPLETE_NOTIFICATION), 267 CMD(CALIBRATION_COMPLETE_NOTIFICATION),
268 CMD(ADD_STA_KEY),
263 CMD(ADD_STA), 269 CMD(ADD_STA),
264 CMD(REMOVE_STA), 270 CMD(REMOVE_STA),
265 CMD(LQ_CMD), 271 CMD(LQ_CMD),
266 CMD(SCAN_OFFLOAD_CONFIG_CMD), 272 CMD(SCAN_OFFLOAD_CONFIG_CMD),
273 CMD(MATCH_FOUND_NOTIFICATION),
267 CMD(SCAN_OFFLOAD_REQUEST_CMD), 274 CMD(SCAN_OFFLOAD_REQUEST_CMD),
268 CMD(SCAN_OFFLOAD_ABORT_CMD), 275 CMD(SCAN_OFFLOAD_ABORT_CMD),
269 CMD(SCAN_OFFLOAD_COMPLETE), 276 CMD(SCAN_OFFLOAD_COMPLETE),
@@ -303,6 +310,7 @@ static const char *iwl_mvm_cmd_strings[REPLY_MAX] = {
303 CMD(REPLY_BEACON_FILTERING_CMD), 310 CMD(REPLY_BEACON_FILTERING_CMD),
304 CMD(REPLY_THERMAL_MNG_BACKOFF), 311 CMD(REPLY_THERMAL_MNG_BACKOFF),
305 CMD(MAC_PM_POWER_TABLE), 312 CMD(MAC_PM_POWER_TABLE),
313 CMD(BT_COEX_CI),
306}; 314};
307#undef CMD 315#undef CMD
308 316
@@ -344,6 +352,14 @@ iwl_op_mode_mvm_start(struct iwl_trans *trans, const struct iwl_cfg *cfg,
344 352
345 mvm->restart_fw = iwlwifi_mod_params.restart_fw ? -1 : 0; 353 mvm->restart_fw = iwlwifi_mod_params.restart_fw ? -1 : 0;
346 354
355 mvm->aux_queue = 15;
356 mvm->first_agg_queue = 16;
357 mvm->last_agg_queue = mvm->cfg->base_params->num_of_queues - 1;
358 if (mvm->cfg->base_params->num_of_queues == 16) {
359 mvm->aux_queue = 11;
360 mvm->first_agg_queue = 12;
361 }
362
347 mutex_init(&mvm->mutex); 363 mutex_init(&mvm->mutex);
348 spin_lock_init(&mvm->async_handlers_lock); 364 spin_lock_init(&mvm->async_handlers_lock);
349 INIT_LIST_HEAD(&mvm->time_event_list); 365 INIT_LIST_HEAD(&mvm->time_event_list);
@@ -401,24 +417,32 @@ iwl_op_mode_mvm_start(struct iwl_trans *trans, const struct iwl_cfg *cfg,
401 IWL_INFO(mvm, "Detected %s, REV=0x%X\n", 417 IWL_INFO(mvm, "Detected %s, REV=0x%X\n",
402 mvm->cfg->name, mvm->trans->hw_rev); 418 mvm->cfg->name, mvm->trans->hw_rev);
403 419
404 err = iwl_trans_start_hw(mvm->trans);
405 if (err)
406 goto out_free;
407
408 iwl_mvm_tt_initialize(mvm); 420 iwl_mvm_tt_initialize(mvm);
409 421
410 mutex_lock(&mvm->mutex); 422 /*
411 err = iwl_run_init_mvm_ucode(mvm, true); 423 * If the NVM exists in an external file,
412 mutex_unlock(&mvm->mutex); 424 * there is no need to unnecessarily power up the NIC at driver load
413 /* returns 0 if successful, 1 if success but in rfkill */ 425 */
414 if (err < 0 && !iwlmvm_mod_params.init_dbg) { 426 if (iwlwifi_mod_params.nvm_file) {
415 IWL_ERR(mvm, "Failed to run INIT ucode: %d\n", err); 427 iwl_nvm_init(mvm);
416 goto out_free; 428 } else {
417 } 429 err = iwl_trans_start_hw(mvm->trans);
430 if (err)
431 goto out_free;
432
433 mutex_lock(&mvm->mutex);
434 err = iwl_run_init_mvm_ucode(mvm, true);
435 mutex_unlock(&mvm->mutex);
436 /* returns 0 if successful, 1 if success but in rfkill */
437 if (err < 0 && !iwlmvm_mod_params.init_dbg) {
438 IWL_ERR(mvm, "Failed to run INIT ucode: %d\n", err);
439 goto out_free;
440 }
418 441
419 /* Stop the hw after the ALIVE and NVM has been read */ 442 /* Stop the hw after the ALIVE and NVM has been read */
420 if (!iwlmvm_mod_params.init_dbg) 443 if (!iwlmvm_mod_params.init_dbg)
421 iwl_trans_stop_hw(mvm->trans, false); 444 iwl_trans_stop_hw(mvm->trans, false);
445 }
422 446
423 scan_size = sizeof(struct iwl_scan_cmd) + 447 scan_size = sizeof(struct iwl_scan_cmd) +
424 mvm->fw->ucode_capa.max_probe_length + 448 mvm->fw->ucode_capa.max_probe_length +
@@ -435,7 +459,7 @@ iwl_op_mode_mvm_start(struct iwl_trans *trans, const struct iwl_cfg *cfg,
435 if (err) 459 if (err)
436 goto out_unregister; 460 goto out_unregister;
437 461
438 if (mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_UAPSD) 462 if (mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_PM_CMD_SUPPORT)
439 mvm->pm_ops = &pm_mac_ops; 463 mvm->pm_ops = &pm_mac_ops;
440 else 464 else
441 mvm->pm_ops = &pm_legacy_ops; 465 mvm->pm_ops = &pm_legacy_ops;
@@ -449,7 +473,8 @@ iwl_op_mode_mvm_start(struct iwl_trans *trans, const struct iwl_cfg *cfg,
449 out_free: 473 out_free:
450 iwl_phy_db_free(mvm->phy_db); 474 iwl_phy_db_free(mvm->phy_db);
451 kfree(mvm->scan_cmd); 475 kfree(mvm->scan_cmd);
452 iwl_trans_stop_hw(trans, true); 476 if (!iwlwifi_mod_params.nvm_file)
477 iwl_trans_stop_hw(trans, true);
453 ieee80211_free_hw(mvm->hw); 478 ieee80211_free_hw(mvm->hw);
454 return NULL; 479 return NULL;
455} 480}
@@ -715,6 +740,9 @@ static void iwl_mvm_nic_restart(struct iwl_mvm *mvm)
715 case IWL_MVM_SCAN_OS: 740 case IWL_MVM_SCAN_OS:
716 ieee80211_scan_completed(mvm->hw, true); 741 ieee80211_scan_completed(mvm->hw, true);
717 break; 742 break;
743 case IWL_MVM_SCAN_SCHED:
744 ieee80211_sched_scan_stopped(mvm->hw);
745 break;
718 } 746 }
719 747
720 if (mvm->restart_fw > 0) 748 if (mvm->restart_fw > 0)
diff --git a/drivers/net/wireless/iwlwifi/mvm/power.c b/drivers/net/wireless/iwlwifi/mvm/power.c
index d58e393324ef..550824aa84ea 100644
--- a/drivers/net/wireless/iwlwifi/mvm/power.c
+++ b/drivers/net/wireless/iwlwifi/mvm/power.c
@@ -300,11 +300,6 @@ static void iwl_mvm_power_build_cmd(struct iwl_mvm *mvm,
300 } 300 }
301 301
302 if (cmd->flags & cpu_to_le16(POWER_FLAGS_ADVANCE_PM_ENA_MSK)) { 302 if (cmd->flags & cpu_to_le16(POWER_FLAGS_ADVANCE_PM_ENA_MSK)) {
303 cmd->rx_data_timeout_uapsd =
304 cpu_to_le32(IWL_MVM_UAPSD_RX_DATA_TIMEOUT);
305 cmd->tx_data_timeout_uapsd =
306 cpu_to_le32(IWL_MVM_UAPSD_TX_DATA_TIMEOUT);
307
308 if (cmd->uapsd_ac_flags == (BIT(IEEE80211_AC_VO) | 303 if (cmd->uapsd_ac_flags == (BIT(IEEE80211_AC_VO) |
309 BIT(IEEE80211_AC_VI) | 304 BIT(IEEE80211_AC_VI) |
310 BIT(IEEE80211_AC_BE) | 305 BIT(IEEE80211_AC_BE) |
@@ -319,10 +314,31 @@ static void iwl_mvm_power_build_cmd(struct iwl_mvm *mvm,
319 } 314 }
320 315
321 cmd->uapsd_max_sp = IWL_UAPSD_MAX_SP; 316 cmd->uapsd_max_sp = IWL_UAPSD_MAX_SP;
322 cmd->heavy_tx_thld_packets = 317
323 IWL_MVM_PS_HEAVY_TX_THLD_PACKETS; 318 if (mvm->cur_ucode == IWL_UCODE_WOWLAN || cmd->flags &
324 cmd->heavy_rx_thld_packets = 319 cpu_to_le16(POWER_FLAGS_SNOOZE_ENA_MSK)) {
325 IWL_MVM_PS_HEAVY_RX_THLD_PACKETS; 320 cmd->rx_data_timeout_uapsd =
321 cpu_to_le32(IWL_MVM_WOWLAN_PS_RX_DATA_TIMEOUT);
322 cmd->tx_data_timeout_uapsd =
323 cpu_to_le32(IWL_MVM_WOWLAN_PS_TX_DATA_TIMEOUT);
324 } else {
325 cmd->rx_data_timeout_uapsd =
326 cpu_to_le32(IWL_MVM_UAPSD_RX_DATA_TIMEOUT);
327 cmd->tx_data_timeout_uapsd =
328 cpu_to_le32(IWL_MVM_UAPSD_TX_DATA_TIMEOUT);
329 }
330
331 if (cmd->flags & cpu_to_le16(POWER_FLAGS_SNOOZE_ENA_MSK)) {
332 cmd->heavy_tx_thld_packets =
333 IWL_MVM_PS_SNOOZE_HEAVY_TX_THLD_PACKETS;
334 cmd->heavy_rx_thld_packets =
335 IWL_MVM_PS_SNOOZE_HEAVY_RX_THLD_PACKETS;
336 } else {
337 cmd->heavy_tx_thld_packets =
338 IWL_MVM_PS_HEAVY_TX_THLD_PACKETS;
339 cmd->heavy_rx_thld_packets =
340 IWL_MVM_PS_HEAVY_RX_THLD_PACKETS;
341 }
326 cmd->heavy_tx_thld_percentage = 342 cmd->heavy_tx_thld_percentage =
327 IWL_MVM_PS_HEAVY_TX_THLD_PERCENT; 343 IWL_MVM_PS_HEAVY_TX_THLD_PERCENT;
328 cmd->heavy_rx_thld_percentage = 344 cmd->heavy_rx_thld_percentage =
@@ -430,6 +446,32 @@ static int iwl_mvm_power_mac_disable(struct iwl_mvm *mvm,
430 sizeof(cmd), &cmd); 446 sizeof(cmd), &cmd);
431} 447}
432 448
449static int iwl_mvm_power_update_device(struct iwl_mvm *mvm)
450{
451 struct iwl_device_power_cmd cmd = {
452 .flags = cpu_to_le16(DEVICE_POWER_FLAGS_POWER_SAVE_ENA_MSK),
453 };
454
455 if (!(mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_DEVICE_PS_CMD))
456 return 0;
457
458 if (iwlmvm_mod_params.power_scheme == IWL_POWER_SCHEME_CAM)
459 cmd.flags |= cpu_to_le16(DEVICE_POWER_FLAGS_CAM_MSK);
460
461#ifdef CONFIG_IWLWIFI_DEBUGFS
462 if ((mvm->cur_ucode == IWL_UCODE_WOWLAN) ? mvm->disable_power_off_d3 :
463 mvm->disable_power_off)
464 cmd.flags &=
465 cpu_to_le16(~DEVICE_POWER_FLAGS_POWER_SAVE_ENA_MSK);
466#endif
467 IWL_DEBUG_POWER(mvm,
468 "Sending device power command with flags = 0x%X\n",
469 cmd.flags);
470
471 return iwl_mvm_send_cmd_pdu(mvm, POWER_TABLE_CMD, CMD_SYNC, sizeof(cmd),
472 &cmd);
473}
474
433#ifdef CONFIG_IWLWIFI_DEBUGFS 475#ifdef CONFIG_IWLWIFI_DEBUGFS
434static int iwl_mvm_power_mac_dbgfs_read(struct iwl_mvm *mvm, 476static int iwl_mvm_power_mac_dbgfs_read(struct iwl_mvm *mvm,
435 struct ieee80211_vif *vif, char *buf, 477 struct ieee80211_vif *vif, char *buf,
@@ -440,10 +482,11 @@ static int iwl_mvm_power_mac_dbgfs_read(struct iwl_mvm *mvm,
440 482
441 iwl_mvm_power_build_cmd(mvm, vif, &cmd); 483 iwl_mvm_power_build_cmd(mvm, vif, &cmd);
442 484
443 pos += scnprintf(buf+pos, bufsz-pos, "disable_power_off = %d\n", 485 if (!(mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_DEVICE_PS_CMD))
444 (cmd.flags & 486 pos += scnprintf(buf+pos, bufsz-pos, "disable_power_off = %d\n",
445 cpu_to_le16(POWER_FLAGS_POWER_SAVE_ENA_MSK)) ? 487 (cmd.flags &
446 0 : 1); 488 cpu_to_le16(POWER_FLAGS_POWER_SAVE_ENA_MSK)) ?
489 0 : 1);
447 pos += scnprintf(buf+pos, bufsz-pos, "power_scheme = %d\n", 490 pos += scnprintf(buf+pos, bufsz-pos, "power_scheme = %d\n",
448 iwlmvm_mod_params.power_scheme); 491 iwlmvm_mod_params.power_scheme);
449 pos += scnprintf(buf+pos, bufsz-pos, "flags = 0x%x\n", 492 pos += scnprintf(buf+pos, bufsz-pos, "flags = 0x%x\n",
@@ -609,6 +652,7 @@ int iwl_mvm_update_beacon_filter(struct iwl_mvm *mvm,
609 652
610const struct iwl_mvm_power_ops pm_mac_ops = { 653const struct iwl_mvm_power_ops pm_mac_ops = {
611 .power_update_mode = iwl_mvm_power_mac_update_mode, 654 .power_update_mode = iwl_mvm_power_mac_update_mode,
655 .power_update_device_mode = iwl_mvm_power_update_device,
612 .power_disable = iwl_mvm_power_mac_disable, 656 .power_disable = iwl_mvm_power_mac_disable,
613#ifdef CONFIG_IWLWIFI_DEBUGFS 657#ifdef CONFIG_IWLWIFI_DEBUGFS
614 .power_dbgfs_read = iwl_mvm_power_mac_dbgfs_read, 658 .power_dbgfs_read = iwl_mvm_power_mac_dbgfs_read,
diff --git a/drivers/net/wireless/iwlwifi/mvm/quota.c b/drivers/net/wireless/iwlwifi/mvm/quota.c
index 5c6ae16ec52b..17e2bc827f9a 100644
--- a/drivers/net/wireless/iwlwifi/mvm/quota.c
+++ b/drivers/net/wireless/iwlwifi/mvm/quota.c
@@ -110,7 +110,8 @@ static void iwl_mvm_quota_iterator(void *_data, u8 *mac,
110 data->n_interfaces[id]++; 110 data->n_interfaces[id]++;
111 break; 111 break;
112 case NL80211_IFTYPE_AP: 112 case NL80211_IFTYPE_AP:
113 if (mvmvif->ap_active) 113 case NL80211_IFTYPE_ADHOC:
114 if (mvmvif->ap_ibss_active)
114 data->n_interfaces[id]++; 115 data->n_interfaces[id]++;
115 break; 116 break;
116 case NL80211_IFTYPE_MONITOR: 117 case NL80211_IFTYPE_MONITOR:
@@ -119,16 +120,45 @@ static void iwl_mvm_quota_iterator(void *_data, u8 *mac,
119 break; 120 break;
120 case NL80211_IFTYPE_P2P_DEVICE: 121 case NL80211_IFTYPE_P2P_DEVICE:
121 break; 122 break;
122 case NL80211_IFTYPE_ADHOC:
123 if (vif->bss_conf.ibss_joined)
124 data->n_interfaces[id]++;
125 break;
126 default: 123 default:
127 WARN_ON_ONCE(1); 124 WARN_ON_ONCE(1);
128 break; 125 break;
129 } 126 }
130} 127}
131 128
129static void iwl_mvm_adjust_quota_for_noa(struct iwl_mvm *mvm,
130 struct iwl_time_quota_cmd *cmd)
131{
132#ifdef CONFIG_NL80211_TESTMODE
133 struct iwl_mvm_vif *mvmvif;
134 int i, phy_id = -1, beacon_int = 0;
135
136 if (!mvm->noa_duration || !mvm->noa_vif)
137 return;
138
139 mvmvif = iwl_mvm_vif_from_mac80211(mvm->noa_vif);
140 if (!mvmvif->ap_ibss_active)
141 return;
142
143 phy_id = mvmvif->phy_ctxt->id;
144 beacon_int = mvm->noa_vif->bss_conf.beacon_int;
145
146 for (i = 0; i < MAX_BINDINGS; i++) {
147 u32 id_n_c = le32_to_cpu(cmd->quotas[i].id_and_color);
148 u32 id = (id_n_c & FW_CTXT_ID_MSK) >> FW_CTXT_ID_POS;
149 u32 quota = le32_to_cpu(cmd->quotas[i].quota);
150
151 if (id != phy_id)
152 continue;
153
154 quota *= (beacon_int - mvm->noa_duration);
155 quota /= beacon_int;
156
157 cmd->quotas[i].quota = cpu_to_le32(quota);
158 }
159#endif
160}
161
132int iwl_mvm_update_quotas(struct iwl_mvm *mvm, struct ieee80211_vif *newvif) 162int iwl_mvm_update_quotas(struct iwl_mvm *mvm, struct ieee80211_vif *newvif)
133{ 163{
134 struct iwl_time_quota_cmd cmd = {}; 164 struct iwl_time_quota_cmd cmd = {};
@@ -196,6 +226,8 @@ int iwl_mvm_update_quotas(struct iwl_mvm *mvm, struct ieee80211_vif *newvif)
196 /* Give the remainder of the session to the first binding */ 226 /* Give the remainder of the session to the first binding */
197 le32_add_cpu(&cmd.quotas[0].quota, quota_rem); 227 le32_add_cpu(&cmd.quotas[0].quota, quota_rem);
198 228
229 iwl_mvm_adjust_quota_for_noa(mvm, &cmd);
230
199 ret = iwl_mvm_send_cmd_pdu(mvm, TIME_QUOTA_CMD, CMD_SYNC, 231 ret = iwl_mvm_send_cmd_pdu(mvm, TIME_QUOTA_CMD, CMD_SYNC,
200 sizeof(cmd), &cmd); 232 sizeof(cmd), &cmd);
201 if (ret) 233 if (ret)
diff --git a/drivers/net/wireless/iwlwifi/mvm/rs.c b/drivers/net/wireless/iwlwifi/mvm/rs.c
index 4ffaa3fa153f..a0b4cc8d9c3b 100644
--- a/drivers/net/wireless/iwlwifi/mvm/rs.c
+++ b/drivers/net/wireless/iwlwifi/mvm/rs.c
@@ -82,13 +82,24 @@ static const u8 ant_toggle_lookup[] = {
82 [ANT_ABC] = ANT_ABC, 82 [ANT_ABC] = ANT_ABC,
83}; 83};
84 84
85#define IWL_DECLARE_RATE_INFO(r, s, rp, rn) \ 85#define IWL_DECLARE_RATE_INFO(r, s, rp, rn) \
86 [IWL_RATE_##r##M_INDEX] = { IWL_RATE_##r##M_PLCP, \ 86 [IWL_RATE_##r##M_INDEX] = { IWL_RATE_##r##M_PLCP, \
87 IWL_RATE_SISO_##s##M_PLCP, \ 87 IWL_RATE_HT_SISO_MCS_##s##_PLCP, \
88 IWL_RATE_MIMO2_##s##M_PLCP,\ 88 IWL_RATE_HT_MIMO2_MCS_##s##_PLCP, \
89 IWL_RATE_##rp##M_INDEX, \ 89 IWL_RATE_VHT_SISO_MCS_##s##_PLCP, \
90 IWL_RATE_VHT_MIMO2_MCS_##s##_PLCP,\
91 IWL_RATE_##rp##M_INDEX, \
90 IWL_RATE_##rn##M_INDEX } 92 IWL_RATE_##rn##M_INDEX }
91 93
94#define IWL_DECLARE_MCS_RATE(s) \
95 [IWL_RATE_MCS_##s##_INDEX] = { IWL_RATE_INVM_PLCP, \
96 IWL_RATE_HT_SISO_MCS_##s##_PLCP, \
97 IWL_RATE_HT_MIMO2_MCS_##s##_PLCP, \
98 IWL_RATE_VHT_SISO_MCS_##s##_PLCP, \
99 IWL_RATE_VHT_MIMO2_MCS_##s##_PLCP, \
100 IWL_RATE_INVM_INDEX, \
101 IWL_RATE_INVM_INDEX }
102
92/* 103/*
93 * Parameter order: 104 * Parameter order:
94 * rate, ht rate, prev rate, next rate 105 * rate, ht rate, prev rate, next rate
@@ -102,16 +113,17 @@ static const struct iwl_rs_rate_info iwl_rates[IWL_RATE_COUNT] = {
102 IWL_DECLARE_RATE_INFO(2, INV, 1, 5), /* 2mbps */ 113 IWL_DECLARE_RATE_INFO(2, INV, 1, 5), /* 2mbps */
103 IWL_DECLARE_RATE_INFO(5, INV, 2, 11), /*5.5mbps */ 114 IWL_DECLARE_RATE_INFO(5, INV, 2, 11), /*5.5mbps */
104 IWL_DECLARE_RATE_INFO(11, INV, 9, 12), /* 11mbps */ 115 IWL_DECLARE_RATE_INFO(11, INV, 9, 12), /* 11mbps */
105 IWL_DECLARE_RATE_INFO(6, 6, 5, 11), /* 6mbps */ 116 IWL_DECLARE_RATE_INFO(6, 0, 5, 11), /* 6mbps ; MCS 0 */
106 IWL_DECLARE_RATE_INFO(9, 6, 6, 11), /* 9mbps */ 117 IWL_DECLARE_RATE_INFO(9, INV, 6, 11), /* 9mbps */
107 IWL_DECLARE_RATE_INFO(12, 12, 11, 18), /* 12mbps */ 118 IWL_DECLARE_RATE_INFO(12, 1, 11, 18), /* 12mbps ; MCS 1 */
108 IWL_DECLARE_RATE_INFO(18, 18, 12, 24), /* 18mbps */ 119 IWL_DECLARE_RATE_INFO(18, 2, 12, 24), /* 18mbps ; MCS 2 */
109 IWL_DECLARE_RATE_INFO(24, 24, 18, 36), /* 24mbps */ 120 IWL_DECLARE_RATE_INFO(24, 3, 18, 36), /* 24mbps ; MCS 3 */
110 IWL_DECLARE_RATE_INFO(36, 36, 24, 48), /* 36mbps */ 121 IWL_DECLARE_RATE_INFO(36, 4, 24, 48), /* 36mbps ; MCS 4 */
111 IWL_DECLARE_RATE_INFO(48, 48, 36, 54), /* 48mbps */ 122 IWL_DECLARE_RATE_INFO(48, 5, 36, 54), /* 48mbps ; MCS 5 */
112 IWL_DECLARE_RATE_INFO(54, 54, 48, INV), /* 54mbps */ 123 IWL_DECLARE_RATE_INFO(54, 6, 48, INV), /* 54mbps ; MCS 6 */
113 IWL_DECLARE_RATE_INFO(60, 60, 48, INV), /* 60mbps */ 124 IWL_DECLARE_MCS_RATE(7), /* MCS 7 */
114 /* FIXME:RS: ^^ should be INV (legacy) */ 125 IWL_DECLARE_MCS_RATE(8), /* MCS 8 */
126 IWL_DECLARE_MCS_RATE(9), /* MCS 9 */
115}; 127};
116 128
117static inline u8 rs_extract_rate(u32 rate_n_flags) 129static inline u8 rs_extract_rate(u32 rate_n_flags)
@@ -124,26 +136,30 @@ static int iwl_hwrate_to_plcp_idx(u32 rate_n_flags)
124{ 136{
125 int idx = 0; 137 int idx = 0;
126 138
127 /* HT rate format */
128 if (rate_n_flags & RATE_MCS_HT_MSK) { 139 if (rate_n_flags & RATE_MCS_HT_MSK) {
129 idx = rs_extract_rate(rate_n_flags); 140 idx = rate_n_flags & RATE_HT_MCS_RATE_CODE_MSK;
130 141 idx += IWL_RATE_MCS_0_INDEX;
131 WARN_ON_ONCE(idx >= IWL_RATE_MIMO3_6M_PLCP);
132 if (idx >= IWL_RATE_MIMO2_6M_PLCP)
133 idx = idx - IWL_RATE_MIMO2_6M_PLCP;
134 142
135 idx += IWL_FIRST_OFDM_RATE; 143 /* skip 9M not supported in HT*/
136 /* skip 9M not supported in ht*/
137 if (idx >= IWL_RATE_9M_INDEX) 144 if (idx >= IWL_RATE_9M_INDEX)
138 idx += 1; 145 idx += 1;
139 if ((idx >= IWL_FIRST_OFDM_RATE) && (idx <= IWL_LAST_OFDM_RATE)) 146 if ((idx >= IWL_FIRST_HT_RATE) && (idx <= IWL_LAST_HT_RATE))
140 return idx; 147 return idx;
148 } else if (rate_n_flags & RATE_MCS_VHT_MSK) {
149 idx = rate_n_flags & RATE_VHT_MCS_RATE_CODE_MSK;
150 idx += IWL_RATE_MCS_0_INDEX;
141 151
142 /* legacy rate format, search for match in table */ 152 /* skip 9M not supported in VHT*/
153 if (idx >= IWL_RATE_9M_INDEX)
154 idx++;
155 if ((idx >= IWL_FIRST_VHT_RATE) && (idx <= IWL_LAST_VHT_RATE))
156 return idx;
143 } else { 157 } else {
158 /* legacy rate format, search for match in table */
159
160 u8 legacy_rate = rs_extract_rate(rate_n_flags);
144 for (idx = 0; idx < ARRAY_SIZE(iwl_rates); idx++) 161 for (idx = 0; idx < ARRAY_SIZE(iwl_rates); idx++)
145 if (iwl_rates[idx].plcp == 162 if (iwl_rates[idx].plcp == legacy_rate)
146 rs_extract_rate(rate_n_flags))
147 return idx; 163 return idx;
148 } 164 }
149 165
@@ -155,6 +171,7 @@ static void rs_rate_scale_perform(struct iwl_mvm *mvm,
155 struct ieee80211_sta *sta, 171 struct ieee80211_sta *sta,
156 struct iwl_lq_sta *lq_sta); 172 struct iwl_lq_sta *lq_sta);
157static void rs_fill_link_cmd(struct iwl_mvm *mvm, 173static void rs_fill_link_cmd(struct iwl_mvm *mvm,
174 struct ieee80211_sta *sta,
158 struct iwl_lq_sta *lq_sta, u32 rate_n_flags); 175 struct iwl_lq_sta *lq_sta, u32 rate_n_flags);
159static void rs_stay_in_table(struct iwl_lq_sta *lq_sta, bool force_search); 176static void rs_stay_in_table(struct iwl_lq_sta *lq_sta, bool force_search);
160 177
@@ -180,35 +197,52 @@ static void rs_dbgfs_set_mcs(struct iwl_lq_sta *lq_sta,
180 */ 197 */
181 198
182static s32 expected_tpt_legacy[IWL_RATE_COUNT] = { 199static s32 expected_tpt_legacy[IWL_RATE_COUNT] = {
183 7, 13, 35, 58, 40, 57, 72, 98, 121, 154, 177, 186, 0 200 7, 13, 35, 58, 40, 57, 72, 98, 121, 154, 177, 186, 0, 0, 0
184}; 201};
185 202
186static s32 expected_tpt_siso20MHz[4][IWL_RATE_COUNT] = { 203/* Expected TpT tables. 4 indexes:
187 {0, 0, 0, 0, 42, 0, 76, 102, 124, 159, 183, 193, 202}, /* Norm */ 204 * 0 - NGI, 1 - SGI, 2 - AGG+NGI, 3 - AGG+SGI
188 {0, 0, 0, 0, 46, 0, 82, 110, 132, 168, 192, 202, 210}, /* SGI */ 205 */
189 {0, 0, 0, 0, 47, 0, 91, 133, 171, 242, 305, 334, 362}, /* AGG */ 206static s32 expected_tpt_siso_20MHz[4][IWL_RATE_COUNT] = {
190 {0, 0, 0, 0, 52, 0, 101, 145, 187, 264, 330, 361, 390}, /* AGG+SGI */ 207 {0, 0, 0, 0, 42, 0, 76, 102, 124, 159, 183, 193, 202, 216, 0},
208 {0, 0, 0, 0, 46, 0, 82, 110, 132, 168, 192, 202, 210, 225, 0},
209 {0, 0, 0, 0, 49, 0, 97, 145, 192, 285, 375, 420, 464, 551, 0},
210 {0, 0, 0, 0, 54, 0, 108, 160, 213, 315, 415, 465, 513, 608, 0},
191}; 211};
192 212
193static s32 expected_tpt_siso40MHz[4][IWL_RATE_COUNT] = { 213static s32 expected_tpt_siso_40MHz[4][IWL_RATE_COUNT] = {
194 {0, 0, 0, 0, 77, 0, 127, 160, 184, 220, 242, 250, 257}, /* Norm */ 214 {0, 0, 0, 0, 77, 0, 127, 160, 184, 220, 242, 250, 257, 269, 275},
195 {0, 0, 0, 0, 83, 0, 135, 169, 193, 229, 250, 257, 264}, /* SGI */ 215 {0, 0, 0, 0, 83, 0, 135, 169, 193, 229, 250, 257, 264, 275, 280},
196 {0, 0, 0, 0, 94, 0, 177, 249, 313, 423, 512, 550, 586}, /* AGG */ 216 {0, 0, 0, 0, 101, 0, 199, 295, 389, 570, 744, 828, 911, 1070, 1173},
197 {0, 0, 0, 0, 104, 0, 193, 270, 338, 454, 545, 584, 620}, /* AGG+SGI */ 217 {0, 0, 0, 0, 112, 0, 220, 326, 429, 629, 819, 912, 1000, 1173, 1284},
218};
219
220static s32 expected_tpt_siso_80MHz[4][IWL_RATE_COUNT] = {
221 {0, 0, 0, 0, 130, 0, 191, 223, 244, 273, 288, 294, 298, 305, 308},
222 {0, 0, 0, 0, 138, 0, 200, 231, 251, 279, 293, 298, 302, 308, 312},
223 {0, 0, 0, 0, 217, 0, 429, 634, 834, 1220, 1585, 1760, 1931, 2258, 2466},
224 {0, 0, 0, 0, 241, 0, 475, 701, 921, 1343, 1741, 1931, 2117, 2468, 2691},
198}; 225};
199 226
200static s32 expected_tpt_mimo2_20MHz[4][IWL_RATE_COUNT] = { 227static s32 expected_tpt_mimo2_20MHz[4][IWL_RATE_COUNT] = {
201 {0, 0, 0, 0, 74, 0, 123, 155, 179, 214, 236, 244, 251}, /* Norm */ 228 {0, 0, 0, 0, 74, 0, 123, 155, 179, 213, 235, 243, 250, 261, 0},
202 {0, 0, 0, 0, 81, 0, 131, 164, 188, 223, 243, 251, 257}, /* SGI */ 229 {0, 0, 0, 0, 81, 0, 131, 164, 187, 221, 242, 250, 256, 267, 0},
203 {0, 0, 0, 0, 89, 0, 167, 235, 296, 402, 488, 526, 560}, /* AGG */ 230 {0, 0, 0, 0, 98, 0, 193, 286, 375, 550, 718, 799, 878, 1032, 0},
204 {0, 0, 0, 0, 97, 0, 182, 255, 320, 431, 520, 558, 593}, /* AGG+SGI*/ 231 {0, 0, 0, 0, 109, 0, 214, 316, 414, 607, 790, 879, 965, 1132, 0},
205}; 232};
206 233
207static s32 expected_tpt_mimo2_40MHz[4][IWL_RATE_COUNT] = { 234static s32 expected_tpt_mimo2_40MHz[4][IWL_RATE_COUNT] = {
208 {0, 0, 0, 0, 123, 0, 182, 214, 235, 264, 279, 285, 289}, /* Norm */ 235 {0, 0, 0, 0, 123, 0, 182, 214, 235, 264, 279, 285, 289, 296, 300},
209 {0, 0, 0, 0, 131, 0, 191, 222, 242, 270, 284, 289, 293}, /* SGI */ 236 {0, 0, 0, 0, 131, 0, 191, 222, 242, 270, 284, 289, 293, 300, 303},
210 {0, 0, 0, 0, 171, 0, 305, 410, 496, 634, 731, 771, 805}, /* AGG */ 237 {0, 0, 0, 0, 200, 0, 390, 571, 741, 1067, 1365, 1505, 1640, 1894, 2053},
211 {0, 0, 0, 0, 186, 0, 329, 439, 527, 667, 764, 803, 838}, /* AGG+SGI */ 238 {0, 0, 0, 0, 221, 0, 430, 630, 816, 1169, 1490, 1641, 1784, 2053, 2221},
239};
240
241static s32 expected_tpt_mimo2_80MHz[4][IWL_RATE_COUNT] = {
242 {0, 0, 0, 0, 182, 0, 240, 264, 278, 299, 308, 311, 313, 317, 319},
243 {0, 0, 0, 0, 190, 0, 247, 269, 282, 302, 310, 313, 315, 319, 320},
244 {0, 0, 0, 0, 428, 0, 833, 1215, 1577, 2254, 2863, 3147, 3418, 3913, 4219},
245 {0, 0, 0, 0, 474, 0, 920, 1338, 1732, 2464, 3116, 3418, 3705, 4225, 4545},
212}; 246};
213 247
214/* mbps, mcs */ 248/* mbps, mcs */
@@ -263,7 +297,7 @@ static void rs_program_fix_rate(struct iwl_mvm *mvm,
263 lq_sta->lq.sta_id, lq_sta->dbg_fixed_rate); 297 lq_sta->lq.sta_id, lq_sta->dbg_fixed_rate);
264 298
265 if (lq_sta->dbg_fixed_rate) { 299 if (lq_sta->dbg_fixed_rate) {
266 rs_fill_link_cmd(NULL, lq_sta, lq_sta->dbg_fixed_rate); 300 rs_fill_link_cmd(NULL, NULL, lq_sta, lq_sta->dbg_fixed_rate);
267 iwl_mvm_send_lq_cmd(lq_sta->drv, &lq_sta->lq, CMD_ASYNC, false); 301 iwl_mvm_send_lq_cmd(lq_sta->drv, &lq_sta->lq, CMD_ASYNC, false);
268 } 302 }
269} 303}
@@ -275,17 +309,6 @@ static int rs_tl_turn_on_agg_for_tid(struct iwl_mvm *mvm,
275{ 309{
276 int ret = -EAGAIN; 310 int ret = -EAGAIN;
277 311
278 /*
279 * Don't create TX aggregation sessions when in high
280 * BT traffic, as they would just be disrupted by BT.
281 */
282 if (BT_MBOX_MSG(&mvm->last_bt_notif, 3, TRAFFIC_LOAD) >= 2) {
283 IWL_DEBUG_COEX(mvm, "BT traffic (%d), no aggregation allowed\n",
284 BT_MBOX_MSG(&mvm->last_bt_notif,
285 3, TRAFFIC_LOAD));
286 return ret;
287 }
288
289 IWL_DEBUG_HT(mvm, "Starting Tx agg: STA: %pM tid: %d\n", 312 IWL_DEBUG_HT(mvm, "Starting Tx agg: STA: %pM tid: %d\n",
290 sta->addr, tid); 313 sta->addr, tid);
291 ret = ieee80211_start_tx_ba_session(sta, tid, 5000); 314 ret = ieee80211_start_tx_ba_session(sta, tid, 5000);
@@ -416,49 +439,54 @@ static int rs_collect_tx_data(struct iwl_scale_tbl_info *tbl,
416 */ 439 */
417/* FIXME:RS:remove this function and put the flags statically in the table */ 440/* FIXME:RS:remove this function and put the flags statically in the table */
418static u32 rate_n_flags_from_tbl(struct iwl_mvm *mvm, 441static u32 rate_n_flags_from_tbl(struct iwl_mvm *mvm,
419 struct iwl_scale_tbl_info *tbl, 442 struct iwl_scale_tbl_info *tbl, int index)
420 int index, u8 use_green)
421{ 443{
422 u32 rate_n_flags = 0; 444 u32 rate_n_flags = 0;
423 445
446 rate_n_flags |= ((tbl->ant_type << RATE_MCS_ANT_POS) &
447 RATE_MCS_ANT_ABC_MSK);
448
424 if (is_legacy(tbl->lq_type)) { 449 if (is_legacy(tbl->lq_type)) {
425 rate_n_flags = iwl_rates[index].plcp; 450 rate_n_flags |= iwl_rates[index].plcp;
426 if (index >= IWL_FIRST_CCK_RATE && index <= IWL_LAST_CCK_RATE) 451 if (index >= IWL_FIRST_CCK_RATE && index <= IWL_LAST_CCK_RATE)
427 rate_n_flags |= RATE_MCS_CCK_MSK; 452 rate_n_flags |= RATE_MCS_CCK_MSK;
428 } else if (is_Ht(tbl->lq_type)) { 453 return rate_n_flags;
429 if (index > IWL_LAST_OFDM_RATE) { 454 }
455
456 if (is_ht(tbl->lq_type)) {
457 if (index < IWL_FIRST_HT_RATE || index > IWL_LAST_HT_RATE) {
430 IWL_ERR(mvm, "Invalid HT rate index %d\n", index); 458 IWL_ERR(mvm, "Invalid HT rate index %d\n", index);
431 index = IWL_LAST_OFDM_RATE; 459 index = IWL_LAST_HT_RATE;
432 } 460 }
433 rate_n_flags = RATE_MCS_HT_MSK; 461 rate_n_flags |= RATE_MCS_HT_MSK;
434 462
435 if (is_siso(tbl->lq_type)) 463 if (is_ht_siso(tbl->lq_type))
436 rate_n_flags |= iwl_rates[index].plcp_siso; 464 rate_n_flags |= iwl_rates[index].plcp_ht_siso;
437 else if (is_mimo2(tbl->lq_type)) 465 else if (is_ht_mimo2(tbl->lq_type))
438 rate_n_flags |= iwl_rates[index].plcp_mimo2; 466 rate_n_flags |= iwl_rates[index].plcp_ht_mimo2;
439 else 467 else
440 WARN_ON_ONCE(1); 468 WARN_ON_ONCE(1);
469 } else if (is_vht(tbl->lq_type)) {
470 if (index < IWL_FIRST_VHT_RATE || index > IWL_LAST_VHT_RATE) {
471 IWL_ERR(mvm, "Invalid VHT rate index %d\n", index);
472 index = IWL_LAST_VHT_RATE;
473 }
474 rate_n_flags |= RATE_MCS_VHT_MSK;
475 if (is_vht_siso(tbl->lq_type))
476 rate_n_flags |= iwl_rates[index].plcp_vht_siso;
477 else if (is_vht_mimo2(tbl->lq_type))
478 rate_n_flags |= iwl_rates[index].plcp_vht_mimo2;
479 else
480 WARN_ON_ONCE(1);
481
441 } else { 482 } else {
442 IWL_ERR(mvm, "Invalid tbl->lq_type %d\n", tbl->lq_type); 483 IWL_ERR(mvm, "Invalid tbl->lq_type %d\n", tbl->lq_type);
443 } 484 }
444 485
445 rate_n_flags |= ((tbl->ant_type << RATE_MCS_ANT_POS) & 486 rate_n_flags |= tbl->bw;
446 RATE_MCS_ANT_ABC_MSK); 487 if (tbl->is_SGI)
447 488 rate_n_flags |= RATE_MCS_SGI_MSK;
448 if (is_Ht(tbl->lq_type)) { 489
449 if (tbl->is_ht40)
450 rate_n_flags |= RATE_MCS_CHAN_WIDTH_40;
451 if (tbl->is_SGI)
452 rate_n_flags |= RATE_MCS_SGI_MSK;
453
454 if (use_green) {
455 rate_n_flags |= RATE_HT_MCS_GF_MSK;
456 if (is_siso(tbl->lq_type) && tbl->is_SGI) {
457 rate_n_flags &= ~RATE_MCS_SGI_MSK;
458 IWL_ERR(mvm, "GF was set with SGI:SISO\n");
459 }
460 }
461 }
462 return rate_n_flags; 490 return rate_n_flags;
463} 491}
464 492
@@ -473,7 +501,7 @@ static int rs_get_tbl_info_from_mcs(const u32 rate_n_flags,
473{ 501{
474 u32 ant_msk = (rate_n_flags & RATE_MCS_ANT_ABC_MSK); 502 u32 ant_msk = (rate_n_flags & RATE_MCS_ANT_ABC_MSK);
475 u8 num_of_ant = get_num_of_ant_from_rate(rate_n_flags); 503 u8 num_of_ant = get_num_of_ant_from_rate(rate_n_flags);
476 u8 mcs; 504 u8 nss;
477 505
478 memset(tbl, 0, offsetof(struct iwl_scale_tbl_info, win)); 506 memset(tbl, 0, offsetof(struct iwl_scale_tbl_info, win));
479 *rate_idx = iwl_hwrate_to_plcp_idx(rate_n_flags); 507 *rate_idx = iwl_hwrate_to_plcp_idx(rate_n_flags);
@@ -483,41 +511,62 @@ static int rs_get_tbl_info_from_mcs(const u32 rate_n_flags,
483 return -EINVAL; 511 return -EINVAL;
484 } 512 }
485 tbl->is_SGI = 0; /* default legacy setup */ 513 tbl->is_SGI = 0; /* default legacy setup */
486 tbl->is_ht40 = 0; 514 tbl->bw = 0;
487 tbl->ant_type = (ant_msk >> RATE_MCS_ANT_POS); 515 tbl->ant_type = (ant_msk >> RATE_MCS_ANT_POS);
488 tbl->lq_type = LQ_NONE; 516 tbl->lq_type = LQ_NONE;
489 tbl->max_search = IWL_MAX_SEARCH; 517 tbl->max_search = IWL_MAX_SEARCH;
490 518
491 /* legacy rate format */ 519 /* Legacy */
492 if (!(rate_n_flags & RATE_MCS_HT_MSK)) { 520 if (!(rate_n_flags & RATE_MCS_HT_MSK) &&
521 !(rate_n_flags & RATE_MCS_VHT_MSK)) {
493 if (num_of_ant == 1) { 522 if (num_of_ant == 1) {
494 if (band == IEEE80211_BAND_5GHZ) 523 if (band == IEEE80211_BAND_5GHZ)
495 tbl->lq_type = LQ_A; 524 tbl->lq_type = LQ_LEGACY_A;
496 else 525 else
497 tbl->lq_type = LQ_G; 526 tbl->lq_type = LQ_LEGACY_G;
498 } 527 }
499 /* HT rate format */ 528
500 } else { 529 return 0;
501 if (rate_n_flags & RATE_MCS_SGI_MSK) 530 }
502 tbl->is_SGI = 1; 531
503 532 /* HT or VHT */
504 if (rate_n_flags & RATE_MCS_CHAN_WIDTH_40) /* TODO */ 533 if (rate_n_flags & RATE_MCS_SGI_MSK)
505 tbl->is_ht40 = 1; 534 tbl->is_SGI = 1;
506 535
507 mcs = rs_extract_rate(rate_n_flags); 536 tbl->bw = rate_n_flags & RATE_MCS_CHAN_WIDTH_MSK;
508 537
509 /* SISO */ 538 if (rate_n_flags & RATE_MCS_HT_MSK) {
510 if (mcs <= IWL_RATE_SISO_60M_PLCP) { 539 nss = ((rate_n_flags & RATE_HT_MCS_NSS_MSK) >>
511 if (num_of_ant == 1) 540 RATE_HT_MCS_NSS_POS) + 1;
512 tbl->lq_type = LQ_SISO; /*else NONE*/ 541
513 /* MIMO2 */ 542 if (nss == 1) {
514 } else if (mcs <= IWL_RATE_MIMO2_60M_PLCP) { 543 tbl->lq_type = LQ_HT_SISO;
515 if (num_of_ant == 2) 544 WARN_ON_ONCE(num_of_ant != 1);
516 tbl->lq_type = LQ_MIMO2; 545 } else if (nss == 2) {
546 tbl->lq_type = LQ_HT_MIMO2;
547 WARN_ON_ONCE(num_of_ant != 2);
548 } else {
549 WARN_ON_ONCE(1);
550 }
551 } else if (rate_n_flags & RATE_MCS_VHT_MSK) {
552 nss = ((rate_n_flags & RATE_VHT_MCS_NSS_MSK) >>
553 RATE_VHT_MCS_NSS_POS) + 1;
554
555 if (nss == 1) {
556 tbl->lq_type = LQ_VHT_SISO;
557 WARN_ON_ONCE(num_of_ant != 1);
558 } else if (nss == 2) {
559 tbl->lq_type = LQ_VHT_MIMO2;
560 WARN_ON_ONCE(num_of_ant != 2);
517 } else { 561 } else {
518 WARN_ON_ONCE(num_of_ant == 3); 562 WARN_ON_ONCE(1);
519 } 563 }
520 } 564 }
565
566 WARN_ON_ONCE(tbl->bw == RATE_MCS_CHAN_WIDTH_160);
567 WARN_ON_ONCE(tbl->bw == RATE_MCS_CHAN_WIDTH_80 &&
568 !is_vht(tbl->lq_type));
569
521 return 0; 570 return 0;
522} 571}
523 572
@@ -550,22 +599,6 @@ static int rs_toggle_antenna(u32 valid_ant, u32 *rate_n_flags,
550} 599}
551 600
552/** 601/**
553 * Green-field mode is valid if the station supports it and
554 * there are no non-GF stations present in the BSS.
555 */
556static bool rs_use_green(struct ieee80211_sta *sta)
557{
558 /*
559 * There's a bug somewhere in this code that causes the
560 * scaling to get stuck because GF+SGI can't be combined
561 * in SISO rates. Until we find that bug, disable GF, it
562 * has only limited benefit and we still interoperate with
563 * GF APs since we can always receive GF transmissions.
564 */
565 return false;
566}
567
568/**
569 * rs_get_supported_rates - get the available rates 602 * rs_get_supported_rates - get the available rates
570 * 603 *
571 * if management frame or broadcast frame only return 604 * if management frame or broadcast frame only return
@@ -576,16 +609,15 @@ static u16 rs_get_supported_rates(struct iwl_lq_sta *lq_sta,
576 struct ieee80211_hdr *hdr, 609 struct ieee80211_hdr *hdr,
577 enum iwl_table_type rate_type) 610 enum iwl_table_type rate_type)
578{ 611{
579 if (is_legacy(rate_type)) { 612 if (is_legacy(rate_type))
580 return lq_sta->active_legacy_rate; 613 return lq_sta->active_legacy_rate;
581 } else { 614 else if (is_siso(rate_type))
582 if (is_siso(rate_type)) 615 return lq_sta->active_siso_rate;
583 return lq_sta->active_siso_rate; 616 else if (is_mimo2(rate_type))
584 else { 617 return lq_sta->active_mimo2_rate;
585 WARN_ON_ONCE(!is_mimo2(rate_type)); 618
586 return lq_sta->active_mimo2_rate; 619 WARN_ON_ONCE(1);
587 } 620 return 0;
588 }
589} 621}
590 622
591static u16 rs_get_adjacent_rate(struct iwl_mvm *mvm, u8 index, u16 rate_mask, 623static u16 rs_get_adjacent_rate(struct iwl_mvm *mvm, u8 index, u16 rate_mask,
@@ -652,7 +684,6 @@ static u32 rs_get_lower_rate(struct iwl_lq_sta *lq_sta,
652 u16 rate_mask; 684 u16 rate_mask;
653 u16 high_low; 685 u16 high_low;
654 u8 switch_to_legacy = 0; 686 u8 switch_to_legacy = 0;
655 u8 is_green = lq_sta->is_green;
656 struct iwl_mvm *mvm = lq_sta->drv; 687 struct iwl_mvm *mvm = lq_sta->drv;
657 688
658 /* check if we need to switch from HT to legacy rates. 689 /* check if we need to switch from HT to legacy rates.
@@ -662,15 +693,15 @@ static u32 rs_get_lower_rate(struct iwl_lq_sta *lq_sta,
662 switch_to_legacy = 1; 693 switch_to_legacy = 1;
663 scale_index = rs_ht_to_legacy[scale_index]; 694 scale_index = rs_ht_to_legacy[scale_index];
664 if (lq_sta->band == IEEE80211_BAND_5GHZ) 695 if (lq_sta->band == IEEE80211_BAND_5GHZ)
665 tbl->lq_type = LQ_A; 696 tbl->lq_type = LQ_LEGACY_A;
666 else 697 else
667 tbl->lq_type = LQ_G; 698 tbl->lq_type = LQ_LEGACY_G;
668 699
669 if (num_of_ant(tbl->ant_type) > 1) 700 if (num_of_ant(tbl->ant_type) > 1)
670 tbl->ant_type = 701 tbl->ant_type =
671 first_antenna(iwl_fw_valid_tx_ant(mvm->fw)); 702 first_antenna(iwl_fw_valid_tx_ant(mvm->fw));
672 703
673 tbl->is_ht40 = 0; 704 tbl->bw = 0;
674 tbl->is_SGI = 0; 705 tbl->is_SGI = 0;
675 tbl->max_search = IWL_MAX_SEARCH; 706 tbl->max_search = IWL_MAX_SEARCH;
676 } 707 }
@@ -701,7 +732,7 @@ static u32 rs_get_lower_rate(struct iwl_lq_sta *lq_sta,
701 low = scale_index; 732 low = scale_index;
702 733
703out: 734out:
704 return rate_n_flags_from_tbl(lq_sta->drv, tbl, low, is_green); 735 return rate_n_flags_from_tbl(lq_sta->drv, tbl, low);
705} 736}
706 737
707/* 738/*
@@ -714,6 +745,18 @@ static bool table_type_matches(struct iwl_scale_tbl_info *a,
714 (a->is_SGI == b->is_SGI); 745 (a->is_SGI == b->is_SGI);
715} 746}
716 747
748static u32 rs_ch_width_from_mac_flags(enum mac80211_rate_control_flags flags)
749{
750 if (flags & IEEE80211_TX_RC_40_MHZ_WIDTH)
751 return RATE_MCS_CHAN_WIDTH_40;
752 else if (flags & IEEE80211_TX_RC_80_MHZ_WIDTH)
753 return RATE_MCS_CHAN_WIDTH_80;
754 else if (flags & IEEE80211_TX_RC_160_MHZ_WIDTH)
755 return RATE_MCS_CHAN_WIDTH_160;
756
757 return RATE_MCS_CHAN_WIDTH_20;
758}
759
717/* 760/*
718 * mac80211 sends us Tx status 761 * mac80211 sends us Tx status
719 */ 762 */
@@ -783,16 +826,23 @@ static void rs_tx_status(void *mvm_r, struct ieee80211_supported_band *sband,
783 */ 826 */
784 if (info->band == IEEE80211_BAND_2GHZ) 827 if (info->band == IEEE80211_BAND_2GHZ)
785 mac_index += IWL_FIRST_OFDM_RATE; 828 mac_index += IWL_FIRST_OFDM_RATE;
829 } else if (mac_flags & IEEE80211_TX_RC_VHT_MCS) {
830 mac_index &= RATE_VHT_MCS_RATE_CODE_MSK;
831 if (mac_index >= (IWL_RATE_9M_INDEX - IWL_FIRST_OFDM_RATE))
832 mac_index++;
786 } 833 }
834
787 /* Here we actually compare this rate to the latest LQ command */ 835 /* Here we actually compare this rate to the latest LQ command */
788 if ((mac_index < 0) || 836 if ((mac_index < 0) ||
789 (tbl_type.is_SGI != !!(mac_flags & IEEE80211_TX_RC_SHORT_GI)) || 837 (tbl_type.is_SGI != !!(mac_flags & IEEE80211_TX_RC_SHORT_GI)) ||
790 (tbl_type.is_ht40 != !!(mac_flags & IEEE80211_TX_RC_40_MHZ_WIDTH)) || 838 (tbl_type.bw != rs_ch_width_from_mac_flags(mac_flags)) ||
791 (tbl_type.ant_type != info->status.antenna) || 839 (tbl_type.ant_type != info->status.antenna) ||
792 (!!(tx_rate & RATE_MCS_HT_MSK) != 840 (!!(tx_rate & RATE_MCS_HT_MSK) !=
793 !!(mac_flags & IEEE80211_TX_RC_MCS)) || 841 !!(mac_flags & IEEE80211_TX_RC_MCS)) ||
842 (!!(tx_rate & RATE_MCS_VHT_MSK) !=
843 !!(mac_flags & IEEE80211_TX_RC_VHT_MCS)) ||
794 (!!(tx_rate & RATE_HT_MCS_GF_MSK) != 844 (!!(tx_rate & RATE_HT_MCS_GF_MSK) !=
795 !!(mac_flags & IEEE80211_TX_RC_GREEN_FIELD)) || 845 !!(mac_flags & IEEE80211_TX_RC_GREEN_FIELD)) ||
796 (rs_index != mac_index)) { 846 (rs_index != mac_index)) {
797 IWL_DEBUG_RATE(mvm, 847 IWL_DEBUG_RATE(mvm,
798 "initial rate %d does not match %d (0x%x)\n", 848 "initial rate %d does not match %d (0x%x)\n",
@@ -947,7 +997,8 @@ static void rs_set_expected_tpt_table(struct iwl_lq_sta *lq_sta,
947 s32 (*ht_tbl_pointer)[IWL_RATE_COUNT]; 997 s32 (*ht_tbl_pointer)[IWL_RATE_COUNT];
948 998
949 /* Check for invalid LQ type */ 999 /* Check for invalid LQ type */
950 if (WARN_ON_ONCE(!is_legacy(tbl->lq_type) && !is_Ht(tbl->lq_type))) { 1000 if (WARN_ON_ONCE(!is_legacy(tbl->lq_type) && !is_ht(tbl->lq_type) &&
1001 !(is_vht(tbl->lq_type)))) {
951 tbl->expected_tpt = expected_tpt_legacy; 1002 tbl->expected_tpt = expected_tpt_legacy;
952 return; 1003 return;
953 } 1004 }
@@ -958,18 +1009,40 @@ static void rs_set_expected_tpt_table(struct iwl_lq_sta *lq_sta,
958 return; 1009 return;
959 } 1010 }
960 1011
1012 ht_tbl_pointer = expected_tpt_mimo2_20MHz;
961 /* Choose among many HT tables depending on number of streams 1013 /* Choose among many HT tables depending on number of streams
962 * (SISO/MIMO2), channel width (20/40), SGI, and aggregation 1014 * (SISO/MIMO2), channel width (20/40/80), SGI, and aggregation
963 * status */ 1015 * status */
964 if (is_siso(tbl->lq_type) && !tbl->is_ht40) 1016 if (is_siso(tbl->lq_type)) {
965 ht_tbl_pointer = expected_tpt_siso20MHz; 1017 switch (tbl->bw) {
966 else if (is_siso(tbl->lq_type)) 1018 case RATE_MCS_CHAN_WIDTH_20:
967 ht_tbl_pointer = expected_tpt_siso40MHz; 1019 ht_tbl_pointer = expected_tpt_siso_20MHz;
968 else if (is_mimo2(tbl->lq_type) && !tbl->is_ht40) 1020 break;
969 ht_tbl_pointer = expected_tpt_mimo2_20MHz; 1021 case RATE_MCS_CHAN_WIDTH_40:
970 else { 1022 ht_tbl_pointer = expected_tpt_siso_40MHz;
971 WARN_ON_ONCE(!is_mimo2(tbl->lq_type)); 1023 break;
972 ht_tbl_pointer = expected_tpt_mimo2_40MHz; 1024 case RATE_MCS_CHAN_WIDTH_80:
1025 ht_tbl_pointer = expected_tpt_siso_80MHz;
1026 break;
1027 default:
1028 WARN_ON_ONCE(1);
1029 }
1030 } else if (is_mimo2(tbl->lq_type)) {
1031 switch (tbl->bw) {
1032 case RATE_MCS_CHAN_WIDTH_20:
1033 ht_tbl_pointer = expected_tpt_mimo2_20MHz;
1034 break;
1035 case RATE_MCS_CHAN_WIDTH_40:
1036 ht_tbl_pointer = expected_tpt_mimo2_40MHz;
1037 break;
1038 case RATE_MCS_CHAN_WIDTH_80:
1039 ht_tbl_pointer = expected_tpt_mimo2_80MHz;
1040 break;
1041 default:
1042 WARN_ON_ONCE(1);
1043 }
1044 } else {
1045 WARN_ON_ONCE(1);
973 } 1046 }
974 1047
975 if (!tbl->is_SGI && !lq_sta->is_agg) /* Normal */ 1048 if (!tbl->is_SGI && !lq_sta->is_agg) /* Normal */
@@ -1084,9 +1157,47 @@ static s32 rs_get_best_rate(struct iwl_mvm *mvm,
1084 return new_rate; 1157 return new_rate;
1085} 1158}
1086 1159
1087static bool iwl_is_ht40_tx_allowed(struct ieee80211_sta *sta) 1160/* Move to the next action and wrap around to the first action in case
1161 * we're at the last action. Assumes actions start at 0.
1162 */
1163static inline void rs_move_next_action(struct iwl_scale_tbl_info *tbl,
1164 u8 last_action)
1165{
1166 BUILD_BUG_ON(IWL_LEGACY_FIRST_ACTION != 0);
1167 BUILD_BUG_ON(IWL_SISO_FIRST_ACTION != 0);
1168 BUILD_BUG_ON(IWL_MIMO2_FIRST_ACTION != 0);
1169
1170 tbl->action = (tbl->action + 1) % (last_action + 1);
1171}
1172
1173static void rs_set_bw_from_sta(struct iwl_scale_tbl_info *tbl,
1174 struct ieee80211_sta *sta)
1175{
1176 if (sta->bandwidth >= IEEE80211_STA_RX_BW_80)
1177 tbl->bw = RATE_MCS_CHAN_WIDTH_80;
1178 else if (sta->bandwidth >= IEEE80211_STA_RX_BW_40)
1179 tbl->bw = RATE_MCS_CHAN_WIDTH_40;
1180 else
1181 tbl->bw = RATE_MCS_CHAN_WIDTH_20;
1182}
1183
1184static bool rs_sgi_allowed(struct iwl_scale_tbl_info *tbl,
1185 struct ieee80211_sta *sta)
1088{ 1186{
1089 return sta->bandwidth >= IEEE80211_STA_RX_BW_40; 1187 struct ieee80211_sta_ht_cap *ht_cap = &sta->ht_cap;
1188 struct ieee80211_sta_vht_cap *vht_cap = &sta->vht_cap;
1189
1190 if (is_ht20(tbl) && (ht_cap->cap &
1191 IEEE80211_HT_CAP_SGI_20))
1192 return true;
1193 if (is_ht40(tbl) && (ht_cap->cap &
1194 IEEE80211_HT_CAP_SGI_40))
1195 return true;
1196 if (is_ht80(tbl) && (vht_cap->cap &
1197 IEEE80211_VHT_CAP_SHORT_GI_80))
1198 return true;
1199
1200 return false;
1090} 1201}
1091 1202
1092/* 1203/*
@@ -1099,7 +1210,6 @@ static int rs_switch_to_mimo2(struct iwl_mvm *mvm,
1099{ 1210{
1100 u16 rate_mask; 1211 u16 rate_mask;
1101 s32 rate; 1212 s32 rate;
1102 s8 is_green = lq_sta->is_green;
1103 1213
1104 if (!sta->ht_cap.ht_supported) 1214 if (!sta->ht_cap.ht_supported)
1105 return -1; 1215 return -1;
@@ -1113,16 +1223,12 @@ static int rs_switch_to_mimo2(struct iwl_mvm *mvm,
1113 1223
1114 IWL_DEBUG_RATE(mvm, "LQ: try to switch to MIMO2\n"); 1224 IWL_DEBUG_RATE(mvm, "LQ: try to switch to MIMO2\n");
1115 1225
1116 tbl->lq_type = LQ_MIMO2; 1226 tbl->lq_type = lq_sta->is_vht ? LQ_VHT_MIMO2 : LQ_HT_MIMO2;
1117 tbl->action = 0; 1227 tbl->action = 0;
1118 tbl->max_search = IWL_MAX_SEARCH; 1228 tbl->max_search = IWL_MAX_SEARCH;
1119 rate_mask = lq_sta->active_mimo2_rate; 1229 rate_mask = lq_sta->active_mimo2_rate;
1120 1230
1121 if (iwl_is_ht40_tx_allowed(sta)) 1231 rs_set_bw_from_sta(tbl, sta);
1122 tbl->is_ht40 = 1;
1123 else
1124 tbl->is_ht40 = 0;
1125
1126 rs_set_expected_tpt_table(lq_sta, tbl); 1232 rs_set_expected_tpt_table(lq_sta, tbl);
1127 1233
1128 rate = rs_get_best_rate(mvm, lq_sta, tbl, rate_mask, index); 1234 rate = rs_get_best_rate(mvm, lq_sta, tbl, rate_mask, index);
@@ -1134,10 +1240,10 @@ static int rs_switch_to_mimo2(struct iwl_mvm *mvm,
1134 rate, rate_mask); 1240 rate, rate_mask);
1135 return -1; 1241 return -1;
1136 } 1242 }
1137 tbl->current_rate = rate_n_flags_from_tbl(mvm, tbl, rate, is_green); 1243 tbl->current_rate = rate_n_flags_from_tbl(mvm, tbl, rate);
1138 1244
1139 IWL_DEBUG_RATE(mvm, "LQ: Switch to new mcs %X index is green %X\n", 1245 IWL_DEBUG_RATE(mvm, "LQ: Switch to new mcs %X index\n",
1140 tbl->current_rate, is_green); 1246 tbl->current_rate);
1141 return 0; 1247 return 0;
1142} 1248}
1143 1249
@@ -1150,7 +1256,6 @@ static int rs_switch_to_siso(struct iwl_mvm *mvm,
1150 struct iwl_scale_tbl_info *tbl, int index) 1256 struct iwl_scale_tbl_info *tbl, int index)
1151{ 1257{
1152 u16 rate_mask; 1258 u16 rate_mask;
1153 u8 is_green = lq_sta->is_green;
1154 s32 rate; 1259 s32 rate;
1155 1260
1156 if (!sta->ht_cap.ht_supported) 1261 if (!sta->ht_cap.ht_supported)
@@ -1158,19 +1263,12 @@ static int rs_switch_to_siso(struct iwl_mvm *mvm,
1158 1263
1159 IWL_DEBUG_RATE(mvm, "LQ: try to switch to SISO\n"); 1264 IWL_DEBUG_RATE(mvm, "LQ: try to switch to SISO\n");
1160 1265
1161 tbl->lq_type = LQ_SISO; 1266 tbl->lq_type = lq_sta->is_vht ? LQ_VHT_SISO : LQ_HT_SISO;
1162 tbl->action = 0; 1267 tbl->action = 0;
1163 tbl->max_search = IWL_MAX_SEARCH; 1268 tbl->max_search = IWL_MAX_SEARCH;
1164 rate_mask = lq_sta->active_siso_rate; 1269 rate_mask = lq_sta->active_siso_rate;
1165 1270
1166 if (iwl_is_ht40_tx_allowed(sta)) 1271 rs_set_bw_from_sta(tbl, sta);
1167 tbl->is_ht40 = 1;
1168 else
1169 tbl->is_ht40 = 0;
1170
1171 if (is_green)
1172 tbl->is_SGI = 0; /*11n spec: no SGI in SISO+Greenfield*/
1173
1174 rs_set_expected_tpt_table(lq_sta, tbl); 1272 rs_set_expected_tpt_table(lq_sta, tbl);
1175 rate = rs_get_best_rate(mvm, lq_sta, tbl, rate_mask, index); 1273 rate = rs_get_best_rate(mvm, lq_sta, tbl, rate_mask, index);
1176 1274
@@ -1181,9 +1279,9 @@ static int rs_switch_to_siso(struct iwl_mvm *mvm,
1181 rate, rate_mask); 1279 rate, rate_mask);
1182 return -1; 1280 return -1;
1183 } 1281 }
1184 tbl->current_rate = rate_n_flags_from_tbl(mvm, tbl, rate, is_green); 1282 tbl->current_rate = rate_n_flags_from_tbl(mvm, tbl, rate);
1185 IWL_DEBUG_RATE(mvm, "LQ: Switch to new mcs %X index is green %X\n", 1283 IWL_DEBUG_RATE(mvm, "LQ: Switch to new mcs %X index\n",
1186 tbl->current_rate, is_green); 1284 tbl->current_rate);
1187 return 0; 1285 return 0;
1188} 1286}
1189 1287
@@ -1211,14 +1309,10 @@ static int rs_move_legacy_other(struct iwl_mvm *mvm,
1211 while (1) { 1309 while (1) {
1212 lq_sta->action_counter++; 1310 lq_sta->action_counter++;
1213 switch (tbl->action) { 1311 switch (tbl->action) {
1214 case IWL_LEGACY_SWITCH_ANTENNA1: 1312 case IWL_LEGACY_SWITCH_ANTENNA:
1215 case IWL_LEGACY_SWITCH_ANTENNA2:
1216 IWL_DEBUG_RATE(mvm, "LQ: Legacy toggle Antenna\n"); 1313 IWL_DEBUG_RATE(mvm, "LQ: Legacy toggle Antenna\n");
1217 1314
1218 if ((tbl->action == IWL_LEGACY_SWITCH_ANTENNA1 && 1315 if (tx_chains_num <= 1)
1219 tx_chains_num <= 1) ||
1220 (tbl->action == IWL_LEGACY_SWITCH_ANTENNA2 &&
1221 tx_chains_num <= 2))
1222 break; 1316 break;
1223 1317
1224 /* Don't change antenna if success has been great */ 1318 /* Don't change antenna if success has been great */
@@ -1273,9 +1367,7 @@ static int rs_move_legacy_other(struct iwl_mvm *mvm,
1273 default: 1367 default:
1274 WARN_ON_ONCE(1); 1368 WARN_ON_ONCE(1);
1275 } 1369 }
1276 tbl->action++; 1370 rs_move_next_action(tbl, IWL_LEGACY_LAST_ACTION);
1277 if (tbl->action > IWL_LEGACY_SWITCH_MIMO2)
1278 tbl->action = IWL_LEGACY_SWITCH_ANTENNA1;
1279 1371
1280 if (tbl->action == start_action) 1372 if (tbl->action == start_action)
1281 break; 1373 break;
@@ -1285,9 +1377,7 @@ static int rs_move_legacy_other(struct iwl_mvm *mvm,
1285 1377
1286out: 1378out:
1287 lq_sta->search_better_tbl = 1; 1379 lq_sta->search_better_tbl = 1;
1288 tbl->action++; 1380 rs_move_next_action(tbl, IWL_LEGACY_LAST_ACTION);
1289 if (tbl->action > IWL_LEGACY_SWITCH_MIMO2)
1290 tbl->action = IWL_LEGACY_SWITCH_ANTENNA1;
1291 if (update_search_tbl_counter) 1381 if (update_search_tbl_counter)
1292 search_tbl->action = tbl->action; 1382 search_tbl->action = tbl->action;
1293 return 0; 1383 return 0;
@@ -1300,12 +1390,10 @@ static int rs_move_siso_to_other(struct iwl_mvm *mvm,
1300 struct iwl_lq_sta *lq_sta, 1390 struct iwl_lq_sta *lq_sta,
1301 struct ieee80211_sta *sta, int index) 1391 struct ieee80211_sta *sta, int index)
1302{ 1392{
1303 u8 is_green = lq_sta->is_green;
1304 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); 1393 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]);
1305 struct iwl_scale_tbl_info *search_tbl = 1394 struct iwl_scale_tbl_info *search_tbl =
1306 &(lq_sta->lq_info[(1 - lq_sta->active_tbl)]); 1395 &(lq_sta->lq_info[(1 - lq_sta->active_tbl)]);
1307 struct iwl_rate_scale_data *window = &(tbl->win[index]); 1396 struct iwl_rate_scale_data *window = &(tbl->win[index]);
1308 struct ieee80211_sta_ht_cap *ht_cap = &sta->ht_cap;
1309 u32 sz = (sizeof(struct iwl_scale_tbl_info) - 1397 u32 sz = (sizeof(struct iwl_scale_tbl_info) -
1310 (sizeof(struct iwl_rate_scale_data) * IWL_RATE_COUNT)); 1398 (sizeof(struct iwl_rate_scale_data) * IWL_RATE_COUNT));
1311 u8 start_action; 1399 u8 start_action;
@@ -1314,40 +1402,17 @@ static int rs_move_siso_to_other(struct iwl_mvm *mvm,
1314 u8 update_search_tbl_counter = 0; 1402 u8 update_search_tbl_counter = 0;
1315 int ret; 1403 int ret;
1316 1404
1317 switch (BT_MBOX_MSG(&mvm->last_bt_notif, 3, TRAFFIC_LOAD)) { 1405 if (tbl->action == IWL_SISO_SWITCH_MIMO2 &&
1318 case IWL_BT_COEX_TRAFFIC_LOAD_NONE: 1406 !iwl_mvm_bt_coex_is_mimo_allowed(mvm, sta))
1319 /* nothing */ 1407 tbl->action = IWL_SISO_SWITCH_ANTENNA;
1320 break;
1321 case IWL_BT_COEX_TRAFFIC_LOAD_LOW:
1322 /* avoid antenna B unless MIMO */
1323 if (tbl->action == IWL_SISO_SWITCH_ANTENNA2)
1324 tbl->action = IWL_SISO_SWITCH_MIMO2;
1325 break;
1326 case IWL_BT_COEX_TRAFFIC_LOAD_HIGH:
1327 case IWL_BT_COEX_TRAFFIC_LOAD_CONTINUOUS:
1328 /* avoid antenna B and MIMO */
1329 valid_tx_ant =
1330 first_antenna(iwl_fw_valid_tx_ant(mvm->fw));
1331 if (tbl->action != IWL_SISO_SWITCH_ANTENNA1)
1332 tbl->action = IWL_SISO_SWITCH_ANTENNA1;
1333 break;
1334 default:
1335 IWL_ERR(mvm, "Invalid BT load %d",
1336 BT_MBOX_MSG(&mvm->last_bt_notif, 3, TRAFFIC_LOAD));
1337 break;
1338 }
1339 1408
1340 start_action = tbl->action; 1409 start_action = tbl->action;
1341 while (1) { 1410 while (1) {
1342 lq_sta->action_counter++; 1411 lq_sta->action_counter++;
1343 switch (tbl->action) { 1412 switch (tbl->action) {
1344 case IWL_SISO_SWITCH_ANTENNA1: 1413 case IWL_SISO_SWITCH_ANTENNA:
1345 case IWL_SISO_SWITCH_ANTENNA2:
1346 IWL_DEBUG_RATE(mvm, "LQ: SISO toggle Antenna\n"); 1414 IWL_DEBUG_RATE(mvm, "LQ: SISO toggle Antenna\n");
1347 if ((tbl->action == IWL_SISO_SWITCH_ANTENNA1 && 1415 if (tx_chains_num <= 1)
1348 tx_chains_num <= 1) ||
1349 (tbl->action == IWL_SISO_SWITCH_ANTENNA2 &&
1350 tx_chains_num <= 2))
1351 break; 1416 break;
1352 1417
1353 if (window->success_ratio >= IWL_RS_GOOD_RATIO && 1418 if (window->success_ratio >= IWL_RS_GOOD_RATIO &&
@@ -1380,23 +1445,12 @@ static int rs_move_siso_to_other(struct iwl_mvm *mvm,
1380 goto out; 1445 goto out;
1381 break; 1446 break;
1382 case IWL_SISO_SWITCH_GI: 1447 case IWL_SISO_SWITCH_GI:
1383 if (!tbl->is_ht40 && !(ht_cap->cap & 1448 if (!rs_sgi_allowed(tbl, sta))
1384 IEEE80211_HT_CAP_SGI_20))
1385 break;
1386 if (tbl->is_ht40 && !(ht_cap->cap &
1387 IEEE80211_HT_CAP_SGI_40))
1388 break; 1449 break;
1389 1450
1390 IWL_DEBUG_RATE(mvm, "LQ: SISO toggle SGI/NGI\n"); 1451 IWL_DEBUG_RATE(mvm, "LQ: SISO toggle SGI/NGI\n");
1391 1452
1392 memcpy(search_tbl, tbl, sz); 1453 memcpy(search_tbl, tbl, sz);
1393 if (is_green) {
1394 if (!tbl->is_SGI)
1395 break;
1396 else
1397 IWL_ERR(mvm,
1398 "SGI was set in GF+SISO\n");
1399 }
1400 search_tbl->is_SGI = !tbl->is_SGI; 1454 search_tbl->is_SGI = !tbl->is_SGI;
1401 rs_set_expected_tpt_table(lq_sta, search_tbl); 1455 rs_set_expected_tpt_table(lq_sta, search_tbl);
1402 if (tbl->is_SGI) { 1456 if (tbl->is_SGI) {
@@ -1405,16 +1459,13 @@ static int rs_move_siso_to_other(struct iwl_mvm *mvm,
1405 break; 1459 break;
1406 } 1460 }
1407 search_tbl->current_rate = 1461 search_tbl->current_rate =
1408 rate_n_flags_from_tbl(mvm, search_tbl, 1462 rate_n_flags_from_tbl(mvm, search_tbl, index);
1409 index, is_green);
1410 update_search_tbl_counter = 1; 1463 update_search_tbl_counter = 1;
1411 goto out; 1464 goto out;
1412 default: 1465 default:
1413 WARN_ON_ONCE(1); 1466 WARN_ON_ONCE(1);
1414 } 1467 }
1415 tbl->action++; 1468 rs_move_next_action(tbl, IWL_SISO_LAST_ACTION);
1416 if (tbl->action > IWL_SISO_SWITCH_GI)
1417 tbl->action = IWL_SISO_SWITCH_ANTENNA1;
1418 1469
1419 if (tbl->action == start_action) 1470 if (tbl->action == start_action)
1420 break; 1471 break;
@@ -1424,9 +1475,7 @@ static int rs_move_siso_to_other(struct iwl_mvm *mvm,
1424 1475
1425 out: 1476 out:
1426 lq_sta->search_better_tbl = 1; 1477 lq_sta->search_better_tbl = 1;
1427 tbl->action++; 1478 rs_move_next_action(tbl, IWL_SISO_LAST_ACTION);
1428 if (tbl->action > IWL_SISO_SWITCH_GI)
1429 tbl->action = IWL_SISO_SWITCH_ANTENNA1;
1430 if (update_search_tbl_counter) 1479 if (update_search_tbl_counter)
1431 search_tbl->action = tbl->action; 1480 search_tbl->action = tbl->action;
1432 1481
@@ -1440,63 +1489,20 @@ static int rs_move_mimo2_to_other(struct iwl_mvm *mvm,
1440 struct iwl_lq_sta *lq_sta, 1489 struct iwl_lq_sta *lq_sta,
1441 struct ieee80211_sta *sta, int index) 1490 struct ieee80211_sta *sta, int index)
1442{ 1491{
1443 s8 is_green = lq_sta->is_green;
1444 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); 1492 struct iwl_scale_tbl_info *tbl = &(lq_sta->lq_info[lq_sta->active_tbl]);
1445 struct iwl_scale_tbl_info *search_tbl = 1493 struct iwl_scale_tbl_info *search_tbl =
1446 &(lq_sta->lq_info[(1 - lq_sta->active_tbl)]); 1494 &(lq_sta->lq_info[(1 - lq_sta->active_tbl)]);
1447 struct iwl_rate_scale_data *window = &(tbl->win[index]);
1448 struct ieee80211_sta_ht_cap *ht_cap = &sta->ht_cap;
1449 u32 sz = (sizeof(struct iwl_scale_tbl_info) - 1495 u32 sz = (sizeof(struct iwl_scale_tbl_info) -
1450 (sizeof(struct iwl_rate_scale_data) * IWL_RATE_COUNT)); 1496 (sizeof(struct iwl_rate_scale_data) * IWL_RATE_COUNT));
1451 u8 start_action; 1497 u8 start_action;
1452 u8 valid_tx_ant = iwl_fw_valid_tx_ant(mvm->fw); 1498 u8 valid_tx_ant = iwl_fw_valid_tx_ant(mvm->fw);
1453 u8 tx_chains_num = num_of_ant(valid_tx_ant);
1454 u8 update_search_tbl_counter = 0; 1499 u8 update_search_tbl_counter = 0;
1455 int ret; 1500 int ret;
1456 1501
1457 switch (BT_MBOX_MSG(&mvm->last_bt_notif, 3, TRAFFIC_LOAD)) {
1458 case IWL_BT_COEX_TRAFFIC_LOAD_NONE:
1459 /* nothing */
1460 break;
1461 case IWL_BT_COEX_TRAFFIC_LOAD_HIGH:
1462 case IWL_BT_COEX_TRAFFIC_LOAD_CONTINUOUS:
1463 /* avoid antenna B and MIMO */
1464 if (tbl->action != IWL_MIMO2_SWITCH_SISO_A)
1465 tbl->action = IWL_MIMO2_SWITCH_SISO_A;
1466 break;
1467 case IWL_BT_COEX_TRAFFIC_LOAD_LOW:
1468 /* avoid antenna B unless MIMO */
1469 if (tbl->action == IWL_MIMO2_SWITCH_SISO_B)
1470 tbl->action = IWL_MIMO2_SWITCH_SISO_A;
1471 break;
1472 default:
1473 IWL_ERR(mvm, "Invalid BT load %d",
1474 BT_MBOX_MSG(&mvm->last_bt_notif, 3, TRAFFIC_LOAD));
1475 break;
1476 }
1477
1478 start_action = tbl->action; 1502 start_action = tbl->action;
1479 while (1) { 1503 while (1) {
1480 lq_sta->action_counter++; 1504 lq_sta->action_counter++;
1481 switch (tbl->action) { 1505 switch (tbl->action) {
1482 case IWL_MIMO2_SWITCH_ANTENNA1:
1483 case IWL_MIMO2_SWITCH_ANTENNA2:
1484 IWL_DEBUG_RATE(mvm, "LQ: MIMO2 toggle Antennas\n");
1485
1486 if (tx_chains_num <= 2)
1487 break;
1488
1489 if (window->success_ratio >= IWL_RS_GOOD_RATIO)
1490 break;
1491
1492 memcpy(search_tbl, tbl, sz);
1493 if (rs_toggle_antenna(valid_tx_ant,
1494 &search_tbl->current_rate,
1495 search_tbl)) {
1496 update_search_tbl_counter = 1;
1497 goto out;
1498 }
1499 break;
1500 case IWL_MIMO2_SWITCH_SISO_A: 1506 case IWL_MIMO2_SWITCH_SISO_A:
1501 case IWL_MIMO2_SWITCH_SISO_B: 1507 case IWL_MIMO2_SWITCH_SISO_B:
1502 IWL_DEBUG_RATE(mvm, "LQ: MIMO2 switch to SISO\n"); 1508 IWL_DEBUG_RATE(mvm, "LQ: MIMO2 switch to SISO\n");
@@ -1521,11 +1527,7 @@ static int rs_move_mimo2_to_other(struct iwl_mvm *mvm,
1521 break; 1527 break;
1522 1528
1523 case IWL_MIMO2_SWITCH_GI: 1529 case IWL_MIMO2_SWITCH_GI:
1524 if (!tbl->is_ht40 && !(ht_cap->cap & 1530 if (!rs_sgi_allowed(tbl, sta))
1525 IEEE80211_HT_CAP_SGI_20))
1526 break;
1527 if (tbl->is_ht40 && !(ht_cap->cap &
1528 IEEE80211_HT_CAP_SGI_40))
1529 break; 1531 break;
1530 1532
1531 IWL_DEBUG_RATE(mvm, "LQ: MIMO2 toggle SGI/NGI\n"); 1533 IWL_DEBUG_RATE(mvm, "LQ: MIMO2 toggle SGI/NGI\n");
@@ -1546,16 +1548,13 @@ static int rs_move_mimo2_to_other(struct iwl_mvm *mvm,
1546 break; 1548 break;
1547 } 1549 }
1548 search_tbl->current_rate = 1550 search_tbl->current_rate =
1549 rate_n_flags_from_tbl(mvm, search_tbl, 1551 rate_n_flags_from_tbl(mvm, search_tbl, index);
1550 index, is_green);
1551 update_search_tbl_counter = 1; 1552 update_search_tbl_counter = 1;
1552 goto out; 1553 goto out;
1553 default: 1554 default:
1554 WARN_ON_ONCE(1); 1555 WARN_ON_ONCE(1);
1555 } 1556 }
1556 tbl->action++; 1557 rs_move_next_action(tbl, IWL_MIMO2_LAST_ACTION);
1557 if (tbl->action > IWL_MIMO2_SWITCH_GI)
1558 tbl->action = IWL_MIMO2_SWITCH_ANTENNA1;
1559 1558
1560 if (tbl->action == start_action) 1559 if (tbl->action == start_action)
1561 break; 1560 break;
@@ -1564,9 +1563,7 @@ static int rs_move_mimo2_to_other(struct iwl_mvm *mvm,
1564 return 0; 1563 return 0;
1565 out: 1564 out:
1566 lq_sta->search_better_tbl = 1; 1565 lq_sta->search_better_tbl = 1;
1567 tbl->action++; 1566 rs_move_next_action(tbl, IWL_MIMO2_LAST_ACTION);
1568 if (tbl->action > IWL_MIMO2_SWITCH_GI)
1569 tbl->action = IWL_MIMO2_SWITCH_ANTENNA1;
1570 if (update_search_tbl_counter) 1567 if (update_search_tbl_counter)
1571 search_tbl->action = tbl->action; 1568 search_tbl->action = tbl->action;
1572 1569
@@ -1660,15 +1657,16 @@ static void rs_stay_in_table(struct iwl_lq_sta *lq_sta, bool force_search)
1660 * setup rate table in uCode 1657 * setup rate table in uCode
1661 */ 1658 */
1662static void rs_update_rate_tbl(struct iwl_mvm *mvm, 1659static void rs_update_rate_tbl(struct iwl_mvm *mvm,
1660 struct ieee80211_sta *sta,
1663 struct iwl_lq_sta *lq_sta, 1661 struct iwl_lq_sta *lq_sta,
1664 struct iwl_scale_tbl_info *tbl, 1662 struct iwl_scale_tbl_info *tbl,
1665 int index, u8 is_green) 1663 int index)
1666{ 1664{
1667 u32 rate; 1665 u32 rate;
1668 1666
1669 /* Update uCode's rate table. */ 1667 /* Update uCode's rate table. */
1670 rate = rate_n_flags_from_tbl(mvm, tbl, index, is_green); 1668 rate = rate_n_flags_from_tbl(mvm, tbl, index);
1671 rs_fill_link_cmd(mvm, lq_sta, rate); 1669 rs_fill_link_cmd(mvm, sta, lq_sta, rate);
1672 iwl_mvm_send_lq_cmd(mvm, &lq_sta->lq, CMD_ASYNC, false); 1670 iwl_mvm_send_lq_cmd(mvm, &lq_sta->lq, CMD_ASYNC, false);
1673} 1671}
1674 1672
@@ -1712,7 +1710,6 @@ static void rs_rate_scale_perform(struct iwl_mvm *mvm,
1712 u8 update_lq = 0; 1710 u8 update_lq = 0;
1713 struct iwl_scale_tbl_info *tbl, *tbl1; 1711 struct iwl_scale_tbl_info *tbl, *tbl1;
1714 u16 rate_scale_index_msk = 0; 1712 u16 rate_scale_index_msk = 0;
1715 u8 is_green = 0;
1716 u8 active_tbl = 0; 1713 u8 active_tbl = 0;
1717 u8 done_search = 0; 1714 u8 done_search = 0;
1718 u16 high_low; 1715 u16 high_low;
@@ -1754,11 +1751,6 @@ static void rs_rate_scale_perform(struct iwl_mvm *mvm,
1754 active_tbl = 1 - lq_sta->active_tbl; 1751 active_tbl = 1 - lq_sta->active_tbl;
1755 1752
1756 tbl = &(lq_sta->lq_info[active_tbl]); 1753 tbl = &(lq_sta->lq_info[active_tbl]);
1757 if (is_legacy(tbl->lq_type))
1758 lq_sta->is_green = 0;
1759 else
1760 lq_sta->is_green = rs_use_green(sta);
1761 is_green = lq_sta->is_green;
1762 1754
1763 /* current tx rate */ 1755 /* current tx rate */
1764 index = lq_sta->last_txrate_idx; 1756 index = lq_sta->last_txrate_idx;
@@ -1797,7 +1789,7 @@ static void rs_rate_scale_perform(struct iwl_mvm *mvm,
1797 tbl = &(lq_sta->lq_info[lq_sta->active_tbl]); 1789 tbl = &(lq_sta->lq_info[lq_sta->active_tbl]);
1798 /* get "active" rate info */ 1790 /* get "active" rate info */
1799 index = iwl_hwrate_to_plcp_idx(tbl->current_rate); 1791 index = iwl_hwrate_to_plcp_idx(tbl->current_rate);
1800 rs_update_rate_tbl(mvm, lq_sta, tbl, index, is_green); 1792 rs_update_rate_tbl(mvm, sta, lq_sta, tbl, index);
1801 } 1793 }
1802 return; 1794 return;
1803 } 1795 }
@@ -1978,24 +1970,24 @@ static void rs_rate_scale_perform(struct iwl_mvm *mvm,
1978 (current_tpt > (100 * tbl->expected_tpt[low])))) 1970 (current_tpt > (100 * tbl->expected_tpt[low]))))
1979 scale_action = 0; 1971 scale_action = 0;
1980 1972
1981 if ((BT_MBOX_MSG(&mvm->last_bt_notif, 3, TRAFFIC_LOAD) >= 1973 if ((le32_to_cpu(mvm->last_bt_notif.bt_activity_grading) >=
1982 IWL_BT_COEX_TRAFFIC_LOAD_HIGH) && (is_mimo(tbl->lq_type))) { 1974 IWL_BT_COEX_TRAFFIC_LOAD_HIGH) && (is_mimo(tbl->lq_type))) {
1983 if (lq_sta->last_bt_traffic > 1975 if (lq_sta->last_bt_traffic >
1984 BT_MBOX_MSG(&mvm->last_bt_notif, 3, TRAFFIC_LOAD)) { 1976 le32_to_cpu(mvm->last_bt_notif.bt_activity_grading)) {
1985 /* 1977 /*
1986 * don't set scale_action, don't want to scale up if 1978 * don't set scale_action, don't want to scale up if
1987 * the rate scale doesn't otherwise think that is a 1979 * the rate scale doesn't otherwise think that is a
1988 * good idea. 1980 * good idea.
1989 */ 1981 */
1990 } else if (lq_sta->last_bt_traffic <= 1982 } else if (lq_sta->last_bt_traffic <=
1991 BT_MBOX_MSG(&mvm->last_bt_notif, 3, TRAFFIC_LOAD)) { 1983 le32_to_cpu(mvm->last_bt_notif.bt_activity_grading)) {
1992 scale_action = -1; 1984 scale_action = -1;
1993 } 1985 }
1994 } 1986 }
1995 lq_sta->last_bt_traffic = 1987 lq_sta->last_bt_traffic =
1996 BT_MBOX_MSG(&mvm->last_bt_notif, 3, TRAFFIC_LOAD); 1988 le32_to_cpu(mvm->last_bt_notif.bt_activity_grading);
1997 1989
1998 if ((BT_MBOX_MSG(&mvm->last_bt_notif, 3, TRAFFIC_LOAD) >= 1990 if ((le32_to_cpu(mvm->last_bt_notif.bt_activity_grading) >=
1999 IWL_BT_COEX_TRAFFIC_LOAD_HIGH) && is_mimo(tbl->lq_type)) { 1991 IWL_BT_COEX_TRAFFIC_LOAD_HIGH) && is_mimo(tbl->lq_type)) {
2000 /* search for a new modulation */ 1992 /* search for a new modulation */
2001 rs_stay_in_table(lq_sta, true); 1993 rs_stay_in_table(lq_sta, true);
@@ -2032,7 +2024,7 @@ static void rs_rate_scale_perform(struct iwl_mvm *mvm,
2032lq_update: 2024lq_update:
2033 /* Replace uCode's rate table for the destination station. */ 2025 /* Replace uCode's rate table for the destination station. */
2034 if (update_lq) 2026 if (update_lq)
2035 rs_update_rate_tbl(mvm, lq_sta, tbl, index, is_green); 2027 rs_update_rate_tbl(mvm, sta, lq_sta, tbl, index);
2036 2028
2037 rs_stay_in_table(lq_sta, false); 2029 rs_stay_in_table(lq_sta, false);
2038 2030
@@ -2071,7 +2063,7 @@ lq_update:
2071 IWL_DEBUG_RATE(mvm, 2063 IWL_DEBUG_RATE(mvm,
2072 "Switch current mcs: %X index: %d\n", 2064 "Switch current mcs: %X index: %d\n",
2073 tbl->current_rate, index); 2065 tbl->current_rate, index);
2074 rs_fill_link_cmd(mvm, lq_sta, tbl->current_rate); 2066 rs_fill_link_cmd(mvm, sta, lq_sta, tbl->current_rate);
2075 iwl_mvm_send_lq_cmd(mvm, &lq_sta->lq, CMD_ASYNC, false); 2067 iwl_mvm_send_lq_cmd(mvm, &lq_sta->lq, CMD_ASYNC, false);
2076 } else { 2068 } else {
2077 done_search = 1; 2069 done_search = 1;
@@ -2113,7 +2105,7 @@ lq_update:
2113 } 2105 }
2114 2106
2115out: 2107out:
2116 tbl->current_rate = rate_n_flags_from_tbl(mvm, tbl, index, is_green); 2108 tbl->current_rate = rate_n_flags_from_tbl(mvm, tbl, index);
2117 lq_sta->last_txrate_idx = index; 2109 lq_sta->last_txrate_idx = index;
2118} 2110}
2119 2111
@@ -2140,7 +2132,6 @@ static void rs_initialize_lq(struct iwl_mvm *mvm,
2140 int rate_idx; 2132 int rate_idx;
2141 int i; 2133 int i;
2142 u32 rate; 2134 u32 rate;
2143 u8 use_green = rs_use_green(sta);
2144 u8 active_tbl = 0; 2135 u8 active_tbl = 0;
2145 u8 valid_tx_ant; 2136 u8 valid_tx_ant;
2146 2137
@@ -2172,10 +2163,10 @@ static void rs_initialize_lq(struct iwl_mvm *mvm,
2172 if (!rs_is_valid_ant(valid_tx_ant, tbl->ant_type)) 2163 if (!rs_is_valid_ant(valid_tx_ant, tbl->ant_type))
2173 rs_toggle_antenna(valid_tx_ant, &rate, tbl); 2164 rs_toggle_antenna(valid_tx_ant, &rate, tbl);
2174 2165
2175 rate = rate_n_flags_from_tbl(mvm, tbl, rate_idx, use_green); 2166 rate = rate_n_flags_from_tbl(mvm, tbl, rate_idx);
2176 tbl->current_rate = rate; 2167 tbl->current_rate = rate;
2177 rs_set_expected_tpt_table(lq_sta, tbl); 2168 rs_set_expected_tpt_table(lq_sta, tbl);
2178 rs_fill_link_cmd(NULL, lq_sta, rate); 2169 rs_fill_link_cmd(NULL, NULL, lq_sta, rate);
2179 /* TODO restore station should remember the lq cmd */ 2170 /* TODO restore station should remember the lq cmd */
2180 iwl_mvm_send_lq_cmd(mvm, &lq_sta->lq, CMD_SYNC, true); 2171 iwl_mvm_send_lq_cmd(mvm, &lq_sta->lq, CMD_SYNC, true);
2181} 2172}
@@ -2190,7 +2181,6 @@ static void rs_get_rate(void *mvm_r, struct ieee80211_sta *sta, void *mvm_sta,
2190 struct iwl_mvm *mvm __maybe_unused = IWL_OP_MODE_GET_MVM(op_mode); 2181 struct iwl_mvm *mvm __maybe_unused = IWL_OP_MODE_GET_MVM(op_mode);
2191 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); 2182 struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb);
2192 struct iwl_lq_sta *lq_sta = mvm_sta; 2183 struct iwl_lq_sta *lq_sta = mvm_sta;
2193 int rate_idx;
2194 2184
2195 IWL_DEBUG_RATE_LIMIT(mvm, "rate scale calculate new rate for skb\n"); 2185 IWL_DEBUG_RATE_LIMIT(mvm, "rate scale calculate new rate for skb\n");
2196 2186
@@ -2215,36 +2205,9 @@ static void rs_get_rate(void *mvm_r, struct ieee80211_sta *sta, void *mvm_sta,
2215 if (rate_control_send_low(sta, mvm_sta, txrc)) 2205 if (rate_control_send_low(sta, mvm_sta, txrc))
2216 return; 2206 return;
2217 2207
2218 rate_idx = lq_sta->last_txrate_idx; 2208 iwl_mvm_hwrate_to_tx_rate(lq_sta->last_rate_n_flags,
2219 2209 info->band, &info->control.rates[0]);
2220 if (lq_sta->last_rate_n_flags & RATE_MCS_HT_MSK) { 2210
2221 rate_idx -= IWL_FIRST_OFDM_RATE;
2222 /* 6M and 9M shared same MCS index */
2223 rate_idx = (rate_idx > 0) ? (rate_idx - 1) : 0;
2224 WARN_ON_ONCE(rs_extract_rate(lq_sta->last_rate_n_flags) >=
2225 IWL_RATE_MIMO3_6M_PLCP);
2226 if (rs_extract_rate(lq_sta->last_rate_n_flags) >=
2227 IWL_RATE_MIMO2_6M_PLCP)
2228 rate_idx = rate_idx + MCS_INDEX_PER_STREAM;
2229 info->control.rates[0].flags = IEEE80211_TX_RC_MCS;
2230 if (lq_sta->last_rate_n_flags & RATE_MCS_SGI_MSK)
2231 info->control.rates[0].flags |= IEEE80211_TX_RC_SHORT_GI;
2232 if (lq_sta->last_rate_n_flags & RATE_MCS_CHAN_WIDTH_40) /* TODO */
2233 info->control.rates[0].flags |= IEEE80211_TX_RC_40_MHZ_WIDTH;
2234 if (lq_sta->last_rate_n_flags & RATE_HT_MCS_GF_MSK)
2235 info->control.rates[0].flags |= IEEE80211_TX_RC_GREEN_FIELD;
2236 } else {
2237 /* Check for invalid rates */
2238 if ((rate_idx < 0) || (rate_idx >= IWL_RATE_COUNT_LEGACY) ||
2239 ((sband->band == IEEE80211_BAND_5GHZ) &&
2240 (rate_idx < IWL_FIRST_OFDM_RATE)))
2241 rate_idx = rate_lowest_index(sband, sta);
2242 /* On valid 5 GHz rate, adjust index */
2243 else if (sband->band == IEEE80211_BAND_5GHZ)
2244 rate_idx -= IWL_FIRST_OFDM_RATE;
2245 info->control.rates[0].flags = 0;
2246 }
2247 info->control.rates[0].idx = rate_idx;
2248 info->control.rates[0].count = 1; 2211 info->control.rates[0].count = 1;
2249} 2212}
2250 2213
@@ -2261,6 +2224,24 @@ static void *rs_alloc_sta(void *mvm_rate, struct ieee80211_sta *sta,
2261 return &sta_priv->lq_sta; 2224 return &sta_priv->lq_sta;
2262} 2225}
2263 2226
2227static int rs_vht_highest_rx_mcs_index(struct ieee80211_sta_vht_cap *vht_cap,
2228 int nss)
2229{
2230 u16 rx_mcs = le16_to_cpu(vht_cap->vht_mcs.rx_mcs_map) &
2231 (0x3 << (2 * (nss - 1)));
2232 rx_mcs >>= (2 * (nss - 1));
2233
2234 if (rx_mcs == IEEE80211_VHT_MCS_SUPPORT_0_7)
2235 return IWL_RATE_MCS_7_INDEX;
2236 else if (rx_mcs == IEEE80211_VHT_MCS_SUPPORT_0_8)
2237 return IWL_RATE_MCS_8_INDEX;
2238 else if (rx_mcs == IEEE80211_VHT_MCS_SUPPORT_0_9)
2239 return IWL_RATE_MCS_9_INDEX;
2240
2241 WARN_ON_ONCE(rx_mcs != IEEE80211_VHT_MCS_NOT_SUPPORTED);
2242 return -1;
2243}
2244
2264/* 2245/*
2265 * Called after adding a new station to initialize rate scaling 2246 * Called after adding a new station to initialize rate scaling
2266 */ 2247 */
@@ -2270,6 +2251,7 @@ void iwl_mvm_rs_rate_init(struct iwl_mvm *mvm, struct ieee80211_sta *sta,
2270 int i, j; 2251 int i, j;
2271 struct ieee80211_hw *hw = mvm->hw; 2252 struct ieee80211_hw *hw = mvm->hw;
2272 struct ieee80211_sta_ht_cap *ht_cap = &sta->ht_cap; 2253 struct ieee80211_sta_ht_cap *ht_cap = &sta->ht_cap;
2254 struct ieee80211_sta_vht_cap *vht_cap = &sta->vht_cap;
2273 struct iwl_mvm_sta *sta_priv; 2255 struct iwl_mvm_sta *sta_priv;
2274 struct iwl_lq_sta *lq_sta; 2256 struct iwl_lq_sta *lq_sta;
2275 struct ieee80211_supported_band *sband; 2257 struct ieee80211_supported_band *sband;
@@ -2298,7 +2280,6 @@ void iwl_mvm_rs_rate_init(struct iwl_mvm *mvm, struct ieee80211_sta *sta,
2298 2280
2299 lq_sta->max_rate_idx = -1; 2281 lq_sta->max_rate_idx = -1;
2300 lq_sta->missed_rate_counter = IWL_MISSED_RATE_MAX; 2282 lq_sta->missed_rate_counter = IWL_MISSED_RATE_MAX;
2301 lq_sta->is_green = rs_use_green(sta);
2302 lq_sta->band = sband->band; 2283 lq_sta->band = sband->band;
2303 /* 2284 /*
2304 * active legacy rates as per supported rates bitmap 2285 * active legacy rates as per supported rates bitmap
@@ -2308,25 +2289,54 @@ void iwl_mvm_rs_rate_init(struct iwl_mvm *mvm, struct ieee80211_sta *sta,
2308 for_each_set_bit(i, &supp, BITS_PER_LONG) 2289 for_each_set_bit(i, &supp, BITS_PER_LONG)
2309 lq_sta->active_legacy_rate |= BIT(sband->bitrates[i].hw_value); 2290 lq_sta->active_legacy_rate |= BIT(sband->bitrates[i].hw_value);
2310 2291
2311 /* 2292 /* TODO: should probably account for rx_highest for both HT/VHT */
2312 * active_siso_rate mask includes 9 MBits (bit 5), and CCK (bits 0-3), 2293 if (!vht_cap || !vht_cap->vht_supported) {
2313 * supp_rates[] does not; shift to convert format, force 9 MBits off. 2294 /* active_siso_rate mask includes 9 MBits (bit 5),
2314 */ 2295 * and CCK (bits 0-3), supp_rates[] does not;
2315 lq_sta->active_siso_rate = ht_cap->mcs.rx_mask[0] << 1; 2296 * shift to convert format, force 9 MBits off.
2316 lq_sta->active_siso_rate |= ht_cap->mcs.rx_mask[0] & 0x1; 2297 */
2317 lq_sta->active_siso_rate &= ~((u16)0x2); 2298 lq_sta->active_siso_rate = ht_cap->mcs.rx_mask[0] << 1;
2318 lq_sta->active_siso_rate <<= IWL_FIRST_OFDM_RATE; 2299 lq_sta->active_siso_rate |= ht_cap->mcs.rx_mask[0] & 0x1;
2300 lq_sta->active_siso_rate &= ~((u16)0x2);
2301 lq_sta->active_siso_rate <<= IWL_FIRST_OFDM_RATE;
2302
2303 /* Same here */
2304 lq_sta->active_mimo2_rate = ht_cap->mcs.rx_mask[1] << 1;
2305 lq_sta->active_mimo2_rate |= ht_cap->mcs.rx_mask[1] & 0x1;
2306 lq_sta->active_mimo2_rate &= ~((u16)0x2);
2307 lq_sta->active_mimo2_rate <<= IWL_FIRST_OFDM_RATE;
2308
2309 lq_sta->is_vht = false;
2310 } else {
2311 int highest_mcs = rs_vht_highest_rx_mcs_index(vht_cap, 1);
2312 if (highest_mcs >= IWL_RATE_MCS_0_INDEX) {
2313 for (i = IWL_RATE_MCS_0_INDEX; i <= highest_mcs; i++) {
2314 if (i == IWL_RATE_9M_INDEX)
2315 continue;
2316
2317 lq_sta->active_siso_rate |= BIT(i);
2318 }
2319 }
2320
2321 highest_mcs = rs_vht_highest_rx_mcs_index(vht_cap, 2);
2322 if (highest_mcs >= IWL_RATE_MCS_0_INDEX) {
2323 for (i = IWL_RATE_MCS_0_INDEX; i <= highest_mcs; i++) {
2324 if (i == IWL_RATE_9M_INDEX)
2325 continue;
2319 2326
2320 /* Same here */ 2327 lq_sta->active_mimo2_rate |= BIT(i);
2321 lq_sta->active_mimo2_rate = ht_cap->mcs.rx_mask[1] << 1; 2328 }
2322 lq_sta->active_mimo2_rate |= ht_cap->mcs.rx_mask[1] & 0x1; 2329 }
2323 lq_sta->active_mimo2_rate &= ~((u16)0x2); 2330
2324 lq_sta->active_mimo2_rate <<= IWL_FIRST_OFDM_RATE; 2331 /* TODO: avoid MCS9 in 20Mhz which isn't valid for 11ac */
2332 lq_sta->is_vht = true;
2333 }
2325 2334
2326 IWL_DEBUG_RATE(mvm, 2335 IWL_DEBUG_RATE(mvm,
2327 "SISO-RATE=%X MIMO2-RATE=%X\n", 2336 "SISO-RATE=%X MIMO2-RATE=%X VHT=%d\n",
2328 lq_sta->active_siso_rate, 2337 lq_sta->active_siso_rate,
2329 lq_sta->active_mimo2_rate); 2338 lq_sta->active_mimo2_rate,
2339 lq_sta->is_vht);
2330 2340
2331 /* These values will be overridden later */ 2341 /* These values will be overridden later */
2332 lq_sta->lq.single_stream_ant_msk = 2342 lq_sta->lq.single_stream_ant_msk =
@@ -2358,6 +2368,7 @@ void iwl_mvm_rs_rate_init(struct iwl_mvm *mvm, struct ieee80211_sta *sta,
2358} 2368}
2359 2369
2360static void rs_fill_link_cmd(struct iwl_mvm *mvm, 2370static void rs_fill_link_cmd(struct iwl_mvm *mvm,
2371 struct ieee80211_sta *sta,
2361 struct iwl_lq_sta *lq_sta, u32 new_rate) 2372 struct iwl_lq_sta *lq_sta, u32 new_rate)
2362{ 2373{
2363 struct iwl_scale_tbl_info tbl_type; 2374 struct iwl_scale_tbl_info tbl_type;
@@ -2429,7 +2440,6 @@ static void rs_fill_link_cmd(struct iwl_mvm *mvm,
2429 rs_get_tbl_info_from_mcs(new_rate, lq_sta->band, &tbl_type, 2440 rs_get_tbl_info_from_mcs(new_rate, lq_sta->band, &tbl_type,
2430 &rate_idx); 2441 &rate_idx);
2431 2442
2432
2433 /* Indicate to uCode which entries might be MIMO. 2443 /* Indicate to uCode which entries might be MIMO.
2434 * If initial rate was MIMO, this will finally end up 2444 * If initial rate was MIMO, this will finally end up
2435 * as (IWL_HT_NUMBER_TRY * 2), after 2nd pass, otherwise 0. */ 2445 * as (IWL_HT_NUMBER_TRY * 2), after 2nd pass, otherwise 0. */
@@ -2455,7 +2465,9 @@ static void rs_fill_link_cmd(struct iwl_mvm *mvm,
2455 } 2465 }
2456 2466
2457 /* Don't allow HT rates after next pass. 2467 /* Don't allow HT rates after next pass.
2458 * rs_get_lower_rate() will change type to LQ_A or LQ_G. */ 2468 * rs_get_lower_rate() will change type to LQ_LEGACY_A
2469 * or LQ_LEGACY_G.
2470 */
2459 use_ht_possible = 0; 2471 use_ht_possible = 0;
2460 2472
2461 /* Override next rate if needed for debug purposes */ 2473 /* Override next rate if needed for debug purposes */
@@ -2474,12 +2486,9 @@ static void rs_fill_link_cmd(struct iwl_mvm *mvm,
2474 lq_cmd->agg_time_limit = 2486 lq_cmd->agg_time_limit =
2475 cpu_to_le16(LINK_QUAL_AGG_TIME_LIMIT_DEF); 2487 cpu_to_le16(LINK_QUAL_AGG_TIME_LIMIT_DEF);
2476 2488
2477 /* 2489 if (sta)
2478 * overwrite if needed, pass aggregation time limit 2490 lq_cmd->agg_time_limit =
2479 * to uCode in uSec - This is racy - but heh, at least it helps... 2491 cpu_to_le16(iwl_mvm_bt_coex_agg_time_limit(mvm, sta));
2480 */
2481 if (mvm && BT_MBOX_MSG(&mvm->last_bt_notif, 3, TRAFFIC_LOAD) >= 2)
2482 lq_cmd->agg_time_limit = cpu_to_le16(1200);
2483} 2492}
2484 2493
2485static void *rs_alloc(struct ieee80211_hw *hw, struct dentry *debugfsdir) 2494static void *rs_alloc(struct ieee80211_hw *hw, struct dentry *debugfsdir)
@@ -2586,16 +2595,18 @@ static ssize_t rs_sta_dbgfs_scale_table_read(struct file *file,
2586 (iwl_fw_valid_tx_ant(mvm->fw) & ANT_B) ? "ANT_B," : "", 2595 (iwl_fw_valid_tx_ant(mvm->fw) & ANT_B) ? "ANT_B," : "",
2587 (iwl_fw_valid_tx_ant(mvm->fw) & ANT_C) ? "ANT_C" : ""); 2596 (iwl_fw_valid_tx_ant(mvm->fw) & ANT_C) ? "ANT_C" : "");
2588 desc += sprintf(buff+desc, "lq type %s\n", 2597 desc += sprintf(buff+desc, "lq type %s\n",
2589 (is_legacy(tbl->lq_type)) ? "legacy" : "HT"); 2598 (is_legacy(tbl->lq_type)) ? "legacy" :
2590 if (is_Ht(tbl->lq_type)) { 2599 is_vht(tbl->lq_type) ? "VHT" : "HT");
2600 if (is_ht(tbl->lq_type)) {
2591 desc += sprintf(buff+desc, " %s", 2601 desc += sprintf(buff+desc, " %s",
2592 (is_siso(tbl->lq_type)) ? "SISO" : "MIMO2"); 2602 (is_siso(tbl->lq_type)) ? "SISO" : "MIMO2");
2593 desc += sprintf(buff+desc, " %s", 2603 desc += sprintf(buff+desc, " %s",
2594 (tbl->is_ht40) ? "40MHz" : "20MHz"); 2604 (is_ht20(tbl)) ? "20MHz" :
2595 desc += sprintf(buff+desc, " %s %s %s\n", 2605 (is_ht40(tbl)) ? "40MHz" :
2606 (is_ht80(tbl)) ? "80Mhz" : "BAD BW");
2607 desc += sprintf(buff+desc, " %s %s\n",
2596 (tbl->is_SGI) ? "SGI" : "", 2608 (tbl->is_SGI) ? "SGI" : "",
2597 (lq_sta->is_green) ? "GF enabled" : "", 2609 (lq_sta->is_agg) ? "AGG on" : "");
2598 (lq_sta->is_agg) ? "AGG on" : "");
2599 } 2610 }
2600 desc += sprintf(buff+desc, "last tx rate=0x%X\n", 2611 desc += sprintf(buff+desc, "last tx rate=0x%X\n",
2601 lq_sta->last_rate_n_flags); 2612 lq_sta->last_rate_n_flags);
@@ -2653,7 +2664,7 @@ static ssize_t rs_sta_dbgfs_stats_table_read(struct file *file,
2653 int desc = 0; 2664 int desc = 0;
2654 int i, j; 2665 int i, j;
2655 ssize_t ret; 2666 ssize_t ret;
2656 2667 struct iwl_scale_tbl_info *tbl;
2657 struct iwl_lq_sta *lq_sta = file->private_data; 2668 struct iwl_lq_sta *lq_sta = file->private_data;
2658 2669
2659 buff = kmalloc(1024, GFP_KERNEL); 2670 buff = kmalloc(1024, GFP_KERNEL);
@@ -2661,21 +2672,23 @@ static ssize_t rs_sta_dbgfs_stats_table_read(struct file *file,
2661 return -ENOMEM; 2672 return -ENOMEM;
2662 2673
2663 for (i = 0; i < LQ_SIZE; i++) { 2674 for (i = 0; i < LQ_SIZE; i++) {
2675 tbl = &(lq_sta->lq_info[i]);
2664 desc += sprintf(buff+desc, 2676 desc += sprintf(buff+desc,
2665 "%s type=%d SGI=%d HT40=%d DUP=0 GF=%d\n" 2677 "%s type=%d SGI=%d BW=%s DUP=0\n"
2666 "rate=0x%X\n", 2678 "rate=0x%X\n",
2667 lq_sta->active_tbl == i ? "*" : "x", 2679 lq_sta->active_tbl == i ? "*" : "x",
2668 lq_sta->lq_info[i].lq_type, 2680 tbl->lq_type,
2669 lq_sta->lq_info[i].is_SGI, 2681 tbl->is_SGI,
2670 lq_sta->lq_info[i].is_ht40, 2682 is_ht20(tbl) ? "20Mhz" :
2671 lq_sta->is_green, 2683 is_ht40(tbl) ? "40Mhz" :
2672 lq_sta->lq_info[i].current_rate); 2684 is_ht80(tbl) ? "80Mhz" : "ERR",
2685 tbl->current_rate);
2673 for (j = 0; j < IWL_RATE_COUNT; j++) { 2686 for (j = 0; j < IWL_RATE_COUNT; j++) {
2674 desc += sprintf(buff+desc, 2687 desc += sprintf(buff+desc,
2675 "counter=%d success=%d %%=%d\n", 2688 "counter=%d success=%d %%=%d\n",
2676 lq_sta->lq_info[i].win[j].counter, 2689 tbl->win[j].counter,
2677 lq_sta->lq_info[i].win[j].success_counter, 2690 tbl->win[j].success_counter,
2678 lq_sta->lq_info[i].win[j].success_ratio); 2691 tbl->win[j].success_ratio);
2679 } 2692 }
2680 } 2693 }
2681 ret = simple_read_from_buffer(user_buf, count, ppos, buff, desc); 2694 ret = simple_read_from_buffer(user_buf, count, ppos, buff, desc);
diff --git a/drivers/net/wireless/iwlwifi/mvm/rs.h b/drivers/net/wireless/iwlwifi/mvm/rs.h
index 335cf1682902..5d5344f7070b 100644
--- a/drivers/net/wireless/iwlwifi/mvm/rs.h
+++ b/drivers/net/wireless/iwlwifi/mvm/rs.h
@@ -35,9 +35,11 @@
35#include "iwl-trans.h" 35#include "iwl-trans.h"
36 36
37struct iwl_rs_rate_info { 37struct iwl_rs_rate_info {
38 u8 plcp; /* uCode API: IWL_RATE_6M_PLCP, etc. */ 38 u8 plcp; /* uCode API: IWL_RATE_6M_PLCP, etc. */
39 u8 plcp_siso; /* uCode API: IWL_RATE_SISO_6M_PLCP, etc. */ 39 u8 plcp_ht_siso; /* uCode API: IWL_RATE_SISO_6M_PLCP, etc. */
40 u8 plcp_mimo2; /* uCode API: IWL_RATE_MIMO2_6M_PLCP, etc. */ 40 u8 plcp_ht_mimo2; /* uCode API: IWL_RATE_MIMO2_6M_PLCP, etc. */
41 u8 plcp_vht_siso;
42 u8 plcp_vht_mimo2;
41 u8 prev_rs; /* previous rate used in rs algo */ 43 u8 prev_rs; /* previous rate used in rs algo */
42 u8 next_rs; /* next rate used in rs algo */ 44 u8 next_rs; /* next rate used in rs algo */
43}; 45};
@@ -83,35 +85,52 @@ enum {
83#define IWL_RATE_11M_MASK (1 << IWL_RATE_11M_INDEX) 85#define IWL_RATE_11M_MASK (1 << IWL_RATE_11M_INDEX)
84 86
85 87
86/* uCode API values for OFDM high-throughput (HT) bit rates */ 88/* uCode API values for HT/VHT bit rates */
87enum { 89enum {
88 IWL_RATE_SISO_6M_PLCP = 0, 90 IWL_RATE_HT_SISO_MCS_0_PLCP = 0,
89 IWL_RATE_SISO_12M_PLCP = 1, 91 IWL_RATE_HT_SISO_MCS_1_PLCP = 1,
90 IWL_RATE_SISO_18M_PLCP = 2, 92 IWL_RATE_HT_SISO_MCS_2_PLCP = 2,
91 IWL_RATE_SISO_24M_PLCP = 3, 93 IWL_RATE_HT_SISO_MCS_3_PLCP = 3,
92 IWL_RATE_SISO_36M_PLCP = 4, 94 IWL_RATE_HT_SISO_MCS_4_PLCP = 4,
93 IWL_RATE_SISO_48M_PLCP = 5, 95 IWL_RATE_HT_SISO_MCS_5_PLCP = 5,
94 IWL_RATE_SISO_54M_PLCP = 6, 96 IWL_RATE_HT_SISO_MCS_6_PLCP = 6,
95 IWL_RATE_SISO_60M_PLCP = 7, 97 IWL_RATE_HT_SISO_MCS_7_PLCP = 7,
96 IWL_RATE_MIMO2_6M_PLCP = 0x8, 98 IWL_RATE_HT_MIMO2_MCS_0_PLCP = 0x8,
97 IWL_RATE_MIMO2_12M_PLCP = 0x9, 99 IWL_RATE_HT_MIMO2_MCS_1_PLCP = 0x9,
98 IWL_RATE_MIMO2_18M_PLCP = 0xa, 100 IWL_RATE_HT_MIMO2_MCS_2_PLCP = 0xA,
99 IWL_RATE_MIMO2_24M_PLCP = 0xb, 101 IWL_RATE_HT_MIMO2_MCS_3_PLCP = 0xB,
100 IWL_RATE_MIMO2_36M_PLCP = 0xc, 102 IWL_RATE_HT_MIMO2_MCS_4_PLCP = 0xC,
101 IWL_RATE_MIMO2_48M_PLCP = 0xd, 103 IWL_RATE_HT_MIMO2_MCS_5_PLCP = 0xD,
102 IWL_RATE_MIMO2_54M_PLCP = 0xe, 104 IWL_RATE_HT_MIMO2_MCS_6_PLCP = 0xE,
103 IWL_RATE_MIMO2_60M_PLCP = 0xf, 105 IWL_RATE_HT_MIMO2_MCS_7_PLCP = 0xF,
104 IWL_RATE_MIMO3_6M_PLCP = 0x10, 106 IWL_RATE_VHT_SISO_MCS_0_PLCP = 0,
105 IWL_RATE_MIMO3_12M_PLCP = 0x11, 107 IWL_RATE_VHT_SISO_MCS_1_PLCP = 1,
106 IWL_RATE_MIMO3_18M_PLCP = 0x12, 108 IWL_RATE_VHT_SISO_MCS_2_PLCP = 2,
107 IWL_RATE_MIMO3_24M_PLCP = 0x13, 109 IWL_RATE_VHT_SISO_MCS_3_PLCP = 3,
108 IWL_RATE_MIMO3_36M_PLCP = 0x14, 110 IWL_RATE_VHT_SISO_MCS_4_PLCP = 4,
109 IWL_RATE_MIMO3_48M_PLCP = 0x15, 111 IWL_RATE_VHT_SISO_MCS_5_PLCP = 5,
110 IWL_RATE_MIMO3_54M_PLCP = 0x16, 112 IWL_RATE_VHT_SISO_MCS_6_PLCP = 6,
111 IWL_RATE_MIMO3_60M_PLCP = 0x17, 113 IWL_RATE_VHT_SISO_MCS_7_PLCP = 7,
112 IWL_RATE_SISO_INVM_PLCP, 114 IWL_RATE_VHT_SISO_MCS_8_PLCP = 8,
113 IWL_RATE_MIMO2_INVM_PLCP = IWL_RATE_SISO_INVM_PLCP, 115 IWL_RATE_VHT_SISO_MCS_9_PLCP = 9,
114 IWL_RATE_MIMO3_INVM_PLCP = IWL_RATE_SISO_INVM_PLCP, 116 IWL_RATE_VHT_MIMO2_MCS_0_PLCP = 0x10,
117 IWL_RATE_VHT_MIMO2_MCS_1_PLCP = 0x11,
118 IWL_RATE_VHT_MIMO2_MCS_2_PLCP = 0x12,
119 IWL_RATE_VHT_MIMO2_MCS_3_PLCP = 0x13,
120 IWL_RATE_VHT_MIMO2_MCS_4_PLCP = 0x14,
121 IWL_RATE_VHT_MIMO2_MCS_5_PLCP = 0x15,
122 IWL_RATE_VHT_MIMO2_MCS_6_PLCP = 0x16,
123 IWL_RATE_VHT_MIMO2_MCS_7_PLCP = 0x17,
124 IWL_RATE_VHT_MIMO2_MCS_8_PLCP = 0x18,
125 IWL_RATE_VHT_MIMO2_MCS_9_PLCP = 0x19,
126 IWL_RATE_HT_SISO_MCS_INV_PLCP,
127 IWL_RATE_HT_MIMO2_MCS_INV_PLCP = IWL_RATE_HT_SISO_MCS_INV_PLCP,
128 IWL_RATE_VHT_SISO_MCS_INV_PLCP = IWL_RATE_HT_SISO_MCS_INV_PLCP,
129 IWL_RATE_VHT_MIMO2_MCS_INV_PLCP = IWL_RATE_HT_SISO_MCS_INV_PLCP,
130 IWL_RATE_HT_SISO_MCS_8_PLCP = IWL_RATE_HT_SISO_MCS_INV_PLCP,
131 IWL_RATE_HT_SISO_MCS_9_PLCP = IWL_RATE_HT_SISO_MCS_INV_PLCP,
132 IWL_RATE_HT_MIMO2_MCS_8_PLCP = IWL_RATE_HT_SISO_MCS_INV_PLCP,
133 IWL_RATE_HT_MIMO2_MCS_9_PLCP = IWL_RATE_HT_SISO_MCS_INV_PLCP,
115}; 134};
116 135
117#define IWL_RATES_MASK ((1 << IWL_RATE_COUNT) - 1) 136#define IWL_RATES_MASK ((1 << IWL_RATE_COUNT) - 1)
@@ -139,25 +158,33 @@ enum {
139#define IWL_RATE_DECREASE_TH 1920 /* 15% */ 158#define IWL_RATE_DECREASE_TH 1920 /* 15% */
140 159
141/* possible actions when in legacy mode */ 160/* possible actions when in legacy mode */
142#define IWL_LEGACY_SWITCH_ANTENNA1 0 161enum {
143#define IWL_LEGACY_SWITCH_ANTENNA2 1 162 IWL_LEGACY_SWITCH_ANTENNA,
144#define IWL_LEGACY_SWITCH_SISO 2 163 IWL_LEGACY_SWITCH_SISO,
145#define IWL_LEGACY_SWITCH_MIMO2 3 164 IWL_LEGACY_SWITCH_MIMO2,
165 IWL_LEGACY_FIRST_ACTION = IWL_LEGACY_SWITCH_ANTENNA,
166 IWL_LEGACY_LAST_ACTION = IWL_LEGACY_SWITCH_MIMO2,
167};
146 168
147/* possible actions when in siso mode */ 169/* possible actions when in siso mode */
148#define IWL_SISO_SWITCH_ANTENNA1 0 170enum {
149#define IWL_SISO_SWITCH_ANTENNA2 1 171 IWL_SISO_SWITCH_ANTENNA,
150#define IWL_SISO_SWITCH_MIMO2 2 172 IWL_SISO_SWITCH_MIMO2,
151#define IWL_SISO_SWITCH_GI 3 173 IWL_SISO_SWITCH_GI,
174 IWL_SISO_FIRST_ACTION = IWL_SISO_SWITCH_ANTENNA,
175 IWL_SISO_LAST_ACTION = IWL_SISO_SWITCH_GI,
176};
152 177
153/* possible actions when in mimo mode */ 178/* possible actions when in mimo mode */
154#define IWL_MIMO2_SWITCH_ANTENNA1 0 179enum {
155#define IWL_MIMO2_SWITCH_ANTENNA2 1 180 IWL_MIMO2_SWITCH_SISO_A,
156#define IWL_MIMO2_SWITCH_SISO_A 2 181 IWL_MIMO2_SWITCH_SISO_B,
157#define IWL_MIMO2_SWITCH_SISO_B 3 182 IWL_MIMO2_SWITCH_GI,
158#define IWL_MIMO2_SWITCH_GI 4 183 IWL_MIMO2_FIRST_ACTION = IWL_MIMO2_SWITCH_SISO_A,
184 IWL_MIMO2_LAST_ACTION = IWL_MIMO2_SWITCH_GI,
185};
159 186
160#define IWL_MAX_SEARCH IWL_MIMO2_SWITCH_GI 187#define IWL_MAX_SEARCH IWL_MIMO2_LAST_ACTION
161 188
162#define IWL_ACTION_LIMIT 3 /* # possible actions */ 189#define IWL_ACTION_LIMIT 3 /* # possible actions */
163 190
@@ -188,20 +215,31 @@ enum {
188 215
189enum iwl_table_type { 216enum iwl_table_type {
190 LQ_NONE, 217 LQ_NONE,
191 LQ_G, /* legacy types */ 218 LQ_LEGACY_G, /* legacy types */
192 LQ_A, 219 LQ_LEGACY_A,
193 LQ_SISO, /* high-throughput types */ 220 LQ_HT_SISO, /* HT types */
194 LQ_MIMO2, 221 LQ_HT_MIMO2,
222 LQ_VHT_SISO, /* VHT types */
223 LQ_VHT_MIMO2,
195 LQ_MAX, 224 LQ_MAX,
196}; 225};
197 226
198#define is_legacy(tbl) (((tbl) == LQ_G) || ((tbl) == LQ_A)) 227#define is_legacy(tbl) (((tbl) == LQ_LEGACY_G) || ((tbl) == LQ_LEGACY_A))
199#define is_siso(tbl) ((tbl) == LQ_SISO) 228#define is_ht_siso(tbl) ((tbl) == LQ_HT_SISO)
200#define is_mimo2(tbl) ((tbl) == LQ_MIMO2) 229#define is_ht_mimo2(tbl) ((tbl) == LQ_HT_MIMO2)
201#define is_mimo(tbl) is_mimo2(tbl) 230#define is_vht_siso(tbl) ((tbl) == LQ_VHT_SISO)
202#define is_Ht(tbl) (is_siso(tbl) || is_mimo(tbl)) 231#define is_vht_mimo2(tbl) ((tbl) == LQ_VHT_MIMO2)
203#define is_a_band(tbl) ((tbl) == LQ_A) 232#define is_siso(tbl) (is_ht_siso(tbl) || is_vht_siso(tbl))
204#define is_g_and(tbl) ((tbl) == LQ_G) 233#define is_mimo2(tbl) (is_ht_mimo2(tbl) || is_vht_mimo2(tbl))
234#define is_mimo(tbl) (is_mimo2(tbl))
235#define is_ht(tbl) (is_ht_siso(tbl) || is_ht_mimo2(tbl))
236#define is_vht(tbl) (is_vht_siso(tbl) || is_vht_mimo2(tbl))
237#define is_a_band(tbl) ((tbl) == LQ_LEGACY_A)
238#define is_g_band(tbl) ((tbl) == LQ_LEGACY_G)
239
240#define is_ht20(tbl) (tbl->bw == RATE_MCS_CHAN_WIDTH_20)
241#define is_ht40(tbl) (tbl->bw == RATE_MCS_CHAN_WIDTH_40)
242#define is_ht80(tbl) (tbl->bw == RATE_MCS_CHAN_WIDTH_80)
205 243
206#define IWL_MAX_MCS_DISPLAY_SIZE 12 244#define IWL_MAX_MCS_DISPLAY_SIZE 12
207 245
@@ -232,7 +270,7 @@ struct iwl_scale_tbl_info {
232 enum iwl_table_type lq_type; 270 enum iwl_table_type lq_type;
233 u8 ant_type; 271 u8 ant_type;
234 u8 is_SGI; /* 1 = short guard interval */ 272 u8 is_SGI; /* 1 = short guard interval */
235 u8 is_ht40; /* 1 = 40 MHz channel width */ 273 u32 bw; /* channel bandwidth; RATE_MCS_CHAN_WIDTH_XX */
236 u8 action; /* change modulation; IWL_[LEGACY/SISO/MIMO]_SWITCH_* */ 274 u8 action; /* change modulation; IWL_[LEGACY/SISO/MIMO]_SWITCH_* */
237 u8 max_search; /* maximun number of tables we can search */ 275 u8 max_search; /* maximun number of tables we can search */
238 s32 *expected_tpt; /* throughput metrics; expected_tpt_G, etc. */ 276 s32 *expected_tpt; /* throughput metrics; expected_tpt_G, etc. */
@@ -262,7 +300,7 @@ struct iwl_lq_sta {
262 u64 flush_timer; /* time staying in mode before new search */ 300 u64 flush_timer; /* time staying in mode before new search */
263 301
264 u8 action_counter; /* # mode-switch actions tried */ 302 u8 action_counter; /* # mode-switch actions tried */
265 u8 is_green; 303 bool is_vht;
266 enum ieee80211_band band; 304 enum ieee80211_band band;
267 305
268 /* The following are bitmaps of rates; IWL_RATE_6M_MASK, etc. */ 306 /* The following are bitmaps of rates; IWL_RATE_6M_MASK, etc. */
@@ -314,9 +352,8 @@ static inline u8 num_of_ant(u8 mask)
314} 352}
315 353
316/* Initialize station's rate scaling information after adding station */ 354/* Initialize station's rate scaling information after adding station */
317extern void iwl_mvm_rs_rate_init(struct iwl_mvm *mvm, 355void iwl_mvm_rs_rate_init(struct iwl_mvm *mvm, struct ieee80211_sta *sta,
318 struct ieee80211_sta *sta, 356 enum ieee80211_band band);
319 enum ieee80211_band band);
320 357
321/** 358/**
322 * iwl_rate_control_register - Register the rate control algorithm callbacks 359 * iwl_rate_control_register - Register the rate control algorithm callbacks
@@ -328,7 +365,7 @@ extern void iwl_mvm_rs_rate_init(struct iwl_mvm *mvm,
328 * ieee80211_register_hw 365 * ieee80211_register_hw
329 * 366 *
330 */ 367 */
331extern int iwl_mvm_rate_control_register(void); 368int iwl_mvm_rate_control_register(void);
332 369
333/** 370/**
334 * iwl_rate_control_unregister - Unregister the rate control callbacks 371 * iwl_rate_control_unregister - Unregister the rate control callbacks
@@ -336,7 +373,7 @@ extern int iwl_mvm_rate_control_register(void);
336 * This should be called after calling ieee80211_unregister_hw, but before 373 * This should be called after calling ieee80211_unregister_hw, but before
337 * the driver is unloaded. 374 * the driver is unloaded.
338 */ 375 */
339extern void iwl_mvm_rate_control_unregister(void); 376void iwl_mvm_rate_control_unregister(void);
340 377
341struct iwl_mvm_sta; 378struct iwl_mvm_sta;
342 379
diff --git a/drivers/net/wireless/iwlwifi/mvm/rx.c b/drivers/net/wireless/iwlwifi/mvm/rx.c
index 2a8cb5a60535..3a1f3982109d 100644
--- a/drivers/net/wireless/iwlwifi/mvm/rx.c
+++ b/drivers/net/wireless/iwlwifi/mvm/rx.c
@@ -300,10 +300,14 @@ int iwl_mvm_rx_rx_mpdu(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb,
300 return 0; 300 return 0;
301 } 301 }
302 302
303 /*
304 * Keep packets with CRC errors (and with overrun) for monitor mode
305 * (otherwise the firmware discards them) but mark them as bad.
306 */
303 if (!(rx_pkt_status & RX_MPDU_RES_STATUS_CRC_OK) || 307 if (!(rx_pkt_status & RX_MPDU_RES_STATUS_CRC_OK) ||
304 !(rx_pkt_status & RX_MPDU_RES_STATUS_OVERRUN_OK)) { 308 !(rx_pkt_status & RX_MPDU_RES_STATUS_OVERRUN_OK)) {
305 IWL_DEBUG_RX(mvm, "Bad CRC or FIFO: 0x%08X.\n", rx_pkt_status); 309 IWL_DEBUG_RX(mvm, "Bad CRC or FIFO: 0x%08X.\n", rx_pkt_status);
306 return 0; 310 rx_status.flag |= RX_FLAG_FAILED_FCS_CRC;
307 } 311 }
308 312
309 /* This will be used in several places later */ 313 /* This will be used in several places later */
@@ -422,6 +426,27 @@ static void iwl_mvm_stat_iterator(void *_data, u8 *mac,
422 426
423 mvmvif->bf_data.ave_beacon_signal = sig; 427 mvmvif->bf_data.ave_beacon_signal = sig;
424 428
429 /* BT Coex */
430 if (mvmvif->bf_data.bt_coex_min_thold !=
431 mvmvif->bf_data.bt_coex_max_thold) {
432 last_event = mvmvif->bf_data.last_bt_coex_event;
433 if (sig > mvmvif->bf_data.bt_coex_max_thold &&
434 (last_event <= mvmvif->bf_data.bt_coex_min_thold ||
435 last_event == 0)) {
436 mvmvif->bf_data.last_bt_coex_event = sig;
437 IWL_DEBUG_RX(mvm, "cqm_iterator bt coex high %d\n",
438 sig);
439 iwl_mvm_bt_rssi_event(mvm, vif, RSSI_EVENT_HIGH);
440 } else if (sig < mvmvif->bf_data.bt_coex_min_thold &&
441 (last_event >= mvmvif->bf_data.bt_coex_max_thold ||
442 last_event == 0)) {
443 mvmvif->bf_data.last_bt_coex_event = sig;
444 IWL_DEBUG_RX(mvm, "cqm_iterator bt coex low %d\n",
445 sig);
446 iwl_mvm_bt_rssi_event(mvm, vif, RSSI_EVENT_LOW);
447 }
448 }
449
425 if (!(vif->driver_flags & IEEE80211_VIF_SUPPORTS_CQM_RSSI)) 450 if (!(vif->driver_flags & IEEE80211_VIF_SUPPORTS_CQM_RSSI))
426 return; 451 return;
427 452
diff --git a/drivers/net/wireless/iwlwifi/mvm/scan.c b/drivers/net/wireless/iwlwifi/mvm/scan.c
index 621fb71f282a..dff7592e1ff8 100644
--- a/drivers/net/wireless/iwlwifi/mvm/scan.c
+++ b/drivers/net/wireless/iwlwifi/mvm/scan.c
@@ -74,8 +74,12 @@
74static inline __le16 iwl_mvm_scan_rx_chain(struct iwl_mvm *mvm) 74static inline __le16 iwl_mvm_scan_rx_chain(struct iwl_mvm *mvm)
75{ 75{
76 u16 rx_chain; 76 u16 rx_chain;
77 u8 rx_ant = iwl_fw_valid_rx_ant(mvm->fw); 77 u8 rx_ant;
78 78
79 if (mvm->scan_rx_ant != ANT_NONE)
80 rx_ant = mvm->scan_rx_ant;
81 else
82 rx_ant = iwl_fw_valid_rx_ant(mvm->fw);
79 rx_chain = rx_ant << PHY_RX_CHAIN_VALID_POS; 83 rx_chain = rx_ant << PHY_RX_CHAIN_VALID_POS;
80 rx_chain |= rx_ant << PHY_RX_CHAIN_FORCE_MIMO_SEL_POS; 84 rx_chain |= rx_ant << PHY_RX_CHAIN_FORCE_MIMO_SEL_POS;
81 rx_chain |= rx_ant << PHY_RX_CHAIN_FORCE_SEL_POS; 85 rx_chain |= rx_ant << PHY_RX_CHAIN_FORCE_SEL_POS;
@@ -93,10 +97,10 @@ static inline __le32 iwl_mvm_scan_max_out_time(struct ieee80211_vif *vif)
93 97
94static inline __le32 iwl_mvm_scan_suspend_time(struct ieee80211_vif *vif) 98static inline __le32 iwl_mvm_scan_suspend_time(struct ieee80211_vif *vif)
95{ 99{
96 if (vif->bss_conf.assoc) 100 if (!vif->bss_conf.assoc)
97 return cpu_to_le32(vif->bss_conf.beacon_int);
98 else
99 return 0; 101 return 0;
102
103 return cpu_to_le32(ieee80211_tu_to_usec(vif->bss_conf.beacon_int));
100} 104}
101 105
102static inline __le32 106static inline __le32
@@ -133,11 +137,12 @@ iwl_mvm_scan_rate_n_flags(struct iwl_mvm *mvm, enum ieee80211_band band,
133 * request. 137 * request.
134 */ 138 */
135static void iwl_mvm_scan_fill_ssids(struct iwl_scan_cmd *cmd, 139static void iwl_mvm_scan_fill_ssids(struct iwl_scan_cmd *cmd,
136 struct cfg80211_scan_request *req) 140 struct cfg80211_scan_request *req,
141 int first)
137{ 142{
138 int fw_idx, req_idx; 143 int fw_idx, req_idx;
139 144
140 for (req_idx = req->n_ssids - 1, fw_idx = 0; req_idx > 0; 145 for (req_idx = req->n_ssids - 1, fw_idx = 0; req_idx >= first;
141 req_idx--, fw_idx++) { 146 req_idx--, fw_idx++) {
142 cmd->direct_scan[fw_idx].id = WLAN_EID_SSID; 147 cmd->direct_scan[fw_idx].id = WLAN_EID_SSID;
143 cmd->direct_scan[fw_idx].len = req->ssids[req_idx].ssid_len; 148 cmd->direct_scan[fw_idx].len = req->ssids[req_idx].ssid_len;
@@ -153,9 +158,9 @@ static void iwl_mvm_scan_fill_ssids(struct iwl_scan_cmd *cmd,
153 * just to notify that this scan is active and not passive. 158 * just to notify that this scan is active and not passive.
154 * In order to notify the FW of the number of SSIDs we wish to scan (including 159 * In order to notify the FW of the number of SSIDs we wish to scan (including
155 * the zero-length one), we need to set the corresponding bits in chan->type, 160 * the zero-length one), we need to set the corresponding bits in chan->type,
156 * one for each SSID, and set the active bit (first). The first SSID is already 161 * one for each SSID, and set the active bit (first). If the first SSID is
157 * included in the probe template, so we need to set only req->n_ssids - 1 bits 162 * already included in the probe template, so we need to set only
158 * in addition to the first bit. 163 * req->n_ssids - 1 bits in addition to the first bit.
159 */ 164 */
160static u16 iwl_mvm_get_active_dwell(enum ieee80211_band band, int n_ssids) 165static u16 iwl_mvm_get_active_dwell(enum ieee80211_band band, int n_ssids)
161{ 166{
@@ -170,7 +175,8 @@ static u16 iwl_mvm_get_passive_dwell(enum ieee80211_band band)
170} 175}
171 176
172static void iwl_mvm_scan_fill_channels(struct iwl_scan_cmd *cmd, 177static void iwl_mvm_scan_fill_channels(struct iwl_scan_cmd *cmd,
173 struct cfg80211_scan_request *req) 178 struct cfg80211_scan_request *req,
179 bool basic_ssid)
174{ 180{
175 u16 passive_dwell = iwl_mvm_get_passive_dwell(req->channels[0]->band); 181 u16 passive_dwell = iwl_mvm_get_passive_dwell(req->channels[0]->band);
176 u16 active_dwell = iwl_mvm_get_active_dwell(req->channels[0]->band, 182 u16 active_dwell = iwl_mvm_get_active_dwell(req->channels[0]->band,
@@ -178,10 +184,14 @@ static void iwl_mvm_scan_fill_channels(struct iwl_scan_cmd *cmd,
178 struct iwl_scan_channel *chan = (struct iwl_scan_channel *) 184 struct iwl_scan_channel *chan = (struct iwl_scan_channel *)
179 (cmd->data + le16_to_cpu(cmd->tx_cmd.len)); 185 (cmd->data + le16_to_cpu(cmd->tx_cmd.len));
180 int i; 186 int i;
187 int type = BIT(req->n_ssids) - 1;
188
189 if (!basic_ssid)
190 type |= BIT(req->n_ssids);
181 191
182 for (i = 0; i < cmd->channel_count; i++) { 192 for (i = 0; i < cmd->channel_count; i++) {
183 chan->channel = cpu_to_le16(req->channels[i]->hw_value); 193 chan->channel = cpu_to_le16(req->channels[i]->hw_value);
184 chan->type = cpu_to_le32(BIT(req->n_ssids) - 1); 194 chan->type = cpu_to_le32(type);
185 if (req->channels[i]->flags & IEEE80211_CHAN_PASSIVE_SCAN) 195 if (req->channels[i]->flags & IEEE80211_CHAN_PASSIVE_SCAN)
186 chan->type &= cpu_to_le32(~SCAN_CHANNEL_TYPE_ACTIVE); 196 chan->type &= cpu_to_le32(~SCAN_CHANNEL_TYPE_ACTIVE);
187 chan->active_dwell = cpu_to_le16(active_dwell); 197 chan->active_dwell = cpu_to_le16(active_dwell);
@@ -268,6 +278,8 @@ int iwl_mvm_scan_request(struct iwl_mvm *mvm,
268 u32 status; 278 u32 status;
269 int ssid_len = 0; 279 int ssid_len = 0;
270 u8 *ssid = NULL; 280 u8 *ssid = NULL;
281 bool basic_ssid = !(mvm->fw->ucode_capa.flags &
282 IWL_UCODE_TLV_FLAGS_NO_BASIC_SSID);
271 283
272 lockdep_assert_held(&mvm->mutex); 284 lockdep_assert_held(&mvm->mutex);
273 BUG_ON(mvm->scan_cmd == NULL); 285 BUG_ON(mvm->scan_cmd == NULL);
@@ -302,14 +314,16 @@ int iwl_mvm_scan_request(struct iwl_mvm *mvm,
302 if (req->n_ssids > 0) { 314 if (req->n_ssids > 0) {
303 cmd->passive2active = cpu_to_le16(1); 315 cmd->passive2active = cpu_to_le16(1);
304 cmd->scan_flags |= SCAN_FLAGS_PASSIVE2ACTIVE; 316 cmd->scan_flags |= SCAN_FLAGS_PASSIVE2ACTIVE;
305 ssid = req->ssids[0].ssid; 317 if (basic_ssid) {
306 ssid_len = req->ssids[0].ssid_len; 318 ssid = req->ssids[0].ssid;
319 ssid_len = req->ssids[0].ssid_len;
320 }
307 } else { 321 } else {
308 cmd->passive2active = 0; 322 cmd->passive2active = 0;
309 cmd->scan_flags &= ~SCAN_FLAGS_PASSIVE2ACTIVE; 323 cmd->scan_flags &= ~SCAN_FLAGS_PASSIVE2ACTIVE;
310 } 324 }
311 325
312 iwl_mvm_scan_fill_ssids(cmd, req); 326 iwl_mvm_scan_fill_ssids(cmd, req, basic_ssid ? 1 : 0);
313 327
314 cmd->tx_cmd.tx_flags = cpu_to_le32(TX_CMD_FLG_SEQ_CTL); 328 cmd->tx_cmd.tx_flags = cpu_to_le32(TX_CMD_FLG_SEQ_CTL);
315 cmd->tx_cmd.sta_id = mvm->aux_sta.sta_id; 329 cmd->tx_cmd.sta_id = mvm->aux_sta.sta_id;
@@ -326,7 +340,7 @@ int iwl_mvm_scan_request(struct iwl_mvm *mvm,
326 req->ie, req->ie_len, 340 req->ie, req->ie_len,
327 mvm->fw->ucode_capa.max_probe_length)); 341 mvm->fw->ucode_capa.max_probe_length));
328 342
329 iwl_mvm_scan_fill_channels(cmd, req); 343 iwl_mvm_scan_fill_channels(cmd, req, basic_ssid);
330 344
331 cmd->len = cpu_to_le16(sizeof(struct iwl_scan_cmd) + 345 cmd->len = cpu_to_le16(sizeof(struct iwl_scan_cmd) +
332 le16_to_cpu(cmd->tx_cmd.len) + 346 le16_to_cpu(cmd->tx_cmd.len) +
@@ -377,6 +391,21 @@ int iwl_mvm_rx_scan_complete(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb,
377 return 0; 391 return 0;
378} 392}
379 393
394int iwl_mvm_rx_sched_scan_results(struct iwl_mvm *mvm,
395 struct iwl_rx_cmd_buffer *rxb,
396 struct iwl_device_cmd *cmd)
397{
398 struct iwl_rx_packet *pkt = rxb_addr(rxb);
399 struct iwl_sched_scan_results *notif = (void *)pkt->data;
400
401 if (notif->client_bitmap & SCAN_CLIENT_SCHED_SCAN) {
402 IWL_DEBUG_SCAN(mvm, "Scheduled scan results\n");
403 ieee80211_sched_scan_results(mvm->hw);
404 }
405
406 return 0;
407}
408
380static bool iwl_mvm_scan_abort_notif(struct iwl_notif_wait_data *notif_wait, 409static bool iwl_mvm_scan_abort_notif(struct iwl_notif_wait_data *notif_wait,
381 struct iwl_rx_packet *pkt, void *data) 410 struct iwl_rx_packet *pkt, void *data)
382{ 411{
@@ -447,3 +476,406 @@ void iwl_mvm_cancel_scan(struct iwl_mvm *mvm)
447out_remove_notif: 476out_remove_notif:
448 iwl_remove_notification(&mvm->notif_wait, &wait_scan_abort); 477 iwl_remove_notification(&mvm->notif_wait, &wait_scan_abort);
449} 478}
479
480int iwl_mvm_rx_scan_offload_complete_notif(struct iwl_mvm *mvm,
481 struct iwl_rx_cmd_buffer *rxb,
482 struct iwl_device_cmd *cmd)
483{
484 struct iwl_rx_packet *pkt = rxb_addr(rxb);
485 struct iwl_scan_offload_complete *scan_notif = (void *)pkt->data;
486
487 IWL_DEBUG_SCAN(mvm, "Scheduled scan completed, status %s\n",
488 scan_notif->status == IWL_SCAN_OFFLOAD_COMPLETED ?
489 "completed" : "aborted");
490
491 mvm->scan_status = IWL_MVM_SCAN_NONE;
492 ieee80211_sched_scan_stopped(mvm->hw);
493
494 return 0;
495}
496
497static void iwl_scan_offload_build_tx_cmd(struct iwl_mvm *mvm,
498 struct ieee80211_vif *vif,
499 struct ieee80211_sched_scan_ies *ies,
500 enum ieee80211_band band,
501 struct iwl_tx_cmd *cmd,
502 u8 *data)
503{
504 u16 cmd_len;
505
506 cmd->tx_flags = cpu_to_le32(TX_CMD_FLG_SEQ_CTL);
507 cmd->life_time = cpu_to_le32(TX_CMD_LIFE_TIME_INFINITE);
508 cmd->sta_id = mvm->aux_sta.sta_id;
509
510 cmd->rate_n_flags = iwl_mvm_scan_rate_n_flags(mvm, band, false);
511
512 cmd_len = iwl_mvm_fill_probe_req((struct ieee80211_mgmt *)data,
513 vif->addr,
514 1, NULL, 0,
515 ies->ie[band], ies->len[band],
516 SCAN_OFFLOAD_PROBE_REQ_SIZE);
517 cmd->len = cpu_to_le16(cmd_len);
518}
519
520static void iwl_build_scan_cmd(struct iwl_mvm *mvm,
521 struct ieee80211_vif *vif,
522 struct cfg80211_sched_scan_request *req,
523 struct iwl_scan_offload_cmd *scan)
524{
525 scan->channel_count =
526 mvm->nvm_data->bands[IEEE80211_BAND_2GHZ].n_channels +
527 mvm->nvm_data->bands[IEEE80211_BAND_5GHZ].n_channels;
528 scan->quiet_time = cpu_to_le16(IWL_ACTIVE_QUIET_TIME);
529 scan->quiet_plcp_th = cpu_to_le16(IWL_PLCP_QUIET_THRESH);
530 scan->good_CRC_th = IWL_GOOD_CRC_TH_DEFAULT;
531 scan->rx_chain = iwl_mvm_scan_rx_chain(mvm);
532 scan->max_out_time = cpu_to_le32(200 * 1024);
533 scan->suspend_time = iwl_mvm_scan_suspend_time(vif);
534 scan->filter_flags |= cpu_to_le32(MAC_FILTER_ACCEPT_GRP |
535 MAC_FILTER_IN_BEACON);
536 scan->scan_type = cpu_to_le32(SCAN_TYPE_BACKGROUND);
537 scan->rep_count = cpu_to_le32(1);
538}
539
540static int iwl_ssid_exist(u8 *ssid, u8 ssid_len, struct iwl_ssid_ie *ssid_list)
541{
542 int i;
543
544 for (i = 0; i < PROBE_OPTION_MAX; i++) {
545 if (!ssid_list[i].len)
546 break;
547 if (ssid_list[i].len == ssid_len &&
548 !memcmp(ssid_list->ssid, ssid, ssid_len))
549 return i;
550 }
551 return -1;
552}
553
554static void iwl_scan_offload_build_ssid(struct cfg80211_sched_scan_request *req,
555 struct iwl_scan_offload_cmd *scan,
556 u32 *ssid_bitmap)
557{
558 int i, j;
559 int index;
560
561 /*
562 * copy SSIDs from match list.
563 * iwl_config_sched_scan_profiles() uses the order of these ssids to
564 * config match list.
565 */
566 for (i = 0; i < req->n_match_sets && i < PROBE_OPTION_MAX; i++) {
567 scan->direct_scan[i].id = WLAN_EID_SSID;
568 scan->direct_scan[i].len = req->match_sets[i].ssid.ssid_len;
569 memcpy(scan->direct_scan[i].ssid, req->match_sets[i].ssid.ssid,
570 scan->direct_scan[i].len);
571 }
572
573 /* add SSIDs from scan SSID list */
574 *ssid_bitmap = 0;
575 for (j = 0; j < req->n_ssids && i < PROBE_OPTION_MAX; j++) {
576 index = iwl_ssid_exist(req->ssids[j].ssid,
577 req->ssids[j].ssid_len,
578 scan->direct_scan);
579 if (index < 0) {
580 if (!req->ssids[j].ssid_len)
581 continue;
582 scan->direct_scan[i].id = WLAN_EID_SSID;
583 scan->direct_scan[i].len = req->ssids[j].ssid_len;
584 memcpy(scan->direct_scan[i].ssid, req->ssids[j].ssid,
585 scan->direct_scan[i].len);
586 *ssid_bitmap |= BIT(i + 1);
587 i++;
588 } else {
589 *ssid_bitmap |= BIT(index + 1);
590 }
591 }
592}
593
594static void iwl_build_channel_cfg(struct iwl_mvm *mvm,
595 struct cfg80211_sched_scan_request *req,
596 struct iwl_scan_channel_cfg *channels,
597 enum ieee80211_band band,
598 int *head, int *tail,
599 u32 ssid_bitmap)
600{
601 struct ieee80211_supported_band *s_band;
602 int n_probes = req->n_ssids;
603 int n_channels = req->n_channels;
604 u8 active_dwell, passive_dwell;
605 int i, j, index = 0;
606 bool partial;
607
608 /*
609 * We have to configure all supported channels, even if we don't want to
610 * scan on them, but we have to send channels in the order that we want
611 * to scan. So add requested channels to head of the list and others to
612 * the end.
613 */
614 active_dwell = iwl_mvm_get_active_dwell(band, n_probes);
615 passive_dwell = iwl_mvm_get_passive_dwell(band);
616 s_band = &mvm->nvm_data->bands[band];
617
618 for (i = 0; i < s_band->n_channels && *head <= *tail; i++) {
619 partial = false;
620 for (j = 0; j < n_channels; j++)
621 if (s_band->channels[i].center_freq ==
622 req->channels[j]->center_freq) {
623 index = *head;
624 (*head)++;
625 /*
626 * Channels that came with the request will be
627 * in partial scan .
628 */
629 partial = true;
630 break;
631 }
632 if (!partial) {
633 index = *tail;
634 (*tail)--;
635 }
636 channels->channel_number[index] =
637 cpu_to_le16(ieee80211_frequency_to_channel(
638 s_band->channels[i].center_freq));
639 channels->dwell_time[index][0] = active_dwell;
640 channels->dwell_time[index][1] = passive_dwell;
641
642 channels->iter_count[index] = cpu_to_le16(1);
643 channels->iter_interval[index] = 0;
644
645 if (!(s_band->channels[i].flags & IEEE80211_CHAN_PASSIVE_SCAN))
646 channels->type[index] |=
647 cpu_to_le32(IWL_SCAN_OFFLOAD_CHANNEL_ACTIVE);
648
649 channels->type[index] |=
650 cpu_to_le32(IWL_SCAN_OFFLOAD_CHANNEL_FULL);
651 if (partial)
652 channels->type[index] |=
653 cpu_to_le32(IWL_SCAN_OFFLOAD_CHANNEL_PARTIAL);
654
655 if (s_band->channels[i].flags & IEEE80211_CHAN_NO_HT40)
656 channels->type[index] |=
657 cpu_to_le32(IWL_SCAN_OFFLOAD_CHANNEL_NARROW);
658
659 /* scan for all SSIDs from req->ssids */
660 channels->type[index] |= cpu_to_le32(ssid_bitmap);
661 }
662}
663
664int iwl_mvm_config_sched_scan(struct iwl_mvm *mvm,
665 struct ieee80211_vif *vif,
666 struct cfg80211_sched_scan_request *req,
667 struct ieee80211_sched_scan_ies *ies)
668{
669 int supported_bands = 0;
670 int band_2ghz = mvm->nvm_data->bands[IEEE80211_BAND_2GHZ].n_channels;
671 int band_5ghz = mvm->nvm_data->bands[IEEE80211_BAND_5GHZ].n_channels;
672 int head = 0;
673 int tail = band_2ghz + band_5ghz;
674 u32 ssid_bitmap;
675 int cmd_len;
676 int ret;
677
678 struct iwl_scan_offload_cfg *scan_cfg;
679 struct iwl_host_cmd cmd = {
680 .id = SCAN_OFFLOAD_CONFIG_CMD,
681 .flags = CMD_SYNC,
682 };
683
684 lockdep_assert_held(&mvm->mutex);
685
686 if (band_2ghz)
687 supported_bands++;
688 if (band_5ghz)
689 supported_bands++;
690
691 cmd_len = sizeof(struct iwl_scan_offload_cfg) +
692 supported_bands * SCAN_OFFLOAD_PROBE_REQ_SIZE;
693
694 scan_cfg = kzalloc(cmd_len, GFP_KERNEL);
695 if (!scan_cfg)
696 return -ENOMEM;
697
698 iwl_build_scan_cmd(mvm, vif, req, &scan_cfg->scan_cmd);
699 scan_cfg->scan_cmd.len = cpu_to_le16(cmd_len);
700
701 iwl_scan_offload_build_ssid(req, &scan_cfg->scan_cmd, &ssid_bitmap);
702 /* build tx frames for supported bands */
703 if (band_2ghz) {
704 iwl_scan_offload_build_tx_cmd(mvm, vif, ies,
705 IEEE80211_BAND_2GHZ,
706 &scan_cfg->scan_cmd.tx_cmd[0],
707 scan_cfg->data);
708 iwl_build_channel_cfg(mvm, req, &scan_cfg->channel_cfg,
709 IEEE80211_BAND_2GHZ, &head, &tail,
710 ssid_bitmap);
711 }
712 if (band_5ghz) {
713 iwl_scan_offload_build_tx_cmd(mvm, vif, ies,
714 IEEE80211_BAND_5GHZ,
715 &scan_cfg->scan_cmd.tx_cmd[1],
716 scan_cfg->data +
717 SCAN_OFFLOAD_PROBE_REQ_SIZE);
718 iwl_build_channel_cfg(mvm, req, &scan_cfg->channel_cfg,
719 IEEE80211_BAND_5GHZ, &head, &tail,
720 ssid_bitmap);
721 }
722
723 cmd.data[0] = scan_cfg;
724 cmd.len[0] = cmd_len;
725 cmd.dataflags[0] = IWL_HCMD_DFL_NOCOPY;
726
727 IWL_DEBUG_SCAN(mvm, "Sending scheduled scan config\n");
728
729 ret = iwl_mvm_send_cmd(mvm, &cmd);
730 kfree(scan_cfg);
731 return ret;
732}
733
734int iwl_mvm_config_sched_scan_profiles(struct iwl_mvm *mvm,
735 struct cfg80211_sched_scan_request *req)
736{
737 struct iwl_scan_offload_profile *profile;
738 struct iwl_scan_offload_profile_cfg *profile_cfg;
739 struct iwl_scan_offload_blacklist *blacklist;
740 struct iwl_host_cmd cmd = {
741 .id = SCAN_OFFLOAD_UPDATE_PROFILES_CMD,
742 .flags = CMD_SYNC,
743 .len[1] = sizeof(*profile_cfg),
744 .dataflags[0] = IWL_HCMD_DFL_NOCOPY,
745 .dataflags[1] = IWL_HCMD_DFL_NOCOPY,
746 };
747 int blacklist_len;
748 int i;
749 int ret;
750
751 if (WARN_ON(req->n_match_sets > IWL_SCAN_MAX_PROFILES))
752 return -EIO;
753
754 if (mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_SHORT_BL)
755 blacklist_len = IWL_SCAN_SHORT_BLACKLIST_LEN;
756 else
757 blacklist_len = IWL_SCAN_MAX_BLACKLIST_LEN;
758
759 blacklist = kzalloc(sizeof(*blacklist) * blacklist_len, GFP_KERNEL);
760 if (!blacklist)
761 return -ENOMEM;
762
763 profile_cfg = kzalloc(sizeof(*profile_cfg), GFP_KERNEL);
764 if (!profile_cfg) {
765 ret = -ENOMEM;
766 goto free_blacklist;
767 }
768
769 cmd.data[0] = blacklist;
770 cmd.len[0] = sizeof(*blacklist) * blacklist_len;
771 cmd.data[1] = profile_cfg;
772
773 /* No blacklist configuration */
774
775 profile_cfg->num_profiles = req->n_match_sets;
776 profile_cfg->active_clients = SCAN_CLIENT_SCHED_SCAN;
777 profile_cfg->pass_match = SCAN_CLIENT_SCHED_SCAN;
778 profile_cfg->match_notify = SCAN_CLIENT_SCHED_SCAN;
779
780 for (i = 0; i < req->n_match_sets; i++) {
781 profile = &profile_cfg->profiles[i];
782 profile->ssid_index = i;
783 /* Support any cipher and auth algorithm */
784 profile->unicast_cipher = 0xff;
785 profile->auth_alg = 0xff;
786 profile->network_type = IWL_NETWORK_TYPE_ANY;
787 profile->band_selection = IWL_SCAN_OFFLOAD_SELECT_ANY;
788 profile->client_bitmap = SCAN_CLIENT_SCHED_SCAN;
789 }
790
791 IWL_DEBUG_SCAN(mvm, "Sending scheduled scan profile config\n");
792
793 ret = iwl_mvm_send_cmd(mvm, &cmd);
794 kfree(profile_cfg);
795free_blacklist:
796 kfree(blacklist);
797
798 return ret;
799}
800
801int iwl_mvm_sched_scan_start(struct iwl_mvm *mvm,
802 struct cfg80211_sched_scan_request *req)
803{
804 struct iwl_scan_offload_req scan_req = {
805 .watchdog = IWL_SCHED_SCAN_WATCHDOG,
806
807 .schedule_line[0].iterations = IWL_FAST_SCHED_SCAN_ITERATIONS,
808 .schedule_line[0].delay = req->interval / 1000,
809 .schedule_line[0].full_scan_mul = 1,
810
811 .schedule_line[1].iterations = 0xff,
812 .schedule_line[1].delay = req->interval / 1000,
813 .schedule_line[1].full_scan_mul = IWL_FULL_SCAN_MULTIPLIER,
814 };
815
816 if (req->n_match_sets && req->match_sets[0].ssid.ssid_len) {
817 IWL_DEBUG_SCAN(mvm,
818 "Sending scheduled scan with filtering, filter len %d\n",
819 req->n_match_sets);
820 scan_req.flags |=
821 cpu_to_le16(IWL_SCAN_OFFLOAD_FLAG_FILTER_SSID);
822 } else {
823 IWL_DEBUG_SCAN(mvm,
824 "Sending Scheduled scan without filtering\n");
825 }
826
827 return iwl_mvm_send_cmd_pdu(mvm, SCAN_OFFLOAD_REQUEST_CMD, CMD_SYNC,
828 sizeof(scan_req), &scan_req);
829}
830
831static int iwl_mvm_send_sched_scan_abort(struct iwl_mvm *mvm)
832{
833 int ret;
834 struct iwl_host_cmd cmd = {
835 .id = SCAN_OFFLOAD_ABORT_CMD,
836 .flags = CMD_SYNC,
837 };
838 u32 status;
839
840 /* Exit instantly with error when device is not ready
841 * to receive scan abort command or it does not perform
842 * scheduled scan currently */
843 if (mvm->scan_status != IWL_MVM_SCAN_SCHED)
844 return -EIO;
845
846 ret = iwl_mvm_send_cmd_status(mvm, &cmd, &status);
847 if (ret)
848 return ret;
849
850 if (status != CAN_ABORT_STATUS) {
851 /*
852 * The scan abort will return 1 for success or
853 * 2 for "failure". A failure condition can be
854 * due to simply not being in an active scan which
855 * can occur if we send the scan abort before the
856 * microcode has notified us that a scan is completed.
857 */
858 IWL_DEBUG_SCAN(mvm, "SCAN OFFLOAD ABORT ret %d.\n", status);
859 ret = -EIO;
860 }
861
862 return ret;
863}
864
865void iwl_mvm_sched_scan_stop(struct iwl_mvm *mvm)
866{
867 int ret;
868
869 lockdep_assert_held(&mvm->mutex);
870
871 if (mvm->scan_status != IWL_MVM_SCAN_SCHED) {
872 IWL_DEBUG_SCAN(mvm, "No offloaded scan to stop\n");
873 return;
874 }
875
876 ret = iwl_mvm_send_sched_scan_abort(mvm);
877 if (ret)
878 IWL_DEBUG_SCAN(mvm, "Send stop offload scan failed %d\n", ret);
879 else
880 IWL_DEBUG_SCAN(mvm, "Successfully sent stop offload scan\n");
881}
diff --git a/drivers/net/wireless/iwlwifi/mvm/sta.c b/drivers/net/wireless/iwlwifi/mvm/sta.c
index 44add291531b..329952363a54 100644
--- a/drivers/net/wireless/iwlwifi/mvm/sta.c
+++ b/drivers/net/wireless/iwlwifi/mvm/sta.c
@@ -66,6 +66,115 @@
66#include "sta.h" 66#include "sta.h"
67#include "rs.h" 67#include "rs.h"
68 68
69static void iwl_mvm_add_sta_cmd_v6_to_v5(struct iwl_mvm_add_sta_cmd_v6 *cmd_v6,
70 struct iwl_mvm_add_sta_cmd_v5 *cmd_v5)
71{
72 memset(cmd_v5, 0, sizeof(*cmd_v5));
73
74 cmd_v5->add_modify = cmd_v6->add_modify;
75 cmd_v5->tid_disable_tx = cmd_v6->tid_disable_tx;
76 cmd_v5->mac_id_n_color = cmd_v6->mac_id_n_color;
77 memcpy(cmd_v5->addr, cmd_v6->addr, ETH_ALEN);
78 cmd_v5->sta_id = cmd_v6->sta_id;
79 cmd_v5->modify_mask = cmd_v6->modify_mask;
80 cmd_v5->station_flags = cmd_v6->station_flags;
81 cmd_v5->station_flags_msk = cmd_v6->station_flags_msk;
82 cmd_v5->add_immediate_ba_tid = cmd_v6->add_immediate_ba_tid;
83 cmd_v5->remove_immediate_ba_tid = cmd_v6->remove_immediate_ba_tid;
84 cmd_v5->add_immediate_ba_ssn = cmd_v6->add_immediate_ba_ssn;
85 cmd_v5->sleep_tx_count = cmd_v6->sleep_tx_count;
86 cmd_v5->sleep_state_flags = cmd_v6->sleep_state_flags;
87 cmd_v5->assoc_id = cmd_v6->assoc_id;
88 cmd_v5->beamform_flags = cmd_v6->beamform_flags;
89 cmd_v5->tfd_queue_msk = cmd_v6->tfd_queue_msk;
90}
91
92static void
93iwl_mvm_add_sta_key_to_add_sta_cmd_v5(struct iwl_mvm_add_sta_key_cmd *key_cmd,
94 struct iwl_mvm_add_sta_cmd_v5 *sta_cmd,
95 u32 mac_id_n_color)
96{
97 memset(sta_cmd, 0, sizeof(*sta_cmd));
98
99 sta_cmd->sta_id = key_cmd->sta_id;
100 sta_cmd->add_modify = STA_MODE_MODIFY;
101 sta_cmd->modify_mask = STA_MODIFY_KEY;
102 sta_cmd->mac_id_n_color = cpu_to_le32(mac_id_n_color);
103
104 sta_cmd->key.key_offset = key_cmd->key_offset;
105 sta_cmd->key.key_flags = key_cmd->key_flags;
106 memcpy(sta_cmd->key.key, key_cmd->key, sizeof(sta_cmd->key.key));
107 sta_cmd->key.tkip_rx_tsc_byte2 = key_cmd->tkip_rx_tsc_byte2;
108 memcpy(sta_cmd->key.tkip_rx_ttak, key_cmd->tkip_rx_ttak,
109 sizeof(sta_cmd->key.tkip_rx_ttak));
110}
111
112static int iwl_mvm_send_add_sta_cmd_status(struct iwl_mvm *mvm,
113 struct iwl_mvm_add_sta_cmd_v6 *cmd,
114 int *status)
115{
116 struct iwl_mvm_add_sta_cmd_v5 cmd_v5;
117
118 if (mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_STA_KEY_CMD)
119 return iwl_mvm_send_cmd_pdu_status(mvm, ADD_STA, sizeof(*cmd),
120 cmd, status);
121
122 iwl_mvm_add_sta_cmd_v6_to_v5(cmd, &cmd_v5);
123
124 return iwl_mvm_send_cmd_pdu_status(mvm, ADD_STA, sizeof(cmd_v5),
125 &cmd_v5, status);
126}
127
128static int iwl_mvm_send_add_sta_cmd(struct iwl_mvm *mvm, u32 flags,
129 struct iwl_mvm_add_sta_cmd_v6 *cmd)
130{
131 struct iwl_mvm_add_sta_cmd_v5 cmd_v5;
132
133 if (mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_STA_KEY_CMD)
134 return iwl_mvm_send_cmd_pdu(mvm, ADD_STA, flags,
135 sizeof(*cmd), cmd);
136
137 iwl_mvm_add_sta_cmd_v6_to_v5(cmd, &cmd_v5);
138
139 return iwl_mvm_send_cmd_pdu(mvm, ADD_STA, flags, sizeof(cmd_v5),
140 &cmd_v5);
141}
142
143static int
144iwl_mvm_send_add_sta_key_cmd_status(struct iwl_mvm *mvm,
145 struct iwl_mvm_add_sta_key_cmd *cmd,
146 u32 mac_id_n_color,
147 int *status)
148{
149 struct iwl_mvm_add_sta_cmd_v5 sta_cmd;
150
151 if (mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_STA_KEY_CMD)
152 return iwl_mvm_send_cmd_pdu_status(mvm, ADD_STA_KEY,
153 sizeof(*cmd), cmd, status);
154
155 iwl_mvm_add_sta_key_to_add_sta_cmd_v5(cmd, &sta_cmd, mac_id_n_color);
156
157 return iwl_mvm_send_cmd_pdu_status(mvm, ADD_STA, sizeof(sta_cmd),
158 &sta_cmd, status);
159}
160
161static int iwl_mvm_send_add_sta_key_cmd(struct iwl_mvm *mvm,
162 u32 flags,
163 struct iwl_mvm_add_sta_key_cmd *cmd,
164 u32 mac_id_n_color)
165{
166 struct iwl_mvm_add_sta_cmd_v5 sta_cmd;
167
168 if (mvm->fw->ucode_capa.flags & IWL_UCODE_TLV_FLAGS_STA_KEY_CMD)
169 return iwl_mvm_send_cmd_pdu(mvm, ADD_STA_KEY, flags,
170 sizeof(*cmd), cmd);
171
172 iwl_mvm_add_sta_key_to_add_sta_cmd_v5(cmd, &sta_cmd, mac_id_n_color);
173
174 return iwl_mvm_send_cmd_pdu(mvm, ADD_STA, flags, sizeof(sta_cmd),
175 &sta_cmd);
176}
177
69static int iwl_mvm_find_free_sta_id(struct iwl_mvm *mvm) 178static int iwl_mvm_find_free_sta_id(struct iwl_mvm *mvm)
70{ 179{
71 int sta_id; 180 int sta_id;
@@ -87,7 +196,7 @@ int iwl_mvm_sta_send_to_fw(struct iwl_mvm *mvm, struct ieee80211_sta *sta,
87 bool update) 196 bool update)
88{ 197{
89 struct iwl_mvm_sta *mvm_sta = (void *)sta->drv_priv; 198 struct iwl_mvm_sta *mvm_sta = (void *)sta->drv_priv;
90 struct iwl_mvm_add_sta_cmd add_sta_cmd; 199 struct iwl_mvm_add_sta_cmd_v6 add_sta_cmd;
91 int ret; 200 int ret;
92 u32 status; 201 u32 status;
93 u32 agg_size = 0, mpdu_dens = 0; 202 u32 agg_size = 0, mpdu_dens = 0;
@@ -175,8 +284,7 @@ int iwl_mvm_sta_send_to_fw(struct iwl_mvm *mvm, struct ieee80211_sta *sta,
175 cpu_to_le32(mpdu_dens << STA_FLG_AGG_MPDU_DENS_SHIFT); 284 cpu_to_le32(mpdu_dens << STA_FLG_AGG_MPDU_DENS_SHIFT);
176 285
177 status = ADD_STA_SUCCESS; 286 status = ADD_STA_SUCCESS;
178 ret = iwl_mvm_send_cmd_pdu_status(mvm, ADD_STA, sizeof(add_sta_cmd), 287 ret = iwl_mvm_send_add_sta_cmd_status(mvm, &add_sta_cmd, &status);
179 &add_sta_cmd, &status);
180 if (ret) 288 if (ret)
181 return ret; 289 return ret;
182 290
@@ -229,8 +337,12 @@ int iwl_mvm_add_sta(struct iwl_mvm *mvm,
229 if (vif->hw_queue[i] != IEEE80211_INVAL_HW_QUEUE) 337 if (vif->hw_queue[i] != IEEE80211_INVAL_HW_QUEUE)
230 mvm_sta->tfd_queue_msk |= BIT(vif->hw_queue[i]); 338 mvm_sta->tfd_queue_msk |= BIT(vif->hw_queue[i]);
231 339
232 /* for HW restart - need to reset the seq_number etc... */ 340 /* for HW restart - reset everything but the sequence number */
233 memset(mvm_sta->tid_data, 0, sizeof(mvm_sta->tid_data)); 341 for (i = 0; i < IWL_MAX_TID_COUNT; i++) {
342 u16 seq = mvm_sta->tid_data[i].seq_number;
343 memset(&mvm_sta->tid_data[i], 0, sizeof(mvm_sta->tid_data[i]));
344 mvm_sta->tid_data[i].seq_number = seq;
345 }
234 346
235 ret = iwl_mvm_sta_send_to_fw(mvm, sta, false); 347 ret = iwl_mvm_sta_send_to_fw(mvm, sta, false);
236 if (ret) 348 if (ret)
@@ -256,7 +368,7 @@ int iwl_mvm_update_sta(struct iwl_mvm *mvm,
256int iwl_mvm_drain_sta(struct iwl_mvm *mvm, struct iwl_mvm_sta *mvmsta, 368int iwl_mvm_drain_sta(struct iwl_mvm *mvm, struct iwl_mvm_sta *mvmsta,
257 bool drain) 369 bool drain)
258{ 370{
259 struct iwl_mvm_add_sta_cmd cmd = {}; 371 struct iwl_mvm_add_sta_cmd_v6 cmd = {};
260 int ret; 372 int ret;
261 u32 status; 373 u32 status;
262 374
@@ -269,8 +381,7 @@ int iwl_mvm_drain_sta(struct iwl_mvm *mvm, struct iwl_mvm_sta *mvmsta,
269 cmd.station_flags_msk = cpu_to_le32(STA_FLG_DRAIN_FLOW); 381 cmd.station_flags_msk = cpu_to_le32(STA_FLG_DRAIN_FLOW);
270 382
271 status = ADD_STA_SUCCESS; 383 status = ADD_STA_SUCCESS;
272 ret = iwl_mvm_send_cmd_pdu_status(mvm, ADD_STA, sizeof(cmd), 384 ret = iwl_mvm_send_add_sta_cmd_status(mvm, &cmd, &status);
273 &cmd, &status);
274 if (ret) 385 if (ret)
275 return ret; 386 return ret;
276 387
@@ -469,13 +580,13 @@ static int iwl_mvm_add_int_sta_common(struct iwl_mvm *mvm,
469 const u8 *addr, 580 const u8 *addr,
470 u16 mac_id, u16 color) 581 u16 mac_id, u16 color)
471{ 582{
472 struct iwl_mvm_add_sta_cmd cmd; 583 struct iwl_mvm_add_sta_cmd_v6 cmd;
473 int ret; 584 int ret;
474 u32 status; 585 u32 status;
475 586
476 lockdep_assert_held(&mvm->mutex); 587 lockdep_assert_held(&mvm->mutex);
477 588
478 memset(&cmd, 0, sizeof(struct iwl_mvm_add_sta_cmd)); 589 memset(&cmd, 0, sizeof(struct iwl_mvm_add_sta_cmd_v6));
479 cmd.sta_id = sta->sta_id; 590 cmd.sta_id = sta->sta_id;
480 cmd.mac_id_n_color = cpu_to_le32(FW_CMD_ID_AND_COLOR(mac_id, 591 cmd.mac_id_n_color = cpu_to_le32(FW_CMD_ID_AND_COLOR(mac_id,
481 color)); 592 color));
@@ -485,8 +596,7 @@ static int iwl_mvm_add_int_sta_common(struct iwl_mvm *mvm,
485 if (addr) 596 if (addr)
486 memcpy(cmd.addr, addr, ETH_ALEN); 597 memcpy(cmd.addr, addr, ETH_ALEN);
487 598
488 ret = iwl_mvm_send_cmd_pdu_status(mvm, ADD_STA, sizeof(cmd), 599 ret = iwl_mvm_send_add_sta_cmd_status(mvm, &cmd, &status);
489 &cmd, &status);
490 if (ret) 600 if (ret)
491 return ret; 601 return ret;
492 602
@@ -534,10 +644,14 @@ int iwl_mvm_send_bcast_sta(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
534 struct iwl_mvm_int_sta *bsta) 644 struct iwl_mvm_int_sta *bsta)
535{ 645{
536 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 646 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif);
537 static const u8 baddr[] = {0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF}; 647 static const u8 _baddr[] = {0xFF, 0xFF, 0xFF, 0xFF, 0xFF, 0xFF};
648 static const u8 *baddr = _baddr;
538 649
539 lockdep_assert_held(&mvm->mutex); 650 lockdep_assert_held(&mvm->mutex);
540 651
652 if (vif->type == NL80211_IFTYPE_ADHOC)
653 baddr = vif->bss_conf.bssid;
654
541 if (WARN_ON_ONCE(bsta->sta_id == IWL_MVM_STATION_COUNT)) 655 if (WARN_ON_ONCE(bsta->sta_id == IWL_MVM_STATION_COUNT))
542 return -ENOSPC; 656 return -ENOSPC;
543 657
@@ -614,7 +728,7 @@ int iwl_mvm_sta_rx_agg(struct iwl_mvm *mvm, struct ieee80211_sta *sta,
614 int tid, u16 ssn, bool start) 728 int tid, u16 ssn, bool start)
615{ 729{
616 struct iwl_mvm_sta *mvm_sta = (void *)sta->drv_priv; 730 struct iwl_mvm_sta *mvm_sta = (void *)sta->drv_priv;
617 struct iwl_mvm_add_sta_cmd cmd = {}; 731 struct iwl_mvm_add_sta_cmd_v6 cmd = {};
618 int ret; 732 int ret;
619 u32 status; 733 u32 status;
620 734
@@ -638,8 +752,7 @@ int iwl_mvm_sta_rx_agg(struct iwl_mvm *mvm, struct ieee80211_sta *sta,
638 STA_MODIFY_REMOVE_BA_TID; 752 STA_MODIFY_REMOVE_BA_TID;
639 753
640 status = ADD_STA_SUCCESS; 754 status = ADD_STA_SUCCESS;
641 ret = iwl_mvm_send_cmd_pdu_status(mvm, ADD_STA, sizeof(cmd), 755 ret = iwl_mvm_send_add_sta_cmd_status(mvm, &cmd, &status);
642 &cmd, &status);
643 if (ret) 756 if (ret)
644 return ret; 757 return ret;
645 758
@@ -674,7 +787,7 @@ static int iwl_mvm_sta_tx_agg(struct iwl_mvm *mvm, struct ieee80211_sta *sta,
674 int tid, u8 queue, bool start) 787 int tid, u8 queue, bool start)
675{ 788{
676 struct iwl_mvm_sta *mvm_sta = (void *)sta->drv_priv; 789 struct iwl_mvm_sta *mvm_sta = (void *)sta->drv_priv;
677 struct iwl_mvm_add_sta_cmd cmd = {}; 790 struct iwl_mvm_add_sta_cmd_v6 cmd = {};
678 int ret; 791 int ret;
679 u32 status; 792 u32 status;
680 793
@@ -696,8 +809,7 @@ static int iwl_mvm_sta_tx_agg(struct iwl_mvm *mvm, struct ieee80211_sta *sta,
696 cmd.tid_disable_tx = cpu_to_le16(mvm_sta->tid_disable_agg); 809 cmd.tid_disable_tx = cpu_to_le16(mvm_sta->tid_disable_agg);
697 810
698 status = ADD_STA_SUCCESS; 811 status = ADD_STA_SUCCESS;
699 ret = iwl_mvm_send_cmd_pdu_status(mvm, ADD_STA, sizeof(cmd), 812 ret = iwl_mvm_send_add_sta_cmd_status(mvm, &cmd, &status);
700 &cmd, &status);
701 if (ret) 813 if (ret)
702 return ret; 814 return ret;
703 815
@@ -743,13 +855,13 @@ int iwl_mvm_sta_tx_agg_start(struct iwl_mvm *mvm, struct ieee80211_vif *vif,
743 855
744 lockdep_assert_held(&mvm->mutex); 856 lockdep_assert_held(&mvm->mutex);
745 857
746 for (txq_id = IWL_MVM_FIRST_AGG_QUEUE; 858 for (txq_id = mvm->first_agg_queue;
747 txq_id <= IWL_MVM_LAST_AGG_QUEUE; txq_id++) 859 txq_id <= mvm->last_agg_queue; txq_id++)
748 if (mvm->queue_to_mac80211[txq_id] == 860 if (mvm->queue_to_mac80211[txq_id] ==
749 IWL_INVALID_MAC80211_QUEUE) 861 IWL_INVALID_MAC80211_QUEUE)
750 break; 862 break;
751 863
752 if (txq_id > IWL_MVM_LAST_AGG_QUEUE) { 864 if (txq_id > mvm->last_agg_queue) {
753 IWL_ERR(mvm, "Failed to allocate agg queue\n"); 865 IWL_ERR(mvm, "Failed to allocate agg queue\n");
754 return -EIO; 866 return -EIO;
755 } 867 }
@@ -987,10 +1099,11 @@ static int iwl_mvm_send_sta_key(struct iwl_mvm *mvm,
987 u32 cmd_flags) 1099 u32 cmd_flags)
988{ 1100{
989 __le16 key_flags; 1101 __le16 key_flags;
990 struct iwl_mvm_add_sta_cmd cmd = {}; 1102 struct iwl_mvm_add_sta_key_cmd cmd = {};
991 int ret, status; 1103 int ret, status;
992 u16 keyidx; 1104 u16 keyidx;
993 int i; 1105 int i;
1106 u32 mac_id_n_color = mvm_sta->mac_id_n_color;
994 1107
995 keyidx = (keyconf->keyidx << STA_KEY_FLG_KEYID_POS) & 1108 keyidx = (keyconf->keyidx << STA_KEY_FLG_KEYID_POS) &
996 STA_KEY_FLG_KEYID_MSK; 1109 STA_KEY_FLG_KEYID_MSK;
@@ -1000,14 +1113,14 @@ static int iwl_mvm_send_sta_key(struct iwl_mvm *mvm,
1000 switch (keyconf->cipher) { 1113 switch (keyconf->cipher) {
1001 case WLAN_CIPHER_SUITE_TKIP: 1114 case WLAN_CIPHER_SUITE_TKIP:
1002 key_flags |= cpu_to_le16(STA_KEY_FLG_TKIP); 1115 key_flags |= cpu_to_le16(STA_KEY_FLG_TKIP);
1003 cmd.key.tkip_rx_tsc_byte2 = tkip_iv32; 1116 cmd.tkip_rx_tsc_byte2 = tkip_iv32;
1004 for (i = 0; i < 5; i++) 1117 for (i = 0; i < 5; i++)
1005 cmd.key.tkip_rx_ttak[i] = cpu_to_le16(tkip_p1k[i]); 1118 cmd.tkip_rx_ttak[i] = cpu_to_le16(tkip_p1k[i]);
1006 memcpy(cmd.key.key, keyconf->key, keyconf->keylen); 1119 memcpy(cmd.key, keyconf->key, keyconf->keylen);
1007 break; 1120 break;
1008 case WLAN_CIPHER_SUITE_CCMP: 1121 case WLAN_CIPHER_SUITE_CCMP:
1009 key_flags |= cpu_to_le16(STA_KEY_FLG_CCM); 1122 key_flags |= cpu_to_le16(STA_KEY_FLG_CCM);
1010 memcpy(cmd.key.key, keyconf->key, keyconf->keylen); 1123 memcpy(cmd.key, keyconf->key, keyconf->keylen);
1011 break; 1124 break;
1012 default: 1125 default:
1013 WARN_ON(1); 1126 WARN_ON(1);
@@ -1017,20 +1130,18 @@ static int iwl_mvm_send_sta_key(struct iwl_mvm *mvm,
1017 if (!(keyconf->flags & IEEE80211_KEY_FLAG_PAIRWISE)) 1130 if (!(keyconf->flags & IEEE80211_KEY_FLAG_PAIRWISE))
1018 key_flags |= cpu_to_le16(STA_KEY_MULTICAST); 1131 key_flags |= cpu_to_le16(STA_KEY_MULTICAST);
1019 1132
1020 cmd.mac_id_n_color = cpu_to_le32(mvm_sta->mac_id_n_color); 1133 cmd.key_offset = keyconf->hw_key_idx;
1021 cmd.key.key_offset = keyconf->hw_key_idx; 1134 cmd.key_flags = key_flags;
1022 cmd.key.key_flags = key_flags;
1023 cmd.add_modify = STA_MODE_MODIFY;
1024 cmd.modify_mask = STA_MODIFY_KEY;
1025 cmd.sta_id = sta_id; 1135 cmd.sta_id = sta_id;
1026 1136
1027 status = ADD_STA_SUCCESS; 1137 status = ADD_STA_SUCCESS;
1028 if (cmd_flags == CMD_SYNC) 1138 if (cmd_flags == CMD_SYNC)
1029 ret = iwl_mvm_send_cmd_pdu_status(mvm, ADD_STA, sizeof(cmd), 1139 ret = iwl_mvm_send_add_sta_key_cmd_status(mvm, &cmd,
1030 &cmd, &status); 1140 mac_id_n_color,
1141 &status);
1031 else 1142 else
1032 ret = iwl_mvm_send_cmd_pdu(mvm, ADD_STA, CMD_ASYNC, 1143 ret = iwl_mvm_send_add_sta_key_cmd(mvm, CMD_ASYNC, &cmd,
1033 sizeof(cmd), &cmd); 1144 mac_id_n_color);
1034 1145
1035 switch (status) { 1146 switch (status) {
1036 case ADD_STA_SUCCESS: 1147 case ADD_STA_SUCCESS:
@@ -1197,7 +1308,7 @@ int iwl_mvm_remove_sta_key(struct iwl_mvm *mvm,
1197 struct ieee80211_key_conf *keyconf) 1308 struct ieee80211_key_conf *keyconf)
1198{ 1309{
1199 struct iwl_mvm_sta *mvm_sta; 1310 struct iwl_mvm_sta *mvm_sta;
1200 struct iwl_mvm_add_sta_cmd cmd = {}; 1311 struct iwl_mvm_add_sta_key_cmd cmd = {};
1201 __le16 key_flags; 1312 __le16 key_flags;
1202 int ret, status; 1313 int ret, status;
1203 u8 sta_id; 1314 u8 sta_id;
@@ -1252,17 +1363,14 @@ int iwl_mvm_remove_sta_key(struct iwl_mvm *mvm,
1252 if (!(keyconf->flags & IEEE80211_KEY_FLAG_PAIRWISE)) 1363 if (!(keyconf->flags & IEEE80211_KEY_FLAG_PAIRWISE))
1253 key_flags |= cpu_to_le16(STA_KEY_MULTICAST); 1364 key_flags |= cpu_to_le16(STA_KEY_MULTICAST);
1254 1365
1255 cmd.mac_id_n_color = cpu_to_le32(mvm_sta->mac_id_n_color); 1366 cmd.key_flags = key_flags;
1256 cmd.key.key_flags = key_flags; 1367 cmd.key_offset = keyconf->hw_key_idx;
1257 cmd.key.key_offset = keyconf->hw_key_idx;
1258 cmd.sta_id = sta_id; 1368 cmd.sta_id = sta_id;
1259 1369
1260 cmd.modify_mask = STA_MODIFY_KEY;
1261 cmd.add_modify = STA_MODE_MODIFY;
1262
1263 status = ADD_STA_SUCCESS; 1370 status = ADD_STA_SUCCESS;
1264 ret = iwl_mvm_send_cmd_pdu_status(mvm, ADD_STA, sizeof(cmd), 1371 ret = iwl_mvm_send_add_sta_key_cmd_status(mvm, &cmd,
1265 &cmd, &status); 1372 mvm_sta->mac_id_n_color,
1373 &status);
1266 1374
1267 switch (status) { 1375 switch (status) {
1268 case ADD_STA_SUCCESS: 1376 case ADD_STA_SUCCESS:
@@ -1309,7 +1417,7 @@ void iwl_mvm_sta_modify_ps_wake(struct iwl_mvm *mvm,
1309 struct ieee80211_sta *sta) 1417 struct ieee80211_sta *sta)
1310{ 1418{
1311 struct iwl_mvm_sta *mvmsta = (void *)sta->drv_priv; 1419 struct iwl_mvm_sta *mvmsta = (void *)sta->drv_priv;
1312 struct iwl_mvm_add_sta_cmd cmd = { 1420 struct iwl_mvm_add_sta_cmd_v6 cmd = {
1313 .add_modify = STA_MODE_MODIFY, 1421 .add_modify = STA_MODE_MODIFY,
1314 .sta_id = mvmsta->sta_id, 1422 .sta_id = mvmsta->sta_id,
1315 .station_flags_msk = cpu_to_le32(STA_FLG_PS), 1423 .station_flags_msk = cpu_to_le32(STA_FLG_PS),
@@ -1317,7 +1425,7 @@ void iwl_mvm_sta_modify_ps_wake(struct iwl_mvm *mvm,
1317 }; 1425 };
1318 int ret; 1426 int ret;
1319 1427
1320 ret = iwl_mvm_send_cmd_pdu(mvm, ADD_STA, CMD_ASYNC, sizeof(cmd), &cmd); 1428 ret = iwl_mvm_send_add_sta_cmd(mvm, CMD_ASYNC, &cmd);
1321 if (ret) 1429 if (ret)
1322 IWL_ERR(mvm, "Failed to send ADD_STA command (%d)\n", ret); 1430 IWL_ERR(mvm, "Failed to send ADD_STA command (%d)\n", ret);
1323} 1431}
@@ -1331,7 +1439,7 @@ void iwl_mvm_sta_modify_sleep_tx_count(struct iwl_mvm *mvm,
1331 (reason == IEEE80211_FRAME_RELEASE_UAPSD) ? 1439 (reason == IEEE80211_FRAME_RELEASE_UAPSD) ?
1332 STA_SLEEP_STATE_UAPSD : STA_SLEEP_STATE_PS_POLL; 1440 STA_SLEEP_STATE_UAPSD : STA_SLEEP_STATE_PS_POLL;
1333 struct iwl_mvm_sta *mvmsta = (void *)sta->drv_priv; 1441 struct iwl_mvm_sta *mvmsta = (void *)sta->drv_priv;
1334 struct iwl_mvm_add_sta_cmd cmd = { 1442 struct iwl_mvm_add_sta_cmd_v6 cmd = {
1335 .add_modify = STA_MODE_MODIFY, 1443 .add_modify = STA_MODE_MODIFY,
1336 .sta_id = mvmsta->sta_id, 1444 .sta_id = mvmsta->sta_id,
1337 .modify_mask = STA_MODIFY_SLEEPING_STA_TX_COUNT, 1445 .modify_mask = STA_MODIFY_SLEEPING_STA_TX_COUNT,
@@ -1346,7 +1454,7 @@ void iwl_mvm_sta_modify_sleep_tx_count(struct iwl_mvm *mvm,
1346 int ret; 1454 int ret;
1347 1455
1348 /* TODO: somehow the fw doesn't seem to take PS_POLL into account */ 1456 /* TODO: somehow the fw doesn't seem to take PS_POLL into account */
1349 ret = iwl_mvm_send_cmd_pdu(mvm, ADD_STA, CMD_ASYNC, sizeof(cmd), &cmd); 1457 ret = iwl_mvm_send_add_sta_cmd(mvm, CMD_ASYNC, &cmd);
1350 if (ret) 1458 if (ret)
1351 IWL_ERR(mvm, "Failed to send ADD_STA command (%d)\n", ret); 1459 IWL_ERR(mvm, "Failed to send ADD_STA command (%d)\n", ret);
1352} 1460}
diff --git a/drivers/net/wireless/iwlwifi/mvm/sta.h b/drivers/net/wireless/iwlwifi/mvm/sta.h
index 94b265eb32b8..4dfc359a4bdd 100644
--- a/drivers/net/wireless/iwlwifi/mvm/sta.h
+++ b/drivers/net/wireless/iwlwifi/mvm/sta.h
@@ -293,10 +293,6 @@ struct iwl_mvm_sta {
293 struct iwl_lq_sta lq_sta; 293 struct iwl_lq_sta lq_sta;
294 struct ieee80211_vif *vif; 294 struct ieee80211_vif *vif;
295 295
296#ifdef CONFIG_PM_SLEEP
297 u16 last_seq_ctl;
298#endif
299
300 /* Temporary, until the new TLC will control the Tx protection */ 296 /* Temporary, until the new TLC will control the Tx protection */
301 s8 tx_protection; 297 s8 tx_protection;
302 bool tt_tx_protection; 298 bool tt_tx_protection;
diff --git a/drivers/net/wireless/iwlwifi/mvm/testmode.h b/drivers/net/wireless/iwlwifi/mvm/testmode.h
new file mode 100644
index 000000000000..eb74391d91ca
--- /dev/null
+++ b/drivers/net/wireless/iwlwifi/mvm/testmode.h
@@ -0,0 +1,95 @@
1/******************************************************************************
2 *
3 * This file is provided under a dual BSD/GPLv2 license. When using or
4 * redistributing this file, you may do so under either license.
5 *
6 * GPL LICENSE SUMMARY
7 *
8 * Copyright(c) 2013 Intel Corporation. All rights reserved.
9 *
10 * This program is free software; you can redistribute it and/or modify
11 * it under the terms of version 2 of the GNU General Public License as
12 * published by the Free Software Foundation.
13 *
14 * This program is distributed in the hope that it will be useful, but
15 * WITHOUT ANY WARRANTY; without even the implied warranty of
16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
17 * General Public License for more details.
18 *
19 * You should have received a copy of the GNU General Public License
20 * along with this program; if not, write to the Free Software
21 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110,
22 * USA
23 *
24 * The full GNU General Public License is included in this distribution
25 * in the file called COPYING.
26 *
27 * Contact Information:
28 * Intel Linux Wireless <ilw@linux.intel.com>
29 * Intel Corporation, 5200 N.E. Elam Young Parkway, Hillsboro, OR 97124-6497
30 *
31 * BSD LICENSE
32 *
33 * Copyright(c) 2013 Intel Corporation. All rights reserved.
34 * All rights reserved.
35 *
36 * Redistribution and use in source and binary forms, with or without
37 * modification, are permitted provided that the following conditions
38 * are met:
39 *
40 * * Redistributions of source code must retain the above copyright
41 * notice, this list of conditions and the following disclaimer.
42 * * Redistributions in binary form must reproduce the above copyright
43 * notice, this list of conditions and the following disclaimer in
44 * the documentation and/or other materials provided with the
45 * distribution.
46 * * Neither the name Intel Corporation nor the names of its
47 * contributors may be used to endorse or promote products derived
48 * from this software without specific prior written permission.
49 *
50 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
51 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
52 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
53 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
54 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
55 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
56 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
57 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
58 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
59 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
60 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
61 *
62 *****************************************************************************/
63
64#ifndef __IWL_MVM_TESTMODE_H__
65#define __IWL_MVM_TESTMODE_H__
66
67/**
68 * enum iwl_mvm_testmode_attrs - testmode attributes inside NL80211_ATTR_TESTDATA
69 * @IWL_MVM_TM_ATTR_UNSPEC: (invalid attribute)
70 * @IWL_MVM_TM_ATTR_CMD: sub command, see &enum iwl_mvm_testmode_commands (u32)
71 * @IWL_MVM_TM_ATTR_NOA_DURATION: requested NoA duration (u32)
72 * @IWL_MVM_TM_ATTR_BEACON_FILTER_STATE: beacon filter state (0 or 1, u32)
73 */
74enum iwl_mvm_testmode_attrs {
75 IWL_MVM_TM_ATTR_UNSPEC,
76 IWL_MVM_TM_ATTR_CMD,
77 IWL_MVM_TM_ATTR_NOA_DURATION,
78 IWL_MVM_TM_ATTR_BEACON_FILTER_STATE,
79
80 /* keep last */
81 NUM_IWL_MVM_TM_ATTRS,
82 IWL_MVM_TM_ATTR_MAX = NUM_IWL_MVM_TM_ATTRS - 1,
83};
84
85/**
86 * enum iwl_mvm_testmode_commands - MVM testmode commands
87 * @IWL_MVM_TM_CMD_SET_NOA: set NoA on GO vif for testing
88 * @IWL_MVM_TM_CMD_SET_BEACON_FILTER: turn beacon filtering off/on
89 */
90enum iwl_mvm_testmode_commands {
91 IWL_MVM_TM_CMD_SET_NOA,
92 IWL_MVM_TM_CMD_SET_BEACON_FILTER,
93};
94
95#endif /* __IWL_MVM_TESTMODE_H__ */
diff --git a/drivers/net/wireless/iwlwifi/mvm/time-event.c b/drivers/net/wireless/iwlwifi/mvm/time-event.c
index 76a3c177e100..33cf56fdfc41 100644
--- a/drivers/net/wireless/iwlwifi/mvm/time-event.c
+++ b/drivers/net/wireless/iwlwifi/mvm/time-event.c
@@ -387,7 +387,8 @@ static int iwl_mvm_time_event_send_add(struct iwl_mvm *mvm,
387 387
388void iwl_mvm_protect_session(struct iwl_mvm *mvm, 388void iwl_mvm_protect_session(struct iwl_mvm *mvm,
389 struct ieee80211_vif *vif, 389 struct ieee80211_vif *vif,
390 u32 duration, u32 min_duration) 390 u32 duration, u32 min_duration,
391 u32 max_delay)
391{ 392{
392 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif); 393 struct iwl_mvm_vif *mvmvif = iwl_mvm_vif_from_mac80211(vif);
393 struct iwl_mvm_time_event_data *te_data = &mvmvif->time_event_data; 394 struct iwl_mvm_time_event_data *te_data = &mvmvif->time_event_data;
@@ -426,7 +427,7 @@ void iwl_mvm_protect_session(struct iwl_mvm *mvm,
426 cpu_to_le32(iwl_read_prph(mvm->trans, DEVICE_SYSTEM_TIME_REG)); 427 cpu_to_le32(iwl_read_prph(mvm->trans, DEVICE_SYSTEM_TIME_REG));
427 428
428 time_cmd.max_frags = TE_V2_FRAG_NONE; 429 time_cmd.max_frags = TE_V2_FRAG_NONE;
429 time_cmd.max_delay = cpu_to_le32(500); 430 time_cmd.max_delay = cpu_to_le32(max_delay);
430 /* TODO: why do we need to interval = bi if it is not periodic? */ 431 /* TODO: why do we need to interval = bi if it is not periodic? */
431 time_cmd.interval = cpu_to_le32(1); 432 time_cmd.interval = cpu_to_le32(1);
432 time_cmd.duration = cpu_to_le32(duration); 433 time_cmd.duration = cpu_to_le32(duration);
diff --git a/drivers/net/wireless/iwlwifi/mvm/time-event.h b/drivers/net/wireless/iwlwifi/mvm/time-event.h
index f86c51065ed3..d9c8d6cfa2db 100644
--- a/drivers/net/wireless/iwlwifi/mvm/time-event.h
+++ b/drivers/net/wireless/iwlwifi/mvm/time-event.h
@@ -123,6 +123,7 @@
123 * @duration: the duration of the session in TU. 123 * @duration: the duration of the session in TU.
124 * @min_duration: will start a new session if the current session will end 124 * @min_duration: will start a new session if the current session will end
125 * in less than min_duration. 125 * in less than min_duration.
126 * @max_delay: maximum delay before starting the time event (in TU)
126 * 127 *
127 * This function can be used to start a session protection which means that the 128 * This function can be used to start a session protection which means that the
128 * fw will stay on the channel for %duration_ms milliseconds. This function 129 * fw will stay on the channel for %duration_ms milliseconds. This function
@@ -133,7 +134,8 @@
133 */ 134 */
134void iwl_mvm_protect_session(struct iwl_mvm *mvm, 135void iwl_mvm_protect_session(struct iwl_mvm *mvm,
135 struct ieee80211_vif *vif, 136 struct ieee80211_vif *vif,
136 u32 duration, u32 min_duration); 137 u32 duration, u32 min_duration,
138 u32 max_delay);
137 139
138/** 140/**
139 * iwl_mvm_stop_session_protection - cancel the session protection. 141 * iwl_mvm_stop_session_protection - cancel the session protection.
diff --git a/drivers/net/wireless/iwlwifi/mvm/tx.c b/drivers/net/wireless/iwlwifi/mvm/tx.c
index e05440d90319..43d97c33a75a 100644
--- a/drivers/net/wireless/iwlwifi/mvm/tx.c
+++ b/drivers/net/wireless/iwlwifi/mvm/tx.c
@@ -417,7 +417,7 @@ int iwl_mvm_tx_skb(struct iwl_mvm *mvm, struct sk_buff *skb,
417 417
418 spin_unlock(&mvmsta->lock); 418 spin_unlock(&mvmsta->lock);
419 419
420 if (txq_id < IWL_MVM_FIRST_AGG_QUEUE) 420 if (txq_id < mvm->first_agg_queue)
421 atomic_inc(&mvm->pending_frames[mvmsta->sta_id]); 421 atomic_inc(&mvm->pending_frames[mvmsta->sta_id]);
422 422
423 return 0; 423 return 0;
@@ -511,16 +511,10 @@ const char *iwl_mvm_get_tx_fail_reason(u32 status)
511} 511}
512#endif /* CONFIG_IWLWIFI_DEBUG */ 512#endif /* CONFIG_IWLWIFI_DEBUG */
513 513
514/** 514void iwl_mvm_hwrate_to_tx_rate(u32 rate_n_flags,
515 * translate ucode response to mac80211 tx status control values 515 enum ieee80211_band band,
516 */ 516 struct ieee80211_tx_rate *r)
517static void iwl_mvm_hwrate_to_tx_control(u32 rate_n_flags,
518 struct ieee80211_tx_info *info)
519{ 517{
520 struct ieee80211_tx_rate *r = &info->status.rates[0];
521
522 info->status.antenna =
523 ((rate_n_flags & RATE_MCS_ANT_ABC_MSK) >> RATE_MCS_ANT_POS);
524 if (rate_n_flags & RATE_HT_MCS_GF_MSK) 518 if (rate_n_flags & RATE_HT_MCS_GF_MSK)
525 r->flags |= IEEE80211_TX_RC_GREEN_FIELD; 519 r->flags |= IEEE80211_TX_RC_GREEN_FIELD;
526 switch (rate_n_flags & RATE_MCS_CHAN_WIDTH_MSK) { 520 switch (rate_n_flags & RATE_MCS_CHAN_WIDTH_MSK) {
@@ -549,10 +543,23 @@ static void iwl_mvm_hwrate_to_tx_control(u32 rate_n_flags,
549 r->flags |= IEEE80211_TX_RC_VHT_MCS; 543 r->flags |= IEEE80211_TX_RC_VHT_MCS;
550 } else { 544 } else {
551 r->idx = iwl_mvm_legacy_rate_to_mac80211_idx(rate_n_flags, 545 r->idx = iwl_mvm_legacy_rate_to_mac80211_idx(rate_n_flags,
552 info->band); 546 band);
553 } 547 }
554} 548}
555 549
550/**
551 * translate ucode response to mac80211 tx status control values
552 */
553static void iwl_mvm_hwrate_to_tx_status(u32 rate_n_flags,
554 struct ieee80211_tx_info *info)
555{
556 struct ieee80211_tx_rate *r = &info->status.rates[0];
557
558 info->status.antenna =
559 ((rate_n_flags & RATE_MCS_ANT_ABC_MSK) >> RATE_MCS_ANT_POS);
560 iwl_mvm_hwrate_to_tx_rate(rate_n_flags, info->band, r);
561}
562
556static void iwl_mvm_rx_tx_cmd_single(struct iwl_mvm *mvm, 563static void iwl_mvm_rx_tx_cmd_single(struct iwl_mvm *mvm,
557 struct iwl_rx_packet *pkt) 564 struct iwl_rx_packet *pkt)
558{ 565{
@@ -602,11 +609,11 @@ static void iwl_mvm_rx_tx_cmd_single(struct iwl_mvm *mvm,
602 } 609 }
603 610
604 info->status.rates[0].count = tx_resp->failure_frame + 1; 611 info->status.rates[0].count = tx_resp->failure_frame + 1;
605 iwl_mvm_hwrate_to_tx_control(le32_to_cpu(tx_resp->initial_rate), 612 iwl_mvm_hwrate_to_tx_status(le32_to_cpu(tx_resp->initial_rate),
606 info); 613 info);
607 614
608 /* Single frame failure in an AMPDU queue => send BAR */ 615 /* Single frame failure in an AMPDU queue => send BAR */
609 if (txq_id >= IWL_MVM_FIRST_AGG_QUEUE && 616 if (txq_id >= mvm->first_agg_queue &&
610 !(info->flags & IEEE80211_TX_STAT_ACK)) 617 !(info->flags & IEEE80211_TX_STAT_ACK))
611 info->flags |= IEEE80211_TX_STAT_AMPDU_NO_BACK; 618 info->flags |= IEEE80211_TX_STAT_AMPDU_NO_BACK;
612 619
@@ -619,7 +626,7 @@ static void iwl_mvm_rx_tx_cmd_single(struct iwl_mvm *mvm,
619 ieee80211_tx_status_ni(mvm->hw, skb); 626 ieee80211_tx_status_ni(mvm->hw, skb);
620 } 627 }
621 628
622 if (txq_id >= IWL_MVM_FIRST_AGG_QUEUE) { 629 if (txq_id >= mvm->first_agg_queue) {
623 /* If this is an aggregation queue, we use the ssn since: 630 /* If this is an aggregation queue, we use the ssn since:
624 * ssn = wifi seq_num % 256. 631 * ssn = wifi seq_num % 256.
625 * The seq_ctl is the sequence control of the packet to which 632 * The seq_ctl is the sequence control of the packet to which
@@ -668,10 +675,6 @@ static void iwl_mvm_rx_tx_cmd_single(struct iwl_mvm *mvm,
668 iwl_mvm_check_ratid_empty(mvm, sta, tid); 675 iwl_mvm_check_ratid_empty(mvm, sta, tid);
669 spin_unlock_bh(&mvmsta->lock); 676 spin_unlock_bh(&mvmsta->lock);
670 } 677 }
671
672#ifdef CONFIG_PM_SLEEP
673 mvmsta->last_seq_ctl = seq_ctl;
674#endif
675 } else { 678 } else {
676 sta = NULL; 679 sta = NULL;
677 mvmsta = NULL; 680 mvmsta = NULL;
@@ -681,7 +684,7 @@ static void iwl_mvm_rx_tx_cmd_single(struct iwl_mvm *mvm,
681 * If the txq is not an AMPDU queue, there is no chance we freed 684 * If the txq is not an AMPDU queue, there is no chance we freed
682 * several skbs. Check that out... 685 * several skbs. Check that out...
683 */ 686 */
684 if (txq_id < IWL_MVM_FIRST_AGG_QUEUE && !WARN_ON(skb_freed > 1) && 687 if (txq_id < mvm->first_agg_queue && !WARN_ON(skb_freed > 1) &&
685 atomic_sub_and_test(skb_freed, &mvm->pending_frames[sta_id])) { 688 atomic_sub_and_test(skb_freed, &mvm->pending_frames[sta_id])) {
686 if (mvmsta) { 689 if (mvmsta) {
687 /* 690 /*
@@ -777,7 +780,7 @@ static void iwl_mvm_rx_tx_cmd_agg(struct iwl_mvm *mvm,
777 u16 sequence = le16_to_cpu(pkt->hdr.sequence); 780 u16 sequence = le16_to_cpu(pkt->hdr.sequence);
778 struct ieee80211_sta *sta; 781 struct ieee80211_sta *sta;
779 782
780 if (WARN_ON_ONCE(SEQ_TO_QUEUE(sequence) < IWL_MVM_FIRST_AGG_QUEUE)) 783 if (WARN_ON_ONCE(SEQ_TO_QUEUE(sequence) < mvm->first_agg_queue))
781 return; 784 return;
782 785
783 if (WARN_ON_ONCE(tid == IWL_TID_NON_QOS)) 786 if (WARN_ON_ONCE(tid == IWL_TID_NON_QOS))
@@ -904,8 +907,8 @@ int iwl_mvm_rx_ba_notif(struct iwl_mvm *mvm, struct iwl_rx_cmd_buffer *rxb,
904 info->flags |= IEEE80211_TX_STAT_AMPDU; 907 info->flags |= IEEE80211_TX_STAT_AMPDU;
905 info->status.ampdu_ack_len = ba_notif->txed_2_done; 908 info->status.ampdu_ack_len = ba_notif->txed_2_done;
906 info->status.ampdu_len = ba_notif->txed; 909 info->status.ampdu_len = ba_notif->txed;
907 iwl_mvm_hwrate_to_tx_control(tid_data->rate_n_flags, 910 iwl_mvm_hwrate_to_tx_status(tid_data->rate_n_flags,
908 info); 911 info);
909 } 912 }
910 } 913 }
911 914
diff --git a/drivers/net/wireless/iwlwifi/mvm/utils.c b/drivers/net/wireless/iwlwifi/mvm/utils.c
index a9c357491434..ed69e9b78e82 100644
--- a/drivers/net/wireless/iwlwifi/mvm/utils.c
+++ b/drivers/net/wireless/iwlwifi/mvm/utils.c
@@ -466,7 +466,7 @@ void iwl_mvm_dump_sram(struct iwl_mvm *mvm)
466 ofs = img->sec[IWL_UCODE_SECTION_DATA].offset; 466 ofs = img->sec[IWL_UCODE_SECTION_DATA].offset;
467 len = img->sec[IWL_UCODE_SECTION_DATA].len; 467 len = img->sec[IWL_UCODE_SECTION_DATA].len;
468 468
469 buf = kzalloc(len, GFP_KERNEL); 469 buf = kzalloc(len, GFP_ATOMIC);
470 if (!buf) 470 if (!buf)
471 return; 471 return;
472 472
diff --git a/drivers/net/wireless/iwlwifi/pcie/drv.c b/drivers/net/wireless/iwlwifi/pcie/drv.c
index 26108a1a29fa..941c0c88f982 100644
--- a/drivers/net/wireless/iwlwifi/pcie/drv.c
+++ b/drivers/net/wireless/iwlwifi/pcie/drv.c
@@ -268,7 +268,7 @@ static DEFINE_PCI_DEVICE_TABLE(iwl_hw_card_ids) = {
268#endif /* CONFIG_IWLDVM */ 268#endif /* CONFIG_IWLDVM */
269 269
270#if IS_ENABLED(CONFIG_IWLMVM) 270#if IS_ENABLED(CONFIG_IWLMVM)
271/* 7000 Series */ 271/* 7260 Series */
272 {IWL_PCI_DEVICE(0x08B1, 0x4070, iwl7260_2ac_cfg)}, 272 {IWL_PCI_DEVICE(0x08B1, 0x4070, iwl7260_2ac_cfg)},
273 {IWL_PCI_DEVICE(0x08B1, 0x4072, iwl7260_2ac_cfg)}, 273 {IWL_PCI_DEVICE(0x08B1, 0x4072, iwl7260_2ac_cfg)},
274 {IWL_PCI_DEVICE(0x08B1, 0x4170, iwl7260_2ac_cfg)}, 274 {IWL_PCI_DEVICE(0x08B1, 0x4170, iwl7260_2ac_cfg)},
@@ -350,6 +350,9 @@ static DEFINE_PCI_DEVICE_TABLE(iwl_hw_card_ids) = {
350 {IWL_PCI_DEVICE(0x08B4, 0x8270, iwl3160_2ac_cfg)}, 350 {IWL_PCI_DEVICE(0x08B4, 0x8270, iwl3160_2ac_cfg)},
351 {IWL_PCI_DEVICE(0x08B3, 0x8470, iwl3160_2ac_cfg)}, 351 {IWL_PCI_DEVICE(0x08B3, 0x8470, iwl3160_2ac_cfg)},
352 {IWL_PCI_DEVICE(0x08B3, 0x8570, iwl3160_2ac_cfg)}, 352 {IWL_PCI_DEVICE(0x08B3, 0x8570, iwl3160_2ac_cfg)},
353
354/* 7265 Series */
355 {IWL_PCI_DEVICE(0x095A, 0x5010, iwl7265_2ac_cfg)},
353#endif /* CONFIG_IWLMVM */ 356#endif /* CONFIG_IWLMVM */
354 357
355 {0} 358 {0}
@@ -391,7 +394,6 @@ out_free_drv:
391 iwl_drv_stop(trans_pcie->drv); 394 iwl_drv_stop(trans_pcie->drv);
392out_free_trans: 395out_free_trans:
393 iwl_trans_pcie_free(iwl_trans); 396 iwl_trans_pcie_free(iwl_trans);
394 pci_set_drvdata(pdev, NULL);
395 return ret; 397 return ret;
396} 398}
397 399
@@ -402,8 +404,6 @@ static void iwl_pci_remove(struct pci_dev *pdev)
402 404
403 iwl_drv_stop(trans_pcie->drv); 405 iwl_drv_stop(trans_pcie->drv);
404 iwl_trans_pcie_free(trans); 406 iwl_trans_pcie_free(trans);
405
406 pci_set_drvdata(pdev, NULL);
407} 407}
408 408
409#ifdef CONFIG_PM_SLEEP 409#ifdef CONFIG_PM_SLEEP
diff --git a/drivers/net/wireless/iwlwifi/pcie/trans.c b/drivers/net/wireless/iwlwifi/pcie/trans.c
index c3f904d422b0..5d9337bec67a 100644
--- a/drivers/net/wireless/iwlwifi/pcie/trans.c
+++ b/drivers/net/wireless/iwlwifi/pcie/trans.c
@@ -220,6 +220,9 @@ static int iwl_pcie_apm_init(struct iwl_trans *trans)
220 iwl_set_bits_prph(trans, APMG_PCIDEV_STT_REG, 220 iwl_set_bits_prph(trans, APMG_PCIDEV_STT_REG,
221 APMG_PCIDEV_STT_VAL_L1_ACT_DIS); 221 APMG_PCIDEV_STT_VAL_L1_ACT_DIS);
222 222
223 /* Clear the interrupt in APMG if the NIC is in RFKILL */
224 iwl_write_prph(trans, APMG_RTC_INT_STT_REG, APMG_RTC_INT_STT_RFKILL);
225
223 set_bit(STATUS_DEVICE_ENABLED, &trans_pcie->status); 226 set_bit(STATUS_DEVICE_ENABLED, &trans_pcie->status);
224 227
225out: 228out:
@@ -443,22 +446,138 @@ static int iwl_pcie_load_section(struct iwl_trans *trans, u8 section_num,
443 return ret; 446 return ret;
444} 447}
445 448
449static int iwl_pcie_secure_set(struct iwl_trans *trans, int cpu)
450{
451 int shift_param;
452 u32 address;
453 int ret = 0;
454
455 if (cpu == 1) {
456 shift_param = 0;
457 address = CSR_SECURE_BOOT_CPU1_STATUS_ADDR;
458 } else {
459 shift_param = 16;
460 address = CSR_SECURE_BOOT_CPU2_STATUS_ADDR;
461 }
462
463 /* set CPU to started */
464 iwl_trans_set_bits_mask(trans,
465 CSR_UCODE_LOAD_STATUS_ADDR,
466 CSR_CPU_STATUS_LOADING_STARTED << shift_param,
467 1);
468
469 /* set last complete descriptor number */
470 iwl_trans_set_bits_mask(trans,
471 CSR_UCODE_LOAD_STATUS_ADDR,
472 CSR_CPU_STATUS_NUM_OF_LAST_COMPLETED
473 << shift_param,
474 1);
475
476 /* set last loaded block */
477 iwl_trans_set_bits_mask(trans,
478 CSR_UCODE_LOAD_STATUS_ADDR,
479 CSR_CPU_STATUS_NUM_OF_LAST_LOADED_BLOCK
480 << shift_param,
481 1);
482
483 /* image loading complete */
484 iwl_trans_set_bits_mask(trans,
485 CSR_UCODE_LOAD_STATUS_ADDR,
486 CSR_CPU_STATUS_LOADING_COMPLETED
487 << shift_param,
488 1);
489
490 /* set FH_TCSR_0_REG */
491 iwl_trans_set_bits_mask(trans, FH_TCSR_0_REG0, 0x00400000, 1);
492
493 /* verify image verification started */
494 ret = iwl_poll_bit(trans, address,
495 CSR_SECURE_BOOT_CPU_STATUS_VERF_STATUS,
496 CSR_SECURE_BOOT_CPU_STATUS_VERF_STATUS,
497 CSR_SECURE_TIME_OUT);
498 if (ret < 0) {
499 IWL_ERR(trans, "secure boot process didn't start\n");
500 return ret;
501 }
502
503 /* wait for image verification to complete */
504 ret = iwl_poll_bit(trans, address,
505 CSR_SECURE_BOOT_CPU_STATUS_VERF_COMPLETED,
506 CSR_SECURE_BOOT_CPU_STATUS_VERF_COMPLETED,
507 CSR_SECURE_TIME_OUT);
508
509 if (ret < 0) {
510 IWL_ERR(trans, "Time out on secure boot process\n");
511 return ret;
512 }
513
514 return 0;
515}
516
446static int iwl_pcie_load_given_ucode(struct iwl_trans *trans, 517static int iwl_pcie_load_given_ucode(struct iwl_trans *trans,
447 const struct fw_img *image) 518 const struct fw_img *image)
448{ 519{
449 int i, ret = 0; 520 int i, ret = 0;
450 521
451 for (i = 0; i < IWL_UCODE_SECTION_MAX; i++) { 522 IWL_DEBUG_FW(trans,
523 "working with %s image\n",
524 image->is_secure ? "Secured" : "Non Secured");
525 IWL_DEBUG_FW(trans,
526 "working with %s CPU\n",
527 image->is_dual_cpus ? "Dual" : "Single");
528
529 /* configure the ucode to be ready to get the secured image */
530 if (image->is_secure) {
531 /* set secure boot inspector addresses */
532 iwl_write32(trans, CSR_SECURE_INSPECTOR_CODE_ADDR, 0);
533 iwl_write32(trans, CSR_SECURE_INSPECTOR_DATA_ADDR, 0);
534
535 /* release CPU1 reset if secure inspector image burned in OTP */
536 iwl_write32(trans, CSR_RESET, 0);
537 }
538
539 /* load to FW the binary sections of CPU1 */
540 IWL_DEBUG_INFO(trans, "Loading CPU1\n");
541 for (i = 0;
542 i < IWL_UCODE_FIRST_SECTION_OF_SECOND_CPU;
543 i++) {
452 if (!image->sec[i].data) 544 if (!image->sec[i].data)
453 break; 545 break;
454
455 ret = iwl_pcie_load_section(trans, i, &image->sec[i]); 546 ret = iwl_pcie_load_section(trans, i, &image->sec[i]);
456 if (ret) 547 if (ret)
457 return ret; 548 return ret;
458 } 549 }
459 550
460 /* Remove all resets to allow NIC to operate */ 551 /* configure the ucode to start secure process on CPU1 */
461 iwl_write32(trans, CSR_RESET, 0); 552 if (image->is_secure) {
553 /* config CPU1 to start secure protocol */
554 ret = iwl_pcie_secure_set(trans, 1);
555 if (ret)
556 return ret;
557 } else {
558 /* Remove all resets to allow NIC to operate */
559 iwl_write32(trans, CSR_RESET, 0);
560 }
561
562 if (image->is_dual_cpus) {
563 /* load to FW the binary sections of CPU2 */
564 IWL_DEBUG_INFO(trans, "working w/ DUAL CPUs - Loading CPU2\n");
565 for (i = IWL_UCODE_FIRST_SECTION_OF_SECOND_CPU;
566 i < IWL_UCODE_SECTION_MAX; i++) {
567 if (!image->sec[i].data)
568 break;
569 ret = iwl_pcie_load_section(trans, i, &image->sec[i]);
570 if (ret)
571 return ret;
572 }
573
574 if (image->is_secure) {
575 /* set CPU2 for secure protocol */
576 ret = iwl_pcie_secure_set(trans, 2);
577 if (ret)
578 return ret;
579 }
580 }
462 581
463 return 0; 582 return 0;
464} 583}
diff --git a/drivers/net/wireless/iwlwifi/pcie/tx.c b/drivers/net/wireless/iwlwifi/pcie/tx.c
index 1424335163b9..059c5acad3a0 100644
--- a/drivers/net/wireless/iwlwifi/pcie/tx.c
+++ b/drivers/net/wireless/iwlwifi/pcie/tx.c
@@ -1465,7 +1465,8 @@ void iwl_pcie_hcmd_complete(struct iwl_trans *trans,
1465 spin_unlock_bh(&txq->lock); 1465 spin_unlock_bh(&txq->lock);
1466} 1466}
1467 1467
1468#define HOST_COMPLETE_TIMEOUT (2 * HZ) 1468#define HOST_COMPLETE_TIMEOUT (2 * HZ)
1469#define COMMAND_POKE_TIMEOUT (HZ / 10)
1469 1470
1470static int iwl_pcie_send_hcmd_async(struct iwl_trans *trans, 1471static int iwl_pcie_send_hcmd_async(struct iwl_trans *trans,
1471 struct iwl_host_cmd *cmd) 1472 struct iwl_host_cmd *cmd)
@@ -1493,16 +1494,16 @@ static int iwl_pcie_send_hcmd_sync(struct iwl_trans *trans,
1493 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans); 1494 struct iwl_trans_pcie *trans_pcie = IWL_TRANS_GET_PCIE_TRANS(trans);
1494 int cmd_idx; 1495 int cmd_idx;
1495 int ret; 1496 int ret;
1497 int timeout = HOST_COMPLETE_TIMEOUT;
1496 1498
1497 IWL_DEBUG_INFO(trans, "Attempting to send sync command %s\n", 1499 IWL_DEBUG_INFO(trans, "Attempting to send sync command %s\n",
1498 get_cmd_string(trans_pcie, cmd->id)); 1500 get_cmd_string(trans_pcie, cmd->id));
1499 1501
1500 if (WARN_ON(test_and_set_bit(STATUS_HCMD_ACTIVE, 1502 if (WARN(test_and_set_bit(STATUS_HCMD_ACTIVE,
1501 &trans_pcie->status))) { 1503 &trans_pcie->status),
1502 IWL_ERR(trans, "Command %s: a command is already active!\n", 1504 "Command %s: a command is already active!\n",
1503 get_cmd_string(trans_pcie, cmd->id)); 1505 get_cmd_string(trans_pcie, cmd->id)))
1504 return -EIO; 1506 return -EIO;
1505 }
1506 1507
1507 IWL_DEBUG_INFO(trans, "Setting HCMD_ACTIVE for command %s\n", 1508 IWL_DEBUG_INFO(trans, "Setting HCMD_ACTIVE for command %s\n",
1508 get_cmd_string(trans_pcie, cmd->id)); 1509 get_cmd_string(trans_pcie, cmd->id));
@@ -1517,10 +1518,29 @@ static int iwl_pcie_send_hcmd_sync(struct iwl_trans *trans,
1517 return ret; 1518 return ret;
1518 } 1519 }
1519 1520
1520 ret = wait_event_timeout(trans_pcie->wait_command_queue, 1521 while (timeout > 0) {
1521 !test_bit(STATUS_HCMD_ACTIVE, 1522 unsigned long flags;
1522 &trans_pcie->status), 1523
1523 HOST_COMPLETE_TIMEOUT); 1524 timeout -= COMMAND_POKE_TIMEOUT;
1525 ret = wait_event_timeout(trans_pcie->wait_command_queue,
1526 !test_bit(STATUS_HCMD_ACTIVE,
1527 &trans_pcie->status),
1528 COMMAND_POKE_TIMEOUT);
1529 if (ret)
1530 break;
1531 /* poke the device - it may have lost the command */
1532 if (iwl_trans_grab_nic_access(trans, true, &flags)) {
1533 iwl_trans_release_nic_access(trans, &flags);
1534 IWL_DEBUG_INFO(trans,
1535 "Tried to wake NIC for command %s\n",
1536 get_cmd_string(trans_pcie, cmd->id));
1537 } else {
1538 IWL_ERR(trans, "Failed to poke NIC for command %s\n",
1539 get_cmd_string(trans_pcie, cmd->id));
1540 break;
1541 }
1542 }
1543
1524 if (!ret) { 1544 if (!ret) {
1525 if (test_bit(STATUS_HCMD_ACTIVE, &trans_pcie->status)) { 1545 if (test_bit(STATUS_HCMD_ACTIVE, &trans_pcie->status)) {
1526 struct iwl_txq *txq = 1546 struct iwl_txq *txq =
@@ -1541,6 +1561,9 @@ static int iwl_pcie_send_hcmd_sync(struct iwl_trans *trans,
1541 "Clearing HCMD_ACTIVE for command %s\n", 1561 "Clearing HCMD_ACTIVE for command %s\n",
1542 get_cmd_string(trans_pcie, cmd->id)); 1562 get_cmd_string(trans_pcie, cmd->id));
1543 ret = -ETIMEDOUT; 1563 ret = -ETIMEDOUT;
1564
1565 iwl_op_mode_nic_error(trans->op_mode);
1566
1544 goto cancel; 1567 goto cancel;
1545 } 1568 }
1546 } 1569 }
diff --git a/drivers/net/wireless/libertas/firmware.c b/drivers/net/wireless/libertas/firmware.c
index c0f9e7e862f6..51b92b5df119 100644
--- a/drivers/net/wireless/libertas/firmware.c
+++ b/drivers/net/wireless/libertas/firmware.c
@@ -53,6 +53,11 @@ static void main_firmware_cb(const struct firmware *firmware, void *context)
53 53
54 /* Firmware found! */ 54 /* Firmware found! */
55 lbs_fw_loaded(priv, 0, priv->helper_fw, firmware); 55 lbs_fw_loaded(priv, 0, priv->helper_fw, firmware);
56 if (priv->helper_fw) {
57 release_firmware (priv->helper_fw);
58 priv->helper_fw = NULL;
59 }
60 release_firmware (firmware);
56} 61}
57 62
58static void helper_firmware_cb(const struct firmware *firmware, void *context) 63static void helper_firmware_cb(const struct firmware *firmware, void *context)
diff --git a/drivers/net/wireless/libertas/if_cs.c b/drivers/net/wireless/libertas/if_cs.c
index c94dd6802672..ef8c98e21098 100644
--- a/drivers/net/wireless/libertas/if_cs.c
+++ b/drivers/net/wireless/libertas/if_cs.c
@@ -754,14 +754,14 @@ static void if_cs_prog_firmware(struct lbs_private *priv, int ret,
754 if (ret == 0 && (card->model != MODEL_8305)) 754 if (ret == 0 && (card->model != MODEL_8305))
755 ret = if_cs_prog_real(card, mainfw); 755 ret = if_cs_prog_real(card, mainfw);
756 if (ret) 756 if (ret)
757 goto out; 757 return;
758 758
759 /* Now actually get the IRQ */ 759 /* Now actually get the IRQ */
760 ret = request_irq(card->p_dev->irq, if_cs_interrupt, 760 ret = request_irq(card->p_dev->irq, if_cs_interrupt,
761 IRQF_SHARED, DRV_NAME, card); 761 IRQF_SHARED, DRV_NAME, card);
762 if (ret) { 762 if (ret) {
763 pr_err("error in request_irq\n"); 763 pr_err("error in request_irq\n");
764 goto out; 764 return;
765 } 765 }
766 766
767 /* 767 /*
@@ -777,10 +777,6 @@ static void if_cs_prog_firmware(struct lbs_private *priv, int ret,
777 pr_err("could not activate card\n"); 777 pr_err("could not activate card\n");
778 free_irq(card->p_dev->irq, card); 778 free_irq(card->p_dev->irq, card);
779 } 779 }
780
781out:
782 release_firmware(helper);
783 release_firmware(mainfw);
784} 780}
785 781
786 782
diff --git a/drivers/net/wireless/libertas/if_sdio.c b/drivers/net/wireless/libertas/if_sdio.c
index 45578335e420..991238afd1b6 100644
--- a/drivers/net/wireless/libertas/if_sdio.c
+++ b/drivers/net/wireless/libertas/if_sdio.c
@@ -708,20 +708,16 @@ static void if_sdio_do_prog_firmware(struct lbs_private *priv, int ret,
708 708
709 ret = if_sdio_prog_helper(card, helper); 709 ret = if_sdio_prog_helper(card, helper);
710 if (ret) 710 if (ret)
711 goto out; 711 return;
712 712
713 lbs_deb_sdio("Helper firmware loaded\n"); 713 lbs_deb_sdio("Helper firmware loaded\n");
714 714
715 ret = if_sdio_prog_real(card, mainfw); 715 ret = if_sdio_prog_real(card, mainfw);
716 if (ret) 716 if (ret)
717 goto out; 717 return;
718 718
719 lbs_deb_sdio("Firmware loaded\n"); 719 lbs_deb_sdio("Firmware loaded\n");
720 if_sdio_finish_power_on(card); 720 if_sdio_finish_power_on(card);
721
722out:
723 release_firmware(helper);
724 release_firmware(mainfw);
725} 721}
726 722
727static int if_sdio_prog_firmware(struct if_sdio_card *card) 723static int if_sdio_prog_firmware(struct if_sdio_card *card)
diff --git a/drivers/net/wireless/libertas/if_spi.c b/drivers/net/wireless/libertas/if_spi.c
index 4bb6574f4073..83669151bb82 100644
--- a/drivers/net/wireless/libertas/if_spi.c
+++ b/drivers/net/wireless/libertas/if_spi.c
@@ -1094,11 +1094,7 @@ static int if_spi_init_card(struct if_spi_card *card)
1094 goto out; 1094 goto out;
1095 1095
1096out: 1096out:
1097 release_firmware(helper);
1098 release_firmware(mainfw);
1099
1100 lbs_deb_leave_args(LBS_DEB_SPI, "err %d\n", err); 1097 lbs_deb_leave_args(LBS_DEB_SPI, "err %d\n", err);
1101
1102 return err; 1098 return err;
1103} 1099}
1104 1100
@@ -1128,7 +1124,7 @@ static int if_spi_probe(struct spi_device *spi)
1128{ 1124{
1129 struct if_spi_card *card; 1125 struct if_spi_card *card;
1130 struct lbs_private *priv = NULL; 1126 struct lbs_private *priv = NULL;
1131 struct libertas_spi_platform_data *pdata = spi->dev.platform_data; 1127 struct libertas_spi_platform_data *pdata = dev_get_platdata(&spi->dev);
1132 int err = 0; 1128 int err = 0;
1133 1129
1134 lbs_deb_enter(LBS_DEB_SPI); 1130 lbs_deb_enter(LBS_DEB_SPI);
diff --git a/drivers/net/wireless/libertas/if_usb.c b/drivers/net/wireless/libertas/if_usb.c
index 27980778d992..dff08a2896a3 100644
--- a/drivers/net/wireless/libertas/if_usb.c
+++ b/drivers/net/wireless/libertas/if_usb.c
@@ -844,7 +844,7 @@ static void if_usb_prog_firmware(struct lbs_private *priv, int ret,
844 cardp->fw = fw; 844 cardp->fw = fw;
845 if (check_fwfile_format(cardp->fw->data, cardp->fw->size)) { 845 if (check_fwfile_format(cardp->fw->data, cardp->fw->size)) {
846 ret = -EINVAL; 846 ret = -EINVAL;
847 goto release_fw; 847 goto done;
848 } 848 }
849 849
850 /* Cancel any pending usb business */ 850 /* Cancel any pending usb business */
@@ -861,7 +861,7 @@ restart:
861 if (if_usb_submit_rx_urb_fwload(cardp) < 0) { 861 if (if_usb_submit_rx_urb_fwload(cardp) < 0) {
862 lbs_deb_usbd(&cardp->udev->dev, "URB submission is failed\n"); 862 lbs_deb_usbd(&cardp->udev->dev, "URB submission is failed\n");
863 ret = -EIO; 863 ret = -EIO;
864 goto release_fw; 864 goto done;
865 } 865 }
866 866
867 cardp->bootcmdresp = 0; 867 cardp->bootcmdresp = 0;
@@ -883,14 +883,14 @@ restart:
883 usb_kill_urb(cardp->tx_urb); 883 usb_kill_urb(cardp->tx_urb);
884 if (if_usb_submit_rx_urb(cardp) < 0) 884 if (if_usb_submit_rx_urb(cardp) < 0)
885 ret = -EIO; 885 ret = -EIO;
886 goto release_fw; 886 goto done;
887 } else if (cardp->bootcmdresp <= 0) { 887 } else if (cardp->bootcmdresp <= 0) {
888 if (--reset_count >= 0) { 888 if (--reset_count >= 0) {
889 if_usb_reset_device(cardp); 889 if_usb_reset_device(cardp);
890 goto restart; 890 goto restart;
891 } 891 }
892 ret = -EIO; 892 ret = -EIO;
893 goto release_fw; 893 goto done;
894 } 894 }
895 895
896 i = 0; 896 i = 0;
@@ -921,14 +921,14 @@ restart:
921 921
922 pr_info("FW download failure, time = %d ms\n", i * 100); 922 pr_info("FW download failure, time = %d ms\n", i * 100);
923 ret = -EIO; 923 ret = -EIO;
924 goto release_fw; 924 goto done;
925 } 925 }
926 926
927 cardp->priv->fw_ready = 1; 927 cardp->priv->fw_ready = 1;
928 if_usb_submit_rx_urb(cardp); 928 if_usb_submit_rx_urb(cardp);
929 929
930 if (lbs_start_card(priv)) 930 if (lbs_start_card(priv))
931 goto release_fw; 931 goto done;
932 932
933 if_usb_setup_firmware(priv); 933 if_usb_setup_firmware(priv);
934 934
@@ -939,11 +939,8 @@ restart:
939 if (lbs_host_sleep_cfg(priv, priv->wol_criteria, NULL)) 939 if (lbs_host_sleep_cfg(priv, priv->wol_criteria, NULL))
940 priv->ehs_remove_supported = false; 940 priv->ehs_remove_supported = false;
941 941
942 release_fw:
943 release_firmware(cardp->fw);
944 cardp->fw = NULL;
945
946 done: 942 done:
943 cardp->fw = NULL;
947 lbs_deb_leave(LBS_DEB_USB); 944 lbs_deb_leave(LBS_DEB_USB);
948} 945}
949 946
diff --git a/drivers/net/wireless/mac80211_hwsim.c b/drivers/net/wireless/mac80211_hwsim.c
index 2cd3f54e1efa..de0df86704e7 100644
--- a/drivers/net/wireless/mac80211_hwsim.c
+++ b/drivers/net/wireless/mac80211_hwsim.c
@@ -167,6 +167,7 @@ struct hwsim_vif_priv {
167 u32 magic; 167 u32 magic;
168 u8 bssid[ETH_ALEN]; 168 u8 bssid[ETH_ALEN];
169 bool assoc; 169 bool assoc;
170 bool bcn_en;
170 u16 aid; 171 u16 aid;
171}; 172};
172 173
@@ -1170,6 +1171,16 @@ static void mac80211_hwsim_configure_filter(struct ieee80211_hw *hw,
1170 *total_flags = data->rx_filter; 1171 *total_flags = data->rx_filter;
1171} 1172}
1172 1173
1174static void mac80211_hwsim_bcn_en_iter(void *data, u8 *mac,
1175 struct ieee80211_vif *vif)
1176{
1177 unsigned int *count = data;
1178 struct hwsim_vif_priv *vp = (void *)vif->drv_priv;
1179
1180 if (vp->bcn_en)
1181 (*count)++;
1182}
1183
1173static void mac80211_hwsim_bss_info_changed(struct ieee80211_hw *hw, 1184static void mac80211_hwsim_bss_info_changed(struct ieee80211_hw *hw,
1174 struct ieee80211_vif *vif, 1185 struct ieee80211_vif *vif,
1175 struct ieee80211_bss_conf *info, 1186 struct ieee80211_bss_conf *info,
@@ -1180,7 +1191,8 @@ static void mac80211_hwsim_bss_info_changed(struct ieee80211_hw *hw,
1180 1191
1181 hwsim_check_magic(vif); 1192 hwsim_check_magic(vif);
1182 1193
1183 wiphy_debug(hw->wiphy, "%s(changed=0x%x)\n", __func__, changed); 1194 wiphy_debug(hw->wiphy, "%s(changed=0x%x vif->addr=%pM)\n",
1195 __func__, changed, vif->addr);
1184 1196
1185 if (changed & BSS_CHANGED_BSSID) { 1197 if (changed & BSS_CHANGED_BSSID) {
1186 wiphy_debug(hw->wiphy, "%s: BSSID changed: %pM\n", 1198 wiphy_debug(hw->wiphy, "%s: BSSID changed: %pM\n",
@@ -1202,6 +1214,7 @@ static void mac80211_hwsim_bss_info_changed(struct ieee80211_hw *hw,
1202 1214
1203 if (changed & BSS_CHANGED_BEACON_ENABLED) { 1215 if (changed & BSS_CHANGED_BEACON_ENABLED) {
1204 wiphy_debug(hw->wiphy, " BCN EN: %d\n", info->enable_beacon); 1216 wiphy_debug(hw->wiphy, " BCN EN: %d\n", info->enable_beacon);
1217 vp->bcn_en = info->enable_beacon;
1205 if (data->started && 1218 if (data->started &&
1206 !hrtimer_is_queued(&data->beacon_timer.timer) && 1219 !hrtimer_is_queued(&data->beacon_timer.timer) &&
1207 info->enable_beacon) { 1220 info->enable_beacon) {
@@ -1215,8 +1228,16 @@ static void mac80211_hwsim_bss_info_changed(struct ieee80211_hw *hw,
1215 tasklet_hrtimer_start(&data->beacon_timer, 1228 tasklet_hrtimer_start(&data->beacon_timer,
1216 ns_to_ktime(until_tbtt * 1000), 1229 ns_to_ktime(until_tbtt * 1000),
1217 HRTIMER_MODE_REL); 1230 HRTIMER_MODE_REL);
1218 } else if (!info->enable_beacon) 1231 } else if (!info->enable_beacon) {
1219 tasklet_hrtimer_cancel(&data->beacon_timer); 1232 unsigned int count = 0;
1233 ieee80211_iterate_active_interfaces(
1234 data->hw, IEEE80211_IFACE_ITER_NORMAL,
1235 mac80211_hwsim_bcn_en_iter, &count);
1236 wiphy_debug(hw->wiphy, " beaconing vifs remaining: %u",
1237 count);
1238 if (count == 0)
1239 tasklet_hrtimer_cancel(&data->beacon_timer);
1240 }
1220 } 1241 }
1221 1242
1222 if (changed & BSS_CHANGED_ERP_CTS_PROT) { 1243 if (changed & BSS_CHANGED_ERP_CTS_PROT) {
diff --git a/drivers/net/wireless/mwifiex/cmdevt.c b/drivers/net/wireless/mwifiex/cmdevt.c
index a6c46f3b6e3a..e47f4e3012b8 100644
--- a/drivers/net/wireless/mwifiex/cmdevt.c
+++ b/drivers/net/wireless/mwifiex/cmdevt.c
@@ -1048,7 +1048,7 @@ mwifiex_cancel_pending_ioctl(struct mwifiex_adapter *adapter)
1048 struct cmd_ctrl_node *cmd_node = NULL, *tmp_node = NULL; 1048 struct cmd_ctrl_node *cmd_node = NULL, *tmp_node = NULL;
1049 unsigned long cmd_flags; 1049 unsigned long cmd_flags;
1050 unsigned long scan_pending_q_flags; 1050 unsigned long scan_pending_q_flags;
1051 uint16_t cancel_scan_cmd = false; 1051 bool cancel_scan_cmd = false;
1052 1052
1053 if ((adapter->curr_cmd) && 1053 if ((adapter->curr_cmd) &&
1054 (adapter->curr_cmd->wait_q_enabled)) { 1054 (adapter->curr_cmd->wait_q_enabled)) {
diff --git a/drivers/net/wireless/mwifiex/join.c b/drivers/net/wireless/mwifiex/join.c
index 37f873bb342f..4e4686e6ac09 100644
--- a/drivers/net/wireless/mwifiex/join.c
+++ b/drivers/net/wireless/mwifiex/join.c
@@ -621,7 +621,7 @@ int mwifiex_ret_802_11_associate(struct mwifiex_private *priv,
621 int ret = 0; 621 int ret = 0;
622 struct ieee_types_assoc_rsp *assoc_rsp; 622 struct ieee_types_assoc_rsp *assoc_rsp;
623 struct mwifiex_bssdescriptor *bss_desc; 623 struct mwifiex_bssdescriptor *bss_desc;
624 u8 enable_data = true; 624 bool enable_data = true;
625 u16 cap_info, status_code; 625 u16 cap_info, status_code;
626 626
627 assoc_rsp = (struct ieee_types_assoc_rsp *) &resp->params; 627 assoc_rsp = (struct ieee_types_assoc_rsp *) &resp->params;
diff --git a/drivers/net/wireless/mwifiex/main.c b/drivers/net/wireless/mwifiex/main.c
index c2b91f566e05..9d7c9d354d34 100644
--- a/drivers/net/wireless/mwifiex/main.c
+++ b/drivers/net/wireless/mwifiex/main.c
@@ -882,7 +882,9 @@ mwifiex_add_card(void *card, struct semaphore *sem,
882 adapter->cmd_wait_q.status = 0; 882 adapter->cmd_wait_q.status = 0;
883 adapter->scan_wait_q_woken = false; 883 adapter->scan_wait_q_woken = false;
884 884
885 adapter->workqueue = create_workqueue("MWIFIEX_WORK_QUEUE"); 885 adapter->workqueue =
886 alloc_workqueue("MWIFIEX_WORK_QUEUE",
887 WQ_HIGHPRI | WQ_MEM_RECLAIM | WQ_UNBOUND, 1);
886 if (!adapter->workqueue) 888 if (!adapter->workqueue)
887 goto err_kmalloc; 889 goto err_kmalloc;
888 890
diff --git a/drivers/net/wireless/mwifiex/pcie.c b/drivers/net/wireless/mwifiex/pcie.c
index 52da8ee7599a..33fa9432b241 100644
--- a/drivers/net/wireless/mwifiex/pcie.c
+++ b/drivers/net/wireless/mwifiex/pcie.c
@@ -93,7 +93,7 @@ static int mwifiex_pcie_suspend(struct device *dev)
93 struct pci_dev *pdev = to_pci_dev(dev); 93 struct pci_dev *pdev = to_pci_dev(dev);
94 94
95 if (pdev) { 95 if (pdev) {
96 card = (struct pcie_service_card *) pci_get_drvdata(pdev); 96 card = pci_get_drvdata(pdev);
97 if (!card || !card->adapter) { 97 if (!card || !card->adapter) {
98 pr_err("Card or adapter structure is not valid\n"); 98 pr_err("Card or adapter structure is not valid\n");
99 return 0; 99 return 0;
@@ -128,7 +128,7 @@ static int mwifiex_pcie_resume(struct device *dev)
128 struct pci_dev *pdev = to_pci_dev(dev); 128 struct pci_dev *pdev = to_pci_dev(dev);
129 129
130 if (pdev) { 130 if (pdev) {
131 card = (struct pcie_service_card *) pci_get_drvdata(pdev); 131 card = pci_get_drvdata(pdev);
132 if (!card || !card->adapter) { 132 if (!card || !card->adapter) {
133 pr_err("Card or adapter structure is not valid\n"); 133 pr_err("Card or adapter structure is not valid\n");
134 return 0; 134 return 0;
@@ -2037,7 +2037,7 @@ static irqreturn_t mwifiex_pcie_interrupt(int irq, void *context)
2037 goto exit; 2037 goto exit;
2038 } 2038 }
2039 2039
2040 card = (struct pcie_service_card *) pci_get_drvdata(pdev); 2040 card = pci_get_drvdata(pdev);
2041 if (!card || !card->adapter) { 2041 if (!card || !card->adapter) {
2042 pr_debug("info: %s: card=%p adapter=%p\n", __func__, card, 2042 pr_debug("info: %s: card=%p adapter=%p\n", __func__, card,
2043 card ? card->adapter : NULL); 2043 card ? card->adapter : NULL);
diff --git a/drivers/net/wireless/mwifiex/sta_cmd.c b/drivers/net/wireless/mwifiex/sta_cmd.c
index c0268b597748..7d66018a2e33 100644
--- a/drivers/net/wireless/mwifiex/sta_cmd.c
+++ b/drivers/net/wireless/mwifiex/sta_cmd.c
@@ -327,7 +327,7 @@ mwifiex_cmd_802_11_hs_cfg(struct mwifiex_private *priv,
327{ 327{
328 struct mwifiex_adapter *adapter = priv->adapter; 328 struct mwifiex_adapter *adapter = priv->adapter;
329 struct host_cmd_ds_802_11_hs_cfg_enh *hs_cfg = &cmd->params.opt_hs_cfg; 329 struct host_cmd_ds_802_11_hs_cfg_enh *hs_cfg = &cmd->params.opt_hs_cfg;
330 u16 hs_activate = false; 330 bool hs_activate = false;
331 331
332 if (!hscfg_param) 332 if (!hscfg_param)
333 /* New Activate command */ 333 /* New Activate command */
diff --git a/drivers/net/wireless/mwifiex/wmm.c b/drivers/net/wireless/mwifiex/wmm.c
index 95fa3599b407..5dd0ccc70b86 100644
--- a/drivers/net/wireless/mwifiex/wmm.c
+++ b/drivers/net/wireless/mwifiex/wmm.c
@@ -708,7 +708,7 @@ int mwifiex_ret_wmm_get_status(struct mwifiex_private *priv,
708{ 708{
709 u8 *curr = (u8 *) &resp->params.get_wmm_status; 709 u8 *curr = (u8 *) &resp->params.get_wmm_status;
710 uint16_t resp_len = le16_to_cpu(resp->size), tlv_len; 710 uint16_t resp_len = le16_to_cpu(resp->size), tlv_len;
711 int valid = true; 711 bool valid = true;
712 712
713 struct mwifiex_ie_types_data *tlv_hdr; 713 struct mwifiex_ie_types_data *tlv_hdr;
714 struct mwifiex_ie_types_wmm_queue_status *tlv_wmm_qstatus; 714 struct mwifiex_ie_types_wmm_queue_status *tlv_wmm_qstatus;
diff --git a/drivers/net/wireless/mwifiex/wmm.h b/drivers/net/wireless/mwifiex/wmm.h
index 644d6e0c51cc..0f129d498fb1 100644
--- a/drivers/net/wireless/mwifiex/wmm.h
+++ b/drivers/net/wireless/mwifiex/wmm.h
@@ -83,11 +83,10 @@ mwifiex_wmm_is_ra_list_empty(struct list_head *ra_list_hhead)
83} 83}
84 84
85void mwifiex_wmm_add_buf_txqueue(struct mwifiex_private *priv, 85void mwifiex_wmm_add_buf_txqueue(struct mwifiex_private *priv,
86 struct sk_buff *skb); 86 struct sk_buff *skb);
87void mwifiex_ralist_add(struct mwifiex_private *priv, u8 *ra); 87void mwifiex_ralist_add(struct mwifiex_private *priv, u8 *ra);
88void mwifiex_rotate_priolists(struct mwifiex_private *priv, 88void mwifiex_rotate_priolists(struct mwifiex_private *priv,
89 struct mwifiex_ra_list_tbl *ra, 89 struct mwifiex_ra_list_tbl *ra, int tid);
90 int tid);
91 90
92int mwifiex_wmm_lists_empty(struct mwifiex_adapter *adapter); 91int mwifiex_wmm_lists_empty(struct mwifiex_adapter *adapter);
93void mwifiex_wmm_process_tx(struct mwifiex_adapter *adapter); 92void mwifiex_wmm_process_tx(struct mwifiex_adapter *adapter);
@@ -95,21 +94,18 @@ int mwifiex_is_ralist_valid(struct mwifiex_private *priv,
95 struct mwifiex_ra_list_tbl *ra_list, int tid); 94 struct mwifiex_ra_list_tbl *ra_list, int tid);
96 95
97u8 mwifiex_wmm_compute_drv_pkt_delay(struct mwifiex_private *priv, 96u8 mwifiex_wmm_compute_drv_pkt_delay(struct mwifiex_private *priv,
98 const struct sk_buff *skb); 97 const struct sk_buff *skb);
99void mwifiex_wmm_init(struct mwifiex_adapter *adapter); 98void mwifiex_wmm_init(struct mwifiex_adapter *adapter);
100 99
101extern u32 mwifiex_wmm_process_association_req(struct mwifiex_private *priv, 100u32 mwifiex_wmm_process_association_req(struct mwifiex_private *priv,
102 u8 **assoc_buf, 101 u8 **assoc_buf,
103 struct ieee_types_wmm_parameter 102 struct ieee_types_wmm_parameter *wmmie,
104 *wmmie, 103 struct ieee80211_ht_cap *htcap);
105 struct ieee80211_ht_cap
106 *htcap);
107 104
108void mwifiex_wmm_setup_queue_priorities(struct mwifiex_private *priv, 105void mwifiex_wmm_setup_queue_priorities(struct mwifiex_private *priv,
109 struct ieee_types_wmm_parameter 106 struct ieee_types_wmm_parameter *wmm_ie);
110 *wmm_ie);
111void mwifiex_wmm_setup_ac_downgrade(struct mwifiex_private *priv); 107void mwifiex_wmm_setup_ac_downgrade(struct mwifiex_private *priv);
112extern int mwifiex_ret_wmm_get_status(struct mwifiex_private *priv, 108int mwifiex_ret_wmm_get_status(struct mwifiex_private *priv,
113 const struct host_cmd_ds_command *resp); 109 const struct host_cmd_ds_command *resp);
114 110
115#endif /* !_MWIFIEX_WMM_H_ */ 111#endif /* !_MWIFIEX_WMM_H_ */
diff --git a/drivers/net/wireless/mwl8k.c b/drivers/net/wireless/mwl8k.c
index a3707fd4ef62..b953ad621e0b 100644
--- a/drivers/net/wireless/mwl8k.c
+++ b/drivers/net/wireless/mwl8k.c
@@ -6093,7 +6093,6 @@ err_iounmap:
6093 if (priv->sram != NULL) 6093 if (priv->sram != NULL)
6094 pci_iounmap(pdev, priv->sram); 6094 pci_iounmap(pdev, priv->sram);
6095 6095
6096 pci_set_drvdata(pdev, NULL);
6097 ieee80211_free_hw(hw); 6096 ieee80211_free_hw(hw);
6098 6097
6099err_free_reg: 6098err_free_reg:
@@ -6147,7 +6146,6 @@ static void mwl8k_remove(struct pci_dev *pdev)
6147unmap: 6146unmap:
6148 pci_iounmap(pdev, priv->regs); 6147 pci_iounmap(pdev, priv->regs);
6149 pci_iounmap(pdev, priv->sram); 6148 pci_iounmap(pdev, priv->sram);
6150 pci_set_drvdata(pdev, NULL);
6151 ieee80211_free_hw(hw); 6149 ieee80211_free_hw(hw);
6152 pci_release_regions(pdev); 6150 pci_release_regions(pdev);
6153 pci_disable_device(pdev); 6151 pci_disable_device(pdev);
diff --git a/drivers/net/wireless/orinoco/orinoco.h b/drivers/net/wireless/orinoco/orinoco.h
index 3bb936b9558c..eebd2be21ee9 100644
--- a/drivers/net/wireless/orinoco/orinoco.h
+++ b/drivers/net/wireless/orinoco/orinoco.h
@@ -182,23 +182,20 @@ extern int orinoco_debug;
182/* Exported prototypes */ 182/* Exported prototypes */
183/********************************************************************/ 183/********************************************************************/
184 184
185extern struct orinoco_private *alloc_orinocodev( 185struct orinoco_private *alloc_orinocodev(int sizeof_card, struct device *device,
186 int sizeof_card, struct device *device, 186 int (*hard_reset)(struct orinoco_private *),
187 int (*hard_reset)(struct orinoco_private *), 187 int (*stop_fw)(struct orinoco_private *, int));
188 int (*stop_fw)(struct orinoco_private *, int)); 188void free_orinocodev(struct orinoco_private *priv);
189extern void free_orinocodev(struct orinoco_private *priv); 189int orinoco_init(struct orinoco_private *priv);
190extern int orinoco_init(struct orinoco_private *priv); 190int orinoco_if_add(struct orinoco_private *priv, unsigned long base_addr,
191extern int orinoco_if_add(struct orinoco_private *priv, 191 unsigned int irq, const struct net_device_ops *ops);
192 unsigned long base_addr, 192void orinoco_if_del(struct orinoco_private *priv);
193 unsigned int irq, 193int orinoco_up(struct orinoco_private *priv);
194 const struct net_device_ops *ops); 194void orinoco_down(struct orinoco_private *priv);
195extern void orinoco_if_del(struct orinoco_private *priv); 195irqreturn_t orinoco_interrupt(int irq, void *dev_id);
196extern int orinoco_up(struct orinoco_private *priv); 196
197extern void orinoco_down(struct orinoco_private *priv); 197void __orinoco_ev_info(struct net_device *dev, struct hermes *hw);
198extern irqreturn_t orinoco_interrupt(int irq, void *dev_id); 198void __orinoco_ev_rx(struct net_device *dev, struct hermes *hw);
199
200extern void __orinoco_ev_info(struct net_device *dev, struct hermes *hw);
201extern void __orinoco_ev_rx(struct net_device *dev, struct hermes *hw);
202 199
203int orinoco_process_xmit_skb(struct sk_buff *skb, 200int orinoco_process_xmit_skb(struct sk_buff *skb,
204 struct net_device *dev, 201 struct net_device *dev,
diff --git a/drivers/net/wireless/orinoco/orinoco_nortel.c b/drivers/net/wireless/orinoco/orinoco_nortel.c
index d73fdf6185a2..ffb2469eb679 100644
--- a/drivers/net/wireless/orinoco/orinoco_nortel.c
+++ b/drivers/net/wireless/orinoco/orinoco_nortel.c
@@ -234,7 +234,6 @@ static int orinoco_nortel_init_one(struct pci_dev *pdev,
234 free_irq(pdev->irq, priv); 234 free_irq(pdev->irq, priv);
235 235
236 fail_irq: 236 fail_irq:
237 pci_set_drvdata(pdev, NULL);
238 free_orinocodev(priv); 237 free_orinocodev(priv);
239 238
240 fail_alloc: 239 fail_alloc:
@@ -265,7 +264,6 @@ static void orinoco_nortel_remove_one(struct pci_dev *pdev)
265 264
266 orinoco_if_del(priv); 265 orinoco_if_del(priv);
267 free_irq(pdev->irq, priv); 266 free_irq(pdev->irq, priv);
268 pci_set_drvdata(pdev, NULL);
269 free_orinocodev(priv); 267 free_orinocodev(priv);
270 pci_iounmap(pdev, priv->hw.iobase); 268 pci_iounmap(pdev, priv->hw.iobase);
271 pci_iounmap(pdev, card->attr_io); 269 pci_iounmap(pdev, card->attr_io);
diff --git a/drivers/net/wireless/orinoco/orinoco_pci.c b/drivers/net/wireless/orinoco/orinoco_pci.c
index 677bf14eca84..5ae1191d2532 100644
--- a/drivers/net/wireless/orinoco/orinoco_pci.c
+++ b/drivers/net/wireless/orinoco/orinoco_pci.c
@@ -184,7 +184,6 @@ static int orinoco_pci_init_one(struct pci_dev *pdev,
184 free_irq(pdev->irq, priv); 184 free_irq(pdev->irq, priv);
185 185
186 fail_irq: 186 fail_irq:
187 pci_set_drvdata(pdev, NULL);
188 free_orinocodev(priv); 187 free_orinocodev(priv);
189 188
190 fail_alloc: 189 fail_alloc:
@@ -205,7 +204,6 @@ static void orinoco_pci_remove_one(struct pci_dev *pdev)
205 204
206 orinoco_if_del(priv); 205 orinoco_if_del(priv);
207 free_irq(pdev->irq, priv); 206 free_irq(pdev->irq, priv);
208 pci_set_drvdata(pdev, NULL);
209 free_orinocodev(priv); 207 free_orinocodev(priv);
210 pci_iounmap(pdev, priv->hw.iobase); 208 pci_iounmap(pdev, priv->hw.iobase);
211 pci_release_regions(pdev); 209 pci_release_regions(pdev);
diff --git a/drivers/net/wireless/orinoco/orinoco_plx.c b/drivers/net/wireless/orinoco/orinoco_plx.c
index 2559dbd6184b..bbd36d1676ff 100644
--- a/drivers/net/wireless/orinoco/orinoco_plx.c
+++ b/drivers/net/wireless/orinoco/orinoco_plx.c
@@ -273,7 +273,6 @@ static int orinoco_plx_init_one(struct pci_dev *pdev,
273 free_irq(pdev->irq, priv); 273 free_irq(pdev->irq, priv);
274 274
275 fail_irq: 275 fail_irq:
276 pci_set_drvdata(pdev, NULL);
277 free_orinocodev(priv); 276 free_orinocodev(priv);
278 277
279 fail_alloc: 278 fail_alloc:
@@ -301,7 +300,6 @@ static void orinoco_plx_remove_one(struct pci_dev *pdev)
301 300
302 orinoco_if_del(priv); 301 orinoco_if_del(priv);
303 free_irq(pdev->irq, priv); 302 free_irq(pdev->irq, priv);
304 pci_set_drvdata(pdev, NULL);
305 free_orinocodev(priv); 303 free_orinocodev(priv);
306 pci_iounmap(pdev, priv->hw.iobase); 304 pci_iounmap(pdev, priv->hw.iobase);
307 pci_iounmap(pdev, card->attr_io); 305 pci_iounmap(pdev, card->attr_io);
diff --git a/drivers/net/wireless/orinoco/orinoco_tmd.c b/drivers/net/wireless/orinoco/orinoco_tmd.c
index 42afeeea2c40..04b08de5fd5d 100644
--- a/drivers/net/wireless/orinoco/orinoco_tmd.c
+++ b/drivers/net/wireless/orinoco/orinoco_tmd.c
@@ -170,7 +170,6 @@ static int orinoco_tmd_init_one(struct pci_dev *pdev,
170 free_irq(pdev->irq, priv); 170 free_irq(pdev->irq, priv);
171 171
172 fail_irq: 172 fail_irq:
173 pci_set_drvdata(pdev, NULL);
174 free_orinocodev(priv); 173 free_orinocodev(priv);
175 174
176 fail_alloc: 175 fail_alloc:
@@ -195,7 +194,6 @@ static void orinoco_tmd_remove_one(struct pci_dev *pdev)
195 194
196 orinoco_if_del(priv); 195 orinoco_if_del(priv);
197 free_irq(pdev->irq, priv); 196 free_irq(pdev->irq, priv);
198 pci_set_drvdata(pdev, NULL);
199 free_orinocodev(priv); 197 free_orinocodev(priv);
200 pci_iounmap(pdev, priv->hw.iobase); 198 pci_iounmap(pdev, priv->hw.iobase);
201 pci_iounmap(pdev, card->bridge_io); 199 pci_iounmap(pdev, card->bridge_io);
diff --git a/drivers/net/wireless/p54/p54pci.c b/drivers/net/wireless/p54/p54pci.c
index 57e3af8ebb4b..f9a07b0d83ac 100644
--- a/drivers/net/wireless/p54/p54pci.c
+++ b/drivers/net/wireless/p54/p54pci.c
@@ -631,7 +631,6 @@ static int p54p_probe(struct pci_dev *pdev,
631 iounmap(priv->map); 631 iounmap(priv->map);
632 632
633 err_free_dev: 633 err_free_dev:
634 pci_set_drvdata(pdev, NULL);
635 p54_free_common(dev); 634 p54_free_common(dev);
636 635
637 err_free_reg: 636 err_free_reg:
diff --git a/drivers/net/wireless/p54/p54spi.c b/drivers/net/wireless/p54/p54spi.c
index 7fc46f26cf2b..de15171e2cd8 100644
--- a/drivers/net/wireless/p54/p54spi.c
+++ b/drivers/net/wireless/p54/p54spi.c
@@ -636,7 +636,7 @@ static int p54spi_probe(struct spi_device *spi)
636 gpio_direction_input(p54spi_gpio_irq); 636 gpio_direction_input(p54spi_gpio_irq);
637 637
638 ret = request_irq(gpio_to_irq(p54spi_gpio_irq), 638 ret = request_irq(gpio_to_irq(p54spi_gpio_irq),
639 p54spi_interrupt, IRQF_DISABLED, "p54spi", 639 p54spi_interrupt, 0, "p54spi",
640 priv->spi); 640 priv->spi);
641 if (ret < 0) { 641 if (ret < 0) {
642 dev_err(&priv->spi->dev, "request_irq() failed"); 642 dev_err(&priv->spi->dev, "request_irq() failed");
diff --git a/drivers/net/wireless/prism54/isl_ioctl.c b/drivers/net/wireless/prism54/isl_ioctl.c
index 1c22b81e6ef3..8863a6cb2388 100644
--- a/drivers/net/wireless/prism54/isl_ioctl.c
+++ b/drivers/net/wireless/prism54/isl_ioctl.c
@@ -183,7 +183,7 @@ prism54_update_stats(struct work_struct *work)
183 data = r.ptr; 183 data = r.ptr;
184 184
185 /* copy this MAC to the bss */ 185 /* copy this MAC to the bss */
186 memcpy(bss.address, data, 6); 186 memcpy(bss.address, data, ETH_ALEN);
187 kfree(data); 187 kfree(data);
188 188
189 /* now ask for the corresponding bss */ 189 /* now ask for the corresponding bss */
@@ -531,7 +531,7 @@ prism54_set_wap(struct net_device *ndev, struct iw_request_info *info,
531 return -EINVAL; 531 return -EINVAL;
532 532
533 /* prepare the structure for the set object */ 533 /* prepare the structure for the set object */
534 memcpy(&bssid[0], awrq->sa_data, 6); 534 memcpy(&bssid[0], awrq->sa_data, ETH_ALEN);
535 535
536 /* set the bssid -- does this make sense when in AP mode? */ 536 /* set the bssid -- does this make sense when in AP mode? */
537 rvalue = mgt_set_request(priv, DOT11_OID_BSSID, 0, &bssid); 537 rvalue = mgt_set_request(priv, DOT11_OID_BSSID, 0, &bssid);
@@ -550,7 +550,7 @@ prism54_get_wap(struct net_device *ndev, struct iw_request_info *info,
550 int rvalue; 550 int rvalue;
551 551
552 rvalue = mgt_get_request(priv, DOT11_OID_BSSID, 0, NULL, &r); 552 rvalue = mgt_get_request(priv, DOT11_OID_BSSID, 0, NULL, &r);
553 memcpy(awrq->sa_data, r.ptr, 6); 553 memcpy(awrq->sa_data, r.ptr, ETH_ALEN);
554 awrq->sa_family = ARPHRD_ETHER; 554 awrq->sa_family = ARPHRD_ETHER;
555 kfree(r.ptr); 555 kfree(r.ptr);
556 556
@@ -582,7 +582,7 @@ prism54_translate_bss(struct net_device *ndev, struct iw_request_info *info,
582 size_t wpa_ie_len; 582 size_t wpa_ie_len;
583 583
584 /* The first entry must be the MAC address */ 584 /* The first entry must be the MAC address */
585 memcpy(iwe.u.ap_addr.sa_data, bss->address, 6); 585 memcpy(iwe.u.ap_addr.sa_data, bss->address, ETH_ALEN);
586 iwe.u.ap_addr.sa_family = ARPHRD_ETHER; 586 iwe.u.ap_addr.sa_family = ARPHRD_ETHER;
587 iwe.cmd = SIOCGIWAP; 587 iwe.cmd = SIOCGIWAP;
588 current_ev = iwe_stream_add_event(info, current_ev, end_buf, 588 current_ev = iwe_stream_add_event(info, current_ev, end_buf,
@@ -2489,7 +2489,7 @@ prism54_set_mac_address(struct net_device *ndev, void *addr)
2489 &((struct sockaddr *) addr)->sa_data); 2489 &((struct sockaddr *) addr)->sa_data);
2490 if (!ret) 2490 if (!ret)
2491 memcpy(priv->ndev->dev_addr, 2491 memcpy(priv->ndev->dev_addr,
2492 &((struct sockaddr *) addr)->sa_data, 6); 2492 &((struct sockaddr *) addr)->sa_data, ETH_ALEN);
2493 2493
2494 return ret; 2494 return ret;
2495} 2495}
diff --git a/drivers/net/wireless/prism54/islpci_dev.c b/drivers/net/wireless/prism54/islpci_dev.c
index 5970ff6f40cc..41a16d30c79c 100644
--- a/drivers/net/wireless/prism54/islpci_dev.c
+++ b/drivers/net/wireless/prism54/islpci_dev.c
@@ -837,7 +837,7 @@ islpci_setup(struct pci_dev *pdev)
837 /* ndev->set_multicast_list = &islpci_set_multicast_list; */ 837 /* ndev->set_multicast_list = &islpci_set_multicast_list; */
838 ndev->addr_len = ETH_ALEN; 838 ndev->addr_len = ETH_ALEN;
839 /* Get a non-zero dummy MAC address for nameif. Jean II */ 839 /* Get a non-zero dummy MAC address for nameif. Jean II */
840 memcpy(ndev->dev_addr, dummy_mac, 6); 840 memcpy(ndev->dev_addr, dummy_mac, ETH_ALEN);
841 841
842 ndev->watchdog_timeo = ISLPCI_TX_TIMEOUT; 842 ndev->watchdog_timeo = ISLPCI_TX_TIMEOUT;
843 843
diff --git a/drivers/net/wireless/prism54/oid_mgt.c b/drivers/net/wireless/prism54/oid_mgt.c
index a01606b36e03..056af38e72e3 100644
--- a/drivers/net/wireless/prism54/oid_mgt.c
+++ b/drivers/net/wireless/prism54/oid_mgt.c
@@ -682,7 +682,7 @@ mgt_update_addr(islpci_private *priv)
682 isl_oid[GEN_OID_MACADDRESS].size, &res); 682 isl_oid[GEN_OID_MACADDRESS].size, &res);
683 683
684 if ((ret == 0) && res && (res->header->operation != PIMFOR_OP_ERROR)) 684 if ((ret == 0) && res && (res->header->operation != PIMFOR_OP_ERROR))
685 memcpy(priv->ndev->dev_addr, res->data, 6); 685 memcpy(priv->ndev->dev_addr, res->data, ETH_ALEN);
686 else 686 else
687 ret = -EIO; 687 ret = -EIO;
688 if (res) 688 if (res)
diff --git a/drivers/net/wireless/rt2x00/Kconfig b/drivers/net/wireless/rt2x00/Kconfig
index 68dbbb9c6d12..006b8bcb2e31 100644
--- a/drivers/net/wireless/rt2x00/Kconfig
+++ b/drivers/net/wireless/rt2x00/Kconfig
@@ -58,11 +58,11 @@ config RT61PCI
58 58
59config RT2800PCI 59config RT2800PCI
60 tristate "Ralink rt27xx/rt28xx/rt30xx (PCI/PCIe/PCMCIA) support" 60 tristate "Ralink rt27xx/rt28xx/rt30xx (PCI/PCIe/PCMCIA) support"
61 depends on PCI || SOC_RT288X || SOC_RT305X 61 depends on PCI
62 select RT2800_LIB 62 select RT2800_LIB
63 select RT2800_LIB_MMIO
63 select RT2X00_LIB_MMIO 64 select RT2X00_LIB_MMIO
64 select RT2X00_LIB_PCI if PCI 65 select RT2X00_LIB_PCI
65 select RT2X00_LIB_SOC if SOC_RT288X || SOC_RT305X
66 select RT2X00_LIB_FIRMWARE 66 select RT2X00_LIB_FIRMWARE
67 select RT2X00_LIB_CRYPTO 67 select RT2X00_LIB_CRYPTO
68 select CRC_CCITT 68 select CRC_CCITT
@@ -199,9 +199,30 @@ config RT2800USB_UNKNOWN
199 199
200endif 200endif
201 201
202config RT2800SOC
203 tristate "Ralink WiSoC support"
204 depends on SOC_RT288X || SOC_RT305X
205 select RT2X00_LIB_SOC
206 select RT2X00_LIB_MMIO
207 select RT2X00_LIB_CRYPTO
208 select RT2X00_LIB_FIRMWARE
209 select RT2800_LIB
210 select RT2800_LIB_MMIO
211 ---help---
212 This adds support for Ralink WiSoC devices.
213 Supported chips: RT2880, RT3050, RT3052, RT3350, RT3352.
214
215 When compiled as a module, this driver will be called rt2800soc.
216
217
202config RT2800_LIB 218config RT2800_LIB
203 tristate 219 tristate
204 220
221config RT2800_LIB_MMIO
222 tristate
223 select RT2X00_LIB_MMIO
224 select RT2800_LIB
225
205config RT2X00_LIB_MMIO 226config RT2X00_LIB_MMIO
206 tristate 227 tristate
207 228
@@ -219,6 +240,7 @@ config RT2X00_LIB_USB
219 240
220config RT2X00_LIB 241config RT2X00_LIB
221 tristate 242 tristate
243 select AVERAGE
222 244
223config RT2X00_LIB_FIRMWARE 245config RT2X00_LIB_FIRMWARE
224 boolean 246 boolean
diff --git a/drivers/net/wireless/rt2x00/Makefile b/drivers/net/wireless/rt2x00/Makefile
index f069d8bc5b67..24a66015a495 100644
--- a/drivers/net/wireless/rt2x00/Makefile
+++ b/drivers/net/wireless/rt2x00/Makefile
@@ -14,6 +14,7 @@ obj-$(CONFIG_RT2X00_LIB_PCI) += rt2x00pci.o
14obj-$(CONFIG_RT2X00_LIB_SOC) += rt2x00soc.o 14obj-$(CONFIG_RT2X00_LIB_SOC) += rt2x00soc.o
15obj-$(CONFIG_RT2X00_LIB_USB) += rt2x00usb.o 15obj-$(CONFIG_RT2X00_LIB_USB) += rt2x00usb.o
16obj-$(CONFIG_RT2800_LIB) += rt2800lib.o 16obj-$(CONFIG_RT2800_LIB) += rt2800lib.o
17obj-$(CONFIG_RT2800_LIB_MMIO) += rt2800mmio.o
17obj-$(CONFIG_RT2400PCI) += rt2400pci.o 18obj-$(CONFIG_RT2400PCI) += rt2400pci.o
18obj-$(CONFIG_RT2500PCI) += rt2500pci.o 19obj-$(CONFIG_RT2500PCI) += rt2500pci.o
19obj-$(CONFIG_RT61PCI) += rt61pci.o 20obj-$(CONFIG_RT61PCI) += rt61pci.o
@@ -21,3 +22,4 @@ obj-$(CONFIG_RT2800PCI) += rt2800pci.o
21obj-$(CONFIG_RT2500USB) += rt2500usb.o 22obj-$(CONFIG_RT2500USB) += rt2500usb.o
22obj-$(CONFIG_RT73USB) += rt73usb.o 23obj-$(CONFIG_RT73USB) += rt73usb.o
23obj-$(CONFIG_RT2800USB) += rt2800usb.o 24obj-$(CONFIG_RT2800USB) += rt2800usb.o
25obj-$(CONFIG_RT2800SOC) += rt2800soc.o
diff --git a/drivers/net/wireless/rt2x00/rt2400pci.c b/drivers/net/wireless/rt2x00/rt2400pci.c
index 3d53a09da5a1..38ed9a3e44c8 100644
--- a/drivers/net/wireless/rt2x00/rt2400pci.c
+++ b/drivers/net/wireless/rt2x00/rt2400pci.c
@@ -1261,7 +1261,7 @@ static void rt2400pci_fill_rxdone(struct queue_entry *entry,
1261 */ 1261 */
1262 rxdesc->timestamp = ((u64)rx_high << 32) | rx_low; 1262 rxdesc->timestamp = ((u64)rx_high << 32) | rx_low;
1263 rxdesc->signal = rt2x00_get_field32(word2, RXD_W2_SIGNAL) & ~0x08; 1263 rxdesc->signal = rt2x00_get_field32(word2, RXD_W2_SIGNAL) & ~0x08;
1264 rxdesc->rssi = rt2x00_get_field32(word2, RXD_W3_RSSI) - 1264 rxdesc->rssi = rt2x00_get_field32(word3, RXD_W3_RSSI) -
1265 entry->queue->rt2x00dev->rssi_offset; 1265 entry->queue->rt2x00dev->rssi_offset;
1266 rxdesc->size = rt2x00_get_field32(word0, RXD_W0_DATABYTE_COUNT); 1266 rxdesc->size = rt2x00_get_field32(word0, RXD_W0_DATABYTE_COUNT);
1267 1267
diff --git a/drivers/net/wireless/rt2x00/rt2800.h b/drivers/net/wireless/rt2x00/rt2800.h
index fa33b5edf931..aab6b5e4f5dd 100644
--- a/drivers/net/wireless/rt2x00/rt2800.h
+++ b/drivers/net/wireless/rt2x00/rt2800.h
@@ -52,6 +52,7 @@
52 * RF3322 2.4G 2T2R(RT3352/RT3371/RT3372/RT3391/RT3392) 52 * RF3322 2.4G 2T2R(RT3352/RT3371/RT3372/RT3391/RT3392)
53 * RF3053 2.4G/5G 3T3R(RT3883/RT3563/RT3573/RT3593/RT3662) 53 * RF3053 2.4G/5G 3T3R(RT3883/RT3563/RT3573/RT3593/RT3662)
54 * RF5592 2.4G/5G 2T2R 54 * RF5592 2.4G/5G 2T2R
55 * RF3070 2.4G 1T1R
55 * RF5360 2.4G 1T1R 56 * RF5360 2.4G 1T1R
56 * RF5370 2.4G 1T1R 57 * RF5370 2.4G 1T1R
57 * RF5390 2.4G 1T1R 58 * RF5390 2.4G 1T1R
@@ -70,6 +71,7 @@
70#define RF3322 0x000c 71#define RF3322 0x000c
71#define RF3053 0x000d 72#define RF3053 0x000d
72#define RF5592 0x000f 73#define RF5592 0x000f
74#define RF3070 0x3070
73#define RF3290 0x3290 75#define RF3290 0x3290
74#define RF5360 0x5360 76#define RF5360 0x5360
75#define RF5370 0x5370 77#define RF5370 0x5370
@@ -122,7 +124,7 @@
122/* 124/*
123 * MAC_CSR0_3290: MAC_CSR0 for RT3290 to identity MAC version number. 125 * MAC_CSR0_3290: MAC_CSR0 for RT3290 to identity MAC version number.
124 */ 126 */
125#define MAC_CSR0_3290 0x0000 127#define MAC_CSR0_3290 0x0000
126 128
127/* 129/*
128 * E2PROM_CSR: PCI EEPROM control register. 130 * E2PROM_CSR: PCI EEPROM control register.
@@ -211,17 +213,17 @@
211/* 213/*
212 * COEX_CFG_0 214 * COEX_CFG_0
213 */ 215 */
214#define COEX_CFG0 0x0040 216#define COEX_CFG0 0x0040
215#define COEX_CFG_ANT FIELD32(0xff000000) 217#define COEX_CFG_ANT FIELD32(0xff000000)
216/* 218/*
217 * COEX_CFG_1 219 * COEX_CFG_1
218 */ 220 */
219#define COEX_CFG1 0x0044 221#define COEX_CFG1 0x0044
220 222
221/* 223/*
222 * COEX_CFG_2 224 * COEX_CFG_2
223 */ 225 */
224#define COEX_CFG2 0x0048 226#define COEX_CFG2 0x0048
225#define BT_COEX_CFG1 FIELD32(0xff000000) 227#define BT_COEX_CFG1 FIELD32(0xff000000)
226#define BT_COEX_CFG0 FIELD32(0x00ff0000) 228#define BT_COEX_CFG0 FIELD32(0x00ff0000)
227#define WL_COEX_CFG1 FIELD32(0x0000ff00) 229#define WL_COEX_CFG1 FIELD32(0x0000ff00)
@@ -235,8 +237,8 @@
235#define PLL_RESERVED_INPUT2 FIELD32(0x0000ff00) 237#define PLL_RESERVED_INPUT2 FIELD32(0x0000ff00)
236#define PLL_CONTROL FIELD32(0x00070000) 238#define PLL_CONTROL FIELD32(0x00070000)
237#define PLL_LPF_R1 FIELD32(0x00080000) 239#define PLL_LPF_R1 FIELD32(0x00080000)
238#define PLL_LPF_C1_CTRL FIELD32(0x00300000) 240#define PLL_LPF_C1_CTRL FIELD32(0x00300000)
239#define PLL_LPF_C2_CTRL FIELD32(0x00c00000) 241#define PLL_LPF_C2_CTRL FIELD32(0x00c00000)
240#define PLL_CP_CURRENT_CTRL FIELD32(0x03000000) 242#define PLL_CP_CURRENT_CTRL FIELD32(0x03000000)
241#define PLL_PFD_DELAY_CTRL FIELD32(0x0c000000) 243#define PLL_PFD_DELAY_CTRL FIELD32(0x0c000000)
242#define PLL_LOCK_CTRL FIELD32(0x70000000) 244#define PLL_LOCK_CTRL FIELD32(0x70000000)
@@ -2164,7 +2166,7 @@ struct mac_iveiv_entry {
2164 */ 2166 */
2165#define RFCSR6_R1 FIELD8(0x03) 2167#define RFCSR6_R1 FIELD8(0x03)
2166#define RFCSR6_R2 FIELD8(0x40) 2168#define RFCSR6_R2 FIELD8(0x40)
2167#define RFCSR6_TXDIV FIELD8(0x0c) 2169#define RFCSR6_TXDIV FIELD8(0x0c)
2168/* bits for RF3053 */ 2170/* bits for RF3053 */
2169#define RFCSR6_VCO_IC FIELD8(0xc0) 2171#define RFCSR6_VCO_IC FIELD8(0xc0)
2170 2172
@@ -2202,13 +2204,13 @@ struct mac_iveiv_entry {
2202 * RFCSR 12: 2204 * RFCSR 12:
2203 */ 2205 */
2204#define RFCSR12_TX_POWER FIELD8(0x1f) 2206#define RFCSR12_TX_POWER FIELD8(0x1f)
2205#define RFCSR12_DR0 FIELD8(0xe0) 2207#define RFCSR12_DR0 FIELD8(0xe0)
2206 2208
2207/* 2209/*
2208 * RFCSR 13: 2210 * RFCSR 13:
2209 */ 2211 */
2210#define RFCSR13_TX_POWER FIELD8(0x1f) 2212#define RFCSR13_TX_POWER FIELD8(0x1f)
2211#define RFCSR13_DR0 FIELD8(0xe0) 2213#define RFCSR13_DR0 FIELD8(0xe0)
2212 2214
2213/* 2215/*
2214 * RFCSR 15: 2216 * RFCSR 15:
@@ -2226,7 +2228,7 @@ struct mac_iveiv_entry {
2226#define RFCSR17_TXMIXER_GAIN FIELD8(0x07) 2228#define RFCSR17_TXMIXER_GAIN FIELD8(0x07)
2227#define RFCSR17_TX_LO1_EN FIELD8(0x08) 2229#define RFCSR17_TX_LO1_EN FIELD8(0x08)
2228#define RFCSR17_R FIELD8(0x20) 2230#define RFCSR17_R FIELD8(0x20)
2229#define RFCSR17_CODE FIELD8(0x7f) 2231#define RFCSR17_CODE FIELD8(0x7f)
2230 2232
2231/* RFCSR 18 */ 2233/* RFCSR 18 */
2232#define RFCSR18_XO_TUNE_BYPASS FIELD8(0x40) 2234#define RFCSR18_XO_TUNE_BYPASS FIELD8(0x40)
@@ -2449,7 +2451,7 @@ enum rt2800_eeprom_word {
2449 */ 2451 */
2450#define EEPROM_NIC_CONF0_RXPATH FIELD16(0x000f) 2452#define EEPROM_NIC_CONF0_RXPATH FIELD16(0x000f)
2451#define EEPROM_NIC_CONF0_TXPATH FIELD16(0x00f0) 2453#define EEPROM_NIC_CONF0_TXPATH FIELD16(0x00f0)
2452#define EEPROM_NIC_CONF0_RF_TYPE FIELD16(0x0f00) 2454#define EEPROM_NIC_CONF0_RF_TYPE FIELD16(0x0f00)
2453 2455
2454/* 2456/*
2455 * EEPROM NIC Configuration 1 2457 * EEPROM NIC Configuration 1
@@ -2471,18 +2473,18 @@ enum rt2800_eeprom_word {
2471 * DAC_TEST: 0: disable, 1: enable 2473 * DAC_TEST: 0: disable, 1: enable
2472 */ 2474 */
2473#define EEPROM_NIC_CONF1_HW_RADIO FIELD16(0x0001) 2475#define EEPROM_NIC_CONF1_HW_RADIO FIELD16(0x0001)
2474#define EEPROM_NIC_CONF1_EXTERNAL_TX_ALC FIELD16(0x0002) 2476#define EEPROM_NIC_CONF1_EXTERNAL_TX_ALC FIELD16(0x0002)
2475#define EEPROM_NIC_CONF1_EXTERNAL_LNA_2G FIELD16(0x0004) 2477#define EEPROM_NIC_CONF1_EXTERNAL_LNA_2G FIELD16(0x0004)
2476#define EEPROM_NIC_CONF1_EXTERNAL_LNA_5G FIELD16(0x0008) 2478#define EEPROM_NIC_CONF1_EXTERNAL_LNA_5G FIELD16(0x0008)
2477#define EEPROM_NIC_CONF1_CARDBUS_ACCEL FIELD16(0x0010) 2479#define EEPROM_NIC_CONF1_CARDBUS_ACCEL FIELD16(0x0010)
2478#define EEPROM_NIC_CONF1_BW40M_SB_2G FIELD16(0x0020) 2480#define EEPROM_NIC_CONF1_BW40M_SB_2G FIELD16(0x0020)
2479#define EEPROM_NIC_CONF1_BW40M_SB_5G FIELD16(0x0040) 2481#define EEPROM_NIC_CONF1_BW40M_SB_5G FIELD16(0x0040)
2480#define EEPROM_NIC_CONF1_WPS_PBC FIELD16(0x0080) 2482#define EEPROM_NIC_CONF1_WPS_PBC FIELD16(0x0080)
2481#define EEPROM_NIC_CONF1_BW40M_2G FIELD16(0x0100) 2483#define EEPROM_NIC_CONF1_BW40M_2G FIELD16(0x0100)
2482#define EEPROM_NIC_CONF1_BW40M_5G FIELD16(0x0200) 2484#define EEPROM_NIC_CONF1_BW40M_5G FIELD16(0x0200)
2483#define EEPROM_NIC_CONF1_BROADBAND_EXT_LNA FIELD16(0x400) 2485#define EEPROM_NIC_CONF1_BROADBAND_EXT_LNA FIELD16(0x400)
2484#define EEPROM_NIC_CONF1_ANT_DIVERSITY FIELD16(0x1800) 2486#define EEPROM_NIC_CONF1_ANT_DIVERSITY FIELD16(0x1800)
2485#define EEPROM_NIC_CONF1_INTERNAL_TX_ALC FIELD16(0x2000) 2487#define EEPROM_NIC_CONF1_INTERNAL_TX_ALC FIELD16(0x2000)
2486#define EEPROM_NIC_CONF1_BT_COEXIST FIELD16(0x4000) 2488#define EEPROM_NIC_CONF1_BT_COEXIST FIELD16(0x4000)
2487#define EEPROM_NIC_CONF1_DAC_TEST FIELD16(0x8000) 2489#define EEPROM_NIC_CONF1_DAC_TEST FIELD16(0x8000)
2488 2490
@@ -2521,9 +2523,9 @@ enum rt2800_eeprom_word {
2521 * TX_STREAM: 0: Reserved, 1: 1 Stream, 2: 2 Stream 2523 * TX_STREAM: 0: Reserved, 1: 1 Stream, 2: 2 Stream
2522 * CRYSTAL: 00: Reserved, 01: One crystal, 10: Two crystal, 11: Reserved 2524 * CRYSTAL: 00: Reserved, 01: One crystal, 10: Two crystal, 11: Reserved
2523 */ 2525 */
2524#define EEPROM_NIC_CONF2_RX_STREAM FIELD16(0x000f) 2526#define EEPROM_NIC_CONF2_RX_STREAM FIELD16(0x000f)
2525#define EEPROM_NIC_CONF2_TX_STREAM FIELD16(0x00f0) 2527#define EEPROM_NIC_CONF2_TX_STREAM FIELD16(0x00f0)
2526#define EEPROM_NIC_CONF2_CRYSTAL FIELD16(0x0600) 2528#define EEPROM_NIC_CONF2_CRYSTAL FIELD16(0x0600)
2527 2529
2528/* 2530/*
2529 * EEPROM LNA 2531 * EEPROM LNA
@@ -2790,7 +2792,7 @@ enum rt2800_eeprom_word {
2790#define MCU_CURRENT 0x36 2792#define MCU_CURRENT 0x36
2791#define MCU_LED 0x50 2793#define MCU_LED 0x50
2792#define MCU_LED_STRENGTH 0x51 2794#define MCU_LED_STRENGTH 0x51
2793#define MCU_LED_AG_CONF 0x52 2795#define MCU_LED_AG_CONF 0x52
2794#define MCU_LED_ACT_CONF 0x53 2796#define MCU_LED_ACT_CONF 0x53
2795#define MCU_LED_LED_POLARITY 0x54 2797#define MCU_LED_LED_POLARITY 0x54
2796#define MCU_RADAR 0x60 2798#define MCU_RADAR 0x60
@@ -2799,7 +2801,7 @@ enum rt2800_eeprom_word {
2799#define MCU_FREQ_OFFSET 0x74 2801#define MCU_FREQ_OFFSET 0x74
2800#define MCU_BBP_SIGNAL 0x80 2802#define MCU_BBP_SIGNAL 0x80
2801#define MCU_POWER_SAVE 0x83 2803#define MCU_POWER_SAVE 0x83
2802#define MCU_BAND_SELECT 0x91 2804#define MCU_BAND_SELECT 0x91
2803 2805
2804/* 2806/*
2805 * MCU mailbox tokens 2807 * MCU mailbox tokens
diff --git a/drivers/net/wireless/rt2x00/rt2800lib.c b/drivers/net/wireless/rt2x00/rt2800lib.c
index 88ce656f96cd..c5738f14c4ba 100644
--- a/drivers/net/wireless/rt2x00/rt2800lib.c
+++ b/drivers/net/wireless/rt2x00/rt2800lib.c
@@ -278,12 +278,9 @@ static const unsigned int rt2800_eeprom_map_ext[EEPROM_WORD_COUNT] = {
278 [EEPROM_LNA] = 0x0026, 278 [EEPROM_LNA] = 0x0026,
279 [EEPROM_EXT_LNA2] = 0x0027, 279 [EEPROM_EXT_LNA2] = 0x0027,
280 [EEPROM_RSSI_BG] = 0x0028, 280 [EEPROM_RSSI_BG] = 0x0028,
281 [EEPROM_TXPOWER_DELTA] = 0x0028, /* Overlaps with RSSI_BG */
282 [EEPROM_RSSI_BG2] = 0x0029, 281 [EEPROM_RSSI_BG2] = 0x0029,
283 [EEPROM_TXMIXER_GAIN_BG] = 0x0029, /* Overlaps with RSSI_BG2 */
284 [EEPROM_RSSI_A] = 0x002a, 282 [EEPROM_RSSI_A] = 0x002a,
285 [EEPROM_RSSI_A2] = 0x002b, 283 [EEPROM_RSSI_A2] = 0x002b,
286 [EEPROM_TXMIXER_GAIN_A] = 0x002b, /* Overlaps with RSSI_A2 */
287 [EEPROM_TXPOWER_BG1] = 0x0030, 284 [EEPROM_TXPOWER_BG1] = 0x0030,
288 [EEPROM_TXPOWER_BG2] = 0x0037, 285 [EEPROM_TXPOWER_BG2] = 0x0037,
289 [EEPROM_EXT_TXPOWER_BG3] = 0x003e, 286 [EEPROM_EXT_TXPOWER_BG3] = 0x003e,
@@ -1783,7 +1780,7 @@ void rt2800_config_ant(struct rt2x00_dev *rt2x00dev, struct antenna_setup *ant)
1783 rt2800_bbp_read(rt2x00dev, 3, &r3); 1780 rt2800_bbp_read(rt2x00dev, 3, &r3);
1784 1781
1785 if (rt2x00_rt(rt2x00dev, RT3572) && 1782 if (rt2x00_rt(rt2x00dev, RT3572) &&
1786 test_bit(CAPABILITY_BT_COEXIST, &rt2x00dev->cap_flags)) 1783 rt2x00_has_cap_bt_coexist(rt2x00dev))
1787 rt2800_config_3572bt_ant(rt2x00dev); 1784 rt2800_config_3572bt_ant(rt2x00dev);
1788 1785
1789 /* 1786 /*
@@ -1795,7 +1792,7 @@ void rt2800_config_ant(struct rt2x00_dev *rt2x00dev, struct antenna_setup *ant)
1795 break; 1792 break;
1796 case 2: 1793 case 2:
1797 if (rt2x00_rt(rt2x00dev, RT3572) && 1794 if (rt2x00_rt(rt2x00dev, RT3572) &&
1798 test_bit(CAPABILITY_BT_COEXIST, &rt2x00dev->cap_flags)) 1795 rt2x00_has_cap_bt_coexist(rt2x00dev))
1799 rt2x00_set_field8(&r1, BBP1_TX_ANTENNA, 1); 1796 rt2x00_set_field8(&r1, BBP1_TX_ANTENNA, 1);
1800 else 1797 else
1801 rt2x00_set_field8(&r1, BBP1_TX_ANTENNA, 2); 1798 rt2x00_set_field8(&r1, BBP1_TX_ANTENNA, 2);
@@ -1825,7 +1822,7 @@ void rt2800_config_ant(struct rt2x00_dev *rt2x00dev, struct antenna_setup *ant)
1825 break; 1822 break;
1826 case 2: 1823 case 2:
1827 if (rt2x00_rt(rt2x00dev, RT3572) && 1824 if (rt2x00_rt(rt2x00dev, RT3572) &&
1828 test_bit(CAPABILITY_BT_COEXIST, &rt2x00dev->cap_flags)) { 1825 rt2x00_has_cap_bt_coexist(rt2x00dev)) {
1829 rt2x00_set_field8(&r3, BBP3_RX_ADC, 1); 1826 rt2x00_set_field8(&r3, BBP3_RX_ADC, 1);
1830 rt2x00_set_field8(&r3, BBP3_RX_ANTENNA, 1827 rt2x00_set_field8(&r3, BBP3_RX_ANTENNA,
1831 rt2x00dev->curr_band == IEEE80211_BAND_5GHZ); 1828 rt2x00dev->curr_band == IEEE80211_BAND_5GHZ);
@@ -2029,13 +2026,6 @@ static void rt2800_config_channel_rf3xxx(struct rt2x00_dev *rt2x00dev,
2029 rt2x00dev->default_ant.tx_chain_num <= 2); 2026 rt2x00dev->default_ant.tx_chain_num <= 2);
2030 rt2800_rfcsr_write(rt2x00dev, 1, rfcsr); 2027 rt2800_rfcsr_write(rt2x00dev, 1, rfcsr);
2031 2028
2032 rt2800_rfcsr_read(rt2x00dev, 30, &rfcsr);
2033 rt2x00_set_field8(&rfcsr, RFCSR30_RF_CALIBRATION, 1);
2034 rt2800_rfcsr_write(rt2x00dev, 30, rfcsr);
2035 msleep(1);
2036 rt2x00_set_field8(&rfcsr, RFCSR30_RF_CALIBRATION, 0);
2037 rt2800_rfcsr_write(rt2x00dev, 30, rfcsr);
2038
2039 rt2800_rfcsr_read(rt2x00dev, 23, &rfcsr); 2029 rt2800_rfcsr_read(rt2x00dev, 23, &rfcsr);
2040 rt2x00_set_field8(&rfcsr, RFCSR23_FREQ_OFFSET, rt2x00dev->freq_offset); 2030 rt2x00_set_field8(&rfcsr, RFCSR23_FREQ_OFFSET, rt2x00dev->freq_offset);
2041 rt2800_rfcsr_write(rt2x00dev, 23, rfcsr); 2031 rt2800_rfcsr_write(rt2x00dev, 23, rfcsr);
@@ -2141,7 +2131,7 @@ static void rt2800_config_channel_rf3052(struct rt2x00_dev *rt2x00dev,
2141 rt2x00_set_field8(&rfcsr, RFCSR1_TX1_PD, 0); 2131 rt2x00_set_field8(&rfcsr, RFCSR1_TX1_PD, 0);
2142 rt2x00_set_field8(&rfcsr, RFCSR1_RX2_PD, 0); 2132 rt2x00_set_field8(&rfcsr, RFCSR1_RX2_PD, 0);
2143 rt2x00_set_field8(&rfcsr, RFCSR1_TX2_PD, 0); 2133 rt2x00_set_field8(&rfcsr, RFCSR1_TX2_PD, 0);
2144 if (test_bit(CAPABILITY_BT_COEXIST, &rt2x00dev->cap_flags)) { 2134 if (rt2x00_has_cap_bt_coexist(rt2x00dev)) {
2145 if (rf->channel <= 14) { 2135 if (rf->channel <= 14) {
2146 rt2x00_set_field8(&rfcsr, RFCSR1_RX0_PD, 1); 2136 rt2x00_set_field8(&rfcsr, RFCSR1_RX0_PD, 1);
2147 rt2x00_set_field8(&rfcsr, RFCSR1_TX0_PD, 1); 2137 rt2x00_set_field8(&rfcsr, RFCSR1_TX0_PD, 1);
@@ -2674,7 +2664,7 @@ static void rt2800_config_channel_rf53xx(struct rt2x00_dev *rt2x00dev,
2674 if (rf->channel <= 14) { 2664 if (rf->channel <= 14) {
2675 int idx = rf->channel-1; 2665 int idx = rf->channel-1;
2676 2666
2677 if (test_bit(CAPABILITY_BT_COEXIST, &rt2x00dev->cap_flags)) { 2667 if (rt2x00_has_cap_bt_coexist(rt2x00dev)) {
2678 if (rt2x00_rt_rev_gte(rt2x00dev, RT5390, REV_RT5390F)) { 2668 if (rt2x00_rt_rev_gte(rt2x00dev, RT5390, REV_RT5390F)) {
2679 /* r55/r59 value array of channel 1~14 */ 2669 /* r55/r59 value array of channel 1~14 */
2680 static const char r55_bt_rev[] = {0x83, 0x83, 2670 static const char r55_bt_rev[] = {0x83, 0x83,
@@ -3152,6 +3142,7 @@ static void rt2800_config_channel(struct rt2x00_dev *rt2x00dev,
3152 case RF3322: 3142 case RF3322:
3153 rt2800_config_channel_rf3322(rt2x00dev, conf, rf, info); 3143 rt2800_config_channel_rf3322(rt2x00dev, conf, rf, info);
3154 break; 3144 break;
3145 case RF3070:
3155 case RF5360: 3146 case RF5360:
3156 case RF5370: 3147 case RF5370:
3157 case RF5372: 3148 case RF5372:
@@ -3166,7 +3157,8 @@ static void rt2800_config_channel(struct rt2x00_dev *rt2x00dev,
3166 rt2800_config_channel_rf2xxx(rt2x00dev, conf, rf, info); 3157 rt2800_config_channel_rf2xxx(rt2x00dev, conf, rf, info);
3167 } 3158 }
3168 3159
3169 if (rt2x00_rf(rt2x00dev, RF3290) || 3160 if (rt2x00_rf(rt2x00dev, RF3070) ||
3161 rt2x00_rf(rt2x00dev, RF3290) ||
3170 rt2x00_rf(rt2x00dev, RF3322) || 3162 rt2x00_rf(rt2x00dev, RF3322) ||
3171 rt2x00_rf(rt2x00dev, RF5360) || 3163 rt2x00_rf(rt2x00dev, RF5360) ||
3172 rt2x00_rf(rt2x00dev, RF5370) || 3164 rt2x00_rf(rt2x00dev, RF5370) ||
@@ -3218,8 +3210,7 @@ static void rt2800_config_channel(struct rt2x00_dev *rt2x00dev,
3218 if (rf->channel <= 14) { 3210 if (rf->channel <= 14) {
3219 if (!rt2x00_rt(rt2x00dev, RT5390) && 3211 if (!rt2x00_rt(rt2x00dev, RT5390) &&
3220 !rt2x00_rt(rt2x00dev, RT5392)) { 3212 !rt2x00_rt(rt2x00dev, RT5392)) {
3221 if (test_bit(CAPABILITY_EXTERNAL_LNA_BG, 3213 if (rt2x00_has_cap_external_lna_bg(rt2x00dev)) {
3222 &rt2x00dev->cap_flags)) {
3223 rt2800_bbp_write(rt2x00dev, 82, 0x62); 3214 rt2800_bbp_write(rt2x00dev, 82, 0x62);
3224 rt2800_bbp_write(rt2x00dev, 75, 0x46); 3215 rt2800_bbp_write(rt2x00dev, 75, 0x46);
3225 } else { 3216 } else {
@@ -3244,7 +3235,7 @@ static void rt2800_config_channel(struct rt2x00_dev *rt2x00dev,
3244 if (rt2x00_rt(rt2x00dev, RT3593)) 3235 if (rt2x00_rt(rt2x00dev, RT3593))
3245 rt2800_bbp_write(rt2x00dev, 83, 0x9a); 3236 rt2800_bbp_write(rt2x00dev, 83, 0x9a);
3246 3237
3247 if (test_bit(CAPABILITY_EXTERNAL_LNA_A, &rt2x00dev->cap_flags)) 3238 if (rt2x00_has_cap_external_lna_a(rt2x00dev))
3248 rt2800_bbp_write(rt2x00dev, 75, 0x46); 3239 rt2800_bbp_write(rt2x00dev, 75, 0x46);
3249 else 3240 else
3250 rt2800_bbp_write(rt2x00dev, 75, 0x50); 3241 rt2800_bbp_write(rt2x00dev, 75, 0x50);
@@ -3280,7 +3271,7 @@ static void rt2800_config_channel(struct rt2x00_dev *rt2x00dev,
3280 /* Turn on primary PAs */ 3271 /* Turn on primary PAs */
3281 rt2x00_set_field32(&tx_pin, TX_PIN_CFG_PA_PE_A0_EN, 3272 rt2x00_set_field32(&tx_pin, TX_PIN_CFG_PA_PE_A0_EN,
3282 rf->channel > 14); 3273 rf->channel > 14);
3283 if (test_bit(CAPABILITY_BT_COEXIST, &rt2x00dev->cap_flags)) 3274 if (rt2x00_has_cap_bt_coexist(rt2x00dev))
3284 rt2x00_set_field32(&tx_pin, TX_PIN_CFG_PA_PE_G0_EN, 1); 3275 rt2x00_set_field32(&tx_pin, TX_PIN_CFG_PA_PE_G0_EN, 1);
3285 else 3276 else
3286 rt2x00_set_field32(&tx_pin, TX_PIN_CFG_PA_PE_G0_EN, 3277 rt2x00_set_field32(&tx_pin, TX_PIN_CFG_PA_PE_G0_EN,
@@ -3311,33 +3302,50 @@ static void rt2800_config_channel(struct rt2x00_dev *rt2x00dev,
3311 3302
3312 rt2800_register_write(rt2x00dev, TX_PIN_CFG, tx_pin); 3303 rt2800_register_write(rt2x00dev, TX_PIN_CFG, tx_pin);
3313 3304
3314 if (rt2x00_rt(rt2x00dev, RT3572)) 3305 if (rt2x00_rt(rt2x00dev, RT3572)) {
3315 rt2800_rfcsr_write(rt2x00dev, 8, 0x80); 3306 rt2800_rfcsr_write(rt2x00dev, 8, 0x80);
3316 3307
3308 /* AGC init */
3309 if (rf->channel <= 14)
3310 reg = 0x1c + (2 * rt2x00dev->lna_gain);
3311 else
3312 reg = 0x22 + ((rt2x00dev->lna_gain * 5) / 3);
3313
3314 rt2800_bbp_write_with_rx_chain(rt2x00dev, 66, reg);
3315 }
3316
3317 if (rt2x00_rt(rt2x00dev, RT3593)) { 3317 if (rt2x00_rt(rt2x00dev, RT3593)) {
3318 if (rt2x00_is_usb(rt2x00dev)) { 3318 rt2800_register_read(rt2x00dev, GPIO_CTRL, &reg);
3319 rt2800_register_read(rt2x00dev, GPIO_CTRL, &reg);
3320 3319
3321 /* Band selection. GPIO #8 controls all paths */ 3320 /* Band selection */
3321 if (rt2x00_is_usb(rt2x00dev) ||
3322 rt2x00_is_pcie(rt2x00dev)) {
3323 /* GPIO #8 controls all paths */
3322 rt2x00_set_field32(&reg, GPIO_CTRL_DIR8, 0); 3324 rt2x00_set_field32(&reg, GPIO_CTRL_DIR8, 0);
3323 if (rf->channel <= 14) 3325 if (rf->channel <= 14)
3324 rt2x00_set_field32(&reg, GPIO_CTRL_VAL8, 1); 3326 rt2x00_set_field32(&reg, GPIO_CTRL_VAL8, 1);
3325 else 3327 else
3326 rt2x00_set_field32(&reg, GPIO_CTRL_VAL8, 0); 3328 rt2x00_set_field32(&reg, GPIO_CTRL_VAL8, 0);
3329 }
3327 3330
3331 /* LNA PE control. */
3332 if (rt2x00_is_usb(rt2x00dev)) {
3333 /* GPIO #4 controls PE0 and PE1,
3334 * GPIO #7 controls PE2
3335 */
3328 rt2x00_set_field32(&reg, GPIO_CTRL_DIR4, 0); 3336 rt2x00_set_field32(&reg, GPIO_CTRL_DIR4, 0);
3329 rt2x00_set_field32(&reg, GPIO_CTRL_DIR7, 0); 3337 rt2x00_set_field32(&reg, GPIO_CTRL_DIR7, 0);
3330 3338
3331 /* LNA PE control.
3332 * GPIO #4 controls PE0 and PE1,
3333 * GPIO #7 controls PE2
3334 */
3335 rt2x00_set_field32(&reg, GPIO_CTRL_VAL4, 1); 3339 rt2x00_set_field32(&reg, GPIO_CTRL_VAL4, 1);
3336 rt2x00_set_field32(&reg, GPIO_CTRL_VAL7, 1); 3340 rt2x00_set_field32(&reg, GPIO_CTRL_VAL7, 1);
3337 3341 } else if (rt2x00_is_pcie(rt2x00dev)) {
3338 rt2800_register_write(rt2x00dev, GPIO_CTRL, reg); 3342 /* GPIO #4 controls PE0, PE1 and PE2 */
3343 rt2x00_set_field32(&reg, GPIO_CTRL_DIR4, 0);
3344 rt2x00_set_field32(&reg, GPIO_CTRL_VAL4, 1);
3339 } 3345 }
3340 3346
3347 rt2800_register_write(rt2x00dev, GPIO_CTRL, reg);
3348
3341 /* AGC init */ 3349 /* AGC init */
3342 if (rf->channel <= 14) 3350 if (rf->channel <= 14)
3343 reg = 0x1c + 2 * rt2x00dev->lna_gain; 3351 reg = 0x1c + 2 * rt2x00dev->lna_gain;
@@ -3565,7 +3573,7 @@ static int rt2800_get_txpower_reg_delta(struct rt2x00_dev *rt2x00dev,
3565{ 3573{
3566 int delta; 3574 int delta;
3567 3575
3568 if (test_bit(CAPABILITY_POWER_LIMIT, &rt2x00dev->cap_flags)) 3576 if (rt2x00_has_cap_power_limit(rt2x00dev))
3569 return 0; 3577 return 0;
3570 3578
3571 /* 3579 /*
@@ -3594,7 +3602,7 @@ static u8 rt2800_compensate_txpower(struct rt2x00_dev *rt2x00dev, int is_rate_b,
3594 if (rt2x00_rt(rt2x00dev, RT3593)) 3602 if (rt2x00_rt(rt2x00dev, RT3593))
3595 return min_t(u8, txpower, 0xc); 3603 return min_t(u8, txpower, 0xc);
3596 3604
3597 if (test_bit(CAPABILITY_POWER_LIMIT, &rt2x00dev->cap_flags)) { 3605 if (rt2x00_has_cap_power_limit(rt2x00dev)) {
3598 /* 3606 /*
3599 * Check if eirp txpower exceed txpower_limit. 3607 * Check if eirp txpower exceed txpower_limit.
3600 * We use OFDM 6M as criterion and its eirp txpower 3608 * We use OFDM 6M as criterion and its eirp txpower
@@ -4264,6 +4272,7 @@ void rt2800_vco_calibration(struct rt2x00_dev *rt2x00dev)
4264 rt2800_rfcsr_write(rt2x00dev, 7, rfcsr); 4272 rt2800_rfcsr_write(rt2x00dev, 7, rfcsr);
4265 break; 4273 break;
4266 case RF3053: 4274 case RF3053:
4275 case RF3070:
4267 case RF3290: 4276 case RF3290:
4268 case RF5360: 4277 case RF5360:
4269 case RF5370: 4278 case RF5370:
@@ -4405,6 +4414,7 @@ static u8 rt2800_get_default_vgc(struct rt2x00_dev *rt2x00dev)
4405 rt2x00_rt(rt2x00dev, RT3290) || 4414 rt2x00_rt(rt2x00dev, RT3290) ||
4406 rt2x00_rt(rt2x00dev, RT3390) || 4415 rt2x00_rt(rt2x00dev, RT3390) ||
4407 rt2x00_rt(rt2x00dev, RT3572) || 4416 rt2x00_rt(rt2x00dev, RT3572) ||
4417 rt2x00_rt(rt2x00dev, RT3593) ||
4408 rt2x00_rt(rt2x00dev, RT5390) || 4418 rt2x00_rt(rt2x00dev, RT5390) ||
4409 rt2x00_rt(rt2x00dev, RT5392) || 4419 rt2x00_rt(rt2x00dev, RT5392) ||
4410 rt2x00_rt(rt2x00dev, RT5592)) 4420 rt2x00_rt(rt2x00dev, RT5592))
@@ -4412,8 +4422,8 @@ static u8 rt2800_get_default_vgc(struct rt2x00_dev *rt2x00dev)
4412 else 4422 else
4413 vgc = 0x2e + rt2x00dev->lna_gain; 4423 vgc = 0x2e + rt2x00dev->lna_gain;
4414 } else { /* 5GHZ band */ 4424 } else { /* 5GHZ band */
4415 if (rt2x00_rt(rt2x00dev, RT3572)) 4425 if (rt2x00_rt(rt2x00dev, RT3593))
4416 vgc = 0x22 + (rt2x00dev->lna_gain * 5) / 3; 4426 vgc = 0x20 + (rt2x00dev->lna_gain * 5) / 3;
4417 else if (rt2x00_rt(rt2x00dev, RT5592)) 4427 else if (rt2x00_rt(rt2x00dev, RT5592))
4418 vgc = 0x24 + (2 * rt2x00dev->lna_gain); 4428 vgc = 0x24 + (2 * rt2x00dev->lna_gain);
4419 else { 4429 else {
@@ -4431,11 +4441,17 @@ static inline void rt2800_set_vgc(struct rt2x00_dev *rt2x00dev,
4431 struct link_qual *qual, u8 vgc_level) 4441 struct link_qual *qual, u8 vgc_level)
4432{ 4442{
4433 if (qual->vgc_level != vgc_level) { 4443 if (qual->vgc_level != vgc_level) {
4434 if (rt2x00_rt(rt2x00dev, RT5592)) { 4444 if (rt2x00_rt(rt2x00dev, RT3572) ||
4445 rt2x00_rt(rt2x00dev, RT3593)) {
4446 rt2800_bbp_write_with_rx_chain(rt2x00dev, 66,
4447 vgc_level);
4448 } else if (rt2x00_rt(rt2x00dev, RT5592)) {
4435 rt2800_bbp_write(rt2x00dev, 83, qual->rssi > -65 ? 0x4a : 0x7a); 4449 rt2800_bbp_write(rt2x00dev, 83, qual->rssi > -65 ? 0x4a : 0x7a);
4436 rt2800_bbp_write_with_rx_chain(rt2x00dev, 66, vgc_level); 4450 rt2800_bbp_write_with_rx_chain(rt2x00dev, 66, vgc_level);
4437 } else 4451 } else {
4438 rt2800_bbp_write(rt2x00dev, 66, vgc_level); 4452 rt2800_bbp_write(rt2x00dev, 66, vgc_level);
4453 }
4454
4439 qual->vgc_level = vgc_level; 4455 qual->vgc_level = vgc_level;
4440 qual->vgc_level_reg = vgc_level; 4456 qual->vgc_level_reg = vgc_level;
4441 } 4457 }
@@ -4454,17 +4470,35 @@ void rt2800_link_tuner(struct rt2x00_dev *rt2x00dev, struct link_qual *qual,
4454 4470
4455 if (rt2x00_rt_rev(rt2x00dev, RT2860, REV_RT2860C)) 4471 if (rt2x00_rt_rev(rt2x00dev, RT2860, REV_RT2860C))
4456 return; 4472 return;
4457 /* 4473
4458 * When RSSI is better then -80 increase VGC level with 0x10, except 4474 /* When RSSI is better than a certain threshold, increase VGC
4459 * for rt5592 chip. 4475 * with a chip specific value in order to improve the balance
4476 * between sensibility and noise isolation.
4460 */ 4477 */
4461 4478
4462 vgc = rt2800_get_default_vgc(rt2x00dev); 4479 vgc = rt2800_get_default_vgc(rt2x00dev);
4463 4480
4464 if (rt2x00_rt(rt2x00dev, RT5592) && qual->rssi > -65) 4481 switch (rt2x00dev->chip.rt) {
4465 vgc += 0x20; 4482 case RT3572:
4466 else if (qual->rssi > -80) 4483 case RT3593:
4467 vgc += 0x10; 4484 if (qual->rssi > -65) {
4485 if (rt2x00dev->curr_band == IEEE80211_BAND_2GHZ)
4486 vgc += 0x20;
4487 else
4488 vgc += 0x10;
4489 }
4490 break;
4491
4492 case RT5592:
4493 if (qual->rssi > -65)
4494 vgc += 0x20;
4495 break;
4496
4497 default:
4498 if (qual->rssi > -80)
4499 vgc += 0x10;
4500 break;
4501 }
4468 4502
4469 rt2800_set_vgc(rt2x00dev, qual, vgc); 4503 rt2800_set_vgc(rt2x00dev, qual, vgc);
4470} 4504}
@@ -5489,7 +5523,7 @@ static void rt2800_init_bbp_53xx(struct rt2x00_dev *rt2x00dev)
5489 ant = (div_mode == 3) ? 1 : 0; 5523 ant = (div_mode == 3) ? 1 : 0;
5490 5524
5491 /* check if this is a Bluetooth combo card */ 5525 /* check if this is a Bluetooth combo card */
5492 if (test_bit(CAPABILITY_BT_COEXIST, &rt2x00dev->cap_flags)) { 5526 if (rt2x00_has_cap_bt_coexist(rt2x00dev)) {
5493 u32 reg; 5527 u32 reg;
5494 5528
5495 rt2800_register_read(rt2x00dev, GPIO_CTRL, &reg); 5529 rt2800_register_read(rt2x00dev, GPIO_CTRL, &reg);
@@ -5798,7 +5832,7 @@ static void rt2800_normal_mode_setup_3xxx(struct rt2x00_dev *rt2x00dev)
5798 rt2x00_rt_rev_lt(rt2x00dev, RT3071, REV_RT3071E) || 5832 rt2x00_rt_rev_lt(rt2x00dev, RT3071, REV_RT3071E) ||
5799 rt2x00_rt_rev_lt(rt2x00dev, RT3090, REV_RT3090E) || 5833 rt2x00_rt_rev_lt(rt2x00dev, RT3090, REV_RT3090E) ||
5800 rt2x00_rt_rev_lt(rt2x00dev, RT3390, REV_RT3390E)) { 5834 rt2x00_rt_rev_lt(rt2x00dev, RT3390, REV_RT3390E)) {
5801 if (!test_bit(CAPABILITY_EXTERNAL_LNA_BG, &rt2x00dev->cap_flags)) 5835 if (!rt2x00_has_cap_external_lna_bg(rt2x00dev))
5802 rt2x00_set_field8(&rfcsr, RFCSR17_R, 1); 5836 rt2x00_set_field8(&rfcsr, RFCSR17_R, 1);
5803 } 5837 }
5804 5838
@@ -5985,7 +6019,7 @@ static void rt2800_init_rfcsr_30xx(struct rt2x00_dev *rt2x00dev)
5985 rt2800_rfcsr_write(rt2x00dev, 20, 0xba); 6019 rt2800_rfcsr_write(rt2x00dev, 20, 0xba);
5986 rt2800_rfcsr_write(rt2x00dev, 21, 0xdb); 6020 rt2800_rfcsr_write(rt2x00dev, 21, 0xdb);
5987 rt2800_rfcsr_write(rt2x00dev, 24, 0x16); 6021 rt2800_rfcsr_write(rt2x00dev, 24, 0x16);
5988 rt2800_rfcsr_write(rt2x00dev, 25, 0x01); 6022 rt2800_rfcsr_write(rt2x00dev, 25, 0x03);
5989 rt2800_rfcsr_write(rt2x00dev, 29, 0x1f); 6023 rt2800_rfcsr_write(rt2x00dev, 29, 0x1f);
5990 6024
5991 if (rt2x00_rt_rev_lt(rt2x00dev, RT3070, REV_RT3070F)) { 6025 if (rt2x00_rt_rev_lt(rt2x00dev, RT3070, REV_RT3070F)) {
@@ -6441,7 +6475,7 @@ static void rt2800_init_rfcsr_5390(struct rt2x00_dev *rt2x00dev)
6441 rt2800_rfcsr_write(rt2x00dev, 28, 0x00); 6475 rt2800_rfcsr_write(rt2x00dev, 28, 0x00);
6442 rt2800_rfcsr_write(rt2x00dev, 29, 0x10); 6476 rt2800_rfcsr_write(rt2x00dev, 29, 0x10);
6443 6477
6444 rt2800_rfcsr_write(rt2x00dev, 30, 0x00); 6478 rt2800_rfcsr_write(rt2x00dev, 30, 0x10);
6445 rt2800_rfcsr_write(rt2x00dev, 31, 0x80); 6479 rt2800_rfcsr_write(rt2x00dev, 31, 0x80);
6446 rt2800_rfcsr_write(rt2x00dev, 32, 0x80); 6480 rt2800_rfcsr_write(rt2x00dev, 32, 0x80);
6447 rt2800_rfcsr_write(rt2x00dev, 33, 0x00); 6481 rt2800_rfcsr_write(rt2x00dev, 33, 0x00);
@@ -6479,7 +6513,7 @@ static void rt2800_init_rfcsr_5390(struct rt2x00_dev *rt2x00dev)
6479 rt2800_rfcsr_write(rt2x00dev, 56, 0x22); 6513 rt2800_rfcsr_write(rt2x00dev, 56, 0x22);
6480 rt2800_rfcsr_write(rt2x00dev, 57, 0x80); 6514 rt2800_rfcsr_write(rt2x00dev, 57, 0x80);
6481 rt2800_rfcsr_write(rt2x00dev, 58, 0x7f); 6515 rt2800_rfcsr_write(rt2x00dev, 58, 0x7f);
6482 rt2800_rfcsr_write(rt2x00dev, 59, 0x63); 6516 rt2800_rfcsr_write(rt2x00dev, 59, 0x8f);
6483 6517
6484 rt2800_rfcsr_write(rt2x00dev, 60, 0x45); 6518 rt2800_rfcsr_write(rt2x00dev, 60, 0x45);
6485 if (rt2x00_rt_rev_gte(rt2x00dev, RT5390, REV_RT5390F)) 6519 if (rt2x00_rt_rev_gte(rt2x00dev, RT5390, REV_RT5390F))
@@ -6499,7 +6533,6 @@ static void rt2800_init_rfcsr_5392(struct rt2x00_dev *rt2x00dev)
6499 rt2800_rf_init_calibration(rt2x00dev, 2); 6533 rt2800_rf_init_calibration(rt2x00dev, 2);
6500 6534
6501 rt2800_rfcsr_write(rt2x00dev, 1, 0x17); 6535 rt2800_rfcsr_write(rt2x00dev, 1, 0x17);
6502 rt2800_rfcsr_write(rt2x00dev, 2, 0x80);
6503 rt2800_rfcsr_write(rt2x00dev, 3, 0x88); 6536 rt2800_rfcsr_write(rt2x00dev, 3, 0x88);
6504 rt2800_rfcsr_write(rt2x00dev, 5, 0x10); 6537 rt2800_rfcsr_write(rt2x00dev, 5, 0x10);
6505 rt2800_rfcsr_write(rt2x00dev, 6, 0xe0); 6538 rt2800_rfcsr_write(rt2x00dev, 6, 0xe0);
@@ -6653,17 +6686,20 @@ int rt2800_enable_radio(struct rt2x00_dev *rt2x00dev)
6653 u16 word; 6686 u16 word;
6654 6687
6655 /* 6688 /*
6656 * Initialize all registers. 6689 * Initialize MAC registers.
6657 */ 6690 */
6658 if (unlikely(rt2800_wait_wpdma_ready(rt2x00dev) || 6691 if (unlikely(rt2800_wait_wpdma_ready(rt2x00dev) ||
6659 rt2800_init_registers(rt2x00dev))) 6692 rt2800_init_registers(rt2x00dev)))
6660 return -EIO; 6693 return -EIO;
6661 6694
6695 /*
6696 * Wait BBP/RF to wake up.
6697 */
6662 if (unlikely(rt2800_wait_bbp_rf_ready(rt2x00dev))) 6698 if (unlikely(rt2800_wait_bbp_rf_ready(rt2x00dev)))
6663 return -EIO; 6699 return -EIO;
6664 6700
6665 /* 6701 /*
6666 * Send signal to firmware during boot time. 6702 * Send signal during boot time to initialize firmware.
6667 */ 6703 */
6668 rt2800_register_write(rt2x00dev, H2M_BBP_AGENT, 0); 6704 rt2800_register_write(rt2x00dev, H2M_BBP_AGENT, 0);
6669 rt2800_register_write(rt2x00dev, H2M_MAILBOX_CSR, 0); 6705 rt2800_register_write(rt2x00dev, H2M_MAILBOX_CSR, 0);
@@ -6672,9 +6708,15 @@ int rt2800_enable_radio(struct rt2x00_dev *rt2x00dev)
6672 rt2800_mcu_request(rt2x00dev, MCU_BOOT_SIGNAL, 0, 0, 0); 6708 rt2800_mcu_request(rt2x00dev, MCU_BOOT_SIGNAL, 0, 0, 0);
6673 msleep(1); 6709 msleep(1);
6674 6710
6711 /*
6712 * Make sure BBP is up and running.
6713 */
6675 if (unlikely(rt2800_wait_bbp_ready(rt2x00dev))) 6714 if (unlikely(rt2800_wait_bbp_ready(rt2x00dev)))
6676 return -EIO; 6715 return -EIO;
6677 6716
6717 /*
6718 * Initialize BBP/RF registers.
6719 */
6678 rt2800_init_bbp(rt2x00dev); 6720 rt2800_init_bbp(rt2x00dev);
6679 rt2800_init_rfcsr(rt2x00dev); 6721 rt2800_init_rfcsr(rt2x00dev);
6680 6722
@@ -7021,6 +7063,7 @@ static int rt2800_init_eeprom(struct rt2x00_dev *rt2x00dev)
7021 case RF3022: 7063 case RF3022:
7022 case RF3052: 7064 case RF3052:
7023 case RF3053: 7065 case RF3053:
7066 case RF3070:
7024 case RF3290: 7067 case RF3290:
7025 case RF3320: 7068 case RF3320:
7026 case RF3322: 7069 case RF3322:
@@ -7203,7 +7246,7 @@ static const struct rf_channel rf_vals[] = {
7203 7246
7204/* 7247/*
7205 * RF value list for rt3xxx 7248 * RF value list for rt3xxx
7206 * Supports: 2.4 GHz (all) & 5.2 GHz (RF3052) 7249 * Supports: 2.4 GHz (all) & 5.2 GHz (RF3052 & RF3053)
7207 */ 7250 */
7208static const struct rf_channel rf_vals_3x[] = { 7251static const struct rf_channel rf_vals_3x[] = {
7209 {1, 241, 2, 2 }, 7252 {1, 241, 2, 2 },
@@ -7399,72 +7442,6 @@ static const struct rf_channel rf_vals_5592_xtal40[] = {
7399 {196, 83, 0, 12, 1}, 7442 {196, 83, 0, 12, 1},
7400}; 7443};
7401 7444
7402static const struct rf_channel rf_vals_3053[] = {
7403 /* Channel, N, R, K */
7404 {1, 241, 2, 2},
7405 {2, 241, 2, 7},
7406 {3, 242, 2, 2},
7407 {4, 242, 2, 7},
7408 {5, 243, 2, 2},
7409 {6, 243, 2, 7},
7410 {7, 244, 2, 2},
7411 {8, 244, 2, 7},
7412 {9, 245, 2, 2},
7413 {10, 245, 2, 7},
7414 {11, 246, 2, 2},
7415 {12, 246, 2, 7},
7416 {13, 247, 2, 2},
7417 {14, 248, 2, 4},
7418
7419 {36, 0x56, 0, 4},
7420 {38, 0x56, 0, 6},
7421 {40, 0x56, 0, 8},
7422 {44, 0x57, 0, 0},
7423 {46, 0x57, 0, 2},
7424 {48, 0x57, 0, 4},
7425 {52, 0x57, 0, 8},
7426 {54, 0x57, 0, 10},
7427 {56, 0x58, 0, 0},
7428 {60, 0x58, 0, 4},
7429 {62, 0x58, 0, 6},
7430 {64, 0x58, 0, 8},
7431
7432 {100, 0x5B, 0, 8},
7433 {102, 0x5B, 0, 10},
7434 {104, 0x5C, 0, 0},
7435 {108, 0x5C, 0, 4},
7436 {110, 0x5C, 0, 6},
7437 {112, 0x5C, 0, 8},
7438
7439 /* NOTE: Channel 114 has been removed intentionally.
7440 * The EEPROM contains no TX power values for that,
7441 * and it is disabled in the vendor driver as well.
7442 */
7443
7444 {116, 0x5D, 0, 0},
7445 {118, 0x5D, 0, 2},
7446 {120, 0x5D, 0, 4},
7447 {124, 0x5D, 0, 8},
7448 {126, 0x5D, 0, 10},
7449 {128, 0x5E, 0, 0},
7450 {132, 0x5E, 0, 4},
7451 {134, 0x5E, 0, 6},
7452 {136, 0x5E, 0, 8},
7453 {140, 0x5F, 0, 0},
7454
7455 {149, 0x5F, 0, 9},
7456 {151, 0x5F, 0, 11},
7457 {153, 0x60, 0, 1},
7458 {157, 0x60, 0, 5},
7459 {159, 0x60, 0, 7},
7460 {161, 0x60, 0, 9},
7461 {165, 0x61, 0, 1},
7462 {167, 0x61, 0, 3},
7463 {169, 0x61, 0, 5},
7464 {171, 0x61, 0, 7},
7465 {173, 0x61, 0, 9},
7466};
7467
7468static int rt2800_probe_hw_mode(struct rt2x00_dev *rt2x00dev) 7445static int rt2800_probe_hw_mode(struct rt2x00_dev *rt2x00dev)
7469{ 7446{
7470 struct hw_mode_spec *spec = &rt2x00dev->spec; 7447 struct hw_mode_spec *spec = &rt2x00dev->spec;
@@ -7473,7 +7450,6 @@ static int rt2800_probe_hw_mode(struct rt2x00_dev *rt2x00dev)
7473 char *default_power2; 7450 char *default_power2;
7474 char *default_power3; 7451 char *default_power3;
7475 unsigned int i; 7452 unsigned int i;
7476 u16 eeprom;
7477 u32 reg; 7453 u32 reg;
7478 7454
7479 /* 7455 /*
@@ -7522,48 +7498,48 @@ static int rt2800_probe_hw_mode(struct rt2x00_dev *rt2x00dev)
7522 rt2x00dev->hw->max_report_rates = 7; 7498 rt2x00dev->hw->max_report_rates = 7;
7523 rt2x00dev->hw->max_rate_tries = 1; 7499 rt2x00dev->hw->max_rate_tries = 1;
7524 7500
7525 rt2800_eeprom_read(rt2x00dev, EEPROM_NIC_CONF0, &eeprom);
7526
7527 /* 7501 /*
7528 * Initialize hw_mode information. 7502 * Initialize hw_mode information.
7529 */ 7503 */
7530 spec->supported_bands = SUPPORT_BAND_2GHZ;
7531 spec->supported_rates = SUPPORT_RATE_CCK | SUPPORT_RATE_OFDM; 7504 spec->supported_rates = SUPPORT_RATE_CCK | SUPPORT_RATE_OFDM;
7532 7505
7533 if (rt2x00_rf(rt2x00dev, RF2820) || 7506 switch (rt2x00dev->chip.rf) {
7534 rt2x00_rf(rt2x00dev, RF2720)) { 7507 case RF2720:
7508 case RF2820:
7535 spec->num_channels = 14; 7509 spec->num_channels = 14;
7536 spec->channels = rf_vals; 7510 spec->channels = rf_vals;
7537 } else if (rt2x00_rf(rt2x00dev, RF2850) || 7511 break;
7538 rt2x00_rf(rt2x00dev, RF2750)) { 7512
7539 spec->supported_bands |= SUPPORT_BAND_5GHZ; 7513 case RF2750:
7514 case RF2850:
7540 spec->num_channels = ARRAY_SIZE(rf_vals); 7515 spec->num_channels = ARRAY_SIZE(rf_vals);
7541 spec->channels = rf_vals; 7516 spec->channels = rf_vals;
7542 } else if (rt2x00_rf(rt2x00dev, RF3020) || 7517 break;
7543 rt2x00_rf(rt2x00dev, RF2020) || 7518
7544 rt2x00_rf(rt2x00dev, RF3021) || 7519 case RF2020:
7545 rt2x00_rf(rt2x00dev, RF3022) || 7520 case RF3020:
7546 rt2x00_rf(rt2x00dev, RF3290) || 7521 case RF3021:
7547 rt2x00_rf(rt2x00dev, RF3320) || 7522 case RF3022:
7548 rt2x00_rf(rt2x00dev, RF3322) || 7523 case RF3070:
7549 rt2x00_rf(rt2x00dev, RF5360) || 7524 case RF3290:
7550 rt2x00_rf(rt2x00dev, RF5370) || 7525 case RF3320:
7551 rt2x00_rf(rt2x00dev, RF5372) || 7526 case RF3322:
7552 rt2x00_rf(rt2x00dev, RF5390) || 7527 case RF5360:
7553 rt2x00_rf(rt2x00dev, RF5392)) { 7528 case RF5370:
7529 case RF5372:
7530 case RF5390:
7531 case RF5392:
7554 spec->num_channels = 14; 7532 spec->num_channels = 14;
7555 spec->channels = rf_vals_3x; 7533 spec->channels = rf_vals_3x;
7556 } else if (rt2x00_rf(rt2x00dev, RF3052)) { 7534 break;
7557 spec->supported_bands |= SUPPORT_BAND_5GHZ; 7535
7536 case RF3052:
7537 case RF3053:
7558 spec->num_channels = ARRAY_SIZE(rf_vals_3x); 7538 spec->num_channels = ARRAY_SIZE(rf_vals_3x);
7559 spec->channels = rf_vals_3x; 7539 spec->channels = rf_vals_3x;
7560 } else if (rt2x00_rf(rt2x00dev, RF3053)) { 7540 break;
7561 spec->supported_bands |= SUPPORT_BAND_5GHZ;
7562 spec->num_channels = ARRAY_SIZE(rf_vals_3053);
7563 spec->channels = rf_vals_3053;
7564 } else if (rt2x00_rf(rt2x00dev, RF5592)) {
7565 spec->supported_bands |= SUPPORT_BAND_5GHZ;
7566 7541
7542 case RF5592:
7567 rt2800_register_read(rt2x00dev, MAC_DEBUG_INDEX, &reg); 7543 rt2800_register_read(rt2x00dev, MAC_DEBUG_INDEX, &reg);
7568 if (rt2x00_get_field32(reg, MAC_DEBUG_INDEX_XTAL)) { 7544 if (rt2x00_get_field32(reg, MAC_DEBUG_INDEX_XTAL)) {
7569 spec->num_channels = ARRAY_SIZE(rf_vals_5592_xtal40); 7545 spec->num_channels = ARRAY_SIZE(rf_vals_5592_xtal40);
@@ -7572,11 +7548,16 @@ static int rt2800_probe_hw_mode(struct rt2x00_dev *rt2x00dev)
7572 spec->num_channels = ARRAY_SIZE(rf_vals_5592_xtal20); 7548 spec->num_channels = ARRAY_SIZE(rf_vals_5592_xtal20);
7573 spec->channels = rf_vals_5592_xtal20; 7549 spec->channels = rf_vals_5592_xtal20;
7574 } 7550 }
7551 break;
7575 } 7552 }
7576 7553
7577 if (WARN_ON_ONCE(!spec->channels)) 7554 if (WARN_ON_ONCE(!spec->channels))
7578 return -ENODEV; 7555 return -ENODEV;
7579 7556
7557 spec->supported_bands = SUPPORT_BAND_2GHZ;
7558 if (spec->num_channels > 14)
7559 spec->supported_bands |= SUPPORT_BAND_5GHZ;
7560
7580 /* 7561 /*
7581 * Initialize HT information. 7562 * Initialize HT information.
7582 */ 7563 */
@@ -7591,22 +7572,21 @@ static int rt2800_probe_hw_mode(struct rt2x00_dev *rt2x00dev)
7591 IEEE80211_HT_CAP_SGI_20 | 7572 IEEE80211_HT_CAP_SGI_20 |
7592 IEEE80211_HT_CAP_SGI_40; 7573 IEEE80211_HT_CAP_SGI_40;
7593 7574
7594 if (rt2x00_get_field16(eeprom, EEPROM_NIC_CONF0_TXPATH) >= 2) 7575 if (rt2x00dev->default_ant.tx_chain_num >= 2)
7595 spec->ht.cap |= IEEE80211_HT_CAP_TX_STBC; 7576 spec->ht.cap |= IEEE80211_HT_CAP_TX_STBC;
7596 7577
7597 spec->ht.cap |= 7578 spec->ht.cap |= rt2x00dev->default_ant.rx_chain_num <<
7598 rt2x00_get_field16(eeprom, EEPROM_NIC_CONF0_RXPATH) << 7579 IEEE80211_HT_CAP_RX_STBC_SHIFT;
7599 IEEE80211_HT_CAP_RX_STBC_SHIFT;
7600 7580
7601 spec->ht.ampdu_factor = 3; 7581 spec->ht.ampdu_factor = 3;
7602 spec->ht.ampdu_density = 4; 7582 spec->ht.ampdu_density = 4;
7603 spec->ht.mcs.tx_params = 7583 spec->ht.mcs.tx_params =
7604 IEEE80211_HT_MCS_TX_DEFINED | 7584 IEEE80211_HT_MCS_TX_DEFINED |
7605 IEEE80211_HT_MCS_TX_RX_DIFF | 7585 IEEE80211_HT_MCS_TX_RX_DIFF |
7606 ((rt2x00_get_field16(eeprom, EEPROM_NIC_CONF0_TXPATH) - 1) << 7586 ((rt2x00dev->default_ant.tx_chain_num - 1) <<
7607 IEEE80211_HT_MCS_TX_MAX_STREAMS_SHIFT); 7587 IEEE80211_HT_MCS_TX_MAX_STREAMS_SHIFT);
7608 7588
7609 switch (rt2x00_get_field16(eeprom, EEPROM_NIC_CONF0_RXPATH)) { 7589 switch (rt2x00dev->default_ant.rx_chain_num) {
7610 case 3: 7590 case 3:
7611 spec->ht.mcs.rx_mask[2] = 0xff; 7591 spec->ht.mcs.rx_mask[2] = 0xff;
7612 case 2: 7592 case 2:
@@ -7671,6 +7651,7 @@ static int rt2800_probe_hw_mode(struct rt2x00_dev *rt2x00dev)
7671 case RF3320: 7651 case RF3320:
7672 case RF3052: 7652 case RF3052:
7673 case RF3053: 7653 case RF3053:
7654 case RF3070:
7674 case RF3290: 7655 case RF3290:
7675 case RF5360: 7656 case RF5360:
7676 case RF5370: 7657 case RF5370:
diff --git a/drivers/net/wireless/rt2x00/rt2800mmio.c b/drivers/net/wireless/rt2x00/rt2800mmio.c
new file mode 100644
index 000000000000..ae152280e071
--- /dev/null
+++ b/drivers/net/wireless/rt2x00/rt2800mmio.c
@@ -0,0 +1,873 @@
1/* Copyright (C) 2009 - 2010 Ivo van Doorn <IvDoorn@gmail.com>
2 * Copyright (C) 2009 Alban Browaeys <prahal@yahoo.com>
3 * Copyright (C) 2009 Felix Fietkau <nbd@openwrt.org>
4 * Copyright (C) 2009 Luis Correia <luis.f.correia@gmail.com>
5 * Copyright (C) 2009 Mattias Nissler <mattias.nissler@gmx.de>
6 * Copyright (C) 2009 Mark Asselstine <asselsm@gmail.com>
7 * Copyright (C) 2009 Xose Vazquez Perez <xose.vazquez@gmail.com>
8 * Copyright (C) 2009 Bart Zolnierkiewicz <bzolnier@gmail.com>
9 * <http://rt2x00.serialmonkey.com>
10 *
11 * This program is free software; you can redistribute it and/or modify
12 * it under the terms of the GNU General Public License as published by
13 * the Free Software Foundation; either version 2 of the License, or
14 * (at your option) any later version.
15 *
16 * This program is distributed in the hope that it will be useful,
17 * but WITHOUT ANY WARRANTY; without even the implied warranty of
18 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
19 * GNU General Public License for more details.
20 *
21 * You should have received a copy of the GNU General Public License
22 * along with this program; if not, write to the
23 * Free Software Foundation, Inc.,
24 * 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
25 */
26
27/* Module: rt2800mmio
28 * Abstract: rt2800 MMIO device routines.
29 */
30
31#include <linux/kernel.h>
32#include <linux/module.h>
33#include <linux/export.h>
34
35#include "rt2x00.h"
36#include "rt2x00mmio.h"
37#include "rt2800.h"
38#include "rt2800lib.h"
39#include "rt2800mmio.h"
40
41/*
42 * TX descriptor initialization
43 */
44__le32 *rt2800mmio_get_txwi(struct queue_entry *entry)
45{
46 return (__le32 *) entry->skb->data;
47}
48EXPORT_SYMBOL_GPL(rt2800mmio_get_txwi);
49
50void rt2800mmio_write_tx_desc(struct queue_entry *entry,
51 struct txentry_desc *txdesc)
52{
53 struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
54 struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
55 __le32 *txd = entry_priv->desc;
56 u32 word;
57 const unsigned int txwi_size = entry->queue->winfo_size;
58
59 /*
60 * The buffers pointed by SD_PTR0/SD_LEN0 and SD_PTR1/SD_LEN1
61 * must contains a TXWI structure + 802.11 header + padding + 802.11
62 * data. We choose to have SD_PTR0/SD_LEN0 only contains TXWI and
63 * SD_PTR1/SD_LEN1 contains 802.11 header + padding + 802.11
64 * data. It means that LAST_SEC0 is always 0.
65 */
66
67 /*
68 * Initialize TX descriptor
69 */
70 word = 0;
71 rt2x00_set_field32(&word, TXD_W0_SD_PTR0, skbdesc->skb_dma);
72 rt2x00_desc_write(txd, 0, word);
73
74 word = 0;
75 rt2x00_set_field32(&word, TXD_W1_SD_LEN1, entry->skb->len);
76 rt2x00_set_field32(&word, TXD_W1_LAST_SEC1,
77 !test_bit(ENTRY_TXD_MORE_FRAG, &txdesc->flags));
78 rt2x00_set_field32(&word, TXD_W1_BURST,
79 test_bit(ENTRY_TXD_BURST, &txdesc->flags));
80 rt2x00_set_field32(&word, TXD_W1_SD_LEN0, txwi_size);
81 rt2x00_set_field32(&word, TXD_W1_LAST_SEC0, 0);
82 rt2x00_set_field32(&word, TXD_W1_DMA_DONE, 0);
83 rt2x00_desc_write(txd, 1, word);
84
85 word = 0;
86 rt2x00_set_field32(&word, TXD_W2_SD_PTR1,
87 skbdesc->skb_dma + txwi_size);
88 rt2x00_desc_write(txd, 2, word);
89
90 word = 0;
91 rt2x00_set_field32(&word, TXD_W3_WIV,
92 !test_bit(ENTRY_TXD_ENCRYPT_IV, &txdesc->flags));
93 rt2x00_set_field32(&word, TXD_W3_QSEL, 2);
94 rt2x00_desc_write(txd, 3, word);
95
96 /*
97 * Register descriptor details in skb frame descriptor.
98 */
99 skbdesc->desc = txd;
100 skbdesc->desc_len = TXD_DESC_SIZE;
101}
102EXPORT_SYMBOL_GPL(rt2800mmio_write_tx_desc);
103
104/*
105 * RX control handlers
106 */
107void rt2800mmio_fill_rxdone(struct queue_entry *entry,
108 struct rxdone_entry_desc *rxdesc)
109{
110 struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
111 __le32 *rxd = entry_priv->desc;
112 u32 word;
113
114 rt2x00_desc_read(rxd, 3, &word);
115
116 if (rt2x00_get_field32(word, RXD_W3_CRC_ERROR))
117 rxdesc->flags |= RX_FLAG_FAILED_FCS_CRC;
118
119 /*
120 * Unfortunately we don't know the cipher type used during
121 * decryption. This prevents us from correct providing
122 * correct statistics through debugfs.
123 */
124 rxdesc->cipher_status = rt2x00_get_field32(word, RXD_W3_CIPHER_ERROR);
125
126 if (rt2x00_get_field32(word, RXD_W3_DECRYPTED)) {
127 /*
128 * Hardware has stripped IV/EIV data from 802.11 frame during
129 * decryption. Unfortunately the descriptor doesn't contain
130 * any fields with the EIV/IV data either, so they can't
131 * be restored by rt2x00lib.
132 */
133 rxdesc->flags |= RX_FLAG_IV_STRIPPED;
134
135 /*
136 * The hardware has already checked the Michael Mic and has
137 * stripped it from the frame. Signal this to mac80211.
138 */
139 rxdesc->flags |= RX_FLAG_MMIC_STRIPPED;
140
141 if (rxdesc->cipher_status == RX_CRYPTO_SUCCESS)
142 rxdesc->flags |= RX_FLAG_DECRYPTED;
143 else if (rxdesc->cipher_status == RX_CRYPTO_FAIL_MIC)
144 rxdesc->flags |= RX_FLAG_MMIC_ERROR;
145 }
146
147 if (rt2x00_get_field32(word, RXD_W3_MY_BSS))
148 rxdesc->dev_flags |= RXDONE_MY_BSS;
149
150 if (rt2x00_get_field32(word, RXD_W3_L2PAD))
151 rxdesc->dev_flags |= RXDONE_L2PAD;
152
153 /*
154 * Process the RXWI structure that is at the start of the buffer.
155 */
156 rt2800_process_rxwi(entry, rxdesc);
157}
158EXPORT_SYMBOL_GPL(rt2800mmio_fill_rxdone);
159
160/*
161 * Interrupt functions.
162 */
163static void rt2800mmio_wakeup(struct rt2x00_dev *rt2x00dev)
164{
165 struct ieee80211_conf conf = { .flags = 0 };
166 struct rt2x00lib_conf libconf = { .conf = &conf };
167
168 rt2800_config(rt2x00dev, &libconf, IEEE80211_CONF_CHANGE_PS);
169}
170
171static bool rt2800mmio_txdone_entry_check(struct queue_entry *entry, u32 status)
172{
173 __le32 *txwi;
174 u32 word;
175 int wcid, tx_wcid;
176
177 wcid = rt2x00_get_field32(status, TX_STA_FIFO_WCID);
178
179 txwi = rt2800_drv_get_txwi(entry);
180 rt2x00_desc_read(txwi, 1, &word);
181 tx_wcid = rt2x00_get_field32(word, TXWI_W1_WIRELESS_CLI_ID);
182
183 return (tx_wcid == wcid);
184}
185
186static bool rt2800mmio_txdone_find_entry(struct queue_entry *entry, void *data)
187{
188 u32 status = *(u32 *)data;
189
190 /*
191 * rt2800pci hardware might reorder frames when exchanging traffic
192 * with multiple BA enabled STAs.
193 *
194 * For example, a tx queue
195 * [ STA1 | STA2 | STA1 | STA2 ]
196 * can result in tx status reports
197 * [ STA1 | STA1 | STA2 | STA2 ]
198 * when the hw decides to aggregate the frames for STA1 into one AMPDU.
199 *
200 * To mitigate this effect, associate the tx status to the first frame
201 * in the tx queue with a matching wcid.
202 */
203 if (rt2800mmio_txdone_entry_check(entry, status) &&
204 !test_bit(ENTRY_DATA_STATUS_SET, &entry->flags)) {
205 /*
206 * Got a matching frame, associate the tx status with
207 * the frame
208 */
209 entry->status = status;
210 set_bit(ENTRY_DATA_STATUS_SET, &entry->flags);
211 return true;
212 }
213
214 /* Check the next frame */
215 return false;
216}
217
218static bool rt2800mmio_txdone_match_first(struct queue_entry *entry, void *data)
219{
220 u32 status = *(u32 *)data;
221
222 /*
223 * Find the first frame without tx status and assign this status to it
224 * regardless if it matches or not.
225 */
226 if (!test_bit(ENTRY_DATA_STATUS_SET, &entry->flags)) {
227 /*
228 * Got a matching frame, associate the tx status with
229 * the frame
230 */
231 entry->status = status;
232 set_bit(ENTRY_DATA_STATUS_SET, &entry->flags);
233 return true;
234 }
235
236 /* Check the next frame */
237 return false;
238}
239static bool rt2800mmio_txdone_release_entries(struct queue_entry *entry,
240 void *data)
241{
242 if (test_bit(ENTRY_DATA_STATUS_SET, &entry->flags)) {
243 rt2800_txdone_entry(entry, entry->status,
244 rt2800mmio_get_txwi(entry));
245 return false;
246 }
247
248 /* No more frames to release */
249 return true;
250}
251
252static bool rt2800mmio_txdone(struct rt2x00_dev *rt2x00dev)
253{
254 struct data_queue *queue;
255 u32 status;
256 u8 qid;
257 int max_tx_done = 16;
258
259 while (kfifo_get(&rt2x00dev->txstatus_fifo, &status)) {
260 qid = rt2x00_get_field32(status, TX_STA_FIFO_PID_QUEUE);
261 if (unlikely(qid >= QID_RX)) {
262 /*
263 * Unknown queue, this shouldn't happen. Just drop
264 * this tx status.
265 */
266 rt2x00_warn(rt2x00dev, "Got TX status report with unexpected pid %u, dropping\n",
267 qid);
268 break;
269 }
270
271 queue = rt2x00queue_get_tx_queue(rt2x00dev, qid);
272 if (unlikely(queue == NULL)) {
273 /*
274 * The queue is NULL, this shouldn't happen. Stop
275 * processing here and drop the tx status
276 */
277 rt2x00_warn(rt2x00dev, "Got TX status for an unavailable queue %u, dropping\n",
278 qid);
279 break;
280 }
281
282 if (unlikely(rt2x00queue_empty(queue))) {
283 /*
284 * The queue is empty. Stop processing here
285 * and drop the tx status.
286 */
287 rt2x00_warn(rt2x00dev, "Got TX status for an empty queue %u, dropping\n",
288 qid);
289 break;
290 }
291
292 /*
293 * Let's associate this tx status with the first
294 * matching frame.
295 */
296 if (!rt2x00queue_for_each_entry(queue, Q_INDEX_DONE,
297 Q_INDEX, &status,
298 rt2800mmio_txdone_find_entry)) {
299 /*
300 * We cannot match the tx status to any frame, so just
301 * use the first one.
302 */
303 if (!rt2x00queue_for_each_entry(queue, Q_INDEX_DONE,
304 Q_INDEX, &status,
305 rt2800mmio_txdone_match_first)) {
306 rt2x00_warn(rt2x00dev, "No frame found for TX status on queue %u, dropping\n",
307 qid);
308 break;
309 }
310 }
311
312 /*
313 * Release all frames with a valid tx status.
314 */
315 rt2x00queue_for_each_entry(queue, Q_INDEX_DONE,
316 Q_INDEX, NULL,
317 rt2800mmio_txdone_release_entries);
318
319 if (--max_tx_done == 0)
320 break;
321 }
322
323 return !max_tx_done;
324}
325
326static inline void rt2800mmio_enable_interrupt(struct rt2x00_dev *rt2x00dev,
327 struct rt2x00_field32 irq_field)
328{
329 u32 reg;
330
331 /*
332 * Enable a single interrupt. The interrupt mask register
333 * access needs locking.
334 */
335 spin_lock_irq(&rt2x00dev->irqmask_lock);
336 rt2x00mmio_register_read(rt2x00dev, INT_MASK_CSR, &reg);
337 rt2x00_set_field32(&reg, irq_field, 1);
338 rt2x00mmio_register_write(rt2x00dev, INT_MASK_CSR, reg);
339 spin_unlock_irq(&rt2x00dev->irqmask_lock);
340}
341
342void rt2800mmio_txstatus_tasklet(unsigned long data)
343{
344 struct rt2x00_dev *rt2x00dev = (struct rt2x00_dev *)data;
345 if (rt2800mmio_txdone(rt2x00dev))
346 tasklet_schedule(&rt2x00dev->txstatus_tasklet);
347
348 /*
349 * No need to enable the tx status interrupt here as we always
350 * leave it enabled to minimize the possibility of a tx status
351 * register overflow. See comment in interrupt handler.
352 */
353}
354EXPORT_SYMBOL_GPL(rt2800mmio_txstatus_tasklet);
355
356void rt2800mmio_pretbtt_tasklet(unsigned long data)
357{
358 struct rt2x00_dev *rt2x00dev = (struct rt2x00_dev *)data;
359 rt2x00lib_pretbtt(rt2x00dev);
360 if (test_bit(DEVICE_STATE_ENABLED_RADIO, &rt2x00dev->flags))
361 rt2800mmio_enable_interrupt(rt2x00dev, INT_MASK_CSR_PRE_TBTT);
362}
363EXPORT_SYMBOL_GPL(rt2800mmio_pretbtt_tasklet);
364
365void rt2800mmio_tbtt_tasklet(unsigned long data)
366{
367 struct rt2x00_dev *rt2x00dev = (struct rt2x00_dev *)data;
368 struct rt2800_drv_data *drv_data = rt2x00dev->drv_data;
369 u32 reg;
370
371 rt2x00lib_beacondone(rt2x00dev);
372
373 if (rt2x00dev->intf_ap_count) {
374 /*
375 * The rt2800pci hardware tbtt timer is off by 1us per tbtt
376 * causing beacon skew and as a result causing problems with
377 * some powersaving clients over time. Shorten the beacon
378 * interval every 64 beacons by 64us to mitigate this effect.
379 */
380 if (drv_data->tbtt_tick == (BCN_TBTT_OFFSET - 2)) {
381 rt2x00mmio_register_read(rt2x00dev, BCN_TIME_CFG, &reg);
382 rt2x00_set_field32(&reg, BCN_TIME_CFG_BEACON_INTERVAL,
383 (rt2x00dev->beacon_int * 16) - 1);
384 rt2x00mmio_register_write(rt2x00dev, BCN_TIME_CFG, reg);
385 } else if (drv_data->tbtt_tick == (BCN_TBTT_OFFSET - 1)) {
386 rt2x00mmio_register_read(rt2x00dev, BCN_TIME_CFG, &reg);
387 rt2x00_set_field32(&reg, BCN_TIME_CFG_BEACON_INTERVAL,
388 (rt2x00dev->beacon_int * 16));
389 rt2x00mmio_register_write(rt2x00dev, BCN_TIME_CFG, reg);
390 }
391 drv_data->tbtt_tick++;
392 drv_data->tbtt_tick %= BCN_TBTT_OFFSET;
393 }
394
395 if (test_bit(DEVICE_STATE_ENABLED_RADIO, &rt2x00dev->flags))
396 rt2800mmio_enable_interrupt(rt2x00dev, INT_MASK_CSR_TBTT);
397}
398EXPORT_SYMBOL_GPL(rt2800mmio_tbtt_tasklet);
399
400void rt2800mmio_rxdone_tasklet(unsigned long data)
401{
402 struct rt2x00_dev *rt2x00dev = (struct rt2x00_dev *)data;
403 if (rt2x00mmio_rxdone(rt2x00dev))
404 tasklet_schedule(&rt2x00dev->rxdone_tasklet);
405 else if (test_bit(DEVICE_STATE_ENABLED_RADIO, &rt2x00dev->flags))
406 rt2800mmio_enable_interrupt(rt2x00dev, INT_MASK_CSR_RX_DONE);
407}
408EXPORT_SYMBOL_GPL(rt2800mmio_rxdone_tasklet);
409
410void rt2800mmio_autowake_tasklet(unsigned long data)
411{
412 struct rt2x00_dev *rt2x00dev = (struct rt2x00_dev *)data;
413 rt2800mmio_wakeup(rt2x00dev);
414 if (test_bit(DEVICE_STATE_ENABLED_RADIO, &rt2x00dev->flags))
415 rt2800mmio_enable_interrupt(rt2x00dev,
416 INT_MASK_CSR_AUTO_WAKEUP);
417}
418EXPORT_SYMBOL_GPL(rt2800mmio_autowake_tasklet);
419
420static void rt2800mmio_txstatus_interrupt(struct rt2x00_dev *rt2x00dev)
421{
422 u32 status;
423 int i;
424
425 /*
426 * The TX_FIFO_STATUS interrupt needs special care. We should
427 * read TX_STA_FIFO but we should do it immediately as otherwise
428 * the register can overflow and we would lose status reports.
429 *
430 * Hence, read the TX_STA_FIFO register and copy all tx status
431 * reports into a kernel FIFO which is handled in the txstatus
432 * tasklet. We use a tasklet to process the tx status reports
433 * because we can schedule the tasklet multiple times (when the
434 * interrupt fires again during tx status processing).
435 *
436 * Furthermore we don't disable the TX_FIFO_STATUS
437 * interrupt here but leave it enabled so that the TX_STA_FIFO
438 * can also be read while the tx status tasklet gets executed.
439 *
440 * Since we have only one producer and one consumer we don't
441 * need to lock the kfifo.
442 */
443 for (i = 0; i < rt2x00dev->tx->limit; i++) {
444 rt2x00mmio_register_read(rt2x00dev, TX_STA_FIFO, &status);
445
446 if (!rt2x00_get_field32(status, TX_STA_FIFO_VALID))
447 break;
448
449 if (!kfifo_put(&rt2x00dev->txstatus_fifo, &status)) {
450 rt2x00_warn(rt2x00dev, "TX status FIFO overrun, drop tx status report\n");
451 break;
452 }
453 }
454
455 /* Schedule the tasklet for processing the tx status. */
456 tasklet_schedule(&rt2x00dev->txstatus_tasklet);
457}
458
459irqreturn_t rt2800mmio_interrupt(int irq, void *dev_instance)
460{
461 struct rt2x00_dev *rt2x00dev = dev_instance;
462 u32 reg, mask;
463
464 /* Read status and ACK all interrupts */
465 rt2x00mmio_register_read(rt2x00dev, INT_SOURCE_CSR, &reg);
466 rt2x00mmio_register_write(rt2x00dev, INT_SOURCE_CSR, reg);
467
468 if (!reg)
469 return IRQ_NONE;
470
471 if (!test_bit(DEVICE_STATE_ENABLED_RADIO, &rt2x00dev->flags))
472 return IRQ_HANDLED;
473
474 /*
475 * Since INT_MASK_CSR and INT_SOURCE_CSR use the same bits
476 * for interrupts and interrupt masks we can just use the value of
477 * INT_SOURCE_CSR to create the interrupt mask.
478 */
479 mask = ~reg;
480
481 if (rt2x00_get_field32(reg, INT_SOURCE_CSR_TX_FIFO_STATUS)) {
482 rt2800mmio_txstatus_interrupt(rt2x00dev);
483 /*
484 * Never disable the TX_FIFO_STATUS interrupt.
485 */
486 rt2x00_set_field32(&mask, INT_MASK_CSR_TX_FIFO_STATUS, 1);
487 }
488
489 if (rt2x00_get_field32(reg, INT_SOURCE_CSR_PRE_TBTT))
490 tasklet_hi_schedule(&rt2x00dev->pretbtt_tasklet);
491
492 if (rt2x00_get_field32(reg, INT_SOURCE_CSR_TBTT))
493 tasklet_hi_schedule(&rt2x00dev->tbtt_tasklet);
494
495 if (rt2x00_get_field32(reg, INT_SOURCE_CSR_RX_DONE))
496 tasklet_schedule(&rt2x00dev->rxdone_tasklet);
497
498 if (rt2x00_get_field32(reg, INT_SOURCE_CSR_AUTO_WAKEUP))
499 tasklet_schedule(&rt2x00dev->autowake_tasklet);
500
501 /*
502 * Disable all interrupts for which a tasklet was scheduled right now,
503 * the tasklet will reenable the appropriate interrupts.
504 */
505 spin_lock(&rt2x00dev->irqmask_lock);
506 rt2x00mmio_register_read(rt2x00dev, INT_MASK_CSR, &reg);
507 reg &= mask;
508 rt2x00mmio_register_write(rt2x00dev, INT_MASK_CSR, reg);
509 spin_unlock(&rt2x00dev->irqmask_lock);
510
511 return IRQ_HANDLED;
512}
513EXPORT_SYMBOL_GPL(rt2800mmio_interrupt);
514
515void rt2800mmio_toggle_irq(struct rt2x00_dev *rt2x00dev,
516 enum dev_state state)
517{
518 u32 reg;
519 unsigned long flags;
520
521 /*
522 * When interrupts are being enabled, the interrupt registers
523 * should clear the register to assure a clean state.
524 */
525 if (state == STATE_RADIO_IRQ_ON) {
526 rt2x00mmio_register_read(rt2x00dev, INT_SOURCE_CSR, &reg);
527 rt2x00mmio_register_write(rt2x00dev, INT_SOURCE_CSR, reg);
528 }
529
530 spin_lock_irqsave(&rt2x00dev->irqmask_lock, flags);
531 reg = 0;
532 if (state == STATE_RADIO_IRQ_ON) {
533 rt2x00_set_field32(&reg, INT_MASK_CSR_RX_DONE, 1);
534 rt2x00_set_field32(&reg, INT_MASK_CSR_TBTT, 1);
535 rt2x00_set_field32(&reg, INT_MASK_CSR_PRE_TBTT, 1);
536 rt2x00_set_field32(&reg, INT_MASK_CSR_TX_FIFO_STATUS, 1);
537 rt2x00_set_field32(&reg, INT_MASK_CSR_AUTO_WAKEUP, 1);
538 }
539 rt2x00mmio_register_write(rt2x00dev, INT_MASK_CSR, reg);
540 spin_unlock_irqrestore(&rt2x00dev->irqmask_lock, flags);
541
542 if (state == STATE_RADIO_IRQ_OFF) {
543 /*
544 * Wait for possibly running tasklets to finish.
545 */
546 tasklet_kill(&rt2x00dev->txstatus_tasklet);
547 tasklet_kill(&rt2x00dev->rxdone_tasklet);
548 tasklet_kill(&rt2x00dev->autowake_tasklet);
549 tasklet_kill(&rt2x00dev->tbtt_tasklet);
550 tasklet_kill(&rt2x00dev->pretbtt_tasklet);
551 }
552}
553EXPORT_SYMBOL_GPL(rt2800mmio_toggle_irq);
554
555/*
556 * Queue handlers.
557 */
558void rt2800mmio_start_queue(struct data_queue *queue)
559{
560 struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
561 u32 reg;
562
563 switch (queue->qid) {
564 case QID_RX:
565 rt2x00mmio_register_read(rt2x00dev, MAC_SYS_CTRL, &reg);
566 rt2x00_set_field32(&reg, MAC_SYS_CTRL_ENABLE_RX, 1);
567 rt2x00mmio_register_write(rt2x00dev, MAC_SYS_CTRL, reg);
568 break;
569 case QID_BEACON:
570 rt2x00mmio_register_read(rt2x00dev, BCN_TIME_CFG, &reg);
571 rt2x00_set_field32(&reg, BCN_TIME_CFG_TSF_TICKING, 1);
572 rt2x00_set_field32(&reg, BCN_TIME_CFG_TBTT_ENABLE, 1);
573 rt2x00_set_field32(&reg, BCN_TIME_CFG_BEACON_GEN, 1);
574 rt2x00mmio_register_write(rt2x00dev, BCN_TIME_CFG, reg);
575
576 rt2x00mmio_register_read(rt2x00dev, INT_TIMER_EN, &reg);
577 rt2x00_set_field32(&reg, INT_TIMER_EN_PRE_TBTT_TIMER, 1);
578 rt2x00mmio_register_write(rt2x00dev, INT_TIMER_EN, reg);
579 break;
580 default:
581 break;
582 }
583}
584EXPORT_SYMBOL_GPL(rt2800mmio_start_queue);
585
586void rt2800mmio_kick_queue(struct data_queue *queue)
587{
588 struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
589 struct queue_entry *entry;
590
591 switch (queue->qid) {
592 case QID_AC_VO:
593 case QID_AC_VI:
594 case QID_AC_BE:
595 case QID_AC_BK:
596 entry = rt2x00queue_get_entry(queue, Q_INDEX);
597 rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX(queue->qid),
598 entry->entry_idx);
599 break;
600 case QID_MGMT:
601 entry = rt2x00queue_get_entry(queue, Q_INDEX);
602 rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX(5),
603 entry->entry_idx);
604 break;
605 default:
606 break;
607 }
608}
609EXPORT_SYMBOL_GPL(rt2800mmio_kick_queue);
610
611void rt2800mmio_stop_queue(struct data_queue *queue)
612{
613 struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
614 u32 reg;
615
616 switch (queue->qid) {
617 case QID_RX:
618 rt2x00mmio_register_read(rt2x00dev, MAC_SYS_CTRL, &reg);
619 rt2x00_set_field32(&reg, MAC_SYS_CTRL_ENABLE_RX, 0);
620 rt2x00mmio_register_write(rt2x00dev, MAC_SYS_CTRL, reg);
621 break;
622 case QID_BEACON:
623 rt2x00mmio_register_read(rt2x00dev, BCN_TIME_CFG, &reg);
624 rt2x00_set_field32(&reg, BCN_TIME_CFG_TSF_TICKING, 0);
625 rt2x00_set_field32(&reg, BCN_TIME_CFG_TBTT_ENABLE, 0);
626 rt2x00_set_field32(&reg, BCN_TIME_CFG_BEACON_GEN, 0);
627 rt2x00mmio_register_write(rt2x00dev, BCN_TIME_CFG, reg);
628
629 rt2x00mmio_register_read(rt2x00dev, INT_TIMER_EN, &reg);
630 rt2x00_set_field32(&reg, INT_TIMER_EN_PRE_TBTT_TIMER, 0);
631 rt2x00mmio_register_write(rt2x00dev, INT_TIMER_EN, reg);
632
633 /*
634 * Wait for current invocation to finish. The tasklet
635 * won't be scheduled anymore afterwards since we disabled
636 * the TBTT and PRE TBTT timer.
637 */
638 tasklet_kill(&rt2x00dev->tbtt_tasklet);
639 tasklet_kill(&rt2x00dev->pretbtt_tasklet);
640
641 break;
642 default:
643 break;
644 }
645}
646EXPORT_SYMBOL_GPL(rt2800mmio_stop_queue);
647
648void rt2800mmio_queue_init(struct data_queue *queue)
649{
650 struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
651 unsigned short txwi_size, rxwi_size;
652
653 rt2800_get_txwi_rxwi_size(rt2x00dev, &txwi_size, &rxwi_size);
654
655 switch (queue->qid) {
656 case QID_RX:
657 queue->limit = 128;
658 queue->data_size = AGGREGATION_SIZE;
659 queue->desc_size = RXD_DESC_SIZE;
660 queue->winfo_size = rxwi_size;
661 queue->priv_size = sizeof(struct queue_entry_priv_mmio);
662 break;
663
664 case QID_AC_VO:
665 case QID_AC_VI:
666 case QID_AC_BE:
667 case QID_AC_BK:
668 queue->limit = 64;
669 queue->data_size = AGGREGATION_SIZE;
670 queue->desc_size = TXD_DESC_SIZE;
671 queue->winfo_size = txwi_size;
672 queue->priv_size = sizeof(struct queue_entry_priv_mmio);
673 break;
674
675 case QID_BEACON:
676 queue->limit = 8;
677 queue->data_size = 0; /* No DMA required for beacons */
678 queue->desc_size = TXD_DESC_SIZE;
679 queue->winfo_size = txwi_size;
680 queue->priv_size = sizeof(struct queue_entry_priv_mmio);
681 break;
682
683 case QID_ATIM:
684 /* fallthrough */
685 default:
686 BUG();
687 break;
688 }
689}
690EXPORT_SYMBOL_GPL(rt2800mmio_queue_init);
691
692/*
693 * Initialization functions.
694 */
695bool rt2800mmio_get_entry_state(struct queue_entry *entry)
696{
697 struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
698 u32 word;
699
700 if (entry->queue->qid == QID_RX) {
701 rt2x00_desc_read(entry_priv->desc, 1, &word);
702
703 return (!rt2x00_get_field32(word, RXD_W1_DMA_DONE));
704 } else {
705 rt2x00_desc_read(entry_priv->desc, 1, &word);
706
707 return (!rt2x00_get_field32(word, TXD_W1_DMA_DONE));
708 }
709}
710EXPORT_SYMBOL_GPL(rt2800mmio_get_entry_state);
711
712void rt2800mmio_clear_entry(struct queue_entry *entry)
713{
714 struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
715 struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
716 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
717 u32 word;
718
719 if (entry->queue->qid == QID_RX) {
720 rt2x00_desc_read(entry_priv->desc, 0, &word);
721 rt2x00_set_field32(&word, RXD_W0_SDP0, skbdesc->skb_dma);
722 rt2x00_desc_write(entry_priv->desc, 0, word);
723
724 rt2x00_desc_read(entry_priv->desc, 1, &word);
725 rt2x00_set_field32(&word, RXD_W1_DMA_DONE, 0);
726 rt2x00_desc_write(entry_priv->desc, 1, word);
727
728 /*
729 * Set RX IDX in register to inform hardware that we have
730 * handled this entry and it is available for reuse again.
731 */
732 rt2x00mmio_register_write(rt2x00dev, RX_CRX_IDX,
733 entry->entry_idx);
734 } else {
735 rt2x00_desc_read(entry_priv->desc, 1, &word);
736 rt2x00_set_field32(&word, TXD_W1_DMA_DONE, 1);
737 rt2x00_desc_write(entry_priv->desc, 1, word);
738 }
739}
740EXPORT_SYMBOL_GPL(rt2800mmio_clear_entry);
741
742int rt2800mmio_init_queues(struct rt2x00_dev *rt2x00dev)
743{
744 struct queue_entry_priv_mmio *entry_priv;
745
746 /*
747 * Initialize registers.
748 */
749 entry_priv = rt2x00dev->tx[0].entries[0].priv_data;
750 rt2x00mmio_register_write(rt2x00dev, TX_BASE_PTR0,
751 entry_priv->desc_dma);
752 rt2x00mmio_register_write(rt2x00dev, TX_MAX_CNT0,
753 rt2x00dev->tx[0].limit);
754 rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX0, 0);
755 rt2x00mmio_register_write(rt2x00dev, TX_DTX_IDX0, 0);
756
757 entry_priv = rt2x00dev->tx[1].entries[0].priv_data;
758 rt2x00mmio_register_write(rt2x00dev, TX_BASE_PTR1,
759 entry_priv->desc_dma);
760 rt2x00mmio_register_write(rt2x00dev, TX_MAX_CNT1,
761 rt2x00dev->tx[1].limit);
762 rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX1, 0);
763 rt2x00mmio_register_write(rt2x00dev, TX_DTX_IDX1, 0);
764
765 entry_priv = rt2x00dev->tx[2].entries[0].priv_data;
766 rt2x00mmio_register_write(rt2x00dev, TX_BASE_PTR2,
767 entry_priv->desc_dma);
768 rt2x00mmio_register_write(rt2x00dev, TX_MAX_CNT2,
769 rt2x00dev->tx[2].limit);
770 rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX2, 0);
771 rt2x00mmio_register_write(rt2x00dev, TX_DTX_IDX2, 0);
772
773 entry_priv = rt2x00dev->tx[3].entries[0].priv_data;
774 rt2x00mmio_register_write(rt2x00dev, TX_BASE_PTR3,
775 entry_priv->desc_dma);
776 rt2x00mmio_register_write(rt2x00dev, TX_MAX_CNT3,
777 rt2x00dev->tx[3].limit);
778 rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX3, 0);
779 rt2x00mmio_register_write(rt2x00dev, TX_DTX_IDX3, 0);
780
781 rt2x00mmio_register_write(rt2x00dev, TX_BASE_PTR4, 0);
782 rt2x00mmio_register_write(rt2x00dev, TX_MAX_CNT4, 0);
783 rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX4, 0);
784 rt2x00mmio_register_write(rt2x00dev, TX_DTX_IDX4, 0);
785
786 rt2x00mmio_register_write(rt2x00dev, TX_BASE_PTR5, 0);
787 rt2x00mmio_register_write(rt2x00dev, TX_MAX_CNT5, 0);
788 rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX5, 0);
789 rt2x00mmio_register_write(rt2x00dev, TX_DTX_IDX5, 0);
790
791 entry_priv = rt2x00dev->rx->entries[0].priv_data;
792 rt2x00mmio_register_write(rt2x00dev, RX_BASE_PTR,
793 entry_priv->desc_dma);
794 rt2x00mmio_register_write(rt2x00dev, RX_MAX_CNT,
795 rt2x00dev->rx[0].limit);
796 rt2x00mmio_register_write(rt2x00dev, RX_CRX_IDX,
797 rt2x00dev->rx[0].limit - 1);
798 rt2x00mmio_register_write(rt2x00dev, RX_DRX_IDX, 0);
799
800 rt2800_disable_wpdma(rt2x00dev);
801
802 rt2x00mmio_register_write(rt2x00dev, DELAY_INT_CFG, 0);
803
804 return 0;
805}
806EXPORT_SYMBOL_GPL(rt2800mmio_init_queues);
807
808int rt2800mmio_init_registers(struct rt2x00_dev *rt2x00dev)
809{
810 u32 reg;
811
812 /*
813 * Reset DMA indexes
814 */
815 rt2x00mmio_register_read(rt2x00dev, WPDMA_RST_IDX, &reg);
816 rt2x00_set_field32(&reg, WPDMA_RST_IDX_DTX_IDX0, 1);
817 rt2x00_set_field32(&reg, WPDMA_RST_IDX_DTX_IDX1, 1);
818 rt2x00_set_field32(&reg, WPDMA_RST_IDX_DTX_IDX2, 1);
819 rt2x00_set_field32(&reg, WPDMA_RST_IDX_DTX_IDX3, 1);
820 rt2x00_set_field32(&reg, WPDMA_RST_IDX_DTX_IDX4, 1);
821 rt2x00_set_field32(&reg, WPDMA_RST_IDX_DTX_IDX5, 1);
822 rt2x00_set_field32(&reg, WPDMA_RST_IDX_DRX_IDX0, 1);
823 rt2x00mmio_register_write(rt2x00dev, WPDMA_RST_IDX, reg);
824
825 rt2x00mmio_register_write(rt2x00dev, PBF_SYS_CTRL, 0x00000e1f);
826 rt2x00mmio_register_write(rt2x00dev, PBF_SYS_CTRL, 0x00000e00);
827
828 if (rt2x00_is_pcie(rt2x00dev) &&
829 (rt2x00_rt(rt2x00dev, RT3090) ||
830 rt2x00_rt(rt2x00dev, RT3390) ||
831 rt2x00_rt(rt2x00dev, RT3572) ||
832 rt2x00_rt(rt2x00dev, RT3593) ||
833 rt2x00_rt(rt2x00dev, RT5390) ||
834 rt2x00_rt(rt2x00dev, RT5392) ||
835 rt2x00_rt(rt2x00dev, RT5592))) {
836 rt2x00mmio_register_read(rt2x00dev, AUX_CTRL, &reg);
837 rt2x00_set_field32(&reg, AUX_CTRL_FORCE_PCIE_CLK, 1);
838 rt2x00_set_field32(&reg, AUX_CTRL_WAKE_PCIE_EN, 1);
839 rt2x00mmio_register_write(rt2x00dev, AUX_CTRL, reg);
840 }
841
842 rt2x00mmio_register_write(rt2x00dev, PWR_PIN_CFG, 0x00000003);
843
844 reg = 0;
845 rt2x00_set_field32(&reg, MAC_SYS_CTRL_RESET_CSR, 1);
846 rt2x00_set_field32(&reg, MAC_SYS_CTRL_RESET_BBP, 1);
847 rt2x00mmio_register_write(rt2x00dev, MAC_SYS_CTRL, reg);
848
849 rt2x00mmio_register_write(rt2x00dev, MAC_SYS_CTRL, 0x00000000);
850
851 return 0;
852}
853EXPORT_SYMBOL_GPL(rt2800mmio_init_registers);
854
855/*
856 * Device state switch handlers.
857 */
858int rt2800mmio_enable_radio(struct rt2x00_dev *rt2x00dev)
859{
860 /* Wait for DMA, ignore error until we initialize queues. */
861 rt2800_wait_wpdma_ready(rt2x00dev);
862
863 if (unlikely(rt2800mmio_init_queues(rt2x00dev)))
864 return -EIO;
865
866 return rt2800_enable_radio(rt2x00dev);
867}
868EXPORT_SYMBOL_GPL(rt2800mmio_enable_radio);
869
870MODULE_AUTHOR(DRV_PROJECT);
871MODULE_VERSION(DRV_VERSION);
872MODULE_DESCRIPTION("rt2800 MMIO library");
873MODULE_LICENSE("GPL");
diff --git a/drivers/net/wireless/rt2x00/rt2800mmio.h b/drivers/net/wireless/rt2x00/rt2800mmio.h
new file mode 100644
index 000000000000..6a10de3eee3e
--- /dev/null
+++ b/drivers/net/wireless/rt2x00/rt2800mmio.h
@@ -0,0 +1,165 @@
1/* Copyright (C) 2009 - 2010 Ivo van Doorn <IvDoorn@gmail.com>
2 * Copyright (C) 2009 Alban Browaeys <prahal@yahoo.com>
3 * Copyright (C) 2009 Felix Fietkau <nbd@openwrt.org>
4 * Copyright (C) 2009 Luis Correia <luis.f.correia@gmail.com>
5 * Copyright (C) 2009 Mattias Nissler <mattias.nissler@gmx.de>
6 * Copyright (C) 2009 Mark Asselstine <asselsm@gmail.com>
7 * Copyright (C) 2009 Xose Vazquez Perez <xose.vazquez@gmail.com>
8 * Copyright (C) 2009 Bart Zolnierkiewicz <bzolnier@gmail.com>
9 * <http://rt2x00.serialmonkey.com>
10 *
11 * This program is free software; you can redistribute it and/or modify
12 * it under the terms of the GNU General Public License as published by
13 * the Free Software Foundation; either version 2 of the License, or
14 * (at your option) any later version.
15 *
16 * This program is distributed in the hope that it will be useful,
17 * but WITHOUT ANY WARRANTY; without even the implied warranty of
18 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
19 * GNU General Public License for more details.
20 *
21 * You should have received a copy of the GNU General Public License
22 * along with this program; if not, write to the
23 * Free Software Foundation, Inc.,
24 * 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
25 */
26
27/* Module: rt2800mmio
28 * Abstract: forward declarations for the rt2800mmio module.
29 */
30
31#ifndef RT2800MMIO_H
32#define RT2800MMIO_H
33
34/*
35 * Queue register offset macros
36 */
37#define TX_QUEUE_REG_OFFSET 0x10
38#define TX_BASE_PTR(__x) (TX_BASE_PTR0 + ((__x) * TX_QUEUE_REG_OFFSET))
39#define TX_MAX_CNT(__x) (TX_MAX_CNT0 + ((__x) * TX_QUEUE_REG_OFFSET))
40#define TX_CTX_IDX(__x) (TX_CTX_IDX0 + ((__x) * TX_QUEUE_REG_OFFSET))
41#define TX_DTX_IDX(__x) (TX_DTX_IDX0 + ((__x) * TX_QUEUE_REG_OFFSET))
42
43/*
44 * DMA descriptor defines.
45 */
46#define TXD_DESC_SIZE (4 * sizeof(__le32))
47#define RXD_DESC_SIZE (4 * sizeof(__le32))
48
49/*
50 * TX descriptor format for TX, PRIO and Beacon Ring.
51 */
52
53/*
54 * Word0
55 */
56#define TXD_W0_SD_PTR0 FIELD32(0xffffffff)
57
58/*
59 * Word1
60 */
61#define TXD_W1_SD_LEN1 FIELD32(0x00003fff)
62#define TXD_W1_LAST_SEC1 FIELD32(0x00004000)
63#define TXD_W1_BURST FIELD32(0x00008000)
64#define TXD_W1_SD_LEN0 FIELD32(0x3fff0000)
65#define TXD_W1_LAST_SEC0 FIELD32(0x40000000)
66#define TXD_W1_DMA_DONE FIELD32(0x80000000)
67
68/*
69 * Word2
70 */
71#define TXD_W2_SD_PTR1 FIELD32(0xffffffff)
72
73/*
74 * Word3
75 * WIV: Wireless Info Valid. 1: Driver filled WI, 0: DMA needs to copy WI
76 * QSEL: Select on-chip FIFO ID for 2nd-stage output scheduler.
77 * 0:MGMT, 1:HCCA 2:EDCA
78 */
79#define TXD_W3_WIV FIELD32(0x01000000)
80#define TXD_W3_QSEL FIELD32(0x06000000)
81#define TXD_W3_TCO FIELD32(0x20000000)
82#define TXD_W3_UCO FIELD32(0x40000000)
83#define TXD_W3_ICO FIELD32(0x80000000)
84
85/*
86 * RX descriptor format for RX Ring.
87 */
88
89/*
90 * Word0
91 */
92#define RXD_W0_SDP0 FIELD32(0xffffffff)
93
94/*
95 * Word1
96 */
97#define RXD_W1_SDL1 FIELD32(0x00003fff)
98#define RXD_W1_SDL0 FIELD32(0x3fff0000)
99#define RXD_W1_LS0 FIELD32(0x40000000)
100#define RXD_W1_DMA_DONE FIELD32(0x80000000)
101
102/*
103 * Word2
104 */
105#define RXD_W2_SDP1 FIELD32(0xffffffff)
106
107/*
108 * Word3
109 * AMSDU: RX with 802.3 header, not 802.11 header.
110 * DECRYPTED: This frame is being decrypted.
111 */
112#define RXD_W3_BA FIELD32(0x00000001)
113#define RXD_W3_DATA FIELD32(0x00000002)
114#define RXD_W3_NULLDATA FIELD32(0x00000004)
115#define RXD_W3_FRAG FIELD32(0x00000008)
116#define RXD_W3_UNICAST_TO_ME FIELD32(0x00000010)
117#define RXD_W3_MULTICAST FIELD32(0x00000020)
118#define RXD_W3_BROADCAST FIELD32(0x00000040)
119#define RXD_W3_MY_BSS FIELD32(0x00000080)
120#define RXD_W3_CRC_ERROR FIELD32(0x00000100)
121#define RXD_W3_CIPHER_ERROR FIELD32(0x00000600)
122#define RXD_W3_AMSDU FIELD32(0x00000800)
123#define RXD_W3_HTC FIELD32(0x00001000)
124#define RXD_W3_RSSI FIELD32(0x00002000)
125#define RXD_W3_L2PAD FIELD32(0x00004000)
126#define RXD_W3_AMPDU FIELD32(0x00008000)
127#define RXD_W3_DECRYPTED FIELD32(0x00010000)
128#define RXD_W3_PLCP_SIGNAL FIELD32(0x00020000)
129#define RXD_W3_PLCP_RSSI FIELD32(0x00040000)
130
131/* TX descriptor initialization */
132__le32 *rt2800mmio_get_txwi(struct queue_entry *entry);
133void rt2800mmio_write_tx_desc(struct queue_entry *entry,
134 struct txentry_desc *txdesc);
135
136/* RX control handlers */
137void rt2800mmio_fill_rxdone(struct queue_entry *entry,
138 struct rxdone_entry_desc *rxdesc);
139
140/* Interrupt functions */
141void rt2800mmio_txstatus_tasklet(unsigned long data);
142void rt2800mmio_pretbtt_tasklet(unsigned long data);
143void rt2800mmio_tbtt_tasklet(unsigned long data);
144void rt2800mmio_rxdone_tasklet(unsigned long data);
145void rt2800mmio_autowake_tasklet(unsigned long data);
146irqreturn_t rt2800mmio_interrupt(int irq, void *dev_instance);
147void rt2800mmio_toggle_irq(struct rt2x00_dev *rt2x00dev,
148 enum dev_state state);
149
150/* Queue handlers */
151void rt2800mmio_start_queue(struct data_queue *queue);
152void rt2800mmio_kick_queue(struct data_queue *queue);
153void rt2800mmio_stop_queue(struct data_queue *queue);
154void rt2800mmio_queue_init(struct data_queue *queue);
155
156/* Initialization functions */
157bool rt2800mmio_get_entry_state(struct queue_entry *entry);
158void rt2800mmio_clear_entry(struct queue_entry *entry);
159int rt2800mmio_init_queues(struct rt2x00_dev *rt2x00dev);
160int rt2800mmio_init_registers(struct rt2x00_dev *rt2x00dev);
161
162/* Device state switch handlers. */
163int rt2800mmio_enable_radio(struct rt2x00_dev *rt2x00dev);
164
165#endif /* RT2800MMIO_H */
diff --git a/drivers/net/wireless/rt2x00/rt2800pci.c b/drivers/net/wireless/rt2x00/rt2800pci.c
index f8f2abbfbb65..b504455b4fec 100644
--- a/drivers/net/wireless/rt2x00/rt2800pci.c
+++ b/drivers/net/wireless/rt2x00/rt2800pci.c
@@ -37,14 +37,13 @@
37#include <linux/kernel.h> 37#include <linux/kernel.h>
38#include <linux/module.h> 38#include <linux/module.h>
39#include <linux/pci.h> 39#include <linux/pci.h>
40#include <linux/platform_device.h>
41#include <linux/eeprom_93cx6.h> 40#include <linux/eeprom_93cx6.h>
42 41
43#include "rt2x00.h" 42#include "rt2x00.h"
44#include "rt2x00mmio.h" 43#include "rt2x00mmio.h"
45#include "rt2x00pci.h" 44#include "rt2x00pci.h"
46#include "rt2x00soc.h"
47#include "rt2800lib.h" 45#include "rt2800lib.h"
46#include "rt2800mmio.h"
48#include "rt2800.h" 47#include "rt2800.h"
49#include "rt2800pci.h" 48#include "rt2800pci.h"
50 49
@@ -90,27 +89,6 @@ static void rt2800pci_mcu_status(struct rt2x00_dev *rt2x00dev, const u8 token)
90 rt2x00mmio_register_write(rt2x00dev, H2M_MAILBOX_CID, ~0); 89 rt2x00mmio_register_write(rt2x00dev, H2M_MAILBOX_CID, ~0);
91} 90}
92 91
93#if defined(CONFIG_SOC_RT288X) || defined(CONFIG_SOC_RT305X)
94static int rt2800pci_read_eeprom_soc(struct rt2x00_dev *rt2x00dev)
95{
96 void __iomem *base_addr = ioremap(0x1F040000, EEPROM_SIZE);
97
98 if (!base_addr)
99 return -ENOMEM;
100
101 memcpy_fromio(rt2x00dev->eeprom, base_addr, EEPROM_SIZE);
102
103 iounmap(base_addr);
104 return 0;
105}
106#else
107static inline int rt2800pci_read_eeprom_soc(struct rt2x00_dev *rt2x00dev)
108{
109 return -ENOMEM;
110}
111#endif /* CONFIG_SOC_RT288X || CONFIG_SOC_RT305X */
112
113#ifdef CONFIG_PCI
114static void rt2800pci_eepromregister_read(struct eeprom_93cx6 *eeprom) 92static void rt2800pci_eepromregister_read(struct eeprom_93cx6 *eeprom)
115{ 93{
116 struct rt2x00_dev *rt2x00dev = eeprom->data; 94 struct rt2x00_dev *rt2x00dev = eeprom->data;
@@ -183,112 +161,6 @@ static inline int rt2800pci_read_eeprom_efuse(struct rt2x00_dev *rt2x00dev)
183{ 161{
184 return rt2800_read_eeprom_efuse(rt2x00dev); 162 return rt2800_read_eeprom_efuse(rt2x00dev);
185} 163}
186#else
187static inline int rt2800pci_read_eeprom_pci(struct rt2x00_dev *rt2x00dev)
188{
189 return -EOPNOTSUPP;
190}
191
192static inline int rt2800pci_efuse_detect(struct rt2x00_dev *rt2x00dev)
193{
194 return 0;
195}
196
197static inline int rt2800pci_read_eeprom_efuse(struct rt2x00_dev *rt2x00dev)
198{
199 return -EOPNOTSUPP;
200}
201#endif /* CONFIG_PCI */
202
203/*
204 * Queue handlers.
205 */
206static void rt2800pci_start_queue(struct data_queue *queue)
207{
208 struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
209 u32 reg;
210
211 switch (queue->qid) {
212 case QID_RX:
213 rt2x00mmio_register_read(rt2x00dev, MAC_SYS_CTRL, &reg);
214 rt2x00_set_field32(&reg, MAC_SYS_CTRL_ENABLE_RX, 1);
215 rt2x00mmio_register_write(rt2x00dev, MAC_SYS_CTRL, reg);
216 break;
217 case QID_BEACON:
218 rt2x00mmio_register_read(rt2x00dev, BCN_TIME_CFG, &reg);
219 rt2x00_set_field32(&reg, BCN_TIME_CFG_TSF_TICKING, 1);
220 rt2x00_set_field32(&reg, BCN_TIME_CFG_TBTT_ENABLE, 1);
221 rt2x00_set_field32(&reg, BCN_TIME_CFG_BEACON_GEN, 1);
222 rt2x00mmio_register_write(rt2x00dev, BCN_TIME_CFG, reg);
223
224 rt2x00mmio_register_read(rt2x00dev, INT_TIMER_EN, &reg);
225 rt2x00_set_field32(&reg, INT_TIMER_EN_PRE_TBTT_TIMER, 1);
226 rt2x00mmio_register_write(rt2x00dev, INT_TIMER_EN, reg);
227 break;
228 default:
229 break;
230 }
231}
232
233static void rt2800pci_kick_queue(struct data_queue *queue)
234{
235 struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
236 struct queue_entry *entry;
237
238 switch (queue->qid) {
239 case QID_AC_VO:
240 case QID_AC_VI:
241 case QID_AC_BE:
242 case QID_AC_BK:
243 entry = rt2x00queue_get_entry(queue, Q_INDEX);
244 rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX(queue->qid),
245 entry->entry_idx);
246 break;
247 case QID_MGMT:
248 entry = rt2x00queue_get_entry(queue, Q_INDEX);
249 rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX(5),
250 entry->entry_idx);
251 break;
252 default:
253 break;
254 }
255}
256
257static void rt2800pci_stop_queue(struct data_queue *queue)
258{
259 struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
260 u32 reg;
261
262 switch (queue->qid) {
263 case QID_RX:
264 rt2x00mmio_register_read(rt2x00dev, MAC_SYS_CTRL, &reg);
265 rt2x00_set_field32(&reg, MAC_SYS_CTRL_ENABLE_RX, 0);
266 rt2x00mmio_register_write(rt2x00dev, MAC_SYS_CTRL, reg);
267 break;
268 case QID_BEACON:
269 rt2x00mmio_register_read(rt2x00dev, BCN_TIME_CFG, &reg);
270 rt2x00_set_field32(&reg, BCN_TIME_CFG_TSF_TICKING, 0);
271 rt2x00_set_field32(&reg, BCN_TIME_CFG_TBTT_ENABLE, 0);
272 rt2x00_set_field32(&reg, BCN_TIME_CFG_BEACON_GEN, 0);
273 rt2x00mmio_register_write(rt2x00dev, BCN_TIME_CFG, reg);
274
275 rt2x00mmio_register_read(rt2x00dev, INT_TIMER_EN, &reg);
276 rt2x00_set_field32(&reg, INT_TIMER_EN_PRE_TBTT_TIMER, 0);
277 rt2x00mmio_register_write(rt2x00dev, INT_TIMER_EN, reg);
278
279 /*
280 * Wait for current invocation to finish. The tasklet
281 * won't be scheduled anymore afterwards since we disabled
282 * the TBTT and PRE TBTT timer.
283 */
284 tasklet_kill(&rt2x00dev->tbtt_tasklet);
285 tasklet_kill(&rt2x00dev->pretbtt_tasklet);
286
287 break;
288 default:
289 break;
290 }
291}
292 164
293/* 165/*
294 * Firmware functions 166 * Firmware functions
@@ -332,217 +204,13 @@ static int rt2800pci_write_firmware(struct rt2x00_dev *rt2x00dev,
332} 204}
333 205
334/* 206/*
335 * Initialization functions.
336 */
337static bool rt2800pci_get_entry_state(struct queue_entry *entry)
338{
339 struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
340 u32 word;
341
342 if (entry->queue->qid == QID_RX) {
343 rt2x00_desc_read(entry_priv->desc, 1, &word);
344
345 return (!rt2x00_get_field32(word, RXD_W1_DMA_DONE));
346 } else {
347 rt2x00_desc_read(entry_priv->desc, 1, &word);
348
349 return (!rt2x00_get_field32(word, TXD_W1_DMA_DONE));
350 }
351}
352
353static void rt2800pci_clear_entry(struct queue_entry *entry)
354{
355 struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
356 struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
357 struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
358 u32 word;
359
360 if (entry->queue->qid == QID_RX) {
361 rt2x00_desc_read(entry_priv->desc, 0, &word);
362 rt2x00_set_field32(&word, RXD_W0_SDP0, skbdesc->skb_dma);
363 rt2x00_desc_write(entry_priv->desc, 0, word);
364
365 rt2x00_desc_read(entry_priv->desc, 1, &word);
366 rt2x00_set_field32(&word, RXD_W1_DMA_DONE, 0);
367 rt2x00_desc_write(entry_priv->desc, 1, word);
368
369 /*
370 * Set RX IDX in register to inform hardware that we have
371 * handled this entry and it is available for reuse again.
372 */
373 rt2x00mmio_register_write(rt2x00dev, RX_CRX_IDX,
374 entry->entry_idx);
375 } else {
376 rt2x00_desc_read(entry_priv->desc, 1, &word);
377 rt2x00_set_field32(&word, TXD_W1_DMA_DONE, 1);
378 rt2x00_desc_write(entry_priv->desc, 1, word);
379 }
380}
381
382static int rt2800pci_init_queues(struct rt2x00_dev *rt2x00dev)
383{
384 struct queue_entry_priv_mmio *entry_priv;
385
386 /*
387 * Initialize registers.
388 */
389 entry_priv = rt2x00dev->tx[0].entries[0].priv_data;
390 rt2x00mmio_register_write(rt2x00dev, TX_BASE_PTR0,
391 entry_priv->desc_dma);
392 rt2x00mmio_register_write(rt2x00dev, TX_MAX_CNT0,
393 rt2x00dev->tx[0].limit);
394 rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX0, 0);
395 rt2x00mmio_register_write(rt2x00dev, TX_DTX_IDX0, 0);
396
397 entry_priv = rt2x00dev->tx[1].entries[0].priv_data;
398 rt2x00mmio_register_write(rt2x00dev, TX_BASE_PTR1,
399 entry_priv->desc_dma);
400 rt2x00mmio_register_write(rt2x00dev, TX_MAX_CNT1,
401 rt2x00dev->tx[1].limit);
402 rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX1, 0);
403 rt2x00mmio_register_write(rt2x00dev, TX_DTX_IDX1, 0);
404
405 entry_priv = rt2x00dev->tx[2].entries[0].priv_data;
406 rt2x00mmio_register_write(rt2x00dev, TX_BASE_PTR2,
407 entry_priv->desc_dma);
408 rt2x00mmio_register_write(rt2x00dev, TX_MAX_CNT2,
409 rt2x00dev->tx[2].limit);
410 rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX2, 0);
411 rt2x00mmio_register_write(rt2x00dev, TX_DTX_IDX2, 0);
412
413 entry_priv = rt2x00dev->tx[3].entries[0].priv_data;
414 rt2x00mmio_register_write(rt2x00dev, TX_BASE_PTR3,
415 entry_priv->desc_dma);
416 rt2x00mmio_register_write(rt2x00dev, TX_MAX_CNT3,
417 rt2x00dev->tx[3].limit);
418 rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX3, 0);
419 rt2x00mmio_register_write(rt2x00dev, TX_DTX_IDX3, 0);
420
421 rt2x00mmio_register_write(rt2x00dev, TX_BASE_PTR4, 0);
422 rt2x00mmio_register_write(rt2x00dev, TX_MAX_CNT4, 0);
423 rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX4, 0);
424 rt2x00mmio_register_write(rt2x00dev, TX_DTX_IDX4, 0);
425
426 rt2x00mmio_register_write(rt2x00dev, TX_BASE_PTR5, 0);
427 rt2x00mmio_register_write(rt2x00dev, TX_MAX_CNT5, 0);
428 rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX5, 0);
429 rt2x00mmio_register_write(rt2x00dev, TX_DTX_IDX5, 0);
430
431 entry_priv = rt2x00dev->rx->entries[0].priv_data;
432 rt2x00mmio_register_write(rt2x00dev, RX_BASE_PTR,
433 entry_priv->desc_dma);
434 rt2x00mmio_register_write(rt2x00dev, RX_MAX_CNT,
435 rt2x00dev->rx[0].limit);
436 rt2x00mmio_register_write(rt2x00dev, RX_CRX_IDX,
437 rt2x00dev->rx[0].limit - 1);
438 rt2x00mmio_register_write(rt2x00dev, RX_DRX_IDX, 0);
439
440 rt2800_disable_wpdma(rt2x00dev);
441
442 rt2x00mmio_register_write(rt2x00dev, DELAY_INT_CFG, 0);
443
444 return 0;
445}
446
447/*
448 * Device state switch handlers. 207 * Device state switch handlers.
449 */ 208 */
450static void rt2800pci_toggle_irq(struct rt2x00_dev *rt2x00dev,
451 enum dev_state state)
452{
453 u32 reg;
454 unsigned long flags;
455
456 /*
457 * When interrupts are being enabled, the interrupt registers
458 * should clear the register to assure a clean state.
459 */
460 if (state == STATE_RADIO_IRQ_ON) {
461 rt2x00mmio_register_read(rt2x00dev, INT_SOURCE_CSR, &reg);
462 rt2x00mmio_register_write(rt2x00dev, INT_SOURCE_CSR, reg);
463 }
464
465 spin_lock_irqsave(&rt2x00dev->irqmask_lock, flags);
466 reg = 0;
467 if (state == STATE_RADIO_IRQ_ON) {
468 rt2x00_set_field32(&reg, INT_MASK_CSR_RX_DONE, 1);
469 rt2x00_set_field32(&reg, INT_MASK_CSR_TBTT, 1);
470 rt2x00_set_field32(&reg, INT_MASK_CSR_PRE_TBTT, 1);
471 rt2x00_set_field32(&reg, INT_MASK_CSR_TX_FIFO_STATUS, 1);
472 rt2x00_set_field32(&reg, INT_MASK_CSR_AUTO_WAKEUP, 1);
473 }
474 rt2x00mmio_register_write(rt2x00dev, INT_MASK_CSR, reg);
475 spin_unlock_irqrestore(&rt2x00dev->irqmask_lock, flags);
476
477 if (state == STATE_RADIO_IRQ_OFF) {
478 /*
479 * Wait for possibly running tasklets to finish.
480 */
481 tasklet_kill(&rt2x00dev->txstatus_tasklet);
482 tasklet_kill(&rt2x00dev->rxdone_tasklet);
483 tasklet_kill(&rt2x00dev->autowake_tasklet);
484 tasklet_kill(&rt2x00dev->tbtt_tasklet);
485 tasklet_kill(&rt2x00dev->pretbtt_tasklet);
486 }
487}
488
489static int rt2800pci_init_registers(struct rt2x00_dev *rt2x00dev)
490{
491 u32 reg;
492
493 /*
494 * Reset DMA indexes
495 */
496 rt2x00mmio_register_read(rt2x00dev, WPDMA_RST_IDX, &reg);
497 rt2x00_set_field32(&reg, WPDMA_RST_IDX_DTX_IDX0, 1);
498 rt2x00_set_field32(&reg, WPDMA_RST_IDX_DTX_IDX1, 1);
499 rt2x00_set_field32(&reg, WPDMA_RST_IDX_DTX_IDX2, 1);
500 rt2x00_set_field32(&reg, WPDMA_RST_IDX_DTX_IDX3, 1);
501 rt2x00_set_field32(&reg, WPDMA_RST_IDX_DTX_IDX4, 1);
502 rt2x00_set_field32(&reg, WPDMA_RST_IDX_DTX_IDX5, 1);
503 rt2x00_set_field32(&reg, WPDMA_RST_IDX_DRX_IDX0, 1);
504 rt2x00mmio_register_write(rt2x00dev, WPDMA_RST_IDX, reg);
505
506 rt2x00mmio_register_write(rt2x00dev, PBF_SYS_CTRL, 0x00000e1f);
507 rt2x00mmio_register_write(rt2x00dev, PBF_SYS_CTRL, 0x00000e00);
508
509 if (rt2x00_is_pcie(rt2x00dev) &&
510 (rt2x00_rt(rt2x00dev, RT3090) ||
511 rt2x00_rt(rt2x00dev, RT3390) ||
512 rt2x00_rt(rt2x00dev, RT3572) ||
513 rt2x00_rt(rt2x00dev, RT3593) ||
514 rt2x00_rt(rt2x00dev, RT5390) ||
515 rt2x00_rt(rt2x00dev, RT5392) ||
516 rt2x00_rt(rt2x00dev, RT5592))) {
517 rt2x00mmio_register_read(rt2x00dev, AUX_CTRL, &reg);
518 rt2x00_set_field32(&reg, AUX_CTRL_FORCE_PCIE_CLK, 1);
519 rt2x00_set_field32(&reg, AUX_CTRL_WAKE_PCIE_EN, 1);
520 rt2x00mmio_register_write(rt2x00dev, AUX_CTRL, reg);
521 }
522
523 rt2x00mmio_register_write(rt2x00dev, PWR_PIN_CFG, 0x00000003);
524
525 reg = 0;
526 rt2x00_set_field32(&reg, MAC_SYS_CTRL_RESET_CSR, 1);
527 rt2x00_set_field32(&reg, MAC_SYS_CTRL_RESET_BBP, 1);
528 rt2x00mmio_register_write(rt2x00dev, MAC_SYS_CTRL, reg);
529
530 rt2x00mmio_register_write(rt2x00dev, MAC_SYS_CTRL, 0x00000000);
531
532 return 0;
533}
534
535static int rt2800pci_enable_radio(struct rt2x00_dev *rt2x00dev) 209static int rt2800pci_enable_radio(struct rt2x00_dev *rt2x00dev)
536{ 210{
537 int retval; 211 int retval;
538 212
539 /* Wait for DMA, ignore error until we initialize queues. */ 213 retval = rt2800mmio_enable_radio(rt2x00dev);
540 rt2800_wait_wpdma_ready(rt2x00dev);
541
542 if (unlikely(rt2800pci_init_queues(rt2x00dev)))
543 return -EIO;
544
545 retval = rt2800_enable_radio(rt2x00dev);
546 if (retval) 214 if (retval)
547 return retval; 215 return retval;
548 216
@@ -559,15 +227,6 @@ static int rt2800pci_enable_radio(struct rt2x00_dev *rt2x00dev)
559 return retval; 227 return retval;
560} 228}
561 229
562static void rt2800pci_disable_radio(struct rt2x00_dev *rt2x00dev)
563{
564 if (rt2x00_is_soc(rt2x00dev)) {
565 rt2800_disable_radio(rt2x00dev);
566 rt2x00mmio_register_write(rt2x00dev, PWR_PIN_CFG, 0);
567 rt2x00mmio_register_write(rt2x00dev, TX_PIN_CFG, 0);
568 }
569}
570
571static int rt2800pci_set_state(struct rt2x00_dev *rt2x00dev, 230static int rt2800pci_set_state(struct rt2x00_dev *rt2x00dev,
572 enum dev_state state) 231 enum dev_state state)
573{ 232{
@@ -601,12 +260,11 @@ static int rt2800pci_set_device_state(struct rt2x00_dev *rt2x00dev,
601 * After the radio has been disabled, the device should 260 * After the radio has been disabled, the device should
602 * be put to sleep for powersaving. 261 * be put to sleep for powersaving.
603 */ 262 */
604 rt2800pci_disable_radio(rt2x00dev);
605 rt2800pci_set_state(rt2x00dev, STATE_SLEEP); 263 rt2800pci_set_state(rt2x00dev, STATE_SLEEP);
606 break; 264 break;
607 case STATE_RADIO_IRQ_ON: 265 case STATE_RADIO_IRQ_ON:
608 case STATE_RADIO_IRQ_OFF: 266 case STATE_RADIO_IRQ_OFF:
609 rt2800pci_toggle_irq(rt2x00dev, state); 267 rt2800mmio_toggle_irq(rt2x00dev, state);
610 break; 268 break;
611 case STATE_DEEP_SLEEP: 269 case STATE_DEEP_SLEEP:
612 case STATE_SLEEP: 270 case STATE_SLEEP:
@@ -627,479 +285,13 @@ static int rt2800pci_set_device_state(struct rt2x00_dev *rt2x00dev,
627} 285}
628 286
629/* 287/*
630 * TX descriptor initialization
631 */
632static __le32 *rt2800pci_get_txwi(struct queue_entry *entry)
633{
634 return (__le32 *) entry->skb->data;
635}
636
637static void rt2800pci_write_tx_desc(struct queue_entry *entry,
638 struct txentry_desc *txdesc)
639{
640 struct skb_frame_desc *skbdesc = get_skb_frame_desc(entry->skb);
641 struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
642 __le32 *txd = entry_priv->desc;
643 u32 word;
644 const unsigned int txwi_size = entry->queue->winfo_size;
645
646 /*
647 * The buffers pointed by SD_PTR0/SD_LEN0 and SD_PTR1/SD_LEN1
648 * must contains a TXWI structure + 802.11 header + padding + 802.11
649 * data. We choose to have SD_PTR0/SD_LEN0 only contains TXWI and
650 * SD_PTR1/SD_LEN1 contains 802.11 header + padding + 802.11
651 * data. It means that LAST_SEC0 is always 0.
652 */
653
654 /*
655 * Initialize TX descriptor
656 */
657 word = 0;
658 rt2x00_set_field32(&word, TXD_W0_SD_PTR0, skbdesc->skb_dma);
659 rt2x00_desc_write(txd, 0, word);
660
661 word = 0;
662 rt2x00_set_field32(&word, TXD_W1_SD_LEN1, entry->skb->len);
663 rt2x00_set_field32(&word, TXD_W1_LAST_SEC1,
664 !test_bit(ENTRY_TXD_MORE_FRAG, &txdesc->flags));
665 rt2x00_set_field32(&word, TXD_W1_BURST,
666 test_bit(ENTRY_TXD_BURST, &txdesc->flags));
667 rt2x00_set_field32(&word, TXD_W1_SD_LEN0, txwi_size);
668 rt2x00_set_field32(&word, TXD_W1_LAST_SEC0, 0);
669 rt2x00_set_field32(&word, TXD_W1_DMA_DONE, 0);
670 rt2x00_desc_write(txd, 1, word);
671
672 word = 0;
673 rt2x00_set_field32(&word, TXD_W2_SD_PTR1,
674 skbdesc->skb_dma + txwi_size);
675 rt2x00_desc_write(txd, 2, word);
676
677 word = 0;
678 rt2x00_set_field32(&word, TXD_W3_WIV,
679 !test_bit(ENTRY_TXD_ENCRYPT_IV, &txdesc->flags));
680 rt2x00_set_field32(&word, TXD_W3_QSEL, 2);
681 rt2x00_desc_write(txd, 3, word);
682
683 /*
684 * Register descriptor details in skb frame descriptor.
685 */
686 skbdesc->desc = txd;
687 skbdesc->desc_len = TXD_DESC_SIZE;
688}
689
690/*
691 * RX control handlers
692 */
693static void rt2800pci_fill_rxdone(struct queue_entry *entry,
694 struct rxdone_entry_desc *rxdesc)
695{
696 struct queue_entry_priv_mmio *entry_priv = entry->priv_data;
697 __le32 *rxd = entry_priv->desc;
698 u32 word;
699
700 rt2x00_desc_read(rxd, 3, &word);
701
702 if (rt2x00_get_field32(word, RXD_W3_CRC_ERROR))
703 rxdesc->flags |= RX_FLAG_FAILED_FCS_CRC;
704
705 /*
706 * Unfortunately we don't know the cipher type used during
707 * decryption. This prevents us from correct providing
708 * correct statistics through debugfs.
709 */
710 rxdesc->cipher_status = rt2x00_get_field32(word, RXD_W3_CIPHER_ERROR);
711
712 if (rt2x00_get_field32(word, RXD_W3_DECRYPTED)) {
713 /*
714 * Hardware has stripped IV/EIV data from 802.11 frame during
715 * decryption. Unfortunately the descriptor doesn't contain
716 * any fields with the EIV/IV data either, so they can't
717 * be restored by rt2x00lib.
718 */
719 rxdesc->flags |= RX_FLAG_IV_STRIPPED;
720
721 /*
722 * The hardware has already checked the Michael Mic and has
723 * stripped it from the frame. Signal this to mac80211.
724 */
725 rxdesc->flags |= RX_FLAG_MMIC_STRIPPED;
726
727 if (rxdesc->cipher_status == RX_CRYPTO_SUCCESS)
728 rxdesc->flags |= RX_FLAG_DECRYPTED;
729 else if (rxdesc->cipher_status == RX_CRYPTO_FAIL_MIC)
730 rxdesc->flags |= RX_FLAG_MMIC_ERROR;
731 }
732
733 if (rt2x00_get_field32(word, RXD_W3_MY_BSS))
734 rxdesc->dev_flags |= RXDONE_MY_BSS;
735
736 if (rt2x00_get_field32(word, RXD_W3_L2PAD))
737 rxdesc->dev_flags |= RXDONE_L2PAD;
738
739 /*
740 * Process the RXWI structure that is at the start of the buffer.
741 */
742 rt2800_process_rxwi(entry, rxdesc);
743}
744
745/*
746 * Interrupt functions.
747 */
748static void rt2800pci_wakeup(struct rt2x00_dev *rt2x00dev)
749{
750 struct ieee80211_conf conf = { .flags = 0 };
751 struct rt2x00lib_conf libconf = { .conf = &conf };
752
753 rt2800_config(rt2x00dev, &libconf, IEEE80211_CONF_CHANGE_PS);
754}
755
756static bool rt2800pci_txdone_entry_check(struct queue_entry *entry, u32 status)
757{
758 __le32 *txwi;
759 u32 word;
760 int wcid, tx_wcid;
761
762 wcid = rt2x00_get_field32(status, TX_STA_FIFO_WCID);
763
764 txwi = rt2800_drv_get_txwi(entry);
765 rt2x00_desc_read(txwi, 1, &word);
766 tx_wcid = rt2x00_get_field32(word, TXWI_W1_WIRELESS_CLI_ID);
767
768 return (tx_wcid == wcid);
769}
770
771static bool rt2800pci_txdone_find_entry(struct queue_entry *entry, void *data)
772{
773 u32 status = *(u32 *)data;
774
775 /*
776 * rt2800pci hardware might reorder frames when exchanging traffic
777 * with multiple BA enabled STAs.
778 *
779 * For example, a tx queue
780 * [ STA1 | STA2 | STA1 | STA2 ]
781 * can result in tx status reports
782 * [ STA1 | STA1 | STA2 | STA2 ]
783 * when the hw decides to aggregate the frames for STA1 into one AMPDU.
784 *
785 * To mitigate this effect, associate the tx status to the first frame
786 * in the tx queue with a matching wcid.
787 */
788 if (rt2800pci_txdone_entry_check(entry, status) &&
789 !test_bit(ENTRY_DATA_STATUS_SET, &entry->flags)) {
790 /*
791 * Got a matching frame, associate the tx status with
792 * the frame
793 */
794 entry->status = status;
795 set_bit(ENTRY_DATA_STATUS_SET, &entry->flags);
796 return true;
797 }
798
799 /* Check the next frame */
800 return false;
801}
802
803static bool rt2800pci_txdone_match_first(struct queue_entry *entry, void *data)
804{
805 u32 status = *(u32 *)data;
806
807 /*
808 * Find the first frame without tx status and assign this status to it
809 * regardless if it matches or not.
810 */
811 if (!test_bit(ENTRY_DATA_STATUS_SET, &entry->flags)) {
812 /*
813 * Got a matching frame, associate the tx status with
814 * the frame
815 */
816 entry->status = status;
817 set_bit(ENTRY_DATA_STATUS_SET, &entry->flags);
818 return true;
819 }
820
821 /* Check the next frame */
822 return false;
823}
824static bool rt2800pci_txdone_release_entries(struct queue_entry *entry,
825 void *data)
826{
827 if (test_bit(ENTRY_DATA_STATUS_SET, &entry->flags)) {
828 rt2800_txdone_entry(entry, entry->status,
829 rt2800pci_get_txwi(entry));
830 return false;
831 }
832
833 /* No more frames to release */
834 return true;
835}
836
837static bool rt2800pci_txdone(struct rt2x00_dev *rt2x00dev)
838{
839 struct data_queue *queue;
840 u32 status;
841 u8 qid;
842 int max_tx_done = 16;
843
844 while (kfifo_get(&rt2x00dev->txstatus_fifo, &status)) {
845 qid = rt2x00_get_field32(status, TX_STA_FIFO_PID_QUEUE);
846 if (unlikely(qid >= QID_RX)) {
847 /*
848 * Unknown queue, this shouldn't happen. Just drop
849 * this tx status.
850 */
851 rt2x00_warn(rt2x00dev, "Got TX status report with unexpected pid %u, dropping\n",
852 qid);
853 break;
854 }
855
856 queue = rt2x00queue_get_tx_queue(rt2x00dev, qid);
857 if (unlikely(queue == NULL)) {
858 /*
859 * The queue is NULL, this shouldn't happen. Stop
860 * processing here and drop the tx status
861 */
862 rt2x00_warn(rt2x00dev, "Got TX status for an unavailable queue %u, dropping\n",
863 qid);
864 break;
865 }
866
867 if (unlikely(rt2x00queue_empty(queue))) {
868 /*
869 * The queue is empty. Stop processing here
870 * and drop the tx status.
871 */
872 rt2x00_warn(rt2x00dev, "Got TX status for an empty queue %u, dropping\n",
873 qid);
874 break;
875 }
876
877 /*
878 * Let's associate this tx status with the first
879 * matching frame.
880 */
881 if (!rt2x00queue_for_each_entry(queue, Q_INDEX_DONE,
882 Q_INDEX, &status,
883 rt2800pci_txdone_find_entry)) {
884 /*
885 * We cannot match the tx status to any frame, so just
886 * use the first one.
887 */
888 if (!rt2x00queue_for_each_entry(queue, Q_INDEX_DONE,
889 Q_INDEX, &status,
890 rt2800pci_txdone_match_first)) {
891 rt2x00_warn(rt2x00dev, "No frame found for TX status on queue %u, dropping\n",
892 qid);
893 break;
894 }
895 }
896
897 /*
898 * Release all frames with a valid tx status.
899 */
900 rt2x00queue_for_each_entry(queue, Q_INDEX_DONE,
901 Q_INDEX, NULL,
902 rt2800pci_txdone_release_entries);
903
904 if (--max_tx_done == 0)
905 break;
906 }
907
908 return !max_tx_done;
909}
910
911static inline void rt2800pci_enable_interrupt(struct rt2x00_dev *rt2x00dev,
912 struct rt2x00_field32 irq_field)
913{
914 u32 reg;
915
916 /*
917 * Enable a single interrupt. The interrupt mask register
918 * access needs locking.
919 */
920 spin_lock_irq(&rt2x00dev->irqmask_lock);
921 rt2x00mmio_register_read(rt2x00dev, INT_MASK_CSR, &reg);
922 rt2x00_set_field32(&reg, irq_field, 1);
923 rt2x00mmio_register_write(rt2x00dev, INT_MASK_CSR, reg);
924 spin_unlock_irq(&rt2x00dev->irqmask_lock);
925}
926
927static void rt2800pci_txstatus_tasklet(unsigned long data)
928{
929 struct rt2x00_dev *rt2x00dev = (struct rt2x00_dev *)data;
930 if (rt2800pci_txdone(rt2x00dev))
931 tasklet_schedule(&rt2x00dev->txstatus_tasklet);
932
933 /*
934 * No need to enable the tx status interrupt here as we always
935 * leave it enabled to minimize the possibility of a tx status
936 * register overflow. See comment in interrupt handler.
937 */
938}
939
940static void rt2800pci_pretbtt_tasklet(unsigned long data)
941{
942 struct rt2x00_dev *rt2x00dev = (struct rt2x00_dev *)data;
943 rt2x00lib_pretbtt(rt2x00dev);
944 if (test_bit(DEVICE_STATE_ENABLED_RADIO, &rt2x00dev->flags))
945 rt2800pci_enable_interrupt(rt2x00dev, INT_MASK_CSR_PRE_TBTT);
946}
947
948static void rt2800pci_tbtt_tasklet(unsigned long data)
949{
950 struct rt2x00_dev *rt2x00dev = (struct rt2x00_dev *)data;
951 struct rt2800_drv_data *drv_data = rt2x00dev->drv_data;
952 u32 reg;
953
954 rt2x00lib_beacondone(rt2x00dev);
955
956 if (rt2x00dev->intf_ap_count) {
957 /*
958 * The rt2800pci hardware tbtt timer is off by 1us per tbtt
959 * causing beacon skew and as a result causing problems with
960 * some powersaving clients over time. Shorten the beacon
961 * interval every 64 beacons by 64us to mitigate this effect.
962 */
963 if (drv_data->tbtt_tick == (BCN_TBTT_OFFSET - 2)) {
964 rt2x00mmio_register_read(rt2x00dev, BCN_TIME_CFG, &reg);
965 rt2x00_set_field32(&reg, BCN_TIME_CFG_BEACON_INTERVAL,
966 (rt2x00dev->beacon_int * 16) - 1);
967 rt2x00mmio_register_write(rt2x00dev, BCN_TIME_CFG, reg);
968 } else if (drv_data->tbtt_tick == (BCN_TBTT_OFFSET - 1)) {
969 rt2x00mmio_register_read(rt2x00dev, BCN_TIME_CFG, &reg);
970 rt2x00_set_field32(&reg, BCN_TIME_CFG_BEACON_INTERVAL,
971 (rt2x00dev->beacon_int * 16));
972 rt2x00mmio_register_write(rt2x00dev, BCN_TIME_CFG, reg);
973 }
974 drv_data->tbtt_tick++;
975 drv_data->tbtt_tick %= BCN_TBTT_OFFSET;
976 }
977
978 if (test_bit(DEVICE_STATE_ENABLED_RADIO, &rt2x00dev->flags))
979 rt2800pci_enable_interrupt(rt2x00dev, INT_MASK_CSR_TBTT);
980}
981
982static void rt2800pci_rxdone_tasklet(unsigned long data)
983{
984 struct rt2x00_dev *rt2x00dev = (struct rt2x00_dev *)data;
985 if (rt2x00mmio_rxdone(rt2x00dev))
986 tasklet_schedule(&rt2x00dev->rxdone_tasklet);
987 else if (test_bit(DEVICE_STATE_ENABLED_RADIO, &rt2x00dev->flags))
988 rt2800pci_enable_interrupt(rt2x00dev, INT_MASK_CSR_RX_DONE);
989}
990
991static void rt2800pci_autowake_tasklet(unsigned long data)
992{
993 struct rt2x00_dev *rt2x00dev = (struct rt2x00_dev *)data;
994 rt2800pci_wakeup(rt2x00dev);
995 if (test_bit(DEVICE_STATE_ENABLED_RADIO, &rt2x00dev->flags))
996 rt2800pci_enable_interrupt(rt2x00dev, INT_MASK_CSR_AUTO_WAKEUP);
997}
998
999static void rt2800pci_txstatus_interrupt(struct rt2x00_dev *rt2x00dev)
1000{
1001 u32 status;
1002 int i;
1003
1004 /*
1005 * The TX_FIFO_STATUS interrupt needs special care. We should
1006 * read TX_STA_FIFO but we should do it immediately as otherwise
1007 * the register can overflow and we would lose status reports.
1008 *
1009 * Hence, read the TX_STA_FIFO register and copy all tx status
1010 * reports into a kernel FIFO which is handled in the txstatus
1011 * tasklet. We use a tasklet to process the tx status reports
1012 * because we can schedule the tasklet multiple times (when the
1013 * interrupt fires again during tx status processing).
1014 *
1015 * Furthermore we don't disable the TX_FIFO_STATUS
1016 * interrupt here but leave it enabled so that the TX_STA_FIFO
1017 * can also be read while the tx status tasklet gets executed.
1018 *
1019 * Since we have only one producer and one consumer we don't
1020 * need to lock the kfifo.
1021 */
1022 for (i = 0; i < rt2x00dev->tx->limit; i++) {
1023 rt2x00mmio_register_read(rt2x00dev, TX_STA_FIFO, &status);
1024
1025 if (!rt2x00_get_field32(status, TX_STA_FIFO_VALID))
1026 break;
1027
1028 if (!kfifo_put(&rt2x00dev->txstatus_fifo, &status)) {
1029 rt2x00_warn(rt2x00dev, "TX status FIFO overrun, drop tx status report\n");
1030 break;
1031 }
1032 }
1033
1034 /* Schedule the tasklet for processing the tx status. */
1035 tasklet_schedule(&rt2x00dev->txstatus_tasklet);
1036}
1037
1038static irqreturn_t rt2800pci_interrupt(int irq, void *dev_instance)
1039{
1040 struct rt2x00_dev *rt2x00dev = dev_instance;
1041 u32 reg, mask;
1042
1043 /* Read status and ACK all interrupts */
1044 rt2x00mmio_register_read(rt2x00dev, INT_SOURCE_CSR, &reg);
1045 rt2x00mmio_register_write(rt2x00dev, INT_SOURCE_CSR, reg);
1046
1047 if (!reg)
1048 return IRQ_NONE;
1049
1050 if (!test_bit(DEVICE_STATE_ENABLED_RADIO, &rt2x00dev->flags))
1051 return IRQ_HANDLED;
1052
1053 /*
1054 * Since INT_MASK_CSR and INT_SOURCE_CSR use the same bits
1055 * for interrupts and interrupt masks we can just use the value of
1056 * INT_SOURCE_CSR to create the interrupt mask.
1057 */
1058 mask = ~reg;
1059
1060 if (rt2x00_get_field32(reg, INT_SOURCE_CSR_TX_FIFO_STATUS)) {
1061 rt2800pci_txstatus_interrupt(rt2x00dev);
1062 /*
1063 * Never disable the TX_FIFO_STATUS interrupt.
1064 */
1065 rt2x00_set_field32(&mask, INT_MASK_CSR_TX_FIFO_STATUS, 1);
1066 }
1067
1068 if (rt2x00_get_field32(reg, INT_SOURCE_CSR_PRE_TBTT))
1069 tasklet_hi_schedule(&rt2x00dev->pretbtt_tasklet);
1070
1071 if (rt2x00_get_field32(reg, INT_SOURCE_CSR_TBTT))
1072 tasklet_hi_schedule(&rt2x00dev->tbtt_tasklet);
1073
1074 if (rt2x00_get_field32(reg, INT_SOURCE_CSR_RX_DONE))
1075 tasklet_schedule(&rt2x00dev->rxdone_tasklet);
1076
1077 if (rt2x00_get_field32(reg, INT_SOURCE_CSR_AUTO_WAKEUP))
1078 tasklet_schedule(&rt2x00dev->autowake_tasklet);
1079
1080 /*
1081 * Disable all interrupts for which a tasklet was scheduled right now,
1082 * the tasklet will reenable the appropriate interrupts.
1083 */
1084 spin_lock(&rt2x00dev->irqmask_lock);
1085 rt2x00mmio_register_read(rt2x00dev, INT_MASK_CSR, &reg);
1086 reg &= mask;
1087 rt2x00mmio_register_write(rt2x00dev, INT_MASK_CSR, reg);
1088 spin_unlock(&rt2x00dev->irqmask_lock);
1089
1090 return IRQ_HANDLED;
1091}
1092
1093/*
1094 * Device probe functions. 288 * Device probe functions.
1095 */ 289 */
1096static int rt2800pci_read_eeprom(struct rt2x00_dev *rt2x00dev) 290static int rt2800pci_read_eeprom(struct rt2x00_dev *rt2x00dev)
1097{ 291{
1098 int retval; 292 int retval;
1099 293
1100 if (rt2x00_is_soc(rt2x00dev)) 294 if (rt2800pci_efuse_detect(rt2x00dev))
1101 retval = rt2800pci_read_eeprom_soc(rt2x00dev);
1102 else if (rt2800pci_efuse_detect(rt2x00dev))
1103 retval = rt2800pci_read_eeprom_efuse(rt2x00dev); 295 retval = rt2800pci_read_eeprom_efuse(rt2x00dev);
1104 else 296 else
1105 retval = rt2800pci_read_eeprom_pci(rt2x00dev); 297 retval = rt2800pci_read_eeprom_pci(rt2x00dev);
@@ -1145,25 +337,25 @@ static const struct rt2800_ops rt2800pci_rt2800_ops = {
1145 .read_eeprom = rt2800pci_read_eeprom, 337 .read_eeprom = rt2800pci_read_eeprom,
1146 .hwcrypt_disabled = rt2800pci_hwcrypt_disabled, 338 .hwcrypt_disabled = rt2800pci_hwcrypt_disabled,
1147 .drv_write_firmware = rt2800pci_write_firmware, 339 .drv_write_firmware = rt2800pci_write_firmware,
1148 .drv_init_registers = rt2800pci_init_registers, 340 .drv_init_registers = rt2800mmio_init_registers,
1149 .drv_get_txwi = rt2800pci_get_txwi, 341 .drv_get_txwi = rt2800mmio_get_txwi,
1150}; 342};
1151 343
1152static const struct rt2x00lib_ops rt2800pci_rt2x00_ops = { 344static const struct rt2x00lib_ops rt2800pci_rt2x00_ops = {
1153 .irq_handler = rt2800pci_interrupt, 345 .irq_handler = rt2800mmio_interrupt,
1154 .txstatus_tasklet = rt2800pci_txstatus_tasklet, 346 .txstatus_tasklet = rt2800mmio_txstatus_tasklet,
1155 .pretbtt_tasklet = rt2800pci_pretbtt_tasklet, 347 .pretbtt_tasklet = rt2800mmio_pretbtt_tasklet,
1156 .tbtt_tasklet = rt2800pci_tbtt_tasklet, 348 .tbtt_tasklet = rt2800mmio_tbtt_tasklet,
1157 .rxdone_tasklet = rt2800pci_rxdone_tasklet, 349 .rxdone_tasklet = rt2800mmio_rxdone_tasklet,
1158 .autowake_tasklet = rt2800pci_autowake_tasklet, 350 .autowake_tasklet = rt2800mmio_autowake_tasklet,
1159 .probe_hw = rt2800_probe_hw, 351 .probe_hw = rt2800_probe_hw,
1160 .get_firmware_name = rt2800pci_get_firmware_name, 352 .get_firmware_name = rt2800pci_get_firmware_name,
1161 .check_firmware = rt2800_check_firmware, 353 .check_firmware = rt2800_check_firmware,
1162 .load_firmware = rt2800_load_firmware, 354 .load_firmware = rt2800_load_firmware,
1163 .initialize = rt2x00mmio_initialize, 355 .initialize = rt2x00mmio_initialize,
1164 .uninitialize = rt2x00mmio_uninitialize, 356 .uninitialize = rt2x00mmio_uninitialize,
1165 .get_entry_state = rt2800pci_get_entry_state, 357 .get_entry_state = rt2800mmio_get_entry_state,
1166 .clear_entry = rt2800pci_clear_entry, 358 .clear_entry = rt2800mmio_clear_entry,
1167 .set_device_state = rt2800pci_set_device_state, 359 .set_device_state = rt2800pci_set_device_state,
1168 .rfkill_poll = rt2800_rfkill_poll, 360 .rfkill_poll = rt2800_rfkill_poll,
1169 .link_stats = rt2800_link_stats, 361 .link_stats = rt2800_link_stats,
@@ -1171,15 +363,15 @@ static const struct rt2x00lib_ops rt2800pci_rt2x00_ops = {
1171 .link_tuner = rt2800_link_tuner, 363 .link_tuner = rt2800_link_tuner,
1172 .gain_calibration = rt2800_gain_calibration, 364 .gain_calibration = rt2800_gain_calibration,
1173 .vco_calibration = rt2800_vco_calibration, 365 .vco_calibration = rt2800_vco_calibration,
1174 .start_queue = rt2800pci_start_queue, 366 .start_queue = rt2800mmio_start_queue,
1175 .kick_queue = rt2800pci_kick_queue, 367 .kick_queue = rt2800mmio_kick_queue,
1176 .stop_queue = rt2800pci_stop_queue, 368 .stop_queue = rt2800mmio_stop_queue,
1177 .flush_queue = rt2x00mmio_flush_queue, 369 .flush_queue = rt2x00mmio_flush_queue,
1178 .write_tx_desc = rt2800pci_write_tx_desc, 370 .write_tx_desc = rt2800mmio_write_tx_desc,
1179 .write_tx_data = rt2800_write_tx_data, 371 .write_tx_data = rt2800_write_tx_data,
1180 .write_beacon = rt2800_write_beacon, 372 .write_beacon = rt2800_write_beacon,
1181 .clear_beacon = rt2800_clear_beacon, 373 .clear_beacon = rt2800_clear_beacon,
1182 .fill_rxdone = rt2800pci_fill_rxdone, 374 .fill_rxdone = rt2800mmio_fill_rxdone,
1183 .config_shared_key = rt2800_config_shared_key, 375 .config_shared_key = rt2800_config_shared_key,
1184 .config_pairwise_key = rt2800_config_pairwise_key, 376 .config_pairwise_key = rt2800_config_pairwise_key,
1185 .config_filter = rt2800_config_filter, 377 .config_filter = rt2800_config_filter,
@@ -1191,49 +383,6 @@ static const struct rt2x00lib_ops rt2800pci_rt2x00_ops = {
1191 .sta_remove = rt2800_sta_remove, 383 .sta_remove = rt2800_sta_remove,
1192}; 384};
1193 385
1194static void rt2800pci_queue_init(struct data_queue *queue)
1195{
1196 struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
1197 unsigned short txwi_size, rxwi_size;
1198
1199 rt2800_get_txwi_rxwi_size(rt2x00dev, &txwi_size, &rxwi_size);
1200
1201 switch (queue->qid) {
1202 case QID_RX:
1203 queue->limit = 128;
1204 queue->data_size = AGGREGATION_SIZE;
1205 queue->desc_size = RXD_DESC_SIZE;
1206 queue->winfo_size = rxwi_size;
1207 queue->priv_size = sizeof(struct queue_entry_priv_mmio);
1208 break;
1209
1210 case QID_AC_VO:
1211 case QID_AC_VI:
1212 case QID_AC_BE:
1213 case QID_AC_BK:
1214 queue->limit = 64;
1215 queue->data_size = AGGREGATION_SIZE;
1216 queue->desc_size = TXD_DESC_SIZE;
1217 queue->winfo_size = txwi_size;
1218 queue->priv_size = sizeof(struct queue_entry_priv_mmio);
1219 break;
1220
1221 case QID_BEACON:
1222 queue->limit = 8;
1223 queue->data_size = 0; /* No DMA required for beacons */
1224 queue->desc_size = TXD_DESC_SIZE;
1225 queue->winfo_size = txwi_size;
1226 queue->priv_size = sizeof(struct queue_entry_priv_mmio);
1227 break;
1228
1229 case QID_ATIM:
1230 /* fallthrough */
1231 default:
1232 BUG();
1233 break;
1234 }
1235}
1236
1237static const struct rt2x00_ops rt2800pci_ops = { 386static const struct rt2x00_ops rt2800pci_ops = {
1238 .name = KBUILD_MODNAME, 387 .name = KBUILD_MODNAME,
1239 .drv_data_size = sizeof(struct rt2800_drv_data), 388 .drv_data_size = sizeof(struct rt2800_drv_data),
@@ -1241,7 +390,7 @@ static const struct rt2x00_ops rt2800pci_ops = {
1241 .eeprom_size = EEPROM_SIZE, 390 .eeprom_size = EEPROM_SIZE,
1242 .rf_size = RF_SIZE, 391 .rf_size = RF_SIZE,
1243 .tx_queues = NUM_TX_QUEUES, 392 .tx_queues = NUM_TX_QUEUES,
1244 .queue_init = rt2800pci_queue_init, 393 .queue_init = rt2800mmio_queue_init,
1245 .lib = &rt2800pci_rt2x00_ops, 394 .lib = &rt2800pci_rt2x00_ops,
1246 .drv = &rt2800pci_rt2800_ops, 395 .drv = &rt2800pci_rt2800_ops,
1247 .hw = &rt2800pci_mac80211_ops, 396 .hw = &rt2800pci_mac80211_ops,
@@ -1253,7 +402,6 @@ static const struct rt2x00_ops rt2800pci_ops = {
1253/* 402/*
1254 * RT2800pci module information. 403 * RT2800pci module information.
1255 */ 404 */
1256#ifdef CONFIG_PCI
1257static DEFINE_PCI_DEVICE_TABLE(rt2800pci_device_table) = { 405static DEFINE_PCI_DEVICE_TABLE(rt2800pci_device_table) = {
1258 { PCI_DEVICE(0x1814, 0x0601) }, 406 { PCI_DEVICE(0x1814, 0x0601) },
1259 { PCI_DEVICE(0x1814, 0x0681) }, 407 { PCI_DEVICE(0x1814, 0x0681) },
@@ -1298,38 +446,15 @@ static DEFINE_PCI_DEVICE_TABLE(rt2800pci_device_table) = {
1298#endif 446#endif
1299 { 0, } 447 { 0, }
1300}; 448};
1301#endif /* CONFIG_PCI */
1302 449
1303MODULE_AUTHOR(DRV_PROJECT); 450MODULE_AUTHOR(DRV_PROJECT);
1304MODULE_VERSION(DRV_VERSION); 451MODULE_VERSION(DRV_VERSION);
1305MODULE_DESCRIPTION("Ralink RT2800 PCI & PCMCIA Wireless LAN driver."); 452MODULE_DESCRIPTION("Ralink RT2800 PCI & PCMCIA Wireless LAN driver.");
1306MODULE_SUPPORTED_DEVICE("Ralink RT2860 PCI & PCMCIA chipset based cards"); 453MODULE_SUPPORTED_DEVICE("Ralink RT2860 PCI & PCMCIA chipset based cards");
1307#ifdef CONFIG_PCI
1308MODULE_FIRMWARE(FIRMWARE_RT2860); 454MODULE_FIRMWARE(FIRMWARE_RT2860);
1309MODULE_DEVICE_TABLE(pci, rt2800pci_device_table); 455MODULE_DEVICE_TABLE(pci, rt2800pci_device_table);
1310#endif /* CONFIG_PCI */
1311MODULE_LICENSE("GPL"); 456MODULE_LICENSE("GPL");
1312 457
1313#if defined(CONFIG_SOC_RT288X) || defined(CONFIG_SOC_RT305X)
1314static int rt2800soc_probe(struct platform_device *pdev)
1315{
1316 return rt2x00soc_probe(pdev, &rt2800pci_ops);
1317}
1318
1319static struct platform_driver rt2800soc_driver = {
1320 .driver = {
1321 .name = "rt2800_wmac",
1322 .owner = THIS_MODULE,
1323 .mod_name = KBUILD_MODNAME,
1324 },
1325 .probe = rt2800soc_probe,
1326 .remove = rt2x00soc_remove,
1327 .suspend = rt2x00soc_suspend,
1328 .resume = rt2x00soc_resume,
1329};
1330#endif /* CONFIG_SOC_RT288X || CONFIG_SOC_RT305X */
1331
1332#ifdef CONFIG_PCI
1333static int rt2800pci_probe(struct pci_dev *pci_dev, 458static int rt2800pci_probe(struct pci_dev *pci_dev,
1334 const struct pci_device_id *id) 459 const struct pci_device_id *id)
1335{ 460{
@@ -1344,39 +469,5 @@ static struct pci_driver rt2800pci_driver = {
1344 .suspend = rt2x00pci_suspend, 469 .suspend = rt2x00pci_suspend,
1345 .resume = rt2x00pci_resume, 470 .resume = rt2x00pci_resume,
1346}; 471};
1347#endif /* CONFIG_PCI */
1348
1349static int __init rt2800pci_init(void)
1350{
1351 int ret = 0;
1352
1353#if defined(CONFIG_SOC_RT288X) || defined(CONFIG_SOC_RT305X)
1354 ret = platform_driver_register(&rt2800soc_driver);
1355 if (ret)
1356 return ret;
1357#endif
1358#ifdef CONFIG_PCI
1359 ret = pci_register_driver(&rt2800pci_driver);
1360 if (ret) {
1361#if defined(CONFIG_SOC_RT288X) || defined(CONFIG_SOC_RT305X)
1362 platform_driver_unregister(&rt2800soc_driver);
1363#endif
1364 return ret;
1365 }
1366#endif
1367
1368 return ret;
1369}
1370
1371static void __exit rt2800pci_exit(void)
1372{
1373#ifdef CONFIG_PCI
1374 pci_unregister_driver(&rt2800pci_driver);
1375#endif
1376#if defined(CONFIG_SOC_RT288X) || defined(CONFIG_SOC_RT305X)
1377 platform_driver_unregister(&rt2800soc_driver);
1378#endif
1379}
1380 472
1381module_init(rt2800pci_init); 473module_pci_driver(rt2800pci_driver);
1382module_exit(rt2800pci_exit);
diff --git a/drivers/net/wireless/rt2x00/rt2800pci.h b/drivers/net/wireless/rt2x00/rt2800pci.h
index ab22a087c50d..a81c9ee281c0 100644
--- a/drivers/net/wireless/rt2x00/rt2800pci.h
+++ b/drivers/net/wireless/rt2x00/rt2800pci.h
@@ -35,107 +35,10 @@
35#define RT2800PCI_H 35#define RT2800PCI_H
36 36
37/* 37/*
38 * Queue register offset macros
39 */
40#define TX_QUEUE_REG_OFFSET 0x10
41#define TX_BASE_PTR(__x) (TX_BASE_PTR0 + ((__x) * TX_QUEUE_REG_OFFSET))
42#define TX_MAX_CNT(__x) (TX_MAX_CNT0 + ((__x) * TX_QUEUE_REG_OFFSET))
43#define TX_CTX_IDX(__x) (TX_CTX_IDX0 + ((__x) * TX_QUEUE_REG_OFFSET))
44#define TX_DTX_IDX(__x) (TX_DTX_IDX0 + ((__x) * TX_QUEUE_REG_OFFSET))
45
46/*
47 * 8051 firmware image. 38 * 8051 firmware image.
48 */ 39 */
49#define FIRMWARE_RT2860 "rt2860.bin" 40#define FIRMWARE_RT2860 "rt2860.bin"
50#define FIRMWARE_RT3290 "rt3290.bin" 41#define FIRMWARE_RT3290 "rt3290.bin"
51#define FIRMWARE_IMAGE_BASE 0x2000 42#define FIRMWARE_IMAGE_BASE 0x2000
52 43
53/*
54 * DMA descriptor defines.
55 */
56#define TXD_DESC_SIZE (4 * sizeof(__le32))
57#define RXD_DESC_SIZE (4 * sizeof(__le32))
58
59/*
60 * TX descriptor format for TX, PRIO and Beacon Ring.
61 */
62
63/*
64 * Word0
65 */
66#define TXD_W0_SD_PTR0 FIELD32(0xffffffff)
67
68/*
69 * Word1
70 */
71#define TXD_W1_SD_LEN1 FIELD32(0x00003fff)
72#define TXD_W1_LAST_SEC1 FIELD32(0x00004000)
73#define TXD_W1_BURST FIELD32(0x00008000)
74#define TXD_W1_SD_LEN0 FIELD32(0x3fff0000)
75#define TXD_W1_LAST_SEC0 FIELD32(0x40000000)
76#define TXD_W1_DMA_DONE FIELD32(0x80000000)
77
78/*
79 * Word2
80 */
81#define TXD_W2_SD_PTR1 FIELD32(0xffffffff)
82
83/*
84 * Word3
85 * WIV: Wireless Info Valid. 1: Driver filled WI, 0: DMA needs to copy WI
86 * QSEL: Select on-chip FIFO ID for 2nd-stage output scheduler.
87 * 0:MGMT, 1:HCCA 2:EDCA
88 */
89#define TXD_W3_WIV FIELD32(0x01000000)
90#define TXD_W3_QSEL FIELD32(0x06000000)
91#define TXD_W3_TCO FIELD32(0x20000000)
92#define TXD_W3_UCO FIELD32(0x40000000)
93#define TXD_W3_ICO FIELD32(0x80000000)
94
95/*
96 * RX descriptor format for RX Ring.
97 */
98
99/*
100 * Word0
101 */
102#define RXD_W0_SDP0 FIELD32(0xffffffff)
103
104/*
105 * Word1
106 */
107#define RXD_W1_SDL1 FIELD32(0x00003fff)
108#define RXD_W1_SDL0 FIELD32(0x3fff0000)
109#define RXD_W1_LS0 FIELD32(0x40000000)
110#define RXD_W1_DMA_DONE FIELD32(0x80000000)
111
112/*
113 * Word2
114 */
115#define RXD_W2_SDP1 FIELD32(0xffffffff)
116
117/*
118 * Word3
119 * AMSDU: RX with 802.3 header, not 802.11 header.
120 * DECRYPTED: This frame is being decrypted.
121 */
122#define RXD_W3_BA FIELD32(0x00000001)
123#define RXD_W3_DATA FIELD32(0x00000002)
124#define RXD_W3_NULLDATA FIELD32(0x00000004)
125#define RXD_W3_FRAG FIELD32(0x00000008)
126#define RXD_W3_UNICAST_TO_ME FIELD32(0x00000010)
127#define RXD_W3_MULTICAST FIELD32(0x00000020)
128#define RXD_W3_BROADCAST FIELD32(0x00000040)
129#define RXD_W3_MY_BSS FIELD32(0x00000080)
130#define RXD_W3_CRC_ERROR FIELD32(0x00000100)
131#define RXD_W3_CIPHER_ERROR FIELD32(0x00000600)
132#define RXD_W3_AMSDU FIELD32(0x00000800)
133#define RXD_W3_HTC FIELD32(0x00001000)
134#define RXD_W3_RSSI FIELD32(0x00002000)
135#define RXD_W3_L2PAD FIELD32(0x00004000)
136#define RXD_W3_AMPDU FIELD32(0x00008000)
137#define RXD_W3_DECRYPTED FIELD32(0x00010000)
138#define RXD_W3_PLCP_SIGNAL FIELD32(0x00020000)
139#define RXD_W3_PLCP_RSSI FIELD32(0x00040000)
140
141#endif /* RT2800PCI_H */ 44#endif /* RT2800PCI_H */
diff --git a/drivers/net/wireless/rt2x00/rt2800soc.c b/drivers/net/wireless/rt2x00/rt2800soc.c
new file mode 100644
index 000000000000..1359227ca411
--- /dev/null
+++ b/drivers/net/wireless/rt2x00/rt2800soc.c
@@ -0,0 +1,263 @@
1/* Copyright (C) 2009 - 2010 Ivo van Doorn <IvDoorn@gmail.com>
2 * Copyright (C) 2009 Alban Browaeys <prahal@yahoo.com>
3 * Copyright (C) 2009 Felix Fietkau <nbd@openwrt.org>
4 * Copyright (C) 2009 Luis Correia <luis.f.correia@gmail.com>
5 * Copyright (C) 2009 Mattias Nissler <mattias.nissler@gmx.de>
6 * Copyright (C) 2009 Mark Asselstine <asselsm@gmail.com>
7 * Copyright (C) 2009 Xose Vazquez Perez <xose.vazquez@gmail.com>
8 * Copyright (C) 2009 Bart Zolnierkiewicz <bzolnier@gmail.com>
9 * <http://rt2x00.serialmonkey.com>
10 *
11 * This program is free software; you can redistribute it and/or modify
12 * it under the terms of the GNU General Public License as published by
13 * the Free Software Foundation; either version 2 of the License, or
14 * (at your option) any later version.
15 *
16 * This program is distributed in the hope that it will be useful,
17 * but WITHOUT ANY WARRANTY; without even the implied warranty of
18 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
19 * GNU General Public License for more details.
20 *
21 * You should have received a copy of the GNU General Public License
22 * along with this program; if not, write to the
23 * Free Software Foundation, Inc.,
24 * 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
25 */
26
27/* Module: rt2800soc
28 * Abstract: rt2800 WiSoC specific routines.
29 */
30
31#include <linux/etherdevice.h>
32#include <linux/init.h>
33#include <linux/kernel.h>
34#include <linux/module.h>
35#include <linux/platform_device.h>
36
37#include "rt2x00.h"
38#include "rt2x00mmio.h"
39#include "rt2x00soc.h"
40#include "rt2800.h"
41#include "rt2800lib.h"
42#include "rt2800mmio.h"
43
44/* Allow hardware encryption to be disabled. */
45static bool modparam_nohwcrypt;
46module_param_named(nohwcrypt, modparam_nohwcrypt, bool, S_IRUGO);
47MODULE_PARM_DESC(nohwcrypt, "Disable hardware encryption.");
48
49static bool rt2800soc_hwcrypt_disabled(struct rt2x00_dev *rt2x00dev)
50{
51 return modparam_nohwcrypt;
52}
53
54static void rt2800soc_disable_radio(struct rt2x00_dev *rt2x00dev)
55{
56 rt2800_disable_radio(rt2x00dev);
57 rt2x00mmio_register_write(rt2x00dev, PWR_PIN_CFG, 0);
58 rt2x00mmio_register_write(rt2x00dev, TX_PIN_CFG, 0);
59}
60
61static int rt2800soc_set_device_state(struct rt2x00_dev *rt2x00dev,
62 enum dev_state state)
63{
64 int retval = 0;
65
66 switch (state) {
67 case STATE_RADIO_ON:
68 retval = rt2800mmio_enable_radio(rt2x00dev);
69 break;
70
71 case STATE_RADIO_OFF:
72 rt2800soc_disable_radio(rt2x00dev);
73 break;
74
75 case STATE_RADIO_IRQ_ON:
76 case STATE_RADIO_IRQ_OFF:
77 rt2800mmio_toggle_irq(rt2x00dev, state);
78 break;
79
80 case STATE_DEEP_SLEEP:
81 case STATE_SLEEP:
82 case STATE_STANDBY:
83 case STATE_AWAKE:
84 /* These states are not supported, but don't report an error */
85 retval = 0;
86 break;
87
88 default:
89 retval = -ENOTSUPP;
90 break;
91 }
92
93 if (unlikely(retval))
94 rt2x00_err(rt2x00dev, "Device failed to enter state %d (%d)\n",
95 state, retval);
96
97 return retval;
98}
99
100static int rt2800soc_read_eeprom(struct rt2x00_dev *rt2x00dev)
101{
102 void __iomem *base_addr = ioremap(0x1F040000, EEPROM_SIZE);
103
104 if (!base_addr)
105 return -ENOMEM;
106
107 memcpy_fromio(rt2x00dev->eeprom, base_addr, EEPROM_SIZE);
108
109 iounmap(base_addr);
110 return 0;
111}
112
113/* Firmware functions */
114static char *rt2800soc_get_firmware_name(struct rt2x00_dev *rt2x00dev)
115{
116 WARN_ON_ONCE(1);
117 return NULL;
118}
119
120static int rt2800soc_load_firmware(struct rt2x00_dev *rt2x00dev,
121 const u8 *data, const size_t len)
122{
123 WARN_ON_ONCE(1);
124 return 0;
125}
126
127static int rt2800soc_check_firmware(struct rt2x00_dev *rt2x00dev,
128 const u8 *data, const size_t len)
129{
130 WARN_ON_ONCE(1);
131 return 0;
132}
133
134static int rt2800soc_write_firmware(struct rt2x00_dev *rt2x00dev,
135 const u8 *data, const size_t len)
136{
137 WARN_ON_ONCE(1);
138 return 0;
139}
140
141static const struct ieee80211_ops rt2800soc_mac80211_ops = {
142 .tx = rt2x00mac_tx,
143 .start = rt2x00mac_start,
144 .stop = rt2x00mac_stop,
145 .add_interface = rt2x00mac_add_interface,
146 .remove_interface = rt2x00mac_remove_interface,
147 .config = rt2x00mac_config,
148 .configure_filter = rt2x00mac_configure_filter,
149 .set_key = rt2x00mac_set_key,
150 .sw_scan_start = rt2x00mac_sw_scan_start,
151 .sw_scan_complete = rt2x00mac_sw_scan_complete,
152 .get_stats = rt2x00mac_get_stats,
153 .get_tkip_seq = rt2800_get_tkip_seq,
154 .set_rts_threshold = rt2800_set_rts_threshold,
155 .sta_add = rt2x00mac_sta_add,
156 .sta_remove = rt2x00mac_sta_remove,
157 .bss_info_changed = rt2x00mac_bss_info_changed,
158 .conf_tx = rt2800_conf_tx,
159 .get_tsf = rt2800_get_tsf,
160 .rfkill_poll = rt2x00mac_rfkill_poll,
161 .ampdu_action = rt2800_ampdu_action,
162 .flush = rt2x00mac_flush,
163 .get_survey = rt2800_get_survey,
164 .get_ringparam = rt2x00mac_get_ringparam,
165 .tx_frames_pending = rt2x00mac_tx_frames_pending,
166};
167
168static const struct rt2800_ops rt2800soc_rt2800_ops = {
169 .register_read = rt2x00mmio_register_read,
170 .register_read_lock = rt2x00mmio_register_read, /* same for SoCs */
171 .register_write = rt2x00mmio_register_write,
172 .register_write_lock = rt2x00mmio_register_write, /* same for SoCs */
173 .register_multiread = rt2x00mmio_register_multiread,
174 .register_multiwrite = rt2x00mmio_register_multiwrite,
175 .regbusy_read = rt2x00mmio_regbusy_read,
176 .read_eeprom = rt2800soc_read_eeprom,
177 .hwcrypt_disabled = rt2800soc_hwcrypt_disabled,
178 .drv_write_firmware = rt2800soc_write_firmware,
179 .drv_init_registers = rt2800mmio_init_registers,
180 .drv_get_txwi = rt2800mmio_get_txwi,
181};
182
183static const struct rt2x00lib_ops rt2800soc_rt2x00_ops = {
184 .irq_handler = rt2800mmio_interrupt,
185 .txstatus_tasklet = rt2800mmio_txstatus_tasklet,
186 .pretbtt_tasklet = rt2800mmio_pretbtt_tasklet,
187 .tbtt_tasklet = rt2800mmio_tbtt_tasklet,
188 .rxdone_tasklet = rt2800mmio_rxdone_tasklet,
189 .autowake_tasklet = rt2800mmio_autowake_tasklet,
190 .probe_hw = rt2800_probe_hw,
191 .get_firmware_name = rt2800soc_get_firmware_name,
192 .check_firmware = rt2800soc_check_firmware,
193 .load_firmware = rt2800soc_load_firmware,
194 .initialize = rt2x00mmio_initialize,
195 .uninitialize = rt2x00mmio_uninitialize,
196 .get_entry_state = rt2800mmio_get_entry_state,
197 .clear_entry = rt2800mmio_clear_entry,
198 .set_device_state = rt2800soc_set_device_state,
199 .rfkill_poll = rt2800_rfkill_poll,
200 .link_stats = rt2800_link_stats,
201 .reset_tuner = rt2800_reset_tuner,
202 .link_tuner = rt2800_link_tuner,
203 .gain_calibration = rt2800_gain_calibration,
204 .vco_calibration = rt2800_vco_calibration,
205 .start_queue = rt2800mmio_start_queue,
206 .kick_queue = rt2800mmio_kick_queue,
207 .stop_queue = rt2800mmio_stop_queue,
208 .flush_queue = rt2x00mmio_flush_queue,
209 .write_tx_desc = rt2800mmio_write_tx_desc,
210 .write_tx_data = rt2800_write_tx_data,
211 .write_beacon = rt2800_write_beacon,
212 .clear_beacon = rt2800_clear_beacon,
213 .fill_rxdone = rt2800mmio_fill_rxdone,
214 .config_shared_key = rt2800_config_shared_key,
215 .config_pairwise_key = rt2800_config_pairwise_key,
216 .config_filter = rt2800_config_filter,
217 .config_intf = rt2800_config_intf,
218 .config_erp = rt2800_config_erp,
219 .config_ant = rt2800_config_ant,
220 .config = rt2800_config,
221 .sta_add = rt2800_sta_add,
222 .sta_remove = rt2800_sta_remove,
223};
224
225static const struct rt2x00_ops rt2800soc_ops = {
226 .name = KBUILD_MODNAME,
227 .drv_data_size = sizeof(struct rt2800_drv_data),
228 .max_ap_intf = 8,
229 .eeprom_size = EEPROM_SIZE,
230 .rf_size = RF_SIZE,
231 .tx_queues = NUM_TX_QUEUES,
232 .queue_init = rt2800mmio_queue_init,
233 .lib = &rt2800soc_rt2x00_ops,
234 .drv = &rt2800soc_rt2800_ops,
235 .hw = &rt2800soc_mac80211_ops,
236#ifdef CONFIG_RT2X00_LIB_DEBUGFS
237 .debugfs = &rt2800_rt2x00debug,
238#endif /* CONFIG_RT2X00_LIB_DEBUGFS */
239};
240
241static int rt2800soc_probe(struct platform_device *pdev)
242{
243 return rt2x00soc_probe(pdev, &rt2800soc_ops);
244}
245
246static struct platform_driver rt2800soc_driver = {
247 .driver = {
248 .name = "rt2800_wmac",
249 .owner = THIS_MODULE,
250 .mod_name = KBUILD_MODNAME,
251 },
252 .probe = rt2800soc_probe,
253 .remove = rt2x00soc_remove,
254 .suspend = rt2x00soc_suspend,
255 .resume = rt2x00soc_resume,
256};
257
258module_platform_driver(rt2800soc_driver);
259
260MODULE_AUTHOR(DRV_PROJECT);
261MODULE_VERSION(DRV_VERSION);
262MODULE_DESCRIPTION("Ralink WiSoC Wireless LAN driver.");
263MODULE_LICENSE("GPL");
diff --git a/drivers/net/wireless/rt2x00/rt2800usb.c b/drivers/net/wireless/rt2x00/rt2800usb.c
index 96961b9a395c..997df03a0c2e 100644
--- a/drivers/net/wireless/rt2x00/rt2800usb.c
+++ b/drivers/net/wireless/rt2x00/rt2800usb.c
@@ -148,6 +148,8 @@ static bool rt2800usb_txstatus_timeout(struct rt2x00_dev *rt2x00dev)
148 return false; 148 return false;
149} 149}
150 150
151#define TXSTATUS_READ_INTERVAL 1000000
152
151static bool rt2800usb_tx_sta_fifo_read_completed(struct rt2x00_dev *rt2x00dev, 153static bool rt2800usb_tx_sta_fifo_read_completed(struct rt2x00_dev *rt2x00dev,
152 int urb_status, u32 tx_status) 154 int urb_status, u32 tx_status)
153{ 155{
@@ -176,8 +178,9 @@ static bool rt2800usb_tx_sta_fifo_read_completed(struct rt2x00_dev *rt2x00dev,
176 queue_work(rt2x00dev->workqueue, &rt2x00dev->txdone_work); 178 queue_work(rt2x00dev->workqueue, &rt2x00dev->txdone_work);
177 179
178 if (rt2800usb_txstatus_pending(rt2x00dev)) { 180 if (rt2800usb_txstatus_pending(rt2x00dev)) {
179 /* Read register after 250 us */ 181 /* Read register after 1 ms */
180 hrtimer_start(&rt2x00dev->txstatus_timer, ktime_set(0, 250000), 182 hrtimer_start(&rt2x00dev->txstatus_timer,
183 ktime_set(0, TXSTATUS_READ_INTERVAL),
181 HRTIMER_MODE_REL); 184 HRTIMER_MODE_REL);
182 return false; 185 return false;
183 } 186 }
@@ -202,8 +205,9 @@ static void rt2800usb_async_read_tx_status(struct rt2x00_dev *rt2x00dev)
202 if (test_and_set_bit(TX_STATUS_READING, &rt2x00dev->flags)) 205 if (test_and_set_bit(TX_STATUS_READING, &rt2x00dev->flags))
203 return; 206 return;
204 207
205 /* Read TX_STA_FIFO register after 500 us */ 208 /* Read TX_STA_FIFO register after 2 ms */
206 hrtimer_start(&rt2x00dev->txstatus_timer, ktime_set(0, 500000), 209 hrtimer_start(&rt2x00dev->txstatus_timer,
210 ktime_set(0, 2*TXSTATUS_READ_INTERVAL),
207 HRTIMER_MODE_REL); 211 HRTIMER_MODE_REL);
208} 212}
209 213
@@ -1176,6 +1180,8 @@ static struct usb_device_id rt2800usb_device_table[] = {
1176 /* Linksys */ 1180 /* Linksys */
1177 { USB_DEVICE(0x13b1, 0x002f) }, 1181 { USB_DEVICE(0x13b1, 0x002f) },
1178 { USB_DEVICE(0x1737, 0x0079) }, 1182 { USB_DEVICE(0x1737, 0x0079) },
1183 /* Logitec */
1184 { USB_DEVICE(0x0789, 0x0170) },
1179 /* Ralink */ 1185 /* Ralink */
1180 { USB_DEVICE(0x148f, 0x3572) }, 1186 { USB_DEVICE(0x148f, 0x3572) },
1181 /* Sitecom */ 1187 /* Sitecom */
@@ -1199,6 +1205,8 @@ static struct usb_device_id rt2800usb_device_table[] = {
1199 { USB_DEVICE(0x050d, 0x1103) }, 1205 { USB_DEVICE(0x050d, 0x1103) },
1200 /* Cameo */ 1206 /* Cameo */
1201 { USB_DEVICE(0x148f, 0xf301) }, 1207 { USB_DEVICE(0x148f, 0xf301) },
1208 /* D-Link */
1209 { USB_DEVICE(0x2001, 0x3c1f) },
1202 /* Edimax */ 1210 /* Edimax */
1203 { USB_DEVICE(0x7392, 0x7733) }, 1211 { USB_DEVICE(0x7392, 0x7733) },
1204 /* Hawking */ 1212 /* Hawking */
@@ -1212,6 +1220,7 @@ static struct usb_device_id rt2800usb_device_table[] = {
1212 { USB_DEVICE(0x0789, 0x016b) }, 1220 { USB_DEVICE(0x0789, 0x016b) },
1213 /* NETGEAR */ 1221 /* NETGEAR */
1214 { USB_DEVICE(0x0846, 0x9012) }, 1222 { USB_DEVICE(0x0846, 0x9012) },
1223 { USB_DEVICE(0x0846, 0x9013) },
1215 { USB_DEVICE(0x0846, 0x9019) }, 1224 { USB_DEVICE(0x0846, 0x9019) },
1216 /* Planex */ 1225 /* Planex */
1217 { USB_DEVICE(0x2019, 0xed19) }, 1226 { USB_DEVICE(0x2019, 0xed19) },
@@ -1220,6 +1229,7 @@ static struct usb_device_id rt2800usb_device_table[] = {
1220 /* Sitecom */ 1229 /* Sitecom */
1221 { USB_DEVICE(0x0df6, 0x0067) }, 1230 { USB_DEVICE(0x0df6, 0x0067) },
1222 { USB_DEVICE(0x0df6, 0x006a) }, 1231 { USB_DEVICE(0x0df6, 0x006a) },
1232 { USB_DEVICE(0x0df6, 0x006e) },
1223 /* ZyXEL */ 1233 /* ZyXEL */
1224 { USB_DEVICE(0x0586, 0x3421) }, 1234 { USB_DEVICE(0x0586, 0x3421) },
1225#endif 1235#endif
@@ -1236,6 +1246,9 @@ static struct usb_device_id rt2800usb_device_table[] = {
1236 { USB_DEVICE(0x2001, 0x3c1c) }, 1246 { USB_DEVICE(0x2001, 0x3c1c) },
1237 { USB_DEVICE(0x2001, 0x3c1d) }, 1247 { USB_DEVICE(0x2001, 0x3c1d) },
1238 { USB_DEVICE(0x2001, 0x3c1e) }, 1248 { USB_DEVICE(0x2001, 0x3c1e) },
1249 { USB_DEVICE(0x2001, 0x3c20) },
1250 { USB_DEVICE(0x2001, 0x3c22) },
1251 { USB_DEVICE(0x2001, 0x3c23) },
1239 /* LG innotek */ 1252 /* LG innotek */
1240 { USB_DEVICE(0x043e, 0x7a22) }, 1253 { USB_DEVICE(0x043e, 0x7a22) },
1241 { USB_DEVICE(0x043e, 0x7a42) }, 1254 { USB_DEVICE(0x043e, 0x7a42) },
@@ -1258,12 +1271,17 @@ static struct usb_device_id rt2800usb_device_table[] = {
1258 { USB_DEVICE(0x043e, 0x7a32) }, 1271 { USB_DEVICE(0x043e, 0x7a32) },
1259 /* AVM GmbH */ 1272 /* AVM GmbH */
1260 { USB_DEVICE(0x057c, 0x8501) }, 1273 { USB_DEVICE(0x057c, 0x8501) },
1261 /* D-Link DWA-160-B2 */ 1274 /* Buffalo */
1275 { USB_DEVICE(0x0411, 0x0241) },
1276 /* D-Link */
1262 { USB_DEVICE(0x2001, 0x3c1a) }, 1277 { USB_DEVICE(0x2001, 0x3c1a) },
1278 { USB_DEVICE(0x2001, 0x3c21) },
1263 /* Proware */ 1279 /* Proware */
1264 { USB_DEVICE(0x043e, 0x7a13) }, 1280 { USB_DEVICE(0x043e, 0x7a13) },
1265 /* Ralink */ 1281 /* Ralink */
1266 { USB_DEVICE(0x148f, 0x5572) }, 1282 { USB_DEVICE(0x148f, 0x5572) },
1283 /* TRENDnet */
1284 { USB_DEVICE(0x20f4, 0x724a) },
1267#endif 1285#endif
1268#ifdef CONFIG_RT2800USB_UNKNOWN 1286#ifdef CONFIG_RT2800USB_UNKNOWN
1269 /* 1287 /*
@@ -1333,6 +1351,7 @@ static struct usb_device_id rt2800usb_device_table[] = {
1333 { USB_DEVICE(0x1d4d, 0x0010) }, 1351 { USB_DEVICE(0x1d4d, 0x0010) },
1334 /* Planex */ 1352 /* Planex */
1335 { USB_DEVICE(0x2019, 0xab24) }, 1353 { USB_DEVICE(0x2019, 0xab24) },
1354 { USB_DEVICE(0x2019, 0xab29) },
1336 /* Qcom */ 1355 /* Qcom */
1337 { USB_DEVICE(0x18e8, 0x6259) }, 1356 { USB_DEVICE(0x18e8, 0x6259) },
1338 /* RadioShack */ 1357 /* RadioShack */
diff --git a/drivers/net/wireless/rt2x00/rt2x00.h b/drivers/net/wireless/rt2x00/rt2x00.h
index fe4c572db52c..e4ba2ce0f212 100644
--- a/drivers/net/wireless/rt2x00/rt2x00.h
+++ b/drivers/net/wireless/rt2x00/rt2x00.h
@@ -39,6 +39,7 @@
39#include <linux/input-polldev.h> 39#include <linux/input-polldev.h>
40#include <linux/kfifo.h> 40#include <linux/kfifo.h>
41#include <linux/hrtimer.h> 41#include <linux/hrtimer.h>
42#include <linux/average.h>
42 43
43#include <net/mac80211.h> 44#include <net/mac80211.h>
44 45
@@ -138,17 +139,6 @@
138#define SHORT_EIFS ( SIFS + SHORT_DIFS + \ 139#define SHORT_EIFS ( SIFS + SHORT_DIFS + \
139 GET_DURATION(IEEE80211_HEADER + ACK_SIZE, 10) ) 140 GET_DURATION(IEEE80211_HEADER + ACK_SIZE, 10) )
140 141
141/*
142 * Structure for average calculation
143 * The avg field contains the actual average value,
144 * but avg_weight is internally used during calculations
145 * to prevent rounding errors.
146 */
147struct avg_val {
148 int avg;
149 int avg_weight;
150};
151
152enum rt2x00_chip_intf { 142enum rt2x00_chip_intf {
153 RT2X00_CHIP_INTF_PCI, 143 RT2X00_CHIP_INTF_PCI,
154 RT2X00_CHIP_INTF_PCIE, 144 RT2X00_CHIP_INTF_PCIE,
@@ -297,7 +287,7 @@ struct link_ant {
297 * Similar to the avg_rssi in the link_qual structure 287 * Similar to the avg_rssi in the link_qual structure
298 * this value is updated by using the walking average. 288 * this value is updated by using the walking average.
299 */ 289 */
300 struct avg_val rssi_ant; 290 struct ewma rssi_ant;
301}; 291};
302 292
303/* 293/*
@@ -326,7 +316,7 @@ struct link {
326 /* 316 /*
327 * Currently active average RSSI value 317 * Currently active average RSSI value
328 */ 318 */
329 struct avg_val avg_rssi; 319 struct ewma avg_rssi;
330 320
331 /* 321 /*
332 * Work structure for scheduling periodic link tuning. 322 * Work structure for scheduling periodic link tuning.
@@ -1179,6 +1169,93 @@ static inline bool rt2x00_is_soc(struct rt2x00_dev *rt2x00dev)
1179 return rt2x00_intf(rt2x00dev, RT2X00_CHIP_INTF_SOC); 1169 return rt2x00_intf(rt2x00dev, RT2X00_CHIP_INTF_SOC);
1180} 1170}
1181 1171
1172/* Helpers for capability flags */
1173
1174static inline bool
1175rt2x00_has_cap_flag(struct rt2x00_dev *rt2x00dev,
1176 enum rt2x00_capability_flags cap_flag)
1177{
1178 return test_bit(cap_flag, &rt2x00dev->cap_flags);
1179}
1180
1181static inline bool
1182rt2x00_has_cap_hw_crypto(struct rt2x00_dev *rt2x00dev)
1183{
1184 return rt2x00_has_cap_flag(rt2x00dev, CAPABILITY_HW_CRYPTO);
1185}
1186
1187static inline bool
1188rt2x00_has_cap_power_limit(struct rt2x00_dev *rt2x00dev)
1189{
1190 return rt2x00_has_cap_flag(rt2x00dev, CAPABILITY_POWER_LIMIT);
1191}
1192
1193static inline bool
1194rt2x00_has_cap_control_filters(struct rt2x00_dev *rt2x00dev)
1195{
1196 return rt2x00_has_cap_flag(rt2x00dev, CAPABILITY_CONTROL_FILTERS);
1197}
1198
1199static inline bool
1200rt2x00_has_cap_control_filter_pspoll(struct rt2x00_dev *rt2x00dev)
1201{
1202 return rt2x00_has_cap_flag(rt2x00dev, CAPABILITY_CONTROL_FILTER_PSPOLL);
1203}
1204
1205static inline bool
1206rt2x00_has_cap_pre_tbtt_interrupt(struct rt2x00_dev *rt2x00dev)
1207{
1208 return rt2x00_has_cap_flag(rt2x00dev, CAPABILITY_PRE_TBTT_INTERRUPT);
1209}
1210
1211static inline bool
1212rt2x00_has_cap_link_tuning(struct rt2x00_dev *rt2x00dev)
1213{
1214 return rt2x00_has_cap_flag(rt2x00dev, CAPABILITY_LINK_TUNING);
1215}
1216
1217static inline bool
1218rt2x00_has_cap_frame_type(struct rt2x00_dev *rt2x00dev)
1219{
1220 return rt2x00_has_cap_flag(rt2x00dev, CAPABILITY_FRAME_TYPE);
1221}
1222
1223static inline bool
1224rt2x00_has_cap_rf_sequence(struct rt2x00_dev *rt2x00dev)
1225{
1226 return rt2x00_has_cap_flag(rt2x00dev, CAPABILITY_RF_SEQUENCE);
1227}
1228
1229static inline bool
1230rt2x00_has_cap_external_lna_a(struct rt2x00_dev *rt2x00dev)
1231{
1232 return rt2x00_has_cap_flag(rt2x00dev, CAPABILITY_EXTERNAL_LNA_A);
1233}
1234
1235static inline bool
1236rt2x00_has_cap_external_lna_bg(struct rt2x00_dev *rt2x00dev)
1237{
1238 return rt2x00_has_cap_flag(rt2x00dev, CAPABILITY_EXTERNAL_LNA_BG);
1239}
1240
1241static inline bool
1242rt2x00_has_cap_double_antenna(struct rt2x00_dev *rt2x00dev)
1243{
1244 return rt2x00_has_cap_flag(rt2x00dev, CAPABILITY_DOUBLE_ANTENNA);
1245}
1246
1247static inline bool
1248rt2x00_has_cap_bt_coexist(struct rt2x00_dev *rt2x00dev)
1249{
1250 return rt2x00_has_cap_flag(rt2x00dev, CAPABILITY_BT_COEXIST);
1251}
1252
1253static inline bool
1254rt2x00_has_cap_vco_recalibration(struct rt2x00_dev *rt2x00dev)
1255{
1256 return rt2x00_has_cap_flag(rt2x00dev, CAPABILITY_VCO_RECALIBRATION);
1257}
1258
1182/** 1259/**
1183 * rt2x00queue_map_txskb - Map a skb into DMA for TX purposes. 1260 * rt2x00queue_map_txskb - Map a skb into DMA for TX purposes.
1184 * @entry: Pointer to &struct queue_entry 1261 * @entry: Pointer to &struct queue_entry
diff --git a/drivers/net/wireless/rt2x00/rt2x00crypto.c b/drivers/net/wireless/rt2x00/rt2x00crypto.c
index 1ca4c7ffc189..3db0d99d9da7 100644
--- a/drivers/net/wireless/rt2x00/rt2x00crypto.c
+++ b/drivers/net/wireless/rt2x00/rt2x00crypto.c
@@ -52,7 +52,7 @@ void rt2x00crypto_create_tx_descriptor(struct rt2x00_dev *rt2x00dev,
52 struct ieee80211_tx_info *tx_info = IEEE80211_SKB_CB(skb); 52 struct ieee80211_tx_info *tx_info = IEEE80211_SKB_CB(skb);
53 struct ieee80211_key_conf *hw_key = tx_info->control.hw_key; 53 struct ieee80211_key_conf *hw_key = tx_info->control.hw_key;
54 54
55 if (!test_bit(CAPABILITY_HW_CRYPTO, &rt2x00dev->cap_flags) || !hw_key) 55 if (!rt2x00_has_cap_hw_crypto(rt2x00dev) || !hw_key)
56 return; 56 return;
57 57
58 __set_bit(ENTRY_TXD_ENCRYPT, &txdesc->flags); 58 __set_bit(ENTRY_TXD_ENCRYPT, &txdesc->flags);
@@ -80,7 +80,7 @@ unsigned int rt2x00crypto_tx_overhead(struct rt2x00_dev *rt2x00dev,
80 struct ieee80211_key_conf *key = tx_info->control.hw_key; 80 struct ieee80211_key_conf *key = tx_info->control.hw_key;
81 unsigned int overhead = 0; 81 unsigned int overhead = 0;
82 82
83 if (!test_bit(CAPABILITY_HW_CRYPTO, &rt2x00dev->cap_flags) || !key) 83 if (!rt2x00_has_cap_hw_crypto(rt2x00dev) || !key)
84 return overhead; 84 return overhead;
85 85
86 /* 86 /*
diff --git a/drivers/net/wireless/rt2x00/rt2x00debug.c b/drivers/net/wireless/rt2x00/rt2x00debug.c
index fe7a7f63a9ed..7f7baae5ae02 100644
--- a/drivers/net/wireless/rt2x00/rt2x00debug.c
+++ b/drivers/net/wireless/rt2x00/rt2x00debug.c
@@ -750,7 +750,7 @@ void rt2x00debug_register(struct rt2x00_dev *rt2x00dev)
750 intf, &rt2x00debug_fop_queue_stats); 750 intf, &rt2x00debug_fop_queue_stats);
751 751
752#ifdef CONFIG_RT2X00_LIB_CRYPTO 752#ifdef CONFIG_RT2X00_LIB_CRYPTO
753 if (test_bit(CAPABILITY_HW_CRYPTO, &rt2x00dev->cap_flags)) 753 if (rt2x00_has_cap_hw_crypto(rt2x00dev))
754 intf->crypto_stats_entry = 754 intf->crypto_stats_entry =
755 debugfs_create_file("crypto", S_IRUGO, intf->queue_folder, 755 debugfs_create_file("crypto", S_IRUGO, intf->queue_folder,
756 intf, &rt2x00debug_fop_crypto_stats); 756 intf, &rt2x00debug_fop_crypto_stats);
diff --git a/drivers/net/wireless/rt2x00/rt2x00dev.c b/drivers/net/wireless/rt2x00/rt2x00dev.c
index 712eea9d398f..080b1fcae5fa 100644
--- a/drivers/net/wireless/rt2x00/rt2x00dev.c
+++ b/drivers/net/wireless/rt2x00/rt2x00dev.c
@@ -88,7 +88,7 @@ int rt2x00lib_enable_radio(struct rt2x00_dev *rt2x00dev)
88 rt2x00queue_start_queues(rt2x00dev); 88 rt2x00queue_start_queues(rt2x00dev);
89 rt2x00link_start_tuner(rt2x00dev); 89 rt2x00link_start_tuner(rt2x00dev);
90 rt2x00link_start_agc(rt2x00dev); 90 rt2x00link_start_agc(rt2x00dev);
91 if (test_bit(CAPABILITY_VCO_RECALIBRATION, &rt2x00dev->cap_flags)) 91 if (rt2x00_has_cap_vco_recalibration(rt2x00dev))
92 rt2x00link_start_vcocal(rt2x00dev); 92 rt2x00link_start_vcocal(rt2x00dev);
93 93
94 /* 94 /*
@@ -113,7 +113,7 @@ void rt2x00lib_disable_radio(struct rt2x00_dev *rt2x00dev)
113 * Stop all queues 113 * Stop all queues
114 */ 114 */
115 rt2x00link_stop_agc(rt2x00dev); 115 rt2x00link_stop_agc(rt2x00dev);
116 if (test_bit(CAPABILITY_VCO_RECALIBRATION, &rt2x00dev->cap_flags)) 116 if (rt2x00_has_cap_vco_recalibration(rt2x00dev))
117 rt2x00link_stop_vcocal(rt2x00dev); 117 rt2x00link_stop_vcocal(rt2x00dev);
118 rt2x00link_stop_tuner(rt2x00dev); 118 rt2x00link_stop_tuner(rt2x00dev);
119 rt2x00queue_stop_queues(rt2x00dev); 119 rt2x00queue_stop_queues(rt2x00dev);
@@ -234,7 +234,7 @@ void rt2x00lib_beacondone(struct rt2x00_dev *rt2x00dev)
234 * here as they will fetch the next beacon directly prior to 234 * here as they will fetch the next beacon directly prior to
235 * transmission. 235 * transmission.
236 */ 236 */
237 if (test_bit(CAPABILITY_PRE_TBTT_INTERRUPT, &rt2x00dev->cap_flags)) 237 if (rt2x00_has_cap_pre_tbtt_interrupt(rt2x00dev))
238 return; 238 return;
239 239
240 /* fetch next beacon */ 240 /* fetch next beacon */
@@ -358,7 +358,7 @@ void rt2x00lib_txdone(struct queue_entry *entry,
358 * mac80211 will expect the same data to be present it the 358 * mac80211 will expect the same data to be present it the
359 * frame as it was passed to us. 359 * frame as it was passed to us.
360 */ 360 */
361 if (test_bit(CAPABILITY_HW_CRYPTO, &rt2x00dev->cap_flags)) 361 if (rt2x00_has_cap_hw_crypto(rt2x00dev))
362 rt2x00crypto_tx_insert_iv(entry->skb, header_length); 362 rt2x00crypto_tx_insert_iv(entry->skb, header_length);
363 363
364 /* 364 /*
diff --git a/drivers/net/wireless/rt2x00/rt2x00link.c b/drivers/net/wireless/rt2x00/rt2x00link.c
index 8368aab86f28..c2b3b6629188 100644
--- a/drivers/net/wireless/rt2x00/rt2x00link.c
+++ b/drivers/net/wireless/rt2x00/rt2x00link.c
@@ -35,50 +35,28 @@
35 */ 35 */
36#define DEFAULT_RSSI -128 36#define DEFAULT_RSSI -128
37 37
38/* 38/* Constants for EWMA calculations. */
39 * Helper struct and macro to work with moving/walking averages. 39#define RT2X00_EWMA_FACTOR 1024
40 * When adding a value to the average value the following calculation 40#define RT2X00_EWMA_WEIGHT 8
41 * is needed: 41
42 * 42static inline int rt2x00link_get_avg_rssi(struct ewma *ewma)
43 * avg_rssi = ((avg_rssi * 7) + rssi) / 8; 43{
44 * 44 unsigned long avg;
45 * The advantage of this approach is that we only need 1 variable 45
46 * to store the average in (No need for a count and a total). 46 avg = ewma_read(ewma);
47 * But more importantly, normal average values will over time 47 if (avg)
48 * move less and less towards newly added values this results 48 return -avg;
49 * that with link tuning, the device can have a very good RSSI 49
50 * for a few minutes but when the device is moved away from the AP 50 return DEFAULT_RSSI;
51 * the average will not decrease fast enough to compensate. 51}
52 * The walking average compensates this and will move towards
53 * the new values correctly allowing a effective link tuning,
54 * the speed of the average moving towards other values depends
55 * on the value for the number of samples. The higher the number
56 * of samples, the slower the average will move.
57 * We use two variables to keep track of the average value to
58 * compensate for the rounding errors. This can be a significant
59 * error (>5dBm) if the factor is too low.
60 */
61#define AVG_SAMPLES 8
62#define AVG_FACTOR 1000
63#define MOVING_AVERAGE(__avg, __val) \
64({ \
65 struct avg_val __new; \
66 __new.avg_weight = \
67 (__avg).avg_weight ? \
68 ((((__avg).avg_weight * ((AVG_SAMPLES) - 1)) + \
69 ((__val) * (AVG_FACTOR))) / \
70 (AVG_SAMPLES)) : \
71 ((__val) * (AVG_FACTOR)); \
72 __new.avg = __new.avg_weight / (AVG_FACTOR); \
73 __new; \
74})
75 52
76static int rt2x00link_antenna_get_link_rssi(struct rt2x00_dev *rt2x00dev) 53static int rt2x00link_antenna_get_link_rssi(struct rt2x00_dev *rt2x00dev)
77{ 54{
78 struct link_ant *ant = &rt2x00dev->link.ant; 55 struct link_ant *ant = &rt2x00dev->link.ant;
79 56
80 if (ant->rssi_ant.avg && rt2x00dev->link.qual.rx_success) 57 if (rt2x00dev->link.qual.rx_success)
81 return ant->rssi_ant.avg; 58 return rt2x00link_get_avg_rssi(&ant->rssi_ant);
59
82 return DEFAULT_RSSI; 60 return DEFAULT_RSSI;
83} 61}
84 62
@@ -100,8 +78,8 @@ static void rt2x00link_antenna_update_rssi_history(struct rt2x00_dev *rt2x00dev,
100 78
101static void rt2x00link_antenna_reset(struct rt2x00_dev *rt2x00dev) 79static void rt2x00link_antenna_reset(struct rt2x00_dev *rt2x00dev)
102{ 80{
103 rt2x00dev->link.ant.rssi_ant.avg = 0; 81 ewma_init(&rt2x00dev->link.ant.rssi_ant, RT2X00_EWMA_FACTOR,
104 rt2x00dev->link.ant.rssi_ant.avg_weight = 0; 82 RT2X00_EWMA_WEIGHT);
105} 83}
106 84
107static void rt2x00lib_antenna_diversity_sample(struct rt2x00_dev *rt2x00dev) 85static void rt2x00lib_antenna_diversity_sample(struct rt2x00_dev *rt2x00dev)
@@ -249,12 +227,12 @@ void rt2x00link_update_stats(struct rt2x00_dev *rt2x00dev,
249 /* 227 /*
250 * Update global RSSI 228 * Update global RSSI
251 */ 229 */
252 link->avg_rssi = MOVING_AVERAGE(link->avg_rssi, rxdesc->rssi); 230 ewma_add(&link->avg_rssi, -rxdesc->rssi);
253 231
254 /* 232 /*
255 * Update antenna RSSI 233 * Update antenna RSSI
256 */ 234 */
257 ant->rssi_ant = MOVING_AVERAGE(ant->rssi_ant, rxdesc->rssi); 235 ewma_add(&ant->rssi_ant, -rxdesc->rssi);
258} 236}
259 237
260void rt2x00link_start_tuner(struct rt2x00_dev *rt2x00dev) 238void rt2x00link_start_tuner(struct rt2x00_dev *rt2x00dev)
@@ -309,6 +287,8 @@ void rt2x00link_reset_tuner(struct rt2x00_dev *rt2x00dev, bool antenna)
309 */ 287 */
310 rt2x00dev->link.count = 0; 288 rt2x00dev->link.count = 0;
311 memset(qual, 0, sizeof(*qual)); 289 memset(qual, 0, sizeof(*qual));
290 ewma_init(&rt2x00dev->link.avg_rssi, RT2X00_EWMA_FACTOR,
291 RT2X00_EWMA_WEIGHT);
312 292
313 /* 293 /*
314 * Restore the VGC level as stored in the registers, 294 * Restore the VGC level as stored in the registers,
@@ -363,17 +343,17 @@ static void rt2x00link_tuner(struct work_struct *work)
363 * collect the RSSI data we could use this. Otherwise we 343 * collect the RSSI data we could use this. Otherwise we
364 * must fallback to the default RSSI value. 344 * must fallback to the default RSSI value.
365 */ 345 */
366 if (!link->avg_rssi.avg || !qual->rx_success) 346 if (!qual->rx_success)
367 qual->rssi = DEFAULT_RSSI; 347 qual->rssi = DEFAULT_RSSI;
368 else 348 else
369 qual->rssi = link->avg_rssi.avg; 349 qual->rssi = rt2x00link_get_avg_rssi(&link->avg_rssi);
370 350
371 /* 351 /*
372 * Check if link tuning is supported by the hardware, some hardware 352 * Check if link tuning is supported by the hardware, some hardware
373 * do not support link tuning at all, while other devices can disable 353 * do not support link tuning at all, while other devices can disable
374 * the feature from the EEPROM. 354 * the feature from the EEPROM.
375 */ 355 */
376 if (test_bit(CAPABILITY_LINK_TUNING, &rt2x00dev->cap_flags)) 356 if (rt2x00_has_cap_link_tuning(rt2x00dev))
377 rt2x00dev->ops->lib->link_tuner(rt2x00dev, qual, link->count); 357 rt2x00dev->ops->lib->link_tuner(rt2x00dev, qual, link->count);
378 358
379 /* 359 /*
@@ -513,7 +493,7 @@ static void rt2x00link_vcocal(struct work_struct *work)
513void rt2x00link_register(struct rt2x00_dev *rt2x00dev) 493void rt2x00link_register(struct rt2x00_dev *rt2x00dev)
514{ 494{
515 INIT_DELAYED_WORK(&rt2x00dev->link.agc_work, rt2x00link_agc); 495 INIT_DELAYED_WORK(&rt2x00dev->link.agc_work, rt2x00link_agc);
516 if (test_bit(CAPABILITY_VCO_RECALIBRATION, &rt2x00dev->cap_flags)) 496 if (rt2x00_has_cap_vco_recalibration(rt2x00dev))
517 INIT_DELAYED_WORK(&rt2x00dev->link.vco_work, rt2x00link_vcocal); 497 INIT_DELAYED_WORK(&rt2x00dev->link.vco_work, rt2x00link_vcocal);
518 INIT_DELAYED_WORK(&rt2x00dev->link.watchdog_work, rt2x00link_watchdog); 498 INIT_DELAYED_WORK(&rt2x00dev->link.watchdog_work, rt2x00link_watchdog);
519 INIT_DELAYED_WORK(&rt2x00dev->link.work, rt2x00link_tuner); 499 INIT_DELAYED_WORK(&rt2x00dev->link.work, rt2x00link_tuner);
diff --git a/drivers/net/wireless/rt2x00/rt2x00mac.c b/drivers/net/wireless/rt2x00/rt2x00mac.c
index f883802f3505..7c157857f5ce 100644
--- a/drivers/net/wireless/rt2x00/rt2x00mac.c
+++ b/drivers/net/wireless/rt2x00/rt2x00mac.c
@@ -382,11 +382,11 @@ void rt2x00mac_configure_filter(struct ieee80211_hw *hw,
382 * of different types, but has no a separate filter for PS Poll frames, 382 * of different types, but has no a separate filter for PS Poll frames,
383 * FIF_CONTROL flag implies FIF_PSPOLL. 383 * FIF_CONTROL flag implies FIF_PSPOLL.
384 */ 384 */
385 if (!test_bit(CAPABILITY_CONTROL_FILTERS, &rt2x00dev->cap_flags)) { 385 if (!rt2x00_has_cap_control_filters(rt2x00dev)) {
386 if (*total_flags & FIF_CONTROL || *total_flags & FIF_PSPOLL) 386 if (*total_flags & FIF_CONTROL || *total_flags & FIF_PSPOLL)
387 *total_flags |= FIF_CONTROL | FIF_PSPOLL; 387 *total_flags |= FIF_CONTROL | FIF_PSPOLL;
388 } 388 }
389 if (!test_bit(CAPABILITY_CONTROL_FILTER_PSPOLL, &rt2x00dev->cap_flags)) { 389 if (!rt2x00_has_cap_control_filter_pspoll(rt2x00dev)) {
390 if (*total_flags & FIF_CONTROL) 390 if (*total_flags & FIF_CONTROL)
391 *total_flags |= FIF_PSPOLL; 391 *total_flags |= FIF_PSPOLL;
392 } 392 }
@@ -469,7 +469,7 @@ int rt2x00mac_set_key(struct ieee80211_hw *hw, enum set_key_cmd cmd,
469 if (!test_bit(DEVICE_STATE_PRESENT, &rt2x00dev->flags)) 469 if (!test_bit(DEVICE_STATE_PRESENT, &rt2x00dev->flags))
470 return 0; 470 return 0;
471 471
472 if (!test_bit(CAPABILITY_HW_CRYPTO, &rt2x00dev->cap_flags)) 472 if (!rt2x00_has_cap_hw_crypto(rt2x00dev))
473 return -EOPNOTSUPP; 473 return -EOPNOTSUPP;
474 474
475 /* 475 /*
@@ -754,6 +754,9 @@ void rt2x00mac_flush(struct ieee80211_hw *hw, u32 queues, bool drop)
754 struct rt2x00_dev *rt2x00dev = hw->priv; 754 struct rt2x00_dev *rt2x00dev = hw->priv;
755 struct data_queue *queue; 755 struct data_queue *queue;
756 756
757 if (!test_bit(DEVICE_STATE_PRESENT, &rt2x00dev->flags))
758 return;
759
757 tx_queue_for_each(rt2x00dev, queue) 760 tx_queue_for_each(rt2x00dev, queue)
758 rt2x00queue_flush_queue(queue, drop); 761 rt2x00queue_flush_queue(queue, drop);
759} 762}
diff --git a/drivers/net/wireless/rt2x00/rt2x00pci.c b/drivers/net/wireless/rt2x00/rt2x00pci.c
index dc49e525ae5e..25da20e7e1f3 100644
--- a/drivers/net/wireless/rt2x00/rt2x00pci.c
+++ b/drivers/net/wireless/rt2x00/rt2x00pci.c
@@ -119,7 +119,7 @@ int rt2x00pci_probe(struct pci_dev *pci_dev, const struct rt2x00_ops *ops)
119 rt2x00dev->ops = ops; 119 rt2x00dev->ops = ops;
120 rt2x00dev->hw = hw; 120 rt2x00dev->hw = hw;
121 rt2x00dev->irq = pci_dev->irq; 121 rt2x00dev->irq = pci_dev->irq;
122 rt2x00dev->name = pci_name(pci_dev); 122 rt2x00dev->name = ops->name;
123 123
124 if (pci_is_pcie(pci_dev)) 124 if (pci_is_pcie(pci_dev))
125 rt2x00_set_chip_intf(rt2x00dev, RT2X00_CHIP_INTF_PCIE); 125 rt2x00_set_chip_intf(rt2x00dev, RT2X00_CHIP_INTF_PCIE);
diff --git a/drivers/net/wireless/rt2x00/rt2x00queue.c b/drivers/net/wireless/rt2x00/rt2x00queue.c
index 6c8a33b6ee22..50590b1420a5 100644
--- a/drivers/net/wireless/rt2x00/rt2x00queue.c
+++ b/drivers/net/wireless/rt2x00/rt2x00queue.c
@@ -61,7 +61,7 @@ struct sk_buff *rt2x00queue_alloc_rxskb(struct queue_entry *entry, gfp_t gfp)
61 * at least 8 bytes bytes available in headroom for IV/EIV 61 * at least 8 bytes bytes available in headroom for IV/EIV
62 * and 8 bytes for ICV data as tailroon. 62 * and 8 bytes for ICV data as tailroon.
63 */ 63 */
64 if (test_bit(CAPABILITY_HW_CRYPTO, &rt2x00dev->cap_flags)) { 64 if (rt2x00_has_cap_hw_crypto(rt2x00dev)) {
65 head_size += 8; 65 head_size += 8;
66 tail_size += 8; 66 tail_size += 8;
67 } 67 }
@@ -1033,38 +1033,21 @@ EXPORT_SYMBOL_GPL(rt2x00queue_stop_queue);
1033 1033
1034void rt2x00queue_flush_queue(struct data_queue *queue, bool drop) 1034void rt2x00queue_flush_queue(struct data_queue *queue, bool drop)
1035{ 1035{
1036 bool started;
1037 bool tx_queue = 1036 bool tx_queue =
1038 (queue->qid == QID_AC_VO) || 1037 (queue->qid == QID_AC_VO) ||
1039 (queue->qid == QID_AC_VI) || 1038 (queue->qid == QID_AC_VI) ||
1040 (queue->qid == QID_AC_BE) || 1039 (queue->qid == QID_AC_BE) ||
1041 (queue->qid == QID_AC_BK); 1040 (queue->qid == QID_AC_BK);
1042 1041
1043 mutex_lock(&queue->status_lock);
1044 1042
1045 /* 1043 /*
1046 * If the queue has been started, we must stop it temporarily 1044 * If we are not supposed to drop any pending
1047 * to prevent any new frames to be queued on the device. If 1045 * frames, this means we must force a start (=kick)
1048 * we are not dropping the pending frames, the queue must 1046 * to the queue to make sure the hardware will
1049 * only be stopped in the software and not the hardware, 1047 * start transmitting.
1050 * otherwise the queue will never become empty on its own.
1051 */ 1048 */
1052 started = test_bit(QUEUE_STARTED, &queue->flags); 1049 if (!drop && tx_queue)
1053 if (started) { 1050 queue->rt2x00dev->ops->lib->kick_queue(queue);
1054 /*
1055 * Pause the queue
1056 */
1057 rt2x00queue_pause_queue(queue);
1058
1059 /*
1060 * If we are not supposed to drop any pending
1061 * frames, this means we must force a start (=kick)
1062 * to the queue to make sure the hardware will
1063 * start transmitting.
1064 */
1065 if (!drop && tx_queue)
1066 queue->rt2x00dev->ops->lib->kick_queue(queue);
1067 }
1068 1051
1069 /* 1052 /*
1070 * Check if driver supports flushing, if that is the case we can 1053 * Check if driver supports flushing, if that is the case we can
@@ -1080,14 +1063,6 @@ void rt2x00queue_flush_queue(struct data_queue *queue, bool drop)
1080 if (unlikely(!rt2x00queue_empty(queue))) 1063 if (unlikely(!rt2x00queue_empty(queue)))
1081 rt2x00_warn(queue->rt2x00dev, "Queue %d failed to flush\n", 1064 rt2x00_warn(queue->rt2x00dev, "Queue %d failed to flush\n",
1082 queue->qid); 1065 queue->qid);
1083
1084 /*
1085 * Restore the queue to the previous status
1086 */
1087 if (started)
1088 rt2x00queue_unpause_queue(queue);
1089
1090 mutex_unlock(&queue->status_lock);
1091} 1066}
1092EXPORT_SYMBOL_GPL(rt2x00queue_flush_queue); 1067EXPORT_SYMBOL_GPL(rt2x00queue_flush_queue);
1093 1068
diff --git a/drivers/net/wireless/rt2x00/rt2x00usb.c b/drivers/net/wireless/rt2x00/rt2x00usb.c
index 88289873c0cf..4e121627925d 100644
--- a/drivers/net/wireless/rt2x00/rt2x00usb.c
+++ b/drivers/net/wireless/rt2x00/rt2x00usb.c
@@ -523,7 +523,9 @@ static void rt2x00usb_watchdog_tx_dma(struct data_queue *queue)
523 rt2x00_warn(queue->rt2x00dev, "TX queue %d DMA timed out, invoke forced forced reset\n", 523 rt2x00_warn(queue->rt2x00dev, "TX queue %d DMA timed out, invoke forced forced reset\n",
524 queue->qid); 524 queue->qid);
525 525
526 rt2x00queue_stop_queue(queue);
526 rt2x00queue_flush_queue(queue, true); 527 rt2x00queue_flush_queue(queue, true);
528 rt2x00queue_start_queue(queue);
527} 529}
528 530
529static int rt2x00usb_dma_timeout(struct data_queue *queue) 531static int rt2x00usb_dma_timeout(struct data_queue *queue)
diff --git a/drivers/net/wireless/rt2x00/rt61pci.c b/drivers/net/wireless/rt2x00/rt61pci.c
index 54d3ddfc9888..a5b69cb49012 100644
--- a/drivers/net/wireless/rt2x00/rt61pci.c
+++ b/drivers/net/wireless/rt2x00/rt61pci.c
@@ -685,7 +685,7 @@ static void rt61pci_config_antenna_2x(struct rt2x00_dev *rt2x00dev,
685 685
686 rt2x00_set_field8(&r3, BBP_R3_SMART_MODE, rt2x00_rf(rt2x00dev, RF2529)); 686 rt2x00_set_field8(&r3, BBP_R3_SMART_MODE, rt2x00_rf(rt2x00dev, RF2529));
687 rt2x00_set_field8(&r4, BBP_R4_RX_FRAME_END, 687 rt2x00_set_field8(&r4, BBP_R4_RX_FRAME_END,
688 !test_bit(CAPABILITY_FRAME_TYPE, &rt2x00dev->cap_flags)); 688 !rt2x00_has_cap_frame_type(rt2x00dev));
689 689
690 /* 690 /*
691 * Configure the RX antenna. 691 * Configure the RX antenna.
@@ -813,10 +813,10 @@ static void rt61pci_config_ant(struct rt2x00_dev *rt2x00dev,
813 813
814 if (rt2x00dev->curr_band == IEEE80211_BAND_5GHZ) { 814 if (rt2x00dev->curr_band == IEEE80211_BAND_5GHZ) {
815 sel = antenna_sel_a; 815 sel = antenna_sel_a;
816 lna = test_bit(CAPABILITY_EXTERNAL_LNA_A, &rt2x00dev->cap_flags); 816 lna = rt2x00_has_cap_external_lna_a(rt2x00dev);
817 } else { 817 } else {
818 sel = antenna_sel_bg; 818 sel = antenna_sel_bg;
819 lna = test_bit(CAPABILITY_EXTERNAL_LNA_BG, &rt2x00dev->cap_flags); 819 lna = rt2x00_has_cap_external_lna_bg(rt2x00dev);
820 } 820 }
821 821
822 for (i = 0; i < ARRAY_SIZE(antenna_sel_a); i++) 822 for (i = 0; i < ARRAY_SIZE(antenna_sel_a); i++)
@@ -836,7 +836,7 @@ static void rt61pci_config_ant(struct rt2x00_dev *rt2x00dev,
836 else if (rt2x00_rf(rt2x00dev, RF2527)) 836 else if (rt2x00_rf(rt2x00dev, RF2527))
837 rt61pci_config_antenna_2x(rt2x00dev, ant); 837 rt61pci_config_antenna_2x(rt2x00dev, ant);
838 else if (rt2x00_rf(rt2x00dev, RF2529)) { 838 else if (rt2x00_rf(rt2x00dev, RF2529)) {
839 if (test_bit(CAPABILITY_DOUBLE_ANTENNA, &rt2x00dev->cap_flags)) 839 if (rt2x00_has_cap_double_antenna(rt2x00dev))
840 rt61pci_config_antenna_2x(rt2x00dev, ant); 840 rt61pci_config_antenna_2x(rt2x00dev, ant);
841 else 841 else
842 rt61pci_config_antenna_2529(rt2x00dev, ant); 842 rt61pci_config_antenna_2529(rt2x00dev, ant);
@@ -850,13 +850,13 @@ static void rt61pci_config_lna_gain(struct rt2x00_dev *rt2x00dev,
850 short lna_gain = 0; 850 short lna_gain = 0;
851 851
852 if (libconf->conf->chandef.chan->band == IEEE80211_BAND_2GHZ) { 852 if (libconf->conf->chandef.chan->band == IEEE80211_BAND_2GHZ) {
853 if (test_bit(CAPABILITY_EXTERNAL_LNA_BG, &rt2x00dev->cap_flags)) 853 if (rt2x00_has_cap_external_lna_bg(rt2x00dev))
854 lna_gain += 14; 854 lna_gain += 14;
855 855
856 rt2x00_eeprom_read(rt2x00dev, EEPROM_RSSI_OFFSET_BG, &eeprom); 856 rt2x00_eeprom_read(rt2x00dev, EEPROM_RSSI_OFFSET_BG, &eeprom);
857 lna_gain -= rt2x00_get_field16(eeprom, EEPROM_RSSI_OFFSET_BG_1); 857 lna_gain -= rt2x00_get_field16(eeprom, EEPROM_RSSI_OFFSET_BG_1);
858 } else { 858 } else {
859 if (test_bit(CAPABILITY_EXTERNAL_LNA_A, &rt2x00dev->cap_flags)) 859 if (rt2x00_has_cap_external_lna_a(rt2x00dev))
860 lna_gain += 14; 860 lna_gain += 14;
861 861
862 rt2x00_eeprom_read(rt2x00dev, EEPROM_RSSI_OFFSET_A, &eeprom); 862 rt2x00_eeprom_read(rt2x00dev, EEPROM_RSSI_OFFSET_A, &eeprom);
@@ -1054,14 +1054,14 @@ static void rt61pci_link_tuner(struct rt2x00_dev *rt2x00dev,
1054 if (rt2x00dev->curr_band == IEEE80211_BAND_5GHZ) { 1054 if (rt2x00dev->curr_band == IEEE80211_BAND_5GHZ) {
1055 low_bound = 0x28; 1055 low_bound = 0x28;
1056 up_bound = 0x48; 1056 up_bound = 0x48;
1057 if (test_bit(CAPABILITY_EXTERNAL_LNA_A, &rt2x00dev->cap_flags)) { 1057 if (rt2x00_has_cap_external_lna_a(rt2x00dev)) {
1058 low_bound += 0x10; 1058 low_bound += 0x10;
1059 up_bound += 0x10; 1059 up_bound += 0x10;
1060 } 1060 }
1061 } else { 1061 } else {
1062 low_bound = 0x20; 1062 low_bound = 0x20;
1063 up_bound = 0x40; 1063 up_bound = 0x40;
1064 if (test_bit(CAPABILITY_EXTERNAL_LNA_BG, &rt2x00dev->cap_flags)) { 1064 if (rt2x00_has_cap_external_lna_bg(rt2x00dev)) {
1065 low_bound += 0x10; 1065 low_bound += 0x10;
1066 up_bound += 0x10; 1066 up_bound += 0x10;
1067 } 1067 }
@@ -2578,7 +2578,7 @@ static int rt61pci_init_eeprom(struct rt2x00_dev *rt2x00dev)
2578 * eeprom word. 2578 * eeprom word.
2579 */ 2579 */
2580 if (rt2x00_rf(rt2x00dev, RF2529) && 2580 if (rt2x00_rf(rt2x00dev, RF2529) &&
2581 !test_bit(CAPABILITY_DOUBLE_ANTENNA, &rt2x00dev->cap_flags)) { 2581 !rt2x00_has_cap_double_antenna(rt2x00dev)) {
2582 rt2x00dev->default_ant.rx = 2582 rt2x00dev->default_ant.rx =
2583 ANTENNA_A + rt2x00_get_field16(eeprom, EEPROM_NIC_RX_FIXED); 2583 ANTENNA_A + rt2x00_get_field16(eeprom, EEPROM_NIC_RX_FIXED);
2584 rt2x00dev->default_ant.tx = 2584 rt2x00dev->default_ant.tx =
@@ -2793,7 +2793,7 @@ static int rt61pci_probe_hw_mode(struct rt2x00_dev *rt2x00dev)
2793 spec->supported_bands = SUPPORT_BAND_2GHZ; 2793 spec->supported_bands = SUPPORT_BAND_2GHZ;
2794 spec->supported_rates = SUPPORT_RATE_CCK | SUPPORT_RATE_OFDM; 2794 spec->supported_rates = SUPPORT_RATE_CCK | SUPPORT_RATE_OFDM;
2795 2795
2796 if (!test_bit(CAPABILITY_RF_SEQUENCE, &rt2x00dev->cap_flags)) { 2796 if (!rt2x00_has_cap_rf_sequence(rt2x00dev)) {
2797 spec->num_channels = 14; 2797 spec->num_channels = 14;
2798 spec->channels = rf_vals_noseq; 2798 spec->channels = rf_vals_noseq;
2799 } else { 2799 } else {
diff --git a/drivers/net/wireless/rt2x00/rt73usb.c b/drivers/net/wireless/rt2x00/rt73usb.c
index 1d3880e09a13..1baf9c896dcd 100644
--- a/drivers/net/wireless/rt2x00/rt73usb.c
+++ b/drivers/net/wireless/rt2x00/rt73usb.c
@@ -595,8 +595,8 @@ static void rt73usb_config_antenna_5x(struct rt2x00_dev *rt2x00dev,
595 switch (ant->rx) { 595 switch (ant->rx) {
596 case ANTENNA_HW_DIVERSITY: 596 case ANTENNA_HW_DIVERSITY:
597 rt2x00_set_field8(&r4, BBP_R4_RX_ANTENNA_CONTROL, 2); 597 rt2x00_set_field8(&r4, BBP_R4_RX_ANTENNA_CONTROL, 2);
598 temp = !test_bit(CAPABILITY_FRAME_TYPE, &rt2x00dev->cap_flags) 598 temp = !rt2x00_has_cap_frame_type(rt2x00dev) &&
599 && (rt2x00dev->curr_band != IEEE80211_BAND_5GHZ); 599 (rt2x00dev->curr_band != IEEE80211_BAND_5GHZ);
600 rt2x00_set_field8(&r4, BBP_R4_RX_FRAME_END, temp); 600 rt2x00_set_field8(&r4, BBP_R4_RX_FRAME_END, temp);
601 break; 601 break;
602 case ANTENNA_A: 602 case ANTENNA_A:
@@ -636,7 +636,7 @@ static void rt73usb_config_antenna_2x(struct rt2x00_dev *rt2x00dev,
636 636
637 rt2x00_set_field8(&r3, BBP_R3_SMART_MODE, 0); 637 rt2x00_set_field8(&r3, BBP_R3_SMART_MODE, 0);
638 rt2x00_set_field8(&r4, BBP_R4_RX_FRAME_END, 638 rt2x00_set_field8(&r4, BBP_R4_RX_FRAME_END,
639 !test_bit(CAPABILITY_FRAME_TYPE, &rt2x00dev->cap_flags)); 639 !rt2x00_has_cap_frame_type(rt2x00dev));
640 640
641 /* 641 /*
642 * Configure the RX antenna. 642 * Configure the RX antenna.
@@ -709,10 +709,10 @@ static void rt73usb_config_ant(struct rt2x00_dev *rt2x00dev,
709 709
710 if (rt2x00dev->curr_band == IEEE80211_BAND_5GHZ) { 710 if (rt2x00dev->curr_band == IEEE80211_BAND_5GHZ) {
711 sel = antenna_sel_a; 711 sel = antenna_sel_a;
712 lna = test_bit(CAPABILITY_EXTERNAL_LNA_A, &rt2x00dev->cap_flags); 712 lna = rt2x00_has_cap_external_lna_a(rt2x00dev);
713 } else { 713 } else {
714 sel = antenna_sel_bg; 714 sel = antenna_sel_bg;
715 lna = test_bit(CAPABILITY_EXTERNAL_LNA_BG, &rt2x00dev->cap_flags); 715 lna = rt2x00_has_cap_external_lna_bg(rt2x00dev);
716 } 716 }
717 717
718 for (i = 0; i < ARRAY_SIZE(antenna_sel_a); i++) 718 for (i = 0; i < ARRAY_SIZE(antenna_sel_a); i++)
@@ -740,7 +740,7 @@ static void rt73usb_config_lna_gain(struct rt2x00_dev *rt2x00dev,
740 short lna_gain = 0; 740 short lna_gain = 0;
741 741
742 if (libconf->conf->chandef.chan->band == IEEE80211_BAND_2GHZ) { 742 if (libconf->conf->chandef.chan->band == IEEE80211_BAND_2GHZ) {
743 if (test_bit(CAPABILITY_EXTERNAL_LNA_BG, &rt2x00dev->cap_flags)) 743 if (rt2x00_has_cap_external_lna_bg(rt2x00dev))
744 lna_gain += 14; 744 lna_gain += 14;
745 745
746 rt2x00_eeprom_read(rt2x00dev, EEPROM_RSSI_OFFSET_BG, &eeprom); 746 rt2x00_eeprom_read(rt2x00dev, EEPROM_RSSI_OFFSET_BG, &eeprom);
@@ -930,7 +930,7 @@ static void rt73usb_link_tuner(struct rt2x00_dev *rt2x00dev,
930 low_bound = 0x28; 930 low_bound = 0x28;
931 up_bound = 0x48; 931 up_bound = 0x48;
932 932
933 if (test_bit(CAPABILITY_EXTERNAL_LNA_A, &rt2x00dev->cap_flags)) { 933 if (rt2x00_has_cap_external_lna_a(rt2x00dev)) {
934 low_bound += 0x10; 934 low_bound += 0x10;
935 up_bound += 0x10; 935 up_bound += 0x10;
936 } 936 }
@@ -946,7 +946,7 @@ static void rt73usb_link_tuner(struct rt2x00_dev *rt2x00dev,
946 up_bound = 0x1c; 946 up_bound = 0x1c;
947 } 947 }
948 948
949 if (test_bit(CAPABILITY_EXTERNAL_LNA_BG, &rt2x00dev->cap_flags)) { 949 if (rt2x00_has_cap_external_lna_bg(rt2x00dev)) {
950 low_bound += 0x14; 950 low_bound += 0x14;
951 up_bound += 0x10; 951 up_bound += 0x10;
952 } 952 }
@@ -1661,7 +1661,7 @@ static int rt73usb_agc_to_rssi(struct rt2x00_dev *rt2x00dev, int rxd_w1)
1661 } 1661 }
1662 1662
1663 if (rt2x00dev->curr_band == IEEE80211_BAND_5GHZ) { 1663 if (rt2x00dev->curr_band == IEEE80211_BAND_5GHZ) {
1664 if (test_bit(CAPABILITY_EXTERNAL_LNA_A, &rt2x00dev->cap_flags)) { 1664 if (rt2x00_has_cap_external_lna_a(rt2x00dev)) {
1665 if (lna == 3 || lna == 2) 1665 if (lna == 3 || lna == 2)
1666 offset += 10; 1666 offset += 10;
1667 } else { 1667 } else {
diff --git a/drivers/net/wireless/rtl818x/rtl8180/dev.c b/drivers/net/wireless/rtl818x/rtl8180/dev.c
index fc207b268e4f..a91506b12a62 100644
--- a/drivers/net/wireless/rtl818x/rtl8180/dev.c
+++ b/drivers/net/wireless/rtl818x/rtl8180/dev.c
@@ -1122,7 +1122,6 @@ static int rtl8180_probe(struct pci_dev *pdev,
1122 iounmap(priv->map); 1122 iounmap(priv->map);
1123 1123
1124 err_free_dev: 1124 err_free_dev:
1125 pci_set_drvdata(pdev, NULL);
1126 ieee80211_free_hw(dev); 1125 ieee80211_free_hw(dev);
1127 1126
1128 err_free_reg: 1127 err_free_reg:
diff --git a/drivers/net/wireless/rtlwifi/base.c b/drivers/net/wireless/rtlwifi/base.c
index 8bb4a9a01a18..9a78e3daf742 100644
--- a/drivers/net/wireless/rtlwifi/base.c
+++ b/drivers/net/wireless/rtlwifi/base.c
@@ -1613,6 +1613,35 @@ err_free:
1613} 1613}
1614EXPORT_SYMBOL(rtl_send_smps_action); 1614EXPORT_SYMBOL(rtl_send_smps_action);
1615 1615
1616void rtl_phy_scan_operation_backup(struct ieee80211_hw *hw, u8 operation)
1617{
1618 struct rtl_priv *rtlpriv = rtl_priv(hw);
1619 struct rtl_hal *rtlhal = rtl_hal(rtl_priv(hw));
1620 enum io_type iotype;
1621
1622 if (!is_hal_stop(rtlhal)) {
1623 switch (operation) {
1624 case SCAN_OPT_BACKUP:
1625 iotype = IO_CMD_PAUSE_DM_BY_SCAN;
1626 rtlpriv->cfg->ops->set_hw_reg(hw,
1627 HW_VAR_IO_CMD,
1628 (u8 *)&iotype);
1629 break;
1630 case SCAN_OPT_RESTORE:
1631 iotype = IO_CMD_RESUME_DM_BY_SCAN;
1632 rtlpriv->cfg->ops->set_hw_reg(hw,
1633 HW_VAR_IO_CMD,
1634 (u8 *)&iotype);
1635 break;
1636 default:
1637 RT_TRACE(rtlpriv, COMP_ERR, DBG_EMERG,
1638 "Unknown Scan Backup operation.\n");
1639 break;
1640 }
1641 }
1642}
1643EXPORT_SYMBOL(rtl_phy_scan_operation_backup);
1644
1616/* There seem to be issues in mac80211 regarding when del ba frames can be 1645/* There seem to be issues in mac80211 regarding when del ba frames can be
1617 * received. As a work around, we make a fake del_ba if we receive a ba_req; 1646 * received. As a work around, we make a fake del_ba if we receive a ba_req;
1618 * however, rx_agg was opened to let mac80211 release some ba related 1647 * however, rx_agg was opened to let mac80211 release some ba related
diff --git a/drivers/net/wireless/rtlwifi/base.h b/drivers/net/wireless/rtlwifi/base.h
index 0e5fe0902daf..0cd07420777a 100644
--- a/drivers/net/wireless/rtlwifi/base.h
+++ b/drivers/net/wireless/rtlwifi/base.h
@@ -114,7 +114,6 @@ void rtl_init_rfkill(struct ieee80211_hw *hw);
114void rtl_deinit_rfkill(struct ieee80211_hw *hw); 114void rtl_deinit_rfkill(struct ieee80211_hw *hw);
115 115
116void rtl_beacon_statistic(struct ieee80211_hw *hw, struct sk_buff *skb); 116void rtl_beacon_statistic(struct ieee80211_hw *hw, struct sk_buff *skb);
117void rtl_watch_dog_timer_callback(unsigned long data);
118void rtl_deinit_deferred_work(struct ieee80211_hw *hw); 117void rtl_deinit_deferred_work(struct ieee80211_hw *hw);
119 118
120bool rtl_action_proc(struct ieee80211_hw *hw, struct sk_buff *skb, u8 is_tx); 119bool rtl_action_proc(struct ieee80211_hw *hw, struct sk_buff *skb, u8 is_tx);
@@ -153,5 +152,6 @@ int rtlwifi_rate_mapping(struct ieee80211_hw *hw,
153bool rtl_tx_mgmt_proc(struct ieee80211_hw *hw, struct sk_buff *skb); 152bool rtl_tx_mgmt_proc(struct ieee80211_hw *hw, struct sk_buff *skb);
154struct sk_buff *rtl_make_del_ba(struct ieee80211_hw *hw, 153struct sk_buff *rtl_make_del_ba(struct ieee80211_hw *hw,
155 u8 *sa, u8 *bssid, u16 tid); 154 u8 *sa, u8 *bssid, u16 tid);
155void rtl_phy_scan_operation_backup(struct ieee80211_hw *hw, u8 operation);
156 156
157#endif 157#endif
diff --git a/drivers/net/wireless/rtlwifi/cam.h b/drivers/net/wireless/rtlwifi/cam.h
index 35e00086a520..0105e6c1901e 100644
--- a/drivers/net/wireless/rtlwifi/cam.h
+++ b/drivers/net/wireless/rtlwifi/cam.h
@@ -41,12 +41,12 @@
41#define CAM_CONFIG_USEDK 1 41#define CAM_CONFIG_USEDK 1
42#define CAM_CONFIG_NO_USEDK 0 42#define CAM_CONFIG_NO_USEDK 0
43 43
44extern void rtl_cam_reset_all_entry(struct ieee80211_hw *hw); 44void rtl_cam_reset_all_entry(struct ieee80211_hw *hw);
45extern u8 rtl_cam_add_one_entry(struct ieee80211_hw *hw, u8 *mac_addr, 45u8 rtl_cam_add_one_entry(struct ieee80211_hw *hw, u8 *mac_addr,
46 u32 ul_key_id, u32 ul_entry_idx, u32 ul_enc_alg, 46 u32 ul_key_id, u32 ul_entry_idx, u32 ul_enc_alg,
47 u32 ul_default_key, u8 *key_content); 47 u32 ul_default_key, u8 *key_content);
48int rtl_cam_delete_one_entry(struct ieee80211_hw *hw, u8 *mac_addr, 48int rtl_cam_delete_one_entry(struct ieee80211_hw *hw, u8 *mac_addr,
49 u32 ul_key_id); 49 u32 ul_key_id);
50void rtl_cam_mark_invalid(struct ieee80211_hw *hw, u8 uc_index); 50void rtl_cam_mark_invalid(struct ieee80211_hw *hw, u8 uc_index);
51void rtl_cam_empty_entry(struct ieee80211_hw *hw, u8 uc_index); 51void rtl_cam_empty_entry(struct ieee80211_hw *hw, u8 uc_index);
52void rtl_cam_reset_sec_info(struct ieee80211_hw *hw); 52void rtl_cam_reset_sec_info(struct ieee80211_hw *hw);
diff --git a/drivers/net/wireless/rtlwifi/core.c b/drivers/net/wireless/rtlwifi/core.c
index 733b7ce7f0e2..210ce7cd94d8 100644
--- a/drivers/net/wireless/rtlwifi/core.c
+++ b/drivers/net/wireless/rtlwifi/core.c
@@ -115,7 +115,7 @@ static void rtl_op_stop(struct ieee80211_hw *hw)
115 mutex_lock(&rtlpriv->locks.conf_mutex); 115 mutex_lock(&rtlpriv->locks.conf_mutex);
116 116
117 mac->link_state = MAC80211_NOLINK; 117 mac->link_state = MAC80211_NOLINK;
118 memset(mac->bssid, 0, 6); 118 memset(mac->bssid, 0, ETH_ALEN);
119 mac->vendor = PEER_UNKNOWN; 119 mac->vendor = PEER_UNKNOWN;
120 120
121 /*reset sec info */ 121 /*reset sec info */
@@ -280,7 +280,7 @@ static void rtl_op_remove_interface(struct ieee80211_hw *hw,
280 mac->p2p = 0; 280 mac->p2p = 0;
281 mac->vif = NULL; 281 mac->vif = NULL;
282 mac->link_state = MAC80211_NOLINK; 282 mac->link_state = MAC80211_NOLINK;
283 memset(mac->bssid, 0, 6); 283 memset(mac->bssid, 0, ETH_ALEN);
284 mac->vendor = PEER_UNKNOWN; 284 mac->vendor = PEER_UNKNOWN;
285 mac->opmode = NL80211_IFTYPE_UNSPECIFIED; 285 mac->opmode = NL80211_IFTYPE_UNSPECIFIED;
286 rtlpriv->cfg->ops->set_network_type(hw, mac->opmode); 286 rtlpriv->cfg->ops->set_network_type(hw, mac->opmode);
@@ -721,7 +721,7 @@ static void rtl_op_bss_info_changed(struct ieee80211_hw *hw,
721 mac->link_state = MAC80211_LINKED; 721 mac->link_state = MAC80211_LINKED;
722 mac->cnt_after_linked = 0; 722 mac->cnt_after_linked = 0;
723 mac->assoc_id = bss_conf->aid; 723 mac->assoc_id = bss_conf->aid;
724 memcpy(mac->bssid, bss_conf->bssid, 6); 724 memcpy(mac->bssid, bss_conf->bssid, ETH_ALEN);
725 725
726 if (rtlpriv->cfg->ops->linked_set_reg) 726 if (rtlpriv->cfg->ops->linked_set_reg)
727 rtlpriv->cfg->ops->linked_set_reg(hw); 727 rtlpriv->cfg->ops->linked_set_reg(hw);
@@ -750,7 +750,7 @@ static void rtl_op_bss_info_changed(struct ieee80211_hw *hw,
750 if (ppsc->p2p_ps_info.p2p_ps_mode > P2P_PS_NONE) 750 if (ppsc->p2p_ps_info.p2p_ps_mode > P2P_PS_NONE)
751 rtl_p2p_ps_cmd(hw, P2P_PS_DISABLE); 751 rtl_p2p_ps_cmd(hw, P2P_PS_DISABLE);
752 mac->link_state = MAC80211_NOLINK; 752 mac->link_state = MAC80211_NOLINK;
753 memset(mac->bssid, 0, 6); 753 memset(mac->bssid, 0, ETH_ALEN);
754 mac->vendor = PEER_UNKNOWN; 754 mac->vendor = PEER_UNKNOWN;
755 755
756 if (rtlpriv->dm.supp_phymode_switch) { 756 if (rtlpriv->dm.supp_phymode_switch) {
@@ -826,7 +826,7 @@ static void rtl_op_bss_info_changed(struct ieee80211_hw *hw,
826 bss_conf->bssid); 826 bss_conf->bssid);
827 827
828 mac->vendor = PEER_UNKNOWN; 828 mac->vendor = PEER_UNKNOWN;
829 memcpy(mac->bssid, bss_conf->bssid, 6); 829 memcpy(mac->bssid, bss_conf->bssid, ETH_ALEN);
830 rtlpriv->cfg->ops->set_network_type(hw, vif->type); 830 rtlpriv->cfg->ops->set_network_type(hw, vif->type);
831 831
832 rcu_read_lock(); 832 rcu_read_lock();
diff --git a/drivers/net/wireless/rtlwifi/efuse.c b/drivers/net/wireless/rtlwifi/efuse.c
index 838a1ed3f194..ae13fb94b2e8 100644
--- a/drivers/net/wireless/rtlwifi/efuse.c
+++ b/drivers/net/wireless/rtlwifi/efuse.c
@@ -1203,20 +1203,18 @@ static void efuse_power_switch(struct ieee80211_hw *hw, u8 write, u8 pwrstate)
1203 1203
1204static u16 efuse_get_current_size(struct ieee80211_hw *hw) 1204static u16 efuse_get_current_size(struct ieee80211_hw *hw)
1205{ 1205{
1206 int continual = true;
1207 u16 efuse_addr = 0; 1206 u16 efuse_addr = 0;
1208 u8 hworden; 1207 u8 hworden;
1209 u8 efuse_data, word_cnts; 1208 u8 efuse_data, word_cnts;
1210 1209
1211 while (continual && efuse_one_byte_read(hw, efuse_addr, &efuse_data) 1210 while (efuse_one_byte_read(hw, efuse_addr, &efuse_data) &&
1212 && (efuse_addr < EFUSE_MAX_SIZE)) { 1211 efuse_addr < EFUSE_MAX_SIZE) {
1213 if (efuse_data != 0xFF) { 1212 if (efuse_data == 0xFF)
1214 hworden = efuse_data & 0x0F; 1213 break;
1215 word_cnts = efuse_calculate_word_cnts(hworden); 1214
1216 efuse_addr = efuse_addr + (word_cnts * 2) + 1; 1215 hworden = efuse_data & 0x0F;
1217 } else { 1216 word_cnts = efuse_calculate_word_cnts(hworden);
1218 continual = false; 1217 efuse_addr = efuse_addr + (word_cnts * 2) + 1;
1219 }
1220 } 1218 }
1221 1219
1222 return efuse_addr; 1220 return efuse_addr;
diff --git a/drivers/net/wireless/rtlwifi/efuse.h b/drivers/net/wireless/rtlwifi/efuse.h
index 395a326acfb4..1663b3afd41e 100644
--- a/drivers/net/wireless/rtlwifi/efuse.h
+++ b/drivers/net/wireless/rtlwifi/efuse.h
@@ -104,20 +104,19 @@ struct efuse_priv {
104 u8 tx_power_g[14]; 104 u8 tx_power_g[14];
105}; 105};
106 106
107extern void read_efuse_byte(struct ieee80211_hw *hw, u16 _offset, u8 *pbuf); 107void read_efuse_byte(struct ieee80211_hw *hw, u16 _offset, u8 *pbuf);
108extern void efuse_initialize(struct ieee80211_hw *hw); 108void efuse_initialize(struct ieee80211_hw *hw);
109extern u8 efuse_read_1byte(struct ieee80211_hw *hw, u16 address); 109u8 efuse_read_1byte(struct ieee80211_hw *hw, u16 address);
110extern void efuse_write_1byte(struct ieee80211_hw *hw, u16 address, u8 value); 110void efuse_write_1byte(struct ieee80211_hw *hw, u16 address, u8 value);
111extern void read_efuse(struct ieee80211_hw *hw, u16 _offset, 111void read_efuse(struct ieee80211_hw *hw, u16 _offset, u16 _size_byte, u8 *pbuf);
112 u16 _size_byte, u8 *pbuf); 112void efuse_shadow_read(struct ieee80211_hw *hw, u8 type, u16 offset,
113extern void efuse_shadow_read(struct ieee80211_hw *hw, u8 type, 113 u32 *value);
114 u16 offset, u32 *value); 114void efuse_shadow_write(struct ieee80211_hw *hw, u8 type, u16 offset,
115extern void efuse_shadow_write(struct ieee80211_hw *hw, u8 type, 115 u32 value);
116 u16 offset, u32 value); 116bool efuse_shadow_update(struct ieee80211_hw *hw);
117extern bool efuse_shadow_update(struct ieee80211_hw *hw); 117bool efuse_shadow_update_chk(struct ieee80211_hw *hw);
118extern bool efuse_shadow_update_chk(struct ieee80211_hw *hw); 118void rtl_efuse_shadow_map_update(struct ieee80211_hw *hw);
119extern void rtl_efuse_shadow_map_update(struct ieee80211_hw *hw); 119void efuse_force_write_vendor_Id(struct ieee80211_hw *hw);
120extern void efuse_force_write_vendor_Id(struct ieee80211_hw *hw); 120void efuse_re_pg_section(struct ieee80211_hw *hw, u8 section_idx);
121extern void efuse_re_pg_section(struct ieee80211_hw *hw, u8 section_idx);
122 121
123#endif 122#endif
diff --git a/drivers/net/wireless/rtlwifi/pci.c b/drivers/net/wireless/rtlwifi/pci.c
index 703f839af6ca..0f494444bcd1 100644
--- a/drivers/net/wireless/rtlwifi/pci.c
+++ b/drivers/net/wireless/rtlwifi/pci.c
@@ -736,7 +736,6 @@ static void _rtl_pci_rx_interrupt(struct ieee80211_hw *hw)
736 736
737 struct rtl_stats stats = { 737 struct rtl_stats stats = {
738 .signal = 0, 738 .signal = 0,
739 .noise = -98,
740 .rate = 0, 739 .rate = 0,
741 }; 740 };
742 int index = rtlpci->rx_ring[rx_queue_idx].idx; 741 int index = rtlpci->rx_ring[rx_queue_idx].idx;
@@ -2009,7 +2008,6 @@ fail2:
2009fail1: 2008fail1:
2010 if (hw) 2009 if (hw)
2011 ieee80211_free_hw(hw); 2010 ieee80211_free_hw(hw);
2012 pci_set_drvdata(pdev, NULL);
2013 pci_disable_device(pdev); 2011 pci_disable_device(pdev);
2014 2012
2015 return err; 2013 return err;
@@ -2064,8 +2062,6 @@ void rtl_pci_disconnect(struct pci_dev *pdev)
2064 2062
2065 rtl_pci_disable_aspm(hw); 2063 rtl_pci_disable_aspm(hw);
2066 2064
2067 pci_set_drvdata(pdev, NULL);
2068
2069 ieee80211_free_hw(hw); 2065 ieee80211_free_hw(hw);
2070} 2066}
2071EXPORT_SYMBOL(rtl_pci_disconnect); 2067EXPORT_SYMBOL(rtl_pci_disconnect);
diff --git a/drivers/net/wireless/rtlwifi/rtl8188ee/hw.c b/drivers/net/wireless/rtlwifi/rtl8188ee/hw.c
index b68cae3024fc..e06971be7df7 100644
--- a/drivers/net/wireless/rtlwifi/rtl8188ee/hw.c
+++ b/drivers/net/wireless/rtlwifi/rtl8188ee/hw.c
@@ -143,6 +143,7 @@ static void _rtl88ee_set_fw_clock_on(struct ieee80211_hw *hw,
143 } else { 143 } else {
144 rtlhal->fw_clk_change_in_progress = false; 144 rtlhal->fw_clk_change_in_progress = false;
145 spin_unlock_bh(&rtlpriv->locks.fw_ps_lock); 145 spin_unlock_bh(&rtlpriv->locks.fw_ps_lock);
146 break;
146 } 147 }
147 } 148 }
148 149
diff --git a/drivers/net/wireless/rtlwifi/rtl8188ee/phy.c b/drivers/net/wireless/rtlwifi/rtl8188ee/phy.c
index e655c0473225..d67f9c731cc4 100644
--- a/drivers/net/wireless/rtlwifi/rtl8188ee/phy.c
+++ b/drivers/net/wireless/rtlwifi/rtl8188ee/phy.c
@@ -1136,34 +1136,6 @@ void rtl88e_phy_set_txpower_level(struct ieee80211_hw *hw, u8 channel)
1136 &bw40_pwr[0], channel); 1136 &bw40_pwr[0], channel);
1137} 1137}
1138 1138
1139void rtl88e_phy_scan_operation_backup(struct ieee80211_hw *hw, u8 operation)
1140{
1141 struct rtl_priv *rtlpriv = rtl_priv(hw);
1142 struct rtl_hal *rtlhal = rtl_hal(rtl_priv(hw));
1143 enum io_type iotype;
1144
1145 if (!is_hal_stop(rtlhal)) {
1146 switch (operation) {
1147 case SCAN_OPT_BACKUP:
1148 iotype = IO_CMD_PAUSE_DM_BY_SCAN;
1149 rtlpriv->cfg->ops->set_hw_reg(hw,
1150 HW_VAR_IO_CMD,
1151 (u8 *)&iotype);
1152 break;
1153 case SCAN_OPT_RESTORE:
1154 iotype = IO_CMD_RESUME_DM_BY_SCAN;
1155 rtlpriv->cfg->ops->set_hw_reg(hw,
1156 HW_VAR_IO_CMD,
1157 (u8 *)&iotype);
1158 break;
1159 default:
1160 RT_TRACE(rtlpriv, COMP_ERR, DBG_EMERG,
1161 "Unknown Scan Backup operation.\n");
1162 break;
1163 }
1164 }
1165}
1166
1167void rtl88e_phy_set_bw_mode_callback(struct ieee80211_hw *hw) 1139void rtl88e_phy_set_bw_mode_callback(struct ieee80211_hw *hw)
1168{ 1140{
1169 struct rtl_priv *rtlpriv = rtl_priv(hw); 1141 struct rtl_priv *rtlpriv = rtl_priv(hw);
diff --git a/drivers/net/wireless/rtlwifi/rtl8188ee/phy.h b/drivers/net/wireless/rtlwifi/rtl8188ee/phy.h
index f1acd6d27e44..89f0f1ef1465 100644
--- a/drivers/net/wireless/rtlwifi/rtl8188ee/phy.h
+++ b/drivers/net/wireless/rtlwifi/rtl8188ee/phy.h
@@ -200,37 +200,35 @@ enum _ANT_DIV_TYPE {
200 CGCS_RX_SW_ANTDIV = 0x05, 200 CGCS_RX_SW_ANTDIV = 0x05,
201}; 201};
202 202
203extern u32 rtl88e_phy_query_bb_reg(struct ieee80211_hw *hw, 203u32 rtl88e_phy_query_bb_reg(struct ieee80211_hw *hw,
204 u32 regaddr, u32 bitmask); 204 u32 regaddr, u32 bitmask);
205extern void rtl88e_phy_set_bb_reg(struct ieee80211_hw *hw, 205void rtl88e_phy_set_bb_reg(struct ieee80211_hw *hw,
206 u32 regaddr, u32 bitmask, u32 data); 206 u32 regaddr, u32 bitmask, u32 data);
207extern u32 rtl88e_phy_query_rf_reg(struct ieee80211_hw *hw, 207u32 rtl88e_phy_query_rf_reg(struct ieee80211_hw *hw,
208 enum radio_path rfpath, u32 regaddr, 208 enum radio_path rfpath, u32 regaddr,
209 u32 bitmask); 209 u32 bitmask);
210extern void rtl88e_phy_set_rf_reg(struct ieee80211_hw *hw, 210void rtl88e_phy_set_rf_reg(struct ieee80211_hw *hw,
211 enum radio_path rfpath, u32 regaddr, 211 enum radio_path rfpath, u32 regaddr,
212 u32 bitmask, u32 data); 212 u32 bitmask, u32 data);
213extern bool rtl88e_phy_mac_config(struct ieee80211_hw *hw); 213bool rtl88e_phy_mac_config(struct ieee80211_hw *hw);
214extern bool rtl88e_phy_bb_config(struct ieee80211_hw *hw); 214bool rtl88e_phy_bb_config(struct ieee80211_hw *hw);
215extern bool rtl88e_phy_rf_config(struct ieee80211_hw *hw); 215bool rtl88e_phy_rf_config(struct ieee80211_hw *hw);
216extern void rtl88e_phy_get_hw_reg_originalvalue(struct ieee80211_hw *hw); 216void rtl88e_phy_get_hw_reg_originalvalue(struct ieee80211_hw *hw);
217extern void rtl88e_phy_get_txpower_level(struct ieee80211_hw *hw, 217void rtl88e_phy_get_txpower_level(struct ieee80211_hw *hw,
218 long *powerlevel); 218 long *powerlevel);
219extern void rtl88e_phy_set_txpower_level(struct ieee80211_hw *hw, u8 channel); 219void rtl88e_phy_set_txpower_level(struct ieee80211_hw *hw, u8 channel);
220extern void rtl88e_phy_scan_operation_backup(struct ieee80211_hw *hw, 220void rtl88e_phy_set_bw_mode_callback(struct ieee80211_hw *hw);
221 u8 operation); 221void rtl88e_phy_set_bw_mode(struct ieee80211_hw *hw,
222extern void rtl88e_phy_set_bw_mode_callback(struct ieee80211_hw *hw); 222 enum nl80211_channel_type ch_type);
223extern void rtl88e_phy_set_bw_mode(struct ieee80211_hw *hw, 223void rtl88e_phy_sw_chnl_callback(struct ieee80211_hw *hw);
224 enum nl80211_channel_type ch_type); 224u8 rtl88e_phy_sw_chnl(struct ieee80211_hw *hw);
225extern void rtl88e_phy_sw_chnl_callback(struct ieee80211_hw *hw); 225void rtl88e_phy_iq_calibrate(struct ieee80211_hw *hw, bool b_recovery);
226extern u8 rtl88e_phy_sw_chnl(struct ieee80211_hw *hw);
227extern void rtl88e_phy_iq_calibrate(struct ieee80211_hw *hw, bool b_recovery);
228void rtl88e_phy_lc_calibrate(struct ieee80211_hw *hw); 226void rtl88e_phy_lc_calibrate(struct ieee80211_hw *hw);
229void rtl88e_phy_set_rfpath_switch(struct ieee80211_hw *hw, bool bmain); 227void rtl88e_phy_set_rfpath_switch(struct ieee80211_hw *hw, bool bmain);
230bool rtl88e_phy_config_rf_with_headerfile(struct ieee80211_hw *hw, 228bool rtl88e_phy_config_rf_with_headerfile(struct ieee80211_hw *hw,
231 enum radio_path rfpath); 229 enum radio_path rfpath);
232bool rtl88e_phy_set_io_cmd(struct ieee80211_hw *hw, enum io_type iotype); 230bool rtl88e_phy_set_io_cmd(struct ieee80211_hw *hw, enum io_type iotype);
233extern bool rtl88e_phy_set_rf_power_state(struct ieee80211_hw *hw, 231bool rtl88e_phy_set_rf_power_state(struct ieee80211_hw *hw,
234 enum rf_pwrstate rfpwr_state); 232 enum rf_pwrstate rfpwr_state);
235 233
236#endif 234#endif
diff --git a/drivers/net/wireless/rtlwifi/rtl8188ee/sw.c b/drivers/net/wireless/rtlwifi/rtl8188ee/sw.c
index c254693a1e6a..347af1e4f438 100644
--- a/drivers/net/wireless/rtlwifi/rtl8188ee/sw.c
+++ b/drivers/net/wireless/rtlwifi/rtl8188ee/sw.c
@@ -30,6 +30,7 @@
30#include "../wifi.h" 30#include "../wifi.h"
31#include "../core.h" 31#include "../core.h"
32#include "../pci.h" 32#include "../pci.h"
33#include "../base.h"
33#include "reg.h" 34#include "reg.h"
34#include "def.h" 35#include "def.h"
35#include "phy.h" 36#include "phy.h"
@@ -244,7 +245,7 @@ static struct rtl_hal_ops rtl8188ee_hal_ops = {
244 .set_bw_mode = rtl88e_phy_set_bw_mode, 245 .set_bw_mode = rtl88e_phy_set_bw_mode,
245 .switch_channel = rtl88e_phy_sw_chnl, 246 .switch_channel = rtl88e_phy_sw_chnl,
246 .dm_watchdog = rtl88e_dm_watchdog, 247 .dm_watchdog = rtl88e_dm_watchdog,
247 .scan_operation_backup = rtl88e_phy_scan_operation_backup, 248 .scan_operation_backup = rtl_phy_scan_operation_backup,
248 .set_rf_power_state = rtl88e_phy_set_rf_power_state, 249 .set_rf_power_state = rtl88e_phy_set_rf_power_state,
249 .led_control = rtl88ee_led_control, 250 .led_control = rtl88ee_led_control,
250 .set_desc = rtl88ee_set_desc, 251 .set_desc = rtl88ee_set_desc,
diff --git a/drivers/net/wireless/rtlwifi/rtl8188ee/trx.c b/drivers/net/wireless/rtlwifi/rtl8188ee/trx.c
index 68685a898257..aece6c9cccf1 100644
--- a/drivers/net/wireless/rtlwifi/rtl8188ee/trx.c
+++ b/drivers/net/wireless/rtlwifi/rtl8188ee/trx.c
@@ -478,7 +478,6 @@ bool rtl88ee_rx_query_desc(struct ieee80211_hw *hw,
478 478
479 /*rx_status->qual = status->signal; */ 479 /*rx_status->qual = status->signal; */
480 rx_status->signal = status->recvsignalpower + 10; 480 rx_status->signal = status->recvsignalpower + 10;
481 /*rx_status->noise = -status->noise; */
482 if (status->packet_report_type == TX_REPORT2) { 481 if (status->packet_report_type == TX_REPORT2) {
483 status->macid_valid_entry[0] = 482 status->macid_valid_entry[0] =
484 GET_RX_RPT2_DESC_MACID_VALID_1(pdesc); 483 GET_RX_RPT2_DESC_MACID_VALID_1(pdesc);
diff --git a/drivers/net/wireless/rtlwifi/rtl8192c/dm_common.c b/drivers/net/wireless/rtlwifi/rtl8192c/dm_common.c
index d2d57a27a7c1..e9caa5d4cff0 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192c/dm_common.c
+++ b/drivers/net/wireless/rtlwifi/rtl8192c/dm_common.c
@@ -541,29 +541,6 @@ EXPORT_SYMBOL(rtl92c_dm_write_dig);
541 541
542static void rtl92c_dm_pwdb_monitor(struct ieee80211_hw *hw) 542static void rtl92c_dm_pwdb_monitor(struct ieee80211_hw *hw)
543{ 543{
544 struct rtl_priv *rtlpriv = rtl_priv(hw);
545 long tmpentry_max_pwdb = 0, tmpentry_min_pwdb = 0xff;
546
547 u8 h2c_parameter[3] = { 0 };
548
549 return;
550
551 if (tmpentry_max_pwdb != 0) {
552 rtlpriv->dm.entry_max_undec_sm_pwdb = tmpentry_max_pwdb;
553 } else {
554 rtlpriv->dm.entry_max_undec_sm_pwdb = 0;
555 }
556
557 if (tmpentry_min_pwdb != 0xff) {
558 rtlpriv->dm.entry_min_undec_sm_pwdb = tmpentry_min_pwdb;
559 } else {
560 rtlpriv->dm.entry_min_undec_sm_pwdb = 0;
561 }
562
563 h2c_parameter[2] = (u8) (rtlpriv->dm.undec_sm_pwdb & 0xFF);
564 h2c_parameter[0] = 0;
565
566 rtl92c_fill_h2c_cmd(hw, H2C_RSSI_REPORT, 3, h2c_parameter);
567} 544}
568 545
569void rtl92c_dm_init_edca_turbo(struct ieee80211_hw *hw) 546void rtl92c_dm_init_edca_turbo(struct ieee80211_hw *hw)
@@ -673,7 +650,7 @@ static void rtl92c_dm_txpower_tracking_callback_thermalmeter(struct ieee80211_hw
673 s8 cck_index = 0; 650 s8 cck_index = 0;
674 int i; 651 int i;
675 bool is2t = IS_92C_SERIAL(rtlhal->version); 652 bool is2t = IS_92C_SERIAL(rtlhal->version);
676 s8 txpwr_level[2] = {0, 0}; 653 s8 txpwr_level[3] = {0, 0, 0};
677 u8 ofdm_min_index = 6, rf; 654 u8 ofdm_min_index = 6, rf;
678 655
679 rtlpriv->dm.txpower_trackinginit = true; 656 rtlpriv->dm.txpower_trackinginit = true;
diff --git a/drivers/net/wireless/rtlwifi/rtl8192c/phy_common.c b/drivers/net/wireless/rtlwifi/rtl8192c/phy_common.c
index 246e5352f2e1..0c0e78263a66 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192c/phy_common.c
+++ b/drivers/net/wireless/rtlwifi/rtl8192c/phy_common.c
@@ -592,36 +592,6 @@ long _rtl92c_phy_txpwr_idx_to_dbm(struct ieee80211_hw *hw,
592} 592}
593EXPORT_SYMBOL(_rtl92c_phy_txpwr_idx_to_dbm); 593EXPORT_SYMBOL(_rtl92c_phy_txpwr_idx_to_dbm);
594 594
595void rtl92c_phy_scan_operation_backup(struct ieee80211_hw *hw, u8 operation)
596{
597 struct rtl_priv *rtlpriv = rtl_priv(hw);
598 struct rtl_hal *rtlhal = rtl_hal(rtl_priv(hw));
599 enum io_type iotype;
600
601 if (!is_hal_stop(rtlhal)) {
602 switch (operation) {
603 case SCAN_OPT_BACKUP:
604 iotype = IO_CMD_PAUSE_DM_BY_SCAN;
605 rtlpriv->cfg->ops->set_hw_reg(hw,
606 HW_VAR_IO_CMD,
607 (u8 *)&iotype);
608
609 break;
610 case SCAN_OPT_RESTORE:
611 iotype = IO_CMD_RESUME_DM_BY_SCAN;
612 rtlpriv->cfg->ops->set_hw_reg(hw,
613 HW_VAR_IO_CMD,
614 (u8 *)&iotype);
615 break;
616 default:
617 RT_TRACE(rtlpriv, COMP_ERR, DBG_EMERG,
618 "Unknown Scan Backup operation\n");
619 break;
620 }
621 }
622}
623EXPORT_SYMBOL(rtl92c_phy_scan_operation_backup);
624
625void rtl92c_phy_set_bw_mode(struct ieee80211_hw *hw, 595void rtl92c_phy_set_bw_mode(struct ieee80211_hw *hw,
626 enum nl80211_channel_type ch_type) 596 enum nl80211_channel_type ch_type)
627{ 597{
diff --git a/drivers/net/wireless/rtlwifi/rtl8192c/phy_common.h b/drivers/net/wireless/rtlwifi/rtl8192c/phy_common.h
index cec10d696492..e79dabe9ba1d 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192c/phy_common.h
+++ b/drivers/net/wireless/rtlwifi/rtl8192c/phy_common.h
@@ -39,9 +39,7 @@
39#define RT_CANNOT_IO(hw) false 39#define RT_CANNOT_IO(hw) false
40#define HIGHPOWER_RADIOA_ARRAYLEN 22 40#define HIGHPOWER_RADIOA_ARRAYLEN 22
41 41
42#define IQK_ADDA_REG_NUM 16
43#define MAX_TOLERANCE 5 42#define MAX_TOLERANCE 5
44#define IQK_DELAY_TIME 1
45 43
46#define APK_BB_REG_NUM 5 44#define APK_BB_REG_NUM 5
47#define APK_AFE_REG_NUM 16 45#define APK_AFE_REG_NUM 16
@@ -205,8 +203,6 @@ void rtl92c_phy_get_txpower_level(struct ieee80211_hw *hw,
205void rtl92c_phy_set_txpower_level(struct ieee80211_hw *hw, u8 channel); 203void rtl92c_phy_set_txpower_level(struct ieee80211_hw *hw, u8 channel);
206bool rtl92c_phy_update_txpower_dbm(struct ieee80211_hw *hw, 204bool rtl92c_phy_update_txpower_dbm(struct ieee80211_hw *hw,
207 long power_indbm); 205 long power_indbm);
208void rtl92c_phy_scan_operation_backup(struct ieee80211_hw *hw,
209 u8 operation);
210void rtl92c_phy_set_bw_mode(struct ieee80211_hw *hw, 206void rtl92c_phy_set_bw_mode(struct ieee80211_hw *hw,
211 enum nl80211_channel_type ch_type); 207 enum nl80211_channel_type ch_type);
212void rtl92c_phy_sw_chnl_callback(struct ieee80211_hw *hw); 208void rtl92c_phy_sw_chnl_callback(struct ieee80211_hw *hw);
diff --git a/drivers/net/wireless/rtlwifi/rtl8192ce/def.h b/drivers/net/wireless/rtlwifi/rtl8192ce/def.h
index 3cfa1bb0f476..fa24de43ce79 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192ce/def.h
+++ b/drivers/net/wireless/rtlwifi/rtl8192ce/def.h
@@ -152,8 +152,6 @@ enum version_8192c {
152#define IS_VENDOR_UMC_A_CUT(version) ((IS_CHIP_VENDOR_UMC(version)) ? \ 152#define IS_VENDOR_UMC_A_CUT(version) ((IS_CHIP_VENDOR_UMC(version)) ? \
153 ((GET_CVID_CUT_VERSION(version)) ? false : true) : false) 153 ((GET_CVID_CUT_VERSION(version)) ? false : true) : false)
154#define IS_CHIP_VER_B(version) ((version & CHIP_VER_B) ? true : false) 154#define IS_CHIP_VER_B(version) ((version & CHIP_VER_B) ? true : false)
155#define IS_VENDOR_UMC_A_CUT(version) ((IS_CHIP_VENDOR_UMC(version)) ? \
156 ((GET_CVID_CUT_VERSION(version)) ? false : true) : false)
157#define IS_92C_SERIAL(version) ((version & CHIP_92C_BITMASK) ? true : false) 155#define IS_92C_SERIAL(version) ((version & CHIP_92C_BITMASK) ? true : false)
158#define IS_CHIP_VENDOR_UMC(version) \ 156#define IS_CHIP_VENDOR_UMC(version) \
159 ((version & CHIP_VENDOR_UMC) ? true : false) 157 ((version & CHIP_VENDOR_UMC) ? true : false)
diff --git a/drivers/net/wireless/rtlwifi/rtl8192ce/phy.h b/drivers/net/wireless/rtlwifi/rtl8192ce/phy.h
index d5e3b704f930..94486cca4000 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192ce/phy.h
+++ b/drivers/net/wireless/rtlwifi/rtl8192ce/phy.h
@@ -39,9 +39,7 @@
39#define RT_CANNOT_IO(hw) false 39#define RT_CANNOT_IO(hw) false
40#define HIGHPOWER_RADIOA_ARRAYLEN 22 40#define HIGHPOWER_RADIOA_ARRAYLEN 22
41 41
42#define IQK_ADDA_REG_NUM 16
43#define MAX_TOLERANCE 5 42#define MAX_TOLERANCE 5
44#define IQK_DELAY_TIME 1
45 43
46#define APK_BB_REG_NUM 5 44#define APK_BB_REG_NUM 5
47#define APK_AFE_REG_NUM 16 45#define APK_AFE_REG_NUM 16
@@ -188,36 +186,29 @@ struct tx_power_struct {
188}; 186};
189 187
190bool rtl92c_phy_bb_config(struct ieee80211_hw *hw); 188bool rtl92c_phy_bb_config(struct ieee80211_hw *hw);
191u32 rtl92c_phy_query_bb_reg(struct ieee80211_hw *hw, 189u32 rtl92c_phy_query_bb_reg(struct ieee80211_hw *hw, u32 regaddr, u32 bitmask);
192 u32 regaddr, u32 bitmask); 190void rtl92c_phy_set_bb_reg(struct ieee80211_hw *hw, u32 regaddr, u32 bitmask,
193void rtl92c_phy_set_bb_reg(struct ieee80211_hw *hw, 191 u32 data);
194 u32 regaddr, u32 bitmask, u32 data); 192u32 rtl92c_phy_query_rf_reg(struct ieee80211_hw *hw, enum radio_path rfpath,
195u32 rtl92c_phy_query_rf_reg(struct ieee80211_hw *hw, 193 u32 regaddr, u32 bitmask);
196 enum radio_path rfpath, u32 regaddr, 194void rtl92ce_phy_set_rf_reg(struct ieee80211_hw *hw, enum radio_path rfpath,
197 u32 bitmask); 195 u32 regaddr, u32 bitmask, u32 data);
198extern void rtl92ce_phy_set_rf_reg(struct ieee80211_hw *hw,
199 enum radio_path rfpath, u32 regaddr,
200 u32 bitmask, u32 data);
201bool rtl92c_phy_mac_config(struct ieee80211_hw *hw); 196bool rtl92c_phy_mac_config(struct ieee80211_hw *hw);
202bool rtl92ce_phy_bb_config(struct ieee80211_hw *hw); 197bool rtl92ce_phy_bb_config(struct ieee80211_hw *hw);
203bool rtl92c_phy_rf_config(struct ieee80211_hw *hw); 198bool rtl92c_phy_rf_config(struct ieee80211_hw *hw);
204bool rtl92c_phy_config_rf_with_feaderfile(struct ieee80211_hw *hw, 199bool rtl92c_phy_config_rf_with_feaderfile(struct ieee80211_hw *hw,
205 enum radio_path rfpath); 200 enum radio_path rfpath);
206void rtl92c_phy_get_hw_reg_originalvalue(struct ieee80211_hw *hw); 201void rtl92c_phy_get_hw_reg_originalvalue(struct ieee80211_hw *hw);
207void rtl92c_phy_get_txpower_level(struct ieee80211_hw *hw, 202void rtl92c_phy_get_txpower_level(struct ieee80211_hw *hw, long *powerlevel);
208 long *powerlevel);
209void rtl92c_phy_set_txpower_level(struct ieee80211_hw *hw, u8 channel); 203void rtl92c_phy_set_txpower_level(struct ieee80211_hw *hw, u8 channel);
210bool rtl92c_phy_update_txpower_dbm(struct ieee80211_hw *hw, 204bool rtl92c_phy_update_txpower_dbm(struct ieee80211_hw *hw,
211 long power_indbm); 205 long power_indbm);
212void rtl92c_phy_scan_operation_backup(struct ieee80211_hw *hw,
213 u8 operation);
214void rtl92c_phy_set_bw_mode(struct ieee80211_hw *hw, 206void rtl92c_phy_set_bw_mode(struct ieee80211_hw *hw,
215 enum nl80211_channel_type ch_type); 207 enum nl80211_channel_type ch_type);
216void rtl92c_phy_sw_chnl_callback(struct ieee80211_hw *hw); 208void rtl92c_phy_sw_chnl_callback(struct ieee80211_hw *hw);
217u8 rtl92c_phy_sw_chnl(struct ieee80211_hw *hw); 209u8 rtl92c_phy_sw_chnl(struct ieee80211_hw *hw);
218void rtl92c_phy_iq_calibrate(struct ieee80211_hw *hw, bool b_recovery); 210void rtl92c_phy_iq_calibrate(struct ieee80211_hw *hw, bool b_recovery);
219void rtl92c_phy_set_beacon_hw_reg(struct ieee80211_hw *hw, 211void rtl92c_phy_set_beacon_hw_reg(struct ieee80211_hw *hw, u16 beaconinterval);
220 u16 beaconinterval);
221void rtl92c_phy_ap_calibrate(struct ieee80211_hw *hw, char delta); 212void rtl92c_phy_ap_calibrate(struct ieee80211_hw *hw, char delta);
222void rtl92c_phy_lc_calibrate(struct ieee80211_hw *hw); 213void rtl92c_phy_lc_calibrate(struct ieee80211_hw *hw);
223void _rtl92ce_phy_lc_calibrate(struct ieee80211_hw *hw, bool is2t); 214void _rtl92ce_phy_lc_calibrate(struct ieee80211_hw *hw, bool is2t);
@@ -225,28 +216,25 @@ void rtl92c_phy_set_rfpath_switch(struct ieee80211_hw *hw, bool bmain);
225bool rtl92c_phy_config_rf_with_headerfile(struct ieee80211_hw *hw, 216bool rtl92c_phy_config_rf_with_headerfile(struct ieee80211_hw *hw,
226 enum radio_path rfpath); 217 enum radio_path rfpath);
227bool rtl8192_phy_check_is_legal_rfpath(struct ieee80211_hw *hw, 218bool rtl8192_phy_check_is_legal_rfpath(struct ieee80211_hw *hw,
228 u32 rfpath); 219 u32 rfpath);
229bool rtl92c_phy_set_io_cmd(struct ieee80211_hw *hw, enum io_type iotype);
230bool rtl92ce_phy_set_rf_power_state(struct ieee80211_hw *hw, 220bool rtl92ce_phy_set_rf_power_state(struct ieee80211_hw *hw,
231 enum rf_pwrstate rfpwr_state); 221 enum rf_pwrstate rfpwr_state);
232void rtl92ce_phy_set_rf_on(struct ieee80211_hw *hw); 222void rtl92ce_phy_set_rf_on(struct ieee80211_hw *hw);
233bool rtl92c_phy_set_io_cmd(struct ieee80211_hw *hw, enum io_type iotype); 223bool rtl92c_phy_set_io_cmd(struct ieee80211_hw *hw, enum io_type iotype);
234void rtl92c_phy_set_io(struct ieee80211_hw *hw); 224void rtl92c_phy_set_io(struct ieee80211_hw *hw);
235void rtl92c_bb_block_on(struct ieee80211_hw *hw); 225void rtl92c_bb_block_on(struct ieee80211_hw *hw);
236u32 _rtl92c_phy_rf_serial_read(struct ieee80211_hw *hw, 226u32 _rtl92c_phy_rf_serial_read(struct ieee80211_hw *hw, enum radio_path rfpath,
237 enum radio_path rfpath, u32 offset); 227 u32 offset);
238u32 _rtl92c_phy_fw_rf_serial_read(struct ieee80211_hw *hw, 228u32 _rtl92c_phy_fw_rf_serial_read(struct ieee80211_hw *hw,
239 enum radio_path rfpath, u32 offset); 229 enum radio_path rfpath, u32 offset);
240u32 _rtl92c_phy_calculate_bit_shift(u32 bitmask); 230u32 _rtl92c_phy_calculate_bit_shift(u32 bitmask);
241void _rtl92c_phy_rf_serial_write(struct ieee80211_hw *hw, 231void _rtl92c_phy_rf_serial_write(struct ieee80211_hw *hw,
242 enum radio_path rfpath, u32 offset, 232 enum radio_path rfpath, u32 offset, u32 data);
243 u32 data);
244void _rtl92c_phy_fw_rf_serial_write(struct ieee80211_hw *hw, 233void _rtl92c_phy_fw_rf_serial_write(struct ieee80211_hw *hw,
245 enum radio_path rfpath, u32 offset, 234 enum radio_path rfpath, u32 offset,
246 u32 data); 235 u32 data);
247void _rtl92c_store_pwrIndex_diffrate_offset(struct ieee80211_hw *hw, 236void _rtl92c_store_pwrIndex_diffrate_offset(struct ieee80211_hw *hw,
248 u32 regaddr, u32 bitmask, 237 u32 regaddr, u32 bitmask, u32 data);
249 u32 data);
250bool _rtl92ce_phy_config_mac_with_headerfile(struct ieee80211_hw *hw); 238bool _rtl92ce_phy_config_mac_with_headerfile(struct ieee80211_hw *hw);
251void _rtl92c_phy_init_bb_rf_register_definition(struct ieee80211_hw *hw); 239void _rtl92c_phy_init_bb_rf_register_definition(struct ieee80211_hw *hw);
252bool _rtl92c_phy_bb8192c_config_parafile(struct ieee80211_hw *hw); 240bool _rtl92c_phy_bb8192c_config_parafile(struct ieee80211_hw *hw);
diff --git a/drivers/net/wireless/rtlwifi/rtl8192ce/reg.h b/drivers/net/wireless/rtlwifi/rtl8192ce/reg.h
index bd4aef74c056..8922ecb47ad2 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192ce/reg.h
+++ b/drivers/net/wireless/rtlwifi/rtl8192ce/reg.h
@@ -560,7 +560,6 @@
560#define EEPROM_DEFAULT_TXPOWERLEVEL 0x22 560#define EEPROM_DEFAULT_TXPOWERLEVEL 0x22
561#define EEPROM_DEFAULT_HT40_2SDIFF 0x0 561#define EEPROM_DEFAULT_HT40_2SDIFF 0x0
562#define EEPROM_DEFAULT_HT20_DIFF 2 562#define EEPROM_DEFAULT_HT20_DIFF 2
563#define EEPROM_DEFAULT_LEGACYHTTXPOWERDIFF 0x3
564#define EEPROM_DEFAULT_HT40_PWRMAXOFFSET 0 563#define EEPROM_DEFAULT_HT40_PWRMAXOFFSET 0
565#define EEPROM_DEFAULT_HT20_PWRMAXOFFSET 0 564#define EEPROM_DEFAULT_HT20_PWRMAXOFFSET 0
566 565
@@ -639,17 +638,8 @@
639 638
640#define EEPROM_TXPWR_GROUP 0x6F 639#define EEPROM_TXPWR_GROUP 0x6F
641 640
642#define EEPROM_TSSI_A 0x76
643#define EEPROM_TSSI_B 0x77
644#define EEPROM_THERMAL_METER 0x78
645
646#define EEPROM_CHANNELPLAN 0x75 641#define EEPROM_CHANNELPLAN 0x75
647 642
648#define RF_OPTION1 0x79
649#define RF_OPTION2 0x7A
650#define RF_OPTION3 0x7B
651#define RF_OPTION4 0x7C
652
653#define STOPBECON BIT(6) 643#define STOPBECON BIT(6)
654#define STOPHIGHT BIT(5) 644#define STOPHIGHT BIT(5)
655#define STOPMGT BIT(4) 645#define STOPMGT BIT(4)
@@ -689,13 +679,6 @@
689#define RSV_CTRL 0x001C 679#define RSV_CTRL 0x001C
690#define RD_CTRL 0x0524 680#define RD_CTRL 0x0524
691 681
692#define REG_USB_INFO 0xFE17
693#define REG_USB_SPECIAL_OPTION 0xFE55
694
695#define REG_USB_DMA_AGG_TO 0xFE5B
696#define REG_USB_AGG_TO 0xFE5C
697#define REG_USB_AGG_TH 0xFE5D
698
699#define REG_USB_VID 0xFE60 682#define REG_USB_VID 0xFE60
700#define REG_USB_PID 0xFE62 683#define REG_USB_PID 0xFE62
701#define REG_USB_OPTIONAL 0xFE64 684#define REG_USB_OPTIONAL 0xFE64
@@ -1196,9 +1179,6 @@
1196#define POLLING_LLT_THRESHOLD 20 1179#define POLLING_LLT_THRESHOLD 20
1197#define POLLING_READY_TIMEOUT_COUNT 1000 1180#define POLLING_READY_TIMEOUT_COUNT 1000
1198 1181
1199#define MAX_MSS_DENSITY_2T 0x13
1200#define MAX_MSS_DENSITY_1T 0x0A
1201
1202#define EPROM_CMD_OPERATING_MODE_MASK ((1<<7)|(1<<6)) 1182#define EPROM_CMD_OPERATING_MODE_MASK ((1<<7)|(1<<6))
1203#define EPROM_CMD_CONFIG 0x3 1183#define EPROM_CMD_CONFIG 0x3
1204#define EPROM_CMD_LOAD 1 1184#define EPROM_CMD_LOAD 1
diff --git a/drivers/net/wireless/rtlwifi/rtl8192ce/rf.h b/drivers/net/wireless/rtlwifi/rtl8192ce/rf.h
index 6c8d56efceae..d8fe68b389d2 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192ce/rf.h
+++ b/drivers/net/wireless/rtlwifi/rtl8192ce/rf.h
@@ -34,11 +34,10 @@
34#define RF6052_MAX_REG 0x3F 34#define RF6052_MAX_REG 0x3F
35#define RF6052_MAX_PATH 2 35#define RF6052_MAX_PATH 2
36 36
37extern void rtl92ce_phy_rf6052_set_bandwidth(struct ieee80211_hw *hw, 37void rtl92ce_phy_rf6052_set_bandwidth(struct ieee80211_hw *hw, u8 bandwidth);
38 u8 bandwidth); 38void rtl92ce_phy_rf6052_set_cck_txpower(struct ieee80211_hw *hw,
39extern void rtl92ce_phy_rf6052_set_cck_txpower(struct ieee80211_hw *hw, 39 u8 *ppowerlevel);
40 u8 *ppowerlevel); 40void rtl92ce_phy_rf6052_set_ofdm_txpower(struct ieee80211_hw *hw,
41extern void rtl92ce_phy_rf6052_set_ofdm_txpower(struct ieee80211_hw *hw, 41 u8 *ppowerlevel, u8 channel);
42 u8 *ppowerlevel, u8 channel); 42bool rtl92ce_phy_rf6052_config(struct ieee80211_hw *hw);
43extern bool rtl92ce_phy_rf6052_config(struct ieee80211_hw *hw);
44#endif 43#endif
diff --git a/drivers/net/wireless/rtlwifi/rtl8192ce/sw.c b/drivers/net/wireless/rtlwifi/rtl8192ce/sw.c
index 14203561b6ee..b790320d2030 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192ce/sw.c
+++ b/drivers/net/wireless/rtlwifi/rtl8192ce/sw.c
@@ -30,6 +30,7 @@
30#include "../wifi.h" 30#include "../wifi.h"
31#include "../core.h" 31#include "../core.h"
32#include "../pci.h" 32#include "../pci.h"
33#include "../base.h"
33#include "reg.h" 34#include "reg.h"
34#include "def.h" 35#include "def.h"
35#include "phy.h" 36#include "phy.h"
@@ -219,7 +220,7 @@ static struct rtl_hal_ops rtl8192ce_hal_ops = {
219 .set_bw_mode = rtl92c_phy_set_bw_mode, 220 .set_bw_mode = rtl92c_phy_set_bw_mode,
220 .switch_channel = rtl92c_phy_sw_chnl, 221 .switch_channel = rtl92c_phy_sw_chnl,
221 .dm_watchdog = rtl92c_dm_watchdog, 222 .dm_watchdog = rtl92c_dm_watchdog,
222 .scan_operation_backup = rtl92c_phy_scan_operation_backup, 223 .scan_operation_backup = rtl_phy_scan_operation_backup,
223 .set_rf_power_state = rtl92c_phy_set_rf_power_state, 224 .set_rf_power_state = rtl92c_phy_set_rf_power_state,
224 .led_control = rtl92ce_led_control, 225 .led_control = rtl92ce_led_control,
225 .set_desc = rtl92ce_set_desc, 226 .set_desc = rtl92ce_set_desc,
diff --git a/drivers/net/wireless/rtlwifi/rtl8192ce/trx.c b/drivers/net/wireless/rtlwifi/rtl8192ce/trx.c
index 6ad23b413eb3..52abf0a862fa 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192ce/trx.c
+++ b/drivers/net/wireless/rtlwifi/rtl8192ce/trx.c
@@ -420,7 +420,6 @@ bool rtl92ce_rx_query_desc(struct ieee80211_hw *hw,
420 420
421 /*rx_status->qual = stats->signal; */ 421 /*rx_status->qual = stats->signal; */
422 rx_status->signal = stats->recvsignalpower + 10; 422 rx_status->signal = stats->recvsignalpower + 10;
423 /*rx_status->noise = -stats->noise; */
424 423
425 return true; 424 return true;
426} 425}
diff --git a/drivers/net/wireless/rtlwifi/rtl8192cu/mac.c b/drivers/net/wireless/rtlwifi/rtl8192cu/mac.c
index da4f587199ee..393685390f3e 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192cu/mac.c
+++ b/drivers/net/wireless/rtlwifi/rtl8192cu/mac.c
@@ -32,6 +32,7 @@
32#include "../usb.h" 32#include "../usb.h"
33#include "../ps.h" 33#include "../ps.h"
34#include "../cam.h" 34#include "../cam.h"
35#include "../stats.h"
35#include "reg.h" 36#include "reg.h"
36#include "def.h" 37#include "def.h"
37#include "phy.h" 38#include "phy.h"
@@ -738,16 +739,6 @@ static u8 _rtl92c_evm_db_to_percentage(char value)
738 return ret_val; 739 return ret_val;
739} 740}
740 741
741static long _rtl92c_translate_todbm(struct ieee80211_hw *hw,
742 u8 signal_strength_index)
743{
744 long signal_power;
745
746 signal_power = (long)((signal_strength_index + 1) >> 1);
747 signal_power -= 95;
748 return signal_power;
749}
750
751static long _rtl92c_signal_scale_mapping(struct ieee80211_hw *hw, 742static long _rtl92c_signal_scale_mapping(struct ieee80211_hw *hw,
752 long currsig) 743 long currsig)
753{ 744{
@@ -913,180 +904,6 @@ static void _rtl92c_query_rxphystatus(struct ieee80211_hw *hw,
913 (hw, total_rssi /= rf_rx_num)); 904 (hw, total_rssi /= rf_rx_num));
914} 905}
915 906
916static void _rtl92c_process_ui_rssi(struct ieee80211_hw *hw,
917 struct rtl_stats *pstats)
918{
919 struct rtl_priv *rtlpriv = rtl_priv(hw);
920 struct rtl_phy *rtlphy = &(rtlpriv->phy);
921 u8 rfpath;
922 u32 last_rssi, tmpval;
923
924 if (pstats->packet_toself || pstats->packet_beacon) {
925 rtlpriv->stats.rssi_calculate_cnt++;
926 if (rtlpriv->stats.ui_rssi.total_num++ >=
927 PHY_RSSI_SLID_WIN_MAX) {
928 rtlpriv->stats.ui_rssi.total_num =
929 PHY_RSSI_SLID_WIN_MAX;
930 last_rssi =
931 rtlpriv->stats.ui_rssi.elements[rtlpriv->
932 stats.ui_rssi.index];
933 rtlpriv->stats.ui_rssi.total_val -= last_rssi;
934 }
935 rtlpriv->stats.ui_rssi.total_val += pstats->signalstrength;
936 rtlpriv->stats.ui_rssi.elements[rtlpriv->stats.ui_rssi.
937 index++] = pstats->signalstrength;
938 if (rtlpriv->stats.ui_rssi.index >= PHY_RSSI_SLID_WIN_MAX)
939 rtlpriv->stats.ui_rssi.index = 0;
940 tmpval = rtlpriv->stats.ui_rssi.total_val /
941 rtlpriv->stats.ui_rssi.total_num;
942 rtlpriv->stats.signal_strength =
943 _rtl92c_translate_todbm(hw, (u8) tmpval);
944 pstats->rssi = rtlpriv->stats.signal_strength;
945 }
946 if (!pstats->is_cck && pstats->packet_toself) {
947 for (rfpath = RF90_PATH_A; rfpath < rtlphy->num_total_rfpath;
948 rfpath++) {
949 if (!rtl8192_phy_check_is_legal_rfpath(hw, rfpath))
950 continue;
951 if (rtlpriv->stats.rx_rssi_percentage[rfpath] == 0) {
952 rtlpriv->stats.rx_rssi_percentage[rfpath] =
953 pstats->rx_mimo_signalstrength[rfpath];
954 }
955 if (pstats->rx_mimo_signalstrength[rfpath] >
956 rtlpriv->stats.rx_rssi_percentage[rfpath]) {
957 rtlpriv->stats.rx_rssi_percentage[rfpath] =
958 ((rtlpriv->stats.
959 rx_rssi_percentage[rfpath] *
960 (RX_SMOOTH_FACTOR - 1)) +
961 (pstats->rx_mimo_signalstrength[rfpath])) /
962 (RX_SMOOTH_FACTOR);
963
964 rtlpriv->stats.rx_rssi_percentage[rfpath] =
965 rtlpriv->stats.rx_rssi_percentage[rfpath] +
966 1;
967 } else {
968 rtlpriv->stats.rx_rssi_percentage[rfpath] =
969 ((rtlpriv->stats.
970 rx_rssi_percentage[rfpath] *
971 (RX_SMOOTH_FACTOR - 1)) +
972 (pstats->rx_mimo_signalstrength[rfpath])) /
973 (RX_SMOOTH_FACTOR);
974 }
975 }
976 }
977}
978
979static void _rtl92c_update_rxsignalstatistics(struct ieee80211_hw *hw,
980 struct rtl_stats *pstats)
981{
982 struct rtl_priv *rtlpriv = rtl_priv(hw);
983 int weighting = 0;
984
985 if (rtlpriv->stats.recv_signal_power == 0)
986 rtlpriv->stats.recv_signal_power = pstats->recvsignalpower;
987 if (pstats->recvsignalpower > rtlpriv->stats.recv_signal_power)
988 weighting = 5;
989 else if (pstats->recvsignalpower < rtlpriv->stats.recv_signal_power)
990 weighting = (-5);
991 rtlpriv->stats.recv_signal_power =
992 (rtlpriv->stats.recv_signal_power * 5 +
993 pstats->recvsignalpower + weighting) / 6;
994}
995
996static void _rtl92c_process_pwdb(struct ieee80211_hw *hw,
997 struct rtl_stats *pstats)
998{
999 struct rtl_priv *rtlpriv = rtl_priv(hw);
1000 struct rtl_mac *mac = rtl_mac(rtl_priv(hw));
1001 long undec_sm_pwdb = 0;
1002
1003 if (mac->opmode == NL80211_IFTYPE_ADHOC) {
1004 return;
1005 } else {
1006 undec_sm_pwdb = rtlpriv->dm.undec_sm_pwdb;
1007 }
1008 if (pstats->packet_toself || pstats->packet_beacon) {
1009 if (undec_sm_pwdb < 0)
1010 undec_sm_pwdb = pstats->rx_pwdb_all;
1011 if (pstats->rx_pwdb_all > (u32) undec_sm_pwdb) {
1012 undec_sm_pwdb = (((undec_sm_pwdb) *
1013 (RX_SMOOTH_FACTOR - 1)) +
1014 (pstats->rx_pwdb_all)) / (RX_SMOOTH_FACTOR);
1015 undec_sm_pwdb += 1;
1016 } else {
1017 undec_sm_pwdb = (((undec_sm_pwdb) *
1018 (RX_SMOOTH_FACTOR - 1)) +
1019 (pstats->rx_pwdb_all)) / (RX_SMOOTH_FACTOR);
1020 }
1021 rtlpriv->dm.undec_sm_pwdb = undec_sm_pwdb;
1022 _rtl92c_update_rxsignalstatistics(hw, pstats);
1023 }
1024}
1025
1026static void _rtl92c_process_LINK_Q(struct ieee80211_hw *hw,
1027 struct rtl_stats *pstats)
1028{
1029 struct rtl_priv *rtlpriv = rtl_priv(hw);
1030 u32 last_evm = 0, n_stream, tmpval;
1031
1032 if (pstats->signalquality != 0) {
1033 if (pstats->packet_toself || pstats->packet_beacon) {
1034 if (rtlpriv->stats.LINK_Q.total_num++ >=
1035 PHY_LINKQUALITY_SLID_WIN_MAX) {
1036 rtlpriv->stats.LINK_Q.total_num =
1037 PHY_LINKQUALITY_SLID_WIN_MAX;
1038 last_evm =
1039 rtlpriv->stats.LINK_Q.elements
1040 [rtlpriv->stats.LINK_Q.index];
1041 rtlpriv->stats.LINK_Q.total_val -=
1042 last_evm;
1043 }
1044 rtlpriv->stats.LINK_Q.total_val +=
1045 pstats->signalquality;
1046 rtlpriv->stats.LINK_Q.elements
1047 [rtlpriv->stats.LINK_Q.index++] =
1048 pstats->signalquality;
1049 if (rtlpriv->stats.LINK_Q.index >=
1050 PHY_LINKQUALITY_SLID_WIN_MAX)
1051 rtlpriv->stats.LINK_Q.index = 0;
1052 tmpval = rtlpriv->stats.LINK_Q.total_val /
1053 rtlpriv->stats.LINK_Q.total_num;
1054 rtlpriv->stats.signal_quality = tmpval;
1055 rtlpriv->stats.last_sigstrength_inpercent = tmpval;
1056 for (n_stream = 0; n_stream < 2;
1057 n_stream++) {
1058 if (pstats->RX_SIGQ[n_stream] != -1) {
1059 if (!rtlpriv->stats.RX_EVM[n_stream]) {
1060 rtlpriv->stats.RX_EVM[n_stream]
1061 = pstats->RX_SIGQ[n_stream];
1062 }
1063 rtlpriv->stats.RX_EVM[n_stream] =
1064 ((rtlpriv->stats.RX_EVM
1065 [n_stream] *
1066 (RX_SMOOTH_FACTOR - 1)) +
1067 (pstats->RX_SIGQ
1068 [n_stream] * 1)) /
1069 (RX_SMOOTH_FACTOR);
1070 }
1071 }
1072 }
1073 } else {
1074 ;
1075 }
1076}
1077
1078static void _rtl92c_process_phyinfo(struct ieee80211_hw *hw,
1079 u8 *buffer,
1080 struct rtl_stats *pcurrent_stats)
1081{
1082 if (!pcurrent_stats->packet_matchbssid &&
1083 !pcurrent_stats->packet_beacon)
1084 return;
1085 _rtl92c_process_ui_rssi(hw, pcurrent_stats);
1086 _rtl92c_process_pwdb(hw, pcurrent_stats);
1087 _rtl92c_process_LINK_Q(hw, pcurrent_stats);
1088}
1089
1090void rtl92c_translate_rx_signal_stuff(struct ieee80211_hw *hw, 907void rtl92c_translate_rx_signal_stuff(struct ieee80211_hw *hw,
1091 struct sk_buff *skb, 908 struct sk_buff *skb,
1092 struct rtl_stats *pstats, 909 struct rtl_stats *pstats,
@@ -1123,5 +940,5 @@ void rtl92c_translate_rx_signal_stuff(struct ieee80211_hw *hw,
1123 _rtl92c_query_rxphystatus(hw, pstats, pdesc, p_drvinfo, 940 _rtl92c_query_rxphystatus(hw, pstats, pdesc, p_drvinfo,
1124 packet_matchbssid, packet_toself, 941 packet_matchbssid, packet_toself,
1125 packet_beacon); 942 packet_beacon);
1126 _rtl92c_process_phyinfo(hw, tmp_buf, pstats); 943 rtl_process_phyinfo(hw, tmp_buf, pstats);
1127} 944}
diff --git a/drivers/net/wireless/rtlwifi/rtl8192cu/rf.h b/drivers/net/wireless/rtlwifi/rtl8192cu/rf.h
index 090fd33a158d..11b439d6b671 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192cu/rf.h
+++ b/drivers/net/wireless/rtlwifi/rtl8192cu/rf.h
@@ -34,15 +34,14 @@
34#define RF6052_MAX_REG 0x3F 34#define RF6052_MAX_REG 0x3F
35#define RF6052_MAX_PATH 2 35#define RF6052_MAX_PATH 2
36 36
37extern void rtl92cu_phy_rf6052_set_bandwidth(struct ieee80211_hw *hw, 37void rtl92cu_phy_rf6052_set_bandwidth(struct ieee80211_hw *hw, u8 bandwidth);
38 u8 bandwidth); 38void rtl92c_phy_rf6052_set_cck_txpower(struct ieee80211_hw *hw,
39extern void rtl92c_phy_rf6052_set_cck_txpower(struct ieee80211_hw *hw, 39 u8 *ppowerlevel);
40 u8 *ppowerlevel); 40void rtl92c_phy_rf6052_set_ofdm_txpower(struct ieee80211_hw *hw,
41extern void rtl92c_phy_rf6052_set_ofdm_txpower(struct ieee80211_hw *hw, 41 u8 *ppowerlevel, u8 channel);
42 u8 *ppowerlevel, u8 channel);
43bool rtl92cu_phy_rf6052_config(struct ieee80211_hw *hw); 42bool rtl92cu_phy_rf6052_config(struct ieee80211_hw *hw);
44bool rtl92cu_phy_config_rf_with_headerfile(struct ieee80211_hw *hw, 43bool rtl92cu_phy_config_rf_with_headerfile(struct ieee80211_hw *hw,
45 enum radio_path rfpath); 44 enum radio_path rfpath);
46void rtl92cu_phy_rf6052_set_cck_txpower(struct ieee80211_hw *hw, 45void rtl92cu_phy_rf6052_set_cck_txpower(struct ieee80211_hw *hw,
47 u8 *ppowerlevel); 46 u8 *ppowerlevel);
48void rtl92cu_phy_rf6052_set_ofdm_txpower(struct ieee80211_hw *hw, 47void rtl92cu_phy_rf6052_set_ofdm_txpower(struct ieee80211_hw *hw,
diff --git a/drivers/net/wireless/rtlwifi/rtl8192cu/sw.c b/drivers/net/wireless/rtlwifi/rtl8192cu/sw.c
index 2bd598526217..9936de716ad5 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192cu/sw.c
+++ b/drivers/net/wireless/rtlwifi/rtl8192cu/sw.c
@@ -31,6 +31,7 @@
31#include "../core.h" 31#include "../core.h"
32#include "../usb.h" 32#include "../usb.h"
33#include "../efuse.h" 33#include "../efuse.h"
34#include "../base.h"
34#include "reg.h" 35#include "reg.h"
35#include "def.h" 36#include "def.h"
36#include "phy.h" 37#include "phy.h"
@@ -117,7 +118,7 @@ static struct rtl_hal_ops rtl8192cu_hal_ops = {
117 .set_bw_mode = rtl92c_phy_set_bw_mode, 118 .set_bw_mode = rtl92c_phy_set_bw_mode,
118 .switch_channel = rtl92c_phy_sw_chnl, 119 .switch_channel = rtl92c_phy_sw_chnl,
119 .dm_watchdog = rtl92c_dm_watchdog, 120 .dm_watchdog = rtl92c_dm_watchdog,
120 .scan_operation_backup = rtl92c_phy_scan_operation_backup, 121 .scan_operation_backup = rtl_phy_scan_operation_backup,
121 .set_rf_power_state = rtl92cu_phy_set_rf_power_state, 122 .set_rf_power_state = rtl92cu_phy_set_rf_power_state,
122 .led_control = rtl92cu_led_control, 123 .led_control = rtl92cu_led_control,
123 .enable_hw_sec = rtl92cu_enable_hw_security_config, 124 .enable_hw_sec = rtl92cu_enable_hw_security_config,
diff --git a/drivers/net/wireless/rtlwifi/rtl8192cu/trx.c b/drivers/net/wireless/rtlwifi/rtl8192cu/trx.c
index 5a060e537fbe..25e50ffc44ec 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192cu/trx.c
+++ b/drivers/net/wireless/rtlwifi/rtl8192cu/trx.c
@@ -350,7 +350,6 @@ bool rtl92cu_rx_query_desc(struct ieee80211_hw *hw,
350 } 350 }
351 /*rx_status->qual = stats->signal; */ 351 /*rx_status->qual = stats->signal; */
352 rx_status->signal = stats->rssi + 10; 352 rx_status->signal = stats->rssi + 10;
353 /*rx_status->noise = -stats->noise; */
354 return true; 353 return true;
355} 354}
356 355
@@ -365,7 +364,6 @@ static void _rtl_rx_process(struct ieee80211_hw *hw, struct sk_buff *skb)
365 u8 *rxdesc; 364 u8 *rxdesc;
366 struct rtl_stats stats = { 365 struct rtl_stats stats = {
367 .signal = 0, 366 .signal = 0,
368 .noise = -98,
369 .rate = 0, 367 .rate = 0,
370 }; 368 };
371 struct rx_fwinfo_92c *p_drvinfo; 369 struct rx_fwinfo_92c *p_drvinfo;
diff --git a/drivers/net/wireless/rtlwifi/rtl8192de/dm.c b/drivers/net/wireless/rtlwifi/rtl8192de/dm.c
index f700f7a614b2..7908e1c85819 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192de/dm.c
+++ b/drivers/net/wireless/rtlwifi/rtl8192de/dm.c
@@ -840,9 +840,9 @@ static void rtl92d_dm_txpower_tracking_callback_thermalmeter(
840 bool internal_pa = false; 840 bool internal_pa = false;
841 long ele_a = 0, ele_d, temp_cck, val_x, value32; 841 long ele_a = 0, ele_d, temp_cck, val_x, value32;
842 long val_y, ele_c = 0; 842 long val_y, ele_c = 0;
843 u8 ofdm_index[2]; 843 u8 ofdm_index[3];
844 s8 cck_index = 0; 844 s8 cck_index = 0;
845 u8 ofdm_index_old[2] = {0, 0}; 845 u8 ofdm_index_old[3] = {0, 0, 0};
846 s8 cck_index_old = 0; 846 s8 cck_index_old = 0;
847 u8 index; 847 u8 index;
848 int i; 848 int i;
@@ -1118,6 +1118,10 @@ static void rtl92d_dm_txpower_tracking_callback_thermalmeter(
1118 val_x, val_y, ele_a, ele_c, ele_d, 1118 val_x, val_y, ele_a, ele_c, ele_d,
1119 val_x, val_y); 1119 val_x, val_y);
1120 1120
1121 if (cck_index >= CCK_TABLE_SIZE)
1122 cck_index = CCK_TABLE_SIZE - 1;
1123 if (cck_index < 0)
1124 cck_index = 0;
1121 if (rtlhal->current_bandtype == BAND_ON_2_4G) { 1125 if (rtlhal->current_bandtype == BAND_ON_2_4G) {
1122 /* Adjust CCK according to IQK result */ 1126 /* Adjust CCK according to IQK result */
1123 if (!rtlpriv->dm.cck_inch14) { 1127 if (!rtlpriv->dm.cck_inch14) {
diff --git a/drivers/net/wireless/rtlwifi/rtl8192de/hw.c b/drivers/net/wireless/rtlwifi/rtl8192de/hw.c
index 7dd8f6de0550..c4a7db9135d6 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192de/hw.c
+++ b/drivers/net/wireless/rtlwifi/rtl8192de/hw.c
@@ -1194,25 +1194,7 @@ void rtl92d_linked_set_reg(struct ieee80211_hw *hw)
1194 * mac80211 will send pkt when scan */ 1194 * mac80211 will send pkt when scan */
1195void rtl92de_set_qos(struct ieee80211_hw *hw, int aci) 1195void rtl92de_set_qos(struct ieee80211_hw *hw, int aci)
1196{ 1196{
1197 struct rtl_priv *rtlpriv = rtl_priv(hw);
1198 rtl92d_dm_init_edca_turbo(hw); 1197 rtl92d_dm_init_edca_turbo(hw);
1199 return;
1200 switch (aci) {
1201 case AC1_BK:
1202 rtl_write_dword(rtlpriv, REG_EDCA_BK_PARAM, 0xa44f);
1203 break;
1204 case AC0_BE:
1205 break;
1206 case AC2_VI:
1207 rtl_write_dword(rtlpriv, REG_EDCA_VI_PARAM, 0x5e4322);
1208 break;
1209 case AC3_VO:
1210 rtl_write_dword(rtlpriv, REG_EDCA_VO_PARAM, 0x2f3222);
1211 break;
1212 default:
1213 RT_ASSERT(false, "invalid aci: %d !\n", aci);
1214 break;
1215 }
1216} 1198}
1217 1199
1218void rtl92de_enable_interrupt(struct ieee80211_hw *hw) 1200void rtl92de_enable_interrupt(struct ieee80211_hw *hw)
diff --git a/drivers/net/wireless/rtlwifi/rtl8192de/hw.h b/drivers/net/wireless/rtlwifi/rtl8192de/hw.h
index 7c9f7a2f1e42..1bc7b1a96d4a 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192de/hw.h
+++ b/drivers/net/wireless/rtlwifi/rtl8192de/hw.h
@@ -55,10 +55,9 @@ void rtl92de_set_key(struct ieee80211_hw *hw, u32 key_index,
55 u8 *p_macaddr, bool is_group, u8 enc_algo, 55 u8 *p_macaddr, bool is_group, u8 enc_algo,
56 bool is_wepkey, bool clear_all); 56 bool is_wepkey, bool clear_all);
57 57
58extern void rtl92de_write_dword_dbi(struct ieee80211_hw *hw, u16 offset, 58void rtl92de_write_dword_dbi(struct ieee80211_hw *hw, u16 offset, u32 value,
59 u32 value, u8 direct); 59 u8 direct);
60extern u32 rtl92de_read_dword_dbi(struct ieee80211_hw *hw, u16 offset, 60u32 rtl92de_read_dword_dbi(struct ieee80211_hw *hw, u16 offset, u8 direct);
61 u8 direct);
62void rtl92de_suspend(struct ieee80211_hw *hw); 61void rtl92de_suspend(struct ieee80211_hw *hw);
63void rtl92de_resume(struct ieee80211_hw *hw); 62void rtl92de_resume(struct ieee80211_hw *hw);
64void rtl92d_linked_set_reg(struct ieee80211_hw *hw); 63void rtl92d_linked_set_reg(struct ieee80211_hw *hw);
diff --git a/drivers/net/wireless/rtlwifi/rtl8192de/phy.c b/drivers/net/wireless/rtlwifi/rtl8192de/phy.c
index 840bac5fa2f8..13196cc4b1d3 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192de/phy.c
+++ b/drivers/net/wireless/rtlwifi/rtl8192de/phy.c
@@ -1022,34 +1022,6 @@ void rtl92d_phy_set_txpower_level(struct ieee80211_hw *hw, u8 channel)
1022 rtl92d_phy_rf6052_set_ofdm_txpower(hw, &ofdmpowerlevel[0], channel); 1022 rtl92d_phy_rf6052_set_ofdm_txpower(hw, &ofdmpowerlevel[0], channel);
1023} 1023}
1024 1024
1025void rtl92d_phy_scan_operation_backup(struct ieee80211_hw *hw, u8 operation)
1026{
1027 struct rtl_priv *rtlpriv = rtl_priv(hw);
1028 struct rtl_hal *rtlhal = rtl_hal(rtl_priv(hw));
1029 enum io_type iotype;
1030
1031 if (!is_hal_stop(rtlhal)) {
1032 switch (operation) {
1033 case SCAN_OPT_BACKUP:
1034 rtlhal->current_bandtypebackup =
1035 rtlhal->current_bandtype;
1036 iotype = IO_CMD_PAUSE_DM_BY_SCAN;
1037 rtlpriv->cfg->ops->set_hw_reg(hw, HW_VAR_IO_CMD,
1038 (u8 *)&iotype);
1039 break;
1040 case SCAN_OPT_RESTORE:
1041 iotype = IO_CMD_RESUME_DM_BY_SCAN;
1042 rtlpriv->cfg->ops->set_hw_reg(hw, HW_VAR_IO_CMD,
1043 (u8 *)&iotype);
1044 break;
1045 default:
1046 RT_TRACE(rtlpriv, COMP_ERR, DBG_EMERG,
1047 "Unknown Scan Backup operation\n");
1048 break;
1049 }
1050 }
1051}
1052
1053void rtl92d_phy_set_bw_mode(struct ieee80211_hw *hw, 1025void rtl92d_phy_set_bw_mode(struct ieee80211_hw *hw,
1054 enum nl80211_channel_type ch_type) 1026 enum nl80211_channel_type ch_type)
1055{ 1027{
diff --git a/drivers/net/wireless/rtlwifi/rtl8192de/phy.h b/drivers/net/wireless/rtlwifi/rtl8192de/phy.h
index f074952bf25c..48d5c6835b6a 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192de/phy.h
+++ b/drivers/net/wireless/rtlwifi/rtl8192de/phy.h
@@ -39,9 +39,7 @@
39#define RT_CANNOT_IO(hw) false 39#define RT_CANNOT_IO(hw) false
40#define HIGHPOWER_RADIOA_ARRAYLEN 22 40#define HIGHPOWER_RADIOA_ARRAYLEN 22
41 41
42#define IQK_ADDA_REG_NUM 16
43#define MAX_TOLERANCE 5 42#define MAX_TOLERANCE 5
44#define IQK_DELAY_TIME 1
45 43
46#define APK_BB_REG_NUM 5 44#define APK_BB_REG_NUM 5
47#define APK_AFE_REG_NUM 16 45#define APK_AFE_REG_NUM 16
@@ -127,34 +125,32 @@ static inline void rtl92d_release_cckandrw_pagea_ctl(struct ieee80211_hw *hw,
127 *flag); 125 *flag);
128} 126}
129 127
130extern u32 rtl92d_phy_query_bb_reg(struct ieee80211_hw *hw, 128u32 rtl92d_phy_query_bb_reg(struct ieee80211_hw *hw,
131 u32 regaddr, u32 bitmask); 129 u32 regaddr, u32 bitmask);
132extern void rtl92d_phy_set_bb_reg(struct ieee80211_hw *hw, 130void rtl92d_phy_set_bb_reg(struct ieee80211_hw *hw,
133 u32 regaddr, u32 bitmask, u32 data); 131 u32 regaddr, u32 bitmask, u32 data);
134extern u32 rtl92d_phy_query_rf_reg(struct ieee80211_hw *hw, 132u32 rtl92d_phy_query_rf_reg(struct ieee80211_hw *hw,
135 enum radio_path rfpath, u32 regaddr, 133 enum radio_path rfpath, u32 regaddr,
136 u32 bitmask); 134 u32 bitmask);
137extern void rtl92d_phy_set_rf_reg(struct ieee80211_hw *hw, 135void rtl92d_phy_set_rf_reg(struct ieee80211_hw *hw,
138 enum radio_path rfpath, u32 regaddr, 136 enum radio_path rfpath, u32 regaddr,
139 u32 bitmask, u32 data); 137 u32 bitmask, u32 data);
140extern bool rtl92d_phy_mac_config(struct ieee80211_hw *hw); 138bool rtl92d_phy_mac_config(struct ieee80211_hw *hw);
141extern bool rtl92d_phy_bb_config(struct ieee80211_hw *hw); 139bool rtl92d_phy_bb_config(struct ieee80211_hw *hw);
142extern bool rtl92d_phy_rf_config(struct ieee80211_hw *hw); 140bool rtl92d_phy_rf_config(struct ieee80211_hw *hw);
143extern bool rtl92c_phy_config_rf_with_feaderfile(struct ieee80211_hw *hw, 141bool rtl92c_phy_config_rf_with_feaderfile(struct ieee80211_hw *hw,
144 enum radio_path rfpath); 142 enum radio_path rfpath);
145extern void rtl92d_phy_get_hw_reg_originalvalue(struct ieee80211_hw *hw); 143void rtl92d_phy_get_hw_reg_originalvalue(struct ieee80211_hw *hw);
146extern void rtl92d_phy_set_txpower_level(struct ieee80211_hw *hw, u8 channel); 144void rtl92d_phy_set_txpower_level(struct ieee80211_hw *hw, u8 channel);
147extern void rtl92d_phy_scan_operation_backup(struct ieee80211_hw *hw, 145void rtl92d_phy_set_bw_mode(struct ieee80211_hw *hw,
148 u8 operation); 146 enum nl80211_channel_type ch_type);
149extern void rtl92d_phy_set_bw_mode(struct ieee80211_hw *hw, 147u8 rtl92d_phy_sw_chnl(struct ieee80211_hw *hw);
150 enum nl80211_channel_type ch_type);
151extern u8 rtl92d_phy_sw_chnl(struct ieee80211_hw *hw);
152bool rtl92d_phy_config_rf_with_headerfile(struct ieee80211_hw *hw, 148bool rtl92d_phy_config_rf_with_headerfile(struct ieee80211_hw *hw,
153 enum rf_content content, 149 enum rf_content content,
154 enum radio_path rfpath); 150 enum radio_path rfpath);
155bool rtl92d_phy_set_io_cmd(struct ieee80211_hw *hw, enum io_type iotype); 151bool rtl92d_phy_set_io_cmd(struct ieee80211_hw *hw, enum io_type iotype);
156extern bool rtl92d_phy_set_rf_power_state(struct ieee80211_hw *hw, 152bool rtl92d_phy_set_rf_power_state(struct ieee80211_hw *hw,
157 enum rf_pwrstate rfpwr_state); 153 enum rf_pwrstate rfpwr_state);
158 154
159void rtl92d_phy_config_macphymode(struct ieee80211_hw *hw); 155void rtl92d_phy_config_macphymode(struct ieee80211_hw *hw);
160void rtl92d_phy_config_macphymode_info(struct ieee80211_hw *hw); 156void rtl92d_phy_config_macphymode_info(struct ieee80211_hw *hw);
@@ -173,6 +169,5 @@ void rtl92d_acquire_cckandrw_pagea_ctl(struct ieee80211_hw *hw,
173 unsigned long *flag); 169 unsigned long *flag);
174u8 rtl92d_get_rightchnlplace_for_iqk(u8 chnl); 170u8 rtl92d_get_rightchnlplace_for_iqk(u8 chnl);
175void rtl92d_phy_reload_iqk_setting(struct ieee80211_hw *hw, u8 channel); 171void rtl92d_phy_reload_iqk_setting(struct ieee80211_hw *hw, u8 channel);
176void rtl92d_phy_iq_calibrate(struct ieee80211_hw *hw);
177 172
178#endif 173#endif
diff --git a/drivers/net/wireless/rtlwifi/rtl8192de/rf.h b/drivers/net/wireless/rtlwifi/rtl8192de/rf.h
index 0fe1a48593e8..7303d12c266f 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192de/rf.h
+++ b/drivers/net/wireless/rtlwifi/rtl8192de/rf.h
@@ -30,15 +30,13 @@
30#ifndef __RTL92D_RF_H__ 30#ifndef __RTL92D_RF_H__
31#define __RTL92D_RF_H__ 31#define __RTL92D_RF_H__
32 32
33extern void rtl92d_phy_rf6052_set_bandwidth(struct ieee80211_hw *hw, 33void rtl92d_phy_rf6052_set_bandwidth(struct ieee80211_hw *hw, u8 bandwidth);
34 u8 bandwidth); 34void rtl92d_phy_rf6052_set_cck_txpower(struct ieee80211_hw *hw,
35extern void rtl92d_phy_rf6052_set_cck_txpower(struct ieee80211_hw *hw, 35 u8 *ppowerlevel);
36 u8 *ppowerlevel); 36void rtl92d_phy_rf6052_set_ofdm_txpower(struct ieee80211_hw *hw,
37extern void rtl92d_phy_rf6052_set_ofdm_txpower(struct ieee80211_hw *hw, 37 u8 *ppowerlevel, u8 channel);
38 u8 *ppowerlevel, u8 channel); 38bool rtl92d_phy_rf6052_config(struct ieee80211_hw *hw);
39extern bool rtl92d_phy_rf6052_config(struct ieee80211_hw *hw); 39bool rtl92d_phy_enable_anotherphy(struct ieee80211_hw *hw, bool bmac0);
40extern bool rtl92d_phy_enable_anotherphy(struct ieee80211_hw *hw, bool bmac0); 40void rtl92d_phy_powerdown_anotherphy(struct ieee80211_hw *hw, bool bmac0);
41extern void rtl92d_phy_powerdown_anotherphy(struct ieee80211_hw *hw,
42 bool bmac0);
43 41
44#endif 42#endif
diff --git a/drivers/net/wireless/rtlwifi/rtl8192de/sw.c b/drivers/net/wireless/rtlwifi/rtl8192de/sw.c
index c18c04bf0c13..edab5a5351b5 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192de/sw.c
+++ b/drivers/net/wireless/rtlwifi/rtl8192de/sw.c
@@ -30,6 +30,7 @@
30#include "../wifi.h" 30#include "../wifi.h"
31#include "../core.h" 31#include "../core.h"
32#include "../pci.h" 32#include "../pci.h"
33#include "../base.h"
33#include "reg.h" 34#include "reg.h"
34#include "def.h" 35#include "def.h"
35#include "phy.h" 36#include "phy.h"
@@ -236,7 +237,7 @@ static struct rtl_hal_ops rtl8192de_hal_ops = {
236 .set_bw_mode = rtl92d_phy_set_bw_mode, 237 .set_bw_mode = rtl92d_phy_set_bw_mode,
237 .switch_channel = rtl92d_phy_sw_chnl, 238 .switch_channel = rtl92d_phy_sw_chnl,
238 .dm_watchdog = rtl92d_dm_watchdog, 239 .dm_watchdog = rtl92d_dm_watchdog,
239 .scan_operation_backup = rtl92d_phy_scan_operation_backup, 240 .scan_operation_backup = rtl_phy_scan_operation_backup,
240 .set_rf_power_state = rtl92d_phy_set_rf_power_state, 241 .set_rf_power_state = rtl92d_phy_set_rf_power_state,
241 .led_control = rtl92de_led_control, 242 .led_control = rtl92de_led_control,
242 .set_desc = rtl92de_set_desc, 243 .set_desc = rtl92de_set_desc,
diff --git a/drivers/net/wireless/rtlwifi/rtl8192de/trx.c b/drivers/net/wireless/rtlwifi/rtl8192de/trx.c
index b8ec718a0fab..945ddecf90c9 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192de/trx.c
+++ b/drivers/net/wireless/rtlwifi/rtl8192de/trx.c
@@ -526,7 +526,6 @@ bool rtl92de_rx_query_desc(struct ieee80211_hw *hw, struct rtl_stats *stats,
526 } 526 }
527 /*rx_status->qual = stats->signal; */ 527 /*rx_status->qual = stats->signal; */
528 rx_status->signal = stats->rssi + 10; 528 rx_status->signal = stats->rssi + 10;
529 /*rx_status->noise = -stats->noise; */
530 return true; 529 return true;
531} 530}
532 531
diff --git a/drivers/net/wireless/rtlwifi/rtl8192se/reg.h b/drivers/net/wireless/rtlwifi/rtl8192se/reg.h
index 84d1181795b8..c81c83591940 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192se/reg.h
+++ b/drivers/net/wireless/rtlwifi/rtl8192se/reg.h
@@ -425,14 +425,9 @@
425#define EXT_IMEM_CODE_DONE BIT(2) 425#define EXT_IMEM_CODE_DONE BIT(2)
426#define IMEM_CHK_RPT BIT(1) 426#define IMEM_CHK_RPT BIT(1)
427#define IMEM_CODE_DONE BIT(0) 427#define IMEM_CODE_DONE BIT(0)
428#define IMEM_CODE_DONE BIT(0)
429#define IMEM_CHK_RPT BIT(1)
430#define EMEM_CODE_DONE BIT(2) 428#define EMEM_CODE_DONE BIT(2)
431#define EMEM_CHK_RPT BIT(3) 429#define EMEM_CHK_RPT BIT(3)
432#define DMEM_CODE_DONE BIT(4)
433#define IMEM_RDY BIT(5) 430#define IMEM_RDY BIT(5)
434#define BASECHG BIT(6)
435#define FWRDY BIT(7)
436#define LOAD_FW_READY (IMEM_CODE_DONE | \ 431#define LOAD_FW_READY (IMEM_CODE_DONE | \
437 IMEM_CHK_RPT | \ 432 IMEM_CHK_RPT | \
438 EMEM_CODE_DONE | \ 433 EMEM_CODE_DONE | \
diff --git a/drivers/net/wireless/rtlwifi/rtl8192se/trx.c b/drivers/net/wireless/rtlwifi/rtl8192se/trx.c
index c7095118de6e..222d2e792ca6 100644
--- a/drivers/net/wireless/rtlwifi/rtl8192se/trx.c
+++ b/drivers/net/wireless/rtlwifi/rtl8192se/trx.c
@@ -330,7 +330,6 @@ bool rtl92se_rx_query_desc(struct ieee80211_hw *hw, struct rtl_stats *stats,
330 330
331 /*rx_status->qual = stats->signal; */ 331 /*rx_status->qual = stats->signal; */
332 rx_status->signal = stats->rssi + 10; 332 rx_status->signal = stats->rssi + 10;
333 /*rx_status->noise = -stats->noise; */
334 333
335 return true; 334 return true;
336} 335}
diff --git a/drivers/net/wireless/rtlwifi/rtl8723ae/phy.c b/drivers/net/wireless/rtlwifi/rtl8723ae/phy.c
index eafbb18dd48e..5d318a85eda4 100644
--- a/drivers/net/wireless/rtlwifi/rtl8723ae/phy.c
+++ b/drivers/net/wireless/rtlwifi/rtl8723ae/phy.c
@@ -934,35 +934,6 @@ static long _phy_txpwr_idx_to_dbm(struct ieee80211_hw *hw,
934 return pwrout_dbm; 934 return pwrout_dbm;
935} 935}
936 936
937void rtl8723ae_phy_scan_operation_backup(struct ieee80211_hw *hw, u8 operation)
938{
939 struct rtl_priv *rtlpriv = rtl_priv(hw);
940 struct rtl_hal *rtlhal = rtl_hal(rtl_priv(hw));
941 enum io_type iotype;
942
943 if (!is_hal_stop(rtlhal)) {
944 switch (operation) {
945 case SCAN_OPT_BACKUP:
946 iotype = IO_CMD_PAUSE_DM_BY_SCAN;
947 rtlpriv->cfg->ops->set_hw_reg(hw,
948 HW_VAR_IO_CMD,
949 (u8 *)&iotype);
950
951 break;
952 case SCAN_OPT_RESTORE:
953 iotype = IO_CMD_RESUME_DM_BY_SCAN;
954 rtlpriv->cfg->ops->set_hw_reg(hw,
955 HW_VAR_IO_CMD,
956 (u8 *)&iotype);
957 break;
958 default:
959 RT_TRACE(rtlpriv, COMP_ERR, DBG_EMERG,
960 "Unknown Scan Backup operation.\n");
961 break;
962 }
963 }
964}
965
966void rtl8723ae_phy_set_bw_mode_callback(struct ieee80211_hw *hw) 937void rtl8723ae_phy_set_bw_mode_callback(struct ieee80211_hw *hw)
967{ 938{
968 struct rtl_priv *rtlpriv = rtl_priv(hw); 939 struct rtl_priv *rtlpriv = rtl_priv(hw);
diff --git a/drivers/net/wireless/rtlwifi/rtl8723ae/phy.h b/drivers/net/wireless/rtlwifi/rtl8723ae/phy.h
index e7a59eba351a..007ebdbbe108 100644
--- a/drivers/net/wireless/rtlwifi/rtl8723ae/phy.h
+++ b/drivers/net/wireless/rtlwifi/rtl8723ae/phy.h
@@ -183,42 +183,40 @@ struct tx_power_struct {
183 u32 mcs_original_offset[4][16]; 183 u32 mcs_original_offset[4][16];
184}; 184};
185 185
186extern u32 rtl8723ae_phy_query_bb_reg(struct ieee80211_hw *hw, 186u32 rtl8723ae_phy_query_bb_reg(struct ieee80211_hw *hw,
187 u32 regaddr, u32 bitmask); 187 u32 regaddr, u32 bitmask);
188extern void rtl8723ae_phy_set_bb_reg(struct ieee80211_hw *hw, 188void rtl8723ae_phy_set_bb_reg(struct ieee80211_hw *hw,
189 u32 regaddr, u32 bitmask, u32 data); 189 u32 regaddr, u32 bitmask, u32 data);
190extern u32 rtl8723ae_phy_query_rf_reg(struct ieee80211_hw *hw, 190u32 rtl8723ae_phy_query_rf_reg(struct ieee80211_hw *hw,
191 enum radio_path rfpath, u32 regaddr, 191 enum radio_path rfpath, u32 regaddr,
192 u32 bitmask); 192 u32 bitmask);
193extern void rtl8723ae_phy_set_rf_reg(struct ieee80211_hw *hw, 193void rtl8723ae_phy_set_rf_reg(struct ieee80211_hw *hw,
194 enum radio_path rfpath, u32 regaddr, 194 enum radio_path rfpath, u32 regaddr,
195 u32 bitmask, u32 data); 195 u32 bitmask, u32 data);
196extern bool rtl8723ae_phy_mac_config(struct ieee80211_hw *hw); 196bool rtl8723ae_phy_mac_config(struct ieee80211_hw *hw);
197extern bool rtl8723ae_phy_bb_config(struct ieee80211_hw *hw); 197bool rtl8723ae_phy_bb_config(struct ieee80211_hw *hw);
198extern bool rtl8723ae_phy_rf_config(struct ieee80211_hw *hw); 198bool rtl8723ae_phy_rf_config(struct ieee80211_hw *hw);
199extern bool rtl92c_phy_config_rf_with_feaderfile(struct ieee80211_hw *hw, 199bool rtl92c_phy_config_rf_with_feaderfile(struct ieee80211_hw *hw,
200 enum radio_path rfpath); 200 enum radio_path rfpath);
201extern void rtl8723ae_phy_get_hw_reg_originalvalue(struct ieee80211_hw *hw); 201void rtl8723ae_phy_get_hw_reg_originalvalue(struct ieee80211_hw *hw);
202extern void rtl8723ae_phy_get_txpower_level(struct ieee80211_hw *hw, 202void rtl8723ae_phy_get_txpower_level(struct ieee80211_hw *hw,
203 long *powerlevel); 203 long *powerlevel);
204extern void rtl8723ae_phy_set_txpower_level(struct ieee80211_hw *hw, 204void rtl8723ae_phy_set_txpower_level(struct ieee80211_hw *hw,
205 u8 channel); 205 u8 channel);
206extern bool rtl8723ae_phy_update_txpower_dbm(struct ieee80211_hw *hw, 206bool rtl8723ae_phy_update_txpower_dbm(struct ieee80211_hw *hw,
207 long power_indbm); 207 long power_indbm);
208extern void rtl8723ae_phy_scan_operation_backup(struct ieee80211_hw *hw, 208void rtl8723ae_phy_set_bw_mode_callback(struct ieee80211_hw *hw);
209 u8 operation); 209void rtl8723ae_phy_set_bw_mode(struct ieee80211_hw *hw,
210extern void rtl8723ae_phy_set_bw_mode_callback(struct ieee80211_hw *hw); 210 enum nl80211_channel_type ch_type);
211extern void rtl8723ae_phy_set_bw_mode(struct ieee80211_hw *hw, 211void rtl8723ae_phy_sw_chnl_callback(struct ieee80211_hw *hw);
212 enum nl80211_channel_type ch_type); 212u8 rtl8723ae_phy_sw_chnl(struct ieee80211_hw *hw);
213extern void rtl8723ae_phy_sw_chnl_callback(struct ieee80211_hw *hw); 213void rtl8723ae_phy_iq_calibrate(struct ieee80211_hw *hw, bool recovery);
214extern u8 rtl8723ae_phy_sw_chnl(struct ieee80211_hw *hw);
215extern void rtl8723ae_phy_iq_calibrate(struct ieee80211_hw *hw, bool recovery);
216void rtl8723ae_phy_lc_calibrate(struct ieee80211_hw *hw); 214void rtl8723ae_phy_lc_calibrate(struct ieee80211_hw *hw);
217void rtl8723ae_phy_set_rfpath_switch(struct ieee80211_hw *hw, bool bmain); 215void rtl8723ae_phy_set_rfpath_switch(struct ieee80211_hw *hw, bool bmain);
218bool rtl8723ae_phy_config_rf_with_headerfile(struct ieee80211_hw *hw, 216bool rtl8723ae_phy_config_rf_with_headerfile(struct ieee80211_hw *hw,
219 enum radio_path rfpath); 217 enum radio_path rfpath);
220bool rtl8723ae_phy_set_io_cmd(struct ieee80211_hw *hw, enum io_type iotype); 218bool rtl8723ae_phy_set_io_cmd(struct ieee80211_hw *hw, enum io_type iotype);
221extern bool rtl8723ae_phy_set_rf_power_state(struct ieee80211_hw *hw, 219bool rtl8723ae_phy_set_rf_power_state(struct ieee80211_hw *hw,
222 enum rf_pwrstate rfpwr_state); 220 enum rf_pwrstate rfpwr_state);
223 221
224#endif 222#endif
diff --git a/drivers/net/wireless/rtlwifi/rtl8723ae/rf.h b/drivers/net/wireless/rtlwifi/rtl8723ae/rf.h
index d0f9dd79abea..57f1933ee663 100644
--- a/drivers/net/wireless/rtlwifi/rtl8723ae/rf.h
+++ b/drivers/net/wireless/rtlwifi/rtl8723ae/rf.h
@@ -32,12 +32,11 @@
32 32
33#define RF6052_MAX_TX_PWR 0x3F 33#define RF6052_MAX_TX_PWR 0x3F
34 34
35extern void rtl8723ae_phy_rf6052_set_bandwidth(struct ieee80211_hw *hw, 35void rtl8723ae_phy_rf6052_set_bandwidth(struct ieee80211_hw *hw, u8 bandwidth);
36 u8 bandwidth); 36void rtl8723ae_phy_rf6052_set_cck_txpower(struct ieee80211_hw *hw,
37extern void rtl8723ae_phy_rf6052_set_cck_txpower(struct ieee80211_hw *hw, 37 u8 *ppowerlevel);
38 u8 *ppowerlevel); 38void rtl8723ae_phy_rf6052_set_ofdm_txpower(struct ieee80211_hw *hw,
39extern void rtl8723ae_phy_rf6052_set_ofdm_txpower(struct ieee80211_hw *hw, 39 u8 *ppowerlevel, u8 channel);
40 u8 *ppowerlevel, u8 channel); 40bool rtl8723ae_phy_rf6052_config(struct ieee80211_hw *hw);
41extern bool rtl8723ae_phy_rf6052_config(struct ieee80211_hw *hw);
42 41
43#endif 42#endif
diff --git a/drivers/net/wireless/rtlwifi/rtl8723ae/sw.c b/drivers/net/wireless/rtlwifi/rtl8723ae/sw.c
index d9ee2efffe5f..62b204faf773 100644
--- a/drivers/net/wireless/rtlwifi/rtl8723ae/sw.c
+++ b/drivers/net/wireless/rtlwifi/rtl8723ae/sw.c
@@ -33,6 +33,7 @@
33 33
34#include "../core.h" 34#include "../core.h"
35#include "../pci.h" 35#include "../pci.h"
36#include "../base.h"
36#include "reg.h" 37#include "reg.h"
37#include "def.h" 38#include "def.h"
38#include "phy.h" 39#include "phy.h"
@@ -220,7 +221,7 @@ static struct rtl_hal_ops rtl8723ae_hal_ops = {
220 .set_bw_mode = rtl8723ae_phy_set_bw_mode, 221 .set_bw_mode = rtl8723ae_phy_set_bw_mode,
221 .switch_channel = rtl8723ae_phy_sw_chnl, 222 .switch_channel = rtl8723ae_phy_sw_chnl,
222 .dm_watchdog = rtl8723ae_dm_watchdog, 223 .dm_watchdog = rtl8723ae_dm_watchdog,
223 .scan_operation_backup = rtl8723ae_phy_scan_operation_backup, 224 .scan_operation_backup = rtl_phy_scan_operation_backup,
224 .set_rf_power_state = rtl8723ae_phy_set_rf_power_state, 225 .set_rf_power_state = rtl8723ae_phy_set_rf_power_state,
225 .led_control = rtl8723ae_led_control, 226 .led_control = rtl8723ae_led_control,
226 .set_desc = rtl8723ae_set_desc, 227 .set_desc = rtl8723ae_set_desc,
diff --git a/drivers/net/wireless/rtlwifi/rtl8723ae/trx.c b/drivers/net/wireless/rtlwifi/rtl8723ae/trx.c
index bcd82a1020a5..50b7be3f3a60 100644
--- a/drivers/net/wireless/rtlwifi/rtl8723ae/trx.c
+++ b/drivers/net/wireless/rtlwifi/rtl8723ae/trx.c
@@ -359,7 +359,6 @@ bool rtl8723ae_rx_query_desc(struct ieee80211_hw *hw,
359 359
360 /*rx_status->qual = status->signal; */ 360 /*rx_status->qual = status->signal; */
361 rx_status->signal = status->recvsignalpower + 10; 361 rx_status->signal = status->recvsignalpower + 10;
362 /*rx_status->noise = -status->noise; */
363 362
364 return true; 363 return true;
365} 364}
diff --git a/drivers/net/wireless/rtlwifi/usb.c b/drivers/net/wireless/rtlwifi/usb.c
index e56778cac9bf..6e2b5c5c83c8 100644
--- a/drivers/net/wireless/rtlwifi/usb.c
+++ b/drivers/net/wireless/rtlwifi/usb.c
@@ -455,7 +455,6 @@ static void _rtl_usb_rx_process_agg(struct ieee80211_hw *hw,
455 struct ieee80211_rx_status rx_status = {0}; 455 struct ieee80211_rx_status rx_status = {0};
456 struct rtl_stats stats = { 456 struct rtl_stats stats = {
457 .signal = 0, 457 .signal = 0,
458 .noise = -98,
459 .rate = 0, 458 .rate = 0,
460 }; 459 };
461 460
@@ -498,7 +497,6 @@ static void _rtl_usb_rx_process_noagg(struct ieee80211_hw *hw,
498 struct ieee80211_rx_status rx_status = {0}; 497 struct ieee80211_rx_status rx_status = {0};
499 struct rtl_stats stats = { 498 struct rtl_stats stats = {
500 .signal = 0, 499 .signal = 0,
501 .noise = -98,
502 .rate = 0, 500 .rate = 0,
503 }; 501 };
504 502
@@ -582,12 +580,15 @@ static void _rtl_rx_work(unsigned long param)
582static unsigned int _rtl_rx_get_padding(struct ieee80211_hdr *hdr, 580static unsigned int _rtl_rx_get_padding(struct ieee80211_hdr *hdr,
583 unsigned int len) 581 unsigned int len)
584{ 582{
583#if NET_IP_ALIGN != 0
585 unsigned int padding = 0; 584 unsigned int padding = 0;
585#endif
586 586
587 /* make function no-op when possible */ 587 /* make function no-op when possible */
588 if (NET_IP_ALIGN == 0 || len < sizeof(*hdr)) 588 if (NET_IP_ALIGN == 0 || len < sizeof(*hdr))
589 return 0; 589 return 0;
590 590
591#if NET_IP_ALIGN != 0
591 /* alignment calculation as in lbtf_rx() / carl9170_rx_copy_data() */ 592 /* alignment calculation as in lbtf_rx() / carl9170_rx_copy_data() */
592 /* TODO: deduplicate common code, define helper function instead? */ 593 /* TODO: deduplicate common code, define helper function instead? */
593 594
@@ -608,6 +609,7 @@ static unsigned int _rtl_rx_get_padding(struct ieee80211_hdr *hdr,
608 padding ^= NET_IP_ALIGN; 609 padding ^= NET_IP_ALIGN;
609 610
610 return padding; 611 return padding;
612#endif
611} 613}
612 614
613#define __RADIO_TAP_SIZE_RSV 32 615#define __RADIO_TAP_SIZE_RSV 32
diff --git a/drivers/net/wireless/rtlwifi/wifi.h b/drivers/net/wireless/rtlwifi/wifi.h
index 703258742d28..d224dc3bb092 100644
--- a/drivers/net/wireless/rtlwifi/wifi.h
+++ b/drivers/net/wireless/rtlwifi/wifi.h
@@ -192,8 +192,6 @@ enum hardware_type {
192(IS_HARDWARE_TYPE_8192DE(rtlhal) || IS_HARDWARE_TYPE_8192DU(rtlhal)) 192(IS_HARDWARE_TYPE_8192DE(rtlhal) || IS_HARDWARE_TYPE_8192DU(rtlhal))
193#define IS_HARDWARE_TYPE_8723(rtlhal) \ 193#define IS_HARDWARE_TYPE_8723(rtlhal) \
194(IS_HARDWARE_TYPE_8723E(rtlhal) || IS_HARDWARE_TYPE_8723U(rtlhal)) 194(IS_HARDWARE_TYPE_8723E(rtlhal) || IS_HARDWARE_TYPE_8723U(rtlhal))
195#define IS_HARDWARE_TYPE_8723U(rtlhal) \
196 (rtlhal->hw_type == HARDWARE_TYPE_RTL8723U)
197 195
198#define RX_HAL_IS_CCK_RATE(_pdesc)\ 196#define RX_HAL_IS_CCK_RATE(_pdesc)\
199 (_pdesc->rxmcs == DESC92_RATE1M || \ 197 (_pdesc->rxmcs == DESC92_RATE1M || \
diff --git a/drivers/net/wireless/ti/wl1251/spi.c b/drivers/net/wireless/ti/wl1251/spi.c
index c7dc6feab2ff..1342f81e683d 100644
--- a/drivers/net/wireless/ti/wl1251/spi.c
+++ b/drivers/net/wireless/ti/wl1251/spi.c
@@ -243,7 +243,7 @@ static int wl1251_spi_probe(struct spi_device *spi)
243 struct wl1251 *wl; 243 struct wl1251 *wl;
244 int ret; 244 int ret;
245 245
246 pdata = spi->dev.platform_data; 246 pdata = dev_get_platdata(&spi->dev);
247 if (!pdata) { 247 if (!pdata) {
248 wl1251_error("no platform data"); 248 wl1251_error("no platform data");
249 return -ENODEV; 249 return -ENODEV;
diff --git a/drivers/net/wireless/ti/wl1251/wl1251.h b/drivers/net/wireless/ti/wl1251/wl1251.h
index fd02060038de..2c3bd1bff3f6 100644
--- a/drivers/net/wireless/ti/wl1251/wl1251.h
+++ b/drivers/net/wireless/ti/wl1251/wl1251.h
@@ -424,8 +424,8 @@ void wl1251_disable_interrupts(struct wl1251 *wl);
424#define CHIP_ID_1271_PG10 (0x4030101) 424#define CHIP_ID_1271_PG10 (0x4030101)
425#define CHIP_ID_1271_PG20 (0x4030111) 425#define CHIP_ID_1271_PG20 (0x4030111)
426 426
427#define WL1251_FW_NAME "wl1251-fw.bin" 427#define WL1251_FW_NAME "ti-connectivity/wl1251-fw.bin"
428#define WL1251_NVS_NAME "wl1251-nvs.bin" 428#define WL1251_NVS_NAME "ti-connectivity/wl1251-nvs.bin"
429 429
430#define WL1251_POWER_ON_SLEEP 10 /* in milliseconds */ 430#define WL1251_POWER_ON_SLEEP 10 /* in milliseconds */
431 431
diff --git a/drivers/net/wireless/ti/wl12xx/main.c b/drivers/net/wireless/ti/wl12xx/main.c
index 1c627da85083..be7129ba16ad 100644
--- a/drivers/net/wireless/ti/wl12xx/main.c
+++ b/drivers/net/wireless/ti/wl12xx/main.c
@@ -333,11 +333,11 @@ static struct wlcore_conf wl12xx_conf = {
333 .always = 0, 333 .always = 0,
334 }, 334 },
335 .fwlog = { 335 .fwlog = {
336 .mode = WL12XX_FWLOG_ON_DEMAND, 336 .mode = WL12XX_FWLOG_CONTINUOUS,
337 .mem_blocks = 2, 337 .mem_blocks = 2,
338 .severity = 0, 338 .severity = 0,
339 .timestamp = WL12XX_FWLOG_TIMESTAMP_DISABLED, 339 .timestamp = WL12XX_FWLOG_TIMESTAMP_DISABLED,
340 .output = WL12XX_FWLOG_OUTPUT_HOST, 340 .output = WL12XX_FWLOG_OUTPUT_DBG_PINS,
341 .threshold = 0, 341 .threshold = 0,
342 }, 342 },
343 .rate = { 343 .rate = {
@@ -717,6 +717,9 @@ static int wl12xx_identify_chip(struct wl1271 *wl)
717 goto out; 717 goto out;
718 } 718 }
719 719
720 wl->fw_mem_block_size = 256;
721 wl->fwlog_end = 0x2000000;
722
720 /* common settings */ 723 /* common settings */
721 wl->scan_templ_id_2_4 = CMD_TEMPL_APP_PROBE_REQ_2_4_LEGACY; 724 wl->scan_templ_id_2_4 = CMD_TEMPL_APP_PROBE_REQ_2_4_LEGACY;
722 wl->scan_templ_id_5 = CMD_TEMPL_APP_PROBE_REQ_5_LEGACY; 725 wl->scan_templ_id_5 = CMD_TEMPL_APP_PROBE_REQ_5_LEGACY;
@@ -1262,9 +1265,10 @@ static int wl12xx_boot(struct wl1271 *wl)
1262 BA_SESSION_RX_CONSTRAINT_EVENT_ID | 1265 BA_SESSION_RX_CONSTRAINT_EVENT_ID |
1263 REMAIN_ON_CHANNEL_COMPLETE_EVENT_ID | 1266 REMAIN_ON_CHANNEL_COMPLETE_EVENT_ID |
1264 INACTIVE_STA_EVENT_ID | 1267 INACTIVE_STA_EVENT_ID |
1265 MAX_TX_RETRY_EVENT_ID |
1266 CHANNEL_SWITCH_COMPLETE_EVENT_ID; 1268 CHANNEL_SWITCH_COMPLETE_EVENT_ID;
1267 1269
1270 wl->ap_event_mask = MAX_TX_RETRY_EVENT_ID;
1271
1268 ret = wlcore_boot_run_firmware(wl); 1272 ret = wlcore_boot_run_firmware(wl);
1269 if (ret < 0) 1273 if (ret < 0)
1270 goto out; 1274 goto out;
@@ -1648,6 +1652,11 @@ static bool wl12xx_lnk_low_prio(struct wl1271 *wl, u8 hlid,
1648 return true; 1652 return true;
1649} 1653}
1650 1654
1655static u32 wl12xx_convert_hwaddr(struct wl1271 *wl, u32 hwaddr)
1656{
1657 return hwaddr << 5;
1658}
1659
1651static int wl12xx_setup(struct wl1271 *wl); 1660static int wl12xx_setup(struct wl1271 *wl);
1652 1661
1653static struct wlcore_ops wl12xx_ops = { 1662static struct wlcore_ops wl12xx_ops = {
@@ -1684,6 +1693,7 @@ static struct wlcore_ops wl12xx_ops = {
1684 .channel_switch = wl12xx_cmd_channel_switch, 1693 .channel_switch = wl12xx_cmd_channel_switch,
1685 .pre_pkt_send = NULL, 1694 .pre_pkt_send = NULL,
1686 .set_peer_cap = wl12xx_set_peer_cap, 1695 .set_peer_cap = wl12xx_set_peer_cap,
1696 .convert_hwaddr = wl12xx_convert_hwaddr,
1687 .lnk_high_prio = wl12xx_lnk_high_prio, 1697 .lnk_high_prio = wl12xx_lnk_high_prio,
1688 .lnk_low_prio = wl12xx_lnk_low_prio, 1698 .lnk_low_prio = wl12xx_lnk_low_prio,
1689}; 1699};
@@ -1704,7 +1714,7 @@ static struct ieee80211_sta_ht_cap wl12xx_ht_cap = {
1704static int wl12xx_setup(struct wl1271 *wl) 1714static int wl12xx_setup(struct wl1271 *wl)
1705{ 1715{
1706 struct wl12xx_priv *priv = wl->priv; 1716 struct wl12xx_priv *priv = wl->priv;
1707 struct wlcore_platdev_data *pdev_data = wl->pdev->dev.platform_data; 1717 struct wlcore_platdev_data *pdev_data = dev_get_platdata(&wl->pdev->dev);
1708 struct wl12xx_platform_data *pdata = pdev_data->pdata; 1718 struct wl12xx_platform_data *pdata = pdev_data->pdata;
1709 1719
1710 wl->rtable = wl12xx_rtable; 1720 wl->rtable = wl12xx_rtable;
diff --git a/drivers/net/wireless/ti/wl18xx/main.c b/drivers/net/wireless/ti/wl18xx/main.c
index 7aa0eb848c5a..ec37b16585df 100644
--- a/drivers/net/wireless/ti/wl18xx/main.c
+++ b/drivers/net/wireless/ti/wl18xx/main.c
@@ -456,11 +456,11 @@ static struct wlcore_conf wl18xx_conf = {
456 .always = 0, 456 .always = 0,
457 }, 457 },
458 .fwlog = { 458 .fwlog = {
459 .mode = WL12XX_FWLOG_ON_DEMAND, 459 .mode = WL12XX_FWLOG_CONTINUOUS,
460 .mem_blocks = 2, 460 .mem_blocks = 2,
461 .severity = 0, 461 .severity = 0,
462 .timestamp = WL12XX_FWLOG_TIMESTAMP_DISABLED, 462 .timestamp = WL12XX_FWLOG_TIMESTAMP_DISABLED,
463 .output = WL12XX_FWLOG_OUTPUT_HOST, 463 .output = WL12XX_FWLOG_OUTPUT_DBG_PINS,
464 .threshold = 0, 464 .threshold = 0,
465 }, 465 },
466 .rate = { 466 .rate = {
@@ -505,7 +505,7 @@ static struct wlcore_conf wl18xx_conf = {
505 505
506static struct wl18xx_priv_conf wl18xx_default_priv_conf = { 506static struct wl18xx_priv_conf wl18xx_default_priv_conf = {
507 .ht = { 507 .ht = {
508 .mode = HT_MODE_DEFAULT, 508 .mode = HT_MODE_WIDE,
509 }, 509 },
510 .phy = { 510 .phy = {
511 .phy_standalone = 0x00, 511 .phy_standalone = 0x00,
@@ -516,7 +516,7 @@ static struct wl18xx_priv_conf wl18xx_default_priv_conf = {
516 .auto_detect = 0x00, 516 .auto_detect = 0x00,
517 .dedicated_fem = FEM_NONE, 517 .dedicated_fem = FEM_NONE,
518 .low_band_component = COMPONENT_3_WAY_SWITCH, 518 .low_band_component = COMPONENT_3_WAY_SWITCH,
519 .low_band_component_type = 0x04, 519 .low_band_component_type = 0x05,
520 .high_band_component = COMPONENT_2_WAY_SWITCH, 520 .high_band_component = COMPONENT_2_WAY_SWITCH,
521 .high_band_component_type = 0x09, 521 .high_band_component_type = 0x09,
522 .tcxo_ldo_voltage = 0x00, 522 .tcxo_ldo_voltage = 0x00,
@@ -556,15 +556,15 @@ static struct wl18xx_priv_conf wl18xx_default_priv_conf = {
556 .per_chan_pwr_limit_arr_11p = { 0xff, 0xff, 0xff, 0xff, 556 .per_chan_pwr_limit_arr_11p = { 0xff, 0xff, 0xff, 0xff,
557 0xff, 0xff, 0xff }, 557 0xff, 0xff, 0xff },
558 .psat = 0, 558 .psat = 0,
559 .low_power_val = 0x08,
560 .med_power_val = 0x12,
561 .high_power_val = 0x18,
562 .low_power_val_2nd = 0x05,
563 .med_power_val_2nd = 0x0a,
564 .high_power_val_2nd = 0x14,
565 .external_pa_dc2dc = 0, 559 .external_pa_dc2dc = 0,
566 .number_of_assembled_ant2_4 = 2, 560 .number_of_assembled_ant2_4 = 2,
567 .number_of_assembled_ant5 = 1, 561 .number_of_assembled_ant5 = 1,
562 .low_power_val = 0xff,
563 .med_power_val = 0xff,
564 .high_power_val = 0xff,
565 .low_power_val_2nd = 0xff,
566 .med_power_val_2nd = 0xff,
567 .high_power_val_2nd = 0xff,
568 .tx_rf_margin = 1, 568 .tx_rf_margin = 1,
569 }, 569 },
570}; 570};
@@ -623,6 +623,18 @@ static const int wl18xx_rtable[REG_TABLE_LEN] = {
623 [REG_RAW_FW_STATUS_ADDR] = WL18XX_FW_STATUS_ADDR, 623 [REG_RAW_FW_STATUS_ADDR] = WL18XX_FW_STATUS_ADDR,
624}; 624};
625 625
626static const struct wl18xx_clk_cfg wl18xx_clk_table_coex[NUM_CLOCK_CONFIGS] = {
627 [CLOCK_CONFIG_16_2_M] = { 8, 121, 0, 0, false },
628 [CLOCK_CONFIG_16_368_M] = { 8, 120, 0, 0, false },
629 [CLOCK_CONFIG_16_8_M] = { 8, 117, 0, 0, false },
630 [CLOCK_CONFIG_19_2_M] = { 10, 128, 0, 0, false },
631 [CLOCK_CONFIG_26_M] = { 11, 104, 0, 0, false },
632 [CLOCK_CONFIG_32_736_M] = { 8, 120, 0, 0, false },
633 [CLOCK_CONFIG_33_6_M] = { 8, 117, 0, 0, false },
634 [CLOCK_CONFIG_38_468_M] = { 10, 128, 0, 0, false },
635 [CLOCK_CONFIG_52_M] = { 11, 104, 0, 0, false },
636};
637
626static const struct wl18xx_clk_cfg wl18xx_clk_table[NUM_CLOCK_CONFIGS] = { 638static const struct wl18xx_clk_cfg wl18xx_clk_table[NUM_CLOCK_CONFIGS] = {
627 [CLOCK_CONFIG_16_2_M] = { 7, 104, 801, 4, true }, 639 [CLOCK_CONFIG_16_2_M] = { 7, 104, 801, 4, true },
628 [CLOCK_CONFIG_16_368_M] = { 9, 132, 3751, 4, true }, 640 [CLOCK_CONFIG_16_368_M] = { 9, 132, 3751, 4, true },
@@ -674,6 +686,9 @@ static int wl18xx_identify_chip(struct wl1271 *wl)
674 goto out; 686 goto out;
675 } 687 }
676 688
689 wl->fw_mem_block_size = 272;
690 wl->fwlog_end = 0x40000000;
691
677 wl->scan_templ_id_2_4 = CMD_TEMPL_CFG_PROBE_REQ_2_4; 692 wl->scan_templ_id_2_4 = CMD_TEMPL_CFG_PROBE_REQ_2_4;
678 wl->scan_templ_id_5 = CMD_TEMPL_CFG_PROBE_REQ_5; 693 wl->scan_templ_id_5 = CMD_TEMPL_CFG_PROBE_REQ_5;
679 wl->sched_scan_templ_id_2_4 = CMD_TEMPL_PROBE_REQ_2_4_PERIODIC; 694 wl->sched_scan_templ_id_2_4 = CMD_TEMPL_PROBE_REQ_2_4_PERIODIC;
@@ -704,6 +719,23 @@ static int wl18xx_set_clk(struct wl1271 *wl)
704 wl18xx_clk_table[clk_freq].p, wl18xx_clk_table[clk_freq].q, 719 wl18xx_clk_table[clk_freq].p, wl18xx_clk_table[clk_freq].q,
705 wl18xx_clk_table[clk_freq].swallow ? "swallow" : "spit"); 720 wl18xx_clk_table[clk_freq].swallow ? "swallow" : "spit");
706 721
722 /* coex PLL configuration */
723 ret = wl18xx_top_reg_write(wl, PLLSH_COEX_PLL_N,
724 wl18xx_clk_table_coex[clk_freq].n);
725 if (ret < 0)
726 goto out;
727
728 ret = wl18xx_top_reg_write(wl, PLLSH_COEX_PLL_M,
729 wl18xx_clk_table_coex[clk_freq].m);
730 if (ret < 0)
731 goto out;
732
733 /* bypass the swallowing logic */
734 ret = wl18xx_top_reg_write(wl, PLLSH_COEX_PLL_SWALLOW_EN,
735 PLLSH_COEX_PLL_SWALLOW_EN_VAL1);
736 if (ret < 0)
737 goto out;
738
707 ret = wl18xx_top_reg_write(wl, PLLSH_WCS_PLL_N, 739 ret = wl18xx_top_reg_write(wl, PLLSH_WCS_PLL_N,
708 wl18xx_clk_table[clk_freq].n); 740 wl18xx_clk_table[clk_freq].n);
709 if (ret < 0) 741 if (ret < 0)
@@ -745,6 +777,30 @@ static int wl18xx_set_clk(struct wl1271 *wl)
745 PLLSH_WCS_PLL_SWALLOW_EN_VAL2); 777 PLLSH_WCS_PLL_SWALLOW_EN_VAL2);
746 } 778 }
747 779
780 /* choose WCS PLL */
781 ret = wl18xx_top_reg_write(wl, PLLSH_WL_PLL_SEL,
782 PLLSH_WL_PLL_SEL_WCS_PLL);
783 if (ret < 0)
784 goto out;
785
786 /* enable both PLLs */
787 ret = wl18xx_top_reg_write(wl, PLLSH_WL_PLL_EN, PLLSH_WL_PLL_EN_VAL1);
788 if (ret < 0)
789 goto out;
790
791 udelay(1000);
792
793 /* disable coex PLL */
794 ret = wl18xx_top_reg_write(wl, PLLSH_WL_PLL_EN, PLLSH_WL_PLL_EN_VAL2);
795 if (ret < 0)
796 goto out;
797
798 /* reset the swallowing logic */
799 ret = wl18xx_top_reg_write(wl, PLLSH_COEX_PLL_SWALLOW_EN,
800 PLLSH_COEX_PLL_SWALLOW_EN_VAL2);
801 if (ret < 0)
802 goto out;
803
748out: 804out:
749 return ret; 805 return ret;
750} 806}
@@ -935,10 +991,11 @@ static int wl18xx_boot(struct wl1271 *wl)
935 BA_SESSION_RX_CONSTRAINT_EVENT_ID | 991 BA_SESSION_RX_CONSTRAINT_EVENT_ID |
936 REMAIN_ON_CHANNEL_COMPLETE_EVENT_ID | 992 REMAIN_ON_CHANNEL_COMPLETE_EVENT_ID |
937 INACTIVE_STA_EVENT_ID | 993 INACTIVE_STA_EVENT_ID |
938 MAX_TX_FAILURE_EVENT_ID |
939 CHANNEL_SWITCH_COMPLETE_EVENT_ID | 994 CHANNEL_SWITCH_COMPLETE_EVENT_ID |
940 DFS_CHANNELS_CONFIG_COMPLETE_EVENT; 995 DFS_CHANNELS_CONFIG_COMPLETE_EVENT;
941 996
997 wl->ap_event_mask = MAX_TX_FAILURE_EVENT_ID;
998
942 ret = wlcore_boot_run_firmware(wl); 999 ret = wlcore_boot_run_firmware(wl);
943 if (ret < 0) 1000 if (ret < 0)
944 goto out; 1001 goto out;
@@ -1175,16 +1232,48 @@ static u32 wl18xx_ap_get_mimo_wide_rate_mask(struct wl1271 *wl,
1175 } 1232 }
1176} 1233}
1177 1234
1235static const char *wl18xx_rdl_name(enum wl18xx_rdl_num rdl_num)
1236{
1237 switch (rdl_num) {
1238 case RDL_1_HP:
1239 return "183xH";
1240 case RDL_2_SP:
1241 return "183x or 180x";
1242 case RDL_3_HP:
1243 return "187xH";
1244 case RDL_4_SP:
1245 return "187x";
1246 case RDL_5_SP:
1247 return "RDL11 - Not Supported";
1248 case RDL_6_SP:
1249 return "180xD";
1250 case RDL_7_SP:
1251 return "RDL13 - Not Supported (1893Q)";
1252 case RDL_8_SP:
1253 return "18xxQ";
1254 case RDL_NONE:
1255 return "UNTRIMMED";
1256 default:
1257 return "UNKNOWN";
1258 }
1259}
1260
1178static int wl18xx_get_pg_ver(struct wl1271 *wl, s8 *ver) 1261static int wl18xx_get_pg_ver(struct wl1271 *wl, s8 *ver)
1179{ 1262{
1180 u32 fuse; 1263 u32 fuse;
1181 s8 rom = 0, metal = 0, pg_ver = 0, rdl_ver = 0; 1264 s8 rom = 0, metal = 0, pg_ver = 0, rdl_ver = 0, package_type = 0;
1182 int ret; 1265 int ret;
1183 1266
1184 ret = wlcore_set_partition(wl, &wl->ptable[PART_TOP_PRCM_ELP_SOC]); 1267 ret = wlcore_set_partition(wl, &wl->ptable[PART_TOP_PRCM_ELP_SOC]);
1185 if (ret < 0) 1268 if (ret < 0)
1186 goto out; 1269 goto out;
1187 1270
1271 ret = wlcore_read32(wl, WL18XX_REG_FUSE_DATA_2_3, &fuse);
1272 if (ret < 0)
1273 goto out;
1274
1275 package_type = (fuse >> WL18XX_PACKAGE_TYPE_OFFSET) & 1;
1276
1188 ret = wlcore_read32(wl, WL18XX_REG_FUSE_DATA_1_3, &fuse); 1277 ret = wlcore_read32(wl, WL18XX_REG_FUSE_DATA_1_3, &fuse);
1189 if (ret < 0) 1278 if (ret < 0)
1190 goto out; 1279 goto out;
@@ -1192,7 +1281,7 @@ static int wl18xx_get_pg_ver(struct wl1271 *wl, s8 *ver)
1192 pg_ver = (fuse & WL18XX_PG_VER_MASK) >> WL18XX_PG_VER_OFFSET; 1281 pg_ver = (fuse & WL18XX_PG_VER_MASK) >> WL18XX_PG_VER_OFFSET;
1193 rom = (fuse & WL18XX_ROM_VER_MASK) >> WL18XX_ROM_VER_OFFSET; 1282 rom = (fuse & WL18XX_ROM_VER_MASK) >> WL18XX_ROM_VER_OFFSET;
1194 1283
1195 if (rom <= 0xE) 1284 if ((rom <= 0xE) && (package_type == WL18XX_PACKAGE_TYPE_WSP))
1196 metal = (fuse & WL18XX_METAL_VER_MASK) >> 1285 metal = (fuse & WL18XX_METAL_VER_MASK) >>
1197 WL18XX_METAL_VER_OFFSET; 1286 WL18XX_METAL_VER_OFFSET;
1198 else 1287 else
@@ -1204,11 +1293,9 @@ static int wl18xx_get_pg_ver(struct wl1271 *wl, s8 *ver)
1204 goto out; 1293 goto out;
1205 1294
1206 rdl_ver = (fuse & WL18XX_RDL_VER_MASK) >> WL18XX_RDL_VER_OFFSET; 1295 rdl_ver = (fuse & WL18XX_RDL_VER_MASK) >> WL18XX_RDL_VER_OFFSET;
1207 if (rdl_ver > RDL_MAX)
1208 rdl_ver = RDL_NONE;
1209 1296
1210 wl1271_info("wl18xx HW: RDL %d, %s, PG %x.%x (ROM %x)", 1297 wl1271_info("wl18xx HW: %s, PG %d.%d (ROM 0x%x)",
1211 rdl_ver, rdl_names[rdl_ver], pg_ver, metal, rom); 1298 wl18xx_rdl_name(rdl_ver), pg_ver, metal, rom);
1212 1299
1213 if (ver) 1300 if (ver)
1214 *ver = pg_ver; 1301 *ver = pg_ver;
@@ -1521,6 +1608,11 @@ static bool wl18xx_lnk_low_prio(struct wl1271 *wl, u8 hlid,
1521 return lnk->allocated_pkts < thold; 1608 return lnk->allocated_pkts < thold;
1522} 1609}
1523 1610
1611static u32 wl18xx_convert_hwaddr(struct wl1271 *wl, u32 hwaddr)
1612{
1613 return hwaddr & ~0x80000000;
1614}
1615
1524static int wl18xx_setup(struct wl1271 *wl); 1616static int wl18xx_setup(struct wl1271 *wl);
1525 1617
1526static struct wlcore_ops wl18xx_ops = { 1618static struct wlcore_ops wl18xx_ops = {
@@ -1558,6 +1650,7 @@ static struct wlcore_ops wl18xx_ops = {
1558 .pre_pkt_send = wl18xx_pre_pkt_send, 1650 .pre_pkt_send = wl18xx_pre_pkt_send,
1559 .sta_rc_update = wl18xx_sta_rc_update, 1651 .sta_rc_update = wl18xx_sta_rc_update,
1560 .set_peer_cap = wl18xx_set_peer_cap, 1652 .set_peer_cap = wl18xx_set_peer_cap,
1653 .convert_hwaddr = wl18xx_convert_hwaddr,
1561 .lnk_high_prio = wl18xx_lnk_high_prio, 1654 .lnk_high_prio = wl18xx_lnk_high_prio,
1562 .lnk_low_prio = wl18xx_lnk_low_prio, 1655 .lnk_low_prio = wl18xx_lnk_low_prio,
1563}; 1656};
diff --git a/drivers/net/wireless/ti/wl18xx/reg.h b/drivers/net/wireless/ti/wl18xx/reg.h
index 05dd8bad2746..a433a75f3cd7 100644
--- a/drivers/net/wireless/ti/wl18xx/reg.h
+++ b/drivers/net/wireless/ti/wl18xx/reg.h
@@ -114,6 +114,11 @@
114#define PLATFORM_DETECTION 0xA0E3E0 114#define PLATFORM_DETECTION 0xA0E3E0
115#define OCS_EN 0xA02080 115#define OCS_EN 0xA02080
116#define PRIMARY_CLK_DETECT 0xA020A6 116#define PRIMARY_CLK_DETECT 0xA020A6
117#define PLLSH_COEX_PLL_N 0xA02384
118#define PLLSH_COEX_PLL_M 0xA02382
119#define PLLSH_COEX_PLL_SWALLOW_EN 0xA0238E
120#define PLLSH_WL_PLL_SEL 0xA02398
121
117#define PLLSH_WCS_PLL_N 0xA02362 122#define PLLSH_WCS_PLL_N 0xA02362
118#define PLLSH_WCS_PLL_M 0xA02360 123#define PLLSH_WCS_PLL_M 0xA02360
119#define PLLSH_WCS_PLL_Q_FACTOR_CFG_1 0xA02364 124#define PLLSH_WCS_PLL_Q_FACTOR_CFG_1 0xA02364
@@ -128,19 +133,30 @@
128#define PLLSH_WCS_PLL_P_FACTOR_CFG_1_MASK 0xFFFF 133#define PLLSH_WCS_PLL_P_FACTOR_CFG_1_MASK 0xFFFF
129#define PLLSH_WCS_PLL_P_FACTOR_CFG_2_MASK 0x000F 134#define PLLSH_WCS_PLL_P_FACTOR_CFG_2_MASK 0x000F
130 135
136#define PLLSH_WL_PLL_EN_VAL1 0x7
137#define PLLSH_WL_PLL_EN_VAL2 0x2
138#define PLLSH_COEX_PLL_SWALLOW_EN_VAL1 0x2
139#define PLLSH_COEX_PLL_SWALLOW_EN_VAL2 0x11
140
131#define PLLSH_WCS_PLL_SWALLOW_EN_VAL1 0x1 141#define PLLSH_WCS_PLL_SWALLOW_EN_VAL1 0x1
132#define PLLSH_WCS_PLL_SWALLOW_EN_VAL2 0x12 142#define PLLSH_WCS_PLL_SWALLOW_EN_VAL2 0x12
133 143
144#define PLLSH_WL_PLL_SEL_WCS_PLL 0x0
145#define PLLSH_WL_PLL_SEL_COEX_PLL 0x1
146
134#define WL18XX_REG_FUSE_DATA_1_3 0xA0260C 147#define WL18XX_REG_FUSE_DATA_1_3 0xA0260C
135#define WL18XX_PG_VER_MASK 0x70 148#define WL18XX_PG_VER_MASK 0x70
136#define WL18XX_PG_VER_OFFSET 4 149#define WL18XX_PG_VER_OFFSET 4
137#define WL18XX_ROM_VER_MASK 0x3 150#define WL18XX_ROM_VER_MASK 0x3e00
138#define WL18XX_ROM_VER_OFFSET 0 151#define WL18XX_ROM_VER_OFFSET 9
139#define WL18XX_METAL_VER_MASK 0xC 152#define WL18XX_METAL_VER_MASK 0xC
140#define WL18XX_METAL_VER_OFFSET 2 153#define WL18XX_METAL_VER_OFFSET 2
141#define WL18XX_NEW_METAL_VER_MASK 0x180 154#define WL18XX_NEW_METAL_VER_MASK 0x180
142#define WL18XX_NEW_METAL_VER_OFFSET 7 155#define WL18XX_NEW_METAL_VER_OFFSET 7
143 156
157#define WL18XX_PACKAGE_TYPE_OFFSET 13
158#define WL18XX_PACKAGE_TYPE_WSP 0
159
144#define WL18XX_REG_FUSE_DATA_2_3 0xA02614 160#define WL18XX_REG_FUSE_DATA_2_3 0xA02614
145#define WL18XX_RDL_VER_MASK 0x1f00 161#define WL18XX_RDL_VER_MASK 0x1f00
146#define WL18XX_RDL_VER_OFFSET 8 162#define WL18XX_RDL_VER_OFFSET 8
@@ -201,24 +217,21 @@ enum {
201 NUM_BOARD_TYPES, 217 NUM_BOARD_TYPES,
202}; 218};
203 219
204enum { 220enum wl18xx_rdl_num {
205 RDL_NONE = 0, 221 RDL_NONE = 0,
206 RDL_1_HP = 1, 222 RDL_1_HP = 1,
207 RDL_2_SP = 2, 223 RDL_2_SP = 2,
208 RDL_3_HP = 3, 224 RDL_3_HP = 3,
209 RDL_4_SP = 4, 225 RDL_4_SP = 4,
226 RDL_5_SP = 0x11,
227 RDL_6_SP = 0x12,
228 RDL_7_SP = 0x13,
229 RDL_8_SP = 0x14,
210 230
211 _RDL_LAST, 231 _RDL_LAST,
212 RDL_MAX = _RDL_LAST - 1, 232 RDL_MAX = _RDL_LAST - 1,
213}; 233};
214 234
215static const char * const rdl_names[] = {
216 [RDL_NONE] = "",
217 [RDL_1_HP] = "1853 SISO",
218 [RDL_2_SP] = "1857 MIMO",
219 [RDL_3_HP] = "1893 SISO",
220 [RDL_4_SP] = "1897 MIMO",
221};
222 235
223/* FPGA_SPARE_1 register - used to change the PHY ATPG clock at boot time */ 236/* FPGA_SPARE_1 register - used to change the PHY ATPG clock at boot time */
224#define WL18XX_PHY_FPGA_SPARE_1 0x8093CA40 237#define WL18XX_PHY_FPGA_SPARE_1 0x8093CA40
diff --git a/drivers/net/wireless/ti/wlcore/acx.c b/drivers/net/wireless/ti/wlcore/acx.c
index 7a970cd9c555..ec83675a2446 100644
--- a/drivers/net/wireless/ti/wlcore/acx.c
+++ b/drivers/net/wireless/ti/wlcore/acx.c
@@ -162,7 +162,8 @@ int wl1271_acx_mem_map(struct wl1271 *wl, struct acx_header *mem_map,
162 162
163 wl1271_debug(DEBUG_ACX, "acx mem map"); 163 wl1271_debug(DEBUG_ACX, "acx mem map");
164 164
165 ret = wl1271_cmd_interrogate(wl, ACX_MEM_MAP, mem_map, len); 165 ret = wl1271_cmd_interrogate(wl, ACX_MEM_MAP, mem_map,
166 sizeof(struct acx_header), len);
166 if (ret < 0) 167 if (ret < 0)
167 return ret; 168 return ret;
168 169
@@ -722,6 +723,7 @@ int wl1271_acx_statistics(struct wl1271 *wl, void *stats)
722 wl1271_debug(DEBUG_ACX, "acx statistics"); 723 wl1271_debug(DEBUG_ACX, "acx statistics");
723 724
724 ret = wl1271_cmd_interrogate(wl, ACX_STATISTICS, stats, 725 ret = wl1271_cmd_interrogate(wl, ACX_STATISTICS, stats,
726 sizeof(struct acx_header),
725 wl->stats.fw_stats_len); 727 wl->stats.fw_stats_len);
726 if (ret < 0) { 728 if (ret < 0) {
727 wl1271_warning("acx statistics failed: %d", ret); 729 wl1271_warning("acx statistics failed: %d", ret);
@@ -1470,8 +1472,8 @@ int wl12xx_acx_tsf_info(struct wl1271 *wl, struct wl12xx_vif *wlvif,
1470 1472
1471 tsf_info->role_id = wlvif->role_id; 1473 tsf_info->role_id = wlvif->role_id;
1472 1474
1473 ret = wl1271_cmd_interrogate(wl, ACX_TSF_INFO, 1475 ret = wl1271_cmd_interrogate(wl, ACX_TSF_INFO, tsf_info,
1474 tsf_info, sizeof(*tsf_info)); 1476 sizeof(struct acx_header), sizeof(*tsf_info));
1475 if (ret < 0) { 1477 if (ret < 0) {
1476 wl1271_warning("acx tsf info interrogate failed"); 1478 wl1271_warning("acx tsf info interrogate failed");
1477 goto out; 1479 goto out;
@@ -1752,7 +1754,7 @@ int wlcore_acx_average_rssi(struct wl1271 *wl, struct wl12xx_vif *wlvif,
1752 1754
1753 acx->role_id = wlvif->role_id; 1755 acx->role_id = wlvif->role_id;
1754 ret = wl1271_cmd_interrogate(wl, ACX_ROAMING_STATISTICS_TBL, 1756 ret = wl1271_cmd_interrogate(wl, ACX_ROAMING_STATISTICS_TBL,
1755 acx, sizeof(*acx)); 1757 acx, sizeof(*acx), sizeof(*acx));
1756 if (ret < 0) { 1758 if (ret < 0) {
1757 wl1271_warning("acx roaming statistics failed: %d", ret); 1759 wl1271_warning("acx roaming statistics failed: %d", ret);
1758 ret = -ENOMEM; 1760 ret = -ENOMEM;
diff --git a/drivers/net/wireless/ti/wlcore/cmd.c b/drivers/net/wireless/ti/wlcore/cmd.c
index c9e060795d13..34d9dfff2ad3 100644
--- a/drivers/net/wireless/ti/wlcore/cmd.c
+++ b/drivers/net/wireless/ti/wlcore/cmd.c
@@ -60,7 +60,8 @@ static int __wlcore_cmd_send(struct wl1271 *wl, u16 id, void *buf,
60 u16 status; 60 u16 status;
61 u16 poll_count = 0; 61 u16 poll_count = 0;
62 62
63 if (WARN_ON(unlikely(wl->state == WLCORE_STATE_RESTARTING))) 63 if (WARN_ON(wl->state == WLCORE_STATE_RESTARTING &&
64 id != CMD_STOP_FWLOGGER))
64 return -EIO; 65 return -EIO;
65 66
66 cmd = buf; 67 cmd = buf;
@@ -845,7 +846,8 @@ EXPORT_SYMBOL_GPL(wl1271_cmd_test);
845 * @buf: buffer for the response, including all headers, must work with dma 846 * @buf: buffer for the response, including all headers, must work with dma
846 * @len: length of buf 847 * @len: length of buf
847 */ 848 */
848int wl1271_cmd_interrogate(struct wl1271 *wl, u16 id, void *buf, size_t len) 849int wl1271_cmd_interrogate(struct wl1271 *wl, u16 id, void *buf,
850 size_t cmd_len, size_t res_len)
849{ 851{
850 struct acx_header *acx = buf; 852 struct acx_header *acx = buf;
851 int ret; 853 int ret;
@@ -854,10 +856,10 @@ int wl1271_cmd_interrogate(struct wl1271 *wl, u16 id, void *buf, size_t len)
854 856
855 acx->id = cpu_to_le16(id); 857 acx->id = cpu_to_le16(id);
856 858
857 /* payload length, does not include any headers */ 859 /* response payload length, does not include any headers */
858 acx->len = cpu_to_le16(len - sizeof(*acx)); 860 acx->len = cpu_to_le16(res_len - sizeof(*acx));
859 861
860 ret = wl1271_cmd_send(wl, CMD_INTERROGATE, acx, sizeof(*acx), len); 862 ret = wl1271_cmd_send(wl, CMD_INTERROGATE, acx, cmd_len, res_len);
861 if (ret < 0) 863 if (ret < 0)
862 wl1271_error("INTERROGATE command failed"); 864 wl1271_error("INTERROGATE command failed");
863 865
@@ -1126,6 +1128,8 @@ int wl12xx_cmd_build_probe_req(struct wl1271 *wl, struct wl12xx_vif *wlvif,
1126 u16 template_id_2_4 = wl->scan_templ_id_2_4; 1128 u16 template_id_2_4 = wl->scan_templ_id_2_4;
1127 u16 template_id_5 = wl->scan_templ_id_5; 1129 u16 template_id_5 = wl->scan_templ_id_5;
1128 1130
1131 wl1271_debug(DEBUG_SCAN, "build probe request band %d", band);
1132
1129 skb = ieee80211_probereq_get(wl->hw, vif, ssid, ssid_len, 1133 skb = ieee80211_probereq_get(wl->hw, vif, ssid, ssid_len,
1130 ie_len); 1134 ie_len);
1131 if (!skb) { 1135 if (!skb) {
@@ -1135,8 +1139,6 @@ int wl12xx_cmd_build_probe_req(struct wl1271 *wl, struct wl12xx_vif *wlvif,
1135 if (ie_len) 1139 if (ie_len)
1136 memcpy(skb_put(skb, ie_len), ie, ie_len); 1140 memcpy(skb_put(skb, ie_len), ie, ie_len);
1137 1141
1138 wl1271_dump(DEBUG_SCAN, "PROBE REQ: ", skb->data, skb->len);
1139
1140 if (sched_scan && 1142 if (sched_scan &&
1141 (wl->quirks & WLCORE_QUIRK_DUAL_PROBE_TMPL)) { 1143 (wl->quirks & WLCORE_QUIRK_DUAL_PROBE_TMPL)) {
1142 template_id_2_4 = wl->sched_scan_templ_id_2_4; 1144 template_id_2_4 = wl->sched_scan_templ_id_2_4;
@@ -1172,7 +1174,7 @@ struct sk_buff *wl1271_cmd_build_ap_probe_req(struct wl1271 *wl,
1172 if (!skb) 1174 if (!skb)
1173 goto out; 1175 goto out;
1174 1176
1175 wl1271_dump(DEBUG_SCAN, "AP PROBE REQ: ", skb->data, skb->len); 1177 wl1271_debug(DEBUG_SCAN, "set ap probe request template");
1176 1178
1177 rate = wl1271_tx_min_rate_get(wl, wlvif->bitrate_masks[wlvif->band]); 1179 rate = wl1271_tx_min_rate_get(wl, wlvif->bitrate_masks[wlvif->band]);
1178 if (wlvif->band == IEEE80211_BAND_2GHZ) 1180 if (wlvif->band == IEEE80211_BAND_2GHZ)
@@ -1607,33 +1609,43 @@ out:
1607 1609
1608static int wlcore_get_reg_conf_ch_idx(enum ieee80211_band band, u16 ch) 1610static int wlcore_get_reg_conf_ch_idx(enum ieee80211_band band, u16 ch)
1609{ 1611{
1610 int idx = -1; 1612 /*
1611 1613 * map the given band/channel to the respective predefined
1614 * bit expected by the fw
1615 */
1612 switch (band) { 1616 switch (band) {
1613 case IEEE80211_BAND_5GHZ:
1614 if (ch >= 8 && ch <= 16)
1615 idx = ((ch-8)/4 + 18);
1616 else if (ch >= 34 && ch <= 64)
1617 idx = ((ch-34)/2 + 3 + 18);
1618 else if (ch >= 100 && ch <= 140)
1619 idx = ((ch-100)/4 + 15 + 18);
1620 else if (ch >= 149 && ch <= 165)
1621 idx = ((ch-149)/4 + 26 + 18);
1622 else
1623 idx = -1;
1624 break;
1625 case IEEE80211_BAND_2GHZ: 1617 case IEEE80211_BAND_2GHZ:
1618 /* channels 1..14 are mapped to 0..13 */
1626 if (ch >= 1 && ch <= 14) 1619 if (ch >= 1 && ch <= 14)
1627 idx = ch - 1; 1620 return ch - 1;
1628 else 1621 break;
1629 idx = -1; 1622 case IEEE80211_BAND_5GHZ:
1623 switch (ch) {
1624 case 8 ... 16:
1625 /* channels 8,12,16 are mapped to 18,19,20 */
1626 return 18 + (ch-8)/4;
1627 case 34 ... 48:
1628 /* channels 34,36..48 are mapped to 21..28 */
1629 return 21 + (ch-34)/2;
1630 case 52 ... 64:
1631 /* channels 52,56..64 are mapped to 29..32 */
1632 return 29 + (ch-52)/4;
1633 case 100 ... 140:
1634 /* channels 100,104..140 are mapped to 33..43 */
1635 return 33 + (ch-100)/4;
1636 case 149 ... 165:
1637 /* channels 149,153..165 are mapped to 44..48 */
1638 return 44 + (ch-149)/4;
1639 default:
1640 break;
1641 }
1630 break; 1642 break;
1631 default: 1643 default:
1632 wl1271_error("get reg conf ch idx - unknown band: %d", 1644 break;
1633 (int)band);
1634 } 1645 }
1635 1646
1636 return idx; 1647 wl1271_error("%s: unknown band/channel: %d/%d", __func__, band, ch);
1648 return -1;
1637} 1649}
1638 1650
1639void wlcore_set_pending_regdomain_ch(struct wl1271 *wl, u16 channel, 1651void wlcore_set_pending_regdomain_ch(struct wl1271 *wl, u16 channel,
@@ -1646,7 +1658,7 @@ void wlcore_set_pending_regdomain_ch(struct wl1271 *wl, u16 channel,
1646 1658
1647 ch_bit_idx = wlcore_get_reg_conf_ch_idx(band, channel); 1659 ch_bit_idx = wlcore_get_reg_conf_ch_idx(band, channel);
1648 1660
1649 if (ch_bit_idx > 0 && ch_bit_idx <= WL1271_MAX_CHANNELS) 1661 if (ch_bit_idx >= 0 && ch_bit_idx <= WL1271_MAX_CHANNELS)
1650 set_bit(ch_bit_idx, (long *)wl->reg_ch_conf_pending); 1662 set_bit(ch_bit_idx, (long *)wl->reg_ch_conf_pending);
1651} 1663}
1652 1664
diff --git a/drivers/net/wireless/ti/wlcore/cmd.h b/drivers/net/wireless/ti/wlcore/cmd.h
index fd34123047cd..323d4a856e4b 100644
--- a/drivers/net/wireless/ti/wlcore/cmd.h
+++ b/drivers/net/wireless/ti/wlcore/cmd.h
@@ -45,7 +45,8 @@ int wl12xx_start_dev(struct wl1271 *wl, struct wl12xx_vif *wlvif,
45 enum ieee80211_band band, int channel); 45 enum ieee80211_band band, int channel);
46int wl12xx_stop_dev(struct wl1271 *wl, struct wl12xx_vif *wlvif); 46int wl12xx_stop_dev(struct wl1271 *wl, struct wl12xx_vif *wlvif);
47int wl1271_cmd_test(struct wl1271 *wl, void *buf, size_t buf_len, u8 answer); 47int wl1271_cmd_test(struct wl1271 *wl, void *buf, size_t buf_len, u8 answer);
48int wl1271_cmd_interrogate(struct wl1271 *wl, u16 id, void *buf, size_t len); 48int wl1271_cmd_interrogate(struct wl1271 *wl, u16 id, void *buf,
49 size_t cmd_len, size_t res_len);
49int wl1271_cmd_configure(struct wl1271 *wl, u16 id, void *buf, size_t len); 50int wl1271_cmd_configure(struct wl1271 *wl, u16 id, void *buf, size_t len);
50int wlcore_cmd_configure_failsafe(struct wl1271 *wl, u16 id, void *buf, 51int wlcore_cmd_configure_failsafe(struct wl1271 *wl, u16 id, void *buf,
51 size_t len, unsigned long valid_rets); 52 size_t len, unsigned long valid_rets);
diff --git a/drivers/net/wireless/ti/wlcore/conf.h b/drivers/net/wireless/ti/wlcore/conf.h
index 2b96ff821341..40995c42bef8 100644
--- a/drivers/net/wireless/ti/wlcore/conf.h
+++ b/drivers/net/wireless/ti/wlcore/conf.h
@@ -1274,6 +1274,9 @@ struct conf_rx_streaming_settings {
1274 u8 always; 1274 u8 always;
1275} __packed; 1275} __packed;
1276 1276
1277#define CONF_FWLOG_MIN_MEM_BLOCKS 2
1278#define CONF_FWLOG_MAX_MEM_BLOCKS 16
1279
1277struct conf_fwlog { 1280struct conf_fwlog {
1278 /* Continuous or on-demand */ 1281 /* Continuous or on-demand */
1279 u8 mode; 1282 u8 mode;
@@ -1281,7 +1284,7 @@ struct conf_fwlog {
1281 /* 1284 /*
1282 * Number of memory blocks dedicated for the FW logger 1285 * Number of memory blocks dedicated for the FW logger
1283 * 1286 *
1284 * Range: 1-3, or 0 to disable the FW logger 1287 * Range: 2-16, or 0 to disable the FW logger
1285 */ 1288 */
1286 u8 mem_blocks; 1289 u8 mem_blocks;
1287 1290
diff --git a/drivers/net/wireless/ti/wlcore/debugfs.c b/drivers/net/wireless/ti/wlcore/debugfs.c
index e17630c2a849..89893c717025 100644
--- a/drivers/net/wireless/ti/wlcore/debugfs.c
+++ b/drivers/net/wireless/ti/wlcore/debugfs.c
@@ -437,6 +437,7 @@ static ssize_t driver_state_read(struct file *file, char __user *user_buf,
437 int res = 0; 437 int res = 0;
438 ssize_t ret; 438 ssize_t ret;
439 char *buf; 439 char *buf;
440 struct wl12xx_vif *wlvif;
440 441
441#define DRIVER_STATE_BUF_LEN 1024 442#define DRIVER_STATE_BUF_LEN 1024
442 443
@@ -450,12 +451,28 @@ static ssize_t driver_state_read(struct file *file, char __user *user_buf,
450 (res += scnprintf(buf + res, DRIVER_STATE_BUF_LEN - res,\ 451 (res += scnprintf(buf + res, DRIVER_STATE_BUF_LEN - res,\
451 #x " = " fmt "\n", wl->x)) 452 #x " = " fmt "\n", wl->x))
452 453
454#define DRIVER_STATE_PRINT_GENERIC(x, fmt, args...) \
455 (res += scnprintf(buf + res, DRIVER_STATE_BUF_LEN - res,\
456 #x " = " fmt "\n", args))
457
453#define DRIVER_STATE_PRINT_LONG(x) DRIVER_STATE_PRINT(x, "%ld") 458#define DRIVER_STATE_PRINT_LONG(x) DRIVER_STATE_PRINT(x, "%ld")
454#define DRIVER_STATE_PRINT_INT(x) DRIVER_STATE_PRINT(x, "%d") 459#define DRIVER_STATE_PRINT_INT(x) DRIVER_STATE_PRINT(x, "%d")
455#define DRIVER_STATE_PRINT_STR(x) DRIVER_STATE_PRINT(x, "%s") 460#define DRIVER_STATE_PRINT_STR(x) DRIVER_STATE_PRINT(x, "%s")
456#define DRIVER_STATE_PRINT_LHEX(x) DRIVER_STATE_PRINT(x, "0x%lx") 461#define DRIVER_STATE_PRINT_LHEX(x) DRIVER_STATE_PRINT(x, "0x%lx")
457#define DRIVER_STATE_PRINT_HEX(x) DRIVER_STATE_PRINT(x, "0x%x") 462#define DRIVER_STATE_PRINT_HEX(x) DRIVER_STATE_PRINT(x, "0x%x")
458 463
464 wl12xx_for_each_wlvif_sta(wl, wlvif) {
465 if (!test_bit(WLVIF_FLAG_STA_ASSOCIATED, &wlvif->flags))
466 continue;
467
468 DRIVER_STATE_PRINT_GENERIC(channel, "%d (%s)", wlvif->channel,
469 wlvif->p2p ? "P2P-CL" : "STA");
470 }
471
472 wl12xx_for_each_wlvif_ap(wl, wlvif)
473 DRIVER_STATE_PRINT_GENERIC(channel, "%d (%s)", wlvif->channel,
474 wlvif->p2p ? "P2P-GO" : "AP");
475
459 DRIVER_STATE_PRINT_INT(tx_blocks_available); 476 DRIVER_STATE_PRINT_INT(tx_blocks_available);
460 DRIVER_STATE_PRINT_INT(tx_allocated_blocks); 477 DRIVER_STATE_PRINT_INT(tx_allocated_blocks);
461 DRIVER_STATE_PRINT_INT(tx_allocated_pkts[0]); 478 DRIVER_STATE_PRINT_INT(tx_allocated_pkts[0]);
@@ -474,7 +491,6 @@ static ssize_t driver_state_read(struct file *file, char __user *user_buf,
474 DRIVER_STATE_PRINT_INT(tx_blocks_freed); 491 DRIVER_STATE_PRINT_INT(tx_blocks_freed);
475 DRIVER_STATE_PRINT_INT(rx_counter); 492 DRIVER_STATE_PRINT_INT(rx_counter);
476 DRIVER_STATE_PRINT_INT(state); 493 DRIVER_STATE_PRINT_INT(state);
477 DRIVER_STATE_PRINT_INT(channel);
478 DRIVER_STATE_PRINT_INT(band); 494 DRIVER_STATE_PRINT_INT(band);
479 DRIVER_STATE_PRINT_INT(power_level); 495 DRIVER_STATE_PRINT_INT(power_level);
480 DRIVER_STATE_PRINT_INT(sg_enabled); 496 DRIVER_STATE_PRINT_INT(sg_enabled);
diff --git a/drivers/net/wireless/ti/wlcore/event.c b/drivers/net/wireless/ti/wlcore/event.c
index 67f61689b49e..8d3b34965db3 100644
--- a/drivers/net/wireless/ti/wlcore/event.c
+++ b/drivers/net/wireless/ti/wlcore/event.c
@@ -266,6 +266,7 @@ int wl1271_event_unmask(struct wl1271 *wl)
266{ 266{
267 int ret; 267 int ret;
268 268
269 wl1271_debug(DEBUG_EVENT, "unmasking event_mask 0x%x", wl->event_mask);
269 ret = wl1271_acx_event_mbox_mask(wl, ~(wl->event_mask)); 270 ret = wl1271_acx_event_mbox_mask(wl, ~(wl->event_mask));
270 if (ret < 0) 271 if (ret < 0)
271 return ret; 272 return ret;
diff --git a/drivers/net/wireless/ti/wlcore/hw_ops.h b/drivers/net/wireless/ti/wlcore/hw_ops.h
index 7fd260c02a0a..51f8d634d32f 100644
--- a/drivers/net/wireless/ti/wlcore/hw_ops.h
+++ b/drivers/net/wireless/ti/wlcore/hw_ops.h
@@ -222,6 +222,15 @@ wlcore_hw_set_peer_cap(struct wl1271 *wl,
222 return 0; 222 return 0;
223} 223}
224 224
225static inline u32
226wlcore_hw_convert_hwaddr(struct wl1271 *wl, u32 hwaddr)
227{
228 if (!wl->ops->convert_hwaddr)
229 BUG_ON(1);
230
231 return wl->ops->convert_hwaddr(wl, hwaddr);
232}
233
225static inline bool 234static inline bool
226wlcore_hw_lnk_high_prio(struct wl1271 *wl, u8 hlid, 235wlcore_hw_lnk_high_prio(struct wl1271 *wl, u8 hlid,
227 struct wl1271_link *lnk) 236 struct wl1271_link *lnk)
diff --git a/drivers/net/wireless/ti/wlcore/init.c b/drivers/net/wireless/ti/wlcore/init.c
index 5c6f11e157d9..7699f9d07e26 100644
--- a/drivers/net/wireless/ti/wlcore/init.c
+++ b/drivers/net/wireless/ti/wlcore/init.c
@@ -571,6 +571,12 @@ int wl1271_init_vif_specific(struct wl1271 *wl, struct ieee80211_vif *vif)
571 ret = wl1271_acx_sleep_auth(wl, WL1271_PSM_CAM); 571 ret = wl1271_acx_sleep_auth(wl, WL1271_PSM_CAM);
572 if (ret < 0) 572 if (ret < 0)
573 return ret; 573 return ret;
574
575 /* unmask ap events */
576 wl->event_mask |= wl->ap_event_mask;
577 ret = wl1271_event_unmask(wl);
578 if (ret < 0)
579 return ret;
574 /* first STA, no APs */ 580 /* first STA, no APs */
575 } else if (wl->sta_count == 0 && wl->ap_count == 0 && !is_ap) { 581 } else if (wl->sta_count == 0 && wl->ap_count == 0 && !is_ap) {
576 u8 sta_auth = wl->conf.conn.sta_sleep_auth; 582 u8 sta_auth = wl->conf.conn.sta_sleep_auth;
diff --git a/drivers/net/wireless/ti/wlcore/io.h b/drivers/net/wireless/ti/wlcore/io.h
index af7d9f9b3b4d..07e3d6a049ad 100644
--- a/drivers/net/wireless/ti/wlcore/io.h
+++ b/drivers/net/wireless/ti/wlcore/io.h
@@ -165,8 +165,8 @@ static inline int __must_check wlcore_read_hwaddr(struct wl1271 *wl, int hwaddr,
165 int physical; 165 int physical;
166 int addr; 166 int addr;
167 167
168 /* Addresses are stored internally as addresses to 32 bytes blocks */ 168 /* Convert from FW internal address which is chip arch dependent */
169 addr = hwaddr << 5; 169 addr = wl->ops->convert_hwaddr(wl, hwaddr);
170 170
171 physical = wlcore_translate_addr(wl, addr); 171 physical = wlcore_translate_addr(wl, addr);
172 172
diff --git a/drivers/net/wireless/ti/wlcore/main.c b/drivers/net/wireless/ti/wlcore/main.c
index 38995f90040d..0368b9cbfb89 100644
--- a/drivers/net/wireless/ti/wlcore/main.c
+++ b/drivers/net/wireless/ti/wlcore/main.c
@@ -44,6 +44,7 @@
44#define WL1271_BOOT_RETRIES 3 44#define WL1271_BOOT_RETRIES 3
45 45
46static char *fwlog_param; 46static char *fwlog_param;
47static int fwlog_mem_blocks = -1;
47static int bug_on_recovery = -1; 48static int bug_on_recovery = -1;
48static int no_recovery = -1; 49static int no_recovery = -1;
49 50
@@ -291,6 +292,18 @@ static void wlcore_adjust_conf(struct wl1271 *wl)
291{ 292{
292 /* Adjust settings according to optional module parameters */ 293 /* Adjust settings according to optional module parameters */
293 294
295 /* Firmware Logger params */
296 if (fwlog_mem_blocks != -1) {
297 if (fwlog_mem_blocks >= CONF_FWLOG_MIN_MEM_BLOCKS &&
298 fwlog_mem_blocks <= CONF_FWLOG_MAX_MEM_BLOCKS) {
299 wl->conf.fwlog.mem_blocks = fwlog_mem_blocks;
300 } else {
301 wl1271_error(
302 "Illegal fwlog_mem_blocks=%d using default %d",
303 fwlog_mem_blocks, wl->conf.fwlog.mem_blocks);
304 }
305 }
306
294 if (fwlog_param) { 307 if (fwlog_param) {
295 if (!strcmp(fwlog_param, "continuous")) { 308 if (!strcmp(fwlog_param, "continuous")) {
296 wl->conf.fwlog.mode = WL12XX_FWLOG_CONTINUOUS; 309 wl->conf.fwlog.mode = WL12XX_FWLOG_CONTINUOUS;
@@ -780,6 +793,7 @@ void wl12xx_queue_recovery_work(struct wl1271 *wl)
780 if (wl->state == WLCORE_STATE_ON) { 793 if (wl->state == WLCORE_STATE_ON) {
781 wl->state = WLCORE_STATE_RESTARTING; 794 wl->state = WLCORE_STATE_RESTARTING;
782 set_bit(WL1271_FLAG_RECOVERY_IN_PROGRESS, &wl->flags); 795 set_bit(WL1271_FLAG_RECOVERY_IN_PROGRESS, &wl->flags);
796 wl1271_ps_elp_wakeup(wl);
783 wlcore_disable_interrupts_nosync(wl); 797 wlcore_disable_interrupts_nosync(wl);
784 ieee80211_queue_work(wl->hw, &wl->recovery_work); 798 ieee80211_queue_work(wl->hw, &wl->recovery_work);
785 } 799 }
@@ -787,19 +801,10 @@ void wl12xx_queue_recovery_work(struct wl1271 *wl)
787 801
788size_t wl12xx_copy_fwlog(struct wl1271 *wl, u8 *memblock, size_t maxlen) 802size_t wl12xx_copy_fwlog(struct wl1271 *wl, u8 *memblock, size_t maxlen)
789{ 803{
790 size_t len = 0; 804 size_t len;
791
792 /* The FW log is a length-value list, find where the log end */
793 while (len < maxlen) {
794 if (memblock[len] == 0)
795 break;
796 if (len + memblock[len] + 1 > maxlen)
797 break;
798 len += memblock[len] + 1;
799 }
800 805
801 /* Make sure we have enough room */ 806 /* Make sure we have enough room */
802 len = min(len, (size_t)(PAGE_SIZE - wl->fwlog_size)); 807 len = min(maxlen, (size_t)(PAGE_SIZE - wl->fwlog_size));
803 808
804 /* Fill the FW log file, consumed by the sysfs fwlog entry */ 809 /* Fill the FW log file, consumed by the sysfs fwlog entry */
805 memcpy(wl->fwlog + wl->fwlog_size, memblock, len); 810 memcpy(wl->fwlog + wl->fwlog_size, memblock, len);
@@ -808,10 +813,9 @@ size_t wl12xx_copy_fwlog(struct wl1271 *wl, u8 *memblock, size_t maxlen)
808 return len; 813 return len;
809} 814}
810 815
811#define WLCORE_FW_LOG_END 0x2000000
812
813static void wl12xx_read_fwlog_panic(struct wl1271 *wl) 816static void wl12xx_read_fwlog_panic(struct wl1271 *wl)
814{ 817{
818 struct wlcore_partition_set part, old_part;
815 u32 addr; 819 u32 addr;
816 u32 offset; 820 u32 offset;
817 u32 end_of_log; 821 u32 end_of_log;
@@ -824,7 +828,7 @@ static void wl12xx_read_fwlog_panic(struct wl1271 *wl)
824 828
825 wl1271_info("Reading FW panic log"); 829 wl1271_info("Reading FW panic log");
826 830
827 block = kmalloc(WL12XX_HW_BLOCK_SIZE, GFP_KERNEL); 831 block = kmalloc(wl->fw_mem_block_size, GFP_KERNEL);
828 if (!block) 832 if (!block)
829 return; 833 return;
830 834
@@ -850,17 +854,31 @@ static void wl12xx_read_fwlog_panic(struct wl1271 *wl)
850 854
851 if (wl->conf.fwlog.mode == WL12XX_FWLOG_CONTINUOUS) { 855 if (wl->conf.fwlog.mode == WL12XX_FWLOG_CONTINUOUS) {
852 offset = sizeof(addr) + sizeof(struct wl1271_rx_descriptor); 856 offset = sizeof(addr) + sizeof(struct wl1271_rx_descriptor);
853 end_of_log = WLCORE_FW_LOG_END; 857 end_of_log = wl->fwlog_end;
854 } else { 858 } else {
855 offset = sizeof(addr); 859 offset = sizeof(addr);
856 end_of_log = addr; 860 end_of_log = addr;
857 } 861 }
858 862
863 old_part = wl->curr_part;
864 memset(&part, 0, sizeof(part));
865
859 /* Traverse the memory blocks linked list */ 866 /* Traverse the memory blocks linked list */
860 do { 867 do {
861 memset(block, 0, WL12XX_HW_BLOCK_SIZE); 868 part.mem.start = wlcore_hw_convert_hwaddr(wl, addr);
862 ret = wlcore_read_hwaddr(wl, addr, block, WL12XX_HW_BLOCK_SIZE, 869 part.mem.size = PAGE_SIZE;
863 false); 870
871 ret = wlcore_set_partition(wl, &part);
872 if (ret < 0) {
873 wl1271_error("%s: set_partition start=0x%X size=%d",
874 __func__, part.mem.start, part.mem.size);
875 goto out;
876 }
877
878 memset(block, 0, wl->fw_mem_block_size);
879 ret = wlcore_read_hwaddr(wl, addr, block,
880 wl->fw_mem_block_size, false);
881
864 if (ret < 0) 882 if (ret < 0)
865 goto out; 883 goto out;
866 884
@@ -871,8 +889,9 @@ static void wl12xx_read_fwlog_panic(struct wl1271 *wl)
871 * on demand mode and is equal to 0x2000000 in continuous mode. 889 * on demand mode and is equal to 0x2000000 in continuous mode.
872 */ 890 */
873 addr = le32_to_cpup((__le32 *)block); 891 addr = le32_to_cpup((__le32 *)block);
892
874 if (!wl12xx_copy_fwlog(wl, block + offset, 893 if (!wl12xx_copy_fwlog(wl, block + offset,
875 WL12XX_HW_BLOCK_SIZE - offset)) 894 wl->fw_mem_block_size - offset))
876 break; 895 break;
877 } while (addr && (addr != end_of_log)); 896 } while (addr && (addr != end_of_log));
878 897
@@ -880,6 +899,7 @@ static void wl12xx_read_fwlog_panic(struct wl1271 *wl)
880 899
881out: 900out:
882 kfree(block); 901 kfree(block);
902 wlcore_set_partition(wl, &old_part);
883} 903}
884 904
885static void wlcore_print_recovery(struct wl1271 *wl) 905static void wlcore_print_recovery(struct wl1271 *wl)
@@ -924,7 +944,8 @@ static void wl1271_recovery_work(struct work_struct *work)
924 goto out_unlock; 944 goto out_unlock;
925 945
926 if (!test_bit(WL1271_FLAG_INTENDED_FW_RECOVERY, &wl->flags)) { 946 if (!test_bit(WL1271_FLAG_INTENDED_FW_RECOVERY, &wl->flags)) {
927 wl12xx_read_fwlog_panic(wl); 947 if (wl->conf.fwlog.output == WL12XX_FWLOG_OUTPUT_HOST)
948 wl12xx_read_fwlog_panic(wl);
928 wlcore_print_recovery(wl); 949 wlcore_print_recovery(wl);
929 } 950 }
930 951
@@ -1062,7 +1083,8 @@ int wl1271_plt_start(struct wl1271 *wl, const enum plt_mode plt_mode)
1062 static const char* const PLT_MODE[] = { 1083 static const char* const PLT_MODE[] = {
1063 "PLT_OFF", 1084 "PLT_OFF",
1064 "PLT_ON", 1085 "PLT_ON",
1065 "PLT_FEM_DETECT" 1086 "PLT_FEM_DETECT",
1087 "PLT_CHIP_AWAKE"
1066 }; 1088 };
1067 1089
1068 int ret; 1090 int ret;
@@ -1088,9 +1110,11 @@ int wl1271_plt_start(struct wl1271 *wl, const enum plt_mode plt_mode)
1088 if (ret < 0) 1110 if (ret < 0)
1089 goto power_off; 1111 goto power_off;
1090 1112
1091 ret = wl->ops->plt_init(wl); 1113 if (plt_mode != PLT_CHIP_AWAKE) {
1092 if (ret < 0) 1114 ret = wl->ops->plt_init(wl);
1093 goto power_off; 1115 if (ret < 0)
1116 goto power_off;
1117 }
1094 1118
1095 wl->state = WLCORE_STATE_ON; 1119 wl->state = WLCORE_STATE_ON;
1096 wl1271_notice("firmware booted in PLT mode %s (%s)", 1120 wl1271_notice("firmware booted in PLT mode %s (%s)",
@@ -1925,8 +1949,10 @@ static void wlcore_op_stop_locked(struct wl1271 *wl)
1925 1949
1926 /* 1950 /*
1927 * FW channels must be re-calibrated after recovery, 1951 * FW channels must be re-calibrated after recovery,
1928 * clear the last Reg-Domain channel configuration. 1952 * save current Reg-Domain channel configuration and clear it.
1929 */ 1953 */
1954 memcpy(wl->reg_ch_conf_pending, wl->reg_ch_conf_last,
1955 sizeof(wl->reg_ch_conf_pending));
1930 memset(wl->reg_ch_conf_last, 0, sizeof(wl->reg_ch_conf_last)); 1956 memset(wl->reg_ch_conf_last, 0, sizeof(wl->reg_ch_conf_last));
1931} 1957}
1932 1958
@@ -2008,6 +2034,47 @@ out:
2008 mutex_unlock(&wl->mutex); 2034 mutex_unlock(&wl->mutex);
2009} 2035}
2010 2036
2037static void wlcore_pending_auth_complete_work(struct work_struct *work)
2038{
2039 struct delayed_work *dwork;
2040 struct wl1271 *wl;
2041 struct wl12xx_vif *wlvif;
2042 unsigned long time_spare;
2043 int ret;
2044
2045 dwork = container_of(work, struct delayed_work, work);
2046 wlvif = container_of(dwork, struct wl12xx_vif,
2047 pending_auth_complete_work);
2048 wl = wlvif->wl;
2049
2050 mutex_lock(&wl->mutex);
2051
2052 if (unlikely(wl->state != WLCORE_STATE_ON))
2053 goto out;
2054
2055 /*
2056 * Make sure a second really passed since the last auth reply. Maybe
2057 * a second auth reply arrived while we were stuck on the mutex.
2058 * Check for a little less than the timeout to protect from scheduler
2059 * irregularities.
2060 */
2061 time_spare = jiffies +
2062 msecs_to_jiffies(WLCORE_PEND_AUTH_ROC_TIMEOUT - 50);
2063 if (!time_after(time_spare, wlvif->pending_auth_reply_time))
2064 goto out;
2065
2066 ret = wl1271_ps_elp_wakeup(wl);
2067 if (ret < 0)
2068 goto out;
2069
2070 /* cancel the ROC if active */
2071 wlcore_update_inconn_sta(wl, wlvif, NULL, false);
2072
2073 wl1271_ps_elp_sleep(wl);
2074out:
2075 mutex_unlock(&wl->mutex);
2076}
2077
2011static int wl12xx_allocate_rate_policy(struct wl1271 *wl, u8 *idx) 2078static int wl12xx_allocate_rate_policy(struct wl1271 *wl, u8 *idx)
2012{ 2079{
2013 u8 policy = find_first_zero_bit(wl->rate_policies_map, 2080 u8 policy = find_first_zero_bit(wl->rate_policies_map,
@@ -2159,6 +2226,8 @@ static int wl12xx_init_vif_data(struct wl1271 *wl, struct ieee80211_vif *vif)
2159 wlcore_channel_switch_work); 2226 wlcore_channel_switch_work);
2160 INIT_DELAYED_WORK(&wlvif->connection_loss_work, 2227 INIT_DELAYED_WORK(&wlvif->connection_loss_work,
2161 wlcore_connection_loss_work); 2228 wlcore_connection_loss_work);
2229 INIT_DELAYED_WORK(&wlvif->pending_auth_complete_work,
2230 wlcore_pending_auth_complete_work);
2162 INIT_LIST_HEAD(&wlvif->list); 2231 INIT_LIST_HEAD(&wlvif->list);
2163 2232
2164 setup_timer(&wlvif->rx_streaming_timer, wl1271_rx_streaming_timer, 2233 setup_timer(&wlvif->rx_streaming_timer, wl1271_rx_streaming_timer,
@@ -2376,6 +2445,11 @@ static int wl1271_op_add_interface(struct ieee80211_hw *hw,
2376 int ret = 0; 2445 int ret = 0;
2377 u8 role_type; 2446 u8 role_type;
2378 2447
2448 if (wl->plt) {
2449 wl1271_error("Adding Interface not allowed while in PLT mode");
2450 return -EBUSY;
2451 }
2452
2379 vif->driver_flags |= IEEE80211_VIF_BEACON_FILTER | 2453 vif->driver_flags |= IEEE80211_VIF_BEACON_FILTER |
2380 IEEE80211_VIF_SUPPORTS_CQM_RSSI; 2454 IEEE80211_VIF_SUPPORTS_CQM_RSSI;
2381 2455
@@ -2572,6 +2646,12 @@ deinit:
2572 !test_bit(WL1271_FLAG_INTENDED_FW_RECOVERY, &wl->flags)) 2646 !test_bit(WL1271_FLAG_INTENDED_FW_RECOVERY, &wl->flags))
2573 goto unlock; 2647 goto unlock;
2574 2648
2649 if (wl->ap_count == 0 && is_ap) {
2650 /* mask ap events */
2651 wl->event_mask &= ~wl->ap_event_mask;
2652 wl1271_event_unmask(wl);
2653 }
2654
2575 if (wl->ap_count == 0 && is_ap && wl->sta_count) { 2655 if (wl->ap_count == 0 && is_ap && wl->sta_count) {
2576 u8 sta_auth = wl->conf.conn.sta_sleep_auth; 2656 u8 sta_auth = wl->conf.conn.sta_sleep_auth;
2577 /* Configure for power according to debugfs */ 2657 /* Configure for power according to debugfs */
@@ -2590,6 +2670,7 @@ unlock:
2590 cancel_work_sync(&wlvif->rx_streaming_disable_work); 2670 cancel_work_sync(&wlvif->rx_streaming_disable_work);
2591 cancel_delayed_work_sync(&wlvif->connection_loss_work); 2671 cancel_delayed_work_sync(&wlvif->connection_loss_work);
2592 cancel_delayed_work_sync(&wlvif->channel_switch_work); 2672 cancel_delayed_work_sync(&wlvif->channel_switch_work);
2673 cancel_delayed_work_sync(&wlvif->pending_auth_complete_work);
2593 2674
2594 mutex_lock(&wl->mutex); 2675 mutex_lock(&wl->mutex);
2595} 2676}
@@ -2875,6 +2956,25 @@ static void wl1271_set_band_rate(struct wl1271 *wl, struct wl12xx_vif *wlvif)
2875 wlvif->rate_set = wlvif->basic_rate_set; 2956 wlvif->rate_set = wlvif->basic_rate_set;
2876} 2957}
2877 2958
2959static void wl1271_sta_handle_idle(struct wl1271 *wl, struct wl12xx_vif *wlvif,
2960 bool idle)
2961{
2962 bool cur_idle = !test_bit(WLVIF_FLAG_ACTIVE, &wlvif->flags);
2963
2964 if (idle == cur_idle)
2965 return;
2966
2967 if (idle) {
2968 clear_bit(WLVIF_FLAG_ACTIVE, &wlvif->flags);
2969 } else {
2970 /* The current firmware only supports sched_scan in idle */
2971 if (wl->sched_vif == wlvif)
2972 wl->ops->sched_scan_stop(wl, wlvif);
2973
2974 set_bit(WLVIF_FLAG_ACTIVE, &wlvif->flags);
2975 }
2976}
2977
2878static int wl12xx_config_vif(struct wl1271 *wl, struct wl12xx_vif *wlvif, 2978static int wl12xx_config_vif(struct wl1271 *wl, struct wl12xx_vif *wlvif,
2879 struct ieee80211_conf *conf, u32 changed) 2979 struct ieee80211_conf *conf, u32 changed)
2880{ 2980{
@@ -3969,6 +4069,13 @@ static void wl1271_bss_info_changed_ap(struct wl1271 *wl,
3969 } 4069 }
3970 } else { 4070 } else {
3971 if (test_bit(WLVIF_FLAG_AP_STARTED, &wlvif->flags)) { 4071 if (test_bit(WLVIF_FLAG_AP_STARTED, &wlvif->flags)) {
4072 /*
4073 * AP might be in ROC in case we have just
4074 * sent auth reply. handle it.
4075 */
4076 if (test_bit(wlvif->role_id, wl->roc_map))
4077 wl12xx_croc(wl, wlvif->role_id);
4078
3972 ret = wl12xx_cmd_role_stop_ap(wl, wlvif); 4079 ret = wl12xx_cmd_role_stop_ap(wl, wlvif);
3973 if (ret < 0) 4080 if (ret < 0)
3974 goto out; 4081 goto out;
@@ -4120,6 +4227,9 @@ static void wl1271_bss_info_changed_sta(struct wl1271 *wl,
4120 do_join = true; 4227 do_join = true;
4121 } 4228 }
4122 4229
4230 if (changed & BSS_CHANGED_IDLE && !is_ibss)
4231 wl1271_sta_handle_idle(wl, wlvif, bss_conf->idle);
4232
4123 if (changed & BSS_CHANGED_CQM) { 4233 if (changed & BSS_CHANGED_CQM) {
4124 bool enable = false; 4234 bool enable = false;
4125 if (bss_conf->cqm_rssi_thold) 4235 if (bss_conf->cqm_rssi_thold)
@@ -4656,29 +4766,49 @@ static void wlcore_roc_if_possible(struct wl1271 *wl,
4656 wl12xx_roc(wl, wlvif, wlvif->role_id, wlvif->band, wlvif->channel); 4766 wl12xx_roc(wl, wlvif, wlvif->role_id, wlvif->band, wlvif->channel);
4657} 4767}
4658 4768
4659static void wlcore_update_inconn_sta(struct wl1271 *wl, 4769/*
4660 struct wl12xx_vif *wlvif, 4770 * when wl_sta is NULL, we treat this call as if coming from a
4661 struct wl1271_station *wl_sta, 4771 * pending auth reply.
4662 bool in_connection) 4772 * wl->mutex must be taken and the FW must be awake when the call
4773 * takes place.
4774 */
4775void wlcore_update_inconn_sta(struct wl1271 *wl, struct wl12xx_vif *wlvif,
4776 struct wl1271_station *wl_sta, bool in_conn)
4663{ 4777{
4664 if (in_connection) { 4778 if (in_conn) {
4665 if (WARN_ON(wl_sta->in_connection)) 4779 if (WARN_ON(wl_sta && wl_sta->in_connection))
4666 return; 4780 return;
4667 wl_sta->in_connection = true; 4781
4668 if (!wlvif->inconn_count++) 4782 if (!wlvif->ap_pending_auth_reply &&
4783 !wlvif->inconn_count)
4669 wlcore_roc_if_possible(wl, wlvif); 4784 wlcore_roc_if_possible(wl, wlvif);
4785
4786 if (wl_sta) {
4787 wl_sta->in_connection = true;
4788 wlvif->inconn_count++;
4789 } else {
4790 wlvif->ap_pending_auth_reply = true;
4791 }
4670 } else { 4792 } else {
4671 if (!wl_sta->in_connection) 4793 if (wl_sta && !wl_sta->in_connection)
4672 return; 4794 return;
4673 4795
4674 wl_sta->in_connection = false; 4796 if (WARN_ON(!wl_sta && !wlvif->ap_pending_auth_reply))
4675 wlvif->inconn_count--;
4676 if (WARN_ON(wlvif->inconn_count < 0))
4677 return; 4797 return;
4678 4798
4679 if (!wlvif->inconn_count) 4799 if (WARN_ON(wl_sta && !wlvif->inconn_count))
4680 if (test_bit(wlvif->role_id, wl->roc_map)) 4800 return;
4681 wl12xx_croc(wl, wlvif->role_id); 4801
4802 if (wl_sta) {
4803 wl_sta->in_connection = false;
4804 wlvif->inconn_count--;
4805 } else {
4806 wlvif->ap_pending_auth_reply = false;
4807 }
4808
4809 if (!wlvif->inconn_count && !wlvif->ap_pending_auth_reply &&
4810 test_bit(wlvif->role_id, wl->roc_map))
4811 wl12xx_croc(wl, wlvif->role_id);
4682 } 4812 }
4683} 4813}
4684 4814
@@ -5313,10 +5443,7 @@ static struct ieee80211_rate wl1271_rates_5ghz[] = {
5313 5443
5314/* 5 GHz band channels for WL1273 */ 5444/* 5 GHz band channels for WL1273 */
5315static struct ieee80211_channel wl1271_channels_5ghz[] = { 5445static struct ieee80211_channel wl1271_channels_5ghz[] = {
5316 { .hw_value = 7, .center_freq = 5035, .max_power = WLCORE_MAX_TXPWR },
5317 { .hw_value = 8, .center_freq = 5040, .max_power = WLCORE_MAX_TXPWR }, 5446 { .hw_value = 8, .center_freq = 5040, .max_power = WLCORE_MAX_TXPWR },
5318 { .hw_value = 9, .center_freq = 5045, .max_power = WLCORE_MAX_TXPWR },
5319 { .hw_value = 11, .center_freq = 5055, .max_power = WLCORE_MAX_TXPWR },
5320 { .hw_value = 12, .center_freq = 5060, .max_power = WLCORE_MAX_TXPWR }, 5447 { .hw_value = 12, .center_freq = 5060, .max_power = WLCORE_MAX_TXPWR },
5321 { .hw_value = 16, .center_freq = 5080, .max_power = WLCORE_MAX_TXPWR }, 5448 { .hw_value = 16, .center_freq = 5080, .max_power = WLCORE_MAX_TXPWR },
5322 { .hw_value = 34, .center_freq = 5170, .max_power = WLCORE_MAX_TXPWR }, 5449 { .hw_value = 34, .center_freq = 5170, .max_power = WLCORE_MAX_TXPWR },
@@ -5896,14 +6023,20 @@ static const struct wiphy_wowlan_support wlcore_wowlan_support = {
5896}; 6023};
5897#endif 6024#endif
5898 6025
6026static irqreturn_t wlcore_hardirq(int irq, void *cookie)
6027{
6028 return IRQ_WAKE_THREAD;
6029}
6030
5899static void wlcore_nvs_cb(const struct firmware *fw, void *context) 6031static void wlcore_nvs_cb(const struct firmware *fw, void *context)
5900{ 6032{
5901 struct wl1271 *wl = context; 6033 struct wl1271 *wl = context;
5902 struct platform_device *pdev = wl->pdev; 6034 struct platform_device *pdev = wl->pdev;
5903 struct wlcore_platdev_data *pdev_data = pdev->dev.platform_data; 6035 struct wlcore_platdev_data *pdev_data = dev_get_platdata(&pdev->dev);
5904 struct wl12xx_platform_data *pdata = pdev_data->pdata; 6036 struct wl12xx_platform_data *pdata = pdev_data->pdata;
5905 unsigned long irqflags; 6037 unsigned long irqflags;
5906 int ret; 6038 int ret;
6039 irq_handler_t hardirq_fn = NULL;
5907 6040
5908 if (fw) { 6041 if (fw) {
5909 wl->nvs = kmemdup(fw->data, fw->size, GFP_KERNEL); 6042 wl->nvs = kmemdup(fw->data, fw->size, GFP_KERNEL);
@@ -5932,12 +6065,14 @@ static void wlcore_nvs_cb(const struct firmware *fw, void *context)
5932 wl->platform_quirks = pdata->platform_quirks; 6065 wl->platform_quirks = pdata->platform_quirks;
5933 wl->if_ops = pdev_data->if_ops; 6066 wl->if_ops = pdev_data->if_ops;
5934 6067
5935 if (wl->platform_quirks & WL12XX_PLATFORM_QUIRK_EDGE_IRQ) 6068 if (wl->platform_quirks & WL12XX_PLATFORM_QUIRK_EDGE_IRQ) {
5936 irqflags = IRQF_TRIGGER_RISING; 6069 irqflags = IRQF_TRIGGER_RISING;
5937 else 6070 hardirq_fn = wlcore_hardirq;
6071 } else {
5938 irqflags = IRQF_TRIGGER_HIGH | IRQF_ONESHOT; 6072 irqflags = IRQF_TRIGGER_HIGH | IRQF_ONESHOT;
6073 }
5939 6074
5940 ret = request_threaded_irq(wl->irq, NULL, wlcore_irq, 6075 ret = request_threaded_irq(wl->irq, hardirq_fn, wlcore_irq,
5941 irqflags, pdev->name, wl); 6076 irqflags, pdev->name, wl);
5942 if (ret < 0) { 6077 if (ret < 0) {
5943 wl1271_error("request_irq() failed: %d", ret); 6078 wl1271_error("request_irq() failed: %d", ret);
@@ -6046,6 +6181,9 @@ module_param_named(fwlog, fwlog_param, charp, 0);
6046MODULE_PARM_DESC(fwlog, 6181MODULE_PARM_DESC(fwlog,
6047 "FW logger options: continuous, ondemand, dbgpins or disable"); 6182 "FW logger options: continuous, ondemand, dbgpins or disable");
6048 6183
6184module_param(fwlog_mem_blocks, int, S_IRUSR | S_IWUSR);
6185MODULE_PARM_DESC(fwlog_mem_blocks, "fwlog mem_blocks");
6186
6049module_param(bug_on_recovery, int, S_IRUSR | S_IWUSR); 6187module_param(bug_on_recovery, int, S_IRUSR | S_IWUSR);
6050MODULE_PARM_DESC(bug_on_recovery, "BUG() on fw recovery"); 6188MODULE_PARM_DESC(bug_on_recovery, "BUG() on fw recovery");
6051 6189
diff --git a/drivers/net/wireless/ti/wlcore/ps.c b/drivers/net/wireless/ti/wlcore/ps.c
index 98066d40c2ad..26bfc365ba70 100644
--- a/drivers/net/wireless/ti/wlcore/ps.c
+++ b/drivers/net/wireless/ti/wlcore/ps.c
@@ -83,6 +83,10 @@ void wl1271_ps_elp_sleep(struct wl1271 *wl)
83 struct wl12xx_vif *wlvif; 83 struct wl12xx_vif *wlvif;
84 u32 timeout; 84 u32 timeout;
85 85
86 /* We do not enter elp sleep in PLT mode */
87 if (wl->plt)
88 return;
89
86 if (wl->sleep_auth != WL1271_PSM_ELP) 90 if (wl->sleep_auth != WL1271_PSM_ELP)
87 return; 91 return;
88 92
diff --git a/drivers/net/wireless/ti/wlcore/scan.c b/drivers/net/wireless/ti/wlcore/scan.c
index f407101e525b..7ed86203304b 100644
--- a/drivers/net/wireless/ti/wlcore/scan.c
+++ b/drivers/net/wireless/ti/wlcore/scan.c
@@ -92,9 +92,31 @@ out:
92static void wlcore_started_vifs_iter(void *data, u8 *mac, 92static void wlcore_started_vifs_iter(void *data, u8 *mac,
93 struct ieee80211_vif *vif) 93 struct ieee80211_vif *vif)
94{ 94{
95 struct wl12xx_vif *wlvif = wl12xx_vif_to_data(vif);
96 bool active = false;
95 int *count = (int *)data; 97 int *count = (int *)data;
96 98
97 if (!vif->bss_conf.idle) 99 /*
100 * count active interfaces according to interface type.
101 * checking only bss_conf.idle is bad for some cases, e.g.
102 * we don't want to count sta in p2p_find as active interface.
103 */
104 switch (wlvif->bss_type) {
105 case BSS_TYPE_STA_BSS:
106 if (test_bit(WLVIF_FLAG_STA_ASSOCIATED, &wlvif->flags))
107 active = true;
108 break;
109
110 case BSS_TYPE_AP_BSS:
111 if (wlvif->wl->active_sta_count > 0)
112 active = true;
113 break;
114
115 default:
116 break;
117 }
118
119 if (active)
98 (*count)++; 120 (*count)++;
99} 121}
100 122
@@ -174,17 +196,7 @@ wlcore_scan_get_channels(struct wl1271 *wl,
174 /* if radar is set, we ignore the passive flag */ 196 /* if radar is set, we ignore the passive flag */
175 (radar || 197 (radar ||
176 !!(flags & IEEE80211_CHAN_PASSIVE_SCAN) == passive)) { 198 !!(flags & IEEE80211_CHAN_PASSIVE_SCAN) == passive)) {
177 wl1271_debug(DEBUG_SCAN, "band %d, center_freq %d ", 199
178 req_channels[i]->band,
179 req_channels[i]->center_freq);
180 wl1271_debug(DEBUG_SCAN, "hw_value %d, flags %X",
181 req_channels[i]->hw_value,
182 req_channels[i]->flags);
183 wl1271_debug(DEBUG_SCAN, "max_power %d",
184 req_channels[i]->max_power);
185 wl1271_debug(DEBUG_SCAN, "min_dwell_time %d max dwell time %d",
186 min_dwell_time_active,
187 max_dwell_time_active);
188 200
189 if (flags & IEEE80211_CHAN_RADAR) { 201 if (flags & IEEE80211_CHAN_RADAR) {
190 channels[j].flags |= SCAN_CHANNEL_FLAGS_DFS; 202 channels[j].flags |= SCAN_CHANNEL_FLAGS_DFS;
@@ -222,6 +234,17 @@ wlcore_scan_get_channels(struct wl1271 *wl,
222 *n_pactive_ch); 234 *n_pactive_ch);
223 } 235 }
224 236
237 wl1271_debug(DEBUG_SCAN, "freq %d, ch. %d, flags 0x%x, power %d, min/max_dwell %d/%d%s%s",
238 req_channels[i]->center_freq,
239 req_channels[i]->hw_value,
240 req_channels[i]->flags,
241 req_channels[i]->max_power,
242 min_dwell_time_active,
243 max_dwell_time_active,
244 flags & IEEE80211_CHAN_RADAR ?
245 ", DFS" : "",
246 flags & IEEE80211_CHAN_PASSIVE_SCAN ?
247 ", PASSIVE" : "");
225 j++; 248 j++;
226 } 249 }
227 } 250 }
@@ -364,7 +387,7 @@ wlcore_scan_sched_scan_ssid_list(struct wl1271 *wl,
364 struct cfg80211_ssid *ssids = req->ssids; 387 struct cfg80211_ssid *ssids = req->ssids;
365 int ret = 0, type, i, j, n_match_ssids = 0; 388 int ret = 0, type, i, j, n_match_ssids = 0;
366 389
367 wl1271_debug(DEBUG_CMD, "cmd sched scan ssid list"); 390 wl1271_debug((DEBUG_CMD | DEBUG_SCAN), "cmd sched scan ssid list");
368 391
369 /* count the match sets that contain SSIDs */ 392 /* count the match sets that contain SSIDs */
370 for (i = 0; i < req->n_match_sets; i++) 393 for (i = 0; i < req->n_match_sets; i++)
@@ -442,8 +465,6 @@ wlcore_scan_sched_scan_ssid_list(struct wl1271 *wl,
442 } 465 }
443 } 466 }
444 467
445 wl1271_dump(DEBUG_SCAN, "SSID_LIST: ", cmd, sizeof(*cmd));
446
447 ret = wl1271_cmd_send(wl, CMD_CONNECTION_SCAN_SSID_CFG, cmd, 468 ret = wl1271_cmd_send(wl, CMD_CONNECTION_SCAN_SSID_CFG, cmd,
448 sizeof(*cmd), 0); 469 sizeof(*cmd), 0);
449 if (ret < 0) { 470 if (ret < 0) {
diff --git a/drivers/net/wireless/ti/wlcore/spi.c b/drivers/net/wireless/ti/wlcore/spi.c
index 1b0cd98e35f1..b2c018dccf18 100644
--- a/drivers/net/wireless/ti/wlcore/spi.c
+++ b/drivers/net/wireless/ti/wlcore/spi.c
@@ -335,7 +335,7 @@ static int wl1271_probe(struct spi_device *spi)
335 if (!pdev_data) 335 if (!pdev_data)
336 goto out; 336 goto out;
337 337
338 pdev_data->pdata = spi->dev.platform_data; 338 pdev_data->pdata = dev_get_platdata(&spi->dev);
339 if (!pdev_data->pdata) { 339 if (!pdev_data->pdata) {
340 dev_err(&spi->dev, "no platform data\n"); 340 dev_err(&spi->dev, "no platform data\n");
341 ret = -ENODEV; 341 ret = -ENODEV;
diff --git a/drivers/net/wireless/ti/wlcore/testmode.c b/drivers/net/wireless/ti/wlcore/testmode.c
index 527590f2adfb..ddad58f614da 100644
--- a/drivers/net/wireless/ti/wlcore/testmode.c
+++ b/drivers/net/wireless/ti/wlcore/testmode.c
@@ -179,7 +179,8 @@ static int wl1271_tm_cmd_interrogate(struct wl1271 *wl, struct nlattr *tb[])
179 goto out_sleep; 179 goto out_sleep;
180 } 180 }
181 181
182 ret = wl1271_cmd_interrogate(wl, ie_id, cmd, sizeof(*cmd)); 182 ret = wl1271_cmd_interrogate(wl, ie_id, cmd,
183 sizeof(struct acx_header), sizeof(*cmd));
183 if (ret < 0) { 184 if (ret < 0) {
184 wl1271_warning("testmode cmd interrogate failed: %d", ret); 185 wl1271_warning("testmode cmd interrogate failed: %d", ret);
185 goto out_free; 186 goto out_free;
@@ -297,7 +298,8 @@ static int wl1271_tm_cmd_set_plt_mode(struct wl1271 *wl, struct nlattr *tb[])
297 ret = wl1271_plt_stop(wl); 298 ret = wl1271_plt_stop(wl);
298 break; 299 break;
299 case PLT_ON: 300 case PLT_ON:
300 ret = wl1271_plt_start(wl, PLT_ON); 301 case PLT_CHIP_AWAKE:
302 ret = wl1271_plt_start(wl, val);
301 break; 303 break;
302 case PLT_FEM_DETECT: 304 case PLT_FEM_DETECT:
303 ret = wl1271_tm_detect_fem(wl, tb); 305 ret = wl1271_tm_detect_fem(wl, tb);
@@ -361,6 +363,7 @@ int wl1271_tm_cmd(struct ieee80211_hw *hw, struct ieee80211_vif *vif,
361{ 363{
362 struct wl1271 *wl = hw->priv; 364 struct wl1271 *wl = hw->priv;
363 struct nlattr *tb[WL1271_TM_ATTR_MAX + 1]; 365 struct nlattr *tb[WL1271_TM_ATTR_MAX + 1];
366 u32 nla_cmd;
364 int err; 367 int err;
365 368
366 err = nla_parse(tb, WL1271_TM_ATTR_MAX, data, len, wl1271_tm_policy); 369 err = nla_parse(tb, WL1271_TM_ATTR_MAX, data, len, wl1271_tm_policy);
@@ -370,7 +373,14 @@ int wl1271_tm_cmd(struct ieee80211_hw *hw, struct ieee80211_vif *vif,
370 if (!tb[WL1271_TM_ATTR_CMD_ID]) 373 if (!tb[WL1271_TM_ATTR_CMD_ID])
371 return -EINVAL; 374 return -EINVAL;
372 375
373 switch (nla_get_u32(tb[WL1271_TM_ATTR_CMD_ID])) { 376 nla_cmd = nla_get_u32(tb[WL1271_TM_ATTR_CMD_ID]);
377
378 /* Only SET_PLT_MODE is allowed in case of mode PLT_CHIP_AWAKE */
379 if (wl->plt_mode == PLT_CHIP_AWAKE &&
380 nla_cmd != WL1271_TM_CMD_SET_PLT_MODE)
381 return -EOPNOTSUPP;
382
383 switch (nla_cmd) {
374 case WL1271_TM_CMD_TEST: 384 case WL1271_TM_CMD_TEST:
375 return wl1271_tm_cmd_test(wl, tb); 385 return wl1271_tm_cmd_test(wl, tb);
376 case WL1271_TM_CMD_INTERROGATE: 386 case WL1271_TM_CMD_INTERROGATE:
diff --git a/drivers/net/wireless/ti/wlcore/tx.c b/drivers/net/wireless/ti/wlcore/tx.c
index 7e93fe63a2c7..87cd707affa2 100644
--- a/drivers/net/wireless/ti/wlcore/tx.c
+++ b/drivers/net/wireless/ti/wlcore/tx.c
@@ -86,19 +86,34 @@ void wl1271_free_tx_id(struct wl1271 *wl, int id)
86EXPORT_SYMBOL(wl1271_free_tx_id); 86EXPORT_SYMBOL(wl1271_free_tx_id);
87 87
88static void wl1271_tx_ap_update_inconnection_sta(struct wl1271 *wl, 88static void wl1271_tx_ap_update_inconnection_sta(struct wl1271 *wl,
89 struct wl12xx_vif *wlvif,
89 struct sk_buff *skb) 90 struct sk_buff *skb)
90{ 91{
91 struct ieee80211_hdr *hdr; 92 struct ieee80211_hdr *hdr;
92 93
94 hdr = (struct ieee80211_hdr *)(skb->data +
95 sizeof(struct wl1271_tx_hw_descr));
96 if (!ieee80211_is_auth(hdr->frame_control))
97 return;
98
93 /* 99 /*
94 * add the station to the known list before transmitting the 100 * add the station to the known list before transmitting the
95 * authentication response. this way it won't get de-authed by FW 101 * authentication response. this way it won't get de-authed by FW
96 * when transmitting too soon. 102 * when transmitting too soon.
97 */ 103 */
98 hdr = (struct ieee80211_hdr *)(skb->data + 104 wl1271_acx_set_inconnection_sta(wl, hdr->addr1);
99 sizeof(struct wl1271_tx_hw_descr)); 105
100 if (ieee80211_is_auth(hdr->frame_control)) 106 /*
101 wl1271_acx_set_inconnection_sta(wl, hdr->addr1); 107 * ROC for 1 second on the AP channel for completing the connection.
108 * Note the ROC will be continued by the update_sta_state callbacks
109 * once the station reaches the associated state.
110 */
111 wlcore_update_inconn_sta(wl, wlvif, NULL, true);
112 wlvif->pending_auth_reply_time = jiffies;
113 cancel_delayed_work(&wlvif->pending_auth_complete_work);
114 ieee80211_queue_delayed_work(wl->hw,
115 &wlvif->pending_auth_complete_work,
116 msecs_to_jiffies(WLCORE_PEND_AUTH_ROC_TIMEOUT));
102} 117}
103 118
104static void wl1271_tx_regulate_link(struct wl1271 *wl, 119static void wl1271_tx_regulate_link(struct wl1271 *wl,
@@ -386,7 +401,7 @@ static int wl1271_prepare_tx_frame(struct wl1271 *wl, struct wl12xx_vif *wlvif,
386 is_wep = (cipher == WLAN_CIPHER_SUITE_WEP40) || 401 is_wep = (cipher == WLAN_CIPHER_SUITE_WEP40) ||
387 (cipher == WLAN_CIPHER_SUITE_WEP104); 402 (cipher == WLAN_CIPHER_SUITE_WEP104);
388 403
389 if (WARN_ON(is_wep && wlvif->default_key != idx)) { 404 if (WARN_ON(is_wep && wlvif && wlvif->default_key != idx)) {
390 ret = wl1271_set_default_wep_key(wl, wlvif, idx); 405 ret = wl1271_set_default_wep_key(wl, wlvif, idx);
391 if (ret < 0) 406 if (ret < 0)
392 return ret; 407 return ret;
@@ -404,7 +419,7 @@ static int wl1271_prepare_tx_frame(struct wl1271 *wl, struct wl12xx_vif *wlvif,
404 wl1271_tx_fill_hdr(wl, wlvif, skb, extra, info, hlid); 419 wl1271_tx_fill_hdr(wl, wlvif, skb, extra, info, hlid);
405 420
406 if (!is_dummy && wlvif && wlvif->bss_type == BSS_TYPE_AP_BSS) { 421 if (!is_dummy && wlvif && wlvif->bss_type == BSS_TYPE_AP_BSS) {
407 wl1271_tx_ap_update_inconnection_sta(wl, skb); 422 wl1271_tx_ap_update_inconnection_sta(wl, wlvif, skb);
408 wl1271_tx_regulate_link(wl, wlvif, hlid); 423 wl1271_tx_regulate_link(wl, wlvif, hlid);
409 } 424 }
410 425
diff --git a/drivers/net/wireless/ti/wlcore/tx.h b/drivers/net/wireless/ti/wlcore/tx.h
index 55aa4acf9105..35489c300da1 100644
--- a/drivers/net/wireless/ti/wlcore/tx.h
+++ b/drivers/net/wireless/ti/wlcore/tx.h
@@ -56,6 +56,9 @@
56/* Used for management frames and dummy packets */ 56/* Used for management frames and dummy packets */
57#define WL1271_TID_MGMT 7 57#define WL1271_TID_MGMT 7
58 58
59/* stop a ROC for pending authentication reply after this time (ms) */
60#define WLCORE_PEND_AUTH_ROC_TIMEOUT 1000
61
59struct wl127x_tx_mem { 62struct wl127x_tx_mem {
60 /* 63 /*
61 * Number of extra memory blocks to allocate for this packet 64 * Number of extra memory blocks to allocate for this packet
diff --git a/drivers/net/wireless/ti/wlcore/wlcore.h b/drivers/net/wireless/ti/wlcore/wlcore.h
index 0034979e97cb..06efc12a39e5 100644
--- a/drivers/net/wireless/ti/wlcore/wlcore.h
+++ b/drivers/net/wireless/ti/wlcore/wlcore.h
@@ -110,6 +110,7 @@ struct wlcore_ops {
110 struct ieee80211_sta_ht_cap *ht_cap, 110 struct ieee80211_sta_ht_cap *ht_cap,
111 bool allow_ht_operation, 111 bool allow_ht_operation,
112 u32 rate_set, u8 hlid); 112 u32 rate_set, u8 hlid);
113 u32 (*convert_hwaddr)(struct wl1271 *wl, u32 hwaddr);
113 bool (*lnk_high_prio)(struct wl1271 *wl, u8 hlid, 114 bool (*lnk_high_prio)(struct wl1271 *wl, u8 hlid,
114 struct wl1271_link *lnk); 115 struct wl1271_link *lnk);
115 bool (*lnk_low_prio)(struct wl1271 *wl, u8 hlid, 116 bool (*lnk_low_prio)(struct wl1271 *wl, u8 hlid,
@@ -290,6 +291,12 @@ struct wl1271 {
290 /* Number of valid bytes in the FW log buffer */ 291 /* Number of valid bytes in the FW log buffer */
291 ssize_t fwlog_size; 292 ssize_t fwlog_size;
292 293
294 /* FW log end marker */
295 u32 fwlog_end;
296
297 /* FW memory block size */
298 u32 fw_mem_block_size;
299
293 /* Sysfs FW log entry readers wait queue */ 300 /* Sysfs FW log entry readers wait queue */
294 wait_queue_head_t fwlog_waitq; 301 wait_queue_head_t fwlog_waitq;
295 302
@@ -307,6 +314,8 @@ struct wl1271 {
307 314
308 /* The mbox event mask */ 315 /* The mbox event mask */
309 u32 event_mask; 316 u32 event_mask;
317 /* events to unmask only when ap interface is up */
318 u32 ap_event_mask;
310 319
311 /* Mailbox pointers */ 320 /* Mailbox pointers */
312 u32 mbox_size; 321 u32 mbox_size;
@@ -481,6 +490,8 @@ int wlcore_set_key(struct wl1271 *wl, enum set_key_cmd cmd,
481 struct ieee80211_sta *sta, 490 struct ieee80211_sta *sta,
482 struct ieee80211_key_conf *key_conf); 491 struct ieee80211_key_conf *key_conf);
483void wlcore_regdomain_config(struct wl1271 *wl); 492void wlcore_regdomain_config(struct wl1271 *wl);
493void wlcore_update_inconn_sta(struct wl1271 *wl, struct wl12xx_vif *wlvif,
494 struct wl1271_station *wl_sta, bool in_conn);
484 495
485static inline void 496static inline void
486wlcore_set_ht_cap(struct wl1271 *wl, enum ieee80211_band band, 497wlcore_set_ht_cap(struct wl1271 *wl, enum ieee80211_band band,
diff --git a/drivers/net/wireless/ti/wlcore/wlcore_i.h b/drivers/net/wireless/ti/wlcore/wlcore_i.h
index e5e146435fe7..ce7261ce8b59 100644
--- a/drivers/net/wireless/ti/wlcore/wlcore_i.h
+++ b/drivers/net/wireless/ti/wlcore/wlcore_i.h
@@ -255,6 +255,7 @@ enum wl12xx_vif_flags {
255 WLVIF_FLAG_CS_PROGRESS, 255 WLVIF_FLAG_CS_PROGRESS,
256 WLVIF_FLAG_AP_PROBE_RESP_SET, 256 WLVIF_FLAG_AP_PROBE_RESP_SET,
257 WLVIF_FLAG_IN_USE, 257 WLVIF_FLAG_IN_USE,
258 WLVIF_FLAG_ACTIVE,
258}; 259};
259 260
260struct wl12xx_vif; 261struct wl12xx_vif;
@@ -307,6 +308,7 @@ enum plt_mode {
307 PLT_OFF = 0, 308 PLT_OFF = 0,
308 PLT_ON = 1, 309 PLT_ON = 1,
309 PLT_FEM_DETECT = 2, 310 PLT_FEM_DETECT = 2,
311 PLT_CHIP_AWAKE = 3
310}; 312};
311 313
312struct wl12xx_rx_filter_field { 314struct wl12xx_rx_filter_field {
@@ -456,6 +458,15 @@ struct wl12xx_vif {
456 */ 458 */
457 int hw_queue_base; 459 int hw_queue_base;
458 460
461 /* do we have a pending auth reply? (and ROC) */
462 bool ap_pending_auth_reply;
463
464 /* time when we sent the pending auth reply */
465 unsigned long pending_auth_reply_time;
466
467 /* work for canceling ROC after pending auth reply */
468 struct delayed_work pending_auth_complete_work;
469
459 /* 470 /*
460 * This struct must be last! 471 * This struct must be last!
461 * data that has to be saved acrossed reconfigs (e.g. recovery) 472 * data that has to be saved acrossed reconfigs (e.g. recovery)
@@ -539,6 +550,4 @@ void wl1271_rx_filter_flatten_fields(struct wl12xx_rx_filter *filter,
539#define HW_HT_RATES_OFFSET 16 550#define HW_HT_RATES_OFFSET 16
540#define HW_MIMO_RATES_OFFSET 24 551#define HW_MIMO_RATES_OFFSET 24
541 552
542#define WL12XX_HW_BLOCK_SIZE 256
543
544#endif /* __WLCORE_I_H__ */ 553#endif /* __WLCORE_I_H__ */
diff --git a/drivers/net/xen-netback/common.h b/drivers/net/xen-netback/common.h
index 400fea1de080..08ae01b41c83 100644
--- a/drivers/net/xen-netback/common.h
+++ b/drivers/net/xen-netback/common.h
@@ -87,9 +87,13 @@ struct pending_tx_info {
87struct xenvif_rx_meta { 87struct xenvif_rx_meta {
88 int id; 88 int id;
89 int size; 89 int size;
90 int gso_type;
90 int gso_size; 91 int gso_size;
91}; 92};
92 93
94#define GSO_BIT(type) \
95 (1 << XEN_NETIF_GSO_TYPE_ ## type)
96
93/* Discriminate from any valid pending_idx value. */ 97/* Discriminate from any valid pending_idx value. */
94#define INVALID_PENDING_IDX 0xFFFF 98#define INVALID_PENDING_IDX 0xFFFF
95 99
@@ -150,10 +154,12 @@ struct xenvif {
150 u8 fe_dev_addr[6]; 154 u8 fe_dev_addr[6];
151 155
152 /* Frontend feature information. */ 156 /* Frontend feature information. */
157 int gso_mask;
158 int gso_prefix_mask;
159
153 u8 can_sg:1; 160 u8 can_sg:1;
154 u8 gso:1; 161 u8 ip_csum:1;
155 u8 gso_prefix:1; 162 u8 ipv6_csum:1;
156 u8 csum:1;
157 163
158 /* Internal feature information. */ 164 /* Internal feature information. */
159 u8 can_queue:1; /* can queue packets for receiver? */ 165 u8 can_queue:1; /* can queue packets for receiver? */
diff --git a/drivers/net/xen-netback/interface.c b/drivers/net/xen-netback/interface.c
index 459935a6bfae..b78ee10a956a 100644
--- a/drivers/net/xen-netback/interface.c
+++ b/drivers/net/xen-netback/interface.c
@@ -214,10 +214,14 @@ static netdev_features_t xenvif_fix_features(struct net_device *dev,
214 214
215 if (!vif->can_sg) 215 if (!vif->can_sg)
216 features &= ~NETIF_F_SG; 216 features &= ~NETIF_F_SG;
217 if (!vif->gso && !vif->gso_prefix) 217 if (~(vif->gso_mask | vif->gso_prefix_mask) & GSO_BIT(TCPV4))
218 features &= ~NETIF_F_TSO; 218 features &= ~NETIF_F_TSO;
219 if (!vif->csum) 219 if (~(vif->gso_mask | vif->gso_prefix_mask) & GSO_BIT(TCPV6))
220 features &= ~NETIF_F_TSO6;
221 if (!vif->ip_csum)
220 features &= ~NETIF_F_IP_CSUM; 222 features &= ~NETIF_F_IP_CSUM;
223 if (!vif->ipv6_csum)
224 features &= ~NETIF_F_IPV6_CSUM;
221 225
222 return features; 226 return features;
223} 227}
@@ -306,7 +310,7 @@ struct xenvif *xenvif_alloc(struct device *parent, domid_t domid,
306 vif->domid = domid; 310 vif->domid = domid;
307 vif->handle = handle; 311 vif->handle = handle;
308 vif->can_sg = 1; 312 vif->can_sg = 1;
309 vif->csum = 1; 313 vif->ip_csum = 1;
310 vif->dev = dev; 314 vif->dev = dev;
311 315
312 vif->credit_bytes = vif->remaining_credit = ~0UL; 316 vif->credit_bytes = vif->remaining_credit = ~0UL;
@@ -315,8 +319,10 @@ struct xenvif *xenvif_alloc(struct device *parent, domid_t domid,
315 vif->credit_window_start = get_jiffies_64(); 319 vif->credit_window_start = get_jiffies_64();
316 320
317 dev->netdev_ops = &xenvif_netdev_ops; 321 dev->netdev_ops = &xenvif_netdev_ops;
318 dev->hw_features = NETIF_F_SG | NETIF_F_IP_CSUM | NETIF_F_TSO; 322 dev->hw_features = NETIF_F_SG |
319 dev->features = dev->hw_features; 323 NETIF_F_IP_CSUM | NETIF_F_IPV6_CSUM |
324 NETIF_F_TSO | NETIF_F_TSO6;
325 dev->features = dev->hw_features | NETIF_F_RXCSUM;
320 SET_ETHTOOL_OPS(dev, &xenvif_ethtool_ops); 326 SET_ETHTOOL_OPS(dev, &xenvif_ethtool_ops);
321 327
322 dev->tx_queue_len = XENVIF_QUEUE_LENGTH; 328 dev->tx_queue_len = XENVIF_QUEUE_LENGTH;
diff --git a/drivers/net/xen-netback/netback.c b/drivers/net/xen-netback/netback.c
index 900da4b243ad..919b6509455c 100644
--- a/drivers/net/xen-netback/netback.c
+++ b/drivers/net/xen-netback/netback.c
@@ -109,15 +109,12 @@ static inline unsigned long idx_to_kaddr(struct xenvif *vif,
109 return (unsigned long)pfn_to_kaddr(idx_to_pfn(vif, idx)); 109 return (unsigned long)pfn_to_kaddr(idx_to_pfn(vif, idx));
110} 110}
111 111
112/* 112/* This is a miniumum size for the linear area to avoid lots of
113 * This is the amount of packet we copy rather than map, so that the 113 * calls to __pskb_pull_tail() as we set up checksum offsets. The
114 * guest can't fiddle with the contents of the headers while we do 114 * value 128 was chosen as it covers all IPv4 and most likely
115 * packet processing on them (netfilter, routing, etc). 115 * IPv6 headers.
116 */ 116 */
117#define PKT_PROT_LEN (ETH_HLEN + \ 117#define PKT_PROT_LEN 128
118 VLAN_HLEN + \
119 sizeof(struct iphdr) + MAX_IPOPTLEN + \
120 sizeof(struct tcphdr) + MAX_TCP_OPTION_SPACE)
121 118
122static u16 frag_get_pending_idx(skb_frag_t *frag) 119static u16 frag_get_pending_idx(skb_frag_t *frag)
123{ 120{
@@ -145,7 +142,7 @@ static int max_required_rx_slots(struct xenvif *vif)
145 int max = DIV_ROUND_UP(vif->dev->mtu, PAGE_SIZE); 142 int max = DIV_ROUND_UP(vif->dev->mtu, PAGE_SIZE);
146 143
147 /* XXX FIXME: RX path dependent on MAX_SKB_FRAGS */ 144 /* XXX FIXME: RX path dependent on MAX_SKB_FRAGS */
148 if (vif->can_sg || vif->gso || vif->gso_prefix) 145 if (vif->can_sg || vif->gso_mask || vif->gso_prefix_mask)
149 max += MAX_SKB_FRAGS + 1; /* extra_info + frags */ 146 max += MAX_SKB_FRAGS + 1; /* extra_info + frags */
150 147
151 return max; 148 return max;
@@ -317,6 +314,7 @@ static struct xenvif_rx_meta *get_next_rx_buffer(struct xenvif *vif,
317 req = RING_GET_REQUEST(&vif->rx, vif->rx.req_cons++); 314 req = RING_GET_REQUEST(&vif->rx, vif->rx.req_cons++);
318 315
319 meta = npo->meta + npo->meta_prod++; 316 meta = npo->meta + npo->meta_prod++;
317 meta->gso_type = XEN_NETIF_GSO_TYPE_NONE;
320 meta->gso_size = 0; 318 meta->gso_size = 0;
321 meta->size = 0; 319 meta->size = 0;
322 meta->id = req->id; 320 meta->id = req->id;
@@ -339,6 +337,7 @@ static void xenvif_gop_frag_copy(struct xenvif *vif, struct sk_buff *skb,
339 struct gnttab_copy *copy_gop; 337 struct gnttab_copy *copy_gop;
340 struct xenvif_rx_meta *meta; 338 struct xenvif_rx_meta *meta;
341 unsigned long bytes; 339 unsigned long bytes;
340 int gso_type;
342 341
343 /* Data must not cross a page boundary. */ 342 /* Data must not cross a page boundary. */
344 BUG_ON(size + offset > PAGE_SIZE<<compound_order(page)); 343 BUG_ON(size + offset > PAGE_SIZE<<compound_order(page));
@@ -397,7 +396,14 @@ static void xenvif_gop_frag_copy(struct xenvif *vif, struct sk_buff *skb,
397 } 396 }
398 397
399 /* Leave a gap for the GSO descriptor. */ 398 /* Leave a gap for the GSO descriptor. */
400 if (*head && skb_shinfo(skb)->gso_size && !vif->gso_prefix) 399 if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV4)
400 gso_type = XEN_NETIF_GSO_TYPE_TCPV4;
401 else if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6)
402 gso_type = XEN_NETIF_GSO_TYPE_TCPV6;
403 else
404 gso_type = XEN_NETIF_GSO_TYPE_NONE;
405
406 if (*head && ((1 << gso_type) & vif->gso_mask))
401 vif->rx.req_cons++; 407 vif->rx.req_cons++;
402 408
403 *head = 0; /* There must be something in this buffer now. */ 409 *head = 0; /* There must be something in this buffer now. */
@@ -428,14 +434,28 @@ static int xenvif_gop_skb(struct sk_buff *skb,
428 unsigned char *data; 434 unsigned char *data;
429 int head = 1; 435 int head = 1;
430 int old_meta_prod; 436 int old_meta_prod;
437 int gso_type;
438 int gso_size;
431 439
432 old_meta_prod = npo->meta_prod; 440 old_meta_prod = npo->meta_prod;
433 441
442 if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV4) {
443 gso_type = XEN_NETIF_GSO_TYPE_TCPV4;
444 gso_size = skb_shinfo(skb)->gso_size;
445 } else if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6) {
446 gso_type = XEN_NETIF_GSO_TYPE_TCPV6;
447 gso_size = skb_shinfo(skb)->gso_size;
448 } else {
449 gso_type = XEN_NETIF_GSO_TYPE_NONE;
450 gso_size = 0;
451 }
452
434 /* Set up a GSO prefix descriptor, if necessary */ 453 /* Set up a GSO prefix descriptor, if necessary */
435 if (skb_shinfo(skb)->gso_size && vif->gso_prefix) { 454 if ((1 << skb_shinfo(skb)->gso_type) & vif->gso_prefix_mask) {
436 req = RING_GET_REQUEST(&vif->rx, vif->rx.req_cons++); 455 req = RING_GET_REQUEST(&vif->rx, vif->rx.req_cons++);
437 meta = npo->meta + npo->meta_prod++; 456 meta = npo->meta + npo->meta_prod++;
438 meta->gso_size = skb_shinfo(skb)->gso_size; 457 meta->gso_type = gso_type;
458 meta->gso_size = gso_size;
439 meta->size = 0; 459 meta->size = 0;
440 meta->id = req->id; 460 meta->id = req->id;
441 } 461 }
@@ -443,10 +463,13 @@ static int xenvif_gop_skb(struct sk_buff *skb,
443 req = RING_GET_REQUEST(&vif->rx, vif->rx.req_cons++); 463 req = RING_GET_REQUEST(&vif->rx, vif->rx.req_cons++);
444 meta = npo->meta + npo->meta_prod++; 464 meta = npo->meta + npo->meta_prod++;
445 465
446 if (!vif->gso_prefix) 466 if ((1 << gso_type) & vif->gso_mask) {
447 meta->gso_size = skb_shinfo(skb)->gso_size; 467 meta->gso_type = gso_type;
448 else 468 meta->gso_size = gso_size;
469 } else {
470 meta->gso_type = XEN_NETIF_GSO_TYPE_NONE;
449 meta->gso_size = 0; 471 meta->gso_size = 0;
472 }
450 473
451 meta->size = 0; 474 meta->size = 0;
452 meta->id = req->id; 475 meta->id = req->id;
@@ -592,7 +615,8 @@ void xenvif_rx_action(struct xenvif *vif)
592 615
593 vif = netdev_priv(skb->dev); 616 vif = netdev_priv(skb->dev);
594 617
595 if (vif->meta[npo.meta_cons].gso_size && vif->gso_prefix) { 618 if ((1 << vif->meta[npo.meta_cons].gso_type) &
619 vif->gso_prefix_mask) {
596 resp = RING_GET_RESPONSE(&vif->rx, 620 resp = RING_GET_RESPONSE(&vif->rx,
597 vif->rx.rsp_prod_pvt++); 621 vif->rx.rsp_prod_pvt++);
598 622
@@ -629,7 +653,8 @@ void xenvif_rx_action(struct xenvif *vif)
629 vif->meta[npo.meta_cons].size, 653 vif->meta[npo.meta_cons].size,
630 flags); 654 flags);
631 655
632 if (vif->meta[npo.meta_cons].gso_size && !vif->gso_prefix) { 656 if ((1 << vif->meta[npo.meta_cons].gso_type) &
657 vif->gso_mask) {
633 struct xen_netif_extra_info *gso = 658 struct xen_netif_extra_info *gso =
634 (struct xen_netif_extra_info *) 659 (struct xen_netif_extra_info *)
635 RING_GET_RESPONSE(&vif->rx, 660 RING_GET_RESPONSE(&vif->rx,
@@ -637,8 +662,8 @@ void xenvif_rx_action(struct xenvif *vif)
637 662
638 resp->flags |= XEN_NETRXF_extra_info; 663 resp->flags |= XEN_NETRXF_extra_info;
639 664
665 gso->u.gso.type = vif->meta[npo.meta_cons].gso_type;
640 gso->u.gso.size = vif->meta[npo.meta_cons].gso_size; 666 gso->u.gso.size = vif->meta[npo.meta_cons].gso_size;
641 gso->u.gso.type = XEN_NETIF_GSO_TYPE_TCPV4;
642 gso->u.gso.pad = 0; 667 gso->u.gso.pad = 0;
643 gso->u.gso.features = 0; 668 gso->u.gso.features = 0;
644 669
@@ -1101,15 +1126,20 @@ static int xenvif_set_skb_gso(struct xenvif *vif,
1101 return -EINVAL; 1126 return -EINVAL;
1102 } 1127 }
1103 1128
1104 /* Currently only TCPv4 S.O. is supported. */ 1129 switch (gso->u.gso.type) {
1105 if (gso->u.gso.type != XEN_NETIF_GSO_TYPE_TCPV4) { 1130 case XEN_NETIF_GSO_TYPE_TCPV4:
1131 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV4;
1132 break;
1133 case XEN_NETIF_GSO_TYPE_TCPV6:
1134 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV6;
1135 break;
1136 default:
1106 netdev_err(vif->dev, "Bad GSO type %d.\n", gso->u.gso.type); 1137 netdev_err(vif->dev, "Bad GSO type %d.\n", gso->u.gso.type);
1107 xenvif_fatal_tx_err(vif); 1138 xenvif_fatal_tx_err(vif);
1108 return -EINVAL; 1139 return -EINVAL;
1109 } 1140 }
1110 1141
1111 skb_shinfo(skb)->gso_size = gso->u.gso.size; 1142 skb_shinfo(skb)->gso_size = gso->u.gso.size;
1112 skb_shinfo(skb)->gso_type = SKB_GSO_TCPV4;
1113 1143
1114 /* Header must be checked, and gso_segs computed. */ 1144 /* Header must be checked, and gso_segs computed. */
1115 skb_shinfo(skb)->gso_type |= SKB_GSO_DODGY; 1145 skb_shinfo(skb)->gso_type |= SKB_GSO_DODGY;
@@ -1118,61 +1148,74 @@ static int xenvif_set_skb_gso(struct xenvif *vif,
1118 return 0; 1148 return 0;
1119} 1149}
1120 1150
1121static int checksum_setup(struct xenvif *vif, struct sk_buff *skb) 1151static inline void maybe_pull_tail(struct sk_buff *skb, unsigned int len)
1152{
1153 if (skb_is_nonlinear(skb) && skb_headlen(skb) < len) {
1154 /* If we need to pullup then pullup to the max, so we
1155 * won't need to do it again.
1156 */
1157 int target = min_t(int, skb->len, MAX_TCP_HEADER);
1158 __pskb_pull_tail(skb, target - skb_headlen(skb));
1159 }
1160}
1161
1162static int checksum_setup_ip(struct xenvif *vif, struct sk_buff *skb,
1163 int recalculate_partial_csum)
1122{ 1164{
1123 struct iphdr *iph; 1165 struct iphdr *iph = (void *)skb->data;
1166 unsigned int header_size;
1167 unsigned int off;
1124 int err = -EPROTO; 1168 int err = -EPROTO;
1125 int recalculate_partial_csum = 0;
1126 1169
1127 /* 1170 off = sizeof(struct iphdr);
1128 * A GSO SKB must be CHECKSUM_PARTIAL. However some buggy
1129 * peers can fail to set NETRXF_csum_blank when sending a GSO
1130 * frame. In this case force the SKB to CHECKSUM_PARTIAL and
1131 * recalculate the partial checksum.
1132 */
1133 if (skb->ip_summed != CHECKSUM_PARTIAL && skb_is_gso(skb)) {
1134 vif->rx_gso_checksum_fixup++;
1135 skb->ip_summed = CHECKSUM_PARTIAL;
1136 recalculate_partial_csum = 1;
1137 }
1138 1171
1139 /* A non-CHECKSUM_PARTIAL SKB does not require setup. */ 1172 header_size = skb->network_header + off + MAX_IPOPTLEN;
1140 if (skb->ip_summed != CHECKSUM_PARTIAL) 1173 maybe_pull_tail(skb, header_size);
1141 return 0;
1142 1174
1143 if (skb->protocol != htons(ETH_P_IP)) 1175 off = iph->ihl * 4;
1144 goto out;
1145 1176
1146 iph = (void *)skb->data;
1147 switch (iph->protocol) { 1177 switch (iph->protocol) {
1148 case IPPROTO_TCP: 1178 case IPPROTO_TCP:
1149 if (!skb_partial_csum_set(skb, 4 * iph->ihl, 1179 if (!skb_partial_csum_set(skb, off,
1150 offsetof(struct tcphdr, check))) 1180 offsetof(struct tcphdr, check)))
1151 goto out; 1181 goto out;
1152 1182
1153 if (recalculate_partial_csum) { 1183 if (recalculate_partial_csum) {
1154 struct tcphdr *tcph = tcp_hdr(skb); 1184 struct tcphdr *tcph = tcp_hdr(skb);
1185
1186 header_size = skb->network_header +
1187 off +
1188 sizeof(struct tcphdr);
1189 maybe_pull_tail(skb, header_size);
1190
1155 tcph->check = ~csum_tcpudp_magic(iph->saddr, iph->daddr, 1191 tcph->check = ~csum_tcpudp_magic(iph->saddr, iph->daddr,
1156 skb->len - iph->ihl*4, 1192 skb->len - off,
1157 IPPROTO_TCP, 0); 1193 IPPROTO_TCP, 0);
1158 } 1194 }
1159 break; 1195 break;
1160 case IPPROTO_UDP: 1196 case IPPROTO_UDP:
1161 if (!skb_partial_csum_set(skb, 4 * iph->ihl, 1197 if (!skb_partial_csum_set(skb, off,
1162 offsetof(struct udphdr, check))) 1198 offsetof(struct udphdr, check)))
1163 goto out; 1199 goto out;
1164 1200
1165 if (recalculate_partial_csum) { 1201 if (recalculate_partial_csum) {
1166 struct udphdr *udph = udp_hdr(skb); 1202 struct udphdr *udph = udp_hdr(skb);
1203
1204 header_size = skb->network_header +
1205 off +
1206 sizeof(struct udphdr);
1207 maybe_pull_tail(skb, header_size);
1208
1167 udph->check = ~csum_tcpudp_magic(iph->saddr, iph->daddr, 1209 udph->check = ~csum_tcpudp_magic(iph->saddr, iph->daddr,
1168 skb->len - iph->ihl*4, 1210 skb->len - off,
1169 IPPROTO_UDP, 0); 1211 IPPROTO_UDP, 0);
1170 } 1212 }
1171 break; 1213 break;
1172 default: 1214 default:
1173 if (net_ratelimit()) 1215 if (net_ratelimit())
1174 netdev_err(vif->dev, 1216 netdev_err(vif->dev,
1175 "Attempting to checksum a non-TCP/UDP packet, dropping a protocol %d packet\n", 1217 "Attempting to checksum a non-TCP/UDP packet, "
1218 "dropping a protocol %d packet\n",
1176 iph->protocol); 1219 iph->protocol);
1177 goto out; 1220 goto out;
1178 } 1221 }
@@ -1183,6 +1226,158 @@ out:
1183 return err; 1226 return err;
1184} 1227}
1185 1228
1229static int checksum_setup_ipv6(struct xenvif *vif, struct sk_buff *skb,
1230 int recalculate_partial_csum)
1231{
1232 int err = -EPROTO;
1233 struct ipv6hdr *ipv6h = (void *)skb->data;
1234 u8 nexthdr;
1235 unsigned int header_size;
1236 unsigned int off;
1237 bool fragment;
1238 bool done;
1239
1240 done = false;
1241
1242 off = sizeof(struct ipv6hdr);
1243
1244 header_size = skb->network_header + off;
1245 maybe_pull_tail(skb, header_size);
1246
1247 nexthdr = ipv6h->nexthdr;
1248
1249 while ((off <= sizeof(struct ipv6hdr) + ntohs(ipv6h->payload_len)) &&
1250 !done) {
1251 switch (nexthdr) {
1252 case IPPROTO_DSTOPTS:
1253 case IPPROTO_HOPOPTS:
1254 case IPPROTO_ROUTING: {
1255 struct ipv6_opt_hdr *hp = (void *)(skb->data + off);
1256
1257 header_size = skb->network_header +
1258 off +
1259 sizeof(struct ipv6_opt_hdr);
1260 maybe_pull_tail(skb, header_size);
1261
1262 nexthdr = hp->nexthdr;
1263 off += ipv6_optlen(hp);
1264 break;
1265 }
1266 case IPPROTO_AH: {
1267 struct ip_auth_hdr *hp = (void *)(skb->data + off);
1268
1269 header_size = skb->network_header +
1270 off +
1271 sizeof(struct ip_auth_hdr);
1272 maybe_pull_tail(skb, header_size);
1273
1274 nexthdr = hp->nexthdr;
1275 off += (hp->hdrlen+2)<<2;
1276 break;
1277 }
1278 case IPPROTO_FRAGMENT:
1279 fragment = true;
1280 /* fall through */
1281 default:
1282 done = true;
1283 break;
1284 }
1285 }
1286
1287 if (!done) {
1288 if (net_ratelimit())
1289 netdev_err(vif->dev, "Failed to parse packet header\n");
1290 goto out;
1291 }
1292
1293 if (fragment) {
1294 if (net_ratelimit())
1295 netdev_err(vif->dev, "Packet is a fragment!\n");
1296 goto out;
1297 }
1298
1299 switch (nexthdr) {
1300 case IPPROTO_TCP:
1301 if (!skb_partial_csum_set(skb, off,
1302 offsetof(struct tcphdr, check)))
1303 goto out;
1304
1305 if (recalculate_partial_csum) {
1306 struct tcphdr *tcph = tcp_hdr(skb);
1307
1308 header_size = skb->network_header +
1309 off +
1310 sizeof(struct tcphdr);
1311 maybe_pull_tail(skb, header_size);
1312
1313 tcph->check = ~csum_ipv6_magic(&ipv6h->saddr,
1314 &ipv6h->daddr,
1315 skb->len - off,
1316 IPPROTO_TCP, 0);
1317 }
1318 break;
1319 case IPPROTO_UDP:
1320 if (!skb_partial_csum_set(skb, off,
1321 offsetof(struct udphdr, check)))
1322 goto out;
1323
1324 if (recalculate_partial_csum) {
1325 struct udphdr *udph = udp_hdr(skb);
1326
1327 header_size = skb->network_header +
1328 off +
1329 sizeof(struct udphdr);
1330 maybe_pull_tail(skb, header_size);
1331
1332 udph->check = ~csum_ipv6_magic(&ipv6h->saddr,
1333 &ipv6h->daddr,
1334 skb->len - off,
1335 IPPROTO_UDP, 0);
1336 }
1337 break;
1338 default:
1339 if (net_ratelimit())
1340 netdev_err(vif->dev,
1341 "Attempting to checksum a non-TCP/UDP packet, "
1342 "dropping a protocol %d packet\n",
1343 nexthdr);
1344 goto out;
1345 }
1346
1347 err = 0;
1348
1349out:
1350 return err;
1351}
1352
1353static int checksum_setup(struct xenvif *vif, struct sk_buff *skb)
1354{
1355 int err = -EPROTO;
1356 int recalculate_partial_csum = 0;
1357
1358 /* A GSO SKB must be CHECKSUM_PARTIAL. However some buggy
1359 * peers can fail to set NETRXF_csum_blank when sending a GSO
1360 * frame. In this case force the SKB to CHECKSUM_PARTIAL and
1361 * recalculate the partial checksum.
1362 */
1363 if (skb->ip_summed != CHECKSUM_PARTIAL && skb_is_gso(skb)) {
1364 vif->rx_gso_checksum_fixup++;
1365 skb->ip_summed = CHECKSUM_PARTIAL;
1366 recalculate_partial_csum = 1;
1367 }
1368
1369 /* A non-CHECKSUM_PARTIAL SKB does not require setup. */
1370 if (skb->ip_summed != CHECKSUM_PARTIAL)
1371 return 0;
1372
1373 if (skb->protocol == htons(ETH_P_IP))
1374 err = checksum_setup_ip(vif, skb, recalculate_partial_csum);
1375 else if (skb->protocol == htons(ETH_P_IPV6))
1376 err = checksum_setup_ipv6(vif, skb, recalculate_partial_csum);
1377
1378 return err;
1379}
1380
1186static bool tx_credit_exceeded(struct xenvif *vif, unsigned size) 1381static bool tx_credit_exceeded(struct xenvif *vif, unsigned size)
1187{ 1382{
1188 u64 now = get_jiffies_64(); 1383 u64 now = get_jiffies_64();
@@ -1428,12 +1623,7 @@ static int xenvif_tx_submit(struct xenvif *vif, int budget)
1428 1623
1429 xenvif_fill_frags(vif, skb); 1624 xenvif_fill_frags(vif, skb);
1430 1625
1431 /* 1626 if (skb_is_nonlinear(skb) && skb_headlen(skb) < PKT_PROT_LEN) {
1432 * If the initial fragment was < PKT_PROT_LEN then
1433 * pull through some bytes from the other fragments to
1434 * increase the linear region to PKT_PROT_LEN bytes.
1435 */
1436 if (skb_headlen(skb) < PKT_PROT_LEN && skb_is_nonlinear(skb)) {
1437 int target = min_t(int, skb->len, PKT_PROT_LEN); 1627 int target = min_t(int, skb->len, PKT_PROT_LEN);
1438 __pskb_pull_tail(skb, target - skb_headlen(skb)); 1628 __pskb_pull_tail(skb, target - skb_headlen(skb));
1439 } 1629 }
diff --git a/drivers/net/xen-netback/xenbus.c b/drivers/net/xen-netback/xenbus.c
index 1b08d8798372..f0358992b04f 100644
--- a/drivers/net/xen-netback/xenbus.c
+++ b/drivers/net/xen-netback/xenbus.c
@@ -105,6 +105,22 @@ static int netback_probe(struct xenbus_device *dev,
105 goto abort_transaction; 105 goto abort_transaction;
106 } 106 }
107 107
108 err = xenbus_printf(xbt, dev->nodename, "feature-gso-tcpv6",
109 "%d", sg);
110 if (err) {
111 message = "writing feature-gso-tcpv6";
112 goto abort_transaction;
113 }
114
115 /* We support partial checksum setup for IPv6 packets */
116 err = xenbus_printf(xbt, dev->nodename,
117 "feature-ipv6-csum-offload",
118 "%d", 1);
119 if (err) {
120 message = "writing feature-ipv6-csum-offload";
121 goto abort_transaction;
122 }
123
108 /* We support rx-copy path. */ 124 /* We support rx-copy path. */
109 err = xenbus_printf(xbt, dev->nodename, 125 err = xenbus_printf(xbt, dev->nodename,
110 "feature-rx-copy", "%d", 1); 126 "feature-rx-copy", "%d", 1);
@@ -561,20 +577,50 @@ static int connect_rings(struct backend_info *be)
561 val = 0; 577 val = 0;
562 vif->can_sg = !!val; 578 vif->can_sg = !!val;
563 579
580 vif->gso_mask = 0;
581 vif->gso_prefix_mask = 0;
582
564 if (xenbus_scanf(XBT_NIL, dev->otherend, "feature-gso-tcpv4", 583 if (xenbus_scanf(XBT_NIL, dev->otherend, "feature-gso-tcpv4",
565 "%d", &val) < 0) 584 "%d", &val) < 0)
566 val = 0; 585 val = 0;
567 vif->gso = !!val; 586 if (val)
587 vif->gso_mask |= GSO_BIT(TCPV4);
568 588
569 if (xenbus_scanf(XBT_NIL, dev->otherend, "feature-gso-tcpv4-prefix", 589 if (xenbus_scanf(XBT_NIL, dev->otherend, "feature-gso-tcpv4-prefix",
570 "%d", &val) < 0) 590 "%d", &val) < 0)
571 val = 0; 591 val = 0;
572 vif->gso_prefix = !!val; 592 if (val)
593 vif->gso_prefix_mask |= GSO_BIT(TCPV4);
594
595 if (xenbus_scanf(XBT_NIL, dev->otherend, "feature-gso-tcpv6",
596 "%d", &val) < 0)
597 val = 0;
598 if (val)
599 vif->gso_mask |= GSO_BIT(TCPV6);
600
601 if (xenbus_scanf(XBT_NIL, dev->otherend, "feature-gso-tcpv6-prefix",
602 "%d", &val) < 0)
603 val = 0;
604 if (val)
605 vif->gso_prefix_mask |= GSO_BIT(TCPV6);
606
607 if (vif->gso_mask & vif->gso_prefix_mask) {
608 xenbus_dev_fatal(dev, err,
609 "%s: gso and gso prefix flags are not "
610 "mutually exclusive",
611 dev->otherend);
612 return -EOPNOTSUPP;
613 }
573 614
574 if (xenbus_scanf(XBT_NIL, dev->otherend, "feature-no-csum-offload", 615 if (xenbus_scanf(XBT_NIL, dev->otherend, "feature-no-csum-offload",
575 "%d", &val) < 0) 616 "%d", &val) < 0)
576 val = 0; 617 val = 0;
577 vif->csum = !val; 618 vif->ip_csum = !val;
619
620 if (xenbus_scanf(XBT_NIL, dev->otherend, "feature-ipv6-csum-offload",
621 "%d", &val) < 0)
622 val = 0;
623 vif->ipv6_csum = !!val;
578 624
579 /* Map the shared frame, irq etc. */ 625 /* Map the shared frame, irq etc. */
580 err = xenvif_connect(vif, tx_ring_ref, rx_ring_ref, 626 err = xenvif_connect(vif, tx_ring_ref, rx_ring_ref,
diff --git a/drivers/net/xen-netfront.c b/drivers/net/xen-netfront.c
index 36808bf25677..dd1011e55cb5 100644
--- a/drivers/net/xen-netfront.c
+++ b/drivers/net/xen-netfront.c
@@ -952,7 +952,7 @@ static int handle_incoming_queue(struct net_device *dev,
952 u64_stats_update_end(&stats->syncp); 952 u64_stats_update_end(&stats->syncp);
953 953
954 /* Pass it up. */ 954 /* Pass it up. */
955 netif_receive_skb(skb); 955 napi_gro_receive(&np->napi, skb);
956 } 956 }
957 957
958 return packets_dropped; 958 return packets_dropped;
@@ -1051,6 +1051,8 @@ err:
1051 if (work_done < budget) { 1051 if (work_done < budget) {
1052 int more_to_do = 0; 1052 int more_to_do = 0;
1053 1053
1054 napi_gro_flush(napi, false);
1055
1054 local_irq_save(flags); 1056 local_irq_save(flags);
1055 1057
1056 RING_FINAL_CHECK_FOR_RESPONSES(&np->rx, more_to_do); 1058 RING_FINAL_CHECK_FOR_RESPONSES(&np->rx, more_to_do);