diff options
author | Linus Torvalds <torvalds@linux-foundation.org> | 2015-04-17 16:31:08 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2015-04-17 16:31:08 -0400 |
commit | 388f997620cb57372c494a194e9698b28cc179b8 (patch) | |
tree | 31f2b7f01793f1711794193450f9047f78ee5370 /drivers | |
parent | e2fdae7e7c5a690b10b2d2891ec819e554dc033d (diff) | |
parent | e3122b7fae7b4e3d1d49fa84f6515bcbe6cbc6fc (diff) |
Merge git://git.kernel.org/pub/scm/linux/kernel/git/davem/net
Pull networking fixes from David Miller:
1) Fix verifier memory corruption and other bugs in BPF layer, from
Alexei Starovoitov.
2) Add a conservative fix for doing BPF properly in the BPF classifier
of the packet scheduler on ingress. Also from Alexei.
3) The SKB scrubber should not clear out the packet MARK and security
label, from Herbert Xu.
4) Fix oops on rmmod in stmmac driver, from Bryan O'Donoghue.
5) Pause handling is not correct in the stmmac driver because it
doesn't take into consideration the RX and TX fifo sizes. From
Vince Bridgers.
6) Failure path missing unlock in FOU driver, from Wang Cong.
* git://git.kernel.org/pub/scm/linux/kernel/git/davem/net: (44 commits)
net: dsa: use DEVICE_ATTR_RW to declare temp1_max
netns: remove BUG_ONs from net_generic()
IB/ipoib: Fix ndo_get_iflink
sfc: Fix memcpy() with const destination compiler warning.
altera tse: Fix network-delays and -retransmissions after high throughput.
net: remove unused 'dev' argument from netif_needs_gso()
act_mirred: Fix bogus header when redirecting from VLAN
inet_diag: fix access to tcp cc information
tcp: tcp_get_info() should fetch socket fields once
net: dsa: mv88e6xxx: Add missing initialization in mv88e6xxx_set_port_state()
skbuff: Do not scrub skb mark within the same name space
Revert "net: Reset secmark when scrubbing packet"
bpf: fix two bugs in verification logic when accessing 'ctx' pointer
bpf: fix bpf helpers to use skb->mac_header relative offsets
stmmac: Configure Flow Control to work correctly based on rxfifo size
stmmac: Enable unicast pause frame detect in GMAC Register 6
stmmac: Read tx-fifo-depth and rx-fifo-depth from the devicetree
stmmac: Add defines and documentation for enabling flow control
stmmac: Add properties for transmit and receive fifo sizes
stmmac: fix oops on rmmod after assigning ip addr
...
Diffstat (limited to 'drivers')
35 files changed, 508 insertions, 290 deletions
diff --git a/drivers/infiniband/hw/cxgb4/mem.c b/drivers/infiniband/hw/cxgb4/mem.c index 6791fd16272c..3ef0cf9f5c44 100644 --- a/drivers/infiniband/hw/cxgb4/mem.c +++ b/drivers/infiniband/hw/cxgb4/mem.c | |||
@@ -73,7 +73,7 @@ static int _c4iw_write_mem_dma_aligned(struct c4iw_rdev *rdev, u32 addr, | |||
73 | c4iw_init_wr_wait(&wr_wait); | 73 | c4iw_init_wr_wait(&wr_wait); |
74 | wr_len = roundup(sizeof(*req) + sizeof(*sgl), 16); | 74 | wr_len = roundup(sizeof(*req) + sizeof(*sgl), 16); |
75 | 75 | ||
76 | skb = alloc_skb(wr_len, GFP_KERNEL | __GFP_NOFAIL); | 76 | skb = alloc_skb(wr_len, GFP_KERNEL); |
77 | if (!skb) | 77 | if (!skb) |
78 | return -ENOMEM; | 78 | return -ENOMEM; |
79 | set_wr_txq(skb, CPL_PRIORITY_CONTROL, 0); | 79 | set_wr_txq(skb, CPL_PRIORITY_CONTROL, 0); |
diff --git a/drivers/infiniband/ulp/ipoib/ipoib_main.c b/drivers/infiniband/ulp/ipoib/ipoib_main.c index 657b89b1d291..915ad04a827e 100644 --- a/drivers/infiniband/ulp/ipoib/ipoib_main.c +++ b/drivers/infiniband/ulp/ipoib/ipoib_main.c | |||
@@ -846,6 +846,11 @@ static int ipoib_get_iflink(const struct net_device *dev) | |||
846 | { | 846 | { |
847 | struct ipoib_dev_priv *priv = netdev_priv(dev); | 847 | struct ipoib_dev_priv *priv = netdev_priv(dev); |
848 | 848 | ||
849 | /* parent interface */ | ||
850 | if (!test_bit(IPOIB_FLAG_SUBINTERFACE, &priv->flags)) | ||
851 | return dev->ifindex; | ||
852 | |||
853 | /* child/vlan interface */ | ||
849 | return priv->parent->ifindex; | 854 | return priv->parent->ifindex; |
850 | } | 855 | } |
851 | 856 | ||
diff --git a/drivers/infiniband/ulp/ipoib/ipoib_vlan.c b/drivers/infiniband/ulp/ipoib/ipoib_vlan.c index 4dd1313056a4..fca1a882de27 100644 --- a/drivers/infiniband/ulp/ipoib/ipoib_vlan.c +++ b/drivers/infiniband/ulp/ipoib/ipoib_vlan.c | |||
@@ -58,6 +58,7 @@ int __ipoib_vlan_add(struct ipoib_dev_priv *ppriv, struct ipoib_dev_priv *priv, | |||
58 | /* MTU will be reset when mcast join happens */ | 58 | /* MTU will be reset when mcast join happens */ |
59 | priv->dev->mtu = IPOIB_UD_MTU(priv->max_ib_mtu); | 59 | priv->dev->mtu = IPOIB_UD_MTU(priv->max_ib_mtu); |
60 | priv->mcast_mtu = priv->admin_mtu = priv->dev->mtu; | 60 | priv->mcast_mtu = priv->admin_mtu = priv->dev->mtu; |
61 | priv->parent = ppriv->dev; | ||
61 | set_bit(IPOIB_FLAG_SUBINTERFACE, &priv->flags); | 62 | set_bit(IPOIB_FLAG_SUBINTERFACE, &priv->flags); |
62 | 63 | ||
63 | result = ipoib_set_dev_features(priv, ppriv->ca); | 64 | result = ipoib_set_dev_features(priv, ppriv->ca); |
@@ -84,8 +85,6 @@ int __ipoib_vlan_add(struct ipoib_dev_priv *ppriv, struct ipoib_dev_priv *priv, | |||
84 | goto register_failed; | 85 | goto register_failed; |
85 | } | 86 | } |
86 | 87 | ||
87 | priv->parent = ppriv->dev; | ||
88 | |||
89 | ipoib_create_debug_files(priv->dev); | 88 | ipoib_create_debug_files(priv->dev); |
90 | 89 | ||
91 | /* RTNL childs don't need proprietary sysfs entries */ | 90 | /* RTNL childs don't need proprietary sysfs entries */ |
diff --git a/drivers/net/dsa/mv88e6xxx.c b/drivers/net/dsa/mv88e6xxx.c index fc8d3b6ffe8e..9f0c2b9d58ae 100644 --- a/drivers/net/dsa/mv88e6xxx.c +++ b/drivers/net/dsa/mv88e6xxx.c | |||
@@ -602,8 +602,6 @@ static void _mv88e6xxx_get_ethtool_stats(struct dsa_switch *ds, | |||
602 | u32 high = 0; | 602 | u32 high = 0; |
603 | 603 | ||
604 | if (s->reg >= 0x100) { | 604 | if (s->reg >= 0x100) { |
605 | int ret; | ||
606 | |||
607 | ret = mv88e6xxx_reg_read(ds, REG_PORT(port), | 605 | ret = mv88e6xxx_reg_read(ds, REG_PORT(port), |
608 | s->reg - 0x100); | 606 | s->reg - 0x100); |
609 | if (ret < 0) | 607 | if (ret < 0) |
@@ -902,14 +900,16 @@ static int _mv88e6xxx_flush_fid(struct dsa_switch *ds, int fid) | |||
902 | static int mv88e6xxx_set_port_state(struct dsa_switch *ds, int port, u8 state) | 900 | static int mv88e6xxx_set_port_state(struct dsa_switch *ds, int port, u8 state) |
903 | { | 901 | { |
904 | struct mv88e6xxx_priv_state *ps = ds_to_priv(ds); | 902 | struct mv88e6xxx_priv_state *ps = ds_to_priv(ds); |
905 | int reg, ret; | 903 | int reg, ret = 0; |
906 | u8 oldstate; | 904 | u8 oldstate; |
907 | 905 | ||
908 | mutex_lock(&ps->smi_mutex); | 906 | mutex_lock(&ps->smi_mutex); |
909 | 907 | ||
910 | reg = _mv88e6xxx_reg_read(ds, REG_PORT(port), PORT_CONTROL); | 908 | reg = _mv88e6xxx_reg_read(ds, REG_PORT(port), PORT_CONTROL); |
911 | if (reg < 0) | 909 | if (reg < 0) { |
910 | ret = reg; | ||
912 | goto abort; | 911 | goto abort; |
912 | } | ||
913 | 913 | ||
914 | oldstate = reg & PORT_CONTROL_STATE_MASK; | 914 | oldstate = reg & PORT_CONTROL_STATE_MASK; |
915 | if (oldstate != state) { | 915 | if (oldstate != state) { |
diff --git a/drivers/net/ethernet/altera/altera_tse_main.c b/drivers/net/ethernet/altera/altera_tse_main.c index 79ea35869e1e..90a76306ad0f 100644 --- a/drivers/net/ethernet/altera/altera_tse_main.c +++ b/drivers/net/ethernet/altera/altera_tse_main.c | |||
@@ -376,8 +376,13 @@ static int tse_rx(struct altera_tse_private *priv, int limit) | |||
376 | u16 pktlength; | 376 | u16 pktlength; |
377 | u16 pktstatus; | 377 | u16 pktstatus; |
378 | 378 | ||
379 | while (((rxstatus = priv->dmaops->get_rx_status(priv)) != 0) && | 379 | /* Check for count < limit first as get_rx_status is changing |
380 | (count < limit)) { | 380 | * the response-fifo so we must process the next packet |
381 | * after calling get_rx_status if a response is pending. | ||
382 | * (reading the last byte of the response pops the value from the fifo.) | ||
383 | */ | ||
384 | while ((count < limit) && | ||
385 | ((rxstatus = priv->dmaops->get_rx_status(priv)) != 0)) { | ||
381 | pktstatus = rxstatus >> 16; | 386 | pktstatus = rxstatus >> 16; |
382 | pktlength = rxstatus & 0xffff; | 387 | pktlength = rxstatus & 0xffff; |
383 | 388 | ||
diff --git a/drivers/net/ethernet/broadcom/bnx2x/bnx2x.h b/drivers/net/ethernet/broadcom/bnx2x/bnx2x.h index 4085c4b31047..355d5fea5be9 100644 --- a/drivers/net/ethernet/broadcom/bnx2x/bnx2x.h +++ b/drivers/net/ethernet/broadcom/bnx2x/bnx2x.h | |||
@@ -531,20 +531,8 @@ struct bnx2x_fastpath { | |||
531 | struct napi_struct napi; | 531 | struct napi_struct napi; |
532 | 532 | ||
533 | #ifdef CONFIG_NET_RX_BUSY_POLL | 533 | #ifdef CONFIG_NET_RX_BUSY_POLL |
534 | unsigned int state; | 534 | unsigned long busy_poll_state; |
535 | #define BNX2X_FP_STATE_IDLE 0 | 535 | #endif |
536 | #define BNX2X_FP_STATE_NAPI (1 << 0) /* NAPI owns this FP */ | ||
537 | #define BNX2X_FP_STATE_POLL (1 << 1) /* poll owns this FP */ | ||
538 | #define BNX2X_FP_STATE_DISABLED (1 << 2) | ||
539 | #define BNX2X_FP_STATE_NAPI_YIELD (1 << 3) /* NAPI yielded this FP */ | ||
540 | #define BNX2X_FP_STATE_POLL_YIELD (1 << 4) /* poll yielded this FP */ | ||
541 | #define BNX2X_FP_OWNED (BNX2X_FP_STATE_NAPI | BNX2X_FP_STATE_POLL) | ||
542 | #define BNX2X_FP_YIELD (BNX2X_FP_STATE_NAPI_YIELD | BNX2X_FP_STATE_POLL_YIELD) | ||
543 | #define BNX2X_FP_LOCKED (BNX2X_FP_OWNED | BNX2X_FP_STATE_DISABLED) | ||
544 | #define BNX2X_FP_USER_PEND (BNX2X_FP_STATE_POLL | BNX2X_FP_STATE_POLL_YIELD) | ||
545 | /* protect state */ | ||
546 | spinlock_t lock; | ||
547 | #endif /* CONFIG_NET_RX_BUSY_POLL */ | ||
548 | 536 | ||
549 | union host_hc_status_block status_blk; | 537 | union host_hc_status_block status_blk; |
550 | /* chip independent shortcuts into sb structure */ | 538 | /* chip independent shortcuts into sb structure */ |
@@ -619,104 +607,83 @@ struct bnx2x_fastpath { | |||
619 | #define bnx2x_fp_qstats(bp, fp) (&((bp)->fp_stats[(fp)->index].eth_q_stats)) | 607 | #define bnx2x_fp_qstats(bp, fp) (&((bp)->fp_stats[(fp)->index].eth_q_stats)) |
620 | 608 | ||
621 | #ifdef CONFIG_NET_RX_BUSY_POLL | 609 | #ifdef CONFIG_NET_RX_BUSY_POLL |
622 | static inline void bnx2x_fp_init_lock(struct bnx2x_fastpath *fp) | 610 | |
611 | enum bnx2x_fp_state { | ||
612 | BNX2X_STATE_FP_NAPI = BIT(0), /* NAPI handler owns the queue */ | ||
613 | |||
614 | BNX2X_STATE_FP_NAPI_REQ_BIT = 1, /* NAPI would like to own the queue */ | ||
615 | BNX2X_STATE_FP_NAPI_REQ = BIT(1), | ||
616 | |||
617 | BNX2X_STATE_FP_POLL_BIT = 2, | ||
618 | BNX2X_STATE_FP_POLL = BIT(2), /* busy_poll owns the queue */ | ||
619 | |||
620 | BNX2X_STATE_FP_DISABLE_BIT = 3, /* queue is dismantled */ | ||
621 | }; | ||
622 | |||
623 | static inline void bnx2x_fp_busy_poll_init(struct bnx2x_fastpath *fp) | ||
623 | { | 624 | { |
624 | spin_lock_init(&fp->lock); | 625 | WRITE_ONCE(fp->busy_poll_state, 0); |
625 | fp->state = BNX2X_FP_STATE_IDLE; | ||
626 | } | 626 | } |
627 | 627 | ||
628 | /* called from the device poll routine to get ownership of a FP */ | 628 | /* called from the device poll routine to get ownership of a FP */ |
629 | static inline bool bnx2x_fp_lock_napi(struct bnx2x_fastpath *fp) | 629 | static inline bool bnx2x_fp_lock_napi(struct bnx2x_fastpath *fp) |
630 | { | 630 | { |
631 | bool rc = true; | 631 | unsigned long prev, old = READ_ONCE(fp->busy_poll_state); |
632 | 632 | ||
633 | spin_lock_bh(&fp->lock); | 633 | while (1) { |
634 | if (fp->state & BNX2X_FP_LOCKED) { | 634 | switch (old) { |
635 | WARN_ON(fp->state & BNX2X_FP_STATE_NAPI); | 635 | case BNX2X_STATE_FP_POLL: |
636 | fp->state |= BNX2X_FP_STATE_NAPI_YIELD; | 636 | /* make sure bnx2x_fp_lock_poll() wont starve us */ |
637 | rc = false; | 637 | set_bit(BNX2X_STATE_FP_NAPI_REQ_BIT, |
638 | } else { | 638 | &fp->busy_poll_state); |
639 | /* we don't care if someone yielded */ | 639 | /* fallthrough */ |
640 | fp->state = BNX2X_FP_STATE_NAPI; | 640 | case BNX2X_STATE_FP_POLL | BNX2X_STATE_FP_NAPI_REQ: |
641 | return false; | ||
642 | default: | ||
643 | break; | ||
644 | } | ||
645 | prev = cmpxchg(&fp->busy_poll_state, old, BNX2X_STATE_FP_NAPI); | ||
646 | if (unlikely(prev != old)) { | ||
647 | old = prev; | ||
648 | continue; | ||
649 | } | ||
650 | return true; | ||
641 | } | 651 | } |
642 | spin_unlock_bh(&fp->lock); | ||
643 | return rc; | ||
644 | } | 652 | } |
645 | 653 | ||
646 | /* returns true is someone tried to get the FP while napi had it */ | 654 | static inline void bnx2x_fp_unlock_napi(struct bnx2x_fastpath *fp) |
647 | static inline bool bnx2x_fp_unlock_napi(struct bnx2x_fastpath *fp) | ||
648 | { | 655 | { |
649 | bool rc = false; | 656 | smp_wmb(); |
650 | 657 | fp->busy_poll_state = 0; | |
651 | spin_lock_bh(&fp->lock); | ||
652 | WARN_ON(fp->state & | ||
653 | (BNX2X_FP_STATE_POLL | BNX2X_FP_STATE_NAPI_YIELD)); | ||
654 | |||
655 | if (fp->state & BNX2X_FP_STATE_POLL_YIELD) | ||
656 | rc = true; | ||
657 | |||
658 | /* state ==> idle, unless currently disabled */ | ||
659 | fp->state &= BNX2X_FP_STATE_DISABLED; | ||
660 | spin_unlock_bh(&fp->lock); | ||
661 | return rc; | ||
662 | } | 658 | } |
663 | 659 | ||
664 | /* called from bnx2x_low_latency_poll() */ | 660 | /* called from bnx2x_low_latency_poll() */ |
665 | static inline bool bnx2x_fp_lock_poll(struct bnx2x_fastpath *fp) | 661 | static inline bool bnx2x_fp_lock_poll(struct bnx2x_fastpath *fp) |
666 | { | 662 | { |
667 | bool rc = true; | 663 | return cmpxchg(&fp->busy_poll_state, 0, BNX2X_STATE_FP_POLL) == 0; |
668 | |||
669 | spin_lock_bh(&fp->lock); | ||
670 | if ((fp->state & BNX2X_FP_LOCKED)) { | ||
671 | fp->state |= BNX2X_FP_STATE_POLL_YIELD; | ||
672 | rc = false; | ||
673 | } else { | ||
674 | /* preserve yield marks */ | ||
675 | fp->state |= BNX2X_FP_STATE_POLL; | ||
676 | } | ||
677 | spin_unlock_bh(&fp->lock); | ||
678 | return rc; | ||
679 | } | 664 | } |
680 | 665 | ||
681 | /* returns true if someone tried to get the FP while it was locked */ | 666 | static inline void bnx2x_fp_unlock_poll(struct bnx2x_fastpath *fp) |
682 | static inline bool bnx2x_fp_unlock_poll(struct bnx2x_fastpath *fp) | ||
683 | { | 667 | { |
684 | bool rc = false; | 668 | smp_mb__before_atomic(); |
685 | 669 | clear_bit(BNX2X_STATE_FP_POLL_BIT, &fp->busy_poll_state); | |
686 | spin_lock_bh(&fp->lock); | ||
687 | WARN_ON(fp->state & BNX2X_FP_STATE_NAPI); | ||
688 | |||
689 | if (fp->state & BNX2X_FP_STATE_POLL_YIELD) | ||
690 | rc = true; | ||
691 | |||
692 | /* state ==> idle, unless currently disabled */ | ||
693 | fp->state &= BNX2X_FP_STATE_DISABLED; | ||
694 | spin_unlock_bh(&fp->lock); | ||
695 | return rc; | ||
696 | } | 670 | } |
697 | 671 | ||
698 | /* true if a socket is polling, even if it did not get the lock */ | 672 | /* true if a socket is polling */ |
699 | static inline bool bnx2x_fp_ll_polling(struct bnx2x_fastpath *fp) | 673 | static inline bool bnx2x_fp_ll_polling(struct bnx2x_fastpath *fp) |
700 | { | 674 | { |
701 | WARN_ON(!(fp->state & BNX2X_FP_OWNED)); | 675 | return READ_ONCE(fp->busy_poll_state) & BNX2X_STATE_FP_POLL; |
702 | return fp->state & BNX2X_FP_USER_PEND; | ||
703 | } | 676 | } |
704 | 677 | ||
705 | /* false if fp is currently owned */ | 678 | /* false if fp is currently owned */ |
706 | static inline bool bnx2x_fp_ll_disable(struct bnx2x_fastpath *fp) | 679 | static inline bool bnx2x_fp_ll_disable(struct bnx2x_fastpath *fp) |
707 | { | 680 | { |
708 | int rc = true; | 681 | set_bit(BNX2X_STATE_FP_DISABLE_BIT, &fp->busy_poll_state); |
709 | 682 | return !bnx2x_fp_ll_polling(fp); | |
710 | spin_lock_bh(&fp->lock); | ||
711 | if (fp->state & BNX2X_FP_OWNED) | ||
712 | rc = false; | ||
713 | fp->state |= BNX2X_FP_STATE_DISABLED; | ||
714 | spin_unlock_bh(&fp->lock); | ||
715 | 683 | ||
716 | return rc; | ||
717 | } | 684 | } |
718 | #else | 685 | #else |
719 | static inline void bnx2x_fp_init_lock(struct bnx2x_fastpath *fp) | 686 | static inline void bnx2x_fp_busy_poll_init(struct bnx2x_fastpath *fp) |
720 | { | 687 | { |
721 | } | 688 | } |
722 | 689 | ||
@@ -725,9 +692,8 @@ static inline bool bnx2x_fp_lock_napi(struct bnx2x_fastpath *fp) | |||
725 | return true; | 692 | return true; |
726 | } | 693 | } |
727 | 694 | ||
728 | static inline bool bnx2x_fp_unlock_napi(struct bnx2x_fastpath *fp) | 695 | static inline void bnx2x_fp_unlock_napi(struct bnx2x_fastpath *fp) |
729 | { | 696 | { |
730 | return false; | ||
731 | } | 697 | } |
732 | 698 | ||
733 | static inline bool bnx2x_fp_lock_poll(struct bnx2x_fastpath *fp) | 699 | static inline bool bnx2x_fp_lock_poll(struct bnx2x_fastpath *fp) |
@@ -735,9 +701,8 @@ static inline bool bnx2x_fp_lock_poll(struct bnx2x_fastpath *fp) | |||
735 | return false; | 701 | return false; |
736 | } | 702 | } |
737 | 703 | ||
738 | static inline bool bnx2x_fp_unlock_poll(struct bnx2x_fastpath *fp) | 704 | static inline void bnx2x_fp_unlock_poll(struct bnx2x_fastpath *fp) |
739 | { | 705 | { |
740 | return false; | ||
741 | } | 706 | } |
742 | 707 | ||
743 | static inline bool bnx2x_fp_ll_polling(struct bnx2x_fastpath *fp) | 708 | static inline bool bnx2x_fp_ll_polling(struct bnx2x_fastpath *fp) |
diff --git a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c index 0a9faa134a9a..2f63467bce46 100644 --- a/drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c +++ b/drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c | |||
@@ -1849,7 +1849,7 @@ static void bnx2x_napi_enable_cnic(struct bnx2x *bp) | |||
1849 | int i; | 1849 | int i; |
1850 | 1850 | ||
1851 | for_each_rx_queue_cnic(bp, i) { | 1851 | for_each_rx_queue_cnic(bp, i) { |
1852 | bnx2x_fp_init_lock(&bp->fp[i]); | 1852 | bnx2x_fp_busy_poll_init(&bp->fp[i]); |
1853 | napi_enable(&bnx2x_fp(bp, i, napi)); | 1853 | napi_enable(&bnx2x_fp(bp, i, napi)); |
1854 | } | 1854 | } |
1855 | } | 1855 | } |
@@ -1859,7 +1859,7 @@ static void bnx2x_napi_enable(struct bnx2x *bp) | |||
1859 | int i; | 1859 | int i; |
1860 | 1860 | ||
1861 | for_each_eth_queue(bp, i) { | 1861 | for_each_eth_queue(bp, i) { |
1862 | bnx2x_fp_init_lock(&bp->fp[i]); | 1862 | bnx2x_fp_busy_poll_init(&bp->fp[i]); |
1863 | napi_enable(&bnx2x_fp(bp, i, napi)); | 1863 | napi_enable(&bnx2x_fp(bp, i, napi)); |
1864 | } | 1864 | } |
1865 | } | 1865 | } |
@@ -3191,9 +3191,10 @@ static int bnx2x_poll(struct napi_struct *napi, int budget) | |||
3191 | } | 3191 | } |
3192 | } | 3192 | } |
3193 | 3193 | ||
3194 | bnx2x_fp_unlock_napi(fp); | ||
3195 | |||
3194 | /* Fall out from the NAPI loop if needed */ | 3196 | /* Fall out from the NAPI loop if needed */ |
3195 | if (!bnx2x_fp_unlock_napi(fp) && | 3197 | if (!(bnx2x_has_rx_work(fp) || bnx2x_has_tx_work(fp))) { |
3196 | !(bnx2x_has_rx_work(fp) || bnx2x_has_tx_work(fp))) { | ||
3197 | 3198 | ||
3198 | /* No need to update SB for FCoE L2 ring as long as | 3199 | /* No need to update SB for FCoE L2 ring as long as |
3199 | * it's connected to the default SB and the SB | 3200 | * it's connected to the default SB and the SB |
diff --git a/drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c b/drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c index 6de054404156..803d91beec6f 100644 --- a/drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c +++ b/drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c | |||
@@ -1140,6 +1140,10 @@ static int set_filter_wr(struct adapter *adapter, int fidx) | |||
1140 | struct fw_filter_wr *fwr; | 1140 | struct fw_filter_wr *fwr; |
1141 | unsigned int ftid; | 1141 | unsigned int ftid; |
1142 | 1142 | ||
1143 | skb = alloc_skb(sizeof(*fwr), GFP_KERNEL); | ||
1144 | if (!skb) | ||
1145 | return -ENOMEM; | ||
1146 | |||
1143 | /* If the new filter requires loopback Destination MAC and/or VLAN | 1147 | /* If the new filter requires loopback Destination MAC and/or VLAN |
1144 | * rewriting then we need to allocate a Layer 2 Table (L2T) entry for | 1148 | * rewriting then we need to allocate a Layer 2 Table (L2T) entry for |
1145 | * the filter. | 1149 | * the filter. |
@@ -1147,19 +1151,21 @@ static int set_filter_wr(struct adapter *adapter, int fidx) | |||
1147 | if (f->fs.newdmac || f->fs.newvlan) { | 1151 | if (f->fs.newdmac || f->fs.newvlan) { |
1148 | /* allocate L2T entry for new filter */ | 1152 | /* allocate L2T entry for new filter */ |
1149 | f->l2t = t4_l2t_alloc_switching(adapter->l2t); | 1153 | f->l2t = t4_l2t_alloc_switching(adapter->l2t); |
1150 | if (f->l2t == NULL) | 1154 | if (f->l2t == NULL) { |
1155 | kfree_skb(skb); | ||
1151 | return -EAGAIN; | 1156 | return -EAGAIN; |
1157 | } | ||
1152 | if (t4_l2t_set_switching(adapter, f->l2t, f->fs.vlan, | 1158 | if (t4_l2t_set_switching(adapter, f->l2t, f->fs.vlan, |
1153 | f->fs.eport, f->fs.dmac)) { | 1159 | f->fs.eport, f->fs.dmac)) { |
1154 | cxgb4_l2t_release(f->l2t); | 1160 | cxgb4_l2t_release(f->l2t); |
1155 | f->l2t = NULL; | 1161 | f->l2t = NULL; |
1162 | kfree_skb(skb); | ||
1156 | return -ENOMEM; | 1163 | return -ENOMEM; |
1157 | } | 1164 | } |
1158 | } | 1165 | } |
1159 | 1166 | ||
1160 | ftid = adapter->tids.ftid_base + fidx; | 1167 | ftid = adapter->tids.ftid_base + fidx; |
1161 | 1168 | ||
1162 | skb = alloc_skb(sizeof(*fwr), GFP_KERNEL | __GFP_NOFAIL); | ||
1163 | fwr = (struct fw_filter_wr *)__skb_put(skb, sizeof(*fwr)); | 1169 | fwr = (struct fw_filter_wr *)__skb_put(skb, sizeof(*fwr)); |
1164 | memset(fwr, 0, sizeof(*fwr)); | 1170 | memset(fwr, 0, sizeof(*fwr)); |
1165 | 1171 | ||
@@ -1257,7 +1263,10 @@ static int del_filter_wr(struct adapter *adapter, int fidx) | |||
1257 | len = sizeof(*fwr); | 1263 | len = sizeof(*fwr); |
1258 | ftid = adapter->tids.ftid_base + fidx; | 1264 | ftid = adapter->tids.ftid_base + fidx; |
1259 | 1265 | ||
1260 | skb = alloc_skb(len, GFP_KERNEL | __GFP_NOFAIL); | 1266 | skb = alloc_skb(len, GFP_KERNEL); |
1267 | if (!skb) | ||
1268 | return -ENOMEM; | ||
1269 | |||
1261 | fwr = (struct fw_filter_wr *)__skb_put(skb, len); | 1270 | fwr = (struct fw_filter_wr *)__skb_put(skb, len); |
1262 | t4_mk_filtdelwr(ftid, fwr, adapter->sge.fw_evtq.abs_id); | 1271 | t4_mk_filtdelwr(ftid, fwr, adapter->sge.fw_evtq.abs_id); |
1263 | 1272 | ||
diff --git a/drivers/net/ethernet/hisilicon/hip04_eth.c b/drivers/net/ethernet/hisilicon/hip04_eth.c index b72d238695d7..3b39fdddeb57 100644 --- a/drivers/net/ethernet/hisilicon/hip04_eth.c +++ b/drivers/net/ethernet/hisilicon/hip04_eth.c | |||
@@ -413,6 +413,15 @@ out: | |||
413 | return count; | 413 | return count; |
414 | } | 414 | } |
415 | 415 | ||
416 | static void hip04_start_tx_timer(struct hip04_priv *priv) | ||
417 | { | ||
418 | unsigned long ns = priv->tx_coalesce_usecs * NSEC_PER_USEC / 2; | ||
419 | |||
420 | /* allow timer to fire after half the time at the earliest */ | ||
421 | hrtimer_start_range_ns(&priv->tx_coalesce_timer, ns_to_ktime(ns), | ||
422 | ns, HRTIMER_MODE_REL); | ||
423 | } | ||
424 | |||
416 | static int hip04_mac_start_xmit(struct sk_buff *skb, struct net_device *ndev) | 425 | static int hip04_mac_start_xmit(struct sk_buff *skb, struct net_device *ndev) |
417 | { | 426 | { |
418 | struct hip04_priv *priv = netdev_priv(ndev); | 427 | struct hip04_priv *priv = netdev_priv(ndev); |
@@ -466,8 +475,7 @@ static int hip04_mac_start_xmit(struct sk_buff *skb, struct net_device *ndev) | |||
466 | } | 475 | } |
467 | } else if (!hrtimer_is_queued(&priv->tx_coalesce_timer)) { | 476 | } else if (!hrtimer_is_queued(&priv->tx_coalesce_timer)) { |
468 | /* cleanup not pending yet, start a new timer */ | 477 | /* cleanup not pending yet, start a new timer */ |
469 | hrtimer_start_expires(&priv->tx_coalesce_timer, | 478 | hip04_start_tx_timer(priv); |
470 | HRTIMER_MODE_REL); | ||
471 | } | 479 | } |
472 | 480 | ||
473 | return NETDEV_TX_OK; | 481 | return NETDEV_TX_OK; |
@@ -549,7 +557,7 @@ done: | |||
549 | /* clean up tx descriptors and start a new timer if necessary */ | 557 | /* clean up tx descriptors and start a new timer if necessary */ |
550 | tx_remaining = hip04_tx_reclaim(ndev, false); | 558 | tx_remaining = hip04_tx_reclaim(ndev, false); |
551 | if (rx < budget && tx_remaining) | 559 | if (rx < budget && tx_remaining) |
552 | hrtimer_start_expires(&priv->tx_coalesce_timer, HRTIMER_MODE_REL); | 560 | hip04_start_tx_timer(priv); |
553 | 561 | ||
554 | return rx; | 562 | return rx; |
555 | } | 563 | } |
@@ -809,7 +817,6 @@ static int hip04_mac_probe(struct platform_device *pdev) | |||
809 | struct hip04_priv *priv; | 817 | struct hip04_priv *priv; |
810 | struct resource *res; | 818 | struct resource *res; |
811 | unsigned int irq; | 819 | unsigned int irq; |
812 | ktime_t txtime; | ||
813 | int ret; | 820 | int ret; |
814 | 821 | ||
815 | ndev = alloc_etherdev(sizeof(struct hip04_priv)); | 822 | ndev = alloc_etherdev(sizeof(struct hip04_priv)); |
@@ -846,9 +853,6 @@ static int hip04_mac_probe(struct platform_device *pdev) | |||
846 | */ | 853 | */ |
847 | priv->tx_coalesce_frames = TX_DESC_NUM * 3 / 4; | 854 | priv->tx_coalesce_frames = TX_DESC_NUM * 3 / 4; |
848 | priv->tx_coalesce_usecs = 200; | 855 | priv->tx_coalesce_usecs = 200; |
849 | /* allow timer to fire after half the time at the earliest */ | ||
850 | txtime = ktime_set(0, priv->tx_coalesce_usecs * NSEC_PER_USEC / 2); | ||
851 | hrtimer_set_expires_range(&priv->tx_coalesce_timer, txtime, txtime); | ||
852 | priv->tx_coalesce_timer.function = tx_done; | 856 | priv->tx_coalesce_timer.function = tx_done; |
853 | 857 | ||
854 | priv->map = syscon_node_to_regmap(arg.np); | 858 | priv->map = syscon_node_to_regmap(arg.np); |
diff --git a/drivers/net/ethernet/intel/i40e/i40e_common.c b/drivers/net/ethernet/intel/i40e/i40e_common.c index d596f6624025..0bae22da014d 100644 --- a/drivers/net/ethernet/intel/i40e/i40e_common.c +++ b/drivers/net/ethernet/intel/i40e/i40e_common.c | |||
@@ -2397,6 +2397,7 @@ i40e_aq_erase_nvm_exit: | |||
2397 | #define I40E_DEV_FUNC_CAP_LED 0x61 | 2397 | #define I40E_DEV_FUNC_CAP_LED 0x61 |
2398 | #define I40E_DEV_FUNC_CAP_SDP 0x62 | 2398 | #define I40E_DEV_FUNC_CAP_SDP 0x62 |
2399 | #define I40E_DEV_FUNC_CAP_MDIO 0x63 | 2399 | #define I40E_DEV_FUNC_CAP_MDIO 0x63 |
2400 | #define I40E_DEV_FUNC_CAP_WR_CSR_PROT 0x64 | ||
2400 | 2401 | ||
2401 | /** | 2402 | /** |
2402 | * i40e_parse_discover_capabilities | 2403 | * i40e_parse_discover_capabilities |
@@ -2541,11 +2542,18 @@ static void i40e_parse_discover_capabilities(struct i40e_hw *hw, void *buff, | |||
2541 | p->fd_filters_guaranteed = number; | 2542 | p->fd_filters_guaranteed = number; |
2542 | p->fd_filters_best_effort = logical_id; | 2543 | p->fd_filters_best_effort = logical_id; |
2543 | break; | 2544 | break; |
2545 | case I40E_DEV_FUNC_CAP_WR_CSR_PROT: | ||
2546 | p->wr_csr_prot = (u64)number; | ||
2547 | p->wr_csr_prot |= (u64)logical_id << 32; | ||
2548 | break; | ||
2544 | default: | 2549 | default: |
2545 | break; | 2550 | break; |
2546 | } | 2551 | } |
2547 | } | 2552 | } |
2548 | 2553 | ||
2554 | if (p->fcoe) | ||
2555 | i40e_debug(hw, I40E_DEBUG_ALL, "device is FCoE capable\n"); | ||
2556 | |||
2549 | /* Software override ensuring FCoE is disabled if npar or mfp | 2557 | /* Software override ensuring FCoE is disabled if npar or mfp |
2550 | * mode because it is not supported in these modes. | 2558 | * mode because it is not supported in these modes. |
2551 | */ | 2559 | */ |
@@ -3503,6 +3511,63 @@ void i40e_set_pci_config_data(struct i40e_hw *hw, u16 link_status) | |||
3503 | } | 3511 | } |
3504 | 3512 | ||
3505 | /** | 3513 | /** |
3514 | * i40e_aq_debug_dump | ||
3515 | * @hw: pointer to the hardware structure | ||
3516 | * @cluster_id: specific cluster to dump | ||
3517 | * @table_id: table id within cluster | ||
3518 | * @start_index: index of line in the block to read | ||
3519 | * @buff_size: dump buffer size | ||
3520 | * @buff: dump buffer | ||
3521 | * @ret_buff_size: actual buffer size returned | ||
3522 | * @ret_next_table: next block to read | ||
3523 | * @ret_next_index: next index to read | ||
3524 | * | ||
3525 | * Dump internal FW/HW data for debug purposes. | ||
3526 | * | ||
3527 | **/ | ||
3528 | i40e_status i40e_aq_debug_dump(struct i40e_hw *hw, u8 cluster_id, | ||
3529 | u8 table_id, u32 start_index, u16 buff_size, | ||
3530 | void *buff, u16 *ret_buff_size, | ||
3531 | u8 *ret_next_table, u32 *ret_next_index, | ||
3532 | struct i40e_asq_cmd_details *cmd_details) | ||
3533 | { | ||
3534 | struct i40e_aq_desc desc; | ||
3535 | struct i40e_aqc_debug_dump_internals *cmd = | ||
3536 | (struct i40e_aqc_debug_dump_internals *)&desc.params.raw; | ||
3537 | struct i40e_aqc_debug_dump_internals *resp = | ||
3538 | (struct i40e_aqc_debug_dump_internals *)&desc.params.raw; | ||
3539 | i40e_status status; | ||
3540 | |||
3541 | if (buff_size == 0 || !buff) | ||
3542 | return I40E_ERR_PARAM; | ||
3543 | |||
3544 | i40e_fill_default_direct_cmd_desc(&desc, | ||
3545 | i40e_aqc_opc_debug_dump_internals); | ||
3546 | /* Indirect Command */ | ||
3547 | desc.flags |= cpu_to_le16((u16)I40E_AQ_FLAG_BUF); | ||
3548 | if (buff_size > I40E_AQ_LARGE_BUF) | ||
3549 | desc.flags |= cpu_to_le16((u16)I40E_AQ_FLAG_LB); | ||
3550 | |||
3551 | cmd->cluster_id = cluster_id; | ||
3552 | cmd->table_id = table_id; | ||
3553 | cmd->idx = cpu_to_le32(start_index); | ||
3554 | |||
3555 | desc.datalen = cpu_to_le16(buff_size); | ||
3556 | |||
3557 | status = i40e_asq_send_command(hw, &desc, buff, buff_size, cmd_details); | ||
3558 | if (!status) { | ||
3559 | if (ret_buff_size) | ||
3560 | *ret_buff_size = le16_to_cpu(desc.datalen); | ||
3561 | if (ret_next_table) | ||
3562 | *ret_next_table = resp->table_id; | ||
3563 | if (ret_next_index) | ||
3564 | *ret_next_index = le32_to_cpu(resp->idx); | ||
3565 | } | ||
3566 | |||
3567 | return status; | ||
3568 | } | ||
3569 | |||
3570 | /** | ||
3506 | * i40e_read_bw_from_alt_ram | 3571 | * i40e_read_bw_from_alt_ram |
3507 | * @hw: pointer to the hardware structure | 3572 | * @hw: pointer to the hardware structure |
3508 | * @max_bw: pointer for max_bw read | 3573 | * @max_bw: pointer for max_bw read |
diff --git a/drivers/net/ethernet/intel/i40e/i40e_dcb.c b/drivers/net/ethernet/intel/i40e/i40e_dcb.c index 6e1466756760..2547aa21b2ca 100644 --- a/drivers/net/ethernet/intel/i40e/i40e_dcb.c +++ b/drivers/net/ethernet/intel/i40e/i40e_dcb.c | |||
@@ -419,7 +419,7 @@ static void i40e_cee_to_dcb_v1_config( | |||
419 | { | 419 | { |
420 | u16 status, tlv_status = le16_to_cpu(cee_cfg->tlv_status); | 420 | u16 status, tlv_status = le16_to_cpu(cee_cfg->tlv_status); |
421 | u16 app_prio = le16_to_cpu(cee_cfg->oper_app_prio); | 421 | u16 app_prio = le16_to_cpu(cee_cfg->oper_app_prio); |
422 | u8 i, tc, err, sync, oper; | 422 | u8 i, tc, err; |
423 | 423 | ||
424 | /* CEE PG data to ETS config */ | 424 | /* CEE PG data to ETS config */ |
425 | dcbcfg->etscfg.maxtcs = cee_cfg->oper_num_tc; | 425 | dcbcfg->etscfg.maxtcs = cee_cfg->oper_num_tc; |
@@ -456,9 +456,7 @@ static void i40e_cee_to_dcb_v1_config( | |||
456 | status = (tlv_status & I40E_AQC_CEE_APP_STATUS_MASK) >> | 456 | status = (tlv_status & I40E_AQC_CEE_APP_STATUS_MASK) >> |
457 | I40E_AQC_CEE_APP_STATUS_SHIFT; | 457 | I40E_AQC_CEE_APP_STATUS_SHIFT; |
458 | err = (status & I40E_TLV_STATUS_ERR) ? 1 : 0; | 458 | err = (status & I40E_TLV_STATUS_ERR) ? 1 : 0; |
459 | sync = (status & I40E_TLV_STATUS_SYNC) ? 1 : 0; | 459 | /* Add APPs if Error is False */ |
460 | oper = (status & I40E_TLV_STATUS_OPER) ? 1 : 0; | ||
461 | /* Add APPs if Error is False and Oper/Sync is True */ | ||
462 | if (!err) { | 460 | if (!err) { |
463 | /* CEE operating configuration supports FCoE/iSCSI/FIP only */ | 461 | /* CEE operating configuration supports FCoE/iSCSI/FIP only */ |
464 | dcbcfg->numapps = I40E_CEE_OPER_MAX_APPS; | 462 | dcbcfg->numapps = I40E_CEE_OPER_MAX_APPS; |
diff --git a/drivers/net/ethernet/intel/i40e/i40e_debugfs.c b/drivers/net/ethernet/intel/i40e/i40e_debugfs.c index daa88263af66..34170eabca7d 100644 --- a/drivers/net/ethernet/intel/i40e/i40e_debugfs.c +++ b/drivers/net/ethernet/intel/i40e/i40e_debugfs.c | |||
@@ -1388,6 +1388,50 @@ static ssize_t i40e_dbg_command_write(struct file *filp, | |||
1388 | r_cfg->app[i].selector, | 1388 | r_cfg->app[i].selector, |
1389 | r_cfg->app[i].protocolid); | 1389 | r_cfg->app[i].protocolid); |
1390 | } | 1390 | } |
1391 | } else if (strncmp(&cmd_buf[5], "debug fwdata", 12) == 0) { | ||
1392 | int cluster_id, table_id; | ||
1393 | int index, ret; | ||
1394 | u16 buff_len = 4096; | ||
1395 | u32 next_index; | ||
1396 | u8 next_table; | ||
1397 | u8 *buff; | ||
1398 | u16 rlen; | ||
1399 | |||
1400 | cnt = sscanf(&cmd_buf[18], "%i %i %i", | ||
1401 | &cluster_id, &table_id, &index); | ||
1402 | if (cnt != 3) { | ||
1403 | dev_info(&pf->pdev->dev, | ||
1404 | "dump debug fwdata <cluster_id> <table_id> <index>\n"); | ||
1405 | goto command_write_done; | ||
1406 | } | ||
1407 | |||
1408 | dev_info(&pf->pdev->dev, | ||
1409 | "AQ debug dump fwdata params %x %x %x %x\n", | ||
1410 | cluster_id, table_id, index, buff_len); | ||
1411 | buff = kzalloc(buff_len, GFP_KERNEL); | ||
1412 | if (!buff) | ||
1413 | goto command_write_done; | ||
1414 | |||
1415 | ret = i40e_aq_debug_dump(&pf->hw, cluster_id, table_id, | ||
1416 | index, buff_len, buff, &rlen, | ||
1417 | &next_table, &next_index, | ||
1418 | NULL); | ||
1419 | if (ret) { | ||
1420 | dev_info(&pf->pdev->dev, | ||
1421 | "debug dump fwdata AQ Failed %d 0x%x\n", | ||
1422 | ret, pf->hw.aq.asq_last_status); | ||
1423 | kfree(buff); | ||
1424 | buff = NULL; | ||
1425 | goto command_write_done; | ||
1426 | } | ||
1427 | dev_info(&pf->pdev->dev, | ||
1428 | "AQ debug dump fwdata rlen=0x%x next_table=0x%x next_index=0x%x\n", | ||
1429 | rlen, next_table, next_index); | ||
1430 | print_hex_dump(KERN_INFO, "AQ buffer WB: ", | ||
1431 | DUMP_PREFIX_OFFSET, 16, 1, | ||
1432 | buff, rlen, true); | ||
1433 | kfree(buff); | ||
1434 | buff = NULL; | ||
1391 | } else { | 1435 | } else { |
1392 | dev_info(&pf->pdev->dev, | 1436 | dev_info(&pf->pdev->dev, |
1393 | "dump desc tx <vsi_seid> <ring_id> [<desc_n>], dump desc rx <vsi_seid> <ring_id> [<desc_n>],\n"); | 1437 | "dump desc tx <vsi_seid> <ring_id> [<desc_n>], dump desc rx <vsi_seid> <ring_id> [<desc_n>],\n"); |
@@ -1903,6 +1947,7 @@ static ssize_t i40e_dbg_command_write(struct file *filp, | |||
1903 | dev_info(&pf->pdev->dev, " dump desc rx <vsi_seid> <ring_id> [<desc_n>]\n"); | 1947 | dev_info(&pf->pdev->dev, " dump desc rx <vsi_seid> <ring_id> [<desc_n>]\n"); |
1904 | dev_info(&pf->pdev->dev, " dump desc aq\n"); | 1948 | dev_info(&pf->pdev->dev, " dump desc aq\n"); |
1905 | dev_info(&pf->pdev->dev, " dump reset stats\n"); | 1949 | dev_info(&pf->pdev->dev, " dump reset stats\n"); |
1950 | dev_info(&pf->pdev->dev, " dump debug fwdata <cluster_id> <table_id> <index>\n"); | ||
1906 | dev_info(&pf->pdev->dev, " msg_enable [level]\n"); | 1951 | dev_info(&pf->pdev->dev, " msg_enable [level]\n"); |
1907 | dev_info(&pf->pdev->dev, " read <reg>\n"); | 1952 | dev_info(&pf->pdev->dev, " read <reg>\n"); |
1908 | dev_info(&pf->pdev->dev, " write <reg> <value>\n"); | 1953 | dev_info(&pf->pdev->dev, " write <reg> <value>\n"); |
diff --git a/drivers/net/ethernet/intel/i40e/i40e_ethtool.c b/drivers/net/ethernet/intel/i40e/i40e_ethtool.c index c848b1862512..4cbaaeb902c4 100644 --- a/drivers/net/ethernet/intel/i40e/i40e_ethtool.c +++ b/drivers/net/ethernet/intel/i40e/i40e_ethtool.c | |||
@@ -356,8 +356,7 @@ static void i40e_get_settings_link_up(struct i40e_hw *hw, | |||
356 | /* Set speed and duplex */ | 356 | /* Set speed and duplex */ |
357 | switch (link_speed) { | 357 | switch (link_speed) { |
358 | case I40E_LINK_SPEED_40GB: | 358 | case I40E_LINK_SPEED_40GB: |
359 | /* need a SPEED_40000 in ethtool.h */ | 359 | ethtool_cmd_speed_set(ecmd, SPEED_40000); |
360 | ethtool_cmd_speed_set(ecmd, 40000); | ||
361 | break; | 360 | break; |
362 | case I40E_LINK_SPEED_20GB: | 361 | case I40E_LINK_SPEED_20GB: |
363 | ethtool_cmd_speed_set(ecmd, SPEED_20000); | 362 | ethtool_cmd_speed_set(ecmd, SPEED_20000); |
@@ -1914,6 +1913,16 @@ static int i40e_get_ethtool_fdir_entry(struct i40e_pf *pf, | |||
1914 | else | 1913 | else |
1915 | fsp->ring_cookie = rule->q_index; | 1914 | fsp->ring_cookie = rule->q_index; |
1916 | 1915 | ||
1916 | if (rule->dest_vsi != pf->vsi[pf->lan_vsi]->id) { | ||
1917 | struct i40e_vsi *vsi; | ||
1918 | |||
1919 | vsi = i40e_find_vsi_from_id(pf, rule->dest_vsi); | ||
1920 | if (vsi && vsi->type == I40E_VSI_SRIOV) { | ||
1921 | fsp->h_ext.data[1] = htonl(vsi->vf_id); | ||
1922 | fsp->m_ext.data[1] = htonl(0x1); | ||
1923 | } | ||
1924 | } | ||
1925 | |||
1917 | return 0; | 1926 | return 0; |
1918 | } | 1927 | } |
1919 | 1928 | ||
@@ -2207,6 +2216,7 @@ static int i40e_add_fdir_ethtool(struct i40e_vsi *vsi, | |||
2207 | struct i40e_fdir_filter *input; | 2216 | struct i40e_fdir_filter *input; |
2208 | struct i40e_pf *pf; | 2217 | struct i40e_pf *pf; |
2209 | int ret = -EINVAL; | 2218 | int ret = -EINVAL; |
2219 | u16 vf_id; | ||
2210 | 2220 | ||
2211 | if (!vsi) | 2221 | if (!vsi) |
2212 | return -EINVAL; | 2222 | return -EINVAL; |
@@ -2267,7 +2277,22 @@ static int i40e_add_fdir_ethtool(struct i40e_vsi *vsi, | |||
2267 | input->dst_ip[0] = fsp->h_u.tcp_ip4_spec.ip4src; | 2277 | input->dst_ip[0] = fsp->h_u.tcp_ip4_spec.ip4src; |
2268 | input->src_ip[0] = fsp->h_u.tcp_ip4_spec.ip4dst; | 2278 | input->src_ip[0] = fsp->h_u.tcp_ip4_spec.ip4dst; |
2269 | 2279 | ||
2280 | if (ntohl(fsp->m_ext.data[1])) { | ||
2281 | if (ntohl(fsp->h_ext.data[1]) >= pf->num_alloc_vfs) { | ||
2282 | netif_info(pf, drv, vsi->netdev, "Invalid VF id\n"); | ||
2283 | goto free_input; | ||
2284 | } | ||
2285 | vf_id = ntohl(fsp->h_ext.data[1]); | ||
2286 | /* Find vsi id from vf id and override dest vsi */ | ||
2287 | input->dest_vsi = pf->vf[vf_id].lan_vsi_id; | ||
2288 | if (input->q_index >= pf->vf[vf_id].num_queue_pairs) { | ||
2289 | netif_info(pf, drv, vsi->netdev, "Invalid queue id\n"); | ||
2290 | goto free_input; | ||
2291 | } | ||
2292 | } | ||
2293 | |||
2270 | ret = i40e_add_del_fdir(vsi, input, true); | 2294 | ret = i40e_add_del_fdir(vsi, input, true); |
2295 | free_input: | ||
2271 | if (ret) | 2296 | if (ret) |
2272 | kfree(input); | 2297 | kfree(input); |
2273 | else | 2298 | else |
diff --git a/drivers/net/ethernet/intel/i40e/i40e_main.c b/drivers/net/ethernet/intel/i40e/i40e_main.c index 63de3f4b7a94..24481cd7e59a 100644 --- a/drivers/net/ethernet/intel/i40e/i40e_main.c +++ b/drivers/net/ethernet/intel/i40e/i40e_main.c | |||
@@ -39,7 +39,7 @@ static const char i40e_driver_string[] = | |||
39 | 39 | ||
40 | #define DRV_VERSION_MAJOR 1 | 40 | #define DRV_VERSION_MAJOR 1 |
41 | #define DRV_VERSION_MINOR 3 | 41 | #define DRV_VERSION_MINOR 3 |
42 | #define DRV_VERSION_BUILD 1 | 42 | #define DRV_VERSION_BUILD 2 |
43 | #define DRV_VERSION __stringify(DRV_VERSION_MAJOR) "." \ | 43 | #define DRV_VERSION __stringify(DRV_VERSION_MAJOR) "." \ |
44 | __stringify(DRV_VERSION_MINOR) "." \ | 44 | __stringify(DRV_VERSION_MINOR) "." \ |
45 | __stringify(DRV_VERSION_BUILD) DRV_KERN | 45 | __stringify(DRV_VERSION_BUILD) DRV_KERN |
@@ -7301,7 +7301,7 @@ err_out: | |||
7301 | * i40e_init_interrupt_scheme - Determine proper interrupt scheme | 7301 | * i40e_init_interrupt_scheme - Determine proper interrupt scheme |
7302 | * @pf: board private structure to initialize | 7302 | * @pf: board private structure to initialize |
7303 | **/ | 7303 | **/ |
7304 | static void i40e_init_interrupt_scheme(struct i40e_pf *pf) | 7304 | static int i40e_init_interrupt_scheme(struct i40e_pf *pf) |
7305 | { | 7305 | { |
7306 | int vectors = 0; | 7306 | int vectors = 0; |
7307 | ssize_t size; | 7307 | ssize_t size; |
@@ -7343,11 +7343,17 @@ static void i40e_init_interrupt_scheme(struct i40e_pf *pf) | |||
7343 | /* set up vector assignment tracking */ | 7343 | /* set up vector assignment tracking */ |
7344 | size = sizeof(struct i40e_lump_tracking) + (sizeof(u16) * vectors); | 7344 | size = sizeof(struct i40e_lump_tracking) + (sizeof(u16) * vectors); |
7345 | pf->irq_pile = kzalloc(size, GFP_KERNEL); | 7345 | pf->irq_pile = kzalloc(size, GFP_KERNEL); |
7346 | if (!pf->irq_pile) { | ||
7347 | dev_err(&pf->pdev->dev, "error allocating irq_pile memory\n"); | ||
7348 | return -ENOMEM; | ||
7349 | } | ||
7346 | pf->irq_pile->num_entries = vectors; | 7350 | pf->irq_pile->num_entries = vectors; |
7347 | pf->irq_pile->search_hint = 0; | 7351 | pf->irq_pile->search_hint = 0; |
7348 | 7352 | ||
7349 | /* track first vector for misc interrupts */ | 7353 | /* track first vector for misc interrupts, ignore return */ |
7350 | (void)i40e_get_lump(pf, pf->irq_pile, 1, I40E_PILE_VALID_BIT - 1); | 7354 | (void)i40e_get_lump(pf, pf->irq_pile, 1, I40E_PILE_VALID_BIT - 1); |
7355 | |||
7356 | return 0; | ||
7351 | } | 7357 | } |
7352 | 7358 | ||
7353 | /** | 7359 | /** |
@@ -9827,7 +9833,9 @@ static int i40e_probe(struct pci_dev *pdev, const struct pci_device_id *ent) | |||
9827 | 9833 | ||
9828 | /* set up the main switch operations */ | 9834 | /* set up the main switch operations */ |
9829 | i40e_determine_queue_usage(pf); | 9835 | i40e_determine_queue_usage(pf); |
9830 | i40e_init_interrupt_scheme(pf); | 9836 | err = i40e_init_interrupt_scheme(pf); |
9837 | if (err) | ||
9838 | goto err_switch_setup; | ||
9831 | 9839 | ||
9832 | /* The number of VSIs reported by the FW is the minimum guaranteed | 9840 | /* The number of VSIs reported by the FW is the minimum guaranteed |
9833 | * to us; HW supports far more and we share the remaining pool with | 9841 | * to us; HW supports far more and we share the remaining pool with |
diff --git a/drivers/net/ethernet/intel/i40e/i40e_nvm.c b/drivers/net/ethernet/intel/i40e/i40e_nvm.c index e49acd2accd3..554e49d02683 100644 --- a/drivers/net/ethernet/intel/i40e/i40e_nvm.c +++ b/drivers/net/ethernet/intel/i40e/i40e_nvm.c | |||
@@ -821,13 +821,12 @@ static enum i40e_nvmupd_cmd i40e_nvmupd_validate_command(struct i40e_hw *hw, | |||
821 | int *errno) | 821 | int *errno) |
822 | { | 822 | { |
823 | enum i40e_nvmupd_cmd upd_cmd; | 823 | enum i40e_nvmupd_cmd upd_cmd; |
824 | u8 transaction, module; | 824 | u8 transaction; |
825 | 825 | ||
826 | /* anything that doesn't match a recognized case is an error */ | 826 | /* anything that doesn't match a recognized case is an error */ |
827 | upd_cmd = I40E_NVMUPD_INVALID; | 827 | upd_cmd = I40E_NVMUPD_INVALID; |
828 | 828 | ||
829 | transaction = i40e_nvmupd_get_transaction(cmd->config); | 829 | transaction = i40e_nvmupd_get_transaction(cmd->config); |
830 | module = i40e_nvmupd_get_module(cmd->config); | ||
831 | 830 | ||
832 | /* limits on data size */ | 831 | /* limits on data size */ |
833 | if ((cmd->data_size < 1) || | 832 | if ((cmd->data_size < 1) || |
diff --git a/drivers/net/ethernet/intel/i40e/i40e_prototype.h b/drivers/net/ethernet/intel/i40e/i40e_prototype.h index fea0d37ecc72..7b34f1e660ea 100644 --- a/drivers/net/ethernet/intel/i40e/i40e_prototype.h +++ b/drivers/net/ethernet/intel/i40e/i40e_prototype.h | |||
@@ -303,4 +303,9 @@ i40e_status i40e_aq_add_rem_control_packet_filter(struct i40e_hw *hw, | |||
303 | u16 vsi_seid, u16 queue, bool is_add, | 303 | u16 vsi_seid, u16 queue, bool is_add, |
304 | struct i40e_control_filter_stats *stats, | 304 | struct i40e_control_filter_stats *stats, |
305 | struct i40e_asq_cmd_details *cmd_details); | 305 | struct i40e_asq_cmd_details *cmd_details); |
306 | i40e_status i40e_aq_debug_dump(struct i40e_hw *hw, u8 cluster_id, | ||
307 | u8 table_id, u32 start_index, u16 buff_size, | ||
308 | void *buff, u16 *ret_buff_size, | ||
309 | u8 *ret_next_table, u32 *ret_next_index, | ||
310 | struct i40e_asq_cmd_details *cmd_details); | ||
306 | #endif /* _I40E_PROTOTYPE_H_ */ | 311 | #endif /* _I40E_PROTOTYPE_H_ */ |
diff --git a/drivers/net/ethernet/intel/i40e/i40e_type.h b/drivers/net/ethernet/intel/i40e/i40e_type.h index 67c7bc9e9c21..568e855da0f3 100644 --- a/drivers/net/ethernet/intel/i40e/i40e_type.h +++ b/drivers/net/ethernet/intel/i40e/i40e_type.h | |||
@@ -242,6 +242,7 @@ struct i40e_hw_capabilities { | |||
242 | u8 rx_buf_chain_len; | 242 | u8 rx_buf_chain_len; |
243 | u32 enabled_tcmap; | 243 | u32 enabled_tcmap; |
244 | u32 maxtc; | 244 | u32 maxtc; |
245 | u64 wr_csr_prot; | ||
245 | }; | 246 | }; |
246 | 247 | ||
247 | struct i40e_mac_info { | 248 | struct i40e_mac_info { |
diff --git a/drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c b/drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c index 4d69e1f04901..78d1c4ff565e 100644 --- a/drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c +++ b/drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c | |||
@@ -26,6 +26,129 @@ | |||
26 | 26 | ||
27 | #include "i40e.h" | 27 | #include "i40e.h" |
28 | 28 | ||
29 | /*********************notification routines***********************/ | ||
30 | |||
31 | /** | ||
32 | * i40e_vc_vf_broadcast | ||
33 | * @pf: pointer to the PF structure | ||
34 | * @opcode: operation code | ||
35 | * @retval: return value | ||
36 | * @msg: pointer to the msg buffer | ||
37 | * @msglen: msg length | ||
38 | * | ||
39 | * send a message to all VFs on a given PF | ||
40 | **/ | ||
41 | static void i40e_vc_vf_broadcast(struct i40e_pf *pf, | ||
42 | enum i40e_virtchnl_ops v_opcode, | ||
43 | i40e_status v_retval, u8 *msg, | ||
44 | u16 msglen) | ||
45 | { | ||
46 | struct i40e_hw *hw = &pf->hw; | ||
47 | struct i40e_vf *vf = pf->vf; | ||
48 | int i; | ||
49 | |||
50 | for (i = 0; i < pf->num_alloc_vfs; i++, vf++) { | ||
51 | int abs_vf_id = vf->vf_id + hw->func_caps.vf_base_id; | ||
52 | /* Not all vfs are enabled so skip the ones that are not */ | ||
53 | if (!test_bit(I40E_VF_STAT_INIT, &vf->vf_states) && | ||
54 | !test_bit(I40E_VF_STAT_ACTIVE, &vf->vf_states)) | ||
55 | continue; | ||
56 | |||
57 | /* Ignore return value on purpose - a given VF may fail, but | ||
58 | * we need to keep going and send to all of them | ||
59 | */ | ||
60 | i40e_aq_send_msg_to_vf(hw, abs_vf_id, v_opcode, v_retval, | ||
61 | msg, msglen, NULL); | ||
62 | } | ||
63 | } | ||
64 | |||
65 | /** | ||
66 | * i40e_vc_notify_link_state | ||
67 | * @vf: pointer to the VF structure | ||
68 | * | ||
69 | * send a link status message to a single VF | ||
70 | **/ | ||
71 | static void i40e_vc_notify_vf_link_state(struct i40e_vf *vf) | ||
72 | { | ||
73 | struct i40e_virtchnl_pf_event pfe; | ||
74 | struct i40e_pf *pf = vf->pf; | ||
75 | struct i40e_hw *hw = &pf->hw; | ||
76 | struct i40e_link_status *ls = &pf->hw.phy.link_info; | ||
77 | int abs_vf_id = vf->vf_id + hw->func_caps.vf_base_id; | ||
78 | |||
79 | pfe.event = I40E_VIRTCHNL_EVENT_LINK_CHANGE; | ||
80 | pfe.severity = I40E_PF_EVENT_SEVERITY_INFO; | ||
81 | if (vf->link_forced) { | ||
82 | pfe.event_data.link_event.link_status = vf->link_up; | ||
83 | pfe.event_data.link_event.link_speed = | ||
84 | (vf->link_up ? I40E_LINK_SPEED_40GB : 0); | ||
85 | } else { | ||
86 | pfe.event_data.link_event.link_status = | ||
87 | ls->link_info & I40E_AQ_LINK_UP; | ||
88 | pfe.event_data.link_event.link_speed = ls->link_speed; | ||
89 | } | ||
90 | i40e_aq_send_msg_to_vf(hw, abs_vf_id, I40E_VIRTCHNL_OP_EVENT, | ||
91 | 0, (u8 *)&pfe, sizeof(pfe), NULL); | ||
92 | } | ||
93 | |||
94 | /** | ||
95 | * i40e_vc_notify_link_state | ||
96 | * @pf: pointer to the PF structure | ||
97 | * | ||
98 | * send a link status message to all VFs on a given PF | ||
99 | **/ | ||
100 | void i40e_vc_notify_link_state(struct i40e_pf *pf) | ||
101 | { | ||
102 | int i; | ||
103 | |||
104 | for (i = 0; i < pf->num_alloc_vfs; i++) | ||
105 | i40e_vc_notify_vf_link_state(&pf->vf[i]); | ||
106 | } | ||
107 | |||
108 | /** | ||
109 | * i40e_vc_notify_reset | ||
110 | * @pf: pointer to the PF structure | ||
111 | * | ||
112 | * indicate a pending reset to all VFs on a given PF | ||
113 | **/ | ||
114 | void i40e_vc_notify_reset(struct i40e_pf *pf) | ||
115 | { | ||
116 | struct i40e_virtchnl_pf_event pfe; | ||
117 | |||
118 | pfe.event = I40E_VIRTCHNL_EVENT_RESET_IMPENDING; | ||
119 | pfe.severity = I40E_PF_EVENT_SEVERITY_CERTAIN_DOOM; | ||
120 | i40e_vc_vf_broadcast(pf, I40E_VIRTCHNL_OP_EVENT, 0, | ||
121 | (u8 *)&pfe, sizeof(struct i40e_virtchnl_pf_event)); | ||
122 | } | ||
123 | |||
124 | /** | ||
125 | * i40e_vc_notify_vf_reset | ||
126 | * @vf: pointer to the VF structure | ||
127 | * | ||
128 | * indicate a pending reset to the given VF | ||
129 | **/ | ||
130 | void i40e_vc_notify_vf_reset(struct i40e_vf *vf) | ||
131 | { | ||
132 | struct i40e_virtchnl_pf_event pfe; | ||
133 | int abs_vf_id; | ||
134 | |||
135 | /* validate the request */ | ||
136 | if (!vf || vf->vf_id >= vf->pf->num_alloc_vfs) | ||
137 | return; | ||
138 | |||
139 | /* verify if the VF is in either init or active before proceeding */ | ||
140 | if (!test_bit(I40E_VF_STAT_INIT, &vf->vf_states) && | ||
141 | !test_bit(I40E_VF_STAT_ACTIVE, &vf->vf_states)) | ||
142 | return; | ||
143 | |||
144 | abs_vf_id = vf->vf_id + vf->pf->hw.func_caps.vf_base_id; | ||
145 | |||
146 | pfe.event = I40E_VIRTCHNL_EVENT_RESET_IMPENDING; | ||
147 | pfe.severity = I40E_PF_EVENT_SEVERITY_CERTAIN_DOOM; | ||
148 | i40e_aq_send_msg_to_vf(&vf->pf->hw, abs_vf_id, I40E_VIRTCHNL_OP_EVENT, | ||
149 | 0, (u8 *)&pfe, | ||
150 | sizeof(struct i40e_virtchnl_pf_event), NULL); | ||
151 | } | ||
29 | /***********************misc routines*****************************/ | 152 | /***********************misc routines*****************************/ |
30 | 153 | ||
31 | /** | 154 | /** |
@@ -689,6 +812,9 @@ void i40e_reset_vf(struct i40e_vf *vf, bool flr) | |||
689 | } | 812 | } |
690 | } | 813 | } |
691 | 814 | ||
815 | if (flr) | ||
816 | usleep_range(10000, 20000); | ||
817 | |||
692 | if (!rsd) | 818 | if (!rsd) |
693 | dev_err(&pf->pdev->dev, "VF reset check timeout on VF %d\n", | 819 | dev_err(&pf->pdev->dev, "VF reset check timeout on VF %d\n", |
694 | vf->vf_id); | 820 | vf->vf_id); |
@@ -733,6 +859,11 @@ void i40e_free_vfs(struct i40e_pf *pf) | |||
733 | while (test_and_set_bit(__I40E_VF_DISABLE, &pf->state)) | 859 | while (test_and_set_bit(__I40E_VF_DISABLE, &pf->state)) |
734 | usleep_range(1000, 2000); | 860 | usleep_range(1000, 2000); |
735 | 861 | ||
862 | for (i = 0; i < pf->num_alloc_vfs; i++) | ||
863 | if (test_bit(I40E_VF_STAT_INIT, &pf->vf[i].vf_states)) | ||
864 | i40e_vsi_control_rings(pf->vsi[pf->vf[i].lan_vsi_idx], | ||
865 | false); | ||
866 | |||
736 | /* Disable IOV before freeing resources. This lets any VF drivers | 867 | /* Disable IOV before freeing resources. This lets any VF drivers |
737 | * running in the host get themselves cleaned up before we yank | 868 | * running in the host get themselves cleaned up before we yank |
738 | * the carpet out from underneath their feet. | 869 | * the carpet out from underneath their feet. |
@@ -1762,6 +1893,7 @@ int i40e_vc_process_vf_msg(struct i40e_pf *pf, u16 vf_id, u32 v_opcode, | |||
1762 | break; | 1893 | break; |
1763 | case I40E_VIRTCHNL_OP_ENABLE_QUEUES: | 1894 | case I40E_VIRTCHNL_OP_ENABLE_QUEUES: |
1764 | ret = i40e_vc_enable_queues_msg(vf, msg, msglen); | 1895 | ret = i40e_vc_enable_queues_msg(vf, msg, msglen); |
1896 | i40e_vc_notify_vf_link_state(vf); | ||
1765 | break; | 1897 | break; |
1766 | case I40E_VIRTCHNL_OP_DISABLE_QUEUES: | 1898 | case I40E_VIRTCHNL_OP_DISABLE_QUEUES: |
1767 | ret = i40e_vc_disable_queues_msg(vf, msg, msglen); | 1899 | ret = i40e_vc_disable_queues_msg(vf, msg, msglen); |
@@ -1835,118 +1967,6 @@ int i40e_vc_process_vflr_event(struct i40e_pf *pf) | |||
1835 | } | 1967 | } |
1836 | 1968 | ||
1837 | /** | 1969 | /** |
1838 | * i40e_vc_vf_broadcast | ||
1839 | * @pf: pointer to the PF structure | ||
1840 | * @opcode: operation code | ||
1841 | * @retval: return value | ||
1842 | * @msg: pointer to the msg buffer | ||
1843 | * @msglen: msg length | ||
1844 | * | ||
1845 | * send a message to all VFs on a given PF | ||
1846 | **/ | ||
1847 | static void i40e_vc_vf_broadcast(struct i40e_pf *pf, | ||
1848 | enum i40e_virtchnl_ops v_opcode, | ||
1849 | i40e_status v_retval, u8 *msg, | ||
1850 | u16 msglen) | ||
1851 | { | ||
1852 | struct i40e_hw *hw = &pf->hw; | ||
1853 | struct i40e_vf *vf = pf->vf; | ||
1854 | int i; | ||
1855 | |||
1856 | for (i = 0; i < pf->num_alloc_vfs; i++, vf++) { | ||
1857 | int abs_vf_id = vf->vf_id + hw->func_caps.vf_base_id; | ||
1858 | /* Not all VFs are enabled so skip the ones that are not */ | ||
1859 | if (!test_bit(I40E_VF_STAT_INIT, &vf->vf_states) && | ||
1860 | !test_bit(I40E_VF_STAT_ACTIVE, &vf->vf_states)) | ||
1861 | continue; | ||
1862 | |||
1863 | /* Ignore return value on purpose - a given VF may fail, but | ||
1864 | * we need to keep going and send to all of them | ||
1865 | */ | ||
1866 | i40e_aq_send_msg_to_vf(hw, abs_vf_id, v_opcode, v_retval, | ||
1867 | msg, msglen, NULL); | ||
1868 | } | ||
1869 | } | ||
1870 | |||
1871 | /** | ||
1872 | * i40e_vc_notify_link_state | ||
1873 | * @pf: pointer to the PF structure | ||
1874 | * | ||
1875 | * send a link status message to all VFs on a given PF | ||
1876 | **/ | ||
1877 | void i40e_vc_notify_link_state(struct i40e_pf *pf) | ||
1878 | { | ||
1879 | struct i40e_virtchnl_pf_event pfe; | ||
1880 | struct i40e_hw *hw = &pf->hw; | ||
1881 | struct i40e_vf *vf = pf->vf; | ||
1882 | struct i40e_link_status *ls = &pf->hw.phy.link_info; | ||
1883 | int i; | ||
1884 | |||
1885 | pfe.event = I40E_VIRTCHNL_EVENT_LINK_CHANGE; | ||
1886 | pfe.severity = I40E_PF_EVENT_SEVERITY_INFO; | ||
1887 | for (i = 0; i < pf->num_alloc_vfs; i++, vf++) { | ||
1888 | int abs_vf_id = vf->vf_id + hw->func_caps.vf_base_id; | ||
1889 | if (vf->link_forced) { | ||
1890 | pfe.event_data.link_event.link_status = vf->link_up; | ||
1891 | pfe.event_data.link_event.link_speed = | ||
1892 | (vf->link_up ? I40E_LINK_SPEED_40GB : 0); | ||
1893 | } else { | ||
1894 | pfe.event_data.link_event.link_status = | ||
1895 | ls->link_info & I40E_AQ_LINK_UP; | ||
1896 | pfe.event_data.link_event.link_speed = ls->link_speed; | ||
1897 | } | ||
1898 | i40e_aq_send_msg_to_vf(hw, abs_vf_id, I40E_VIRTCHNL_OP_EVENT, | ||
1899 | 0, (u8 *)&pfe, sizeof(pfe), | ||
1900 | NULL); | ||
1901 | } | ||
1902 | } | ||
1903 | |||
1904 | /** | ||
1905 | * i40e_vc_notify_reset | ||
1906 | * @pf: pointer to the PF structure | ||
1907 | * | ||
1908 | * indicate a pending reset to all VFs on a given PF | ||
1909 | **/ | ||
1910 | void i40e_vc_notify_reset(struct i40e_pf *pf) | ||
1911 | { | ||
1912 | struct i40e_virtchnl_pf_event pfe; | ||
1913 | |||
1914 | pfe.event = I40E_VIRTCHNL_EVENT_RESET_IMPENDING; | ||
1915 | pfe.severity = I40E_PF_EVENT_SEVERITY_CERTAIN_DOOM; | ||
1916 | i40e_vc_vf_broadcast(pf, I40E_VIRTCHNL_OP_EVENT, I40E_SUCCESS, | ||
1917 | (u8 *)&pfe, sizeof(struct i40e_virtchnl_pf_event)); | ||
1918 | } | ||
1919 | |||
1920 | /** | ||
1921 | * i40e_vc_notify_vf_reset | ||
1922 | * @vf: pointer to the VF structure | ||
1923 | * | ||
1924 | * indicate a pending reset to the given VF | ||
1925 | **/ | ||
1926 | void i40e_vc_notify_vf_reset(struct i40e_vf *vf) | ||
1927 | { | ||
1928 | struct i40e_virtchnl_pf_event pfe; | ||
1929 | int abs_vf_id; | ||
1930 | |||
1931 | /* validate the request */ | ||
1932 | if (!vf || vf->vf_id >= vf->pf->num_alloc_vfs) | ||
1933 | return; | ||
1934 | |||
1935 | /* verify if the VF is in either init or active before proceeding */ | ||
1936 | if (!test_bit(I40E_VF_STAT_INIT, &vf->vf_states) && | ||
1937 | !test_bit(I40E_VF_STAT_ACTIVE, &vf->vf_states)) | ||
1938 | return; | ||
1939 | |||
1940 | abs_vf_id = vf->vf_id + vf->pf->hw.func_caps.vf_base_id; | ||
1941 | |||
1942 | pfe.event = I40E_VIRTCHNL_EVENT_RESET_IMPENDING; | ||
1943 | pfe.severity = I40E_PF_EVENT_SEVERITY_CERTAIN_DOOM; | ||
1944 | i40e_aq_send_msg_to_vf(&vf->pf->hw, abs_vf_id, I40E_VIRTCHNL_OP_EVENT, | ||
1945 | I40E_SUCCESS, (u8 *)&pfe, | ||
1946 | sizeof(struct i40e_virtchnl_pf_event), NULL); | ||
1947 | } | ||
1948 | |||
1949 | /** | ||
1950 | * i40e_ndo_set_vf_mac | 1970 | * i40e_ndo_set_vf_mac |
1951 | * @netdev: network interface device structure | 1971 | * @netdev: network interface device structure |
1952 | * @vf_id: VF identifier | 1972 | * @vf_id: VF identifier |
diff --git a/drivers/net/ethernet/intel/i40evf/i40e_type.h b/drivers/net/ethernet/intel/i40evf/i40e_type.h index 9c79cb6abb2b..ec9d83a93379 100644 --- a/drivers/net/ethernet/intel/i40evf/i40e_type.h +++ b/drivers/net/ethernet/intel/i40evf/i40e_type.h | |||
@@ -242,6 +242,7 @@ struct i40e_hw_capabilities { | |||
242 | u8 rx_buf_chain_len; | 242 | u8 rx_buf_chain_len; |
243 | u32 enabled_tcmap; | 243 | u32 enabled_tcmap; |
244 | u32 maxtc; | 244 | u32 maxtc; |
245 | u64 wr_csr_prot; | ||
245 | }; | 246 | }; |
246 | 247 | ||
247 | struct i40e_mac_info { | 248 | struct i40e_mac_info { |
diff --git a/drivers/net/ethernet/intel/i40evf/i40evf.h b/drivers/net/ethernet/intel/i40evf/i40evf.h index 34c8565031f6..1b98c25b3092 100644 --- a/drivers/net/ethernet/intel/i40evf/i40evf.h +++ b/drivers/net/ethernet/intel/i40evf/i40evf.h | |||
@@ -225,7 +225,6 @@ struct i40evf_adapter { | |||
225 | #define I40E_FLAG_RX_CSUM_ENABLED I40EVF_FLAG_RX_CSUM_ENABLED | 225 | #define I40E_FLAG_RX_CSUM_ENABLED I40EVF_FLAG_RX_CSUM_ENABLED |
226 | /* flags for admin queue service task */ | 226 | /* flags for admin queue service task */ |
227 | u32 aq_required; | 227 | u32 aq_required; |
228 | u32 aq_pending; | ||
229 | #define I40EVF_FLAG_AQ_ENABLE_QUEUES (u32)(1) | 228 | #define I40EVF_FLAG_AQ_ENABLE_QUEUES (u32)(1) |
230 | #define I40EVF_FLAG_AQ_DISABLE_QUEUES (u32)(1 << 1) | 229 | #define I40EVF_FLAG_AQ_DISABLE_QUEUES (u32)(1 << 1) |
231 | #define I40EVF_FLAG_AQ_ADD_MAC_FILTER (u32)(1 << 2) | 230 | #define I40EVF_FLAG_AQ_ADD_MAC_FILTER (u32)(1 << 2) |
diff --git a/drivers/net/ethernet/intel/i40evf/i40evf_main.c b/drivers/net/ethernet/intel/i40evf/i40evf_main.c index 6d5f3b21c68a..7c53aca4b5a6 100644 --- a/drivers/net/ethernet/intel/i40evf/i40evf_main.c +++ b/drivers/net/ethernet/intel/i40evf/i40evf_main.c | |||
@@ -1008,7 +1008,6 @@ void i40evf_down(struct i40evf_adapter *adapter) | |||
1008 | adapter->state != __I40EVF_RESETTING) { | 1008 | adapter->state != __I40EVF_RESETTING) { |
1009 | /* cancel any current operation */ | 1009 | /* cancel any current operation */ |
1010 | adapter->current_op = I40E_VIRTCHNL_OP_UNKNOWN; | 1010 | adapter->current_op = I40E_VIRTCHNL_OP_UNKNOWN; |
1011 | adapter->aq_pending = 0; | ||
1012 | /* Schedule operations to close down the HW. Don't wait | 1011 | /* Schedule operations to close down the HW. Don't wait |
1013 | * here for this to complete. The watchdog is still running | 1012 | * here for this to complete. The watchdog is still running |
1014 | * and it will take care of this. | 1013 | * and it will take care of this. |
@@ -1335,7 +1334,6 @@ static void i40evf_watchdog_task(struct work_struct *work) | |||
1335 | */ | 1334 | */ |
1336 | return; | 1335 | return; |
1337 | } | 1336 | } |
1338 | adapter->aq_pending = 0; | ||
1339 | adapter->aq_required = 0; | 1337 | adapter->aq_required = 0; |
1340 | adapter->current_op = I40E_VIRTCHNL_OP_UNKNOWN; | 1338 | adapter->current_op = I40E_VIRTCHNL_OP_UNKNOWN; |
1341 | goto watchdog_done; | 1339 | goto watchdog_done; |
@@ -1355,7 +1353,6 @@ static void i40evf_watchdog_task(struct work_struct *work) | |||
1355 | adapter->flags |= I40EVF_FLAG_RESET_PENDING; | 1353 | adapter->flags |= I40EVF_FLAG_RESET_PENDING; |
1356 | dev_err(&adapter->pdev->dev, "Hardware reset detected\n"); | 1354 | dev_err(&adapter->pdev->dev, "Hardware reset detected\n"); |
1357 | schedule_work(&adapter->reset_task); | 1355 | schedule_work(&adapter->reset_task); |
1358 | adapter->aq_pending = 0; | ||
1359 | adapter->aq_required = 0; | 1356 | adapter->aq_required = 0; |
1360 | adapter->current_op = I40E_VIRTCHNL_OP_UNKNOWN; | 1357 | adapter->current_op = I40E_VIRTCHNL_OP_UNKNOWN; |
1361 | goto watchdog_done; | 1358 | goto watchdog_done; |
@@ -1364,7 +1361,7 @@ static void i40evf_watchdog_task(struct work_struct *work) | |||
1364 | /* Process admin queue tasks. After init, everything gets done | 1361 | /* Process admin queue tasks. After init, everything gets done |
1365 | * here so we don't race on the admin queue. | 1362 | * here so we don't race on the admin queue. |
1366 | */ | 1363 | */ |
1367 | if (adapter->aq_pending) { | 1364 | if (adapter->current_op) { |
1368 | if (!i40evf_asq_done(hw)) { | 1365 | if (!i40evf_asq_done(hw)) { |
1369 | dev_dbg(&adapter->pdev->dev, "Admin queue timeout\n"); | 1366 | dev_dbg(&adapter->pdev->dev, "Admin queue timeout\n"); |
1370 | i40evf_send_api_ver(adapter); | 1367 | i40evf_send_api_ver(adapter); |
@@ -2029,7 +2026,7 @@ static void i40evf_init_task(struct work_struct *work) | |||
2029 | if (err) { | 2026 | if (err) { |
2030 | dev_err(&pdev->dev, "Failed to set MAC type (%d)\n", | 2027 | dev_err(&pdev->dev, "Failed to set MAC type (%d)\n", |
2031 | err); | 2028 | err); |
2032 | goto err; | 2029 | goto err; |
2033 | } | 2030 | } |
2034 | err = i40evf_check_reset_complete(hw); | 2031 | err = i40evf_check_reset_complete(hw); |
2035 | if (err) { | 2032 | if (err) { |
@@ -2249,7 +2246,6 @@ static void i40evf_shutdown(struct pci_dev *pdev) | |||
2249 | /* Prevent the watchdog from running. */ | 2246 | /* Prevent the watchdog from running. */ |
2250 | adapter->state = __I40EVF_REMOVE; | 2247 | adapter->state = __I40EVF_REMOVE; |
2251 | adapter->aq_required = 0; | 2248 | adapter->aq_required = 0; |
2252 | adapter->aq_pending = 0; | ||
2253 | 2249 | ||
2254 | #ifdef CONFIG_PM | 2250 | #ifdef CONFIG_PM |
2255 | pci_save_state(pdev); | 2251 | pci_save_state(pdev); |
@@ -2467,7 +2463,6 @@ static void i40evf_remove(struct pci_dev *pdev) | |||
2467 | /* Shut down all the garbage mashers on the detention level */ | 2463 | /* Shut down all the garbage mashers on the detention level */ |
2468 | adapter->state = __I40EVF_REMOVE; | 2464 | adapter->state = __I40EVF_REMOVE; |
2469 | adapter->aq_required = 0; | 2465 | adapter->aq_required = 0; |
2470 | adapter->aq_pending = 0; | ||
2471 | i40evf_request_reset(adapter); | 2466 | i40evf_request_reset(adapter); |
2472 | msleep(20); | 2467 | msleep(20); |
2473 | /* If the FW isn't responding, kick it once, but only once. */ | 2468 | /* If the FW isn't responding, kick it once, but only once. */ |
diff --git a/drivers/net/ethernet/intel/i40evf/i40evf_virtchnl.c b/drivers/net/ethernet/intel/i40evf/i40evf_virtchnl.c index 4240a496dc50..61e090558f31 100644 --- a/drivers/net/ethernet/intel/i40evf/i40evf_virtchnl.c +++ b/drivers/net/ethernet/intel/i40evf/i40evf_virtchnl.c | |||
@@ -250,7 +250,6 @@ void i40evf_configure_queues(struct i40evf_adapter *adapter) | |||
250 | vqpi++; | 250 | vqpi++; |
251 | } | 251 | } |
252 | 252 | ||
253 | adapter->aq_pending |= I40EVF_FLAG_AQ_CONFIGURE_QUEUES; | ||
254 | adapter->aq_required &= ~I40EVF_FLAG_AQ_CONFIGURE_QUEUES; | 253 | adapter->aq_required &= ~I40EVF_FLAG_AQ_CONFIGURE_QUEUES; |
255 | i40evf_send_pf_msg(adapter, I40E_VIRTCHNL_OP_CONFIG_VSI_QUEUES, | 254 | i40evf_send_pf_msg(adapter, I40E_VIRTCHNL_OP_CONFIG_VSI_QUEUES, |
256 | (u8 *)vqci, len); | 255 | (u8 *)vqci, len); |
@@ -277,7 +276,6 @@ void i40evf_enable_queues(struct i40evf_adapter *adapter) | |||
277 | vqs.vsi_id = adapter->vsi_res->vsi_id; | 276 | vqs.vsi_id = adapter->vsi_res->vsi_id; |
278 | vqs.tx_queues = (1 << adapter->num_active_queues) - 1; | 277 | vqs.tx_queues = (1 << adapter->num_active_queues) - 1; |
279 | vqs.rx_queues = vqs.tx_queues; | 278 | vqs.rx_queues = vqs.tx_queues; |
280 | adapter->aq_pending |= I40EVF_FLAG_AQ_ENABLE_QUEUES; | ||
281 | adapter->aq_required &= ~I40EVF_FLAG_AQ_ENABLE_QUEUES; | 279 | adapter->aq_required &= ~I40EVF_FLAG_AQ_ENABLE_QUEUES; |
282 | i40evf_send_pf_msg(adapter, I40E_VIRTCHNL_OP_ENABLE_QUEUES, | 280 | i40evf_send_pf_msg(adapter, I40E_VIRTCHNL_OP_ENABLE_QUEUES, |
283 | (u8 *)&vqs, sizeof(vqs)); | 281 | (u8 *)&vqs, sizeof(vqs)); |
@@ -303,7 +301,6 @@ void i40evf_disable_queues(struct i40evf_adapter *adapter) | |||
303 | vqs.vsi_id = adapter->vsi_res->vsi_id; | 301 | vqs.vsi_id = adapter->vsi_res->vsi_id; |
304 | vqs.tx_queues = (1 << adapter->num_active_queues) - 1; | 302 | vqs.tx_queues = (1 << adapter->num_active_queues) - 1; |
305 | vqs.rx_queues = vqs.tx_queues; | 303 | vqs.rx_queues = vqs.tx_queues; |
306 | adapter->aq_pending |= I40EVF_FLAG_AQ_DISABLE_QUEUES; | ||
307 | adapter->aq_required &= ~I40EVF_FLAG_AQ_DISABLE_QUEUES; | 304 | adapter->aq_required &= ~I40EVF_FLAG_AQ_DISABLE_QUEUES; |
308 | i40evf_send_pf_msg(adapter, I40E_VIRTCHNL_OP_DISABLE_QUEUES, | 305 | i40evf_send_pf_msg(adapter, I40E_VIRTCHNL_OP_DISABLE_QUEUES, |
309 | (u8 *)&vqs, sizeof(vqs)); | 306 | (u8 *)&vqs, sizeof(vqs)); |
@@ -354,7 +351,6 @@ void i40evf_map_queues(struct i40evf_adapter *adapter) | |||
354 | vimi->vecmap[v_idx].txq_map = 0; | 351 | vimi->vecmap[v_idx].txq_map = 0; |
355 | vimi->vecmap[v_idx].rxq_map = 0; | 352 | vimi->vecmap[v_idx].rxq_map = 0; |
356 | 353 | ||
357 | adapter->aq_pending |= I40EVF_FLAG_AQ_MAP_VECTORS; | ||
358 | adapter->aq_required &= ~I40EVF_FLAG_AQ_MAP_VECTORS; | 354 | adapter->aq_required &= ~I40EVF_FLAG_AQ_MAP_VECTORS; |
359 | i40evf_send_pf_msg(adapter, I40E_VIRTCHNL_OP_CONFIG_IRQ_MAP, | 355 | i40evf_send_pf_msg(adapter, I40E_VIRTCHNL_OP_CONFIG_IRQ_MAP, |
360 | (u8 *)vimi, len); | 356 | (u8 *)vimi, len); |
@@ -415,7 +411,6 @@ void i40evf_add_ether_addrs(struct i40evf_adapter *adapter) | |||
415 | f->add = false; | 411 | f->add = false; |
416 | } | 412 | } |
417 | } | 413 | } |
418 | adapter->aq_pending |= I40EVF_FLAG_AQ_ADD_MAC_FILTER; | ||
419 | adapter->aq_required &= ~I40EVF_FLAG_AQ_ADD_MAC_FILTER; | 414 | adapter->aq_required &= ~I40EVF_FLAG_AQ_ADD_MAC_FILTER; |
420 | i40evf_send_pf_msg(adapter, I40E_VIRTCHNL_OP_ADD_ETHER_ADDRESS, | 415 | i40evf_send_pf_msg(adapter, I40E_VIRTCHNL_OP_ADD_ETHER_ADDRESS, |
421 | (u8 *)veal, len); | 416 | (u8 *)veal, len); |
@@ -476,7 +471,6 @@ void i40evf_del_ether_addrs(struct i40evf_adapter *adapter) | |||
476 | kfree(f); | 471 | kfree(f); |
477 | } | 472 | } |
478 | } | 473 | } |
479 | adapter->aq_pending |= I40EVF_FLAG_AQ_DEL_MAC_FILTER; | ||
480 | adapter->aq_required &= ~I40EVF_FLAG_AQ_DEL_MAC_FILTER; | 474 | adapter->aq_required &= ~I40EVF_FLAG_AQ_DEL_MAC_FILTER; |
481 | i40evf_send_pf_msg(adapter, I40E_VIRTCHNL_OP_DEL_ETHER_ADDRESS, | 475 | i40evf_send_pf_msg(adapter, I40E_VIRTCHNL_OP_DEL_ETHER_ADDRESS, |
482 | (u8 *)veal, len); | 476 | (u8 *)veal, len); |
@@ -537,7 +531,6 @@ void i40evf_add_vlans(struct i40evf_adapter *adapter) | |||
537 | f->add = false; | 531 | f->add = false; |
538 | } | 532 | } |
539 | } | 533 | } |
540 | adapter->aq_pending |= I40EVF_FLAG_AQ_ADD_VLAN_FILTER; | ||
541 | adapter->aq_required &= ~I40EVF_FLAG_AQ_ADD_VLAN_FILTER; | 534 | adapter->aq_required &= ~I40EVF_FLAG_AQ_ADD_VLAN_FILTER; |
542 | i40evf_send_pf_msg(adapter, I40E_VIRTCHNL_OP_ADD_VLAN, (u8 *)vvfl, len); | 535 | i40evf_send_pf_msg(adapter, I40E_VIRTCHNL_OP_ADD_VLAN, (u8 *)vvfl, len); |
543 | kfree(vvfl); | 536 | kfree(vvfl); |
@@ -598,7 +591,6 @@ void i40evf_del_vlans(struct i40evf_adapter *adapter) | |||
598 | kfree(f); | 591 | kfree(f); |
599 | } | 592 | } |
600 | } | 593 | } |
601 | adapter->aq_pending |= I40EVF_FLAG_AQ_DEL_VLAN_FILTER; | ||
602 | adapter->aq_required &= ~I40EVF_FLAG_AQ_DEL_VLAN_FILTER; | 594 | adapter->aq_required &= ~I40EVF_FLAG_AQ_DEL_VLAN_FILTER; |
603 | i40evf_send_pf_msg(adapter, I40E_VIRTCHNL_OP_DEL_VLAN, (u8 *)vvfl, len); | 595 | i40evf_send_pf_msg(adapter, I40E_VIRTCHNL_OP_DEL_VLAN, (u8 *)vvfl, len); |
604 | kfree(vvfl); | 596 | kfree(vvfl); |
@@ -720,9 +712,6 @@ void i40evf_virtchnl_completion(struct i40evf_adapter *adapter, | |||
720 | __func__, v_retval, v_opcode); | 712 | __func__, v_retval, v_opcode); |
721 | } | 713 | } |
722 | switch (v_opcode) { | 714 | switch (v_opcode) { |
723 | case I40E_VIRTCHNL_OP_VERSION: | ||
724 | /* no action, but also not an error */ | ||
725 | break; | ||
726 | case I40E_VIRTCHNL_OP_GET_STATS: { | 715 | case I40E_VIRTCHNL_OP_GET_STATS: { |
727 | struct i40e_eth_stats *stats = | 716 | struct i40e_eth_stats *stats = |
728 | (struct i40e_eth_stats *)msg; | 717 | (struct i40e_eth_stats *)msg; |
@@ -740,39 +729,30 @@ void i40evf_virtchnl_completion(struct i40evf_adapter *adapter, | |||
740 | adapter->current_stats = *stats; | 729 | adapter->current_stats = *stats; |
741 | } | 730 | } |
742 | break; | 731 | break; |
743 | case I40E_VIRTCHNL_OP_ADD_ETHER_ADDRESS: | ||
744 | adapter->aq_pending &= ~(I40EVF_FLAG_AQ_ADD_MAC_FILTER); | ||
745 | break; | ||
746 | case I40E_VIRTCHNL_OP_DEL_ETHER_ADDRESS: | ||
747 | adapter->aq_pending &= ~(I40EVF_FLAG_AQ_DEL_MAC_FILTER); | ||
748 | break; | ||
749 | case I40E_VIRTCHNL_OP_ADD_VLAN: | ||
750 | adapter->aq_pending &= ~(I40EVF_FLAG_AQ_ADD_VLAN_FILTER); | ||
751 | break; | ||
752 | case I40E_VIRTCHNL_OP_DEL_VLAN: | ||
753 | adapter->aq_pending &= ~(I40EVF_FLAG_AQ_DEL_VLAN_FILTER); | ||
754 | break; | ||
755 | case I40E_VIRTCHNL_OP_ENABLE_QUEUES: | 732 | case I40E_VIRTCHNL_OP_ENABLE_QUEUES: |
756 | adapter->aq_pending &= ~(I40EVF_FLAG_AQ_ENABLE_QUEUES); | ||
757 | /* enable transmits */ | 733 | /* enable transmits */ |
758 | i40evf_irq_enable(adapter, true); | 734 | i40evf_irq_enable(adapter, true); |
759 | netif_tx_start_all_queues(adapter->netdev); | 735 | netif_tx_start_all_queues(adapter->netdev); |
760 | netif_carrier_on(adapter->netdev); | 736 | netif_carrier_on(adapter->netdev); |
761 | break; | 737 | break; |
762 | case I40E_VIRTCHNL_OP_DISABLE_QUEUES: | 738 | case I40E_VIRTCHNL_OP_DISABLE_QUEUES: |
763 | adapter->aq_pending &= ~(I40EVF_FLAG_AQ_DISABLE_QUEUES); | ||
764 | i40evf_free_all_tx_resources(adapter); | 739 | i40evf_free_all_tx_resources(adapter); |
765 | i40evf_free_all_rx_resources(adapter); | 740 | i40evf_free_all_rx_resources(adapter); |
766 | break; | 741 | break; |
767 | case I40E_VIRTCHNL_OP_CONFIG_VSI_QUEUES: | 742 | case I40E_VIRTCHNL_OP_VERSION: |
768 | adapter->aq_pending &= ~(I40EVF_FLAG_AQ_CONFIGURE_QUEUES); | 743 | case I40E_VIRTCHNL_OP_GET_VF_RESOURCES: |
769 | break; | ||
770 | case I40E_VIRTCHNL_OP_CONFIG_IRQ_MAP: | 744 | case I40E_VIRTCHNL_OP_CONFIG_IRQ_MAP: |
771 | adapter->aq_pending &= ~(I40EVF_FLAG_AQ_MAP_VECTORS); | 745 | /* Don't display an error if we get these out of sequence. |
746 | * If the firmware needed to get kicked, we'll get these and | ||
747 | * it's no problem. | ||
748 | */ | ||
749 | if (v_opcode != adapter->current_op) | ||
750 | return; | ||
772 | break; | 751 | break; |
773 | default: | 752 | default: |
774 | dev_info(&adapter->pdev->dev, "Received unexpected message %d from PF\n", | 753 | if (v_opcode != adapter->current_op) |
775 | v_opcode); | 754 | dev_warn(&adapter->pdev->dev, "Expected response %d from PF, received %d\n", |
755 | adapter->current_op, v_opcode); | ||
776 | break; | 756 | break; |
777 | } /* switch v_opcode */ | 757 | } /* switch v_opcode */ |
778 | adapter->current_op = I40E_VIRTCHNL_OP_UNKNOWN; | 758 | adapter->current_op = I40E_VIRTCHNL_OP_UNKNOWN; |
diff --git a/drivers/net/ethernet/qualcomm/qca_spi.c b/drivers/net/ethernet/qualcomm/qca_spi.c index 4a42e960d331..f66641d961e3 100644 --- a/drivers/net/ethernet/qualcomm/qca_spi.c +++ b/drivers/net/ethernet/qualcomm/qca_spi.c | |||
@@ -41,7 +41,6 @@ | |||
41 | #include <linux/skbuff.h> | 41 | #include <linux/skbuff.h> |
42 | #include <linux/spi/spi.h> | 42 | #include <linux/spi/spi.h> |
43 | #include <linux/types.h> | 43 | #include <linux/types.h> |
44 | #include <linux/version.h> | ||
45 | 44 | ||
46 | #include "qca_7k.h" | 45 | #include "qca_7k.h" |
47 | #include "qca_debug.h" | 46 | #include "qca_debug.h" |
diff --git a/drivers/net/ethernet/rocker/rocker.c b/drivers/net/ethernet/rocker/rocker.c index a87b177bd723..a570a60533be 100644 --- a/drivers/net/ethernet/rocker/rocker.c +++ b/drivers/net/ethernet/rocker/rocker.c | |||
@@ -4759,6 +4759,7 @@ static int rocker_probe(struct pci_dev *pdev, const struct pci_device_id *id) | |||
4759 | 4759 | ||
4760 | if (pci_resource_len(pdev, 0) < ROCKER_PCI_BAR0_SIZE) { | 4760 | if (pci_resource_len(pdev, 0) < ROCKER_PCI_BAR0_SIZE) { |
4761 | dev_err(&pdev->dev, "invalid PCI region size\n"); | 4761 | dev_err(&pdev->dev, "invalid PCI region size\n"); |
4762 | err = -EINVAL; | ||
4762 | goto err_pci_resource_len_check; | 4763 | goto err_pci_resource_len_check; |
4763 | } | 4764 | } |
4764 | 4765 | ||
diff --git a/drivers/net/ethernet/sfc/selftest.c b/drivers/net/ethernet/sfc/selftest.c index 10b6173d557d..b605dfd5c7bc 100644 --- a/drivers/net/ethernet/sfc/selftest.c +++ b/drivers/net/ethernet/sfc/selftest.c | |||
@@ -46,7 +46,7 @@ struct efx_loopback_payload { | |||
46 | struct iphdr ip; | 46 | struct iphdr ip; |
47 | struct udphdr udp; | 47 | struct udphdr udp; |
48 | __be16 iteration; | 48 | __be16 iteration; |
49 | const char msg[64]; | 49 | char msg[64]; |
50 | } __packed; | 50 | } __packed; |
51 | 51 | ||
52 | /* Loopback test source MAC address */ | 52 | /* Loopback test source MAC address */ |
diff --git a/drivers/net/ethernet/stmicro/stmmac/common.h b/drivers/net/ethernet/stmicro/stmmac/common.h index cd77289c3cfe..623c6ed8764a 100644 --- a/drivers/net/ethernet/stmicro/stmmac/common.h +++ b/drivers/net/ethernet/stmicro/stmmac/common.h | |||
@@ -150,7 +150,7 @@ struct stmmac_extra_stats { | |||
150 | #define MAC_CSR_H_FRQ_MASK 0x20 | 150 | #define MAC_CSR_H_FRQ_MASK 0x20 |
151 | 151 | ||
152 | #define HASH_TABLE_SIZE 64 | 152 | #define HASH_TABLE_SIZE 64 |
153 | #define PAUSE_TIME 0x200 | 153 | #define PAUSE_TIME 0xffff |
154 | 154 | ||
155 | /* Flow Control defines */ | 155 | /* Flow Control defines */ |
156 | #define FLOW_OFF 0 | 156 | #define FLOW_OFF 0 |
@@ -357,7 +357,8 @@ struct stmmac_dma_ops { | |||
357 | void (*dump_regs) (void __iomem *ioaddr); | 357 | void (*dump_regs) (void __iomem *ioaddr); |
358 | /* Set tx/rx threshold in the csr6 register | 358 | /* Set tx/rx threshold in the csr6 register |
359 | * An invalid value enables the store-and-forward mode */ | 359 | * An invalid value enables the store-and-forward mode */ |
360 | void (*dma_mode) (void __iomem *ioaddr, int txmode, int rxmode); | 360 | void (*dma_mode)(void __iomem *ioaddr, int txmode, int rxmode, |
361 | int rxfifosz); | ||
361 | /* To track extra statistic (if supported) */ | 362 | /* To track extra statistic (if supported) */ |
362 | void (*dma_diagnostic_fr) (void *data, struct stmmac_extra_stats *x, | 363 | void (*dma_diagnostic_fr) (void *data, struct stmmac_extra_stats *x, |
363 | void __iomem *ioaddr); | 364 | void __iomem *ioaddr); |
diff --git a/drivers/net/ethernet/stmicro/stmmac/dwmac1000.h b/drivers/net/ethernet/stmicro/stmmac/dwmac1000.h index 64d8f56a9c17..b3fe0575ff6b 100644 --- a/drivers/net/ethernet/stmicro/stmmac/dwmac1000.h +++ b/drivers/net/ethernet/stmicro/stmmac/dwmac1000.h | |||
@@ -172,6 +172,7 @@ enum inter_frame_gap { | |||
172 | /* GMAC FLOW CTRL defines */ | 172 | /* GMAC FLOW CTRL defines */ |
173 | #define GMAC_FLOW_CTRL_PT_MASK 0xffff0000 /* Pause Time Mask */ | 173 | #define GMAC_FLOW_CTRL_PT_MASK 0xffff0000 /* Pause Time Mask */ |
174 | #define GMAC_FLOW_CTRL_PT_SHIFT 16 | 174 | #define GMAC_FLOW_CTRL_PT_SHIFT 16 |
175 | #define GMAC_FLOW_CTRL_UP 0x00000008 /* Unicast pause frame enable */ | ||
175 | #define GMAC_FLOW_CTRL_RFE 0x00000004 /* Rx Flow Control Enable */ | 176 | #define GMAC_FLOW_CTRL_RFE 0x00000004 /* Rx Flow Control Enable */ |
176 | #define GMAC_FLOW_CTRL_TFE 0x00000002 /* Tx Flow Control Enable */ | 177 | #define GMAC_FLOW_CTRL_TFE 0x00000002 /* Tx Flow Control Enable */ |
177 | #define GMAC_FLOW_CTRL_FCB_BPA 0x00000001 /* Flow Control Busy ... */ | 178 | #define GMAC_FLOW_CTRL_FCB_BPA 0x00000001 /* Flow Control Busy ... */ |
@@ -246,6 +247,56 @@ enum ttc_control { | |||
246 | #define DMA_CONTROL_FEF 0x00000080 | 247 | #define DMA_CONTROL_FEF 0x00000080 |
247 | #define DMA_CONTROL_FUF 0x00000040 | 248 | #define DMA_CONTROL_FUF 0x00000040 |
248 | 249 | ||
250 | /* Receive flow control activation field | ||
251 | * RFA field in DMA control register, bits 23,10:9 | ||
252 | */ | ||
253 | #define DMA_CONTROL_RFA_MASK 0x00800600 | ||
254 | |||
255 | /* Receive flow control deactivation field | ||
256 | * RFD field in DMA control register, bits 22,12:11 | ||
257 | */ | ||
258 | #define DMA_CONTROL_RFD_MASK 0x00401800 | ||
259 | |||
260 | /* RFD and RFA fields are encoded as follows | ||
261 | * | ||
262 | * Bit Field | ||
263 | * 0,00 - Full minus 1KB (only valid when rxfifo >= 4KB and EFC enabled) | ||
264 | * 0,01 - Full minus 2KB (only valid when rxfifo >= 4KB and EFC enabled) | ||
265 | * 0,10 - Full minus 3KB (only valid when rxfifo >= 4KB and EFC enabled) | ||
266 | * 0,11 - Full minus 4KB (only valid when rxfifo > 4KB and EFC enabled) | ||
267 | * 1,00 - Full minus 5KB (only valid when rxfifo > 8KB and EFC enabled) | ||
268 | * 1,01 - Full minus 6KB (only valid when rxfifo > 8KB and EFC enabled) | ||
269 | * 1,10 - Full minus 7KB (only valid when rxfifo > 8KB and EFC enabled) | ||
270 | * 1,11 - Reserved | ||
271 | * | ||
272 | * RFD should always be > RFA for a given FIFO size. RFD == RFA may work, | ||
273 | * but packet throughput performance may not be as expected. | ||
274 | * | ||
275 | * Be sure that bit 3 in GMAC Register 6 is set for Unicast Pause frame | ||
276 | * detection (IEEE Specification Requirement, Annex 31B, 31B.1, Pause | ||
277 | * Description). | ||
278 | * | ||
279 | * Be sure that DZPA (bit 7 in Flow Control Register, GMAC Register 6), | ||
280 | * is set to 0. This allows pause frames with a quanta of 0 to be sent | ||
281 | * as an XOFF message to the link peer. | ||
282 | */ | ||
283 | |||
284 | #define RFA_FULL_MINUS_1K 0x00000000 | ||
285 | #define RFA_FULL_MINUS_2K 0x00000200 | ||
286 | #define RFA_FULL_MINUS_3K 0x00000400 | ||
287 | #define RFA_FULL_MINUS_4K 0x00000600 | ||
288 | #define RFA_FULL_MINUS_5K 0x00800000 | ||
289 | #define RFA_FULL_MINUS_6K 0x00800200 | ||
290 | #define RFA_FULL_MINUS_7K 0x00800400 | ||
291 | |||
292 | #define RFD_FULL_MINUS_1K 0x00000000 | ||
293 | #define RFD_FULL_MINUS_2K 0x00000800 | ||
294 | #define RFD_FULL_MINUS_3K 0x00001000 | ||
295 | #define RFD_FULL_MINUS_4K 0x00001800 | ||
296 | #define RFD_FULL_MINUS_5K 0x00400000 | ||
297 | #define RFD_FULL_MINUS_6K 0x00400800 | ||
298 | #define RFD_FULL_MINUS_7K 0x00401000 | ||
299 | |||
249 | enum rtc_control { | 300 | enum rtc_control { |
250 | DMA_CONTROL_RTC_64 = 0x00000000, | 301 | DMA_CONTROL_RTC_64 = 0x00000000, |
251 | DMA_CONTROL_RTC_32 = 0x00000008, | 302 | DMA_CONTROL_RTC_32 = 0x00000008, |
diff --git a/drivers/net/ethernet/stmicro/stmmac/dwmac1000_core.c b/drivers/net/ethernet/stmicro/stmmac/dwmac1000_core.c index 0adcf73cf722..371a669d69fd 100644 --- a/drivers/net/ethernet/stmicro/stmmac/dwmac1000_core.c +++ b/drivers/net/ethernet/stmicro/stmmac/dwmac1000_core.c | |||
@@ -201,7 +201,10 @@ static void dwmac1000_flow_ctrl(struct mac_device_info *hw, unsigned int duplex, | |||
201 | unsigned int fc, unsigned int pause_time) | 201 | unsigned int fc, unsigned int pause_time) |
202 | { | 202 | { |
203 | void __iomem *ioaddr = hw->pcsr; | 203 | void __iomem *ioaddr = hw->pcsr; |
204 | unsigned int flow = 0; | 204 | /* Set flow such that DZPQ in Mac Register 6 is 0, |
205 | * and unicast pause detect is enabled. | ||
206 | */ | ||
207 | unsigned int flow = GMAC_FLOW_CTRL_UP; | ||
205 | 208 | ||
206 | pr_debug("GMAC Flow-Control:\n"); | 209 | pr_debug("GMAC Flow-Control:\n"); |
207 | if (fc & FLOW_RX) { | 210 | if (fc & FLOW_RX) { |
diff --git a/drivers/net/ethernet/stmicro/stmmac/dwmac1000_dma.c b/drivers/net/ethernet/stmicro/stmmac/dwmac1000_dma.c index 59d92e811750..0e8937c1184a 100644 --- a/drivers/net/ethernet/stmicro/stmmac/dwmac1000_dma.c +++ b/drivers/net/ethernet/stmicro/stmmac/dwmac1000_dma.c | |||
@@ -106,8 +106,29 @@ static int dwmac1000_dma_init(void __iomem *ioaddr, int pbl, int fb, int mb, | |||
106 | return 0; | 106 | return 0; |
107 | } | 107 | } |
108 | 108 | ||
109 | static u32 dwmac1000_configure_fc(u32 csr6, int rxfifosz) | ||
110 | { | ||
111 | csr6 &= ~DMA_CONTROL_RFA_MASK; | ||
112 | csr6 &= ~DMA_CONTROL_RFD_MASK; | ||
113 | |||
114 | /* Leave flow control disabled if receive fifo size is less than | ||
115 | * 4K or 0. Otherwise, send XOFF when fifo is 1K less than full, | ||
116 | * and send XON when 2K less than full. | ||
117 | */ | ||
118 | if (rxfifosz < 4096) { | ||
119 | csr6 &= ~DMA_CONTROL_EFC; | ||
120 | pr_debug("GMAC: disabling flow control, rxfifo too small(%d)\n", | ||
121 | rxfifosz); | ||
122 | } else { | ||
123 | csr6 |= DMA_CONTROL_EFC; | ||
124 | csr6 |= RFA_FULL_MINUS_1K; | ||
125 | csr6 |= RFD_FULL_MINUS_2K; | ||
126 | } | ||
127 | return csr6; | ||
128 | } | ||
129 | |||
109 | static void dwmac1000_dma_operation_mode(void __iomem *ioaddr, int txmode, | 130 | static void dwmac1000_dma_operation_mode(void __iomem *ioaddr, int txmode, |
110 | int rxmode) | 131 | int rxmode, int rxfifosz) |
111 | { | 132 | { |
112 | u32 csr6 = readl(ioaddr + DMA_CONTROL); | 133 | u32 csr6 = readl(ioaddr + DMA_CONTROL); |
113 | 134 | ||
@@ -153,6 +174,9 @@ static void dwmac1000_dma_operation_mode(void __iomem *ioaddr, int txmode, | |||
153 | csr6 |= DMA_CONTROL_RTC_128; | 174 | csr6 |= DMA_CONTROL_RTC_128; |
154 | } | 175 | } |
155 | 176 | ||
177 | /* Configure flow control based on rx fifo size */ | ||
178 | csr6 = dwmac1000_configure_fc(csr6, rxfifosz); | ||
179 | |||
156 | writel(csr6, ioaddr + DMA_CONTROL); | 180 | writel(csr6, ioaddr + DMA_CONTROL); |
157 | } | 181 | } |
158 | 182 | ||
diff --git a/drivers/net/ethernet/stmicro/stmmac/dwmac100_dma.c b/drivers/net/ethernet/stmicro/stmmac/dwmac100_dma.c index 7d1dce9e7ffc..9d0971c1c2ee 100644 --- a/drivers/net/ethernet/stmicro/stmmac/dwmac100_dma.c +++ b/drivers/net/ethernet/stmicro/stmmac/dwmac100_dma.c | |||
@@ -72,7 +72,7 @@ static int dwmac100_dma_init(void __iomem *ioaddr, int pbl, int fb, int mb, | |||
72 | * control register. | 72 | * control register. |
73 | */ | 73 | */ |
74 | static void dwmac100_dma_operation_mode(void __iomem *ioaddr, int txmode, | 74 | static void dwmac100_dma_operation_mode(void __iomem *ioaddr, int txmode, |
75 | int rxmode) | 75 | int rxmode, int rxfifosz) |
76 | { | 76 | { |
77 | u32 csr6 = readl(ioaddr + DMA_CONTROL); | 77 | u32 csr6 = readl(ioaddr + DMA_CONTROL); |
78 | 78 | ||
diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c index 06103cad7c77..05c146f718a3 100644 --- a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c +++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c | |||
@@ -1277,8 +1277,10 @@ static void free_dma_desc_resources(struct stmmac_priv *priv) | |||
1277 | */ | 1277 | */ |
1278 | static void stmmac_dma_operation_mode(struct stmmac_priv *priv) | 1278 | static void stmmac_dma_operation_mode(struct stmmac_priv *priv) |
1279 | { | 1279 | { |
1280 | int rxfifosz = priv->plat->rx_fifo_size; | ||
1281 | |||
1280 | if (priv->plat->force_thresh_dma_mode) | 1282 | if (priv->plat->force_thresh_dma_mode) |
1281 | priv->hw->dma->dma_mode(priv->ioaddr, tc, tc); | 1283 | priv->hw->dma->dma_mode(priv->ioaddr, tc, tc, rxfifosz); |
1282 | else if (priv->plat->force_sf_dma_mode || priv->plat->tx_coe) { | 1284 | else if (priv->plat->force_sf_dma_mode || priv->plat->tx_coe) { |
1283 | /* | 1285 | /* |
1284 | * In case of GMAC, SF mode can be enabled | 1286 | * In case of GMAC, SF mode can be enabled |
@@ -1287,10 +1289,12 @@ static void stmmac_dma_operation_mode(struct stmmac_priv *priv) | |||
1287 | * 2) There is no bugged Jumbo frame support | 1289 | * 2) There is no bugged Jumbo frame support |
1288 | * that needs to not insert csum in the TDES. | 1290 | * that needs to not insert csum in the TDES. |
1289 | */ | 1291 | */ |
1290 | priv->hw->dma->dma_mode(priv->ioaddr, SF_DMA_MODE, SF_DMA_MODE); | 1292 | priv->hw->dma->dma_mode(priv->ioaddr, SF_DMA_MODE, SF_DMA_MODE, |
1293 | rxfifosz); | ||
1291 | priv->xstats.threshold = SF_DMA_MODE; | 1294 | priv->xstats.threshold = SF_DMA_MODE; |
1292 | } else | 1295 | } else |
1293 | priv->hw->dma->dma_mode(priv->ioaddr, tc, SF_DMA_MODE); | 1296 | priv->hw->dma->dma_mode(priv->ioaddr, tc, SF_DMA_MODE, |
1297 | rxfifosz); | ||
1294 | } | 1298 | } |
1295 | 1299 | ||
1296 | /** | 1300 | /** |
@@ -1442,6 +1446,7 @@ static void stmmac_tx_err(struct stmmac_priv *priv) | |||
1442 | static void stmmac_dma_interrupt(struct stmmac_priv *priv) | 1446 | static void stmmac_dma_interrupt(struct stmmac_priv *priv) |
1443 | { | 1447 | { |
1444 | int status; | 1448 | int status; |
1449 | int rxfifosz = priv->plat->rx_fifo_size; | ||
1445 | 1450 | ||
1446 | status = priv->hw->dma->dma_interrupt(priv->ioaddr, &priv->xstats); | 1451 | status = priv->hw->dma->dma_interrupt(priv->ioaddr, &priv->xstats); |
1447 | if (likely((status & handle_rx)) || (status & handle_tx)) { | 1452 | if (likely((status & handle_rx)) || (status & handle_tx)) { |
@@ -1456,10 +1461,11 @@ static void stmmac_dma_interrupt(struct stmmac_priv *priv) | |||
1456 | (tc <= 256)) { | 1461 | (tc <= 256)) { |
1457 | tc += 64; | 1462 | tc += 64; |
1458 | if (priv->plat->force_thresh_dma_mode) | 1463 | if (priv->plat->force_thresh_dma_mode) |
1459 | priv->hw->dma->dma_mode(priv->ioaddr, tc, tc); | 1464 | priv->hw->dma->dma_mode(priv->ioaddr, tc, tc, |
1465 | rxfifosz); | ||
1460 | else | 1466 | else |
1461 | priv->hw->dma->dma_mode(priv->ioaddr, tc, | 1467 | priv->hw->dma->dma_mode(priv->ioaddr, tc, |
1462 | SF_DMA_MODE); | 1468 | SF_DMA_MODE, rxfifosz); |
1463 | priv->xstats.threshold = tc; | 1469 | priv->xstats.threshold = tc; |
1464 | } | 1470 | } |
1465 | } else if (unlikely(status == tx_hard_error)) | 1471 | } else if (unlikely(status == tx_hard_error)) |
@@ -2970,15 +2976,15 @@ int stmmac_dvr_remove(struct net_device *ndev) | |||
2970 | priv->hw->dma->stop_tx(priv->ioaddr); | 2976 | priv->hw->dma->stop_tx(priv->ioaddr); |
2971 | 2977 | ||
2972 | stmmac_set_mac(priv->ioaddr, false); | 2978 | stmmac_set_mac(priv->ioaddr, false); |
2973 | if (priv->pcs != STMMAC_PCS_RGMII && priv->pcs != STMMAC_PCS_TBI && | ||
2974 | priv->pcs != STMMAC_PCS_RTBI) | ||
2975 | stmmac_mdio_unregister(ndev); | ||
2976 | netif_carrier_off(ndev); | 2979 | netif_carrier_off(ndev); |
2977 | unregister_netdev(ndev); | 2980 | unregister_netdev(ndev); |
2978 | if (priv->stmmac_rst) | 2981 | if (priv->stmmac_rst) |
2979 | reset_control_assert(priv->stmmac_rst); | 2982 | reset_control_assert(priv->stmmac_rst); |
2980 | clk_disable_unprepare(priv->pclk); | 2983 | clk_disable_unprepare(priv->pclk); |
2981 | clk_disable_unprepare(priv->stmmac_clk); | 2984 | clk_disable_unprepare(priv->stmmac_clk); |
2985 | if (priv->pcs != STMMAC_PCS_RGMII && priv->pcs != STMMAC_PCS_TBI && | ||
2986 | priv->pcs != STMMAC_PCS_RTBI) | ||
2987 | stmmac_mdio_unregister(ndev); | ||
2982 | free_netdev(ndev); | 2988 | free_netdev(ndev); |
2983 | 2989 | ||
2984 | return 0; | 2990 | return 0; |
diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c b/drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c index f9b42f11950f..705bbdf93940 100644 --- a/drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c +++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c | |||
@@ -181,6 +181,10 @@ static int stmmac_probe_config_dt(struct platform_device *pdev, | |||
181 | sizeof(struct stmmac_mdio_bus_data), | 181 | sizeof(struct stmmac_mdio_bus_data), |
182 | GFP_KERNEL); | 182 | GFP_KERNEL); |
183 | 183 | ||
184 | of_property_read_u32(np, "tx-fifo-depth", &plat->tx_fifo_size); | ||
185 | |||
186 | of_property_read_u32(np, "rx-fifo-depth", &plat->rx_fifo_size); | ||
187 | |||
184 | plat->force_sf_dma_mode = | 188 | plat->force_sf_dma_mode = |
185 | of_property_read_bool(np, "snps,force_sf_dma_mode"); | 189 | of_property_read_bool(np, "snps,force_sf_dma_mode"); |
186 | 190 | ||
diff --git a/drivers/net/ifb.c b/drivers/net/ifb.c index 34f846b4bd05..94570aace241 100644 --- a/drivers/net/ifb.c +++ b/drivers/net/ifb.c | |||
@@ -105,7 +105,7 @@ static void ri_tasklet(unsigned long dev) | |||
105 | if (from & AT_EGRESS) { | 105 | if (from & AT_EGRESS) { |
106 | dev_queue_xmit(skb); | 106 | dev_queue_xmit(skb); |
107 | } else if (from & AT_INGRESS) { | 107 | } else if (from & AT_INGRESS) { |
108 | skb_pull(skb, skb->dev->hard_header_len); | 108 | skb_pull(skb, skb->mac_len); |
109 | netif_receive_skb(skb); | 109 | netif_receive_skb(skb); |
110 | } else | 110 | } else |
111 | BUG(); | 111 | BUG(); |
diff --git a/drivers/net/macvtap.c b/drivers/net/macvtap.c index 9c91ff872485..8c350c5d54ad 100644 --- a/drivers/net/macvtap.c +++ b/drivers/net/macvtap.c | |||
@@ -313,7 +313,7 @@ static rx_handler_result_t macvtap_handle_frame(struct sk_buff **pskb) | |||
313 | */ | 313 | */ |
314 | if (q->flags & IFF_VNET_HDR) | 314 | if (q->flags & IFF_VNET_HDR) |
315 | features |= vlan->tap_features; | 315 | features |= vlan->tap_features; |
316 | if (netif_needs_gso(dev, skb, features)) { | 316 | if (netif_needs_gso(skb, features)) { |
317 | struct sk_buff *segs = __skb_gso_segment(skb, features, false); | 317 | struct sk_buff *segs = __skb_gso_segment(skb, features, false); |
318 | 318 | ||
319 | if (IS_ERR(segs)) | 319 | if (IS_ERR(segs)) |
diff --git a/drivers/net/xen-netfront.c b/drivers/net/xen-netfront.c index 4c08f98f4484..3f45afd4382e 100644 --- a/drivers/net/xen-netfront.c +++ b/drivers/net/xen-netfront.c | |||
@@ -560,7 +560,7 @@ static int xennet_start_xmit(struct sk_buff *skb, struct net_device *dev) | |||
560 | 560 | ||
561 | if (unlikely(!netif_carrier_ok(dev) || | 561 | if (unlikely(!netif_carrier_ok(dev) || |
562 | (slots > 1 && !xennet_can_sg(dev)) || | 562 | (slots > 1 && !xennet_can_sg(dev)) || |
563 | netif_needs_gso(dev, skb, netif_skb_features(skb)))) { | 563 | netif_needs_gso(skb, netif_skb_features(skb)))) { |
564 | spin_unlock_irqrestore(&queue->tx_lock, flags); | 564 | spin_unlock_irqrestore(&queue->tx_lock, flags); |
565 | goto drop; | 565 | goto drop; |
566 | } | 566 | } |