aboutsummaryrefslogtreecommitdiffstats
path: root/net/sched
diff options
context:
space:
mode:
authorEric Dumazet <edumazet@google.com>2014-10-05 21:38:35 -0400
committerDavid S. Miller <davem@davemloft.net>2014-10-07 13:22:11 -0400
commit0287587884b15041203b3a362d485e1ab1f24445 (patch)
tree675ae57663c1ba3ee8768e65e7fb0e6d0259e04c /net/sched
parentfe971b95c22578456ff7198537827841c726d3f7 (diff)
net: better IFF_XMIT_DST_RELEASE support
Testing xmit_more support with netperf and connected UDP sockets, I found strange dst refcount false sharing. Current handling of IFF_XMIT_DST_RELEASE is not optimal. Dropping dst in validate_xmit_skb() is certainly too late in case packet was queued by cpu X but dequeued by cpu Y The logical point to take care of drop/force is in __dev_queue_xmit() before even taking qdisc lock. As Julian Anastasov pointed out, need for skb_dst() might come from some packet schedulers or classifiers. This patch adds new helper to cleanly express needs of various drivers or qdiscs/classifiers. Drivers that need skb_dst() in their ndo_start_xmit() should call following helper in their setup instead of the prior : dev->priv_flags &= ~IFF_XMIT_DST_RELEASE; -> netif_keep_dst(dev); Instead of using a single bit, we use two bits, one being eventually rebuilt in bonding/team drivers. The other one, is permanent and blocks IFF_XMIT_DST_RELEASE being rebuilt in bonding/team. Eventually, we could add something smarter later. Signed-off-by: Eric Dumazet <edumazet@google.com> Cc: Julian Anastasov <ja@ssi.bg> Signed-off-by: David S. Miller <davem@davemloft.net>
Diffstat (limited to 'net/sched')
-rw-r--r--net/sched/cls_flow.c2
-rw-r--r--net/sched/cls_route.c1
-rw-r--r--net/sched/sch_generic.c3
-rw-r--r--net/sched/sch_teql.c2
4 files changed, 4 insertions, 4 deletions
diff --git a/net/sched/cls_flow.c b/net/sched/cls_flow.c
index a5d2b20db560..4ac515f2a6ce 100644
--- a/net/sched/cls_flow.c
+++ b/net/sched/cls_flow.c
@@ -493,6 +493,8 @@ static int flow_change(struct net *net, struct sk_buff *in_skb,
493 tcf_exts_change(tp, &fnew->exts, &e); 493 tcf_exts_change(tp, &fnew->exts, &e);
494 tcf_em_tree_change(tp, &fnew->ematches, &t); 494 tcf_em_tree_change(tp, &fnew->ematches, &t);
495 495
496 netif_keep_dst(qdisc_dev(tp->q));
497
496 if (tb[TCA_FLOW_KEYS]) { 498 if (tb[TCA_FLOW_KEYS]) {
497 fnew->keymask = keymask; 499 fnew->keymask = keymask;
498 fnew->nkeys = nkeys; 500 fnew->nkeys = nkeys;
diff --git a/net/sched/cls_route.c b/net/sched/cls_route.c
index 6f22baae0afa..109a329b7198 100644
--- a/net/sched/cls_route.c
+++ b/net/sched/cls_route.c
@@ -524,6 +524,7 @@ static int route4_change(struct net *net, struct sk_buff *in_skb,
524 if (f->handle < f1->handle) 524 if (f->handle < f1->handle)
525 break; 525 break;
526 526
527 netif_keep_dst(qdisc_dev(tp->q));
527 rcu_assign_pointer(f->next, f1); 528 rcu_assign_pointer(f->next, f1);
528 rcu_assign_pointer(*fp, f); 529 rcu_assign_pointer(*fp, f);
529 530
diff --git a/net/sched/sch_generic.c b/net/sched/sch_generic.c
index 2b349a4de3c8..38d58e6cef07 100644
--- a/net/sched/sch_generic.c
+++ b/net/sched/sch_generic.c
@@ -47,7 +47,6 @@ EXPORT_SYMBOL(default_qdisc_ops);
47 47
48static inline int dev_requeue_skb(struct sk_buff *skb, struct Qdisc *q) 48static inline int dev_requeue_skb(struct sk_buff *skb, struct Qdisc *q)
49{ 49{
50 skb_dst_force(skb);
51 q->gso_skb = skb; 50 q->gso_skb = skb;
52 q->qstats.requeues++; 51 q->qstats.requeues++;
53 q->q.qlen++; /* it's still part of the queue */ 52 q->q.qlen++; /* it's still part of the queue */
@@ -218,8 +217,6 @@ static inline int qdisc_restart(struct Qdisc *q)
218 if (unlikely(!skb)) 217 if (unlikely(!skb))
219 return 0; 218 return 0;
220 219
221 WARN_ON_ONCE(skb_dst_is_noref(skb));
222
223 root_lock = qdisc_lock(q); 220 root_lock = qdisc_lock(q);
224 dev = qdisc_dev(q); 221 dev = qdisc_dev(q);
225 txq = skb_get_tx_queue(dev, skb); 222 txq = skb_get_tx_queue(dev, skb);
diff --git a/net/sched/sch_teql.c b/net/sched/sch_teql.c
index 5cd291bd00e4..6ada42396a24 100644
--- a/net/sched/sch_teql.c
+++ b/net/sched/sch_teql.c
@@ -470,7 +470,7 @@ static __init void teql_master_setup(struct net_device *dev)
470 dev->tx_queue_len = 100; 470 dev->tx_queue_len = 100;
471 dev->flags = IFF_NOARP; 471 dev->flags = IFF_NOARP;
472 dev->hard_header_len = LL_MAX_HEADER; 472 dev->hard_header_len = LL_MAX_HEADER;
473 dev->priv_flags &= ~IFF_XMIT_DST_RELEASE; 473 netif_keep_dst(dev);
474} 474}
475 475
476static LIST_HEAD(master_dev_list); 476static LIST_HEAD(master_dev_list);