diff options
author | Tom Herbert <therbert@google.com> | 2010-05-20 14:37:59 -0400 |
---|---|---|
committer | David S. Miller <davem@davemloft.net> | 2010-05-21 03:38:33 -0400 |
commit | 76cc8b13a6e41b537fd262b600da1571314add62 (patch) | |
tree | a8a25a4f23be5c0eeed77dd9fcb30dc9b315e519 /net | |
parent | 1f01bfd202bc539bccd282befa2bbdb8d6ad80ee (diff) |
net: fix problem in dequeuing from input_pkt_queue
Fix some issues introduced in batch skb dequeuing for input_pkt_queue.
The primary issue it that the queue head must be incremented only
after a packet has been processed, that is only after
__netif_receive_skb has been called. This is needed for the mechanism
to prevent OOO packet in RFS. Also when flushing the input_pkt_queue
and process_queue, the process queue should be done first to prevent
OOO packets.
Because the input_pkt_queue has been effectively split into two queues,
the calculation of the tail ptr is no longer correct. The correct value
would be head+input_pkt_queue->len+process_queue->len. To avoid
this calculation we added an explict input_queue_tail in softnet_data.
The tail value is simply incremented when queuing to input_pkt_queue.
Signed-off-by: Tom Herbert <therbert@google.com>
Acked-by: Eric Dumazet <eric.dumazet@gmail.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
Diffstat (limited to 'net')
-rw-r--r-- | net/core/dev.c | 28 |
1 files changed, 15 insertions, 13 deletions
diff --git a/net/core/dev.c b/net/core/dev.c index 6c820650b80f..0aab66d68b19 100644 --- a/net/core/dev.c +++ b/net/core/dev.c | |||
@@ -2426,10 +2426,7 @@ static int enqueue_to_backlog(struct sk_buff *skb, int cpu, | |||
2426 | if (skb_queue_len(&sd->input_pkt_queue)) { | 2426 | if (skb_queue_len(&sd->input_pkt_queue)) { |
2427 | enqueue: | 2427 | enqueue: |
2428 | __skb_queue_tail(&sd->input_pkt_queue, skb); | 2428 | __skb_queue_tail(&sd->input_pkt_queue, skb); |
2429 | #ifdef CONFIG_RPS | 2429 | input_queue_tail_incr_save(sd, qtail); |
2430 | *qtail = sd->input_queue_head + | ||
2431 | skb_queue_len(&sd->input_pkt_queue); | ||
2432 | #endif | ||
2433 | rps_unlock(sd); | 2430 | rps_unlock(sd); |
2434 | local_irq_restore(flags); | 2431 | local_irq_restore(flags); |
2435 | return NET_RX_SUCCESS; | 2432 | return NET_RX_SUCCESS; |
@@ -2964,7 +2961,7 @@ static void flush_backlog(void *arg) | |||
2964 | if (skb->dev == dev) { | 2961 | if (skb->dev == dev) { |
2965 | __skb_unlink(skb, &sd->input_pkt_queue); | 2962 | __skb_unlink(skb, &sd->input_pkt_queue); |
2966 | kfree_skb(skb); | 2963 | kfree_skb(skb); |
2967 | input_queue_head_add(sd, 1); | 2964 | input_queue_head_incr(sd); |
2968 | } | 2965 | } |
2969 | } | 2966 | } |
2970 | rps_unlock(sd); | 2967 | rps_unlock(sd); |
@@ -2973,6 +2970,7 @@ static void flush_backlog(void *arg) | |||
2973 | if (skb->dev == dev) { | 2970 | if (skb->dev == dev) { |
2974 | __skb_unlink(skb, &sd->process_queue); | 2971 | __skb_unlink(skb, &sd->process_queue); |
2975 | kfree_skb(skb); | 2972 | kfree_skb(skb); |
2973 | input_queue_head_incr(sd); | ||
2976 | } | 2974 | } |
2977 | } | 2975 | } |
2978 | } | 2976 | } |
@@ -3328,18 +3326,20 @@ static int process_backlog(struct napi_struct *napi, int quota) | |||
3328 | while ((skb = __skb_dequeue(&sd->process_queue))) { | 3326 | while ((skb = __skb_dequeue(&sd->process_queue))) { |
3329 | local_irq_enable(); | 3327 | local_irq_enable(); |
3330 | __netif_receive_skb(skb); | 3328 | __netif_receive_skb(skb); |
3331 | if (++work >= quota) | ||
3332 | return work; | ||
3333 | local_irq_disable(); | 3329 | local_irq_disable(); |
3330 | input_queue_head_incr(sd); | ||
3331 | if (++work >= quota) { | ||
3332 | local_irq_enable(); | ||
3333 | return work; | ||
3334 | } | ||
3334 | } | 3335 | } |
3335 | 3336 | ||
3336 | rps_lock(sd); | 3337 | rps_lock(sd); |
3337 | qlen = skb_queue_len(&sd->input_pkt_queue); | 3338 | qlen = skb_queue_len(&sd->input_pkt_queue); |
3338 | if (qlen) { | 3339 | if (qlen) |
3339 | input_queue_head_add(sd, qlen); | ||
3340 | skb_queue_splice_tail_init(&sd->input_pkt_queue, | 3340 | skb_queue_splice_tail_init(&sd->input_pkt_queue, |
3341 | &sd->process_queue); | 3341 | &sd->process_queue); |
3342 | } | 3342 | |
3343 | if (qlen < quota - work) { | 3343 | if (qlen < quota - work) { |
3344 | /* | 3344 | /* |
3345 | * Inline a custom version of __napi_complete(). | 3345 | * Inline a custom version of __napi_complete(). |
@@ -5679,12 +5679,14 @@ static int dev_cpu_callback(struct notifier_block *nfb, | |||
5679 | local_irq_enable(); | 5679 | local_irq_enable(); |
5680 | 5680 | ||
5681 | /* Process offline CPU's input_pkt_queue */ | 5681 | /* Process offline CPU's input_pkt_queue */ |
5682 | while ((skb = __skb_dequeue(&oldsd->input_pkt_queue))) { | 5682 | while ((skb = __skb_dequeue(&oldsd->process_queue))) { |
5683 | netif_rx(skb); | 5683 | netif_rx(skb); |
5684 | input_queue_head_add(oldsd, 1); | 5684 | input_queue_head_incr(oldsd); |
5685 | } | 5685 | } |
5686 | while ((skb = __skb_dequeue(&oldsd->process_queue))) | 5686 | while ((skb = __skb_dequeue(&oldsd->input_pkt_queue))) { |
5687 | netif_rx(skb); | 5687 | netif_rx(skb); |
5688 | input_queue_head_incr(oldsd); | ||
5689 | } | ||
5688 | 5690 | ||
5689 | return NOTIFY_OK; | 5691 | return NOTIFY_OK; |
5690 | } | 5692 | } |