aboutsummaryrefslogtreecommitdiffstats
path: root/net/core
diff options
context:
space:
mode:
authorEric Dumazet <eric.dumazet@gmail.com>2010-04-19 17:17:14 -0400
committerDavid S. Miller <davem@davemloft.net>2010-04-20 04:18:05 -0400
commite36fa2f7e92f25aab2e3d787dcfe3590817f19d3 (patch)
treeb1bd60b14131dacc72a80efbd2f062935cc43e25 /net/core
parentf5acb907dc24c3822f408211bad1cd6e5d0433cf (diff)
rps: cleanups
struct softnet_data holds many queues, so consistent use "sd" name instead of "queue" is better. Adds a rps_ipi_queued() helper to cleanup enqueue_to_backlog() Adds a _and_irq_disable suffix to net_rps_action() name, as David suggested. incr_input_queue_head() becomes input_queue_head_incr() Signed-off-by: Eric Dumazet <eric.dumazet@gmail.com> Signed-off-by: David S. Miller <davem@davemloft.net>
Diffstat (limited to 'net/core')
-rw-r--r--net/core/dev.c149
1 files changed, 80 insertions, 69 deletions
diff --git a/net/core/dev.c b/net/core/dev.c
index 05a2b294906b..7f5755b0a57c 100644
--- a/net/core/dev.c
+++ b/net/core/dev.c
@@ -208,17 +208,17 @@ static inline struct hlist_head *dev_index_hash(struct net *net, int ifindex)
208 return &net->dev_index_head[ifindex & (NETDEV_HASHENTRIES - 1)]; 208 return &net->dev_index_head[ifindex & (NETDEV_HASHENTRIES - 1)];
209} 209}
210 210
211static inline void rps_lock(struct softnet_data *queue) 211static inline void rps_lock(struct softnet_data *sd)
212{ 212{
213#ifdef CONFIG_RPS 213#ifdef CONFIG_RPS
214 spin_lock(&queue->input_pkt_queue.lock); 214 spin_lock(&sd->input_pkt_queue.lock);
215#endif 215#endif
216} 216}
217 217
218static inline void rps_unlock(struct softnet_data *queue) 218static inline void rps_unlock(struct softnet_data *sd)
219{ 219{
220#ifdef CONFIG_RPS 220#ifdef CONFIG_RPS
221 spin_unlock(&queue->input_pkt_queue.lock); 221 spin_unlock(&sd->input_pkt_queue.lock);
222#endif 222#endif
223} 223}
224 224
@@ -2346,63 +2346,74 @@ done:
2346} 2346}
2347 2347
2348/* Called from hardirq (IPI) context */ 2348/* Called from hardirq (IPI) context */
2349static void trigger_softirq(void *data) 2349static void rps_trigger_softirq(void *data)
2350{ 2350{
2351 struct softnet_data *queue = data; 2351 struct softnet_data *sd = data;
2352 __napi_schedule(&queue->backlog); 2352
2353 __napi_schedule(&sd->backlog);
2353 __get_cpu_var(netdev_rx_stat).received_rps++; 2354 __get_cpu_var(netdev_rx_stat).received_rps++;
2354} 2355}
2356
2355#endif /* CONFIG_RPS */ 2357#endif /* CONFIG_RPS */
2356 2358
2357/* 2359/*
2360 * Check if this softnet_data structure is another cpu one
2361 * If yes, queue it to our IPI list and return 1
2362 * If no, return 0
2363 */
2364static int rps_ipi_queued(struct softnet_data *sd)
2365{
2366#ifdef CONFIG_RPS
2367 struct softnet_data *mysd = &__get_cpu_var(softnet_data);
2368
2369 if (sd != mysd) {
2370 sd->rps_ipi_next = mysd->rps_ipi_list;
2371 mysd->rps_ipi_list = sd;
2372
2373 __raise_softirq_irqoff(NET_RX_SOFTIRQ);
2374 return 1;
2375 }
2376#endif /* CONFIG_RPS */
2377 return 0;
2378}
2379
2380/*
2358 * enqueue_to_backlog is called to queue an skb to a per CPU backlog 2381 * enqueue_to_backlog is called to queue an skb to a per CPU backlog
2359 * queue (may be a remote CPU queue). 2382 * queue (may be a remote CPU queue).
2360 */ 2383 */
2361static int enqueue_to_backlog(struct sk_buff *skb, int cpu, 2384static int enqueue_to_backlog(struct sk_buff *skb, int cpu,
2362 unsigned int *qtail) 2385 unsigned int *qtail)
2363{ 2386{
2364 struct softnet_data *queue; 2387 struct softnet_data *sd;
2365 unsigned long flags; 2388 unsigned long flags;
2366 2389
2367 queue = &per_cpu(softnet_data, cpu); 2390 sd = &per_cpu(softnet_data, cpu);
2368 2391
2369 local_irq_save(flags); 2392 local_irq_save(flags);
2370 __get_cpu_var(netdev_rx_stat).total++; 2393 __get_cpu_var(netdev_rx_stat).total++;
2371 2394
2372 rps_lock(queue); 2395 rps_lock(sd);
2373 if (queue->input_pkt_queue.qlen <= netdev_max_backlog) { 2396 if (sd->input_pkt_queue.qlen <= netdev_max_backlog) {
2374 if (queue->input_pkt_queue.qlen) { 2397 if (sd->input_pkt_queue.qlen) {
2375enqueue: 2398enqueue:
2376 __skb_queue_tail(&queue->input_pkt_queue, skb); 2399 __skb_queue_tail(&sd->input_pkt_queue, skb);
2377#ifdef CONFIG_RPS 2400#ifdef CONFIG_RPS
2378 *qtail = queue->input_queue_head + 2401 *qtail = sd->input_queue_head + sd->input_pkt_queue.qlen;
2379 queue->input_pkt_queue.qlen;
2380#endif 2402#endif
2381 rps_unlock(queue); 2403 rps_unlock(sd);
2382 local_irq_restore(flags); 2404 local_irq_restore(flags);
2383 return NET_RX_SUCCESS; 2405 return NET_RX_SUCCESS;
2384 } 2406 }
2385 2407
2386 /* Schedule NAPI for backlog device */ 2408 /* Schedule NAPI for backlog device */
2387 if (napi_schedule_prep(&queue->backlog)) { 2409 if (napi_schedule_prep(&sd->backlog)) {
2388#ifdef CONFIG_RPS 2410 if (!rps_ipi_queued(sd))
2389 if (cpu != smp_processor_id()) { 2411 __napi_schedule(&sd->backlog);
2390 struct softnet_data *myqueue;
2391
2392 myqueue = &__get_cpu_var(softnet_data);
2393 queue->rps_ipi_next = myqueue->rps_ipi_list;
2394 myqueue->rps_ipi_list = queue;
2395
2396 __raise_softirq_irqoff(NET_RX_SOFTIRQ);
2397 goto enqueue;
2398 }
2399#endif
2400 __napi_schedule(&queue->backlog);
2401 } 2412 }
2402 goto enqueue; 2413 goto enqueue;
2403 } 2414 }
2404 2415
2405 rps_unlock(queue); 2416 rps_unlock(sd);
2406 2417
2407 __get_cpu_var(netdev_rx_stat).dropped++; 2418 __get_cpu_var(netdev_rx_stat).dropped++;
2408 local_irq_restore(flags); 2419 local_irq_restore(flags);
@@ -2903,17 +2914,17 @@ EXPORT_SYMBOL(netif_receive_skb);
2903static void flush_backlog(void *arg) 2914static void flush_backlog(void *arg)
2904{ 2915{
2905 struct net_device *dev = arg; 2916 struct net_device *dev = arg;
2906 struct softnet_data *queue = &__get_cpu_var(softnet_data); 2917 struct softnet_data *sd = &__get_cpu_var(softnet_data);
2907 struct sk_buff *skb, *tmp; 2918 struct sk_buff *skb, *tmp;
2908 2919
2909 rps_lock(queue); 2920 rps_lock(sd);
2910 skb_queue_walk_safe(&queue->input_pkt_queue, skb, tmp) 2921 skb_queue_walk_safe(&sd->input_pkt_queue, skb, tmp)
2911 if (skb->dev == dev) { 2922 if (skb->dev == dev) {
2912 __skb_unlink(skb, &queue->input_pkt_queue); 2923 __skb_unlink(skb, &sd->input_pkt_queue);
2913 kfree_skb(skb); 2924 kfree_skb(skb);
2914 incr_input_queue_head(queue); 2925 input_queue_head_incr(sd);
2915 } 2926 }
2916 rps_unlock(queue); 2927 rps_unlock(sd);
2917} 2928}
2918 2929
2919static int napi_gro_complete(struct sk_buff *skb) 2930static int napi_gro_complete(struct sk_buff *skb)
@@ -3219,23 +3230,23 @@ EXPORT_SYMBOL(napi_gro_frags);
3219static int process_backlog(struct napi_struct *napi, int quota) 3230static int process_backlog(struct napi_struct *napi, int quota)
3220{ 3231{
3221 int work = 0; 3232 int work = 0;
3222 struct softnet_data *queue = &__get_cpu_var(softnet_data); 3233 struct softnet_data *sd = &__get_cpu_var(softnet_data);
3223 3234
3224 napi->weight = weight_p; 3235 napi->weight = weight_p;
3225 do { 3236 do {
3226 struct sk_buff *skb; 3237 struct sk_buff *skb;
3227 3238
3228 local_irq_disable(); 3239 local_irq_disable();
3229 rps_lock(queue); 3240 rps_lock(sd);
3230 skb = __skb_dequeue(&queue->input_pkt_queue); 3241 skb = __skb_dequeue(&sd->input_pkt_queue);
3231 if (!skb) { 3242 if (!skb) {
3232 __napi_complete(napi); 3243 __napi_complete(napi);
3233 rps_unlock(queue); 3244 rps_unlock(sd);
3234 local_irq_enable(); 3245 local_irq_enable();
3235 break; 3246 break;
3236 } 3247 }
3237 incr_input_queue_head(queue); 3248 input_queue_head_incr(sd);
3238 rps_unlock(queue); 3249 rps_unlock(sd);
3239 local_irq_enable(); 3250 local_irq_enable();
3240 3251
3241 __netif_receive_skb(skb); 3252 __netif_receive_skb(skb);
@@ -3331,24 +3342,25 @@ EXPORT_SYMBOL(netif_napi_del);
3331 * net_rps_action sends any pending IPI's for rps. 3342 * net_rps_action sends any pending IPI's for rps.
3332 * Note: called with local irq disabled, but exits with local irq enabled. 3343 * Note: called with local irq disabled, but exits with local irq enabled.
3333 */ 3344 */
3334static void net_rps_action(void) 3345static void net_rps_action_and_irq_disable(void)
3335{ 3346{
3336#ifdef CONFIG_RPS 3347#ifdef CONFIG_RPS
3337 struct softnet_data *locqueue = &__get_cpu_var(softnet_data); 3348 struct softnet_data *sd = &__get_cpu_var(softnet_data);
3338 struct softnet_data *remqueue = locqueue->rps_ipi_list; 3349 struct softnet_data *remsd = sd->rps_ipi_list;
3339 3350
3340 if (remqueue) { 3351 if (remsd) {
3341 locqueue->rps_ipi_list = NULL; 3352 sd->rps_ipi_list = NULL;
3342 3353
3343 local_irq_enable(); 3354 local_irq_enable();
3344 3355
3345 /* Send pending IPI's to kick RPS processing on remote cpus. */ 3356 /* Send pending IPI's to kick RPS processing on remote cpus. */
3346 while (remqueue) { 3357 while (remsd) {
3347 struct softnet_data *next = remqueue->rps_ipi_next; 3358 struct softnet_data *next = remsd->rps_ipi_next;
3348 if (cpu_online(remqueue->cpu)) 3359
3349 __smp_call_function_single(remqueue->cpu, 3360 if (cpu_online(remsd->cpu))
3350 &remqueue->csd, 0); 3361 __smp_call_function_single(remsd->cpu,
3351 remqueue = next; 3362 &remsd->csd, 0);
3363 remsd = next;
3352 } 3364 }
3353 } else 3365 } else
3354#endif 3366#endif
@@ -3423,7 +3435,7 @@ static void net_rx_action(struct softirq_action *h)
3423 netpoll_poll_unlock(have); 3435 netpoll_poll_unlock(have);
3424 } 3436 }
3425out: 3437out:
3426 net_rps_action(); 3438 net_rps_action_and_irq_disable();
3427 3439
3428#ifdef CONFIG_NET_DMA 3440#ifdef CONFIG_NET_DMA
3429 /* 3441 /*
@@ -5595,7 +5607,7 @@ static int dev_cpu_callback(struct notifier_block *nfb,
5595 /* Process offline CPU's input_pkt_queue */ 5607 /* Process offline CPU's input_pkt_queue */
5596 while ((skb = __skb_dequeue(&oldsd->input_pkt_queue))) { 5608 while ((skb = __skb_dequeue(&oldsd->input_pkt_queue))) {
5597 netif_rx(skb); 5609 netif_rx(skb);
5598 incr_input_queue_head(oldsd); 5610 input_queue_head_incr(oldsd);
5599 } 5611 }
5600 5612
5601 return NOTIFY_OK; 5613 return NOTIFY_OK;
@@ -5812,24 +5824,23 @@ static int __init net_dev_init(void)
5812 */ 5824 */
5813 5825
5814 for_each_possible_cpu(i) { 5826 for_each_possible_cpu(i) {
5815 struct softnet_data *queue; 5827 struct softnet_data *sd = &per_cpu(softnet_data, i);
5816 5828
5817 queue = &per_cpu(softnet_data, i); 5829 skb_queue_head_init(&sd->input_pkt_queue);
5818 skb_queue_head_init(&queue->input_pkt_queue); 5830 sd->completion_queue = NULL;
5819 queue->completion_queue = NULL; 5831 INIT_LIST_HEAD(&sd->poll_list);
5820 INIT_LIST_HEAD(&queue->poll_list);
5821 5832
5822#ifdef CONFIG_RPS 5833#ifdef CONFIG_RPS
5823 queue->csd.func = trigger_softirq; 5834 sd->csd.func = rps_trigger_softirq;
5824 queue->csd.info = queue; 5835 sd->csd.info = sd;
5825 queue->csd.flags = 0; 5836 sd->csd.flags = 0;
5826 queue->cpu = i; 5837 sd->cpu = i;
5827#endif 5838#endif
5828 5839
5829 queue->backlog.poll = process_backlog; 5840 sd->backlog.poll = process_backlog;
5830 queue->backlog.weight = weight_p; 5841 sd->backlog.weight = weight_p;
5831 queue->backlog.gro_list = NULL; 5842 sd->backlog.gro_list = NULL;
5832 queue->backlog.gro_count = 0; 5843 sd->backlog.gro_count = 0;
5833 } 5844 }
5834 5845
5835 dev_boot_phase = 0; 5846 dev_boot_phase = 0;