diff options
Diffstat (limited to 'net')
85 files changed, 1015 insertions, 1023 deletions
diff --git a/net/802/tr.c b/net/802/tr.c index a755e880f4ba..1bb7dc1b85cd 100644 --- a/net/802/tr.c +++ b/net/802/tr.c | |||
@@ -251,10 +251,11 @@ void tr_source_route(struct sk_buff *skb,struct trh_hdr *trh,struct net_device * | |||
251 | unsigned int hash; | 251 | unsigned int hash; |
252 | struct rif_cache *entry; | 252 | struct rif_cache *entry; |
253 | unsigned char *olddata; | 253 | unsigned char *olddata; |
254 | unsigned long flags; | ||
254 | static const unsigned char mcast_func_addr[] | 255 | static const unsigned char mcast_func_addr[] |
255 | = {0xC0,0x00,0x00,0x04,0x00,0x00}; | 256 | = {0xC0,0x00,0x00,0x04,0x00,0x00}; |
256 | 257 | ||
257 | spin_lock_bh(&rif_lock); | 258 | spin_lock_irqsave(&rif_lock, flags); |
258 | 259 | ||
259 | /* | 260 | /* |
260 | * Broadcasts are single route as stated in RFC 1042 | 261 | * Broadcasts are single route as stated in RFC 1042 |
@@ -323,7 +324,7 @@ printk("source routing for %02X:%02X:%02X:%02X:%02X:%02X\n",trh->daddr[0], | |||
323 | else | 324 | else |
324 | slack = 18 - ((ntohs(trh->rcf) & TR_RCF_LEN_MASK)>>8); | 325 | slack = 18 - ((ntohs(trh->rcf) & TR_RCF_LEN_MASK)>>8); |
325 | olddata = skb->data; | 326 | olddata = skb->data; |
326 | spin_unlock_bh(&rif_lock); | 327 | spin_unlock_irqrestore(&rif_lock, flags); |
327 | 328 | ||
328 | skb_pull(skb, slack); | 329 | skb_pull(skb, slack); |
329 | memmove(skb->data, olddata, sizeof(struct trh_hdr) - slack); | 330 | memmove(skb->data, olddata, sizeof(struct trh_hdr) - slack); |
@@ -337,10 +338,11 @@ printk("source routing for %02X:%02X:%02X:%02X:%02X:%02X\n",trh->daddr[0], | |||
337 | static void tr_add_rif_info(struct trh_hdr *trh, struct net_device *dev) | 338 | static void tr_add_rif_info(struct trh_hdr *trh, struct net_device *dev) |
338 | { | 339 | { |
339 | unsigned int hash, rii_p = 0; | 340 | unsigned int hash, rii_p = 0; |
341 | unsigned long flags; | ||
340 | struct rif_cache *entry; | 342 | struct rif_cache *entry; |
341 | 343 | ||
342 | 344 | ||
343 | spin_lock_bh(&rif_lock); | 345 | spin_lock_irqsave(&rif_lock, flags); |
344 | 346 | ||
345 | /* | 347 | /* |
346 | * Firstly see if the entry exists | 348 | * Firstly see if the entry exists |
@@ -378,7 +380,7 @@ printk("adding rif_entry: addr:%02X:%02X:%02X:%02X:%02X:%02X rcf:%04X\n", | |||
378 | if(!entry) | 380 | if(!entry) |
379 | { | 381 | { |
380 | printk(KERN_DEBUG "tr.c: Couldn't malloc rif cache entry !\n"); | 382 | printk(KERN_DEBUG "tr.c: Couldn't malloc rif cache entry !\n"); |
381 | spin_unlock_bh(&rif_lock); | 383 | spin_unlock_irqrestore(&rif_lock, flags); |
382 | return; | 384 | return; |
383 | } | 385 | } |
384 | 386 | ||
@@ -420,7 +422,7 @@ printk("updating rif_entry: addr:%02X:%02X:%02X:%02X:%02X:%02X rcf:%04X\n", | |||
420 | } | 422 | } |
421 | entry->last_used=jiffies; | 423 | entry->last_used=jiffies; |
422 | } | 424 | } |
423 | spin_unlock_bh(&rif_lock); | 425 | spin_unlock_irqrestore(&rif_lock, flags); |
424 | } | 426 | } |
425 | 427 | ||
426 | /* | 428 | /* |
@@ -430,9 +432,9 @@ printk("updating rif_entry: addr:%02X:%02X:%02X:%02X:%02X:%02X rcf:%04X\n", | |||
430 | static void rif_check_expire(unsigned long dummy) | 432 | static void rif_check_expire(unsigned long dummy) |
431 | { | 433 | { |
432 | int i; | 434 | int i; |
433 | unsigned long next_interval = jiffies + sysctl_tr_rif_timeout/2; | 435 | unsigned long flags, next_interval = jiffies + sysctl_tr_rif_timeout/2; |
434 | 436 | ||
435 | spin_lock_bh(&rif_lock); | 437 | spin_lock_irqsave(&rif_lock, flags); |
436 | 438 | ||
437 | for(i =0; i < RIF_TABLE_SIZE; i++) { | 439 | for(i =0; i < RIF_TABLE_SIZE; i++) { |
438 | struct rif_cache *entry, **pentry; | 440 | struct rif_cache *entry, **pentry; |
@@ -454,7 +456,7 @@ static void rif_check_expire(unsigned long dummy) | |||
454 | } | 456 | } |
455 | } | 457 | } |
456 | 458 | ||
457 | spin_unlock_bh(&rif_lock); | 459 | spin_unlock_irqrestore(&rif_lock, flags); |
458 | 460 | ||
459 | mod_timer(&rif_timer, next_interval); | 461 | mod_timer(&rif_timer, next_interval); |
460 | 462 | ||
@@ -485,7 +487,7 @@ static struct rif_cache *rif_get_idx(loff_t pos) | |||
485 | 487 | ||
486 | static void *rif_seq_start(struct seq_file *seq, loff_t *pos) | 488 | static void *rif_seq_start(struct seq_file *seq, loff_t *pos) |
487 | { | 489 | { |
488 | spin_lock_bh(&rif_lock); | 490 | spin_lock_irq(&rif_lock); |
489 | 491 | ||
490 | return *pos ? rif_get_idx(*pos - 1) : SEQ_START_TOKEN; | 492 | return *pos ? rif_get_idx(*pos - 1) : SEQ_START_TOKEN; |
491 | } | 493 | } |
@@ -516,7 +518,7 @@ static void *rif_seq_next(struct seq_file *seq, void *v, loff_t *pos) | |||
516 | 518 | ||
517 | static void rif_seq_stop(struct seq_file *seq, void *v) | 519 | static void rif_seq_stop(struct seq_file *seq, void *v) |
518 | { | 520 | { |
519 | spin_unlock_bh(&rif_lock); | 521 | spin_unlock_irq(&rif_lock); |
520 | } | 522 | } |
521 | 523 | ||
522 | static int rif_seq_show(struct seq_file *seq, void *v) | 524 | static int rif_seq_show(struct seq_file *seq, void *v) |
diff --git a/net/Kconfig b/net/Kconfig index 2684e809a649..40a31ba86d2c 100644 --- a/net/Kconfig +++ b/net/Kconfig | |||
@@ -209,22 +209,6 @@ endmenu | |||
209 | 209 | ||
210 | endmenu | 210 | endmenu |
211 | 211 | ||
212 | config NETPOLL | ||
213 | def_bool NETCONSOLE | ||
214 | |||
215 | config NETPOLL_RX | ||
216 | bool "Netpoll support for trapping incoming packets" | ||
217 | default n | ||
218 | depends on NETPOLL | ||
219 | |||
220 | config NETPOLL_TRAP | ||
221 | bool "Netpoll traffic trapping" | ||
222 | default n | ||
223 | depends on NETPOLL | ||
224 | |||
225 | config NET_POLL_CONTROLLER | ||
226 | def_bool NETPOLL | ||
227 | |||
228 | source "net/ax25/Kconfig" | 212 | source "net/ax25/Kconfig" |
229 | source "net/irda/Kconfig" | 213 | source "net/irda/Kconfig" |
230 | source "net/bluetooth/Kconfig" | 214 | source "net/bluetooth/Kconfig" |
diff --git a/net/atm/Kconfig b/net/atm/Kconfig index bea2426229b1..21ff276b2d80 100644 --- a/net/atm/Kconfig +++ b/net/atm/Kconfig | |||
@@ -60,7 +60,7 @@ config ATM_BR2684 | |||
60 | tristate "RFC1483/2684 Bridged protocols" | 60 | tristate "RFC1483/2684 Bridged protocols" |
61 | depends on ATM && INET | 61 | depends on ATM && INET |
62 | help | 62 | help |
63 | ATM PVCs can carry ethernet PDUs according to rfc2684 (formerly 1483) | 63 | ATM PVCs can carry ethernet PDUs according to RFC2684 (formerly 1483) |
64 | This device will act like an ethernet from the kernels point of view, | 64 | This device will act like an ethernet from the kernels point of view, |
65 | with the traffic being carried by ATM PVCs (currently 1 PVC/device). | 65 | with the traffic being carried by ATM PVCs (currently 1 PVC/device). |
66 | This is sometimes used over DSL lines. If in doubt, say N. | 66 | This is sometimes used over DSL lines. If in doubt, say N. |
@@ -69,6 +69,6 @@ config ATM_BR2684_IPFILTER | |||
69 | bool "Per-VC IP filter kludge" | 69 | bool "Per-VC IP filter kludge" |
70 | depends on ATM_BR2684 | 70 | depends on ATM_BR2684 |
71 | help | 71 | help |
72 | This is an experimental mechanism for users who need to terminating a | 72 | This is an experimental mechanism for users who need to terminate a |
73 | large number of IP-only vcc's. Do not enable this unless you are sure | 73 | large number of IP-only vcc's. Do not enable this unless you are sure |
74 | you know what you are doing. | 74 | you know what you are doing. |
diff --git a/net/atm/svc.c b/net/atm/svc.c index 02f5374a51f2..08e46052a3e4 100644 --- a/net/atm/svc.c +++ b/net/atm/svc.c | |||
@@ -118,10 +118,6 @@ static int svc_bind(struct socket *sock,struct sockaddr *sockaddr, | |||
118 | goto out; | 118 | goto out; |
119 | } | 119 | } |
120 | vcc = ATM_SD(sock); | 120 | vcc = ATM_SD(sock); |
121 | if (test_bit(ATM_VF_SESSION, &vcc->flags)) { | ||
122 | error = -EINVAL; | ||
123 | goto out; | ||
124 | } | ||
125 | addr = (struct sockaddr_atmsvc *) sockaddr; | 121 | addr = (struct sockaddr_atmsvc *) sockaddr; |
126 | if (addr->sas_family != AF_ATMSVC) { | 122 | if (addr->sas_family != AF_ATMSVC) { |
127 | error = -EAFNOSUPPORT; | 123 | error = -EAFNOSUPPORT; |
diff --git a/net/ax25/af_ax25.c b/net/ax25/af_ax25.c index 707097deac3d..a5c94f11547c 100644 --- a/net/ax25/af_ax25.c +++ b/net/ax25/af_ax25.c | |||
@@ -875,12 +875,7 @@ struct sock *ax25_make_new(struct sock *osk, struct ax25_dev *ax25_dev) | |||
875 | sk->sk_sndbuf = osk->sk_sndbuf; | 875 | sk->sk_sndbuf = osk->sk_sndbuf; |
876 | sk->sk_state = TCP_ESTABLISHED; | 876 | sk->sk_state = TCP_ESTABLISHED; |
877 | sk->sk_sleep = osk->sk_sleep; | 877 | sk->sk_sleep = osk->sk_sleep; |
878 | 878 | sock_copy_flags(sk, osk); | |
879 | if (sock_flag(osk, SOCK_DBG)) | ||
880 | sock_set_flag(sk, SOCK_DBG); | ||
881 | |||
882 | if (sock_flag(osk, SOCK_ZAPPED)) | ||
883 | sock_set_flag(sk, SOCK_ZAPPED); | ||
884 | 879 | ||
885 | oax25 = ax25_sk(osk); | 880 | oax25 = ax25_sk(osk); |
886 | 881 | ||
@@ -1007,7 +1002,8 @@ static int ax25_bind(struct socket *sock, struct sockaddr *uaddr, int addr_len) | |||
1007 | struct sock *sk = sock->sk; | 1002 | struct sock *sk = sock->sk; |
1008 | struct full_sockaddr_ax25 *addr = (struct full_sockaddr_ax25 *)uaddr; | 1003 | struct full_sockaddr_ax25 *addr = (struct full_sockaddr_ax25 *)uaddr; |
1009 | ax25_dev *ax25_dev = NULL; | 1004 | ax25_dev *ax25_dev = NULL; |
1010 | ax25_address *call; | 1005 | ax25_uid_assoc *user; |
1006 | ax25_address call; | ||
1011 | ax25_cb *ax25; | 1007 | ax25_cb *ax25; |
1012 | int err = 0; | 1008 | int err = 0; |
1013 | 1009 | ||
@@ -1026,9 +1022,15 @@ static int ax25_bind(struct socket *sock, struct sockaddr *uaddr, int addr_len) | |||
1026 | if (addr->fsa_ax25.sax25_family != AF_AX25) | 1022 | if (addr->fsa_ax25.sax25_family != AF_AX25) |
1027 | return -EINVAL; | 1023 | return -EINVAL; |
1028 | 1024 | ||
1029 | call = ax25_findbyuid(current->euid); | 1025 | user = ax25_findbyuid(current->euid); |
1030 | if (call == NULL && ax25_uid_policy && !capable(CAP_NET_ADMIN)) { | 1026 | if (user) { |
1031 | return -EACCES; | 1027 | call = user->call; |
1028 | ax25_uid_put(user); | ||
1029 | } else { | ||
1030 | if (ax25_uid_policy && !capable(CAP_NET_ADMIN)) | ||
1031 | return -EACCES; | ||
1032 | |||
1033 | call = addr->fsa_ax25.sax25_call; | ||
1032 | } | 1034 | } |
1033 | 1035 | ||
1034 | lock_sock(sk); | 1036 | lock_sock(sk); |
@@ -1039,10 +1041,7 @@ static int ax25_bind(struct socket *sock, struct sockaddr *uaddr, int addr_len) | |||
1039 | goto out; | 1041 | goto out; |
1040 | } | 1042 | } |
1041 | 1043 | ||
1042 | if (call == NULL) | 1044 | ax25->source_addr = call; |
1043 | ax25->source_addr = addr->fsa_ax25.sax25_call; | ||
1044 | else | ||
1045 | ax25->source_addr = *call; | ||
1046 | 1045 | ||
1047 | /* | 1046 | /* |
1048 | * User already set interface with SO_BINDTODEVICE | 1047 | * User already set interface with SO_BINDTODEVICE |
diff --git a/net/ax25/ax25_route.c b/net/ax25/ax25_route.c index 44b99b1ff9f8..c288526da4ce 100644 --- a/net/ax25/ax25_route.c +++ b/net/ax25/ax25_route.c | |||
@@ -422,8 +422,8 @@ static inline void ax25_adjust_path(ax25_address *addr, ax25_digi *digipeat) | |||
422 | */ | 422 | */ |
423 | int ax25_rt_autobind(ax25_cb *ax25, ax25_address *addr) | 423 | int ax25_rt_autobind(ax25_cb *ax25, ax25_address *addr) |
424 | { | 424 | { |
425 | ax25_uid_assoc *user; | ||
425 | ax25_route *ax25_rt; | 426 | ax25_route *ax25_rt; |
426 | ax25_address *call; | ||
427 | int err; | 427 | int err; |
428 | 428 | ||
429 | if ((ax25_rt = ax25_get_route(addr, NULL)) == NULL) | 429 | if ((ax25_rt = ax25_get_route(addr, NULL)) == NULL) |
@@ -434,16 +434,18 @@ int ax25_rt_autobind(ax25_cb *ax25, ax25_address *addr) | |||
434 | goto put; | 434 | goto put; |
435 | } | 435 | } |
436 | 436 | ||
437 | if ((call = ax25_findbyuid(current->euid)) == NULL) { | 437 | user = ax25_findbyuid(current->euid); |
438 | if (user) { | ||
439 | ax25->source_addr = user->call; | ||
440 | ax25_uid_put(user); | ||
441 | } else { | ||
438 | if (ax25_uid_policy && !capable(CAP_NET_BIND_SERVICE)) { | 442 | if (ax25_uid_policy && !capable(CAP_NET_BIND_SERVICE)) { |
439 | err = -EPERM; | 443 | err = -EPERM; |
440 | goto put; | 444 | goto put; |
441 | } | 445 | } |
442 | call = (ax25_address *)ax25->ax25_dev->dev->dev_addr; | 446 | ax25->source_addr = *(ax25_address *)ax25->ax25_dev->dev->dev_addr; |
443 | } | 447 | } |
444 | 448 | ||
445 | ax25->source_addr = *call; | ||
446 | |||
447 | if (ax25_rt->digipeat != NULL) { | 449 | if (ax25_rt->digipeat != NULL) { |
448 | if ((ax25->digipeat = kmalloc(sizeof(ax25_digi), GFP_ATOMIC)) == NULL) { | 450 | if ((ax25->digipeat = kmalloc(sizeof(ax25_digi), GFP_ATOMIC)) == NULL) { |
449 | err = -ENOMEM; | 451 | err = -ENOMEM; |
diff --git a/net/ax25/ax25_uid.c b/net/ax25/ax25_uid.c index cea6b7d19729..a8b3822f3ee4 100644 --- a/net/ax25/ax25_uid.c +++ b/net/ax25/ax25_uid.c | |||
@@ -28,6 +28,7 @@ | |||
28 | #include <linux/fcntl.h> | 28 | #include <linux/fcntl.h> |
29 | #include <linux/mm.h> | 29 | #include <linux/mm.h> |
30 | #include <linux/interrupt.h> | 30 | #include <linux/interrupt.h> |
31 | #include <linux/list.h> | ||
31 | #include <linux/notifier.h> | 32 | #include <linux/notifier.h> |
32 | #include <linux/proc_fs.h> | 33 | #include <linux/proc_fs.h> |
33 | #include <linux/seq_file.h> | 34 | #include <linux/seq_file.h> |
@@ -41,38 +42,41 @@ | |||
41 | * Callsign/UID mapper. This is in kernel space for security on multi-amateur machines. | 42 | * Callsign/UID mapper. This is in kernel space for security on multi-amateur machines. |
42 | */ | 43 | */ |
43 | 44 | ||
44 | static ax25_uid_assoc *ax25_uid_list; | 45 | HLIST_HEAD(ax25_uid_list); |
45 | static DEFINE_RWLOCK(ax25_uid_lock); | 46 | static DEFINE_RWLOCK(ax25_uid_lock); |
46 | 47 | ||
47 | int ax25_uid_policy = 0; | 48 | int ax25_uid_policy = 0; |
48 | 49 | ||
49 | ax25_address *ax25_findbyuid(uid_t uid) | 50 | ax25_uid_assoc *ax25_findbyuid(uid_t uid) |
50 | { | 51 | { |
51 | ax25_uid_assoc *ax25_uid; | 52 | ax25_uid_assoc *ax25_uid, *res = NULL; |
52 | ax25_address *res = NULL; | 53 | struct hlist_node *node; |
53 | 54 | ||
54 | read_lock(&ax25_uid_lock); | 55 | read_lock(&ax25_uid_lock); |
55 | for (ax25_uid = ax25_uid_list; ax25_uid != NULL; ax25_uid = ax25_uid->next) { | 56 | ax25_uid_for_each(ax25_uid, node, &ax25_uid_list) { |
56 | if (ax25_uid->uid == uid) { | 57 | if (ax25_uid->uid == uid) { |
57 | res = &ax25_uid->call; | 58 | ax25_uid_hold(ax25_uid); |
59 | res = ax25_uid; | ||
58 | break; | 60 | break; |
59 | } | 61 | } |
60 | } | 62 | } |
61 | read_unlock(&ax25_uid_lock); | 63 | read_unlock(&ax25_uid_lock); |
62 | 64 | ||
63 | return NULL; | 65 | return res; |
64 | } | 66 | } |
65 | 67 | ||
66 | int ax25_uid_ioctl(int cmd, struct sockaddr_ax25 *sax) | 68 | int ax25_uid_ioctl(int cmd, struct sockaddr_ax25 *sax) |
67 | { | 69 | { |
68 | ax25_uid_assoc *s, *ax25_uid; | 70 | ax25_uid_assoc *ax25_uid; |
71 | struct hlist_node *node; | ||
72 | ax25_uid_assoc *user; | ||
69 | unsigned long res; | 73 | unsigned long res; |
70 | 74 | ||
71 | switch (cmd) { | 75 | switch (cmd) { |
72 | case SIOCAX25GETUID: | 76 | case SIOCAX25GETUID: |
73 | res = -ENOENT; | 77 | res = -ENOENT; |
74 | read_lock(&ax25_uid_lock); | 78 | read_lock(&ax25_uid_lock); |
75 | for (ax25_uid = ax25_uid_list; ax25_uid != NULL; ax25_uid = ax25_uid->next) { | 79 | ax25_uid_for_each(ax25_uid, node, &ax25_uid_list) { |
76 | if (ax25cmp(&sax->sax25_call, &ax25_uid->call) == 0) { | 80 | if (ax25cmp(&sax->sax25_call, &ax25_uid->call) == 0) { |
77 | res = ax25_uid->uid; | 81 | res = ax25_uid->uid; |
78 | break; | 82 | break; |
@@ -85,19 +89,22 @@ int ax25_uid_ioctl(int cmd, struct sockaddr_ax25 *sax) | |||
85 | case SIOCAX25ADDUID: | 89 | case SIOCAX25ADDUID: |
86 | if (!capable(CAP_NET_ADMIN)) | 90 | if (!capable(CAP_NET_ADMIN)) |
87 | return -EPERM; | 91 | return -EPERM; |
88 | if (ax25_findbyuid(sax->sax25_uid)) | 92 | user = ax25_findbyuid(sax->sax25_uid); |
93 | if (user) { | ||
94 | ax25_uid_put(user); | ||
89 | return -EEXIST; | 95 | return -EEXIST; |
96 | } | ||
90 | if (sax->sax25_uid == 0) | 97 | if (sax->sax25_uid == 0) |
91 | return -EINVAL; | 98 | return -EINVAL; |
92 | if ((ax25_uid = kmalloc(sizeof(*ax25_uid), GFP_KERNEL)) == NULL) | 99 | if ((ax25_uid = kmalloc(sizeof(*ax25_uid), GFP_KERNEL)) == NULL) |
93 | return -ENOMEM; | 100 | return -ENOMEM; |
94 | 101 | ||
102 | atomic_set(&ax25_uid->refcount, 1); | ||
95 | ax25_uid->uid = sax->sax25_uid; | 103 | ax25_uid->uid = sax->sax25_uid; |
96 | ax25_uid->call = sax->sax25_call; | 104 | ax25_uid->call = sax->sax25_call; |
97 | 105 | ||
98 | write_lock(&ax25_uid_lock); | 106 | write_lock(&ax25_uid_lock); |
99 | ax25_uid->next = ax25_uid_list; | 107 | hlist_add_head(&ax25_uid->uid_node, &ax25_uid_list); |
100 | ax25_uid_list = ax25_uid; | ||
101 | write_unlock(&ax25_uid_lock); | 108 | write_unlock(&ax25_uid_lock); |
102 | 109 | ||
103 | return 0; | 110 | return 0; |
@@ -106,34 +113,21 @@ int ax25_uid_ioctl(int cmd, struct sockaddr_ax25 *sax) | |||
106 | if (!capable(CAP_NET_ADMIN)) | 113 | if (!capable(CAP_NET_ADMIN)) |
107 | return -EPERM; | 114 | return -EPERM; |
108 | 115 | ||
116 | ax25_uid = NULL; | ||
109 | write_lock(&ax25_uid_lock); | 117 | write_lock(&ax25_uid_lock); |
110 | for (ax25_uid = ax25_uid_list; ax25_uid != NULL; ax25_uid = ax25_uid->next) { | 118 | ax25_uid_for_each(ax25_uid, node, &ax25_uid_list) { |
111 | if (ax25cmp(&sax->sax25_call, &ax25_uid->call) == 0) { | 119 | if (ax25cmp(&sax->sax25_call, &ax25_uid->call) == 0) |
112 | break; | 120 | break; |
113 | } | ||
114 | } | 121 | } |
115 | if (ax25_uid == NULL) { | 122 | if (ax25_uid == NULL) { |
116 | write_unlock(&ax25_uid_lock); | 123 | write_unlock(&ax25_uid_lock); |
117 | return -ENOENT; | 124 | return -ENOENT; |
118 | } | 125 | } |
119 | if ((s = ax25_uid_list) == ax25_uid) { | 126 | hlist_del_init(&ax25_uid->uid_node); |
120 | ax25_uid_list = s->next; | 127 | ax25_uid_put(ax25_uid); |
121 | write_unlock(&ax25_uid_lock); | ||
122 | kfree(ax25_uid); | ||
123 | return 0; | ||
124 | } | ||
125 | while (s != NULL && s->next != NULL) { | ||
126 | if (s->next == ax25_uid) { | ||
127 | s->next = ax25_uid->next; | ||
128 | write_unlock(&ax25_uid_lock); | ||
129 | kfree(ax25_uid); | ||
130 | return 0; | ||
131 | } | ||
132 | s = s->next; | ||
133 | } | ||
134 | write_unlock(&ax25_uid_lock); | 128 | write_unlock(&ax25_uid_lock); |
135 | 129 | ||
136 | return -ENOENT; | 130 | return 0; |
137 | 131 | ||
138 | default: | 132 | default: |
139 | return -EINVAL; | 133 | return -EINVAL; |
@@ -147,13 +141,11 @@ int ax25_uid_ioctl(int cmd, struct sockaddr_ax25 *sax) | |||
147 | static void *ax25_uid_seq_start(struct seq_file *seq, loff_t *pos) | 141 | static void *ax25_uid_seq_start(struct seq_file *seq, loff_t *pos) |
148 | { | 142 | { |
149 | struct ax25_uid_assoc *pt; | 143 | struct ax25_uid_assoc *pt; |
150 | int i = 1; | 144 | struct hlist_node *node; |
145 | int i = 0; | ||
151 | 146 | ||
152 | read_lock(&ax25_uid_lock); | 147 | read_lock(&ax25_uid_lock); |
153 | if (*pos == 0) | 148 | ax25_uid_for_each(pt, node, &ax25_uid_list) { |
154 | return SEQ_START_TOKEN; | ||
155 | |||
156 | for (pt = ax25_uid_list; pt != NULL; pt = pt->next) { | ||
157 | if (i == *pos) | 149 | if (i == *pos) |
158 | return pt; | 150 | return pt; |
159 | ++i; | 151 | ++i; |
@@ -164,8 +156,9 @@ static void *ax25_uid_seq_start(struct seq_file *seq, loff_t *pos) | |||
164 | static void *ax25_uid_seq_next(struct seq_file *seq, void *v, loff_t *pos) | 156 | static void *ax25_uid_seq_next(struct seq_file *seq, void *v, loff_t *pos) |
165 | { | 157 | { |
166 | ++*pos; | 158 | ++*pos; |
167 | return (v == SEQ_START_TOKEN) ? ax25_uid_list : | 159 | |
168 | ((struct ax25_uid_assoc *) v)->next; | 160 | return hlist_entry(((ax25_uid_assoc *)v)->uid_node.next, |
161 | ax25_uid_assoc, uid_node); | ||
169 | } | 162 | } |
170 | 163 | ||
171 | static void ax25_uid_seq_stop(struct seq_file *seq, void *v) | 164 | static void ax25_uid_seq_stop(struct seq_file *seq, void *v) |
@@ -179,7 +172,6 @@ static int ax25_uid_seq_show(struct seq_file *seq, void *v) | |||
179 | seq_printf(seq, "Policy: %d\n", ax25_uid_policy); | 172 | seq_printf(seq, "Policy: %d\n", ax25_uid_policy); |
180 | else { | 173 | else { |
181 | struct ax25_uid_assoc *pt = v; | 174 | struct ax25_uid_assoc *pt = v; |
182 | |||
183 | 175 | ||
184 | seq_printf(seq, "%6d %s\n", pt->uid, ax2asc(&pt->call)); | 176 | seq_printf(seq, "%6d %s\n", pt->uid, ax2asc(&pt->call)); |
185 | } | 177 | } |
@@ -213,16 +205,13 @@ struct file_operations ax25_uid_fops = { | |||
213 | */ | 205 | */ |
214 | void __exit ax25_uid_free(void) | 206 | void __exit ax25_uid_free(void) |
215 | { | 207 | { |
216 | ax25_uid_assoc *s, *ax25_uid; | 208 | ax25_uid_assoc *ax25_uid; |
209 | struct hlist_node *node; | ||
217 | 210 | ||
218 | write_lock(&ax25_uid_lock); | 211 | write_lock(&ax25_uid_lock); |
219 | ax25_uid = ax25_uid_list; | 212 | ax25_uid_for_each(ax25_uid, node, &ax25_uid_list) { |
220 | while (ax25_uid != NULL) { | 213 | hlist_del_init(&ax25_uid->uid_node); |
221 | s = ax25_uid; | 214 | ax25_uid_put(ax25_uid); |
222 | ax25_uid = ax25_uid->next; | ||
223 | |||
224 | kfree(s); | ||
225 | } | 215 | } |
226 | ax25_uid_list = NULL; | ||
227 | write_unlock(&ax25_uid_lock); | 216 | write_unlock(&ax25_uid_lock); |
228 | } | 217 | } |
diff --git a/net/bluetooth/hci_core.c b/net/bluetooth/hci_core.c index fb5524365bc2..ffa26c10bfe8 100644 --- a/net/bluetooth/hci_core.c +++ b/net/bluetooth/hci_core.c | |||
@@ -299,7 +299,6 @@ struct hci_dev *hci_dev_get(int index) | |||
299 | read_unlock(&hci_dev_list_lock); | 299 | read_unlock(&hci_dev_list_lock); |
300 | return hdev; | 300 | return hdev; |
301 | } | 301 | } |
302 | EXPORT_SYMBOL(hci_dev_get); | ||
303 | 302 | ||
304 | /* ---- Inquiry support ---- */ | 303 | /* ---- Inquiry support ---- */ |
305 | static void inquiry_cache_flush(struct hci_dev *hdev) | 304 | static void inquiry_cache_flush(struct hci_dev *hdev) |
@@ -1042,7 +1041,6 @@ int hci_send_cmd(struct hci_dev *hdev, __u16 ogf, __u16 ocf, __u32 plen, void *p | |||
1042 | 1041 | ||
1043 | return 0; | 1042 | return 0; |
1044 | } | 1043 | } |
1045 | EXPORT_SYMBOL(hci_send_cmd); | ||
1046 | 1044 | ||
1047 | /* Get data from the previously sent command */ | 1045 | /* Get data from the previously sent command */ |
1048 | void *hci_sent_cmd_data(struct hci_dev *hdev, __u16 ogf, __u16 ocf) | 1046 | void *hci_sent_cmd_data(struct hci_dev *hdev, __u16 ogf, __u16 ocf) |
diff --git a/net/bluetooth/hci_event.c b/net/bluetooth/hci_event.c index c4b592b4ef10..46367bd129c3 100644 --- a/net/bluetooth/hci_event.c +++ b/net/bluetooth/hci_event.c | |||
@@ -1035,9 +1035,11 @@ void hci_si_event(struct hci_dev *hdev, int type, int dlen, void *data) | |||
1035 | ev->type = type; | 1035 | ev->type = type; |
1036 | memcpy(ev->data, data, dlen); | 1036 | memcpy(ev->data, data, dlen); |
1037 | 1037 | ||
1038 | bt_cb(skb)->incoming = 1; | ||
1039 | do_gettimeofday(&skb->stamp); | ||
1040 | |||
1038 | skb->pkt_type = HCI_EVENT_PKT; | 1041 | skb->pkt_type = HCI_EVENT_PKT; |
1039 | skb->dev = (void *) hdev; | 1042 | skb->dev = (void *) hdev; |
1040 | hci_send_to_sock(hdev, skb); | 1043 | hci_send_to_sock(hdev, skb); |
1041 | kfree_skb(skb); | 1044 | kfree_skb(skb); |
1042 | } | 1045 | } |
1043 | EXPORT_SYMBOL(hci_si_event); | ||
diff --git a/net/bluetooth/lib.c b/net/bluetooth/lib.c index 9efb0a093612..ee6a66979913 100644 --- a/net/bluetooth/lib.c +++ b/net/bluetooth/lib.c | |||
@@ -34,31 +34,6 @@ | |||
34 | 34 | ||
35 | #include <net/bluetooth/bluetooth.h> | 35 | #include <net/bluetooth/bluetooth.h> |
36 | 36 | ||
37 | void bt_dump(char *pref, __u8 *buf, int count) | ||
38 | { | ||
39 | char *ptr; | ||
40 | char line[100]; | ||
41 | unsigned int i; | ||
42 | |||
43 | printk(KERN_INFO "%s: dump, len %d\n", pref, count); | ||
44 | |||
45 | ptr = line; | ||
46 | *ptr = 0; | ||
47 | for (i = 0; i < count; i++) { | ||
48 | ptr += sprintf(ptr, " %2.2X", buf[i]); | ||
49 | |||
50 | if (i && !((i + 1) % 20)) { | ||
51 | printk(KERN_INFO "%s:%s\n", pref, line); | ||
52 | ptr = line; | ||
53 | *ptr = 0; | ||
54 | } | ||
55 | } | ||
56 | |||
57 | if (line[0]) | ||
58 | printk(KERN_INFO "%s:%s\n", pref, line); | ||
59 | } | ||
60 | EXPORT_SYMBOL(bt_dump); | ||
61 | |||
62 | void baswap(bdaddr_t *dst, bdaddr_t *src) | 37 | void baswap(bdaddr_t *dst, bdaddr_t *src) |
63 | { | 38 | { |
64 | unsigned char *d = (unsigned char *) dst; | 39 | unsigned char *d = (unsigned char *) dst; |
diff --git a/net/bluetooth/rfcomm/core.c b/net/bluetooth/rfcomm/core.c index e9e6fda66f1a..27bf5047cd33 100644 --- a/net/bluetooth/rfcomm/core.c +++ b/net/bluetooth/rfcomm/core.c | |||
@@ -389,8 +389,6 @@ static int __rfcomm_dlc_close(struct rfcomm_dlc *d, int err) | |||
389 | rfcomm_dlc_unlock(d); | 389 | rfcomm_dlc_unlock(d); |
390 | 390 | ||
391 | skb_queue_purge(&d->tx_queue); | 391 | skb_queue_purge(&d->tx_queue); |
392 | rfcomm_session_put(s); | ||
393 | |||
394 | rfcomm_dlc_unlink(d); | 392 | rfcomm_dlc_unlink(d); |
395 | } | 393 | } |
396 | 394 | ||
@@ -600,8 +598,6 @@ static struct rfcomm_session *rfcomm_session_create(bdaddr_t *src, bdaddr_t *dst | |||
600 | goto failed; | 598 | goto failed; |
601 | } | 599 | } |
602 | 600 | ||
603 | rfcomm_session_hold(s); | ||
604 | |||
605 | s->initiator = 1; | 601 | s->initiator = 1; |
606 | 602 | ||
607 | bacpy(&addr.l2_bdaddr, dst); | 603 | bacpy(&addr.l2_bdaddr, dst); |
diff --git a/net/bridge/netfilter/Kconfig b/net/bridge/netfilter/Kconfig index 68ccef507b49..c70b3be23026 100644 --- a/net/bridge/netfilter/Kconfig +++ b/net/bridge/netfilter/Kconfig | |||
@@ -138,7 +138,7 @@ config BRIDGE_EBT_VLAN | |||
138 | # | 138 | # |
139 | config BRIDGE_EBT_ARPREPLY | 139 | config BRIDGE_EBT_ARPREPLY |
140 | tristate "ebt: arp reply target support" | 140 | tristate "ebt: arp reply target support" |
141 | depends on BRIDGE_NF_EBTABLES | 141 | depends on BRIDGE_NF_EBTABLES && INET |
142 | help | 142 | help |
143 | This option adds the arp reply target, which allows | 143 | This option adds the arp reply target, which allows |
144 | automatically sending arp replies to arp requests. | 144 | automatically sending arp replies to arp requests. |
diff --git a/net/compat.c b/net/compat.c index be5d936dc423..d99ab9695893 100644 --- a/net/compat.c +++ b/net/compat.c | |||
@@ -91,20 +91,11 @@ int verify_compat_iovec(struct msghdr *kern_msg, struct iovec *kern_iov, | |||
91 | } else | 91 | } else |
92 | kern_msg->msg_name = NULL; | 92 | kern_msg->msg_name = NULL; |
93 | 93 | ||
94 | if(kern_msg->msg_iovlen > UIO_FASTIOV) { | ||
95 | kern_iov = kmalloc(kern_msg->msg_iovlen * sizeof(struct iovec), | ||
96 | GFP_KERNEL); | ||
97 | if(!kern_iov) | ||
98 | return -ENOMEM; | ||
99 | } | ||
100 | |||
101 | tot_len = iov_from_user_compat_to_kern(kern_iov, | 94 | tot_len = iov_from_user_compat_to_kern(kern_iov, |
102 | (struct compat_iovec __user *)kern_msg->msg_iov, | 95 | (struct compat_iovec __user *)kern_msg->msg_iov, |
103 | kern_msg->msg_iovlen); | 96 | kern_msg->msg_iovlen); |
104 | if(tot_len >= 0) | 97 | if(tot_len >= 0) |
105 | kern_msg->msg_iov = kern_iov; | 98 | kern_msg->msg_iov = kern_iov; |
106 | else if(kern_msg->msg_iovlen > UIO_FASTIOV) | ||
107 | kfree(kern_iov); | ||
108 | 99 | ||
109 | return tot_len; | 100 | return tot_len; |
110 | } | 101 | } |
diff --git a/net/core/Makefile b/net/core/Makefile index 5e0c56b7f607..f5f5e58943e8 100644 --- a/net/core/Makefile +++ b/net/core/Makefile | |||
@@ -7,9 +7,10 @@ obj-y := sock.o request_sock.o skbuff.o iovec.o datagram.o stream.o scm.o \ | |||
7 | 7 | ||
8 | obj-$(CONFIG_SYSCTL) += sysctl_net_core.o | 8 | obj-$(CONFIG_SYSCTL) += sysctl_net_core.o |
9 | 9 | ||
10 | obj-y += flow.o dev.o ethtool.o dev_mcast.o dst.o \ | 10 | obj-y += dev.o ethtool.o dev_mcast.o dst.o \ |
11 | neighbour.o rtnetlink.o utils.o link_watch.o filter.o | 11 | neighbour.o rtnetlink.o utils.o link_watch.o filter.o |
12 | 12 | ||
13 | obj-$(CONFIG_XFRM) += flow.o | ||
13 | obj-$(CONFIG_SYSFS) += net-sysfs.o | 14 | obj-$(CONFIG_SYSFS) += net-sysfs.o |
14 | obj-$(CONFIG_NETFILTER) += netfilter.o | 15 | obj-$(CONFIG_NETFILTER) += netfilter.o |
15 | obj-$(CONFIG_NET_DIVERT) += dv.o | 16 | obj-$(CONFIG_NET_DIVERT) += dv.o |
diff --git a/net/core/dev.c b/net/core/dev.c index ff9dc029233a..faf59b02c4bf 100644 --- a/net/core/dev.c +++ b/net/core/dev.c | |||
@@ -901,8 +901,7 @@ int dev_close(struct net_device *dev) | |||
901 | smp_mb__after_clear_bit(); /* Commit netif_running(). */ | 901 | smp_mb__after_clear_bit(); /* Commit netif_running(). */ |
902 | while (test_bit(__LINK_STATE_RX_SCHED, &dev->state)) { | 902 | while (test_bit(__LINK_STATE_RX_SCHED, &dev->state)) { |
903 | /* No hurry. */ | 903 | /* No hurry. */ |
904 | current->state = TASK_INTERRUPTIBLE; | 904 | msleep(1); |
905 | schedule_timeout(1); | ||
906 | } | 905 | } |
907 | 906 | ||
908 | /* | 907 | /* |
@@ -1697,7 +1696,8 @@ static void net_rx_action(struct softirq_action *h) | |||
1697 | struct softnet_data *queue = &__get_cpu_var(softnet_data); | 1696 | struct softnet_data *queue = &__get_cpu_var(softnet_data); |
1698 | unsigned long start_time = jiffies; | 1697 | unsigned long start_time = jiffies; |
1699 | int budget = netdev_budget; | 1698 | int budget = netdev_budget; |
1700 | 1699 | void *have; | |
1700 | |||
1701 | local_irq_disable(); | 1701 | local_irq_disable(); |
1702 | 1702 | ||
1703 | while (!list_empty(&queue->poll_list)) { | 1703 | while (!list_empty(&queue->poll_list)) { |
@@ -1710,10 +1710,10 @@ static void net_rx_action(struct softirq_action *h) | |||
1710 | 1710 | ||
1711 | dev = list_entry(queue->poll_list.next, | 1711 | dev = list_entry(queue->poll_list.next, |
1712 | struct net_device, poll_list); | 1712 | struct net_device, poll_list); |
1713 | netpoll_poll_lock(dev); | 1713 | have = netpoll_poll_lock(dev); |
1714 | 1714 | ||
1715 | if (dev->quota <= 0 || dev->poll(dev, &budget)) { | 1715 | if (dev->quota <= 0 || dev->poll(dev, &budget)) { |
1716 | netpoll_poll_unlock(dev); | 1716 | netpoll_poll_unlock(have); |
1717 | local_irq_disable(); | 1717 | local_irq_disable(); |
1718 | list_del(&dev->poll_list); | 1718 | list_del(&dev->poll_list); |
1719 | list_add_tail(&dev->poll_list, &queue->poll_list); | 1719 | list_add_tail(&dev->poll_list, &queue->poll_list); |
@@ -1722,7 +1722,7 @@ static void net_rx_action(struct softirq_action *h) | |||
1722 | else | 1722 | else |
1723 | dev->quota = dev->weight; | 1723 | dev->quota = dev->weight; |
1724 | } else { | 1724 | } else { |
1725 | netpoll_poll_unlock(dev); | 1725 | netpoll_poll_unlock(have); |
1726 | dev_put(dev); | 1726 | dev_put(dev); |
1727 | local_irq_disable(); | 1727 | local_irq_disable(); |
1728 | } | 1728 | } |
diff --git a/net/core/dst.c b/net/core/dst.c index fc434ade5270..334790da9f16 100644 --- a/net/core/dst.c +++ b/net/core/dst.c | |||
@@ -45,6 +45,7 @@ static struct timer_list dst_gc_timer = | |||
45 | static void dst_run_gc(unsigned long dummy) | 45 | static void dst_run_gc(unsigned long dummy) |
46 | { | 46 | { |
47 | int delayed = 0; | 47 | int delayed = 0; |
48 | int work_performed; | ||
48 | struct dst_entry * dst, **dstp; | 49 | struct dst_entry * dst, **dstp; |
49 | 50 | ||
50 | if (!spin_trylock(&dst_lock)) { | 51 | if (!spin_trylock(&dst_lock)) { |
@@ -52,9 +53,9 @@ static void dst_run_gc(unsigned long dummy) | |||
52 | return; | 53 | return; |
53 | } | 54 | } |
54 | 55 | ||
55 | |||
56 | del_timer(&dst_gc_timer); | 56 | del_timer(&dst_gc_timer); |
57 | dstp = &dst_garbage_list; | 57 | dstp = &dst_garbage_list; |
58 | work_performed = 0; | ||
58 | while ((dst = *dstp) != NULL) { | 59 | while ((dst = *dstp) != NULL) { |
59 | if (atomic_read(&dst->__refcnt)) { | 60 | if (atomic_read(&dst->__refcnt)) { |
60 | dstp = &dst->next; | 61 | dstp = &dst->next; |
@@ -62,6 +63,7 @@ static void dst_run_gc(unsigned long dummy) | |||
62 | continue; | 63 | continue; |
63 | } | 64 | } |
64 | *dstp = dst->next; | 65 | *dstp = dst->next; |
66 | work_performed = 1; | ||
65 | 67 | ||
66 | dst = dst_destroy(dst); | 68 | dst = dst_destroy(dst); |
67 | if (dst) { | 69 | if (dst) { |
@@ -86,9 +88,14 @@ static void dst_run_gc(unsigned long dummy) | |||
86 | dst_gc_timer_inc = DST_GC_MAX; | 88 | dst_gc_timer_inc = DST_GC_MAX; |
87 | goto out; | 89 | goto out; |
88 | } | 90 | } |
89 | if ((dst_gc_timer_expires += dst_gc_timer_inc) > DST_GC_MAX) | 91 | if (!work_performed) { |
90 | dst_gc_timer_expires = DST_GC_MAX; | 92 | if ((dst_gc_timer_expires += dst_gc_timer_inc) > DST_GC_MAX) |
91 | dst_gc_timer_inc += DST_GC_INC; | 93 | dst_gc_timer_expires = DST_GC_MAX; |
94 | dst_gc_timer_inc += DST_GC_INC; | ||
95 | } else { | ||
96 | dst_gc_timer_inc = DST_GC_INC; | ||
97 | dst_gc_timer_expires = DST_GC_MIN; | ||
98 | } | ||
92 | dst_gc_timer.expires = jiffies + dst_gc_timer_expires; | 99 | dst_gc_timer.expires = jiffies + dst_gc_timer_expires; |
93 | #if RT_CACHE_DEBUG >= 2 | 100 | #if RT_CACHE_DEBUG >= 2 |
94 | printk("dst_total: %d/%d %ld\n", | 101 | printk("dst_total: %d/%d %ld\n", |
diff --git a/net/core/netpoll.c b/net/core/netpoll.c index c327c9edadc5..a1a9a7abff50 100644 --- a/net/core/netpoll.c +++ b/net/core/netpoll.c | |||
@@ -33,6 +33,7 @@ | |||
33 | #define MAX_UDP_CHUNK 1460 | 33 | #define MAX_UDP_CHUNK 1460 |
34 | #define MAX_SKBS 32 | 34 | #define MAX_SKBS 32 |
35 | #define MAX_QUEUE_DEPTH (MAX_SKBS / 2) | 35 | #define MAX_QUEUE_DEPTH (MAX_SKBS / 2) |
36 | #define MAX_RETRIES 20000 | ||
36 | 37 | ||
37 | static DEFINE_SPINLOCK(skb_list_lock); | 38 | static DEFINE_SPINLOCK(skb_list_lock); |
38 | static int nr_skbs; | 39 | static int nr_skbs; |
@@ -248,14 +249,14 @@ static void netpoll_send_skb(struct netpoll *np, struct sk_buff *skb) | |||
248 | int status; | 249 | int status; |
249 | struct netpoll_info *npinfo; | 250 | struct netpoll_info *npinfo; |
250 | 251 | ||
251 | repeat: | 252 | if (!np || !np->dev || !netif_running(np->dev)) { |
252 | if(!np || !np->dev || !netif_running(np->dev)) { | ||
253 | __kfree_skb(skb); | 253 | __kfree_skb(skb); |
254 | return; | 254 | return; |
255 | } | 255 | } |
256 | 256 | ||
257 | /* avoid recursion */ | ||
258 | npinfo = np->dev->npinfo; | 257 | npinfo = np->dev->npinfo; |
258 | |||
259 | /* avoid recursion */ | ||
259 | if (npinfo->poll_owner == smp_processor_id() || | 260 | if (npinfo->poll_owner == smp_processor_id() || |
260 | np->dev->xmit_lock_owner == smp_processor_id()) { | 261 | np->dev->xmit_lock_owner == smp_processor_id()) { |
261 | if (np->drop) | 262 | if (np->drop) |
@@ -265,30 +266,37 @@ repeat: | |||
265 | return; | 266 | return; |
266 | } | 267 | } |
267 | 268 | ||
268 | spin_lock(&np->dev->xmit_lock); | 269 | do { |
269 | np->dev->xmit_lock_owner = smp_processor_id(); | 270 | npinfo->tries--; |
271 | spin_lock(&np->dev->xmit_lock); | ||
272 | np->dev->xmit_lock_owner = smp_processor_id(); | ||
270 | 273 | ||
271 | /* | 274 | /* |
272 | * network drivers do not expect to be called if the queue is | 275 | * network drivers do not expect to be called if the queue is |
273 | * stopped. | 276 | * stopped. |
274 | */ | 277 | */ |
275 | if (netif_queue_stopped(np->dev)) { | 278 | if (netif_queue_stopped(np->dev)) { |
279 | np->dev->xmit_lock_owner = -1; | ||
280 | spin_unlock(&np->dev->xmit_lock); | ||
281 | netpoll_poll(np); | ||
282 | udelay(50); | ||
283 | continue; | ||
284 | } | ||
285 | |||
286 | status = np->dev->hard_start_xmit(skb, np->dev); | ||
276 | np->dev->xmit_lock_owner = -1; | 287 | np->dev->xmit_lock_owner = -1; |
277 | spin_unlock(&np->dev->xmit_lock); | 288 | spin_unlock(&np->dev->xmit_lock); |
278 | 289 | ||
279 | netpoll_poll(np); | 290 | /* success */ |
280 | goto repeat; | 291 | if(!status) { |
281 | } | 292 | npinfo->tries = MAX_RETRIES; /* reset */ |
282 | 293 | return; | |
283 | status = np->dev->hard_start_xmit(skb, np->dev); | 294 | } |
284 | np->dev->xmit_lock_owner = -1; | ||
285 | spin_unlock(&np->dev->xmit_lock); | ||
286 | 295 | ||
287 | /* transmit busy */ | 296 | /* transmit busy */ |
288 | if(status) { | ||
289 | netpoll_poll(np); | 297 | netpoll_poll(np); |
290 | goto repeat; | 298 | udelay(50); |
291 | } | 299 | } while (npinfo->tries > 0); |
292 | } | 300 | } |
293 | 301 | ||
294 | void netpoll_send_udp(struct netpoll *np, const char *msg, int len) | 302 | void netpoll_send_udp(struct netpoll *np, const char *msg, int len) |
@@ -349,15 +357,11 @@ static void arp_reply(struct sk_buff *skb) | |||
349 | unsigned char *arp_ptr; | 357 | unsigned char *arp_ptr; |
350 | int size, type = ARPOP_REPLY, ptype = ETH_P_ARP; | 358 | int size, type = ARPOP_REPLY, ptype = ETH_P_ARP; |
351 | u32 sip, tip; | 359 | u32 sip, tip; |
352 | unsigned long flags; | ||
353 | struct sk_buff *send_skb; | 360 | struct sk_buff *send_skb; |
354 | struct netpoll *np = NULL; | 361 | struct netpoll *np = NULL; |
355 | 362 | ||
356 | spin_lock_irqsave(&npinfo->rx_lock, flags); | ||
357 | if (npinfo->rx_np && npinfo->rx_np->dev == skb->dev) | 363 | if (npinfo->rx_np && npinfo->rx_np->dev == skb->dev) |
358 | np = npinfo->rx_np; | 364 | np = npinfo->rx_np; |
359 | spin_unlock_irqrestore(&npinfo->rx_lock, flags); | ||
360 | |||
361 | if (!np) | 365 | if (!np) |
362 | return; | 366 | return; |
363 | 367 | ||
@@ -639,9 +643,11 @@ int netpoll_setup(struct netpoll *np) | |||
639 | if (!npinfo) | 643 | if (!npinfo) |
640 | goto release; | 644 | goto release; |
641 | 645 | ||
646 | npinfo->rx_flags = 0; | ||
642 | npinfo->rx_np = NULL; | 647 | npinfo->rx_np = NULL; |
643 | npinfo->poll_lock = SPIN_LOCK_UNLOCKED; | 648 | npinfo->poll_lock = SPIN_LOCK_UNLOCKED; |
644 | npinfo->poll_owner = -1; | 649 | npinfo->poll_owner = -1; |
650 | npinfo->tries = MAX_RETRIES; | ||
645 | npinfo->rx_lock = SPIN_LOCK_UNLOCKED; | 651 | npinfo->rx_lock = SPIN_LOCK_UNLOCKED; |
646 | } else | 652 | } else |
647 | npinfo = ndev->npinfo; | 653 | npinfo = ndev->npinfo; |
@@ -718,9 +724,16 @@ int netpoll_setup(struct netpoll *np) | |||
718 | npinfo->rx_np = np; | 724 | npinfo->rx_np = np; |
719 | spin_unlock_irqrestore(&npinfo->rx_lock, flags); | 725 | spin_unlock_irqrestore(&npinfo->rx_lock, flags); |
720 | } | 726 | } |
727 | |||
728 | /* fill up the skb queue */ | ||
729 | refill_skbs(); | ||
730 | |||
721 | /* last thing to do is link it to the net device structure */ | 731 | /* last thing to do is link it to the net device structure */ |
722 | ndev->npinfo = npinfo; | 732 | ndev->npinfo = npinfo; |
723 | 733 | ||
734 | /* avoid racing with NAPI reading npinfo */ | ||
735 | synchronize_rcu(); | ||
736 | |||
724 | return 0; | 737 | return 0; |
725 | 738 | ||
726 | release: | 739 | release: |
diff --git a/net/core/pktgen.c b/net/core/pktgen.c index 975d651312dc..8eb083b6041a 100644 --- a/net/core/pktgen.c +++ b/net/core/pktgen.c | |||
@@ -363,7 +363,7 @@ struct pktgen_thread { | |||
363 | * All Rights Reserved. | 363 | * All Rights Reserved. |
364 | * | 364 | * |
365 | */ | 365 | */ |
366 | inline static s64 divremdi3(s64 x, s64 y, int type) | 366 | static inline s64 divremdi3(s64 x, s64 y, int type) |
367 | { | 367 | { |
368 | u64 a = (x < 0) ? -x : x; | 368 | u64 a = (x < 0) ? -x : x; |
369 | u64 b = (y < 0) ? -y : y; | 369 | u64 b = (y < 0) ? -y : y; |
diff --git a/net/core/skbuff.c b/net/core/skbuff.c index d9f7b06fe886..7eab867ede59 100644 --- a/net/core/skbuff.c +++ b/net/core/skbuff.c | |||
@@ -377,8 +377,8 @@ struct sk_buff *skb_clone(struct sk_buff *skb, unsigned int __nocast gfp_mask) | |||
377 | C(tc_index); | 377 | C(tc_index); |
378 | #ifdef CONFIG_NET_CLS_ACT | 378 | #ifdef CONFIG_NET_CLS_ACT |
379 | n->tc_verd = SET_TC_VERD(skb->tc_verd,0); | 379 | n->tc_verd = SET_TC_VERD(skb->tc_verd,0); |
380 | n->tc_verd = CLR_TC_OK2MUNGE(skb->tc_verd); | 380 | n->tc_verd = CLR_TC_OK2MUNGE(n->tc_verd); |
381 | n->tc_verd = CLR_TC_MUNGED(skb->tc_verd); | 381 | n->tc_verd = CLR_TC_MUNGED(n->tc_verd); |
382 | C(input_dev); | 382 | C(input_dev); |
383 | C(tc_classid); | 383 | C(tc_classid); |
384 | #endif | 384 | #endif |
diff --git a/net/core/sock.c b/net/core/sock.c index 8b35ccdc2b3b..12f6d9a2a522 100644 --- a/net/core/sock.c +++ b/net/core/sock.c | |||
@@ -206,13 +206,14 @@ int sock_setsockopt(struct socket *sock, int level, int optname, | |||
206 | */ | 206 | */ |
207 | 207 | ||
208 | #ifdef SO_DONTLINGER /* Compatibility item... */ | 208 | #ifdef SO_DONTLINGER /* Compatibility item... */ |
209 | switch (optname) { | 209 | if (optname == SO_DONTLINGER) { |
210 | case SO_DONTLINGER: | 210 | lock_sock(sk); |
211 | sock_reset_flag(sk, SOCK_LINGER); | 211 | sock_reset_flag(sk, SOCK_LINGER); |
212 | return 0; | 212 | release_sock(sk); |
213 | return 0; | ||
213 | } | 214 | } |
214 | #endif | 215 | #endif |
215 | 216 | ||
216 | if(optlen<sizeof(int)) | 217 | if(optlen<sizeof(int)) |
217 | return(-EINVAL); | 218 | return(-EINVAL); |
218 | 219 | ||
diff --git a/net/core/utils.c b/net/core/utils.c index e11a8654f363..88eb8b68e26b 100644 --- a/net/core/utils.c +++ b/net/core/utils.c | |||
@@ -23,10 +23,10 @@ | |||
23 | #include <linux/percpu.h> | 23 | #include <linux/percpu.h> |
24 | #include <linux/init.h> | 24 | #include <linux/init.h> |
25 | 25 | ||
26 | #include <asm/byteorder.h> | ||
26 | #include <asm/system.h> | 27 | #include <asm/system.h> |
27 | #include <asm/uaccess.h> | 28 | #include <asm/uaccess.h> |
28 | 29 | ||
29 | |||
30 | /* | 30 | /* |
31 | This is a maximally equidistributed combined Tausworthe generator | 31 | This is a maximally equidistributed combined Tausworthe generator |
32 | based on code from GNU Scientific Library 1.5 (30 Jun 2004) | 32 | based on code from GNU Scientific Library 1.5 (30 Jun 2004) |
@@ -153,3 +153,38 @@ int net_ratelimit(void) | |||
153 | EXPORT_SYMBOL(net_random); | 153 | EXPORT_SYMBOL(net_random); |
154 | EXPORT_SYMBOL(net_ratelimit); | 154 | EXPORT_SYMBOL(net_ratelimit); |
155 | EXPORT_SYMBOL(net_srandom); | 155 | EXPORT_SYMBOL(net_srandom); |
156 | |||
157 | /* | ||
158 | * Convert an ASCII string to binary IP. | ||
159 | * This is outside of net/ipv4/ because various code that uses IP addresses | ||
160 | * is otherwise not dependent on the TCP/IP stack. | ||
161 | */ | ||
162 | |||
163 | __u32 in_aton(const char *str) | ||
164 | { | ||
165 | unsigned long l; | ||
166 | unsigned int val; | ||
167 | int i; | ||
168 | |||
169 | l = 0; | ||
170 | for (i = 0; i < 4; i++) | ||
171 | { | ||
172 | l <<= 8; | ||
173 | if (*str != '\0') | ||
174 | { | ||
175 | val = 0; | ||
176 | while (*str != '\0' && *str != '.') | ||
177 | { | ||
178 | val *= 10; | ||
179 | val += *str - '0'; | ||
180 | str++; | ||
181 | } | ||
182 | l |= val; | ||
183 | if (*str != '\0') | ||
184 | str++; | ||
185 | } | ||
186 | } | ||
187 | return(htonl(l)); | ||
188 | } | ||
189 | |||
190 | EXPORT_SYMBOL(in_aton); | ||
diff --git a/net/decnet/af_decnet.c b/net/decnet/af_decnet.c index 96a02800cd28..acdd18e6adb2 100644 --- a/net/decnet/af_decnet.c +++ b/net/decnet/af_decnet.c | |||
@@ -1876,15 +1876,6 @@ static inline unsigned int dn_current_mss(struct sock *sk, int flags) | |||
1876 | return mss_now; | 1876 | return mss_now; |
1877 | } | 1877 | } |
1878 | 1878 | ||
1879 | static int dn_error(struct sock *sk, int flags, int err) | ||
1880 | { | ||
1881 | if (err == -EPIPE) | ||
1882 | err = sock_error(sk) ? : -EPIPE; | ||
1883 | if (err == -EPIPE && !(flags & MSG_NOSIGNAL)) | ||
1884 | send_sig(SIGPIPE, current, 0); | ||
1885 | return err; | ||
1886 | } | ||
1887 | |||
1888 | static int dn_sendmsg(struct kiocb *iocb, struct socket *sock, | 1879 | static int dn_sendmsg(struct kiocb *iocb, struct socket *sock, |
1889 | struct msghdr *msg, size_t size) | 1880 | struct msghdr *msg, size_t size) |
1890 | { | 1881 | { |
@@ -2045,7 +2036,7 @@ out: | |||
2045 | return sent ? sent : err; | 2036 | return sent ? sent : err; |
2046 | 2037 | ||
2047 | out_err: | 2038 | out_err: |
2048 | err = dn_error(sk, flags, err); | 2039 | err = sk_stream_error(sk, flags, err); |
2049 | release_sock(sk); | 2040 | release_sock(sk); |
2050 | return err; | 2041 | return err; |
2051 | } | 2042 | } |
diff --git a/net/decnet/dn_neigh.c b/net/decnet/dn_neigh.c index f32dba9e26fe..8d0cc3cf3e49 100644 --- a/net/decnet/dn_neigh.c +++ b/net/decnet/dn_neigh.c | |||
@@ -148,12 +148,12 @@ static int dn_neigh_construct(struct neighbour *neigh) | |||
148 | 148 | ||
149 | __neigh_parms_put(neigh->parms); | 149 | __neigh_parms_put(neigh->parms); |
150 | neigh->parms = neigh_parms_clone(parms); | 150 | neigh->parms = neigh_parms_clone(parms); |
151 | rcu_read_unlock(); | ||
152 | 151 | ||
153 | if (dn_db->use_long) | 152 | if (dn_db->use_long) |
154 | neigh->ops = &dn_long_ops; | 153 | neigh->ops = &dn_long_ops; |
155 | else | 154 | else |
156 | neigh->ops = &dn_short_ops; | 155 | neigh->ops = &dn_short_ops; |
156 | rcu_read_unlock(); | ||
157 | 157 | ||
158 | if (dn->flags & DN_NDFLAG_P3) | 158 | if (dn->flags & DN_NDFLAG_P3) |
159 | neigh->ops = &dn_phase3_ops; | 159 | neigh->ops = &dn_phase3_ops; |
diff --git a/net/ipv4/Kconfig b/net/ipv4/Kconfig index df5386885a90..0b3d9f1d8069 100644 --- a/net/ipv4/Kconfig +++ b/net/ipv4/Kconfig | |||
@@ -54,9 +54,9 @@ config IP_ADVANCED_ROUTER | |||
54 | choice | 54 | choice |
55 | prompt "Choose IP: FIB lookup algorithm (choose FIB_HASH if unsure)" | 55 | prompt "Choose IP: FIB lookup algorithm (choose FIB_HASH if unsure)" |
56 | depends on IP_ADVANCED_ROUTER | 56 | depends on IP_ADVANCED_ROUTER |
57 | default IP_FIB_HASH | 57 | default ASK_IP_FIB_HASH |
58 | 58 | ||
59 | config IP_FIB_HASH | 59 | config ASK_IP_FIB_HASH |
60 | bool "FIB_HASH" | 60 | bool "FIB_HASH" |
61 | ---help--- | 61 | ---help--- |
62 | Current FIB is very proven and good enough for most users. | 62 | Current FIB is very proven and good enough for most users. |
@@ -82,12 +82,8 @@ config IP_FIB_TRIE | |||
82 | 82 | ||
83 | endchoice | 83 | endchoice |
84 | 84 | ||
85 | # If the user does not enable advanced routing, he gets the safe | ||
86 | # default of the fib-hash algorithm. | ||
87 | config IP_FIB_HASH | 85 | config IP_FIB_HASH |
88 | bool | 86 | def_bool ASK_IP_FIB_HASH || !IP_ADVANCED_ROUTER |
89 | depends on !IP_ADVANCED_ROUTER | ||
90 | default y | ||
91 | 87 | ||
92 | config IP_MULTIPLE_TABLES | 88 | config IP_MULTIPLE_TABLES |
93 | bool "IP: policy routing" | 89 | bool "IP: policy routing" |
@@ -128,7 +124,7 @@ config IP_ROUTE_MULTIPATH | |||
128 | 124 | ||
129 | config IP_ROUTE_MULTIPATH_CACHED | 125 | config IP_ROUTE_MULTIPATH_CACHED |
130 | bool "IP: equal cost multipath with caching support (EXPERIMENTAL)" | 126 | bool "IP: equal cost multipath with caching support (EXPERIMENTAL)" |
131 | depends on: IP_ROUTE_MULTIPATH | 127 | depends on IP_ROUTE_MULTIPATH |
132 | help | 128 | help |
133 | Normally, equal cost multipath routing is not supported by the | 129 | Normally, equal cost multipath routing is not supported by the |
134 | routing cache. If you say Y here, alternative routes are cached | 130 | routing cache. If you say Y here, alternative routes are cached |
@@ -239,7 +235,6 @@ config IP_PNP_RARP | |||
239 | # bool ' IP: ARP support' CONFIG_IP_PNP_ARP | 235 | # bool ' IP: ARP support' CONFIG_IP_PNP_ARP |
240 | config NET_IPIP | 236 | config NET_IPIP |
241 | tristate "IP: tunneling" | 237 | tristate "IP: tunneling" |
242 | select INET_TUNNEL | ||
243 | ---help--- | 238 | ---help--- |
244 | Tunneling means encapsulating data of one protocol type within | 239 | Tunneling means encapsulating data of one protocol type within |
245 | another protocol and sending it over a channel that understands the | 240 | another protocol and sending it over a channel that understands the |
@@ -256,7 +251,6 @@ config NET_IPIP | |||
256 | 251 | ||
257 | config NET_IPGRE | 252 | config NET_IPGRE |
258 | tristate "IP: GRE tunnels over IP" | 253 | tristate "IP: GRE tunnels over IP" |
259 | select XFRM | ||
260 | help | 254 | help |
261 | Tunneling means encapsulating data of one protocol type within | 255 | Tunneling means encapsulating data of one protocol type within |
262 | another protocol and sending it over a channel that understands the | 256 | another protocol and sending it over a channel that understands the |
diff --git a/net/ipv4/Makefile b/net/ipv4/Makefile index 5718cdb3a61e..55dc6cca1e7b 100644 --- a/net/ipv4/Makefile +++ b/net/ipv4/Makefile | |||
@@ -2,7 +2,7 @@ | |||
2 | # Makefile for the Linux TCP/IP (INET) layer. | 2 | # Makefile for the Linux TCP/IP (INET) layer. |
3 | # | 3 | # |
4 | 4 | ||
5 | obj-y := utils.o route.o inetpeer.o protocol.o \ | 5 | obj-y := route.o inetpeer.o protocol.o \ |
6 | ip_input.o ip_fragment.o ip_forward.o ip_options.o \ | 6 | ip_input.o ip_fragment.o ip_forward.o ip_options.o \ |
7 | ip_output.o ip_sockglue.o \ | 7 | ip_output.o ip_sockglue.o \ |
8 | tcp.o tcp_input.o tcp_output.o tcp_timer.o tcp_ipv4.o \ | 8 | tcp.o tcp_input.o tcp_output.o tcp_timer.o tcp_ipv4.o \ |
diff --git a/net/ipv4/af_inet.c b/net/ipv4/af_inet.c index ef7468376ae6..163ae4068b5f 100644 --- a/net/ipv4/af_inet.c +++ b/net/ipv4/af_inet.c | |||
@@ -1157,7 +1157,7 @@ static int __init ipv4_proc_init(void) | |||
1157 | #ifdef CONFIG_IP_FIB_TRIE | 1157 | #ifdef CONFIG_IP_FIB_TRIE |
1158 | if (fib_stat_proc_init()) | 1158 | if (fib_stat_proc_init()) |
1159 | goto out_fib_stat; | 1159 | goto out_fib_stat; |
1160 | #endif | 1160 | #endif |
1161 | if (ip_misc_proc_init()) | 1161 | if (ip_misc_proc_init()) |
1162 | goto out_misc; | 1162 | goto out_misc; |
1163 | out: | 1163 | out: |
diff --git a/net/ipv4/fib_semantics.c b/net/ipv4/fib_semantics.c index c886b28ba9f5..e278cb9d0075 100644 --- a/net/ipv4/fib_semantics.c +++ b/net/ipv4/fib_semantics.c | |||
@@ -593,10 +593,13 @@ static void fib_hash_move(struct hlist_head *new_info_hash, | |||
593 | struct hlist_head *new_laddrhash, | 593 | struct hlist_head *new_laddrhash, |
594 | unsigned int new_size) | 594 | unsigned int new_size) |
595 | { | 595 | { |
596 | struct hlist_head *old_info_hash, *old_laddrhash; | ||
596 | unsigned int old_size = fib_hash_size; | 597 | unsigned int old_size = fib_hash_size; |
597 | unsigned int i; | 598 | unsigned int i, bytes; |
598 | 599 | ||
599 | write_lock(&fib_info_lock); | 600 | write_lock(&fib_info_lock); |
601 | old_info_hash = fib_info_hash; | ||
602 | old_laddrhash = fib_info_laddrhash; | ||
600 | fib_hash_size = new_size; | 603 | fib_hash_size = new_size; |
601 | 604 | ||
602 | for (i = 0; i < old_size; i++) { | 605 | for (i = 0; i < old_size; i++) { |
@@ -636,6 +639,10 @@ static void fib_hash_move(struct hlist_head *new_info_hash, | |||
636 | fib_info_laddrhash = new_laddrhash; | 639 | fib_info_laddrhash = new_laddrhash; |
637 | 640 | ||
638 | write_unlock(&fib_info_lock); | 641 | write_unlock(&fib_info_lock); |
642 | |||
643 | bytes = old_size * sizeof(struct hlist_head *); | ||
644 | fib_hash_free(old_info_hash, bytes); | ||
645 | fib_hash_free(old_laddrhash, bytes); | ||
639 | } | 646 | } |
640 | 647 | ||
641 | struct fib_info * | 648 | struct fib_info * |
diff --git a/net/ipv4/fib_trie.c b/net/ipv4/fib_trie.c index 4be234c7d8c3..45efd5f4741b 100644 --- a/net/ipv4/fib_trie.c +++ b/net/ipv4/fib_trie.c | |||
@@ -90,14 +90,14 @@ typedef unsigned int t_key; | |||
90 | #define T_LEAF 1 | 90 | #define T_LEAF 1 |
91 | #define NODE_TYPE_MASK 0x1UL | 91 | #define NODE_TYPE_MASK 0x1UL |
92 | #define NODE_PARENT(_node) \ | 92 | #define NODE_PARENT(_node) \ |
93 | ((struct tnode *)((_node)->_parent & ~NODE_TYPE_MASK)) | 93 | ((struct tnode *)((_node)->_parent & ~NODE_TYPE_MASK)) |
94 | #define NODE_SET_PARENT(_node, _ptr) \ | 94 | #define NODE_SET_PARENT(_node, _ptr) \ |
95 | ((_node)->_parent = (((unsigned long)(_ptr)) | \ | 95 | ((_node)->_parent = (((unsigned long)(_ptr)) | \ |
96 | ((_node)->_parent & NODE_TYPE_MASK))) | 96 | ((_node)->_parent & NODE_TYPE_MASK))) |
97 | #define NODE_INIT_PARENT(_node, _type) \ | 97 | #define NODE_INIT_PARENT(_node, _type) \ |
98 | ((_node)->_parent = (_type)) | 98 | ((_node)->_parent = (_type)) |
99 | #define NODE_TYPE(_node) \ | 99 | #define NODE_TYPE(_node) \ |
100 | ((_node)->_parent & NODE_TYPE_MASK) | 100 | ((_node)->_parent & NODE_TYPE_MASK) |
101 | 101 | ||
102 | #define IS_TNODE(n) (!(n->_parent & T_LEAF)) | 102 | #define IS_TNODE(n) (!(n->_parent & T_LEAF)) |
103 | #define IS_LEAF(n) (n->_parent & T_LEAF) | 103 | #define IS_LEAF(n) (n->_parent & T_LEAF) |
@@ -147,7 +147,7 @@ struct trie_stat { | |||
147 | unsigned int leaves; | 147 | unsigned int leaves; |
148 | unsigned int nullpointers; | 148 | unsigned int nullpointers; |
149 | unsigned int nodesizes[MAX_CHILDS]; | 149 | unsigned int nodesizes[MAX_CHILDS]; |
150 | }; | 150 | }; |
151 | 151 | ||
152 | struct trie { | 152 | struct trie { |
153 | struct node *trie; | 153 | struct node *trie; |
@@ -185,9 +185,9 @@ static void trie_bug(char *err) | |||
185 | BUG(); | 185 | BUG(); |
186 | } | 186 | } |
187 | 187 | ||
188 | static inline struct node *tnode_get_child(struct tnode *tn, int i) | 188 | static inline struct node *tnode_get_child(struct tnode *tn, int i) |
189 | { | 189 | { |
190 | if (i >= 1<<tn->bits) | 190 | if (i >= 1<<tn->bits) |
191 | trie_bug("tnode_get_child"); | 191 | trie_bug("tnode_get_child"); |
192 | 192 | ||
193 | return tn->child[i]; | 193 | return tn->child[i]; |
@@ -202,7 +202,7 @@ static inline int tnode_child_length(struct tnode *tn) | |||
202 | _________________________________________________________________ | 202 | _________________________________________________________________ |
203 | | i | i | i | i | i | i | i | N | N | N | S | S | S | S | S | C | | 203 | | i | i | i | i | i | i | i | N | N | N | S | S | S | S | S | C | |
204 | ---------------------------------------------------------------- | 204 | ---------------------------------------------------------------- |
205 | 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 | 205 | 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 |
206 | 206 | ||
207 | _________________________________________________________________ | 207 | _________________________________________________________________ |
208 | | C | C | C | u | u | u | u | u | u | u | u | u | u | u | u | u | | 208 | | C | C | C | u | u | u | u | u | u | u | u | u | u | u | u | u | |
@@ -226,25 +226,25 @@ static inline t_key tkey_extract_bits(t_key a, int offset, int bits) | |||
226 | 226 | ||
227 | static inline int tkey_equals(t_key a, t_key b) | 227 | static inline int tkey_equals(t_key a, t_key b) |
228 | { | 228 | { |
229 | return a == b; | 229 | return a == b; |
230 | } | 230 | } |
231 | 231 | ||
232 | static inline int tkey_sub_equals(t_key a, int offset, int bits, t_key b) | 232 | static inline int tkey_sub_equals(t_key a, int offset, int bits, t_key b) |
233 | { | 233 | { |
234 | if (bits == 0 || offset >= KEYLENGTH) | 234 | if (bits == 0 || offset >= KEYLENGTH) |
235 | return 1; | 235 | return 1; |
236 | bits = bits > KEYLENGTH ? KEYLENGTH : bits; | 236 | bits = bits > KEYLENGTH ? KEYLENGTH : bits; |
237 | return ((a ^ b) << offset) >> (KEYLENGTH - bits) == 0; | 237 | return ((a ^ b) << offset) >> (KEYLENGTH - bits) == 0; |
238 | } | 238 | } |
239 | 239 | ||
240 | static inline int tkey_mismatch(t_key a, int offset, t_key b) | 240 | static inline int tkey_mismatch(t_key a, int offset, t_key b) |
241 | { | 241 | { |
242 | t_key diff = a ^ b; | 242 | t_key diff = a ^ b; |
243 | int i = offset; | 243 | int i = offset; |
244 | 244 | ||
245 | if(!diff) | 245 | if (!diff) |
246 | return 0; | 246 | return 0; |
247 | while((diff << i) >> (KEYLENGTH-1) == 0) | 247 | while ((diff << i) >> (KEYLENGTH-1) == 0) |
248 | i++; | 248 | i++; |
249 | return i; | 249 | return i; |
250 | } | 250 | } |
@@ -314,6 +314,7 @@ static void fn_free_alias(struct fib_alias *fa) | |||
314 | The bits from (n->pos) to (n->pos + n->bits - 1) - "C" - are the index into | 314 | The bits from (n->pos) to (n->pos + n->bits - 1) - "C" - are the index into |
315 | n's child array, and will of course be different for each child. | 315 | n's child array, and will of course be different for each child. |
316 | 316 | ||
317 | |||
317 | The rest of the bits, from (n->pos + n->bits) onward, are completely unknown | 318 | The rest of the bits, from (n->pos + n->bits) onward, are completely unknown |
318 | at this point. | 319 | at this point. |
319 | 320 | ||
@@ -321,7 +322,7 @@ static void fn_free_alias(struct fib_alias *fa) | |||
321 | 322 | ||
322 | static void check_tnode(struct tnode *tn) | 323 | static void check_tnode(struct tnode *tn) |
323 | { | 324 | { |
324 | if(tn && tn->pos+tn->bits > 32) { | 325 | if (tn && tn->pos+tn->bits > 32) { |
325 | printk("TNODE ERROR tn=%p, pos=%d, bits=%d\n", tn, tn->pos, tn->bits); | 326 | printk("TNODE ERROR tn=%p, pos=%d, bits=%d\n", tn, tn->pos, tn->bits); |
326 | } | 327 | } |
327 | } | 328 | } |
@@ -332,7 +333,7 @@ static int inflate_threshold = 50; | |||
332 | static struct leaf *leaf_new(void) | 333 | static struct leaf *leaf_new(void) |
333 | { | 334 | { |
334 | struct leaf *l = kmalloc(sizeof(struct leaf), GFP_KERNEL); | 335 | struct leaf *l = kmalloc(sizeof(struct leaf), GFP_KERNEL); |
335 | if(l) { | 336 | if (l) { |
336 | NODE_INIT_PARENT(l, T_LEAF); | 337 | NODE_INIT_PARENT(l, T_LEAF); |
337 | INIT_HLIST_HEAD(&l->list); | 338 | INIT_HLIST_HEAD(&l->list); |
338 | } | 339 | } |
@@ -342,7 +343,7 @@ static struct leaf *leaf_new(void) | |||
342 | static struct leaf_info *leaf_info_new(int plen) | 343 | static struct leaf_info *leaf_info_new(int plen) |
343 | { | 344 | { |
344 | struct leaf_info *li = kmalloc(sizeof(struct leaf_info), GFP_KERNEL); | 345 | struct leaf_info *li = kmalloc(sizeof(struct leaf_info), GFP_KERNEL); |
345 | if(li) { | 346 | if (li) { |
346 | li->plen = plen; | 347 | li->plen = plen; |
347 | INIT_LIST_HEAD(&li->falh); | 348 | INIT_LIST_HEAD(&li->falh); |
348 | } | 349 | } |
@@ -365,7 +366,7 @@ static struct tnode *tnode_alloc(unsigned int size) | |||
365 | return kmalloc(size, GFP_KERNEL); | 366 | return kmalloc(size, GFP_KERNEL); |
366 | } else { | 367 | } else { |
367 | return (struct tnode *) | 368 | return (struct tnode *) |
368 | __get_free_pages(GFP_KERNEL, get_order(size)); | 369 | __get_free_pages(GFP_KERNEL, get_order(size)); |
369 | } | 370 | } |
370 | } | 371 | } |
371 | 372 | ||
@@ -386,7 +387,7 @@ static struct tnode* tnode_new(t_key key, int pos, int bits) | |||
386 | int sz = sizeof(struct tnode) + nchildren * sizeof(struct node *); | 387 | int sz = sizeof(struct tnode) + nchildren * sizeof(struct node *); |
387 | struct tnode *tn = tnode_alloc(sz); | 388 | struct tnode *tn = tnode_alloc(sz); |
388 | 389 | ||
389 | if(tn) { | 390 | if (tn) { |
390 | memset(tn, 0, sz); | 391 | memset(tn, 0, sz); |
391 | NODE_INIT_PARENT(tn, T_TNODE); | 392 | NODE_INIT_PARENT(tn, T_TNODE); |
392 | tn->pos = pos; | 393 | tn->pos = pos; |
@@ -395,7 +396,8 @@ static struct tnode* tnode_new(t_key key, int pos, int bits) | |||
395 | tn->full_children = 0; | 396 | tn->full_children = 0; |
396 | tn->empty_children = 1<<bits; | 397 | tn->empty_children = 1<<bits; |
397 | } | 398 | } |
398 | if(trie_debug > 0) | 399 | |
400 | if (trie_debug > 0) | ||
399 | printk("AT %p s=%u %u\n", tn, (unsigned int) sizeof(struct tnode), | 401 | printk("AT %p s=%u %u\n", tn, (unsigned int) sizeof(struct tnode), |
400 | (unsigned int) (sizeof(struct node) * 1<<bits)); | 402 | (unsigned int) (sizeof(struct node) * 1<<bits)); |
401 | return tn; | 403 | return tn; |
@@ -403,17 +405,17 @@ static struct tnode* tnode_new(t_key key, int pos, int bits) | |||
403 | 405 | ||
404 | static void tnode_free(struct tnode *tn) | 406 | static void tnode_free(struct tnode *tn) |
405 | { | 407 | { |
406 | if(!tn) { | 408 | if (!tn) { |
407 | trie_bug("tnode_free\n"); | 409 | trie_bug("tnode_free\n"); |
408 | } | 410 | } |
409 | if(IS_LEAF(tn)) { | 411 | if (IS_LEAF(tn)) { |
410 | free_leaf((struct leaf *)tn); | 412 | free_leaf((struct leaf *)tn); |
411 | if(trie_debug > 0 ) | 413 | if (trie_debug > 0 ) |
412 | printk("FL %p \n", tn); | 414 | printk("FL %p \n", tn); |
413 | } | 415 | } |
414 | else if(IS_TNODE(tn)) { | 416 | else if (IS_TNODE(tn)) { |
415 | __tnode_free(tn); | 417 | __tnode_free(tn); |
416 | if(trie_debug > 0 ) | 418 | if (trie_debug > 0 ) |
417 | printk("FT %p \n", tn); | 419 | printk("FT %p \n", tn); |
418 | } | 420 | } |
419 | else { | 421 | else { |
@@ -428,58 +430,58 @@ static void tnode_free(struct tnode *tn) | |||
428 | 430 | ||
429 | static inline int tnode_full(struct tnode *tn, struct node *n) | 431 | static inline int tnode_full(struct tnode *tn, struct node *n) |
430 | { | 432 | { |
431 | if(n == NULL || IS_LEAF(n)) | 433 | if (n == NULL || IS_LEAF(n)) |
432 | return 0; | 434 | return 0; |
433 | 435 | ||
434 | return ((struct tnode *) n)->pos == tn->pos + tn->bits; | 436 | return ((struct tnode *) n)->pos == tn->pos + tn->bits; |
435 | } | 437 | } |
436 | 438 | ||
437 | static inline void put_child(struct trie *t, struct tnode *tn, int i, struct node *n) | 439 | static inline void put_child(struct trie *t, struct tnode *tn, int i, struct node *n) |
438 | { | 440 | { |
439 | tnode_put_child_reorg(tn, i, n, -1); | 441 | tnode_put_child_reorg(tn, i, n, -1); |
440 | } | 442 | } |
441 | 443 | ||
442 | /* | 444 | /* |
443 | * Add a child at position i overwriting the old value. | 445 | * Add a child at position i overwriting the old value. |
444 | * Update the value of full_children and empty_children. | 446 | * Update the value of full_children and empty_children. |
445 | */ | 447 | */ |
446 | 448 | ||
447 | static void tnode_put_child_reorg(struct tnode *tn, int i, struct node *n, int wasfull) | 449 | static void tnode_put_child_reorg(struct tnode *tn, int i, struct node *n, int wasfull) |
448 | { | 450 | { |
449 | struct node *chi; | 451 | struct node *chi; |
450 | int isfull; | 452 | int isfull; |
451 | 453 | ||
452 | if(i >= 1<<tn->bits) { | 454 | if (i >= 1<<tn->bits) { |
453 | printk("bits=%d, i=%d\n", tn->bits, i); | 455 | printk("bits=%d, i=%d\n", tn->bits, i); |
454 | trie_bug("tnode_put_child_reorg bits"); | 456 | trie_bug("tnode_put_child_reorg bits"); |
455 | } | 457 | } |
456 | write_lock_bh(&fib_lock); | 458 | write_lock_bh(&fib_lock); |
457 | chi = tn->child[i]; | 459 | chi = tn->child[i]; |
458 | 460 | ||
459 | /* update emptyChildren */ | 461 | /* update emptyChildren */ |
460 | if (n == NULL && chi != NULL) | 462 | if (n == NULL && chi != NULL) |
461 | tn->empty_children++; | 463 | tn->empty_children++; |
462 | else if (n != NULL && chi == NULL) | 464 | else if (n != NULL && chi == NULL) |
463 | tn->empty_children--; | 465 | tn->empty_children--; |
464 | 466 | ||
465 | /* update fullChildren */ | 467 | /* update fullChildren */ |
466 | if (wasfull == -1) | 468 | if (wasfull == -1) |
467 | wasfull = tnode_full(tn, chi); | 469 | wasfull = tnode_full(tn, chi); |
468 | 470 | ||
469 | isfull = tnode_full(tn, n); | 471 | isfull = tnode_full(tn, n); |
470 | if (wasfull && !isfull) | 472 | if (wasfull && !isfull) |
471 | tn->full_children--; | 473 | tn->full_children--; |
472 | 474 | ||
473 | else if (!wasfull && isfull) | 475 | else if (!wasfull && isfull) |
474 | tn->full_children++; | 476 | tn->full_children++; |
475 | if(n) | 477 | if (n) |
476 | NODE_SET_PARENT(n, tn); | 478 | NODE_SET_PARENT(n, tn); |
477 | 479 | ||
478 | tn->child[i] = n; | 480 | tn->child[i] = n; |
479 | write_unlock_bh(&fib_lock); | 481 | write_unlock_bh(&fib_lock); |
480 | } | 482 | } |
481 | 483 | ||
482 | static struct node *resize(struct trie *t, struct tnode *tn) | 484 | static struct node *resize(struct trie *t, struct tnode *tn) |
483 | { | 485 | { |
484 | int i; | 486 | int i; |
485 | int err = 0; | 487 | int err = 0; |
@@ -487,8 +489,8 @@ static struct node *resize(struct trie *t, struct tnode *tn) | |||
487 | if (!tn) | 489 | if (!tn) |
488 | return NULL; | 490 | return NULL; |
489 | 491 | ||
490 | if(trie_debug) | 492 | if (trie_debug) |
491 | printk("In tnode_resize %p inflate_threshold=%d threshold=%d\n", | 493 | printk("In tnode_resize %p inflate_threshold=%d threshold=%d\n", |
492 | tn, inflate_threshold, halve_threshold); | 494 | tn, inflate_threshold, halve_threshold); |
493 | 495 | ||
494 | /* No children */ | 496 | /* No children */ |
@@ -505,7 +507,7 @@ static struct node *resize(struct trie *t, struct tnode *tn) | |||
505 | 507 | ||
506 | /* compress one level */ | 508 | /* compress one level */ |
507 | struct node *n = tn->child[i]; | 509 | struct node *n = tn->child[i]; |
508 | if(n) | 510 | if (n) |
509 | NODE_INIT_PARENT(n, NODE_TYPE(n)); | 511 | NODE_INIT_PARENT(n, NODE_TYPE(n)); |
510 | 512 | ||
511 | write_unlock_bh(&fib_lock); | 513 | write_unlock_bh(&fib_lock); |
@@ -514,72 +516,72 @@ static struct node *resize(struct trie *t, struct tnode *tn) | |||
514 | } | 516 | } |
515 | write_unlock_bh(&fib_lock); | 517 | write_unlock_bh(&fib_lock); |
516 | } | 518 | } |
517 | /* | 519 | /* |
518 | * Double as long as the resulting node has a number of | 520 | * Double as long as the resulting node has a number of |
519 | * nonempty nodes that are above the threshold. | 521 | * nonempty nodes that are above the threshold. |
520 | */ | 522 | */ |
521 | 523 | ||
522 | /* | 524 | /* |
523 | * From "Implementing a dynamic compressed trie" by Stefan Nilsson of | 525 | * From "Implementing a dynamic compressed trie" by Stefan Nilsson of |
524 | * the Helsinki University of Technology and Matti Tikkanen of Nokia | 526 | * the Helsinki University of Technology and Matti Tikkanen of Nokia |
525 | * Telecommunications, page 6: | 527 | * Telecommunications, page 6: |
526 | * "A node is doubled if the ratio of non-empty children to all | 528 | * "A node is doubled if the ratio of non-empty children to all |
527 | * children in the *doubled* node is at least 'high'." | 529 | * children in the *doubled* node is at least 'high'." |
528 | * | 530 | * |
529 | * 'high' in this instance is the variable 'inflate_threshold'. It | 531 | * 'high' in this instance is the variable 'inflate_threshold'. It |
530 | * is expressed as a percentage, so we multiply it with | 532 | * is expressed as a percentage, so we multiply it with |
531 | * tnode_child_length() and instead of multiplying by 2 (since the | 533 | * tnode_child_length() and instead of multiplying by 2 (since the |
532 | * child array will be doubled by inflate()) and multiplying | 534 | * child array will be doubled by inflate()) and multiplying |
533 | * the left-hand side by 100 (to handle the percentage thing) we | 535 | * the left-hand side by 100 (to handle the percentage thing) we |
534 | * multiply the left-hand side by 50. | 536 | * multiply the left-hand side by 50. |
535 | * | 537 | * |
536 | * The left-hand side may look a bit weird: tnode_child_length(tn) | 538 | * The left-hand side may look a bit weird: tnode_child_length(tn) |
537 | * - tn->empty_children is of course the number of non-null children | 539 | * - tn->empty_children is of course the number of non-null children |
538 | * in the current node. tn->full_children is the number of "full" | 540 | * in the current node. tn->full_children is the number of "full" |
539 | * children, that is non-null tnodes with a skip value of 0. | 541 | * children, that is non-null tnodes with a skip value of 0. |
540 | * All of those will be doubled in the resulting inflated tnode, so | 542 | * All of those will be doubled in the resulting inflated tnode, so |
541 | * we just count them one extra time here. | 543 | * we just count them one extra time here. |
542 | * | 544 | * |
543 | * A clearer way to write this would be: | 545 | * A clearer way to write this would be: |
544 | * | 546 | * |
545 | * to_be_doubled = tn->full_children; | 547 | * to_be_doubled = tn->full_children; |
546 | * not_to_be_doubled = tnode_child_length(tn) - tn->empty_children - | 548 | * not_to_be_doubled = tnode_child_length(tn) - tn->empty_children - |
547 | * tn->full_children; | 549 | * tn->full_children; |
548 | * | 550 | * |
549 | * new_child_length = tnode_child_length(tn) * 2; | 551 | * new_child_length = tnode_child_length(tn) * 2; |
550 | * | 552 | * |
551 | * new_fill_factor = 100 * (not_to_be_doubled + 2*to_be_doubled) / | 553 | * new_fill_factor = 100 * (not_to_be_doubled + 2*to_be_doubled) / |
552 | * new_child_length; | 554 | * new_child_length; |
553 | * if (new_fill_factor >= inflate_threshold) | 555 | * if (new_fill_factor >= inflate_threshold) |
554 | * | 556 | * |
555 | * ...and so on, tho it would mess up the while() loop. | 557 | * ...and so on, tho it would mess up the while () loop. |
556 | * | 558 | * |
557 | * anyway, | 559 | * anyway, |
558 | * 100 * (not_to_be_doubled + 2*to_be_doubled) / new_child_length >= | 560 | * 100 * (not_to_be_doubled + 2*to_be_doubled) / new_child_length >= |
559 | * inflate_threshold | 561 | * inflate_threshold |
560 | * | 562 | * |
561 | * avoid a division: | 563 | * avoid a division: |
562 | * 100 * (not_to_be_doubled + 2*to_be_doubled) >= | 564 | * 100 * (not_to_be_doubled + 2*to_be_doubled) >= |
563 | * inflate_threshold * new_child_length | 565 | * inflate_threshold * new_child_length |
564 | * | 566 | * |
565 | * expand not_to_be_doubled and to_be_doubled, and shorten: | 567 | * expand not_to_be_doubled and to_be_doubled, and shorten: |
566 | * 100 * (tnode_child_length(tn) - tn->empty_children + | 568 | * 100 * (tnode_child_length(tn) - tn->empty_children + |
567 | * tn->full_children ) >= inflate_threshold * new_child_length | 569 | * tn->full_children ) >= inflate_threshold * new_child_length |
568 | * | 570 | * |
569 | * expand new_child_length: | 571 | * expand new_child_length: |
570 | * 100 * (tnode_child_length(tn) - tn->empty_children + | 572 | * 100 * (tnode_child_length(tn) - tn->empty_children + |
571 | * tn->full_children ) >= | 573 | * tn->full_children ) >= |
572 | * inflate_threshold * tnode_child_length(tn) * 2 | 574 | * inflate_threshold * tnode_child_length(tn) * 2 |
573 | * | 575 | * |
574 | * shorten again: | 576 | * shorten again: |
575 | * 50 * (tn->full_children + tnode_child_length(tn) - | 577 | * 50 * (tn->full_children + tnode_child_length(tn) - |
576 | * tn->empty_children ) >= inflate_threshold * | 578 | * tn->empty_children ) >= inflate_threshold * |
577 | * tnode_child_length(tn) | 579 | * tnode_child_length(tn) |
578 | * | 580 | * |
579 | */ | 581 | */ |
580 | 582 | ||
581 | check_tnode(tn); | 583 | check_tnode(tn); |
582 | 584 | ||
583 | err = 0; | 585 | err = 0; |
584 | while ((tn->full_children > 0 && | 586 | while ((tn->full_children > 0 && |
585 | 50 * (tn->full_children + tnode_child_length(tn) - tn->empty_children) >= | 587 | 50 * (tn->full_children + tnode_child_length(tn) - tn->empty_children) >= |
@@ -587,7 +589,7 @@ static struct node *resize(struct trie *t, struct tnode *tn) | |||
587 | 589 | ||
588 | tn = inflate(t, tn, &err); | 590 | tn = inflate(t, tn, &err); |
589 | 591 | ||
590 | if(err) { | 592 | if (err) { |
591 | #ifdef CONFIG_IP_FIB_TRIE_STATS | 593 | #ifdef CONFIG_IP_FIB_TRIE_STATS |
592 | t->stats.resize_node_skipped++; | 594 | t->stats.resize_node_skipped++; |
593 | #endif | 595 | #endif |
@@ -609,7 +611,7 @@ static struct node *resize(struct trie *t, struct tnode *tn) | |||
609 | 611 | ||
610 | tn = halve(t, tn, &err); | 612 | tn = halve(t, tn, &err); |
611 | 613 | ||
612 | if(err) { | 614 | if (err) { |
613 | #ifdef CONFIG_IP_FIB_TRIE_STATS | 615 | #ifdef CONFIG_IP_FIB_TRIE_STATS |
614 | t->stats.resize_node_skipped++; | 616 | t->stats.resize_node_skipped++; |
615 | #endif | 617 | #endif |
@@ -617,18 +619,18 @@ static struct node *resize(struct trie *t, struct tnode *tn) | |||
617 | } | 619 | } |
618 | } | 620 | } |
619 | 621 | ||
620 | 622 | ||
621 | /* Only one child remains */ | 623 | /* Only one child remains */ |
622 | 624 | ||
623 | if (tn->empty_children == tnode_child_length(tn) - 1) | 625 | if (tn->empty_children == tnode_child_length(tn) - 1) |
624 | for (i = 0; i < tnode_child_length(tn); i++) { | 626 | for (i = 0; i < tnode_child_length(tn); i++) { |
625 | 627 | ||
626 | write_lock_bh(&fib_lock); | 628 | write_lock_bh(&fib_lock); |
627 | if (tn->child[i] != NULL) { | 629 | if (tn->child[i] != NULL) { |
628 | /* compress one level */ | 630 | /* compress one level */ |
629 | struct node *n = tn->child[i]; | 631 | struct node *n = tn->child[i]; |
630 | 632 | ||
631 | if(n) | 633 | if (n) |
632 | NODE_INIT_PARENT(n, NODE_TYPE(n)); | 634 | NODE_INIT_PARENT(n, NODE_TYPE(n)); |
633 | 635 | ||
634 | write_unlock_bh(&fib_lock); | 636 | write_unlock_bh(&fib_lock); |
@@ -648,7 +650,7 @@ static struct tnode *inflate(struct trie *t, struct tnode *tn, int *err) | |||
648 | int olen = tnode_child_length(tn); | 650 | int olen = tnode_child_length(tn); |
649 | int i; | 651 | int i; |
650 | 652 | ||
651 | if(trie_debug) | 653 | if (trie_debug) |
652 | printk("In inflate\n"); | 654 | printk("In inflate\n"); |
653 | 655 | ||
654 | tn = tnode_new(oldtnode->key, oldtnode->pos, oldtnode->bits + 1); | 656 | tn = tnode_new(oldtnode->key, oldtnode->pos, oldtnode->bits + 1); |
@@ -659,12 +661,12 @@ static struct tnode *inflate(struct trie *t, struct tnode *tn, int *err) | |||
659 | } | 661 | } |
660 | 662 | ||
661 | /* | 663 | /* |
662 | * Preallocate and store tnodes before the actual work so we | 664 | * Preallocate and store tnodes before the actual work so we |
663 | * don't get into an inconsistent state if memory allocation | 665 | * don't get into an inconsistent state if memory allocation |
664 | * fails. In case of failure we return the oldnode and inflate | 666 | * fails. In case of failure we return the oldnode and inflate |
665 | * of tnode is ignored. | 667 | * of tnode is ignored. |
666 | */ | 668 | */ |
667 | 669 | ||
668 | for(i = 0; i < olen; i++) { | 670 | for(i = 0; i < olen; i++) { |
669 | struct tnode *inode = (struct tnode *) tnode_get_child(oldtnode, i); | 671 | struct tnode *inode = (struct tnode *) tnode_get_child(oldtnode, i); |
670 | 672 | ||
@@ -675,20 +677,20 @@ static struct tnode *inflate(struct trie *t, struct tnode *tn, int *err) | |||
675 | struct tnode *left, *right; | 677 | struct tnode *left, *right; |
676 | 678 | ||
677 | t_key m = TKEY_GET_MASK(inode->pos, 1); | 679 | t_key m = TKEY_GET_MASK(inode->pos, 1); |
678 | 680 | ||
679 | left = tnode_new(inode->key&(~m), inode->pos + 1, | 681 | left = tnode_new(inode->key&(~m), inode->pos + 1, |
680 | inode->bits - 1); | 682 | inode->bits - 1); |
681 | 683 | ||
682 | if(!left) { | 684 | if (!left) { |
683 | *err = -ENOMEM; | 685 | *err = -ENOMEM; |
684 | break; | 686 | break; |
685 | } | 687 | } |
686 | 688 | ||
687 | right = tnode_new(inode->key|m, inode->pos + 1, | 689 | right = tnode_new(inode->key|m, inode->pos + 1, |
688 | inode->bits - 1); | 690 | inode->bits - 1); |
689 | 691 | ||
690 | if(!right) { | 692 | if (!right) { |
691 | *err = -ENOMEM; | 693 | *err = -ENOMEM; |
692 | break; | 694 | break; |
693 | } | 695 | } |
694 | 696 | ||
@@ -697,32 +699,32 @@ static struct tnode *inflate(struct trie *t, struct tnode *tn, int *err) | |||
697 | } | 699 | } |
698 | } | 700 | } |
699 | 701 | ||
700 | if(*err) { | 702 | if (*err) { |
701 | int size = tnode_child_length(tn); | 703 | int size = tnode_child_length(tn); |
702 | int j; | 704 | int j; |
703 | 705 | ||
704 | for(j = 0; j < size; j++) | 706 | for(j = 0; j < size; j++) |
705 | if( tn->child[j]) | 707 | if (tn->child[j]) |
706 | tnode_free((struct tnode *)tn->child[j]); | 708 | tnode_free((struct tnode *)tn->child[j]); |
707 | 709 | ||
708 | tnode_free(tn); | 710 | tnode_free(tn); |
709 | 711 | ||
710 | *err = -ENOMEM; | 712 | *err = -ENOMEM; |
711 | return oldtnode; | 713 | return oldtnode; |
712 | } | 714 | } |
713 | 715 | ||
714 | for(i = 0; i < olen; i++) { | 716 | for(i = 0; i < olen; i++) { |
715 | struct node *node = tnode_get_child(oldtnode, i); | 717 | struct node *node = tnode_get_child(oldtnode, i); |
716 | 718 | ||
717 | /* An empty child */ | 719 | /* An empty child */ |
718 | if (node == NULL) | 720 | if (node == NULL) |
719 | continue; | 721 | continue; |
720 | 722 | ||
721 | /* A leaf or an internal node with skipped bits */ | 723 | /* A leaf or an internal node with skipped bits */ |
722 | 724 | ||
723 | if(IS_LEAF(node) || ((struct tnode *) node)->pos > | 725 | if (IS_LEAF(node) || ((struct tnode *) node)->pos > |
724 | tn->pos + tn->bits - 1) { | 726 | tn->pos + tn->bits - 1) { |
725 | if(tkey_extract_bits(node->key, oldtnode->pos + oldtnode->bits, | 727 | if (tkey_extract_bits(node->key, oldtnode->pos + oldtnode->bits, |
726 | 1) == 0) | 728 | 1) == 0) |
727 | put_child(t, tn, 2*i, node); | 729 | put_child(t, tn, 2*i, node); |
728 | else | 730 | else |
@@ -745,37 +747,37 @@ static struct tnode *inflate(struct trie *t, struct tnode *tn, int *err) | |||
745 | struct tnode *left, *right; | 747 | struct tnode *left, *right; |
746 | int size, j; | 748 | int size, j; |
747 | 749 | ||
748 | /* We will replace this node 'inode' with two new | 750 | /* We will replace this node 'inode' with two new |
749 | * ones, 'left' and 'right', each with half of the | 751 | * ones, 'left' and 'right', each with half of the |
750 | * original children. The two new nodes will have | 752 | * original children. The two new nodes will have |
751 | * a position one bit further down the key and this | 753 | * a position one bit further down the key and this |
752 | * means that the "significant" part of their keys | 754 | * means that the "significant" part of their keys |
753 | * (see the discussion near the top of this file) | 755 | * (see the discussion near the top of this file) |
754 | * will differ by one bit, which will be "0" in | 756 | * will differ by one bit, which will be "0" in |
755 | * left's key and "1" in right's key. Since we are | 757 | * left's key and "1" in right's key. Since we are |
756 | * moving the key position by one step, the bit that | 758 | * moving the key position by one step, the bit that |
757 | * we are moving away from - the bit at position | 759 | * we are moving away from - the bit at position |
758 | * (inode->pos) - is the one that will differ between | 760 | * (inode->pos) - is the one that will differ between |
759 | * left and right. So... we synthesize that bit in the | 761 | * left and right. So... we synthesize that bit in the |
760 | * two new keys. | 762 | * two new keys. |
761 | * The mask 'm' below will be a single "one" bit at | 763 | * The mask 'm' below will be a single "one" bit at |
762 | * the position (inode->pos) | 764 | * the position (inode->pos) |
763 | */ | 765 | */ |
764 | 766 | ||
765 | /* Use the old key, but set the new significant | 767 | /* Use the old key, but set the new significant |
766 | * bit to zero. | 768 | * bit to zero. |
767 | */ | 769 | */ |
768 | 770 | ||
769 | left = (struct tnode *) tnode_get_child(tn, 2*i); | 771 | left = (struct tnode *) tnode_get_child(tn, 2*i); |
770 | put_child(t, tn, 2*i, NULL); | 772 | put_child(t, tn, 2*i, NULL); |
771 | 773 | ||
772 | if(!left) | 774 | if (!left) |
773 | BUG(); | 775 | BUG(); |
774 | 776 | ||
775 | right = (struct tnode *) tnode_get_child(tn, 2*i+1); | 777 | right = (struct tnode *) tnode_get_child(tn, 2*i+1); |
776 | put_child(t, tn, 2*i+1, NULL); | 778 | put_child(t, tn, 2*i+1, NULL); |
777 | 779 | ||
778 | if(!right) | 780 | if (!right) |
779 | BUG(); | 781 | BUG(); |
780 | 782 | ||
781 | size = tnode_child_length(left); | 783 | size = tnode_child_length(left); |
@@ -800,9 +802,9 @@ static struct tnode *halve(struct trie *t, struct tnode *tn, int *err) | |||
800 | int i; | 802 | int i; |
801 | int olen = tnode_child_length(tn); | 803 | int olen = tnode_child_length(tn); |
802 | 804 | ||
803 | if(trie_debug) printk("In halve\n"); | 805 | if (trie_debug) printk("In halve\n"); |
804 | 806 | ||
805 | tn=tnode_new(oldtnode->key, oldtnode->pos, oldtnode->bits - 1); | 807 | tn = tnode_new(oldtnode->key, oldtnode->pos, oldtnode->bits - 1); |
806 | 808 | ||
807 | if (!tn) { | 809 | if (!tn) { |
808 | *err = -ENOMEM; | 810 | *err = -ENOMEM; |
@@ -810,39 +812,39 @@ static struct tnode *halve(struct trie *t, struct tnode *tn, int *err) | |||
810 | } | 812 | } |
811 | 813 | ||
812 | /* | 814 | /* |
813 | * Preallocate and store tnodes before the actual work so we | 815 | * Preallocate and store tnodes before the actual work so we |
814 | * don't get into an inconsistent state if memory allocation | 816 | * don't get into an inconsistent state if memory allocation |
815 | * fails. In case of failure we return the oldnode and halve | 817 | * fails. In case of failure we return the oldnode and halve |
816 | * of tnode is ignored. | 818 | * of tnode is ignored. |
817 | */ | 819 | */ |
818 | 820 | ||
819 | for(i = 0; i < olen; i += 2) { | 821 | for(i = 0; i < olen; i += 2) { |
820 | left = tnode_get_child(oldtnode, i); | 822 | left = tnode_get_child(oldtnode, i); |
821 | right = tnode_get_child(oldtnode, i+1); | 823 | right = tnode_get_child(oldtnode, i+1); |
822 | 824 | ||
823 | /* Two nonempty children */ | 825 | /* Two nonempty children */ |
824 | if( left && right) { | 826 | if (left && right) { |
825 | struct tnode *newBinNode = | 827 | struct tnode *newBinNode = |
826 | tnode_new(left->key, tn->pos + tn->bits, 1); | 828 | tnode_new(left->key, tn->pos + tn->bits, 1); |
827 | 829 | ||
828 | if(!newBinNode) { | 830 | if (!newBinNode) { |
829 | *err = -ENOMEM; | 831 | *err = -ENOMEM; |
830 | break; | 832 | break; |
831 | } | 833 | } |
832 | put_child(t, tn, i/2, (struct node *)newBinNode); | 834 | put_child(t, tn, i/2, (struct node *)newBinNode); |
833 | } | 835 | } |
834 | } | 836 | } |
835 | 837 | ||
836 | if(*err) { | 838 | if (*err) { |
837 | int size = tnode_child_length(tn); | 839 | int size = tnode_child_length(tn); |
838 | int j; | 840 | int j; |
839 | 841 | ||
840 | for(j = 0; j < size; j++) | 842 | for(j = 0; j < size; j++) |
841 | if( tn->child[j]) | 843 | if (tn->child[j]) |
842 | tnode_free((struct tnode *)tn->child[j]); | 844 | tnode_free((struct tnode *)tn->child[j]); |
843 | 845 | ||
844 | tnode_free(tn); | 846 | tnode_free(tn); |
845 | 847 | ||
846 | *err = -ENOMEM; | 848 | *err = -ENOMEM; |
847 | return oldtnode; | 849 | return oldtnode; |
848 | } | 850 | } |
@@ -850,7 +852,7 @@ static struct tnode *halve(struct trie *t, struct tnode *tn, int *err) | |||
850 | for(i = 0; i < olen; i += 2) { | 852 | for(i = 0; i < olen; i += 2) { |
851 | left = tnode_get_child(oldtnode, i); | 853 | left = tnode_get_child(oldtnode, i); |
852 | right = tnode_get_child(oldtnode, i+1); | 854 | right = tnode_get_child(oldtnode, i+1); |
853 | 855 | ||
854 | /* At least one of the children is empty */ | 856 | /* At least one of the children is empty */ |
855 | if (left == NULL) { | 857 | if (left == NULL) { |
856 | if (right == NULL) /* Both are empty */ | 858 | if (right == NULL) /* Both are empty */ |
@@ -858,14 +860,14 @@ static struct tnode *halve(struct trie *t, struct tnode *tn, int *err) | |||
858 | put_child(t, tn, i/2, right); | 860 | put_child(t, tn, i/2, right); |
859 | } else if (right == NULL) | 861 | } else if (right == NULL) |
860 | put_child(t, tn, i/2, left); | 862 | put_child(t, tn, i/2, left); |
861 | 863 | ||
862 | /* Two nonempty children */ | 864 | /* Two nonempty children */ |
863 | else { | 865 | else { |
864 | struct tnode *newBinNode = | 866 | struct tnode *newBinNode = |
865 | (struct tnode *) tnode_get_child(tn, i/2); | 867 | (struct tnode *) tnode_get_child(tn, i/2); |
866 | put_child(t, tn, i/2, NULL); | 868 | put_child(t, tn, i/2, NULL); |
867 | 869 | ||
868 | if(!newBinNode) | 870 | if (!newBinNode) |
869 | BUG(); | 871 | BUG(); |
870 | 872 | ||
871 | put_child(t, newBinNode, 0, left); | 873 | put_child(t, newBinNode, 0, left); |
@@ -879,7 +881,7 @@ static struct tnode *halve(struct trie *t, struct tnode *tn, int *err) | |||
879 | 881 | ||
880 | static void *trie_init(struct trie *t) | 882 | static void *trie_init(struct trie *t) |
881 | { | 883 | { |
882 | if(t) { | 884 | if (t) { |
883 | t->size = 0; | 885 | t->size = 0; |
884 | t->trie = NULL; | 886 | t->trie = NULL; |
885 | t->revision = 0; | 887 | t->revision = 0; |
@@ -896,8 +898,7 @@ static struct leaf_info *find_leaf_info(struct hlist_head *head, int plen) | |||
896 | struct leaf_info *li; | 898 | struct leaf_info *li; |
897 | 899 | ||
898 | hlist_for_each_entry(li, node, head, hlist) { | 900 | hlist_for_each_entry(li, node, head, hlist) { |
899 | 901 | if (li->plen == plen) | |
900 | if ( li->plen == plen ) | ||
901 | return li; | 902 | return li; |
902 | } | 903 | } |
903 | return NULL; | 904 | return NULL; |
@@ -905,35 +906,35 @@ static struct leaf_info *find_leaf_info(struct hlist_head *head, int plen) | |||
905 | 906 | ||
906 | static inline struct list_head * get_fa_head(struct leaf *l, int plen) | 907 | static inline struct list_head * get_fa_head(struct leaf *l, int plen) |
907 | { | 908 | { |
908 | struct list_head *fa_head=NULL; | 909 | struct list_head *fa_head = NULL; |
909 | struct leaf_info *li = find_leaf_info(&l->list, plen); | 910 | struct leaf_info *li = find_leaf_info(&l->list, plen); |
910 | 911 | ||
911 | if(li) | 912 | if (li) |
912 | fa_head = &li->falh; | 913 | fa_head = &li->falh; |
913 | 914 | ||
914 | return fa_head; | 915 | return fa_head; |
915 | } | 916 | } |
916 | 917 | ||
917 | static void insert_leaf_info(struct hlist_head *head, struct leaf_info *new) | 918 | static void insert_leaf_info(struct hlist_head *head, struct leaf_info *new) |
918 | { | 919 | { |
919 | struct leaf_info *li=NULL, *last=NULL; | 920 | struct leaf_info *li = NULL, *last = NULL; |
920 | struct hlist_node *node, *tmp; | 921 | struct hlist_node *node, *tmp; |
921 | 922 | ||
922 | write_lock_bh(&fib_lock); | 923 | write_lock_bh(&fib_lock); |
923 | 924 | ||
924 | if(hlist_empty(head)) | 925 | if (hlist_empty(head)) |
925 | hlist_add_head(&new->hlist, head); | 926 | hlist_add_head(&new->hlist, head); |
926 | else { | 927 | else { |
927 | hlist_for_each_entry_safe(li, node, tmp, head, hlist) { | 928 | hlist_for_each_entry_safe(li, node, tmp, head, hlist) { |
928 | 929 | ||
929 | if (new->plen > li->plen) | 930 | if (new->plen > li->plen) |
930 | break; | 931 | break; |
931 | 932 | ||
932 | last = li; | 933 | last = li; |
933 | } | 934 | } |
934 | if(last) | 935 | if (last) |
935 | hlist_add_after(&last->hlist, &new->hlist); | 936 | hlist_add_after(&last->hlist, &new->hlist); |
936 | else | 937 | else |
937 | hlist_add_before(&new->hlist, &li->hlist); | 938 | hlist_add_before(&new->hlist, &li->hlist); |
938 | } | 939 | } |
939 | write_unlock_bh(&fib_lock); | 940 | write_unlock_bh(&fib_lock); |
@@ -947,14 +948,14 @@ fib_find_node(struct trie *t, u32 key) | |||
947 | struct node *n; | 948 | struct node *n; |
948 | 949 | ||
949 | pos = 0; | 950 | pos = 0; |
950 | n=t->trie; | 951 | n = t->trie; |
951 | 952 | ||
952 | while (n != NULL && NODE_TYPE(n) == T_TNODE) { | 953 | while (n != NULL && NODE_TYPE(n) == T_TNODE) { |
953 | tn = (struct tnode *) n; | 954 | tn = (struct tnode *) n; |
954 | 955 | ||
955 | check_tnode(tn); | 956 | check_tnode(tn); |
956 | 957 | ||
957 | if(tkey_sub_equals(tn->key, pos, tn->pos-pos, key)) { | 958 | if (tkey_sub_equals(tn->key, pos, tn->pos-pos, key)) { |
958 | pos=tn->pos + tn->bits; | 959 | pos=tn->pos + tn->bits; |
959 | n = tnode_get_child(tn, tkey_extract_bits(key, tn->pos, tn->bits)); | 960 | n = tnode_get_child(tn, tkey_extract_bits(key, tn->pos, tn->bits)); |
960 | } | 961 | } |
@@ -977,23 +978,23 @@ static struct node *trie_rebalance(struct trie *t, struct tnode *tn) | |||
977 | t_key cindex, key; | 978 | t_key cindex, key; |
978 | struct tnode *tp = NULL; | 979 | struct tnode *tp = NULL; |
979 | 980 | ||
980 | if(!tn) | 981 | if (!tn) |
981 | BUG(); | 982 | BUG(); |
982 | 983 | ||
983 | key = tn->key; | 984 | key = tn->key; |
984 | i = 0; | 985 | i = 0; |
985 | 986 | ||
986 | while (tn != NULL && NODE_PARENT(tn) != NULL) { | 987 | while (tn != NULL && NODE_PARENT(tn) != NULL) { |
987 | 988 | ||
988 | if( i > 10 ) { | 989 | if (i > 10) { |
989 | printk("Rebalance tn=%p \n", tn); | 990 | printk("Rebalance tn=%p \n", tn); |
990 | if(tn) printk("tn->parent=%p \n", NODE_PARENT(tn)); | 991 | if (tn) printk("tn->parent=%p \n", NODE_PARENT(tn)); |
991 | 992 | ||
992 | printk("Rebalance tp=%p \n", tp); | 993 | printk("Rebalance tp=%p \n", tp); |
993 | if(tp) printk("tp->parent=%p \n", NODE_PARENT(tp)); | 994 | if (tp) printk("tp->parent=%p \n", NODE_PARENT(tp)); |
994 | } | 995 | } |
995 | 996 | ||
996 | if( i > 12 ) BUG(); | 997 | if (i > 12) BUG(); |
997 | i++; | 998 | i++; |
998 | 999 | ||
999 | tp = NODE_PARENT(tn); | 1000 | tp = NODE_PARENT(tn); |
@@ -1001,14 +1002,14 @@ static struct node *trie_rebalance(struct trie *t, struct tnode *tn) | |||
1001 | wasfull = tnode_full(tp, tnode_get_child(tp, cindex)); | 1002 | wasfull = tnode_full(tp, tnode_get_child(tp, cindex)); |
1002 | tn = (struct tnode *) resize (t, (struct tnode *)tn); | 1003 | tn = (struct tnode *) resize (t, (struct tnode *)tn); |
1003 | tnode_put_child_reorg((struct tnode *)tp, cindex,(struct node*)tn, wasfull); | 1004 | tnode_put_child_reorg((struct tnode *)tp, cindex,(struct node*)tn, wasfull); |
1004 | 1005 | ||
1005 | if(!NODE_PARENT(tn)) | 1006 | if (!NODE_PARENT(tn)) |
1006 | break; | 1007 | break; |
1007 | 1008 | ||
1008 | tn = NODE_PARENT(tn); | 1009 | tn = NODE_PARENT(tn); |
1009 | } | 1010 | } |
1010 | /* Handle last (top) tnode */ | 1011 | /* Handle last (top) tnode */ |
1011 | if (IS_TNODE(tn)) | 1012 | if (IS_TNODE(tn)) |
1012 | tn = (struct tnode*) resize(t, (struct tnode *)tn); | 1013 | tn = (struct tnode*) resize(t, (struct tnode *)tn); |
1013 | 1014 | ||
1014 | return (struct node*) tn; | 1015 | return (struct node*) tn; |
@@ -1022,42 +1023,42 @@ fib_insert_node(struct trie *t, int *err, u32 key, int plen) | |||
1022 | struct node *n; | 1023 | struct node *n; |
1023 | struct leaf *l; | 1024 | struct leaf *l; |
1024 | int missbit; | 1025 | int missbit; |
1025 | struct list_head *fa_head=NULL; | 1026 | struct list_head *fa_head = NULL; |
1026 | struct leaf_info *li; | 1027 | struct leaf_info *li; |
1027 | t_key cindex; | 1028 | t_key cindex; |
1028 | 1029 | ||
1029 | pos = 0; | 1030 | pos = 0; |
1030 | n=t->trie; | 1031 | n = t->trie; |
1031 | 1032 | ||
1032 | /* If we point to NULL, stop. Either the tree is empty and we should | 1033 | /* If we point to NULL, stop. Either the tree is empty and we should |
1033 | * just put a new leaf in if, or we have reached an empty child slot, | 1034 | * just put a new leaf in if, or we have reached an empty child slot, |
1034 | * and we should just put our new leaf in that. | 1035 | * and we should just put our new leaf in that. |
1035 | * If we point to a T_TNODE, check if it matches our key. Note that | 1036 | * If we point to a T_TNODE, check if it matches our key. Note that |
1036 | * a T_TNODE might be skipping any number of bits - its 'pos' need | 1037 | * a T_TNODE might be skipping any number of bits - its 'pos' need |
1037 | * not be the parent's 'pos'+'bits'! | 1038 | * not be the parent's 'pos'+'bits'! |
1038 | * | 1039 | * |
1039 | * If it does match the current key, get pos/bits from it, extract | 1040 | * If it does match the current key, get pos/bits from it, extract |
1040 | * the index from our key, push the T_TNODE and walk the tree. | 1041 | * the index from our key, push the T_TNODE and walk the tree. |
1041 | * | 1042 | * |
1042 | * If it doesn't, we have to replace it with a new T_TNODE. | 1043 | * If it doesn't, we have to replace it with a new T_TNODE. |
1043 | * | 1044 | * |
1044 | * If we point to a T_LEAF, it might or might not have the same key | 1045 | * If we point to a T_LEAF, it might or might not have the same key |
1045 | * as we do. If it does, just change the value, update the T_LEAF's | 1046 | * as we do. If it does, just change the value, update the T_LEAF's |
1046 | * value, and return it. | 1047 | * value, and return it. |
1047 | * If it doesn't, we need to replace it with a T_TNODE. | 1048 | * If it doesn't, we need to replace it with a T_TNODE. |
1048 | */ | 1049 | */ |
1049 | 1050 | ||
1050 | while (n != NULL && NODE_TYPE(n) == T_TNODE) { | 1051 | while (n != NULL && NODE_TYPE(n) == T_TNODE) { |
1051 | tn = (struct tnode *) n; | 1052 | tn = (struct tnode *) n; |
1052 | |||
1053 | check_tnode(tn); | ||
1054 | 1053 | ||
1055 | if(tkey_sub_equals(tn->key, pos, tn->pos-pos, key)) { | 1054 | check_tnode(tn); |
1055 | |||
1056 | if (tkey_sub_equals(tn->key, pos, tn->pos-pos, key)) { | ||
1056 | tp = tn; | 1057 | tp = tn; |
1057 | pos=tn->pos + tn->bits; | 1058 | pos=tn->pos + tn->bits; |
1058 | n = tnode_get_child(tn, tkey_extract_bits(key, tn->pos, tn->bits)); | 1059 | n = tnode_get_child(tn, tkey_extract_bits(key, tn->pos, tn->bits)); |
1059 | 1060 | ||
1060 | if(n && NODE_PARENT(n) != tn) { | 1061 | if (n && NODE_PARENT(n) != tn) { |
1061 | printk("BUG tn=%p, n->parent=%p\n", tn, NODE_PARENT(n)); | 1062 | printk("BUG tn=%p, n->parent=%p\n", tn, NODE_PARENT(n)); |
1062 | BUG(); | 1063 | BUG(); |
1063 | } | 1064 | } |
@@ -1069,21 +1070,21 @@ fib_insert_node(struct trie *t, int *err, u32 key, int plen) | |||
1069 | /* | 1070 | /* |
1070 | * n ----> NULL, LEAF or TNODE | 1071 | * n ----> NULL, LEAF or TNODE |
1071 | * | 1072 | * |
1072 | * tp is n's (parent) ----> NULL or TNODE | 1073 | * tp is n's (parent) ----> NULL or TNODE |
1073 | */ | 1074 | */ |
1074 | 1075 | ||
1075 | if(tp && IS_LEAF(tp)) | 1076 | if (tp && IS_LEAF(tp)) |
1076 | BUG(); | 1077 | BUG(); |
1077 | 1078 | ||
1078 | 1079 | ||
1079 | /* Case 1: n is a leaf. Compare prefixes */ | 1080 | /* Case 1: n is a leaf. Compare prefixes */ |
1080 | 1081 | ||
1081 | if (n != NULL && IS_LEAF(n) && tkey_equals(key, n->key)) { | 1082 | if (n != NULL && IS_LEAF(n) && tkey_equals(key, n->key)) { |
1082 | struct leaf *l = ( struct leaf *) n; | 1083 | struct leaf *l = ( struct leaf *) n; |
1083 | 1084 | ||
1084 | li = leaf_info_new(plen); | 1085 | li = leaf_info_new(plen); |
1085 | 1086 | ||
1086 | if(! li) { | 1087 | if (!li) { |
1087 | *err = -ENOMEM; | 1088 | *err = -ENOMEM; |
1088 | goto err; | 1089 | goto err; |
1089 | } | 1090 | } |
@@ -1095,7 +1096,7 @@ fib_insert_node(struct trie *t, int *err, u32 key, int plen) | |||
1095 | t->size++; | 1096 | t->size++; |
1096 | l = leaf_new(); | 1097 | l = leaf_new(); |
1097 | 1098 | ||
1098 | if(! l) { | 1099 | if (!l) { |
1099 | *err = -ENOMEM; | 1100 | *err = -ENOMEM; |
1100 | goto err; | 1101 | goto err; |
1101 | } | 1102 | } |
@@ -1103,7 +1104,7 @@ fib_insert_node(struct trie *t, int *err, u32 key, int plen) | |||
1103 | l->key = key; | 1104 | l->key = key; |
1104 | li = leaf_info_new(plen); | 1105 | li = leaf_info_new(plen); |
1105 | 1106 | ||
1106 | if(! li) { | 1107 | if (!li) { |
1107 | tnode_free((struct tnode *) l); | 1108 | tnode_free((struct tnode *) l); |
1108 | *err = -ENOMEM; | 1109 | *err = -ENOMEM; |
1109 | goto err; | 1110 | goto err; |
@@ -1116,8 +1117,8 @@ fib_insert_node(struct trie *t, int *err, u32 key, int plen) | |||
1116 | if (t->trie && n == NULL) { | 1117 | if (t->trie && n == NULL) { |
1117 | 1118 | ||
1118 | NODE_SET_PARENT(l, tp); | 1119 | NODE_SET_PARENT(l, tp); |
1119 | 1120 | ||
1120 | if (!tp) | 1121 | if (!tp) |
1121 | BUG(); | 1122 | BUG(); |
1122 | 1123 | ||
1123 | else { | 1124 | else { |
@@ -1127,8 +1128,8 @@ fib_insert_node(struct trie *t, int *err, u32 key, int plen) | |||
1127 | } | 1128 | } |
1128 | /* Case 3: n is a LEAF or a TNODE and the key doesn't match. */ | 1129 | /* Case 3: n is a LEAF or a TNODE and the key doesn't match. */ |
1129 | else { | 1130 | else { |
1130 | /* | 1131 | /* |
1131 | * Add a new tnode here | 1132 | * Add a new tnode here |
1132 | * first tnode need some special handling | 1133 | * first tnode need some special handling |
1133 | */ | 1134 | */ |
1134 | 1135 | ||
@@ -1136,39 +1137,39 @@ fib_insert_node(struct trie *t, int *err, u32 key, int plen) | |||
1136 | pos=tp->pos+tp->bits; | 1137 | pos=tp->pos+tp->bits; |
1137 | else | 1138 | else |
1138 | pos=0; | 1139 | pos=0; |
1139 | if(n) { | 1140 | if (n) { |
1140 | newpos = tkey_mismatch(key, pos, n->key); | 1141 | newpos = tkey_mismatch(key, pos, n->key); |
1141 | tn = tnode_new(n->key, newpos, 1); | 1142 | tn = tnode_new(n->key, newpos, 1); |
1142 | } | 1143 | } |
1143 | else { | 1144 | else { |
1144 | newpos = 0; | 1145 | newpos = 0; |
1145 | tn = tnode_new(key, newpos, 1); /* First tnode */ | 1146 | tn = tnode_new(key, newpos, 1); /* First tnode */ |
1146 | } | 1147 | } |
1147 | 1148 | ||
1148 | if(!tn) { | 1149 | if (!tn) { |
1149 | free_leaf_info(li); | 1150 | free_leaf_info(li); |
1150 | tnode_free((struct tnode *) l); | 1151 | tnode_free((struct tnode *) l); |
1151 | *err = -ENOMEM; | 1152 | *err = -ENOMEM; |
1152 | goto err; | 1153 | goto err; |
1153 | } | 1154 | } |
1154 | 1155 | ||
1155 | NODE_SET_PARENT(tn, tp); | 1156 | NODE_SET_PARENT(tn, tp); |
1156 | 1157 | ||
1157 | missbit=tkey_extract_bits(key, newpos, 1); | 1158 | missbit=tkey_extract_bits(key, newpos, 1); |
1158 | put_child(t, tn, missbit, (struct node *)l); | 1159 | put_child(t, tn, missbit, (struct node *)l); |
1159 | put_child(t, tn, 1-missbit, n); | 1160 | put_child(t, tn, 1-missbit, n); |
1160 | 1161 | ||
1161 | if(tp) { | 1162 | if (tp) { |
1162 | cindex = tkey_extract_bits(key, tp->pos, tp->bits); | 1163 | cindex = tkey_extract_bits(key, tp->pos, tp->bits); |
1163 | put_child(t, (struct tnode *)tp, cindex, (struct node *)tn); | 1164 | put_child(t, (struct tnode *)tp, cindex, (struct node *)tn); |
1164 | } | 1165 | } |
1165 | else { | 1166 | else { |
1166 | t->trie = (struct node*) tn; /* First tnode */ | 1167 | t->trie = (struct node*) tn; /* First tnode */ |
1167 | tp = tn; | 1168 | tp = tn; |
1168 | } | 1169 | } |
1169 | } | 1170 | } |
1170 | if(tp && tp->pos+tp->bits > 32) { | 1171 | if (tp && tp->pos+tp->bits > 32) { |
1171 | printk("ERROR tp=%p pos=%d, bits=%d, key=%0x plen=%d\n", | 1172 | printk("ERROR tp=%p pos=%d, bits=%d, key=%0x plen=%d\n", |
1172 | tp, tp->pos, tp->bits, key, plen); | 1173 | tp, tp->pos, tp->bits, key, plen); |
1173 | } | 1174 | } |
1174 | /* Rebalance the trie */ | 1175 | /* Rebalance the trie */ |
@@ -1185,7 +1186,7 @@ fn_trie_insert(struct fib_table *tb, struct rtmsg *r, struct kern_rta *rta, | |||
1185 | { | 1186 | { |
1186 | struct trie *t = (struct trie *) tb->tb_data; | 1187 | struct trie *t = (struct trie *) tb->tb_data; |
1187 | struct fib_alias *fa, *new_fa; | 1188 | struct fib_alias *fa, *new_fa; |
1188 | struct list_head *fa_head=NULL; | 1189 | struct list_head *fa_head = NULL; |
1189 | struct fib_info *fi; | 1190 | struct fib_info *fi; |
1190 | int plen = r->rtm_dst_len; | 1191 | int plen = r->rtm_dst_len; |
1191 | int type = r->rtm_type; | 1192 | int type = r->rtm_type; |
@@ -1198,17 +1199,17 @@ fn_trie_insert(struct fib_table *tb, struct rtmsg *r, struct kern_rta *rta, | |||
1198 | return -EINVAL; | 1199 | return -EINVAL; |
1199 | 1200 | ||
1200 | key = 0; | 1201 | key = 0; |
1201 | if (rta->rta_dst) | 1202 | if (rta->rta_dst) |
1202 | memcpy(&key, rta->rta_dst, 4); | 1203 | memcpy(&key, rta->rta_dst, 4); |
1203 | 1204 | ||
1204 | key = ntohl(key); | 1205 | key = ntohl(key); |
1205 | 1206 | ||
1206 | if(trie_debug) | 1207 | if (trie_debug) |
1207 | printk("Insert table=%d %08x/%d\n", tb->tb_id, key, plen); | 1208 | printk("Insert table=%d %08x/%d\n", tb->tb_id, key, plen); |
1208 | 1209 | ||
1209 | mask = ntohl( inet_make_mask(plen) ); | 1210 | mask = ntohl( inet_make_mask(plen) ); |
1210 | 1211 | ||
1211 | if(key & ~mask) | 1212 | if (key & ~mask) |
1212 | return -EINVAL; | 1213 | return -EINVAL; |
1213 | 1214 | ||
1214 | key = key & mask; | 1215 | key = key & mask; |
@@ -1217,9 +1218,9 @@ fn_trie_insert(struct fib_table *tb, struct rtmsg *r, struct kern_rta *rta, | |||
1217 | goto err; | 1218 | goto err; |
1218 | 1219 | ||
1219 | l = fib_find_node(t, key); | 1220 | l = fib_find_node(t, key); |
1220 | fa = NULL; | 1221 | fa = NULL; |
1221 | 1222 | ||
1222 | if(l) { | 1223 | if (l) { |
1223 | fa_head = get_fa_head(l, plen); | 1224 | fa_head = get_fa_head(l, plen); |
1224 | fa = fib_find_alias(fa_head, tos, fi->fib_priority); | 1225 | fa = fib_find_alias(fa_head, tos, fi->fib_priority); |
1225 | } | 1226 | } |
@@ -1298,16 +1299,16 @@ fn_trie_insert(struct fib_table *tb, struct rtmsg *r, struct kern_rta *rta, | |||
1298 | new_fa->fa_scope = r->rtm_scope; | 1299 | new_fa->fa_scope = r->rtm_scope; |
1299 | new_fa->fa_state = 0; | 1300 | new_fa->fa_state = 0; |
1300 | #if 0 | 1301 | #if 0 |
1301 | new_fa->dst = NULL; | 1302 | new_fa->dst = NULL; |
1302 | #endif | 1303 | #endif |
1303 | /* | 1304 | /* |
1304 | * Insert new entry to the list. | 1305 | * Insert new entry to the list. |
1305 | */ | 1306 | */ |
1306 | 1307 | ||
1307 | if(!fa_head) { | 1308 | if (!fa_head) { |
1308 | fa_head = fib_insert_node(t, &err, key, plen); | 1309 | fa_head = fib_insert_node(t, &err, key, plen); |
1309 | err = 0; | 1310 | err = 0; |
1310 | if(err) | 1311 | if (err) |
1311 | goto out_free_new_fa; | 1312 | goto out_free_new_fa; |
1312 | } | 1313 | } |
1313 | 1314 | ||
@@ -1327,38 +1328,38 @@ out_free_new_fa: | |||
1327 | kmem_cache_free(fn_alias_kmem, new_fa); | 1328 | kmem_cache_free(fn_alias_kmem, new_fa); |
1328 | out: | 1329 | out: |
1329 | fib_release_info(fi); | 1330 | fib_release_info(fi); |
1330 | err:; | 1331 | err:; |
1331 | return err; | 1332 | return err; |
1332 | } | 1333 | } |
1333 | 1334 | ||
1334 | static inline int check_leaf(struct trie *t, struct leaf *l, t_key key, int *plen, const struct flowi *flp, | 1335 | static inline int check_leaf(struct trie *t, struct leaf *l, t_key key, int *plen, const struct flowi *flp, |
1335 | struct fib_result *res, int *err) | 1336 | struct fib_result *res) |
1336 | { | 1337 | { |
1337 | int i; | 1338 | int err, i; |
1338 | t_key mask; | 1339 | t_key mask; |
1339 | struct leaf_info *li; | 1340 | struct leaf_info *li; |
1340 | struct hlist_head *hhead = &l->list; | 1341 | struct hlist_head *hhead = &l->list; |
1341 | struct hlist_node *node; | 1342 | struct hlist_node *node; |
1342 | 1343 | ||
1343 | hlist_for_each_entry(li, node, hhead, hlist) { | 1344 | hlist_for_each_entry(li, node, hhead, hlist) { |
1344 | 1345 | ||
1345 | i = li->plen; | 1346 | i = li->plen; |
1346 | mask = ntohl(inet_make_mask(i)); | 1347 | mask = ntohl(inet_make_mask(i)); |
1347 | if (l->key != (key & mask)) | 1348 | if (l->key != (key & mask)) |
1348 | continue; | 1349 | continue; |
1349 | 1350 | ||
1350 | if (((*err) = fib_semantic_match(&li->falh, flp, res, l->key, mask, i)) == 0) { | 1351 | if ((err = fib_semantic_match(&li->falh, flp, res, l->key, mask, i)) <= 0) { |
1351 | *plen = i; | 1352 | *plen = i; |
1352 | #ifdef CONFIG_IP_FIB_TRIE_STATS | 1353 | #ifdef CONFIG_IP_FIB_TRIE_STATS |
1353 | t->stats.semantic_match_passed++; | 1354 | t->stats.semantic_match_passed++; |
1354 | #endif | 1355 | #endif |
1355 | return 1; | 1356 | return err; |
1356 | } | 1357 | } |
1357 | #ifdef CONFIG_IP_FIB_TRIE_STATS | 1358 | #ifdef CONFIG_IP_FIB_TRIE_STATS |
1358 | t->stats.semantic_match_miss++; | 1359 | t->stats.semantic_match_miss++; |
1359 | #endif | 1360 | #endif |
1360 | } | 1361 | } |
1361 | return 0; | 1362 | return 1; |
1362 | } | 1363 | } |
1363 | 1364 | ||
1364 | static int | 1365 | static int |
@@ -1376,7 +1377,7 @@ fn_trie_lookup(struct fib_table *tb, const struct flowi *flp, struct fib_result | |||
1376 | n = t->trie; | 1377 | n = t->trie; |
1377 | 1378 | ||
1378 | read_lock(&fib_lock); | 1379 | read_lock(&fib_lock); |
1379 | if(!n) | 1380 | if (!n) |
1380 | goto failed; | 1381 | goto failed; |
1381 | 1382 | ||
1382 | #ifdef CONFIG_IP_FIB_TRIE_STATS | 1383 | #ifdef CONFIG_IP_FIB_TRIE_STATS |
@@ -1385,19 +1386,19 @@ fn_trie_lookup(struct fib_table *tb, const struct flowi *flp, struct fib_result | |||
1385 | 1386 | ||
1386 | /* Just a leaf? */ | 1387 | /* Just a leaf? */ |
1387 | if (IS_LEAF(n)) { | 1388 | if (IS_LEAF(n)) { |
1388 | if( check_leaf(t, (struct leaf *)n, key, &plen, flp, res, &ret) ) | 1389 | if ((ret = check_leaf(t, (struct leaf *)n, key, &plen, flp, res)) <= 0) |
1389 | goto found; | 1390 | goto found; |
1390 | goto failed; | 1391 | goto failed; |
1391 | } | 1392 | } |
1392 | pn = (struct tnode *) n; | 1393 | pn = (struct tnode *) n; |
1393 | chopped_off = 0; | 1394 | chopped_off = 0; |
1394 | 1395 | ||
1395 | while (pn) { | 1396 | while (pn) { |
1396 | 1397 | ||
1397 | pos = pn->pos; | 1398 | pos = pn->pos; |
1398 | bits = pn->bits; | 1399 | bits = pn->bits; |
1399 | 1400 | ||
1400 | if(!chopped_off) | 1401 | if (!chopped_off) |
1401 | cindex = tkey_extract_bits(MASK_PFX(key, current_prefix_length), pos, bits); | 1402 | cindex = tkey_extract_bits(MASK_PFX(key, current_prefix_length), pos, bits); |
1402 | 1403 | ||
1403 | n = tnode_get_child(pn, cindex); | 1404 | n = tnode_get_child(pn, cindex); |
@@ -1417,33 +1418,33 @@ fn_trie_lookup(struct fib_table *tb, const struct flowi *flp, struct fib_result | |||
1417 | int mp; | 1418 | int mp; |
1418 | 1419 | ||
1419 | /* | 1420 | /* |
1420 | * It's a tnode, and we can do some extra checks here if we | 1421 | * It's a tnode, and we can do some extra checks here if we |
1421 | * like, to avoid descending into a dead-end branch. | 1422 | * like, to avoid descending into a dead-end branch. |
1422 | * This tnode is in the parent's child array at index | 1423 | * This tnode is in the parent's child array at index |
1423 | * key[p_pos..p_pos+p_bits] but potentially with some bits | 1424 | * key[p_pos..p_pos+p_bits] but potentially with some bits |
1424 | * chopped off, so in reality the index may be just a | 1425 | * chopped off, so in reality the index may be just a |
1425 | * subprefix, padded with zero at the end. | 1426 | * subprefix, padded with zero at the end. |
1426 | * We can also take a look at any skipped bits in this | 1427 | * We can also take a look at any skipped bits in this |
1427 | * tnode - everything up to p_pos is supposed to be ok, | 1428 | * tnode - everything up to p_pos is supposed to be ok, |
1428 | * and the non-chopped bits of the index (se previous | 1429 | * and the non-chopped bits of the index (se previous |
1429 | * paragraph) are also guaranteed ok, but the rest is | 1430 | * paragraph) are also guaranteed ok, but the rest is |
1430 | * considered unknown. | 1431 | * considered unknown. |
1431 | * | 1432 | * |
1432 | * The skipped bits are key[pos+bits..cn->pos]. | 1433 | * The skipped bits are key[pos+bits..cn->pos]. |
1433 | */ | 1434 | */ |
1434 | 1435 | ||
1435 | /* If current_prefix_length < pos+bits, we are already doing | 1436 | /* If current_prefix_length < pos+bits, we are already doing |
1436 | * actual prefix matching, which means everything from | 1437 | * actual prefix matching, which means everything from |
1437 | * pos+(bits-chopped_off) onward must be zero along some | 1438 | * pos+(bits-chopped_off) onward must be zero along some |
1438 | * branch of this subtree - otherwise there is *no* valid | 1439 | * branch of this subtree - otherwise there is *no* valid |
1439 | * prefix present. Here we can only check the skipped | 1440 | * prefix present. Here we can only check the skipped |
1440 | * bits. Remember, since we have already indexed into the | 1441 | * bits. Remember, since we have already indexed into the |
1441 | * parent's child array, we know that the bits we chopped of | 1442 | * parent's child array, we know that the bits we chopped of |
1442 | * *are* zero. | 1443 | * *are* zero. |
1443 | */ | 1444 | */ |
1444 | 1445 | ||
1445 | /* NOTA BENE: CHECKING ONLY SKIPPED BITS FOR THE NEW NODE HERE */ | 1446 | /* NOTA BENE: CHECKING ONLY SKIPPED BITS FOR THE NEW NODE HERE */ |
1446 | 1447 | ||
1447 | if (current_prefix_length < pos+bits) { | 1448 | if (current_prefix_length < pos+bits) { |
1448 | if (tkey_extract_bits(cn->key, current_prefix_length, | 1449 | if (tkey_extract_bits(cn->key, current_prefix_length, |
1449 | cn->pos - current_prefix_length) != 0 || | 1450 | cn->pos - current_prefix_length) != 0 || |
@@ -1452,13 +1453,13 @@ fn_trie_lookup(struct fib_table *tb, const struct flowi *flp, struct fib_result | |||
1452 | } | 1453 | } |
1453 | 1454 | ||
1454 | /* | 1455 | /* |
1455 | * If chopped_off=0, the index is fully validated and we | 1456 | * If chopped_off=0, the index is fully validated and we |
1456 | * only need to look at the skipped bits for this, the new, | 1457 | * only need to look at the skipped bits for this, the new, |
1457 | * tnode. What we actually want to do is to find out if | 1458 | * tnode. What we actually want to do is to find out if |
1458 | * these skipped bits match our key perfectly, or if we will | 1459 | * these skipped bits match our key perfectly, or if we will |
1459 | * have to count on finding a matching prefix further down, | 1460 | * have to count on finding a matching prefix further down, |
1460 | * because if we do, we would like to have some way of | 1461 | * because if we do, we would like to have some way of |
1461 | * verifying the existence of such a prefix at this point. | 1462 | * verifying the existence of such a prefix at this point. |
1462 | */ | 1463 | */ |
1463 | 1464 | ||
1464 | /* The only thing we can do at this point is to verify that | 1465 | /* The only thing we can do at this point is to verify that |
@@ -1470,22 +1471,22 @@ fn_trie_lookup(struct fib_table *tb, const struct flowi *flp, struct fib_result | |||
1470 | * new tnode's key. | 1471 | * new tnode's key. |
1471 | */ | 1472 | */ |
1472 | 1473 | ||
1473 | /* Note: We aren't very concerned about the piece of the key | 1474 | /* Note: We aren't very concerned about the piece of the key |
1474 | * that precede pn->pos+pn->bits, since these have already been | 1475 | * that precede pn->pos+pn->bits, since these have already been |
1475 | * checked. The bits after cn->pos aren't checked since these are | 1476 | * checked. The bits after cn->pos aren't checked since these are |
1476 | * by definition "unknown" at this point. Thus, what we want to | 1477 | * by definition "unknown" at this point. Thus, what we want to |
1477 | * see is if we are about to enter the "prefix matching" state, | 1478 | * see is if we are about to enter the "prefix matching" state, |
1478 | * and in that case verify that the skipped bits that will prevail | 1479 | * and in that case verify that the skipped bits that will prevail |
1479 | * throughout this subtree are zero, as they have to be if we are | 1480 | * throughout this subtree are zero, as they have to be if we are |
1480 | * to find a matching prefix. | 1481 | * to find a matching prefix. |
1481 | */ | 1482 | */ |
1482 | 1483 | ||
1483 | node_prefix = MASK_PFX(cn->key, cn->pos); | 1484 | node_prefix = MASK_PFX(cn->key, cn->pos); |
1484 | key_prefix = MASK_PFX(key, cn->pos); | 1485 | key_prefix = MASK_PFX(key, cn->pos); |
1485 | pref_mismatch = key_prefix^node_prefix; | 1486 | pref_mismatch = key_prefix^node_prefix; |
1486 | mp = 0; | 1487 | mp = 0; |
1487 | 1488 | ||
1488 | /* In short: If skipped bits in this node do not match the search | 1489 | /* In short: If skipped bits in this node do not match the search |
1489 | * key, enter the "prefix matching" state.directly. | 1490 | * key, enter the "prefix matching" state.directly. |
1490 | */ | 1491 | */ |
1491 | if (pref_mismatch) { | 1492 | if (pref_mismatch) { |
@@ -1494,7 +1495,7 @@ fn_trie_lookup(struct fib_table *tb, const struct flowi *flp, struct fib_result | |||
1494 | pref_mismatch = pref_mismatch <<1; | 1495 | pref_mismatch = pref_mismatch <<1; |
1495 | } | 1496 | } |
1496 | key_prefix = tkey_extract_bits(cn->key, mp, cn->pos-mp); | 1497 | key_prefix = tkey_extract_bits(cn->key, mp, cn->pos-mp); |
1497 | 1498 | ||
1498 | if (key_prefix != 0) | 1499 | if (key_prefix != 0) |
1499 | goto backtrace; | 1500 | goto backtrace; |
1500 | 1501 | ||
@@ -1505,9 +1506,9 @@ fn_trie_lookup(struct fib_table *tb, const struct flowi *flp, struct fib_result | |||
1505 | pn = (struct tnode *)n; /* Descend */ | 1506 | pn = (struct tnode *)n; /* Descend */ |
1506 | chopped_off = 0; | 1507 | chopped_off = 0; |
1507 | continue; | 1508 | continue; |
1508 | } | 1509 | } |
1509 | if (IS_LEAF(n)) { | 1510 | if (IS_LEAF(n)) { |
1510 | if( check_leaf(t, (struct leaf *)n, key, &plen, flp, res, &ret)) | 1511 | if ((ret = check_leaf(t, (struct leaf *)n, key, &plen, flp, res)) <= 0) |
1511 | goto found; | 1512 | goto found; |
1512 | } | 1513 | } |
1513 | backtrace: | 1514 | backtrace: |
@@ -1521,18 +1522,18 @@ backtrace: | |||
1521 | /* Decrease current_... with bits chopped off */ | 1522 | /* Decrease current_... with bits chopped off */ |
1522 | if (current_prefix_length > pn->pos + pn->bits - chopped_off) | 1523 | if (current_prefix_length > pn->pos + pn->bits - chopped_off) |
1523 | current_prefix_length = pn->pos + pn->bits - chopped_off; | 1524 | current_prefix_length = pn->pos + pn->bits - chopped_off; |
1524 | 1525 | ||
1525 | /* | 1526 | /* |
1526 | * Either we do the actual chop off according or if we have | 1527 | * Either we do the actual chop off according or if we have |
1527 | * chopped off all bits in this tnode walk up to our parent. | 1528 | * chopped off all bits in this tnode walk up to our parent. |
1528 | */ | 1529 | */ |
1529 | 1530 | ||
1530 | if(chopped_off <= pn->bits) | 1531 | if (chopped_off <= pn->bits) |
1531 | cindex &= ~(1 << (chopped_off-1)); | 1532 | cindex &= ~(1 << (chopped_off-1)); |
1532 | else { | 1533 | else { |
1533 | if( NODE_PARENT(pn) == NULL) | 1534 | if (NODE_PARENT(pn) == NULL) |
1534 | goto failed; | 1535 | goto failed; |
1535 | 1536 | ||
1536 | /* Get Child's index */ | 1537 | /* Get Child's index */ |
1537 | cindex = tkey_extract_bits(pn->key, NODE_PARENT(pn)->pos, NODE_PARENT(pn)->bits); | 1538 | cindex = tkey_extract_bits(pn->key, NODE_PARENT(pn)->pos, NODE_PARENT(pn)->bits); |
1538 | pn = NODE_PARENT(pn); | 1539 | pn = NODE_PARENT(pn); |
@@ -1542,10 +1543,10 @@ backtrace: | |||
1542 | t->stats.backtrack++; | 1543 | t->stats.backtrack++; |
1543 | #endif | 1544 | #endif |
1544 | goto backtrace; | 1545 | goto backtrace; |
1545 | } | 1546 | } |
1546 | } | 1547 | } |
1547 | failed: | 1548 | failed: |
1548 | ret = 1; | 1549 | ret = 1; |
1549 | found: | 1550 | found: |
1550 | read_unlock(&fib_lock); | 1551 | read_unlock(&fib_lock); |
1551 | return ret; | 1552 | return ret; |
@@ -1558,11 +1559,11 @@ static int trie_leaf_remove(struct trie *t, t_key key) | |||
1558 | struct node *n = t->trie; | 1559 | struct node *n = t->trie; |
1559 | struct leaf *l; | 1560 | struct leaf *l; |
1560 | 1561 | ||
1561 | if(trie_debug) | 1562 | if (trie_debug) |
1562 | printk("entering trie_leaf_remove(%p)\n", n); | 1563 | printk("entering trie_leaf_remove(%p)\n", n); |
1563 | 1564 | ||
1564 | /* Note that in the case skipped bits, those bits are *not* checked! | 1565 | /* Note that in the case skipped bits, those bits are *not* checked! |
1565 | * When we finish this, we will have NULL or a T_LEAF, and the | 1566 | * When we finish this, we will have NULL or a T_LEAF, and the |
1566 | * T_LEAF may or may not match our key. | 1567 | * T_LEAF may or may not match our key. |
1567 | */ | 1568 | */ |
1568 | 1569 | ||
@@ -1571,19 +1572,19 @@ static int trie_leaf_remove(struct trie *t, t_key key) | |||
1571 | check_tnode(tn); | 1572 | check_tnode(tn); |
1572 | n = tnode_get_child(tn ,tkey_extract_bits(key, tn->pos, tn->bits)); | 1573 | n = tnode_get_child(tn ,tkey_extract_bits(key, tn->pos, tn->bits)); |
1573 | 1574 | ||
1574 | if(n && NODE_PARENT(n) != tn) { | 1575 | if (n && NODE_PARENT(n) != tn) { |
1575 | printk("BUG tn=%p, n->parent=%p\n", tn, NODE_PARENT(n)); | 1576 | printk("BUG tn=%p, n->parent=%p\n", tn, NODE_PARENT(n)); |
1576 | BUG(); | 1577 | BUG(); |
1577 | } | 1578 | } |
1578 | } | 1579 | } |
1579 | l = (struct leaf *) n; | 1580 | l = (struct leaf *) n; |
1580 | 1581 | ||
1581 | if(!n || !tkey_equals(l->key, key)) | 1582 | if (!n || !tkey_equals(l->key, key)) |
1582 | return 0; | 1583 | return 0; |
1583 | 1584 | ||
1584 | /* | 1585 | /* |
1585 | * Key found. | 1586 | * Key found. |
1586 | * Remove the leaf and rebalance the tree | 1587 | * Remove the leaf and rebalance the tree |
1587 | */ | 1588 | */ |
1588 | 1589 | ||
1589 | t->revision++; | 1590 | t->revision++; |
@@ -1592,7 +1593,7 @@ static int trie_leaf_remove(struct trie *t, t_key key) | |||
1592 | tp = NODE_PARENT(n); | 1593 | tp = NODE_PARENT(n); |
1593 | tnode_free((struct tnode *) n); | 1594 | tnode_free((struct tnode *) n); |
1594 | 1595 | ||
1595 | if(tp) { | 1596 | if (tp) { |
1596 | cindex = tkey_extract_bits(key, tp->pos, tp->bits); | 1597 | cindex = tkey_extract_bits(key, tp->pos, tp->bits); |
1597 | put_child(t, (struct tnode *)tp, cindex, NULL); | 1598 | put_child(t, (struct tnode *)tp, cindex, NULL); |
1598 | t->trie = trie_rebalance(t, tp); | 1599 | t->trie = trie_rebalance(t, tp); |
@@ -1615,23 +1616,23 @@ fn_trie_delete(struct fib_table *tb, struct rtmsg *r, struct kern_rta *rta, | |||
1615 | struct list_head *fa_head; | 1616 | struct list_head *fa_head; |
1616 | struct leaf *l; | 1617 | struct leaf *l; |
1617 | 1618 | ||
1618 | if (plen > 32) | 1619 | if (plen > 32) |
1619 | return -EINVAL; | 1620 | return -EINVAL; |
1620 | 1621 | ||
1621 | key = 0; | 1622 | key = 0; |
1622 | if (rta->rta_dst) | 1623 | if (rta->rta_dst) |
1623 | memcpy(&key, rta->rta_dst, 4); | 1624 | memcpy(&key, rta->rta_dst, 4); |
1624 | 1625 | ||
1625 | key = ntohl(key); | 1626 | key = ntohl(key); |
1626 | mask = ntohl( inet_make_mask(plen) ); | 1627 | mask = ntohl( inet_make_mask(plen) ); |
1627 | 1628 | ||
1628 | if(key & ~mask) | 1629 | if (key & ~mask) |
1629 | return -EINVAL; | 1630 | return -EINVAL; |
1630 | 1631 | ||
1631 | key = key & mask; | 1632 | key = key & mask; |
1632 | l = fib_find_node(t, key); | 1633 | l = fib_find_node(t, key); |
1633 | 1634 | ||
1634 | if(!l) | 1635 | if (!l) |
1635 | return -ESRCH; | 1636 | return -ESRCH; |
1636 | 1637 | ||
1637 | fa_head = get_fa_head(l, plen); | 1638 | fa_head = get_fa_head(l, plen); |
@@ -1677,16 +1678,16 @@ fn_trie_delete(struct fib_table *tb, struct rtmsg *r, struct kern_rta *rta, | |||
1677 | 1678 | ||
1678 | list_del(&fa->fa_list); | 1679 | list_del(&fa->fa_list); |
1679 | 1680 | ||
1680 | if(list_empty(fa_head)) { | 1681 | if (list_empty(fa_head)) { |
1681 | hlist_del(&li->hlist); | 1682 | hlist_del(&li->hlist); |
1682 | kill_li = 1; | 1683 | kill_li = 1; |
1683 | } | 1684 | } |
1684 | write_unlock_bh(&fib_lock); | 1685 | write_unlock_bh(&fib_lock); |
1685 | 1686 | ||
1686 | if(kill_li) | 1687 | if (kill_li) |
1687 | free_leaf_info(li); | 1688 | free_leaf_info(li); |
1688 | 1689 | ||
1689 | if(hlist_empty(&l->list)) | 1690 | if (hlist_empty(&l->list)) |
1690 | trie_leaf_remove(t, key); | 1691 | trie_leaf_remove(t, key); |
1691 | 1692 | ||
1692 | if (fa->fa_state & FA_S_ACCESSED) | 1693 | if (fa->fa_state & FA_S_ACCESSED) |
@@ -1705,12 +1706,12 @@ static int trie_flush_list(struct trie *t, struct list_head *head) | |||
1705 | 1706 | ||
1706 | list_for_each_entry_safe(fa, fa_node, head, fa_list) { | 1707 | list_for_each_entry_safe(fa, fa_node, head, fa_list) { |
1707 | struct fib_info *fi = fa->fa_info; | 1708 | struct fib_info *fi = fa->fa_info; |
1708 | 1709 | ||
1709 | if (fi && (fi->fib_flags&RTNH_F_DEAD)) { | 1710 | if (fi && (fi->fib_flags&RTNH_F_DEAD)) { |
1710 | 1711 | ||
1711 | write_lock_bh(&fib_lock); | 1712 | write_lock_bh(&fib_lock); |
1712 | list_del(&fa->fa_list); | 1713 | list_del(&fa->fa_list); |
1713 | write_unlock_bh(&fib_lock); | 1714 | write_unlock_bh(&fib_lock); |
1714 | 1715 | ||
1715 | fn_free_alias(fa); | 1716 | fn_free_alias(fa); |
1716 | found++; | 1717 | found++; |
@@ -1727,14 +1728,14 @@ static int trie_flush_leaf(struct trie *t, struct leaf *l) | |||
1727 | struct leaf_info *li = NULL; | 1728 | struct leaf_info *li = NULL; |
1728 | 1729 | ||
1729 | hlist_for_each_entry_safe(li, node, tmp, lih, hlist) { | 1730 | hlist_for_each_entry_safe(li, node, tmp, lih, hlist) { |
1730 | 1731 | ||
1731 | found += trie_flush_list(t, &li->falh); | 1732 | found += trie_flush_list(t, &li->falh); |
1732 | 1733 | ||
1733 | if (list_empty(&li->falh)) { | 1734 | if (list_empty(&li->falh)) { |
1734 | 1735 | ||
1735 | write_lock_bh(&fib_lock); | 1736 | write_lock_bh(&fib_lock); |
1736 | hlist_del(&li->hlist); | 1737 | hlist_del(&li->hlist); |
1737 | write_unlock_bh(&fib_lock); | 1738 | write_unlock_bh(&fib_lock); |
1738 | 1739 | ||
1739 | free_leaf_info(li); | 1740 | free_leaf_info(li); |
1740 | } | 1741 | } |
@@ -1748,8 +1749,8 @@ static struct leaf *nextleaf(struct trie *t, struct leaf *thisleaf) | |||
1748 | struct tnode *p; | 1749 | struct tnode *p; |
1749 | int idx; | 1750 | int idx; |
1750 | 1751 | ||
1751 | if(c == NULL) { | 1752 | if (c == NULL) { |
1752 | if(t->trie == NULL) | 1753 | if (t->trie == NULL) |
1753 | return NULL; | 1754 | return NULL; |
1754 | 1755 | ||
1755 | if (IS_LEAF(t->trie)) /* trie w. just a leaf */ | 1756 | if (IS_LEAF(t->trie)) /* trie w. just a leaf */ |
@@ -1757,33 +1758,34 @@ static struct leaf *nextleaf(struct trie *t, struct leaf *thisleaf) | |||
1757 | 1758 | ||
1758 | p = (struct tnode*) t->trie; /* Start */ | 1759 | p = (struct tnode*) t->trie; /* Start */ |
1759 | } | 1760 | } |
1760 | else | 1761 | else |
1761 | p = (struct tnode *) NODE_PARENT(c); | 1762 | p = (struct tnode *) NODE_PARENT(c); |
1763 | |||
1762 | while (p) { | 1764 | while (p) { |
1763 | int pos, last; | 1765 | int pos, last; |
1764 | 1766 | ||
1765 | /* Find the next child of the parent */ | 1767 | /* Find the next child of the parent */ |
1766 | if(c) | 1768 | if (c) |
1767 | pos = 1 + tkey_extract_bits(c->key, p->pos, p->bits); | 1769 | pos = 1 + tkey_extract_bits(c->key, p->pos, p->bits); |
1768 | else | 1770 | else |
1769 | pos = 0; | 1771 | pos = 0; |
1770 | 1772 | ||
1771 | last = 1 << p->bits; | 1773 | last = 1 << p->bits; |
1772 | for(idx = pos; idx < last ; idx++) { | 1774 | for(idx = pos; idx < last ; idx++) { |
1773 | if( p->child[idx]) { | 1775 | if (p->child[idx]) { |
1774 | 1776 | ||
1775 | /* Decend if tnode */ | 1777 | /* Decend if tnode */ |
1776 | 1778 | ||
1777 | while (IS_TNODE(p->child[idx])) { | 1779 | while (IS_TNODE(p->child[idx])) { |
1778 | p = (struct tnode*) p->child[idx]; | 1780 | p = (struct tnode*) p->child[idx]; |
1779 | idx = 0; | 1781 | idx = 0; |
1780 | 1782 | ||
1781 | /* Rightmost non-NULL branch */ | 1783 | /* Rightmost non-NULL branch */ |
1782 | if( p && IS_TNODE(p) ) | 1784 | if (p && IS_TNODE(p)) |
1783 | while ( p->child[idx] == NULL && idx < (1 << p->bits) ) idx++; | 1785 | while (p->child[idx] == NULL && idx < (1 << p->bits)) idx++; |
1784 | 1786 | ||
1785 | /* Done with this tnode? */ | 1787 | /* Done with this tnode? */ |
1786 | if( idx >= (1 << p->bits) || p->child[idx] == NULL ) | 1788 | if (idx >= (1 << p->bits) || p->child[idx] == NULL ) |
1787 | goto up; | 1789 | goto up; |
1788 | } | 1790 | } |
1789 | return (struct leaf*) p->child[idx]; | 1791 | return (struct leaf*) p->child[idx]; |
@@ -1816,7 +1818,7 @@ static int fn_trie_flush(struct fib_table *tb) | |||
1816 | if (ll && hlist_empty(&ll->list)) | 1818 | if (ll && hlist_empty(&ll->list)) |
1817 | trie_leaf_remove(t, ll->key); | 1819 | trie_leaf_remove(t, ll->key); |
1818 | 1820 | ||
1819 | if(trie_debug) | 1821 | if (trie_debug) |
1820 | printk("trie_flush found=%d\n", found); | 1822 | printk("trie_flush found=%d\n", found); |
1821 | return found; | 1823 | return found; |
1822 | } | 1824 | } |
@@ -1839,32 +1841,32 @@ fn_trie_select_default(struct fib_table *tb, const struct flowi *flp, struct fib | |||
1839 | order = -1; | 1841 | order = -1; |
1840 | 1842 | ||
1841 | read_lock(&fib_lock); | 1843 | read_lock(&fib_lock); |
1842 | 1844 | ||
1843 | l = fib_find_node(t, 0); | 1845 | l = fib_find_node(t, 0); |
1844 | if(!l) | 1846 | if (!l) |
1845 | goto out; | 1847 | goto out; |
1846 | 1848 | ||
1847 | fa_head = get_fa_head(l, 0); | 1849 | fa_head = get_fa_head(l, 0); |
1848 | if(!fa_head) | 1850 | if (!fa_head) |
1849 | goto out; | 1851 | goto out; |
1850 | 1852 | ||
1851 | if (list_empty(fa_head)) | 1853 | if (list_empty(fa_head)) |
1852 | goto out; | 1854 | goto out; |
1853 | 1855 | ||
1854 | list_for_each_entry(fa, fa_head, fa_list) { | 1856 | list_for_each_entry(fa, fa_head, fa_list) { |
1855 | struct fib_info *next_fi = fa->fa_info; | 1857 | struct fib_info *next_fi = fa->fa_info; |
1856 | 1858 | ||
1857 | if (fa->fa_scope != res->scope || | 1859 | if (fa->fa_scope != res->scope || |
1858 | fa->fa_type != RTN_UNICAST) | 1860 | fa->fa_type != RTN_UNICAST) |
1859 | continue; | 1861 | continue; |
1860 | 1862 | ||
1861 | if (next_fi->fib_priority > res->fi->fib_priority) | 1863 | if (next_fi->fib_priority > res->fi->fib_priority) |
1862 | break; | 1864 | break; |
1863 | if (!next_fi->fib_nh[0].nh_gw || | 1865 | if (!next_fi->fib_nh[0].nh_gw || |
1864 | next_fi->fib_nh[0].nh_scope != RT_SCOPE_LINK) | 1866 | next_fi->fib_nh[0].nh_scope != RT_SCOPE_LINK) |
1865 | continue; | 1867 | continue; |
1866 | fa->fa_state |= FA_S_ACCESSED; | 1868 | fa->fa_state |= FA_S_ACCESSED; |
1867 | 1869 | ||
1868 | if (fi == NULL) { | 1870 | if (fi == NULL) { |
1869 | if (next_fi != res->fi) | 1871 | if (next_fi != res->fi) |
1870 | break; | 1872 | break; |
@@ -1902,10 +1904,10 @@ fn_trie_select_default(struct fib_table *tb, const struct flowi *flp, struct fib | |||
1902 | } | 1904 | } |
1903 | trie_last_dflt = last_idx; | 1905 | trie_last_dflt = last_idx; |
1904 | out:; | 1906 | out:; |
1905 | read_unlock(&fib_lock); | 1907 | read_unlock(&fib_lock); |
1906 | } | 1908 | } |
1907 | 1909 | ||
1908 | static int fn_trie_dump_fa(t_key key, int plen, struct list_head *fah, struct fib_table *tb, | 1910 | static int fn_trie_dump_fa(t_key key, int plen, struct list_head *fah, struct fib_table *tb, |
1909 | struct sk_buff *skb, struct netlink_callback *cb) | 1911 | struct sk_buff *skb, struct netlink_callback *cb) |
1910 | { | 1912 | { |
1911 | int i, s_i; | 1913 | int i, s_i; |
@@ -1951,7 +1953,7 @@ static int fn_trie_dump_fa(t_key key, int plen, struct list_head *fah, struct fi | |||
1951 | return skb->len; | 1953 | return skb->len; |
1952 | } | 1954 | } |
1953 | 1955 | ||
1954 | static int fn_trie_dump_plen(struct trie *t, int plen, struct fib_table *tb, struct sk_buff *skb, | 1956 | static int fn_trie_dump_plen(struct trie *t, int plen, struct fib_table *tb, struct sk_buff *skb, |
1955 | struct netlink_callback *cb) | 1957 | struct netlink_callback *cb) |
1956 | { | 1958 | { |
1957 | int h, s_h; | 1959 | int h, s_h; |
@@ -1968,11 +1970,11 @@ static int fn_trie_dump_plen(struct trie *t, int plen, struct fib_table *tb, str | |||
1968 | sizeof(cb->args) - 3*sizeof(cb->args[0])); | 1970 | sizeof(cb->args) - 3*sizeof(cb->args[0])); |
1969 | 1971 | ||
1970 | fa_head = get_fa_head(l, plen); | 1972 | fa_head = get_fa_head(l, plen); |
1971 | 1973 | ||
1972 | if(!fa_head) | 1974 | if (!fa_head) |
1973 | continue; | 1975 | continue; |
1974 | 1976 | ||
1975 | if(list_empty(fa_head)) | 1977 | if (list_empty(fa_head)) |
1976 | continue; | 1978 | continue; |
1977 | 1979 | ||
1978 | if (fn_trie_dump_fa(l->key, plen, fa_head, tb, skb, cb)<0) { | 1980 | if (fn_trie_dump_fa(l->key, plen, fa_head, tb, skb, cb)<0) { |
@@ -2048,10 +2050,10 @@ struct fib_table * __init fib_hash_init(int id) | |||
2048 | 2050 | ||
2049 | trie_init(t); | 2051 | trie_init(t); |
2050 | 2052 | ||
2051 | if (id == RT_TABLE_LOCAL) | 2053 | if (id == RT_TABLE_LOCAL) |
2052 | trie_local=t; | 2054 | trie_local = t; |
2053 | else if (id == RT_TABLE_MAIN) | 2055 | else if (id == RT_TABLE_MAIN) |
2054 | trie_main=t; | 2056 | trie_main = t; |
2055 | 2057 | ||
2056 | if (id == RT_TABLE_LOCAL) | 2058 | if (id == RT_TABLE_LOCAL) |
2057 | printk("IPv4 FIB: Using LC-trie version %s\n", VERSION); | 2059 | printk("IPv4 FIB: Using LC-trie version %s\n", VERSION); |
@@ -2072,7 +2074,7 @@ static void printbin_seq(struct seq_file *seq, unsigned int v, int bits) | |||
2072 | seq_printf(seq, "%s", (v & (1<<bits))?"1":"0"); | 2074 | seq_printf(seq, "%s", (v & (1<<bits))?"1":"0"); |
2073 | } | 2075 | } |
2074 | 2076 | ||
2075 | static void printnode_seq(struct seq_file *seq, int indent, struct node *n, | 2077 | static void printnode_seq(struct seq_file *seq, int indent, struct node *n, |
2076 | int pend, int cindex, int bits) | 2078 | int pend, int cindex, int bits) |
2077 | { | 2079 | { |
2078 | putspace_seq(seq, indent); | 2080 | putspace_seq(seq, indent); |
@@ -2090,12 +2092,12 @@ static void printnode_seq(struct seq_file *seq, int indent, struct node *n, | |||
2090 | seq_printf(seq, "%s:%p ", IS_LEAF(n)?"Leaf":"Internal node", n); | 2092 | seq_printf(seq, "%s:%p ", IS_LEAF(n)?"Leaf":"Internal node", n); |
2091 | 2093 | ||
2092 | if (IS_LEAF(n)) | 2094 | if (IS_LEAF(n)) |
2093 | seq_printf(seq, "key=%d.%d.%d.%d\n", | 2095 | seq_printf(seq, "key=%d.%d.%d.%d\n", |
2094 | n->key >> 24, (n->key >> 16) % 256, (n->key >> 8) % 256, n->key % 256); | 2096 | n->key >> 24, (n->key >> 16) % 256, (n->key >> 8) % 256, n->key % 256); |
2095 | else { | 2097 | else { |
2096 | int plen=((struct tnode *)n)->pos; | 2098 | int plen = ((struct tnode *)n)->pos; |
2097 | t_key prf=MASK_PFX(n->key, plen); | 2099 | t_key prf=MASK_PFX(n->key, plen); |
2098 | seq_printf(seq, "key=%d.%d.%d.%d/%d\n", | 2100 | seq_printf(seq, "key=%d.%d.%d.%d/%d\n", |
2099 | prf >> 24, (prf >> 16) % 256, (prf >> 8) % 256, prf % 256, plen); | 2101 | prf >> 24, (prf >> 16) % 256, (prf >> 8) % 256, prf % 256, plen); |
2100 | } | 2102 | } |
2101 | if (IS_LEAF(n)) { | 2103 | if (IS_LEAF(n)) { |
@@ -2103,14 +2105,14 @@ static void printnode_seq(struct seq_file *seq, int indent, struct node *n, | |||
2103 | struct fib_alias *fa; | 2105 | struct fib_alias *fa; |
2104 | int i; | 2106 | int i; |
2105 | for (i=32; i>=0; i--) | 2107 | for (i=32; i>=0; i--) |
2106 | if(find_leaf_info(&l->list, i)) { | 2108 | if (find_leaf_info(&l->list, i)) { |
2107 | 2109 | ||
2108 | struct list_head *fa_head = get_fa_head(l, i); | 2110 | struct list_head *fa_head = get_fa_head(l, i); |
2109 | 2111 | ||
2110 | if(!fa_head) | 2112 | if (!fa_head) |
2111 | continue; | 2113 | continue; |
2112 | 2114 | ||
2113 | if(list_empty(fa_head)) | 2115 | if (list_empty(fa_head)) |
2114 | continue; | 2116 | continue; |
2115 | 2117 | ||
2116 | putspace_seq(seq, indent+2); | 2118 | putspace_seq(seq, indent+2); |
@@ -2136,7 +2138,7 @@ static void printnode_seq(struct seq_file *seq, int indent, struct node *n, | |||
2136 | } | 2138 | } |
2137 | } | 2139 | } |
2138 | else if (IS_TNODE(n)) { | 2140 | else if (IS_TNODE(n)) { |
2139 | struct tnode *tn=(struct tnode *)n; | 2141 | struct tnode *tn = (struct tnode *)n; |
2140 | putspace_seq(seq, indent); seq_printf(seq, "| "); | 2142 | putspace_seq(seq, indent); seq_printf(seq, "| "); |
2141 | seq_printf(seq, "{key prefix=%08x/", tn->key&TKEY_GET_MASK(0, tn->pos)); | 2143 | seq_printf(seq, "{key prefix=%08x/", tn->key&TKEY_GET_MASK(0, tn->pos)); |
2142 | printbin_seq(seq, tkey_extract_bits(tn->key, 0, tn->pos), tn->pos); | 2144 | printbin_seq(seq, tkey_extract_bits(tn->key, 0, tn->pos), tn->pos); |
@@ -2152,7 +2154,7 @@ static void printnode_seq(struct seq_file *seq, int indent, struct node *n, | |||
2152 | 2154 | ||
2153 | static void trie_dump_seq(struct seq_file *seq, struct trie *t) | 2155 | static void trie_dump_seq(struct seq_file *seq, struct trie *t) |
2154 | { | 2156 | { |
2155 | struct node *n=t->trie; | 2157 | struct node *n = t->trie; |
2156 | int cindex=0; | 2158 | int cindex=0; |
2157 | int indent=1; | 2159 | int indent=1; |
2158 | int pend=0; | 2160 | int pend=0; |
@@ -2164,7 +2166,7 @@ static void trie_dump_seq(struct seq_file *seq, struct trie *t) | |||
2164 | if (n) { | 2166 | if (n) { |
2165 | printnode_seq(seq, indent, n, pend, cindex, 0); | 2167 | printnode_seq(seq, indent, n, pend, cindex, 0); |
2166 | if (IS_TNODE(n)) { | 2168 | if (IS_TNODE(n)) { |
2167 | struct tnode *tn=(struct tnode *)n; | 2169 | struct tnode *tn = (struct tnode *)n; |
2168 | pend = tn->pos+tn->bits; | 2170 | pend = tn->pos+tn->bits; |
2169 | putspace_seq(seq, indent); seq_printf(seq, "\\--\n"); | 2171 | putspace_seq(seq, indent); seq_printf(seq, "\\--\n"); |
2170 | indent += 3; | 2172 | indent += 3; |
@@ -2172,42 +2174,42 @@ static void trie_dump_seq(struct seq_file *seq, struct trie *t) | |||
2172 | 2174 | ||
2173 | while (tn && cindex < (1 << tn->bits)) { | 2175 | while (tn && cindex < (1 << tn->bits)) { |
2174 | if (tn->child[cindex]) { | 2176 | if (tn->child[cindex]) { |
2175 | 2177 | ||
2176 | /* Got a child */ | 2178 | /* Got a child */ |
2177 | 2179 | ||
2178 | printnode_seq(seq, indent, tn->child[cindex], pend, cindex, tn->bits); | 2180 | printnode_seq(seq, indent, tn->child[cindex], pend, cindex, tn->bits); |
2179 | if (IS_LEAF(tn->child[cindex])) { | 2181 | if (IS_LEAF(tn->child[cindex])) { |
2180 | cindex++; | 2182 | cindex++; |
2181 | 2183 | ||
2182 | } | 2184 | } |
2183 | else { | 2185 | else { |
2184 | /* | 2186 | /* |
2185 | * New tnode. Decend one level | 2187 | * New tnode. Decend one level |
2186 | */ | 2188 | */ |
2187 | 2189 | ||
2188 | depth++; | 2190 | depth++; |
2189 | n=tn->child[cindex]; | 2191 | n = tn->child[cindex]; |
2190 | tn=(struct tnode *)n; | 2192 | tn = (struct tnode *)n; |
2191 | pend=tn->pos+tn->bits; | 2193 | pend = tn->pos+tn->bits; |
2192 | putspace_seq(seq, indent); seq_printf(seq, "\\--\n"); | 2194 | putspace_seq(seq, indent); seq_printf(seq, "\\--\n"); |
2193 | indent+=3; | 2195 | indent+=3; |
2194 | cindex=0; | 2196 | cindex=0; |
2195 | } | 2197 | } |
2196 | } | 2198 | } |
2197 | else | 2199 | else |
2198 | cindex++; | 2200 | cindex++; |
2199 | 2201 | ||
2200 | /* | 2202 | /* |
2201 | * Test if we are done | 2203 | * Test if we are done |
2202 | */ | 2204 | */ |
2203 | 2205 | ||
2204 | while (cindex >= (1 << tn->bits)) { | 2206 | while (cindex >= (1 << tn->bits)) { |
2205 | 2207 | ||
2206 | /* | 2208 | /* |
2207 | * Move upwards and test for root | 2209 | * Move upwards and test for root |
2208 | * pop off all traversed nodes | 2210 | * pop off all traversed nodes |
2209 | */ | 2211 | */ |
2210 | 2212 | ||
2211 | if (NODE_PARENT(tn) == NULL) { | 2213 | if (NODE_PARENT(tn) == NULL) { |
2212 | tn = NULL; | 2214 | tn = NULL; |
2213 | n = NULL; | 2215 | n = NULL; |
@@ -2217,8 +2219,8 @@ static void trie_dump_seq(struct seq_file *seq, struct trie *t) | |||
2217 | cindex = tkey_extract_bits(tn->key, NODE_PARENT(tn)->pos, NODE_PARENT(tn)->bits); | 2219 | cindex = tkey_extract_bits(tn->key, NODE_PARENT(tn)->pos, NODE_PARENT(tn)->bits); |
2218 | tn = NODE_PARENT(tn); | 2220 | tn = NODE_PARENT(tn); |
2219 | cindex++; | 2221 | cindex++; |
2220 | n=(struct node *)tn; | 2222 | n = (struct node *)tn; |
2221 | pend=tn->pos+tn->bits; | 2223 | pend = tn->pos+tn->bits; |
2222 | indent-=3; | 2224 | indent-=3; |
2223 | depth--; | 2225 | depth--; |
2224 | } | 2226 | } |
@@ -2236,36 +2238,36 @@ static struct trie_stat *trie_stat_new(void) | |||
2236 | { | 2238 | { |
2237 | struct trie_stat *s = kmalloc(sizeof(struct trie_stat), GFP_KERNEL); | 2239 | struct trie_stat *s = kmalloc(sizeof(struct trie_stat), GFP_KERNEL); |
2238 | int i; | 2240 | int i; |
2239 | 2241 | ||
2240 | if(s) { | 2242 | if (s) { |
2241 | s->totdepth = 0; | 2243 | s->totdepth = 0; |
2242 | s->maxdepth = 0; | 2244 | s->maxdepth = 0; |
2243 | s->tnodes = 0; | 2245 | s->tnodes = 0; |
2244 | s->leaves = 0; | 2246 | s->leaves = 0; |
2245 | s->nullpointers = 0; | 2247 | s->nullpointers = 0; |
2246 | 2248 | ||
2247 | for(i=0; i< MAX_CHILDS; i++) | 2249 | for(i=0; i< MAX_CHILDS; i++) |
2248 | s->nodesizes[i] = 0; | 2250 | s->nodesizes[i] = 0; |
2249 | } | 2251 | } |
2250 | return s; | 2252 | return s; |
2251 | } | 2253 | } |
2252 | 2254 | ||
2253 | static struct trie_stat *trie_collect_stats(struct trie *t) | 2255 | static struct trie_stat *trie_collect_stats(struct trie *t) |
2254 | { | 2256 | { |
2255 | struct node *n=t->trie; | 2257 | struct node *n = t->trie; |
2256 | struct trie_stat *s = trie_stat_new(); | 2258 | struct trie_stat *s = trie_stat_new(); |
2257 | int cindex = 0; | 2259 | int cindex = 0; |
2258 | int indent = 1; | 2260 | int indent = 1; |
2259 | int pend = 0; | 2261 | int pend = 0; |
2260 | int depth = 0; | 2262 | int depth = 0; |
2261 | 2263 | ||
2262 | read_lock(&fib_lock); | 2264 | read_lock(&fib_lock); |
2263 | 2265 | ||
2264 | if (s) { | 2266 | if (s) { |
2265 | if (n) { | 2267 | if (n) { |
2266 | if (IS_TNODE(n)) { | 2268 | if (IS_TNODE(n)) { |
2267 | struct tnode *tn = (struct tnode *)n; | 2269 | struct tnode *tn = (struct tnode *)n; |
2268 | pend=tn->pos+tn->bits; | 2270 | pend = tn->pos+tn->bits; |
2269 | indent += 3; | 2271 | indent += 3; |
2270 | s->nodesizes[tn->bits]++; | 2272 | s->nodesizes[tn->bits]++; |
2271 | depth++; | 2273 | depth++; |
@@ -2273,26 +2275,26 @@ static struct trie_stat *trie_collect_stats(struct trie *t) | |||
2273 | while (tn && cindex < (1 << tn->bits)) { | 2275 | while (tn && cindex < (1 << tn->bits)) { |
2274 | if (tn->child[cindex]) { | 2276 | if (tn->child[cindex]) { |
2275 | /* Got a child */ | 2277 | /* Got a child */ |
2276 | 2278 | ||
2277 | if (IS_LEAF(tn->child[cindex])) { | 2279 | if (IS_LEAF(tn->child[cindex])) { |
2278 | cindex++; | 2280 | cindex++; |
2279 | 2281 | ||
2280 | /* stats */ | 2282 | /* stats */ |
2281 | if (depth > s->maxdepth) | 2283 | if (depth > s->maxdepth) |
2282 | s->maxdepth = depth; | 2284 | s->maxdepth = depth; |
2283 | s->totdepth += depth; | 2285 | s->totdepth += depth; |
2284 | s->leaves++; | 2286 | s->leaves++; |
2285 | } | 2287 | } |
2286 | 2288 | ||
2287 | else { | 2289 | else { |
2288 | /* | 2290 | /* |
2289 | * New tnode. Decend one level | 2291 | * New tnode. Decend one level |
2290 | */ | 2292 | */ |
2291 | 2293 | ||
2292 | s->tnodes++; | 2294 | s->tnodes++; |
2293 | s->nodesizes[tn->bits]++; | 2295 | s->nodesizes[tn->bits]++; |
2294 | depth++; | 2296 | depth++; |
2295 | 2297 | ||
2296 | n = tn->child[cindex]; | 2298 | n = tn->child[cindex]; |
2297 | tn = (struct tnode *)n; | 2299 | tn = (struct tnode *)n; |
2298 | pend = tn->pos+tn->bits; | 2300 | pend = tn->pos+tn->bits; |
@@ -2303,13 +2305,13 @@ static struct trie_stat *trie_collect_stats(struct trie *t) | |||
2303 | } | 2305 | } |
2304 | else { | 2306 | else { |
2305 | cindex++; | 2307 | cindex++; |
2306 | s->nullpointers++; | 2308 | s->nullpointers++; |
2307 | } | 2309 | } |
2308 | 2310 | ||
2309 | /* | 2311 | /* |
2310 | * Test if we are done | 2312 | * Test if we are done |
2311 | */ | 2313 | */ |
2312 | 2314 | ||
2313 | while (cindex >= (1 << tn->bits)) { | 2315 | while (cindex >= (1 << tn->bits)) { |
2314 | 2316 | ||
2315 | /* | 2317 | /* |
@@ -2317,7 +2319,7 @@ static struct trie_stat *trie_collect_stats(struct trie *t) | |||
2317 | * pop off all traversed nodes | 2319 | * pop off all traversed nodes |
2318 | */ | 2320 | */ |
2319 | 2321 | ||
2320 | 2322 | ||
2321 | if (NODE_PARENT(tn) == NULL) { | 2323 | if (NODE_PARENT(tn) == NULL) { |
2322 | tn = NULL; | 2324 | tn = NULL; |
2323 | n = NULL; | 2325 | n = NULL; |
@@ -2326,9 +2328,9 @@ static struct trie_stat *trie_collect_stats(struct trie *t) | |||
2326 | else { | 2328 | else { |
2327 | cindex = tkey_extract_bits(tn->key, NODE_PARENT(tn)->pos, NODE_PARENT(tn)->bits); | 2329 | cindex = tkey_extract_bits(tn->key, NODE_PARENT(tn)->pos, NODE_PARENT(tn)->bits); |
2328 | tn = NODE_PARENT(tn); | 2330 | tn = NODE_PARENT(tn); |
2329 | cindex++; | 2331 | cindex++; |
2330 | n = (struct node *)tn; | 2332 | n = (struct node *)tn; |
2331 | pend=tn->pos+tn->bits; | 2333 | pend = tn->pos+tn->bits; |
2332 | indent -= 3; | 2334 | indent -= 3; |
2333 | depth--; | 2335 | depth--; |
2334 | } | 2336 | } |
@@ -2339,7 +2341,7 @@ static struct trie_stat *trie_collect_stats(struct trie *t) | |||
2339 | } | 2341 | } |
2340 | } | 2342 | } |
2341 | 2343 | ||
2342 | read_unlock(&fib_lock); | 2344 | read_unlock(&fib_lock); |
2343 | return s; | 2345 | return s; |
2344 | } | 2346 | } |
2345 | 2347 | ||
@@ -2375,7 +2377,7 @@ static void fib_triestat_seq_stop(struct seq_file *seq, void *v) | |||
2375 | 2377 | ||
2376 | } | 2378 | } |
2377 | 2379 | ||
2378 | /* | 2380 | /* |
2379 | * This outputs /proc/net/fib_triestats | 2381 | * This outputs /proc/net/fib_triestats |
2380 | * | 2382 | * |
2381 | * It always works in backward compatibility mode. | 2383 | * It always works in backward compatibility mode. |
@@ -2401,7 +2403,7 @@ static void collect_and_show(struct trie *t, struct seq_file *seq) | |||
2401 | avdepth=0; | 2403 | avdepth=0; |
2402 | seq_printf(seq, "Aver depth: %d.%02d\n", avdepth / 100, avdepth % 100 ); | 2404 | seq_printf(seq, "Aver depth: %d.%02d\n", avdepth / 100, avdepth % 100 ); |
2403 | seq_printf(seq, "Max depth: %4d\n", stat->maxdepth); | 2405 | seq_printf(seq, "Max depth: %4d\n", stat->maxdepth); |
2404 | 2406 | ||
2405 | seq_printf(seq, "Leaves: %d\n", stat->leaves); | 2407 | seq_printf(seq, "Leaves: %d\n", stat->leaves); |
2406 | bytes += sizeof(struct leaf) * stat->leaves; | 2408 | bytes += sizeof(struct leaf) * stat->leaves; |
2407 | seq_printf(seq, "Internal nodes: %d\n", stat->tnodes); | 2409 | seq_printf(seq, "Internal nodes: %d\n", stat->tnodes); |
@@ -2413,7 +2415,7 @@ static void collect_and_show(struct trie *t, struct seq_file *seq) | |||
2413 | max--; | 2415 | max--; |
2414 | pointers = 0; | 2416 | pointers = 0; |
2415 | 2417 | ||
2416 | for (i = 1; i <= max; i++) | 2418 | for (i = 1; i <= max; i++) |
2417 | if (stat->nodesizes[i] != 0) { | 2419 | if (stat->nodesizes[i] != 0) { |
2418 | seq_printf(seq, " %d: %d", i, stat->nodesizes[i]); | 2420 | seq_printf(seq, " %d: %d", i, stat->nodesizes[i]); |
2419 | pointers += (1<<i) * stat->nodesizes[i]; | 2421 | pointers += (1<<i) * stat->nodesizes[i]; |
@@ -2444,30 +2446,30 @@ static void collect_and_show(struct trie *t, struct seq_file *seq) | |||
2444 | static int fib_triestat_seq_show(struct seq_file *seq, void *v) | 2446 | static int fib_triestat_seq_show(struct seq_file *seq, void *v) |
2445 | { | 2447 | { |
2446 | char bf[128]; | 2448 | char bf[128]; |
2447 | 2449 | ||
2448 | if (v == SEQ_START_TOKEN) { | 2450 | if (v == SEQ_START_TOKEN) { |
2449 | seq_printf(seq, "Basic info: size of leaf: %Zd bytes, size of tnode: %Zd bytes.\n", | 2451 | seq_printf(seq, "Basic info: size of leaf: %Zd bytes, size of tnode: %Zd bytes.\n", |
2450 | sizeof(struct leaf), sizeof(struct tnode)); | 2452 | sizeof(struct leaf), sizeof(struct tnode)); |
2451 | if (trie_local) | 2453 | if (trie_local) |
2452 | collect_and_show(trie_local, seq); | 2454 | collect_and_show(trie_local, seq); |
2453 | 2455 | ||
2454 | if (trie_main) | 2456 | if (trie_main) |
2455 | collect_and_show(trie_main, seq); | 2457 | collect_and_show(trie_main, seq); |
2456 | } | 2458 | } |
2457 | else { | 2459 | else { |
2458 | snprintf(bf, sizeof(bf), | 2460 | snprintf(bf, sizeof(bf), |
2459 | "*\t%08X\t%08X", 200, 400); | 2461 | "*\t%08X\t%08X", 200, 400); |
2460 | 2462 | ||
2461 | seq_printf(seq, "%-127s\n", bf); | 2463 | seq_printf(seq, "%-127s\n", bf); |
2462 | } | 2464 | } |
2463 | return 0; | 2465 | return 0; |
2464 | } | 2466 | } |
2465 | 2467 | ||
2466 | static struct seq_operations fib_triestat_seq_ops = { | 2468 | static struct seq_operations fib_triestat_seq_ops = { |
2467 | .start = fib_triestat_seq_start, | 2469 | .start = fib_triestat_seq_start, |
2468 | .next = fib_triestat_seq_next, | 2470 | .next = fib_triestat_seq_next, |
2469 | .stop = fib_triestat_seq_stop, | 2471 | .stop = fib_triestat_seq_stop, |
2470 | .show = fib_triestat_seq_show, | 2472 | .show = fib_triestat_seq_show, |
2471 | }; | 2473 | }; |
2472 | 2474 | ||
2473 | static int fib_triestat_seq_open(struct inode *inode, struct file *file) | 2475 | static int fib_triestat_seq_open(struct inode *inode, struct file *file) |
@@ -2479,7 +2481,7 @@ static int fib_triestat_seq_open(struct inode *inode, struct file *file) | |||
2479 | if (rc) | 2481 | if (rc) |
2480 | goto out_kfree; | 2482 | goto out_kfree; |
2481 | 2483 | ||
2482 | seq = file->private_data; | 2484 | seq = file->private_data; |
2483 | out: | 2485 | out: |
2484 | return rc; | 2486 | return rc; |
2485 | out_kfree: | 2487 | out_kfree: |
@@ -2487,11 +2489,11 @@ out_kfree: | |||
2487 | } | 2489 | } |
2488 | 2490 | ||
2489 | static struct file_operations fib_triestat_seq_fops = { | 2491 | static struct file_operations fib_triestat_seq_fops = { |
2490 | .owner = THIS_MODULE, | 2492 | .owner = THIS_MODULE, |
2491 | .open = fib_triestat_seq_open, | 2493 | .open = fib_triestat_seq_open, |
2492 | .read = seq_read, | 2494 | .read = seq_read, |
2493 | .llseek = seq_lseek, | 2495 | .llseek = seq_lseek, |
2494 | .release = seq_release_private, | 2496 | .release = seq_release_private, |
2495 | }; | 2497 | }; |
2496 | 2498 | ||
2497 | int __init fib_stat_proc_init(void) | 2499 | int __init fib_stat_proc_init(void) |
@@ -2536,7 +2538,7 @@ static void fib_trie_seq_stop(struct seq_file *seq, void *v) | |||
2536 | 2538 | ||
2537 | } | 2539 | } |
2538 | 2540 | ||
2539 | /* | 2541 | /* |
2540 | * This outputs /proc/net/fib_trie. | 2542 | * This outputs /proc/net/fib_trie. |
2541 | * | 2543 | * |
2542 | * It always works in backward compatibility mode. | 2544 | * It always works in backward compatibility mode. |
@@ -2548,10 +2550,10 @@ static int fib_trie_seq_show(struct seq_file *seq, void *v) | |||
2548 | char bf[128]; | 2550 | char bf[128]; |
2549 | 2551 | ||
2550 | if (v == SEQ_START_TOKEN) { | 2552 | if (v == SEQ_START_TOKEN) { |
2551 | if (trie_local) | 2553 | if (trie_local) |
2552 | trie_dump_seq(seq, trie_local); | 2554 | trie_dump_seq(seq, trie_local); |
2553 | 2555 | ||
2554 | if (trie_main) | 2556 | if (trie_main) |
2555 | trie_dump_seq(seq, trie_main); | 2557 | trie_dump_seq(seq, trie_main); |
2556 | } | 2558 | } |
2557 | 2559 | ||
@@ -2565,10 +2567,10 @@ static int fib_trie_seq_show(struct seq_file *seq, void *v) | |||
2565 | } | 2567 | } |
2566 | 2568 | ||
2567 | static struct seq_operations fib_trie_seq_ops = { | 2569 | static struct seq_operations fib_trie_seq_ops = { |
2568 | .start = fib_trie_seq_start, | 2570 | .start = fib_trie_seq_start, |
2569 | .next = fib_trie_seq_next, | 2571 | .next = fib_trie_seq_next, |
2570 | .stop = fib_trie_seq_stop, | 2572 | .stop = fib_trie_seq_stop, |
2571 | .show = fib_trie_seq_show, | 2573 | .show = fib_trie_seq_show, |
2572 | }; | 2574 | }; |
2573 | 2575 | ||
2574 | static int fib_trie_seq_open(struct inode *inode, struct file *file) | 2576 | static int fib_trie_seq_open(struct inode *inode, struct file *file) |
@@ -2580,7 +2582,7 @@ static int fib_trie_seq_open(struct inode *inode, struct file *file) | |||
2580 | if (rc) | 2582 | if (rc) |
2581 | goto out_kfree; | 2583 | goto out_kfree; |
2582 | 2584 | ||
2583 | seq = file->private_data; | 2585 | seq = file->private_data; |
2584 | out: | 2586 | out: |
2585 | return rc; | 2587 | return rc; |
2586 | out_kfree: | 2588 | out_kfree: |
@@ -2588,11 +2590,11 @@ out_kfree: | |||
2588 | } | 2590 | } |
2589 | 2591 | ||
2590 | static struct file_operations fib_trie_seq_fops = { | 2592 | static struct file_operations fib_trie_seq_fops = { |
2591 | .owner = THIS_MODULE, | 2593 | .owner = THIS_MODULE, |
2592 | .open = fib_trie_seq_open, | 2594 | .open = fib_trie_seq_open, |
2593 | .read = seq_read, | 2595 | .read = seq_read, |
2594 | .llseek = seq_lseek, | 2596 | .llseek = seq_lseek, |
2595 | .release = seq_release_private, | 2597 | .release= seq_release_private, |
2596 | }; | 2598 | }; |
2597 | 2599 | ||
2598 | int __init fib_proc_init(void) | 2600 | int __init fib_proc_init(void) |
diff --git a/net/ipv4/icmp.c b/net/ipv4/icmp.c index 279f57abfecb..badfc5849973 100644 --- a/net/ipv4/icmp.c +++ b/net/ipv4/icmp.c | |||
@@ -349,12 +349,12 @@ static void icmp_push_reply(struct icmp_bxm *icmp_param, | |||
349 | { | 349 | { |
350 | struct sk_buff *skb; | 350 | struct sk_buff *skb; |
351 | 351 | ||
352 | ip_append_data(icmp_socket->sk, icmp_glue_bits, icmp_param, | 352 | if (ip_append_data(icmp_socket->sk, icmp_glue_bits, icmp_param, |
353 | icmp_param->data_len+icmp_param->head_len, | 353 | icmp_param->data_len+icmp_param->head_len, |
354 | icmp_param->head_len, | 354 | icmp_param->head_len, |
355 | ipc, rt, MSG_DONTWAIT); | 355 | ipc, rt, MSG_DONTWAIT) < 0) |
356 | 356 | ip_flush_pending_frames(icmp_socket->sk); | |
357 | if ((skb = skb_peek(&icmp_socket->sk->sk_write_queue)) != NULL) { | 357 | else if ((skb = skb_peek(&icmp_socket->sk->sk_write_queue)) != NULL) { |
358 | struct icmphdr *icmph = skb->h.icmph; | 358 | struct icmphdr *icmph = skb->h.icmph; |
359 | unsigned int csum = 0; | 359 | unsigned int csum = 0; |
360 | struct sk_buff *skb1; | 360 | struct sk_buff *skb1; |
@@ -936,8 +936,7 @@ int icmp_rcv(struct sk_buff *skb) | |||
936 | case CHECKSUM_HW: | 936 | case CHECKSUM_HW: |
937 | if (!(u16)csum_fold(skb->csum)) | 937 | if (!(u16)csum_fold(skb->csum)) |
938 | break; | 938 | break; |
939 | NETDEBUG(if (net_ratelimit()) | 939 | LIMIT_NETDEBUG(printk(KERN_DEBUG "icmp v4 hw csum failure\n")); |
940 | printk(KERN_DEBUG "icmp v4 hw csum failure\n")); | ||
941 | case CHECKSUM_NONE: | 940 | case CHECKSUM_NONE: |
942 | if ((u16)csum_fold(skb_checksum(skb, 0, skb->len, 0))) | 941 | if ((u16)csum_fold(skb_checksum(skb, 0, skb->len, 0))) |
943 | goto error; | 942 | goto error; |
diff --git a/net/ipv4/inetpeer.c b/net/ipv4/inetpeer.c index 95473953c406..ab18a853d7ce 100644 --- a/net/ipv4/inetpeer.c +++ b/net/ipv4/inetpeer.c | |||
@@ -450,10 +450,13 @@ static void peer_check_expire(unsigned long dummy) | |||
450 | /* Trigger the timer after inet_peer_gc_mintime .. inet_peer_gc_maxtime | 450 | /* Trigger the timer after inet_peer_gc_mintime .. inet_peer_gc_maxtime |
451 | * interval depending on the total number of entries (more entries, | 451 | * interval depending on the total number of entries (more entries, |
452 | * less interval). */ | 452 | * less interval). */ |
453 | peer_periodic_timer.expires = jiffies | 453 | if (peer_total >= inet_peer_threshold) |
454 | + inet_peer_gc_maxtime | 454 | peer_periodic_timer.expires = jiffies + inet_peer_gc_mintime; |
455 | - (inet_peer_gc_maxtime - inet_peer_gc_mintime) / HZ * | 455 | else |
456 | peer_total / inet_peer_threshold * HZ; | 456 | peer_periodic_timer.expires = jiffies |
457 | + inet_peer_gc_maxtime | ||
458 | - (inet_peer_gc_maxtime - inet_peer_gc_mintime) / HZ * | ||
459 | peer_total / inet_peer_threshold * HZ; | ||
457 | add_timer(&peer_periodic_timer); | 460 | add_timer(&peer_periodic_timer); |
458 | } | 461 | } |
459 | 462 | ||
diff --git a/net/ipv4/ip_fragment.c b/net/ipv4/ip_fragment.c index 7f68e27eb4ea..eb377ae15305 100644 --- a/net/ipv4/ip_fragment.c +++ b/net/ipv4/ip_fragment.c | |||
@@ -377,7 +377,7 @@ static struct ipq *ip_frag_create(unsigned hash, struct iphdr *iph, u32 user) | |||
377 | return ip_frag_intern(hash, qp); | 377 | return ip_frag_intern(hash, qp); |
378 | 378 | ||
379 | out_nomem: | 379 | out_nomem: |
380 | NETDEBUG(if (net_ratelimit()) printk(KERN_ERR "ip_frag_create: no memory left !\n")); | 380 | LIMIT_NETDEBUG(printk(KERN_ERR "ip_frag_create: no memory left !\n")); |
381 | return NULL; | 381 | return NULL; |
382 | } | 382 | } |
383 | 383 | ||
@@ -625,10 +625,8 @@ static struct sk_buff *ip_frag_reasm(struct ipq *qp, struct net_device *dev) | |||
625 | return head; | 625 | return head; |
626 | 626 | ||
627 | out_nomem: | 627 | out_nomem: |
628 | NETDEBUG(if (net_ratelimit()) | 628 | LIMIT_NETDEBUG(printk(KERN_ERR "IP: queue_glue: no memory for gluing " |
629 | printk(KERN_ERR | 629 | "queue %p\n", qp)); |
630 | "IP: queue_glue: no memory for gluing queue %p\n", | ||
631 | qp)); | ||
632 | goto out_fail; | 630 | goto out_fail; |
633 | out_oversize: | 631 | out_oversize: |
634 | if (net_ratelimit()) | 632 | if (net_ratelimit()) |
diff --git a/net/ipv4/ip_gre.c b/net/ipv4/ip_gre.c index 884835522224..f0d5740d7e22 100644 --- a/net/ipv4/ip_gre.c +++ b/net/ipv4/ip_gre.c | |||
@@ -290,7 +290,6 @@ static struct ip_tunnel * ipgre_tunnel_locate(struct ip_tunnel_parm *parms, int | |||
290 | 290 | ||
291 | dev_hold(dev); | 291 | dev_hold(dev); |
292 | ipgre_tunnel_link(nt); | 292 | ipgre_tunnel_link(nt); |
293 | /* Do not decrement MOD_USE_COUNT here. */ | ||
294 | return nt; | 293 | return nt; |
295 | 294 | ||
296 | failed: | 295 | failed: |
@@ -1277,12 +1276,28 @@ err1: | |||
1277 | goto out; | 1276 | goto out; |
1278 | } | 1277 | } |
1279 | 1278 | ||
1280 | static void ipgre_fini(void) | 1279 | static void __exit ipgre_destroy_tunnels(void) |
1280 | { | ||
1281 | int prio; | ||
1282 | |||
1283 | for (prio = 0; prio < 4; prio++) { | ||
1284 | int h; | ||
1285 | for (h = 0; h < HASH_SIZE; h++) { | ||
1286 | struct ip_tunnel *t; | ||
1287 | while ((t = tunnels[prio][h]) != NULL) | ||
1288 | unregister_netdevice(t->dev); | ||
1289 | } | ||
1290 | } | ||
1291 | } | ||
1292 | |||
1293 | static void __exit ipgre_fini(void) | ||
1281 | { | 1294 | { |
1282 | if (inet_del_protocol(&ipgre_protocol, IPPROTO_GRE) < 0) | 1295 | if (inet_del_protocol(&ipgre_protocol, IPPROTO_GRE) < 0) |
1283 | printk(KERN_INFO "ipgre close: can't remove protocol\n"); | 1296 | printk(KERN_INFO "ipgre close: can't remove protocol\n"); |
1284 | 1297 | ||
1285 | unregister_netdev(ipgre_fb_tunnel_dev); | 1298 | rtnl_lock(); |
1299 | ipgre_destroy_tunnels(); | ||
1300 | rtnl_unlock(); | ||
1286 | } | 1301 | } |
1287 | 1302 | ||
1288 | module_init(ipgre_init); | 1303 | module_init(ipgre_init); |
diff --git a/net/ipv4/ip_sockglue.c b/net/ipv4/ip_sockglue.c index fc7c481d0d79..ff4bd067b397 100644 --- a/net/ipv4/ip_sockglue.c +++ b/net/ipv4/ip_sockglue.c | |||
@@ -848,6 +848,9 @@ mc_msf_out: | |||
848 | 848 | ||
849 | case IP_IPSEC_POLICY: | 849 | case IP_IPSEC_POLICY: |
850 | case IP_XFRM_POLICY: | 850 | case IP_XFRM_POLICY: |
851 | err = -EPERM; | ||
852 | if (!capable(CAP_NET_ADMIN)) | ||
853 | break; | ||
851 | err = xfrm_user_policy(sk, optname, optval, optlen); | 854 | err = xfrm_user_policy(sk, optname, optval, optlen); |
852 | break; | 855 | break; |
853 | 856 | ||
diff --git a/net/ipv4/ipcomp.c b/net/ipv4/ipcomp.c index 2065944fd9e5..7ded6e60f43a 100644 --- a/net/ipv4/ipcomp.c +++ b/net/ipv4/ipcomp.c | |||
@@ -358,7 +358,7 @@ static struct crypto_tfm **ipcomp_alloc_tfms(const char *alg_name) | |||
358 | int cpu; | 358 | int cpu; |
359 | 359 | ||
360 | /* This can be any valid CPU ID so we don't need locking. */ | 360 | /* This can be any valid CPU ID so we don't need locking. */ |
361 | cpu = smp_processor_id(); | 361 | cpu = raw_smp_processor_id(); |
362 | 362 | ||
363 | list_for_each_entry(pos, &ipcomp_tfms_list, list) { | 363 | list_for_each_entry(pos, &ipcomp_tfms_list, list) { |
364 | struct crypto_tfm *tfm; | 364 | struct crypto_tfm *tfm; |
diff --git a/net/ipv4/ipip.c b/net/ipv4/ipip.c index 68a78731f722..c05c1df0bb04 100644 --- a/net/ipv4/ipip.c +++ b/net/ipv4/ipip.c | |||
@@ -255,7 +255,6 @@ static struct ip_tunnel * ipip_tunnel_locate(struct ip_tunnel_parm *parms, int c | |||
255 | 255 | ||
256 | dev_hold(dev); | 256 | dev_hold(dev); |
257 | ipip_tunnel_link(nt); | 257 | ipip_tunnel_link(nt); |
258 | /* Do not decrement MOD_USE_COUNT here. */ | ||
259 | return nt; | 258 | return nt; |
260 | 259 | ||
261 | failed: | 260 | failed: |
@@ -273,7 +272,7 @@ static void ipip_tunnel_uninit(struct net_device *dev) | |||
273 | dev_put(dev); | 272 | dev_put(dev); |
274 | } | 273 | } |
275 | 274 | ||
276 | static void ipip_err(struct sk_buff *skb, void *__unused) | 275 | static void ipip_err(struct sk_buff *skb, u32 info) |
277 | { | 276 | { |
278 | #ifndef I_WISH_WORLD_WERE_PERFECT | 277 | #ifndef I_WISH_WORLD_WERE_PERFECT |
279 | 278 | ||
@@ -852,11 +851,39 @@ static int __init ipip_fb_tunnel_init(struct net_device *dev) | |||
852 | return 0; | 851 | return 0; |
853 | } | 852 | } |
854 | 853 | ||
854 | #ifdef CONFIG_INET_TUNNEL | ||
855 | static struct xfrm_tunnel ipip_handler = { | 855 | static struct xfrm_tunnel ipip_handler = { |
856 | .handler = ipip_rcv, | 856 | .handler = ipip_rcv, |
857 | .err_handler = ipip_err, | 857 | .err_handler = ipip_err, |
858 | }; | 858 | }; |
859 | 859 | ||
860 | static inline int ipip_register(void) | ||
861 | { | ||
862 | return xfrm4_tunnel_register(&ipip_handler); | ||
863 | } | ||
864 | |||
865 | static inline int ipip_unregister(void) | ||
866 | { | ||
867 | return xfrm4_tunnel_deregister(&ipip_handler); | ||
868 | } | ||
869 | #else | ||
870 | static struct net_protocol ipip_protocol = { | ||
871 | .handler = ipip_rcv, | ||
872 | .err_handler = ipip_err, | ||
873 | .no_policy = 1, | ||
874 | }; | ||
875 | |||
876 | static inline int ipip_register(void) | ||
877 | { | ||
878 | return inet_add_protocol(&ipip_protocol, IPPROTO_IPIP); | ||
879 | } | ||
880 | |||
881 | static inline int ipip_unregister(void) | ||
882 | { | ||
883 | return inet_del_protocol(&ipip_protocol, IPPROTO_IPIP); | ||
884 | } | ||
885 | #endif | ||
886 | |||
860 | static char banner[] __initdata = | 887 | static char banner[] __initdata = |
861 | KERN_INFO "IPv4 over IPv4 tunneling driver\n"; | 888 | KERN_INFO "IPv4 over IPv4 tunneling driver\n"; |
862 | 889 | ||
@@ -866,7 +893,7 @@ static int __init ipip_init(void) | |||
866 | 893 | ||
867 | printk(banner); | 894 | printk(banner); |
868 | 895 | ||
869 | if (xfrm4_tunnel_register(&ipip_handler) < 0) { | 896 | if (ipip_register() < 0) { |
870 | printk(KERN_INFO "ipip init: can't register tunnel\n"); | 897 | printk(KERN_INFO "ipip init: can't register tunnel\n"); |
871 | return -EAGAIN; | 898 | return -EAGAIN; |
872 | } | 899 | } |
@@ -888,16 +915,33 @@ static int __init ipip_init(void) | |||
888 | err2: | 915 | err2: |
889 | free_netdev(ipip_fb_tunnel_dev); | 916 | free_netdev(ipip_fb_tunnel_dev); |
890 | err1: | 917 | err1: |
891 | xfrm4_tunnel_deregister(&ipip_handler); | 918 | ipip_unregister(); |
892 | goto out; | 919 | goto out; |
893 | } | 920 | } |
894 | 921 | ||
922 | static void __exit ipip_destroy_tunnels(void) | ||
923 | { | ||
924 | int prio; | ||
925 | |||
926 | for (prio = 1; prio < 4; prio++) { | ||
927 | int h; | ||
928 | for (h = 0; h < HASH_SIZE; h++) { | ||
929 | struct ip_tunnel *t; | ||
930 | while ((t = tunnels[prio][h]) != NULL) | ||
931 | unregister_netdevice(t->dev); | ||
932 | } | ||
933 | } | ||
934 | } | ||
935 | |||
895 | static void __exit ipip_fini(void) | 936 | static void __exit ipip_fini(void) |
896 | { | 937 | { |
897 | if (xfrm4_tunnel_deregister(&ipip_handler) < 0) | 938 | if (ipip_unregister() < 0) |
898 | printk(KERN_INFO "ipip close: can't deregister tunnel\n"); | 939 | printk(KERN_INFO "ipip close: can't deregister tunnel\n"); |
899 | 940 | ||
900 | unregister_netdev(ipip_fb_tunnel_dev); | 941 | rtnl_lock(); |
942 | ipip_destroy_tunnels(); | ||
943 | unregister_netdevice(ipip_fb_tunnel_dev); | ||
944 | rtnl_unlock(); | ||
901 | } | 945 | } |
902 | 946 | ||
903 | module_init(ipip_init); | 947 | module_init(ipip_init); |
diff --git a/net/ipv4/ipmr.c b/net/ipv4/ipmr.c index 7833d920bdba..dc806b578427 100644 --- a/net/ipv4/ipmr.c +++ b/net/ipv4/ipmr.c | |||
@@ -362,7 +362,7 @@ out: | |||
362 | 362 | ||
363 | /* Fill oifs list. It is called under write locked mrt_lock. */ | 363 | /* Fill oifs list. It is called under write locked mrt_lock. */ |
364 | 364 | ||
365 | static void ipmr_update_threshoulds(struct mfc_cache *cache, unsigned char *ttls) | 365 | static void ipmr_update_thresholds(struct mfc_cache *cache, unsigned char *ttls) |
366 | { | 366 | { |
367 | int vifi; | 367 | int vifi; |
368 | 368 | ||
@@ -727,7 +727,7 @@ static int ipmr_mfc_add(struct mfcctl *mfc, int mrtsock) | |||
727 | if (c != NULL) { | 727 | if (c != NULL) { |
728 | write_lock_bh(&mrt_lock); | 728 | write_lock_bh(&mrt_lock); |
729 | c->mfc_parent = mfc->mfcc_parent; | 729 | c->mfc_parent = mfc->mfcc_parent; |
730 | ipmr_update_threshoulds(c, mfc->mfcc_ttls); | 730 | ipmr_update_thresholds(c, mfc->mfcc_ttls); |
731 | if (!mrtsock) | 731 | if (!mrtsock) |
732 | c->mfc_flags |= MFC_STATIC; | 732 | c->mfc_flags |= MFC_STATIC; |
733 | write_unlock_bh(&mrt_lock); | 733 | write_unlock_bh(&mrt_lock); |
@@ -744,7 +744,7 @@ static int ipmr_mfc_add(struct mfcctl *mfc, int mrtsock) | |||
744 | c->mfc_origin=mfc->mfcc_origin.s_addr; | 744 | c->mfc_origin=mfc->mfcc_origin.s_addr; |
745 | c->mfc_mcastgrp=mfc->mfcc_mcastgrp.s_addr; | 745 | c->mfc_mcastgrp=mfc->mfcc_mcastgrp.s_addr; |
746 | c->mfc_parent=mfc->mfcc_parent; | 746 | c->mfc_parent=mfc->mfcc_parent; |
747 | ipmr_update_threshoulds(c, mfc->mfcc_ttls); | 747 | ipmr_update_thresholds(c, mfc->mfcc_ttls); |
748 | if (!mrtsock) | 748 | if (!mrtsock) |
749 | c->mfc_flags |= MFC_STATIC; | 749 | c->mfc_flags |= MFC_STATIC; |
750 | 750 | ||
diff --git a/net/ipv4/netfilter/ip_conntrack_amanda.c b/net/ipv4/netfilter/ip_conntrack_amanda.c index a78a320eee08..01e1b58322a9 100644 --- a/net/ipv4/netfilter/ip_conntrack_amanda.c +++ b/net/ipv4/netfilter/ip_conntrack_amanda.c | |||
@@ -101,14 +101,13 @@ static int help(struct sk_buff **pskb, | |||
101 | if (port == 0 || len > 5) | 101 | if (port == 0 || len > 5) |
102 | break; | 102 | break; |
103 | 103 | ||
104 | exp = ip_conntrack_expect_alloc(); | 104 | exp = ip_conntrack_expect_alloc(ct); |
105 | if (exp == NULL) { | 105 | if (exp == NULL) { |
106 | ret = NF_DROP; | 106 | ret = NF_DROP; |
107 | goto out; | 107 | goto out; |
108 | } | 108 | } |
109 | 109 | ||
110 | exp->expectfn = NULL; | 110 | exp->expectfn = NULL; |
111 | exp->master = ct; | ||
112 | 111 | ||
113 | exp->tuple.src.ip = ct->tuplehash[IP_CT_DIR_ORIGINAL].tuple.src.ip; | 112 | exp->tuple.src.ip = ct->tuplehash[IP_CT_DIR_ORIGINAL].tuple.src.ip; |
114 | exp->tuple.src.u.tcp.port = 0; | 113 | exp->tuple.src.u.tcp.port = 0; |
@@ -126,10 +125,9 @@ static int help(struct sk_buff **pskb, | |||
126 | ret = ip_nat_amanda_hook(pskb, ctinfo, | 125 | ret = ip_nat_amanda_hook(pskb, ctinfo, |
127 | tmp - amanda_buffer, | 126 | tmp - amanda_buffer, |
128 | len, exp); | 127 | len, exp); |
129 | else if (ip_conntrack_expect_related(exp) != 0) { | 128 | else if (ip_conntrack_expect_related(exp) != 0) |
130 | ip_conntrack_expect_free(exp); | ||
131 | ret = NF_DROP; | 129 | ret = NF_DROP; |
132 | } | 130 | ip_conntrack_expect_put(exp); |
133 | } | 131 | } |
134 | 132 | ||
135 | out: | 133 | out: |
diff --git a/net/ipv4/netfilter/ip_conntrack_core.c b/net/ipv4/netfilter/ip_conntrack_core.c index 4b78ebeb6635..a7f0c821a9b2 100644 --- a/net/ipv4/netfilter/ip_conntrack_core.c +++ b/net/ipv4/netfilter/ip_conntrack_core.c | |||
@@ -137,19 +137,12 @@ ip_ct_invert_tuple(struct ip_conntrack_tuple *inverse, | |||
137 | 137 | ||
138 | 138 | ||
139 | /* ip_conntrack_expect helper functions */ | 139 | /* ip_conntrack_expect helper functions */ |
140 | static void destroy_expect(struct ip_conntrack_expect *exp) | ||
141 | { | ||
142 | ip_conntrack_put(exp->master); | ||
143 | IP_NF_ASSERT(!timer_pending(&exp->timeout)); | ||
144 | kmem_cache_free(ip_conntrack_expect_cachep, exp); | ||
145 | CONNTRACK_STAT_INC(expect_delete); | ||
146 | } | ||
147 | |||
148 | static void unlink_expect(struct ip_conntrack_expect *exp) | 140 | static void unlink_expect(struct ip_conntrack_expect *exp) |
149 | { | 141 | { |
150 | ASSERT_WRITE_LOCK(&ip_conntrack_lock); | 142 | ASSERT_WRITE_LOCK(&ip_conntrack_lock); |
143 | IP_NF_ASSERT(!timer_pending(&exp->timeout)); | ||
151 | list_del(&exp->list); | 144 | list_del(&exp->list); |
152 | /* Logically in destroy_expect, but we hold the lock here. */ | 145 | CONNTRACK_STAT_INC(expect_delete); |
153 | exp->master->expecting--; | 146 | exp->master->expecting--; |
154 | } | 147 | } |
155 | 148 | ||
@@ -160,7 +153,7 @@ static void expectation_timed_out(unsigned long ul_expect) | |||
160 | write_lock_bh(&ip_conntrack_lock); | 153 | write_lock_bh(&ip_conntrack_lock); |
161 | unlink_expect(exp); | 154 | unlink_expect(exp); |
162 | write_unlock_bh(&ip_conntrack_lock); | 155 | write_unlock_bh(&ip_conntrack_lock); |
163 | destroy_expect(exp); | 156 | ip_conntrack_expect_put(exp); |
164 | } | 157 | } |
165 | 158 | ||
166 | /* If an expectation for this connection is found, it gets delete from | 159 | /* If an expectation for this connection is found, it gets delete from |
@@ -198,7 +191,7 @@ static void remove_expectations(struct ip_conntrack *ct) | |||
198 | list_for_each_entry_safe(i, tmp, &ip_conntrack_expect_list, list) { | 191 | list_for_each_entry_safe(i, tmp, &ip_conntrack_expect_list, list) { |
199 | if (i->master == ct && del_timer(&i->timeout)) { | 192 | if (i->master == ct && del_timer(&i->timeout)) { |
200 | unlink_expect(i); | 193 | unlink_expect(i); |
201 | destroy_expect(i); | 194 | ip_conntrack_expect_put(i); |
202 | } | 195 | } |
203 | } | 196 | } |
204 | } | 197 | } |
@@ -517,9 +510,14 @@ init_conntrack(const struct ip_conntrack_tuple *tuple, | |||
517 | /* Welcome, Mr. Bond. We've been expecting you... */ | 510 | /* Welcome, Mr. Bond. We've been expecting you... */ |
518 | __set_bit(IPS_EXPECTED_BIT, &conntrack->status); | 511 | __set_bit(IPS_EXPECTED_BIT, &conntrack->status); |
519 | conntrack->master = exp->master; | 512 | conntrack->master = exp->master; |
520 | #if CONFIG_IP_NF_CONNTRACK_MARK | 513 | #ifdef CONFIG_IP_NF_CONNTRACK_MARK |
521 | conntrack->mark = exp->master->mark; | 514 | conntrack->mark = exp->master->mark; |
522 | #endif | 515 | #endif |
516 | #if defined(CONFIG_IP_NF_TARGET_MASQUERADE) || \ | ||
517 | defined(CONFIG_IP_NF_TARGET_MASQUERADE_MODULE) | ||
518 | /* this is ugly, but there is no other place where to put it */ | ||
519 | conntrack->nat.masq_index = exp->master->nat.masq_index; | ||
520 | #endif | ||
523 | nf_conntrack_get(&conntrack->master->ct_general); | 521 | nf_conntrack_get(&conntrack->master->ct_general); |
524 | CONNTRACK_STAT_INC(expect_new); | 522 | CONNTRACK_STAT_INC(expect_new); |
525 | } else { | 523 | } else { |
@@ -537,7 +535,7 @@ init_conntrack(const struct ip_conntrack_tuple *tuple, | |||
537 | if (exp) { | 535 | if (exp) { |
538 | if (exp->expectfn) | 536 | if (exp->expectfn) |
539 | exp->expectfn(conntrack, exp); | 537 | exp->expectfn(conntrack, exp); |
540 | destroy_expect(exp); | 538 | ip_conntrack_expect_put(exp); |
541 | } | 539 | } |
542 | 540 | ||
543 | return &conntrack->tuplehash[IP_CT_DIR_ORIGINAL]; | 541 | return &conntrack->tuplehash[IP_CT_DIR_ORIGINAL]; |
@@ -729,14 +727,14 @@ void ip_conntrack_unexpect_related(struct ip_conntrack_expect *exp) | |||
729 | if (expect_matches(i, exp) && del_timer(&i->timeout)) { | 727 | if (expect_matches(i, exp) && del_timer(&i->timeout)) { |
730 | unlink_expect(i); | 728 | unlink_expect(i); |
731 | write_unlock_bh(&ip_conntrack_lock); | 729 | write_unlock_bh(&ip_conntrack_lock); |
732 | destroy_expect(i); | 730 | ip_conntrack_expect_put(i); |
733 | return; | 731 | return; |
734 | } | 732 | } |
735 | } | 733 | } |
736 | write_unlock_bh(&ip_conntrack_lock); | 734 | write_unlock_bh(&ip_conntrack_lock); |
737 | } | 735 | } |
738 | 736 | ||
739 | struct ip_conntrack_expect *ip_conntrack_expect_alloc(void) | 737 | struct ip_conntrack_expect *ip_conntrack_expect_alloc(struct ip_conntrack *me) |
740 | { | 738 | { |
741 | struct ip_conntrack_expect *new; | 739 | struct ip_conntrack_expect *new; |
742 | 740 | ||
@@ -745,18 +743,23 @@ struct ip_conntrack_expect *ip_conntrack_expect_alloc(void) | |||
745 | DEBUGP("expect_related: OOM allocating expect\n"); | 743 | DEBUGP("expect_related: OOM allocating expect\n"); |
746 | return NULL; | 744 | return NULL; |
747 | } | 745 | } |
748 | new->master = NULL; | 746 | new->master = me; |
747 | atomic_inc(&new->master->ct_general.use); | ||
748 | atomic_set(&new->use, 1); | ||
749 | return new; | 749 | return new; |
750 | } | 750 | } |
751 | 751 | ||
752 | void ip_conntrack_expect_free(struct ip_conntrack_expect *expect) | 752 | void ip_conntrack_expect_put(struct ip_conntrack_expect *exp) |
753 | { | 753 | { |
754 | kmem_cache_free(ip_conntrack_expect_cachep, expect); | 754 | if (atomic_dec_and_test(&exp->use)) { |
755 | ip_conntrack_put(exp->master); | ||
756 | kmem_cache_free(ip_conntrack_expect_cachep, exp); | ||
757 | } | ||
755 | } | 758 | } |
756 | 759 | ||
757 | static void ip_conntrack_expect_insert(struct ip_conntrack_expect *exp) | 760 | static void ip_conntrack_expect_insert(struct ip_conntrack_expect *exp) |
758 | { | 761 | { |
759 | atomic_inc(&exp->master->ct_general.use); | 762 | atomic_inc(&exp->use); |
760 | exp->master->expecting++; | 763 | exp->master->expecting++; |
761 | list_add(&exp->list, &ip_conntrack_expect_list); | 764 | list_add(&exp->list, &ip_conntrack_expect_list); |
762 | 765 | ||
@@ -778,7 +781,7 @@ static void evict_oldest_expect(struct ip_conntrack *master) | |||
778 | if (i->master == master) { | 781 | if (i->master == master) { |
779 | if (del_timer(&i->timeout)) { | 782 | if (del_timer(&i->timeout)) { |
780 | unlink_expect(i); | 783 | unlink_expect(i); |
781 | destroy_expect(i); | 784 | ip_conntrack_expect_put(i); |
782 | } | 785 | } |
783 | break; | 786 | break; |
784 | } | 787 | } |
@@ -810,8 +813,6 @@ int ip_conntrack_expect_related(struct ip_conntrack_expect *expect) | |||
810 | /* Refresh timer: if it's dying, ignore.. */ | 813 | /* Refresh timer: if it's dying, ignore.. */ |
811 | if (refresh_timer(i)) { | 814 | if (refresh_timer(i)) { |
812 | ret = 0; | 815 | ret = 0; |
813 | /* We don't need the one they've given us. */ | ||
814 | ip_conntrack_expect_free(expect); | ||
815 | goto out; | 816 | goto out; |
816 | } | 817 | } |
817 | } else if (expect_clash(i, expect)) { | 818 | } else if (expect_clash(i, expect)) { |
@@ -881,7 +882,7 @@ void ip_conntrack_helper_unregister(struct ip_conntrack_helper *me) | |||
881 | list_for_each_entry_safe(exp, tmp, &ip_conntrack_expect_list, list) { | 882 | list_for_each_entry_safe(exp, tmp, &ip_conntrack_expect_list, list) { |
882 | if (exp->master->helper == me && del_timer(&exp->timeout)) { | 883 | if (exp->master->helper == me && del_timer(&exp->timeout)) { |
883 | unlink_expect(exp); | 884 | unlink_expect(exp); |
884 | destroy_expect(exp); | 885 | ip_conntrack_expect_put(exp); |
885 | } | 886 | } |
886 | } | 887 | } |
887 | /* Get rid of expecteds, set helpers to NULL. */ | 888 | /* Get rid of expecteds, set helpers to NULL. */ |
@@ -1111,6 +1112,9 @@ void ip_conntrack_cleanup(void) | |||
1111 | schedule(); | 1112 | schedule(); |
1112 | goto i_see_dead_people; | 1113 | goto i_see_dead_people; |
1113 | } | 1114 | } |
1115 | /* wait until all references to ip_conntrack_untracked are dropped */ | ||
1116 | while (atomic_read(&ip_conntrack_untracked.ct_general.use) > 1) | ||
1117 | schedule(); | ||
1114 | 1118 | ||
1115 | kmem_cache_destroy(ip_conntrack_cachep); | 1119 | kmem_cache_destroy(ip_conntrack_cachep); |
1116 | kmem_cache_destroy(ip_conntrack_expect_cachep); | 1120 | kmem_cache_destroy(ip_conntrack_expect_cachep); |
diff --git a/net/ipv4/netfilter/ip_conntrack_ftp.c b/net/ipv4/netfilter/ip_conntrack_ftp.c index fea6dd2a00b6..7a3b773be3f9 100644 --- a/net/ipv4/netfilter/ip_conntrack_ftp.c +++ b/net/ipv4/netfilter/ip_conntrack_ftp.c | |||
@@ -376,7 +376,7 @@ static int help(struct sk_buff **pskb, | |||
376 | fb_ptr + matchoff, matchlen, ntohl(th->seq) + matchoff); | 376 | fb_ptr + matchoff, matchlen, ntohl(th->seq) + matchoff); |
377 | 377 | ||
378 | /* Allocate expectation which will be inserted */ | 378 | /* Allocate expectation which will be inserted */ |
379 | exp = ip_conntrack_expect_alloc(); | 379 | exp = ip_conntrack_expect_alloc(ct); |
380 | if (exp == NULL) { | 380 | if (exp == NULL) { |
381 | ret = NF_DROP; | 381 | ret = NF_DROP; |
382 | goto out; | 382 | goto out; |
@@ -403,8 +403,7 @@ static int help(struct sk_buff **pskb, | |||
403 | networks, or the packet filter itself). */ | 403 | networks, or the packet filter itself). */ |
404 | if (!loose) { | 404 | if (!loose) { |
405 | ret = NF_ACCEPT; | 405 | ret = NF_ACCEPT; |
406 | ip_conntrack_expect_free(exp); | 406 | goto out_put_expect; |
407 | goto out_update_nl; | ||
408 | } | 407 | } |
409 | exp->tuple.dst.ip = htonl((array[0] << 24) | (array[1] << 16) | 408 | exp->tuple.dst.ip = htonl((array[0] << 24) | (array[1] << 16) |
410 | | (array[2] << 8) | array[3]); | 409 | | (array[2] << 8) | array[3]); |
@@ -419,7 +418,6 @@ static int help(struct sk_buff **pskb, | |||
419 | { 0xFFFFFFFF, { .tcp = { 0xFFFF } }, 0xFF }}); | 418 | { 0xFFFFFFFF, { .tcp = { 0xFFFF } }, 0xFF }}); |
420 | 419 | ||
421 | exp->expectfn = NULL; | 420 | exp->expectfn = NULL; |
422 | exp->master = ct; | ||
423 | 421 | ||
424 | /* Now, NAT might want to mangle the packet, and register the | 422 | /* Now, NAT might want to mangle the packet, and register the |
425 | * (possibly changed) expectation itself. */ | 423 | * (possibly changed) expectation itself. */ |
@@ -428,13 +426,15 @@ static int help(struct sk_buff **pskb, | |||
428 | matchoff, matchlen, exp, &seq); | 426 | matchoff, matchlen, exp, &seq); |
429 | else { | 427 | else { |
430 | /* Can't expect this? Best to drop packet now. */ | 428 | /* Can't expect this? Best to drop packet now. */ |
431 | if (ip_conntrack_expect_related(exp) != 0) { | 429 | if (ip_conntrack_expect_related(exp) != 0) |
432 | ip_conntrack_expect_free(exp); | ||
433 | ret = NF_DROP; | 430 | ret = NF_DROP; |
434 | } else | 431 | else |
435 | ret = NF_ACCEPT; | 432 | ret = NF_ACCEPT; |
436 | } | 433 | } |
437 | 434 | ||
435 | out_put_expect: | ||
436 | ip_conntrack_expect_put(exp); | ||
437 | |||
438 | out_update_nl: | 438 | out_update_nl: |
439 | /* Now if this ends in \n, update ftp info. Seq may have been | 439 | /* Now if this ends in \n, update ftp info. Seq may have been |
440 | * adjusted by NAT code. */ | 440 | * adjusted by NAT code. */ |
diff --git a/net/ipv4/netfilter/ip_conntrack_irc.c b/net/ipv4/netfilter/ip_conntrack_irc.c index cd98772cc332..4a28f297d502 100644 --- a/net/ipv4/netfilter/ip_conntrack_irc.c +++ b/net/ipv4/netfilter/ip_conntrack_irc.c | |||
@@ -197,7 +197,7 @@ static int help(struct sk_buff **pskb, | |||
197 | continue; | 197 | continue; |
198 | } | 198 | } |
199 | 199 | ||
200 | exp = ip_conntrack_expect_alloc(); | 200 | exp = ip_conntrack_expect_alloc(ct); |
201 | if (exp == NULL) { | 201 | if (exp == NULL) { |
202 | ret = NF_DROP; | 202 | ret = NF_DROP; |
203 | goto out; | 203 | goto out; |
@@ -221,16 +221,14 @@ static int help(struct sk_buff **pskb, | |||
221 | { { 0, { 0 } }, | 221 | { { 0, { 0 } }, |
222 | { 0xFFFFFFFF, { .tcp = { 0xFFFF } }, 0xFF }}); | 222 | { 0xFFFFFFFF, { .tcp = { 0xFFFF } }, 0xFF }}); |
223 | exp->expectfn = NULL; | 223 | exp->expectfn = NULL; |
224 | exp->master = ct; | ||
225 | if (ip_nat_irc_hook) | 224 | if (ip_nat_irc_hook) |
226 | ret = ip_nat_irc_hook(pskb, ctinfo, | 225 | ret = ip_nat_irc_hook(pskb, ctinfo, |
227 | addr_beg_p - ib_ptr, | 226 | addr_beg_p - ib_ptr, |
228 | addr_end_p - addr_beg_p, | 227 | addr_end_p - addr_beg_p, |
229 | exp); | 228 | exp); |
230 | else if (ip_conntrack_expect_related(exp) != 0) { | 229 | else if (ip_conntrack_expect_related(exp) != 0) |
231 | ip_conntrack_expect_free(exp); | ||
232 | ret = NF_DROP; | 230 | ret = NF_DROP; |
233 | } | 231 | ip_conntrack_expect_put(exp); |
234 | goto out; | 232 | goto out; |
235 | } /* for .. NUM_DCCPROTO */ | 233 | } /* for .. NUM_DCCPROTO */ |
236 | } /* while data < ... */ | 234 | } /* while data < ... */ |
diff --git a/net/ipv4/netfilter/ip_conntrack_standalone.c b/net/ipv4/netfilter/ip_conntrack_standalone.c index 1dd824f3cf0a..61798c46e91d 100644 --- a/net/ipv4/netfilter/ip_conntrack_standalone.c +++ b/net/ipv4/netfilter/ip_conntrack_standalone.c | |||
@@ -985,7 +985,7 @@ EXPORT_SYMBOL(ip_ct_refresh_acct); | |||
985 | EXPORT_SYMBOL(ip_ct_protos); | 985 | EXPORT_SYMBOL(ip_ct_protos); |
986 | EXPORT_SYMBOL(ip_ct_find_proto); | 986 | EXPORT_SYMBOL(ip_ct_find_proto); |
987 | EXPORT_SYMBOL(ip_conntrack_expect_alloc); | 987 | EXPORT_SYMBOL(ip_conntrack_expect_alloc); |
988 | EXPORT_SYMBOL(ip_conntrack_expect_free); | 988 | EXPORT_SYMBOL(ip_conntrack_expect_put); |
989 | EXPORT_SYMBOL(ip_conntrack_expect_related); | 989 | EXPORT_SYMBOL(ip_conntrack_expect_related); |
990 | EXPORT_SYMBOL(ip_conntrack_unexpect_related); | 990 | EXPORT_SYMBOL(ip_conntrack_unexpect_related); |
991 | EXPORT_SYMBOL(ip_conntrack_tuple_taken); | 991 | EXPORT_SYMBOL(ip_conntrack_tuple_taken); |
diff --git a/net/ipv4/netfilter/ip_conntrack_tftp.c b/net/ipv4/netfilter/ip_conntrack_tftp.c index 992fac3e36ee..f8ff170f390a 100644 --- a/net/ipv4/netfilter/ip_conntrack_tftp.c +++ b/net/ipv4/netfilter/ip_conntrack_tftp.c | |||
@@ -65,7 +65,7 @@ static int tftp_help(struct sk_buff **pskb, | |||
65 | DUMP_TUPLE(&ct->tuplehash[IP_CT_DIR_ORIGINAL].tuple); | 65 | DUMP_TUPLE(&ct->tuplehash[IP_CT_DIR_ORIGINAL].tuple); |
66 | DUMP_TUPLE(&ct->tuplehash[IP_CT_DIR_REPLY].tuple); | 66 | DUMP_TUPLE(&ct->tuplehash[IP_CT_DIR_REPLY].tuple); |
67 | 67 | ||
68 | exp = ip_conntrack_expect_alloc(); | 68 | exp = ip_conntrack_expect_alloc(ct); |
69 | if (exp == NULL) | 69 | if (exp == NULL) |
70 | return NF_DROP; | 70 | return NF_DROP; |
71 | 71 | ||
@@ -75,17 +75,15 @@ static int tftp_help(struct sk_buff **pskb, | |||
75 | exp->mask.dst.u.udp.port = 0xffff; | 75 | exp->mask.dst.u.udp.port = 0xffff; |
76 | exp->mask.dst.protonum = 0xff; | 76 | exp->mask.dst.protonum = 0xff; |
77 | exp->expectfn = NULL; | 77 | exp->expectfn = NULL; |
78 | exp->master = ct; | ||
79 | 78 | ||
80 | DEBUGP("expect: "); | 79 | DEBUGP("expect: "); |
81 | DUMP_TUPLE(&exp->tuple); | 80 | DUMP_TUPLE(&exp->tuple); |
82 | DUMP_TUPLE(&exp->mask); | 81 | DUMP_TUPLE(&exp->mask); |
83 | if (ip_nat_tftp_hook) | 82 | if (ip_nat_tftp_hook) |
84 | ret = ip_nat_tftp_hook(pskb, ctinfo, exp); | 83 | ret = ip_nat_tftp_hook(pskb, ctinfo, exp); |
85 | else if (ip_conntrack_expect_related(exp) != 0) { | 84 | else if (ip_conntrack_expect_related(exp) != 0) |
86 | ip_conntrack_expect_free(exp); | ||
87 | ret = NF_DROP; | 85 | ret = NF_DROP; |
88 | } | 86 | ip_conntrack_expect_put(exp); |
89 | break; | 87 | break; |
90 | case TFTP_OPCODE_DATA: | 88 | case TFTP_OPCODE_DATA: |
91 | case TFTP_OPCODE_ACK: | 89 | case TFTP_OPCODE_ACK: |
diff --git a/net/ipv4/netfilter/ip_nat_amanda.c b/net/ipv4/netfilter/ip_nat_amanda.c index da1f412583ed..706c8074f422 100644 --- a/net/ipv4/netfilter/ip_nat_amanda.c +++ b/net/ipv4/netfilter/ip_nat_amanda.c | |||
@@ -56,10 +56,8 @@ static unsigned int help(struct sk_buff **pskb, | |||
56 | break; | 56 | break; |
57 | } | 57 | } |
58 | 58 | ||
59 | if (port == 0) { | 59 | if (port == 0) |
60 | ip_conntrack_expect_free(exp); | ||
61 | return NF_DROP; | 60 | return NF_DROP; |
62 | } | ||
63 | 61 | ||
64 | sprintf(buffer, "%u", port); | 62 | sprintf(buffer, "%u", port); |
65 | ret = ip_nat_mangle_udp_packet(pskb, exp->master, ctinfo, | 63 | ret = ip_nat_mangle_udp_packet(pskb, exp->master, ctinfo, |
diff --git a/net/ipv4/netfilter/ip_nat_ftp.c b/net/ipv4/netfilter/ip_nat_ftp.c index c6000e794ad6..d83757a70d9f 100644 --- a/net/ipv4/netfilter/ip_nat_ftp.c +++ b/net/ipv4/netfilter/ip_nat_ftp.c | |||
@@ -143,10 +143,8 @@ static unsigned int ip_nat_ftp(struct sk_buff **pskb, | |||
143 | break; | 143 | break; |
144 | } | 144 | } |
145 | 145 | ||
146 | if (port == 0) { | 146 | if (port == 0) |
147 | ip_conntrack_expect_free(exp); | ||
148 | return NF_DROP; | 147 | return NF_DROP; |
149 | } | ||
150 | 148 | ||
151 | if (!mangle[type](pskb, newip, port, matchoff, matchlen, ct, ctinfo, | 149 | if (!mangle[type](pskb, newip, port, matchoff, matchlen, ct, ctinfo, |
152 | seq)) { | 150 | seq)) { |
diff --git a/net/ipv4/netfilter/ip_nat_irc.c b/net/ipv4/netfilter/ip_nat_irc.c index 9c1ca3381d56..de31942babe3 100644 --- a/net/ipv4/netfilter/ip_nat_irc.c +++ b/net/ipv4/netfilter/ip_nat_irc.c | |||
@@ -65,10 +65,8 @@ static unsigned int help(struct sk_buff **pskb, | |||
65 | break; | 65 | break; |
66 | } | 66 | } |
67 | 67 | ||
68 | if (port == 0) { | 68 | if (port == 0) |
69 | ip_conntrack_expect_free(exp); | ||
70 | return NF_DROP; | 69 | return NF_DROP; |
71 | } | ||
72 | 70 | ||
73 | /* strlen("\1DCC CHAT chat AAAAAAAA P\1\n")=27 | 71 | /* strlen("\1DCC CHAT chat AAAAAAAA P\1\n")=27 |
74 | * strlen("\1DCC SCHAT chat AAAAAAAA P\1\n")=28 | 72 | * strlen("\1DCC SCHAT chat AAAAAAAA P\1\n")=28 |
diff --git a/net/ipv4/netfilter/ip_nat_proto_icmp.c b/net/ipv4/netfilter/ip_nat_proto_icmp.c index a558cf0eee8a..6596c9ee1655 100644 --- a/net/ipv4/netfilter/ip_nat_proto_icmp.c +++ b/net/ipv4/netfilter/ip_nat_proto_icmp.c | |||
@@ -35,16 +35,17 @@ icmp_unique_tuple(struct ip_conntrack_tuple *tuple, | |||
35 | const struct ip_conntrack *conntrack) | 35 | const struct ip_conntrack *conntrack) |
36 | { | 36 | { |
37 | static u_int16_t id; | 37 | static u_int16_t id; |
38 | unsigned int range_size | 38 | unsigned int range_size; |
39 | = (unsigned int)range->max.icmp.id - range->min.icmp.id + 1; | ||
40 | unsigned int i; | 39 | unsigned int i; |
41 | 40 | ||
41 | range_size = ntohs(range->max.icmp.id) - ntohs(range->min.icmp.id) + 1; | ||
42 | /* If no range specified... */ | 42 | /* If no range specified... */ |
43 | if (!(range->flags & IP_NAT_RANGE_PROTO_SPECIFIED)) | 43 | if (!(range->flags & IP_NAT_RANGE_PROTO_SPECIFIED)) |
44 | range_size = 0xFFFF; | 44 | range_size = 0xFFFF; |
45 | 45 | ||
46 | for (i = 0; i < range_size; i++, id++) { | 46 | for (i = 0; i < range_size; i++, id++) { |
47 | tuple->src.u.icmp.id = range->min.icmp.id + (id % range_size); | 47 | tuple->src.u.icmp.id = htons(ntohs(range->min.icmp.id) + |
48 | (id % range_size)); | ||
48 | if (!ip_nat_used_tuple(tuple, conntrack)) | 49 | if (!ip_nat_used_tuple(tuple, conntrack)) |
49 | return 1; | 50 | return 1; |
50 | } | 51 | } |
diff --git a/net/ipv4/netfilter/ip_nat_proto_tcp.c b/net/ipv4/netfilter/ip_nat_proto_tcp.c index a91cfceff272..a98e36d2b3c6 100644 --- a/net/ipv4/netfilter/ip_nat_proto_tcp.c +++ b/net/ipv4/netfilter/ip_nat_proto_tcp.c | |||
@@ -40,7 +40,8 @@ tcp_unique_tuple(struct ip_conntrack_tuple *tuple, | |||
40 | enum ip_nat_manip_type maniptype, | 40 | enum ip_nat_manip_type maniptype, |
41 | const struct ip_conntrack *conntrack) | 41 | const struct ip_conntrack *conntrack) |
42 | { | 42 | { |
43 | static u_int16_t port, *portptr; | 43 | static u_int16_t port; |
44 | u_int16_t *portptr; | ||
44 | unsigned int range_size, min, i; | 45 | unsigned int range_size, min, i; |
45 | 46 | ||
46 | if (maniptype == IP_NAT_MANIP_SRC) | 47 | if (maniptype == IP_NAT_MANIP_SRC) |
diff --git a/net/ipv4/netfilter/ip_nat_proto_udp.c b/net/ipv4/netfilter/ip_nat_proto_udp.c index c669e3b5f5d0..9f66e5625664 100644 --- a/net/ipv4/netfilter/ip_nat_proto_udp.c +++ b/net/ipv4/netfilter/ip_nat_proto_udp.c | |||
@@ -41,7 +41,8 @@ udp_unique_tuple(struct ip_conntrack_tuple *tuple, | |||
41 | enum ip_nat_manip_type maniptype, | 41 | enum ip_nat_manip_type maniptype, |
42 | const struct ip_conntrack *conntrack) | 42 | const struct ip_conntrack *conntrack) |
43 | { | 43 | { |
44 | static u_int16_t port, *portptr; | 44 | static u_int16_t port; |
45 | u_int16_t *portptr; | ||
45 | unsigned int range_size, min, i; | 46 | unsigned int range_size, min, i; |
46 | 47 | ||
47 | if (maniptype == IP_NAT_MANIP_SRC) | 48 | if (maniptype == IP_NAT_MANIP_SRC) |
diff --git a/net/ipv4/netfilter/ip_nat_standalone.c b/net/ipv4/netfilter/ip_nat_standalone.c index bc59d0d6e89e..91d5ea1dbbc9 100644 --- a/net/ipv4/netfilter/ip_nat_standalone.c +++ b/net/ipv4/netfilter/ip_nat_standalone.c | |||
@@ -102,6 +102,10 @@ ip_nat_fn(unsigned int hooknum, | |||
102 | return NF_ACCEPT; | 102 | return NF_ACCEPT; |
103 | } | 103 | } |
104 | 104 | ||
105 | /* Don't try to NAT if this packet is not conntracked */ | ||
106 | if (ct == &ip_conntrack_untracked) | ||
107 | return NF_ACCEPT; | ||
108 | |||
105 | switch (ctinfo) { | 109 | switch (ctinfo) { |
106 | case IP_CT_RELATED: | 110 | case IP_CT_RELATED: |
107 | case IP_CT_RELATED+IP_CT_IS_REPLY: | 111 | case IP_CT_RELATED+IP_CT_IS_REPLY: |
diff --git a/net/ipv4/netfilter/ip_nat_tftp.c b/net/ipv4/netfilter/ip_nat_tftp.c index 0343e0d64674..2215317c76b7 100644 --- a/net/ipv4/netfilter/ip_nat_tftp.c +++ b/net/ipv4/netfilter/ip_nat_tftp.c | |||
@@ -45,10 +45,8 @@ static unsigned int help(struct sk_buff **pskb, | |||
45 | exp->saved_proto.udp.port = exp->tuple.dst.u.tcp.port; | 45 | exp->saved_proto.udp.port = exp->tuple.dst.u.tcp.port; |
46 | exp->dir = IP_CT_DIR_REPLY; | 46 | exp->dir = IP_CT_DIR_REPLY; |
47 | exp->expectfn = ip_nat_follow_master; | 47 | exp->expectfn = ip_nat_follow_master; |
48 | if (ip_conntrack_expect_related(exp) != 0) { | 48 | if (ip_conntrack_expect_related(exp) != 0) |
49 | ip_conntrack_expect_free(exp); | ||
50 | return NF_DROP; | 49 | return NF_DROP; |
51 | } | ||
52 | return NF_ACCEPT; | 50 | return NF_ACCEPT; |
53 | } | 51 | } |
54 | 52 | ||
diff --git a/net/ipv4/netfilter/ip_queue.c b/net/ipv4/netfilter/ip_queue.c index eda1fba431a4..c6baa8174389 100644 --- a/net/ipv4/netfilter/ip_queue.c +++ b/net/ipv4/netfilter/ip_queue.c | |||
@@ -214,6 +214,12 @@ ipq_build_packet_message(struct ipq_queue_entry *entry, int *errp) | |||
214 | break; | 214 | break; |
215 | 215 | ||
216 | case IPQ_COPY_PACKET: | 216 | case IPQ_COPY_PACKET: |
217 | if (entry->skb->ip_summed == CHECKSUM_HW && | ||
218 | (*errp = skb_checksum_help(entry->skb, | ||
219 | entry->info->outdev == NULL))) { | ||
220 | read_unlock_bh(&queue_lock); | ||
221 | return NULL; | ||
222 | } | ||
217 | if (copy_range == 0 || copy_range > entry->skb->len) | 223 | if (copy_range == 0 || copy_range > entry->skb->len) |
218 | data_len = entry->skb->len; | 224 | data_len = entry->skb->len; |
219 | else | 225 | else |
@@ -385,6 +391,7 @@ ipq_mangle_ipv4(ipq_verdict_msg_t *v, struct ipq_queue_entry *e) | |||
385 | if (!skb_ip_make_writable(&e->skb, v->data_len)) | 391 | if (!skb_ip_make_writable(&e->skb, v->data_len)) |
386 | return -ENOMEM; | 392 | return -ENOMEM; |
387 | memcpy(e->skb->data, v->payload, v->data_len); | 393 | memcpy(e->skb->data, v->payload, v->data_len); |
394 | e->skb->ip_summed = CHECKSUM_NONE; | ||
388 | e->skb->nfcache |= NFC_ALTERED; | 395 | e->skb->nfcache |= NFC_ALTERED; |
389 | 396 | ||
390 | /* | 397 | /* |
diff --git a/net/ipv4/netfilter/ipt_ECN.c b/net/ipv4/netfilter/ipt_ECN.c index ada9911118e9..94a0ce1c1c9d 100644 --- a/net/ipv4/netfilter/ipt_ECN.c +++ b/net/ipv4/netfilter/ipt_ECN.c | |||
@@ -61,16 +61,20 @@ set_ect_tcp(struct sk_buff **pskb, const struct ipt_ECN_info *einfo, int inward) | |||
61 | if (!tcph) | 61 | if (!tcph) |
62 | return 0; | 62 | return 0; |
63 | 63 | ||
64 | if (!(einfo->operation & IPT_ECN_OP_SET_ECE | 64 | if ((!(einfo->operation & IPT_ECN_OP_SET_ECE) || |
65 | || tcph->ece == einfo->proto.tcp.ece) | 65 | tcph->ece == einfo->proto.tcp.ece) && |
66 | && (!(einfo->operation & IPT_ECN_OP_SET_CWR | 66 | ((!(einfo->operation & IPT_ECN_OP_SET_CWR) || |
67 | || tcph->cwr == einfo->proto.tcp.cwr))) | 67 | tcph->cwr == einfo->proto.tcp.cwr))) |
68 | return 1; | 68 | return 1; |
69 | 69 | ||
70 | if (!skb_ip_make_writable(pskb, (*pskb)->nh.iph->ihl*4+sizeof(*tcph))) | 70 | if (!skb_ip_make_writable(pskb, (*pskb)->nh.iph->ihl*4+sizeof(*tcph))) |
71 | return 0; | 71 | return 0; |
72 | tcph = (void *)(*pskb)->nh.iph + (*pskb)->nh.iph->ihl*4; | 72 | tcph = (void *)(*pskb)->nh.iph + (*pskb)->nh.iph->ihl*4; |
73 | 73 | ||
74 | if ((*pskb)->ip_summed == CHECKSUM_HW && | ||
75 | skb_checksum_help(*pskb, inward)) | ||
76 | return 0; | ||
77 | |||
74 | diffs[0] = ((u_int16_t *)tcph)[6]; | 78 | diffs[0] = ((u_int16_t *)tcph)[6]; |
75 | if (einfo->operation & IPT_ECN_OP_SET_ECE) | 79 | if (einfo->operation & IPT_ECN_OP_SET_ECE) |
76 | tcph->ece = einfo->proto.tcp.ece; | 80 | tcph->ece = einfo->proto.tcp.ece; |
@@ -79,13 +83,10 @@ set_ect_tcp(struct sk_buff **pskb, const struct ipt_ECN_info *einfo, int inward) | |||
79 | diffs[1] = ((u_int16_t *)tcph)[6]; | 83 | diffs[1] = ((u_int16_t *)tcph)[6]; |
80 | diffs[0] = diffs[0] ^ 0xFFFF; | 84 | diffs[0] = diffs[0] ^ 0xFFFF; |
81 | 85 | ||
82 | if ((*pskb)->ip_summed != CHECKSUM_HW) | 86 | if ((*pskb)->ip_summed != CHECKSUM_UNNECESSARY) |
83 | tcph->check = csum_fold(csum_partial((char *)diffs, | 87 | tcph->check = csum_fold(csum_partial((char *)diffs, |
84 | sizeof(diffs), | 88 | sizeof(diffs), |
85 | tcph->check^0xFFFF)); | 89 | tcph->check^0xFFFF)); |
86 | else | ||
87 | if (skb_checksum_help(*pskb, inward)) | ||
88 | return 0; | ||
89 | (*pskb)->nfcache |= NFC_ALTERED; | 90 | (*pskb)->nfcache |= NFC_ALTERED; |
90 | return 1; | 91 | return 1; |
91 | } | 92 | } |
diff --git a/net/ipv4/netfilter/ipt_TCPMSS.c b/net/ipv4/netfilter/ipt_TCPMSS.c index 1049050b2bfb..7b84a254440e 100644 --- a/net/ipv4/netfilter/ipt_TCPMSS.c +++ b/net/ipv4/netfilter/ipt_TCPMSS.c | |||
@@ -61,6 +61,10 @@ ipt_tcpmss_target(struct sk_buff **pskb, | |||
61 | if (!skb_ip_make_writable(pskb, (*pskb)->len)) | 61 | if (!skb_ip_make_writable(pskb, (*pskb)->len)) |
62 | return NF_DROP; | 62 | return NF_DROP; |
63 | 63 | ||
64 | if ((*pskb)->ip_summed == CHECKSUM_HW && | ||
65 | skb_checksum_help(*pskb, out == NULL)) | ||
66 | return NF_DROP; | ||
67 | |||
64 | iph = (*pskb)->nh.iph; | 68 | iph = (*pskb)->nh.iph; |
65 | tcplen = (*pskb)->len - iph->ihl*4; | 69 | tcplen = (*pskb)->len - iph->ihl*4; |
66 | 70 | ||
@@ -186,9 +190,6 @@ ipt_tcpmss_target(struct sk_buff **pskb, | |||
186 | newmss); | 190 | newmss); |
187 | 191 | ||
188 | retmodified: | 192 | retmodified: |
189 | /* We never hw checksum SYN packets. */ | ||
190 | BUG_ON((*pskb)->ip_summed == CHECKSUM_HW); | ||
191 | |||
192 | (*pskb)->nfcache |= NFC_UNKNOWN | NFC_ALTERED; | 193 | (*pskb)->nfcache |= NFC_UNKNOWN | NFC_ALTERED; |
193 | return IPT_CONTINUE; | 194 | return IPT_CONTINUE; |
194 | } | 195 | } |
diff --git a/net/ipv4/tcp.c b/net/ipv4/tcp.c index ddb6ce4ecff2..69b1fcf70077 100644 --- a/net/ipv4/tcp.c +++ b/net/ipv4/tcp.c | |||
@@ -584,7 +584,7 @@ static inline void skb_entail(struct sock *sk, struct tcp_sock *tp, | |||
584 | sk_charge_skb(sk, skb); | 584 | sk_charge_skb(sk, skb); |
585 | if (!sk->sk_send_head) | 585 | if (!sk->sk_send_head) |
586 | sk->sk_send_head = skb; | 586 | sk->sk_send_head = skb; |
587 | else if (tp->nonagle&TCP_NAGLE_PUSH) | 587 | if (tp->nonagle & TCP_NAGLE_PUSH) |
588 | tp->nonagle &= ~TCP_NAGLE_PUSH; | 588 | tp->nonagle &= ~TCP_NAGLE_PUSH; |
589 | } | 589 | } |
590 | 590 | ||
diff --git a/net/ipv4/tcp_ipv4.c b/net/ipv4/tcp_ipv4.c index 62f62bb05c2a..67c670886c1f 100644 --- a/net/ipv4/tcp_ipv4.c +++ b/net/ipv4/tcp_ipv4.c | |||
@@ -242,9 +242,14 @@ static int tcp_v4_get_port(struct sock *sk, unsigned short snum) | |||
242 | tcp_port_rover = rover; | 242 | tcp_port_rover = rover; |
243 | spin_unlock(&tcp_portalloc_lock); | 243 | spin_unlock(&tcp_portalloc_lock); |
244 | 244 | ||
245 | /* Exhausted local port range during search? */ | 245 | /* Exhausted local port range during search? It is not |
246 | * possible for us to be holding one of the bind hash | ||
247 | * locks if this test triggers, because if 'remaining' | ||
248 | * drops to zero, we broke out of the do/while loop at | ||
249 | * the top level, not from the 'break;' statement. | ||
250 | */ | ||
246 | ret = 1; | 251 | ret = 1; |
247 | if (remaining <= 0) | 252 | if (unlikely(remaining <= 0)) |
248 | goto fail; | 253 | goto fail; |
249 | 254 | ||
250 | /* OK, here is the one we will use. HEAD is | 255 | /* OK, here is the one we will use. HEAD is |
@@ -1494,12 +1499,11 @@ int tcp_v4_conn_request(struct sock *sk, struct sk_buff *skb) | |||
1494 | * to destinations, already remembered | 1499 | * to destinations, already remembered |
1495 | * to the moment of synflood. | 1500 | * to the moment of synflood. |
1496 | */ | 1501 | */ |
1497 | NETDEBUG(if (net_ratelimit()) \ | 1502 | LIMIT_NETDEBUG(printk(KERN_DEBUG "TCP: drop open " |
1498 | printk(KERN_DEBUG "TCP: drop open " | 1503 | "request from %u.%u." |
1499 | "request from %u.%u." | 1504 | "%u.%u/%u\n", |
1500 | "%u.%u/%u\n", \ | 1505 | NIPQUAD(saddr), |
1501 | NIPQUAD(saddr), | 1506 | ntohs(skb->h.th->source))); |
1502 | ntohs(skb->h.th->source))); | ||
1503 | dst_release(dst); | 1507 | dst_release(dst); |
1504 | goto drop_and_free; | 1508 | goto drop_and_free; |
1505 | } | 1509 | } |
@@ -1627,8 +1631,7 @@ static int tcp_v4_checksum_init(struct sk_buff *skb) | |||
1627 | skb->nh.iph->daddr, skb->csum)) | 1631 | skb->nh.iph->daddr, skb->csum)) |
1628 | return 0; | 1632 | return 0; |
1629 | 1633 | ||
1630 | NETDEBUG(if (net_ratelimit()) | 1634 | LIMIT_NETDEBUG(printk(KERN_DEBUG "hw tcp v4 csum failed\n")); |
1631 | printk(KERN_DEBUG "hw tcp v4 csum failed\n")); | ||
1632 | skb->ip_summed = CHECKSUM_NONE; | 1635 | skb->ip_summed = CHECKSUM_NONE; |
1633 | } | 1636 | } |
1634 | if (skb->len <= 76) { | 1637 | if (skb->len <= 76) { |
diff --git a/net/ipv4/tcp_output.c b/net/ipv4/tcp_output.c index e3f8ea1bfa9c..dd30dd137b74 100644 --- a/net/ipv4/tcp_output.c +++ b/net/ipv4/tcp_output.c | |||
@@ -403,11 +403,9 @@ static void tcp_queue_skb(struct sock *sk, struct sk_buff *skb) | |||
403 | sk->sk_send_head = skb; | 403 | sk->sk_send_head = skb; |
404 | } | 404 | } |
405 | 405 | ||
406 | static void tcp_set_skb_tso_segs(struct sock *sk, struct sk_buff *skb) | 406 | static void tcp_set_skb_tso_segs(struct sock *sk, struct sk_buff *skb, unsigned int mss_now) |
407 | { | 407 | { |
408 | struct tcp_sock *tp = tcp_sk(sk); | 408 | if (skb->len <= mss_now || |
409 | |||
410 | if (skb->len <= tp->mss_cache || | ||
411 | !(sk->sk_route_caps & NETIF_F_TSO)) { | 409 | !(sk->sk_route_caps & NETIF_F_TSO)) { |
412 | /* Avoid the costly divide in the normal | 410 | /* Avoid the costly divide in the normal |
413 | * non-TSO case. | 411 | * non-TSO case. |
@@ -417,10 +415,10 @@ static void tcp_set_skb_tso_segs(struct sock *sk, struct sk_buff *skb) | |||
417 | } else { | 415 | } else { |
418 | unsigned int factor; | 416 | unsigned int factor; |
419 | 417 | ||
420 | factor = skb->len + (tp->mss_cache - 1); | 418 | factor = skb->len + (mss_now - 1); |
421 | factor /= tp->mss_cache; | 419 | factor /= mss_now; |
422 | skb_shinfo(skb)->tso_segs = factor; | 420 | skb_shinfo(skb)->tso_segs = factor; |
423 | skb_shinfo(skb)->tso_size = tp->mss_cache; | 421 | skb_shinfo(skb)->tso_size = mss_now; |
424 | } | 422 | } |
425 | } | 423 | } |
426 | 424 | ||
@@ -429,7 +427,7 @@ static void tcp_set_skb_tso_segs(struct sock *sk, struct sk_buff *skb) | |||
429 | * packet to the list. This won't be called frequently, I hope. | 427 | * packet to the list. This won't be called frequently, I hope. |
430 | * Remember, these are still headerless SKBs at this point. | 428 | * Remember, these are still headerless SKBs at this point. |
431 | */ | 429 | */ |
432 | static int tcp_fragment(struct sock *sk, struct sk_buff *skb, u32 len) | 430 | static int tcp_fragment(struct sock *sk, struct sk_buff *skb, u32 len, unsigned int mss_now) |
433 | { | 431 | { |
434 | struct tcp_sock *tp = tcp_sk(sk); | 432 | struct tcp_sock *tp = tcp_sk(sk); |
435 | struct sk_buff *buff; | 433 | struct sk_buff *buff; |
@@ -492,8 +490,8 @@ static int tcp_fragment(struct sock *sk, struct sk_buff *skb, u32 len) | |||
492 | } | 490 | } |
493 | 491 | ||
494 | /* Fix up tso_factor for both original and new SKB. */ | 492 | /* Fix up tso_factor for both original and new SKB. */ |
495 | tcp_set_skb_tso_segs(sk, skb); | 493 | tcp_set_skb_tso_segs(sk, skb, mss_now); |
496 | tcp_set_skb_tso_segs(sk, buff); | 494 | tcp_set_skb_tso_segs(sk, buff, mss_now); |
497 | 495 | ||
498 | if (TCP_SKB_CB(skb)->sacked & TCPCB_LOST) { | 496 | if (TCP_SKB_CB(skb)->sacked & TCPCB_LOST) { |
499 | tp->lost_out += tcp_skb_pcount(skb); | 497 | tp->lost_out += tcp_skb_pcount(skb); |
@@ -569,7 +567,7 @@ int tcp_trim_head(struct sock *sk, struct sk_buff *skb, u32 len) | |||
569 | * factor and mss. | 567 | * factor and mss. |
570 | */ | 568 | */ |
571 | if (tcp_skb_pcount(skb) > 1) | 569 | if (tcp_skb_pcount(skb) > 1) |
572 | tcp_set_skb_tso_segs(sk, skb); | 570 | tcp_set_skb_tso_segs(sk, skb, tcp_current_mss(sk, 1)); |
573 | 571 | ||
574 | return 0; | 572 | return 0; |
575 | } | 573 | } |
@@ -734,12 +732,14 @@ static inline unsigned int tcp_cwnd_test(struct tcp_sock *tp, struct sk_buff *sk | |||
734 | /* This must be invoked the first time we consider transmitting | 732 | /* This must be invoked the first time we consider transmitting |
735 | * SKB onto the wire. | 733 | * SKB onto the wire. |
736 | */ | 734 | */ |
737 | static inline int tcp_init_tso_segs(struct sock *sk, struct sk_buff *skb) | 735 | static inline int tcp_init_tso_segs(struct sock *sk, struct sk_buff *skb, unsigned int mss_now) |
738 | { | 736 | { |
739 | int tso_segs = tcp_skb_pcount(skb); | 737 | int tso_segs = tcp_skb_pcount(skb); |
740 | 738 | ||
741 | if (!tso_segs) { | 739 | if (!tso_segs || |
742 | tcp_set_skb_tso_segs(sk, skb); | 740 | (tso_segs > 1 && |
741 | skb_shinfo(skb)->tso_size != mss_now)) { | ||
742 | tcp_set_skb_tso_segs(sk, skb, mss_now); | ||
743 | tso_segs = tcp_skb_pcount(skb); | 743 | tso_segs = tcp_skb_pcount(skb); |
744 | } | 744 | } |
745 | return tso_segs; | 745 | return tso_segs; |
@@ -817,7 +817,7 @@ static unsigned int tcp_snd_test(struct sock *sk, struct sk_buff *skb, | |||
817 | struct tcp_sock *tp = tcp_sk(sk); | 817 | struct tcp_sock *tp = tcp_sk(sk); |
818 | unsigned int cwnd_quota; | 818 | unsigned int cwnd_quota; |
819 | 819 | ||
820 | tcp_init_tso_segs(sk, skb); | 820 | tcp_init_tso_segs(sk, skb, cur_mss); |
821 | 821 | ||
822 | if (!tcp_nagle_test(tp, skb, cur_mss, nonagle)) | 822 | if (!tcp_nagle_test(tp, skb, cur_mss, nonagle)) |
823 | return 0; | 823 | return 0; |
@@ -854,14 +854,15 @@ int tcp_may_send_now(struct sock *sk, struct tcp_sock *tp) | |||
854 | * know that all the data is in scatter-gather pages, and that the | 854 | * know that all the data is in scatter-gather pages, and that the |
855 | * packet has never been sent out before (and thus is not cloned). | 855 | * packet has never been sent out before (and thus is not cloned). |
856 | */ | 856 | */ |
857 | static int tso_fragment(struct sock *sk, struct sk_buff *skb, unsigned int len) | 857 | static int tso_fragment(struct sock *sk, struct sk_buff *skb, unsigned int len, unsigned int mss_now) |
858 | { | 858 | { |
859 | struct sk_buff *buff; | 859 | struct sk_buff *buff; |
860 | int nlen = skb->len - len; | 860 | int nlen = skb->len - len; |
861 | u16 flags; | 861 | u16 flags; |
862 | 862 | ||
863 | /* All of a TSO frame must be composed of paged data. */ | 863 | /* All of a TSO frame must be composed of paged data. */ |
864 | BUG_ON(skb->len != skb->data_len); | 864 | if (skb->len != skb->data_len) |
865 | return tcp_fragment(sk, skb, len, mss_now); | ||
865 | 866 | ||
866 | buff = sk_stream_alloc_pskb(sk, 0, 0, GFP_ATOMIC); | 867 | buff = sk_stream_alloc_pskb(sk, 0, 0, GFP_ATOMIC); |
867 | if (unlikely(buff == NULL)) | 868 | if (unlikely(buff == NULL)) |
@@ -887,8 +888,8 @@ static int tso_fragment(struct sock *sk, struct sk_buff *skb, unsigned int len) | |||
887 | skb_split(skb, buff, len); | 888 | skb_split(skb, buff, len); |
888 | 889 | ||
889 | /* Fix up tso_factor for both original and new SKB. */ | 890 | /* Fix up tso_factor for both original and new SKB. */ |
890 | tcp_set_skb_tso_segs(sk, skb); | 891 | tcp_set_skb_tso_segs(sk, skb, mss_now); |
891 | tcp_set_skb_tso_segs(sk, buff); | 892 | tcp_set_skb_tso_segs(sk, buff, mss_now); |
892 | 893 | ||
893 | /* Link BUFF into the send queue. */ | 894 | /* Link BUFF into the send queue. */ |
894 | skb_header_release(buff); | 895 | skb_header_release(buff); |
@@ -924,10 +925,6 @@ static int tcp_tso_should_defer(struct sock *sk, struct tcp_sock *tp, struct sk_ | |||
924 | 925 | ||
925 | limit = min(send_win, cong_win); | 926 | limit = min(send_win, cong_win); |
926 | 927 | ||
927 | /* If sk_send_head can be sent fully now, just do it. */ | ||
928 | if (skb->len <= limit) | ||
929 | return 0; | ||
930 | |||
931 | if (sysctl_tcp_tso_win_divisor) { | 928 | if (sysctl_tcp_tso_win_divisor) { |
932 | u32 chunk = min(tp->snd_wnd, tp->snd_cwnd * tp->mss_cache); | 929 | u32 chunk = min(tp->snd_wnd, tp->snd_cwnd * tp->mss_cache); |
933 | 930 | ||
@@ -972,19 +969,20 @@ static int tcp_write_xmit(struct sock *sk, unsigned int mss_now, int nonagle) | |||
972 | if (unlikely(sk->sk_state == TCP_CLOSE)) | 969 | if (unlikely(sk->sk_state == TCP_CLOSE)) |
973 | return 0; | 970 | return 0; |
974 | 971 | ||
975 | skb = sk->sk_send_head; | ||
976 | if (unlikely(!skb)) | ||
977 | return 0; | ||
978 | |||
979 | tso_segs = tcp_init_tso_segs(sk, skb); | ||
980 | cwnd_quota = tcp_cwnd_test(tp, skb); | ||
981 | if (unlikely(!cwnd_quota)) | ||
982 | goto out; | ||
983 | |||
984 | sent_pkts = 0; | 972 | sent_pkts = 0; |
985 | while (likely(tcp_snd_wnd_test(tp, skb, mss_now))) { | 973 | while ((skb = sk->sk_send_head)) { |
974 | unsigned int limit; | ||
975 | |||
976 | tso_segs = tcp_init_tso_segs(sk, skb, mss_now); | ||
986 | BUG_ON(!tso_segs); | 977 | BUG_ON(!tso_segs); |
987 | 978 | ||
979 | cwnd_quota = tcp_cwnd_test(tp, skb); | ||
980 | if (!cwnd_quota) | ||
981 | break; | ||
982 | |||
983 | if (unlikely(!tcp_snd_wnd_test(tp, skb, mss_now))) | ||
984 | break; | ||
985 | |||
988 | if (tso_segs == 1) { | 986 | if (tso_segs == 1) { |
989 | if (unlikely(!tcp_nagle_test(tp, skb, mss_now, | 987 | if (unlikely(!tcp_nagle_test(tp, skb, mss_now, |
990 | (tcp_skb_is_last(sk, skb) ? | 988 | (tcp_skb_is_last(sk, skb) ? |
@@ -995,9 +993,10 @@ static int tcp_write_xmit(struct sock *sk, unsigned int mss_now, int nonagle) | |||
995 | break; | 993 | break; |
996 | } | 994 | } |
997 | 995 | ||
996 | limit = mss_now; | ||
998 | if (tso_segs > 1) { | 997 | if (tso_segs > 1) { |
999 | u32 limit = tcp_window_allows(tp, skb, | 998 | limit = tcp_window_allows(tp, skb, |
1000 | mss_now, cwnd_quota); | 999 | mss_now, cwnd_quota); |
1001 | 1000 | ||
1002 | if (skb->len < limit) { | 1001 | if (skb->len < limit) { |
1003 | unsigned int trim = skb->len % mss_now; | 1002 | unsigned int trim = skb->len % mss_now; |
@@ -1005,15 +1004,12 @@ static int tcp_write_xmit(struct sock *sk, unsigned int mss_now, int nonagle) | |||
1005 | if (trim) | 1004 | if (trim) |
1006 | limit = skb->len - trim; | 1005 | limit = skb->len - trim; |
1007 | } | 1006 | } |
1008 | if (skb->len > limit) { | ||
1009 | if (tso_fragment(sk, skb, limit)) | ||
1010 | break; | ||
1011 | } | ||
1012 | } else if (unlikely(skb->len > mss_now)) { | ||
1013 | if (unlikely(tcp_fragment(sk, skb, mss_now))) | ||
1014 | break; | ||
1015 | } | 1007 | } |
1016 | 1008 | ||
1009 | if (skb->len > limit && | ||
1010 | unlikely(tso_fragment(sk, skb, limit, mss_now))) | ||
1011 | break; | ||
1012 | |||
1017 | TCP_SKB_CB(skb)->when = tcp_time_stamp; | 1013 | TCP_SKB_CB(skb)->when = tcp_time_stamp; |
1018 | 1014 | ||
1019 | if (unlikely(tcp_transmit_skb(sk, skb_clone(skb, GFP_ATOMIC)))) | 1015 | if (unlikely(tcp_transmit_skb(sk, skb_clone(skb, GFP_ATOMIC)))) |
@@ -1026,27 +1022,12 @@ static int tcp_write_xmit(struct sock *sk, unsigned int mss_now, int nonagle) | |||
1026 | 1022 | ||
1027 | tcp_minshall_update(tp, mss_now, skb); | 1023 | tcp_minshall_update(tp, mss_now, skb); |
1028 | sent_pkts++; | 1024 | sent_pkts++; |
1029 | |||
1030 | /* Do not optimize this to use tso_segs. If we chopped up | ||
1031 | * the packet above, tso_segs will no longer be valid. | ||
1032 | */ | ||
1033 | cwnd_quota -= tcp_skb_pcount(skb); | ||
1034 | |||
1035 | BUG_ON(cwnd_quota < 0); | ||
1036 | if (!cwnd_quota) | ||
1037 | break; | ||
1038 | |||
1039 | skb = sk->sk_send_head; | ||
1040 | if (!skb) | ||
1041 | break; | ||
1042 | tso_segs = tcp_init_tso_segs(sk, skb); | ||
1043 | } | 1025 | } |
1044 | 1026 | ||
1045 | if (likely(sent_pkts)) { | 1027 | if (likely(sent_pkts)) { |
1046 | tcp_cwnd_validate(sk, tp); | 1028 | tcp_cwnd_validate(sk, tp); |
1047 | return 0; | 1029 | return 0; |
1048 | } | 1030 | } |
1049 | out: | ||
1050 | return !tp->packets_out && sk->sk_send_head; | 1031 | return !tp->packets_out && sk->sk_send_head; |
1051 | } | 1032 | } |
1052 | 1033 | ||
@@ -1076,15 +1057,18 @@ void tcp_push_one(struct sock *sk, unsigned int mss_now) | |||
1076 | 1057 | ||
1077 | BUG_ON(!skb || skb->len < mss_now); | 1058 | BUG_ON(!skb || skb->len < mss_now); |
1078 | 1059 | ||
1079 | tso_segs = tcp_init_tso_segs(sk, skb); | 1060 | tso_segs = tcp_init_tso_segs(sk, skb, mss_now); |
1080 | cwnd_quota = tcp_snd_test(sk, skb, mss_now, TCP_NAGLE_PUSH); | 1061 | cwnd_quota = tcp_snd_test(sk, skb, mss_now, TCP_NAGLE_PUSH); |
1081 | 1062 | ||
1082 | if (likely(cwnd_quota)) { | 1063 | if (likely(cwnd_quota)) { |
1064 | unsigned int limit; | ||
1065 | |||
1083 | BUG_ON(!tso_segs); | 1066 | BUG_ON(!tso_segs); |
1084 | 1067 | ||
1068 | limit = mss_now; | ||
1085 | if (tso_segs > 1) { | 1069 | if (tso_segs > 1) { |
1086 | u32 limit = tcp_window_allows(tp, skb, | 1070 | limit = tcp_window_allows(tp, skb, |
1087 | mss_now, cwnd_quota); | 1071 | mss_now, cwnd_quota); |
1088 | 1072 | ||
1089 | if (skb->len < limit) { | 1073 | if (skb->len < limit) { |
1090 | unsigned int trim = skb->len % mss_now; | 1074 | unsigned int trim = skb->len % mss_now; |
@@ -1092,15 +1076,12 @@ void tcp_push_one(struct sock *sk, unsigned int mss_now) | |||
1092 | if (trim) | 1076 | if (trim) |
1093 | limit = skb->len - trim; | 1077 | limit = skb->len - trim; |
1094 | } | 1078 | } |
1095 | if (skb->len > limit) { | ||
1096 | if (unlikely(tso_fragment(sk, skb, limit))) | ||
1097 | return; | ||
1098 | } | ||
1099 | } else if (unlikely(skb->len > mss_now)) { | ||
1100 | if (unlikely(tcp_fragment(sk, skb, mss_now))) | ||
1101 | return; | ||
1102 | } | 1079 | } |
1103 | 1080 | ||
1081 | if (skb->len > limit && | ||
1082 | unlikely(tso_fragment(sk, skb, limit, mss_now))) | ||
1083 | return; | ||
1084 | |||
1104 | /* Send it out now. */ | 1085 | /* Send it out now. */ |
1105 | TCP_SKB_CB(skb)->when = tcp_time_stamp; | 1086 | TCP_SKB_CB(skb)->when = tcp_time_stamp; |
1106 | 1087 | ||
@@ -1386,15 +1367,21 @@ int tcp_retransmit_skb(struct sock *sk, struct sk_buff *skb) | |||
1386 | 1367 | ||
1387 | if (skb->len > cur_mss) { | 1368 | if (skb->len > cur_mss) { |
1388 | int old_factor = tcp_skb_pcount(skb); | 1369 | int old_factor = tcp_skb_pcount(skb); |
1389 | int new_factor; | 1370 | int diff; |
1390 | 1371 | ||
1391 | if (tcp_fragment(sk, skb, cur_mss)) | 1372 | if (tcp_fragment(sk, skb, cur_mss, cur_mss)) |
1392 | return -ENOMEM; /* We'll try again later. */ | 1373 | return -ENOMEM; /* We'll try again later. */ |
1393 | 1374 | ||
1394 | /* New SKB created, account for it. */ | 1375 | /* New SKB created, account for it. */ |
1395 | new_factor = tcp_skb_pcount(skb); | 1376 | diff = old_factor - tcp_skb_pcount(skb) - |
1396 | tp->packets_out -= old_factor - new_factor; | 1377 | tcp_skb_pcount(skb->next); |
1397 | tp->packets_out += tcp_skb_pcount(skb->next); | 1378 | tp->packets_out -= diff; |
1379 | |||
1380 | if (diff > 0) { | ||
1381 | tp->fackets_out -= diff; | ||
1382 | if ((int)tp->fackets_out < 0) | ||
1383 | tp->fackets_out = 0; | ||
1384 | } | ||
1398 | } | 1385 | } |
1399 | 1386 | ||
1400 | /* Collapse two adjacent packets if worthwhile and we can. */ | 1387 | /* Collapse two adjacent packets if worthwhile and we can. */ |
@@ -1991,7 +1978,7 @@ int tcp_write_wakeup(struct sock *sk) | |||
1991 | skb->len > mss) { | 1978 | skb->len > mss) { |
1992 | seg_size = min(seg_size, mss); | 1979 | seg_size = min(seg_size, mss); |
1993 | TCP_SKB_CB(skb)->flags |= TCPCB_FLAG_PSH; | 1980 | TCP_SKB_CB(skb)->flags |= TCPCB_FLAG_PSH; |
1994 | if (tcp_fragment(sk, skb, seg_size)) | 1981 | if (tcp_fragment(sk, skb, seg_size, mss)) |
1995 | return -1; | 1982 | return -1; |
1996 | /* SWS override triggered forced fragmentation. | 1983 | /* SWS override triggered forced fragmentation. |
1997 | * Disable TSO, the connection is too sick. */ | 1984 | * Disable TSO, the connection is too sick. */ |
@@ -2000,7 +1987,7 @@ int tcp_write_wakeup(struct sock *sk) | |||
2000 | sk->sk_route_caps &= ~NETIF_F_TSO; | 1987 | sk->sk_route_caps &= ~NETIF_F_TSO; |
2001 | } | 1988 | } |
2002 | } else if (!tcp_skb_pcount(skb)) | 1989 | } else if (!tcp_skb_pcount(skb)) |
2003 | tcp_set_skb_tso_segs(sk, skb); | 1990 | tcp_set_skb_tso_segs(sk, skb, mss); |
2004 | 1991 | ||
2005 | TCP_SKB_CB(skb)->flags |= TCPCB_FLAG_PSH; | 1992 | TCP_SKB_CB(skb)->flags |= TCPCB_FLAG_PSH; |
2006 | TCP_SKB_CB(skb)->when = tcp_time_stamp; | 1993 | TCP_SKB_CB(skb)->when = tcp_time_stamp; |
diff --git a/net/ipv4/udp.c b/net/ipv4/udp.c index 7c24e64b443f..dc4d07357e3a 100644 --- a/net/ipv4/udp.c +++ b/net/ipv4/udp.c | |||
@@ -628,7 +628,7 @@ back_from_confirm: | |||
628 | /* ... which is an evident application bug. --ANK */ | 628 | /* ... which is an evident application bug. --ANK */ |
629 | release_sock(sk); | 629 | release_sock(sk); |
630 | 630 | ||
631 | NETDEBUG(if (net_ratelimit()) printk(KERN_DEBUG "udp cork app bug 2\n")); | 631 | LIMIT_NETDEBUG(printk(KERN_DEBUG "udp cork app bug 2\n")); |
632 | err = -EINVAL; | 632 | err = -EINVAL; |
633 | goto out; | 633 | goto out; |
634 | } | 634 | } |
@@ -693,7 +693,7 @@ static int udp_sendpage(struct sock *sk, struct page *page, int offset, | |||
693 | if (unlikely(!up->pending)) { | 693 | if (unlikely(!up->pending)) { |
694 | release_sock(sk); | 694 | release_sock(sk); |
695 | 695 | ||
696 | NETDEBUG(if (net_ratelimit()) printk(KERN_DEBUG "udp cork app bug 3\n")); | 696 | LIMIT_NETDEBUG(printk(KERN_DEBUG "udp cork app bug 3\n")); |
697 | return -EINVAL; | 697 | return -EINVAL; |
698 | } | 698 | } |
699 | 699 | ||
@@ -1102,7 +1102,7 @@ static int udp_checksum_init(struct sk_buff *skb, struct udphdr *uh, | |||
1102 | skb->ip_summed = CHECKSUM_UNNECESSARY; | 1102 | skb->ip_summed = CHECKSUM_UNNECESSARY; |
1103 | if (!udp_check(uh, ulen, saddr, daddr, skb->csum)) | 1103 | if (!udp_check(uh, ulen, saddr, daddr, skb->csum)) |
1104 | return 0; | 1104 | return 0; |
1105 | NETDEBUG(if (net_ratelimit()) printk(KERN_DEBUG "udp v4 hw csum failure.\n")); | 1105 | LIMIT_NETDEBUG(printk(KERN_DEBUG "udp v4 hw csum failure.\n")); |
1106 | skb->ip_summed = CHECKSUM_NONE; | 1106 | skb->ip_summed = CHECKSUM_NONE; |
1107 | } | 1107 | } |
1108 | if (skb->ip_summed != CHECKSUM_UNNECESSARY) | 1108 | if (skb->ip_summed != CHECKSUM_UNNECESSARY) |
@@ -1181,14 +1181,13 @@ int udp_rcv(struct sk_buff *skb) | |||
1181 | return(0); | 1181 | return(0); |
1182 | 1182 | ||
1183 | short_packet: | 1183 | short_packet: |
1184 | NETDEBUG(if (net_ratelimit()) | 1184 | LIMIT_NETDEBUG(printk(KERN_DEBUG "UDP: short packet: From %u.%u.%u.%u:%u %d/%d to %u.%u.%u.%u:%u\n", |
1185 | printk(KERN_DEBUG "UDP: short packet: From %u.%u.%u.%u:%u %d/%d to %u.%u.%u.%u:%u\n", | 1185 | NIPQUAD(saddr), |
1186 | NIPQUAD(saddr), | 1186 | ntohs(uh->source), |
1187 | ntohs(uh->source), | 1187 | ulen, |
1188 | ulen, | 1188 | len, |
1189 | len, | 1189 | NIPQUAD(daddr), |
1190 | NIPQUAD(daddr), | 1190 | ntohs(uh->dest))); |
1191 | ntohs(uh->dest))); | ||
1192 | no_header: | 1191 | no_header: |
1193 | UDP_INC_STATS_BH(UDP_MIB_INERRORS); | 1192 | UDP_INC_STATS_BH(UDP_MIB_INERRORS); |
1194 | kfree_skb(skb); | 1193 | kfree_skb(skb); |
@@ -1199,13 +1198,12 @@ csum_error: | |||
1199 | * RFC1122: OK. Discards the bad packet silently (as far as | 1198 | * RFC1122: OK. Discards the bad packet silently (as far as |
1200 | * the network is concerned, anyway) as per 4.1.3.4 (MUST). | 1199 | * the network is concerned, anyway) as per 4.1.3.4 (MUST). |
1201 | */ | 1200 | */ |
1202 | NETDEBUG(if (net_ratelimit()) | 1201 | LIMIT_NETDEBUG(printk(KERN_DEBUG "UDP: bad checksum. From %d.%d.%d.%d:%d to %d.%d.%d.%d:%d ulen %d\n", |
1203 | printk(KERN_DEBUG "UDP: bad checksum. From %d.%d.%d.%d:%d to %d.%d.%d.%d:%d ulen %d\n", | 1202 | NIPQUAD(saddr), |
1204 | NIPQUAD(saddr), | 1203 | ntohs(uh->source), |
1205 | ntohs(uh->source), | 1204 | NIPQUAD(daddr), |
1206 | NIPQUAD(daddr), | 1205 | ntohs(uh->dest), |
1207 | ntohs(uh->dest), | 1206 | ulen)); |
1208 | ulen)); | ||
1209 | drop: | 1207 | drop: |
1210 | UDP_INC_STATS_BH(UDP_MIB_INERRORS); | 1208 | UDP_INC_STATS_BH(UDP_MIB_INERRORS); |
1211 | kfree_skb(skb); | 1209 | kfree_skb(skb); |
diff --git a/net/ipv4/utils.c b/net/ipv4/utils.c deleted file mode 100644 index 6aecd7a43534..000000000000 --- a/net/ipv4/utils.c +++ /dev/null | |||
@@ -1,59 +0,0 @@ | |||
1 | /* | ||
2 | * INET An implementation of the TCP/IP protocol suite for the LINUX | ||
3 | * operating system. INET is implemented using the BSD Socket | ||
4 | * interface as the means of communication with the user level. | ||
5 | * | ||
6 | * Various kernel-resident INET utility functions; mainly | ||
7 | * for format conversion and debugging output. | ||
8 | * | ||
9 | * Version: $Id: utils.c,v 1.8 2000/10/03 07:29:01 anton Exp $ | ||
10 | * | ||
11 | * Author: Fred N. van Kempen, <waltje@uWalt.NL.Mugnet.ORG> | ||
12 | * | ||
13 | * Fixes: | ||
14 | * Alan Cox : verify_area check. | ||
15 | * Alan Cox : removed old debugging. | ||
16 | * Andi Kleen : add net_ratelimit() | ||
17 | * | ||
18 | * This program is free software; you can redistribute it and/or | ||
19 | * modify it under the terms of the GNU General Public License | ||
20 | * as published by the Free Software Foundation; either version | ||
21 | * 2 of the License, or (at your option) any later version. | ||
22 | */ | ||
23 | |||
24 | #include <linux/module.h> | ||
25 | #include <linux/types.h> | ||
26 | #include <asm/byteorder.h> | ||
27 | |||
28 | /* | ||
29 | * Convert an ASCII string to binary IP. | ||
30 | */ | ||
31 | |||
32 | __u32 in_aton(const char *str) | ||
33 | { | ||
34 | unsigned long l; | ||
35 | unsigned int val; | ||
36 | int i; | ||
37 | |||
38 | l = 0; | ||
39 | for (i = 0; i < 4; i++) | ||
40 | { | ||
41 | l <<= 8; | ||
42 | if (*str != '\0') | ||
43 | { | ||
44 | val = 0; | ||
45 | while (*str != '\0' && *str != '.') | ||
46 | { | ||
47 | val *= 10; | ||
48 | val += *str - '0'; | ||
49 | str++; | ||
50 | } | ||
51 | l |= val; | ||
52 | if (*str != '\0') | ||
53 | str++; | ||
54 | } | ||
55 | } | ||
56 | return(htonl(l)); | ||
57 | } | ||
58 | |||
59 | EXPORT_SYMBOL(in_aton); | ||
diff --git a/net/ipv4/xfrm4_tunnel.c b/net/ipv4/xfrm4_tunnel.c index e1fe360ed27a..afbb0d4cc305 100644 --- a/net/ipv4/xfrm4_tunnel.c +++ b/net/ipv4/xfrm4_tunnel.c | |||
@@ -78,10 +78,9 @@ static int ipip_rcv(struct sk_buff *skb) | |||
78 | static void ipip_err(struct sk_buff *skb, u32 info) | 78 | static void ipip_err(struct sk_buff *skb, u32 info) |
79 | { | 79 | { |
80 | struct xfrm_tunnel *handler = ipip_handler; | 80 | struct xfrm_tunnel *handler = ipip_handler; |
81 | u32 arg = info; | ||
82 | 81 | ||
83 | if (handler) | 82 | if (handler) |
84 | handler->err_handler(skb, &arg); | 83 | handler->err_handler(skb, info); |
85 | } | 84 | } |
86 | 85 | ||
87 | static int ipip_init_state(struct xfrm_state *x) | 86 | static int ipip_init_state(struct xfrm_state *x) |
diff --git a/net/ipv6/Kconfig b/net/ipv6/Kconfig index 95163cd52ae0..ab7a9124f985 100644 --- a/net/ipv6/Kconfig +++ b/net/ipv6/Kconfig | |||
@@ -91,7 +91,6 @@ config INET6_TUNNEL | |||
91 | config IPV6_TUNNEL | 91 | config IPV6_TUNNEL |
92 | tristate "IPv6: IPv6-in-IPv6 tunnel" | 92 | tristate "IPv6: IPv6-in-IPv6 tunnel" |
93 | depends on IPV6 | 93 | depends on IPV6 |
94 | select INET6_TUNNEL | ||
95 | ---help--- | 94 | ---help--- |
96 | Support for IPv6-in-IPv6 tunnels described in RFC 2473. | 95 | Support for IPv6-in-IPv6 tunnels described in RFC 2473. |
97 | 96 | ||
diff --git a/net/ipv6/ip6_input.c b/net/ipv6/ip6_input.c index 866f10726c58..10fbb50daea4 100644 --- a/net/ipv6/ip6_input.c +++ b/net/ipv6/ip6_input.c | |||
@@ -198,12 +198,13 @@ resubmit: | |||
198 | if (!raw_sk) { | 198 | if (!raw_sk) { |
199 | if (xfrm6_policy_check(NULL, XFRM_POLICY_IN, skb)) { | 199 | if (xfrm6_policy_check(NULL, XFRM_POLICY_IN, skb)) { |
200 | IP6_INC_STATS_BH(IPSTATS_MIB_INUNKNOWNPROTOS); | 200 | IP6_INC_STATS_BH(IPSTATS_MIB_INUNKNOWNPROTOS); |
201 | icmpv6_param_prob(skb, ICMPV6_UNK_NEXTHDR, nhoff); | 201 | icmpv6_send(skb, ICMPV6_PARAMPROB, |
202 | ICMPV6_UNK_NEXTHDR, nhoff, | ||
203 | skb->dev); | ||
202 | } | 204 | } |
203 | } else { | 205 | } else |
204 | IP6_INC_STATS_BH(IPSTATS_MIB_INDELIVERS); | 206 | IP6_INC_STATS_BH(IPSTATS_MIB_INDELIVERS); |
205 | kfree_skb(skb); | 207 | kfree_skb(skb); |
206 | } | ||
207 | } | 208 | } |
208 | rcu_read_unlock(); | 209 | rcu_read_unlock(); |
209 | return 0; | 210 | return 0; |
diff --git a/net/ipv6/ip6_output.c b/net/ipv6/ip6_output.c index 1f2c2f9e353f..ae652ca14bc9 100644 --- a/net/ipv6/ip6_output.c +++ b/net/ipv6/ip6_output.c | |||
@@ -792,13 +792,8 @@ int ip6_dst_lookup(struct sock *sk, struct dst_entry **dst, struct flowi *fl) | |||
792 | if (ipv6_addr_any(&fl->fl6_src)) { | 792 | if (ipv6_addr_any(&fl->fl6_src)) { |
793 | err = ipv6_get_saddr(*dst, &fl->fl6_dst, &fl->fl6_src); | 793 | err = ipv6_get_saddr(*dst, &fl->fl6_dst, &fl->fl6_src); |
794 | 794 | ||
795 | if (err) { | 795 | if (err) |
796 | #if IP6_DEBUG >= 2 | ||
797 | printk(KERN_DEBUG "ip6_dst_lookup: " | ||
798 | "no available source address\n"); | ||
799 | #endif | ||
800 | goto out_err_release; | 796 | goto out_err_release; |
801 | } | ||
802 | } | 797 | } |
803 | 798 | ||
804 | return 0; | 799 | return 0; |
diff --git a/net/ipv6/ip6_tunnel.c b/net/ipv6/ip6_tunnel.c index ba3b0c267f75..09613729404c 100644 --- a/net/ipv6/ip6_tunnel.c +++ b/net/ipv6/ip6_tunnel.c | |||
@@ -1110,11 +1110,39 @@ ip6ip6_fb_tnl_dev_init(struct net_device *dev) | |||
1110 | return 0; | 1110 | return 0; |
1111 | } | 1111 | } |
1112 | 1112 | ||
1113 | #ifdef CONFIG_INET6_TUNNEL | ||
1113 | static struct xfrm6_tunnel ip6ip6_handler = { | 1114 | static struct xfrm6_tunnel ip6ip6_handler = { |
1114 | .handler = ip6ip6_rcv, | 1115 | .handler = ip6ip6_rcv, |
1115 | .err_handler = ip6ip6_err, | 1116 | .err_handler = ip6ip6_err, |
1116 | }; | 1117 | }; |
1117 | 1118 | ||
1119 | static inline int ip6ip6_register(void) | ||
1120 | { | ||
1121 | return xfrm6_tunnel_register(&ip6ip6_handler); | ||
1122 | } | ||
1123 | |||
1124 | static inline int ip6ip6_unregister(void) | ||
1125 | { | ||
1126 | return xfrm6_tunnel_deregister(&ip6ip6_handler); | ||
1127 | } | ||
1128 | #else | ||
1129 | static struct inet6_protocol xfrm6_tunnel_protocol = { | ||
1130 | .handler = ip6ip6_rcv, | ||
1131 | .err_handler = ip6ip6_err, | ||
1132 | .flags = INET6_PROTO_NOPOLICY|INET6_PROTO_FINAL, | ||
1133 | }; | ||
1134 | |||
1135 | static inline int ip6ip6_register(void) | ||
1136 | { | ||
1137 | return inet6_add_protocol(&xfrm6_tunnel_protocol, IPPROTO_IPV6); | ||
1138 | } | ||
1139 | |||
1140 | static inline int ip6ip6_unregister(void) | ||
1141 | { | ||
1142 | return inet6_del_protocol(&xfrm6_tunnel_protocol, IPPROTO_IPV6); | ||
1143 | } | ||
1144 | #endif | ||
1145 | |||
1118 | /** | 1146 | /** |
1119 | * ip6_tunnel_init - register protocol and reserve needed resources | 1147 | * ip6_tunnel_init - register protocol and reserve needed resources |
1120 | * | 1148 | * |
@@ -1125,7 +1153,7 @@ static int __init ip6_tunnel_init(void) | |||
1125 | { | 1153 | { |
1126 | int err; | 1154 | int err; |
1127 | 1155 | ||
1128 | if (xfrm6_tunnel_register(&ip6ip6_handler) < 0) { | 1156 | if (ip6ip6_register() < 0) { |
1129 | printk(KERN_ERR "ip6ip6 init: can't register tunnel\n"); | 1157 | printk(KERN_ERR "ip6ip6 init: can't register tunnel\n"); |
1130 | return -EAGAIN; | 1158 | return -EAGAIN; |
1131 | } | 1159 | } |
@@ -1144,7 +1172,7 @@ static int __init ip6_tunnel_init(void) | |||
1144 | } | 1172 | } |
1145 | return 0; | 1173 | return 0; |
1146 | fail: | 1174 | fail: |
1147 | xfrm6_tunnel_deregister(&ip6ip6_handler); | 1175 | ip6ip6_unregister(); |
1148 | return err; | 1176 | return err; |
1149 | } | 1177 | } |
1150 | 1178 | ||
@@ -1154,7 +1182,7 @@ fail: | |||
1154 | 1182 | ||
1155 | static void __exit ip6_tunnel_cleanup(void) | 1183 | static void __exit ip6_tunnel_cleanup(void) |
1156 | { | 1184 | { |
1157 | if (xfrm6_tunnel_deregister(&ip6ip6_handler) < 0) | 1185 | if (ip6ip6_unregister() < 0) |
1158 | printk(KERN_INFO "ip6ip6 close: can't deregister tunnel\n"); | 1186 | printk(KERN_INFO "ip6ip6 close: can't deregister tunnel\n"); |
1159 | 1187 | ||
1160 | unregister_netdev(ip6ip6_fb_tnl_dev); | 1188 | unregister_netdev(ip6ip6_fb_tnl_dev); |
diff --git a/net/ipv6/ipcomp6.c b/net/ipv6/ipcomp6.c index 423feb46ccc0..135383ef538f 100644 --- a/net/ipv6/ipcomp6.c +++ b/net/ipv6/ipcomp6.c | |||
@@ -354,7 +354,7 @@ static struct crypto_tfm **ipcomp6_alloc_tfms(const char *alg_name) | |||
354 | int cpu; | 354 | int cpu; |
355 | 355 | ||
356 | /* This can be any valid CPU ID so we don't need locking. */ | 356 | /* This can be any valid CPU ID so we don't need locking. */ |
357 | cpu = smp_processor_id(); | 357 | cpu = raw_smp_processor_id(); |
358 | 358 | ||
359 | list_for_each_entry(pos, &ipcomp6_tfms_list, list) { | 359 | list_for_each_entry(pos, &ipcomp6_tfms_list, list) { |
360 | struct crypto_tfm *tfm; | 360 | struct crypto_tfm *tfm; |
diff --git a/net/ipv6/ipv6_sockglue.c b/net/ipv6/ipv6_sockglue.c index f3ef4c38d315..3bc144a79fa5 100644 --- a/net/ipv6/ipv6_sockglue.c +++ b/net/ipv6/ipv6_sockglue.c | |||
@@ -504,6 +504,9 @@ done: | |||
504 | break; | 504 | break; |
505 | case IPV6_IPSEC_POLICY: | 505 | case IPV6_IPSEC_POLICY: |
506 | case IPV6_XFRM_POLICY: | 506 | case IPV6_XFRM_POLICY: |
507 | retv = -EPERM; | ||
508 | if (!capable(CAP_NET_ADMIN)) | ||
509 | break; | ||
507 | retv = xfrm_user_policy(sk, optname, optval, optlen); | 510 | retv = xfrm_user_policy(sk, optname, optval, optlen); |
508 | break; | 511 | break; |
509 | 512 | ||
diff --git a/net/ipv6/netfilter/ip6_queue.c b/net/ipv6/netfilter/ip6_queue.c index 750943e2d34e..a16df5b27c84 100644 --- a/net/ipv6/netfilter/ip6_queue.c +++ b/net/ipv6/netfilter/ip6_queue.c | |||
@@ -76,7 +76,9 @@ static DECLARE_MUTEX(ipqnl_sem); | |||
76 | static void | 76 | static void |
77 | ipq_issue_verdict(struct ipq_queue_entry *entry, int verdict) | 77 | ipq_issue_verdict(struct ipq_queue_entry *entry, int verdict) |
78 | { | 78 | { |
79 | local_bh_disable(); | ||
79 | nf_reinject(entry->skb, entry->info, verdict); | 80 | nf_reinject(entry->skb, entry->info, verdict); |
81 | local_bh_enable(); | ||
80 | kfree(entry); | 82 | kfree(entry); |
81 | } | 83 | } |
82 | 84 | ||
@@ -209,6 +211,12 @@ ipq_build_packet_message(struct ipq_queue_entry *entry, int *errp) | |||
209 | break; | 211 | break; |
210 | 212 | ||
211 | case IPQ_COPY_PACKET: | 213 | case IPQ_COPY_PACKET: |
214 | if (entry->skb->ip_summed == CHECKSUM_HW && | ||
215 | (*errp = skb_checksum_help(entry->skb, | ||
216 | entry->info->outdev == NULL))) { | ||
217 | read_unlock_bh(&queue_lock); | ||
218 | return NULL; | ||
219 | } | ||
212 | if (copy_range == 0 || copy_range > entry->skb->len) | 220 | if (copy_range == 0 || copy_range > entry->skb->len) |
213 | data_len = entry->skb->len; | 221 | data_len = entry->skb->len; |
214 | else | 222 | else |
@@ -379,6 +387,7 @@ ipq_mangle_ipv6(ipq_verdict_msg_t *v, struct ipq_queue_entry *e) | |||
379 | if (!skb_ip_make_writable(&e->skb, v->data_len)) | 387 | if (!skb_ip_make_writable(&e->skb, v->data_len)) |
380 | return -ENOMEM; | 388 | return -ENOMEM; |
381 | memcpy(e->skb->data, v->payload, v->data_len); | 389 | memcpy(e->skb->data, v->payload, v->data_len); |
390 | e->skb->ip_summed = CHECKSUM_NONE; | ||
382 | e->skb->nfcache |= NFC_ALTERED; | 391 | e->skb->nfcache |= NFC_ALTERED; |
383 | 392 | ||
384 | /* | 393 | /* |
diff --git a/net/ipv6/netfilter/ip6t_LOG.c b/net/ipv6/netfilter/ip6t_LOG.c index c44685e391b7..a692e26a4fa3 100644 --- a/net/ipv6/netfilter/ip6t_LOG.c +++ b/net/ipv6/netfilter/ip6t_LOG.c | |||
@@ -373,9 +373,10 @@ ip6t_log_packet(unsigned int hooknum, | |||
373 | in ? in->name : "", | 373 | in ? in->name : "", |
374 | out ? out->name : ""); | 374 | out ? out->name : ""); |
375 | if (in && !out) { | 375 | if (in && !out) { |
376 | unsigned int len; | ||
376 | /* MAC logging for input chain only. */ | 377 | /* MAC logging for input chain only. */ |
377 | printk("MAC="); | 378 | printk("MAC="); |
378 | if (skb->dev && skb->dev->hard_header_len && | 379 | if (skb->dev && (len = skb->dev->hard_header_len) && |
379 | skb->mac.raw != skb->nh.raw) { | 380 | skb->mac.raw != skb->nh.raw) { |
380 | unsigned char *p = skb->mac.raw; | 381 | unsigned char *p = skb->mac.raw; |
381 | int i; | 382 | int i; |
@@ -384,9 +385,11 @@ ip6t_log_packet(unsigned int hooknum, | |||
384 | (p -= ETH_HLEN) < skb->head) | 385 | (p -= ETH_HLEN) < skb->head) |
385 | p = NULL; | 386 | p = NULL; |
386 | 387 | ||
387 | if (p != NULL) | 388 | if (p != NULL) { |
388 | for (i = 0; i < skb->dev->hard_header_len; i++) | 389 | for (i = 0; i < len; i++) |
389 | printk("%02x", p[i]); | 390 | printk("%02x%s", p[i], |
391 | i == len - 1 ? "" : ":"); | ||
392 | } | ||
390 | printk(" "); | 393 | printk(" "); |
391 | 394 | ||
392 | if (skb->dev->type == ARPHRD_SIT) { | 395 | if (skb->dev->type == ARPHRD_SIT) { |
diff --git a/net/ipv6/raw.c b/net/ipv6/raw.c index e2b848ec9851..1d4d75b34d32 100644 --- a/net/ipv6/raw.c +++ b/net/ipv6/raw.c | |||
@@ -328,6 +328,8 @@ int rawv6_rcv(struct sock *sk, struct sk_buff *skb) | |||
328 | 328 | ||
329 | if (skb->ip_summed != CHECKSUM_UNNECESSARY) { | 329 | if (skb->ip_summed != CHECKSUM_UNNECESSARY) { |
330 | if (skb->ip_summed == CHECKSUM_HW) { | 330 | if (skb->ip_summed == CHECKSUM_HW) { |
331 | skb_postpull_rcsum(skb, skb->nh.raw, | ||
332 | skb->h.raw - skb->nh.raw); | ||
331 | skb->ip_summed = CHECKSUM_UNNECESSARY; | 333 | skb->ip_summed = CHECKSUM_UNNECESSARY; |
332 | if (csum_ipv6_magic(&skb->nh.ipv6h->saddr, | 334 | if (csum_ipv6_magic(&skb->nh.ipv6h->saddr, |
333 | &skb->nh.ipv6h->daddr, | 335 | &skb->nh.ipv6h->daddr, |
diff --git a/net/ipv6/sit.c b/net/ipv6/sit.c index b788f55e139b..e553e5b80d6e 100644 --- a/net/ipv6/sit.c +++ b/net/ipv6/sit.c | |||
@@ -195,7 +195,6 @@ static struct ip_tunnel * ipip6_tunnel_locate(struct ip_tunnel_parm *parms, int | |||
195 | dev_hold(dev); | 195 | dev_hold(dev); |
196 | 196 | ||
197 | ipip6_tunnel_link(nt); | 197 | ipip6_tunnel_link(nt); |
198 | /* Do not decrement MOD_USE_COUNT here. */ | ||
199 | return nt; | 198 | return nt; |
200 | 199 | ||
201 | failed: | 200 | failed: |
@@ -794,10 +793,28 @@ static struct net_protocol sit_protocol = { | |||
794 | .err_handler = ipip6_err, | 793 | .err_handler = ipip6_err, |
795 | }; | 794 | }; |
796 | 795 | ||
796 | static void __exit sit_destroy_tunnels(void) | ||
797 | { | ||
798 | int prio; | ||
799 | |||
800 | for (prio = 1; prio < 4; prio++) { | ||
801 | int h; | ||
802 | for (h = 0; h < HASH_SIZE; h++) { | ||
803 | struct ip_tunnel *t; | ||
804 | while ((t = tunnels[prio][h]) != NULL) | ||
805 | unregister_netdevice(t->dev); | ||
806 | } | ||
807 | } | ||
808 | } | ||
809 | |||
797 | void __exit sit_cleanup(void) | 810 | void __exit sit_cleanup(void) |
798 | { | 811 | { |
799 | inet_del_protocol(&sit_protocol, IPPROTO_IPV6); | 812 | inet_del_protocol(&sit_protocol, IPPROTO_IPV6); |
800 | unregister_netdev(ipip6_fb_tunnel_dev); | 813 | |
814 | rtnl_lock(); | ||
815 | sit_destroy_tunnels(); | ||
816 | unregister_netdevice(ipip6_fb_tunnel_dev); | ||
817 | rtnl_unlock(); | ||
801 | } | 818 | } |
802 | 819 | ||
803 | int __init sit_init(void) | 820 | int __init sit_init(void) |
diff --git a/net/ipv6/tcp_ipv6.c b/net/ipv6/tcp_ipv6.c index f6e288dc116e..ef29cfd936d3 100644 --- a/net/ipv6/tcp_ipv6.c +++ b/net/ipv6/tcp_ipv6.c | |||
@@ -158,9 +158,14 @@ static int tcp_v6_get_port(struct sock *sk, unsigned short snum) | |||
158 | tcp_port_rover = rover; | 158 | tcp_port_rover = rover; |
159 | spin_unlock(&tcp_portalloc_lock); | 159 | spin_unlock(&tcp_portalloc_lock); |
160 | 160 | ||
161 | /* Exhausted local port range during search? */ | 161 | /* Exhausted local port range during search? It is not |
162 | * possible for us to be holding one of the bind hash | ||
163 | * locks if this test triggers, because if 'remaining' | ||
164 | * drops to zero, we broke out of the do/while loop at | ||
165 | * the top level, not from the 'break;' statement. | ||
166 | */ | ||
162 | ret = 1; | 167 | ret = 1; |
163 | if (remaining <= 0) | 168 | if (unlikely(remaining <= 0)) |
164 | goto fail; | 169 | goto fail; |
165 | 170 | ||
166 | /* OK, here is the one we will use. */ | 171 | /* OK, here is the one we will use. */ |
diff --git a/net/netlink/af_netlink.c b/net/netlink/af_netlink.c index 3405fdf41b93..ff774a06c89d 100644 --- a/net/netlink/af_netlink.c +++ b/net/netlink/af_netlink.c | |||
@@ -648,7 +648,8 @@ void netlink_detachskb(struct sock *sk, struct sk_buff *skb) | |||
648 | sock_put(sk); | 648 | sock_put(sk); |
649 | } | 649 | } |
650 | 650 | ||
651 | static inline struct sk_buff *netlink_trim(struct sk_buff *skb, int allocation) | 651 | static inline struct sk_buff *netlink_trim(struct sk_buff *skb, |
652 | unsigned int __nocast allocation) | ||
652 | { | 653 | { |
653 | int delta; | 654 | int delta; |
654 | 655 | ||
@@ -717,7 +718,7 @@ struct netlink_broadcast_data { | |||
717 | int failure; | 718 | int failure; |
718 | int congested; | 719 | int congested; |
719 | int delivered; | 720 | int delivered; |
720 | int allocation; | 721 | unsigned int allocation; |
721 | struct sk_buff *skb, *skb2; | 722 | struct sk_buff *skb, *skb2; |
722 | }; | 723 | }; |
723 | 724 | ||
diff --git a/net/netrom/af_netrom.c b/net/netrom/af_netrom.c index 31ed4a9a1d06..162a85fed150 100644 --- a/net/netrom/af_netrom.c +++ b/net/netrom/af_netrom.c | |||
@@ -459,12 +459,7 @@ static struct sock *nr_make_new(struct sock *osk) | |||
459 | sk->sk_sndbuf = osk->sk_sndbuf; | 459 | sk->sk_sndbuf = osk->sk_sndbuf; |
460 | sk->sk_state = TCP_ESTABLISHED; | 460 | sk->sk_state = TCP_ESTABLISHED; |
461 | sk->sk_sleep = osk->sk_sleep; | 461 | sk->sk_sleep = osk->sk_sleep; |
462 | 462 | sock_copy_flags(sk, osk); | |
463 | if (sock_flag(osk, SOCK_ZAPPED)) | ||
464 | sock_set_flag(sk, SOCK_ZAPPED); | ||
465 | |||
466 | if (sock_flag(osk, SOCK_DBG)) | ||
467 | sock_set_flag(sk, SOCK_DBG); | ||
468 | 463 | ||
469 | skb_queue_head_init(&nr->ack_queue); | 464 | skb_queue_head_init(&nr->ack_queue); |
470 | skb_queue_head_init(&nr->reseq_queue); | 465 | skb_queue_head_init(&nr->reseq_queue); |
@@ -541,7 +536,8 @@ static int nr_bind(struct socket *sock, struct sockaddr *uaddr, int addr_len) | |||
541 | struct nr_sock *nr = nr_sk(sk); | 536 | struct nr_sock *nr = nr_sk(sk); |
542 | struct full_sockaddr_ax25 *addr = (struct full_sockaddr_ax25 *)uaddr; | 537 | struct full_sockaddr_ax25 *addr = (struct full_sockaddr_ax25 *)uaddr; |
543 | struct net_device *dev; | 538 | struct net_device *dev; |
544 | ax25_address *user, *source; | 539 | ax25_uid_assoc *user; |
540 | ax25_address *source; | ||
545 | 541 | ||
546 | lock_sock(sk); | 542 | lock_sock(sk); |
547 | if (!sock_flag(sk, SOCK_ZAPPED)) { | 543 | if (!sock_flag(sk, SOCK_ZAPPED)) { |
@@ -580,16 +576,19 @@ static int nr_bind(struct socket *sock, struct sockaddr *uaddr, int addr_len) | |||
580 | } else { | 576 | } else { |
581 | source = &addr->fsa_ax25.sax25_call; | 577 | source = &addr->fsa_ax25.sax25_call; |
582 | 578 | ||
583 | if ((user = ax25_findbyuid(current->euid)) == NULL) { | 579 | user = ax25_findbyuid(current->euid); |
580 | if (user) { | ||
581 | nr->user_addr = user->call; | ||
582 | ax25_uid_put(user); | ||
583 | } else { | ||
584 | if (ax25_uid_policy && !capable(CAP_NET_BIND_SERVICE)) { | 584 | if (ax25_uid_policy && !capable(CAP_NET_BIND_SERVICE)) { |
585 | release_sock(sk); | 585 | release_sock(sk); |
586 | dev_put(dev); | 586 | dev_put(dev); |
587 | return -EPERM; | 587 | return -EPERM; |
588 | } | 588 | } |
589 | user = source; | 589 | nr->user_addr = *source; |
590 | } | 590 | } |
591 | 591 | ||
592 | nr->user_addr = *user; | ||
593 | nr->source_addr = *source; | 592 | nr->source_addr = *source; |
594 | } | 593 | } |
595 | 594 | ||
@@ -609,7 +608,8 @@ static int nr_connect(struct socket *sock, struct sockaddr *uaddr, | |||
609 | struct sock *sk = sock->sk; | 608 | struct sock *sk = sock->sk; |
610 | struct nr_sock *nr = nr_sk(sk); | 609 | struct nr_sock *nr = nr_sk(sk); |
611 | struct sockaddr_ax25 *addr = (struct sockaddr_ax25 *)uaddr; | 610 | struct sockaddr_ax25 *addr = (struct sockaddr_ax25 *)uaddr; |
612 | ax25_address *user, *source = NULL; | 611 | ax25_address *source = NULL; |
612 | ax25_uid_assoc *user; | ||
613 | struct net_device *dev; | 613 | struct net_device *dev; |
614 | 614 | ||
615 | lock_sock(sk); | 615 | lock_sock(sk); |
@@ -650,16 +650,19 @@ static int nr_connect(struct socket *sock, struct sockaddr *uaddr, | |||
650 | } | 650 | } |
651 | source = (ax25_address *)dev->dev_addr; | 651 | source = (ax25_address *)dev->dev_addr; |
652 | 652 | ||
653 | if ((user = ax25_findbyuid(current->euid)) == NULL) { | 653 | user = ax25_findbyuid(current->euid); |
654 | if (user) { | ||
655 | nr->user_addr = user->call; | ||
656 | ax25_uid_put(user); | ||
657 | } else { | ||
654 | if (ax25_uid_policy && !capable(CAP_NET_ADMIN)) { | 658 | if (ax25_uid_policy && !capable(CAP_NET_ADMIN)) { |
655 | dev_put(dev); | 659 | dev_put(dev); |
656 | release_sock(sk); | 660 | release_sock(sk); |
657 | return -EPERM; | 661 | return -EPERM; |
658 | } | 662 | } |
659 | user = source; | 663 | nr->user_addr = *source; |
660 | } | 664 | } |
661 | 665 | ||
662 | nr->user_addr = *user; | ||
663 | nr->source_addr = *source; | 666 | nr->source_addr = *source; |
664 | nr->device = dev; | 667 | nr->device = dev; |
665 | 668 | ||
diff --git a/net/rose/af_rose.c b/net/rose/af_rose.c index 7eb6a5bf93ea..5480caf8ccc2 100644 --- a/net/rose/af_rose.c +++ b/net/rose/af_rose.c | |||
@@ -556,12 +556,7 @@ static struct sock *rose_make_new(struct sock *osk) | |||
556 | sk->sk_sndbuf = osk->sk_sndbuf; | 556 | sk->sk_sndbuf = osk->sk_sndbuf; |
557 | sk->sk_state = TCP_ESTABLISHED; | 557 | sk->sk_state = TCP_ESTABLISHED; |
558 | sk->sk_sleep = osk->sk_sleep; | 558 | sk->sk_sleep = osk->sk_sleep; |
559 | 559 | sock_copy_flags(sk, osk); | |
560 | if (sock_flag(osk, SOCK_ZAPPED)) | ||
561 | sock_set_flag(sk, SOCK_ZAPPED); | ||
562 | |||
563 | if (sock_flag(osk, SOCK_DBG)) | ||
564 | sock_set_flag(sk, SOCK_DBG); | ||
565 | 560 | ||
566 | init_timer(&rose->timer); | 561 | init_timer(&rose->timer); |
567 | init_timer(&rose->idletimer); | 562 | init_timer(&rose->idletimer); |
@@ -631,7 +626,8 @@ static int rose_bind(struct socket *sock, struct sockaddr *uaddr, int addr_len) | |||
631 | struct rose_sock *rose = rose_sk(sk); | 626 | struct rose_sock *rose = rose_sk(sk); |
632 | struct sockaddr_rose *addr = (struct sockaddr_rose *)uaddr; | 627 | struct sockaddr_rose *addr = (struct sockaddr_rose *)uaddr; |
633 | struct net_device *dev; | 628 | struct net_device *dev; |
634 | ax25_address *user, *source; | 629 | ax25_address *source; |
630 | ax25_uid_assoc *user; | ||
635 | int n; | 631 | int n; |
636 | 632 | ||
637 | if (!sock_flag(sk, SOCK_ZAPPED)) | 633 | if (!sock_flag(sk, SOCK_ZAPPED)) |
@@ -656,14 +652,17 @@ static int rose_bind(struct socket *sock, struct sockaddr *uaddr, int addr_len) | |||
656 | 652 | ||
657 | source = &addr->srose_call; | 653 | source = &addr->srose_call; |
658 | 654 | ||
659 | if ((user = ax25_findbyuid(current->euid)) == NULL) { | 655 | user = ax25_findbyuid(current->euid); |
656 | if (user) { | ||
657 | rose->source_call = user->call; | ||
658 | ax25_uid_put(user); | ||
659 | } else { | ||
660 | if (ax25_uid_policy && !capable(CAP_NET_BIND_SERVICE)) | 660 | if (ax25_uid_policy && !capable(CAP_NET_BIND_SERVICE)) |
661 | return -EACCES; | 661 | return -EACCES; |
662 | user = source; | 662 | rose->source_call = *source; |
663 | } | 663 | } |
664 | 664 | ||
665 | rose->source_addr = addr->srose_addr; | 665 | rose->source_addr = addr->srose_addr; |
666 | rose->source_call = *user; | ||
667 | rose->device = dev; | 666 | rose->device = dev; |
668 | rose->source_ndigis = addr->srose_ndigis; | 667 | rose->source_ndigis = addr->srose_ndigis; |
669 | 668 | ||
@@ -690,8 +689,8 @@ static int rose_connect(struct socket *sock, struct sockaddr *uaddr, int addr_le | |||
690 | struct rose_sock *rose = rose_sk(sk); | 689 | struct rose_sock *rose = rose_sk(sk); |
691 | struct sockaddr_rose *addr = (struct sockaddr_rose *)uaddr; | 690 | struct sockaddr_rose *addr = (struct sockaddr_rose *)uaddr; |
692 | unsigned char cause, diagnostic; | 691 | unsigned char cause, diagnostic; |
693 | ax25_address *user; | ||
694 | struct net_device *dev; | 692 | struct net_device *dev; |
693 | ax25_uid_assoc *user; | ||
695 | int n; | 694 | int n; |
696 | 695 | ||
697 | if (sk->sk_state == TCP_ESTABLISHED && sock->state == SS_CONNECTING) { | 696 | if (sk->sk_state == TCP_ESTABLISHED && sock->state == SS_CONNECTING) { |
@@ -741,12 +740,14 @@ static int rose_connect(struct socket *sock, struct sockaddr *uaddr, int addr_le | |||
741 | if ((dev = rose_dev_first()) == NULL) | 740 | if ((dev = rose_dev_first()) == NULL) |
742 | return -ENETUNREACH; | 741 | return -ENETUNREACH; |
743 | 742 | ||
744 | if ((user = ax25_findbyuid(current->euid)) == NULL) | 743 | user = ax25_findbyuid(current->euid); |
744 | if (!user) | ||
745 | return -EINVAL; | 745 | return -EINVAL; |
746 | 746 | ||
747 | memcpy(&rose->source_addr, dev->dev_addr, ROSE_ADDR_LEN); | 747 | memcpy(&rose->source_addr, dev->dev_addr, ROSE_ADDR_LEN); |
748 | rose->source_call = *user; | 748 | rose->source_call = user->call; |
749 | rose->device = dev; | 749 | rose->device = dev; |
750 | ax25_uid_put(user); | ||
750 | 751 | ||
751 | rose_insert_socket(sk); /* Finish the bind */ | 752 | rose_insert_socket(sk); /* Finish the bind */ |
752 | } | 753 | } |
diff --git a/net/rose/rose_route.c b/net/rose/rose_route.c index ff73ebb912b8..25da6f699fd0 100644 --- a/net/rose/rose_route.c +++ b/net/rose/rose_route.c | |||
@@ -994,8 +994,10 @@ int rose_route_frame(struct sk_buff *skb, ax25_cb *ax25) | |||
994 | * 1. The frame isn't for us, | 994 | * 1. The frame isn't for us, |
995 | * 2. It isn't "owned" by any existing route. | 995 | * 2. It isn't "owned" by any existing route. |
996 | */ | 996 | */ |
997 | if (frametype != ROSE_CALL_REQUEST) /* XXX */ | 997 | if (frametype != ROSE_CALL_REQUEST) { /* XXX */ |
998 | return 0; | 998 | res = 0; |
999 | goto out; | ||
1000 | } | ||
999 | 1001 | ||
1000 | len = (((skb->data[3] >> 4) & 0x0F) + 1) / 2; | 1002 | len = (((skb->data[3] >> 4) & 0x0F) + 1) / 2; |
1001 | len += (((skb->data[3] >> 0) & 0x0F) + 1) / 2; | 1003 | len += (((skb->data[3] >> 0) & 0x0F) + 1) / 2; |
diff --git a/net/sched/em_meta.c b/net/sched/em_meta.c index 53d98f8d3d80..00eae5f9a01a 100644 --- a/net/sched/em_meta.c +++ b/net/sched/em_meta.c | |||
@@ -27,17 +27,17 @@ | |||
27 | * lvalue rvalue | 27 | * lvalue rvalue |
28 | * +-----------+ +-----------+ | 28 | * +-----------+ +-----------+ |
29 | * | type: INT | | type: INT | | 29 | * | type: INT | | type: INT | |
30 | * def | id: INDEV | | id: VALUE | | 30 | * def | id: DEV | | id: VALUE | |
31 | * | data: | | data: 3 | | 31 | * | data: | | data: 3 | |
32 | * +-----------+ +-----------+ | 32 | * +-----------+ +-----------+ |
33 | * | | | 33 | * | | |
34 | * ---> meta_ops[INT][INDEV](...) | | 34 | * ---> meta_ops[INT][DEV](...) | |
35 | * | | | 35 | * | | |
36 | * ----------- | | 36 | * ----------- | |
37 | * V V | 37 | * V V |
38 | * +-----------+ +-----------+ | 38 | * +-----------+ +-----------+ |
39 | * | type: INT | | type: INT | | 39 | * | type: INT | | type: INT | |
40 | * obj | id: INDEV | | id: VALUE | | 40 | * obj | id: DEV | | id: VALUE | |
41 | * | data: 2 |<--data got filled out | data: 3 | | 41 | * | data: 2 |<--data got filled out | data: 3 | |
42 | * +-----------+ +-----------+ | 42 | * +-----------+ +-----------+ |
43 | * | | | 43 | * | | |
@@ -170,26 +170,6 @@ META_COLLECTOR(var_dev) | |||
170 | *err = var_dev(skb->dev, dst); | 170 | *err = var_dev(skb->dev, dst); |
171 | } | 171 | } |
172 | 172 | ||
173 | META_COLLECTOR(int_indev) | ||
174 | { | ||
175 | *err = int_dev(skb->input_dev, dst); | ||
176 | } | ||
177 | |||
178 | META_COLLECTOR(var_indev) | ||
179 | { | ||
180 | *err = var_dev(skb->input_dev, dst); | ||
181 | } | ||
182 | |||
183 | META_COLLECTOR(int_realdev) | ||
184 | { | ||
185 | *err = int_dev(skb->real_dev, dst); | ||
186 | } | ||
187 | |||
188 | META_COLLECTOR(var_realdev) | ||
189 | { | ||
190 | *err = var_dev(skb->real_dev, dst); | ||
191 | } | ||
192 | |||
193 | /************************************************************************** | 173 | /************************************************************************** |
194 | * skb attributes | 174 | * skb attributes |
195 | **************************************************************************/ | 175 | **************************************************************************/ |
@@ -229,12 +209,14 @@ META_COLLECTOR(int_maclen) | |||
229 | * Netfilter | 209 | * Netfilter |
230 | **************************************************************************/ | 210 | **************************************************************************/ |
231 | 211 | ||
232 | #ifdef CONFIG_NETFILTER | ||
233 | META_COLLECTOR(int_nfmark) | 212 | META_COLLECTOR(int_nfmark) |
234 | { | 213 | { |
214 | #ifdef CONFIG_NETFILTER | ||
235 | dst->value = skb->nfmark; | 215 | dst->value = skb->nfmark; |
236 | } | 216 | #else |
217 | dst->value = 0; | ||
237 | #endif | 218 | #endif |
219 | } | ||
238 | 220 | ||
239 | /************************************************************************** | 221 | /************************************************************************** |
240 | * Traffic Control | 222 | * Traffic Control |
@@ -245,31 +227,21 @@ META_COLLECTOR(int_tcindex) | |||
245 | dst->value = skb->tc_index; | 227 | dst->value = skb->tc_index; |
246 | } | 228 | } |
247 | 229 | ||
248 | #ifdef CONFIG_NET_CLS_ACT | ||
249 | META_COLLECTOR(int_tcverd) | ||
250 | { | ||
251 | dst->value = skb->tc_verd; | ||
252 | } | ||
253 | |||
254 | META_COLLECTOR(int_tcclassid) | ||
255 | { | ||
256 | dst->value = skb->tc_classid; | ||
257 | } | ||
258 | #endif | ||
259 | |||
260 | /************************************************************************** | 230 | /************************************************************************** |
261 | * Routing | 231 | * Routing |
262 | **************************************************************************/ | 232 | **************************************************************************/ |
263 | 233 | ||
264 | #ifdef CONFIG_NET_CLS_ROUTE | ||
265 | META_COLLECTOR(int_rtclassid) | 234 | META_COLLECTOR(int_rtclassid) |
266 | { | 235 | { |
267 | if (unlikely(skb->dst == NULL)) | 236 | if (unlikely(skb->dst == NULL)) |
268 | *err = -1; | 237 | *err = -1; |
269 | else | 238 | else |
239 | #ifdef CONFIG_NET_CLS_ROUTE | ||
270 | dst->value = skb->dst->tclassid; | 240 | dst->value = skb->dst->tclassid; |
271 | } | 241 | #else |
242 | dst->value = 0; | ||
272 | #endif | 243 | #endif |
244 | } | ||
273 | 245 | ||
274 | META_COLLECTOR(int_rtiif) | 246 | META_COLLECTOR(int_rtiif) |
275 | { | 247 | { |
@@ -505,8 +477,6 @@ struct meta_ops | |||
505 | static struct meta_ops __meta_ops[TCF_META_TYPE_MAX+1][TCF_META_ID_MAX+1] = { | 477 | static struct meta_ops __meta_ops[TCF_META_TYPE_MAX+1][TCF_META_ID_MAX+1] = { |
506 | [TCF_META_TYPE_VAR] = { | 478 | [TCF_META_TYPE_VAR] = { |
507 | [META_ID(DEV)] = META_FUNC(var_dev), | 479 | [META_ID(DEV)] = META_FUNC(var_dev), |
508 | [META_ID(INDEV)] = META_FUNC(var_indev), | ||
509 | [META_ID(REALDEV)] = META_FUNC(var_realdev), | ||
510 | [META_ID(SK_BOUND_IF)] = META_FUNC(var_sk_bound_if), | 480 | [META_ID(SK_BOUND_IF)] = META_FUNC(var_sk_bound_if), |
511 | }, | 481 | }, |
512 | [TCF_META_TYPE_INT] = { | 482 | [TCF_META_TYPE_INT] = { |
@@ -515,25 +485,15 @@ static struct meta_ops __meta_ops[TCF_META_TYPE_MAX+1][TCF_META_ID_MAX+1] = { | |||
515 | [META_ID(LOADAVG_1)] = META_FUNC(int_loadavg_1), | 485 | [META_ID(LOADAVG_1)] = META_FUNC(int_loadavg_1), |
516 | [META_ID(LOADAVG_2)] = META_FUNC(int_loadavg_2), | 486 | [META_ID(LOADAVG_2)] = META_FUNC(int_loadavg_2), |
517 | [META_ID(DEV)] = META_FUNC(int_dev), | 487 | [META_ID(DEV)] = META_FUNC(int_dev), |
518 | [META_ID(INDEV)] = META_FUNC(int_indev), | ||
519 | [META_ID(REALDEV)] = META_FUNC(int_realdev), | ||
520 | [META_ID(PRIORITY)] = META_FUNC(int_priority), | 488 | [META_ID(PRIORITY)] = META_FUNC(int_priority), |
521 | [META_ID(PROTOCOL)] = META_FUNC(int_protocol), | 489 | [META_ID(PROTOCOL)] = META_FUNC(int_protocol), |
522 | [META_ID(PKTTYPE)] = META_FUNC(int_pkttype), | 490 | [META_ID(PKTTYPE)] = META_FUNC(int_pkttype), |
523 | [META_ID(PKTLEN)] = META_FUNC(int_pktlen), | 491 | [META_ID(PKTLEN)] = META_FUNC(int_pktlen), |
524 | [META_ID(DATALEN)] = META_FUNC(int_datalen), | 492 | [META_ID(DATALEN)] = META_FUNC(int_datalen), |
525 | [META_ID(MACLEN)] = META_FUNC(int_maclen), | 493 | [META_ID(MACLEN)] = META_FUNC(int_maclen), |
526 | #ifdef CONFIG_NETFILTER | ||
527 | [META_ID(NFMARK)] = META_FUNC(int_nfmark), | 494 | [META_ID(NFMARK)] = META_FUNC(int_nfmark), |
528 | #endif | ||
529 | [META_ID(TCINDEX)] = META_FUNC(int_tcindex), | 495 | [META_ID(TCINDEX)] = META_FUNC(int_tcindex), |
530 | #ifdef CONFIG_NET_CLS_ACT | ||
531 | [META_ID(TCVERDICT)] = META_FUNC(int_tcverd), | ||
532 | [META_ID(TCCLASSID)] = META_FUNC(int_tcclassid), | ||
533 | #endif | ||
534 | #ifdef CONFIG_NET_CLS_ROUTE | ||
535 | [META_ID(RTCLASSID)] = META_FUNC(int_rtclassid), | 496 | [META_ID(RTCLASSID)] = META_FUNC(int_rtclassid), |
536 | #endif | ||
537 | [META_ID(RTIIF)] = META_FUNC(int_rtiif), | 497 | [META_ID(RTIIF)] = META_FUNC(int_rtiif), |
538 | [META_ID(SK_FAMILY)] = META_FUNC(int_sk_family), | 498 | [META_ID(SK_FAMILY)] = META_FUNC(int_sk_family), |
539 | [META_ID(SK_STATE)] = META_FUNC(int_sk_state), | 499 | [META_ID(SK_STATE)] = META_FUNC(int_sk_state), |
diff --git a/net/sched/em_text.c b/net/sched/em_text.c index 873840d8d072..77beabc91fa3 100644 --- a/net/sched/em_text.c +++ b/net/sched/em_text.c | |||
@@ -55,9 +55,6 @@ static int em_text_change(struct tcf_proto *tp, void *data, int len, | |||
55 | struct ts_config *ts_conf; | 55 | struct ts_config *ts_conf; |
56 | int flags = 0; | 56 | int flags = 0; |
57 | 57 | ||
58 | printk("Configuring text: %s from %d:%d to %d:%d len %d\n", conf->algo, conf->from_offset, | ||
59 | conf->from_layer, conf->to_offset, conf->to_layer, conf->pattern_len); | ||
60 | |||
61 | if (len < sizeof(*conf) || len < (sizeof(*conf) + conf->pattern_len)) | 58 | if (len < sizeof(*conf) || len < (sizeof(*conf) + conf->pattern_len)) |
62 | return -EINVAL; | 59 | return -EINVAL; |
63 | 60 | ||
diff --git a/net/sched/sch_generic.c b/net/sched/sch_generic.c index 73e218e646ac..0d066c965342 100644 --- a/net/sched/sch_generic.c +++ b/net/sched/sch_generic.c | |||
@@ -331,11 +331,10 @@ static struct sk_buff *pfifo_fast_dequeue(struct Qdisc* qdisc) | |||
331 | int prio; | 331 | int prio; |
332 | struct sk_buff_head *list = qdisc_priv(qdisc); | 332 | struct sk_buff_head *list = qdisc_priv(qdisc); |
333 | 333 | ||
334 | for (prio = 0; prio < PFIFO_FAST_BANDS; prio++, list++) { | 334 | for (prio = 0; prio < PFIFO_FAST_BANDS; prio++) { |
335 | struct sk_buff *skb = __qdisc_dequeue_head(qdisc, list); | 335 | if (!skb_queue_empty(list + prio)) { |
336 | if (skb) { | ||
337 | qdisc->q.qlen--; | 336 | qdisc->q.qlen--; |
338 | return skb; | 337 | return __qdisc_dequeue_head(qdisc, list + prio); |
339 | } | 338 | } |
340 | } | 339 | } |
341 | 340 | ||
@@ -439,6 +438,7 @@ struct Qdisc * qdisc_create_dflt(struct net_device *dev, struct Qdisc_ops *ops) | |||
439 | if (!ops->init || ops->init(sch, NULL) == 0) | 438 | if (!ops->init || ops->init(sch, NULL) == 0) |
440 | return sch; | 439 | return sch; |
441 | 440 | ||
441 | qdisc_destroy(sch); | ||
442 | errout: | 442 | errout: |
443 | return NULL; | 443 | return NULL; |
444 | } | 444 | } |
diff --git a/net/sctp/input.c b/net/sctp/input.c index 5e085e041a6e..742be9171b7d 100644 --- a/net/sctp/input.c +++ b/net/sctp/input.c | |||
@@ -351,7 +351,6 @@ void sctp_icmp_frag_needed(struct sock *sk, struct sctp_association *asoc, | |||
351 | * | 351 | * |
352 | */ | 352 | */ |
353 | void sctp_icmp_proto_unreachable(struct sock *sk, | 353 | void sctp_icmp_proto_unreachable(struct sock *sk, |
354 | struct sctp_endpoint *ep, | ||
355 | struct sctp_association *asoc, | 354 | struct sctp_association *asoc, |
356 | struct sctp_transport *t) | 355 | struct sctp_transport *t) |
357 | { | 356 | { |
@@ -367,7 +366,6 @@ void sctp_icmp_proto_unreachable(struct sock *sk, | |||
367 | /* Common lookup code for icmp/icmpv6 error handler. */ | 366 | /* Common lookup code for icmp/icmpv6 error handler. */ |
368 | struct sock *sctp_err_lookup(int family, struct sk_buff *skb, | 367 | struct sock *sctp_err_lookup(int family, struct sk_buff *skb, |
369 | struct sctphdr *sctphdr, | 368 | struct sctphdr *sctphdr, |
370 | struct sctp_endpoint **epp, | ||
371 | struct sctp_association **app, | 369 | struct sctp_association **app, |
372 | struct sctp_transport **tpp) | 370 | struct sctp_transport **tpp) |
373 | { | 371 | { |
@@ -375,11 +373,10 @@ struct sock *sctp_err_lookup(int family, struct sk_buff *skb, | |||
375 | union sctp_addr daddr; | 373 | union sctp_addr daddr; |
376 | struct sctp_af *af; | 374 | struct sctp_af *af; |
377 | struct sock *sk = NULL; | 375 | struct sock *sk = NULL; |
378 | struct sctp_endpoint *ep = NULL; | ||
379 | struct sctp_association *asoc = NULL; | 376 | struct sctp_association *asoc = NULL; |
380 | struct sctp_transport *transport = NULL; | 377 | struct sctp_transport *transport = NULL; |
381 | 378 | ||
382 | *app = NULL; *epp = NULL; *tpp = NULL; | 379 | *app = NULL; *tpp = NULL; |
383 | 380 | ||
384 | af = sctp_get_af_specific(family); | 381 | af = sctp_get_af_specific(family); |
385 | if (unlikely(!af)) { | 382 | if (unlikely(!af)) { |
@@ -394,26 +391,15 @@ struct sock *sctp_err_lookup(int family, struct sk_buff *skb, | |||
394 | * packet. | 391 | * packet. |
395 | */ | 392 | */ |
396 | asoc = __sctp_lookup_association(&saddr, &daddr, &transport); | 393 | asoc = __sctp_lookup_association(&saddr, &daddr, &transport); |
397 | if (!asoc) { | 394 | if (!asoc) |
398 | /* If there is no matching association, see if it matches any | 395 | return NULL; |
399 | * endpoint. This may happen for an ICMP error generated in | ||
400 | * response to an INIT_ACK. | ||
401 | */ | ||
402 | ep = __sctp_rcv_lookup_endpoint(&daddr); | ||
403 | if (!ep) { | ||
404 | return NULL; | ||
405 | } | ||
406 | } | ||
407 | 396 | ||
408 | if (asoc) { | 397 | sk = asoc->base.sk; |
409 | sk = asoc->base.sk; | ||
410 | 398 | ||
411 | if (ntohl(sctphdr->vtag) != asoc->c.peer_vtag) { | 399 | if (ntohl(sctphdr->vtag) != asoc->c.peer_vtag) { |
412 | ICMP_INC_STATS_BH(ICMP_MIB_INERRORS); | 400 | ICMP_INC_STATS_BH(ICMP_MIB_INERRORS); |
413 | goto out; | 401 | goto out; |
414 | } | 402 | } |
415 | } else | ||
416 | sk = ep->base.sk; | ||
417 | 403 | ||
418 | sctp_bh_lock_sock(sk); | 404 | sctp_bh_lock_sock(sk); |
419 | 405 | ||
@@ -423,7 +409,6 @@ struct sock *sctp_err_lookup(int family, struct sk_buff *skb, | |||
423 | if (sock_owned_by_user(sk)) | 409 | if (sock_owned_by_user(sk)) |
424 | NET_INC_STATS_BH(LINUX_MIB_LOCKDROPPEDICMPS); | 410 | NET_INC_STATS_BH(LINUX_MIB_LOCKDROPPEDICMPS); |
425 | 411 | ||
426 | *epp = ep; | ||
427 | *app = asoc; | 412 | *app = asoc; |
428 | *tpp = transport; | 413 | *tpp = transport; |
429 | return sk; | 414 | return sk; |
@@ -432,21 +417,16 @@ out: | |||
432 | sock_put(sk); | 417 | sock_put(sk); |
433 | if (asoc) | 418 | if (asoc) |
434 | sctp_association_put(asoc); | 419 | sctp_association_put(asoc); |
435 | if (ep) | ||
436 | sctp_endpoint_put(ep); | ||
437 | return NULL; | 420 | return NULL; |
438 | } | 421 | } |
439 | 422 | ||
440 | /* Common cleanup code for icmp/icmpv6 error handler. */ | 423 | /* Common cleanup code for icmp/icmpv6 error handler. */ |
441 | void sctp_err_finish(struct sock *sk, struct sctp_endpoint *ep, | 424 | void sctp_err_finish(struct sock *sk, struct sctp_association *asoc) |
442 | struct sctp_association *asoc) | ||
443 | { | 425 | { |
444 | sctp_bh_unlock_sock(sk); | 426 | sctp_bh_unlock_sock(sk); |
445 | sock_put(sk); | 427 | sock_put(sk); |
446 | if (asoc) | 428 | if (asoc) |
447 | sctp_association_put(asoc); | 429 | sctp_association_put(asoc); |
448 | if (ep) | ||
449 | sctp_endpoint_put(ep); | ||
450 | } | 430 | } |
451 | 431 | ||
452 | /* | 432 | /* |
@@ -471,7 +451,6 @@ void sctp_v4_err(struct sk_buff *skb, __u32 info) | |||
471 | int type = skb->h.icmph->type; | 451 | int type = skb->h.icmph->type; |
472 | int code = skb->h.icmph->code; | 452 | int code = skb->h.icmph->code; |
473 | struct sock *sk; | 453 | struct sock *sk; |
474 | struct sctp_endpoint *ep; | ||
475 | struct sctp_association *asoc; | 454 | struct sctp_association *asoc; |
476 | struct sctp_transport *transport; | 455 | struct sctp_transport *transport; |
477 | struct inet_sock *inet; | 456 | struct inet_sock *inet; |
@@ -488,7 +467,7 @@ void sctp_v4_err(struct sk_buff *skb, __u32 info) | |||
488 | savesctp = skb->h.raw; | 467 | savesctp = skb->h.raw; |
489 | skb->nh.iph = iph; | 468 | skb->nh.iph = iph; |
490 | skb->h.raw = (char *)sh; | 469 | skb->h.raw = (char *)sh; |
491 | sk = sctp_err_lookup(AF_INET, skb, sh, &ep, &asoc, &transport); | 470 | sk = sctp_err_lookup(AF_INET, skb, sh, &asoc, &transport); |
492 | /* Put back, the original pointers. */ | 471 | /* Put back, the original pointers. */ |
493 | skb->nh.raw = saveip; | 472 | skb->nh.raw = saveip; |
494 | skb->h.raw = savesctp; | 473 | skb->h.raw = savesctp; |
@@ -515,7 +494,7 @@ void sctp_v4_err(struct sk_buff *skb, __u32 info) | |||
515 | } | 494 | } |
516 | else { | 495 | else { |
517 | if (ICMP_PROT_UNREACH == code) { | 496 | if (ICMP_PROT_UNREACH == code) { |
518 | sctp_icmp_proto_unreachable(sk, ep, asoc, | 497 | sctp_icmp_proto_unreachable(sk, asoc, |
519 | transport); | 498 | transport); |
520 | goto out_unlock; | 499 | goto out_unlock; |
521 | } | 500 | } |
@@ -544,7 +523,7 @@ void sctp_v4_err(struct sk_buff *skb, __u32 info) | |||
544 | } | 523 | } |
545 | 524 | ||
546 | out_unlock: | 525 | out_unlock: |
547 | sctp_err_finish(sk, ep, asoc); | 526 | sctp_err_finish(sk, asoc); |
548 | } | 527 | } |
549 | 528 | ||
550 | /* | 529 | /* |
diff --git a/net/sctp/ipv6.c b/net/sctp/ipv6.c index c7e42d125b9c..e9b2fd480d61 100644 --- a/net/sctp/ipv6.c +++ b/net/sctp/ipv6.c | |||
@@ -91,7 +91,6 @@ SCTP_STATIC void sctp_v6_err(struct sk_buff *skb, struct inet6_skb_parm *opt, | |||
91 | struct ipv6hdr *iph = (struct ipv6hdr *)skb->data; | 91 | struct ipv6hdr *iph = (struct ipv6hdr *)skb->data; |
92 | struct sctphdr *sh = (struct sctphdr *)(skb->data + offset); | 92 | struct sctphdr *sh = (struct sctphdr *)(skb->data + offset); |
93 | struct sock *sk; | 93 | struct sock *sk; |
94 | struct sctp_endpoint *ep; | ||
95 | struct sctp_association *asoc; | 94 | struct sctp_association *asoc; |
96 | struct sctp_transport *transport; | 95 | struct sctp_transport *transport; |
97 | struct ipv6_pinfo *np; | 96 | struct ipv6_pinfo *np; |
@@ -105,7 +104,7 @@ SCTP_STATIC void sctp_v6_err(struct sk_buff *skb, struct inet6_skb_parm *opt, | |||
105 | savesctp = skb->h.raw; | 104 | savesctp = skb->h.raw; |
106 | skb->nh.ipv6h = iph; | 105 | skb->nh.ipv6h = iph; |
107 | skb->h.raw = (char *)sh; | 106 | skb->h.raw = (char *)sh; |
108 | sk = sctp_err_lookup(AF_INET6, skb, sh, &ep, &asoc, &transport); | 107 | sk = sctp_err_lookup(AF_INET6, skb, sh, &asoc, &transport); |
109 | /* Put back, the original pointers. */ | 108 | /* Put back, the original pointers. */ |
110 | skb->nh.raw = saveip; | 109 | skb->nh.raw = saveip; |
111 | skb->h.raw = savesctp; | 110 | skb->h.raw = savesctp; |
@@ -124,7 +123,7 @@ SCTP_STATIC void sctp_v6_err(struct sk_buff *skb, struct inet6_skb_parm *opt, | |||
124 | goto out_unlock; | 123 | goto out_unlock; |
125 | case ICMPV6_PARAMPROB: | 124 | case ICMPV6_PARAMPROB: |
126 | if (ICMPV6_UNK_NEXTHDR == code) { | 125 | if (ICMPV6_UNK_NEXTHDR == code) { |
127 | sctp_icmp_proto_unreachable(sk, ep, asoc, transport); | 126 | sctp_icmp_proto_unreachable(sk, asoc, transport); |
128 | goto out_unlock; | 127 | goto out_unlock; |
129 | } | 128 | } |
130 | break; | 129 | break; |
@@ -142,7 +141,7 @@ SCTP_STATIC void sctp_v6_err(struct sk_buff *skb, struct inet6_skb_parm *opt, | |||
142 | } | 141 | } |
143 | 142 | ||
144 | out_unlock: | 143 | out_unlock: |
145 | sctp_err_finish(sk, ep, asoc); | 144 | sctp_err_finish(sk, asoc); |
146 | out: | 145 | out: |
147 | if (likely(idev != NULL)) | 146 | if (likely(idev != NULL)) |
148 | in6_dev_put(idev); | 147 | in6_dev_put(idev); |
diff --git a/net/sctp/objcnt.c b/net/sctp/objcnt.c index 0781e5d509fd..8ff588f0d76a 100644 --- a/net/sctp/objcnt.c +++ b/net/sctp/objcnt.c | |||
@@ -127,8 +127,12 @@ done: | |||
127 | /* Initialize the objcount in the proc filesystem. */ | 127 | /* Initialize the objcount in the proc filesystem. */ |
128 | void sctp_dbg_objcnt_init(void) | 128 | void sctp_dbg_objcnt_init(void) |
129 | { | 129 | { |
130 | create_proc_read_entry("sctp_dbg_objcnt", 0, proc_net_sctp, | 130 | struct proc_dir_entry *ent; |
131 | ent = create_proc_read_entry("sctp_dbg_objcnt", 0, proc_net_sctp, | ||
131 | sctp_dbg_objcnt_read, NULL); | 132 | sctp_dbg_objcnt_read, NULL); |
133 | if (!ent) | ||
134 | printk(KERN_WARNING | ||
135 | "sctp_dbg_objcnt: Unable to create /proc entry.\n"); | ||
132 | } | 136 | } |
133 | 137 | ||
134 | /* Cleanup the objcount entry in the proc filesystem. */ | 138 | /* Cleanup the objcount entry in the proc filesystem. */ |
diff --git a/net/sctp/proc.c b/net/sctp/proc.c index 98d49ec9b74b..b74f7772b576 100644 --- a/net/sctp/proc.c +++ b/net/sctp/proc.c | |||
@@ -57,6 +57,7 @@ static struct snmp_mib sctp_snmp_list[] = { | |||
57 | SNMP_MIB_ITEM("SctpReasmUsrMsgs", SCTP_MIB_REASMUSRMSGS), | 57 | SNMP_MIB_ITEM("SctpReasmUsrMsgs", SCTP_MIB_REASMUSRMSGS), |
58 | SNMP_MIB_ITEM("SctpOutSCTPPacks", SCTP_MIB_OUTSCTPPACKS), | 58 | SNMP_MIB_ITEM("SctpOutSCTPPacks", SCTP_MIB_OUTSCTPPACKS), |
59 | SNMP_MIB_ITEM("SctpInSCTPPacks", SCTP_MIB_INSCTPPACKS), | 59 | SNMP_MIB_ITEM("SctpInSCTPPacks", SCTP_MIB_INSCTPPACKS), |
60 | SNMP_MIB_SENTINEL | ||
60 | }; | 61 | }; |
61 | 62 | ||
62 | /* Return the current value of a particular entry in the mib by adding its | 63 | /* Return the current value of a particular entry in the mib by adding its |
diff --git a/net/sunrpc/auth_gss/gss_krb5_crypto.c b/net/sunrpc/auth_gss/gss_krb5_crypto.c index 24c21f2a33a7..5a7265aeaf83 100644 --- a/net/sunrpc/auth_gss/gss_krb5_crypto.c +++ b/net/sunrpc/auth_gss/gss_krb5_crypto.c | |||
@@ -185,9 +185,7 @@ make_checksum(s32 cksumtype, char *header, int hdrlen, struct xdr_buf *body, | |||
185 | sg->page = body->pages[i]; | 185 | sg->page = body->pages[i]; |
186 | sg->offset = offset; | 186 | sg->offset = offset; |
187 | sg->length = thislen; | 187 | sg->length = thislen; |
188 | kmap(sg->page); /* XXX kmap_atomic? */ | ||
189 | crypto_digest_update(tfm, sg, 1); | 188 | crypto_digest_update(tfm, sg, 1); |
190 | kunmap(sg->page); | ||
191 | len -= thislen; | 189 | len -= thislen; |
192 | i++; | 190 | i++; |
193 | offset = 0; | 191 | offset = 0; |
diff --git a/net/sunrpc/svcsock.c b/net/sunrpc/svcsock.c index 56db8f13e6cb..d0c3120d0233 100644 --- a/net/sunrpc/svcsock.c +++ b/net/sunrpc/svcsock.c | |||
@@ -586,7 +586,7 @@ svc_udp_recvfrom(struct svc_rqst *rqstp) | |||
586 | } | 586 | } |
587 | if (skb->stamp.tv_sec == 0) { | 587 | if (skb->stamp.tv_sec == 0) { |
588 | skb->stamp.tv_sec = xtime.tv_sec; | 588 | skb->stamp.tv_sec = xtime.tv_sec; |
589 | skb->stamp.tv_usec = xtime.tv_nsec * 1000; | 589 | skb->stamp.tv_usec = xtime.tv_nsec / NSEC_PER_USEC; |
590 | /* Don't enable netstamp, sunrpc doesn't | 590 | /* Don't enable netstamp, sunrpc doesn't |
591 | need that much accuracy */ | 591 | need that much accuracy */ |
592 | } | 592 | } |
diff --git a/net/sunrpc/xdr.c b/net/sunrpc/xdr.c index 8a4d9c106af1..fde16f40a581 100644 --- a/net/sunrpc/xdr.c +++ b/net/sunrpc/xdr.c | |||
@@ -993,6 +993,7 @@ xdr_xcode_array2(struct xdr_buf *buf, unsigned int base, | |||
993 | return -EINVAL; | 993 | return -EINVAL; |
994 | } else { | 994 | } else { |
995 | if (xdr_decode_word(buf, base, &desc->array_len) != 0 || | 995 | if (xdr_decode_word(buf, base, &desc->array_len) != 0 || |
996 | desc->array_len > desc->array_maxlen || | ||
996 | (unsigned long) base + 4 + desc->array_len * | 997 | (unsigned long) base + 4 + desc->array_len * |
997 | desc->elem_size > buf->len) | 998 | desc->elem_size > buf->len) |
998 | return -EINVAL; | 999 | return -EINVAL; |
diff --git a/net/xfrm/xfrm_user.c b/net/xfrm/xfrm_user.c index ecade4893a13..8da3e25b2c4c 100644 --- a/net/xfrm/xfrm_user.c +++ b/net/xfrm/xfrm_user.c | |||
@@ -1350,6 +1350,9 @@ static struct xfrm_policy *xfrm_compile_policy(u16 family, int opt, | |||
1350 | if (nr > XFRM_MAX_DEPTH) | 1350 | if (nr > XFRM_MAX_DEPTH) |
1351 | return NULL; | 1351 | return NULL; |
1352 | 1352 | ||
1353 | if (p->dir > XFRM_POLICY_OUT) | ||
1354 | return NULL; | ||
1355 | |||
1353 | xp = xfrm_policy_alloc(GFP_KERNEL); | 1356 | xp = xfrm_policy_alloc(GFP_KERNEL); |
1354 | if (xp == NULL) { | 1357 | if (xp == NULL) { |
1355 | *dir = -ENOBUFS; | 1358 | *dir = -ENOBUFS; |