diff options
Diffstat (limited to 'net')
48 files changed, 816 insertions, 578 deletions
diff --git a/net/8021q/vlan.c b/net/8021q/vlan.c index c0c7bb8e9f07..bd93c45778d4 100644 --- a/net/8021q/vlan.c +++ b/net/8021q/vlan.c | |||
@@ -117,8 +117,7 @@ static void __exit vlan_cleanup_devices(void) | |||
117 | struct net_device *dev, *nxt; | 117 | struct net_device *dev, *nxt; |
118 | 118 | ||
119 | rtnl_lock(); | 119 | rtnl_lock(); |
120 | for (dev = dev_base; dev; dev = nxt) { | 120 | for_each_netdev_safe(dev, nxt) { |
121 | nxt = dev->next; | ||
122 | if (dev->priv_flags & IFF_802_1Q_VLAN) { | 121 | if (dev->priv_flags & IFF_802_1Q_VLAN) { |
123 | unregister_vlan_dev(VLAN_DEV_INFO(dev)->real_dev, | 122 | unregister_vlan_dev(VLAN_DEV_INFO(dev)->real_dev, |
124 | VLAN_DEV_INFO(dev)->vlan_id); | 123 | VLAN_DEV_INFO(dev)->vlan_id); |
diff --git a/net/8021q/vlanproc.c b/net/8021q/vlanproc.c index 5e24f72602a1..d216a64421cd 100644 --- a/net/8021q/vlanproc.c +++ b/net/8021q/vlanproc.c | |||
@@ -237,13 +237,9 @@ int vlan_proc_rem_dev(struct net_device *vlandev) | |||
237 | * The following few functions build the content of /proc/net/vlan/config | 237 | * The following few functions build the content of /proc/net/vlan/config |
238 | */ | 238 | */ |
239 | 239 | ||
240 | /* starting at dev, find a VLAN device */ | 240 | static inline int is_vlan_dev(struct net_device *dev) |
241 | static struct net_device *vlan_skip(struct net_device *dev) | ||
242 | { | 241 | { |
243 | while (dev && !(dev->priv_flags & IFF_802_1Q_VLAN)) | 242 | return dev->priv_flags & IFF_802_1Q_VLAN; |
244 | dev = dev->next; | ||
245 | |||
246 | return dev; | ||
247 | } | 243 | } |
248 | 244 | ||
249 | /* start read of /proc/net/vlan/config */ | 245 | /* start read of /proc/net/vlan/config */ |
@@ -257,19 +253,35 @@ static void *vlan_seq_start(struct seq_file *seq, loff_t *pos) | |||
257 | if (*pos == 0) | 253 | if (*pos == 0) |
258 | return SEQ_START_TOKEN; | 254 | return SEQ_START_TOKEN; |
259 | 255 | ||
260 | for (dev = vlan_skip(dev_base); dev && i < *pos; | 256 | for_each_netdev(dev) { |
261 | dev = vlan_skip(dev->next), ++i); | 257 | if (!is_vlan_dev(dev)) |
258 | continue; | ||
259 | |||
260 | if (i++ == *pos) | ||
261 | return dev; | ||
262 | } | ||
262 | 263 | ||
263 | return (i == *pos) ? dev : NULL; | 264 | return NULL; |
264 | } | 265 | } |
265 | 266 | ||
266 | static void *vlan_seq_next(struct seq_file *seq, void *v, loff_t *pos) | 267 | static void *vlan_seq_next(struct seq_file *seq, void *v, loff_t *pos) |
267 | { | 268 | { |
269 | struct net_device *dev; | ||
270 | |||
268 | ++*pos; | 271 | ++*pos; |
269 | 272 | ||
270 | return vlan_skip((v == SEQ_START_TOKEN) | 273 | dev = (struct net_device *)v; |
271 | ? dev_base | 274 | if (v == SEQ_START_TOKEN) |
272 | : ((struct net_device *)v)->next); | 275 | dev = net_device_entry(&dev_base_head); |
276 | |||
277 | for_each_netdev_continue(dev) { | ||
278 | if (!is_vlan_dev(dev)) | ||
279 | continue; | ||
280 | |||
281 | return dev; | ||
282 | } | ||
283 | |||
284 | return NULL; | ||
273 | } | 285 | } |
274 | 286 | ||
275 | static void vlan_seq_stop(struct seq_file *seq, void *v) | 287 | static void vlan_seq_stop(struct seq_file *seq, void *v) |
diff --git a/net/bridge/br_if.c b/net/bridge/br_if.c index 690573bbf012..849deaf14108 100644 --- a/net/bridge/br_if.c +++ b/net/bridge/br_if.c | |||
@@ -475,11 +475,9 @@ void __exit br_cleanup_bridges(void) | |||
475 | struct net_device *dev, *nxt; | 475 | struct net_device *dev, *nxt; |
476 | 476 | ||
477 | rtnl_lock(); | 477 | rtnl_lock(); |
478 | for (dev = dev_base; dev; dev = nxt) { | 478 | for_each_netdev_safe(dev, nxt) |
479 | nxt = dev->next; | ||
480 | if (dev->priv_flags & IFF_EBRIDGE) | 479 | if (dev->priv_flags & IFF_EBRIDGE) |
481 | del_br(dev->priv); | 480 | del_br(dev->priv); |
482 | } | ||
483 | rtnl_unlock(); | 481 | rtnl_unlock(); |
484 | 482 | ||
485 | } | 483 | } |
diff --git a/net/bridge/br_ioctl.c b/net/bridge/br_ioctl.c index eda0fbfc923a..bb15e9e259b1 100644 --- a/net/bridge/br_ioctl.c +++ b/net/bridge/br_ioctl.c | |||
@@ -27,7 +27,9 @@ static int get_bridge_ifindices(int *indices, int num) | |||
27 | struct net_device *dev; | 27 | struct net_device *dev; |
28 | int i = 0; | 28 | int i = 0; |
29 | 29 | ||
30 | for (dev = dev_base; dev && i < num; dev = dev->next) { | 30 | for_each_netdev(dev) { |
31 | if (i >= num) | ||
32 | break; | ||
31 | if (dev->priv_flags & IFF_EBRIDGE) | 33 | if (dev->priv_flags & IFF_EBRIDGE) |
32 | indices[i++] = dev->ifindex; | 34 | indices[i++] = dev->ifindex; |
33 | } | 35 | } |
diff --git a/net/bridge/br_netfilter.c b/net/bridge/br_netfilter.c index 9b2986b182ba..fa779874b9dd 100644 --- a/net/bridge/br_netfilter.c +++ b/net/bridge/br_netfilter.c | |||
@@ -142,14 +142,33 @@ static inline struct nf_bridge_info *nf_bridge_alloc(struct sk_buff *skb) | |||
142 | return skb->nf_bridge; | 142 | return skb->nf_bridge; |
143 | } | 143 | } |
144 | 144 | ||
145 | static inline void nf_bridge_save_header(struct sk_buff *skb) | 145 | static inline void nf_bridge_push_encap_header(struct sk_buff *skb) |
146 | { | ||
147 | unsigned int len = nf_bridge_encap_header_len(skb); | ||
148 | |||
149 | skb_push(skb, len); | ||
150 | skb->network_header -= len; | ||
151 | } | ||
152 | |||
153 | static inline void nf_bridge_pull_encap_header(struct sk_buff *skb) | ||
146 | { | 154 | { |
147 | int header_size = ETH_HLEN; | 155 | unsigned int len = nf_bridge_encap_header_len(skb); |
156 | |||
157 | skb_pull(skb, len); | ||
158 | skb->network_header += len; | ||
159 | } | ||
148 | 160 | ||
149 | if (skb->protocol == htons(ETH_P_8021Q)) | 161 | static inline void nf_bridge_pull_encap_header_rcsum(struct sk_buff *skb) |
150 | header_size += VLAN_HLEN; | 162 | { |
151 | else if (skb->protocol == htons(ETH_P_PPP_SES)) | 163 | unsigned int len = nf_bridge_encap_header_len(skb); |
152 | header_size += PPPOE_SES_HLEN; | 164 | |
165 | skb_pull_rcsum(skb, len); | ||
166 | skb->network_header += len; | ||
167 | } | ||
168 | |||
169 | static inline void nf_bridge_save_header(struct sk_buff *skb) | ||
170 | { | ||
171 | int header_size = ETH_HLEN + nf_bridge_encap_header_len(skb); | ||
153 | 172 | ||
154 | skb_copy_from_linear_data_offset(skb, -header_size, | 173 | skb_copy_from_linear_data_offset(skb, -header_size, |
155 | skb->nf_bridge->data, header_size); | 174 | skb->nf_bridge->data, header_size); |
@@ -162,12 +181,7 @@ static inline void nf_bridge_save_header(struct sk_buff *skb) | |||
162 | int nf_bridge_copy_header(struct sk_buff *skb) | 181 | int nf_bridge_copy_header(struct sk_buff *skb) |
163 | { | 182 | { |
164 | int err; | 183 | int err; |
165 | int header_size = ETH_HLEN; | 184 | int header_size = ETH_HLEN + nf_bridge_encap_header_len(skb); |
166 | |||
167 | if (skb->protocol == htons(ETH_P_8021Q)) | ||
168 | header_size += VLAN_HLEN; | ||
169 | else if (skb->protocol == htons(ETH_P_PPP_SES)) | ||
170 | header_size += PPPOE_SES_HLEN; | ||
171 | 185 | ||
172 | err = skb_cow(skb, header_size); | 186 | err = skb_cow(skb, header_size); |
173 | if (err) | 187 | if (err) |
@@ -175,11 +189,7 @@ int nf_bridge_copy_header(struct sk_buff *skb) | |||
175 | 189 | ||
176 | skb_copy_to_linear_data_offset(skb, -header_size, | 190 | skb_copy_to_linear_data_offset(skb, -header_size, |
177 | skb->nf_bridge->data, header_size); | 191 | skb->nf_bridge->data, header_size); |
178 | 192 | __skb_push(skb, nf_bridge_encap_header_len(skb)); | |
179 | if (skb->protocol == htons(ETH_P_8021Q)) | ||
180 | __skb_push(skb, VLAN_HLEN); | ||
181 | else if (skb->protocol == htons(ETH_P_PPP_SES)) | ||
182 | __skb_push(skb, PPPOE_SES_HLEN); | ||
183 | return 0; | 193 | return 0; |
184 | } | 194 | } |
185 | 195 | ||
@@ -200,13 +210,7 @@ static int br_nf_pre_routing_finish_ipv6(struct sk_buff *skb) | |||
200 | dst_hold(skb->dst); | 210 | dst_hold(skb->dst); |
201 | 211 | ||
202 | skb->dev = nf_bridge->physindev; | 212 | skb->dev = nf_bridge->physindev; |
203 | if (skb->protocol == htons(ETH_P_8021Q)) { | 213 | nf_bridge_push_encap_header(skb); |
204 | skb_push(skb, VLAN_HLEN); | ||
205 | skb->network_header -= VLAN_HLEN; | ||
206 | } else if (skb->protocol == htons(ETH_P_PPP_SES)) { | ||
207 | skb_push(skb, PPPOE_SES_HLEN); | ||
208 | skb->network_header -= PPPOE_SES_HLEN; | ||
209 | } | ||
210 | NF_HOOK_THRESH(PF_BRIDGE, NF_BR_PRE_ROUTING, skb, skb->dev, NULL, | 214 | NF_HOOK_THRESH(PF_BRIDGE, NF_BR_PRE_ROUTING, skb, skb->dev, NULL, |
211 | br_handle_frame_finish, 1); | 215 | br_handle_frame_finish, 1); |
212 | 216 | ||
@@ -284,13 +288,7 @@ static int br_nf_pre_routing_finish_bridge(struct sk_buff *skb) | |||
284 | if (!skb->dev) | 288 | if (!skb->dev) |
285 | kfree_skb(skb); | 289 | kfree_skb(skb); |
286 | else { | 290 | else { |
287 | if (skb->protocol == htons(ETH_P_8021Q)) { | 291 | nf_bridge_pull_encap_header(skb); |
288 | skb_pull(skb, VLAN_HLEN); | ||
289 | skb->network_header += VLAN_HLEN; | ||
290 | } else if (skb->protocol == htons(ETH_P_PPP_SES)) { | ||
291 | skb_pull(skb, PPPOE_SES_HLEN); | ||
292 | skb->network_header += PPPOE_SES_HLEN; | ||
293 | } | ||
294 | skb->dst->output(skb); | 292 | skb->dst->output(skb); |
295 | } | 293 | } |
296 | return 0; | 294 | return 0; |
@@ -356,15 +354,7 @@ bridged_dnat: | |||
356 | * bridged frame */ | 354 | * bridged frame */ |
357 | nf_bridge->mask |= BRNF_BRIDGED_DNAT; | 355 | nf_bridge->mask |= BRNF_BRIDGED_DNAT; |
358 | skb->dev = nf_bridge->physindev; | 356 | skb->dev = nf_bridge->physindev; |
359 | if (skb->protocol == | 357 | nf_bridge_push_encap_header(skb); |
360 | htons(ETH_P_8021Q)) { | ||
361 | skb_push(skb, VLAN_HLEN); | ||
362 | skb->network_header -= VLAN_HLEN; | ||
363 | } else if(skb->protocol == | ||
364 | htons(ETH_P_PPP_SES)) { | ||
365 | skb_push(skb, PPPOE_SES_HLEN); | ||
366 | skb->network_header -= PPPOE_SES_HLEN; | ||
367 | } | ||
368 | NF_HOOK_THRESH(PF_BRIDGE, NF_BR_PRE_ROUTING, | 358 | NF_HOOK_THRESH(PF_BRIDGE, NF_BR_PRE_ROUTING, |
369 | skb, skb->dev, NULL, | 359 | skb, skb->dev, NULL, |
370 | br_nf_pre_routing_finish_bridge, | 360 | br_nf_pre_routing_finish_bridge, |
@@ -380,13 +370,7 @@ bridged_dnat: | |||
380 | } | 370 | } |
381 | 371 | ||
382 | skb->dev = nf_bridge->physindev; | 372 | skb->dev = nf_bridge->physindev; |
383 | if (skb->protocol == htons(ETH_P_8021Q)) { | 373 | nf_bridge_push_encap_header(skb); |
384 | skb_push(skb, VLAN_HLEN); | ||
385 | skb->network_header -= VLAN_HLEN; | ||
386 | } else if (skb->protocol == htons(ETH_P_PPP_SES)) { | ||
387 | skb_push(skb, PPPOE_SES_HLEN); | ||
388 | skb->network_header -= PPPOE_SES_HLEN; | ||
389 | } | ||
390 | NF_HOOK_THRESH(PF_BRIDGE, NF_BR_PRE_ROUTING, skb, skb->dev, NULL, | 374 | NF_HOOK_THRESH(PF_BRIDGE, NF_BR_PRE_ROUTING, skb, skb->dev, NULL, |
391 | br_handle_frame_finish, 1); | 375 | br_handle_frame_finish, 1); |
392 | 376 | ||
@@ -536,14 +520,7 @@ static unsigned int br_nf_pre_routing(unsigned int hook, struct sk_buff **pskb, | |||
536 | #endif | 520 | #endif |
537 | if ((skb = skb_share_check(*pskb, GFP_ATOMIC)) == NULL) | 521 | if ((skb = skb_share_check(*pskb, GFP_ATOMIC)) == NULL) |
538 | goto out; | 522 | goto out; |
539 | 523 | nf_bridge_pull_encap_header_rcsum(skb); | |
540 | if (skb->protocol == htons(ETH_P_8021Q)) { | ||
541 | skb_pull_rcsum(skb, VLAN_HLEN); | ||
542 | skb->network_header += VLAN_HLEN; | ||
543 | } else if (skb->protocol == htons(ETH_P_PPP_SES)) { | ||
544 | skb_pull_rcsum(skb, PPPOE_SES_HLEN); | ||
545 | skb->network_header += PPPOE_SES_HLEN; | ||
546 | } | ||
547 | return br_nf_pre_routing_ipv6(hook, skb, in, out, okfn); | 524 | return br_nf_pre_routing_ipv6(hook, skb, in, out, okfn); |
548 | } | 525 | } |
549 | #ifdef CONFIG_SYSCTL | 526 | #ifdef CONFIG_SYSCTL |
@@ -557,14 +534,7 @@ static unsigned int br_nf_pre_routing(unsigned int hook, struct sk_buff **pskb, | |||
557 | 534 | ||
558 | if ((skb = skb_share_check(*pskb, GFP_ATOMIC)) == NULL) | 535 | if ((skb = skb_share_check(*pskb, GFP_ATOMIC)) == NULL) |
559 | goto out; | 536 | goto out; |
560 | 537 | nf_bridge_pull_encap_header_rcsum(skb); | |
561 | if (skb->protocol == htons(ETH_P_8021Q)) { | ||
562 | skb_pull_rcsum(skb, VLAN_HLEN); | ||
563 | skb->network_header += VLAN_HLEN; | ||
564 | } else if (skb->protocol == htons(ETH_P_PPP_SES)) { | ||
565 | skb_pull_rcsum(skb, PPPOE_SES_HLEN); | ||
566 | skb->network_header += PPPOE_SES_HLEN; | ||
567 | } | ||
568 | 538 | ||
569 | if (!pskb_may_pull(skb, sizeof(struct iphdr))) | 539 | if (!pskb_may_pull(skb, sizeof(struct iphdr))) |
570 | goto inhdr_error; | 540 | goto inhdr_error; |
@@ -642,13 +612,7 @@ static int br_nf_forward_finish(struct sk_buff *skb) | |||
642 | } else { | 612 | } else { |
643 | in = *((struct net_device **)(skb->cb)); | 613 | in = *((struct net_device **)(skb->cb)); |
644 | } | 614 | } |
645 | if (skb->protocol == htons(ETH_P_8021Q)) { | 615 | nf_bridge_push_encap_header(skb); |
646 | skb_push(skb, VLAN_HLEN); | ||
647 | skb->network_header -= VLAN_HLEN; | ||
648 | } else if (skb->protocol == htons(ETH_P_PPP_SES)) { | ||
649 | skb_push(skb, PPPOE_SES_HLEN); | ||
650 | skb->network_header -= PPPOE_SES_HLEN; | ||
651 | } | ||
652 | NF_HOOK_THRESH(PF_BRIDGE, NF_BR_FORWARD, skb, in, | 616 | NF_HOOK_THRESH(PF_BRIDGE, NF_BR_FORWARD, skb, in, |
653 | skb->dev, br_forward_finish, 1); | 617 | skb->dev, br_forward_finish, 1); |
654 | return 0; | 618 | return 0; |
@@ -682,13 +646,7 @@ static unsigned int br_nf_forward_ip(unsigned int hook, struct sk_buff **pskb, | |||
682 | else | 646 | else |
683 | pf = PF_INET6; | 647 | pf = PF_INET6; |
684 | 648 | ||
685 | if (skb->protocol == htons(ETH_P_8021Q)) { | 649 | nf_bridge_pull_encap_header(*pskb); |
686 | skb_pull(*pskb, VLAN_HLEN); | ||
687 | (*pskb)->network_header += VLAN_HLEN; | ||
688 | } else if (skb->protocol == htons(ETH_P_PPP_SES)) { | ||
689 | skb_pull(*pskb, PPPOE_SES_HLEN); | ||
690 | (*pskb)->network_header += PPPOE_SES_HLEN; | ||
691 | } | ||
692 | 650 | ||
693 | nf_bridge = skb->nf_bridge; | 651 | nf_bridge = skb->nf_bridge; |
694 | if (skb->pkt_type == PACKET_OTHERHOST) { | 652 | if (skb->pkt_type == PACKET_OTHERHOST) { |
@@ -722,15 +680,12 @@ static unsigned int br_nf_forward_arp(unsigned int hook, struct sk_buff **pskb, | |||
722 | if (skb->protocol != htons(ETH_P_ARP)) { | 680 | if (skb->protocol != htons(ETH_P_ARP)) { |
723 | if (!IS_VLAN_ARP(skb)) | 681 | if (!IS_VLAN_ARP(skb)) |
724 | return NF_ACCEPT; | 682 | return NF_ACCEPT; |
725 | skb_pull(*pskb, VLAN_HLEN); | 683 | nf_bridge_pull_encap_header(*pskb); |
726 | (*pskb)->network_header += VLAN_HLEN; | ||
727 | } | 684 | } |
728 | 685 | ||
729 | if (arp_hdr(skb)->ar_pln != 4) { | 686 | if (arp_hdr(skb)->ar_pln != 4) { |
730 | if (IS_VLAN_ARP(skb)) { | 687 | if (IS_VLAN_ARP(skb)) |
731 | skb_push(*pskb, VLAN_HLEN); | 688 | nf_bridge_push_encap_header(*pskb); |
732 | (*pskb)->network_header -= VLAN_HLEN; | ||
733 | } | ||
734 | return NF_ACCEPT; | 689 | return NF_ACCEPT; |
735 | } | 690 | } |
736 | *d = (struct net_device *)in; | 691 | *d = (struct net_device *)in; |
@@ -777,13 +732,7 @@ static unsigned int br_nf_local_out(unsigned int hook, struct sk_buff **pskb, | |||
777 | skb->pkt_type = PACKET_OTHERHOST; | 732 | skb->pkt_type = PACKET_OTHERHOST; |
778 | nf_bridge->mask ^= BRNF_PKT_TYPE; | 733 | nf_bridge->mask ^= BRNF_PKT_TYPE; |
779 | } | 734 | } |
780 | if (skb->protocol == htons(ETH_P_8021Q)) { | 735 | nf_bridge_push_encap_header(skb); |
781 | skb_push(skb, VLAN_HLEN); | ||
782 | skb->network_header -= VLAN_HLEN; | ||
783 | } else if (skb->protocol == htons(ETH_P_PPP_SES)) { | ||
784 | skb_push(skb, PPPOE_SES_HLEN); | ||
785 | skb->network_header -= PPPOE_SES_HLEN; | ||
786 | } | ||
787 | 736 | ||
788 | NF_HOOK(PF_BRIDGE, NF_BR_FORWARD, skb, realindev, skb->dev, | 737 | NF_HOOK(PF_BRIDGE, NF_BR_FORWARD, skb, realindev, skb->dev, |
789 | br_forward_finish); | 738 | br_forward_finish); |
@@ -848,14 +797,7 @@ static unsigned int br_nf_post_routing(unsigned int hook, struct sk_buff **pskb, | |||
848 | nf_bridge->mask |= BRNF_PKT_TYPE; | 797 | nf_bridge->mask |= BRNF_PKT_TYPE; |
849 | } | 798 | } |
850 | 799 | ||
851 | if (skb->protocol == htons(ETH_P_8021Q)) { | 800 | nf_bridge_pull_encap_header(skb); |
852 | skb_pull(skb, VLAN_HLEN); | ||
853 | skb->network_header += VLAN_HLEN; | ||
854 | } else if (skb->protocol == htons(ETH_P_PPP_SES)) { | ||
855 | skb_pull(skb, PPPOE_SES_HLEN); | ||
856 | skb->network_header += PPPOE_SES_HLEN; | ||
857 | } | ||
858 | |||
859 | nf_bridge_save_header(skb); | 801 | nf_bridge_save_header(skb); |
860 | 802 | ||
861 | #if defined(CONFIG_VLAN_8021Q) || defined(CONFIG_VLAN_8021Q_MODULE) | 803 | #if defined(CONFIG_VLAN_8021Q) || defined(CONFIG_VLAN_8021Q_MODULE) |
diff --git a/net/bridge/br_netlink.c b/net/bridge/br_netlink.c index 35facc0c11c2..0fcf6f073064 100644 --- a/net/bridge/br_netlink.c +++ b/net/bridge/br_netlink.c | |||
@@ -109,7 +109,8 @@ static int br_dump_ifinfo(struct sk_buff *skb, struct netlink_callback *cb) | |||
109 | struct net_device *dev; | 109 | struct net_device *dev; |
110 | int idx; | 110 | int idx; |
111 | 111 | ||
112 | for (dev = dev_base, idx = 0; dev; dev = dev->next) { | 112 | idx = 0; |
113 | for_each_netdev(dev) { | ||
113 | /* not a bridge port */ | 114 | /* not a bridge port */ |
114 | if (dev->br_port == NULL || idx < cb->args[0]) | 115 | if (dev->br_port == NULL || idx < cb->args[0]) |
115 | goto skip; | 116 | goto skip; |
diff --git a/net/core/dev.c b/net/core/dev.c index eb999003bbb7..f27d4ab181e6 100644 --- a/net/core/dev.c +++ b/net/core/dev.c | |||
@@ -156,13 +156,13 @@ static spinlock_t net_dma_event_lock; | |||
156 | #endif | 156 | #endif |
157 | 157 | ||
158 | /* | 158 | /* |
159 | * The @dev_base list is protected by @dev_base_lock and the rtnl | 159 | * The @dev_base_head list is protected by @dev_base_lock and the rtnl |
160 | * semaphore. | 160 | * semaphore. |
161 | * | 161 | * |
162 | * Pure readers hold dev_base_lock for reading. | 162 | * Pure readers hold dev_base_lock for reading. |
163 | * | 163 | * |
164 | * Writers must hold the rtnl semaphore while they loop through the | 164 | * Writers must hold the rtnl semaphore while they loop through the |
165 | * dev_base list, and hold dev_base_lock for writing when they do the | 165 | * dev_base_head list, and hold dev_base_lock for writing when they do the |
166 | * actual updates. This allows pure readers to access the list even | 166 | * actual updates. This allows pure readers to access the list even |
167 | * while a writer is preparing to update it. | 167 | * while a writer is preparing to update it. |
168 | * | 168 | * |
@@ -174,11 +174,10 @@ static spinlock_t net_dma_event_lock; | |||
174 | * unregister_netdevice(), which must be called with the rtnl | 174 | * unregister_netdevice(), which must be called with the rtnl |
175 | * semaphore held. | 175 | * semaphore held. |
176 | */ | 176 | */ |
177 | struct net_device *dev_base; | 177 | LIST_HEAD(dev_base_head); |
178 | static struct net_device **dev_tail = &dev_base; | ||
179 | DEFINE_RWLOCK(dev_base_lock); | 178 | DEFINE_RWLOCK(dev_base_lock); |
180 | 179 | ||
181 | EXPORT_SYMBOL(dev_base); | 180 | EXPORT_SYMBOL(dev_base_head); |
182 | EXPORT_SYMBOL(dev_base_lock); | 181 | EXPORT_SYMBOL(dev_base_lock); |
183 | 182 | ||
184 | #define NETDEV_HASHBITS 8 | 183 | #define NETDEV_HASHBITS 8 |
@@ -567,26 +566,38 @@ struct net_device *dev_getbyhwaddr(unsigned short type, char *ha) | |||
567 | 566 | ||
568 | ASSERT_RTNL(); | 567 | ASSERT_RTNL(); |
569 | 568 | ||
570 | for (dev = dev_base; dev; dev = dev->next) | 569 | for_each_netdev(dev) |
571 | if (dev->type == type && | 570 | if (dev->type == type && |
572 | !memcmp(dev->dev_addr, ha, dev->addr_len)) | 571 | !memcmp(dev->dev_addr, ha, dev->addr_len)) |
573 | break; | 572 | return dev; |
574 | return dev; | 573 | |
574 | return NULL; | ||
575 | } | 575 | } |
576 | 576 | ||
577 | EXPORT_SYMBOL(dev_getbyhwaddr); | 577 | EXPORT_SYMBOL(dev_getbyhwaddr); |
578 | 578 | ||
579 | struct net_device *__dev_getfirstbyhwtype(unsigned short type) | ||
580 | { | ||
581 | struct net_device *dev; | ||
582 | |||
583 | ASSERT_RTNL(); | ||
584 | for_each_netdev(dev) | ||
585 | if (dev->type == type) | ||
586 | return dev; | ||
587 | |||
588 | return NULL; | ||
589 | } | ||
590 | |||
591 | EXPORT_SYMBOL(__dev_getfirstbyhwtype); | ||
592 | |||
579 | struct net_device *dev_getfirstbyhwtype(unsigned short type) | 593 | struct net_device *dev_getfirstbyhwtype(unsigned short type) |
580 | { | 594 | { |
581 | struct net_device *dev; | 595 | struct net_device *dev; |
582 | 596 | ||
583 | rtnl_lock(); | 597 | rtnl_lock(); |
584 | for (dev = dev_base; dev; dev = dev->next) { | 598 | dev = __dev_getfirstbyhwtype(type); |
585 | if (dev->type == type) { | 599 | if (dev) |
586 | dev_hold(dev); | 600 | dev_hold(dev); |
587 | break; | ||
588 | } | ||
589 | } | ||
590 | rtnl_unlock(); | 601 | rtnl_unlock(); |
591 | return dev; | 602 | return dev; |
592 | } | 603 | } |
@@ -606,17 +617,19 @@ EXPORT_SYMBOL(dev_getfirstbyhwtype); | |||
606 | 617 | ||
607 | struct net_device * dev_get_by_flags(unsigned short if_flags, unsigned short mask) | 618 | struct net_device * dev_get_by_flags(unsigned short if_flags, unsigned short mask) |
608 | { | 619 | { |
609 | struct net_device *dev; | 620 | struct net_device *dev, *ret; |
610 | 621 | ||
622 | ret = NULL; | ||
611 | read_lock(&dev_base_lock); | 623 | read_lock(&dev_base_lock); |
612 | for (dev = dev_base; dev != NULL; dev = dev->next) { | 624 | for_each_netdev(dev) { |
613 | if (((dev->flags ^ if_flags) & mask) == 0) { | 625 | if (((dev->flags ^ if_flags) & mask) == 0) { |
614 | dev_hold(dev); | 626 | dev_hold(dev); |
627 | ret = dev; | ||
615 | break; | 628 | break; |
616 | } | 629 | } |
617 | } | 630 | } |
618 | read_unlock(&dev_base_lock); | 631 | read_unlock(&dev_base_lock); |
619 | return dev; | 632 | return ret; |
620 | } | 633 | } |
621 | 634 | ||
622 | /** | 635 | /** |
@@ -682,7 +695,7 @@ int dev_alloc_name(struct net_device *dev, const char *name) | |||
682 | if (!inuse) | 695 | if (!inuse) |
683 | return -ENOMEM; | 696 | return -ENOMEM; |
684 | 697 | ||
685 | for (d = dev_base; d; d = d->next) { | 698 | for_each_netdev(d) { |
686 | if (!sscanf(d->name, name, &i)) | 699 | if (!sscanf(d->name, name, &i)) |
687 | continue; | 700 | continue; |
688 | if (i < 0 || i >= max_netdevices) | 701 | if (i < 0 || i >= max_netdevices) |
@@ -964,7 +977,7 @@ int register_netdevice_notifier(struct notifier_block *nb) | |||
964 | rtnl_lock(); | 977 | rtnl_lock(); |
965 | err = raw_notifier_chain_register(&netdev_chain, nb); | 978 | err = raw_notifier_chain_register(&netdev_chain, nb); |
966 | if (!err) { | 979 | if (!err) { |
967 | for (dev = dev_base; dev; dev = dev->next) { | 980 | for_each_netdev(dev) { |
968 | nb->notifier_call(nb, NETDEV_REGISTER, dev); | 981 | nb->notifier_call(nb, NETDEV_REGISTER, dev); |
969 | 982 | ||
970 | if (dev->flags & IFF_UP) | 983 | if (dev->flags & IFF_UP) |
@@ -2038,7 +2051,7 @@ static int dev_ifconf(char __user *arg) | |||
2038 | */ | 2051 | */ |
2039 | 2052 | ||
2040 | total = 0; | 2053 | total = 0; |
2041 | for (dev = dev_base; dev; dev = dev->next) { | 2054 | for_each_netdev(dev) { |
2042 | for (i = 0; i < NPROTO; i++) { | 2055 | for (i = 0; i < NPROTO; i++) { |
2043 | if (gifconf_list[i]) { | 2056 | if (gifconf_list[i]) { |
2044 | int done; | 2057 | int done; |
@@ -2070,26 +2083,28 @@ static int dev_ifconf(char __user *arg) | |||
2070 | * This is invoked by the /proc filesystem handler to display a device | 2083 | * This is invoked by the /proc filesystem handler to display a device |
2071 | * in detail. | 2084 | * in detail. |
2072 | */ | 2085 | */ |
2073 | static struct net_device *dev_get_idx(loff_t pos) | 2086 | void *dev_seq_start(struct seq_file *seq, loff_t *pos) |
2074 | { | 2087 | { |
2088 | loff_t off; | ||
2075 | struct net_device *dev; | 2089 | struct net_device *dev; |
2076 | loff_t i; | ||
2077 | 2090 | ||
2078 | for (i = 0, dev = dev_base; dev && i < pos; ++i, dev = dev->next); | 2091 | read_lock(&dev_base_lock); |
2092 | if (!*pos) | ||
2093 | return SEQ_START_TOKEN; | ||
2079 | 2094 | ||
2080 | return i == pos ? dev : NULL; | 2095 | off = 1; |
2081 | } | 2096 | for_each_netdev(dev) |
2097 | if (off++ == *pos) | ||
2098 | return dev; | ||
2082 | 2099 | ||
2083 | void *dev_seq_start(struct seq_file *seq, loff_t *pos) | 2100 | return NULL; |
2084 | { | ||
2085 | read_lock(&dev_base_lock); | ||
2086 | return *pos ? dev_get_idx(*pos - 1) : SEQ_START_TOKEN; | ||
2087 | } | 2101 | } |
2088 | 2102 | ||
2089 | void *dev_seq_next(struct seq_file *seq, void *v, loff_t *pos) | 2103 | void *dev_seq_next(struct seq_file *seq, void *v, loff_t *pos) |
2090 | { | 2104 | { |
2091 | ++*pos; | 2105 | ++*pos; |
2092 | return v == SEQ_START_TOKEN ? dev_base : ((struct net_device *)v)->next; | 2106 | return v == SEQ_START_TOKEN ? |
2107 | first_net_device() : next_net_device((struct net_device *)v); | ||
2093 | } | 2108 | } |
2094 | 2109 | ||
2095 | void dev_seq_stop(struct seq_file *seq, void *v) | 2110 | void dev_seq_stop(struct seq_file *seq, void *v) |
@@ -3071,11 +3086,9 @@ int register_netdevice(struct net_device *dev) | |||
3071 | 3086 | ||
3072 | set_bit(__LINK_STATE_PRESENT, &dev->state); | 3087 | set_bit(__LINK_STATE_PRESENT, &dev->state); |
3073 | 3088 | ||
3074 | dev->next = NULL; | ||
3075 | dev_init_scheduler(dev); | 3089 | dev_init_scheduler(dev); |
3076 | write_lock_bh(&dev_base_lock); | 3090 | write_lock_bh(&dev_base_lock); |
3077 | *dev_tail = dev; | 3091 | list_add_tail(&dev->dev_list, &dev_base_head); |
3078 | dev_tail = &dev->next; | ||
3079 | hlist_add_head(&dev->name_hlist, head); | 3092 | hlist_add_head(&dev->name_hlist, head); |
3080 | hlist_add_head(&dev->index_hlist, dev_index_hash(dev->ifindex)); | 3093 | hlist_add_head(&dev->index_hlist, dev_index_hash(dev->ifindex)); |
3081 | dev_hold(dev); | 3094 | dev_hold(dev); |
@@ -3349,8 +3362,6 @@ void synchronize_net(void) | |||
3349 | 3362 | ||
3350 | void unregister_netdevice(struct net_device *dev) | 3363 | void unregister_netdevice(struct net_device *dev) |
3351 | { | 3364 | { |
3352 | struct net_device *d, **dp; | ||
3353 | |||
3354 | BUG_ON(dev_boot_phase); | 3365 | BUG_ON(dev_boot_phase); |
3355 | ASSERT_RTNL(); | 3366 | ASSERT_RTNL(); |
3356 | 3367 | ||
@@ -3370,19 +3381,11 @@ void unregister_netdevice(struct net_device *dev) | |||
3370 | dev_close(dev); | 3381 | dev_close(dev); |
3371 | 3382 | ||
3372 | /* And unlink it from device chain. */ | 3383 | /* And unlink it from device chain. */ |
3373 | for (dp = &dev_base; (d = *dp) != NULL; dp = &d->next) { | 3384 | write_lock_bh(&dev_base_lock); |
3374 | if (d == dev) { | 3385 | list_del(&dev->dev_list); |
3375 | write_lock_bh(&dev_base_lock); | 3386 | hlist_del(&dev->name_hlist); |
3376 | hlist_del(&dev->name_hlist); | 3387 | hlist_del(&dev->index_hlist); |
3377 | hlist_del(&dev->index_hlist); | 3388 | write_unlock_bh(&dev_base_lock); |
3378 | if (dev_tail == &dev->next) | ||
3379 | dev_tail = dp; | ||
3380 | *dp = d->next; | ||
3381 | write_unlock_bh(&dev_base_lock); | ||
3382 | break; | ||
3383 | } | ||
3384 | } | ||
3385 | BUG_ON(!d); | ||
3386 | 3389 | ||
3387 | dev->reg_state = NETREG_UNREGISTERING; | 3390 | dev->reg_state = NETREG_UNREGISTERING; |
3388 | 3391 | ||
diff --git a/net/core/dev_mcast.c b/net/core/dev_mcast.c index 7d57bf77f3a3..5a54053386c8 100644 --- a/net/core/dev_mcast.c +++ b/net/core/dev_mcast.c | |||
@@ -223,7 +223,7 @@ static void *dev_mc_seq_start(struct seq_file *seq, loff_t *pos) | |||
223 | loff_t off = 0; | 223 | loff_t off = 0; |
224 | 224 | ||
225 | read_lock(&dev_base_lock); | 225 | read_lock(&dev_base_lock); |
226 | for (dev = dev_base; dev; dev = dev->next) { | 226 | for_each_netdev(dev) { |
227 | if (off++ == *pos) | 227 | if (off++ == *pos) |
228 | return dev; | 228 | return dev; |
229 | } | 229 | } |
@@ -232,9 +232,8 @@ static void *dev_mc_seq_start(struct seq_file *seq, loff_t *pos) | |||
232 | 232 | ||
233 | static void *dev_mc_seq_next(struct seq_file *seq, void *v, loff_t *pos) | 233 | static void *dev_mc_seq_next(struct seq_file *seq, void *v, loff_t *pos) |
234 | { | 234 | { |
235 | struct net_device *dev = v; | ||
236 | ++*pos; | 235 | ++*pos; |
237 | return dev->next; | 236 | return next_net_device((struct net_device *)v); |
238 | } | 237 | } |
239 | 238 | ||
240 | static void dev_mc_seq_stop(struct seq_file *seq, void *v) | 239 | static void dev_mc_seq_stop(struct seq_file *seq, void *v) |
diff --git a/net/core/rtnetlink.c b/net/core/rtnetlink.c index cec111109155..8c971a2efe2a 100644 --- a/net/core/rtnetlink.c +++ b/net/core/rtnetlink.c | |||
@@ -539,13 +539,16 @@ static int rtnl_dump_ifinfo(struct sk_buff *skb, struct netlink_callback *cb) | |||
539 | int s_idx = cb->args[0]; | 539 | int s_idx = cb->args[0]; |
540 | struct net_device *dev; | 540 | struct net_device *dev; |
541 | 541 | ||
542 | for (dev=dev_base, idx=0; dev; dev = dev->next, idx++) { | 542 | idx = 0; |
543 | for_each_netdev(dev) { | ||
543 | if (idx < s_idx) | 544 | if (idx < s_idx) |
544 | continue; | 545 | goto cont; |
545 | if (rtnl_fill_ifinfo(skb, dev, NULL, 0, RTM_NEWLINK, | 546 | if (rtnl_fill_ifinfo(skb, dev, NULL, 0, RTM_NEWLINK, |
546 | NETLINK_CB(cb->skb).pid, | 547 | NETLINK_CB(cb->skb).pid, |
547 | cb->nlh->nlmsg_seq, 0, NLM_F_MULTI) <= 0) | 548 | cb->nlh->nlmsg_seq, 0, NLM_F_MULTI) <= 0) |
548 | break; | 549 | break; |
550 | cont: | ||
551 | idx++; | ||
549 | } | 552 | } |
550 | cb->args[0] = idx; | 553 | cb->args[0] = idx; |
551 | 554 | ||
diff --git a/net/decnet/af_decnet.c b/net/decnet/af_decnet.c index a205eaa87f52..9fbe87c93802 100644 --- a/net/decnet/af_decnet.c +++ b/net/decnet/af_decnet.c | |||
@@ -721,7 +721,7 @@ static int dn_bind(struct socket *sock, struct sockaddr *uaddr, int addr_len) | |||
721 | struct sock *sk = sock->sk; | 721 | struct sock *sk = sock->sk; |
722 | struct dn_scp *scp = DN_SK(sk); | 722 | struct dn_scp *scp = DN_SK(sk); |
723 | struct sockaddr_dn *saddr = (struct sockaddr_dn *)uaddr; | 723 | struct sockaddr_dn *saddr = (struct sockaddr_dn *)uaddr; |
724 | struct net_device *dev; | 724 | struct net_device *dev, *ldev; |
725 | int rv; | 725 | int rv; |
726 | 726 | ||
727 | if (addr_len != sizeof(struct sockaddr_dn)) | 727 | if (addr_len != sizeof(struct sockaddr_dn)) |
@@ -746,14 +746,17 @@ static int dn_bind(struct socket *sock, struct sockaddr *uaddr, int addr_len) | |||
746 | if (!(saddr->sdn_flags & SDF_WILD)) { | 746 | if (!(saddr->sdn_flags & SDF_WILD)) { |
747 | if (dn_ntohs(saddr->sdn_nodeaddrl)) { | 747 | if (dn_ntohs(saddr->sdn_nodeaddrl)) { |
748 | read_lock(&dev_base_lock); | 748 | read_lock(&dev_base_lock); |
749 | for(dev = dev_base; dev; dev = dev->next) { | 749 | ldev = NULL; |
750 | for_each_netdev(dev) { | ||
750 | if (!dev->dn_ptr) | 751 | if (!dev->dn_ptr) |
751 | continue; | 752 | continue; |
752 | if (dn_dev_islocal(dev, dn_saddr2dn(saddr))) | 753 | if (dn_dev_islocal(dev, dn_saddr2dn(saddr))) { |
754 | ldev = dev; | ||
753 | break; | 755 | break; |
756 | } | ||
754 | } | 757 | } |
755 | read_unlock(&dev_base_lock); | 758 | read_unlock(&dev_base_lock); |
756 | if (dev == NULL) | 759 | if (ldev == NULL) |
757 | return -EADDRNOTAVAIL; | 760 | return -EADDRNOTAVAIL; |
758 | } | 761 | } |
759 | } | 762 | } |
diff --git a/net/decnet/dn_dev.c b/net/decnet/dn_dev.c index 5c2a9951b638..764a56a13e38 100644 --- a/net/decnet/dn_dev.c +++ b/net/decnet/dn_dev.c | |||
@@ -799,9 +799,10 @@ static int dn_nl_dump_ifaddr(struct sk_buff *skb, struct netlink_callback *cb) | |||
799 | skip_ndevs = cb->args[0]; | 799 | skip_ndevs = cb->args[0]; |
800 | skip_naddr = cb->args[1]; | 800 | skip_naddr = cb->args[1]; |
801 | 801 | ||
802 | for (dev = dev_base, idx = 0; dev; dev = dev->next, idx++) { | 802 | idx = 0; |
803 | for_each_netdev(dev) { | ||
803 | if (idx < skip_ndevs) | 804 | if (idx < skip_ndevs) |
804 | continue; | 805 | goto cont; |
805 | else if (idx > skip_ndevs) { | 806 | else if (idx > skip_ndevs) { |
806 | /* Only skip over addresses for first dev dumped | 807 | /* Only skip over addresses for first dev dumped |
807 | * in this iteration (idx == skip_ndevs) */ | 808 | * in this iteration (idx == skip_ndevs) */ |
@@ -809,18 +810,20 @@ static int dn_nl_dump_ifaddr(struct sk_buff *skb, struct netlink_callback *cb) | |||
809 | } | 810 | } |
810 | 811 | ||
811 | if ((dn_db = dev->dn_ptr) == NULL) | 812 | if ((dn_db = dev->dn_ptr) == NULL) |
812 | continue; | 813 | goto cont; |
813 | 814 | ||
814 | for (ifa = dn_db->ifa_list, dn_idx = 0; ifa; | 815 | for (ifa = dn_db->ifa_list, dn_idx = 0; ifa; |
815 | ifa = ifa->ifa_next, dn_idx++) { | 816 | ifa = ifa->ifa_next, dn_idx++) { |
816 | if (dn_idx < skip_naddr) | 817 | if (dn_idx < skip_naddr) |
817 | continue; | 818 | goto cont; |
818 | 819 | ||
819 | if (dn_nl_fill_ifaddr(skb, ifa, NETLINK_CB(cb->skb).pid, | 820 | if (dn_nl_fill_ifaddr(skb, ifa, NETLINK_CB(cb->skb).pid, |
820 | cb->nlh->nlmsg_seq, RTM_NEWADDR, | 821 | cb->nlh->nlmsg_seq, RTM_NEWADDR, |
821 | NLM_F_MULTI) < 0) | 822 | NLM_F_MULTI) < 0) |
822 | goto done; | 823 | goto done; |
823 | } | 824 | } |
825 | cont: | ||
826 | idx++; | ||
824 | } | 827 | } |
825 | done: | 828 | done: |
826 | cb->args[0] = idx; | 829 | cb->args[0] = idx; |
@@ -1296,7 +1299,7 @@ void dn_dev_devices_off(void) | |||
1296 | struct net_device *dev; | 1299 | struct net_device *dev; |
1297 | 1300 | ||
1298 | rtnl_lock(); | 1301 | rtnl_lock(); |
1299 | for(dev = dev_base; dev; dev = dev->next) | 1302 | for_each_netdev(dev) |
1300 | dn_dev_down(dev); | 1303 | dn_dev_down(dev); |
1301 | rtnl_unlock(); | 1304 | rtnl_unlock(); |
1302 | 1305 | ||
@@ -1307,7 +1310,7 @@ void dn_dev_devices_on(void) | |||
1307 | struct net_device *dev; | 1310 | struct net_device *dev; |
1308 | 1311 | ||
1309 | rtnl_lock(); | 1312 | rtnl_lock(); |
1310 | for(dev = dev_base; dev; dev = dev->next) { | 1313 | for_each_netdev(dev) { |
1311 | if (dev->flags & IFF_UP) | 1314 | if (dev->flags & IFF_UP) |
1312 | dn_dev_up(dev); | 1315 | dn_dev_up(dev); |
1313 | } | 1316 | } |
@@ -1325,62 +1328,56 @@ int unregister_dnaddr_notifier(struct notifier_block *nb) | |||
1325 | } | 1328 | } |
1326 | 1329 | ||
1327 | #ifdef CONFIG_PROC_FS | 1330 | #ifdef CONFIG_PROC_FS |
1328 | static inline struct net_device *dn_dev_get_next(struct seq_file *seq, struct net_device *dev) | 1331 | static inline int is_dn_dev(struct net_device *dev) |
1329 | { | 1332 | { |
1330 | do { | 1333 | return dev->dn_ptr != NULL; |
1331 | dev = dev->next; | ||
1332 | } while(dev && !dev->dn_ptr); | ||
1333 | |||
1334 | return dev; | ||
1335 | } | 1334 | } |
1336 | 1335 | ||
1337 | static struct net_device *dn_dev_get_idx(struct seq_file *seq, loff_t pos) | 1336 | static void *dn_dev_seq_start(struct seq_file *seq, loff_t *pos) |
1338 | { | 1337 | { |
1338 | int i; | ||
1339 | struct net_device *dev; | 1339 | struct net_device *dev; |
1340 | 1340 | ||
1341 | dev = dev_base; | 1341 | read_lock(&dev_base_lock); |
1342 | if (dev && !dev->dn_ptr) | ||
1343 | dev = dn_dev_get_next(seq, dev); | ||
1344 | if (pos) { | ||
1345 | while(dev && (dev = dn_dev_get_next(seq, dev))) | ||
1346 | --pos; | ||
1347 | } | ||
1348 | return dev; | ||
1349 | } | ||
1350 | 1342 | ||
1351 | static void *dn_dev_seq_start(struct seq_file *seq, loff_t *pos) | 1343 | if (*pos == 0) |
1352 | { | 1344 | return SEQ_START_TOKEN; |
1353 | if (*pos) { | 1345 | |
1354 | struct net_device *dev; | 1346 | i = 1; |
1355 | read_lock(&dev_base_lock); | 1347 | for_each_netdev(dev) { |
1356 | dev = dn_dev_get_idx(seq, *pos - 1); | 1348 | if (!is_dn_dev(dev)) |
1357 | if (dev == NULL) | 1349 | continue; |
1358 | read_unlock(&dev_base_lock); | 1350 | |
1359 | return dev; | 1351 | if (i++ == *pos) |
1352 | return dev; | ||
1360 | } | 1353 | } |
1361 | return SEQ_START_TOKEN; | 1354 | |
1355 | return NULL; | ||
1362 | } | 1356 | } |
1363 | 1357 | ||
1364 | static void *dn_dev_seq_next(struct seq_file *seq, void *v, loff_t *pos) | 1358 | static void *dn_dev_seq_next(struct seq_file *seq, void *v, loff_t *pos) |
1365 | { | 1359 | { |
1366 | struct net_device *dev = v; | 1360 | struct net_device *dev; |
1367 | loff_t one = 1; | ||
1368 | 1361 | ||
1369 | if (v == SEQ_START_TOKEN) { | ||
1370 | dev = dn_dev_seq_start(seq, &one); | ||
1371 | } else { | ||
1372 | dev = dn_dev_get_next(seq, dev); | ||
1373 | if (dev == NULL) | ||
1374 | read_unlock(&dev_base_lock); | ||
1375 | } | ||
1376 | ++*pos; | 1362 | ++*pos; |
1377 | return dev; | 1363 | |
1364 | dev = (struct net_device *)v; | ||
1365 | if (v == SEQ_START_TOKEN) | ||
1366 | dev = net_device_entry(&dev_base_head); | ||
1367 | |||
1368 | for_each_netdev_continue(dev) { | ||
1369 | if (!is_dn_dev(dev)) | ||
1370 | continue; | ||
1371 | |||
1372 | return dev; | ||
1373 | } | ||
1374 | |||
1375 | return NULL; | ||
1378 | } | 1376 | } |
1379 | 1377 | ||
1380 | static void dn_dev_seq_stop(struct seq_file *seq, void *v) | 1378 | static void dn_dev_seq_stop(struct seq_file *seq, void *v) |
1381 | { | 1379 | { |
1382 | if (v && v != SEQ_START_TOKEN) | 1380 | read_unlock(&dev_base_lock); |
1383 | read_unlock(&dev_base_lock); | ||
1384 | } | 1381 | } |
1385 | 1382 | ||
1386 | static char *dn_type2asc(char type) | 1383 | static char *dn_type2asc(char type) |
diff --git a/net/decnet/dn_fib.c b/net/decnet/dn_fib.c index 310a86268d2b..d2bc19d47950 100644 --- a/net/decnet/dn_fib.c +++ b/net/decnet/dn_fib.c | |||
@@ -602,7 +602,7 @@ static void dn_fib_del_ifaddr(struct dn_ifaddr *ifa) | |||
602 | 602 | ||
603 | /* Scan device list */ | 603 | /* Scan device list */ |
604 | read_lock(&dev_base_lock); | 604 | read_lock(&dev_base_lock); |
605 | for(dev = dev_base; dev; dev = dev->next) { | 605 | for_each_netdev(dev) { |
606 | dn_db = dev->dn_ptr; | 606 | dn_db = dev->dn_ptr; |
607 | if (dn_db == NULL) | 607 | if (dn_db == NULL) |
608 | continue; | 608 | continue; |
diff --git a/net/decnet/dn_route.c b/net/decnet/dn_route.c index 5d7337bcf0fe..a8bf106b7a61 100644 --- a/net/decnet/dn_route.c +++ b/net/decnet/dn_route.c | |||
@@ -886,7 +886,7 @@ static int dn_route_output_slow(struct dst_entry **pprt, const struct flowi *old | |||
886 | .iif = loopback_dev.ifindex, | 886 | .iif = loopback_dev.ifindex, |
887 | .oif = oldflp->oif }; | 887 | .oif = oldflp->oif }; |
888 | struct dn_route *rt = NULL; | 888 | struct dn_route *rt = NULL; |
889 | struct net_device *dev_out = NULL; | 889 | struct net_device *dev_out = NULL, *dev; |
890 | struct neighbour *neigh = NULL; | 890 | struct neighbour *neigh = NULL; |
891 | unsigned hash; | 891 | unsigned hash; |
892 | unsigned flags = 0; | 892 | unsigned flags = 0; |
@@ -925,15 +925,17 @@ static int dn_route_output_slow(struct dst_entry **pprt, const struct flowi *old | |||
925 | goto out; | 925 | goto out; |
926 | } | 926 | } |
927 | read_lock(&dev_base_lock); | 927 | read_lock(&dev_base_lock); |
928 | for(dev_out = dev_base; dev_out; dev_out = dev_out->next) { | 928 | for_each_netdev(dev) { |
929 | if (!dev_out->dn_ptr) | 929 | if (!dev->dn_ptr) |
930 | continue; | 930 | continue; |
931 | if (!dn_dev_islocal(dev_out, oldflp->fld_src)) | 931 | if (!dn_dev_islocal(dev, oldflp->fld_src)) |
932 | continue; | 932 | continue; |
933 | if ((dev_out->flags & IFF_LOOPBACK) && | 933 | if ((dev->flags & IFF_LOOPBACK) && |
934 | oldflp->fld_dst && | 934 | oldflp->fld_dst && |
935 | !dn_dev_islocal(dev_out, oldflp->fld_dst)) | 935 | !dn_dev_islocal(dev, oldflp->fld_dst)) |
936 | continue; | 936 | continue; |
937 | |||
938 | dev_out = dev; | ||
937 | break; | 939 | break; |
938 | } | 940 | } |
939 | read_unlock(&dev_base_lock); | 941 | read_unlock(&dev_base_lock); |
diff --git a/net/ipv4/devinet.c b/net/ipv4/devinet.c index 088888db8b3d..7f95e6e9beeb 100644 --- a/net/ipv4/devinet.c +++ b/net/ipv4/devinet.c | |||
@@ -910,7 +910,7 @@ no_in_dev: | |||
910 | */ | 910 | */ |
911 | read_lock(&dev_base_lock); | 911 | read_lock(&dev_base_lock); |
912 | rcu_read_lock(); | 912 | rcu_read_lock(); |
913 | for (dev = dev_base; dev; dev = dev->next) { | 913 | for_each_netdev(dev) { |
914 | if ((in_dev = __in_dev_get_rcu(dev)) == NULL) | 914 | if ((in_dev = __in_dev_get_rcu(dev)) == NULL) |
915 | continue; | 915 | continue; |
916 | 916 | ||
@@ -989,7 +989,7 @@ __be32 inet_confirm_addr(const struct net_device *dev, __be32 dst, __be32 local, | |||
989 | 989 | ||
990 | read_lock(&dev_base_lock); | 990 | read_lock(&dev_base_lock); |
991 | rcu_read_lock(); | 991 | rcu_read_lock(); |
992 | for (dev = dev_base; dev; dev = dev->next) { | 992 | for_each_netdev(dev) { |
993 | if ((in_dev = __in_dev_get_rcu(dev))) { | 993 | if ((in_dev = __in_dev_get_rcu(dev))) { |
994 | addr = confirm_addr_indev(in_dev, dst, local, scope); | 994 | addr = confirm_addr_indev(in_dev, dst, local, scope); |
995 | if (addr) | 995 | if (addr) |
@@ -1182,23 +1182,26 @@ static int inet_dump_ifaddr(struct sk_buff *skb, struct netlink_callback *cb) | |||
1182 | int s_ip_idx, s_idx = cb->args[0]; | 1182 | int s_ip_idx, s_idx = cb->args[0]; |
1183 | 1183 | ||
1184 | s_ip_idx = ip_idx = cb->args[1]; | 1184 | s_ip_idx = ip_idx = cb->args[1]; |
1185 | for (dev = dev_base, idx = 0; dev; dev = dev->next, idx++) { | 1185 | idx = 0; |
1186 | for_each_netdev(dev) { | ||
1186 | if (idx < s_idx) | 1187 | if (idx < s_idx) |
1187 | continue; | 1188 | goto cont; |
1188 | if (idx > s_idx) | 1189 | if (idx > s_idx) |
1189 | s_ip_idx = 0; | 1190 | s_ip_idx = 0; |
1190 | if ((in_dev = __in_dev_get_rtnl(dev)) == NULL) | 1191 | if ((in_dev = __in_dev_get_rtnl(dev)) == NULL) |
1191 | continue; | 1192 | goto cont; |
1192 | 1193 | ||
1193 | for (ifa = in_dev->ifa_list, ip_idx = 0; ifa; | 1194 | for (ifa = in_dev->ifa_list, ip_idx = 0; ifa; |
1194 | ifa = ifa->ifa_next, ip_idx++) { | 1195 | ifa = ifa->ifa_next, ip_idx++) { |
1195 | if (ip_idx < s_ip_idx) | 1196 | if (ip_idx < s_ip_idx) |
1196 | continue; | 1197 | goto cont; |
1197 | if (inet_fill_ifaddr(skb, ifa, NETLINK_CB(cb->skb).pid, | 1198 | if (inet_fill_ifaddr(skb, ifa, NETLINK_CB(cb->skb).pid, |
1198 | cb->nlh->nlmsg_seq, | 1199 | cb->nlh->nlmsg_seq, |
1199 | RTM_NEWADDR, NLM_F_MULTI) <= 0) | 1200 | RTM_NEWADDR, NLM_F_MULTI) <= 0) |
1200 | goto done; | 1201 | goto done; |
1201 | } | 1202 | } |
1203 | cont: | ||
1204 | idx++; | ||
1202 | } | 1205 | } |
1203 | 1206 | ||
1204 | done: | 1207 | done: |
@@ -1243,7 +1246,7 @@ void inet_forward_change(void) | |||
1243 | ipv4_devconf_dflt.forwarding = on; | 1246 | ipv4_devconf_dflt.forwarding = on; |
1244 | 1247 | ||
1245 | read_lock(&dev_base_lock); | 1248 | read_lock(&dev_base_lock); |
1246 | for (dev = dev_base; dev; dev = dev->next) { | 1249 | for_each_netdev(dev) { |
1247 | struct in_device *in_dev; | 1250 | struct in_device *in_dev; |
1248 | rcu_read_lock(); | 1251 | rcu_read_lock(); |
1249 | in_dev = __in_dev_get_rcu(dev); | 1252 | in_dev = __in_dev_get_rcu(dev); |
diff --git a/net/ipv4/igmp.c b/net/ipv4/igmp.c index 2506021c2935..f4dd47453108 100644 --- a/net/ipv4/igmp.c +++ b/net/ipv4/igmp.c | |||
@@ -2288,9 +2288,8 @@ static inline struct ip_mc_list *igmp_mc_get_first(struct seq_file *seq) | |||
2288 | struct ip_mc_list *im = NULL; | 2288 | struct ip_mc_list *im = NULL; |
2289 | struct igmp_mc_iter_state *state = igmp_mc_seq_private(seq); | 2289 | struct igmp_mc_iter_state *state = igmp_mc_seq_private(seq); |
2290 | 2290 | ||
2291 | for (state->dev = dev_base, state->in_dev = NULL; | 2291 | state->in_dev = NULL; |
2292 | state->dev; | 2292 | for_each_netdev(state->dev) { |
2293 | state->dev = state->dev->next) { | ||
2294 | struct in_device *in_dev; | 2293 | struct in_device *in_dev; |
2295 | in_dev = in_dev_get(state->dev); | 2294 | in_dev = in_dev_get(state->dev); |
2296 | if (!in_dev) | 2295 | if (!in_dev) |
@@ -2316,7 +2315,7 @@ static struct ip_mc_list *igmp_mc_get_next(struct seq_file *seq, struct ip_mc_li | |||
2316 | read_unlock(&state->in_dev->mc_list_lock); | 2315 | read_unlock(&state->in_dev->mc_list_lock); |
2317 | in_dev_put(state->in_dev); | 2316 | in_dev_put(state->in_dev); |
2318 | } | 2317 | } |
2319 | state->dev = state->dev->next; | 2318 | state->dev = next_net_device(state->dev); |
2320 | if (!state->dev) { | 2319 | if (!state->dev) { |
2321 | state->in_dev = NULL; | 2320 | state->in_dev = NULL; |
2322 | break; | 2321 | break; |
@@ -2450,9 +2449,9 @@ static inline struct ip_sf_list *igmp_mcf_get_first(struct seq_file *seq) | |||
2450 | struct ip_mc_list *im = NULL; | 2449 | struct ip_mc_list *im = NULL; |
2451 | struct igmp_mcf_iter_state *state = igmp_mcf_seq_private(seq); | 2450 | struct igmp_mcf_iter_state *state = igmp_mcf_seq_private(seq); |
2452 | 2451 | ||
2453 | for (state->dev = dev_base, state->idev = NULL, state->im = NULL; | 2452 | state->idev = NULL; |
2454 | state->dev; | 2453 | state->im = NULL; |
2455 | state->dev = state->dev->next) { | 2454 | for_each_netdev(state->dev) { |
2456 | struct in_device *idev; | 2455 | struct in_device *idev; |
2457 | idev = in_dev_get(state->dev); | 2456 | idev = in_dev_get(state->dev); |
2458 | if (unlikely(idev == NULL)) | 2457 | if (unlikely(idev == NULL)) |
@@ -2488,7 +2487,7 @@ static struct ip_sf_list *igmp_mcf_get_next(struct seq_file *seq, struct ip_sf_l | |||
2488 | read_unlock(&state->idev->mc_list_lock); | 2487 | read_unlock(&state->idev->mc_list_lock); |
2489 | in_dev_put(state->idev); | 2488 | in_dev_put(state->idev); |
2490 | } | 2489 | } |
2491 | state->dev = state->dev->next; | 2490 | state->dev = next_net_device(state->dev); |
2492 | if (!state->dev) { | 2491 | if (!state->dev) { |
2493 | state->idev = NULL; | 2492 | state->idev = NULL; |
2494 | goto out; | 2493 | goto out; |
diff --git a/net/ipv4/ipconfig.c b/net/ipv4/ipconfig.c index 597c800b2fdc..342ca8d89458 100644 --- a/net/ipv4/ipconfig.c +++ b/net/ipv4/ipconfig.c | |||
@@ -192,7 +192,7 @@ static int __init ic_open_devs(void) | |||
192 | if (dev_change_flags(&loopback_dev, loopback_dev.flags | IFF_UP) < 0) | 192 | if (dev_change_flags(&loopback_dev, loopback_dev.flags | IFF_UP) < 0) |
193 | printk(KERN_ERR "IP-Config: Failed to open %s\n", loopback_dev.name); | 193 | printk(KERN_ERR "IP-Config: Failed to open %s\n", loopback_dev.name); |
194 | 194 | ||
195 | for (dev = dev_base; dev; dev = dev->next) { | 195 | for_each_netdev(dev) { |
196 | if (dev == &loopback_dev) | 196 | if (dev == &loopback_dev) |
197 | continue; | 197 | continue; |
198 | if (user_dev_name[0] ? !strcmp(dev->name, user_dev_name) : | 198 | if (user_dev_name[0] ? !strcmp(dev->name, user_dev_name) : |
diff --git a/net/ipv4/netfilter/nf_nat_proto_gre.c b/net/ipv4/netfilter/nf_nat_proto_gre.c index e5a34c17d927..c3908bc5a709 100644 --- a/net/ipv4/netfilter/nf_nat_proto_gre.c +++ b/net/ipv4/netfilter/nf_nat_proto_gre.c | |||
@@ -72,6 +72,11 @@ gre_unique_tuple(struct nf_conntrack_tuple *tuple, | |||
72 | __be16 *keyptr; | 72 | __be16 *keyptr; |
73 | unsigned int min, i, range_size; | 73 | unsigned int min, i, range_size; |
74 | 74 | ||
75 | /* If there is no master conntrack we are not PPTP, | ||
76 | do not change tuples */ | ||
77 | if (!conntrack->master) | ||
78 | return 0; | ||
79 | |||
75 | if (maniptype == IP_NAT_MANIP_SRC) | 80 | if (maniptype == IP_NAT_MANIP_SRC) |
76 | keyptr = &tuple->src.u.gre.key; | 81 | keyptr = &tuple->src.u.gre.key; |
77 | else | 82 | else |
@@ -122,18 +127,9 @@ gre_manip_pkt(struct sk_buff **pskb, unsigned int iphdroff, | |||
122 | if (maniptype != IP_NAT_MANIP_DST) | 127 | if (maniptype != IP_NAT_MANIP_DST) |
123 | return 1; | 128 | return 1; |
124 | switch (greh->version) { | 129 | switch (greh->version) { |
125 | case 0: | 130 | case GRE_VERSION_1701: |
126 | if (!greh->key) { | 131 | /* We do not currently NAT any GREv0 packets. |
127 | DEBUGP("can't nat GRE w/o key\n"); | 132 | * Try to behave like "nf_nat_proto_unknown" */ |
128 | break; | ||
129 | } | ||
130 | if (greh->csum) { | ||
131 | /* FIXME: Never tested this code... */ | ||
132 | nf_proto_csum_replace4(gre_csum(greh), *pskb, | ||
133 | *(gre_key(greh)), | ||
134 | tuple->dst.u.gre.key, 0); | ||
135 | } | ||
136 | *(gre_key(greh)) = tuple->dst.u.gre.key; | ||
137 | break; | 133 | break; |
138 | case GRE_VERSION_PPTP: | 134 | case GRE_VERSION_PPTP: |
139 | DEBUGP("call_id -> 0x%04x\n", ntohs(tuple->dst.u.gre.key)); | 135 | DEBUGP("call_id -> 0x%04x\n", ntohs(tuple->dst.u.gre.key)); |
diff --git a/net/ipv4/netfilter/nf_nat_rule.c b/net/ipv4/netfilter/nf_nat_rule.c index 2a283397a8b6..2534f718ab92 100644 --- a/net/ipv4/netfilter/nf_nat_rule.c +++ b/net/ipv4/netfilter/nf_nat_rule.c | |||
@@ -226,10 +226,6 @@ static int ipt_dnat_checkentry(const char *tablename, | |||
226 | printk("DNAT: multiple ranges no longer supported\n"); | 226 | printk("DNAT: multiple ranges no longer supported\n"); |
227 | return 0; | 227 | return 0; |
228 | } | 228 | } |
229 | if (mr->range[0].flags & IP_NAT_RANGE_PROTO_RANDOM) { | ||
230 | printk("DNAT: port randomization not supported\n"); | ||
231 | return 0; | ||
232 | } | ||
233 | return 1; | 229 | return 1; |
234 | } | 230 | } |
235 | 231 | ||
diff --git a/net/ipv4/netfilter/nf_nat_sip.c b/net/ipv4/netfilter/nf_nat_sip.c index bfd88e4e0685..fac97cf51ae5 100644 --- a/net/ipv4/netfilter/nf_nat_sip.c +++ b/net/ipv4/netfilter/nf_nat_sip.c | |||
@@ -222,6 +222,29 @@ static unsigned int mangle_sdp(struct sk_buff **pskb, | |||
222 | return mangle_content_len(pskb, ctinfo, ct, dptr); | 222 | return mangle_content_len(pskb, ctinfo, ct, dptr); |
223 | } | 223 | } |
224 | 224 | ||
225 | static void ip_nat_sdp_expect(struct nf_conn *ct, | ||
226 | struct nf_conntrack_expect *exp) | ||
227 | { | ||
228 | struct nf_nat_range range; | ||
229 | |||
230 | /* This must be a fresh one. */ | ||
231 | BUG_ON(ct->status & IPS_NAT_DONE_MASK); | ||
232 | |||
233 | /* Change src to where master sends to */ | ||
234 | range.flags = IP_NAT_RANGE_MAP_IPS; | ||
235 | range.min_ip = range.max_ip | ||
236 | = ct->master->tuplehash[!exp->dir].tuple.dst.u3.ip; | ||
237 | /* hook doesn't matter, but it has to do source manip */ | ||
238 | nf_nat_setup_info(ct, &range, NF_IP_POST_ROUTING); | ||
239 | |||
240 | /* For DST manip, map port here to where it's expected. */ | ||
241 | range.flags = (IP_NAT_RANGE_MAP_IPS | IP_NAT_RANGE_PROTO_SPECIFIED); | ||
242 | range.min = range.max = exp->saved_proto; | ||
243 | range.min_ip = range.max_ip = exp->saved_ip; | ||
244 | /* hook doesn't matter, but it has to do destination manip */ | ||
245 | nf_nat_setup_info(ct, &range, NF_IP_PRE_ROUTING); | ||
246 | } | ||
247 | |||
225 | /* So, this packet has hit the connection tracking matching code. | 248 | /* So, this packet has hit the connection tracking matching code. |
226 | Mangle it, and change the expectation to match the new version. */ | 249 | Mangle it, and change the expectation to match the new version. */ |
227 | static unsigned int ip_nat_sdp(struct sk_buff **pskb, | 250 | static unsigned int ip_nat_sdp(struct sk_buff **pskb, |
@@ -239,13 +262,14 @@ static unsigned int ip_nat_sdp(struct sk_buff **pskb, | |||
239 | /* Connection will come from reply */ | 262 | /* Connection will come from reply */ |
240 | newip = ct->tuplehash[!dir].tuple.dst.u3.ip; | 263 | newip = ct->tuplehash[!dir].tuple.dst.u3.ip; |
241 | 264 | ||
265 | exp->saved_ip = exp->tuple.dst.u3.ip; | ||
242 | exp->tuple.dst.u3.ip = newip; | 266 | exp->tuple.dst.u3.ip = newip; |
243 | exp->saved_proto.udp.port = exp->tuple.dst.u.udp.port; | 267 | exp->saved_proto.udp.port = exp->tuple.dst.u.udp.port; |
244 | exp->dir = !dir; | 268 | exp->dir = !dir; |
245 | 269 | ||
246 | /* When you see the packet, we need to NAT it the same as the | 270 | /* When you see the packet, we need to NAT it the same as the |
247 | this one. */ | 271 | this one. */ |
248 | exp->expectfn = nf_nat_follow_master; | 272 | exp->expectfn = ip_nat_sdp_expect; |
249 | 273 | ||
250 | /* Try to get same port: if not, try to change it. */ | 274 | /* Try to get same port: if not, try to change it. */ |
251 | for (port = ntohs(exp->saved_proto.udp.port); port != 0; port++) { | 275 | for (port = ntohs(exp->saved_proto.udp.port); port != 0; port++) { |
diff --git a/net/ipv4/tcp.c b/net/ipv4/tcp.c index d6e488668171..8b124eafbb90 100644 --- a/net/ipv4/tcp.c +++ b/net/ipv4/tcp.c | |||
@@ -1760,8 +1760,7 @@ int tcp_disconnect(struct sock *sk, int flags) | |||
1760 | tcp_clear_retrans(tp); | 1760 | tcp_clear_retrans(tp); |
1761 | inet_csk_delack_init(sk); | 1761 | inet_csk_delack_init(sk); |
1762 | tcp_init_send_head(sk); | 1762 | tcp_init_send_head(sk); |
1763 | tp->rx_opt.saw_tstamp = 0; | 1763 | memset(&tp->rx_opt, 0, sizeof(tp->rx_opt)); |
1764 | tcp_sack_reset(&tp->rx_opt); | ||
1765 | __sk_dst_reset(sk); | 1764 | __sk_dst_reset(sk); |
1766 | 1765 | ||
1767 | BUG_TRAP(!inet->num || icsk->icsk_bind_hash); | 1766 | BUG_TRAP(!inet->num || icsk->icsk_bind_hash); |
diff --git a/net/ipv4/tcp_highspeed.c b/net/ipv4/tcp_highspeed.c index a291097fcc0a..43d624e5043c 100644 --- a/net/ipv4/tcp_highspeed.c +++ b/net/ipv4/tcp_highspeed.c | |||
@@ -97,10 +97,6 @@ struct hstcp { | |||
97 | u32 ai; | 97 | u32 ai; |
98 | }; | 98 | }; |
99 | 99 | ||
100 | static int max_ssthresh = 100; | ||
101 | module_param(max_ssthresh, int, 0644); | ||
102 | MODULE_PARM_DESC(max_ssthresh, "limited slow start threshold (RFC3742)"); | ||
103 | |||
104 | static void hstcp_init(struct sock *sk) | 100 | static void hstcp_init(struct sock *sk) |
105 | { | 101 | { |
106 | struct tcp_sock *tp = tcp_sk(sk); | 102 | struct tcp_sock *tp = tcp_sk(sk); |
@@ -122,23 +118,9 @@ static void hstcp_cong_avoid(struct sock *sk, u32 adk, u32 rtt, | |||
122 | if (!tcp_is_cwnd_limited(sk, in_flight)) | 118 | if (!tcp_is_cwnd_limited(sk, in_flight)) |
123 | return; | 119 | return; |
124 | 120 | ||
125 | if (tp->snd_cwnd <= tp->snd_ssthresh) { | 121 | if (tp->snd_cwnd <= tp->snd_ssthresh) |
126 | /* RFC3742: limited slow start | 122 | tcp_slow_start(tp); |
127 | * the window is increased by 1/K MSS for each arriving ACK, | 123 | else { |
128 | * for K = int(cwnd/(0.5 max_ssthresh)) | ||
129 | */ | ||
130 | if (max_ssthresh > 0 && tp->snd_cwnd > max_ssthresh) { | ||
131 | u32 k = max(tp->snd_cwnd / (max_ssthresh >> 1), 1U); | ||
132 | if (++tp->snd_cwnd_cnt >= k) { | ||
133 | if (tp->snd_cwnd < tp->snd_cwnd_clamp) | ||
134 | tp->snd_cwnd++; | ||
135 | tp->snd_cwnd_cnt = 0; | ||
136 | } | ||
137 | } else { | ||
138 | if (tp->snd_cwnd < tp->snd_cwnd_clamp) | ||
139 | tp->snd_cwnd++; | ||
140 | } | ||
141 | } else { | ||
142 | /* Update AIMD parameters. | 124 | /* Update AIMD parameters. |
143 | * | 125 | * |
144 | * We want to guarantee that: | 126 | * We want to guarantee that: |
diff --git a/net/ipv4/tcp_yeah.h b/net/ipv4/tcp_yeah.h deleted file mode 100644 index ed3b7198f23c..000000000000 --- a/net/ipv4/tcp_yeah.h +++ /dev/null | |||
@@ -1,7 +0,0 @@ | |||
1 | #include <linux/mm.h> | ||
2 | #include <linux/module.h> | ||
3 | #include <linux/skbuff.h> | ||
4 | #include <linux/inet_diag.h> | ||
5 | #include <asm/div64.h> | ||
6 | |||
7 | #include <net/tcp.h> | ||
diff --git a/net/ipv6/addrconf.c b/net/ipv6/addrconf.c index 3452433cbc96..d02685c6bc69 100644 --- a/net/ipv6/addrconf.c +++ b/net/ipv6/addrconf.c | |||
@@ -449,7 +449,7 @@ static void addrconf_forward_change(void) | |||
449 | struct inet6_dev *idev; | 449 | struct inet6_dev *idev; |
450 | 450 | ||
451 | read_lock(&dev_base_lock); | 451 | read_lock(&dev_base_lock); |
452 | for (dev=dev_base; dev; dev=dev->next) { | 452 | for_each_netdev(dev) { |
453 | rcu_read_lock(); | 453 | rcu_read_lock(); |
454 | idev = __in6_dev_get(dev); | 454 | idev = __in6_dev_get(dev); |
455 | if (idev) { | 455 | if (idev) { |
@@ -911,7 +911,7 @@ int ipv6_dev_get_saddr(struct net_device *daddr_dev, | |||
911 | read_lock(&dev_base_lock); | 911 | read_lock(&dev_base_lock); |
912 | rcu_read_lock(); | 912 | rcu_read_lock(); |
913 | 913 | ||
914 | for (dev = dev_base; dev; dev=dev->next) { | 914 | for_each_netdev(dev) { |
915 | struct inet6_dev *idev; | 915 | struct inet6_dev *idev; |
916 | struct inet6_ifaddr *ifa; | 916 | struct inet6_ifaddr *ifa; |
917 | 917 | ||
@@ -2064,7 +2064,7 @@ static void sit_add_v4_addrs(struct inet6_dev *idev) | |||
2064 | return; | 2064 | return; |
2065 | } | 2065 | } |
2066 | 2066 | ||
2067 | for (dev = dev_base; dev != NULL; dev = dev->next) { | 2067 | for_each_netdev(dev) { |
2068 | struct in_device * in_dev = __in_dev_get_rtnl(dev); | 2068 | struct in_device * in_dev = __in_dev_get_rtnl(dev); |
2069 | if (in_dev && (dev->flags & IFF_UP)) { | 2069 | if (in_dev && (dev->flags & IFF_UP)) { |
2070 | struct in_ifaddr * ifa; | 2070 | struct in_ifaddr * ifa; |
@@ -2225,7 +2225,7 @@ static void ip6_tnl_add_linklocal(struct inet6_dev *idev) | |||
2225 | return; | 2225 | return; |
2226 | } | 2226 | } |
2227 | /* then try to inherit it from any device */ | 2227 | /* then try to inherit it from any device */ |
2228 | for (link_dev = dev_base; link_dev; link_dev = link_dev->next) { | 2228 | for_each_netdev(link_dev) { |
2229 | if (!ipv6_inherit_linklocal(idev, link_dev)) | 2229 | if (!ipv6_inherit_linklocal(idev, link_dev)) |
2230 | return; | 2230 | return; |
2231 | } | 2231 | } |
@@ -3257,14 +3257,15 @@ static int inet6_dump_addr(struct sk_buff *skb, struct netlink_callback *cb, | |||
3257 | s_idx = cb->args[0]; | 3257 | s_idx = cb->args[0]; |
3258 | s_ip_idx = ip_idx = cb->args[1]; | 3258 | s_ip_idx = ip_idx = cb->args[1]; |
3259 | 3259 | ||
3260 | for (dev = dev_base, idx = 0; dev; dev = dev->next, idx++) { | 3260 | idx = 0; |
3261 | for_each_netdev(dev) { | ||
3261 | if (idx < s_idx) | 3262 | if (idx < s_idx) |
3262 | continue; | 3263 | goto cont; |
3263 | if (idx > s_idx) | 3264 | if (idx > s_idx) |
3264 | s_ip_idx = 0; | 3265 | s_ip_idx = 0; |
3265 | ip_idx = 0; | 3266 | ip_idx = 0; |
3266 | if ((idev = in6_dev_get(dev)) == NULL) | 3267 | if ((idev = in6_dev_get(dev)) == NULL) |
3267 | continue; | 3268 | goto cont; |
3268 | read_lock_bh(&idev->lock); | 3269 | read_lock_bh(&idev->lock); |
3269 | switch (type) { | 3270 | switch (type) { |
3270 | case UNICAST_ADDR: | 3271 | case UNICAST_ADDR: |
@@ -3311,6 +3312,8 @@ static int inet6_dump_addr(struct sk_buff *skb, struct netlink_callback *cb, | |||
3311 | } | 3312 | } |
3312 | read_unlock_bh(&idev->lock); | 3313 | read_unlock_bh(&idev->lock); |
3313 | in6_dev_put(idev); | 3314 | in6_dev_put(idev); |
3315 | cont: | ||
3316 | idx++; | ||
3314 | } | 3317 | } |
3315 | done: | 3318 | done: |
3316 | if (err <= 0) { | 3319 | if (err <= 0) { |
@@ -3575,16 +3578,19 @@ static int inet6_dump_ifinfo(struct sk_buff *skb, struct netlink_callback *cb) | |||
3575 | struct inet6_dev *idev; | 3578 | struct inet6_dev *idev; |
3576 | 3579 | ||
3577 | read_lock(&dev_base_lock); | 3580 | read_lock(&dev_base_lock); |
3578 | for (dev=dev_base, idx=0; dev; dev = dev->next, idx++) { | 3581 | idx = 0; |
3582 | for_each_netdev(dev) { | ||
3579 | if (idx < s_idx) | 3583 | if (idx < s_idx) |
3580 | continue; | 3584 | goto cont; |
3581 | if ((idev = in6_dev_get(dev)) == NULL) | 3585 | if ((idev = in6_dev_get(dev)) == NULL) |
3582 | continue; | 3586 | goto cont; |
3583 | err = inet6_fill_ifinfo(skb, idev, NETLINK_CB(cb->skb).pid, | 3587 | err = inet6_fill_ifinfo(skb, idev, NETLINK_CB(cb->skb).pid, |
3584 | cb->nlh->nlmsg_seq, RTM_NEWLINK, NLM_F_MULTI); | 3588 | cb->nlh->nlmsg_seq, RTM_NEWLINK, NLM_F_MULTI); |
3585 | in6_dev_put(idev); | 3589 | in6_dev_put(idev); |
3586 | if (err <= 0) | 3590 | if (err <= 0) |
3587 | break; | 3591 | break; |
3592 | cont: | ||
3593 | idx++; | ||
3588 | } | 3594 | } |
3589 | read_unlock(&dev_base_lock); | 3595 | read_unlock(&dev_base_lock); |
3590 | cb->args[0] = idx; | 3596 | cb->args[0] = idx; |
@@ -4247,7 +4253,7 @@ void __exit addrconf_cleanup(void) | |||
4247 | * clean dev list. | 4253 | * clean dev list. |
4248 | */ | 4254 | */ |
4249 | 4255 | ||
4250 | for (dev=dev_base; dev; dev=dev->next) { | 4256 | for_each_netdev(dev) { |
4251 | if ((idev = __in6_dev_get(dev)) == NULL) | 4257 | if ((idev = __in6_dev_get(dev)) == NULL) |
4252 | continue; | 4258 | continue; |
4253 | addrconf_ifdown(dev, 1); | 4259 | addrconf_ifdown(dev, 1); |
diff --git a/net/ipv6/anycast.c b/net/ipv6/anycast.c index 09117d63256f..9b81264eb78f 100644 --- a/net/ipv6/anycast.c +++ b/net/ipv6/anycast.c | |||
@@ -423,14 +423,18 @@ static int ipv6_chk_acast_dev(struct net_device *dev, struct in6_addr *addr) | |||
423 | */ | 423 | */ |
424 | int ipv6_chk_acast_addr(struct net_device *dev, struct in6_addr *addr) | 424 | int ipv6_chk_acast_addr(struct net_device *dev, struct in6_addr *addr) |
425 | { | 425 | { |
426 | int found = 0; | ||
427 | |||
426 | if (dev) | 428 | if (dev) |
427 | return ipv6_chk_acast_dev(dev, addr); | 429 | return ipv6_chk_acast_dev(dev, addr); |
428 | read_lock(&dev_base_lock); | 430 | read_lock(&dev_base_lock); |
429 | for (dev=dev_base; dev; dev=dev->next) | 431 | for_each_netdev(dev) |
430 | if (ipv6_chk_acast_dev(dev, addr)) | 432 | if (ipv6_chk_acast_dev(dev, addr)) { |
433 | found = 1; | ||
431 | break; | 434 | break; |
435 | } | ||
432 | read_unlock(&dev_base_lock); | 436 | read_unlock(&dev_base_lock); |
433 | return dev != 0; | 437 | return found; |
434 | } | 438 | } |
435 | 439 | ||
436 | 440 | ||
@@ -447,9 +451,8 @@ static inline struct ifacaddr6 *ac6_get_first(struct seq_file *seq) | |||
447 | struct ifacaddr6 *im = NULL; | 451 | struct ifacaddr6 *im = NULL; |
448 | struct ac6_iter_state *state = ac6_seq_private(seq); | 452 | struct ac6_iter_state *state = ac6_seq_private(seq); |
449 | 453 | ||
450 | for (state->dev = dev_base, state->idev = NULL; | 454 | state->idev = NULL; |
451 | state->dev; | 455 | for_each_netdev(state->dev) { |
452 | state->dev = state->dev->next) { | ||
453 | struct inet6_dev *idev; | 456 | struct inet6_dev *idev; |
454 | idev = in6_dev_get(state->dev); | 457 | idev = in6_dev_get(state->dev); |
455 | if (!idev) | 458 | if (!idev) |
@@ -476,7 +479,7 @@ static struct ifacaddr6 *ac6_get_next(struct seq_file *seq, struct ifacaddr6 *im | |||
476 | read_unlock_bh(&state->idev->lock); | 479 | read_unlock_bh(&state->idev->lock); |
477 | in6_dev_put(state->idev); | 480 | in6_dev_put(state->idev); |
478 | } | 481 | } |
479 | state->dev = state->dev->next; | 482 | state->dev = next_net_device(state->dev); |
480 | if (!state->dev) { | 483 | if (!state->dev) { |
481 | state->idev = NULL; | 484 | state->idev = NULL; |
482 | break; | 485 | break; |
diff --git a/net/ipv6/mcast.c b/net/ipv6/mcast.c index 6c2758951d60..3e308fb41b49 100644 --- a/net/ipv6/mcast.c +++ b/net/ipv6/mcast.c | |||
@@ -2331,9 +2331,8 @@ static inline struct ifmcaddr6 *igmp6_mc_get_first(struct seq_file *seq) | |||
2331 | struct ifmcaddr6 *im = NULL; | 2331 | struct ifmcaddr6 *im = NULL; |
2332 | struct igmp6_mc_iter_state *state = igmp6_mc_seq_private(seq); | 2332 | struct igmp6_mc_iter_state *state = igmp6_mc_seq_private(seq); |
2333 | 2333 | ||
2334 | for (state->dev = dev_base, state->idev = NULL; | 2334 | state->idev = NULL; |
2335 | state->dev; | 2335 | for_each_netdev(state->dev) { |
2336 | state->dev = state->dev->next) { | ||
2337 | struct inet6_dev *idev; | 2336 | struct inet6_dev *idev; |
2338 | idev = in6_dev_get(state->dev); | 2337 | idev = in6_dev_get(state->dev); |
2339 | if (!idev) | 2338 | if (!idev) |
@@ -2360,7 +2359,7 @@ static struct ifmcaddr6 *igmp6_mc_get_next(struct seq_file *seq, struct ifmcaddr | |||
2360 | read_unlock_bh(&state->idev->lock); | 2359 | read_unlock_bh(&state->idev->lock); |
2361 | in6_dev_put(state->idev); | 2360 | in6_dev_put(state->idev); |
2362 | } | 2361 | } |
2363 | state->dev = state->dev->next; | 2362 | state->dev = next_net_device(state->dev); |
2364 | if (!state->dev) { | 2363 | if (!state->dev) { |
2365 | state->idev = NULL; | 2364 | state->idev = NULL; |
2366 | break; | 2365 | break; |
@@ -2475,9 +2474,9 @@ static inline struct ip6_sf_list *igmp6_mcf_get_first(struct seq_file *seq) | |||
2475 | struct ifmcaddr6 *im = NULL; | 2474 | struct ifmcaddr6 *im = NULL; |
2476 | struct igmp6_mcf_iter_state *state = igmp6_mcf_seq_private(seq); | 2475 | struct igmp6_mcf_iter_state *state = igmp6_mcf_seq_private(seq); |
2477 | 2476 | ||
2478 | for (state->dev = dev_base, state->idev = NULL, state->im = NULL; | 2477 | state->idev = NULL; |
2479 | state->dev; | 2478 | state->im = NULL; |
2480 | state->dev = state->dev->next) { | 2479 | for_each_netdev(state->dev) { |
2481 | struct inet6_dev *idev; | 2480 | struct inet6_dev *idev; |
2482 | idev = in6_dev_get(state->dev); | 2481 | idev = in6_dev_get(state->dev); |
2483 | if (unlikely(idev == NULL)) | 2482 | if (unlikely(idev == NULL)) |
@@ -2513,7 +2512,7 @@ static struct ip6_sf_list *igmp6_mcf_get_next(struct seq_file *seq, struct ip6_s | |||
2513 | read_unlock_bh(&state->idev->lock); | 2512 | read_unlock_bh(&state->idev->lock); |
2514 | in6_dev_put(state->idev); | 2513 | in6_dev_put(state->idev); |
2515 | } | 2514 | } |
2516 | state->dev = state->dev->next; | 2515 | state->dev = next_net_device(state->dev); |
2517 | if (!state->dev) { | 2516 | if (!state->dev) { |
2518 | state->idev = NULL; | 2517 | state->idev = NULL; |
2519 | goto out; | 2518 | goto out; |
diff --git a/net/iucv/af_iucv.c b/net/iucv/af_iucv.c index e84c924a81ee..2f1373855a8b 100644 --- a/net/iucv/af_iucv.c +++ b/net/iucv/af_iucv.c | |||
@@ -45,7 +45,8 @@ static struct proto iucv_proto = { | |||
45 | static void iucv_callback_rx(struct iucv_path *, struct iucv_message *); | 45 | static void iucv_callback_rx(struct iucv_path *, struct iucv_message *); |
46 | static void iucv_callback_txdone(struct iucv_path *, struct iucv_message *); | 46 | static void iucv_callback_txdone(struct iucv_path *, struct iucv_message *); |
47 | static void iucv_callback_connack(struct iucv_path *, u8 ipuser[16]); | 47 | static void iucv_callback_connack(struct iucv_path *, u8 ipuser[16]); |
48 | static int iucv_callback_connreq(struct iucv_path *, u8 ipvmid[8], u8 ipuser[16]); | 48 | static int iucv_callback_connreq(struct iucv_path *, u8 ipvmid[8], |
49 | u8 ipuser[16]); | ||
49 | static void iucv_callback_connrej(struct iucv_path *, u8 ipuser[16]); | 50 | static void iucv_callback_connrej(struct iucv_path *, u8 ipuser[16]); |
50 | 51 | ||
51 | static struct iucv_sock_list iucv_sk_list = { | 52 | static struct iucv_sock_list iucv_sk_list = { |
@@ -147,11 +148,12 @@ static void iucv_sock_close(struct sock *sk) | |||
147 | unsigned char user_data[16]; | 148 | unsigned char user_data[16]; |
148 | struct iucv_sock *iucv = iucv_sk(sk); | 149 | struct iucv_sock *iucv = iucv_sk(sk); |
149 | int err; | 150 | int err; |
151 | unsigned long timeo; | ||
150 | 152 | ||
151 | iucv_sock_clear_timer(sk); | 153 | iucv_sock_clear_timer(sk); |
152 | lock_sock(sk); | 154 | lock_sock(sk); |
153 | 155 | ||
154 | switch(sk->sk_state) { | 156 | switch (sk->sk_state) { |
155 | case IUCV_LISTEN: | 157 | case IUCV_LISTEN: |
156 | iucv_sock_cleanup_listen(sk); | 158 | iucv_sock_cleanup_listen(sk); |
157 | break; | 159 | break; |
@@ -159,6 +161,21 @@ static void iucv_sock_close(struct sock *sk) | |||
159 | case IUCV_CONNECTED: | 161 | case IUCV_CONNECTED: |
160 | case IUCV_DISCONN: | 162 | case IUCV_DISCONN: |
161 | err = 0; | 163 | err = 0; |
164 | |||
165 | sk->sk_state = IUCV_CLOSING; | ||
166 | sk->sk_state_change(sk); | ||
167 | |||
168 | if (!skb_queue_empty(&iucv->send_skb_q)) { | ||
169 | if (sock_flag(sk, SOCK_LINGER) && sk->sk_lingertime) | ||
170 | timeo = sk->sk_lingertime; | ||
171 | else | ||
172 | timeo = IUCV_DISCONN_TIMEOUT; | ||
173 | err = iucv_sock_wait_state(sk, IUCV_CLOSED, 0, timeo); | ||
174 | } | ||
175 | |||
176 | sk->sk_state = IUCV_CLOSED; | ||
177 | sk->sk_state_change(sk); | ||
178 | |||
162 | if (iucv->path) { | 179 | if (iucv->path) { |
163 | low_nmcpy(user_data, iucv->src_name); | 180 | low_nmcpy(user_data, iucv->src_name); |
164 | high_nmcpy(user_data, iucv->dst_name); | 181 | high_nmcpy(user_data, iucv->dst_name); |
@@ -168,12 +185,11 @@ static void iucv_sock_close(struct sock *sk) | |||
168 | iucv->path = NULL; | 185 | iucv->path = NULL; |
169 | } | 186 | } |
170 | 187 | ||
171 | sk->sk_state = IUCV_CLOSED; | ||
172 | sk->sk_state_change(sk); | ||
173 | sk->sk_err = ECONNRESET; | 188 | sk->sk_err = ECONNRESET; |
174 | sk->sk_state_change(sk); | 189 | sk->sk_state_change(sk); |
175 | 190 | ||
176 | skb_queue_purge(&iucv->send_skb_q); | 191 | skb_queue_purge(&iucv->send_skb_q); |
192 | skb_queue_purge(&iucv->backlog_skb_q); | ||
177 | 193 | ||
178 | sock_set_flag(sk, SOCK_ZAPPED); | 194 | sock_set_flag(sk, SOCK_ZAPPED); |
179 | break; | 195 | break; |
@@ -204,6 +220,7 @@ static struct sock *iucv_sock_alloc(struct socket *sock, int proto, gfp_t prio) | |||
204 | sock_init_data(sock, sk); | 220 | sock_init_data(sock, sk); |
205 | INIT_LIST_HEAD(&iucv_sk(sk)->accept_q); | 221 | INIT_LIST_HEAD(&iucv_sk(sk)->accept_q); |
206 | skb_queue_head_init(&iucv_sk(sk)->send_skb_q); | 222 | skb_queue_head_init(&iucv_sk(sk)->send_skb_q); |
223 | skb_queue_head_init(&iucv_sk(sk)->backlog_skb_q); | ||
207 | iucv_sk(sk)->send_tag = 0; | 224 | iucv_sk(sk)->send_tag = 0; |
208 | 225 | ||
209 | sk->sk_destruct = iucv_sock_destruct; | 226 | sk->sk_destruct = iucv_sock_destruct; |
@@ -276,7 +293,7 @@ struct sock *iucv_accept_dequeue(struct sock *parent, struct socket *newsock) | |||
276 | struct iucv_sock *isk, *n; | 293 | struct iucv_sock *isk, *n; |
277 | struct sock *sk; | 294 | struct sock *sk; |
278 | 295 | ||
279 | list_for_each_entry_safe(isk, n, &iucv_sk(parent)->accept_q, accept_q){ | 296 | list_for_each_entry_safe(isk, n, &iucv_sk(parent)->accept_q, accept_q) { |
280 | sk = (struct sock *) isk; | 297 | sk = (struct sock *) isk; |
281 | lock_sock(sk); | 298 | lock_sock(sk); |
282 | 299 | ||
@@ -510,7 +527,7 @@ static int iucv_sock_accept(struct socket *sock, struct socket *newsock, | |||
510 | long timeo; | 527 | long timeo; |
511 | int err = 0; | 528 | int err = 0; |
512 | 529 | ||
513 | lock_sock(sk); | 530 | lock_sock_nested(sk, SINGLE_DEPTH_NESTING); |
514 | 531 | ||
515 | if (sk->sk_state != IUCV_LISTEN) { | 532 | if (sk->sk_state != IUCV_LISTEN) { |
516 | err = -EBADFD; | 533 | err = -EBADFD; |
@@ -521,7 +538,7 @@ static int iucv_sock_accept(struct socket *sock, struct socket *newsock, | |||
521 | 538 | ||
522 | /* Wait for an incoming connection */ | 539 | /* Wait for an incoming connection */ |
523 | add_wait_queue_exclusive(sk->sk_sleep, &wait); | 540 | add_wait_queue_exclusive(sk->sk_sleep, &wait); |
524 | while (!(nsk = iucv_accept_dequeue(sk, newsock))){ | 541 | while (!(nsk = iucv_accept_dequeue(sk, newsock))) { |
525 | set_current_state(TASK_INTERRUPTIBLE); | 542 | set_current_state(TASK_INTERRUPTIBLE); |
526 | if (!timeo) { | 543 | if (!timeo) { |
527 | err = -EAGAIN; | 544 | err = -EAGAIN; |
@@ -530,7 +547,7 @@ static int iucv_sock_accept(struct socket *sock, struct socket *newsock, | |||
530 | 547 | ||
531 | release_sock(sk); | 548 | release_sock(sk); |
532 | timeo = schedule_timeout(timeo); | 549 | timeo = schedule_timeout(timeo); |
533 | lock_sock(sk); | 550 | lock_sock_nested(sk, SINGLE_DEPTH_NESTING); |
534 | 551 | ||
535 | if (sk->sk_state != IUCV_LISTEN) { | 552 | if (sk->sk_state != IUCV_LISTEN) { |
536 | err = -EBADFD; | 553 | err = -EBADFD; |
@@ -602,13 +619,13 @@ static int iucv_sock_sendmsg(struct kiocb *iocb, struct socket *sock, | |||
602 | goto out; | 619 | goto out; |
603 | } | 620 | } |
604 | 621 | ||
605 | if (sk->sk_state == IUCV_CONNECTED){ | 622 | if (sk->sk_state == IUCV_CONNECTED) { |
606 | if(!(skb = sock_alloc_send_skb(sk, len, | 623 | if (!(skb = sock_alloc_send_skb(sk, len, |
607 | msg->msg_flags & MSG_DONTWAIT, | 624 | msg->msg_flags & MSG_DONTWAIT, |
608 | &err))) | 625 | &err))) |
609 | return err; | 626 | goto out; |
610 | 627 | ||
611 | if (memcpy_fromiovec(skb_put(skb, len), msg->msg_iov, len)){ | 628 | if (memcpy_fromiovec(skb_put(skb, len), msg->msg_iov, len)) { |
612 | err = -EFAULT; | 629 | err = -EFAULT; |
613 | goto fail; | 630 | goto fail; |
614 | } | 631 | } |
@@ -647,10 +664,16 @@ static int iucv_sock_recvmsg(struct kiocb *iocb, struct socket *sock, | |||
647 | { | 664 | { |
648 | int noblock = flags & MSG_DONTWAIT; | 665 | int noblock = flags & MSG_DONTWAIT; |
649 | struct sock *sk = sock->sk; | 666 | struct sock *sk = sock->sk; |
667 | struct iucv_sock *iucv = iucv_sk(sk); | ||
650 | int target, copied = 0; | 668 | int target, copied = 0; |
651 | struct sk_buff *skb; | 669 | struct sk_buff *skb, *rskb, *cskb; |
652 | int err = 0; | 670 | int err = 0; |
653 | 671 | ||
672 | if ((sk->sk_state == IUCV_DISCONN || sk->sk_state == IUCV_SEVERED) && | ||
673 | skb_queue_empty(&iucv->backlog_skb_q) && | ||
674 | skb_queue_empty(&sk->sk_receive_queue)) | ||
675 | return 0; | ||
676 | |||
654 | if (flags & (MSG_OOB)) | 677 | if (flags & (MSG_OOB)) |
655 | return -EOPNOTSUPP; | 678 | return -EOPNOTSUPP; |
656 | 679 | ||
@@ -665,10 +688,12 @@ static int iucv_sock_recvmsg(struct kiocb *iocb, struct socket *sock, | |||
665 | 688 | ||
666 | copied = min_t(unsigned int, skb->len, len); | 689 | copied = min_t(unsigned int, skb->len, len); |
667 | 690 | ||
668 | if (memcpy_toiovec(msg->msg_iov, skb->data, copied)) { | 691 | cskb = skb; |
692 | if (memcpy_toiovec(msg->msg_iov, cskb->data, copied)) { | ||
669 | skb_queue_head(&sk->sk_receive_queue, skb); | 693 | skb_queue_head(&sk->sk_receive_queue, skb); |
670 | if (copied == 0) | 694 | if (copied == 0) |
671 | return -EFAULT; | 695 | return -EFAULT; |
696 | goto done; | ||
672 | } | 697 | } |
673 | 698 | ||
674 | len -= copied; | 699 | len -= copied; |
@@ -683,6 +708,18 @@ static int iucv_sock_recvmsg(struct kiocb *iocb, struct socket *sock, | |||
683 | } | 708 | } |
684 | 709 | ||
685 | kfree_skb(skb); | 710 | kfree_skb(skb); |
711 | |||
712 | /* Queue backlog skbs */ | ||
713 | rskb = skb_dequeue(&iucv_sk(sk)->backlog_skb_q); | ||
714 | while (rskb) { | ||
715 | if (sock_queue_rcv_skb(sk, rskb)) { | ||
716 | skb_queue_head(&iucv_sk(sk)->backlog_skb_q, | ||
717 | rskb); | ||
718 | break; | ||
719 | } else { | ||
720 | rskb = skb_dequeue(&iucv_sk(sk)->backlog_skb_q); | ||
721 | } | ||
722 | } | ||
686 | } else | 723 | } else |
687 | skb_queue_head(&sk->sk_receive_queue, skb); | 724 | skb_queue_head(&sk->sk_receive_queue, skb); |
688 | 725 | ||
@@ -695,7 +732,7 @@ static inline unsigned int iucv_accept_poll(struct sock *parent) | |||
695 | struct iucv_sock *isk, *n; | 732 | struct iucv_sock *isk, *n; |
696 | struct sock *sk; | 733 | struct sock *sk; |
697 | 734 | ||
698 | list_for_each_entry_safe(isk, n, &iucv_sk(parent)->accept_q, accept_q){ | 735 | list_for_each_entry_safe(isk, n, &iucv_sk(parent)->accept_q, accept_q) { |
699 | sk = (struct sock *) isk; | 736 | sk = (struct sock *) isk; |
700 | 737 | ||
701 | if (sk->sk_state == IUCV_CONNECTED) | 738 | if (sk->sk_state == IUCV_CONNECTED) |
@@ -726,12 +763,15 @@ unsigned int iucv_sock_poll(struct file *file, struct socket *sock, | |||
726 | mask |= POLLHUP; | 763 | mask |= POLLHUP; |
727 | 764 | ||
728 | if (!skb_queue_empty(&sk->sk_receive_queue) || | 765 | if (!skb_queue_empty(&sk->sk_receive_queue) || |
729 | (sk->sk_shutdown & RCV_SHUTDOWN)) | 766 | (sk->sk_shutdown & RCV_SHUTDOWN)) |
730 | mask |= POLLIN | POLLRDNORM; | 767 | mask |= POLLIN | POLLRDNORM; |
731 | 768 | ||
732 | if (sk->sk_state == IUCV_CLOSED) | 769 | if (sk->sk_state == IUCV_CLOSED) |
733 | mask |= POLLHUP; | 770 | mask |= POLLHUP; |
734 | 771 | ||
772 | if (sk->sk_state == IUCV_DISCONN || sk->sk_state == IUCV_SEVERED) | ||
773 | mask |= POLLIN; | ||
774 | |||
735 | if (sock_writeable(sk)) | 775 | if (sock_writeable(sk)) |
736 | mask |= POLLOUT | POLLWRNORM | POLLWRBAND; | 776 | mask |= POLLOUT | POLLWRNORM | POLLWRBAND; |
737 | else | 777 | else |
@@ -754,7 +794,7 @@ static int iucv_sock_shutdown(struct socket *sock, int how) | |||
754 | return -EINVAL; | 794 | return -EINVAL; |
755 | 795 | ||
756 | lock_sock(sk); | 796 | lock_sock(sk); |
757 | switch(sk->sk_state) { | 797 | switch (sk->sk_state) { |
758 | case IUCV_CLOSED: | 798 | case IUCV_CLOSED: |
759 | err = -ENOTCONN; | 799 | err = -ENOTCONN; |
760 | goto fail; | 800 | goto fail; |
@@ -770,7 +810,7 @@ static int iucv_sock_shutdown(struct socket *sock, int how) | |||
770 | err = iucv_message_send(iucv->path, &txmsg, IUCV_IPRMDATA, 0, | 810 | err = iucv_message_send(iucv->path, &txmsg, IUCV_IPRMDATA, 0, |
771 | (void *) prmmsg, 8); | 811 | (void *) prmmsg, 8); |
772 | if (err) { | 812 | if (err) { |
773 | switch(err) { | 813 | switch (err) { |
774 | case 1: | 814 | case 1: |
775 | err = -ENOTCONN; | 815 | err = -ENOTCONN; |
776 | break; | 816 | break; |
@@ -817,13 +857,6 @@ static int iucv_sock_release(struct socket *sock) | |||
817 | iucv_sk(sk)->path = NULL; | 857 | iucv_sk(sk)->path = NULL; |
818 | } | 858 | } |
819 | 859 | ||
820 | if (sock_flag(sk, SOCK_LINGER) && sk->sk_lingertime){ | ||
821 | lock_sock(sk); | ||
822 | err = iucv_sock_wait_state(sk, IUCV_CLOSED, 0, | ||
823 | sk->sk_lingertime); | ||
824 | release_sock(sk); | ||
825 | } | ||
826 | |||
827 | sock_orphan(sk); | 860 | sock_orphan(sk); |
828 | iucv_sock_kill(sk); | 861 | iucv_sock_kill(sk); |
829 | return err; | 862 | return err; |
@@ -880,7 +913,7 @@ static int iucv_callback_connreq(struct iucv_path *path, | |||
880 | 913 | ||
881 | /* Create the new socket */ | 914 | /* Create the new socket */ |
882 | nsk = iucv_sock_alloc(NULL, SOCK_STREAM, GFP_ATOMIC); | 915 | nsk = iucv_sock_alloc(NULL, SOCK_STREAM, GFP_ATOMIC); |
883 | if (!nsk){ | 916 | if (!nsk) { |
884 | err = iucv_path_sever(path, user_data); | 917 | err = iucv_path_sever(path, user_data); |
885 | goto fail; | 918 | goto fail; |
886 | } | 919 | } |
@@ -903,7 +936,7 @@ static int iucv_callback_connreq(struct iucv_path *path, | |||
903 | 936 | ||
904 | path->msglim = IUCV_QUEUELEN_DEFAULT; | 937 | path->msglim = IUCV_QUEUELEN_DEFAULT; |
905 | err = iucv_path_accept(path, &af_iucv_handler, nuser_data, nsk); | 938 | err = iucv_path_accept(path, &af_iucv_handler, nuser_data, nsk); |
906 | if (err){ | 939 | if (err) { |
907 | err = iucv_path_sever(path, user_data); | 940 | err = iucv_path_sever(path, user_data); |
908 | goto fail; | 941 | goto fail; |
909 | } | 942 | } |
@@ -927,18 +960,53 @@ static void iucv_callback_connack(struct iucv_path *path, u8 ipuser[16]) | |||
927 | sk->sk_state_change(sk); | 960 | sk->sk_state_change(sk); |
928 | } | 961 | } |
929 | 962 | ||
963 | static int iucv_fragment_skb(struct sock *sk, struct sk_buff *skb, int len, | ||
964 | struct sk_buff_head fragmented_skb_q) | ||
965 | { | ||
966 | int dataleft, size, copied = 0; | ||
967 | struct sk_buff *nskb; | ||
968 | |||
969 | dataleft = len; | ||
970 | while (dataleft) { | ||
971 | if (dataleft >= sk->sk_rcvbuf / 4) | ||
972 | size = sk->sk_rcvbuf / 4; | ||
973 | else | ||
974 | size = dataleft; | ||
975 | |||
976 | nskb = alloc_skb(size, GFP_ATOMIC | GFP_DMA); | ||
977 | if (!nskb) | ||
978 | return -ENOMEM; | ||
979 | |||
980 | memcpy(nskb->data, skb->data + copied, size); | ||
981 | copied += size; | ||
982 | dataleft -= size; | ||
983 | |||
984 | nskb->h.raw = nskb->data; | ||
985 | nskb->nh.raw = nskb->data; | ||
986 | nskb->len = size; | ||
987 | |||
988 | skb_queue_tail(fragmented_skb_q, nskb); | ||
989 | } | ||
990 | |||
991 | return 0; | ||
992 | } | ||
993 | |||
930 | static void iucv_callback_rx(struct iucv_path *path, struct iucv_message *msg) | 994 | static void iucv_callback_rx(struct iucv_path *path, struct iucv_message *msg) |
931 | { | 995 | { |
932 | struct sock *sk = path->private; | 996 | struct sock *sk = path->private; |
933 | struct sk_buff *skb; | 997 | struct iucv_sock *iucv = iucv_sk(sk); |
998 | struct sk_buff *skb, *fskb; | ||
999 | struct sk_buff_head fragmented_skb_q; | ||
934 | int rc; | 1000 | int rc; |
935 | 1001 | ||
1002 | skb_queue_head_init(&fragmented_skb_q); | ||
1003 | |||
936 | if (sk->sk_shutdown & RCV_SHUTDOWN) | 1004 | if (sk->sk_shutdown & RCV_SHUTDOWN) |
937 | return; | 1005 | return; |
938 | 1006 | ||
939 | skb = alloc_skb(msg->length, GFP_ATOMIC | GFP_DMA); | 1007 | skb = alloc_skb(msg->length, GFP_ATOMIC | GFP_DMA); |
940 | if (!skb) { | 1008 | if (!skb) { |
941 | iucv_message_reject(path, msg); | 1009 | iucv_path_sever(path, NULL); |
942 | return; | 1010 | return; |
943 | } | 1011 | } |
944 | 1012 | ||
@@ -952,14 +1020,39 @@ static void iucv_callback_rx(struct iucv_path *path, struct iucv_message *msg) | |||
952 | kfree_skb(skb); | 1020 | kfree_skb(skb); |
953 | return; | 1021 | return; |
954 | } | 1022 | } |
1023 | if (skb->truesize >= sk->sk_rcvbuf / 4) { | ||
1024 | rc = iucv_fragment_skb(sk, skb, msg->length, | ||
1025 | &fragmented_skb_q); | ||
1026 | kfree_skb(skb); | ||
1027 | skb = NULL; | ||
1028 | if (rc) { | ||
1029 | iucv_path_sever(path, NULL); | ||
1030 | return; | ||
1031 | } | ||
1032 | } else { | ||
1033 | skb_reset_transport_header(skb); | ||
1034 | skb_reset_network_header(skb); | ||
1035 | skb->len = msg->length; | ||
1036 | } | ||
1037 | } | ||
1038 | /* Queue the fragmented skb */ | ||
1039 | fskb = skb_dequeue(&fragmented_skb_q); | ||
1040 | while (fskb) { | ||
1041 | if (!skb_queue_empty(&iucv->backlog_skb_q)) | ||
1042 | skb_queue_tail(&iucv->backlog_skb_q, fskb); | ||
1043 | else if (sock_queue_rcv_skb(sk, fskb)) | ||
1044 | skb_queue_tail(&iucv_sk(sk)->backlog_skb_q, fskb); | ||
1045 | fskb = skb_dequeue(&fragmented_skb_q); | ||
1046 | } | ||
955 | 1047 | ||
956 | skb_reset_transport_header(skb); | 1048 | /* Queue the original skb if it exists (was not fragmented) */ |
957 | skb_reset_network_header(skb); | 1049 | if (skb) { |
958 | skb->len = msg->length; | 1050 | if (!skb_queue_empty(&iucv->backlog_skb_q)) |
1051 | skb_queue_tail(&iucv_sk(sk)->backlog_skb_q, skb); | ||
1052 | else if (sock_queue_rcv_skb(sk, skb)) | ||
1053 | skb_queue_tail(&iucv_sk(sk)->backlog_skb_q, skb); | ||
959 | } | 1054 | } |
960 | 1055 | ||
961 | if (sock_queue_rcv_skb(sk, skb)) | ||
962 | kfree_skb(skb); | ||
963 | } | 1056 | } |
964 | 1057 | ||
965 | static void iucv_callback_txdone(struct iucv_path *path, | 1058 | static void iucv_callback_txdone(struct iucv_path *path, |
@@ -971,17 +1064,27 @@ static void iucv_callback_txdone(struct iucv_path *path, | |||
971 | struct sk_buff *list_skb = list->next; | 1064 | struct sk_buff *list_skb = list->next; |
972 | unsigned long flags; | 1065 | unsigned long flags; |
973 | 1066 | ||
974 | spin_lock_irqsave(&list->lock, flags); | 1067 | if (list_skb) { |
1068 | spin_lock_irqsave(&list->lock, flags); | ||
1069 | |||
1070 | do { | ||
1071 | this = list_skb; | ||
1072 | list_skb = list_skb->next; | ||
1073 | } while (memcmp(&msg->tag, this->cb, 4) && list_skb); | ||
1074 | |||
1075 | spin_unlock_irqrestore(&list->lock, flags); | ||
975 | 1076 | ||
976 | do { | 1077 | skb_unlink(this, &iucv_sk(sk)->send_skb_q); |
977 | this = list_skb; | 1078 | kfree_skb(this); |
978 | list_skb = list_skb->next; | 1079 | } |
979 | } while (memcmp(&msg->tag, this->cb, 4)); | ||
980 | 1080 | ||
981 | spin_unlock_irqrestore(&list->lock, flags); | 1081 | if (sk->sk_state == IUCV_CLOSING) { |
1082 | if (skb_queue_empty(&iucv_sk(sk)->send_skb_q)) { | ||
1083 | sk->sk_state = IUCV_CLOSED; | ||
1084 | sk->sk_state_change(sk); | ||
1085 | } | ||
1086 | } | ||
982 | 1087 | ||
983 | skb_unlink(this, &iucv_sk(sk)->send_skb_q); | ||
984 | kfree_skb(this); | ||
985 | } | 1088 | } |
986 | 1089 | ||
987 | static void iucv_callback_connrej(struct iucv_path *path, u8 ipuser[16]) | 1090 | static void iucv_callback_connrej(struct iucv_path *path, u8 ipuser[16]) |
@@ -1022,7 +1125,7 @@ static struct net_proto_family iucv_sock_family_ops = { | |||
1022 | .create = iucv_sock_create, | 1125 | .create = iucv_sock_create, |
1023 | }; | 1126 | }; |
1024 | 1127 | ||
1025 | static int afiucv_init(void) | 1128 | static int __init afiucv_init(void) |
1026 | { | 1129 | { |
1027 | int err; | 1130 | int err; |
1028 | 1131 | ||
diff --git a/net/iucv/iucv.c b/net/iucv/iucv.c index 903bdb6eaaa1..fb3faf72e850 100644 --- a/net/iucv/iucv.c +++ b/net/iucv/iucv.c | |||
@@ -32,7 +32,6 @@ | |||
32 | 32 | ||
33 | #include <linux/module.h> | 33 | #include <linux/module.h> |
34 | #include <linux/moduleparam.h> | 34 | #include <linux/moduleparam.h> |
35 | |||
36 | #include <linux/spinlock.h> | 35 | #include <linux/spinlock.h> |
37 | #include <linux/kernel.h> | 36 | #include <linux/kernel.h> |
38 | #include <linux/slab.h> | 37 | #include <linux/slab.h> |
@@ -69,7 +68,7 @@ | |||
69 | #define IUCV_IPNORPY 0x10 | 68 | #define IUCV_IPNORPY 0x10 |
70 | #define IUCV_IPALL 0x80 | 69 | #define IUCV_IPALL 0x80 |
71 | 70 | ||
72 | static int iucv_bus_match (struct device *dev, struct device_driver *drv) | 71 | static int iucv_bus_match(struct device *dev, struct device_driver *drv) |
73 | { | 72 | { |
74 | return 0; | 73 | return 0; |
75 | } | 74 | } |
@@ -78,8 +77,11 @@ struct bus_type iucv_bus = { | |||
78 | .name = "iucv", | 77 | .name = "iucv", |
79 | .match = iucv_bus_match, | 78 | .match = iucv_bus_match, |
80 | }; | 79 | }; |
80 | EXPORT_SYMBOL(iucv_bus); | ||
81 | 81 | ||
82 | struct device *iucv_root; | 82 | struct device *iucv_root; |
83 | EXPORT_SYMBOL(iucv_root); | ||
84 | |||
83 | static int iucv_available; | 85 | static int iucv_available; |
84 | 86 | ||
85 | /* General IUCV interrupt structure */ | 87 | /* General IUCV interrupt structure */ |
@@ -405,7 +407,7 @@ static void iucv_declare_cpu(void *data) | |||
405 | rc = iucv_call_b2f0(IUCV_DECLARE_BUFFER, parm); | 407 | rc = iucv_call_b2f0(IUCV_DECLARE_BUFFER, parm); |
406 | if (rc) { | 408 | if (rc) { |
407 | char *err = "Unknown"; | 409 | char *err = "Unknown"; |
408 | switch(rc) { | 410 | switch (rc) { |
409 | case 0x03: | 411 | case 0x03: |
410 | err = "Directory error"; | 412 | err = "Directory error"; |
411 | break; | 413 | break; |
@@ -588,7 +590,7 @@ static int __cpuinit iucv_cpu_notify(struct notifier_block *self, | |||
588 | return NOTIFY_OK; | 590 | return NOTIFY_OK; |
589 | } | 591 | } |
590 | 592 | ||
591 | static struct notifier_block iucv_cpu_notifier = { | 593 | static struct notifier_block __cpuinitdata iucv_cpu_notifier = { |
592 | .notifier_call = iucv_cpu_notify, | 594 | .notifier_call = iucv_cpu_notify, |
593 | }; | 595 | }; |
594 | 596 | ||
@@ -691,6 +693,7 @@ out_mutex: | |||
691 | mutex_unlock(&iucv_register_mutex); | 693 | mutex_unlock(&iucv_register_mutex); |
692 | return rc; | 694 | return rc; |
693 | } | 695 | } |
696 | EXPORT_SYMBOL(iucv_register); | ||
694 | 697 | ||
695 | /** | 698 | /** |
696 | * iucv_unregister | 699 | * iucv_unregister |
@@ -723,6 +726,7 @@ void iucv_unregister(struct iucv_handler *handler, int smp) | |||
723 | iucv_setmask_mp(); | 726 | iucv_setmask_mp(); |
724 | mutex_unlock(&iucv_register_mutex); | 727 | mutex_unlock(&iucv_register_mutex); |
725 | } | 728 | } |
729 | EXPORT_SYMBOL(iucv_unregister); | ||
726 | 730 | ||
727 | /** | 731 | /** |
728 | * iucv_path_accept | 732 | * iucv_path_accept |
@@ -761,6 +765,7 @@ int iucv_path_accept(struct iucv_path *path, struct iucv_handler *handler, | |||
761 | local_bh_enable(); | 765 | local_bh_enable(); |
762 | return rc; | 766 | return rc; |
763 | } | 767 | } |
768 | EXPORT_SYMBOL(iucv_path_accept); | ||
764 | 769 | ||
765 | /** | 770 | /** |
766 | * iucv_path_connect | 771 | * iucv_path_connect |
@@ -824,6 +829,7 @@ int iucv_path_connect(struct iucv_path *path, struct iucv_handler *handler, | |||
824 | spin_unlock_bh(&iucv_table_lock); | 829 | spin_unlock_bh(&iucv_table_lock); |
825 | return rc; | 830 | return rc; |
826 | } | 831 | } |
832 | EXPORT_SYMBOL(iucv_path_connect); | ||
827 | 833 | ||
828 | /** | 834 | /** |
829 | * iucv_path_quiesce: | 835 | * iucv_path_quiesce: |
@@ -850,6 +856,7 @@ int iucv_path_quiesce(struct iucv_path *path, u8 userdata[16]) | |||
850 | local_bh_enable(); | 856 | local_bh_enable(); |
851 | return rc; | 857 | return rc; |
852 | } | 858 | } |
859 | EXPORT_SYMBOL(iucv_path_quiesce); | ||
853 | 860 | ||
854 | /** | 861 | /** |
855 | * iucv_path_resume: | 862 | * iucv_path_resume: |
@@ -890,7 +897,6 @@ int iucv_path_sever(struct iucv_path *path, u8 userdata[16]) | |||
890 | { | 897 | { |
891 | int rc; | 898 | int rc; |
892 | 899 | ||
893 | |||
894 | preempt_disable(); | 900 | preempt_disable(); |
895 | if (iucv_active_cpu != smp_processor_id()) | 901 | if (iucv_active_cpu != smp_processor_id()) |
896 | spin_lock_bh(&iucv_table_lock); | 902 | spin_lock_bh(&iucv_table_lock); |
@@ -904,6 +910,7 @@ int iucv_path_sever(struct iucv_path *path, u8 userdata[16]) | |||
904 | preempt_enable(); | 910 | preempt_enable(); |
905 | return rc; | 911 | return rc; |
906 | } | 912 | } |
913 | EXPORT_SYMBOL(iucv_path_sever); | ||
907 | 914 | ||
908 | /** | 915 | /** |
909 | * iucv_message_purge | 916 | * iucv_message_purge |
@@ -936,6 +943,7 @@ int iucv_message_purge(struct iucv_path *path, struct iucv_message *msg, | |||
936 | local_bh_enable(); | 943 | local_bh_enable(); |
937 | return rc; | 944 | return rc; |
938 | } | 945 | } |
946 | EXPORT_SYMBOL(iucv_message_purge); | ||
939 | 947 | ||
940 | /** | 948 | /** |
941 | * iucv_message_receive | 949 | * iucv_message_receive |
@@ -1006,6 +1014,7 @@ int iucv_message_receive(struct iucv_path *path, struct iucv_message *msg, | |||
1006 | local_bh_enable(); | 1014 | local_bh_enable(); |
1007 | return rc; | 1015 | return rc; |
1008 | } | 1016 | } |
1017 | EXPORT_SYMBOL(iucv_message_receive); | ||
1009 | 1018 | ||
1010 | /** | 1019 | /** |
1011 | * iucv_message_reject | 1020 | * iucv_message_reject |
@@ -1034,6 +1043,7 @@ int iucv_message_reject(struct iucv_path *path, struct iucv_message *msg) | |||
1034 | local_bh_enable(); | 1043 | local_bh_enable(); |
1035 | return rc; | 1044 | return rc; |
1036 | } | 1045 | } |
1046 | EXPORT_SYMBOL(iucv_message_reject); | ||
1037 | 1047 | ||
1038 | /** | 1048 | /** |
1039 | * iucv_message_reply | 1049 | * iucv_message_reply |
@@ -1077,6 +1087,7 @@ int iucv_message_reply(struct iucv_path *path, struct iucv_message *msg, | |||
1077 | local_bh_enable(); | 1087 | local_bh_enable(); |
1078 | return rc; | 1088 | return rc; |
1079 | } | 1089 | } |
1090 | EXPORT_SYMBOL(iucv_message_reply); | ||
1080 | 1091 | ||
1081 | /** | 1092 | /** |
1082 | * iucv_message_send | 1093 | * iucv_message_send |
@@ -1125,6 +1136,7 @@ int iucv_message_send(struct iucv_path *path, struct iucv_message *msg, | |||
1125 | local_bh_enable(); | 1136 | local_bh_enable(); |
1126 | return rc; | 1137 | return rc; |
1127 | } | 1138 | } |
1139 | EXPORT_SYMBOL(iucv_message_send); | ||
1128 | 1140 | ||
1129 | /** | 1141 | /** |
1130 | * iucv_message_send2way | 1142 | * iucv_message_send2way |
@@ -1181,6 +1193,7 @@ int iucv_message_send2way(struct iucv_path *path, struct iucv_message *msg, | |||
1181 | local_bh_enable(); | 1193 | local_bh_enable(); |
1182 | return rc; | 1194 | return rc; |
1183 | } | 1195 | } |
1196 | EXPORT_SYMBOL(iucv_message_send2way); | ||
1184 | 1197 | ||
1185 | /** | 1198 | /** |
1186 | * iucv_path_pending | 1199 | * iucv_path_pending |
@@ -1572,7 +1585,7 @@ static void iucv_external_interrupt(u16 code) | |||
1572 | * | 1585 | * |
1573 | * Allocates and initializes various data structures. | 1586 | * Allocates and initializes various data structures. |
1574 | */ | 1587 | */ |
1575 | static int iucv_init(void) | 1588 | static int __init iucv_init(void) |
1576 | { | 1589 | { |
1577 | int rc; | 1590 | int rc; |
1578 | 1591 | ||
@@ -1583,7 +1596,7 @@ static int iucv_init(void) | |||
1583 | rc = iucv_query_maxconn(); | 1596 | rc = iucv_query_maxconn(); |
1584 | if (rc) | 1597 | if (rc) |
1585 | goto out; | 1598 | goto out; |
1586 | rc = register_external_interrupt (0x4000, iucv_external_interrupt); | 1599 | rc = register_external_interrupt(0x4000, iucv_external_interrupt); |
1587 | if (rc) | 1600 | if (rc) |
1588 | goto out; | 1601 | goto out; |
1589 | rc = bus_register(&iucv_bus); | 1602 | rc = bus_register(&iucv_bus); |
@@ -1594,7 +1607,7 @@ static int iucv_init(void) | |||
1594 | rc = PTR_ERR(iucv_root); | 1607 | rc = PTR_ERR(iucv_root); |
1595 | goto out_bus; | 1608 | goto out_bus; |
1596 | } | 1609 | } |
1597 | /* Note: GFP_DMA used used to get memory below 2G */ | 1610 | /* Note: GFP_DMA used to get memory below 2G */ |
1598 | iucv_irq_data = percpu_alloc(sizeof(struct iucv_irq_data), | 1611 | iucv_irq_data = percpu_alloc(sizeof(struct iucv_irq_data), |
1599 | GFP_KERNEL|GFP_DMA); | 1612 | GFP_KERNEL|GFP_DMA); |
1600 | if (!iucv_irq_data) { | 1613 | if (!iucv_irq_data) { |
@@ -1632,7 +1645,7 @@ out: | |||
1632 | * | 1645 | * |
1633 | * Frees everything allocated from iucv_init. | 1646 | * Frees everything allocated from iucv_init. |
1634 | */ | 1647 | */ |
1635 | static void iucv_exit(void) | 1648 | static void __exit iucv_exit(void) |
1636 | { | 1649 | { |
1637 | struct iucv_irq_list *p, *n; | 1650 | struct iucv_irq_list *p, *n; |
1638 | 1651 | ||
@@ -1653,24 +1666,6 @@ static void iucv_exit(void) | |||
1653 | subsys_initcall(iucv_init); | 1666 | subsys_initcall(iucv_init); |
1654 | module_exit(iucv_exit); | 1667 | module_exit(iucv_exit); |
1655 | 1668 | ||
1656 | /** | ||
1657 | * Export all public stuff | ||
1658 | */ | ||
1659 | EXPORT_SYMBOL (iucv_bus); | ||
1660 | EXPORT_SYMBOL (iucv_root); | ||
1661 | EXPORT_SYMBOL (iucv_register); | ||
1662 | EXPORT_SYMBOL (iucv_unregister); | ||
1663 | EXPORT_SYMBOL (iucv_path_accept); | ||
1664 | EXPORT_SYMBOL (iucv_path_connect); | ||
1665 | EXPORT_SYMBOL (iucv_path_quiesce); | ||
1666 | EXPORT_SYMBOL (iucv_path_sever); | ||
1667 | EXPORT_SYMBOL (iucv_message_purge); | ||
1668 | EXPORT_SYMBOL (iucv_message_receive); | ||
1669 | EXPORT_SYMBOL (iucv_message_reject); | ||
1670 | EXPORT_SYMBOL (iucv_message_reply); | ||
1671 | EXPORT_SYMBOL (iucv_message_send); | ||
1672 | EXPORT_SYMBOL (iucv_message_send2way); | ||
1673 | |||
1674 | MODULE_AUTHOR("(C) 2001 IBM Corp. by Fritz Elfert (felfert@millenux.com)"); | 1669 | MODULE_AUTHOR("(C) 2001 IBM Corp. by Fritz Elfert (felfert@millenux.com)"); |
1675 | MODULE_DESCRIPTION("Linux for S/390 IUCV lowlevel driver"); | 1670 | MODULE_DESCRIPTION("Linux for S/390 IUCV lowlevel driver"); |
1676 | MODULE_LICENSE("GPL"); | 1671 | MODULE_LICENSE("GPL"); |
diff --git a/net/llc/llc_core.c b/net/llc/llc_core.c index d12413cff5bd..d4b13a031fd5 100644 --- a/net/llc/llc_core.c +++ b/net/llc/llc_core.c | |||
@@ -160,8 +160,14 @@ static struct packet_type llc_tr_packet_type = { | |||
160 | 160 | ||
161 | static int __init llc_init(void) | 161 | static int __init llc_init(void) |
162 | { | 162 | { |
163 | if (dev_base->next) | 163 | struct net_device *dev; |
164 | memcpy(llc_station_mac_sa, dev_base->next->dev_addr, ETH_ALEN); | 164 | |
165 | dev = first_net_device(); | ||
166 | if (dev != NULL) | ||
167 | dev = next_net_device(dev); | ||
168 | |||
169 | if (dev != NULL) | ||
170 | memcpy(llc_station_mac_sa, dev->dev_addr, ETH_ALEN); | ||
165 | else | 171 | else |
166 | memset(llc_station_mac_sa, 0, ETH_ALEN); | 172 | memset(llc_station_mac_sa, 0, ETH_ALEN); |
167 | dev_add_pack(&llc_packet_type); | 173 | dev_add_pack(&llc_packet_type); |
diff --git a/net/netlink/af_netlink.c b/net/netlink/af_netlink.c index 42d2fb94eff1..507828d7d4ae 100644 --- a/net/netlink/af_netlink.c +++ b/net/netlink/af_netlink.c | |||
@@ -140,6 +140,14 @@ static struct hlist_head *nl_pid_hashfn(struct nl_pid_hash *hash, u32 pid) | |||
140 | 140 | ||
141 | static void netlink_sock_destruct(struct sock *sk) | 141 | static void netlink_sock_destruct(struct sock *sk) |
142 | { | 142 | { |
143 | struct netlink_sock *nlk = nlk_sk(sk); | ||
144 | |||
145 | if (nlk->cb) { | ||
146 | if (nlk->cb->done) | ||
147 | nlk->cb->done(nlk->cb); | ||
148 | netlink_destroy_callback(nlk->cb); | ||
149 | } | ||
150 | |||
143 | skb_queue_purge(&sk->sk_receive_queue); | 151 | skb_queue_purge(&sk->sk_receive_queue); |
144 | 152 | ||
145 | if (!sock_flag(sk, SOCK_DEAD)) { | 153 | if (!sock_flag(sk, SOCK_DEAD)) { |
@@ -148,7 +156,6 @@ static void netlink_sock_destruct(struct sock *sk) | |||
148 | } | 156 | } |
149 | BUG_TRAP(!atomic_read(&sk->sk_rmem_alloc)); | 157 | BUG_TRAP(!atomic_read(&sk->sk_rmem_alloc)); |
150 | BUG_TRAP(!atomic_read(&sk->sk_wmem_alloc)); | 158 | BUG_TRAP(!atomic_read(&sk->sk_wmem_alloc)); |
151 | BUG_TRAP(!nlk_sk(sk)->cb); | ||
152 | BUG_TRAP(!nlk_sk(sk)->groups); | 159 | BUG_TRAP(!nlk_sk(sk)->groups); |
153 | } | 160 | } |
154 | 161 | ||
@@ -456,17 +463,10 @@ static int netlink_release(struct socket *sock) | |||
456 | sock_orphan(sk); | 463 | sock_orphan(sk); |
457 | nlk = nlk_sk(sk); | 464 | nlk = nlk_sk(sk); |
458 | 465 | ||
459 | mutex_lock(nlk->cb_mutex); | 466 | /* |
460 | if (nlk->cb) { | 467 | * OK. Socket is unlinked, any packets that arrive now |
461 | if (nlk->cb->done) | 468 | * will be purged. |
462 | nlk->cb->done(nlk->cb); | 469 | */ |
463 | netlink_destroy_callback(nlk->cb); | ||
464 | nlk->cb = NULL; | ||
465 | } | ||
466 | mutex_unlock(nlk->cb_mutex); | ||
467 | |||
468 | /* OK. Socket is unlinked, and, therefore, | ||
469 | no new packets will arrive */ | ||
470 | 470 | ||
471 | sock->sk = NULL; | 471 | sock->sk = NULL; |
472 | wake_up_interruptible_all(&nlk->wait); | 472 | wake_up_interruptible_all(&nlk->wait); |
@@ -1245,16 +1245,14 @@ static int netlink_recvmsg(struct kiocb *kiocb, struct socket *sock, | |||
1245 | siocb->scm = &scm; | 1245 | siocb->scm = &scm; |
1246 | } | 1246 | } |
1247 | siocb->scm->creds = *NETLINK_CREDS(skb); | 1247 | siocb->scm->creds = *NETLINK_CREDS(skb); |
1248 | if (flags & MSG_TRUNC) | ||
1249 | copied = skb->len; | ||
1248 | skb_free_datagram(sk, skb); | 1250 | skb_free_datagram(sk, skb); |
1249 | 1251 | ||
1250 | if (nlk->cb && atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf / 2) | 1252 | if (nlk->cb && atomic_read(&sk->sk_rmem_alloc) <= sk->sk_rcvbuf / 2) |
1251 | netlink_dump(sk); | 1253 | netlink_dump(sk); |
1252 | 1254 | ||
1253 | scm_recv(sock, msg, siocb->scm, flags); | 1255 | scm_recv(sock, msg, siocb->scm, flags); |
1254 | |||
1255 | if (flags & MSG_TRUNC) | ||
1256 | copied = skb->len; | ||
1257 | |||
1258 | out: | 1256 | out: |
1259 | netlink_rcv_wake(sk); | 1257 | netlink_rcv_wake(sk); |
1260 | return err ? : copied; | 1258 | return err ? : copied; |
@@ -1426,9 +1424,9 @@ int netlink_dump_start(struct sock *ssk, struct sk_buff *skb, | |||
1426 | return -ECONNREFUSED; | 1424 | return -ECONNREFUSED; |
1427 | } | 1425 | } |
1428 | nlk = nlk_sk(sk); | 1426 | nlk = nlk_sk(sk); |
1429 | /* A dump or destruction is in progress... */ | 1427 | /* A dump is in progress... */ |
1430 | mutex_lock(nlk->cb_mutex); | 1428 | mutex_lock(nlk->cb_mutex); |
1431 | if (nlk->cb || sock_flag(sk, SOCK_DEAD)) { | 1429 | if (nlk->cb) { |
1432 | mutex_unlock(nlk->cb_mutex); | 1430 | mutex_unlock(nlk->cb_mutex); |
1433 | netlink_destroy_callback(cb); | 1431 | netlink_destroy_callback(cb); |
1434 | sock_put(sk); | 1432 | sock_put(sk); |
diff --git a/net/netrom/nr_route.c b/net/netrom/nr_route.c index 8e6bd4e9d82c..2f76e062609d 100644 --- a/net/netrom/nr_route.c +++ b/net/netrom/nr_route.c | |||
@@ -598,7 +598,7 @@ struct net_device *nr_dev_first(void) | |||
598 | struct net_device *dev, *first = NULL; | 598 | struct net_device *dev, *first = NULL; |
599 | 599 | ||
600 | read_lock(&dev_base_lock); | 600 | read_lock(&dev_base_lock); |
601 | for (dev = dev_base; dev != NULL; dev = dev->next) { | 601 | for_each_netdev(dev) { |
602 | if ((dev->flags & IFF_UP) && dev->type == ARPHRD_NETROM) | 602 | if ((dev->flags & IFF_UP) && dev->type == ARPHRD_NETROM) |
603 | if (first == NULL || strncmp(dev->name, first->name, 3) < 0) | 603 | if (first == NULL || strncmp(dev->name, first->name, 3) < 0) |
604 | first = dev; | 604 | first = dev; |
@@ -618,12 +618,13 @@ struct net_device *nr_dev_get(ax25_address *addr) | |||
618 | struct net_device *dev; | 618 | struct net_device *dev; |
619 | 619 | ||
620 | read_lock(&dev_base_lock); | 620 | read_lock(&dev_base_lock); |
621 | for (dev = dev_base; dev != NULL; dev = dev->next) { | 621 | for_each_netdev(dev) { |
622 | if ((dev->flags & IFF_UP) && dev->type == ARPHRD_NETROM && ax25cmp(addr, (ax25_address *)dev->dev_addr) == 0) { | 622 | if ((dev->flags & IFF_UP) && dev->type == ARPHRD_NETROM && ax25cmp(addr, (ax25_address *)dev->dev_addr) == 0) { |
623 | dev_hold(dev); | 623 | dev_hold(dev); |
624 | goto out; | 624 | goto out; |
625 | } | 625 | } |
626 | } | 626 | } |
627 | dev = NULL; | ||
627 | out: | 628 | out: |
628 | read_unlock(&dev_base_lock); | 629 | read_unlock(&dev_base_lock); |
629 | return dev; | 630 | return dev; |
diff --git a/net/rose/rose_route.c b/net/rose/rose_route.c index 1f9aefd95a99..929a784a86d7 100644 --- a/net/rose/rose_route.c +++ b/net/rose/rose_route.c | |||
@@ -596,7 +596,7 @@ struct net_device *rose_dev_first(void) | |||
596 | struct net_device *dev, *first = NULL; | 596 | struct net_device *dev, *first = NULL; |
597 | 597 | ||
598 | read_lock(&dev_base_lock); | 598 | read_lock(&dev_base_lock); |
599 | for (dev = dev_base; dev != NULL; dev = dev->next) { | 599 | for_each_netdev(dev) { |
600 | if ((dev->flags & IFF_UP) && dev->type == ARPHRD_ROSE) | 600 | if ((dev->flags & IFF_UP) && dev->type == ARPHRD_ROSE) |
601 | if (first == NULL || strncmp(dev->name, first->name, 3) < 0) | 601 | if (first == NULL || strncmp(dev->name, first->name, 3) < 0) |
602 | first = dev; | 602 | first = dev; |
@@ -614,12 +614,13 @@ struct net_device *rose_dev_get(rose_address *addr) | |||
614 | struct net_device *dev; | 614 | struct net_device *dev; |
615 | 615 | ||
616 | read_lock(&dev_base_lock); | 616 | read_lock(&dev_base_lock); |
617 | for (dev = dev_base; dev != NULL; dev = dev->next) { | 617 | for_each_netdev(dev) { |
618 | if ((dev->flags & IFF_UP) && dev->type == ARPHRD_ROSE && rosecmp(addr, (rose_address *)dev->dev_addr) == 0) { | 618 | if ((dev->flags & IFF_UP) && dev->type == ARPHRD_ROSE && rosecmp(addr, (rose_address *)dev->dev_addr) == 0) { |
619 | dev_hold(dev); | 619 | dev_hold(dev); |
620 | goto out; | 620 | goto out; |
621 | } | 621 | } |
622 | } | 622 | } |
623 | dev = NULL; | ||
623 | out: | 624 | out: |
624 | read_unlock(&dev_base_lock); | 625 | read_unlock(&dev_base_lock); |
625 | return dev; | 626 | return dev; |
@@ -630,10 +631,11 @@ static int rose_dev_exists(rose_address *addr) | |||
630 | struct net_device *dev; | 631 | struct net_device *dev; |
631 | 632 | ||
632 | read_lock(&dev_base_lock); | 633 | read_lock(&dev_base_lock); |
633 | for (dev = dev_base; dev != NULL; dev = dev->next) { | 634 | for_each_netdev(dev) { |
634 | if ((dev->flags & IFF_UP) && dev->type == ARPHRD_ROSE && rosecmp(addr, (rose_address *)dev->dev_addr) == 0) | 635 | if ((dev->flags & IFF_UP) && dev->type == ARPHRD_ROSE && rosecmp(addr, (rose_address *)dev->dev_addr) == 0) |
635 | goto out; | 636 | goto out; |
636 | } | 637 | } |
638 | dev = NULL; | ||
637 | out: | 639 | out: |
638 | read_unlock(&dev_base_lock); | 640 | read_unlock(&dev_base_lock); |
639 | return dev != NULL; | 641 | return dev != NULL; |
diff --git a/net/rxrpc/Kconfig b/net/rxrpc/Kconfig index 8750f6da6bc7..91b3d52f6f1a 100644 --- a/net/rxrpc/Kconfig +++ b/net/rxrpc/Kconfig | |||
@@ -5,6 +5,7 @@ | |||
5 | config AF_RXRPC | 5 | config AF_RXRPC |
6 | tristate "RxRPC session sockets" | 6 | tristate "RxRPC session sockets" |
7 | depends on EXPERIMENTAL | 7 | depends on EXPERIMENTAL |
8 | select KEYS | ||
8 | help | 9 | help |
9 | Say Y or M here to include support for RxRPC session sockets (just | 10 | Say Y or M here to include support for RxRPC session sockets (just |
10 | the transport part, not the presentation part: (un)marshalling is | 11 | the transport part, not the presentation part: (un)marshalling is |
@@ -29,7 +30,7 @@ config AF_RXRPC_DEBUG | |||
29 | 30 | ||
30 | config RXKAD | 31 | config RXKAD |
31 | tristate "RxRPC Kerberos security" | 32 | tristate "RxRPC Kerberos security" |
32 | depends on AF_RXRPC && KEYS | 33 | depends on AF_RXRPC |
33 | select CRYPTO | 34 | select CRYPTO |
34 | select CRYPTO_MANAGER | 35 | select CRYPTO_MANAGER |
35 | select CRYPTO_BLKCIPHER | 36 | select CRYPTO_BLKCIPHER |
diff --git a/net/rxrpc/ar-ack.c b/net/rxrpc/ar-ack.c index fc07a926df56..657ee69f2133 100644 --- a/net/rxrpc/ar-ack.c +++ b/net/rxrpc/ar-ack.c | |||
@@ -543,6 +543,38 @@ static void rxrpc_zap_tx_window(struct rxrpc_call *call) | |||
543 | } | 543 | } |
544 | 544 | ||
545 | /* | 545 | /* |
546 | * process the extra information that may be appended to an ACK packet | ||
547 | */ | ||
548 | static void rxrpc_extract_ackinfo(struct rxrpc_call *call, struct sk_buff *skb, | ||
549 | unsigned latest, int nAcks) | ||
550 | { | ||
551 | struct rxrpc_ackinfo ackinfo; | ||
552 | struct rxrpc_peer *peer; | ||
553 | unsigned mtu; | ||
554 | |||
555 | if (skb_copy_bits(skb, nAcks + 3, &ackinfo, sizeof(ackinfo)) < 0) { | ||
556 | _leave(" [no ackinfo]"); | ||
557 | return; | ||
558 | } | ||
559 | |||
560 | _proto("Rx ACK %%%u Info { rx=%u max=%u rwin=%u jm=%u }", | ||
561 | latest, | ||
562 | ntohl(ackinfo.rxMTU), ntohl(ackinfo.maxMTU), | ||
563 | ntohl(ackinfo.rwind), ntohl(ackinfo.jumbo_max)); | ||
564 | |||
565 | mtu = min(ntohl(ackinfo.rxMTU), ntohl(ackinfo.maxMTU)); | ||
566 | |||
567 | peer = call->conn->trans->peer; | ||
568 | if (mtu < peer->maxdata) { | ||
569 | spin_lock_bh(&peer->lock); | ||
570 | peer->maxdata = mtu; | ||
571 | peer->mtu = mtu + peer->hdrsize; | ||
572 | spin_unlock_bh(&peer->lock); | ||
573 | _net("Net MTU %u (maxdata %u)", peer->mtu, peer->maxdata); | ||
574 | } | ||
575 | } | ||
576 | |||
577 | /* | ||
546 | * process packets in the reception queue | 578 | * process packets in the reception queue |
547 | */ | 579 | */ |
548 | static int rxrpc_process_rx_queue(struct rxrpc_call *call, | 580 | static int rxrpc_process_rx_queue(struct rxrpc_call *call, |
@@ -606,6 +638,8 @@ process_further: | |||
606 | rxrpc_acks[ack.reason], | 638 | rxrpc_acks[ack.reason], |
607 | ack.nAcks); | 639 | ack.nAcks); |
608 | 640 | ||
641 | rxrpc_extract_ackinfo(call, skb, latest, ack.nAcks); | ||
642 | |||
609 | if (ack.reason == RXRPC_ACK_PING) { | 643 | if (ack.reason == RXRPC_ACK_PING) { |
610 | _proto("Rx ACK %%%u PING Request", latest); | 644 | _proto("Rx ACK %%%u PING Request", latest); |
611 | rxrpc_propose_ACK(call, RXRPC_ACK_PING_RESPONSE, | 645 | rxrpc_propose_ACK(call, RXRPC_ACK_PING_RESPONSE, |
@@ -801,9 +835,9 @@ void rxrpc_process_call(struct work_struct *work) | |||
801 | struct msghdr msg; | 835 | struct msghdr msg; |
802 | struct kvec iov[5]; | 836 | struct kvec iov[5]; |
803 | unsigned long bits; | 837 | unsigned long bits; |
804 | __be32 data; | 838 | __be32 data, pad; |
805 | size_t len; | 839 | size_t len; |
806 | int genbit, loop, nbit, ioc, ret; | 840 | int genbit, loop, nbit, ioc, ret, mtu; |
807 | u32 abort_code = RX_PROTOCOL_ERROR; | 841 | u32 abort_code = RX_PROTOCOL_ERROR; |
808 | u8 *acks = NULL; | 842 | u8 *acks = NULL; |
809 | 843 | ||
@@ -899,9 +933,30 @@ void rxrpc_process_call(struct work_struct *work) | |||
899 | } | 933 | } |
900 | 934 | ||
901 | if (test_bit(RXRPC_CALL_ACK_FINAL, &call->events)) { | 935 | if (test_bit(RXRPC_CALL_ACK_FINAL, &call->events)) { |
902 | hdr.type = RXRPC_PACKET_TYPE_ACKALL; | ||
903 | genbit = RXRPC_CALL_ACK_FINAL; | 936 | genbit = RXRPC_CALL_ACK_FINAL; |
904 | goto send_message; | 937 | |
938 | ack.bufferSpace = htons(8); | ||
939 | ack.maxSkew = 0; | ||
940 | ack.serial = 0; | ||
941 | ack.reason = RXRPC_ACK_IDLE; | ||
942 | ack.nAcks = 0; | ||
943 | call->ackr_reason = 0; | ||
944 | |||
945 | spin_lock_bh(&call->lock); | ||
946 | ack.serial = call->ackr_serial; | ||
947 | ack.previousPacket = call->ackr_prev_seq; | ||
948 | ack.firstPacket = htonl(call->rx_data_eaten + 1); | ||
949 | spin_unlock_bh(&call->lock); | ||
950 | |||
951 | pad = 0; | ||
952 | |||
953 | iov[1].iov_base = &ack; | ||
954 | iov[1].iov_len = sizeof(ack); | ||
955 | iov[2].iov_base = &pad; | ||
956 | iov[2].iov_len = 3; | ||
957 | iov[3].iov_base = &ackinfo; | ||
958 | iov[3].iov_len = sizeof(ackinfo); | ||
959 | goto send_ACK; | ||
905 | } | 960 | } |
906 | 961 | ||
907 | if (call->events & ((1 << RXRPC_CALL_RCVD_BUSY) | | 962 | if (call->events & ((1 << RXRPC_CALL_RCVD_BUSY) | |
@@ -971,8 +1026,6 @@ void rxrpc_process_call(struct work_struct *work) | |||
971 | 1026 | ||
972 | /* consider sending an ordinary ACK */ | 1027 | /* consider sending an ordinary ACK */ |
973 | if (test_bit(RXRPC_CALL_ACK, &call->events)) { | 1028 | if (test_bit(RXRPC_CALL_ACK, &call->events)) { |
974 | __be32 pad; | ||
975 | |||
976 | _debug("send ACK: window: %d - %d { %lx }", | 1029 | _debug("send ACK: window: %d - %d { %lx }", |
977 | call->rx_data_eaten, call->ackr_win_top, | 1030 | call->rx_data_eaten, call->ackr_win_top, |
978 | call->ackr_window[0]); | 1031 | call->ackr_window[0]); |
@@ -997,12 +1050,6 @@ void rxrpc_process_call(struct work_struct *work) | |||
997 | ack.serial = 0; | 1050 | ack.serial = 0; |
998 | ack.reason = 0; | 1051 | ack.reason = 0; |
999 | 1052 | ||
1000 | ackinfo.rxMTU = htonl(5692); | ||
1001 | // ackinfo.rxMTU = htonl(call->conn->trans->peer->maxdata); | ||
1002 | ackinfo.maxMTU = htonl(call->conn->trans->peer->maxdata); | ||
1003 | ackinfo.rwind = htonl(32); | ||
1004 | ackinfo.jumbo_max = htonl(4); | ||
1005 | |||
1006 | spin_lock_bh(&call->lock); | 1053 | spin_lock_bh(&call->lock); |
1007 | ack.reason = call->ackr_reason; | 1054 | ack.reason = call->ackr_reason; |
1008 | ack.serial = call->ackr_serial; | 1055 | ack.serial = call->ackr_serial; |
@@ -1116,6 +1163,15 @@ send_ACK_with_skew: | |||
1116 | ack.maxSkew = htons(atomic_read(&call->conn->hi_serial) - | 1163 | ack.maxSkew = htons(atomic_read(&call->conn->hi_serial) - |
1117 | ntohl(ack.serial)); | 1164 | ntohl(ack.serial)); |
1118 | send_ACK: | 1165 | send_ACK: |
1166 | mtu = call->conn->trans->peer->if_mtu; | ||
1167 | mtu -= call->conn->trans->peer->hdrsize; | ||
1168 | ackinfo.maxMTU = htonl(mtu); | ||
1169 | ackinfo.rwind = htonl(32); | ||
1170 | |||
1171 | /* permit the peer to send us jumbo packets if it wants to */ | ||
1172 | ackinfo.rxMTU = htonl(5692); | ||
1173 | ackinfo.jumbo_max = htonl(4); | ||
1174 | |||
1119 | hdr.serial = htonl(atomic_inc_return(&call->conn->serial)); | 1175 | hdr.serial = htonl(atomic_inc_return(&call->conn->serial)); |
1120 | _proto("Tx ACK %%%u { m=%hu f=#%u p=#%u s=%%%u r=%s n=%u }", | 1176 | _proto("Tx ACK %%%u { m=%hu f=#%u p=#%u s=%%%u r=%s n=%u }", |
1121 | ntohl(hdr.serial), | 1177 | ntohl(hdr.serial), |
diff --git a/net/rxrpc/ar-error.c b/net/rxrpc/ar-error.c index 2c27df1ffa17..6cb3e8890e7e 100644 --- a/net/rxrpc/ar-error.c +++ b/net/rxrpc/ar-error.c | |||
@@ -100,8 +100,10 @@ void rxrpc_UDP_error_report(struct sock *sk) | |||
100 | } | 100 | } |
101 | 101 | ||
102 | if (mtu < peer->mtu) { | 102 | if (mtu < peer->mtu) { |
103 | spin_lock_bh(&peer->lock); | ||
103 | peer->mtu = mtu; | 104 | peer->mtu = mtu; |
104 | peer->maxdata = peer->mtu - peer->hdrsize; | 105 | peer->maxdata = peer->mtu - peer->hdrsize; |
106 | spin_unlock_bh(&peer->lock); | ||
105 | _net("Net MTU %u (maxdata %u)", | 107 | _net("Net MTU %u (maxdata %u)", |
106 | peer->mtu, peer->maxdata); | 108 | peer->mtu, peer->maxdata); |
107 | } | 109 | } |
diff --git a/net/rxrpc/ar-output.c b/net/rxrpc/ar-output.c index 5cdde4a48ed1..591c4422205e 100644 --- a/net/rxrpc/ar-output.c +++ b/net/rxrpc/ar-output.c | |||
@@ -582,7 +582,7 @@ static int rxrpc_send_data(struct kiocb *iocb, | |||
582 | max &= ~(call->conn->size_align - 1UL); | 582 | max &= ~(call->conn->size_align - 1UL); |
583 | 583 | ||
584 | chunk = max; | 584 | chunk = max; |
585 | if (chunk > len) | 585 | if (chunk > len && !more) |
586 | chunk = len; | 586 | chunk = len; |
587 | 587 | ||
588 | space = chunk + call->conn->size_align; | 588 | space = chunk + call->conn->size_align; |
diff --git a/net/rxrpc/ar-peer.c b/net/rxrpc/ar-peer.c index d399de4a7fe2..ce08b78647ce 100644 --- a/net/rxrpc/ar-peer.c +++ b/net/rxrpc/ar-peer.c | |||
@@ -19,6 +19,7 @@ | |||
19 | #include <net/sock.h> | 19 | #include <net/sock.h> |
20 | #include <net/af_rxrpc.h> | 20 | #include <net/af_rxrpc.h> |
21 | #include <net/ip.h> | 21 | #include <net/ip.h> |
22 | #include <net/route.h> | ||
22 | #include "ar-internal.h" | 23 | #include "ar-internal.h" |
23 | 24 | ||
24 | static LIST_HEAD(rxrpc_peers); | 25 | static LIST_HEAD(rxrpc_peers); |
@@ -28,6 +29,47 @@ static DECLARE_WAIT_QUEUE_HEAD(rxrpc_peer_wq); | |||
28 | static void rxrpc_destroy_peer(struct work_struct *work); | 29 | static void rxrpc_destroy_peer(struct work_struct *work); |
29 | 30 | ||
30 | /* | 31 | /* |
32 | * assess the MTU size for the network interface through which this peer is | ||
33 | * reached | ||
34 | */ | ||
35 | static void rxrpc_assess_MTU_size(struct rxrpc_peer *peer) | ||
36 | { | ||
37 | struct rtable *rt; | ||
38 | struct flowi fl; | ||
39 | int ret; | ||
40 | |||
41 | peer->if_mtu = 1500; | ||
42 | |||
43 | memset(&fl, 0, sizeof(fl)); | ||
44 | |||
45 | switch (peer->srx.transport.family) { | ||
46 | case AF_INET: | ||
47 | fl.oif = 0; | ||
48 | fl.proto = IPPROTO_UDP, | ||
49 | fl.nl_u.ip4_u.saddr = 0; | ||
50 | fl.nl_u.ip4_u.daddr = peer->srx.transport.sin.sin_addr.s_addr; | ||
51 | fl.nl_u.ip4_u.tos = 0; | ||
52 | /* assume AFS.CM talking to AFS.FS */ | ||
53 | fl.uli_u.ports.sport = htons(7001); | ||
54 | fl.uli_u.ports.dport = htons(7000); | ||
55 | break; | ||
56 | default: | ||
57 | BUG(); | ||
58 | } | ||
59 | |||
60 | ret = ip_route_output_key(&rt, &fl); | ||
61 | if (ret < 0) { | ||
62 | kleave(" [route err %d]", ret); | ||
63 | return; | ||
64 | } | ||
65 | |||
66 | peer->if_mtu = dst_mtu(&rt->u.dst); | ||
67 | dst_release(&rt->u.dst); | ||
68 | |||
69 | kleave(" [if_mtu %u]", peer->if_mtu); | ||
70 | } | ||
71 | |||
72 | /* | ||
31 | * allocate a new peer | 73 | * allocate a new peer |
32 | */ | 74 | */ |
33 | static struct rxrpc_peer *rxrpc_alloc_peer(struct sockaddr_rxrpc *srx, | 75 | static struct rxrpc_peer *rxrpc_alloc_peer(struct sockaddr_rxrpc *srx, |
@@ -47,7 +89,8 @@ static struct rxrpc_peer *rxrpc_alloc_peer(struct sockaddr_rxrpc *srx, | |||
47 | peer->debug_id = atomic_inc_return(&rxrpc_debug_id); | 89 | peer->debug_id = atomic_inc_return(&rxrpc_debug_id); |
48 | memcpy(&peer->srx, srx, sizeof(*srx)); | 90 | memcpy(&peer->srx, srx, sizeof(*srx)); |
49 | 91 | ||
50 | peer->mtu = peer->if_mtu = 65535; | 92 | rxrpc_assess_MTU_size(peer); |
93 | peer->mtu = peer->if_mtu; | ||
51 | 94 | ||
52 | if (srx->transport.family == AF_INET) { | 95 | if (srx->transport.family == AF_INET) { |
53 | peer->hdrsize = sizeof(struct iphdr); | 96 | peer->hdrsize = sizeof(struct iphdr); |
diff --git a/net/sched/sch_api.c b/net/sched/sch_api.c index 8699e7006d80..bec600af03ca 100644 --- a/net/sched/sch_api.c +++ b/net/sched/sch_api.c | |||
@@ -894,9 +894,10 @@ static int tc_dump_qdisc(struct sk_buff *skb, struct netlink_callback *cb) | |||
894 | s_idx = cb->args[0]; | 894 | s_idx = cb->args[0]; |
895 | s_q_idx = q_idx = cb->args[1]; | 895 | s_q_idx = q_idx = cb->args[1]; |
896 | read_lock(&dev_base_lock); | 896 | read_lock(&dev_base_lock); |
897 | for (dev=dev_base, idx=0; dev; dev = dev->next, idx++) { | 897 | idx = 0; |
898 | for_each_netdev(dev) { | ||
898 | if (idx < s_idx) | 899 | if (idx < s_idx) |
899 | continue; | 900 | goto cont; |
900 | if (idx > s_idx) | 901 | if (idx > s_idx) |
901 | s_q_idx = 0; | 902 | s_q_idx = 0; |
902 | q_idx = 0; | 903 | q_idx = 0; |
@@ -910,6 +911,8 @@ static int tc_dump_qdisc(struct sk_buff *skb, struct netlink_callback *cb) | |||
910 | goto done; | 911 | goto done; |
911 | q_idx++; | 912 | q_idx++; |
912 | } | 913 | } |
914 | cont: | ||
915 | idx++; | ||
913 | } | 916 | } |
914 | 917 | ||
915 | done: | 918 | done: |
diff --git a/net/sctp/associola.c b/net/sctp/associola.c index db73ef97485a..df94e3cdfba3 100644 --- a/net/sctp/associola.c +++ b/net/sctp/associola.c | |||
@@ -1103,6 +1103,13 @@ void sctp_assoc_update(struct sctp_association *asoc, | |||
1103 | asoc->ssnmap = new->ssnmap; | 1103 | asoc->ssnmap = new->ssnmap; |
1104 | new->ssnmap = NULL; | 1104 | new->ssnmap = NULL; |
1105 | } | 1105 | } |
1106 | |||
1107 | if (!asoc->assoc_id) { | ||
1108 | /* get a new association id since we don't have one | ||
1109 | * yet. | ||
1110 | */ | ||
1111 | sctp_assoc_set_id(asoc, GFP_ATOMIC); | ||
1112 | } | ||
1106 | } | 1113 | } |
1107 | } | 1114 | } |
1108 | 1115 | ||
@@ -1375,3 +1382,25 @@ out: | |||
1375 | sctp_read_unlock(&asoc->base.addr_lock); | 1382 | sctp_read_unlock(&asoc->base.addr_lock); |
1376 | return found; | 1383 | return found; |
1377 | } | 1384 | } |
1385 | |||
1386 | /* Set an association id for a given association */ | ||
1387 | int sctp_assoc_set_id(struct sctp_association *asoc, gfp_t gfp) | ||
1388 | { | ||
1389 | int assoc_id; | ||
1390 | int error = 0; | ||
1391 | retry: | ||
1392 | if (unlikely(!idr_pre_get(&sctp_assocs_id, gfp))) | ||
1393 | return -ENOMEM; | ||
1394 | |||
1395 | spin_lock_bh(&sctp_assocs_id_lock); | ||
1396 | error = idr_get_new_above(&sctp_assocs_id, (void *)asoc, | ||
1397 | 1, &assoc_id); | ||
1398 | spin_unlock_bh(&sctp_assocs_id_lock); | ||
1399 | if (error == -EAGAIN) | ||
1400 | goto retry; | ||
1401 | else if (error) | ||
1402 | return error; | ||
1403 | |||
1404 | asoc->assoc_id = (sctp_assoc_t) assoc_id; | ||
1405 | return error; | ||
1406 | } | ||
diff --git a/net/sctp/ipv6.c b/net/sctp/ipv6.c index ca527a27dd05..84cd53635fe8 100644 --- a/net/sctp/ipv6.c +++ b/net/sctp/ipv6.c | |||
@@ -992,45 +992,52 @@ static struct sctp_pf sctp_pf_inet6_specific = { | |||
992 | .af = &sctp_ipv6_specific, | 992 | .af = &sctp_ipv6_specific, |
993 | }; | 993 | }; |
994 | 994 | ||
995 | /* Initialize IPv6 support and register with inet6 stack. */ | 995 | /* Initialize IPv6 support and register with socket layer. */ |
996 | int sctp_v6_init(void) | 996 | int sctp_v6_init(void) |
997 | { | 997 | { |
998 | int rc = proto_register(&sctpv6_prot, 1); | 998 | int rc; |
999 | 999 | ||
1000 | /* Register the SCTP specific PF_INET6 functions. */ | ||
1001 | sctp_register_pf(&sctp_pf_inet6_specific, PF_INET6); | ||
1002 | |||
1003 | /* Register the SCTP specific AF_INET6 functions. */ | ||
1004 | sctp_register_af(&sctp_ipv6_specific); | ||
1005 | |||
1006 | rc = proto_register(&sctpv6_prot, 1); | ||
1000 | if (rc) | 1007 | if (rc) |
1001 | goto out; | 1008 | return rc; |
1002 | /* Register inet6 protocol. */ | ||
1003 | rc = -EAGAIN; | ||
1004 | if (inet6_add_protocol(&sctpv6_protocol, IPPROTO_SCTP) < 0) | ||
1005 | goto out_unregister_sctp_proto; | ||
1006 | 1009 | ||
1007 | /* Add SCTPv6(UDP and TCP style) to inetsw6 linked list. */ | 1010 | /* Add SCTPv6(UDP and TCP style) to inetsw6 linked list. */ |
1008 | inet6_register_protosw(&sctpv6_seqpacket_protosw); | 1011 | inet6_register_protosw(&sctpv6_seqpacket_protosw); |
1009 | inet6_register_protosw(&sctpv6_stream_protosw); | 1012 | inet6_register_protosw(&sctpv6_stream_protosw); |
1010 | 1013 | ||
1011 | /* Register the SCTP specific PF_INET6 functions. */ | 1014 | return 0; |
1012 | sctp_register_pf(&sctp_pf_inet6_specific, PF_INET6); | 1015 | } |
1013 | |||
1014 | /* Register the SCTP specific AF_INET6 functions. */ | ||
1015 | sctp_register_af(&sctp_ipv6_specific); | ||
1016 | 1016 | ||
1017 | /* Register with inet6 layer. */ | ||
1018 | int sctp_v6_add_protocol(void) | ||
1019 | { | ||
1017 | /* Register notifier for inet6 address additions/deletions. */ | 1020 | /* Register notifier for inet6 address additions/deletions. */ |
1018 | register_inet6addr_notifier(&sctp_inet6addr_notifier); | 1021 | register_inet6addr_notifier(&sctp_inet6addr_notifier); |
1019 | rc = 0; | 1022 | |
1020 | out: | 1023 | if (inet6_add_protocol(&sctpv6_protocol, IPPROTO_SCTP) < 0) |
1021 | return rc; | 1024 | return -EAGAIN; |
1022 | out_unregister_sctp_proto: | 1025 | |
1023 | proto_unregister(&sctpv6_prot); | 1026 | return 0; |
1024 | goto out; | ||
1025 | } | 1027 | } |
1026 | 1028 | ||
1027 | /* IPv6 specific exit support. */ | 1029 | /* IPv6 specific exit support. */ |
1028 | void sctp_v6_exit(void) | 1030 | void sctp_v6_exit(void) |
1029 | { | 1031 | { |
1030 | list_del(&sctp_ipv6_specific.list); | ||
1031 | inet6_del_protocol(&sctpv6_protocol, IPPROTO_SCTP); | ||
1032 | inet6_unregister_protosw(&sctpv6_seqpacket_protosw); | 1032 | inet6_unregister_protosw(&sctpv6_seqpacket_protosw); |
1033 | inet6_unregister_protosw(&sctpv6_stream_protosw); | 1033 | inet6_unregister_protosw(&sctpv6_stream_protosw); |
1034 | unregister_inet6addr_notifier(&sctp_inet6addr_notifier); | ||
1035 | proto_unregister(&sctpv6_prot); | 1034 | proto_unregister(&sctpv6_prot); |
1035 | list_del(&sctp_ipv6_specific.list); | ||
1036 | } | ||
1037 | |||
1038 | /* Unregister with inet6 layer. */ | ||
1039 | void sctp_v6_del_protocol(void) | ||
1040 | { | ||
1041 | inet6_del_protocol(&sctpv6_protocol, IPPROTO_SCTP); | ||
1042 | unregister_inet6addr_notifier(&sctp_inet6addr_notifier); | ||
1036 | } | 1043 | } |
diff --git a/net/sctp/protocol.c b/net/sctp/protocol.c index c361deb6cea9..34bab36637ac 100644 --- a/net/sctp/protocol.c +++ b/net/sctp/protocol.c | |||
@@ -170,7 +170,7 @@ static void sctp_get_local_addr_list(void) | |||
170 | struct sctp_af *af; | 170 | struct sctp_af *af; |
171 | 171 | ||
172 | read_lock(&dev_base_lock); | 172 | read_lock(&dev_base_lock); |
173 | for (dev = dev_base; dev; dev = dev->next) { | 173 | for_each_netdev(dev) { |
174 | __list_for_each(pos, &sctp_address_families) { | 174 | __list_for_each(pos, &sctp_address_families) { |
175 | af = list_entry(pos, struct sctp_af, list); | 175 | af = list_entry(pos, struct sctp_af, list); |
176 | af->copy_addrlist(&sctp_local_addr_list, dev); | 176 | af->copy_addrlist(&sctp_local_addr_list, dev); |
@@ -975,28 +975,14 @@ SCTP_STATIC __init int sctp_init(void) | |||
975 | if (!sctp_sanity_check()) | 975 | if (!sctp_sanity_check()) |
976 | goto out; | 976 | goto out; |
977 | 977 | ||
978 | status = proto_register(&sctp_prot, 1); | 978 | /* Allocate bind_bucket and chunk caches. */ |
979 | if (status) | ||
980 | goto out; | ||
981 | |||
982 | /* Add SCTP to inet_protos hash table. */ | ||
983 | status = -EAGAIN; | ||
984 | if (inet_add_protocol(&sctp_protocol, IPPROTO_SCTP) < 0) | ||
985 | goto err_add_protocol; | ||
986 | |||
987 | /* Add SCTP(TCP and UDP style) to inetsw linked list. */ | ||
988 | inet_register_protosw(&sctp_seqpacket_protosw); | ||
989 | inet_register_protosw(&sctp_stream_protosw); | ||
990 | |||
991 | /* Allocate a cache pools. */ | ||
992 | status = -ENOBUFS; | 979 | status = -ENOBUFS; |
993 | sctp_bucket_cachep = kmem_cache_create("sctp_bind_bucket", | 980 | sctp_bucket_cachep = kmem_cache_create("sctp_bind_bucket", |
994 | sizeof(struct sctp_bind_bucket), | 981 | sizeof(struct sctp_bind_bucket), |
995 | 0, SLAB_HWCACHE_ALIGN, | 982 | 0, SLAB_HWCACHE_ALIGN, |
996 | NULL, NULL); | 983 | NULL, NULL); |
997 | |||
998 | if (!sctp_bucket_cachep) | 984 | if (!sctp_bucket_cachep) |
999 | goto err_bucket_cachep; | 985 | goto out; |
1000 | 986 | ||
1001 | sctp_chunk_cachep = kmem_cache_create("sctp_chunk", | 987 | sctp_chunk_cachep = kmem_cache_create("sctp_chunk", |
1002 | sizeof(struct sctp_chunk), | 988 | sizeof(struct sctp_chunk), |
@@ -1153,6 +1139,14 @@ SCTP_STATIC __init int sctp_init(void) | |||
1153 | INIT_LIST_HEAD(&sctp_address_families); | 1139 | INIT_LIST_HEAD(&sctp_address_families); |
1154 | sctp_register_af(&sctp_ipv4_specific); | 1140 | sctp_register_af(&sctp_ipv4_specific); |
1155 | 1141 | ||
1142 | status = proto_register(&sctp_prot, 1); | ||
1143 | if (status) | ||
1144 | goto err_proto_register; | ||
1145 | |||
1146 | /* Register SCTP(UDP and TCP style) with socket layer. */ | ||
1147 | inet_register_protosw(&sctp_seqpacket_protosw); | ||
1148 | inet_register_protosw(&sctp_stream_protosw); | ||
1149 | |||
1156 | status = sctp_v6_init(); | 1150 | status = sctp_v6_init(); |
1157 | if (status) | 1151 | if (status) |
1158 | goto err_v6_init; | 1152 | goto err_v6_init; |
@@ -1166,19 +1160,39 @@ SCTP_STATIC __init int sctp_init(void) | |||
1166 | 1160 | ||
1167 | /* Initialize the local address list. */ | 1161 | /* Initialize the local address list. */ |
1168 | INIT_LIST_HEAD(&sctp_local_addr_list); | 1162 | INIT_LIST_HEAD(&sctp_local_addr_list); |
1169 | |||
1170 | sctp_get_local_addr_list(); | 1163 | sctp_get_local_addr_list(); |
1171 | 1164 | ||
1172 | /* Register notifier for inet address additions/deletions. */ | 1165 | /* Register notifier for inet address additions/deletions. */ |
1173 | register_inetaddr_notifier(&sctp_inetaddr_notifier); | 1166 | register_inetaddr_notifier(&sctp_inetaddr_notifier); |
1174 | 1167 | ||
1168 | /* Register SCTP with inet layer. */ | ||
1169 | if (inet_add_protocol(&sctp_protocol, IPPROTO_SCTP) < 0) { | ||
1170 | status = -EAGAIN; | ||
1171 | goto err_add_protocol; | ||
1172 | } | ||
1173 | |||
1174 | /* Register SCTP with inet6 layer. */ | ||
1175 | status = sctp_v6_add_protocol(); | ||
1176 | if (status) | ||
1177 | goto err_v6_add_protocol; | ||
1178 | |||
1175 | __unsafe(THIS_MODULE); | 1179 | __unsafe(THIS_MODULE); |
1176 | status = 0; | 1180 | status = 0; |
1177 | out: | 1181 | out: |
1178 | return status; | 1182 | return status; |
1183 | err_v6_add_protocol: | ||
1184 | inet_del_protocol(&sctp_protocol, IPPROTO_SCTP); | ||
1185 | unregister_inetaddr_notifier(&sctp_inetaddr_notifier); | ||
1186 | err_add_protocol: | ||
1187 | sctp_free_local_addr_list(); | ||
1188 | sock_release(sctp_ctl_socket); | ||
1179 | err_ctl_sock_init: | 1189 | err_ctl_sock_init: |
1180 | sctp_v6_exit(); | 1190 | sctp_v6_exit(); |
1181 | err_v6_init: | 1191 | err_v6_init: |
1192 | inet_unregister_protosw(&sctp_stream_protosw); | ||
1193 | inet_unregister_protosw(&sctp_seqpacket_protosw); | ||
1194 | proto_unregister(&sctp_prot); | ||
1195 | err_proto_register: | ||
1182 | sctp_sysctl_unregister(); | 1196 | sctp_sysctl_unregister(); |
1183 | list_del(&sctp_ipv4_specific.list); | 1197 | list_del(&sctp_ipv4_specific.list); |
1184 | free_pages((unsigned long)sctp_port_hashtable, | 1198 | free_pages((unsigned long)sctp_port_hashtable, |
@@ -1192,19 +1206,13 @@ err_ehash_alloc: | |||
1192 | sizeof(struct sctp_hashbucket))); | 1206 | sizeof(struct sctp_hashbucket))); |
1193 | err_ahash_alloc: | 1207 | err_ahash_alloc: |
1194 | sctp_dbg_objcnt_exit(); | 1208 | sctp_dbg_objcnt_exit(); |
1195 | err_init_proc: | ||
1196 | sctp_proc_exit(); | 1209 | sctp_proc_exit(); |
1210 | err_init_proc: | ||
1197 | cleanup_sctp_mibs(); | 1211 | cleanup_sctp_mibs(); |
1198 | err_init_mibs: | 1212 | err_init_mibs: |
1199 | kmem_cache_destroy(sctp_chunk_cachep); | 1213 | kmem_cache_destroy(sctp_chunk_cachep); |
1200 | err_chunk_cachep: | 1214 | err_chunk_cachep: |
1201 | kmem_cache_destroy(sctp_bucket_cachep); | 1215 | kmem_cache_destroy(sctp_bucket_cachep); |
1202 | err_bucket_cachep: | ||
1203 | inet_del_protocol(&sctp_protocol, IPPROTO_SCTP); | ||
1204 | inet_unregister_protosw(&sctp_seqpacket_protosw); | ||
1205 | inet_unregister_protosw(&sctp_stream_protosw); | ||
1206 | err_add_protocol: | ||
1207 | proto_unregister(&sctp_prot); | ||
1208 | goto out; | 1216 | goto out; |
1209 | } | 1217 | } |
1210 | 1218 | ||
@@ -1215,8 +1223,9 @@ SCTP_STATIC __exit void sctp_exit(void) | |||
1215 | * up all the remaining associations and all that memory. | 1223 | * up all the remaining associations and all that memory. |
1216 | */ | 1224 | */ |
1217 | 1225 | ||
1218 | /* Unregister notifier for inet address additions/deletions. */ | 1226 | /* Unregister with inet6/inet layers. */ |
1219 | unregister_inetaddr_notifier(&sctp_inetaddr_notifier); | 1227 | sctp_v6_del_protocol(); |
1228 | inet_del_protocol(&sctp_protocol, IPPROTO_SCTP); | ||
1220 | 1229 | ||
1221 | /* Free the local address list. */ | 1230 | /* Free the local address list. */ |
1222 | sctp_free_local_addr_list(); | 1231 | sctp_free_local_addr_list(); |
@@ -1224,7 +1233,16 @@ SCTP_STATIC __exit void sctp_exit(void) | |||
1224 | /* Free the control endpoint. */ | 1233 | /* Free the control endpoint. */ |
1225 | sock_release(sctp_ctl_socket); | 1234 | sock_release(sctp_ctl_socket); |
1226 | 1235 | ||
1236 | /* Cleanup v6 initializations. */ | ||
1227 | sctp_v6_exit(); | 1237 | sctp_v6_exit(); |
1238 | |||
1239 | /* Unregister with socket layer. */ | ||
1240 | inet_unregister_protosw(&sctp_stream_protosw); | ||
1241 | inet_unregister_protosw(&sctp_seqpacket_protosw); | ||
1242 | |||
1243 | /* Unregister notifier for inet address additions/deletions. */ | ||
1244 | unregister_inetaddr_notifier(&sctp_inetaddr_notifier); | ||
1245 | |||
1228 | sctp_sysctl_unregister(); | 1246 | sctp_sysctl_unregister(); |
1229 | list_del(&sctp_ipv4_specific.list); | 1247 | list_del(&sctp_ipv4_specific.list); |
1230 | 1248 | ||
@@ -1236,16 +1254,13 @@ SCTP_STATIC __exit void sctp_exit(void) | |||
1236 | get_order(sctp_port_hashsize * | 1254 | get_order(sctp_port_hashsize * |
1237 | sizeof(struct sctp_bind_hashbucket))); | 1255 | sizeof(struct sctp_bind_hashbucket))); |
1238 | 1256 | ||
1239 | kmem_cache_destroy(sctp_chunk_cachep); | ||
1240 | kmem_cache_destroy(sctp_bucket_cachep); | ||
1241 | |||
1242 | sctp_dbg_objcnt_exit(); | 1257 | sctp_dbg_objcnt_exit(); |
1243 | sctp_proc_exit(); | 1258 | sctp_proc_exit(); |
1244 | cleanup_sctp_mibs(); | 1259 | cleanup_sctp_mibs(); |
1245 | 1260 | ||
1246 | inet_del_protocol(&sctp_protocol, IPPROTO_SCTP); | 1261 | kmem_cache_destroy(sctp_chunk_cachep); |
1247 | inet_unregister_protosw(&sctp_seqpacket_protosw); | 1262 | kmem_cache_destroy(sctp_bucket_cachep); |
1248 | inet_unregister_protosw(&sctp_stream_protosw); | 1263 | |
1249 | proto_unregister(&sctp_prot); | 1264 | proto_unregister(&sctp_prot); |
1250 | } | 1265 | } |
1251 | 1266 | ||
diff --git a/net/sctp/sm_make_chunk.c b/net/sctp/sm_make_chunk.c index be783a3761c4..8d18f570c2e6 100644 --- a/net/sctp/sm_make_chunk.c +++ b/net/sctp/sm_make_chunk.c | |||
@@ -1939,7 +1939,6 @@ int sctp_process_init(struct sctp_association *asoc, sctp_cid_t cid, | |||
1939 | * association. | 1939 | * association. |
1940 | */ | 1940 | */ |
1941 | if (!asoc->temp) { | 1941 | if (!asoc->temp) { |
1942 | int assoc_id; | ||
1943 | int error; | 1942 | int error; |
1944 | 1943 | ||
1945 | asoc->ssnmap = sctp_ssnmap_new(asoc->c.sinit_max_instreams, | 1944 | asoc->ssnmap = sctp_ssnmap_new(asoc->c.sinit_max_instreams, |
@@ -1947,19 +1946,9 @@ int sctp_process_init(struct sctp_association *asoc, sctp_cid_t cid, | |||
1947 | if (!asoc->ssnmap) | 1946 | if (!asoc->ssnmap) |
1948 | goto clean_up; | 1947 | goto clean_up; |
1949 | 1948 | ||
1950 | retry: | 1949 | error = sctp_assoc_set_id(asoc, gfp); |
1951 | if (unlikely(!idr_pre_get(&sctp_assocs_id, gfp))) | 1950 | if (error) |
1952 | goto clean_up; | 1951 | goto clean_up; |
1953 | spin_lock_bh(&sctp_assocs_id_lock); | ||
1954 | error = idr_get_new_above(&sctp_assocs_id, (void *)asoc, 1, | ||
1955 | &assoc_id); | ||
1956 | spin_unlock_bh(&sctp_assocs_id_lock); | ||
1957 | if (error == -EAGAIN) | ||
1958 | goto retry; | ||
1959 | else if (error) | ||
1960 | goto clean_up; | ||
1961 | |||
1962 | asoc->assoc_id = (sctp_assoc_t) assoc_id; | ||
1963 | } | 1952 | } |
1964 | 1953 | ||
1965 | /* ADDIP Section 4.1 ASCONF Chunk Procedures | 1954 | /* ADDIP Section 4.1 ASCONF Chunk Procedures |
diff --git a/net/sctp/sm_sideeffect.c b/net/sctp/sm_sideeffect.c index b37a7adeb150..d9fad4f6ffc3 100644 --- a/net/sctp/sm_sideeffect.c +++ b/net/sctp/sm_sideeffect.c | |||
@@ -862,6 +862,33 @@ static void sctp_cmd_set_sk_err(struct sctp_association *asoc, int error) | |||
862 | sk->sk_err = error; | 862 | sk->sk_err = error; |
863 | } | 863 | } |
864 | 864 | ||
865 | /* Helper function to generate an association change event */ | ||
866 | static void sctp_cmd_assoc_change(sctp_cmd_seq_t *commands, | ||
867 | struct sctp_association *asoc, | ||
868 | u8 state) | ||
869 | { | ||
870 | struct sctp_ulpevent *ev; | ||
871 | |||
872 | ev = sctp_ulpevent_make_assoc_change(asoc, 0, state, 0, | ||
873 | asoc->c.sinit_num_ostreams, | ||
874 | asoc->c.sinit_max_instreams, | ||
875 | NULL, GFP_ATOMIC); | ||
876 | if (ev) | ||
877 | sctp_ulpq_tail_event(&asoc->ulpq, ev); | ||
878 | } | ||
879 | |||
880 | /* Helper function to generate an adaptation indication event */ | ||
881 | static void sctp_cmd_adaptation_ind(sctp_cmd_seq_t *commands, | ||
882 | struct sctp_association *asoc) | ||
883 | { | ||
884 | struct sctp_ulpevent *ev; | ||
885 | |||
886 | ev = sctp_ulpevent_make_adaptation_indication(asoc, GFP_ATOMIC); | ||
887 | |||
888 | if (ev) | ||
889 | sctp_ulpq_tail_event(&asoc->ulpq, ev); | ||
890 | } | ||
891 | |||
865 | /* These three macros allow us to pull the debugging code out of the | 892 | /* These three macros allow us to pull the debugging code out of the |
866 | * main flow of sctp_do_sm() to keep attention focused on the real | 893 | * main flow of sctp_do_sm() to keep attention focused on the real |
867 | * functionality there. | 894 | * functionality there. |
@@ -1485,6 +1512,14 @@ static int sctp_cmd_interpreter(sctp_event_t event_type, | |||
1485 | case SCTP_CMD_SET_SK_ERR: | 1512 | case SCTP_CMD_SET_SK_ERR: |
1486 | sctp_cmd_set_sk_err(asoc, cmd->obj.error); | 1513 | sctp_cmd_set_sk_err(asoc, cmd->obj.error); |
1487 | break; | 1514 | break; |
1515 | case SCTP_CMD_ASSOC_CHANGE: | ||
1516 | sctp_cmd_assoc_change(commands, asoc, | ||
1517 | cmd->obj.u8); | ||
1518 | break; | ||
1519 | case SCTP_CMD_ADAPTATION_IND: | ||
1520 | sctp_cmd_adaptation_ind(commands, asoc); | ||
1521 | break; | ||
1522 | |||
1488 | default: | 1523 | default: |
1489 | printk(KERN_WARNING "Impossible command: %u, %p\n", | 1524 | printk(KERN_WARNING "Impossible command: %u, %p\n", |
1490 | cmd->verb, cmd->obj.ptr); | 1525 | cmd->verb, cmd->obj.ptr); |
diff --git a/net/sctp/sm_statefuns.c b/net/sctp/sm_statefuns.c index 9e28a5d51200..f02ce3dddb7b 100644 --- a/net/sctp/sm_statefuns.c +++ b/net/sctp/sm_statefuns.c | |||
@@ -1656,7 +1656,6 @@ static sctp_disposition_t sctp_sf_do_dupcook_b(const struct sctp_endpoint *ep, | |||
1656 | struct sctp_association *new_asoc) | 1656 | struct sctp_association *new_asoc) |
1657 | { | 1657 | { |
1658 | sctp_init_chunk_t *peer_init; | 1658 | sctp_init_chunk_t *peer_init; |
1659 | struct sctp_ulpevent *ev; | ||
1660 | struct sctp_chunk *repl; | 1659 | struct sctp_chunk *repl; |
1661 | 1660 | ||
1662 | /* new_asoc is a brand-new association, so these are not yet | 1661 | /* new_asoc is a brand-new association, so these are not yet |
@@ -1687,34 +1686,28 @@ static sctp_disposition_t sctp_sf_do_dupcook_b(const struct sctp_endpoint *ep, | |||
1687 | * D) IMPLEMENTATION NOTE: An implementation may choose to | 1686 | * D) IMPLEMENTATION NOTE: An implementation may choose to |
1688 | * send the Communication Up notification to the SCTP user | 1687 | * send the Communication Up notification to the SCTP user |
1689 | * upon reception of a valid COOKIE ECHO chunk. | 1688 | * upon reception of a valid COOKIE ECHO chunk. |
1689 | * | ||
1690 | * Sadly, this needs to be implemented as a side-effect, because | ||
1691 | * we are not guaranteed to have set the association id of the real | ||
1692 | * association and so these notifications need to be delayed until | ||
1693 | * the association id is allocated. | ||
1690 | */ | 1694 | */ |
1691 | ev = sctp_ulpevent_make_assoc_change(asoc, 0, SCTP_COMM_UP, 0, | ||
1692 | new_asoc->c.sinit_num_ostreams, | ||
1693 | new_asoc->c.sinit_max_instreams, | ||
1694 | NULL, GFP_ATOMIC); | ||
1695 | if (!ev) | ||
1696 | goto nomem_ev; | ||
1697 | 1695 | ||
1698 | sctp_add_cmd_sf(commands, SCTP_CMD_EVENT_ULP, SCTP_ULPEVENT(ev)); | 1696 | sctp_add_cmd_sf(commands, SCTP_CMD_ASSOC_CHANGE, SCTP_U8(SCTP_COMM_UP)); |
1699 | 1697 | ||
1700 | /* Sockets API Draft Section 5.3.1.6 | 1698 | /* Sockets API Draft Section 5.3.1.6 |
1701 | * When a peer sends a Adaptation Layer Indication parameter , SCTP | 1699 | * When a peer sends a Adaptation Layer Indication parameter , SCTP |
1702 | * delivers this notification to inform the application that of the | 1700 | * delivers this notification to inform the application that of the |
1703 | * peers requested adaptation layer. | 1701 | * peers requested adaptation layer. |
1702 | * | ||
1703 | * This also needs to be done as a side effect for the same reason as | ||
1704 | * above. | ||
1704 | */ | 1705 | */ |
1705 | if (asoc->peer.adaptation_ind) { | 1706 | if (asoc->peer.adaptation_ind) |
1706 | ev = sctp_ulpevent_make_adaptation_indication(asoc, GFP_ATOMIC); | 1707 | sctp_add_cmd_sf(commands, SCTP_CMD_ADAPTATION_IND, SCTP_NULL()); |
1707 | if (!ev) | ||
1708 | goto nomem_ev; | ||
1709 | |||
1710 | sctp_add_cmd_sf(commands, SCTP_CMD_EVENT_ULP, | ||
1711 | SCTP_ULPEVENT(ev)); | ||
1712 | } | ||
1713 | 1708 | ||
1714 | return SCTP_DISPOSITION_CONSUME; | 1709 | return SCTP_DISPOSITION_CONSUME; |
1715 | 1710 | ||
1716 | nomem_ev: | ||
1717 | sctp_chunk_free(repl); | ||
1718 | nomem: | 1711 | nomem: |
1719 | return SCTP_DISPOSITION_NOMEM; | 1712 | return SCTP_DISPOSITION_NOMEM; |
1720 | } | 1713 | } |
diff --git a/net/sctp/socket.c b/net/sctp/socket.c index 2fc0a92caa78..9f1a908776de 100644 --- a/net/sctp/socket.c +++ b/net/sctp/socket.c | |||
@@ -972,6 +972,7 @@ static int __sctp_connect(struct sock* sk, | |||
972 | int walk_size = 0; | 972 | int walk_size = 0; |
973 | union sctp_addr *sa_addr; | 973 | union sctp_addr *sa_addr; |
974 | void *addr_buf; | 974 | void *addr_buf; |
975 | unsigned short port; | ||
975 | 976 | ||
976 | sp = sctp_sk(sk); | 977 | sp = sctp_sk(sk); |
977 | ep = sp->ep; | 978 | ep = sp->ep; |
@@ -992,6 +993,7 @@ static int __sctp_connect(struct sock* sk, | |||
992 | while (walk_size < addrs_size) { | 993 | while (walk_size < addrs_size) { |
993 | sa_addr = (union sctp_addr *)addr_buf; | 994 | sa_addr = (union sctp_addr *)addr_buf; |
994 | af = sctp_get_af_specific(sa_addr->sa.sa_family); | 995 | af = sctp_get_af_specific(sa_addr->sa.sa_family); |
996 | port = ntohs(sa_addr->v4.sin_port); | ||
995 | 997 | ||
996 | /* If the address family is not supported or if this address | 998 | /* If the address family is not supported or if this address |
997 | * causes the address buffer to overflow return EINVAL. | 999 | * causes the address buffer to overflow return EINVAL. |
@@ -1005,6 +1007,12 @@ static int __sctp_connect(struct sock* sk, | |||
1005 | if (err) | 1007 | if (err) |
1006 | goto out_free; | 1008 | goto out_free; |
1007 | 1009 | ||
1010 | /* Make sure the destination port is correctly set | ||
1011 | * in all addresses. | ||
1012 | */ | ||
1013 | if (asoc && asoc->peer.port && asoc->peer.port != port) | ||
1014 | goto out_free; | ||
1015 | |||
1008 | memcpy(&to, sa_addr, af->sockaddr_len); | 1016 | memcpy(&to, sa_addr, af->sockaddr_len); |
1009 | 1017 | ||
1010 | /* Check if there already is a matching association on the | 1018 | /* Check if there already is a matching association on the |
@@ -5012,7 +5020,8 @@ pp_found: | |||
5012 | struct hlist_node *node; | 5020 | struct hlist_node *node; |
5013 | 5021 | ||
5014 | SCTP_DEBUG_PRINTK("sctp_get_port() found a possible match\n"); | 5022 | SCTP_DEBUG_PRINTK("sctp_get_port() found a possible match\n"); |
5015 | if (pp->fastreuse && sk->sk_reuse) | 5023 | if (pp->fastreuse && sk->sk_reuse && |
5024 | sk->sk_state != SCTP_SS_LISTENING) | ||
5016 | goto success; | 5025 | goto success; |
5017 | 5026 | ||
5018 | /* Run through the list of sockets bound to the port | 5027 | /* Run through the list of sockets bound to the port |
@@ -5029,7 +5038,8 @@ pp_found: | |||
5029 | struct sctp_endpoint *ep2; | 5038 | struct sctp_endpoint *ep2; |
5030 | ep2 = sctp_sk(sk2)->ep; | 5039 | ep2 = sctp_sk(sk2)->ep; |
5031 | 5040 | ||
5032 | if (reuse && sk2->sk_reuse) | 5041 | if (reuse && sk2->sk_reuse && |
5042 | sk2->sk_state != SCTP_SS_LISTENING) | ||
5033 | continue; | 5043 | continue; |
5034 | 5044 | ||
5035 | if (sctp_bind_addr_match(&ep2->base.bind_addr, addr, | 5045 | if (sctp_bind_addr_match(&ep2->base.bind_addr, addr, |
@@ -5050,9 +5060,13 @@ pp_not_found: | |||
5050 | * if sk->sk_reuse is too (that is, if the caller requested | 5060 | * if sk->sk_reuse is too (that is, if the caller requested |
5051 | * SO_REUSEADDR on this socket -sk-). | 5061 | * SO_REUSEADDR on this socket -sk-). |
5052 | */ | 5062 | */ |
5053 | if (hlist_empty(&pp->owner)) | 5063 | if (hlist_empty(&pp->owner)) { |
5054 | pp->fastreuse = sk->sk_reuse ? 1 : 0; | 5064 | if (sk->sk_reuse && sk->sk_state != SCTP_SS_LISTENING) |
5055 | else if (pp->fastreuse && !sk->sk_reuse) | 5065 | pp->fastreuse = 1; |
5066 | else | ||
5067 | pp->fastreuse = 0; | ||
5068 | } else if (pp->fastreuse && | ||
5069 | (!sk->sk_reuse || sk->sk_state == SCTP_SS_LISTENING)) | ||
5056 | pp->fastreuse = 0; | 5070 | pp->fastreuse = 0; |
5057 | 5071 | ||
5058 | /* We are set, so fill up all the data in the hash table | 5072 | /* We are set, so fill up all the data in the hash table |
@@ -5060,8 +5074,8 @@ pp_not_found: | |||
5060 | * sockets FIXME: Blurry, NPI (ipg). | 5074 | * sockets FIXME: Blurry, NPI (ipg). |
5061 | */ | 5075 | */ |
5062 | success: | 5076 | success: |
5063 | inet_sk(sk)->num = snum; | ||
5064 | if (!sctp_sk(sk)->bind_hash) { | 5077 | if (!sctp_sk(sk)->bind_hash) { |
5078 | inet_sk(sk)->num = snum; | ||
5065 | sk_add_bind_node(sk, &pp->owner); | 5079 | sk_add_bind_node(sk, &pp->owner); |
5066 | sctp_sk(sk)->bind_hash = pp; | 5080 | sctp_sk(sk)->bind_hash = pp; |
5067 | } | 5081 | } |
@@ -5134,12 +5148,16 @@ SCTP_STATIC int sctp_seqpacket_listen(struct sock *sk, int backlog) | |||
5134 | * This is not currently spelled out in the SCTP sockets | 5148 | * This is not currently spelled out in the SCTP sockets |
5135 | * extensions draft, but follows the practice as seen in TCP | 5149 | * extensions draft, but follows the practice as seen in TCP |
5136 | * sockets. | 5150 | * sockets. |
5151 | * | ||
5152 | * Additionally, turn off fastreuse flag since we are not listening | ||
5137 | */ | 5153 | */ |
5154 | sk->sk_state = SCTP_SS_LISTENING; | ||
5138 | if (!ep->base.bind_addr.port) { | 5155 | if (!ep->base.bind_addr.port) { |
5139 | if (sctp_autobind(sk)) | 5156 | if (sctp_autobind(sk)) |
5140 | return -EAGAIN; | 5157 | return -EAGAIN; |
5141 | } | 5158 | } else |
5142 | sk->sk_state = SCTP_SS_LISTENING; | 5159 | sctp_sk(sk)->bind_hash->fastreuse = 0; |
5160 | |||
5143 | sctp_hash_endpoint(ep); | 5161 | sctp_hash_endpoint(ep); |
5144 | return 0; | 5162 | return 0; |
5145 | } | 5163 | } |
@@ -5177,11 +5195,13 @@ SCTP_STATIC int sctp_stream_listen(struct sock *sk, int backlog) | |||
5177 | * extensions draft, but follows the practice as seen in TCP | 5195 | * extensions draft, but follows the practice as seen in TCP |
5178 | * sockets. | 5196 | * sockets. |
5179 | */ | 5197 | */ |
5198 | sk->sk_state = SCTP_SS_LISTENING; | ||
5180 | if (!ep->base.bind_addr.port) { | 5199 | if (!ep->base.bind_addr.port) { |
5181 | if (sctp_autobind(sk)) | 5200 | if (sctp_autobind(sk)) |
5182 | return -EAGAIN; | 5201 | return -EAGAIN; |
5183 | } | 5202 | } else |
5184 | sk->sk_state = SCTP_SS_LISTENING; | 5203 | sctp_sk(sk)->bind_hash->fastreuse = 0; |
5204 | |||
5185 | sk->sk_max_ack_backlog = backlog; | 5205 | sk->sk_max_ack_backlog = backlog; |
5186 | sctp_hash_endpoint(ep); | 5206 | sctp_hash_endpoint(ep); |
5187 | return 0; | 5207 | return 0; |
diff --git a/net/tipc/eth_media.c b/net/tipc/eth_media.c index 67bb29b44d1b..0ee6ded18f3a 100644 --- a/net/tipc/eth_media.c +++ b/net/tipc/eth_media.c | |||
@@ -120,16 +120,18 @@ static int recv_msg(struct sk_buff *buf, struct net_device *dev, | |||
120 | 120 | ||
121 | static int enable_bearer(struct tipc_bearer *tb_ptr) | 121 | static int enable_bearer(struct tipc_bearer *tb_ptr) |
122 | { | 122 | { |
123 | struct net_device *dev = dev_base; | 123 | struct net_device *dev, *pdev; |
124 | struct eth_bearer *eb_ptr = ð_bearers[0]; | 124 | struct eth_bearer *eb_ptr = ð_bearers[0]; |
125 | struct eth_bearer *stop = ð_bearers[MAX_ETH_BEARERS]; | 125 | struct eth_bearer *stop = ð_bearers[MAX_ETH_BEARERS]; |
126 | char *driver_name = strchr((const char *)tb_ptr->name, ':') + 1; | 126 | char *driver_name = strchr((const char *)tb_ptr->name, ':') + 1; |
127 | 127 | ||
128 | /* Find device with specified name */ | 128 | /* Find device with specified name */ |
129 | 129 | dev = NULL; | |
130 | while (dev && dev->name && strncmp(dev->name, driver_name, IFNAMSIZ)) { | 130 | for_each_netdev(pdev) |
131 | dev = dev->next; | 131 | if (!strncmp(dev->name, driver_name, IFNAMSIZ)) { |
132 | } | 132 | dev = pdev; |
133 | break; | ||
134 | } | ||
133 | if (!dev) | 135 | if (!dev) |
134 | return -ENODEV; | 136 | return -ENODEV; |
135 | 137 | ||
diff --git a/net/xfrm/xfrm_policy.c b/net/xfrm/xfrm_policy.c index 263e34e45265..95271e8426a1 100644 --- a/net/xfrm/xfrm_policy.c +++ b/net/xfrm/xfrm_policy.c | |||
@@ -579,7 +579,7 @@ static inline int xfrm_byidx_should_resize(int total) | |||
579 | return 0; | 579 | return 0; |
580 | } | 580 | } |
581 | 581 | ||
582 | void xfrm_spd_getinfo(struct xfrm_spdinfo *si) | 582 | void xfrm_spd_getinfo(struct xfrmk_spdinfo *si) |
583 | { | 583 | { |
584 | read_lock_bh(&xfrm_policy_lock); | 584 | read_lock_bh(&xfrm_policy_lock); |
585 | si->incnt = xfrm_policy_count[XFRM_POLICY_IN]; | 585 | si->incnt = xfrm_policy_count[XFRM_POLICY_IN]; |
diff --git a/net/xfrm/xfrm_state.c b/net/xfrm/xfrm_state.c index f3a61ebd8d65..9955ff4da0a2 100644 --- a/net/xfrm/xfrm_state.c +++ b/net/xfrm/xfrm_state.c | |||
@@ -421,7 +421,7 @@ restart: | |||
421 | } | 421 | } |
422 | EXPORT_SYMBOL(xfrm_state_flush); | 422 | EXPORT_SYMBOL(xfrm_state_flush); |
423 | 423 | ||
424 | void xfrm_sad_getinfo(struct xfrm_sadinfo *si) | 424 | void xfrm_sad_getinfo(struct xfrmk_sadinfo *si) |
425 | { | 425 | { |
426 | spin_lock_bh(&xfrm_state_lock); | 426 | spin_lock_bh(&xfrm_state_lock); |
427 | si->sadcnt = xfrm_state_num; | 427 | si->sadcnt = xfrm_state_num; |
diff --git a/net/xfrm/xfrm_user.c b/net/xfrm/xfrm_user.c index 4210d91624cd..b14c7e590c31 100644 --- a/net/xfrm/xfrm_user.c +++ b/net/xfrm/xfrm_user.c | |||
@@ -674,7 +674,9 @@ static struct sk_buff *xfrm_state_netlink(struct sk_buff *in_skb, | |||
674 | 674 | ||
675 | static int build_spdinfo(struct sk_buff *skb, u32 pid, u32 seq, u32 flags) | 675 | static int build_spdinfo(struct sk_buff *skb, u32 pid, u32 seq, u32 flags) |
676 | { | 676 | { |
677 | struct xfrm_spdinfo si; | 677 | struct xfrmk_spdinfo si; |
678 | struct xfrmu_spdinfo spc; | ||
679 | struct xfrmu_spdhinfo sph; | ||
678 | struct nlmsghdr *nlh; | 680 | struct nlmsghdr *nlh; |
679 | u32 *f; | 681 | u32 *f; |
680 | 682 | ||
@@ -685,23 +687,17 @@ static int build_spdinfo(struct sk_buff *skb, u32 pid, u32 seq, u32 flags) | |||
685 | f = nlmsg_data(nlh); | 687 | f = nlmsg_data(nlh); |
686 | *f = flags; | 688 | *f = flags; |
687 | xfrm_spd_getinfo(&si); | 689 | xfrm_spd_getinfo(&si); |
688 | 690 | spc.incnt = si.incnt; | |
689 | if (flags & XFRM_SPD_HMASK) | 691 | spc.outcnt = si.outcnt; |
690 | NLA_PUT_U32(skb, XFRMA_SPDHMASK, si.spdhcnt); | 692 | spc.fwdcnt = si.fwdcnt; |
691 | if (flags & XFRM_SPD_HMAX) | 693 | spc.inscnt = si.inscnt; |
692 | NLA_PUT_U32(skb, XFRMA_SPDHMAX, si.spdhmcnt); | 694 | spc.outscnt = si.outscnt; |
693 | if (flags & XFRM_SPD_ICNT) | 695 | spc.fwdscnt = si.fwdscnt; |
694 | NLA_PUT_U32(skb, XFRMA_SPDICNT, si.incnt); | 696 | sph.spdhcnt = si.spdhcnt; |
695 | if (flags & XFRM_SPD_OCNT) | 697 | sph.spdhmcnt = si.spdhmcnt; |
696 | NLA_PUT_U32(skb, XFRMA_SPDOCNT, si.outcnt); | 698 | |
697 | if (flags & XFRM_SPD_FCNT) | 699 | NLA_PUT(skb, XFRMA_SPD_INFO, sizeof(spc), &spc); |
698 | NLA_PUT_U32(skb, XFRMA_SPDFCNT, si.fwdcnt); | 700 | NLA_PUT(skb, XFRMA_SPD_HINFO, sizeof(sph), &sph); |
699 | if (flags & XFRM_SPD_ISCNT) | ||
700 | NLA_PUT_U32(skb, XFRMA_SPDISCNT, si.inscnt); | ||
701 | if (flags & XFRM_SPD_OSCNT) | ||
702 | NLA_PUT_U32(skb, XFRMA_SPDOSCNT, si.inscnt); | ||
703 | if (flags & XFRM_SPD_FSCNT) | ||
704 | NLA_PUT_U32(skb, XFRMA_SPDFSCNT, si.inscnt); | ||
705 | 701 | ||
706 | return nlmsg_end(skb, nlh); | 702 | return nlmsg_end(skb, nlh); |
707 | 703 | ||
@@ -719,23 +715,8 @@ static int xfrm_get_spdinfo(struct sk_buff *skb, struct nlmsghdr *nlh, | |||
719 | u32 seq = nlh->nlmsg_seq; | 715 | u32 seq = nlh->nlmsg_seq; |
720 | int len = NLMSG_LENGTH(sizeof(u32)); | 716 | int len = NLMSG_LENGTH(sizeof(u32)); |
721 | 717 | ||
722 | 718 | len += RTA_SPACE(sizeof(struct xfrmu_spdinfo)); | |
723 | if (*flags & XFRM_SPD_HMASK) | 719 | len += RTA_SPACE(sizeof(struct xfrmu_spdhinfo)); |
724 | len += RTA_SPACE(sizeof(u32)); | ||
725 | if (*flags & XFRM_SPD_HMAX) | ||
726 | len += RTA_SPACE(sizeof(u32)); | ||
727 | if (*flags & XFRM_SPD_ICNT) | ||
728 | len += RTA_SPACE(sizeof(u32)); | ||
729 | if (*flags & XFRM_SPD_OCNT) | ||
730 | len += RTA_SPACE(sizeof(u32)); | ||
731 | if (*flags & XFRM_SPD_FCNT) | ||
732 | len += RTA_SPACE(sizeof(u32)); | ||
733 | if (*flags & XFRM_SPD_ISCNT) | ||
734 | len += RTA_SPACE(sizeof(u32)); | ||
735 | if (*flags & XFRM_SPD_OSCNT) | ||
736 | len += RTA_SPACE(sizeof(u32)); | ||
737 | if (*flags & XFRM_SPD_FSCNT) | ||
738 | len += RTA_SPACE(sizeof(u32)); | ||
739 | 720 | ||
740 | r_skb = alloc_skb(len, GFP_ATOMIC); | 721 | r_skb = alloc_skb(len, GFP_ATOMIC); |
741 | if (r_skb == NULL) | 722 | if (r_skb == NULL) |
@@ -749,7 +730,8 @@ static int xfrm_get_spdinfo(struct sk_buff *skb, struct nlmsghdr *nlh, | |||
749 | 730 | ||
750 | static int build_sadinfo(struct sk_buff *skb, u32 pid, u32 seq, u32 flags) | 731 | static int build_sadinfo(struct sk_buff *skb, u32 pid, u32 seq, u32 flags) |
751 | { | 732 | { |
752 | struct xfrm_sadinfo si; | 733 | struct xfrmk_sadinfo si; |
734 | struct xfrmu_sadhinfo sh; | ||
753 | struct nlmsghdr *nlh; | 735 | struct nlmsghdr *nlh; |
754 | u32 *f; | 736 | u32 *f; |
755 | 737 | ||
@@ -761,12 +743,11 @@ static int build_sadinfo(struct sk_buff *skb, u32 pid, u32 seq, u32 flags) | |||
761 | *f = flags; | 743 | *f = flags; |
762 | xfrm_sad_getinfo(&si); | 744 | xfrm_sad_getinfo(&si); |
763 | 745 | ||
764 | if (flags & XFRM_SAD_HMASK) | 746 | sh.sadhmcnt = si.sadhmcnt; |
765 | NLA_PUT_U32(skb, XFRMA_SADHMASK, si.sadhcnt); | 747 | sh.sadhcnt = si.sadhcnt; |
766 | if (flags & XFRM_SAD_HMAX) | 748 | |
767 | NLA_PUT_U32(skb, XFRMA_SADHMAX, si.sadhmcnt); | 749 | NLA_PUT_U32(skb, XFRMA_SAD_CNT, si.sadcnt); |
768 | if (flags & XFRM_SAD_CNT) | 750 | NLA_PUT(skb, XFRMA_SAD_HINFO, sizeof(sh), &sh); |
769 | NLA_PUT_U32(skb, XFRMA_SADCNT, si.sadcnt); | ||
770 | 751 | ||
771 | return nlmsg_end(skb, nlh); | 752 | return nlmsg_end(skb, nlh); |
772 | 753 | ||
@@ -784,12 +765,8 @@ static int xfrm_get_sadinfo(struct sk_buff *skb, struct nlmsghdr *nlh, | |||
784 | u32 seq = nlh->nlmsg_seq; | 765 | u32 seq = nlh->nlmsg_seq; |
785 | int len = NLMSG_LENGTH(sizeof(u32)); | 766 | int len = NLMSG_LENGTH(sizeof(u32)); |
786 | 767 | ||
787 | if (*flags & XFRM_SAD_HMASK) | 768 | len += RTA_SPACE(sizeof(struct xfrmu_sadhinfo)); |
788 | len += RTA_SPACE(sizeof(u32)); | 769 | len += RTA_SPACE(sizeof(u32)); |
789 | if (*flags & XFRM_SAD_HMAX) | ||
790 | len += RTA_SPACE(sizeof(u32)); | ||
791 | if (*flags & XFRM_SAD_CNT) | ||
792 | len += RTA_SPACE(sizeof(u32)); | ||
793 | 770 | ||
794 | r_skb = alloc_skb(len, GFP_ATOMIC); | 771 | r_skb = alloc_skb(len, GFP_ATOMIC); |
795 | 772 | ||