diff options
-rw-r--r-- | net/tipc/bcast.c | 10 | ||||
-rw-r--r-- | net/tipc/group.c | 4 | ||||
-rw-r--r-- | net/tipc/link.c | 14 | ||||
-rw-r--r-- | net/tipc/name_distr.c | 2 | ||||
-rw-r--r-- | net/tipc/node.c | 7 | ||||
-rw-r--r-- | net/tipc/socket.c | 14 |
6 files changed, 26 insertions, 25 deletions
diff --git a/net/tipc/bcast.c b/net/tipc/bcast.c index 34f3e5641438..6ef1abdd525f 100644 --- a/net/tipc/bcast.c +++ b/net/tipc/bcast.c | |||
@@ -185,7 +185,7 @@ static void tipc_bcbase_xmit(struct net *net, struct sk_buff_head *xmitq) | |||
185 | } | 185 | } |
186 | 186 | ||
187 | /* We have to transmit across all bearers */ | 187 | /* We have to transmit across all bearers */ |
188 | skb_queue_head_init(&_xmitq); | 188 | __skb_queue_head_init(&_xmitq); |
189 | for (bearer_id = 0; bearer_id < MAX_BEARERS; bearer_id++) { | 189 | for (bearer_id = 0; bearer_id < MAX_BEARERS; bearer_id++) { |
190 | if (!bb->dests[bearer_id]) | 190 | if (!bb->dests[bearer_id]) |
191 | continue; | 191 | continue; |
@@ -256,7 +256,7 @@ static int tipc_bcast_xmit(struct net *net, struct sk_buff_head *pkts, | |||
256 | struct sk_buff_head xmitq; | 256 | struct sk_buff_head xmitq; |
257 | int rc = 0; | 257 | int rc = 0; |
258 | 258 | ||
259 | skb_queue_head_init(&xmitq); | 259 | __skb_queue_head_init(&xmitq); |
260 | tipc_bcast_lock(net); | 260 | tipc_bcast_lock(net); |
261 | if (tipc_link_bc_peers(l)) | 261 | if (tipc_link_bc_peers(l)) |
262 | rc = tipc_link_xmit(l, pkts, &xmitq); | 262 | rc = tipc_link_xmit(l, pkts, &xmitq); |
@@ -286,7 +286,7 @@ static int tipc_rcast_xmit(struct net *net, struct sk_buff_head *pkts, | |||
286 | u32 dnode, selector; | 286 | u32 dnode, selector; |
287 | 287 | ||
288 | selector = msg_link_selector(buf_msg(skb_peek(pkts))); | 288 | selector = msg_link_selector(buf_msg(skb_peek(pkts))); |
289 | skb_queue_head_init(&_pkts); | 289 | __skb_queue_head_init(&_pkts); |
290 | 290 | ||
291 | list_for_each_entry_safe(dst, tmp, &dests->list, list) { | 291 | list_for_each_entry_safe(dst, tmp, &dests->list, list) { |
292 | dnode = dst->node; | 292 | dnode = dst->node; |
@@ -344,7 +344,7 @@ static int tipc_mcast_send_sync(struct net *net, struct sk_buff *skb, | |||
344 | msg_set_size(_hdr, MCAST_H_SIZE); | 344 | msg_set_size(_hdr, MCAST_H_SIZE); |
345 | msg_set_is_rcast(_hdr, !msg_is_rcast(hdr)); | 345 | msg_set_is_rcast(_hdr, !msg_is_rcast(hdr)); |
346 | 346 | ||
347 | skb_queue_head_init(&tmpq); | 347 | __skb_queue_head_init(&tmpq); |
348 | __skb_queue_tail(&tmpq, _skb); | 348 | __skb_queue_tail(&tmpq, _skb); |
349 | if (method->rcast) | 349 | if (method->rcast) |
350 | tipc_bcast_xmit(net, &tmpq, cong_link_cnt); | 350 | tipc_bcast_xmit(net, &tmpq, cong_link_cnt); |
@@ -378,7 +378,7 @@ int tipc_mcast_xmit(struct net *net, struct sk_buff_head *pkts, | |||
378 | int rc = 0; | 378 | int rc = 0; |
379 | 379 | ||
380 | skb_queue_head_init(&inputq); | 380 | skb_queue_head_init(&inputq); |
381 | skb_queue_head_init(&localq); | 381 | __skb_queue_head_init(&localq); |
382 | 382 | ||
383 | /* Clone packets before they are consumed by next call */ | 383 | /* Clone packets before they are consumed by next call */ |
384 | if (dests->local && !tipc_msg_reassemble(pkts, &localq)) { | 384 | if (dests->local && !tipc_msg_reassemble(pkts, &localq)) { |
diff --git a/net/tipc/group.c b/net/tipc/group.c index 5f98d38bcf08..89257e2a980d 100644 --- a/net/tipc/group.c +++ b/net/tipc/group.c | |||
@@ -199,7 +199,7 @@ void tipc_group_join(struct net *net, struct tipc_group *grp, int *sk_rcvbuf) | |||
199 | struct tipc_member *m, *tmp; | 199 | struct tipc_member *m, *tmp; |
200 | struct sk_buff_head xmitq; | 200 | struct sk_buff_head xmitq; |
201 | 201 | ||
202 | skb_queue_head_init(&xmitq); | 202 | __skb_queue_head_init(&xmitq); |
203 | rbtree_postorder_for_each_entry_safe(m, tmp, tree, tree_node) { | 203 | rbtree_postorder_for_each_entry_safe(m, tmp, tree, tree_node) { |
204 | tipc_group_proto_xmit(grp, m, GRP_JOIN_MSG, &xmitq); | 204 | tipc_group_proto_xmit(grp, m, GRP_JOIN_MSG, &xmitq); |
205 | tipc_group_update_member(m, 0); | 205 | tipc_group_update_member(m, 0); |
@@ -435,7 +435,7 @@ bool tipc_group_cong(struct tipc_group *grp, u32 dnode, u32 dport, | |||
435 | return true; | 435 | return true; |
436 | if (state == MBR_PENDING && adv == ADV_IDLE) | 436 | if (state == MBR_PENDING && adv == ADV_IDLE) |
437 | return true; | 437 | return true; |
438 | skb_queue_head_init(&xmitq); | 438 | __skb_queue_head_init(&xmitq); |
439 | tipc_group_proto_xmit(grp, m, GRP_ADV_MSG, &xmitq); | 439 | tipc_group_proto_xmit(grp, m, GRP_ADV_MSG, &xmitq); |
440 | tipc_node_distr_xmit(grp->net, &xmitq); | 440 | tipc_node_distr_xmit(grp->net, &xmitq); |
441 | return true; | 441 | return true; |
diff --git a/net/tipc/link.c b/net/tipc/link.c index dd3155b14654..289e848084ac 100644 --- a/net/tipc/link.c +++ b/net/tipc/link.c | |||
@@ -959,7 +959,7 @@ int tipc_link_xmit(struct tipc_link *l, struct sk_buff_head *list, | |||
959 | pr_warn("Too large msg, purging xmit list %d %d %d %d %d!\n", | 959 | pr_warn("Too large msg, purging xmit list %d %d %d %d %d!\n", |
960 | skb_queue_len(list), msg_user(hdr), | 960 | skb_queue_len(list), msg_user(hdr), |
961 | msg_type(hdr), msg_size(hdr), mtu); | 961 | msg_type(hdr), msg_size(hdr), mtu); |
962 | skb_queue_purge(list); | 962 | __skb_queue_purge(list); |
963 | return -EMSGSIZE; | 963 | return -EMSGSIZE; |
964 | } | 964 | } |
965 | 965 | ||
@@ -988,7 +988,7 @@ int tipc_link_xmit(struct tipc_link *l, struct sk_buff_head *list, | |||
988 | if (likely(skb_queue_len(transmq) < maxwin)) { | 988 | if (likely(skb_queue_len(transmq) < maxwin)) { |
989 | _skb = skb_clone(skb, GFP_ATOMIC); | 989 | _skb = skb_clone(skb, GFP_ATOMIC); |
990 | if (!_skb) { | 990 | if (!_skb) { |
991 | skb_queue_purge(list); | 991 | __skb_queue_purge(list); |
992 | return -ENOBUFS; | 992 | return -ENOBUFS; |
993 | } | 993 | } |
994 | __skb_dequeue(list); | 994 | __skb_dequeue(list); |
@@ -1668,7 +1668,7 @@ void tipc_link_create_dummy_tnl_msg(struct tipc_link *l, | |||
1668 | struct sk_buff *skb; | 1668 | struct sk_buff *skb; |
1669 | u32 dnode = l->addr; | 1669 | u32 dnode = l->addr; |
1670 | 1670 | ||
1671 | skb_queue_head_init(&tnlq); | 1671 | __skb_queue_head_init(&tnlq); |
1672 | skb = tipc_msg_create(TUNNEL_PROTOCOL, FAILOVER_MSG, | 1672 | skb = tipc_msg_create(TUNNEL_PROTOCOL, FAILOVER_MSG, |
1673 | INT_H_SIZE, BASIC_H_SIZE, | 1673 | INT_H_SIZE, BASIC_H_SIZE, |
1674 | dnode, onode, 0, 0, 0); | 1674 | dnode, onode, 0, 0, 0); |
@@ -1708,9 +1708,9 @@ void tipc_link_tnl_prepare(struct tipc_link *l, struct tipc_link *tnl, | |||
1708 | if (!tnl) | 1708 | if (!tnl) |
1709 | return; | 1709 | return; |
1710 | 1710 | ||
1711 | skb_queue_head_init(&tnlq); | 1711 | __skb_queue_head_init(&tnlq); |
1712 | skb_queue_head_init(&tmpxq); | 1712 | __skb_queue_head_init(&tmpxq); |
1713 | skb_queue_head_init(&frags); | 1713 | __skb_queue_head_init(&frags); |
1714 | 1714 | ||
1715 | /* At least one packet required for safe algorithm => add dummy */ | 1715 | /* At least one packet required for safe algorithm => add dummy */ |
1716 | skb = tipc_msg_create(TIPC_LOW_IMPORTANCE, TIPC_DIRECT_MSG, | 1716 | skb = tipc_msg_create(TIPC_LOW_IMPORTANCE, TIPC_DIRECT_MSG, |
@@ -1720,7 +1720,7 @@ void tipc_link_tnl_prepare(struct tipc_link *l, struct tipc_link *tnl, | |||
1720 | pr_warn("%sunable to create tunnel packet\n", link_co_err); | 1720 | pr_warn("%sunable to create tunnel packet\n", link_co_err); |
1721 | return; | 1721 | return; |
1722 | } | 1722 | } |
1723 | skb_queue_tail(&tnlq, skb); | 1723 | __skb_queue_tail(&tnlq, skb); |
1724 | tipc_link_xmit(l, &tnlq, &tmpxq); | 1724 | tipc_link_xmit(l, &tnlq, &tmpxq); |
1725 | __skb_queue_purge(&tmpxq); | 1725 | __skb_queue_purge(&tmpxq); |
1726 | 1726 | ||
diff --git a/net/tipc/name_distr.c b/net/tipc/name_distr.c index 44abc8e9c990..61219f0b9677 100644 --- a/net/tipc/name_distr.c +++ b/net/tipc/name_distr.c | |||
@@ -190,7 +190,7 @@ void tipc_named_node_up(struct net *net, u32 dnode) | |||
190 | struct name_table *nt = tipc_name_table(net); | 190 | struct name_table *nt = tipc_name_table(net); |
191 | struct sk_buff_head head; | 191 | struct sk_buff_head head; |
192 | 192 | ||
193 | skb_queue_head_init(&head); | 193 | __skb_queue_head_init(&head); |
194 | 194 | ||
195 | read_lock_bh(&nt->cluster_scope_lock); | 195 | read_lock_bh(&nt->cluster_scope_lock); |
196 | named_distribute(net, &head, dnode, &nt->cluster_scope); | 196 | named_distribute(net, &head, dnode, &nt->cluster_scope); |
diff --git a/net/tipc/node.c b/net/tipc/node.c index 1bdcf0fc1a4d..c8f6177dd5a2 100644 --- a/net/tipc/node.c +++ b/net/tipc/node.c | |||
@@ -1444,13 +1444,14 @@ int tipc_node_xmit(struct net *net, struct sk_buff_head *list, | |||
1444 | 1444 | ||
1445 | if (in_own_node(net, dnode)) { | 1445 | if (in_own_node(net, dnode)) { |
1446 | tipc_loopback_trace(net, list); | 1446 | tipc_loopback_trace(net, list); |
1447 | spin_lock_init(&list->lock); | ||
1447 | tipc_sk_rcv(net, list); | 1448 | tipc_sk_rcv(net, list); |
1448 | return 0; | 1449 | return 0; |
1449 | } | 1450 | } |
1450 | 1451 | ||
1451 | n = tipc_node_find(net, dnode); | 1452 | n = tipc_node_find(net, dnode); |
1452 | if (unlikely(!n)) { | 1453 | if (unlikely(!n)) { |
1453 | skb_queue_purge(list); | 1454 | __skb_queue_purge(list); |
1454 | return -EHOSTUNREACH; | 1455 | return -EHOSTUNREACH; |
1455 | } | 1456 | } |
1456 | 1457 | ||
@@ -1459,7 +1460,7 @@ int tipc_node_xmit(struct net *net, struct sk_buff_head *list, | |||
1459 | if (unlikely(bearer_id == INVALID_BEARER_ID)) { | 1460 | if (unlikely(bearer_id == INVALID_BEARER_ID)) { |
1460 | tipc_node_read_unlock(n); | 1461 | tipc_node_read_unlock(n); |
1461 | tipc_node_put(n); | 1462 | tipc_node_put(n); |
1462 | skb_queue_purge(list); | 1463 | __skb_queue_purge(list); |
1463 | return -EHOSTUNREACH; | 1464 | return -EHOSTUNREACH; |
1464 | } | 1465 | } |
1465 | 1466 | ||
@@ -1491,7 +1492,7 @@ int tipc_node_xmit_skb(struct net *net, struct sk_buff *skb, u32 dnode, | |||
1491 | { | 1492 | { |
1492 | struct sk_buff_head head; | 1493 | struct sk_buff_head head; |
1493 | 1494 | ||
1494 | skb_queue_head_init(&head); | 1495 | __skb_queue_head_init(&head); |
1495 | __skb_queue_tail(&head, skb); | 1496 | __skb_queue_tail(&head, skb); |
1496 | tipc_node_xmit(net, &head, dnode, selector); | 1497 | tipc_node_xmit(net, &head, dnode, selector); |
1497 | return 0; | 1498 | return 0; |
diff --git a/net/tipc/socket.c b/net/tipc/socket.c index 83ae41d7e554..3b9f8cc328f5 100644 --- a/net/tipc/socket.c +++ b/net/tipc/socket.c | |||
@@ -809,7 +809,7 @@ static int tipc_sendmcast(struct socket *sock, struct tipc_name_seq *seq, | |||
809 | msg_set_nameupper(hdr, seq->upper); | 809 | msg_set_nameupper(hdr, seq->upper); |
810 | 810 | ||
811 | /* Build message as chain of buffers */ | 811 | /* Build message as chain of buffers */ |
812 | skb_queue_head_init(&pkts); | 812 | __skb_queue_head_init(&pkts); |
813 | rc = tipc_msg_build(hdr, msg, 0, dlen, mtu, &pkts); | 813 | rc = tipc_msg_build(hdr, msg, 0, dlen, mtu, &pkts); |
814 | 814 | ||
815 | /* Send message if build was successful */ | 815 | /* Send message if build was successful */ |
@@ -853,7 +853,7 @@ static int tipc_send_group_msg(struct net *net, struct tipc_sock *tsk, | |||
853 | msg_set_grp_bc_seqno(hdr, bc_snd_nxt); | 853 | msg_set_grp_bc_seqno(hdr, bc_snd_nxt); |
854 | 854 | ||
855 | /* Build message as chain of buffers */ | 855 | /* Build message as chain of buffers */ |
856 | skb_queue_head_init(&pkts); | 856 | __skb_queue_head_init(&pkts); |
857 | mtu = tipc_node_get_mtu(net, dnode, tsk->portid); | 857 | mtu = tipc_node_get_mtu(net, dnode, tsk->portid); |
858 | rc = tipc_msg_build(hdr, m, 0, dlen, mtu, &pkts); | 858 | rc = tipc_msg_build(hdr, m, 0, dlen, mtu, &pkts); |
859 | if (unlikely(rc != dlen)) | 859 | if (unlikely(rc != dlen)) |
@@ -1058,7 +1058,7 @@ static int tipc_send_group_bcast(struct socket *sock, struct msghdr *m, | |||
1058 | msg_set_grp_bc_ack_req(hdr, ack); | 1058 | msg_set_grp_bc_ack_req(hdr, ack); |
1059 | 1059 | ||
1060 | /* Build message as chain of buffers */ | 1060 | /* Build message as chain of buffers */ |
1061 | skb_queue_head_init(&pkts); | 1061 | __skb_queue_head_init(&pkts); |
1062 | rc = tipc_msg_build(hdr, m, 0, dlen, mtu, &pkts); | 1062 | rc = tipc_msg_build(hdr, m, 0, dlen, mtu, &pkts); |
1063 | if (unlikely(rc != dlen)) | 1063 | if (unlikely(rc != dlen)) |
1064 | return rc; | 1064 | return rc; |
@@ -1387,7 +1387,7 @@ static int __tipc_sendmsg(struct socket *sock, struct msghdr *m, size_t dlen) | |||
1387 | if (unlikely(rc)) | 1387 | if (unlikely(rc)) |
1388 | return rc; | 1388 | return rc; |
1389 | 1389 | ||
1390 | skb_queue_head_init(&pkts); | 1390 | __skb_queue_head_init(&pkts); |
1391 | mtu = tipc_node_get_mtu(net, dnode, tsk->portid); | 1391 | mtu = tipc_node_get_mtu(net, dnode, tsk->portid); |
1392 | rc = tipc_msg_build(hdr, m, 0, dlen, mtu, &pkts); | 1392 | rc = tipc_msg_build(hdr, m, 0, dlen, mtu, &pkts); |
1393 | if (unlikely(rc != dlen)) | 1393 | if (unlikely(rc != dlen)) |
@@ -1445,7 +1445,7 @@ static int __tipc_sendstream(struct socket *sock, struct msghdr *m, size_t dlen) | |||
1445 | int send, sent = 0; | 1445 | int send, sent = 0; |
1446 | int rc = 0; | 1446 | int rc = 0; |
1447 | 1447 | ||
1448 | skb_queue_head_init(&pkts); | 1448 | __skb_queue_head_init(&pkts); |
1449 | 1449 | ||
1450 | if (unlikely(dlen > INT_MAX)) | 1450 | if (unlikely(dlen > INT_MAX)) |
1451 | return -EMSGSIZE; | 1451 | return -EMSGSIZE; |
@@ -1805,7 +1805,7 @@ static int tipc_recvmsg(struct socket *sock, struct msghdr *m, | |||
1805 | 1805 | ||
1806 | /* Send group flow control advertisement when applicable */ | 1806 | /* Send group flow control advertisement when applicable */ |
1807 | if (tsk->group && msg_in_group(hdr) && !grp_evt) { | 1807 | if (tsk->group && msg_in_group(hdr) && !grp_evt) { |
1808 | skb_queue_head_init(&xmitq); | 1808 | __skb_queue_head_init(&xmitq); |
1809 | tipc_group_update_rcv_win(tsk->group, tsk_blocks(hlen + dlen), | 1809 | tipc_group_update_rcv_win(tsk->group, tsk_blocks(hlen + dlen), |
1810 | msg_orignode(hdr), msg_origport(hdr), | 1810 | msg_orignode(hdr), msg_origport(hdr), |
1811 | &xmitq); | 1811 | &xmitq); |
@@ -2674,7 +2674,7 @@ static void tipc_sk_timeout(struct timer_list *t) | |||
2674 | struct sk_buff_head list; | 2674 | struct sk_buff_head list; |
2675 | int rc = 0; | 2675 | int rc = 0; |
2676 | 2676 | ||
2677 | skb_queue_head_init(&list); | 2677 | __skb_queue_head_init(&list); |
2678 | bh_lock_sock(sk); | 2678 | bh_lock_sock(sk); |
2679 | 2679 | ||
2680 | /* Try again later if socket is busy */ | 2680 | /* Try again later if socket is busy */ |