diff options
Diffstat (limited to 'net/tipc/link.c')
-rw-r--r-- | net/tipc/link.c | 3167 |
1 files changed, 3167 insertions, 0 deletions
diff --git a/net/tipc/link.c b/net/tipc/link.c new file mode 100644 index 000000000000..7265f4be4766 --- /dev/null +++ b/net/tipc/link.c | |||
@@ -0,0 +1,3167 @@ | |||
1 | /* | ||
2 | * net/tipc/link.c: TIPC link code | ||
3 | * | ||
4 | * Copyright (c) 1996-2006, Ericsson AB | ||
5 | * Copyright (c) 2004-2005, Wind River Systems | ||
6 | * All rights reserved. | ||
7 | * | ||
8 | * Redistribution and use in source and binary forms, with or without | ||
9 | * modification, are permitted provided that the following conditions are met: | ||
10 | * | ||
11 | * 1. Redistributions of source code must retain the above copyright | ||
12 | * notice, this list of conditions and the following disclaimer. | ||
13 | * 2. Redistributions in binary form must reproduce the above copyright | ||
14 | * notice, this list of conditions and the following disclaimer in the | ||
15 | * documentation and/or other materials provided with the distribution. | ||
16 | * 3. Neither the names of the copyright holders nor the names of its | ||
17 | * contributors may be used to endorse or promote products derived from | ||
18 | * this software without specific prior written permission. | ||
19 | * | ||
20 | * Alternatively, this software may be distributed under the terms of the | ||
21 | * GNU General Public License ("GPL") version 2 as published by the Free | ||
22 | * Software Foundation. | ||
23 | * | ||
24 | * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" | ||
25 | * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE | ||
26 | * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE | ||
27 | * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE | ||
28 | * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR | ||
29 | * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF | ||
30 | * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS | ||
31 | * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN | ||
32 | * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) | ||
33 | * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE | ||
34 | * POSSIBILITY OF SUCH DAMAGE. | ||
35 | */ | ||
36 | |||
37 | #include "core.h" | ||
38 | #include "dbg.h" | ||
39 | #include "link.h" | ||
40 | #include "net.h" | ||
41 | #include "node.h" | ||
42 | #include "port.h" | ||
43 | #include "addr.h" | ||
44 | #include "node_subscr.h" | ||
45 | #include "name_distr.h" | ||
46 | #include "bearer.h" | ||
47 | #include "name_table.h" | ||
48 | #include "discover.h" | ||
49 | #include "config.h" | ||
50 | #include "bcast.h" | ||
51 | |||
52 | |||
53 | /* | ||
54 | * Limit for deferred reception queue: | ||
55 | */ | ||
56 | |||
57 | #define DEF_QUEUE_LIMIT 256u | ||
58 | |||
59 | /* | ||
60 | * Link state events: | ||
61 | */ | ||
62 | |||
63 | #define STARTING_EVT 856384768 /* link processing trigger */ | ||
64 | #define TRAFFIC_MSG_EVT 560815u /* rx'd ??? */ | ||
65 | #define TIMEOUT_EVT 560817u /* link timer expired */ | ||
66 | |||
67 | /* | ||
68 | * The following two 'message types' is really just implementation | ||
69 | * data conveniently stored in the message header. | ||
70 | * They must not be considered part of the protocol | ||
71 | */ | ||
72 | #define OPEN_MSG 0 | ||
73 | #define CLOSED_MSG 1 | ||
74 | |||
75 | /* | ||
76 | * State value stored in 'exp_msg_count' | ||
77 | */ | ||
78 | |||
79 | #define START_CHANGEOVER 100000u | ||
80 | |||
81 | /** | ||
82 | * struct link_name - deconstructed link name | ||
83 | * @addr_local: network address of node at this end | ||
84 | * @if_local: name of interface at this end | ||
85 | * @addr_peer: network address of node at far end | ||
86 | * @if_peer: name of interface at far end | ||
87 | */ | ||
88 | |||
89 | struct link_name { | ||
90 | u32 addr_local; | ||
91 | char if_local[TIPC_MAX_IF_NAME]; | ||
92 | u32 addr_peer; | ||
93 | char if_peer[TIPC_MAX_IF_NAME]; | ||
94 | }; | ||
95 | |||
96 | #if 0 | ||
97 | |||
98 | /* LINK EVENT CODE IS NOT SUPPORTED AT PRESENT */ | ||
99 | |||
100 | /** | ||
101 | * struct link_event - link up/down event notification | ||
102 | */ | ||
103 | |||
104 | struct link_event { | ||
105 | u32 addr; | ||
106 | int up; | ||
107 | void (*fcn)(u32, char *, int); | ||
108 | char name[TIPC_MAX_LINK_NAME]; | ||
109 | }; | ||
110 | |||
111 | #endif | ||
112 | |||
113 | static void link_handle_out_of_seq_msg(struct link *l_ptr, | ||
114 | struct sk_buff *buf); | ||
115 | static void link_recv_proto_msg(struct link *l_ptr, struct sk_buff *buf); | ||
116 | static int link_recv_changeover_msg(struct link **l_ptr, struct sk_buff **buf); | ||
117 | static void link_set_supervision_props(struct link *l_ptr, u32 tolerance); | ||
118 | static int link_send_sections_long(struct port *sender, | ||
119 | struct iovec const *msg_sect, | ||
120 | u32 num_sect, u32 destnode); | ||
121 | static void link_check_defragm_bufs(struct link *l_ptr); | ||
122 | static void link_state_event(struct link *l_ptr, u32 event); | ||
123 | static void link_reset_statistics(struct link *l_ptr); | ||
124 | static void link_print(struct link *l_ptr, struct print_buf *buf, | ||
125 | const char *str); | ||
126 | |||
127 | /* | ||
128 | * Debugging code used by link routines only | ||
129 | * | ||
130 | * When debugging link problems on a system that has multiple links, | ||
131 | * the standard TIPC debugging routines may not be useful since they | ||
132 | * allow the output from multiple links to be intermixed. For this reason | ||
133 | * routines of the form "dbg_link_XXX()" have been created that will capture | ||
134 | * debug info into a link's personal print buffer, which can then be dumped | ||
135 | * into the TIPC system log (LOG) upon request. | ||
136 | * | ||
137 | * To enable per-link debugging, use LINK_LOG_BUF_SIZE to specify the size | ||
138 | * of the print buffer used by each link. If LINK_LOG_BUF_SIZE is set to 0, | ||
139 | * the dbg_link_XXX() routines simply send their output to the standard | ||
140 | * debug print buffer (DBG_OUTPUT), if it has been defined; this can be useful | ||
141 | * when there is only a single link in the system being debugged. | ||
142 | * | ||
143 | * Notes: | ||
144 | * - When enabled, LINK_LOG_BUF_SIZE should be set to at least 1000 (bytes) | ||
145 | * - "l_ptr" must be valid when using dbg_link_XXX() macros | ||
146 | */ | ||
147 | |||
148 | #define LINK_LOG_BUF_SIZE 0 | ||
149 | |||
150 | #define dbg_link(fmt, arg...) do {if (LINK_LOG_BUF_SIZE) tipc_printf(&l_ptr->print_buf, fmt, ## arg); } while(0) | ||
151 | #define dbg_link_msg(msg, txt) do {if (LINK_LOG_BUF_SIZE) msg_print(&l_ptr->print_buf, msg, txt); } while(0) | ||
152 | #define dbg_link_state(txt) do {if (LINK_LOG_BUF_SIZE) link_print(l_ptr, &l_ptr->print_buf, txt); } while(0) | ||
153 | #define dbg_link_dump() do { \ | ||
154 | if (LINK_LOG_BUF_SIZE) { \ | ||
155 | tipc_printf(LOG, "\n\nDumping link <%s>:\n", l_ptr->name); \ | ||
156 | printbuf_move(LOG, &l_ptr->print_buf); \ | ||
157 | } \ | ||
158 | } while (0) | ||
159 | |||
160 | static inline void dbg_print_link(struct link *l_ptr, const char *str) | ||
161 | { | ||
162 | if (DBG_OUTPUT) | ||
163 | link_print(l_ptr, DBG_OUTPUT, str); | ||
164 | } | ||
165 | |||
166 | static inline void dbg_print_buf_chain(struct sk_buff *root_buf) | ||
167 | { | ||
168 | if (DBG_OUTPUT) { | ||
169 | struct sk_buff *buf = root_buf; | ||
170 | |||
171 | while (buf) { | ||
172 | msg_dbg(buf_msg(buf), "In chain: "); | ||
173 | buf = buf->next; | ||
174 | } | ||
175 | } | ||
176 | } | ||
177 | |||
178 | /* | ||
179 | * Simple inlined link routines | ||
180 | */ | ||
181 | |||
182 | static inline unsigned int align(unsigned int i) | ||
183 | { | ||
184 | return (i + 3) & ~3u; | ||
185 | } | ||
186 | |||
187 | static inline int link_working_working(struct link *l_ptr) | ||
188 | { | ||
189 | return (l_ptr->state == WORKING_WORKING); | ||
190 | } | ||
191 | |||
192 | static inline int link_working_unknown(struct link *l_ptr) | ||
193 | { | ||
194 | return (l_ptr->state == WORKING_UNKNOWN); | ||
195 | } | ||
196 | |||
197 | static inline int link_reset_unknown(struct link *l_ptr) | ||
198 | { | ||
199 | return (l_ptr->state == RESET_UNKNOWN); | ||
200 | } | ||
201 | |||
202 | static inline int link_reset_reset(struct link *l_ptr) | ||
203 | { | ||
204 | return (l_ptr->state == RESET_RESET); | ||
205 | } | ||
206 | |||
207 | static inline int link_blocked(struct link *l_ptr) | ||
208 | { | ||
209 | return (l_ptr->exp_msg_count || l_ptr->blocked); | ||
210 | } | ||
211 | |||
212 | static inline int link_congested(struct link *l_ptr) | ||
213 | { | ||
214 | return (l_ptr->out_queue_size >= l_ptr->queue_limit[0]); | ||
215 | } | ||
216 | |||
217 | static inline u32 link_max_pkt(struct link *l_ptr) | ||
218 | { | ||
219 | return l_ptr->max_pkt; | ||
220 | } | ||
221 | |||
222 | static inline void link_init_max_pkt(struct link *l_ptr) | ||
223 | { | ||
224 | u32 max_pkt; | ||
225 | |||
226 | max_pkt = (l_ptr->b_ptr->publ.mtu & ~3); | ||
227 | if (max_pkt > MAX_MSG_SIZE) | ||
228 | max_pkt = MAX_MSG_SIZE; | ||
229 | |||
230 | l_ptr->max_pkt_target = max_pkt; | ||
231 | if (l_ptr->max_pkt_target < MAX_PKT_DEFAULT) | ||
232 | l_ptr->max_pkt = l_ptr->max_pkt_target; | ||
233 | else | ||
234 | l_ptr->max_pkt = MAX_PKT_DEFAULT; | ||
235 | |||
236 | l_ptr->max_pkt_probes = 0; | ||
237 | } | ||
238 | |||
239 | static inline u32 link_next_sent(struct link *l_ptr) | ||
240 | { | ||
241 | if (l_ptr->next_out) | ||
242 | return msg_seqno(buf_msg(l_ptr->next_out)); | ||
243 | return mod(l_ptr->next_out_no); | ||
244 | } | ||
245 | |||
246 | static inline u32 link_last_sent(struct link *l_ptr) | ||
247 | { | ||
248 | return mod(link_next_sent(l_ptr) - 1); | ||
249 | } | ||
250 | |||
251 | /* | ||
252 | * Simple non-inlined link routines (i.e. referenced outside this file) | ||
253 | */ | ||
254 | |||
255 | int link_is_up(struct link *l_ptr) | ||
256 | { | ||
257 | if (!l_ptr) | ||
258 | return 0; | ||
259 | return (link_working_working(l_ptr) || link_working_unknown(l_ptr)); | ||
260 | } | ||
261 | |||
262 | int link_is_active(struct link *l_ptr) | ||
263 | { | ||
264 | return ((l_ptr->owner->active_links[0] == l_ptr) || | ||
265 | (l_ptr->owner->active_links[1] == l_ptr)); | ||
266 | } | ||
267 | |||
268 | /** | ||
269 | * link_name_validate - validate & (optionally) deconstruct link name | ||
270 | * @name - ptr to link name string | ||
271 | * @name_parts - ptr to area for link name components (or NULL if not needed) | ||
272 | * | ||
273 | * Returns 1 if link name is valid, otherwise 0. | ||
274 | */ | ||
275 | |||
276 | static int link_name_validate(const char *name, struct link_name *name_parts) | ||
277 | { | ||
278 | char name_copy[TIPC_MAX_LINK_NAME]; | ||
279 | char *addr_local; | ||
280 | char *if_local; | ||
281 | char *addr_peer; | ||
282 | char *if_peer; | ||
283 | char dummy; | ||
284 | u32 z_local, c_local, n_local; | ||
285 | u32 z_peer, c_peer, n_peer; | ||
286 | u32 if_local_len; | ||
287 | u32 if_peer_len; | ||
288 | |||
289 | /* copy link name & ensure length is OK */ | ||
290 | |||
291 | name_copy[TIPC_MAX_LINK_NAME - 1] = 0; | ||
292 | /* need above in case non-Posix strncpy() doesn't pad with nulls */ | ||
293 | strncpy(name_copy, name, TIPC_MAX_LINK_NAME); | ||
294 | if (name_copy[TIPC_MAX_LINK_NAME - 1] != 0) | ||
295 | return 0; | ||
296 | |||
297 | /* ensure all component parts of link name are present */ | ||
298 | |||
299 | addr_local = name_copy; | ||
300 | if ((if_local = strchr(addr_local, ':')) == NULL) | ||
301 | return 0; | ||
302 | *(if_local++) = 0; | ||
303 | if ((addr_peer = strchr(if_local, '-')) == NULL) | ||
304 | return 0; | ||
305 | *(addr_peer++) = 0; | ||
306 | if_local_len = addr_peer - if_local; | ||
307 | if ((if_peer = strchr(addr_peer, ':')) == NULL) | ||
308 | return 0; | ||
309 | *(if_peer++) = 0; | ||
310 | if_peer_len = strlen(if_peer) + 1; | ||
311 | |||
312 | /* validate component parts of link name */ | ||
313 | |||
314 | if ((sscanf(addr_local, "%u.%u.%u%c", | ||
315 | &z_local, &c_local, &n_local, &dummy) != 3) || | ||
316 | (sscanf(addr_peer, "%u.%u.%u%c", | ||
317 | &z_peer, &c_peer, &n_peer, &dummy) != 3) || | ||
318 | (z_local > 255) || (c_local > 4095) || (n_local > 4095) || | ||
319 | (z_peer > 255) || (c_peer > 4095) || (n_peer > 4095) || | ||
320 | (if_local_len <= 1) || (if_local_len > TIPC_MAX_IF_NAME) || | ||
321 | (if_peer_len <= 1) || (if_peer_len > TIPC_MAX_IF_NAME) || | ||
322 | (strspn(if_local, tipc_alphabet) != (if_local_len - 1)) || | ||
323 | (strspn(if_peer, tipc_alphabet) != (if_peer_len - 1))) | ||
324 | return 0; | ||
325 | |||
326 | /* return link name components, if necessary */ | ||
327 | |||
328 | if (name_parts) { | ||
329 | name_parts->addr_local = tipc_addr(z_local, c_local, n_local); | ||
330 | strcpy(name_parts->if_local, if_local); | ||
331 | name_parts->addr_peer = tipc_addr(z_peer, c_peer, n_peer); | ||
332 | strcpy(name_parts->if_peer, if_peer); | ||
333 | } | ||
334 | return 1; | ||
335 | } | ||
336 | |||
337 | /** | ||
338 | * link_timeout - handle expiration of link timer | ||
339 | * @l_ptr: pointer to link | ||
340 | * | ||
341 | * This routine must not grab "net_lock" to avoid a potential deadlock conflict | ||
342 | * with link_delete(). (There is no risk that the node will be deleted by | ||
343 | * another thread because link_delete() always cancels the link timer before | ||
344 | * node_delete() is called.) | ||
345 | */ | ||
346 | |||
347 | static void link_timeout(struct link *l_ptr) | ||
348 | { | ||
349 | node_lock(l_ptr->owner); | ||
350 | |||
351 | /* update counters used in statistical profiling of send traffic */ | ||
352 | |||
353 | l_ptr->stats.accu_queue_sz += l_ptr->out_queue_size; | ||
354 | l_ptr->stats.queue_sz_counts++; | ||
355 | |||
356 | if (l_ptr->out_queue_size > l_ptr->stats.max_queue_sz) | ||
357 | l_ptr->stats.max_queue_sz = l_ptr->out_queue_size; | ||
358 | |||
359 | if (l_ptr->first_out) { | ||
360 | struct tipc_msg *msg = buf_msg(l_ptr->first_out); | ||
361 | u32 length = msg_size(msg); | ||
362 | |||
363 | if ((msg_user(msg) == MSG_FRAGMENTER) | ||
364 | && (msg_type(msg) == FIRST_FRAGMENT)) { | ||
365 | length = msg_size(msg_get_wrapped(msg)); | ||
366 | } | ||
367 | if (length) { | ||
368 | l_ptr->stats.msg_lengths_total += length; | ||
369 | l_ptr->stats.msg_length_counts++; | ||
370 | if (length <= 64) | ||
371 | l_ptr->stats.msg_length_profile[0]++; | ||
372 | else if (length <= 256) | ||
373 | l_ptr->stats.msg_length_profile[1]++; | ||
374 | else if (length <= 1024) | ||
375 | l_ptr->stats.msg_length_profile[2]++; | ||
376 | else if (length <= 4096) | ||
377 | l_ptr->stats.msg_length_profile[3]++; | ||
378 | else if (length <= 16384) | ||
379 | l_ptr->stats.msg_length_profile[4]++; | ||
380 | else if (length <= 32768) | ||
381 | l_ptr->stats.msg_length_profile[5]++; | ||
382 | else | ||
383 | l_ptr->stats.msg_length_profile[6]++; | ||
384 | } | ||
385 | } | ||
386 | |||
387 | /* do all other link processing performed on a periodic basis */ | ||
388 | |||
389 | link_check_defragm_bufs(l_ptr); | ||
390 | |||
391 | link_state_event(l_ptr, TIMEOUT_EVT); | ||
392 | |||
393 | if (l_ptr->next_out) | ||
394 | link_push_queue(l_ptr); | ||
395 | |||
396 | node_unlock(l_ptr->owner); | ||
397 | } | ||
398 | |||
399 | static inline void link_set_timer(struct link *l_ptr, u32 time) | ||
400 | { | ||
401 | k_start_timer(&l_ptr->timer, time); | ||
402 | } | ||
403 | |||
404 | /** | ||
405 | * link_create - create a new link | ||
406 | * @b_ptr: pointer to associated bearer | ||
407 | * @peer: network address of node at other end of link | ||
408 | * @media_addr: media address to use when sending messages over link | ||
409 | * | ||
410 | * Returns pointer to link. | ||
411 | */ | ||
412 | |||
413 | struct link *link_create(struct bearer *b_ptr, const u32 peer, | ||
414 | const struct tipc_media_addr *media_addr) | ||
415 | { | ||
416 | struct link *l_ptr; | ||
417 | struct tipc_msg *msg; | ||
418 | char *if_name; | ||
419 | |||
420 | l_ptr = (struct link *)kmalloc(sizeof(*l_ptr), GFP_ATOMIC); | ||
421 | if (!l_ptr) { | ||
422 | warn("Memory squeeze; Failed to create link\n"); | ||
423 | return NULL; | ||
424 | } | ||
425 | memset(l_ptr, 0, sizeof(*l_ptr)); | ||
426 | |||
427 | l_ptr->addr = peer; | ||
428 | if_name = strchr(b_ptr->publ.name, ':') + 1; | ||
429 | sprintf(l_ptr->name, "%u.%u.%u:%s-%u.%u.%u:", | ||
430 | tipc_zone(tipc_own_addr), tipc_cluster(tipc_own_addr), | ||
431 | tipc_node(tipc_own_addr), | ||
432 | if_name, | ||
433 | tipc_zone(peer), tipc_cluster(peer), tipc_node(peer)); | ||
434 | /* note: peer i/f is appended to link name by reset/activate */ | ||
435 | memcpy(&l_ptr->media_addr, media_addr, sizeof(*media_addr)); | ||
436 | k_init_timer(&l_ptr->timer, (Handler)link_timeout, (unsigned long)l_ptr); | ||
437 | list_add_tail(&l_ptr->link_list, &b_ptr->links); | ||
438 | l_ptr->checkpoint = 1; | ||
439 | l_ptr->b_ptr = b_ptr; | ||
440 | link_set_supervision_props(l_ptr, b_ptr->media->tolerance); | ||
441 | l_ptr->state = RESET_UNKNOWN; | ||
442 | |||
443 | l_ptr->pmsg = (struct tipc_msg *)&l_ptr->proto_msg; | ||
444 | msg = l_ptr->pmsg; | ||
445 | msg_init(msg, LINK_PROTOCOL, RESET_MSG, TIPC_OK, INT_H_SIZE, l_ptr->addr); | ||
446 | msg_set_size(msg, sizeof(l_ptr->proto_msg)); | ||
447 | msg_set_session(msg, tipc_random); | ||
448 | msg_set_bearer_id(msg, b_ptr->identity); | ||
449 | strcpy((char *)msg_data(msg), if_name); | ||
450 | |||
451 | l_ptr->priority = b_ptr->priority; | ||
452 | link_set_queue_limits(l_ptr, b_ptr->media->window); | ||
453 | |||
454 | link_init_max_pkt(l_ptr); | ||
455 | |||
456 | l_ptr->next_out_no = 1; | ||
457 | INIT_LIST_HEAD(&l_ptr->waiting_ports); | ||
458 | |||
459 | link_reset_statistics(l_ptr); | ||
460 | |||
461 | l_ptr->owner = node_attach_link(l_ptr); | ||
462 | if (!l_ptr->owner) { | ||
463 | kfree(l_ptr); | ||
464 | return NULL; | ||
465 | } | ||
466 | |||
467 | if (LINK_LOG_BUF_SIZE) { | ||
468 | char *pb = kmalloc(LINK_LOG_BUF_SIZE, GFP_ATOMIC); | ||
469 | |||
470 | if (!pb) { | ||
471 | kfree(l_ptr); | ||
472 | warn("Memory squeeze; Failed to create link\n"); | ||
473 | return NULL; | ||
474 | } | ||
475 | printbuf_init(&l_ptr->print_buf, pb, LINK_LOG_BUF_SIZE); | ||
476 | } | ||
477 | |||
478 | k_signal((Handler)link_start, (unsigned long)l_ptr); | ||
479 | |||
480 | dbg("link_create(): tolerance = %u,cont intv = %u, abort_limit = %u\n", | ||
481 | l_ptr->tolerance, l_ptr->continuity_interval, l_ptr->abort_limit); | ||
482 | |||
483 | return l_ptr; | ||
484 | } | ||
485 | |||
486 | /** | ||
487 | * link_delete - delete a link | ||
488 | * @l_ptr: pointer to link | ||
489 | * | ||
490 | * Note: 'net_lock' is write_locked, bearer is locked. | ||
491 | * This routine must not grab the node lock until after link timer cancellation | ||
492 | * to avoid a potential deadlock situation. | ||
493 | */ | ||
494 | |||
495 | void link_delete(struct link *l_ptr) | ||
496 | { | ||
497 | if (!l_ptr) { | ||
498 | err("Attempt to delete non-existent link\n"); | ||
499 | return; | ||
500 | } | ||
501 | |||
502 | dbg("link_delete()\n"); | ||
503 | |||
504 | k_cancel_timer(&l_ptr->timer); | ||
505 | |||
506 | node_lock(l_ptr->owner); | ||
507 | link_reset(l_ptr); | ||
508 | node_detach_link(l_ptr->owner, l_ptr); | ||
509 | link_stop(l_ptr); | ||
510 | list_del_init(&l_ptr->link_list); | ||
511 | if (LINK_LOG_BUF_SIZE) | ||
512 | kfree(l_ptr->print_buf.buf); | ||
513 | node_unlock(l_ptr->owner); | ||
514 | k_term_timer(&l_ptr->timer); | ||
515 | kfree(l_ptr); | ||
516 | } | ||
517 | |||
518 | void link_start(struct link *l_ptr) | ||
519 | { | ||
520 | dbg("link_start %x\n", l_ptr); | ||
521 | link_state_event(l_ptr, STARTING_EVT); | ||
522 | } | ||
523 | |||
524 | /** | ||
525 | * link_schedule_port - schedule port for deferred sending | ||
526 | * @l_ptr: pointer to link | ||
527 | * @origport: reference to sending port | ||
528 | * @sz: amount of data to be sent | ||
529 | * | ||
530 | * Schedules port for renewed sending of messages after link congestion | ||
531 | * has abated. | ||
532 | */ | ||
533 | |||
534 | static int link_schedule_port(struct link *l_ptr, u32 origport, u32 sz) | ||
535 | { | ||
536 | struct port *p_ptr; | ||
537 | |||
538 | spin_lock_bh(&port_list_lock); | ||
539 | p_ptr = port_lock(origport); | ||
540 | if (p_ptr) { | ||
541 | if (!p_ptr->wakeup) | ||
542 | goto exit; | ||
543 | if (!list_empty(&p_ptr->wait_list)) | ||
544 | goto exit; | ||
545 | p_ptr->congested_link = l_ptr; | ||
546 | p_ptr->publ.congested = 1; | ||
547 | p_ptr->waiting_pkts = 1 + ((sz - 1) / link_max_pkt(l_ptr)); | ||
548 | list_add_tail(&p_ptr->wait_list, &l_ptr->waiting_ports); | ||
549 | l_ptr->stats.link_congs++; | ||
550 | exit: | ||
551 | port_unlock(p_ptr); | ||
552 | } | ||
553 | spin_unlock_bh(&port_list_lock); | ||
554 | return -ELINKCONG; | ||
555 | } | ||
556 | |||
557 | void link_wakeup_ports(struct link *l_ptr, int all) | ||
558 | { | ||
559 | struct port *p_ptr; | ||
560 | struct port *temp_p_ptr; | ||
561 | int win = l_ptr->queue_limit[0] - l_ptr->out_queue_size; | ||
562 | |||
563 | if (all) | ||
564 | win = 100000; | ||
565 | if (win <= 0) | ||
566 | return; | ||
567 | if (!spin_trylock_bh(&port_list_lock)) | ||
568 | return; | ||
569 | if (link_congested(l_ptr)) | ||
570 | goto exit; | ||
571 | list_for_each_entry_safe(p_ptr, temp_p_ptr, &l_ptr->waiting_ports, | ||
572 | wait_list) { | ||
573 | if (win <= 0) | ||
574 | break; | ||
575 | list_del_init(&p_ptr->wait_list); | ||
576 | p_ptr->congested_link = 0; | ||
577 | assert(p_ptr->wakeup); | ||
578 | spin_lock_bh(p_ptr->publ.lock); | ||
579 | p_ptr->publ.congested = 0; | ||
580 | p_ptr->wakeup(&p_ptr->publ); | ||
581 | win -= p_ptr->waiting_pkts; | ||
582 | spin_unlock_bh(p_ptr->publ.lock); | ||
583 | } | ||
584 | |||
585 | exit: | ||
586 | spin_unlock_bh(&port_list_lock); | ||
587 | } | ||
588 | |||
589 | /** | ||
590 | * link_release_outqueue - purge link's outbound message queue | ||
591 | * @l_ptr: pointer to link | ||
592 | */ | ||
593 | |||
594 | static void link_release_outqueue(struct link *l_ptr) | ||
595 | { | ||
596 | struct sk_buff *buf = l_ptr->first_out; | ||
597 | struct sk_buff *next; | ||
598 | |||
599 | while (buf) { | ||
600 | next = buf->next; | ||
601 | buf_discard(buf); | ||
602 | buf = next; | ||
603 | } | ||
604 | l_ptr->first_out = NULL; | ||
605 | l_ptr->out_queue_size = 0; | ||
606 | } | ||
607 | |||
608 | /** | ||
609 | * link_reset_fragments - purge link's inbound message fragments queue | ||
610 | * @l_ptr: pointer to link | ||
611 | */ | ||
612 | |||
613 | void link_reset_fragments(struct link *l_ptr) | ||
614 | { | ||
615 | struct sk_buff *buf = l_ptr->defragm_buf; | ||
616 | struct sk_buff *next; | ||
617 | |||
618 | while (buf) { | ||
619 | next = buf->next; | ||
620 | buf_discard(buf); | ||
621 | buf = next; | ||
622 | } | ||
623 | l_ptr->defragm_buf = NULL; | ||
624 | } | ||
625 | |||
626 | /** | ||
627 | * link_stop - purge all inbound and outbound messages associated with link | ||
628 | * @l_ptr: pointer to link | ||
629 | */ | ||
630 | |||
631 | void link_stop(struct link *l_ptr) | ||
632 | { | ||
633 | struct sk_buff *buf; | ||
634 | struct sk_buff *next; | ||
635 | |||
636 | buf = l_ptr->oldest_deferred_in; | ||
637 | while (buf) { | ||
638 | next = buf->next; | ||
639 | buf_discard(buf); | ||
640 | buf = next; | ||
641 | } | ||
642 | |||
643 | buf = l_ptr->first_out; | ||
644 | while (buf) { | ||
645 | next = buf->next; | ||
646 | buf_discard(buf); | ||
647 | buf = next; | ||
648 | } | ||
649 | |||
650 | link_reset_fragments(l_ptr); | ||
651 | |||
652 | buf_discard(l_ptr->proto_msg_queue); | ||
653 | l_ptr->proto_msg_queue = NULL; | ||
654 | } | ||
655 | |||
656 | #if 0 | ||
657 | |||
658 | /* LINK EVENT CODE IS NOT SUPPORTED AT PRESENT */ | ||
659 | |||
660 | static void link_recv_event(struct link_event *ev) | ||
661 | { | ||
662 | ev->fcn(ev->addr, ev->name, ev->up); | ||
663 | kfree(ev); | ||
664 | } | ||
665 | |||
666 | static void link_send_event(void (*fcn)(u32 a, char *n, int up), | ||
667 | struct link *l_ptr, int up) | ||
668 | { | ||
669 | struct link_event *ev; | ||
670 | |||
671 | ev = kmalloc(sizeof(*ev), GFP_ATOMIC); | ||
672 | if (!ev) { | ||
673 | warn("Link event allocation failure\n"); | ||
674 | return; | ||
675 | } | ||
676 | ev->addr = l_ptr->addr; | ||
677 | ev->up = up; | ||
678 | ev->fcn = fcn; | ||
679 | memcpy(ev->name, l_ptr->name, TIPC_MAX_LINK_NAME); | ||
680 | k_signal((Handler)link_recv_event, (unsigned long)ev); | ||
681 | } | ||
682 | |||
683 | #else | ||
684 | |||
685 | #define link_send_event(fcn, l_ptr, up) do { } while (0) | ||
686 | |||
687 | #endif | ||
688 | |||
689 | void link_reset(struct link *l_ptr) | ||
690 | { | ||
691 | struct sk_buff *buf; | ||
692 | u32 prev_state = l_ptr->state; | ||
693 | u32 checkpoint = l_ptr->next_in_no; | ||
694 | |||
695 | msg_set_session(l_ptr->pmsg, msg_session(l_ptr->pmsg) + 1); | ||
696 | |||
697 | /* Link is down, accept any session: */ | ||
698 | l_ptr->peer_session = 0; | ||
699 | |||
700 | /* Prepare for max packet size negotiation */ | ||
701 | link_init_max_pkt(l_ptr); | ||
702 | |||
703 | l_ptr->state = RESET_UNKNOWN; | ||
704 | dbg_link_state("Resetting Link\n"); | ||
705 | |||
706 | if ((prev_state == RESET_UNKNOWN) || (prev_state == RESET_RESET)) | ||
707 | return; | ||
708 | |||
709 | node_link_down(l_ptr->owner, l_ptr); | ||
710 | bearer_remove_dest(l_ptr->b_ptr, l_ptr->addr); | ||
711 | #if 0 | ||
712 | tipc_printf(CONS, "\nReset link <%s>\n", l_ptr->name); | ||
713 | dbg_link_dump(); | ||
714 | #endif | ||
715 | if (node_has_active_links(l_ptr->owner) && | ||
716 | l_ptr->owner->permit_changeover) { | ||
717 | l_ptr->reset_checkpoint = checkpoint; | ||
718 | l_ptr->exp_msg_count = START_CHANGEOVER; | ||
719 | } | ||
720 | |||
721 | /* Clean up all queues: */ | ||
722 | |||
723 | link_release_outqueue(l_ptr); | ||
724 | buf_discard(l_ptr->proto_msg_queue); | ||
725 | l_ptr->proto_msg_queue = NULL; | ||
726 | buf = l_ptr->oldest_deferred_in; | ||
727 | while (buf) { | ||
728 | struct sk_buff *next = buf->next; | ||
729 | buf_discard(buf); | ||
730 | buf = next; | ||
731 | } | ||
732 | if (!list_empty(&l_ptr->waiting_ports)) | ||
733 | link_wakeup_ports(l_ptr, 1); | ||
734 | |||
735 | l_ptr->retransm_queue_head = 0; | ||
736 | l_ptr->retransm_queue_size = 0; | ||
737 | l_ptr->last_out = NULL; | ||
738 | l_ptr->first_out = NULL; | ||
739 | l_ptr->next_out = NULL; | ||
740 | l_ptr->unacked_window = 0; | ||
741 | l_ptr->checkpoint = 1; | ||
742 | l_ptr->next_out_no = 1; | ||
743 | l_ptr->deferred_inqueue_sz = 0; | ||
744 | l_ptr->oldest_deferred_in = NULL; | ||
745 | l_ptr->newest_deferred_in = NULL; | ||
746 | l_ptr->fsm_msg_cnt = 0; | ||
747 | l_ptr->stale_count = 0; | ||
748 | link_reset_statistics(l_ptr); | ||
749 | |||
750 | link_send_event(cfg_link_event, l_ptr, 0); | ||
751 | if (!in_own_cluster(l_ptr->addr)) | ||
752 | link_send_event(disc_link_event, l_ptr, 0); | ||
753 | } | ||
754 | |||
755 | |||
756 | static void link_activate(struct link *l_ptr) | ||
757 | { | ||
758 | l_ptr->next_in_no = 1; | ||
759 | node_link_up(l_ptr->owner, l_ptr); | ||
760 | bearer_add_dest(l_ptr->b_ptr, l_ptr->addr); | ||
761 | link_send_event(cfg_link_event, l_ptr, 1); | ||
762 | if (!in_own_cluster(l_ptr->addr)) | ||
763 | link_send_event(disc_link_event, l_ptr, 1); | ||
764 | } | ||
765 | |||
766 | /** | ||
767 | * link_state_event - link finite state machine | ||
768 | * @l_ptr: pointer to link | ||
769 | * @event: state machine event to process | ||
770 | */ | ||
771 | |||
772 | static void link_state_event(struct link *l_ptr, unsigned event) | ||
773 | { | ||
774 | struct link *other; | ||
775 | u32 cont_intv = l_ptr->continuity_interval; | ||
776 | |||
777 | if (!l_ptr->started && (event != STARTING_EVT)) | ||
778 | return; /* Not yet. */ | ||
779 | |||
780 | if (link_blocked(l_ptr)) { | ||
781 | if (event == TIMEOUT_EVT) { | ||
782 | link_set_timer(l_ptr, cont_intv); | ||
783 | } | ||
784 | return; /* Changeover going on */ | ||
785 | } | ||
786 | dbg_link("STATE_EV: <%s> ", l_ptr->name); | ||
787 | |||
788 | switch (l_ptr->state) { | ||
789 | case WORKING_WORKING: | ||
790 | dbg_link("WW/"); | ||
791 | switch (event) { | ||
792 | case TRAFFIC_MSG_EVT: | ||
793 | dbg_link("TRF-"); | ||
794 | /* fall through */ | ||
795 | case ACTIVATE_MSG: | ||
796 | dbg_link("ACT\n"); | ||
797 | break; | ||
798 | case TIMEOUT_EVT: | ||
799 | dbg_link("TIM "); | ||
800 | if (l_ptr->next_in_no != l_ptr->checkpoint) { | ||
801 | l_ptr->checkpoint = l_ptr->next_in_no; | ||
802 | if (bclink_acks_missing(l_ptr->owner)) { | ||
803 | link_send_proto_msg(l_ptr, STATE_MSG, | ||
804 | 0, 0, 0, 0, 0); | ||
805 | l_ptr->fsm_msg_cnt++; | ||
806 | } else if (l_ptr->max_pkt < l_ptr->max_pkt_target) { | ||
807 | link_send_proto_msg(l_ptr, STATE_MSG, | ||
808 | 1, 0, 0, 0, 0); | ||
809 | l_ptr->fsm_msg_cnt++; | ||
810 | } | ||
811 | link_set_timer(l_ptr, cont_intv); | ||
812 | break; | ||
813 | } | ||
814 | dbg_link(" -> WU\n"); | ||
815 | l_ptr->state = WORKING_UNKNOWN; | ||
816 | l_ptr->fsm_msg_cnt = 0; | ||
817 | link_send_proto_msg(l_ptr, STATE_MSG, 1, 0, 0, 0, 0); | ||
818 | l_ptr->fsm_msg_cnt++; | ||
819 | link_set_timer(l_ptr, cont_intv / 4); | ||
820 | break; | ||
821 | case RESET_MSG: | ||
822 | dbg_link("RES -> RR\n"); | ||
823 | link_reset(l_ptr); | ||
824 | l_ptr->state = RESET_RESET; | ||
825 | l_ptr->fsm_msg_cnt = 0; | ||
826 | link_send_proto_msg(l_ptr, ACTIVATE_MSG, 0, 0, 0, 0, 0); | ||
827 | l_ptr->fsm_msg_cnt++; | ||
828 | link_set_timer(l_ptr, cont_intv); | ||
829 | break; | ||
830 | default: | ||
831 | err("Unknown link event %u in WW state\n", event); | ||
832 | } | ||
833 | break; | ||
834 | case WORKING_UNKNOWN: | ||
835 | dbg_link("WU/"); | ||
836 | switch (event) { | ||
837 | case TRAFFIC_MSG_EVT: | ||
838 | dbg_link("TRF-"); | ||
839 | case ACTIVATE_MSG: | ||
840 | dbg_link("ACT -> WW\n"); | ||
841 | l_ptr->state = WORKING_WORKING; | ||
842 | l_ptr->fsm_msg_cnt = 0; | ||
843 | link_set_timer(l_ptr, cont_intv); | ||
844 | break; | ||
845 | case RESET_MSG: | ||
846 | dbg_link("RES -> RR\n"); | ||
847 | link_reset(l_ptr); | ||
848 | l_ptr->state = RESET_RESET; | ||
849 | l_ptr->fsm_msg_cnt = 0; | ||
850 | link_send_proto_msg(l_ptr, ACTIVATE_MSG, 0, 0, 0, 0, 0); | ||
851 | l_ptr->fsm_msg_cnt++; | ||
852 | link_set_timer(l_ptr, cont_intv); | ||
853 | break; | ||
854 | case TIMEOUT_EVT: | ||
855 | dbg_link("TIM "); | ||
856 | if (l_ptr->next_in_no != l_ptr->checkpoint) { | ||
857 | dbg_link("-> WW \n"); | ||
858 | l_ptr->state = WORKING_WORKING; | ||
859 | l_ptr->fsm_msg_cnt = 0; | ||
860 | l_ptr->checkpoint = l_ptr->next_in_no; | ||
861 | if (bclink_acks_missing(l_ptr->owner)) { | ||
862 | link_send_proto_msg(l_ptr, STATE_MSG, | ||
863 | 0, 0, 0, 0, 0); | ||
864 | l_ptr->fsm_msg_cnt++; | ||
865 | } | ||
866 | link_set_timer(l_ptr, cont_intv); | ||
867 | } else if (l_ptr->fsm_msg_cnt < l_ptr->abort_limit) { | ||
868 | dbg_link("Probing %u/%u,timer = %u ms)\n", | ||
869 | l_ptr->fsm_msg_cnt, l_ptr->abort_limit, | ||
870 | cont_intv / 4); | ||
871 | link_send_proto_msg(l_ptr, STATE_MSG, | ||
872 | 1, 0, 0, 0, 0); | ||
873 | l_ptr->fsm_msg_cnt++; | ||
874 | link_set_timer(l_ptr, cont_intv / 4); | ||
875 | } else { /* Link has failed */ | ||
876 | dbg_link("-> RU (%u probes unanswered)\n", | ||
877 | l_ptr->fsm_msg_cnt); | ||
878 | link_reset(l_ptr); | ||
879 | l_ptr->state = RESET_UNKNOWN; | ||
880 | l_ptr->fsm_msg_cnt = 0; | ||
881 | link_send_proto_msg(l_ptr, RESET_MSG, | ||
882 | 0, 0, 0, 0, 0); | ||
883 | l_ptr->fsm_msg_cnt++; | ||
884 | link_set_timer(l_ptr, cont_intv); | ||
885 | } | ||
886 | break; | ||
887 | default: | ||
888 | err("Unknown link event %u in WU state\n", event); | ||
889 | } | ||
890 | break; | ||
891 | case RESET_UNKNOWN: | ||
892 | dbg_link("RU/"); | ||
893 | switch (event) { | ||
894 | case TRAFFIC_MSG_EVT: | ||
895 | dbg_link("TRF-\n"); | ||
896 | break; | ||
897 | case ACTIVATE_MSG: | ||
898 | other = l_ptr->owner->active_links[0]; | ||
899 | if (other && link_working_unknown(other)) { | ||
900 | dbg_link("ACT\n"); | ||
901 | break; | ||
902 | } | ||
903 | dbg_link("ACT -> WW\n"); | ||
904 | l_ptr->state = WORKING_WORKING; | ||
905 | l_ptr->fsm_msg_cnt = 0; | ||
906 | link_activate(l_ptr); | ||
907 | link_send_proto_msg(l_ptr, STATE_MSG, 1, 0, 0, 0, 0); | ||
908 | l_ptr->fsm_msg_cnt++; | ||
909 | link_set_timer(l_ptr, cont_intv); | ||
910 | break; | ||
911 | case RESET_MSG: | ||
912 | dbg_link("RES \n"); | ||
913 | dbg_link(" -> RR\n"); | ||
914 | l_ptr->state = RESET_RESET; | ||
915 | l_ptr->fsm_msg_cnt = 0; | ||
916 | link_send_proto_msg(l_ptr, ACTIVATE_MSG, 1, 0, 0, 0, 0); | ||
917 | l_ptr->fsm_msg_cnt++; | ||
918 | link_set_timer(l_ptr, cont_intv); | ||
919 | break; | ||
920 | case STARTING_EVT: | ||
921 | dbg_link("START-"); | ||
922 | l_ptr->started = 1; | ||
923 | /* fall through */ | ||
924 | case TIMEOUT_EVT: | ||
925 | dbg_link("TIM \n"); | ||
926 | link_send_proto_msg(l_ptr, RESET_MSG, 0, 0, 0, 0, 0); | ||
927 | l_ptr->fsm_msg_cnt++; | ||
928 | link_set_timer(l_ptr, cont_intv); | ||
929 | break; | ||
930 | default: | ||
931 | err("Unknown link event %u in RU state\n", event); | ||
932 | } | ||
933 | break; | ||
934 | case RESET_RESET: | ||
935 | dbg_link("RR/ "); | ||
936 | switch (event) { | ||
937 | case TRAFFIC_MSG_EVT: | ||
938 | dbg_link("TRF-"); | ||
939 | /* fall through */ | ||
940 | case ACTIVATE_MSG: | ||
941 | other = l_ptr->owner->active_links[0]; | ||
942 | if (other && link_working_unknown(other)) { | ||
943 | dbg_link("ACT\n"); | ||
944 | break; | ||
945 | } | ||
946 | dbg_link("ACT -> WW\n"); | ||
947 | l_ptr->state = WORKING_WORKING; | ||
948 | l_ptr->fsm_msg_cnt = 0; | ||
949 | link_activate(l_ptr); | ||
950 | link_send_proto_msg(l_ptr, STATE_MSG, 1, 0, 0, 0, 0); | ||
951 | l_ptr->fsm_msg_cnt++; | ||
952 | link_set_timer(l_ptr, cont_intv); | ||
953 | break; | ||
954 | case RESET_MSG: | ||
955 | dbg_link("RES\n"); | ||
956 | break; | ||
957 | case TIMEOUT_EVT: | ||
958 | dbg_link("TIM\n"); | ||
959 | link_send_proto_msg(l_ptr, ACTIVATE_MSG, 0, 0, 0, 0, 0); | ||
960 | l_ptr->fsm_msg_cnt++; | ||
961 | link_set_timer(l_ptr, cont_intv); | ||
962 | dbg_link("fsm_msg_cnt %u\n", l_ptr->fsm_msg_cnt); | ||
963 | break; | ||
964 | default: | ||
965 | err("Unknown link event %u in RR state\n", event); | ||
966 | } | ||
967 | break; | ||
968 | default: | ||
969 | err("Unknown link state %u/%u\n", l_ptr->state, event); | ||
970 | } | ||
971 | } | ||
972 | |||
973 | /* | ||
974 | * link_bundle_buf(): Append contents of a buffer to | ||
975 | * the tail of an existing one. | ||
976 | */ | ||
977 | |||
978 | static int link_bundle_buf(struct link *l_ptr, | ||
979 | struct sk_buff *bundler, | ||
980 | struct sk_buff *buf) | ||
981 | { | ||
982 | struct tipc_msg *bundler_msg = buf_msg(bundler); | ||
983 | struct tipc_msg *msg = buf_msg(buf); | ||
984 | u32 size = msg_size(msg); | ||
985 | u32 to_pos = align(msg_size(bundler_msg)); | ||
986 | u32 rest = link_max_pkt(l_ptr) - to_pos; | ||
987 | |||
988 | if (msg_user(bundler_msg) != MSG_BUNDLER) | ||
989 | return 0; | ||
990 | if (msg_type(bundler_msg) != OPEN_MSG) | ||
991 | return 0; | ||
992 | if (rest < align(size)) | ||
993 | return 0; | ||
994 | |||
995 | skb_put(bundler, (to_pos - msg_size(bundler_msg)) + size); | ||
996 | memcpy(bundler->data + to_pos, buf->data, size); | ||
997 | msg_set_size(bundler_msg, to_pos + size); | ||
998 | msg_set_msgcnt(bundler_msg, msg_msgcnt(bundler_msg) + 1); | ||
999 | dbg("Packed msg # %u(%u octets) into pos %u in buf(#%u)\n", | ||
1000 | msg_msgcnt(bundler_msg), size, to_pos, msg_seqno(bundler_msg)); | ||
1001 | msg_dbg(msg, "PACKD:"); | ||
1002 | buf_discard(buf); | ||
1003 | l_ptr->stats.sent_bundled++; | ||
1004 | return 1; | ||
1005 | } | ||
1006 | |||
1007 | static inline void link_add_to_outqueue(struct link *l_ptr, | ||
1008 | struct sk_buff *buf, | ||
1009 | struct tipc_msg *msg) | ||
1010 | { | ||
1011 | u32 ack = mod(l_ptr->next_in_no - 1); | ||
1012 | u32 seqno = mod(l_ptr->next_out_no++); | ||
1013 | |||
1014 | msg_set_word(msg, 2, ((ack << 16) | seqno)); | ||
1015 | msg_set_bcast_ack(msg, l_ptr->owner->bclink.last_in); | ||
1016 | buf->next = NULL; | ||
1017 | if (l_ptr->first_out) { | ||
1018 | l_ptr->last_out->next = buf; | ||
1019 | l_ptr->last_out = buf; | ||
1020 | } else | ||
1021 | l_ptr->first_out = l_ptr->last_out = buf; | ||
1022 | l_ptr->out_queue_size++; | ||
1023 | } | ||
1024 | |||
1025 | /* | ||
1026 | * link_send_buf() is the 'full path' for messages, called from | ||
1027 | * inside TIPC when the 'fast path' in tipc_send_buf | ||
1028 | * has failed, and from link_send() | ||
1029 | */ | ||
1030 | |||
1031 | int link_send_buf(struct link *l_ptr, struct sk_buff *buf) | ||
1032 | { | ||
1033 | struct tipc_msg *msg = buf_msg(buf); | ||
1034 | u32 size = msg_size(msg); | ||
1035 | u32 dsz = msg_data_sz(msg); | ||
1036 | u32 queue_size = l_ptr->out_queue_size; | ||
1037 | u32 imp = msg_tot_importance(msg); | ||
1038 | u32 queue_limit = l_ptr->queue_limit[imp]; | ||
1039 | u32 max_packet = link_max_pkt(l_ptr); | ||
1040 | |||
1041 | msg_set_prevnode(msg, tipc_own_addr); /* If routed message */ | ||
1042 | |||
1043 | /* Match msg importance against queue limits: */ | ||
1044 | |||
1045 | if (unlikely(queue_size >= queue_limit)) { | ||
1046 | if (imp <= TIPC_CRITICAL_IMPORTANCE) { | ||
1047 | return link_schedule_port(l_ptr, msg_origport(msg), | ||
1048 | size); | ||
1049 | } | ||
1050 | msg_dbg(msg, "TIPC: Congestion, throwing away\n"); | ||
1051 | buf_discard(buf); | ||
1052 | if (imp > CONN_MANAGER) { | ||
1053 | warn("Resetting <%s>, send queue full", l_ptr->name); | ||
1054 | link_reset(l_ptr); | ||
1055 | } | ||
1056 | return dsz; | ||
1057 | } | ||
1058 | |||
1059 | /* Fragmentation needed ? */ | ||
1060 | |||
1061 | if (size > max_packet) | ||
1062 | return link_send_long_buf(l_ptr, buf); | ||
1063 | |||
1064 | /* Packet can be queued or sent: */ | ||
1065 | |||
1066 | if (queue_size > l_ptr->stats.max_queue_sz) | ||
1067 | l_ptr->stats.max_queue_sz = queue_size; | ||
1068 | |||
1069 | if (likely(!bearer_congested(l_ptr->b_ptr, l_ptr) && | ||
1070 | !link_congested(l_ptr))) { | ||
1071 | link_add_to_outqueue(l_ptr, buf, msg); | ||
1072 | |||
1073 | if (likely(bearer_send(l_ptr->b_ptr, buf, &l_ptr->media_addr))) { | ||
1074 | l_ptr->unacked_window = 0; | ||
1075 | } else { | ||
1076 | bearer_schedule(l_ptr->b_ptr, l_ptr); | ||
1077 | l_ptr->stats.bearer_congs++; | ||
1078 | l_ptr->next_out = buf; | ||
1079 | } | ||
1080 | return dsz; | ||
1081 | } | ||
1082 | /* Congestion: can message be bundled ?: */ | ||
1083 | |||
1084 | if ((msg_user(msg) != CHANGEOVER_PROTOCOL) && | ||
1085 | (msg_user(msg) != MSG_FRAGMENTER)) { | ||
1086 | |||
1087 | /* Try adding message to an existing bundle */ | ||
1088 | |||
1089 | if (l_ptr->next_out && | ||
1090 | link_bundle_buf(l_ptr, l_ptr->last_out, buf)) { | ||
1091 | bearer_resolve_congestion(l_ptr->b_ptr, l_ptr); | ||
1092 | return dsz; | ||
1093 | } | ||
1094 | |||
1095 | /* Try creating a new bundle */ | ||
1096 | |||
1097 | if (size <= max_packet * 2 / 3) { | ||
1098 | struct sk_buff *bundler = buf_acquire(max_packet); | ||
1099 | struct tipc_msg bundler_hdr; | ||
1100 | |||
1101 | if (bundler) { | ||
1102 | msg_init(&bundler_hdr, MSG_BUNDLER, OPEN_MSG, | ||
1103 | TIPC_OK, INT_H_SIZE, l_ptr->addr); | ||
1104 | memcpy(bundler->data, (unchar *)&bundler_hdr, | ||
1105 | INT_H_SIZE); | ||
1106 | skb_trim(bundler, INT_H_SIZE); | ||
1107 | link_bundle_buf(l_ptr, bundler, buf); | ||
1108 | buf = bundler; | ||
1109 | msg = buf_msg(buf); | ||
1110 | l_ptr->stats.sent_bundles++; | ||
1111 | } | ||
1112 | } | ||
1113 | } | ||
1114 | if (!l_ptr->next_out) | ||
1115 | l_ptr->next_out = buf; | ||
1116 | link_add_to_outqueue(l_ptr, buf, msg); | ||
1117 | bearer_resolve_congestion(l_ptr->b_ptr, l_ptr); | ||
1118 | return dsz; | ||
1119 | } | ||
1120 | |||
1121 | /* | ||
1122 | * link_send(): same as link_send_buf(), but the link to use has | ||
1123 | * not been selected yet, and the the owner node is not locked | ||
1124 | * Called by TIPC internal users, e.g. the name distributor | ||
1125 | */ | ||
1126 | |||
1127 | int link_send(struct sk_buff *buf, u32 dest, u32 selector) | ||
1128 | { | ||
1129 | struct link *l_ptr; | ||
1130 | struct node *n_ptr; | ||
1131 | int res = -ELINKCONG; | ||
1132 | |||
1133 | read_lock_bh(&net_lock); | ||
1134 | n_ptr = node_select(dest, selector); | ||
1135 | if (n_ptr) { | ||
1136 | node_lock(n_ptr); | ||
1137 | l_ptr = n_ptr->active_links[selector & 1]; | ||
1138 | dbg("link_send: found link %x for dest %x\n", l_ptr, dest); | ||
1139 | if (l_ptr) { | ||
1140 | res = link_send_buf(l_ptr, buf); | ||
1141 | } | ||
1142 | node_unlock(n_ptr); | ||
1143 | } else { | ||
1144 | dbg("Attempt to send msg to unknown node:\n"); | ||
1145 | msg_dbg(buf_msg(buf),">>>"); | ||
1146 | buf_discard(buf); | ||
1147 | } | ||
1148 | read_unlock_bh(&net_lock); | ||
1149 | return res; | ||
1150 | } | ||
1151 | |||
1152 | /* | ||
1153 | * link_send_buf_fast: Entry for data messages where the | ||
1154 | * destination link is known and the header is complete, | ||
1155 | * inclusive total message length. Very time critical. | ||
1156 | * Link is locked. Returns user data length. | ||
1157 | */ | ||
1158 | |||
1159 | static inline int link_send_buf_fast(struct link *l_ptr, struct sk_buff *buf, | ||
1160 | u32 *used_max_pkt) | ||
1161 | { | ||
1162 | struct tipc_msg *msg = buf_msg(buf); | ||
1163 | int res = msg_data_sz(msg); | ||
1164 | |||
1165 | if (likely(!link_congested(l_ptr))) { | ||
1166 | if (likely(msg_size(msg) <= link_max_pkt(l_ptr))) { | ||
1167 | if (likely(list_empty(&l_ptr->b_ptr->cong_links))) { | ||
1168 | link_add_to_outqueue(l_ptr, buf, msg); | ||
1169 | if (likely(bearer_send(l_ptr->b_ptr, buf, | ||
1170 | &l_ptr->media_addr))) { | ||
1171 | l_ptr->unacked_window = 0; | ||
1172 | msg_dbg(msg,"SENT_FAST:"); | ||
1173 | return res; | ||
1174 | } | ||
1175 | dbg("failed sent fast...\n"); | ||
1176 | bearer_schedule(l_ptr->b_ptr, l_ptr); | ||
1177 | l_ptr->stats.bearer_congs++; | ||
1178 | l_ptr->next_out = buf; | ||
1179 | return res; | ||
1180 | } | ||
1181 | } | ||
1182 | else | ||
1183 | *used_max_pkt = link_max_pkt(l_ptr); | ||
1184 | } | ||
1185 | return link_send_buf(l_ptr, buf); /* All other cases */ | ||
1186 | } | ||
1187 | |||
1188 | /* | ||
1189 | * tipc_send_buf_fast: Entry for data messages where the | ||
1190 | * destination node is known and the header is complete, | ||
1191 | * inclusive total message length. | ||
1192 | * Returns user data length. | ||
1193 | */ | ||
1194 | int tipc_send_buf_fast(struct sk_buff *buf, u32 destnode) | ||
1195 | { | ||
1196 | struct link *l_ptr; | ||
1197 | struct node *n_ptr; | ||
1198 | int res; | ||
1199 | u32 selector = msg_origport(buf_msg(buf)) & 1; | ||
1200 | u32 dummy; | ||
1201 | |||
1202 | if (destnode == tipc_own_addr) | ||
1203 | return port_recv_msg(buf); | ||
1204 | |||
1205 | read_lock_bh(&net_lock); | ||
1206 | n_ptr = node_select(destnode, selector); | ||
1207 | if (likely(n_ptr)) { | ||
1208 | node_lock(n_ptr); | ||
1209 | l_ptr = n_ptr->active_links[selector]; | ||
1210 | dbg("send_fast: buf %x selected %x, destnode = %x\n", | ||
1211 | buf, l_ptr, destnode); | ||
1212 | if (likely(l_ptr)) { | ||
1213 | res = link_send_buf_fast(l_ptr, buf, &dummy); | ||
1214 | node_unlock(n_ptr); | ||
1215 | read_unlock_bh(&net_lock); | ||
1216 | return res; | ||
1217 | } | ||
1218 | node_unlock(n_ptr); | ||
1219 | } | ||
1220 | read_unlock_bh(&net_lock); | ||
1221 | res = msg_data_sz(buf_msg(buf)); | ||
1222 | tipc_reject_msg(buf, TIPC_ERR_NO_NODE); | ||
1223 | return res; | ||
1224 | } | ||
1225 | |||
1226 | |||
1227 | /* | ||
1228 | * link_send_sections_fast: Entry for messages where the | ||
1229 | * destination processor is known and the header is complete, | ||
1230 | * except for total message length. | ||
1231 | * Returns user data length or errno. | ||
1232 | */ | ||
1233 | int link_send_sections_fast(struct port *sender, | ||
1234 | struct iovec const *msg_sect, | ||
1235 | const u32 num_sect, | ||
1236 | u32 destaddr) | ||
1237 | { | ||
1238 | struct tipc_msg *hdr = &sender->publ.phdr; | ||
1239 | struct link *l_ptr; | ||
1240 | struct sk_buff *buf; | ||
1241 | struct node *node; | ||
1242 | int res; | ||
1243 | u32 selector = msg_origport(hdr) & 1; | ||
1244 | |||
1245 | assert(destaddr != tipc_own_addr); | ||
1246 | |||
1247 | again: | ||
1248 | /* | ||
1249 | * Try building message using port's max_pkt hint. | ||
1250 | * (Must not hold any locks while building message.) | ||
1251 | */ | ||
1252 | |||
1253 | res = msg_build(hdr, msg_sect, num_sect, sender->max_pkt, | ||
1254 | !sender->user_port, &buf); | ||
1255 | |||
1256 | read_lock_bh(&net_lock); | ||
1257 | node = node_select(destaddr, selector); | ||
1258 | if (likely(node)) { | ||
1259 | node_lock(node); | ||
1260 | l_ptr = node->active_links[selector]; | ||
1261 | if (likely(l_ptr)) { | ||
1262 | if (likely(buf)) { | ||
1263 | res = link_send_buf_fast(l_ptr, buf, | ||
1264 | &sender->max_pkt); | ||
1265 | if (unlikely(res < 0)) | ||
1266 | buf_discard(buf); | ||
1267 | exit: | ||
1268 | node_unlock(node); | ||
1269 | read_unlock_bh(&net_lock); | ||
1270 | return res; | ||
1271 | } | ||
1272 | |||
1273 | /* Exit if build request was invalid */ | ||
1274 | |||
1275 | if (unlikely(res < 0)) | ||
1276 | goto exit; | ||
1277 | |||
1278 | /* Exit if link (or bearer) is congested */ | ||
1279 | |||
1280 | if (link_congested(l_ptr) || | ||
1281 | !list_empty(&l_ptr->b_ptr->cong_links)) { | ||
1282 | res = link_schedule_port(l_ptr, | ||
1283 | sender->publ.ref, res); | ||
1284 | goto exit; | ||
1285 | } | ||
1286 | |||
1287 | /* | ||
1288 | * Message size exceeds max_pkt hint; update hint, | ||
1289 | * then re-try fast path or fragment the message | ||
1290 | */ | ||
1291 | |||
1292 | sender->max_pkt = link_max_pkt(l_ptr); | ||
1293 | node_unlock(node); | ||
1294 | read_unlock_bh(&net_lock); | ||
1295 | |||
1296 | |||
1297 | if ((msg_hdr_sz(hdr) + res) <= sender->max_pkt) | ||
1298 | goto again; | ||
1299 | |||
1300 | return link_send_sections_long(sender, msg_sect, | ||
1301 | num_sect, destaddr); | ||
1302 | } | ||
1303 | node_unlock(node); | ||
1304 | } | ||
1305 | read_unlock_bh(&net_lock); | ||
1306 | |||
1307 | /* Couldn't find a link to the destination node */ | ||
1308 | |||
1309 | if (buf) | ||
1310 | return tipc_reject_msg(buf, TIPC_ERR_NO_NODE); | ||
1311 | if (res >= 0) | ||
1312 | return port_reject_sections(sender, hdr, msg_sect, num_sect, | ||
1313 | TIPC_ERR_NO_NODE); | ||
1314 | return res; | ||
1315 | } | ||
1316 | |||
1317 | /* | ||
1318 | * link_send_sections_long(): Entry for long messages where the | ||
1319 | * destination node is known and the header is complete, | ||
1320 | * inclusive total message length. | ||
1321 | * Link and bearer congestion status have been checked to be ok, | ||
1322 | * and are ignored if they change. | ||
1323 | * | ||
1324 | * Note that fragments do not use the full link MTU so that they won't have | ||
1325 | * to undergo refragmentation if link changeover causes them to be sent | ||
1326 | * over another link with an additional tunnel header added as prefix. | ||
1327 | * (Refragmentation will still occur if the other link has a smaller MTU.) | ||
1328 | * | ||
1329 | * Returns user data length or errno. | ||
1330 | */ | ||
1331 | static int link_send_sections_long(struct port *sender, | ||
1332 | struct iovec const *msg_sect, | ||
1333 | u32 num_sect, | ||
1334 | u32 destaddr) | ||
1335 | { | ||
1336 | struct link *l_ptr; | ||
1337 | struct node *node; | ||
1338 | struct tipc_msg *hdr = &sender->publ.phdr; | ||
1339 | u32 dsz = msg_data_sz(hdr); | ||
1340 | u32 max_pkt,fragm_sz,rest; | ||
1341 | struct tipc_msg fragm_hdr; | ||
1342 | struct sk_buff *buf,*buf_chain,*prev; | ||
1343 | u32 fragm_crs,fragm_rest,hsz,sect_rest; | ||
1344 | const unchar *sect_crs; | ||
1345 | int curr_sect; | ||
1346 | u32 fragm_no; | ||
1347 | |||
1348 | again: | ||
1349 | fragm_no = 1; | ||
1350 | max_pkt = sender->max_pkt - INT_H_SIZE; | ||
1351 | /* leave room for tunnel header in case of link changeover */ | ||
1352 | fragm_sz = max_pkt - INT_H_SIZE; | ||
1353 | /* leave room for fragmentation header in each fragment */ | ||
1354 | rest = dsz; | ||
1355 | fragm_crs = 0; | ||
1356 | fragm_rest = 0; | ||
1357 | sect_rest = 0; | ||
1358 | sect_crs = 0; | ||
1359 | curr_sect = -1; | ||
1360 | |||
1361 | /* Prepare reusable fragment header: */ | ||
1362 | |||
1363 | msg_dbg(hdr, ">FRAGMENTING>"); | ||
1364 | msg_init(&fragm_hdr, MSG_FRAGMENTER, FIRST_FRAGMENT, | ||
1365 | TIPC_OK, INT_H_SIZE, msg_destnode(hdr)); | ||
1366 | msg_set_link_selector(&fragm_hdr, sender->publ.ref); | ||
1367 | msg_set_size(&fragm_hdr, max_pkt); | ||
1368 | msg_set_fragm_no(&fragm_hdr, 1); | ||
1369 | |||
1370 | /* Prepare header of first fragment: */ | ||
1371 | |||
1372 | buf_chain = buf = buf_acquire(max_pkt); | ||
1373 | if (!buf) | ||
1374 | return -ENOMEM; | ||
1375 | buf->next = NULL; | ||
1376 | memcpy(buf->data, (unchar *)&fragm_hdr, INT_H_SIZE); | ||
1377 | hsz = msg_hdr_sz(hdr); | ||
1378 | memcpy(buf->data + INT_H_SIZE, (unchar *)hdr, hsz); | ||
1379 | msg_dbg(buf_msg(buf), ">BUILD>"); | ||
1380 | |||
1381 | /* Chop up message: */ | ||
1382 | |||
1383 | fragm_crs = INT_H_SIZE + hsz; | ||
1384 | fragm_rest = fragm_sz - hsz; | ||
1385 | |||
1386 | do { /* For all sections */ | ||
1387 | u32 sz; | ||
1388 | |||
1389 | if (!sect_rest) { | ||
1390 | sect_rest = msg_sect[++curr_sect].iov_len; | ||
1391 | sect_crs = (const unchar *)msg_sect[curr_sect].iov_base; | ||
1392 | } | ||
1393 | |||
1394 | if (sect_rest < fragm_rest) | ||
1395 | sz = sect_rest; | ||
1396 | else | ||
1397 | sz = fragm_rest; | ||
1398 | |||
1399 | if (likely(!sender->user_port)) { | ||
1400 | if (copy_from_user(buf->data + fragm_crs, sect_crs, sz)) { | ||
1401 | error: | ||
1402 | for (; buf_chain; buf_chain = buf) { | ||
1403 | buf = buf_chain->next; | ||
1404 | buf_discard(buf_chain); | ||
1405 | } | ||
1406 | return -EFAULT; | ||
1407 | } | ||
1408 | } else | ||
1409 | memcpy(buf->data + fragm_crs, sect_crs, sz); | ||
1410 | |||
1411 | sect_crs += sz; | ||
1412 | sect_rest -= sz; | ||
1413 | fragm_crs += sz; | ||
1414 | fragm_rest -= sz; | ||
1415 | rest -= sz; | ||
1416 | |||
1417 | if (!fragm_rest && rest) { | ||
1418 | |||
1419 | /* Initiate new fragment: */ | ||
1420 | if (rest <= fragm_sz) { | ||
1421 | fragm_sz = rest; | ||
1422 | msg_set_type(&fragm_hdr,LAST_FRAGMENT); | ||
1423 | } else { | ||
1424 | msg_set_type(&fragm_hdr, FRAGMENT); | ||
1425 | } | ||
1426 | msg_set_size(&fragm_hdr, fragm_sz + INT_H_SIZE); | ||
1427 | msg_set_fragm_no(&fragm_hdr, ++fragm_no); | ||
1428 | prev = buf; | ||
1429 | buf = buf_acquire(fragm_sz + INT_H_SIZE); | ||
1430 | if (!buf) | ||
1431 | goto error; | ||
1432 | |||
1433 | buf->next = NULL; | ||
1434 | prev->next = buf; | ||
1435 | memcpy(buf->data, (unchar *)&fragm_hdr, INT_H_SIZE); | ||
1436 | fragm_crs = INT_H_SIZE; | ||
1437 | fragm_rest = fragm_sz; | ||
1438 | msg_dbg(buf_msg(buf)," >BUILD>"); | ||
1439 | } | ||
1440 | } | ||
1441 | while (rest > 0); | ||
1442 | |||
1443 | /* | ||
1444 | * Now we have a buffer chain. Select a link and check | ||
1445 | * that packet size is still OK | ||
1446 | */ | ||
1447 | node = node_select(destaddr, sender->publ.ref & 1); | ||
1448 | if (likely(node)) { | ||
1449 | node_lock(node); | ||
1450 | l_ptr = node->active_links[sender->publ.ref & 1]; | ||
1451 | if (!l_ptr) { | ||
1452 | node_unlock(node); | ||
1453 | goto reject; | ||
1454 | } | ||
1455 | if (link_max_pkt(l_ptr) < max_pkt) { | ||
1456 | sender->max_pkt = link_max_pkt(l_ptr); | ||
1457 | node_unlock(node); | ||
1458 | for (; buf_chain; buf_chain = buf) { | ||
1459 | buf = buf_chain->next; | ||
1460 | buf_discard(buf_chain); | ||
1461 | } | ||
1462 | goto again; | ||
1463 | } | ||
1464 | } else { | ||
1465 | reject: | ||
1466 | for (; buf_chain; buf_chain = buf) { | ||
1467 | buf = buf_chain->next; | ||
1468 | buf_discard(buf_chain); | ||
1469 | } | ||
1470 | return port_reject_sections(sender, hdr, msg_sect, num_sect, | ||
1471 | TIPC_ERR_NO_NODE); | ||
1472 | } | ||
1473 | |||
1474 | /* Append whole chain to send queue: */ | ||
1475 | |||
1476 | buf = buf_chain; | ||
1477 | l_ptr->long_msg_seq_no = mod(l_ptr->long_msg_seq_no + 1); | ||
1478 | if (!l_ptr->next_out) | ||
1479 | l_ptr->next_out = buf_chain; | ||
1480 | l_ptr->stats.sent_fragmented++; | ||
1481 | while (buf) { | ||
1482 | struct sk_buff *next = buf->next; | ||
1483 | struct tipc_msg *msg = buf_msg(buf); | ||
1484 | |||
1485 | l_ptr->stats.sent_fragments++; | ||
1486 | msg_set_long_msgno(msg, l_ptr->long_msg_seq_no); | ||
1487 | link_add_to_outqueue(l_ptr, buf, msg); | ||
1488 | msg_dbg(msg, ">ADD>"); | ||
1489 | buf = next; | ||
1490 | } | ||
1491 | |||
1492 | /* Send it, if possible: */ | ||
1493 | |||
1494 | link_push_queue(l_ptr); | ||
1495 | node_unlock(node); | ||
1496 | return dsz; | ||
1497 | } | ||
1498 | |||
1499 | /* | ||
1500 | * link_push_packet: Push one unsent packet to the media | ||
1501 | */ | ||
1502 | u32 link_push_packet(struct link *l_ptr) | ||
1503 | { | ||
1504 | struct sk_buff *buf = l_ptr->first_out; | ||
1505 | u32 r_q_size = l_ptr->retransm_queue_size; | ||
1506 | u32 r_q_head = l_ptr->retransm_queue_head; | ||
1507 | |||
1508 | /* Step to position where retransmission failed, if any, */ | ||
1509 | /* consider that buffers may have been released in meantime */ | ||
1510 | |||
1511 | if (r_q_size && buf) { | ||
1512 | u32 last = lesser(mod(r_q_head + r_q_size), | ||
1513 | link_last_sent(l_ptr)); | ||
1514 | u32 first = msg_seqno(buf_msg(buf)); | ||
1515 | |||
1516 | while (buf && less(first, r_q_head)) { | ||
1517 | first = mod(first + 1); | ||
1518 | buf = buf->next; | ||
1519 | } | ||
1520 | l_ptr->retransm_queue_head = r_q_head = first; | ||
1521 | l_ptr->retransm_queue_size = r_q_size = mod(last - first); | ||
1522 | } | ||
1523 | |||
1524 | /* Continue retransmission now, if there is anything: */ | ||
1525 | |||
1526 | if (r_q_size && buf && !skb_cloned(buf)) { | ||
1527 | msg_set_ack(buf_msg(buf), mod(l_ptr->next_in_no - 1)); | ||
1528 | msg_set_bcast_ack(buf_msg(buf), l_ptr->owner->bclink.last_in); | ||
1529 | if (bearer_send(l_ptr->b_ptr, buf, &l_ptr->media_addr)) { | ||
1530 | msg_dbg(buf_msg(buf), ">DEF-RETR>"); | ||
1531 | l_ptr->retransm_queue_head = mod(++r_q_head); | ||
1532 | l_ptr->retransm_queue_size = --r_q_size; | ||
1533 | l_ptr->stats.retransmitted++; | ||
1534 | return TIPC_OK; | ||
1535 | } else { | ||
1536 | l_ptr->stats.bearer_congs++; | ||
1537 | msg_dbg(buf_msg(buf), "|>DEF-RETR>"); | ||
1538 | return PUSH_FAILED; | ||
1539 | } | ||
1540 | } | ||
1541 | |||
1542 | /* Send deferred protocol message, if any: */ | ||
1543 | |||
1544 | buf = l_ptr->proto_msg_queue; | ||
1545 | if (buf) { | ||
1546 | msg_set_ack(buf_msg(buf), mod(l_ptr->next_in_no - 1)); | ||
1547 | msg_set_bcast_ack(buf_msg(buf),l_ptr->owner->bclink.last_in); | ||
1548 | if (bearer_send(l_ptr->b_ptr, buf, &l_ptr->media_addr)) { | ||
1549 | msg_dbg(buf_msg(buf), ">DEF-PROT>"); | ||
1550 | l_ptr->unacked_window = 0; | ||
1551 | buf_discard(buf); | ||
1552 | l_ptr->proto_msg_queue = 0; | ||
1553 | return TIPC_OK; | ||
1554 | } else { | ||
1555 | msg_dbg(buf_msg(buf), "|>DEF-PROT>"); | ||
1556 | l_ptr->stats.bearer_congs++; | ||
1557 | return PUSH_FAILED; | ||
1558 | } | ||
1559 | } | ||
1560 | |||
1561 | /* Send one deferred data message, if send window not full: */ | ||
1562 | |||
1563 | buf = l_ptr->next_out; | ||
1564 | if (buf) { | ||
1565 | struct tipc_msg *msg = buf_msg(buf); | ||
1566 | u32 next = msg_seqno(msg); | ||
1567 | u32 first = msg_seqno(buf_msg(l_ptr->first_out)); | ||
1568 | |||
1569 | if (mod(next - first) < l_ptr->queue_limit[0]) { | ||
1570 | msg_set_ack(msg, mod(l_ptr->next_in_no - 1)); | ||
1571 | msg_set_bcast_ack(msg, l_ptr->owner->bclink.last_in); | ||
1572 | if (bearer_send(l_ptr->b_ptr, buf, &l_ptr->media_addr)) { | ||
1573 | if (msg_user(msg) == MSG_BUNDLER) | ||
1574 | msg_set_type(msg, CLOSED_MSG); | ||
1575 | msg_dbg(msg, ">PUSH-DATA>"); | ||
1576 | l_ptr->next_out = buf->next; | ||
1577 | return TIPC_OK; | ||
1578 | } else { | ||
1579 | msg_dbg(msg, "|PUSH-DATA|"); | ||
1580 | l_ptr->stats.bearer_congs++; | ||
1581 | return PUSH_FAILED; | ||
1582 | } | ||
1583 | } | ||
1584 | } | ||
1585 | return PUSH_FINISHED; | ||
1586 | } | ||
1587 | |||
1588 | /* | ||
1589 | * push_queue(): push out the unsent messages of a link where | ||
1590 | * congestion has abated. Node is locked | ||
1591 | */ | ||
1592 | void link_push_queue(struct link *l_ptr) | ||
1593 | { | ||
1594 | u32 res; | ||
1595 | |||
1596 | if (bearer_congested(l_ptr->b_ptr, l_ptr)) | ||
1597 | return; | ||
1598 | |||
1599 | do { | ||
1600 | res = link_push_packet(l_ptr); | ||
1601 | } | ||
1602 | while (res == TIPC_OK); | ||
1603 | if (res == PUSH_FAILED) | ||
1604 | bearer_schedule(l_ptr->b_ptr, l_ptr); | ||
1605 | } | ||
1606 | |||
1607 | void link_retransmit(struct link *l_ptr, struct sk_buff *buf, | ||
1608 | u32 retransmits) | ||
1609 | { | ||
1610 | struct tipc_msg *msg; | ||
1611 | |||
1612 | dbg("Retransmitting %u in link %x\n", retransmits, l_ptr); | ||
1613 | |||
1614 | if (bearer_congested(l_ptr->b_ptr, l_ptr) && buf && !skb_cloned(buf)) { | ||
1615 | msg_dbg(buf_msg(buf), ">NO_RETR->BCONG>"); | ||
1616 | dbg_print_link(l_ptr, " "); | ||
1617 | l_ptr->retransm_queue_head = msg_seqno(buf_msg(buf)); | ||
1618 | l_ptr->retransm_queue_size = retransmits; | ||
1619 | return; | ||
1620 | } | ||
1621 | while (retransmits && (buf != l_ptr->next_out) && buf && !skb_cloned(buf)) { | ||
1622 | msg = buf_msg(buf); | ||
1623 | msg_set_ack(msg, mod(l_ptr->next_in_no - 1)); | ||
1624 | msg_set_bcast_ack(msg, l_ptr->owner->bclink.last_in); | ||
1625 | if (bearer_send(l_ptr->b_ptr, buf, &l_ptr->media_addr)) { | ||
1626 | /* Catch if retransmissions fail repeatedly: */ | ||
1627 | if (l_ptr->last_retransmitted == msg_seqno(msg)) { | ||
1628 | if (++l_ptr->stale_count > 100) { | ||
1629 | msg_print(CONS, buf_msg(buf), ">RETR>"); | ||
1630 | info("...Retransmitted %u times\n", | ||
1631 | l_ptr->stale_count); | ||
1632 | link_print(l_ptr, CONS, "Resetting Link\n");; | ||
1633 | link_reset(l_ptr); | ||
1634 | break; | ||
1635 | } | ||
1636 | } else { | ||
1637 | l_ptr->stale_count = 0; | ||
1638 | } | ||
1639 | l_ptr->last_retransmitted = msg_seqno(msg); | ||
1640 | |||
1641 | msg_dbg(buf_msg(buf), ">RETR>"); | ||
1642 | buf = buf->next; | ||
1643 | retransmits--; | ||
1644 | l_ptr->stats.retransmitted++; | ||
1645 | } else { | ||
1646 | bearer_schedule(l_ptr->b_ptr, l_ptr); | ||
1647 | l_ptr->stats.bearer_congs++; | ||
1648 | l_ptr->retransm_queue_head = msg_seqno(buf_msg(buf)); | ||
1649 | l_ptr->retransm_queue_size = retransmits; | ||
1650 | return; | ||
1651 | } | ||
1652 | } | ||
1653 | l_ptr->retransm_queue_head = l_ptr->retransm_queue_size = 0; | ||
1654 | } | ||
1655 | |||
1656 | /* | ||
1657 | * link_recv_non_seq: Receive packets which are outside | ||
1658 | * the link sequence flow | ||
1659 | */ | ||
1660 | |||
1661 | static void link_recv_non_seq(struct sk_buff *buf) | ||
1662 | { | ||
1663 | struct tipc_msg *msg = buf_msg(buf); | ||
1664 | |||
1665 | if (msg_user(msg) == LINK_CONFIG) | ||
1666 | disc_recv_msg(buf); | ||
1667 | else | ||
1668 | bclink_recv_pkt(buf); | ||
1669 | } | ||
1670 | |||
1671 | /** | ||
1672 | * link_insert_deferred_queue - insert deferred messages back into receive chain | ||
1673 | */ | ||
1674 | |||
1675 | static struct sk_buff *link_insert_deferred_queue(struct link *l_ptr, | ||
1676 | struct sk_buff *buf) | ||
1677 | { | ||
1678 | u32 seq_no; | ||
1679 | |||
1680 | if (l_ptr->oldest_deferred_in == NULL) | ||
1681 | return buf; | ||
1682 | |||
1683 | seq_no = msg_seqno(buf_msg(l_ptr->oldest_deferred_in)); | ||
1684 | if (seq_no == mod(l_ptr->next_in_no)) { | ||
1685 | l_ptr->newest_deferred_in->next = buf; | ||
1686 | buf = l_ptr->oldest_deferred_in; | ||
1687 | l_ptr->oldest_deferred_in = NULL; | ||
1688 | l_ptr->deferred_inqueue_sz = 0; | ||
1689 | } | ||
1690 | return buf; | ||
1691 | } | ||
1692 | |||
1693 | void tipc_recv_msg(struct sk_buff *head, struct tipc_bearer *tb_ptr) | ||
1694 | { | ||
1695 | read_lock_bh(&net_lock); | ||
1696 | while (head) { | ||
1697 | struct bearer *b_ptr; | ||
1698 | struct node *n_ptr; | ||
1699 | struct link *l_ptr; | ||
1700 | struct sk_buff *crs; | ||
1701 | struct sk_buff *buf = head; | ||
1702 | struct tipc_msg *msg = buf_msg(buf); | ||
1703 | u32 seq_no = msg_seqno(msg); | ||
1704 | u32 ackd = msg_ack(msg); | ||
1705 | u32 released = 0; | ||
1706 | int type; | ||
1707 | |||
1708 | b_ptr = (struct bearer *)tb_ptr; | ||
1709 | TIPC_SKB_CB(buf)->handle = b_ptr; | ||
1710 | |||
1711 | head = head->next; | ||
1712 | if (unlikely(msg_version(msg) != TIPC_VERSION)) | ||
1713 | goto cont; | ||
1714 | #if 0 | ||
1715 | if (msg_user(msg) != LINK_PROTOCOL) | ||
1716 | #endif | ||
1717 | msg_dbg(msg,"<REC<"); | ||
1718 | |||
1719 | if (unlikely(msg_non_seq(msg))) { | ||
1720 | link_recv_non_seq(buf); | ||
1721 | continue; | ||
1722 | } | ||
1723 | n_ptr = node_find(msg_prevnode(msg)); | ||
1724 | if (unlikely(!n_ptr)) | ||
1725 | goto cont; | ||
1726 | |||
1727 | node_lock(n_ptr); | ||
1728 | l_ptr = n_ptr->links[b_ptr->identity]; | ||
1729 | if (unlikely(!l_ptr)) { | ||
1730 | node_unlock(n_ptr); | ||
1731 | goto cont; | ||
1732 | } | ||
1733 | /* | ||
1734 | * Release acked messages | ||
1735 | */ | ||
1736 | if (less(n_ptr->bclink.acked, msg_bcast_ack(msg))) { | ||
1737 | if (node_is_up(n_ptr) && n_ptr->bclink.supported) | ||
1738 | bclink_acknowledge(n_ptr, msg_bcast_ack(msg)); | ||
1739 | } | ||
1740 | |||
1741 | crs = l_ptr->first_out; | ||
1742 | while ((crs != l_ptr->next_out) && | ||
1743 | less_eq(msg_seqno(buf_msg(crs)), ackd)) { | ||
1744 | struct sk_buff *next = crs->next; | ||
1745 | |||
1746 | buf_discard(crs); | ||
1747 | crs = next; | ||
1748 | released++; | ||
1749 | } | ||
1750 | if (released) { | ||
1751 | l_ptr->first_out = crs; | ||
1752 | l_ptr->out_queue_size -= released; | ||
1753 | } | ||
1754 | if (unlikely(l_ptr->next_out)) | ||
1755 | link_push_queue(l_ptr); | ||
1756 | if (unlikely(!list_empty(&l_ptr->waiting_ports))) | ||
1757 | link_wakeup_ports(l_ptr, 0); | ||
1758 | if (unlikely(++l_ptr->unacked_window >= TIPC_MIN_LINK_WIN)) { | ||
1759 | l_ptr->stats.sent_acks++; | ||
1760 | link_send_proto_msg(l_ptr, STATE_MSG, 0, 0, 0, 0, 0); | ||
1761 | } | ||
1762 | |||
1763 | protocol_check: | ||
1764 | if (likely(link_working_working(l_ptr))) { | ||
1765 | if (likely(seq_no == mod(l_ptr->next_in_no))) { | ||
1766 | l_ptr->next_in_no++; | ||
1767 | if (unlikely(l_ptr->oldest_deferred_in)) | ||
1768 | head = link_insert_deferred_queue(l_ptr, | ||
1769 | head); | ||
1770 | if (likely(msg_is_dest(msg, tipc_own_addr))) { | ||
1771 | deliver: | ||
1772 | if (likely(msg_isdata(msg))) { | ||
1773 | node_unlock(n_ptr); | ||
1774 | port_recv_msg(buf); | ||
1775 | continue; | ||
1776 | } | ||
1777 | switch (msg_user(msg)) { | ||
1778 | case MSG_BUNDLER: | ||
1779 | l_ptr->stats.recv_bundles++; | ||
1780 | l_ptr->stats.recv_bundled += | ||
1781 | msg_msgcnt(msg); | ||
1782 | node_unlock(n_ptr); | ||
1783 | link_recv_bundle(buf); | ||
1784 | continue; | ||
1785 | case ROUTE_DISTRIBUTOR: | ||
1786 | node_unlock(n_ptr); | ||
1787 | cluster_recv_routing_table(buf); | ||
1788 | continue; | ||
1789 | case NAME_DISTRIBUTOR: | ||
1790 | node_unlock(n_ptr); | ||
1791 | named_recv(buf); | ||
1792 | continue; | ||
1793 | case CONN_MANAGER: | ||
1794 | node_unlock(n_ptr); | ||
1795 | port_recv_proto_msg(buf); | ||
1796 | continue; | ||
1797 | case MSG_FRAGMENTER: | ||
1798 | l_ptr->stats.recv_fragments++; | ||
1799 | if (link_recv_fragment( | ||
1800 | &l_ptr->defragm_buf, | ||
1801 | &buf, &msg)) { | ||
1802 | l_ptr->stats.recv_fragmented++; | ||
1803 | goto deliver; | ||
1804 | } | ||
1805 | break; | ||
1806 | case CHANGEOVER_PROTOCOL: | ||
1807 | type = msg_type(msg); | ||
1808 | if (link_recv_changeover_msg( | ||
1809 | &l_ptr, &buf)) { | ||
1810 | msg = buf_msg(buf); | ||
1811 | seq_no = msg_seqno(msg); | ||
1812 | TIPC_SKB_CB(buf)->handle | ||
1813 | = b_ptr; | ||
1814 | if (type == ORIGINAL_MSG) | ||
1815 | goto deliver; | ||
1816 | goto protocol_check; | ||
1817 | } | ||
1818 | break; | ||
1819 | } | ||
1820 | } | ||
1821 | node_unlock(n_ptr); | ||
1822 | net_route_msg(buf); | ||
1823 | continue; | ||
1824 | } | ||
1825 | link_handle_out_of_seq_msg(l_ptr, buf); | ||
1826 | head = link_insert_deferred_queue(l_ptr, head); | ||
1827 | node_unlock(n_ptr); | ||
1828 | continue; | ||
1829 | } | ||
1830 | |||
1831 | if (msg_user(msg) == LINK_PROTOCOL) { | ||
1832 | link_recv_proto_msg(l_ptr, buf); | ||
1833 | head = link_insert_deferred_queue(l_ptr, head); | ||
1834 | node_unlock(n_ptr); | ||
1835 | continue; | ||
1836 | } | ||
1837 | msg_dbg(msg,"NSEQ<REC<"); | ||
1838 | link_state_event(l_ptr, TRAFFIC_MSG_EVT); | ||
1839 | |||
1840 | if (link_working_working(l_ptr)) { | ||
1841 | /* Re-insert in front of queue */ | ||
1842 | msg_dbg(msg,"RECV-REINS:"); | ||
1843 | buf->next = head; | ||
1844 | head = buf; | ||
1845 | node_unlock(n_ptr); | ||
1846 | continue; | ||
1847 | } | ||
1848 | node_unlock(n_ptr); | ||
1849 | cont: | ||
1850 | buf_discard(buf); | ||
1851 | } | ||
1852 | read_unlock_bh(&net_lock); | ||
1853 | } | ||
1854 | |||
1855 | /* | ||
1856 | * link_defer_buf(): Sort a received out-of-sequence packet | ||
1857 | * into the deferred reception queue. | ||
1858 | * Returns the increase of the queue length,i.e. 0 or 1 | ||
1859 | */ | ||
1860 | |||
1861 | u32 link_defer_pkt(struct sk_buff **head, | ||
1862 | struct sk_buff **tail, | ||
1863 | struct sk_buff *buf) | ||
1864 | { | ||
1865 | struct sk_buff *prev = 0; | ||
1866 | struct sk_buff *crs = *head; | ||
1867 | u32 seq_no = msg_seqno(buf_msg(buf)); | ||
1868 | |||
1869 | buf->next = NULL; | ||
1870 | |||
1871 | /* Empty queue ? */ | ||
1872 | if (*head == NULL) { | ||
1873 | *head = *tail = buf; | ||
1874 | return 1; | ||
1875 | } | ||
1876 | |||
1877 | /* Last ? */ | ||
1878 | if (less(msg_seqno(buf_msg(*tail)), seq_no)) { | ||
1879 | (*tail)->next = buf; | ||
1880 | *tail = buf; | ||
1881 | return 1; | ||
1882 | } | ||
1883 | |||
1884 | /* Scan through queue and sort it in */ | ||
1885 | do { | ||
1886 | struct tipc_msg *msg = buf_msg(crs); | ||
1887 | |||
1888 | if (less(seq_no, msg_seqno(msg))) { | ||
1889 | buf->next = crs; | ||
1890 | if (prev) | ||
1891 | prev->next = buf; | ||
1892 | else | ||
1893 | *head = buf; | ||
1894 | return 1; | ||
1895 | } | ||
1896 | if (seq_no == msg_seqno(msg)) { | ||
1897 | break; | ||
1898 | } | ||
1899 | prev = crs; | ||
1900 | crs = crs->next; | ||
1901 | } | ||
1902 | while (crs); | ||
1903 | |||
1904 | /* Message is a duplicate of an existing message */ | ||
1905 | |||
1906 | buf_discard(buf); | ||
1907 | return 0; | ||
1908 | } | ||
1909 | |||
1910 | /** | ||
1911 | * link_handle_out_of_seq_msg - handle arrival of out-of-sequence packet | ||
1912 | */ | ||
1913 | |||
1914 | static void link_handle_out_of_seq_msg(struct link *l_ptr, | ||
1915 | struct sk_buff *buf) | ||
1916 | { | ||
1917 | u32 seq_no = msg_seqno(buf_msg(buf)); | ||
1918 | |||
1919 | if (likely(msg_user(buf_msg(buf)) == LINK_PROTOCOL)) { | ||
1920 | link_recv_proto_msg(l_ptr, buf); | ||
1921 | return; | ||
1922 | } | ||
1923 | |||
1924 | dbg("rx OOS msg: seq_no %u, expecting %u (%u)\n", | ||
1925 | seq_no, mod(l_ptr->next_in_no), l_ptr->next_in_no); | ||
1926 | |||
1927 | /* Record OOS packet arrival (force mismatch on next timeout) */ | ||
1928 | |||
1929 | l_ptr->checkpoint--; | ||
1930 | |||
1931 | /* | ||
1932 | * Discard packet if a duplicate; otherwise add it to deferred queue | ||
1933 | * and notify peer of gap as per protocol specification | ||
1934 | */ | ||
1935 | |||
1936 | if (less(seq_no, mod(l_ptr->next_in_no))) { | ||
1937 | l_ptr->stats.duplicates++; | ||
1938 | buf_discard(buf); | ||
1939 | return; | ||
1940 | } | ||
1941 | |||
1942 | if (link_defer_pkt(&l_ptr->oldest_deferred_in, | ||
1943 | &l_ptr->newest_deferred_in, buf)) { | ||
1944 | l_ptr->deferred_inqueue_sz++; | ||
1945 | l_ptr->stats.deferred_recv++; | ||
1946 | if ((l_ptr->deferred_inqueue_sz % 16) == 1) | ||
1947 | link_send_proto_msg(l_ptr, STATE_MSG, 0, 0, 0, 0, 0); | ||
1948 | } else | ||
1949 | l_ptr->stats.duplicates++; | ||
1950 | } | ||
1951 | |||
1952 | /* | ||
1953 | * Send protocol message to the other endpoint. | ||
1954 | */ | ||
1955 | void link_send_proto_msg(struct link *l_ptr, u32 msg_typ, int probe_msg, | ||
1956 | u32 gap, u32 tolerance, u32 priority, u32 ack_mtu) | ||
1957 | { | ||
1958 | struct sk_buff *buf = 0; | ||
1959 | struct tipc_msg *msg = l_ptr->pmsg; | ||
1960 | u32 msg_size = sizeof(l_ptr->proto_msg); | ||
1961 | |||
1962 | if (link_blocked(l_ptr)) | ||
1963 | return; | ||
1964 | msg_set_type(msg, msg_typ); | ||
1965 | msg_set_net_plane(msg, l_ptr->b_ptr->net_plane); | ||
1966 | msg_set_bcast_ack(msg, mod(l_ptr->owner->bclink.last_in)); | ||
1967 | msg_set_last_bcast(msg, bclink_get_last_sent()); | ||
1968 | |||
1969 | if (msg_typ == STATE_MSG) { | ||
1970 | u32 next_sent = mod(l_ptr->next_out_no); | ||
1971 | |||
1972 | if (!link_is_up(l_ptr)) | ||
1973 | return; | ||
1974 | if (l_ptr->next_out) | ||
1975 | next_sent = msg_seqno(buf_msg(l_ptr->next_out)); | ||
1976 | msg_set_next_sent(msg, next_sent); | ||
1977 | if (l_ptr->oldest_deferred_in) { | ||
1978 | u32 rec = msg_seqno(buf_msg(l_ptr->oldest_deferred_in)); | ||
1979 | gap = mod(rec - mod(l_ptr->next_in_no)); | ||
1980 | } | ||
1981 | msg_set_seq_gap(msg, gap); | ||
1982 | if (gap) | ||
1983 | l_ptr->stats.sent_nacks++; | ||
1984 | msg_set_link_tolerance(msg, tolerance); | ||
1985 | msg_set_linkprio(msg, priority); | ||
1986 | msg_set_max_pkt(msg, ack_mtu); | ||
1987 | msg_set_ack(msg, mod(l_ptr->next_in_no - 1)); | ||
1988 | msg_set_probe(msg, probe_msg != 0); | ||
1989 | if (probe_msg) { | ||
1990 | u32 mtu = l_ptr->max_pkt; | ||
1991 | |||
1992 | if ((mtu < l_ptr->max_pkt_target) && | ||
1993 | link_working_working(l_ptr) && | ||
1994 | l_ptr->fsm_msg_cnt) { | ||
1995 | msg_size = (mtu + (l_ptr->max_pkt_target - mtu)/2 + 2) & ~3; | ||
1996 | if (l_ptr->max_pkt_probes == 10) { | ||
1997 | l_ptr->max_pkt_target = (msg_size - 4); | ||
1998 | l_ptr->max_pkt_probes = 0; | ||
1999 | msg_size = (mtu + (l_ptr->max_pkt_target - mtu)/2 + 2) & ~3; | ||
2000 | } | ||
2001 | l_ptr->max_pkt_probes++; | ||
2002 | } | ||
2003 | |||
2004 | l_ptr->stats.sent_probes++; | ||
2005 | } | ||
2006 | l_ptr->stats.sent_states++; | ||
2007 | } else { /* RESET_MSG or ACTIVATE_MSG */ | ||
2008 | msg_set_ack(msg, mod(l_ptr->reset_checkpoint - 1)); | ||
2009 | msg_set_seq_gap(msg, 0); | ||
2010 | msg_set_next_sent(msg, 1); | ||
2011 | msg_set_link_tolerance(msg, l_ptr->tolerance); | ||
2012 | msg_set_linkprio(msg, l_ptr->priority); | ||
2013 | msg_set_max_pkt(msg, l_ptr->max_pkt_target); | ||
2014 | } | ||
2015 | |||
2016 | if (node_has_redundant_links(l_ptr->owner)) { | ||
2017 | msg_set_redundant_link(msg); | ||
2018 | } else { | ||
2019 | msg_clear_redundant_link(msg); | ||
2020 | } | ||
2021 | msg_set_linkprio(msg, l_ptr->priority); | ||
2022 | |||
2023 | /* Ensure sequence number will not fit : */ | ||
2024 | |||
2025 | msg_set_seqno(msg, mod(l_ptr->next_out_no + (0xffff/2))); | ||
2026 | |||
2027 | /* Congestion? */ | ||
2028 | |||
2029 | if (bearer_congested(l_ptr->b_ptr, l_ptr)) { | ||
2030 | if (!l_ptr->proto_msg_queue) { | ||
2031 | l_ptr->proto_msg_queue = | ||
2032 | buf_acquire(sizeof(l_ptr->proto_msg)); | ||
2033 | } | ||
2034 | buf = l_ptr->proto_msg_queue; | ||
2035 | if (!buf) | ||
2036 | return; | ||
2037 | memcpy(buf->data, (unchar *)msg, sizeof(l_ptr->proto_msg)); | ||
2038 | return; | ||
2039 | } | ||
2040 | msg_set_timestamp(msg, jiffies_to_msecs(jiffies)); | ||
2041 | |||
2042 | /* Message can be sent */ | ||
2043 | |||
2044 | msg_dbg(msg, ">>"); | ||
2045 | |||
2046 | buf = buf_acquire(msg_size); | ||
2047 | if (!buf) | ||
2048 | return; | ||
2049 | |||
2050 | memcpy(buf->data, (unchar *)msg, sizeof(l_ptr->proto_msg)); | ||
2051 | msg_set_size(buf_msg(buf), msg_size); | ||
2052 | |||
2053 | if (bearer_send(l_ptr->b_ptr, buf, &l_ptr->media_addr)) { | ||
2054 | l_ptr->unacked_window = 0; | ||
2055 | buf_discard(buf); | ||
2056 | return; | ||
2057 | } | ||
2058 | |||
2059 | /* New congestion */ | ||
2060 | bearer_schedule(l_ptr->b_ptr, l_ptr); | ||
2061 | l_ptr->proto_msg_queue = buf; | ||
2062 | l_ptr->stats.bearer_congs++; | ||
2063 | } | ||
2064 | |||
2065 | /* | ||
2066 | * Receive protocol message : | ||
2067 | * Note that network plane id propagates through the network, and may | ||
2068 | * change at any time. The node with lowest address rules | ||
2069 | */ | ||
2070 | |||
2071 | static void link_recv_proto_msg(struct link *l_ptr, struct sk_buff *buf) | ||
2072 | { | ||
2073 | u32 rec_gap = 0; | ||
2074 | u32 max_pkt_info; | ||
2075 | u32 max_pkt_ack; | ||
2076 | u32 msg_tol; | ||
2077 | struct tipc_msg *msg = buf_msg(buf); | ||
2078 | |||
2079 | dbg("AT(%u):", jiffies_to_msecs(jiffies)); | ||
2080 | msg_dbg(msg, "<<"); | ||
2081 | if (link_blocked(l_ptr)) | ||
2082 | goto exit; | ||
2083 | |||
2084 | /* record unnumbered packet arrival (force mismatch on next timeout) */ | ||
2085 | |||
2086 | l_ptr->checkpoint--; | ||
2087 | |||
2088 | if (l_ptr->b_ptr->net_plane != msg_net_plane(msg)) | ||
2089 | if (tipc_own_addr > msg_prevnode(msg)) | ||
2090 | l_ptr->b_ptr->net_plane = msg_net_plane(msg); | ||
2091 | |||
2092 | l_ptr->owner->permit_changeover = msg_redundant_link(msg); | ||
2093 | |||
2094 | switch (msg_type(msg)) { | ||
2095 | |||
2096 | case RESET_MSG: | ||
2097 | if (!link_working_unknown(l_ptr) && l_ptr->peer_session) { | ||
2098 | if (msg_session(msg) == l_ptr->peer_session) { | ||
2099 | dbg("Duplicate RESET: %u<->%u\n", | ||
2100 | msg_session(msg), l_ptr->peer_session); | ||
2101 | break; /* duplicate: ignore */ | ||
2102 | } | ||
2103 | } | ||
2104 | /* fall thru' */ | ||
2105 | case ACTIVATE_MSG: | ||
2106 | /* Update link settings according other endpoint's values */ | ||
2107 | |||
2108 | strcpy((strrchr(l_ptr->name, ':') + 1), (char *)msg_data(msg)); | ||
2109 | |||
2110 | if ((msg_tol = msg_link_tolerance(msg)) && | ||
2111 | (msg_tol > l_ptr->tolerance)) | ||
2112 | link_set_supervision_props(l_ptr, msg_tol); | ||
2113 | |||
2114 | if (msg_linkprio(msg) > l_ptr->priority) | ||
2115 | l_ptr->priority = msg_linkprio(msg); | ||
2116 | |||
2117 | max_pkt_info = msg_max_pkt(msg); | ||
2118 | if (max_pkt_info) { | ||
2119 | if (max_pkt_info < l_ptr->max_pkt_target) | ||
2120 | l_ptr->max_pkt_target = max_pkt_info; | ||
2121 | if (l_ptr->max_pkt > l_ptr->max_pkt_target) | ||
2122 | l_ptr->max_pkt = l_ptr->max_pkt_target; | ||
2123 | } else { | ||
2124 | l_ptr->max_pkt = l_ptr->max_pkt_target; | ||
2125 | } | ||
2126 | l_ptr->owner->bclink.supported = (max_pkt_info != 0); | ||
2127 | |||
2128 | link_state_event(l_ptr, msg_type(msg)); | ||
2129 | |||
2130 | l_ptr->peer_session = msg_session(msg); | ||
2131 | l_ptr->peer_bearer_id = msg_bearer_id(msg); | ||
2132 | |||
2133 | /* Synchronize broadcast sequence numbers */ | ||
2134 | if (!node_has_redundant_links(l_ptr->owner)) { | ||
2135 | l_ptr->owner->bclink.last_in = mod(msg_last_bcast(msg)); | ||
2136 | } | ||
2137 | break; | ||
2138 | case STATE_MSG: | ||
2139 | |||
2140 | if ((msg_tol = msg_link_tolerance(msg))) | ||
2141 | link_set_supervision_props(l_ptr, msg_tol); | ||
2142 | |||
2143 | if (msg_linkprio(msg) && | ||
2144 | (msg_linkprio(msg) != l_ptr->priority)) { | ||
2145 | warn("Changing prio <%s>: %u->%u\n", | ||
2146 | l_ptr->name, l_ptr->priority, msg_linkprio(msg)); | ||
2147 | l_ptr->priority = msg_linkprio(msg); | ||
2148 | link_reset(l_ptr); /* Enforce change to take effect */ | ||
2149 | break; | ||
2150 | } | ||
2151 | link_state_event(l_ptr, TRAFFIC_MSG_EVT); | ||
2152 | l_ptr->stats.recv_states++; | ||
2153 | if (link_reset_unknown(l_ptr)) | ||
2154 | break; | ||
2155 | |||
2156 | if (less_eq(mod(l_ptr->next_in_no), msg_next_sent(msg))) { | ||
2157 | rec_gap = mod(msg_next_sent(msg) - | ||
2158 | mod(l_ptr->next_in_no)); | ||
2159 | } | ||
2160 | |||
2161 | max_pkt_ack = msg_max_pkt(msg); | ||
2162 | if (max_pkt_ack > l_ptr->max_pkt) { | ||
2163 | dbg("Link <%s> updated MTU %u -> %u\n", | ||
2164 | l_ptr->name, l_ptr->max_pkt, max_pkt_ack); | ||
2165 | l_ptr->max_pkt = max_pkt_ack; | ||
2166 | l_ptr->max_pkt_probes = 0; | ||
2167 | } | ||
2168 | |||
2169 | max_pkt_ack = 0; | ||
2170 | if (msg_probe(msg)) { | ||
2171 | l_ptr->stats.recv_probes++; | ||
2172 | if (msg_size(msg) > sizeof(l_ptr->proto_msg)) { | ||
2173 | max_pkt_ack = msg_size(msg); | ||
2174 | } | ||
2175 | } | ||
2176 | |||
2177 | /* Protocol message before retransmits, reduce loss risk */ | ||
2178 | |||
2179 | bclink_check_gap(l_ptr->owner, msg_last_bcast(msg)); | ||
2180 | |||
2181 | if (rec_gap || (msg_probe(msg))) { | ||
2182 | link_send_proto_msg(l_ptr, STATE_MSG, | ||
2183 | 0, rec_gap, 0, 0, max_pkt_ack); | ||
2184 | } | ||
2185 | if (msg_seq_gap(msg)) { | ||
2186 | msg_dbg(msg, "With Gap:"); | ||
2187 | l_ptr->stats.recv_nacks++; | ||
2188 | link_retransmit(l_ptr, l_ptr->first_out, | ||
2189 | msg_seq_gap(msg)); | ||
2190 | } | ||
2191 | break; | ||
2192 | default: | ||
2193 | msg_dbg(buf_msg(buf), "<DISCARDING UNKNOWN<"); | ||
2194 | } | ||
2195 | exit: | ||
2196 | buf_discard(buf); | ||
2197 | } | ||
2198 | |||
2199 | |||
2200 | /* | ||
2201 | * link_tunnel(): Send one message via a link belonging to | ||
2202 | * another bearer. Owner node is locked. | ||
2203 | */ | ||
2204 | void link_tunnel(struct link *l_ptr, | ||
2205 | struct tipc_msg *tunnel_hdr, | ||
2206 | struct tipc_msg *msg, | ||
2207 | u32 selector) | ||
2208 | { | ||
2209 | struct link *tunnel; | ||
2210 | struct sk_buff *buf; | ||
2211 | u32 length = msg_size(msg); | ||
2212 | |||
2213 | tunnel = l_ptr->owner->active_links[selector & 1]; | ||
2214 | if (!link_is_up(tunnel)) | ||
2215 | return; | ||
2216 | msg_set_size(tunnel_hdr, length + INT_H_SIZE); | ||
2217 | buf = buf_acquire(length + INT_H_SIZE); | ||
2218 | if (!buf) | ||
2219 | return; | ||
2220 | memcpy(buf->data, (unchar *)tunnel_hdr, INT_H_SIZE); | ||
2221 | memcpy(buf->data + INT_H_SIZE, (unchar *)msg, length); | ||
2222 | dbg("%c->%c:", l_ptr->b_ptr->net_plane, tunnel->b_ptr->net_plane); | ||
2223 | msg_dbg(buf_msg(buf), ">SEND>"); | ||
2224 | assert(tunnel); | ||
2225 | link_send_buf(tunnel, buf); | ||
2226 | } | ||
2227 | |||
2228 | |||
2229 | |||
2230 | /* | ||
2231 | * changeover(): Send whole message queue via the remaining link | ||
2232 | * Owner node is locked. | ||
2233 | */ | ||
2234 | |||
2235 | void link_changeover(struct link *l_ptr) | ||
2236 | { | ||
2237 | u32 msgcount = l_ptr->out_queue_size; | ||
2238 | struct sk_buff *crs = l_ptr->first_out; | ||
2239 | struct link *tunnel = l_ptr->owner->active_links[0]; | ||
2240 | int split_bundles = node_has_redundant_links(l_ptr->owner); | ||
2241 | struct tipc_msg tunnel_hdr; | ||
2242 | |||
2243 | if (!tunnel) | ||
2244 | return; | ||
2245 | |||
2246 | if (!l_ptr->owner->permit_changeover) | ||
2247 | return; | ||
2248 | |||
2249 | msg_init(&tunnel_hdr, CHANGEOVER_PROTOCOL, | ||
2250 | ORIGINAL_MSG, TIPC_OK, INT_H_SIZE, l_ptr->addr); | ||
2251 | msg_set_bearer_id(&tunnel_hdr, l_ptr->peer_bearer_id); | ||
2252 | msg_set_msgcnt(&tunnel_hdr, msgcount); | ||
2253 | if (!l_ptr->first_out) { | ||
2254 | struct sk_buff *buf; | ||
2255 | |||
2256 | assert(!msgcount); | ||
2257 | buf = buf_acquire(INT_H_SIZE); | ||
2258 | if (buf) { | ||
2259 | memcpy(buf->data, (unchar *)&tunnel_hdr, INT_H_SIZE); | ||
2260 | msg_set_size(&tunnel_hdr, INT_H_SIZE); | ||
2261 | dbg("%c->%c:", l_ptr->b_ptr->net_plane, | ||
2262 | tunnel->b_ptr->net_plane); | ||
2263 | msg_dbg(&tunnel_hdr, "EMPTY>SEND>"); | ||
2264 | link_send_buf(tunnel, buf); | ||
2265 | } else { | ||
2266 | warn("Memory squeeze; link changeover failed\n"); | ||
2267 | } | ||
2268 | return; | ||
2269 | } | ||
2270 | while (crs) { | ||
2271 | struct tipc_msg *msg = buf_msg(crs); | ||
2272 | |||
2273 | if ((msg_user(msg) == MSG_BUNDLER) && split_bundles) { | ||
2274 | u32 msgcount = msg_msgcnt(msg); | ||
2275 | struct tipc_msg *m = msg_get_wrapped(msg); | ||
2276 | unchar* pos = (unchar*)m; | ||
2277 | |||
2278 | while (msgcount--) { | ||
2279 | msg_set_seqno(m,msg_seqno(msg)); | ||
2280 | link_tunnel(l_ptr, &tunnel_hdr, m, | ||
2281 | msg_link_selector(m)); | ||
2282 | pos += align(msg_size(m)); | ||
2283 | m = (struct tipc_msg *)pos; | ||
2284 | } | ||
2285 | } else { | ||
2286 | link_tunnel(l_ptr, &tunnel_hdr, msg, | ||
2287 | msg_link_selector(msg)); | ||
2288 | } | ||
2289 | crs = crs->next; | ||
2290 | } | ||
2291 | } | ||
2292 | |||
2293 | void link_send_duplicate(struct link *l_ptr, struct link *tunnel) | ||
2294 | { | ||
2295 | struct sk_buff *iter; | ||
2296 | struct tipc_msg tunnel_hdr; | ||
2297 | |||
2298 | msg_init(&tunnel_hdr, CHANGEOVER_PROTOCOL, | ||
2299 | DUPLICATE_MSG, TIPC_OK, INT_H_SIZE, l_ptr->addr); | ||
2300 | msg_set_msgcnt(&tunnel_hdr, l_ptr->out_queue_size); | ||
2301 | msg_set_bearer_id(&tunnel_hdr, l_ptr->peer_bearer_id); | ||
2302 | iter = l_ptr->first_out; | ||
2303 | while (iter) { | ||
2304 | struct sk_buff *outbuf; | ||
2305 | struct tipc_msg *msg = buf_msg(iter); | ||
2306 | u32 length = msg_size(msg); | ||
2307 | |||
2308 | if (msg_user(msg) == MSG_BUNDLER) | ||
2309 | msg_set_type(msg, CLOSED_MSG); | ||
2310 | msg_set_ack(msg, mod(l_ptr->next_in_no - 1)); /* Update */ | ||
2311 | msg_set_bcast_ack(msg, l_ptr->owner->bclink.last_in); | ||
2312 | msg_set_size(&tunnel_hdr, length + INT_H_SIZE); | ||
2313 | outbuf = buf_acquire(length + INT_H_SIZE); | ||
2314 | if (outbuf == NULL) { | ||
2315 | warn("Memory squeeze; buffer duplication failed\n"); | ||
2316 | return; | ||
2317 | } | ||
2318 | memcpy(outbuf->data, (unchar *)&tunnel_hdr, INT_H_SIZE); | ||
2319 | memcpy(outbuf->data + INT_H_SIZE, iter->data, length); | ||
2320 | dbg("%c->%c:", l_ptr->b_ptr->net_plane, | ||
2321 | tunnel->b_ptr->net_plane); | ||
2322 | msg_dbg(buf_msg(outbuf), ">SEND>"); | ||
2323 | link_send_buf(tunnel, outbuf); | ||
2324 | if (!link_is_up(l_ptr)) | ||
2325 | return; | ||
2326 | iter = iter->next; | ||
2327 | } | ||
2328 | } | ||
2329 | |||
2330 | |||
2331 | |||
2332 | /** | ||
2333 | * buf_extract - extracts embedded TIPC message from another message | ||
2334 | * @skb: encapsulating message buffer | ||
2335 | * @from_pos: offset to extract from | ||
2336 | * | ||
2337 | * Returns a new message buffer containing an embedded message. The | ||
2338 | * encapsulating message itself is left unchanged. | ||
2339 | */ | ||
2340 | |||
2341 | static struct sk_buff *buf_extract(struct sk_buff *skb, u32 from_pos) | ||
2342 | { | ||
2343 | struct tipc_msg *msg = (struct tipc_msg *)(skb->data + from_pos); | ||
2344 | u32 size = msg_size(msg); | ||
2345 | struct sk_buff *eb; | ||
2346 | |||
2347 | eb = buf_acquire(size); | ||
2348 | if (eb) | ||
2349 | memcpy(eb->data, (unchar *)msg, size); | ||
2350 | return eb; | ||
2351 | } | ||
2352 | |||
2353 | /* | ||
2354 | * link_recv_changeover_msg(): Receive tunneled packet sent | ||
2355 | * via other link. Node is locked. Return extracted buffer. | ||
2356 | */ | ||
2357 | |||
2358 | static int link_recv_changeover_msg(struct link **l_ptr, | ||
2359 | struct sk_buff **buf) | ||
2360 | { | ||
2361 | struct sk_buff *tunnel_buf = *buf; | ||
2362 | struct link *dest_link; | ||
2363 | struct tipc_msg *msg; | ||
2364 | struct tipc_msg *tunnel_msg = buf_msg(tunnel_buf); | ||
2365 | u32 msg_typ = msg_type(tunnel_msg); | ||
2366 | u32 msg_count = msg_msgcnt(tunnel_msg); | ||
2367 | |||
2368 | dest_link = (*l_ptr)->owner->links[msg_bearer_id(tunnel_msg)]; | ||
2369 | assert(dest_link != *l_ptr); | ||
2370 | if (!dest_link) { | ||
2371 | msg_dbg(tunnel_msg, "NOLINK/<REC<"); | ||
2372 | goto exit; | ||
2373 | } | ||
2374 | dbg("%c<-%c:", dest_link->b_ptr->net_plane, | ||
2375 | (*l_ptr)->b_ptr->net_plane); | ||
2376 | *l_ptr = dest_link; | ||
2377 | msg = msg_get_wrapped(tunnel_msg); | ||
2378 | |||
2379 | if (msg_typ == DUPLICATE_MSG) { | ||
2380 | if (less(msg_seqno(msg), mod(dest_link->next_in_no))) { | ||
2381 | msg_dbg(tunnel_msg, "DROP/<REC<"); | ||
2382 | goto exit; | ||
2383 | } | ||
2384 | *buf = buf_extract(tunnel_buf,INT_H_SIZE); | ||
2385 | if (*buf == NULL) { | ||
2386 | warn("Memory squeeze; failed to extract msg\n"); | ||
2387 | goto exit; | ||
2388 | } | ||
2389 | msg_dbg(tunnel_msg, "TNL<REC<"); | ||
2390 | buf_discard(tunnel_buf); | ||
2391 | return 1; | ||
2392 | } | ||
2393 | |||
2394 | /* First original message ?: */ | ||
2395 | |||
2396 | if (link_is_up(dest_link)) { | ||
2397 | msg_dbg(tunnel_msg, "UP/FIRST/<REC<"); | ||
2398 | link_reset(dest_link); | ||
2399 | dest_link->exp_msg_count = msg_count; | ||
2400 | if (!msg_count) | ||
2401 | goto exit; | ||
2402 | } else if (dest_link->exp_msg_count == START_CHANGEOVER) { | ||
2403 | msg_dbg(tunnel_msg, "BLK/FIRST/<REC<"); | ||
2404 | dest_link->exp_msg_count = msg_count; | ||
2405 | if (!msg_count) | ||
2406 | goto exit; | ||
2407 | } | ||
2408 | |||
2409 | /* Receive original message */ | ||
2410 | |||
2411 | if (dest_link->exp_msg_count == 0) { | ||
2412 | msg_dbg(tunnel_msg, "OVERDUE/DROP/<REC<"); | ||
2413 | dbg_print_link(dest_link, "LINK:"); | ||
2414 | goto exit; | ||
2415 | } | ||
2416 | dest_link->exp_msg_count--; | ||
2417 | if (less(msg_seqno(msg), dest_link->reset_checkpoint)) { | ||
2418 | msg_dbg(tunnel_msg, "DROP/DUPL/<REC<"); | ||
2419 | goto exit; | ||
2420 | } else { | ||
2421 | *buf = buf_extract(tunnel_buf, INT_H_SIZE); | ||
2422 | if (*buf != NULL) { | ||
2423 | msg_dbg(tunnel_msg, "TNL<REC<"); | ||
2424 | buf_discard(tunnel_buf); | ||
2425 | return 1; | ||
2426 | } else { | ||
2427 | warn("Memory squeeze; dropped incoming msg\n"); | ||
2428 | } | ||
2429 | } | ||
2430 | exit: | ||
2431 | *buf = 0; | ||
2432 | buf_discard(tunnel_buf); | ||
2433 | return 0; | ||
2434 | } | ||
2435 | |||
2436 | /* | ||
2437 | * Bundler functionality: | ||
2438 | */ | ||
2439 | void link_recv_bundle(struct sk_buff *buf) | ||
2440 | { | ||
2441 | u32 msgcount = msg_msgcnt(buf_msg(buf)); | ||
2442 | u32 pos = INT_H_SIZE; | ||
2443 | struct sk_buff *obuf; | ||
2444 | |||
2445 | msg_dbg(buf_msg(buf), "<BNDL<: "); | ||
2446 | while (msgcount--) { | ||
2447 | obuf = buf_extract(buf, pos); | ||
2448 | if (obuf == NULL) { | ||
2449 | char addr_string[16]; | ||
2450 | |||
2451 | warn("Buffer allocation failure;\n"); | ||
2452 | warn(" incoming message(s) from %s lost\n", | ||
2453 | addr_string_fill(addr_string, | ||
2454 | msg_orignode(buf_msg(buf)))); | ||
2455 | return; | ||
2456 | }; | ||
2457 | pos += align(msg_size(buf_msg(obuf))); | ||
2458 | msg_dbg(buf_msg(obuf), " /"); | ||
2459 | net_route_msg(obuf); | ||
2460 | } | ||
2461 | buf_discard(buf); | ||
2462 | } | ||
2463 | |||
2464 | /* | ||
2465 | * Fragmentation/defragmentation: | ||
2466 | */ | ||
2467 | |||
2468 | |||
2469 | /* | ||
2470 | * link_send_long_buf: Entry for buffers needing fragmentation. | ||
2471 | * The buffer is complete, inclusive total message length. | ||
2472 | * Returns user data length. | ||
2473 | */ | ||
2474 | int link_send_long_buf(struct link *l_ptr, struct sk_buff *buf) | ||
2475 | { | ||
2476 | struct tipc_msg *inmsg = buf_msg(buf); | ||
2477 | struct tipc_msg fragm_hdr; | ||
2478 | u32 insize = msg_size(inmsg); | ||
2479 | u32 dsz = msg_data_sz(inmsg); | ||
2480 | unchar *crs = buf->data; | ||
2481 | u32 rest = insize; | ||
2482 | u32 pack_sz = link_max_pkt(l_ptr); | ||
2483 | u32 fragm_sz = pack_sz - INT_H_SIZE; | ||
2484 | u32 fragm_no = 1; | ||
2485 | u32 destaddr = msg_destnode(inmsg); | ||
2486 | |||
2487 | if (msg_short(inmsg)) | ||
2488 | destaddr = l_ptr->addr; | ||
2489 | |||
2490 | if (msg_routed(inmsg)) | ||
2491 | msg_set_prevnode(inmsg, tipc_own_addr); | ||
2492 | |||
2493 | /* Prepare reusable fragment header: */ | ||
2494 | |||
2495 | msg_init(&fragm_hdr, MSG_FRAGMENTER, FIRST_FRAGMENT, | ||
2496 | TIPC_OK, INT_H_SIZE, destaddr); | ||
2497 | msg_set_link_selector(&fragm_hdr, msg_link_selector(inmsg)); | ||
2498 | msg_set_long_msgno(&fragm_hdr, mod(l_ptr->long_msg_seq_no++)); | ||
2499 | msg_set_fragm_no(&fragm_hdr, fragm_no); | ||
2500 | l_ptr->stats.sent_fragmented++; | ||
2501 | |||
2502 | /* Chop up message: */ | ||
2503 | |||
2504 | while (rest > 0) { | ||
2505 | struct sk_buff *fragm; | ||
2506 | |||
2507 | if (rest <= fragm_sz) { | ||
2508 | fragm_sz = rest; | ||
2509 | msg_set_type(&fragm_hdr, LAST_FRAGMENT); | ||
2510 | } | ||
2511 | fragm = buf_acquire(fragm_sz + INT_H_SIZE); | ||
2512 | if (fragm == NULL) { | ||
2513 | warn("Memory squeeze; failed to fragment msg\n"); | ||
2514 | dsz = -ENOMEM; | ||
2515 | goto exit; | ||
2516 | } | ||
2517 | msg_set_size(&fragm_hdr, fragm_sz + INT_H_SIZE); | ||
2518 | memcpy(fragm->data, (unchar *)&fragm_hdr, INT_H_SIZE); | ||
2519 | memcpy(fragm->data + INT_H_SIZE, crs, fragm_sz); | ||
2520 | |||
2521 | /* Send queued messages first, if any: */ | ||
2522 | |||
2523 | l_ptr->stats.sent_fragments++; | ||
2524 | link_send_buf(l_ptr, fragm); | ||
2525 | if (!link_is_up(l_ptr)) | ||
2526 | return dsz; | ||
2527 | msg_set_fragm_no(&fragm_hdr, ++fragm_no); | ||
2528 | rest -= fragm_sz; | ||
2529 | crs += fragm_sz; | ||
2530 | msg_set_type(&fragm_hdr, FRAGMENT); | ||
2531 | } | ||
2532 | exit: | ||
2533 | buf_discard(buf); | ||
2534 | return dsz; | ||
2535 | } | ||
2536 | |||
2537 | /* | ||
2538 | * A pending message being re-assembled must store certain values | ||
2539 | * to handle subsequent fragments correctly. The following functions | ||
2540 | * help storing these values in unused, available fields in the | ||
2541 | * pending message. This makes dynamic memory allocation unecessary. | ||
2542 | */ | ||
2543 | |||
2544 | static inline u32 get_long_msg_seqno(struct sk_buff *buf) | ||
2545 | { | ||
2546 | return msg_seqno(buf_msg(buf)); | ||
2547 | } | ||
2548 | |||
2549 | static inline void set_long_msg_seqno(struct sk_buff *buf, u32 seqno) | ||
2550 | { | ||
2551 | msg_set_seqno(buf_msg(buf), seqno); | ||
2552 | } | ||
2553 | |||
2554 | static inline u32 get_fragm_size(struct sk_buff *buf) | ||
2555 | { | ||
2556 | return msg_ack(buf_msg(buf)); | ||
2557 | } | ||
2558 | |||
2559 | static inline void set_fragm_size(struct sk_buff *buf, u32 sz) | ||
2560 | { | ||
2561 | msg_set_ack(buf_msg(buf), sz); | ||
2562 | } | ||
2563 | |||
2564 | static inline u32 get_expected_frags(struct sk_buff *buf) | ||
2565 | { | ||
2566 | return msg_bcast_ack(buf_msg(buf)); | ||
2567 | } | ||
2568 | |||
2569 | static inline void set_expected_frags(struct sk_buff *buf, u32 exp) | ||
2570 | { | ||
2571 | msg_set_bcast_ack(buf_msg(buf), exp); | ||
2572 | } | ||
2573 | |||
2574 | static inline u32 get_timer_cnt(struct sk_buff *buf) | ||
2575 | { | ||
2576 | return msg_reroute_cnt(buf_msg(buf)); | ||
2577 | } | ||
2578 | |||
2579 | static inline void incr_timer_cnt(struct sk_buff *buf) | ||
2580 | { | ||
2581 | msg_incr_reroute_cnt(buf_msg(buf)); | ||
2582 | } | ||
2583 | |||
2584 | /* | ||
2585 | * link_recv_fragment(): Called with node lock on. Returns | ||
2586 | * the reassembled buffer if message is complete. | ||
2587 | */ | ||
2588 | int link_recv_fragment(struct sk_buff **pending, struct sk_buff **fb, | ||
2589 | struct tipc_msg **m) | ||
2590 | { | ||
2591 | struct sk_buff *prev = 0; | ||
2592 | struct sk_buff *fbuf = *fb; | ||
2593 | struct tipc_msg *fragm = buf_msg(fbuf); | ||
2594 | struct sk_buff *pbuf = *pending; | ||
2595 | u32 long_msg_seq_no = msg_long_msgno(fragm); | ||
2596 | |||
2597 | *fb = 0; | ||
2598 | msg_dbg(fragm,"FRG<REC<"); | ||
2599 | |||
2600 | /* Is there an incomplete message waiting for this fragment? */ | ||
2601 | |||
2602 | while (pbuf && ((msg_seqno(buf_msg(pbuf)) != long_msg_seq_no) | ||
2603 | || (msg_orignode(fragm) != msg_orignode(buf_msg(pbuf))))) { | ||
2604 | prev = pbuf; | ||
2605 | pbuf = pbuf->next; | ||
2606 | } | ||
2607 | |||
2608 | if (!pbuf && (msg_type(fragm) == FIRST_FRAGMENT)) { | ||
2609 | struct tipc_msg *imsg = (struct tipc_msg *)msg_data(fragm); | ||
2610 | u32 msg_sz = msg_size(imsg); | ||
2611 | u32 fragm_sz = msg_data_sz(fragm); | ||
2612 | u32 exp_fragm_cnt = msg_sz/fragm_sz + !!(msg_sz % fragm_sz); | ||
2613 | u32 max = TIPC_MAX_USER_MSG_SIZE + LONG_H_SIZE; | ||
2614 | if (msg_type(imsg) == TIPC_MCAST_MSG) | ||
2615 | max = TIPC_MAX_USER_MSG_SIZE + MCAST_H_SIZE; | ||
2616 | if (msg_size(imsg) > max) { | ||
2617 | msg_dbg(fragm,"<REC<Oversized: "); | ||
2618 | buf_discard(fbuf); | ||
2619 | return 0; | ||
2620 | } | ||
2621 | pbuf = buf_acquire(msg_size(imsg)); | ||
2622 | if (pbuf != NULL) { | ||
2623 | pbuf->next = *pending; | ||
2624 | *pending = pbuf; | ||
2625 | memcpy(pbuf->data, (unchar *)imsg, msg_data_sz(fragm)); | ||
2626 | |||
2627 | /* Prepare buffer for subsequent fragments. */ | ||
2628 | |||
2629 | set_long_msg_seqno(pbuf, long_msg_seq_no); | ||
2630 | set_fragm_size(pbuf,fragm_sz); | ||
2631 | set_expected_frags(pbuf,exp_fragm_cnt - 1); | ||
2632 | } else { | ||
2633 | warn("Memory squeeze; got no defragmenting buffer\n"); | ||
2634 | } | ||
2635 | buf_discard(fbuf); | ||
2636 | return 0; | ||
2637 | } else if (pbuf && (msg_type(fragm) != FIRST_FRAGMENT)) { | ||
2638 | u32 dsz = msg_data_sz(fragm); | ||
2639 | u32 fsz = get_fragm_size(pbuf); | ||
2640 | u32 crs = ((msg_fragm_no(fragm) - 1) * fsz); | ||
2641 | u32 exp_frags = get_expected_frags(pbuf) - 1; | ||
2642 | memcpy(pbuf->data + crs, msg_data(fragm), dsz); | ||
2643 | buf_discard(fbuf); | ||
2644 | |||
2645 | /* Is message complete? */ | ||
2646 | |||
2647 | if (exp_frags == 0) { | ||
2648 | if (prev) | ||
2649 | prev->next = pbuf->next; | ||
2650 | else | ||
2651 | *pending = pbuf->next; | ||
2652 | msg_reset_reroute_cnt(buf_msg(pbuf)); | ||
2653 | *fb = pbuf; | ||
2654 | *m = buf_msg(pbuf); | ||
2655 | return 1; | ||
2656 | } | ||
2657 | set_expected_frags(pbuf,exp_frags); | ||
2658 | return 0; | ||
2659 | } | ||
2660 | dbg(" Discarding orphan fragment %x\n",fbuf); | ||
2661 | msg_dbg(fragm,"ORPHAN:"); | ||
2662 | dbg("Pending long buffers:\n"); | ||
2663 | dbg_print_buf_chain(*pending); | ||
2664 | buf_discard(fbuf); | ||
2665 | return 0; | ||
2666 | } | ||
2667 | |||
2668 | /** | ||
2669 | * link_check_defragm_bufs - flush stale incoming message fragments | ||
2670 | * @l_ptr: pointer to link | ||
2671 | */ | ||
2672 | |||
2673 | static void link_check_defragm_bufs(struct link *l_ptr) | ||
2674 | { | ||
2675 | struct sk_buff *prev = 0; | ||
2676 | struct sk_buff *next = 0; | ||
2677 | struct sk_buff *buf = l_ptr->defragm_buf; | ||
2678 | |||
2679 | if (!buf) | ||
2680 | return; | ||
2681 | if (!link_working_working(l_ptr)) | ||
2682 | return; | ||
2683 | while (buf) { | ||
2684 | u32 cnt = get_timer_cnt(buf); | ||
2685 | |||
2686 | next = buf->next; | ||
2687 | if (cnt < 4) { | ||
2688 | incr_timer_cnt(buf); | ||
2689 | prev = buf; | ||
2690 | } else { | ||
2691 | dbg(" Discarding incomplete long buffer\n"); | ||
2692 | msg_dbg(buf_msg(buf), "LONG:"); | ||
2693 | dbg_print_link(l_ptr, "curr:"); | ||
2694 | dbg("Pending long buffers:\n"); | ||
2695 | dbg_print_buf_chain(l_ptr->defragm_buf); | ||
2696 | if (prev) | ||
2697 | prev->next = buf->next; | ||
2698 | else | ||
2699 | l_ptr->defragm_buf = buf->next; | ||
2700 | buf_discard(buf); | ||
2701 | } | ||
2702 | buf = next; | ||
2703 | } | ||
2704 | } | ||
2705 | |||
2706 | |||
2707 | |||
2708 | static void link_set_supervision_props(struct link *l_ptr, u32 tolerance) | ||
2709 | { | ||
2710 | l_ptr->tolerance = tolerance; | ||
2711 | l_ptr->continuity_interval = | ||
2712 | ((tolerance / 4) > 500) ? 500 : tolerance / 4; | ||
2713 | l_ptr->abort_limit = tolerance / (l_ptr->continuity_interval / 4); | ||
2714 | } | ||
2715 | |||
2716 | |||
2717 | void link_set_queue_limits(struct link *l_ptr, u32 window) | ||
2718 | { | ||
2719 | /* Data messages from this node, inclusive FIRST_FRAGM */ | ||
2720 | l_ptr->queue_limit[DATA_LOW] = window; | ||
2721 | l_ptr->queue_limit[DATA_MEDIUM] = (window / 3) * 4; | ||
2722 | l_ptr->queue_limit[DATA_HIGH] = (window / 3) * 5; | ||
2723 | l_ptr->queue_limit[DATA_CRITICAL] = (window / 3) * 6; | ||
2724 | /* Transiting data messages,inclusive FIRST_FRAGM */ | ||
2725 | l_ptr->queue_limit[DATA_LOW + 4] = 300; | ||
2726 | l_ptr->queue_limit[DATA_MEDIUM + 4] = 600; | ||
2727 | l_ptr->queue_limit[DATA_HIGH + 4] = 900; | ||
2728 | l_ptr->queue_limit[DATA_CRITICAL + 4] = 1200; | ||
2729 | l_ptr->queue_limit[CONN_MANAGER] = 1200; | ||
2730 | l_ptr->queue_limit[ROUTE_DISTRIBUTOR] = 1200; | ||
2731 | l_ptr->queue_limit[CHANGEOVER_PROTOCOL] = 2500; | ||
2732 | l_ptr->queue_limit[NAME_DISTRIBUTOR] = 3000; | ||
2733 | /* FRAGMENT and LAST_FRAGMENT packets */ | ||
2734 | l_ptr->queue_limit[MSG_FRAGMENTER] = 4000; | ||
2735 | } | ||
2736 | |||
2737 | /** | ||
2738 | * link_find_link - locate link by name | ||
2739 | * @name - ptr to link name string | ||
2740 | * @node - ptr to area to be filled with ptr to associated node | ||
2741 | * | ||
2742 | * Caller must hold 'net_lock' to ensure node and bearer are not deleted; | ||
2743 | * this also prevents link deletion. | ||
2744 | * | ||
2745 | * Returns pointer to link (or 0 if invalid link name). | ||
2746 | */ | ||
2747 | |||
2748 | static struct link *link_find_link(const char *name, struct node **node) | ||
2749 | { | ||
2750 | struct link_name link_name_parts; | ||
2751 | struct bearer *b_ptr; | ||
2752 | struct link *l_ptr; | ||
2753 | |||
2754 | if (!link_name_validate(name, &link_name_parts)) | ||
2755 | return 0; | ||
2756 | |||
2757 | b_ptr = bearer_find_interface(link_name_parts.if_local); | ||
2758 | if (!b_ptr) | ||
2759 | return 0; | ||
2760 | |||
2761 | *node = node_find(link_name_parts.addr_peer); | ||
2762 | if (!*node) | ||
2763 | return 0; | ||
2764 | |||
2765 | l_ptr = (*node)->links[b_ptr->identity]; | ||
2766 | if (!l_ptr || strcmp(l_ptr->name, name)) | ||
2767 | return 0; | ||
2768 | |||
2769 | return l_ptr; | ||
2770 | } | ||
2771 | |||
2772 | struct sk_buff *link_cmd_config(const void *req_tlv_area, int req_tlv_space, | ||
2773 | u16 cmd) | ||
2774 | { | ||
2775 | struct tipc_link_config *args; | ||
2776 | u32 new_value; | ||
2777 | struct link *l_ptr; | ||
2778 | struct node *node; | ||
2779 | int res; | ||
2780 | |||
2781 | if (!TLV_CHECK(req_tlv_area, req_tlv_space, TIPC_TLV_LINK_CONFIG)) | ||
2782 | return cfg_reply_error_string(TIPC_CFG_TLV_ERROR); | ||
2783 | |||
2784 | args = (struct tipc_link_config *)TLV_DATA(req_tlv_area); | ||
2785 | new_value = ntohl(args->value); | ||
2786 | |||
2787 | if (!strcmp(args->name, bc_link_name)) { | ||
2788 | if ((cmd == TIPC_CMD_SET_LINK_WINDOW) && | ||
2789 | (bclink_set_queue_limits(new_value) == 0)) | ||
2790 | return cfg_reply_none(); | ||
2791 | return cfg_reply_error_string(TIPC_CFG_NOT_SUPPORTED | ||
2792 | " (cannot change setting on broadcast link)"); | ||
2793 | } | ||
2794 | |||
2795 | read_lock_bh(&net_lock); | ||
2796 | l_ptr = link_find_link(args->name, &node); | ||
2797 | if (!l_ptr) { | ||
2798 | read_unlock_bh(&net_lock); | ||
2799 | return cfg_reply_error_string("link not found"); | ||
2800 | } | ||
2801 | |||
2802 | node_lock(node); | ||
2803 | res = -EINVAL; | ||
2804 | switch (cmd) { | ||
2805 | case TIPC_CMD_SET_LINK_TOL: | ||
2806 | if ((new_value >= TIPC_MIN_LINK_TOL) && | ||
2807 | (new_value <= TIPC_MAX_LINK_TOL)) { | ||
2808 | link_set_supervision_props(l_ptr, new_value); | ||
2809 | link_send_proto_msg(l_ptr, STATE_MSG, | ||
2810 | 0, 0, new_value, 0, 0); | ||
2811 | res = TIPC_OK; | ||
2812 | } | ||
2813 | break; | ||
2814 | case TIPC_CMD_SET_LINK_PRI: | ||
2815 | if (new_value < TIPC_NUM_LINK_PRI) { | ||
2816 | l_ptr->priority = new_value; | ||
2817 | link_send_proto_msg(l_ptr, STATE_MSG, | ||
2818 | 0, 0, 0, new_value, 0); | ||
2819 | res = TIPC_OK; | ||
2820 | } | ||
2821 | break; | ||
2822 | case TIPC_CMD_SET_LINK_WINDOW: | ||
2823 | if ((new_value >= TIPC_MIN_LINK_WIN) && | ||
2824 | (new_value <= TIPC_MAX_LINK_WIN)) { | ||
2825 | link_set_queue_limits(l_ptr, new_value); | ||
2826 | res = TIPC_OK; | ||
2827 | } | ||
2828 | break; | ||
2829 | } | ||
2830 | node_unlock(node); | ||
2831 | |||
2832 | read_unlock_bh(&net_lock); | ||
2833 | if (res) | ||
2834 | return cfg_reply_error_string("cannot change link setting"); | ||
2835 | |||
2836 | return cfg_reply_none(); | ||
2837 | } | ||
2838 | |||
2839 | /** | ||
2840 | * link_reset_statistics - reset link statistics | ||
2841 | * @l_ptr: pointer to link | ||
2842 | */ | ||
2843 | |||
2844 | static void link_reset_statistics(struct link *l_ptr) | ||
2845 | { | ||
2846 | memset(&l_ptr->stats, 0, sizeof(l_ptr->stats)); | ||
2847 | l_ptr->stats.sent_info = l_ptr->next_out_no; | ||
2848 | l_ptr->stats.recv_info = l_ptr->next_in_no; | ||
2849 | } | ||
2850 | |||
2851 | struct sk_buff *link_cmd_reset_stats(const void *req_tlv_area, int req_tlv_space) | ||
2852 | { | ||
2853 | char *link_name; | ||
2854 | struct link *l_ptr; | ||
2855 | struct node *node; | ||
2856 | |||
2857 | if (!TLV_CHECK(req_tlv_area, req_tlv_space, TIPC_TLV_LINK_NAME)) | ||
2858 | return cfg_reply_error_string(TIPC_CFG_TLV_ERROR); | ||
2859 | |||
2860 | link_name = (char *)TLV_DATA(req_tlv_area); | ||
2861 | if (!strcmp(link_name, bc_link_name)) { | ||
2862 | if (bclink_reset_stats()) | ||
2863 | return cfg_reply_error_string("link not found"); | ||
2864 | return cfg_reply_none(); | ||
2865 | } | ||
2866 | |||
2867 | read_lock_bh(&net_lock); | ||
2868 | l_ptr = link_find_link(link_name, &node); | ||
2869 | if (!l_ptr) { | ||
2870 | read_unlock_bh(&net_lock); | ||
2871 | return cfg_reply_error_string("link not found"); | ||
2872 | } | ||
2873 | |||
2874 | node_lock(node); | ||
2875 | link_reset_statistics(l_ptr); | ||
2876 | node_unlock(node); | ||
2877 | read_unlock_bh(&net_lock); | ||
2878 | return cfg_reply_none(); | ||
2879 | } | ||
2880 | |||
2881 | /** | ||
2882 | * percent - convert count to a percentage of total (rounding up or down) | ||
2883 | */ | ||
2884 | |||
2885 | static u32 percent(u32 count, u32 total) | ||
2886 | { | ||
2887 | return (count * 100 + (total / 2)) / total; | ||
2888 | } | ||
2889 | |||
2890 | /** | ||
2891 | * link_stats - print link statistics | ||
2892 | * @name: link name | ||
2893 | * @buf: print buffer area | ||
2894 | * @buf_size: size of print buffer area | ||
2895 | * | ||
2896 | * Returns length of print buffer data string (or 0 if error) | ||
2897 | */ | ||
2898 | |||
2899 | static int link_stats(const char *name, char *buf, const u32 buf_size) | ||
2900 | { | ||
2901 | struct print_buf pb; | ||
2902 | struct link *l_ptr; | ||
2903 | struct node *node; | ||
2904 | char *status; | ||
2905 | u32 profile_total = 0; | ||
2906 | |||
2907 | if (!strcmp(name, bc_link_name)) | ||
2908 | return bclink_stats(buf, buf_size); | ||
2909 | |||
2910 | printbuf_init(&pb, buf, buf_size); | ||
2911 | |||
2912 | read_lock_bh(&net_lock); | ||
2913 | l_ptr = link_find_link(name, &node); | ||
2914 | if (!l_ptr) { | ||
2915 | read_unlock_bh(&net_lock); | ||
2916 | return 0; | ||
2917 | } | ||
2918 | node_lock(node); | ||
2919 | |||
2920 | if (link_is_active(l_ptr)) | ||
2921 | status = "ACTIVE"; | ||
2922 | else if (link_is_up(l_ptr)) | ||
2923 | status = "STANDBY"; | ||
2924 | else | ||
2925 | status = "DEFUNCT"; | ||
2926 | tipc_printf(&pb, "Link <%s>\n" | ||
2927 | " %s MTU:%u Priority:%u Tolerance:%u ms" | ||
2928 | " Window:%u packets\n", | ||
2929 | l_ptr->name, status, link_max_pkt(l_ptr), | ||
2930 | l_ptr->priority, l_ptr->tolerance, l_ptr->queue_limit[0]); | ||
2931 | tipc_printf(&pb, " RX packets:%u fragments:%u/%u bundles:%u/%u\n", | ||
2932 | l_ptr->next_in_no - l_ptr->stats.recv_info, | ||
2933 | l_ptr->stats.recv_fragments, | ||
2934 | l_ptr->stats.recv_fragmented, | ||
2935 | l_ptr->stats.recv_bundles, | ||
2936 | l_ptr->stats.recv_bundled); | ||
2937 | tipc_printf(&pb, " TX packets:%u fragments:%u/%u bundles:%u/%u\n", | ||
2938 | l_ptr->next_out_no - l_ptr->stats.sent_info, | ||
2939 | l_ptr->stats.sent_fragments, | ||
2940 | l_ptr->stats.sent_fragmented, | ||
2941 | l_ptr->stats.sent_bundles, | ||
2942 | l_ptr->stats.sent_bundled); | ||
2943 | profile_total = l_ptr->stats.msg_length_counts; | ||
2944 | if (!profile_total) | ||
2945 | profile_total = 1; | ||
2946 | tipc_printf(&pb, " TX profile sample:%u packets average:%u octets\n" | ||
2947 | " 0-64:%u%% -256:%u%% -1024:%u%% -4096:%u%% " | ||
2948 | "-16354:%u%% -32768:%u%% -66000:%u%%\n", | ||
2949 | l_ptr->stats.msg_length_counts, | ||
2950 | l_ptr->stats.msg_lengths_total / profile_total, | ||
2951 | percent(l_ptr->stats.msg_length_profile[0], profile_total), | ||
2952 | percent(l_ptr->stats.msg_length_profile[1], profile_total), | ||
2953 | percent(l_ptr->stats.msg_length_profile[2], profile_total), | ||
2954 | percent(l_ptr->stats.msg_length_profile[3], profile_total), | ||
2955 | percent(l_ptr->stats.msg_length_profile[4], profile_total), | ||
2956 | percent(l_ptr->stats.msg_length_profile[5], profile_total), | ||
2957 | percent(l_ptr->stats.msg_length_profile[6], profile_total)); | ||
2958 | tipc_printf(&pb, " RX states:%u probes:%u naks:%u defs:%u dups:%u\n", | ||
2959 | l_ptr->stats.recv_states, | ||
2960 | l_ptr->stats.recv_probes, | ||
2961 | l_ptr->stats.recv_nacks, | ||
2962 | l_ptr->stats.deferred_recv, | ||
2963 | l_ptr->stats.duplicates); | ||
2964 | tipc_printf(&pb, " TX states:%u probes:%u naks:%u acks:%u dups:%u\n", | ||
2965 | l_ptr->stats.sent_states, | ||
2966 | l_ptr->stats.sent_probes, | ||
2967 | l_ptr->stats.sent_nacks, | ||
2968 | l_ptr->stats.sent_acks, | ||
2969 | l_ptr->stats.retransmitted); | ||
2970 | tipc_printf(&pb, " Congestion bearer:%u link:%u Send queue max:%u avg:%u\n", | ||
2971 | l_ptr->stats.bearer_congs, | ||
2972 | l_ptr->stats.link_congs, | ||
2973 | l_ptr->stats.max_queue_sz, | ||
2974 | l_ptr->stats.queue_sz_counts | ||
2975 | ? (l_ptr->stats.accu_queue_sz / l_ptr->stats.queue_sz_counts) | ||
2976 | : 0); | ||
2977 | |||
2978 | node_unlock(node); | ||
2979 | read_unlock_bh(&net_lock); | ||
2980 | return printbuf_validate(&pb); | ||
2981 | } | ||
2982 | |||
2983 | #define MAX_LINK_STATS_INFO 2000 | ||
2984 | |||
2985 | struct sk_buff *link_cmd_show_stats(const void *req_tlv_area, int req_tlv_space) | ||
2986 | { | ||
2987 | struct sk_buff *buf; | ||
2988 | struct tlv_desc *rep_tlv; | ||
2989 | int str_len; | ||
2990 | |||
2991 | if (!TLV_CHECK(req_tlv_area, req_tlv_space, TIPC_TLV_LINK_NAME)) | ||
2992 | return cfg_reply_error_string(TIPC_CFG_TLV_ERROR); | ||
2993 | |||
2994 | buf = cfg_reply_alloc(TLV_SPACE(MAX_LINK_STATS_INFO)); | ||
2995 | if (!buf) | ||
2996 | return NULL; | ||
2997 | |||
2998 | rep_tlv = (struct tlv_desc *)buf->data; | ||
2999 | |||
3000 | str_len = link_stats((char *)TLV_DATA(req_tlv_area), | ||
3001 | (char *)TLV_DATA(rep_tlv), MAX_LINK_STATS_INFO); | ||
3002 | if (!str_len) { | ||
3003 | buf_discard(buf); | ||
3004 | return cfg_reply_error_string("link not found"); | ||
3005 | } | ||
3006 | |||
3007 | skb_put(buf, TLV_SPACE(str_len)); | ||
3008 | TLV_SET(rep_tlv, TIPC_TLV_ULTRA_STRING, NULL, str_len); | ||
3009 | |||
3010 | return buf; | ||
3011 | } | ||
3012 | |||
3013 | #if 0 | ||
3014 | int link_control(const char *name, u32 op, u32 val) | ||
3015 | { | ||
3016 | int res = -EINVAL; | ||
3017 | struct link *l_ptr; | ||
3018 | u32 bearer_id; | ||
3019 | struct node * node; | ||
3020 | u32 a; | ||
3021 | |||
3022 | a = link_name2addr(name, &bearer_id); | ||
3023 | read_lock_bh(&net_lock); | ||
3024 | node = node_find(a); | ||
3025 | if (node) { | ||
3026 | node_lock(node); | ||
3027 | l_ptr = node->links[bearer_id]; | ||
3028 | if (l_ptr) { | ||
3029 | if (op == TIPC_REMOVE_LINK) { | ||
3030 | struct bearer *b_ptr = l_ptr->b_ptr; | ||
3031 | spin_lock_bh(&b_ptr->publ.lock); | ||
3032 | link_delete(l_ptr); | ||
3033 | spin_unlock_bh(&b_ptr->publ.lock); | ||
3034 | } | ||
3035 | if (op == TIPC_CMD_BLOCK_LINK) { | ||
3036 | link_reset(l_ptr); | ||
3037 | l_ptr->blocked = 1; | ||
3038 | } | ||
3039 | if (op == TIPC_CMD_UNBLOCK_LINK) { | ||
3040 | l_ptr->blocked = 0; | ||
3041 | } | ||
3042 | res = TIPC_OK; | ||
3043 | } | ||
3044 | node_unlock(node); | ||
3045 | } | ||
3046 | read_unlock_bh(&net_lock); | ||
3047 | return res; | ||
3048 | } | ||
3049 | #endif | ||
3050 | |||
3051 | /** | ||
3052 | * link_get_max_pkt - get maximum packet size to use when sending to destination | ||
3053 | * @dest: network address of destination node | ||
3054 | * @selector: used to select from set of active links | ||
3055 | * | ||
3056 | * If no active link can be found, uses default maximum packet size. | ||
3057 | */ | ||
3058 | |||
3059 | u32 link_get_max_pkt(u32 dest, u32 selector) | ||
3060 | { | ||
3061 | struct node *n_ptr; | ||
3062 | struct link *l_ptr; | ||
3063 | u32 res = MAX_PKT_DEFAULT; | ||
3064 | |||
3065 | if (dest == tipc_own_addr) | ||
3066 | return MAX_MSG_SIZE; | ||
3067 | |||
3068 | read_lock_bh(&net_lock); | ||
3069 | n_ptr = node_select(dest, selector); | ||
3070 | if (n_ptr) { | ||
3071 | node_lock(n_ptr); | ||
3072 | l_ptr = n_ptr->active_links[selector & 1]; | ||
3073 | if (l_ptr) | ||
3074 | res = link_max_pkt(l_ptr); | ||
3075 | node_unlock(n_ptr); | ||
3076 | } | ||
3077 | read_unlock_bh(&net_lock); | ||
3078 | return res; | ||
3079 | } | ||
3080 | |||
3081 | #if 0 | ||
3082 | static void link_dump_rec_queue(struct link *l_ptr) | ||
3083 | { | ||
3084 | struct sk_buff *crs; | ||
3085 | |||
3086 | if (!l_ptr->oldest_deferred_in) { | ||
3087 | info("Reception queue empty\n"); | ||
3088 | return; | ||
3089 | } | ||
3090 | info("Contents of Reception queue:\n"); | ||
3091 | crs = l_ptr->oldest_deferred_in; | ||
3092 | while (crs) { | ||
3093 | if (crs->data == (void *)0x0000a3a3) { | ||
3094 | info("buffer %x invalid\n", crs); | ||
3095 | return; | ||
3096 | } | ||
3097 | msg_dbg(buf_msg(crs), "In rec queue: \n"); | ||
3098 | crs = crs->next; | ||
3099 | } | ||
3100 | } | ||
3101 | #endif | ||
3102 | |||
3103 | static void link_dump_send_queue(struct link *l_ptr) | ||
3104 | { | ||
3105 | if (l_ptr->next_out) { | ||
3106 | info("\nContents of unsent queue:\n"); | ||
3107 | dbg_print_buf_chain(l_ptr->next_out); | ||
3108 | } | ||
3109 | info("\nContents of send queue:\n"); | ||
3110 | if (l_ptr->first_out) { | ||
3111 | dbg_print_buf_chain(l_ptr->first_out); | ||
3112 | } | ||
3113 | info("Empty send queue\n"); | ||
3114 | } | ||
3115 | |||
3116 | static void link_print(struct link *l_ptr, struct print_buf *buf, | ||
3117 | const char *str) | ||
3118 | { | ||
3119 | tipc_printf(buf, str); | ||
3120 | if (link_reset_reset(l_ptr) || link_reset_unknown(l_ptr)) | ||
3121 | return; | ||
3122 | tipc_printf(buf, "Link %x<%s>:", | ||
3123 | l_ptr->addr, l_ptr->b_ptr->publ.name); | ||
3124 | tipc_printf(buf, ": NXO(%u):", mod(l_ptr->next_out_no)); | ||
3125 | tipc_printf(buf, "NXI(%u):", mod(l_ptr->next_in_no)); | ||
3126 | tipc_printf(buf, "SQUE"); | ||
3127 | if (l_ptr->first_out) { | ||
3128 | tipc_printf(buf, "[%u..", msg_seqno(buf_msg(l_ptr->first_out))); | ||
3129 | if (l_ptr->next_out) | ||
3130 | tipc_printf(buf, "%u..", | ||
3131 | msg_seqno(buf_msg(l_ptr->next_out))); | ||
3132 | tipc_printf(buf, "%u]", | ||
3133 | msg_seqno(buf_msg | ||
3134 | (l_ptr->last_out)), l_ptr->out_queue_size); | ||
3135 | if ((mod(msg_seqno(buf_msg(l_ptr->last_out)) - | ||
3136 | msg_seqno(buf_msg(l_ptr->first_out))) | ||
3137 | != (l_ptr->out_queue_size - 1)) | ||
3138 | || (l_ptr->last_out->next != 0)) { | ||
3139 | tipc_printf(buf, "\nSend queue inconsistency\n"); | ||
3140 | tipc_printf(buf, "first_out= %x ", l_ptr->first_out); | ||
3141 | tipc_printf(buf, "next_out= %x ", l_ptr->next_out); | ||
3142 | tipc_printf(buf, "last_out= %x ", l_ptr->last_out); | ||
3143 | link_dump_send_queue(l_ptr); | ||
3144 | } | ||
3145 | } else | ||
3146 | tipc_printf(buf, "[]"); | ||
3147 | tipc_printf(buf, "SQSIZ(%u)", l_ptr->out_queue_size); | ||
3148 | if (l_ptr->oldest_deferred_in) { | ||
3149 | u32 o = msg_seqno(buf_msg(l_ptr->oldest_deferred_in)); | ||
3150 | u32 n = msg_seqno(buf_msg(l_ptr->newest_deferred_in)); | ||
3151 | tipc_printf(buf, ":RQUE[%u..%u]", o, n); | ||
3152 | if (l_ptr->deferred_inqueue_sz != mod((n + 1) - o)) { | ||
3153 | tipc_printf(buf, ":RQSIZ(%u)", | ||
3154 | l_ptr->deferred_inqueue_sz); | ||
3155 | } | ||
3156 | } | ||
3157 | if (link_working_unknown(l_ptr)) | ||
3158 | tipc_printf(buf, ":WU"); | ||
3159 | if (link_reset_reset(l_ptr)) | ||
3160 | tipc_printf(buf, ":RR"); | ||
3161 | if (link_reset_unknown(l_ptr)) | ||
3162 | tipc_printf(buf, ":RU"); | ||
3163 | if (link_working_working(l_ptr)) | ||
3164 | tipc_printf(buf, ":WW"); | ||
3165 | tipc_printf(buf, "\n"); | ||
3166 | } | ||
3167 | |||