aboutsummaryrefslogtreecommitdiffstats
path: root/include/net
diff options
context:
space:
mode:
authorDave Jones <davej@redhat.com>2005-08-18 01:56:07 -0400
committerDave Jones <davej@redhat.com>2005-08-18 01:56:07 -0400
commita8b3e6f10f08f66ae1072efd087b30966a3654f6 (patch)
tree1d1409855f8ad5beabafe061c6453edd84ba94c8 /include/net
parent46acac3b4fd8ef66eec63b51de8d556a17c7d4f7 (diff)
parent099d44e869f1886b5eb02a5145ca97b5e4142e28 (diff)
Merge /pub/scm/linux/kernel/git/torvalds/linux-2.6
Diffstat (limited to 'include/net')
-rw-r--r--include/net/ax25.h2
-rw-r--r--include/net/bluetooth/bluetooth.h8
-rw-r--r--include/net/ieee80211.h856
-rw-r--r--include/net/ip.h1
-rw-r--r--include/net/ip6_fib.h9
-rw-r--r--include/net/ip6_route.h9
-rw-r--r--include/net/ip_fib.h14
-rw-r--r--include/net/ipv6.h1
-rw-r--r--include/net/irda/irda_device.h2
-rw-r--r--include/net/neighbour.h7
-rw-r--r--include/net/pkt_sched.h17
-rw-r--r--include/net/request_sock.h255
-rw-r--r--include/net/route.h2
-rw-r--r--include/net/sch_generic.h135
-rw-r--r--include/net/sctp/command.h8
-rw-r--r--include/net/sctp/constants.h25
-rw-r--r--include/net/sctp/sctp.h27
-rw-r--r--include/net/sctp/sm.h17
-rw-r--r--include/net/sctp/structs.h93
-rw-r--r--include/net/sctp/ulpevent.h16
-rw-r--r--include/net/sctp/ulpqueue.h11
-rw-r--r--include/net/sctp/user.h3
-rw-r--r--include/net/slhc_vj.h21
-rw-r--r--include/net/snmp.h14
-rw-r--r--include/net/sock.h29
-rw-r--r--include/net/tcp.h562
-rw-r--r--include/net/tcp_ecn.h13
-rw-r--r--include/net/x25.h9
-rw-r--r--include/net/x25device.h3
-rw-r--r--include/net/xfrm.h30
30 files changed, 1599 insertions, 600 deletions
diff --git a/include/net/ax25.h b/include/net/ax25.h
index 9e6368a54547..828a3a93dda1 100644
--- a/include/net/ax25.h
+++ b/include/net/ax25.h
@@ -220,7 +220,7 @@ static __inline__ void ax25_cb_put(ax25_cb *ax25)
220 } 220 }
221} 221}
222 222
223static inline unsigned short ax25_type_trans(struct sk_buff *skb, struct net_device *dev) 223static inline __be16 ax25_type_trans(struct sk_buff *skb, struct net_device *dev)
224{ 224{
225 skb->dev = dev; 225 skb->dev = dev;
226 skb->pkt_type = PACKET_HOST; 226 skb->pkt_type = PACKET_HOST;
diff --git a/include/net/bluetooth/bluetooth.h b/include/net/bluetooth/bluetooth.h
index 42a84c53678b..06b24f637026 100644
--- a/include/net/bluetooth/bluetooth.h
+++ b/include/net/bluetooth/bluetooth.h
@@ -57,12 +57,6 @@
57#define BT_DBG(fmt, arg...) printk(KERN_INFO "%s: " fmt "\n" , __FUNCTION__ , ## arg) 57#define BT_DBG(fmt, arg...) printk(KERN_INFO "%s: " fmt "\n" , __FUNCTION__ , ## arg)
58#define BT_ERR(fmt, arg...) printk(KERN_ERR "%s: " fmt "\n" , __FUNCTION__ , ## arg) 58#define BT_ERR(fmt, arg...) printk(KERN_ERR "%s: " fmt "\n" , __FUNCTION__ , ## arg)
59 59
60#ifdef HCI_DATA_DUMP
61#define BT_DMP(buf, len) bt_dump(__FUNCTION__, buf, len)
62#else
63#define BT_DMP(D...)
64#endif
65
66extern struct proc_dir_entry *proc_bt; 60extern struct proc_dir_entry *proc_bt;
67 61
68/* Connection and socket states */ 62/* Connection and socket states */
@@ -174,8 +168,6 @@ static inline int skb_frags_no(struct sk_buff *skb)
174 return n; 168 return n;
175} 169}
176 170
177void bt_dump(char *pref, __u8 *buf, int count);
178
179int bt_err(__u16 code); 171int bt_err(__u16 code);
180 172
181#endif /* __BLUETOOTH_H */ 173#endif /* __BLUETOOTH_H */
diff --git a/include/net/ieee80211.h b/include/net/ieee80211.h
new file mode 100644
index 000000000000..db09580ad14b
--- /dev/null
+++ b/include/net/ieee80211.h
@@ -0,0 +1,856 @@
1/*
2 * Merged with mainline ieee80211.h in Aug 2004. Original ieee802_11
3 * remains copyright by the original authors
4 *
5 * Portions of the merged code are based on Host AP (software wireless
6 * LAN access point) driver for Intersil Prism2/2.5/3.
7 *
8 * Copyright (c) 2001-2002, SSH Communications Security Corp and Jouni Malinen
9 * <jkmaline@cc.hut.fi>
10 * Copyright (c) 2002-2003, Jouni Malinen <jkmaline@cc.hut.fi>
11 *
12 * Adaption to a generic IEEE 802.11 stack by James Ketrenos
13 * <jketreno@linux.intel.com>
14 * Copyright (c) 2004, Intel Corporation
15 *
16 * This program is free software; you can redistribute it and/or modify
17 * it under the terms of the GNU General Public License version 2 as
18 * published by the Free Software Foundation. See README and COPYING for
19 * more details.
20 */
21#ifndef IEEE80211_H
22#define IEEE80211_H
23
24#include <linux/if_ether.h> /* ETH_ALEN */
25#include <linux/kernel.h> /* ARRAY_SIZE */
26
27#if WIRELESS_EXT < 17
28#define IW_QUAL_QUAL_INVALID 0x10
29#define IW_QUAL_LEVEL_INVALID 0x20
30#define IW_QUAL_NOISE_INVALID 0x40
31#define IW_QUAL_QUAL_UPDATED 0x1
32#define IW_QUAL_LEVEL_UPDATED 0x2
33#define IW_QUAL_NOISE_UPDATED 0x4
34#endif
35
36#define IEEE80211_DATA_LEN 2304
37/* Maximum size for the MA-UNITDATA primitive, 802.11 standard section
38 6.2.1.1.2.
39
40 The figure in section 7.1.2 suggests a body size of up to 2312
41 bytes is allowed, which is a bit confusing, I suspect this
42 represents the 2304 bytes of real data, plus a possible 8 bytes of
43 WEP IV and ICV. (this interpretation suggested by Ramiro Barreiro) */
44
45
46#define IEEE80211_HLEN 30
47#define IEEE80211_FRAME_LEN (IEEE80211_DATA_LEN + IEEE80211_HLEN)
48
49struct ieee80211_hdr {
50 u16 frame_ctl;
51 u16 duration_id;
52 u8 addr1[ETH_ALEN];
53 u8 addr2[ETH_ALEN];
54 u8 addr3[ETH_ALEN];
55 u16 seq_ctl;
56 u8 addr4[ETH_ALEN];
57} __attribute__ ((packed));
58
59struct ieee80211_hdr_3addr {
60 u16 frame_ctl;
61 u16 duration_id;
62 u8 addr1[ETH_ALEN];
63 u8 addr2[ETH_ALEN];
64 u8 addr3[ETH_ALEN];
65 u16 seq_ctl;
66} __attribute__ ((packed));
67
68enum eap_type {
69 EAP_PACKET = 0,
70 EAPOL_START,
71 EAPOL_LOGOFF,
72 EAPOL_KEY,
73 EAPOL_ENCAP_ASF_ALERT
74};
75
76static const char *eap_types[] = {
77 [EAP_PACKET] = "EAP-Packet",
78 [EAPOL_START] = "EAPOL-Start",
79 [EAPOL_LOGOFF] = "EAPOL-Logoff",
80 [EAPOL_KEY] = "EAPOL-Key",
81 [EAPOL_ENCAP_ASF_ALERT] = "EAPOL-Encap-ASF-Alert"
82};
83
84static inline const char *eap_get_type(int type)
85{
86 return (type >= ARRAY_SIZE(eap_types)) ? "Unknown" : eap_types[type];
87}
88
89struct eapol {
90 u8 snap[6];
91 u16 ethertype;
92 u8 version;
93 u8 type;
94 u16 length;
95} __attribute__ ((packed));
96
97#define IEEE80211_1ADDR_LEN 10
98#define IEEE80211_2ADDR_LEN 16
99#define IEEE80211_3ADDR_LEN 24
100#define IEEE80211_4ADDR_LEN 30
101#define IEEE80211_FCS_LEN 4
102
103#define MIN_FRAG_THRESHOLD 256U
104#define MAX_FRAG_THRESHOLD 2346U
105
106/* Frame control field constants */
107#define IEEE80211_FCTL_VERS 0x0002
108#define IEEE80211_FCTL_FTYPE 0x000c
109#define IEEE80211_FCTL_STYPE 0x00f0
110#define IEEE80211_FCTL_TODS 0x0100
111#define IEEE80211_FCTL_FROMDS 0x0200
112#define IEEE80211_FCTL_MOREFRAGS 0x0400
113#define IEEE80211_FCTL_RETRY 0x0800
114#define IEEE80211_FCTL_PM 0x1000
115#define IEEE80211_FCTL_MOREDATA 0x2000
116#define IEEE80211_FCTL_WEP 0x4000
117#define IEEE80211_FCTL_ORDER 0x8000
118
119#define IEEE80211_FTYPE_MGMT 0x0000
120#define IEEE80211_FTYPE_CTL 0x0004
121#define IEEE80211_FTYPE_DATA 0x0008
122
123/* management */
124#define IEEE80211_STYPE_ASSOC_REQ 0x0000
125#define IEEE80211_STYPE_ASSOC_RESP 0x0010
126#define IEEE80211_STYPE_REASSOC_REQ 0x0020
127#define IEEE80211_STYPE_REASSOC_RESP 0x0030
128#define IEEE80211_STYPE_PROBE_REQ 0x0040
129#define IEEE80211_STYPE_PROBE_RESP 0x0050
130#define IEEE80211_STYPE_BEACON 0x0080
131#define IEEE80211_STYPE_ATIM 0x0090
132#define IEEE80211_STYPE_DISASSOC 0x00A0
133#define IEEE80211_STYPE_AUTH 0x00B0
134#define IEEE80211_STYPE_DEAUTH 0x00C0
135
136/* control */
137#define IEEE80211_STYPE_PSPOLL 0x00A0
138#define IEEE80211_STYPE_RTS 0x00B0
139#define IEEE80211_STYPE_CTS 0x00C0
140#define IEEE80211_STYPE_ACK 0x00D0
141#define IEEE80211_STYPE_CFEND 0x00E0
142#define IEEE80211_STYPE_CFENDACK 0x00F0
143
144/* data */
145#define IEEE80211_STYPE_DATA 0x0000
146#define IEEE80211_STYPE_DATA_CFACK 0x0010
147#define IEEE80211_STYPE_DATA_CFPOLL 0x0020
148#define IEEE80211_STYPE_DATA_CFACKPOLL 0x0030
149#define IEEE80211_STYPE_NULLFUNC 0x0040
150#define IEEE80211_STYPE_CFACK 0x0050
151#define IEEE80211_STYPE_CFPOLL 0x0060
152#define IEEE80211_STYPE_CFACKPOLL 0x0070
153
154#define IEEE80211_SCTL_FRAG 0x000F
155#define IEEE80211_SCTL_SEQ 0xFFF0
156
157
158/* debug macros */
159
160#ifdef CONFIG_IEEE80211_DEBUG
161extern u32 ieee80211_debug_level;
162#define IEEE80211_DEBUG(level, fmt, args...) \
163do { if (ieee80211_debug_level & (level)) \
164 printk(KERN_DEBUG "ieee80211: %c %s " fmt, \
165 in_interrupt() ? 'I' : 'U', __FUNCTION__ , ## args); } while (0)
166#else
167#define IEEE80211_DEBUG(level, fmt, args...) do {} while (0)
168#endif /* CONFIG_IEEE80211_DEBUG */
169
170/*
171 * To use the debug system;
172 *
173 * If you are defining a new debug classification, simply add it to the #define
174 * list here in the form of:
175 *
176 * #define IEEE80211_DL_xxxx VALUE
177 *
178 * shifting value to the left one bit from the previous entry. xxxx should be
179 * the name of the classification (for example, WEP)
180 *
181 * You then need to either add a IEEE80211_xxxx_DEBUG() macro definition for your
182 * classification, or use IEEE80211_DEBUG(IEEE80211_DL_xxxx, ...) whenever you want
183 * to send output to that classification.
184 *
185 * To add your debug level to the list of levels seen when you perform
186 *
187 * % cat /proc/net/ipw/debug_level
188 *
189 * you simply need to add your entry to the ipw_debug_levels array.
190 *
191 * If you do not see debug_level in /proc/net/ipw then you do not have
192 * CONFIG_IEEE80211_DEBUG defined in your kernel configuration
193 *
194 */
195
196#define IEEE80211_DL_INFO (1<<0)
197#define IEEE80211_DL_WX (1<<1)
198#define IEEE80211_DL_SCAN (1<<2)
199#define IEEE80211_DL_STATE (1<<3)
200#define IEEE80211_DL_MGMT (1<<4)
201#define IEEE80211_DL_FRAG (1<<5)
202#define IEEE80211_DL_EAP (1<<6)
203#define IEEE80211_DL_DROP (1<<7)
204
205#define IEEE80211_DL_TX (1<<8)
206#define IEEE80211_DL_RX (1<<9)
207
208#define IEEE80211_ERROR(f, a...) printk(KERN_ERR "ieee80211: " f, ## a)
209#define IEEE80211_WARNING(f, a...) printk(KERN_WARNING "ieee80211: " f, ## a)
210#define IEEE80211_DEBUG_INFO(f, a...) IEEE80211_DEBUG(IEEE80211_DL_INFO, f, ## a)
211
212#define IEEE80211_DEBUG_WX(f, a...) IEEE80211_DEBUG(IEEE80211_DL_WX, f, ## a)
213#define IEEE80211_DEBUG_SCAN(f, a...) IEEE80211_DEBUG(IEEE80211_DL_SCAN, f, ## a)
214#define IEEE80211_DEBUG_STATE(f, a...) IEEE80211_DEBUG(IEEE80211_DL_STATE, f, ## a)
215#define IEEE80211_DEBUG_MGMT(f, a...) IEEE80211_DEBUG(IEEE80211_DL_MGMT, f, ## a)
216#define IEEE80211_DEBUG_FRAG(f, a...) IEEE80211_DEBUG(IEEE80211_DL_FRAG, f, ## a)
217#define IEEE80211_DEBUG_EAP(f, a...) IEEE80211_DEBUG(IEEE80211_DL_EAP, f, ## a)
218#define IEEE80211_DEBUG_DROP(f, a...) IEEE80211_DEBUG(IEEE80211_DL_DROP, f, ## a)
219#define IEEE80211_DEBUG_TX(f, a...) IEEE80211_DEBUG(IEEE80211_DL_TX, f, ## a)
220#define IEEE80211_DEBUG_RX(f, a...) IEEE80211_DEBUG(IEEE80211_DL_RX, f, ## a)
221#include <linux/netdevice.h>
222#include <linux/wireless.h>
223#include <linux/if_arp.h> /* ARPHRD_ETHER */
224
225#ifndef WIRELESS_SPY
226#define WIRELESS_SPY // enable iwspy support
227#endif
228#include <net/iw_handler.h> // new driver API
229
230#ifndef ETH_P_PAE
231#define ETH_P_PAE 0x888E /* Port Access Entity (IEEE 802.1X) */
232#endif /* ETH_P_PAE */
233
234#define ETH_P_PREAUTH 0x88C7 /* IEEE 802.11i pre-authentication */
235
236#ifndef ETH_P_80211_RAW
237#define ETH_P_80211_RAW (ETH_P_ECONET + 1)
238#endif
239
240/* IEEE 802.11 defines */
241
242#define P80211_OUI_LEN 3
243
244struct ieee80211_snap_hdr {
245
246 u8 dsap; /* always 0xAA */
247 u8 ssap; /* always 0xAA */
248 u8 ctrl; /* always 0x03 */
249 u8 oui[P80211_OUI_LEN]; /* organizational universal id */
250
251} __attribute__ ((packed));
252
253#define SNAP_SIZE sizeof(struct ieee80211_snap_hdr)
254
255#define WLAN_FC_GET_TYPE(fc) ((fc) & IEEE80211_FCTL_FTYPE)
256#define WLAN_FC_GET_STYPE(fc) ((fc) & IEEE80211_FCTL_STYPE)
257
258#define WLAN_GET_SEQ_FRAG(seq) ((seq) & IEEE80211_SCTL_FRAG)
259#define WLAN_GET_SEQ_SEQ(seq) ((seq) & IEEE80211_SCTL_SEQ)
260
261/* Authentication algorithms */
262#define WLAN_AUTH_OPEN 0
263#define WLAN_AUTH_SHARED_KEY 1
264
265#define WLAN_AUTH_CHALLENGE_LEN 128
266
267#define WLAN_CAPABILITY_BSS (1<<0)
268#define WLAN_CAPABILITY_IBSS (1<<1)
269#define WLAN_CAPABILITY_CF_POLLABLE (1<<2)
270#define WLAN_CAPABILITY_CF_POLL_REQUEST (1<<3)
271#define WLAN_CAPABILITY_PRIVACY (1<<4)
272#define WLAN_CAPABILITY_SHORT_PREAMBLE (1<<5)
273#define WLAN_CAPABILITY_PBCC (1<<6)
274#define WLAN_CAPABILITY_CHANNEL_AGILITY (1<<7)
275
276/* Status codes */
277#define WLAN_STATUS_SUCCESS 0
278#define WLAN_STATUS_UNSPECIFIED_FAILURE 1
279#define WLAN_STATUS_CAPS_UNSUPPORTED 10
280#define WLAN_STATUS_REASSOC_NO_ASSOC 11
281#define WLAN_STATUS_ASSOC_DENIED_UNSPEC 12
282#define WLAN_STATUS_NOT_SUPPORTED_AUTH_ALG 13
283#define WLAN_STATUS_UNKNOWN_AUTH_TRANSACTION 14
284#define WLAN_STATUS_CHALLENGE_FAIL 15
285#define WLAN_STATUS_AUTH_TIMEOUT 16
286#define WLAN_STATUS_AP_UNABLE_TO_HANDLE_NEW_STA 17
287#define WLAN_STATUS_ASSOC_DENIED_RATES 18
288/* 802.11b */
289#define WLAN_STATUS_ASSOC_DENIED_NOSHORT 19
290#define WLAN_STATUS_ASSOC_DENIED_NOPBCC 20
291#define WLAN_STATUS_ASSOC_DENIED_NOAGILITY 21
292
293/* Reason codes */
294#define WLAN_REASON_UNSPECIFIED 1
295#define WLAN_REASON_PREV_AUTH_NOT_VALID 2
296#define WLAN_REASON_DEAUTH_LEAVING 3
297#define WLAN_REASON_DISASSOC_DUE_TO_INACTIVITY 4
298#define WLAN_REASON_DISASSOC_AP_BUSY 5
299#define WLAN_REASON_CLASS2_FRAME_FROM_NONAUTH_STA 6
300#define WLAN_REASON_CLASS3_FRAME_FROM_NONASSOC_STA 7
301#define WLAN_REASON_DISASSOC_STA_HAS_LEFT 8
302#define WLAN_REASON_STA_REQ_ASSOC_WITHOUT_AUTH 9
303
304
305#define IEEE80211_STATMASK_SIGNAL (1<<0)
306#define IEEE80211_STATMASK_RSSI (1<<1)
307#define IEEE80211_STATMASK_NOISE (1<<2)
308#define IEEE80211_STATMASK_RATE (1<<3)
309#define IEEE80211_STATMASK_WEMASK 0x7
310
311
312#define IEEE80211_CCK_MODULATION (1<<0)
313#define IEEE80211_OFDM_MODULATION (1<<1)
314
315#define IEEE80211_24GHZ_BAND (1<<0)
316#define IEEE80211_52GHZ_BAND (1<<1)
317
318#define IEEE80211_CCK_RATE_1MB 0x02
319#define IEEE80211_CCK_RATE_2MB 0x04
320#define IEEE80211_CCK_RATE_5MB 0x0B
321#define IEEE80211_CCK_RATE_11MB 0x16
322#define IEEE80211_OFDM_RATE_6MB 0x0C
323#define IEEE80211_OFDM_RATE_9MB 0x12
324#define IEEE80211_OFDM_RATE_12MB 0x18
325#define IEEE80211_OFDM_RATE_18MB 0x24
326#define IEEE80211_OFDM_RATE_24MB 0x30
327#define IEEE80211_OFDM_RATE_36MB 0x48
328#define IEEE80211_OFDM_RATE_48MB 0x60
329#define IEEE80211_OFDM_RATE_54MB 0x6C
330#define IEEE80211_BASIC_RATE_MASK 0x80
331
332#define IEEE80211_CCK_RATE_1MB_MASK (1<<0)
333#define IEEE80211_CCK_RATE_2MB_MASK (1<<1)
334#define IEEE80211_CCK_RATE_5MB_MASK (1<<2)
335#define IEEE80211_CCK_RATE_11MB_MASK (1<<3)
336#define IEEE80211_OFDM_RATE_6MB_MASK (1<<4)
337#define IEEE80211_OFDM_RATE_9MB_MASK (1<<5)
338#define IEEE80211_OFDM_RATE_12MB_MASK (1<<6)
339#define IEEE80211_OFDM_RATE_18MB_MASK (1<<7)
340#define IEEE80211_OFDM_RATE_24MB_MASK (1<<8)
341#define IEEE80211_OFDM_RATE_36MB_MASK (1<<9)
342#define IEEE80211_OFDM_RATE_48MB_MASK (1<<10)
343#define IEEE80211_OFDM_RATE_54MB_MASK (1<<11)
344
345#define IEEE80211_CCK_RATES_MASK 0x0000000F
346#define IEEE80211_CCK_BASIC_RATES_MASK (IEEE80211_CCK_RATE_1MB_MASK | \
347 IEEE80211_CCK_RATE_2MB_MASK)
348#define IEEE80211_CCK_DEFAULT_RATES_MASK (IEEE80211_CCK_BASIC_RATES_MASK | \
349 IEEE80211_CCK_RATE_5MB_MASK | \
350 IEEE80211_CCK_RATE_11MB_MASK)
351
352#define IEEE80211_OFDM_RATES_MASK 0x00000FF0
353#define IEEE80211_OFDM_BASIC_RATES_MASK (IEEE80211_OFDM_RATE_6MB_MASK | \
354 IEEE80211_OFDM_RATE_12MB_MASK | \
355 IEEE80211_OFDM_RATE_24MB_MASK)
356#define IEEE80211_OFDM_DEFAULT_RATES_MASK (IEEE80211_OFDM_BASIC_RATES_MASK | \
357 IEEE80211_OFDM_RATE_9MB_MASK | \
358 IEEE80211_OFDM_RATE_18MB_MASK | \
359 IEEE80211_OFDM_RATE_36MB_MASK | \
360 IEEE80211_OFDM_RATE_48MB_MASK | \
361 IEEE80211_OFDM_RATE_54MB_MASK)
362#define IEEE80211_DEFAULT_RATES_MASK (IEEE80211_OFDM_DEFAULT_RATES_MASK | \
363 IEEE80211_CCK_DEFAULT_RATES_MASK)
364
365#define IEEE80211_NUM_OFDM_RATES 8
366#define IEEE80211_NUM_CCK_RATES 4
367#define IEEE80211_OFDM_SHIFT_MASK_A 4
368
369
370
371
372/* NOTE: This data is for statistical purposes; not all hardware provides this
373 * information for frames received. Not setting these will not cause
374 * any adverse affects. */
375struct ieee80211_rx_stats {
376 u32 mac_time;
377 s8 rssi;
378 u8 signal;
379 u8 noise;
380 u16 rate; /* in 100 kbps */
381 u8 received_channel;
382 u8 control;
383 u8 mask;
384 u8 freq;
385 u16 len;
386};
387
388/* IEEE 802.11 requires that STA supports concurrent reception of at least
389 * three fragmented frames. This define can be increased to support more
390 * concurrent frames, but it should be noted that each entry can consume about
391 * 2 kB of RAM and increasing cache size will slow down frame reassembly. */
392#define IEEE80211_FRAG_CACHE_LEN 4
393
394struct ieee80211_frag_entry {
395 unsigned long first_frag_time;
396 unsigned int seq;
397 unsigned int last_frag;
398 struct sk_buff *skb;
399 u8 src_addr[ETH_ALEN];
400 u8 dst_addr[ETH_ALEN];
401};
402
403struct ieee80211_stats {
404 unsigned int tx_unicast_frames;
405 unsigned int tx_multicast_frames;
406 unsigned int tx_fragments;
407 unsigned int tx_unicast_octets;
408 unsigned int tx_multicast_octets;
409 unsigned int tx_deferred_transmissions;
410 unsigned int tx_single_retry_frames;
411 unsigned int tx_multiple_retry_frames;
412 unsigned int tx_retry_limit_exceeded;
413 unsigned int tx_discards;
414 unsigned int rx_unicast_frames;
415 unsigned int rx_multicast_frames;
416 unsigned int rx_fragments;
417 unsigned int rx_unicast_octets;
418 unsigned int rx_multicast_octets;
419 unsigned int rx_fcs_errors;
420 unsigned int rx_discards_no_buffer;
421 unsigned int tx_discards_wrong_sa;
422 unsigned int rx_discards_undecryptable;
423 unsigned int rx_message_in_msg_fragments;
424 unsigned int rx_message_in_bad_msg_fragments;
425};
426
427struct ieee80211_device;
428
429#if 0 /* for later */
430#include "ieee80211_crypt.h"
431#endif
432
433#define SEC_KEY_1 (1<<0)
434#define SEC_KEY_2 (1<<1)
435#define SEC_KEY_3 (1<<2)
436#define SEC_KEY_4 (1<<3)
437#define SEC_ACTIVE_KEY (1<<4)
438#define SEC_AUTH_MODE (1<<5)
439#define SEC_UNICAST_GROUP (1<<6)
440#define SEC_LEVEL (1<<7)
441#define SEC_ENABLED (1<<8)
442
443#define SEC_LEVEL_0 0 /* None */
444#define SEC_LEVEL_1 1 /* WEP 40 and 104 bit */
445#define SEC_LEVEL_2 2 /* Level 1 + TKIP */
446#define SEC_LEVEL_2_CKIP 3 /* Level 1 + CKIP */
447#define SEC_LEVEL_3 4 /* Level 2 + CCMP */
448
449#define WEP_KEYS 4
450#define WEP_KEY_LEN 13
451
452struct ieee80211_security {
453 u16 active_key:2,
454 enabled:1,
455 auth_mode:2,
456 auth_algo:4,
457 unicast_uses_group:1;
458 u8 key_sizes[WEP_KEYS];
459 u8 keys[WEP_KEYS][WEP_KEY_LEN];
460 u8 level;
461 u16 flags;
462} __attribute__ ((packed));
463
464
465/*
466
467 802.11 data frame from AP
468
469 ,-------------------------------------------------------------------.
470Bytes | 2 | 2 | 6 | 6 | 6 | 2 | 0..2312 | 4 |
471 |------|------|---------|---------|---------|------|---------|------|
472Desc. | ctrl | dura | DA/RA | TA | SA | Sequ | frame | fcs |
473 | | tion | (BSSID) | | | ence | data | |
474 `-------------------------------------------------------------------'
475
476Total: 28-2340 bytes
477
478*/
479
480#define BEACON_PROBE_SSID_ID_POSITION 12
481
482/* Management Frame Information Element Types */
483#define MFIE_TYPE_SSID 0
484#define MFIE_TYPE_RATES 1
485#define MFIE_TYPE_FH_SET 2
486#define MFIE_TYPE_DS_SET 3
487#define MFIE_TYPE_CF_SET 4
488#define MFIE_TYPE_TIM 5
489#define MFIE_TYPE_IBSS_SET 6
490#define MFIE_TYPE_CHALLENGE 16
491#define MFIE_TYPE_RSN 48
492#define MFIE_TYPE_RATES_EX 50
493#define MFIE_TYPE_GENERIC 221
494
495struct ieee80211_info_element_hdr {
496 u8 id;
497 u8 len;
498} __attribute__ ((packed));
499
500struct ieee80211_info_element {
501 u8 id;
502 u8 len;
503 u8 data[0];
504} __attribute__ ((packed));
505
506/*
507 * These are the data types that can make up management packets
508 *
509 u16 auth_algorithm;
510 u16 auth_sequence;
511 u16 beacon_interval;
512 u16 capability;
513 u8 current_ap[ETH_ALEN];
514 u16 listen_interval;
515 struct {
516 u16 association_id:14, reserved:2;
517 } __attribute__ ((packed));
518 u32 time_stamp[2];
519 u16 reason;
520 u16 status;
521*/
522
523struct ieee80211_authentication {
524 struct ieee80211_hdr_3addr header;
525 u16 algorithm;
526 u16 transaction;
527 u16 status;
528 struct ieee80211_info_element info_element;
529} __attribute__ ((packed));
530
531
532struct ieee80211_probe_response {
533 struct ieee80211_hdr_3addr header;
534 u32 time_stamp[2];
535 u16 beacon_interval;
536 u16 capability;
537 struct ieee80211_info_element info_element;
538} __attribute__ ((packed));
539
540struct ieee80211_assoc_request_frame {
541 u16 capability;
542 u16 listen_interval;
543 u8 current_ap[ETH_ALEN];
544 struct ieee80211_info_element info_element;
545} __attribute__ ((packed));
546
547struct ieee80211_assoc_response_frame {
548 struct ieee80211_hdr_3addr header;
549 u16 capability;
550 u16 status;
551 u16 aid;
552 struct ieee80211_info_element info_element; /* supported rates */
553} __attribute__ ((packed));
554
555
556struct ieee80211_txb {
557 u8 nr_frags;
558 u8 encrypted;
559 u16 reserved;
560 u16 frag_size;
561 u16 payload_size;
562 struct sk_buff *fragments[0];
563};
564
565
566/* SWEEP TABLE ENTRIES NUMBER*/
567#define MAX_SWEEP_TAB_ENTRIES 42
568#define MAX_SWEEP_TAB_ENTRIES_PER_PACKET 7
569/* MAX_RATES_LENGTH needs to be 12. The spec says 8, and many APs
570 * only use 8, and then use extended rates for the remaining supported
571 * rates. Other APs, however, stick all of their supported rates on the
572 * main rates information element... */
573#define MAX_RATES_LENGTH ((u8)12)
574#define MAX_RATES_EX_LENGTH ((u8)16)
575#define MAX_NETWORK_COUNT 128
576
577#define CRC_LENGTH 4U
578
579#define MAX_WPA_IE_LEN 64
580
581#define NETWORK_EMPTY_ESSID (1<<0)
582#define NETWORK_HAS_OFDM (1<<1)
583#define NETWORK_HAS_CCK (1<<2)
584
585struct ieee80211_network {
586 /* These entries are used to identify a unique network */
587 u8 bssid[ETH_ALEN];
588 u8 channel;
589 /* Ensure null-terminated for any debug msgs */
590 u8 ssid[IW_ESSID_MAX_SIZE + 1];
591 u8 ssid_len;
592
593 /* These are network statistics */
594 struct ieee80211_rx_stats stats;
595 u16 capability;
596 u8 rates[MAX_RATES_LENGTH];
597 u8 rates_len;
598 u8 rates_ex[MAX_RATES_EX_LENGTH];
599 u8 rates_ex_len;
600 unsigned long last_scanned;
601 u8 mode;
602 u8 flags;
603 u32 last_associate;
604 u32 time_stamp[2];
605 u16 beacon_interval;
606 u16 listen_interval;
607 u16 atim_window;
608 u8 wpa_ie[MAX_WPA_IE_LEN];
609 size_t wpa_ie_len;
610 u8 rsn_ie[MAX_WPA_IE_LEN];
611 size_t rsn_ie_len;
612 struct list_head list;
613};
614
615enum ieee80211_state {
616 IEEE80211_UNINITIALIZED = 0,
617 IEEE80211_INITIALIZED,
618 IEEE80211_ASSOCIATING,
619 IEEE80211_ASSOCIATED,
620 IEEE80211_AUTHENTICATING,
621 IEEE80211_AUTHENTICATED,
622 IEEE80211_SHUTDOWN
623};
624
625#define DEFAULT_MAX_SCAN_AGE (15 * HZ)
626#define DEFAULT_FTS 2346
627#define MAC_FMT "%02x:%02x:%02x:%02x:%02x:%02x"
628#define MAC_ARG(x) ((u8*)(x))[0],((u8*)(x))[1],((u8*)(x))[2],((u8*)(x))[3],((u8*)(x))[4],((u8*)(x))[5]
629
630
631#define CFG_IEEE80211_RESERVE_FCS (1<<0)
632#define CFG_IEEE80211_COMPUTE_FCS (1<<1)
633
634struct ieee80211_device {
635 struct net_device *dev;
636
637 /* Bookkeeping structures */
638 struct net_device_stats stats;
639 struct ieee80211_stats ieee_stats;
640
641 /* Probe / Beacon management */
642 struct list_head network_free_list;
643 struct list_head network_list;
644 struct ieee80211_network *networks;
645 int scans;
646 int scan_age;
647
648 int iw_mode; /* operating mode (IW_MODE_*) */
649
650 spinlock_t lock;
651
652 int tx_headroom; /* Set to size of any additional room needed at front
653 * of allocated Tx SKBs */
654 u32 config;
655
656 /* WEP and other encryption related settings at the device level */
657 int open_wep; /* Set to 1 to allow unencrypted frames */
658
659 int reset_on_keychange; /* Set to 1 if the HW needs to be reset on
660 * WEP key changes */
661
662 /* If the host performs {en,de}cryption, then set to 1 */
663 int host_encrypt;
664 int host_decrypt;
665 int ieee802_1x; /* is IEEE 802.1X used */
666
667 /* WPA data */
668 int wpa_enabled;
669 int drop_unencrypted;
670 int tkip_countermeasures;
671 int privacy_invoked;
672 size_t wpa_ie_len;
673 u8 *wpa_ie;
674
675 struct list_head crypt_deinit_list;
676 struct ieee80211_crypt_data *crypt[WEP_KEYS];
677 int tx_keyidx; /* default TX key index (crypt[tx_keyidx]) */
678 struct timer_list crypt_deinit_timer;
679
680 int bcrx_sta_key; /* use individual keys to override default keys even
681 * with RX of broad/multicast frames */
682
683 /* Fragmentation structures */
684 struct ieee80211_frag_entry frag_cache[IEEE80211_FRAG_CACHE_LEN];
685 unsigned int frag_next_idx;
686 u16 fts; /* Fragmentation Threshold */
687
688 /* Association info */
689 u8 bssid[ETH_ALEN];
690
691 enum ieee80211_state state;
692
693 int mode; /* A, B, G */
694 int modulation; /* CCK, OFDM */
695 int freq_band; /* 2.4Ghz, 5.2Ghz, Mixed */
696 int abg_ture; /* ABG flag */
697
698 /* Callback functions */
699 void (*set_security)(struct net_device *dev,
700 struct ieee80211_security *sec);
701 int (*hard_start_xmit)(struct ieee80211_txb *txb,
702 struct net_device *dev);
703 int (*reset_port)(struct net_device *dev);
704
705 /* This must be the last item so that it points to the data
706 * allocated beyond this structure by alloc_ieee80211 */
707 u8 priv[0];
708};
709
710#define IEEE_A (1<<0)
711#define IEEE_B (1<<1)
712#define IEEE_G (1<<2)
713#define IEEE_MODE_MASK (IEEE_A|IEEE_B|IEEE_G)
714
715extern inline void *ieee80211_priv(struct net_device *dev)
716{
717 return ((struct ieee80211_device *)netdev_priv(dev))->priv;
718}
719
720extern inline int ieee80211_is_empty_essid(const char *essid, int essid_len)
721{
722 /* Single white space is for Linksys APs */
723 if (essid_len == 1 && essid[0] == ' ')
724 return 1;
725
726 /* Otherwise, if the entire essid is 0, we assume it is hidden */
727 while (essid_len) {
728 essid_len--;
729 if (essid[essid_len] != '\0')
730 return 0;
731 }
732
733 return 1;
734}
735
736extern inline int ieee80211_is_valid_mode(struct ieee80211_device *ieee, int mode)
737{
738 /*
739 * It is possible for both access points and our device to support
740 * combinations of modes, so as long as there is one valid combination
741 * of ap/device supported modes, then return success
742 *
743 */
744 if ((mode & IEEE_A) &&
745 (ieee->modulation & IEEE80211_OFDM_MODULATION) &&
746 (ieee->freq_band & IEEE80211_52GHZ_BAND))
747 return 1;
748
749 if ((mode & IEEE_G) &&
750 (ieee->modulation & IEEE80211_OFDM_MODULATION) &&
751 (ieee->freq_band & IEEE80211_24GHZ_BAND))
752 return 1;
753
754 if ((mode & IEEE_B) &&
755 (ieee->modulation & IEEE80211_CCK_MODULATION) &&
756 (ieee->freq_band & IEEE80211_24GHZ_BAND))
757 return 1;
758
759 return 0;
760}
761
762extern inline int ieee80211_get_hdrlen(u16 fc)
763{
764 int hdrlen = IEEE80211_3ADDR_LEN;
765
766 switch (WLAN_FC_GET_TYPE(fc)) {
767 case IEEE80211_FTYPE_DATA:
768 if ((fc & IEEE80211_FCTL_FROMDS) && (fc & IEEE80211_FCTL_TODS))
769 hdrlen = IEEE80211_4ADDR_LEN;
770 break;
771 case IEEE80211_FTYPE_CTL:
772 switch (WLAN_FC_GET_STYPE(fc)) {
773 case IEEE80211_STYPE_CTS:
774 case IEEE80211_STYPE_ACK:
775 hdrlen = IEEE80211_1ADDR_LEN;
776 break;
777 default:
778 hdrlen = IEEE80211_2ADDR_LEN;
779 break;
780 }
781 break;
782 }
783
784 return hdrlen;
785}
786
787
788
789/* ieee80211.c */
790extern void free_ieee80211(struct net_device *dev);
791extern struct net_device *alloc_ieee80211(int sizeof_priv);
792
793extern int ieee80211_set_encryption(struct ieee80211_device *ieee);
794
795/* ieee80211_tx.c */
796
797
798extern int ieee80211_xmit(struct sk_buff *skb,
799 struct net_device *dev);
800extern void ieee80211_txb_free(struct ieee80211_txb *);
801
802
803/* ieee80211_rx.c */
804extern int ieee80211_rx(struct ieee80211_device *ieee, struct sk_buff *skb,
805 struct ieee80211_rx_stats *rx_stats);
806extern void ieee80211_rx_mgt(struct ieee80211_device *ieee,
807 struct ieee80211_hdr *header,
808 struct ieee80211_rx_stats *stats);
809
810/* iee80211_wx.c */
811extern int ieee80211_wx_get_scan(struct ieee80211_device *ieee,
812 struct iw_request_info *info,
813 union iwreq_data *wrqu, char *key);
814extern int ieee80211_wx_set_encode(struct ieee80211_device *ieee,
815 struct iw_request_info *info,
816 union iwreq_data *wrqu, char *key);
817extern int ieee80211_wx_get_encode(struct ieee80211_device *ieee,
818 struct iw_request_info *info,
819 union iwreq_data *wrqu, char *key);
820
821
822extern inline void ieee80211_increment_scans(struct ieee80211_device *ieee)
823{
824 ieee->scans++;
825}
826
827extern inline int ieee80211_get_scans(struct ieee80211_device *ieee)
828{
829 return ieee->scans;
830}
831
832static inline const char *escape_essid(const char *essid, u8 essid_len) {
833 static char escaped[IW_ESSID_MAX_SIZE * 2 + 1];
834 const char *s = essid;
835 char *d = escaped;
836
837 if (ieee80211_is_empty_essid(essid, essid_len)) {
838 memcpy(escaped, "<hidden>", sizeof("<hidden>"));
839 return escaped;
840 }
841
842 essid_len = min(essid_len, (u8)IW_ESSID_MAX_SIZE);
843 while (essid_len--) {
844 if (*s == '\0') {
845 *d++ = '\\';
846 *d++ = '0';
847 s++;
848 } else {
849 *d++ = *s++;
850 }
851 }
852 *d = '\0';
853 return escaped;
854}
855
856#endif /* IEEE80211_H */
diff --git a/include/net/ip.h b/include/net/ip.h
index 3f63992eb712..32360bbe143f 100644
--- a/include/net/ip.h
+++ b/include/net/ip.h
@@ -163,6 +163,7 @@ DECLARE_SNMP_STAT(struct linux_mib, net_statistics);
163 163
164extern int sysctl_local_port_range[2]; 164extern int sysctl_local_port_range[2];
165extern int sysctl_ip_default_ttl; 165extern int sysctl_ip_default_ttl;
166extern int sysctl_ip_nonlocal_bind;
166 167
167#ifdef CONFIG_INET 168#ifdef CONFIG_INET
168/* The function in 2.2 was invalid, producing wrong result for 169/* The function in 2.2 was invalid, producing wrong result for
diff --git a/include/net/ip6_fib.h b/include/net/ip6_fib.h
index 319904518194..a66e9de16a6c 100644
--- a/include/net/ip6_fib.h
+++ b/include/net/ip6_fib.h
@@ -167,14 +167,17 @@ extern int fib6_walk_continue(struct fib6_walker_t *w);
167extern int fib6_add(struct fib6_node *root, 167extern int fib6_add(struct fib6_node *root,
168 struct rt6_info *rt, 168 struct rt6_info *rt,
169 struct nlmsghdr *nlh, 169 struct nlmsghdr *nlh,
170 void *rtattr); 170 void *rtattr,
171 struct netlink_skb_parms *req);
171 172
172extern int fib6_del(struct rt6_info *rt, 173extern int fib6_del(struct rt6_info *rt,
173 struct nlmsghdr *nlh, 174 struct nlmsghdr *nlh,
174 void *rtattr); 175 void *rtattr,
176 struct netlink_skb_parms *req);
175 177
176extern void inet6_rt_notify(int event, struct rt6_info *rt, 178extern void inet6_rt_notify(int event, struct rt6_info *rt,
177 struct nlmsghdr *nlh); 179 struct nlmsghdr *nlh,
180 struct netlink_skb_parms *req);
178 181
179extern void fib6_run_gc(unsigned long dummy); 182extern void fib6_run_gc(unsigned long dummy);
180 183
diff --git a/include/net/ip6_route.h b/include/net/ip6_route.h
index d5d1dd10cdb8..f920706d526b 100644
--- a/include/net/ip6_route.h
+++ b/include/net/ip6_route.h
@@ -41,13 +41,16 @@ extern int ipv6_route_ioctl(unsigned int cmd, void __user *arg);
41 41
42extern int ip6_route_add(struct in6_rtmsg *rtmsg, 42extern int ip6_route_add(struct in6_rtmsg *rtmsg,
43 struct nlmsghdr *, 43 struct nlmsghdr *,
44 void *rtattr); 44 void *rtattr,
45 struct netlink_skb_parms *req);
45extern int ip6_ins_rt(struct rt6_info *, 46extern int ip6_ins_rt(struct rt6_info *,
46 struct nlmsghdr *, 47 struct nlmsghdr *,
47 void *rtattr); 48 void *rtattr,
49 struct netlink_skb_parms *req);
48extern int ip6_del_rt(struct rt6_info *, 50extern int ip6_del_rt(struct rt6_info *,
49 struct nlmsghdr *, 51 struct nlmsghdr *,
50 void *rtattr); 52 void *rtattr,
53 struct netlink_skb_parms *req);
51 54
52extern int ip6_rt_addr_add(struct in6_addr *addr, 55extern int ip6_rt_addr_add(struct in6_addr *addr,
53 struct net_device *dev, 56 struct net_device *dev,
diff --git a/include/net/ip_fib.h b/include/net/ip_fib.h
index e5a5f6b62f88..a4208a336ac0 100644
--- a/include/net/ip_fib.h
+++ b/include/net/ip_fib.h
@@ -109,6 +109,20 @@ struct fib_result {
109#endif 109#endif
110}; 110};
111 111
112struct fib_result_nl {
113 u32 fl_addr; /* To be looked up*/
114 u32 fl_fwmark;
115 unsigned char fl_tos;
116 unsigned char fl_scope;
117 unsigned char tb_id_in;
118
119 unsigned char tb_id; /* Results */
120 unsigned char prefixlen;
121 unsigned char nh_sel;
122 unsigned char type;
123 unsigned char scope;
124 int err;
125};
112 126
113#ifdef CONFIG_IP_ROUTE_MULTIPATH 127#ifdef CONFIG_IP_ROUTE_MULTIPATH
114 128
diff --git a/include/net/ipv6.h b/include/net/ipv6.h
index 771b47e30f86..69324465e8b3 100644
--- a/include/net/ipv6.h
+++ b/include/net/ipv6.h
@@ -183,7 +183,6 @@ struct ipv6_txoptions
183 struct ipv6_opt_hdr *hopopt; 183 struct ipv6_opt_hdr *hopopt;
184 struct ipv6_opt_hdr *dst0opt; 184 struct ipv6_opt_hdr *dst0opt;
185 struct ipv6_rt_hdr *srcrt; /* Routing Header */ 185 struct ipv6_rt_hdr *srcrt; /* Routing Header */
186 struct ipv6_opt_hdr *auth;
187 struct ipv6_opt_hdr *dst1opt; 186 struct ipv6_opt_hdr *dst1opt;
188 187
189 /* Option buffer, as read by IPV6_PKTOPTIONS, starts here. */ 188 /* Option buffer, as read by IPV6_PKTOPTIONS, starts here. */
diff --git a/include/net/irda/irda_device.h b/include/net/irda/irda_device.h
index 71d6af83b631..92c828029cd8 100644
--- a/include/net/irda/irda_device.h
+++ b/include/net/irda/irda_device.h
@@ -224,7 +224,7 @@ int irda_device_is_receiving(struct net_device *dev);
224/* Interface for internal use */ 224/* Interface for internal use */
225static inline int irda_device_txqueue_empty(const struct net_device *dev) 225static inline int irda_device_txqueue_empty(const struct net_device *dev)
226{ 226{
227 return (skb_queue_len(&dev->qdisc->q) == 0); 227 return skb_queue_empty(&dev->qdisc->q);
228} 228}
229int irda_device_set_raw_mode(struct net_device* self, int status); 229int irda_device_set_raw_mode(struct net_device* self, int status);
230struct net_device *alloc_irdadev(int sizeof_priv); 230struct net_device *alloc_irdadev(int sizeof_priv);
diff --git a/include/net/neighbour.h b/include/net/neighbour.h
index 4f33bbc21e7f..89809891e5ab 100644
--- a/include/net/neighbour.h
+++ b/include/net/neighbour.h
@@ -65,11 +65,10 @@ struct neighbour;
65 65
66struct neigh_parms 66struct neigh_parms
67{ 67{
68 struct net_device *dev;
68 struct neigh_parms *next; 69 struct neigh_parms *next;
69 int (*neigh_setup)(struct neighbour *); 70 int (*neigh_setup)(struct neighbour *);
70 struct neigh_table *tbl; 71 struct neigh_table *tbl;
71 int entries;
72 void *priv;
73 72
74 void *sysctl_table; 73 void *sysctl_table;
75 74
@@ -192,7 +191,6 @@ struct neigh_table
192 atomic_t entries; 191 atomic_t entries;
193 rwlock_t lock; 192 rwlock_t lock;
194 unsigned long last_rand; 193 unsigned long last_rand;
195 struct neigh_parms *parms_list;
196 kmem_cache_t *kmem_cachep; 194 kmem_cache_t *kmem_cachep;
197 struct neigh_statistics *stats; 195 struct neigh_statistics *stats;
198 struct neighbour **hash_buckets; 196 struct neighbour **hash_buckets;
@@ -252,6 +250,9 @@ extern int neigh_add(struct sk_buff *skb, struct nlmsghdr *nlh, void *arg);
252extern int neigh_delete(struct sk_buff *skb, struct nlmsghdr *nlh, void *arg); 250extern int neigh_delete(struct sk_buff *skb, struct nlmsghdr *nlh, void *arg);
253extern void neigh_app_ns(struct neighbour *n); 251extern void neigh_app_ns(struct neighbour *n);
254 252
253extern int neightbl_dump_info(struct sk_buff *skb, struct netlink_callback *cb);
254extern int neightbl_set(struct sk_buff *skb, struct nlmsghdr *nlh, void *arg);
255
255extern void neigh_for_each(struct neigh_table *tbl, void (*cb)(struct neighbour *, void *), void *cookie); 256extern void neigh_for_each(struct neigh_table *tbl, void (*cb)(struct neighbour *, void *), void *cookie);
256extern void __neigh_for_each_release(struct neigh_table *tbl, int (*cb)(struct neighbour *)); 257extern void __neigh_for_each_release(struct neigh_table *tbl, int (*cb)(struct neighbour *));
257extern void pneigh_for_each(struct neigh_table *tbl, void (*cb)(struct pneigh_entry *)); 258extern void pneigh_for_each(struct neigh_table *tbl, void (*cb)(struct pneigh_entry *));
diff --git a/include/net/pkt_sched.h b/include/net/pkt_sched.h
index fcb05a387dbe..6492e7363d84 100644
--- a/include/net/pkt_sched.h
+++ b/include/net/pkt_sched.h
@@ -13,13 +13,12 @@ struct qdisc_walker
13 13
14extern rwlock_t qdisc_tree_lock; 14extern rwlock_t qdisc_tree_lock;
15 15
16#define QDISC_ALIGN 32 16#define QDISC_ALIGNTO 32
17#define QDISC_ALIGN_CONST (QDISC_ALIGN - 1) 17#define QDISC_ALIGN(len) (((len) + QDISC_ALIGNTO-1) & ~(QDISC_ALIGNTO-1))
18 18
19static inline void *qdisc_priv(struct Qdisc *q) 19static inline void *qdisc_priv(struct Qdisc *q)
20{ 20{
21 return (char *)q + ((sizeof(struct Qdisc) + QDISC_ALIGN_CONST) 21 return (char *) q + QDISC_ALIGN(sizeof(struct Qdisc));
22 & ~QDISC_ALIGN_CONST);
23} 22}
24 23
25/* 24/*
@@ -207,8 +206,6 @@ psched_tod_diff(int delta_sec, int bound)
207 206
208#endif /* !CONFIG_NET_SCH_CLK_GETTIMEOFDAY */ 207#endif /* !CONFIG_NET_SCH_CLK_GETTIMEOFDAY */
209 208
210extern struct Qdisc noop_qdisc;
211extern struct Qdisc_ops noop_qdisc_ops;
212extern struct Qdisc_ops pfifo_qdisc_ops; 209extern struct Qdisc_ops pfifo_qdisc_ops;
213extern struct Qdisc_ops bfifo_qdisc_ops; 210extern struct Qdisc_ops bfifo_qdisc_ops;
214 211
@@ -216,14 +213,6 @@ extern int register_qdisc(struct Qdisc_ops *qops);
216extern int unregister_qdisc(struct Qdisc_ops *qops); 213extern int unregister_qdisc(struct Qdisc_ops *qops);
217extern struct Qdisc *qdisc_lookup(struct net_device *dev, u32 handle); 214extern struct Qdisc *qdisc_lookup(struct net_device *dev, u32 handle);
218extern struct Qdisc *qdisc_lookup_class(struct net_device *dev, u32 handle); 215extern struct Qdisc *qdisc_lookup_class(struct net_device *dev, u32 handle);
219extern void dev_init_scheduler(struct net_device *dev);
220extern void dev_shutdown(struct net_device *dev);
221extern void dev_activate(struct net_device *dev);
222extern void dev_deactivate(struct net_device *dev);
223extern void qdisc_reset(struct Qdisc *qdisc);
224extern void qdisc_destroy(struct Qdisc *qdisc);
225extern struct Qdisc * qdisc_create_dflt(struct net_device *dev,
226 struct Qdisc_ops *ops);
227extern struct qdisc_rate_table *qdisc_get_rtab(struct tc_ratespec *r, 216extern struct qdisc_rate_table *qdisc_get_rtab(struct tc_ratespec *r,
228 struct rtattr *tab); 217 struct rtattr *tab);
229extern void qdisc_put_rtab(struct qdisc_rate_table *tab); 218extern void qdisc_put_rtab(struct qdisc_rate_table *tab);
diff --git a/include/net/request_sock.h b/include/net/request_sock.h
new file mode 100644
index 000000000000..72fd6f5e86b1
--- /dev/null
+++ b/include/net/request_sock.h
@@ -0,0 +1,255 @@
1/*
2 * NET Generic infrastructure for Network protocols.
3 *
4 * Definitions for request_sock
5 *
6 * Authors: Arnaldo Carvalho de Melo <acme@conectiva.com.br>
7 *
8 * From code originally in include/net/tcp.h
9 *
10 * This program is free software; you can redistribute it and/or
11 * modify it under the terms of the GNU General Public License
12 * as published by the Free Software Foundation; either version
13 * 2 of the License, or (at your option) any later version.
14 */
15#ifndef _REQUEST_SOCK_H
16#define _REQUEST_SOCK_H
17
18#include <linux/slab.h>
19#include <linux/spinlock.h>
20#include <linux/types.h>
21
22#include <net/sock.h>
23
24struct request_sock;
25struct sk_buff;
26struct dst_entry;
27struct proto;
28
29struct request_sock_ops {
30 int family;
31 kmem_cache_t *slab;
32 int obj_size;
33 int (*rtx_syn_ack)(struct sock *sk,
34 struct request_sock *req,
35 struct dst_entry *dst);
36 void (*send_ack)(struct sk_buff *skb,
37 struct request_sock *req);
38 void (*send_reset)(struct sk_buff *skb);
39 void (*destructor)(struct request_sock *req);
40};
41
42/* struct request_sock - mini sock to represent a connection request
43 */
44struct request_sock {
45 struct request_sock *dl_next; /* Must be first member! */
46 u16 mss;
47 u8 retrans;
48 u8 __pad;
49 /* The following two fields can be easily recomputed I think -AK */
50 u32 window_clamp; /* window clamp at creation time */
51 u32 rcv_wnd; /* rcv_wnd offered first time */
52 u32 ts_recent;
53 unsigned long expires;
54 struct request_sock_ops *rsk_ops;
55 struct sock *sk;
56};
57
58static inline struct request_sock *reqsk_alloc(struct request_sock_ops *ops)
59{
60 struct request_sock *req = kmem_cache_alloc(ops->slab, SLAB_ATOMIC);
61
62 if (req != NULL)
63 req->rsk_ops = ops;
64
65 return req;
66}
67
68static inline void __reqsk_free(struct request_sock *req)
69{
70 kmem_cache_free(req->rsk_ops->slab, req);
71}
72
73static inline void reqsk_free(struct request_sock *req)
74{
75 req->rsk_ops->destructor(req);
76 __reqsk_free(req);
77}
78
79extern int sysctl_max_syn_backlog;
80
81/** struct listen_sock - listen state
82 *
83 * @max_qlen_log - log_2 of maximal queued SYNs/REQUESTs
84 */
85struct listen_sock {
86 u8 max_qlen_log;
87 /* 3 bytes hole, try to use */
88 int qlen;
89 int qlen_young;
90 int clock_hand;
91 u32 hash_rnd;
92 struct request_sock *syn_table[0];
93};
94
95/** struct request_sock_queue - queue of request_socks
96 *
97 * @rskq_accept_head - FIFO head of established children
98 * @rskq_accept_tail - FIFO tail of established children
99 * @syn_wait_lock - serializer
100 *
101 * %syn_wait_lock is necessary only to avoid proc interface having to grab the main
102 * lock sock while browsing the listening hash (otherwise it's deadlock prone).
103 *
104 * This lock is acquired in read mode only from listening_get_next() seq_file
105 * op and it's acquired in write mode _only_ from code that is actively
106 * changing rskq_accept_head. All readers that are holding the master sock lock
107 * don't need to grab this lock in read mode too as rskq_accept_head. writes
108 * are always protected from the main sock lock.
109 */
110struct request_sock_queue {
111 struct request_sock *rskq_accept_head;
112 struct request_sock *rskq_accept_tail;
113 rwlock_t syn_wait_lock;
114 struct listen_sock *listen_opt;
115};
116
117extern int reqsk_queue_alloc(struct request_sock_queue *queue,
118 const int nr_table_entries);
119
120static inline struct listen_sock *reqsk_queue_yank_listen_sk(struct request_sock_queue *queue)
121{
122 struct listen_sock *lopt;
123
124 write_lock_bh(&queue->syn_wait_lock);
125 lopt = queue->listen_opt;
126 queue->listen_opt = NULL;
127 write_unlock_bh(&queue->syn_wait_lock);
128
129 return lopt;
130}
131
132static inline void reqsk_queue_destroy(struct request_sock_queue *queue)
133{
134 kfree(reqsk_queue_yank_listen_sk(queue));
135}
136
137static inline struct request_sock *
138 reqsk_queue_yank_acceptq(struct request_sock_queue *queue)
139{
140 struct request_sock *req = queue->rskq_accept_head;
141
142 queue->rskq_accept_head = queue->rskq_accept_head = NULL;
143 return req;
144}
145
146static inline int reqsk_queue_empty(struct request_sock_queue *queue)
147{
148 return queue->rskq_accept_head == NULL;
149}
150
151static inline void reqsk_queue_unlink(struct request_sock_queue *queue,
152 struct request_sock *req,
153 struct request_sock **prev_req)
154{
155 write_lock(&queue->syn_wait_lock);
156 *prev_req = req->dl_next;
157 write_unlock(&queue->syn_wait_lock);
158}
159
160static inline void reqsk_queue_add(struct request_sock_queue *queue,
161 struct request_sock *req,
162 struct sock *parent,
163 struct sock *child)
164{
165 req->sk = child;
166 sk_acceptq_added(parent);
167
168 if (queue->rskq_accept_head == NULL)
169 queue->rskq_accept_head = req;
170 else
171 queue->rskq_accept_tail->dl_next = req;
172
173 queue->rskq_accept_tail = req;
174 req->dl_next = NULL;
175}
176
177static inline struct request_sock *reqsk_queue_remove(struct request_sock_queue *queue)
178{
179 struct request_sock *req = queue->rskq_accept_head;
180
181 BUG_TRAP(req != NULL);
182
183 queue->rskq_accept_head = req->dl_next;
184 if (queue->rskq_accept_head == NULL)
185 queue->rskq_accept_tail = NULL;
186
187 return req;
188}
189
190static inline struct sock *reqsk_queue_get_child(struct request_sock_queue *queue,
191 struct sock *parent)
192{
193 struct request_sock *req = reqsk_queue_remove(queue);
194 struct sock *child = req->sk;
195
196 BUG_TRAP(child != NULL);
197
198 sk_acceptq_removed(parent);
199 __reqsk_free(req);
200 return child;
201}
202
203static inline int reqsk_queue_removed(struct request_sock_queue *queue,
204 struct request_sock *req)
205{
206 struct listen_sock *lopt = queue->listen_opt;
207
208 if (req->retrans == 0)
209 --lopt->qlen_young;
210
211 return --lopt->qlen;
212}
213
214static inline int reqsk_queue_added(struct request_sock_queue *queue)
215{
216 struct listen_sock *lopt = queue->listen_opt;
217 const int prev_qlen = lopt->qlen;
218
219 lopt->qlen_young++;
220 lopt->qlen++;
221 return prev_qlen;
222}
223
224static inline int reqsk_queue_len(struct request_sock_queue *queue)
225{
226 return queue->listen_opt != NULL ? queue->listen_opt->qlen : 0;
227}
228
229static inline int reqsk_queue_len_young(struct request_sock_queue *queue)
230{
231 return queue->listen_opt->qlen_young;
232}
233
234static inline int reqsk_queue_is_full(struct request_sock_queue *queue)
235{
236 return queue->listen_opt->qlen >> queue->listen_opt->max_qlen_log;
237}
238
239static inline void reqsk_queue_hash_req(struct request_sock_queue *queue,
240 u32 hash, struct request_sock *req,
241 unsigned timeout)
242{
243 struct listen_sock *lopt = queue->listen_opt;
244
245 req->expires = jiffies + timeout;
246 req->retrans = 0;
247 req->sk = NULL;
248 req->dl_next = lopt->syn_table[hash];
249
250 write_lock(&queue->syn_wait_lock);
251 lopt->syn_table[hash] = req;
252 write_unlock(&queue->syn_wait_lock);
253}
254
255#endif /* _REQUEST_SOCK_H */
diff --git a/include/net/route.h b/include/net/route.h
index d34ca8fc6756..c3cd069a9aca 100644
--- a/include/net/route.h
+++ b/include/net/route.h
@@ -107,7 +107,7 @@ struct rt_cache_stat
107 107
108extern struct rt_cache_stat *rt_cache_stat; 108extern struct rt_cache_stat *rt_cache_stat;
109#define RT_CACHE_STAT_INC(field) \ 109#define RT_CACHE_STAT_INC(field) \
110 (per_cpu_ptr(rt_cache_stat, _smp_processor_id())->field++) 110 (per_cpu_ptr(rt_cache_stat, raw_smp_processor_id())->field++)
111 111
112extern struct ip_rt_acct *ip_rt_acct; 112extern struct ip_rt_acct *ip_rt_acct;
113 113
diff --git a/include/net/sch_generic.h b/include/net/sch_generic.h
index c57504b3b518..7b6ec9986715 100644
--- a/include/net/sch_generic.h
+++ b/include/net/sch_generic.h
@@ -164,6 +164,19 @@ extern void qdisc_unlock_tree(struct net_device *dev);
164#define tcf_tree_lock(tp) qdisc_lock_tree((tp)->q->dev) 164#define tcf_tree_lock(tp) qdisc_lock_tree((tp)->q->dev)
165#define tcf_tree_unlock(tp) qdisc_unlock_tree((tp)->q->dev) 165#define tcf_tree_unlock(tp) qdisc_unlock_tree((tp)->q->dev)
166 166
167extern struct Qdisc noop_qdisc;
168extern struct Qdisc_ops noop_qdisc_ops;
169
170extern void dev_init_scheduler(struct net_device *dev);
171extern void dev_shutdown(struct net_device *dev);
172extern void dev_activate(struct net_device *dev);
173extern void dev_deactivate(struct net_device *dev);
174extern void qdisc_reset(struct Qdisc *qdisc);
175extern void qdisc_destroy(struct Qdisc *qdisc);
176extern struct Qdisc *qdisc_alloc(struct net_device *dev, struct Qdisc_ops *ops);
177extern struct Qdisc *qdisc_create_dflt(struct net_device *dev,
178 struct Qdisc_ops *ops);
179
167static inline void 180static inline void
168tcf_destroy(struct tcf_proto *tp) 181tcf_destroy(struct tcf_proto *tp)
169{ 182{
@@ -172,4 +185,126 @@ tcf_destroy(struct tcf_proto *tp)
172 kfree(tp); 185 kfree(tp);
173} 186}
174 187
188static inline int __qdisc_enqueue_tail(struct sk_buff *skb, struct Qdisc *sch,
189 struct sk_buff_head *list)
190{
191 __skb_queue_tail(list, skb);
192 sch->qstats.backlog += skb->len;
193 sch->bstats.bytes += skb->len;
194 sch->bstats.packets++;
195
196 return NET_XMIT_SUCCESS;
197}
198
199static inline int qdisc_enqueue_tail(struct sk_buff *skb, struct Qdisc *sch)
200{
201 return __qdisc_enqueue_tail(skb, sch, &sch->q);
202}
203
204static inline struct sk_buff *__qdisc_dequeue_head(struct Qdisc *sch,
205 struct sk_buff_head *list)
206{
207 struct sk_buff *skb = __skb_dequeue(list);
208
209 if (likely(skb != NULL))
210 sch->qstats.backlog -= skb->len;
211
212 return skb;
213}
214
215static inline struct sk_buff *qdisc_dequeue_head(struct Qdisc *sch)
216{
217 return __qdisc_dequeue_head(sch, &sch->q);
218}
219
220static inline struct sk_buff *__qdisc_dequeue_tail(struct Qdisc *sch,
221 struct sk_buff_head *list)
222{
223 struct sk_buff *skb = __skb_dequeue_tail(list);
224
225 if (likely(skb != NULL))
226 sch->qstats.backlog -= skb->len;
227
228 return skb;
229}
230
231static inline struct sk_buff *qdisc_dequeue_tail(struct Qdisc *sch)
232{
233 return __qdisc_dequeue_tail(sch, &sch->q);
234}
235
236static inline int __qdisc_requeue(struct sk_buff *skb, struct Qdisc *sch,
237 struct sk_buff_head *list)
238{
239 __skb_queue_head(list, skb);
240 sch->qstats.backlog += skb->len;
241 sch->qstats.requeues++;
242
243 return NET_XMIT_SUCCESS;
244}
245
246static inline int qdisc_requeue(struct sk_buff *skb, struct Qdisc *sch)
247{
248 return __qdisc_requeue(skb, sch, &sch->q);
249}
250
251static inline void __qdisc_reset_queue(struct Qdisc *sch,
252 struct sk_buff_head *list)
253{
254 /*
255 * We do not know the backlog in bytes of this list, it
256 * is up to the caller to correct it
257 */
258 skb_queue_purge(list);
259}
260
261static inline void qdisc_reset_queue(struct Qdisc *sch)
262{
263 __qdisc_reset_queue(sch, &sch->q);
264 sch->qstats.backlog = 0;
265}
266
267static inline unsigned int __qdisc_queue_drop(struct Qdisc *sch,
268 struct sk_buff_head *list)
269{
270 struct sk_buff *skb = __qdisc_dequeue_tail(sch, list);
271
272 if (likely(skb != NULL)) {
273 unsigned int len = skb->len;
274 kfree_skb(skb);
275 return len;
276 }
277
278 return 0;
279}
280
281static inline unsigned int qdisc_queue_drop(struct Qdisc *sch)
282{
283 return __qdisc_queue_drop(sch, &sch->q);
284}
285
286static inline int qdisc_drop(struct sk_buff *skb, struct Qdisc *sch)
287{
288 kfree_skb(skb);
289 sch->qstats.drops++;
290
291 return NET_XMIT_DROP;
292}
293
294static inline int qdisc_reshape_fail(struct sk_buff *skb, struct Qdisc *sch)
295{
296 sch->qstats.drops++;
297
298#ifdef CONFIG_NET_CLS_POLICE
299 if (sch->reshape_fail == NULL || sch->reshape_fail(skb, sch))
300 goto drop;
301
302 return NET_XMIT_SUCCESS;
303
304drop:
305#endif
306 kfree_skb(skb);
307 return NET_XMIT_DROP;
308}
309
175#endif 310#endif
diff --git a/include/net/sctp/command.h b/include/net/sctp/command.h
index ebc5282e6d58..dc107ffad483 100644
--- a/include/net/sctp/command.h
+++ b/include/net/sctp/command.h
@@ -65,9 +65,11 @@ typedef enum {
65 SCTP_CMD_TIMER_START, /* Start a timer. */ 65 SCTP_CMD_TIMER_START, /* Start a timer. */
66 SCTP_CMD_TIMER_RESTART, /* Restart a timer. */ 66 SCTP_CMD_TIMER_RESTART, /* Restart a timer. */
67 SCTP_CMD_TIMER_STOP, /* Stop a timer. */ 67 SCTP_CMD_TIMER_STOP, /* Stop a timer. */
68 SCTP_CMD_COUNTER_RESET, /* Reset a counter. */ 68 SCTP_CMD_INIT_CHOOSE_TRANSPORT, /* Choose transport for an INIT. */
69 SCTP_CMD_COUNTER_INC, /* Increment a counter. */ 69 SCTP_CMD_INIT_COUNTER_RESET, /* Reset init counter. */
70 SCTP_CMD_INIT_COUNTER_INC, /* Increment init counter. */
70 SCTP_CMD_INIT_RESTART, /* High level, do init timer work. */ 71 SCTP_CMD_INIT_RESTART, /* High level, do init timer work. */
72 SCTP_CMD_COOKIEECHO_RESTART, /* High level, do cookie-echo timer work. */
71 SCTP_CMD_INIT_FAILED, /* High level, do init failure work. */ 73 SCTP_CMD_INIT_FAILED, /* High level, do init failure work. */
72 SCTP_CMD_REPORT_DUP, /* Report a duplicate TSN. */ 74 SCTP_CMD_REPORT_DUP, /* Report a duplicate TSN. */
73 SCTP_CMD_STRIKE, /* Mark a strike against a transport. */ 75 SCTP_CMD_STRIKE, /* Mark a strike against a transport. */
@@ -118,7 +120,6 @@ typedef union {
118 int error; 120 int error;
119 sctp_state_t state; 121 sctp_state_t state;
120 sctp_event_timeout_t to; 122 sctp_event_timeout_t to;
121 sctp_counter_t counter;
122 void *ptr; 123 void *ptr;
123 struct sctp_chunk *chunk; 124 struct sctp_chunk *chunk;
124 struct sctp_association *asoc; 125 struct sctp_association *asoc;
@@ -165,7 +166,6 @@ SCTP_ARG_CONSTRUCTOR(U16, __u16, u16)
165SCTP_ARG_CONSTRUCTOR(U8, __u8, u8) 166SCTP_ARG_CONSTRUCTOR(U8, __u8, u8)
166SCTP_ARG_CONSTRUCTOR(ERROR, int, error) 167SCTP_ARG_CONSTRUCTOR(ERROR, int, error)
167SCTP_ARG_CONSTRUCTOR(STATE, sctp_state_t, state) 168SCTP_ARG_CONSTRUCTOR(STATE, sctp_state_t, state)
168SCTP_ARG_CONSTRUCTOR(COUNTER, sctp_counter_t, counter)
169SCTP_ARG_CONSTRUCTOR(TO, sctp_event_timeout_t, to) 169SCTP_ARG_CONSTRUCTOR(TO, sctp_event_timeout_t, to)
170SCTP_ARG_CONSTRUCTOR(PTR, void *, ptr) 170SCTP_ARG_CONSTRUCTOR(PTR, void *, ptr)
171SCTP_ARG_CONSTRUCTOR(CHUNK, struct sctp_chunk *, chunk) 171SCTP_ARG_CONSTRUCTOR(CHUNK, struct sctp_chunk *, chunk)
diff --git a/include/net/sctp/constants.h b/include/net/sctp/constants.h
index 2b76c0f4babc..5999e5684bbf 100644
--- a/include/net/sctp/constants.h
+++ b/include/net/sctp/constants.h
@@ -263,30 +263,11 @@ enum { SCTP_MIN_PMTU = 576 };
263enum { SCTP_MAX_DUP_TSNS = 16 }; 263enum { SCTP_MAX_DUP_TSNS = 16 };
264enum { SCTP_MAX_GABS = 16 }; 264enum { SCTP_MAX_GABS = 16 };
265 265
266typedef enum { 266/* Heartbeat interval - 30 secs */
267 SCTP_COUNTER_INIT_ERROR, 267#define SCTP_DEFAULT_TIMEOUT_HEARTBEAT (30 * HZ)
268} sctp_counter_t;
269
270/* How many counters does an association need? */
271#define SCTP_NUMBER_COUNTERS 5
272
273/* Here we define the default timers. */
274
275/* cookie timer def = ? seconds */
276#define SCTP_DEFAULT_TIMEOUT_T1_COOKIE (3 * HZ)
277
278/* init timer def = 3 seconds */
279#define SCTP_DEFAULT_TIMEOUT_T1_INIT (3 * HZ)
280
281/* shutdown timer def = 300 ms */
282#define SCTP_DEFAULT_TIMEOUT_T2_SHUTDOWN ((300 * HZ) / 1000)
283
284/* 0 seconds + RTO */
285#define SCTP_DEFAULT_TIMEOUT_HEARTBEAT (10 * HZ)
286 268
287/* recv timer def = 200ms (in usec) */ 269/* Delayed sack timer - 200ms */
288#define SCTP_DEFAULT_TIMEOUT_SACK ((200 * HZ) / 1000) 270#define SCTP_DEFAULT_TIMEOUT_SACK ((200 * HZ) / 1000)
289#define SCTP_DEFAULT_TIMEOUT_SACK_MAX ((500 * HZ) / 1000) /* 500 ms */
290 271
291/* RTO.Initial - 3 seconds 272/* RTO.Initial - 3 seconds
292 * RTO.Min - 1 second 273 * RTO.Min - 1 second
diff --git a/include/net/sctp/sctp.h b/include/net/sctp/sctp.h
index 960abfa48d68..e1d5ec1c23c0 100644
--- a/include/net/sctp/sctp.h
+++ b/include/net/sctp/sctp.h
@@ -125,7 +125,8 @@
125 */ 125 */
126extern struct sock *sctp_get_ctl_sock(void); 126extern struct sock *sctp_get_ctl_sock(void);
127extern int sctp_copy_local_addr_list(struct sctp_bind_addr *, 127extern int sctp_copy_local_addr_list(struct sctp_bind_addr *,
128 sctp_scope_t, int gfp, int flags); 128 sctp_scope_t, unsigned int __nocast gfp,
129 int flags);
129extern struct sctp_pf *sctp_get_pf_specific(sa_family_t family); 130extern struct sctp_pf *sctp_get_pf_specific(sa_family_t family);
130extern int sctp_register_pf(struct sctp_pf *, sa_family_t); 131extern int sctp_register_pf(struct sctp_pf *, sa_family_t);
131 132
@@ -166,15 +167,12 @@ void sctp_unhash_established(struct sctp_association *);
166void sctp_hash_endpoint(struct sctp_endpoint *); 167void sctp_hash_endpoint(struct sctp_endpoint *);
167void sctp_unhash_endpoint(struct sctp_endpoint *); 168void sctp_unhash_endpoint(struct sctp_endpoint *);
168struct sock *sctp_err_lookup(int family, struct sk_buff *, 169struct sock *sctp_err_lookup(int family, struct sk_buff *,
169 struct sctphdr *, struct sctp_endpoint **, 170 struct sctphdr *, struct sctp_association **,
170 struct sctp_association **,
171 struct sctp_transport **); 171 struct sctp_transport **);
172void sctp_err_finish(struct sock *, struct sctp_endpoint *, 172void sctp_err_finish(struct sock *, struct sctp_association *);
173 struct sctp_association *);
174void sctp_icmp_frag_needed(struct sock *, struct sctp_association *, 173void sctp_icmp_frag_needed(struct sock *, struct sctp_association *,
175 struct sctp_transport *t, __u32 pmtu); 174 struct sctp_transport *t, __u32 pmtu);
176void sctp_icmp_proto_unreachable(struct sock *sk, 175void sctp_icmp_proto_unreachable(struct sock *sk,
177 struct sctp_endpoint *ep,
178 struct sctp_association *asoc, 176 struct sctp_association *asoc,
179 struct sctp_transport *t); 177 struct sctp_transport *t);
180 178
@@ -223,6 +221,22 @@ DECLARE_SNMP_STAT(struct sctp_mib, sctp_statistics);
223extern int sctp_debug_flag; 221extern int sctp_debug_flag;
224#define SCTP_DEBUG_PRINTK(whatever...) \ 222#define SCTP_DEBUG_PRINTK(whatever...) \
225 ((void) (sctp_debug_flag && printk(KERN_DEBUG whatever))) 223 ((void) (sctp_debug_flag && printk(KERN_DEBUG whatever)))
224#define SCTP_DEBUG_PRINTK_IPADDR(lead, trail, leadparm, saddr, otherparms...) \
225 if (sctp_debug_flag) { \
226 if (saddr->sa.sa_family == AF_INET6) { \
227 printk(KERN_DEBUG \
228 lead "%04x:%04x:%04x:%04x:%04x:%04x:%04x:%04x" trail, \
229 leadparm, \
230 NIP6(saddr->v6.sin6_addr), \
231 otherparms); \
232 } else { \
233 printk(KERN_DEBUG \
234 lead "%u.%u.%u.%u" trail, \
235 leadparm, \
236 NIPQUAD(saddr->v4.sin_addr.s_addr), \
237 otherparms); \
238 } \
239 }
226#define SCTP_ENABLE_DEBUG { sctp_debug_flag = 1; } 240#define SCTP_ENABLE_DEBUG { sctp_debug_flag = 1; }
227#define SCTP_DISABLE_DEBUG { sctp_debug_flag = 0; } 241#define SCTP_DISABLE_DEBUG { sctp_debug_flag = 0; }
228 242
@@ -236,6 +250,7 @@ extern int sctp_debug_flag;
236#else /* SCTP_DEBUG */ 250#else /* SCTP_DEBUG */
237 251
238#define SCTP_DEBUG_PRINTK(whatever...) 252#define SCTP_DEBUG_PRINTK(whatever...)
253#define SCTP_DEBUG_PRINTK_IPADDR(whatever...)
239#define SCTP_ENABLE_DEBUG 254#define SCTP_ENABLE_DEBUG
240#define SCTP_DISABLE_DEBUG 255#define SCTP_DISABLE_DEBUG
241#define SCTP_ASSERT(expr, str, func) 256#define SCTP_ASSERT(expr, str, func)
diff --git a/include/net/sctp/sm.h b/include/net/sctp/sm.h
index f4fcee104707..58462164d960 100644
--- a/include/net/sctp/sm.h
+++ b/include/net/sctp/sm.h
@@ -116,7 +116,8 @@ sctp_state_fn_t sctp_sf_eat_data_fast_4_4;
116sctp_state_fn_t sctp_sf_eat_sack_6_2; 116sctp_state_fn_t sctp_sf_eat_sack_6_2;
117sctp_state_fn_t sctp_sf_tabort_8_4_8; 117sctp_state_fn_t sctp_sf_tabort_8_4_8;
118sctp_state_fn_t sctp_sf_operr_notify; 118sctp_state_fn_t sctp_sf_operr_notify;
119sctp_state_fn_t sctp_sf_t1_timer_expire; 119sctp_state_fn_t sctp_sf_t1_init_timer_expire;
120sctp_state_fn_t sctp_sf_t1_cookie_timer_expire;
120sctp_state_fn_t sctp_sf_t2_timer_expire; 121sctp_state_fn_t sctp_sf_t2_timer_expire;
121sctp_state_fn_t sctp_sf_t4_timer_expire; 122sctp_state_fn_t sctp_sf_t4_timer_expire;
122sctp_state_fn_t sctp_sf_t5_timer_expire; 123sctp_state_fn_t sctp_sf_t5_timer_expire;
@@ -130,7 +131,6 @@ sctp_state_fn_t sctp_sf_do_ecne;
130sctp_state_fn_t sctp_sf_ootb; 131sctp_state_fn_t sctp_sf_ootb;
131sctp_state_fn_t sctp_sf_pdiscard; 132sctp_state_fn_t sctp_sf_pdiscard;
132sctp_state_fn_t sctp_sf_violation; 133sctp_state_fn_t sctp_sf_violation;
133sctp_state_fn_t sctp_sf_violation_chunklen;
134sctp_state_fn_t sctp_sf_discard_chunk; 134sctp_state_fn_t sctp_sf_discard_chunk;
135sctp_state_fn_t sctp_sf_do_5_2_1_siminit; 135sctp_state_fn_t sctp_sf_do_5_2_1_siminit;
136sctp_state_fn_t sctp_sf_do_5_2_2_dupinit; 136sctp_state_fn_t sctp_sf_do_5_2_2_dupinit;
@@ -181,17 +181,17 @@ const sctp_sm_table_entry_t *sctp_sm_lookup_event(sctp_event_t,
181int sctp_chunk_iif(const struct sctp_chunk *); 181int sctp_chunk_iif(const struct sctp_chunk *);
182struct sctp_association *sctp_make_temp_asoc(const struct sctp_endpoint *, 182struct sctp_association *sctp_make_temp_asoc(const struct sctp_endpoint *,
183 struct sctp_chunk *, 183 struct sctp_chunk *,
184 int gfp); 184 unsigned int __nocast gfp);
185__u32 sctp_generate_verification_tag(void); 185__u32 sctp_generate_verification_tag(void);
186void sctp_populate_tie_tags(__u8 *cookie, __u32 curTag, __u32 hisTag); 186void sctp_populate_tie_tags(__u8 *cookie, __u32 curTag, __u32 hisTag);
187 187
188/* Prototypes for chunk-building functions. */ 188/* Prototypes for chunk-building functions. */
189struct sctp_chunk *sctp_make_init(const struct sctp_association *, 189struct sctp_chunk *sctp_make_init(const struct sctp_association *,
190 const struct sctp_bind_addr *, 190 const struct sctp_bind_addr *,
191 int gfp, int vparam_len); 191 unsigned int __nocast gfp, int vparam_len);
192struct sctp_chunk *sctp_make_init_ack(const struct sctp_association *, 192struct sctp_chunk *sctp_make_init_ack(const struct sctp_association *,
193 const struct sctp_chunk *, 193 const struct sctp_chunk *,
194 const int gfp, 194 const unsigned int __nocast gfp,
195 const int unkparam_len); 195 const int unkparam_len);
196struct sctp_chunk *sctp_make_cookie_echo(const struct sctp_association *, 196struct sctp_chunk *sctp_make_cookie_echo(const struct sctp_association *,
197 const struct sctp_chunk *); 197 const struct sctp_chunk *);
@@ -258,8 +258,6 @@ struct sctp_chunk *sctp_make_fwdtsn(const struct sctp_association *asoc,
258void sctp_chunk_assign_tsn(struct sctp_chunk *); 258void sctp_chunk_assign_tsn(struct sctp_chunk *);
259void sctp_chunk_assign_ssn(struct sctp_chunk *); 259void sctp_chunk_assign_ssn(struct sctp_chunk *);
260 260
261void sctp_stop_t1_and_abort(sctp_cmd_seq_t *commands, __u16 error);
262
263/* Prototypes for statetable processing. */ 261/* Prototypes for statetable processing. */
264 262
265int sctp_do_sm(sctp_event_t event_type, sctp_subtype_t subtype, 263int sctp_do_sm(sctp_event_t event_type, sctp_subtype_t subtype,
@@ -267,7 +265,7 @@ int sctp_do_sm(sctp_event_t event_type, sctp_subtype_t subtype,
267 struct sctp_endpoint *, 265 struct sctp_endpoint *,
268 struct sctp_association *asoc, 266 struct sctp_association *asoc,
269 void *event_arg, 267 void *event_arg,
270 int gfp); 268 unsigned int __nocast gfp);
271 269
272/* 2nd level prototypes */ 270/* 2nd level prototypes */
273void sctp_generate_t3_rtx_event(unsigned long peer); 271void sctp_generate_t3_rtx_event(unsigned long peer);
@@ -277,7 +275,8 @@ void sctp_ootb_pkt_free(struct sctp_packet *);
277 275
278struct sctp_association *sctp_unpack_cookie(const struct sctp_endpoint *, 276struct sctp_association *sctp_unpack_cookie(const struct sctp_endpoint *,
279 const struct sctp_association *, 277 const struct sctp_association *,
280 struct sctp_chunk *, int gfp, int *err, 278 struct sctp_chunk *,
279 unsigned int __nocast gfp, int *err,
281 struct sctp_chunk **err_chk_p); 280 struct sctp_chunk **err_chk_p);
282int sctp_addip_addr_config(struct sctp_association *, sctp_param_t, 281int sctp_addip_addr_config(struct sctp_association *, sctp_param_t,
283 struct sockaddr_storage*, int); 282 struct sockaddr_storage*, int);
diff --git a/include/net/sctp/structs.h b/include/net/sctp/structs.h
index 6c24d9cd3d66..994009bbe3b4 100644
--- a/include/net/sctp/structs.h
+++ b/include/net/sctp/structs.h
@@ -161,6 +161,9 @@ extern struct sctp_globals {
161 */ 161 */
162 int sndbuf_policy; 162 int sndbuf_policy;
163 163
164 /* Delayed SACK timeout 200ms default*/
165 int sack_timeout;
166
164 /* HB.interval - 30 seconds */ 167 /* HB.interval - 30 seconds */
165 int hb_interval; 168 int hb_interval;
166 169
@@ -217,6 +220,7 @@ extern struct sctp_globals {
217#define sctp_sndbuf_policy (sctp_globals.sndbuf_policy) 220#define sctp_sndbuf_policy (sctp_globals.sndbuf_policy)
218#define sctp_max_retrans_path (sctp_globals.max_retrans_path) 221#define sctp_max_retrans_path (sctp_globals.max_retrans_path)
219#define sctp_max_retrans_init (sctp_globals.max_retrans_init) 222#define sctp_max_retrans_init (sctp_globals.max_retrans_init)
223#define sctp_sack_timeout (sctp_globals.sack_timeout)
220#define sctp_hb_interval (sctp_globals.hb_interval) 224#define sctp_hb_interval (sctp_globals.hb_interval)
221#define sctp_max_instreams (sctp_globals.max_instreams) 225#define sctp_max_instreams (sctp_globals.max_instreams)
222#define sctp_max_outstreams (sctp_globals.max_outstreams) 226#define sctp_max_outstreams (sctp_globals.max_outstreams)
@@ -441,7 +445,8 @@ struct sctp_ssnmap {
441 int malloced; 445 int malloced;
442}; 446};
443 447
444struct sctp_ssnmap *sctp_ssnmap_new(__u16 in, __u16 out, int gfp); 448struct sctp_ssnmap *sctp_ssnmap_new(__u16 in, __u16 out,
449 unsigned int __nocast gfp);
445void sctp_ssnmap_free(struct sctp_ssnmap *map); 450void sctp_ssnmap_free(struct sctp_ssnmap *map);
446void sctp_ssnmap_clear(struct sctp_ssnmap *map); 451void sctp_ssnmap_clear(struct sctp_ssnmap *map);
447 452
@@ -578,7 +583,6 @@ void sctp_datamsg_track(struct sctp_chunk *);
578void sctp_chunk_fail(struct sctp_chunk *, int error); 583void sctp_chunk_fail(struct sctp_chunk *, int error);
579int sctp_chunk_abandoned(struct sctp_chunk *); 584int sctp_chunk_abandoned(struct sctp_chunk *);
580 585
581
582/* RFC2960 1.4 Key Terms 586/* RFC2960 1.4 Key Terms
583 * 587 *
584 * o Chunk: A unit of information within an SCTP packet, consisting of 588 * o Chunk: A unit of information within an SCTP packet, consisting of
@@ -588,13 +592,8 @@ int sctp_chunk_abandoned(struct sctp_chunk *);
588 * each chunk as well as a few other header pointers... 592 * each chunk as well as a few other header pointers...
589 */ 593 */
590struct sctp_chunk { 594struct sctp_chunk {
591 /* These first three elements MUST PRECISELY match the first 595 struct list_head list;
592 * three elements of struct sk_buff. This allows us to reuse 596
593 * all the skb_* queue management functions.
594 */
595 struct sctp_chunk *next;
596 struct sctp_chunk *prev;
597 struct sk_buff_head *list;
598 atomic_t refcnt; 597 atomic_t refcnt;
599 598
600 /* This is our link to the per-transport transmitted list. */ 599 /* This is our link to the per-transport transmitted list. */
@@ -713,7 +712,7 @@ struct sctp_packet {
713 __u32 vtag; 712 __u32 vtag;
714 713
715 /* This contains the payload chunks. */ 714 /* This contains the payload chunks. */
716 struct sk_buff_head chunks; 715 struct list_head chunk_list;
717 716
718 /* This is the overhead of the sctp and ip headers. */ 717 /* This is the overhead of the sctp and ip headers. */
719 size_t overhead; 718 size_t overhead;
@@ -867,10 +866,13 @@ struct sctp_transport {
867 */ 866 */
868 unsigned long last_time_ecne_reduced; 867 unsigned long last_time_ecne_reduced;
869 868
870 /* active : The current active state of this destination, 869 /* The number of times INIT has been sent on this transport. */
871 * : i.e. DOWN, UP, etc. 870 int init_sent_count;
871
872 /* state : The current state of this destination,
873 * : i.e. SCTP_ACTIVE, SCTP_INACTIVE, SCTP_UNKOWN.
872 */ 874 */
873 int active; 875 int state;
874 876
875 /* hb_allowed : The current heartbeat state of this destination, 877 /* hb_allowed : The current heartbeat state of this destination,
876 * : i.e. ALLOW-HB, NO-HEARTBEAT, etc. 878 * : i.e. ALLOW-HB, NO-HEARTBEAT, etc.
@@ -944,7 +946,8 @@ struct sctp_transport {
944 } cacc; 946 } cacc;
945}; 947};
946 948
947struct sctp_transport *sctp_transport_new(const union sctp_addr *, int); 949struct sctp_transport *sctp_transport_new(const union sctp_addr *,
950 unsigned int __nocast);
948void sctp_transport_set_owner(struct sctp_transport *, 951void sctp_transport_set_owner(struct sctp_transport *,
949 struct sctp_association *); 952 struct sctp_association *);
950void sctp_transport_route(struct sctp_transport *, union sctp_addr *, 953void sctp_transport_route(struct sctp_transport *, union sctp_addr *,
@@ -967,7 +970,7 @@ struct sctp_inq {
967 /* This is actually a queue of sctp_chunk each 970 /* This is actually a queue of sctp_chunk each
968 * containing a partially decoded packet. 971 * containing a partially decoded packet.
969 */ 972 */
970 struct sk_buff_head in; 973 struct list_head in_chunk_list;
971 /* This is the packet which is currently off the in queue and is 974 /* This is the packet which is currently off the in queue and is
972 * being worked on through the inbound chunk processing. 975 * being worked on through the inbound chunk processing.
973 */ 976 */
@@ -1010,7 +1013,7 @@ struct sctp_outq {
1010 struct sctp_association *asoc; 1013 struct sctp_association *asoc;
1011 1014
1012 /* Data pending that has never been transmitted. */ 1015 /* Data pending that has never been transmitted. */
1013 struct sk_buff_head out; 1016 struct list_head out_chunk_list;
1014 1017
1015 unsigned out_qlen; /* Total length of queued data chunks. */ 1018 unsigned out_qlen; /* Total length of queued data chunks. */
1016 1019
@@ -1018,7 +1021,7 @@ struct sctp_outq {
1018 unsigned error; 1021 unsigned error;
1019 1022
1020 /* These are control chunks we want to send. */ 1023 /* These are control chunks we want to send. */
1021 struct sk_buff_head control; 1024 struct list_head control_chunk_list;
1022 1025
1023 /* These are chunks that have been sacked but are above the 1026 /* These are chunks that have been sacked but are above the
1024 * CTSN, or cumulative tsn ack point. 1027 * CTSN, or cumulative tsn ack point.
@@ -1092,9 +1095,10 @@ void sctp_bind_addr_init(struct sctp_bind_addr *, __u16 port);
1092void sctp_bind_addr_free(struct sctp_bind_addr *); 1095void sctp_bind_addr_free(struct sctp_bind_addr *);
1093int sctp_bind_addr_copy(struct sctp_bind_addr *dest, 1096int sctp_bind_addr_copy(struct sctp_bind_addr *dest,
1094 const struct sctp_bind_addr *src, 1097 const struct sctp_bind_addr *src,
1095 sctp_scope_t scope, int gfp,int flags); 1098 sctp_scope_t scope, unsigned int __nocast gfp,
1099 int flags);
1096int sctp_add_bind_addr(struct sctp_bind_addr *, union sctp_addr *, 1100int sctp_add_bind_addr(struct sctp_bind_addr *, union sctp_addr *,
1097 int gfp); 1101 unsigned int __nocast gfp);
1098int sctp_del_bind_addr(struct sctp_bind_addr *, union sctp_addr *); 1102int sctp_del_bind_addr(struct sctp_bind_addr *, union sctp_addr *);
1099int sctp_bind_addr_match(struct sctp_bind_addr *, const union sctp_addr *, 1103int sctp_bind_addr_match(struct sctp_bind_addr *, const union sctp_addr *,
1100 struct sctp_sock *); 1104 struct sctp_sock *);
@@ -1103,9 +1107,10 @@ union sctp_addr *sctp_find_unmatch_addr(struct sctp_bind_addr *bp,
1103 int addrcnt, 1107 int addrcnt,
1104 struct sctp_sock *opt); 1108 struct sctp_sock *opt);
1105union sctp_params sctp_bind_addrs_to_raw(const struct sctp_bind_addr *bp, 1109union sctp_params sctp_bind_addrs_to_raw(const struct sctp_bind_addr *bp,
1106 int *addrs_len, int gfp); 1110 int *addrs_len,
1111 unsigned int __nocast gfp);
1107int sctp_raw_to_bind_addrs(struct sctp_bind_addr *bp, __u8 *raw, int len, 1112int sctp_raw_to_bind_addrs(struct sctp_bind_addr *bp, __u8 *raw, int len,
1108 __u16 port, int gfp); 1113 __u16 port, unsigned int __nocast gfp);
1109 1114
1110sctp_scope_t sctp_scope(const union sctp_addr *); 1115sctp_scope_t sctp_scope(const union sctp_addr *);
1111int sctp_in_scope(const union sctp_addr *addr, const sctp_scope_t scope); 1116int sctp_in_scope(const union sctp_addr *addr, const sctp_scope_t scope);
@@ -1222,9 +1227,6 @@ struct sctp_endpoint {
1222 1227
1223 /* sendbuf acct. policy. */ 1228 /* sendbuf acct. policy. */
1224 __u32 sndbuf_policy; 1229 __u32 sndbuf_policy;
1225
1226 /* Name for debugging output... */
1227 char *debug_name;
1228}; 1230};
1229 1231
1230/* Recover the outter endpoint structure. */ 1232/* Recover the outter endpoint structure. */
@@ -1237,7 +1239,7 @@ static inline struct sctp_endpoint *sctp_ep(struct sctp_ep_common *base)
1237} 1239}
1238 1240
1239/* These are function signatures for manipulating endpoints. */ 1241/* These are function signatures for manipulating endpoints. */
1240struct sctp_endpoint *sctp_endpoint_new(struct sock *, int); 1242struct sctp_endpoint *sctp_endpoint_new(struct sock *, unsigned int __nocast);
1241void sctp_endpoint_free(struct sctp_endpoint *); 1243void sctp_endpoint_free(struct sctp_endpoint *);
1242void sctp_endpoint_put(struct sctp_endpoint *); 1244void sctp_endpoint_put(struct sctp_endpoint *);
1243void sctp_endpoint_hold(struct sctp_endpoint *); 1245void sctp_endpoint_hold(struct sctp_endpoint *);
@@ -1258,7 +1260,7 @@ int sctp_verify_init(const struct sctp_association *asoc, sctp_cid_t,
1258 struct sctp_chunk **err_chunk); 1260 struct sctp_chunk **err_chunk);
1259int sctp_process_init(struct sctp_association *, sctp_cid_t cid, 1261int sctp_process_init(struct sctp_association *, sctp_cid_t cid,
1260 const union sctp_addr *peer, 1262 const union sctp_addr *peer,
1261 sctp_init_chunk_t *init, int gfp); 1263 sctp_init_chunk_t *init, unsigned int __nocast gfp);
1262__u32 sctp_generate_tag(const struct sctp_endpoint *); 1264__u32 sctp_generate_tag(const struct sctp_endpoint *);
1263__u32 sctp_generate_tsn(const struct sctp_endpoint *); 1265__u32 sctp_generate_tsn(const struct sctp_endpoint *);
1264 1266
@@ -1314,11 +1316,23 @@ struct sctp_association {
1314 * : association. Normally this information is 1316 * : association. Normally this information is
1315 * : hashed or keyed for quick lookup and access 1317 * : hashed or keyed for quick lookup and access
1316 * : of the TCB. 1318 * : of the TCB.
1319 * : The list is also initialized with the list
1320 * : of addresses passed with the sctp_connectx()
1321 * : call.
1317 * 1322 *
1318 * It is a list of SCTP_transport's. 1323 * It is a list of SCTP_transport's.
1319 */ 1324 */
1320 struct list_head transport_addr_list; 1325 struct list_head transport_addr_list;
1321 1326
1327 /* transport_count
1328 *
1329 * Peer : A count of the number of peer addresses
1330 * Transport : in the Peer Transport Address List.
1331 * Address :
1332 * Count :
1333 */
1334 __u16 transport_count;
1335
1322 /* port 1336 /* port
1323 * The transport layer port number. 1337 * The transport layer port number.
1324 */ 1338 */
@@ -1486,6 +1500,9 @@ struct sctp_association {
1486 /* Transport to which SHUTDOWN chunk was last sent. */ 1500 /* Transport to which SHUTDOWN chunk was last sent. */
1487 struct sctp_transport *shutdown_last_sent_to; 1501 struct sctp_transport *shutdown_last_sent_to;
1488 1502
1503 /* Transport to which INIT chunk was last sent. */
1504 struct sctp_transport *init_last_sent_to;
1505
1489 /* Next TSN : The next TSN number to be assigned to a new 1506 /* Next TSN : The next TSN number to be assigned to a new
1490 * : DATA chunk. This is sent in the INIT or INIT 1507 * : DATA chunk. This is sent in the INIT or INIT
1491 * : ACK chunk to the peer and incremented each 1508 * : ACK chunk to the peer and incremented each
@@ -1549,8 +1566,11 @@ struct sctp_association {
1549 /* The message size at which SCTP fragmentation will occur. */ 1566 /* The message size at which SCTP fragmentation will occur. */
1550 __u32 frag_point; 1567 __u32 frag_point;
1551 1568
1552 /* Currently only one counter is used to count INIT errors. */ 1569 /* Counter used to count INIT errors. */
1553 int counters[SCTP_NUMBER_COUNTERS]; 1570 int init_err_counter;
1571
1572 /* Count the number of INIT cycles (for doubling timeout). */
1573 int init_cycle;
1554 1574
1555 /* Default send parameters. */ 1575 /* Default send parameters. */
1556 __u16 default_stream; 1576 __u16 default_stream;
@@ -1650,7 +1670,7 @@ struct sctp_association {
1650 * which already resides in sctp_outq. Please move this 1670 * which already resides in sctp_outq. Please move this
1651 * queue and its supporting logic down there. --piggy] 1671 * queue and its supporting logic down there. --piggy]
1652 */ 1672 */
1653 struct sk_buff_head addip_chunks; 1673 struct list_head addip_chunk_list;
1654 1674
1655 /* ADDIP Section 4.1 ASCONF Chunk Procedures 1675 /* ADDIP Section 4.1 ASCONF Chunk Procedures
1656 * 1676 *
@@ -1703,11 +1723,13 @@ static inline struct sctp_association *sctp_assoc(struct sctp_ep_common *base)
1703 1723
1704struct sctp_association * 1724struct sctp_association *
1705sctp_association_new(const struct sctp_endpoint *, const struct sock *, 1725sctp_association_new(const struct sctp_endpoint *, const struct sock *,
1706 sctp_scope_t scope, int gfp); 1726 sctp_scope_t scope, unsigned int __nocast gfp);
1707void sctp_association_free(struct sctp_association *); 1727void sctp_association_free(struct sctp_association *);
1708void sctp_association_put(struct sctp_association *); 1728void sctp_association_put(struct sctp_association *);
1709void sctp_association_hold(struct sctp_association *); 1729void sctp_association_hold(struct sctp_association *);
1710 1730
1731struct sctp_transport *sctp_assoc_choose_init_transport(
1732 struct sctp_association *);
1711struct sctp_transport *sctp_assoc_choose_shutdown_transport( 1733struct sctp_transport *sctp_assoc_choose_shutdown_transport(
1712 struct sctp_association *); 1734 struct sctp_association *);
1713void sctp_assoc_update_retran_path(struct sctp_association *); 1735void sctp_assoc_update_retran_path(struct sctp_association *);
@@ -1717,9 +1739,12 @@ int sctp_assoc_lookup_laddr(struct sctp_association *asoc,
1717 const union sctp_addr *laddr); 1739 const union sctp_addr *laddr);
1718struct sctp_transport *sctp_assoc_add_peer(struct sctp_association *, 1740struct sctp_transport *sctp_assoc_add_peer(struct sctp_association *,
1719 const union sctp_addr *address, 1741 const union sctp_addr *address,
1720 const int gfp); 1742 const unsigned int __nocast gfp,
1743 const int peer_state);
1721void sctp_assoc_del_peer(struct sctp_association *asoc, 1744void sctp_assoc_del_peer(struct sctp_association *asoc,
1722 const union sctp_addr *addr); 1745 const union sctp_addr *addr);
1746void sctp_assoc_rm_peer(struct sctp_association *asoc,
1747 struct sctp_transport *peer);
1723void sctp_assoc_control_transport(struct sctp_association *, 1748void sctp_assoc_control_transport(struct sctp_association *,
1724 struct sctp_transport *, 1749 struct sctp_transport *,
1725 sctp_transport_cmd_t, sctp_sn_error_t); 1750 sctp_transport_cmd_t, sctp_sn_error_t);
@@ -1738,9 +1763,11 @@ void sctp_assoc_rwnd_increase(struct sctp_association *, unsigned);
1738void sctp_assoc_rwnd_decrease(struct sctp_association *, unsigned); 1763void sctp_assoc_rwnd_decrease(struct sctp_association *, unsigned);
1739void sctp_assoc_set_primary(struct sctp_association *, 1764void sctp_assoc_set_primary(struct sctp_association *,
1740 struct sctp_transport *); 1765 struct sctp_transport *);
1741int sctp_assoc_set_bind_addr_from_ep(struct sctp_association *, int); 1766int sctp_assoc_set_bind_addr_from_ep(struct sctp_association *,
1767 unsigned int __nocast);
1742int sctp_assoc_set_bind_addr_from_cookie(struct sctp_association *, 1768int sctp_assoc_set_bind_addr_from_cookie(struct sctp_association *,
1743 struct sctp_cookie*, int gfp); 1769 struct sctp_cookie*,
1770 unsigned int __nocast gfp);
1744 1771
1745int sctp_cmp_addr_exact(const union sctp_addr *ss1, 1772int sctp_cmp_addr_exact(const union sctp_addr *ss1,
1746 const union sctp_addr *ss2); 1773 const union sctp_addr *ss2);
diff --git a/include/net/sctp/ulpevent.h b/include/net/sctp/ulpevent.h
index 1019d83a580a..90fe4bf6754f 100644
--- a/include/net/sctp/ulpevent.h
+++ b/include/net/sctp/ulpevent.h
@@ -88,7 +88,7 @@ struct sctp_ulpevent *sctp_ulpevent_make_assoc_change(
88 __u16 error, 88 __u16 error,
89 __u16 outbound, 89 __u16 outbound,
90 __u16 inbound, 90 __u16 inbound,
91 int gfp); 91 unsigned int __nocast gfp);
92 92
93struct sctp_ulpevent *sctp_ulpevent_make_peer_addr_change( 93struct sctp_ulpevent *sctp_ulpevent_make_peer_addr_change(
94 const struct sctp_association *asoc, 94 const struct sctp_association *asoc,
@@ -96,35 +96,35 @@ struct sctp_ulpevent *sctp_ulpevent_make_peer_addr_change(
96 int flags, 96 int flags,
97 int state, 97 int state,
98 int error, 98 int error,
99 int gfp); 99 unsigned int __nocast gfp);
100 100
101struct sctp_ulpevent *sctp_ulpevent_make_remote_error( 101struct sctp_ulpevent *sctp_ulpevent_make_remote_error(
102 const struct sctp_association *asoc, 102 const struct sctp_association *asoc,
103 struct sctp_chunk *chunk, 103 struct sctp_chunk *chunk,
104 __u16 flags, 104 __u16 flags,
105 int gfp); 105 unsigned int __nocast gfp);
106struct sctp_ulpevent *sctp_ulpevent_make_send_failed( 106struct sctp_ulpevent *sctp_ulpevent_make_send_failed(
107 const struct sctp_association *asoc, 107 const struct sctp_association *asoc,
108 struct sctp_chunk *chunk, 108 struct sctp_chunk *chunk,
109 __u16 flags, 109 __u16 flags,
110 __u32 error, 110 __u32 error,
111 int gfp); 111 unsigned int __nocast gfp);
112 112
113struct sctp_ulpevent *sctp_ulpevent_make_shutdown_event( 113struct sctp_ulpevent *sctp_ulpevent_make_shutdown_event(
114 const struct sctp_association *asoc, 114 const struct sctp_association *asoc,
115 __u16 flags, 115 __u16 flags,
116 int gfp); 116 unsigned int __nocast gfp);
117 117
118struct sctp_ulpevent *sctp_ulpevent_make_pdapi( 118struct sctp_ulpevent *sctp_ulpevent_make_pdapi(
119 const struct sctp_association *asoc, 119 const struct sctp_association *asoc,
120 __u32 indication, int gfp); 120 __u32 indication, unsigned int __nocast gfp);
121 121
122struct sctp_ulpevent *sctp_ulpevent_make_adaption_indication( 122struct sctp_ulpevent *sctp_ulpevent_make_adaption_indication(
123 const struct sctp_association *asoc, int gfp); 123 const struct sctp_association *asoc, unsigned int __nocast gfp);
124 124
125struct sctp_ulpevent *sctp_ulpevent_make_rcvmsg(struct sctp_association *asoc, 125struct sctp_ulpevent *sctp_ulpevent_make_rcvmsg(struct sctp_association *asoc,
126 struct sctp_chunk *chunk, 126 struct sctp_chunk *chunk,
127 int gfp); 127 unsigned int __nocast gfp);
128 128
129void sctp_ulpevent_read_sndrcvinfo(const struct sctp_ulpevent *event, 129void sctp_ulpevent_read_sndrcvinfo(const struct sctp_ulpevent *event,
130 struct msghdr *); 130 struct msghdr *);
diff --git a/include/net/sctp/ulpqueue.h b/include/net/sctp/ulpqueue.h
index 961736d29d21..1a60c6d943c1 100644
--- a/include/net/sctp/ulpqueue.h
+++ b/include/net/sctp/ulpqueue.h
@@ -62,19 +62,22 @@ struct sctp_ulpq *sctp_ulpq_init(struct sctp_ulpq *,
62void sctp_ulpq_free(struct sctp_ulpq *); 62void sctp_ulpq_free(struct sctp_ulpq *);
63 63
64/* Add a new DATA chunk for processing. */ 64/* Add a new DATA chunk for processing. */
65int sctp_ulpq_tail_data(struct sctp_ulpq *, struct sctp_chunk *, int); 65int sctp_ulpq_tail_data(struct sctp_ulpq *, struct sctp_chunk *,
66 unsigned int __nocast);
66 67
67/* Add a new event for propagation to the ULP. */ 68/* Add a new event for propagation to the ULP. */
68int sctp_ulpq_tail_event(struct sctp_ulpq *, struct sctp_ulpevent *ev); 69int sctp_ulpq_tail_event(struct sctp_ulpq *, struct sctp_ulpevent *ev);
69 70
70/* Renege previously received chunks. */ 71/* Renege previously received chunks. */
71void sctp_ulpq_renege(struct sctp_ulpq *, struct sctp_chunk *, int); 72void sctp_ulpq_renege(struct sctp_ulpq *, struct sctp_chunk *,
73 unsigned int __nocast);
72 74
73/* Perform partial delivery. */ 75/* Perform partial delivery. */
74void sctp_ulpq_partial_delivery(struct sctp_ulpq *, struct sctp_chunk *, int); 76void sctp_ulpq_partial_delivery(struct sctp_ulpq *, struct sctp_chunk *,
77 unsigned int __nocast);
75 78
76/* Abort the partial delivery. */ 79/* Abort the partial delivery. */
77void sctp_ulpq_abort_pd(struct sctp_ulpq *, int); 80void sctp_ulpq_abort_pd(struct sctp_ulpq *, unsigned int __nocast);
78 81
79/* Clear the partial data delivery condition on this socket. */ 82/* Clear the partial data delivery condition on this socket. */
80int sctp_clear_pd(struct sock *sk); 83int sctp_clear_pd(struct sock *sk);
diff --git a/include/net/sctp/user.h b/include/net/sctp/user.h
index 2758e8ce4f25..f6328aeddcce 100644
--- a/include/net/sctp/user.h
+++ b/include/net/sctp/user.h
@@ -111,6 +111,8 @@ enum sctp_optname {
111#define SCTP_GET_LOCAL_ADDRS_NUM SCTP_GET_LOCAL_ADDRS_NUM 111#define SCTP_GET_LOCAL_ADDRS_NUM SCTP_GET_LOCAL_ADDRS_NUM
112 SCTP_GET_LOCAL_ADDRS, /* Get all local addresss. */ 112 SCTP_GET_LOCAL_ADDRS, /* Get all local addresss. */
113#define SCTP_GET_LOCAL_ADDRS SCTP_GET_LOCAL_ADDRS 113#define SCTP_GET_LOCAL_ADDRS SCTP_GET_LOCAL_ADDRS
114 SCTP_SOCKOPT_CONNECTX, /* CONNECTX requests. */
115#define SCTP_SOCKOPT_CONNECTX SCTP_SOCKOPT_CONNECTX
114}; 116};
115 117
116/* 118/*
@@ -527,6 +529,7 @@ struct sctp_paddrinfo {
527enum sctp_spinfo_state { 529enum sctp_spinfo_state {
528 SCTP_INACTIVE, 530 SCTP_INACTIVE,
529 SCTP_ACTIVE, 531 SCTP_ACTIVE,
532 SCTP_UNKNOWN = 0xffff /* Value used for transport state unknown */
530}; 533};
531 534
532/* 535/*
diff --git a/include/net/slhc_vj.h b/include/net/slhc_vj.h
index 0b2c2784f333..8716d5942b65 100644
--- a/include/net/slhc_vj.h
+++ b/include/net/slhc_vj.h
@@ -170,19 +170,14 @@ struct slcompress {
170}; 170};
171#define NULLSLCOMPR (struct slcompress *)0 171#define NULLSLCOMPR (struct slcompress *)0
172 172
173#define __ARGS(x) x
174
175/* In slhc.c: */ 173/* In slhc.c: */
176struct slcompress *slhc_init __ARGS((int rslots, int tslots)); 174struct slcompress *slhc_init(int rslots, int tslots);
177void slhc_free __ARGS((struct slcompress *comp)); 175void slhc_free(struct slcompress *comp);
178 176
179int slhc_compress __ARGS((struct slcompress *comp, unsigned char *icp, 177int slhc_compress(struct slcompress *comp, unsigned char *icp, int isize,
180 int isize, unsigned char *ocp, unsigned char **cpp, 178 unsigned char *ocp, unsigned char **cpp, int compress_cid);
181 int compress_cid)); 179int slhc_uncompress(struct slcompress *comp, unsigned char *icp, int isize);
182int slhc_uncompress __ARGS((struct slcompress *comp, unsigned char *icp, 180int slhc_remember(struct slcompress *comp, unsigned char *icp, int isize);
183 int isize)); 181int slhc_toss(struct slcompress *comp);
184int slhc_remember __ARGS((struct slcompress *comp, unsigned char *icp,
185 int isize));
186int slhc_toss __ARGS((struct slcompress *comp));
187 182
188#endif /* _SLHC_H */ 183#endif /* _SLHC_H */
diff --git a/include/net/snmp.h b/include/net/snmp.h
index a15ab256276e..a36bed8ea210 100644
--- a/include/net/snmp.h
+++ b/include/net/snmp.h
@@ -128,18 +128,18 @@ struct linux_mib {
128#define SNMP_STAT_USRPTR(name) (name[1]) 128#define SNMP_STAT_USRPTR(name) (name[1])
129 129
130#define SNMP_INC_STATS_BH(mib, field) \ 130#define SNMP_INC_STATS_BH(mib, field) \
131 (per_cpu_ptr(mib[0], _smp_processor_id())->mibs[field]++) 131 (per_cpu_ptr(mib[0], raw_smp_processor_id())->mibs[field]++)
132#define SNMP_INC_STATS_OFFSET_BH(mib, field, offset) \ 132#define SNMP_INC_STATS_OFFSET_BH(mib, field, offset) \
133 (per_cpu_ptr(mib[0], _smp_processor_id())->mibs[field + (offset)]++) 133 (per_cpu_ptr(mib[0], raw_smp_processor_id())->mibs[field + (offset)]++)
134#define SNMP_INC_STATS_USER(mib, field) \ 134#define SNMP_INC_STATS_USER(mib, field) \
135 (per_cpu_ptr(mib[1], _smp_processor_id())->mibs[field]++) 135 (per_cpu_ptr(mib[1], raw_smp_processor_id())->mibs[field]++)
136#define SNMP_INC_STATS(mib, field) \ 136#define SNMP_INC_STATS(mib, field) \
137 (per_cpu_ptr(mib[!in_softirq()], _smp_processor_id())->mibs[field]++) 137 (per_cpu_ptr(mib[!in_softirq()], raw_smp_processor_id())->mibs[field]++)
138#define SNMP_DEC_STATS(mib, field) \ 138#define SNMP_DEC_STATS(mib, field) \
139 (per_cpu_ptr(mib[!in_softirq()], _smp_processor_id())->mibs[field]--) 139 (per_cpu_ptr(mib[!in_softirq()], raw_smp_processor_id())->mibs[field]--)
140#define SNMP_ADD_STATS_BH(mib, field, addend) \ 140#define SNMP_ADD_STATS_BH(mib, field, addend) \
141 (per_cpu_ptr(mib[0], _smp_processor_id())->mibs[field] += addend) 141 (per_cpu_ptr(mib[0], raw_smp_processor_id())->mibs[field] += addend)
142#define SNMP_ADD_STATS_USER(mib, field, addend) \ 142#define SNMP_ADD_STATS_USER(mib, field, addend) \
143 (per_cpu_ptr(mib[1], _smp_processor_id())->mibs[field] += addend) 143 (per_cpu_ptr(mib[1], raw_smp_processor_id())->mibs[field] += addend)
144 144
145#endif 145#endif
diff --git a/include/net/sock.h b/include/net/sock.h
index a9ef3a6a13f3..a1042d08becd 100644
--- a/include/net/sock.h
+++ b/include/net/sock.h
@@ -484,6 +484,8 @@ extern void sk_stream_kill_queues(struct sock *sk);
484 484
485extern int sk_wait_data(struct sock *sk, long *timeo); 485extern int sk_wait_data(struct sock *sk, long *timeo);
486 486
487struct request_sock_ops;
488
487/* Networking protocol blocks we attach to sockets. 489/* Networking protocol blocks we attach to sockets.
488 * socket layer -> transport layer interface 490 * socket layer -> transport layer interface
489 * transport -> network interface is defined by struct inet_proto 491 * transport -> network interface is defined by struct inet_proto
@@ -547,6 +549,8 @@ struct proto {
547 kmem_cache_t *slab; 549 kmem_cache_t *slab;
548 unsigned int obj_size; 550 unsigned int obj_size;
549 551
552 struct request_sock_ops *rsk_prot;
553
550 struct module *owner; 554 struct module *owner;
551 555
552 char name[32]; 556 char name[32];
@@ -680,16 +684,17 @@ extern void FASTCALL(release_sock(struct sock *sk));
680#define bh_lock_sock(__sk) spin_lock(&((__sk)->sk_lock.slock)) 684#define bh_lock_sock(__sk) spin_lock(&((__sk)->sk_lock.slock))
681#define bh_unlock_sock(__sk) spin_unlock(&((__sk)->sk_lock.slock)) 685#define bh_unlock_sock(__sk) spin_unlock(&((__sk)->sk_lock.slock))
682 686
683extern struct sock *sk_alloc(int family, int priority, 687extern struct sock *sk_alloc(int family,
688 unsigned int __nocast priority,
684 struct proto *prot, int zero_it); 689 struct proto *prot, int zero_it);
685extern void sk_free(struct sock *sk); 690extern void sk_free(struct sock *sk);
686 691
687extern struct sk_buff *sock_wmalloc(struct sock *sk, 692extern struct sk_buff *sock_wmalloc(struct sock *sk,
688 unsigned long size, int force, 693 unsigned long size, int force,
689 int priority); 694 unsigned int __nocast priority);
690extern struct sk_buff *sock_rmalloc(struct sock *sk, 695extern struct sk_buff *sock_rmalloc(struct sock *sk,
691 unsigned long size, int force, 696 unsigned long size, int force,
692 int priority); 697 unsigned int __nocast priority);
693extern void sock_wfree(struct sk_buff *skb); 698extern void sock_wfree(struct sk_buff *skb);
694extern void sock_rfree(struct sk_buff *skb); 699extern void sock_rfree(struct sk_buff *skb);
695 700
@@ -704,7 +709,8 @@ extern struct sk_buff *sock_alloc_send_skb(struct sock *sk,
704 unsigned long size, 709 unsigned long size,
705 int noblock, 710 int noblock,
706 int *errcode); 711 int *errcode);
707extern void *sock_kmalloc(struct sock *sk, int size, int priority); 712extern void *sock_kmalloc(struct sock *sk, int size,
713 unsigned int __nocast priority);
708extern void sock_kfree_s(struct sock *sk, void *mem, int size); 714extern void sock_kfree_s(struct sock *sk, void *mem, int size);
709extern void sk_send_sigurg(struct sock *sk); 715extern void sk_send_sigurg(struct sock *sk);
710 716
@@ -1128,15 +1134,19 @@ static inline void sk_stream_moderate_sndbuf(struct sock *sk)
1128} 1134}
1129 1135
1130static inline struct sk_buff *sk_stream_alloc_pskb(struct sock *sk, 1136static inline struct sk_buff *sk_stream_alloc_pskb(struct sock *sk,
1131 int size, int mem, int gfp) 1137 int size, int mem,
1138 unsigned int __nocast gfp)
1132{ 1139{
1133 struct sk_buff *skb = alloc_skb(size + sk->sk_prot->max_header, gfp); 1140 struct sk_buff *skb;
1141 int hdr_len;
1134 1142
1143 hdr_len = SKB_DATA_ALIGN(sk->sk_prot->max_header);
1144 skb = alloc_skb(size + hdr_len, gfp);
1135 if (skb) { 1145 if (skb) {
1136 skb->truesize += mem; 1146 skb->truesize += mem;
1137 if (sk->sk_forward_alloc >= (int)skb->truesize || 1147 if (sk->sk_forward_alloc >= (int)skb->truesize ||
1138 sk_stream_mem_schedule(sk, skb->truesize, 0)) { 1148 sk_stream_mem_schedule(sk, skb->truesize, 0)) {
1139 skb_reserve(skb, sk->sk_prot->max_header); 1149 skb_reserve(skb, hdr_len);
1140 return skb; 1150 return skb;
1141 } 1151 }
1142 __kfree_skb(skb); 1152 __kfree_skb(skb);
@@ -1148,7 +1158,8 @@ static inline struct sk_buff *sk_stream_alloc_pskb(struct sock *sk,
1148} 1158}
1149 1159
1150static inline struct sk_buff *sk_stream_alloc_skb(struct sock *sk, 1160static inline struct sk_buff *sk_stream_alloc_skb(struct sock *sk,
1151 int size, int gfp) 1161 int size,
1162 unsigned int __nocast gfp)
1152{ 1163{
1153 return sk_stream_alloc_pskb(sk, size, 0, gfp); 1164 return sk_stream_alloc_pskb(sk, size, 0, gfp);
1154} 1165}
@@ -1181,7 +1192,7 @@ static inline int sock_writeable(const struct sock *sk)
1181 return atomic_read(&sk->sk_wmem_alloc) < (sk->sk_sndbuf / 2); 1192 return atomic_read(&sk->sk_wmem_alloc) < (sk->sk_sndbuf / 2);
1182} 1193}
1183 1194
1184static inline int gfp_any(void) 1195static inline unsigned int __nocast gfp_any(void)
1185{ 1196{
1186 return in_softirq() ? GFP_ATOMIC : GFP_KERNEL; 1197 return in_softirq() ? GFP_ATOMIC : GFP_KERNEL;
1187} 1198}
diff --git a/include/net/tcp.h b/include/net/tcp.h
index e71f8ba3e101..5010f0c5a56e 100644
--- a/include/net/tcp.h
+++ b/include/net/tcp.h
@@ -31,6 +31,7 @@
31#include <linux/cache.h> 31#include <linux/cache.h>
32#include <linux/percpu.h> 32#include <linux/percpu.h>
33#include <net/checksum.h> 33#include <net/checksum.h>
34#include <net/request_sock.h>
34#include <net/sock.h> 35#include <net/sock.h>
35#include <net/snmp.h> 36#include <net/snmp.h>
36#include <net/ip.h> 37#include <net/ip.h>
@@ -504,25 +505,6 @@ static __inline__ int tcp_sk_listen_hashfn(struct sock *sk)
504#else 505#else
505# define TCP_TW_RECYCLE_TICK (12+2-TCP_TW_RECYCLE_SLOTS_LOG) 506# define TCP_TW_RECYCLE_TICK (12+2-TCP_TW_RECYCLE_SLOTS_LOG)
506#endif 507#endif
507
508#define BICTCP_BETA_SCALE 1024 /* Scale factor beta calculation
509 * max_cwnd = snd_cwnd * beta
510 */
511#define BICTCP_MAX_INCREMENT 32 /*
512 * Limit on the amount of
513 * increment allowed during
514 * binary search.
515 */
516#define BICTCP_FUNC_OF_MIN_INCR 11 /*
517 * log(B/Smin)/log(B/(B-1))+1,
518 * Smin:min increment
519 * B:log factor
520 */
521#define BICTCP_B 4 /*
522 * In binary search,
523 * go to point (max+min)/N
524 */
525
526/* 508/*
527 * TCP option 509 * TCP option
528 */ 510 */
@@ -563,7 +545,6 @@ static __inline__ int tcp_sk_listen_hashfn(struct sock *sk)
563#define TCP_NAGLE_PUSH 4 /* Cork is overriden for already queued data */ 545#define TCP_NAGLE_PUSH 4 /* Cork is overriden for already queued data */
564 546
565/* sysctl variables for tcp */ 547/* sysctl variables for tcp */
566extern int sysctl_max_syn_backlog;
567extern int sysctl_tcp_timestamps; 548extern int sysctl_tcp_timestamps;
568extern int sysctl_tcp_window_scaling; 549extern int sysctl_tcp_window_scaling;
569extern int sysctl_tcp_sack; 550extern int sysctl_tcp_sack;
@@ -596,16 +577,7 @@ extern int sysctl_tcp_adv_win_scale;
596extern int sysctl_tcp_tw_reuse; 577extern int sysctl_tcp_tw_reuse;
597extern int sysctl_tcp_frto; 578extern int sysctl_tcp_frto;
598extern int sysctl_tcp_low_latency; 579extern int sysctl_tcp_low_latency;
599extern int sysctl_tcp_westwood;
600extern int sysctl_tcp_vegas_cong_avoid;
601extern int sysctl_tcp_vegas_alpha;
602extern int sysctl_tcp_vegas_beta;
603extern int sysctl_tcp_vegas_gamma;
604extern int sysctl_tcp_nometrics_save; 580extern int sysctl_tcp_nometrics_save;
605extern int sysctl_tcp_bic;
606extern int sysctl_tcp_bic_fast_convergence;
607extern int sysctl_tcp_bic_low_window;
608extern int sysctl_tcp_bic_beta;
609extern int sysctl_tcp_moderate_rcvbuf; 581extern int sysctl_tcp_moderate_rcvbuf;
610extern int sysctl_tcp_tso_win_divisor; 582extern int sysctl_tcp_tso_win_divisor;
611 583
@@ -613,74 +585,6 @@ extern atomic_t tcp_memory_allocated;
613extern atomic_t tcp_sockets_allocated; 585extern atomic_t tcp_sockets_allocated;
614extern int tcp_memory_pressure; 586extern int tcp_memory_pressure;
615 587
616struct open_request;
617
618struct or_calltable {
619 int family;
620 int (*rtx_syn_ack) (struct sock *sk, struct open_request *req, struct dst_entry*);
621 void (*send_ack) (struct sk_buff *skb, struct open_request *req);
622 void (*destructor) (struct open_request *req);
623 void (*send_reset) (struct sk_buff *skb);
624};
625
626struct tcp_v4_open_req {
627 __u32 loc_addr;
628 __u32 rmt_addr;
629 struct ip_options *opt;
630};
631
632#if defined(CONFIG_IPV6) || defined (CONFIG_IPV6_MODULE)
633struct tcp_v6_open_req {
634 struct in6_addr loc_addr;
635 struct in6_addr rmt_addr;
636 struct sk_buff *pktopts;
637 int iif;
638};
639#endif
640
641/* this structure is too big */
642struct open_request {
643 struct open_request *dl_next; /* Must be first member! */
644 __u32 rcv_isn;
645 __u32 snt_isn;
646 __u16 rmt_port;
647 __u16 mss;
648 __u8 retrans;
649 __u8 __pad;
650 __u16 snd_wscale : 4,
651 rcv_wscale : 4,
652 tstamp_ok : 1,
653 sack_ok : 1,
654 wscale_ok : 1,
655 ecn_ok : 1,
656 acked : 1;
657 /* The following two fields can be easily recomputed I think -AK */
658 __u32 window_clamp; /* window clamp at creation time */
659 __u32 rcv_wnd; /* rcv_wnd offered first time */
660 __u32 ts_recent;
661 unsigned long expires;
662 struct or_calltable *class;
663 struct sock *sk;
664 union {
665 struct tcp_v4_open_req v4_req;
666#if defined(CONFIG_IPV6) || defined (CONFIG_IPV6_MODULE)
667 struct tcp_v6_open_req v6_req;
668#endif
669 } af;
670};
671
672/* SLAB cache for open requests. */
673extern kmem_cache_t *tcp_openreq_cachep;
674
675#define tcp_openreq_alloc() kmem_cache_alloc(tcp_openreq_cachep, SLAB_ATOMIC)
676#define tcp_openreq_fastfree(req) kmem_cache_free(tcp_openreq_cachep, req)
677
678static inline void tcp_openreq_free(struct open_request *req)
679{
680 req->class->destructor(req);
681 tcp_openreq_fastfree(req);
682}
683
684#if defined(CONFIG_IPV6) || defined(CONFIG_IPV6_MODULE) 588#if defined(CONFIG_IPV6) || defined(CONFIG_IPV6_MODULE)
685#define TCP_INET_FAMILY(fam) ((fam) == AF_INET) 589#define TCP_INET_FAMILY(fam) ((fam) == AF_INET)
686#else 590#else
@@ -708,7 +612,7 @@ struct tcp_func {
708 612
709 struct sock * (*syn_recv_sock) (struct sock *sk, 613 struct sock * (*syn_recv_sock) (struct sock *sk,
710 struct sk_buff *skb, 614 struct sk_buff *skb,
711 struct open_request *req, 615 struct request_sock *req,
712 struct dst_entry *dst); 616 struct dst_entry *dst);
713 617
714 int (*remember_stamp) (struct sock *sk); 618 int (*remember_stamp) (struct sock *sk);
@@ -817,11 +721,16 @@ static inline int tcp_ack_scheduled(struct tcp_sock *tp)
817 return tp->ack.pending&TCP_ACK_SCHED; 721 return tp->ack.pending&TCP_ACK_SCHED;
818} 722}
819 723
820static __inline__ void tcp_dec_quickack_mode(struct tcp_sock *tp) 724static __inline__ void tcp_dec_quickack_mode(struct tcp_sock *tp, unsigned int pkts)
821{ 725{
822 if (tp->ack.quick && --tp->ack.quick == 0) { 726 if (tp->ack.quick) {
823 /* Leaving quickack mode we deflate ATO. */ 727 if (pkts >= tp->ack.quick) {
824 tp->ack.ato = TCP_ATO_MIN; 728 tp->ack.quick = 0;
729
730 /* Leaving quickack mode we deflate ATO. */
731 tp->ack.ato = TCP_ATO_MIN;
732 } else
733 tp->ack.quick -= pkts;
825 } 734 }
826} 735}
827 736
@@ -852,8 +761,8 @@ extern enum tcp_tw_status tcp_timewait_state_process(struct tcp_tw_bucket *tw,
852 unsigned len); 761 unsigned len);
853 762
854extern struct sock * tcp_check_req(struct sock *sk,struct sk_buff *skb, 763extern struct sock * tcp_check_req(struct sock *sk,struct sk_buff *skb,
855 struct open_request *req, 764 struct request_sock *req,
856 struct open_request **prev); 765 struct request_sock **prev);
857extern int tcp_child_process(struct sock *parent, 766extern int tcp_child_process(struct sock *parent,
858 struct sock *child, 767 struct sock *child,
859 struct sk_buff *skb); 768 struct sk_buff *skb);
@@ -903,12 +812,12 @@ extern int tcp_v4_conn_request(struct sock *sk,
903 struct sk_buff *skb); 812 struct sk_buff *skb);
904 813
905extern struct sock * tcp_create_openreq_child(struct sock *sk, 814extern struct sock * tcp_create_openreq_child(struct sock *sk,
906 struct open_request *req, 815 struct request_sock *req,
907 struct sk_buff *skb); 816 struct sk_buff *skb);
908 817
909extern struct sock * tcp_v4_syn_recv_sock(struct sock *sk, 818extern struct sock * tcp_v4_syn_recv_sock(struct sock *sk,
910 struct sk_buff *skb, 819 struct sk_buff *skb,
911 struct open_request *req, 820 struct request_sock *req,
912 struct dst_entry *dst); 821 struct dst_entry *dst);
913 822
914extern int tcp_v4_do_rcv(struct sock *sk, 823extern int tcp_v4_do_rcv(struct sock *sk,
@@ -922,7 +831,7 @@ extern int tcp_connect(struct sock *sk);
922 831
923extern struct sk_buff * tcp_make_synack(struct sock *sk, 832extern struct sk_buff * tcp_make_synack(struct sock *sk,
924 struct dst_entry *dst, 833 struct dst_entry *dst,
925 struct open_request *req); 834 struct request_sock *req);
926 835
927extern int tcp_disconnect(struct sock *sk, int flags); 836extern int tcp_disconnect(struct sock *sk, int flags);
928 837
@@ -939,7 +848,9 @@ extern __u32 cookie_v4_init_sequence(struct sock *sk, struct sk_buff *skb,
939 848
940/* tcp_output.c */ 849/* tcp_output.c */
941 850
942extern int tcp_write_xmit(struct sock *, int nonagle); 851extern void __tcp_push_pending_frames(struct sock *sk, struct tcp_sock *tp,
852 unsigned int cur_mss, int nonagle);
853extern int tcp_may_send_now(struct sock *sk, struct tcp_sock *tp);
943extern int tcp_retransmit_skb(struct sock *, struct sk_buff *); 854extern int tcp_retransmit_skb(struct sock *, struct sk_buff *);
944extern void tcp_xmit_retransmit_queue(struct sock *); 855extern void tcp_xmit_retransmit_queue(struct sock *);
945extern void tcp_simple_retransmit(struct sock *); 856extern void tcp_simple_retransmit(struct sock *);
@@ -949,12 +860,16 @@ extern void tcp_send_probe0(struct sock *);
949extern void tcp_send_partial(struct sock *); 860extern void tcp_send_partial(struct sock *);
950extern int tcp_write_wakeup(struct sock *); 861extern int tcp_write_wakeup(struct sock *);
951extern void tcp_send_fin(struct sock *sk); 862extern void tcp_send_fin(struct sock *sk);
952extern void tcp_send_active_reset(struct sock *sk, int priority); 863extern void tcp_send_active_reset(struct sock *sk,
864 unsigned int __nocast priority);
953extern int tcp_send_synack(struct sock *); 865extern int tcp_send_synack(struct sock *);
954extern void tcp_push_one(struct sock *, unsigned mss_now); 866extern void tcp_push_one(struct sock *, unsigned int mss_now);
955extern void tcp_send_ack(struct sock *sk); 867extern void tcp_send_ack(struct sock *sk);
956extern void tcp_send_delayed_ack(struct sock *sk); 868extern void tcp_send_delayed_ack(struct sock *sk);
957 869
870/* tcp_input.c */
871extern void tcp_cwnd_application_limited(struct sock *sk);
872
958/* tcp_timer.c */ 873/* tcp_timer.c */
959extern void tcp_init_xmit_timers(struct sock *); 874extern void tcp_init_xmit_timers(struct sock *);
960extern void tcp_clear_xmit_timers(struct sock *); 875extern void tcp_clear_xmit_timers(struct sock *);
@@ -1054,7 +969,7 @@ static inline void tcp_reset_xmit_timer(struct sock *sk, int what, unsigned long
1054static inline void tcp_initialize_rcv_mss(struct sock *sk) 969static inline void tcp_initialize_rcv_mss(struct sock *sk)
1055{ 970{
1056 struct tcp_sock *tp = tcp_sk(sk); 971 struct tcp_sock *tp = tcp_sk(sk);
1057 unsigned int hint = min(tp->advmss, tp->mss_cache_std); 972 unsigned int hint = min_t(unsigned int, tp->advmss, tp->mss_cache);
1058 973
1059 hint = min(hint, tp->rcv_wnd/2); 974 hint = min(hint, tp->rcv_wnd/2);
1060 hint = min(hint, TCP_MIN_RCVMSS); 975 hint = min(hint, TCP_MIN_RCVMSS);
@@ -1077,7 +992,7 @@ static __inline__ void tcp_fast_path_on(struct tcp_sock *tp)
1077 992
1078static inline void tcp_fast_path_check(struct sock *sk, struct tcp_sock *tp) 993static inline void tcp_fast_path_check(struct sock *sk, struct tcp_sock *tp)
1079{ 994{
1080 if (skb_queue_len(&tp->out_of_order_queue) == 0 && 995 if (skb_queue_empty(&tp->out_of_order_queue) &&
1081 tp->rcv_wnd && 996 tp->rcv_wnd &&
1082 atomic_read(&sk->sk_rmem_alloc) < sk->sk_rcvbuf && 997 atomic_read(&sk->sk_rmem_alloc) < sk->sk_rcvbuf &&
1083 !tp->urg_data) 998 !tp->urg_data)
@@ -1204,6 +1119,82 @@ static inline void tcp_packets_out_dec(struct tcp_sock *tp,
1204 tp->packets_out -= tcp_skb_pcount(skb); 1119 tp->packets_out -= tcp_skb_pcount(skb);
1205} 1120}
1206 1121
1122/* Events passed to congestion control interface */
1123enum tcp_ca_event {
1124 CA_EVENT_TX_START, /* first transmit when no packets in flight */
1125 CA_EVENT_CWND_RESTART, /* congestion window restart */
1126 CA_EVENT_COMPLETE_CWR, /* end of congestion recovery */
1127 CA_EVENT_FRTO, /* fast recovery timeout */
1128 CA_EVENT_LOSS, /* loss timeout */
1129 CA_EVENT_FAST_ACK, /* in sequence ack */
1130 CA_EVENT_SLOW_ACK, /* other ack */
1131};
1132
1133/*
1134 * Interface for adding new TCP congestion control handlers
1135 */
1136#define TCP_CA_NAME_MAX 16
1137struct tcp_congestion_ops {
1138 struct list_head list;
1139
1140 /* initialize private data (optional) */
1141 void (*init)(struct tcp_sock *tp);
1142 /* cleanup private data (optional) */
1143 void (*release)(struct tcp_sock *tp);
1144
1145 /* return slow start threshold (required) */
1146 u32 (*ssthresh)(struct tcp_sock *tp);
1147 /* lower bound for congestion window (optional) */
1148 u32 (*min_cwnd)(struct tcp_sock *tp);
1149 /* do new cwnd calculation (required) */
1150 void (*cong_avoid)(struct tcp_sock *tp, u32 ack,
1151 u32 rtt, u32 in_flight, int good_ack);
1152 /* round trip time sample per acked packet (optional) */
1153 void (*rtt_sample)(struct tcp_sock *tp, u32 usrtt);
1154 /* call before changing ca_state (optional) */
1155 void (*set_state)(struct tcp_sock *tp, u8 new_state);
1156 /* call when cwnd event occurs (optional) */
1157 void (*cwnd_event)(struct tcp_sock *tp, enum tcp_ca_event ev);
1158 /* new value of cwnd after loss (optional) */
1159 u32 (*undo_cwnd)(struct tcp_sock *tp);
1160 /* hook for packet ack accounting (optional) */
1161 void (*pkts_acked)(struct tcp_sock *tp, u32 num_acked);
1162 /* get info for tcp_diag (optional) */
1163 void (*get_info)(struct tcp_sock *tp, u32 ext, struct sk_buff *skb);
1164
1165 char name[TCP_CA_NAME_MAX];
1166 struct module *owner;
1167};
1168
1169extern int tcp_register_congestion_control(struct tcp_congestion_ops *type);
1170extern void tcp_unregister_congestion_control(struct tcp_congestion_ops *type);
1171
1172extern void tcp_init_congestion_control(struct tcp_sock *tp);
1173extern void tcp_cleanup_congestion_control(struct tcp_sock *tp);
1174extern int tcp_set_default_congestion_control(const char *name);
1175extern void tcp_get_default_congestion_control(char *name);
1176extern int tcp_set_congestion_control(struct tcp_sock *tp, const char *name);
1177
1178extern struct tcp_congestion_ops tcp_init_congestion_ops;
1179extern u32 tcp_reno_ssthresh(struct tcp_sock *tp);
1180extern void tcp_reno_cong_avoid(struct tcp_sock *tp, u32 ack,
1181 u32 rtt, u32 in_flight, int flag);
1182extern u32 tcp_reno_min_cwnd(struct tcp_sock *tp);
1183extern struct tcp_congestion_ops tcp_reno;
1184
1185static inline void tcp_set_ca_state(struct tcp_sock *tp, u8 ca_state)
1186{
1187 if (tp->ca_ops->set_state)
1188 tp->ca_ops->set_state(tp, ca_state);
1189 tp->ca_state = ca_state;
1190}
1191
1192static inline void tcp_ca_event(struct tcp_sock *tp, enum tcp_ca_event event)
1193{
1194 if (tp->ca_ops->cwnd_event)
1195 tp->ca_ops->cwnd_event(tp, event);
1196}
1197
1207/* This determines how many packets are "in the network" to the best 1198/* This determines how many packets are "in the network" to the best
1208 * of our knowledge. In many cases it is conservative, but where 1199 * of our knowledge. In many cases it is conservative, but where
1209 * detailed information is available from the receiver (via SACK 1200 * detailed information is available from the receiver (via SACK
@@ -1223,91 +1214,6 @@ static __inline__ unsigned int tcp_packets_in_flight(const struct tcp_sock *tp)
1223 return (tp->packets_out - tp->left_out + tp->retrans_out); 1214 return (tp->packets_out - tp->left_out + tp->retrans_out);
1224} 1215}
1225 1216
1226/*
1227 * Which congestion algorithim is in use on the connection.
1228 */
1229#define tcp_is_vegas(__tp) ((__tp)->adv_cong == TCP_VEGAS)
1230#define tcp_is_westwood(__tp) ((__tp)->adv_cong == TCP_WESTWOOD)
1231#define tcp_is_bic(__tp) ((__tp)->adv_cong == TCP_BIC)
1232
1233/* Recalculate snd_ssthresh, we want to set it to:
1234 *
1235 * Reno:
1236 * one half the current congestion window, but no
1237 * less than two segments
1238 *
1239 * BIC:
1240 * behave like Reno until low_window is reached,
1241 * then increase congestion window slowly
1242 */
1243static inline __u32 tcp_recalc_ssthresh(struct tcp_sock *tp)
1244{
1245 if (tcp_is_bic(tp)) {
1246 if (sysctl_tcp_bic_fast_convergence &&
1247 tp->snd_cwnd < tp->bictcp.last_max_cwnd)
1248 tp->bictcp.last_max_cwnd = (tp->snd_cwnd *
1249 (BICTCP_BETA_SCALE
1250 + sysctl_tcp_bic_beta))
1251 / (2 * BICTCP_BETA_SCALE);
1252 else
1253 tp->bictcp.last_max_cwnd = tp->snd_cwnd;
1254
1255 if (tp->snd_cwnd > sysctl_tcp_bic_low_window)
1256 return max((tp->snd_cwnd * sysctl_tcp_bic_beta)
1257 / BICTCP_BETA_SCALE, 2U);
1258 }
1259
1260 return max(tp->snd_cwnd >> 1U, 2U);
1261}
1262
1263/* Stop taking Vegas samples for now. */
1264#define tcp_vegas_disable(__tp) ((__tp)->vegas.doing_vegas_now = 0)
1265
1266static inline void tcp_vegas_enable(struct tcp_sock *tp)
1267{
1268 /* There are several situations when we must "re-start" Vegas:
1269 *
1270 * o when a connection is established
1271 * o after an RTO
1272 * o after fast recovery
1273 * o when we send a packet and there is no outstanding
1274 * unacknowledged data (restarting an idle connection)
1275 *
1276 * In these circumstances we cannot do a Vegas calculation at the
1277 * end of the first RTT, because any calculation we do is using
1278 * stale info -- both the saved cwnd and congestion feedback are
1279 * stale.
1280 *
1281 * Instead we must wait until the completion of an RTT during
1282 * which we actually receive ACKs.
1283 */
1284
1285 /* Begin taking Vegas samples next time we send something. */
1286 tp->vegas.doing_vegas_now = 1;
1287
1288 /* Set the beginning of the next send window. */
1289 tp->vegas.beg_snd_nxt = tp->snd_nxt;
1290
1291 tp->vegas.cntRTT = 0;
1292 tp->vegas.minRTT = 0x7fffffff;
1293}
1294
1295/* Should we be taking Vegas samples right now? */
1296#define tcp_vegas_enabled(__tp) ((__tp)->vegas.doing_vegas_now)
1297
1298extern void tcp_ca_init(struct tcp_sock *tp);
1299
1300static inline void tcp_set_ca_state(struct tcp_sock *tp, u8 ca_state)
1301{
1302 if (tcp_is_vegas(tp)) {
1303 if (ca_state == TCP_CA_Open)
1304 tcp_vegas_enable(tp);
1305 else
1306 tcp_vegas_disable(tp);
1307 }
1308 tp->ca_state = ca_state;
1309}
1310
1311/* If cwnd > ssthresh, we may raise ssthresh to be half-way to cwnd. 1217/* If cwnd > ssthresh, we may raise ssthresh to be half-way to cwnd.
1312 * The exception is rate halving phase, when cwnd is decreasing towards 1218 * The exception is rate halving phase, when cwnd is decreasing towards
1313 * ssthresh. 1219 * ssthresh.
@@ -1330,33 +1236,11 @@ static inline void tcp_sync_left_out(struct tcp_sock *tp)
1330 tp->left_out = tp->sacked_out + tp->lost_out; 1236 tp->left_out = tp->sacked_out + tp->lost_out;
1331} 1237}
1332 1238
1333extern void tcp_cwnd_application_limited(struct sock *sk); 1239/* Set slow start threshold and cwnd not falling to slow start */
1334
1335/* Congestion window validation. (RFC2861) */
1336
1337static inline void tcp_cwnd_validate(struct sock *sk, struct tcp_sock *tp)
1338{
1339 __u32 packets_out = tp->packets_out;
1340
1341 if (packets_out >= tp->snd_cwnd) {
1342 /* Network is feed fully. */
1343 tp->snd_cwnd_used = 0;
1344 tp->snd_cwnd_stamp = tcp_time_stamp;
1345 } else {
1346 /* Network starves. */
1347 if (tp->packets_out > tp->snd_cwnd_used)
1348 tp->snd_cwnd_used = tp->packets_out;
1349
1350 if ((s32)(tcp_time_stamp - tp->snd_cwnd_stamp) >= tp->rto)
1351 tcp_cwnd_application_limited(sk);
1352 }
1353}
1354
1355/* Set slow start threshould and cwnd not falling to slow start */
1356static inline void __tcp_enter_cwr(struct tcp_sock *tp) 1240static inline void __tcp_enter_cwr(struct tcp_sock *tp)
1357{ 1241{
1358 tp->undo_marker = 0; 1242 tp->undo_marker = 0;
1359 tp->snd_ssthresh = tcp_recalc_ssthresh(tp); 1243 tp->snd_ssthresh = tp->ca_ops->ssthresh(tp);
1360 tp->snd_cwnd = min(tp->snd_cwnd, 1244 tp->snd_cwnd = min(tp->snd_cwnd,
1361 tcp_packets_in_flight(tp) + 1U); 1245 tcp_packets_in_flight(tp) + 1U);
1362 tp->snd_cwnd_cnt = 0; 1246 tp->snd_cwnd_cnt = 0;
@@ -1384,12 +1268,6 @@ static __inline__ __u32 tcp_max_burst(const struct tcp_sock *tp)
1384 return 3; 1268 return 3;
1385} 1269}
1386 1270
1387static __inline__ int tcp_minshall_check(const struct tcp_sock *tp)
1388{
1389 return after(tp->snd_sml,tp->snd_una) &&
1390 !after(tp->snd_sml, tp->snd_nxt);
1391}
1392
1393static __inline__ void tcp_minshall_update(struct tcp_sock *tp, int mss, 1271static __inline__ void tcp_minshall_update(struct tcp_sock *tp, int mss,
1394 const struct sk_buff *skb) 1272 const struct sk_buff *skb)
1395{ 1273{
@@ -1397,122 +1275,18 @@ static __inline__ void tcp_minshall_update(struct tcp_sock *tp, int mss,
1397 tp->snd_sml = TCP_SKB_CB(skb)->end_seq; 1275 tp->snd_sml = TCP_SKB_CB(skb)->end_seq;
1398} 1276}
1399 1277
1400/* Return 0, if packet can be sent now without violation Nagle's rules:
1401 1. It is full sized.
1402 2. Or it contains FIN.
1403 3. Or TCP_NODELAY was set.
1404 4. Or TCP_CORK is not set, and all sent packets are ACKed.
1405 With Minshall's modification: all sent small packets are ACKed.
1406 */
1407
1408static __inline__ int
1409tcp_nagle_check(const struct tcp_sock *tp, const struct sk_buff *skb,
1410 unsigned mss_now, int nonagle)
1411{
1412 return (skb->len < mss_now &&
1413 !(TCP_SKB_CB(skb)->flags & TCPCB_FLAG_FIN) &&
1414 ((nonagle&TCP_NAGLE_CORK) ||
1415 (!nonagle &&
1416 tp->packets_out &&
1417 tcp_minshall_check(tp))));
1418}
1419
1420extern void tcp_set_skb_tso_segs(struct sock *, struct sk_buff *);
1421
1422/* This checks if the data bearing packet SKB (usually sk->sk_send_head)
1423 * should be put on the wire right now.
1424 */
1425static __inline__ int tcp_snd_test(struct sock *sk,
1426 struct sk_buff *skb,
1427 unsigned cur_mss, int nonagle)
1428{
1429 struct tcp_sock *tp = tcp_sk(sk);
1430 int pkts = tcp_skb_pcount(skb);
1431
1432 if (!pkts) {
1433 tcp_set_skb_tso_segs(sk, skb);
1434 pkts = tcp_skb_pcount(skb);
1435 }
1436
1437 /* RFC 1122 - section 4.2.3.4
1438 *
1439 * We must queue if
1440 *
1441 * a) The right edge of this frame exceeds the window
1442 * b) There are packets in flight and we have a small segment
1443 * [SWS avoidance and Nagle algorithm]
1444 * (part of SWS is done on packetization)
1445 * Minshall version sounds: there are no _small_
1446 * segments in flight. (tcp_nagle_check)
1447 * c) We have too many packets 'in flight'
1448 *
1449 * Don't use the nagle rule for urgent data (or
1450 * for the final FIN -DaveM).
1451 *
1452 * Also, Nagle rule does not apply to frames, which
1453 * sit in the middle of queue (they have no chances
1454 * to get new data) and if room at tail of skb is
1455 * not enough to save something seriously (<32 for now).
1456 */
1457
1458 /* Don't be strict about the congestion window for the
1459 * final FIN frame. -DaveM
1460 */
1461 return (((nonagle&TCP_NAGLE_PUSH) || tp->urg_mode
1462 || !tcp_nagle_check(tp, skb, cur_mss, nonagle)) &&
1463 (((tcp_packets_in_flight(tp) + (pkts-1)) < tp->snd_cwnd) ||
1464 (TCP_SKB_CB(skb)->flags & TCPCB_FLAG_FIN)) &&
1465 !after(TCP_SKB_CB(skb)->end_seq, tp->snd_una + tp->snd_wnd));
1466}
1467
1468static __inline__ void tcp_check_probe_timer(struct sock *sk, struct tcp_sock *tp) 1278static __inline__ void tcp_check_probe_timer(struct sock *sk, struct tcp_sock *tp)
1469{ 1279{
1470 if (!tp->packets_out && !tp->pending) 1280 if (!tp->packets_out && !tp->pending)
1471 tcp_reset_xmit_timer(sk, TCP_TIME_PROBE0, tp->rto); 1281 tcp_reset_xmit_timer(sk, TCP_TIME_PROBE0, tp->rto);
1472} 1282}
1473 1283
1474static __inline__ int tcp_skb_is_last(const struct sock *sk,
1475 const struct sk_buff *skb)
1476{
1477 return skb->next == (struct sk_buff *)&sk->sk_write_queue;
1478}
1479
1480/* Push out any pending frames which were held back due to
1481 * TCP_CORK or attempt at coalescing tiny packets.
1482 * The socket must be locked by the caller.
1483 */
1484static __inline__ void __tcp_push_pending_frames(struct sock *sk,
1485 struct tcp_sock *tp,
1486 unsigned cur_mss,
1487 int nonagle)
1488{
1489 struct sk_buff *skb = sk->sk_send_head;
1490
1491 if (skb) {
1492 if (!tcp_skb_is_last(sk, skb))
1493 nonagle = TCP_NAGLE_PUSH;
1494 if (!tcp_snd_test(sk, skb, cur_mss, nonagle) ||
1495 tcp_write_xmit(sk, nonagle))
1496 tcp_check_probe_timer(sk, tp);
1497 }
1498 tcp_cwnd_validate(sk, tp);
1499}
1500
1501static __inline__ void tcp_push_pending_frames(struct sock *sk, 1284static __inline__ void tcp_push_pending_frames(struct sock *sk,
1502 struct tcp_sock *tp) 1285 struct tcp_sock *tp)
1503{ 1286{
1504 __tcp_push_pending_frames(sk, tp, tcp_current_mss(sk, 1), tp->nonagle); 1287 __tcp_push_pending_frames(sk, tp, tcp_current_mss(sk, 1), tp->nonagle);
1505} 1288}
1506 1289
1507static __inline__ int tcp_may_send_now(struct sock *sk, struct tcp_sock *tp)
1508{
1509 struct sk_buff *skb = sk->sk_send_head;
1510
1511 return (skb &&
1512 tcp_snd_test(sk, skb, tcp_current_mss(sk, 1),
1513 tcp_skb_is_last(sk, skb) ? TCP_NAGLE_PUSH : tp->nonagle));
1514}
1515
1516static __inline__ void tcp_init_wl(struct tcp_sock *tp, u32 ack, u32 seq) 1290static __inline__ void tcp_init_wl(struct tcp_sock *tp, u32 ack, u32 seq)
1517{ 1291{
1518 tp->snd_wl1 = seq; 1292 tp->snd_wl1 = seq;
@@ -1750,99 +1524,71 @@ static inline int tcp_full_space(const struct sock *sk)
1750 return tcp_win_from_space(sk->sk_rcvbuf); 1524 return tcp_win_from_space(sk->sk_rcvbuf);
1751} 1525}
1752 1526
1753static inline void tcp_acceptq_queue(struct sock *sk, struct open_request *req, 1527static inline void tcp_acceptq_queue(struct sock *sk, struct request_sock *req,
1754 struct sock *child) 1528 struct sock *child)
1755{ 1529{
1756 struct tcp_sock *tp = tcp_sk(sk); 1530 reqsk_queue_add(&tcp_sk(sk)->accept_queue, req, sk, child);
1757
1758 req->sk = child;
1759 sk_acceptq_added(sk);
1760
1761 if (!tp->accept_queue_tail) {
1762 tp->accept_queue = req;
1763 } else {
1764 tp->accept_queue_tail->dl_next = req;
1765 }
1766 tp->accept_queue_tail = req;
1767 req->dl_next = NULL;
1768} 1531}
1769 1532
1770struct tcp_listen_opt
1771{
1772 u8 max_qlen_log; /* log_2 of maximal queued SYNs */
1773 int qlen;
1774 int qlen_young;
1775 int clock_hand;
1776 u32 hash_rnd;
1777 struct open_request *syn_table[TCP_SYNQ_HSIZE];
1778};
1779
1780static inline void 1533static inline void
1781tcp_synq_removed(struct sock *sk, struct open_request *req) 1534tcp_synq_removed(struct sock *sk, struct request_sock *req)
1782{ 1535{
1783 struct tcp_listen_opt *lopt = tcp_sk(sk)->listen_opt; 1536 if (reqsk_queue_removed(&tcp_sk(sk)->accept_queue, req) == 0)
1784
1785 if (--lopt->qlen == 0)
1786 tcp_delete_keepalive_timer(sk); 1537 tcp_delete_keepalive_timer(sk);
1787 if (req->retrans == 0)
1788 lopt->qlen_young--;
1789} 1538}
1790 1539
1791static inline void tcp_synq_added(struct sock *sk) 1540static inline void tcp_synq_added(struct sock *sk)
1792{ 1541{
1793 struct tcp_listen_opt *lopt = tcp_sk(sk)->listen_opt; 1542 if (reqsk_queue_added(&tcp_sk(sk)->accept_queue) == 0)
1794
1795 if (lopt->qlen++ == 0)
1796 tcp_reset_keepalive_timer(sk, TCP_TIMEOUT_INIT); 1543 tcp_reset_keepalive_timer(sk, TCP_TIMEOUT_INIT);
1797 lopt->qlen_young++;
1798} 1544}
1799 1545
1800static inline int tcp_synq_len(struct sock *sk) 1546static inline int tcp_synq_len(struct sock *sk)
1801{ 1547{
1802 return tcp_sk(sk)->listen_opt->qlen; 1548 return reqsk_queue_len(&tcp_sk(sk)->accept_queue);
1803} 1549}
1804 1550
1805static inline int tcp_synq_young(struct sock *sk) 1551static inline int tcp_synq_young(struct sock *sk)
1806{ 1552{
1807 return tcp_sk(sk)->listen_opt->qlen_young; 1553 return reqsk_queue_len_young(&tcp_sk(sk)->accept_queue);
1808} 1554}
1809 1555
1810static inline int tcp_synq_is_full(struct sock *sk) 1556static inline int tcp_synq_is_full(struct sock *sk)
1811{ 1557{
1812 return tcp_synq_len(sk) >> tcp_sk(sk)->listen_opt->max_qlen_log; 1558 return reqsk_queue_is_full(&tcp_sk(sk)->accept_queue);
1813} 1559}
1814 1560
1815static inline void tcp_synq_unlink(struct tcp_sock *tp, struct open_request *req, 1561static inline void tcp_synq_unlink(struct tcp_sock *tp, struct request_sock *req,
1816 struct open_request **prev) 1562 struct request_sock **prev)
1817{ 1563{
1818 write_lock(&tp->syn_wait_lock); 1564 reqsk_queue_unlink(&tp->accept_queue, req, prev);
1819 *prev = req->dl_next;
1820 write_unlock(&tp->syn_wait_lock);
1821} 1565}
1822 1566
1823static inline void tcp_synq_drop(struct sock *sk, struct open_request *req, 1567static inline void tcp_synq_drop(struct sock *sk, struct request_sock *req,
1824 struct open_request **prev) 1568 struct request_sock **prev)
1825{ 1569{
1826 tcp_synq_unlink(tcp_sk(sk), req, prev); 1570 tcp_synq_unlink(tcp_sk(sk), req, prev);
1827 tcp_synq_removed(sk, req); 1571 tcp_synq_removed(sk, req);
1828 tcp_openreq_free(req); 1572 reqsk_free(req);
1829} 1573}
1830 1574
1831static __inline__ void tcp_openreq_init(struct open_request *req, 1575static __inline__ void tcp_openreq_init(struct request_sock *req,
1832 struct tcp_options_received *rx_opt, 1576 struct tcp_options_received *rx_opt,
1833 struct sk_buff *skb) 1577 struct sk_buff *skb)
1834{ 1578{
1579 struct inet_request_sock *ireq = inet_rsk(req);
1580
1835 req->rcv_wnd = 0; /* So that tcp_send_synack() knows! */ 1581 req->rcv_wnd = 0; /* So that tcp_send_synack() knows! */
1836 req->rcv_isn = TCP_SKB_CB(skb)->seq; 1582 tcp_rsk(req)->rcv_isn = TCP_SKB_CB(skb)->seq;
1837 req->mss = rx_opt->mss_clamp; 1583 req->mss = rx_opt->mss_clamp;
1838 req->ts_recent = rx_opt->saw_tstamp ? rx_opt->rcv_tsval : 0; 1584 req->ts_recent = rx_opt->saw_tstamp ? rx_opt->rcv_tsval : 0;
1839 req->tstamp_ok = rx_opt->tstamp_ok; 1585 ireq->tstamp_ok = rx_opt->tstamp_ok;
1840 req->sack_ok = rx_opt->sack_ok; 1586 ireq->sack_ok = rx_opt->sack_ok;
1841 req->snd_wscale = rx_opt->snd_wscale; 1587 ireq->snd_wscale = rx_opt->snd_wscale;
1842 req->wscale_ok = rx_opt->wscale_ok; 1588 ireq->wscale_ok = rx_opt->wscale_ok;
1843 req->acked = 0; 1589 ireq->acked = 0;
1844 req->ecn_ok = 0; 1590 ireq->ecn_ok = 0;
1845 req->rmt_port = skb->h.th->source; 1591 ireq->rmt_port = skb->h.th->source;
1846} 1592}
1847 1593
1848extern void tcp_enter_memory_pressure(void); 1594extern void tcp_enter_memory_pressure(void);
@@ -1972,52 +1718,4 @@ struct tcp_iter_state {
1972extern int tcp_proc_register(struct tcp_seq_afinfo *afinfo); 1718extern int tcp_proc_register(struct tcp_seq_afinfo *afinfo);
1973extern void tcp_proc_unregister(struct tcp_seq_afinfo *afinfo); 1719extern void tcp_proc_unregister(struct tcp_seq_afinfo *afinfo);
1974 1720
1975/* TCP Westwood functions and constants */
1976
1977#define TCP_WESTWOOD_INIT_RTT (20*HZ) /* maybe too conservative?! */
1978#define TCP_WESTWOOD_RTT_MIN (HZ/20) /* 50ms */
1979
1980static inline void tcp_westwood_update_rtt(struct tcp_sock *tp, __u32 rtt_seq)
1981{
1982 if (tcp_is_westwood(tp))
1983 tp->westwood.rtt = rtt_seq;
1984}
1985
1986static inline __u32 __tcp_westwood_bw_rttmin(const struct tcp_sock *tp)
1987{
1988 return max((tp->westwood.bw_est) * (tp->westwood.rtt_min) /
1989 (__u32) (tp->mss_cache_std),
1990 2U);
1991}
1992
1993static inline __u32 tcp_westwood_bw_rttmin(const struct tcp_sock *tp)
1994{
1995 return tcp_is_westwood(tp) ? __tcp_westwood_bw_rttmin(tp) : 0;
1996}
1997
1998static inline int tcp_westwood_ssthresh(struct tcp_sock *tp)
1999{
2000 __u32 ssthresh = 0;
2001
2002 if (tcp_is_westwood(tp)) {
2003 ssthresh = __tcp_westwood_bw_rttmin(tp);
2004 if (ssthresh)
2005 tp->snd_ssthresh = ssthresh;
2006 }
2007
2008 return (ssthresh != 0);
2009}
2010
2011static inline int tcp_westwood_cwnd(struct tcp_sock *tp)
2012{
2013 __u32 cwnd = 0;
2014
2015 if (tcp_is_westwood(tp)) {
2016 cwnd = __tcp_westwood_bw_rttmin(tp);
2017 if (cwnd)
2018 tp->snd_cwnd = cwnd;
2019 }
2020
2021 return (cwnd != 0);
2022}
2023#endif /* _TCP_H */ 1721#endif /* _TCP_H */
diff --git a/include/net/tcp_ecn.h b/include/net/tcp_ecn.h
index dc1456389a97..64980ee8c92a 100644
--- a/include/net/tcp_ecn.h
+++ b/include/net/tcp_ecn.h
@@ -2,6 +2,7 @@
2#define _NET_TCP_ECN_H_ 1 2#define _NET_TCP_ECN_H_ 1
3 3
4#include <net/inet_ecn.h> 4#include <net/inet_ecn.h>
5#include <net/request_sock.h>
5 6
6#define TCP_HP_BITS (~(TCP_RESERVED_BITS|TCP_FLAG_PSH)) 7#define TCP_HP_BITS (~(TCP_RESERVED_BITS|TCP_FLAG_PSH))
7 8
@@ -38,9 +39,9 @@ static inline void TCP_ECN_send_syn(struct sock *sk, struct tcp_sock *tp,
38} 39}
39 40
40static __inline__ void 41static __inline__ void
41TCP_ECN_make_synack(struct open_request *req, struct tcphdr *th) 42TCP_ECN_make_synack(struct request_sock *req, struct tcphdr *th)
42{ 43{
43 if (req->ecn_ok) 44 if (inet_rsk(req)->ecn_ok)
44 th->ece = 1; 45 th->ece = 1;
45} 46}
46 47
@@ -111,16 +112,16 @@ static inline int TCP_ECN_rcv_ecn_echo(struct tcp_sock *tp, struct tcphdr *th)
111} 112}
112 113
113static inline void TCP_ECN_openreq_child(struct tcp_sock *tp, 114static inline void TCP_ECN_openreq_child(struct tcp_sock *tp,
114 struct open_request *req) 115 struct request_sock *req)
115{ 116{
116 tp->ecn_flags = req->ecn_ok ? TCP_ECN_OK : 0; 117 tp->ecn_flags = inet_rsk(req)->ecn_ok ? TCP_ECN_OK : 0;
117} 118}
118 119
119static __inline__ void 120static __inline__ void
120TCP_ECN_create_request(struct open_request *req, struct tcphdr *th) 121TCP_ECN_create_request(struct request_sock *req, struct tcphdr *th)
121{ 122{
122 if (sysctl_tcp_ecn && th->ece && th->cwr) 123 if (sysctl_tcp_ecn && th->ece && th->cwr)
123 req->ecn_ok = 1; 124 inet_rsk(req)->ecn_ok = 1;
124} 125}
125 126
126#endif 127#endif
diff --git a/include/net/x25.h b/include/net/x25.h
index 7a1ba5bbb868..8b39b98876e8 100644
--- a/include/net/x25.h
+++ b/include/net/x25.h
@@ -79,6 +79,8 @@ enum {
79#define X25_DEFAULT_PACKET_SIZE X25_PS128 /* Default Packet Size */ 79#define X25_DEFAULT_PACKET_SIZE X25_PS128 /* Default Packet Size */
80#define X25_DEFAULT_THROUGHPUT 0x0A /* Deafult Throughput */ 80#define X25_DEFAULT_THROUGHPUT 0x0A /* Deafult Throughput */
81#define X25_DEFAULT_REVERSE 0x00 /* Default Reverse Charging */ 81#define X25_DEFAULT_REVERSE 0x00 /* Default Reverse Charging */
82#define X25_DENY_ACCPT_APPRV 0x01 /* Default value */
83#define X25_ALLOW_ACCPT_APPRV 0x00 /* Control enabled */
82 84
83#define X25_SMODULUS 8 85#define X25_SMODULUS 8
84#define X25_EMODULUS 128 86#define X25_EMODULUS 128
@@ -94,7 +96,7 @@ enum {
94#define X25_FAC_CLASS_C 0x80 96#define X25_FAC_CLASS_C 0x80
95#define X25_FAC_CLASS_D 0xC0 97#define X25_FAC_CLASS_D 0xC0
96 98
97#define X25_FAC_REVERSE 0x01 99#define X25_FAC_REVERSE 0x01 /* also fast select */
98#define X25_FAC_THROUGHPUT 0x02 100#define X25_FAC_THROUGHPUT 0x02
99#define X25_FAC_PACKET_SIZE 0x42 101#define X25_FAC_PACKET_SIZE 0x42
100#define X25_FAC_WINDOW_SIZE 0x43 102#define X25_FAC_WINDOW_SIZE 0x43
@@ -134,8 +136,8 @@ struct x25_sock {
134 struct sock sk; 136 struct sock sk;
135 struct x25_address source_addr, dest_addr; 137 struct x25_address source_addr, dest_addr;
136 struct x25_neigh *neighbour; 138 struct x25_neigh *neighbour;
137 unsigned int lci; 139 unsigned int lci, cudmatchlength;
138 unsigned char state, condition, qbitincl, intflag; 140 unsigned char state, condition, qbitincl, intflag, accptapprv;
139 unsigned short vs, vr, va, vl; 141 unsigned short vs, vr, va, vl;
140 unsigned long t2, t21, t22, t23; 142 unsigned long t2, t21, t22, t23;
141 unsigned short fraglen; 143 unsigned short fraglen;
@@ -242,7 +244,6 @@ extern int x25_validate_nr(struct sock *, unsigned short);
242extern void x25_write_internal(struct sock *, int); 244extern void x25_write_internal(struct sock *, int);
243extern int x25_decode(struct sock *, struct sk_buff *, int *, int *, int *, int *, int *); 245extern int x25_decode(struct sock *, struct sk_buff *, int *, int *, int *, int *, int *);
244extern void x25_disconnect(struct sock *, int, unsigned char, unsigned char); 246extern void x25_disconnect(struct sock *, int, unsigned char, unsigned char);
245extern int x25_check_calluserdata(struct x25_calluserdata *,struct x25_calluserdata *);
246 247
247/* x25_timer.c */ 248/* x25_timer.c */
248extern void x25_start_heartbeat(struct sock *); 249extern void x25_start_heartbeat(struct sock *);
diff --git a/include/net/x25device.h b/include/net/x25device.h
index cf36a20ea3c5..d45ae883bd1d 100644
--- a/include/net/x25device.h
+++ b/include/net/x25device.h
@@ -5,8 +5,7 @@
5#include <linux/if_packet.h> 5#include <linux/if_packet.h>
6#include <linux/skbuff.h> 6#include <linux/skbuff.h>
7 7
8static inline unsigned short x25_type_trans(struct sk_buff *skb, 8static inline __be16 x25_type_trans(struct sk_buff *skb, struct net_device *dev)
9 struct net_device *dev)
10{ 9{
11 skb->mac.raw = skb->data; 10 skb->mac.raw = skb->data;
12 skb->input_dev = skb->dev = dev; 11 skb->input_dev = skb->dev = dev;
diff --git a/include/net/xfrm.h b/include/net/xfrm.h
index d675836ba6c3..868ef88ef971 100644
--- a/include/net/xfrm.h
+++ b/include/net/xfrm.h
@@ -158,6 +158,20 @@ enum {
158 XFRM_STATE_DEAD 158 XFRM_STATE_DEAD
159}; 159};
160 160
161/* callback structure passed from either netlink or pfkey */
162struct km_event
163{
164 union {
165 u32 hard;
166 u32 proto;
167 u32 byid;
168 } data;
169
170 u32 seq;
171 u32 pid;
172 u32 event;
173};
174
161struct xfrm_type; 175struct xfrm_type;
162struct xfrm_dst; 176struct xfrm_dst;
163struct xfrm_policy_afinfo { 177struct xfrm_policy_afinfo {
@@ -179,6 +193,8 @@ struct xfrm_policy_afinfo {
179 193
180extern int xfrm_policy_register_afinfo(struct xfrm_policy_afinfo *afinfo); 194extern int xfrm_policy_register_afinfo(struct xfrm_policy_afinfo *afinfo);
181extern int xfrm_policy_unregister_afinfo(struct xfrm_policy_afinfo *afinfo); 195extern int xfrm_policy_unregister_afinfo(struct xfrm_policy_afinfo *afinfo);
196extern void km_policy_notify(struct xfrm_policy *xp, int dir, struct km_event *c);
197extern void km_state_notify(struct xfrm_state *x, struct km_event *c);
182 198
183#define XFRM_ACQ_EXPIRES 30 199#define XFRM_ACQ_EXPIRES 30
184 200
@@ -188,6 +204,7 @@ struct xfrm_state_afinfo {
188 rwlock_t lock; 204 rwlock_t lock;
189 struct list_head *state_bydst; 205 struct list_head *state_bydst;
190 struct list_head *state_byspi; 206 struct list_head *state_byspi;
207 int (*init_flags)(struct xfrm_state *x);
191 void (*init_tempsel)(struct xfrm_state *x, struct flowi *fl, 208 void (*init_tempsel)(struct xfrm_state *x, struct flowi *fl,
192 struct xfrm_tmpl *tmpl, 209 struct xfrm_tmpl *tmpl,
193 xfrm_address_t *daddr, xfrm_address_t *saddr); 210 xfrm_address_t *daddr, xfrm_address_t *saddr);
@@ -209,7 +226,7 @@ struct xfrm_type
209 struct module *owner; 226 struct module *owner;
210 __u8 proto; 227 __u8 proto;
211 228
212 int (*init_state)(struct xfrm_state *x, void *args); 229 int (*init_state)(struct xfrm_state *x);
213 void (*destructor)(struct xfrm_state *); 230 void (*destructor)(struct xfrm_state *);
214 int (*input)(struct xfrm_state *, struct xfrm_decap_state *, struct sk_buff *skb); 231 int (*input)(struct xfrm_state *, struct xfrm_decap_state *, struct sk_buff *skb);
215 int (*post_input)(struct xfrm_state *, struct xfrm_decap_state *, struct sk_buff *skb); 232 int (*post_input)(struct xfrm_state *, struct xfrm_decap_state *, struct sk_buff *skb);
@@ -290,11 +307,11 @@ struct xfrm_mgr
290{ 307{
291 struct list_head list; 308 struct list_head list;
292 char *id; 309 char *id;
293 int (*notify)(struct xfrm_state *x, int event); 310 int (*notify)(struct xfrm_state *x, struct km_event *c);
294 int (*acquire)(struct xfrm_state *x, struct xfrm_tmpl *, struct xfrm_policy *xp, int dir); 311 int (*acquire)(struct xfrm_state *x, struct xfrm_tmpl *, struct xfrm_policy *xp, int dir);
295 struct xfrm_policy *(*compile_policy)(u16 family, int opt, u8 *data, int len, int *dir); 312 struct xfrm_policy *(*compile_policy)(u16 family, int opt, u8 *data, int len, int *dir);
296 int (*new_mapping)(struct xfrm_state *x, xfrm_address_t *ipaddr, u16 sport); 313 int (*new_mapping)(struct xfrm_state *x, xfrm_address_t *ipaddr, u16 sport);
297 int (*notify_policy)(struct xfrm_policy *x, int dir, int event); 314 int (*notify_policy)(struct xfrm_policy *x, int dir, struct km_event *c);
298}; 315};
299 316
300extern int xfrm_register_km(struct xfrm_mgr *km); 317extern int xfrm_register_km(struct xfrm_mgr *km);
@@ -656,7 +673,7 @@ static inline int xfrm_sk_clone_policy(struct sock *sk)
656 return 0; 673 return 0;
657} 674}
658 675
659extern void xfrm_policy_delete(struct xfrm_policy *pol, int dir); 676extern int xfrm_policy_delete(struct xfrm_policy *pol, int dir);
660 677
661static inline void xfrm_sk_free_policy(struct sock *sk) 678static inline void xfrm_sk_free_policy(struct sock *sk)
662{ 679{
@@ -786,7 +803,7 @@ struct xfrm_algo_desc {
786/* XFRM tunnel handlers. */ 803/* XFRM tunnel handlers. */
787struct xfrm_tunnel { 804struct xfrm_tunnel {
788 int (*handler)(struct sk_buff *skb); 805 int (*handler)(struct sk_buff *skb);
789 void (*err_handler)(struct sk_buff *skb, void *info); 806 void (*err_handler)(struct sk_buff *skb, __u32 info);
790}; 807};
791 808
792struct xfrm6_tunnel { 809struct xfrm6_tunnel {
@@ -817,12 +834,13 @@ extern int xfrm_state_add(struct xfrm_state *x);
817extern int xfrm_state_update(struct xfrm_state *x); 834extern int xfrm_state_update(struct xfrm_state *x);
818extern struct xfrm_state *xfrm_state_lookup(xfrm_address_t *daddr, u32 spi, u8 proto, unsigned short family); 835extern struct xfrm_state *xfrm_state_lookup(xfrm_address_t *daddr, u32 spi, u8 proto, unsigned short family);
819extern struct xfrm_state *xfrm_find_acq_byseq(u32 seq); 836extern struct xfrm_state *xfrm_find_acq_byseq(u32 seq);
820extern void xfrm_state_delete(struct xfrm_state *x); 837extern int xfrm_state_delete(struct xfrm_state *x);
821extern void xfrm_state_flush(u8 proto); 838extern void xfrm_state_flush(u8 proto);
822extern int xfrm_replay_check(struct xfrm_state *x, u32 seq); 839extern int xfrm_replay_check(struct xfrm_state *x, u32 seq);
823extern void xfrm_replay_advance(struct xfrm_state *x, u32 seq); 840extern void xfrm_replay_advance(struct xfrm_state *x, u32 seq);
824extern int xfrm_state_check(struct xfrm_state *x, struct sk_buff *skb); 841extern int xfrm_state_check(struct xfrm_state *x, struct sk_buff *skb);
825extern int xfrm_state_mtu(struct xfrm_state *x, int mtu); 842extern int xfrm_state_mtu(struct xfrm_state *x, int mtu);
843extern int xfrm_init_state(struct xfrm_state *x);
826extern int xfrm4_rcv(struct sk_buff *skb); 844extern int xfrm4_rcv(struct sk_buff *skb);
827extern int xfrm4_output(struct sk_buff *skb); 845extern int xfrm4_output(struct sk_buff *skb);
828extern int xfrm4_tunnel_register(struct xfrm_tunnel *handler); 846extern int xfrm4_tunnel_register(struct xfrm_tunnel *handler);