diff options
Diffstat (limited to 'net/ipv4/tcp_ipv4.c')
-rw-r--r-- | net/ipv4/tcp_ipv4.c | 232 |
1 files changed, 162 insertions, 70 deletions
diff --git a/net/ipv4/tcp_ipv4.c b/net/ipv4/tcp_ipv4.c index 7cda24b53f61..3c23e70885f4 100644 --- a/net/ipv4/tcp_ipv4.c +++ b/net/ipv4/tcp_ipv4.c | |||
@@ -60,6 +60,7 @@ | |||
60 | #include <linux/jhash.h> | 60 | #include <linux/jhash.h> |
61 | #include <linux/init.h> | 61 | #include <linux/init.h> |
62 | #include <linux/times.h> | 62 | #include <linux/times.h> |
63 | #include <linux/slab.h> | ||
63 | 64 | ||
64 | #include <net/net_namespace.h> | 65 | #include <net/net_namespace.h> |
65 | #include <net/icmp.h> | 66 | #include <net/icmp.h> |
@@ -165,10 +166,10 @@ int tcp_v4_connect(struct sock *sk, struct sockaddr *uaddr, int addr_len) | |||
165 | nexthop = inet->opt->faddr; | 166 | nexthop = inet->opt->faddr; |
166 | } | 167 | } |
167 | 168 | ||
168 | tmp = ip_route_connect(&rt, nexthop, inet->saddr, | 169 | tmp = ip_route_connect(&rt, nexthop, inet->inet_saddr, |
169 | RT_CONN_FLAGS(sk), sk->sk_bound_dev_if, | 170 | RT_CONN_FLAGS(sk), sk->sk_bound_dev_if, |
170 | IPPROTO_TCP, | 171 | IPPROTO_TCP, |
171 | inet->sport, usin->sin_port, sk, 1); | 172 | inet->inet_sport, usin->sin_port, sk, 1); |
172 | if (tmp < 0) { | 173 | if (tmp < 0) { |
173 | if (tmp == -ENETUNREACH) | 174 | if (tmp == -ENETUNREACH) |
174 | IP_INC_STATS_BH(sock_net(sk), IPSTATS_MIB_OUTNOROUTES); | 175 | IP_INC_STATS_BH(sock_net(sk), IPSTATS_MIB_OUTNOROUTES); |
@@ -183,11 +184,11 @@ int tcp_v4_connect(struct sock *sk, struct sockaddr *uaddr, int addr_len) | |||
183 | if (!inet->opt || !inet->opt->srr) | 184 | if (!inet->opt || !inet->opt->srr) |
184 | daddr = rt->rt_dst; | 185 | daddr = rt->rt_dst; |
185 | 186 | ||
186 | if (!inet->saddr) | 187 | if (!inet->inet_saddr) |
187 | inet->saddr = rt->rt_src; | 188 | inet->inet_saddr = rt->rt_src; |
188 | inet->rcv_saddr = inet->saddr; | 189 | inet->inet_rcv_saddr = inet->inet_saddr; |
189 | 190 | ||
190 | if (tp->rx_opt.ts_recent_stamp && inet->daddr != daddr) { | 191 | if (tp->rx_opt.ts_recent_stamp && inet->inet_daddr != daddr) { |
191 | /* Reset inherited state */ | 192 | /* Reset inherited state */ |
192 | tp->rx_opt.ts_recent = 0; | 193 | tp->rx_opt.ts_recent = 0; |
193 | tp->rx_opt.ts_recent_stamp = 0; | 194 | tp->rx_opt.ts_recent_stamp = 0; |
@@ -204,20 +205,20 @@ int tcp_v4_connect(struct sock *sk, struct sockaddr *uaddr, int addr_len) | |||
204 | * when trying new connection. | 205 | * when trying new connection. |
205 | */ | 206 | */ |
206 | if (peer != NULL && | 207 | if (peer != NULL && |
207 | peer->tcp_ts_stamp + TCP_PAWS_MSL >= get_seconds()) { | 208 | (u32)get_seconds() - peer->tcp_ts_stamp <= TCP_PAWS_MSL) { |
208 | tp->rx_opt.ts_recent_stamp = peer->tcp_ts_stamp; | 209 | tp->rx_opt.ts_recent_stamp = peer->tcp_ts_stamp; |
209 | tp->rx_opt.ts_recent = peer->tcp_ts; | 210 | tp->rx_opt.ts_recent = peer->tcp_ts; |
210 | } | 211 | } |
211 | } | 212 | } |
212 | 213 | ||
213 | inet->dport = usin->sin_port; | 214 | inet->inet_dport = usin->sin_port; |
214 | inet->daddr = daddr; | 215 | inet->inet_daddr = daddr; |
215 | 216 | ||
216 | inet_csk(sk)->icsk_ext_hdr_len = 0; | 217 | inet_csk(sk)->icsk_ext_hdr_len = 0; |
217 | if (inet->opt) | 218 | if (inet->opt) |
218 | inet_csk(sk)->icsk_ext_hdr_len = inet->opt->optlen; | 219 | inet_csk(sk)->icsk_ext_hdr_len = inet->opt->optlen; |
219 | 220 | ||
220 | tp->rx_opt.mss_clamp = 536; | 221 | tp->rx_opt.mss_clamp = TCP_MSS_DEFAULT; |
221 | 222 | ||
222 | /* Socket identity is still unknown (sport may be zero). | 223 | /* Socket identity is still unknown (sport may be zero). |
223 | * However we set state to SYN-SENT and not releasing socket | 224 | * However we set state to SYN-SENT and not releasing socket |
@@ -230,7 +231,7 @@ int tcp_v4_connect(struct sock *sk, struct sockaddr *uaddr, int addr_len) | |||
230 | goto failure; | 231 | goto failure; |
231 | 232 | ||
232 | err = ip_route_newports(&rt, IPPROTO_TCP, | 233 | err = ip_route_newports(&rt, IPPROTO_TCP, |
233 | inet->sport, inet->dport, sk); | 234 | inet->inet_sport, inet->inet_dport, sk); |
234 | if (err) | 235 | if (err) |
235 | goto failure; | 236 | goto failure; |
236 | 237 | ||
@@ -239,12 +240,12 @@ int tcp_v4_connect(struct sock *sk, struct sockaddr *uaddr, int addr_len) | |||
239 | sk_setup_caps(sk, &rt->u.dst); | 240 | sk_setup_caps(sk, &rt->u.dst); |
240 | 241 | ||
241 | if (!tp->write_seq) | 242 | if (!tp->write_seq) |
242 | tp->write_seq = secure_tcp_sequence_number(inet->saddr, | 243 | tp->write_seq = secure_tcp_sequence_number(inet->inet_saddr, |
243 | inet->daddr, | 244 | inet->inet_daddr, |
244 | inet->sport, | 245 | inet->inet_sport, |
245 | usin->sin_port); | 246 | usin->sin_port); |
246 | 247 | ||
247 | inet->id = tp->write_seq ^ jiffies; | 248 | inet->inet_id = tp->write_seq ^ jiffies; |
248 | 249 | ||
249 | err = tcp_connect(sk); | 250 | err = tcp_connect(sk); |
250 | rt = NULL; | 251 | rt = NULL; |
@@ -261,7 +262,7 @@ failure: | |||
261 | tcp_set_state(sk, TCP_CLOSE); | 262 | tcp_set_state(sk, TCP_CLOSE); |
262 | ip_rt_put(rt); | 263 | ip_rt_put(rt); |
263 | sk->sk_route_caps = 0; | 264 | sk->sk_route_caps = 0; |
264 | inet->dport = 0; | 265 | inet->inet_dport = 0; |
265 | return err; | 266 | return err; |
266 | } | 267 | } |
267 | 268 | ||
@@ -370,6 +371,11 @@ void tcp_v4_err(struct sk_buff *icmp_skb, u32 info) | |||
370 | if (sk->sk_state == TCP_CLOSE) | 371 | if (sk->sk_state == TCP_CLOSE) |
371 | goto out; | 372 | goto out; |
372 | 373 | ||
374 | if (unlikely(iph->ttl < inet_sk(sk)->min_ttl)) { | ||
375 | NET_INC_STATS_BH(net, LINUX_MIB_TCPMINTTLDROP); | ||
376 | goto out; | ||
377 | } | ||
378 | |||
373 | icsk = inet_csk(sk); | 379 | icsk = inet_csk(sk); |
374 | tp = tcp_sk(sk); | 380 | tp = tcp_sk(sk); |
375 | seq = ntohl(th->seq); | 381 | seq = ntohl(th->seq); |
@@ -520,12 +526,13 @@ void tcp_v4_send_check(struct sock *sk, int len, struct sk_buff *skb) | |||
520 | struct tcphdr *th = tcp_hdr(skb); | 526 | struct tcphdr *th = tcp_hdr(skb); |
521 | 527 | ||
522 | if (skb->ip_summed == CHECKSUM_PARTIAL) { | 528 | if (skb->ip_summed == CHECKSUM_PARTIAL) { |
523 | th->check = ~tcp_v4_check(len, inet->saddr, | 529 | th->check = ~tcp_v4_check(len, inet->inet_saddr, |
524 | inet->daddr, 0); | 530 | inet->inet_daddr, 0); |
525 | skb->csum_start = skb_transport_header(skb) - skb->head; | 531 | skb->csum_start = skb_transport_header(skb) - skb->head; |
526 | skb->csum_offset = offsetof(struct tcphdr, check); | 532 | skb->csum_offset = offsetof(struct tcphdr, check); |
527 | } else { | 533 | } else { |
528 | th->check = tcp_v4_check(len, inet->saddr, inet->daddr, | 534 | th->check = tcp_v4_check(len, inet->inet_saddr, |
535 | inet->inet_daddr, | ||
529 | csum_partial(th, | 536 | csum_partial(th, |
530 | th->doff << 2, | 537 | th->doff << 2, |
531 | skb->csum)); | 538 | skb->csum)); |
@@ -741,8 +748,9 @@ static void tcp_v4_reqsk_send_ack(struct sock *sk, struct sk_buff *skb, | |||
741 | * This still operates on a request_sock only, not on a big | 748 | * This still operates on a request_sock only, not on a big |
742 | * socket. | 749 | * socket. |
743 | */ | 750 | */ |
744 | static int __tcp_v4_send_synack(struct sock *sk, struct request_sock *req, | 751 | static int tcp_v4_send_synack(struct sock *sk, struct dst_entry *dst, |
745 | struct dst_entry *dst) | 752 | struct request_sock *req, |
753 | struct request_values *rvp) | ||
746 | { | 754 | { |
747 | const struct inet_request_sock *ireq = inet_rsk(req); | 755 | const struct inet_request_sock *ireq = inet_rsk(req); |
748 | int err = -1; | 756 | int err = -1; |
@@ -752,7 +760,7 @@ static int __tcp_v4_send_synack(struct sock *sk, struct request_sock *req, | |||
752 | if (!dst && (dst = inet_csk_route_req(sk, req)) == NULL) | 760 | if (!dst && (dst = inet_csk_route_req(sk, req)) == NULL) |
753 | return -1; | 761 | return -1; |
754 | 762 | ||
755 | skb = tcp_make_synack(sk, dst, req); | 763 | skb = tcp_make_synack(sk, dst, req, rvp); |
756 | 764 | ||
757 | if (skb) { | 765 | if (skb) { |
758 | struct tcphdr *th = tcp_hdr(skb); | 766 | struct tcphdr *th = tcp_hdr(skb); |
@@ -773,9 +781,11 @@ static int __tcp_v4_send_synack(struct sock *sk, struct request_sock *req, | |||
773 | return err; | 781 | return err; |
774 | } | 782 | } |
775 | 783 | ||
776 | static int tcp_v4_send_synack(struct sock *sk, struct request_sock *req) | 784 | static int tcp_v4_rtx_synack(struct sock *sk, struct request_sock *req, |
785 | struct request_values *rvp) | ||
777 | { | 786 | { |
778 | return __tcp_v4_send_synack(sk, req, NULL); | 787 | TCP_INC_STATS_BH(sock_net(sk), TCP_MIB_RETRANSSEGS); |
788 | return tcp_v4_send_synack(sk, NULL, req, rvp); | ||
779 | } | 789 | } |
780 | 790 | ||
781 | /* | 791 | /* |
@@ -848,7 +858,7 @@ static struct tcp_md5sig_key * | |||
848 | struct tcp_md5sig_key *tcp_v4_md5_lookup(struct sock *sk, | 858 | struct tcp_md5sig_key *tcp_v4_md5_lookup(struct sock *sk, |
849 | struct sock *addr_sk) | 859 | struct sock *addr_sk) |
850 | { | 860 | { |
851 | return tcp_v4_md5_do_lookup(sk, inet_sk(addr_sk)->daddr); | 861 | return tcp_v4_md5_do_lookup(sk, inet_sk(addr_sk)->inet_daddr); |
852 | } | 862 | } |
853 | 863 | ||
854 | EXPORT_SYMBOL(tcp_v4_md5_lookup); | 864 | EXPORT_SYMBOL(tcp_v4_md5_lookup); |
@@ -923,7 +933,7 @@ EXPORT_SYMBOL(tcp_v4_md5_do_add); | |||
923 | static int tcp_v4_md5_add_func(struct sock *sk, struct sock *addr_sk, | 933 | static int tcp_v4_md5_add_func(struct sock *sk, struct sock *addr_sk, |
924 | u8 *newkey, u8 newkeylen) | 934 | u8 *newkey, u8 newkeylen) |
925 | { | 935 | { |
926 | return tcp_v4_md5_do_add(sk, inet_sk(addr_sk)->daddr, | 936 | return tcp_v4_md5_do_add(sk, inet_sk(addr_sk)->inet_daddr, |
927 | newkey, newkeylen); | 937 | newkey, newkeylen); |
928 | } | 938 | } |
929 | 939 | ||
@@ -1089,8 +1099,8 @@ int tcp_v4_md5_hash_skb(char *md5_hash, struct tcp_md5sig_key *key, | |||
1089 | __be32 saddr, daddr; | 1099 | __be32 saddr, daddr; |
1090 | 1100 | ||
1091 | if (sk) { | 1101 | if (sk) { |
1092 | saddr = inet_sk(sk)->saddr; | 1102 | saddr = inet_sk(sk)->inet_saddr; |
1093 | daddr = inet_sk(sk)->daddr; | 1103 | daddr = inet_sk(sk)->inet_daddr; |
1094 | } else if (req) { | 1104 | } else if (req) { |
1095 | saddr = inet_rsk(req)->loc_addr; | 1105 | saddr = inet_rsk(req)->loc_addr; |
1096 | daddr = inet_rsk(req)->rmt_addr; | 1106 | daddr = inet_rsk(req)->rmt_addr; |
@@ -1189,10 +1199,11 @@ static int tcp_v4_inbound_md5_hash(struct sock *sk, struct sk_buff *skb) | |||
1189 | struct request_sock_ops tcp_request_sock_ops __read_mostly = { | 1199 | struct request_sock_ops tcp_request_sock_ops __read_mostly = { |
1190 | .family = PF_INET, | 1200 | .family = PF_INET, |
1191 | .obj_size = sizeof(struct tcp_request_sock), | 1201 | .obj_size = sizeof(struct tcp_request_sock), |
1192 | .rtx_syn_ack = tcp_v4_send_synack, | 1202 | .rtx_syn_ack = tcp_v4_rtx_synack, |
1193 | .send_ack = tcp_v4_reqsk_send_ack, | 1203 | .send_ack = tcp_v4_reqsk_send_ack, |
1194 | .destructor = tcp_v4_reqsk_destructor, | 1204 | .destructor = tcp_v4_reqsk_destructor, |
1195 | .send_reset = tcp_v4_send_reset, | 1205 | .send_reset = tcp_v4_send_reset, |
1206 | .syn_ack_timeout = tcp_syn_ack_timeout, | ||
1196 | }; | 1207 | }; |
1197 | 1208 | ||
1198 | #ifdef CONFIG_TCP_MD5SIG | 1209 | #ifdef CONFIG_TCP_MD5SIG |
@@ -1210,13 +1221,16 @@ static struct timewait_sock_ops tcp_timewait_sock_ops = { | |||
1210 | 1221 | ||
1211 | int tcp_v4_conn_request(struct sock *sk, struct sk_buff *skb) | 1222 | int tcp_v4_conn_request(struct sock *sk, struct sk_buff *skb) |
1212 | { | 1223 | { |
1213 | struct inet_request_sock *ireq; | 1224 | struct tcp_extend_values tmp_ext; |
1214 | struct tcp_options_received tmp_opt; | 1225 | struct tcp_options_received tmp_opt; |
1226 | u8 *hash_location; | ||
1215 | struct request_sock *req; | 1227 | struct request_sock *req; |
1228 | struct inet_request_sock *ireq; | ||
1229 | struct tcp_sock *tp = tcp_sk(sk); | ||
1230 | struct dst_entry *dst = NULL; | ||
1216 | __be32 saddr = ip_hdr(skb)->saddr; | 1231 | __be32 saddr = ip_hdr(skb)->saddr; |
1217 | __be32 daddr = ip_hdr(skb)->daddr; | 1232 | __be32 daddr = ip_hdr(skb)->daddr; |
1218 | __u32 isn = TCP_SKB_CB(skb)->when; | 1233 | __u32 isn = TCP_SKB_CB(skb)->when; |
1219 | struct dst_entry *dst = NULL; | ||
1220 | #ifdef CONFIG_SYN_COOKIES | 1234 | #ifdef CONFIG_SYN_COOKIES |
1221 | int want_cookie = 0; | 1235 | int want_cookie = 0; |
1222 | #else | 1236 | #else |
@@ -1257,16 +1271,50 @@ int tcp_v4_conn_request(struct sock *sk, struct sk_buff *skb) | |||
1257 | #endif | 1271 | #endif |
1258 | 1272 | ||
1259 | tcp_clear_options(&tmp_opt); | 1273 | tcp_clear_options(&tmp_opt); |
1260 | tmp_opt.mss_clamp = 536; | 1274 | tmp_opt.mss_clamp = TCP_MSS_DEFAULT; |
1261 | tmp_opt.user_mss = tcp_sk(sk)->rx_opt.user_mss; | 1275 | tmp_opt.user_mss = tp->rx_opt.user_mss; |
1276 | tcp_parse_options(skb, &tmp_opt, &hash_location, 0); | ||
1277 | |||
1278 | if (tmp_opt.cookie_plus > 0 && | ||
1279 | tmp_opt.saw_tstamp && | ||
1280 | !tp->rx_opt.cookie_out_never && | ||
1281 | (sysctl_tcp_cookie_size > 0 || | ||
1282 | (tp->cookie_values != NULL && | ||
1283 | tp->cookie_values->cookie_desired > 0))) { | ||
1284 | u8 *c; | ||
1285 | u32 *mess = &tmp_ext.cookie_bakery[COOKIE_DIGEST_WORDS]; | ||
1286 | int l = tmp_opt.cookie_plus - TCPOLEN_COOKIE_BASE; | ||
1287 | |||
1288 | if (tcp_cookie_generator(&tmp_ext.cookie_bakery[0]) != 0) | ||
1289 | goto drop_and_release; | ||
1290 | |||
1291 | /* Secret recipe starts with IP addresses */ | ||
1292 | *mess++ ^= daddr; | ||
1293 | *mess++ ^= saddr; | ||
1262 | 1294 | ||
1263 | tcp_parse_options(skb, &tmp_opt, 0); | 1295 | /* plus variable length Initiator Cookie */ |
1296 | c = (u8 *)mess; | ||
1297 | while (l-- > 0) | ||
1298 | *c++ ^= *hash_location++; | ||
1299 | |||
1300 | #ifdef CONFIG_SYN_COOKIES | ||
1301 | want_cookie = 0; /* not our kind of cookie */ | ||
1302 | #endif | ||
1303 | tmp_ext.cookie_out_never = 0; /* false */ | ||
1304 | tmp_ext.cookie_plus = tmp_opt.cookie_plus; | ||
1305 | } else if (!tp->rx_opt.cookie_in_always) { | ||
1306 | /* redundant indications, but ensure initialization. */ | ||
1307 | tmp_ext.cookie_out_never = 1; /* true */ | ||
1308 | tmp_ext.cookie_plus = 0; | ||
1309 | } else { | ||
1310 | goto drop_and_release; | ||
1311 | } | ||
1312 | tmp_ext.cookie_in_always = tp->rx_opt.cookie_in_always; | ||
1264 | 1313 | ||
1265 | if (want_cookie && !tmp_opt.saw_tstamp) | 1314 | if (want_cookie && !tmp_opt.saw_tstamp) |
1266 | tcp_clear_options(&tmp_opt); | 1315 | tcp_clear_options(&tmp_opt); |
1267 | 1316 | ||
1268 | tmp_opt.tstamp_ok = tmp_opt.saw_tstamp; | 1317 | tmp_opt.tstamp_ok = tmp_opt.saw_tstamp; |
1269 | |||
1270 | tcp_openreq_init(req, &tmp_opt, skb); | 1318 | tcp_openreq_init(req, &tmp_opt, skb); |
1271 | 1319 | ||
1272 | ireq = inet_rsk(req); | 1320 | ireq = inet_rsk(req); |
@@ -1304,7 +1352,7 @@ int tcp_v4_conn_request(struct sock *sk, struct sk_buff *skb) | |||
1304 | (dst = inet_csk_route_req(sk, req)) != NULL && | 1352 | (dst = inet_csk_route_req(sk, req)) != NULL && |
1305 | (peer = rt_get_peer((struct rtable *)dst)) != NULL && | 1353 | (peer = rt_get_peer((struct rtable *)dst)) != NULL && |
1306 | peer->v4daddr == saddr) { | 1354 | peer->v4daddr == saddr) { |
1307 | if (get_seconds() < peer->tcp_ts_stamp + TCP_PAWS_MSL && | 1355 | if ((u32)get_seconds() - peer->tcp_ts_stamp < TCP_PAWS_MSL && |
1308 | (s32)(peer->tcp_ts - req->ts_recent) > | 1356 | (s32)(peer->tcp_ts - req->ts_recent) > |
1309 | TCP_PAWS_WINDOW) { | 1357 | TCP_PAWS_WINDOW) { |
1310 | NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_PAWSPASSIVEREJECTED); | 1358 | NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_PAWSPASSIVEREJECTED); |
@@ -1333,7 +1381,9 @@ int tcp_v4_conn_request(struct sock *sk, struct sk_buff *skb) | |||
1333 | } | 1381 | } |
1334 | tcp_rsk(req)->snt_isn = isn; | 1382 | tcp_rsk(req)->snt_isn = isn; |
1335 | 1383 | ||
1336 | if (__tcp_v4_send_synack(sk, req, dst) || want_cookie) | 1384 | if (tcp_v4_send_synack(sk, dst, req, |
1385 | (struct request_values *)&tmp_ext) || | ||
1386 | want_cookie) | ||
1337 | goto drop_and_free; | 1387 | goto drop_and_free; |
1338 | 1388 | ||
1339 | inet_csk_reqsk_queue_hash_add(sk, req, TCP_TIMEOUT_INIT); | 1389 | inet_csk_reqsk_queue_hash_add(sk, req, TCP_TIMEOUT_INIT); |
@@ -1380,9 +1430,9 @@ struct sock *tcp_v4_syn_recv_sock(struct sock *sk, struct sk_buff *skb, | |||
1380 | newtp = tcp_sk(newsk); | 1430 | newtp = tcp_sk(newsk); |
1381 | newinet = inet_sk(newsk); | 1431 | newinet = inet_sk(newsk); |
1382 | ireq = inet_rsk(req); | 1432 | ireq = inet_rsk(req); |
1383 | newinet->daddr = ireq->rmt_addr; | 1433 | newinet->inet_daddr = ireq->rmt_addr; |
1384 | newinet->rcv_saddr = ireq->loc_addr; | 1434 | newinet->inet_rcv_saddr = ireq->loc_addr; |
1385 | newinet->saddr = ireq->loc_addr; | 1435 | newinet->inet_saddr = ireq->loc_addr; |
1386 | newinet->opt = ireq->opt; | 1436 | newinet->opt = ireq->opt; |
1387 | ireq->opt = NULL; | 1437 | ireq->opt = NULL; |
1388 | newinet->mc_index = inet_iif(skb); | 1438 | newinet->mc_index = inet_iif(skb); |
@@ -1390,7 +1440,7 @@ struct sock *tcp_v4_syn_recv_sock(struct sock *sk, struct sk_buff *skb, | |||
1390 | inet_csk(newsk)->icsk_ext_hdr_len = 0; | 1440 | inet_csk(newsk)->icsk_ext_hdr_len = 0; |
1391 | if (newinet->opt) | 1441 | if (newinet->opt) |
1392 | inet_csk(newsk)->icsk_ext_hdr_len = newinet->opt->optlen; | 1442 | inet_csk(newsk)->icsk_ext_hdr_len = newinet->opt->optlen; |
1393 | newinet->id = newtp->write_seq ^ jiffies; | 1443 | newinet->inet_id = newtp->write_seq ^ jiffies; |
1394 | 1444 | ||
1395 | tcp_mtup_init(newsk); | 1445 | tcp_mtup_init(newsk); |
1396 | tcp_sync_mss(newsk, dst_mtu(dst)); | 1446 | tcp_sync_mss(newsk, dst_mtu(dst)); |
@@ -1403,7 +1453,8 @@ struct sock *tcp_v4_syn_recv_sock(struct sock *sk, struct sk_buff *skb, | |||
1403 | 1453 | ||
1404 | #ifdef CONFIG_TCP_MD5SIG | 1454 | #ifdef CONFIG_TCP_MD5SIG |
1405 | /* Copy over the MD5 key from the original socket */ | 1455 | /* Copy over the MD5 key from the original socket */ |
1406 | if ((key = tcp_v4_md5_do_lookup(sk, newinet->daddr)) != NULL) { | 1456 | key = tcp_v4_md5_do_lookup(sk, newinet->inet_daddr); |
1457 | if (key != NULL) { | ||
1407 | /* | 1458 | /* |
1408 | * We're using one, so create a matching key | 1459 | * We're using one, so create a matching key |
1409 | * on the newsk structure. If we fail to get | 1460 | * on the newsk structure. If we fail to get |
@@ -1412,13 +1463,13 @@ struct sock *tcp_v4_syn_recv_sock(struct sock *sk, struct sk_buff *skb, | |||
1412 | */ | 1463 | */ |
1413 | char *newkey = kmemdup(key->key, key->keylen, GFP_ATOMIC); | 1464 | char *newkey = kmemdup(key->key, key->keylen, GFP_ATOMIC); |
1414 | if (newkey != NULL) | 1465 | if (newkey != NULL) |
1415 | tcp_v4_md5_do_add(newsk, newinet->daddr, | 1466 | tcp_v4_md5_do_add(newsk, newinet->inet_daddr, |
1416 | newkey, key->keylen); | 1467 | newkey, key->keylen); |
1417 | newsk->sk_route_caps &= ~NETIF_F_GSO_MASK; | 1468 | newsk->sk_route_caps &= ~NETIF_F_GSO_MASK; |
1418 | } | 1469 | } |
1419 | #endif | 1470 | #endif |
1420 | 1471 | ||
1421 | __inet_hash_nolisten(newsk); | 1472 | __inet_hash_nolisten(newsk, NULL); |
1422 | __inet_inherit_port(sk, newsk); | 1473 | __inet_inherit_port(sk, newsk); |
1423 | 1474 | ||
1424 | return newsk; | 1475 | return newsk; |
@@ -1610,6 +1661,11 @@ process: | |||
1610 | if (sk->sk_state == TCP_TIME_WAIT) | 1661 | if (sk->sk_state == TCP_TIME_WAIT) |
1611 | goto do_time_wait; | 1662 | goto do_time_wait; |
1612 | 1663 | ||
1664 | if (unlikely(iph->ttl < inet_sk(sk)->min_ttl)) { | ||
1665 | NET_INC_STATS_BH(net, LINUX_MIB_TCPMINTTLDROP); | ||
1666 | goto discard_and_relse; | ||
1667 | } | ||
1668 | |||
1613 | if (!xfrm4_policy_check(sk, XFRM_POLICY_IN, skb)) | 1669 | if (!xfrm4_policy_check(sk, XFRM_POLICY_IN, skb)) |
1614 | goto discard_and_relse; | 1670 | goto discard_and_relse; |
1615 | nf_reset(skb); | 1671 | nf_reset(skb); |
@@ -1634,8 +1690,11 @@ process: | |||
1634 | if (!tcp_prequeue(sk, skb)) | 1690 | if (!tcp_prequeue(sk, skb)) |
1635 | ret = tcp_v4_do_rcv(sk, skb); | 1691 | ret = tcp_v4_do_rcv(sk, skb); |
1636 | } | 1692 | } |
1637 | } else | 1693 | } else if (unlikely(sk_add_backlog(sk, skb))) { |
1638 | sk_add_backlog(sk, skb); | 1694 | bh_unlock_sock(sk); |
1695 | NET_INC_STATS_BH(net, LINUX_MIB_TCPBACKLOGDROP); | ||
1696 | goto discard_and_relse; | ||
1697 | } | ||
1639 | bh_unlock_sock(sk); | 1698 | bh_unlock_sock(sk); |
1640 | 1699 | ||
1641 | sock_put(sk); | 1700 | sock_put(sk); |
@@ -1711,8 +1770,8 @@ int tcp_v4_remember_stamp(struct sock *sk) | |||
1711 | struct inet_peer *peer = NULL; | 1770 | struct inet_peer *peer = NULL; |
1712 | int release_it = 0; | 1771 | int release_it = 0; |
1713 | 1772 | ||
1714 | if (!rt || rt->rt_dst != inet->daddr) { | 1773 | if (!rt || rt->rt_dst != inet->inet_daddr) { |
1715 | peer = inet_getpeer(inet->daddr, 1); | 1774 | peer = inet_getpeer(inet->inet_daddr, 1); |
1716 | release_it = 1; | 1775 | release_it = 1; |
1717 | } else { | 1776 | } else { |
1718 | if (!rt->peer) | 1777 | if (!rt->peer) |
@@ -1722,9 +1781,9 @@ int tcp_v4_remember_stamp(struct sock *sk) | |||
1722 | 1781 | ||
1723 | if (peer) { | 1782 | if (peer) { |
1724 | if ((s32)(peer->tcp_ts - tp->rx_opt.ts_recent) <= 0 || | 1783 | if ((s32)(peer->tcp_ts - tp->rx_opt.ts_recent) <= 0 || |
1725 | (peer->tcp_ts_stamp + TCP_PAWS_MSL < get_seconds() && | 1784 | ((u32)get_seconds() - peer->tcp_ts_stamp > TCP_PAWS_MSL && |
1726 | peer->tcp_ts_stamp <= tp->rx_opt.ts_recent_stamp)) { | 1785 | peer->tcp_ts_stamp <= (u32)tp->rx_opt.ts_recent_stamp)) { |
1727 | peer->tcp_ts_stamp = tp->rx_opt.ts_recent_stamp; | 1786 | peer->tcp_ts_stamp = (u32)tp->rx_opt.ts_recent_stamp; |
1728 | peer->tcp_ts = tp->rx_opt.ts_recent; | 1787 | peer->tcp_ts = tp->rx_opt.ts_recent; |
1729 | } | 1788 | } |
1730 | if (release_it) | 1789 | if (release_it) |
@@ -1743,9 +1802,9 @@ int tcp_v4_tw_remember_stamp(struct inet_timewait_sock *tw) | |||
1743 | const struct tcp_timewait_sock *tcptw = tcp_twsk((struct sock *)tw); | 1802 | const struct tcp_timewait_sock *tcptw = tcp_twsk((struct sock *)tw); |
1744 | 1803 | ||
1745 | if ((s32)(peer->tcp_ts - tcptw->tw_ts_recent) <= 0 || | 1804 | if ((s32)(peer->tcp_ts - tcptw->tw_ts_recent) <= 0 || |
1746 | (peer->tcp_ts_stamp + TCP_PAWS_MSL < get_seconds() && | 1805 | ((u32)get_seconds() - peer->tcp_ts_stamp > TCP_PAWS_MSL && |
1747 | peer->tcp_ts_stamp <= tcptw->tw_ts_recent_stamp)) { | 1806 | peer->tcp_ts_stamp <= (u32)tcptw->tw_ts_recent_stamp)) { |
1748 | peer->tcp_ts_stamp = tcptw->tw_ts_recent_stamp; | 1807 | peer->tcp_ts_stamp = (u32)tcptw->tw_ts_recent_stamp; |
1749 | peer->tcp_ts = tcptw->tw_ts_recent; | 1808 | peer->tcp_ts = tcptw->tw_ts_recent; |
1750 | } | 1809 | } |
1751 | inet_putpeer(peer); | 1810 | inet_putpeer(peer); |
@@ -1810,7 +1869,7 @@ static int tcp_v4_init_sock(struct sock *sk) | |||
1810 | */ | 1869 | */ |
1811 | tp->snd_ssthresh = TCP_INFINITE_SSTHRESH; | 1870 | tp->snd_ssthresh = TCP_INFINITE_SSTHRESH; |
1812 | tp->snd_cwnd_clamp = ~0; | 1871 | tp->snd_cwnd_clamp = ~0; |
1813 | tp->mss_cache = 536; | 1872 | tp->mss_cache = TCP_MSS_DEFAULT; |
1814 | 1873 | ||
1815 | tp->reordering = sysctl_tcp_reordering; | 1874 | tp->reordering = sysctl_tcp_reordering; |
1816 | icsk->icsk_ca_ops = &tcp_init_congestion_ops; | 1875 | icsk->icsk_ca_ops = &tcp_init_congestion_ops; |
@@ -1826,6 +1885,19 @@ static int tcp_v4_init_sock(struct sock *sk) | |||
1826 | tp->af_specific = &tcp_sock_ipv4_specific; | 1885 | tp->af_specific = &tcp_sock_ipv4_specific; |
1827 | #endif | 1886 | #endif |
1828 | 1887 | ||
1888 | /* TCP Cookie Transactions */ | ||
1889 | if (sysctl_tcp_cookie_size > 0) { | ||
1890 | /* Default, cookies without s_data_payload. */ | ||
1891 | tp->cookie_values = | ||
1892 | kzalloc(sizeof(*tp->cookie_values), | ||
1893 | sk->sk_allocation); | ||
1894 | if (tp->cookie_values != NULL) | ||
1895 | kref_init(&tp->cookie_values->kref); | ||
1896 | } | ||
1897 | /* Presumed zeroed, in order of appearance: | ||
1898 | * cookie_in_always, cookie_out_never, | ||
1899 | * s_data_constant, s_data_in, s_data_out | ||
1900 | */ | ||
1829 | sk->sk_sndbuf = sysctl_tcp_wmem[1]; | 1901 | sk->sk_sndbuf = sysctl_tcp_wmem[1]; |
1830 | sk->sk_rcvbuf = sysctl_tcp_rmem[1]; | 1902 | sk->sk_rcvbuf = sysctl_tcp_rmem[1]; |
1831 | 1903 | ||
@@ -1879,6 +1951,13 @@ void tcp_v4_destroy_sock(struct sock *sk) | |||
1879 | sk->sk_sndmsg_page = NULL; | 1951 | sk->sk_sndmsg_page = NULL; |
1880 | } | 1952 | } |
1881 | 1953 | ||
1954 | /* TCP Cookie Transactions */ | ||
1955 | if (tp->cookie_values != NULL) { | ||
1956 | kref_put(&tp->cookie_values->kref, | ||
1957 | tcp_cookie_values_release); | ||
1958 | tp->cookie_values = NULL; | ||
1959 | } | ||
1960 | |||
1882 | percpu_counter_dec(&tcp_sockets_allocated); | 1961 | percpu_counter_dec(&tcp_sockets_allocated); |
1883 | } | 1962 | } |
1884 | 1963 | ||
@@ -2000,7 +2079,7 @@ static void *established_get_first(struct seq_file *seq) | |||
2000 | struct net *net = seq_file_net(seq); | 2079 | struct net *net = seq_file_net(seq); |
2001 | void *rc = NULL; | 2080 | void *rc = NULL; |
2002 | 2081 | ||
2003 | for (st->bucket = 0; st->bucket < tcp_hashinfo.ehash_size; ++st->bucket) { | 2082 | for (st->bucket = 0; st->bucket <= tcp_hashinfo.ehash_mask; ++st->bucket) { |
2004 | struct sock *sk; | 2083 | struct sock *sk; |
2005 | struct hlist_nulls_node *node; | 2084 | struct hlist_nulls_node *node; |
2006 | struct inet_timewait_sock *tw; | 2085 | struct inet_timewait_sock *tw; |
@@ -2061,10 +2140,10 @@ get_tw: | |||
2061 | st->state = TCP_SEQ_STATE_ESTABLISHED; | 2140 | st->state = TCP_SEQ_STATE_ESTABLISHED; |
2062 | 2141 | ||
2063 | /* Look for next non empty bucket */ | 2142 | /* Look for next non empty bucket */ |
2064 | while (++st->bucket < tcp_hashinfo.ehash_size && | 2143 | while (++st->bucket <= tcp_hashinfo.ehash_mask && |
2065 | empty_bucket(st)) | 2144 | empty_bucket(st)) |
2066 | ; | 2145 | ; |
2067 | if (st->bucket >= tcp_hashinfo.ehash_size) | 2146 | if (st->bucket > tcp_hashinfo.ehash_mask) |
2068 | return NULL; | 2147 | return NULL; |
2069 | 2148 | ||
2070 | spin_lock_bh(inet_ehash_lockp(&tcp_hashinfo, st->bucket)); | 2149 | spin_lock_bh(inet_ehash_lockp(&tcp_hashinfo, st->bucket)); |
@@ -2225,7 +2304,7 @@ static void get_openreq4(struct sock *sk, struct request_sock *req, | |||
2225 | " %02X %08X:%08X %02X:%08lX %08X %5d %8d %u %d %p%n", | 2304 | " %02X %08X:%08X %02X:%08lX %08X %5d %8d %u %d %p%n", |
2226 | i, | 2305 | i, |
2227 | ireq->loc_addr, | 2306 | ireq->loc_addr, |
2228 | ntohs(inet_sk(sk)->sport), | 2307 | ntohs(inet_sk(sk)->inet_sport), |
2229 | ireq->rmt_addr, | 2308 | ireq->rmt_addr, |
2230 | ntohs(ireq->rmt_port), | 2309 | ntohs(ireq->rmt_port), |
2231 | TCP_SYN_RECV, | 2310 | TCP_SYN_RECV, |
@@ -2248,10 +2327,11 @@ static void get_tcp4_sock(struct sock *sk, struct seq_file *f, int i, int *len) | |||
2248 | struct tcp_sock *tp = tcp_sk(sk); | 2327 | struct tcp_sock *tp = tcp_sk(sk); |
2249 | const struct inet_connection_sock *icsk = inet_csk(sk); | 2328 | const struct inet_connection_sock *icsk = inet_csk(sk); |
2250 | struct inet_sock *inet = inet_sk(sk); | 2329 | struct inet_sock *inet = inet_sk(sk); |
2251 | __be32 dest = inet->daddr; | 2330 | __be32 dest = inet->inet_daddr; |
2252 | __be32 src = inet->rcv_saddr; | 2331 | __be32 src = inet->inet_rcv_saddr; |
2253 | __u16 destp = ntohs(inet->dport); | 2332 | __u16 destp = ntohs(inet->inet_dport); |
2254 | __u16 srcp = ntohs(inet->sport); | 2333 | __u16 srcp = ntohs(inet->inet_sport); |
2334 | int rx_queue; | ||
2255 | 2335 | ||
2256 | if (icsk->icsk_pending == ICSK_TIME_RETRANS) { | 2336 | if (icsk->icsk_pending == ICSK_TIME_RETRANS) { |
2257 | timer_active = 1; | 2337 | timer_active = 1; |
@@ -2267,12 +2347,19 @@ static void get_tcp4_sock(struct sock *sk, struct seq_file *f, int i, int *len) | |||
2267 | timer_expires = jiffies; | 2347 | timer_expires = jiffies; |
2268 | } | 2348 | } |
2269 | 2349 | ||
2350 | if (sk->sk_state == TCP_LISTEN) | ||
2351 | rx_queue = sk->sk_ack_backlog; | ||
2352 | else | ||
2353 | /* | ||
2354 | * because we dont lock socket, we might find a transient negative value | ||
2355 | */ | ||
2356 | rx_queue = max_t(int, tp->rcv_nxt - tp->copied_seq, 0); | ||
2357 | |||
2270 | seq_printf(f, "%4d: %08X:%04X %08X:%04X %02X %08X:%08X %02X:%08lX " | 2358 | seq_printf(f, "%4d: %08X:%04X %08X:%04X %02X %08X:%08X %02X:%08lX " |
2271 | "%08X %5d %8d %lu %d %p %lu %lu %u %u %d%n", | 2359 | "%08X %5d %8d %lu %d %p %lu %lu %u %u %d%n", |
2272 | i, src, srcp, dest, destp, sk->sk_state, | 2360 | i, src, srcp, dest, destp, sk->sk_state, |
2273 | tp->write_seq - tp->snd_una, | 2361 | tp->write_seq - tp->snd_una, |
2274 | sk->sk_state == TCP_LISTEN ? sk->sk_ack_backlog : | 2362 | rx_queue, |
2275 | (tp->rcv_nxt - tp->copied_seq), | ||
2276 | timer_active, | 2363 | timer_active, |
2277 | jiffies_to_clock_t(timer_expires - jiffies), | 2364 | jiffies_to_clock_t(timer_expires - jiffies), |
2278 | icsk->icsk_retransmits, | 2365 | icsk->icsk_retransmits, |
@@ -2354,12 +2441,12 @@ static struct tcp_seq_afinfo tcp4_seq_afinfo = { | |||
2354 | }, | 2441 | }, |
2355 | }; | 2442 | }; |
2356 | 2443 | ||
2357 | static int tcp4_proc_init_net(struct net *net) | 2444 | static int __net_init tcp4_proc_init_net(struct net *net) |
2358 | { | 2445 | { |
2359 | return tcp_proc_register(net, &tcp4_seq_afinfo); | 2446 | return tcp_proc_register(net, &tcp4_seq_afinfo); |
2360 | } | 2447 | } |
2361 | 2448 | ||
2362 | static void tcp4_proc_exit_net(struct net *net) | 2449 | static void __net_exit tcp4_proc_exit_net(struct net *net) |
2363 | { | 2450 | { |
2364 | tcp_proc_unregister(net, &tcp4_seq_afinfo); | 2451 | tcp_proc_unregister(net, &tcp4_seq_afinfo); |
2365 | } | 2452 | } |
@@ -2463,12 +2550,17 @@ static int __net_init tcp_sk_init(struct net *net) | |||
2463 | static void __net_exit tcp_sk_exit(struct net *net) | 2550 | static void __net_exit tcp_sk_exit(struct net *net) |
2464 | { | 2551 | { |
2465 | inet_ctl_sock_destroy(net->ipv4.tcp_sock); | 2552 | inet_ctl_sock_destroy(net->ipv4.tcp_sock); |
2466 | inet_twsk_purge(net, &tcp_hashinfo, &tcp_death_row, AF_INET); | 2553 | } |
2554 | |||
2555 | static void __net_exit tcp_sk_exit_batch(struct list_head *net_exit_list) | ||
2556 | { | ||
2557 | inet_twsk_purge(&tcp_hashinfo, &tcp_death_row, AF_INET); | ||
2467 | } | 2558 | } |
2468 | 2559 | ||
2469 | static struct pernet_operations __net_initdata tcp_sk_ops = { | 2560 | static struct pernet_operations __net_initdata tcp_sk_ops = { |
2470 | .init = tcp_sk_init, | 2561 | .init = tcp_sk_init, |
2471 | .exit = tcp_sk_exit, | 2562 | .exit = tcp_sk_exit, |
2563 | .exit_batch = tcp_sk_exit_batch, | ||
2472 | }; | 2564 | }; |
2473 | 2565 | ||
2474 | void __init tcp_v4_init(void) | 2566 | void __init tcp_v4_init(void) |