diff options
-rw-r--r-- | include/linux/tcp.h | 5 | ||||
-rw-r--r-- | include/net/tcp.h | 15 | ||||
-rw-r--r-- | net/ipv4/tcp_fastopen.c | 219 | ||||
-rw-r--r-- | net/ipv4/tcp_ipv4.c | 277 | ||||
-rw-r--r-- | net/ipv4/tcp_minisocks.c | 31 | ||||
-rw-r--r-- | net/ipv4/tcp_output.c | 23 | ||||
-rw-r--r-- | net/ipv6/tcp_ipv6.c | 62 |
7 files changed, 323 insertions, 309 deletions
diff --git a/include/linux/tcp.h b/include/linux/tcp.h index 4e37c71ecd74..bc35e4709e8e 100644 --- a/include/linux/tcp.h +++ b/include/linux/tcp.h | |||
@@ -366,11 +366,6 @@ static inline bool tcp_passive_fastopen(const struct sock *sk) | |||
366 | tcp_sk(sk)->fastopen_rsk != NULL); | 366 | tcp_sk(sk)->fastopen_rsk != NULL); |
367 | } | 367 | } |
368 | 368 | ||
369 | static inline bool fastopen_cookie_present(struct tcp_fastopen_cookie *foc) | ||
370 | { | ||
371 | return foc->len != -1; | ||
372 | } | ||
373 | |||
374 | extern void tcp_sock_destruct(struct sock *sk); | 369 | extern void tcp_sock_destruct(struct sock *sk); |
375 | 370 | ||
376 | static inline int fastopen_init_queue(struct sock *sk, int backlog) | 371 | static inline int fastopen_init_queue(struct sock *sk, int backlog) |
diff --git a/include/net/tcp.h b/include/net/tcp.h index 3c9418456640..f5d6ca4a9d28 100644 --- a/include/net/tcp.h +++ b/include/net/tcp.h | |||
@@ -220,8 +220,6 @@ void tcp_time_wait(struct sock *sk, int state, int timeo); | |||
220 | #define TFO_SERVER_ENABLE 2 | 220 | #define TFO_SERVER_ENABLE 2 |
221 | #define TFO_CLIENT_NO_COOKIE 4 /* Data in SYN w/o cookie option */ | 221 | #define TFO_CLIENT_NO_COOKIE 4 /* Data in SYN w/o cookie option */ |
222 | 222 | ||
223 | /* Process SYN data but skip cookie validation */ | ||
224 | #define TFO_SERVER_COOKIE_NOT_CHKED 0x100 | ||
225 | /* Accept SYN data w/o any cookie option */ | 223 | /* Accept SYN data w/o any cookie option */ |
226 | #define TFO_SERVER_COOKIE_NOT_REQD 0x200 | 224 | #define TFO_SERVER_COOKIE_NOT_REQD 0x200 |
227 | 225 | ||
@@ -230,10 +228,6 @@ void tcp_time_wait(struct sock *sk, int state, int timeo); | |||
230 | */ | 228 | */ |
231 | #define TFO_SERVER_WO_SOCKOPT1 0x400 | 229 | #define TFO_SERVER_WO_SOCKOPT1 0x400 |
232 | #define TFO_SERVER_WO_SOCKOPT2 0x800 | 230 | #define TFO_SERVER_WO_SOCKOPT2 0x800 |
233 | /* Always create TFO child sockets on a TFO listener even when | ||
234 | * cookie/data not present. (For testing purpose!) | ||
235 | */ | ||
236 | #define TFO_SERVER_ALWAYS 0x1000 | ||
237 | 231 | ||
238 | extern struct inet_timewait_death_row tcp_death_row; | 232 | extern struct inet_timewait_death_row tcp_death_row; |
239 | 233 | ||
@@ -1120,6 +1114,9 @@ static inline void tcp_openreq_init(struct request_sock *req, | |||
1120 | ireq->ir_num = ntohs(tcp_hdr(skb)->dest); | 1114 | ireq->ir_num = ntohs(tcp_hdr(skb)->dest); |
1121 | } | 1115 | } |
1122 | 1116 | ||
1117 | extern void tcp_openreq_init_rwin(struct request_sock *req, | ||
1118 | struct sock *sk, struct dst_entry *dst); | ||
1119 | |||
1123 | void tcp_enter_memory_pressure(struct sock *sk); | 1120 | void tcp_enter_memory_pressure(struct sock *sk); |
1124 | 1121 | ||
1125 | static inline int keepalive_intvl_when(const struct tcp_sock *tp) | 1122 | static inline int keepalive_intvl_when(const struct tcp_sock *tp) |
@@ -1329,8 +1326,10 @@ void tcp_free_fastopen_req(struct tcp_sock *tp); | |||
1329 | 1326 | ||
1330 | extern struct tcp_fastopen_context __rcu *tcp_fastopen_ctx; | 1327 | extern struct tcp_fastopen_context __rcu *tcp_fastopen_ctx; |
1331 | int tcp_fastopen_reset_cipher(void *key, unsigned int len); | 1328 | int tcp_fastopen_reset_cipher(void *key, unsigned int len); |
1332 | void tcp_fastopen_cookie_gen(__be32 src, __be32 dst, | 1329 | bool tcp_try_fastopen(struct sock *sk, struct sk_buff *skb, |
1333 | struct tcp_fastopen_cookie *foc); | 1330 | struct request_sock *req, |
1331 | struct tcp_fastopen_cookie *foc, | ||
1332 | struct dst_entry *dst); | ||
1334 | void tcp_fastopen_init_key_once(bool publish); | 1333 | void tcp_fastopen_init_key_once(bool publish); |
1335 | #define TCP_FASTOPEN_KEY_LENGTH 16 | 1334 | #define TCP_FASTOPEN_KEY_LENGTH 16 |
1336 | 1335 | ||
diff --git a/net/ipv4/tcp_fastopen.c b/net/ipv4/tcp_fastopen.c index f195d9316e55..62e48cf84e60 100644 --- a/net/ipv4/tcp_fastopen.c +++ b/net/ipv4/tcp_fastopen.c | |||
@@ -72,25 +72,224 @@ error: kfree(ctx); | |||
72 | return err; | 72 | return err; |
73 | } | 73 | } |
74 | 74 | ||
75 | /* Computes the fastopen cookie for the IP path. | 75 | static bool __tcp_fastopen_cookie_gen(const void *path, |
76 | * The path is a 128 bits long (pad with zeros for IPv4). | 76 | struct tcp_fastopen_cookie *foc) |
77 | * | ||
78 | * The caller must check foc->len to determine if a valid cookie | ||
79 | * has been generated successfully. | ||
80 | */ | ||
81 | void tcp_fastopen_cookie_gen(__be32 src, __be32 dst, | ||
82 | struct tcp_fastopen_cookie *foc) | ||
83 | { | 77 | { |
84 | __be32 path[4] = { src, dst, 0, 0 }; | ||
85 | struct tcp_fastopen_context *ctx; | 78 | struct tcp_fastopen_context *ctx; |
79 | bool ok = false; | ||
86 | 80 | ||
87 | tcp_fastopen_init_key_once(true); | 81 | tcp_fastopen_init_key_once(true); |
88 | 82 | ||
89 | rcu_read_lock(); | 83 | rcu_read_lock(); |
90 | ctx = rcu_dereference(tcp_fastopen_ctx); | 84 | ctx = rcu_dereference(tcp_fastopen_ctx); |
91 | if (ctx) { | 85 | if (ctx) { |
92 | crypto_cipher_encrypt_one(ctx->tfm, foc->val, (__u8 *)path); | 86 | crypto_cipher_encrypt_one(ctx->tfm, foc->val, path); |
93 | foc->len = TCP_FASTOPEN_COOKIE_SIZE; | 87 | foc->len = TCP_FASTOPEN_COOKIE_SIZE; |
88 | ok = true; | ||
94 | } | 89 | } |
95 | rcu_read_unlock(); | 90 | rcu_read_unlock(); |
91 | return ok; | ||
92 | } | ||
93 | |||
94 | /* Generate the fastopen cookie by doing aes128 encryption on both | ||
95 | * the source and destination addresses. Pad 0s for IPv4 or IPv4-mapped-IPv6 | ||
96 | * addresses. For the longer IPv6 addresses use CBC-MAC. | ||
97 | * | ||
98 | * XXX (TFO) - refactor when TCP_FASTOPEN_COOKIE_SIZE != AES_BLOCK_SIZE. | ||
99 | */ | ||
100 | static bool tcp_fastopen_cookie_gen(struct request_sock *req, | ||
101 | struct sk_buff *syn, | ||
102 | struct tcp_fastopen_cookie *foc) | ||
103 | { | ||
104 | if (req->rsk_ops->family == AF_INET) { | ||
105 | const struct iphdr *iph = ip_hdr(syn); | ||
106 | |||
107 | __be32 path[4] = { iph->saddr, iph->daddr, 0, 0 }; | ||
108 | return __tcp_fastopen_cookie_gen(path, foc); | ||
109 | } | ||
110 | |||
111 | #if IS_ENABLED(CONFIG_IPV6) | ||
112 | if (req->rsk_ops->family == AF_INET6) { | ||
113 | const struct ipv6hdr *ip6h = ipv6_hdr(syn); | ||
114 | struct tcp_fastopen_cookie tmp; | ||
115 | |||
116 | if (__tcp_fastopen_cookie_gen(&ip6h->saddr, &tmp)) { | ||
117 | struct in6_addr *buf = (struct in6_addr *) tmp.val; | ||
118 | int i = 4; | ||
119 | |||
120 | for (i = 0; i < 4; i++) | ||
121 | buf->s6_addr32[i] ^= ip6h->daddr.s6_addr32[i]; | ||
122 | return __tcp_fastopen_cookie_gen(buf, foc); | ||
123 | } | ||
124 | } | ||
125 | #endif | ||
126 | return false; | ||
127 | } | ||
128 | |||
129 | static bool tcp_fastopen_create_child(struct sock *sk, | ||
130 | struct sk_buff *skb, | ||
131 | struct dst_entry *dst, | ||
132 | struct request_sock *req) | ||
133 | { | ||
134 | struct tcp_sock *tp = tcp_sk(sk); | ||
135 | struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue; | ||
136 | struct sock *child; | ||
137 | |||
138 | req->num_retrans = 0; | ||
139 | req->num_timeout = 0; | ||
140 | req->sk = NULL; | ||
141 | |||
142 | child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL); | ||
143 | if (child == NULL) | ||
144 | return false; | ||
145 | |||
146 | spin_lock(&queue->fastopenq->lock); | ||
147 | queue->fastopenq->qlen++; | ||
148 | spin_unlock(&queue->fastopenq->lock); | ||
149 | |||
150 | /* Initialize the child socket. Have to fix some values to take | ||
151 | * into account the child is a Fast Open socket and is created | ||
152 | * only out of the bits carried in the SYN packet. | ||
153 | */ | ||
154 | tp = tcp_sk(child); | ||
155 | |||
156 | tp->fastopen_rsk = req; | ||
157 | /* Do a hold on the listner sk so that if the listener is being | ||
158 | * closed, the child that has been accepted can live on and still | ||
159 | * access listen_lock. | ||
160 | */ | ||
161 | sock_hold(sk); | ||
162 | tcp_rsk(req)->listener = sk; | ||
163 | |||
164 | /* RFC1323: The window in SYN & SYN/ACK segments is never | ||
165 | * scaled. So correct it appropriately. | ||
166 | */ | ||
167 | tp->snd_wnd = ntohs(tcp_hdr(skb)->window); | ||
168 | |||
169 | /* Activate the retrans timer so that SYNACK can be retransmitted. | ||
170 | * The request socket is not added to the SYN table of the parent | ||
171 | * because it's been added to the accept queue directly. | ||
172 | */ | ||
173 | inet_csk_reset_xmit_timer(child, ICSK_TIME_RETRANS, | ||
174 | TCP_TIMEOUT_INIT, TCP_RTO_MAX); | ||
175 | |||
176 | /* Add the child socket directly into the accept queue */ | ||
177 | inet_csk_reqsk_queue_add(sk, req, child); | ||
178 | |||
179 | /* Now finish processing the fastopen child socket. */ | ||
180 | inet_csk(child)->icsk_af_ops->rebuild_header(child); | ||
181 | tcp_init_congestion_control(child); | ||
182 | tcp_mtup_init(child); | ||
183 | tcp_init_metrics(child); | ||
184 | tcp_init_buffer_space(child); | ||
185 | |||
186 | /* Queue the data carried in the SYN packet. We need to first | ||
187 | * bump skb's refcnt because the caller will attempt to free it. | ||
188 | * | ||
189 | * XXX (TFO) - we honor a zero-payload TFO request for now, | ||
190 | * (any reason not to?) but no need to queue the skb since | ||
191 | * there is no data. How about SYN+FIN? | ||
192 | */ | ||
193 | if (TCP_SKB_CB(skb)->end_seq != TCP_SKB_CB(skb)->seq + 1) { | ||
194 | skb = skb_get(skb); | ||
195 | skb_dst_drop(skb); | ||
196 | __skb_pull(skb, tcp_hdr(skb)->doff * 4); | ||
197 | skb_set_owner_r(skb, child); | ||
198 | __skb_queue_tail(&child->sk_receive_queue, skb); | ||
199 | tp->syn_data_acked = 1; | ||
200 | } | ||
201 | tcp_rsk(req)->rcv_nxt = tp->rcv_nxt = TCP_SKB_CB(skb)->end_seq; | ||
202 | sk->sk_data_ready(sk); | ||
203 | bh_unlock_sock(child); | ||
204 | sock_put(child); | ||
205 | WARN_ON(req->sk == NULL); | ||
206 | return true; | ||
207 | } | ||
208 | EXPORT_SYMBOL(tcp_fastopen_create_child); | ||
209 | |||
210 | static bool tcp_fastopen_queue_check(struct sock *sk) | ||
211 | { | ||
212 | struct fastopen_queue *fastopenq; | ||
213 | |||
214 | /* Make sure the listener has enabled fastopen, and we don't | ||
215 | * exceed the max # of pending TFO requests allowed before trying | ||
216 | * to validating the cookie in order to avoid burning CPU cycles | ||
217 | * unnecessarily. | ||
218 | * | ||
219 | * XXX (TFO) - The implication of checking the max_qlen before | ||
220 | * processing a cookie request is that clients can't differentiate | ||
221 | * between qlen overflow causing Fast Open to be disabled | ||
222 | * temporarily vs a server not supporting Fast Open at all. | ||
223 | */ | ||
224 | fastopenq = inet_csk(sk)->icsk_accept_queue.fastopenq; | ||
225 | if (fastopenq == NULL || fastopenq->max_qlen == 0) | ||
226 | return false; | ||
227 | |||
228 | if (fastopenq->qlen >= fastopenq->max_qlen) { | ||
229 | struct request_sock *req1; | ||
230 | spin_lock(&fastopenq->lock); | ||
231 | req1 = fastopenq->rskq_rst_head; | ||
232 | if ((req1 == NULL) || time_after(req1->expires, jiffies)) { | ||
233 | spin_unlock(&fastopenq->lock); | ||
234 | NET_INC_STATS_BH(sock_net(sk), | ||
235 | LINUX_MIB_TCPFASTOPENLISTENOVERFLOW); | ||
236 | return false; | ||
237 | } | ||
238 | fastopenq->rskq_rst_head = req1->dl_next; | ||
239 | fastopenq->qlen--; | ||
240 | spin_unlock(&fastopenq->lock); | ||
241 | reqsk_free(req1); | ||
242 | } | ||
243 | return true; | ||
244 | } | ||
245 | |||
246 | /* Returns true if we should perform Fast Open on the SYN. The cookie (foc) | ||
247 | * may be updated and return the client in the SYN-ACK later. E.g., Fast Open | ||
248 | * cookie request (foc->len == 0). | ||
249 | */ | ||
250 | bool tcp_try_fastopen(struct sock *sk, struct sk_buff *skb, | ||
251 | struct request_sock *req, | ||
252 | struct tcp_fastopen_cookie *foc, | ||
253 | struct dst_entry *dst) | ||
254 | { | ||
255 | struct tcp_fastopen_cookie valid_foc = { .len = -1 }; | ||
256 | bool syn_data = TCP_SKB_CB(skb)->end_seq != TCP_SKB_CB(skb)->seq + 1; | ||
257 | |||
258 | if (!((sysctl_tcp_fastopen & TFO_SERVER_ENABLE) && | ||
259 | (syn_data || foc->len >= 0) && | ||
260 | tcp_fastopen_queue_check(sk))) { | ||
261 | foc->len = -1; | ||
262 | return false; | ||
263 | } | ||
264 | |||
265 | if (syn_data && (sysctl_tcp_fastopen & TFO_SERVER_COOKIE_NOT_REQD)) | ||
266 | goto fastopen; | ||
267 | |||
268 | if (tcp_fastopen_cookie_gen(req, skb, &valid_foc) && | ||
269 | foc->len == TCP_FASTOPEN_COOKIE_SIZE && | ||
270 | foc->len == valid_foc.len && | ||
271 | !memcmp(foc->val, valid_foc.val, foc->len)) { | ||
272 | /* Cookie is valid. Create a (full) child socket to accept | ||
273 | * the data in SYN before returning a SYN-ACK to ack the | ||
274 | * data. If we fail to create the socket, fall back and | ||
275 | * ack the ISN only but includes the same cookie. | ||
276 | * | ||
277 | * Note: Data-less SYN with valid cookie is allowed to send | ||
278 | * data in SYN_RECV state. | ||
279 | */ | ||
280 | fastopen: | ||
281 | if (tcp_fastopen_create_child(sk, skb, dst, req)) { | ||
282 | foc->len = -1; | ||
283 | NET_INC_STATS_BH(sock_net(sk), | ||
284 | LINUX_MIB_TCPFASTOPENPASSIVE); | ||
285 | return true; | ||
286 | } | ||
287 | } | ||
288 | |||
289 | NET_INC_STATS_BH(sock_net(sk), foc->len ? | ||
290 | LINUX_MIB_TCPFASTOPENPASSIVEFAIL : | ||
291 | LINUX_MIB_TCPFASTOPENCOOKIEREQD); | ||
292 | *foc = valid_foc; | ||
293 | return false; | ||
96 | } | 294 | } |
295 | EXPORT_SYMBOL(tcp_try_fastopen); | ||
diff --git a/net/ipv4/tcp_ipv4.c b/net/ipv4/tcp_ipv4.c index ad166dcc278f..a2780e5334c9 100644 --- a/net/ipv4/tcp_ipv4.c +++ b/net/ipv4/tcp_ipv4.c | |||
@@ -336,8 +336,8 @@ void tcp_v4_err(struct sk_buff *icmp_skb, u32 info) | |||
336 | const int code = icmp_hdr(icmp_skb)->code; | 336 | const int code = icmp_hdr(icmp_skb)->code; |
337 | struct sock *sk; | 337 | struct sock *sk; |
338 | struct sk_buff *skb; | 338 | struct sk_buff *skb; |
339 | struct request_sock *req; | 339 | struct request_sock *fastopen; |
340 | __u32 seq; | 340 | __u32 seq, snd_una; |
341 | __u32 remaining; | 341 | __u32 remaining; |
342 | int err; | 342 | int err; |
343 | struct net *net = dev_net(icmp_skb->dev); | 343 | struct net *net = dev_net(icmp_skb->dev); |
@@ -378,12 +378,12 @@ void tcp_v4_err(struct sk_buff *icmp_skb, u32 info) | |||
378 | 378 | ||
379 | icsk = inet_csk(sk); | 379 | icsk = inet_csk(sk); |
380 | tp = tcp_sk(sk); | 380 | tp = tcp_sk(sk); |
381 | req = tp->fastopen_rsk; | ||
382 | seq = ntohl(th->seq); | 381 | seq = ntohl(th->seq); |
382 | /* XXX (TFO) - tp->snd_una should be ISN (tcp_create_openreq_child() */ | ||
383 | fastopen = tp->fastopen_rsk; | ||
384 | snd_una = fastopen ? tcp_rsk(fastopen)->snt_isn : tp->snd_una; | ||
383 | if (sk->sk_state != TCP_LISTEN && | 385 | if (sk->sk_state != TCP_LISTEN && |
384 | !between(seq, tp->snd_una, tp->snd_nxt) && | 386 | !between(seq, snd_una, tp->snd_nxt)) { |
385 | (req == NULL || seq != tcp_rsk(req)->snt_isn)) { | ||
386 | /* For a Fast Open socket, allow seq to be snt_isn. */ | ||
387 | NET_INC_STATS_BH(net, LINUX_MIB_OUTOFWINDOWICMPS); | 387 | NET_INC_STATS_BH(net, LINUX_MIB_OUTOFWINDOWICMPS); |
388 | goto out; | 388 | goto out; |
389 | } | 389 | } |
@@ -426,11 +426,9 @@ void tcp_v4_err(struct sk_buff *icmp_skb, u32 info) | |||
426 | if (code != ICMP_NET_UNREACH && code != ICMP_HOST_UNREACH) | 426 | if (code != ICMP_NET_UNREACH && code != ICMP_HOST_UNREACH) |
427 | break; | 427 | break; |
428 | if (seq != tp->snd_una || !icsk->icsk_retransmits || | 428 | if (seq != tp->snd_una || !icsk->icsk_retransmits || |
429 | !icsk->icsk_backoff) | 429 | !icsk->icsk_backoff || fastopen) |
430 | break; | 430 | break; |
431 | 431 | ||
432 | /* XXX (TFO) - revisit the following logic for TFO */ | ||
433 | |||
434 | if (sock_owned_by_user(sk)) | 432 | if (sock_owned_by_user(sk)) |
435 | break; | 433 | break; |
436 | 434 | ||
@@ -462,14 +460,6 @@ void tcp_v4_err(struct sk_buff *icmp_skb, u32 info) | |||
462 | goto out; | 460 | goto out; |
463 | } | 461 | } |
464 | 462 | ||
465 | /* XXX (TFO) - if it's a TFO socket and has been accepted, rather | ||
466 | * than following the TCP_SYN_RECV case and closing the socket, | ||
467 | * we ignore the ICMP error and keep trying like a fully established | ||
468 | * socket. Is this the right thing to do? | ||
469 | */ | ||
470 | if (req && req->sk == NULL) | ||
471 | goto out; | ||
472 | |||
473 | switch (sk->sk_state) { | 463 | switch (sk->sk_state) { |
474 | struct request_sock *req, **prev; | 464 | struct request_sock *req, **prev; |
475 | case TCP_LISTEN: | 465 | case TCP_LISTEN: |
@@ -502,10 +492,13 @@ void tcp_v4_err(struct sk_buff *icmp_skb, u32 info) | |||
502 | goto out; | 492 | goto out; |
503 | 493 | ||
504 | case TCP_SYN_SENT: | 494 | case TCP_SYN_SENT: |
505 | case TCP_SYN_RECV: /* Cannot happen. | 495 | case TCP_SYN_RECV: |
506 | It can f.e. if SYNs crossed, | 496 | /* Only in fast or simultaneous open. If a fast open socket is |
507 | or Fast Open. | 497 | * is already accepted it is treated as a connected one below. |
508 | */ | 498 | */ |
499 | if (fastopen && fastopen->sk == NULL) | ||
500 | break; | ||
501 | |||
509 | if (!sock_owned_by_user(sk)) { | 502 | if (!sock_owned_by_user(sk)) { |
510 | sk->sk_err = err; | 503 | sk->sk_err = err; |
511 | 504 | ||
@@ -822,7 +815,8 @@ static void tcp_v4_reqsk_send_ack(struct sock *sk, struct sk_buff *skb, | |||
822 | */ | 815 | */ |
823 | static int tcp_v4_send_synack(struct sock *sk, struct dst_entry *dst, | 816 | static int tcp_v4_send_synack(struct sock *sk, struct dst_entry *dst, |
824 | struct request_sock *req, | 817 | struct request_sock *req, |
825 | u16 queue_mapping) | 818 | u16 queue_mapping, |
819 | struct tcp_fastopen_cookie *foc) | ||
826 | { | 820 | { |
827 | const struct inet_request_sock *ireq = inet_rsk(req); | 821 | const struct inet_request_sock *ireq = inet_rsk(req); |
828 | struct flowi4 fl4; | 822 | struct flowi4 fl4; |
@@ -833,7 +827,7 @@ static int tcp_v4_send_synack(struct sock *sk, struct dst_entry *dst, | |||
833 | if (!dst && (dst = inet_csk_route_req(sk, &fl4, req)) == NULL) | 827 | if (!dst && (dst = inet_csk_route_req(sk, &fl4, req)) == NULL) |
834 | return -1; | 828 | return -1; |
835 | 829 | ||
836 | skb = tcp_make_synack(sk, dst, req, NULL); | 830 | skb = tcp_make_synack(sk, dst, req, foc); |
837 | 831 | ||
838 | if (skb) { | 832 | if (skb) { |
839 | __tcp_v4_send_check(skb, ireq->ir_loc_addr, ireq->ir_rmt_addr); | 833 | __tcp_v4_send_check(skb, ireq->ir_loc_addr, ireq->ir_rmt_addr); |
@@ -852,7 +846,7 @@ static int tcp_v4_send_synack(struct sock *sk, struct dst_entry *dst, | |||
852 | 846 | ||
853 | static int tcp_v4_rtx_synack(struct sock *sk, struct request_sock *req) | 847 | static int tcp_v4_rtx_synack(struct sock *sk, struct request_sock *req) |
854 | { | 848 | { |
855 | int res = tcp_v4_send_synack(sk, NULL, req, 0); | 849 | int res = tcp_v4_send_synack(sk, NULL, req, 0, NULL); |
856 | 850 | ||
857 | if (!res) { | 851 | if (!res) { |
858 | TCP_INC_STATS_BH(sock_net(sk), TCP_MIB_RETRANSSEGS); | 852 | TCP_INC_STATS_BH(sock_net(sk), TCP_MIB_RETRANSSEGS); |
@@ -1260,187 +1254,6 @@ static const struct tcp_request_sock_ops tcp_request_sock_ipv4_ops = { | |||
1260 | }; | 1254 | }; |
1261 | #endif | 1255 | #endif |
1262 | 1256 | ||
1263 | static bool tcp_fastopen_check(struct sock *sk, struct sk_buff *skb, | ||
1264 | struct request_sock *req, | ||
1265 | struct tcp_fastopen_cookie *foc, | ||
1266 | struct tcp_fastopen_cookie *valid_foc) | ||
1267 | { | ||
1268 | bool skip_cookie = false; | ||
1269 | struct fastopen_queue *fastopenq; | ||
1270 | |||
1271 | if (likely(!fastopen_cookie_present(foc))) { | ||
1272 | /* See include/net/tcp.h for the meaning of these knobs */ | ||
1273 | if ((sysctl_tcp_fastopen & TFO_SERVER_ALWAYS) || | ||
1274 | ((sysctl_tcp_fastopen & TFO_SERVER_COOKIE_NOT_REQD) && | ||
1275 | (TCP_SKB_CB(skb)->end_seq != TCP_SKB_CB(skb)->seq + 1))) | ||
1276 | skip_cookie = true; /* no cookie to validate */ | ||
1277 | else | ||
1278 | return false; | ||
1279 | } | ||
1280 | fastopenq = inet_csk(sk)->icsk_accept_queue.fastopenq; | ||
1281 | /* A FO option is present; bump the counter. */ | ||
1282 | NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_TCPFASTOPENPASSIVE); | ||
1283 | |||
1284 | /* Make sure the listener has enabled fastopen, and we don't | ||
1285 | * exceed the max # of pending TFO requests allowed before trying | ||
1286 | * to validating the cookie in order to avoid burning CPU cycles | ||
1287 | * unnecessarily. | ||
1288 | * | ||
1289 | * XXX (TFO) - The implication of checking the max_qlen before | ||
1290 | * processing a cookie request is that clients can't differentiate | ||
1291 | * between qlen overflow causing Fast Open to be disabled | ||
1292 | * temporarily vs a server not supporting Fast Open at all. | ||
1293 | */ | ||
1294 | if ((sysctl_tcp_fastopen & TFO_SERVER_ENABLE) == 0 || | ||
1295 | fastopenq == NULL || fastopenq->max_qlen == 0) | ||
1296 | return false; | ||
1297 | |||
1298 | if (fastopenq->qlen >= fastopenq->max_qlen) { | ||
1299 | struct request_sock *req1; | ||
1300 | spin_lock(&fastopenq->lock); | ||
1301 | req1 = fastopenq->rskq_rst_head; | ||
1302 | if ((req1 == NULL) || time_after(req1->expires, jiffies)) { | ||
1303 | spin_unlock(&fastopenq->lock); | ||
1304 | NET_INC_STATS_BH(sock_net(sk), | ||
1305 | LINUX_MIB_TCPFASTOPENLISTENOVERFLOW); | ||
1306 | /* Avoid bumping LINUX_MIB_TCPFASTOPENPASSIVEFAIL*/ | ||
1307 | foc->len = -1; | ||
1308 | return false; | ||
1309 | } | ||
1310 | fastopenq->rskq_rst_head = req1->dl_next; | ||
1311 | fastopenq->qlen--; | ||
1312 | spin_unlock(&fastopenq->lock); | ||
1313 | reqsk_free(req1); | ||
1314 | } | ||
1315 | if (skip_cookie) { | ||
1316 | tcp_rsk(req)->rcv_nxt = TCP_SKB_CB(skb)->end_seq; | ||
1317 | return true; | ||
1318 | } | ||
1319 | |||
1320 | if (foc->len == TCP_FASTOPEN_COOKIE_SIZE) { | ||
1321 | if ((sysctl_tcp_fastopen & TFO_SERVER_COOKIE_NOT_CHKED) == 0) { | ||
1322 | tcp_fastopen_cookie_gen(ip_hdr(skb)->saddr, | ||
1323 | ip_hdr(skb)->daddr, valid_foc); | ||
1324 | if ((valid_foc->len != TCP_FASTOPEN_COOKIE_SIZE) || | ||
1325 | memcmp(&foc->val[0], &valid_foc->val[0], | ||
1326 | TCP_FASTOPEN_COOKIE_SIZE) != 0) | ||
1327 | return false; | ||
1328 | valid_foc->len = -1; | ||
1329 | } | ||
1330 | /* Acknowledge the data received from the peer. */ | ||
1331 | tcp_rsk(req)->rcv_nxt = TCP_SKB_CB(skb)->end_seq; | ||
1332 | return true; | ||
1333 | } else if (foc->len == 0) { /* Client requesting a cookie */ | ||
1334 | tcp_fastopen_cookie_gen(ip_hdr(skb)->saddr, | ||
1335 | ip_hdr(skb)->daddr, valid_foc); | ||
1336 | NET_INC_STATS_BH(sock_net(sk), | ||
1337 | LINUX_MIB_TCPFASTOPENCOOKIEREQD); | ||
1338 | } else { | ||
1339 | /* Client sent a cookie with wrong size. Treat it | ||
1340 | * the same as invalid and return a valid one. | ||
1341 | */ | ||
1342 | tcp_fastopen_cookie_gen(ip_hdr(skb)->saddr, | ||
1343 | ip_hdr(skb)->daddr, valid_foc); | ||
1344 | } | ||
1345 | return false; | ||
1346 | } | ||
1347 | |||
1348 | static int tcp_v4_conn_req_fastopen(struct sock *sk, | ||
1349 | struct sk_buff *skb, | ||
1350 | struct sk_buff *skb_synack, | ||
1351 | struct request_sock *req) | ||
1352 | { | ||
1353 | struct tcp_sock *tp = tcp_sk(sk); | ||
1354 | struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue; | ||
1355 | const struct inet_request_sock *ireq = inet_rsk(req); | ||
1356 | struct sock *child; | ||
1357 | int err; | ||
1358 | |||
1359 | req->num_retrans = 0; | ||
1360 | req->num_timeout = 0; | ||
1361 | req->sk = NULL; | ||
1362 | |||
1363 | child = inet_csk(sk)->icsk_af_ops->syn_recv_sock(sk, skb, req, NULL); | ||
1364 | if (child == NULL) { | ||
1365 | NET_INC_STATS_BH(sock_net(sk), | ||
1366 | LINUX_MIB_TCPFASTOPENPASSIVEFAIL); | ||
1367 | kfree_skb(skb_synack); | ||
1368 | return -1; | ||
1369 | } | ||
1370 | err = ip_build_and_send_pkt(skb_synack, sk, ireq->ir_loc_addr, | ||
1371 | ireq->ir_rmt_addr, ireq->opt); | ||
1372 | err = net_xmit_eval(err); | ||
1373 | if (!err) | ||
1374 | tcp_rsk(req)->snt_synack = tcp_time_stamp; | ||
1375 | /* XXX (TFO) - is it ok to ignore error and continue? */ | ||
1376 | |||
1377 | spin_lock(&queue->fastopenq->lock); | ||
1378 | queue->fastopenq->qlen++; | ||
1379 | spin_unlock(&queue->fastopenq->lock); | ||
1380 | |||
1381 | /* Initialize the child socket. Have to fix some values to take | ||
1382 | * into account the child is a Fast Open socket and is created | ||
1383 | * only out of the bits carried in the SYN packet. | ||
1384 | */ | ||
1385 | tp = tcp_sk(child); | ||
1386 | |||
1387 | tp->fastopen_rsk = req; | ||
1388 | /* Do a hold on the listner sk so that if the listener is being | ||
1389 | * closed, the child that has been accepted can live on and still | ||
1390 | * access listen_lock. | ||
1391 | */ | ||
1392 | sock_hold(sk); | ||
1393 | tcp_rsk(req)->listener = sk; | ||
1394 | |||
1395 | /* RFC1323: The window in SYN & SYN/ACK segments is never | ||
1396 | * scaled. So correct it appropriately. | ||
1397 | */ | ||
1398 | tp->snd_wnd = ntohs(tcp_hdr(skb)->window); | ||
1399 | |||
1400 | /* Activate the retrans timer so that SYNACK can be retransmitted. | ||
1401 | * The request socket is not added to the SYN table of the parent | ||
1402 | * because it's been added to the accept queue directly. | ||
1403 | */ | ||
1404 | inet_csk_reset_xmit_timer(child, ICSK_TIME_RETRANS, | ||
1405 | TCP_TIMEOUT_INIT, TCP_RTO_MAX); | ||
1406 | |||
1407 | /* Add the child socket directly into the accept queue */ | ||
1408 | inet_csk_reqsk_queue_add(sk, req, child); | ||
1409 | |||
1410 | /* Now finish processing the fastopen child socket. */ | ||
1411 | inet_csk(child)->icsk_af_ops->rebuild_header(child); | ||
1412 | tcp_init_congestion_control(child); | ||
1413 | tcp_mtup_init(child); | ||
1414 | tcp_init_metrics(child); | ||
1415 | tcp_init_buffer_space(child); | ||
1416 | |||
1417 | /* Queue the data carried in the SYN packet. We need to first | ||
1418 | * bump skb's refcnt because the caller will attempt to free it. | ||
1419 | * | ||
1420 | * XXX (TFO) - we honor a zero-payload TFO request for now. | ||
1421 | * (Any reason not to?) | ||
1422 | */ | ||
1423 | if (TCP_SKB_CB(skb)->end_seq == TCP_SKB_CB(skb)->seq + 1) { | ||
1424 | /* Don't queue the skb if there is no payload in SYN. | ||
1425 | * XXX (TFO) - How about SYN+FIN? | ||
1426 | */ | ||
1427 | tp->rcv_nxt = TCP_SKB_CB(skb)->end_seq; | ||
1428 | } else { | ||
1429 | skb = skb_get(skb); | ||
1430 | skb_dst_drop(skb); | ||
1431 | __skb_pull(skb, tcp_hdr(skb)->doff * 4); | ||
1432 | skb_set_owner_r(skb, child); | ||
1433 | __skb_queue_tail(&child->sk_receive_queue, skb); | ||
1434 | tp->rcv_nxt = TCP_SKB_CB(skb)->end_seq; | ||
1435 | tp->syn_data_acked = 1; | ||
1436 | } | ||
1437 | sk->sk_data_ready(sk); | ||
1438 | bh_unlock_sock(child); | ||
1439 | sock_put(child); | ||
1440 | WARN_ON(req->sk == NULL); | ||
1441 | return 0; | ||
1442 | } | ||
1443 | |||
1444 | int tcp_v4_conn_request(struct sock *sk, struct sk_buff *skb) | 1257 | int tcp_v4_conn_request(struct sock *sk, struct sk_buff *skb) |
1445 | { | 1258 | { |
1446 | struct tcp_options_received tmp_opt; | 1259 | struct tcp_options_received tmp_opt; |
@@ -1451,12 +1264,10 @@ int tcp_v4_conn_request(struct sock *sk, struct sk_buff *skb) | |||
1451 | __be32 saddr = ip_hdr(skb)->saddr; | 1264 | __be32 saddr = ip_hdr(skb)->saddr; |
1452 | __be32 daddr = ip_hdr(skb)->daddr; | 1265 | __be32 daddr = ip_hdr(skb)->daddr; |
1453 | __u32 isn = TCP_SKB_CB(skb)->when; | 1266 | __u32 isn = TCP_SKB_CB(skb)->when; |
1454 | bool want_cookie = false; | 1267 | bool want_cookie = false, fastopen; |
1455 | struct flowi4 fl4; | 1268 | struct flowi4 fl4; |
1456 | struct tcp_fastopen_cookie foc = { .len = -1 }; | 1269 | struct tcp_fastopen_cookie foc = { .len = -1 }; |
1457 | struct tcp_fastopen_cookie valid_foc = { .len = -1 }; | 1270 | int err; |
1458 | struct sk_buff *skb_synack; | ||
1459 | int do_fastopen; | ||
1460 | 1271 | ||
1461 | /* Never answer to SYNs send to broadcast or multicast */ | 1272 | /* Never answer to SYNs send to broadcast or multicast */ |
1462 | if (skb_rtable(skb)->rt_flags & (RTCF_BROADCAST | RTCF_MULTICAST)) | 1273 | if (skb_rtable(skb)->rt_flags & (RTCF_BROADCAST | RTCF_MULTICAST)) |
@@ -1555,52 +1366,24 @@ int tcp_v4_conn_request(struct sock *sk, struct sk_buff *skb) | |||
1555 | 1366 | ||
1556 | isn = tcp_v4_init_sequence(skb); | 1367 | isn = tcp_v4_init_sequence(skb); |
1557 | } | 1368 | } |
1558 | tcp_rsk(req)->snt_isn = isn; | 1369 | if (!dst && (dst = inet_csk_route_req(sk, &fl4, req)) == NULL) |
1559 | |||
1560 | if (dst == NULL) { | ||
1561 | dst = inet_csk_route_req(sk, &fl4, req); | ||
1562 | if (dst == NULL) | ||
1563 | goto drop_and_free; | ||
1564 | } | ||
1565 | do_fastopen = tcp_fastopen_check(sk, skb, req, &foc, &valid_foc); | ||
1566 | |||
1567 | /* We don't call tcp_v4_send_synack() directly because we need | ||
1568 | * to make sure a child socket can be created successfully before | ||
1569 | * sending back synack! | ||
1570 | * | ||
1571 | * XXX (TFO) - Ideally one would simply call tcp_v4_send_synack() | ||
1572 | * (or better yet, call tcp_send_synack() in the child context | ||
1573 | * directly, but will have to fix bunch of other code first) | ||
1574 | * after syn_recv_sock() except one will need to first fix the | ||
1575 | * latter to remove its dependency on the current implementation | ||
1576 | * of tcp_v4_send_synack()->tcp_select_initial_window(). | ||
1577 | */ | ||
1578 | skb_synack = tcp_make_synack(sk, dst, req, | ||
1579 | fastopen_cookie_present(&valid_foc) ? &valid_foc : NULL); | ||
1580 | |||
1581 | if (skb_synack) { | ||
1582 | __tcp_v4_send_check(skb_synack, ireq->ir_loc_addr, ireq->ir_rmt_addr); | ||
1583 | skb_set_queue_mapping(skb_synack, skb_get_queue_mapping(skb)); | ||
1584 | } else | ||
1585 | goto drop_and_free; | 1370 | goto drop_and_free; |
1586 | 1371 | ||
1587 | if (likely(!do_fastopen)) { | 1372 | tcp_rsk(req)->snt_isn = isn; |
1588 | int err; | 1373 | tcp_rsk(req)->snt_synack = tcp_time_stamp; |
1589 | err = ip_build_and_send_pkt(skb_synack, sk, ireq->ir_loc_addr, | 1374 | tcp_openreq_init_rwin(req, sk, dst); |
1590 | ireq->ir_rmt_addr, ireq->opt); | 1375 | fastopen = !want_cookie && |
1591 | err = net_xmit_eval(err); | 1376 | tcp_try_fastopen(sk, skb, req, &foc, dst); |
1377 | err = tcp_v4_send_synack(sk, dst, req, | ||
1378 | skb_get_queue_mapping(skb), &foc); | ||
1379 | if (!fastopen) { | ||
1592 | if (err || want_cookie) | 1380 | if (err || want_cookie) |
1593 | goto drop_and_free; | 1381 | goto drop_and_free; |
1594 | 1382 | ||
1595 | tcp_rsk(req)->snt_synack = tcp_time_stamp; | 1383 | tcp_rsk(req)->snt_synack = tcp_time_stamp; |
1596 | tcp_rsk(req)->listener = NULL; | 1384 | tcp_rsk(req)->listener = NULL; |
1597 | /* Add the request_sock to the SYN table */ | ||
1598 | inet_csk_reqsk_queue_hash_add(sk, req, TCP_TIMEOUT_INIT); | 1385 | inet_csk_reqsk_queue_hash_add(sk, req, TCP_TIMEOUT_INIT); |
1599 | if (fastopen_cookie_present(&foc) && foc.len != 0) | 1386 | } |
1600 | NET_INC_STATS_BH(sock_net(sk), | ||
1601 | LINUX_MIB_TCPFASTOPENPASSIVEFAIL); | ||
1602 | } else if (tcp_v4_conn_req_fastopen(sk, skb, skb_synack, req)) | ||
1603 | goto drop_and_free; | ||
1604 | 1387 | ||
1605 | return 0; | 1388 | return 0; |
1606 | 1389 | ||
diff --git a/net/ipv4/tcp_minisocks.c b/net/ipv4/tcp_minisocks.c index 05c1b155251d..e68e0d4af6c9 100644 --- a/net/ipv4/tcp_minisocks.c +++ b/net/ipv4/tcp_minisocks.c | |||
@@ -362,6 +362,37 @@ void tcp_twsk_destructor(struct sock *sk) | |||
362 | } | 362 | } |
363 | EXPORT_SYMBOL_GPL(tcp_twsk_destructor); | 363 | EXPORT_SYMBOL_GPL(tcp_twsk_destructor); |
364 | 364 | ||
365 | void tcp_openreq_init_rwin(struct request_sock *req, | ||
366 | struct sock *sk, struct dst_entry *dst) | ||
367 | { | ||
368 | struct inet_request_sock *ireq = inet_rsk(req); | ||
369 | struct tcp_sock *tp = tcp_sk(sk); | ||
370 | __u8 rcv_wscale; | ||
371 | int mss = dst_metric_advmss(dst); | ||
372 | |||
373 | if (tp->rx_opt.user_mss && tp->rx_opt.user_mss < mss) | ||
374 | mss = tp->rx_opt.user_mss; | ||
375 | |||
376 | /* Set this up on the first call only */ | ||
377 | req->window_clamp = tp->window_clamp ? : dst_metric(dst, RTAX_WINDOW); | ||
378 | |||
379 | /* limit the window selection if the user enforce a smaller rx buffer */ | ||
380 | if (sk->sk_userlocks & SOCK_RCVBUF_LOCK && | ||
381 | (req->window_clamp > tcp_full_space(sk) || req->window_clamp == 0)) | ||
382 | req->window_clamp = tcp_full_space(sk); | ||
383 | |||
384 | /* tcp_full_space because it is guaranteed to be the first packet */ | ||
385 | tcp_select_initial_window(tcp_full_space(sk), | ||
386 | mss - (ireq->tstamp_ok ? TCPOLEN_TSTAMP_ALIGNED : 0), | ||
387 | &req->rcv_wnd, | ||
388 | &req->window_clamp, | ||
389 | ireq->wscale_ok, | ||
390 | &rcv_wscale, | ||
391 | dst_metric(dst, RTAX_INITRWND)); | ||
392 | ireq->rcv_wscale = rcv_wscale; | ||
393 | } | ||
394 | EXPORT_SYMBOL(tcp_openreq_init_rwin); | ||
395 | |||
365 | static inline void TCP_ECN_openreq_child(struct tcp_sock *tp, | 396 | static inline void TCP_ECN_openreq_child(struct tcp_sock *tp, |
366 | struct request_sock *req) | 397 | struct request_sock *req) |
367 | { | 398 | { |
diff --git a/net/ipv4/tcp_output.c b/net/ipv4/tcp_output.c index 694711a140d4..3d61c52bdf79 100644 --- a/net/ipv4/tcp_output.c +++ b/net/ipv4/tcp_output.c | |||
@@ -627,7 +627,7 @@ static unsigned int tcp_synack_options(struct sock *sk, | |||
627 | if (unlikely(!ireq->tstamp_ok)) | 627 | if (unlikely(!ireq->tstamp_ok)) |
628 | remaining -= TCPOLEN_SACKPERM_ALIGNED; | 628 | remaining -= TCPOLEN_SACKPERM_ALIGNED; |
629 | } | 629 | } |
630 | if (foc != NULL) { | 630 | if (foc != NULL && foc->len >= 0) { |
631 | u32 need = TCPOLEN_EXP_FASTOPEN_BASE + foc->len; | 631 | u32 need = TCPOLEN_EXP_FASTOPEN_BASE + foc->len; |
632 | need = (need + 3) & ~3U; /* Align to 32 bits */ | 632 | need = (need + 3) & ~3U; /* Align to 32 bits */ |
633 | if (remaining >= need) { | 633 | if (remaining >= need) { |
@@ -2803,27 +2803,6 @@ struct sk_buff *tcp_make_synack(struct sock *sk, struct dst_entry *dst, | |||
2803 | if (tp->rx_opt.user_mss && tp->rx_opt.user_mss < mss) | 2803 | if (tp->rx_opt.user_mss && tp->rx_opt.user_mss < mss) |
2804 | mss = tp->rx_opt.user_mss; | 2804 | mss = tp->rx_opt.user_mss; |
2805 | 2805 | ||
2806 | if (req->rcv_wnd == 0) { /* ignored for retransmitted syns */ | ||
2807 | __u8 rcv_wscale; | ||
2808 | /* Set this up on the first call only */ | ||
2809 | req->window_clamp = tp->window_clamp ? : dst_metric(dst, RTAX_WINDOW); | ||
2810 | |||
2811 | /* limit the window selection if the user enforce a smaller rx buffer */ | ||
2812 | if (sk->sk_userlocks & SOCK_RCVBUF_LOCK && | ||
2813 | (req->window_clamp > tcp_full_space(sk) || req->window_clamp == 0)) | ||
2814 | req->window_clamp = tcp_full_space(sk); | ||
2815 | |||
2816 | /* tcp_full_space because it is guaranteed to be the first packet */ | ||
2817 | tcp_select_initial_window(tcp_full_space(sk), | ||
2818 | mss - (ireq->tstamp_ok ? TCPOLEN_TSTAMP_ALIGNED : 0), | ||
2819 | &req->rcv_wnd, | ||
2820 | &req->window_clamp, | ||
2821 | ireq->wscale_ok, | ||
2822 | &rcv_wscale, | ||
2823 | dst_metric(dst, RTAX_INITRWND)); | ||
2824 | ireq->rcv_wscale = rcv_wscale; | ||
2825 | } | ||
2826 | |||
2827 | memset(&opts, 0, sizeof(opts)); | 2806 | memset(&opts, 0, sizeof(opts)); |
2828 | #ifdef CONFIG_SYN_COOKIES | 2807 | #ifdef CONFIG_SYN_COOKIES |
2829 | if (unlikely(req->cookie_ts)) | 2808 | if (unlikely(req->cookie_ts)) |
diff --git a/net/ipv6/tcp_ipv6.c b/net/ipv6/tcp_ipv6.c index 7fa67439f4d6..3a267bf14f2f 100644 --- a/net/ipv6/tcp_ipv6.c +++ b/net/ipv6/tcp_ipv6.c | |||
@@ -340,7 +340,8 @@ static void tcp_v6_err(struct sk_buff *skb, struct inet6_skb_parm *opt, | |||
340 | struct sock *sk; | 340 | struct sock *sk; |
341 | int err; | 341 | int err; |
342 | struct tcp_sock *tp; | 342 | struct tcp_sock *tp; |
343 | __u32 seq; | 343 | struct request_sock *fastopen; |
344 | __u32 seq, snd_una; | ||
344 | struct net *net = dev_net(skb->dev); | 345 | struct net *net = dev_net(skb->dev); |
345 | 346 | ||
346 | sk = inet6_lookup(net, &tcp_hashinfo, &hdr->daddr, | 347 | sk = inet6_lookup(net, &tcp_hashinfo, &hdr->daddr, |
@@ -371,8 +372,11 @@ static void tcp_v6_err(struct sk_buff *skb, struct inet6_skb_parm *opt, | |||
371 | 372 | ||
372 | tp = tcp_sk(sk); | 373 | tp = tcp_sk(sk); |
373 | seq = ntohl(th->seq); | 374 | seq = ntohl(th->seq); |
375 | /* XXX (TFO) - tp->snd_una should be ISN (tcp_create_openreq_child() */ | ||
376 | fastopen = tp->fastopen_rsk; | ||
377 | snd_una = fastopen ? tcp_rsk(fastopen)->snt_isn : tp->snd_una; | ||
374 | if (sk->sk_state != TCP_LISTEN && | 378 | if (sk->sk_state != TCP_LISTEN && |
375 | !between(seq, tp->snd_una, tp->snd_nxt)) { | 379 | !between(seq, snd_una, tp->snd_nxt)) { |
376 | NET_INC_STATS_BH(net, LINUX_MIB_OUTOFWINDOWICMPS); | 380 | NET_INC_STATS_BH(net, LINUX_MIB_OUTOFWINDOWICMPS); |
377 | goto out; | 381 | goto out; |
378 | } | 382 | } |
@@ -436,8 +440,13 @@ static void tcp_v6_err(struct sk_buff *skb, struct inet6_skb_parm *opt, | |||
436 | goto out; | 440 | goto out; |
437 | 441 | ||
438 | case TCP_SYN_SENT: | 442 | case TCP_SYN_SENT: |
439 | case TCP_SYN_RECV: /* Cannot happen. | 443 | case TCP_SYN_RECV: |
440 | It can, it SYNs are crossed. --ANK */ | 444 | /* Only in fast or simultaneous open. If a fast open socket is |
445 | * is already accepted it is treated as a connected one below. | ||
446 | */ | ||
447 | if (fastopen && fastopen->sk == NULL) | ||
448 | break; | ||
449 | |||
441 | if (!sock_owned_by_user(sk)) { | 450 | if (!sock_owned_by_user(sk)) { |
442 | sk->sk_err = err; | 451 | sk->sk_err = err; |
443 | sk->sk_error_report(sk); /* Wake people up to see the error (see connect in sock.c) */ | 452 | sk->sk_error_report(sk); /* Wake people up to see the error (see connect in sock.c) */ |
@@ -463,7 +472,8 @@ out: | |||
463 | static int tcp_v6_send_synack(struct sock *sk, struct dst_entry *dst, | 472 | static int tcp_v6_send_synack(struct sock *sk, struct dst_entry *dst, |
464 | struct flowi6 *fl6, | 473 | struct flowi6 *fl6, |
465 | struct request_sock *req, | 474 | struct request_sock *req, |
466 | u16 queue_mapping) | 475 | u16 queue_mapping, |
476 | struct tcp_fastopen_cookie *foc) | ||
467 | { | 477 | { |
468 | struct inet_request_sock *ireq = inet_rsk(req); | 478 | struct inet_request_sock *ireq = inet_rsk(req); |
469 | struct ipv6_pinfo *np = inet6_sk(sk); | 479 | struct ipv6_pinfo *np = inet6_sk(sk); |
@@ -474,7 +484,7 @@ static int tcp_v6_send_synack(struct sock *sk, struct dst_entry *dst, | |||
474 | if (!dst && (dst = inet6_csk_route_req(sk, fl6, req)) == NULL) | 484 | if (!dst && (dst = inet6_csk_route_req(sk, fl6, req)) == NULL) |
475 | goto done; | 485 | goto done; |
476 | 486 | ||
477 | skb = tcp_make_synack(sk, dst, req, NULL); | 487 | skb = tcp_make_synack(sk, dst, req, foc); |
478 | 488 | ||
479 | if (skb) { | 489 | if (skb) { |
480 | __tcp_v6_send_check(skb, &ireq->ir_v6_loc_addr, | 490 | __tcp_v6_send_check(skb, &ireq->ir_v6_loc_addr, |
@@ -498,7 +508,7 @@ static int tcp_v6_rtx_synack(struct sock *sk, struct request_sock *req) | |||
498 | struct flowi6 fl6; | 508 | struct flowi6 fl6; |
499 | int res; | 509 | int res; |
500 | 510 | ||
501 | res = tcp_v6_send_synack(sk, NULL, &fl6, req, 0); | 511 | res = tcp_v6_send_synack(sk, NULL, &fl6, req, 0, NULL); |
502 | if (!res) { | 512 | if (!res) { |
503 | TCP_INC_STATS_BH(sock_net(sk), TCP_MIB_RETRANSSEGS); | 513 | TCP_INC_STATS_BH(sock_net(sk), TCP_MIB_RETRANSSEGS); |
504 | NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_TCPSYNRETRANS); | 514 | NET_INC_STATS_BH(sock_net(sk), LINUX_MIB_TCPSYNRETRANS); |
@@ -917,7 +927,12 @@ static void tcp_v6_timewait_ack(struct sock *sk, struct sk_buff *skb) | |||
917 | static void tcp_v6_reqsk_send_ack(struct sock *sk, struct sk_buff *skb, | 927 | static void tcp_v6_reqsk_send_ack(struct sock *sk, struct sk_buff *skb, |
918 | struct request_sock *req) | 928 | struct request_sock *req) |
919 | { | 929 | { |
920 | tcp_v6_send_ack(skb, tcp_rsk(req)->snt_isn + 1, tcp_rsk(req)->rcv_isn + 1, | 930 | /* sk->sk_state == TCP_LISTEN -> for regular TCP_SYN_RECV |
931 | * sk->sk_state == TCP_SYN_RECV -> for Fast Open. | ||
932 | */ | ||
933 | tcp_v6_send_ack(skb, (sk->sk_state == TCP_LISTEN) ? | ||
934 | tcp_rsk(req)->snt_isn + 1 : tcp_sk(sk)->snd_nxt, | ||
935 | tcp_rsk(req)->rcv_nxt, | ||
921 | req->rcv_wnd, tcp_time_stamp, req->ts_recent, sk->sk_bound_dev_if, | 936 | req->rcv_wnd, tcp_time_stamp, req->ts_recent, sk->sk_bound_dev_if, |
922 | tcp_v6_md5_do_lookup(sk, &ipv6_hdr(skb)->daddr), | 937 | tcp_v6_md5_do_lookup(sk, &ipv6_hdr(skb)->daddr), |
923 | 0, 0); | 938 | 0, 0); |
@@ -969,8 +984,10 @@ static int tcp_v6_conn_request(struct sock *sk, struct sk_buff *skb) | |||
969 | struct tcp_sock *tp = tcp_sk(sk); | 984 | struct tcp_sock *tp = tcp_sk(sk); |
970 | __u32 isn = TCP_SKB_CB(skb)->when; | 985 | __u32 isn = TCP_SKB_CB(skb)->when; |
971 | struct dst_entry *dst = NULL; | 986 | struct dst_entry *dst = NULL; |
987 | struct tcp_fastopen_cookie foc = { .len = -1 }; | ||
988 | bool want_cookie = false, fastopen; | ||
972 | struct flowi6 fl6; | 989 | struct flowi6 fl6; |
973 | bool want_cookie = false; | 990 | int err; |
974 | 991 | ||
975 | if (skb->protocol == htons(ETH_P_IP)) | 992 | if (skb->protocol == htons(ETH_P_IP)) |
976 | return tcp_v4_conn_request(sk, skb); | 993 | return tcp_v4_conn_request(sk, skb); |
@@ -1001,7 +1018,7 @@ static int tcp_v6_conn_request(struct sock *sk, struct sk_buff *skb) | |||
1001 | tcp_clear_options(&tmp_opt); | 1018 | tcp_clear_options(&tmp_opt); |
1002 | tmp_opt.mss_clamp = IPV6_MIN_MTU - sizeof(struct tcphdr) - sizeof(struct ipv6hdr); | 1019 | tmp_opt.mss_clamp = IPV6_MIN_MTU - sizeof(struct tcphdr) - sizeof(struct ipv6hdr); |
1003 | tmp_opt.user_mss = tp->rx_opt.user_mss; | 1020 | tmp_opt.user_mss = tp->rx_opt.user_mss; |
1004 | tcp_parse_options(skb, &tmp_opt, 0, NULL); | 1021 | tcp_parse_options(skb, &tmp_opt, 0, want_cookie ? NULL : &foc); |
1005 | 1022 | ||
1006 | if (want_cookie && !tmp_opt.saw_tstamp) | 1023 | if (want_cookie && !tmp_opt.saw_tstamp) |
1007 | tcp_clear_options(&tmp_opt); | 1024 | tcp_clear_options(&tmp_opt); |
@@ -1074,19 +1091,27 @@ static int tcp_v6_conn_request(struct sock *sk, struct sk_buff *skb) | |||
1074 | isn = tcp_v6_init_sequence(skb); | 1091 | isn = tcp_v6_init_sequence(skb); |
1075 | } | 1092 | } |
1076 | have_isn: | 1093 | have_isn: |
1077 | tcp_rsk(req)->snt_isn = isn; | ||
1078 | 1094 | ||
1079 | if (security_inet_conn_request(sk, skb, req)) | 1095 | if (security_inet_conn_request(sk, skb, req)) |
1080 | goto drop_and_release; | 1096 | goto drop_and_release; |
1081 | 1097 | ||
1082 | if (tcp_v6_send_synack(sk, dst, &fl6, req, | 1098 | if (!dst && (dst = inet6_csk_route_req(sk, &fl6, req)) == NULL) |
1083 | skb_get_queue_mapping(skb)) || | ||
1084 | want_cookie) | ||
1085 | goto drop_and_free; | 1099 | goto drop_and_free; |
1086 | 1100 | ||
1101 | tcp_rsk(req)->snt_isn = isn; | ||
1087 | tcp_rsk(req)->snt_synack = tcp_time_stamp; | 1102 | tcp_rsk(req)->snt_synack = tcp_time_stamp; |
1088 | tcp_rsk(req)->listener = NULL; | 1103 | tcp_openreq_init_rwin(req, sk, dst); |
1089 | inet6_csk_reqsk_queue_hash_add(sk, req, TCP_TIMEOUT_INIT); | 1104 | fastopen = !want_cookie && |
1105 | tcp_try_fastopen(sk, skb, req, &foc, dst); | ||
1106 | err = tcp_v6_send_synack(sk, dst, &fl6, req, | ||
1107 | skb_get_queue_mapping(skb), &foc); | ||
1108 | if (!fastopen) { | ||
1109 | if (err || want_cookie) | ||
1110 | goto drop_and_free; | ||
1111 | |||
1112 | tcp_rsk(req)->listener = NULL; | ||
1113 | inet6_csk_reqsk_queue_hash_add(sk, req, TCP_TIMEOUT_INIT); | ||
1114 | } | ||
1090 | return 0; | 1115 | return 0; |
1091 | 1116 | ||
1092 | drop_and_release: | 1117 | drop_and_release: |
@@ -1760,6 +1785,7 @@ static void get_tcp6_sock(struct seq_file *seq, struct sock *sp, int i) | |||
1760 | const struct inet_sock *inet = inet_sk(sp); | 1785 | const struct inet_sock *inet = inet_sk(sp); |
1761 | const struct tcp_sock *tp = tcp_sk(sp); | 1786 | const struct tcp_sock *tp = tcp_sk(sp); |
1762 | const struct inet_connection_sock *icsk = inet_csk(sp); | 1787 | const struct inet_connection_sock *icsk = inet_csk(sp); |
1788 | struct fastopen_queue *fastopenq = icsk->icsk_accept_queue.fastopenq; | ||
1763 | 1789 | ||
1764 | dest = &sp->sk_v6_daddr; | 1790 | dest = &sp->sk_v6_daddr; |
1765 | src = &sp->sk_v6_rcv_saddr; | 1791 | src = &sp->sk_v6_rcv_saddr; |
@@ -1802,7 +1828,9 @@ static void get_tcp6_sock(struct seq_file *seq, struct sock *sp, int i) | |||
1802 | jiffies_to_clock_t(icsk->icsk_ack.ato), | 1828 | jiffies_to_clock_t(icsk->icsk_ack.ato), |
1803 | (icsk->icsk_ack.quick << 1) | icsk->icsk_ack.pingpong, | 1829 | (icsk->icsk_ack.quick << 1) | icsk->icsk_ack.pingpong, |
1804 | tp->snd_cwnd, | 1830 | tp->snd_cwnd, |
1805 | tcp_in_initial_slowstart(tp) ? -1 : tp->snd_ssthresh | 1831 | sp->sk_state == TCP_LISTEN ? |
1832 | (fastopenq ? fastopenq->max_qlen : 0) : | ||
1833 | (tcp_in_initial_slowstart(tp) ? -1 : tp->snd_ssthresh) | ||
1806 | ); | 1834 | ); |
1807 | } | 1835 | } |
1808 | 1836 | ||