aboutsummaryrefslogtreecommitdiffstats
path: root/net
diff options
context:
space:
mode:
authorStephen Hemminger <shemminger@osdl.org>2005-11-10 19:56:12 -0500
committerDavid S. Miller <davem@davemloft.net>2005-11-10 19:56:12 -0500
commit2d2abbab63f6726a147ae61ada39bf2c9ee0db9a (patch)
tree3063abb67ee163a5b565b125796ba2f90582618b /net
parentf4805eded7d38c4e42bf473dc5eb2f34853beb06 (diff)
[TCP]: simplify microsecond rtt sampling
Simplify the code that comuputes microsecond rtt estimate used by TCP Vegas. Move the callback out of the RTT sampler and into the end of the ack cleanup. Signed-off-by: Stephen Hemminger <shemminger@osdl.org> Signed-off-by: David S. Miller <davem@davemloft.net>
Diffstat (limited to 'net')
-rw-r--r--net/ipv4/tcp_input.c62
1 files changed, 30 insertions, 32 deletions
diff --git a/net/ipv4/tcp_input.c b/net/ipv4/tcp_input.c
index 3e98b57578dc..e43065654930 100644
--- a/net/ipv4/tcp_input.c
+++ b/net/ipv4/tcp_input.c
@@ -548,10 +548,9 @@ static void tcp_event_data_recv(struct sock *sk, struct tcp_sock *tp, struct sk_
548 * To save cycles in the RFC 1323 implementation it was better to break 548 * To save cycles in the RFC 1323 implementation it was better to break
549 * it up into three procedures. -- erics 549 * it up into three procedures. -- erics
550 */ 550 */
551static void tcp_rtt_estimator(struct sock *sk, const __u32 mrtt, u32 *usrtt) 551static void tcp_rtt_estimator(struct sock *sk, const __u32 mrtt)
552{ 552{
553 struct tcp_sock *tp = tcp_sk(sk); 553 struct tcp_sock *tp = tcp_sk(sk);
554 const struct inet_connection_sock *icsk = inet_csk(sk);
555 long m = mrtt; /* RTT */ 554 long m = mrtt; /* RTT */
556 555
557 /* The following amusing code comes from Jacobson's 556 /* The following amusing code comes from Jacobson's
@@ -610,9 +609,6 @@ static void tcp_rtt_estimator(struct sock *sk, const __u32 mrtt, u32 *usrtt)
610 tp->mdev_max = tp->rttvar = max(tp->mdev, TCP_RTO_MIN); 609 tp->mdev_max = tp->rttvar = max(tp->mdev, TCP_RTO_MIN);
611 tp->rtt_seq = tp->snd_nxt; 610 tp->rtt_seq = tp->snd_nxt;
612 } 611 }
613
614 if (icsk->icsk_ca_ops->rtt_sample)
615 icsk->icsk_ca_ops->rtt_sample(sk, *usrtt);
616} 612}
617 613
618/* Calculate rto without backoff. This is the second half of Van Jacobson's 614/* Calculate rto without backoff. This is the second half of Van Jacobson's
@@ -1921,7 +1917,7 @@ tcp_fastretrans_alert(struct sock *sk, u32 prior_snd_una,
1921/* Read draft-ietf-tcplw-high-performance before mucking 1917/* Read draft-ietf-tcplw-high-performance before mucking
1922 * with this code. (Superceeds RFC1323) 1918 * with this code. (Superceeds RFC1323)
1923 */ 1919 */
1924static void tcp_ack_saw_tstamp(struct sock *sk, u32 *usrtt, int flag) 1920static void tcp_ack_saw_tstamp(struct sock *sk, int flag)
1925{ 1921{
1926 /* RTTM Rule: A TSecr value received in a segment is used to 1922 /* RTTM Rule: A TSecr value received in a segment is used to
1927 * update the averaged RTT measurement only if the segment 1923 * update the averaged RTT measurement only if the segment
@@ -1940,13 +1936,13 @@ static void tcp_ack_saw_tstamp(struct sock *sk, u32 *usrtt, int flag)
1940 */ 1936 */
1941 struct tcp_sock *tp = tcp_sk(sk); 1937 struct tcp_sock *tp = tcp_sk(sk);
1942 const __u32 seq_rtt = tcp_time_stamp - tp->rx_opt.rcv_tsecr; 1938 const __u32 seq_rtt = tcp_time_stamp - tp->rx_opt.rcv_tsecr;
1943 tcp_rtt_estimator(sk, seq_rtt, usrtt); 1939 tcp_rtt_estimator(sk, seq_rtt);
1944 tcp_set_rto(sk); 1940 tcp_set_rto(sk);
1945 inet_csk(sk)->icsk_backoff = 0; 1941 inet_csk(sk)->icsk_backoff = 0;
1946 tcp_bound_rto(sk); 1942 tcp_bound_rto(sk);
1947} 1943}
1948 1944
1949static void tcp_ack_no_tstamp(struct sock *sk, u32 seq_rtt, u32 *usrtt, int flag) 1945static void tcp_ack_no_tstamp(struct sock *sk, u32 seq_rtt, int flag)
1950{ 1946{
1951 /* We don't have a timestamp. Can only use 1947 /* We don't have a timestamp. Can only use
1952 * packets that are not retransmitted to determine 1948 * packets that are not retransmitted to determine
@@ -1960,21 +1956,21 @@ static void tcp_ack_no_tstamp(struct sock *sk, u32 seq_rtt, u32 *usrtt, int flag
1960 if (flag & FLAG_RETRANS_DATA_ACKED) 1956 if (flag & FLAG_RETRANS_DATA_ACKED)
1961 return; 1957 return;
1962 1958
1963 tcp_rtt_estimator(sk, seq_rtt, usrtt); 1959 tcp_rtt_estimator(sk, seq_rtt);
1964 tcp_set_rto(sk); 1960 tcp_set_rto(sk);
1965 inet_csk(sk)->icsk_backoff = 0; 1961 inet_csk(sk)->icsk_backoff = 0;
1966 tcp_bound_rto(sk); 1962 tcp_bound_rto(sk);
1967} 1963}
1968 1964
1969static inline void tcp_ack_update_rtt(struct sock *sk, const int flag, 1965static inline void tcp_ack_update_rtt(struct sock *sk, const int flag,
1970 const s32 seq_rtt, u32 *usrtt) 1966 const s32 seq_rtt)
1971{ 1967{
1972 const struct tcp_sock *tp = tcp_sk(sk); 1968 const struct tcp_sock *tp = tcp_sk(sk);
1973 /* Note that peer MAY send zero echo. In this case it is ignored. (rfc1323) */ 1969 /* Note that peer MAY send zero echo. In this case it is ignored. (rfc1323) */
1974 if (tp->rx_opt.saw_tstamp && tp->rx_opt.rcv_tsecr) 1970 if (tp->rx_opt.saw_tstamp && tp->rx_opt.rcv_tsecr)
1975 tcp_ack_saw_tstamp(sk, usrtt, flag); 1971 tcp_ack_saw_tstamp(sk, flag);
1976 else if (seq_rtt >= 0) 1972 else if (seq_rtt >= 0)
1977 tcp_ack_no_tstamp(sk, seq_rtt, usrtt, flag); 1973 tcp_ack_no_tstamp(sk, seq_rtt, flag);
1978} 1974}
1979 1975
1980static inline void tcp_cong_avoid(struct sock *sk, u32 ack, u32 rtt, 1976static inline void tcp_cong_avoid(struct sock *sk, u32 ack, u32 rtt,
@@ -2054,20 +2050,27 @@ static int tcp_tso_acked(struct sock *sk, struct sk_buff *skb,
2054 return acked; 2050 return acked;
2055} 2051}
2056 2052
2053static inline u32 tcp_usrtt(const struct sk_buff *skb)
2054{
2055 struct timeval tv, now;
2056
2057 do_gettimeofday(&now);
2058 skb_get_timestamp(skb, &tv);
2059 return (now.tv_sec - tv.tv_sec) * 1000000 + (now.tv_usec - tv.tv_usec);
2060}
2057 2061
2058/* Remove acknowledged frames from the retransmission queue. */ 2062/* Remove acknowledged frames from the retransmission queue. */
2059static int tcp_clean_rtx_queue(struct sock *sk, __s32 *seq_rtt_p, s32 *seq_usrtt) 2063static int tcp_clean_rtx_queue(struct sock *sk, __s32 *seq_rtt_p)
2060{ 2064{
2061 struct tcp_sock *tp = tcp_sk(sk); 2065 struct tcp_sock *tp = tcp_sk(sk);
2066 const struct inet_connection_sock *icsk = inet_csk(sk);
2062 struct sk_buff *skb; 2067 struct sk_buff *skb;
2063 __u32 now = tcp_time_stamp; 2068 __u32 now = tcp_time_stamp;
2064 int acked = 0; 2069 int acked = 0;
2065 __s32 seq_rtt = -1; 2070 __s32 seq_rtt = -1;
2066 struct timeval usnow;
2067 u32 pkts_acked = 0; 2071 u32 pkts_acked = 0;
2068 2072 void (*rtt_sample)(struct sock *sk, u32 usrtt)
2069 if (seq_usrtt) 2073 = icsk->icsk_ca_ops->rtt_sample;
2070 do_gettimeofday(&usnow);
2071 2074
2072 while ((skb = skb_peek(&sk->sk_write_queue)) && 2075 while ((skb = skb_peek(&sk->sk_write_queue)) &&
2073 skb != sk->sk_send_head) { 2076 skb != sk->sk_send_head) {
@@ -2107,16 +2110,11 @@ static int tcp_clean_rtx_queue(struct sock *sk, __s32 *seq_rtt_p, s32 *seq_usrtt
2107 tp->retrans_out -= tcp_skb_pcount(skb); 2110 tp->retrans_out -= tcp_skb_pcount(skb);
2108 acked |= FLAG_RETRANS_DATA_ACKED; 2111 acked |= FLAG_RETRANS_DATA_ACKED;
2109 seq_rtt = -1; 2112 seq_rtt = -1;
2110 } else if (seq_rtt < 0) 2113 } else if (seq_rtt < 0) {
2111 seq_rtt = now - scb->when; 2114 seq_rtt = now - scb->when;
2112 if (seq_usrtt) { 2115 if (rtt_sample)
2113 struct timeval tv; 2116 (*rtt_sample)(sk, tcp_usrtt(skb));
2114
2115 skb_get_timestamp(skb, &tv);
2116 *seq_usrtt = (usnow.tv_sec - tv.tv_sec) * 1000000
2117 + (usnow.tv_usec - tv.tv_usec);
2118 } 2117 }
2119
2120 if (sacked & TCPCB_SACKED_ACKED) 2118 if (sacked & TCPCB_SACKED_ACKED)
2121 tp->sacked_out -= tcp_skb_pcount(skb); 2119 tp->sacked_out -= tcp_skb_pcount(skb);
2122 if (sacked & TCPCB_LOST) 2120 if (sacked & TCPCB_LOST)
@@ -2126,8 +2124,11 @@ static int tcp_clean_rtx_queue(struct sock *sk, __s32 *seq_rtt_p, s32 *seq_usrtt
2126 !before(scb->end_seq, tp->snd_up)) 2124 !before(scb->end_seq, tp->snd_up))
2127 tp->urg_mode = 0; 2125 tp->urg_mode = 0;
2128 } 2126 }
2129 } else if (seq_rtt < 0) 2127 } else if (seq_rtt < 0) {
2130 seq_rtt = now - scb->when; 2128 seq_rtt = now - scb->when;
2129 if (rtt_sample)
2130 (*rtt_sample)(sk, tcp_usrtt(skb));
2131 }
2131 tcp_dec_pcount_approx(&tp->fackets_out, skb); 2132 tcp_dec_pcount_approx(&tp->fackets_out, skb);
2132 tcp_packets_out_dec(tp, skb); 2133 tcp_packets_out_dec(tp, skb);
2133 __skb_unlink(skb, &sk->sk_write_queue); 2134 __skb_unlink(skb, &sk->sk_write_queue);
@@ -2135,8 +2136,7 @@ static int tcp_clean_rtx_queue(struct sock *sk, __s32 *seq_rtt_p, s32 *seq_usrtt
2135 } 2136 }
2136 2137
2137 if (acked&FLAG_ACKED) { 2138 if (acked&FLAG_ACKED) {
2138 const struct inet_connection_sock *icsk = inet_csk(sk); 2139 tcp_ack_update_rtt(sk, acked, seq_rtt);
2139 tcp_ack_update_rtt(sk, acked, seq_rtt, seq_usrtt);
2140 tcp_ack_packets_out(sk, tp); 2140 tcp_ack_packets_out(sk, tp);
2141 2141
2142 if (icsk->icsk_ca_ops->pkts_acked) 2142 if (icsk->icsk_ca_ops->pkts_acked)
@@ -2299,7 +2299,6 @@ static int tcp_ack(struct sock *sk, struct sk_buff *skb, int flag)
2299 u32 ack = TCP_SKB_CB(skb)->ack_seq; 2299 u32 ack = TCP_SKB_CB(skb)->ack_seq;
2300 u32 prior_in_flight; 2300 u32 prior_in_flight;
2301 s32 seq_rtt; 2301 s32 seq_rtt;
2302 s32 seq_usrtt = 0;
2303 int prior_packets; 2302 int prior_packets;
2304 2303
2305 /* If the ack is newer than sent or older than previous acks 2304 /* If the ack is newer than sent or older than previous acks
@@ -2352,8 +2351,7 @@ static int tcp_ack(struct sock *sk, struct sk_buff *skb, int flag)
2352 prior_in_flight = tcp_packets_in_flight(tp); 2351 prior_in_flight = tcp_packets_in_flight(tp);
2353 2352
2354 /* See if we can take anything off of the retransmit queue. */ 2353 /* See if we can take anything off of the retransmit queue. */
2355 flag |= tcp_clean_rtx_queue(sk, &seq_rtt, 2354 flag |= tcp_clean_rtx_queue(sk, &seq_rtt);
2356 icsk->icsk_ca_ops->rtt_sample ? &seq_usrtt : NULL);
2357 2355
2358 if (tp->frto_counter) 2356 if (tp->frto_counter)
2359 tcp_process_frto(sk, prior_snd_una); 2357 tcp_process_frto(sk, prior_snd_una);
@@ -4242,7 +4240,7 @@ int tcp_rcv_state_process(struct sock *sk, struct sk_buff *skb,
4242 */ 4240 */
4243 if (tp->rx_opt.saw_tstamp && tp->rx_opt.rcv_tsecr && 4241 if (tp->rx_opt.saw_tstamp && tp->rx_opt.rcv_tsecr &&
4244 !tp->srtt) 4242 !tp->srtt)
4245 tcp_ack_saw_tstamp(sk, NULL, 0); 4243 tcp_ack_saw_tstamp(sk, 0);
4246 4244
4247 if (tp->rx_opt.tstamp_ok) 4245 if (tp->rx_opt.tstamp_ok)
4248 tp->advmss -= TCPOLEN_TSTAMP_ALIGNED; 4246 tp->advmss -= TCPOLEN_TSTAMP_ALIGNED;