[e2e] Are we doing sliding window in the Internet?
Fred Baker
fred at cisco.com
Sun Dec 31 16:29:00 PST 2006
yes and no.
A large percentage of sessions are very short - count the bytes in
this email and consider how many TCP segments are required to carry
it, for example, or look through your web cache to see the sizes of
objects it stores. We are doing the sliding window algorithm, but it
cuts very short when the TCP session abruptly closes.
For longer exchanges - p2p and many others - yes, we indeed do
sliding window.
I don't see any reason to believe that TCPs tune themselves to have
exactly RTT/MSS segments outstanding. That would be the optimal
number to have ourstanding, but generally they will have the smallest
of { the offered window, the sender's maximum window, and the used
window at which they start dropping traffic }. If they never see
loss, they can keep an incredibly large amount of data outstanding
regardless of the values of RTT and MSS.
I wonder where you got the notion that a typical session had a 10 ms
RTT. In a LAN environment where the servers are in the same building,
that is probably the case. But consider these rather more typical
examples: across my VPN to a machine at work, across the US to MIT,
and across the Atlantic to you:
[stealth-10-32-244-218:~] fred% traceroute irp-view7
traceroute to irp-view7.cisco.com (171.70.65.144), 64 hops max, 40
byte packets
1 fred-vpn (10.32.244.217) 1.486 ms 1.047 ms 1.034 ms
2 n003-000-000-000.static.ge.com (3.7.12.1) 22.360 ms 20.962 ms
22.194 ms
3 10.34.251.137 (10.34.251.137) 23.559 ms 22.586 ms 22.236 ms
4 sjc20-a5-gw2 (10.34.250.78) 21.465 ms 22.544 ms 20.748 ms
5 sjc20-sbb5-gw1 (128.107.180.105) 22.294 ms 22.351 ms 22.803 ms
6 sjc20-rbb-gw5 (128.107.180.22) 21.583 ms 22.517 ms 24.190 ms
7 sjc12-rbb-gw4 (128.107.180.2) 22.115 ms 23.143 ms 21.478 ms
8 sjc5-sbb4-gw1 (171.71.241.253) 26.550 ms 23.122 ms 21.569 ms
9 sjc12-dc5-gw2 (171.71.241.66) 22.115 ms 22.435 ms 22.185 ms
10 sjc5-dc3-gw2 (171.71.243.46) 22.031 ms 21.846 ms 22.185 ms
11 irp-view7 (171.70.65.144) 22.760 ms 22.912 ms 21.941 ms
[stealth-10-32-244-218:~] fred% traceroute www.mit.edu
traceroute to www.mit.edu (18.7.22.83), 64 hops max, 40 byte packets
1 fred-vpn (10.32.244.217) 1.468 ms 1.108 ms 1.083 ms
2 172.16.16.1 (172.16.16.1) 11.994 ms 10.351 ms 10.858 ms
3 cbshost-68-111-47-251.sbcox.net (68.111.47.251) 9.238 ms 19.517
ms 9.857 ms
4 12.125.98.101 (12.125.98.101) 11.849 ms 11.913 ms 12.086 ms
5 gbr1-p100.la2ca.ip.att.net (12.123.28.130) 12.348 ms 11.736 ms
12.891 ms
6 tbr2-p013502.la2ca.ip.att.net (12.122.11.145) 15.071 ms 13.462
ms 13.453 ms
7 12.127.3.221 (12.127.3.221) 12.643 ms 13.761 ms 14.345 ms
8 br1-a3110s9.attga.ip.att.net (192.205.33.230) 13.842 ms 12.414
ms 12.647 ms
9 ae-32-54.ebr2.losangeles1.level3.net (4.68.102.126) 16.651 ms
ae-32-56.ebr2.losangeles1.level3.net (4.68.102.190) 20.154 ms *
10 * * *
11 ae-2.ebr1.sanjose1.level3.net (4.69.132.9) 28.222 ms 24.319 ms
ae-1-100.ebr2.sanjose1.level3.net (4.69.132.2) 35.417 ms
12 ae-1-100.ebr2.sanjose1.level3.net (4.69.132.2) 25.640 ms 22.567
ms *
13 ae-3.ebr1.denver1.level3.net (4.69.132.58) 52.275 ms 60.821 ms
54.384 ms
14 ae-3.ebr1.chicago1.level3.net (4.69.132.62) 68.285 ms
ae-1-100.ebr2.denver1.level3.net (4.69.132.38) 59.113 ms 68.779 ms
15 * * *
16 * ae-7-7.car1.boston1.level3.net (4.69.132.241) 94.977 ms *
17 ae-7-7.car1.boston1.level3.net (4.69.132.241) 95.821 ms
ae-11-11.car2.boston1.level3.net (4.69.132.246) 93.856 ms
ae-7-7.car1.boston1.level3.net (4.69.132.241) 96.735 ms
18 ae-11-11.car2.boston1.level3.net (4.69.132.246) 91.093 ms
92.125 ms 4.79.2.2 (4.79.2.2) 95.802 ms
19 4.79.2.2 (4.79.2.2) 93.945 ms 95.336 ms 97.301 ms
20 w92-rtr-1-backbone.mit.edu (18.168.0.25) 98.246 ms www.mit.edu
(18.7.22.83) 93.657 ms w92-rtr-1-backbone.mit.edu (18.168.0.25)
92.610 ms
[stealth-10-32-244-218:~] fred% traceroute web.de
traceroute to web.de (217.72.195.42), 64 hops max, 40 byte packets
1 fred-vpn (10.32.244.217) 1.482 ms 1.078 ms 1.093 ms
2 172.16.16.1 (172.16.16.1) 12.131 ms 9.318 ms 8.140 ms
3 cbshost-68-111-47-251.sbcox.net (68.111.47.251) 10.790 ms 9.051
ms 10.564 ms
4 12.125.98.101 (12.125.98.101) 13.580 ms 21.643 ms 12.206 ms
5 gbr2-p100.la2ca.ip.att.net (12.123.28.134) 12.446 ms 12.914 ms
12.006 ms
6 tbr2-p013602.la2ca.ip.att.net (12.122.11.149) 13.463 ms 12.711
ms 12.187 ms
7 12.127.3.213 (12.127.3.213) 185.324 ms 11.845 ms 12.189 ms
8 192.205.33.226 (192.205.33.226) 12.008 ms 11.665 ms 25.390 ms
9 ae-1-53.bbr1.losangeles1.level3.net (4.68.102.65) 13.695 ms
ae-1-51.bbr1.losangeles1.level3.net (4.68.102.1) 11.645 ms
ae-1-53.bbr1.losangeles1.level3.net (4.68.102.65) 12.517 ms
10 ae-1-0.bbr1.frankfurt1.level3.net (212.187.128.30) 171.886 ms
as-2-0.bbr2.frankfurt1.level3.net (4.68.128.169) 167.640 ms 168.895 ms
11 ge-10-0.ipcolo1.frankfurt1.level3.net (4.68.118.9) 170.336 ms
ge-11-1.ipcolo1.frankfurt1.level3.net (4.68.118.105) 174.211 ms
ge-10-1.ipcolo1.frankfurt1.level3.net (4.68.118.73) 169.730 ms
12 gw-megaspace.frankfurt.eu.level3.net (212.162.44.158) 169.276
ms 170.110 ms 168.099 ms
13 te-2-3.gw-backbone-d.bs.ka.schlund.net (212.227.120.17) 171.412
ms 171.820 ms 170.265 ms
14 a0kac2.gw-distwe-a.bs.ka.schlund.net (212.227.121.218) 175.416
ms 173.653 ms 174.007 ms
15 ha-42.web.de (217.72.195.42) 174.908 ms 174.921 ms 175.821 ms
On Dec 31, 2006, at 11:15 AM, Detlef Bosau wrote:
> Happy New Year, Miss Sophy My Dear!
>
> (Although this sketch is in Englisch, it is hardly known outside
> Germay to my knowledge.)
>
> I wonder whether we´re really doing sliding window in TCP
> connections all the time or whether a number of connections have
> congestion windows of only one segment, i.e. behave like stop´n
> wait in reality.
>
> When I assume an Ethernet like MTU, i.e. 1500 byte = 12000 bit,
> and 10 ms RTT the throughput is roughly 12000 bit / 10 ms = 1.2 Mbps.
>
> From this I would expect that in quite a few cases a TCP connection
> will have a congestion window of 1 MSS or even less.
>
> In addition, some weeks ago I read a paper, I don´t remember were,
> that we should reconsider and perhaps resize our MTUs to larger
> values for networks with large bandwidth. The rationale was simply
> as follows: The MTU size is always a tradeoff between overhead and
> jitter. From Ethernet we know that we can accept a maximum packet
> duration of 12000 bit / (10 Mbps) = 1.2 ms and the resultig
> jitter. For Gigabit Ethernet
> a maximum packet duration of 1.2 ms would result in a MTU size of
> 1500 kbyte = 1.5 Mbyte.
>
> If so, we would see "stop´n wait like" connections much more
> frequently than today.
>
> Is this view correct?
>
More information about the end2end-interest
mailing list