In our discussion of TCP futures in Section 3. 7, we noted that to achieve a throughput of 10 Gbps, TCP could only tolerate a segment loss probability of 2 · 10-10(or equivalently, one loss event for every 5,000,000,000 segments). Show the derivation for the values of
2 ·10-10 1-out-of-5.000,000 for the RTT and MSS values given in Section 3.7. If TCP needed to support a 100 Gbps connection, what would the tolerable loss be?
