Suppose that TCP is measuring RTTs of 1.0 second, with a mean deviation of 0.1 second. Suddenly
Question:
Suppose that TCP is measuring RTTs of 1.0 second, with a mean deviation of 0.1 second. Suddenly the RTT jumps to 5.0 seconds, with no deviation. Compare the behaviors of the original and Jacobson/Karels algorithms for computing TimeOut. Specifically, how many timeouts are encountered with each algorithm? What is the largest TimeOut calculated? Use δ = 1/8.
Fantastic news! We've Found the answer you've been seeking!
Step by Step Answer:
Related Book For
Computer Networks A Systems Approach
ISBN: 9780128182000
6th Edition
Authors: Larry L. Peterson, Bruce S. Davie
Question Posted: