Begin forwarded message:

> From: Srinivasan Keshav <kes...@uwaterloo.ca>
> Subject: [e2e] Why do we need congestion control?
> Date: March 5, 2013 15:04:48 GMT+01:00
> To: "<end2end-inter...@postel.org>" <end2end-inter...@postel.org>
> 
> To answer this question, I put together some slides for a presentation at the 
> IRTF ICCRG Workshop in 2007 [1]. In a nutshell, to save costs, we always size 
> a shared resource (such as a link or a router) smaller than the sum of peak 
> demands. This can result in transient or persistent overloads, reducing 
> user-perceived performance. Transient overloads are easily relieved by a 
> buffer, but persistent overload requires reductions of source loads, which is 
> the role of congestion control. Lacking congestion control, or worse, with an 
> inappropriate response to a performance problem (such as by increasing the 
> load), shared network resources are always overloaded leading to delays, 
> losses, and eventually collapse, where every packet that is sent is a 
> retransmission and no source makes progress. A more detailed description can 
> also be found in chapter 1 of my PhD thesis [2].
> 
> Incidentally, the distributed optimization approach that Jon mentioned is 
> described beautifully in [3]. 
> 
> hope this helps, 
> keshav
> 
> [1] Congestion and Congestion Control, Presentation at IRTF ICCRG Workshop, 
> PFLDnet, 2007, Los Angeles (California), USA, February 2007. 
> http://blizzard.cs.uwaterloo.ca/keshav/home/Papers/data/07/congestion.pdf
> 
> [2] S. Keshav, Congestion Control in Computer Networks PhD Thesis, published 
> as UC Berkeley TR-654, September 1991
> http://blizzard.cs.uwaterloo.ca/keshav/home/Papers/data/91/thesis/ch1.pdf
> 
> [3] Palomar, Daniel P., and Mung Chiang. "A tutorial on decomposition methods 
> for network utility maximization." Selected Areas in Communications, IEEE 
> Journal on 24.8 (2006): 1439-1451.
> http://www.princeton.edu/~chiangm/decomptutorial.pdf
> 
> 

Reply via email to