Hello Rodney, yes I agree. Elasticity may mean different things.
BTW, I hope I made the point about incentives to cheat, and the risks for unresponsive traffic for L4S when using ECT(1) as a trusted input. My comments are mostly related to my experience with Cisco WebEx as a Cisco employee working on that. I do care about these applications! But I do care of all other on-line meetings apps. I'm sure we all do lately. These apps may have incentives to cheat. Other apps too. I know for sure Cisco WebEx does not cheat. Incentives to cheat will force everyone to cheat. I believe we do not want that. So I was talking about real-time video (media in general, including content sharing), transported using RTP/UDP. Not DAS video transported using HTTP/TCP. Typically the latter do not use FEC while the former do have FEC streams over defined RTP payload types. Luca On Wed, Apr 29, 2020 at 10:44 AM Rodney W. Grimes <i...@gndrsh.dnsmgr.net> wrote: > Hello Luca, tsvwg'ers, > > I believe that there is some confusion around about how video > conference streams, and video *streams* in general differ from other > forms of traffic. I believe some of that confusion comes about not > only becasue of the FEC nature that many use but also over the terms > "elastic", "greedy" and "capacity seeking." > > Though video streams *do* adapt to network conditions, they > do so at fixed consumption steps, this is the elastic nature of a > video stream. They do not continually seek to find full bandwidth, > that is a greedy or capacity seeking flow which video streams are *not*. > > There is a differenece between watching a video vs downloading > a video on the internet. > > The above are *rough* statements, as the details are much more > involved with things like traffic burst of next frame chunk, and other > techniques that have come onto the market. I would love to hear from > an expert on the current true nature, but there was certainly some > mis-statements about video conference streams during the meeting. > > Regards, > Rod > > > Hi Jake, > > > > Thanks for the notes. Very useful. > > The other issue with the meeting was that the virtual mic queue control > > channel was the WebEx Meeting chat that does not exist in WebEx Teams. > So, > > I had to switch to Meetings and lost some pieces of the discussion. > > > > Yes there might be a terminology difference. Elastic traffic is usually > > used in the sense of bandwidth sharing not just to define variable bit > > rates. > > > > The point is that there are incentives to cheat in L4S. > > > > There is a priority queue that my application can enter by providing as > > input ECT(1). > > Applications such as on-line meetings will have a relatively low and > highly > > paced rate. > > > > This traffic is conformant to dualQ L queue but is unresponsive to > > congestion notifications. > > > > This is especially true for FEC streams which could be used to ameliorate > > the media quality in presence of losses(e.g. Wi-Fi) > > or increased jitter. > > > > > > That was one more point on why using ECT(1) as input assumes trust or a > > black list after being caught. > > > > In both cases the ECT(1) as input is DoSable. > > > > > > > > On Tue, Apr 28, 2020 at 7:12 PM Holland, Jake <jholl...@akamai.com> > wrote: > > > > > Hi Luca, > > > > > > > > > > > > To your point about the discussion being difficult to follow: I tried > to > > > capture the intent of everyone who commented while taking notes: > > > > > > https://etherpad.ietf.org:9009/p/notes-ietf-interim-2020-tsvwg-03 > > > > > > > > > > > > I think this was intended to take the place of a need for everyone to > > > re-send the same points to the list, but of course some of the most > crucial > > > points could probably use fleshing out with on-list follow up. > > > > > > > > > > > > It got a bit rough in places because I was disconnected a few times and > > > had to cut over to a local text file, and I may have failed to > correctly > > > understand or summarize some of the comments, so there?s chances I > might > > > have missed something, but I did my best to capture them all. > > > > > > > > > > > > I encourage people to review comments and check whether they came out > more > > > or less correct, and to offer formatting and cleanup suggestions if > there?s > > > a good way to make it easier to follow. > > > > > > > > > > > > I had timestamps at the beginning of each main point of discussion, > with > > > the intent that after the video is published it would be easier to go > back > > > and check precisely what was said. It looks like someone has been > making > > > cleanup edits that removed the first half of those so far, but my local > > > text file still has most of those and I can go back and re-insert them > if > > > it seems useful. > > > > > > > > > > > > @Luca: during your comments in particular I think there might have > been a > > > disruption--I had a ?first comment missed, please check video? > placeholder > > > and I may have misunderstood the part about video elasticity, but my > > > interpretation at the time was that Stuart was claiming that video was > > > elastic in that it would adjust downward to avoid overflowing a loaded > > > link, and I thought you were claiming that it was not elastic in that > it > > > would not exceed a maximum rate, which I summarized as perhaps a > semantic > > > disagreement, but if you?d like to help clean that up, it might be > useful. > > > > > > > > > > > > From this message, it sounds like the key point you were making was > that > > > it also will not go below a certain rate, and perhaps that quality can > stay > > > relatively good in spite of high network loss? > > > > > > > > > > > > Best regards, > > > > > > Jake > > > > > > > > > > > > *From: *Luca Muscariello <muscarie...@ieee.org> > > > *Date: *Tuesday, April 28, 2020 at 1:54 AM > > > *To: *Dave Taht <dave.t...@gmail.com> > > > *Cc: *tsvwg IETF list <ts...@ietf.org>, bloat < > bloat@lists.bufferbloat.net > > > > > > > *Subject: *Re: [Bloat] my backlogged comments on the ECT(1) interim > call > > > > > > > > > > > > Hi Dave and list members, > > > > > > > > > > > > It was difficult to follow the discussion at the meeting yesterday. > > > > > > Who said what in the first place. > > > > > > > > > > > > There have been a lot of non-technical comments such as: this solution > > > > > > is better than another in my opinion. "better" has often been used > > > > > > as when evaluating the taste of an ice cream: White chocolate vs black > > > chocolate. > > > > > > This has taken a significant amount of time at the meeting. I haven't > > > learned > > > > > > much from that kind of discussion and I do not think that helped to > make > > > > > > much progress. > > > > > > > > > > > > If people can re-make their points in the list it would help the > debate. > > > > > > > > > > > > Another point that a few raised is that we have to make a decision as > fast > > > as possible. > > > > > > I dismissed entirely that argument. Trading off latency with > resilience of > > > the Internet > > > > > > is entirely against the design principle of the Internet architecture > > > itself. > > > > > > Risk analysis is something that we should keep in mind even when > > > deploying any experiment > > > > > > and should be a substantial part of it. > > > > > > > > > > > > Someone claimed that on-line meeting traffic is elastic. This is not > true, > > > I tried to > > > > > > clarify this. These applications (WebEx/Zoom) are low rate, a typical > > > maximum upstream > > > > > > rate is 2Mbps and is not elastic. These applications have often a > > > stand-alone app > > > > > > that is not using the browser WebRTC stack (the standalone app > typically > > > works better). > > > > > > > > > > > > A client sends upstream one or two video qualities unless the video > camera > > > is switched off. > > > > > > In presence of losses, FEC is used but it is still non elastic. > > > > > > Someone claimed (at yesterday's meeting) that fairness is not an issue > > > (who cares, I heard!) > > > > > > Well, fairness can constitute a differentiation advantage between two > > > companies that are > > > > > > commercializing on-line meetings products. Unless at the IETF we accept > > > > > > "law-of-the-jungle" behaviours from Internet applications developers, > we > > > should be careful > > > > > > about making such claims. > > > > > > Any opportunity to cheat, that brings a business advantage WILL be > used. > > > > > > > > > > > > /Luca > > > > > > > > > > > > TL;DR > > > > > > To Dave: you asked several times what Cisco does on latency reduction > in > > > > > > network equipment. I tend to be very shy when replying on these > questions > > > > > > as this is not vendor neutral. If chairs think this is not appropriate > for > > > > > > the list, please say it and I'll reply privately only. > > > > > > > > > > > > What I write below can be found in Cisco products data sheets and is > not > > > > > > trade secret. There are very good blog posts explaining details. > > > > > > Not surprisingly Cisco implements the state of the art on the topic > > > > > > and it is totally feasible to do-the-right-thing in software and > hardware.. > > > > > > > > > > > > Cisco implements AFD (one queue + a flow table) accompanied by a > priority > > > queue for > > > > > > flows that have a certain profile in rate and size. The concept is well > > > known and well > > > > > > studied in the literature. AFD is safe and can well serve a complex > > > traffic mix when > > > > > > accompanied by a priority queue. This prio-queue should not be confused > > > with a strict > > > > > > priority queue (e.g. EF in diffserv). There are subtleties related to > the > > > DOCSIS > > > > > > shared medium which would be too long to describe here. > > > > > > > > > > > > This is available in Cisco CMTS for the DOCSIS segment. Bottleneck > traffic > > > > > > does not negatively impact non-bottlenecked-traffic such as an on-line > > > meeting like > > > > > > the WebEx call we had yesterday. It is safe from a network neutrality > > > point-of-view > > > > > > and no applications get hurt. > > > > > > > > > > > > Cisco implements AFD+prio also for some DC switches such as the Nexus > 9k. > > > There > > > > > > is a blog post written by Tom Edsal online that explains pretty well > how > > > that works. > > > > > > This includes mechanisms such as p-fabric to approximate SRPT (shortest > > > remaining processing time) > > > > > > and minimize flow completion time for many DC workloads. The mix of > the two > > > > > > brings FCT minimization AND latency minimization. This is silicon and > > > scales at any speed. > > > > > > For those who are not familiar with these concepts, please search the > > > research work of Balaji > > > > > > Prabhakar and Ron Pang at Stanford. > > > > > > > > > > > > Wi-Fi: Cisco does airtime fairness in Aironet but I think in the Meraki > > > series too. > > > > > > The concept is similar to what described above but there are several > > > queues, one per STA. > > > > > > Packets are enqueued in the access (category) queue at dequeue time > from > > > the air-time > > > > > > packet scheduler. > > > > > > > > > > > > On Mon, Apr 27, 2020 at 9:24 PM Dave Taht <dave.t...@gmail.com> wrote: > > > > > > It looks like the majority of what I say below is not related to the > > > fate of the "bit". The push to take the bit was > > > strong with this one, and me... can't we deploy more of what we > > > already got in places where it matters? > > > > > > ... > > > > > > so: A) PLEA: From 10 years now, of me working on bufferbloat, working > > > on real end-user and wifi traffic and real networks.... > > > > > > I would like folk here to stop benchmarking two flows that run for a > long > > > time > > > and in one direction only... and thus exclusively in tcp congestion > > > avoidance mode. > > > > > > Please. just. stop. Real traffic looks nothing like that. The internet > > > looks nothing like that. > > > The netops folk I know just roll their eyes up at benchmarks like this > > > that prove nothing and tell me to go to ripe meetings instead. > > > When y'all talk about "not looking foolish for not mandating ecn now", > > > you've already lost that audience with benchmarks like these. > > > > > > Sure, setup a background flow(s) like that, but then hit the result > > > with a mix of > > > far more normal traffic? Please? networks are never used > unidirectionally > > > and both directions congesting is frequent. To illustrate that > problem... > > > > > > I have a really robust benchmark that we have used throughout the > > > bufferbloat > > > project that I would like everyone to run in their environments, the > flent > > > "rrul" test. Everybody on both sides has big enough testbeds setup > that a > > > few > > > hours spent on doing that - and please add in asymmetric networks > > > especially - > > > and perusing the results ought to be enlightening to everyone as to the > > > kind > > > of problems real people have, on real networks. > > > > > > Can the L4S and SCE folk run the rrul test some day soon? Please? > > > > > > I rather liked this benchmark that tested another traffic mix, > > > > > > ( > > > > https://www.cablelabs.com/wp-content/uploads/2014/06/DOCSIS-AQM_May2014.pdf > > > < > https://urldefense.proofpoint.com/v2/url?u=https-3A__www.cablelabs.com_wp-2Dcontent_uploads_2014_06_DOCSIS-2DAQM-5FMay2014.pdf&d=DwMFaQ&c=96ZbZZcaMF4w0F4jpN6LZg&r=bqnFROivDo_4iF8Z3R4DyNWKbbMeXr0LOgLnElT1Ook&m=j5nEJ3W8fRmqjnBSWapTVKj6dNbpegl4kSeynebCQT4&s=DrB4ENWjWbVu9SqtIh7lXKJj96fwm6TqESC6E8_IdnY&e= > > > > > ) > > > > > > although it had many flaws (like not doing dns lookups), I wish it > > > could be dusted off and used to compare this > > > new fangled ecn enabled stuff with the kind of results you can merely > get > > > with packet loss and rtt awareness. It would be so great to be able > > > to directly compare all these new algorithms against this benchmark. > > > > > > Adding in a non ecn'd udp based routing protocol on heavily > > > oversubscribed 100mbit link is also enlightening. > > > > > > I'd rather like to see that benchmark improved for a more modernized > > > home traffic mix > > > where it is projected there may be 30 devices on the network on > average, > > > in a few years. > > > > > > If there is any one thing y'all can do to reduce my blood pressure and > > > keep me engaged here whilst you > > > debate the end of the internet as I understand it, it would be to run > > > the rrul test as part of all your benchmarks. > > > > > > thank you. > > > > > > B) Stuart Cheshire regaled us with several anecdotes - one concerning > > > his problems > > > with comcast's 1Gbit/35mbit service being unusable, under load, for > > > videoconferencing. This is true. The overbuffering at the CMTSes > > > still, has to be seen to be believed, at all rates. At lower rates > > > it's possible to shape this, with another device (which is what > > > the entire SQM deployment does in self defense and why cake has a > > > specific docsis ingress mode), but it is cpu intensive > > > and requires x86 hardware to do well at rates above 500Mbits, > presently. > > > > > > So I wish CMTS makers (Arris and Cisco) were in this room. are they? > > > > > > (Stuart, if you'd like a box that can make your comcast link > pleasurable > > > under all workloads, whenever you get back to los gatos, I've got a few > > > lying around. Was so happy to get a few ietfers this past week to apply > > > what's off the shelf for end users today. :) > > > > > > C) I am glad bob said the L4S is finally looking at asymmetric > > > networks, and starting to tackle ack-filtering and accecn issues > > > there. > > > > > > But... I would have *started there*. Asymmetric access is the > predominate > > > form > > > of all edge technologies. > > > > > > I would love to see flent rrul test results for 1gig/35mbit, 100/10, > 200/10 > > > services, in particular. (from SCE also!). "lifeline" service (11/2) > > > would be good > > > to have results on. It would be especially good to have baseline > > > comparison data from the measured, current deployment > > > of the CMTSes at these rates, to start with, with no queue management > in > > > play, then pie on the uplink, then fq_codel on the uplink, and then > > > this ecn stuff, and so on. > > > > > > D) The two CPE makers in the room have dismissed both fq and sce as > > > being too difficult to implement. They did say that dualpi was > > > actually implemented in software, not hardware. > > > > > > I would certainly like them to benchmark what they plan to offer in L4S > > > vs what is already available in the edgerouter X, as one low end > > > example among thousands. > > > > > > I also have to note, at higher speeds, all the buffering moves into > > > the wifi and the results are currently ugly. I imagine > > > they are exploring how to fix their wifi stacks also? I wish more folk > > > were using RVR + latency benchmarks like this one: > > > > > > > > > > http://flent-newark.bufferbloat.net/~d/Airtime%20based%20queue%20limit%20for%20FQ_CoDel%20in%20wireless%20interface.pdf > > > < > https://urldefense.proofpoint.com/v2/url?u=http-3A__flent-2Dnewark.bufferbloat.net_-7Ed_Airtime-2520based-2520queue-2520limit-2520for-2520FQ-5FCoDel-2520in-2520wireless-2520interface.pdf&d=DwMFaQ&c=96ZbZZcaMF4w0F4jpN6LZg&r=bqnFROivDo_4iF8Z3R4DyNWKbbMeXr0LOgLnElT1Ook&m=j5nEJ3W8fRmqjnBSWapTVKj6dNbpegl4kSeynebCQT4&s=UEzrGb3xL5zElDhYxB7wHpux1_SLFHGUcEkgTNMOe2Q&e= > > > > > > > > Same goes for the LTE folk. > > > > > > E) Andrew mcgregor mentioned how great it would be for a closeted > musician > > > to > > > be able to play in real time with someone across town. that has been my > > > goal > > > for nearly 30 years now!! And although I rather enjoyed his > participation > > > in > > > my last talk on the subject ( > > > > > > > https://blog.apnic.net/2020/01/22/bufferbloat-may-be-solved-but-its-not-over-yet/ > > > < > https://urldefense.proofpoint.com/v2/url?u=https-3A__blog.apnic.net_2020_01_22_bufferbloat-2Dmay-2Dbe-2Dsolved-2Dbut-2Dits-2Dnot-2Dover-2Dyet_&d=DwMFaQ&c=96ZbZZcaMF4w0F4jpN6LZg&r=bqnFROivDo_4iF8Z3R4DyNWKbbMeXr0LOgLnElT1Ook&m=j5nEJ3W8fRmqjnBSWapTVKj6dNbpegl4kSeynebCQT4&s=BSDbzxnB7k7krFmkHv9id0BeDC6Vh39LgPNxyHUIg34&e= > > > > > ) conflating > > > a need for ecn and l4s signalling for low latency audio applications > > > with what I actually said in that talk, kind of hurt. I achieved > > > "my 2ms fiber based guitarist to fiber based drummer dream" 4+ years > > > back with fq_codel and diffserv, no ecn required, > > > no changes to the specs, no mandating packets be undroppable" and > > > would like to rip the opus codec out of that mix one day. > > > > > > F) I agree with jana that changing the definition of RFC3168 to suit > > > the RED algorithm (which is not pi or anything fancy) often present in > > > network switches, > > > today to suit dctcp, works. But you should say "configuring red to > > > have l4s marking style" and document that. > > > > > > Sometimes I try to point out many switches have a form of DRR in them, > > > and it's helpful to use that in conjunction with whatever diffserv > > > markings you trust in your network. > > > > > > To this day I wish someone would publish how much they use DCTCP style > > > signalling on a dc network relative to their other traffic. > > > > > > To this day I keep hoping that someone will publish a suitable > > > set of RED parameters for a wide variety of switches and routers - > > > for the most common switches and ethernet chips, for correct DCTCP > usage. > > > > > > Mellonox's example: > > > ( > > > > https://community.mellanox.com/s/article/howto-configure-ecn-on-mellanox-ethernet-switches--spectrum-x > > > < > https://urldefense.proofpoint.com/v2/url?u=https-3A__community.mellanox.com_s_article_howto-2Dconfigure-2Decn-2Don-2Dmellanox-2Dethernet-2Dswitches-2D-2Dspectrum-2Dx&d=DwMFaQ&c=96ZbZZcaMF4w0F4jpN6LZg&r=bqnFROivDo_4iF8Z3R4DyNWKbbMeXr0LOgLnElT1Ook&m=j5nEJ3W8fRmqjnBSWapTVKj6dNbpegl4kSeynebCQT4&s=nEIW1DhRXOHu3F5tMwpyO5rQUBMfCZx3Hs4wVvkVFIQ&e= > > > > > ) is not dctcp specific. > > > > > > many switches have a form of DRR in them, and it's helpful to use that > > > in conjunction with whatever diffserv markings you trust in your > > > network, > > > and, as per the above example, segregate two red queues that way. From > > > what I see > > > above there is no way to differentiate ECT(0) from ECT(1) in that > switch. > > > (?) > > > > > > I do keep trying to point out the size of the end user ecn enabled > > > deployment, starting with the data I have from free.fr > > > < > https://urldefense.proofpoint.com/v2/url?u=http-3A__free.fr&d=DwMFaQ&c=96ZbZZcaMF4w0F4jpN6LZg&r=bqnFROivDo_4iF8Z3R4DyNWKbbMeXr0LOgLnElT1Ook&m=j5nEJ3W8fRmqjnBSWapTVKj6dNbpegl4kSeynebCQT4&s=7gswGhl21lejSnIiu3yyUTPZEArHqQG6hD64BoW2Zco&e= > >. > > > Are we > > > building a network for AIs or people? > > > > > > G) Jana also made a point about 2 queues "being enough" (I might be > > > mis-remembering the exact point). Mellonoxes ethernet chips at 10Gig > expose > > > 64 hardware queues, some new intel hardware exposes 2000+. How do these > > > queues work relative to these algorithms? > > > > > > We have generally found hw mq to be far less of a benefit than the > > > manufacturers think, especially as regard to > > > lower latency or reduced cpu usage (as cache crossing is a bear). > > > There is a lot of software work in this area left to be done, however > > > they are needed to match queues to cpus (and tenants) > > > > > > Until sch_pie gained timestamping support recently, the rate estimator > > > did not work correctly in a hw mq environment. Haven't looked over > > > dualpi in this respect. > > > > > > > > > > > > > > > > > > -- > > > Make Music, Not War > > > > > > Dave T?ht > > > CTO, TekLibre, LLC > > > http://www.teklibre.com > > > < > https://urldefense.proofpoint.com/v2/url?u=http-3A__www.teklibre.com&d=DwMFaQ&c=96ZbZZcaMF4w0F4jpN6LZg&r=bqnFROivDo_4iF8Z3R4DyNWKbbMeXr0LOgLnElT1Ook&m=j5nEJ3W8fRmqjnBSWapTVKj6dNbpegl4kSeynebCQT4&s=DqPVjNVWDmF4_cwubNhhJS4Y1jCj71szPiBn9pmDZ70&e= > > > > > Tel: 1-831-435-0729 > > > _______________________________________________ > > > Bloat mailing list > > > Bloat@lists.bufferbloat.net > > > https://lists.bufferbloat.net/listinfo/bloat > > > < > https://urldefense.proofpoint.com/v2/url?u=https-3A__lists.bufferbloat..net_listinfo_bloat&d=DwMFaQ&c=96ZbZZcaMF4w0F4jpN6LZg&r=bqnFROivDo_4iF8Z3R4DyNWKbbMeXr0LOgLnElT1Ook&m=j5nEJ3W8fRmqjnBSWapTVKj6dNbpegl4kSeynebCQT4&s=DBDxIR6eSYcBOh7rqZx0PWzsHOfvvJeqioI3r2IQOA4&e= > > > > > > > > > > -- > Rod Grimes > rgri...@freebsd.org >
_______________________________________________ Bloat mailing list Bloat@lists.bufferbloat.net https://lists.bufferbloat.net/listinfo/bloat