Thank you Jonathan for this clear description of the issues and their history. I wonder if there's a fourth one - privacy.
Rosenworcel's talk https://docs.fcc.gov/public/attachments/DOC-397257A1.pdf also points out that ISPs might want to monetize our traffic patterns and location data. (This is less of an issue in the EU, but the US remains a Wild West in this regard.) I am hopeful that the FCC will include this in their NPRM (which must be available by now but I haven't looked...) - Rich Brown > On Sep 29, 2023, at 12:54 AM, Jonathan Morton via Rpm > <r...@lists.bufferbloat.net> wrote: > >> On 29 Sep, 2023, at 1:19 am, David Lang via Bloat >> <bloat@lists.bufferbloat.net> wrote: >> >> Dave T called out earlier that the rise of bittorrent was a large part of >> the inital NN discussion here in the US. But a second large portion was a >> money grab from ISPs thinking that they could hold up large paid websites >> (netflix for example) for additional fees by threatening to make their >> service less useful to their users (viewing their users as an asset to be >> marketed to the websites rather than customers to be satisfied by providing >> them access to the websites) >> >> I don't know if a new round of "it's not fair that Netflix doesn't pay us >> for the bandwidth to service them" would fall flat at this point or not. > > I think there were three more-or-less separate concerns which have, over > time, fallen under the same umbrella: > > > 1: Capacity-seeking flows tend to interfere with latency-sensitive flows, > and the "induced demand" phenomenon means that increases in link rate do not > in themselves solve this problem, even though they may be sold as doing so. > > This is directly addressed by properly-sized buffers and/or AQM, and even > better by FQ and SQM. It's a solved problem, so long as the solutions are > deployed. It's not usually necessary, for example, to specifically enhance > service for latency-sensitive traffic, if FQ does a sufficiently good job. > An increased link rate *does* enhance service quality for both > latency-sensitive and capacity-seeking traffic, provided FQ is in use. > > > 2: Swarm traffic tends to drown out conventional traffic, due to congestion > control algorithms which try to be more-or-less fair on a per-flow basis, and > the substantially larger number of parallel flows used by swarm traffic. > This also caused subscribers using swarm traffic to impair the service of > subscribers who had nothing to do with it. > > FQ on a per-flow basis (see problem 1) actually amplifies this effect, and I > think it was occasionally used as an argument for *not* deploying FQ. ISPs' > initial response was to outright block swarm traffic where they could > identify it, which was then softened to merely throttling it heavily, before > NN regulations intervened. Usage quotas also showed up around this time, and > were probably related to this problem. > > This has since been addressed by several means. ISPs may use FQ on a > per-subscriber basis to prevent one subscriber's heavy traffic from degrading > service for another. Swarm applications nowadays tend to employ altruistic > congestion control which deliberately compensates for the large number of > flows, and/or mark them with one or more of the Least Effort class DSCPs. > Hence, swarm applications are no longer as damaging to service quality as > they used to be. Usage quotas, however, still remain in use as a profit > centre, to the point where an "unlimited" service is a rare and precious > specimen in many jurisdictions. > > > 3: ISPs merged with media distribution companies, creating a conflict of > interest in which the media side of the business wanted the internet side to > actively favour "their own" media traffic at the expense of "the > competition". Some ISPs began to actively degrade Netflix traffic, in > particular by refusing to provision adequate peering capacity at the nodes > through which Netflix traffic predominated, or by zero-rating (for the > purpose of usage quotas) traffic from their own media empire while refusing > to do the same for Netflix traffic. > > **THIS** was the true core of Net Neutrality. NN regulations forced ISPs to > carry Netflix traffic with reasonable levels of service, even though they > didn't want to for purely selfish and greedy commercial reasons. NN > succeeded in curbing an anti-competitive and consumer-hostile practice, which > I am perfectly sure would resume just as soon as NN regulations were repealed. > > And this type of practice is just the sort of thing that technologies like > L4S are designed to support. The ISPs behind L4S actively do not want a > technology that works end-to-end over the general Internet. They want > something that can provide a domination service within their own walled > gardens. That's why L4S is a NN hazard, and why they actively resisted all > attempts to displace it with SCE. > > > All of the above were made more difficult to solve by the monopolistic nature > of the Internet service industry. It is actively difficult for Internet > users to move to a truly different service, especially one based on a > different link technology. When attempts are made to increase competition, > for example by deploying a publicly-funded network, the incumbents actively > sabotage those attempts by any means they can. Monopolies are inherently > customer-hostile, and arguments based on market forces fail in their presence. > > - Jonathan Morton > > _______________________________________________ > Rpm mailing list > r...@lists.bufferbloat.net > https://lists.bufferbloat.net/listinfo/rpm _______________________________________________ Bloat mailing list Bloat@lists.bufferbloat.net https://lists.bufferbloat.net/listinfo/bloat