I know this is a small side note but I felt compelled to speak up in defense of 
online gaming. I’m not a gamer at all and up till a year or two ago, would have 
agreed with Dick’s take about benefit to “society as a whole.” However, lately 
I’ve started hearing some research on the benefits of groups of friends using 
online games to socialize together, effectively using the game primarily as a 
group call.

There’s also this project, where people have collected banned/censored books 
into a library in Minecraft. Specifically as a solution to contexts where 
regulators/censors ban and monitor content through other channels (websites 
etc) but don’t surveil Minecraft... Presumably because they share Dick’s 
opinion ;-) https://www.uncensoredlibrary.com/en

> On Oct 17, 2023, at 03:26, Sebastian Moeller via Nnagain 
> <nnagain@lists.bufferbloat.net> wrote:
> 
> Hi Richard,
> 
> 
>> On Oct 16, 2023, at 20:04, Dick Roy <dick...@alum.mit.edu> wrote:
>> 
>> Good points all, Sebastien.  How to "trade-off" a fixed capacity amongst 
>> many users is ultimately a game theoretic problem when users are allowed to 
>> make choices, which is certainly the case here.  Secondly, any network that 
>> can and does generate "more traffic" (aka overhead such as ACKs NACKs and 
>> retries) reduces the capacity of the network, and ultimately can lead to the 
>> "user" capacity going to zero!  Such is life in the fast lane (aka the 
>> internet).
>> 
>> Lastly, on the issue of low-latency real-time experience, there are many 
>> applications that need/want such capabilities that actually have a net 
>> benefit to the individuals involved AND to society as a whole.  IMO, 
>> interactive gaming is NOT one of those.
> 
>       [SM] Yes, gaming is one obvious example of a class of uses that work 
> best with low latency and low jitter, not necessarily an example for a 
> use-case worthy enough to justify the work required to increase the 
> responsiveness of the internet. Other examples are video conferences, VoIP, 
> in extension of both musical collaboration over the internet, and surprising 
> to some even plain old web browsing (it often needs to first read a page 
> before being able to follow links and load resources, and every read takes at 
> best a single RTT). None of these are inherently beneficial or detrimental to 
> individuals or society, but most can be used to improve the status quo... I 
> would argue that in the last 4 years the relevance of interactive use-cases 
> has been made quite clear to a lot of folks...
> 
> 
>> OK, so now you know I don't engage in these time sinks with no redeeming 
>> social value.:)
> 
>       [SM] Duly noted ;)
> 
>> Since it is not hard to argue that just like power distribution, information 
>> exchange/dissemination is "in the public interest", the question becomes "Do 
>> we allow any and all forms of information exchange/dissemination over what 
>> is becoming something akin to a public utility?"  FWIW, I don't know the 
>> answer to this question! :)
> 
>       [SM] This is an interesting question and one (only) tangentially 
> related to network neutrality... it is more related to freedom of speech and 
> limits thereof. Maybe a question for another mailing list? Certainly one 
> meriting a topic change...
> 
> 
> Regards
>       Sebastian
> 
>> 
>> Cheers,
>> 
>> RR
>> 
>> -----Original Message-----
>> From: Sebastian Moeller [mailto:moell...@gmx.de] 
>> Sent: Monday, October 16, 2023 10:36 AM
>> To: dick...@alum.mit.edu; Network Neutrality is back! Let´s make the 
>> technical aspects heard this time!
>> Subject: Re: [NNagain] transit and peering costs projections
>> 
>> Hi Richard,
>> 
>> 
>>> On Oct 16, 2023, at 19:01, Dick Roy via Nnagain 
>>> <nnagain@lists.bufferbloat.net> wrote:
>>> 
>>> Just an observation:  ANY type of congestion control that changes 
>>> application behavior in response to congestion, or predicted congestion 
>>> (ENC), begs the question "How does throttling of application information 
>>> exchange rate (aka behavior) affect the user experience and will the user 
>>> tolerate it?" 
>> 
>>      [SM] The trade-off here is, if the application does not respond (or 
>> rather if no application would respond) we would end up with congestion 
>> collapse where no application would gain much of anything as the network 
>> busies itself trying to re-transmit dropped packets without making much head 
>> way... Simplistic game theory application might imply that individual 
>> applications could try to game this, and generally that seems to be true, 
>> but we have remedies for that available..
>> 
>> 
>>> 
>>> Given any (complex and packet-switched) network topology of interconnected 
>>> nodes and links, each with possible a different capacity and 
>>> characteristics, such as the internet today, IMO the two fundamental 
>>> questions are:
>>> 
>>> 1) How can a given network be operated/configured so as to maximize 
>>> aggregate throughput (i.e. achieve its theoretical capacity), and
>>> 2) What things in the network need to change to increase the throughput 
>>> (aka parameters in the network with the largest Lagrange multipliers 
>>> associated with them)?
>> 
>>      [SM] The thing is we generally know how to maximize (average) 
>> throughput, just add (over-)generous amounts of buffering, the problem is 
>> that this screws up the other important quality axis, latency... We ideally 
>> want low latency and even more low latency variance (aka jitter) AND high 
>> throughput... Turns out though that above a certain throughput threshold* 
>> many users do not seem to care all that much for more throughput as long as 
>> interactive use cases are sufficiently responsive... but high responsiveness 
>> requires low latency and low jitter... This is actually a good thing, as 
>> that means we do not necessarily aim for 100% utilization (almost requiring 
>> deep buffers and hence resulting in compromised latency) but can get away 
>> with say 80-90% where shallow buffers will do (or rather where buffer 
>> filling stays shallow, there is IMHO still value in having deep buffers for 
>> rare events that need it).
>> 
>> 
>> 
>> *) This is not a hard physical law so the exact threshold is not set in 
>> stone, but unless one has many parallel users, something in the 20-50 Mbps 
>> range is plenty and that is only needed in the "loaded" direction, that is 
>> for pure consumers the upload can be thinner, for pure producers the 
>> download can be thinner.
>> 
>> 
>>> 
>>> I am not an expert in this field,
>> 
>>      [SM] Nor am I, I come from the wet-ware side of things so not even 
>> soft- or hard-ware ;)
>> 
>> 
>>> however it seems to me that answers to these questions would be useful, 
>>> assuming they are not yet available!
>>> 
>>> Cheers,
>>> 
>>> RR
>>> 
>>> 
>>> 
>>> -----Original Message-----
>>> From: Nnagain [mailto:nnagain-boun...@lists.bufferbloat.net] On Behalf Of 
>>> rjmcmahon via Nnagain
>>> Sent: Sunday, October 15, 2023 1:39 PM
>>> To: Network Neutrality is back! Let´s make the technical aspects heard this 
>>> time!
>>> Cc: rjmcmahon
>>> Subject: Re: [NNagain] transit and peering costs projections
>>> 
>>> Hi Jack,
>>> 
>>> Thanks again for sharing. It's very interesting to me.
>>> 
>>> Today, the networks are shifting from capacity constrained to latency 
>>> constrained, as can be seen in the IX discussions about how the speed of 
>>> light over fiber is too slow even between Houston & Dallas.
>>> 
>>> The mitigations against standing queues (which cause bloat today) are:
>>> 
>>> o) Shrink the e2e bottleneck queue so it will drop packets in a flow and 
>>> TCP will respond to that "signal"
>>> o) Use some form of ECN marking where the network forwarding plane 
>>> ultimately informs the TCP source state machine so it can slow down or 
>>> pace effectively. This can be an earlier feedback signal and, if done 
>>> well, can inform the sources to avoid bottleneck queuing. There are 
>>> couple of approaches with ECN. Comcast is trialing L4S now which seems 
>>> interesting to me as a WiFi test & measurement engineer. The jury is 
>>> still out on this and measurements are needed.
>>> o) Mitigate source side bloat via TCP_NOTSENT_LOWAT
>>> 
>>> The QoS priority approach per congestion is orthogonal by my judgment as 
>>> it's typically not supported e2e, many networks will bleach DSCP 
>>> markings. And it's really too late by my judgment.
>>> 
>>> Also, on clock sync, yes your generation did us both a service and 
>>> disservice by getting rid of the PSTN TDM clock ;) So IP networking 
>>> devices kinda ignored clock sync, which makes e2e one way delay (OWD) 
>>> measurements impossible. Thankfully, the GPS atomic clock is now 
>>> available mostly everywhere and many devices use TCXO oscillators so 
>>> it's possible to get clock sync and use oscillators that can minimize 
>>> drift. I pay $14 for a Rpi4 GPS chip with pulse per second as an 
>>> example.
>>> 
>>> It seems silly to me that clocks aren't synced to the GPS atomic clock 
>>> even if by a proxy even if only for measurement and monitoring.
>>> 
>>> Note: As Richard Roy will point out, there really is no such thing as 
>>> synchronized clocks across geographies per general relativity - so those 
>>> syncing clocks need to keep those effects in mind. I limited the iperf 2 
>>> timestamps to microsecond precision in hopes avoiding those issues.
>>> 
>>> Note: With WiFi, a packet drop can occur because an intermittent RF 
>>> channel condition. TCP can't tell the difference between an RF drop vs a 
>>> congested queue drop. That's another reason ECN markings from network 
>>> devices may be better than dropped packets.
>>> 
>>> Note: I've added some iperf 2 test support around pacing as that seems 
>>> to be the direction the industry is heading as networks are less and 
>>> less capacity strained and user quality of experience is being driven by 
>>> tail latencies. One can also test with the Prague CCA for the L4S 
>>> scenarios. (This is a fun project: https://www.l4sgear.com/ and fairly 
>>> low cost)
>>> 
>>> --fq-rate n[kmgKMG]
>>> Set a rate to be used with fair-queuing based socket-level pacing, in 
>>> bytes or bits per second. Only available on platforms supporting the 
>>> SO_MAX_PACING_RATE socket option. (Note: Here the suffixes indicate 
>>> bytes/sec or bits/sec per use of uppercase or lowercase, respectively)
>>> 
>>> --fq-rate-step n[kmgKMG]
>>> Set a step of rate to be used with fair-queuing based socket-level 
>>> pacing, in bytes or bits per second. Step occurs every 
>>> fq-rate-step-interval (defaults to one second)
>>> 
>>> --fq-rate-step-interval n
>>> Time in seconds before stepping the fq-rate
>>> 
>>> Bob
>>> 
>>> PS. Iperf 2 man page https://iperf2.sourceforge.io/iperf-manpage.html
>>> 
>>>> The "VGV User" (Voice, Gaming, Videoconferencing) cares a lot about
>>>> latency.   It's not just "rewarding" to have lower latencies; high
>>>> latencies may make VGV unusable.   Average (or "typical") latency as
>>>> the FCC label proposes isn't a good metric to judge usability.  A path
>>>> which has high variance in latency can be unusable even if the average
>>>> is quite low.   Having your voice or video or gameplay "break up"
>>>> every minute or so when latency spikes to 500 msec makes the "user
>>>> experience" intolerable.
>>>> 
>>>> A few years ago, I ran some simple "ping" tests to help a friend who
>>>> was trying to use a gaming app.  My data was only for one specific
>>>> path so it's anecdotal.  What I saw was surprising - zero data loss,
>>>> every datagram was delivered, but occasionally a datagram would take
>>>> up to 30 seconds to arrive.  I didn't have the ability to poke around
>>>> inside, but I suspected it was an experience of "bufferbloat", enabled
>>>> by the dramatic drop in price of memory over the decades.
>>>> 
>>>> It's been a long time since I was involved in operating any part of
>>>> the Internet, so I don't know much about the inner workings today.
>>>> Apologies for my ignorance....
>>>> 
>>>> There was a scenario in the early days of the Internet for which we
>>>> struggled to find a technical solution.  Imagine some node in the
>>>> bowels of the network, with 3 connected "circuits" to some other
>>>> nodes.  On two of those inputs, traffic is arriving to be forwarded
>>>> out the third circuit.  The incoming flows are significantly more than
>>>> the outgoing path can accept.
>>>> 
>>>> What happens?   How is "backpressure" generated so that the incoming
>>>> flows are reduced to the point that the outgoing circuit can handle
>>>> the traffic?
>>>> 
>>>> About 45 years ago, while we were defining TCPV4, we struggled with
>>>> this issue, but didn't find any consensus solutions.  So "placeholder"
>>>> mechanisms were defined in TCPV4, to be replaced as research continued
>>>> and found a good solution.
>>>> 
>>>> In that "placeholder" scheme, the "Source Quench" (SQ) IP message was
>>>> defined; it was to be sent by a switching node back toward the sender
>>>> of any datagram that had to be discarded because there wasn't any
>>>> place to put it.
>>>> 
>>>> In addition, the TOS (Type Of Service) and TTL (Time To Live) fields
>>>> were defined in IP.
>>>> 
>>>> TOS would allow the sender to distinguish datagrams based on their
>>>> needs.  For example, we thought "Interactive" service might be needed
>>>> for VGV traffic, where timeliness of delivery was most important. 
>>>> "Bulk" service might be useful for activities like file transfers,
>>>> backups, et al.   "Normal" service might now mean activities like
>>>> using the Web.
>>>> 
>>>> The TTL field was an attempt to inform each switching node about the
>>>> "expiration date" for a datagram.   If a node somehow knew that a
>>>> particular datagram was unlikely to reach its destination in time to
>>>> be useful (such as a video datagram for a frame that has already been
>>>> displayed), the node could, and should, discard that datagram to free
>>>> up resources for useful traffic.  Sadly we had no mechanisms for
>>>> measuring delay, either in transit or in queuing, so TTL was defined
>>>> in terms of "hops", which is not an accurate proxy for time.   But
>>>> it's all we had.
>>>> 
>>>> Part of the complexity was that the "flow control" mechanism of the
>>>> Internet had put much of the mechanism in the users' computers' TCP
>>>> implementations, rather than the switches which handle only IP.
>>>> Without mechanisms in the users' computers, all a switch could do is
>>>> order more circuits, and add more memory to the switches for queuing. 
>>>> Perhaps that led to "bufferbloat".
>>>> 
>>>> So TOS, SQ, and TTL were all placeholders, for some mechanism in a
>>>> future release that would introduce a "real" form of Backpressure and
>>>> the ability to handle different types of traffic.   Meanwhile, these
>>>> rudimentary mechanisms would provide some flow control. Hopefully the
>>>> users' computers sending the flows would respond to the SQ
>>>> backpressure, and switches would prioritize traffic using the TTL and
>>>> TOS information.
>>>> 
>>>> But, being way out of touch, I don't know what actually happens
>>>> today.  Perhaps the current operators and current government watchers
>>>> can answer?:git clone https://rjmcma...@git.code.sf.net/p/iperf2/code 
>>>> iperf2-code
>>>> 
>>>> 1/ How do current switches exert Backpressure to  reduce competing
>>>> traffic flows?  Do they still send SQs?
>>>> 
>>>> 2/ How do the current and proposed government regulations treat the
>>>> different needs of different types of traffic, e.g., "Bulk" versus
>>>> "Interactive" versus "Normal"?  Are Internet carriers permitted to
>>>> treat traffic types differently?  Are they permitted to charge
>>>> different amounts for different types of service?
>>>> 
>>>> Jack Haverty
>>>> 
>>>> On 10/15/23 09:45, Dave Taht via Nnagain wrote:
>>>>> For starters I would like to apologize for cc-ing both nanog and my
>>>>> new nn list. (I will add sender filters)
>>>>> 
>>>>> A bit more below.
>>>>> 
>>>>> On Sun, Oct 15, 2023 at 9:32 AM Tom Beecher <beec...@beecher.cc> 
>>>>> wrote:
>>>>>>> So for now, we'll keep paying for transit to get to the others 
>>>>>>> (since it’s about as much as transporting IXP from Dallas), and 
>>>>>>> hoping someone at Google finally sees Houston as more than a third 
>>>>>>> rate city hanging off of Dallas. Or… someone finally brings a 
>>>>>>> worthwhile IX to Houston that gets us more than peering to Kansas 
>>>>>>> City. Yeah, I think the former is more likely. 😊
>>>>>> 
>>>>>> There is often a chicken/egg scenario here with the economics. As an 
>>>>>> eyeball network, your costs to build out and connect to Dallas are 
>>>>>> greater than your transit cost, so you do that. Totally fair.
>>>>>> 
>>>>>> However think about it from the content side. Say I want to build 
>>>>>> into to Houston. I have to put routers in, and a bunch of cache 
>>>>>> servers, so I have capital outlay , plus opex for space, power, 
>>>>>> IX/backhaul/transit costs. That's not cheap, so there's a lot of 
>>>>>> calculations that go into it. Is there enough total eyeball traffic 
>>>>>> there to make it worth it? Is saving 8-10ms enough of a performance 
>>>>>> boost to justify the spend? What are the long term trends in that 
>>>>>> market? These answers are of course different for a company running 
>>>>>> their own CDN vs the commercial CDNs.
>>>>>> 
>>>>>> I don't work for Google and obviously don't speak for them, but I 
>>>>>> would suspect that they're happy to eat a 8-10ms performance hit to 
>>>>>> serve from Dallas , versus the amount of capital outlay to build out 
>>>>>> there right now.
>>>>> The three forms of traffic I care most about are voip, gaming, and
>>>>> videoconferencing, which are rewarding to have at lower latencies.
>>>>> When I was a kid, we had switched phone networks, and while the sound
>>>>> quality was poorer than today, the voice latency cross-town was just
>>>>> like "being there". Nowadays we see 500+ms latencies for this kind of
>>>>> traffic.
>>>>> 
>>>>> As to how to make calls across town work that well again, cost-wise, I
>>>>> do not know, but the volume of traffic that would be better served by
>>>>> these interconnects quite low, respective to the overall gains in
>>>>> lower latency experiences for them.
>>>>> 
>>>>> 
>>>>> 
>>>>>> On Sat, Oct 14, 2023 at 11:47 PM Tim Burke <t...@mid.net> wrote:
>>>>>>> I would say that a 1Gbit IP transit in a carrier neutral DC can be 
>>>>>>> had for a good bit less than $900 on the wholesale market.
>>>>>>> 
>>>>>>> Sadly, IXP’s are seemingly turning into a pay to play game, with 
>>>>>>> rates almost costing as much as transit in many cases after you 
>>>>>>> factor in loop costs.
>>>>>>> 
>>>>>>> For example, in the Houston market (one of the largest and fastest 
>>>>>>> growing regions in the US!), we do not have a major IX, so to get up 
>>>>>>> to Dallas it’s several thousand for a 100g wave, plus several 
>>>>>>> thousand for a 100g port on one of those major IXes. Or, a better 
>>>>>>> option, we can get a 100g flat internet transit for just a little 
>>>>>>> bit more.
>>>>>>> 
>>>>>>> Fortunately, for us as an eyeball network, there are a good number 
>>>>>>> of major content networks that are allowing for private peering in 
>>>>>>> markets like Houston for just the cost of a cross connect and a QSFP 
>>>>>>> if you’re in the right DC, with Google and some others being the 
>>>>>>> outliers.
>>>>>>> 
>>>>>>> So for now, we'll keep paying for transit to get to the others 
>>>>>>> (since it’s about as much as transporting IXP from Dallas), and 
>>>>>>> hoping someone at Google finally sees Houston as more than a third 
>>>>>>> rate city hanging off of Dallas. Or… someone finally brings a 
>>>>>>> worthwhile IX to Houston that gets us more than peering to Kansas 
>>>>>>> City. Yeah, I think the former is more likely. 😊
>>>>>>> 
>>>>>>> See y’all in San Diego this week,
>>>>>>> Tim
>>>>>>> 
>>>>>>> On Oct 14, 2023, at 18:04, Dave Taht <dave.t...@gmail.com> wrote:
>>>>>>>> This set of trendlines was very interesting. Unfortunately the 
>>>>>>>> data
>>>>>>>> stops in 2015. Does anyone have more recent data?
>>>>>>>> 
>>>>>>>> https://drpeering.net/white-papers/Internet-Transit-Pricing-Historical-And-Projected.php
>>>>>>>> 
>>>>>>>> I believe a gbit circuit that an ISP can resell still runs at about
>>>>>>>> $900 - $1.4k (?) in the usa? How about elsewhere?
>>>>>>>> 
>>>>>>>> ...
>>>>>>>> 
>>>>>>>> I am under the impression that many IXPs remain very successful,
>>>>>>>> states without them suffer, and I also find the concept of doing 
>>>>>>>> micro
>>>>>>>> IXPs at the city level, appealing, and now achievable with cheap 
>>>>>>>> gear.
>>>>>>>> Finer grained cross connects between telco and ISP and IXP would 
>>>>>>>> lower
>>>>>>>> latencies across town quite hugely...
>>>>>>>> 
>>>>>>>> PS I hear ARIN is planning on dropping the price for, and bundling 
>>>>>>>> 3
>>>>>>>> BGP AS numbers at a time, as of the end of this year, also.
>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>>>> --
>>>>>>>> Oct 30: 
>>>>>>>> https://netdevconf.info/0x17/news/the-maestro-and-the-music-bof.html
>>>>>>>> Dave Täht CSO, LibreQos
>>>>> 
>>>>> 
>>>> 
>>>> _______________________________________________
>>>> Nnagain mailing list
>>>> Nnagain@lists.bufferbloat.net
>>>> https://lists.bufferbloat.net/listinfo/nnagain
>>> _______________________________________________
>>> Nnagain mailing list
>>> Nnagain@lists.bufferbloat.net
>>> https://lists.bufferbloat.net/listinfo/nnagain
>>> 
>>> _______________________________________________
>>> Nnagain mailing list
>>> Nnagain@lists.bufferbloat.net
>>> https://lists.bufferbloat.net/listinfo/nnagain
>> 
> 
> _______________________________________________
> Nnagain mailing list
> Nnagain@lists.bufferbloat.net
> https://lists.bufferbloat.net/listinfo/nnagain

_______________________________________________
Nnagain mailing list
Nnagain@lists.bufferbloat.net
https://lists.bufferbloat.net/listinfo/nnagain

Reply via email to