Let me clarify: We have a certain amount of latency when streaming games from both local and internet servers. In either case, how do we improve that latency and what limits will we run in to as the technology progresses?

  • Dark Arc@social.packetloss.gg
    link
    fedilink
    English
    arrow-up
    1
    ·
    9 months ago

    I wear a lot of hats professionally; mostly programming. I don’t do networking on a day-to-day basis though if that’s what you’re asking.

    If you’ve got something actually substantive to back up your claim that (if money was no object) the current topology is totally optimal for traffic from an arbitrary point A <-> B on that map though… have at it.

    This all started with:

    you’re often back-tracking across the continent before your traffic makes it to the end destination, with ISPs caring more about saving money than routing traffic quickly

    And that’s absolutely true … depending on your location, you will travel an unnecessary distance to get to your destination … because there just aren’t wires connecting A <-> B. Just like a GPS will take you on a non-direct path to your destination because there’s not a road directly to it.

    A very simple example where the current topology results in routing all the way out to Seattle only to backtrack: https://geotraceroute.com/?node=0&amp;host=umt.edu#

    • Blake [he/him]@feddit.uk
      link
      fedilink
      arrow-up
      1
      ·
      9 months ago

      The problem that I’m having (and why I asked that) is because I was assuming that you would have some knowledge which you don’t seem to have with a lot of my comments. I’m really not trying to be rude, but it makes it a lot more difficult to explain the flaws in your reasoning when you’re talking about topics that are beyond your knowledge as if you know them well.

      I have explained the realities of the situation to you, if you don’t want to accept them, that’s fine, but you’re basically arguing with an expert about something you don’t really understand very well. I’m happy to explain stuff but you should just ask rather than assume you know better because it makes it much more difficult for me to understand the gaps in your understanding/knowledge.

      So ultimately, for routers, we have a number of limited resources. Firstly, yes, interfaces, but also the usual stuff - CPU, RAM, etc.

      Now, I mentioned before that routing protocols are very complex - they have many metrics which are taken into account to determine what path is ultimately best for each packet. This is a process which can be quite intensive on CPU and RAM - because the router needs to “remember” all of the possible routes/destinations a packet can travel, as well as all of the metrics for each destination - distance, delays, administrative distance, TTL, dropped packets, etc. and then make a decision about processing it. And it needs to make these decisions billions of times a second. Slowing it down, even a tiny bit, can hugely impact the total throughout of the router.

      When you add another connection to a router, you’re not just increasing the load for that one router, but for the routers which connect to the routers which connect to those routers which route to the routers that route to that router… you get the idea. It increases the number of options available, and so it places additional burden on memory and processing. When the ultimate difference in distance even an extra 100 miles, that’s less than a millisecond of travelling time. It’s not worth the added complexity.

      That’s what I meant when I said that an extra hop isn’t worth worrying about, but adding additional connections is inefficient.

      • Dark Arc@social.packetloss.gg
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        9 months ago

        but you’re basically arguing with an expert about something you don’t really understand very well. I’m happy to explain stuff but you should just ask rather than assume you know better because it makes it much more difficult for me to understand the gaps in your understanding/knowledge.

        Okay, I’ll apologize… For context though, in general, it’s the internet and it’s hard to take “expert” at its word (and even outside of an online context, “expert” is a title I’m often skeptical of … even when it’s assigned to me :) ). I’ve argued with plenty of people (more so on Reddit) that are CS students… It’s just the price of being on the internet I guess, ha

        I’m still not sure I agree with your conclusions, but that’s mostly healthy skepticism… because your argument isn’t tracking with … well … physics or distributed computing… more direct “routes” and taking load off “routes” that aren’t the optimal route typically is a great way to speed up a system. It’s definitely true that doing that adds overhead vs just having a few “better” systems do the work (at least from some perspectives), but it’s hard for me to imagine that with sufficient funds it truly makes it worse to give routing algorithms more direct options and/or cut out unnecessary hops entirely.

        Reducing “hops” and travel time is kind of the bread and butter of performance work when it comes to all kinds of optimizations in software engineering…

        If you want me to ask a question … what’s your explanation for why there are so many more connections in the north east and west coast if more connections slows the whole system down? Why not just have a handful of routes?