Comcast says it represents a 10 Gigabit cable internet network they are building (it doesn’t exist) so they are basically changing the meaning of the g from generation to gig to act like 10g is 5 generations better (or twice as fast)…or that they have a 10 gigabit network. Neither is accurate. It’s still just cable internet that people have to use because they have no other option.

Fuck Comcast.

I read online they are abandoning the “confusing” 10g branding but I just saw a commercial for it. They think all of their customers are morons and count on folks having no other choices in a lot of cases.

Apologies to anyone outside the United States, this is just complaining about our poor internet options and deceptive advertising by greedy corporations.

  • BorgDrone@lemmy.one
    link
    fedilink
    English
    arrow-up
    3
    ·
    10 months ago

    Not just internet providers. Data communication speeds have always been in bits per second. Historically it makes perfect sense.

    Specifying speed in bytes per second would be inconvenient because while we settled on 8 bits per byte in the early days of computing this was not the case. 6-bit bytes were common, but other sizes were used too, 7,8, 9, 10 and sometimes even larger.

    So when you’re talking about communication between different types of computers with different size bytes, it would be confusing to use bytes/second as a unit.

    • vithigar@lemmy.ca
      link
      fedilink
      English
      arrow-up
      1
      ·
      edit-2
      10 months ago

      Even now that we’ve by and large settled on 8 bits per byte it’s still useful to call out the communication rate as distinct from the actual payload data transfer rate, as there are other sources of overhead.

      You’ll never actually see a 1MB/s transfer over an 8Mbps connection because some of those bits are going to be used for things like packet headers, keep alive messages, etc.