- cross-posted to:
- technology@lemmy.world
- cross-posted to:
- technology@lemmy.world
These companies are constantly mislabeling these networks every generation. They are addicted to advertising bullshit.
See AT&T labeling HSPA as “4Ge” and enhanced LTE as “5Ge”.
Neither of which is actually 4G or 5G.
This shit started with 3g. None of the carriers were even close to supporting 3g, which was designed to support 100mbps.
Tmo and vzw had a modest improvement from 2g even though it was still slow as shit but AT&T simple did nothing and started calling theirs 3g.
4G LTE was the point of no return. It was supposed to mean “it’s not 4G yet but we have an upgrade plan to get there”, but when they finally did, marketing found out that to the average person, going from 4G LTE to 4G sounded like a downgrade, so they rebranded it to 5G.
What kills me is it’s the same with CPUs. A 5nm chip is larger than 5nm.
I thought it meant the transistor size.
It doesn’t really mean anything anymore. The transistors are not 5nm either. It’s just marketing.
Quoth Wikipedia:
The term “5 nm” has no relation to any actual physical feature (such as gate length, metal pitch or gate pitch) of the transistors being 5 nanometers in size. According to the projections contained in the 2021 update of the International Roadmap for Devices and Systems published by IEEE Standards Association Industry Connection, a “5 nm node is expected to have a contacted gate pitch of 51 nanometers and a tightest metal pitch of 30 nanometers”. However, in real world commercial practice, “5 nm” is used primarily as a marketing term by individual microchip manufacturers to refer to a new, improved generation of silicon semiconductor chips in terms of increased transistor density (i.e. a higher degree of miniaturization), increased speed and reduced power consumption compared to the previous 7 nm process.
Doesn’t it stand for generation? Like 3rd generation and 5th generation? This is referencing a wifi router is it not?
So sort of: the 3g is part of a standard for data rates, but the difficulty, comes in that networks are not homogeneous. Similarly to how you might be familiar with 5Ghz and 2.4Ghz WiFi signals. As a general rule of thumb the higher the frequency the more data you can send but with more attenuation so the signal can be blocked more easily and cannot travel as far, whereas the inverse is true for lower frequencies. So while the generations did make some changes in terms of protocols— it also came with higher frequency emitters which can theoretically carry more data. Other changes include MIMO antennas which do beam forming to make more of the energy go in the direction of a user using constructive and destructive interference from an array of antennas to accomplish this. However marketing people are always very eager to adopt technical terms and inflate them into oblivion. However some of this can be attributed to honest misunderstanding within a company.
Cool. Now add the modem I just bought 6 months ago, that supports the speeds I pay for and you had on your approved list before your recent updates back to the approved list.
Who the fuck cares about 10Gbit/s? With data caps, there is nothing I am downloading on a mobile device that is perceptibly faster than downloading it at 1/1000 of that speed
Imagine having data caps in 2024
And this a problem only in the monopolistic culture of the US. I can’t name a single EU State with broadband data cap.
Fine I’ll stop misleading people about 10g.
In other news have you heard about our 11g network?