Hard to believe it's been 24 years since Y2K (2000) And it feels like we've come such a long way, but this decade started off very poorly with one of the worst pandemics the modern world has ever seen, and technology in general is looking very bleak in several ways
I'm a PC gamer, and it looks like things are stagnating massively in our space. So many gaming companies are incapable of putting out a successful AAA title because people are either too poor, don't want to play a live service AAA disaster like every single one that has been released lately, Call of Duty, battlefield, anything electronic arts or Ubisoft puts out is almost entirely a failure or undersales. So many gaming studios have been shuttered and are being shuttered, Microsoft is basically one member of an oligopoly with Sony and a couple other companies.
Hardware is stagnating. Nvidia is putting on the brakes for developing their next line of GPUs, we're not going to see huge gains in performance anymore because AMD isn't caught up yet and they have no reason to innovate. So they are just going to sell their next line of cards for $1,500 a pop for the top ones, with 10% increase in performance rather than 50 or 60% like we really need. We still don't have the capability to play games in full native 4K 144 Hertz. That's at least a decade away
Virtual reality is on the verge of collapse because meta is basically the only real player in that space, they have a monopoly with them and valve index, pico from China is on the verge of developing something incredible as well, and Apple just revealed a mixed reality headset but the price is so extraordinary that barely anyone has it so use isn't very widespread. We're again a decade away from seeing anything really substantial in terms of performance
Artificial intelligence is really, really fucking things up in general and the discussions about AI look almost as bad as the news about the latest election in the USA. It's so clowny and ridiculous and over-the-top hearing any news about AI. The latest news is that open AI is going to go from a non-profit to a for-profit company after they promised they were operating for the good of humanity and broke countless laws stealing copyrighted information, supposedly for the public good, but now they're just going to snap their fingers and morph into a for-profit company. So they can just basically steal anything they want that's copyrighted, but claim it's for the public good, and then randomly swap to a for-profit model. Doesn't make any sense and just looks like they're going to be a vessel for widespread economic poverty...
It just seems like there's a lot of bubbles that are about to burst all at the same time, like I don't see how things are going to possibly get better for a while now?
Well, that's the doomer take.
The rumors are that the 80 series card is 10% faster than the 90 series card from last gen: that's not a '10%' improvement, assuming the prices are the same, that's more like a 40% improvement. I think a LOT of people don't realize how shitty the 4080 was compared to the 4090 and are vastly mis-valuing that rumor.
I'd also argue the 'GAMES MUST BE ULTRA AT 4K144 OR DONT BOTHER' take is wrong. My gaming has moved almost entirely to my Rog Ally and you know what? Shit is just as fun and way more convenient than the 7700x/3080 12gb desktop even if it's 1080p low and not 1440p120. If the only thing the game has going for it is 'ooh it's pretty' then it's unlikely to be one of those games people care about in six months.
And anyways, who gives a crap about AAAAAAAAAAAAA games? Indie games are rocking it in every genre you could care to mention, and the higher budget stuff like BG 3 is, well, probably the best RPG since FO:NV (fight me!).
And yes, VR is in a shitty place because nobody gives a crap about it. I've got a Rift, Rift S, Quest, and a Quest 2 and you know what? It's not interesting. It's a fun toy that, but it has zero sticking power and that's frankly due to two things:
If you could justify spending the kind of money that would lead to having a cool VR experience, then yeah, it might be more compelling but that's been tried and nobody bought anything. Will say that Beat Saber is great, but one stellar experience will not sell anyone on anything.
And AI is this year's crypto which was last year's whatever and it's bubbles and VC scams all the way down and pretty much always has been. Tech hops from thing to thing that they go all in on because they can hype it and cash out. Good for them, and be skeptical of shit, but if it sticks it sticks, and if it doesn't it doesn't.
The 5080 is rumored to be 10% faster, but also use 90% the power. While performance has a normal generational leap, power consumption has gone up to match leaving you with a much smaller actual improvement.
Power consumption numbers like that are expected, though.
One thing to keep in mind is how big the die is and how many transistors are in a GPU.
As a direct-ish comparison, there's about 25 billion transistors in a 14900k, and 76 billion in a 4090.
Big die + lots and lots of transistors = bigly power usage.
I wouldn't imagine that the 5000-series GPUs are going to be smaller or have less transistors, so I'd expect this to be in the die shrink lowers power usage, but more transistors increase power usage zone.
You can also get big power consumption from turning up the voltage and cranking the clock speeds well past their efficient zone. You see that right now with most 40 series cards where turning the clock speeds down a smidge gives you huge power savings at almost no loss in performance.
Cost per MM^2 of die space has only gone up with each process node these last 10 years, so unless you're paying big money don't expect big chip.