PC graphics cards to get more expensive again "thanks" to AI boom
PC graphics cards to get more expensive again "thanks" to AI boom
I do have to say that I’m skeptical of this article. It smells a lot like GPU corporation shareholder drumbeating…
I can find every RTX 4000 SKU on Amazon, ready to ship, at (and usually below) MSRP…
It makes sense to me; AI needs GPUs, and there’s an AI boom.
Probably won’t be as bad as the crypto situation, but I imagine it will have an effect on GPU availability at some point.
People are skeptical about AI, but companies seem to be less so.
They see an opportunity to save money and they’re gonna go for it, with the ship steered by shareholders who like the “AI” buzzword.
Maybe it won’t have any effect on GPU prices though. Crypto was a much more accessible market for anyone who wanted to have a go at it, whereas you do need to actually have an idea to make use of AI.
Ai cards need to be efficient so they need tsmc 3n and 5n fabs. Desktop cards don’t so they could use Samsung or older cheap tsmc fabs. When we had the shortages before it was the smaller components like vrm stages and capacitors that had a shortage. Those are now over supplied. There is no reason for the price hikes other than Nvidia seeing what people paid to scalpers and wanting it for themselves.
Nvidia has continued to push for more clock speed on lower end parts and charging higher prices. There is no reason a mid range die should be clocked to 3ghz on a super expensive pcb like the 4080 has or a low end part doing it on the 4070. Those both have pcb that cost more than the die for a die that would historically be used in $150-400 parts when adjusted for inflation. They also both should be clocked around 1.8-2ghz as that has a 60-70% reduction their power consumption for a 30% performance loss (see the mobile parts for what those parts should be close to for their base sku.)
Data centers care a lot about power. The ai products run around 2ghz in the sweet spot. Consumer cards target 3ghz this gen and use 3-4x the power that they do at ~2ghz. The die in the 4080 is a mid range size. It is what used to be in things like the 60 series or maybe a 70 series card. They have been overclocking the snot out of them stock and putting them on massively expensive pcb instead of giving us the larger dies we used to get. That shifts the costs to the board partners and lets them get away with selling the dies at a huge profit compared to their older products.
Back to data centers. You pay a lot for your spot based on power and location. If they stay efficient and pack lots of chips in, that is the cheapest way over the life of the server. If you save 10 or 20% power due to using a new node that is worth a huge reduction in data center fees. On the consumer desktop side, they can overclock to double the power instead of using a larger more expensive die and pocket the difference with no one really caring.
Remember when PC gaming used to be hands down better than console?
Anyone who still believes that is stuck in the past.
PC certainly has its benefits and is my platform of choice, but if somebody was getting into gaming or just casually interested then a PS5 is a much better choice these days.
a PS5 is a much better choice these days.
if you upgrade your PC to “PS5 level of graphic detail” your whole library of games (starting from the original Doom/Quake in ray tracing, upscaled texture pack etc.) get upgraded.
If you buy a PS5… then only the games you buy specifically for the PS5 will run at PS5 graphics.
To this add that, on PC, you can buy second hand GPU that matches PS5 fidelity (both Nvidia or AMD)
The point is that even a second hand high-end card will cost as much as a PS5. For a single component.
PC gaming has grown a lot, but its peak was the PS3/PS4 era when it could be noticably superior to average people. I'd say it's going back to being a platform for entheusiasts who care about those things you mentioned, which most people don't. At least, not enough to spend twice as much or more for a console equivilent.
But aren’t the GPUs used by AI different than the GPUs used by gamers? 8GB of RAM isn’t enough to run even the smaller LLMs, you need specialized GPUs with 80+GB like A100s and H100s.
Right now GPU prices aren’t extremely low, but you can actually but them from retailers at market price. That wasn’t the case when crypto-mining was popular
They're not that different, really. CUDA processing cores are the most used in AI training, and those are the main processors used in both Nvidia's consumer desktop cards and machine learning enterprise cards. As "AI" is on the rise, more and more of the supply of CUDA processors and VRAM chips will be diverted to enterprise solutions that will fetch a higher price from deals with corporations. Meaning there will be less materials available for the consumer-level GPU supply, which will drive prices up for normal consumers. NVIDIA has been banking on this for a long time; that's why they don't care about overpricing the consumer market and have been trying to push people towards cloud-based GeForce Now subscription models where you don't even own the hardware and just basically rent the processing power to play games.
Also just to be anal, the 3090 and 4090 have 24Gb of vram, not 32Gb. And unlike gaming nowadays you can distribute the workload to multiple GPU's in one system, or over a network of machines.