Friday, October 11th 2024
NVIDIA Might Consider Major Design Shift for Future 300 GPU Series
NVIDIA is reportedly considering a significant design change for its GPU products, shifting from the current on-board solution to an independent GPU socket design following the GB200 shipment in Q4, according to reports from MoneyDJ and the Economic Daily News quoted by TrendForce. This move is not new in the industry, AMD has already introduced socket design in 2023 with their MI300A series via Supermicro dedicated servers. The B300 series, expected to become NVIDIA's mainstream product in the second half of 2025, is rumored to be the main beneficiary of this design change that could improve yield rates, though it may come with some performance trade-offs.
According to the Economic Daily News, the socket design will simplify after-sales service and server board maintenance, allowing users to replace or upgrade the GPUs quickly. The report also pointed out that based on the slot design, boards will contain up to four NVIDIA GPUs and a CPU, with each GPU having its dedicated slot. This will bring benefits for Taiwanese manufacturers like Foxconn and LOTES, who will supply different components and connectors. The move seems logical since with the current on-board design, once a GPU becomes faulty, the entire motherboard needs to be replaced, leading to significant downtime and high operational and maintenance costs.
Sources:
MoneyDJ, Economic Daily News, TrendForce
According to the Economic Daily News, the socket design will simplify after-sales service and server board maintenance, allowing users to replace or upgrade the GPUs quickly. The report also pointed out that based on the slot design, boards will contain up to four NVIDIA GPUs and a CPU, with each GPU having its dedicated slot. This will bring benefits for Taiwanese manufacturers like Foxconn and LOTES, who will supply different components and connectors. The move seems logical since with the current on-board design, once a GPU becomes faulty, the entire motherboard needs to be replaced, leading to significant downtime and high operational and maintenance costs.
28 Comments on NVIDIA Might Consider Major Design Shift for Future 300 GPU Series
Seriously? Why the hell hasn't GPU sockets been a thing? VDIMM modules? It's not like it would all that difficult..
VRM difference. Traces difference on the board for different memory buses and configs (and PCI-e lanes nowadays). Different sized chips requiring different cold plates on coolers since there is no unified IHS. Lots of reasons, really. They are POTENTIALLY solvable, but you would have to have a board and cooler combo that would be interchangeable between a hypothetical 4060 and 4090. This would be hilariously costly. Do we really want GPUs coating even more?
Cost is hard to gauge as you could have two 600W PSU's (one for CPU daughter board and one for GPU daughter board) which can be cheaper than one 1.2 kW power supply. The ability to upgrade just the VRAM and GPU itself could generate cost savings over time.
Either way, we are stuck in the socketed CPU, DIMM slot, multiple PCIe slots off of one motherboard era as it has been since the 80386 days (circa late 80s). Only the integration of I/O to the motherboard has really changed from those days requiring less expansion slots. Oh and I'm glad the era of jumpers is over. I hated those little plastic pieces of you know what. :)
Because it doesn’t work this way even with CPUs? Good luck cooling a 14900KS with a stock cooler. Same here - if you have a hypothetical 600W 5090 and standardize with THAT in mind you’d have to make it so that “any cooler” can reasonably keep THAT cool. Which is costly. I suppose the argument can be made for establishing a ceiling for power usage on GPUs and going from there, but it wouldn’t fly.
Essentially, socketed GPUs, be it for AIBs or directly onto MoBos, would require a complete overhaul of everything about the modern PC form-factor, as @Daven has mentioned. And it’s just not really feasible as it stands right now - ATX and its offspring has a profoundly ingrained effect on how every single PC part is created and functions. And it has been this way for decades now. Shifting the entire ecosystem just for some potential upgradeability benefits wouldn’t fly with any of the current players.
The same should apply to GPUs.
I mean, sure, in that you can theoretically put the 14900K into the shittiest board possible and slap an Intel stock cooler on top of it. You won’t be able to actually access the full potential of the chip, but I suppose by the metric of “it can be physically done” it works.
You also haven’t provided any actual argument other than “it should be like this just cause” which… okay. I will be sure to forward your feedback to whoever it may concern.
90's had their own chip(s) with each and every function. Dozens of examples, 3DFX for example. Chips and tech got advanced, now pretty much house everything inside of it, but Nvidia's approach right now is extremely inefficient. High failure rates on wafers is expensive. AMD on the other hand now makes the compute dies seperate from the memory cache dies - yielding much better on wafers and thus get more out of it.
Stop laughing pls!
Let's say you did make a socketed GPU. By the time you are ready to upgrade your GPU, guess what? The VRAM is going to need an upgrade too. Which means a new board. So you get that "upgradeability" and never use it, OR waste tons of $$$ upgrading inside a generation or every generation, like the people who bought a ryzen 1600, 2600, 3600, and 5600 for the same PC and spent twice what it would have cost to simply buy an i7 PC originally then replace it with a 5600 PC later.
how many people out there are salivating at the idea of putting a 9800x on a x370 board with 2400 mhz DDR4? not many, I'd reckon.
Speaking of AMD, there's another great reason. AMD controlled their chipsets and firmware, yet getting the ryzen 3000 and 5000 on the 300 series chipsets was a major clusterfuck. To date we STILL dont have full support. There are 400 series and 500 series chips that never got the X3d update and cannot run those chips. YAY, what a wonderful support nightmare! Intel couldnt pull it off either, remember the issues with PCIe 4.0 and rocket lake?
That's before getting into issues with clock scaling with current DIMM design, or the extra space it would require. And the cost, you'd have to either have multiple socket designs OR even the cheapest boards would have to support a 4090. So all the people whining about GPU costs now get to pay 4090 board price for 4060 GPUs! WOOOHHOOOOO!! :slap:
Most people dont upgrade every gen anyway. Much like CPUs, GPUs often go several generations before being upgraded. The primary market for Pascal was Fermi holdouts and first gen Kepler buyers, not those who owned Maxwell. And if you wait 5-6 years, well, you gotta upgrade everything anyway. The only difference between MXM and desktop GPUs is that one is mobile sized and has the video out circuit wired to the motherboard. It's not anymore modular then a 4090. But why? What does this gain you? You buy a 4060, then later buy a 4070 chip wasting a ton of cash, but WHOOPS you've choked your new 4070 with a smaller bus, time for a new mobo!
Or maybe you get a 5060, but WHOOPS, you've again choked it, this time with slower DDR6, and to fix that, you need a new GPU mobo! Just LOOk at the savings!
How many people do you know want to run an alder lake, or even coffee lake, chips on a ivy bridge boards? Or haswell?
The same reason we dont put 10+ CPU gens on one motherboard, GPU tech advances quickly, and you cant plug newer tech into an older board. You need updated traces and power components, so socketing the GPU only introduces a new point of failure with thermal expansions.
This is talking about their GB200 chip, which has TWO blackwell GPUs (with HBM memory built-in) along with a Grace CPU (with LPDDR5X soldered next to it).
The current device has that entire board as a single piece.
This article is likely proposing that the Grace CPU will be socketed instead of built in a single board alongside the GPUs.
They even compare it to the Instinct MI300A, which is an APU (with HBM on top of it) that can be socketed.
Memory won't be something apart, it'll still be soldered, you can't just pack LPDDR5X on a socket easily (I don't think Nvidia will be using those CAMM modules), nor can HBM.
Nvidia's GPU-only offerings are already socketed, either in SXM or PCIe. That rumor is NOT referring to those.
The G in GB200 refers to the Grace chip that lives alongisde the 2 Blackwell GPUs on the board.
EDIT: Also funny to see people comparing this to desktop products when those have nothing to do with one another lol
The products being discussed in the original OP also have nothing to do with your regular ATX desktop products, so there's no close link whatsoever. This is not about the SXM lineup, but rather their GB offerings that have a Grace CPU on the same board.
As for the economics, they're not only viable but also represent a serious oversight of missed opportunities and the potential for higher profits is extensive. You see, when a user buys a single card, then chooses a GPU and VRAM to go on it, the expense of the card is just an infrequent purchase. But if it can support multiple generations of GPUs and the VRAM to match, makers only have to make the GPU and the VDIMMs, cutting manufacturing expenses dramatically and reducing the environmental impacts of making whole cards as well.
So arguing economics is very narrow thinking, typical of the bean-counting sector who can barely see 5 paces in front of them. The vision of possibilities is what drives progress and socketed consumers GPUs is LONG over due. So much money wasted, so much missed opportunity..
on a serious note, the more you buy the more you save.. :roll:
You can scream and keep dreaming all you want, that idea is plainly bad in every sense and most people are aware of that anyway.
Others already gave many great examples of why this is a bad idea, but you do you, feel free to build your own modular GPU with blackjack and hookers lol