Friday, October 11th 2024

NVIDIA Might Consider Major Design Shift for Future 300 GPU Series

NVIDIA is reportedly considering a significant design change for its GPU products, shifting from the current on-board solution to an independent GPU socket design following the GB200 shipment in Q4, according to reports from MoneyDJ and the Economic Daily News quoted by TrendForce. This move is not new in the industry, AMD has already introduced socket design in 2023 with their MI300A series via Supermicro dedicated servers. The B300 series, expected to become NVIDIA's mainstream product in the second half of 2025, is rumored to be the main beneficiary of this design change that could improve yield rates, though it may come with some performance trade-offs.

According to the Economic Daily News, the socket design will simplify after-sales service and server board maintenance, allowing users to replace or upgrade the GPUs quickly. The report also pointed out that based on the slot design, boards will contain up to four NVIDIA GPUs and a CPU, with each GPU having its dedicated slot. This will bring benefits for Taiwanese manufacturers like Foxconn and LOTES, who will supply different components and connectors. The move seems logical since with the current on-board design, once a GPU becomes faulty, the entire motherboard needs to be replaced, leading to significant downtime and high operational and maintenance costs.
Sources: MoneyDJ, Economic Daily News, TrendForce
Add your own comment

28 Comments on NVIDIA Might Consider Major Design Shift for Future 300 GPU Series

#1
SL2
Waiting for a comment asking why this isn't a thing on consumer cards.
Posted on Reply
#2
lexluthermiester
SL2Waiting for a comment asking why this isn't a thing on consumer cards.
I'll fall on that grenade...

Seriously? Why the hell hasn't GPU sockets been a thing? VDIMM modules? It's not like it would all that difficult..
Posted on Reply
#3
Onasi
@lexluthermiester
VRM difference. Traces difference on the board for different memory buses and configs (and PCI-e lanes nowadays). Different sized chips requiring different cold plates on coolers since there is no unified IHS. Lots of reasons, really. They are POTENTIALLY solvable, but you would have to have a board and cooler combo that would be interchangeable between a hypothetical 4060 and 4090. This would be hilariously costly. Do we really want GPUs coating even more?
Posted on Reply
#4
Philaphlous
lexluthermiesterI'll fall on that grenade...

Seriously? Why the hell hasn't GPU sockets been a thing? VDIMM modules? It's not like it would all that difficult..
Money? Nvidia selling directly with a GPU will grant far less money than a full board and chip from a 3rd party... Plus then NVIDIA gets to control the wholesale GPU chip price and 3rd party gets to set the margins..win win for the middleman...
Posted on Reply
#5
Daven
Onasi@lexluthermiester
VRM difference. Traces difference on the board for different memory buses and configs (and PCI-e lanes nowadays). Different sized chips requiring different cold plates on coolers since there is no unified IHS. Lots of reasons, really. They are POTENTIALLY solvable, but you would have to have a board and cooler combo that would be interchangeable between a hypothetical 4060 and 4090. This would be hilariously costly. Do we really want GPUs coating even more?
Traditional PC architectures haven't been significantly revamped since the beginning of consumer client desktops. While you are correct, that just adding another socket to existing motherboards is probably not the answer, we can create a two daughter board standard connection through x16 PCIe Gen 5 interlink. Each daughter board can have up its own socket, memory slots, I/O connections and even separate power supplies/delivery.

Cost is hard to gauge as you could have two 600W PSU's (one for CPU daughter board and one for GPU daughter board) which can be cheaper than one 1.2 kW power supply. The ability to upgrade just the VRAM and GPU itself could generate cost savings over time.

Either way, we are stuck in the socketed CPU, DIMM slot, multiple PCIe slots off of one motherboard era as it has been since the 80386 days (circa late 80s). Only the integration of I/O to the motherboard has really changed from those days requiring less expansion slots. Oh and I'm glad the era of jumpers is over. I hated those little plastic pieces of you know what. :)
Posted on Reply
#6
Caring1
Onasi@lexluthermiester
VRM difference. Traces difference on the board for different memory buses and configs (and PCI-e lanes nowadays). Different sized chips requiring different cold plates on coolers since there is no unified IHS. Lots of reasons, really. They are POTENTIALLY solvable, but you would have to have a board and cooler combo that would be interchangeable between a hypothetical 4060 and 4090. This would be hilariously costly. Do we really want GPUs coating even more?
Why not make the GPU socket a standard and all GPUs have to comply so one cooler would suffice for any unit used?
Posted on Reply
#7
Onasi
@Caring1
Because it doesn’t work this way even with CPUs? Good luck cooling a 14900KS with a stock cooler. Same here - if you have a hypothetical 600W 5090 and standardize with THAT in mind you’d have to make it so that “any cooler” can reasonably keep THAT cool. Which is costly. I suppose the argument can be made for establishing a ceiling for power usage on GPUs and going from there, but it wouldn’t fly.

Essentially, socketed GPUs, be it for AIBs or directly onto MoBos, would require a complete overhaul of everything about the modern PC form-factor, as @Daven has mentioned. And it’s just not really feasible as it stands right now - ATX and its offspring has a profoundly ingrained effect on how every single PC part is created and functions. And it has been this way for decades now. Shifting the entire ecosystem just for some potential upgradeability benefits wouldn’t fly with any of the current players.
Posted on Reply
#8
Caring1
Onasi@Caring1
Because it doesn’t work this way even with CPUs? Good luck cooling a 14900KS with a stock cooler. Same here - if you have a hypothetical 600W 5090 and standardize with THAT in mind you’d have to make it so that “any cooler” can reasonably keep THAT cool. Which is costly. I suppose the argument can be made for establishing a ceiling for power usage on GPUs and going from there, but it wouldn’t fly.

Essentially, socketed GPUs, be it for AIBs or directly onto MoBos would require a complete overhaul of everything about the modern PC form-factor, as @Daven has mentioned. And it’s just not really feasible as it stands right now - ATX and its offspring has a profoundly ingrained effect on how every single PC part is created and functions. And it has been this way for decades now. Shifting the entire ecosystem just for some potential upgradeability benefits wouldn’t fly with any of the current players.
You can use the same socket for anything from an i3 to an i9, the cooler can be whatever you want it to be.
The same should apply to GPUs.
Posted on Reply
#9
Vya Domus
Yeah not happening, part of Nvidia's strategy is making upgrading the hardware as expensive as possible because customers don't have a choice.
Posted on Reply
#10
Onasi
@Caring1
I mean, sure, in that you can theoretically put the 14900K into the shittiest board possible and slap an Intel stock cooler on top of it. You won’t be able to actually access the full potential of the chip, but I suppose by the metric of “it can be physically done” it works.
You also haven’t provided any actual argument other than “it should be like this just cause” which… okay. I will be sure to forward your feedback to whoever it may concern.
Posted on Reply
#11
lexluthermiester
I'm not going to respond to anyone separately, instead offering the following thoughts to all: Anything we can do with CPUs, we can do just as easily with GPUs. It's a design shift and spec that should have happened decades ago. Build a GPU daughter-card that has a socket and VDIMMs, have it plug into the motherboard with power circuitry to support whatever you want to install, and you heatsink of choice and Bob's your uncle. I mean after all, we have the MXM standard, and that's fairly close.
Posted on Reply
#12
Jism
SL2Waiting for a comment asking why this isn't a thing on consumer cards.
It was.

90's had their own chip(s) with each and every function. Dozens of examples, 3DFX for example. Chips and tech got advanced, now pretty much house everything inside of it, but Nvidia's approach right now is extremely inefficient. High failure rates on wafers is expensive. AMD on the other hand now makes the compute dies seperate from the memory cache dies - yielding much better on wafers and thus get more out of it.
Posted on Reply
#13
SL2
lexluthermiesterI'm not going to respond to anyone separately, instead offering the following thought to all: Anything we can do with CPUs, we can do just as easily with GPUs.
As if anyone questioned the hypothetical possibility to do it.
Posted on Reply
#14
OSdevr
Surprised this is possible given the speed of modern VRAM, unless they're thinking everything will use HBM. Not sure how practical GPU sockets would be on consumer GPUs with separate VRAM for signal integrity reasons, that was ostensibly the reason for ditching upgradable VRAM so long ago (it used to be a thing).
Posted on Reply
#15
Vayra86
Ah, improved yields! Cheaper GPUs!

Stop laughing pls!
Posted on Reply
#16
TheinsanegamerN
JismIt was.

90's had their own chip(s) with each and every function. Dozens of examples, 3DFX for example. Chips and tech got advanced, now pretty much house everything inside of it, but Nvidia's approach right now is extremely inefficient. High failure rates on wafers is expensive. AMD on the other hand now makes the compute dies seperate from the memory cache dies - yielding much better on wafers and thus get more out of it.
Yeah, but wafer yields and centralization of components has nothing to do with why GPus are not socketed.
lexluthermiesterI'm not going to respond to anyone separately, instead offering the following thought to all: Anything we can do with CPUs, we can do just as easily with GPUs. It's a design shift and spec that should have happened decades ago. Build a GPU daughter-card that has a socket and VDIMMs, have it plug into the motherboard with power circuitry to support whatever you want to install.
But why? To what end?

Let's say you did make a socketed GPU. By the time you are ready to upgrade your GPU, guess what? The VRAM is going to need an upgrade too. Which means a new board. So you get that "upgradeability" and never use it, OR waste tons of $$$ upgrading inside a generation or every generation, like the people who bought a ryzen 1600, 2600, 3600, and 5600 for the same PC and spent twice what it would have cost to simply buy an i7 PC originally then replace it with a 5600 PC later.

how many people out there are salivating at the idea of putting a 9800x on a x370 board with 2400 mhz DDR4? not many, I'd reckon.

Speaking of AMD, there's another great reason. AMD controlled their chipsets and firmware, yet getting the ryzen 3000 and 5000 on the 300 series chipsets was a major clusterfuck. To date we STILL dont have full support. There are 400 series and 500 series chips that never got the X3d update and cannot run those chips. YAY, what a wonderful support nightmare! Intel couldnt pull it off either, remember the issues with PCIe 4.0 and rocket lake?

That's before getting into issues with clock scaling with current DIMM design, or the extra space it would require. And the cost, you'd have to either have multiple socket designs OR even the cheapest boards would have to support a 4090. So all the people whining about GPU costs now get to pay 4090 board price for 4060 GPUs! WOOOHHOOOOO!! :slap:

Most people dont upgrade every gen anyway. Much like CPUs, GPUs often go several generations before being upgraded. The primary market for Pascal was Fermi holdouts and first gen Kepler buyers, not those who owned Maxwell. And if you wait 5-6 years, well, you gotta upgrade everything anyway.
lexluthermiesterI mean after all, we have the MXM standard, and that's fairly close.
The only difference between MXM and desktop GPUs is that one is mobile sized and has the video out circuit wired to the motherboard. It's not anymore modular then a 4090.
Caring1You can use the same socket for anything from an i3 to an i9, the cooler can be whatever you want it to be.
The same should apply to GPUs.
But why? What does this gain you? You buy a 4060, then later buy a 4070 chip wasting a ton of cash, but WHOOPS you've choked your new 4070 with a smaller bus, time for a new mobo!

Or maybe you get a 5060, but WHOOPS, you've again choked it, this time with slower DDR6, and to fix that, you need a new GPU mobo! Just LOOk at the savings!

How many people do you know want to run an alder lake, or even coffee lake, chips on a ivy bridge boards? Or haswell?

The same reason we dont put 10+ CPU gens on one motherboard, GPU tech advances quickly, and you cant plug newer tech into an older board. You need updated traces and power components, so socketing the GPU only introduces a new point of failure with thermal expansions.
Posted on Reply
#17
igormp
Uhhhh, is everyone really missing the point, or did I really missread the rumor???

This is talking about their GB200 chip, which has TWO blackwell GPUs (with HBM memory built-in) along with a Grace CPU (with LPDDR5X soldered next to it).
The current device has that entire board as a single piece.

This article is likely proposing that the Grace CPU will be socketed instead of built in a single board alongside the GPUs.

They even compare it to the Instinct MI300A, which is an APU (with HBM on top of it) that can be socketed.
Memory won't be something apart, it'll still be soldered, you can't just pack LPDDR5X on a socket easily (I don't think Nvidia will be using those CAMM modules), nor can HBM.

Nvidia's GPU-only offerings are already socketed, either in SXM or PCIe. That rumor is NOT referring to those.
The G in GB200 refers to the Grace chip that lives alongisde the 2 Blackwell GPUs on the board.

EDIT: Also funny to see people comparing this to desktop products when those have nothing to do with one another lol
Posted on Reply
#18
SL2
igormpEDIT: Also funny to see people comparing this to desktop products when those have nothing to do with one another lol
I don't see much comparing going on, just a few fantasies
Posted on Reply
#19
kapone32
Back in the day some what we call GPUs came with upgrade able memory slots.
Posted on Reply
#20
lexluthermiester
TheinsanegamerNBut why? To what end?
If you have to ask those questions, you either haven't thought the matter through or my answer would be nothing to you. Either, ask yourself why and let the gears in your head turn for a while.
SL2I don't see much comparing going on, just a few fantasies
Well grounded fantasies.. Practical and logical.. and let's face facts, the enterprise and consumer sectors are for more closely linked than most people realize.
Posted on Reply
#21
Solaris17
Super Dainty Moderator
I wonder if they just mean a smaller package/interposer. These GPUs are already socketed. Maybe they mean going LGA instead of going SXM.
Posted on Reply
#22
igormp
lexluthermiesterWell grounded fantasies.. Practical and logical.. and let's face facts, the enterprise and consumer sectors are for more closely linked than most people realize.
No, they're just wishful thinking. Doable in theory? Sure. Do they make any economical sense, both from the consumer and manufacturers perspective? No.
The products being discussed in the original OP also have nothing to do with your regular ATX desktop products, so there's no close link whatsoever.
Solaris17I wonder if they just mean a smaller package/interposer. These GPUs are already socketed. Maybe they mean going LGA instead of going SXM.
This is not about the SXM lineup, but rather their GB offerings that have a Grace CPU on the same board.
Posted on Reply
#23
lexluthermiester
igormpNo, they're just wishful thinking.
To YOU.
igormpDoable in theory? Sure.
No, doable in reality.
igormpDo they make any economical sense, both from the consumer and manufacturers perspective? No.
I don't give a flying rats-butt about notions like that. Flexiblity and choice is what I'm suggesting. A socketed GPU daughter board would offer the same flexibility and choice and a CPU motherboard. Choice of multiple GPU models, various choices of VRAM and various choices of cooling.

As for the economics, they're not only viable but also represent a serious oversight of missed opportunities and the potential for higher profits is extensive. You see, when a user buys a single card, then chooses a GPU and VRAM to go on it, the expense of the card is just an infrequent purchase. But if it can support multiple generations of GPUs and the VRAM to match, makers only have to make the GPU and the VDIMMs, cutting manufacturing expenses dramatically and reducing the environmental impacts of making whole cards as well.

So arguing economics is very narrow thinking, typical of the bean-counting sector who can barely see 5 paces in front of them. The vision of possibilities is what drives progress and socketed consumers GPUs is LONG over due. So much money wasted, so much missed opportunity..
Posted on Reply
#24
RootinTootinPootin
lexluthermiestera flying rats-butt
do they really fly?

on a serious note, the more you buy the more you save.. :roll:
Posted on Reply
#25
igormp
lexluthermiesterTo YOU.
Sure, all the market is drooling over such product, companies are just ignoring this pile of money freely available to them :laugh:
lexluthermiesterNo, doable in reality.
lexluthermiesterI don't give a flying rats-butt about notions like that.
Do you know that when someone speaks about the real life, economics are a really important topic? If it's not doable in the economic sense, it's not doable in reality.
lexluthermiesterChoice of multiple GPU models, various choices of VRAM and various choices of cooling.
Aaaand shitty performance for a way higher cost and higher power consumption. I personally wouldn't touch such thing at all.
lexluthermiesterThe vision of possibilities is what drives progress and socketed consumers GPUs is LONG over due. So much money wasted, so much missed opportunity..
If you haven't noticed, most of the current computing hardware is trending more and more into the tightly integrated space for better performance and higher efficiency.

You can scream and keep dreaming all you want, that idea is plainly bad in every sense and most people are aware of that anyway.
Others already gave many great examples of why this is a bad idea, but you do you, feel free to build your own modular GPU with blackjack and hookers lol
Posted on Reply
Add your own comment
Dec 11th, 2024 20:28 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts