Tuesday, March 19th 2024

NVIDIA "Blackwell" GeForce RTX to Feature Same 5nm-based TSMC 4N Foundry Node as GB100 AI GPU

Following Monday's blockbuster announcements of the "Blackwell" architecture and NVIDIA's B100, B200, and GB200 AI GPUs, all eyes are now on its client graphics derivatives, or the GeForce RTX GPUs that implement "Blackwell" as a graphics architecture. Leading the effort will be the new GB202 ASIC, a successor to the AD102 powering the current RTX 4090. This will be NVIDIA's biggest GPU with raster graphics and ray tracing capabilities. The GB202 is rumored to be followed by the GB203 in the premium segment, the GB205 a notch lower, and the GB206 further down the stack. Kopite7kimi, a reliable source with NVIDIA leaks, says that the GB202 silicon will be built on the same TSMC 4N foundry node as the GB100.

TSMC 4N is a derivative of the company's mainline N4P node, the "N" in 4N stands for NVIDIA. This is a nodelet that TSMC designed with optimization for NVIDIA SoCs. TSMC still considers the 4N as a derivative of the 5 nm EUV node. There is very little public information on the power- and transistor density improvements of the TSMC 4N over TSMC N5. For reference, the N4P, which TSMC regards as a 5 nm derivative, offers a 6% transistor-density improvement, and a 22% power efficiency improvement. In related news, Kopite7kimi says that with "Blackwell," NVIDIA is focusing on enlarging the L1 caches of the streaming multiprocessors (SM), which suggests a design focus on increasing the performance at an SM-level.
Sources: Kopite7kimi (Twitter), #2, VideoCardz
Add your own comment

60 Comments on NVIDIA "Blackwell" GeForce RTX to Feature Same 5nm-based TSMC 4N Foundry Node as GB100 AI GPU

#26
gffermari
If they manage to extract 20-25% more performance using the same node, would be a success.
Pricewise they will be cheaper than Ada according to tick tock policy (cheap Pascal, expensive Turing, cheap Ampere, expensive Ada....)

What I don't want to see is bullshXt software tweaks, just for the charts.
Ok, they may introduce a DLSS 4+, available only on 5000, but I don't want to see a performance jump because of this only.
Posted on Reply
#27
evernessince
OnasiFor their GPUs? Highly unlikely.
Hmm, depends. AMD's GCD die is 304mm2, which isn't huge. When you factor the node shrink in you are talking about a lot of performance on a pretty manufacturable die size. One of the advantages to the chiplet based approach is that each individual die is smaller and thus will yield better on nodes with higher defect rates (newer nodes). The only reason GPUs tend to use more mature nodes is due to their size but a chiplet based architecture may enable them to have an advantage over Nvidia in regards to node.
Posted on Reply
#28
AnotherReader
They have two levers to increase performance with a bigger die:
  1. Increase power limit to just a hair under 600 W so that clocks don't decrease
  2. keep power limit the same as the 4090, but increase die size to 2080 Ti proportions. This would allow more cores, perhaps up to the 192 in the rumours
Another option is to use Chip-On-Wafer-On-Substrate-L (CoWoS-L) which is used for GB200, but that seems unlikely for a gaming GPU.
Posted on Reply
#29
N/A
OC models could definitely benefit from 600W power limit. The same case of 980 Ti and 2080 Ti +25% performance by OC alone, both second gen on same node. CUDA cores can be increased at a very little die area cost 1536 to 2048 in each GPC for example 750mm2. But if kopite still fantasizes about 512 bit memory GB202 is definitely a maxed out reticle sized GPU.
Posted on Reply
#30
phints
Wait isn't TSMC 4N what Lovelace is on? If Blackwell is on this too it'll be a very weak new generation. Probably large increases in TGP as you can only do so much with layout optimization and GDDR7. It's too early to tell but I might be waiting for a RTX 6000 now.
Posted on Reply
#31
Onasi
evernessinceHmm, depends. AMD's GCD die is 304mm2, which isn't huge. When you factor the node shrink in you are talking about a lot of performance on a pretty manufacturable die size. One of the advantages to the chiplet based approach is that each individual die is smaller and thus will yield better on nodes with higher defect rates (newer nodes). The only reason GPUs tend to use more mature nodes is due to their size but a chiplet based architecture may enable them to have an advantage over Nvidia in regards to node.
That’s absolutely true, but that’s not why I was skeptical. I just have strong doubts that any 3nm allocation will be left this year for AMD after Apple places the orders it needs.
Posted on Reply
#32
Unregistered
dgianstefaniPrices went down with Supers so there's hope for at least price/performance to not just continue scaling. I guess we'll see.

$1000 5080 would be nice, and possible, unlike the $600 5080 some people seem to expect/want.
The 3080 was $700 MSRP, that's ~$836 today - Jensen shifted the goal post with the 4080 big time asking $1200.
Four figures for an 80-series card is still too high IMO, you shouldn't break that barrier until the 90-series.
Posted on Edit | Reply
#33
evernessince
OnasiThat’s absolutely true, but that’s not why I was skeptical. I just have strong doubts that any 3nm allocation will be left this year for AMD after Apple places the orders it needs.
That's a valid concern although the way the industry works is that silicon design companies like AMD sign contracts years in advance guarantying a certain amount of allocation at a specified rate. This ensures that years spent designing a product aren't foiled by market demand swings or competitors and enables foundries time to build production to meet demand. It wouldn't make much sense for TSMC or their design parters if allocation was determined dynamically as the cost of designing a product is extremely high and time investment of silicon products and fabs takes years. No company is going to want to invest years and billions in designing a chip they cannot gaurantee wafers for.

In otherwords AMD would have to specifically forget to include 3nm capacity for it's GPUs in order for them to not get any allocation (assuming they are going 3nm of course), which would be the highest level of incompetence possible given securing wafers is essential to operating a silicon design business. A company like AMD does not simply forget to secure wafers for it's products and continue to ignore that deficiency over the multi-year design period of a product.

What could be a problem for AMD though is if they didn't purchase enough allocation in advance, as in their sales exceed their expectations. In that instance they would be competiting with other companies like Apple for any unearmarked capacity TSMC has. There's also the possibility that TSMC would not be able to meet AMDs required 3nm wafer allotment as a reason AMD could not go 3nm, although I doubt this potentialitiy given TSMC's capacity has increased while simulatniously demand has decreased.
Posted on Reply
#34
Onasi
@evernessince
That’s how it usually works, yes, but Apple is a special case in terms of their relationship with TSMC. They are a VVIP customer and they get first dibs on any new node, no ifs or buts. Everyone else has to fight for scraps and if there ARE no scraps… tough. The fact that NV didn’t manage to get any or elected not to speaks volumes to me. Even if AMD does get some allocation, there is absolutely no way they would have decided to spend it on the GPU part of the business and not what actually makes them money - CPUs and/or CDNA HPC accelerators. As such, while it’s theoretically possible that we will see a case of NV being on an older node while new AMD GPUs get their chiplets made on 3nm, I just don’t see it. Especially since in the Ada vs RDNA 3 it was AMD who used an older (if very marginally) node for their GPUs.
Posted on Reply
#35
evernessince
Onasi@evernessince
That’s how it usually works, yes, but Apple is a special case in terms of their relationship with TSMC. They are a VVIP customer and they get first dibs on any new node, no ifs or buts. Everyone else has to fight for scraps and if there ARE no scraps… tough. The fact that NV didn’t manage to get any or elected not to speaks volumes to me. Even if AMD does get some allocation, there is absolutely no way they would have decided to spend it on the GPU part of the business and not what actually makes them money - CPUs and/or CDNA HPC accelerators. As such, while it’s theoretically possible that we will see a case of NV being on an older node while new AMD GPUs get their chiplets made on 3nm, I just don’t see it. Especially since in the Ada vs RDNA 3 it was AMD who used an older (if very marginally) node for their GPUs.
I wouldn't say the GPU segment makes them no money, the MI300 and MI300X are both big margin parts and include GCDs. This is particularly important if AMD wants to break into the AI market and just so happens to carry huge margins for them. It's also counter-intuitive to approach the market like "well we don't earn much from it now so it's not worth investing in or pushing for 3nm". That's a mentallity that begets loosing. AMD has massive revenue earning potential in both the AI and consumer graphics markets.

In addition, a good part of AMD's investment into GPUs is shared with their CPUs. Any improvements to infinity fabric and additional modularization that come with GPU chiplets undoubtedly help advance the packaging of their CPUs as well. Even if AMD only performs so-so in AI / Gaming GPUs, that technical expertise and knowledge is invaluable to the company as a whole. The MI300 and MI300X are excellent examples of that.

At the end of the day, regardless of how Vs Apple has in addition to being a VIP, AMD and TSMC would have known years in advance wafer allotments. If TSMC told AMD straight up years back that they couldn't fill a theoretical 3nm allotment for GPUs, that would be a failure on TSMCs part for sure. Again Apple can eat up all the uncontracted capacity but a lack of an ability for TSMC to build out capacity? I'm not seeing it given the lowered demand and massive investments by TSMC to do just that, build out capacity.

Again this all assumes that AMD wanted to go with 3nm for it's GPUs but I tink if they wanted to they could.
Posted on Reply
#36
AnotherReader
Onasi@evernessince
That’s how it usually works, yes, but Apple is a special case in terms of their relationship with TSMC. They are a VVIP customer and they get first dibs on any new node, no ifs or buts. Everyone else has to fight for scraps and if there ARE no scraps… tough. The fact that NV didn’t manage to get any or elected not to speaks volumes to me. Even if AMD does get some allocation, there is absolutely no way they would have decided to spend it on the GPU part of the business and not what actually makes them money - CPUs and/or CDNA HPC accelerators. As such, while it’s theoretically possible that we will see a case of NV being on an older node while new AMD GPUs get their chiplets made on 3nm, I just don’t see it. Especially since in the Ada vs RDNA 3 it was AMD who used an older (if very marginally) node for their GPUs.
I suspect N3 yields are still too low for anyone but Apple to bother with it, but we'll know by the time Zen 5c rolls out as that's rumoured to be on N3. If that rumour turns out to be true, then that would lend more weight to the hypothesis that N3 isn't mature enough for Nvidia's giant datacenter dies.
Posted on Reply
#37
Onasi
evernessinceI wouldn't say the GPU segment makes them no money, the MI300 and MI300X are both big margin parts and include GCDs. This is particularly important if AMD wants to break into the AI market and just so happens to carry huge margins for them. It's also counter-intuitive to approach the market like "well we don't earn much from it now so it's not worth investing in or pushing for 3nm". That's a mentallity that begets loosing. AMD has massive revenue earning potential in both the AI and consumer graphics markets.
I specifically mentioned CDNA accelerators in my post, I am not sure what you are disagreeing with. I remind you that AMDs HPC cards run explicitly different architecture to their gaming offerings. Yes, even the GCDs are different. They would not be in the same equation for AMD.
evernessinceAt the end of the day, regardless of how Vs Apple has in addition to being a VIP, AMD and TSMC would have known years in advance wafer allotments. If TSMC told AMD straight up years back that they couldn't fill a theoretical 3nm allotment for GPUs, that would be a failure on TSMCs part for sure. Again Apple can eat up all the uncontracted capacity but a lack of an ability for TSMC to build out capacity? I'm not seeing it given the lowered demand and massive investments by TSMC to do just that, build out capacity.
I am actually not sure about this. TSMC had yield issues with 3N up until last summer and I have no idea if they still are having them. There might not have been any offers to anyone except Apple at all.
AnotherReaderI suspect N3 yields are still too low for anyone but Apple to bother with it, but we'll know by the time Zen 5c rolls out as that's rumoured to be on N3. If that rumour turns out to be true, then that would lend more weight to the hypothesis that N3 isn't mature enough for Nvidia's giant datacenter dies.
That, and historically early node iterations aren’t good for anything that’s not fairly low-power. This is fine for Apple with their SoCs. This is absolutely not fine for GPUs.
Posted on Reply
#38
Broken Processor
Man the doom and gloom on this thread is hilarious. Hopefully the product won't be bad and if it is don't buy it.
Posted on Reply
#39
Vayra86
SlizzoGuys, node wars don't really matter anymore. It's the performance they can extract out of the nodes they're using.

On that point, I have no doubt NVIDIA will be able to extract enough performance from the node.
This was always the case. There are no node wars. Old node, you can make a bigger GPU at better yields and put it in the market cheaper. New node, you make a smaller die because its f'ing expensive. If you can keep some semblance of performance parity you're good.

There are however, bad nodes; like Samsung's 8nm. As a result of that, while Ampere wasn't bad in performance, it was absolutely stellar in TDP. Not in a good sense.
Posted on Reply
#40
N/A
Broken ProcessorMan the doom and gloom on this thread is hilarious. Hopefully the product won't be bad and if it is don't buy it.
Kopite said back in November TSMC N3, now this, let's see if the 512 bit bus holds true. But he flip flopped on that too already maybe twice. the future is written in the end remains for events to play out. i don't mind as long as there is a good performing card at $799.
Posted on Reply
#41
AnotherReader
Vayra86This was always the case. There are no node wars. Old node, you can make a bigger GPU at better yields and put it in the market cheaper. New node, you make a smaller die because its f'ing expensive. If you can keep some semblance of performance parity you're good.

There are however, bad nodes; like Samsung's 8nm. As a result of that, while Ampere wasn't bad in performance, it was absolutely stellar in TDP. Not in a good sense.
I'm surprised by how many of our forum members think that process nodes are unimportant and can be worked around. There's no substitute for smaller, faster, and lower power transistors. None of Nvidia, AMD or Apple would be where they are now without continuously improving nodes. As far as the flatlining cost per transistor is concerned, that is a new phenomenon. For most of the history of making microprocessors, newer processors have also brought cost reductions.
Posted on Reply
#42
evernessince
OnasiI am actually not sure about this. TSMC had yield issues with 3N up until last summer and I have no idea if they still are having them. There might not have been any offers to anyone except Apple at all.
Apple was not the only one provided access given MediaTek has already publicly announced a 3nm chip slated for sometime in 2024: www.neowin.net/news/mediatek-develops-its-first-3nm-chip-using-tsmc-process-coming-in-2024/

TSMC is targeting 80% yield for 3nm: www.androidheadlines.com/2024/02/tsmc-double-3nm-production-2024.html

They are also looking at doubling capacity with Qualcomm, MediaTek, and others having 3nm chips in the pipe.

A GPU in late 2024 would not be infeasible at an 80% yield rate, particularly when we are talking about a chiplet based GPU where the individual dies are smaller and thus work better on lower yield nodes. There are no official die size numbers for the M3 Max, Apple's largest 3nm product, but it has 3.7 times the transistors of the base M3. The base M3 of which has a approximate die size of 150 - 170mm2. Based on the lowest estimate, a rough guess of the die size would be 555mm2. Even if that estimately is significantly off, it's definitely possible to see that if Apple is able to get high enough yield to make the M3 Max possible, it should definitely be possible to get a 304mm2 GPU die working. I'm also thinking it's possible that the rumors that AMD is ceding the high end could potentially be people confusing a smaller GCD size with them giving up the high end. It could be that AMD lowers the GCD size to even smaller than 304mm2 to further increase yields and then simply adds multiple to it's higher end products. AMD could just keep the MCDs on an older node given those don't really benefit much, although they are extremely tiny so they would yield well on newer nodes. Using different nodes though would allow them to leverage more capacity.

I don't know the probability that AMD goes 3nm, without access to both TSMC's numbers and AMD's numbers it's very hard to say.
Posted on Reply
#43
Onasi
@evernessince
We have no idea what the yields are like on Apple silicon, true, but Apple did announce that they are buying50% more 3nm capacity in 2024. I have no idea what this tells us about yields (do they expect more demand or are the yields still poor even for what demand is there), but it means that a lot of extra capacity they will gobble up. Then there is Intel who also already booked quite a bit, apparently. MediaTek and Qualcomm are in line, I guess that they will get what is left. Again, the fact that after all the rumors and leaks it turns out that NV was unable or unwilling to get some for themselves is, IMO, telling. It’s obviously not a question of money, not for NV. Whatever is the reason - lack of capacity or current 3nm being unsuitable for their needs is a separate question. I do want to note that even the M3 Max is a fairly low power chip. It very well might be that, until N3E is off the ground, producing high wattage parts which desktop GPUs absolutely are is just off the table. If AMD can design the chiplets in a way that would make them suitable they MIGHT have a chance to use 3nm, but those parts will be quite limited in how capable they would be, I think. So far all the AMD focused 3nm rumors were about Zen5c, another low power part. I think that we will know more in approximately May when it’s speculated that Zen 5 proper will be revealed. If regular Zen 5 isn’t on 3nm (or at least initially isn’t, shrinks are possible) then I’d say no way in hell is RDNA 4 on 3nm.
It doesn’t really matter in the end. I have no faith that even with a hypothetical node advantage AMD could actually make NV stumble and present themselves as a peer competitor. It would take a miracle at this point, AMD’s last time they could be one was what, Hawaii, arguably? Or hell, the HD5000 series where they capitalized on NV eff up that was OG Fermi?
Posted on Reply
#44
Gucky
It really doesn't look good. Ada is already close to B100 in density.
What could happen is that GB202 uses 2x GB203. :D That does make a bit sense, if GB202 is 512bit and GB203 is 256bit... But I doubt it.
How much extra performance is there with a 30% density and shader increase? More VRAM bandwidth with GDDR7 also gives some performance, but +70% performance in total like some rumors said???
I won't buy a GPU that uses more power then the 4090, I even undervolted it below 300W...

If the latest rumors are true, Blackwell might be dead for consumers...
gffermariIf they manage to extract 20-25% more performance using the same node, would be a success.
I doubt many enthusiasts would buy a 5090 at that point for over 1700$...
Posted on Reply
#45
kondamin
If they changed the architecture for better rasterization they might just be able to pull it off
Posted on Reply
#46
RogueSix
IF the rumors are true: Very smart move by nVidia. They are going to make dozens of billions by allocating all 3nm capacities to the professional AI/datacenter stuff while gaming GPUs will not compete over scarce 3nm capacities because they are going to be made on the older 4N node. At the same time, the 4N node is super-mature so nVidia will make very nice and juicy margins from the gaming products.

It would be pretty unfortunate for us gamers if these rumors are really true but it is kind of inevitable that gaming is taking a backseat for now if you look at the sheer numbers. nVidia made more than SIX times more revenue from AI/datacenter than from gaming ($18.404bn vs. $2.865bn). And yet another ~$0.75bn was made from Automotive and Professional Visualization.

nVidia have certainly almost exclusively been focusing on datacenter for quite a while by now and moved all of their top talent (engineering *and* software) to the datacenter segment. That is what makes this rumor pretty plausible. They are probably putting a pretty low effort into the next gaming generation. Compared to the previous gens that had lots of surprises on the feature/software side (DLSS3/FG/Remix etc.) the RTX 5000 series will probably be a disappointment but that's how it goes when the spotlight shifts in such a MASSIVE way...
Posted on Reply
#47
Prima.Vera
BwazeWe will see if this new "Jensen's Law" looks like this now:

2020, RTX 3080 - $700
2022, RTX 4080 - $1200
2024, RTX 5080 - $2060
No gaming card is worth more than 1000$, with all taxes included. Not even the top dog.
But hey, people are paying 1000+ $ for the same phones every year, so there is no lack of idiots with more money than common sense...
Posted on Reply
#48
kondamin
RogueSixIF the rumors are true: Very smart move by nVidia. They are going to make dozens of billions by allocating all 3nm capacities to the professional AI/datacenter stuff while gaming GPUs will not compete over scarce 3nm capacities because they are going to be made on the older 4N node. At the same time, the 4N node is super-mature so nVidia will make very nice and juicy margins from the gaming products.

It would be pretty unfortunate for us gamers if these rumors are really true but it is kind of inevitable that gaming is taking a backseat for now if you look at the sheer numbers. nVidia made more than SIX times more revenue from AI/datacenter than from gaming ($18.404bn vs. $2.865bn). And yet another ~$0.75bn was made from Automotive and Professional Visualization.

nVidia have certainly almost exclusively been focusing on datacenter for quite a while by now and moved all of their top talent (engineering *and* software) to the datacenter segment. That is what makes this rumor pretty plausible. They are probably putting a pretty low effort into the next gaming generation. Compared to the previous gens that had lots of surprises on the feature/software side (DLSS3/FG/Remix etc.) the RTX 5000 series will probably be a disappointment but that's how it goes when the spotlight shifts in such a MASSIVE way...
That might be a good thing, splitting the segments completely and allowing them to diverge.
3bn isn't peanuts
Posted on Reply
#49
Bwaze
Double-ClickThe 3080 was $700 MSRP, that's ~$836 today - Jensen shifted the goal post with the 4080 big time asking $1200.
Four figures for an 80-series card is still too high IMO, you shouldn't break that barrier until the 90-series.
Yeah, but the usual Nvidia apologizing line goes:

- "RTX 3080 was never really $700 card, for most of it's lifetime it was sold for over $1500, and people (not gamers, though) were buying it!"

Even original TechPowerUP review told us to compare it to cryptoinflated RTX 3090 Ti, which officialy launched for insane $2000! And then crypto collapsed.

And when RTX 5080 hits at close to $2000, we'll hear that we can't compare it to RTX 4080 SUPER's $1000 MSRP or other discounts but with original generation releases $1200, and add two years worth of inflation - which in this economy is what people think it is. So in view of many a $2000 card in start of 2025 will be perfectly same value as $1200 card in 2022, and they'll have charts to prove it to you! It's not even more expensive!
Posted on Reply
#50
Fouquin
DemonicRyzen666Nvidia can't come close to 40% increase on the same node, & has never achieved this.
Citation needed. Actually, let's just debunk this one right here and now.

TSMC 150nm, NV20 to NV25: 44% aggregate increase.
TSMC 130nm, NV38 to NV40: 63% aggregate increase.
TSMC 90nm, G71 to G80: 88.7% aggregate increase.
TSMC 65nm, G92 to GT200: 49.8% aggregate increase.
TSMC 28nm, GK110 to GM200: 49.3% aggregate increase.
Posted on Reply
Add your own comment
Jan 11th, 2025 01:18 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts