Friday, May 17th 2024

TSMC Unveils Next-Generation HBM4 Base Dies, Built on 12 nm and 5 nm Nodes

During the European Technology Symposium 2024, TSMC has announced its readiness to manufacture next-generation HBM4 base dies using both 12 nm and 5 nm nodes. This significant development is expected to substantially improve the performance, power consumption, and logic density of HBM4 memory, catering to the demands of high-performance computing (HPC) and artificial intelligence (AI) applications. The shift from a traditional 1024-bit interface to an ultra-wide 2048-bit interface is a key aspect of the new HBM4 standard. This change will enable the integration of more logic and higher performance while reducing power consumption. TSMC's N12FFC+ and N5 processes will be used to produce these base dies, with the N12FFC+ process offering a cost-effective solution for achieving HBM4 performance and the N5 process providing even more logic and lower power consumption at HBM4 speeds.

The company is collaborating with major HBM memory partners, including Micron, Samsung, and SK Hynix, to integrate advanced nodes for HBM4 full-stack integration. TSMC's base die, fabricated using the N12FFC+ process, will be used to install HBM4 memory stacks on a silicon interposer alongside system-on-chips (SoCs). This setup will enable the creation of 12-Hi (48 GB) and 16-Hi (64 GB) stacks with per-stack bandwidth exceeding 2 TB/s. TSMC's collaboration with EDA partners like Cadence, Synopsys, and Ansys ensures the integrity of HBM4 channel signals, thermal accuracy, and electromagnetic interference (EMI) in the new HBM4 base dies. TSMC is also optimizing CoWoS-L and CoWoS-R for HBM4 integration, meaning that massive high-performance chips are already utilizing this technology and getting ready for volume manufacturing.
Source: AnandTech
Add your own comment

13 Comments on TSMC Unveils Next-Generation HBM4 Base Dies, Built on 12 nm and 5 nm Nodes

#1
Wirko
It's not at all clear what logic the HBM4 base (logic) die contains. The memory controller is located on the GPU or whatever the main processor is, so what remains are some signal buffers, testing logic and maybe power management. Nothing that would require 12nm. This, basically:



Or is this old architecture about to change substantially? The guys at SemiEngineering have a thorough understanding of what they write about, and they say, "SK hynix said the collaboration will enable breakthroughs in memory performance with increased density of the memory controller at the base of the HBM stack."
Posted on Reply
#2
AnarchoPrimitiv
64GB stacks? So then we could see 6x stacks for a total.of 384GB of HBM4? That's wild, can you imagine what the next super APU from AMD will contain? It really will be an entire system on an AIC.

It's a shame AI is eating up all the HBM, can you imagine what that new Strix Halo APU would be capable of with 16GB of HBM on package?
Posted on Reply
#3
SOAREVERSOR
AnarchoPrimitiv64GB stacks? So then we could see 6x stacks for a total.of 384GB of HBM4? That's wild, can you imagine what the next super APU from AMD will contain? It really will be an entire system on an AIC.

It's a shame AI is eating up all the HBM, can you imagine what that new Strix Halo APU would be capable of with 16GB of HBM on package?
AI is not eating all the HBM it's also available on pro cards aka quadro. The issue is that the cost of it is not really a good idea for consumer cards. And for all the footsies stompsies people already throw about the cost of GPUs imagine the billow biting and tantrums with the added costs of HBM!
Posted on Reply
#4
Jism
AnarchoPrimitiv64GB stacks? So then we could see 6x stacks for a total.of 384GB of HBM4? That's wild, can you imagine what the next super APU from AMD will contain? It really will be an entire system on an AIC.

It's a shame AI is eating up all the HBM, can you imagine what that new Strix Halo APU would be capable of with 16GB of HBM on package?
Adding more HBM to a APU is not going to change it's performance just like that. You still need to have a GPU that's capable of taking advantage of all that extra memory bandwidth.

And if a GPU is rated for over 150W of power, what do you keep left for the CPU? Your starting to get into absurd cooling territory here if it was just for an APU.
Posted on Reply
#5
Panther_Seraphin
WirkoIt's not at all clear what logic the HBM4 base (logic) die contains. The memory controller is located on the GPU or whatever the main processor is, so what remains are some signal buffers, testing logic and maybe power management. Nothing that would require 12nm. This, basically:



Or is this old architecture about to change substantially? The guys at SemiEngineering have a thorough understanding of what they write about, and they say, "SK hynix said the collaboration will enable breakthroughs in memory performance with increased density of the memory controller at the base of the HBM stack."
Correct me if I am wrong but I was under impression the Memory controller of the device spoke to the logic die and the logic die spoke to the HBM banks and there wasnt a direct connection between the memory and the memory controller.

upcommons.upc.edu/bitstream/handle/2117/362010/HBM.pdf
HBM STRUCTURE AND FEATURES
Point B



I wonder then if they are going to move from the 1024bit interface to say 1536 or 2048 per die? This just means you can go >8 "stacks" per die is what SK Hynix was alluding to?
Posted on Reply
#6
overclockedamd
JismAdding more HBM to a APU is not going to change it's performance just like that. You still need to have a GPU that's capable of taking advantage of all that extra memory bandwidth.


AMD's 8000g series seem to scale pretty linearly with memory bandwidth. I know HBM4 would be massive overkill for this but fast ram does help.
Posted on Reply
#7
evernessince
SOAREVERSORThe issue is that the cost of it is not really a good idea for consumer cards.
It costs more than GDDR but to the point of being not logical for consumer products? I don't think so, Vega 56 was $400 at launch. If AMD could slip HBM into cards at that price point, with the economies of scale being much more in favor of HBM today, it should absolutely be possible to put on consumer cards. Especially after you consider the price of current video cards.

Consumer cards don't need anywhere near the latest HBM either. They'd be fine with going with last gen to save money.
SOAREVERSORAnd for all the footsies stompsies people already throw about the cost of GPUs imagine the billow biting and tantrums with the added costs of HBM!
People are mad at the unjustified inflation of GPU prices, which makes sense given performance per dollar metrics have hardly been improving (particularly on the "low" end). Meanwhile game requirements are increasing, which has translated to a regression in how much performance people's budgets get them in new games.

Demeaning these complaints as "billow biting" and "tantrums" is simply ignoring these facts and is nothing more than a Ad Hominen attack designed to discredit anyone that disagrees with your argument without actually having to debate the facts.

I really don't think a price increase would be required at all given Nvidia operates at a gross margin of 64% (vastly higher than even Apple) but I also don't think people would be mad at a $30 price increase on their $1,000 GPU if it meant more memory bandwidth and lower power usage. Nvidia has already make the argument that new nodes are a large part of why pricing has increased (an argument that is dubious in part because their profit margins have increased in tune with their price increases and the fact that wafer cost is only a portion of their total expenses). Following Nvidia's own logic, nodes are expensive and therefore the cost to implment HBM becomes increasingly trivial as compared to the rising cost of nodes. Especially when you consider a cash starved AMD did it back when they were only making 1/3rd of the margins Nvidia is now.
Posted on Reply
#8
Random_User
SOAREVERSORAI is not eating all the HBM it's also available on pro cards aka quadro. The issue is that the cost of it is not really a good idea for consumer cards. And for all the footsies stompsies people already throw about the cost of GPUs imagine the billow biting and tantrums with the added costs of HBM!
I mean, at the price of these top consumer solutions akin 4090, or other almost two grand worth product, they better have HBM on-board, instead GDDR6X. A limited series of 4090/7900XTX HBM versions, could easilly find their buyers (these are already go in way lower amounts anyway). Not to mention with amounts of memory these top cards use, the HBM would suit even more, especially by reducing the temperature footprint, along with much wider bus. And it was beneficial to GCN, this also might benefit ADA/Blackwell and RDNA3/4/5. Especially, that RDNA MCDs can be altered to any memory type/configuration.

evernessinceIt costs more than GDDR but to the point of being not logical for consumer products? I don't think so, Vega 56 was $400 at launch. If AMD could slip HBM into cards at that price point, with the economies of scale being much more in favor of HBM today, it should absolutely be possible to put on consumer cards. Especially after you consider the price of current video cards.

Consumer cards don't need anywhere near the latest HBM either. They'd be fine with going with last gen to save money.




People are mad at the unjustified inflation of GPU prices, which makes sense given performance per dollar metrics have hardly been improving (particularly on the "low" end). Meanwhile game requirements are increasing, which has translated to a regression in how much performance people's budgets get them in new games.

Demeaning these complaints as "billow biting" and "tantrums" is simply ignoring these facts and is nothing more than a Ad Hominen attack designed to discredit anyone that disagrees with your argument without actually having to debate the facts.

I really don't think a price increase would be required at all given Nvidia operates at a gross margin of 64% (vastly higher than even Apple) but I also don't think people would be mad at a $30 price increase on their $1,000 GPU if it meant more memory bandwidth and lower power usage. Nvidia has already make the argument that new nodes are a large part of why pricing has increased (an argument that is dubious in part because their profit margins have increased in tune with their price increases and the fact that wafer cost is only a portion of their total expenses). Following Nvidia's own logic, nodes are expensive and therefore the cost to implment HBM becomes increasingly trivial as compared to the rising cost of nodes. Especially when you consider a cash starved AMD did it back when they were only making 1/3rd of the margins Nvidia is now.
People not only complain. They are no longer able to afford low and medium end cards, they used to buy for decades. Because those cards suddengly had their price tags became premium. Considering the margins, the entire range of GPUs, is highly overpriced. As if the BOM has risen that much, that should have eaten significant portions of those said margins. But it didn't.

And this isn't only demeaning. This is discarding any reasoning. Kinda, like no one is allowed asking these companies about their behavior, just take everything they say and report as given.

P.S.: Intel used to make Kaby Lake-G with single HBM2 stack on board. These still being used in specific embedded systems, like this, for around $300. This is obvious, that these APU's are old, and the HBM2 stacks have been installed more than five years ago, when the HBM prices were nowhere as expensive. But there's nothing wrong in this concept, and the NUCs based on KBL-G used to have impressive performance. I don't see why AMD can't revive this approach, since they benefit from it like no one else.
Posted on Reply
#9
Panther_Seraphin
HBM when Vega/Kay Lake-G were being released didnt really have a mass market appeal anywhere at the time so it didnt command a massive premium over GDDR like it does now.

HBM now due to the AI craze etc now cannot keep up with demand. I mean all of SKs 2025 production has been purchased if I remember correctly from a few weeks ago plus
www.trendforce.com/news/2024/05/14/news-samsung-and-sk-hynix-urgently-reallocate-20-of-dram-production-capacity-to-support-rising-hbm-demand/

So that "cheap" HBM that was offered back at the HBM1/2 days is NOT there anymore and it asks a hell of a premium vs GDDR.

Do you want to sell a $70k B200 or a $3k 5090........ You see where I am getting at.
Posted on Reply
#10
Random_User
Panther_SeraphinHBM when Vega/Kay Lake-G were being released didnt really have a mass market appeal anywhere at the time so it didnt command a massive premium over GDDR like it does now.

HBM now due to the AI craze etc now cannot keep up with demand. I mean all of SKs 2025 production has been purchased if I remember correctly from a few weeks ago plus
www.trendforce.com/news/2024/05/14/news-samsung-and-sk-hynix-urgently-reallocate-20-of-dram-production-capacity-to-support-rising-hbm-demand/

So that "cheap" HBM that was offered back at the HBM1/2 days is NOT there anymore and it asks a hell of a premium vs GDDR.

Do you want to sell a $70k B200 or a $3k 5090........ You see where I am getting at.
This is true. But the part of the "issue" is that AMD didn't secure even tiny amounts of HBM, even for the "experimental" APU series. I mean, Intel Kaby Lake-G didn't live long, and made only a bunch of them. But they still find the place in the custom products, years later.

Of course, HMB isn't viable inside desktop APUs, and it's clear, as those are direct derivatives of mobile silicon. But still, there could be some laptops and mini-PCs, where the there's plenty of space for bigger package substrate/die area (and it isn't restricted by IHS/socket), which would benefit such devices, since it will lower (not remove) the need of fast and expensive SODIMMs, at least for an iGPU.

But it's hard to tell if this approach is worth or not, since neither AMD, nor Intel got any consumer grade products on the market right now, that would have had an APU+HBM concept, even in limited amounts. Because, yeah, the entire HBM stocks went solely into the enterprise stuff.
I mean, if they'd had at least some amounts Hawk-Point+HBM products, it might have shown their viability, and consumer interest.

But that's my take.

P.S.: "3K for 5090" is still too overpriced. There's no reason consumer cards with GDDR having such price tags. I doubt, that even with that that pricey exclusive GDD6x, costs a grand. And it's even more ridiculous, to see AMD cards being neck-and-neck in price with nVidia counterparts, while having "regular" slower GRDDR6.
Posted on Reply
#11
evernessince
Panther_SeraphinHBM when Vega/Kay Lake-G were being released didnt really have a mass market appeal anywhere at the time so it didnt command a massive premium over GDDR like it does now.

HBM now due to the AI craze etc now cannot keep up with demand. I mean all of SKs 2025 production has been purchased if I remember correctly from a few weeks ago plus
www.trendforce.com/news/2024/05/14/news-samsung-and-sk-hynix-urgently-reallocate-20-of-dram-production-capacity-to-support-rising-hbm-demand/

So that "cheap" HBM that was offered back at the HBM1/2 days is NOT there anymore and it asks a hell of a premium vs GDDR.
The AI craze started after the 4000 series was released and especially after the 3000 series release. By extension it's almost irrelevant to point out the fact that HBM is supply constrained today as it simply wasn't a factor when these generations were being designed.

Consider also that HBM allocation is secured years in advance to ensure partners are able to build out production to meet customer requirements. The current shortfall is due to unexpected demand and that's where you start seeing supply constraints in the semi-conductor industry as production cannot just be picked up willy-nilly. IF Nvidia wanted some 4000 series cards to have HBM, it would have secured allocation at least 1 year or more ahead of time. They would have signed a contract agreeing to reasonable rates and allotment at that time, meaning they would have gotten much better pricing than you do in the current market. Pretty much what you are saying would only be relevant if Nvidia wanted some 5000 series cards to have HBM, which indeed is unlikely due to the pricing increase and short supply. That said it doesn't apply to the 4000 series, were Nvidia signed contracts securing their wafers and memory way before HBM supply became an issue.
Panther_SeraphinDo you want to sell a $70k B200 or a $3k 5090........ You see where I am getting at.
I'd assume this is indeed the reason why Nvidia is putting everything into enterprise / AI. The price they are fetching for those cards is insane.
Posted on Reply
#12
Panther_Seraphin
GDDR6x was a pricey way of getting extra bandwidth from the same width bus. GDDR7 is a halfway house between 6x and 6. These sorts of measures keeps the "cheaper" GDDR relevant in modern times and has meant things like HBM/HMC get sidelined in the consumer space.
Random_UserOf course, HMB isn't viable inside desktop APUs, and it's clear, that those are direct derivatives of mobile silicon.
What would have been a big push for it is if they had decided to go for HBM on say the PS5/Xbox Series X. AMD already saw the benefits in HBM and had experience with it from Vega but again costing/supply would have been the biggest constratint for it in the Console space.

So what do you do. Make it a halo only product again? 69xx maybe having HBM? So you either go for the top end of HBM2 which was available at the time as it was limited to 8Gb per package or your looking at 4 dies of the more common 4Gb layout to make it up to the 16Gb of the eventual 6900xt. So now you are talking a 4096bit bus on a consumer GPU. How big is this die going to end up? HBM2e/HBM3 upped the dies per package to 12 drastically increasing the memory density available hence why were are now seeing insane memory on the MI250/H100 and their upcoming replacements but they had only really just been announced when the 6900xt was being released.
evernessinceThe AI craze started after the 4000 series was released and especially after the 3000 series release.
AI went into HYPERDRIVE at that point but there was a build up of AI focus since 2016 when nVidia gifted their 1st DGX1 to OpenAI and it was always going to happen but was a question of when it would become viable outside of supercomputer. 3080 and especially the 3090s gave people the ability to do LLM on their home systems and the 4080/90s are even better at it due to increased power and VRAM.
evernessinceI'd assume this is indeed the reason why Nvidia is putting everything into enterprise / AI. The price they are fetching for those cards is insane.
Its why everyone not just nVidia is focusing so heavily on it. Its got similar demand as the crypto boom but the difference is that people are willing to add 000s to cheques to guarantee they get what they want when they want it. Its why your seeing things like AMD helping TinyCorp trying to get RDNA3 into the datacenter as getting more AMD products into the datacenter means more EPYC and Instinct accelerators in there as well in the longer term.
Posted on Reply
Add your own comment
Nov 23rd, 2024 03:22 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts