• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.

Micron Announces 12-high HBM3E Memory, Bringing 36 GB Capacity and 1.2 TB/s Bandwidth

GFreeman

News Editor
Staff member
Joined
Mar 6, 2023
Messages
1,352 (2.45/day)
As AI workloads continue to evolve and expand, memory bandwidth and capacity are increasingly critical for system performance. The latest GPUs in the industry need the highest performance high bandwidth memory (HBM), significant memory capacity, as well as improved power efficiency. Micron is at the forefront of memory innovation to meet these needs and is now shipping production-capable HBM3E 12-high to key industry partners for qualification across the AI ecosystem.

Micron's industry-leading HBM3E 12-high 36 GB delivers significantly lower power consumption than our competitors' 8-high 24 GB offerings, despite having 50% more DRAM capacity in the package
Micron HBM3E 12-high boasts an impressive 36 GB capacity, a 50% increase over current HBM3E 8-high offerings, allowing larger AI models like Llama 2 with 70 billion parameters to run on a single processor. This capacity increase allows faster time to insight by avoiding CPU offload and GPU-GPU communication delays. Micron HBM3E 12-high 36 GB delivers significantly lower power consumption than the competitors' HBM3E 8-high 24 GB solutions. Micron HBM3E 12-high 36 GB offers more than 1.2 terabytes per second (TB/s) of memory bandwidth at a pin speed greater than 9.2 gigabits per second (Gb/s). These combined advantages of Micron HBM3E offer maximum throughput with the lowest power consumption can ensure optimal outcomes for power-hungry data centers. Additionally, Micron HBM3E 12-high incorporates fully programmable MBIST that can run system representative traffic at full spec speed, providing improved test coverage for expedited validation and enabling faster time to market and enhancing system reliability.



Robust ecosystem support
Micron is now shipping production-capable HBM3E 12-high units to key industry partners for qualification across the AI ecosystem. This HBM3E 12-high milestone demonstrates Micron's innovations to meet the data-intensive demands of the evolving AI infrastructure.

Micron is also a proud partner in TSMC's 3DFabric Alliance, which helps shape the future of semiconductor and system innovations. AI system manufacturing is complex, and HBM3E integration requires close collaboration between memory suppliers, customers and outsourced semiconductor assembly and test (OSAT) players.

In a recent exchange, Dan Kochpatcharin, head of the Ecosystem and Alliance Management Division at TSMC, commented, "TSMC and Micron have enjoyed a long-term strategic partnership. As part of the OIP ecosystem, we have worked closely to enable Micron's HBM3E-based system and chip-on-wafer-on-substrate (CoWoS) packaging design to support our customer's AI innovation."

In summary, here are the Micron HBM3E 12-high 36 GB highlights:
  • Undergoing multiple customer qualifications: Micron is shipping production-capable 12-high units to key industry partners to enable qualifications across the AI ecosystem.
  • Seamless scalability: With 36 GB of capacity (a 50% increase in capacity over current HBM3E offerings), Micron HBM3E 12-high allows data centers to scale their increasing AI workloads seamlessly.
  • Exceptional efficiency: Micron HBM3E 12-high 36 GB delivers significantly lower power consumption than the competitive HBM3E 8-high 24 GB solution!
  • Superior performance: With pin speed greater than 9.2 gigabits per second (Gb/s), Micron HBM3E 12-high 36 GB delivers more than 1.2 TB/s of memory bandwidth, enabling lightning-fast data access for AI accelerators, supercomputers and data centers.
  • Expedited validation: Fully programmable MBIST capabilities can run at speeds representative of system traffic, providing improved test coverage for expedited validation, enabling faster time to market and enhancing system reliability.

Looking ahead
Micron's leading-edge data center memory and storage portfolio is designed to meet the evolving demands of generative AI workloads. From near memory (HBM) and main memory (high-capacity server RDIMMs) to Gen 5 PCIe NVMe SSDs and data lake SSDs, Micron offers market-leading products that scale AI workloads efficiently and effectively.

As Micron continues to focus on extending its industry leadership, the company is already looking toward the future with its HBM4 and HBM4E roadmap. This forward-thinking approach ensures that Micron remains at the forefront of memory and storage development, driving the next wave of advancements in data center technology.

For more information, visit Micron's HBM3E page.

View at TechPowerUp Main Site | Source
 
Joined
Jan 3, 2021
Messages
3,183 (2.37/day)
Location
Slovenia
Processor i5-6600K
Motherboard Asus Z170A
Cooling some cheap Cooler Master Hyper 103 or similar
Memory 16GB DDR4-2400
Video Card(s) IGP
Storage Samsung 850 EVO 250GB
Display(s) 2x Oldell 24" 1920x1200
Case Bitfenix Nova white windowless non-mesh
Audio Device(s) E-mu 1212m PCI
Power Supply Seasonic G-360
Mouse Logitech Marble trackball, never had a mouse
Keyboard Key Tronic KT2000, no Win key because 1994
Software Oldwin
So Micron outsoutces a crucial (muh) part of their process, the bonding of stacked dies, to TSMC? That's surprising.
 
Joined
Dec 28, 2012
Messages
3,699 (0.87/day)
System Name Skunkworks
Processor 5800x3d
Motherboard x570 unify
Cooling Noctua NH-U12A
Memory 32GB 3600 mhz
Video Card(s) asrock 6800xt challenger D
Storage Sabarent rocket 4.0 2TB, MX 500 2TB
Display(s) Asus 1440p144 27"
Case Old arse cooler master 932
Power Supply Corsair 1200w platinum
Mouse *squeak*
Keyboard Some old office thing
Software openSUSE tumbleweed/Mint 21.2
So Micron outsoutces a crucial (muh) part of their process, the bonding of stacked dies, to TSMC? That's surprising.
Wait, so its all TSMC?

....always has been.
 

bug

Joined
May 22, 2015
Messages
13,538 (3.99/day)
Processor Intel i5-12600k
Motherboard Asus H670 TUF
Cooling Arctic Freezer 34
Memory 2x16GB DDR4 3600 G.Skill Ripjaws V
Video Card(s) EVGA GTX 1060 SC
Storage 500GB Samsung 970 EVO, 500GB Samsung 850 EVO, 1TB Crucial MX300 and 2TB Crucial MX500
Display(s) Dell U3219Q + HP ZR24w
Case Raijintek Thetis
Audio Device(s) Audioquest Dragonfly Red :D
Power Supply Seasonic 620W M12
Mouse Logitech G502 Proteus Core
Keyboard G.Skill KM780R
Software Arch Linux + Win10
Not gonna come anywhere near the consumer space, so meh...
 
Joined
Nov 26, 2021
Messages
1,517 (1.49/day)
Location
Mississauga, Canada
Processor Ryzen 7 5700X
Motherboard ASUS TUF Gaming X570-PRO (WiFi 6)
Cooling Noctua NH-C14S (two fans)
Memory 2x16GB DDR4 3200
Video Card(s) Reference Vega 64
Storage Intel 665p 1TB, WD Black SN850X 2TB, Crucial MX300 1TB SATA, Samsung 830 256 GB SATA
Display(s) Nixeus NX-EDG27, and Samsung S23A700
Case Fractal Design R5
Power Supply Seasonic PRIME TITANIUM 850W
Mouse Logitech
VR HMD Oculus Rift
Software Windows 11 Pro, and Ubuntu 20.04
Joined
Nov 4, 2005
Messages
11,895 (1.73/day)
System Name Compy 386
Processor 7800X3D
Motherboard Asus
Cooling Air for now.....
Memory 64 GB DDR5 6400Mhz
Video Card(s) 7900XTX 310 Merc
Storage Samsung 990 2TB, 2 SP 2TB SSDs, 24TB Enterprise drives
Display(s) 55" Samsung 4K HDR
Audio Device(s) ATI HDMI
Mouse Logitech MX518
Keyboard Razer
Software A lot.
Benchmark Scores Its fast. Enough.
I was confused by the 50% more when they stacked 12 instead of the 8, I'm glad they were able to point out that 12 is 50% more than 8, my life is now complete.
 

bug

Joined
May 22, 2015
Messages
13,538 (3.99/day)
Processor Intel i5-12600k
Motherboard Asus H670 TUF
Cooling Arctic Freezer 34
Memory 2x16GB DDR4 3600 G.Skill Ripjaws V
Video Card(s) EVGA GTX 1060 SC
Storage 500GB Samsung 970 EVO, 500GB Samsung 850 EVO, 1TB Crucial MX300 and 2TB Crucial MX500
Display(s) Dell U3219Q + HP ZR24w
Case Raijintek Thetis
Audio Device(s) Audioquest Dragonfly Red :D
Power Supply Seasonic 620W M12
Mouse Logitech G502 Proteus Core
Keyboard G.Skill KM780R
Software Arch Linux + Win10
I was confused by the 50% more when they stacked 12 instead of the 8, I'm glad they were able to point out that 12 is 50% more than 8, my life is now complete.
That's how it should always be (% diff compared to the old value), though some will play fast and loose with that.
 

bug

Joined
May 22, 2015
Messages
13,538 (3.99/day)
Processor Intel i5-12600k
Motherboard Asus H670 TUF
Cooling Arctic Freezer 34
Memory 2x16GB DDR4 3600 G.Skill Ripjaws V
Video Card(s) EVGA GTX 1060 SC
Storage 500GB Samsung 970 EVO, 500GB Samsung 850 EVO, 1TB Crucial MX300 and 2TB Crucial MX500
Display(s) Dell U3219Q + HP ZR24w
Case Raijintek Thetis
Audio Device(s) Audioquest Dragonfly Red :D
Power Supply Seasonic 620W M12
Mouse Logitech G502 Proteus Core
Keyboard G.Skill KM780R
Software Arch Linux + Win10
Joined
Nov 26, 2021
Messages
1,517 (1.49/day)
Location
Mississauga, Canada
Processor Ryzen 7 5700X
Motherboard ASUS TUF Gaming X570-PRO (WiFi 6)
Cooling Noctua NH-C14S (two fans)
Memory 2x16GB DDR4 3200
Video Card(s) Reference Vega 64
Storage Intel 665p 1TB, WD Black SN850X 2TB, Crucial MX300 1TB SATA, Samsung 830 256 GB SATA
Display(s) Nixeus NX-EDG27, and Samsung S23A700
Case Fractal Design R5
Power Supply Seasonic PRIME TITANIUM 850W
Mouse Logitech
VR HMD Oculus Rift
Software Windows 11 Pro, and Ubuntu 20.04
High-latency, huge bandwidth, iirc, which isn't a great fit for consumer GPUs.
GPUs don't care about latency as much as CPUs. GDDR6 has higher latency than garden variety DDR4. The difference in latency between HBM and DDR4, despite being large, isn't as stark (link to PDF). HBM is vastly superior to GDDR6 on technical merits, but the latter is much cheaper. The reason HBM isn't used for consumer GPUs is that it's extremely expensive.

1725641018361.png
 

Aquinus

Resident Wat-man
Joined
Jan 28, 2012
Messages
13,156 (2.86/day)
Location
Concord, NH, USA
System Name Apollo
Processor Intel Core i9 9880H
Motherboard Some proprietary Apple thing.
Memory 64GB DDR4-2667
Video Card(s) AMD Radeon Pro 5600M, 8GB HBM2
Storage 1TB Apple NVMe, 4TB External
Display(s) Laptop @ 3072x1920 + 2x LG 5k Ultrafine TB3 displays
Case MacBook Pro (16", 2019)
Audio Device(s) AirPods Pro, Sennheiser HD 380s w/ FIIO Alpen 2, or Logitech 2.1 Speakers
Power Supply 96w Power Adapter
Mouse Logitech MX Master 3
Keyboard Logitech G915, GL Clicky
Software MacOS 12.1
High-latency, huge bandwidth, iirc, which isn't a great fit for consumer GPUs.
It's an excellent fit for how GPUs work. The problem is the added cost isn't worth it for desktop GPUs because less expensive options can get the same results. I'd argue that HBM's advantage isn't bandwidth because we already have plenty of it, but the power efficiency and size demands compared to traditional DRAM. That makes it far more suitable for higher performance mobile applications in my opinion because you can get the same done with less power and in less space. Both of which are precious commodities for mobile devices and the server space.

So I agree that it's not a great fit for desktop GPUs. It's a great fit for mobile and server GPUs simply because of the power consumption and space advantage it has. We already see this in the server market with these server GPUs that nVidia has been producing for AI and whatnot. The disadvantage of HBM is all of the costs (money) associated with it.
 
Joined
Jun 27, 2011
Messages
6,755 (1.40/day)
Processor 7800x3d
Motherboard Gigabyte B650 Auros Elite AX
Cooling Custom Water
Memory GSKILL 2x16gb 6000mhz Cas 30 with custom timings
Video Card(s) MSI RX 6750 XT MECH 2X 12G OC
Storage Adata SX8200 1tb with Windows, Samsung 990 Pro 2tb with games
Display(s) HP Omen 27q QHD 165hz
Case ThermalTake P3
Power Supply SuperFlower Leadex Titanium
Software Windows 11 64 Bit
Benchmark Scores CB23: 1811 / 19424 CB24: 1136 / 7687
This would never exist but I think it would be cool. Imagine an APU with 220w TDP and 32gb on die HBM memory, in addition to normal dimm slots separate from the HBM. You could have the benefits of a massive L4 cache for the CPU and more than sufficient on die VRAM for the GPU.
 

Aquinus

Resident Wat-man
Joined
Jan 28, 2012
Messages
13,156 (2.86/day)
Location
Concord, NH, USA
System Name Apollo
Processor Intel Core i9 9880H
Motherboard Some proprietary Apple thing.
Memory 64GB DDR4-2667
Video Card(s) AMD Radeon Pro 5600M, 8GB HBM2
Storage 1TB Apple NVMe, 4TB External
Display(s) Laptop @ 3072x1920 + 2x LG 5k Ultrafine TB3 displays
Case MacBook Pro (16", 2019)
Audio Device(s) AirPods Pro, Sennheiser HD 380s w/ FIIO Alpen 2, or Logitech 2.1 Speakers
Power Supply 96w Power Adapter
Mouse Logitech MX Master 3
Keyboard Logitech G915, GL Clicky
Software MacOS 12.1
This would never exist but I think it would be cool. Imagine an APU with 220w TDP and 32gb on die HBM memory, in addition to normal dimm slots separate from the HBM. You could have the benefits of a massive L4 cache for the CPU and more than sufficient on die VRAM for the GPU.
You don't have to look that far to find a CPU with HBM memory onboard.

Or reviews to see how it fares on HPC applications.

I agree though. I'd like to see an APU-like device with a stack or two of this new HBM3e, at the very least to see how it fares.
 
Top