• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.

Blackwell Shipments Imminent, Total CoWoS Capacity Expected to Surge by Over 70% in 2025

TheLostSwede

News Editor
Joined
Nov 11, 2004
Messages
16,711 (2.33/day)
Location
Sweden
System Name Overlord Mk MLI
Processor AMD Ryzen 7 7800X3D
Motherboard Gigabyte X670E Aorus Master
Cooling Noctua NH-D15 SE with offsets
Memory 32GB Team T-Create Expert DDR5 6000 MHz @ CL30-34-34-68
Video Card(s) Gainward GeForce RTX 4080 Phantom GS
Storage 1TB Solidigm P44 Pro, 2 TB Corsair MP600 Pro, 2TB Kingston KC3000
Display(s) Acer XV272K LVbmiipruzx 4K@160Hz
Case Fractal Design Torrent Compact
Audio Device(s) Corsair Virtuoso SE
Power Supply be quiet! Pure Power 12 M 850 W
Mouse Logitech G502 Lightspeed
Keyboard Corsair K70 Max
Software Windows 10 Pro
Benchmark Scores https://valid.x86.fr/yfsd9w
TrendForce reports that NVIDIA's Hopper H100 began to see a reduction in shortages in 1Q24. The new H200 from the same platform is expected to gradually ramp in Q2, with the Blackwell platform entering the market in Q3 and expanding to data center customers in Q4. However, this year will still primarily focus on the Hopper platform, which includes the H100 and H200 product lines. The Blackwell platform—based on how far supply chain integration has progressed—is expected to start ramping up in Q4, accounting for less than 10% of the total high-end GPU market.

The die size of Blackwell platform chips like the B100 is twice that of the H100. As Blackwell becomes mainstream in 2025, the total capacity of TSMC's CoWoS is projected to grow by 150% in 2024 and by over 70% in 2025, with NVIDIA's demand occupying nearly half of this capacity. For HBM, the NVIDIA GPU platform's evolution sees the H100 primarily using 80 GB of HBM3, while the 2025 B200 will feature 288 GB of HBM3e—a 3-4 fold increase in capacity per chip. The three major manufacturers' expansion plans indicate that HBM production volume will likely double by 2025.




Microsoft, Meta, and AWS to be the first to adopt GB200
CSPs such as Microsoft, Meta, and AWS are taking the lead in adopting GB200 and their combined shipment volume is expected to exceed 30,000 racks by 2025. In contrast, Google is likely to prioritize expanding its TPU AI infrastructure. The Blackwell platform has been highly anticipated in the market because it differs from the traditional HGX or MGX single AI server approach. Instead, it promotes a whole-rack configuration, integrating NVIDIA's own CPU, GPU, NVLink, and InfiniBand high-speed networking technologies. The GB200 is divided into two configurations: NVL36 and NVL72, with each rack featuring 36 and 72 GPUs, respectively. While NVL72 is the primary configuration that NVIDIA is pushing, the complexity of its design suggests that the NVL36 will be introduced first by the end of 2024 for initial trials to ensure a quick market entry.

AI server demand from tier-2 data centers and the government set to rise to 50% in 2024
NVIDIA's FY1Q25 data highlights a noticeable increase in demand from enterprise customers, sovereign AI clouds, and tier-2 data centers. TrendForce notes that in 2023, NVIDIA prioritized its high-end GPUs for hyper CSPs due to limited CoWoS and HBM supply. In 2023, CSPs accounted for approximately 65% of the global high-end AI server market demand.

With the supply of GPU easing in 1H24, demand from server brands like Dell and HPE will expand to 19%. Other sources, including tier-2 data centers such as CoreWeave and Tesla, as well as sovereign AI clouds and supercomputing projects, will see their combined demand share rise to 31%. This growth is driven by the sensitive and privacy-focused nature of AI training and applications. This segment is expected to play a crucial role in driving the growth of shipments of AI servers in the future.

View at TechPowerUp Main Site | Source
 
Joined
May 3, 2018
Messages
2,550 (1.13/day)
It makes me wonder about Nvidia's PR telling us Rubin is already coming in 2025, just as Blackwell shipments really get going. Rubin will be a paper launch with shipments starting well into 2026. So why not just stick to the actual real world effective 2 year cadence rather than hyping a 1 year cadence for updates which is in no way true.
 
Top