Monday, November 13th 2023

Supermicro Expands AI Solutions with the Upcoming NVIDIA HGX H200 and MGX Grace Hopper Platforms Featuring HBM3e Memory

Supermicro, Inc., a Total IT Solution Provider for AI, Cloud, Storage, and 5G/Edge, is expanding its AI reach with the upcoming support for the new NVIDIA HGX H200 built with H200 Tensor Core GPUs. Supermicro's industry leading AI platforms, including 8U and 4U Universal GPU Systems, are drop-in ready for the HGX H200 8-GPU, 4-GPU, and with nearly 2x capacity and 1.4x higher bandwidth HBM3e memory compared to the NVIDIA H100 Tensor Core GPU. In addition, the broadest portfolio of Supermicro NVIDIA MGX systems supports the upcoming NVIDIA Grace Hopper Superchip with HBM3e memory. With unprecedented performance, scalability, and reliability, Supermicro's rack scale AI solutions accelerate the performance of computationally intensive generative AI, large language Model (LLM) training, and HPC applications while meeting the evolving demands of growing model sizes. Using the building block architecture, Supermicro can quickly bring new technology to market, enabling customers to become more productive sooner.

Supermicro is also introducing the industry's highest density server with NVIDIA HGX H100 8-GPUs systems in a liquid cooled 4U system, utilizing the latest Supermicro liquid cooling solution. The industry's most compact high performance GPU server enables data center operators to reduce footprints and energy costs while offering the highest performance AI training capacity available in a single rack. With the highest density GPU systems, organizations can reduce their TCO by leveraging cutting-edge liquid cooling solutions.
"Supermicro partners with NVIDIA to design the most advanced systems for AI training and HPC applications," said Charles Liang, president and CEO of Supermicro. "Our building block architecture enables us to be first to market with the latest technology, allowing customers to deploy generative AI faster than ever before. We can deliver these new systems to customers faster with our worldwide manufacturing facilities. The new systems, using the NVIDIA H200 GPU with NVIDIA NVLink and NVSwitch high-speed GPU-GPU interconnects at 900 GB/s, now provide up to 1.1 TB of high-bandwidth HBM3e memory per node in our rack scale AI solutions to deliver the highest performance of model parallelism for today's LLMs and generative AI. We are also excited to offer the world's most compact NVIDIA HGX 8-GPU liquid cooled server, which doubles the density of our rack scale AI solutions and reduces energy costs to achieve green computing for today's accelerated data center."

Supermicro designs and manufactures a broad portfolio of AI servers with different form factors. The popular 8U and 4U Universal GPU systems featuring four-way and eight-way NVIDIA HGX H100 GPUs are now drop-in ready for the new H200 GPUs to train even larger language models in less time. Each NVIDIA H200 GPU contains 141 GB of memory with a bandwidth of 4.8 TB/s.

"Supermicro's upcoming server designs using NVIDIA HGX H200 will help accelerate generative AI and HPC workloads, so that enterprises and organizations can get the most out of their AI infrastructure," said Dion Harris, director of data center product solutions for HPC, AI, and quantum computing at NVIDIA. "The NVIDIA H200 GPU with high-speed HBM3e memory will be able to handle massive amounts of data for a variety of workloads."

Additionally, the recently launched Supermicro MGX servers with the NVIDIA GH200 Grace Hopper Superchips are engineered to incorporate the NVIDIA H200 GPU with HBM3e memory.

The new NVIDIA GPUs allow acceleration of today's and future large language models (LLMs) with 100s of billions of parameters to fit in more compact and efficient clusters to train Generative AI with less time and also allow multiple larger models to fit in one system for real-time LLM inference to serve Generative AI for millions of users.

At SC23, Supermicro is showcasing the latest offering, a 4U Universal GPU System featuring the eight-way NVIDIA HGX H100 with its latest liquid-cooling innovations that further improve density and efficiency to drive the evolution of AI. With Supermicro's industry leading GPU and CPU cold plates, CDU (cooling distribution unit), and CDM (cooling distribution manifold) designed for green computing, the new liquid-cooled 4U Universal GPU System is also ready for the eight-way NVIDIA HGX H200, which will dramatically reduce data center footprints, power cost, and deployment hurdles through Supermicro's fully integrated liquid-cooling rack solutions and our L10, L11 and L12 validation testing.

For more information, visit the Supermicro booth at SC23
Source: Supermicro
Add your own comment

3 Comments on Supermicro Expands AI Solutions with the Upcoming NVIDIA HGX H200 and MGX Grace Hopper Platforms Featuring HBM3e Memory

#1
Minus Infinity
But why would you buy this with Blackwell based B100 out late next year and will probably be even faster than H200.
Posted on Reply
#2
mahirzukic2
Minus InfinityBut why would you buy this with Blackwell based B100 out late next year and will probably be even faster than H200.
Because you need it right now for business purposes. Are you going to wait for a year in the off-chance that you get a better and faster product?
Are you going to sit on your ass doing nothing and waiting for the next best thing while the business suffers?
Probably not. "Perfect is the enemy of good" or good enough.
Posted on Reply
#3
Minus Infinity
mahirzukic2Because you need it right now for business purposes. Are you going to wait for a year in the off-chance that you get a better and faster product?
Are you going to sit on your ass doing nothing and waiting for the next best thing while the business suffers?
Probably not. "Perfect is the enemy of good" or good enough.
LOL so you are saying there's nothing currently available for these companies or they "must" upgrade everytime a new product appears. I'm sure they can slum it on A100 or H100.
Posted on Reply
Dec 21st, 2024 23:51 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts