Tuesday, March 14th 2023
Microsoft Azure Announces New Scalable Generative AI VMs Featuring NVIDIA H100
Microsoft Azure announced their new ND H100 v5 virtual machine which packs Intel's Sapphire Rapids Xeon Scalable processors with NVIDIA's Hopper H100 GPUs, as well as NVIDIA's Quantum-2 CX7 interconnect. Inside each physical machine sits eight H100s—presumably the SXM5 variant packing a whopping 132 SMs and 528 4th generation tensor cores—interconnected by NVLink 4.0 which ties them all together with 3.6 TB/s bisectional bandwidth. Outside each local machine is a network of thousands more H100s connected together with 400 GB/s Quantum-2 CX7 InfiniBand, which Microsoft says allows 3.2 Tb/s per VM for on-demand scaling to accelerate the largest AI training workloads.
Generative AI solutions like ChatGPT have accelerated demand for multi-ExaOP cloud services that can handle the large training sets and utilize the latest development tools. Azure's new ND H100 v5 VMs offer that capability to organizations of any size, whether you're a smaller startup or a larger company looking to implement large-scale AI training deployments. While Microsoft is not making any direct claims for performance, NVIDIA has advertised H100 as running up to 30x faster than the preceding Ampere architecture that is currently offered with the ND A100 v4 VMs.Microsoft Azure provides the following technical specifications for the new VMs:
Microsoft Azure has opened up the preview of the ND H100 v5 VM service and you can sign up to request access here.
Source:
Microsoft
Generative AI solutions like ChatGPT have accelerated demand for multi-ExaOP cloud services that can handle the large training sets and utilize the latest development tools. Azure's new ND H100 v5 VMs offer that capability to organizations of any size, whether you're a smaller startup or a larger company looking to implement large-scale AI training deployments. While Microsoft is not making any direct claims for performance, NVIDIA has advertised H100 as running up to 30x faster than the preceding Ampere architecture that is currently offered with the ND A100 v4 VMs.Microsoft Azure provides the following technical specifications for the new VMs:
- 8x NVIDIA H100 Tensor Core GPUs interconnected via next gen NVSwitch and NVLink 4.0
- 400 Gb/s NVIDIA Quantum-2 CX7 InfiniBand per GPU with 3.2 Tb/s per VM in a non-blocking fat-tree network
- NVSwitch and NVLink 4.0 with 3.6 TB/s bisectional bandwidth between 8 local GPUs within each VM
- 4th Gen Intel Xeon Scalable processors
- PCIE Gen 5 host to GPU interconnect with 64 GB/s bandwidth per GPU
- 16 Channels of 4800 MHz DDR5 DIMMs
Microsoft Azure has opened up the preview of the ND H100 v5 VM service and you can sign up to request access here.
5 Comments on Microsoft Azure Announces New Scalable Generative AI VMs Featuring NVIDIA H100
Considering the Xeon loses in pretty much every possible way, I think it's a pretty good joke. Maybe intel has returned to the strategy of giving generous discounts, I hope it doesn't suffer any more lawsuits. :P