Tuesday, March 18th 2025

NVIDIA Announces DGX Spark and DGX Station Personal AI Computers

NVIDIA today unveiled NVIDIA DGX personal AI supercomputers powered by the NVIDIA Grace Blackwell platform. DGX Spark—formerly Project DIGITS—and DGX Station, a new high-performance NVIDIA Grace Blackwell desktop supercomputer powered by the NVIDIA Blackwell Ultra platform, enable AI developers, researchers, data scientists and students to prototype, fine-tune and inference large models on desktops. Users can run these models locally or deploy them on NVIDIA DGX Cloud or any other accelerated cloud or data center infrastructure.

DGX Spark and DGX Station bring the power of the Grace Blackwell architecture, previously only available in the data center, to the desktop. Global system builders to develop DGX Spark and DGX Station include ASUS, Dell, HP Inc. and Lenovo.
"AI has transformed every layer of the computing stack. It stands to reason a new class of computers would emerge—designed for AI-native developers and to run AI-native applications," said Jensen Huang, founder and CEO of NVIDIA. "With these new DGX personal AI computers, AI can span from cloud services to desktop and edge applications."
Igniting Innovation With DGX Spark
DGX Spark is the world's smallest AI supercomputer, empowering millions of researchers, data scientists, robotics developers and students to push the boundaries of generative and physical AI with massive performance and capabilities.

At the heart of DGX Spark is the NVIDIA GB10 Grace Blackwell Superchip, optimized for a desktop form factor. GB10 features a powerful NVIDIA Blackwell GPU with fifth-generation Tensor Cores and FP4 support, delivering up to 1,000 trillion operations per second of AI compute for fine-tuning and inference with the latest AI reasoning models, including the NVIDIA Cosmos Reason world foundation model and NVIDIA GR00T N1 robot foundation model.

The GB10 Superchip uses NVIDIA NVLink -C2C interconnect technology to deliver a CPU+GPU-coherent memory model with 5x the bandwidth of fifth-generation PCIe. This lets the superchip access data between a GPU and CPU to optimize performance for memory-intensive AI developer workloads.

NVIDIA's full-stack AI platform enables DGX Spark users to seamlessly move their models from their desktops to DGX Cloud or any accelerated cloud or data center infrastructure—with virtually no code changes—making it easier than ever to prototype, fine-tune and iterate on their workflows.

Full Speed Ahead With DGX Station
NVIDIA DGX Station brings data-center-level performance to desktops for AI development. The first desktop system to be built with the NVIDIA GB300 Grace Blackwell Ultra Desktop Superchip, DGX Station features a massive 784 GB of coherent memory space to accelerate large-scale training and inferencing workloads. The GB300 Desktop Superchip features an NVIDIA Blackwell Ultra GPU with latest-generation Tensor Cores and FP4 precision—connected to a high-performance NVIDIA Grace CPU via NVLink-C2C—delivering best-in-class system communication and performance.

DGX Station also features the NVIDIA ConnectX -8 SuperNIC, optimized to supercharge hyperscale AI computing workloads. With support for networking at up to 800 Gb/s, the ConnectX-8 SuperNIC delivers extremely fast, efficient network connectivity, enabling high-speed connectivity of multiple DGX Stations for even larger workloads, and network-accelerated data transfers for AI workloads.

Combining these state-of-the-art DGX Station capabilities with the NVIDIA CUDA-X AI platform, teams can achieve exceptional desktop AI development performance.

In addition, users gain access to NVIDIA NIM microservices with the NVIDIA AI Enterprise software platform, which offers highly optimized, easy-to-deploy inference microservices backed by enterprise support.

Availability
Reservations for DGX Spark systems open today at nvidia.com.

DGX Station is expected to be available from manufacturing partners like ASUS, BOXX, Dell, HP, Lambda and Supermicro later this year.
Source: NVIDIA
Add your own comment

9 Comments on NVIDIA Announces DGX Spark and DGX Station Personal AI Computers

#1
abysal
Oooo nice, is that 3x 12VHPWR? 3x chance of fire!
Posted on Reply
#2
trsttte
abysalOooo nice, is that 3x 12VHPWR? 3x chance of fire!
Right next to a larger standard ATX 24 pin on a regular ATX sized board, just why????
Posted on Reply
#4
igormp
HDBitdataIs the CPU ARM or x86?
ARM.
Posted on Reply
#5
Rover4444
igormpARM.
Should've been RISC-V if they really wanted to be wild with it.
Posted on Reply
#6
tpa-pr
I'm a layman for the Nvidia AI side of things: what's the better OS for AI workloads? Windows or Linux? I know on the AMD side it's Linux because of the ROCm support.
Posted on Reply
#7
10tothemin9volts
Many say that for 3000 $ the DGX Spark (128GB RAM at 273 GB/s memory bandwidth) is DOA/obsolete already:
  • Arm CPU, not x86 CPU, so playing (certain) games is going to be an issue
  • Supposedly requires a specialized NVIDIA Arm OS
  • Can't run serious models like the real DeepSeek-R1 quants, because can't fit even the smallest quant ("DeepSeek-R1-IQ1_S.gguf" is 133.56GB)
  • The overpriced Framework Desktop using AMD Max+ 395 128GB is still 1000 $ cheaper for similar bandwidth (256 GB/s) and has all the advantages of x86 compatibility
  • It's a given that it can't be used for training (maybe models in the millions of parameters, not billions), but even fine-tuning it isn't going to be great with the 273 GB/s memory bandwidth (maybe small models)
  • Soon there are going to be many AMD Max+ 395 128GB x86 mini-PC devices at half the price with all the compatibility advantages that x86 brings
What would be really needed are 256GB RAM devices with, say, at least twice the memory bandwidth (>500GB/s) to be able to run serious LLM like the real DeepSeek-R1 (luckily as a MOE model, it runs faster than dense models at the same size).
Maybe when AMD and NVIDIA designed the 128 GB RAM, 256-bit, quad-channel, SOCs (and to also catch up no APPLE's offerings), DeepSeek-R1 wasn't a thing (though maybe at that time there was Llama-3.1-405B, but as a dense model it would run much slower), but it is now.
Posted on Reply
#8
igormp
Rover4444Should've been RISC-V if they really wanted to be wild with it.
I don't think they wanted to be wild, just have good software support and slap their custom interconnect into a CPU. ARM already has a pretty good software support out there, especially in the python ecosystem.
Vera seems to be a custom core as well, unlike grace which was a neoverse design.
tpa-prI'm a layman for the Nvidia AI side of things: what's the better OS for AI workloads? Windows or Linux? I know on the AMD side it's Linux because of the ROCm support.
Linux for sure. Even on windows Nvidia is recommending people to use WSL instead of trying to do things natively.
10tothemin9voltsMany say that for 3000 $ the DGX Spark (128GB RAM at 273 GB/s memory bandwidth) is DOA/obsolete already:
  • Arm CPU, not x86 CPU, so playing (certain) games is going to be an issue
  • Supposedly requires a specialized NVIDIA Arm OS
  • Can't run serious models like the real DeepSeek-R1 quants, because can't fit even the smallest quant ("DeepSeek-R1-IQ1_S.gguf" is 133.56GB)
  • The overpriced Framework Desktop using AMD Max+ 395 128GB is still 1000 $ cheaper for similar bandwidth (256 GB/s) and has all the advantages of x86 compatibility
  • It's a given that it can't be used for training (maybe models in the millions of parameters, not billions), but even fine-tuning it isn't going to be great with the 273 GB/s memory bandwidth (maybe small models)
  • Soon there are going to be many AMD Max+ 395 128GB x86 mini-PC devices at half the price with all the compatibility advantages that x86 brings
What would be really needed are 256GB RAM devices with, say, at least twice the memory bandwidth (>500GB/s) to be able to run serious LLM like the real DeepSeek-R1 (luckily as a MOE model, it runs faster than dense models at the same size).
Maybe when AMD and NVIDIA designed the 128 GB RAM, 256-bit, quad-channel, SOCs (and to also catch up no APPLE's offerings), DeepSeek-R1 wasn't a thing (though maybe at that time there was Llama-3.1-405B, but as a dense model it would run much slower), but it is now.
Only saving point I see for spark is if you want to cluster multiple of those, given that it has connect-x. Not sure why one would spend that amount of money in such a setup, but it's an option.
Posted on Reply
#9
Rover4444
igormpI don't think they wanted to be wild, just have good software support and slap their custom interconnect into a CPU. ARM already has a pretty good software support out there, especially in the python ecosystem.
Vera seems to be a custom core as well, unlike grace which was a neoverse design.

Linux for sure. Even on windows Nvidia is recommending people to use WSL instead of trying to do things natively.

Only saving point I see for spark is if you want to cluster multiple of those, given that it has connect-x. Not sure why one would spend that amount of money in such a setup, but it's an option.
The networking's totally useless aside from loading/offloading onto a remote machine... which is pretty important when you're loading 100+ GB models, but it's something you only have to do once. For most people it's not worth the price.
Posted on Reply
Mar 19th, 2025 12:26 EDT change timezone

New Forum Posts

Popular Reviews

Controversial News Posts