Monday, January 6th 2025

NVIDIA Puts Grace Blackwell on Every Desk and at Every AI Developer's Fingertips

NVIDIA today unveiled NVIDIA Project DIGITS, a personal AI supercomputer that provides AI researchers, data scientists and students worldwide with access to the power of the NVIDIA Grace Blackwell platform. Project DIGITS features the new NVIDIA GB10 Grace Blackwell Superchip, offering a petaflop of AI computing performance for prototyping, fine-tuning and running large AI models.

With Project DIGITS, users can develop and run inference on models using their own desktop system, then seamlessly deploy the models on accelerated cloud or data center infrastructure. "AI will be mainstream in every application for every industry. With Project DIGITS, the Grace Blackwell Superchip comes to millions of developers," said Jensen Huang, founder and CEO of NVIDIA. "Placing an AI supercomputer on the desks of every data scientist, AI researcher and student empowers them to engage and shape the age of AI."
GB10 Superchip Provides a Petaflop of Power-Efficient AI Performance
The GB10 Superchip is a system-on-a-chip (SoC) based on the NVIDIA Grace Blackwell architecture and delivers up to 1 petaflop of AI performance at FP4 precision.

GB10 features an NVIDIA Blackwell GPU with latest-generation CUDA cores and fifth-generation Tensor Cores, connected via NVLink -C2C chip-to-chip interconnect to a high-performance NVIDIA Grace CPU, which includes 20 power-efficient cores built with the Arm architecture. MediaTek, a market leader in Arm-based SoC designs, collaborated on the design of GB10, contributing to its best-in-class power efficiency, performance and connectivity.

The GB10 Superchip enables Project DIGITS to deliver powerful performance using only a standard electrical outlet. Each Project DIGITS features 128 GB of unified, coherent memory and up to 4 TB of NVMe storage. With the supercomputer, developers can run up to 200-billion-parameter large language models to supercharge AI innovation. In addition, using NVIDIA ConnectX networking, two Project DIGITS AI supercomputers can be linked to run up to 405-billion-parameter models.

Grace Blackwell AI Supercomputing Within Reach
With the Grace Blackwell architecture, enterprises and researchers can prototype, fine-tune and test models on local Project DIGITS systems running Linux-based NVIDIA DGX OS, and then deploy them seamlessly on NVIDIA DGX Cloud, accelerated cloud instances or data center infrastructure.

This allows developers to prototype AI on Project DIGITS and then scale on cloud or data center infrastructure, using the same Grace Blackwell architecture and the NVIDIA AI Enterprise software platform.

Project DIGITS users can access an extensive library of NVIDIA AI software for experimentation and prototyping, including software development kits, orchestration tools, frameworks and models available in the NVIDIA NGC catalog and on the NVIDIA Developer portal. Developers can fine-tune models with the NVIDIA NeMo framework, accelerate data science with NVIDIA RAPIDS libraries and run common frameworks such as PyTorch, Python and Jupyter notebooks.

To build agentic AI applications, users can also harness NVIDIA Blueprints and NVIDIA NIM microservices, which are available for research, development and testing via the NVIDIA Developer Program. When AI applications are ready to move from experimentation to production environments, the NVIDIA AI Enterprise license provides enterprise-grade security, support and product releases of NVIDIA AI software.

Availability
Project DIGITS will be available in May from NVIDIA and top partners, starting at $3,000. Sign up for notifications today.
Add your own comment

13 Comments on NVIDIA Puts Grace Blackwell on Every Desk and at Every AI Developer's Fingertips

#1
JohH
I wonder if something like this will be marketed for Windows too.
Something like a RTX 5070 + 16 ARM cores would be pretty killer. But probably too low margin for Nvidia to market, maybe they'll let MediaTek do it.
Posted on Reply
#2
john_
Mediatek in a $3000+ system.......
Nvidia can make Mediatek look like the hi end brand in the ARM ecosystem.
Posted on Reply
#3
hsew
john_Mediatek in a $3000+ system.......
Nvidia can make Mediatek look like the hi end brand in the ARM ecosystem.
I wouldn't be surprised if nVidia bought Mediatek soon.
JohHI wonder if something like this will be marketed for Windows too.
Something like a RTX 5070 + 16 ARM cores would be pretty killer. But probably too low margin for Nvidia to market, maybe they'll let MediaTek do it.
After a certain point (probably $2500), (V)RAM capacity becomes significantly more important than TOPS. So unless that 5070 was also outfitted with 128GB of G7 I doubt there would be a market for it. If you're set on running a 5070 you'd want to install it into a system like a LattePanda Mu + carrier board, which also has the advantage of supporting every major OS available today.

www.dfrobot.com/kit-004.html?tracking=LATTEPANDAMU
Posted on Reply
#4
igormp
Dang, I was hoping that it'd be cheaper than $2k, guess I won't be getting one :(
Still a reasonable price given that it's cheaper than an apple equivalent.
john_Mediatek in a $3000+ system.......
Nvidia can make Mediatek look like the hi end brand in the ARM ecosystem.
The product at hand has a Grace CPU, not a mediatek one. They helped with the SoC design and likely fabbing, but it's still a neoverse CPU under the hood.
hsewI wouldn't be surprised if nVidia bought Mediatek soon.



After a certain point (probably $2500), (V)RAM capacity becomes significantly more important than TOPS. So unless that 5070 was also outfitted with 128GB of G7 I doubt there would be a market for it. If you're set on running a 5070 you'd want to install it into a system like a LattePanda Mu + carrier board, which also has the advantage of supporting every major OS available today.

www.dfrobot.com/kit-004.html?tracking=LATTEPANDAMU
A GB205-like chip with a bigger memory bus and unified LPDDR5X memory wouldn't be bad at all, it'd be similar to Strix Halo even.
Instslling a regular GPU in an x86 SBCs kills all the fun of it, and lacks the unified memory.
Posted on Reply
#5
JohH
hsewAfter a certain point (probably $2500), (V)RAM capacity becomes significantly more important than TOPS. So unless that 5070 was also outfitted with 128GB of G7 I doubt there would be a market for it. If you're set on running a 5070 you'd want to install it into a system like a LattePanda Mu + carrier board, which also has the advantage of supporting every major OS available today.
?? I mean for real human use like an Apple M4. With a normal amount of memory. Not AI stuff. A NVlink C2C connection of MediaTek SoC with Nvidia GPU.
Posted on Reply
#6
Dr. Dro
igormpDang, I was hoping that it'd be cheaper than $2k, guess I won't be getting one :(
Still a reasonable price given that it's cheaper than an apple equivalent.

The product at hand has a Grace CPU, not a mediatek one. They helped with the SoC design and likely fabbing, but it's still a neoverse CPU under the hood.

A GB205-like chip with a bigger memory bus and unified LPDDR5X memory wouldn't be bad at all, it'd be similar to Strix Halo even.
Instslling a regular GPU in an x86 SBCs kills all the fun of it, and lacks the unified memory.
I mean, you work with this stuff. Wait for the release, see if the performance is really all that it's cracked up to be, and perhaps even at this 50% higher price it might still make sense. 128 GB of unified memory and 4 TB SSD contribute to the price a lot. It's possible they could come up with lower end models that have 64 GB of RAM and say, 2 TB SSD instead for less cash, though I'm not sure it'd be worth it.

I'm of the opinion that for a good craftsman, a good tool is always worth the price, it might cost more up front but if you can recover the costs by creating better, or more products as a result, why not.
Posted on Reply
#7
igormp
Dr. DroI mean, you work with this stuff. Wait for the release, see if the performance is really all that it's cracked up to be, and perhaps even at this 50% higher price it might still make sense. 128 GB of unified memory and 4 TB SSD contribute to the price a lot. It's possible they could come up with lower end models that have 64 GB of RAM and say, 2 TB SSD instead for less cash, though I'm not sure it'd be worth it.

I'm of the opinion that for a good craftsman, a good tool is always worth the price, it might cost more up front but if you can recover the costs by creating better, or more products as a result, why not.
But at this price it won't make sense whatsoever for me.
I can rent a GH200 with 96GB of vram and 480GB of RAM with C2C for $1.5/hour. That Digits won't be anywhere close in performance nor that useful to me to justify that price. Same reason as I haven't upgraded my 3090s.
At $2k I'd be more willing to grab one just because it's a fun device that I'd like to tinker with, so it'd be more of a hobby excuse than actual utility.
Posted on Reply
#8
hsew
JohH?? I mean for real human use like an Apple M4. With a normal about of memory. Not AI stuff. A NVlink C2C connection of MediaTek SoC with Nvidia GPU.
Oh sorry, been thinking in terms of AI workloads for everything these days.

It's doubtful that nVidia feels motivated to grab more consumer market share these days, seeing as they already own 90% of it where they compete and it is barely worth 1/30th of their enterprise division. And you know they're closely watching Qualcomm struggle with its own x86 competitor ambitions.
Posted on Reply
#9
mb194dc
They'll need to, maybe will actually find a decent front end application to go with the mountain of bullshit.
Posted on Reply
#10
SIGSEGV
I really hate their closed ecosystem. It's like a cancer.
Posted on Reply
#11
SOAREVERSOR
SIGSEGVI really hate their closed ecosystem. It's like a cancer.
CUDA is so dominant now they've won. For all intents and purposes nvidia is CUDA and made that decision with the 8800 GTX.
igormpBut at this price it won't make sense whatsoever for me.
I can rent a GH200 with 96GB of vram and 480GB of RAM with C2C for $1.5/hour. That Digits won't be anywhere close in performance nor that useful to me to justify that price. Same reason as I haven't upgraded my 3090s.
At $2k I'd be more willing to grab one just because it's a fun device that I'd like to tinker with, so it'd be more of a hobby excuse than actual utility.
Meh do what I do. Sell the 3090s to another CUDA person and pick up some 5090s and off you go.
Posted on Reply
#12
Visible Noise
They are going to sell these by the truckload to educational institutions.

Genius idea.
Posted on Reply
#13
igormp
SOAREVERSORMeh do what I do. Sell the 3090s to another CUDA person and pick up some 5090s and off you go.
Not many people willing to buy one here, and I'd likely only be able to sell mine for ~$600 whereas a single 5090 will cost ~$3.3k here.
I don't think it's worth it to let go of over $5k grand compared to the cost x benefit of a gh200 (that still has more vram).
Posted on Reply
Add your own comment
Jan 8th, 2025 02:24 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts