• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.

NVIDIA Reveals Secret Weapon Behind DLSS Evolution: Dedicated Supercomputer Running for Six Years

AleksandarK

News Editor
Staff member
Joined
Aug 19, 2017
Messages
2,723 (1.01/day)
At the RTX "Blackwell" Editor's Day during CES 2025, NVIDIA pulled back the curtain on one of its most powerful tools: a dedicated supercomputer that has been continuously improving DLSS (Deep Learning Super Sampling) for the past six years. Brian Catanzaro, NVIDIA's VP of applied deep learning research, disclosed that thousands of the company's latest GPUs have been working round-the-clock, analyzing and perfecting the technology that has revolutionized gaming graphics. "We have a big supercomputer at NVIDIA that is running 24/7, 365 days a year improving DLSS," Catanzaro explained during his presentation on DLSS 4. The supercomputer's primary task involves analyzing failures in DLSS performance, such as ghosting, flickering, or blurriness across hundreds of games. When issues are identified, the system augments its training data sets with new examples of optimal graphics and challenging scenarios that DLSS needs to address.

DLSS 4 is the first move from convolutional neural networks to a transformer model that runs locally on client PCs. The continuous learning process has been crucial in refining the technology, with the dedicated supercomputer serving as the backbone of this evolution. The scale of resources allocated to DLSS development is massive, as the entire pipeline for a self-improving DLSS model must consist of not only thousands but tens of thousands of GPUs. Of course, a company making 100,000 GPU data centers (xAI's Colossus) must save some for itself and is proactively using it to improve its software stack. NVIDIA's CEO Jensen Huang famously said that DLSS can predict the future. Of course, these statements are to be tested when the Blackwell series launches. However, the approach of using massive data centers to improve DLSS is quite interesting, and with each new GPU generation NVIDIA release, the process is getting significantly sped up.



View at TechPowerUp Main Site | Source
 

Space Lynx

Astronaut
Joined
Oct 17, 2014
Messages
17,497 (4.67/day)
Location
Kepler-186f
Processor 7800X3D -25 all core
Motherboard B650 Steel Legend
Cooling Frost Commander 140
Video Card(s) Merc 310 7900 XT @3100 core -.75v
Display(s) Agon 27" QD-OLED Glossy 240hz 1440p
Case NZXT H710 (Red/Black)
Audio Device(s) Asgard 2, Modi 3, HD58X
Power Supply Corsair RM850x Gold
I hate it, but it is the future most likely. My next card probably will be Nvidia, even if FSR4 ends up being great, it probably will only be a tiny handful of games. I don't plan to upgrade for a couple years, maybe even three or four. Don't forget your backlog games are fun too lads.
 
Joined
Nov 15, 2010
Messages
153 (0.03/day)
Processor 5700X
Motherboard ASRock Gaming X
Memory 2x8 3600 CL14
Video Card(s) 1080 SEAHAWK
Display(s) ViewSonic XG2703-GS
Tell me how it ends
 
Joined
Jul 24, 2024
Messages
341 (1.92/day)
System Name AM4_TimeKiller
Processor AMD Ryzen 5 5600X @ all-core 4.7 GHz
Motherboard ASUS ROG Strix B550-E Gaming
Cooling Arctic Freezer II 420 rev.7 (push-pull)
Memory G.Skill TridentZ RGB, 2x16 GB DDR4, B-Die, 3800 MHz @ CL14-15-14-29-43 1T, 53.2 ns
Video Card(s) ASRock Radeon RX 7800 XT Phantom Gaming
Storage Samsung 990 PRO 1 TB, Kingston KC3000 1 TB, Kingston KC3000 2 TB
Case Corsair 7000D Airflow
Audio Device(s) Creative Sound Blaster X-Fi Titanium
Power Supply Seasonic Prime TX-850
Mouse Logitech wireless mouse
Keyboard Logitech wireless keyboard
I kind of expected this. Was wondering for long time, how the hell can they advance in DLSS so quickly, meaning how does the model keep improving itself?

A supercomputer purely dedicated to grinding away on images that just aren't quite good enough.

"What we're doing during this process," Catanzaro continues, "is we're analysing failures. When the DLSS model fails it looks like ghosting or flickering or blurriness. And, you know, we find failures in many of the games we're looking at and we try to figure out what's going on, why does the model make the wrong choice about how to draw the image there?

"We then find ways to augment our training data set. Our training data sets are always growing. We're compiling examples of what good graphics looks like and what difficult problems DLSS needs to solve.

"We put those in our training set, and then we retrain the model, and then we test across hundreds of games in order to figure out how to make DLSS better. So, that's the process."
Source: https://www.yahoo.com/tech/turns-theres-big-supercomputer-nvidia-140554206.html

Now it's pretty clear why nowadays middle class RTX costs a kidney.
It's pretty clear why Nvidia so strongly encourages game devs to include DLSS.
 
Last edited:
Joined
Nov 15, 2024
Messages
60 (0.94/day)
I kind of expected this. Was wondering for long time, how the hell can they advance in DLSS so quickly, meaning how does the model keep improving itself? Is there some dedicated algorithm working when DLSS is turned on in games to search for problematic sections/defects in rendered picture?


Source: https://www.yahoo.com/tech/turns-theres-big-supercomputer-nvidia-140554206.html

Now it's pretty clear why nowadays middle class RTX costs a kidney.
It's pretty clear why Nvidia so strongly encourages game devs to include DLSS.
"We can't do computer graphics anymore without artificial intelligence," he said. "We compute one pixel, we infer the other 32. I mean, it's incredible... And so we hallucinate, if you will, the other 32, and it looks temporally stable, it looks photorealistic, and the image quality is incredible, the performance is incredible."

If that saves energy then I'm all for it, but not sure how much the training aspect offsets that. Therein lies the rub...

 
Last edited:
Joined
Nov 11, 2016
Messages
3,517 (1.18/day)
System Name The de-ploughminator Mk-III
Processor 9800X3D
Motherboard Gigabyte X870E Aorus Master
Cooling DeepCool AK620
Memory 2x32GB G.SKill 6400MT Cas32
Video Card(s) Asus RTX4090 TUF
Storage 4TB Samsung 990 Pro
Display(s) 48" LG OLED C4
Case Corsair 5000D Air
Audio Device(s) KEF LSX II LT speakers + KEF KC62 Subwoofer
Power Supply Corsair HX850
Mouse Razor Death Adder v3
Keyboard Razor Huntsman V3 Pro TKL
Software win11
I'm quite interested in how the Transformer model will look vs the old CNN model.

The best thing is that all that advance in visuals can be backported to all games that support DLSS2.x, meaning over 600 games
 

Easy Rhino

Linux Advocate
Staff member
Joined
Nov 13, 2006
Messages
15,632 (2.35/day)
Location
Mid-Atlantic
System Name Desktop
Processor i5 13600KF
Motherboard AsRock B760M Steel Legend Wifi
Cooling Noctua NH-U9S
Memory 4x 16 Gb Gskill S5 DDR5 @6000
Video Card(s) Gigabyte Gaming OC 6750 XT 12GB
Storage WD_BLACK 4TB SN850x
Display(s) Gigabye M32U
Case Corsair Carbide 400C
Audio Device(s) On Board
Power Supply EVGA Supernova 650 P2
Mouse MX Master 3s
Keyboard Logitech G915 Wireless Clicky
Software Fedora KDE Spin
The next time some politician or Hollywood a-hole preaches to you about your carbon footprint and how you can't have a wood stove anymore to heat your home, remind them that a trillion dollar company used untold megawatts training AI so you could get a few extra FPS on a crappy upscaled 4K video game.
 
Joined
Jun 14, 2020
Messages
3,931 (2.34/day)
System Name Mean machine
Processor 12900k
Motherboard MSI Unify X
Cooling Noctua U12A
Memory 7600c34
Video Card(s) 4090 Gamerock oc
Storage 980 pro 2tb
Display(s) Samsung crg90
Case Fractal Torent
Audio Device(s) Hifiman Arya / a30 - d30 pro stack
Power Supply Be quiet dark power pro 1200
Mouse Viper ultimate
Keyboard Blackwidow 65%
I'm quite interested in how the Transformer model will look vs the old CNN model.

The best thing is that all that advance in visuals can be backported to all games that support DLSS2.x, meaning over 600 games
And for gpus as far back as turing. That is wild.
 
Joined
Nov 6, 2016
Messages
1,812 (0.60/day)
Location
NH, USA
System Name Lightbringer
Processor Ryzen 7 2700X
Motherboard Asus ROG Strix X470-F Gaming
Cooling Enermax Liqmax Iii 360mm AIO
Memory G.Skill Trident Z RGB 32GB (8GBx4) 3200Mhz CL 14
Video Card(s) Sapphire RX 5700XT Nitro+
Storage Hp EX950 2TB NVMe M.2, HP EX950 1TB NVMe M.2, Samsung 860 EVO 2TB
Display(s) LG 34BK95U-W 34" 5120 x 2160
Case Lian Li PC-O11 Dynamic (White)
Power Supply BeQuiet Straight Power 11 850w Gold Rated PSU
Mouse Glorious Model O (Matte White)
Keyboard Royal Kludge RK71
Software Windows 10
I think one of the biggest issues here, and I never see it mentioned, is that implementations like the one Nvidia is using detailed in this article, basically create a runaway effect with respect to competition. Nvidia is basically leveraging it's unparalleled resources to exponentially accelerate the gulf between them and AMD/Intel, and because their competitors do not have that level of resources, they will perpetually be "behind" and they'll never be able to catch up.

Now, what I'm about to say could be wrong, I don't have expert knowledge of how GPUs are designed, but it seemed like 20 years ago, if you had a brilliant individual or a few of them, you could compete because in the end, every company is more or less working with and is limited by, the same tool: the human brain.

Now, with machine learning and AI, that limitation has been breached, and it has basically turned into an arms race with who can amass the most compute power. In a reality wholly shaped by the dictates of capitalism and the profit motive, the competition has basically been reduced to who can buy the most hardware. It then basically turns into a positive feedback loop: Nvidia has the most resources so they have access to more compute power, this compute power let's them create faster products, the products sell more and Nvidia gets more resources....repeat. With the use of AI/ML in the design process, I feel like Nvidia has literally gained an insurmountable advantage and it will never be "corrected" by market forces.
 
Joined
Nov 2, 2016
Messages
130 (0.04/day)
I've been wondering for a while if eventually every game will be an AI model running on the GPU that takes the player's input as a "prompt" and outputs the game's visuals realtime. Eventually everything could be procedurally generated, assuming AI will be able to write a captivating story as opposed to just outputting visuals. Just take the "one to rule them all" model and prompt it for what the game should be like.
 
Joined
Dec 12, 2016
Messages
2,082 (0.70/day)
Yawn. Wake me when the singularity happens. In the meantime, I’ll be playing Broforce in my dreams and having fun.

Nvidia is just way too serious when it comes to gaming.
 
Joined
Nov 15, 2024
Messages
60 (0.94/day)
I've been wondering for a while if eventually every game will be an AI model running on the GPU that takes the player's input as a "prompt" and outputs the game's visuals realtime. Eventually everything could be procedurally generated, assuming AI will be able to write a captivating story as opposed to just outputting visuals. Just take the "one to rule them all" model and prompt it for what the game should be like.

I think one of the biggest issues here, and I never see it mentioned, is that implementations like the one Nvidia is using detailed in this article, basically create a runaway effect with respect to competition. Nvidia is basically leveraging it's unparalleled resources to exponentially accelerate the gulf between them and AMD/Intel, and because their competitors do not have that level of resources, they will perpetually be "behind" and they'll never be able to catch up.

Now, what I'm about to say could be wrong, I don't have expert knowledge of how GPUs are designed, but it seemed like 20 years ago, if you had a brilliant individual or a few of them, you could compete because in the end, every company is more or less working with and is limited by, the same tool: the human brain.

Now, with machine learning and AI, that limitation has been breached, and it has basically turned into an arms race with who can amass the most compute power. In a reality wholly shaped by the dictates of capitalism and the profit motive, the competition has basically been reduced to who can buy the most hardware. It then basically turns into a positive feedback loop: Nvidia has the most resources so they have access to more compute power, this compute power let's them create faster products, the products sell more and Nvidia gets more resources....repeat. With the use of AI/ML in the design process, I feel like Nvidia has literally gained an insurmountable advantage and it will never be "corrected" by market forces.
They weren't happy with Biden for a reason. Not sure if "ethics washing" is a term but they certainly played the humanity card, although in fairness I did read the argument about Leather being better for the environment rather than pleather, so there's that at least.
 
Joined
Jun 14, 2020
Messages
3,931 (2.34/day)
System Name Mean machine
Processor 12900k
Motherboard MSI Unify X
Cooling Noctua U12A
Memory 7600c34
Video Card(s) 4090 Gamerock oc
Storage 980 pro 2tb
Display(s) Samsung crg90
Case Fractal Torent
Audio Device(s) Hifiman Arya / a30 - d30 pro stack
Power Supply Be quiet dark power pro 1200
Mouse Viper ultimate
Keyboard Blackwidow 65%
I think one of the biggest issues here, and I never see it mentioned, is that implementations like the one Nvidia is using detailed in this article, basically create a runaway effect with respect to competition. Nvidia is basically leveraging it's unparalleled resources to exponentially accelerate the gulf between them and AMD/Intel, and because their competitors do not have that level of resources, they will perpetually be "behind" and they'll never be able to catch up.
Nothing stops the competition from releasing a 450$ card that annihilates eg. a 5080 in raster. Especially considering how anemic nvidias last 2 gens have been, i'd say a 450$ card that beats the 5080 in raster is not even something to write home about. It should be trivial
 
Joined
Jul 24, 2024
Messages
341 (1.92/day)
System Name AM4_TimeKiller
Processor AMD Ryzen 5 5600X @ all-core 4.7 GHz
Motherboard ASUS ROG Strix B550-E Gaming
Cooling Arctic Freezer II 420 rev.7 (push-pull)
Memory G.Skill TridentZ RGB, 2x16 GB DDR4, B-Die, 3800 MHz @ CL14-15-14-29-43 1T, 53.2 ns
Video Card(s) ASRock Radeon RX 7800 XT Phantom Gaming
Storage Samsung 990 PRO 1 TB, Kingston KC3000 1 TB, Kingston KC3000 2 TB
Case Corsair 7000D Airflow
Audio Device(s) Creative Sound Blaster X-Fi Titanium
Power Supply Seasonic Prime TX-850
Mouse Logitech wireless mouse
Keyboard Logitech wireless keyboard
I think one of the biggest issues here, and I never see it mentioned, is that implementations like the one Nvidia is using detailed in this article, basically create a runaway effect with respect to competition. Nvidia is basically leveraging it's unparalleled resources to exponentially accelerate the gulf between them and AMD/Intel, and because their competitors do not have that level of resources, they will perpetually be "behind" and they'll never be able to catch up.

Now, what I'm about to say could be wrong, I don't have expert knowledge of how GPUs are designed, but it seemed like 20 years ago, if you had a brilliant individual or a few of them, you could compete because in the end, every company is more or less working with and is limited by, the same tool: the human brain.

Now, with machine learning and AI, that limitation has been breached, and it has basically turned into an arms race with who can amass the most compute power. In a reality wholly shaped by the dictates of capitalism and the profit motive, the competition has basically been reduced to who can buy the most hardware. It then basically turns into a positive feedback loop: Nvidia has the most resources so they have access to more compute power, this compute power let's them create faster products, the products sell more and Nvidia gets more resources....repeat. With the use of AI/ML in the design process, I feel like Nvidia has literally gained an insurmountable advantage and it will never be "corrected" by market forces.
Yes but you can't blame Nvidia for re-applying their success from elsewhere. They used their own tech to their advantage.
Nvidia's success is heavily dependent on TSMC's success. Maybe it will get corrected through there if you know what I mean.
 
Joined
Aug 31, 2021
Messages
22 (0.02/day)
Processor Ryzen 9 5900X
Motherboard Gigabyte Aorus B550i pro ax
Cooling Noctua NH-D15 chromax.black (with original fans)
Memory G.skill 32GB 3200MHz
Video Card(s) 4060ti 16GB
Storage 1TB Samsung PM9A1, 256GB Toshiba pcie3 (from a laptop), 512GB crucial MX500, 2x 1TB Toshiba HDD 2.5
Display(s) Mateview GT 34''
Case Thermaltake the tower 100, 1 noctua NF-A14 ippc3000 on top, 2 x Arctic F14
Power Supply Seasonic focus-GX 750W
Software Windows 10, Ubuntu when needed.
Now, what I'm about to say could be wrong, I don't have expert knowledge of how GPUs are designed, but it seemed like 20 years ago, if you had a brilliant individual or a few of them, you could compete because in the end, every company is more or less working with and is limited by, the same tool: the human brain.
I strongly disagree on that you said. IMO things haven't change, it was almost impossible to compete before and it is almost impossible to compete now. If what you say is correct, then where are the dozens of intel competitors back in the 2000? There were not, because it is not a matter of having "brilliant individuals" to design the thing, it has always been about the know how and data that you have on the thing. And unfortunately intel had (and has) 30 years of testing methodology for their processors such that the code you wrote it is guaranteed to do the same thing on their different CPUs. Try design another CPU with the same ISA than intel and prove that it is doing the right computation, good luck with that! And what is even funnier is that the same is true even if you redesign everything from zero. Try developing a CPU from scratch, even with your own libraries and code, and convince people that the thing have been tested and it is guaranteed that the code they write (even corner cases) are correctly handled. And there is no AI (so far) in there.
 

Cybebe

New Member
Joined
Jan 17, 2025
Messages
1 (1.00/day)
with all the discussions and blablabla, the only thing that hurts is the final price of the product. Ngreedia knowing their current position as the leader in AI and DLSS is cashing on this advantage, knowing that no one in the industry right now can challenge them. Well AMD has better graphics cards generation 7000 in terms of rasterization but what happened?? hypocrites turning on DLSS and FG for their games and letting the public know how FSR sucks
 
Joined
Jan 9, 2025
Messages
5 (0.56/day)
Processor AMD Ryzen 7 5800X3D
Motherboard MSI MPG B550 Gaming Plus
Cooling Deepcool LT720 360mm
Memory 32GB Corsair Dominator Platinum DDR4 3600Mhz
Video Card(s) Zotac Nvidia RTX 3090
Storage Western Digital SN850X
Display(s) Acer Nitro 34" Ultrawide
Case Lianli O11 Dynamic Evo White
Audio Device(s) HiVi Swans OS-10 Bookshelf Speakers
Power Supply Seasonic Prime GX-1000 80+ Gold
Mouse Razer Ultimate | Razer Naga Pro
Keyboard Razer Huntsman Tournament Edition
The next time some politician or Hollywood a-hole preaches to you about your carbon footprint and how you can't have a wood stove anymore to heat your home, remind them that a trillion dollar company used untold megawatts training AI so you could get a few extra FPS on a crappy upscaled 4K video game.
What do you mean by "a wood stove"? My 380w graphics card is all the heating I ever need for my room!
 
Joined
Jun 12, 2023
Messages
66 (0.11/day)
I'm quite interested in how the Transformer model will look vs the old CNN model.

The best thing is that all that advance in visuals can be backported to all games that support DLSS2.x, meaning over 600 games
"Can" is imperative here. Is it really going to be made available for those older models of DLSS? Or is NVIDIA going to lock them behind newer versions to taut the newest hardware?

I think one of the biggest issues here, and I never see it mentioned, is that implementations like the one Nvidia is using detailed in this article, basically create a runaway effect with respect to competition. Nvidia is basically leveraging it's unparalleled resources to exponentially accelerate the gulf between them and AMD/Intel, and because their competitors do not have that level of resources, they will perpetually be "behind" and they'll never be able to catch up.

Now, what I'm about to say could be wrong, I don't have expert knowledge of how GPUs are designed, but it seemed like 20 years ago, if you had a brilliant individual or a few of them, you could compete because in the end, every company is more or less working with and is limited by, the same tool: the human brain.

Now, with machine learning and AI, that limitation has been breached, and it has basically turned into an arms race with who can amass the most compute power. In a reality wholly shaped by the dictates of capitalism and the profit motive, the competition has basically been reduced to who can buy the most hardware. It then basically turns into a positive feedback loop: Nvidia has the most resources so they have access to more compute power, this compute power let's them create faster products, the products sell more and Nvidia gets more resources....repeat. With the use of AI/ML in the design process, I feel like Nvidia has literally gained an insurmountable advantage and it will never be "corrected" by market forces.
I agree that it's a runaway cycle. Even if some genius were to create a new method of coding that made rasterization faster, easier, or more efficient; Nvidia would still gain by using their current method. I'm a proponent of regulation for fairness but we can't impose a limit in how Nvidia manages their business to improve their products. In some respects, they should be praised for seeing the writing on the wall years ago about the impending physical limitations in hardware improvements.

Other than maybe a blanket tax on any company that uses exorbitant amounts of energy with no offsets, I don't see what could be done. There would also have to be exceptions such as for steel mills. Electric Arc furnaces are amazing for recycling steel but they are so energy intensive.

Yes but you can't blame Nvidia for re-applying their success from elsewhere. They used their own tech to their advantage.
Nvidia's success is heavily dependent on TSMC's success. Maybe it will get corrected through there if you know what I mean.
No, I don't know what you mean. Please elaborate lol
 

Chrisy

New Member
Joined
Jan 8, 2025
Messages
3 (0.30/day)
I am wondering if running the optimization locally means that the driver will upload these optimizations to Nvidia servers for aggregating the improvements.
Which effectively would be kind of moving the calculations they had to manage within their infrastructure to the clients now.
That would be a lot of off-loading (time and money saving) for Nvidia. Not sure if clever in evil-way or not :D
 
Joined
Sep 17, 2014
Messages
22,914 (6.07/day)
Location
The Washing Machine
System Name Tiny the White Yeti
Processor 7800X3D
Motherboard MSI MAG Mortar b650m wifi
Cooling CPU: Thermalright Peerless Assassin / Case: Phanteks T30-120 x3
Memory 32GB Corsair Vengeance 30CL6000
Video Card(s) ASRock RX7900XT Phantom Gaming
Storage Lexar NM790 4TB + Samsung 850 EVO 1TB + Samsung 980 1TB + Crucial BX100 250GB
Display(s) Gigabyte G34QWC (3440x1440)
Case Lian Li A3 mATX White
Audio Device(s) Harman Kardon AVR137 + 2.1
Power Supply EVGA Supernova G2 750W
Mouse Steelseries Aerox 5
Keyboard Lenovo Thinkpad Trackpoint II
VR HMD HD 420 - Green Edition ;)
Software W11 IoT Enterprise LTSC
Benchmark Scores Over 9000
I think one of the biggest issues here, and I never see it mentioned, is that implementations like the one Nvidia is using detailed in this article, basically create a runaway effect with respect to competition. Nvidia is basically leveraging it's unparalleled resources to exponentially accelerate the gulf between them and AMD/Intel, and because their competitors do not have that level of resources, they will perpetually be "behind" and they'll never be able to catch up.

Now, what I'm about to say could be wrong, I don't have expert knowledge of how GPUs are designed, but it seemed like 20 years ago, if you had a brilliant individual or a few of them, you could compete because in the end, every company is more or less working with and is limited by, the same tool: the human brain.

Now, with machine learning and AI, that limitation has been breached, and it has basically turned into an arms race with who can amass the most compute power. In a reality wholly shaped by the dictates of capitalism and the profit motive, the competition has basically been reduced to who can buy the most hardware. It then basically turns into a positive feedback loop: Nvidia has the most resources so they have access to more compute power, this compute power let's them create faster products, the products sell more and Nvidia gets more resources....repeat. With the use of AI/ML in the design process, I feel like Nvidia has literally gained an insurmountable advantage and it will never be "corrected" by market forces.
But... are they that far behind? XeSS is pretty good and FSR strongly improved, without harping about all this.
 
Joined
Nov 15, 2024
Messages
60 (0.94/day)
But... are they that far behind? XeSS is pretty good and FSR strongly improved, without harping about all this.
I wondering how far away we are from the point of it being "good enough" for most people. Diminishing returns and all that. The "Reasoning Era" being the new paradigm at OpenAI.
 
Top