• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.

NVIDIA Reveals Secret Weapon Behind DLSS Evolution: Dedicated Supercomputer Running for Six Years

Joined
Sep 17, 2014
Messages
22,945 (6.07/day)
Location
The Washing Machine
System Name Tiny the White Yeti
Processor 7800X3D
Motherboard MSI MAG Mortar b650m wifi
Cooling CPU: Thermalright Peerless Assassin / Case: Phanteks T30-120 x3
Memory 32GB Corsair Vengeance 30CL6000
Video Card(s) ASRock RX7900XT Phantom Gaming
Storage Lexar NM790 4TB + Samsung 850 EVO 1TB + Samsung 980 1TB + Crucial BX100 250GB
Display(s) Gigabyte G34QWC (3440x1440)
Case Lian Li A3 mATX White
Audio Device(s) Harman Kardon AVR137 + 2.1
Power Supply EVGA Supernova G2 750W
Mouse Steelseries Aerox 5
Keyboard Lenovo Thinkpad Trackpoint II
VR HMD HD 420 - Green Edition ;)
Software W11 IoT Enterprise LTSC
Benchmark Scores Over 9000
I am wondering if running the optimization locally means that the driver will upload these optimizations to Nvidia servers for aggregating the improvements.
Which effectively would be kind of moving the calculations they had to manage within their infrastructure to the clients now.
That would be a lot of off-loading (time and money saving) for Nvidia. Not sure if clever in evil-way or not :D
Well yeah, then you are literally mining for Nvidia :D They're not going to pay you dividends, so that would be dirty AF.

But of course they'll try to sell that differently. You're helping games improve! :p

I think one of the biggest issues here, and I never see it mentioned, is that implementations like the one Nvidia is using detailed in this article, basically create a runaway effect with respect to competition. Nvidia is basically leveraging it's unparalleled resources to exponentially accelerate the gulf between them and AMD/Intel, and because their competitors do not have that level of resources, they will perpetually be "behind" and they'll never be able to catch up.

Now, what I'm about to say could be wrong, I don't have expert knowledge of how GPUs are designed, but it seemed like 20 years ago, if you had a brilliant individual or a few of them, you could compete because in the end, every company is more or less working with and is limited by, the same tool: the human brain.

Now, with machine learning and AI, that limitation has been breached, and it has basically turned into an arms race with who can amass the most compute power. In a reality wholly shaped by the dictates of capitalism and the profit motive, the competition has basically been reduced to who can buy the most hardware. It then basically turns into a positive feedback loop: Nvidia has the most resources so they have access to more compute power, this compute power let's them create faster products, the products sell more and Nvidia gets more resources....repeat. With the use of AI/ML in the design process, I feel like Nvidia has literally gained an insurmountable advantage and it will never be "corrected" by market forces.
Nah, conceptual things only get discovered once, and then we all copy them, and that's that. Look at FSR4. That's also why its folly to be paying for proprietary bullshit. Just wait. It'll come. And if it won't, it simply will not survive.
 
Joined
Dec 25, 2020
Messages
7,332 (4.93/day)
Location
São Paulo, Brazil
System Name "Icy Resurrection"
Processor 13th Gen Intel Core i9-13900KS
Motherboard ASUS ROG Maximus Z790 Apex Encore
Cooling Noctua NH-D15S upgraded with 2x NF-F12 iPPC-3000 fans and Honeywell PTM7950 TIM
Memory 32 GB G.SKILL Trident Z5 RGB F5-6800J3445G16GX2-TZ5RK @ 7600 MT/s 36-44-44-52-96 1.4V
Video Card(s) NVIDIA RTX A2000
Storage 500 GB WD Black SN750 SE NVMe SSD + 4 TB WD Red Plus WD40EFPX HDD
Display(s) 55-inch LG G3 OLED
Case Pichau Mancer CV500 White Edition
Power Supply EVGA 1300 G2 1.3kW 80+ Gold
Mouse Microsoft Classic IntelliMouse (2017)
Keyboard IBM Model M type 1391405
Software Windows 10 Pro 22H2
Benchmark Scores I pulled a Qiqi~
Maybe AMD will run GPT-4 on a supercomputer to write them some drivers :laugh:
 
Joined
Mar 11, 2024
Messages
91 (0.29/day)
I've been wondering for a while if eventually every game will be an AI model running on the GPU that takes the player's input as a "prompt" and outputs the game's visuals realtime. Eventually everything could be procedurally generated, assuming AI will be able to write a captivating story as opposed to just outputting visuals. Just take the "one to rule them all" model and prompt it for what the game should be like.
The required processing power would be madness but if AI eventually surpass hupans in every area...
 
Joined
Jan 11, 2022
Messages
1,009 (0.91/day)
I've been wondering for a while if eventually every game will be an AI model running on the GPU that takes the player's input as a "prompt" and outputs the game's visuals realtime. Eventually everything could be procedurally generated, assuming AI will be able to write a captivating story as opposed to just outputting visuals. Just take the "one to rule them all" model and prompt it for what the game should be like.
maybe there will be some further evolution, but it’s become pretty clear what was written by ai and what was written by a real person after being exposed To it for a good while now.
especially if it’s about a topic I’m at home at.

i think it’s going to become very bland very quickly playing ai generated stories.
 
Joined
Aug 22, 2007
Messages
3,616 (0.57/day)
Location
Terra
System Name :)
Processor Intel 13700k
Motherboard Gigabyte z790 UD AC
Cooling Noctua NH-D15
Memory 64GB GSKILL DDR5
Video Card(s) Gigabyte RTX 4090 Gaming OC
Storage 960GB Optane 905P U.2 SSD + 4TB PCIe4 U.2 SSD
Display(s) Alienware AW3423DW 175Hz QD-OLED + AOC Agon Pro AG276QZD2 240Hz QD-OLED
Case Fractal Design Torrent
Audio Device(s) MOTU M4 - JBL 305P MKII w/2x JL Audio 10 Sealed --- X-Fi Titanium HD - Presonus Eris E5 - JBL 4412
Power Supply Silverstone 1000W
Mouse Roccat Kain 122 AIMO
Keyboard KBD67 Lite / Mammoth75
VR HMD Reverb G2 V2
Software Win 11 Pro
I thought people already knew this.... like when RTX/DLSS came out they mentioned that it was trained to do this/that. What do people think trained that rendering network? a Raspberry Pi? :kookoo:
 

Rightness_1

New Member
Joined
Dec 30, 2024
Messages
26 (1.18/day)
If A.I. was so great, why is it not done in the GPU in real-time? And why is some "supercomputer" allegedly doing it offline? I just don't believe this approach is completely necessary in 2025. It's just marketing lies to tie DLSS to nv hardware.

FSR 4.0 is going to be very interesting.
 
Joined
Dec 25, 2020
Messages
7,332 (4.93/day)
Location
São Paulo, Brazil
System Name "Icy Resurrection"
Processor 13th Gen Intel Core i9-13900KS
Motherboard ASUS ROG Maximus Z790 Apex Encore
Cooling Noctua NH-D15S upgraded with 2x NF-F12 iPPC-3000 fans and Honeywell PTM7950 TIM
Memory 32 GB G.SKILL Trident Z5 RGB F5-6800J3445G16GX2-TZ5RK @ 7600 MT/s 36-44-44-52-96 1.4V
Video Card(s) NVIDIA RTX A2000
Storage 500 GB WD Black SN750 SE NVMe SSD + 4 TB WD Red Plus WD40EFPX HDD
Display(s) 55-inch LG G3 OLED
Case Pichau Mancer CV500 White Edition
Power Supply EVGA 1300 G2 1.3kW 80+ Gold
Mouse Microsoft Classic IntelliMouse (2017)
Keyboard IBM Model M type 1391405
Software Windows 10 Pro 22H2
Benchmark Scores I pulled a Qiqi~
If A.I. was so great, why is it not done in the GPU in real-time? And why is some "supercomputer" allegedly doing it offline? I just don't believe this approach is completely necessary in 2025. It's just marketing lies to tie DLSS to nv hardware.

FSR 4.0 is going to be very interesting.

>Posts in here trashing AI and discrediting the work of engineers with decades of experience developing an incredibly complex transformative algorithm because you don't like the company which did it
>Proceeds to praise an attempt at doing the same thing from the competitor because you like them

The hallmark of the average AMD fan
 
Joined
May 10, 2023
Messages
522 (0.84/day)
Location
Brazil
Processor 5950x
Motherboard B550 ProArt
Cooling Fuma 2
Memory 4x32GB 3200MHz Corsair LPX
Video Card(s) 2x RTX 3090
Display(s) LG 42" C2 4k OLED
Power Supply XPG Core Reactor 850W
Software I use Arch btw
If A.I. was so great, why is it not done in the GPU in real-time? And why is some "supercomputer" allegedly doing it offline? I just don't believe this approach is completely necessary in 2025. It's just marketing lies to tie DLSS to nv hardware.

FSR 4.0 is going to be very interesting.
Training is different from inference.
The "real-time" part that runs on the gpu is what's called inference, where you run a model that was trained to do something.
The training part often takes way longer, and you need to keep iterating on it as time goes in order to improve its performance over time.
 
Joined
Dec 29, 2020
Messages
220 (0.15/day)
Throwing supercomputers at the problem for years, now that's what I call brute-force rendering.
 
Joined
Sep 3, 2019
Messages
3,751 (1.91/day)
Location
Thessaloniki, Greece
System Name PC on since Aug 2019, 1st CPU R5 3600 + ASUS ROG RX580 8GB >> MSI Gaming X RX5700XT (Jan 2020)
Processor Ryzen 9 5900X (July 2022), 200W PPT limit, 80C temp limit, CO -6-14, +50MHz (up to 5.0GHz)
Motherboard Gigabyte X570 Aorus Pro (Rev1.0), BIOS F39b, AGESA V2 1.2.0.C
Cooling Arctic Liquid Freezer II 420mm Rev7 (Jan 2024) with off-center mount for Ryzen, TIM: Kryonaut
Memory 2x16GB G.Skill Trident Z Neo GTZN (July 2022) 3600MT/s 1.38V CL16-16-16-16-32-48 1T, tRFC:280, B-die
Video Card(s) Sapphire Nitro+ RX 7900XTX (Dec 2023) 314~467W (382W current) PowerLimit, 1060mV, Adrenalin v24.12.1
Storage Samsung NVMe: 980Pro 1TB(OS 2022), 970Pro 512GB(2019) / SATA-III: 850Pro 1TB(2015) 860Evo 1TB(2020)
Display(s) Dell Alienware AW3423DW 34" QD-OLED curved (1800R), 3440x1440 144Hz (max 175Hz) HDR400/1000, VRR on
Case None... naked on desk
Audio Device(s) Astro A50 headset
Power Supply Corsair HX750i, ATX v2.4, 80+ Platinum, 93% (250~700W), modular, single/dual rail (switch)
Mouse Logitech MX Master (Gen1)
Keyboard Logitech G15 (Gen2) w/ LCDSirReal applet
Software Windows 11 Home 64bit (v24H2, OSBuild 26100.2605), upgraded from Win10 to Win11 on Jan 2024
Humans are using AI for many many years now in a lot of aspects. From general research to more specific like medical research, physics, astrophysics, chemistry, even quantum mechanics and so on.

It just so far the capabilities of AI was limited by hardware running the models.
Now we have come to the point that AI can use the “new” current hardware to improve everything. Models, algorithms, data acquisition and even it self. Large supercomputers are needed still just like 40-50 years ago in order to have the compute power that now we all have in our pockets.

The real breakthrough will come once compute power grows exponentially. Something else than today’s silicon based chips. The wall of improvement is coming fast and for next years AI/ML alone is the way around it with these “new” hardware that will keep growing in quantities.

Quantum computers is one strong candidate for replacing current tech. Just one of them will be able to replace entire rooms of servers. But it’s not a tech for personal/individual usage as it requires isolated near absolute zero conditions to work properly without any “outside” particle interference.
This type of compute power will improve AI vastly.
At some point, after them, almost everything will be cloud serviced. In the mean time IoT is going to continue to evolve as it is required as the base infrastructure for all this to happen.
One example is that cars will be all connected to each other and cross talk.

Tesla’s self driving software up to v11 was code written by men. Hundreds of of thousands of code lines. From v12+ AI took over and now there is no need for men to write any code. It acquires all data from existing human driven vehicles on the road, asses them into safe and unsafe based on the after come results, and use the “best” to improve the model, in a nutshell.
If you see how the older (pre v12) versions behave in comparison with the latest, the difference is night and day. And this is improved fast by every single step.

I’m not trying to paint an all pink happy cloud picture. Just stating the obvious. Like anything else humans ever created it has the goods and the bads.
Every aspect (good/bad) though will grow exponentially just like AI itself.
There are people and teams that are working daily to predict the (positive/negative) implications upon the society. It’s not a simple matter at all.
Unless someone explain it to us we can’t even begin to imagine the potential (positive/negative) impact.
And be assured that on the opposite side there are teams that researching how to exploit the negatives just like anything else.

I do try to keep up with subject.
 

Rightness_1

New Member
Joined
Dec 30, 2024
Messages
26 (1.18/day)
>Posts in here trashing AI and discrediting the work of engineers with decades of experience developing an incredibly complex transformative algorithm because you don't like the company which did it
>Proceeds to praise an attempt at doing the same thing from the competitor because you like them

The hallmark of the average AMD fan
Erm... wth?

Reel your neck in and mute me so I don't have to deal with your crazy unhinged religious attacks like this.

You come across as some kind of amateur bully boy, attacking anyone who doesn't agree with your religious beliefs, you should see what my comment was saying is that how come nv needs allegedly 6 years of supercomputing time, when FSR and Sony's PSSR is supposed to use on-chip A.I. to render its frames? I offer the possibility that it's only to lock devs into nv's propriety algorithms, which they have to pay for.
 
Last edited:

A_macholl

New Member
Joined
Jun 14, 2024
Messages
8 (0.04/day)
Quantum computers is one strong candidate for replacing current tech. Just one of them will be able to replace entire rooms of servers. But it’s not a tech for personal/individual usage as it requires isolated near absolute zero conditions to work properly without any “outside” particle interference.
This type of compute power will improve AI vastly.
At some point, after them, almost everything will be cloud serviced. In the mean time IoT is going to continue to evolve as it is required as the base infrastructure for all this to happen.
One example is that cars will be all connected to each other and cross talk.
The moment where there will be AI implemented on quantum computer humans will become a secondary race.
 
Joined
Dec 25, 2020
Messages
7,332 (4.93/day)
Location
São Paulo, Brazil
System Name "Icy Resurrection"
Processor 13th Gen Intel Core i9-13900KS
Motherboard ASUS ROG Maximus Z790 Apex Encore
Cooling Noctua NH-D15S upgraded with 2x NF-F12 iPPC-3000 fans and Honeywell PTM7950 TIM
Memory 32 GB G.SKILL Trident Z5 RGB F5-6800J3445G16GX2-TZ5RK @ 7600 MT/s 36-44-44-52-96 1.4V
Video Card(s) NVIDIA RTX A2000
Storage 500 GB WD Black SN750 SE NVMe SSD + 4 TB WD Red Plus WD40EFPX HDD
Display(s) 55-inch LG G3 OLED
Case Pichau Mancer CV500 White Edition
Power Supply EVGA 1300 G2 1.3kW 80+ Gold
Mouse Microsoft Classic IntelliMouse (2017)
Keyboard IBM Model M type 1391405
Software Windows 10 Pro 22H2
Benchmark Scores I pulled a Qiqi~
Erm... wth?

Reel your neck in and mute me so I don't have to deal with your crazy unhinged religious attacks like this.

You come across as some kind of amateur bully boy, attacking anyone who doesn't agree with your religious beliefs, you should see what my comment was saying is that how come nv needs allegedly 6 years of supercomputing time, when FSR and Sony's PSSR is supposed to use on-chip A.I. to render its frames? I offer the possibility that it's only to lock devs into nv's propriety algorithms, which they have to pay for.

I am not bullying you... however, you simply seem not to know how any of this works, and you're passing judgment right away. Training and inference are two different things and you're also falling into a false equivalence trap by assuming all upscalers work the same way.

Take a quick look at how FSR works as an example. FSR 1 started as a simple CAS shader, you could load it through ReShade on any GPU from any vendor before AMD even added it tô games.

It eventually grew into a more complex upscaling solution but it never leveraged AI or matrix multiplication, not because they are nice or zeal for openness but because AMD's hardware is the only one in the industry which is not capable of it. And FSR 4, which allegedly does leverage machine learning algorithms, will be gated to the RX 9070 series, so much for that defense of open compatibility.

PSSR, as everything Sony, is fully proprietary, poorly documented to the public and apparently has been relatively poorly received so far. I don't believe it has any particular need for ML hardware since the PS5 Pro's graphics are still based on RDNA 2, which does not have this capability. Unless there is a semicustom solution, but I don't believe this to be the case.

Meanwhile, DLSS has been an ML trained model designed to reconstruct the image from less pixels from the very start, when it was introduced 7 years ago alongside the RTX 20 series.

The same applies to XeSS 1, but Intel went a step beyond and allowed it to run (albeit much slower) on any hardware they supports DP4A instructions. Which includes Nvidia Pascal and newer, but excludes RX Vega (exception of Radeon VII) and the original RDNA architecture (5700 XT).

I might have come off as harsh (yes I'll take the blame for it), and apologize if there was genuinely no malice in your initial remarks.
 
Last edited:
Joined
Sep 3, 2019
Messages
3,751 (1.91/day)
Location
Thessaloniki, Greece
System Name PC on since Aug 2019, 1st CPU R5 3600 + ASUS ROG RX580 8GB >> MSI Gaming X RX5700XT (Jan 2020)
Processor Ryzen 9 5900X (July 2022), 200W PPT limit, 80C temp limit, CO -6-14, +50MHz (up to 5.0GHz)
Motherboard Gigabyte X570 Aorus Pro (Rev1.0), BIOS F39b, AGESA V2 1.2.0.C
Cooling Arctic Liquid Freezer II 420mm Rev7 (Jan 2024) with off-center mount for Ryzen, TIM: Kryonaut
Memory 2x16GB G.Skill Trident Z Neo GTZN (July 2022) 3600MT/s 1.38V CL16-16-16-16-32-48 1T, tRFC:280, B-die
Video Card(s) Sapphire Nitro+ RX 7900XTX (Dec 2023) 314~467W (382W current) PowerLimit, 1060mV, Adrenalin v24.12.1
Storage Samsung NVMe: 980Pro 1TB(OS 2022), 970Pro 512GB(2019) / SATA-III: 850Pro 1TB(2015) 860Evo 1TB(2020)
Display(s) Dell Alienware AW3423DW 34" QD-OLED curved (1800R), 3440x1440 144Hz (max 175Hz) HDR400/1000, VRR on
Case None... naked on desk
Audio Device(s) Astro A50 headset
Power Supply Corsair HX750i, ATX v2.4, 80+ Platinum, 93% (250~700W), modular, single/dual rail (switch)
Mouse Logitech MX Master (Gen1)
Keyboard Logitech G15 (Gen2) w/ LCDSirReal applet
Software Windows 11 Home 64bit (v24H2, OSBuild 26100.2605), upgraded from Win10 to Win11 on Jan 2024
If A.I. was so great, why is it not done in the GPU in real-time? And why is some "supercomputer" allegedly doing it offline? I just don't believe this approach is completely necessary in 2025. It's just marketing lies to tie DLSS to nv hardware.

FSR 4.0 is going to be very interesting.
I am not bullying you... however, you simply seem not to know how any of this works, and you're passing judgment right away. Training and inference are two different things and you're also falling into a false equivalence trap by assuming all upscalers work the same way.

Take a quick look at how FSR works as an example. FSR 1 started as a simple CAS shader, you could load it through ReShade on any GPU from any vendor before AMD even added it tô games.

It eventually grew into a more complex upscaling solution but it never leveraged AI or matrix multiplication, not because they are nice or zeal for openness but because AMD's hardware is the only one in the industry which is not capable of it. And FSR 4, which allegedly does leverage machine learning algorithms, will be gated to the RX 9070 series, so much for that defense of open compatibility.

PSSR, as everything Sony, is fully proprietary, poorly documented to the public and apparently has been relatively poorly received so far. I don't believe it has any particular need for ML hardware since the PS5 Pro's graphics are still based on RDNA 2, which does not have this capability. Unless there is a semicustom solution, but I don't believe this to be the case.

Meanwhile, DLSS has been an ML trained model designed to reconstruct the image from less pixels from the very start, when it was introduced 7 years ago alongside the RTX 20 series.

The same applies to XeSS 1, but Intel went a step beyond and allowed it to run (albeit much slower) on any hardware they supports DP4A instructions. Which includes Nvidia Pascal and newer, but excludes RX Vega (exception of Radeon VII) and the original RDNA architecture (5700 XT).

I might have come off as harsh (yes I'll take the blame for it), and apologize if there was genuinely no malice in your initial remarks.
English is not my native language but here is what I understand from OP and correct me if I'm wrong please...

What this supercomputer does is separate of what an individual GPU is doing on the end user PC. This "server" simulates gaming on a wide variety of games and searching to find image errors after the upscaling and DLSS application. Then it tries to improve the model of the DLSS reconstruction. Every new version of DLSS reconstruction model with new enhancements is distributed through drivers to all end users.
So the reconstruction model is indeed running locally on every GPU but on the background the server is keep improving it.

How is that?
 
Joined
Nov 27, 2023
Messages
2,658 (6.31/day)
System Name The Workhorse
Processor AMD Ryzen R9 5900X
Motherboard Gigabyte Aorus B550 Pro
Cooling CPU - Noctua NH-D15S Case - 3 Noctua NF-A14 PWM at the bottom, 2 Fractal Design 180mm at the front
Memory GSkill Trident Z 3200CL14
Video Card(s) NVidia GTX 1070 MSI QuickSilver
Storage Adata SX8200Pro
Display(s) LG 32GK850G
Case Fractal Design Torrent (Solid)
Audio Device(s) FiiO E-10K DAC/Amp, Samson Meteorite USB Microphone
Power Supply Corsair RMx850 (2018)
Mouse Razer Viper (Original) on a X-Raypad Equate Plus V2
Keyboard Cooler Master QuickFire Rapid TKL keyboard (Cherry MX Black)
Software Windows 11 Pro (24H2)
@Zach_01
Essentially correct. Each local users GPU uses a model that was created on that supercomputer. I assume the improvements are delivered via updates to DLSS profiles which the NV driver searches for and pulls when launching a DLSS enabled title (it does that). Works both ways, from how they word it - PCs with telemetry enabled send back the information on what I assume are considered errors and weak points of the model usage in each supported title to facilitate improvements. They ARE being somewhat vague on what EXACTLY is done behind the scenes.
 
Joined
Dec 25, 2020
Messages
7,332 (4.93/day)
Location
São Paulo, Brazil
System Name "Icy Resurrection"
Processor 13th Gen Intel Core i9-13900KS
Motherboard ASUS ROG Maximus Z790 Apex Encore
Cooling Noctua NH-D15S upgraded with 2x NF-F12 iPPC-3000 fans and Honeywell PTM7950 TIM
Memory 32 GB G.SKILL Trident Z5 RGB F5-6800J3445G16GX2-TZ5RK @ 7600 MT/s 36-44-44-52-96 1.4V
Video Card(s) NVIDIA RTX A2000
Storage 500 GB WD Black SN750 SE NVMe SSD + 4 TB WD Red Plus WD40EFPX HDD
Display(s) 55-inch LG G3 OLED
Case Pichau Mancer CV500 White Edition
Power Supply EVGA 1300 G2 1.3kW 80+ Gold
Mouse Microsoft Classic IntelliMouse (2017)
Keyboard IBM Model M type 1391405
Software Windows 10 Pro 22H2
Benchmark Scores I pulled a Qiqi~
English is not my native language but here is what I understand from OP and correct me if I'm wrong please...

What this supercomputer does is separate of what an individual GPU is doing on the end user PC. This "server" simulates gaming on a wide variety of games and searching to find image errors after the upscaling and DLSS application. Then it tries to improve the model of the DLSS reconstruction. Every new version of DLSS reconstruction model with new enhancements is distributed through drivers to all end users.
So the reconstruction model is indeed running locally on every GPU but on the background the server is keep improving it.

How is that?

You are correct, it is as @igormp explained it: this bespoke "supercomputer" constantly runs simulations out of a huge data set, comparing between them and "learning" how to perfect the algorithm, it is always finding errors and more efficient ways of achieving a perfect image with as little of that data as possible, thus enabling the inference to make accurate predictions and conclusions. This is called the "training" stage.

During the inference stage, the model will then be run, producing predictions and conclusions based on the data from the training stage, thus maximizing the performance of the AI model as all available computing power will be used to apply the model instead of thinking about how it should work.

It's generative AI 101
 
Joined
Sep 3, 2019
Messages
3,751 (1.91/day)
Location
Thessaloniki, Greece
System Name PC on since Aug 2019, 1st CPU R5 3600 + ASUS ROG RX580 8GB >> MSI Gaming X RX5700XT (Jan 2020)
Processor Ryzen 9 5900X (July 2022), 200W PPT limit, 80C temp limit, CO -6-14, +50MHz (up to 5.0GHz)
Motherboard Gigabyte X570 Aorus Pro (Rev1.0), BIOS F39b, AGESA V2 1.2.0.C
Cooling Arctic Liquid Freezer II 420mm Rev7 (Jan 2024) with off-center mount for Ryzen, TIM: Kryonaut
Memory 2x16GB G.Skill Trident Z Neo GTZN (July 2022) 3600MT/s 1.38V CL16-16-16-16-32-48 1T, tRFC:280, B-die
Video Card(s) Sapphire Nitro+ RX 7900XTX (Dec 2023) 314~467W (382W current) PowerLimit, 1060mV, Adrenalin v24.12.1
Storage Samsung NVMe: 980Pro 1TB(OS 2022), 970Pro 512GB(2019) / SATA-III: 850Pro 1TB(2015) 860Evo 1TB(2020)
Display(s) Dell Alienware AW3423DW 34" QD-OLED curved (1800R), 3440x1440 144Hz (max 175Hz) HDR400/1000, VRR on
Case None... naked on desk
Audio Device(s) Astro A50 headset
Power Supply Corsair HX750i, ATX v2.4, 80+ Platinum, 93% (250~700W), modular, single/dual rail (switch)
Mouse Logitech MX Master (Gen1)
Keyboard Logitech G15 (Gen2) w/ LCDSirReal applet
Software Windows 11 Home 64bit (v24H2, OSBuild 26100.2605), upgraded from Win10 to Win11 on Jan 2024
Its pretty clear to me (I think) and I quote from OP

"The supercomputer's primary task involves analyzing failures in DLSS performance, such as ghosting, flickering, or blurriness across hundreds of games. When issues are identified, the system augments its training data sets with new examples of optimal graphics and challenging scenarios that DLSS needs to address."

Sounds like it runs and resets simulation gaming to continuously improve the model of prediction.
 
Joined
Dec 25, 2020
Messages
7,332 (4.93/day)
Location
São Paulo, Brazil
System Name "Icy Resurrection"
Processor 13th Gen Intel Core i9-13900KS
Motherboard ASUS ROG Maximus Z790 Apex Encore
Cooling Noctua NH-D15S upgraded with 2x NF-F12 iPPC-3000 fans and Honeywell PTM7950 TIM
Memory 32 GB G.SKILL Trident Z5 RGB F5-6800J3445G16GX2-TZ5RK @ 7600 MT/s 36-44-44-52-96 1.4V
Video Card(s) NVIDIA RTX A2000
Storage 500 GB WD Black SN750 SE NVMe SSD + 4 TB WD Red Plus WD40EFPX HDD
Display(s) 55-inch LG G3 OLED
Case Pichau Mancer CV500 White Edition
Power Supply EVGA 1300 G2 1.3kW 80+ Gold
Mouse Microsoft Classic IntelliMouse (2017)
Keyboard IBM Model M type 1391405
Software Windows 10 Pro 22H2
Benchmark Scores I pulled a Qiqi~
Its pretty clear to me (I think) and I quote from OP

"The supercomputer's primary task involves analyzing failures in DLSS performance, such as ghosting, flickering, or blurriness across hundreds of games. When issues are identified, the system augments its training data sets with new examples of optimal graphics and challenging scenarios that DLSS needs to address."

Sounds like it runs and resets simulation gaming to continuously improve the model of prediction.

That is correct and not mere marketing talk. I'm personally not a big fan of LLMs and "AI assistants", (training and inference is essentially how LLMs learn to make sense and sound coherent from a human context) but in the context of graphics it's really as big as it sounds.

The stuff JHH showed at CES is what allowed full scene pathtracing to be even feasible.
 

Rightness_1

New Member
Joined
Dec 30, 2024
Messages
26 (1.18/day)
English is not my native language but here is what I understand from OP and correct me if I'm wrong please...

What this supercomputer does is separate of what an individual GPU is doing on the end user PC. This "server" simulates gaming on a wide variety of games and searching to find image errors after the upscaling and DLSS application. Then it tries to improve the model of the DLSS reconstruction. Every new version of DLSS reconstruction model with new enhancements is distributed through drivers to all end users.
So the reconstruction model is indeed running locally on every GPU but on the background the server is keep improving it.

How is that?
Exactly, but why? If others are going to be doing all the work on the local GPU, why is nv still trying to say that they need a supercomputer running for 6 years to enhance DLSS if A.I. is as great as they keep saying it is? Surely, it's only to keep the nv eco system full of money?

I can see all the various problems playing Cyberpunk for instance, and the problems do not seem to get any better with newer DLSS DLLs. So why is it not better now than it was 2 years ago?

I'm not bashing nv personally, I do not use AMD graphics, and likely never will. But all this A.I. supercomputer nonsense makes me annoyed because it's a term that's just plastered all over everything and is 99.9% of the time, a complete lie. I think if A.I. was as advanced as nv claims, then why are they still using a remote supercomputer to render DLSS and not locally on the card itself?

I have heard that AMD are using pure A.I. in its upcoming FSR 4, and Sony is already using it on the PS5 Pro. And for the record, I'm not bashing DLSS, I enjoy it and find it a plus to my old RTX2070!
 
Last edited:
Joined
Sep 3, 2019
Messages
3,751 (1.91/day)
Location
Thessaloniki, Greece
System Name PC on since Aug 2019, 1st CPU R5 3600 + ASUS ROG RX580 8GB >> MSI Gaming X RX5700XT (Jan 2020)
Processor Ryzen 9 5900X (July 2022), 200W PPT limit, 80C temp limit, CO -6-14, +50MHz (up to 5.0GHz)
Motherboard Gigabyte X570 Aorus Pro (Rev1.0), BIOS F39b, AGESA V2 1.2.0.C
Cooling Arctic Liquid Freezer II 420mm Rev7 (Jan 2024) with off-center mount for Ryzen, TIM: Kryonaut
Memory 2x16GB G.Skill Trident Z Neo GTZN (July 2022) 3600MT/s 1.38V CL16-16-16-16-32-48 1T, tRFC:280, B-die
Video Card(s) Sapphire Nitro+ RX 7900XTX (Dec 2023) 314~467W (382W current) PowerLimit, 1060mV, Adrenalin v24.12.1
Storage Samsung NVMe: 980Pro 1TB(OS 2022), 970Pro 512GB(2019) / SATA-III: 850Pro 1TB(2015) 860Evo 1TB(2020)
Display(s) Dell Alienware AW3423DW 34" QD-OLED curved (1800R), 3440x1440 144Hz (max 175Hz) HDR400/1000, VRR on
Case None... naked on desk
Audio Device(s) Astro A50 headset
Power Supply Corsair HX750i, ATX v2.4, 80+ Platinum, 93% (250~700W), modular, single/dual rail (switch)
Mouse Logitech MX Master (Gen1)
Keyboard Logitech G15 (Gen2) w/ LCDSirReal applet
Software Windows 11 Home 64bit (v24H2, OSBuild 26100.2605), upgraded from Win10 to Win11 on Jan 2024
Exactly, but why? If others are doing all the work on the local GPU, why is nv still trying to say that they need a supercomputer running for 6 years to enhance DLSS if A.I. is as great as they keep saying it is? Surely, it's only to keep the nv eco system full of money?
I think you are missing the point.
The individual end user GPU is running the model of reconstruction and prediction on the game that the user is running. It doesn't improve the model of reconstruction/prediction, only runs it.
The model improvement is done be the server. They are not necessarily communicate with each other on the fly.
 
Joined
Dec 25, 2020
Messages
7,332 (4.93/day)
Location
São Paulo, Brazil
System Name "Icy Resurrection"
Processor 13th Gen Intel Core i9-13900KS
Motherboard ASUS ROG Maximus Z790 Apex Encore
Cooling Noctua NH-D15S upgraded with 2x NF-F12 iPPC-3000 fans and Honeywell PTM7950 TIM
Memory 32 GB G.SKILL Trident Z5 RGB F5-6800J3445G16GX2-TZ5RK @ 7600 MT/s 36-44-44-52-96 1.4V
Video Card(s) NVIDIA RTX A2000
Storage 500 GB WD Black SN750 SE NVMe SSD + 4 TB WD Red Plus WD40EFPX HDD
Display(s) 55-inch LG G3 OLED
Case Pichau Mancer CV500 White Edition
Power Supply EVGA 1300 G2 1.3kW 80+ Gold
Mouse Microsoft Classic IntelliMouse (2017)
Keyboard IBM Model M type 1391405
Software Windows 10 Pro 22H2
Benchmark Scores I pulled a Qiqi~
Exactly, but why? If others are doing all the work on the local GPU, why is nv still trying to say that they need a supercomputer running for 6 years to enhance DLSS if A.I. is as great as they keep saying it is? Surely, it's only to keep the nv eco system full of money?

I can see all the various problems playing Cyberpunk for instance, and the problems do not seem to get any better with newer DLSS DLLs. So why is it not better now than it was 2 years ago?

DLSS DLLs are only a small part of the system. These are basically just the runtime, not the code itself. Training data just won't fit into a small DLL.

What you're implying is akin to considering the DirectX DLLs to basically contain all the code necessary to display any game on their own, for example.

Without the software providing such updated training data for inference, updating the runtime is of questionable benefit at best. That's why updating the DLSS DLL won't increase image quality, it might improve performance ever so slightly - and even then, that is also anecdotal.

You might get a small improvement, you might not. Anything of quantifiable substance will require updated data which can only come with a software update.
 

Rightness_1

New Member
Joined
Dec 30, 2024
Messages
26 (1.18/day)
DLSS DLLs are only a small part of the system. These are basically just the runtime, not the code itself. Training data just won't fit into a small DLL.

What you're implying is akin to considering the DirectX DLLs to basically contain all the code necessary to display any game on their own, for example.

Without the software providing such updated training data for inference, updating the runtime is of questionable benefit at best. That's why updating the DLSS DLL won't increase image quality, it might improve performance ever so slightly - and even then, that is also anecdotal.

You might get a small improvement, you might not. Anything of quantifiable substance will require updated data which can only come with a software update.
Got you... But what I'm wondering is how can AMD/Sony (allegedly) do this in FSR4 without some "supercomputer" doing the work for them to upscale the image with minimal artifacts?
 
Joined
Dec 25, 2020
Messages
7,332 (4.93/day)
Location
São Paulo, Brazil
System Name "Icy Resurrection"
Processor 13th Gen Intel Core i9-13900KS
Motherboard ASUS ROG Maximus Z790 Apex Encore
Cooling Noctua NH-D15S upgraded with 2x NF-F12 iPPC-3000 fans and Honeywell PTM7950 TIM
Memory 32 GB G.SKILL Trident Z5 RGB F5-6800J3445G16GX2-TZ5RK @ 7600 MT/s 36-44-44-52-96 1.4V
Video Card(s) NVIDIA RTX A2000
Storage 500 GB WD Black SN750 SE NVMe SSD + 4 TB WD Red Plus WD40EFPX HDD
Display(s) 55-inch LG G3 OLED
Case Pichau Mancer CV500 White Edition
Power Supply EVGA 1300 G2 1.3kW 80+ Gold
Mouse Microsoft Classic IntelliMouse (2017)
Keyboard IBM Model M type 1391405
Software Windows 10 Pro 22H2
Benchmark Scores I pulled a Qiqi~
Got you... But what I'm wondering is how can AMD/Sony (allegedly) do this in FSR4 without some "supercomputer" doing the work for them to upscale the image with minimal artifacts?

Because it isn't the same method, and FSR has markedly inferior image quality when compared to DLSS in 9 out of 10 supported games. As for FSR 4 specifically? We don't know. It has not released yet.
 
Top