Tuesday, February 28th 2023

NVIDIA RTX Video Super Resolution Tested, AI Enhanced Streaming That Barely Makes a Difference

NVIDIA has leveraged their expertise in neural networks and deep learning to release an interesting new feature with their R530 driver branch, an AI video stream upscaler designed to take advantage of RTX Tensor Cores when playing video content within Chromium based browsers. Our previous news article on RTX Video Super Resolution (VSR) covered the release of Chrome 110 stable, which included support for this technology. The latest version of Microsoft Edge, based on Chromium, also officially supports RTX VSR. Owners of NVIDIA RTX graphics cards may have been puzzled by exactly how to enable this feature however, either in Chrome 110 or in the NVIDIA Control Panel, since the relevant 'NvidiaVpSuperResolution' setting is enabled by default within Chrome, but the required accompanying driver has only just been released, three weeks later.
To use RTX VSR, you'll need a RTX 30 or 40-series graphics card, be running the latest NVIDIA GeForce Graphics Driver, and have enabled the "RTX Video Enhancement" option within the NVIDIA Control Panel, under the "Adjust video image settings" submenu. There are four quality presets, with "1" being the lowest and "4" being the highest, while also using the most GPU resources. Owners of RTX 20-series cards will have to wait for NVIDIA to enable this functionality for their GPUs, once the engineering work is completed for that architecture.
Some comparison screenshots, taken on my personal system with a 3080 Ti and a 1440p monitor, it seems the technology is most noticeable when applied to videos at 720p resolution and below.

Similar to the well received NVIDIA Shield TV, which could take 720p or 1080p content and upscale it to 4K at up to 30 frames per second using the AI hardware within the Tegra X1+, RTX VSR is a further, more advanced development. Using the more powerful hardware on modern RTX graphics cards, RTX VSR automatically upscales content played from within your browser between 360p and 1440p, to 4K, improving detail and removing the compression artifacts streamed content is known for.

NVIDIA's RTX VSR FAQ and blog post answers some common questions and provides further details on how the technology works.

You can take a look at NVIDIA's comparison video or try enabling the feature yourself to decide how well NVIDIA's efforts have paid off. As we've seen with other AI based deep learning solutions, the technology will continue to improve with time. In it's current state, RTX VSR seems particularly well suited for increasing the clarity of videos uploaded at lower resolutions or bitrates, such as older videos or live streamed content from Twitch or YouTube. Those using capped or slower network connections limiting their streaming options should also appreciate being able to efficiently consume content without sacrificing too much in image quality. I can't wait to see where the iterative path leads, as this technology could be as impactful in video media as AI based upscalers were for gaming!


Update: After further testing of a YouTube stream of in game content, set to 480p and 720p, isolated differences between RTX VSR enabled at setting '4' and disabled can be shown.
480p enabled 480p disabled
720p enabled 720 disabled
Looking closely at the barrels, trees, textures on the surfaces, text on the container, and straight lines for example the roof of the service station, we can see image quality improvements with RTX VSR enabled and set to '4' quality.

While these image quality improvements certainly exist, I have some questions as to how many owners of RTX 30 and 40-series graphics cards spend their time watching low resolution streams, since improvements are much less obvious when using higher resolution source material. This technology seems ideally suited to portable applications, where there is limited internet bandwidth available, such as smartphones on mobile networks, or laptops on the go using slow wireless connections. Unfortunately, NVIDIA requires laptops to be plugged into mains power to use RTX VSR, due to the additional power drawn by the Tensor Cores required for image processing (most laptops would use iGPU via Optimus under light graphics loads, and RTX VSR requires the discrete GPU to be active), and there are no smartphones that have RTX features. The way I see it, it's a zero effort (after initial setup, which takes a minute or so) way to get slightly better image quality, scaling less as you go up in source resolution, with a negligible draw on system resources. There is also the case where many older videos from the earlier days of the internet tend to be only available in relatively low resolution, so this technology can certainly come into play to offer a more contemporary image quality.
Add your own comment

114 Comments on NVIDIA RTX Video Super Resolution Tested, AI Enhanced Streaming That Barely Makes a Difference

#76
TheoneandonlyMrK
rolachazWe need some CSIs tech enhanced video streaming!

Best meme ever, I actually belly laughed :)
Posted on Reply
#77
R-T-B
I mean sure, it's something and no one should act like a free feature is bad. But amazing this is not.
Posted on Reply
#78
R0H1T
It's not really free though, those Tensor cores probably cost a lot of die space & $$$ overall. Better put them to good use!
Posted on Reply
#79
Mussels
Freshwater Moderator
Testing this with an 8K video on YT, set to 144p cause why nots?



Before:




After (will edit in after browser restart)



I cant see much change with these 'bears eating grass'



Now we do 720p:
before:


After:



Things like the bees antennae look like they have AA turned on, but that's about it

720p - lvl 4 - 4K native
Posted on Reply
#80
Solaris17
Super Dainty Moderator
MusselsI cant see much change with these 'bears eating grass'
are you kidding?? the AI is magnificent. It literally changed from bears to a bunch of bees, I cant even tell where the grass and bears were supposed to be!
Posted on Reply
#81
Mussels
Freshwater Moderator
Solaris17are you kidding?? the AI is magnificent. It literally changed from bears to a bunch of bees, I cant even tell where the grass and bears were supposed to be!
Realtime Bee Tracing


I wonder where this works best - videos with film grain and artifacting? High quality videos downsampled? Small screens, big screens?


Ahah, one of the quotes covers it: Basically, good source material over-compressed can be 'restored'

It's GPU accelerated madVR
RTX VSR reduces or eliminates artifacts caused by compressing video — such as blockiness, ringing artifacts around edges, washout of high-frequency details and banding on flat areas — while reducing lost textures. It also sharpens edges and details.
Posted on Reply
#83
ancelotti
Looks excellent to me. I've been using it on my Shield TV Pro for several years and it's amazing how well it can work on the big screen.

I think mileage is going to really depend upon the source and what you're viewing it on. A decent 720p source is going to look a lot better on a large 4k monitor. On a smaller 1080p screen the difference may be negligible.

I can totally understand how someone on a regular 24" 1080p lcd thinks the whole thing is a waste of time, and how someone on a 45" 4K oled thinks it's the greatest tech in the world.

Most people will be somewhere in the middle, where it gives a nice little bump in quality, but nothing to buy a new video card over.
Posted on Reply
#85
Selaya
R-T-BI mean sure, it's something and no one should act like a free feature is bad. But amazing this is not.
i wouldnt exactly call 240w of power draw waste a free feature but ye
Posted on Reply
#86
Wasteland
ancelottiLooks excellent to me. I've been using it on my Shield TV Pro for several years and it's amazing how well it can work on the big screen.

I think mileage is going to really depend upon the source and what you're viewing it on. A decent 720p source is going to look a lot better on a large 4k monitor. On a smaller 1080p screen the difference may be negligible.

I can totally understand how someone on a regular 24" 1080p lcd thinks the whole thing is a waste of time, and how someone on a 45" 4K oled thinks it's the greatest tech in the world.

Most people will be somewhere in the middle, where it gives a nice little bump in quality, but nothing to buy a new video card over.
I use the "AI upscaler" on my Shield TV too. It's great, but I don't think it's the same tech at all. The Shield certainly isn't pulling an extra 200 watts with the feature active, lol. The Shield's version is more like a glorified sharpening filter, which can work wonders but it also regularly creates quite aggressive artifacts at lower resolutions. The tradeoff is almost always worth it, IMO, but a lot of people would likely disagree.

Anyway, this new feature is much more subtle--subtle to the point of clownishness at the moment, but it seems Nvidia is striving for something here that is substantially more advanced. I look forward to seeing how this thing develops.
Posted on Reply
#87
loracle706
Watching a youtube video with 300+ watts, what a dumb idea even if it looks a little better, but the cost of energy and heat are inacceptable, only if you are one of those idiots who wants to fuck the planet quickly !!
Posted on Reply
#88
ThrashZone
MusselsRealtime Bee Tracing


I wonder where this works best - videos with film grain and artifacting? High quality videos downsampled? Small screens, big screens?


Ahah, one of the quotes covers it: Basically, good source material over-compressed can be 'restored'

It's GPU accelerated madVR
Hi,
Maybe youtube/... should stop compressing videos instead that wouldn't eat up a couple hundred watts and everyone could benefit ;)
Posted on Reply
#89
TheoneandonlyMrK
ancelottiLooks excellent to me. I've been using it on my Shield TV Pro for several years and it's amazing how well it can work on the big screen.

I think mileage is going to really depend upon the source and what you're viewing it on. A decent 720p source is going to look a lot better on a large 4k monitor. On a smaller 1080p screen the difference may be negligible.

I can totally understand how someone on a regular 24" 1080p lcd thinks the whole thing is a waste of time, and how someone on a 45" 4K oled thinks it's the greatest tech in the world.

Most people will be somewhere in the middle, where it gives a nice little bump in quality, but nothing to buy a new video card over.
Well done you beat it's invention by several year's then, it's new this week?!?

It's only just been made, you were probably using something else?!


Tried it, meh, game changing it isn't, and I have the worst network of anyone I know, a sim card equipped mobile WiFi system on 3 in the UK, DL in the low Mb.
It's marginally better yes but not that noticeable and I am not buying a Nvidia GPU for this feature.
As the years roll and DL speeds increase past dial up it will become less required, wait,,,,. were a fair amount past dialup, ,, aren't you all.
Posted on Reply
#90
R-T-B
Selayai wouldnt exactly call 240w of power draw waste a free feature but ye
That's actually a fair point.
Posted on Reply
#91
Super Firm Tofu
R-T-BThat's actually a fair point.
And it's not actually a free feature - the cost for R&D is baked into what we pay for nvidia cards.
Posted on Reply
#92
loregnum
loracle706Watching a youtube video with 300+ watts, what a dumb idea even if it looks a little better, but the cost of energy and heat are inacceptable, only if you are one of those idiots who wants to fuck the planet quickly !!
Every single person that has a capable gpu could do what you said and NOTHING would happen to the planet. NOTHING. May I suggest getting out of the climate cult you are apparently in and understanding you have been sold a bunch of bullshit by people who want more control of your life and more of your money to do so. May I suggest actually looking into the history of the world (like the fact the planet right now isn't anywhere close to being as warm as it has been long before the 20th/21st centuries and somehow it still managed to keep on chugging and life prospered!) and simple reality and realizing how insane the climate cult fear mongering is. Maybe I suggest actully realizing the people in your cult have been making doom predictions for decades that have never come true so that in itself should show nobody should trust anything they say.

I do have to wonder though....what if the people you speak of are using solar and wind to power their homes? They still "fucking the planet quickly" or nah?
Posted on Reply
#93
Fluffmeister
Hmm yeah not impressed, if I wanted higher power consumption to watch a YouTube video I'd buy a NAVI 31 based card.

I'll leave this option off thanks.
Posted on Reply
#94
R-T-B
Super Firm TofuAnd it's not actually a free feature - the cost for R&D is baked into what we pay for nvidia cards.
I doubt this one feature factors in much at all on the bill, especially considering it doesn't even really appear high effort.
Posted on Reply
#95
Laykun
For brighter videos this seems to work quite well, but for darker content the AI has a hard time distinguishing what's "noise" and what's "detail". The two screenshots post show 1) No VSR 2) VSR level 4. The content is running from my plex server, it's the Aliens from 1986. At the time of taking the screenshot, in both scenarios, the server is transcoding the content to my web browser, so some contrast is lost. But you can clearly see VSR is crushing the details in the blacks, even straight up "grime" is gone, the window shutter horizontal lines become non-existent. When streaming the content without transcoding, you get more contrast and the AI tends to handle it better.

No VSR

VSR Level 4


Weirdly, it seems to add an almost uniform / patterned "noise" to the image that I'm not a fan of, and can again create strange artifacts. The second set of screenshots show deep space nine, first episode, jake sisko's eyes. The first screenshot is VSR level 4, second is off. Notice the vertical patterns on the left side of the image, on Jake's right eye.
VSR:

No VSR


Afraid there's just too many weird quirks, not for me.
Posted on Reply
#97
loracle706
loregnumEvery single person that has a capable gpu could do what you said and NOTHING would happen to the planet. NOTHING. May I suggest getting out of the climate cult you are apparently in and understanding you have been sold a bunch of bullshit by people who want more control of your life and more of your money to do so. May I suggest actually looking into the history of the world (like the fact the planet right now isn't anywhere close to being as warm as it has been long before the 20th/21st centuries and somehow it still managed to keep on chugging and life prospered!) and simple reality and realizing how insane the climate cult fear mongering is. Maybe I suggest actully realizing the people in your cult have been making doom predictions for decades that have never come true so that in itself should show nobody should trust anything they say.

I do have to wonder though....what if the people you speak of are using solar and wind to power their homes? They still "fucking the planet quickly" or nah?
I am climate neutral, but i believe thats a dumb idea.
Posted on Reply
#98
chrcoluk
nguyen
Difference seems tiny, and I did watch that video in as high res as possible. If it really consumes watts in the hundreds then not worth it in my opinion.
Posted on Reply
#99
nguyen
chrcolukDifference seems tiny, and I did watch that video in as high res as possible. If it really consumes watts in the hundreds then not worth it in my opinion.
Dont use Quality setting 4, 1 is good enough and it raise power consumption by 10-30W depending on res
Posted on Reply
#100
Redwoodz
I just want to know who complained to Nvidia that their youtube video's were not the sharpest they could be. lmao
Posted on Reply
Add your own comment
Dec 13th, 2024 22:48 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts