Tuesday, March 11th 2025

AMD Radeon RX 9070 XT Reportedly Outperforms RTX 5080 Through Undervolting
AMD's Radeon RX 9070 XT is demonstrating unexpected performance gains through aggressive undervolting, with overclocking specialists documenting significant improvements that push the GPU past NVIDIA's pricier GeForce RTX 5080 in specific benchmarks. Recent tests by Der8auer using a PowerColor Red Devil RX 9070 XT revealed a 10% frame rate increase in Cyberpunk 2077 at 4K Ultra settings by applying a -170 mV voltage offset while increasing the power target to 110%. This modification enabled the GPU to reach clock speeds of 3.36 GHz, compared to 2.90 GHz at stock settings, resulting in 66 FPS versus the RTX 5080's 65 FPS in identical testing environments. The undervolting phenomenon appears consistent across the product line, with YouTuber Alva Jonathan achieving similar 10% performance improvements on the standard RX 9070 using ASRock's Steel Legend model.
Both testers discovered that traditional core clock overclocking yielded negligible results, suggesting these factory-overclocked cards are already operating near their architectural limits. The voltage-frequency curve adjustments effectively lower the voltage required for higher frequencies. Memory overclocking proved counterproductive, with error correction mechanisms actually reducing in-game performance when pushed beyond stable parameters. These results come with important caveats—both tested units are premium variants with enhanced power delivery and cooling solutions that sell significantly above AMD's MSRP. The PowerColor Red Devil commanded a $200 premium over the RX 9070 XT's $599 launch price, while the ASRock Steel Legend carried a $90 markup over the RX 9070's base $550 MSRP. Even with these premiums, however, the high-end RX 9070 XT models remain approximately $200 less expensive than NVIDIA's RTX 5080 while delivering comparable rasterization performance after optimization, despite NVIDIA's ongoing advantages in ray tracing capabilities and software ecosystem.
Source:
via Tom's Hardware
Both testers discovered that traditional core clock overclocking yielded negligible results, suggesting these factory-overclocked cards are already operating near their architectural limits. The voltage-frequency curve adjustments effectively lower the voltage required for higher frequencies. Memory overclocking proved counterproductive, with error correction mechanisms actually reducing in-game performance when pushed beyond stable parameters. These results come with important caveats—both tested units are premium variants with enhanced power delivery and cooling solutions that sell significantly above AMD's MSRP. The PowerColor Red Devil commanded a $200 premium over the RX 9070 XT's $599 launch price, while the ASRock Steel Legend carried a $90 markup over the RX 9070's base $550 MSRP. Even with these premiums, however, the high-end RX 9070 XT models remain approximately $200 less expensive than NVIDIA's RTX 5080 while delivering comparable rasterization performance after optimization, despite NVIDIA's ongoing advantages in ray tracing capabilities and software ecosystem.
86 Comments on AMD Radeon RX 9070 XT Reportedly Outperforms RTX 5080 Through Undervolting
Of course demand is high when supply is basically nonexistent.
OC vs OC
Stock Vs Stock
That how it should go
But if a $600 MSRP can catch the tail of a $1000 MSRP one, that's kind of bad, isn't it? Or you oc the 5080 and settle with +10% performance for nearly double the price?
But lets not let that get in the way of a good old fashion yay AMD boo Nvidia right :rolleyes:
No, Timespy is not a game. But what it is is a fair representation of equalized scaling, especially on a equalized architecture and wrt minimum framerates when there is not a clear limitation.
Therefore, this is not a surprise to see in practical applications. I really do try to explain this to people, but some are in extreme denial about 5080's limitations (stock clock and buffer).
It's cool, wait for others. But, you know, don't bash me when I try to explain it (sometimes briefly, other-times very much not) earlier. :P
nVIDIA disguises this limitation a little by the boatload of extra bw (5080 only needs 22gbps to operate at stock; that extra bw can translate into ~6% extra perf in many cases).
This is why you should laugh when (something like a proposed) 18GB card that surpass it in many ways. The buffer no long the primary bottleneck at higher speeds, and slightly higher compute at stock (both important to keeping 60fps in many games) to match.
Like I say, you can believe it or not, but this is what nVIDIA does. Some people will eventually understand and some never will. That is their prerogative, but it's still true, and why people should understand.
5080 has a stock (especially important wrt RT) clock of 2640mhz. Yes, it can overclock to ~3150mhz or so, but 16GB becomes a limitation,where 120fps may be approx equal to stable ~1440p60RT, it can't.
This is why benchmarks matter, and 100fps (and next 120fps) important in that bench. It is also why ~20k is important in Unigine (like in that under-volting video)...they all roughly equal out to standards.
This is why 9070 xt clocks to where it does, and has 16GB...and hits 100fps/20k...which probably similar to stable 1080p60RT. Many cards have the 5080 problem, including 9070/5070/5070ti, for different reasons.
9070 XT knows what it is, and knows what it isn't. It is a well-matched card that will hit the spec it needs to, which is keeping 60fps in 1080pRT or 1440p 'quality' up-scaling (960p->1440p), perhaps with a lil' tweak.
On a 5080 1080p->4k up-scaling will be better native, and that likely will be the point of a '9080xt' or something, but it's still alright with a 9070 xt if you use something like frame-gen. Native FR/lag still okay.
That is why 9070 xt is a good product. It hits all those targets and is inexpensive. 5080 (besides 1080p->4k) isn't really going to give a much better stabilized experience; it's not a 1440pRT card...not *really*.
You can be amazed by the 'tremendous architectural advancements' and whatever they try to sell cheaper cards on next gen that surpass those that were formerly more expensive.
All-the-while I will have already explained exactly what the former limitations are (that were fixable within silicon choices/ram limitations) but nVIDIA chose to limit them to sell later as new product(s).
Sometimes several times to showcase different areas a product was originally limited. Like I say, there are instances >16GB will matter, but other instances greater than most 5080s can OC on the core.
Again, I have explained why 9070 XT is *perfectly* oriented towards 1080pRT or 1440p 'quality' up-scaling. 5080 is NOT a 1440pRT card, or one step up, as 4090 currently is, hence 5080 is not a great card.
It is faster, but not generally in any way that will give you a tangible better experience. This is why the cheaper Radeon makes sense; for all intents and purposes it can do the same things well.
This pressure would largely be relieved (and equalized) on something like 9216sp @ ~3700-3780/36gbps and 18GB, and can almost guarentee to you however far in advance will hit 120fps in Time Spy. Which would be both a better product (in terms of keeping 60fps in many instances at 1440pRT), and perhaps most importantly for you to understand, cheaper for nVIDIA to make. This will also manifest as better 1080p->4k up-scaling capability (than 9070xt/5070ti, but perhaps similar to 5080, depending upon if >16GB ram needed to keep 60fps).
In that case, a '6070' would be a '4090'...except when you need a larger buffer (which is largely true in keeping 60fps at 4k, but will be more true in instances of 1440p fairly soon, especially with up-scaling/rt/fg).
That is why I call 16GB cards (and 9070 xt) a 1080p card. It is a 1080p RT card (That can upscale). It is the bottom of this up-coming era, where more ram needed etc for higher-rez and (eventually standard) RT.
Again, we can pretty much assume the next-gen stack will be 1080p, 1440p, and 4k RT...with a card like 9070 xt at the bottom (~6144sp @ 4200mhz/40gbps/16GB? Maybe configured differently). And then scaled.
Where 1440p is ~4090 or a high-clocked ~12288sp card that replaces it. 5080 (to it's replacement) a 1080p->4k card. Then a halo for native 4kRT or 1440p (and 4k up-scaling) path tracing.
It all lines up. The only cards that currently fit are ~4090 (to a 12288sp part, perhaps clocked lower at first and then higher later) and 9070 xt to a low-end part (with higher clocks as AMD generally uses).
It's possible we see this exact same battle with AMD using 6144sp/16GB and nVIDIA 9216sp/18GB, doing this exact same thing for the exact relative prices but one tier down.
nVIDIA will look better in *some* instances, but things will change as RAM/compute requirements increase and things *actually* stabalize to the next-gen consoles, and how that manifests IDK.
That's why 1440p requirements are murky. Will 5080 level (with more ram and/compute) be enough? I don't know. That's why I say think of it as 1080p->4k, sometimes 1440p (as that product *kind of* is now).
Next tier up 1440pRT, and 4k raster or quality up-scaling, as 4090 is now (but may not always be capable of keeping the later at 60fps). This is why a faster part (nvidia refresh or maybe AMD right-away) is needed.
This is why I've said in order for 5080 to not be outdated (because of the difference in compute potential between such parts that they could then outdate through software demands, if not new games) the 24GB 5080 needs to clock to ~3.23-3.24ghz. The process *will* allow this. nVIDIA will almost certainly not, and it is for these reasons. To sell the more-or-less same part multiple times, and outdate each one through different limitations.
Follow? At the end of the day 5080 16GB becomes just as much a 1080p card as 9070 XT, and to me already is, which to me is humorous. It just has better up-scaling performance at 1080p->4k.
Again, AMD is already prepping for this by having a lower price and not selling it for something it is not. It is a 1080pRT card with the ability upscale to 1440p. 5080 can upscale to 4k. Paying for that is your choice.
And likely will continue to be. I have no doubt AMD will sell a better-matched card for actual next-gen 1440p that actually performs better than what nVIDIA does to replace 5080. BC this is what they do.
Yeah...I know. Wall of text. Sometimes I just edit rather than making a new post. Bad habit of mine and I apologize...It just keeps everything together (although I know it becomes a lot in one post).
www.techpowerup.com/255221/amd-radeon-rx-3080-xt-navi-to-challenge-rtx-2070-at-usd-330 On Navi III and Navi IV? Other than a rumor of a back-ported Overdrive API in Linux for Navi III, AFAIK no.
www.phoronix.com/news/AMD-OverDrive-SMU13-RDNA3
For Vega 10, I've used OverdriveNTool's softpowerplaytable editor, extensively.
Thanks @DR4G00N, I was being very obtuse until he pointed out how to use the tool. :laugh:
For Navi I and Navi II, I think MorePowerTool supported such.
MoreClockTool on Navi III, is extremely basic. Great as an el strippo replacement for Radeon Settings/Wattman (in the case of a 'minimal install' of the AMD drivers), but that's about it.
AFAIK, MCT offers 0 over Radeon Settings.
Obviously voltages are set very conservatively to ensure stability, some chips will be able to do much higher clocks at lower voltages giving 20% more in the best case scenarios.
So we're talking about 5070 Ti @ 750 and 9070 XT @ 650 MSRP, and not the fake MSRP with 50 discount for the reviews.
I will take nothing away from the 5080's though they do clock well, it's the efficiency benchmarks that would make for a more interesting outlook.
Go Radeons!! :pimp:
Felt like for multiple generations in a row the 70 class could overclock to the 80 class and they had large price disparities, it was borderline expected. Vocal AMD supporter and Radeon buyer and GeForce naysayer believes GeForce equivalent product (or higher tier) not worth it, more news at 10.
Edit: What is this "naysayer" thing anyway? The holy inquisition?