• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.

NVIDIA GeForce RTX 5070 Ti Specs Leak: Same Die as RTX 5080, 300 W TDP

AleksandarK

News Editor
Staff member
Joined
Aug 19, 2017
Messages
2,999 (1.07/day)
Recent leaks have unveiled specifications for NVIDIA's upcoming RTX 5070 Ti graphics card, suggesting an increase in power consumption. According to industry leaker Kopite7kimi, the RTX 5070 Ti will feature 8,960 CUDA cores and operate at a 300 W TDP. In a departure from previous generations, the RTX 5070 Ti will reportedly share the same GB203 die with its higher-tier sibling, the RTX 5080. This architectural decision differs from the RTX 40-series lineup, where the 4070 Ti and 4080 utilized different dies (AD104 and AD103, respectively). This shared die approach could potentially keep NVIDIA's manufacturing costs lower. Performance-wise, the RTX 5070 Ti shows promising improvements over its predecessor. The leaked specifications indicate a 16% increase in CUDA cores compared to the RTX 4070 Ti, though this advantage shrinks to 6% when measured against the RTX 4070 Ti Super.

Power consumption sees a modest 5% increase to 300 W, suggesting improved efficiency despite the enhanced capabilities. Memory configurations remain unconfirmed, but speculations about the card indicate that it could feature 16 GB of memory on a 256-bit interface, distinguishing it from the RTX 5080's rumored 24 GB configuration. The positioning across the 50-series GPU stack of this RTX 5070 Ti appears carefully calculated, with its 8,960 CUDA cores sitting approximately 20% below the RTX 5080's 10,752 cores. This larger performance gap between tiers contrasts with the previous generation's approach, potentially indicating a more defined product hierarchy in the Blackwell lineup. NVIDIA is expected to unveil its Blackwell gaming graphics cards at CES 2025, with the RTX 5090, 5080, and 5070 series leading the announcement.



View at TechPowerUp Main Site | Source
 
Wait till competition commits suicide
@
"Design" the "new" architecture for a couple months longer than usual
@
Sell a slightly overclocked and tuned mid-tier product for even longer dollars than last time
@
"You definitely need that, trust me"
 
While on paper this sounds a lot better than the 4070 Ti was design-wise, I don't like the power consumption increase (even if modest) and certainly won't like its price. Lucky me I'm not in the market for a new graphics card in the foreseeable future.
 
Looks like the upper part of the line-up is coming into focus

5070 $600 250W Slightly higher perforamance than 4070 Super
5070 Ti $800 300W Slightly higher performance than 4070 Ti Super
5080 $1000 400W Slightly higher performance than 4080 Super
5090 $2000 600W 40% higher performance than 4090

Nothing too exciting given the same 4 nm die process except for the 5090. I have no idea how this thing is going to work at 600W if your rig isn't perfectly up to snuff.

As for the lower part of the line-up, Nvidia is definitely waiting to see how Battlemage and RNDA4 performs.
 
I have my eye on the 5070Ti, however, if it arrives with 12GB VRAM, there is no way. Also, the MSRP should be reasonable, you can't just keep pushing prices up nVidia, the world doesn't work that way, salaries doesn't just go up ad-infinitum, eventually you will price yourself out of the tier/market you used to target.
 
Last edited:
Looks like the upper part of the line-up is coming into focus

5070 $600 250W Slightly higher perforamance than 4070 Super
5070 Ti $800 300W Slightly higher performance than 4070 Ti Super
5080 $1000 400W Slightly higher performance than 4080 Super
5090 $2000 600W 40% higher performance than 4090

Nothing too exciting given the same 4 nm die process except for the 5090. I have no idea how this thing is going to work at 600W if your rig isn't perfectly up to snuff.

As for the lower part of the line-up, Nvidia is definitely waiting to see how Battlemage and RNDA4 performs.
Replace "slightly higher" with "same" and you will nail them.
Also add $100 to 5070 and $200 to 70ti and 80
 
I have my eye on the 5070Ti, however, if it arrives with 12GB VRAM, there is no way. Also, the MSRP should be reasonable, you can't just keep pushing prices up nVidia, the world doesn't work that way, salaries doesn't just go up infinitum, eventually you will price yourself out of the tier/market you used to target.
It's rumored to be 16GB. About MSRP I'm afraid same as 4070ti/ti super at best.
 
It's rumored to be 16GB. About MSRP I'm afraid same as 4070ti/ti super at best.

Hopefully some semblance of sanity can be restored. I was happy with the RTX3070Ti GPU performance itself, after all, I am only on 1440p, however, that damn VRAM ceiling of only 8GB, the thing cries out in pain with AAA games on this resolution, I can't stand stuttering/1% lows, irks me to the extreme.
 
Wait till competition commits suicide
@
"Design" the "new" architecture for a couple months longer than usual
@
Sell a slightly overclocked and tuned mid-tier product for even longer dollars than last time
@
"You definitely need that, trust me"
Of course the competition committed suicide. Journalists and reviewers banged the drum of how much better Nvidia is in terms of performance even when just edging out AMD, banged another drum of how ray tracing is "the future" (TM). People gobbled that up, went for Nvidia even when AMD was a perfectly decent (almost equivalent) or cheaper alternative. The demand for either cheaper (at the mid/low end) or more efficient (at the high end) GPUs (I think AMD never had these together in a single product) dried up in favor of "Nvidia at every end".

Only now that Nvidia is squeezing every $ it can with last year's overclocked products from a market they ended up dominating have many of the same journalists and reviewers realized the consequences and I started seeing articles how "Nvidia's GPUs don't get better, they just trade more power for more performance", or users started voting in polls that they care about raster performance nor ray tracing.
 
Hopefully some semblance of sanity can be restored. I was happy with the RTX3070Ti GPU performance itself, after all, I am only on 1440p, however, that damn VRAM ceiling of only 8GB, the thing cries out in pain with AAA games on this resolution, I can't stand stuttering/1% lows, irks me to the extreme.
Naaah that can't be, 'its just allocation, not utilisation'

:roll::oops::(

But yeah, that's what's about to happen to the crop of 12GB cards in about 2 years, with the occasional game popping up in 2025 already. If you upgrade now, better make it 16 or don't bother.
 
Of course the competition committed suicide. Journalists and reviewers banged the drum of how much better Nvidia is in terms of performance even when just edging out AMD, banged another drum of how ray tracing is "the future" (TM). People gobbled that up, went for Nvidia even when AMD was a perfectly decent (almost equivalent) or cheaper alternative. The demand for either cheaper (at the mid/low end) or more efficient (at the high end) GPUs (I think AMD never had these together in a single product) dried up in favor of "Nvidia at every end".

Only now that Nvidia is squeezing every $ it can with last year's overclocked products from a market they ended up dominating have many of the same journalists and reviewers realized the consequences and I started seeing articles how "Nvidia's GPUs don't get better, they just trade more power for more performance", or users started voting in polls that they care about raster performance nor ray tracing.
But... Nvidia was always better. Simple as that. Even during GCN; AMD drew more power for a slightly better bang for buck, offered more VRAM for a slightly better bang for buck. And that's all AMD wrote. Not ONCE did they take the leading position, either in featureset or in software altogether. Driver regime has been spotty. GPU time to market has no real fixed cadence, its 'whatever happens with AMD' every single time and it never happens to be just a smooth launch. The list of issues goes on and on and on.

The only thing to applaud is AMD brought RDNA2/3 to a good, stable situation. Too bad the products don't sell. Because AMD chose to price them in parity with Nvidia... So at that point, they didn't have the consistency, nor the trust factor or brand image, nor the bang for buck price win.... and guess what. RDNA4 is a bugfix and then they're going back to the drawing board. Again: no consistency, they even admitted themselves that they failed now.
 
This has got some strong 4080 performance for $999 price vibes written all over it; with potential to be much worse value than the 4070ti super depending on tariffs.

Another cycle of Nvidia giving you less or the same for more.
 
Correct, I was confused by the 5070 - it has significantly lower core count than the 4070 super, but for the prices I believe they will be higher
RTX 4070 - 5888 cores
RTX 5070 - 6400 cores
RTX 4070 super - 7168 cores
Whoops, you are right. I was mistaken that the 4070 Super had 5888 cores.
 
This has got some strong 4080 performance for $999 price vibes written all over it; with potential to be much worse value than the 4070ti super depending on tariffs.

Another cycle of Nvidia giving you less or the same for more.
999$ for the always out of stock "founders edition" and probably 1100$+ for the non-hairdryer models
 
But... Nvidia was always better. Simple as that. Even during GCN; AMD drew more power for a slightly better bang for buck, offered more VRAM for a slightly better bang for buck. And that's all AMD wrote. Not ONCE did they take the leading position, either in featureset or in software altogether. Driver regime has been spotty. GPU time to market has no real fixed cadence, its 'whatever happens with AMD' every single time and it never happens to be just a smooth launch. The list of issues goes on and on and on.

The only thing to applaud is AMD brought RDNA2/3 to a good, stable situation. Too bad the products don't sell. Because AMD chose to price them in parity with Nvidia... So at that point, they didn't have the consistency, nor the trust factor or brand image, nor the bang for buck price win.... and guess what. RDNA4 is a bugfix and then they're going back to the drawing board. Again: no consistency, they even admitted themselves that they failed now.
Not quite true. AMD had good wins with the R300, Fury and all RDNA products. Also their All-in-wonder products were good. And iGPUs were always a step ahead.

In between those, AMD and Nvidia traded blows with the Nvidia FX5000 series being a notable flop. Only recently starting with the release of Maxwell did Nvidia start really pulling ahead. And yes they are killing it in the last six years.
 
As for the lower part of the line-up, Nvidia is definitely waiting to see how Battlemage and RNDA4 performs.
I expect both to suck as well. AMD is distracted with AI and not focused on gaming, despite what they are saying, and Intel may *wish* to gain market share with Battlemage but they have no money left and Celestial looks dead in the water.
 
But... Nvidia was always better. Simple as that.
Maybe you are to young to remember, but definitely nVidia wasn't always better.
Just for your homework, search for AMD Radeon HD 5870 card. It was so good, that it was almost beating the dual GPU card from nVidia, while wiping the floor with whole nvidia gen cards. Also the 5850 was a monster too, and could work in pair with the 5870. I remember that was my last SLI setup ever, but it was a blast. Good ol' times.
 
But... Nvidia was always better. Simple as that. Even during GCN; AMD drew more power for a slightly better bang for buck, offered more VRAM for a slightly better bang for buck. And that's all AMD wrote. Not ONCE did they take the leading position, either in featureset or in software altogether. Driver regime has been spotty. GPU time to market has no real fixed cadence, its 'whatever happens with AMD' every single time and it never happens to be just a smooth launch. The list of issues goes on and on and on.
Were you even around when the HD 5870 came out and Nvidia's answer was the Fermi toaster?
 
Maybe you are to young to remember, but definitely nVidia wasn't always better.
Just for your homework, search for AMD Radeon HD 5870 card. It was so good, that it was almost beating the dual GPU card from nVidia, while wiping the floor with whole nvidia gen cards. Also the 5850 was a monster too, and could work in pair with the 5870. I remember that was my last SLI setup ever, but it was a blast. Good ol' times.
And earlier than that in the ATI era, the Radeon 9700 was a game-changer.
 
But... Nvidia was always better. Simple as that. Even during GCN; AMD drew more power for a slightly better bang for buck, offered more VRAM for a slightly better bang for buck. And that's all AMD wrote. Not ONCE did they take the leading position, either in featureset or in software altogether. Driver regime has been spotty. GPU time to market has no real fixed cadence, its 'whatever happens with AMD' every single time and it never happens to be just a smooth launch. The list of issues goes on and on and on.

The only thing to applaud is AMD brought RDNA2/3 to a good, stable situation. Too bad the products don't sell. Because AMD chose to price them in parity with Nvidia... So at that point, they didn't have the consistency, nor the trust factor or brand image, nor the bang for buck price win.... and guess what. RDNA4 is a bugfix and then they're going back to the drawing board. Again: no consistency, they even admitted themselves that they failed now.
AMD was beating nVidia, since 4870 'till Polaris. Even the dumb GTX260 was a hot furnace. 5870 was the first DX11 card, and the first gen of HW tesselation. Also, the HW of AMD/ATi was much of higher quality. But drivers were hit-and miss, for decades- that's true. The picture quality was a bit better on AMD/ATi either, much like not gimping the colour preset like "green" company did since ever.
P.S.: AMD made Mantle, which became Vulcan. The SW RT back in 2016.
And earlier than that in the ATI era, the Radeon 9700 was a game-changer.
There was also the Radeon 9600 Pro. A great, affordable low end card. that made gaming possible for alot of people. And it had a quite nice OC room. It was like "2500+ Barton" of videocard.
 
Also their All-in-wonder products were good.
Just to be clear, this was all ATi, AMD had nothing to do with that :)

I will probably still aim for 5080 this time around. My oldest will get my 4070Ti, youngest gets my 3070Ti.

Although I could be swayed into a 5070Ti, because while the internet was hating on the Ti series, I was enjoying them :D
 
AMD was beating nVidia, since 4870 'till Polaris. Even the dumb GTX260 was a hot furnace. 5870 was the first DX11 card, and the first gen of HW tesselation. Also, the HW of AMD/ATi was much of higher quality. But drivers were hit-and miss, for decades- that's true. The picture quality was a bit better on AMD/ATi either, much like not gimping the colour preset like "green" company did since ever.
P.S.: AMD made Mantle, which became Vulcan. The SW RT back in 2016.

There was also the Radeon 9600 Pro. A great, affordable low end card. that made gaming possible for alot of people. And it had a quite nice OC room. It was like "2500+ Barton" of videocard.
So pray tell where it went wrong then. They were 'beating' Nvidia with what? Tech that met the end of its dev cycle. They had every opportunity to obtain true leadership but AMD was thinking 'meh, we're good, this is fine, we don't need to chase the cutting edge continuously, 50% market is all we can do'? And then they thought, 'beating Nvidia': 'Let's release Nvidia's 970*(Edited) 2 years after the fact and kill this market!' I mean... what?! They weren't beating Nvidia at all. They traded punches, but never answered Nvidia Titan, and Hawaii XT failed miserably - a way too hungry dead end forcing them into Fury X and the capital loss against Maxwell. AMD's death of GCN happened somewhere between the great release of a 7970 and the birth of Tonga, which proved the arch was a dead end, but pushing out 290(x) anyway on a whoppin 512 bit bus. And then Fury had to happen, because how else do you above and beyond moar VRAM 512 bit? And then they got their 1000 bit hbm ass kicked by a 384 bit 980ti.

AMD made Mantle, which became Vulkan. And then what? What is the overarching strategy here, console access? We can applaud their many successes but the key to those events is that you use them to increase your market share and control, to the detriment of other key players. That's commerce.

Its one thing to make the occasional 'good card' (which is really nothing more than pricing a product correctly / in a way people buy it!) that sells, its another to actually execute on a strategy. Over several decades of AMD GPUs I haven't discovered what it is. If we go buy the marketing its some wild mix of making fun of the others while failing yourself (Poor Volta and a string of other events), going unified arch first and then not, and then yes, we might as well unify this again after dropping under 20% share convincingly; going 'midrange with Polaris' to lose key market share and brand recognition earned on GCN (which had a few 'good cards') only to claw back into the high end with RDNA2/3 and then back to midrange again?

There's just no rhyme or reason to it, and that is why it can't get ever get consistently good.

Maybe you are to young to remember, but definitely nVidia wasn't always better.
Just for your homework, search for AMD Radeon HD 5870 card. It was so good, that it was almost beating the dual GPU card from nVidia, while wiping the floor with whole nvidia gen cards. Also the 5850 was a monster too, and could work in pair with the 5870. I remember that was my last SLI setup ever, but it was a blast. Good ol' times.
I had a console age in those years, for some reason it was PS3 at that point, not PC :D
 
Last edited:
My prediction for this 5070Ti is 4080S equivalent for 799. If it manages to beat the 4080S by 10-15%, 899.

So pray tell where it went wrong then. They were 'beating' Nvidia with what? Tech that met the end of its dev cycle. They had every opportunity to obtain true leadership but AMD was thinking 'meh, we're good, this is fine, we don't need to chase the cutting edge continuously, 50% market is all we can do'? And then they thought, 'beating Nvidia': 'Let's release Nvidia's 970*(Edited) 2 years after the fact and kill this market!' I mean... what?! They weren't beating Nvidia at all. They traded punches, but never answered Nvidia Titan, and Hawaii XT failed miserably - a way too hungry dead end forcing them into Fury X and the capital loss against Maxwell. AMD's death of GCN happened somewhere between the great release of a 7970 and the birth of Tonga, which proved the arch was a dead end, but pushing out 290(x) anyway on a whoppin 512 bit bus. And then Fury had to happen, because how else do you above and beyond moar VRAM 512 bit? And then they got their 1000 bit hbm ass kicked by a 384 bit 980ti.

AMD made Mantle, which became Vulkan. And then what? What is the overarching strategy here, console access? We can applaud their many successes but the key to those events is that you use them to increase your market share and control, to the detriment of other key players. That's commerce.

Its one thing to make the occasional 'good card' (which is really nothing more than pricing a product correctly / in a way people buy it!) that sells, its another to actually execute on a strategy. Over several decades of AMD GPUs I haven't discovered what it is. If we go buy the marketing its some wild mix of making fun of the others while failing yourself (Poor Volta and a string of other events), going unified arch first and then not, and then yes, we might as well unify this again after dropping under 20% share convincingly; going 'midrange with Polaris' to lose key market share and brand recognition earned on GCN (which had a few 'good cards') only to claw back into the high end with RDNA2/3 and then back to midrange again?

There's just no rhyme or reason to it, and that is why it can't get ever get consistently good.


I had a console age in those years, for some reason it was PS3 at that point, not PC :D
Not moving forwards=going backwards
 
Last edited:
Back
Top