• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.

NVIDIA GeForce RTX 4070 isn't a Rebadged RTX 4080 12GB, To Be Cut Down

btarunr

Editor & Senior Moderator
Staff member
Joined
Oct 9, 2007
Messages
47,878 (7.38/day)
Location
Dublin, Ireland
System Name RBMK-1000
Processor AMD Ryzen 7 5700G
Motherboard Gigabyte B550 AORUS Elite V2
Cooling DeepCool Gammax L240 V2
Memory 2x 16GB DDR4-3200
Video Card(s) Galax RTX 4070 Ti EX
Storage Samsung 990 1TB
Display(s) BenQ 1440p 60 Hz 27-inch
Case Corsair Carbide 100R
Audio Device(s) ASUS SupremeFX S1220A
Power Supply Cooler Master MWE Gold 650W
Mouse ASUS ROG Strix Impact
Keyboard Gamdias Hermes E2
Software Windows 11 Pro
It turns out that NVIDIA didn't just cancel (unlaunch) the GeForce RTX 4080 12 GB last week, but also shelved the SKU until it is needed in the product stack. This is probably because NVIDIA intended to sell it at $900, and will find it difficult to justify a xx70-class SKU at this price-point. A Moore's Law is Dead report goes into the possible reasons NVIDIA shelved the RTX 4080 12 GB, and why it won't be rebadged as the RTX 4070.

The RTX 4070, although expected to be based on the same AD104 silicon as the RTX 4080 12 GB, won't have the same configuration. The RTX 4080 12 GB maxed out the AD104, enabling all 7,680 CUDA cores on the silicon. It's likely that the RTX 4070 will have fewer CUDA cores, even if it retains the 192-bit memory interface and 12 GB memory size. The memory clock could be changed, too. The RTX 4080 12 GB was essentially NVIDIA trying to upsell the successor of the RTX 3070 Ti (maxed out GA104) as an xx80-class SKU, at a higher price-point. Moore's Law is Dead also showed off possible designs of the RTX 4070 Founders Edition, revealing a compact design with many of the same design improvements implemented with the RTX 4090 FE. This card comes in a strictly 2-slot design.



View at TechPowerUp Main Site | Source
 
"The RTX 4080 12 GB maxed out the AD102, enabling all 7,680 CUDA cores on the silicon."

this should say 'AD104', not 'AD102'
 
It doesn't really surprise me that they'd launch the 4070 before the 4070 Ti. I assume they'll bring the latter out with the 4080 12G specs around this time next year for a slightly lower price.
 
Replying to Source 4: So even the 4070TI wouldn't use a full AD104?
 
I can smell a "cheap" 4070 coming with a severely cut-down chip for $700, then a 4070 Ti with the same config as the 4080 12 GB would have been for $850. And then, Jensen won't understand why isn't everybody happy.

It really puzzles me how Nvidia doesn't have a well-thought out plan for the whole product stack before the launch of the flagship the way AMD and Intel do.
 
they really did shelved the 4080 12GB, huh... vendors are gonna waste their money, time and effort again for damage control. Guess that EVGA saying nope to the 40 Series was probably a "good move" on their part. It's no wonder this launch was utter chaos.
 
I can smell a "cheap" 4070 coming with a severely cut-down chip for $700, then a 4070 Ti with the same config as the 4080 12 GB would have been for $850. And then, Jensen won't understand why isn't everybody happy.

It really puzzles me how Nvidia doesn't have a well-thought out plan for the whole product stack before the launch of the flagship the way AMD and Intel do.

I mean the 4080 12GB was already cut down to 46% of a 4090. How much more can you cut that down and charge north of $700? At that point you are better off with numerous other cards already on the market. Heck right now you can get 6900XTs for under $700.
 
Replying to Source 4: So even the 4070TI wouldn't use a full AD104?
Same result: they'd be admitting they tried to scam people if they just change the name or price with the same die.
Honestly since they went so far to reveal it at their launch announcement, Nvidia probably would have been better off just going ahead with it. This wishy-washy approach puts them in an even more awkward situation than they were already in. Amazing that so many well paid individuals in their company couldn't come to the conclusion that this was a terrible idea from the start.
 
they really did shelved the 4080 12GB, huh... vendors are gonna waste their money, time and effort again for damage control. Guess that EVGA saying nope to the 40 Series was probably a "good move" on their part. It's no wonder this launch was utter chaos.
It appears to me that they didn't launch Ada - they only launched the 4090 with zero idea what the rest of the product stack will look like. It shouldn't be so hard to figure out since Ada is just Ampere Refresh which is just Turing Refresh, but for some reason, it is. Maybe the hard part is figuring out how to milk people for small updates on the same architecture again and again while brute forcing some improvements in them to keep everybody quiet and praise them for their "efforts".

I mean the 4080 12GB was already cut down to 46% of a 4090. How much more can you cut that down and charge north of $700? At that point you are better off with numerous other cards already on the market. Heck right now you can get 6900XTs for under $700.
It wasn't cut down. It was a fully enabled AD104.
 
It wasn't cut down. It was a fully enabled AD104.

It's a fully enabled AD104, which is a cutdown version of AD102. Pointless semantics, you know what I meant.

Instead of replying constructively you decided to reply with that instead. The question remains, how much more can you cut down the die compared to the 4090 and sell it for north of $700; disagreements on my usages of cutdown aside.
 
Last edited:
Normally the xx70' performance is equal to last highest performing gaming card, yet this generation it seems no to be the case.
Only the 4090 is a reasonable uplift in performance but due to the high price it isn't impressive.
AMD should take this opportunity and crush nVidia's disappointing Alda Lovelace GPUs.
 
Unless Nvidua convinces the reviewers to only look at the frame doubled performance.
They don't even need to try. They are already convinced. Tech sites are turning again pro Nvidia in their articles. Slowly back they are.

I believe that after AM5's fail to become a clear winner over Intel and the power of RTX 4090 seen in reviews combined with DLSS3 and the lack of information about RDNA3 cards, plus Intel's enter in the GPU market, they see a lack of ability from AMD to continue as an effective and strong competitor to the other two in the retail market. So, they are starting to play it nice with Nvidia. See the latest articles about GPU prices. Nvidia still sells above MSRP most of it's cards, RTX 4090 comes at $1600 to over $2000 and tech sites post articles with titles saying that GPU prices are going down and that normality in GPU pricing is back. They do say somewhere that AMD is undercutting Nvidia in pricing and that mid range Radeon cards are better, but those remarks look more like necessary small text lost in the articles than something to clearly point to and turn the reader's focus at it.
 
I'd like 4070 based on AD103, sealing 25% off the CUDA count for a 8064-8704 /w regular GD6 and price it 35% lower at 799.
Use AD104 only for 4060 Ti 60 and 50. with 192 160 and 128 bit bus.
 
Last edited:
The renamed 12GB at $700 would have pushed it, but they are not even releasing it at those specs? Fuck them.
 
This whole thing is bizarre.

NVIDIA knows for each Ampere card, the price, COGS, gross margin, number of each type of core, memory size, memory bus width, memory bandwidth, pixel fillrate, texture fillrate, peak FP32, etc.

And they made Ada Lovelace GPUs, they know the same performance figure for the new generation silicon. It's their freakin' design.

They know what VRAM capacities are available to a particular memory bus width.

All they need to do is put together a simple spreadsheet with model numbers from 3050 to 3090 Ti and an adjacent cells listing model numbers from 4050 to 4090 Ti. Look for anomalies, try to fix those and price accordingly.
 
I mean the 4080 12GB was already cut down to 46% of a 4090. How much more can you cut that down and charge north of $700? At that point you are better off with numerous other cards already on the market. Heck right now you can get 6900XTs for under $700.
The 4080 (both) is positioned too far away from the 4090 really; and that happened because the 4090 really isnt that big of a jump where it matters: raster perf.

Now they have no real space above the Ampere stack to do anything meaningful gen to gen without looking like clowns. And it appears AMD is being more competent about it. And precisely: last gen is turning highly competitive now all of a sudden. Working as intended?! Not if there are sub 700 dollar top end cards destroying your new offering...

As said before, my popcorn is ready :D
 
Last edited:
It's a fully enabled AD104, which is a cutdown version of AD102. Pointless semantics, you know what I meant.

Instead of replying constructively you decided to reply with that instead. The question remains, how much more can you cut down the die compared to the 4090 and sell it for north of $700; disagreements on my usages of cutdown aside.
It's not pointless semantics. AD104 is not a cut down version of the GA102. It is a much smaller chip to begin with, and as such, a lot more of it can be manufactured per wafer. The defective parts of each fully manufactured chip are cut down to make lesser performing ones. Either you don't understand how chip design and manufacturing works and really believe that every single GPU is a cut down version of another, or it's you who decided to pick a pointless fight about semantics when you knew very well what I meant. If "cutting down" means what you seem to think it means (using the same architecture to make smaller chips), then what is the 1030 compared to the 1080 Ti or the 710 compared to the 780 Ti? C'mon...

If only products based on chips with defective parts form the basis of the initial launch, then I want to know where fully working chips go - probably to storage to be sold for an even higher price later. This is why I was scratching my head during the Ampere launch, and this is why I'm scratching my head now.
 
This whole thing is bizarre.

NVIDIA knows for each Ampere card, the price, COGS, gross margin, number of each type of core, memory size, memory bus width, memory bandwidth, pixel fillrate, texture fillrate, peak FP32, etc.

And they made Ada Lovelace GPUs, they know the same performance figure for the new generation silicon. It's their freakin' design.

They know what VRAM capacities are available to a particular memory bus width.

All they need to do is put together a simple spreadsheet with model numbers from 3050 to 3090 Ti and an adjacent cells listing model numbers from 4050 to 4090 Ti. Look for anomalies, try to fix those and price accordingly.
This is the struggle of having to cut back on margins and you can see the internal fight of the company bleed out into the product stack. Greed, is what happened here, followed by sharp market realities.
 
The 4080 (both) is positioned too far away from the 4090 really; and that happened because the 4090 really isnt that big of a jump where it matters: raster perf.
It's not that big of a jump, because it's the same architecture with improvements in the RT and Tensor cores. The rest of the design is just a node shrink and brute force.
 
MLID doesn't do "reports". They only speculate out loud stuff that seems logical to the average viewer. Tbh after their Arc "report" everyone should just ignore this "source" forever.
 
Same result: they'd be admitting they tried to scam people if they just change the name or price with the same die.
Honestly since they went so far to reveal it at their launch announcement, Nvidia probably would have been better off just going ahead with it. This wishy-washy approach puts them in an even more awkward situation than they were already in. Amazing that so many well paid individuals in their company couldn't come to the conclusion that this was a terrible idea from the start.
I ended up watching his video, I caught that part. But it seems really odd that they wouldn't use it as the TI version.
 
MLID doesn't do "reports". They only speculate out loud stuff that seems logical to the average viewer. Tbh after their Arc "report" everyone should just ignore this "source" forever.
I couldn't agree more - except that I don't even look at them as "source" (maybe source of pointless speculation, but nothing else).
 
It seems obvious to me that the full die will be reserved for the RTX 4070 Ti to be released in the mid-gen refresh in about a year's time. The same as the RTX 3070 Ti using the full GA104 over the 3070's moderately cut down configuration.
 
Back
Top