Monday, March 27th 2023
12GB Confirmed to be GeForce RTX 4070 Standard Memory Size in MSI and GIGABYTE Regulatory Filings
It looks like 12 GB will be the standard memory size for the NVIDIA GeForce RTX 4070 graphics card the company plans to launch in mid-April 2023. It is very likely that the card has 12 GB of memory across the 192-bit memory bus width of the "AD104" silicon the SKU is based on. The RTX 4070 is already heavily cut down from the RTX 4070 Ti that maxes out the "AD104," with the upcoming SKU featuring just 5,888 CUDA cores, compared to the 7,680 of the RTX 4070 Ti. The memory sub-system, however, could see NVIDIA use the same 21 Gbps-rated GDDR6X memory chips, which across the 192-bit memory interface, produce 504 GB/s of memory bandwidth. Confirmation of the memory size came from regulatory filings of several upcoming custom-design RTX 4070 board models by MSI and GIGABYTE, with the Eurasian Economic Commission (EEC), and Korean NRRA.
Sources:
harukaze5719 (Twitter), VideoCardz
62 Comments on 12GB Confirmed to be GeForce RTX 4070 Standard Memory Size in MSI and GIGABYTE Regulatory Filings
the xx50 will have 128-bit
I think that calling it a 4060 is actually generous, because it'd be a 4060 at best if the market was healthy. This chart from back when Ada was revealed is a relative of execution units enabled in the processor per released product over the architectures since Kepler, in percent relative to the full die possible:
(credit to whoever made and posted this, I saved it from some post here on TPU - mad useful btw)
As you can see: 4090 is only 88% enabled despite being the top product, there is an extreme gap between it that would fit practically the entire Ampere and Turing lineups between it and the 4080, 4070 Ti was still referred to as 4080 12G here, but yes, its full AD104 configuration is only 42.5% relative to a full AD102. It's probably not accurate to 4070 and below as these were very far when this chart was created, but it's quite possible to update it without major difficulty.
All in all, this is a very, very lukewarm midranger that JHH is positioning as a performance segment product, thanks to lack of competition from AMD, and the worst market conditions you could imagine - it's probably worse now than it was during the crypto boom because at least back then they had an excuse to price gouge.
By the next financial report I fully expect they will include this even in sector naming. "Gaming" all of a sudden won't be adequate for a sector that will sell accelerators to small AI servers and home AI users, so I expect they'll rename it to something it will cover both gaming and AI acceleration, and they'll have full mouths of AI, even if it will only be a small portion of their sales. The potential for growth is almost endless.
Of course AI won't bring much revenue for home users or small nodes, or at least it won't look as lucrative as home mining, so I doubt we'll see a very quick adoption. But while at mining Nvidia had to appear as if the miners are abusing their products and using it for something it wasn't intended to do, with AI they can wholly embrace it!
So the next generation you'll try to buy a graphics card, it won't be a gaming card any more. It will be a Home Accelerator for AI and Gaming.
I will probably get an RTX7070Ti in the future, that is, if the price is within acceptable parameters at the time, and I, of course, am still on this greed-filled dustball.
But yeah, similar thoughts. Anything above 500 is running into some psychological barrier here. The one called common sense.
Lets say it is an eye opener.
This would allow an eventual full AD102 Titan/4090 Ti card to significantly exceed the original 4090's performance (something the 3090 Ti was never able to do vs. the 3090 because the improvement lies in raising the power limit and halving the amount of memory chips used, lowering memory power consumption significantly - 3090 had 82 out of 84 and 3080 Ti 80 out of 84 SMs enabled, meaning there was never a significant gap in execution units between the three of these). The addition of 16 SMs (4090 has 128 out of 144 enabled), better memory and 24 MB of L2 disabled on the 4090 would potentially create a card that is at least 30% faster before any power limit was ever raised.
GDDR6X optimization (by using newer generation, faster modules), as well as low-level architectural tweaks can still be leveraged to bring improvements throughout the entire segment, and this is what infuriates me with the RTX 40 series, they have a brilliant architecture yet carved a product stack that is designed from the bottom up to entirely forgo generational performance per dollar improvements and make its lower budget options look so bad against the top tier one that people would be just compelled to buy the 4090 on a value proposition alone (this is just evil!) - and it looks like they've succeeded, the 4090 has outsold every other GPU from both them and AMD combined this generation.
The 30 series was on the terrible Samsung 10nm+ they called 8nm a big reason Nvidia was able to offer semi decent pricing although you could say the 3090 was way overpriced given how much cheaper the process was vs TSMC 4n, The 4080/4090 are vastly superior products vs high end ampere in almost every way the issue is just pricing especially at the 80tier and lower. I like my 3080ti but it's always felt like a meh product even though it offered 40% more performance over my 2080ti although I will say part of that is how terrible it feels 4k vs 4k RT vs my 4090 after just one generation.
At the end of the day all that really matters is performance and the last 2 flagship vs flagship have been some of the largest increases over the last decade with the 4090 being the largest increase in performance since the 1080ti released over a half decade ago yet everyone is still crying because they are expensive. We've had 3 generations of terrible pricing in a row at this point so anyone who thinks we are going to go back to 10 series pricing I hope is not holding their breath. Regardless of how much more cut down the 4080 is compared to my 3080ti in my secondary pc I would still feel way better spending $1200 on it vs the $1400 I spent on my 3080ti even though that was nearly two years ago.
Although anyone who bought a 3090ti near launch really got shafted 2000 usd for that is just comical regardless of if it uses the full die or not even at it's eventual 1100 MSRP less than 6 months after release the 4080 is way better.
I do expect the 5000 series to be priced a little better especially if AMD figures out how to make dual GCD gpu's work for gaming. I'm still not holding my breath though.
Regardless this is a thread about the 4070 and unfortunately for most people that card is going to be a joke but at least Nvidia didn't cheap out to the point of the 30 series and give it 8GB of Vram again.
The competition is doing the only thing that they can do, produce great video cards. The thing is, if everyone and their mother is buying nVidia, what difference does it make?
If you have a Radeon or an Arc GPU under the hood, then you can be proud because it means that you're doing what is needed to make nVidia accountable. I've been doing the same thing since 2008 because I would feel like a pathetic loser if I complained about nVidia while supporting them at the same time. At least is right. The RX 7900 XT shouldn't be over $600USD. Seeing as the RTX 4080 and RTX 4090 both came out before RDNA3, I don't see how RDNA3 could have had any effect on nVidia's pricing. It was already in the stratosphere without anyone else's help. It's true, but the thing is that Radeons historically have obliterated nVidia from a value standpoint but people still bought nVidia. All that the lower prices did was reduce AMD's graphics revenue. Can you really blame them for not bothering with consumers who turned their backs on them over and over again? I'm honestly surprised that it took this long for them to do it. Their attitude is probably "If you sheep want to pay too much for a video card, by all means, we'll get on that bus!" and it's an attitude that consumers have earned by voting with their wallets over the years and choosing the most over-priced and bad-value cards on the market. If consumers had rewarded AMD for their better offerings, we wouldn't be in this mess to begin with. Everyone with a GeForce card under the hood has helped the current situation to form and has nobody but themselves to blame. At least people with Radeons (and Arcs for that matter) did their part to try to prevent the current situation from forming. At this point, it doesn't matter what AMD does. It only matters if people become willing to stop paying through the nose for nVidia cards. I used to think like you do, that all AMD has to do is make Radeons far more attractive than GeForce. AMD did that very thing for years but people are stupid and wanted the brand that had "The fastest card in the world" (like that even matters when your budget is less than $300). Consider tha the RTX 3050, the worst value video card on the market today, is actually selling well!
If people are willing to pay through the nose for a card as weak as an RTX 3050, then all is already lost.
Both these gpu makers know long before what the competition performance targets are it isn't by chance that RDNA 3 almost performs the same as the 4070ti/4080 that was likely AMD target all along. ADA/RDNA3 were likely mostly finalized 2-3 years ago as far as performance targets although looking at AMD own presentation they missed their mark by quite a bit.
That all changed when Zen kicked Intel in the cojones. The last time that nVidia was kicked in the cojones was when ATi released the HD 5000-series. It's been a very long time...
There are more games with the same issue not just this one I suppose.
Nvidia also has several cards in recent history that had issues on the VRAM department. GTX 970; 660; for example, with asymmetrical buses, fell off in performance faster than their VRAM cap would indicate. Both have 0,5 GB wired to lower bandwidth. On the other end of the spectrum, Nvidia's best/most loved cards are always the ones that sport a higher VRAM than anything else in the stack: 980ti (6GB), 1080ti (11GB) being the best examples.
Today Nvidia is 'fixing' very low VRAM amounts with a bunch of cache, and its already showing a lot of difference between games in how that performs.