Thursday, January 19th 2023
NVIDIA GeForce RTX 4060 Ti Possible Specs Surface—160 W Power, Debuts AD106 Silicon
NVIDIA's next GeForce RTX 40-series "Ada" graphics card launch is widely expected to be the GeForce RTX 4070 (non-Ti), and as we approach Spring 2023, the company is expected to ramp up to the meat of its new generation, with xx60-segment, beginning with the GeForce RTX 4060 Ti. This new performance-segment SKU debuts the 4 nm "AD106" silicon. A set of leaks by kopite7kimi, a reliable source with NVIDIA leaks, shed light on possible specifications.
The RTX 4060 Ti is based on the AD106 silicon, which is expected to be much smaller than the AD104 powering the RTX 4070 series. The reference board developed at NVIDIA, codenamed PG190, is reportedly tiny, and yet it features the 16-pin ATX 12VHPWR connector. This is probably set for 300 W at its signal pins, and adapters included with graphics cards could convert two 8-pin PCIe into one 300 W 16-pin connector. The RTX 4060 Ti is expected to come with a typical graphics power value of 160 W.At this point we don't know whether the RTX 4060 Ti maxes out the AD106, but its rumored specs read as follows: 4,352 CUDA cores across 34 streaming multiprocessors (SM), 34 RT cores, 136 Tensor cores, 136 TMUs, and an unknown ROP count. The GPU is expected to feature a 128-bit wide GDDR6/X memory interface, and 8 GB could remain the standard memory size. NVIDIA is expected to use JEDEC-standard 18 Gbps GDDR6 memory, which should yield 288 GB/s of memory bandwidth. It will be very interesting to see how much faster the RTX 4060 Ti is over its predecessor, the RTX 3060 Ti, given that it has barely two-thirds the memory bandwidth. NVIDIA has made several architectural improvements to the memory sub-system with "Ada," and the AD106 is expected to get a large 32 MB L2 cache.
Sources:
kopite7kimi (Twitter), VideoCardz
The RTX 4060 Ti is based on the AD106 silicon, which is expected to be much smaller than the AD104 powering the RTX 4070 series. The reference board developed at NVIDIA, codenamed PG190, is reportedly tiny, and yet it features the 16-pin ATX 12VHPWR connector. This is probably set for 300 W at its signal pins, and adapters included with graphics cards could convert two 8-pin PCIe into one 300 W 16-pin connector. The RTX 4060 Ti is expected to come with a typical graphics power value of 160 W.At this point we don't know whether the RTX 4060 Ti maxes out the AD106, but its rumored specs read as follows: 4,352 CUDA cores across 34 streaming multiprocessors (SM), 34 RT cores, 136 Tensor cores, 136 TMUs, and an unknown ROP count. The GPU is expected to feature a 128-bit wide GDDR6/X memory interface, and 8 GB could remain the standard memory size. NVIDIA is expected to use JEDEC-standard 18 Gbps GDDR6 memory, which should yield 288 GB/s of memory bandwidth. It will be very interesting to see how much faster the RTX 4060 Ti is over its predecessor, the RTX 3060 Ti, given that it has barely two-thirds the memory bandwidth. NVIDIA has made several architectural improvements to the memory sub-system with "Ada," and the AD106 is expected to get a large 32 MB L2 cache.
164 Comments on NVIDIA GeForce RTX 4060 Ti Possible Specs Surface—160 W Power, Debuts AD106 Silicon
The biggest consumer graphics dies have ranged wildly from 104 to 110 to 200 to 102 over the last decade. The smallest have ranged from 119 to 208 to 117 to 107.
There's simply no precedent for trying to create a rule and apply to GPU die internal codenames. They're constantly in flux and both AMD and Nvidia have proven, quite consistently, that they'll mix and match multiple dies to single products on a whim, whilst often doing something differently to the previous generation. The only pattern that really holds any truth is that lower last numbers are bigger dies within each generation.
This classification corresponding with a performance tier is entirely arbitrary and barely conforms to any pattern because within a couple of generations the tentative pattern is broken.
Meanwhile knowledgeable people on tech sites were calling the 680 a high end GPU just because it had the 8 in it's name which in the past signified high end like with the 280/285/480/580. Some of those people bought a Kepler Titan because they thought it was the Kepler Flagship. Buyers remorse was painful when the 780 Ti came out and was not only faster than the Titan with more shaders but also $300 less.
The 3080/3080 Ti were both high end GPUs but the 3080 was a gimped 3080 Ti.
Okay, does that work in reverse as well? Say they release the 5090, the full x02 die, but it's slower than Amd's small 8600xt, should they keep asking 2k for it even though performance is terrible cause it uses the big die? How does that make any senses to you? Are you paying for performance or for die sizes? What difference does the die size make to you? I don't care if they sell me the smaller die humanly possible, if the performance is good enough that's all that matters
en.m.wikipedia.org/wiki/List_of_Nvidia_graphics_processing_units
Nvidia are not consistent today and historically have been changing every aspect of their silicon codenames. Any pattern lasts for 2-3 architectures at most before being altered beyond recognition.
The Ti models in particular are even less consistent than other models, if there's a Ti at all, and then how does "Super" fit into all this?!
You're insisting that a 4060Ti should use AD104, but on what precedent? It happened for Ampere, but it didn't happen for either the RTX or GTX lines of Turing, there wasn't one at all for Pascal, or Maxwell.
Your point is that xx60 Ti should use '104' die, but there's no pattern in the last decade to support that, why are you trying to force this '104' rule on something that has quite literally happened only once before (Ampere) in the last 8.5 years?
I'm only bringing up older cards to prove that there is no pattern, it's arbitrary based on Nvidia's own choice and has very little in the way of consistency. Nvidia has become more consistent since Pascal at least, but only if you ignore Ti models, Supers, and the fact that Turing was split into two distinct RTX and GTX lines.
Geforce 4 Ti4600 = NV25
Geforce FX 5600 = NV31
Geforce 6600 = NV43
Geforce 7600 = G73
Geforce 8600GT = G84
Geforce 9600 = G96 and G94a and b
Geforce GT 260 = GT200-100 and 103
Geforce GTX 460 = GF104
Geforce GT 560 = GF114
Geforce GTX 660 = GK106
Geforce GTX 760 = GK104
Geforce GTX 960 = GM204
Geforce GTX 1060 = GP106
Geforce GTX 1660 = GP106
Geforce RTX 2060 = TU106
Geforce RTX 3060 = GA106
The Supers changed core configuration. 2060Super gained CUDA cores, ROPs, bus-width etc. 2070Super was entirely different silicon, using the 104 die instead of the 106.
Unlike the rebrand of the GTX680 -> GT770, 20-series Supers were different core/ROP/bus/RT/VRAM configurations across the entire product stack with way more than just a 30MHz clock bump and new sticker on the box. Yes, they used the same silicon but it was chopped up very differently with configurations that didn't match any of the prior vanilla 20-series models.
R
T
How are we liking that price of RT yet? Glorious innit And faster VRAM, which is what held the 680 back. I owned a 770, was a pretty neat card, basically GK104 in optimal form. Experience was better than the SLI GTX 660's I had prior to it, and that was a free 'upgrade', so to speak :D Good times... You actually had choices. Now its 'okay, where does my budget land in the stack, and how much crap am I getting along with it that I really don't want to pay for but still do?'
That's literally what Nvidia has been doing since they implemented RTX. To even sell something that looks like we need it, they need to pass their old boundaries every gen. Its a clear break from the norm of several decades, where you might have the occasional 'step away from norm' because competition, etc. - the norm is now that things escalate further every gen, because die size is at a feasible max with 600~ish mm2, shrinks are not enough, raster performance is basically maxed out per sq/mm of die space, and stronger RT eats away at it directly for Nvidia. They're quickly heading for a brick wall, and I think they know it. Numbers might change, but the stack order doesn't. And the stack order relates to chip SKUs. The 104, for example, was for a long time the second biggest chip, and generally populated the x70/x80 slots, and only one other chip was above it. Now there's a 103 slotted in between, so what this reads as, is that the performance delta across the whole stack has expanded a bit. The same thing happened when Nvidia added Titan to Kepler. We just got a new performance level, above the 104. At the same time, the 104 was clearly not all Kepler gave; they sold a dual chip 690 with two 104's inside as well. The gist: 104 was never truly the top of an Nvidia card stack. Ever.
When Pascal released they also started off with maxed out 104's - not much changed at all since Kepler or Maxwell. The 1080ti came later, to succeed the 980ti, and both were cut down 'Titans', another 'rule' that existed since the Titan got pushed to 780s.
This principle still holds true, and you can literally see Nvidia's struggle when they tried to position ADA. The immense gap from the 102 to 103 is annoying. That's why they had trouble selling an even worse than 103 chip as an x80, but in fact, that's what they really do strive to do: 104 = x80. Its called x70ti now, but its clear Nvidia miscalculated here. And they still do, given the price points.
I consider Samsung 8nm an odd one out because its very clear now those chips/node were pretty crappy. The foundry basically forced the reality on Nvidia there, you can bet your ass they really didn't plan on pushing x80 on 102 there with an abysmal 1700mhz clock and still horrible perf/w. The 3090/ti story is of similar ridiculousness, ti barely has a right to exist.
The waters do become very muddy when you go below the 106's, but at and above that, there is a pattern and its pretty clear.
What's your timeframe at which point you consider this consistent ? 20, 30 Years ? Centuries ?
Based on this metric you can compare the cards and tell which one is which ones successor in Nvidia's mind so to speak. Whoever said the GTX 680 was supposed to be a midrange card (at least based on the TDP of previous generations) is imo right but market circumstances "promoted" it into the 80 class (AMD was slower).
As it happens I made an excel table based on TDP. I have another one with AMD cards if you're interested.
I think this table shows that the cards above 250W are more expensive now exactly because they relatively perform better than "old day&age" high-end SKUs (those had 250W TDP).
As you can see Nvidia moved up the numbers so at the same TDP you either get a "lower class" GPU for the same price or you have to pay more for the "same" class with the same naming. People usually upgrade within the same class so when they see the lower class card for the same money they rather buy the more expensive card. Brilliant and very simple marketing by Nvidia.
But this is just my thinking.
I scribbled some lines.