Monday, April 29th 2024
NVIDIA Builds Exotic RTX 4070 From Larger AD103 by Disabling Nearly Half its Shaders
A few batches of GeForce RTX 4070 graphics cards are based on the 5 nm "AD103" silicon, a significantly larger chip than the "AD104" that powers the original RTX 4070. A reader has reached out to us with a curiously named MSI RTX 4070 Ventus 3X E 12 GB OC graphics card, saying that TechPowerUp GPU-Z wasn't able to detect it correctly. When we took a closer look at their GPU-Z submission data, we found that the card was based on the larger "AD103" silicon, looking at its device ID. Interestingly, current NVIDIA drivers, such as the 552.22 WHQL used here, are able to seamlessly present the card to the user as an RTX 4070. We dug through older versions of GeForce drivers, and found that the oldest driver to support this card is 551.86, which NVIDIA released in early-March 2024.
The original GeForce RTX 4070 was created by NVIDIA by enabling 46 out of 60 streaming multiprocessors (SM), or a little over 76% of the available shaders. To create an RTX 4070 out of an "AD103," NVIDIA would have to enable 46 out of 80, or just 57% of the available shaders, and just 36 MB out of the 64 MB available on-die L2 cache. The company would also have to narrow the memory bus down to 192-bit from the available 256-bit, to drive the 12 GB of memory. The PCB footprint, pin-map, and package size of both the "AD103" and "AD104" are similar, so board partners are able to seamlessly integrate the chip with their existing AD104-based RTX 4070 board designs. End-users would probably not even notice the change until they fire up diagnostic utilities and find them surprised.Why NVIDIA would make RTX 4070 using the significantly larger "AD103" silicon, is anyone's guess—the company probably has a stash of chips that are good enough to match the specs of the RTX 4070, so it would make sense to harvest the RTX 4070 out of them, which could sell for at least $500 in the market. This also opens up the possibility of RTX 4070 SUPER cards based on this chip, all NVIDIA has to do is dial up the SM count to 56, and increase the L2 cache available to 48 MB. How the switch to AD103 affects power and thermals, is an interesting thing to look out for.
Our next update of TechPowerUp GPU-Z will be able to correctly detect RTX 4070 cards based on AD103 chips.
The original GeForce RTX 4070 was created by NVIDIA by enabling 46 out of 60 streaming multiprocessors (SM), or a little over 76% of the available shaders. To create an RTX 4070 out of an "AD103," NVIDIA would have to enable 46 out of 80, or just 57% of the available shaders, and just 36 MB out of the 64 MB available on-die L2 cache. The company would also have to narrow the memory bus down to 192-bit from the available 256-bit, to drive the 12 GB of memory. The PCB footprint, pin-map, and package size of both the "AD103" and "AD104" are similar, so board partners are able to seamlessly integrate the chip with their existing AD104-based RTX 4070 board designs. End-users would probably not even notice the change until they fire up diagnostic utilities and find them surprised.Why NVIDIA would make RTX 4070 using the significantly larger "AD103" silicon, is anyone's guess—the company probably has a stash of chips that are good enough to match the specs of the RTX 4070, so it would make sense to harvest the RTX 4070 out of them, which could sell for at least $500 in the market. This also opens up the possibility of RTX 4070 SUPER cards based on this chip, all NVIDIA has to do is dial up the SM count to 56, and increase the L2 cache available to 48 MB. How the switch to AD103 affects power and thermals, is an interesting thing to look out for.
Our next update of TechPowerUp GPU-Z will be able to correctly detect RTX 4070 cards based on AD103 chips.
57 Comments on NVIDIA Builds Exotic RTX 4070 From Larger AD103 by Disabling Nearly Half its Shaders
RT performance is improving faster than non-RT performance, since general performance per gen including all the raster/shader/FLOPS stuff is more like 25-50% depending on which models you pick to compare.
So even relative to the non-RT performance gains per generation, the RT hardware is improving at about double that rate. How is that a "glacial pace" of RT improvement?
Feel free to explain it in a way that makes sense, but as far as I can see it's pretty black and white; RT performance is improving faster than anything else in the GPU space right now. It's the opposite of glacial.