Monday, April 29th 2024

NVIDIA Builds Exotic RTX 4070 From Larger AD103 by Disabling Nearly Half its Shaders

A few batches of GeForce RTX 4070 graphics cards are based on the 5 nm "AD103" silicon, a significantly larger chip than the "AD104" that powers the original RTX 4070. A reader has reached out to us with a curiously named MSI RTX 4070 Ventus 3X E 12 GB OC graphics card, saying that TechPowerUp GPU-Z wasn't able to detect it correctly. When we took a closer look at their GPU-Z submission data, we found that the card was based on the larger "AD103" silicon, looking at its device ID. Interestingly, current NVIDIA drivers, such as the 552.22 WHQL used here, are able to seamlessly present the card to the user as an RTX 4070. We dug through older versions of GeForce drivers, and found that the oldest driver to support this card is 551.86, which NVIDIA released in early-March 2024.

The original GeForce RTX 4070 was created by NVIDIA by enabling 46 out of 60 streaming multiprocessors (SM), or a little over 76% of the available shaders. To create an RTX 4070 out of an "AD103," NVIDIA would have to enable 46 out of 80, or just 57% of the available shaders, and just 36 MB out of the 64 MB available on-die L2 cache. The company would also have to narrow the memory bus down to 192-bit from the available 256-bit, to drive the 12 GB of memory. The PCB footprint, pin-map, and package size of both the "AD103" and "AD104" are similar, so board partners are able to seamlessly integrate the chip with their existing AD104-based RTX 4070 board designs. End-users would probably not even notice the change until they fire up diagnostic utilities and find them surprised.
Why NVIDIA would make RTX 4070 using the significantly larger "AD103" silicon, is anyone's guess—the company probably has a stash of chips that are good enough to match the specs of the RTX 4070, so it would make sense to harvest the RTX 4070 out of them, which could sell for at least $500 in the market. This also opens up the possibility of RTX 4070 SUPER cards based on this chip, all NVIDIA has to do is dial up the SM count to 56, and increase the L2 cache available to 48 MB. How the switch to AD103 affects power and thermals, is an interesting thing to look out for.

Our next update of TechPowerUp GPU-Z will be able to correctly detect RTX 4070 cards based on AD103 chips.
Add your own comment

57 Comments on NVIDIA Builds Exotic RTX 4070 From Larger AD103 by Disabling Nearly Half its Shaders

#26
AvrageGamr
thesmokingmanLmao, EXOTIC?
Higher boost clock. My 4070 overclocked just matches it.
Posted on Reply
#27
TheDeeGee
64KCan't happen. It's all about pushing Ray Tracing with Nvidia for years now. Hence the RTX branding.
You know path tracing is the future right? Eventually an iGPU will be able to render it.
Posted on Reply
#29
64K
TheDeeGeeYou know path tracing is the future right? Eventually an iGPU will be able to render it.
Yes, but bear in mind that games only get more and more demanding on hardware as the years go by. That is what has held back 4K adoption since 4K came on the market 12 years ago. The price for a 4K monitor had been down to reasonable levels for years now but the adoption rate is still only around 3.5%
Posted on Reply
#30
HeadRusch1
Nvidia is anxious to sell us all the latest and greatest expensive hardware so we can REALLY get excited about playing:.........Fortnight?.....GTAV Online?.....Witcher III? Sorry, I'm neglecting the big games.....Starfield.....errr....Cyberpunk 2077's benchmark mode.....uhhhhh. HellDivers 2? No wait that runs even on AMD APU's......

Release all the hardware you want, there reaches a point where nobody cares anymore when the software is garbage, and software is garbage because instead of making good games based on previous sales figures, all they are doing is forcing devs to focus on the buzzwords. "Our game NEEDS AI and it needs PATH TRACING......and NFT's!!!....huh?...Oh..wait....<side conversation>...ok the NFT's can be patched-in later!".....

Smells like 83/84 again......"so all the new software is shovelware but they still want us to buy new systems every 2 years to re-play the old games but with upgraded graphics, you say?" :)
Posted on Reply
#31
GoldenTiger
AnotherReaderGiven how poorly even today's highest end hardware performs in path traced games and taking into account that non Apple iGPUs selling today aren't even as fast as a 290X or a 780 Ti from over ten years ago, that's a rather optimistic prediction.

... and then you turn on frame generation and it's nice and smooth with an imperceptible latency difference. That "useless" ai stuff people love to whine about. Also dlss, but I don't know how it looks on 1080p dlss quality as I haven't run that low a resolution since 2004. Been on 1600p since 2008, and 4k since 2014. Dlss q at those resolutions is so close to flawless on every game I've tried it on, it's basically a free performance boost.
Posted on Reply
#32
DarkDreams
Chrispy_So about 8-10% more power-hungry in the case of the 2060KO. Not a disaster, but also not great either.
Keep in mind that the powerlimits are different for the cards in that graphic:

So it's less than 5%
Posted on Reply
#33
mate123
If i remember correctly the 2060 KO was inexplicably a small bit faster than the regular 2060 in certain workloads (beyond the margin of error) despite having been cut down to the same specs. It would be interesting to see if this also the case with these
Posted on Reply
#34
stimpy88
Have nGreedia stopped manufactoring the smaller 40x0 chips to get ready for an early 50x0 launch? Probably not, but fun to speculate.
Posted on Reply
#35
Chrispy_
AnotherReaderGiven how poorly even today's highest end hardware performs in path traced games and taking into account that non Apple iGPUs selling today aren't even as fast as a 290X or a 780 Ti from over ten years ago, that's a rather optimistic prediction.

Yeah, I've been on the RTX train since launch and threw a 3090 at path tracing when it first popped up in CP2077, which was a total disaster.
So I tried it on a 4070 (non-super) and it was pretty lousy even with balanced DLSS, ray-reconstruction and frame-gen. Even if I was getting a reasonable framerate thanks to frame-gen, the number of actual raytracing samples it was working with at a reduced resolution every other frame made the resulting image quality absolute shit-tier garbage in motion.

Sure, path-tracing looks great in still screenshots once you've had a couple of seconds of static image to fill in all the missing data over the course of 100+ frames, but the in-motion, per-frame level of noise and distortion is abysmal on the 4070. With the 4090 having roughly double the raytracing performance in CP2077 according to this chart, I can't imagine the 4090 experience is much better. Ray reconstruction makes the reflections look pretty decent but everything else is a laggy, shadow-crawling mess of delayed lighting and out-of-date temporal information in the scene.



Honestly, I liked the raytraced reflections in CP2077 but the shadow and GI stuff isn't great even without path-tracing, it simply doesn't work in motion, only in static screenshots (by which I mean screenshots taken when the camera is static). With path tracing and ray-reconstruction, the GI and reflections are fantastic but the low framerate means that the already dubious RT shadows and ambient occlusion delay are an even worse, unmissable eyesore that stick out so badly that it completely ruins the entire image in motion.

There's simply not enough raw information per frame to handle shadows in realtime. "Slow light" simply just work in an FPS game where even when you're stood still the shadows crawly like a Lovecraftian ooze and every moving object has a bright trail absent of shadow where shadow should be at its darkest - as well as any new part of the image coming into view as you turn your view to glance sideways always being fullbright for several frames as the temporal-filter slowly fills in shadows and occlusion. It's hard to ignore when it's often half of your entire screen lit incorrectly in something as simple as turning your head to look left at any street junction or turn of any corridor.

I'm at work, but I will take a screenshot at home if I remember of CP2077 at 4K with DLSS3.5 FG RR path-traced as I turn my camera to check a street junction. From memory, these screenshots look like total ass for around 1/3rd of the image that's just come into view. Anyone accepting that as a superior image quality to baked-and-faked is crazy, in my opinion.
Posted on Reply
#36
ARF
64KYes, but bear in mind that games only get more and more demanding on hardware as the years go by. That is what has held back 4K adoption since 4K came on the market 12 years ago. The price for a 4K monitor had been down to reasonable levels for years now but the adoption rate is still only around 3.5%
No. There is a very strong anti-4K propaganda. You actually get a higher image quality if you switch from 1K to 4K, instead of staying at 1K with some glossy effects aka ray-tracing...
Posted on Reply
#37
Chrispy_
So I said I'd post two path-traced screenshots - This is raytracing in motion vs a static scene. I took a screenshot during a camera move that represents actual gameplay and then after pasting the image into PhotoShop I took a second screenshot without movement as close to the output of the first screenshot as possible.

In-motion screenshot as I turn to to face the street on the stairs to V's Megacomplex apartment:

This screenhot is 100% authentic gameplay and highlights the shortcomings of both DLSS and raytracing with it's jaggies and godawful splotchy lighting. There's obvious visible crawl of any and all lighting in the scene and you can repeat this experiment yourself by simply enabling overdrive and hitting PrintcSreen in typical gameplay motion. The uneven shadow/lighting mess in the screenshot above isn't static, it crawls like a tentacle-endowed Lovecraftian horror so an animated .gif of this in slow-mo would look orders of magnitude worse!

What an RT Overdrive screenshot looks like if you give it a second or so (40-50 frames of temporal data):

This is what RT frames should look like, but this is the temporal average of several hundred frames and doesn't represent real gameplay! In other words, it's a hoax. The lighting is a total mess in motion and it looks like ass - nothing like this second static screenshot. Yes, I said 'ass' because I lack the vocabulary to better describe this splotchy, irregular mess, but if the 4070 isn't remotely close to a reasonable RT lighting output, I'm going to guess that the 4090 is only half as bad, and "semi-ass" is no compliment for $2000 of cutting-edge hardware that's beyond the reach of 99% of all gamers.

Meanwhile, here is what "raster" looks like (in motion) to compare against the first screenshot in this post:

This "fake" lighting will run at 160+ FPS without framegen-lag on my 4070. Youtube proves that you can get decent, true 60+fps results from entry-level GPUs costing $200, which is a miles better experience than the approximately 30fps input lag of framegen with path-tracing on far more expensive hardware on my $600 4070. Sure, the lighting is different, but it isn't anywhere near as bad as the RTX splotchy, blurry, mess. Anyone with half-decent hardware can make their own decision but I'm not personally keen on the inaccurate, Lovecraftian nightmare of RTX realtime* lighting. It's low quality, inconsistent garbage that requires you to spend quadruple on hardware for the low-quality result.

Realistically, human vision isn't a 1:1 translation of screenshots, but you need to be practically blind to miss the artifacts and problems with realtime RT. We're so far off the 'realtime' RT reality that Nvidia aspire to that I don't ever think we'll get there. Game developers increase scene complexity to match GPU capabilities. Unless a GPU magically appears with 20x more horsepower than the current expectations, the in-game scenes will always be too complex to render on any given GPU of the same generation.

* - intentional obvious sarcasm.

Let us not forget that this example I'm using is the definitive Nvidia-sponsored, Nvidia-funded, total RTX experience in its best possible light. No other game comes close to this level of DLSS and raytracing development effort.
Posted on Reply
#38
stimpy88
Chrispy_So I said I'd post two path-traced screenshots - This is raytracing in motion vs a static scene. I took a screenshot during a camera move that represents actual gameplay and then after pasting the image into PhotoShop I took a second screenshot without movement as close to the output of the first screenshot as possible.

In-motion screenshot as I turn to to face the street on the stairs to V's Megacomplex apartment:

This screenhot is 100% authentic gameplay and highlights the shortcomings of both DLSS and raytracing with it's jaggies and godawful splotchy lighting. There's obvious visible crawl of any and all lighting in the scene and you can repeat this experiment yourself by simply enabling overdrive and hitting PrintcSreen in typical gameplay motion. The uneven shadow/lighting mess in the screenshot above isn't static, it crawls like a tentacle-endowed Lovecraftian horror so an animated .gif of this in slow-mo would look orders of magnitude worse!

What an RT Overdrive screenshot looks like if you give it a second or so (40-50 frames of temporal data):

This is what RT frames should look like, but this is the temporal average of several hundred frames and doesn't represent real gameplay! In other words, it's a hoax. The lighting is a total mess in motion and it looks like ass - nothing like this second static screenshot. Yes, I said 'ass' because I lack the vocabulary to better describe this splotchy, irregular mess, but if the 4070 isn't remotely close to a reasonable RT lighting output, I'm going to guess that the 4090 is only half as bad, and "semi-ass" is no compliment for $2000 of cutting-edge hardware that's beyond the reach of 99% of all gamers.

Meanwhile, here is what "raster" looks like (in motion) to compare against the first screenshot in this post:

This "fake" lighting will run at 160+ FPS without framegen-lag on my 4070. Youtube proves that you can get decent, true 60+fps results from entry-level GPUs costing $200, which is a miles better experience than the approximately 30fps input lag of framegen with path-tracing on far more expensive hardware on my $600 4070. Sure, the lighting is different, but it isn't anywhere near as bad as the RTX splotchy, blurry, mess. Anyone with half-decent hardware can make their own decision but I'm not personally keen on the inaccurate, Lovecraftian nightmare of RTX realtime* lighting. It's low quality, inconsistent garbage that requires you to spend quadruple on hardware for the low-quality result.

Realistically, human vision isn't a 1:1 translation of screenshots, but you need to be practically blind to miss the artifacts and problems with realtime RT. We're so far off the 'realtime' RT reality that Nvidia aspire to that I don't ever think we'll get there. Game developers increase scene complexity to match GPU capabilities. Unless a GPU magically appears with 20x more horsepower than the current expectations, the in-game scenes will always be too complex to render on any given GPU of the same generation.

* - intentional obvious sarcasm.

Let us not forget that this example I'm using is the definitive Nvidia-sponsored, Nvidia-funded, total RTX experience in its best possible light. No other game comes close to this level of DLSS and raytracing development effort.
Absolutely correct. True RT is about 12-15 years away, at least, maybe. Everything so far is tricks, and not very good ones at that.

I would guess the higer end cards of the RTX 8000 series might be the first to do real-time RT at 720p 30fps.

TBH we need a new way of rendering games, I'm not sure current style RT rendering can ever be truly realtime.
Posted on Reply
#39
Chrispy_
stimpy88I would guess the higher end cards of the RTX 8000 series might be the first to do real-time RT at 720p 30fps.
In a 2020 game like CP2077? Sure I can believe that. They'll still need to use temporal dithering but perhaps they have enough samples in just 2 frames to make a decent image, rather than requiring 20 frames to get anywhere close to a decent output.

The problem is that game graphics keep getting pushed forward to match the capabilities of current GPUs, so once the RTX 8000 series arrives, we will probably have games that push an order of magnitude more polygons...
Posted on Reply
#40
Assimilator
Chrispy_Unless a GPU magically appears with 20x more horsepower than the current expectations, the in-game scenes will always be too complex to render on any given GPU of the same generation.
The same claim is made literally every time a new GPU generation is released, and it's always disproven a few generations later. Especially given that this is only the third generation of GPUs with dedicated ray-tracing hardware, it seems rather premature to claim that real-time ray-tracing will never be possible.
Posted on Reply
#41
Chrispy_
AssimilatorThe same claim is made literally every time a new GPU generation is released, and it's always disproven a few generations later. Especially given that this is only the third generation of GPUs with dedicated ray-tracing hardware, it seems rather premature to claim that real-time ray-tracing will never be possible.
My point is that game requirements increase to keep pace with hardware, so it may be a loooong time until RT is the default render path.
When the RTX 11070TiSuper comes out in 2034, you can be sure there will be game developers making games that will bring it to its knees and max out its VRAM.
Posted on Reply
#42
Assimilator
Chrispy_My point is that game requirements increase to keep pace with hardware, so it may be a loooong time until RT is the default render path.
When the RTX 11070TiSuper comes out in 2034, you can be sure there will be game developers making games that will bring it to its knees and max out its VRAM.
And Cyberpunk is one such game, yet you are using this outlier to justify your argument that real-time RT/PT will in general never be the default renderer. This basic fallacy entirely invalidates the claim being made.
Posted on Reply
#43
Chrispy_
AssimilatorAnd Cyberpunk is one such game, yet you are using this outlier to justify your argument that real-time RT/PT will in general never be the default renderer. This basic fallacy entirely invalidates the claim being made.
I picked Cyberpunk because it's clearly the game that's had the most work and effort put into it by Nvidia for RTX features. It's definitely an outlier, but it's the BEST CASE SCENARIO for current raytracing, not a poor example that doesn't represent raytracing fairly.

More realistic examples of fully-raytraced games are Portal RTX and Quake II RTX. For Quake II, it took a 2080Ti to get 4K60 path-traced, in a 22-year-old game at that point in time.
Posted on Reply
#44
stimpy88
Another thing that doesn't help is nGreedias glacial pace at improving RT performance. It's only really the current gen series that has moved the needle substantially on RT performance. I sincerely hope that the Blackwell series also has a marked uptick in RT performance. RT is almost useless on anything less than a 4090. That level of RT performance needs to be available on the 5070 series cards, to help bring playable RT to the "mid range", obviously I know we are not talking of full RT here.
Posted on Reply
#45
Chrispy_
stimpy88Another thing that doesn't help is nGreedias glacial pace at improving RT performance. It's only really the current gen series that has moved the needle substantially on RT performance.
What? 2080Ti > 3080Ti was a bigger jump in RT performance than 3090 to 4090. Maybe what you say is true for other parts of the product stack but nobody is doing the latest detailed benchmarks with mid-tier 20-series by the looks of it.



I'm not sure the 2080Ti can path trace or not. I'm seeing no benchmarks anywhere so maybe it's a "30-series-and-above" feature in CP2077, the single heaviest RT title I can think of.
Posted on Reply
#46
Assimilator
Chrispy_What? 2080Ti > 3080Ti was a bigger jump in RT performance than 3090 to 4090. Maybe what you say is true for other parts of the product stack but nobody is doing the latest detailed benchmarks with mid-tier 20-series by the looks of it.



I'm not sure the 2080Ti can path trace or not. I'm seeing no benchmarks anywhere so maybe it's a "30-series-and-above" feature in CP2077, the single heaviest RT title I can think of.
You're responding to someone who unironically uses the phrase "ngreedia", don't expect much in the way of capability to absorb facts.
Posted on Reply
#47
mubarak ahmed

Not only MSI, I guess.
I bought this GIGABYTE WINDFORCE RTX 4070 and it has an AD103 chip too.
Posted on Reply
#48
Dr. Dro
mubarak ahmed
Not only MSI, I guess.
I bought this GIGABYTE WINDFORCE RTX 4070 and it has an AD103 chip too.
Nice find. Are there any clues regarding the approximate manufacturing date of your card in its box, BIOS or documentation?
Posted on Reply
#49
mubarak ahmed
Dr. DroNice find. Are there any clues regarding the approximate manufacturing date of your card in its box, BIOS or documentation?
I tried searching the box for any manufacturing date but I didn't find anything, and the BIOS is not available in the database
Posted on Reply
#50
stimpy88
Chrispy_What? 2080Ti > 3080Ti was a bigger jump in RT performance than 3090 to 4090. Maybe what you say is true for other parts of the product stack but nobody is doing the latest detailed benchmarks with mid-tier 20-series by the looks of it.



I'm not sure the 2080Ti can path trace or not. I'm seeing no benchmarks anywhere so maybe it's a "30-series-and-above" feature in CP2077, the single heaviest RT title I can think of.
And the performance of the non-RT hardware in the newer cards is not having an impact... :kookoo: Clockspeeds, shaders, VRAM bandwidth etc? Try looking at the performance hit from enabling RT.
Posted on Reply
Add your own comment
Nov 22nd, 2024 00:30 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts