Monday, December 5th 2022

January 5 Release Date Predicted for NVIDIA GeForce RTX 4070 Ti

A January 5, 2023 release date is being mooted by retailers for NVIDIA's upcoming GeForce RTX 4070 Ti graphics card, which is widely expected to be a re-branding of what would have been the RTX 4080 12 GB. Italian retailer Drako started a countdown for an ASUS TUF Gaming RTX 4070 Ti O12G custom-design graphics card, which winds down to January 5, and aligns with the rumored January 3 announcement of the card. It is also expected that reviews of the RTX 4070 Ti will be allowed to go live on January 4.

The GeForce RTX 4080 12 GB was supposed to max out the 4 nm AD104 silicon, featuring 7,680 CUDA cores across 60 SM (streaming multiprocessors), 60 RT cores, 240 Tensor cores, 240 TMUs, and 80 ROPs. The GPU features a 192-bit wide GDDR6X memory interface, to which NVIDIA is giving 21 Gbps-rated memory, yielding 504 GB/s of memory bandwidth. Its most interesting aspect is its power configuration, with a typical board power of 285 W, which makes it technically possible for board partners to use two 8-pin PCIe power connectors, unless they've been asked nicely to implement the 16-pin 12VHPWR connector.
Sources: Drako.it, Wccftech, VideoCardz
Add your own comment

59 Comments on January 5 Release Date Predicted for NVIDIA GeForce RTX 4070 Ti

#51
AusWolf
Vayra86Right, because they didn't learn from Pascal?


I mean, I wouldn't blame them. Who thought Etherum was going to switch to Proof-of-Stake, effectively crashing the mining market? It was hanging in the air for years, but nobody knew when it would actually happen, or if it would at all.
Posted on Reply
#52
Bwaze
Etherum switched to Proof-of-Stake well after the crypto crash - in September 2022. And no, crypto didn't crash because Ethereum announced change to Proof-of-Stake - that was planned and postponed since 2019. Home mining on gaming GPUs has very little impact on overall crypto market, all it does is it fucks up the GPU market. People used to praise it publicly on forums like this one, now they know better.
Posted on Reply
#53
AusWolf
BwazeEtherum switched to Proof-of-Stake well after the crypto crash - in September 2022. And no, crypto didn't crash because Ethereum announced change to Proof-of-Stake - that was planned and postponed since 2019. Home mining on gaming GPUs has very little impact on overall crypto market, all it does is it fucks up the GPU market. People used to praise it publicly on forums like this one, now they know better.
There's no denying that, but I'm talking about the GPU market, which was hugely f-ed up due to crypto mining. The other thing is that a GPU architecture's roadmap spans across several years. When Nvidia was only making plans for the transition from Ampere to Ada was well within the crypto boom. They could have easily counted on Etherum not making the switch for a couple more years back then.

Edit: I find it supported by the fact that Ada is basically a larger, more efficient Ampere made on a smaller node. Not much effort has been made to improve on the architecture, which makes me believe that gaming wasn't the main focus in development. Sure, there's DLSS 3 and stuff, but those improvements are usually a side note, not the main focus.
Posted on Reply
#54
Bwaze
Well the way they did Ada was going for the low hanging fruits, new process at TSMC really enabled quite a jump from inefficient Samsung one, so they got to pocket the money from 2020-2022 crypto high and save the money on redesign!

And then have the audacity to tell us that's really expensive, so we should be glad it's only 70% price increase.
Posted on Reply
#55
Vayra86
AusWolfThere's no denying that, but I'm talking about the GPU market, which was hugely f-ed up due to crypto mining. The other thing is that a GPU architecture's roadmap spans across several years. When Nvidia was only making plans for the transition from Ampere to Ada was well within the crypto boom. They could have easily counted on Etherum not making the switch for a couple more years back then.

Edit: I find it supported by the fact that Ada is basically a larger, more efficient Ampere made on a smaller node. Not much effort has been made to improve on the architecture, which makes me believe that gaming wasn't the main focus in development. Sure, there's DLSS 3 and stuff, but those improvements are usually a side note, not the main focus.
Interesting point. But: gaming is the main focus for Geforce and Ada is a gaming arch. They have Hopper and other for datacenter/enterprise, they have quadro RTX cards for pro. But its all more of the same, with tweaks, additions or omissions.

In the end its a chip that works and improves through iterative technological upgrades/adjustments. Both Nvidia and AMD have historically been doing a dance with what they do or don't enable on these chips, and how this suits different purposes or markets. But in the end its a huge amount of programmable cores that can work in parallel. We've seen Nvidia's CUDA based arch evolve over time, and Geforce was already thinned down to pure gaming (Pascal), GPGPU capability removed; then we got Volta and they borrowed tensor cores from it to transplant into the Pascal 'base' (=Turing) and added some cache and sauce so now we have an RT core too. But in the usage/purpose, there are similarities between everything Nvidia releases, and always have been. The usage of machine learning/'AI' to produce frames is effectively also a gaming improvement that originates from another corner of the market.

I think they've reached the end of the line with the fruits they can harvest for a more efficient gaming GPU based on raster alone. The marketing on Geforce kinda follows that logic, its getting separated further and further from reality, just like with Intel's chaotic explanation and changes to how turbo and frequency works, resulting in product stacks that become largely unusable or provide parts you'd rather undervolt; or priced way out of comfort because they must adhere to inflated marketing claims. You're right, what used to 'seem' side note like DLSS version upgrade is now front and center; but compare it to Maxwell > Pascal! That was similarly mostly just a good shrink and some improvements to power delivery (Ada has almost the same powerpoint slides :D) enabling higher frequencies. Except today, they barely pay off on their own; diminishing returns happen, especially as lots of raster components in the pipeline are becoming more dynamic.

TL DR I agree with the idea the focus has indeed shifted since Volta, but it remains plausible that still was the best way forward for their gaming lineup.
Posted on Reply
#56
AusWolf
Vayra86Interesting point. But: gaming is the main focus for Geforce and Ada is a gaming arch. They have Hopper and other for datacenter/enterprise, they have quadro RTX cards for pro. But its all more of the same, with tweaks, additions or omissions.
Yeah, but they didn't have a separation between gaming and mining cards since the unpopular (and as I've seen, unavailable) CMP series. When you say "Ada is a gaming arch", it is kind of synonymous with "Ada is a mining arch". Considering how similar Ada is to Ampere in its physical design, I would say mining was a more likely target than gaming, as miners were the ones who needed more of the same. Gamer architectures tend to undergo more changes as games change in their approach to graphics.
Vayra86I think they've reached the end of the line with the fruits they can harvest for a more efficient gaming GPU based on raster alone. The marketing on Geforce kinda follows that logic, its getting separated further and further from reality, just like with Intel's chaotic explanation and changes to how turbo and frequency works, resulting in product stacks that become largely unusable or provide parts you'd rather undervolt; or priced way out of comfort because they must adhere to inflated marketing claims.
I agree - but then logic dictates that they should have focused on things that aren't at the end of the line, like RT for example. Ada seems to show exactly the same performance drop when RT is on compared to off as Ampere, which is a real shame, imo. Cramming more RT cores on the die relative to traditional shaders, or just making the existing ones faster would have paid off much more, especially since AMD doesn't seem to be focusing on improving RT with RDNA 3, either. Nobody cares if you have 600 FPS in CS:GO instead of 400, so having more of the same is again, a choice that would have catered to miners much more than gamers.
Posted on Reply
#57
Bwaze
AusWolfConsidering how similar Ada is to Ampere in its physical design, I would say mining was a more likely target than gaming, as miners were the ones who needed more of the same. Gamer architectures tend to undergo more changes as games change in their approach to graphics.
I'd say gaming changed right about 0 from arrival of Ampere to Ada. We don't even have a single AA game that they would promote as a showcase of new RTX, DLSS iterations, they are showing them in patches to old games, or in Portal fully path-traced RTX mod - which only shows we're not there by a long shot...

Remember the GeForce RTX 2080 Ti Cyberpunk 2077 Edition, that came out months before game was finally launched, and was actually then too slow to run the ray-tracing in game properly?
:-D
Posted on Reply
#58
Vayra86
AusWolfYeah, but they didn't have a separation between gaming and mining cards since the unpopular (and as I've seen, unavailable) CMP series. When you say "Ada is a gaming arch", it is kind of synonymous with "Ada is a mining arch". Considering how similar Ada is to Ampere in its physical design, I would say mining was a more likely target than gaming, as miners were the ones who needed more of the same. Gamer architectures tend to undergo more changes as games change in their approach to graphics.


I agree - but then logic dictates that they should have focused on things that aren't at the end of the line, like RT for example. Ada seems to show exactly the same performance drop when RT is on compared to off as Ampere, which is a real shame, imo. Cramming more RT cores on the die relative to traditional shaders, or just making the existing ones faster would have paid off much more, especially since AMD doesn't seem to be focusing on improving RT with RDNA 3, either. Nobody cares if you have 600 FPS in CS:GO instead of 400, so having more of the same is again, a choice that would have catered to miners much more than gamers.
Interesting points! I'm more of the opinion (we'll never really know will we...) Nvidia's mining performance is a side effect, not a target for them. It would present major risk to their continued business strategy to make mining oriented GPUs. They just repurposed existing cards.
Posted on Reply
#59
bug
Vayra86Interesting points! I'm more of the opinion (we'll never really know will we...) Nvidia's mining performance is a side effect, not a target for them. It would present major risk to their continued business strategy to make mining oriented GPUs. They just repurposed existing cards.
I believe you are correct. GPUs only started doing compute when rendering shaders became flexible enough to tackle compute tasks. And later on, someone figured since the compute capabilities on a GPU are highly parallel, they could use that for mining.
If Nvidia was actively trying to target miners, there's a lot of GPU silicon they could remove, building dedicated, smaller dies. As it stands, their compute cards only use the same dies meant for gaming, albeit with (a lot of) disabled shaders.
Posted on Reply
Add your own comment
Dec 18th, 2024 02:11 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts