Friday, February 9th 2024

Widespread GeForce RTX 4080 SUPER Card Shortage Reported in North America

NVIDIA's decision to shave off $200 from its GeForce RTX 4080 GPU tier has caused a run on retail since the launch of SUPER variants late last monthVideoCardz has investigated an apparent North American supply shortage. The adjusted $999 base MSRP appears to be an irresistible prospect for discerning US buyers—today's report explains how: "a week after its release, that GeForce RTX 4080 SUPER cards are not available at any major US retailer for online orders." At the time of writing, no $999 models are available to purchase via e-tailers (for delivery)—BestBuy and Micro Center have a smattering of baseline MSRP cards (including the Founders Edition), but for in-store pickup only. Across the pond, AD103 SUPER's supply status is a bit different: "On the other hand, in Europe, the situation appears to be more favorable, with several retailers listing the cards at or near the MSRP of €1109."

The cheapest custom GeForce RTX 4080 SUPER SKU, at $1123, seems to be listed by Amazon.com. Almost all of Newegg's product pages are displaying an "Out of Stock" notice—ZOTAC GAMING's GeForce RTX 4080 SUPER Trinity OC White Edition model is on "back order" for $1049.99, while the only "in stock" option is MSI's GeForce RTX 4080 Super Expert card (at $1149.99). VideoCardz notes that GeForce RTX 4070 SUPER and RTX 4070 TI SUPER models are in plentiful supply, which highlights a big contrast in market conditions for NVIDIA's latest Ada Lovelace families. The report also mentions an ongoing shortage of GeForce RTX 4080 (Non-SUPER) cards, going back weeks prior to the official January 31 rollout: "Similar to the RTX 4090, finding the RTX 4080 at its $1200 price point has proven challenging." Exact sales figures are not available to media outlets—it is unusual to see official metrics presented a week or two after a product's launch—so we will have to wait a little longer to find out whether demand has far outstripped supply in the USA.
Source: VideoCardz
Add your own comment

95 Comments on Widespread GeForce RTX 4080 SUPER Card Shortage Reported in North America

#51
Dr. Dro
Vya DomusThe reason this is wrong is because Nvidia adds more memory to their cards all the time, point in case the 4070ti super.

So how am I supposed to interpret this ? Nvidia needs to stand out as well now ? Why ?

Nah, that's not it. Nvidia is simply penny pinching and crippling the VRAM on their cards intentionally. They've been at it since the early 2000s, that's why their cards have always had these bizarre VRAM configurations, 384mb, 768mb, 896mb, 1.2GB, 1.5GB, etc.

AMD isn't trying to stand out as much as Nvidia is trying to skimp on VRAM, that much is clear.
The 4070 Ti Super is built out of a crippled AD103, not AD104. The AD104 on the 4070 Ti is maxed out. The 192-bit memory interface means that it can have either 6 or 12 memory chips in clamshell, and that translates to real memory capacities of 6, 12 and 24 GB with currently available chips. It's not 6 (not enough) or 24 GB (power, thermal, and market segmentation) for obvious reasons. On the other hand, the AD103 has a 256-bit interface, which translates to 8 or 16 chips and 8, 16 and 32 GB memory capacities being possible. The 4070 Ti Super simple uses 8 of the slower 21Gbps variant of GDDR6X instead to reach its performance and capacity targets. This is impossible to achieve on the AD104 4070 Ti.

As the party that holds both the market and mind share, it isn't Nvidia that needs to do something to stand out, it's the other way around.
Posted on Reply
#52
trsttte
Dr. DroThe 192-bit memory interface means that it can have either 6 or 12 memory chips in clamshell, and that translates to real memory capacities of 6, 12 and 24 GB with currently available chips. It's not 6 (not enough) or 24 GB (power, thermal, and market segmentation) for obvious reasons. On the other hand, the AD103 has a 256-bit interface, which translates to 8 or 16 chips and 8, 16 and 32 GB memory capacities being possible. The 4070 Ti Super simple uses 8 of the slower 21Gbps variant of GDDR6X instead to reach its performance and capacity targets. This is impossible to achieve on the AD104 4070 Ti.
That's their problem, not mine. Asking for $600 and above for a card with 12gb of vram is unreasonable. The 4060ti has a 16gb version (and should only have that version, the regular 4060ti with 8gb is wildly inferior), so why does the card above end up with less? Nvidia was the one to choose to design AD104 with 192bits, hell they even went as far as calling it 4080 and trying to sell it for 900$, it's like they are actively trying to mock consumers to their faces.

They are the ones in control of the design, I think even they get surprised with what they're getting away sometimes, sometimes they push their luck and have to pull back - like the 4060ti and 4080 12gb - but the 4070 seems like a clear example throwing shit into the market and getting away with it as it's still selling anyway.
Posted on Reply
#53
Dr. Dro
trsttteThat's their problem, not mine. Asking for $600 and above for a card with 12gb of vram is unreasonable. The 4060ti has a 16gb version (and should only have that version, the regular 4060ti with 8gb is wildly inferior), so why does the card above end up with less? Nvidia was the one to choose to design AD104 with 192bits, hell they even went as far as calling it 4080 and trying to sell it for 900$, it's like they are actively trying to mock consumers to their faces.

They are the ones in control of the design, I think even they get surprised with what they're getting away sometimes, sometimes they push their luck and have to pull back - like the 4060ti and 4080 12gb - but the 4070 seems like a clear example throwing shit into the market and getting away with it as it's still selling anyway.
Not so unreasonable apparently since it's selling anyway. The only thing you can do is vote with your wallet, and it's a losing battle if you ask me.
Posted on Reply
#54
The Von Matrices
trsttteThat's their problem, not mine. Asking for $600 and above for a card with 12gb of vram is unreasonable. The 4060ti has a 16gb version (and should only have that version, the regular 4060ti with 8gb is wildly inferior), so why does the card above end up with less? Nvidia was the one to choose to design AD104 with 192bits, hell they even went as far as calling it 4080 and trying to sell it for 900$, it's like they are actively trying to mock consumers to their faces.

They are the ones in control of the design, I think even they get surprised with what they're getting away sometimes, sometimes they push their luck and have to pull back - like the 4060ti and 4080 12gb - but the 4070 seems like a clear example throwing shit into the market and getting away with it as it's still selling anyway.
I agree with Dr. Dro in that you should vote with your wallet and (theoretically) companies will respond to that feedback. But also realize that if it is selling well then it could also be that your needs are just different than the needs of most buyers.
Posted on Reply
#55
stimpy88
Dr. Drowww.digikey.com/en/products/detail/micron-technology-inc/MT61K512M32KPA-24-U-TR/17632186

This is the exact memory chip used on the RTX 4080. At $55,700, the per unit cost is $27.85 in a 2000 unit bulk.

$27.85 * 8 = $222.80 in memory IC costs alone.

www.digikey.com/en/products/detail/micron-technology-inc/MT61K512M32KPA-21-U/17631914

The same situation with the slower 21 Gbps chip used in the 4070 Ti and 4090, at $25.31 per unit at bulk cost

$25.31 * 6 = $151.86 in memory costs alone

Not only does the cost add up, but factor in PCB costs, power delivery component costs, etc. - losses, the fact that everyone needs profit, yada yada, you get what I'm going at. I'm fairly sure the BoM for a 4070 Ti must be somewhere in the vicinity of $500 as a rough estimate. The rest makes up for profit, distribution cost, driver software development costs, etc.
Yeah, nGreedia orders millions of memory chips from Digikey every month...
Posted on Reply
#56
Dr. Dro
stimpy88Yeah, nGreedia orders millions of memory chips from Digikey every month...
You overestimate Digikey's margins. Of course Nvidia gets the chips a little cheaper - but here's the kicker, both them and AMD sell AIB partners a validated set comprised of GPU ASIC + memory ICs, so both of them charge basically whatever they want for the memory chips.
Posted on Reply
#57
evernessince
The Von MatricesI agree with Dr. Dro in that you should vote with your wallet and (theoretically) companies will respond to that feedback. But also realize that if it is selling well then it could also be that your needs are just different than the needs of most buyers.
The idea that one can "vote with their wallet" implies the customer has enough power to make a difference to begin with. The reality is we exist in a duoplistic market in which Nvidia is increasingly earning more and more of it's revenue from sources outside the gaming market. By extension consumer power to push for changes is greatly dimminished. This argument might, and I very much stress might, have more pull if Intel becomes competitive but that assumes Intel too doesn't just join AMD and Nvidia in their pricing structure to maintain high margins. The customer has the right to choose in which way they'd like to be taken over a barrel.
Dr. DroDo any of you even remotely believe what you've been saying? It's been proven time and time again that you need to push unrealistic, unreasonable settings to run out of VRAM on 12 GB GPUs today. Certainly settings far beyond the means of an RTX 4070 Ti, and certainly far beyond the means of AMD's equivalents.

There's two games in W1zzard's suite, that at 4K resolution and ultra high settings, require more than 12 GB. if you're playing ultra high 4K, you aren't running a midranger card.
Do you often purchase $800 graphics cards with the expectation that day 1 there will be games that already exceed it's VRAM buffer? Since when did some PC gamers start finding that acceptable? Oh that's right, during the 7 year time period Nvidia failed to increase VRAM allowance across their GPU stack. Sure are fools willing to fight against their own best interests to defend a company that doesn't care about them. That's before considering the 4070 Ti is the better offering than the initial 4070. The 4070 only has 12GB of VRAM, you were picking the best possible example of the two. Such a failure to increase VRAM would have never been acceptable in the past as would paying that kind of money to not have enough VRAM day one, let alone into the future.

I had a 1080 Ti for nearly 7 years and it wasn't until the tail end of that did a few games start showing up that challenged the VRAM. None of this day 1 lacking VRAM nonsense. That card cost me $650 and was the top end card of that generation. Now that much money doesn't even get you an xx70 Ti class card, you get a 4070 with a mere 1GB more VRAM over a 7 year period. Historically VRAM capacity would have doubled two to three times in the same timespan. 780 Ti had 3GB, 980 Ti had 6GB, and 1080 Ti had 11GB. Imagine that, VRAM nearly doubling or doubling gen over gen was pretty normal and there was a mere $50 price difference between 3 generations.

The 1080 Ti also does a good job of demonstrating that if VRAM allowances on cards don't increase, devs won't create games that utilize that additoinal VRAM. 8GB was the standard that game devs were forced to target for a whopping 7 years and heck even today still have to heavily consider it given the high price tag of 12GB+ card on the Nvidia side (which sells some 80%+ of all video cards).

Also lest we not forget another major factor:

"Hardware leads software, not the other way around. It's backwards to imply that because games don't use x amount of VRAM today, x amount of VRAM isn't worthwhile. That kind of logic only perpetuates stagnation of VRAM capacity. Devs aren't going to make games with VRAM requirements the vast majority of cards can't meet. It's up to the hardware vendors to increase GPU VRAM allowance so that game devs can then create games that utilize that extra capacity.

I will never understand people who fight against the idea of increasing VRAM amounts of video cards. VRAM is cheap, enables devs to do more with their games, and increases card longevity. Whether it benefits you at this very moment is an extremely narrow way to look at things."
Posted on Reply
#58
Dr. Dro
evernessinceDo you often purchase $800 graphics cards with the expectation that day 1 there will be games that already exceed it's VRAM buffer? Since when did some PC gamers start finding that acceptable? Oh that's right, during the 7 year time period Nvidia failed to increase VRAM allowance across their GPU stack. Sure are fools willing to fight against their own best interests to defend a company that doesn't care about them. Such a failure to increase VRAM would have never been acceptable in the past as would paying that kind of money to not have enough VRAM day one, let alone into the future.

I had a 1080 Ti for nearly 7 years and it wasn't until the tail end of that did a few games start showing up that challenged the VRAM. None of this day 1 lacking VRAM nonsense. That card cost me $650 and was the top end card of that generation. Now that much money doesn't even get you an xx70 Ti class card, you get a 4070 with a mere 1GB more VRAM over a 7 year period. Historically VRAM capacity would have doubled two to three times in the same timespan. 780 Ti had 3GB, 980 Ti had 6GB, and 1080 Ti had 11GB. Imagine that, VRAM nearly doubling or doubling gen over gen was pretty normal and there was a mere $50 price difference between 3 generations.

The 1080 Ti also does a good job of demonstrating that if VRAM allowances on cards don't increase, devs won't create games that utilize that additoinal VRAM. 8GB was the standard that game devs were forced to target for a whopping 7 years and heck even today still have to heavily consider it given the high price tag of 12GB+ card on the Nvidia side (which sells some 80%+ of all video cards).

Also lest we not forget another major factor:

"Hardware leads software, not the other way around. It's backwards to imply that because games don't use x amount of VRAM today, x amount of VRAM isn't worthwhile. That kind of logic only perpetuates stagnation of VRAM capacity. Devs aren't going to make games with VRAM requirements the vast majority of cards can't meet. It's up to the hardware vendors to increase GPU VRAM allowance so that game devs can then create games that utilize that extra capacity.

I will never understand people who fight against the idea of increasing VRAM amounts of video cards. VRAM is cheap, enables devs to do more with their games, and increases card longevity. Whether it benefits you at this very moment is an extremely narrow way to look at things."
I haven't run into VRAM limitations with any of my GPU purchases in the past 8 years with the sole exception of the R9 Fury X (for obvious reasons), so I couldn't really answer. I normally don't buy lower-end GPUs and that I opted for a 4080 this time is most unusual (and frankly motivated by money, as I wanted to buy an expensive display and overhaul the rest of my PC as well), I always buy cards at the halo tier. The few games I played on a 4 GB GPU recently (my laptop's 3050M) did require compromises in texture quality to be playable, but they were playable nonetheless, at settings that I personally consider acceptable for a low-end laptop GPU.

Anyway, let's be realistic for a moment here. When the 1080 Ti launched 7 years ago in the Pascal refresh cycle, 11 GB was truly massive. And we all know that 11 GB was because Nvidia only removed a memory chip to lower the bandwidth and ensure that the 2016 Titan X retained a similar level of performance (despite having slower memory) and that there was a significant enough gap between it and the then-new Titan Xp, which came with the complete GP102 chip and the fastest memory that they had at the time. For most of its lifetime you could simply enable all settings and never worry about that.

GPU memory capacities began to rise after the Turing and RDNA generation but we were still targeting 8 GB for the performance segment parts (RTX 2080, 5700 XT), and these cards weren't really all that badly choked by these. Then came RDNA 2 and started to give everyone 12 and 16 GB like candy, but that still didn't give them a very distinct advantage over Ampere, that was still lingering on 8-10 GB, and then eventually 12 GB for the high end. Nvidia positioned the RTX 3090 with 24 GB to basically assume both the role of gaming flagship and a replacement for the Titan RTX, while removing a few perks of the latter and technically lowering the MSRP by $1000 (which was tbh an acceptable tradeoff). We know reality is different, but if we exclude the RTX 3090, then only AMD really offered a high VRAM capacity.

This generation started to bring higher VRAM capacities on both sides, and it's only now with games that were designed first and foremost for the PS5 and Xbox Series X starting to become part of gamers' rosters and benchmark suites, all featuring advanced graphics, new texture streaming systems, tons of high-resolution assets, etc. that we've begun to see that being utilized to a bigger extent. IMO, this places cards like the RTX 3080 in a bad spot, but is it really a deal breaker? With the exception of a noteworthy game I'm about to mention, we've yet to see even the 3080 fall dramatically behind its AMD counterpart despite having 6 GB less.

I agree, more VRAM is better. But realistically speaking, it's not a life or death situation, at least not yet, and I don't think it'll be one for some time to come, especially if you're willing to drop textures from "ultra" to "high" and not run extreme raytracing settings. Perhaps the only game where 10-12 GB has become a true limitation at this point is The Last of Us, which seems to use VRAM so aggressively that it's not going to perform on a GPU that has less than 16 GB of VRAM, as evident here given the RX 6800 XT wins out against the otherwise far superior 4070 Ti.



Frankly, a game that *genuinely* needs so much VRAM probably has more than a few optimization issues, but I digress. We're not "fighting against memory increases", I'm just making the case that VRAM amounts are currently adequate, if not slightly lacking across most segments, as you see, even then the aforementioned scenario involves 4K and very high graphics settings.
Posted on Reply
#59
GhostRyder
Vya DomusIt's really not. I'll remind you that Nvidia originally intended for this thing to be a 900$ "4080" and no "super" models most likely (not that they really change anything). They moved the labels around but that doesn't change what these products were originally intended to be.

Not that it matters, your comment is trying to imply that it's unreasonable to expect such a card to be used for 4K gaming. 700$ is a truckload of money, no matter how spin it that's not mid range, nor is isn't unreasonable to expect it to not be limited by VRAM at 4K ultra.


Strange how AMD despite being the budget option can always include a truckload of VRAM.
I agree, its not unreasonable anymore to assume people will 4k game with the 4070ti class of cards.
Dr. DroI'm well aware, and like I said, I agree, $700 is a lot of money, but it remains the middle option in Nvidia's product stack. We just have to accept that the days for a $699 flagship aren't coming back, as much as we'd want that to happen.

AMD adds VRAM because they have to stand out somehow, it's not a big secret. Also, they've been discontinuing driver support for their previous generation hardware mercilessly as of late. That alone would justify a cheaper asking price, IMHO.
No its intentionally making their products obsolete sooner. Nvidia is very careful when it comes to making cards last for a certain amount of time as that their business strategy in getting people to upgrade. VRAM is an easy way to limit a cards future prospects as we have seen VRAM usage grow exponentially and its a reason Nvidia forced vendors to stop putting out special versions of cards with more VRAM. Only the very top of the product stack has reasonable VRAM amounts that would allow you to use the card effectively longer. Hence why the card had a 192bit bus at launch and lower vram (But now we have a newer version with a 256 bit bus and 16gb near the end of the cycle) because they are trying to gin up a reason to purchase it for the hold outs.
Dr. DroI haven't run into VRAM limitations with any of my GPU purchases in the past 8 years with the sole exception of the R9 Fury X (for obvious reasons), so I couldn't really answer. I normally don't buy lower-end GPUs and that I opted for a 4080 this time is most unusual (and frankly motivated by money, as I wanted to buy an expensive display and overhaul the rest of my PC as well), I always buy cards at the halo tier. The few games I played on a 4 GB GPU recently (my laptop's 3050M) did require compromises in texture quality to be playable, but they were playable nonetheless, at settings that I personally consider acceptable for a low-end laptop GPU.

Anyway, let's be realistic for a moment here. When the 1080 Ti launched 7 years ago in the Pascal refresh cycle, 11 GB was truly massive. And we all know that 11 GB was because Nvidia only removed a memory chip to lower the bandwidth and ensure that the 2016 Titan X retained a similar level of performance (despite having slower memory) and that there was a significant enough gap between it and the then-new Titan Xp, which came with the complete GP102 chip and the fastest memory that they had at the time. For most of its lifetime you could simply enable all settings and never worry about that.

GPU memory capacities began to rise after the Turing and RDNA generation but we were still targeting 8 GB for the performance segment parts (RTX 2080, 5700 XT), and these cards weren't really all that badly choked by these. Then came RDNA 2 and started to give everyone 12 and 16 GB like candy, but that still didn't give them a very distinct advantage over Ampere, that was still lingering on 8-10 GB, and then eventually 12 GB for the high end. Nvidia positioned the RTX 3090 with 24 GB to basically assume both the role of gaming flagship and a replacement for the Titan RTX, while removing a few perks of the latter and technically lowering the MSRP by $1000 (which was tbh an acceptable tradeoff). We know reality is different, but if we exclude the RTX 3090, then only AMD really offered a high VRAM capacity.

This generation started to bring higher VRAM capacities on both sides, and it's only now with games that were designed first and foremost for the PS5 and Xbox Series X starting to become part of gamers' rosters and benchmark suites, all featuring advanced graphics, new texture streaming systems, tons of high-resolution assets, etc. that we've begun to see that being utilized to a bigger extent. IMO, this places cards like the RTX 3080 in a bad spot, but is it really a deal breaker? With the exception of a noteworthy game I'm about to mention, we've yet to see even the 3080 fall dramatically behind its AMD counterpart despite having 6 GB less.

I agree, more VRAM is better. But realistically speaking, it's not a life or death situation, at least not yet, and I don't think it'll be one for some time to come, especially if you're willing to drop textures from "ultra" to "high" and not run extreme raytracing settings. Perhaps the only game where 10-12 GB has become a true limitation at this point is The Last of Us, which seems to use VRAM so aggressively that it's not going to perform on a GPU that has less than 16 GB of VRAM, as evident here given the RX 6800 XT wins out against the otherwise far superior 4070 Ti.



Frankly, a game that *genuinely* needs so much VRAM probably has more than a few optimization issues, but I digress. We're not "fighting against memory increases", I'm just making the case that VRAM amounts are currently adequate, if not slightly lacking across most segments, as you see, even then the aforementioned scenario involves 4K and very high graphics settings.
I agree to a point on the optimization argument as it has become more common place to just dump all textures into the GPU memory for developers to avoid having to do work harder in pull those assets when needed. But that unfortunately is the reality and we now have to contend with that fact when purchasing a GPU.
Posted on Reply
#60
trsttte
The Von MatricesI agree with Dr. Dro in that you should vote with your wallet and (theoretically) companies will respond to that feedback. But also realize that if it is selling well then it could also be that your needs are just different than the needs of most buyers.
Dr. DroGPU memory capacities began to rise after the Turing and RDNA generation but we were still targeting 8 GB for the performance segment parts (RTX 2080, 5700 XT), and these cards weren't really all that badly choked by these. Then came RDNA 2 and started to give everyone 12 and 16 GB like candy, but that still didn't give them a very distinct advantage over Ampere, that was still lingering on 8-10 GB, and then eventually 12 GB for the high end. Nvidia positioned the RTX 3090 with 24 GB to basically assume both the role of gaming flagship and a replacement for the Titan RTX, while removing a few perks of the latter and technically lowering the MSRP by $1000 (which was tbh an acceptable tradeoff). We know reality is different, but if we exclude the RTX 3090, then only AMD really offered a high VRAM capacity.

This generation started to bring higher VRAM capacities on both sides, and it's only now with games that were designed first and foremost for the PS5 and Xbox Series X starting to become part of gamers' rosters and benchmark suites, all featuring advanced graphics, new texture streaming systems, tons of high-resolution assets, etc. that we've begun to see that being utilized to a bigger extent. IMO, this places cards like the RTX 3080 in a bad spot, but is it really a deal breaker? With the exception of a noteworthy game I'm about to mention, we've yet to see even the 3080 fall dramatically behind its AMD counterpart despite having 6 GB less.
Why was 8gb enough for so long, want to take a guess at how much vram the ps4 had? The PS5 and Xbox Series X launched in November 2020 with 16gb, we're in 2024, 12gb is inexcusable and will be an obvious limitation unless you upgrade often - which you shouldn't need to when you're spending more than 600$. It's ridiculous and nvidia is doing it's best to kill pc gaming and make people turn to consoles.

It's cool that you mention the 3090 as well, that was problematic for an entirely different set of reasons: the titan cards used to have professional features enabled, bringing back the x90 class card was supposed to replace the titan and was even advertised often as such but they cuf off access to those.

Vote with your wallet and all that, doesn't matter all that much when mindless people will buy whatever is available but anyway... here's to waiting the 50 series isn't as bad as this one but i'm not holding my breath, given the current AI boom we'll need to wait for AI winter.
Posted on Reply
#61
evernessince
Dr. DroAnyway, let's be realistic for a moment here. When the 1080 Ti launched 7 years ago in the Pascal refresh cycle, 11 GB was truly massive. And we all know that 11 GB was because Nvidia only removed a memory chip to lower the bandwidth and ensure that the 2016 Titan X retained a similar level of performance (despite having slower memory) and that there was a significant enough gap between it and the then-new Titan Xp, which came with the complete GP102 chip and the fastest memory that they had at the time. For most of its lifetime you could simply enable all settings and never worry about that.
Compared to the jump in VRAM of prior gens including from the 780 Ti to 980 Ti, 11GB was slightly below normal. The 780 Ti to 980 Ti was a straight doubling, the 1080 Ti a near doubling. The 2080 Ti was zero increase and the 3080 Ti was a 1GB increase to 12GB. It seems revisionist to call the 1080 Ti's VRAM allotance massive, no it was a pretty normal increase at the time. Only through 7 years of Nvidia cuckery does it appear big.
Dr. DroGPU memory capacities began to rise after the Turing and RDNA generation but we were still targeting 8 GB for the performance segment parts (RTX 2080, 5700 XT), and these cards weren't really all that badly choked by these. Then came RDNA 2 and started to give everyone 12 and 16 GB like candy, but that still didn't give them a very distinct advantage over Ampere, that was still lingering on 8-10 GB, and then eventually 12 GB for the high end. Nvidia positioned the RTX 3090 with 24 GB to basically assume both the role of gaming flagship and a replacement for the Titan RTX, while removing a few perks of the latter and technically lowering the MSRP by $1000 (which was tbh an acceptable tradeoff). We know reality is different, but if we exclude the RTX 3090, then only AMD really offered a high VRAM capacity.
2080 and 5700 XT are not in the same class. x700 is distinctively AMD's midrange where xx80 is Nvidia's high end. That the 2080 only having 8GB of VRAM was disgraceful, the 2000 series was a joke of a generation.

The 3090 (and it's varients) is the only Nvidia card that actually increased VRAM on the Nvidia side. the rest of the stack saw little to no movement, especially if you factored in price (with or without inflation). The 3090 was also super expensive.
Dr. DroI agree, more VRAM is better. But realistically speaking, it's not a life or death situation, at least not yet, and I don't think it'll be one for some time to come, especially if you're willing to drop textures from "ultra" to "high" and not run extreme raytracing settings. Perhaps the only game where 10-12 GB has become a true limitation at this point is The Last of Us, which seems to use VRAM so aggressively that it's not going to perform on a GPU that has less than 16 GB of VRAM, as evident here given the RX 6800 XT wins out against the otherwise far superior 4070 Ti.
You don't seem to understand that my argument has never been "hey look at this chart, it shows more VRAM = better". Yes you can show that but the point I've been making, and have repeated 4 times now is:

"Hardware leads software, not the other way around. It's backwards to imply that because games don't use x amount of VRAM today, x amount of VRAM isn't worthwhile. That kind of logic only perpetuates stagnation of VRAM capacity. Devs aren't going to make games with VRAM requirements the vast majority of cards can't meet. It's up to the hardware vendors to increase GPU VRAM allowance so that game devs can then create games that utilize that extra capacity."

There isn't ever going to be a point where games are releasing en mass with requirements that make most people's experience unenjoyable. You don't seem to understand that one has to come before the other, instead arguing that because the latter hasn't come the former is fine. Again, backwords logic.
Posted on Reply
#62
Dr. Dro
evernessince2080 and 5700 XT are not in the same class. x700 is distinctively AMD's midrange where xx80 is Nvidia's high end. That the 2080 only having 8GB of VRAM was disgraceful, the 2000 series was a joke of a generation.

The 3090 (and it's varients) is the only Nvidia card that actually increased VRAM on the Nvidia side. the rest of the stack saw little to no movement, especially if you factored in price (with or without inflation). The 3090 was also super expensive.

You don't seem to understand that my argument has never been "hey look at this chart, it shows more VRAM = better". Yes you can show that but the point I've been making, and have repeated 4 times now is:

"Hardware leads software, not the other way around. It's backwards to imply that because games don't use x amount of VRAM today, x amount of VRAM isn't worthwhile. That kind of logic only perpetuates stagnation of VRAM capacity. Devs aren't going to make games with VRAM requirements the vast majority of cards can't meet. It's up to the hardware vendors to increase GPU VRAM allowance so that game devs can then create games that utilize that extra capacity."

There isn't ever going to be a point where games are releasing en mass with requirements that make most people's experience unenjoyable. You don't seem to understand that one has to come before the other, instead arguing that because the latter hasn't come the former is fine. Again, backwords logic.
The 5700 XT performs similarly to the 2080 (in the things that it can do, anyway), and the 2080 similarly uses the midrange TU104 processor. So in this specific generation, no. As for the 3090, it was supposed to take over the market segment that they extinguished after the Titan RTX (which also had 24 GB), so in that sense, it was a net zero improvement, other than the $1000 decrease in MSRP. Which didn't mean anything during that season anyhow.

And again, let's be honest. Games are generally developed with average hardware in mind, and primarily target consoles.
trsttteWhy was 8gb enough for so long, want to take a guess at how much vram the ps4 had? The PS5 and Xbox Series X launched in November 2020 with 16gb, we're in 2024, 12gb is inexcusable and will be an obvious limitation unless you upgrade often - which you shouldn't need to when you're spending more than 600$. It's ridiculous and nvidia is doing it's best to kill pc gaming and make people turn to consoles.

It's cool that you mention the 3090 as well, that was problematic for an entirely different set of reasons: the titan cards used to have professional features enabled, bringing back the x90 class card was supposed to replace the titan and was even advertised often as such but they cuf off access to those.

Vote with your wallet and all that, doesn't matter all that much when mindless people will buy whatever is available but anyway... here's to waiting the 50 series isn't as bad as this one but i'm not holding my breath, given the current AI boom we'll need to wait for AI winter.
Yes, I brought all of that up and i'm in agreement. But remember that the consoles also have unified memory pool, those 16 GB count for system and graphics memory simultaneously.
Posted on Reply
#63
Vya Domus
Dr. Droand the 2080 similarly uses the midrange TU104 processor.
TU104 was 545mm^2, pretty huge as far as GPU dies go, definitely not midrange. Navi 10 was a much smaller die but on 7nm, this is not a good comparison at all.
Posted on Reply
#64
trsttte
Dr. Drothose 16 GB count for system and graphics memory simultaneously.
Architectures vary, but they don't, the PS5 has 512mb of DDR4 for system functions and the Xbox also has something similar besides it's weird tiered memory system. But the way it's shared is not comparable with how a desktop computer would share memory, even DDR5 is dogshit slow if the gpu ends up needing to use it.
Posted on Reply
#65
Dr. Dro
Vya DomusTU104 was 545mm^2, pretty huge as far as GPU dies go, definitely not midrange. Navi 10 was a much smaller die but on 7nm, this is not a good comparison at all.
Yes, it was 545mm², but it also had several hardware capabilities (such as dedicated tensor and raytracing cores), and it was built on 12 nm, while Navi 10 completely lacks those areas and was built on 7 nm. So there's that.
trsttteArchitectures vary, but they don't, the PS5 has 512mb of DDR4 for system functions and the Xbox also has something similar besides it's weird tiered memory system. But the way it's shared is not comparable with how a desktop computer would share memory, even DDR5 is dogshit slow if the gpu ends up needing to use it.
That handles the core OS at best, though.
Posted on Reply
#66
evernessince
Dr. DroThe 5700 XT performs similarly to the 2080 and the 2080 similarly uses the midrange TU104 processor. So in this specific generation, no.
As Vya pointed out, the 2080 was a 545mm2 die. It also cost $800.
(in the things that it can do, anyway),
Sorta of like how the 2080 isn't going to be doing RT on any modern game either? Yeah you can technically do RT on either the 5700 XT and 2080 and both are going to give you are terrible experience. The 3070 is already half in the same boat already as well.
Dr. DroYes, it was 545mm², but it also had several hardware capabilities (such as dedicated tensor and raytracing cores), and it was built on 12 nm, while Navi 10 completely lacks those areas and was built on 7 nm. So there's that.
You are trying to make this hypothetical argument that the 2080 WOULD be a midrange card in your hypothetical situation but it doesn't workout because 1) It would be even more expensive in that scenario given the increased cost of 7nm 2) It'd still be at least an $800+ video card 3) Your hypothetical is not reality. The 2080 in reality has the price tag, name, and die size of a higher card. Let's call a pig a pig.
Dr. DroAs for the 3090, it was supposed to take over the market segment that they extinguished after the Titan RTX (which also had 24 GB), so in that sense, it was a net zero improvement, other than the $1000 decrease in MSRP. Which didn't mean anything during that season anyhow.
xx90 cards do FP64 at half the rate of Titan cards (1:64 vs 1:32). Nvidia wanted people doing scientific calculations, one of the major uses of Titan cards, to spend more money so they got rid of Titan cards. The xx90 class is a more gimped version of the Titan cards focused specifically on the gaming aspect. People didn't get a discount, most people who legimitimately needed the features of the Titan cards eneded up having to spend more and buy up.
Dr. DroAnd again, let's be honest. Games are generally developed with average hardware in mind, and primarily target consoles.
Entirely depends if the consoles are considered average at the point of measure in their lifecycle. Console performance tends to go down over time as games become more demanding. Games requirements do not remain fixed and often games released later in a console gen tend to perform more poorly as we reach the end of a cycle.
Dr. DroYes, I brought all of that up and i'm in agreement. But remember that the consoles also have unified memory pool, those 16 GB count for system and graphics memory simultaneously.
Couple of factors you are forgetting:

1) Both consoles have additional reserved memory just for the OS

2) Consoles have less overhead

3) Both consoles have proprietary texture decompression chips that allow them to dynamically stream data off their SSDs at high speeds and low latency, drastically reducing the amount of data that has to be stored in memory / VRAM.
Posted on Reply
#67
Vya Domus
evernessincexx90 cards do FP64 at half the rate of Titan cards (1:64 vs 1:32).
To be fair 1:32 is just as useless as 1:64. My 7900XT can do roughly 1 TFLOP of FP64 (1:32) which is actually about the same as my CPU. Only the very first Titan and Titan V had good FP64 performance.
Posted on Reply
#68
evernessince
Vya DomusTo be fair 1:32 is just as useless as 1:64. My 7900XT can do roughly 1 TFLOP of FP64 (1:32) which is actually about the same as my CPU. Only the very first Titan and Titan V had good FP64 performance.
Agreed, it's really a shame that the Titan cards went from 1:2 to 1:32 in FP64 when the Titan branding was in part built on their scientific and engineering chops.
Posted on Reply
#69
umeng2002
Dr. Drowww.digikey.com/en/products/detail/micron-technology-inc/MT61K512M32KPA-24-U-TR/17632186

This is the exact memory chip used on the RTX 4080. At $55,700, the per unit cost is $27.85 in a 2000 unit bulk.

$27.85 * 8 = $222.80 in memory IC costs alone.

www.digikey.com/en/products/detail/micron-technology-inc/MT61K512M32KPA-21-U/17631914

The same situation with the slower 21 Gbps chip used in the 4070 Ti and 4090, at $25.31 per unit at bulk cost

$25.31 * 6 = $151.86 in memory costs alone

Not only does the cost add up, but factor in PCB costs, power delivery component costs, etc. - losses, the fact that everyone needs profit, yada yada, you get what I'm going at. I'm fairly sure the BoM for a 4070 Ti must be somewhere in the vicinity of $500 as a rough estimate. The rest makes up for profit, distribution cost, driver software development costs, etc.
If consumers knew the volume discount on these things, there would be riots in the streets.

But, for all we know, nVidia™ might not be able to design a memory controller as well as others; and bad yields in the controller area might be making die costs way higher than AMD or Intel.
Posted on Reply
#70
stimpy88
umeng2002If consumers knew the volume discount on these things, there would be riots in the streets.

But, for all we know, nVidia™ might not be able to design a memory controller as well as others; and bad yields in the controller area might be making die costs way higher than AMD or Intel.
I believe that nGreedia feels that they do good compression work, and that negates the need for a larger frame buffer. In nGreedias eyes. But the facts are slightly different. nGreedia only cares about profit, like every other corporation, but nGreedia does this to the next level, as they simply do not care about offering any kind of value to their customers.

VRAM is the second most costly part of a graphics card, and that right there is the reason we don't see 16GB as standard on all but the lowest end cards in 2024. I truly worry about Blackwell, as if nGreedia doesn't up the VRAM for that generation, then we will begin to see a stutter fest on all AAA games next year onwards. But I suspect that we will see only the unobtanium card getting 24 or possibly 32GB frame buffers, while the rest will top out at 16GB, and the midrange and lower will top out at 12GB. If so, that generation will be very bad value, even worse than the bad value of this awful generation.
Posted on Reply
#71
GhostRyder
stimpy88I believe that nGreedia feels that they do good compression work, and that negates the need for a larger frame buffer. In nGreedias eyes. But the facts are slightly different. nGreedia only cares about profit, like every other corporation, but nGreedia does this to the next level, as they simply do not care about offering any kind of value to their customers.

VRAM is the second most costly part of a graphics card, and that right there is the reason we don't see 16GB as standard on all but the lowest end cards in 2024. I truly worry about Blackwell, as if nGreedia doesn't up the VRAM for that generation, then we will begin to see a stutter fest on all AAA games next year onwards. But I suspect that we will see only the unobtanium card getting 24 or possibly 32GB frame buffers, while the rest will top out at 16GB, and the midrange and lower will top out at 12GB. If so, that generation will be very bad value, even worse than the bad value of this awful generation.
AMD and Nvidia are no different in that they are focused on profits as that is what they are there to do. We can't say that either is objectively worse in wanting to make profits off their cards. Nvidia's major sin is the fact they have gotten complacent with whatever they release. Who can blame them, their cards always sell very well regardless of what they do and its caused the market to be very stagnant/uncompetitive. If people want change, they need to start reading the gaming reviews and buying based on that instead of what Nvidia card they can afford. Does it absolve AMD when they make a bad card, no it should not sell well if its not a good card (R9 Fury's for instance) but currently they are pretty competitive in more areas than naught. If people sent a message in those areas by voting with their wallet, then Nvidia will change and improve.
evernessinceAgreed, it's really a shame that the Titan cards went from 1:2 to 1:32 in FP64 when the Titan branding was in part built on their scientific and engineering chops.
Yea, its funny how the Titan series went from a hybrid professional card, to the ultimate gaming card, to just being dropped entirely and the XX90 series branding coming back.
Posted on Reply
#72
neatfeatguy
Was at Micro Center yesterday picking up a few items and a guy checking out in front of me was excited he was able to pickup the last 4080 Super (priced at $1000 for the model he got) they had in stock. Damn....after tax that guy spent close to $1100 on a 4080 Super. Good for him and all, but I don't know how you could be excited about dropping that much money on a GPU.

How he wants to spend his money is his propagative, but if you ask me, a fool and his money are soon parted. The card alone cost almost the same as the rest of the items he was getting (SSD, MB, CPU, case, RAM, PSU). In all he was looking at around $2300 for everything after taxes, based on the prices on the items (it is possible some were priced lower than the sticker cost).

To be fair, I asked a guy in the DIY department how many 4080S they see come in and he said just a couple here and there and within a few days the couple that come do sell. They are selling, but by no means are they coming in by the truckload.
Posted on Reply
#73
rv8000
GhostRyderAMD and Nvidia are no different in that they are focused on profits as that is what they are there to do. We can't say that either is objectively worse in wanting to make profits off their cards. Nvidia's major sin is the fact they have gotten complacent with whatever they release. Who can blame them, their cards always sell very well regardless of what they do and its caused the market to be very stagnant/uncompetitive. If people want change, they need to start reading the gaming reviews and buying based on that instead of what Nvidia card they can afford. Does it absolve AMD when they make a bad card, no it should not sell well if its not a good card (R9 Fury's for instance) but currently they are pretty competitive in more areas than naught. If people sent a message in those areas by voting with their wallet, then Nvidia will change and improve.

Yea, its funny how the Titan series went from a hybrid professional card, to the ultimate gaming card, to just being dropped entirely and the XX90 series branding coming back.
The problem with that is most buyers end up purchasing Nvidia regardless of the fact that for 95% of games (without RT) AMD offers a better value product in the midrange. Voting with your “wallet” requires thinking about a purchase rather than looking at a halo product and automatically jumping to the conclusion everything down the stack is the best you can get. Eventually consumers are going to price themselves out of PC gaming by accepting whatever ridiculous price Nvidia sets.

Proves time and time again, most people make pretty poor choices when it comes to their “needs” when it comes to tech.
Posted on Reply
#74
AnotherReader
Dr. Drowww.digikey.com/en/products/detail/micron-technology-inc/MT61K512M32KPA-24-U-TR/17632186

This is the exact memory chip used on the RTX 4080. At $55,700, the per unit cost is $27.85 in a 2000 unit bulk.

$27.85 * 8 = $222.80 in memory IC costs alone.

www.digikey.com/en/products/detail/micron-technology-inc/MT61K512M32KPA-21-U/17631914

The same situation with the slower 21 Gbps chip used in the 4070 Ti and 4090, at $25.31 per unit at bulk cost

$25.31 * 6 = $151.86 in memory costs alone

Not only does the cost add up, but factor in PCB costs, power delivery component costs, etc. - losses, the fact that everyone needs profit, yada yada, you get what I'm going at. I'm fairly sure the BoM for a 4070 Ti must be somewhere in the vicinity of $500 as a rough estimate. The rest makes up for profit, distribution cost, driver software development costs, etc.
DRAMExchange shows otherwise. This also lines up with other reports.



Another data point is the price of DDR5 on Digikey. Notice it says 16Gbit. Now we all know that this price is much higher than the actual price of DRAM. DDR5 is now starting at $73 for two 16 GB DIMMs. That is less than a quarter the price of the Digikey quote if they are selling it one IC at a time. The logical conclusion is that the reel, i.e. one unit, doesn't correspond to one DRAM IC.


It goes down to $13.5 when ordering in large quantities.

Posted on Reply
#75
GhostRyder
neatfeatguyWas at Micro Center yesterday picking up a few items and a guy checking out in front of me was excited he was able to pickup the last 4080 Super (priced at $1000 for the model he got) they had in stock. Damn....after tax that guy spent close to $1100 on a 4080 Super. Good for him and all, but I don't know how you could be excited about dropping that much money on a GPU.

How he wants to spend his money is his propagative, but if you ask me, a fool and his money are soon parted. The card alone cost almost the same as the rest of the items he was getting (SSD, MB, CPU, case, RAM, PSU). In all he was looking at around $2300 for everything after taxes, based on the prices on the items (it is possible some were priced lower than the sticker cost).

To be fair, I asked a guy in the DIY department how many 4080S they see come in and he said just a couple here and there and within a few days the couple that come do sell. They are selling, but by no means are they coming in by the truckload.
That is why I bet there is this "Shortage", its artificial to make people buy them when they see them in stock. Helps sales with people who see a card sold out constantly being willing to buy it the moment they see it in stock. I was at Micro-Center as well in Dallas this week and they told me it was a very low amount of stock they got.
rv8000The problem with that is most buyers end up purchasing Nvidia regardless of the fact that for 95% of games (without RT) AMD offers a better value product in the midrange. Voting with your “wallet” requires thinking about a purchase rather than looking at a halo product and automatically jumping to the conclusion everything down the stack is the best you can get. Eventually consumers are going to price themselves out of PC gaming by accepting whatever ridiculous price Nvidia sets.

Proves time and time again, most people make pretty poor choices when it comes to their “needs” when it comes to tech.
I don't disagree with that statement, though I would not say 95% of the time as I think its a much closer battle than that depending on which cards you are comparing. That being said I stand by saying the 7900 XTX is a better value at $900 than buying a RTX 4080 S at $1000.
Posted on Reply
Add your own comment
Dec 18th, 2024 07:38 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts