Considering the 5080 chip: 370mm²
Approximate dies per wafer (ignoring defects) = 181 dies x 0.85 (85% yield rate) = 153 chips
Original N4P = $17,000, $111/chip.
N4C = $15,555 = $101/chip
AMD should use 20Gbps (cheap, and in good supply), Samsung never put 24Gbps chips into mass production... AMD only sells chips, so its profit margin must be calculated per unit. While the percentage margin may be high, the absolute profit per chip is relatively small. For example, a
100% margin on a $100 chip still results in just $100 of profit. However, AMD has hundreds of millions to billions in R&D, manufacturing, and operational costs that must be recouped through these sales. Then, $600 is possible, but AMD needs to convince everyone in the chain to tighten margins for this to work.
On Nvidia’s side, they weren’t satisfied with just selling chips. Now, they’ve secured massive quantities of GDDR7 to bundle with their GPUs, ensuring AIBs buy the complete set—further increasing their profits.
I am figuring ~70% yield. It is likely higher and therefore chips cheaper. I am being conservative and showing how these products are figured out (roughly).
Yes, 20gbps cheap and why using it. 24gbps is a BIG question mark. It was made, 27gbps was not (bc it couldn't yield). Don't trust the 'sample' listed by Samsung; they do that to protect their customers.
Just like GDDR7 pricing, which isn't as high as you may think. They want you to think it is, but it isn't, esp for nvidia. They ofc use cheap Micron as leverage against Samsung/Hynix which make better chips.
It doesn't even matter (GDDR7) anyway bc it doesn't effect the performance tiers. It's all a gimmick. 5080 literally can't take advantage of more than ~18GB/~26.3gbps on avg OCed...but has 16GB.
They totally didn't clock it (and limit it's potential) to around where 16GB 24gbps GDDR6 could OC. Totally not gonna refresh it above the actual limitation of 9070xtx with 24GB and actually use 18-20GB.
Just like GDDR6x (made cheaply by Micron using PAM4 versus more-expensive normal GDDR6 from Sam/Hynix). It's super weird how GDDR6x clocks less than actual 20gbps GDDR6. Oh wait. Micron.
nVIDIA's margins are insane. Freaking. Not. Okay. DO NOT BUY BC BAD VALUE AND THEY STILL CUT RAM/ABSOLUTE PERF BELOW EACH TIER. That is what I keep trying to explain.
It's genius marketing on 5D-chess levels, but awful for consumers whom fall for these gimmicks.
Many don't understand the limitations/planned obsolesence (until later). It's not shown correctly in reviews.
They sell the hype (32GB with 5090, GDDR7 as whole with 5080, high clocks with 5070), but in doing that miss the mark of actually matching bw/buffer/compute perf (which AMD actually does).
I'm truly hoping one of these days people actually test all these cards (minimum frameate to show limitations, absolute perf from OCing) in ways to show my point.
AMD does so many things that are smart to keep prices down for consumers (while still allowing the same absolute performance as a tier or more up from nVIDIA) and people just don't get it.
This includes bin at threshold clocks/high voltage, but leave that voltage/power limit for the capability of clocks on the table so you can OC it while matching absolute perf for buffer/bw/compute/process.
nVIDIA does the literal exact opposite which is make units that are a very good match at stock clocks, and efficient, but maxed out for RAM capabilties so overclocking doesn't make sense. Limted by buffer.
Typically...5070 will probably clock pretty high to get close to maxing out the capability of 12GB. It is yet another gimmick (high clocks) so they could (probably) cut the unit/engine count down to 4/6144sp.
They sell these cards that are on the wrong side of the general performance expectations (look at RAM/minimum framerates) on snake oil software...sometimes marketed as hardware.
Hey, cool. They can make 4/8-bit ops that take less space on the die than 16/32. They can make them do simple things...cool! They're also not FP32, and can't do FP32.
IOW, they are selling you less for more. They have trained people that because normally a FP32 unit would have to do that calc, but they can do it with lesser, it's a feature.
It's actually a way to save money (for them) AND sell it for more.
I know so many people just don't understand, and that's fine...it's just sad.
I'm not arguing the uses they've created and/or their software ecosystem they've built. While a lot of it is built on innovation from elsewhere, they did take the time/money to get it off the ground.
I am, however, explaining to you how their business (and products) is built on marketing/selling perpetual units.
Not actually smart engineering wrt to what's best for the customer.
On top of that their margins are literally out of control. It is apparently AMD's job to show people this.