Tuesday, June 17th 2008

ATI Believes GeForce GTX 200 Will be NVIDIA's Last Monolithic GPU.

The head of ATI Technologies claims that the recently introduced NVIDIA GeForce GTX 200 GPU will be the last monolithic "megachip" because they are simply too expensive to manufacture. The statement was made after NVIDIA executives vowed to keep producing large single chip GPUs. The size of the G200 GPU is about 600mm2¬¬ which means only about 97 can fit on a 300mm wafer costing thousands of dollars. Earlier this year NVIDIA's chief scientist said that AMD is unable to develop a large monolithic graphics processor due to lack of resources. However, Mr. Bergman said that smaller chips allow easier adoption of them for mobile computers.
Source: X-bit Labs
Add your own comment

116 Comments on ATI Believes GeForce GTX 200 Will be NVIDIA's Last Monolithic GPU.

#51
btarunr
Editor & Senior Moderator
DarkMatterI would like to know in which facts are you guys basing your claims that a die shrink won't do anything to help lowering heat output and power consumption?
A 65nm single GPU requires 6 + 8 pin power input (obviously for higher power input at peak). How much of that input can be reduced with a die shrink to 55nm? Enough to make a GX2? Without say three 6-pin connectors?
Posted on Reply
#52
Megasty
DarkMatterI would like to know in which facts are you guys basing your claims that a die shrink won't do anything to help lowering heat output and power consumption? It has always helped A LOT. It is helping Ati and surely will help Nvidia. Thinking that the lower power consumption of RV670 and RV770 is based on architecture ehancements alone is naive. I'm talking about peak power, in comparison to what R600 was, idle power WAS improved indeed, and so has GT200.
Of course it'll lower the heat & power. The only point was that it still wouldn't allow for a GX2, not to mention that the card would cost around $1200 :wtf: :shadedshu

However, a faster & cheaper 400mm² die does have EVERY advantage over a slower, more costly 576mm² die.
Posted on Reply
#54
tkpenalty
DarkMatterI would like to know in which facts are you guys basing your claims that a die shrink won't do anything to help lowering heat output and power consumption? It has always helped A LOT. It is helping Ati and surely will help Nvidia. Thinking that the lower power consumption of RV670 and RV770 is based on architecture ehancements alone is naive. I'm talking about peak power, in comparison to what R600 was compared to competition, idle power WAS improved indeed, and so has GT200.
90nm > 65nm (R600 to RV670), was a HUGE leap in the drop of the transistor size, moreover remember the RV670 has a massive chunk of it; half of the 512bit memory controller effectively removed.

Die shrinks do something but under around 65nm the usefulness of die shrinking insn't really significant. Nvidia's CEO admitted that dieshrinking the GTX280 wouldnt help its extreme heat output a lot. Its fairly reasonable as to why, transistor count is more of a factor. In all cases, G80 > G92, R600 > RV670, its due to the cutting down of the memory controller.

By the way, the reason why AMD's cards use more power is simple; their cards use more phases in contrast to Nvidia. More phases = more power used but phases subject to less current, as well as generating less heat.
Posted on Reply
#55
newconroer
tkpenaltyI agree with your view.

GT200s, well Nvidia are shearing down their profits just to get these things to sell, AMD on the otherhand enjoy not having to reinforce their cards and put high end air cooling on-they are way better off. If these 4850s sell well, as well as the RV770, the GT200s look like an awful flop.
Ya because ATi's been looooooving the way things have turned out the last two and a half years.

Yep, they don't have put 'high end air cooling' on their products, what a wonderful relief for them!

~
Posted on Reply
#56
Megasty
newconroerYa because ATi's been looooooving the way things have turned out the last two and a half years.

Yep, they don't have put 'high end air cooling' on their products, what a wonderful relief for them!

~
Not to be negative or anything but none of the stock cards from NV or ATi have high-end cooling fans. The stock casing only restrict most of the fans anyway :(
Posted on Reply
#57
Kreij
Senior Monkey Moderator
Polaris573The head of ATI Technologies claims that the recently introduced NVIDIA GeForce GTX 200 GPU will be the last monolithic “megachip” because they are simply too expensive to manufacture. The statement was made after NVIDIA executives vowed to keep producing large single chip GPUs. The size of the G200 GPU is about 600mm2¬¬ which means only about 97 can fit on a 300mm wafer costing thousands of dollars.
Why are the wafers limited to 300mm? Can't they use a 600mm wafer and get four times the processors out of it?
Is it just because all the FABs are set up to use that size or is there some kind of physical limit?
Posted on Reply
#58
[I.R.A]_FBi
MegastyNot to be negative or anything but none of the stock cards from NV or ATi have high-end cooling fans. The stock casing only restrict most of the fans anyway :(
orly?
im sure i read somewhere about teh gtx 280 cooler being designed by CM or sumpn.
Posted on Reply
#59
DarkMatter
IMO a die shrink to 55nm could enable the posibility of doing a GX2. Maybe not a GTX280 GX2 but yes one with slightly lower clocks or one cluster dissabled and with enough performance to crunch the X2, of course it would require more power, but it would still be within the 6+8 pin envelop. I have three "facts", but of course are only based on my opinion:

1- You have to take into account how power consumtion works. It's exponential, not linear, so a slower part would consume a lot less and the same can be applied to voltages. Nvidia because GT200 was worse than expected in this area had to lower the clocks, but probably they have kept it as high as possible within the selected power envelope. There's always a hot spot for performance-per-watt for any chip and GTX 280 is probably quite higher than that spot. FACT: look at Wizzard's Zotac AMP! GTX280 it consumes a lot more than what you should expect from that overclock. Aim a bit lower than that said spot and you have a "low power" chip. For example a GTX280 GX2 @ 500 Mhz would consume a lot less and still leave the HD4870 X2 behind in performance.

2- Nvidia has implemented the abbility to shut down parts of the chip in the GT200, and it really works very well. Again look at Wizzard's power consumption charts and how it consumes a lot less than the X2 on average, even though its maximum is almost the same. That would make the card to probably never reach the maximum power consumtion in the GX2 card. There's no way you are going to be able to make a total of 64 ROPs work at the same time, for example.

3- Continuing with the above argument, IMO if Nvidia did a GX2 it wouldn't be based on the GTX 280, but on the 8800 GS substitute. Nvidia will surely make a 16-20-24 ROP card while mantaining a high shader count (maybe 192/168 SP, same or one less cluster than GTX260 for example), they would be stupid if they didn't, as it makes more sense than ever. The GS is "weak" because it has 12 ROPs but 16 on the other hand are enough for high-def gaming. 16 ROP x 2 is more than enough as the X2/GX2 can testify, 32 x 2 is just over-over-overkill and silly.
My bet is that Nvidia will do a 20 ROP 168/192 SP card for high mainstream no matter what and they could use that for the GX2. Final specs for that hypothetical GX2 would be: 40 ROP, 336/384 SP, 112/128 TMU and 2 x 320 bit memory controler, that if they can't make the card use the same memory pool as R700 seems to be going to do. The above card would leave the X2 well behind performance wise and still be within the power envelop IMO. Of course that envelop would be higher than that on the X2 but reachable IMHO and still within the 6+8 pin layout that's 300W, the GTX 280 needs 6+8 pins just by a hair.
Posted on Reply
#60
wolf
Better Than Native
interesting
Posted on Reply
#61
DarkMatter
KreijWhy are the wafers limited to 300mm? Can't they use a 600mm wafer and get four times the processors out of it?
Is it just because all the FABs are set up to use that size or is there some kind of physical limit?
Exactly. Wafer size is to Fabs and manufacturers like CPU sockets are for motherboard and CPU makers (interms of compatibility) or ATX standard if you prefer. 450 mm wafers are on track already BTW, 600mm is too much to handle right now.

EDIT: And yes, there is some physical limit too. Take in mind wafers are done by slicing silicon bars at very thin width (less than 1 mm IIRC) and have to mantain the same width all over their area. To that you have to add that the alloy of silicon has to be homogeneous throughout the whole wafer too.
Posted on Reply
#62
btarunr
Editor & Senior Moderator
DarkMatterExactly. Wafer size is to Fabs and manufacturers like CPU sockets are for motherboard and CPU makers (interms of compatibility) or ATX standard if you prefer. 450 mm wafers are on track already BTW, 600mm is too much to handle right now.
That's because of yields. Big wafer, wafer fails, loss of more yield. Keeping wafer sizes limited is a precautionary measure (while compromising on manufacturing expenditure).
Posted on Reply
#63
Nyte
tkpenalty90nm > 65nm (R600 to RV670), was a HUGE leap in the drop of the transistor size, moreover remember the RV670 has a massive chunk of it; half of the 512bit memory controller effectively removed.

Die shrinks do something but under around 65nm the usefulness of die shrinking insn't really significant. Nvidia's CEO admitted that dieshrinking the GTX280 wouldnt help its extreme heat output a lot. Its fairly reasonable as to why, transistor count is more of a factor. In all cases, G80 > G92, R600 > RV670, its due to the cutting down of the memory controller.

By the way, the reason why AMD's cards use more power is simple; their cards use more phases in contrast to Nvidia. More phases = more power used but phases subject to less current, as well as generating less heat.
670/620/635 = 55 nm
630/610 = 65 nm
Posted on Reply
#64
Unregistered
Whats the average yield for a 300mm wafer then? Does it differ with differant manufacturers or is it totally dependant on the size of the wafer?
#65
HTC
tigger69Whats the average yield for a 300mm wafer then? Does it differ with differant manufacturers or is it totally dependant on the size of the wafer?
It depends on the die size: the bigger the die size, the less units a wafer yields.

That's why ATI is ahead of nVidia (in this respect, atm): they manage to make their die size much smaller then nVidia.
Posted on Reply
#66
btarunr
Editor & Senior Moderator
tigger69Whats the average yield for a 300mm wafer then? Does it differ with differant manufacturers or is it totally dependant on the size of the wafer?
Of course, articles from The Inquirer are so full of it, but in one such article, it was mentioned that on 300mm wafer, for the GT200 yields could be as low as 40%. Somewhere else is said that the die costs $110 to manufacture and assemble into the package (package as in electronics, not logistics) will send the cost upto $120. With increase in wafer sizes, you're increasing the risk of yield loss.
Posted on Reply
#67
DarkMatter
btarunrThat's because of yields. Big wafer, wafer fails, loss of more yield. Keeping wafer sizes limited is a precautionary measure (while compromising on manufacturing expenditure).
Yeah that's what I wanted to say when I said physical limit, as there is no absolute physical limit for that.
Also because bigger wafers are possible I said it works like standards. Nvidia will surely want bigger wafers for GT200 in expense of waffer yields, because probably the loss in those yields would be smaller than the gains in die yields, but since it's like an standard they can't. I don't know if I have explained that well.

EDIT: Also I highly doubt those Inquirer yield numbers. Probably are on the high 40s and were told so, and they just slapped that 40% number. Also that number seems extremely low without knowing how high are other GPU yields. Probably are never higher than 75%, and much lower in new high-end chips. For example RV770. Difference from say 60% and 50% is already very high.
Posted on Reply
#68
spud107
so when nvidia see this they say o rly?
next gfx card will be 2pcb's, one for the gpu, other for the rest of the components:D
Posted on Reply
#69
DanishDevil
^ THAT was a good laugh. It wouldn't surprise me...
Posted on Reply
#70
tkpenalty
spud107so when nvidia see this they say o rly?
next gfx card will be 2pcb's, one for the gpu, other for the rest of the components:D
Explain the logic behind that. You're only mfgr prices up by doing that, ever heard of multi layered PCBs man?...

HD4850 > 9800GTX by 25% According to AMD, this is fairly believeable.

A dual GTX280 is technically impossible, between two slots, why? 65nm to 55nm doesn't boast much of a change in TDP! Nvidia's CEO even admitted it, do I have to repeat this? GX2 would be viable, with say a GT200 variant that is similar to the G92 in die size. It was mentioned that a die shrink would only drop the GTX280's heat ouput down to what, around 200W, which is still ridiculously high (400W+ GX2). Who gives about Idle when the card is ridiculously hot at load.

Nvidia really stabbed themselves in the foot, while it is powerful as such, the HD4870X2 will be a more successful product.
Posted on Reply
#71
spud107
the logic is "taking the piss", wheres your sense of humour?
Posted on Reply
#72
tkpenalty
spud107the logic is "taking the piss", wheres your sense of humour?
Oh, um it broke when my cousin dropped my guitar.

This is serious.
Posted on Reply
#73
Megasty
tkpenaltyOh, um it broke when my cousin dropped my guitar.

This is serious.
The only thing that's serious about it is how NV bet the farm on this thing. I'll be collecting that farm when I buy my 4870x2 :p
Posted on Reply
#74
EastCoasthandle


gtx 200 series gpu (from what I've found so far)

A wafer from the 4800 series gpu will offer a whole lot more. However, I haven't found one yet. Anyone have a 55nm wafer pic?
Posted on Reply
#75
DarkMatter
tkpenaltyExplain the logic behind that. You're only mfgr prices up by doing that, ever heard of multi layered PCBs man?...

HD4850 > 9800GTX by 25% According to AMD, this is fairly believeable.

A dual GTX280 is technically impossible, between two slots, why? 65nm to 55nm doesn't boast much of a change in TDP! Nvidia's CEO even admitted it, do I have to repeat this? GX2 would be viable, with say a GT200 variant that is similar to the G92 in die size. It was mentioned that a die shrink would only drop the GTX280's heat ouput down to what, around 200W, which is still ridiculously high (400W+ GX2). Who gives about Idle when the card is ridiculously hot at load.

Nvidia really stabbed themselves in the foot, while it is powerful as such, the HD4870X2 will be a more successful product.
Could you post a link to where Nvidia's CEO said that please? And where those power numbers were mentioned, though I suppose it's the same. I highly doubt going to 55nm won't make the card consume less than 200W.
Also as I mentioned, Nvidia doesn't need 2 280s to crush Ati's X2, not even 2 260s. By only shrinking the chip to 55nm it would be 400mm2, take some ROPs out and you will get a die size close to G92. No one has said GT200 GX2 is possible but GT200b IS, and you will see it soon if Ati's X2 happens to be quite faster than GTX280.

Also real power consumption of GTX280 is nowhere near those 236W, while the older cards are close to their claimed TDP. It's temperatures are far better than on G92 and RV670 too, despite being a lot bigger, so there's some room left there. If GT200b can't improve the performance beyond that of the X2 a GX2 of GT200b WILL come, but it's nature is not so defined. In fact a card with 2x the performance of GTX280 doesn't make sense AT ALL. If it did, because games in the near future could take advantage of it, then Ati would be LOST.

In the end it will all depend on the real performance of the RV770. AFAIK HD4870 > 9800 GTX by 25% and HD4850 > 8800 GT by 25%. That also means HD4850 > 9800 GTX but by 5-10 %. ANYWAY forget about that if the performance boost of newer drivers happens to be true.
Posted on Reply
Add your own comment
Nov 28th, 2024 15:48 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts