Sunday, April 20th 2014

NVIDIA GM204 and GM206 to Tape-Out in April, Products to Launch in Q4?

It looks like things are going horribly wrong at TSMC, NVIDIA and AMD's principal foundry partner, with its 20 nm manufacturing process, which is throwing a wrench into the works at NVIDIA, forcing it to re-engineer an entire lineup of "Maxwell" GPUs based on existing 28 nm process. Either that, or NVIDIA is confident of delivering an efficiency leap using Maxwell on existing/mature 28 nm process, and saving costs in the process. NVIDIA is probably drawing comfort from the excellent energy-efficiency demonstrated by its Maxwell-based GeForce GTX 750 series. According to a 3DCenter.org report, NVIDIA's next mainline GPUs, the GM204 and GM206, which will be built on the 28 nm process, and "Maxwell" architecture, will tape out later this month. Products based on the two, however, can't be expected before Q4 2014, as late as December, or even as late as January 2015.

GM204 succeeds GK104 as the company's next workhorse performance-segment silicon, which could power graphics card SKUs ranging all the way from US $250 to $500. An older report suggests that it could feature as many as 3,200 CUDA cores. The GM204 could be taped out in April 2014, and the first GeForce products based on it could launch no sooner than December 2014. The GM206 is the company's next mid-range silicon, which succeeds GK106. It will tape out in April, alongside the GM204, but products based on it will launch only in January 2015. The GM200 is a different beast altogether. There's no mention of which process the chip will be based on, but it will succeed the GK110, and should offer performance increments worthy of being a successor. For that, it has to be based on the 20 nm process. It will tape-out in June 2014, and products based on it will launch only in or after Q2 2015.
Source: 3DCenter.org
Add your own comment

51 Comments on NVIDIA GM204 and GM206 to Tape-Out in April, Products to Launch in Q4?

#1
matar
28nm I am not buying.
Posted on Reply
#2
Razorfang
Or it could be a conspiracy to extend the product line on both sides.
Posted on Reply
#3
LAN_deRf_HA
Would this then explain the specs for the "880" not being as impressive as expected?
Posted on Reply
#4
JTristam
If this is true then it sucks. Q4 2014/Q1 2015 is way too long. I was expecting Maxwell to be released at least this summer.
Posted on Reply
#5
mroofie
waaaat ??????

what's going to be released this year ? This article is not making sense -_-

and TSMc can go *** themselves
NVIDIA should look for a new partner because this is ridiculous :D
Posted on Reply
#6
mroofie
RazorfangOr it could be a conspiracy to extend the product line on both sides.
the gtx750 750 ti has been a success so im not sure where your getting this idea about aconspiracy
Posted on Reply
#7
seronx
My numbers;
  • 28 nm GM104 silicon
  • ~7 billion transistors
  • 3,072 CUDA cores
  • 192 TMUs
  • 48 ROPs
  • 6.1 single-precision TFLOP/s - 2.8 double-precision TFLOP/s
  • 384-bit wide GDDR5 memory interface
  • 6 GB standard memory amount
  • 384 GB/s memory bandwidth
  • Clock speeds of 900 MHz core, 1000 MHz GPU Boost, 8 GHz memory
  • 250W board power
  • 28 nm GM106 silicon
  • ~5 billion transistors
  • 1,792 CUDA cores
  • 112 TMUs
  • 32 ROPs
  • 3.9 single-precision TFLOP/s - 0.9 double-precision TFLOP/s
  • 256-bit wide GDDR5 memory interface
  • 4 GB standard memory amount
  • 224 GB/s memory bandwidth
  • Clock speeds of 1000 MHz core, 1100 MHz GPU Boost, 7 GHz memory
  • 150W board power
Posted on Reply
#8
mroofie
seronxMy numbers;
  • 28 nm GM104 silicon
  • ~7 billion transistors
  • 3,072 CUDA cores
  • 192 TMUs
  • 48 ROPs
  • 6.1 single-precision TFLOP/s - 2.8 double-precision TFLOP/s
  • 384-bit wide GDDR5 memory interface
  • 6 GB standard memory amount
  • 384 GB/s memory bandwidth
  • Clock speeds of 900 MHz core, 1000 MHz GPU Boost, 8 GHz memory
  • 250W board power
  • 28 nm GM106 silicon
  • ~5 billion transistors
  • 1,792 CUDA cores
  • 112 TMUs
  • 32 ROPs
  • 3.9 single-precision TFLOP/s - 0.9 double-precision TFLOP/s
  • 256-bit wide GDDR5 memory interface
  • 4 GB standard memory amount
  • 224 GB/s memory bandwidth
  • Clock speeds of 1000 MHz core, 1100 MHz GPU Boost, 7 GHz memory
  • 150W board power
150 w and 250 w ?
please go look agian at the 750 ti
and comment back with the correct results :)
Posted on Reply
#9
seronx
mroofie150 w and 250 w ?
please go look agian at the 750 ti
and comment back with the correct results :)
GK106 -> 140 Watts
GM107 -> 60 Watts

GK104 -> 230 Watts
GM106 -> 150 Watts

GK110 = GM104
Posted on Reply
#10
Relayer
Better tape out soon if it's going to be this month.
Posted on Reply
#11
HumanSmoke
seronxMy numbers;
  • 28 nm GM104 silicon
  • ~7 billion transistors
  • 3,072 CUDA cores
  • 192 TMUs
  • 48 ROPs
  • 6.1 single-precision TFLOP/s - 2.8 double-precision TFLOP/s
  • 384-bit wide GDDR5 memory interface
  • 6 GB standard memory amount
  • 384 GB/s memory bandwidth
  • Clock speeds of 900 MHz core, 1000 MHz GPU Boost, 8 GHz memory
  • 250W board power
Not sure how you arrived at that calculation. Highly unlikely that Nvidia would offer a 1:2 rate for FP64 on the GM204 any more than it did with GK104 and GF114/104 before it. Double precision is 1. Unneeded for the gaming segment, 2. Adds to the power budget, and 3. Adds die space.
If the GM204 is an analogue of the previous 104 boards then FP64 will be culled. It was 1:12 in the GF104/104, and 1:24 in GK104. Keeping the FP64 ability at a nominal level would also protect Nvidia's margins on existing Titan/K6000/K20/K40 product lines- and more appropriately, keep them relevant since there's no way Nvidia make a GK 110 replacement on 28nm - which means holding out for the 16nm FinFET node (20nm BEOL+16nm FEOL) for a successor.
Posted on Reply
#12
seronx
HumanSmokeNot sure how you arrived at that calculation. Highly unlikely that Nvidia would offer a 1:2 rate for FP64 on the GM204 any more than it did with GK104 and GF114/104 before it. Double precision is 1. Unneeded for the gaming segment, 2. Adds to the power budget, and 3. Adds die space.
If the GM204 is an analogue of the previous 104 boards then FP64 will be culled. It was 1:12 in the GF104/104, and 1:24 in GK104.
GM107 => 1/8th
GM106 => 1/4th
GM104 => 1/2th
GM200 => Full DP.

The future is compute shading which will be reliant on 64-bit maths.
Posted on Reply
#13
hardcore_gamer
matar28nm I am not buying.
Does the process node matter if the card delivers good performance and power efficiency ?
Posted on Reply
#14
MxPhenom 216
ASIC Engineer
seronxGM107 => 1/8th
GM106 => 1/4th
GM104 => 1/2th
GM200 => Full DP.

The future is compute shading which will be reliant on 64-bit maths.
You are pulling so much of this out of your ass. Unless you have some insider info.
Posted on Reply
#15
mroofie
RelayerBetter tape out soon if it's going to be this month.
9 days left lol then we have to wait until next year for mid-range :(
Posted on Reply
#16
HumanSmoke
seronxGM107 => 1/8th
GM106 => 1/4th
GM104 => 1/2th
GM200 => Full DP.
The future is compute shading which will be reliant on 64-bit maths.
Really? I always thought that compute shading tended to only use FP64 for professional simulations and the like. Gaming compute - ambient occlusion, global illumination, motion blur, particle/water/smoke/fog effects, and depth of field etc. were almost entirely single precision based. If they were double precision based then wouldn't it stand to reason (as an example) that a R9 290X's (704 GFlops FP64) ability at applying compute shader image quality options would make it markedly inferior to the HD 7970 (1075 GFlops) ?
mroofie9 days left lol then we have to wait until next year for mid-range :(
FWIW, the original forum post this article is based on is dated 15th April.
Posted on Reply
#17
LAN_deRf_HA
This happened not long ago where we were stuck on a node for awhile and it sucked for us, but with the efficiency of Maxwell might make up for it. The thing that really sucks is this Q4 nonsense.
Posted on Reply
#18
mroofie
HumanSmokeReally? I always thought that compute shading tended to only use FP64 for professional simulations and the like. Gaming compute - ambient occlusion, global illumination, motion blur, particle/water/smoke/fog effects, and depth of field etc. were almost entirely single precision based. If they were double precision based then wouldn't it stand to reason (as an example) that a R9 290X's (704 GFlops FP64) ability at applying compute shader image quality options would make it markedly inferior to the HD 7970 (1075 GFlops) ?

FWIW, the original forum post this article is based on is dated 15th April.
April 15th? for wat the tape out or release ? xD
Posted on Reply
#19
HumanSmoke
mroofieApril 15th? for wat the tape out or release ? xD
15th April is the date of the original post(16th April local time- my time zone is 10 hours ahead of Germany) stating tape out this month.
So, if the tape out hadn't happened at that stage, it left 15 days in the month for it to happen at that stage....assuming tape out hadn't already occurred- then you're in the realms of trying to disprove a negative.
Posted on Reply
#20
pjl321
I know this could probably never happen but wouldn't it be amazing if either nVidia or AMD (even less likely) signed up to used Intel's foundries giving us Maxwell or Pirate Islands at a pretty much ready 14nm!

It actually makes a lot of sense for all parties, Intel needs more of a reason than its own chips to really push forward with 14nm and for nVidia and AMD its a highly advanced and relatively mature/tested process.

Win, freakin win baby!
Posted on Reply
#22
xenocide
pjl321I know this could probably never happen but wouldn't it be amazing if either nVidia or AMD (even less likely) signed up to used Intel's foundries giving us Maxwell or Pirate Islands at a pretty much ready 14nm!

It actually makes a lot of sense for all parties, Intel needs more of a reason than its own chips to really push forward with 14nm and for nVidia and AMD its a highly advanced and relatively mature/tested process.

Win, freakin win baby!
That will never happen. Hell, Intel barely can get their 14nm process running correctly, and they have the best Engineers in the industry. Not to mention Intel has literally nothing to gain from openning their top of the line fabs to competitors. Business aside, you can't just take a microprocessor design and slap it on a process node that's 30% smaller, it doesn't work like that. They would have to spend a few monthes redesigning and testing it to ensure it's functioning correctly, efficient, and cost effective.
Posted on Reply
#23
librin.so.1
pjl321I know this could probably never happen but wouldn't it be amazing if either nVidia or AMD (even less likely) signed up to used Intel's foundries giving us Maxwell or Pirate Islands at a pretty much ready 14nm!

It actually makes a lot of sense for all parties, Intel needs more of a reason than its own chips to really push forward with 14nm and for nVidia and AMD its a highly advanced and relatively mature/tested process.

Win, freakin win baby!
>implying CPUs and GPUs use the same kind of process

The transistors / chips for CPUs and for GPUs are done in a different way to cater to the ways each of these kinds of ICs work.

As a very good example to illustrate this, IF You know / remember, this was a major hindrance for AMD when they made their latest APUs to keep the GPU part good – using a process meant for CPUs would have non-trivially harmed the performance of the GPU side and vice versa. So they had to compromise. Which is also the reason why the CPU part on their latest APUs don't OC as good any more, compared to their previous APUs.
So yeah, using Intel's fabs for those GPUs could mean actually worse performance and power efficiency despite being 14nm.
Posted on Reply
#24
refillable
TheBrainyOneI will literally drop a Hydrogen bomb on TSMC's Foundries if even one their spokesperson says, "Moore's Law is still being followed today."
Lol that was ridiculously funny.
Posted on Reply
#25
HumanSmoke
TheBrainyOneI will literally drop a Hydrogen bomb on TSMC's Foundries if even one their spokesperson says, "Moore's Law is still being followed today."
Why would TSMC say that now, considering they know full well that processes they timelined for are falling behind schedule due to litho tools and energy demands slipping?
Back when people were a little more confident of EUV's ramp - a year or more ago, people might have seen a business as usual scenario, but ASML's delays in wafer and validation tooling (which caused an influx of funding from their customers), as well as TSMC's own well publicised false start recentlyhave certainly stopped any talk of the continuation of transistor density per dollar.
Posted on Reply
Add your own comment
Nov 18th, 2024 22:23 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts