Wednesday, August 22nd 2018
AMD 7nm "Vega" by December, Not a Die-shrink of "Vega 10"
AMD is reportedly prioritizing its first 7 nanometer silicon fabrication allocations to two chips - "Rome," and "Vega 20." Rome, as you'll recall, is the first CPU die based on the company's "Zen 2" architecture, which will build the company's 2nd generation EPYC enterprise processors. "Vega 20," on the other hand, could be the world's first 7 nm GPU.
"Vega 20" is not a mere die-shrink of the "Vega 10" GPU die to the 7 nm node. For starters, it is flanked by four HBM2 memory stacks, confirming it will feature a wider memory interface, and support for up to 32 GB of memory. AMD at its Computex event confirmed that "Vega 20" will build Radeon Instinct and Radeon Pro graphics cards, and that it has no plans to bring it to the client-segment. That distinction will be reserved for "Navi," which could only debut in 2019, if not later.
Source:
VideoCardz
"Vega 20" is not a mere die-shrink of the "Vega 10" GPU die to the 7 nm node. For starters, it is flanked by four HBM2 memory stacks, confirming it will feature a wider memory interface, and support for up to 32 GB of memory. AMD at its Computex event confirmed that "Vega 20" will build Radeon Instinct and Radeon Pro graphics cards, and that it has no plans to bring it to the client-segment. That distinction will be reserved for "Navi," which could only debut in 2019, if not later.
194 Comments on AMD 7nm "Vega" by December, Not a Die-shrink of "Vega 10"
130W more power for around the same performance as GTX1080. Also, considering the custom cards coming to market late, a lot of Vega cards are reference. That cooler is noisy as hell even compared to Geforce's reference blower.
Mining and GPGPU stuff is where Vega is awesome. But where gaming is concerned, Vega is simply outclassed.
Now when GTX1080s and GTX1070Tis are around 50€ cheaper than Vega 64 and Vega56 respectively (at least in Europe) I honestly cannot see the appeal.
Turing is out and seems to have the usual 20-25% generational performance boost with a huge price increase. AMD might still reconsider bringing 7nm Vega to consumer space. We have not heard anything about the frequency potential but 7nm should bring a considerable uplift so there is a real chance :)
Vega 64 is another story. But in all honesty, anyone who does a little bit of research before purchase knew that wit a simple bios flash you get Vega 64 performance...
And normally, the AMD cards shiuld not be more expensive. Mining hype makes Vega 56 more expensive... Vega outclassed? Not really. People just expected it to beat Nvidia wich is not realistic with AMD's budget.
2014: Maxwell - GTX980/GTX970
2016: Pascal - GTX1080
2017: Vega - Vega64/Vega56
1080Ti is a good 25-30% faster than Vega64. 30-35% compared to Vega56 (Edit: TPU reviews' performance summary says 43% faster than Vega56 and 28% faster than Vega64 at 1440p).
1080Ti also uses less power than Vega64, about 50W less.
Mining hype was not and is not what makes Vega more expensive. AMD needs to retain some profit margin for Vega. GTX1080/GTX1070Ti are cheaper to produce.
Budget is not all that relevant here. All the GPUs are manufactured in the same foundry. Just look at the objective measurements and specs of Vega compared to GP102/GP104 and you see why it was expected to perform roughly at the level of 1080Ti.
Edit:
Specs. Homework is to figure out which GPU is which.
Die size: 314 mm² vs 486 mm² (vs 471 mm²)
Transistors: 7.2 b vs 12.5 b (vs 12 b)
TDP/Power: 180 W vs 295 W (vs 250 W)
RAM bandwidth: 320 GB/s vs 484 GB/s (vs 484 GB/s)
FP32 Compute: 8.2 TFLOPS vs 10.2 TFLOPS (vs 10.6 TFLOPS)
Edit2: 7nm would be able to change the die size situation as well as power and hopefully clocks but we do not know how manufacturing prices compare. The last semi-confirmed accounts said 7nm is still more expensive. More expensive is OK when dealing with high price/margin scenarios like Instinct or Radeon Pro but is prohibitive in getting it to consumers.
oc vs oc 980Ti will be better
btw look where gtx 1070 is on 16nm and ddr5, 1.5x over a 14nm hbm2 amd card.... Do you understand the concept of time and node size ? Maxwell was lightyears ahead of amd with maxwell
that forced amd to use hbm on fury x,that's why it had a measly 4gb vram and cost $650, and even then it was good 20% behind 980ti in perf/wat
Now, to emphasize: this sucks for us end-users. It really sucks. I would much rather live in a world where this wasn't the situation. But it's been pretty clear since the launch of Vega that this is the planned way forward, which makes sense in terms of AMD only recently returning to profitability and thus having to prioritize heavily what they spend their R&D money on.
But here's how I see this: AMD has a compute-centric GPU architecture, which still beats Nvidia (at least Pascal) in certain perf/w and perf/$ metrics when it comes to compute. At the very least, they're far more competitive there than they are in perf/W for gaming (which again limits their ability to compete in the high end, where cards are either power or silicon area limited). They've decided to play to their strengths with the current arch, and pitch it as an alternative to the Quadros and Teslas of the world. Which, as it looks right now, they're having reasonable success with, even with the added challenge that the vast majority of enterprise compute software is written for CUDA. Their consistent focus on promoting open-source software and open standards for writing software has obviously helped this. The key here, though, is that Vega - as it stands today - is a decently compelling product for this type of workload.
The question of what they could have done to improve gaming performance, as this is obviously where Vega lags behind Nvidia the most. This is an extremely complicated question. According to AMD around launch time, the majority of the increase in transistor count between Polaris and Vega was spent on increasing clock speeds, which ... well, didn't really do all that much. Around 200 MHz (1400-ish to 1600-ish), or ~14%. It's pretty clear they'd struggle to go further here. Now, I've also seen postings about 4096 SPs being a "hard" limit of the GCN architecture for whatever reason. I can't back that up, but at least it would seem to make sense in light of there being no increase between the Fury X and Vega 64. So, the architecture might need significant reworking to accommodate a wider layout (though I can't find any confirmation that this is actually the case). They're not starved for memory bandwidth (given that the Vega 56 and 64 match or exceed Nvidia's best). So what can they improve without investing a massive amount of money into R&D? We know multi-chip GPUs aren't ready yet, so ... there doesn't seem to be much. They'll improve power draw and possibly clock speeds by moving to new process nodes, but that's about it.
In other words: it seems very likely that AMD needs an architecture update far bigger than anything we've seen since the launch of GCN. This is a wildly expensive and massively time-consuming endeavor. Also note that AMD has about 1/10 the resources of Nvidia (if that), and have until recently been preoccupied with reducing debt and returning to profitability, all while designing a from-scratch X86 architecture. In other words: they haven't had the resources to do this. Yet.
But things are starting to come together. Zen is extremely successful with consumers, and is looking like it will be the same in the enterprise market. Which will give AMD much-needed funds to increase R&D spending. Vega was highly competitive in compute when it launched, and still seems to do quite well, even if their market share is a fraction of Nvidia's. It's still bringing in some money. All the while, this situation has essentially forced AMD to abandon the high-end gaming market. Is this a nice decision? No, as a gamer, right now, I don't like it at all. But for the future of both gaming and competition in the GPU market in general, I think they're doing the right thing. Hold off today, so that they can compete tomorrow. Investing what little R&D money they had into putting some proverbial lipstick on Vega to sell to gamers (which likely still wouldn't have let them compete in the high end) would have been crazy expensive, but not given them much back, and gained them nothing in the long run. Yet another "it can keep up with 2nd-tier Nvidia, but at 50W more at $100 less" card wouldn't have gained AMD much of an increased user base, given Nvidia's mindshare advantage. But if prioritizing high-margin compute markets for Vega for now, and focusing on Zen for really making money allows them to produce a properly improved architecture in a year? That's the right way to go, even if it leaves me using my Fury X for a while longer than I had originally planned to.
Of course, it's entirely possible that the new arch will fall flat on its face. I don't think so, but it's possible. But it's far more certain that yet another limited-resource, short-term GCN refresh would be even worse.
I don't think Navi is GCN based. I think it's a new architecture which is why AMD has been quiet other than about Vega 20. Vega 20 will likely be AMD's last compute oriented card for a long while. Navi is focused on consumers and consoles.
And you're entirely right: RTG could have put out a cheaper-to-produce "big Polaris" with GDDR5X, which would likely have been a very compelling gaming card - but inherently inferior to Vega in the higher-margin enterprise segment (lack of RPM/FP16 support, no HBCC). Not to mention that - even with AMD's lego-like architectures - designing and getting this chip into production (including designing a GDDR5X controller for the first time, which would likely only see use in that one product line) would have been very expensive. Not even remotely as expensive as Vega or a new arch, but enough to make a serious dent - and thus push development of a new arch back even further. Short-term gains for long-term losses, or at least postponing long-term gains? Yeah, not the best strategy.
Some leaks for 7nm Vega have hinted to additional specialized compute units, similarly to tensor cores in Nvidia's Volta/Turing. These are suspected to be aimed at AI (training). That would actually make a lot of sense, especially with the roughly 2x transistor density of 7nm process as well as not having to alter the base architecture and its limits (yet).
If AMD/RTG can live out this lull by selling Polaris 10/20/30/whatever-minor-tweak for ever-lower prices at roughly performance-per-$ parity with Nvidia, while putting as much effort and money as possible into making their upcoming arch as good as possible, that's a far better solution than churning out half-assed attempts at putting lipstick on Polaris by spreading their limited R&D funds thinner. Doubtful. If the reports I've seen of 64 CUs being a hard limit in GCN are to be trusted, that just means a smaller die with higher clocks or less power (or both). If we can trust Nvidia's numbers somewhat (and ignore their faux-AA tensor core stuff), AMD would need a 50%+ performance increase to beat the 2080, let alone the Ti. That's not happening, even with a 14-to-7nm transition.
Also, the process isn't the key issue - the GTX 1050 and 1050Ti are made by GloFo on the same process as AMD, and roughly match the other Pascal cards for perf/W. This is mainly an arch issue, not a process issue. Why? The lower price of the RAM would likely be offset by designing a new chip with a new memory controller and going through the ~1/2-year process of getting it fabbed. Zero performance gain, at best a $100 price drop, and that's if AMD eats the entire design-to-silicon cost. Not to mention that total board power draw would increase, forcing them to lower clocks.
the point is AMD GPUS need MORE ROPS, (TMUS are usually high cont (Fury X) &Vega 64), but also with 2-3x the theoretical Fil Rates doubled @ same base/boost clocks with a ROP,TMU count equivalent say the ...
@ A ..... 64 ROP'S AND 256 TMU'S COUNT!!!!
fury x is 67 GPPS (GPixels per /)
267 GTPS (GTexels Per /s..... (i forgot the texel fillrates but its just an example!)
@ A 1050 mhz gpu CLOCK
now
VEGA20 7nm
NOT 64, but more like 96 rops remember this number!!! 96 ROPS....
but at 64 rops the fillrates are only slightly higher because of the core gpu clock/boost clocks...
64 rops @ 1ghz gpu clock should be at least doubled from the fury x on the Pixels fillrate like 64 X1ghz + 64.0 Gtexels p/s but with a x2 perf. it be 64 x2 x1ghz = 128.0 Gtexels p/s thAT what we need then make it up to 96 rops then amd would have a monster card
\\
MOST CARDS have much higher texels fillrates, but pixel rates need to be almost just as high to run all the pixels especially on 4k and up like 8k resolutions need that raw pixel fillrate power up in the hundreds 2-300 or even higher would make a difference on a cardlike the fury x with the same 267 gtexels /s texture rate but if we could just double the pixel fillrate i guarantee you the gpu would perform 100% better, especialy on 4k and higher resolutions!!!
However bad the situation is for those who buy xx80 nvidia cards,it is still a lot better than for those who stick to AMD in that segment. You can question the pricing of new cards and how useful rtx will be in the early adoption days, but nvidia has a new architecture out and they make cards available for gamers instantly. Vega 7nm will be out this year, but will take a friggin year for gamers to see one. I still think - better shoot rays with nvidia than shoot yourself in the foot by waiting for amd and getting disappointed again.