Friday, September 1st 2017
On AMD's Raja Koduri RX Vega Tweetstorm
In what is usually described as a tweetstorm, AMD's RTG leader Raja Koduri weighed in on AMD's RX Vega reception and perception from both the public and reviewers. There are some interesting tidbits there; namely, AMD's option of setting the RX vega parts at frequencies and voltages outside the optimal curve for power/performance ratios, in a bid to increase attractiveness towards the performance/$ crowds.
However, it can be said that if AMD had done otherwise, neither gamers nor reviewers would have been impressed with cards that potentially delivered less performance than their NVIDIA counterparts, while consuming more power all the same (even if consuming significantly less wattage). At the rated MSRP (and that's a whole new discussion), this RTG decision was the best one towards increasing attractiveness of RX Vega offerings. However, Raja Koduri does stress Vega's dynamic performance/watt ratios, due to the usage of specially defined power profiles.To our forum-walkers: this piece is marked as an editorial
Raja also touched an interesting subject; namely, that "Folks doing perf/mm2 comparisons need to account for Vega10 features competing with 3 different competitive SOCs (GP100, GP102 and GP104)". This is interesting, and one of the points we have stressed here on TPU: AMD's Vega architecture makes great strides towards being the architecture AMD needs for its server/AI/computing needs, but it's a far cry from what AMD gamers deserve. it's always a thin line to be threaded by chip designers on which features and workloads to implement/improve upon with new architectures. AMD, due to its lack of a high-performance graphics chip, was in dire need of a solution that addressed both the high-performance gaming market and the server/computing market where the highest profits are to be made.This round, the company decided to focus its developments more towards the server side of the equation - as Raja Koduri himself puts it relative to Infinity Fabric: "Infinity fabric on Vega is optimized for server. It's a very scalable fabric and you will see consumer optimized versions of it in future." Likely, this is Raja's way of telling us to expect MCMs (Multi-Chip-Modules) supported by AMD's Infinity Fabric, much as we see with Ryzen. This design philosophy allows companies to design smaller, more scalable and cheaper chips, which are more resistant to yield issues, and effectively improve every metric, from performance, to power and yield, compared to a monolithic design. NVIDIA themselves have said that MCM chip design is the future, and this little tidbit by Raja could be a pointer towards AMD Radeon's future in that department.That Infinity Fabric is optimized for servers is a reflection of the entire Vega architecture; AMD's RX Vega may be gaming-oriented graphics cards, but most of the architecture improvements aren't capable of being tapped by already-existent gaming workloads. AMD tooled its architecture for a professional/computing push, in a bid to fight NVIDIA's ever-increasing entrenchment in that market segment, and it shows on the architecture. Does this equate to disappointing (current) gaming performance? It does. And it equates to especially disappointing price/performance ratios with the current supply/pricing situation, which the company still hasn't clearly and forcefully defined. Not that they can, anyway - it's not like AMD can point the finger towards the distributors and retailers that carry their products with no repercussion, now can they?Raja Koduri also addressed the current mining/gamer arguments, but in a slightly deflective way: in truth, every sale counts as a sale for the company in terms of profit (and/or loss, since rumors abound that AMD is losing up to $100 with every RX Vega graphics card that is being sold at MSRP). An argument can be made that AMD's retreating market share on the PC graphics card market would be a never-ending cycle of lesser developer attention towards architectures that aren't that prominent in their user bases; however, I think that AMD considers they're relatively insulated from that particular issue due to the fact that the company's graphics solutions are embedded into the PS4, PS4 Pro, Xbox One S consoles, as well as the upcoming Xbox One X. Developers code for AMD hardware from the beginning in most games; AMD can certainly put some faith in that as being a factor to tide their GPU performance over until they have the time to properly refine an architecture that shines at both computing and gaming environments. Perhaps when we see AMD's Navi, bolstered by AMD's increased R&D budget due to a (hopefully) successful wager in the professional and computing markets, will we see an AMD that can bring the competition to NVIDIA as well as they have done for Intel.
Sources:
Raja Koduri's Twitter, via ETeknix
However, it can be said that if AMD had done otherwise, neither gamers nor reviewers would have been impressed with cards that potentially delivered less performance than their NVIDIA counterparts, while consuming more power all the same (even if consuming significantly less wattage). At the rated MSRP (and that's a whole new discussion), this RTG decision was the best one towards increasing attractiveness of RX Vega offerings. However, Raja Koduri does stress Vega's dynamic performance/watt ratios, due to the usage of specially defined power profiles.To our forum-walkers: this piece is marked as an editorial
Raja also touched an interesting subject; namely, that "Folks doing perf/mm2 comparisons need to account for Vega10 features competing with 3 different competitive SOCs (GP100, GP102 and GP104)". This is interesting, and one of the points we have stressed here on TPU: AMD's Vega architecture makes great strides towards being the architecture AMD needs for its server/AI/computing needs, but it's a far cry from what AMD gamers deserve. it's always a thin line to be threaded by chip designers on which features and workloads to implement/improve upon with new architectures. AMD, due to its lack of a high-performance graphics chip, was in dire need of a solution that addressed both the high-performance gaming market and the server/computing market where the highest profits are to be made.This round, the company decided to focus its developments more towards the server side of the equation - as Raja Koduri himself puts it relative to Infinity Fabric: "Infinity fabric on Vega is optimized for server. It's a very scalable fabric and you will see consumer optimized versions of it in future." Likely, this is Raja's way of telling us to expect MCMs (Multi-Chip-Modules) supported by AMD's Infinity Fabric, much as we see with Ryzen. This design philosophy allows companies to design smaller, more scalable and cheaper chips, which are more resistant to yield issues, and effectively improve every metric, from performance, to power and yield, compared to a monolithic design. NVIDIA themselves have said that MCM chip design is the future, and this little tidbit by Raja could be a pointer towards AMD Radeon's future in that department.That Infinity Fabric is optimized for servers is a reflection of the entire Vega architecture; AMD's RX Vega may be gaming-oriented graphics cards, but most of the architecture improvements aren't capable of being tapped by already-existent gaming workloads. AMD tooled its architecture for a professional/computing push, in a bid to fight NVIDIA's ever-increasing entrenchment in that market segment, and it shows on the architecture. Does this equate to disappointing (current) gaming performance? It does. And it equates to especially disappointing price/performance ratios with the current supply/pricing situation, which the company still hasn't clearly and forcefully defined. Not that they can, anyway - it's not like AMD can point the finger towards the distributors and retailers that carry their products with no repercussion, now can they?Raja Koduri also addressed the current mining/gamer arguments, but in a slightly deflective way: in truth, every sale counts as a sale for the company in terms of profit (and/or loss, since rumors abound that AMD is losing up to $100 with every RX Vega graphics card that is being sold at MSRP). An argument can be made that AMD's retreating market share on the PC graphics card market would be a never-ending cycle of lesser developer attention towards architectures that aren't that prominent in their user bases; however, I think that AMD considers they're relatively insulated from that particular issue due to the fact that the company's graphics solutions are embedded into the PS4, PS4 Pro, Xbox One S consoles, as well as the upcoming Xbox One X. Developers code for AMD hardware from the beginning in most games; AMD can certainly put some faith in that as being a factor to tide their GPU performance over until they have the time to properly refine an architecture that shines at both computing and gaming environments. Perhaps when we see AMD's Navi, bolstered by AMD's increased R&D budget due to a (hopefully) successful wager in the professional and computing markets, will we see an AMD that can bring the competition to NVIDIA as well as they have done for Intel.
131 Comments on On AMD's Raja Koduri RX Vega Tweetstorm
AMD has required more compute units to get ahead of Maxwell. Vega only has about 512 more compute units compared to GTX 1080 Ti but it has 1,536 more than GTX 1080. Vega was always meant to take on GTX 1080, and it does, but AMD was blindsided by GTX 1080 Ti. This huge difference in compute units versus performance is where AMD gets it's significant lead in hashing and also why GTX 1080 has a huge margin in their favor in terms of performance per watt.
There's two paths forward for AMD:
1) Debut a Fiji-like chip: as big as the interposer can fit. It will be expensive, it will be power hungry, but it will be king until Volta comes.
2) Create a new, lean microarchitecture that attempts to squeeze every frame it can out of every watt.
Since Ryzen has been pushed out the door, my hope is that AMD fully commits to #2.
"On top of all that, RTG locked Vega's BIOS. Meaning there will be no way to implement any actual BIOS based modifications. RX Vega is a failure no matter what way you spin the story."
Lol, no
www.techpowerup.com/236632/amd-rx-vega-56-to-vega-64-bios-flash-no-unlocked-shaders-improved-performance
forum.ethereum.org/discussion/15024/hows-it-hashin-vega-people
It hasn't even been very long and people can easily flash their Vega BIOS.
There were plenty of other things you could have shit on Vega for but you literally choose non-issues.
- special units for integers and special for float vectors, opposed to each cuda core having both alus inside
- too many special purpose decode hardware blocks, opposed to one unit that knows to decode all and shares internal logic for all
- too many special purpose cache units connected to its special purpose block, opposed to more flexible approach with bigger unified shared cache pool and bigger multipurpose and unified local caches
Basically it's a low-latency throughput favoring design that is wasteful and inflexible. Based on the type of the code running, at some particular moment, bunch of the units are doing nothing still being fully powered on to maybe do something useful in the next clock cycle. To gracefully saturate GCN (both peak efficiency and 100% usage) you should have right ratio of int/float instructions and right amount of memory operations sprinkled through code :laugh: ... which is incidentally easier to do using async computeAMD effectively has a 4096 core co-processor which is why it is fantastic at compute workloads (async and otherwise). Problem is, rendering operates on wave fronts that are very synchronous. I think you're fundamentally right: these two things are at odds with each other. AMD needs to make a new architecture that is mostly synchronous with only some cores capable of async work.
The design mistake with Vega is making the cores even more complex than Fiji, which requires higher voltage to operate, and a longer pipeline which makes it less efficient in dynamic workloads such as gaming. This is also why Vega has lower "IPC" than Fiji.
This is precisely why I'm looking at an i7 5775c instead. It almost matches 7700k, with exceptions actually *favoring* the 5775c by a serious margin, and just needs 4.2 Ghz to do so (versus a 5 Ghz OC on the other) and runs 15-20C cooler. Its curious though that when its a CPU, nobody pays it any mind, when its an AMD GPU, its a horrible product ;) The entire Kaby Lake gen consists of Skylake CPUs with a clock bump that puts them outside their ideal perf/watt.
Food for thought...
Why did AMD not see this? If they did why weren't memory clocks increased?
This probably is probably where Vega would shine in perf/watt. Can you imagine VEGA @1400-1600 core speed with HBM speeds in the same range?
I think that would've been a more compelling release.
What do you all think?
I'd favor covfefe lake in your situation. You need a new platform anyway. Good,new Z97X boards cost a lot now since they're scarce, they severely limit m.2 nvme drives as well with pci-e 2.0 m.2 slots and dmi 2.0. Z170/270 have better nvme ssd support since you can run two 32gb/s m.2 ssds on a decent z270 boards but 5775c outperforms 6700k/7700k in terms of performance/thermals and power draw. 6c/12t i7 will have very good efficiency due to what I described in one of me pervious posts. Even if 8700K will have marginal performance improvement over 7700K it will be able to maintain better efficiency due to lower load on cores in gaming.
No actual diagram for pascal but here is similar setup in maxwell gpu - with all the arrows:
and the cuda core itself didn't change much since fermi afaik:
As you can see, Nvidia doesn't have fetch/decode/dispatch machinery around every 1 scalar + 4 simd units ... they have it around 32 versatile simd/scalar cores.
There is another benefit for having a small GCN compute unit beside async and that's having better yields when salvaging dies for lesser skus. When silicon is bad inside nvidia SM, the whole SM goes away (unless it's gtx 970 as we all know :laugh:)
The trend with Ryzen is fundamentally breaking that mold but even then i think people are more tired of Intels games more than there actual appeal of Ryzen. Odd way to look at things and maybe even small minded of me, but i can't not think about the Athlon Era of cpus. AMD clearly had the better product yet ppl willingly bought Intel. Same goes for the 5870/5970 Era of gpus. They were the best at just about every level yet consumers still bought Nvidia.
i7 5775c hovers around the 360 EUR price point
Cheaper boards make up the difference, and cheaper DDR3 RAMs make the Broadwell build actually a bit cheaper, or you can use that money to buy faster ones. Its really 1=1 across the board, so that's why its still in my mind too :) Another real consideration is that the i7 7700k needs to be pushed to 5Ghz which will increase the cost of proper cooling + little guarantees + delid risk, while a typical 5775c OC @ 4.2 Ghz is very easy to achieve and runs a LOT cooler by default, while being on par in terms of gaming. Total cost of the Broadwell rig will likely be lower, and the OC more feasible.
Really the only thing less good about it is the platform - but then again I'm also fine with just SATA SSDs, they're fast enough tbh and have good price/GB.
Sorry this is really offtopic. :o We are to blame for making AMD suck? Naahh its the other way around. AMD sells us 'tomorrow'. Nvidia and Intel sell us 'today' - and by the time AMD's 'tomorrow' proposition has landed, Nvidia and Intel have moved on to a new 'today'. Of course we have seen launch problems with every company, but with AMD, its a guarantee, and you get the fine wine argument along with that. This combination of drawbacks just doesn't instill confidence in the brand.
Ryzen had a similar issue, and most people called that 'a good launch for AMD'. Go figure! Gaming performance was all over the place, boards weren't ready, RAM was and still is a special breed. Yes yes, new platform and all. But again, these things don't help the brand at all. If its an exception (like Intel's X299 kneejerk response with a completely screwed up product stack), you'll just be surprised at it. If it becomes the rule, many quickly avoid it.
Ryzen could also have been sold so much better - imagine the difference if AMD had been up front about Ryzens' pretty hard 4 Ghz wall, and limited OC potential, but then went on to sell Ryzen on its extreme efficiency and core count advantage below or AT the 4 Ghz wall. They didn't do this explicitly at all, we had to analyze reviews to get there ourselves. I mean, ITS ON THE BOX. The TDPs are very good, especially for a competitor that is reintroducing itself to high end CPUs after a long time, that is and was a massive achievement. They just drop five slides with pricing, architecture and product stacks and some one-liners, and then drop the ball. Every. Single. Fucking. Time. Just releasing one solid product after half a decade does not work to instill confidence in a brand once its been lost. They will have to remain consistent with solid releases and SELL those releases to the public, and Vega's release doesn't help this at all.
Really the longer I watch them, the more I am convinced AMD is its own greatest competitor.