Wednesday, July 14th 2021
AMD Zen 4 Desktop Processors Likely Limited to 16 Cores, 170 W TDP
We have recently seen several reputable rumors confirming that AMD's Zen 4 Raphael desktop processors will be limited to 16 cores with 2 compute units. There were previous rumors of a 24 core model with 3 compute units however that now seems unlikely. While the core counts won't increase some skews may see a TDP increase up to 170 W which should offer some performance uplift. AMD is expected to debut their 5 nm Zen 4 Raphael desktop processors in 2022 which will come with support for PCIe 5.0 and DDR5. The processors will switch to a new AM5 LGA1718 socket and will compete with Intel's Alder Lake-S successor Raptor Lake which could feature 24 cores.
Source:
@patrickschur_
75 Comments on AMD Zen 4 Desktop Processors Likely Limited to 16 Cores, 170 W TDP
We aren't going to see 32 cores on the desktop until Apple or Intel start creeping back up.
Also, regarding the news: at the moment, they should focus on single core performance and efficiency for the first gen on AM5 and increase the core count on second/third gen. At the current moment, I simply can't see how pushing more cores would benefit the average consumer. If you need more cores, why not go Threadripper?
A key selling point for desktop CPU right now is still gaming performance. I suppose that AMD want to use 2 CCD (2x8 cores) and gave them more power to go higher clock than having another chiplet that would need to use a significant portion of the power envelope preventing the others core to achieve the higher frequency.
Single thread performance is still the key performance metrics everyone need to get on no matter what people say here. The higher it is, the higher performance you will get with each core you add. You have to always balance working on getting your single thread performance as good as possible then leveraging to more cores.
Also, no matter what architecture or ISA you are on (x86-64, ARM, RISC-V), the key is to feed the big cores. 2 channels (or even the 4 channel (2x2x32) of DDR5 might not be enough to feed another full CCD. Larger cache might help but not fix the problem as a whole.
Also AMD is benefits a lot from its chiplet architecture, but there are drawback. I suppose that they would have released smaller cpu with Zen 2 (like 10-12 core) but they had to go to 2 full CCD. The next step is 24 witch is another big jump.
We will see, pretty sure that if that become a problem, they will adapt, but I am not sure if that it would be something to keep the crown but not really something widely available. AMD already have to give the best chips to the *950 and *900 cpu to stay in the AM4 power envelope. They would have to do crazy binning to do the same thing with 3 chiplet instead of 2.
But i would be more interested as a someone that primarily need performance for gaming at 8 core with 96MB of L3 than a 24 cores with the same amount of L3 cache. Even more if that 8 cores run at higher frequency (or is just buyable).
I had a 6 core i7-970 however you are incorrect when saying it offered no performance improvement. That cpu can still play most recent BF games with a decent gpu due to the extra cores a i7-920 cannot. Anything that can use those extra cores will including handbrake, any compression applications etc.
What is going on with CPUs? Since the Pentium 4 stupidity (I know, NetBurst was not bad in some scenarios, but for consumers it was mostly not great.) and then Bloomfield being a horrible space heater it was, not to mention Bulldozer which was actually one of the worst CPUs I ever owned, after all this we were going in a good direction with TDP going down, efficiency going up and got to a point where even the top end consumer CPUs were reasonable for home use - I know several people who still use slightly overclocked 6700k/7700k/8700k with something like RTX 3060/3070 with no disturbing bottlenecking. Nowadays it's somehow acceptable to sell a 170W, 16 core monstrosity as a consumer product, so we're back to the days of pumping more power and creating more heat just to get a higher number on a cardboard box. I know there are people who want this, I'm sure AMD did their due diligence in market research, so: Who are you people and what do you use such consumer hardware for? This is honest curiosity, not condescension.
I know there is e-peen enlargement, also known as "bragging rights" (or conspicuous consumption known otherwise as a Veblen effect if you're more academically inclined), which is probably the main reason for high-end consumer hardware existence since before home computers were a thing. If you need performance for work you don't fart around, buy big boy stuff and write it off your taxes as a business expense, that's where your Xeons, Threadrippers and such live. I know people with multi-CPU servers with Tesla cards in their basements, but they use them to earn money or do scientific research, but a 170W 16 core consumer CPU? To do what, games can't even saturate eight cores properly without GPU becoming the limiting factor and consumer software is just not that demanding. What, you do time-critical building modeling with Revit as a hobby or something? I'd like to again stress the fact that his is honest curiosity. I personally have a Xeon/Quadro machine at home, but I use it do do actual work and earn money, and that's not consumer hardware, hence my curiosity regarding the viability of such products in a consumer setting. Maybe I'm just oblivious to a fundamental change in consumer usage patterns and people really, really need to have six hundred browser windows opened at once or a full suite of Autodesk software is now a must-have for everyone.
intel sandy to skylake (3 newer architectures) is no where near that
And yes they can use those vacant threads. Browser, stream(s) , some transcode and a bunchbof other stuff will likely be used concurrently.
If they release chips on mainstream platform with more cores thats getting into TR territory and will totally castrate their own product line.
And now with gaming studios taking advantage of more cores, scaling mostly up to 8, with a rare few offering options up to 16, the bigger livestreamers and content creators who make their living off such now start buying 12 and 16 core CPUs both to future-proof their rigs for a few years, and still have reserve cores for the real-time editing/encoding and other software they run. Sure, GPU makes have helped to off-load some of the encoding from the CPU to the GPU (NVENC notably comes to mind), but not everyone uses GPU encoding for one reason or another.
I myself am not familiar with all that goes into Livestreaming so I don't know the full extent of CPU reqs for a high-end, do-everything rig, but I have a passing familiarity with using consumer level hardware for "small, local business" ventures, since one of my friends uses a 5950X + 6700XT desktop for their 3-man photography and videography side-business. As they're the one mostly doing the editing and post-processing for photos and videos taken by the group, they like not having to wait for extended periods trying to scrub through 4k video footage or wait just as long for sampling and testing applied effects to large photos.
A less-used, but noted case was one Linus made; being able to take a high-end CPU + multi GPU combo, and split its cores to separate workstations. While Linus meme'd using a TR or EPYC CPU and multiple GPUs, he mentioned the viability of using say, a 16c CPU split 4 ways into "4c terminals" in an environment that doesn't need much more than CPU processing and basic graphics (provided by the terminal's basic display output), or a 16c CPU + 2 GPU rig split into 2 "8c+GPU" terminals.
Just an FYI 32 core mainstream chips should become a reality come zen5/6 on 3nm TSMC.
AMD doesn't need to pull an Intel & I doubt they'll do so given the competitive landscape today, not to mention real threat of ARM eating x86 everywhere for breakfast, lunch & dinner. If ARM was even remotely competitive back in 2005 (on PC) then you can bet every dollar in crypto exchanges that Intel would either have released 8 core chips early 2010's or gone bankrupt today!
1981: You really need more than 640K of memory on your daily home and gaming machine?
What I'm saying is we need to keep an eye on what AMD is or isn't doing. Sure, eight cores may be overkill nowadays, but what about in a few years?
Intel perpetuated the idea that four cores were plenty from at least 2007 all the way up until Zen happened in 2017. Also notice that when Zen came out, all of a sudden, "regular" programs could take advantage of it.
I hate Intel as much as the next guy, but AMD isn't perfect either. They're both companies, and when they've gotten far enough ahead of their competitors, they'll slow down innovation until we get another Intel from Sandy Bridge to Kaby Lake situation.
The price increase Zen 3 received was the first red flag for me (but that didn't stop me from buying it!).
But there is a rules about multicore/multithread scaling. Amdahl law en.wikipedia.org/wiki/Amdahl%27s_law
The thing is a lot of the thing that can easily be parallelized can be done by GPU, or can be done by larger cores with larger SIMD. So for desktop users, there won't be a lot of benefits of having very large number of cores.
it have been shown in the past and it will be shown again in the future, you better have less faster core than more slower cores. At least on desktop. On server, well depending on the workload and stuff that might be another story.
In the end, all that are child game, it do not really matter. The only things that really matter is how much performance you get from the chips and that all it matter. If someone could release a 1 core CPU that blow out of the water everything we have currently, who would really care?
When you can't scale any higher, you make the work harder. Back in the 00s, why would you ever go above 768p ?? But today, we are doing 1080p regularly and enthusiasts are all in 4k. With literally 16x the pixels, you suddenly have 16x more work and computers can scale to 16x the size.
4k seems to be the limit for how far is reasonable (at least for now). But there's also raytracing coming up: instead of making more pixels, we're simply making "each pixel" much harder to compute. I expect Gustafson's law to continue to apply until people are satisfied with computer-generated graphics. (and considering how much money Disney spends on Pixar rendering farms... you might be surprised at how much compute power is needed to get a "good cartoony" look like Pixar's Wreck it Ralph, let alone CGI Thanos in the movies).
Video game graphics are very far away from what people want still. The amount of work done per frame can continue to increase according to Gustafson's law for the foreseeable future. More realistic shadows, even on non-realistic games (ex: Minecraft with Raytracing or Quake with Raytracing) wows people.
--------------
For those who aren't in the know: Amdahl's law roughly states that a given task can only be parallelized to a certain amount. For example: if you're looking forward 1-move in Chess, the maximum parallelization you can get is roughly 20 (there are only ~20 moves available at any given position).
Therefore, when the "chess problem" is described as "improve the speed of looking forward 1-move ahead", you run into Ahmdal's law. You literally can't get any faster than 20x speedup (one core looking at each of the 20-moves available).
But... Gustafson's law is the opposite, and is the secret to scaling. If the "chess problem" is described as "improve the number of positions you look at within 30 seconds", this is Gustafson's law. If you have 400-cores, you have those 400-cores analyze 2-moves ahead (20x for the 1st move, 20x that for the 2nd move). All 400-cores will have work to do for the time period. When you get 8000 cores, you have one core look at 3 moves ahead (20x20x20 positions), and now you can look at 8000 positions in a given timeframe. Etc. etc. As you can see: "increasing the work done" leads to successful scaling.
----------
Some problems are Amdahl style (render this 768p image as fast as possible). Some other problems are Gustafson's (render as many pixels as you can within 1/60th of a second). The truth for any given problem lies somewhere in between the two laws.
And while your 970 can still play games, I'd bet good money that a current gen i3 quad core would absolutely smoke it. Single core performance has come a long way in the last decade, as has multicore performance. One can look up benchmarks for an i3-10320 and see it is ranked slightly higher then a i7-980. You suggested that the fundamental change for intel performance wise was due to DRAM speed increases instead of core changes. That, you yourself, have proven incorrect, as both haswell and the core 2 quad can use DDR3. You said it, not me. Well, when 8 cores finally isnt enough, AMD already has 12 and 16 cores available for you! and by that time, AMD will have likely have higher then 16 core parts available.
Funny how when the 2000 series came out and still features 8 cores nobody was loosing their mind on AMD "stagnating", yet after the same amount of time at 16 cores suddenly AMD is worse then intel. This would be like people complaining that intel was stagnating right after sandy bridge came out because it was still 4 cores when the vast majority of games were still single threaded. By the time it's relevant this AM4/AM5 platform will be ancient history.
Regular models like the vanilla 11700 are fine, the moment Intel tries to force 14nm to be more than it is, you get CPUs that can't be cooled with a 360mm radiator.
2006 was already ripe for multithreaded software and still took like 2 or 3 years to take advantage. So the 640k memory limit. Is not even close to the situation we are now, we are far from exploiting actively 8 cores for daily tasks or even gaming.