Monday, December 18th 2023
AMD 5th Gen EPYC "Turin" Pictured: Who Needs Accelerators When You Have 192 Cores?
AMD's upcoming server processor, the 5th Gen EPYC "Turin," has been pictured as an engineering sample is probably being evaluated by the company's data-center or cloud customers. The processor has a mammoth core-count of 192-core/384-thread in its high-density cloud-focused variant that uses "Zen 5c" CPU cores. Its regular version that uses larger "Zen 5" cores that can sustain higher clock speeds, also comes with a fairly high core-count of 128-core/256-thread, up from the 96-core/192-thread of the "Zen 4" based EPYC "Genoa."
The EPYC "Turin" server processor based on "Zen 5" comes with an updated sIOD (server I/O die), surrounded by as many as 16 CCDs (CPU complex dies). AMD is expected to build these CCDs on the TSMC N4P foundry node, which is a more advanced version of the TSMC N4 node the company currently uses for its "Phoenix" client processors, and the TSMC N5 node it uses for its "Zen 4" CCD. TSMC claims that the N4P node offers an up to 22% improvement in power efficiency over N5, as well as a 6% increase in transistor density. Each of the "Zen 5" CCDs is confirmed to have 8 CPU cores sharing 32 MB L3 cache memory. A total of 16 such CCDs add up to the processor's 128-core/256-thread number. The high-density "Turin" meant for cloud data-centers, is a whole different beast.The "Turin" processor uses 12 "Zen 5c" CCDs. Much like "Zen 4c" is to "Zen 4," the "Zen 5c" is a physically compacted version of the larger "Zen 5" core that has the same ISA (instruction sets) and IPC, but typically runs at lower clock speeds than the regular "Zen 5" cores. It's meant for high core-count processors. The high-density "Turin" MCM has the same sIOD as the regular "Turin," but with 12 "Zen 5c" CCDs. Each CCD has 16 "Zen 5c" cores share a 32 MB L3 cache. This is interesting—if you recall, the current "Zen 4c" CCD has two CCX (CPU core complexes) that each has 8 "Zen 4c" cores share a 16 MB L3 cache. The "Zen 5c" CCD doubles the total addressable L3 cache.
Sources:
YuuKi_AnS (Twitter), VideoCardz
The EPYC "Turin" server processor based on "Zen 5" comes with an updated sIOD (server I/O die), surrounded by as many as 16 CCDs (CPU complex dies). AMD is expected to build these CCDs on the TSMC N4P foundry node, which is a more advanced version of the TSMC N4 node the company currently uses for its "Phoenix" client processors, and the TSMC N5 node it uses for its "Zen 4" CCD. TSMC claims that the N4P node offers an up to 22% improvement in power efficiency over N5, as well as a 6% increase in transistor density. Each of the "Zen 5" CCDs is confirmed to have 8 CPU cores sharing 32 MB L3 cache memory. A total of 16 such CCDs add up to the processor's 128-core/256-thread number. The high-density "Turin" meant for cloud data-centers, is a whole different beast.The "Turin" processor uses 12 "Zen 5c" CCDs. Much like "Zen 4c" is to "Zen 4," the "Zen 5c" is a physically compacted version of the larger "Zen 5" core that has the same ISA (instruction sets) and IPC, but typically runs at lower clock speeds than the regular "Zen 5" cores. It's meant for high core-count processors. The high-density "Turin" MCM has the same sIOD as the regular "Turin," but with 12 "Zen 5c" CCDs. Each CCD has 16 "Zen 5c" cores share a 32 MB L3 cache. This is interesting—if you recall, the current "Zen 4c" CCD has two CCX (CPU core complexes) that each has 8 "Zen 4c" cores share a 16 MB L3 cache. The "Zen 5c" CCD doubles the total addressable L3 cache.
40 Comments on AMD 5th Gen EPYC "Turin" Pictured: Who Needs Accelerators When You Have 192 Cores?
Some high-performance cores on a standard CCD and a bunch of Zen4c cores with a decent IGP and RAM controller.
If you mean using Phoenix to put cores inside the IO die (like Intel is doing for Meteor Lake), it's an idea, but AMD has lots of work to do to make its Ryzen chiplet design more efficient. Currently it's just a cost saver and is an order of magnitude worse than its own Navi31 Fabric and Intel EMIB/Foveros. Whole point of island E-cores is to enable shutting down the entire compute die for power. IFOP is way too primitive.
If AMD is going to make Strix Halo, disaggregation is already a requirement. Costs/yields already forced Navi31 that way, it will happen eventually for APUs.
Dragon Range isn't even an APU, it's a minimal effort repackaging of Raphael for FP7/FP8 without an IHS, and has the same shit iGPU. Has nothing to do with producing Strix Halo.
Not to mention, massive server IO die takes up most of the space and will also undoubtedly change if the node changes. I don't disagree, most of the leaks probably came from MLID lmao
But even if it's not 40CU, the writing is on the wall for monolithic APU. Regardless of Meteor Lake, Intel is no longer as constrained by die size the same way AMD has been. Just remains to be seen what high perf interconnect solution AMD is planning for chiplet APUs.
although since Z4c/Z5c are not like Intel E core, it would be interesting to see it in desktop
it was never about core dense CPU on desktop ;) at least for me, it was for the person i quoted ;)