Monday, May 20th 2024
AMD to Present "Zen 5" Microarchitecture Deep-dive at Hot Chips 2024
AMD is slated to deliver a "Zen 5" microarchitecture deep-dive at the Hot Chips 2024 conference, on August 25. The company is widely expected to either unveil or announce its next-generation processors based on the architecture, in its 2024 Computex keynote on June 3, so it remains to be seen if the deep-dive follows a product launch, or predates it. Either way, Hot Chips talks tend to be significantly more detailed than the product launch pre-briefs that we get; and so we hope to learn a lot more about the architecture.
A lot rides on the continued success of "Zen 5" to deliver a double-digit percentage IPC increase over its predecessor, while also introducing new microarchitecture-level features; and leveraging new foundry processes at TSMC, to deliver competitive processors to Intel. Unlike Intel, which has implemented hybrid CPU cores across its product stack, AMD continues to make traditional multicore processors, and refuses to level even the chips that contain regular and high-density versions of its "Zen 4" cores as "hybrid."
Sources:
Hot Chips, Wccftech
A lot rides on the continued success of "Zen 5" to deliver a double-digit percentage IPC increase over its predecessor, while also introducing new microarchitecture-level features; and leveraging new foundry processes at TSMC, to deliver competitive processors to Intel. Unlike Intel, which has implemented hybrid CPU cores across its product stack, AMD continues to make traditional multicore processors, and refuses to level even the chips that contain regular and high-density versions of its "Zen 4" cores as "hybrid."
14 Comments on AMD to Present "Zen 5" Microarchitecture Deep-dive at Hot Chips 2024
AMD chips are the same microarchitecture with just a physical space optimisation and reduction of cache.
Intel is a different archtetcure between E core and P Core which did and can lead to weird interactions with certain software.
I think you meant ‘label’ not ‘level’. How about ‘Chips that Matter’?
the Scheduler was there so that things that needed performance (games, video encoding etc) were pushed onto the P cores but if you didnt need that performance/were only running background tasks it would push things to E cores and power down the P cores aggressively to save power etc. Most of this was meant to be done in the hardware scheduler built into the CPU directly and all windows was doing was supplying info regarding the app and if it was a background task etc.
This - sequential RAM read bandwidth - mattered for AI workloads, which is apparently the newest thing everyone is capitalizing on. Lunar Lake is apparently going after that with its LPDDR5X-on-package design.
EDIT: To illustrate the problem, my current system does LLM inference on CPU at an indicated 60GB/s, while the system I upgraded from, an AMD Cezanne-based setup with DDR4-3200, did it at around 43GB/s. The performance improvement is noticeable, but at first less than expected. If I set my RAM to JEDEC speed with default FCLK(DDR5-4800, 1600MHz FCLK, same as the Cezanne), then there's ~zero improvement.
For every token generated, an LLM would need to iterate through the entire active model weight once, and those are several GB minimum for models of useful performance. It's going to be a problem, if powerful client-side AI ever works out for the general consumer.
Not all AI workloads are LLM, but LLM and Transformer-based models in other modalities are the stars of the current AI boom.
A better name would be "Innovation & Chips", "New Chips", "Modern Chips", "Fast Chips", or whatever else, but not literally "hot".
We have a brand of beer in Australia called Minimum Chips, which I quite like.