Monday, August 14th 2023
Intel Arrow Lake-S to Feature 3 MB of L2 Cache per Performance Core
Intel's next-generation designs are nearing launch, and we are already getting information about the upcoming generations. Today, we have the information that Intel's Arrow Lake-S desktop/client implementations of the Arrow Lake family will feature as much as 3 MB of level two (L2) cache for each performance core. Currently, Intel's latest 13th-generation Raptor Lake and 14th-generation Raptor Lake Refresh feature 2 MB of L2 cache per performance core. However, the 15th generation Arrow Lake, scheduled for launch in 2024, will bump that up by 50% and reach 3 MB. Given that P-cores are getting a boost in capacity, we expect E-cores to do so as well, but at a smaller size.
Arrow Lake will utilize Lion Cove P-core microarchitecture, while the E-core design will be based on Skymont. Intel plans to use a 20A node for this CPU, and more details will be presented next year.
Source:
via VideoCardz
Arrow Lake will utilize Lion Cove P-core microarchitecture, while the E-core design will be based on Skymont. Intel plans to use a 20A node for this CPU, and more details will be presented next year.
36 Comments on Intel Arrow Lake-S to Feature 3 MB of L2 Cache per Performance Core
8 arrow lake p cores with E cores turned off... That might be my next cpu. All depends on benchmarks VS 8800x3d though. Will be an interesting match up next year.
www.igorslab.de/en/intels-schnellere-raptor-lake-refresh-cpus-koennten-bald-auf-den-markt-kommen/#:~:text=It%20is%20assumed%20that%20Arrow,a%20working%20title)%20in%202026. You shouldn't be buying intel CPU's...
And it's not like a 13700k get hot in games anyway. If you do stuff that's going to stress the CPU that much, chances are you'll be needing the e-cores. Otherwise you might as well just get a Ryzen 9 7900.
This may lead to increase performance but maybe not up to the point people think it would. We will see. They might have improved a lot the lookup mechanism to reduce the performance hit of a larger cache.
Also since cache do not scale very well with lower nodes, i wonder if it will be worth the die space.
And yes, larger caches mean more time to look up things, but engineers are well aware of that. A cache is only increased when it can be built so that the added cache hits will mitigate most/all of the latency hit, while also not burning through unsustainable amounts of power. A lot of simulation and real-world data goes into a decision to increase a particular cache or add another level. That's why we don't have 4MB 1st level caches or 16 caching level already.
e-cores activating in this new game according to the bottom of the review, ruining the experience... and how many more games out there that simply aren't played have the same issue? e-cores are a terrible concept, and if I buy Arrow Lake, I will be turning them off.
That won't do much good to the consumer prices.
You keep saying "I'm getting this CPU" and changing your mind next post, or put in "I'm disabling e-cores" without ever understanding why they're there.
Like pick a side already jfc.
Let's hope Intel deliver on their claims for IPC uplifts in both Meteor and Arrow lake. They'll need to as Zen 5 is looking very good and may be out by April 2024.
Looking forward to Zen 5 vs Arrow lake, but I hope Meteor Lake gives us a good idea of how Intel is moving beyond Raptor lake and if they really have made large gains in efficiency too. If they have Meteor Lake mobile will be strong alternative to Phoenix at least even on iGPU.
No, it have to look up, and a simple way represent it is the larger the cache, the longer it take to look it up.
But to details this a bit, Cache do not cache data, that is a misconception. The CPU at that level is just aware of instruction and memory address. The way the cache work is by caching memory region.
A ultra fast lookup would be to just cache one contiguous memory region. The lookup would be just, is that memory address is in this region? yes/no, then done. The thing is caching a single 3 MB region would have a disastrous cache hit ratio so it wouldn't make sense to do it. Instead they cache smaller region of memory and this is were the trade off happen.
The smaller the region, the higher will be the hit ratio but at the same time, the longer it will take to see if the data is in there. Working with cache isn't just more is better. it's a balance you do when you design a CPU. By example, AMD frequently went with Larger L1 and L2 but had slower cache speed. And by example, Core 2 Duo had 3 MB of L2 per core (merged into a 6 MB shared L2). So 3 MB L2 isn't new. But at that time they didn't had L3.
The thing with cache is you have to look it up every time. If you get a L1 hit, perfect, you just had to look up at that. if you have a cache miss at all level, you still have to check if the data was in L1, then L2, then L3, Then you access it from memory. There is a cache miss penalty over not having cache at all, but if you make your stuff properly, it can way outperform accessing the memory all the time. But the way you do it can greatly impact performance. You need to find the right balance for your architecture.
Generally, L1 is ultra fast and very low latency and very close to the core itself. The L2 is generally dedicated to the core, contain a fair bit much of data needed by that core and L3 is shared across all core and is generally a victim cache (It contain the data that got evicted from L2). That setup worked well.
Intel isn't stupid, so they must think that the new core need a larger cache to be fed. It's possible that they are mitigating the larger cache size with longer pipeline or other technique. In the end it take transistor and the more you put, the larger your chip is and the more expensive it cost.
Designing is always about tradeoff and there, I was just wondering if that was the way to go. In the past, architecture that were near their last redesign frequently had larger cache than their successors because at that point, adding cache was the best thing to do without a full redesign.