• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.

Editorial x86 Lacks Innovation, Arm is Catching up. Enough to Replace the Giant?

Yet AMD is showing that they can do more work with more efficiency. Goes to show you that Intel badly needs a new architecture, the current one is showing that it's not up to the task.
No. AMD's lead comes only from the node advantage.
Zen+ efficiency was on par with 8-9th gen Core. Zen2 efficiency seems roughly on par with Ice Lake (not a big niche to compare...). Same x86 realm, compatible instruction set and 1/0 bits that need to be switched. Smaller transistors mean less energy. There's really no magic inside.

Intel already has a few more modern architectures. They're just not in the product line you're interested in.
 
For extremely old software, there is emulation as an option. For more modern games, it's a little more tricky.

Tricky when it's supposed to control hardware. Like industrial control systems and the like. Possible, but no one is interested in taking the cost, which frankly makes sense.
 
It's not about compliting for ARM being hard or not. It's about the cost of recompliting everything we need. And more importantly: of supporting and optimizing for both architectures.
Most tool chains, standard libraries etc. already have excellent support for ARM. Most applications today don't have x86 optimizations at all, so porting them is easy, but making them perform well might be harder if the ISA is inherently slower. ARM is a very customized ISA, and if you want better performance you have to leverage the custom accelerated features of the specific chip. Even if you just target Apple's chips, in two-three years time they will replace some of it and you have to rewrite your core application. Now imagine if you want to support all the major ARM chips at the market (current and 2-3 years back), you have to rewrite the software over and over again. The other option is to only stick to the basic ARM ISA, and the lacking performance resulting from it. Having an ISA that is very stable and backwards compatible like x86 is a huge advantage for software.

If we now ask companies to make the same software for both x86 and ARM, they'll just ask more for it. And software is already way more expensive than hardware (at least in the enterprise segment).
Developing software is very expensive, especially maintaining it over time.
All the more reason not to prioritize a dozen or so different ARM variations.

There's still a lot that can be done to improve x86 efficiency and flexibility. The heterogeneous architecture idea (e.g. Intel Lakefield) is the most obvious - and for some reason: very underestimated on this forum...
If ARM didn't have big.LITTLE, we probably wouldn't be having this discussion. ARM SoCs would suck just like they did 10 years ago.
Sure, but this is microarchtecture, not ISA.
 
Intel needs a new manufacturing process. They have architecture(s).
To some degree. Dies can only still get so big without yields going to crap. The MCM approach really is the way to solve this problem and we're seeing it a little bit from Intel on that front with their really high core count Xeons, but we're talking about progress that AMD had already made years ago. Even if they get their CPUs on a smaller process, there is still a barrier to how many cores they can throw on a die and that's going to continue to hamstring Intel with both consumer chips and server chips because they're at an extreme disadvantage when it comes to producing CPUs with more than 8 cores.
 
To some degree. Dies can only still get so big without yields going to crap. The MCM approach really is the way to solve this problem and we're seeing it a little bit from Intel on that front with their really high core count Xeons, but we're talking about progress that AMD had already made years ago. Even if they get their CPUs on a smaller process, there is still a barrier to how many cores they can throw on a die and that's going to continue to hamstring Intel with both consumer chips and server chips because they're at an extreme disadvantage when it comes to producing CPUs with more than 8 cores.
People here are too eager to call Intel's lineup unattractive just because it can't make the same core density as AMD/TSMC.
If Intel isn't able to make chips larger than say 150mm2 profitable, they just won't fight in these segments.
But below, with the right pricing, Intel will easily keep the market share leadership.

High-core server chips are a problem and this is something Intel will have to address one way or another.
Everything else is either doable one node behind or not that significant volume-wise. Intel makes many other products - they'll utilize their fab capacity anyway.

And yeah, there's obviously a market share threshold that will trigger Intel to go harder on AMD - e.g. taking over TSMC node supply. At this point they don't see this as necessary.
 
To some degree. Dies can only still get so big without yields going to crap. The MCM approach really is the way to solve this problem and we're seeing it a little bit from Intel on that front with their really high core count Xeons, but we're talking about progress that AMD had already made years ago. Even if they get their CPUs on a smaller process, there is still a barrier to how many cores they can throw on a die and that's going to continue to hamstring Intel with both consumer chips and server chips because they're at an extreme disadvantage when it comes to producing CPUs with more than 8 cores.
MCM on Zen/Zen+ level would not require architectural change from Intel. Their problem from last year compared to Zen/Zen+ is primarily in the lack of QPI links on their smaller dies for optimal topology with many dies. Xeon 9200 is literally two 28-core dies on one package. Xeons are meant for up to 8-socket configurations and there is a topological reason why Zen/Zen+ EPYCs were up to 2-socket.

IO Die is kind of both new and old at the same time. Architecturally speaking this is good old pure chipset + bus configuration. Packaging and bus organization is new and different. No doubt Intel will follow AMD's lead on this one.

None of this is on topic though. Core count and managing core-to-core traffic has different approach for the manycore ARM and other CPUs as well. Mesh (not Intel one but generic mesh principle) does seem to be an up and coming one there, primarily due to the number of cores it seems.
 
They certainly would need something to change, you can't just slap "IF" like links onto any *cove & call it a day! Suspect that's why Keller is there & we'll probably see the first fruits of his work in Sapphire Rapids, likely the same MCM approach? And by the time Intel do go MCM, AMD will have DDR5 & what IF v4 or v5 by then & before anyone says CXL ~ yeah no that won't work right off the bat, if at all.
 
They certainly would need something to change, you can't just slap "IF" like links onto any *cove & call it a day!
Intel has QPI that is used to connect CPUs in multi-CPU system. Its parameters are very-very similar to AMD's inter-die IF links. They would not even need to slap these on, it is already there. They might need a couple additional links though.

Edit:
Sorry, it is called UPI now. Same thing.
 
UPI is inter die connect so not quite the same as IF, from what I remember anyway. There are other differences as well, I'll see if I can find something better on UPI.
 
UPI is inter die connect so not quite the same as IF, from what I remember anyway. There are other differences as well, I'll see if I can find something better on UPI.
UPI is primarily used for inter-socket connection. Using it for inter-die connections is trivial (see Xeon 9200). It should also be flexible enough to make it faster and/or wider if physical connection permits (say, EMIB).
I do not understand exactly what you mean by interconnect for inter-die connect not being quite the same as IF. IF is used throughout Zen precisely for inter-die connections - from one die to another in some topology in Zen/Zen+ and from CCD to IOD on Zen2.
 
I do not understand exactly what you mean by interconnect for inter-die connect not being quite the same as IF. IF is used throughout Zen precisely for inter-die connections - from one die to another in some topology in Zen/Zen+ and from CCD to IOD on Zen2.
IF is also present on the die itself, this is a Zen layout & IIRC it's still the same way Zen2 operates.
ISSCC%207_575px.jpg


The Ryzen Die
 
IF is also present on the die itself, this is a Zen layout & IIRC it's still the same way Zen2 operates.
There is SDF, think of it as an IF switch. This is what CCXs connect to. Traffic inside CCX does not use IF. That one basically goes through L3$.
 
IF includes both SCF & SDF, unless you want to elaborate on that?
The Infinity Fabric consists of two separate communication planes - Infinity Scalable Data Fabric (SDF) and the Infinity Scalable Control Fabric (SCF).
 
No. AMD's lead comes only from the node advantage.
Zen+ efficiency was on par with 8-9th gen Core. Zen2 efficiency seems roughly on par with Ice Lake (not a big niche to compare...). Same x86 realm, compatible instruction set and 1/0 bits that need to be switched. Smaller transistors mean less energy. There's really no magic inside.
Now this may be based upon rumors, but it's said that Zen 3 will actually surpass Intel in terms of IPC. This will allow AMD, in some workloads, be able to do more work with less clock speed which of course leads to less power usage and less heat. It not the GHz that matters, it's what you can do with those GHz that really counts. Higher IPC contributes to better efficiency which is where Intel really hasn't shown any progress as of late. Sure, they've been giving us a few percentages lately but the leaps that AMD has been making makes Intel look like they're standing still.
 
Now this may be based upon rumors, but it's said that Zen 3 will actually surpass Intel in terms of IPC.
Considering Zen2 is already trading blows with current Intel offerings (as fara as IPC is concerned), it would be really surprising if Zen3 didn't surpass Intel.
This will allow AMD, in some workloads, be able to do more work with less clock speed which of course leads to less power usage and less heat.
Lower clocks help, alot at times, but they don't tell the whole story. Die size matters, architecture itself matters, voltage matters (though I don't see voltage going up this round).
 
Zen 2 is already ahead of Skylake or the last iterative "Whiskey" lake or what ever it was, even I don't remember at this point :laugh:

Not Icelake though or Sunny Cove to be precise, even if it's only relegated to laptops.
 
Lower clocks help, a lot at times, but they don't tell the whole story. Die size matters, architecture itself matters, voltage matters (though I don't see voltage going up this round).
It does seem funny that as you ramp up the clock speeds in the Intel camp the voltages required to get those higher clocks increases in ways that aren't exactly nice. Sure, you might win the silicon lottery and get a chip that overclocks nicely while not having to pump quite so much voltage into it to achieve stability but there's a reason why they call it the silicon lottery; most people don't win it. So more often than not, you have to push some hideously high voltages to get those high clock speeds in the Intel camp which leads to decreased overall efficiency.

Never mind the fact that as you push past 5 GHz on Intel it appears that the performance gains drop off a cliff. Sure, you achieved 5.2 GHz... but does it really matter? Has it helped you? Nope. All it has given you is bragging rights at that point, no real performance gains while doing nothing but suck power.
Considering Zen2 is already trading blows with current Intel offerings (as far as IPC is concerned), it would be really surprising if Zen3 didn't surpass Intel.
Exactly. AMD doesn't appear to be taking the foot up off the gas pedal so look forward to even more IPC increases as AMD moves forward. We may very well see an AMD chip that's clocked significantly lower (while using less power) than an Intel chip yet is trouncing them (Intel).
 
Last edited:
No they're not. They're a great idea.


There is no problem with ARM in a laptop form factor - actually imagine a laptop with the TDP of a smartphone and still running its screen perfectly smooth at resolutions like 3200 x 1440, and smooth apps, too.
 
There is no problem with ARM in a laptop form factor - actually imagine a laptop with the TDP of a smartphone
You have a laptop with 2500U (15W). Do you run any kind of tasks that use most of that SoC?
Maybe some of them can be replicated on your smartphone. Try it.
If it's interactive - you'll see if it's as smooth and responsive as on your laptop.
If it's more about processing, time it.

ARM can be a viable alternative for x86 in *some* scenarios and *some* people. And I'll be the first to admit it because I'm very much in that group (I'm mostly cloud-based).
But your suggestions that it's a solution for everyone is absurd.
Seriously, spend some time learning how computers work and how they are used.
still running its screen perfectly smooth at resolutions like 3200 x 1440, and smooth apps, too.
There's something seriously wrong with your PCs if they can't run a smooth video output. :eek:
 
Last edited:
I have and screw that crap. Call me a screen snob if you wish but I just can not do general computing on a 5.5" phone screen.
I agree. i have an iPhone 11 Pro Max with a 6.5" screen and even if I connected it to a monitor, I'm pretty sure I wouldn't want to use it for my entire job. There are some things that make sense, like management-related apps like JIRA, or for communication like email, Slack, or Zoom. It however is grossly insufficient to dev upon because I need the JVM, VIM, Node, and a Postgres server. At the end of the day, it's not unreasonable for me to be using 11-14GB of memory on my tower or laptop and that's not going to change if I were capable of doing it on my phone, yet my phone has only 4GB.

On the other hand, I have a laptop with an i7 8550u in it and its single core performance more or less matches the performance of my overclocked 3930k and also has 16GB of ram, not to mention the capability to drive two 4k displays in addition to the built-in 1080p display. My phone isn't doing that.
 
You are not target market for ARM notebook. While large amounts of RAM is not something that depends on architecture, you are more than likely to run into performance issues on "phone-class TDP" ARM laptop. You are producer or creator, you create content (in its wider meaning).

Target market kind of overlaps with the people who are OK doing things with the phone, except maybe proper screen and keyboard. Consumer is a pretty accurate term here.

For a consumer, ARM is enough. So is an x86 CPU from last decade. The primary place where lowest end of such CPUs run into trouble for consumer is media playback and recording. This is usually resolved with ASIC blocks in the SoC - things like encoders/decoders and image processing in phones or similar blocks on x86 CPUs, usually integrated into iGPU for the latter. As long as iGPU or SoC can decode video stream from Youtube and Netflix, average user can make do with ARM (or Atom) as CPU.
 
Last edited:
  • Like
Reactions: ARF
You are not target market for ARM notebook. While large amounts of RAM is not something that depends on architecture, you are more than likely to run into performance issues on "phone-class TDP" ARM laptop. You are producer or creator, you create content (in its wider meaning).

Target market kind of overlaps with the people who are OK doing things with the phone, except maybe proper screen and keyboard. Consumer is a pretty accurate term here.

For a consumer, ARM is enough. So is an x86 CPU from last decade. The primary place where lowest end of such CPUs run into trouble for consumer is media playback and recording. This is usually resolved with ASIC blocks in the SoC - things like encoders/decoders and image processing in phones or similar blocks on x86 CPUs, usually integrated into iGPU for the latter. As long as iGPU or SoC can decode video stream from Youtube and Netflix, average user can make do with ARM (or Atom) as CPU.


Yes, we are speaking about consumer apps, he shifts to offtopic workstations...
 
Back
Top