Monday, October 17th 2022

AMD Ryzen 9 7950X Posts Significantly Higher Gaming Performance with a CCD Disabled

AMD Ryzen 9 7950X 16-core processor exhibits some strange behavior with regards to the max boost frequency spread among its cores. A multi-chip module with two 8-core CCDs (CPU complex dies); we noticed early on in our review that the cores located in CCD-1 boost to a higher frequency than the ones in CCD-2, with differences as high as 300 MHz. CapFrameX noticed that when CCD-2 is disabled on a machine running Windows 11 22H2, the processor actually puts out higher gaming performance, by as much as 10%. This is mainly because the cores in CCD-2, with a lower maximum boost frequency no longer handle processing load from the game; and with CCD-2 disabled, CCD-1 has all of the processor's power budget—up to 230 W—to itself, giving it much higher boost residency across its 8 cores.
Source: CapFrameX
Add your own comment

47 Comments on AMD Ryzen 9 7950X Posts Significantly Higher Gaming Performance with a CCD Disabled

#26
stimpy88
I'll put money on this being a Windows bug, or some kind of AGESA issue.
Posted on Reply
#27
AusWolf
The KingSo is the 7700X not single CCD? If it is should it not be outperforming the 7950X in benchmarks?

When you disable 1 CDD on the 7950X does that mean the L3 cache available is 32MB or is it still at 64MB?

1CDD with access to 64MB of L3 should be faster in games??
Yes, but the 7950X has a higher max. boost clock, so supposedly a more aggressive boosting algorithm, not to mention higher default TDP as well.
WirkoRegarding the concentration of heat, wouldn't it be better to disable half of the cores in each CCD? Also the full L3 would remain active, and even if it's not really acting as one 64MB block, it still seems better to have two 32MB blocks than only one.
That's a nice theory, but you're also gaining core-to-core latency. It needs to be tested, I guess.
Posted on Reply
#28
The King
AusWolfYes, but the 7950X has a higher max. boost clock, so supposedly a more aggressive boosting algorithm, not to mention higher default TDP as well
Yes, that is pretty obvious from just looking at the specs. the 7700X is unlocked you can get it to boost higher and to a higher TDP provided you can keep it cool using PBO etc.
In Single thread you are not going to use up all your TDP!

So to clear things up, why are we not seeing overclocked 7700X beating 7950X in games? 30% increasing in peformance is pretty high.

AMD could just cut the 7950X in half make 2 CPUs and double their sales instead of selling the 7700X. :p

Ok maybe not double their sales but a 8 core 16T CPU with that peformance would outsell most CPUs on the market today if priced right? Yes/No?
Posted on Reply
#29
Max(IT)
7950X is not a gaming CPU, even if AMD marketing is pushing to it (because it is easier to sell to enthusiasts).
You don't need 16 cores to play a game, and with 2 different quality CCDs these are the results.
I experienced something similar with 3900X and 5900X, but the situation was less noticeable because even if CCD2 was worse, the boost was quite similar.
Posted on Reply
#30
InVasMani
The KingSo is the 7700X not single CCD? If it is should it not be outperforming the 7950X in benchmarks?

When you disable 1 CDD on the 7950X does that mean the L3 cache available is 32MB or is it still at 64MB?

1CDD with access to 64MB of L3 should be faster in games??
Disabled it'll still be 64MB and the CCD is binned higher as well. The L3 cache is shared it's labeled 2x32MB due to half the cache being shared from the other CCD, but even if you disable the cores you aren't disabling the shared cache.

I think what's happening is you've got less thread contention and more thermal headroom primarily, but the same amount of cache and also outside of thread contention where scheduling is dynamically changing around thread assignments the second CCD is going to cause some latency even though AMD has minimized it you can't eliminate it outright same goes with E cores which is why some have preferred disabling them for Intel likewise.

What is a little strange is why Intel/AMD aren't just including the full shared L3 cache across the lineup regardless of cut down core count parts. I wonder what the cost of the larger L3 cache is relative to additional cores to manufacturer and how that trade off balances out.
Posted on Reply
#31
Soupladel
DrGrossmanIs not the boost, is the latency penalty between CCDs. My 7950x is locked on all cores at 5.5 on all games and exhibits the same behavior when you set the affinity on just one CCD. For example on Riftbreaker the difference is ~30 fps. Other games respond well with affinity on physical cores only, like Cyberpunk, Battlefield 2042. In most of the games I saw an uplift in average and 1% by manually setting the affinity. The issue is old but for some reason is accentuated on this platform. Windows doesn't care, AMD doesn't care, developers doesn't care either....so don't cripple your CPU, just use Process Lasso :rolleyes:
I have noticed the 5.5GHz cap and it only seem to start happening after they introduced a BIOS for my board (Crosshair X670E Hero) that included the ComboAM5PI 1.0.0.3 patch A. Prior to this, all my cores were capable of boosting to 5.75 even on the slower CCD when gaming.
Posted on Reply
#32
Wirko
InVasManiDisabled it'll still be 64MB and the CCD is binned higher as well. The L3 cache is shared it's labeled 2x32MB due to half the cache being shared from the other CCD, but even if you disable the cores you aren't disabling the shared cache.

I think what's happening is you've got less thread contention and more thermal headroom primarily, but the same amount of cache and also outside of thread contention where scheduling is dynamically changing around thread assignments the second CCD is going to cause some latency even though AMD has minimized it you can't eliminate it outright same goes with E cores which is why some have preferred disabling them for Intel likewise.

What is a little strange is why Intel/AMD aren't just including the full shared L3 cache across the lineup regardless of cut down core count parts. I wonder what the cost of the larger L3 cache is relative to additional cores to manufacturer and how that trade off balances out.
Manufacturing defects in cores are common. Defects in L3 are also common as its surface area is similar to that of the cores. I assume that an L3 slice with a defect or two is still usable and remains active at 99% capacity - but if it isn't the case then there may be some technical reason to disable the core next to it.
Posted on Reply
#33
SirMaster
Why disable a CCD? Why not just run the game with CPU affinity for CCD1 cores only?

That way you still have the other CCD operating for other background processes and such.
Posted on Reply
#34
ARF
AMD needs to release a new chipset driver to fix this. And tell M$ to get their act together with OS power management plans.

You are not supposed to lose gaming performance because someone at that ecosystem doesn't know what to do.
Posted on Reply
#35
InVasMani
I'm wondering if AMD could configure the second CCD get parked and just enabled for AVX-512 workloads with or without the other CCD and if there would be a latency advantage to configuring a option for that. It seems like there very much could be given it's not a monolithic core so even under perfect circumstances you'll have slight latency between CCD's though it's been minimized a good deal over earlier Ryzen chips by AMD and should continue to be refined and improved further.
Posted on Reply
#36
DrGrossman
SoupladelI have noticed the 5.5GHz cap and it only seem to start happening after they introduced a BIOS for my board (Crosshair X670E Hero) that included the ComboAM5PI 1.0.0.3 patch A. Prior to this, all my cores were capable of boosting to 5.75 even on the slower CCD when gaming.
Yes, weird behavior. Its only happens in gaming. On desktop environment it boost properly with loads that exceed 10% but in games that barely use 10% is just caped at 5.5. (X670E AORUS MASTER - AGESA 1.0.0.3 A)
Posted on Reply
#37
nangu
This AMD's CCX behaviour is as old as with the 3900X/3950X.
Posted on Reply
#38
Soupladel
DrGrossmanYes, weird behavior. Its only happens in gaming. On desktop environment it boost properly with loads that exceed 10% but in games that barely use 10% is just caped at 5.5. (X670E AORUS MASTER - AGESA 1.0.0.3 A)
Only Core 0 of mine boosts correctly regardless of the activity or load, the other cores will not go above 5.5 regardless of what i do
Posted on Reply
#39
Wirko
InVasManiWhat is a little strange is why Intel/AMD aren't just including the full shared L3 cache across the lineup regardless of cut down core count parts.
Actually, AMD does. I checked all parts with Zen 3 and Zen 4 chiplets, and the only exceptions that have L3 halved seem to be the Epyc 7513 (8 chiplets by 4 cores) and the weirder case of Epyc 7453 (4 chiplets by 7 cores). Also two Zen 4 Epycs, purportedly.
InVasManiIt seems like there very much could be given it's not a monolithic core so even under perfect circumstances you'll have slight latency between CCD's though it's been minimized a good deal over earlier Ryzen chips by AMD and should continue to be refined and improved further.
Sadly, this doesn't seem to be true. Anandtech did their measurements and found the inter-CCD latency to be 75-80 ns in the 7950X, which is only about 5 ns less than in the 5950X. The way it is, it doesn't make sense to ask the other CCD for data, it's slower than RAM access.

Note: what is important, and affects performance, is the latency to access slices of L3, which are located at various distances from the core that wants to access data. It's impossible to reliably measure that directly. But there's an indirect way, which is to measure core-to-core latency, and we know that each core is paired with a slice of L3 (at least in x86, E-core clusters being an exception).
Posted on Reply
#40
Silvinjo
ARFAMD needs to release a new chipset driver to fix this. And tell M$ to get their act together with OS power management plans.

You are not supposed to lose gaming performance because someone at that ecosystem doesn't know what to do.
This is Microsoft problem not AGESA or chipset.
Posted on Reply
#41
SethNW
Interesting, maybe AMD should take advantage of whatever backend changes Microsoft made for Alder Lakes performance and efficiency cores and use the thing in a way where faster CCD is treated as performance cores and slower as efficiency cores. So lightly threaded stuluff run only on faster CCD, while slower one handled background stuff. Since vast majority of games barely scale to 8 cores and already get most of performance at 6 cores, both SMT enabled. Of course assuming those changes MS made can be used for that. But Alder Lake does have no issues with games using efficiency cores by mistake, so system seems to work, at least after some initial fixes.
Posted on Reply
#42
Dirt Chip
So basically a 7700x with better binning.
This is beyond undervolt to up your performance, it`s undercore and underCCD..
The KingSo to clear things up, why are we not seeing overclocked 7700X beating 7950X in games? 30% increasing in peformance is pretty high.
Better CCD binning for 7950x.
It appears that more than 1 CCD hurts FPS that is countered by higher frequencies with 5900\5950\7900\7950.
Might be why the 5800X3d is so much faster in some gaming beside the extra cache- just 1 CCD.
If AMD will successful engineer the 7800X3d without the frequencies limit it will dominate gaming easily.
Posted on Reply
#43
medi01
Selayaany plans for a tpu review/bench of that?
The plan that I'd have would be certain select companies addressing that in their software.
Posted on Reply
#44
regorwin56
tabascosauzI noticed @W1zzard tested on Win 11 - even on Zen 3 the 11 scheduler is far less accommodating of 2CCD. I've been saying for a long time that it not only treats 2CCD like a big.Little CPU (CCD2 acting as "little"), but also regularly disrespects CCX hierarchy by juggling load from CCD1 preferred cores all the way onto Windows' designated CCD2 background core. Which inevitably incurs inter-CCD performance penalties. Windows 10 at the very least still kept loads within CCDs. Wouldn't be surprised if the 7950X isn't the only CPU suffering this way. Gamersnexus' review with the 7950X's abysmal showing in a few games seems to suggest that.

Can avoid some scheduler behaviours by disabling CPPC Preferred Cores on 1CCD CPUs, but for 2CCD it doesn't do much to avoid Windows picking some CCD2 core.



The 5800X3D result is not the same.

Uncore in CPU-Z is Fabric FCLK for Ryzens.

L3 runs on its own clock that usually (but not always, especially for X3D) mirrors core clocks. It doesn't share clock domain, nor voltage domain with Fabric.
I am thinking that the test is performed under Win11 22H2. If I change the system to Win10 and do not turn off the CCD, can I achieve the performance mentioned in the article (after Win11 22H2 turns off the CCD)
Posted on Reply
#45
SL2
CapFrameX have more interesting updates, this time with a 5900X.

"Windows 11 22H2 can cause performance issues on PCs with Ryzen CPUs. This is a comparison of feature update 22H2 vs reinstallation OS (including 22H2)."

Dirt ChipBetter CCD binning for 7950x.
It appears that more than 1 CCD hurts FPS that is countered by higher frequencies with 5900\5950\7900\7950.
I'm not so sure about that, since limiting power draw even down to 65 W (see below), which should affect clock frequency much more than binning IMO, doesn't make much of a difference in 1080 gaming benchmarks compared to what's happening in the OP.

Although the CPU doesn't run nowhere near max power draw in games to begin with, TPU measured 87 W. That's why I doubt that the fully enabled 7950X is power limited in games, compared to when disabling one CCD.

A 5,6 % higher clock speed (comparing 7700X & 7950X) won't translate into the 16 % higher FPS in the OP.


Posted on Reply
#46
regorwin56
MatsCapFrameX have more interesting updates, this time with a 5900X.

"Windows 11 22H2 can cause performance issues on PCs with Ryzen CPUs. This is a comparison of feature update 22H2 vs reinstallation OS (including 22H2)."



I'm not so sure about that, since limiting power draw even down to 65 W (see below), which should affect clock frequency much more than binning IMO, doesn't make much of a difference in 1080 gaming benchmarks compared to what's happening in the OP.

Although the CPU doesn't run nowhere near max power draw in games to begin with, TPU measured 87 W. That's why I doubt that the fully enabled 7950X is power limited in games, compared to when disabling one CCD.

A 5,6 % higher clock speed (comparing 7700X & 7950X) won't translate into the 16 % higher FPS in the OP.


I guess the original article test was done under Win11 22H2. If I change the system to Win10 without turning off the CCD, can I achieve the 10% performance improvement mentioned in the article (Win11 22H2 7950X after turning off the CCD)

I would like to know if the performance boost from closing the CCD happens on Win10 (7950X)
Posted on Reply
#47
SL2
regorwin56I guess the original article test was done under Win11 22H2. If I change the system to Win10 without turning off the CCD, can I achieve the 10% performance improvement mentioned in the article (Win11 22H2 7950X after turning off the CCD)

I would like to know if the performance boost from closing the CCD happens on Win10 (7950X)
I dunno, since they're blaming the upgrade alone and not 22H2 itself, you might as well reinstall W11 22H2.

Recently I've seen AMD claiming that there's no issue here, but I still doubt that..
Posted on Reply
Add your own comment
Nov 21st, 2024 06:29 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts