• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.

GeForce GTX 980 PCI-Express Scaling

W1zzard

Administrator
Staff member
Joined
May 14, 2004
Messages
28,959 (3.75/day)
Processor Ryzen 7 5700X
Memory 48 GB
Video Card(s) RTX 4080
Storage 2x HDD RAID 1, 3x M.2 NVMe
Display(s) 30" 2560x1600 + 19" 1280x1024
Software Windows 10 64-bit
PCI-Express x16 3.0 is well established in the market, and the majority of gamers are using the interface. But what happens if you end up in a slot-bandwidth-constrained situation? We are testing NVIDIA's latest GeForce GTX 980 flagship in 17 games, at four resolutions, including 4K, to assess what performance to expect.

Show full review
 
Last edited:
Thanks for doing an update W1zzard.
 
Interesting to see that 3.0 8x is sometimes faster than 3.0 16x.
 
idk it seems extremely odd on how dynamic and ryse, wow, and wolfenstein differences are. It really seems like there is a frame limiter detecting the speed of the bus and adjusting the limit accordingly.
 
Nice review still no real difference.

Wow big boost from WoW MoP to WoW WoD. 161FPS to 231FPS at 1080p
 
I get the Id Tech 5 engine with it's constant mega texture streaming, but I don't get what Ryse has to move over the PCIE bus other than draw calls. All resources that kind of games need are usually preloaded in VRAM at level loading.
 
Thanks for the post W1zzard, it was a good read.
 
All resources that kind of games need are usually preloaded in VRAM at level loading.
Not anymore

It really seems like there is a frame limiter detecting the speed of the bus and adjusting the limit accordingly.
I see no mechanism how a game could do that (detecting PCIe bandwidth is not trivial, I know from GPU-Z). Also why would a game do that, and why would a gamedev invest time for it :)
 
Great article.
As someone with a dual-gpu card, a 7990, is there any chance you could benchmark just a few games with a dual-gpu card?
The reason why I'm asking, is that I'm still running a x58 system with an overclocked i7 920 @ 4.2ghz but it's still pcie 2.0.
 
Nice article, but it does prove that it's pointless to waste money on a socket 2011 platform for a gaming machine, or, that 3 way 8x SLI on a Z97 platform is going to only show a 0-5% decrease in performance for a CPU, RAM and MOBO setup that costs at least 50% less. In fact, that loss in performance would likely be mitigated by the faster clock speeds native to the Devil's Canyon chips.

Socket 2011 gaming rigs are for people with more money than sense.
 
Last edited:
Socket 2011 gaming rigs are for people with more money than sense

Strictly speaking about pci-e lanes, yes but those 4 or 8 extra threads of a 2011v3 chip might come in handy in the future, quite a few game engines multithread pretty well already.
 
I wonder if AMD cards would behave differently? I was thinking of buying the next high end card from them for my pci-e 2.0 board and was just thinking about this a few days ago. If new games like Ryse and Wolfenstein start showing a difference it might be finally time to start planning for an update for my 2500K.
 
Excellent article guys! thanks!
 
I wonder if AMD cards would behave differently? I was thinking of buying the next high end card from them for my pci-e 2.0 board and was just thinking about this a few days ago. If new games like Ryse and Wolfenstein start showing a difference it might be finally time to start planning for an update for my 2500K.
Great article.
As someone with a dual-gpu card, a 7990, is there any chance you could benchmark just a few games with a dual-gpu card?
The reason why I'm asking, is that I'm still running a x58 system with an overclocked i7 920 @ 4.2ghz but it's still pcie 2.0.
I have no plans for any other PCIe scaling tests, not until new cards are released from AMD.
 
First page:
"While PCI-Express 1.0 pushes 250 MB/s per direction, PCI-Express 2.0 pushes 500 MB/s, and PCI-Express 3.0 doubles that to 1 GB/s. While the resulting absolute bandwidth of PCI-Express 3.0 x16, 32 GB/s, might seem like overkill, the ability to push that much data per lane could come to the rescue of configurations such as 8-lanes (x8) and 4-lanes (x4)."


PCI-Express 3.0 at 16x has a ~16GB/s bandwidth, not 32GB/s
 
W1z is right, it's 32GB/s.

1GB/s x 16 x 2 = 32GB/s.

PCI-E is a duplex connection, so each lane is 1GB/s in two directions, so the total bandwidth is 2GB/s per lane(1GB/s in each direction). So the total bandwidth for an x16 3.0 slot is 32GB/s.
 
Last edited:
Thank you for this article. It makes me feel more at ease that my i7 2700k which only supports x16 2.0 is not going to restrict current high end cards.

I don't get any restriction with my i7-2600 System. I have a pair of R9-280X-OC cards in it.
I just bought a third 280X-OC card, but I'll have to swap out my motherboard to run Tri-Crossfire with it.
 
Wow great article @W1zzard, its nice to see some formal testing updated with a recent card about PCI-E bandwidth. Its such an odd subject to get into because there are not many areas that testing to this extent is done to show people when they question it.
 
∆ IIRC, the story was pretty much the same with AGP...

:lovetpu:
 
@ wizard,

"The most surprising find to me is the huge performance hit some of the latest games take when running on limited PCIe bandwidth. The real shocker here is certainly Ryse: Son of Rome, based on Crytek's latest CryEngine 4. The game seems to constantly stream large amounts of data between the CPU and GPU, taking a large 10% performance hit by switching to the second-fastest x16 3.0 configuration. At x4 1.1, the slowest setting we tested, performance is torn down to less than a third, while running lower resolutions! Shocking!

Based on id's idTech5 engine, another noteworthy title with large drops in performance is Wolfenstein: The New Order. Virtual Textures certainly look great in-game, providing highly detailed, non-repeating textures, but they also put a significant load on the PCI-Express bus. One key challenge here is to have texture data ready for display in-time. Sometimes too late, it manifests as the dreaded texture pop-in some users have been reporting.

Last but not least, World of Warcraft has received a new rendering engine for its latest expansion Warlords of Draenor. While the game doesn't look much different visually, Blizzard made large changes under the hood, changing to a deferred rendering engine which not only disallows MSAA in-game, but also requires much improved PCI-Express bandwidth."


All you're really proven is there is a slight gain or loss in certain scenarios, but you don't go into depth as to why PCIe 16x 2.0 performs equal to or better than 3.0. Doubt it really matters. Though, I agree with some points of your message, but you aren't really proving any more than a drop or gain in average of what, around 10% in any of the other scenarios. It seems informative, but also a waste of your own time. In addition to that, games that are either MMOs or highly-progressed games like Crysis 3, BF4, Wolfenstein 3D, and others, will make better use of 3.0 over 2.0. One good example of this will probably be Star Citizens in the not to distant future. I would highly suggest using Planetside 2 and EQN for upcoming benches. For the higher resolutions (above 1080p), you'll probably see a higher use in 3.0 if you enabled more AA at 4k resolutions.

Here's an idea. Instead of sitting in the Shrine in World of Warcraft, why don't you conduct test during a Garrosh 25 man fight at Ultra Settings. Tell us what the results are of the PCIe Lane Saturations after that. I would think that would be more vital information than just staring at a wall to stare at the in-game FPS meter to see how high your FPS can get. Also, why don't you measure the same games with 2-way, 3-way, and 4-way SLI. It's not like NVidia has anything to hide right...
 
Interesting to see that 3.0 8x is sometimes faster than 3.0 16x.

I probably within the realm of error you could say something in video slightly diff happened cause the small difference or cpu usage.

I wonder if AMD cards would behave differently? I was thinking of buying the next high end card from them for my pci-e 2.0 board and was just thinking about this a few days ago. If new games like Ryse and Wolfenstein start showing a difference it might be finally time to start planning for an update for my 2500K.

I would probably expect AMD cards would yield pretty much same kinda results.
 
Back
Top