System Name | Silent/X1 Yoga/S25U-1TB |
---|---|
Processor | Ryzen 9800X3D @ 5.575ghz all core 1.24 V, Thermal Grizzly AM5 High Performance Heatspreader/1185 G7 |
Motherboard | ASUS ROG Strix X670E-I, chipset fans replaced with Noctua A14x25 G2 |
Cooling | Optimus Block, HWLabs Copper 240/40 + 240/30, D5/Res, 4x Noctua A12x25, 1x A14G2, Mayhems Ultra Pure |
Memory | 64 GB Dominator Titanium White 6000 MT, 130 ns tRFC, active cooled |
Video Card(s) | RTX 3080 Ti Founders Edition, Conductonaut Extreme, 18 W/mK MinusPad Extreme, Corsair XG7 Waterblock |
Storage | Intel Optane DC P1600X 118 GB, Samsung 990 Pro 2 TB |
Display(s) | 32" 240 Hz 1440p Samsung G7, 31.5" 165 Hz 1440p LG NanoIPS Ultragear, MX900 dual gas VESA mount |
Case | Sliger SM570 CNC Aluminium 13-Litre, 3D printed feet, custom front, LINKUP Ultra PCIe 4.0 x16 White |
Audio Device(s) | Audeze Maxwell Ultraviolet w/upgrade pads & LCD headband, Galaxy Buds 3 Pro, Razer Nommo Pro |
Power Supply | SF1000 Plat, full transparent custom cables, Sentinel Pro 1500 Online Double Conversion UPS w/Noctua |
Mouse | Razer Viper V3 Pro 8 KHz Mercury White & Pulsar Supergrip tape, Razer Atlas, Razer Strider Chroma |
Keyboard | Wooting 60HE+ module, TOFU-R CNC Alu/Brass, SS Prismcaps W+Jellykey, LekkerV2 mod, TLabs Leath/Suede |
Software | Windows 11 IoT Enterprise LTSC 24H2 |
Benchmark Scores | Legendary |
You're still missing the point.If the difference is so little you would hardly notice or tell then I don't think it really matters. Power Efficiency simply doesn't really matter for this comparison in my opinion. Just because the 7900XTX is a smidge bit faster than the 4080S (usually only by 1% to 4%), suddenly didnt mean people were recommending the 7900XTX over the 4080S in raster. I think the same logic applies here.
Not to mention, OP mentioned overclocking a potential 5070Ti to beat a 4080 (which it already matches as is) which opens up another can of worms for power efficiency, and ATP if your gonna overclock power efficiency probably isnt your concern anymore. And even in this regard, an overclocked 5070Ti from what im reading is still well within 5% difference range. Cooler matters too in that regard but it seems pretty consistent across all the charts I saw for different models (avoiding FE for comparison due to no FE card for 5070Ti)
TL;DR, power efficiency, and by extension, other stuff such as raster, really dont matter in this arguement. I dont see a point in bringing them up. Your choice is gonna entirely hinge on the features. (and price, if you can get a good price of course.)
5070Ti if its at a price similar to a 4080S (which the cheapest ive seen are), if you already own a 4080S, you should pass unless you value the features exclusive to the 5070Ti and can actually get it. If you don't, and were considering getting a 4080S, then, in that specific instance, go ahead. Even more so if you can buy a 5070Ti at its actual MSRP. If your not an american thats a factor too, due to stuff like VAT, but thats getting into a whole other can of worms.
He doesn't need to sell it, it's within return window.I would say to get it if you can at MSRP. It's worth it for MFG IMO and you can likely sell your 4080s for the same price or more.
System Name | Don't do thermal paste, kids |
---|---|
Processor | Ryzen 7 5800X |
Motherboard | ASUS PRIME B550-PLUS AC-HES |
Cooling | Thermalright Peerless Assassin 120 SE |
Memory | Silicon Power 32GB (2 x 16GB) DDR4 3200 |
Video Card(s) | GTX 1060 3GB (temporarily) |
Display(s) | Gigabyte G27Q (1440p / 170hz DP) |
Case | SAMA SV01 |
Power Supply | Firehazard in the making |
Mouse | Corsair Nightsword |
Keyboard | Steelseries Apex Pro |
You're still missing the point.
You are exactly highlighting more of what I'm talking about. Nobody bought a 7900XTX because its faster in raster. They bought it for other reasons. This comparison is no different, except its actually a very close comparison. So why should power efficiency even matter? Just because the 7900XTX on paper wins in raster doesn't mean it actually matters.. you can still value that small, microscopic win, that nobody will really realistically care about or even notice, but the comparison is so close here it doesnt matter.7900XTX vs 4080 had pros/cons. 7900XTX ~5% faster in raster, bit more VRAM (debatable if a real positive beyond bigger number better, since no CUDA for professional applications and games don't use 24 GB), slightly better in Linux (at the time), 40% slower in RT, no DLSS, more power draw etc.
well dlss4 is also on 4000 cards, what are the other features? mfg? i don't know if i ever will use, i have a 120hz oled tv for gameHave 4080S? Keep, unless;
Value DLSS4 and MFG? Get 5070Ti.
Return window open? If you have a guaranteed way of getting a 5070Ti, get 5070Ti. But do NOT pay more than 4080S MSRP. Ideally, pay less.
Don't have one but wanted it, now 5070Ti out? Get 5070Ti.
This is basically what I would recommend OP. 50 Series launch has been rough, just don't be a idiot when buying a 5070Ti and you should be okay if your doing that. OC, Power Effiency, Raster, etc are not what you should worry about. Just focus on the features, or other small things the 5070Ti has over the 4080 if it benefits you. Or if you already can get one for same / smaller price.
![]()
You are exactly highlighting more of what I'm talking about. Nobody bought a 7900XTX because its faster in raster. They bought it for other reasons. This comparison is no different, except its actually a very close comparison. So why should power efficiency even matter? Just because the 7900XTX on paper wins in raster doesn't mean it actually matters.. you can still value that small, microscopic win, that nobody will really realistically care about or even notice, but the comparison is so close here it doesnt matter.
I think we both are saying the same thing but just applying it in different ways. And for me, a microscopic win might as well be a draw.
DLSS4 works just fine on a 4080SValue DLSS4 and MFG? Get 5070Ti.
System Name | Mean machine |
---|---|
Processor | AMD 6900HS |
Memory | 2x16 GB 4800C40 |
Video Card(s) | AMD Radeon 6700S |
If I ever dared mentioned that I prefer nvidia over amd cards for the physX support at any point in the last 5 years in these forums, people would nail me on the cross. and would suggest that im just trying to find excuses not to buy amd. Now that nvidia removes support for it, it became a news worthy issue."condoning such actions"
My dude it's 32 bit software for which a direct replacement has existed for more than 15 years, and this is 2025. Should hardware manufacturers keep supporting all standards for infinity? Seems like a waste of die space.
Besides, it's not like the game is unplayable without PhysX, or that it will magically make a 15 year old game look contemporary.
Seems like a very similar issue to how modern fast computers have problems running very old games, as framerate tied to physics etc and the sheer speed of modern hardware causes stuff to break. If turning on Physx on a game from the 2000s is critical, I'm sure there's millions of old GPUs/systems on Ebay or in your garage gathering dust.
Could NVIDIA write a translation layer similar to what Intel did for old DX games with their discrete GPU release? Maybe. Is it worth it? Unlikely.
My bet is this whole "issue" was the first time most people even remembered PhysX existed.
I'm guessing NVIDIA made this move to 64 bit only CUDA/PhysX for the same reason Apple went 64 bit only, Intel tried to, and Android is in the process of doing, it simplifies things and allows more focus on currently important things.
Correction, the newest ARM processors for Android already only support 64 bit code.
Started with the Pixel 7. I didn't hear much outcry then.
![]()
Pixel 7, the first 64-bit-only Android phone
With 64-bit-only devices now reaching users, developers should start paying extra attention to testing their apps and updates for 64-bit-only devicesandroid-developers.googleblog.com
But hey, NVIDIA bad right?
System Name | Don't do thermal paste, kids |
---|---|
Processor | Ryzen 7 5800X |
Motherboard | ASUS PRIME B550-PLUS AC-HES |
Cooling | Thermalright Peerless Assassin 120 SE |
Memory | Silicon Power 32GB (2 x 16GB) DDR4 3200 |
Video Card(s) | GTX 1060 3GB (temporarily) |
Display(s) | Gigabyte G27Q (1440p / 170hz DP) |
Case | SAMA SV01 |
Power Supply | Firehazard in the making |
Mouse | Corsair Nightsword |
Keyboard | Steelseries Apex Pro |
DLSS4 works just fine on a 4080S
well dlss4 is also on 4000 cards
System Name | Tiny the White Yeti |
---|---|
Processor | 7800X3D |
Motherboard | MSI MAG Mortar b650m wifi |
Cooling | CPU: Thermalright Peerless Assassin / Case: Phanteks T30-120 x3 |
Memory | 32GB Corsair Vengeance 30CL6000 |
Video Card(s) | ASRock RX7900XT Phantom Gaming |
Storage | Lexar NM790 4TB + Samsung 850 EVO 1TB + Samsung 980 1TB + Crucial BX100 250GB |
Display(s) | Gigabyte G34QWC (3440x1440) |
Case | Lian Li A3 mATX White |
Audio Device(s) | Harman Kardon AVR137 + 2.1 |
Power Supply | EVGA Supernova G2 750W |
Mouse | Steelseries Aerox 5 |
Keyboard | Lenovo Thinkpad Trackpoint II |
VR HMD | HD 420 - Green Edition ;) |
Software | W11 IoT Enterprise LTSC |
Benchmark Scores | Over 9000 |
Yes, and the featureset is slightly changed, you might miss some ROPs, and the driver branch is exhibiting major problems.It pretty much is, same perf, more features, better power efficiency. If he can get it cheaper too and return the 4080 what exactly is your problem?
You're comparing PC gaming to an Android phone. You don't get it at all."condoning such actions"
My dude it's 32 bit software for which a direct replacement has existed for more than 15 years, and this is 2025. Should hardware manufacturers keep supporting all standards for infinity? Seems like a waste of die space.
Besides, it's not like the game is unplayable without PhysX, or that it will magically make a 15 year old game look contemporary.
Seems like a very similar issue to how modern fast computers have problems running very old games, as framerate tied to physics etc and the sheer speed of modern hardware causes stuff to break. If turning on Physx on a game from the 2000s is critical, I'm sure there's millions of old GPUs/systems on Ebay or in your garage gathering dust.
Could NVIDIA write a translation layer similar to what Intel did for old DX games with their discrete GPU release? Maybe. Is it worth it? Unlikely.
My bet is this whole "issue" was the first time most people even remembered PhysX existed.
I'm guessing NVIDIA made this move to 64 bit only CUDA/PhysX for the same reason Apple went 64 bit only, Intel tried to, and Android is in the process of doing, it simplifies things and allows more focus on currently important things.
Correction, the newest ARM processors for Android already only support 64 bit code.
Started with the Pixel 7. I didn't hear much outcry then.
![]()
Pixel 7, the first 64-bit-only Android phone
With 64-bit-only devices now reaching users, developers should start paying extra attention to testing their apps and updates for 64-bit-only devicesandroid-developers.googleblog.com
But hey, NVIDIA bad right?
System Name | Silent/X1 Yoga/S25U-1TB |
---|---|
Processor | Ryzen 9800X3D @ 5.575ghz all core 1.24 V, Thermal Grizzly AM5 High Performance Heatspreader/1185 G7 |
Motherboard | ASUS ROG Strix X670E-I, chipset fans replaced with Noctua A14x25 G2 |
Cooling | Optimus Block, HWLabs Copper 240/40 + 240/30, D5/Res, 4x Noctua A12x25, 1x A14G2, Mayhems Ultra Pure |
Memory | 64 GB Dominator Titanium White 6000 MT, 130 ns tRFC, active cooled |
Video Card(s) | RTX 3080 Ti Founders Edition, Conductonaut Extreme, 18 W/mK MinusPad Extreme, Corsair XG7 Waterblock |
Storage | Intel Optane DC P1600X 118 GB, Samsung 990 Pro 2 TB |
Display(s) | 32" 240 Hz 1440p Samsung G7, 31.5" 165 Hz 1440p LG NanoIPS Ultragear, MX900 dual gas VESA mount |
Case | Sliger SM570 CNC Aluminium 13-Litre, 3D printed feet, custom front, LINKUP Ultra PCIe 4.0 x16 White |
Audio Device(s) | Audeze Maxwell Ultraviolet w/upgrade pads & LCD headband, Galaxy Buds 3 Pro, Razer Nommo Pro |
Power Supply | SF1000 Plat, full transparent custom cables, Sentinel Pro 1500 Online Double Conversion UPS w/Noctua |
Mouse | Razer Viper V3 Pro 8 KHz Mercury White & Pulsar Supergrip tape, Razer Atlas, Razer Strider Chroma |
Keyboard | Wooting 60HE+ module, TOFU-R CNC Alu/Brass, SS Prismcaps W+Jellykey, LekkerV2 mod, TLabs Leath/Suede |
Software | Windows 11 IoT Enterprise LTSC 24H2 |
Benchmark Scores | Legendary |
To be clear, it still has PhysX, just not the 32 bit version that is only used in ~15 year old games.well dlss4 is also on 4000 cards, what are the other features? mfg? i don't know if i ever will use, i have a 120hz oled tv for game
and a feature i will miss is physix
Lol, I'm simply giving an example of why NVIDIA did this and how this kind of thing is completely normal. The card generation still has backwards compatibility, just not for a niche toggleable feature on a small set of games almost two decades old, which still run just fine without it and have a CPU fallback or processing on a second GPU if it's really so important to you (doubt.jpg).You're comparing PC gaming to an Android phone. You don't get it at all.
Its the same reason Windows still has backwards compatibility.
Yes, when everyone remembered it existed since it's a vector to attack NVIDIA.If I ever dared mentioned that I prefer nvidia over amd cards for the physX support at any point in the last 5 years in these forums, people would nail me on the cross. and would suggest that im just trying to find excuses not to buy amd. Now that nvidia removes support for it, it became a news worthy issue.
To be clear, it still has PhysX, just not the 32 bit version that is only used in ~15 year old games.
Yes, when everyone remembered it existed since it's a vector to attack NVIDIA.
System Name | Silent/X1 Yoga/S25U-1TB |
---|---|
Processor | Ryzen 9800X3D @ 5.575ghz all core 1.24 V, Thermal Grizzly AM5 High Performance Heatspreader/1185 G7 |
Motherboard | ASUS ROG Strix X670E-I, chipset fans replaced with Noctua A14x25 G2 |
Cooling | Optimus Block, HWLabs Copper 240/40 + 240/30, D5/Res, 4x Noctua A12x25, 1x A14G2, Mayhems Ultra Pure |
Memory | 64 GB Dominator Titanium White 6000 MT, 130 ns tRFC, active cooled |
Video Card(s) | RTX 3080 Ti Founders Edition, Conductonaut Extreme, 18 W/mK MinusPad Extreme, Corsair XG7 Waterblock |
Storage | Intel Optane DC P1600X 118 GB, Samsung 990 Pro 2 TB |
Display(s) | 32" 240 Hz 1440p Samsung G7, 31.5" 165 Hz 1440p LG NanoIPS Ultragear, MX900 dual gas VESA mount |
Case | Sliger SM570 CNC Aluminium 13-Litre, 3D printed feet, custom front, LINKUP Ultra PCIe 4.0 x16 White |
Audio Device(s) | Audeze Maxwell Ultraviolet w/upgrade pads & LCD headband, Galaxy Buds 3 Pro, Razer Nommo Pro |
Power Supply | SF1000 Plat, full transparent custom cables, Sentinel Pro 1500 Online Double Conversion UPS w/Noctua |
Mouse | Razer Viper V3 Pro 8 KHz Mercury White & Pulsar Supergrip tape, Razer Atlas, Razer Strider Chroma |
Keyboard | Wooting 60HE+ module, TOFU-R CNC Alu/Brass, SS Prismcaps W+Jellykey, LekkerV2 mod, TLabs Leath/Suede |
Software | Windows 11 IoT Enterprise LTSC 24H2 |
Benchmark Scores | Legendary |
So do any of these "most played" games in the world use 32 bit PhysX? Or have they, perhaps, been updated, and don't use the same engine they did 20 years ago. Making the argument that they're "20 year old" games a bit... forced. To my knowledge it's mainly singleplayer games that implemented PhysX, and I can't think of many singleplayer games from 15-20 years ago that have consistently high player counts.You keep mentioning the age of the games as if old games should not be playable due to their age, which I think is hilarious because some of the most played games in the world are 10-20 years old. WoW, Runescape, DOTA 2, GTA:V, etc.
Borderlands 2 is the worst case scenario here as it still gets around 5000 players each day on Steam and is quite frequently sold in the trilogy pack to new buyers who will probably want to play it. On the 50 series without a dedicated card for 32-bit PhysX it tanks into the single digit FPS range during combat. Not fun, and an 18 year old PhysX accelerator is faster and smoother. Mirror's Edge without PhysX is also an entirely different game. There's an entire physical interaction system with glass impacting the player which gets turned off entirely, but very few people still play that game and it is still playable in the 'PhysX off' mode without the extra gameplay variety.
An optional feature most devs didn't even bother to implement due to being vendor locked, that was completely replaced 15 years ago with a 64 bit version, is not comparable to a complete OS rewrite to a new architecture, requiring a translation layer without which literally all legacy apps would not work, at all.It's objectively a bad move to remove compatibility features without notice and without alternatives. Even the most hated company on this forum, Apple, made an effort with Rosetta to bring legacy functionality forward both times they dropped an enter architecture. NVIDIA could have, didn't, and it shows where they stand.
System Name | Tiny the White Yeti |
---|---|
Processor | 7800X3D |
Motherboard | MSI MAG Mortar b650m wifi |
Cooling | CPU: Thermalright Peerless Assassin / Case: Phanteks T30-120 x3 |
Memory | 32GB Corsair Vengeance 30CL6000 |
Video Card(s) | ASRock RX7900XT Phantom Gaming |
Storage | Lexar NM790 4TB + Samsung 850 EVO 1TB + Samsung 980 1TB + Crucial BX100 250GB |
Display(s) | Gigabyte G34QWC (3440x1440) |
Case | Lian Li A3 mATX White |
Audio Device(s) | Harman Kardon AVR137 + 2.1 |
Power Supply | EVGA Supernova G2 750W |
Mouse | Steelseries Aerox 5 |
Keyboard | Lenovo Thinkpad Trackpoint II |
VR HMD | HD 420 - Green Edition ;) |
Software | W11 IoT Enterprise LTSC |
Benchmark Scores | Over 9000 |
Understood, and I don't think the comparison flies just quite so well, because legacy gaming is a thing, and keeping featuresets up for gaming is also a thing. So I view this as a loss, too. And no, a lot of older games do not get remade or updated to newer engines, and qpeople don't have that desire either. The precedent here is a dangerous one. Its crazy there is no backwards compatibility for something so simple.To be clear, it still has PhysX, just not the 32 bit version that is only used in ~15 year old games.
Lol, I'm simply giving an example of why NVIDIA did this and how this kind of thing is completely normal. The card generation still has backwards compatibility, just not for a niche toggleable feature on a small set of games almost two decades old, which still run just fine without it and have a CPU fallback or processing on a second GPU if it's really so important to you (doubt.jpg).
Yes, when everyone remembered it existed since it's a vector to attack NVIDIA.
Processor | Ryzen 7800X3D |
---|---|
Motherboard | ASRock X670E Taichi |
Cooling | Noctua NH-D15 Chromax |
Memory | 32GB DDR5 6000 CL30 |
Video Card(s) | MSI RTX 4090 Trio |
Storage | P5800X 1.6TB 4x 15.36TB Micron 9300 Pro 4x WD Black 8TB M.2 |
Display(s) | Acer Predator XB3 27" 240 Hz |
Case | Thermaltake Core X9 |
Audio Device(s) | JDS Element IV, DCA Aeon II |
Power Supply | Seasonic Prime Titanium 850w |
Mouse | PMM P-305 |
Keyboard | Wooting HE60 |
VR HMD | Valve Index |
Software | Win 10 |
An optional feature most devs didn't even bother to implement due to being vendor locked, that was completely replaced 15 years ago with a 64 bit version, is not comparable to a complete OS rewrite to a new architecture, requiring a translation layer without which literally all legacy apps would not work, at all.
NVIDIA not bothering to write a translation layer isn't ideal, but it's far from equivalent to if Apple didn't make Rosetta. It's more surprising considering NVIDIA software support is typically the benchmark.