Thursday, December 11th 2008
PhysX will Die, Says AMD
In an interview with Godfrey Cheng, Director of Technical Marketing in AMD's Graphics Products Group, Bit-Tech.net has quoted him saying that standards such as PhysX would die due to their proprietary and closed nature. Says Mr. Cheng:
"There is no plan for closed and proprietary standards like PhysX," said Cheng. "As we have emphasised with our support for OpenCL and DX11, closed and proprietary standards will die."Bit-Tech.net interviewed the AMD person to get the company's take on EA and 2K's decision to adopt NVIDIA PhysX across all of their worldwide studios, earlier this week. Interestingly, when asked about how the major publishers such as EA adopting PhysX across all of their studios would impact the propagation of the API, Cheng responded with saying that monetary incentives provided to publishing houses alone won't help a great deal in propagating the API, and that the product (PhysX) must be competitive, and that AMD viewed Havoc and its physics simulation technologies as leaders. "Games developers share this view. We will also invest in technologies and partnerships beyond Havok that enhances gameplay." he added. PhysX is a proprietary physics simulation API created by Ageia technologies, which was acquired and developed by NVIDIA. You can read the full Bit-Tech.net interview with Godfrey Cheng here.
67 Comments on PhysX will Die, Says AMD
Time will tell i guess.
We shall see.
I think they may be counting their chickens here.
whoops dbl post my bad.
The main issue stems from a lack of developers even bothering to conform to such proprietary standards; they want to do it their own way. CPU Based physics engines in game engines such as the CryEngine 2, or even the latest Source engine generally are sufficient.
Sure AMD is being bold and attempting to scare away nvidia shareholders, but its true. Havok basically rips Physx in terms of how much its implemented. Cannot agree more.
The performance hit is going to happen regardless of what API is used to create the physics. If they both are creating the same level of physic, the performance hit will be the same, as the graphics cards are asked to render more on the screen due to the physics.
Edit: Of course I hope AMD realizes that they have kind of screwed themselves by saying that. History shows that when a company says their competitions product will fail, the product usually becomes wildly popular.
Since the GRAW 2 production days, PhysX has come a long way. This can be witnessed via the many examples out there on the internet. Whether it be a fluid demo, particle demo, a ripping flag or my balls bouncing off each other. Either way, the realism it provides is a vital step. EA and 2K seem to think so.
PhysX enabled reduces performance on lower end systems, and/or systems missing required hardware. Ofcourse we can get the CPU to run the PhysX stuff, but whats going to run everything else . . . .
Cheng and all of AMD is scared that PhysX will evolve to the only next step it has. To become a part of the A.I, and the game play.
PhysX cant get worse, and we all know that this technology will eventually evolve. Simulating, and ripping is second grade and will never sum up to be the best.
I wonder how the 295GTX will cope with all this.
PhysX is an nvidia only way of doing this
DirectX 11 is doing an open (any video card) version of this.
ATI/AMD are saying that nvidias closd one will die, and the open version will live on.
"We won't support a standard where Nvidia is faster until we have something to compete."
And open your eyes guys Nvidia IS and would probably be faster at physics calculations because they made changes to the architecture, like more CPU-like cahing system, ultra-light branching prediction, etc. Not exactly branching prediction but something that paliates the effects of lacking one. THAT's why Nvidia cards are faster in F@H for example, where more than number crunching is required. At simple number crunching Ati is faster: video conversion.
That advantage Nvidia has would apply to PhysX, OpenCL, DX11 physics or ANY hardware physics API you'd want to throw in. Their claim is just an excuse until they can prepare something. For instance they say they support HAvok, Intel OWNED Havok. Open standards? Yeah sure.
One more thing is that, PhysX is a physics API and middleware, with some bits of an engine here and there just as Havok, that can RUN on various platforms unchanged: Ageia PPUs, x86 cpus, Cell Microprocessor, and CUDA and potentially any PowerPC. It does not run directly on Nvidia GPUs, as you may remember CUDA is a x86 emulation API that runs on Nvidia cards. Once OpenCL is out, PhysX will be possible to do through OpenCL just as well as through CUDA. As long as Ati has good OpenCL there shouldn't be any problems, until then they could make PhysX run through CAL/Stream for FREE, BUT they don't want to, because it would be slower. IT'S at simple as that.
Another lie there, which you have to love, is that they claim that PhysX is just being used for eye candy. IT IS being used only for that AMD, yeah, but tell WHY. Because developers have been said hardware physics will not be supported on Ati GPUs until DX11, that's why. Because they are working hard along with Intel to make that statement true. Nvidia has many demos where PhysX are being used for a lot more, so it can be done.
AMD is just double acting. It's a shame Ati/AMD, a shame, I remember the days you were honest. I know bad times and experiences make personalities change, but this is inexcusable as well as the fact that all the advertising campaign has been based on bashing every initiative made by Nvidia instead of making your's better.
This sentence resumes it all (spealing of Havok on their GPU): Like back in the 90's, because they are facing competition in something they can't compete, they are downplaying it. They know it's a great thing, they know it's the future, but they don't want that future to kick start yet. YOU SIMPLY CAN'T DOWNPLAY SOMETHING AND SAY IT WILL DIE, WHILE AT THE SAME TIME YOU'RE HARD WORKING ON YOUR OWN BEHIND THE CURTAINS!! AND USING INTEL'S HAVOK!!!
We know how accelerated graphics history evolved, despite what they said back then the GPU has become the most important thing and so will the hardware physics. Just as back then, they are just HOLDING BACK the revolution until they can be part of it. Clever, from a marketing stand point, but DISHONEST. You won't have my support Ati, you already pulled down another thing that I liked a lot: Stereo 3D. You can cheat me once, but not more.
I for one think this could kill AMD for good , if 2-3 big games launch with some physx thing and the difference is big bettwen them it could kill AMD graphics departement forever.
Nvidia could continue to battle in 3dmakrs and games perf. with AMD but it seems to go on for a long time and one advantage like this could end the competition a little faster.
I feel sorry for them , intel is kicking their asses , now Nvidia too , second place forever for AMD.
Some of you are talkiing about studios and games adopting physx but what have you seen so far? Mirror's edge seems to do a bit of showcase work for physx but how many would really buy that game? Being an EA game through and though I personally have a hard time believing the game would offer anything more than what I could get watching the trailer or some demos. Otherwise I haven't seen physx do anything that havok wasn't already doing. There's no extra edge here. There's probably some incentives from Nvidia - but then that comes back to what this guy was saying in the first place;
"We cannot provide comments on our competitor's business model except that it is awfully hard to sustain support by monetary incentives. The product itself must be competitive. We view Havok technologies and products to be the leaders in physics simulation and this is why we are working with them. Games developers share this view." This lends itself to why this stuff won't kill ATI... end of the list is Apple, didn't they put together the OpenCL standard? Which do you think they'll be pushing, CUDA or their own standard? Microsoft will be pushing it's own thing with DX11 down the road. Adobe recently took advantage of Nvidia's CUDA and ATI's Stream if I'm not mistaken... but do you think they'll want to keep on making 2 versions of their products when other people are pushing for a unified standard? I guess this is all moot anyways then... if AMD responding to an Nvidia announcement for a reporter will guarantee success for Physx then surely the grandstanding Nvidia took part in vs. Intel will have Larrabee burying all competition.
At the end of the day people may not like what this guy is saying, why, or how, but it's true. AMD is not going to support Nvidia's proprietary APIs (and why the hell would they?), and with out that support other companies will have less incentive to get on board unless Nvidia provides it. That requires either a superior product or probably cash incentive. Now realisticly.. when the immediate alternatives to Nvidia's systems seem to be OpenCL (Apple - but open), DirectX 11 (Microsoft), and Havok (Intel), do you think these other standards won't have the resources behind them to provide both those things moreso than Nvidia? If you were in AMD's shoes who would you side with? They could do all but seriously... why? It'd just confuse other efforts and probably waste their own resources, and for what? To better prop up the competition that they can beat? So they can claim some moral high ground when they recall how Nv made DX10.1 completely moot?
As a matter of interest, NVidia's Market share has dropped by quite a bit this year, but is still has a much bigger market share than AMD/ATI, this is quite an interesting little piece.....NVidia actually admitting that they were caught "on the hop" by AMD, similar things have popped up around the web on the same subject but this seems to quite nicely bring all the angles together...........
www.crn.com/it-channel/212001134