Friday, September 1st 2017

On AMD's Raja Koduri RX Vega Tweetstorm

In what is usually described as a tweetstorm, AMD's RTG leader Raja Koduri weighed in on AMD's RX Vega reception and perception from both the public and reviewers. There are some interesting tidbits there; namely, AMD's option of setting the RX vega parts at frequencies and voltages outside the optimal curve for power/performance ratios, in a bid to increase attractiveness towards the performance/$ crowds.

However, it can be said that if AMD had done otherwise, neither gamers nor reviewers would have been impressed with cards that potentially delivered less performance than their NVIDIA counterparts, while consuming more power all the same (even if consuming significantly less wattage). At the rated MSRP (and that's a whole new discussion), this RTG decision was the best one towards increasing attractiveness of RX Vega offerings. However, Raja Koduri does stress Vega's dynamic performance/watt ratios, due to the usage of specially defined power profiles.
To our forum-walkers: this piece is marked as an editorial

Raja also touched an interesting subject; namely, that "Folks doing perf/mm2 comparisons need to account for Vega10 features competing with 3 different competitive SOCs (GP100, GP102 and GP104)". This is interesting, and one of the points we have stressed here on TPU: AMD's Vega architecture makes great strides towards being the architecture AMD needs for its server/AI/computing needs, but it's a far cry from what AMD gamers deserve. it's always a thin line to be threaded by chip designers on which features and workloads to implement/improve upon with new architectures. AMD, due to its lack of a high-performance graphics chip, was in dire need of a solution that addressed both the high-performance gaming market and the server/computing market where the highest profits are to be made.
This round, the company decided to focus its developments more towards the server side of the equation - as Raja Koduri himself puts it relative to Infinity Fabric: "Infinity fabric on Vega is optimized for server. It's a very scalable fabric and you will see consumer optimized versions of it in future." Likely, this is Raja's way of telling us to expect MCMs (Multi-Chip-Modules) supported by AMD's Infinity Fabric, much as we see with Ryzen. This design philosophy allows companies to design smaller, more scalable and cheaper chips, which are more resistant to yield issues, and effectively improve every metric, from performance, to power and yield, compared to a monolithic design. NVIDIA themselves have said that MCM chip design is the future, and this little tidbit by Raja could be a pointer towards AMD Radeon's future in that department.
That Infinity Fabric is optimized for servers is a reflection of the entire Vega architecture; AMD's RX Vega may be gaming-oriented graphics cards, but most of the architecture improvements aren't capable of being tapped by already-existent gaming workloads. AMD tooled its architecture for a professional/computing push, in a bid to fight NVIDIA's ever-increasing entrenchment in that market segment, and it shows on the architecture. Does this equate to disappointing (current) gaming performance? It does. And it equates to especially disappointing price/performance ratios with the current supply/pricing situation, which the company still hasn't clearly and forcefully defined. Not that they can, anyway - it's not like AMD can point the finger towards the distributors and retailers that carry their products with no repercussion, now can they?
Raja Koduri also addressed the current mining/gamer arguments, but in a slightly deflective way: in truth, every sale counts as a sale for the company in terms of profit (and/or loss, since rumors abound that AMD is losing up to $100 with every RX Vega graphics card that is being sold at MSRP). An argument can be made that AMD's retreating market share on the PC graphics card market would be a never-ending cycle of lesser developer attention towards architectures that aren't that prominent in their user bases; however, I think that AMD considers they're relatively insulated from that particular issue due to the fact that the company's graphics solutions are embedded into the PS4, PS4 Pro, Xbox One S consoles, as well as the upcoming Xbox One X. Developers code for AMD hardware from the beginning in most games; AMD can certainly put some faith in that as being a factor to tide their GPU performance over until they have the time to properly refine an architecture that shines at both computing and gaming environments. Perhaps when we see AMD's Navi, bolstered by AMD's increased R&D budget due to a (hopefully) successful wager in the professional and computing markets, will we see an AMD that can bring the competition to NVIDIA as well as they have done for Intel.
Sources: Raja Koduri's Twitter, via ETeknix
Add your own comment

131 Comments on On AMD's Raja Koduri RX Vega Tweetstorm

#76
PowerPC
Really the longer I watch them, the more I am convinced AMD is its own greatest competitor.
Just shut up, ok? AMD is our only chance for 4K gaming sooner rather than later. At least they are trying to provide competition against the two top companies in the PC arena. If all we had were Intel and Nvidia, they'd be milking 1440p for the next 10 years. All these two companies care about is money and status quo you call "today" like a lamb. At least AMD tries something new. Trying new things will always lead to some failure, but don't worry, sooner or later it will pay off.
Posted on Reply
#77
springs113
Vayra86i7 7700K in Netherlands comes in at around 315-329 EUR, not sure where you're pulling the 200 EUR number from, they've never been that cheap, that's the i5-K
i7 5775c hovers around the 360 EUR price point

Cheaper boards make up the difference, and cheaper DDR3 RAMs make the Broadwell build actually a bit cheaper, or you can use that money to buy faster ones. Its really 1=1 across the board, so that's why its still in my mind too :) Another real consideration is that the i7 7700k needs to be pushed to 5Ghz which will increase the cost of proper cooling + little guarantees + delid risk, while a typical 5775c OC @ 4.2 Ghz is very easy to achieve and runs a LOT cooler by default, while being on par in terms of gaming. Total cost of the Broadwell rig will likely be lower, and the OC more feasible.

Really the only thing less good about it is the platform - but then again I'm also fine with just SATA SSDs, they're fast enough tbh and have good price/GB.

Sorry this is really offtopic. :eek:



We are to blame for making AMD suck? Naahh its the other way around. AMD sells us 'tomorrow'. Nvidia and Intel sell us 'today' - and by the time AMD's 'tomorrow' proposition has landed, Nvidia and Intel have moved on to a new 'today'. Of course we have seen launch problems with every company, but with AMD, its a guarantee, and you get the fine wine argument along with that. This combination of drawbacks just doesn't instill confidence in the brand.

Ryzen had a similar issue, and most people called that 'a good launch for AMD'. Go figure! Gaming performance was all over the place, boards weren't ready, RAM was and still is a special breed. Yes yes, new platform and all. But again, these things don't help the brand at all. If its an exception (like Intel's X299 kneejerk response with a completely screwed up product stack), you'll just be surprised at it. If it becomes the rule, many quickly avoid it.

Ryzen could also have been sold so much better - imagine the difference if AMD had been up front about Ryzens' pretty hard 4 Ghz wall, and limited OC potential, but then went on to sell Ryzen on its extreme efficiency and core count advantage below or AT the 4 Ghz wall. They didn't do this explicitly at all, we had to analyze reviews to get there ourselves. I mean, ITS ON THE BOX. The TDPs are very good, especially for a competitor that is reintroducing itself to high end CPUs after a long time, that is and was a massive achievement. They just drop five slides with pricing, architecture and product stacks and some one-liners, and then drop the ball. Every. Single. Fucking. Time. Just releasing one solid product after half a decade does not work to instill confidence in a brand once its been lost. They will have to remain consistent with solid releases and SELL those releases to the public, and Vega's release doesn't help this at all.

Really the longer I watch them, the more I am convinced AMD is its own greatest competitor.
While i agree with some of your points, we are still part of the blame, the 4800/5800 cards were way better than the nvidia counterparts and cheaper. What did us consumers do, still bought nvidia. AMD never sold us an overclockers dream. People with high hopes caused their own let down. We see it here in our own forum, Ryzen decimates just about the whole i7 line up ranging from $300-700 yet ppl in these same forums still wouldn't support AMD to save their own lives. If consumers want to pay $1700 for a 10 core processor then so be it...I'm out. I will not support that 1 bit. If AMD had the same budget as Nv/Intel I'm quite sure they'd pull off similar feats. You guys are failing to realize that this "lil" company is taking on 2 big giants at the same time with nowhere those companies money at hand.

At the end of the day is Vega hot? Yes. Is it competing with its target (1070/1080)? Yes. Despite all the issues/controversy surrounding Vega is it selling? Yes.
Is Ryzen selling? Yes.

It's a consumers market, we determine a products outcome. People flocked for the original titan @$1000 guess what Nvidia did? Increase the new one to $1200. Did ppl still buy it? Yup.
Posted on Reply
#78
NeoGalaxy
neko77025seems overpriced too me .. /pass ... will be getting A 1080 ti in A few months .
Better get it right now, since the price will increase with at least 30%.
Posted on Reply
#79
EarthDog
Decimates intel with more cores for your dollar...3 years from now, this situation would be awesome. Too bad 4c/8t is perfectly fine for those next 3 years. Some prefer paying a premium for the faster ipc and better overclocking chip.

Not that it makes it much better, but the 7900x is $999, not $1700 ;)
Posted on Reply
#80
NeoGalaxy
If I were Raja Koduri all I would have to say would be: "I am sorry. I've written my resignation" Have a nice day".
Posted on Reply
#81
cucker tarlson
You must be hallucinating. Vega is a compute monster. It's just not good as good as geforce for gaming enthusiasts.
Posted on Reply
#82
RejZoR
EarthDogDecimates intel with more cores for your dollar...3 years from now, this situation would be awesome. Too bad 4c/8t is perfectly fine for those next 3 years. Some prefer paying a premium for the faster ipc and better overclocking chip.

Not that it makes it much better, but the 7900x is $999, not $1700 ;)
So is a Threadripper 1950X with 16 cores and 32 threads...
Posted on Reply
#83
efikkan
FordGT90ConceptVega has about 10% higher IPC than Fiji with some 50% higher clockspeed to boot.
I guess you mean theoretically. In gaming Vega performs worse per clock than Fiji.
(Analogy: theoretically Ryzen has ~50% higher "IPC" than Skylake, but in real life Skylake IPC is much higher)
Posted on Reply
#84
Frick
Fishfaced Nincompoop
Vayra86i7 7700K in Netherlands comes in at around 315-329 EUR, not sure where you're pulling the 200 EUR number from, they've never been that cheap, that's the i5-K
i7 5775c hovers around the 360 EUR price point
I meant used Haswell i7Ks.
Posted on Reply
#85
cucker tarlson
Not many ppl sell their 5775c in the first place. I know I won't sell mine in 2-3 years when I finally upgrade to moar cores. I'll use it for a console-killer HTPC to hook up to my living room TV along with downvolted GTX 1080.
Posted on Reply
#86
cadaveca
My name is Dave
evernessince"On top of all that, RTG locked Vega's BIOS. Meaning there will be no way to implement any actual BIOS based modifications. RX Vega is a failure no matter what way you spin the story."

Lol, no

www.techpowerup.com/236632/amd-rx-vega-56-to-vega-64-bios-flash-no-unlocked-shaders-improved-performance
forum.ethereum.org/discussion/15024/hows-it-hashin-vega-people

It hasn't even been very long and people can easily flash their Vega BIOS.

There were plenty of other things you could have shit on Vega for but you literally choose non-issues.
Flashing an unmodded BIOS is not modding the BIOS. :p Driver verifies that BIOS has valid checksum at boot. Modding Linux to make it work doesn't actually make it work...
Posted on Reply
#87
xkm1948
cadaveca[QUOTE="evernessince, post: 3719683, member: 165920
"On top of all that, RTG locked Vega's BIOS. Meaning there will be no way to implement any actual BIOS based modifications. RX Vega is a failure no matter what way you spin the story."

Lol, no

www.techpowerup.com/236632/amd-rx-vega-56-to-vega-64-bios-flash-no-unlocked-shaders-improved-performance
forum.ethereum.org/discussion/15024/hows-it-hashin-vega-people

It hasn't even been very long and people can easily flash their Vega BIOS.

There were plenty of other things you could have shit on Vega for but you literally choose non-issues.
Flashing an unmodded BIOS is not modding the BIOS. :p Driver verifies that BIOS has valid checksum at boot. Modding Linux to make it work doesn't actually make it work...[/QUOTE]


Don't burst his imagination bubble goat!
Posted on Reply
#88
Th3pwn3r
Vayra86Another real consideration is that the i7 7700k needs to be pushed to 5Ghz which will increase the cost of proper cooling + little guarantees + delid risk, while a typical 5775c OC @ 4.2 Ghz is very easy to achieve and runs a LOT cooler by default, while being on par in terms of gaming. Total cost of the Broadwell rig will likely be lower, and the OC more feasible.
Calling the i7 7700k the Vega of Intel is really way out of left field and makes no sense. The 7700k is actually a huge bang for the buck, had Intel not been retarded with the TIM they used under the lid it would have easily been the best processor for the money in a long time in my opinion. Also, I'm curious why you have it in your head that the 7700k "needs to be pushed to 5ghz"? Plenty of people run them at standard clocks and they work just fine. Anyhow, I like you said, this is off topic and I have no idea why you chose to discuss the 7700k really and try to smear it. Right now they're selling for $299.99 at Fry's Electronics. I bought mine for a bit over $300 almost a year ago?
Posted on Reply
#89
dyonoctis
People keep buying Nvidia/Intel, because AMD is still being seen as the budget solution, it doesn't matter if the pentium 4 was hot and slow for such a high GHZ, or fermi hot and power hungry, those were still sold by the companies being seen as market leader. This effect is even worse when the buyer isn't a tech nerd.

For the average joe, the P4 with an higher clock had to be better, architecture optimization isn't something they are aware of. They also don't care/know about the fact that AMD was first on 64bit.
Intel also got a nice little eco-system going on :

Thunderbolt being intel locked means that apple won't go full AMD ever, and also means giving up on that on any ryzen based laptop, so no egpu with a Ryzen system.



As for nvidia, they got an habit of selling the fastest thing on earth wich do a lot for brand fidelity, and brand equity (geforce 256, GTX 8800, Titan) and when they are launching new "toys" they are doing it right:
  • 3D vision was better, easier to set up, 3D with AMD was a mess.
  • Phys-x may be a gimmick, but it's still a little extra nice to have,and with all the succesfull batman games having it, it was another reason to get an Nvidia card.
  • G-sync was first and heavily marketed, freesync came later and have poor marketing,
  • CUDA is easier to work with, the developpers get better support for every platform, meanwhile open cl can be really troublesome, some of the devellopers working on opencl cycle renderer for mac osx gave up out of disgust, because they couldn't get any support to fix an issue, open cl on mac osx is broken, and even tho apple was the creator of open cl they didn't care.
The fine wine argument isn't something marketable, and if they ditch GCN it may not be true anymore. When the GTX 980/ti were launched, it was a fact that they were the fastest gpu, that's how you are selling product to people, you don't say: in 2 years our product will go toes to toes with the old fastest gpu in the market.

AMD may have been the first to give HBM to the consummers, the fact that it wasn't the fastest thing on earth failed to impress part of the consummers. People in general are not reasonable, fully rational buyers, they like dreams, and buying something from the company selling the fastest gpu or the makers of the 1st cpu sounds better than buying the stuff from the budget company. Being seen as a premium brand does a lot, even if the client doesn't buy the high end product, they still feel like they are a part of it.



It wasn't long ago that i had to show some benchmark to a guy who was making a meh face when i mentionned that the new AMD cpu were good product. People don't know that AMD is back in the game, and all the high end laptop still rocking an Intel cpu/ Nvidia gpu won't help.
Posted on Reply
#90
BiggieShady
FordGT90ConceptThat's not a weakness.
Huh, where is this notion of weakness coming from? ... it's far from weak, just not as efficient ... and if you mean weakness as a shortcoming, sounds better if you call it a trade off
FordGT90ConceptBut NVIDIA doesn't really give us any details on what's inside the cores except the obvious.
I don't think anything is missing, all the parts are there although naming convention differs.
I guess you are wondering, for example, where is the branch predictor in nvidia arch ... there isn't one, warp is executing both parts of conditional statement (first "if" block and "else" block after) and portion of cuda cores sleep through the "if" part and other portion sleep through the "else" part. Brute forcing branching in shader code like this seems inefficient and a weakness but it's a trade off for simple and robust parallel design that saves die space and wins efficiency in total.
Posted on Reply
#91
Vya Domus
BiggieShadyBasically it's a low-latency throughput favoring design that is wasteful and inflexible.
I don't know if that's true , higher instruction throughput is exactly what you want if you desire efficiency.

Actually that's exactly what you want in general with a GPU where you need to keep the core fed. The fact that Vega wins a lot more from memory bandwidth suggests there is nothing wrong with the way instructions are handled. Which isn't surprising at all , one of the biggest bottlenecks inside any of these GPUs was and still is memory bandwidth/latency. AMD should have kept the 4096-bit memory bus.
BiggieShadyI don't think anything is missing, all the parts are there although naming convention differs.
I guess you are wondering, for example, where is the branch predictor in nvidia arch ... there isn't one, warp is executing both parts of conditional statement (first "if" block and "else" block after) and portion of cuda cores sleep through the "if" part and other portion sleep through the "else" part. Brute forcing branching in shader code like this seems inefficient and a weakness but it's a trade off for simple and robust parallel design that saves die space and wins efficiency in total.
Lack of branch prediction means more burden on compilers/drivers and more CPU overhead in general at run-time. And I would say that's not robust at all , in fact it's the reason why until Pascal they have been struggling with async. When they said Maxwell can do async but it just needs enabling , they were BSing. What they meant is that they can only do that through the help of the driver but that would have ended up being ridiculously inefficient and would have hurt performance most likely and that's why we never saw "a magic async enabling driver".

Nvidia offloaded a lot of the control logic on the driver side of things , meanwhile AMD relies almost entirely on the silicon itself to properly dispatch work. That should have been an advantage if anything for AMD , however Nvidia struck gold when they managed to support multi threaded draw calls automatically through their drivers with DX11 and AMD for some reason couldn't.

And this is why DX12/Vulkan was supposed to help AMD since now draw calls are handled exclusively by the programmer and Nvidia's drivers can't do jack now. But all we got were lousy implementations since it requires a lot of work. In fact it can hurt performance because even the few DX11 optimizations AMD had going are being nullified , see Rise of the Tomb Raider.

In the end my point is that the design language AMD uses for it's architecture is perfectly fine , it's just that they have to deal with an antiquated software landscape that isn't on the same page with them.
Posted on Reply
#92
Anymal
slehmannGot my Vega 64 in the first week they were sold. Waited 2 weeks more for the pre-ordered EK watercooler. Now its at >1600Mhz all the time and >50% faster than my Fury X before and reaches hardly 50°C. Im satisfied with the increase of gaming performance at 1440p so far. So, tnx AMD, you delivered.
1080ti almost doubles it, so 100% if you compare custom 1080ti vs Fury X. Satisfied? Maybe with temperature.
Posted on Reply
#93
efikkan
Vya DomusI don't know if that's true , higher instruction throughput is exactly what you want if you desire efficiency.

Actually that's exactly what you want in general with a GPU where you need to keep the core fed. The fact that Vega wins a lot more from memory bandwidth suggests there is nothing wrong with the way instructions are handled. Which isn't surprising at all , one of the biggest bottlenecks inside any of these GPUs was and still is memory bandwidth/latency. AMD should have kept the 4096-bit memory bus.
Nonesene.
Memory bandwidth is not holding Vega back, it has the same bandwidth as GTX 1080 Ti. The problem is the GPU's scheduler and a longer pipeline.
Vya DomusLack of branch prediction means more burden on compilers/drivers and more CPU overhead in general at run-time…
Nonesene.
Execution of shader code and low level scheduling is done in the GPU, not the CPU. Conditionals in the shader code doesn't stress the compiler or the driver. Start by learning the basics first.
Vya DomusNvidia offloaded a lot of the control logic on the driver side of things , meanwhile AMD relies almost entirely on the silicon itself to properly dispatch work…
Nonsense.
The drivers translate a queue of API calls into native operations, the GPU schedules clusters, analyzes dependencies, etc. in real time.
Vya DomusAnd this is why DX12/Vulkan was supposed to help AMD since now draw calls are handled exclusively by the programmer and Nvidia's drivers can't do jack now.
Nonsense.
If you knew anything about graphics programming, you'd know the programmer has always controlled the draw calls.
Posted on Reply
#94
FordGT90Concept
"I go fast!1!11!1!"
BiggieShadyI don't think anything is missing, all the parts are there although naming convention differs.
Case in point: how many SIMD's does each CUDA core have? How many scalar operations can each core perform per clock?

Hmm... www.tomshardware.com/reviews/nvidia-cuda-gpu,1954-7.html

If I'm following this correctly. AMD has compute unit flexibility where CUDA only has flexibility on the warp level (64-512 threads). If CUDA is told to execute just a single SIMD operation on behalf of CUDA, 98.4-99.8% of the block will result in idle hardware. I suspect GCN is the opposite: the ACE will reserve a single compute unit to deal with that task while the rest of the GPU continues with other work.

The design philosophy is fundamentally opposite: GCN is about GPGPU where CUDA is about GPU with the ability to switch context to GPGPU. GCN demands saturation (like a CPU) where CUDA demands strict control (like a GPU).

Which is better going forward? Which should computers 10 years from now have? I think GCN because DirectPhysics is coming. We're finally going to have a reason to get serious about physics as a gameplay mechanic and NVIDIA can't get serious because of the massive penalty in framerate CUDA incurs when context switching. It makes far more sense for physics to occupy idle compute resources than disrupting the render wave front. CUDA, as it is designed now, will hold engines back from using physics as they should.
Posted on Reply
#95
Vayra86
Th3pwn3rCalling the i7 7700k the Vega of Intel is really way out of left field and makes no sense. The 7700k is actually a huge bang for the buck, had Intel not been retarded with the TIM they used under the lid it would have easily been the best processor for the money in a long time in my opinion. Also, I'm curious why you have it in your head that the 7700k "needs to be pushed to 5ghz"? Plenty of people run them at standard clocks and they work just fine. Anyhow, I like you said, this is off topic and I have no idea why you chose to discuss the 7700k really and try to smear it. Right now they're selling for $299.99 at Fry's Electronics. I bought mine for a bit over $300 almost a year ago?
There is nothing wrong with the CPU, but what I am looking for in my next upgrade is the most solid CPU for gaming - and only that. Have you seen the comparisons of OCd 7700k and 5775c next to each other, and their temps, and vCores? Its quite a gap and it shows that Intel has pushed core to/past its limit. The fact is, the 7700k needs some serious work for that similar-performing OC too.
Posted on Reply
#96
springs113
@efikkan Vega is definitely bandwidth starved. Increasing the memory clocks give you a better boat with lower power draw than doing so with the core.
Posted on Reply
#97
BiggieShady
FordGT90ConceptCase in point: how many SIMD's does each CUDA core have? How many scalar operations can each core perform per clock?
Dude, cuda core is much more lightweight than that and it's completely scalar ... it's an execution unit that can retire 1 single precision floating point operation per clock (2 for a fused-multiply-add).
So basically it takes several of them to do a job of single SIMD unit in GCN ... this is what I meant by finer granularity
FordGT90ConceptIf CUDA is told to execute just a single SIMD operation on behalf of CUDA, 98.4-99.8% of the block will result in idle hardware.
Single SIMD operation still has multiple data, so it would saturate much more than that ... what you are describing is using SIMD as SISD (having data array consist of only 1 element), and in that context, yes, GCN would handle it more gracefully.
Nvidia's model is actually SIMT and its efficiency depends on having enough concurrent warps - even when those warps are not saturating SM enough by themselves, if there is enough of them for scheduler to choose from, all is well (good read yosefk.com/blog/simd-simt-smt-parallelism-in-nvidia-gpus.html)
FordGT90ConceptThe design philosophy is fundamentally opposite: GCN is about GPGPU where CUDA is about GPU with the ability to switch context to GPGPU. GCN demands saturation (like a CPU) where CUDA demands strict control (like a GPU).
Ditto. Nvidia kinda expects copious amount of SIMD/SIMT parallelism in their workload, because they are making GPUs.
With GPGPU it's always the question "How much general purpose you want it to be?"
FordGT90ConceptWhich is better going forward? Which should computers 10 years from now have? I think GCN because DirectPhysics is coming. We're finally going to have a reason to get serious about physics as a gameplay mechanic and NVIDIA can't get serious because of the massive penalty in framerate CUDA incurs when context switching. It makes far more sense for physics to occupy idle compute resources than disrupting the render wave front. CUDA, as it is designed now, will hold engines back from using physics as they should.
Yeah well, physics solvers are extremely dynamic in their load ... one frame nothing is happening, the next frame one collision sets hundreds of objects in motion. If you simply must update all positions by the end of the frame, and you do, async is the way to go. I guess that's why cuda physx is exclusively of a simd/simt kind (huge number of elements behaving by the same algorithm - fluid, hair, grass). Too bad we don't have some alpha version previewed somewhere by now since async compute is a thing.
Posted on Reply
#98
FordGT90Concept
"I go fast!1!11!1!"
Intel abandoned Havok after they bought it so Microsoft has a lot of work to do before they debut it. Not just integrated into DirectX 12 but also make it competitive in the market with PhysX which has been getting constant updates for the last...almost decade. Quite depressing but it's something Microsoft can't afford to pass up.

It's going to take GPGPU to make physics mean something in games. Specifically, they need to prioritize the compute tasks and balance the GPU work with it. PhysX was always built off the foundation of having a separate card handling it. DirectPhysics can't do that.


Edit: trademarks.justia.com/871/86/directphysics-87186880.html
DIRECTPHYSICS/DIRECT PHYSICSComputer software development tools; Computer software development tools for game programs, three-dimensional application programs and other application programs; Computer software, namely, software development kits (SDK); Computer software for use in the design, development and execution of game programs, three-dimensional interactive application programs, and other application programs; Computer software, namely, computer operating programs for games, three-dimensional interactive application programs, and other application programs; Computer software for developing and interfacing with virtual reality software, hardware and peripherals; Computer software, namely, 3D content and motion content rendering software for use in producing videos and animation
Posted on Reply
#99
RejZoR
But I see a bright future if Havok sometime soon becomes DirectPhysics as part of DirectX, that's the moment we'll see huge leap in game physics. It won't be just ragdolls anymore. It'll actually be able to affect gameplay.
Posted on Reply
#100
InVasMani
Vega also has less bandwidth than Fiji sadly which is probably one of the bigger things holding it back currently I think with a die shrink a higher core clock speed and a memory bus that's more like 640 bit it's successor will be in good shape provided Nvidia doesn't go crazy on R&D and make it look like trash comparatively which could happen I suppose though I don't there there is major demand for a $2000 GPU for gaming as it is GPU's are too damn expensive.
RejZoRBut I see a bright future if Havok sometime soon becomes DirectPhysics as part of DirectX, that's the moment we'll see huge leap in game physics. It won't be just ragdolls anymore. It'll actually be able to affect gameplay.
I imagine it's quite hard to do complex physics in a game where it's already pushing the limits and taxing the system to death just rendering games at the resolutions and image quality settings we as game players often demand.
Posted on Reply
Add your own comment
Nov 18th, 2024 23:30 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts