• Welcome to TechPowerUp Forums, Guest! Please check out our forum guidelines for info related to our community.
  • The forums have been upgraded with support for dark mode. By default it will follow the setting on your system/browser. You may override it by scrolling to the end of the page and clicking the gears icon.

Editorial On AMD's Raja Koduri RX Vega Tweetstorm

I think I've said this, but Poland seems to be the only place where Broadwell a) exists and b) wasn't/isn't €100+ more than an i7K. In the vast majority of the market they were non existent, and getting an LGA1150 platform these days is just dumb, unless you get a good deal on an old system and somehow get a hold of the 5775c for less than the €200ish the i7K's usually go for these days.

i7 7700K in Netherlands comes in at around 315-329 EUR, not sure where you're pulling the 200 EUR number from, they've never been that cheap, that's the i5-K
i7 5775c hovers around the 360 EUR price point

Cheaper boards make up the difference, and cheaper DDR3 RAMs make the Broadwell build actually a bit cheaper, or you can use that money to buy faster ones. Its really 1=1 across the board, so that's why its still in my mind too :) Another real consideration is that the i7 7700k needs to be pushed to 5Ghz which will increase the cost of proper cooling + little guarantees + delid risk, while a typical 5775c OC @ 4.2 Ghz is very easy to achieve and runs a LOT cooler by default, while being on par in terms of gaming. Total cost of the Broadwell rig will likely be lower, and the OC more feasible.

Really the only thing less good about it is the platform - but then again I'm also fine with just SATA SSDs, they're fast enough tbh and have good price/GB.

Sorry this is really offtopic. :o

I'll give them til' Navi to see, as it seems Navi is the RTGs Ryzen. Vega is an awesome card. I was able to get 2, both at the $500 price tag. It is smooth on my non freesync panel. I'm currently undergoing a system overhaul on both my pcs and can't bench or test anything which sucks. But in due time i think Vega will be at least 10%faster given or take 3 to 6 months. Sadly AMD keeps repeating this cycle, i thought they learnt from the Polaris release and even so Ryzen. We all knew something was up when Vega was suppose to be released right after TR but the reviewers got the cards like literally 3 days before release date. Then being told to focus on the 56 and not the 64 being released first. Shameful on AMDs part. I just hope they get their act together sooner than later. I (we all) need more/better competition. I'll rock out with Vega until Navi shows itself though. Another sad part is because of AMDs marketing or lack there of even when they have the better option we as consumers don't support them. To some degree i feel us add commanders are partly to blame for this.

The trend with Ryzen is fundamentally breaking that mold but even then i think people are more tired of Intels games more than there actual appeal of Ryzen. Odd way to look at things and maybe even small minded of me, but i can't not think about the Athlon Era of cpus. AMD clearly had the better product yet ppl willingly bought Intel. Same goes for the 5870/5970 Era of gpus. They were the best at just about every level yet consumers still bought Nvidia.

We are to blame for making AMD suck? Naahh its the other way around. AMD sells us 'tomorrow'. Nvidia and Intel sell us 'today' - and by the time AMD's 'tomorrow' proposition has landed, Nvidia and Intel have moved on to a new 'today'. Of course we have seen launch problems with every company, but with AMD, its a guarantee, and you get the fine wine argument along with that. This combination of drawbacks just doesn't instill confidence in the brand.

Ryzen had a similar issue, and most people called that 'a good launch for AMD'. Go figure! Gaming performance was all over the place, boards weren't ready, RAM was and still is a special breed. Yes yes, new platform and all. But again, these things don't help the brand at all. If its an exception (like Intel's X299 kneejerk response with a completely screwed up product stack), you'll just be surprised at it. If it becomes the rule, many quickly avoid it.

Ryzen could also have been sold so much better - imagine the difference if AMD had been up front about Ryzens' pretty hard 4 Ghz wall, and limited OC potential, but then went on to sell Ryzen on its extreme efficiency and core count advantage below or AT the 4 Ghz wall. They didn't do this explicitly at all, we had to analyze reviews to get there ourselves. I mean, ITS ON THE BOX. The TDPs are very good, especially for a competitor that is reintroducing itself to high end CPUs after a long time, that is and was a massive achievement. They just drop five slides with pricing, architecture and product stacks and some one-liners, and then drop the ball. Every. Single. Fucking. Time. Just releasing one solid product after half a decade does not work to instill confidence in a brand once its been lost. They will have to remain consistent with solid releases and SELL those releases to the public, and Vega's release doesn't help this at all.

Really the longer I watch them, the more I am convinced AMD is its own greatest competitor.
 
Last edited:
Really the longer I watch them, the more I am convinced AMD is its own greatest competitor.
Just shut up, ok? AMD is our only chance for 4K gaming sooner rather than later. At least they are trying to provide competition against the two top companies in the PC arena. If all we had were Intel and Nvidia, they'd be milking 1440p for the next 10 years. All these two companies care about is money and status quo you call "today" like a lamb. At least AMD tries something new. Trying new things will always lead to some failure, but don't worry, sooner or later it will pay off.
 
i7 7700K in Netherlands comes in at around 315-329 EUR, not sure where you're pulling the 200 EUR number from, they've never been that cheap, that's the i5-K
i7 5775c hovers around the 360 EUR price point

Cheaper boards make up the difference, and cheaper DDR3 RAMs make the Broadwell build actually a bit cheaper, or you can use that money to buy faster ones. Its really 1=1 across the board, so that's why its still in my mind too :) Another real consideration is that the i7 7700k needs to be pushed to 5Ghz which will increase the cost of proper cooling + little guarantees + delid risk, while a typical 5775c OC @ 4.2 Ghz is very easy to achieve and runs a LOT cooler by default, while being on par in terms of gaming. Total cost of the Broadwell rig will likely be lower, and the OC more feasible.

Really the only thing less good about it is the platform - but then again I'm also fine with just SATA SSDs, they're fast enough tbh and have good price/GB.

Sorry this is really offtopic. :eek:



We are to blame for making AMD suck? Naahh its the other way around. AMD sells us 'tomorrow'. Nvidia and Intel sell us 'today' - and by the time AMD's 'tomorrow' proposition has landed, Nvidia and Intel have moved on to a new 'today'. Of course we have seen launch problems with every company, but with AMD, its a guarantee, and you get the fine wine argument along with that. This combination of drawbacks just doesn't instill confidence in the brand.

Ryzen had a similar issue, and most people called that 'a good launch for AMD'. Go figure! Gaming performance was all over the place, boards weren't ready, RAM was and still is a special breed. Yes yes, new platform and all. But again, these things don't help the brand at all. If its an exception (like Intel's X299 kneejerk response with a completely screwed up product stack), you'll just be surprised at it. If it becomes the rule, many quickly avoid it.

Ryzen could also have been sold so much better - imagine the difference if AMD had been up front about Ryzens' pretty hard 4 Ghz wall, and limited OC potential, but then went on to sell Ryzen on its extreme efficiency and core count advantage below or AT the 4 Ghz wall. They didn't do this explicitly at all, we had to analyze reviews to get there ourselves. I mean, ITS ON THE BOX. The TDPs are very good, especially for a competitor that is reintroducing itself to high end CPUs after a long time, that is and was a massive achievement. They just drop five slides with pricing, architecture and product stacks and some one-liners, and then drop the ball. Every. Single. Fucking. Time. Just releasing one solid product after half a decade does not work to instill confidence in a brand once its been lost. They will have to remain consistent with solid releases and SELL those releases to the public, and Vega's release doesn't help this at all.

Really the longer I watch them, the more I am convinced AMD is its own greatest competitor.
While i agree with some of your points, we are still part of the blame, the 4800/5800 cards were way better than the nvidia counterparts and cheaper. What did us consumers do, still bought nvidia. AMD never sold us an overclockers dream. People with high hopes caused their own let down. We see it here in our own forum, Ryzen decimates just about the whole i7 line up ranging from $300-700 yet ppl in these same forums still wouldn't support AMD to save their own lives. If consumers want to pay $1700 for a 10 core processor then so be it...I'm out. I will not support that 1 bit. If AMD had the same budget as Nv/Intel I'm quite sure they'd pull off similar feats. You guys are failing to realize that this "lil" company is taking on 2 big giants at the same time with nowhere those companies money at hand.

At the end of the day is Vega hot? Yes. Is it competing with its target (1070/1080)? Yes. Despite all the issues/controversy surrounding Vega is it selling? Yes.
Is Ryzen selling? Yes.

It's a consumers market, we determine a products outcome. People flocked for the original titan @$1000 guess what Nvidia did? Increase the new one to $1200. Did ppl still buy it? Yup.
 
Last edited:
seems overpriced too me .. /pass ... will be getting A 1080 ti in A few months .

Better get it right now, since the price will increase with at least 30%.
 
Decimates intel with more cores for your dollar...3 years from now, this situation would be awesome. Too bad 4c/8t is perfectly fine for those next 3 years. Some prefer paying a premium for the faster ipc and better overclocking chip.

Not that it makes it much better, but the 7900x is $999, not $1700 ;)
 
If I were Raja Koduri all I would have to say would be: "I am sorry. I've written my resignation" Have a nice day".
 
You must be hallucinating. Vega is a compute monster. It's just not good as good as geforce for gaming enthusiasts.
 
Decimates intel with more cores for your dollar...3 years from now, this situation would be awesome. Too bad 4c/8t is perfectly fine for those next 3 years. Some prefer paying a premium for the faster ipc and better overclocking chip.

Not that it makes it much better, but the 7900x is $999, not $1700 ;)

So is a Threadripper 1950X with 16 cores and 32 threads...
 
Vega has about 10% higher IPC than Fiji with some 50% higher clockspeed to boot.
I guess you mean theoretically. In gaming Vega performs worse per clock than Fiji.
(Analogy: theoretically Ryzen has ~50% higher "IPC" than Skylake, but in real life Skylake IPC is much higher)
 
i7 7700K in Netherlands comes in at around 315-329 EUR, not sure where you're pulling the 200 EUR number from, they've never been that cheap, that's the i5-K
i7 5775c hovers around the 360 EUR price point

I meant used Haswell i7Ks.
 
Not many ppl sell their 5775c in the first place. I know I won't sell mine in 2-3 years when I finally upgrade to moar cores. I'll use it for a console-killer HTPC to hook up to my living room TV along with downvolted GTX 1080.
 
evernessince said:
"On top of all that, RTG locked Vega's BIOS. Meaning there will be no way to implement any actual BIOS based modifications. RX Vega is a failure no matter what way you spin the story."

Lol, no

https://www.techpowerup.com/236632/...lash-no-unlocked-shaders-improved-performance
https://forum.ethereum.org/discussion/15024/hows-it-hashin-vega-people

It hasn't even been very long and people can easily flash their Vega BIOS.

There were plenty of other things you could have shit on Vega for but you literally choose non-issues.


Flashing an unmodded BIOS is not modding the BIOS. :p Driver verifies that BIOS has valid checksum at boot. Modding Linux to make it work doesn't actually make it work...
 
Last edited:
[QUOTE="evernessince, post: 3719683, member: 165920
"On top of all that, RTG locked Vega's BIOS. Meaning there will be no way to implement any actual BIOS based modifications. RX Vega is a failure no matter what way you spin the story."

Lol, no

https://www.techpowerup.com/236632/...lash-no-unlocked-shaders-improved-performance
https://forum.ethereum.org/discussion/15024/hows-it-hashin-vega-people

It hasn't even been very long and people can easily flash their Vega BIOS.

There were plenty of other things you could have shit on Vega for but you literally choose non-issues.


Flashing an unmodded BIOS is not modding the BIOS. :p Driver verifies that BIOS has valid checksum at boot. Modding Linux to make it work doesn't actually make it work...[/QUOTE]


Don't burst his imagination bubble goat!
 
Another real consideration is that the i7 7700k needs to be pushed to 5Ghz which will increase the cost of proper cooling + little guarantees + delid risk, while a typical 5775c OC @ 4.2 Ghz is very easy to achieve and runs a LOT cooler by default, while being on par in terms of gaming. Total cost of the Broadwell rig will likely be lower, and the OC more feasible.

Calling the i7 7700k the Vega of Intel is really way out of left field and makes no sense. The 7700k is actually a huge bang for the buck, had Intel not been retarded with the TIM they used under the lid it would have easily been the best processor for the money in a long time in my opinion. Also, I'm curious why you have it in your head that the 7700k "needs to be pushed to 5ghz"? Plenty of people run them at standard clocks and they work just fine. Anyhow, I like you said, this is off topic and I have no idea why you chose to discuss the 7700k really and try to smear it. Right now they're selling for $299.99 at Fry's Electronics. I bought mine for a bit over $300 almost a year ago?
 
People keep buying Nvidia/Intel, because AMD is still being seen as the budget solution, it doesn't matter if the pentium 4 was hot and slow for such a high GHZ, or fermi hot and power hungry, those were still sold by the companies being seen as market leader. This effect is even worse when the buyer isn't a tech nerd.

For the average joe, the P4 with an higher clock had to be better, architecture optimization isn't something they are aware of. They also don't care/know about the fact that AMD was first on 64bit.
Intel also got a nice little eco-system going on :

Thunderbolt being intel locked means that apple won't go full AMD ever, and also means giving up on that on any ryzen based laptop, so no egpu with a Ryzen system.



As for nvidia, they got an habit of selling the fastest thing on earth wich do a lot for brand fidelity, and brand equity (geforce 256, GTX 8800, Titan) and when they are launching new "toys" they are doing it right:
  • 3D vision was better, easier to set up, 3D with AMD was a mess.
  • Phys-x may be a gimmick, but it's still a little extra nice to have,and with all the succesfull batman games having it, it was another reason to get an Nvidia card.
  • G-sync was first and heavily marketed, freesync came later and have poor marketing,
  • CUDA is easier to work with, the developpers get better support for every platform, meanwhile open cl can be really troublesome, some of the devellopers working on opencl cycle renderer for mac osx gave up out of disgust, because they couldn't get any support to fix an issue, open cl on mac osx is broken, and even tho apple was the creator of open cl they didn't care.
The fine wine argument isn't something marketable, and if they ditch GCN it may not be true anymore. When the GTX 980/ti were launched, it was a fact that they were the fastest gpu, that's how you are selling product to people, you don't say: in 2 years our product will go toes to toes with the old fastest gpu in the market.

AMD may have been the first to give HBM to the consummers, the fact that it wasn't the fastest thing on earth failed to impress part of the consummers. People in general are not reasonable, fully rational buyers, they like dreams, and buying something from the company selling the fastest gpu or the makers of the 1st cpu sounds better than buying the stuff from the budget company. Being seen as a premium brand does a lot, even if the client doesn't buy the high end product, they still feel like they are a part of it.



It wasn't long ago that i had to show some benchmark to a guy who was making a meh face when i mentionned that the new AMD cpu were good product. People don't know that AMD is back in the game, and all the high end laptop still rocking an Intel cpu/ Nvidia gpu won't help.
 
That's not a weakness.
Huh, where is this notion of weakness coming from? ... it's far from weak, just not as efficient ... and if you mean weakness as a shortcoming, sounds better if you call it a trade off
But NVIDIA doesn't really give us any details on what's inside the cores except the obvious.
I don't think anything is missing, all the parts are there although naming convention differs.
I guess you are wondering, for example, where is the branch predictor in nvidia arch ... there isn't one, warp is executing both parts of conditional statement (first "if" block and "else" block after) and portion of cuda cores sleep through the "if" part and other portion sleep through the "else" part. Brute forcing branching in shader code like this seems inefficient and a weakness but it's a trade off for simple and robust parallel design that saves die space and wins efficiency in total.
 
Basically it's a low-latency throughput favoring design that is wasteful and inflexible.

I don't know if that's true , higher instruction throughput is exactly what you want if you desire efficiency.

Actually that's exactly what you want in general with a GPU where you need to keep the core fed. The fact that Vega wins a lot more from memory bandwidth suggests there is nothing wrong with the way instructions are handled. Which isn't surprising at all , one of the biggest bottlenecks inside any of these GPUs was and still is memory bandwidth/latency. AMD should have kept the 4096-bit memory bus.

I don't think anything is missing, all the parts are there although naming convention differs.
I guess you are wondering, for example, where is the branch predictor in nvidia arch ... there isn't one, warp is executing both parts of conditional statement (first "if" block and "else" block after) and portion of cuda cores sleep through the "if" part and other portion sleep through the "else" part. Brute forcing branching in shader code like this seems inefficient and a weakness but it's a trade off for simple and robust parallel design that saves die space and wins efficiency in total.


Lack of branch prediction means more burden on compilers/drivers and more CPU overhead in general at run-time. And I would say that's not robust at all , in fact it's the reason why until Pascal they have been struggling with async. When they said Maxwell can do async but it just needs enabling , they were BSing. What they meant is that they can only do that through the help of the driver but that would have ended up being ridiculously inefficient and would have hurt performance most likely and that's why we never saw "a magic async enabling driver".

Nvidia offloaded a lot of the control logic on the driver side of things , meanwhile AMD relies almost entirely on the silicon itself to properly dispatch work. That should have been an advantage if anything for AMD , however Nvidia struck gold when they managed to support multi threaded draw calls automatically through their drivers with DX11 and AMD for some reason couldn't.

And this is why DX12/Vulkan was supposed to help AMD since now draw calls are handled exclusively by the programmer and Nvidia's drivers can't do jack now. But all we got were lousy implementations since it requires a lot of work. In fact it can hurt performance because even the few DX11 optimizations AMD had going are being nullified , see Rise of the Tomb Raider.

In the end my point is that the design language AMD uses for it's architecture is perfectly fine , it's just that they have to deal with an antiquated software landscape that isn't on the same page with them.
 
Last edited:
Got my Vega 64 in the first week they were sold. Waited 2 weeks more for the pre-ordered EK watercooler. Now its at >1600Mhz all the time and >50% faster than my Fury X before and reaches hardly 50°C. Im satisfied with the increase of gaming performance at 1440p so far. So, tnx AMD, you delivered.
1080ti almost doubles it, so 100% if you compare custom 1080ti vs Fury X. Satisfied? Maybe with temperature.
 
I don't know if that's true , higher instruction throughput is exactly what you want if you desire efficiency.

Actually that's exactly what you want in general with a GPU where you need to keep the core fed. The fact that Vega wins a lot more from memory bandwidth suggests there is nothing wrong with the way instructions are handled. Which isn't surprising at all , one of the biggest bottlenecks inside any of these GPUs was and still is memory bandwidth/latency. AMD should have kept the 4096-bit memory bus.
Nonesene.
Memory bandwidth is not holding Vega back, it has the same bandwidth as GTX 1080 Ti. The problem is the GPU's scheduler and a longer pipeline.

Lack of branch prediction means more burden on compilers/drivers and more CPU overhead in general at run-time…
Nonesene.
Execution of shader code and low level scheduling is done in the GPU, not the CPU. Conditionals in the shader code doesn't stress the compiler or the driver. Start by learning the basics first.

Nvidia offloaded a lot of the control logic on the driver side of things , meanwhile AMD relies almost entirely on the silicon itself to properly dispatch work…
Nonsense.
The drivers translate a queue of API calls into native operations, the GPU schedules clusters, analyzes dependencies, etc. in real time.

And this is why DX12/Vulkan was supposed to help AMD since now draw calls are handled exclusively by the programmer and Nvidia's drivers can't do jack now.
Nonsense.
If you knew anything about graphics programming, you'd know the programmer has always controlled the draw calls.
 
I don't think anything is missing, all the parts are there although naming convention differs.
Case in point: how many SIMD's does each CUDA core have? How many scalar operations can each core perform per clock?

Hmm... http://www.tomshardware.com/reviews/nvidia-cuda-gpu,1954-7.html

If I'm following this correctly. AMD has compute unit flexibility where CUDA only has flexibility on the warp level (64-512 threads). If CUDA is told to execute just a single SIMD operation on behalf of CUDA, 98.4-99.8% of the block will result in idle hardware. I suspect GCN is the opposite: the ACE will reserve a single compute unit to deal with that task while the rest of the GPU continues with other work.

The design philosophy is fundamentally opposite: GCN is about GPGPU where CUDA is about GPU with the ability to switch context to GPGPU. GCN demands saturation (like a CPU) where CUDA demands strict control (like a GPU).

Which is better going forward? Which should computers 10 years from now have? I think GCN because DirectPhysics is coming. We're finally going to have a reason to get serious about physics as a gameplay mechanic and NVIDIA can't get serious because of the massive penalty in framerate CUDA incurs when context switching. It makes far more sense for physics to occupy idle compute resources than disrupting the render wave front. CUDA, as it is designed now, will hold engines back from using physics as they should.
 
Calling the i7 7700k the Vega of Intel is really way out of left field and makes no sense. The 7700k is actually a huge bang for the buck, had Intel not been retarded with the TIM they used under the lid it would have easily been the best processor for the money in a long time in my opinion. Also, I'm curious why you have it in your head that the 7700k "needs to be pushed to 5ghz"? Plenty of people run them at standard clocks and they work just fine. Anyhow, I like you said, this is off topic and I have no idea why you chose to discuss the 7700k really and try to smear it. Right now they're selling for $299.99 at Fry's Electronics. I bought mine for a bit over $300 almost a year ago?

There is nothing wrong with the CPU, but what I am looking for in my next upgrade is the most solid CPU for gaming - and only that. Have you seen the comparisons of OCd 7700k and 5775c next to each other, and their temps, and vCores? Its quite a gap and it shows that Intel has pushed core to/past its limit. The fact is, the 7700k needs some serious work for that similar-performing OC too.
 
@efikkan Vega is definitely bandwidth starved. Increasing the memory clocks give you a better boat with lower power draw than doing so with the core.
 
Case in point: how many SIMD's does each CUDA core have? How many scalar operations can each core perform per clock?
Dude, cuda core is much more lightweight than that and it's completely scalar ... it's an execution unit that can retire 1 single precision floating point operation per clock (2 for a fused-multiply-add).
So basically it takes several of them to do a job of single SIMD unit in GCN ... this is what I meant by finer granularity
If CUDA is told to execute just a single SIMD operation on behalf of CUDA, 98.4-99.8% of the block will result in idle hardware.
Single SIMD operation still has multiple data, so it would saturate much more than that ... what you are describing is using SIMD as SISD (having data array consist of only 1 element), and in that context, yes, GCN would handle it more gracefully.
Nvidia's model is actually SIMT and its efficiency depends on having enough concurrent warps - even when those warps are not saturating SM enough by themselves, if there is enough of them for scheduler to choose from, all is well (good read http://yosefk.com/blog/simd-simt-smt-parallelism-in-nvidia-gpus.html)
The design philosophy is fundamentally opposite: GCN is about GPGPU where CUDA is about GPU with the ability to switch context to GPGPU. GCN demands saturation (like a CPU) where CUDA demands strict control (like a GPU).
Ditto. Nvidia kinda expects copious amount of SIMD/SIMT parallelism in their workload, because they are making GPUs.
With GPGPU it's always the question "How much general purpose you want it to be?"
Which is better going forward? Which should computers 10 years from now have? I think GCN because DirectPhysics is coming. We're finally going to have a reason to get serious about physics as a gameplay mechanic and NVIDIA can't get serious because of the massive penalty in framerate CUDA incurs when context switching. It makes far more sense for physics to occupy idle compute resources than disrupting the render wave front. CUDA, as it is designed now, will hold engines back from using physics as they should.
Yeah well, physics solvers are extremely dynamic in their load ... one frame nothing is happening, the next frame one collision sets hundreds of objects in motion. If you simply must update all positions by the end of the frame, and you do, async is the way to go. I guess that's why cuda physx is exclusively of a simd/simt kind (huge number of elements behaving by the same algorithm - fluid, hair, grass). Too bad we don't have some alpha version previewed somewhere by now since async compute is a thing.
 
Intel abandoned Havok after they bought it so Microsoft has a lot of work to do before they debut it. Not just integrated into DirectX 12 but also make it competitive in the market with PhysX which has been getting constant updates for the last...almost decade. Quite depressing but it's something Microsoft can't afford to pass up.

It's going to take GPGPU to make physics mean something in games. Specifically, they need to prioritize the compute tasks and balance the GPU work with it. PhysX was always built off the foundation of having a separate card handling it. DirectPhysics can't do that.


Edit: https://trademarks.justia.com/871/86/directphysics-87186880.html
DIRECTPHYSICS/DIRECT PHYSICS said:
Computer software development tools; Computer software development tools for game programs, three-dimensional application programs and other application programs; Computer software, namely, software development kits (SDK); Computer software for use in the design, development and execution of game programs, three-dimensional interactive application programs, and other application programs; Computer software, namely, computer operating programs for games, three-dimensional interactive application programs, and other application programs; Computer software for developing and interfacing with virtual reality software, hardware and peripherals; Computer software, namely, 3D content and motion content rendering software for use in producing videos and animation
 
Last edited:
But I see a bright future if Havok sometime soon becomes DirectPhysics as part of DirectX, that's the moment we'll see huge leap in game physics. It won't be just ragdolls anymore. It'll actually be able to affect gameplay.
 
Back
Top