News Posts matching #DirectX 12

Return to Keyword Browsing

Futuremark Showcases DirectX Raytracing Demo, Teases Upcoming 3D Benchmark Test

DirectX Raytracing (DXR) is a new feature in DirectX 12 that opens the door to a new class of real-time graphics techniques for games. We were thrilled to join Microsoft onstage for the announcement, which we followed with a presentation of our own work in developing practical real-time applications for this exciting new tech.

Accurate real-time reflections with DirectX Raytracing
Rendering accurate reflections in real-time is difficult. There are many challenges and limitations when using the existing methods. For the past few months, we've been exploring ways of combining DirectX Raytracing with existing methods to solve some of these challenges. While much of our presentation went deep into the math for our solution, I would like to show you some examples of our new technique in action.

Lesson from the Crypto/DRAM Plagues: Build Future-Proof

As someone who does not mine crypto-currency, loves fast computers, and gaming on them, I find the current crypto-currency mining craze using graphics cards nothing short of a plague. It's like war broke out, and your government took away all the things you love from the market. All difficult times teach valuable lessons, and in this case, it is "Save up and build future-proof."

When NVIDIA launched its "Pascal" GPU architecture way back in Summer 2016, and AMD followed up, as a user of 2x GeForce GTX 970 SLI, I did not feel the need to upgrade anything, and planned to skip the Pascal/Polaris/Vega generation, and only upgrade when "Volta" or "Navi" offered something interesting. My pair of GTX 970 cards are backed by a Core i7-4770K processor, and 16 GB of dual-channel DDR3-1866 memory, both of which were considered high-end when I bought them, around 2014-15.

Throughout 2016, my GTX 970 pair ate AAA titles for breakfast. With NVIDIA investing on advancing SLI with the new SLI-HB, and DirectX 12 promising a mixed multi-GPU utopia, I had calculated a rather rosy future for my cards (at least to the point where NVIDIA would keep adding SLI profiles for newer games for my cards to chew through). What I didn't see coming was the inflection point between the decline of multi-GPU and crypto-plague eating away availability of high-end graphics cards at sane prices. That is where we are today.

NVIDIA Asks Retailers to Ensure Gamers Get Graphics Cards

The crypto-currency plague continues to nibble away at the PC gaming ecosystem, driving gamers to consoles. This affects the long-term prospects of graphics processor manufacturers like NVIDIA, who will find it difficult to pull gamers lost to consoles back to the PC ecosystem, should newer cost-effective ASICs arrest the viability of using GPUs to mine crypto-currency (as it did on several occasions in the past). The company is reportedly writing to PC hardware retailers to take steps to ensure that PC gamers have access to graphics cards first, and only then crypto-currency miners. The request doesn't prescribe any measures to control graphics card prices that have clearly gone out of hand.

Retailers in Germany responded lukewarmly, by setting quantity limitations, ranging between 2 to 5 units per order (and not per-household). Retailers responding to ComputerBase.de commented that they sometimes receive orders by e-mail with quantities running over 100 units. NVIDIA emphasizes that this request is a "recommendation" on its part, and not a directive, and that it would never interfere in freedom of commerce (by letting market-forces sell $200 graphics cards at $1000). The only people standing to lose are PC gamers who haven't upgraded their graphics cards in over two years, who are not able to play today's AAA games at recommended settings, especially when newer games are implementing the new DirectX 12 API, and their eye-candy have leaped to levels 2+ year old high-end graphics cards struggle to keep up with.

PowerColor Officially Launches Radeon RX Vega Red Devil Series

TUL Corporation, a leading and innovative manufacturer of AMD graphic cards since 1997, has launched a new PowerColor Red Devil RX VEGA 64 and Red Devil RX VEGA 56 and opened up a new generation of the graphics cards market. The VEGA series are for the extreme gamers looking for the highest resolutions, highest framerates in maximum video setting.

The PowerColor RX VEGA graphics are designed to deliver exhilarating performance in the latest DirectX 12 and Vulkan game titles. With a dedicated High-Bandwidth Cache, the VEGA utilizes HBM2, the latest in graphic memory technology, to provide incredible levels of power efficiency and memory performance. The Next-Gen Pixel Engine found in the Vega GPU is designed to boost shading performance more efficiently to bring the latest VR and extreme resolution games to life.

Futuremark Introduces "Cyan Room" DX12 VR Benchmark to VRMark

Adding to its staple of benchmarking suites, Futuremark has announced the upcoming release of their latest benchmark, dubbed "Cyan Room". this is a Bioshock-esque benchmark that was made for DX12-powered, VR-enabled workloads, and should let users know just how much processing grunt they have at their disposal. It uses a pure DirectX 12 engine built in-house and optimized for VR, besides featuring "a large, complex environment and many eye-catching effects."

The Cyan Room can be explored at the users' leisure through its "Experience Mode"; it's a benchmarking experience where users can actually change the rendering resolution and other settings to make the scene more or less demanding, on the fly. This should allow users to truly gauge the difference in experience according to achieved performance in the benchmark - the company says "using Experience mode with a VR headset is a great way to see how system performance affects your VR experience." With its massive 5K rendering resolution and spectacular volumetric lighting effects, the company says it sets a high bar for future hardware generations. Cyan Room will be released on November 22 as a free update for VRMark Advanced Edition and VRMark Professional Edition.

Latest Intel Graphics Driver Enables Netflix HDR

Intel today released its latest Graphics Driver for Windows (GDW). Version 15.60 WHQL (15.60.0.4849), which is applicable for integrated graphics embedded into 6th generation "Skylake," 7th generation "Kaby Lake," and 8th generation "Coffee Lake" processors. The drivers are WDDM 2.3 compliant (Windows 10 Fall Creators Update), and add support for Netflix HDR and YouTube HDR on Windows 10. The drivers also add support for 10-bpc (1.07 billion colors) displays over HDMI, and adds video decode hardware acceleration for several formats introduced after DirectX 12.

For those with beefier Iris Pro graphics, Intel GDW 15.60 adds optimization for "Middle-earth: Shadow of War," "Pro Evolution Soccer 2018," "Call of Duty: WWII," "Destiny 2," and "Divinity: Original Sin." As a WDDM 2.3 compliant driver, version 15.60 enables Windows Mixed Reality headsets plugged into the integrated graphics connectors. Download the driver from the link below.
DOWNLOAD: Intel Graphics Driver for Windows 15.60

Futuremark Releases 3DMark v2.4.3819 with "Time Spy Extreme" Benchmark

Futuremark today released the latest update to the 3DMark graphics benchmark suite. Version 2.4.3819, released to the public today, introduces the new "Time Spy Extreme" benchmark for machines running Windows 10 and DirectX 12 compatible graphics cards. With a rendering resolution of 4K Ultra HD (3840 x 2160 pixels), the new benchmark applies enough stress to put today's 4K UHD gaming PCs through their paces. You don't require a 4K monitor to run the test, however, your graphics card must feature at least 4 GB of video memory.

Time Spy Extreme also comes with a new CPU benchmark that is up to 3 times more taxing than the older CPU tests. It can take advantage of practically any number of CPU cores you can throw at it, and benefits from the the AVX2 instruction-set. "Time Spy Extreme," isn't available on the free version of 3DMark. You will require at least 3DMark Advanced, with a license purchased after July 14, 2016, to get it as a free upgrade. The update also improves the API overhead tests.
DOWNLOAD: Futuremark 3DMark v2.4.3819

The change-log follows.

NVIDIA Releases the GeForce 387.92 Game Ready Graphics Drivers

NVIDIA today released the GeForce 387.92 Game Ready Graphics Drivers, which presents a more comprehensive package than most Game Ready drivers. This time, the changes aren't mostly limited to new, upcoming games support, though that's still part of the package. On that front, you can count on these drivers being optimized for the upcoming Middle-earth: Shadow of War and The Evil Within 2, as well as Forza Motorsport 7, on which NVIDIA expects anywhere between 15% and 25% performance improvement (which should help improve NVIDIA's situation). On the VR space, this driver package brings the optimal experience for 4A Games' Artika.1 (if you recognize the studios' name, it's probably because they're the ones behind the Metro game series.

Also interesting, however, are the feature updates that come with this driver package. Fast Sync is now supported for SLI configurations, though Maxwell architecture users should steer clear of 4K resolution, which isn't supported for this feature. There's also NVIDIA GameStream support for HDR under Windows 10. On the renderer side, we see added support for OpenGL 4.6, as well as for DirectX Intermediate Language (DXIL). This includes full support for DirectX 12 Shader Model 6.0 ( for features such as Wavemath), and the DirectX Shader Compiler, though it's only supported on NVIDIA Kepler and later GPUs. There are also some improvements to full-screen Vulkan applications, though this particular change may cause more errors while leaving and entering full-screen mode (Alt-Tab). Grab the drivers from the link below right here on TPU, and find the change-log after the break.
DOWNLOAD: NVIDIA GeForce 387.92 Game Ready Drivers

Futuremark Readies 3DMark TimeSpy Extreme Benchmark

Futuremark is giving final touches to its top-tier GPU benchmark, 3DMark "TimeSpy Extreme." This benchmark tests your graphics hardware's performance at the 4K Ultra HD (3840 x 2160 pixels) resolution, with the latest DirectX 12 API. You will need a graphics card with at least 4 GB of video memory to run the test. The benchmark will put not just the fastest graphics cards through their paces, but is also designed to take advantage of today's multi-core processors.

3DMark "TimeSpy" Extreme can take advantage of processors with 8 or more CPU cores, and will benefit from the processors supporting the AVX2 instruction-set. Futuremark claims that the CPU tests of "TimeSpy" Extreme will be 3 times more demanding. The company also mentions that it developed the new benchmark while taking inputs from AMD, NVIDIA, and Intel. The benchmark is expected to launch on the 11th of October as an update.

AMD Phasing Out CrossFire Brand With DX 12 Adoption, Favors mGPU

An AMD representative recently answered PC World's query regarding the absence of "CrossFire" branding on their latest Radeon Software release, which introduced multi-GPU support for AMD's Vega line of graphics cards. According to the AMD representative, it goes down to a technicality, in that "CrossFire isn't mentioned because it technically refers to DX11 applications. In DirectX 12, we reference multi-GPU as applications must support mGPU, whereas AMD has to create the profiles for DX11. We've accordingly moved away from using the CrossFire tag for multi-GPU gaming."

PNY Announces GeForce GTX 1070 XLR8 OC Graphics Card

Make every gaming session an immersive and thrilling experience with PNY GTX 1070 XLR8 Gaming Overclocked Edition. PNY Technologies (PNY) considered one of the worldwide leaders in consumer electronics market and flash memory products, has launched the GeForce GTX 1070 8GB XLR8 graphics card that transforms your PC into a super-charged gaming computer.

Designed specifically to support next-generation PC games, the GeForce provides real-time 4K graphics powered by the new NVIDIA Pascal architecture. The GeForce GTX 1070 is built with 8 GB GDDR5 memory clocked at 8 Gbps with data transfer rate of 256 GB/s and 256-bit wide memory bus which provides ultra-graphics settings for an outstanding performance and gaming experience. Enjoy smooth gameplay while playing the most challenging, graphics-intensive games without any interruption or glitches.

PNY GeForce GTX 1060 6GB XLR8 Gaming OC Edition

PNY Technologies (PNY) considered one of the worldwide leaders in consumer electronics market and flash memory products, is proud to present the new PNY GeForce GTX 1060 6GB XLR8 Gaming OC Edition graphics card. Aimed at hardcore gamers and enthusiasts, PNY raises the bar with this new graphics card by overclocking it right at the factory, ensuring that it delivers every drop of graphics juice.

Powered by the raw silicon muscle power of NVIDIA's latest Pascal architecture and ultra-fast FinFET technology coupled with PNY's XLR8 Dual Fan Cooler means silent but accelerated gaming at the highest graphics settings with blistering fast frame rates. The 6GB GDDR5 memory clocked at 8 Gbps with a data transfer rate of 192 GB/s and with a 192-bit wide memory data bus guarantees that there will never be any bottleneck while transferring large game textures and huge maps.

NVIDIA GeForce 384 Series Driver Removes Need for New CPUs for 4K Netflix

NVIDIA's GeForce 384 series drivers seem to have quite a few secrets, beginning with DirectX 12 API support on 5-plus year old GeForce "Fermi" GPUs, and now 4K Ultra HD support for Netflix UWP app without the need of new-generation CPUs (namely Intel "Kaby Lake," AMD "Summit Ridge," and AMD "Bristol Ridge."). The new-generation CPUs feature a host of hardware-level DRM features which the Netflix app needs to playback 4K Ultra HD content. The new GeForce 384 series drivers let you circumvent that requirement.

Reddit user aethervisor discovered that the Windows Store (UWP platform) app of Netflix could play back content at full 4K Ultra HD resolution on their machine with an older CPU and GeForce GTX 1080 graphics. New-generation CPUs had become a requirement for this to happen, besides the latest Windows 10 version, an HDCP 2.2-compliant 4K display (and no active secondary displays that don't satisfy HDCP 2.2), a powerful enough GPU, and either the UWP app or the Netflix website on Microsoft's Edge web-browser. NVIDIA struck down a big requirement that opens up Netflix 4K to a much wider user-base.

NVIDIA Adds DirectX 12 Support to GeForce "Fermi" Architecture

With its latest GeForce 384 series graphics drivers, NVIDIA quietly added DirectX 12 API support for GPUs based on its "Fermi" architecture, as discovered by keen-eyed users on the Guru3D Forums. These include the GeForce 400-series and 500-series graphics cards. The support appears to be sufficient to run today's Direct3D feature-level 12_0 games or applications, and completes WDDM 2.2 compliance for GeForce "Fermi" graphics cards on Windows 10 Creators Update (version 1703), which could be NVIDIA's motivation for extending DirectX 12 support to these 5+ year old chips. Whether they meet your games' minimum system requirements is an entirely different matter.

Vega Frontier Ed Beats TITAN Xp in Compute, Formidable Game Performance: Preview

PC World posted a preview of an AMD Radeon Pro Vega Frontier Edition graphics card, and reported some interesting observations about the card ahead of its review NDA. The tech publication compared the air-cooled Pro Vega Frontier Edition against NVIDIA's fastest consumer graphics card, the TITAN Xp. It did reveal performance numbers of the two cards in two compute-heavy tests, SPECViewPerf 12.1 and Cinebench R15 (OpenGL test), where the Vega FE significantly outperforms the TITAN Xp. This shouldn't come as a shocker because AMD GPUs tend to have a strong footing with GPU compute performance, particularly with open standards.

It's PC World's comments on the Vega card's gaming performance that might pique your interest. In its report, the publication comments that the Radeon Pro Vega Frontier Edition offers gaming performance that is faster than NVIDIA's GeForce GTX 1080, but slightly slower than its GTX 1080 Ti graphics card. To back its statement, PC World claims to have run the Vega Frontier Edition and TITAN Xp in "Doom" with Vulkan API, "Prey" with DirectX 11, and "Sniper Elite 4" with DirectX 12. You must also take into account that the Radeon Pro Vega Frontier Edition could command a four-figure price, in the league of the TITAN Xp; and that gamers should look forward to the Radeon RX Vega series, bound for a late-July/early-August launch, at price-points more appropriate to their competitive positioning. The RX Vega is also expected to have 8 GB of memory compared to 16 GB on the Frontier Edition. Watch PC World's video presentation in the source link below.

CryEngine to Support Vulkan Renderer in Upcoming 5.4 Update

CryEngine, the rendering prodigy responsible for some of the most visually impressive titles ever to grace our personal computing and gaming shores, is getting a Vulkan renderer. The news were broken down by the team at Crytek through a blog post, where they reaffirmed their commitment to proper GitHub support and updates for their game engine. The company puts it this way:

"Vulkan renderer
Following on from the renderer refactoring and DirectX 12 implementation, the team has been hard at work implementing a Vulkan renderer. The code can be seen in Code/RenderDll/XRenderD3D9/Vulkan/… although the feature is not functional, yet. We want to make these changes available to you for review whilst we are currently stabilizing the engine for our 5.4 release. So you can track our progress on GitHub until 5.4 is finally here by the end of July."

SAPPHIRE Introduces GPRO E-Series Professional Graphics

SAPPHIRE Technology is launching its new GPRO E-Series graphics cards series for professional and industrial usage. A comprehensive lineup of models, ranging from the best-performing to most compact and power-efficient units, will suit the varying needs of businesses. Increased computational through put and further power optimizations allow for more immersive visuals and sophisticated applications - from digital gaming and casinos to industrial automation, medical, aerospace, and defence.

Available starting in Q3 2017, the SAPPHIRE GPRO E-Series includes 4 models with different power levels and outputs. The E-Series lineup starts with the SAPPHIRE GPRO E9260, the fastest of the debuting cards, with 2.5 TFLOPS of computing power in single precision operations. SAPPHIRE has included 8 GB of GDDR5 memory and 4 DisplayPort 1.3 outputs. Thanks to the latest generation of Polaris GPU architecture, with updated display engine, the card can drive up to 2 screens in 5K resolution and 60 Hz refresh rate or up to 4 screens in 4K and 60 Hz. The applied thermal solution is active, with a single efficient, dual ball bearing fan. The GPRO E9260 model is dedicated to the most graphically-demanding and compute-heavy scenarios such us high-end digital gaming, advanced Real-time medical imaging as well as military.

Microsoft Trademarks Direct Physics - HAVOK Rebranded?

As you might recall, Microsoft bought HAVOK from Intel back in 2015, promising to "add Havok's IP to its existing tools and platforms, including DirectX 12, Visual Studio, and Azure." Well, it would seem we are seeing the fruits of that particular seeding, with Microsoft having trademarked "Direct Physics".

With Microsoft having previously talked about integrating HAVOK with its DX12 API, that is probably the most probable scenario for this trademark. A tighter, in-DX12 integration could possibly allow for the physics workflow to have increased performance under the API, which is something we can all get behind of. However, this also begs the question as to what exactly happens to HAVOK licensing in the process. I myself wouldn't expect Microsoft to put its HAVOK tools and libraries behind a DX12 implementation wall - the number of companies who license those libraries aren't few in number. So my guess is that Microsoft is simply rebranding the HAVOK middleware for integration under its DX12 API, which could mean opening up its libraries to any game that makes use of DX12.

NVIDIA Demonstrates GameWorks Flow Tech Under DirectX 12

NVIDIA Flow, was previously announced by the company in 2016's GDC as the new GameWorks implementation for combustible fluid, fire and smoke simulation (superseding NVIDIA's Turbulence and FlameWorks.) It makes use of an adaptive sparse voxel grid for maximum flexibility with the least memory impact, being optimized for use of Volume Tiled Resources when available. With this technology being implemented on the Unreal Engine 4 soon, the company is now looking to increase developer awareness of the tool by showcasing its capabilities.

In the video below, the company is showing off its DirectX 12 implementation of the technology, which showcases gas combustion that results into real-time simulation of fire and smoke in the air.

Futuremark Releases 3DMark v2.3.3682

Futuremark today released an update to 3DMark benchmark suite. Version 2.3.3682 resolves a critical bug in the API overhead test, which causes the test to fail on some machines, towards the end of the DirectX 12 API test. The API overhead benchmark tests the amount of API draw-calls your machine can handle, over popular 3D graphics APIs, including DirectX 11 (single-threaded and multi-threaded), DirectX 12, and Vulkan. The rest of the feature-set is consistent with the older version 2.3.3663, in which the company introduced a brand new Vulkan API overhead test, replacing the Mantle API test. Grab the update from the link below.
DOWNLOAD: Futuremark 3DMark + Time Spy v2.3.3682

Futuremark Releases 3DMark v2.3.3663 - Adds Vulkan Support

Futuremark has just released a major update to its 3DMark benchmarking suite, adding Vulkan support while simultaneously axing its cousin, Mantle. This means that the API Overhead test now uses a Vulkan path instead of its previous Mantle one, which is sure to lead several enthusiasts into a frenzy of benchmarking under the Khronos's API (which has just recently been announced will offer support for multi-GPU in Windows 10, 8.x, 7, and Linux operating systems.)

Check some of the new features, improvements and fixes on the new version right after the break. You can download this piece of software right here on TPU - just follow the link below.
Download: Futuremark 3DMark + TimeSpy v2.3.3663

NVIDIA Releases the GeForce 378.78 Drivers

Remember that NVIDIA driver update that I mentioned yesterday? NVIDIA has just released into the wild. Version 378.78 of its GeForce driver suite provides the optimal gaming experience for Tom Clancy's Ghost Recon Wildlands and includes DirectX 12 optimizations which provide additional performance increases for a variety of titles. This release also adds official support for the GeForce GTX 1080 Ti, and updates the SLI profiles for Deus Ex: Mankind Divided, Titanfall 2, and Tom Clancy's The Division. ANSEL support for Ghost Recon Wildlands is also baked into this update, which joins the likes of Dishonored 2, Mirror's Edge Catalyst, Watch Dogs 2, The Witcher 3: Wild Hunt and The Witness, with Mass Effect: Andromeda support coming in at the game's launch.

Update: According to NVIDIA, average performance gains across all titles stands at around 16%. NVIDIA boasts of a 33% (no, that isn't a typo) on Tomb Raider; 16% on Hitman; Gears of War 4 sees a boost of 10%; Ashes of the Singularity makes do with 9%; and Tom Clancy's The Division will get a 4% increase.

As always, you can get your driver fix right here on TPU. Just follow the download link below.
DOWNLOAD: NVIDIA GeForce 378.78 Drivers

NVIDIA Working on a Major DirectX 12 Focused Driver Update

NVIDIA is reportedly working on a major driver update that improves performance across the board on games that take advantage of DirectX 12. The company is working with major game studios to improve DirectX 12 performance by up to 16 percent on recent AAA titles, such as "Gears of War 4," "Hitman" (2016), "Rise of the Tomb Raider" and "Tom Clancy's The Division." Even "Ashes of the Singularity" is getting a performance improvement.

At its GeForce GTX 1080 Ti launch event, NVIDIA revealed that its "Pascal" and "Maxwell" GPU architectures are capable of tile-based rendering, a feature that significantly improves video memory bandwidth utilization. With the GTX 1080 Ti going on sale later this month, one can expect NVIDIA to release new drivers.

Is DirectX 12 Worth the Trouble?

We are at the 2017 Game Developers Conference, and were invited to one of the many enlightening tech sessions, titled "Is DirectX 12 Worth it," by Jurjen Katsman, CEO of Nixxes, a company credited with several successful PC ports of console games (Rise of the Tomb Raider, Deus Ex Mankind Divided). Over the past 18 months, DirectX 12 has become the selling point to PC gamers, of everything from Windows 10 (free upgrade) to new graphics cards, and even games, with the lack of DirectX 12 support even denting the PR of certain new AAA game launches, until the developers hashed out support for the new API through patches. Game developers are asking the dev community at large to manage their expectations from DirectX 12, with the underlying point being that it isn't a silver-bullet to all the tech limitations developers have to cope with, and that to reap all its performance rewards, a proportionate amount of effort has to be put in by developers.

The presentation begins with the speaker talking about the disillusionment consumers have about DirectX 12, and how they're yet to see the kind of console-rivaling performance gains DirectX 12 was purported to bring. Besides lack of huge performance gains, consumers eagerly await the multi-GPU utopia that was promised to them, in which not only can you mix and match GPUs of your choice across models and brands, but also have them stack up their video memory - a theoretical possibility with by DirectX 12, but which developers argue is easier said than done, in the real world. One of the key areas where DirectX 12 is designed to improve performance is by distributing rendering overhead evenly among many CPU cores, in a multi-core CPU. For high-performance desktop users with reasonably fast CPUs, the gains are negligible. This also goes for people gaming on higher resolutions, such as 1440p and 4K Ultra HD, where the frame-rates are low, and the performance tends to be more GPU-limited.

Futuremark Readies New Vulkan and DirectX 12 Benchmarks

Futuremark is working on new game-tests for its 3DMark benchmark suite. One of these is a game test that takes advantage of DirectX 12, but isn't as taxing on the hardware as "Time Spy." Its target hardware is notebook graphics and entry-mainstream graphics cards. It will be to "Time Spy" what "Sky Diver" is to "Fire Strike."

The next, more interesting move by Futuremark is a benchmark that takes advantage of the Vulkan 3D graphics API. The company will release this Vulkan-based benchmark for both Windows and Android platforms. Lastly we've learned that development of the company's VR benchmarks are coming along nicely, and the company hopes to release new VR benchmarks for PC and mobile platforms soon. Futuremark is expected to reveal these new game-tests and benchmarks at its 2017 International CES booth, early January.
Return to Keyword Browsing
Jan 5th, 2025 11:44 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts