Wednesday, November 18th 2015
NVIDIA Details "Pascal" Some More at GTC Japan
NVIDIA revealed more details of its upcoming "Pascal" GPU architecture at the Japanese edition of the Graphics Technology Conference. The architecture will be designed to nearly double performance/Watt over the current "Maxwell" architecture, by implementing the latest tech. This begins with stacked HBM2 (high-bandwidth memory 2). The top "Pascal" based product will feature four 4-gigabyte HBM2 stacks, totaling 16 GB of memory. The combined memory bandwidth for the chip will be 1 TB/s. Internally, bandwidths can touch as high as 2 TB/s. The chip itself will support up to 32 GB of memory, and so enterprise variants (Quadro, Tesla), could max out the capacity. The consumer GeForce variant is expected to serve up 16 GB.
It's also becoming clear that NVIDIA will build its "Pascal" chips on the 16 nanometer FinFET process (AMD will build its next-gen chips on more advanced 14 nm process). NVIDIA is innovating a new interconnect called NVLink, which will change the way the company has been building dual-GPU graphics cards. Currently, dual-GPU cards are essentially two graphics cards on a common PCB, with PCIe bandwidth from the slot shared by a bridge-chip, and an internal SLI bridge connecting the two GPUs. With NVLink, the two GPUs will be interconnected with an 80 GB/s bi-directional data path, letting each GPU directly address memory controlled by the other. This should greatly improve memory management in games that take advantage of newer APIs such as DirectX 12 and Vulkan; and prime the graphics card for higher display resolutions. NVIDIA is expected to launch its first "Pascal" based products in the first half of 2016.
Source:
VR World
It's also becoming clear that NVIDIA will build its "Pascal" chips on the 16 nanometer FinFET process (AMD will build its next-gen chips on more advanced 14 nm process). NVIDIA is innovating a new interconnect called NVLink, which will change the way the company has been building dual-GPU graphics cards. Currently, dual-GPU cards are essentially two graphics cards on a common PCB, with PCIe bandwidth from the slot shared by a bridge-chip, and an internal SLI bridge connecting the two GPUs. With NVLink, the two GPUs will be interconnected with an 80 GB/s bi-directional data path, letting each GPU directly address memory controlled by the other. This should greatly improve memory management in games that take advantage of newer APIs such as DirectX 12 and Vulkan; and prime the graphics card for higher display resolutions. NVIDIA is expected to launch its first "Pascal" based products in the first half of 2016.
67 Comments on NVIDIA Details "Pascal" Some More at GTC Japan
We are at the point of needing 3GB of vmem with good memory management and 4GB for textures that may not be optomized, not 16 with piss poor management. 8GB would be enough with better timings to reduce latency.
1. Can't say I've seen anything that actually backs up your assertion that 14LPP is more advanced than 16FF+/16FFC
2. I also haven't seen it confirmed anywhere that AMD will tap GloFo exclusively for GPUs. Somewhere in the last 7-8 weeks"could" has turned into "will". Some sources seem to think that GloFo will be tapped for lower end GPUs and Zen, with TSMC tasked with producing the larger GPUs.
Instead of, oh idk, having a single damn gpu capable of running games at 4k @ 120hz, they seem to focus and plan on us getting 2...
I already find it massively bs with the current gen of cards, 600 - 1000 euro for a card that feels outdated right away....
AC Unity on a mere 1080p with a freaking 1000 dollar Titan X only manages 48 fps...
BF 4 , 4k, same 1000 dollar card...a mere 41 fps....
I could go on, if I am laying down that kind of cash for the latest and greatest gpu it better damn well run atleast current ffing games at the highest settings and then some.
Its ridiculous the ask you to put down that kind of money twice.
I mean Im not even talking about the more demanding more advanced games in the future, im talking about the here and now and those cards cannot do that?
I feel we should demand better as consumers tbh but yeah that focus on better SLI using that NVlink....not a good thing to focus on imo.
Its sorta how the internet originated in Military development, everything seems to start with Military or Space development and then finds its way to the consumer.
Sure this is first for industry but yeah will make its way to the gamers soon enough.
And even if it does not, statement still stands, hate this focus and "need" for dual gpu setups to get anything decent going.
also on a side note....dear gawd what a presentation...
Did they fix the VR preemption problem?Nvidia VR preemption "possibly catastrophic"
Did they add
freesyncadaptive sync compatibility?And hopefully we get 980Ti performance in 970mini/Nano/or_even_smaller form factor. :)
And does anybody know yet Arctic Islands actual product availability (not paper) launch date/time?
Well, damn...
2) adaptive sync is THE ONLY standard, there is no "freesync" standard
3) nothing stops any manufacturer out there to use adaptive sync (dp 1.2a), no need to involve AMD or any of its "freesync" stuff in there
Basically many non-single play , non "GPU tests" scenarious just like much more VRAM
Thanks for that clean up. Do we have this as confirmation? While sure it could see starting delivery's for HPC customer initiatives first (Exascale/IBM, Cray, etc), then professional products (Tesla/Quadro), while GeForce use of the GP100 should be out a ways.
But then again Fallout 4 happened so even though Activision we're the first to get their game out. its Bethesda thats gettin all the pussy
:::EDIT:::
Oh, and not to forget about battlefront of course which has already been available in many countries apart from the UK. People are either playing one of the two games
Adaptive sync IS FreeSync.
FreeSync is the brand name for an adaptive synchronization technology for LCD displays that support a dynamic refresh rate aimed at reducing screen tearing.[2] FreeSync was initially developed by AMD in response to NVidia's G-Sync. FreeSync is royalty-free, free to use, and has no performance penalty.[3] As of 2015, VESA has adopted FreeSync as an optional component of the DisplayPort 1.2a specification.[4] FreeSync has a dynamic refresh rate range of 9–240 Hz.[3] As of August 2015, Intel also plan to support VESA's adaptive-sync with the next generation of GPU.[5]
en.wikipedia.org/wiki/FreeSync