NVIDIA launched its GeForce GTX 1660 graphics card, doubling down on its idea of Turing-based GeForce GTX graphics cards that lack ray-tracing capabilities, but feature performance uplifts at normal raster graphics. At $280, the GTX 1660 Ti from last month ended up leagues ahead of the similarly priced and recently launched Radeon RX 590 and could play any game at 1080p with details maxed out, and 1440p with a little tweaking. At its price point, though, the GTX 1660 Ti wasn't really a successor to the current mainstream market leader in terms of sheer sales, the GTX 1060 6 GB. That distinction now goes to the GTX 1660.
At $220, the GTX 1660 is built to a cost, and NVIDIA has made sure it has ample headroom to cut costs further in the future if AMD comes out with competitive products in this segment, such as the fabled "Navi." It is carved out of the same 12 nm "TU116" silicon as the GTX 1660 Ti with fewer CUDA cores and slower 8 Gbps GDDR5 memory replacing 12 Gbps GDDR6. NVIDIA is hence looking to offer a product that's incrementally faster than the GTX 1060 6 GB and anything AMD has to offer in this segment, which can still deliver on Full HD gameplay with maximum quality.
As we detailed the "TU116" in our GTX 1660 Ti reviews, this silicon is derived from the "Turing" architecture by removing RT cores and tensor cores, leaving just the CUDA cores, which have the same IPC and clock-speed uplifts as any other RTX 20-series card. The target audience for the GTX 1660 is that colossal mass of gamers into online multiplayer e-Sports titles and just in need a card that can keep them ticking at Full HD, perhaps even at high refresh rates.
NVIDIA carved the GTX 1660 out of the "TU116" silicon by disabling 2 out of 24 streaming multiprocessors, resulting in a CUDA core count of 1,408 and 88 TMUs, which is still higher than what the "Pascal" based GTX 1060 6 GB packs. With 48 ROPs and a 192-bit GDDR5 memory bus driving 6 GB of memory, the rendering and memory subsystem is practically carried over.
Today, we have with us the MSI GeForce GTX 1660 Gaming X, the company's premium offering based on this GPU, featuring its Twin Frozr 7 cooling solution, a back-plate, idle fan-stop, and factory-overclocked speeds.
GeForce GTX 1660 Market Segment Analysis
Price
Shader Units
ROPs
Core Clock
Boost Clock
Memory Clock
GPU
Transistors
Memory
GTX 1050
$140
640
32
1354 MHz
1455 MHz
1752 MHz
GP107
3300M
2 GB, GDDR5, 128-bit
GTX 1050 Ti
$180
768
32
1290 MHz
1392 MHz
1752 MHz
GP107
3300M
4 GB, GDDR5, 128-bit
RX 570
$130
2048
32
1168 MHz
1244 MHz
1750 MHz
Ellesmere
5700M
4 GB, GDDR5, 256-bit
RX 580
$170
2304
32
1257 MHz
1340 MHz
2000 MHz
Ellesmere
5700M
8 GB, GDDR5, 256-bit
GTX 1060 3 GB
$185
1152
48
1506 MHz
1708 MHz
2002 MHz
GP106
4400M
3 GB, GDDR5, 192-bit
GTX 1060
$210
1280
48
1506 MHz
1708 MHz
2002 MHz
GP106
4400M
6 GB, GDDR5, 192-bit
RX 590
$240
2304
32
1469 MHz
1545 MHz
2000 MHz
Polaris 30
5700M
8 GB, GDDR5, 256-bit
GTX 1660
$220
1408
48
1530 MHz
1785 MHz
2000 MHz
TU116
6600M
6 GB, GDDR5, 192-bit
MSI GTX 1660 Gaming X
$250
1408
48
1530 MHz
1860 MHz
2000 MHz
TU116
6600M
6 GB, GDDR5, 192-bit
GTX 1070
$310
1920
64
1506 MHz
1683 MHz
2002 MHz
GP104
7200M
8 GB, GDDR5, 256-bit
RX Vega 56
$320
3584
64
1156 MHz
1471 MHz
800 MHz
Vega 10
12500M
8 GB, HBM2, 2048-bit
GTX 1660 Ti
$280
1536
48
1500 MHz
1770 MHz
1500 MHz
TU116
6600M
6 GB, GDDR6, 192-bit
GTX 1070 Ti
$450
2432
64
1607 MHz
1683 MHz
2000 MHz
GP104
7200M
8 GB, GDDR5, 256-bit
RTX 2060 FE
$350
1920
48
1365 MHz
1680 MHz
1750 MHz
TU106
10800M
6 GB, GDDR6, 192-bit
Architecture
The GeForce GTX 1660 is based on the same 12 nm "TU116" silicon as the GTX 1660 Ti launched last month. NVIDIA has carved out the GTX 1660 by disabling two out of 24 streaming multiprocessors (SMs), and pairing the GPU with 8 Gbps GDDR5 memory instead of 12 Gbps GDDR6. Memory bus width itself is unchanged at 192-bit wide, but the switch to GDDR5 has reduced memory bandwidth by a third (192 GB/s vs. 288 GB/s). GPU clock speeds are negligibly increased if not largely the same with 1530 MHz base and 1785 MHz GPU Boost, compared to the 1500/1770 MHz of the GTX 1660 Ti.
NVIDIA has significantly re-engineered the Graphics Processing Clusters (GPCs) of the silicon to lack RT cores and tensor cores. The chip's hierarchy is similar to other "Turing" GPUs. The GigaThread Engine and L2 cache are town-square for the GPU, which bind three GPCs with the chip's PCI-Express 3.0 x16 host and 192-bit GDDR5 memory interfaces. Each GPC has four indivisible TPCs (Texture Processing Cluster) that share a Polymorph Engine between two streaming multiprocessors (SM). Each Turing SM packs 64 CUDA cores, and thus, we end up with 128 CUDA cores per TPC, 512 per GPC, and 1,536 across the silicon. On the GTX 1660, there are 22 out of 24 SMs (or 11 out of 12 TPCs) enabled, which results in 1,408 CUDA cores. This is still more than the 1,280 of the GTX 1060 6 GB, and one has to also consider the increased IPC of the "Turing" architecture.
Much of NVIDIA's CUDA core specific innovation for Turing centers on improving the architecture's concurrent-execution capabilities. This is not the same as asynchronous compute, but the two concepts aren't too far out from each other. Turing CUDA cores are designed to execute integer and floating-point instructions per clock cycle in parallel, while older architectures, such as Pascal, can only handle one kind of execution at a time. Asynchronous compute is more of a macro concept and concerns the GPU's ability to handle various graphics and compute workloads in tandem.
Cushioning the CUDA cores is an improved L1 cache subsystem. The L1 caches are enlarged three-fold, with a four-fold increase in load/store bandwidth. The caches are configurable on the fly as either two 32 KB partitions per SM or a unified 64 KB block per TPC. NVIDIA has also substituted tensor cores with dedicated FP16 cores per SM to execute FP16 operations. These are physically separate components to the 64 FP32 and 64 INT32 cores per SM and execute FP16 at double the speed of FP32 cores. On the RTX 2060, for example, there are no dedicated FP16 cores per SM, and the tensor cores are configured to handle FP16 ops at an enormous rate.
NVIDIA has deployed the older generation GDDR5 memory on the GTX 1660, clocked at 8 Gbps data rate compared to 12 Gbps GDDR6 on the GTX 1660 Ti. This is a massive 33 percent decrease in memory bandwidth compared to the GTX 1660 Ti and exactly the same as on the GTX 1060 6 GB. The memory amount is unchanged at 6 GB.
Features
Let's talk about the two elephants in the room first. The GTX 1660 will not give you real-time raytracing because it lacks RT cores, and won't give you DLSS for want of tensor cores. What you will get is Variable Rate Shading. The Adaptive Shading (aka variable-rate shading) feature introduced with Turing is carried over to the GTX 1660. Both its key algorithms, content-adaptive shading (CAS) and motion-adaptive shading (MAS), are available. CAS senses color or spatial coherence in scenes to minimize repetitive shading of details in pursuit of increasing detail where it matters. MAS senses high motion in a scene (e.g.: race simulators) and minimizes shading of details in favor of performance.
Packaging and Contents
You will receive:
Graphics card
Documentation
Driver disc
The Card
MSI's card follows the theme set by their previous GeForce 20 cards. This is the same cooler as on the MSI GTX 1660 Ti Gaming X. A backplate is included, too. Dimensions of the card are 25.0 x 13.0 cm.
Installation requires two slots in your system.
Display connectivity options include three standard DisplayPort 1.4a and one HDMI 2.0b.
NVIDIA has updated their display engine with the Turing microarchitecture, which now supports DisplayPort 1.4a with support for VESA's nearly lossless Display Stream Compression (DSC). Combined, this enables support for 8K@30Hz with a single cable, or 8K@60Hz when DSC is turned on. For context, DisplayPort 1.4a is the latest version of the standard that was published in April, 2018.
At CES 2019, NVIDIA announced that all their graphics cards will now support VESA Adaptive Sync (aka FreeSync). While only a small number of FreeSync monitors have been fully qualified for G-SYNC, users can enable the feature in NVIDIA's control panel, no matter whether the monitor is certified or not.
The board uses a single 8-pin power connector. This input configuration is specified for up to 225 watts of power draw.