Lexar NM610 1 TB M.2 NVMe SSD Review 6

Lexar NM610 1 TB M.2 NVMe SSD Review

Write Intensive Usage / SLC Cache »

DRAM-Less Testing

The weakness of DRAM-less SSDs is random write with relatively small block sizes over large areas. We tested this by sending random 4K block-sized writes to a file of varying size to control the locality of the writes, which should put different levels of stress on the flash-translation layer and reveal additional structure.



The results follow an asymptotic curve nearly exactly, which suggests that some kind of small, fixed-sized buffer is used (in the controller) that gets overwhelmed by requests that are far apart more and more the larger the test area is set.

Compared to other DRAM-less SSDs, the NM610 performs very well for all random writes within 10 GB or so. Once you go beyond that, it's similar to other SM2263XT drives, only the WD Blue NVMe with its SanDisk controller can beat it.

IO Latency

In this section, we take a closer look at the IO latencies of our SSDs, which helps quantify the time it takes for a data transfer to travel through the OS and the SSD controller, get executed, and report its completion back to the application. The numbers presented are the 99th percentile, recording an upper latency limit (=worst case) you can expect from the drive with the given IO load. The 99th percentile was chosen to eliminate outliers caused by random events, like OS processor scheduling and background processes using up CPU time. Latency is an important factor for enterprise sectors that need to achieve certain quality-of-service levels, but ends up playing an important role to us enthusiasts as well. Our goal here is to identify bottlenecks in the controller or flash cell erase process.

4K Random Read Latency
512K Sequential Read Latency


4K Random Write Latency
512K Sequential Write Latency


Mixed Accesses Patterns

Our final synthetic test workload examines IO performance with various mixed read/write ratios. On the horizontal axis, we start with a 100% read (0% write) operation on the left, moving through various read/write ratios until we reach 100% write (0% read) on the right. The 99% ratio values are especially important data points here since it's rare to only send read or write operations to a drive. It is much more common to have reads and writes interspaced in between, one source of which is disk "noise" created by the operating system or background programs. The other read/write ratios are useful because they help determine the performance you can expect from various application scenarios.

Mixed Access Patterns IOPS
Next Page »Write Intensive Usage / SLC Cache
View as single page
Dec 23rd, 2024 19:25 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts