ADATA Swordfish 1 TB M.2 NVMe SSD Review - No Excuse not to NVMe 36

ADATA Swordfish 1 TB M.2 NVMe SSD Review - No Excuse not to NVMe

Write Intensive Usage & SLC Cache »

DRAM-Less Testing

The weakness of DRAM-less SSDs is random write with relatively small block sizes over large areas. We tested this by sending random 4K block-sized writes to a file of varying size to control the locality of the writes, which should put different levels of stress on the flash-translation layer and reveal additional structure.



The results follow an asymptotic curve nearly exactly, which suggests the use of some kind of small, fixed-sized buffer in the controller that gets overwhelmed by requests that are further apart the larger the test area is set.

Compared to other DRAM-less SSDs, the Swordfish definitely performs worse. Out of the NVMe drives it's actually the worst performer; only the SATA-based BX500 is slower.

IO Latency

In this section, we take a closer look at the IO latencies of our SSDs, which helps quantify the time it takes for a data transfer to travel through the OS and the SSD controller, get executed, and report its completion back to the application. The numbers presented are the 99th percentile, recording an upper latency limit (=worst case) you can expect from the drive with the given IO load. The 99th percentile was chosen to eliminate outliers caused by random events, like OS processor scheduling and background processes using up CPU time. Latency is an important factor for enterprise sectors that need to achieve certain quality-of-service levels, but ends up playing an important role to us enthusiasts as well. Our goal here is to identify bottlenecks in the controller or flash cell erase process.

4K Random Read Latency
512K Sequential Read Latency


4K Random Write Latency
512K Sequential Write Latency


Mixed Accesses Patterns

Our final synthetic test workload examines IO performance with various mixed read/write ratios. On the horizontal axis, we start with a 100% read (0% write) operation on the left, moving through various read/write ratios until we reach 100% write (0% read) on the right. The 99% ratio values are especially important data points here since it's rare to only send read or write operations to a drive. It is much more common to have reads and writes interspaced in between, one source of which is disk "noise" created by the operating system or background programs. The other read/write ratios are useful because they help determine the performance you can expect from various application scenarios.

Mixed Access Patterns IOPS
Next Page »Write Intensive Usage & SLC Cache
View as single page
Nov 27th, 2024 11:50 EST change timezone

New Forum Posts

Popular Reviews

Controversial News Posts