AnandTech Storage Bench 2013

When I built the AnandTech Heavy and Light Storage Bench suites in 2011 I did so because we didn't have any good tools at the time that would begin to stress a drive's garbage collection routines. Once all blocks have a sufficient number of used pages, all further writes will inevitably trigger some sort of garbage collection/block recycling algorithm. Our Heavy 2011 test in particular was designed to do just this. By hitting the test SSD with a large enough and write intensive enough workload, we could ensure that some amount of GC would happen.

There were a couple of issues with our 2011 tests that I've been wanting to rectify however. First off, all of our 2011 tests were built using Windows 7 x64 pre-SP1, which meant there were potentially some 4K alignment issues that wouldn't exist had we built the trace on a system with SP1. This didn't really impact most SSDs but it proved to be a problem with some hard drives. Secondly, and more recently, I've shifted focus from simply triggering GC routines to really looking at worst case scenario performance after prolonged random IO. For years I'd felt the negative impacts of inconsistent IO performance with all SSDs, but until the S3700 showed up I didn't think to actually measure and visualize IO consistency. The problem with our IO consistency tests are they are very focused on 4KB random writes at high queue depths and full LBA spans, not exactly a real world client usage model. The aspects of SSD architecture that those tests stress however are very important, and none of our existing tests were doing a good job of quantifying that.

I needed an updated heavy test, one that dealt with an even larger set of data and one that somehow incorporated IO consistency into its metrics. I think I have that test. The new benchmark doesn't even have a name, I've just been calling it The Destroyer (although AnandTech Storage Bench 2013 is likely a better fit for PR reasons).

Everything about this new test is bigger and better. The test platform moves to Windows 8 Pro x64. The workload is far more realistic. Just as before, this is an application trace based test - I record all IO requests made to a test system, then play them back on the drive I'm measuring and run statistical analysis on the drive's responses.

Imitating most modern benchmarks I crafted the Destroyer out of a series of scenarios. For this benchmark I focused heavily on Photo editing, Gaming, Virtualization, General Productivity, Video Playback and Application Development. Rough descriptions of the various scenarios are in the table below:

AnandTech Storage Bench 2013 Preview - The Destroyer
Workload Description Applications Used
Photo Sync/Editing Import images, edit, export Adobe Photoshop CS6, Adobe Lightroom 4, Dropbox
Gaming Download/install games, play games Steam, Deus Ex, Skyrim, Starcraft 2, BioShock Infinite
Virtualization Run/manage VM, use general apps inside VM VirtualBox
General Productivity Browse the web, manage local email, copy files, encrypt/decrypt files, backup system, download content, virus/malware scan Chrome, IE10, Outlook, Windows 8, AxCrypt, uTorrent, AdAware
Video Playback Copy and watch movies Windows 8
Application Development Compile projects, check out code, download code samples Visual Studio 2012

While some tasks remained independent, many were stitched together (e.g. system backups would take place while other scenarios were taking place). The overall stats give some justification to what I've been calling this test internally:

AnandTech Storage Bench 2013 Preview - The Destroyer, Specs
  The Destroyer (2013) Heavy 2011
Reads 38.83 million 2.17 million
Writes 10.98 million 1.78 million
Total IO Operations 49.8 million 3.99 million
Total GB Read 1583.02 GB 48.63 GB
Total GB Written 875.62 GB 106.32 GB
Average Queue Depth ~5.5 ~4.6
Focus Worst case multitasking, IO consistency Peak IO, basic GC routines

SSDs have grown in their performance abilities over the years, so I wanted a new test that could really push high queue depths at times. The average queue depth is still realistic for a client workload, but the Destroyer has some very demanding peaks. When I first introduced the Heavy 2011 test, some drives would take multiple hours to complete it - today most high performance SSDs can finish the test in under 90 minutes. The Destroyer? So far the fastest I've seen it go is 10 hours. Most high performance I've tested seem to need around 12 - 13 hours per run, with mainstream drives taking closer to 24 hours. The read/write balance is also a lot more realistic than in the Heavy 2011 test. Back in 2011 I just needed something that had a ton of writes so I could start separating the good from the bad. Now that the drives have matured, I felt a test that was a bit more balanced would be a better idea.

Despite the balance recalibration, there's just a ton of data moving around in this test. Ultimately the sheer volume of data here and the fact that there's a good amount of random IO courtesy of all of the multitasking (e.g. background VM work, background photo exports/syncs, etc...) makes the Destroyer do a far better job of giving credit for performance consistency than the old Heavy 2011 test. Both tests are valid, they just stress/showcase different things. As the days of begging for better random IO performance and basic GC intelligence are over, I wanted a test that would give me a bit more of what I'm interested in these days. As I mentioned in the S3700 review - having good worst case IO performance and consistency matters just as much to client users as it does to enterprise users.

Given the sheer amount of time it takes to run through the Destroyer, and the fact that the test was only completed a little over a week ago, I don't have many results to share. I'll be populating this database over the coming weeks/months. I'm still hunting for any issues/weirdness with the test so I'm not ready to remove the "Preview" label from it just yet. But the results thus far are very telling.

I'm reporting two primary metrics with the Destroyer: average data rate in MB/s and average service time in microseconds. The former gives you an idea of the throughput of the drive during the time that it was running the Destroyer workload. This can be a very good indication of overall performance. What average data rate doesn't do a good job of is taking into account response time of very bursty (read: high queue depth) IO. By reporting average service time we heavily weigh latency for queued IOs. You'll note that this is a metric I've been reporting in our enterprise benchmarks for a while now. With the client tests maturing, the time was right for a little convergence.

AnandTech Storage Bench 2013 - The Destroyer

As you'd expect, the combination of great performance consistency and competitive peak performance drives the Extreme II to the top of our Destroyer charts. I didn't expect to see anyone put out an SSD faster than the Seagate 600 so soon but it looks like SanDisk did it.

AnandTech Storage Bench 2013 - The Destroyer

 

Performance Consistency Random & Sequential Performance
Comments Locked

51 Comments

View All Comments

  • dsumanik - Tuesday, June 4, 2013 - link

    The benches on this drive are good.....not great, and I don think the opening bias is necessary. Who runs any disk at capacity 24/7? Perhaps some people temporarily... But 24/7 drive full???

    Only a fool.

    Kudos to sandisk for making a competitive offering, but please anandtech keep the bias out of the reviews....specially when it's not warranted.

    Storage bench is great, but it's not the only metric.

    Haswell is good, not great. But if your rocking a 2600k from 2 years ago? Meh.

    Where are the legendary power savings? Why don't we have 4 ghz + skus? 8 cores? 64gb ram support? Quick sync degraded lol!! Good job on iris pro. Why can't I buy it and slap it into an enthusiast board?

    Yet you read this review and the haswell review and come away feeling positive.

    Real life:

    Intel,
    A mild upgrade in IPC, higher In use TDP, 2 year old CPU's are still competitive

    Sandisk,
    Mixed bag of results, on unproven firmware.
    .
  • Death666Angel - Tuesday, June 4, 2013 - link

    Why do you keep ignoring the Samsung 840 Pro with spare area increased when it comes to consistency. It seems to me to be the best drive around. And if you value and know about consistency it seems pretty straight forward to increase the spare area and you should have the abilities to do so as well.
  • seapeople - Wednesday, June 5, 2013 - link

    Agreed, it looks like a Samsung 840 Pro that's not completely full would be the performance king in every aspect - most consistent (check the 25% spare area graphs!), fastest in every test, good reliability history, and the best all around power consumption numbers, especially in the idle state which is presumably the most important.

    Yet this drive is virtually ignored in the review, other than the ancillary mention in all the performance benchmarks it still wins, "The SanDisk did great here! Only a little behind all the Samsung drives... and as long as the Samsung drives are completely full, then the SanDisk gets better consistency, too! The SanDisk is my FAVORITE!"

    The prevailing theme of this review should probably be "The SanDisk gives you performance nearly as good as a Samsung at a lower price." Not, "OMG I HAVE A NEW FAV0RIT3 DRIVE! Look at the contrived benchmark I came up with to punish all the other drives being used in ways that nobody would actually use them in..."

    Seriously, anybody doing all that junk with their SSD would know to partition 25% of spare area into it, which then makes the Samsung Pro the clear winner, albeit at a higher cost per usable GB.
  • FunBunny2 - Tuesday, June 4, 2013 - link

    To the extent that "cloud" (re-)creates server-dense/client-thin computing, how well an SSD behaves in today's "client" doesn't matter much. Server workloads, with lots o random operations, will be where storage happens. Anand is correct to test SSDs under loads more server-like. As many have figured out, HDD in the enterprise are little different from consumer parts. "Cloud" vendors, in order to make money, will segue to "consumer" SSD. Thus, we do need to know how well they behave doing "server" loads; they will in any case. Clients will come with some amount flash (not necessarily even on current file system protocols).
  • joel4565 - Tuesday, June 4, 2013 - link

    Any word on whether this drive will be offered in a 960 GB capacity for a reasonable price in the near future?

    This looks like the best performing drive yet reviewed, but I doubt I will see that big of difference from my 120 GB Crucial M4 in day to day usage. I really don't think most of us will see a large difference until we go to a faster interface.

    So unless this drastically change in the next few months, I think my next drive will be the Crucial M500 960GB. Yes it will not be as consistent or quite as fast as the SanDisk Extreme II, but I won't have to worry about splitting my files, or moving steam games from my 7200 rpm drive to the SSD if they have long load times.
  • clepsydrae - Wednesday, June 5, 2013 - link

    Question for those more knowledgeable: I'm building a new DAW (4770k, win 8) which will also be used for development (Eclipse in linux). Based on earlier anandtech reviews I ordered a 128GB 840P Pro for use as the OS drive and eclipse workspace directory and the like. Reading this article, i'm not sure if I should return the 840P for the SanDisk... the 840P leads it in almost all the metrics except the one that is the most "real-world" and which seems to mimic what I'll be using it for (i.e. Eclipse.)

    Opinions?
  • bmgoodman - Wednesday, June 5, 2013 - link

    I gave up on SanDisk after they totally botched TRIM on their previous generation drive. They did such a poor job admitting it and finally fixing it that it left a bad taste in my mouth. They'd have to *give* me a drive for me to try their products again.
  • samster712 - Friday, June 7, 2013 - link

    So would anyone recommend this drive over the 840pro 256? Im very indecisive about buying a new drive.
  • Rumboogy - Thursday, July 11, 2013 - link

    Quick question. You mentioned a method to create an unused block of storage that could be used by the controller by creating a new partition (I assume fully formatting it) and then deleting it. This assumes TRIM marks the whole set of LBAs that covered the partition as being available. What is the comparable procedure on a Mac? Particularly if you don't get TRIM by default. And if you do turn it would it work in this case? Is there a way to guarantee you are allocating a block of LBAs to non-use on the Mac?
  • pcmax - Monday, August 12, 2013 - link

    Would have been really nice to compare it to their previous gen the Extreme I?

Log in

Don't have an account? Sign up now