I have added a new page with scenes from office. Take a look.
All posts by Hetzler
Flash Memory Summit 2015
I have posted my presentations from the 2015 Flash Memory Summit in the Library.
In my Touch Rate talk I showed how Touch Rate can be applied to hybrid storage systems. As mentioned in the talk, Tom Coughlin and I will be working on a 2nd Touch Rate whitepaper covering this new material. I will post the new paper when it’s complete. This will include an updated version of the spreadsheet supporting hybrid (cached or tiered) systems.
Touch Rate whitepaper released
Steve Hetzler and Tom Coughlin have just released a whitepaer on Touch Rate, which is a scale-independent metric for classifying the performance of storage systems under different workloads. You can download the paper here, and the accompanying Touch Rate Spreadsheet.
Flash Temperature Testing and Modeling
Temperature Testing
This an expanded version of the temperature modeling section from my Flash Memory Summit 2014 Tutorial T1.
An accurate temperature model is vital for flash devices as most vendors rely on accelerate temperature testing to verify retention capabilities. I tested flash at the SSD level as this is how the devices are integrated in to storage systems. All tests were performed using devices supporting a host-managed interface. Continue reading
Flash Memory Summit 2014 Spreadsheet
I am posting the spreadsheet that accompanies the Flash Memory Summit 2014 Seminar A chart deck. This is a LibreOffice 4 spreadsheet. Note that it includes a macro that extends the range of the cumulative binomial distribution. Interestingly, LibreOffice suffers from the same lack of precision that Excel does. The macro is provided as-is, and is not guaranteed to work in all cases. [Translation: I haven’t had time to debug it and the error handling in LibreOffice basic is abysmal.]
The spreadsheet shows how to compute reliability for flash-based storage systems, and compares various RAID architectures when using a DNR ECC approach (described in the chart deck).
Note: I have seen a lot of talk in the industry about how new systems use erasure codes instead of RAID. Technically, all RAID designs use erasure codes – even mirroring (that’s called a replication code). I use the term RAID loosely to describe a system utilizing an erasure code to protect against data loss.
The PMDS codes described here are erasure codes designed to protect against simultaneous device loss and sector loss with high data efficiency. You can read about them here, here and here.
You can download the RAID Reliability Spreadsheet.
The spreadsheet is provide under GPLv2.
Flash Memory Summit 2014 Tutorial T1 Charts
I am posting the charts from my Flash Memory Summit 2014 Tutorial T1:
Flash Memory Summit 2014 Seminar A Charts
You can download my Flash Memory Summit 2014 chart deck from here. This is my section of Seminar A – Making Error Correcting Codes Work for Flash Memory presented August 4, 2014.
Making ECC work for flash part 2
I will post a spreadsheet containing the calculations soon.
Flash Memory Summit 2014
FMS 2014
I will be at the 2014 Flash Memory Summit presenting in two tutorials.
The first is: Pre-Conference Seminar A Making Error Correcting Codes Work for Flash Memory
The second is Tutorial T-1 Measuring Reliability in SSD Storage Systems.
I think you’ll find them both enlightening. In the first, I will be presenting on the benefits of optimizing error correction at the flash device level in concert with a RAID system. In the second, I will cover material not yet presented here on SSD reliability measurements and presenting an empirical temperature acceleration model for flash, derived from device-level measurements. Don’t miss it!

We’ve moved!
Smorgastor has moved to: http://smorgastor.drhetzler.com. Please make a note.
Site remodel
Scraping off the rust
It’s taken some time, but I have remodeled the site, hopefully it’s all working now.
I have moved the site — please note we are now at: http://smorgastor.drhetzler.com.
I have updated some of the prior posts to reflect new information and new data. There will be new content coming shortly. The long break since the last post is due to the remodel (and re-architecting the site), and the fact that this is an interesting big data problem. I have many TB of data to sift through and analyze, which has required a significant coding effort.