I am an early adopter of solid state drives, having paid $621 last November for an Intel X25-M 80 GB SSD, and $189 last month for an OCZ Vertex 60 GB SSD, both from Newegg. I made each purchase after closely studying Anandtech's storage articles on solid state drives:
http://www.anandtech.com/storage/
which are required reading for anyone considering such a purchase. My Intel X25-M was used (and will again be used) as the internal drive in my MacBook.
Here is an issue that has received little attention in these articles: How many times can a hard drive survive a power failure while writing?
Over the three decades I have owned computers, I have experienced many power failures with mechanical drives. Often, I was lucky and the drive was fine. Often, I needed to reformat the drive, and then it was fine. I have never lost a drive this way, although I have read andecdotal reports of others losing drives this way.
On my very first power failure, I bricked my Intel X25-M solid state drive. On an airplane, I put my MacBook to sleep, and then changed the battery. Apparently I didn't wait long enough, and the MacBook was still writing "sleep" data when the power was lost.
I am thrilled with Intel's support, under the circumstances. They offered me a free swap under warranty. I instead elected to pay $25 for a Cross Ship, along with accepting a $279.45 hold on my card to guarantee the return. (Of course I asked if I could have two at this price! They said no.)
I am less thrilled with Intel's engineering and software support. I may be no expert, but to discuss this one needs to be up on how solid state drives work, specifically the existence of "wear-leveling" algorithms and so forth.
There is an abstraction layer in solid state drives that doesn't exist in mechanical drives, allowing the solid state drive to simulate a mechanical drive from the operating system's point of view, while handling the bookkeeping of rewriting data in blocks, with wear-leveling movement of data. There is nothing a user can do to penetrate this abstaction layer, at least for an Intel X25-M drive. If the data tables used by this abstraction layer get hosed by a power failure, the effect on the user is indistinguishable from a hardware failure.
There is a notion of "atomic writes" that comes up both in file system design (see the radical ZFS file system and its brethren) and in symmetric multiprocessor parallelism (so other cores don't see inconsistent data). Intel could have specified atomic writes as a design requirement, for the abstraction layers hidden from the user. I have no idea if they did or not, but my experience suggests that they sacrificed this for a bit of extra speed. Atomic writes would protect against hosing data, through thousands of power failures. (Good luck getting Intel to comment on this...)
Likewise, the drive itself is aware of a reformat request, which can be distinguished from daily use. It could elect to do a hard reset as part of handling such requests. Unless the wear-leveling algorithms work really well, one would want to preserve past history counts, another reason to want atomic writes, and why a reset could be complicated. As far as I can tell, there is instead no way for a user to reset the Intel X25-M. Tech support confirmed this; my drive simply needs to be replaced.
It is interesting that Intel, and other vendors, offer warranty support under these circumstances. One could instead imagine selling solid state drives with the user taking full responsibility for power failures, perhaps buying insurance, as one does when taking responsibility for crashing a car. This would however have a dampening affect on the solid state market.
Here is my challenge to AnandTech:
It would be relatively easy to build a computer that tested hard drives by deliberately causing power failures during writes, then reformatting and repeating the test. (Put the drive under testing on a different power supply, which turns on and off via a digital timer like used to cycle house lights, and write a clever script.) Part of any storage article evaluation should be to expose each hard drive to hundreds of power failures while writing, and determine which drives end up bricked and how quickly.
My guess? Mechanical drives can survive 1000 "duty cycles" of power failures while writing, with loss of data but no physical damage. Solid state drives can reliably be bricked by similar testing, in the first few cycles. If this is true, it deserves to be widely known.
I still swear by the speed of a solid state drive, and my backup strategies expect loss of data as a regularly recurring event. I'd think twice about recommending a solid state drive to a less technically aware friend, give my experience and the above conjecture.
http://www.anandtech.com/storage/
which are required reading for anyone considering such a purchase. My Intel X25-M was used (and will again be used) as the internal drive in my MacBook.
Here is an issue that has received little attention in these articles: How many times can a hard drive survive a power failure while writing?
Over the three decades I have owned computers, I have experienced many power failures with mechanical drives. Often, I was lucky and the drive was fine. Often, I needed to reformat the drive, and then it was fine. I have never lost a drive this way, although I have read andecdotal reports of others losing drives this way.
On my very first power failure, I bricked my Intel X25-M solid state drive. On an airplane, I put my MacBook to sleep, and then changed the battery. Apparently I didn't wait long enough, and the MacBook was still writing "sleep" data when the power was lost.
I am thrilled with Intel's support, under the circumstances. They offered me a free swap under warranty. I instead elected to pay $25 for a Cross Ship, along with accepting a $279.45 hold on my card to guarantee the return. (Of course I asked if I could have two at this price! They said no.)
I am less thrilled with Intel's engineering and software support. I may be no expert, but to discuss this one needs to be up on how solid state drives work, specifically the existence of "wear-leveling" algorithms and so forth.
There is an abstraction layer in solid state drives that doesn't exist in mechanical drives, allowing the solid state drive to simulate a mechanical drive from the operating system's point of view, while handling the bookkeeping of rewriting data in blocks, with wear-leveling movement of data. There is nothing a user can do to penetrate this abstaction layer, at least for an Intel X25-M drive. If the data tables used by this abstraction layer get hosed by a power failure, the effect on the user is indistinguishable from a hardware failure.
There is a notion of "atomic writes" that comes up both in file system design (see the radical ZFS file system and its brethren) and in symmetric multiprocessor parallelism (so other cores don't see inconsistent data). Intel could have specified atomic writes as a design requirement, for the abstraction layers hidden from the user. I have no idea if they did or not, but my experience suggests that they sacrificed this for a bit of extra speed. Atomic writes would protect against hosing data, through thousands of power failures. (Good luck getting Intel to comment on this...)
Likewise, the drive itself is aware of a reformat request, which can be distinguished from daily use. It could elect to do a hard reset as part of handling such requests. Unless the wear-leveling algorithms work really well, one would want to preserve past history counts, another reason to want atomic writes, and why a reset could be complicated. As far as I can tell, there is instead no way for a user to reset the Intel X25-M. Tech support confirmed this; my drive simply needs to be replaced.
It is interesting that Intel, and other vendors, offer warranty support under these circumstances. One could instead imagine selling solid state drives with the user taking full responsibility for power failures, perhaps buying insurance, as one does when taking responsibility for crashing a car. This would however have a dampening affect on the solid state market.
Here is my challenge to AnandTech:
It would be relatively easy to build a computer that tested hard drives by deliberately causing power failures during writes, then reformatting and repeating the test. (Put the drive under testing on a different power supply, which turns on and off via a digital timer like used to cycle house lights, and write a clever script.) Part of any storage article evaluation should be to expose each hard drive to hundreds of power failures while writing, and determine which drives end up bricked and how quickly.
My guess? Mechanical drives can survive 1000 "duty cycles" of power failures while writing, with loss of data but no physical damage. Solid state drives can reliably be bricked by similar testing, in the first few cycles. If this is true, it deserves to be widely known.
I still swear by the speed of a solid state drive, and my backup strategies expect loss of data as a regularly recurring event. I'd think twice about recommending a solid state drive to a less technically aware friend, give my experience and the above conjecture.
