Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

RAID is not a backup solution. Anyone that seriously uses RAID knows what it is and when to use it. If you are worried about the integrity of your data than you need a real backup solution. RAIDn won't restore your data incase of shock, fire, or hurricanes.

A thought that occurred to me is I don't know anyone that replaces a hard drive when it gets too old. They replace them when SMART says failure is imminent or the drive has already crashed. Hard drives should be treated like tapes. No one replaces tape when they are too ragged to be used. Tapes are replaced on a scheduled (like 30-40 writes). Replace drives after a duty cycle of 5,000 hours and you can minimize your exposure.



"RAID is not a backup solution. Anyone that seriously uses RAID knows what it is and when to use it. If you are worried about the integrity of your data than you need a real backup solution. RAIDn won't restore your data incase of shock, fire, or hurricanes."

You say that like it has something to do with the article. It doesn't and it is not implied in any way.



5000 is just something I made up as an example. What should be done is HD manufacturers should rate their drives up to a certain amount of hours they have tested themselves. I'm not saying the MTBF of drives. But the chances of data integrity along the life of the drive. Like after 5000 hours you will have 99% integrity, 10000 hours is 95% integrity, 20000 hours is 80%, and so on.

They same way tire manufacturers rate their tires for so many miles or kilometers. No logical person waits for a car tire to self destruct before replacement. You note the odometer and make a record of when to change them. So you're not the unfortunate schmuck stuck in snow storm with shredded radials.

Data storage is the antithesis of good maintenance. Instead of proaction its reaction. You buy and install a drive, wait for it to die, then replace it.


Actually, I'm not so sure your advice is very good.

The likelihood of a drive failing does not linearly increase with it's age. For relatively young drives, the opposite is true -- if the drive has survived a few weeks of heavy use, it is much LESS likely to fail in the future than a fresh drive off a shelf. Tires are changed routinely because they wear -- hard drives don't so much wear as spontaneously fail. Changing them "before failure" costs you money and increases the amount of drives that blow up in your face.


I think this has more to do with manufacturing consistency than anything else. If a hard drive fails in a few weeks than it's quite likely the lot it came from is also likely to fail.

Unless something fundamental has changed about hard drives I think my point is still correct. Hard drives are electromechanical devices that also wear; bearings wear, fluids evaporate, and heads hit the platter. They spontaneously fail like light bulbs spontaneously fail. It's not spontaneous at all. If you know a light bulb is about to blow or is near the end of it's life cycle you change it.

In the greater scheme of things the cost of the hardware is miniscule compared to the data on it. If drives start randomly blowing up in your face it's time to get a different model.

What I'm getting at is we put more care into the maintenance of our cars than we do our data. If you think hard drives are expensive than what do you think of the cost of productivity while an office full of people wait for a RAID to rebuild. Instead of waiting for the inevitable failure wouldn't it be better to cycle old drives on a friday evening when usage is low.

RAID6 is just a bandaid on a bandaid. It came about because drives are of the same age when a RAID is built. And start failing around the same time. RAID 5 can recover from one failed drive, RAID6 can do 2. But you are still vulnerable if a 3rd. But this is still reactive thinking, "I will replace them as they fail", rather than using proactive solutions.


For relatively young drives, the opposite is true -- if the drive has survived a few weeks of heavy use, it is much LESS likely to fail in the future than a fresh drive off a shelf

References, please.



Dumb question here, how do I get at the smart readings? How will I know if it's warning me? I'd like to know for Ubuntu and for Windows XP.


In Google's whitepaper on hard drives a while back, they showed that, basically, SMART is useless.

So, take SMART your data with grains of salt and all that.


"man smartctl" is the place to start.

I don't know how to do it on Windows. There's probably a GUI. Although it seems that smartmontools has Windows support.


Of course RAID is a backup solution. What is a redundancy if not an immediately available backup? Having a RAID 1 or 5 array ensures that you can continue to serve your data when one of your disks crashes. Saying RAID is not a backup solution is just playing wordgames by redefining 'backup'.


It's not a backup because if you delete a file from the "main" filesystem, it is immediately deleted from the "backup" as well.


I think the important thing here is that the term "backup" only makes sense within the context of what kind of thing you're protecting against.

RAID is a useful backup for drive failures, if you're able to replace failed drives within a certain amount of time, but not for manual deletions. DVDs and tapes are useful backups for manual deletions, but not for fires. DVDs and tapes, stored at an offsite location, are useful backups for manual deletions and fires, but possibly not for earthquakes.

Since all of the different backup and archival methods have major pros and cons, the only way to say something meaningful about how useful they are is to talk about them within context.


We've solved this semantic issue. Redundancy is used to refer to availability while backup is used to refer to recovery from data loss. Hence the R in RAID.


"I know this sounds like a semantic quibble, but words mean things."

RAID is a redundancy mechanism. It's there in the name. rdiff-backup is a backup tool.


RAID is not a backup solution, it's an availability solution. RAID (with the exception or RAID 0 of course) helps you stay available and running in the event of a hardware failure. Raid is no more a "backup" mechanism then load balancing!




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: