Actually, found a couple web documents here that may prove to be interesting to others creating arrays:
Draft write up of "silent" data errors (not only for disk but the entire chain of the process):
http://indico.cern.ch/getFile.py/acc...r&confId=13797
Failure rates on drives:
http://www.usenix.org/events/fast07/...tml/index.html
Another failure rate study:
http://209.85.163.132/papers/disk_failures.pdf
Pretty much shows that:
1) if you need to have data accurate you need to _ALWAYS_ verify it before, during, after, use as no subsystem will give you correct data 100% of the time.
2) MTTF, AFR, are not good paper specs to "buy" off of.
3) Older drives than rated lifetime or drives that have reported an error/reallocation should be replaced/removed due to the dramatically higher failure rates.



Reply With Quote
i7 920@2.8
X3220@3.0
X3220@2.4
E8400@4.05
E6600@2.4
and two the documents show the statistical error/failure issues on drives (and barring a similar study for home-only drives to contradict due to environmental reasons all we can do is take this as indicative of the base drive failure rates) which would apply equally to the enterprise and homes.



). It's an unfortunate trend that I'm seeing over the past 5-7 years where people are not truly understanding their apps nor their subsystems. 

Bookmarks