drive failure during rebuild causes page fault
asym
asym at rfnj.org
Wed Dec 15 16:16:37 PST 2004
At 18:57 12/15/2004, Gianluca wrote:
>actually all the data I plan to keep on that server is gonna be backed up,
>either to cdr/dvdr or in the original audio cds that I still have. what I
>meant by integrity is trying to avoid having to go back to the backups to
>restore 120G (or more in this case) that were on a dead drive. I've done
>that before, and even if it's no mission-critical data, it remains a huge
>PITA :)
That's true. Restoring is always a pain in the ass, no matter the media
you use.
>thanks for the detailed explanation of how RAID5 works, somehow I didn't
>really catch the distinction between the normal and degraded operations on
>the array.
>
>what would be your recommendations for this particular (and very limited)
>application?
Honestly I'd probably go for a RAID1+0 setup. It wastes half the space in
total for mirroring, but it has none of the performance penalties of
RAID-5, and upto half the drives in the array can fail without anything but
speed being degraded. You can sort of think of this as having a second
dedicated array for 'backups' if you want, with the normal caveats --
namely that "destroyed" data cannot be recovered, such as things purposely
deleted.
RAID5 sacrifices write speed and redundancy for the sake of space. Since
you're using IDE and the drives are pretty cheap, I don't see the need for
such a sacrifice.
Just make sure the controller can do "real" 1+0. Several vendoers are
confused about what the differences are between 1+0, 0+1, and 10 -- they
mistakenly call their raid 0+1 support "RAID-10".
The difference is pretty important though. If you have say 8 drives, in
RAID 1+0 (aka 10) you would first create 4 RAID-1 mirrors with 2 disks
each, and then use these 4 virtual disks in a RAID-0 stripe setup. This
would be optimal, as any 4 drives could fail provided they all came from
different RAID-1 pairs.
In 0+1, you first create two 4-disk RAID-0 arrays and then use one as a
mirror of the other to create one large RAID-1 disk. In this setup, which
has *no* benefits over 1+0, if any drive fails the entire 4-disk RAID-0
stripe set that the disk is in goes offline and you are left with no
redundancy -- the entire array is degraded running off the remaining 4-disk
RAID-0 array, and if any of the drives in that array fail, you're smoked.
If you want redundancy to avoid having to possibly restore data, and you
can afford more disks, go 1+0. If you can't afford more disks, then one of
the striped+parity solutions (-3, -4, -5) are all you can do.. but be ready
to see write performance anywhere from "ok" on a $1500 controller, to
"annoying" on a sub $500 controller, to "downright retardedly slow" on
anything down in the cheap end -- including most IDE controllers -- Look up
the controller, find out what I/O chip it's using (most are intel based,
either StrongARM or i960) and see if the chip supports hardware XOR. If it
doesn't, you'll really wish it did.
More information about the freebsd-stable
mailing list