Nice page.
Just recently been dealing with a HW card that udev doesn't like (doesn't have
SERIAL or SHORT_SERIAL), so I've been thinking about this a lot recently.
* * *
As for pros v cons...
I would think the main advantage to modern HW RAID systems is the ability to
hot-plug. SW RAID has many advantages, but being able to detect a failure, and
then physically see, replace, and rebuild a degraded array while the array is
alive has absolutely got to be the unequivocally primary benefit of a complete
HW RAID setup.
SW RAID is great. Generally faster (at least for RAID-0, back when I used to
benchmark this sort of thing). But, to be fair, while SW has the benefit of
being open-sourced, it does suffer from version skew, too. I have no idea if
new kernel versions make old MD devices unrecognizable, or if everything is
always backwards-compatible. That's worth finding out & mentioning. And, even
if the kernel is currently backwards-compatible ATM, who's providing the
guarantee that newer versions will also be? Sure, it's open-sourced, but,
realistically, most RAID users aren't going to be able to kernel-hack
(driver-hack) in the event that the kernel eventually deprecates a version of
the MD driver. To me, that's just as bad a problem as not being able to find
the same HW card.
It's also worth saying that in software RAID, you have to shut down the machine
to do any repairs, even if the array is running in a degraded state. Unless
you have PCI- or SATA-hotplug in your kernel (is this widely supported or
stable)...and even then, you'd have to be able to put those drives in a
hot-plug bay.
Might also want to mention hot spares.
And...(again, still trying to be constructive, not a jerk)...a page about RAID
absolutely has to have a recovery HOWTO. It's just dangerous not to include
it, lest someone get a machine running, and has no idea how to recover from it.
And, in addition to the "normal" recovery scenarios, point out how it might be
worth using with udev (disk/by-id) long names lest they reorder devices (or the
kernel does it on a version change). I personally just went through this the
hard way on a colo server...
Also, it's probably worth mentioning that most HW RAID setups (provided they
have a Linux driver) make recovery much easier. Just slide a new drive in,
maybe issue a command or two, and the card's firmware will take care of the
rebuild, all while the machine continues to run. With mdadm, I think the whole
recovery process is harder (or, more involved & dangerous to someone who might
be new).
Finally, might want to combine the RAID and LVM pages.
Q
On Feb 8, 2012, at 4:11 PM, Bruce Dubbs wrote:
> I've added mdadm and a new page, About RAID, to the book.
>
> I'd appreciate feedback on ways to improve it.
>
> http://www.linuxfromscratch.org/blfs/view/svn/postlfs/raid.html
>
> -- Bruce
> --
> http://linuxfromscratch.org/mailman/listinfo/blfs-dev
> FAQ: http://www.linuxfromscratch.org/blfs/faq.html
> Unsubscribe: See the above information page
--
http://linuxfromscratch.org/mailman/listinfo/blfs-dev
FAQ: http://www.linuxfromscratch.org/blfs/faq.html
Unsubscribe: See the above information page