Thanks for the replys,
It looks as though it should be easy.
I want to be able to reboot using md0 for the root boot partition,
then I can dispose of the ext2 fs on the /dev/sda1 and raidhotadd
sda1 to md0.
Is this possible, what does this lilo error message mean?
[root@otherweb /root]# df
Filesystem 1k-blocks Used Available Use% Mounted on
/dev/sda1 127902 40984 80314 34% /
/dev/md1 10079980 37028 9530908 0% /home
/dev/md2 4032448 416792 3410812 11% /usr
/dev/md4 254699 15 241532 0% /tmp
/dev/md5 254699 6844 234703 3% /var
/dev/md3 2016656 1992 1912220 0% /usr/local
/dev/md0 127790 40988 80204 34% /stage
[root@otherweb /]# lilo -V
LILO version 21.4-3
[root@otherweb /]# lilo -r /stage
boot = /dev/sdb, map = /boot/map.0811
Added bzImage *
Syntax error near line 2 in file /etc/lilo.conf
[root@otherweb /]# cat /stage/etc/lilo.conf
boot=/dev/md0
default=Linux
delay=50
vga=normal
read-only
image=/boot/bzImage
label=Linux
root=/dev/md0
image=/boot/vmlinuz
label=lin
root=/dev/md0
[root@otherweb boot]# head -9 /etc/raidtab
raiddev /dev/md0
raid-level 1
nr-raid-disks 2
persistent-superblock 1
chunk-size 4
device /dev/sda1
failed-disk 0
device /dev/sdb1
raid-disk 1
I want to be able to reboot using md0 for the root boot partition,
then I can dispose of the files on the /dev/sda1 and raidhotadd
sda1 to md0.
Is this possible, what does the lilo error message mean?
-------- Original Message --------
Subject: RE: root+boot on Raid1 + lilo
Date: Wed, 28 Jun 2000 09:46:44 -0500
From: "Diegmueller, Jason (I.T. Dept)" <[EMAIL PROTECTED]>
To: 'Hugh Bragg' <[EMAIL PROTECTED]>
Hugh--
It's actually relatively easy to move a root filesystem over to a
RAID. I don't have a HOWTO or step-by-step, but here's what I did
just one week ago for a 3-disk RAID1:
- Install System as usual
- Establish /dev/md0, but mark the used disk as "failed-disk" in
the /etc/raidtab (as opposed to raid-disk). This will have the
software RAID mark this disk as bad, but since it's RAID1, it will
continue in degraded mode. Also, make sure you ARE using the
persistent-superblocks.
- Format the filesystem (I use 4096-byte block e2fs, benchmarks show
that to be quicker)
- Mount the /dev/md0 filesystem. I always use "/newraid".
- Copy the current filesystem over to the new filesystem (I used
cp -a /bin /dev /usr /var etc etc etc /newraid". No need to
copy /proc.
- mkdir a /newraid/proc (so there is a /proc under the new filesystem
for when it boots up)
- Modify the /newraid/etc/fstab to utilize /dev/md0 as the root
(as opposed to /dev/sda1 or whatever)
Here's where it gets subjective. Some people use the RedHat-patched
RAID1-bootable lilo, me personally .. I do this:
- I have separate 16MB boot partitions, /mnt/boot1, /mnt/boot2, and
/mnt/boot3. These are exactly identical (/tmp, /etc, /boot, /dev)
with /etc/lilo.conf's underneath which tell lilo to install on it's
corresponding drive (ie, /mnt/boot1 == /dev/sda1, and has a
"boot = /dev/sda" line, /mnt/boot2 == /dev/sdb1, and has a
"boot = /dev/sdb" line, etc). This way, any drive can die, and
booting
will occur.
- I then run "lilo -r /mnt/boot1", "lilo -r /mnt/boot2", and
"lilo -r /mnt/boot3". The -r tells lilo to chroot, which is why you
must have a directory structure under /mnt/boot1, etc. You must have
a /mnt/boot1/dev with the devices referred to in
/mnt/boot1/etc/lilo.conf
(such as /mnt/boot1/dev/md0, /mnt/boot1/dev/sda
Also ..
- Make sure the RAID'd partitions are marked type "fd", for RAID
autodetect. Otherwise you'll have trouble booting.
Reboot.
- If you're up and running on the /dev/md0 root, you now need to
integrate
the "old root" in to your RAID. Modify /etc/raidtab and change the
"failed-disk" to "raid-disk", then "raidhotadd /dev/md0 /dev/sda5"
replacing
sda5 with the old root. It might make you --force it, I can't
remember.
When it finishing rebuilding, mark the old root partition as fd as well
so
it will autodetect. You should then be good.
I realize this is a long diatrabe and not a simple HOWTO, but hopefully
it
will help. Feel free to email ANY questions ... I'd love to help. I'm
not
a RAID maintainer, just a happy user. =) I'm doing root-RAID5,
root-RAID1,
and even root-RAID0 on 5 different machines, 3 of those I'd consider
"production" machines .. one of those being up over 300 days now without
a
problem.