Public bug reported:

System: Edubuntu 7.04 Kernel 2.6.20-16-generic x86_64

The server has 4 scsi disks.
There is a combination of sw raid1 and raid10.
/boot and swap are raid1
/ and /home are raid10.

I simulate an totally fault of one disk.
Then the system doesn't reboot correctly with 3 of 4 devices.
I got stuck in busybox/initramfs. There i were able to reassemble the md arrays
with mdadm.
But the server does not reboot with only 3 disks. And that should not be.

I've put a script with mdadm -A -s in the various script sections of 
initramfs and generated similar new initrd images. Also put raid1, raid10
an md/scsi modules in these initrds.
As an effect i see that during boot of initrd the md devices become "up" with 
the 3 of 4 devices.
But further mount of the root fs and starting init is still not possibly, i run 
only in busybox.

Before busybox i see these in tty-Logs (Sorry, only notices on paper, i am not 
on the server)
Trying to resume from /dev/md0
Then i see an modprobe call but only the usage text of modprobe-
Then an entry that /etc/fstab could not be read-
Then that /proc and /sys could not mounted
Then: Target filesystem doesn't have /sbin/init
Then i got in busybox.

I have tested the same scenario on an other distribution in viritual machine 
and there the system
boots after a "disk fault" only with 3 of 4 devices.

Sorry for my bad english.
I could provide more information if needed.

Thank you.

** Affects: Ubuntu
     Importance: Undecided
         Status: New

-- 
No complete reboot with sw-raid when one disk get faulty
https://bugs.launchpad.net/bugs/125561
You received this bug notification because you are a member of Ubuntu
Bugs, which is the bug contact for Ubuntu.

-- 
ubuntu-bugs mailing list
[email protected]
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs

Reply via email to