I just experienced this bug with Ubuntu Lucid Alpha 2.
Version: mdadm 2.6.7.1-1ubuntu15
I have 3 RAID 6 arrays: /dev/md0, /dev/md1 and /dev/md2
/dev/md0 was created by the installer and was ok. md1 and md2 were
created after installation.
At boot, my /proc/mdstat looked as follows:
md_d2 : inactive sdi2[3](S)
871895680 blocks
md_d1 : inactive sdf1[4](S)
1465047552 blocks
md0 : active raid6 sdb1[0] sdh1[2] sdi1[3] sdg1[1] sdj1[4]
314592576 blocks level 6, 64k chunk, algorithm 2 [5/5] [UUUUU]
I then proceeded to try and assemble the missing RAID arrays with the
following command:
mdadm --assemble --scan
After that, md1 and md2 were visible, but degraded:
Personalities : [raid6] [raid5] [raid4] [linear] [multipath] [raid0] [raid1]
[raid10]
md2 : active raid6 sdb2[0] sdj2[4] sdh2[2] sdg2[1]
2615687040 blocks level 6, 64k chunk, algorithm 2 [5/4] [UUU_U]
md1 : active raid6 sda1[0] sde1[3] sdd1[2] sdc1[1]
4395142656 blocks level 6, 64k chunk, algorithm 2 [5/4] [UUUU_]
md_d2 : inactive sdi2[3](S)
871895680 blocks
md_d1 : inactive sdf1[4](S)
1465047552 blocks
md0 : active raid6 sdb1[0] sdh1[2] sdi1[3] sdg1[1] sdj1[4]
314592576 blocks level 6, 64k chunk, algorithm 2 [5/5] [UUUUU]
Following some of the comments, I recovered without rebuilding the arrays with
the following commands:
1. Stop the degraded arrays and the strange md_d* arrays:
mdadm -S /dev/md1
mdadm -S /dev/md2
mdadm -S /dev/md_d1
mdadm -S /dev/md_d2
2. I checked to see that my mdstat looked clean:
cat /proc/mdstat
Personalities : [raid6] [raid5] [raid4] [linear] [multipath] [raid0] [raid1]
[raid10]
md0 : active raid6 sdb1[0] sdh1[2] sdi1[3] sdg1[1] sdj1[4]
314592576 blocks level 6, 64k chunk, algorithm 2 [5/5] [UUUUU]
unused devices: <none>
Yep, just md0, so that's all good.
3. I tried to restart md1 and md2:
mdadm --assemble --scan
4. Check mdstat
cat /proc/mdstat
Personalities : [raid6] [raid5] [raid4] [linear] [multipath] [raid0] [raid1]
[raid10]
md1 : active raid6 sda1[0] sdf1[4] sde1[3] sdd1[2] sdc1[1]
4395142656 blocks level 6, 64k chunk, algorithm 2 [5/5] [UUUUU]
md2 : active raid6 sdb2[0] sdj2[4] sdi2[3] sdh2[2] sdg2[1]
2615687040 blocks level 6, 64k chunk, algorithm 2 [5/5] [UUUUU]
md0 : active raid6 sdb1[0] sdh1[2] sdi1[3] sdg1[1] sdj1[4]
314592576 blocks level 6, 64k chunk, algorithm 2 [5/5] [UUUUU]
unused devices: <none>
5. Make sure mdadm.conf is correct.
I previously only had an entry for md0 in mdadm.conf, so I needed to add
definitions for md1 and md2:
mdadm -Es | grep md1 >> /etc/mdadm/mdadm.conf
mdadm -Es | grep md2 >> /etc/mdadm/mdadm.conf
6. For good measure I updated my initrd in case the mdadm conf is stored there:
update-initramfs -u ALL
All seems good :) Thanks for all the comments, it helped me fix a 7TB
array quickly (after having a heart attack).
--
mdadm cannot assemble array as cannot open drive with O_EXCL
https://bugs.launchpad.net/bugs/27037
You received this bug notification because you are a member of Ubuntu
Bugs, which is a direct subscriber.
--
ubuntu-bugs mailing list
[email protected]
https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs