Hi,
I found out that my storage drive was gone and I went to my server to
check out what wrong.
I've got 3 400GB disks wich form the array.
I found out I had one spare and one faulty drive, and the RAID 5 array
was not able to recover.
After a reboot because of some stuff with Xen my main rootdisk (hda)
was also failing, and the whole machine was not able to boot anymore.
And there I was...
After I tried to commit suicide and did not succeed, I went back to my
server to try something out.
I booted with Knoppix 4.02 and edited the mdadm.conf as follows:
DEVICE /dev/hd[bcd]1
ARRAY /dev/md0 devices=/dev/hdb1,/dev/hdc1,/dev/hdd1
I executed mdrun and the following messages appeared:
Forcing event count in /dev/hdd1(2) from 81190986 upto 88231796
clearing FAULTY flag for device 2 in /dev/md0 for /dev/hdd1
/dev/md0 has been started with 2 drives (out of 3) and 1 spare.
So I thought I was lucky enough, to get back my data, maybe a bit lost
concerning the event count which is missing some. Am I right?
But, when I tried to mount it the next day, this was also not
happening. I ended up with one faulty, one spare and one active. After
stopping and starting the array sometimes the array was rebuilding
again. I found out that the disk that it needs to rebuilt the array
(hdd1 that is) is
getting errors and falls back to faulty again.
Number Major Minor RaidDevice State
0 3 65 0 active sync
1 0 0 - removed
2 22 65 2 active sync
3 22 1 1 spare rebuilding
and then this:
Rebuild Status : 1% complete
Number Major Minor RaidDevice State
0 3 65 0 active sync
1 0 0 - removed
2 0 0 - removed
3 22 1 1 spare rebuilding
4 22 65 2 faulty
And my dmesg is full of these errors coming from the faulty hdd:
end_request: I/O error, dev hdd, sector 13614775
hdd: dma_intr: status=0x51 { DriveReady SeekComplete Error }
hdd: dma_intr: error=0x40 { UncorrectableError }, LBAsect=13615063,
high=0, low=13615063, sector=13614783
ide: failed opcode was: unknown
end_request: I/O error, dev hdd, sector 13614783
I guess this will never succeed...
Is there away to get this data back from the individual disks perhaps?
FYI:
[EMAIL PROTECTED] cat /proc/mdstat
Personalities : [raid5]
md0 : active raid5 hdb1[0] hdc1[3] hdd1[4](F)
781417472 blocks level 5, 64k chunk, algorithm 2 [3/1] [U__]
[>....................] recovery = 1.7% (6807460/390708736)
finish=3626.9min speed=1764K/sec
unused devices: <none>
Krekna
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at http://vger.kernel.org/majordomo-info.html