Re: Raid-10 mount at startup always has problem
Daniel L. Miller wrote: Doug Ledford wrote: Nah. Even if we had concluded that udev was to blame here, I'm not entirely certain that we hadn't left Daniel with the impression that we suspected it versus blamed it, so reiterating it doesn't hurt. And I'm sure no one has given him a fix for the problem (although Neil did request a change that will give debug output, but not solve the problem), so not dropping it entirely would seem appropriate as well. I've opened a bug report on Ubuntu's Launchpad.net. Scott James Remnant asked me to cc him on Neil's incremental reference - we'll see what happens from here. Thanks for the help guys. At the moment, I've changed my mdadm.conf to explicitly list the drives, instead of the auto=partition parameter. We'll see what happens on the next reboot. I don't know if it means anything, but I'm using a self-compiled 2.6.22 kernel - with initrd. At least I THINK I'm using initrd - I have an image, but I don't see an initrd line in my grub config. HmmI'm going to add a stanza that includes the initrd and see what happens also. Wow. Been a while since I asked about this - I just realized a reboot or two has come and gone. I checked my md status - everything was online! Cool. My current dmesg output: sata_nv :00:07.0: version 3.4 ACPI: PCI Interrupt Link [LTID] enabled at IRQ 23 ACPI: PCI Interrupt :00:07.0[A] - Link [LTID] - GSI 23 (level, high) - IRQ 23 sata_nv :00:07.0: Using ADMA mode PCI: Setting latency timer of device :00:07.0 to 64 scsi0 : sata_nv scsi1 : sata_nv ata1: SATA max UDMA/133 cmd 0xc20001428480 ctl 0xc200014284a0 bmdma 0x00 011410 irq 23 ata2: SATA max UDMA/133 cmd 0xc20001428580 ctl 0xc200014285a0 bmdma 0x00 011418 irq 23 ata1: SATA link up 1.5 Gbps (SStatus 113 SControl 300) ata1.00: ATA-7: ST3160811AS, 3.AAE, max UDMA/133 ata1.00: 312581808 sectors, multi 16: LBA48 NCQ (depth 31/32) ata1.00: configured for UDMA/133 ata2: SATA link up 1.5 Gbps (SStatus 113 SControl 300) ata2.00: ATA-7: ST3160811AS, 3.AAE, max UDMA/133 ata2.00: 312581808 sectors, multi 16: LBA48 NCQ (depth 31/32) ata2.00: configured for UDMA/133 scsi 0:0:0:0: Direct-Access ATA ST3160811AS 3.AA PQ: 0 ANSI: 5 ata1: bounce limit 0x, segment boundary 0x, hw segs 61 scsi 1:0:0:0: Direct-Access ATA ST3160811AS 3.AA PQ: 0 ANSI: 5 ata2: bounce limit 0x, segment boundary 0x, hw segs 61 ACPI: PCI Interrupt Link [LSI1] enabled at IRQ 22 ACPI: PCI Interrupt :00:08.0[A] - Link [LSI1] - GSI 22 (level, high) - IRQ 22 sata_nv :00:08.0: Using ADMA mode PCI: Setting latency timer of device :00:08.0 to 64 scsi2 : sata_nv scsi3 : sata_nv ata3: SATA max UDMA/133 cmd 0xc2000142a480 ctl 0xc2000142a4a0 bmdma 0x00 011420 irq 22 ata4: SATA max UDMA/133 cmd 0xc2000142a580 ctl 0xc2000142a5a0 bmdma 0x00 011428 irq 22 ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) ata3.00: ATA-7: ST3160811AS, 3.AAE, max UDMA/133 ata3.00: 312581808 sectors, multi 16: LBA48 NCQ (depth 31/32) ata3.00: configured for UDMA/133 ata4: SATA link up 1.5 Gbps (SStatus 113 SControl 300) ata4.00: ATA-7: ST3160811AS, 3.AAE, max UDMA/133 ata4.00: 312581808 sectors, multi 16: LBA48 NCQ (depth 31/32) ata4.00: configured for UDMA/133 scsi 2:0:0:0: Direct-Access ATA ST3160811AS 3.AA PQ: 0 ANSI: 5 ata3: bounce limit 0x, segment boundary 0x, hw segs 61 scsi 3:0:0:0: Direct-Access ATA ST3160811AS 3.AA PQ: 0 ANSI: 5 ata4: bounce limit 0x, segment boundary 0x, hw segs 61 sd 0:0:0:0: [sda] 312581808 512-byte hardware sectors (160042 MB) sd 0:0:0:0: [sda] Write Protect is off sd 0:0:0:0: [sda] Mode Sense: 00 3a 00 00 sd 0:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA sd 0:0:0:0: [sda] 312581808 512-byte hardware sectors (160042 MB) sd 0:0:0:0: [sda] Write Protect is off sd 0:0:0:0: [sda] Mode Sense: 00 3a 00 00 sd 0:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA sda: unknown partition table sd 0:0:0:0: [sda] Attached SCSI disk sd 1:0:0:0: [sdb] 312581808 512-byte hardware sectors (160042 MB) sd 1:0:0:0: [sdb] Write Protect is off sd 1:0:0:0: [sdb] Mode Sense: 00 3a 00 00 sd 1:0:0:0: [sdb] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA sd 1:0:0:0: [sdb] 312581808
Cannot re-assemble Degraded RAID6 after crash
My system has crashed a couple of times, each time the two drives have dropped off of the RAID. Previously I simply did the following, which would take all night: mdadm -a --re-add /dev/md2 /dev/sde3 mdadm -a --re-add /dev/md2 /dev/sdf3 mdadm -a --re-add /dev/md3 /dev/sde5 mdadm -a --re-add /dev/md3 /dev/sde5 When I woke up in the morning, everything was happy...until it crashed again yesterday. This time, I get a message: /dev/md3 assembled from 4 drives - not enough to start the array while not clean - consider --force. I can re-assemble /dev/md3 (sda5, sdb5, sdc5, sdd5, sde5 and sdf5) if I use -f, although all the other sets seem fine. I cannot --re-add the other partitions. I'm afraid of losing data and would like advice before I proceed. Below I've detailed my geometry and included the data from mdadm --detail /dev/md2 (showing happy root partition), mdadm --detail /dev/md3 (showing unhappy /srv partition) as well as mdadm --examine /dev/sdX5 where X is the drive letter. It seems that sde5 and sdf5 are somehow confused. They think of themselves as: /dev/.static/dev/sdj5 and /dev/.static/dev/sdk5 as revealed by mdadm --examine /dev/sde5 I have 4 raid partitions per disk: swap, tmp, root and srv (sda6 is extra...result of adding 300gb disk into group of 250gb) cfdisk (util-linux-ng 2.13) Disk Drive: /dev/sda Size: 300090728448 bytes, 300.0 GB Heads: 255 Sectors per Track: 63 Cylinders: 36483 NameFlags Part Type FS Type [Label]Size (MB) -- sda1Primary Linux raid autodetect 501.75 sda2Primary Linux raid autodetect 1003.49 sda3Primary Linux raid autodetect10001.95 sda5Logical Linux raid autodetect 239503.71 sda6Logical Linux raid autodetect49072.03 /dev/md2: Version : 00.90.03 Creation Time : Sun Dec 3 20:30:54 2006 Raid Level : raid6 Array Size : 39069696 (37.26 GiB 40.01 GB) Used Dev Size : 9767424 (9.31 GiB 10.00 GB) Raid Devices : 6 Total Devices : 6 Preferred Minor : 2 Persistence : Superblock is persistent Update Time : Mon Dec 17 19:39:30 2007 State : clean Active Devices : 6 Working Devices : 6 Failed Devices : 0 Spare Devices : 0 Chunk Size : 64K UUID : 3bc78325:00787a3d:645931e8:f69b1082 Events : 0.5781192 Number Major Minor RaidDevice State 0 830 active sync /dev/sda3 1 8 191 active sync /dev/sdb3 2 8 352 active sync /dev/sdc3 3 8 513 active sync /dev/sdd3 4 8 674 active sync /dev/sde3 5 8 835 active sync /dev/sdf3 /dev/md3: Version : 00.90.03 Creation Time : Sun Dec 3 20:31:16 2006 Raid Level : raid6 Array Size : 935496448 (892.16 GiB 957.95 GB) Used Dev Size : 233874112 (223.04 GiB 239.49 GB) Raid Devices : 6 Total Devices : 4 Preferred Minor : 3 Persistence : Superblock is persistent Update Time : Mon Dec 17 19:29:53 2007 State : clean, degraded Active Devices : 4 Working Devices : 4 Failed Devices : 0 Spare Devices : 0 Chunk Size : 64K UUID : ae3b47bd:ac100bfe:1bed3458:8d6915d8 Events : 0.992402 Number Major Minor RaidDevice State 0 850 active sync /dev/sda5 1 8 211 active sync /dev/sdb5 2 8 372 active sync /dev/sdc5 3 8 533 active sync /dev/sdd5 4 004 removed 5 005 removed /dev/sda5: Magic : a92b4efc Version : 00.90.03 UUID : ae3b47bd:ac100bfe:1bed3458:8d6915d8 Creation Time : Sun Dec 3 20:31:16 2006 Raid Level : raid6 Used Dev Size : 233874112 (223.04 GiB 239.49 GB) Array Size : 935496448 (892.16 GiB 957.95 GB) Raid Devices : 6 Total Devices : 4 Preferred Minor : 3 Update Time : Mon Dec 17 19:33:55 2007 State : clean Active Devices : 4 Working Devices : 4 Failed Devices : 2 Spare Devices : 0 Checksum : 47b7a255 - correct Events : 0.992404 Chunk Size : 64K Number Major Minor RaidDevice State this 0 850 active sync /dev/sda5 0 0 850 active sync /dev/sda5 1 1 8 211 active sync /dev/sdb5 2 2 8 372 active sync /dev/sdc5 3 3 8 533 active sync
Re: Cannot re-assemble Degraded RAID6 after crash
On Monday December 17, [EMAIL PROTECTED] wrote: My system has crashed a couple of times, each time the two drives have dropped off of the RAID. Previously I simply did the following, which would take all night: mdadm -a --re-add /dev/md2 /dev/sde3 mdadm -a --re-add /dev/md2 /dev/sdf3 mdadm -a --re-add /dev/md3 /dev/sde5 mdadm -a --re-add /dev/md3 /dev/sde5 When I woke up in the morning, everything was happy...until it crashed again yesterday. This time, I get a message: /dev/md3 assembled from 4 drives - not enough to start the array while not clean - consider --force. I can re-assemble /dev/md3 (sda5, sdb5, sdc5, sdd5, sde5 and sdf5) if I use -f, although all the other sets seem fine. I cannot --re-add the other partitions. What happens when you try to re-add those devices? How about just --add. --re-add is only need for arrays without metadata, in your case it should behave the same as --add. NeilBrown - To unsubscribe from this list: send the line unsubscribe linux-raid in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html