Re: Raid-10 mount at startup always has problem

2007-12-17 Thread Daniel L. Miller

Daniel L. Miller wrote:

Doug Ledford wrote:

Nah.  Even if we had concluded that udev was to blame here, I'm not
entirely certain that we hadn't left Daniel with the impression that we
suspected it versus blamed it, so reiterating it doesn't hurt.  And I'm
sure no one has given him a fix for the problem (although Neil did
request a change that will give debug output, but not solve the
problem), so not dropping it entirely would seem appropriate as well.
  
I've opened a bug report on Ubuntu's Launchpad.net.  Scott James 
Remnant asked me to cc him on Neil's incremental reference - we'll see 
what happens from here.


Thanks for the help guys.  At the moment, I've changed my mdadm.conf 
to explicitly list the drives, instead of the auto=partition 
parameter.  We'll see what happens on the next reboot.


I don't know if it means anything, but I'm using a self-compiled 
2.6.22 kernel - with initrd.  At least I THINK I'm using initrd - I 
have an image, but I don't see an initrd line in my grub config.  
HmmI'm going to add a stanza that includes the initrd and see what 
happens also.


Wow.  Been a while since I asked about this - I just realized a reboot 
or two has come and gone.  I checked my md status - everything was 
online!  Cool.


My current dmesg output:
sata_nv :00:07.0: version 3.4
ACPI: PCI Interrupt Link [LTID] enabled at IRQ 23
ACPI: PCI Interrupt :00:07.0[A] - Link [LTID] - GSI 23 (level, 
high) - IRQ 23

sata_nv :00:07.0: Using ADMA mode
PCI: Setting latency timer of device :00:07.0 to 64
scsi0 : sata_nv
scsi1 : sata_nv
ata1: SATA max UDMA/133 cmd 0xc20001428480 ctl 0xc200014284a0 
bmdma 0x00
011410 irq 23
ata2: SATA max UDMA/133 cmd 0xc20001428580 ctl 0xc200014285a0 
bmdma 0x00
011418 irq 23

ata1: SATA link up 1.5 Gbps (SStatus 113 SControl 300)
ata1.00: ATA-7: ST3160811AS, 3.AAE, max UDMA/133
ata1.00: 312581808 sectors, multi 16: LBA48 NCQ (depth 31/32)
ata1.00: configured for UDMA/133
ata2: SATA link up 1.5 Gbps (SStatus 113 SControl 300)
ata2.00: ATA-7: ST3160811AS, 3.AAE, max UDMA/133
ata2.00: 312581808 sectors, multi 16: LBA48 NCQ (depth 31/32)
ata2.00: configured for UDMA/133
scsi 0:0:0:0: Direct-Access ATA  ST3160811AS  3.AA PQ: 0 ANSI: 5
ata1: bounce limit 0x, segment boundary 0x, hw 
segs 61

scsi 1:0:0:0: Direct-Access ATA  ST3160811AS  3.AA PQ: 0 ANSI: 5
ata2: bounce limit 0x, segment boundary 0x, hw 
segs 61

ACPI: PCI Interrupt Link [LSI1] enabled at IRQ 22
ACPI: PCI Interrupt :00:08.0[A] - Link [LSI1] - GSI 22 (level, 
high) - IRQ 22

sata_nv :00:08.0: Using ADMA mode
PCI: Setting latency timer of device :00:08.0 to 64
scsi2 : sata_nv
scsi3 : sata_nv
ata3: SATA max UDMA/133 cmd 0xc2000142a480 ctl 0xc2000142a4a0 
bmdma 0x00
011420 irq 22
ata4: SATA max UDMA/133 cmd 0xc2000142a580 ctl 0xc2000142a5a0 
bmdma 0x00
011428 irq 22

ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300)
ata3.00: ATA-7: ST3160811AS, 3.AAE, max UDMA/133
ata3.00: 312581808 sectors, multi 16: LBA48 NCQ (depth 31/32)
ata3.00: configured for UDMA/133
ata4: SATA link up 1.5 Gbps (SStatus 113 SControl 300)
ata4.00: ATA-7: ST3160811AS, 3.AAE, max UDMA/133
ata4.00: 312581808 sectors, multi 16: LBA48 NCQ (depth 31/32)
ata4.00: configured for UDMA/133
scsi 2:0:0:0: Direct-Access ATA  ST3160811AS  3.AA PQ: 0 ANSI: 5
ata3: bounce limit 0x, segment boundary 0x, hw 
segs 61

scsi 3:0:0:0: Direct-Access ATA  ST3160811AS  3.AA PQ: 0 ANSI: 5
ata4: bounce limit 0x, segment boundary 0x, hw 
segs 61

sd 0:0:0:0: [sda] 312581808 512-byte hardware sectors (160042 MB)
sd 0:0:0:0: [sda] Write Protect is off
sd 0:0:0:0: [sda] Mode Sense: 00 3a 00 00
sd 0:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't 
support DPO or FUA

sd 0:0:0:0: [sda] 312581808 512-byte hardware sectors (160042 MB)
sd 0:0:0:0: [sda] Write Protect is off
sd 0:0:0:0: [sda] Mode Sense: 00 3a 00 00
sd 0:0:0:0: [sda] Write cache: enabled, read cache: enabled, doesn't 
support DPO or FUA

sda: unknown partition table
sd 0:0:0:0: [sda] Attached SCSI disk
sd 1:0:0:0: [sdb] 312581808 512-byte hardware sectors (160042 MB)
sd 1:0:0:0: [sdb] Write Protect is off
sd 1:0:0:0: [sdb] Mode Sense: 00 3a 00 00
sd 1:0:0:0: [sdb] Write cache: enabled, read cache: enabled, doesn't 
support DPO or FUA

sd 1:0:0:0: [sdb] 312581808 

Cannot re-assemble Degraded RAID6 after crash

2007-12-17 Thread Erich Newell
My system has crashed a couple of times, each time the two drives have
dropped off of the RAID.

Previously I simply did the following, which would take all night:

mdadm -a --re-add /dev/md2 /dev/sde3
mdadm -a --re-add /dev/md2 /dev/sdf3
mdadm -a --re-add /dev/md3 /dev/sde5
mdadm -a --re-add /dev/md3 /dev/sde5

When I woke up in the morning, everything was happy...until it crashed
again yesterday. This time, I get a message: /dev/md3 assembled from
4 drives - not enough to start the array while not clean - consider
--force.

I can re-assemble /dev/md3 (sda5, sdb5, sdc5, sdd5, sde5 and sdf5) if
I use -f, although all the other sets seem fine. I cannot --re-add
the other partitions. I'm afraid of losing data and would like advice
before I proceed. Below I've detailed my geometry and included the
data from mdadm --detail /dev/md2 (showing happy root partition),
mdadm --detail /dev/md3 (showing unhappy /srv partition) as well as
mdadm --examine /dev/sdX5 where X is the drive letter. It seems
that sde5 and sdf5 are somehow confused. They think of themselves
as: /dev/.static/dev/sdj5 and /dev/.static/dev/sdk5 as revealed by
mdadm --examine /dev/sde5

I have 4 raid partitions per disk: swap, tmp, root and srv (sda6 is
extra...result of adding 300gb disk into group of 250gb)

  cfdisk (util-linux-ng 2.13)

  Disk Drive: /dev/sda
   Size: 300090728448 bytes, 300.0 GB
 Heads: 255   Sectors per Track: 63   Cylinders: 36483

NameFlags  Part Type  FS Type  [Label]Size (MB)
 --
sda1Primary   Linux raid autodetect  501.75
sda2Primary   Linux raid autodetect 1003.49
sda3Primary   Linux raid autodetect10001.95
sda5Logical   Linux raid autodetect   239503.71
sda6Logical   Linux raid autodetect49072.03

/dev/md2:
Version : 00.90.03
  Creation Time : Sun Dec  3 20:30:54 2006
 Raid Level : raid6
 Array Size : 39069696 (37.26 GiB 40.01 GB)
  Used Dev Size : 9767424 (9.31 GiB 10.00 GB)
   Raid Devices : 6
  Total Devices : 6
Preferred Minor : 2
Persistence : Superblock is persistent

Update Time : Mon Dec 17 19:39:30 2007
  State : clean
 Active Devices : 6
Working Devices : 6
 Failed Devices : 0
  Spare Devices : 0

 Chunk Size : 64K

   UUID : 3bc78325:00787a3d:645931e8:f69b1082
 Events : 0.5781192

Number   Major   Minor   RaidDevice State
   0   830  active sync   /dev/sda3
   1   8   191  active sync   /dev/sdb3
   2   8   352  active sync   /dev/sdc3
   3   8   513  active sync   /dev/sdd3
   4   8   674  active sync   /dev/sde3
   5   8   835  active sync   /dev/sdf3

/dev/md3:
Version : 00.90.03
  Creation Time : Sun Dec  3 20:31:16 2006
 Raid Level : raid6
 Array Size : 935496448 (892.16 GiB 957.95 GB)
  Used Dev Size : 233874112 (223.04 GiB 239.49 GB)
   Raid Devices : 6
  Total Devices : 4
Preferred Minor : 3
Persistence : Superblock is persistent

Update Time : Mon Dec 17 19:29:53 2007
  State : clean, degraded
 Active Devices : 4
Working Devices : 4
 Failed Devices : 0
  Spare Devices : 0

 Chunk Size : 64K

   UUID : ae3b47bd:ac100bfe:1bed3458:8d6915d8
 Events : 0.992402

Number   Major   Minor   RaidDevice State
   0   850  active sync   /dev/sda5
   1   8   211  active sync   /dev/sdb5
   2   8   372  active sync   /dev/sdc5
   3   8   533  active sync   /dev/sdd5
   4   004  removed
   5   005  removed

/dev/sda5:
  Magic : a92b4efc
Version : 00.90.03
   UUID : ae3b47bd:ac100bfe:1bed3458:8d6915d8
  Creation Time : Sun Dec  3 20:31:16 2006
 Raid Level : raid6
  Used Dev Size : 233874112 (223.04 GiB 239.49 GB)
 Array Size : 935496448 (892.16 GiB 957.95 GB)
   Raid Devices : 6
  Total Devices : 4
Preferred Minor : 3

Update Time : Mon Dec 17 19:33:55 2007
  State : clean
 Active Devices : 4
Working Devices : 4
 Failed Devices : 2
  Spare Devices : 0
   Checksum : 47b7a255 - correct
 Events : 0.992404

 Chunk Size : 64K

  Number   Major   Minor   RaidDevice State
this 0   850  active sync   /dev/sda5

   0 0   850  active sync   /dev/sda5
   1 1   8   211  active sync   /dev/sdb5
   2 2   8   372  active sync   /dev/sdc5
   3 3   8   533  active sync   

Re: Cannot re-assemble Degraded RAID6 after crash

2007-12-17 Thread Neil Brown
On Monday December 17, [EMAIL PROTECTED] wrote:
 My system has crashed a couple of times, each time the two drives have
 dropped off of the RAID.
 
 Previously I simply did the following, which would take all night:
 
 mdadm -a --re-add /dev/md2 /dev/sde3
 mdadm -a --re-add /dev/md2 /dev/sdf3
 mdadm -a --re-add /dev/md3 /dev/sde5
 mdadm -a --re-add /dev/md3 /dev/sde5
 
 When I woke up in the morning, everything was happy...until it crashed
 again yesterday. This time, I get a message: /dev/md3 assembled from
 4 drives - not enough to start the array while not clean - consider
 --force.
 
 I can re-assemble /dev/md3 (sda5, sdb5, sdc5, sdd5, sde5 and sdf5) if
 I use -f, although all the other sets seem fine. I cannot --re-add
 the other partitions. 

What happens when you try to re-add those devices?
How about just --add.  --re-add is only need for arrays without
metadata, in your case it should behave the same as --add.

NeilBrown
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html