Cannot re-assemble Degraded RAID6 after crash

2007-12-17 Thread Erich Newell
My system has crashed a couple of times, each time the two drives have
dropped off of the RAID.

Previously I simply did the following, which would take all night:

mdadm -a --re-add /dev/md2 /dev/sde3
mdadm -a --re-add /dev/md2 /dev/sdf3
mdadm -a --re-add /dev/md3 /dev/sde5
mdadm -a --re-add /dev/md3 /dev/sde5

When I woke up in the morning, everything was happy...until it crashed
again yesterday. This time, I get a message: /dev/md3 assembled from
4 drives - not enough to start the array while not clean - consider
--force.

I can re-assemble /dev/md3 (sda5, sdb5, sdc5, sdd5, sde5 and sdf5) if
I use -f, although all the other sets seem fine. I cannot --re-add
the other partitions. I'm afraid of losing data and would like advice
before I proceed. Below I've detailed my geometry and included the
data from mdadm --detail /dev/md2 (showing happy root partition),
mdadm --detail /dev/md3 (showing unhappy /srv partition) as well as
mdadm --examine /dev/sdX5 where X is the drive letter. It seems
that sde5 and sdf5 are somehow confused. They think of themselves
as: /dev/.static/dev/sdj5 and /dev/.static/dev/sdk5 as revealed by
mdadm --examine /dev/sde5

I have 4 raid partitions per disk: swap, tmp, root and srv (sda6 is
extra...result of adding 300gb disk into group of 250gb)

  cfdisk (util-linux-ng 2.13)

  Disk Drive: /dev/sda
   Size: 300090728448 bytes, 300.0 GB
 Heads: 255   Sectors per Track: 63   Cylinders: 36483

NameFlags  Part Type  FS Type  [Label]Size (MB)
 --
sda1Primary   Linux raid autodetect  501.75
sda2Primary   Linux raid autodetect 1003.49
sda3Primary   Linux raid autodetect10001.95
sda5Logical   Linux raid autodetect   239503.71
sda6Logical   Linux raid autodetect49072.03

/dev/md2:
Version : 00.90.03
  Creation Time : Sun Dec  3 20:30:54 2006
 Raid Level : raid6
 Array Size : 39069696 (37.26 GiB 40.01 GB)
  Used Dev Size : 9767424 (9.31 GiB 10.00 GB)
   Raid Devices : 6
  Total Devices : 6
Preferred Minor : 2
Persistence : Superblock is persistent

Update Time : Mon Dec 17 19:39:30 2007
  State : clean
 Active Devices : 6
Working Devices : 6
 Failed Devices : 0
  Spare Devices : 0

 Chunk Size : 64K

   UUID : 3bc78325:00787a3d:645931e8:f69b1082
 Events : 0.5781192

Number   Major   Minor   RaidDevice State
   0   830  active sync   /dev/sda3
   1   8   191  active sync   /dev/sdb3
   2   8   352  active sync   /dev/sdc3
   3   8   513  active sync   /dev/sdd3
   4   8   674  active sync   /dev/sde3
   5   8   835  active sync   /dev/sdf3

/dev/md3:
Version : 00.90.03
  Creation Time : Sun Dec  3 20:31:16 2006
 Raid Level : raid6
 Array Size : 935496448 (892.16 GiB 957.95 GB)
  Used Dev Size : 233874112 (223.04 GiB 239.49 GB)
   Raid Devices : 6
  Total Devices : 4
Preferred Minor : 3
Persistence : Superblock is persistent

Update Time : Mon Dec 17 19:29:53 2007
  State : clean, degraded
 Active Devices : 4
Working Devices : 4
 Failed Devices : 0
  Spare Devices : 0

 Chunk Size : 64K

   UUID : ae3b47bd:ac100bfe:1bed3458:8d6915d8
 Events : 0.992402

Number   Major   Minor   RaidDevice State
   0   850  active sync   /dev/sda5
   1   8   211  active sync   /dev/sdb5
   2   8   372  active sync   /dev/sdc5
   3   8   533  active sync   /dev/sdd5
   4   004  removed
   5   005  removed

/dev/sda5:
  Magic : a92b4efc
Version : 00.90.03
   UUID : ae3b47bd:ac100bfe:1bed3458:8d6915d8
  Creation Time : Sun Dec  3 20:31:16 2006
 Raid Level : raid6
  Used Dev Size : 233874112 (223.04 GiB 239.49 GB)
 Array Size : 935496448 (892.16 GiB 957.95 GB)
   Raid Devices : 6
  Total Devices : 4
Preferred Minor : 3

Update Time : Mon Dec 17 19:33:55 2007
  State : clean
 Active Devices : 4
Working Devices : 4
 Failed Devices : 2
  Spare Devices : 0
   Checksum : 47b7a255 - correct
 Events : 0.992404

 Chunk Size : 64K

  Number   Major   Minor   RaidDevice State
this 0   850  active sync   /dev/sda5

   0 0   850  active sync   /dev/sda5
   1 1   8   211  active sync   /dev/sdb5
   2 2   8   372  active sync   /dev/sdc5
   3 3   8   533  active sync   

Re: Cannot re-assemble Degraded RAID6 after crash

2007-12-17 Thread Neil Brown
On Monday December 17, [EMAIL PROTECTED] wrote:
 My system has crashed a couple of times, each time the two drives have
 dropped off of the RAID.
 
 Previously I simply did the following, which would take all night:
 
 mdadm -a --re-add /dev/md2 /dev/sde3
 mdadm -a --re-add /dev/md2 /dev/sdf3
 mdadm -a --re-add /dev/md3 /dev/sde5
 mdadm -a --re-add /dev/md3 /dev/sde5
 
 When I woke up in the morning, everything was happy...until it crashed
 again yesterday. This time, I get a message: /dev/md3 assembled from
 4 drives - not enough to start the array while not clean - consider
 --force.
 
 I can re-assemble /dev/md3 (sda5, sdb5, sdc5, sdd5, sde5 and sdf5) if
 I use -f, although all the other sets seem fine. I cannot --re-add
 the other partitions. 

What happens when you try to re-add those devices?
How about just --add.  --re-add is only need for arrays without
metadata, in your case it should behave the same as --add.

NeilBrown
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html