On Friday April 13, [EMAIL PROTECTED] wrote:
> Dear All,
> 
> I have an 8-drive raid-5 array running under 2.6.11. This morning it 
> bombed out, and when I brought
> it up again, two drives had incorrect event counts:
> 
> 
> sda1: 0.8258715
> sdb1: 0.8258715
> sdc1: 0.8258715
> sdd1: 0.8258715
> sde1: 0.8258715
> sdf1: 0.8258715
> sdg1: 0.8258708
> sdh1: 0.8258716
> 
> 
> sdg1 is out of date (expected), but sdh1 has received an extra event.
> 
> Any attempt to restart with mdadm --assemble --force, results in an an
> un-startable array with an event count of 0.8258715.
> 
> Can anybody advise on the correct command to use to get it started again?
> I'm assuming I'll need to use mdadm --create --assume-clean - but I'm 
> not sure
> which drives should be included/excluded when I do this.

A difference of 1 in event counts is not supposed to cause a problem.
Have you tried simply assembling the array without including sdg1.
e.g.
  mdadm -A /dev/md0 /dev/sd[abcdefh]1

??

NeilBrown
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to