proactive-raid-disk-replacement

2006-09-08 Thread Michael Tokarev
Recently Dean Gaudet, in thread titled 'Feature
Request/Suggestion - Drive Linking', mentioned his
document, http://arctic.org/~dean/proactive-raid5-disk-replacement.txt

I've read it, and have some umm.. concerns.  Here's why:


 mdadm -Gb internal --bitmap-chunk=1024 /dev/md4
 mdadm /dev/md4 -r /dev/sdh1
 mdadm /dev/md4 -f /dev/sde1 -r /dev/sde1
 mdadm --build /dev/md5 -ayes --level=1 --raid-devices=2 /dev/sde1 missing
 mdadm /dev/md4 --re-add /dev/md5
 mdadm /dev/md5 -a /dev/sdh1

 ... wait a few hours for md5 resync...

And here's the problem.  While new disk, sdh1, are resynced from
old, probably failing disk sde1, chances are high that there will
be an unreadable block on sde1.  And this means the whole thing
will not work -- md5 initially contained one working drive (sde1)
and one spare (sdh1) which is being converted (resynced) to working
disk.  But after read error on sde1, md5 will contain one failed
drive and one spare -- for raid1 it's fatal combination.

While at the same time, it's perfectly easy to reconstruct this
failing block from other component devices of md4.

That to say: this way of replacing disk in a software raid array
isn't much better than just removing old drive and adding new one.
And if the drive you're replacing is failing (according to SMART
for example), this method is more likely to fail.

/mjt
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: proactive-raid-disk-replacement

2006-09-08 Thread dean gaudet
On Fri, 8 Sep 2006, Michael Tokarev wrote:

 Recently Dean Gaudet, in thread titled 'Feature
 Request/Suggestion - Drive Linking', mentioned his
 document, http://arctic.org/~dean/proactive-raid5-disk-replacement.txt
 
 I've read it, and have some umm.. concerns.  Here's why:
 
 
  mdadm -Gb internal --bitmap-chunk=1024 /dev/md4
  mdadm /dev/md4 -r /dev/sdh1
  mdadm /dev/md4 -f /dev/sde1 -r /dev/sde1
  mdadm --build /dev/md5 -ayes --level=1 --raid-devices=2 /dev/sde1 missing
  mdadm /dev/md4 --re-add /dev/md5
  mdadm /dev/md5 -a /dev/sdh1
 
  ... wait a few hours for md5 resync...
 
 And here's the problem.  While new disk, sdh1, are resynced from
 old, probably failing disk sde1, chances are high that there will
 be an unreadable block on sde1.  And this means the whole thing
 will not work -- md5 initially contained one working drive (sde1)
 and one spare (sdh1) which is being converted (resynced) to working
 disk.  But after read error on sde1, md5 will contain one failed
 drive and one spare -- for raid1 it's fatal combination.
 
 While at the same time, it's perfectly easy to reconstruct this
 failing block from other component devices of md4.

this statement is an argument for native support for this type of activity 
in md itself.

 That to say: this way of replacing disk in a software raid array
 isn't much better than just removing old drive and adding new one.

hmm... i'm not sure i agree.  in your proposal you're guaranteed to have 
no redundancy while you wait for the new disk to sync in the raid5.

in my proposal the probability that you'll retain redundancy through the 
entire process is non-zero.  we can debate how non-zero it is, but 
non-zero is greater than zero.

i'll admit it depends a heck of a lot on how long you wait to replace your 
disks, but i prefer to replace mine well before they get to the point 
where just reading the entire disk is guaranteed to result in problems.


 And if the drive you're replacing is failing (according to SMART
 for example), this method is more likely to fail.

my practice is to run regular SMART long self tests, which tend to find 
Current_Pending_Sectors (which are generally read errors waiting to 
happen) and then launch a repair sync action... that generally drops the 
Current_Pending_Sector back to zero.  either through a realloc or just 
simply rewriting the block.  if it's a realloc then i consider if there's 
enough of them to warrant replacing the disk...

so for me the chances of a read error while doing the raid1 thing aren't 
as high as they could be...

but yeah you've convinced me this solution isn't good enough.

-dean
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: Interesting RAID checking observations - I'm getting it too

2006-09-08 Thread Mark Smith
Hi Bill,

On Mon, 04 Sep 2006 12:42:53 -0400
Bill Davidsen [EMAIL PROTECTED] wrote:

 
 
 
 Mark Smith wrote:
 
 
 Interesting, but do you run other stuff at that time? Several 
 distributions run various things in the middle of the night which really 
 bog the machine.
 

Doesn't look like it. Shifted the check to 3.00am, 4.00am period is now
all nothing :

00:01:01  CPU %user %nice   %system   %iowait%steal 
%idle
.
.
03:56:02  all  0.85  0.00  0.38  0.00  0.00 
98.77
04:01:01  all  0.75  0.00  0.34  0.00  0.00 
98.91
04:06:01  all  0.73  0.00  0.35  0.00  0.00 
98.92
04:11:01  all  0.73  0.00  0.38  0.00  0.00 
98.89
04:16:01  all  0.84  0.00  0.37  0.00  0.00 
98.79
04:21:01  all  0.68  0.00  0.34  0.00  0.00 
98.98
04:26:01  all  0.78  0.00  0.39  0.00  0.00 
98.83
04:31:01  all  1.18  0.00  0.40  0.00  0.00 
98.43
04:36:01  all  0.88  0.00  0.33  0.00  0.00 
98.79
04:41:01  all  0.72  0.00  0.38  0.00  0.00 
98.90
04:46:01  all  0.80  0.00  0.42  0.00  0.00 
98.78
.
.

Regards,
Mark.

-- 

Sheep are slow and tasty, and therefore must remain constantly
 alert.
   - Bruce Schneier, Beyond Fear
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


RAID5 fill up?

2006-09-08 Thread Lars Schimmer
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

Hi!

I´ve got a software RAiD5 with 6 250GB HDs.
Now I changed one disk after another to a 400GB HD and resynced the
raid5 after each change.
Now the RAID5 has got 6 400GB HDs and still uses only 6*250GB space.
How can I grow the md0 device to use 6*400GB?

MfG,
Lars Schimmer
- --
- -
TU Graz, Institut für ComputerGraphik  WissensVisualisierung
Tel: +43 316 873-5405   E-Mail: [EMAIL PROTECTED]
Fax: +43 316 873-5402   PGP-Key-ID: 0x4A9B1723
-BEGIN PGP SIGNATURE-
Version: GnuPG v1.4.1 (GNU/Linux)
Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org

iD8DBQFFAThzmWhuE0qbFyMRApImAJ9j3CJRyfoGeg2IthY6JCMDRs5GugCfQZIY
ENAC2gqh3fc8/HZQs1dUhAs=
=guH4
-END PGP SIGNATURE-
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


RAID5 fill up?

2006-09-08 Thread Lars Schimmer
Hi!

I´ve got a software RAiD5 with 6 250GB HDs.
Now I changed one disk after another to a 400GB HD and resynced the
raid5 after each change.
Now the RAID5 has got 6 400GB HDs and still uses only 6*250GB space.
How can I grow the md0 device to use 6*400GB?

MfG,
Lars Schimmer
-- 
-
TU Graz, Institut für ComputerGraphik  WissensVisualisierung
Tel: +43 316 873-5405   E-Mail: [EMAIL PROTECTED]
Fax: +43 316 873-5402   PGP-Key-ID: 0x4A9B1723
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: RAID5 fill up?

2006-09-08 Thread Michael Tokarev
Lars Schimmer wrote:
 Hi!
 
 I´ve got a software RAiD5 with 6 250GB HDs.
 Now I changed one disk after another to a 400GB HD and resynced the
 raid5 after each change.
 Now the RAID5 has got 6 400GB HDs and still uses only 6*250GB space.
 How can I grow the md0 device to use 6*400GB?

mdadm --grow is your friend.

/mjt

-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: proactive-raid-disk-replacement

2006-09-08 Thread Michael Tokarev
dean gaudet wrote:
 On Fri, 8 Sep 2006, Michael Tokarev wrote:
 
 Recently Dean Gaudet, in thread titled 'Feature
 Request/Suggestion - Drive Linking', mentioned his
 document, http://arctic.org/~dean/proactive-raid5-disk-replacement.txt

 I've read it, and have some umm.. concerns.  Here's why:

 
 mdadm -Gb internal --bitmap-chunk=1024 /dev/md4

By the way, don't specify bitmap-chunk for internal bitmap.
It's needed for file-based (external) bitmap.  With internal
bitmap, we have fixed size in superblock for it, so bitmap-chunk
is determined by dividing that size by size of the array.

 mdadm /dev/md4 -r /dev/sdh1
 mdadm /dev/md4 -f /dev/sde1 -r /dev/sde1
 mdadm --build /dev/md5 -ayes --level=1 --raid-devices=2 /dev/sde1 missing
 mdadm /dev/md4 --re-add /dev/md5
 mdadm /dev/md5 -a /dev/sdh1

 ... wait a few hours for md5 resync...
 And here's the problem.  While new disk, sdh1, are resynced from
 old, probably failing disk sde1, chances are high that there will
 be an unreadable block on sde1.  And this means the whole thing
 will not work -- md5 initially contained one working drive (sde1)
 and one spare (sdh1) which is being converted (resynced) to working
 disk.  But after read error on sde1, md5 will contain one failed
 drive and one spare -- for raid1 it's fatal combination.

 While at the same time, it's perfectly easy to reconstruct this
 failing block from other component devices of md4.
 
 this statement is an argument for native support for this type of activity 
 in md itself.

Yes, definitely.

 That to say: this way of replacing disk in a software raid array
 isn't much better than just removing old drive and adding new one.
 
 hmm... i'm not sure i agree.  in your proposal you're guaranteed to have 
 no redundancy while you wait for the new disk to sync in the raid5.

It's not a proposal per se, it's just another possible way (used by majority
of users I think, because it's way simpler ;)

 in my proposal the probability that you'll retain redundancy through the 
 entire process is non-zero.  we can debate how non-zero it is, but 
 non-zero is greater than zero.

Yes there will be no redundancy in my variant, guaranteed.  And yes,
there is probability to complete the whole your process without a glitch.

 i'll admit it depends a heck of a lot on how long you wait to replace your 
 disks, but i prefer to replace mine well before they get to the point 
 where just reading the entire disk is guaranteed to result in problems.
 
 And if the drive you're replacing is failing (according to SMART
 for example), this method is more likely to fail.
 
 my practice is to run regular SMART long self tests, which tend to find 
 Current_Pending_Sectors (which are generally read errors waiting to 
 happen) and then launch a repair sync action... that generally drops the 
 Current_Pending_Sector back to zero.  either through a realloc or just 
 simply rewriting the block.  if it's a realloc then i consider if there's 
 enough of them to warrant replacing the disk...
 
 so for me the chances of a read error while doing the raid1 thing aren't 
 as high as they could be...

So the whole thing goes this way:
  0) do a SMART selftest ;)
  1) do repair for the whole array
  2) copy data from failing to new drive
(using temporary superblock-less array)
  2a) if step 2 failed still, probably due to new bad sectors,
  go the old way, removing the failing drive and adding
  new one.

That's 2x or 3x (or 4x counting the selftest, but that should be
done regardless) more work than just going the old way from the
beginning, but still some chances to have it completed flawlessly
in 2 steps, without losing redundancy.

Too complicated and too long for most people I'd say ;)

I can come to yet another way, which is only somewhat possible with
current md code. In 3 variants.

1)  Offline the array, stop it.
Make a copy of the drive using dd with error=skip (or how it is),
 noticing the bad blocks
Mark those bad blocks in bitmap as dirty
Assemble the array with new drive, letting it to resync the blocks
to new drive which we were unable to copy previously.

This variant does not lose redundancy at all, but requires the array to
be off-line during the whole copy procedure.  What's missing (which has
been discussed on linux-raid@ recently too) is the ability to mark those
bad blocks in bitmap.

2)  The same, but not offlining the array.  Hot-remove a drive, make copy
   of it to new drive, flip necessary bitmap bits, and re-add the new drive,
   and let raid code to resync changed (during copy, while the array was
   still active, something might has changed) and missing blocks.

This variant still loses redundancy, but not much of it, provided the bitmap
code works correctly.

3)  The same as your way, with the difference that we tell md to *skip* and
  ignore possible errors during resync (which is also not possible currently).

 but yeah you've convinced me this solution isn't good 

Re: RAID5 fill up?

2006-09-08 Thread Lars Schimmer
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

Michael Tokarev wrote:
 Lars Schimmer wrote:
 Hi!

 I´ve got a software RAiD5 with 6 250GB HDs.
 Now I changed one disk after another to a 400GB HD and resynced the
 raid5 after each change.
 Now the RAID5 has got 6 400GB HDs and still uses only 6*250GB space.
 How can I grow the md0 device to use 6*400GB?
 
 mdadm --grow is your friend.

Oh, damn, right. I was focussed on --grow to add a new HD to the RAID
But isn´t there a switch to grow to max possible value?
Do I always have to search for the biggest value and type it in by hand?

 /mjt
 


MfG,
Lars Schimmer
- --
- -
TU Graz, Institut für ComputerGraphik  WissensVisualisierung
Tel: +43 316 873-5405   E-Mail: [EMAIL PROTECTED]
Fax: +43 316 873-5402   PGP-Key-ID: 0x4A9B1723
-BEGIN PGP SIGNATURE-
Version: GnuPG v1.4.1 (GNU/Linux)
Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org

iD8DBQFFAWF3mWhuE0qbFyMRAn7QAJ9nZ7I+n/bxM22AoKbfSXoMovup8ACfS+i6
nYASA51BlYQaSavD7CWLL8c=
=S0dO
-END PGP SIGNATURE-
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: RAID5 fill up?

2006-09-08 Thread Luca Berra

On Fri, Sep 08, 2006 at 02:26:31PM +0200, Lars Schimmer wrote:

-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

Michael Tokarev wrote:

Lars Schimmer wrote:

Hi!

I´ve got a software RAiD5 with 6 250GB HDs.
Now I changed one disk after another to a 400GB HD and resynced the
raid5 after each change.
Now the RAID5 has got 6 400GB HDs and still uses only 6*250GB space.
How can I grow the md0 device to use 6*400GB?


mdadm --grow is your friend.


Oh, damn, right. I was focussed on --grow to add a new HD to the RAID
But isn´t there a switch to grow to max possible value?
Do I always have to search for the biggest value and type it in by hand?

man mdadm

  -z, --size=

 This value can be set with --grow for RAID level 1/4/5/6. If the
 array  was created with a size smaller than the currently active
 drives, the extra space can be accessed using --grow.  The  size
 can  be given as max which means to choose the largest size that
 fits on all current drives.


--
Luca Berra -- [EMAIL PROTECTED]
   Communication Media  Services S.r.l.
/\
\ / ASCII RIBBON CAMPAIGN
 XAGAINST HTML MAIL
/ \
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: RAID5 fill up?

2006-09-08 Thread Lars Schimmer
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

Michael Tokarev wrote:
 Lars Schimmer wrote:
 []
 mdadm --grow is your friend.
 Oh, damn, right. I was focussed on --grow to add a new HD to the RAID
 But isn´t there a switch to grow to max possible value?
 Do I always have to search for the biggest value and type it in by hand?
 
 How about actually reading the fine manual? ;-P
 
   -z, --size=
 
   This value can be set with --grow for RAID level 1/4/5/6. If the
   array  was created with a size smaller than the currently active
   drives, the extra space can be accessed using --grow.  The  size
   can  be given as max which means to choose the largest size that
   ^^^
   fits on all current drives.

*damn* This must be the friday... Or I´m in weekend to fast.

THANKS so far.

 /mjt


MfG,
Lars Schimmer
- --
- -
TU Graz, Institut für ComputerGraphik  WissensVisualisierung
Tel: +43 316 873-5405   E-Mail: [EMAIL PROTECTED]
Fax: +43 316 873-5402   PGP-Key-ID: 0x4A9B1723
-BEGIN PGP SIGNATURE-
Version: GnuPG v1.4.1 (GNU/Linux)
Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org

iD8DBQFFAXMUmWhuE0qbFyMRAtqjAJ96irdwtHjIEpvXbrq60a01VLAsMQCdHhCn
YIPwr2pZQ7cErVc9NxenxGA=
=Kgy0
-END PGP SIGNATURE-
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Messed up creating new array...

2006-09-08 Thread Ruth Ivimey-Cook
Folks,

I messed up slightly when creating a new 6-disk raid6 array, and am wondering
if there is a simple answer. The problem is that I didn't partition the drives,
but simply used the whole drive. All drives are of the same type and using the
Supermicro SAT2-MV8 controller.

This is a problem because mdadm doesn't find the component drives... I can only
start the array by using mdadm --assemble /dev/md3 /dev/sd{drives}. While
this works it means my system doesn't boot very easily :-(

For fixes I can think of the possible solutions:

 - Modify the startup script to do the mdadm --assemble /dev/md3
/dev/sd{a,b,c,d,e}. Is there any other disadvantage of this?

 - Use fdisk to create whole-disk partitions - but would this damage the
already stored data?

 - Start again doing it right But this will take 12 hours and has a chance
of failure!

Any thoughts?

Regards,

Ruth

-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Please help me save my data

2006-09-08 Thread martin . kihlgren

Hello list.

I have a spot of trouble with a RAID5 array of mine, and I thought maybe you 
could help me.

This is the story so far:

 * I bought 10 external USB drives. This seemed like a good idea, they are
   cheap, they are hot-pluggable and they are fast enough.
 * I set them up in two RAID5 arrays, which I set up as LVM pv's. Then I
   created an LVM vg out of these and an LVM lv out of the vg.
 * I encrypted this lv and formatted it with an xfs fs.

This all worked perfectly fine, until I realized how bad these drives and this 
USB controller work with ehci_hcd.

In short, the devices get reset all the time. And each time they get reset, 
everything stops for a while. Nothing strange and no showstopper 
here.

But the really bad thing is when they reset in some extra-bad way, and get 
dropped completely. What happens then is that the RAID5 system 
drops them, and they get reincarnated with a new device name. /dev/sdi becomes 
suddenly /dev/sdn or something similarly horrbile.

And since I (up until yesterday) didnt know about write-intent bitmaps each 
resync took around 10 hours. Plenty of time for ANOTHER disk to 
fail and get dropped.

This I usually solved by doing mdadm -S and then mdadm -A -f.

Yesterday, however, I was feeling extra clever, and I just did mdadm -a 
/dev/md1 /dev/sdn1.

This was a huge mistake.

What had happened, I now realized, was this:

 * /dev/md1 is fine
 * /dev/sdX1 drops, and /dev/md1 is degraded
 * I re-add /dev/sdX1 in its new guise, and /dev/md1 is resyncing with
   4 working drives and one spare
 * /dev/sdY1 drops, and /dev/md1 stops
 * I re-add /dev/sdY1 in its new guise, and mdadm marks it as a SPARE. * I 
suddenly have an array with 3 working drives and 2 spares where
   I know that one spare is in fact synced and ready to go, since
   the array stopped the moment it failed.

Also, I dont know any longer WHERE in the array the synced but
spare-marked drive should go. I know that the working drives are 0, 2 and 4, 
but not where the synced spare drive should go.

So, what I want to do is:

 * Mark the synced spare drive as working and in position 1
 * Assemble the array without the unsynced spare and check if this
   provides consistent data
 * If it didnt, I want to mark the synced spare as working and in
   position 3, and try the same thing again
 * When I have it working, I just want to add the unsynced spare and
   let it sync normally
 * Then I will create a write-intent bitmap to avoid the dangerously
   long sync times, and also buy a new USB controller hoping that it will solve 
my problems

So, do you guys have any idea how I can do this? mdadm doesnt support changing 
the superblock in such a free hand manner...

Please help me save this data :/ It is precious to me :.(

regards,
//Martin Kihlgren

-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Please help me save my data

2006-09-08 Thread martin . kihlgren

Hello list.

I have a spot of trouble with a RAID5 array of mine, and I thought maybe
you could help me.

This is the story so far:

 * I bought 10 external USB drives. This seemed like a good idea, they are
   cheap, they are hot-pluggable and they are fast enough.
 * I set them up in two RAID5 arrays, which I set up as LVM pv's. Then I
   created an LVM vg out of these and an LVM lv out of the vg.
 * I encrypted this lv and formatted it with an xfs fs.

This all worked perfectly fine, until I realized how bad these drives and
this USB controller work with ehci_hcd.

In short, the devices get reset all the time. And each time they get
reset, everything stops for a while. Nothing strange and no showstopper
here.

But the really bad thing is when they reset in some extra-bad way, and get
dropped completely. What happens then is that the RAID5 system drops them,
and they get reincarnated with a new device name. /dev/sdi becomes
suddenly /dev/sdn or something similarly horrbile.

And since I (up until yesterday) didnt know about write-intent bitmaps
each resync took around 10 hours. Plenty of time for ANOTHER disk to fail
and get dropped.

This I usually solved by doing mdadm -S and then mdadm -A -f.

Yesterday, however, I was feeling extra clever, and I just did mdadm -a
/dev/md1 /dev/sdn1.

This was a huge mistake.

What had happened, I now realized, was this:

 * /dev/md1 is fine
 * /dev/sdX1 drops, and /dev/md1 is degraded
 * I re-add /dev/sdX1 in its new guise, and /dev/md1 is resyncing with
   4 working drives and one spare
 * /dev/sdY1 drops, and /dev/md1 stops
 * I re-add /dev/sdY1 in its new guise, and mdadm marks it as a SPARE.
 * I suddenly have an array with 3 working drives and 2 spares where
   I know that one spare is in fact synced and ready to go, since
   the array stopped the moment it failed.

Also, I dont know any longer WHERE in the array the synced but
spare-marked drive should go. I know that the working drives are 0, 2 and
4, but not where the synced spare drive should go.

So, what I want to do is:

 * Mark the synced spare drive as working and in position 1
 * Assemble the array without the unsynced spare and check if this
   provides consistent data
 * If it didnt, I want to mark the synced spare as working and in
   position 3, and try the same thing again
 * When I have it working, I just want to add the unsynced spare and
   let it sync normally
 * Then I will create a write-intent bitmap to avoid the dangerously
   long sync times, and also buy a new USB controller hoping that it
   will solve my problems

So, do you guys have any idea how I can do this? mdadm doesnt support
changing the superblock in such a free hand manner...

Please help me save this data :/ It is precious to me :.(

regards,
//Martin Kihlgren
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: RAID5 fill up?

2006-09-08 Thread Mr. James W. Laferriere

Hello Neil  Luca ,

On Fri, 8 Sep 2006, Luca Berra wrote:

On Fri, Sep 08, 2006 at 02:26:31PM +0200, Lars Schimmer wrote:

-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

Michael Tokarev wrote:

Lars Schimmer wrote:

Hi!

I´ve got a software RAiD5 with 6 250GB HDs.
Now I changed one disk after another to a 400GB HD and resynced the
raid5 after each change.
Now the RAID5 has got 6 400GB HDs and still uses only 6*250GB space.
How can I grow the md0 device to use 6*400GB?


mdadm --grow is your friend.


Oh, damn, right. I was focussed on --grow to add a new HD to the RAID
But isn´t there a switch to grow to max possible value?
Do I always have to search for the biggest value and type it in by hand?

man mdadm

 -z, --size=

This value can be set with --grow for RAID level 1/4/5/6. If the
array  was created with a size smaller than the currently active
drives, the extra space can be accessed using --grow.  The  size
can  be given as max which means to choose the largest size that
fits on all current drives.


Kuca ,  Thank you for posting this snippet .
Neil ,  Might changing ,

 can  be given as max which means to choose the largest size that
To
 can  be given as 'max' which means to choose the largest size that

help those reading this be aware that this is a 'string' to add to the
end of --size= ?  Also if there are other keywords not quoted ('') this
might be a good opertunity .  ;-)  Tia ,  JimL

--
+--+
| James   W.   Laferriere |   SystemTechniques   | Give me VMS |
| NetworkEngineer | 3600 14th Ave SE #20-103 |  Give me Linux  |
| [EMAIL PROTECTED] |  Olympia ,  WA.   98501  |   only  on  AXP |
+--+

Re: Messed up creating new array...

2006-09-08 Thread David Rees

On 9/8/06, Ruth Ivimey-Cook [EMAIL PROTECTED] wrote:

I messed up slightly when creating a new 6-disk raid6 array, and am wondering
if there is a simple answer. The problem is that I didn't partition the drives,
but simply used the whole drive. All drives are of the same type and using the
Supermicro SAT2-MV8 controller.


This should work:

1. Unmount filesystem.
2. Shrink file system to something a bit smaller. Since it's a big
array, 1GB should give you plenty of room.
3. Shrink raid array to something in between the new fs size and old
fs size. Make sure you don't shrink it smaller than the filesystem!
4. Remove a disk from the array (fail/remove)
5. Partition disk
6. Add partition back to array
7. Repeat steps 4-6 for all disks in the array.
8. Now the whole array should be on partitions. Grow the raid array
back to max.
9. Grow filesystem to the partition size.

-Dave
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


UUID's

2006-09-08 Thread Richard Scobie

If I have specified an array in mdadm.conf using UUID's:

ARRAY /dev/md0 UUID=3aaa0122:29827cfa:5331ad66:ca767371

and I replace a failed drive in the array, will the new drive be given 
the previous UUID, or do I need to upate the mdadm.conf entry?


Regards,

Richard
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: UUID's

2006-09-08 Thread dean gaudet


On Sat, 9 Sep 2006, Richard Scobie wrote:

 If I have specified an array in mdadm.conf using UUID's:
 
 ARRAY /dev/md0 UUID=3aaa0122:29827cfa:5331ad66:ca767371
 
 and I replace a failed drive in the array, will the new drive be given the
 previous UUID, or do I need to upate the mdadm.conf entry?

once you do the mdadm /dev/mdX -a /dev/newdrive the new drive will have 
the UUID.  no need to update the mdadm.conf for the UUID...

however if you're using DEVICE foo where foo is not partitions then 
you should make sure foo includes the new drive.  (DEVICE partitions is 
recommended.)

-dean
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: UUID's

2006-09-08 Thread Richard Scobie

dean gaudet wrote:


On Sat, 9 Sep 2006, Richard Scobie wrote:



If I have specified an array in mdadm.conf using UUID's:

ARRAY /dev/md0 UUID=3aaa0122:29827cfa:5331ad66:ca767371

and I replace a failed drive in the array, will the new drive be given the
previous UUID, or do I need to upate the mdadm.conf entry?



once you do the mdadm /dev/mdX -a /dev/newdrive the new drive will have 
the UUID.  no need to update the mdadm.conf for the UUID...


however if you're using DEVICE foo where foo is not partitions then 
you should make sure foo includes the new drive.  (DEVICE partitions is 
recommended.)


Thanks Dean,

I am setting up a Fedora 5 machine, which I configured to use RAID1 for 
all partions during the install and the resulting madadm.conf it 
generated is:


DEVICE partitions
MAILADDR root
ARRAY /dev/md3 super-minor=3
ARRAY /dev/md1 super-minor=1
ARRAY /dev/md4 super-minor=4
ARRAY /dev/md0 super-minor=0
ARRAY /dev/md2 super-minor=2

To remove all doubt about what is assembled where, I though going to:

DEVICE partitions
MAILADDR root
ARRAY /dev/md3 UUID=xyz etc.

would be more secure.

Is this correct thinking on my part?

Regards,

Richard
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Check/repair on composite RAID

2006-09-08 Thread Richard Scobie
If I have a RAID 10, comprising a RAID 0, /dev/md3 made up of RAID1, 
/dev/md1 and RAID1, /dev/md2 and I do an:


echo repair  /sys/block/md3/md/sync_action

will this run simultaneous repairs on the the underlying RAID 1's, or 
should seperate repairs be done to md1 and 2?


Thanks for any assistence.

Regards,

Richard
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


2.6.18-rc5-mm1 - bd_claim_by_disk oops

2006-09-08 Thread John Stoffel

I got the following on 2.6.18-rc5-mm1 when trying to lvextend a test
logical volume that I had just created.  This came about because I
have been trying to expand some LVs on my system, which are based on a
VG ontop of an MD mirror pair.  It's an SMP box too if that means
anything.  

device-mapper: table: 253:3: linear: dm-linear: Device lookup failed
device-mapper: ioctl: error adding target to table
device-mapper: table: 253:3: linear: dm-linear: Device lookup failed
device-mapper: ioctl: error adding target to table
device-mapper: table: 253:2: linear: dm-linear: Device lookup failed
device-mapper: ioctl: error adding target to table
BUG: unable to handle kernel paging request at virtual address
6f72702e
 printing eip:
c018505b
*pde = 
Oops:  [#1]
4K_STACKS SMP 
last sysfs file: /block/dm-2/removable
Modules linked in: rfcomm l2cap bluetooth sbp2 i2c_piix4 ohci1394
evdev snd_ens1
371 snd_rawmidi snd_ac97_codec snd_ac97_bus i2c_core snd_pcm snd_timer
snd sound
core snd_page_alloc ieee1394 uhci_hcd ohci_hcd usb_storage
CPU:1
EIP:0060:[c018505b]Not tainted VLI
EFLAGS: 00010292   (2.6.18-rc5-mm1 #1) 
EIP is at bd_claim_by_disk+0x9b/0x1c0
eax: e6622ba0   ebx: 6f72702e   ecx: c16decbc   edx: 6f72702e
esi:    edi: e4a11220   ebp: c16dec80   esp: c62e1ca4
ds: 007b   es: 007b   ss: 0068
Process lvextend (pid: 2926, ti=c62e1000 task=dccfa580
task.ti=c62e1000)
Stack: c047cefa c16dec8c c16dec80 e3e31920 d1130e80 c16dec80 c035f9c8
fff4 
   e128d160 0090 e3e31920 c036040c f0a5915c c042bf66 c62e1d50
c62e1d4c 
   c0500f0c e128d160 f0a5915c f0a1b080 e128d0c0  0001
c0500f0c 
Call Trace:
 [c035f9c8] open_dev+0x48/0x80
 [c036040c] dm_get_device+0x20c/0x480
 [c0243537] vsscanf+0x447/0x490
 [c0360ca9] linear_ctr+0x99/0x110
 [c035ff0d] dm_table_add_target+0x17d/0x2e0
 [c0361e6a] table_load+0xba/0x1d0
 [c0362b9d] ctl_ioctl+0x22d/0x2a0
 [c028b594] n_tty_receive_buf+0x324/0x1070
 [c0361db0] table_load+0x0/0x1d0
 [c016c2d8] do_ioctl+0x78/0x90
 [c016c34c] vfs_ioctl+0x5c/0x2b0
 [c016c5dd] sys_ioctl+0x3d/0x70
 [c0102f33] syscall_call+0x7/0xb
 ===
Code: 3c 8b 13 0f 18 02 90 8d 4d 3c 39 cb 0f 84 24 01 00 00 8b 47 0c
3b 43 0c 75
 0f e9 ab 00 00 00 90 3b 43 0c 0f 84 a1 00 00 00 89 d3 8b 12 0f 18
02 90 39 cb
 75 eb e8 d6 a9 0b 00 85 c0 89 47 0c 0f 
EIP: [c018505b] bd_claim_by_disk+0x9b/0x1c0 SS:ESP 0068:c62e1ca4
 
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html