RE: FW: Ugh . . .

1999-12-20 Thread Eric Jorgensen


*sigh* raidtools on the system in question reports that it is V0.90. Well,
specifically --

-
[root@charlotte /]# /sbin/raidstart --version
/sbin/raidstart v0.3d compiled for md raidtools-0.90
-

 - Eric

 -Original Message-
 From: Jakob Østergaard [mailto:[EMAIL PROTECTED]]
 Sent: Friday, December 17, 1999 6:34 PM
 To: Eric Jorgensen
 Cc: [EMAIL PROTECTED]
 Subject: Re: FW: Ugh . . .


 On Fri, Dec 17, 1999 at 04:32:14PM -0700, Eric Jorgensen wrote:
 

 My last mail to you bounced, so I'm trying again.  Sorry if anyone
 gets this twice...

 On Fri, Dec 17, 1999 at 03:41:37PM -0700, Eric Jorgensen wrote:
  Well, I tried, something seems to be wrong. I had to update raidtools to
  include the failed-disk directive. that took a while to figure
 out. someone
  needs to tap linuxdoc.org on the shoulder and inform them their
  software-raid-howto is painfully out of date. I'd do it myself
 but there are
  too many blunt objects handy.

 http://ostenfeld.dk/~jakob/Software-RAID.HOWTO/  is the place for
 the current
 0.90 software RAID howto.

  ANYway, here's what happens. Sensitive argument replaced per request.
 
  
  [root@charlotte /root]# ./mkraid --truly-foolish /dev/md0
  DESTROYING the contents of /dev/md0 in 5 seconds, Ctrl-C if unsure!
  handling MD device /dev/md0
  analyzing super-block
  disk 0: /dev/sdd1, 8883913kB, raid superblock at 8883840kB
  disk 1: /dev/sde1, 8964238kB, raid superblock at 8964160kB
  disk 2: /dev/sdb1, failed
  disk 3: /dev/sdc1, 8883913kB, raid superblock at 8883840kB
  mkraid: aborted, see the syslog and /proc/mdstat for potential clues.
 
  ---
 
And here's what dmesg reveals:
 
  ---
 
  bindsdd1,1
  bindsde1,2
  blkdev_open() failed: -19
  md: [dev 00:00] has zero size, marking faulty!
  md: error, md_import_device() returned -22
 
  ---
 
  And here's my raidtab. Sorry for the confusion, sdb is visually
 marked "4"
  on the front of the case. Longer story.

 Gosh, something is just coming to my mind here...  I was
 convinced that you
 were running 0.90 RAID, since most people posting on the list are  (stupid
 assumptions come easy).  But I guess you aren't...   Right ?

 You're running a kernel with standard RAID, not an -ac or
 raid-patched kernel I
 guess...   That means the new raidtools (which understand
 "failed-disk") will
 not talk to your kernel.

 I see one way out:  Patch your old raidtools (version 0.42 or so ?) to
 understand the failed-disk directive.   This may involve manual
 inclusion of
 some patch rejects.  Maybe not.  Don't know.

 If I'm really right that you're running the old code, you probably want to
 upgrade to 0.90 once your data are back   :)   The new code is
 stable, and the
 old code isn't  (you can usually crash a RAID-5 box by stressing
 the RAID with
 the old code).


 Another way out would involve upgrading your old array to the new
 format, using
 the --upgrade switch, but I doubt that it is a very clever thing
 to do with the
 current state of your array...

 The failed-disk patch is fairly small. I guess you can apply it
 pretty quickly
 even if it doesn't apply cleanly to the older raidtools.


 --
 
 : [EMAIL PROTECTED]  : And I see the elder races, :
 :.: putrid forms of man:
 :   Jakob Østergaard  : See him rise and claim the earth,  :
 :OZ9ABN   : his downfall is at hand.   :
 :.:{Konkhra}...:




Re: FW: Ugh . . .

1999-12-18 Thread Eric Jorgensen



That would be a correct assesment. I've gotten the new howto, and new 
raidtools - which kernel should i be using, or who's patch should i apply to standard 
kernel source? 

Sorry about the bounce, momentary sendmail misconfiguration while upgrading 
email server. 

 - Eric

-- Original Message --
From: Jakob Østergaard [EMAIL PROTECTED]
Date: Sat, 18 Dec 1999 02:34:03 +0100

On Fri, Dec 17, 1999 at 04:32:14PM -0700, Eric Jorgensen wrote:
 

My last mail to you bounced, so I'm trying again.  Sorry if anyone
gets this twice...

On Fri, Dec 17, 1999 at 03:41:37PM -0700, Eric Jorgensen wrote:
 Well, I tried, something seems to be wrong. I had to update raidtools to
 include the failed-disk directive. that took a while to figure out. someone
 needs to tap linuxdoc.org on the shoulder and inform them their
 software-raid-howto is painfully out of date. I'd do it myself but there are
 too many blunt objects handy.

http://ostenfeld.dk/~jakob/Software-RAID.HOWTO/  is the place for the current
0.90 software RAID howto.

 ANYway, here's what happens. Sensitive argument replaced per request.
 
 
 [root@charlotte /root]# ./mkraid --truly-foolish /dev/md0
 DESTROYING the contents of /dev/md0 in 5 seconds, Ctrl-C if unsure!
 handling MD device /dev/md0
 analyzing super-block
 disk 0: /dev/sdd1, 8883913kB, raid superblock at 8883840kB
 disk 1: /dev/sde1, 8964238kB, raid superblock at 8964160kB
 disk 2: /dev/sdb1, failed
 disk 3: /dev/sdc1, 8883913kB, raid superblock at 8883840kB
 mkraid: aborted, see the syslog and /proc/mdstat for potential clues.
 
 ---
 
   And here's what dmesg reveals:
 
 ---
 
 bindsdd1,1
 bindsde1,2
 blkdev_open() failed: -19
 md: [dev 00:00] has zero size, marking faulty!
 md: error, md_import_device() returned -22
 
 ---
 
 And here's my raidtab. Sorry for the confusion, sdb is visually marked "4"
 on the front of the case. Longer story.

Gosh, something is just coming to my mind here...  I was convinced that you
were running 0.90 RAID, since most people posting on the list are  (stupid
assumptions come easy).  But I guess you aren't...   Right ?

You're running a kernel with standard RAID, not an -ac or raid-patched kernel I
guess...   That means the new raidtools (which understand "failed-disk") will
not talk to your kernel.

I see one way out:  Patch your old raidtools (version 0.42 or so ?) to
understand the failed-disk directive.   This may involve manual inclusion of
some patch rejects.  Maybe not.  Don't know.

If I'm really right that you're running the old code, you probably want to
upgrade to 0.90 once your data are back   :)   The new code is stable, and the
old code isn't  (you can usually crash a RAID-5 box by stressing the RAID with
the old code).


Another way out would involve upgrading your old array to the new format, using
the --upgrade switch, but I doubt that it is a very clever thing to do with the
current state of your array...

The failed-disk patch is fairly small. I guess you can apply it pretty quickly
even if it doesn't apply cleanly to the older raidtools.


-- 

: [EMAIL PROTECTED]  : And I see the elder races, :
:.: putrid forms of man:
:   Jakob Østergaard  : See him rise and claim the earth,  :
:OZ9ABN   : his downfall is at hand.   :
:.:{Konkhra}...:




Re: FW: Ugh . . .

1999-12-18 Thread Jakob Østergaard

On Sat, Dec 18, 1999 at 09:13:39AM -0700, Eric Jorgensen wrote:
 
 
   That would be a correct assesment. I've gotten the new howto, and new 
raidtools - which kernel should i be using, or who's patch should i apply to standard 
kernel source? 

Problem is, if your current array was built with the old raidtools/kernel, you should
not upgrade the raidtools/kernel before the recovery, but instead get old raidtools to
go with your current kernel and array.  Problem is that the failed-disk patch is for
the newer raidtools.  Apply the new patch to the old raidtools (manually if required).

-- 

: [EMAIL PROTECTED]  : And I see the elder races, :
:.: putrid forms of man:
:   Jakob Østergaard  : See him rise and claim the earth,  :
:OZ9ABN   : his downfall is at hand.   :
:.:{Konkhra}...:



FW: Ugh . . .

1999-12-17 Thread Eric Jorgensen



-Original Message-
From: Eric Jorgensen [mailto:[EMAIL PROTECTED]]
Sent: Friday, December 17, 1999 3:42 PM
To: Jakob Østergaard
Subject: RE: Ugh . . .


Well, I tried, something seems to be wrong. I had to update raidtools to
include the failed-disk directive. that took a while to figure out. someone
needs to tap linuxdoc.org on the shoulder and inform them their
software-raid-howto is painfully out of date. I'd do it myself but there are
too many blunt objects handy.

ANYway, here's what happens. Sensitive argument replaced per request.


[root@charlotte /root]# ./mkraid --truly-foolish /dev/md0
DESTROYING the contents of /dev/md0 in 5 seconds, Ctrl-C if unsure!
handling MD device /dev/md0
analyzing super-block
disk 0: /dev/sdd1, 8883913kB, raid superblock at 8883840kB
disk 1: /dev/sde1, 8964238kB, raid superblock at 8964160kB
disk 2: /dev/sdb1, failed
disk 3: /dev/sdc1, 8883913kB, raid superblock at 8883840kB
mkraid: aborted, see the syslog and /proc/mdstat for potential clues.

---

And here's what dmesg reveals:

---

bindsdd1,1
bindsde1,2
blkdev_open() failed: -19
md: [dev 00:00] has zero size, marking faulty!
md: error, md_import_device() returned -22

---

And here's my raidtab. Sorry for the confusion, sdb is visually marked "4"
on the front of the case. Longer story.

---

raiddev /dev/md0
raid-level  5
nr-raid-disks   4
nr-spare-disks  0
chunk-size  512

device  /dev/sdd1
raid-disk   0
device  /dev/sde1
raid-disk   1
device  /dev/sdb1
failed-disk 2
device  /dev/sdc1
raid-disk   3




Thanks, very much, for any help at this time.

 - Eric

 -Original Message-
 From: Jakob Østergaard [mailto:[EMAIL PROTECTED]]
 Sent: Friday, December 17, 1999 11:30 AM
 To: Eric Jorgensen
 Subject: Re: Ugh . . .


 On Fri, Dec 17, 1999 at 11:10:42AM -0700, Eric Jorgensen wrote:
I have, well, had, 4 drives in a raid 5 array. drive 4 went
   south, drive 3
is desynced.
  
   What ?  How ?  You lost two drives in your RAID-5 ?
 
  Box crashed for other reasons while it was in degraded
 mode. Long story.
  Wish RedHat had thought to include LOUD alarms for degraded drives.

 If it has been in degraded mode for long, you won't save all your
 data, or more
 specifically, all you data will be saved but some of them will be older
 versions.  That might or might not work well with e2fsck.

 
  Ahh, sounds rational. I knew i was missing something,
 thanks for pointing
  out manually ensuring the blank drive is failed out.

 If it works out (and nobody gets to scream STOP in time ;)
 please let me know.

 
   PleasePlease:  Before you actually do this, wait for a little
   while to allow someone
   on this list to jump in and say "DON'T!", if appropriate.
 
  Agreed.
 
  Thanks very much. Lemme know if you need any web hosting
 space (heh heh .
  . )

 Thanks, I think I'm covered for now, but I'll keep it in mind;)

 Cheers,
 --
 
 : [EMAIL PROTECTED]  : And I see the elder races, :
 :.: putrid forms of man:
 :   Jakob Østergaard  : See him rise and claim the earth,  :
 :OZ9ABN   : his downfall is at hand.   :
 :.:{Konkhra}...:




Re: FW: Ugh . . .

1999-12-17 Thread Jakob Østergaard

On Fri, Dec 17, 1999 at 04:32:14PM -0700, Eric Jorgensen wrote:
 

My last mail to you bounced, so I'm trying again.  Sorry if anyone
gets this twice...

On Fri, Dec 17, 1999 at 03:41:37PM -0700, Eric Jorgensen wrote:
 Well, I tried, something seems to be wrong. I had to update raidtools to
 include the failed-disk directive. that took a while to figure out. someone
 needs to tap linuxdoc.org on the shoulder and inform them their
 software-raid-howto is painfully out of date. I'd do it myself but there are
 too many blunt objects handy.

http://ostenfeld.dk/~jakob/Software-RAID.HOWTO/  is the place for the current
0.90 software RAID howto.

 ANYway, here's what happens. Sensitive argument replaced per request.
 
 
 [root@charlotte /root]# ./mkraid --truly-foolish /dev/md0
 DESTROYING the contents of /dev/md0 in 5 seconds, Ctrl-C if unsure!
 handling MD device /dev/md0
 analyzing super-block
 disk 0: /dev/sdd1, 8883913kB, raid superblock at 8883840kB
 disk 1: /dev/sde1, 8964238kB, raid superblock at 8964160kB
 disk 2: /dev/sdb1, failed
 disk 3: /dev/sdc1, 8883913kB, raid superblock at 8883840kB
 mkraid: aborted, see the syslog and /proc/mdstat for potential clues.
 
 ---
 
   And here's what dmesg reveals:
 
 ---
 
 bindsdd1,1
 bindsde1,2
 blkdev_open() failed: -19
 md: [dev 00:00] has zero size, marking faulty!
 md: error, md_import_device() returned -22
 
 ---
 
 And here's my raidtab. Sorry for the confusion, sdb is visually marked "4"
 on the front of the case. Longer story.

Gosh, something is just coming to my mind here...  I was convinced that you
were running 0.90 RAID, since most people posting on the list are  (stupid
assumptions come easy).  But I guess you aren't...   Right ?

You're running a kernel with standard RAID, not an -ac or raid-patched kernel I
guess...   That means the new raidtools (which understand "failed-disk") will
not talk to your kernel.

I see one way out:  Patch your old raidtools (version 0.42 or so ?) to
understand the failed-disk directive.   This may involve manual inclusion of
some patch rejects.  Maybe not.  Don't know.

If I'm really right that you're running the old code, you probably want to
upgrade to 0.90 once your data are back   :)   The new code is stable, and the
old code isn't  (you can usually crash a RAID-5 box by stressing the RAID with
the old code).


Another way out would involve upgrading your old array to the new format, using
the --upgrade switch, but I doubt that it is a very clever thing to do with the
current state of your array...

The failed-disk patch is fairly small. I guess you can apply it pretty quickly
even if it doesn't apply cleanly to the older raidtools.


-- 

: [EMAIL PROTECTED]  : And I see the elder races, :
:.: putrid forms of man:
:   Jakob Østergaard  : See him rise and claim the earth,  :
:OZ9ABN   : his downfall is at hand.   :
:.:{Konkhra}...: