Re: In Trouble--Please Help! (was Re: Can't add disk to failed raid array)

2006-07-24 Thread Dan Williams

On 7/23/06, Paul Waldo [EMAIL PROTECTED] wrote:

Here is the dmesg output.  No log files are created with the FC5 rescue disk.
Thanks!

I ran into this as well, I believe at this point you want to set:

md-mod.start_dirty_degraded=1

as part of your boot options.  Understand you may see some filesystem
corruption as noted in the documentation.

See:
http://www.linux-m32r.org/lxr/http/source/Documentation/md.txt?v=2.6.17#L54

Regards,

Dan
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: In Trouble--Please Help! (was Re: Can't add disk to failed raid array)

2006-07-24 Thread Paul Waldo

Dan Williams wrote:

On 7/23/06, Paul Waldo [EMAIL PROTECTED] wrote:
Here is the dmesg output.  No log files are created with the FC5 
rescue disk.

Thanks!

I ran into this as well, I believe at this point you want to set:

md-mod.start_dirty_degraded=1

as part of your boot options.  Understand you may see some filesystem
corruption as noted in the documentation.

See:
http://www.linux-m32r.org/lxr/http/source/Documentation/md.txt?v=2.6.17#L54

Regards,

Dan


I'll certainly give that a try later on, as I need physical access to 
the box.


The corruption part is worrisome...  When you did this, did you 
experience corruption?  I'm running RAID6 with 7 disks; presumably even 
with two disks out of whack, I should be in good shape...???


Paul
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: In Trouble--Please Help! (was Re: Can't add disk to failed raid array)

2006-07-24 Thread Dan Williams

I'll certainly give that a try later on, as I need physical access to
the box.

The corruption part is worrisome...  When you did this, did you
experience corruption?  I'm running RAID6 with 7 disks; presumably even
with two disks out of whack, I should be in good shape...???


I was running a 5 disk RAID-5 and did not detect any corruption.  Neil
correct me if I am wrong, but I believe that since your failure
occured without power loss that the chances for data corruption in
this case are small.

Dan
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


In Trouble--Please Help! (was Re: Can't add disk to failed raid array)

2006-07-23 Thread Paul Waldo
Please, please! I am dead in the water!

To recap, I have a RAID6 array on a Fedora Core 5 system, using /dev/hd[acdeg]
2 and /dev/sd[ab]2.  /dev/hdd went bad so I replaced the drive and tried to 
add it back to the array.  Here is what happens:

#mdadm --assemble /dev/md1 /dev/hd[acdeg]2 /dev/sd[ab]2
mdadm: no RAID superblock on /dev/hdd2
mdadm: /dev/hdd2 has no superblock - assembly aborted

I also tried assembling without the new drive:
#mdadm --assemble /dev/md1 /dev/hd[aceg]2 /dev/sd[ab]2
mdadm: failed to RUN_ARRAY /dev/md1: Input/output error

Forgive me if the messages are not exactly correct, as I am booted into the FC 
rescue disk and I am transcribing what I see to another computer.

How can I get mdadm to use the new drive?  Am I completely fscked?  Thanks in 
advance for any help

Paul


On Tuesday 18 July 2006 01:36, Neil Brown wrote:
 On Sunday July 16, [EMAIL PROTECTED] wrote:
  Thanks for the reply, Neil.  Here is my version:
  [EMAIL PROTECTED] log]# mdadm --version
  mdadm - v2.3.1 - 6 February 2006

 Positively ancient :-)  Nothing obvious in the change log since then.

 Can you show me the output of
   mdadm -E /dev/hdd2
   mdadm -E /dev/hda2

 immediately after the failed attempt to add hdd2.

 Thanks,
 NeilBrown
 -
 To unsubscribe from this list: send the line unsubscribe linux-raid in
 the body of a message to [EMAIL PROTECTED]
 More majordomo info at  http://vger.kernel.org/majordomo-info.html
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: In Trouble--Please Help! (was Re: Can't add disk to failed raid array)

2006-07-23 Thread Neil Brown
On Sunday July 23, [EMAIL PROTECTED] wrote:
 Please, please! I am dead in the water!
 
 To recap, I have a RAID6 array on a Fedora Core 5 system, using /dev/hd[acdeg]
 2 and /dev/sd[ab]2.  /dev/hdd went bad so I replaced the drive and tried to 
 add it back to the array.  Here is what happens:
 
 #mdadm --assemble /dev/md1 /dev/hd[acdeg]2 /dev/sd[ab]2
 mdadm: no RAID superblock on /dev/hdd2
 mdadm: /dev/hdd2 has no superblock - assembly aborted
 
 I also tried assembling without the new drive:
 #mdadm --assemble /dev/md1 /dev/hd[aceg]2 /dev/sd[ab]2
 mdadm: failed to RUN_ARRAY /dev/md1: Input/output error
 
 Forgive me if the messages are not exactly correct, as I am booted into the 
 FC 
 rescue disk and I am transcribing what I see to another computer.
 
 How can I get mdadm to use the new drive?  Am I completely fscked?  Thanks in 
 advance for any help

Sorry for not following through on this earlier.
I think I know what the problem is.

 From the -E output you have me :
 Device Size : 155308288 (148.11 GiB 159.04 GB)

 From the fdisk output in the original email:

[EMAIL PROTECTED] ~]# fdisk -l /dev/hdd

Disk /dev/hdd: 160.0 GB, 16002616 bytes
255 heads, 63 sectors/track, 19455 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device Boot  Start End  Blocks   Id  System
/dev/hdd1   1 122  979933+  fd  Linux raid autodetect
/dev/hdd2 123   19455   155292322+  fd  Linux raid autodetect


Notice that hdd2 is 155292322, but the device needs to be 
155308288.  It isn't quite big enough.
mdadm should pick this up, but obviously doesn't, and the error message
isn't at all helpful.
I will fix that in the next release.
For now, repartition your drive so that hdd2 is slightly larger.

NeilBrown
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: In Trouble--Please Help! (was Re: Can't add disk to failed raid array)

2006-07-23 Thread Paul Waldo
On Sunday 23 July 2006 07:25, Neil Brown wrote:
 On Sunday July 23, [EMAIL PROTECTED] wrote:
  Please, please! I am dead in the water!
 
  To recap, I have a RAID6 array on a Fedora Core 5 system, using
  /dev/hd[acdeg] 2 and /dev/sd[ab]2.  /dev/hdd went bad so I replaced the
  drive and tried to add it back to the array.  Here is what happens:
 
  #mdadm --assemble /dev/md1 /dev/hd[acdeg]2 /dev/sd[ab]2
  mdadm: no RAID superblock on /dev/hdd2
  mdadm: /dev/hdd2 has no superblock - assembly aborted
 
  I also tried assembling without the new drive:
  #mdadm --assemble /dev/md1 /dev/hd[aceg]2 /dev/sd[ab]2
  mdadm: failed to RUN_ARRAY /dev/md1: Input/output error
 
  Forgive me if the messages are not exactly correct, as I am booted into
  the FC rescue disk and I am transcribing what I see to another computer.
 
  How can I get mdadm to use the new drive?  Am I completely fscked? 
  Thanks in advance for any help

 Sorry for not following through on this earlier.
 I think I know what the problem is.

  From the -E output you have me :
  Device Size : 155308288 (148.11 GiB 159.04 GB)

  From the fdisk output in the original email:

 [EMAIL PROTECTED] ~]# fdisk -l /dev/hdd

 Disk /dev/hdd: 160.0 GB, 16002616 bytes
 255 heads, 63 sectors/track, 19455 cylinders
 Units = cylinders of 16065 * 512 = 8225280 bytes

Device Boot  Start End  Blocks   Id  System
 /dev/hdd1   1 122  979933+  fd  Linux raid
 autodetect /dev/hdd2 123   19455   155292322+  fd  Linux
 raid autodetect


 Notice that hdd2 is 155292322, but the device needs to be
 155308288.  It isn't quite big enough.
 mdadm should pick this up, but obviously doesn't, and the error message
 isn't at all helpful.
 I will fix that in the next release.
 For now, repartition your drive so that hdd2 is slightly larger.

 NeilBrown

Hi Neil,

Thanks for the quick reply.  I tried your suggestion.  I now see 
that /dev/hda2 (good partition) is 155308387 blocks.  I 
repartitioned /dev/hdd such that /dev/hdd2 is also 155308387, but I still get 
the message about no superblock on hdd2 :-(.  I used --force.

Paul
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: In Trouble--Please Help! (was Re: Can't add disk to failed raid array)

2006-07-23 Thread Paul Waldo
On Sunday 23 July 2006 07:25, Neil Brown wrote:
 On Sunday July 23, [EMAIL PROTECTED] wrote:
  Please, please! I am dead in the water!
 
  To recap, I have a RAID6 array on a Fedora Core 5 system, using
  /dev/hd[acdeg] 2 and /dev/sd[ab]2.  /dev/hdd went bad so I replaced the
  drive and tried to add it back to the array.  Here is what happens:
 
  #mdadm --assemble /dev/md1 /dev/hd[acdeg]2 /dev/sd[ab]2
  mdadm: no RAID superblock on /dev/hdd2
  mdadm: /dev/hdd2 has no superblock - assembly aborted
 
  I also tried assembling without the new drive:
  #mdadm --assemble /dev/md1 /dev/hd[aceg]2 /dev/sd[ab]2
  mdadm: failed to RUN_ARRAY /dev/md1: Input/output error
 
  Forgive me if the messages are not exactly correct, as I am booted into
  the FC rescue disk and I am transcribing what I see to another computer.
 
  How can I get mdadm to use the new drive?  Am I completely fscked? 
  Thanks in advance for any help

 Sorry for not following through on this earlier.
 I think I know what the problem is.

  From the -E output you have me :
  Device Size : 155308288 (148.11 GiB 159.04 GB)

  From the fdisk output in the original email:

 [EMAIL PROTECTED] ~]# fdisk -l /dev/hdd

 Disk /dev/hdd: 160.0 GB, 16002616 bytes
 255 heads, 63 sectors/track, 19455 cylinders
 Units = cylinders of 16065 * 512 = 8225280 bytes

Device Boot  Start End  Blocks   Id  System
 /dev/hdd1   1 122  979933+  fd  Linux raid
 autodetect /dev/hdd2 123   19455   155292322+  fd  Linux
 raid autodetect


 Notice that hdd2 is 155292322, but the device needs to be
 155308288.  It isn't quite big enough.
 mdadm should pick this up, but obviously doesn't, and the error message
 isn't at all helpful.
 I will fix that in the next release.
 For now, repartition your drive so that hdd2 is slightly larger.

 NeilBrown


At this point, I'd just be happy to be able to get the degraded array back up 
and running.  Is there any way to do that?  Thanks!

Paul
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: In Trouble--Please Help! (was Re: Can't add disk to failed raid array)

2006-07-23 Thread Neil Brown
On Sunday July 23, [EMAIL PROTECTED] wrote:
 
 
 At this point, I'd just be happy to be able to get the degraded array back up 
 and running.  Is there any way to do that?  Thanks!

 mdadm --assemble --force /dev/md1 /dev/hd[aceg]2 /dev/sd[ab]2

should get you the degraded array.  If not, what kernel log messages
do you get?


-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: In Trouble--Please Help! (was Re: Can't add disk to failed raid array)

2006-07-23 Thread Paul Waldo
Here is the dmesg output.  No log files are created with the FC5 rescue disk.  
Thanks!



On Sunday 23 July 2006 07:59, Neil Brown wrote:
 On Sunday July 23, [EMAIL PROTECTED] wrote:
  At this point, I'd just be happy to be able to get the degraded array
  back up and running.  Is there any way to do that?  Thanks!

  mdadm --assemble --force /dev/md1 /dev/hd[aceg]2 /dev/sd[ab]2

 should get you the degraded array.  If not, what kernel log messages
 do you get?
[APCG] enabled at IRQ 20
ACPI: PCI Interrupt :00:02.1[B] - Link [APCG] - GSI 20 (level, high) - IRQ 19
PCI: Setting latency timer of device :00:02.1 to 64
ohci_hcd :00:02.1: OHCI Host Controller
ohci_hcd :00:02.1: new USB bus registered, assigned bus number 3
ohci_hcd :00:02.1: irq 19, io mem 0xe2082000
usb usb3: configuration #1 chosen from 1 choice
hub 3-0:1.0: USB hub found
hub 3-0:1.0: 3 ports detected
usb 1-3: configuration #1 chosen from 1 choice
Initializing USB Mass Storage driver...
scsi0 : SCSI emulation for USB Mass Storage devices
usb-storage: device found at 2
usb-storage: waiting for device to settle before scanning
usbcore: registered new driver usb-storage
USB Mass Storage support registered.
  Vendor: HL-DT-ST  Model: DVDRRW GSA-2166D  Rev: 1.01
  Type:   CD-ROM ANSI SCSI revision: 00
sr0: scsi3-mmc drive: 48x/48x writer dvd-ram cd/rw xa/form2 cdda tray
Uniform CD-ROM driver Revision: 3.20
sr 0:0:0:0: Attached scsi CD-ROM sr0
usb-storage: device scan complete
libata version 1.20 loaded.
sata_sil :01:0d.0: version 0.9
ACPI: PCI Interrupt Link [APC3] enabled at IRQ 18
ACPI: PCI Interrupt :01:0d.0[A] - Link [APC3] - GSI 18 (level, high) - IRQ 20
ata1: SATA max UDMA/100 cmd 0xE090C080 ctl 0xE090C08A bmdma 0xE090C000 irq 20
ata2: SATA max UDMA/100 cmd 0xE090C0C0 ctl 0xE090C0CA bmdma 0xE090C008 irq 20
ata1: SATA link up 1.5 Gbps (SStatus 113)
ata1: dev 0 cfg 49:2f00 82:346b 83:7f01 84:4003 85:3c69 86:3c01 87:4003 88:20ff
ata1: dev 0 ATA-7, max UDMA7, 312581808 sectors: LBA48
ata1: dev 0 configured for UDMA/100
scsi1 : sata_sil
ata2: SATA link up 1.5 Gbps (SStatus 113)
ata2: dev 0 cfg 49:2f00 82:346b 83:7f01 84:4003 85:3c69 86:3c01 87:4003 88:20ff
ata2: dev 0 ATA-7, max UDMA7, 312581808 sectors: LBA48
ata2: dev 0 configured for UDMA/100
scsi2 : sata_sil
  Vendor: ATA   Model: SAMSUNG SP1614C   Rev: SW10
  Type:   Direct-Access  ANSI SCSI revision: 05
SCSI device sda: 312581808 512-byte hdwr sectors (160042 MB)
sda: Write Protect is off
sda: Mode Sense: 00 3a 00 00
SCSI device sda: drive cache: write back
SCSI device sda: 312581808 512-byte hdwr sectors (160042 MB)
sda: Write Protect is off
sda: Mode Sense: 00 3a 00 00
SCSI device sda: drive cache: write back
 sda: sda1 sda2
sd 1:0:0:0: Attached scsi disk sda
  Vendor: ATA   Model: SAMSUNG SP1614C   Rev: SW10
  Type:   Direct-Access  ANSI SCSI revision: 05
SCSI device sdb: 312581808 512-byte hdwr sectors (160042 MB)
sdb: Write Protect is off
sdb: Mode Sense: 00 3a 00 00
SCSI device sdb: drive cache: write back
SCSI device sdb: 312581808 512-byte hdwr sectors (160042 MB)
sdb: Write Protect is off
sdb: Mode Sense: 00 3a 00 00
SCSI device sdb: drive cache: write back
 sdb: sdb1 sdb2
sd 2:0:0:0: Attached scsi disk sdb
forcedeth.c: Reverse Engineered nForce ethernet driver. Version 0.49.
ACPI: PCI Interrupt Link [APCH] enabled at IRQ 22
ACPI: PCI Interrupt :00:04.0[A] - Link [APCH] - GSI 22 (level, high) - IRQ 17
PCI: Setting latency timer of device :00:04.0 to 64
eth0: forcedeth.c: subsystem: 010de:05b2 bound to :00:04.0
ISO 9660 Extensions: Microsoft Joliet Level 3
Unable to load NLS charset utf8
Unable to load NLS charset utf8
ISO 9660 Extensions: RRIP_1991A
Unable to identify CD-ROM format.
VFS: Can't find an ext2 filesystem on dev loop0.
security:  3 users, 6 roles, 1161 types, 135 bools, 1 sens, 256 cats
security:  55 classes, 38679 rules
SELinux:  Completing initialization.
SELinux:  Setting up existing superblocks.
SELinux: initialized (dev loop0, type squashfs), not configured for labeling
SELinux: initialized (dev usbfs, type usbfs), uses genfs_contexts
SELinux: initialized (dev ramfs, type ramfs), uses genfs_contexts
SELinux: initialized (dev ramfs, type ramfs), uses genfs_contexts
SELinux: initialized (dev tmpfs, type tmpfs), uses transition SIDs
SELinux: initialized (dev debugfs, type debugfs), uses genfs_contexts
SELinux: initialized (dev selinuxfs, type selinuxfs), uses genfs_contexts
SELinux: initialized (dev mqueue, type mqueue), uses transition SIDs
SELinux: initialized (dev hugetlbfs, type hugetlbfs), uses genfs_contexts
SELinux: initialized (dev devpts, type devpts), uses transition SIDs
SELinux: initialized (dev eventpollfs, type eventpollfs), uses genfs_contexts
SELinux: initialized (dev inotifyfs, type inotifyfs), uses genfs_contexts
SELinux: initialized (dev tmpfs, type tmpfs), uses transition SIDs
SELinux: initialized (dev futexfs, type futexfs), uses