raid1 boot issues

2005-08-25 Thread Lewis Shobbrook
Hi All,
I have a problem attempting to boot a raid 1 system from lilo.  I have 
succeeded at this many times in the past, but am completely stumped as to 
what the issue is in this instance. I have the boot and root partitions 
seperate on /dev/md0  /dev/md1 respectively, both raid1.  
The mdadm examination  for  the components is clean and the superblocks appear 
as they ought.  I'm using Debian unstable with std. apt kernel-image.  The 
correct modules are in place for the initrd.
The boot partition fails to mount during  the boot process...
fsck.ext3: invalid argument while trying to open /dev/md0
The superblock can not be read or does not describe a correct ext2 
filesystem omitted usual stuff...
Root password for maintenance or Control-D to continue...

The bizarre thing is that it appears perfectly clean, I've re-zero'd the 
superblocks and completely recreated the device, but the result is always the 
same.  
mdrun loads /dev/md0 in a clean state straight away and it mounts 
cleanly.  /dev/md1 is no problem. 
mdadm -E of the components is clean, as is mdadm -d for the device.
My fstab  mtab are the same as systems running the same kernel that work 
fine.

I found the system laying around from about 12 months ago which had originally 
been set-up using raidtools.  I upgraded the system using dist-upgrade 
installed mdadm; after zeroing all superblocks for both drives and component 
partitions, I created the devices with the following ...

mdadm -C /dev/md0 -l1 -n2 /dev/hda1 missing
reformatted ext3 and restored the files before adding the missing devices and 
resyncing.  
I'm stumped!
Anyone got any ideas?

Cheers,

Lewis
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: raid1 boot issues

2005-08-25 Thread Tyler

Fdisk it and set partitions to Raid Autodetect (0xfd) possibly?

Tyler.

Lewis Shobbrook wrote:


Hi All,
I have a problem attempting to boot a raid 1 system from lilo.  I have 
succeeded at this many times in the past, but am completely stumped as to 
what the issue is in this instance. I have the boot and root partitions 
seperate on /dev/md0  /dev/md1 respectively, both raid1.  
The mdadm examination  for  the components is clean and the superblocks appear 
as they ought.  I'm using Debian unstable with std. apt kernel-image.  The 
correct modules are in place for the initrd.

The boot partition fails to mount during  the boot process...
fsck.ext3: invalid argument while trying to open /dev/md0
The superblock can not be read or does not describe a correct ext2 
filesystem omitted usual stuff...

Root password for maintenance or Control-D to continue...

The bizarre thing is that it appears perfectly clean, I've re-zero'd the 
superblocks and completely recreated the device, but the result is always the 
same.  
mdrun loads /dev/md0 in a clean state straight away and it mounts 
cleanly.  /dev/md1 is no problem. 
mdadm -E of the components is clean, as is mdadm -d for the device.
My fstab  mtab are the same as systems running the same kernel that work 
fine.


I found the system laying around from about 12 months ago which had originally 
been set-up using raidtools.  I upgraded the system using dist-upgrade 
installed mdadm; after zeroing all superblocks for both drives and component 
partitions, I created the devices with the following ...


mdadm -C /dev/md0 -l1 -n2 /dev/hda1 missing
reformatted ext3 and restored the files before adding the missing devices and 
resyncing.  
I'm stumped!

Anyone got any ideas?

Cheers,

Lewis
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

 




--
No virus found in this outgoing message.
Checked by AVG Anti-Virus.
Version: 7.0.344 / Virus Database: 267.10.15/80 - Release Date: 8/23/2005

-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: RAID 5 write performance advice

2005-08-25 Thread Mirko Benz

Hello,

We intend to export a lvm/md volume via iSCSI or SRP using InfiniBand to 
remote clients. There is no local file system processing on the storage 
platform. The clients may have a variety of file systems including ext3, 
GFS.


Single disk write performance is: 58,5 MB/s. With large sequential write 
operations I would expect something like 90% of n-1 * 
single_disk_performance if stripe write can be utilized. So roughly 400 
MB/s – which the HW RAID devices achieve.


RAID setup:
Personalities : [raid0] [raid5]
md0 : active raid5 sdi[7] sdh[6] sdg[5] sdf[4] sde[3] sdd[2] sdc[1] sdb[0]
1094035712 blocks level 5, 64k chunk, algorithm 2 [8/8] []

We have assigned the deadline scheduler to every disk in the RAID. The 
default scheduler gives much lower results.


*** dd TEST ***

time dd if=/dev/zero of=/dev/md0 bs=1M
5329911808 bytes transferred in 28,086199 seconds (189769779 bytes/sec)

iostat 5 output:
avg-cpu: %user %nice %sys %iowait %idle
0,10 0,00 87,80 7,30 4,80

Device: tps Blk_read/s Blk_wrtn/s Blk_read Blk_wrtn
hda 0,00 0,00 0,00 0 0
sda 0,00 0,00 0,00 0 0
sdb 1976,10 1576,10 53150,60 7912 266816
sdc 2072,31 1478,88 53150,60 7424 266816
sdd 2034,06 1525,10 53150,60 7656 266816
sde 1988,05 1439,04 53147,41 7224 266800
sdf 1975,10 1499,60 53147,41 7528 266800
sdg 1383,07 1485,26 53145,82 7456 266792
sdh 1562,55 1311,55 53145,82 6584 266792
sdi 1586,85 1295,62 53145,82 6504 266792
sdj 0,00 0,00 0,00 0 0
sdk 0,00 0,00 0,00 0 0
sdl 0,00 0,00 0,00 0 0
sdm 0,00 0,00 0,00 0 0
sdn 0,00 0,00 0,00 0 0
md0 46515,54 0,00 372124,30 0 1868064

Comments: Large write should not see any read operations. But there are 
some???



*** disktest ***

disktest -w -PT -T30 -h1 -K8 -B512k -ID /dev/md0

| 2005/08/25-17:27:04 | STAT | 4072 | v1.1.12 | /dev/md0 | Write 
throughput: 160152507.7B/s (152.73MB/s), IOPS 305.7/s.
| 2005/08/25-17:27:05 | STAT | 4072 | v1.1.12 | /dev/md0 | Write 
throughput: 160694272.0B/s (153.25MB/s), IOPS 306.6/s.
| 2005/08/25-17:27:06 | STAT | 4072 | v1.1.12 | /dev/md0 | Write 
throughput: 160339606.6B/s (152.91MB/s), IOPS 305.8/s.


iostat 5 output:
avg-cpu: %user %nice %sys %iowait %idle
38,96 0,00 50,25 5,29 5,49

Device: tps Blk_read/s Blk_wrtn/s Blk_read Blk_wrtn
hda 0,00 0,00 0,00 0 0
sda 1,20 0,00 11,18 0 56
sdb 986,43 0,00 39702,99 0 198912
sdc 922,75 0,00 39728,54 0 199040
sdd 895,81 0,00 39728,54 0 199040
sde 880,84 0,00 39728,54 0 199040
sdf 839,92 0,00 39728,54 0 199040
sdg 842,91 0,00 39728,54 0 199040
sdh 1557,49 0,00 79431,54 0 397952
sdi 2246,71 0,00 104411,98 0 523104
sdj 0,00 0,00 0,00 0 0
sdk 0,00 0,00 0,00 0 0
sdl 0,00 0,00 0,00 0 0
sdm 0,00 0,00 0,00 0 0
sdn 0,00 0,00 0,00 0 0
md0 1550,70 0,00 317574,45 0 1591048

Comments:
Zero read requests – as it should be. But the write requests are not 
proportional. sdh and sdi have significantly more requests???
The write requests to the disks of the RAID should be 1/7 higher than to 
the md device.

But there are significantly more write operations.

All these operations are to the raw device. Setting up a ext3 fs we get 
around 127 MB/s with dd.


Any idea?

--Mirko

-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: RAID 5 write performance advice

2005-08-25 Thread Ming Zhang
On Thu, 2005-08-25 at 18:38 +0200, Mirko Benz wrote:
 Hello,
 
 We intend to export a lvm/md volume via iSCSI or SRP using InfiniBand to 
 remote clients. There is no local file system processing on the storage 
 platform. The clients may have a variety of file systems including ext3, 
 GFS.
 
 Single disk write performance is: 58,5 MB/s. With large sequential write 
 operations I would expect something like 90% of n-1 * 
 single_disk_performance if stripe write can be utilized. So roughly 400 
 MB/s – which the HW RAID devices achieve.

change to RAID0 and test to see if u controller will be a bottleneck.



 
 RAID setup:
 Personalities : [raid0] [raid5]
 md0 : active raid5 sdi[7] sdh[6] sdg[5] sdf[4] sde[3] sdd[2] sdc[1] sdb[0]
 1094035712 blocks level 5, 64k chunk, algorithm 2 [8/8] []
 
 We have assigned the deadline scheduler to every disk in the RAID. The 
 default scheduler gives much lower results.
 
 *** dd TEST ***
 
 time dd if=/dev/zero of=/dev/md0 bs=1M
 5329911808 bytes transferred in 28,086199 seconds (189769779 bytes/sec)
 
 iostat 5 output:
 avg-cpu: %user %nice %sys %iowait %idle
 0,10 0,00 87,80 7,30 4,80
 
 Device: tps Blk_read/s Blk_wrtn/s Blk_read Blk_wrtn
 hda 0,00 0,00 0,00 0 0
 sda 0,00 0,00 0,00 0 0
 sdb 1976,10 1576,10 53150,60 7912 266816
 sdc 2072,31 1478,88 53150,60 7424 266816
 sdd 2034,06 1525,10 53150,60 7656 266816
 sde 1988,05 1439,04 53147,41 7224 266800
 sdf 1975,10 1499,60 53147,41 7528 266800
 sdg 1383,07 1485,26 53145,82 7456 266792
 sdh 1562,55 1311,55 53145,82 6584 266792
 sdi 1586,85 1295,62 53145,82 6504 266792
 sdj 0,00 0,00 0,00 0 0
 sdk 0,00 0,00 0,00 0 0
 sdl 0,00 0,00 0,00 0 0
 sdm 0,00 0,00 0,00 0 0
 sdn 0,00 0,00 0,00 0 0
 md0 46515,54 0,00 372124,30 0 1868064
 
 Comments: Large write should not see any read operations. But there are 
 some???

i always saw those small number reads and i feel it is reasonable since
u stripe is 7 * 64KB


 
 
 *** disktest ***
 
 disktest -w -PT -T30 -h1 -K8 -B512k -ID /dev/md0
 
 | 2005/08/25-17:27:04 | STAT | 4072 | v1.1.12 | /dev/md0 | Write 
 throughput: 160152507.7B/s (152.73MB/s), IOPS 305.7/s.
 | 2005/08/25-17:27:05 | STAT | 4072 | v1.1.12 | /dev/md0 | Write 
 throughput: 160694272.0B/s (153.25MB/s), IOPS 306.6/s.
 | 2005/08/25-17:27:06 | STAT | 4072 | v1.1.12 | /dev/md0 | Write 
 throughput: 160339606.6B/s (152.91MB/s), IOPS 305.8/s.

so here 152/7 = 21, large than what u sdc sdd got.


 
 iostat 5 output:
 avg-cpu: %user %nice %sys %iowait %idle
 38,96 0,00 50,25 5,29 5,49
 
 Device: tps Blk_read/s Blk_wrtn/s Blk_read Blk_wrtn
 hda 0,00 0,00 0,00 0 0
 sda 1,20 0,00 11,18 0 56
 sdb 986,43 0,00 39702,99 0 198912
 sdc 922,75 0,00 39728,54 0 199040
 sdd 895,81 0,00 39728,54 0 199040
 sde 880,84 0,00 39728,54 0 199040
 sdf 839,92 0,00 39728,54 0 199040
 sdg 842,91 0,00 39728,54 0 199040
 sdh 1557,49 0,00 79431,54 0 397952
 sdi 2246,71 0,00 104411,98 0 523104
 sdj 0,00 0,00 0,00 0 0
 sdk 0,00 0,00 0,00 0 0
 sdl 0,00 0,00 0,00 0 0
 sdm 0,00 0,00 0,00 0 0
 sdn 0,00 0,00 0,00 0 0
 md0 1550,70 0,00 317574,45 0 1591048
 
 Comments:
 Zero read requests – as it should be. But the write requests are not 
 proportional. sdh and sdi have significantly more requests???

yes, interesting.


 The write requests to the disks of the RAID should be 1/7 higher than to 
 the md device.
 But there are significantly more write operations.
 
 All these operations are to the raw device. Setting up a ext3 fs we get 
 around 127 MB/s with dd.
 
 Any idea?
 
 --Mirko
 

-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: raid1 boot issues

2005-08-25 Thread Lewis Shobbrook
On Thursday 25 August 2005 7:14 pm, you wrote:
 Fdisk it and set partitions to Raid Autodetect (0xfd) possibly?

 Tyler.
Nope already set fd
 Lewis Shobbrook wrote:
 Hi All,
 I have a problem attempting to boot a raid 1 system from lilo.  I have
 succeeded at this many times in the past, but am completely stumped as to
 what the issue is in this instance. I have the boot and root partitions
 seperate on /dev/md0  /dev/md1 respectively, both raid1.
 The mdadm examination  for  the components is clean and the superblocks
  appear as they ought.  I'm using Debian unstable with std. apt
  kernel-image.  The correct modules are in place for the initrd.
 The boot partition fails to mount during  the boot process...
 fsck.ext3: invalid argument while trying to open /dev/md0
 The superblock can not be read or does not describe a correct ext2
 filesystem omitted usual stuff...
 Root password for maintenance or Control-D to continue...
 
 The bizarre thing is that it appears perfectly clean, I've re-zero'd the
 superblocks and completely recreated the device, but the result is always
  the same.
 mdrun loads /dev/md0 in a clean state straight away and it mounts
 cleanly.  /dev/md1 is no problem.
 mdadm -E of the components is clean, as is mdadm -d for the device.
 My fstab  mtab are the same as systems running the same kernel that work
 fine.
 
 I found the system laying around from about 12 months ago which had
  originally been set-up using raidtools.  I upgraded the system using
  dist-upgrade installed mdadm; after zeroing all superblocks for both
  drives and component partitions, I created the devices with the following
  ...
 
 mdadm -C /dev/md0 -l1 -n2 /dev/hda1 missing
 reformatted ext3 and restored the files before adding the missing devices
  and resyncing.
 I'm stumped!
 Anyone got any ideas?
 
 Cheers,
 
 Lewis
 -
 To unsubscribe from this list: send the line unsubscribe linux-raid in
 the body of a message to [EMAIL PROTECTED]
 More majordomo info at  http://vger.kernel.org/majordomo-info.html
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: raid1 boot issues

2005-08-25 Thread Berk Walker

Lewis Shobbrook wrote:


On Thursday 25 August 2005 7:14 pm, you wrote:
 


Fdisk it and set partitions to Raid Autodetect (0xfd) possibly?

Tyler.
   


Nope already set fd
 


Lewis Shobbrook wrote:
   


Hi All,
I have a problem attempting to boot a raid 1 system from lilo.  I have
succeeded at this many times in the past, but am completely stumped as to
what the issue is in this instance. I have the boot and root partitions
seperate on /dev/md0  /dev/md1 respectively, both raid1.
The mdadm examination  for  the components is clean and the superblocks
appear as they ought.  I'm using Debian unstable with std. apt
kernel-image.  The correct modules are in place for the initrd.
The boot partition fails to mount during  the boot process...
fsck.ext3: invalid argument while trying to open /dev/md0
The superblock can not be read or does not describe a correct ext2
filesystem omitted usual stuff...
Root password for maintenance or Control-D to continue...

The bizarre thing is that it appears perfectly clean, I've re-zero'd the
superblocks and completely recreated the device, but the result is always
the same.
mdrun loads /dev/md0 in a clean state straight away and it mounts
cleanly.  /dev/md1 is no problem.
mdadm -E of the components is clean, as is mdadm -d for the device.
My fstab  mtab are the same as systems running the same kernel that work
fine.

I found the system laying around from about 12 months ago which had
originally been set-up using raidtools.  I upgraded the system using
dist-upgrade installed mdadm; after zeroing all superblocks for both
drives and component partitions, I created the devices with the following
...

mdadm -C /dev/md0 -l1 -n2 /dev/hda1 missing
reformatted ext3 and restored the files before adding the missing devices
and resyncing.
I'm stumped!
Anyone got any ideas?

Cheers,

Lewis
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
 


-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

 

Just a silly thought, but does your kernel have raid compiled in?  
'cause if it's a module it will never have a chance to see it.  How far 
in the boot process does your system go?


b-
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: raid1 boot issues

2005-08-25 Thread Lewis Shobbrook
On Thursday 25 August 2005 6:37 pm, you wrote:
Thanks for the reply Rui,

 Hi Lewis,

 Are the boot sectors of your HD's clean? Or have you copied lilo
 onto them?
 If not, are you shure that both you md0 partitions are marked as
 bootable?
 Are the partitions marked 0xFD ?
The bootsectors are clean.  The system boots fine it just won't mount /dev/md0 
correctly during startup.  What surprises me is that /dev/md1 mounts.  Often 
these type of problems are associated with the initrd modules, and result in 
kernel panic earlier in the boot process.  I'd expect that if one raid1 
device is mounted then all required modules are present.  

H
 Rui Santos

 Lewis Shobbrook wrote:
 Hi All,
 I have a problem attempting to boot a raid 1 system from lilo.  I have
 succeeded at this many times in the past, but am completely stumped as to
 what the issue is in this instance. I have the boot and root partitions
 seperate on /dev/md0  /dev/md1 respectively, both raid1.
 The mdadm examination  for  the components is clean and the superblocks
  appear as they ought.  I'm using Debian unstable with std. apt
  kernel-image.  The correct modules are in place for the initrd.
 The boot partition fails to mount during  the boot process...
 fsck.ext3: invalid argument while trying to open /dev/md0
 The superblock can not be read or does not describe a correct ext2
 filesystem omitted usual stuff...
 Root password for maintenance or Control-D to continue...
 
 The bizarre thing is that it appears perfectly clean, I've re-zero'd the
 superblocks and completely recreated the device, but the result is always
  the same.
 mdrun loads /dev/md0 in a clean state straight away and it mounts
 cleanly.  /dev/md1 is no problem.
 mdadm -E of the components is clean, as is mdadm -d for the device.
 My fstab  mtab are the same as systems running the same kernel that work
 fine.
 
 I found the system laying around from about 12 months ago which had
  originally been set-up using raidtools.  I upgraded the system using
  dist-upgrade installed mdadm; after zeroing all superblocks for both
  drives and component partitions, I created the devices with the following
  ...
 
 mdadm -C /dev/md0 -l1 -n2 /dev/hda1 missing
 reformatted ext3 and restored the files before adding the missing devices
  and resyncing.
 I'm stumped!
 Anyone got any ideas?
 
 Cheers,
 
 Lewis
 -
 To unsubscribe from this list: send the line unsubscribe linux-raid in
 the body of a message to [EMAIL PROTECTED]
 More majordomo info at  http://vger.kernel.org/majordomo-info.html
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: raid1 boot issues

2005-08-25 Thread Lewis Shobbrook
On Friday 26 August 2005 11:32 am, you wrote:

Thanks for the help guys,
I've given up for now, cp'd the boot to the md1 device and reflagged the 
partitions as bootable and just not mounting /dev/md0 anymore.  It's most 
probably some strange artefact from the raidtools  mkraid when the devices 
were originally created.  It's only a testing system.  I've not deleted 
the /dev/md0 just toggled the boot flags off for the hd[ac]1 partitions.  All 
the info should int theory still be present if anyone is curious to find out 
more info about this.

 
  Are the boot sectors of your HD's clean? Or have you copied lilo
  onto them?
  If not, are you shure that both you md0 partitions are marked as
  bootable?
  Are the partitions marked 0xFD ?

 The bootsectors are clean.  The system boots fine it just won't mount
 /dev/md0 correctly during startup.  What surprises me is that /dev/md1
 mounts.  Often these type of problems are associated with the initrd
 modules, and result in kernel panic earlier in the boot process.  I'd
 expect that if one raid1 device is mounted then all required modules are
 present.

 H

  Rui Santos
 
  Lewis Shobbrook wrote:
  Hi All,
  I have a problem attempting to boot a raid 1 system from lilo.  I have
  succeeded at this many times in the past, but am completely stumped as
   to what the issue is in this instance. I have the boot and root
   partitions seperate on /dev/md0  /dev/md1 respectively, both raid1.
  The mdadm examination  for  the components is clean and the superblocks
   appear as they ought.  I'm using Debian unstable with std. apt
   kernel-image.  The correct modules are in place for the initrd.
  The boot partition fails to mount during  the boot process...
  fsck.ext3: invalid argument while trying to open /dev/md0
  The superblock can not be read or does not describe a correct ext2
  filesystem omitted usual stuff...
  Root password for maintenance or Control-D to continue...
  
  The bizarre thing is that it appears perfectly clean, I've re-zero'd the
  superblocks and completely recreated the device, but the result is
   always the same.
  mdrun loads /dev/md0 in a clean state straight away and it mounts
  cleanly.  /dev/md1 is no problem.
  mdadm -E of the components is clean, as is mdadm -d for the device.
  My fstab  mtab are the same as systems running the same kernel that
   work fine.
  
  I found the system laying around from about 12 months ago which had
   originally been set-up using raidtools.  I upgraded the system using
   dist-upgrade installed mdadm; after zeroing all superblocks for both
   drives and component partitions, I created the devices with the
   following ...
  
  mdadm -C /dev/md0 -l1 -n2 /dev/hda1 missing
  reformatted ext3 and restored the files before adding the missing
   devices and resyncing.
  I'm stumped!
  Anyone got any ideas?
  
  Cheers,
  
  Lewis
  
Lewis
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html