Re: recover data from linear raid

2006-06-26 Thread Neil Brown
On Monday June 26, [EMAIL PROTECTED] wrote:
> 
>   This is what I get now, after creating with fdisk /dev/hdb1 and 
> /dev/hdc1 as linux raid autodetect partitions

So I'm totally confused now.

You said it was 'linear', but the boot log showed 'raid0'.

The drives didn't have a partition table on them, yet it is clear from
the old boot log that the did.  Are you sure they are the same drives,
'cause it doesn't seem like it.

You could try hunting for ext3 superblocks on the device.  There might
be an easier way but

  od -x /dev/hdb | grep '^.60     ef53 '

should find them.  Once you have this information we might be able to
make something work.  But I feel the chances are dwindling.


NeilBrown
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: recover data from linear raid

2006-06-26 Thread Neil Brown
On Monday June 26, [EMAIL PROTECTED] wrote:
> 
>   This is what I get now, after creating with fdisk /dev/hdb1 and 
> /dev/hdc1 as linux raid autodetect partitions

So I'm totally confused now.

You said it was 'linear', but the boot log showed 'raid0'.

The drives didn't have a partition table on them, yet it is clear from
the old boot log that the did.  Are you sure they are the same drives,
'cause it doesn't seem like it.

You could try hunting for ext3 superblocks on the device.  There might
be an easier way but

  od -x /dev/hdb | grep '^.60     ef53 '

should find them.  Once you have this information we might be able to
make something work.  But I feel the chances are dwindling.


NeilBrown
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: Bug in 2.6.17 / mdadm 2.5.1

2006-06-26 Thread Neil Brown
On Monday June 26, [EMAIL PROTECTED] wrote:
> Neil Brown wrote:
> 
> > Alternately you can apply the following patch to the kernel and
> > version-1 superblocks should work better.
> 
> -stable material?

Maybe.  I'm not sure it exactly qualifies, but I might try sending it
to them and see what they think.

NeilBrown
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: Bug in 2.6.17 / mdadm 2.5.1

2006-06-26 Thread Andre Tomt

Neil Brown wrote:


Alternately you can apply the following patch to the kernel and
version-1 superblocks should work better.


-stable material?
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[-mm patch] make variables static after klibc merge

2006-06-26 Thread Adrian Bunk
We can now make the following variables static:
- drivers/md/md.c: mdp_major
- init/main.c: envp_init[]

Signed-off-by: Adrian Bunk <[EMAIL PROTECTED]>

---

This patch was already sent on:
- 16 May 2006

 drivers/md/md.c |2 +-
 init/main.c |2 +-
 2 files changed, 2 insertions(+), 2 deletions(-)

--- linux-2.6.17-rc4-mm1-full/drivers/md/md.c.old   2006-05-16 
12:49:59.0 +0200
+++ linux-2.6.17-rc4-mm1-full/drivers/md/md.c   2006-05-16 12:50:17.0 
+0200
@@ -2563,7 +2563,7 @@
.default_attrs  = md_default_attrs,
 };
 
-int mdp_major = 0;
+static int mdp_major = 0;
 
 static struct kobject *md_probe(dev_t dev, int *part, void *data)
 {
--- linux-2.6.17-rc4-mm1-full/init/main.c.old   2006-05-16 13:20:22.0 
+0200
+++ linux-2.6.17-rc4-mm1-full/init/main.c   2006-05-16 13:20:43.0 
+0200
@@ -161,7 +161,7 @@
 __setup("maxcpus=", maxcpus);
 
 static char * argv_init[MAX_INIT_ARGS+2] = { "init", NULL, };
-char * envp_init[MAX_INIT_ENVS+2] = { "HOME=/", "TERM=linux", NULL, };
+static char * envp_init[MAX_INIT_ENVS+2] = { "HOME=/", "TERM=linux", NULL, };
 static const char *panic_later, *panic_param;
 
 extern struct obs_kernel_param __setup_start[], __setup_end[];

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[-mm patch] drivers/md/raid5.c: remove an unused variable

2006-06-26 Thread Adrian Bunk
This patch removes an unused variable.

Signed-off-by: Adrian Bunk <[EMAIL PROTECTED]>

--- linux-2.6.17-mm2-full/drivers/md/raid5.c.old2006-06-26 
21:17:13.0 +0200
+++ linux-2.6.17-mm2-full/drivers/md/raid5.c2006-06-26 21:17:20.0 
+0200
@@ -2827,7 +2827,6 @@
struct stripe_head *sh;
int pd_idx;
int raid_disks = conf->raid_disks;
-   int data_disks = raid_disks - conf->max_degraded;
sector_t max_sector = mddev->size << 1;
int sync_blocks;
int still_degraded = 0;

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: recover data from linear raid

2006-06-26 Thread Dimitris Zilaskos


	This is what I get now, after creating with fdisk /dev/hdb1 and 
/dev/hdc1 as linux raid autodetect partitions


mdadm -E /dev/hdb1
/dev/hdb1:
  Magic : a92b4efc
Version : 00.90.00
   UUID : a7e90d4b:f347bd0e:07ebf941:e718f695
  Creation Time : Wed Mar 16 18:14:25 2005
 Raid Level : raid0
Device Size : 244195904 (232.88 GiB 250.06 GB)
   Raid Devices : 2
  Total Devices : 2
Preferred Minor : 0

Update Time : Wed Jun 21 21:57:37 2006
  State : clean, no-errors
 Active Devices : 2
Working Devices : 2
 Failed Devices : 0
  Spare Devices : 0
   Checksum : c8c21ed8 - correct
 Events : 0.82

 Chunk Size : 64K

  Number   Major   Minor   RaidDevice State
this 0   3   650  active sync   /dev/hdb1
   0 0   3   650  active sync   /dev/hdb1
   1 1  2211  active sync   /dev/hdc1
[EMAIL PROTECTED] root]# mdadm -E /dev/hdc1
/dev/hdc1:
  Magic : a92b4efc
Version : 00.90.00
   UUID : a7e90d4b:f347bd0e:07ebf941:e718f695
  Creation Time : Wed Mar 16 18:14:25 2005
 Raid Level : raid0
Device Size : 244195904 (232.88 GiB 250.06 GB)
   Raid Devices : 2
  Total Devices : 2
Preferred Minor : 0

Update Time : Wed Jun 21 21:57:37 2006
  State : clean, no-errors
 Active Devices : 2
Working Devices : 2
 Failed Devices : 0
  Spare Devices : 0
   Checksum : c8c21ead - correct
 Events : 0.82

 Chunk Size : 64K

  Number   Major   Minor   RaidDevice State
this 1  2211  active sync   /dev/hdc1
   0 0   3   650  active sync   /dev/hdb1
   1 1  2211  active sync   /dev/hdc1


--


Dimitris Zilaskos

Department of Physics @ Aristotle University of Thessaloniki , Greece
PGP key : http://tassadar.physics.auth.gr/~dzila/pgp_public_key.asc
  http://egnatia.ee.auth.gr/~dzila/pgp_public_key.asc
MD5sum  : de2bd8f73d545f0e4caf3096894ad83f  pgp_public_key.asc


-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: recover data from linear raid

2006-06-26 Thread Dimitris Zilaskos


	I managed to get the hard disk of the retired system and this is 
its raid-related boot log:


md: Autodetecting RAID arrays.
 [events: 004d]
 [events: 004d]
md: autorun ...
md: considering hdb1 ...
md:  adding hdb1 ...
md:  adding hdc1 ...
md: created md0
md: bind
md: bind
md: running: 
md: hdb1's event counter: 004d
md: hdc1's event counter: 004d
md0: max total readahead window set to 512k
md0: 2 data-disks, max readahead per data-disk: 256k
raid0: looking at hdb1
raid0:   comparing hdb1(244195904) with hdb1(244195904)
raid0:   END
raid0:   ==> UNIQUE
raid0: 1 zones
raid0: looking at hdc1
raid0:   comparing hdc1(293049600) with hdb1(244195904)
raid0:   NOT EQUAL
raid0:   comparing hdc1(293049600) with hdc1(293049600)
raid0:   END
raid0:   ==> UNIQUE
raid0: 2 zones
raid0: FINAL 2 zones
raid0: zone 0
raid0: checking hdb1 ... contained as device 0
  (244195904) is smallest!.
raid0: checking hdc1 ... contained as device 1
raid0: zone->nb_dev: 2, size: 488391808
raid0: current zone offset: 244195904
raid0: zone 1
raid0: checking hdb1 ... nope.
raid0: checking hdc1 ... contained as device 0
  (293049600) is smallest!.
raid0: zone->nb_dev: 1, size: 48853696
raid0: current zone offset: 293049600
raid0: done.
raid0 : md_size is 537245504 blocks.
raid0 : conf->smallest->size is 48853696 blocks.
raid0 : nb_zone is 11.
raid0 : Allocating 88 bytes for hash.
md: ... autorun DONE.




As Christian said, specific error message help a lot.
Assume the two devices are hdc and hde,

 fdisk -l /dev/hdc
 fdisk -l /dev/hde
 mdadm -E /dev/hdc
 mdadm -E /dev/hde

and my best guess

  mdadm --build /dev/md0 --level linear --raid-disks 2 /dev/hdc /dev/hde
  fsck -n /dev/md0

(and linux-raid@vger.kernel.org might be a better mailing list for
this particular sort of problem).


Disk /dev/hdb: 250.0 GB, 250059350016 bytes
255 heads, 63 sectors/track, 30401 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

  Device BootStart   EndBlocks   Id  System
[EMAIL PROTECTED] root]# fdisk -l /dev/hdc

Disk /dev/hdc: 300.0 GB, 300090728448 bytes
255 heads, 63 sectors/track, 36483 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

  Device BootStart   EndBlocks   Id  System
[EMAIL PROTECTED] root]# mdadm -E /dev/hdb
/dev/hdb:
 Magic : a92b4efc
   Version : 00.90.00
  UUID : 293be3e8:5a7ac6e7:adefc469:84f8aefb
 Creation Time : Fri Jun 23 15:47:10 2006
Raid Level : linear
   Device Size : 244198464 (232.89 GiB 250.06 GB)
  Raid Devices : 2
 Total Devices : 2
Preferred Minor : 0

   Update Time : Fri Jun 23 15:48:43 2006
 State : clean, no-errors
Active Devices : 2
Working Devices : 2
Failed Devices : 0
 Spare Devices : 0
  Checksum : f790e07f - correct
Events : 0.2

  Rounding : 32K

 Number   Major   Minor   RaidDevice State
this 0   3   640  active sync   /dev/hdb
  0 0   3   640  active sync   /dev/hdb
  1 1  2201  active sync   /dev/hdc
[EMAIL PROTECTED] root]# mdadm -E /dev/hdc
/dev/hdc:
 Magic : a92b4efc
   Version : 00.90.00
  UUID : 293be3e8:5a7ac6e7:adefc469:84f8aefb
 Creation Time : Fri Jun 23 15:47:10 2006
Raid Level : linear
   Device Size : 244198464 (232.89 GiB 250.06 GB)
  Raid Devices : 2
 Total Devices : 2
Preferred Minor : 0

   Update Time : Fri Jun 23 15:48:43 2006
 State : clean, no-errors
Active Devices : 2
Working Devices : 2
Failed Devices : 0
 Spare Devices : 0
  Checksum : f790e054 - correct
Events : 0.2

  Rounding : 32K

 Number   Major   Minor   RaidDevice State
this 1  2201  active sync   /dev/hdc
  0 0   3   640  active sync   /dev/hdb
  1 1  2201  active sync   /dev/hdc
[EMAIL PROTECTED] root]# mdadm --build /dev/md0 --level linear --raid-disks 2 
/dev/hdb /dev/hdc

mdadm: array /dev/md0 built and started.

fsck -n /dev/md0
fsck 1.32 (09-Nov-2002)
e2fsck 1.32 (09-Nov-2002)
Couldn't find ext2 superblock, trying backup blocks...
fsck.ext2: Bad magic number in super-block while trying to open /dev/md0

The superblock could not be read or does not describe a correct ext2
filesystem.  If the device is valid and it really contains an ext2
filesystem (and not swap or ufs or something else), then the superblock
is corrupt, and you might try running e2fsck with an alternate superblock:
   e2fsck -b 8193 

fsck -b 8193 /dev/md0
fsck 1.32 (09-Nov-2002)
e2fsck 1.32 (09-Nov-2002)
fsck.ext2: Bad magic number in super-block while trying to open /dev/md0

The superblock could not be read or does not describe a correct ext2
filesystem.  If the device is valid and it really contains an ext2
filesystem (and not swap or ufs or something else), then the superblock
is corrupt, and you might try running e2fsck with an alternate superblock:
   e2fsck -b 8193 


	During a recovery attemp today by mistake I created a mirror a

Re: IBM xSeries stop responding during RAID1 reconstruction

2006-06-26 Thread Bill Davidsen

Mr. James W. Laferriere wrote:


Hello Gabor ,

On Tue, 20 Jun 2006, Gabor Gombas wrote:


On Tue, Jun 20, 2006 at 03:08:59PM +0200, Niccolo Rigacci wrote:


Do you know if it is possible to switch the scheduler at runtime?


echo cfq > /sys/block//queue/scheduler



At least one can do a ls of the /sys/block area & then do an 
automated
echo cfq down the tree .  Does anyone know of a method to set a 
default
scheduler ?  Scanning down a list or manually maintaining a list 
seems

to be a bug in the waiting .  Tia ,  JimL


Thought I posted this... it can be set in kernel build or on the bloot 
parameters from grub/lilo.


2nd thought: set it to cfq by default, then at the END of rc.local, if 
there are no arrays rebuilding, change to something else if you like.


--
bill davidsen <[EMAIL PROTECTED]>
 CTO TMR Associates, Inc
 Doing interesting things with small computers since 1979

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: RAID5 degraded after mdadm -S, mdadm --assemble (everytime)

2006-06-26 Thread Bill Davidsen

Ronald Lembcke wrote:


Hi!

I set up a RAID5 array of 4 disks. I initially created a degraded array
and added the fourth disk (sda1) later.

The array is "clean", but when I do  
 mdadm -S /dev/md0 
 mdadm --assemble /dev/md0 /dev/sd[abcd]1

it won't start. It always says sda1 is "failed".

When I remove sda1 and add it again everything seems to be fine until I
stop the array. 


Below is the output of /proc/mdstat, mdadm -D -Q, mdadm -E and a piece of the
kernel log.
The output of mdadm -E looks strange for /dev/sd[bcd]1, saying "1 failed".

What can I do about this?
How could this happen? I mixed up the syntax when adding the fourth disk and
tried these two commands (at least one didn't yield an error message):
mdadm --manage -a /dev/md0 /dev/sda1
mdadm --manage -a /dev/sda1 /dev/md0


Thanks in advance ...
 Roni



ganges:~# cat /proc/mdstat 
Personalities : [raid5] [raid4] 
md0 : active raid5 sda1[4] sdc1[0] sdb1[2] sdd1[1]

 691404864 blocks super 1.0 level 5, 64k chunk, algorithm 2 [4/4] []
 
unused devices: 


I will just comment that the 0 1 2   4 numbering on the devices is 
unusual. When you created this did you do something which made md think 
there was another device, failed or missing, which was device[3]? I just 
looked at a bunch of my arrays and found no similar examples.


--
bill davidsen <[EMAIL PROTECTED]>
 CTO TMR Associates, Inc
 Doing interesting things with small computers since 1979

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: Large single raid and XFS or two small ones and EXT3?

2006-06-26 Thread Bill Davidsen

Adam Talbot wrote:

Not exactly sure how to tune for stripe size. 
What would you advise?

-Adam
 



See the -R option of mke2fs. I don't have a number for the performance 
impact of this, but I bet someone else on the list will. Depending on 
what posts you read, reports range from "measurable" to "significant," 
without quantifying.


Note, next month I will set up either a 2x750 RAID-1 or 4x250 RAID-5 
array, and if I got RAID-5 I will have the chance to run some metrics 
before putting the hardware into production service. I'll report on the 
-R option if I have any data.




Bill Davidsen wrote:
 


winspeareAdam Talbot wrote:

   


OK, this topic I relay need to get in on.
I have spent the last few week bench marking my new 1.2TB, 6 disk, RAID6
array. I wanted real numbers, not "This FS is faster because..." I have
moved over 100TB of data on my new array running the bench mark
testing.  I have yet to have any major problems with ReiserFS, EXT2/3,
JFS, or XFS.  I have done extensive testing on all, including just
trying to break the file system with billions of 1k files, or a 1TB
file. Was able to cause some problems with EXT3 and RiserFS with the 1KB
and 1TB tests, respectively. but both were fixed with a fsck. My basic
test is to move all data from my old server to my new server
(whitequeen2) and clock the transfer time.  Whitequeen2 has very little
storage.  The NAS's 1.2TB of storage is attached via iSCSI and a cross
over cable to the back of whitequeen2.  The data is 100GB of user's
files(1KB~2MB), 50GB of MP3's (1MB~5MB) and the rest is movies and
system backups 600MB~2GB.  Here is a copy of my current data sheet,
including specs on the servers and copy times, my numbers are not
perfect, but they should give you a clue about speeds...  XFS wins.


 


In many (most?) cases I'm a lot more concerned about filesystem
stability than performance. That is, I want the fastest 
filesystem. With ext2 and ext3 I've run multiple multi-TB machines
spread over four time zones, and not had a f/s problem updating ~1TB/day.

   


Did you tune the extN filesystems to the stripe size of the raid?

   




--
bill davidsen <[EMAIL PROTECTED]>
 CTO TMR Associates, Inc
 Doing interesting things with small computers since 1979

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: recover data from linear raid

2006-06-26 Thread Dimitris Zilaskos


As Christian said, specific error message help a lot.
Assume the two devices are hdc and hde,

 fdisk -l /dev/hdc
 fdisk -l /dev/hde
 mdadm -E /dev/hdc
 mdadm -E /dev/hde

and my best guess

  mdadm --build /dev/md0 --level linear --raid-disks 2 /dev/hdc /dev/hde
  fsck -n /dev/md0

(and linux-raid@vger.kernel.org might be a better mailing list for
this particular sort of problem).


Disk /dev/hdb: 250.0 GB, 250059350016 bytes
255 heads, 63 sectors/track, 30401 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device BootStart   EndBlocks   Id  System
[EMAIL PROTECTED] root]# fdisk -l /dev/hdc

Disk /dev/hdc: 300.0 GB, 300090728448 bytes
255 heads, 63 sectors/track, 36483 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device BootStart   EndBlocks   Id  System
[EMAIL PROTECTED] root]# mdadm -E /dev/hdb
/dev/hdb:
  Magic : a92b4efc
Version : 00.90.00
   UUID : 293be3e8:5a7ac6e7:adefc469:84f8aefb
  Creation Time : Fri Jun 23 15:47:10 2006
 Raid Level : linear
Device Size : 244198464 (232.89 GiB 250.06 GB)
   Raid Devices : 2
  Total Devices : 2
Preferred Minor : 0

Update Time : Fri Jun 23 15:48:43 2006
  State : clean, no-errors
 Active Devices : 2
Working Devices : 2
 Failed Devices : 0
  Spare Devices : 0
   Checksum : f790e07f - correct
 Events : 0.2

   Rounding : 32K

  Number   Major   Minor   RaidDevice State
this 0   3   640  active sync   /dev/hdb
   0 0   3   640  active sync   /dev/hdb
   1 1  2201  active sync   /dev/hdc
[EMAIL PROTECTED] root]# mdadm -E /dev/hdc
/dev/hdc:
  Magic : a92b4efc
Version : 00.90.00
   UUID : 293be3e8:5a7ac6e7:adefc469:84f8aefb
  Creation Time : Fri Jun 23 15:47:10 2006
 Raid Level : linear
Device Size : 244198464 (232.89 GiB 250.06 GB)
   Raid Devices : 2
  Total Devices : 2
Preferred Minor : 0

Update Time : Fri Jun 23 15:48:43 2006
  State : clean, no-errors
 Active Devices : 2
Working Devices : 2
 Failed Devices : 0
  Spare Devices : 0
   Checksum : f790e054 - correct
 Events : 0.2

   Rounding : 32K

  Number   Major   Minor   RaidDevice State
this 1  2201  active sync   /dev/hdc
   0 0   3   640  active sync   /dev/hdb
   1 1  2201  active sync   /dev/hdc
[EMAIL PROTECTED] root]# mdadm --build /dev/md0 --level linear --raid-disks 2 
/dev/hdb /dev/hdc

mdadm: array /dev/md0 built and started.

fsck -n /dev/md0
fsck 1.32 (09-Nov-2002)
e2fsck 1.32 (09-Nov-2002)
Couldn't find ext2 superblock, trying backup blocks...
fsck.ext2: Bad magic number in super-block while trying to open /dev/md0

The superblock could not be read or does not describe a correct ext2
filesystem.  If the device is valid and it really contains an ext2
filesystem (and not swap or ufs or something else), then the superblock
is corrupt, and you might try running e2fsck with an alternate superblock:
e2fsck -b 8193 

fsck -b 8193 /dev/md0
fsck 1.32 (09-Nov-2002)
e2fsck 1.32 (09-Nov-2002)
fsck.ext2: Bad magic number in super-block while trying to open /dev/md0

The superblock could not be read or does not describe a correct ext2
filesystem.  If the device is valid and it really contains an ext2
filesystem (and not swap or ufs or something else), then the superblock
is corrupt, and you might try running e2fsck with an alternate superblock:
e2fsck -b 8193 


	During a recovery attemp today by mistake I created a mirror array 
with hdb as the primary and hdc as the secondary. I interrupted the array 
creation almost immediately, but part of the hdc was overwritten. However 
the array never held more than 70 gbytes of data, so I hope everything is 
intact on hdb :/



Thank you all for your kind help:)
--


Dimitris Zilaskos

Department of Physics @ Aristotle University of Thessaloniki , Greece
PGP key : http://tassadar.physics.auth.gr/~dzila/pgp_public_key.asc
  http://egnatia.ee.auth.gr/~dzila/pgp_public_key.asc
MD5sum  : de2bd8f73d545f0e4caf3096894ad83f  pgp_public_key.asc

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: Is shrinking raid5 possible?

2006-06-26 Thread Christian Pernegger

This is shrinking an array by removing drives.  We were talking about
shrinking an array by reducing the size of drives - a very different
think.


Yes I know - I just wanted to get this in as an alternative shrinking semantic.

As for reducing the RAID (partition) size on the individual drives I
can only see two reasons:

1) One wants to replace / add a disk but the new disk is slightly
smaller than the existing ones. Actual capacity varies a lot for the
same nominal capacity, especially across brands.

2) One wants to do something else with some of the space, although for
a RAID5 I don't quite see the point - you'd end up with n small
partitions. Shrinking a 2-way mirror or stripe this way sounds far
more useful.

Having RAID tightly integrated with volume management and partitions
would be nice, but that's a pipe dream. (EVMS just integrates the UI
somewhat, you still have to mess about with layers.)


I'm not sure it is really worth the effort I'm afraid.


Neither am I ... it just would have come in handy a few times, that's all.

Thanks,

C.
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: Multiple raids on one machine?

2006-06-26 Thread Chris Allen



Gordon Henderson wrote:

I use option 2 (above) all the time, and I've never noticed any
performance issues. (not issues with recovery after a power failure) I'd
like to think that on a modern processor the CPU can handle the parity,
etc. calculations several orders of magnitude faster than the hardware can
chug data to & from the drives, so all it's really adding is a tiny bit of
latency...

  


Thanks for this, and for the very informed responses from other 
subscribers. We have gone from
100GB of storage in May 2000 (which lasted us over a year) to 50TB today 
(which will last us
three months), and a forecast of 200TB by this time next year! I'm 
learning fast :-)



-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: Large single raid and XFS or two small ones and EXT3?

2006-06-26 Thread Justin Piszcz



On Sun, 25 Jun 2006, Bill Davidsen wrote:


Justin Piszcz wrote:



On Sat, 24 Jun 2006, Neil Brown wrote:


On Friday June 23, [EMAIL PROTECTED] wrote:


The problem is that there is no cost effective backup available.



One-liner questions :
- How does Google make backups ?



No, Google ARE the backups :-)


- Aren't tapes dead yet ?



LTO-3 does 300Gig, and LTO-4 is planned.
They may not cope with tera-byte arrays in one hit, but they still
have real value.


- What about a NUMA principle applied to storage ?



You mean an Hierarchical Storage Manager?  Yep, they exist.  I'm sure
SGI, EMC and assorted other TLAs could sell you one.



LTO3 is 400GB native and we've seen very good compression, so 800GB-1TB per 
tape. 


The problem is in small business use, LTO3 is costly in the 1-10TB range, and 
takes a lot of media changes as well. A TB of RAID-5 is ~$500, and at that 
small size the cost of drives and media is disproportionally high. Using more 
drives is cost effective, but they are not good for long term off site 
storage, because they're large and fragile.


No obvious solutions in that price and application range that I see.

--
bill davidsen <[EMAIL PROTECTED]>
CTO TMR Associates, Inc
Doing interesting things with small computers since 1979



In the 1-10TB range you are probably correct, as the numbers increase 
however, many LTO2/LTO3 drives + robotics become justifiable.

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: Multiple raids on one machine?

2006-06-26 Thread Gordon Henderson
On Sun, 25 Jun 2006, Chris Allen wrote:

> Back to my 12 terabyte fileserver, I have decided to split the storage
> into four partitions
> each of 3TB. This way I can choose between XFS and EXT3 later on.
>
> So now, my options are between the following:
>
> 1. Single 12TB /dev/md0, partitioned into four 3TB partitions. But how do
> I do this? fdisk won't handle it. Can GNU Parted handle partitions this big?
>
> 2. Partition the raw disks into four partitions and make
> /dev/md0,md1,md2,md3.
> But am I heading for problems here? Is there going to be a big
> performance hit
> with four raid5 arrays on the same machine? Am I likely to have dataloss
> problems
> if my machine crashes?

I use option 2 (above) all the time, and I've never noticed any
performance issues. (not issues with recovery after a power failure) I'd
like to think that on a modern processor the CPU can handle the parity,
etc. calculations several orders of magnitude faster than the hardware can
chug data to & from the drives, so all it's really adding is a tiny bit of
latency...

Someone some time back on this list posted a hint that I've been using
though - you might find it handy - name the md? devices after the
partition number, if possible. So md1 would be made up from /dev/sda1,
/dev/sdb1, /dev/sdc1, etc. md2 made up from /dev/sda2, /dev/sdb2, etc. It
might just save any confusion when hot adding/removing drives, etc.

The down-side is that if you do have to remove a drive, you have to
manually 'fail' each other md device+partition for that drive, then
manually remove them before you can hot-remove the physical drive. (or
cold remove it/whatever)

So if you have /dev/md{1,2.3,4} and /dev/md3 (etc) is made from /dev/sda3,
/dev/sdb3, /dev/sdc3, and md3 (eg. /dev/sdc3) has a failure, then you need
to:

  mdadm --fail /dev/md1 /dev/sdc1
  mdadm --fail /dev/md2 /dev/sdc2
# mdadm --fail /dev/md3 /dev/sdc3   # already failed
  mdadm --fail /dev/md4 /dev/sdc4

Then repeat, s/fail/remove/ then you can echo the right runes to
/proc/scsi/scsi and hot-remove /dev/sdc and plug a new one in.

At least, thats what I do when I've done it 'hot'. Doing it cold doesn't
really matter as the server will boot with a blank partition table in the
replaced disk and just kick it out of the array - you can then
re-partition, and mdadm --add ... the partition back into each array.

I like to keep each partition identical, if I can - heres an example:

Personalities : [raid1] [raid6]
md1 : active raid1 sdf1[5] sde1[4] sdd1[3] sdc1[2] sdb1[1] sda1[0]
  248896 blocks [6/6] [UU]

md2 : active raid6 sdf2[5] sde2[4] sdd2[3] sdc2[2] sdb2[1] sda2[0]
  1991680 blocks level 6, 64k chunk, algorithm 2 [6/6] [UU]

md3 : active raid6 sdf3[5] sde3[4] sdd3[3] sdc3[2] sdb3[1] sda3[0]
  1991680 blocks level 6, 64k chunk, algorithm 2 [6/6] [UU]

md5 : active raid6 sdf5[5] sde5[4] sdd5[3] sdc5[2] sdb5[1] sda5[0]
  3983616 blocks level 6, 64k chunk, algorithm 2 [6/6] [UU]

md6 : active raid6 sdf6[5] sde6[4] sdd6[3] sdc6[2] sdb6[1] sda6[0]
  277345536 blocks level 6, 64k chunk, algorithm 2 [6/6] [UU]

md7 : active raid6 sdf7[5] sde7[4] sdd7[3] sdc7[2] sdb7[1] sda7[0]
  287177472 blocks level 6, 64k chunk, algorithm 2 [6/6] [UU]

Each drive is partitioned like:

Disk /dev/sda: 255 heads, 63 sectors, 17849 cylinders
Units = cylinders of 16065 * 512 bytes

   Device BootStart   EndBlocks   Id  System
/dev/sda1   * 131248976   fd  Linux raid autodetect
/dev/sda23293498015   fd  Linux raid autodetect
/dev/sda394   155498015   fd  Linux raid autodetect
/dev/sda4   156 17849 1421270555  Extended
/dev/sda5   156   279995998+  fd  Linux raid autodetect
/dev/sda6   280  8911  69336508+  fd  Linux raid autodetect
/dev/sda7  8912 17849  71794453+  fd  Linux raid autodetect


(Yes, md1 is a RAID-1 striped over 6 drives! write performance might be
"sub optimal", but it's only the root partition and hardly ever written
to, and yes, swap /dev/md2 on this box is under R-6)

Gordon
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: Is shrinking raid5 possible?

2006-06-26 Thread Neil Brown
On Friday June 23, [EMAIL PROTECTED] wrote:
> > Why would you ever want to reduce the size of a raid5 in this way?
> 
> A feature that would have been useful to me a few times is the ability
> to shrink an array by whole disks.
> 
> Example:
> 
> 8x 300 GB disks -> 2100 GB raw capacity
> 
> shrink file system, remove 2 disks =
> 
> 6x 300 GB disks --> 1500 GB raw capacity
> 

This is shrinking an array by removing drives.  We were talking about
shrinking an array by reducing the size of drives - a very different
think.

Yes, it might be sometimes useful to reduce the number of drives in a
raid5.  This would be similar to adding a drive to a raid5 (now
possible), but the data copy would have to go in a different
direction, so there would need to be substantial changes to the code.
I'm not sure it is really worth the effort I'm afraid, but it might
get done, one day, especially if someone volunteers some code ... ;-)

NeilBrown


> Why?
> 
> If you're not backed up by a company budget, moving data to an new
> array (extra / larger disks) is extremely difficult. A lot of cases
> will hold 8 disks but not 16, never mind the extra RAID controller.
> Building another temporary server and moving the data via Gigabit is
> slow and expensive as well.
> 
> Shrinking the array step-by-step and unloading data onto a regular
> filesystem on the freed disks would be a cheap (if time consuming) way
> to migrate, because the data could be copied back to the new array a
> disk at a time.
> 
> Thanks,
> 
> C.
> -
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to [EMAIL PROTECTED]
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html