Re: Problem with 5disk RAID5 array - two drives lost

2006-04-23 Thread Arthur Britto
On Sun, 2006-04-23 at 17:17 -0700, Tim Bostrom wrote:
> I bought two extra 250GB drives - I'll try using dd_rescue as  
> recommended and see if I can get a "good" copy of hdf online.

You might want to use dd_rhelp:
http://www.kalysto.org/utilities/dd_rhelp/index.en.html

-Arthur


-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


proactive raid5 disk replacement success (using bitmap + raid1)

2006-04-23 Thread dean gaudet
i had a disk in a raid5 which i wanted to clone onto the hot spare... 
without going offline and without long periods without redundancy.  a few 
folks have discussed using bitmaps and temporary (superblockless) raid1 
mappings to do this... i'm not sure anyone has tried / reported success 
though.  this is my success report.

setup info:

- kernel version 2.6.16.9 (as packaged by debian)
- mdadm version 2.4.1
- /dev/md4 is the raid5
- /dev/sde1 is the disk in md4 i want to clone from
- /dev/sdh1 is the hot spare from md4, and is the clone target
- /dev/md5 is an unused md device name

here are the exact commands i issued:

mdadm -Gb internal --bitmap-chunk=1024 /dev/md4
mdadm /dev/md4 -r /dev/sdh1
mdadm /dev/md4 -f /dev/sde1 -r /dev/sde1
mdadm --build /dev/md5 -ayes --level=1 --raid-devices=2 /dev/sde1 missing
mdadm /dev/md4 --re-add /dev/md5
mdadm /dev/md5 -a /dev/sdh1

... wait a few hours for md5 resync...

mdadm /dev/md4 -f /dev/md5 -r /dev/md5
mdadm --stop /dev/md5
mdadm /dev/md4 --re-add /dev/sdh1
mdadm --zero-superblock /dev/sde1
mdadm /dev/md4 -a /dev/sde1

this sort of thing shouldn't be hard to script :)

the only times i was without full redundancy was briefly between the "-r" 
and "--re-add" commands... and with bitmap support the raid5 resync for 
each of those --re-adds was essentially zero.

thanks Neil (and others)!

-dean

p.s. it's absolutely necessary to use "--build" for the temporary raid1 
... if you use --create mdadm will rightfully tell you it's already a raid 
component and if you --force it then you'll trash the raid5 superblock and 
it won't fit into the raid5 any more...
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: Problem with 5disk RAID5 array - two drives lost

2006-04-23 Thread Tim Bostrom
First let me say - thank you for responding.  I'm still trying to  
figure out this problem.



On Apr 21, 2006, at 8:54 PM  Apr 21, 2006, Molle Bestefich wrote:


Tim Bostrom wrote:
It appears that /dev/hdf1 failed this past week and /dev/hdh1  
failed  back in February.


An obvious question would be, how much have you been altering the
contents of the array since February?



This is a video backup drive for my MythTV system.  I just backup old  
TV shows and movies from my system.  There's maybe 3-4 GB of data  
that's been stored there since February and no other data's been  
moved or deleted.  Pretty much when something is backed up here, it  
stays.  I'm willing to lose the February - April data.





hdf: dma_intr: error=0x40 { UncorrectableError }, LBAsect=6720




Actually, there's a lot of sequential sector numbers in the output  
you posted.
I think it's unusual for a drive to develop that many bad blocks in  
a row.

I could be wrong, and it could be a head crash or something (have you
been moving the system around much?).

But if I had to guess, I'd say that there's a real likelihood that
it's a loose cable or a controller problem or a driver issue.

Could you try and run:
# dd if=/dev/hdf of=/dev/null bs=1M count=100 skip=1234567
You can play around with different random numbers instead of 1234567.
If it craps out *immediately*, then I'd say it's a cable problem or
so, and not a problem with what's on the platters.




Tried running this.  It doesn't crap out immediately.  Goes along,  
but I see a bunch of the {Uncorrectable Error} {Drive Ready Seek  
Complete} errors in dmesg that I posted before.  I see these messages  
in dmesg when I run the above command.


I bought two extra 250GB drives - I'll try using dd_rescue as  
recommended and see if I can get a "good" copy of hdf online.






No, get the array running first, then fix the filesystem.

You can initiate array checks and repairs like this:
# cd /sys/block/md0/md/
# echo check > sync_action
or
# echo repair > sync_action



-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: Recovery speed at 1MB/s/device, unable to change

2006-04-23 Thread Anssi Hannula
(resend, prev post missed the list)

Neil Brown wrote:
> On Monday April 24, [EMAIL PROTECTED] wrote:
> 
>># mdadm --grow /dev/md_d1 --raid-devices=3 --backup-file backupfile
>>mdadm: Need to backup 128K of critical section..
>>mdadm: /dev/md_d1: Cannot get array details from sysfs
>>
>>Strace shows that it's trying to access
>>"/sys/block/md_d4/md/component_size".
>>
>>Why is this?
> 
> 
> Because I didn't test my code properly :-(
> 
> Following patch should fix it.
> 

It appears you missed another occurrence (patch attached).

However, it got stuck after
mdadm: Need to backup 128K of critical section..

I did ctrl-c, and according to /proc/mdstat it was successfully
reshaped, nor can I restart the reshape.

But "umount /mnt/test" (/mnt/test is where /dev/md_d1p1 is mounted)
blocks, and this time it becomes unkillable.

/proc/mdstat reports:
md_d1 : active raid5 loop1[1] loop0[0]
  4992 blocks level 5, 64k chunk, algorithm 2 [3/2] [UU_]


-- 
Anssi Hannula


--- sysfs.c.old	2006-04-24 01:24:09.0 +0300
+++ sysfs.c	2006-04-24 01:33:48.0 +0300
@@ -69,7 +69,7 @@
 			sprintf(sra->name, "md%d", minor(stb.st_rdev));
 		else
 			sprintf(sra->name, "md_d%d",
-minor(stb.st_rdev)/16);
+minor(stb.st_rdev)/64);
 	} else {
 		if (devnum >= 0)
 			sprintf(sra->name, "md%d", devnum);



Re: Recovery speed at 1MB/s/device, unable to change

2006-04-23 Thread Anssi Hannula
(sorry, prev post missed the list)

Neil Brown wrote:
> On Monday April 24, [EMAIL PROTECTED] wrote:
> 
>># mdadm --grow /dev/md_d1 --raid-devices=3 --backup-file backupfile
>>mdadm: Need to backup 128K of critical section..
>>mdadm: /dev/md_d1: Cannot get array details from sysfs
>>
>>Strace shows that it's trying to access
>>"/sys/block/md_d4/md/component_size".
>>
>>Why is this?
> 
> 
> Because I didn't test my code properly :-(
> 
> Following patch should fix it.

Okay, will test.

> Re: the recovery speed problem:  I would try dding from one drive to
> the other and see what sort of speed you get then:
>   dd if=/dev/sda of=/dev/sdb bs=1024k count=1000
> or something like that.


# export LANGUAGE=C; export LANG=C; export LC_ALL=C; date; dd
if=/dev/sda skip=2 of=/dev/sdb bs=1024k count=1000; date
Mon Apr 24 01:14:48 EEST 2006
1000+0 records in
1000+0 records out
Mon Apr 24 01:15:42 EEST 2006

So the speed is about 20MB/s.

-- 
Anssi Hannula


-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: EVMS causing problems with mdadm?

2006-04-23 Thread Luca Berra

On Mon, Apr 24, 2006 at 07:48:00AM +1000, Neil Brown wrote:

On Sunday April 23, [EMAIL PROTECTED] wrote:

Did my latest updates for my Kubuntu (Ubuntu KDE variant) this
morning, and noticed that EVMS has now "taken control" of my RAID
array. Didn't think much about it until I tried to make a RAID-1 array
with two disks I've just added to the system. Trying to do a create
verbose tells me that device /dev/md1 (or 2 or 3 - I tried a couple
just to see) doesn't exist. And in fact, there are no block devices
listed beyond md0.


Sounds like udev is in use rather than a static /dev.

Add '--auto=md' to the mdadm command line, and it will create the
devices for you.  Or --auto=part if you want partitioned arrays.  See
man page for more details.

I suspect this might need to be come the default in another year or so


i was tkinking about stat()ing "/dev/.udev" and automatically enabling
--auto if found

WDYT?

L.

--
Luca Berra -- [EMAIL PROTECTED]
   Communication Media & Services S.r.l.
/"\
\ / ASCII RIBBON CAMPAIGN
 XAGAINST HTML MAIL
/ \
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: Recovery speed at 1MB/s/device, unable to change

2006-04-23 Thread Neil Brown
On Monday April 24, [EMAIL PROTECTED] wrote:
> # mdadm --grow /dev/md_d1 --raid-devices=3 --backup-file backupfile
> mdadm: Need to backup 128K of critical section..
> mdadm: /dev/md_d1: Cannot get array details from sysfs
> 
> Strace shows that it's trying to access
> "/sys/block/md_d4/md/component_size".
> 
> Why is this?

Because I didn't test my code properly :-(

Following patch should fix it.

Re: the recovery speed problem:  I would try dding from one drive to
the other and see what sort of speed you get then:
  dd if=/dev/sda of=/dev/sdb bs=1024k count=1000
or something like that.

Thanks,
NeilBrown

Signed-off-by: Neil Brown <[EMAIL PROTECTED]>

### Diffstat output
 ./sysfs.c |2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff ./sysfs.c~current~ ./sysfs.c
--- ./sysfs.c~current~  2006-03-14 15:33:12.0 +1100
+++ ./sysfs.c   2006-04-24 07:56:56.0 +1000
@@ -206,7 +206,7 @@ unsigned long long get_component_size(in
minor(stb.st_rdev));
else
sprintf(fname, "/sys/block/md_d%d/md/component_size",
-   minor(stb.st_rdev)/16);
+   minor(stb.st_rdev)/64);
fd = open(fname, O_RDONLY);
if (fd < 0)
return 0;
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: EVMS causing problems with mdadm?

2006-04-23 Thread Neil Brown
On Sunday April 23, [EMAIL PROTECTED] wrote:
> Did my latest updates for my Kubuntu (Ubuntu KDE variant) this
> morning, and noticed that EVMS has now "taken control" of my RAID
> array. Didn't think much about it until I tried to make a RAID-1 array
> with two disks I've just added to the system. Trying to do a create
> verbose tells me that device /dev/md1 (or 2 or 3 - I tried a couple
> just to see) doesn't exist. And in fact, there are no block devices
> listed beyond md0.

Sounds like udev is in use rather than a static /dev.

Add '--auto=md' to the mdadm command line, and it will create the
devices for you.  Or --auto=part if you want partitioned arrays.  See
man page for more details.

I suspect this might need to be come the default in another year or so

NeilBrown

> 
> Trying to go into EVMS to create a new Volume brings up the screen for
> doing so, but doesn't actually let you create a new one.
> 
> Any suggestion of what I might be doing wrong, or is this a bug I
> should be reporting to the Ubuntu folks?
> -
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to [EMAIL PROTECTED]
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: to be or not to be...

2006-04-23 Thread Neil Brown
On Sunday April 23, [EMAIL PROTECTED] wrote:
> Hi all,
>   to make a long story very very shorty:
>   a) I create /dev/md1, kernel latest rc-2-git4 and mdadm-2.4.1.tgz,
>  with this command:
>  /root/mdadm -Cv /dev/.static/dev/.static/dev/.static/dev/md1 \
>  --bitmap-chunk=1024 --chunk=256 --assume-clean --bitmap=internal 
> \
   ^^
>  -l5 -n5 /dev/hda2 /dev/hdb2 /dev/hde2 /dev/hdf2 missing
>   

 From the man page:
   --assume-clean
  Tell mdadm that the array pre-existed and is known to be  clean.
  It  can be useful when trying to recover from a major failure as
  you can be sure that no data will be affected unless  you  actu-
  ally  write  to  the array.  It can also be used when creating a
  RAID1 or RAID10 if you want to avoid the initial resync, however
  this  practice - while normally safe - is not recommended.   Use
   ^^^
  this ony if you really know what you are doing.
  ^^

So presumably you know what you are doing, and I wonder why you bother
to ask us :-)
Ofcourse, if you don't know what you are doing, then I suggest
dropping the --assume-clean.  
In correct use of this flag can lead to data corruption.  This is
particularly true if your array goes degraded, but is also true while
your array isn't degraded.  In this case it is (I think) very unusual
and may not be the cause of your corruption, but you should avoid
using the flag anyway.


>   b) dm-encrypt /dev/md1
>   
>   c) create fs with:
>  mkfs.ext3 -O dir_index -L 'tritone' -i 256000 /dev/mapper/raidone
>   
>   d) export it via nfs (mounting /dev/mapper/raidone as ext2)
  

Why not ext3?

> 
>   e) start to cp-ing files
> 
>   f) after 1 TB of written data, with no problem/warning, one of the
>   not-in-raid-array HD freeze

This could signal a bad controller.  If it does, then you cannot trust
any drives.

NeilBrown
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: Recovery speed at 1MB/s/device, unable to change

2006-04-23 Thread Anssi Hannula
Anssi Hannula wrote:
> The speed is only 2000K/sec, even after I set:
> ---
> # cat /proc/sys/dev/raid/speed_limit_min
> 1
> # cat /proc/sys/dev/raid/speed_limit_max
> 40
> ---
> 
> The system is about 90% idle, so there should be more bandwidth.
> 
> ---
> # cat /proc/version
> Linux version 2.6.17-rc2-git4 ([EMAIL PROTECTED]) (gcc version 4.0.1
> (4.0.1-5mdk for Mandriva Linux release 2006.0)) #1 Sun Apr 23 00:56:30
> EEST 2006
> ---

Hmm... I don't know if this is related, but something seems to be really
wrong.

I run the following set of commands:

# dd if=/dev/zero of=ohase count=5K
# dd if=/dev/zero of=ohase2 count=5K
# losetup /dev/loop0 ohase
# losetup /dev/loop1 ohase2
# mdadm --create /dev/md_d1 -ap --level=5 --raid-devices=2 /dev/loop0
missing
# fdisk /dev/md_d1
(created one partition)
# mkfs.ext3 /dev/md_d1p1
# mount /dev/md_d1p1 /mnt/test
# echo jopajoo > /mnt/iso/test
# mdadm --manage /dev/md_d1 --add /dev/loop1
# mdadm --grow /dev/md_d1 --raid-devices=3 --backup-file backupfile
mdadm: Need to backup 128K of critical section..
mdadm: /dev/md_d1: Cannot get array details from sysfs

Strace shows that it's trying to access
"/sys/block/md_d4/md/component_size".

Why is this?

-- 
Anssi Hannula

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Recovery speed at 1MB/s/device, unable to change

2006-04-23 Thread Anssi Hannula
I created a raid array:
---
# mdadm --create /dev/md_d0 -ap --level=5 --raid-devices=2 \
/dev/sda1 missing
---

Then partitioned it:
---
# LANGUAGE=C fdisk -l /dev/md_d0

Disk /dev/md_d0: 250.0 GB, 250056605696 bytes
2 heads, 4 sectors/track, 61048976 cylinders
Units = cylinders of 8 * 512 = 4096 bytes

  Device Boot  Start End  Blocks   Id  System
/dev/md_d0p1   161048976   244195902   83  Linux
---

Then made a ext3 filesystem in /dev/md_d0p1 with mke2fs.

Then I added another device:
---
# mdadm --manage /dev/md_d0 --add /dev/sdb1
---

Currently status is:
---
# cat /proc/mdstat
Personalities : [raid0] [raid5] [raid4]
md_d0 : active raid5 sdb1[2] sda1[0]
  244195904 blocks level 5, 64k chunk, algorithm 2 [2/1] [U_]
  [>]  recovery =  1.0% (2588800/244195904)
finish=2100.4min speed=1916K/sec

unused devices: 
---

The speed is only 2000K/sec, even after I set:
---
# cat /proc/sys/dev/raid/speed_limit_min
1
# cat /proc/sys/dev/raid/speed_limit_max
40
---

The system is about 90% idle, so there should be more bandwidth.

---
# cat /proc/version
Linux version 2.6.17-rc2-git4 ([EMAIL PROTECTED]) (gcc version 4.0.1
(4.0.1-5mdk for Mandriva Linux release 2006.0)) #1 Sun Apr 23 00:56:30
EEST 2006
---


Any tips?

-- 
Anssi Hannula

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: disks becoming slow but not explicitly failing anyone?

2006-04-23 Thread Mark Hahn
> I've seen a lot of cheap disks say (generally deep in the data sheet
> that's only available online after much searching and that nobody ever
> reads) that they are only reliable if used for a maximum of twelve hours
> a day, or 90 hours a week, or something of that nature. Even server

I haven't, and I read lots of specs.  they _will_ sometimes say that 
non-enterprise drives are "intended" or "designed" for a 8x5 desktop-like
usage pattern.  to the normal way of thinking about reliability, this would 
simply mean a factor of 4.2x lower reliability - say from 1M to 250K hours
MTBF.  that's still many times lower rate of failure than power supplies or 
fans.

> It still stuns me that anyone would ever voluntarily buy drives that
> can't be left switched on (which is perhaps why the manufacturers hide

I've definitely never seen any spec that stated that the drive had to be 
switched off.  the issue is really just "what is the designed duty-cycle?"

I run a number of servers which are used as compute clusters.  load is
definitely 24x7, since my users always keep the queues full.  but the servers
are not maxed out 24x7, and do work quite nicely with desktop drives
for years at a time.  it's certainly also significant that these are in a 
decent machineroom environment.

it's unfortunate that disk vendors aren't more forthcoming with their drive
stats.  for instance, it's obvious that "wear" in MTBF terms would depend 
nonlinearly on the duty cycle.  it's important for a customer to know where 
that curve bends, and to try to stay in the low-wear zone.  similarly, disk
specs often just give a max operating temperature (often 60C!), which is 
almost disingenuous, since temperature has a superlinear effect on reliability.

a system designer needs to evaluate the expected duty cycle when choosing
disks, as well as many other factors which are probably more important.
for instance, an earlier thread concerned a vast amount of read traffic 
to disks resulting from atime updates.  obviously, just mounting noatime 
will improve the system's reliability.  providing a bit more memory on a 
fileserver to cache and eliminate IOs is another great way to help out.
simply using more disks also decreases the load per disk, though this is 
clearly only a win if it's the difference in staying out of the disks 
"duty-cycle danger zone" (since more disks divide system MTBF).

regards, mark hahn.

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: replace disk in raid5 without linux noticing?

2006-04-23 Thread Martin Cracauer
Carlos Carvalho wrote on Sat, Apr 22, 2006 at 02:48:23PM -0300: 
> Martin Cracauer (cracauer@cons.org) wrote on 22 April 2006 11:08:
>  >> stop the array
>  >> dd warning disk => new one
>  >> remove warning disk
>  >> assemble the array again with the new disk
>  >> 
>  >> The inconvenience is that you don't have the array during the copy.
>  >
>  >Stopping the array and restarting it as readonly will give you access
>  >to the data while that copy is in progress.
> 
> Yes but then you could just switch it to read-only without stopping.

I believe that would be fine to do the whole operation.  Filesystem
read-only, then md read-only, copy disk, then you need to unmount and
stop the md to restart it with the new disk.

If the final disk change involves a powerdown and putting the new disk
on the physical interface that the old one was on it should be
transparent.

%%

BTW, last time I tested a Linux software RAID-5 by ripping out an
active disk I noticed that while the filesystem stayed up and usable,
a currently ongoing system call would not return and block forever.

Is that a know behaviour?

Martin
-- 
%%%
Martin Cracauerhttp://www.cons.org/cracauer/
FreeBSD - where you want to go, today.  http://www.freebsd.org/
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


EVMS causing problems with mdadm?

2006-04-23 Thread Ewan Grantham
Did my latest updates for my Kubuntu (Ubuntu KDE variant) this
morning, and noticed that EVMS has now "taken control" of my RAID
array. Didn't think much about it until I tried to make a RAID-1 array
with two disks I've just added to the system. Trying to do a create
verbose tells me that device /dev/md1 (or 2 or 3 - I tried a couple
just to see) doesn't exist. And in fact, there are no block devices
listed beyond md0.

Trying to go into EVMS to create a new Volume brings up the screen for
doing so, but doesn't actually let you create a new one.

Any suggestion of what I might be doing wrong, or is this a bug I
should be reporting to the Ubuntu folks?
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: to be or not to be...

2006-04-23 Thread Molle Bestefich
gelma wrote:
> first run: lot of strange errors report about impossible i_size
> values, duplicated blocks, and so on

You mention only filesystem errors, no block device related errors.
In this case, I'd say that it's more likely that dm-crypt is to blame
rather than MD.

I think you should try the dm-devel mailing list.  Posting a complete
log of everything that has happened would probably be a good thing.

I have no experience with dm-crypt, but I do have experience with
another dm target (dm-snapshot), which iss very good at destroying my
data.

If you want a stable solution for encrypting your files, I can
recommend loop-aes.
loop-aes has very well thought-through security, the docs are concise
but have wide coverage,
it has good backwards compatibility - probably not your biggest
concern right now, but it is *very* nice to know that your data is
accessible, in the future as well as now - etc..  I've been using it
for a couple of years now, since the 2.2 or 2.4 days (can't remember),
and I've had nothing short of an absolutely *brilliant* experience
with it.

Enough propaganda for now, hope that you get your problem solved :-).
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: disks becoming slow but not explicitly failing anyone?

2006-04-23 Thread Nix
On 23 Apr 2006, Mark Hahn said:
>   some people claim that if you put a normal (desktop)
> drive into a 24x7 server (with real round-the-clock load), you should 
> expect failures quite promptly.  I'm inclined to believe that with 
> MTBF's upwards of 1M hour, vendors would not claim a 3-5yr warranty
> unless the actual failure rate was low, even if only running 8/24.

I've seen a lot of cheap disks say (generally deep in the data sheet
that's only available online after much searching and that nobody ever
reads) that they are only reliable if used for a maximum of twelve hours
a day, or 90 hours a week, or something of that nature. Even server
disks generally seem to say something like that, but the figure given is
more like `168 hours a week', i.e., constant use.

It still stuns me that anyone would ever voluntarily buy drives that
can't be left switched on (which is perhaps why the manufacturers hide
the info in such an obscure place), and I don't know what might go wrong
if you use the disk `too much': overheating?

But still it seems that there are crappy disks out there with very
silly limits on the time they can safely be used for.

(But this *is* the RAID list: we know that disks suck, right?)

-- 
`On a scale of 1-10, X's "brokenness rating" is 1.1, but that's only
 because bringing Windows into the picture rescaled "brokenness" by
 a factor of 10.' --- Peter da Silva
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


to be or not to be...

2006-04-23 Thread gelma
Hi all,
to make a long story very very shorty:
a) I create /dev/md1, kernel latest rc-2-git4 and mdadm-2.4.1.tgz,
   with this command:
   /root/mdadm -Cv /dev/.static/dev/.static/dev/.static/dev/md1 
--bitmap-chunk=1024 --chunk=256 --assume-clean --bitmap=internal -l5 -n5 
/dev/hda2 /dev/hdb2 /dev/hde2 /dev/hdf2 missing

b) dm-encrypt /dev/md1

c) create fs with:
   mkfs.ext3 -O dir_index -L 'tritone' -i 256000 /dev/mapper/raidone

d) export it via nfs (mounting /dev/mapper/raidone as ext2)

e) start to cp-ing files

f) after 1 TB of written data, with no problem/warning, one of the
not-in-raid-array HD freeze

g) reboot, check with:
   fsck -C -D -y /dev/mapper/raidone
   
   a) first run: lot of strange errors report about impossible
   i_size values, duplicated blocks, and so on, but it ends without
   complain, saying everything is fixed.
   b) mount it (as ext3), everything, at first glance, seems good
   (I will check checksum tomorrow) as
   number/size/filename/directory place of files. In /lost+found
   some files, but nothing "real". I mean, special files/devices,
   that never were on that fs, with giga/tera size (holes, of
   course), with chattr bits randomly setted.
   when I try to remove them I've got a kernel complain about
   offset in dir /lost+found.
   c) fsck again, after everything is fine

Now the cloning from old storage is going on, and now I'm wondering
if "--assume-clean" could be the reason of what happens.
btw, hardware passed usual test (memtest, cpuburn, ecc).

thanks a lot for your time and sorry for my terrible english,
gelma
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html