Re: question : raid bio sector size

2006-03-29 Thread Raz Ben-Jehuda(caro)
I was refering to bios reaching make_request in raid5.c . I would be more precise. I am dd'ing dd if=/dev/md1 of=/dev/zero bs=1M count=1 skip=10 I have added the following printk in make_request printk (%d:,bio-bi_size) I am getting sector sizes. 512:512:512:512:512 I suppose they gathered in

Re: question : raid bio sector size

2006-03-29 Thread Neil Brown
On Wednesday March 29, [EMAIL PROTECTED] wrote: I was refering to bios reaching make_request in raid5.c . I would be more precise. I am dd'ing dd if=/dev/md1 of=/dev/zero bs=1M count=1 skip=10 I have added the following printk in make_request printk (%d:,bio-bi_size) I am getting sector

(X)FS corruption on 2 SATA disk RAID 1

2006-03-29 Thread JaniD++
Hello, list, I think, this is generally hardware error, but looks like software problem too. At this point there is no dirty data in memory! Cheers, Janos [EMAIL PROTECTED] /]# cmp -b /dev/sda1 /dev/sdb1 /dev/sda1 /dev/sdb1 differ: byte 68881481729, line 308395510 is 301 M-A 74 [EMAIL

Re: question : raid bio sector size

2006-03-29 Thread Raz Ben-Jehuda(caro)
man .. very very good. blockdev --getsz says 512. On 3/29/06, Neil Brown [EMAIL PROTECTED] wrote: On Wednesday March 29, [EMAIL PROTECTED] wrote: I was refering to bios reaching make_request in raid5.c . I would be more precise. I am dd'ing dd if=/dev/md1 of=/dev/zero bs=1M count=1

Re: making raid5 more robust after a crash?

2006-03-29 Thread Chris Allen
On Sat, Mar 18, 2006 at 08:13:48AM +1100, Neil Brown wrote: On Friday March 17, [EMAIL PROTECTED] wrote: Dear All, We have a number of machines running 4TB raid5 arrays. Occasionally one of these machines will lock up solid and will need power cycling. Often when this happens, the

Re: making raid5 more robust after a crash?

2006-03-29 Thread Neil Brown
On Wednesday March 29, [EMAIL PROTECTED] wrote: Thanks for your reply. As you guessed, this was a problem with our hardware/config and nothing to do with the raid software. I'm glad you have found your problem! Can anybody point me to the syntax I could use for saying: force rebuild the

addendum: was Re: recovering data on a failed raid-0 installation

2006-03-29 Thread Technomage
ok, guy and others. this is a followup to the case I am currently trying (still) to solve. synopsis: the general consensus is that raid0 writes in a striping fashion. However, the test case I have here doesn't appear to operate in the above described manner. what was observed was this: on

ANNOUNCE: mdadm 2.4 - A tool for managing Soft RAID under Linux

2006-03-29 Thread Neil Brown
I am pleased to announce the availability of mdadm version 2.4 It is available at the usual places: http://www.cse.unsw.edu.au/~neilb/source/mdadm/ and http://www.{countrycode}.kernel.org/pub/linux/utils/raid/mdadm/ mdadm is a tool for creating, managing and monitoring device arrays

RE: addendum: was Re: recovering data on a failed raid-0 installation

2006-03-29 Thread Guy
If what you say is true, then it was not a RAID0. It sounds like LINEAR. Do you have the original command used to create the array? Or the output from mdadm before you tried any recovery methods. The output must be from before you re-created the array. Output from commands like mdadm -D /dev/md0

Re: [PATCH] Add stripe cache entries to raid6 sysfs

2006-03-29 Thread Neil Brown
On Saturday March 25, [EMAIL PROTECTED] wrote: Raid-6 did not create sysfs entries for stripe cache Signed-off-by: Brad Campbell [EMAIL PROTECTED] --- diff -u vanilla/linux-2.6.16/drivers/md/raid6main.c linux-2.6.16/drivers/md/raid6main.c --- vanilla/linux-2.6.16/drivers/md/raid6main.c

[PATCH 001 of 3] md: Don't clear bits in bitmap when writing to one device fails during recovery.

2006-03-29 Thread NeilBrown
Currently a device failure during recovery leaves bits set in the bitmap. This normally isn't a problem as the offending device will be rejected because of errors. However if device re-adding is being used with non-persistent bitmaps, this can be a problem. Signed-off-by: Neil Brown [EMAIL

[PATCH 002 of 3] md: Remove some code that can sleep from under a spinlock.

2006-03-29 Thread NeilBrown
And remove the comments that were put in inplace of a fix too Signed-off-by: Neil Brown [EMAIL PROTECTED] ### Diffstat output ./drivers/md/md.c |8 +++- 1 file changed, 3 insertions(+), 5 deletions(-) diff ./drivers/md/md.c~current~ ./drivers/md/md.c --- ./drivers/md/md.c~current~

[PATCH 003 of 3] md: Raid-6 did not create sysfs entries for stripe cache

2006-03-29 Thread NeilBrown
Signed-off-by: Brad Campbell [EMAIL PROTECTED] Signed-off-by: Neil Brown [EMAIL PROTECTED] ### Diffstat output ./drivers/md/raid6main.c |2 ++ 1 file changed, 2 insertions(+) diff ./drivers/md/raid6main.c~current~ ./drivers/md/raid6main.c --- ./drivers/md/raid6main.c~current~ 2006-03-30

[PATCH 000 of 3] md: Introduction - assorted fixed for 2.6.16

2006-03-29 Thread NeilBrown
Following are three patches for md. The first fixes a problem that can cause corruption in fairly unusual circumstances (re-adding a device to a raid1 and suffering write-errors that are subsequntly fixed and the device is re-added again). The other two fix minor problems The are suitable to go

Re: [PATCH 001 of 3] md: Don't clear bits in bitmap when writing to one device fails during recovery.

2006-03-29 Thread Andrew Morton
NeilBrown [EMAIL PROTECTED] wrote: + if (!uptodate) { +int sync_blocks = 0; +sector_t s = r1_bio-sector; +long sectors_to_go = r1_bio-sectors; +/* make sure these bits doesn't get cleared. */ +do { +

Re: [PATCH 001 of 3] md: Don't clear bits in bitmap when writing to one device fails during recovery.

2006-03-29 Thread Neil Brown
On Wednesday March 29, [EMAIL PROTECTED] wrote: NeilBrown [EMAIL PROTECTED] wrote: + if (!uptodate) { + int sync_blocks = 0; + sector_t s = r1_bio-sector; + long sectors_to_go = r1_bio-sectors; + /* make sure these bits doesn't get cleared. */

making raid5 more robust against block errors

2006-03-29 Thread Mikael Abrahamsson
Is there any work going on to handle readerrors on a raid5 disk being handled by recreating the faulty block from the other disks and just rewriting the block, instead of kicking the disk out? I've problems on several occasions where two disks in a raid5 will have single sector errors and