Re: spare group

2007-06-13 Thread Tomka Gergely

Neil Brown írta:

On Tuesday June 12, [EMAIL PROTECTED] wrote:
I am very sorry, but it wont works with .9 superblocks also :( We 
missing something small, but important here. Before you start to code. 
mdadm was running in monitor mode, and reported a Fail. mdadm is the 
latest version, 2.6.2.


tg


Hmmm. 
[tests code]


Yes, you are right.  It looks like a bug was introduced in 2.6 which
broke various aspects of --monitor.  I guess I need to add some
--monitor tests to my regression test suite.

This patch should fix it.

Thanks again,
NeilBrown


Thanks, the patch is working.

tg
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: spare group

2007-06-12 Thread Tomka Gergely

Neil Brown írta:

(reads code).

Ahhh. You are using version-1 superblocks aren't you?  That code only
works for version-0.90 superblocks.  That was careless of me.  It
shouldn't be hard to make it work more generally, but it looks like it
will be slightly more than trivial.  I'll try to get you a patch in
the next day or so (feel free to remind me if I seem to have
forgotten).

Thanks for testing and reporting this problem.

NeilBrown


# mdadm26 --detail /dev/md0
/dev/md0:
Version : 00.90.03
  Creation Time : Tue Jun 12 10:31:08 2007
 Raid Level : raid5
 Array Size : 19534848 (18.63 GiB 20.00 GB)
  Used Dev Size : 9767424 (9.31 GiB 10.00 GB)
   Raid Devices : 3
  Total Devices : 4
Preferred Minor : 0
Persistence : Superblock is persistent

Update Time : Tue Jun 12 10:33:35 2007
  State : clean
 Active Devices : 3
Working Devices : 4
 Failed Devices : 0
  Spare Devices : 1

 Layout : left-symmetric
 Chunk Size : 64K

   UUID : 5fd83926:01739a55:36458d87:119f8994 (local to host 
ursula)

 Events : 0.4

Number   Major   Minor   RaidDevice State
   0   810  active sync   /dev/sda1
   1   8   171  active sync   /dev/sdb1
   2   8   332  active sync   /dev/sdc1

   3   8   49-  spare   /dev/sdd1

/dev/md1:
Version : 00.90.03
  Creation Time : Tue Jun 12 10:31:29 2007
 Raid Level : raid5
 Array Size : 19534848 (18.63 GiB 20.00 GB)
  Used Dev Size : 9767424 (9.31 GiB 10.00 GB)
   Raid Devices : 3
  Total Devices : 3
Preferred Minor : 1
Persistence : Superblock is persistent

Update Time : Tue Jun 12 10:36:18 2007
  State : clean, degraded
 Active Devices : 2
Working Devices : 2
 Failed Devices : 1
  Spare Devices : 0

 Layout : left-symmetric
 Chunk Size : 64K

   UUID : 815d6fc4:a55c2602:36458d87:119f8994 (local to host 
ursula)

 Events : 0.6

Number   Major   Minor   RaidDevice State
   0   8   650  active sync   /dev/sde1
   1   8   811  active sync   /dev/sdf1
   2   002  removed

   3   8   97-  faulty spare   /dev/sdg1

ARRAY /dev/md1 level=raid5 num-devices=3 spare-group=ubul 
UUID=815d6fc4:a55c2602:36458d87:119f8994
ARRAY /dev/md0 level=raid5 num-devices=3 spares=1 spare-group=ubul 
UUID=5fd83926:01739a55:36458d87:119f8994


I am very sorry, but it wont works with .9 superblocks also :( We 
missing something small, but important here. Before you start to code. 
mdadm was running in monitor mode, and reported a Fail. mdadm is the 
latest version, 2.6.2.


tg
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


strange test results

2007-03-19 Thread Tomka Gergely
Hi!

I am running tests on our new test device. The device has 2x2 core Xeon, 
intel 5000 chipset, two 3ware sata raid card on pcie, and 15 sata2 disks, 
running debian etch. More info at the bottom.

The first phase of the test is probing various raid levels. So i 
configured the cards to 15 JBOD disks, and hacked together a testing 
script. The script builds raid arrays, waits for sync, and then runs this 
command:

iozone -eM -s 4g -r 1024 -i0 -i1 -i2 -i8 -t16 -+u

The graphs of the results here:

http://gergely.tomka.hu/dt/index.html

And i have a lots of questions.

http://gergely.tomka.hu/dt/1.html

This graph is crazy, like thunderbolts. But the raid50 is generally slower 
than raid5. Why?

http://gergely.tomka.hu/dt/3.html

This is the only graph i can explain :)

http://gergely.tomka.hu/dt/4.html

With random readers, why raid0 slowing down? And why raid10 faster than 
raid0?

http://gergely.tomka.hu/dt/2.html

Why raid6 cant became faster, with multiple disks, as raid5  50?

So lots of questions. I am generally surprised by the non-linearity of 
some results and the lack of acceleration with more disks on other 
results. And now, the details:

Hardware:

Base Board Information
Manufacturer: Supermicro
Product Name: X7DB8
Processor Information
Socket Designation: LGA771/CPU1
Type: Central Processor
Family: Xeon
Manufacturer: Intel
ID: 64 0F 00 00 FF FB EB BF
Signature: Type 0, Family 15, Model 6, Stepping 4
(two cpus)
Memory Device
Array Handle: 0x0017
Error Information Handle: No Error
Total Width: 72 bits
Data Width: 64 bits
Size: 1024 MB
Form Factor: DIMM
Set: 1
Locator: DIMM x 4
Bank Locator: Bank1
Type: DDR2
Type Detail: Synchronous
Speed: 533 MHz (1.9 ns)
Manufacturer: Not Specified
Serial Number: Not Specified
Asset Tag: Not Specified
Part Number: Not Specified
(two of this also)

ursula:~# tw_cli show

Ctl   ModelPorts   Drives   Units   NotOpt   RRate   VRate   BBU

c09590SE-8ML   8   77   01   1   -
c19590SE-8ML   8   88   01   1   -

The tests generally:
mdadm
mkfs.xfs
blockdev --setra 524288 md (maybe not a good idea for multiple arrays)
do iozone test

raid10 is two disks raid1s in raid0 and raid50 is three disk raid6s in 
raid0.

These test runs for a week, and now slowly finishing. For this reason, 
replicatong the test to filter out accidents not a good option.

Any comments?

-- 
Tomka Gergely, [EMAIL PROTECTED]
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


sw raid0 read bottleneck

2007-03-13 Thread Tomka Gergely
Hi!

I am currently testing 3ware raid cards. Now i have 15 disks, and on these 
a swraid0. The write speed seems good (700 MBps), but the read performance 
only 350 MBps. Another problem when i try to read with two process, then 
the _sum_ of the read speeds fall back to 200 MBps. So there is a 
bottleneck, or something i need to know, but i dont have ideas. 

The details:

/dev/md0:
Version : 00.90.03
  Creation Time : Tue Mar 13 16:57:32 2007
 Raid Level : raid0
 Array Size : 7325797440 (6986.43 GiB 7501.62 GB)
   Raid Devices : 15
  Total Devices : 15
Preferred Minor : 0
Persistence : Superblock is persistent

Update Time : Tue Mar 13 16:57:32 2007
  State : clean
 Active Devices : 15
Working Devices : 15
 Failed Devices : 0
  Spare Devices : 0

 Chunk Size : 64K

# uname -a
Linux ursula 2.6.18-4-686-bigmem #1 SMP Wed Feb 21 17:30:22 UTC 2007 i686 
GNU/Linux

# xfs_info /mnt/
meta-data=/dev/md0   isize=256agcount=32, agsize=57232784 
blks
 =   sectsz=512   attr=0
data =   bsize=4096   blocks=1831449088, 
imaxpct=25
 =   sunit=16 swidth=240 blks, unwritten=1
naming   =version 2  bsize=4096  
log  =internal   bsize=4096   blocks=32768, version=1
 =   sectsz=512   sunit=0 blks
realtime =none   extsz=983040 blocks=0, rtextents=0

(all software Debian Etch)

four Intel(R) Xeon(TM) CPU 3.00GHz

two 3ware 9590SE-8ML on PCIe

Intel Corporation 5000P Chipset


-- 
Tomka Gergely, [EMAIL PROTECTED]
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: sw raid0 read bottleneck

2007-03-13 Thread Tomka Gergely
On Tue, 13 Mar 2007, Justin Piszcz wrote:

 Have you tried increasing your readahead values for the md device?

Yes. No real change. According to my humble mental image, readahead not a 
too useful thing, when we read 1-4 thread with sdd. The io subsystem 
already reading with the possible maximum speed, so don't have time to 
read ahead. Correct me, if i wrong. 

-- 
Tomka Gergely, [EMAIL PROTECTED]
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: sw raid0 read bottleneck

2007-03-13 Thread Tomka Gergely
On Tue, 13 Mar 2007, Tomka Gergely wrote:

 On Tue, 13 Mar 2007, Justin Piszcz wrote:
 
  Have you tried increasing your readahead values for the md device?
 
 Yes. No real change. According to my humble mental image, readahead not a 
 too useful thing, when we read 1-4 thread with sdd. The io subsystem 
 already reading with the possible maximum speed, so don't have time to 
 read ahead. Correct me, if i wrong. 

I was wrong, readahead can speed things up, to 450 MBps.

-- 
Tomka Gergely, [EMAIL PROTECTED]
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


Re: sw raid0 read bottleneck

2007-03-13 Thread Tomka Gergely
On Wed, 14 Mar 2007, Neil Brown wrote:

 On Tuesday March 13, [EMAIL PROTECTED] wrote:
  On Tue, 13 Mar 2007, Tomka Gergely wrote:
  
   On Tue, 13 Mar 2007, Justin Piszcz wrote:
   
Have you tried increasing your readahead values for the md device?
   
   Yes. No real change. According to my humble mental image, readahead not a 
   too useful thing, when we read 1-4 thread with sdd. The io subsystem 
   already reading with the possible maximum speed, so don't have time to 
   read ahead. Correct me, if i wrong. 
  
  I was wrong, readahead can speed things up, to 450 MBps.
 
 Can you tell use what read-ahead size you needed?
 
 15 drives and 64K chunks gives 960K per stripe.
 The raid0 code should set the read-ahead to twice that: 1920K
 which I would have thought would be enough, but apparently not.

blockdev --setra 262144 /dev/md0 gives me 650+ MB/s with 4 threads 
(paralell running sdd). Lower values give lower speeds, greater values not 
giving higher speeds.

-- 
Tomka Gergely, [EMAIL PROTECTED]
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html


beginner error detection

2007-02-23 Thread Tomka Gergely
Hi!

I have a simple sw raid1, over two sata disks. One of the disks started to 
complain (s.m.a.r.t. errors). I think in the near future i witness a disk 
failure. But i don't know how this thing is happening with raid1, so i 
have some questions. If these questions answered somewhere (faq, manpage, 
url), then feel free to redirect me to this source(s). 

Can the raid1 detect and handle disk errors? If one block goes wrong, how 
can the raid1 driver choose which was the correct, original value?

Sata systems can die gracefully? When in a good scsi system happens a 
total disk failure, then the scsi makes the disk fail, mdadm removes 
the disk from the array, and in the morning i see a nice e-mail. When a 
PATA disk dies, the system goes down, so i need to call a cab. I dont know 
how the sata behaves in this situation.

The kernel 2.6.20 (with skas patch), the controller :
nVidia Corporation CK804 Serial ATA Controller (rev f3)

Thanks.

-- 
Tomka Gergely, [EMAIL PROTECTED]
-
To unsubscribe from this list: send the line unsubscribe linux-raid in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html