WRT seek performance, we're doing 2500 seeks per second on the
Sun/Thumper on 36 disks.
Luke,
Have you had time to run benchmarksql against it yet? I'm just curious
about the IO seeks/s vs. transactions/minute correlation...
/Mikael
---(end of
On 8/3/06, Luke Lonergan [EMAIL PROTECTED] wrote:
Merlin,
moving a gigabyte around/sec on the server, attached or no,
is pretty heavy lifting on x86 hardware.
Maybe so, but we're doing 2GB/s plus on Sun/Thumper with software RAID
and 36 disks and 1GB/s on a HW RAID with 16 disks, all SATA.
On 7/18/06, Alex Turner [EMAIL PROTECTED] wrote:
Remember when it comes to OLTP, massive serial throughput is not gonna help
you, it's low seek times, which is why people still buy 15k RPM drives, and
why you don't necessarily need a honking SAS/SATA controller which can
harness the full
Merlin,
moving a gigabyte around/sec on the server, attached or no,
is pretty heavy lifting on x86 hardware.
Maybe so, but we're doing 2GB/s plus on Sun/Thumper with software RAID
and 36 disks and 1GB/s on a HW RAID with 16 disks, all SATA.
WRT seek performance, we're doing 2500 seeks per
From: Alex Turner [EMAIL PROTECTED]
Sent: Jul 18, 2006 12:21 AM
To: Ron Peacetree [EMAIL PROTECTED]
Cc: Mikael Carneholm [EMAIL PROTECTED], pgsql-performance@postgresql.org
Subject: Re: [PERFORM] RAID stripe size question
On 7/17/06, Ron Peacetree [EMAIL PROTECTED] wrote:
-Original Message
This is a relatively low end HBA with 1 4Gb FC on it. Max sustained
IO on it is going to be ~320MBps. Or ~ enough for an 8 HD RAID 10 set
made of 75MBps ASTR HD's.
Looking at http://h30094.www3.hp.com/product.asp?sku=2260908extended=1,
I notice that the controller has a Ultra160 SCSI interface
Title: Re: [PERFORM] RAID stripe size question
Mikael,
On 7/18/06 6:34 AM, Mikael Carneholm [EMAIL PROTECTED] wrote:
However, what's more important is the seeks/s - ~530/s on a 28 disk
array is quite lousy compared to the 1400/s on a 12 x 15Kdisk array
I'm getting 2500 seeks/second on a 36
This is a great testament to the fact that very often software RAID will seriously outperform hardware RAID because the OS guys who implemented it took the time to do it right, as compared with some controller manufacturers who seem to think it's okay to provided sub-standard performance.
Based on
On Tue, 2006-07-18 at 14:27, Alex Turner wrote:
This is a great testament to the fact that very often software RAID
will seriously outperform hardware RAID because the OS guys who
implemented it took the time to do it right, as compared with some
controller manufacturers who seem to think it's
], Mikael Carneholm [EMAIL PROTECTED],
Ron Peacetree [EMAIL PROTECTED], pgsql-performance@postgresql.org
Subject: Re: [PERFORM] RAID stripe size question
On Tue, 2006-07-18 at 14:27, Alex Turner wrote:
This is a great testament to the fact that very often software RAID
will seriously outperform
[EMAIL PROTECTED]
Sent: Jul 18, 2006 3:37 PM
To: Alex Turner [EMAIL PROTECTED]
Cc: Luke Lonergan [EMAIL PROTECTED], Mikael Carneholm [EMAIL PROTECTED],
Ron Peacetree [EMAIL PROTECTED], pgsql-performance@postgresql.org
Subject: Re: [PERFORM] RAID stripe size question
On Tue, 2006-07-18 at 14:27
According to
http://www.pcguide.com/ref/hdd/perf/raid/concepts/perfStripe-c.html, it seems
to be the other way around?
(As stripe size is decreased, files are broken into smaller and smaller
pieces. This increases the number of drives
that an average file will use to hold all the blocks
With 18 disks dedicated to data, you could make 100/7*9 seeks/second (7ms av seeks time, 9 independant units) which is 128seeks/second writing on average 64kb of data, which is 4.1MB/sec throughput worst case, probably 10x best case so 40Mb/sec - you might want to take more disks for your data and
Message-
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Michael
Stone
Sent: den 17 juli 2006 02:04
To: pgsql-performance@postgresql.org
Subject: Re: [PERFORM] RAID stripe size question
On Mon, Jul 17, 2006 at 12:52:17AM +0200, Mikael Carneholm wrote:
I have finally gotten my hands
Hi, Mikael,
Mikael Carneholm wrote:
An 0+1 array of 4 disks *could* be enough, but I'm still unsure how WAL
activity correlates to normal data activity (is it 1:1, 1:2, 1:4,
...?)
I think the main difference is that the WAL activity is mostly linear,
where the normal data activity is rather
I think the main difference is that the WAL activity is mostly linear,
where the normal data activity is rather random access.
That was what I was expecting, and after reading
http://www.pcguide.com/ref/hdd/perf/raid/concepts/perfStripe-c.html I
figured that a different stripe size for the WAL
Hi, Mikael,
Mikael Carneholm wrote:
This is something I'd also would like to test, as a common best-practice
these days is to go for a SAME (stripe all, mirror everything) setup.
From a development perspective it's easier to use SAME as the developers
won't have to think about physical
This is something I'd also would like to test, as a common
best-practice these days is to go for a SAME (stripe all, mirror
everything) setup.
From a development perspective it's easier to use SAME as the
developers won't have to think about physical location for new
tables/indices, so if
From: Mikael Carneholm [EMAIL PROTECTED]
Sent: Jul 16, 2006 6:52 PM
To: pgsql-performance@postgresql.org
Subject: [PERFORM] RAID stripe size question
I have finally gotten my hands on the MSA1500 that we ordered some time
ago. It has 28 x 10K 146Gb drives,
Unless I'm missing something, the only
On Mon, Jul 17, 2006 at 09:40:30AM -0400, Ron Peacetree wrote:
Unless I'm missing something, the only FC or SCSI HDs of ~147GB capacity are
15K, not 10K.
(unless they are old?)
There are still 146GB SCSI 1rpm disks being sold here, at least.
/* Steinar */
--
Homepage:
On 7/17/06, Mikael Carneholm [EMAIL PROTECTED] wrote:
This is something I'd also would like to test, as a common best-practice these days is to go for a SAME (stripe all, mirroreverything) setup. From a development perspective it's easier to use SAME as the
developers won't have to think about
Unless I'm missing something, the only FC or SCSI HDs of ~147GB capacity are
15K, not 10K.
In the spec we got from HP, they are listed as model 286716-B22
(http://www.dealtime.com/xPF-Compaq_HP_146_8_GB_286716_B22) which seems to run
at 10K. Don't know how old those are, but that's what we got
Mikael Carneholm wrote:
Btw, here's the bonnie++ results from two different array sets (10+18,
4+24) on the MSA1500:
LUN: DATA, 24 disks, stripe size 64K
-
Version 1.03 --Sequential Output-- --Sequential Input-
--Random-
-Original Message-
From: Mikael Carneholm [EMAIL PROTECTED]
Sent: Jul 17, 2006 5:16 PM
To: Ron Peacetree [EMAIL PROTECTED], pgsql-performance@postgresql.org
Subject: RE: [PERFORM] RAID stripe size question
15Krpm HDs will have average access times of 5-6ms. 10Krpm ones of 7-8ms
On 7/17/06, Ron Peacetree [EMAIL PROTECTED] wrote:
-Original Message-From: Mikael Carneholm [EMAIL PROTECTED]Sent: Jul 17, 2006 5:16 PMTo: RonPeacetree
[EMAIL PROTECTED], pgsql-performance@postgresql.orgSubject: RE: [PERFORM] RAID stripe size question15Krpm HDs will have average access
On Mon, Jul 17, 2006 at 12:52:17AM +0200, Mikael Carneholm wrote:
Now to the interesting part: would it make sense to use different stripe
sizes on the separate disk arrays? In theory, a smaller stripe size
(8-32K) should increase sequential write throughput at the cost of
decreased
On Mon, Jul 17, 2006 at 12:52:17AM +0200, Mikael Carneholm wrote:
I have finally gotten my hands on the MSA1500 that we ordered some time
ago. It has 28 x 10K 146Gb drives, currently grouped as 10 (for wal) +
18 (for data). There's only one controller (an emulex), but I hope
You've got 1.4TB
27 matches
Mail list logo