Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-18 Thread Kenji Morishige
Regarding the DL585 etc boxes from HP, they seem to require external JBOD or
SCSI/SAS enclosures.  Does anyone have any particular preference on how these
units should be configured or speced?  I'm guessing I'll use the onboard SCSI
RAID 1 with the onboard drives for the OS, but will need 2 external channels
for the data and xlog.  Any recommendations there?

Sincerely,
Kenji

On Sat, Aug 12, 2006 at 04:20:19PM +, Roman Neuhauser wrote:
 # [EMAIL PROTECTED] / 2006-08-08 14:49:21 -0700:
  I am considering a setup such as this:
- At least dual cpu (possibly with 2 cores each)
- 4GB of RAM
- 2 disk RAID 1 array for root disk
- 4 disk RAID 1+0 array for PGDATA
- 2 disk RAID 1 array for pg_xlog
  
  Does anyone know a vendor that might be able provide such setup?
  
  I would look at the HP DL 385 or 585. The 385 is going to max a (2) dual 
  core cpus. The 585 is (4) dual core cpus.
 
 I don't know about DL385 or DL585, but DL380 seem to go south within
 1 year of heavy hitting; precisely the Smart Array RAID controllers
 (4 out of 6 disks suddenly red; insert new disks, ooops red as
 well).
 
 I've seen this happen several times, and came away with a conclusion
 that DL380 is sexy, but you don't want to marry it. Then again,
 maybe the DL385 is different, though I seem to remember that both
 G3 (Smart Array 5i) and G4 (6i) did this.
 
 -- 
 How many Vietnam vets does it take to screw in a light bulb?
 You don't know, man.  You don't KNOW.
 Cause you weren't THERE. http://bash.org/?255991

---(end of broadcast)---
TIP 1: if posting/reading through Usenet, please send an appropriate
   subscribe-nomail command to [EMAIL PROTECTED] so that your
   message can get through to the mailing list cleanly


Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-18 Thread Kenji Morishige
Thanks Arjen, 
I have unlimited rack space if I really need it.  Is serial/SAS really the
better route to go than SCSI these days? I'm so used to ordering SCSI that
I've been out of the loop with new disk enclosures and disk tech.  I been
trying to price out a HP DL585, but those are considerably more than the
Dells.  Is it worth waiting a few more weeks/months for Dell to release
something newer?

-Kenji

On Wed, Aug 09, 2006 at 07:35:22AM +0200, Arjen van der Meijden wrote:
 With such a budget you should easily be able to get something like:
 - A 1U high-performance server (for instance the Dell 1950 with 2x 
 Woodcrest 5160, 16GB of FB-Dimm memory, one 5i and one 5e perc raid 
 controller and some disks internally)
 - An external SAS direct attached disks storage enclosure full with 15k 
 rpm 36GB disks (for instance the MD1000, with 15x 36GB 15k disks)
 
 Going for the dell-solution would set you back only (including 
 savings) about $13-$14k. HP offers a similar solutions (a HP DL360G5 or 
 a DL380G5/DL385 with two MSA50's for instance) which also fit in your 
 budget afaik. The other players tend to be (a bit) more expensive, force 
 you to go with Fibre Channel or ancient SCSI external storage ;)
 
 If you'd like to have a product by a generic vendor, have a look at the 
 Adaptec JS50 SAS Jbod enclosure or Promise's Vtrak 300 (both offer 12 
 sas/sata bays in 2U) for storage.
 
 If you're limited to only 2U of rack space, its a bit more difficult to 
 get maximum I/O in your budget (you have basically space for about 8 or 
 12 3.5 disks (with generic suppliers) or 16 2.5 sff disks (with HP)).
 But you should still be able to have two top-off-the-line x86 cpu's (amd 
 opteron 285 or intel woorcrest 5160) and 16GB of memory (even FB Dimm, 
 which is pretty expensive).
 
 Best regards,
 
 Arjen van der Meijden
 
 
 On 8-8-2006 22:43, Kenji Morishige wrote:
 I've asked for some help here a few months ago and got some really helpfull
 answers regarding RAID controllers and server configuration.  Up until
 recently I've been running PostgreSQL on a two year old Dual Xeon 3.06Ghz
 machine with a single channel RAID controller (previously Adaptec 2200S, 
 but
 now changed to LSI MegaRAID). The 2U unit is from a generic vendor using 
 what
 I believe is a SuperMicro motherboard.  In the last week after upgrading 
 the
 RAID controller, the machine has had disk failure and some other issues. I
 would like to build a very reliable dedicated postgreSQL server that has 
 the
 ultimate possible performance and reliabily for around $20,000.  The data 
 set
 size is only currently about 4GB, but is increasing by approximately 50MB
 daily.  The server also requires about 500 connections and I have been
 monitoring about 100-200 queries per second at the moment.  I am planning 
 to
 run FreeBSD 6.1 if possible, but I am open to any other suggestions if it
 improves performance.
 
 I am considering a setup such as this:
   - At least dual cpu (possibly with 2 cores each)
   - 4GB of RAM
   - 2 disk RAID 1 array for root disk
   - 4 disk RAID 1+0 array for PGDATA
   - 2 disk RAID 1 array for pg_xlog
 
 Does anyone know a vendor that might be able provide such setup?  Any
 critique in this design? I'm thinking having a 2 channel RAID controller to
 seperate the PGDATA, root and pg_xlog.
 
 Sincerely,
 Kenji
 
 ---(end of broadcast)---
 TIP 2: Don't 'kill -9' the postmaster
 

---(end of broadcast)---
TIP 1: if posting/reading through Usenet, please send an appropriate
   subscribe-nomail command to [EMAIL PROTECTED] so that your
   message can get through to the mailing list cleanly


Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-18 Thread Arjen van der Meijden

Hi Kenji,

I'm not sure what you mean by 'something newer'? The intel 
woodcrest-cpu's are brand-new compared to the amd opterons. But if you 
need a 4-cpu config (I take it you want 8-cores in that case), Dell 
doesn't offer much. Whether something new will come, I don't know. I'm 
not sure when (or if?) a MP-Woodcrest will arrive and/or when Dell will 
start offering Opteron-servers.


Sas has been designed as the successor to SCSI.

As I see it, SAS has currently one major disadvantage. Lots of new 
servers are equipped with SAS-drives, a few nice SAS-raidcontrollers 
exist, but the availability of external enclosures for SAS is not 
widespread yet. So your options of going beyond (say) 8 disks per system 
are a bit limited.


There are of course advantages as well. The bus is much wider (you can 
have 4 lanes of 3Gbps each to an enclosure). You can mix sas and sata 
disks, so you could have two arrays in the same enclosure, one big 
storage bin and a very fast array or just use only sata disks on a sas 
controller. The cabling itself is also much simpler/more flexible 
(although using a hot-plug enclosure of course shields you mostly from 
that).
But whether its the right choice to make now? I'm not sure. We weren't 
to fond of investing a lot of money in an end-of-life system. And since 
we're a tech-website, we also had to worry about our being modern 
image, of course ;)


The main disadvantage I see in this case is, as said, the limited 
availability of external enclosures in comparison to SCSI and Fibre 
Channel. HP currently only offers their MSA50 (for the rather expensive 
SFF disks) while their MSA60 (normal disks) will not be available until 
somewhere in 2007 and Dell also only offers one enclosure, the MD1000.
The other big players offer nothing yet, as far as I know, while they 
normally offer several SCSI and/or FC-enclosures.
There are also some third-party enclosures (adaptec and promise for 
instance) available of course.


Best regards,

Arjen

On 18-8-2006 21:07, Kenji Morishige wrote:
Thanks Arjen, 
I have unlimited rack space if I really need it.  Is serial/SAS really the

better route to go than SCSI these days? I'm so used to ordering SCSI that
I've been out of the loop with new disk enclosures and disk tech.  I been
trying to price out a HP DL585, but those are considerably more than the
Dells.  Is it worth waiting a few more weeks/months for Dell to release
something newer?

-Kenji

On Wed, Aug 09, 2006 at 07:35:22AM +0200, Arjen van der Meijden wrote:

With such a budget you should easily be able to get something like:
- A 1U high-performance server (for instance the Dell 1950 with 2x 
Woodcrest 5160, 16GB of FB-Dimm memory, one 5i and one 5e perc raid 
controller and some disks internally)
- An external SAS direct attached disks storage enclosure full with 15k 
rpm 36GB disks (for instance the MD1000, with 15x 36GB 15k disks)


Going for the dell-solution would set you back only (including 
savings) about $13-$14k. HP offers a similar solutions (a HP DL360G5 or 
a DL380G5/DL385 with two MSA50's for instance) which also fit in your 
budget afaik. The other players tend to be (a bit) more expensive, force 
you to go with Fibre Channel or ancient SCSI external storage ;)


If you'd like to have a product by a generic vendor, have a look at the 
Adaptec JS50 SAS Jbod enclosure or Promise's Vtrak 300 (both offer 12 
sas/sata bays in 2U) for storage.


If you're limited to only 2U of rack space, its a bit more difficult to 
get maximum I/O in your budget (you have basically space for about 8 or 
12 3.5 disks (with generic suppliers) or 16 2.5 sff disks (with HP)).
But you should still be able to have two top-off-the-line x86 cpu's (amd 
opteron 285 or intel woorcrest 5160) and 16GB of memory (even FB Dimm, 
which is pretty expensive).


Best regards,

Arjen van der Meijden


On 8-8-2006 22:43, Kenji Morishige wrote:

I've asked for some help here a few months ago and got some really helpfull
answers regarding RAID controllers and server configuration.  Up until
recently I've been running PostgreSQL on a two year old Dual Xeon 3.06Ghz
machine with a single channel RAID controller (previously Adaptec 2200S, 
but
now changed to LSI MegaRAID). The 2U unit is from a generic vendor using 
what
I believe is a SuperMicro motherboard.  In the last week after upgrading 
the

RAID controller, the machine has had disk failure and some other issues. I
would like to build a very reliable dedicated postgreSQL server that has 
the
ultimate possible performance and reliabily for around $20,000.  The data 
set

size is only currently about 4GB, but is increasing by approximately 50MB
daily.  The server also requires about 500 connections and I have been
monitoring about 100-200 queries per second at the moment.  I am planning 
to

run FreeBSD 6.1 if possible, but I am open to any other suggestions if it
improves performance.

I am considering a setup such as this:
 - At least dual cpu (possibly 

Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-18 Thread Kenji Morishige
Thanks Arjen for your reply, this is definitely something to consider. I
think in our case, we are not too concerned with the tech image as much as if
the machine will allow us to scale the loads we need. I'm not sure if we
should worry so much about the IO bandwidth as we are not even close to
saturating 320MB/s.  I think stability, reliability, and ease-of-use and
recovery is our main concern at the moment.  I currently am runing a load
average of about .5 on a dual Xeon 3.06Ghz P4 setup.  How much CPU
performance improvement do you think the new woodcrest cpus are over these?

-Kenji

On Fri, Aug 18, 2006 at 09:41:55PM +0200, Arjen van der Meijden wrote:
 Hi Kenji,
 
 I'm not sure what you mean by 'something newer'? The intel 
 woodcrest-cpu's are brand-new compared to the amd opterons. But if you 
 need a 4-cpu config (I take it you want 8-cores in that case), Dell 
 doesn't offer much. Whether something new will come, I don't know. I'm 
 not sure when (or if?) a MP-Woodcrest will arrive and/or when Dell will 
 start offering Opteron-servers.
 
 Sas has been designed as the successor to SCSI.
 
 As I see it, SAS has currently one major disadvantage. Lots of new 
 servers are equipped with SAS-drives, a few nice SAS-raidcontrollers 
 exist, but the availability of external enclosures for SAS is not 
 widespread yet. So your options of going beyond (say) 8 disks per system 
 are a bit limited.
 
 There are of course advantages as well. The bus is much wider (you can 
 have 4 lanes of 3Gbps each to an enclosure). You can mix sas and sata 
 disks, so you could have two arrays in the same enclosure, one big 
 storage bin and a very fast array or just use only sata disks on a sas 
 controller. The cabling itself is also much simpler/more flexible 
 (although using a hot-plug enclosure of course shields you mostly from 
 that).
 But whether its the right choice to make now? I'm not sure. We weren't 
 to fond of investing a lot of money in an end-of-life system. And since 
 we're a tech-website, we also had to worry about our being modern 
 image, of course ;)
 
 The main disadvantage I see in this case is, as said, the limited 
 availability of external enclosures in comparison to SCSI and Fibre 
 Channel. HP currently only offers their MSA50 (for the rather expensive 
 SFF disks) while their MSA60 (normal disks) will not be available until 
 somewhere in 2007 and Dell also only offers one enclosure, the MD1000.
 The other big players offer nothing yet, as far as I know, while they 
 normally offer several SCSI and/or FC-enclosures.
 There are also some third-party enclosures (adaptec and promise for 
 instance) available of course.
 
 Best regards,
 
 Arjen
 
 On 18-8-2006 21:07, Kenji Morishige wrote:
 Thanks Arjen, 
 I have unlimited rack space if I really need it.  Is serial/SAS really the
 better route to go than SCSI these days? I'm so used to ordering SCSI that
 I've been out of the loop with new disk enclosures and disk tech.  I been
 trying to price out a HP DL585, but those are considerably more than the
 Dells.  Is it worth waiting a few more weeks/months for Dell to release
 something newer?
 
 -Kenji
 
 On Wed, Aug 09, 2006 at 07:35:22AM +0200, Arjen van der Meijden wrote:
 With such a budget you should easily be able to get something like:
 - A 1U high-performance server (for instance the Dell 1950 with 2x 
 Woodcrest 5160, 16GB of FB-Dimm memory, one 5i and one 5e perc raid 
 controller and some disks internally)
 - An external SAS direct attached disks storage enclosure full with 15k 
 rpm 36GB disks (for instance the MD1000, with 15x 36GB 15k disks)
 
 Going for the dell-solution would set you back only (including 
 savings) about $13-$14k. HP offers a similar solutions (a HP DL360G5 or 
 a DL380G5/DL385 with two MSA50's for instance) which also fit in your 
 budget afaik. The other players tend to be (a bit) more expensive, force 
 you to go with Fibre Channel or ancient SCSI external storage ;)
 
 If you'd like to have a product by a generic vendor, have a look at the 
 Adaptec JS50 SAS Jbod enclosure or Promise's Vtrak 300 (both offer 12 
 sas/sata bays in 2U) for storage.
 
 If you're limited to only 2U of rack space, its a bit more difficult to 
 get maximum I/O in your budget (you have basically space for about 8 or 
 12 3.5 disks (with generic suppliers) or 16 2.5 sff disks (with HP)).
 But you should still be able to have two top-off-the-line x86 cpu's (amd 
 opteron 285 or intel woorcrest 5160) and 16GB of memory (even FB Dimm, 
 which is pretty expensive).
 
 Best regards,
 
 Arjen van der Meijden
 
 
 On 8-8-2006 22:43, Kenji Morishige wrote:
 I've asked for some help here a few months ago and got some really 
 helpfull
 answers regarding RAID controllers and server configuration.  Up until
 recently I've been running PostgreSQL on a two year old Dual Xeon 3.06Ghz
 machine with a single channel RAID controller (previously Adaptec 2200S, 
 but
 now changed to LSI MegaRAID). The 2U 

Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-18 Thread Bucky Jordan

We've been doing some research in this area (the new Woodcrest from
Intel, the Opterons from Dell, and SAS).

In a nutshell, here's what I'm aware of:

Dell does provide a 15 disk external SAS enclosure- the performance
numbers they claim look pretty good (of course, go figure) and as far as
I can tell, the Perc5/I (the new SAS controller) actually has reasonable
performance. I've been playing around with a 2950 with 6x300 GB 10k RPM
SAS drives, but no enclosure yet. You can also apparently daisy chain up
to 3 enclosures and use multiple perc cards.

Dell originally was planning to only support 4 socket opteron boxes, but
now they have also apparently decided to support 2 socket ones also.
According to this article, they're saying before end of the year.
http://www.eweek.com/article2/0,1895,2004886,00.asp

Some say that the Woodcrest performs just as well, if not better than
the opteron, but I have been unable to do specific tests as of yet. If
anyone has a comparable Opteron box (to a PE2950 2x3.0 8 GB RAM
Woodcrest), I'd be happy to run some benchmarks. 

Lastly, Sun just came out with their new X4600. 48 drives, 24 TB
storage, 4 U rack space. http://www.sun.com/servers/x64/x4500/getit.jsp
That's over your $20k limit though, but looks like it'd be a great DB
box.

For $20k with dell, you could probably get a 2 CPU 2950, with an
external drive cage and 15 SAS drives (just large/med business pricing
on their website). I know I would be very curious about the performance
of this setup if anyone got their hands on it.

We're a Dell shop, so it looks like we'll be settling in on the 2950
Woodcrest for a while, but I have managed to get some people interested
in the Sun box and the 4-way opteron from Dell if the need for more
performance should arise.

HTH,

Bucky



-Original Message-
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Arjen van
der Meijden
Sent: Friday, August 18, 2006 3:42 PM
To: Kenji Morishige
Cc: pgsql-performance@postgresql.org
Subject: Re: [PERFORM] most bang for buck with ~ $20,000

Hi Kenji,

I'm not sure what you mean by 'something newer'? The intel 
woodcrest-cpu's are brand-new compared to the amd opterons. But if you 
need a 4-cpu config (I take it you want 8-cores in that case), Dell 
doesn't offer much. Whether something new will come, I don't know. I'm 
not sure when (or if?) a MP-Woodcrest will arrive and/or when Dell will 
start offering Opteron-servers.

Sas has been designed as the successor to SCSI.

As I see it, SAS has currently one major disadvantage. Lots of new 
servers are equipped with SAS-drives, a few nice SAS-raidcontrollers 
exist, but the availability of external enclosures for SAS is not 
widespread yet. So your options of going beyond (say) 8 disks per system

are a bit limited.

There are of course advantages as well. The bus is much wider (you can 
have 4 lanes of 3Gbps each to an enclosure). You can mix sas and sata 
disks, so you could have two arrays in the same enclosure, one big 
storage bin and a very fast array or just use only sata disks on a sas 
controller. The cabling itself is also much simpler/more flexible 
(although using a hot-plug enclosure of course shields you mostly from 
that).
But whether its the right choice to make now? I'm not sure. We weren't 
to fond of investing a lot of money in an end-of-life system. And since 
we're a tech-website, we also had to worry about our being modern 
image, of course ;)

The main disadvantage I see in this case is, as said, the limited 
availability of external enclosures in comparison to SCSI and Fibre 
Channel. HP currently only offers their MSA50 (for the rather expensive 
SFF disks) while their MSA60 (normal disks) will not be available until 
somewhere in 2007 and Dell also only offers one enclosure, the MD1000.
The other big players offer nothing yet, as far as I know, while they 
normally offer several SCSI and/or FC-enclosures.
There are also some third-party enclosures (adaptec and promise for 
instance) available of course.

Best regards,

Arjen

On 18-8-2006 21:07, Kenji Morishige wrote:
 Thanks Arjen, 
 I have unlimited rack space if I really need it.  Is serial/SAS really
the
 better route to go than SCSI these days? I'm so used to ordering SCSI
that
 I've been out of the loop with new disk enclosures and disk tech.  I
been
 trying to price out a HP DL585, but those are considerably more than
the
 Dells.  Is it worth waiting a few more weeks/months for Dell to
release
 something newer?
 
 -Kenji
 
 On Wed, Aug 09, 2006 at 07:35:22AM +0200, Arjen van der Meijden wrote:
 With such a budget you should easily be able to get something like:
 - A 1U high-performance server (for instance the Dell 1950 with 2x 
 Woodcrest 5160, 16GB of FB-Dimm memory, one 5i and one 5e perc raid 
 controller and some disks internally)
 - An external SAS direct attached disks storage enclosure full with
15k 
 rpm 36GB disks (for instance the MD1000, with 15x 36GB 15k disks)

 Going

Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-18 Thread Arjen van der Meijden
Well, that's of course really hard to tell. From personal experience in 
a read-mostly environment, the subtop woodcrest 5150 (2.6Ghz) 
outperforms the top dempsey 5080 (3.7Ghz, in the same system) by quite a 
nice margin. But that dempsey already has the faster FB-Dimm memory and 
a much wider FSB compared to your 3.06Ghz Xeons.
But if we assume that the 3.7Ghz 5080 is just the extra mhz faster (~ 
25%), for a single (dual core) 3Ghz Woodcrest you might already be 
talking about a 50% improvement in terms of cpu-power over your current 
set-up. Of course depending on workload and its scalability etc etc.


In a perfect world, with linear scalability (note, a read-mostly 
postgresql can actually do that on a Sun fire T2000 with solaris) that 
would yield a 200% improvement when going form 2 to 4 cores. A 70-80% 
scaling is more reasonable and would still imply you'd improve more than 
150% over your current set-up.
Please note that this is partially based on internal testing and partial 
on assumptions and would at least require more real-world testing for a 
app more similar to yours.


As soon as we're publishing some numbers on this (and I don't forget), 
I'll let you know on the list. That will include postgresql and recent 
x86 cpu's on linux and should be ready soon.


Best regards,

Arjen


On 18-8-2006 21:51, Kenji Morishige wrote:

Thanks Arjen for your reply, this is definitely something to consider. I
think in our case, we are not too concerned with the tech image as much as if
the machine will allow us to scale the loads we need. I'm not sure if we
should worry so much about the IO bandwidth as we are not even close to
saturating 320MB/s.  I think stability, reliability, and ease-of-use and
recovery is our main concern at the moment.  I currently am runing a load
average of about .5 on a dual Xeon 3.06Ghz P4 setup.  How much CPU
performance improvement do you think the new woodcrest cpus are over these?

-Kenji

On Fri, Aug 18, 2006 at 09:41:55PM +0200, Arjen van der Meijden wrote:

Hi Kenji,

I'm not sure what you mean by 'something newer'? The intel 
woodcrest-cpu's are brand-new compared to the amd opterons. But if you 
need a 4-cpu config (I take it you want 8-cores in that case), Dell 
doesn't offer much. Whether something new will come, I don't know. I'm 
not sure when (or if?) a MP-Woodcrest will arrive and/or when Dell will 
start offering Opteron-servers.


Sas has been designed as the successor to SCSI.

As I see it, SAS has currently one major disadvantage. Lots of new 
servers are equipped with SAS-drives, a few nice SAS-raidcontrollers 
exist, but the availability of external enclosures for SAS is not 
widespread yet. So your options of going beyond (say) 8 disks per system 
are a bit limited.


There are of course advantages as well. The bus is much wider (you can 
have 4 lanes of 3Gbps each to an enclosure). You can mix sas and sata 
disks, so you could have two arrays in the same enclosure, one big 
storage bin and a very fast array or just use only sata disks on a sas 
controller. The cabling itself is also much simpler/more flexible 
(although using a hot-plug enclosure of course shields you mostly from 
that).
But whether its the right choice to make now? I'm not sure. We weren't 
to fond of investing a lot of money in an end-of-life system. And since 
we're a tech-website, we also had to worry about our being modern 
image, of course ;)


The main disadvantage I see in this case is, as said, the limited 
availability of external enclosures in comparison to SCSI and Fibre 
Channel. HP currently only offers their MSA50 (for the rather expensive 
SFF disks) while their MSA60 (normal disks) will not be available until 
somewhere in 2007 and Dell also only offers one enclosure, the MD1000.
The other big players offer nothing yet, as far as I know, while they 
normally offer several SCSI and/or FC-enclosures.
There are also some third-party enclosures (adaptec and promise for 
instance) available of course.


Best regards,

Arjen

On 18-8-2006 21:07, Kenji Morishige wrote:
Thanks Arjen, 
I have unlimited rack space if I really need it.  Is serial/SAS really the

better route to go than SCSI these days? I'm so used to ordering SCSI that
I've been out of the loop with new disk enclosures and disk tech.  I been
trying to price out a HP DL585, but those are considerably more than the
Dells.  Is it worth waiting a few more weeks/months for Dell to release
something newer?

-Kenji

On Wed, Aug 09, 2006 at 07:35:22AM +0200, Arjen van der Meijden wrote:

With such a budget you should easily be able to get something like:
- A 1U high-performance server (for instance the Dell 1950 with 2x 
Woodcrest 5160, 16GB of FB-Dimm memory, one 5i and one 5e perc raid 
controller and some disks internally)
- An external SAS direct attached disks storage enclosure full with 15k 
rpm 36GB disks (for instance the MD1000, with 15x 36GB 15k disks)


Going for the dell-solution would set you back only 

Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-10 Thread Jeff Trout


On Aug 9, 2006, at 5:35 PM, Jim C. Nasby wrote:

Note that some controllers (such as 3ware) need to periodically  
test the

life of the BBU, and they disable write caching when they do so, which
would tank performance.


Yep. I did the battery capacity test before I went live with our  
9550sx controller.
The only downside I see by not doing it is its estimated battery  
lifetime number may be inaccurate, and once a week you get an alarm  
message about the capacity test being overdue.


It does seem like a big design flaw needing to do it, but if you  
think about it, you don't want to have data in the cache while seeing  
how long it takes for the battery to drain :)


--
Jeff Trout [EMAIL PROTECTED]
http://www.dellsmartexitin.com/
http://www.stuarthamm.net/




---(end of broadcast)---
TIP 3: Have you checked our extensive FAQ?

  http://www.postgresql.org/docs/faq


Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-09 Thread Kenji Morishige
I have unlimited rack space, so 2U is not the issue. The boxes are stored in
our lab for internal software tools.  I'm going to research those boxes you
mention.  Regarding the JBOD enclosures, are these generally just 2U or 4U
units with SCSI interface connectors?  I didn't see these types of boxes
availble on Dell website, I'll look again.
-Kenji

On Wed, Aug 09, 2006 at 07:35:22AM +0200, Arjen van der Meijden wrote:
 With such a budget you should easily be able to get something like:
 - A 1U high-performance server (for instance the Dell 1950 with 2x 
 Woodcrest 5160, 16GB of FB-Dimm memory, one 5i and one 5e perc raid 
 controller and some disks internally)
 - An external SAS direct attached disks storage enclosure full with 15k 
 rpm 36GB disks (for instance the MD1000, with 15x 36GB 15k disks)
 
 Going for the dell-solution would set you back only (including 
 savings) about $13-$14k. HP offers a similar solutions (a HP DL360G5 or 
 a DL380G5/DL385 with two MSA50's for instance) which also fit in your 
 budget afaik. The other players tend to be (a bit) more expensive, force 
 you to go with Fibre Channel or ancient SCSI external storage ;)
 
 If you'd like to have a product by a generic vendor, have a look at the 
 Adaptec JS50 SAS Jbod enclosure or Promise's Vtrak 300 (both offer 12 
 sas/sata bays in 2U) for storage.
 
 If you're limited to only 2U of rack space, its a bit more difficult to 
 get maximum I/O in your budget (you have basically space for about 8 or 
 12 3.5 disks (with generic suppliers) or 16 2.5 sff disks (with HP)).
 But you should still be able to have two top-off-the-line x86 cpu's (amd 
 opteron 285 or intel woorcrest 5160) and 16GB of memory (even FB Dimm, 
 which is pretty expensive).
 
 Best regards,
 
 Arjen van der Meijden
 
 
 On 8-8-2006 22:43, Kenji Morishige wrote:
 I've asked for some help here a few months ago and got some really helpfull
 answers regarding RAID controllers and server configuration.  Up until
 recently I've been running PostgreSQL on a two year old Dual Xeon 3.06Ghz
 machine with a single channel RAID controller (previously Adaptec 2200S, 
 but
 now changed to LSI MegaRAID). The 2U unit is from a generic vendor using 
 what
 I believe is a SuperMicro motherboard.  In the last week after upgrading 
 the
 RAID controller, the machine has had disk failure and some other issues. I
 would like to build a very reliable dedicated postgreSQL server that has 
 the
 ultimate possible performance and reliabily for around $20,000.  The data 
 set
 size is only currently about 4GB, but is increasing by approximately 50MB
 daily.  The server also requires about 500 connections and I have been
 monitoring about 100-200 queries per second at the moment.  I am planning 
 to
 run FreeBSD 6.1 if possible, but I am open to any other suggestions if it
 improves performance.
 
 I am considering a setup such as this:
   - At least dual cpu (possibly with 2 cores each)
   - 4GB of RAM
   - 2 disk RAID 1 array for root disk
   - 4 disk RAID 1+0 array for PGDATA
   - 2 disk RAID 1 array for pg_xlog
 
 Does anyone know a vendor that might be able provide such setup?  Any
 critique in this design? I'm thinking having a 2 channel RAID controller to
 seperate the PGDATA, root and pg_xlog.
 
 Sincerely,
 Kenji
 
 ---(end of broadcast)---
 TIP 2: Don't 'kill -9' the postmaster
 

---(end of broadcast)---
TIP 6: explain analyze is your friend


Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-09 Thread Arjen van der Meijden
We were in a similar situation with a similar budget. But we had two 
requirements, no deprecated scsi while the successor SAS is available 
and preferrably only 3 or 4U of rack space. And it had to have 
reasonable amounts of disks (at least 12).


The two options we finally choose between where a Dell 1U 1950 with two 
woodcrests 5160 (I don't think the older dempsey 50x0's are a good idea) 
and 16GB of memory combined with a PowerVault MD1000 external storage 
SAS JBOD unit, with 15 36GB 15k rpm disks and from HP a similar 
configured DL360G5 (also 1U) combined with two MSA50 SFF SAS JBOD 
enclosures with 20 36GB 10k rpm SFF disks.


Both enclosures offer has SAS-connectivity (serial attached scsi), i.e. 
the next generation scsi. Which is supposed to be the successor to 
scsi, but unfortunately its not yet as widely available.


The Dell MD1000 is 3U high and can be fitted with 15 3.5 disks, the 
MSA50 is 1U and can be fitted with 10 2.5 disks.
In terms of performance you'll likely need two MSA50's to be up to par 
with one MD1000. The SFF disks are about as expensive as the 15k 3.5 
disks... so its mostly interesting for packing a lot of I/O in a small 
enclosure. HP is going to offer a 3.5 SAS-enclosure (MSA60) but that 
one won't be available until Q1 2007 or something like that.
As said Promise and Adaptec also offer SAS enclosures, both are 2U and 
can be fitted with 12 disks. There are more available, but they are 
generally quite bit hard to find.


Good luck with your search.

Best regards,

Arjen


Kenji Morishige wrote:

I have unlimited rack space, so 2U is not the issue. The boxes are stored in
our lab for internal software tools.  I'm going to research those boxes you
mention.  Regarding the JBOD enclosures, are these generally just 2U or 4U
units with SCSI interface connectors?  I didn't see these types of boxes
availble on Dell website, I'll look again.
-Kenji

On Wed, Aug 09, 2006 at 07:35:22AM +0200, Arjen van der Meijden wrote:

With such a budget you should easily be able to get something like:
- A 1U high-performance server (for instance the Dell 1950 with 2x 
Woodcrest 5160, 16GB of FB-Dimm memory, one 5i and one 5e perc raid 
controller and some disks internally)
- An external SAS direct attached disks storage enclosure full with 15k 
rpm 36GB disks (for instance the MD1000, with 15x 36GB 15k disks)


Going for the dell-solution would set you back only (including 
savings) about $13-$14k. HP offers a similar solutions (a HP DL360G5 or 
a DL380G5/DL385 with two MSA50's for instance) which also fit in your 
budget afaik. The other players tend to be (a bit) more expensive, force 
you to go with Fibre Channel or ancient SCSI external storage ;)


If you'd like to have a product by a generic vendor, have a look at the 
Adaptec JS50 SAS Jbod enclosure or Promise's Vtrak 300 (both offer 12 
sas/sata bays in 2U) for storage.


If you're limited to only 2U of rack space, its a bit more difficult to 
get maximum I/O in your budget (you have basically space for about 8 or 
12 3.5 disks (with generic suppliers) or 16 2.5 sff disks (with HP)).
But you should still be able to have two top-off-the-line x86 cpu's (amd 
opteron 285 or intel woorcrest 5160) and 16GB of memory (even FB Dimm, 
which is pretty expensive).


Best regards,

Arjen van der Meijden


On 8-8-2006 22:43, Kenji Morishige wrote:

I've asked for some help here a few months ago and got some really helpfull
answers regarding RAID controllers and server configuration.  Up until
recently I've been running PostgreSQL on a two year old Dual Xeon 3.06Ghz
machine with a single channel RAID controller (previously Adaptec 2200S, 
but
now changed to LSI MegaRAID). The 2U unit is from a generic vendor using 
what
I believe is a SuperMicro motherboard.  In the last week after upgrading 
the

RAID controller, the machine has had disk failure and some other issues. I
would like to build a very reliable dedicated postgreSQL server that has 
the
ultimate possible performance and reliabily for around $20,000.  The data 
set

size is only currently about 4GB, but is increasing by approximately 50MB
daily.  The server also requires about 500 connections and I have been
monitoring about 100-200 queries per second at the moment.  I am planning 
to

run FreeBSD 6.1 if possible, but I am open to any other suggestions if it
improves performance.

I am considering a setup such as this:
 - At least dual cpu (possibly with 2 cores each)
 - 4GB of RAM
 - 2 disk RAID 1 array for root disk
 - 4 disk RAID 1+0 array for PGDATA
 - 2 disk RAID 1 array for pg_xlog

Does anyone know a vendor that might be able provide such setup?  Any
critique in this design? I'm thinking having a 2 channel RAID controller to
seperate the PGDATA, root and pg_xlog.

Sincerely,
Kenji

---(end of broadcast)---
TIP 2: Don't 'kill -9' the postmaster





---(end of broadcast)---
TIP 5: 

Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-09 Thread Scott Marlowe
On Tue, 2006-08-08 at 17:53, Thomas F. O'Connell wrote:
 On Aug 8, 2006, at 5:28 PM, Joshua D. Drake wrote:
 
  Thomas F. O'Connell wrote:
  On Aug 8, 2006, at 4:49 PM, Joshua D. Drake wrote:
  I am considering a setup such as this:
- At least dual cpu (possibly with 2 cores each)
- 4GB of RAM
- 2 disk RAID 1 array for root disk
- 4 disk RAID 1+0 array for PGDATA
- 2 disk RAID 1 array for pg_xlog
  Does anyone know a vendor that might be able provide such setup?
  Wouldn't it be preferable to put WAL on a multi-disk RAID 10 if  
  you had the opportunity? This gives you the redundancy of RAID 1  
  but approaches the performance of RAID 0, especially as you add  
  disks to the array. In benchmarking, I've seen consistent success  
  with this approach.
 
  WALL is written in order so RAID 1 is usually fine. We also don't  
  need journaling for WAL so the speed is even faster.
 
 In which case, which is theoretically better (since I don't have a  
 convenient test bed at the moment) for WAL in a write-heavy  
 environment? More disks in a RAID 10 (which should theoretically  
 improve write throughput in general, to a point) or a 2-disk RAID 1?  
 Does it become a price/performance question, or is there virtually no  
 benefit to throwing more disks at RAID 10 for WAL if you turn off  
 journaling on the filesystem?

Actually, the BIGGEST win comes when you've got battery backed cache on
your RAID controller.  In fact, I'd spend money on a separate RAID
controller for xlog with its own cache hitting a simple mirror set
before I'd spring for more drives on pg_xlog.  The battery backed cache
on the pg_xlog likely wouldn't need to be big, just there and set to
write-back.

Then put all the rest of your cash into disks on a big RAID 10 config,
and as big of a battery backed cache as you can afford for it and memory
for the machine.

---(end of broadcast)---
TIP 6: explain analyze is your friend


Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-09 Thread Merlin Moncure

On 8/9/06, Kenji Morishige [EMAIL PROTECTED] wrote:

I have unlimited rack space, so 2U is not the issue. The boxes are stored in
our lab for internal software tools.  I'm going to research those boxes you
mention.  Regarding the JBOD enclosures, are these generally just 2U or 4U
units with SCSI interface connectors?  I didn't see these types of boxes
availble on Dell website, I'll look again.
-Kenji

On Wed, Aug 09, 2006 at 07:35:22AM +0200, Arjen van der Meijden wrote:
 With such a budget you should easily be able to get something like:
 - A 1U high-performance server (for instance the Dell 1950 with 2x
 Woodcrest 5160, 16GB of FB-Dimm memory, one 5i and one 5e perc raid
 controller and some disks internally)
 - An external SAS direct attached disks storage enclosure full with 15k
 rpm 36GB disks (for instance the MD1000, with 15x 36GB 15k disks)

 Going for the dell-solution would set you back only (including
 savings) about $13-$14k. HP offers a similar solutions (a HP DL360G5 or
 a DL380G5/DL385 with two MSA50's for instance) which also fit in your
 budget afaik. The other players tend to be (a bit) more expensive, force
 you to go with Fibre Channel or ancient SCSI external storage ;)

 If you'd like to have a product by a generic vendor, have a look at the
 Adaptec JS50 SAS Jbod enclosure or Promise's Vtrak 300 (both offer 12
 sas/sata bays in 2U) for storage.


I am really curious about the Adaptec SAS product to see what it can
do.  If you don't know what SAS is, it is Sata Attached SCSI.  SAS
cables use 4 sata lanes (3gb/sec each) bonded together in a single
cable.  The raid is handled with via the raid controller or the o/s in
a software configuration.

SAS is the evolution of SCSI and I think will ultimately replace scsi
in enterprise setups becuase it is faster, cheaper, and more flexible.
SAS enclosures generally accept sata or sas drives in mix/match
configurations.  so, you get to choose between cheap, large, 7200 rpm
sata drives or small, expensive sas 10k or 15k rpm drives  *in the
same enclosure*.  You also get a compromise drive in the form of the
raptor which is 10k rpm sata drive.

You could buy a 2u 12 drive SAS encloure (3000$), 12 150g raptors
(3000$) and spend another grand on cables/controller and have a
hellishly performing raid system for the money assuming sas performs
like it does on paper.  note that i would not be trying this with my
own money unless I was guaranteed a money back rma for a 30 day
evaluation period.

that would leave you with 12 grand or so to pick up a quad (8 core)
opeteron if you bought it right.

regards,
merlin

---(end of broadcast)---
TIP 4: Have you searched our list archives?

  http://archives.postgresql.org


Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-09 Thread Scott Marlowe
On Wed, 2006-08-09 at 16:35, Jim C. Nasby wrote:
 On Wed, Aug 09, 2006 at 10:15:27AM -0500, Scott Marlowe wrote:
  Actually, the BIGGEST win comes when you've got battery backed cache on
  your RAID controller.  In fact, I'd spend money on a separate RAID
  controller for xlog with its own cache hitting a simple mirror set
  before I'd spring for more drives on pg_xlog.  The battery backed cache
  on the pg_xlog likely wouldn't need to be big, just there and set to
  write-back.
  
  Then put all the rest of your cash into disks on a big RAID 10 config,
  and as big of a battery backed cache as you can afford for it and memory
  for the machine.
 
 Actually, my (limited) testing has show than on a good battery-backed
 controller, there's no penalty to leaving pg_xlog in with the rest of
 PGDATA. This means that the OP could pile all 8 drives into a RAID10,
 which would almost certainly do better than 6+2.

I've seen a few posts that said that before.  I wonder if there's a
point where the single RAID array / controller would get saturated and a
second one would help.  I think most of the testing I've seen so far has
been multiple RAID arrays under the same controller, hasn't it?

 Note that some controllers (such as 3ware) need to periodically test the
 life of the BBU, and they disable write caching when they do so, which
 would tank performance.

ugh, that's a scary thing.  Can you at least schedule it?

---(end of broadcast)---
TIP 1: if posting/reading through Usenet, please send an appropriate
   subscribe-nomail command to [EMAIL PROTECTED] so that your
   message can get through to the mailing list cleanly


Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-09 Thread Jim C. Nasby
On Wed, Aug 09, 2006 at 04:50:30PM -0500, Scott Marlowe wrote:
 On Wed, 2006-08-09 at 16:35, Jim C. Nasby wrote:
  On Wed, Aug 09, 2006 at 10:15:27AM -0500, Scott Marlowe wrote:
   Actually, the BIGGEST win comes when you've got battery backed cache on
   your RAID controller.  In fact, I'd spend money on a separate RAID
   controller for xlog with its own cache hitting a simple mirror set
   before I'd spring for more drives on pg_xlog.  The battery backed cache
   on the pg_xlog likely wouldn't need to be big, just there and set to
   write-back.
   
   Then put all the rest of your cash into disks on a big RAID 10 config,
   and as big of a battery backed cache as you can afford for it and memory
   for the machine.
  
  Actually, my (limited) testing has show than on a good battery-backed
  controller, there's no penalty to leaving pg_xlog in with the rest of
  PGDATA. This means that the OP could pile all 8 drives into a RAID10,
  which would almost certainly do better than 6+2.
 
 I've seen a few posts that said that before.  I wonder if there's a
 point where the single RAID array / controller would get saturated and a
 second one would help.  I think most of the testing I've seen so far has
 been multiple RAID arrays under the same controller, hasn't it?
 
Yeah. I've had one client try it so far, but it was a pretty small array
(8 drives, IIRC).

I suspect that by the time you get to a size where you're saturating a
controller, you're looking at enough drives where having two extra
(instead of dedicating them to pg_xlog) won't make much difference.

  Note that some controllers (such as 3ware) need to periodically test the
  life of the BBU, and they disable write caching when they do so, which
  would tank performance.
 
 ugh, that's a scary thing.  Can you at least schedule it?

Yeah, it's not automatic at all. Which itself is somewhat scarry
-- 
Jim C. Nasby, Sr. Engineering Consultant  [EMAIL PROTECTED]
Pervasive Software  http://pervasive.comwork: 512-231-6117
vcard: http://jim.nasby.net/pervasive.vcf   cell: 512-569-9461

---(end of broadcast)---
TIP 3: Have you checked our extensive FAQ?

   http://www.postgresql.org/docs/faq


Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-08 Thread Joshua D. Drake


I am considering a setup such as this:
  - At least dual cpu (possibly with 2 cores each)
  - 4GB of RAM
  - 2 disk RAID 1 array for root disk
  - 4 disk RAID 1+0 array for PGDATA
  - 2 disk RAID 1 array for pg_xlog

Does anyone know a vendor that might be able provide such setup?


I would look at the HP DL 385 or 585. The 385 is going to max a (2) dual 
core cpus. The 585 is (4) dual core cpus.


For 20k you should be able to go with a bit more then what you have 
above, specifically in the RAM department.


Joshua D. Drake




  Any

critique in this design? I'm thinking having a 2 channel RAID controller to
seperate the PGDATA, root and pg_xlog.

Sincerely,
Kenji

---(end of broadcast)---
TIP 2: Don't 'kill -9' the postmaster




--

   === The PostgreSQL Company: Command Prompt, Inc. ===
Sales/Support: +1.503.667.4564 || 24x7/Emergency: +1.800.492.2240
   Providing the most comprehensive  PostgreSQL solutions since 1997
 http://www.commandprompt.com/



---(end of broadcast)---
TIP 1: if posting/reading through Usenet, please send an appropriate
  subscribe-nomail command to [EMAIL PROTECTED] so that your
  message can get through to the mailing list cleanly


Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-08 Thread Thomas F. O'Connell


On Aug 8, 2006, at 4:49 PM, Joshua D. Drake wrote:


I am considering a setup such as this:
  - At least dual cpu (possibly with 2 cores each)
  - 4GB of RAM
  - 2 disk RAID 1 array for root disk
  - 4 disk RAID 1+0 array for PGDATA
  - 2 disk RAID 1 array for pg_xlog
Does anyone know a vendor that might be able provide such setup?


Wouldn't it be preferable to put WAL on a multi-disk RAID 10 if you  
had the opportunity? This gives you the redundancy of RAID 1 but  
approaches the performance of RAID 0, especially as you add disks to  
the array. In benchmarking, I've seen consistent success with this  
approach.


--
Thomas F. O'Connell
Sitening, LLC

http://www.sitening.com/
3004B Poston Avenue
Nashville, TN 37203-1314
615-469-5150 x802
615-469-5151 (fax)

---(end of broadcast)---
TIP 4: Have you searched our list archives?

  http://archives.postgresql.org


Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-08 Thread Kenji Morishige
The 1+0 on the WAL is better than on PGDATA? I guess I'm confused about the
write sequence of the data. I will research more, thank you!
-Kenji

On Tue, Aug 08, 2006 at 04:59:09PM -0500, Thomas F. O'Connell wrote:
 
 On Aug 8, 2006, at 4:49 PM, Joshua D. Drake wrote:
 
 I am considering a setup such as this:
   - At least dual cpu (possibly with 2 cores each)
   - 4GB of RAM
   - 2 disk RAID 1 array for root disk
   - 4 disk RAID 1+0 array for PGDATA
   - 2 disk RAID 1 array for pg_xlog
 Does anyone know a vendor that might be able provide such setup?
 
 Wouldn't it be preferable to put WAL on a multi-disk RAID 10 if you  
 had the opportunity? This gives you the redundancy of RAID 1 but  
 approaches the performance of RAID 0, especially as you add disks to  
 the array. In benchmarking, I've seen consistent success with this  
 approach.
 
 --
 Thomas F. O'Connell
 Sitening, LLC
 
 http://www.sitening.com/
 3004B Poston Avenue
 Nashville, TN 37203-1314
 615-469-5150 x802
 615-469-5151 (fax)

---(end of broadcast)---
TIP 1: if posting/reading through Usenet, please send an appropriate
   subscribe-nomail command to [EMAIL PROTECTED] so that your
   message can get through to the mailing list cleanly


Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-08 Thread Scott Marlowe
On Tue, 2006-08-08 at 15:43, Kenji Morishige wrote:
 I've asked for some help here a few months ago and got some really helpfull
 answers regarding RAID controllers and server configuration.  Up until
 recently I've been running PostgreSQL on a two year old Dual Xeon 3.06Ghz
 machine with a single channel RAID controller (previously Adaptec 2200S, but
 now changed to LSI MegaRAID). The 2U unit is from a generic vendor using what
 I believe is a SuperMicro motherboard.  In the last week after upgrading the
 RAID controller, the machine has had disk failure and some other issues. I
 would like to build a very reliable dedicated postgreSQL server that has the
 ultimate possible performance and reliabily for around $20,000.  The data set
 size is only currently about 4GB, but is increasing by approximately 50MB
 daily.  The server also requires about 500 connections and I have been
 monitoring about 100-200 queries per second at the moment.  I am planning to
 run FreeBSD 6.1 if possible, but I am open to any other suggestions if it
 improves performance.

This really depends on your usage patterns.

OLAP or OLTP workloads?  Do you need 24/7 reliability and therefore a
two machine setup?  There's a lot of variety in load.

Generally, you spend your money on disks, then memory, then CPU, in that
order.

Look at the Areca cards, they've come highly recommended here.  Look at
LOTS of drives.  Given the size of your db, you can go with LOTS of
smaller drives and get good performance.  If you can find a good box to
hold 12 to 16 drives and fill it with 37 gig 15k RPM drives, you'll have
lots of storage, even in RAID 1+0 config.  That's aiming at
transactional throughput.

Toss as much memory as is reasonably affordable at it.  That's normally
in the 4 to 8 gig range.  After that things start to get expensive fast.

Multiple - dual core CPUs are a good idea.  Opterons seem to be better
data pumps with large memory and 2 CPUs than Intels right now. 
Better to have a 2xdual core opteron with slower processors than a
single dual core or dual single core CPU(s) with a faster clock speed. 
As long as the memory access is equivalent, the more CPUs the better in
Opterons, where their interconnect speed increases as you increase the
number of CPUs.  Intel Xeons are the opposite.  Better with fewer faster
CPUs / cores.

I just ran through a configurator on a site selling quad dual core
opteron servers.  8 Seagate cheetah 15k rpm drives, 8 gig ram, and the
slowest (1.8 GHz) AMD dual core CPUs (4 of them) for 8 cores, came out
to $13,500 or so.

I'd take the other $7.5 grand and buy a backup server that can old as
much but isn't quite as beefy and set up slony to have a live hot spare
sitting ready.  Oh, and maybe to buy some spare parts to sit in the desk
drawer in case things break.

---(end of broadcast)---
TIP 1: if posting/reading through Usenet, please send an appropriate
   subscribe-nomail command to [EMAIL PROTECTED] so that your
   message can get through to the mailing list cleanly


Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-08 Thread Kenji Morishige
Great info, which vendor were you looking at for these Opterons? I am goign
to be purchasing 2 of these. :) I do need 24/7 reliability.

On Tue, Aug 08, 2006 at 05:08:29PM -0500, Scott Marlowe wrote:
 On Tue, 2006-08-08 at 15:43, Kenji Morishige wrote:
  I've asked for some help here a few months ago and got some really helpfull
  answers regarding RAID controllers and server configuration.  Up until
  recently I've been running PostgreSQL on a two year old Dual Xeon 3.06Ghz
  machine with a single channel RAID controller (previously Adaptec 2200S, but
  now changed to LSI MegaRAID). The 2U unit is from a generic vendor using 
  what
  I believe is a SuperMicro motherboard.  In the last week after upgrading the
  RAID controller, the machine has had disk failure and some other issues. I
  would like to build a very reliable dedicated postgreSQL server that has the
  ultimate possible performance and reliabily for around $20,000.  The data 
  set
  size is only currently about 4GB, but is increasing by approximately 50MB
  daily.  The server also requires about 500 connections and I have been
  monitoring about 100-200 queries per second at the moment.  I am planning to
  run FreeBSD 6.1 if possible, but I am open to any other suggestions if it
  improves performance.
 
 This really depends on your usage patterns.
 
 OLAP or OLTP workloads?  Do you need 24/7 reliability and therefore a
 two machine setup?  There's a lot of variety in load.
 
 Generally, you spend your money on disks, then memory, then CPU, in that
 order.
 
 Look at the Areca cards, they've come highly recommended here.  Look at
 LOTS of drives.  Given the size of your db, you can go with LOTS of
 smaller drives and get good performance.  If you can find a good box to
 hold 12 to 16 drives and fill it with 37 gig 15k RPM drives, you'll have
 lots of storage, even in RAID 1+0 config.  That's aiming at
 transactional throughput.
 
 Toss as much memory as is reasonably affordable at it.  That's normally
 in the 4 to 8 gig range.  After that things start to get expensive fast.
 
 Multiple - dual core CPUs are a good idea.  Opterons seem to be better
 data pumps with large memory and 2 CPUs than Intels right now. 
 Better to have a 2xdual core opteron with slower processors than a
 single dual core or dual single core CPU(s) with a faster clock speed. 
 As long as the memory access is equivalent, the more CPUs the better in
 Opterons, where their interconnect speed increases as you increase the
 number of CPUs.  Intel Xeons are the opposite.  Better with fewer faster
 CPUs / cores.
 
 I just ran through a configurator on a site selling quad dual core
 opteron servers.  8 Seagate cheetah 15k rpm drives, 8 gig ram, and the
 slowest (1.8 GHz) AMD dual core CPUs (4 of them) for 8 cores, came out
 to $13,500 or so.
 
 I'd take the other $7.5 grand and buy a backup server that can old as
 much but isn't quite as beefy and set up slony to have a live hot spare
 sitting ready.  Oh, and maybe to buy some spare parts to sit in the desk
 drawer in case things break.

---(end of broadcast)---
TIP 1: if posting/reading through Usenet, please send an appropriate
   subscribe-nomail command to [EMAIL PROTECTED] so that your
   message can get through to the mailing list cleanly


Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-08 Thread Steve Atkins


On Aug 8, 2006, at 1:43 PM, Kenji Morishige wrote:




I am considering a setup such as this:
  - At least dual cpu (possibly with 2 cores each)
  - 4GB of RAM
  - 2 disk RAID 1 array for root disk
  - 4 disk RAID 1+0 array for PGDATA
  - 2 disk RAID 1 array for pg_xlog

Does anyone know a vendor that might be able provide such setup?  Any
critique in this design? I'm thinking having a 2 channel RAID  
controller to

seperate the PGDATA, root and pg_xlog.


It's fairly similar to the system I'm using - I have 6 spindles for
PGDATA rather than 4. It's a 9550SX SATA based box with
26 bays (16 empty right now) dual Opterons and 4 gigs of
RAM. About $7k from asacomputers.com. Performance is
good, drastically better than the SCSI Dell server it replaced,
but I've not benchmarked it.

You mention 500 simultaneous connections. If that's really
500 simultaneous database connections (rather than, say,
500 simultaneous web users feeding into a pool of fewer
database connections) then that's a lot. I might go with a
bit more RAM than 4GB.

You should have a lot left over from
$20k, and spending some of it on more RAM would be a good
investment (as might a better RAID controller than the 9550SX,
some dual core CPUs and maybe some more PGDATA spindles).

Cheers,
  Steve


---(end of broadcast)---
TIP 6: explain analyze is your friend


Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-08 Thread Joshua D. Drake

Thomas F. O'Connell wrote:


On Aug 8, 2006, at 4:49 PM, Joshua D. Drake wrote:


I am considering a setup such as this:
  - At least dual cpu (possibly with 2 cores each)
  - 4GB of RAM
  - 2 disk RAID 1 array for root disk
  - 4 disk RAID 1+0 array for PGDATA
  - 2 disk RAID 1 array for pg_xlog
Does anyone know a vendor that might be able provide such setup?


Wouldn't it be preferable to put WAL on a multi-disk RAID 10 if you had 
the opportunity? This gives you the redundancy of RAID 1 but approaches 
the performance of RAID 0, especially as you add disks to the array. In 
benchmarking, I've seen consistent success with this approach.


WALL is written in order so RAID 1 is usually fine. We also don't need 
journaling for WAL so the speed is even faster.


Joshua D. Drake



--
Thomas F. O'Connell
Sitening, LLC

http://www.sitening.com/
3004B Poston Avenue
Nashville, TN 37203-1314
615-469-5150 x802
615-469-5151 (fax)

---(end of broadcast)---
TIP 4: Have you searched our list archives?

  http://archives.postgresql.org




--

   === The PostgreSQL Company: Command Prompt, Inc. ===
Sales/Support: +1.503.667.4564 || 24x7/Emergency: +1.800.492.2240
   Providing the most comprehensive  PostgreSQL solutions since 1997
 http://www.commandprompt.com/



---(end of broadcast)---
TIP 1: if posting/reading through Usenet, please send an appropriate
  subscribe-nomail command to [EMAIL PROTECTED] so that your
  message can get through to the mailing list cleanly


Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-08 Thread Thomas F. O'Connell


On Aug 8, 2006, at 5:28 PM, Joshua D. Drake wrote:


Thomas F. O'Connell wrote:

On Aug 8, 2006, at 4:49 PM, Joshua D. Drake wrote:

I am considering a setup such as this:
  - At least dual cpu (possibly with 2 cores each)
  - 4GB of RAM
  - 2 disk RAID 1 array for root disk
  - 4 disk RAID 1+0 array for PGDATA
  - 2 disk RAID 1 array for pg_xlog
Does anyone know a vendor that might be able provide such setup?
Wouldn't it be preferable to put WAL on a multi-disk RAID 10 if  
you had the opportunity? This gives you the redundancy of RAID 1  
but approaches the performance of RAID 0, especially as you add  
disks to the array. In benchmarking, I've seen consistent success  
with this approach.


WALL is written in order so RAID 1 is usually fine. We also don't  
need journaling for WAL so the speed is even faster.


In which case, which is theoretically better (since I don't have a  
convenient test bed at the moment) for WAL in a write-heavy  
environment? More disks in a RAID 10 (which should theoretically  
improve write throughput in general, to a point) or a 2-disk RAID 1?  
Does it become a price/performance question, or is there virtually no  
benefit to throwing more disks at RAID 10 for WAL if you turn off  
journaling on the filesystem?


--
Thomas F. O'Connell
Sitening, LLC

http://www.sitening.com/
3004B Poston Avenue
Nashville, TN 37203-1314
615-469-5150 x802
615-469-5151 (fax)

---(end of broadcast)---
TIP 3: Have you checked our extensive FAQ?

  http://www.postgresql.org/docs/faq


Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-08 Thread Joshua D. Drake


In which case, which is theoretically better (since I don't have a 
convenient test bed at the moment) for WAL in a write-heavy environment? 
More disks in a RAID 10 (which should theoretically improve write 
throughput in general, to a point) or a 2-disk RAID 1? Does it become a 
price/performance question, or is there virtually no benefit to throwing 
more disks at RAID 10 for WAL if you turn off journaling on the filesystem?


Over 4 drives, I would gather that RAID 10 wouldn't gain you anything. 
Possibly over 6 or 8 however, it may be faster because you are writing 
smaller chunks of data, even if two copies of each.


Sincerely,

Joshua D. Drake


--

   === The PostgreSQL Company: Command Prompt, Inc. ===
Sales/Support: +1.503.667.4564 || 24x7/Emergency: +1.800.492.2240
   Providing the most comprehensive  PostgreSQL solutions since 1997
 http://www.commandprompt.com/



---(end of broadcast)---
TIP 2: Don't 'kill -9' the postmaster


Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-08 Thread Thomas F. O'Connell


On Aug 8, 2006, at 6:24 PM, Joshua D. Drake wrote:

In which case, which is theoretically better (since I don't have a  
convenient test bed at the moment) for WAL in a write-heavy  
environment? More disks in a RAID 10 (which should theoretically  
improve write throughput in general, to a point) or a 2-disk RAID  
1? Does it become a price/performance question, or is there  
virtually no benefit to throwing more disks at RAID 10 for WAL if  
you turn off journaling on the filesystem?


Over 4 drives, I would gather that RAID 10 wouldn't gain you  
anything. Possibly over 6 or 8 however, it may be faster because  
you are writing smaller chunks of data, even if two copies of each.


Yeah, where I've seen the benefits in practice, the scenarios have  
involved the availability of a minimum of 6 drives for a RAID 10 for  
WAL. I really should do a comparison of a 2-disk RAID 1 with a  
variety of multi-disk RAID 10 configurations at some point.


--
Thomas F. O'Connell
Sitening, LLC

http://www.sitening.com/
3004B Poston Avenue
Nashville, TN 37203-1314
615-469-5150 x802
615-469-5151 (fax)

---(end of broadcast)---
TIP 4: Have you searched our list archives?

  http://archives.postgresql.org


Re: [PERFORM] most bang for buck with ~ $20,000

2006-08-08 Thread Arjen van der Meijden

With such a budget you should easily be able to get something like:
- A 1U high-performance server (for instance the Dell 1950 with 2x 
Woodcrest 5160, 16GB of FB-Dimm memory, one 5i and one 5e perc raid 
controller and some disks internally)
- An external SAS direct attached disks storage enclosure full with 15k 
rpm 36GB disks (for instance the MD1000, with 15x 36GB 15k disks)


Going for the dell-solution would set you back only (including 
savings) about $13-$14k. HP offers a similar solutions (a HP DL360G5 or 
a DL380G5/DL385 with two MSA50's for instance) which also fit in your 
budget afaik. The other players tend to be (a bit) more expensive, force 
you to go with Fibre Channel or ancient SCSI external storage ;)


If you'd like to have a product by a generic vendor, have a look at the 
Adaptec JS50 SAS Jbod enclosure or Promise's Vtrak 300 (both offer 12 
sas/sata bays in 2U) for storage.


If you're limited to only 2U of rack space, its a bit more difficult to 
get maximum I/O in your budget (you have basically space for about 8 or 
12 3.5 disks (with generic suppliers) or 16 2.5 sff disks (with HP)).
But you should still be able to have two top-off-the-line x86 cpu's (amd 
opteron 285 or intel woorcrest 5160) and 16GB of memory (even FB Dimm, 
which is pretty expensive).


Best regards,

Arjen van der Meijden


On 8-8-2006 22:43, Kenji Morishige wrote:

I've asked for some help here a few months ago and got some really helpfull
answers regarding RAID controllers and server configuration.  Up until
recently I've been running PostgreSQL on a two year old Dual Xeon 3.06Ghz
machine with a single channel RAID controller (previously Adaptec 2200S, but
now changed to LSI MegaRAID). The 2U unit is from a generic vendor using what
I believe is a SuperMicro motherboard.  In the last week after upgrading the
RAID controller, the machine has had disk failure and some other issues. I
would like to build a very reliable dedicated postgreSQL server that has the
ultimate possible performance and reliabily for around $20,000.  The data set
size is only currently about 4GB, but is increasing by approximately 50MB
daily.  The server also requires about 500 connections and I have been
monitoring about 100-200 queries per second at the moment.  I am planning to
run FreeBSD 6.1 if possible, but I am open to any other suggestions if it
improves performance.

I am considering a setup such as this:
  - At least dual cpu (possibly with 2 cores each)
  - 4GB of RAM
  - 2 disk RAID 1 array for root disk
  - 4 disk RAID 1+0 array for PGDATA
  - 2 disk RAID 1 array for pg_xlog

Does anyone know a vendor that might be able provide such setup?  Any
critique in this design? I'm thinking having a 2 channel RAID controller to
seperate the PGDATA, root and pg_xlog.

Sincerely,
Kenji

---(end of broadcast)---
TIP 2: Don't 'kill -9' the postmaster



---(end of broadcast)---
TIP 6: explain analyze is your friend