----- Original Message -----
From: "Claus Guttesen" <[EMAIL PROTECTED]>
To: "Vivek Khera" <[EMAIL PROTECTED]>
Cc: "FreeBSD Stable" <[email protected]>
Sent: Friday, August 17, 2007 11:10 PM
Subject: Re: large RAID volume partition strategy
I have a shiny new big RAID array. 16x500GB SATA 300+NCQ drives
connected to the host via 4Gb fibre channel. This gives me 6.5Tb of
raw disk.
I've come up with three possibilities on organizing this disk. My
needs are really for a single 1Tb file system on which I will run
postgres. However, in the future I'm not sure what I'll really need.
I don't plan to ever connect any other servers to this RAID unit.
The three choices I've come with so far are:
1) Make one RAID volume of 6.5Tb (in a RAID6 + hot spare
configuration), and make one FreeBSD file system on the whole partition.
2) Make one RAID volume of 6.5Tb (in a RAID6 + hot spare
configuration), and make 6 FreeBSD partitions with one file system each.
3) Make 6 RAID volumes and expose them to FreeBSD as multiple drives,
then make one partition + file system on each "disk". Each RAID
volume would span across all 16 drives, and I could make the volumes
of differing RAID levels, if needed, but I'd probably stick with RAID6
+spare.
I'm not keen on option 1 because of the potentially long fsck times
after a crash.
If you want to avoid the long fsck-times your remaining options are a
journaling filesystem or zfs, either requires an upgrade from freebsd
6.2. I have used zfs and had a serverstop due to powerutage in out
area. Our zfs-samba-server came up fine with no data corruption. So I
will suggest freebsd 7.0 with zfs.
Short fsck-times and ufs2 don't do well together. I know there is
background-fsck but for me that is not an option.
--
regards
Claus
When lenity and cruelty play for a kingdom,
the gentlest gamester is the soonest winner.
Shakespeare
If you goal is speed and obviously as little possibility of a fail
(RAID6+spare) then RAID6 is the wrong way to go...
RAID6's read speeds are great but the write speeds are not.
If you want awesome performance and reliability the real way to go is RAID10
(or more correctly RAID 0+1).
You will of course lose a lot more space than you will with RAID6 but the
write speeds will be astronomically higher.
How would you feel with 16 drives in RAID10 with 2 hot spares? This will
give you 3.5TB and if you're using a good RAID controller you should be
getting write speeds of around 400MB/s to the array.
I've got an Areca 1120 RAID controller with 4 320G drives in a stripe set
and I'm writing at 280MB/s to that. With 7 500G drives you should be getting
around 400MB/s because hte RAID10 doesn't have to calculate reconstrust
data. The theoretical max you're ever going to get from the array is 500MB/s
anyways with a 4Gb fibre channel controller.
What it really boils down to is how much space are you willing to sacrifice
for performance...
Another thing you really have to do is make sure you have a good backup
system. I've seen more than one customer crying because their RAID system
with hot spares went on the blink and they lost their data.
-Clay
_______________________________________________
[email protected] mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "[EMAIL PROTECTED]"