On 06/01/2012 12:53 AM, Brian Candler wrote:
On Thu, May 31, 2012 at 11:30:28AM -0700, Mike Seda wrote:
Hi All,
Does the following GlusterFS config seem solid for a small (24
compute-node) HPC cluster with a mixed I/O pattern?
- 4 storage-nodes (distribute-only) with 3 bricks
- Each storage-node will be HP DL360 G8 (32 GB RAM, 16 core)
- Each storage-node will have 1 HP D2600 shelf with 12 x 2 TB drives
(1 RAID 10 given to OS as 1 LVM PV)
Comments and things to consider:
* I'd say CPU and RAM are massive overkill for a storage node.
Right. I figured. I got my original numbers from here though:
http://download.gluster.com/pub/gluster/RHSSA/3.2/Documentation/UG/html/sect-User_Guide-gssa_prepare-chec_min_req.html
IMO 4 core
and 8GB would be fine, unless your working data set is so small it would
fit into 32GB of cache.
* RAID 10 for the OS? Do you mean dedicating *four* disks for the OS?
Nope. I mean one RAID 10 for the entire D2600 shelf. Apparently, you can
have a RAID 10 (1+0 or whatever) that spans more than 4 drives. Never
done it, but I've heard that it's possible.
Also
seems overkill. Two disks of RAID1 would be more than enough.
That's my plan.
Or if you
are using LVM, a small LV for the OS (plus a small boot partition).
Yep. That's the plan.
The DL360 appears to have 8 x 2.5" drive slots, are they part of the data
array or just for OS?
I will only have 2 drives there just for that RAID 1.
* What form of RAID are you planning to use for the data? Hardware raid
controller or software md raid? RAID10, RAID6, ..? For a mixed I/O pattern
which contains more than a tiny amount of writes, don't even consider RAID5
or RAID6.
Hardware RAID 10.
* Why three bricks per node? Are you planning to split the 12 x 2 TB drives
into three 4-drive arrays?
Well I was going to split up the large RAID 10 LUN with LVM. Each brick
would map to it's own dedicated LV on each node.
* What sort of network are you linking this to the compute nodes with?
GigE, 10gigE, Infiniband, something else?
1GbE unfortunately.
If so, do you think I get away with only 2 storage-nodes?
That's entirely dependent on your application. Since you are doing
distribution, not replication, maybe you could get away with one storage
node? What if you had 3TB drives instead of 2TB drives?
My hands are tied there. 2 TB is what we have.
Thanks for the response.
_______________________________________________
Gluster-users mailing list
[email protected]
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users