Hi Folks,

I find myself having four servers, each with 4 large disks, that I'm 
trying to assemble into a high-availability cluster.  (Note: I've got 4 
gigE ports on each box, 2 set aside for outside access, 2 for inter-node 
clustering)

Now it's easy enough to RAID disks on each server, and/or mirror disks 
pair-wise with DRBD, but DRBD doesn't work as well with >2 servers.

Now what I really should do is separate storage nodes from compute nodes 
- but I'm limited by rack space and chassis configuration of the 
hardware I've got, and I've been thinking through various configurations 
to make use of the resources at hand.

One option is to put all the drives into one large pool managed by 
gluster - but I expect that would result in some serious performance 
hits (and gluster's replicated/distributed mode is fairly new).

It's late at night and a thought occurred to me that is probably 
wrongheaded (or at least silly) - but maybe I'm too tired to see any 
obvious problems.  So I'd welcome 2nd (and 3rd) opinions.

The basic notion:
- mount all 16 drives as network block devices via iSCSI or AoE
- build 4 RAID10 volumes - each volume consisting of one drive from each 
server
- run LVM on top of the RAID volumes
- then use NFS or maybe OCFS2 to make volumes available across nodes
- of course md would be running on only one node (for each array), so if 
a node goes down, use pacemaker to startup md on another node, 
reassemble the array, and remount everything

Does this make sense, or is it totally crazy?

Thanks much,

Miles Fidelman

-- 
In theory, there is no difference between theory and practice.
In practice, there is.   .... Yogi Berra

_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to