On Fri, Jan 17, 2014 at 6:46 AM,  <[email protected]> wrote:
> Hi,
>
> I am currently testing GlusterFS and am looking for some advice. My setup 
> uses the latest 3.4.2 packages from www.gluster.org  SLES11-SP3.
>
> I currently have a storage pool shared read-write across 2 gluster server 
> nodes. This seems to work fine. However, I would also like to mount this pool 
> on 4 further client machines running a legacy web application. Because of 
> some limitations in the application, I would like to be able to tell these 
> servers to mount the storage pool from one particular gluster server node, 
> but to fail over to the second node if and only the first node becomes 
> unavailable. I can mount the storage on the client nodes with both gluster 
> nodes specified or with only one node specified but cannot see a way in the 
> documentation of preferring one particular node and having the second node 
> configured as a fail over. Is this possible? What am I missing?

You do realize that the initial connection is just for retrieving the
volfiles, and then all hosts are used, right? If so, carrying on:

You can use VRRP and a VIP to specify which host to mount from. An
example of this is done in my Puppet-Gluster setup:
https://ttboj.wordpress.com/2014/01/08/automatically-deploying-glusterfs-with-puppet-gluster-vagrant/

You can specify ordering of the VIP with priority arguments to
keepalived for example.

You can also specify more than one server on the mount command for
glusterfs. I forget the syntax for that, but it's easy to google.

I hope this answers your questions!

James



 My current gluster vol file on my clients is as follows:
>
> volume remote1
>   type protocol/client
>   option transport-type tcp
>   option remote-host GLUSTER-01
>   option remote-subvolume /data/gv0/brick1
> end-volume
>
> volume remote2
>   type protocol/client
>   option transport-type tcp
>   option remote-host GLUSTER-02
>   option remote-subvolume /data/gv0/brick1
> end-volume
>
> volume replicate
>   type cluster/replicate
>   subvolumes remote1 remote2
> end-volume
>
> volume writebehind
>   type performance/write-behind
>   option window-size 1MB
>   subvolumes replicate
> end-volume
>
> volume cache
>   type performance/io-cache
>   option cache-size 512MB
>   subvolumes writebehind
> end-volume
>
> And I have the following in the client's fstab:
>
> /etc/glusterfs/glusterfs.vol /opt/shared_files glusterfs 
> rw,allow_other,default_permissions,max_read=131072 0 0
>
>
> Thanks in advance for any help,
>
> Mike Peters
> _______________________________________________
> Gluster-users mailing list
> [email protected]
> http://supercolony.gluster.org/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Reply via email to