Hi Roberto,
Gluster Storage Platform provides client volume spec file through server for
created volumes.
You can mount using
mount -t glusterfs <server>:<volume>-<transport> <your-mount-point>
for example,
mount -t glusterfs node01:gluster01-tcp /mnt/gluster01
Its not required to write your own spec file for mounting it.
Thanks,
Regards,
Bala
Roberto Lucignani wrote:
Hi all,
I installed two Gluster Storage Platorm 3.0.4 on two servers node01 e
node02.
I created a volume called gluster01 than I mounted it on a Debian box in
this way:
mount -t glusterfs /etc/glusterfs/gluster01-tcp.vol /mnt/gluster01/
the gluster01-tcp.vol is the following:
volume 192.168.0.200-1
type protocol/client
option transport-type tcp
option remote-host 192.168.0.200
option transport.socket.nodelay on
option transport.remote-port 10012
option remote-subvolume brick1
end-volume
volume 192.168.0.200-2
type protocol/client
option transport-type tcp
option remote-host 192.168.0.200
option transport.socket.nodelay on
option transport.remote-port 10012
option remote-subvolume brick2
end-volume
volume 192.168.0.201-1
type protocol/client
option transport-type tcp
option remote-host 192.168.0.201
option transport.socket.nodelay on
option transport.remote-port 10012
option remote-subvolume brick1
end-volume
volume 192.168.0.201-2
type protocol/client
option transport-type tcp
option remote-host 192.168.0.201
option transport.socket.nodelay on
option transport.remote-port 10012
option remote-subvolume brick2
end-volume
volume mirror-0
type cluster/replicate
subvolumes 192.168.0.201-1 192.168.0.200-1
end-volume
volume mirror-1
type cluster/replicate
subvolumes 192.168.0.201-2 192.168.0.200-2
end-volume
volume distribute
type cluster/distribute
subvolumes mirror-0 mirror-1
end-volume
volume readahead
type performance/read-ahead
option page-count 4
subvolumes distribute
end-volume
volume iocache
type performance/io-cache
option cache-size `echo $(( $(grep 'MemTotal' /proc/meminfo | sed
's/[^0-9]//g') / 5120 ))`MB
option cache-timeout 1
subvolumes readahead
end-volume
volume quickread
type performance/quick-read
option cache-timeout 1
option max-file-size 64kB
subvolumes iocache
end-volume
volume writebehind
type performance/write-behind
option cache-size 4MB
subvolumes quickread
end-volume
volume statprefetch
type performance/stat-prefetch
subvolumes writebehind
end-volume
all works fine and smooth, I can write and read on that volume without any
problem.
The problem is when the node01 is unavailable, I can't access the volume via
mount on the Debian box. This doesn't happen if it is the node02 to be
unavailable.
I expected the same behavior in the two cases, in this way the node01
represents an SPOF, am I wrong ? am I missing something the configuration ?
Tnx in advance
Rpberto
------------------------------------------------------------------------
_______________________________________________
Gluster-users mailing list
[email protected]
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
[email protected]
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users