Hi,
OK, the problem with the VIPs not starting is due to the ganesha_mon heartbeat 
script looking for a pid file called /var/run/ganesha.nfsd.pid, while by 
default ganesha.nfsd v.2.2.0 is creating /var/run/ganesha.pid, this needs to be 
corrected. The file is in glusterfs-ganesha-3.7.1-1.el7.x86_64, in my case.
For the moment I have created a symlink in this way and it works:

ln -s /var/run/ganesha.pid /var/run/ganesha.nfsd.pid

So far so good, the VIPs are up and pingable, but still there is the problem of 
the hanging showmount (i.e. hanging RPC).
Still, I see a lot of errors like this in /var/log/messages:

Jun  9 11:15:20 atlas-node1 lrmd[31221]:   notice: operation_finished: 
nfs-mon_monitor_10000:29292:stderr [ Error: Resource does not exist. ]

While ganesha.log shows the server is not in grace:

09/06/2015 11:16:20 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29964[main] 
main :MAIN :EVENT :ganesha.nfsd Starting: Ganesha Version 
/builddir/build/BUILD/nfs-ganesha-2.2.0/src, built at May 18 2015 14:17:18 on 
buildhw-09.phx2.fedoraproject.org
09/06/2015 11:16:20 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[main] 
nfs_set_param_from_conf :NFS STARTUP :EVENT :Configuration file successfully 
parsed
09/06/2015 11:16:20 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[main] 
init_server_pkgs :NFS STARTUP :EVENT :Initializing ID Mapper.
09/06/2015 11:16:20 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[main] 
init_server_pkgs :NFS STARTUP :EVENT :ID Mapper successfully initialized.
09/06/2015 11:16:20 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[main] 
main :NFS STARTUP :WARN :No export entries found in configuration file !!!
09/06/2015 11:16:20 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[main] 
config_errs_to_log :CONFIG :WARN :Config File ((null):0): Empty configuration 
file
09/06/2015 11:16:20 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[main] 
lower_my_caps :NFS STARTUP :EVENT :CAP_SYS_RESOURCE was successfully removed 
for proper quota management in FSAL
09/06/2015 11:16:20 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[main] 
lower_my_caps :NFS STARTUP :EVENT :currenty set capabilities are: = 
cap_chown,cap_dac_override,cap_dac_read_search,cap_fowner,cap_fsetid,cap_kill,cap_setgid,cap_setuid,cap_setpcap,cap_linux_immutable,cap_net_bind_service,cap_net_broadcast,cap_net_admin,cap_net_raw,cap_ipc_lock,cap_ipc_owner,cap_sys_module,cap_sys_rawio,cap_sys_chroot,cap_sys_ptrace,cap_sys_pacct,cap_sys_admin,cap_sys_boot,cap_sys_nice,cap_sys_time,cap_sys_tty_config,cap_mknod,cap_lease,cap_audit_write,cap_audit_control,cap_setfcap+ep
09/06/2015 11:16:21 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[main] 
nfs_Init_svc :DISP :CRIT :Cannot acquire credentials for principal nfs
09/06/2015 11:16:21 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[main] 
nfs_Init_admin_thread :NFS CB :EVENT :Admin thread initialized
09/06/2015 11:16:21 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[main] 
nfs4_start_grace :STATE :EVENT :NFS Server Now IN GRACE, duration 60
09/06/2015 11:16:21 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[main] 
nfs_rpc_cb_init_ccache :NFS STARTUP :EVENT :Callback creds directory 
(/var/run/ganesha) already exists
09/06/2015 11:16:21 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[main] 
nfs_rpc_cb_init_ccache :NFS STARTUP :WARN :gssd_refresh_krb5_machine_credential 
failed (2:2)
09/06/2015 11:16:21 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[main] 
nfs_Start_threads :THREAD :EVENT :Starting delayed executor.
09/06/2015 11:16:22 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[main] 
nfs_Start_threads :THREAD :EVENT :9P/TCP dispatcher thread was started 
successfully
09/06/2015 11:16:22 : epoch 5576aee4 : atlas-node1 : 
ganesha.nfsd-29965[_9p_disp] _9p_dispatcher_thread :9P DISP :EVENT :9P 
dispatcher started
09/06/2015 11:16:22 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[main] 
nfs_Start_threads :THREAD :EVENT :gsh_dbusthread was started successfully
09/06/2015 11:16:22 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[main] 
nfs_Start_threads :THREAD :EVENT :admin thread was started successfully
09/06/2015 11:16:22 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[main] 
nfs_Start_threads :THREAD :EVENT :reaper thread was started successfully
09/06/2015 11:16:22 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[reaper] 
nfs_in_grace :STATE :EVENT :NFS Server Now IN GRACE
09/06/2015 11:16:22 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[main] 
nfs_Start_threads :THREAD :EVENT :General fridge was started successfully
09/06/2015 11:16:22 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[main] 
nfs_start :NFS STARTUP :EVENT :-------------------------------------------------
09/06/2015 11:16:22 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[main] 
nfs_start :NFS STARTUP :EVENT :             NFS SERVER INITIALIZED
09/06/2015 11:16:22 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[main] 
nfs_start :NFS STARTUP :EVENT :-------------------------------------------------
09/06/2015 11:17:22 : epoch 5576aee4 : atlas-node1 : ganesha.nfsd-29965[reaper] 
nfs_in_grace :STATE :EVENT :NFS Server Now NOT IN GRACE


Cheers,

        Alessandro


> Il giorno 09/giu/2015, alle ore 10:36, Alessandro De Salvo 
> <alessandro.desa...@roma1.infn.it> ha scritto:
> 
> Hi Soumya,
> 
>> Il giorno 09/giu/2015, alle ore 08:06, Soumya Koduri <skod...@redhat.com> ha 
>> scritto:
>> 
>> 
>> 
>> On 06/09/2015 01:31 AM, Alessandro De Salvo wrote:
>>> OK, I found at least one of the bugs.
>>> The /usr/libexec/ganesha/ganesha.sh has the following lines:
>>> 
>>>    if [ -e /etc/os-release ]; then
>>>        RHEL6_PCS_CNAME_OPTION=""
>>>    fi
>>> 
>>> This is OK for RHEL < 7, but does not work for >= 7. I have changed it to 
>>> the following, to make it working:
>>> 
>>>    if [ -e /etc/os-release ]; then
>>>        eval $(grep -F "REDHAT_SUPPORT_PRODUCT=" /etc/os-release)
>>>        [ "$REDHAT_SUPPORT_PRODUCT" == "Fedora" ] && 
>>> RHEL6_PCS_CNAME_OPTION=""
>>>    fi
>>> 
>> Oh..Thanks for the fix. Could you please file a bug for the same (and 
>> probably submit your fix as well). We shall have it corrected.
> 
> Just did it, https://bugzilla.redhat.com/show_bug.cgi?id=1229601 
> <https://bugzilla.redhat.com/show_bug.cgi?id=1229601>
> 
>> 
>>> Apart from that, the VIP_<node> I was using were wrong, and I should have 
>>> converted all the “-“ to underscores, maybe this could be mentioned in the 
>>> documentation when you will have it ready.
>>> Now, the cluster starts, but the VIPs apparently not:
>>> 
>> Sure. Thanks again for pointing it out. We shall make a note of it.
>> 
>>> Online: [ atlas-node1 atlas-node2 ]
>>> 
>>> Full list of resources:
>>> 
>>> Clone Set: nfs-mon-clone [nfs-mon]
>>>     Started: [ atlas-node1 atlas-node2 ]
>>> Clone Set: nfs-grace-clone [nfs-grace]
>>>     Started: [ atlas-node1 atlas-node2 ]
>>> atlas-node1-cluster_ip-1  (ocf::heartbeat:IPaddr):        Stopped
>>> atlas-node1-trigger_ip-1  (ocf::heartbeat:Dummy): Started atlas-node1
>>> atlas-node2-cluster_ip-1  (ocf::heartbeat:IPaddr):        Stopped
>>> atlas-node2-trigger_ip-1  (ocf::heartbeat:Dummy): Started atlas-node2
>>> atlas-node1-dead_ip-1     (ocf::heartbeat:Dummy): Started atlas-node1
>>> atlas-node2-dead_ip-1     (ocf::heartbeat:Dummy): Started atlas-node2
>>> 
>>> PCSD Status:
>>>  atlas-node1: Online
>>>  atlas-node2: Online
>>> 
>>> Daemon Status:
>>>  corosync: active/disabled
>>>  pacemaker: active/disabled
>>>  pcsd: active/enabled
>>> 
>>> 
>> Here corosync and pacemaker shows 'disabled' state. Can you check the status 
>> of their services. They should be running prior to cluster creation. We need 
>> to include that step in document as well.
> 
> Ah, OK, you’re right, I have added it to my puppet modules (we install and 
> configure ganesha via puppet, I’ll put the module on puppetforge soon, in 
> case anyone is interested).
> 
>> 
>>> But the issue that is puzzling me more is the following:
>>> 
>>> # showmount -e localhost
>>> rpc mount export: RPC: Timed out
>>> 
>>> And when I try to enable the ganesha exports on a volume I get this error:
>>> 
>>> # gluster volume set atlas-home-01 ganesha.enable on
>>> volume set: failed: Failed to create NFS-Ganesha export config file.
>>> 
>>> But I see the file created in /etc/ganesha/exports/*.conf
>>> Still, showmount hangs and times out.
>>> Any help?
>>> Thanks,
>>> 
>> Hmm that's strange. Sometimes, in case if there was no proper cleanup done 
>> while trying to re-create the cluster, we have seen such issues.
>> 
>> https://bugzilla.redhat.com/show_bug.cgi?id=1227709
>> 
>> http://review.gluster.org/#/c/11093/
>> 
>> Can you please unexport all the volumes, teardown the cluster using
>> 'gluster vol set <volname> ganesha.enable off’
> 
> OK:
> 
> # gluster vol set atlas-home-01 ganesha.enable off
> volume set: failed: ganesha.enable is already 'off'.
> 
> # gluster vol set atlas-data-01 ganesha.enable off
> volume set: failed: ganesha.enable is already 'off'.
> 
> 
>> 'gluster ganesha disable' command.
> 
> I’m assuming you wanted to write nfs-ganesha instead?
> 
> # gluster nfs-ganesha disable
> ganesha enable : success
> 
> 
> A side note (not really important): it’s strange that when I do a disable the 
> message is “ganesha enable” :-)
> 
>> 
>> Verify if the following files have been deleted on all the nodes-
>> '/etc/cluster/cluster.conf’
> 
> this file is not present at all, I think it’s not needed in CentOS 7
> 
>> '/etc/ganesha/ganesha.conf’,
> 
> it’s still there, but empty, and I guess it should be OK, right?
> 
>> '/etc/ganesha/exports/*’
> 
> no more files there
> 
>> '/var/lib/pacemaker/cib’
> 
> it’s empty
> 
>> 
>> Verify if the ganesha service is stopped on all the nodes.
> 
> nope, it’s still running, I will stop it.
> 
>> 
>> start/restart the services - corosync, pcs.
> 
> In the node where I issued the nfs-ganesha disable there is no more any 
> /etc/corosync/corosync.conf so corosync won’t start. The other node instead 
> still has the file, it’s strange.
> 
>> 
>> And re-try the HA cluster creation
>> 'gluster ganesha enable’
> 
> This time (repeated twice) it did not work at all:
> 
> # pcs status
> Cluster name: ATLAS_GANESHA_01
> Last updated: Tue Jun  9 10:13:43 2015
> Last change: Tue Jun  9 10:13:22 2015
> Stack: corosync
> Current DC: atlas-node1 (1) - partition with quorum
> Version: 1.1.12-a14efad
> 2 Nodes configured
> 6 Resources configured
> 
> 
> Online: [ atlas-node1 atlas-node2 ]
> 
> Full list of resources:
> 
> Clone Set: nfs-mon-clone [nfs-mon]
>     Started: [ atlas-node1 atlas-node2 ]
> Clone Set: nfs-grace-clone [nfs-grace]
>     Started: [ atlas-node1 atlas-node2 ]
> atlas-node2-dead_ip-1     (ocf::heartbeat:Dummy): Started atlas-node1 
> atlas-node1-dead_ip-1     (ocf::heartbeat:Dummy): Started atlas-node2 
> 
> PCSD Status:
>  atlas-node1: Online
>  atlas-node2: Online
> 
> Daemon Status:
>  corosync: active/enabled
>  pacemaker: active/enabled
>  pcsd: active/enabled
> 
> 
> 
> I tried then "pcs cluster destroy" on both nodes, and then again nfs-ganesha 
> enable, but now I’m back to the old problem:
> 
> # pcs status
> Cluster name: ATLAS_GANESHA_01
> Last updated: Tue Jun  9 10:22:27 2015
> Last change: Tue Jun  9 10:17:00 2015
> Stack: corosync
> Current DC: atlas-node2 (2) - partition with quorum
> Version: 1.1.12-a14efad
> 2 Nodes configured
> 10 Resources configured
> 
> 
> Online: [ atlas-node1 atlas-node2 ]
> 
> Full list of resources:
> 
> Clone Set: nfs-mon-clone [nfs-mon]
>     Started: [ atlas-node1 atlas-node2 ]
> Clone Set: nfs-grace-clone [nfs-grace]
>     Started: [ atlas-node1 atlas-node2 ]
> atlas-node1-cluster_ip-1       (ocf::heartbeat:IPaddr):        Stopped 
> atlas-node1-trigger_ip-1       (ocf::heartbeat:Dummy): Started atlas-node1 
> atlas-node2-cluster_ip-1       (ocf::heartbeat:IPaddr):        Stopped 
> atlas-node2-trigger_ip-1       (ocf::heartbeat:Dummy): Started atlas-node2 
> atlas-node1-dead_ip-1  (ocf::heartbeat:Dummy): Started atlas-node1 
> atlas-node2-dead_ip-1  (ocf::heartbeat:Dummy): Started atlas-node2 
> 
> PCSD Status:
>  atlas-node1: Online
>  atlas-node2: Online
> 
> Daemon Status:
>  corosync: active/enabled
>  pacemaker: active/enabled
>  pcsd: active/enabled
> 
> 
> Cheers,
> 
>       Alessandro
> 
>> 
>> 
>> Thanks,
>> Soumya
>> 
>>>     Alessandro
>>> 
>>>> Il giorno 08/giu/2015, alle ore 20:00, Alessandro De Salvo 
>>>> <alessandro.desa...@roma1.infn.it> ha scritto:
>>>> 
>>>> Hi,
>>>> indeed, it does not work :-)
>>>> OK, this is what I did, with 2 machines, running CentOS 7.1, Glusterfs 
>>>> 3.7.1 and nfs-ganesha 2.2.0:
>>>> 
>>>> 1) ensured that the machines are able to resolve their IPs (but this was 
>>>> already true since they were in the DNS);
>>>> 2) disabled NetworkManager and enabled network on both machines;
>>>> 3) created a gluster shared volume 'gluster_shared_storage' and mounted it 
>>>> on '/run/gluster/shared_storage' on all the cluster nodes using glusterfs 
>>>> native mount (on CentOS 7.1 there is a link by default /var/run -> ../run)
>>>> 4) created an empty /etc/ganesha/ganesha.conf;
>>>> 5) installed pacemaker pcs resource-agents corosync on all cluster 
>>>> machines;
>>>> 6) set the ‘hacluster’ user the same password on all machines;
>>>> 7) pcs cluster auth <hostname> -u hacluster -p <pass> on all the nodes (on 
>>>> both nodes I issued the commands for both nodes)
>>>> 8) IPv6 is configured by default on all nodes, although the infrastructure 
>>>> is not ready for IPv6
>>>> 9) enabled pcsd and started it on all nodes
>>>> 10) populated /etc/ganesha/ganesha-ha.conf with the following contents, 
>>>> one per machine:
>>>> 
>>>> 
>>>> ===> atlas-node1
>>>> # Name of the HA cluster created.
>>>> HA_NAME="ATLAS_GANESHA_01"
>>>> # The server from which you intend to mount
>>>> # the shared volume.
>>>> HA_VOL_SERVER=“atlas-node1"
>>>> # The subset of nodes of the Gluster Trusted Pool
>>>> # that forms the ganesha HA cluster. IP/Hostname
>>>> # is specified.
>>>> HA_CLUSTER_NODES=“atlas-node1,atlas-node2"
>>>> # Virtual IPs of each of the nodes specified above.
>>>> VIP_atlas-node1=“x.x.x.1"
>>>> VIP_atlas-node2=“x.x.x.2"
>>>> 
>>>> ===> atlas-node2
>>>> # Name of the HA cluster created.
>>>> HA_NAME="ATLAS_GANESHA_01"
>>>> # The server from which you intend to mount
>>>> # the shared volume.
>>>> HA_VOL_SERVER=“atlas-node2"
>>>> # The subset of nodes of the Gluster Trusted Pool
>>>> # that forms the ganesha HA cluster. IP/Hostname
>>>> # is specified.
>>>> HA_CLUSTER_NODES=“atlas-node1,atlas-node2"
>>>> # Virtual IPs of each of the nodes specified above.
>>>> VIP_atlas-node1=“x.x.x.1"
>>>> VIP_atlas-node2=“x.x.x.2”
>>>> 
>>>> 11) issued gluster nfs-ganesha enable, but it fails with a cryptic message:
>>>> 
>>>> # gluster nfs-ganesha enable
>>>> Enabling NFS-Ganesha requires Gluster-NFS to be disabled across the 
>>>> trusted pool. Do you still want to continue? (y/n) y
>>>> nfs-ganesha: failed: Failed to set up HA config for NFS-Ganesha. Please 
>>>> check the log file for details
>>>> 
>>>> Looking at the logs I found nothing really special but this:
>>>> 
>>>> ==> /var/log/glusterfs/etc-glusterfs-glusterd.vol.log <==
>>>> [2015-06-08 17:57:15.672844] I [MSGID: 106132] 
>>>> [glusterd-proc-mgmt.c:83:glusterd_proc_stop] 0-management: nfs already 
>>>> stopped
>>>> [2015-06-08 17:57:15.675395] I [glusterd-ganesha.c:386:check_host_list] 
>>>> 0-management: ganesha host found Hostname is atlas-node2
>>>> [2015-06-08 17:57:15.720692] I [glusterd-ganesha.c:386:check_host_list] 
>>>> 0-management: ganesha host found Hostname is atlas-node2
>>>> [2015-06-08 17:57:15.721161] I [glusterd-ganesha.c:335:is_ganesha_host] 
>>>> 0-management: ganesha host found Hostname is atlas-node2
>>>> [2015-06-08 17:57:16.633048] E 
>>>> [glusterd-ganesha.c:254:glusterd_op_set_ganesha] 0-management: Initial 
>>>> NFS-Ganesha set up failed
>>>> [2015-06-08 17:57:16.641563] E [glusterd-syncop.c:1396:gd_commit_op_phase] 
>>>> 0-management: Commit of operation 'Volume (null)' failed on localhost : 
>>>> Failed to set up HA config for NFS-Ganesha. Please check the log file for 
>>>> details
>>>> 
>>>> ==> /var/log/glusterfs/cmd_history.log <==
>>>> [2015-06-08 17:57:16.643615]  : nfs-ganesha enable : FAILED : Failed to 
>>>> set up HA config for NFS-Ganesha. Please check the log file for details
>>>> 
>>>> ==> /var/log/glusterfs/cli.log <==
>>>> [2015-06-08 17:57:16.643839] I [input.c:36:cli_batch] 0-: Exiting with: -1
>>>> 
>>>> 
>>>> Also, pcs seems to be fine for the auth part, although it obviously tells 
>>>> me the cluster is not running.
>>>> 
>>>> I, [2015-06-08T19:57:16.305323 #7223]  INFO -- : Running: 
>>>> /usr/sbin/corosync-cmapctl totem.cluster_name
>>>> I, [2015-06-08T19:57:16.345457 #7223]  INFO -- : Running: /usr/sbin/pcs 
>>>> cluster token-nodes
>>>> ::ffff:141.108.38.46 - - [08/Jun/2015 19:57:16] "GET /remote/check_auth 
>>>> HTTP/1.1" 200 68 0.1919
>>>> ::ffff:141.108.38.46 - - [08/Jun/2015 19:57:16] "GET /remote/check_auth 
>>>> HTTP/1.1" 200 68 0.1920
>>>> atlas-node1.mydomain - - [08/Jun/2015:19:57:16 CEST] "GET 
>>>> /remote/check_auth HTTP/1.1" 200 68
>>>> - -> /remote/check_auth
>>>> 
>>>> 
>>>> What am I doing wrong?
>>>> Thanks,
>>>> 
>>>>    Alessandro
>>>> 
>>>>> Il giorno 08/giu/2015, alle ore 19:30, Soumya Koduri <skod...@redhat.com> 
>>>>> ha scritto:
>>>>> 
>>>>> 
>>>>> 
>>>>> 
>>>>> On 06/08/2015 08:20 PM, Alessandro De Salvo wrote:
>>>>>> Sorry, just another question:
>>>>>> 
>>>>>> - in my installation of gluster 3.7.1 the command gluster 
>>>>>> features.ganesha enable does not work:
>>>>>> 
>>>>>> # gluster features.ganesha enable
>>>>>> unrecognized word: features.ganesha (position 0)
>>>>>> 
>>>>>> Which version has full support for it?
>>>>> 
>>>>> Sorry. This option has recently been changed. It is now
>>>>> 
>>>>> $ gluster nfs-ganesha enable
>>>>> 
>>>>> 
>>>>>> 
>>>>>> - in the documentation the ccs and cman packages are required, but they 
>>>>>> seems not to be available anymore on CentOS 7 and similar, I guess they 
>>>>>> are not really required anymore, as pcs should do the full job
>>>>>> 
>>>>>> Thanks,
>>>>>> 
>>>>>>  Alessandro
>>>>> 
>>>>> Looks like so from http://clusterlabs.org/quickstart-redhat.html. Let us 
>>>>> know if it doesn't work.
>>>>> 
>>>>> Thanks,
>>>>> Soumya
>>>>> 
>>>>>> 
>>>>>>> Il giorno 08/giu/2015, alle ore 15:09, Alessandro De Salvo 
>>>>>>> <alessandro.desa...@roma1.infn.it> ha scritto:
>>>>>>> 
>>>>>>> Great, many thanks Soumya!
>>>>>>> Cheers,
>>>>>>> 
>>>>>>>         Alessandro
>>>>>>> 
>>>>>>>> Il giorno 08/giu/2015, alle ore 13:53, Soumya Koduri 
>>>>>>>> <skod...@redhat.com> ha scritto:
>>>>>>>> 
>>>>>>>> Hi,
>>>>>>>> 
>>>>>>>> Please find the slides of the demo video at [1]
>>>>>>>> 
>>>>>>>> We recommend to have a distributed replica volume as a shared volume 
>>>>>>>> for better data-availability.
>>>>>>>> 
>>>>>>>> Size of the volume depends on the workload you may have. Since it is 
>>>>>>>> used to maintain states of NLM/NFSv4 clients, you may calculate the 
>>>>>>>> size of the volume to be minimum of aggregate of
>>>>>>>> (typical_size_of'/var/lib/nfs'_directory + 
>>>>>>>> ~4k*no_of_clients_connected_to_each_of_the_nfs_servers_at_any_point)
>>>>>>>> 
>>>>>>>> We shall document about this feature sooner in the gluster docs as 
>>>>>>>> well.
>>>>>>>> 
>>>>>>>> Thanks,
>>>>>>>> Soumya
>>>>>>>> 
>>>>>>>> [1] - http://www.slideshare.net/SoumyaKoduri/high-49117846
>>>>>>>> 
>>>>>>>> On 06/08/2015 04:34 PM, Alessandro De Salvo wrote:
>>>>>>>>> Hi,
>>>>>>>>> I have seen the demo video on ganesha HA, 
>>>>>>>>> https://www.youtube.com/watch?v=Z4mvTQC-efM
>>>>>>>>> However there is no advice on the appropriate size of the shared 
>>>>>>>>> volume. How is it really used, and what should be a reasonable size 
>>>>>>>>> for it?
>>>>>>>>> Also, are the slides from the video available somewhere, as well as a 
>>>>>>>>> documentation on all this? I did not manage to find them.
>>>>>>>>> Thanks,
>>>>>>>>> 
>>>>>>>>>       Alessandro
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> _______________________________________________
>>>>>>>>> Gluster-users mailing list
>>>>>>>>> Gluster-users@gluster.org
>>>>>>>>> http://www.gluster.org/mailman/listinfo/gluster-users
>>>>>>>>> 
>>>>>>> 
>>>>>> 
>>>> 
>>>> _______________________________________________
>>>> Gluster-users mailing list
>>>> Gluster-users@gluster.org
>>>> http://www.gluster.org/mailman/listinfo/gluster-users
>>> 
> 
> _______________________________________________
> Gluster-users mailing list
> Gluster-users@gluster.org <mailto:Gluster-users@gluster.org>
> http://www.gluster.org/mailman/listinfo/gluster-users 
> <http://www.gluster.org/mailman/listinfo/gluster-users>

Attachment: smime.p7s
Description: S/MIME cryptographic signature

_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Reply via email to