Hi hvjunk

>From the logs it seems like the ganesha cluster is not yet created/running and 
>performing refresh config is failing.As the cluster is not running and volume 
>"ganesha" is not yet exported due to this performing refresh config on volume 
>will fail.
For creating ganesha cluster,there are few pre-requisite of adding services 
like  nlm,nfs,rpc-bind,high-availability,mountd,rquota to firewalld.
After executing create-cluster block ,It will validate the HA Status of ganesha 
via gdeploy.If the ganesha cluster is up and running it will reflect HA status 
as "HEALTHY".
After then performing refresh-config on nfs exported volume will succeed.


Attaching the sample config file to create ganesha cluster and exporting 
ganesha volume to client -

[hosts]
dhcp37-102.lab.eng.blr.redhat.com
dhcp37-92.lab.eng.blr.redhat.com
dhcp37-119.lab.eng.blr.redhat.com
dhcp37-122.lab.eng.blr.redhat.com


[backend-setup]
devices=/dev/sdb,/dev/sdc,/dev/sdd
vgs=vg1,vg2,vg3
pools=pool1,pool2,pool3
lvs=lv1,lv2,lv3
mountpoints=/gluster/brick1,/gluster/brick2,/gluster/brick3
brick_dirs=/gluster/brick1/1,/gluster/brick2/1,/gluster/brick3/1

[yum]
action=install
repolist=
gpgcheck=no
update=no
packages=glusterfs-ganesha

[firewalld]
action=add
ports=111/tcp,2049/tcp,54321/tcp,5900/tcp,5900-6923/tcp,5666/tcp,16514/tcp
services=glusterfs,nlm,nfs,rpc-bind,high-availability,mountd,rquota

[volume]
action=create
volname=ganesha
transport=tcp
replica_count=2
force=yes

[nfs-ganesha]
action=create-cluster
ha-name=ganesha-ha-360
cluster-nodes=dhcp37-102.lab.eng.blr.redhat.com,dhcp37-92.lab.eng.blr.redhat.com,dhcp37-119.lab.eng.blr.redhat.com,dhcp37-122.lab.eng.blr.redhat.com
vip=10.70.36.217,10.70.36.218,10.70.36.219,10.70.36.220
volname=ganesha  #This will export volume via nfs-ganesha

[clients]
action=mount
volname=10.70.37.153:/ganesha  #VIP from which the volume needs to be mounted 
to client
hosts=10.70.46.30    #Client IP
client_mount_points=/mnt/ganesha
fstype=nfs
options=rw,relatime,vers=4.0,rsize=1048576,wsize=1048576,namlen=255,hard,proto=tcp,port=0,timeo=600,retrans=2,sec=sys,clientaddr=10.70.44.153,local_lock=none,addr=10.70.46.30

Regards,
Manisha Saini

----- Original Message -----
From: "Soumya Koduri" <[email protected]>
To: "hvjunk" <[email protected]>, [email protected], "Sachidananda URS" 
<[email protected]>, "Manisha Saini" <[email protected]>, "Arthy Loganathan" 
<[email protected]>
Sent: Wednesday, May 10, 2017 10:45:57 AM
Subject: Re: [Gluster-users] gdeploy not starting all the daemons for 
NFS-ganesha :(

CCin Sac, Manisha, Arthy who could help with troubleshooting.

Thanks,
Soumya


On 05/09/2017 08:31 PM, hvjunk wrote:
> Hi there,
>
> Given the following config file, what am I doing wrong causing the error at 
> the bottom  & no mounted /gluster_shared_storage?
>
> Hendrik
>
> [root@linked-clone-of-centos-linux ~]# cat t.conf
> [hosts]
> 10.10.10.11
> 10.10.10.12
> 10.10.10.13
>
> [backend-setup]
> devices=/dev/sdb
> mountpoints=/gluster/brick1
> brick_dirs=/gluster/brick1/one
> pools=pool1
>
> #Installing nfs-ganesha
> [yum]
> action=install
> repolist=
> gpgcheck=no
> update=no
> packages=glusterfs-ganesha
>
> [service]
> action=start
> service=glusterd
>
> [service]
> action=enable
> service=glusterd
>
> #This will create a volume. Skip this section if your volume already exists
> [volume]
> action=create
> volname=ganesha
> transport=tcp
> replica_count=3
> arbiter_count=1
> force=yes
>
> [clients]
> action=mount
> volname=ganesha
> hosts=10.10.10.11,10.10.10.12,10.10.10.13
> fstype=glusterfs
> client_mount_points=/mnt/ganesha_mnt
>
>
> #Creating a high availability cluster and exporting the volume
> [nfs-ganesha]
> action=create-cluster
> ha-name=ganesha-ha-360
> cluster-nodes=10.10.10.11,10.10.10.12
> vip=10.10.10.31,10.10.10.41
> volname=ganesha
>
>
> [nfs-ganesha]
> action=export-volume
> volname=ganesha
>
> [nfs-ganesha]
> action=refresh-config
> volname=ganesha
> [root@linked-clone-of-centos-linux ~]#
>
>
> [root@linked-clone-of-centos-linux ~]# gdeploy -c t.conf -k
>
> PLAY [gluster_servers] 
> *************************************************************************************
>
> TASK [Clean up filesystem signature] 
> ***********************************************************************
> skipping: [10.10.10.11] => (item=/dev/sdb)
> skipping: [10.10.10.12] => (item=/dev/sdb)
> skipping: [10.10.10.13] => (item=/dev/sdb)
>
> TASK [Create Physical Volume] 
> ******************************************************************************
> changed: [10.10.10.13] => (item=/dev/sdb)
> changed: [10.10.10.11] => (item=/dev/sdb)
> changed: [10.10.10.12] => (item=/dev/sdb)
>
> PLAY RECAP 
> *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] 
> *************************************************************************************
>
> TASK [Create volume group on the disks] 
> ********************************************************************
> changed: [10.10.10.12] => (item={u'brick': u'/dev/sdb', u'vg': 
> u'GLUSTER_vg1'})
> changed: [10.10.10.13] => (item={u'brick': u'/dev/sdb', u'vg': 
> u'GLUSTER_vg1'})
> changed: [10.10.10.11] => (item={u'brick': u'/dev/sdb', u'vg': 
> u'GLUSTER_vg1'})
>
> PLAY RECAP 
> *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] 
> *************************************************************************************
>
> TASK [Create logical volume named metadata] 
> ****************************************************************
> changed: [10.10.10.12] => (item=GLUSTER_vg1)
> changed: [10.10.10.13] => (item=GLUSTER_vg1)
> changed: [10.10.10.11] => (item=GLUSTER_vg1)
>
> TASK [create data LV that has a size which is a multiple of stripe width] 
> **********************************
> changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', 
> u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', 
> u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', 
> u'vg': u'GLUSTER_vg1'})
>
> TASK [Convert the logical volume] 
> **************************************************************************
> changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', 
> u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', 
> u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', 
> u'vg': u'GLUSTER_vg1'})
>
> TASK [create stripe-aligned thin volume] 
> *******************************************************************
> changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', 
> u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', 
> u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', 
> u'vg': u'GLUSTER_vg1'})
>
> TASK [Change the attributes of the logical volume] 
> *********************************************************
> changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', 
> u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', 
> u'vg': u'GLUSTER_vg1'})
> changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', 
> u'vg': u'GLUSTER_vg1'})
>
> PLAY RECAP 
> *************************************************************************************************
> 10.10.10.11                : ok=5    changed=5    unreachable=0    failed=0
> 10.10.10.12                : ok=5    changed=5    unreachable=0    failed=0
> 10.10.10.13                : ok=5    changed=5    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] 
> *************************************************************************************
>
> TASK [Create an xfs filesystem] 
> ****************************************************************************
> changed: [10.10.10.13] => (item=/dev/GLUSTER_vg1/GLUSTER_lv1)
> changed: [10.10.10.12] => (item=/dev/GLUSTER_vg1/GLUSTER_lv1)
> changed: [10.10.10.11] => (item=/dev/GLUSTER_vg1/GLUSTER_lv1)
>
> PLAY RECAP 
> *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] 
> *************************************************************************************
>
> TASK [Create the backend disks, skips if present] 
> **********************************************************
> changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', 
> u'path': u'/gluster/brick1'})
> changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', 
> u'path': u'/gluster/brick1'})
> changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', 
> u'path': u'/gluster/brick1'})
>
> TASK [Mount the volumes] 
> ***********************************************************************************
> changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', 
> u'path': u'/gluster/brick1'})
> changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', 
> u'path': u'/gluster/brick1'})
> changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', 
> u'path': u'/gluster/brick1'})
>
> PLAY RECAP 
> *************************************************************************************************
> 10.10.10.11                : ok=2    changed=2    unreachable=0    failed=0
> 10.10.10.12                : ok=2    changed=2    unreachable=0    failed=0
> 10.10.10.13                : ok=2    changed=2    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] 
> *************************************************************************************
>
> TASK [Set SELinux labels on the bricks] 
> ********************************************************************
> changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', 
> u'path': u'/gluster/brick1'})
> changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', 
> u'path': u'/gluster/brick1'})
> changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', 
> u'path': u'/gluster/brick1'})
>
> TASK [Restore the SELinux context] 
> *************************************************************************
> changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', 
> u'path': u'/gluster/brick1'})
> changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', 
> u'path': u'/gluster/brick1'})
> changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', 
> u'path': u'/gluster/brick1'})
>
> PLAY RECAP 
> *************************************************************************************************
> 10.10.10.11                : ok=2    changed=2    unreachable=0    failed=0
> 10.10.10.12                : ok=2    changed=2    unreachable=0    failed=0
> 10.10.10.13                : ok=2    changed=2    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] 
> *************************************************************************************
>
> TASK [Create/Enabling yum repos] 
> ***************************************************************************
> skipping: [10.10.10.12] => (item=)
> skipping: [10.10.10.11] => (item=)
> skipping: [10.10.10.13] => (item=)
>
> TASK [Clean up the metadata] 
> *******************************************************************************
> skipping: [10.10.10.12]
> skipping: [10.10.10.11]
> skipping: [10.10.10.13]
>
> TASK [Yum update] 
> ******************************************************************************************
> skipping: [10.10.10.12]
> skipping: [10.10.10.11]
> skipping: [10.10.10.13]
>
> TASK [Installs/Removes a package using yum] 
> ****************************************************************
> changed: [10.10.10.12]
> changed: [10.10.10.13]
> changed: [10.10.10.11]
>
> PLAY RECAP 
> *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] 
> *************************************************************************************
>
> TASK [Enable or disable services] 
> **************************************************************************
> changed: [10.10.10.12] => (item=glusterd)
> changed: [10.10.10.11] => (item=glusterd)
> changed: [10.10.10.13] => (item=glusterd)
>
> PLAY RECAP 
> *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] 
> *************************************************************************************
>
> TASK [start/stop/restart/reload services] 
> ******************************************************************
> changed: [10.10.10.11] => (item=glusterd)
> changed: [10.10.10.12] => (item=glusterd)
> changed: [10.10.10.13] => (item=glusterd)
>
> PLAY RECAP 
> *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [gluster_servers] 
> *************************************************************************************
>
> TASK [Create the brick dirs, skips if present] 
> *************************************************************
> changed: [10.10.10.11] => (item=/gluster/brick1/one)
> changed: [10.10.10.12] => (item=/gluster/brick1/one)
> changed: [10.10.10.13] => (item=/gluster/brick1/one)
>
> PLAY RECAP 
> *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [master] 
> **********************************************************************************************
>
> TASK [Creates a Trusted Storage Pool] 
> **********************************************************************
> changed: [10.10.10.11]
>
> TASK [Pause for a few seconds] 
> *****************************************************************************
> Pausing for 5 seconds
> (ctrl+C then 'C' = continue early, ctrl+C then 'A' = abort)
> ok: [10.10.10.11]
>
> PLAY RECAP 
> *************************************************************************************************
> 10.10.10.11                : ok=2    changed=1    unreachable=0    failed=0
>
>
> PLAY [master] 
> **********************************************************************************************
>
> TASK [Creates a volume] 
> ************************************************************************************
> changed: [10.10.10.11]
>
> PLAY RECAP 
> *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [master] 
> **********************************************************************************************
>
> TASK [Starts a volume] 
> *************************************************************************************
> changed: [10.10.10.11]
>
> PLAY RECAP 
> *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [clients] 
> *********************************************************************************************
>
> TASK [Create the dir to mount the volume, skips if present] 
> ************************************************
> changed: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> changed: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> changed: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
>
> PLAY RECAP 
> *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [clients] 
> *********************************************************************************************
>
> TASK [Mount the volumes, if fstype is glusterfs] 
> ***********************************************************
> changed: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> changed: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> changed: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
>
> PLAY RECAP 
> *************************************************************************************************
> 10.10.10.11                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=1    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=1    unreachable=0    failed=0
>
>
> PLAY [clients] 
> *********************************************************************************************
>
> TASK [Gathering Facts] 
> *************************************************************************************
> ok: [10.10.10.11]
> ok: [10.10.10.12]
> ok: [10.10.10.13]
>
> TASK [Uncomment STATD_PORT for rpc.statd to listen on] 
> *****************************************************
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
>
> TASK [Uncomment LOCKD_TCPPORT for rpc.lockd to listen on] 
> **************************************************
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
>
> TASK [Uncomment LOCKD_UDPPORT for rpc.lockd to listen on] 
> **************************************************
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
>
> TASK [Uncomment MOUNTD_PORT for rpc.mountd to listen on] 
> ***************************************************
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
>
> TASK [Restart nfs service (RHEL 6 only)] 
> *******************************************************************
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
>
> TASK [Restart rpc-statd service] 
> ***************************************************************************
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
>
> TASK [Restart nfs-config service] 
> **************************************************************************
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
>
> TASK [Restart nfs-mountd service] 
> **************************************************************************
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
>
> TASK [Restart nfslock service (RHEL 6 & 7)] 
> ****************************************************************
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
>
> TASK [Mount the volumes if fstype is NFS] 
> ******************************************************************
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
>
> PLAY RECAP 
> *************************************************************************************************
> 10.10.10.11                : ok=1    changed=0    unreachable=0    failed=0
> 10.10.10.12                : ok=1    changed=0    unreachable=0    failed=0
> 10.10.10.13                : ok=1    changed=0    unreachable=0    failed=0
>
>
> PLAY [clients] 
> *********************************************************************************************
>
> TASK [Mount the volumes, if fstype is CIFS] 
> ****************************************************************
> skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
> skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', 
> u'fstype': u'fuse'})
>
> PLAY RECAP 
> *************************************************************************************************
> 10.10.10.11                : ok=0    changed=0    unreachable=0    failed=0
> 10.10.10.12                : ok=0    changed=0    unreachable=0    failed=0
> 10.10.10.13                : ok=0    changed=0    unreachable=0    failed=0
>
>
> PLAY [master] 
> **********************************************************************************************
>
> TASK [Delete lines from Ganesha export file] 
> ***************************************************************
>
> TASK [Add lines to Ganesha export file] 
> ********************************************************************
>
> TASK [Update lines to Ganesha export file] 
> *****************************************************************
>
> TASK [Add a block to Ganesha export file] 
> ******************************************************************
>
> TASK [Refresh NFS-Ganesha config] 
> **************************************************************************
> fatal: [10.10.10.11]: FAILED! => {"changed": true, "cmd": 
> "/usr/libexec/ganesha/ganesha-ha.sh --refresh-config 
> \"/var/run/gluster/shared_storage/nfs-ganesha\" \"ganesha\"", "delta": 
> "0:00:00.227074", "end": "2017-05-09 16:54:45.442219", "failed": true, "rc": 
> 1, "start": "2017-05-09 16:54:45.215145", "stderr": "grep: 
> /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or 
> directory\ngrep: /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: 
> No such file or directory\nError: cluster is not currently running on this 
> node\ngrep: 
> /var/run/gluster/shared_storage/nfs-ganesha/exports/export.ganesha.conf: No 
> such file or directory", "stderr_lines": ["grep: 
> /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or 
> directory", "grep: 
> /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or 
> directory", "Error: cluster is not currently running on this node", "grep: 
> /var/run/gluster/shared_storage/nfs-ganesha/exports/exp
 or
>  t.ganesha.conf: No such file or directory"], "stdout": "Error: 
> refresh-config failed on localhost.", "stdout_lines": ["Error: refresh-config 
> failed on localhost."]}
>       to retry, use: --limit @/tmp/tmpacx8yi/ganesha-refresh-config.retry
>
> PLAY RECAP 
> *************************************************************************************************
> 10.10.10.11                : ok=0    changed=0    unreachable=0    failed=1
>
> Ignoring errors...
> You can view the generated configuration files inside /tmp/tmpacx8yi
> [root@linked-clone-of-centos-linux ~]#
> _______________________________________________
> Gluster-users mailing list
> [email protected]
> http://lists.gluster.org/mailman/listinfo/gluster-users
>
_______________________________________________
Gluster-users mailing list
[email protected]
http://lists.gluster.org/mailman/listinfo/gluster-users

Reply via email to