Hi hvjunk >From the logs it seems like the ganesha cluster is not yet created/running and >performing refresh config is failing.As the cluster is not running and volume >"ganesha" is not yet exported due to this performing refresh config on volume >will fail. For creating ganesha cluster,there are few pre-requisite of adding services like nlm,nfs,rpc-bind,high-availability,mountd,rquota to firewalld. After executing create-cluster block ,It will validate the HA Status of ganesha via gdeploy.If the ganesha cluster is up and running it will reflect HA status as "HEALTHY". After then performing refresh-config on nfs exported volume will succeed.
Attaching the sample config file to create ganesha cluster and exporting ganesha volume to client - [hosts] dhcp37-102.lab.eng.blr.redhat.com dhcp37-92.lab.eng.blr.redhat.com dhcp37-119.lab.eng.blr.redhat.com dhcp37-122.lab.eng.blr.redhat.com [backend-setup] devices=/dev/sdb,/dev/sdc,/dev/sdd vgs=vg1,vg2,vg3 pools=pool1,pool2,pool3 lvs=lv1,lv2,lv3 mountpoints=/gluster/brick1,/gluster/brick2,/gluster/brick3 brick_dirs=/gluster/brick1/1,/gluster/brick2/1,/gluster/brick3/1 [yum] action=install repolist= gpgcheck=no update=no packages=glusterfs-ganesha [firewalld] action=add ports=111/tcp,2049/tcp,54321/tcp,5900/tcp,5900-6923/tcp,5666/tcp,16514/tcp services=glusterfs,nlm,nfs,rpc-bind,high-availability,mountd,rquota [volume] action=create volname=ganesha transport=tcp replica_count=2 force=yes [nfs-ganesha] action=create-cluster ha-name=ganesha-ha-360 cluster-nodes=dhcp37-102.lab.eng.blr.redhat.com,dhcp37-92.lab.eng.blr.redhat.com,dhcp37-119.lab.eng.blr.redhat.com,dhcp37-122.lab.eng.blr.redhat.com vip=10.70.36.217,10.70.36.218,10.70.36.219,10.70.36.220 volname=ganesha #This will export volume via nfs-ganesha [clients] action=mount volname=10.70.37.153:/ganesha #VIP from which the volume needs to be mounted to client hosts=10.70.46.30 #Client IP client_mount_points=/mnt/ganesha fstype=nfs options=rw,relatime,vers=4.0,rsize=1048576,wsize=1048576,namlen=255,hard,proto=tcp,port=0,timeo=600,retrans=2,sec=sys,clientaddr=10.70.44.153,local_lock=none,addr=10.70.46.30 Regards, Manisha Saini ----- Original Message ----- From: "Soumya Koduri" <[email protected]> To: "hvjunk" <[email protected]>, [email protected], "Sachidananda URS" <[email protected]>, "Manisha Saini" <[email protected]>, "Arthy Loganathan" <[email protected]> Sent: Wednesday, May 10, 2017 10:45:57 AM Subject: Re: [Gluster-users] gdeploy not starting all the daemons for NFS-ganesha :( CCin Sac, Manisha, Arthy who could help with troubleshooting. Thanks, Soumya On 05/09/2017 08:31 PM, hvjunk wrote: > Hi there, > > Given the following config file, what am I doing wrong causing the error at > the bottom & no mounted /gluster_shared_storage? > > Hendrik > > [root@linked-clone-of-centos-linux ~]# cat t.conf > [hosts] > 10.10.10.11 > 10.10.10.12 > 10.10.10.13 > > [backend-setup] > devices=/dev/sdb > mountpoints=/gluster/brick1 > brick_dirs=/gluster/brick1/one > pools=pool1 > > #Installing nfs-ganesha > [yum] > action=install > repolist= > gpgcheck=no > update=no > packages=glusterfs-ganesha > > [service] > action=start > service=glusterd > > [service] > action=enable > service=glusterd > > #This will create a volume. Skip this section if your volume already exists > [volume] > action=create > volname=ganesha > transport=tcp > replica_count=3 > arbiter_count=1 > force=yes > > [clients] > action=mount > volname=ganesha > hosts=10.10.10.11,10.10.10.12,10.10.10.13 > fstype=glusterfs > client_mount_points=/mnt/ganesha_mnt > > > #Creating a high availability cluster and exporting the volume > [nfs-ganesha] > action=create-cluster > ha-name=ganesha-ha-360 > cluster-nodes=10.10.10.11,10.10.10.12 > vip=10.10.10.31,10.10.10.41 > volname=ganesha > > > [nfs-ganesha] > action=export-volume > volname=ganesha > > [nfs-ganesha] > action=refresh-config > volname=ganesha > [root@linked-clone-of-centos-linux ~]# > > > [root@linked-clone-of-centos-linux ~]# gdeploy -c t.conf -k > > PLAY [gluster_servers] > ************************************************************************************* > > TASK [Clean up filesystem signature] > *********************************************************************** > skipping: [10.10.10.11] => (item=/dev/sdb) > skipping: [10.10.10.12] => (item=/dev/sdb) > skipping: [10.10.10.13] => (item=/dev/sdb) > > TASK [Create Physical Volume] > ****************************************************************************** > changed: [10.10.10.13] => (item=/dev/sdb) > changed: [10.10.10.11] => (item=/dev/sdb) > changed: [10.10.10.12] => (item=/dev/sdb) > > PLAY RECAP > ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [gluster_servers] > ************************************************************************************* > > TASK [Create volume group on the disks] > ******************************************************************** > changed: [10.10.10.12] => (item={u'brick': u'/dev/sdb', u'vg': > u'GLUSTER_vg1'}) > changed: [10.10.10.13] => (item={u'brick': u'/dev/sdb', u'vg': > u'GLUSTER_vg1'}) > changed: [10.10.10.11] => (item={u'brick': u'/dev/sdb', u'vg': > u'GLUSTER_vg1'}) > > PLAY RECAP > ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [gluster_servers] > ************************************************************************************* > > TASK [Create logical volume named metadata] > **************************************************************** > changed: [10.10.10.12] => (item=GLUSTER_vg1) > changed: [10.10.10.13] => (item=GLUSTER_vg1) > changed: [10.10.10.11] => (item=GLUSTER_vg1) > > TASK [create data LV that has a size which is a multiple of stripe width] > ********************************** > changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', > u'vg': u'GLUSTER_vg1'}) > changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', > u'vg': u'GLUSTER_vg1'}) > changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', > u'vg': u'GLUSTER_vg1'}) > > TASK [Convert the logical volume] > ************************************************************************** > changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', > u'vg': u'GLUSTER_vg1'}) > changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', > u'vg': u'GLUSTER_vg1'}) > changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', > u'vg': u'GLUSTER_vg1'}) > > TASK [create stripe-aligned thin volume] > ******************************************************************* > changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', > u'vg': u'GLUSTER_vg1'}) > changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', > u'vg': u'GLUSTER_vg1'}) > changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', > u'vg': u'GLUSTER_vg1'}) > > TASK [Change the attributes of the logical volume] > ********************************************************* > changed: [10.10.10.11] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', > u'vg': u'GLUSTER_vg1'}) > changed: [10.10.10.13] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', > u'vg': u'GLUSTER_vg1'}) > changed: [10.10.10.12] => (item={u'lv': u'GLUSTER_lv1', u'pool': u'pool1', > u'vg': u'GLUSTER_vg1'}) > > PLAY RECAP > ************************************************************************************************* > 10.10.10.11 : ok=5 changed=5 unreachable=0 failed=0 > 10.10.10.12 : ok=5 changed=5 unreachable=0 failed=0 > 10.10.10.13 : ok=5 changed=5 unreachable=0 failed=0 > > > PLAY [gluster_servers] > ************************************************************************************* > > TASK [Create an xfs filesystem] > **************************************************************************** > changed: [10.10.10.13] => (item=/dev/GLUSTER_vg1/GLUSTER_lv1) > changed: [10.10.10.12] => (item=/dev/GLUSTER_vg1/GLUSTER_lv1) > changed: [10.10.10.11] => (item=/dev/GLUSTER_vg1/GLUSTER_lv1) > > PLAY RECAP > ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [gluster_servers] > ************************************************************************************* > > TASK [Create the backend disks, skips if present] > ********************************************************** > changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', > u'path': u'/gluster/brick1'}) > changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', > u'path': u'/gluster/brick1'}) > changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', > u'path': u'/gluster/brick1'}) > > TASK [Mount the volumes] > *********************************************************************************** > changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', > u'path': u'/gluster/brick1'}) > changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', > u'path': u'/gluster/brick1'}) > changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', > u'path': u'/gluster/brick1'}) > > PLAY RECAP > ************************************************************************************************* > 10.10.10.11 : ok=2 changed=2 unreachable=0 failed=0 > 10.10.10.12 : ok=2 changed=2 unreachable=0 failed=0 > 10.10.10.13 : ok=2 changed=2 unreachable=0 failed=0 > > > PLAY [gluster_servers] > ************************************************************************************* > > TASK [Set SELinux labels on the bricks] > ******************************************************************** > changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', > u'path': u'/gluster/brick1'}) > changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', > u'path': u'/gluster/brick1'}) > changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', > u'path': u'/gluster/brick1'}) > > TASK [Restore the SELinux context] > ************************************************************************* > changed: [10.10.10.11] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', > u'path': u'/gluster/brick1'}) > changed: [10.10.10.12] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', > u'path': u'/gluster/brick1'}) > changed: [10.10.10.13] => (item={u'device': u'/dev/GLUSTER_vg1/GLUSTER_lv1', > u'path': u'/gluster/brick1'}) > > PLAY RECAP > ************************************************************************************************* > 10.10.10.11 : ok=2 changed=2 unreachable=0 failed=0 > 10.10.10.12 : ok=2 changed=2 unreachable=0 failed=0 > 10.10.10.13 : ok=2 changed=2 unreachable=0 failed=0 > > > PLAY [gluster_servers] > ************************************************************************************* > > TASK [Create/Enabling yum repos] > *************************************************************************** > skipping: [10.10.10.12] => (item=) > skipping: [10.10.10.11] => (item=) > skipping: [10.10.10.13] => (item=) > > TASK [Clean up the metadata] > ******************************************************************************* > skipping: [10.10.10.12] > skipping: [10.10.10.11] > skipping: [10.10.10.13] > > TASK [Yum update] > ****************************************************************************************** > skipping: [10.10.10.12] > skipping: [10.10.10.11] > skipping: [10.10.10.13] > > TASK [Installs/Removes a package using yum] > **************************************************************** > changed: [10.10.10.12] > changed: [10.10.10.13] > changed: [10.10.10.11] > > PLAY RECAP > ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [gluster_servers] > ************************************************************************************* > > TASK [Enable or disable services] > ************************************************************************** > changed: [10.10.10.12] => (item=glusterd) > changed: [10.10.10.11] => (item=glusterd) > changed: [10.10.10.13] => (item=glusterd) > > PLAY RECAP > ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [gluster_servers] > ************************************************************************************* > > TASK [start/stop/restart/reload services] > ****************************************************************** > changed: [10.10.10.11] => (item=glusterd) > changed: [10.10.10.12] => (item=glusterd) > changed: [10.10.10.13] => (item=glusterd) > > PLAY RECAP > ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [gluster_servers] > ************************************************************************************* > > TASK [Create the brick dirs, skips if present] > ************************************************************* > changed: [10.10.10.11] => (item=/gluster/brick1/one) > changed: [10.10.10.12] => (item=/gluster/brick1/one) > changed: [10.10.10.13] => (item=/gluster/brick1/one) > > PLAY RECAP > ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [master] > ********************************************************************************************** > > TASK [Creates a Trusted Storage Pool] > ********************************************************************** > changed: [10.10.10.11] > > TASK [Pause for a few seconds] > ***************************************************************************** > Pausing for 5 seconds > (ctrl+C then 'C' = continue early, ctrl+C then 'A' = abort) > ok: [10.10.10.11] > > PLAY RECAP > ************************************************************************************************* > 10.10.10.11 : ok=2 changed=1 unreachable=0 failed=0 > > > PLAY [master] > ********************************************************************************************** > > TASK [Creates a volume] > ************************************************************************************ > changed: [10.10.10.11] > > PLAY RECAP > ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [master] > ********************************************************************************************** > > TASK [Starts a volume] > ************************************************************************************* > changed: [10.10.10.11] > > PLAY RECAP > ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [clients] > ********************************************************************************************* > > TASK [Create the dir to mount the volume, skips if present] > ************************************************ > changed: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > changed: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > changed: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > > PLAY RECAP > ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [clients] > ********************************************************************************************* > > TASK [Mount the volumes, if fstype is glusterfs] > *********************************************************** > changed: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > changed: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > changed: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > > PLAY RECAP > ************************************************************************************************* > 10.10.10.11 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.12 : ok=1 changed=1 unreachable=0 failed=0 > 10.10.10.13 : ok=1 changed=1 unreachable=0 failed=0 > > > PLAY [clients] > ********************************************************************************************* > > TASK [Gathering Facts] > ************************************************************************************* > ok: [10.10.10.11] > ok: [10.10.10.12] > ok: [10.10.10.13] > > TASK [Uncomment STATD_PORT for rpc.statd to listen on] > ***************************************************** > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > > TASK [Uncomment LOCKD_TCPPORT for rpc.lockd to listen on] > ************************************************** > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > > TASK [Uncomment LOCKD_UDPPORT for rpc.lockd to listen on] > ************************************************** > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > > TASK [Uncomment MOUNTD_PORT for rpc.mountd to listen on] > *************************************************** > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > > TASK [Restart nfs service (RHEL 6 only)] > ******************************************************************* > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > > TASK [Restart rpc-statd service] > *************************************************************************** > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > > TASK [Restart nfs-config service] > ************************************************************************** > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > > TASK [Restart nfs-mountd service] > ************************************************************************** > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > > TASK [Restart nfslock service (RHEL 6 & 7)] > **************************************************************** > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > > TASK [Mount the volumes if fstype is NFS] > ****************************************************************** > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > > PLAY RECAP > ************************************************************************************************* > 10.10.10.11 : ok=1 changed=0 unreachable=0 failed=0 > 10.10.10.12 : ok=1 changed=0 unreachable=0 failed=0 > 10.10.10.13 : ok=1 changed=0 unreachable=0 failed=0 > > > PLAY [clients] > ********************************************************************************************* > > TASK [Mount the volumes, if fstype is CIFS] > **************************************************************** > skipping: [10.10.10.12] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.11] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > skipping: [10.10.10.13] => (item={u'mountpoint': u'/mnt/ganesha_mnt', > u'fstype': u'fuse'}) > > PLAY RECAP > ************************************************************************************************* > 10.10.10.11 : ok=0 changed=0 unreachable=0 failed=0 > 10.10.10.12 : ok=0 changed=0 unreachable=0 failed=0 > 10.10.10.13 : ok=0 changed=0 unreachable=0 failed=0 > > > PLAY [master] > ********************************************************************************************** > > TASK [Delete lines from Ganesha export file] > *************************************************************** > > TASK [Add lines to Ganesha export file] > ******************************************************************** > > TASK [Update lines to Ganesha export file] > ***************************************************************** > > TASK [Add a block to Ganesha export file] > ****************************************************************** > > TASK [Refresh NFS-Ganesha config] > ************************************************************************** > fatal: [10.10.10.11]: FAILED! => {"changed": true, "cmd": > "/usr/libexec/ganesha/ganesha-ha.sh --refresh-config > \"/var/run/gluster/shared_storage/nfs-ganesha\" \"ganesha\"", "delta": > "0:00:00.227074", "end": "2017-05-09 16:54:45.442219", "failed": true, "rc": > 1, "start": "2017-05-09 16:54:45.215145", "stderr": "grep: > /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or > directory\ngrep: /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: > No such file or directory\nError: cluster is not currently running on this > node\ngrep: > /var/run/gluster/shared_storage/nfs-ganesha/exports/export.ganesha.conf: No > such file or directory", "stderr_lines": ["grep: > /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or > directory", "grep: > /var/run/gluster/shared_storage/nfs-ganesha/ganesha-ha.conf: No such file or > directory", "Error: cluster is not currently running on this node", "grep: > /var/run/gluster/shared_storage/nfs-ganesha/exports/exp or > t.ganesha.conf: No such file or directory"], "stdout": "Error: > refresh-config failed on localhost.", "stdout_lines": ["Error: refresh-config > failed on localhost."]} > to retry, use: --limit @/tmp/tmpacx8yi/ganesha-refresh-config.retry > > PLAY RECAP > ************************************************************************************************* > 10.10.10.11 : ok=0 changed=0 unreachable=0 failed=1 > > Ignoring errors... > You can view the generated configuration files inside /tmp/tmpacx8yi > [root@linked-clone-of-centos-linux ~]# > _______________________________________________ > Gluster-users mailing list > [email protected] > http://lists.gluster.org/mailman/listinfo/gluster-users > _______________________________________________ Gluster-users mailing list [email protected] http://lists.gluster.org/mailman/listinfo/gluster-users
