Thanks! After adding the workaround, I was able to complete the deployment.

On Fri, May 10, 2019 at 1:39 AM Parth Dhanjal <dpa...@redhat.com> wrote:

> Hey!
>
> oVirt 4.3.3 uses gluster-ansible-roles to deploy the storage.
> There are multiple checks during a deployment.
> This particular check which is failing is a part of the
> gluster-ansible-featues (
> https://github.com/gluster/gluster-ansible-features/tree/master/roles/gluster_hci
> )
>
> A simple workaround can be to skip the test, by editing the finally
> generated inventory file in the last step before deployment and adding 
> gluster_features_force_varlogsizecheck:
> false under the vars section of the file.
>
> Regards
> Parth Dhanjal
>
> On Fri, May 10, 2019 at 5:58 AM Edward Berger <edwber...@gmail.com> wrote:
>
>> I'm trying to bring up a single node hyperconverged with the current
>> node-ng ISO installation,
>> but it ends with this failure message.
>>
>> TASK [gluster.features/roles/gluster_hci : Check if /var/log has enough
>> disk space] ***
>> fatal: [br014.bridges.psc.edu]: FAILED! => {"changed": true, "cmd": "df
>> -m /var/log | awk '/[0-9]%/ {print $4}'", "delta": "0:00:00.008513", "end":
>> "2019-05-09 20:09:27.914400", "failed_when_result": true, "rc": 0, "start":
>> "2019-05-09 20:09:27.905887", "stderr": "", "stderr_lines": [], "stdout":
>> "7470", "stdout_lines": ["7470"]}
>>
>> I have what the installer created by default for /var/log, so I don't
>> know why its complaining.
>>
>> [root@br014 ~]# df -kh
>> Filesystem                                                      Size
>> Used Avail Use% Mounted on
>> /dev/mapper/onn_br014-ovirt--node--ng--4.3.3.1--0.20190417.0+1  3.5T
>> 2.1G  3.3T   1% /
>> devtmpfs                                                         63G
>> 0   63G   0% /dev
>> tmpfs                                                            63G
>> 4.0K   63G   1% /dev/shm
>> tmpfs                                                            63G
>> 18M   63G   1% /run
>> tmpfs                                                            63G
>> 0   63G   0% /sys/fs/cgroup
>> /dev/mapper/onn_br014-home                                      976M
>> 2.6M  907M   1% /home
>> /dev/mapper/onn_br014-tmp                                       976M
>> 2.8M  906M   1% /tmp
>> /dev/mapper/onn_br014-var                                        15G
>> 42M   14G   1% /var
>> /dev/sda2                                                       976M
>> 173M  737M  19% /boot
>> /dev/mapper/onn_br014-var_log                                   7.8G
>> 41M  7.3G   1% /var/log
>> /dev/mapper/onn_br014-var_log_audit                             2.0G
>> 7.6M  1.8G   1% /var/log/audit
>> /dev/mapper/onn_br014-var_crash                                 9.8G
>> 37M  9.2G   1% /var/crash
>> /dev/sda1                                                       200M
>> 12M  189M   6% /boot/efi
>> tmpfs                                                            13G
>> 0   13G   0% /run/user/1000
>> tmpfs                                                            13G
>> 0   13G   0% /run/user/0
>> /dev/mapper/gluster_vg_sdb-gluster_lv_engine                    3.7T
>> 33M  3.7T   1% /gluster_bricks/engine
>> /dev/mapper/gluster_vg_sdc-gluster_lv_data                      3.7T
>> 34M  3.7T   1% /gluster_bricks/data
>> /dev/mapper/gluster_vg_sdd-gluster_lv_vmstore                   3.7T
>> 34M  3.7T   1% /gluster_bricks/vmstore
>>
>> The machine had 4 4TB disks, so sda is the installation for oVirt
>> node-ng, the other 3 disks for the gluster volumes.
>>
>> root@br014 ~]# pvs
>>   PV         VG             Fmt  Attr PSize  PFree
>>   /dev/sda3  onn_br014      lvm2 a--  <3.64t 100.00g
>>   /dev/sdb   gluster_vg_sdb lvm2 a--  <3.64t <26.02g
>>   /dev/sdc   gluster_vg_sdc lvm2 a--  <3.64t      0
>>   /dev/sdd   gluster_vg_sdd lvm2 a--  <3.64t      0
>>
>> [root@br014 ~]# vgs
>>   VG             #PV #LV #SN Attr   VSize  VFree
>>   gluster_vg_sdb   1   1   0 wz--n- <3.64t <26.02g
>>   gluster_vg_sdc   1   2   0 wz--n- <3.64t      0
>>   gluster_vg_sdd   1   2   0 wz--n- <3.64t      0
>>   onn_br014        1  11   0 wz--n- <3.64t 100.00g
>>
>> [root@br014 ~]# lvs
>>   LV                                   VG             Attr       LSize
>> Pool                            Origin                             Data%
>> Meta%  Move Log Cpy%Sync Convert
>>   gluster_lv_engine                    gluster_vg_sdb -wi-ao----
>> 3.61t
>>
>>   gluster_lv_data                      gluster_vg_sdc Vwi-aot---  3.61t
>> gluster_thinpool_gluster_vg_sdc
>> 0.05
>>   gluster_thinpool_gluster_vg_sdc      gluster_vg_sdc twi-aot---
>> <3.61t
>> 0.05   0.13
>>   gluster_lv_vmstore                   gluster_vg_sdd Vwi-aot---  3.61t
>> gluster_thinpool_gluster_vg_sdd
>> 0.05
>>   gluster_thinpool_gluster_vg_sdd      gluster_vg_sdd twi-aot---
>> <3.61t
>> 0.05   0.13
>>   home                                 onn_br014      Vwi-aotz--  1.00g
>> pool00
>> 4.79
>>   ovirt-node-ng-4.3.3.1-0.20190417.0   onn_br014      Vwi---tz-k <3.51t
>> pool00
>> root
>>   ovirt-node-ng-4.3.3.1-0.20190417.0+1 onn_br014      Vwi-aotz-- <3.51t
>> pool00                          ovirt-node-ng-4.3.3.1-0.20190417.0
>> 0.13
>>   pool00                               onn_br014      twi-aotz--
>> 3.53t
>> 0.19   1.86
>>   root                                 onn_br014      Vri---tz-k <3.51t
>> pool00
>>
>>   swap                                 onn_br014      -wi-ao----
>> 4.00g
>>
>>   tmp                                  onn_br014      Vwi-aotz--  1.00g
>> pool00
>> 4.84
>>   var                                  onn_br014      Vwi-aotz-- 15.00g
>> pool00
>> 3.67
>>   var_crash                            onn_br014      Vwi-aotz-- 10.00g
>> pool00
>> 2.86
>>   var_log                              onn_br014      Vwi-aotz--  8.00g
>> pool00
>> 3.25
>>   var_log_audit                        onn_br014      Vwi-aotz--  2.00g
>> pool00
>> 4.86
>>
>>
>>
>> Here's the full deploy log from the UI.  Let me know if you need specific
>> logs.
>>
>>
>> PLAY [Setup backend]
>> ***********************************************************
>>
>> TASK [Gathering Facts]
>> *********************************************************
>> ok: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/firewall_config : Start firewalld if not
>> already started] ***
>> ok: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/firewall_config : check if required variables
>> are set] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/firewall_config : Open/Close firewalld ports]
>> ********
>> ok: [br014.bridges.psc.edu] => (item=2049/tcp)
>> ok: [br014.bridges.psc.edu] => (item=54321/tcp)
>> ok: [br014.bridges.psc.edu] => (item=5900/tcp)
>> ok: [br014.bridges.psc.edu] => (item=5900-6923/tcp)
>> ok: [br014.bridges.psc.edu] => (item=5666/tcp)
>> ok: [br014.bridges.psc.edu] => (item=16514/tcp)
>>
>> TASK [gluster.infra/roles/firewall_config : Add/Delete services to
>> firewalld rules] ***
>> ok: [br014.bridges.psc.edu] => (item=glusterfs)
>>
>> TASK [gluster.infra/roles/backend_setup : Gather facts to determine the
>> OS distribution] ***
>> ok: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for
>> debian systems.] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for
>> RHEL systems.] ***
>> ok: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Install python-yaml package for
>> Debian systems] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Initialize vdo_devs array]
>> ***********
>> ok: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Record VDO devices (if any)]
>> *********
>> skipping: [br014.bridges.psc.edu] => (item={u'vgname':
>> u'gluster_vg_sdb', u'pvname': u'/dev/sdb'})
>> skipping: [br014.bridges.psc.edu] => (item={u'vgname':
>> u'gluster_vg_sdc', u'pvname': u'/dev/sdc'})
>> skipping: [br014.bridges.psc.edu] => (item={u'vgname':
>> u'gluster_vg_sdd', u'pvname': u'/dev/sdd'})
>>
>> TASK [gluster.infra/roles/backend_setup : Enable and start vdo service]
>> ********
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Create VDO with specified size]
>> ******
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Check if valid disktype is
>> provided] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Set PV data alignment for JBOD]
>> ******
>> ok: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Set PV data alignment for RAID]
>> ******
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Set VG physical extent size for
>> RAID] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Create volume groups]
>> ****************
>> ok: [br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdb',
>> u'pvname': u'/dev/sdb'})
>> ok: [br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdc',
>> u'pvname': u'/dev/sdc'})
>> ok: [br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdd',
>> u'pvname': u'/dev/sdd'})
>>
>> TASK [gluster.infra/roles/backend_setup : Create thick logical volume]
>> *********
>> ok: [br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_engine',
>> u'vgname': u'gluster_vg_sdb', u'size': u'3700G'})
>>
>> TASK [gluster.infra/roles/backend_setup : Calculate chunksize for
>> RAID6/RAID10/RAID5] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Set chunksize for JBOD]
>> **************
>> ok: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Create a LV thinpool]
>> ****************
>> ok: [br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdc',
>> u'thinpoolname': u'gluster_thinpool_gluster_vg_sdc', u'poolmetadatasize':
>> u'16G'})
>> ok: [br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdd',
>> u'thinpoolname': u'gluster_thinpool_gluster_vg_sdd', u'poolmetadatasize':
>> u'16G'})
>>
>> TASK [gluster.infra/roles/backend_setup : Create thin logical volume]
>> **********
>> ok: [br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_data',
>> u'vgname': u'gluster_vg_sdc', u'thinpool':
>> u'gluster_thinpool_gluster_vg_sdc', u'lvsize': u'3700G'})
>> ok: [br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_vmstore',
>> u'vgname': u'gluster_vg_sdd', u'thinpool':
>> u'gluster_thinpool_gluster_vg_sdd', u'lvsize': u'3700G'})
>>
>> TASK [gluster.infra/roles/backend_setup : Extend volume group]
>> *****************
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Change attributes of LV]
>> *************
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Create LV for cache]
>> *****************
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Create metadata LV for cache]
>> ********
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Convert logical volume to a
>> cache pool LV] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Convert logical volume to a
>> cache pool LV without cachemetalvname] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Convert an existing logical
>> volume to a cache LV] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Set XFS options for JBOD]
>> ************
>> ok: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Set XFS options for RAID
>> devices] ****
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Create filesystem on thin
>> logical vols] ***
>> ok: [br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_data',
>> u'vgname': u'gluster_vg_sdc', u'thinpool':
>> u'gluster_thinpool_gluster_vg_sdc', u'lvsize': u'3700G'})
>> ok: [br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_vmstore',
>> u'vgname': u'gluster_vg_sdd', u'thinpool':
>> u'gluster_thinpool_gluster_vg_sdd', u'lvsize': u'3700G'})
>>
>> TASK [gluster.infra/roles/backend_setup : Create filesystem on thick
>> logical vols] ***
>> ok: [br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_engine',
>> u'vgname': u'gluster_vg_sdb', u'size': u'3700G'})
>>
>> TASK [gluster.infra/roles/backend_setup : Create mount directories if not
>> already present] ***
>> ok: [br014.bridges.psc.edu] => (item={u'path':
>> u'/gluster_bricks/engine', u'vgname': u'gluster_vg_sdb', u'lvname':
>> u'gluster_lv_engine'})
>> ok: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/data',
>> u'vgname': u'gluster_vg_sdc', u'lvname': u'gluster_lv_data'})
>> ok: [br014.bridges.psc.edu] => (item={u'path':
>> u'/gluster_bricks/vmstore', u'vgname': u'gluster_vg_sdd', u'lvname':
>> u'gluster_lv_vmstore'})
>>
>> TASK [gluster.infra/roles/backend_setup : Set mount options for VDO]
>> ***********
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_setup : Mount the vdo devices (If any)]
>> ******
>> skipping: [br014.bridges.psc.edu] => (item={u'path':
>> u'/gluster_bricks/engine', u'vgname': u'gluster_vg_sdb', u'lvname':
>> u'gluster_lv_engine'})
>> skipping: [br014.bridges.psc.edu] => (item={u'path':
>> u'/gluster_bricks/data', u'vgname': u'gluster_vg_sdc', u'lvname':
>> u'gluster_lv_data'})
>> skipping: [br014.bridges.psc.edu] => (item={u'path':
>> u'/gluster_bricks/vmstore', u'vgname': u'gluster_vg_sdd', u'lvname':
>> u'gluster_lv_vmstore'})
>>
>> TASK [gluster.infra/roles/backend_setup : Mount the devices]
>> *******************
>> ok: [br014.bridges.psc.edu] => (item={u'path':
>> u'/gluster_bricks/engine', u'vgname': u'gluster_vg_sdb', u'lvname':
>> u'gluster_lv_engine'})
>> ok: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/data',
>> u'vgname': u'gluster_vg_sdc', u'lvname': u'gluster_lv_data'})
>> ok: [br014.bridges.psc.edu] => (item={u'path':
>> u'/gluster_bricks/vmstore', u'vgname': u'gluster_vg_sdd', u'lvname':
>> u'gluster_lv_vmstore'})
>>
>> TASK [gluster.infra/roles/backend_setup : Set Gluster specific SeLinux
>> context on the bricks] ***
>> ok: [br014.bridges.psc.edu] => (item={u'path':
>> u'/gluster_bricks/engine', u'vgname': u'gluster_vg_sdb', u'lvname':
>> u'gluster_lv_engine'})
>> ok: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/data',
>> u'vgname': u'gluster_vg_sdc', u'lvname': u'gluster_lv_data'})
>> ok: [br014.bridges.psc.edu] => (item={u'path':
>> u'/gluster_bricks/vmstore', u'vgname': u'gluster_vg_sdd', u'lvname':
>> u'gluster_lv_vmstore'})
>>
>> TASK [gluster.infra/roles/backend_setup : restore file(s) default SELinux
>> security contexts] ***
>> changed: [br014.bridges.psc.edu] => (item={u'path':
>> u'/gluster_bricks/engine', u'vgname': u'gluster_vg_sdb', u'lvname':
>> u'gluster_lv_engine'})
>> changed: [br014.bridges.psc.edu] => (item={u'path':
>> u'/gluster_bricks/data', u'vgname': u'gluster_vg_sdc', u'lvname':
>> u'gluster_lv_data'})
>> changed: [br014.bridges.psc.edu] => (item={u'path':
>> u'/gluster_bricks/vmstore', u'vgname': u'gluster_vg_sdd', u'lvname':
>> u'gluster_lv_vmstore'})
>>
>> TASK [gluster.infra/roles/backend_reset : unmount the directories (if
>> mounted)] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_reset : Delete volume groups]
>> ****************
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.infra/roles/backend_reset : Remove VDO devices]
>> ******************
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Create temporary storage
>> directory] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Get the name of the directory
>> created] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : check if
>> gluster_features_ganesha_clusternodes is set] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Define service port]
>> ****************
>> skipping: [br014.bridges.psc.edu] => (item=^#(STATD_PORT=.*))
>> skipping: [br014.bridges.psc.edu] => (item=^#(LOCKD_TCPPORT=.*))
>> skipping: [br014.bridges.psc.edu] => (item=^#(LOCKD_UDPPORT=.*))
>>
>> TASK [gluster.features/roles/nfs_ganesha : Check packages installed, if
>> not install] ***
>> skipping: [br014.bridges.psc.edu] => (item=glusterfs-ganesha)
>> skipping: [br014.bridges.psc.edu] => (item=nfs-ganesha)
>> skipping: [br014.bridges.psc.edu] => (item=corosync)
>> skipping: [br014.bridges.psc.edu] => (item=pacemaker)
>> skipping: [br014.bridges.psc.edu] => (item=libntirpc)
>> skipping: [br014.bridges.psc.edu] => (item=pcs)
>>
>> TASK [gluster.features/roles/nfs_ganesha : Restart services]
>> *******************
>> skipping: [br014.bridges.psc.edu] => (item=nfslock)
>> skipping: [br014.bridges.psc.edu] => (item=nfs-config)
>> skipping: [br014.bridges.psc.edu] => (item=rpc-statd)
>>
>> TASK [gluster.features/roles/nfs_ganesha : Stop services]
>> **********************
>> skipping: [br014.bridges.psc.edu] => (item=nfs-server)
>>
>> TASK [gluster.features/roles/nfs_ganesha : Disable service]
>> ********************
>> skipping: [br014.bridges.psc.edu] => (item=nfs-server)
>>
>> TASK [gluster.features/roles/nfs_ganesha : Enable services]
>> ********************
>> skipping: [br014.bridges.psc.edu] => (item=glusterfssharedstorage)
>> skipping: [br014.bridges.psc.edu] => (item=nfs-ganesha)
>> skipping: [br014.bridges.psc.edu] => (item=network)
>> skipping: [br014.bridges.psc.edu] => (item=pcsd)
>> skipping: [br014.bridges.psc.edu] => (item=pacemaker)
>>
>> TASK [gluster.features/roles/nfs_ganesha : Start services]
>> *********************
>> skipping: [br014.bridges.psc.edu] => (item=network)
>> skipping: [br014.bridges.psc.edu] => (item=pcsd)
>>
>> TASK [gluster.features/roles/nfs_ganesha : Create a user hacluster if not
>> already present] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Set the password for
>> hacluster] *****
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Set the hacluster user the
>> same password on new nodes] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Pcs cluster authenticate the
>> hacluster on new nodes] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Pause for a few seconds after
>> pcs auth] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Set gluster_use_execmem flag
>> on and keep it persistent] ***
>> skipping: [br014.bridges.psc.edu] => (item=gluster_use_execmem)
>> skipping: [br014.bridges.psc.edu] => (item=ganesha_use_fusefs)
>>
>> TASK [gluster.features/roles/nfs_ganesha : check if
>> gluster_features_ganesha_masternode is set] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Copy the ssh keys to the local
>> machine] ***
>> skipping: [br014.bridges.psc.edu] => (item=secret.pem.pub)
>> skipping: [br014.bridges.psc.edu] => (item=secret.pem)
>>
>> TASK [gluster.features/roles/nfs_ganesha : check if
>> gluster_features_ganesha_newnodes_vip is set] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Copy the public key to remote
>> nodes] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Copy the private key to remote
>> node] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Deploy the pubkey on all
>> nodes] *****
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Make the volume a gluster
>> shared volume] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Generate ssh key in one of the
>> nodes in HA cluster] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Copy the ssh keys to the local
>> machine] ***
>> skipping: [br014.bridges.psc.edu] => (item=secret.pem.pub)
>> skipping: [br014.bridges.psc.edu] => (item=secret.pem)
>>
>> TASK [gluster.features/roles/nfs_ganesha : Create configuration directory
>> for nfs_ganesha] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Copy ganesha.conf to config
>> directory on shared volume] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Create ganesha-ha.conf file]
>> ********
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Enable NFS Ganesha]
>> *****************
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Pause for 30 seconds (takes a
>> while to enable NFS Ganesha)] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Check NFS Ganesha status]
>> ***********
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Report NFS Ganesha status]
>> **********
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Report NFS Ganesha status (If
>> any errors)] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : check if
>> gluster_features_ganesha_volume is set] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Export the NFS Ganesha volume]
>> ******
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Copy the public key to remote
>> nodes] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Copy the private key to remote
>> node] ***
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Deploy the pubkey on all
>> nodes] *****
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Adds a node to the cluster]
>> *********
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Report ganesha add-node
>> status] *****
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/nfs_ganesha : Delete the temporary
>> directory] *****
>> skipping: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/gluster_hci : Check if packages are
>> installed, if not install] ***
>> ok: [br014.bridges.psc.edu] => (item=vdsm)
>> ok: [br014.bridges.psc.edu] => (item=vdsm-gluster)
>> ok: [br014.bridges.psc.edu] => (item=ovirt-host)
>> ok: [br014.bridges.psc.edu] => (item=screen)
>>
>> TASK [gluster.features/roles/gluster_hci : Enable and start glusterd and
>> chronyd] ***
>> ok: [br014.bridges.psc.edu] => (item=chronyd)
>> ok: [br014.bridges.psc.edu] => (item=glusterd)
>> ok: [br014.bridges.psc.edu] => (item=firewalld)
>>
>> TASK [gluster.features/roles/gluster_hci : Add user qemu to gluster
>> group] *****
>> ok: [br014.bridges.psc.edu]
>>
>> TASK [gluster.features/roles/gluster_hci : Disable the hook scripts]
>> ***********
>> changed: [br014.bridges.psc.edu] =>
>> (item=/var/lib/glusterd/hooks/1/set/post/S30samba-set.sh)
>> changed: [br014.bridges.psc.edu] =>
>> (item=/var/lib/glusterd/hooks/1/start/post/S30samba-start.sh)
>> changed: [br014.bridges.psc.edu] =>
>> (item=/var/lib/glusterd/hooks/1/stop/pre/S30samba-stop.sh)
>> changed: [br014.bridges.psc.edu] =>
>> (item=/var/lib/glusterd/hooks/1/reset/post/S31ganesha-reset.sh)
>> changed: [br014.bridges.psc.edu] =>
>> (item=/var/lib/glusterd/hooks/1//start/post/S31ganesha-start.sh)
>> changed: [br014.bridges.psc.edu] =>
>> (item=/var/lib/glusterd/hooks/1/start/post/S29CTDBsetup.sh)
>> changed: [br014.bridges.psc.edu] =>
>> (item=/var/lib/glusterd/hooks/1/stop/pre/S29CTDB-teardown.sh)
>>
>> TASK [gluster.features/roles/gluster_hci : Check if valid FQDN is
>> provided] ****
>> changed: [br014.bridges.psc.edu -> localhost] => (item=
>> br014.bridges.psc.edu)
>>
>> TASK [gluster.features/roles/gluster_hci : Check if /var/log has enough
>> disk space] ***
>> fatal: [br014.bridges.psc.edu]: FAILED! => {"changed": true, "cmd": "df
>> -m /var/log | awk '/[0-9]%/ {print $4}'", "delta": "0:00:00.008513", "end":
>> "2019-05-09 20:09:27.914400", "failed_when_result": true, "rc": 0, "start":
>> "2019-05-09 20:09:27.905887", "stderr": "", "stderr_lines": [], "stdout":
>> "7470", "stdout_lines": ["7470"]}
>>
>> NO MORE HOSTS LEFT
>> *************************************************************
>>
>> NO MORE HOSTS LEFT
>> *************************************************************
>>     to retry, use: --limit
>> @/usr/share/cockpit/ovirt-dashboard/ansible/hc_wizard.retry
>>
>> PLAY RECAP
>> *********************************************************************
>> br014.bridges.psc.edu      : ok=25   changed=3    unreachable=0
>> failed=1
>>
>>
>> _______________________________________________
>> Users mailing list -- users@ovirt.org
>> To unsubscribe send an email to users-le...@ovirt.org
>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>> oVirt Code of Conduct:
>> https://www.ovirt.org/community/about/community-guidelines/
>> List Archives:
>> https://lists.ovirt.org/archives/list/users@ovirt.org/message/P2WRSR6U67BOQYKMEJFNMSRJFRMJFECN/
>>
>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/KILGRYYIXS7N5ZFUENNPWELJCIHRSF7G/

Reply via email to