Hey! oVirt 4.3.3 uses gluster-ansible-roles to deploy the storage. There are multiple checks during a deployment. This particular check which is failing is a part of the gluster-ansible-featues ( https://github.com/gluster/gluster-ansible-features/tree/master/roles/gluster_hci )
A simple workaround can be to skip the test, by editing the finally generated inventory file in the last step before deployment and adding gluster_features_force_varlogsizecheck: false under the vars section of the file. Regards Parth Dhanjal On Fri, May 10, 2019 at 5:58 AM Edward Berger <edwber...@gmail.com> wrote: > I'm trying to bring up a single node hyperconverged with the current > node-ng ISO installation, > but it ends with this failure message. > > TASK [gluster.features/roles/gluster_hci : Check if /var/log has enough > disk space] *** > fatal: [br014.bridges.psc.edu]: FAILED! => {"changed": true, "cmd": "df > -m /var/log | awk '/[0-9]%/ {print $4}'", "delta": "0:00:00.008513", "end": > "2019-05-09 20:09:27.914400", "failed_when_result": true, "rc": 0, "start": > "2019-05-09 20:09:27.905887", "stderr": "", "stderr_lines": [], "stdout": > "7470", "stdout_lines": ["7470"]} > > I have what the installer created by default for /var/log, so I don't know > why its complaining. > > [root@br014 ~]# df -kh > Filesystem Size Used > Avail Use% Mounted on > /dev/mapper/onn_br014-ovirt--node--ng--4.3.3.1--0.20190417.0+1 3.5T > 2.1G 3.3T 1% / > devtmpfs 63G > 0 63G 0% /dev > tmpfs 63G > 4.0K 63G 1% /dev/shm > tmpfs 63G > 18M 63G 1% /run > tmpfs 63G > 0 63G 0% /sys/fs/cgroup > /dev/mapper/onn_br014-home 976M > 2.6M 907M 1% /home > /dev/mapper/onn_br014-tmp 976M > 2.8M 906M 1% /tmp > /dev/mapper/onn_br014-var 15G > 42M 14G 1% /var > /dev/sda2 976M > 173M 737M 19% /boot > /dev/mapper/onn_br014-var_log 7.8G > 41M 7.3G 1% /var/log > /dev/mapper/onn_br014-var_log_audit 2.0G > 7.6M 1.8G 1% /var/log/audit > /dev/mapper/onn_br014-var_crash 9.8G > 37M 9.2G 1% /var/crash > /dev/sda1 200M > 12M 189M 6% /boot/efi > tmpfs 13G > 0 13G 0% /run/user/1000 > tmpfs 13G > 0 13G 0% /run/user/0 > /dev/mapper/gluster_vg_sdb-gluster_lv_engine 3.7T > 33M 3.7T 1% /gluster_bricks/engine > /dev/mapper/gluster_vg_sdc-gluster_lv_data 3.7T > 34M 3.7T 1% /gluster_bricks/data > /dev/mapper/gluster_vg_sdd-gluster_lv_vmstore 3.7T > 34M 3.7T 1% /gluster_bricks/vmstore > > The machine had 4 4TB disks, so sda is the installation for oVirt node-ng, > the other 3 disks for the gluster volumes. > > root@br014 ~]# pvs > PV VG Fmt Attr PSize PFree > /dev/sda3 onn_br014 lvm2 a-- <3.64t 100.00g > /dev/sdb gluster_vg_sdb lvm2 a-- <3.64t <26.02g > /dev/sdc gluster_vg_sdc lvm2 a-- <3.64t 0 > /dev/sdd gluster_vg_sdd lvm2 a-- <3.64t 0 > > [root@br014 ~]# vgs > VG #PV #LV #SN Attr VSize VFree > gluster_vg_sdb 1 1 0 wz--n- <3.64t <26.02g > gluster_vg_sdc 1 2 0 wz--n- <3.64t 0 > gluster_vg_sdd 1 2 0 wz--n- <3.64t 0 > onn_br014 1 11 0 wz--n- <3.64t 100.00g > > [root@br014 ~]# lvs > LV VG Attr LSize > Pool Origin Data% > Meta% Move Log Cpy%Sync Convert > gluster_lv_engine gluster_vg_sdb -wi-ao---- > 3.61t > > gluster_lv_data gluster_vg_sdc Vwi-aot--- 3.61t > gluster_thinpool_gluster_vg_sdc > 0.05 > gluster_thinpool_gluster_vg_sdc gluster_vg_sdc twi-aot--- > <3.61t > 0.05 0.13 > gluster_lv_vmstore gluster_vg_sdd Vwi-aot--- 3.61t > gluster_thinpool_gluster_vg_sdd > 0.05 > gluster_thinpool_gluster_vg_sdd gluster_vg_sdd twi-aot--- > <3.61t > 0.05 0.13 > home onn_br014 Vwi-aotz-- 1.00g > pool00 > 4.79 > ovirt-node-ng-4.3.3.1-0.20190417.0 onn_br014 Vwi---tz-k <3.51t > pool00 > root > ovirt-node-ng-4.3.3.1-0.20190417.0+1 onn_br014 Vwi-aotz-- <3.51t > pool00 ovirt-node-ng-4.3.3.1-0.20190417.0 > 0.13 > pool00 onn_br014 twi-aotz-- > 3.53t > 0.19 1.86 > root onn_br014 Vri---tz-k <3.51t > pool00 > > swap onn_br014 -wi-ao---- > 4.00g > > tmp onn_br014 Vwi-aotz-- 1.00g > pool00 > 4.84 > var onn_br014 Vwi-aotz-- 15.00g > pool00 > 3.67 > var_crash onn_br014 Vwi-aotz-- 10.00g > pool00 > 2.86 > var_log onn_br014 Vwi-aotz-- 8.00g > pool00 > 3.25 > var_log_audit onn_br014 Vwi-aotz-- 2.00g > pool00 > 4.86 > > > > Here's the full deploy log from the UI. Let me know if you need specific > logs. > > > PLAY [Setup backend] > *********************************************************** > > TASK [Gathering Facts] > ********************************************************* > ok: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/firewall_config : Start firewalld if not already > started] *** > ok: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/firewall_config : check if required variables > are set] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/firewall_config : Open/Close firewalld ports] > ******** > ok: [br014.bridges.psc.edu] => (item=2049/tcp) > ok: [br014.bridges.psc.edu] => (item=54321/tcp) > ok: [br014.bridges.psc.edu] => (item=5900/tcp) > ok: [br014.bridges.psc.edu] => (item=5900-6923/tcp) > ok: [br014.bridges.psc.edu] => (item=5666/tcp) > ok: [br014.bridges.psc.edu] => (item=16514/tcp) > > TASK [gluster.infra/roles/firewall_config : Add/Delete services to > firewalld rules] *** > ok: [br014.bridges.psc.edu] => (item=glusterfs) > > TASK [gluster.infra/roles/backend_setup : Gather facts to determine the OS > distribution] *** > ok: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for > debian systems.] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for > RHEL systems.] *** > ok: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Install python-yaml package for > Debian systems] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Initialize vdo_devs array] > *********** > ok: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Record VDO devices (if any)] > ********* > skipping: [br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdb', > u'pvname': u'/dev/sdb'}) > skipping: [br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdc', > u'pvname': u'/dev/sdc'}) > skipping: [br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdd', > u'pvname': u'/dev/sdd'}) > > TASK [gluster.infra/roles/backend_setup : Enable and start vdo service] > ******** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Create VDO with specified size] > ****** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Check if valid disktype is > provided] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Set PV data alignment for JBOD] > ****** > ok: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Set PV data alignment for RAID] > ****** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Set VG physical extent size for > RAID] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Create volume groups] > **************** > ok: [br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdb', > u'pvname': u'/dev/sdb'}) > ok: [br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdc', > u'pvname': u'/dev/sdc'}) > ok: [br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdd', > u'pvname': u'/dev/sdd'}) > > TASK [gluster.infra/roles/backend_setup : Create thick logical volume] > ********* > ok: [br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_engine', > u'vgname': u'gluster_vg_sdb', u'size': u'3700G'}) > > TASK [gluster.infra/roles/backend_setup : Calculate chunksize for > RAID6/RAID10/RAID5] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Set chunksize for JBOD] > ************** > ok: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Create a LV thinpool] > **************** > ok: [br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdc', > u'thinpoolname': u'gluster_thinpool_gluster_vg_sdc', u'poolmetadatasize': > u'16G'}) > ok: [br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdd', > u'thinpoolname': u'gluster_thinpool_gluster_vg_sdd', u'poolmetadatasize': > u'16G'}) > > TASK [gluster.infra/roles/backend_setup : Create thin logical volume] > ********** > ok: [br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_data', > u'vgname': u'gluster_vg_sdc', u'thinpool': > u'gluster_thinpool_gluster_vg_sdc', u'lvsize': u'3700G'}) > ok: [br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_vmstore', > u'vgname': u'gluster_vg_sdd', u'thinpool': > u'gluster_thinpool_gluster_vg_sdd', u'lvsize': u'3700G'}) > > TASK [gluster.infra/roles/backend_setup : Extend volume group] > ***************** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Change attributes of LV] > ************* > skipping: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Create LV for cache] > ***************** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Create metadata LV for cache] > ******** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Convert logical volume to a > cache pool LV] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Convert logical volume to a > cache pool LV without cachemetalvname] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Convert an existing logical > volume to a cache LV] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Set XFS options for JBOD] > ************ > ok: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Set XFS options for RAID > devices] **** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Create filesystem on thin > logical vols] *** > ok: [br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_data', > u'vgname': u'gluster_vg_sdc', u'thinpool': > u'gluster_thinpool_gluster_vg_sdc', u'lvsize': u'3700G'}) > ok: [br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_vmstore', > u'vgname': u'gluster_vg_sdd', u'thinpool': > u'gluster_thinpool_gluster_vg_sdd', u'lvsize': u'3700G'}) > > TASK [gluster.infra/roles/backend_setup : Create filesystem on thick > logical vols] *** > ok: [br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_engine', > u'vgname': u'gluster_vg_sdb', u'size': u'3700G'}) > > TASK [gluster.infra/roles/backend_setup : Create mount directories if not > already present] *** > ok: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/engine', > u'vgname': u'gluster_vg_sdb', u'lvname': u'gluster_lv_engine'}) > ok: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/data', > u'vgname': u'gluster_vg_sdc', u'lvname': u'gluster_lv_data'}) > ok: [br014.bridges.psc.edu] => (item={u'path': > u'/gluster_bricks/vmstore', u'vgname': u'gluster_vg_sdd', u'lvname': > u'gluster_lv_vmstore'}) > > TASK [gluster.infra/roles/backend_setup : Set mount options for VDO] > *********** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_setup : Mount the vdo devices (If any)] > ****** > skipping: [br014.bridges.psc.edu] => (item={u'path': > u'/gluster_bricks/engine', u'vgname': u'gluster_vg_sdb', u'lvname': > u'gluster_lv_engine'}) > skipping: [br014.bridges.psc.edu] => (item={u'path': > u'/gluster_bricks/data', u'vgname': u'gluster_vg_sdc', u'lvname': > u'gluster_lv_data'}) > skipping: [br014.bridges.psc.edu] => (item={u'path': > u'/gluster_bricks/vmstore', u'vgname': u'gluster_vg_sdd', u'lvname': > u'gluster_lv_vmstore'}) > > TASK [gluster.infra/roles/backend_setup : Mount the devices] > ******************* > ok: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/engine', > u'vgname': u'gluster_vg_sdb', u'lvname': u'gluster_lv_engine'}) > ok: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/data', > u'vgname': u'gluster_vg_sdc', u'lvname': u'gluster_lv_data'}) > ok: [br014.bridges.psc.edu] => (item={u'path': > u'/gluster_bricks/vmstore', u'vgname': u'gluster_vg_sdd', u'lvname': > u'gluster_lv_vmstore'}) > > TASK [gluster.infra/roles/backend_setup : Set Gluster specific SeLinux > context on the bricks] *** > ok: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/engine', > u'vgname': u'gluster_vg_sdb', u'lvname': u'gluster_lv_engine'}) > ok: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/data', > u'vgname': u'gluster_vg_sdc', u'lvname': u'gluster_lv_data'}) > ok: [br014.bridges.psc.edu] => (item={u'path': > u'/gluster_bricks/vmstore', u'vgname': u'gluster_vg_sdd', u'lvname': > u'gluster_lv_vmstore'}) > > TASK [gluster.infra/roles/backend_setup : restore file(s) default SELinux > security contexts] *** > changed: [br014.bridges.psc.edu] => (item={u'path': > u'/gluster_bricks/engine', u'vgname': u'gluster_vg_sdb', u'lvname': > u'gluster_lv_engine'}) > changed: [br014.bridges.psc.edu] => (item={u'path': > u'/gluster_bricks/data', u'vgname': u'gluster_vg_sdc', u'lvname': > u'gluster_lv_data'}) > changed: [br014.bridges.psc.edu] => (item={u'path': > u'/gluster_bricks/vmstore', u'vgname': u'gluster_vg_sdd', u'lvname': > u'gluster_lv_vmstore'}) > > TASK [gluster.infra/roles/backend_reset : unmount the directories (if > mounted)] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_reset : Delete volume groups] > **************** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.infra/roles/backend_reset : Remove VDO devices] > ****************** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Create temporary storage > directory] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Get the name of the directory > created] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : check if > gluster_features_ganesha_clusternodes is set] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Define service port] > **************** > skipping: [br014.bridges.psc.edu] => (item=^#(STATD_PORT=.*)) > skipping: [br014.bridges.psc.edu] => (item=^#(LOCKD_TCPPORT=.*)) > skipping: [br014.bridges.psc.edu] => (item=^#(LOCKD_UDPPORT=.*)) > > TASK [gluster.features/roles/nfs_ganesha : Check packages installed, if > not install] *** > skipping: [br014.bridges.psc.edu] => (item=glusterfs-ganesha) > skipping: [br014.bridges.psc.edu] => (item=nfs-ganesha) > skipping: [br014.bridges.psc.edu] => (item=corosync) > skipping: [br014.bridges.psc.edu] => (item=pacemaker) > skipping: [br014.bridges.psc.edu] => (item=libntirpc) > skipping: [br014.bridges.psc.edu] => (item=pcs) > > TASK [gluster.features/roles/nfs_ganesha : Restart services] > ******************* > skipping: [br014.bridges.psc.edu] => (item=nfslock) > skipping: [br014.bridges.psc.edu] => (item=nfs-config) > skipping: [br014.bridges.psc.edu] => (item=rpc-statd) > > TASK [gluster.features/roles/nfs_ganesha : Stop services] > ********************** > skipping: [br014.bridges.psc.edu] => (item=nfs-server) > > TASK [gluster.features/roles/nfs_ganesha : Disable service] > ******************** > skipping: [br014.bridges.psc.edu] => (item=nfs-server) > > TASK [gluster.features/roles/nfs_ganesha : Enable services] > ******************** > skipping: [br014.bridges.psc.edu] => (item=glusterfssharedstorage) > skipping: [br014.bridges.psc.edu] => (item=nfs-ganesha) > skipping: [br014.bridges.psc.edu] => (item=network) > skipping: [br014.bridges.psc.edu] => (item=pcsd) > skipping: [br014.bridges.psc.edu] => (item=pacemaker) > > TASK [gluster.features/roles/nfs_ganesha : Start services] > ********************* > skipping: [br014.bridges.psc.edu] => (item=network) > skipping: [br014.bridges.psc.edu] => (item=pcsd) > > TASK [gluster.features/roles/nfs_ganesha : Create a user hacluster if not > already present] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Set the password for hacluster] > ***** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Set the hacluster user the same > password on new nodes] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Pcs cluster authenticate the > hacluster on new nodes] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Pause for a few seconds after > pcs auth] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Set gluster_use_execmem flag on > and keep it persistent] *** > skipping: [br014.bridges.psc.edu] => (item=gluster_use_execmem) > skipping: [br014.bridges.psc.edu] => (item=ganesha_use_fusefs) > > TASK [gluster.features/roles/nfs_ganesha : check if > gluster_features_ganesha_masternode is set] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Copy the ssh keys to the local > machine] *** > skipping: [br014.bridges.psc.edu] => (item=secret.pem.pub) > skipping: [br014.bridges.psc.edu] => (item=secret.pem) > > TASK [gluster.features/roles/nfs_ganesha : check if > gluster_features_ganesha_newnodes_vip is set] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Copy the public key to remote > nodes] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Copy the private key to remote > node] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Deploy the pubkey on all nodes] > ***** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Make the volume a gluster > shared volume] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Generate ssh key in one of the > nodes in HA cluster] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Copy the ssh keys to the local > machine] *** > skipping: [br014.bridges.psc.edu] => (item=secret.pem.pub) > skipping: [br014.bridges.psc.edu] => (item=secret.pem) > > TASK [gluster.features/roles/nfs_ganesha : Create configuration directory > for nfs_ganesha] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Copy ganesha.conf to config > directory on shared volume] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Create ganesha-ha.conf file] > ******** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Enable NFS Ganesha] > ***************** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Pause for 30 seconds (takes a > while to enable NFS Ganesha)] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Check NFS Ganesha status] > *********** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Report NFS Ganesha status] > ********** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Report NFS Ganesha status (If > any errors)] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : check if > gluster_features_ganesha_volume is set] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Export the NFS Ganesha volume] > ****** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Copy the public key to remote > nodes] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Copy the private key to remote > node] *** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Deploy the pubkey on all nodes] > ***** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Adds a node to the cluster] > ********* > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Report ganesha add-node status] > ***** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/nfs_ganesha : Delete the temporary directory] > ***** > skipping: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/gluster_hci : Check if packages are > installed, if not install] *** > ok: [br014.bridges.psc.edu] => (item=vdsm) > ok: [br014.bridges.psc.edu] => (item=vdsm-gluster) > ok: [br014.bridges.psc.edu] => (item=ovirt-host) > ok: [br014.bridges.psc.edu] => (item=screen) > > TASK [gluster.features/roles/gluster_hci : Enable and start glusterd and > chronyd] *** > ok: [br014.bridges.psc.edu] => (item=chronyd) > ok: [br014.bridges.psc.edu] => (item=glusterd) > ok: [br014.bridges.psc.edu] => (item=firewalld) > > TASK [gluster.features/roles/gluster_hci : Add user qemu to gluster group] > ***** > ok: [br014.bridges.psc.edu] > > TASK [gluster.features/roles/gluster_hci : Disable the hook scripts] > *********** > changed: [br014.bridges.psc.edu] => > (item=/var/lib/glusterd/hooks/1/set/post/S30samba-set.sh) > changed: [br014.bridges.psc.edu] => > (item=/var/lib/glusterd/hooks/1/start/post/S30samba-start.sh) > changed: [br014.bridges.psc.edu] => > (item=/var/lib/glusterd/hooks/1/stop/pre/S30samba-stop.sh) > changed: [br014.bridges.psc.edu] => > (item=/var/lib/glusterd/hooks/1/reset/post/S31ganesha-reset.sh) > changed: [br014.bridges.psc.edu] => > (item=/var/lib/glusterd/hooks/1//start/post/S31ganesha-start.sh) > changed: [br014.bridges.psc.edu] => > (item=/var/lib/glusterd/hooks/1/start/post/S29CTDBsetup.sh) > changed: [br014.bridges.psc.edu] => > (item=/var/lib/glusterd/hooks/1/stop/pre/S29CTDB-teardown.sh) > > TASK [gluster.features/roles/gluster_hci : Check if valid FQDN is > provided] **** > changed: [br014.bridges.psc.edu -> localhost] => (item= > br014.bridges.psc.edu) > > TASK [gluster.features/roles/gluster_hci : Check if /var/log has enough > disk space] *** > fatal: [br014.bridges.psc.edu]: FAILED! => {"changed": true, "cmd": "df > -m /var/log | awk '/[0-9]%/ {print $4}'", "delta": "0:00:00.008513", "end": > "2019-05-09 20:09:27.914400", "failed_when_result": true, "rc": 0, "start": > "2019-05-09 20:09:27.905887", "stderr": "", "stderr_lines": [], "stdout": > "7470", "stdout_lines": ["7470"]} > > NO MORE HOSTS LEFT > ************************************************************* > > NO MORE HOSTS LEFT > ************************************************************* > to retry, use: --limit > @/usr/share/cockpit/ovirt-dashboard/ansible/hc_wizard.retry > > PLAY RECAP > ********************************************************************* > br014.bridges.psc.edu : ok=25 changed=3 unreachable=0 > failed=1 > > > _______________________________________________ > Users mailing list -- users@ovirt.org > To unsubscribe send an email to users-le...@ovirt.org > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/users@ovirt.org/message/P2WRSR6U67BOQYKMEJFNMSRJFRMJFECN/ >
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/JZAX3ETMTDWC4MWK7NW4T4ZZ52ETD5GB/