I'm trying to bring up a single node hyperconverged with the current node-ng ISO installation, but it ends with this failure message.
TASK [gluster.features/roles/gluster_hci : Check if /var/log has enough disk space] *** fatal: [br014.bridges.psc.edu]: FAILED! => {"changed": true, "cmd": "df -m /var/log | awk '/[0-9]%/ {print $4}'", "delta": "0:00:00.008513", "end": "2019-05-09 20:09:27.914400", "failed_when_result": true, "rc": 0, "start": "2019-05-09 20:09:27.905887", "stderr": "", "stderr_lines": [], "stdout": "7470", "stdout_lines": ["7470"]} I have what the installer created by default for /var/log, so I don't know why its complaining. [root@br014 ~]# df -kh Filesystem Size Used Avail Use% Mounted on /dev/mapper/onn_br014-ovirt--node--ng--4.3.3.1--0.20190417.0+1 3.5T 2.1G 3.3T 1% / devtmpfs 63G 0 63G 0% /dev tmpfs 63G 4.0K 63G 1% /dev/shm tmpfs 63G 18M 63G 1% /run tmpfs 63G 0 63G 0% /sys/fs/cgroup /dev/mapper/onn_br014-home 976M 2.6M 907M 1% /home /dev/mapper/onn_br014-tmp 976M 2.8M 906M 1% /tmp /dev/mapper/onn_br014-var 15G 42M 14G 1% /var /dev/sda2 976M 173M 737M 19% /boot /dev/mapper/onn_br014-var_log 7.8G 41M 7.3G 1% /var/log /dev/mapper/onn_br014-var_log_audit 2.0G 7.6M 1.8G 1% /var/log/audit /dev/mapper/onn_br014-var_crash 9.8G 37M 9.2G 1% /var/crash /dev/sda1 200M 12M 189M 6% /boot/efi tmpfs 13G 0 13G 0% /run/user/1000 tmpfs 13G 0 13G 0% /run/user/0 /dev/mapper/gluster_vg_sdb-gluster_lv_engine 3.7T 33M 3.7T 1% /gluster_bricks/engine /dev/mapper/gluster_vg_sdc-gluster_lv_data 3.7T 34M 3.7T 1% /gluster_bricks/data /dev/mapper/gluster_vg_sdd-gluster_lv_vmstore 3.7T 34M 3.7T 1% /gluster_bricks/vmstore The machine had 4 4TB disks, so sda is the installation for oVirt node-ng, the other 3 disks for the gluster volumes. root@br014 ~]# pvs PV VG Fmt Attr PSize PFree /dev/sda3 onn_br014 lvm2 a-- <3.64t 100.00g /dev/sdb gluster_vg_sdb lvm2 a-- <3.64t <26.02g /dev/sdc gluster_vg_sdc lvm2 a-- <3.64t 0 /dev/sdd gluster_vg_sdd lvm2 a-- <3.64t 0 [root@br014 ~]# vgs VG #PV #LV #SN Attr VSize VFree gluster_vg_sdb 1 1 0 wz--n- <3.64t <26.02g gluster_vg_sdc 1 2 0 wz--n- <3.64t 0 gluster_vg_sdd 1 2 0 wz--n- <3.64t 0 onn_br014 1 11 0 wz--n- <3.64t 100.00g [root@br014 ~]# lvs LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert gluster_lv_engine gluster_vg_sdb -wi-ao---- 3.61t gluster_lv_data gluster_vg_sdc Vwi-aot--- 3.61t gluster_thinpool_gluster_vg_sdc 0.05 gluster_thinpool_gluster_vg_sdc gluster_vg_sdc twi-aot--- <3.61t 0.05 0.13 gluster_lv_vmstore gluster_vg_sdd Vwi-aot--- 3.61t gluster_thinpool_gluster_vg_sdd 0.05 gluster_thinpool_gluster_vg_sdd gluster_vg_sdd twi-aot--- <3.61t 0.05 0.13 home onn_br014 Vwi-aotz-- 1.00g pool00 4.79 ovirt-node-ng-4.3.3.1-0.20190417.0 onn_br014 Vwi---tz-k <3.51t pool00 root ovirt-node-ng-4.3.3.1-0.20190417.0+1 onn_br014 Vwi-aotz-- <3.51t pool00 ovirt-node-ng-4.3.3.1-0.20190417.0 0.13 pool00 onn_br014 twi-aotz-- 3.53t 0.19 1.86 root onn_br014 Vri---tz-k <3.51t pool00 swap onn_br014 -wi-ao---- 4.00g tmp onn_br014 Vwi-aotz-- 1.00g pool00 4.84 var onn_br014 Vwi-aotz-- 15.00g pool00 3.67 var_crash onn_br014 Vwi-aotz-- 10.00g pool00 2.86 var_log onn_br014 Vwi-aotz-- 8.00g pool00 3.25 var_log_audit onn_br014 Vwi-aotz-- 2.00g pool00 4.86 Here's the full deploy log from the UI. Let me know if you need specific logs. PLAY [Setup backend] *********************************************************** TASK [Gathering Facts] ********************************************************* ok: [br014.bridges.psc.edu] TASK [gluster.infra/roles/firewall_config : Start firewalld if not already started] *** ok: [br014.bridges.psc.edu] TASK [gluster.infra/roles/firewall_config : check if required variables are set] *** skipping: [br014.bridges.psc.edu] TASK [gluster.infra/roles/firewall_config : Open/Close firewalld ports] ******** ok: [br014.bridges.psc.edu] => (item=2049/tcp) ok: [br014.bridges.psc.edu] => (item=54321/tcp) ok: [br014.bridges.psc.edu] => (item=5900/tcp) ok: [br014.bridges.psc.edu] => (item=5900-6923/tcp) ok: [br014.bridges.psc.edu] => (item=5666/tcp) ok: [br014.bridges.psc.edu] => (item=16514/tcp) TASK [gluster.infra/roles/firewall_config : Add/Delete services to firewalld rules] *** ok: [br014.bridges.psc.edu] => (item=glusterfs) TASK [gluster.infra/roles/backend_setup : Gather facts to determine the OS distribution] *** ok: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for debian systems.] *** skipping: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Change to Install lvm tools for RHEL systems.] *** ok: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Install python-yaml package for Debian systems] *** skipping: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Initialize vdo_devs array] *********** ok: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Record VDO devices (if any)] ********* skipping: [br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}) skipping: [br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdc', u'pvname': u'/dev/sdc'}) skipping: [br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdd', u'pvname': u'/dev/sdd'}) TASK [gluster.infra/roles/backend_setup : Enable and start vdo service] ******** skipping: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Create VDO with specified size] ****** skipping: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Check if valid disktype is provided] *** skipping: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Set PV data alignment for JBOD] ****** ok: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Set PV data alignment for RAID] ****** skipping: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Set VG physical extent size for RAID] *** skipping: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Create volume groups] **************** ok: [br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}) ok: [br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdc', u'pvname': u'/dev/sdc'}) ok: [br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdd', u'pvname': u'/dev/sdd'}) TASK [gluster.infra/roles/backend_setup : Create thick logical volume] ********* ok: [br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'3700G'}) TASK [gluster.infra/roles/backend_setup : Calculate chunksize for RAID6/RAID10/RAID5] *** skipping: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Set chunksize for JBOD] ************** ok: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Create a LV thinpool] **************** ok: [br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdc', u'thinpoolname': u'gluster_thinpool_gluster_vg_sdc', u'poolmetadatasize': u'16G'}) ok: [br014.bridges.psc.edu] => (item={u'vgname': u'gluster_vg_sdd', u'thinpoolname': u'gluster_thinpool_gluster_vg_sdd', u'poolmetadatasize': u'16G'}) TASK [gluster.infra/roles/backend_setup : Create thin logical volume] ********** ok: [br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_data', u'vgname': u'gluster_vg_sdc', u'thinpool': u'gluster_thinpool_gluster_vg_sdc', u'lvsize': u'3700G'}) ok: [br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_vmstore', u'vgname': u'gluster_vg_sdd', u'thinpool': u'gluster_thinpool_gluster_vg_sdd', u'lvsize': u'3700G'}) TASK [gluster.infra/roles/backend_setup : Extend volume group] ***************** skipping: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Change attributes of LV] ************* skipping: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Create LV for cache] ***************** skipping: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Create metadata LV for cache] ******** skipping: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Convert logical volume to a cache pool LV] *** skipping: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Convert logical volume to a cache pool LV without cachemetalvname] *** skipping: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Convert an existing logical volume to a cache LV] *** skipping: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Set XFS options for JBOD] ************ ok: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Set XFS options for RAID devices] **** skipping: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Create filesystem on thin logical vols] *** ok: [br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_data', u'vgname': u'gluster_vg_sdc', u'thinpool': u'gluster_thinpool_gluster_vg_sdc', u'lvsize': u'3700G'}) ok: [br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_vmstore', u'vgname': u'gluster_vg_sdd', u'thinpool': u'gluster_thinpool_gluster_vg_sdd', u'lvsize': u'3700G'}) TASK [gluster.infra/roles/backend_setup : Create filesystem on thick logical vols] *** ok: [br014.bridges.psc.edu] => (item={u'lvname': u'gluster_lv_engine', u'vgname': u'gluster_vg_sdb', u'size': u'3700G'}) TASK [gluster.infra/roles/backend_setup : Create mount directories if not already present] *** ok: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/engine', u'vgname': u'gluster_vg_sdb', u'lvname': u'gluster_lv_engine'}) ok: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/data', u'vgname': u'gluster_vg_sdc', u'lvname': u'gluster_lv_data'}) ok: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/vmstore', u'vgname': u'gluster_vg_sdd', u'lvname': u'gluster_lv_vmstore'}) TASK [gluster.infra/roles/backend_setup : Set mount options for VDO] *********** skipping: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_setup : Mount the vdo devices (If any)] ****** skipping: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/engine', u'vgname': u'gluster_vg_sdb', u'lvname': u'gluster_lv_engine'}) skipping: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/data', u'vgname': u'gluster_vg_sdc', u'lvname': u'gluster_lv_data'}) skipping: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/vmstore', u'vgname': u'gluster_vg_sdd', u'lvname': u'gluster_lv_vmstore'}) TASK [gluster.infra/roles/backend_setup : Mount the devices] ******************* ok: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/engine', u'vgname': u'gluster_vg_sdb', u'lvname': u'gluster_lv_engine'}) ok: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/data', u'vgname': u'gluster_vg_sdc', u'lvname': u'gluster_lv_data'}) ok: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/vmstore', u'vgname': u'gluster_vg_sdd', u'lvname': u'gluster_lv_vmstore'}) TASK [gluster.infra/roles/backend_setup : Set Gluster specific SeLinux context on the bricks] *** ok: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/engine', u'vgname': u'gluster_vg_sdb', u'lvname': u'gluster_lv_engine'}) ok: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/data', u'vgname': u'gluster_vg_sdc', u'lvname': u'gluster_lv_data'}) ok: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/vmstore', u'vgname': u'gluster_vg_sdd', u'lvname': u'gluster_lv_vmstore'}) TASK [gluster.infra/roles/backend_setup : restore file(s) default SELinux security contexts] *** changed: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/engine', u'vgname': u'gluster_vg_sdb', u'lvname': u'gluster_lv_engine'}) changed: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/data', u'vgname': u'gluster_vg_sdc', u'lvname': u'gluster_lv_data'}) changed: [br014.bridges.psc.edu] => (item={u'path': u'/gluster_bricks/vmstore', u'vgname': u'gluster_vg_sdd', u'lvname': u'gluster_lv_vmstore'}) TASK [gluster.infra/roles/backend_reset : unmount the directories (if mounted)] *** skipping: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_reset : Delete volume groups] **************** skipping: [br014.bridges.psc.edu] TASK [gluster.infra/roles/backend_reset : Remove VDO devices] ****************** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Create temporary storage directory] *** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Get the name of the directory created] *** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : check if gluster_features_ganesha_clusternodes is set] *** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Define service port] **************** skipping: [br014.bridges.psc.edu] => (item=^#(STATD_PORT=.*)) skipping: [br014.bridges.psc.edu] => (item=^#(LOCKD_TCPPORT=.*)) skipping: [br014.bridges.psc.edu] => (item=^#(LOCKD_UDPPORT=.*)) TASK [gluster.features/roles/nfs_ganesha : Check packages installed, if not install] *** skipping: [br014.bridges.psc.edu] => (item=glusterfs-ganesha) skipping: [br014.bridges.psc.edu] => (item=nfs-ganesha) skipping: [br014.bridges.psc.edu] => (item=corosync) skipping: [br014.bridges.psc.edu] => (item=pacemaker) skipping: [br014.bridges.psc.edu] => (item=libntirpc) skipping: [br014.bridges.psc.edu] => (item=pcs) TASK [gluster.features/roles/nfs_ganesha : Restart services] ******************* skipping: [br014.bridges.psc.edu] => (item=nfslock) skipping: [br014.bridges.psc.edu] => (item=nfs-config) skipping: [br014.bridges.psc.edu] => (item=rpc-statd) TASK [gluster.features/roles/nfs_ganesha : Stop services] ********************** skipping: [br014.bridges.psc.edu] => (item=nfs-server) TASK [gluster.features/roles/nfs_ganesha : Disable service] ******************** skipping: [br014.bridges.psc.edu] => (item=nfs-server) TASK [gluster.features/roles/nfs_ganesha : Enable services] ******************** skipping: [br014.bridges.psc.edu] => (item=glusterfssharedstorage) skipping: [br014.bridges.psc.edu] => (item=nfs-ganesha) skipping: [br014.bridges.psc.edu] => (item=network) skipping: [br014.bridges.psc.edu] => (item=pcsd) skipping: [br014.bridges.psc.edu] => (item=pacemaker) TASK [gluster.features/roles/nfs_ganesha : Start services] ********************* skipping: [br014.bridges.psc.edu] => (item=network) skipping: [br014.bridges.psc.edu] => (item=pcsd) TASK [gluster.features/roles/nfs_ganesha : Create a user hacluster if not already present] *** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Set the password for hacluster] ***** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Set the hacluster user the same password on new nodes] *** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Pcs cluster authenticate the hacluster on new nodes] *** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Pause for a few seconds after pcs auth] *** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Set gluster_use_execmem flag on and keep it persistent] *** skipping: [br014.bridges.psc.edu] => (item=gluster_use_execmem) skipping: [br014.bridges.psc.edu] => (item=ganesha_use_fusefs) TASK [gluster.features/roles/nfs_ganesha : check if gluster_features_ganesha_masternode is set] *** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Copy the ssh keys to the local machine] *** skipping: [br014.bridges.psc.edu] => (item=secret.pem.pub) skipping: [br014.bridges.psc.edu] => (item=secret.pem) TASK [gluster.features/roles/nfs_ganesha : check if gluster_features_ganesha_newnodes_vip is set] *** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Copy the public key to remote nodes] *** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Copy the private key to remote node] *** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Deploy the pubkey on all nodes] ***** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Make the volume a gluster shared volume] *** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Generate ssh key in one of the nodes in HA cluster] *** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Copy the ssh keys to the local machine] *** skipping: [br014.bridges.psc.edu] => (item=secret.pem.pub) skipping: [br014.bridges.psc.edu] => (item=secret.pem) TASK [gluster.features/roles/nfs_ganesha : Create configuration directory for nfs_ganesha] *** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Copy ganesha.conf to config directory on shared volume] *** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Create ganesha-ha.conf file] ******** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Enable NFS Ganesha] ***************** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Pause for 30 seconds (takes a while to enable NFS Ganesha)] *** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Check NFS Ganesha status] *********** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Report NFS Ganesha status] ********** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Report NFS Ganesha status (If any errors)] *** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : check if gluster_features_ganesha_volume is set] *** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Export the NFS Ganesha volume] ****** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Copy the public key to remote nodes] *** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Copy the private key to remote node] *** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Deploy the pubkey on all nodes] ***** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Adds a node to the cluster] ********* skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Report ganesha add-node status] ***** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/nfs_ganesha : Delete the temporary directory] ***** skipping: [br014.bridges.psc.edu] TASK [gluster.features/roles/gluster_hci : Check if packages are installed, if not install] *** ok: [br014.bridges.psc.edu] => (item=vdsm) ok: [br014.bridges.psc.edu] => (item=vdsm-gluster) ok: [br014.bridges.psc.edu] => (item=ovirt-host) ok: [br014.bridges.psc.edu] => (item=screen) TASK [gluster.features/roles/gluster_hci : Enable and start glusterd and chronyd] *** ok: [br014.bridges.psc.edu] => (item=chronyd) ok: [br014.bridges.psc.edu] => (item=glusterd) ok: [br014.bridges.psc.edu] => (item=firewalld) TASK [gluster.features/roles/gluster_hci : Add user qemu to gluster group] ***** ok: [br014.bridges.psc.edu] TASK [gluster.features/roles/gluster_hci : Disable the hook scripts] *********** changed: [br014.bridges.psc.edu] => (item=/var/lib/glusterd/hooks/1/set/post/S30samba-set.sh) changed: [br014.bridges.psc.edu] => (item=/var/lib/glusterd/hooks/1/start/post/S30samba-start.sh) changed: [br014.bridges.psc.edu] => (item=/var/lib/glusterd/hooks/1/stop/pre/S30samba-stop.sh) changed: [br014.bridges.psc.edu] => (item=/var/lib/glusterd/hooks/1/reset/post/S31ganesha-reset.sh) changed: [br014.bridges.psc.edu] => (item=/var/lib/glusterd/hooks/1//start/post/S31ganesha-start.sh) changed: [br014.bridges.psc.edu] => (item=/var/lib/glusterd/hooks/1/start/post/S29CTDBsetup.sh) changed: [br014.bridges.psc.edu] => (item=/var/lib/glusterd/hooks/1/stop/pre/S29CTDB-teardown.sh) TASK [gluster.features/roles/gluster_hci : Check if valid FQDN is provided] **** changed: [br014.bridges.psc.edu -> localhost] => (item=br014.bridges.psc.edu ) TASK [gluster.features/roles/gluster_hci : Check if /var/log has enough disk space] *** fatal: [br014.bridges.psc.edu]: FAILED! => {"changed": true, "cmd": "df -m /var/log | awk '/[0-9]%/ {print $4}'", "delta": "0:00:00.008513", "end": "2019-05-09 20:09:27.914400", "failed_when_result": true, "rc": 0, "start": "2019-05-09 20:09:27.905887", "stderr": "", "stderr_lines": [], "stdout": "7470", "stdout_lines": ["7470"]} NO MORE HOSTS LEFT ************************************************************* NO MORE HOSTS LEFT ************************************************************* to retry, use: --limit @/usr/share/cockpit/ovirt-dashboard/ansible/hc_wizard.retry PLAY RECAP ********************************************************************* br014.bridges.psc.edu : ok=25 changed=3 unreachable=0 failed=1
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/P2WRSR6U67BOQYKMEJFNMSRJFRMJFECN/