Re: [ovirt-users] Confused / Hosted-Engine won't start

2015-07-02 Thread Sandro Bonazzola
Il 02/07/2015 02:36, Christopher Young ha scritto:
 I'm sure I have worked through this before, but I've been banging my head 
 against this one for a while now, and I think I'm too close to the issue.
 
 Basically, my hosted-engine won't start anymore.  I do recall attempting to 
 migrate it to a new gluster-based NFS share recently, but I could have
 sworn that was successful and working.  I _think_ I have some sort of id 
 issue with storage/volume/whatever id's, but I need some help digging through
 it if someone would be so kind.
 
 I have the following error in /var/log/libvirt/qemu/HostedEngine.log:
 
 --
 2015-07-02T00:01:13.080952Z qemu-kvm: -drive
 file=/var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8-7057fcee412c/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432-4df6-8856-fdb14df260e3,if=none,id=drive-virtio-disk0,format=raw,serial=5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836,cache=none,werror=stop,rerror=stop,aio=threads:
 could not open disk image
 /var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8-7057fcee412c/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432-4df6-8856-fdb14df260e3:
  Could not
 refresh total sector count: Input/output error
 --


Please check selinux: ausearch -m avc
it might be selinux preventing access to the disk image.



 
 I also am including a few command outputs that I'm sure might help:
 
 --
 [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8-7057fcee412c]# ls -la 
 /var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8-7057fcee412c/
 total 8
 drwxr-xr-x. 2 vdsm kvm  80 Jul  1 20:04 .
 drwxr-xr-x. 3 vdsm kvm  60 Jul  1 20:04 ..
 lrwxrwxrwx. 1 vdsm kvm 128 Jul  1 20:04 1d80a60c-8f26-4448-9460-2c7b00ff75bf 
 -
 /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062-4ad1-9df8-7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf
 lrwxrwxrwx. 1 vdsm kvm 128 Jul  1 20:04 23ac8897-b0c7-41d6-a7de-19f46ed78400 
 -
 /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062-4ad1-9df8-7057fcee412c/images/23ac8897-b0c7-41d6-a7de-19f46ed78400
 
 --
 [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8-7057fcee412c]# cat 
 /etc/ovirt-hosted-engine/vm.conf
 vmId=6b7329f9-518a-4488-b1c4-2cd809f2f580
 memSize=5120
 display=vnc
 devices={index:2,iface:ide,address:{ controller:0, target:0,unit:0, bus:1,
 type:drive},specParams:{},readonly:true,deviceId:77924fc2-c5c9-408b-97d3-cd0b0d11a62c,path:/home/tmp/CentOS-6.6-x86_64-minimal.iso,device:cdrom,shared:false,type:disk}
 devices={index:0,iface:virtio,format:raw,poolID:----,volumeID:eeb2d821-a432-4df6-8856-fdb14df260e3,imageID:5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836,specParams:{},readonly:false,domainID:4e3017eb-d062-4ad1-9df8-7057fcee412c,optional:false,deviceId:5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836,address:{bus:0x00,
 slot:0x06, domain:0x, type:pci, 
 function:0x0},device:disk,shared:exclusive,propagateErrors:off,type:disk,bootOrder:1}
 devices={device:scsi,model:virtio-scsi,type:controller}
 devices={nicModel:pv,macAddr:00:16:3e:0e:d0:68,linkActive:true,network:ovirtmgmt,filter:vdsm-no-mac-spoofing,specParams:{},deviceId:f70ba622-6ac8-4c06-a005-0ebd940a15b2,address:{bus:0x00,
 slot:0x03, domain:0x, type:pci, 
 function:0x0},device:bridge,type:interface}
 devices={device:console,specParams:{},type:console,deviceId:98386e6c-ae56-4b6d-9bfb-c72bbd299ad1,alias:console0}
 vmName=HostedEngine
 spiceSecureChannels=smain,sdisplay,sinputs,scursor,splayback,srecord,ssmartcard,susbredir
 smp=2
 cpuType=Westmere
 emulatedMachine=pc
 --
 
 [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8-7057fcee412c]# cat 
 /etc/ovirt-hosted-engine/hosted-engine.conf
 fqdn=orldc-dev-vengine01.***
 vm_disk_id=5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836
 vmid=6b7329f9-518a-4488-b1c4-2cd809f2f580
 storage=ovirt-gluster-nfs:/engine
 conf=/etc/ovirt-hosted-engine/vm.conf
 service_start_time=0
 host_id=1
 console=vnc
 domainType=nfs3
 spUUID=379cf161-d5b1-4c20-bb71-e3ca5d2ccd6b
 sdUUID=4e3017eb-d062-4ad1-9df8-7057fcee412c
 connectionUUID=0d1b50ac-cf3f-4cd7-90df-3c3a6d11a984
 ca_cert=/etc/pki/vdsm/libvirt-spice/ca-cert.pem
 ca_subject=C=EN, L=Test, O=Test, CN=Test
 vdsm_use_ssl=true
 gateway=10.16.3.1
 bridge=ovirtmgmt
 metadata_volume_UUID=dd9f373c-d161-4fa0-aab1-3cb52305dba7
 metadata_image_UUID=23ac8897-b0c7-41d6-a7de-19f46ed78400
 lockspace_volume_UUID=d9bacbf6-c2f4-4f74-a91f-3a3a52f255bf
 lockspace_image_UUID=1d80a60c-8f26-4448-9460-2c7b00ff75bf
 
 # The following are used only for iSCSI storage
 iqn=
 portal=
 user=
 password=
 port=
 --
 
 (mount output for the NFS share this should be running on (gluster-based):
 
 ovirt-gluster-nfs:/engine on /rhev/data-center/mnt/ovirt-gluster-nfs:_engine 
 type nfs
 (rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,soft,nolock,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mountaddr=10.16.3.30,mountvers=3,mountport=38465,mountproto=tcp,local_lock=all,addr=10.16.3.30)
 
 --
 
 [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8-7057fcee412c]# 

Re: [ovirt-users] Confused / Hosted-Engine won't start

2015-07-02 Thread Sandro Bonazzola
Il 02/07/2015 16:04, Christopher Young ha scritto:
 [root@orldc-dev-vnode02 ~]# getenforce 
 Permissive
 
 It looks like it isn't SELinux.

Checked also on the storage server ovirt-gluster-nfs?


 
 On Thu, 2015-07-02 at 09:53 +0200, Sandro Bonazzola wrote:
 Il 02/07/2015 02:36, Christopher Young ha scritto:
 I'm sure I have worked through this before, but I've been banging 
 my head against this one for a while now, and I think I'm too close 
 to the issue.

 Basically, my hosted-engine won't start anymore.  I do recall 
 attempting to migrate it to a new gluster-based NFS share recently, 
 but I could have
 sworn that was successful and working.  I _think_ I have some sort 
 of id issue with storage/volume/whatever id's, but I need some help 
 digging through
 it if someone would be so kind.

 I have the following error in 
 /var/log/libvirt/qemu/HostedEngine.log:

 --
 2015-07-02T00:01:13.080952Z qemu-kvm: -drive
 file=/var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8
 -7057fcee412c/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432
 -4df6-8856-fdb14df260e3,if=none,id=drive-virtio
 -disk0,format=raw,serial=5ead7b5d-50e8-4d6c-a0e5
 -bbe6d93dd836,cache=none,werror=stop,rerror=stop,aio=threads:
 could not open disk image
 /var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8-7057fcee412c/5ead7b5d
 -50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432-4df6-8856-fdb14df260e3: 
 Could not
 refresh total sector count: Input/output error
 --


 Please check selinux: ausearch -m avc
 it might be selinux preventing access to the disk image.




 I also am including a few command outputs that I'm sure might help:

 --
 [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8-7057fcee412c]# ls 
 -la /var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8-7057fcee412c/
 total 8
 drwxr-xr-x. 2 vdsm kvm  80 Jul  1 20:04 .
 drwxr-xr-x. 3 vdsm kvm  60 Jul  1 20:04 ..
 lrwxrwxrwx. 1 vdsm kvm 128 Jul  1 20:04 1d80a60c-8f26-4448-9460
 -2c7b00ff75bf -
 /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062-4ad1
 -9df8-7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf
 lrwxrwxrwx. 1 vdsm kvm 128 Jul  1 20:04 23ac8897-b0c7-41d6-a7de
 -19f46ed78400 -
 /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062-4ad1
 -9df8-7057fcee412c/images/23ac8897-b0c7-41d6-a7de-19f46ed78400

 --
 [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8-7057fcee412c]# cat 
 /etc/ovirt-hosted-engine/vm.conf
 vmId=6b7329f9-518a-4488-b1c4-2cd809f2f580
 memSize=5120
 display=vnc
 devices={index:2,iface:ide,address:{ controller:0, target:0,unit:0, 
 bus:1,
 type:drive},specParams:{},readonly:true,deviceId:77924fc2-c5c9-408b
 -97d3-cd0b0d11a62c,path:/home/tmp/CentOS-6.6-x86_64
 -minimal.iso,device:cdrom,shared:false,type:disk}
 devices={index:0,iface:virtio,format:raw,poolID:--
 --,volumeID:eeb2d821-a432-4df6-8856
 -fdb14df260e3,imageID:5ead7b5d-50e8-4d6c-a0e5
 -bbe6d93dd836,specParams:{},readonly:false,domainID:4e3017eb-d062
 -4ad1-9df8-7057fcee412c,optional:false,deviceId:5ead7b5d-50e8-4d6c
 -a0e5-bbe6d93dd836,address:{bus:0x00,
 slot:0x06, domain:0x, type:pci, 
 function:0x0},device:disk,shared:exclusive,propagateErrors:off,type
 :disk,bootOrder:1}
 devices={device:scsi,model:virtio-scsi,type:controller}
 devices={nicModel:pv,macAddr:00:16:3e:0e:d0:68,linkActive:true,netw
 ork:ovirtmgmt,filter:vdsm-no-mac
 -spoofing,specParams:{},deviceId:f70ba622-6ac8-4c06-a005
 -0ebd940a15b2,address:{bus:0x00,
 slot:0x03, domain:0x, type:pci, 
 function:0x0},device:bridge,type:interface}
 devices={device:console,specParams:{},type:console,deviceId:98386e6
 c-ae56-4b6d-9bfb-c72bbd299ad1,alias:console0}
 vmName=HostedEngine
 spiceSecureChannels=smain,sdisplay,sinputs,scursor,splayback,srecor
 d,ssmartcard,susbredir
 smp=2
 cpuType=Westmere
 emulatedMachine=pc
 --

 [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8-7057fcee412c]# cat 
 /etc/ovirt-hosted-engine/hosted-engine.conf
 fqdn=orldc-dev-vengine01.***
 vm_disk_id=5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836
 vmid=6b7329f9-518a-4488-b1c4-2cd809f2f580
 storage=ovirt-gluster-nfs:/engine
 conf=/etc/ovirt-hosted-engine/vm.conf
 service_start_time=0
 host_id=1
 console=vnc
 domainType=nfs3
 spUUID=379cf161-d5b1-4c20-bb71-e3ca5d2ccd6b
 sdUUID=4e3017eb-d062-4ad1-9df8-7057fcee412c
 connectionUUID=0d1b50ac-cf3f-4cd7-90df-3c3a6d11a984
 ca_cert=/etc/pki/vdsm/libvirt-spice/ca-cert.pem
 ca_subject=C=EN, L=Test, O=Test, CN=Test
 vdsm_use_ssl=true
 gateway=10.16.3.1
 bridge=ovirtmgmt
 metadata_volume_UUID=dd9f373c-d161-4fa0-aab1-3cb52305dba7
 metadata_image_UUID=23ac8897-b0c7-41d6-a7de-19f46ed78400
 lockspace_volume_UUID=d9bacbf6-c2f4-4f74-a91f-3a3a52f255bf
 lockspace_image_UUID=1d80a60c-8f26-4448-9460-2c7b00ff75bf

 # The following are used only for iSCSI storage
 iqn=
 portal=
 user=
 password=
 port=
 --

 (mount output for the NFS share this should be running on (gluster
 -based):

 ovirt-gluster-nfs:/engine on /rhev/data-center/mnt/ovirt-gluster
 -nfs:_engine type nfs

Re: [ovirt-users] Confused / Hosted-Engine won't start

2015-07-02 Thread Christopher Young
That's actually a local NFS implementation (running on Gluster) so I
wouldn't expect that it would factor in, but I just verified that all
gluster hosts are in Permissive mode.  Like I mentioned, I think I may
have some ids wrong or bad links or something since this had previously
been working fine, though I must be missing it. 

On Thu, 2015-07-02 at 16:08 +0200, Sandro Bonazzola wrote:
 Il 02/07/2015 16:04, Christopher Young ha scritto:
  [root@orldc-dev-vnode02 ~]# getenforce 
  Permissive
  
  It looks like it isn't SELinux.
 
 Checked also on the storage server ovirt-gluster-nfs?
 
 
  
  On Thu, 2015-07-02 at 09:53 +0200, Sandro Bonazzola wrote:
   Il 02/07/2015 02:36, Christopher Young ha scritto:
I'm sure I have worked through this before, but I've been 
banging 
my head against this one for a while now, and I think I'm too 
close 
to the issue.

Basically, my hosted-engine won't start anymore.  I do recall 
attempting to migrate it to a new gluster-based NFS share 
recently, 
but I could have
sworn that was successful and working.  I _think_ I have some 
sort 
of id issue with storage/volume/whatever id's, but I need some 
help 
digging through
it if someone would be so kind.

I have the following error in 
/var/log/libvirt/qemu/HostedEngine.log:

--
2015-07-02T00:01:13.080952Z qemu-kvm: -drive
file=/var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8
-7057fcee412c/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821
-a432
-4df6-8856-fdb14df260e3,if=none,id=drive-virtio
-disk0,format=raw,serial=5ead7b5d-50e8-4d6c-a0e5
-bbe6d93dd836,cache=none,werror=stop,rerror=stop,aio=threads:
could not open disk image
/var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8
-7057fcee412c/5ead7b5d
-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432-4df6-8856
-fdb14df260e3: 
Could not
refresh total sector count: Input/output error
--
   
   
   Please check selinux: ausearch -m avc
   it might be selinux preventing access to the disk image.
   
   
   

I also am including a few command outputs that I'm sure might 
help:

--
[root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8-7057fcee412c]# 
ls 
-la /var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8-7057fcee412c/
total 8
drwxr-xr-x. 2 vdsm kvm  80 Jul  1 20:04 .
drwxr-xr-x. 3 vdsm kvm  60 Jul  1 20:04 ..
lrwxrwxrwx. 1 vdsm kvm 128 Jul  1 20:04 1d80a60c-8f26-4448-9460
-2c7b00ff75bf -
/rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062
-4ad1
-9df8-7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf
lrwxrwxrwx. 1 vdsm kvm 128 Jul  1 20:04 23ac8897-b0c7-41d6-a7de
-19f46ed78400 -
/rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062
-4ad1
-9df8-7057fcee412c/images/23ac8897-b0c7-41d6-a7de-19f46ed78400

--
[root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8-7057fcee412c]# 
cat 
/etc/ovirt-hosted-engine/vm.conf
vmId=6b7329f9-518a-4488-b1c4-2cd809f2f580
memSize=5120
display=vnc
devices={index:2,iface:ide,address:{ controller:0, 
target:0,unit:0, 
bus:1,
type:drive},specParams:{},readonly:true,deviceId:77924fc2-c5c9
-408b
-97d3-cd0b0d11a62c,path:/home/tmp/CentOS-6.6-x86_64
-minimal.iso,device:cdrom,shared:false,type:disk}
devices={index:0,iface:virtio,format:raw,poolID:-
-
--,volumeID:eeb2d821-a432-4df6-8856
-fdb14df260e3,imageID:5ead7b5d-50e8-4d6c-a0e5
-bbe6d93dd836,specParams:{},readonly:false,domainID:4e3017eb
-d062
-4ad1-9df8-7057fcee412c,optional:false,deviceId:5ead7b5d-50e8
-4d6c
-a0e5-bbe6d93dd836,address:{bus:0x00,
slot:0x06, domain:0x, type:pci, 
function:0x0},device:disk,shared:exclusive,propagateErrors:off,
type
:disk,bootOrder:1}
devices={device:scsi,model:virtio-scsi,type:controller}
devices={nicModel:pv,macAddr:00:16:3e:0e:d0:68,linkActive:true,
netw
ork:ovirtmgmt,filter:vdsm-no-mac
-spoofing,specParams:{},deviceId:f70ba622-6ac8-4c06-a005
-0ebd940a15b2,address:{bus:0x00,
slot:0x03, domain:0x, type:pci, 
function:0x0},device:bridge,type:interface}
devices={device:console,specParams:{},type:console,deviceId:983
86e6
c-ae56-4b6d-9bfb-c72bbd299ad1,alias:console0}
vmName=HostedEngine
spiceSecureChannels=smain,sdisplay,sinputs,scursor,splayback,sr
ecor
d,ssmartcard,susbredir
smp=2
cpuType=Westmere
emulatedMachine=pc
--

[root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8-7057fcee412c]# 
cat 
/etc/ovirt-hosted-engine/hosted-engine.conf
fqdn=orldc-dev-vengine01.***
vm_disk_id=5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836
vmid=6b7329f9-518a-4488-b1c4-2cd809f2f580
storage=ovirt-gluster-nfs:/engine
conf=/etc/ovirt-hosted-engine/vm.conf
service_start_time=0
host_id=1

Re: [ovirt-users] Confused / Hosted-Engine won't start

2015-07-02 Thread Christopher Young
If anyone has an experience of the various IDs in hosted-engine.conf
and vm.conf for the Hosted Engine, I believe I need to just verify
everything.  I tried a couple of changes, but I feel like I'm just
making this worse, so I've reverted them.

One thing I do not understand well is how a gluster-based (NFS) storage
domain for the hosted-engine has so many entries:

-

[root@orldc-dev-vnode02 ovirt-gluster-nfs:_engine]# find . -type f |
xargs ls -lah
ls: cannot access ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432
-4df6-8856-fdb14df260e3: Input/output error
-rw-rw. 1 vdsm kvm  1.0M Jul  2 11:20 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/dom_md/ids
-rw-rw. 1 vdsm kvm   16M Jul  1 19:54 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/dom_md/inbox
-rw-rw. 1 vdsm kvm  2.0M Jul  1 19:50 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/dom_md/leases
-rw-r--r--. 1 vdsm kvm   482 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/dom_md/metadata
-rw-rw. 1 vdsm kvm   16M Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/dom_md/outbox
-rw-rw. 1 vdsm kvm  1.0M Jul  2 11:32 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf/d9bacbf6-c2f4
-4f74-a91f-3a3a52f255bf
-rw-rw. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf/d9bacbf6-c2f4
-4f74-a91f-3a3a52f255bf.lease
-rw-r--r--. 1 vdsm kvm   284 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf/d9bacbf6-c2f4
-4f74-a91f-3a3a52f255bf.meta
-rw-rw. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/23ac8897-b0c7-41d6-a7de-19f46ed78400/dd9f373c-d161
-4fa0-aab1-3cb52305dba7.lease
-rw-r--r--. 1 vdsm kvm   283 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/23ac8897-b0c7-41d6-a7de-19f46ed78400/dd9f373c-d161
-4fa0-aab1-3cb52305dba7.meta
-rw-rw. 1 vdsm kvm   25G Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/3278c444-d92a-4cb9-87d6-9669c6e4993e/1a4b6a5d-9c8f
-4d54-91a7-3dd87377c362
-rw-rw. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/3278c444-d92a-4cb9-87d6-9669c6e4993e/1a4b6a5d-9c8f
-4d54-91a7-3dd87377c362.lease
-rw-r--r--. 1 vdsm kvm   278 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/3278c444-d92a-4cb9-87d6-9669c6e4993e/1a4b6a5d-9c8f
-4d54-91a7-3dd87377c362.meta
-rw-rw. 1 vdsm kvm  1.0M Jul  1 17:50 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432
-4df6-8856-fdb14df260e3.lease
-rw-r--r--. 1 vdsm kvm   278 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432
-4df6-8856-fdb14df260e3.meta
-rw-rw. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/6064179f-2720-4db9-a7c4-a97e044c2238/05afaa26-95af
-4226-9a00-5383d8937a83
-rw-rw. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/6064179f-2720-4db9-a7c4-a97e044c2238/05afaa26-95af
-4226-9a00-5383d8937a83.lease
-rw-r--r--. 1 vdsm kvm   284 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/6064179f-2720-4db9-a7c4-a97e044c2238/05afaa26-95af
-4226-9a00-5383d8937a83.meta
-rw-rw. 1 vdsm kvm 1004K Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/bb9d9a37-4f91-4973-ba9e-72ee81aed0b6/5acb27b3-62c5
-46ac-8978-576a8a4a0399
-rw-rw. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/bb9d9a37-4f91-4973-ba9e-72ee81aed0b6/5acb27b3-62c5
-46ac-8978-576a8a4a0399.lease
-rw-r--r--. 1 vdsm kvm   283 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/images/bb9d9a37-4f91-4973-ba9e-72ee81aed0b6/5acb27b3-62c5
-46ac-8978-576a8a4a0399.meta
-rw-r--r--. 1 vdsm kvm   384 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
-092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.job.0
-rw-r--r--. 1 vdsm kvm   277 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
-092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.recover.0
-rw-r--r--. 1 vdsm kvm   417 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
-092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.recover.1
-rw-r--r--. 1 vdsm kvm   107 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
-092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.result
-rw-r--r--. 1 vdsm kvm   299 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
-7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
-092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.task
-rwxr-xr-x. 1 vdsm kvm 0 Jul  2 11:22 ./__DIRECT_IO_TEST__
-



On Thu, 2015-07-02 at 11:05 -0400, Christopher Young wrote:
 That's actually a local NFS implementation (running on Gluster) so I
 wouldn't expect that it would factor in, but I just verified that all
 gluster hosts are in Permissive 

Re: [ovirt-users] Confused / Hosted-Engine won't start

2015-07-02 Thread Christopher Young
Ok.  It turns out that I had a split-brain on that volume file and
didn't even realize it.  Thank both of you guys for keeping me thinking
and putting me on track.  I had so many thoughts running through my
head and had run 'gluster volume status engine' (etc.) without checking
for heal status.

In the process, I learned more about glusters healing options, so I
suppose I'm going to put this all in the category of 'learning
experience'.

I believe that I eventually want to build out a new hosted-engine VM
since this one had its configuration changed a few times and getting an
additional, redundant hosted-engine fails (primarily due to me missing
the original answers file).  I had tried recreating it using my limited
knowledge, but I could never get the 2nd hosted-engine deploy to work.

If anyone has any advise on that, I'd be appreciative and open to it. 
 I'm debating either continuing with this hosted-engine VM or backing
things up and building a new one (though I don't know the optimal
procedure for that).  Any direction is always helpful.

Thanks again!

On Thu, 2015-07-02 at 11:40 -0500, Darrell Budic wrote:
 Looks normal, hosted engine uses some extra files/leases to track 
 some of it’s stuff.
 
 Looks like you might have a gluster problem though, that IO error 
 appears to be on your hosted engines disk image. Check for split 
 brains and try and initiate a heal on the files, see what you get.
 
   -Darrell
 
  On Jul 2, 2015, at 11:33 AM, Christopher Young 
  mexigaba...@gmail.com wrote:
  
  If anyone has an experience of the various IDs in hosted
  -engine.conf
  and vm.conf for the Hosted Engine, I believe I need to just verify
  everything.  I tried a couple of changes, but I feel like I'm just
  making this worse, so I've reverted them.
  
  One thing I do not understand well is how a gluster-based (NFS) 
  storage
  domain for the hosted-engine has so many entries:
  
  -
  
  [root@orldc-dev-vnode02 ovirt-gluster-nfs:_engine]# find . -type f 
  |
  xargs ls -lah
  ls: cannot access ./4e3017eb-d062-4ad1-9df8
  -7057fcee412c/images/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821
  -a432
  -4df6-8856-fdb14df260e3: Input/output error
  -rw-rw. 1 vdsm kvm  1.0M Jul  2 11:20 ./4e3017eb-d062-4ad1-9df8
  -7057fcee412c/dom_md/ids
  -rw-rw. 1 vdsm kvm   16M Jul  1 19:54 ./4e3017eb-d062-4ad1-9df8
  -7057fcee412c/dom_md/inbox
  -rw-rw. 1 vdsm kvm  2.0M Jul  1 19:50 ./4e3017eb-d062-4ad1-9df8
  -7057fcee412c/dom_md/leases
  -rw-r--r--. 1 vdsm kvm   482 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
  -7057fcee412c/dom_md/metadata
  -rw-rw. 1 vdsm kvm   16M Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
  -7057fcee412c/dom_md/outbox
  -rw-rw. 1 vdsm kvm  1.0M Jul  2 11:32 ./4e3017eb-d062-4ad1-9df8
  -7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf/d9bacbf6
  -c2f4
  -4f74-a91f-3a3a52f255bf
  -rw-rw. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
  -7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf/d9bacbf6
  -c2f4
  -4f74-a91f-3a3a52f255bf.lease
  -rw-r--r--. 1 vdsm kvm   284 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
  -7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf/d9bacbf6
  -c2f4
  -4f74-a91f-3a3a52f255bf.meta
  -rw-rw. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
  -7057fcee412c/images/23ac8897-b0c7-41d6-a7de-19f46ed78400/dd9f373c
  -d161
  -4fa0-aab1-3cb52305dba7.lease
  -rw-r--r--. 1 vdsm kvm   283 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
  -7057fcee412c/images/23ac8897-b0c7-41d6-a7de-19f46ed78400/dd9f373c
  -d161
  -4fa0-aab1-3cb52305dba7.meta
  -rw-rw. 1 vdsm kvm   25G Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
  -7057fcee412c/images/3278c444-d92a-4cb9-87d6-9669c6e4993e/1a4b6a5d
  -9c8f
  -4d54-91a7-3dd87377c362
  -rw-rw. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
  -7057fcee412c/images/3278c444-d92a-4cb9-87d6-9669c6e4993e/1a4b6a5d
  -9c8f
  -4d54-91a7-3dd87377c362.lease
  -rw-r--r--. 1 vdsm kvm   278 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
  -7057fcee412c/images/3278c444-d92a-4cb9-87d6-9669c6e4993e/1a4b6a5d
  -9c8f
  -4d54-91a7-3dd87377c362.meta
  -rw-rw. 1 vdsm kvm  1.0M Jul  1 17:50 ./4e3017eb-d062-4ad1-9df8
  -7057fcee412c/images/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821
  -a432
  -4df6-8856-fdb14df260e3.lease
  -rw-r--r--. 1 vdsm kvm   278 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
  -7057fcee412c/images/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821
  -a432
  -4df6-8856-fdb14df260e3.meta
  -rw-rw. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
  -7057fcee412c/images/6064179f-2720-4db9-a7c4-a97e044c2238/05afaa26
  -95af
  -4226-9a00-5383d8937a83
  -rw-rw. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
  -7057fcee412c/images/6064179f-2720-4db9-a7c4-a97e044c2238/05afaa26
  -95af
  -4226-9a00-5383d8937a83.lease
  -rw-r--r--. 1 vdsm kvm   284 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
  -7057fcee412c/images/6064179f-2720-4db9-a7c4-a97e044c2238/05afaa26
  -95af
  -4226-9a00-5383d8937a83.meta
  

Re: [ovirt-users] Confused / Hosted-Engine won't start

2015-07-02 Thread Darrell Budic
Looks normal, hosted engine uses some extra files/leases to track some of it’s 
stuff.

Looks like you might have a gluster problem though, that IO error appears to be 
on your hosted engines disk image. Check for split brains and try and initiate 
a heal on the files, see what you get.

  -Darrell

 On Jul 2, 2015, at 11:33 AM, Christopher Young mexigaba...@gmail.com wrote:
 
 If anyone has an experience of the various IDs in hosted-engine.conf
 and vm.conf for the Hosted Engine, I believe I need to just verify
 everything.  I tried a couple of changes, but I feel like I'm just
 making this worse, so I've reverted them.
 
 One thing I do not understand well is how a gluster-based (NFS) storage
 domain for the hosted-engine has so many entries:
 
 -
 
 [root@orldc-dev-vnode02 ovirt-gluster-nfs:_engine]# find . -type f |
 xargs ls -lah
 ls: cannot access ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/images/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432
 -4df6-8856-fdb14df260e3: Input/output error
 -rw-rw. 1 vdsm kvm  1.0M Jul  2 11:20 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/dom_md/ids
 -rw-rw. 1 vdsm kvm   16M Jul  1 19:54 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/dom_md/inbox
 -rw-rw. 1 vdsm kvm  2.0M Jul  1 19:50 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/dom_md/leases
 -rw-r--r--. 1 vdsm kvm   482 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/dom_md/metadata
 -rw-rw. 1 vdsm kvm   16M Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/dom_md/outbox
 -rw-rw. 1 vdsm kvm  1.0M Jul  2 11:32 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf/d9bacbf6-c2f4
 -4f74-a91f-3a3a52f255bf
 -rw-rw. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf/d9bacbf6-c2f4
 -4f74-a91f-3a3a52f255bf.lease
 -rw-r--r--. 1 vdsm kvm   284 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf/d9bacbf6-c2f4
 -4f74-a91f-3a3a52f255bf.meta
 -rw-rw. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/images/23ac8897-b0c7-41d6-a7de-19f46ed78400/dd9f373c-d161
 -4fa0-aab1-3cb52305dba7.lease
 -rw-r--r--. 1 vdsm kvm   283 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/images/23ac8897-b0c7-41d6-a7de-19f46ed78400/dd9f373c-d161
 -4fa0-aab1-3cb52305dba7.meta
 -rw-rw. 1 vdsm kvm   25G Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/images/3278c444-d92a-4cb9-87d6-9669c6e4993e/1a4b6a5d-9c8f
 -4d54-91a7-3dd87377c362
 -rw-rw. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/images/3278c444-d92a-4cb9-87d6-9669c6e4993e/1a4b6a5d-9c8f
 -4d54-91a7-3dd87377c362.lease
 -rw-r--r--. 1 vdsm kvm   278 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/images/3278c444-d92a-4cb9-87d6-9669c6e4993e/1a4b6a5d-9c8f
 -4d54-91a7-3dd87377c362.meta
 -rw-rw. 1 vdsm kvm  1.0M Jul  1 17:50 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/images/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432
 -4df6-8856-fdb14df260e3.lease
 -rw-r--r--. 1 vdsm kvm   278 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/images/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432
 -4df6-8856-fdb14df260e3.meta
 -rw-rw. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/images/6064179f-2720-4db9-a7c4-a97e044c2238/05afaa26-95af
 -4226-9a00-5383d8937a83
 -rw-rw. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/images/6064179f-2720-4db9-a7c4-a97e044c2238/05afaa26-95af
 -4226-9a00-5383d8937a83.lease
 -rw-r--r--. 1 vdsm kvm   284 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/images/6064179f-2720-4db9-a7c4-a97e044c2238/05afaa26-95af
 -4226-9a00-5383d8937a83.meta
 -rw-rw. 1 vdsm kvm 1004K Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/images/bb9d9a37-4f91-4973-ba9e-72ee81aed0b6/5acb27b3-62c5
 -46ac-8978-576a8a4a0399
 -rw-rw. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/images/bb9d9a37-4f91-4973-ba9e-72ee81aed0b6/5acb27b3-62c5
 -46ac-8978-576a8a4a0399.lease
 -rw-r--r--. 1 vdsm kvm   283 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/images/bb9d9a37-4f91-4973-ba9e-72ee81aed0b6/5acb27b3-62c5
 -46ac-8978-576a8a4a0399.meta
 -rw-r--r--. 1 vdsm kvm   384 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
 -092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.job.0
 -rw-r--r--. 1 vdsm kvm   277 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
 -092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.recover.0
 -rw-r--r--. 1 vdsm kvm   417 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
 -092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.recover.1
 -rw-r--r--. 1 vdsm kvm   107 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
 -7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
 -092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.result
 -rw-r--r--. 1 vdsm kvm   

[ovirt-users] Confused / Hosted-Engine won't start

2015-07-01 Thread Christopher Young
I'm sure I have worked through this before, but I've been banging my head
against this one for a while now, and I think I'm too close to the issue.

Basically, my hosted-engine won't start anymore.  I do recall attempting to
migrate it to a new gluster-based NFS share recently, but I could have
sworn that was successful and working.  I _think_ I have some sort of id
issue with storage/volume/whatever id's, but I need some help digging
through it if someone would be so kind.

I have the following error in /var/log/libvirt/qemu/HostedEngine.log:

--
2015-07-02T00:01:13.080952Z qemu-kvm: -drive
file=/var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8-7057fcee412c/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432-4df6-8856-fdb14df260e3,if=none,id=drive-virtio-disk0,format=raw,serial=5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836,cache=none,werror=stop,rerror=stop,aio=threads:
could not open disk image
/var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8-7057fcee412c/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432-4df6-8856-fdb14df260e3:
Could not refresh total sector count: Input/output error
--

I also am including a few command outputs that I'm sure might help:

--
[root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8-7057fcee412c]# ls -la
/var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8-7057fcee412c/
total 8
drwxr-xr-x. 2 vdsm kvm  80 Jul  1 20:04 .
drwxr-xr-x. 3 vdsm kvm  60 Jul  1 20:04 ..
lrwxrwxrwx. 1 vdsm kvm 128 Jul  1 20:04
1d80a60c-8f26-4448-9460-2c7b00ff75bf -
/rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062-4ad1-9df8-7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf
lrwxrwxrwx. 1 vdsm kvm 128 Jul  1 20:04
23ac8897-b0c7-41d6-a7de-19f46ed78400 -
/rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062-4ad1-9df8-7057fcee412c/images/23ac8897-b0c7-41d6-a7de-19f46ed78400

--
[root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8-7057fcee412c]# cat
/etc/ovirt-hosted-engine/vm.conf
vmId=6b7329f9-518a-4488-b1c4-2cd809f2f580
memSize=5120
display=vnc
devices={index:2,iface:ide,address:{ controller:0, target:0,unit:0, bus:1,
type:drive},specParams:{},readonly:true,deviceId:77924fc2-c5c9-408b-97d3-cd0b0d11a62c,path:/home/tmp/CentOS-6.6-x86_64-minimal.iso,device:cdrom,shared:false,type:disk}
devices={index:0,iface:virtio,format:raw,poolID:----,volumeID:eeb2d821-a432-4df6-8856-fdb14df260e3,imageID:5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836,specParams:{},readonly:false,domainID:4e3017eb-d062-4ad1-9df8-7057fcee412c,optional:false,deviceId:5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836,address:{bus:0x00,
slot:0x06, domain:0x, type:pci,
function:0x0},device:disk,shared:exclusive,propagateErrors:off,type:disk,bootOrder:1}
devices={device:scsi,model:virtio-scsi,type:controller}
devices={nicModel:pv,macAddr:00:16:3e:0e:d0:68,linkActive:true,network:ovirtmgmt,filter:vdsm-no-mac-spoofing,specParams:{},deviceId:f70ba622-6ac8-4c06-a005-0ebd940a15b2,address:{bus:0x00,
slot:0x03, domain:0x, type:pci,
function:0x0},device:bridge,type:interface}
devices={device:console,specParams:{},type:console,deviceId:98386e6c-ae56-4b6d-9bfb-c72bbd299ad1,alias:console0}
vmName=HostedEngine
spiceSecureChannels=smain,sdisplay,sinputs,scursor,splayback,srecord,ssmartcard,susbredir
smp=2
cpuType=Westmere
emulatedMachine=pc
--

[root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8-7057fcee412c]# cat
/etc/ovirt-hosted-engine/hosted-engine.conf
fqdn=orldc-dev-vengine01.***
vm_disk_id=5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836
vmid=6b7329f9-518a-4488-b1c4-2cd809f2f580
storage=ovirt-gluster-nfs:/engine
conf=/etc/ovirt-hosted-engine/vm.conf
service_start_time=0
host_id=1
console=vnc
domainType=nfs3
spUUID=379cf161-d5b1-4c20-bb71-e3ca5d2ccd6b
sdUUID=4e3017eb-d062-4ad1-9df8-7057fcee412c
connectionUUID=0d1b50ac-cf3f-4cd7-90df-3c3a6d11a984
ca_cert=/etc/pki/vdsm/libvirt-spice/ca-cert.pem
ca_subject=C=EN, L=Test, O=Test, CN=Test
vdsm_use_ssl=true
gateway=10.16.3.1
bridge=ovirtmgmt
metadata_volume_UUID=dd9f373c-d161-4fa0-aab1-3cb52305dba7
metadata_image_UUID=23ac8897-b0c7-41d6-a7de-19f46ed78400
lockspace_volume_UUID=d9bacbf6-c2f4-4f74-a91f-3a3a52f255bf
lockspace_image_UUID=1d80a60c-8f26-4448-9460-2c7b00ff75bf

# The following are used only for iSCSI storage
iqn=
portal=
user=
password=
port=
--

(mount output for the NFS share this should be running on (gluster-based):

ovirt-gluster-nfs:/engine on
/rhev/data-center/mnt/ovirt-gluster-nfs:_engine type nfs
(rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,soft,nolock,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mountaddr=10.16.3.30,mountvers=3,mountport=38465,mountproto=tcp,local_lock=all,addr=10.16.3.30)

--

[root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8-7057fcee412c]# tree
/rhev/data-center/mnt/ovirt-gluster-nfs\:_engine/
/rhev/data-center/mnt/ovirt-gluster-nfs:_engine/
├── 4e3017eb-d062-4ad1-9df8-7057fcee412c
│   ├── dom_md
│   │   ├── ids
│   │   ├── inbox
│   │   ├── leases
│   │   ├── metadata
│   │   └── outbox
│   ├──