Looks normal, hosted engine uses some extra files/leases to track some of it’s 
stuff.

Looks like you might have a gluster problem though, that IO error appears to be 
on your hosted engines disk image. Check for split brains and try and initiate 
a heal on the files, see what you get.

  -Darrell

> On Jul 2, 2015, at 11:33 AM, Christopher Young <mexigaba...@gmail.com> wrote:
> 
> If anyone has an experience of the various IDs in hosted-engine.conf
> and vm.conf for the Hosted Engine, I believe I need to just verify
> everything.  I tried a couple of changes, but I feel like I'm just
> making this worse, so I've reverted them.
> 
> One thing I do not understand well is how a gluster-based (NFS) storage
> domain for the hosted-engine has so many entries:
> 
> -------------
> 
> [root@orldc-dev-vnode02 ovirt-gluster-nfs:_engine]# find . -type f |
> xargs ls -lah
> ls: cannot access ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/images/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432
> -4df6-8856-fdb14df260e3: Input/output error
> -rw-rw----. 1 vdsm kvm  1.0M Jul  2 11:20 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/dom_md/ids
> -rw-rw----. 1 vdsm kvm   16M Jul  1 19:54 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/dom_md/inbox
> -rw-rw----. 1 vdsm kvm  2.0M Jul  1 19:50 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/dom_md/leases
> -rw-r--r--. 1 vdsm kvm   482 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/dom_md/metadata
> -rw-rw----. 1 vdsm kvm   16M Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/dom_md/outbox
> -rw-rw----. 1 vdsm kvm  1.0M Jul  2 11:32 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf/d9bacbf6-c2f4
> -4f74-a91f-3a3a52f255bf
> -rw-rw----. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf/d9bacbf6-c2f4
> -4f74-a91f-3a3a52f255bf.lease
> -rw-r--r--. 1 vdsm kvm   284 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf/d9bacbf6-c2f4
> -4f74-a91f-3a3a52f255bf.meta
> -rw-rw----. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/images/23ac8897-b0c7-41d6-a7de-19f46ed78400/dd9f373c-d161
> -4fa0-aab1-3cb52305dba7.lease
> -rw-r--r--. 1 vdsm kvm   283 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/images/23ac8897-b0c7-41d6-a7de-19f46ed78400/dd9f373c-d161
> -4fa0-aab1-3cb52305dba7.meta
> -rw-rw----. 1 vdsm kvm   25G Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/images/3278c444-d92a-4cb9-87d6-9669c6e4993e/1a4b6a5d-9c8f
> -4d54-91a7-3dd87377c362
> -rw-rw----. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/images/3278c444-d92a-4cb9-87d6-9669c6e4993e/1a4b6a5d-9c8f
> -4d54-91a7-3dd87377c362.lease
> -rw-r--r--. 1 vdsm kvm   278 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/images/3278c444-d92a-4cb9-87d6-9669c6e4993e/1a4b6a5d-9c8f
> -4d54-91a7-3dd87377c362.meta
> -rw-rw----. 1 vdsm kvm  1.0M Jul  1 17:50 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/images/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432
> -4df6-8856-fdb14df260e3.lease
> -rw-r--r--. 1 vdsm kvm   278 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/images/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432
> -4df6-8856-fdb14df260e3.meta
> -rw-rw----. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/images/6064179f-2720-4db9-a7c4-a97e044c2238/05afaa26-95af
> -4226-9a00-5383d8937a83
> -rw-rw----. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/images/6064179f-2720-4db9-a7c4-a97e044c2238/05afaa26-95af
> -4226-9a00-5383d8937a83.lease
> -rw-r--r--. 1 vdsm kvm   284 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/images/6064179f-2720-4db9-a7c4-a97e044c2238/05afaa26-95af
> -4226-9a00-5383d8937a83.meta
> -rw-rw----. 1 vdsm kvm 1004K Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/images/bb9d9a37-4f91-4973-ba9e-72ee81aed0b6/5acb27b3-62c5
> -46ac-8978-576a8a4a0399
> -rw-rw----. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/images/bb9d9a37-4f91-4973-ba9e-72ee81aed0b6/5acb27b3-62c5
> -46ac-8978-576a8a4a0399.lease
> -rw-r--r--. 1 vdsm kvm   283 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/images/bb9d9a37-4f91-4973-ba9e-72ee81aed0b6/5acb27b3-62c5
> -46ac-8978-576a8a4a0399.meta
> -rw-r--r--. 1 vdsm kvm   384 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
> -092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.job.0
> -rw-r--r--. 1 vdsm kvm   277 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
> -092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.recover.0
> -rw-r--r--. 1 vdsm kvm   417 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
> -092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.recover.1
> -rw-r--r--. 1 vdsm kvm   107 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
> -092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.result
> -rw-r--r--. 1 vdsm kvm   299 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
> -7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
> -092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.task
> -rwxr-xr-x. 1 vdsm kvm     0 Jul  2 11:22 ./__DIRECT_IO_TEST__
> -------------
> 
> 
> 
> On Thu, 2015-07-02 at 11:05 -0400, Christopher Young wrote:
>> That's actually a local NFS implementation (running on Gluster) so I
>> wouldn't expect that it would factor in, but I just verified that all
>> gluster hosts are in Permissive mode.  Like I mentioned, I think I 
>> may
>> have some ids wrong or bad links or something since this had 
>> previously
>> been working fine, though I must be missing it. 
>> 
>> On Thu, 2015-07-02 at 16:08 +0200, Sandro Bonazzola wrote:
>>> Il 02/07/2015 16:04, Christopher Young ha scritto:
>>>> [root@orldc-dev-vnode02 ~]# getenforce 
>>>> Permissive
>>>> 
>>>> It looks like it isn't SELinux.
>>> 
>>> Checked also on the storage server ovirt-gluster-nfs?
>>> 
>>> 
>>>> 
>>>> On Thu, 2015-07-02 at 09:53 +0200, Sandro Bonazzola wrote:
>>>>> Il 02/07/2015 02:36, Christopher Young ha scritto:
>>>>>> I'm sure I have worked through this before, but I've been 
>>>>>> banging 
>>>>>> my head against this one for a while now, and I think I'm too 
>>>>>> 
>>>>>> close 
>>>>>> to the issue.
>>>>>> 
>>>>>> Basically, my hosted-engine won't start anymore.  I do recall 
>>>>>> 
>>>>>> attempting to migrate it to a new gluster-based NFS share 
>>>>>> recently, 
>>>>>> but I could have
>>>>>> sworn that was successful and working.  I _think_ I have some 
>>>>>> 
>>>>>> sort 
>>>>>> of id issue with storage/volume/whatever id's, but I need 
>>>>>> some 
>>>>>> help 
>>>>>> digging through
>>>>>> it if someone would be so kind.
>>>>>> 
>>>>>> I have the following error in 
>>>>>> /var/log/libvirt/qemu/HostedEngine.log:
>>>>>> 
>>>>>> ----------
>>>>>> 2015-07-02T00:01:13.080952Z qemu-kvm: -drive
>>>>>> file=/var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8
>>>>>> -7057fcee412c/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821
>>>>>> -a432
>>>>>> -4df6-8856-fdb14df260e3,if=none,id=drive-virtio
>>>>>> -disk0,format=raw,serial=5ead7b5d-50e8-4d6c-a0e5
>>>>>> -bbe6d93dd836,cache=none,werror=stop,rerror=stop,aio=threads:
>>>>>> could not open disk image
>>>>>> /var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8
>>>>>> -7057fcee412c/5ead7b5d
>>>>>> -50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432-4df6-8856
>>>>>> -fdb14df260e3: 
>>>>>> Could not
>>>>>> refresh total sector count: Input/output error
>>>>>> ----------
>>>>> 
>>>>> 
>>>>> Please check selinux: ausearch -m avc
>>>>> it might be selinux preventing access to the disk image.
>>>>> 
>>>>> 
>>>>> 
>>>>>> 
>>>>>> I also am including a few command outputs that I'm sure might 
>>>>>> 
>>>>>> help:
>>>>>> 
>>>>>> ----------
>>>>>> [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8
>>>>>> -7057fcee412c]# 
>>>>>> ls 
>>>>>> -la /var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8
>>>>>> -7057fcee412c/
>>>>>> total 8
>>>>>> drwxr-xr-x. 2 vdsm kvm  80 Jul  1 20:04 .
>>>>>> drwxr-xr-x. 3 vdsm kvm  60 Jul  1 20:04 ..
>>>>>> lrwxrwxrwx. 1 vdsm kvm 128 Jul  1 20:04 1d80a60c-8f26-4448
>>>>>> -9460
>>>>>> -2c7b00ff75bf ->
>>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062
>>>>>> -4ad1
>>>>>> -9df8-7057fcee412c/images/1d80a60c-8f26-4448-9460
>>>>>> -2c7b00ff75bf
>>>>>> lrwxrwxrwx. 1 vdsm kvm 128 Jul  1 20:04 23ac8897-b0c7-41d6
>>>>>> -a7de
>>>>>> -19f46ed78400 ->
>>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062
>>>>>> -4ad1
>>>>>> -9df8-7057fcee412c/images/23ac8897-b0c7-41d6-a7de
>>>>>> -19f46ed78400
>>>>>> 
>>>>>> ----------
>>>>>> [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8
>>>>>> -7057fcee412c]# 
>>>>>> cat 
>>>>>> /etc/ovirt-hosted-engine/vm.conf
>>>>>> vmId=6b7329f9-518a-4488-b1c4-2cd809f2f580
>>>>>> memSize=5120
>>>>>> display=vnc
>>>>>> devices={index:2,iface:ide,address:{ controller:0, 
>>>>>> target:0,unit:0, 
>>>>>> bus:1,
>>>>>> type:drive},specParams:{},readonly:true,deviceId:77924fc2
>>>>>> -c5c9
>>>>>> -408b
>>>>>> -97d3-cd0b0d11a62c,path:/home/tmp/CentOS-6.6-x86_64
>>>>>> -minimal.iso,device:cdrom,shared:false,type:disk}
>>>>>> devices={index:0,iface:virtio,format:raw,poolID:00000000-0000
>>>>>> -0000
>>>>>> -0000-000000000000,volumeID:eeb2d821-a432-4df6-8856
>>>>>> -fdb14df260e3,imageID:5ead7b5d-50e8-4d6c-a0e5
>>>>>> -bbe6d93dd836,specParams:{},readonly:false,domainID:4e3017eb
>>>>>> -d062
>>>>>> -4ad1-9df8-7057fcee412c,optional:false,deviceId:5ead7b5d-50e8
>>>>>> -4d6c
>>>>>> -a0e5-bbe6d93dd836,address:{bus:0x00,
>>>>>> slot:0x06, domain:0x0000, type:pci, 
>>>>>> function:0x0},device:disk,shared:exclusive,propagateErrors:of
>>>>>> f,
>>>>>> type
>>>>>> :disk,bootOrder:1}
>>>>>> devices={device:scsi,model:virtio-scsi,type:controller}
>>>>>> devices={nicModel:pv,macAddr:00:16:3e:0e:d0:68,linkActive:tru
>>>>>> e,
>>>>>> netw
>>>>>> ork:ovirtmgmt,filter:vdsm-no-mac
>>>>>> -spoofing,specParams:{},deviceId:f70ba622-6ac8-4c06-a005
>>>>>> -0ebd940a15b2,address:{bus:0x00,
>>>>>> slot:0x03, domain:0x0000, type:pci, 
>>>>>> function:0x0},device:bridge,type:interface}
>>>>>> devices={device:console,specParams:{},type:console,deviceId:9
>>>>>> 83
>>>>>> 86e6
>>>>>> c-ae56-4b6d-9bfb-c72bbd299ad1,alias:console0}
>>>>>> vmName=HostedEngine
>>>>>> spiceSecureChannels=smain,sdisplay,sinputs,scursor,splayback,
>>>>>> sr
>>>>>> ecor
>>>>>> d,ssmartcard,susbredir
>>>>>> smp=2
>>>>>> cpuType=Westmere
>>>>>> emulatedMachine=pc
>>>>>> ----------
>>>>>> 
>>>>>> [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8
>>>>>> -7057fcee412c]# 
>>>>>> cat 
>>>>>> /etc/ovirt-hosted-engine/hosted-engine.conf
>>>>>> fqdn=orldc-dev-vengine01.*******
>>>>>> vm_disk_id=5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836
>>>>>> vmid=6b7329f9-518a-4488-b1c4-2cd809f2f580
>>>>>> storage=ovirt-gluster-nfs:/engine
>>>>>> conf=/etc/ovirt-hosted-engine/vm.conf
>>>>>> service_start_time=0
>>>>>> host_id=1
>>>>>> console=vnc
>>>>>> domainType=nfs3
>>>>>> spUUID=379cf161-d5b1-4c20-bb71-e3ca5d2ccd6b
>>>>>> sdUUID=4e3017eb-d062-4ad1-9df8-7057fcee412c
>>>>>> connectionUUID=0d1b50ac-cf3f-4cd7-90df-3c3a6d11a984
>>>>>> ca_cert=/etc/pki/vdsm/libvirt-spice/ca-cert.pem
>>>>>> ca_subject="C=EN, L=Test, O=Test, CN=Test"
>>>>>> vdsm_use_ssl=true
>>>>>> gateway=10.16.3.1
>>>>>> bridge=ovirtmgmt
>>>>>> metadata_volume_UUID=dd9f373c-d161-4fa0-aab1-3cb52305dba7
>>>>>> metadata_image_UUID=23ac8897-b0c7-41d6-a7de-19f46ed78400
>>>>>> lockspace_volume_UUID=d9bacbf6-c2f4-4f74-a91f-3a3a52f255bf
>>>>>> lockspace_image_UUID=1d80a60c-8f26-4448-9460-2c7b00ff75bf
>>>>>> 
>>>>>> # The following are used only for iSCSI storage
>>>>>> iqn=
>>>>>> portal=
>>>>>> user=
>>>>>> password=
>>>>>> port=
>>>>>> ----------
>>>>>> 
>>>>>> (mount output for the NFS share this should be running on 
>>>>>> (gluster
>>>>>> -based):
>>>>>> 
>>>>>> ovirt-gluster-nfs:/engine on /rhev/data-center/mnt/ovirt
>>>>>> -gluster
>>>>>> -nfs:_engine type nfs
>>>>>> (rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so
>>>>>> ft
>>>>>> ,nol
>>>>>> ock,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta
>>>>>> dd
>>>>>> r=10
>>>>>> .16.3.30,mountvers=3,mountport=38465,mountproto=tcp,local_loc
>>>>>> k=
>>>>>> all,
>>>>>> addr=10.16.3.30)
>>>>>> 
>>>>>> ----------
>>>>>> 
>>>>>> [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8
>>>>>> -7057fcee412c]# 
>>>>>> tree 
>>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs\:_engine/
>>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/
>>>>>> ├── 4e3017eb-d062-4ad1-9df8-7057fcee412c
>>>>>> │   ├── dom_md
>>>>>> │   │   ├── ids
>>>>>> │   │   ├── inbox
>>>>>> │   │   ├── leases
>>>>>> │   │   ├── metadata
>>>>>> │   │   └── outbox
>>>>>> │   ├── ha_agent
>>>>>> │   │   ├── hosted-engine.lockspace ->
>>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062
>>>>>> -4ad1
>>>>>> -9df8-7057fcee412c/images/1d80a60c-8f26-4448-9460
>>>>>> -2c7b00ff75bf/d9bacbf6-c2f4-4f74-a91f-3a3a52f255bf
>>>>>> │   │   └── hosted-engine.metadata ->
>>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062
>>>>>> -4ad1
>>>>>> -9df8-7057fcee412c/images/23ac8897-b0c7-41d6-a7de
>>>>>> -19f46ed78400/dd9f373c-d161-4fa0-aab1-3cb52305dba7
>>>>>> │   ├── images
>>>>>> │   │   ├── 1d80a60c-8f26-4448-9460-2c7b00ff75bf
>>>>>> │   │   │   ├── d9bacbf6-c2f4-4f74-a91f-3a3a52f255bf
>>>>>> │   │   │   ├── d9bacbf6-c2f4-4f74-a91f-3a3a52f255bf.lease
>>>>>> │   │   │   └── d9bacbf6-c2f4-4f74-a91f-3a3a52f255bf.meta
>>>>>> │   │   ├── 23ac8897-b0c7-41d6-a7de-19f46ed78400
>>>>>> │   │   │   ├── dd9f373c-d161-4fa0-aab1-3cb52305dba7 ->
>>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062
>>>>>> -4ad1
>>>>>> -9df8-7057fcee412c/images/23ac8897-b0c7-41d6-a7de
>>>>>> -19f46ed78400
>>>>>> │   │   │   ├── dd9f373c-d161-4fa0-aab1-3cb52305dba7.lease
>>>>>> │   │   │   └── dd9f373c-d161-4fa0-aab1-3cb52305dba7.meta
>>>>>> │   │   ├── 3278c444-d92a-4cb9-87d6-9669c6e4993e
>>>>>> │   │   │   ├── 1a4b6a5d-9c8f-4d54-91a7-3dd87377c362
>>>>>> │   │   │   ├── 1a4b6a5d-9c8f-4d54-91a7-3dd87377c362.lease
>>>>>> │   │   │   └── 1a4b6a5d-9c8f-4d54-91a7-3dd87377c362.meta
>>>>>> │   │   ├── 5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836
>>>>>> │   │   │   ├── eeb2d821-a432-4df6-8856-fdb14df260e3.lease
>>>>>> │   │   │   └── eeb2d821-a432-4df6-8856-fdb14df260e3.meta
>>>>>> │   │   ├── 6064179f-2720-4db9-a7c4-a97e044c2238
>>>>>> │   │   │   ├── 05afaa26-95af-4226-9a00-5383d8937a83
>>>>>> │   │   │   ├── 05afaa26-95af-4226-9a00-5383d8937a83.lease
>>>>>> │   │   │   └── 05afaa26-95af-4226-9a00-5383d8937a83.meta
>>>>>> │   │   └── bb9d9a37-4f91-4973-ba9e-72ee81aed0b6
>>>>>> │   │       ├── 5acb27b3-62c5-46ac-8978-576a8a4a0399
>>>>>> │   │       ├── 5acb27b3-62c5-46ac-8978-576a8a4a0399.lease
>>>>>> │   │       └── 5acb27b3-62c5-46ac-8978-576a8a4a0399.meta
>>>>>> │   └── master
>>>>>> │       ├── tasks
>>>>>> │       │   └── fef13299-0e7f-4c7a-a399-092a1235faab
>>>>>> │       │       ├── fef13299-0e7f-4c7a-a399
>>>>>> -092a1235faab.job.0
>>>>>> │       │       ├── fef13299-0e7f-4c7a-a399
>>>>>> -092a1235faab.recover.0
>>>>>> │       │       ├── fef13299-0e7f-4c7a-a399
>>>>>> -092a1235faab.recover.1
>>>>>> │       │       ├── fef13299-0e7f-4c7a-a399
>>>>>> -092a1235faab.result
>>>>>> │       │       └── fef13299-0e7f-4c7a-a399-092a1235faab.task
>>>>>> │       └── vms
>>>>>> └── __DIRECT_IO_TEST__
>>>>>> 
>>>>>> 16 directories, 28 files
>>>>>> ----------
>>>>>> 
>>>>>> [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8
>>>>>> -7057fcee412c]# 
>>>>>> tree 
>>>>>> /var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8-7057fcee412c
>>>>>> /var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8-7057fcee412c
>>>>>> ├── 1d80a60c-8f26-4448-9460-2c7b00ff75bf ->
>>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062
>>>>>> -4ad1
>>>>>> -9df8-7057fcee412c/images/1d80a60c-8f26-4448-9460
>>>>>> -2c7b00ff75bf
>>>>>> └── 23ac8897-b0c7-41d6-a7de-19f46ed78400 ->
>>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062
>>>>>> -4ad1
>>>>>> -9df8-7057fcee412c/images/23ac8897-b0c7-41d6-a7de
>>>>>> -19f46ed78400
>>>>>> 
>>>>>> 2 directories, 0 files
>>>>>> ----------
>>>>>> 
>>>>>> 
>>>>>> Can you anyone point me in the right direction here? 
>>>>>> Eventually, 
>>>>>> I'd like to get this hosted-engine running in HA mode (I had 
>>>>>> started this process
>>>>>> with ctdb/gluster/etc. via the well-known guide out there), 
>>>>>> but 
>>>>>> I'm 
>>>>>> thinking that at this point, I should get this up and running 
>>>>>> 
>>>>>> again, export
>>>>>> everything, and perhaps just build a new hosted-engine and 
>>>>>> import 
>>>>>> things into (if that is possible).  Right now, though, I just 
>>>>>> 
>>>>>> need 
>>>>>> to get this
>>>>>> hosted-engine run.
>>>>>> 
>>>>>> Many many thanks!
>>>>>> 
>>>>>> Chris
>>>>>> 
>>>>>> 
>>>>>> _______________________________________________
>>>>>> Users mailing list
>>>>>> Users@ovirt.org
>>>>>> http://lists.ovirt.org/mailman/listinfo/users
>>>>>> 
>>>>> 
>>>>> 
>>> 
>>> 
> _______________________________________________
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users

_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to