Ok.  It turns out that I had a split-brain on that volume file and
didn't even realize it.  Thank both of you guys for keeping me thinking
and putting me on track.  I had so many thoughts running through my
head and had run 'gluster volume status engine' (etc.) without checking
for heal status.

In the process, I learned more about glusters healing options, so I
suppose I'm going to put this all in the category of 'learning
experience'.

I believe that I eventually want to build out a new hosted-engine VM
since this one had its configuration changed a few times and getting an
additional, redundant hosted-engine fails (primarily due to me missing
the original answers file).  I had tried recreating it using my limited
knowledge, but I could never get the 2nd hosted-engine deploy to work.

If anyone has any advise on that, I'd be appreciative and open to it. 
 I'm debating either continuing with this hosted-engine VM or backing
things up and building a new one (though I don't know the optimal
procedure for that).  Any direction is always helpful.

Thanks again!

On Thu, 2015-07-02 at 11:40 -0500, Darrell Budic wrote:
> Looks normal, hosted engine uses some extra files/leases to track 
> some of it’s stuff.
> 
> Looks like you might have a gluster problem though, that IO error 
> appears to be on your hosted engines disk image. Check for split 
> brains and try and initiate a heal on the files, see what you get.
> 
>   -Darrell
> 
> > On Jul 2, 2015, at 11:33 AM, Christopher Young <
> > mexigaba...@gmail.com> wrote:
> > 
> > If anyone has an experience of the various IDs in hosted
> > -engine.conf
> > and vm.conf for the Hosted Engine, I believe I need to just verify
> > everything.  I tried a couple of changes, but I feel like I'm just
> > making this worse, so I've reverted them.
> > 
> > One thing I do not understand well is how a gluster-based (NFS) 
> > storage
> > domain for the hosted-engine has so many entries:
> > 
> > -------------
> > 
> > [root@orldc-dev-vnode02 ovirt-gluster-nfs:_engine]# find . -type f 
> > |
> > xargs ls -lah
> > ls: cannot access ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/images/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821
> > -a432
> > -4df6-8856-fdb14df260e3: Input/output error
> > -rw-rw----. 1 vdsm kvm  1.0M Jul  2 11:20 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/dom_md/ids
> > -rw-rw----. 1 vdsm kvm   16M Jul  1 19:54 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/dom_md/inbox
> > -rw-rw----. 1 vdsm kvm  2.0M Jul  1 19:50 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/dom_md/leases
> > -rw-r--r--. 1 vdsm kvm   482 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/dom_md/metadata
> > -rw-rw----. 1 vdsm kvm   16M Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/dom_md/outbox
> > -rw-rw----. 1 vdsm kvm  1.0M Jul  2 11:32 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf/d9bacbf6
> > -c2f4
> > -4f74-a91f-3a3a52f255bf
> > -rw-rw----. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf/d9bacbf6
> > -c2f4
> > -4f74-a91f-3a3a52f255bf.lease
> > -rw-r--r--. 1 vdsm kvm   284 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf/d9bacbf6
> > -c2f4
> > -4f74-a91f-3a3a52f255bf.meta
> > -rw-rw----. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/images/23ac8897-b0c7-41d6-a7de-19f46ed78400/dd9f373c
> > -d161
> > -4fa0-aab1-3cb52305dba7.lease
> > -rw-r--r--. 1 vdsm kvm   283 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/images/23ac8897-b0c7-41d6-a7de-19f46ed78400/dd9f373c
> > -d161
> > -4fa0-aab1-3cb52305dba7.meta
> > -rw-rw----. 1 vdsm kvm   25G Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/images/3278c444-d92a-4cb9-87d6-9669c6e4993e/1a4b6a5d
> > -9c8f
> > -4d54-91a7-3dd87377c362
> > -rw-rw----. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/images/3278c444-d92a-4cb9-87d6-9669c6e4993e/1a4b6a5d
> > -9c8f
> > -4d54-91a7-3dd87377c362.lease
> > -rw-r--r--. 1 vdsm kvm   278 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/images/3278c444-d92a-4cb9-87d6-9669c6e4993e/1a4b6a5d
> > -9c8f
> > -4d54-91a7-3dd87377c362.meta
> > -rw-rw----. 1 vdsm kvm  1.0M Jul  1 17:50 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/images/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821
> > -a432
> > -4df6-8856-fdb14df260e3.lease
> > -rw-r--r--. 1 vdsm kvm   278 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/images/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821
> > -a432
> > -4df6-8856-fdb14df260e3.meta
> > -rw-rw----. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/images/6064179f-2720-4db9-a7c4-a97e044c2238/05afaa26
> > -95af
> > -4226-9a00-5383d8937a83
> > -rw-rw----. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/images/6064179f-2720-4db9-a7c4-a97e044c2238/05afaa26
> > -95af
> > -4226-9a00-5383d8937a83.lease
> > -rw-r--r--. 1 vdsm kvm   284 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/images/6064179f-2720-4db9-a7c4-a97e044c2238/05afaa26
> > -95af
> > -4226-9a00-5383d8937a83.meta
> > -rw-rw----. 1 vdsm kvm 1004K Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/images/bb9d9a37-4f91-4973-ba9e-72ee81aed0b6/5acb27b3
> > -62c5
> > -46ac-8978-576a8a4a0399
> > -rw-rw----. 1 vdsm kvm  1.0M Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/images/bb9d9a37-4f91-4973-ba9e-72ee81aed0b6/5acb27b3
> > -62c5
> > -46ac-8978-576a8a4a0399.lease
> > -rw-r--r--. 1 vdsm kvm   283 Dec 23  2014 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/images/bb9d9a37-4f91-4973-ba9e-72ee81aed0b6/5acb27b3
> > -62c5
> > -46ac-8978-576a8a4a0399.meta
> > -rw-r--r--. 1 vdsm kvm   384 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
> > -092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.job.0
> > -rw-r--r--. 1 vdsm kvm   277 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
> > -092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.recover.0
> > -rw-r--r--. 1 vdsm kvm   417 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
> > -092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.recover.1
> > -rw-r--r--. 1 vdsm kvm   107 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
> > -092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.result
> > -rw-r--r--. 1 vdsm kvm   299 Jul  1 19:49 ./4e3017eb-d062-4ad1-9df8
> > -7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399
> > -092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.task
> > -rwxr-xr-x. 1 vdsm kvm     0 Jul  2 11:22 ./__DIRECT_IO_TEST__
> > -------------
> > 
> > 
> > 
> > On Thu, 2015-07-02 at 11:05 -0400, Christopher Young wrote:
> > > That's actually a local NFS implementation (running on Gluster) 
> > > so I
> > > wouldn't expect that it would factor in, but I just verified that 
> > > all
> > > gluster hosts are in Permissive mode.  Like I mentioned, I think 
> > > I 
> > > may
> > > have some ids wrong or bad links or something since this had 
> > > previously
> > > been working fine, though I must be missing it. 
> > > 
> > > On Thu, 2015-07-02 at 16:08 +0200, Sandro Bonazzola wrote:
> > > > Il 02/07/2015 16:04, Christopher Young ha scritto:
> > > > > [root@orldc-dev-vnode02 ~]# getenforce 
> > > > > Permissive
> > > > > 
> > > > > It looks like it isn't SELinux.
> > > > 
> > > > Checked also on the storage server ovirt-gluster-nfs?
> > > > 
> > > > 
> > > > > 
> > > > > On Thu, 2015-07-02 at 09:53 +0200, Sandro Bonazzola wrote:
> > > > > > Il 02/07/2015 02:36, Christopher Young ha scritto:
> > > > > > > I'm sure I have worked through this before, but I've been 
> > > > > > > 
> > > > > > > banging 
> > > > > > > my head against this one for a while now, and I think I'm 
> > > > > > > too 
> > > > > > > 
> > > > > > > close 
> > > > > > > to the issue.
> > > > > > > 
> > > > > > > Basically, my hosted-engine won't start anymore.  I do 
> > > > > > > recall 
> > > > > > > 
> > > > > > > attempting to migrate it to a new gluster-based NFS share 
> > > > > > > 
> > > > > > > recently, 
> > > > > > > but I could have
> > > > > > > sworn that was successful and working.  I _think_ I have 
> > > > > > > some 
> > > > > > > 
> > > > > > > sort 
> > > > > > > of id issue with storage/volume/whatever id's, but I need 
> > > > > > > 
> > > > > > > some 
> > > > > > > help 
> > > > > > > digging through
> > > > > > > it if someone would be so kind.
> > > > > > > 
> > > > > > > I have the following error in 
> > > > > > > /var/log/libvirt/qemu/HostedEngine.log:
> > > > > > > 
> > > > > > > ----------
> > > > > > > 2015-07-02T00:01:13.080952Z qemu-kvm: -drive
> > > > > > > file=/var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8
> > > > > > > -7057fcee412c/5ead7b5d-50e8-4d6c-a0e5
> > > > > > > -bbe6d93dd836/eeb2d821
> > > > > > > -a432
> > > > > > > -4df6-8856-fdb14df260e3,if=none,id=drive-virtio
> > > > > > > -disk0,format=raw,serial=5ead7b5d-50e8-4d6c-a0e5
> > > > > > > -bbe6d93dd836,cache=none,werror=stop,rerror=stop,aio=thre
> > > > > > > ads:
> > > > > > > could not open disk image
> > > > > > > /var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8
> > > > > > > -7057fcee412c/5ead7b5d
> > > > > > > -50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432-4df6-8856
> > > > > > > -fdb14df260e3: 
> > > > > > > Could not
> > > > > > > refresh total sector count: Input/output error
> > > > > > > ----------
> > > > > > 
> > > > > > 
> > > > > > Please check selinux: ausearch -m avc
> > > > > > it might be selinux preventing access to the disk image.
> > > > > > 
> > > > > > 
> > > > > > 
> > > > > > > 
> > > > > > > I also am including a few command outputs that I'm sure 
> > > > > > > might 
> > > > > > > 
> > > > > > > help:
> > > > > > > 
> > > > > > > ----------
> > > > > > > [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8
> > > > > > > -7057fcee412c]# 
> > > > > > > ls 
> > > > > > > -la /var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8
> > > > > > > -7057fcee412c/
> > > > > > > total 8
> > > > > > > drwxr-xr-x. 2 vdsm kvm  80 Jul  1 20:04 .
> > > > > > > drwxr-xr-x. 3 vdsm kvm  60 Jul  1 20:04 ..
> > > > > > > lrwxrwxrwx. 1 vdsm kvm 128 Jul  1 20:04 1d80a60c-8f26
> > > > > > > -4448
> > > > > > > -9460
> > > > > > > -2c7b00ff75bf ->
> > > > > > > /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb
> > > > > > > -d062
> > > > > > > -4ad1
> > > > > > > -9df8-7057fcee412c/images/1d80a60c-8f26-4448-9460
> > > > > > > -2c7b00ff75bf
> > > > > > > lrwxrwxrwx. 1 vdsm kvm 128 Jul  1 20:04 23ac8897-b0c7
> > > > > > > -41d6
> > > > > > > -a7de
> > > > > > > -19f46ed78400 ->
> > > > > > > /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb
> > > > > > > -d062
> > > > > > > -4ad1
> > > > > > > -9df8-7057fcee412c/images/23ac8897-b0c7-41d6-a7de
> > > > > > > -19f46ed78400
> > > > > > > 
> > > > > > > ----------
> > > > > > > [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8
> > > > > > > -7057fcee412c]# 
> > > > > > > cat 
> > > > > > > /etc/ovirt-hosted-engine/vm.conf
> > > > > > > vmId=6b7329f9-518a-4488-b1c4-2cd809f2f580
> > > > > > > memSize=5120
> > > > > > > display=vnc
> > > > > > > devices={index:2,iface:ide,address:{ controller:0, 
> > > > > > > target:0,unit:0, 
> > > > > > > bus:1,
> > > > > > > type:drive},specParams:{},readonly:true,deviceId:77924fc2
> > > > > > > -c5c9
> > > > > > > -408b
> > > > > > > -97d3-cd0b0d11a62c,path:/home/tmp/CentOS-6.6-x86_64
> > > > > > > -minimal.iso,device:cdrom,shared:false,type:disk}
> > > > > > > devices={index:0,iface:virtio,format:raw,poolID:00000000
> > > > > > > -0000
> > > > > > > -0000
> > > > > > > -0000-000000000000,volumeID:eeb2d821-a432-4df6-8856
> > > > > > > -fdb14df260e3,imageID:5ead7b5d-50e8-4d6c-a0e5
> > > > > > > -bbe6d93dd836,specParams:{},readonly:false,domainID:4e301
> > > > > > > 7eb
> > > > > > > -d062
> > > > > > > -4ad1-9df8-7057fcee412c,optional:false,deviceId:5ead7b5d
> > > > > > > -50e8
> > > > > > > -4d6c
> > > > > > > -a0e5-bbe6d93dd836,address:{bus:0x00,
> > > > > > > slot:0x06, domain:0x0000, type:pci, 
> > > > > > > function:0x0},device:disk,shared:exclusive,propagateError
> > > > > > > s:of
> > > > > > > f,
> > > > > > > type
> > > > > > > :disk,bootOrder:1}
> > > > > > > devices={device:scsi,model:virtio-scsi,type:controller}
> > > > > > > devices={nicModel:pv,macAddr:00:16:3e:0e:d0:68,linkActive
> > > > > > > :tru
> > > > > > > e,
> > > > > > > netw
> > > > > > > ork:ovirtmgmt,filter:vdsm-no-mac
> > > > > > > -spoofing,specParams:{},deviceId:f70ba622-6ac8-4c06-a005
> > > > > > > -0ebd940a15b2,address:{bus:0x00,
> > > > > > > slot:0x03, domain:0x0000, type:pci, 
> > > > > > > function:0x0},device:bridge,type:interface}
> > > > > > > devices={device:console,specParams:{},type:console,device
> > > > > > > Id:9
> > > > > > > 83
> > > > > > > 86e6
> > > > > > > c-ae56-4b6d-9bfb-c72bbd299ad1,alias:console0}
> > > > > > > vmName=HostedEngine
> > > > > > > spiceSecureChannels=smain,sdisplay,sinputs,scursor,splayb
> > > > > > > ack,
> > > > > > > sr
> > > > > > > ecor
> > > > > > > d,ssmartcard,susbredir
> > > > > > > smp=2
> > > > > > > cpuType=Westmere
> > > > > > > emulatedMachine=pc
> > > > > > > ----------
> > > > > > > 
> > > > > > > [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8
> > > > > > > -7057fcee412c]# 
> > > > > > > cat 
> > > > > > > /etc/ovirt-hosted-engine/hosted-engine.conf
> > > > > > > fqdn=orldc-dev-vengine01.*******
> > > > > > > vm_disk_id=5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836
> > > > > > > vmid=6b7329f9-518a-4488-b1c4-2cd809f2f580
> > > > > > > storage=ovirt-gluster-nfs:/engine
> > > > > > > conf=/etc/ovirt-hosted-engine/vm.conf
> > > > > > > service_start_time=0
> > > > > > > host_id=1
> > > > > > > console=vnc
> > > > > > > domainType=nfs3
> > > > > > > spUUID=379cf161-d5b1-4c20-bb71-e3ca5d2ccd6b
> > > > > > > sdUUID=4e3017eb-d062-4ad1-9df8-7057fcee412c
> > > > > > > connectionUUID=0d1b50ac-cf3f-4cd7-90df-3c3a6d11a984
> > > > > > > ca_cert=/etc/pki/vdsm/libvirt-spice/ca-cert.pem
> > > > > > > ca_subject="C=EN, L=Test, O=Test, CN=Test"
> > > > > > > vdsm_use_ssl=true
> > > > > > > gateway=10.16.3.1
> > > > > > > bridge=ovirtmgmt
> > > > > > > metadata_volume_UUID=dd9f373c-d161-4fa0-aab1-3cb52305dba7
> > > > > > > metadata_image_UUID=23ac8897-b0c7-41d6-a7de-19f46ed78400
> > > > > > > lockspace_volume_UUID=d9bacbf6-c2f4-4f74-a91f
> > > > > > > -3a3a52f255bf
> > > > > > > lockspace_image_UUID=1d80a60c-8f26-4448-9460-2c7b00ff75bf
> > > > > > > 
> > > > > > > # The following are used only for iSCSI storage
> > > > > > > iqn=
> > > > > > > portal=
> > > > > > > user=
> > > > > > > password=
> > > > > > > port=
> > > > > > > ----------
> > > > > > > 
> > > > > > > (mount output for the NFS share this should be running on 
> > > > > > > 
> > > > > > > (gluster
> > > > > > > -based):
> > > > > > > 
> > > > > > > ovirt-gluster-nfs:/engine on /rhev/data-center/mnt/ovirt
> > > > > > > -gluster
> > > > > > > -nfs:_engine type nfs
> > > > > > > (rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=25
> > > > > > > 5,so
> > > > > > > ft
> > > > > > > ,nol
> > > > > > > ock,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mo
> > > > > > > unta
> > > > > > > dd
> > > > > > > r=10
> > > > > > > .16.3.30,mountvers=3,mountport=38465,mountproto=tcp,local
> > > > > > > _loc
> > > > > > > k=
> > > > > > > all,
> > > > > > > addr=10.16.3.30)
> > > > > > > 
> > > > > > > ----------
> > > > > > > 
> > > > > > > [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8
> > > > > > > -7057fcee412c]# 
> > > > > > > tree 
> > > > > > > /rhev/data-center/mnt/ovirt-gluster-nfs\:_engine/
> > > > > > > /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/
> > > > > > > ├── 4e3017eb-d062-4ad1-9df8-7057fcee412c
> > > > > > > │   ├── dom_md
> > > > > > > │   │   ├── ids
> > > > > > > │   │   ├── inbox
> > > > > > > │   │   ├── leases
> > > > > > > │   │   ├── metadata
> > > > > > > │   │   └── outbox
> > > > > > > │   ├── ha_agent
> > > > > > > │   │   ├── hosted-engine.lockspace ->
> > > > > > > /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb
> > > > > > > -d062
> > > > > > > -4ad1
> > > > > > > -9df8-7057fcee412c/images/1d80a60c-8f26-4448-9460
> > > > > > > -2c7b00ff75bf/d9bacbf6-c2f4-4f74-a91f-3a3a52f255bf
> > > > > > > │   │   └── hosted-engine.metadata ->
> > > > > > > /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb
> > > > > > > -d062
> > > > > > > -4ad1
> > > > > > > -9df8-7057fcee412c/images/23ac8897-b0c7-41d6-a7de
> > > > > > > -19f46ed78400/dd9f373c-d161-4fa0-aab1-3cb52305dba7
> > > > > > > │   ├── images
> > > > > > > │   │   ├── 1d80a60c-8f26-4448-9460-2c7b00ff75bf
> > > > > > > │   │   │   ├── d9bacbf6-c2f4-4f74-a91f-3a3a52f255bf
> > > > > > > │   │   │   ├── d9bacbf6-c2f4-4f74-a91f
> > > > > > > -3a3a52f255bf.lease
> > > > > > > │   │   │   └── d9bacbf6-c2f4-4f74-a91f-3a3a52f255bf.meta
> > > > > > > │   │   ├── 23ac8897-b0c7-41d6-a7de-19f46ed78400
> > > > > > > │   │   │   ├── dd9f373c-d161-4fa0-aab1-3cb52305dba7 ->
> > > > > > > /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb
> > > > > > > -d062
> > > > > > > -4ad1
> > > > > > > -9df8-7057fcee412c/images/23ac8897-b0c7-41d6-a7de
> > > > > > > -19f46ed78400
> > > > > > > │   │   │   ├── dd9f373c-d161-4fa0-aab1
> > > > > > > -3cb52305dba7.lease
> > > > > > > │   │   │   └── dd9f373c-d161-4fa0-aab1-3cb52305dba7.meta
> > > > > > > │   │   ├── 3278c444-d92a-4cb9-87d6-9669c6e4993e
> > > > > > > │   │   │   ├── 1a4b6a5d-9c8f-4d54-91a7-3dd87377c362
> > > > > > > │   │   │   ├── 1a4b6a5d-9c8f-4d54-91a7
> > > > > > > -3dd87377c362.lease
> > > > > > > │   │   │   └── 1a4b6a5d-9c8f-4d54-91a7-3dd87377c362.meta
> > > > > > > │   │   ├── 5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836
> > > > > > > │   │   │   ├── eeb2d821-a432-4df6-8856
> > > > > > > -fdb14df260e3.lease
> > > > > > > │   │   │   └── eeb2d821-a432-4df6-8856-fdb14df260e3.meta
> > > > > > > │   │   ├── 6064179f-2720-4db9-a7c4-a97e044c2238
> > > > > > > │   │   │   ├── 05afaa26-95af-4226-9a00-5383d8937a83
> > > > > > > │   │   │   ├── 05afaa26-95af-4226-9a00
> > > > > > > -5383d8937a83.lease
> > > > > > > │   │   │   └── 05afaa26-95af-4226-9a00-5383d8937a83.meta
> > > > > > > │   │   └── bb9d9a37-4f91-4973-ba9e-72ee81aed0b6
> > > > > > > │   │       ├── 5acb27b3-62c5-46ac-8978-576a8a4a0399
> > > > > > > │   │       ├── 5acb27b3-62c5-46ac-8978
> > > > > > > -576a8a4a0399.lease
> > > > > > > │   │       └── 5acb27b3-62c5-46ac-8978-576a8a4a0399.meta
> > > > > > > │   └── master
> > > > > > > │       ├── tasks
> > > > > > > │       │   └── fef13299-0e7f-4c7a-a399-092a1235faab
> > > > > > > │       │       ├── fef13299-0e7f-4c7a-a399
> > > > > > > -092a1235faab.job.0
> > > > > > > │       │       ├── fef13299-0e7f-4c7a-a399
> > > > > > > -092a1235faab.recover.0
> > > > > > > │       │       ├── fef13299-0e7f-4c7a-a399
> > > > > > > -092a1235faab.recover.1
> > > > > > > │       │       ├── fef13299-0e7f-4c7a-a399
> > > > > > > -092a1235faab.result
> > > > > > > │       │       └── fef13299-0e7f-4c7a-a399
> > > > > > > -092a1235faab.task
> > > > > > > │       └── vms
> > > > > > > └── __DIRECT_IO_TEST__
> > > > > > > 
> > > > > > > 16 directories, 28 files
> > > > > > > ----------
> > > > > > > 
> > > > > > > [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8
> > > > > > > -7057fcee412c]# 
> > > > > > > tree 
> > > > > > > /var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8
> > > > > > > -7057fcee412c
> > > > > > > /var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8
> > > > > > > -7057fcee412c
> > > > > > > ├── 1d80a60c-8f26-4448-9460-2c7b00ff75bf ->
> > > > > > > /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb
> > > > > > > -d062
> > > > > > > -4ad1
> > > > > > > -9df8-7057fcee412c/images/1d80a60c-8f26-4448-9460
> > > > > > > -2c7b00ff75bf
> > > > > > > └── 23ac8897-b0c7-41d6-a7de-19f46ed78400 ->
> > > > > > > /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb
> > > > > > > -d062
> > > > > > > -4ad1
> > > > > > > -9df8-7057fcee412c/images/23ac8897-b0c7-41d6-a7de
> > > > > > > -19f46ed78400
> > > > > > > 
> > > > > > > 2 directories, 0 files
> > > > > > > ----------
> > > > > > > 
> > > > > > > 
> > > > > > > Can you anyone point me in the right direction here? 
> > > > > > > Eventually, 
> > > > > > > I'd like to get this hosted-engine running in HA mode (I 
> > > > > > > had 
> > > > > > > started this process
> > > > > > > with ctdb/gluster/etc. via the well-known guide out 
> > > > > > > there), 
> > > > > > > but 
> > > > > > > I'm 
> > > > > > > thinking that at this point, I should get this up and 
> > > > > > > running 
> > > > > > > 
> > > > > > > again, export
> > > > > > > everything, and perhaps just build a new hosted-engine 
> > > > > > > and 
> > > > > > > import 
> > > > > > > things into (if that is possible).  Right now, though, I 
> > > > > > > just 
> > > > > > > 
> > > > > > > need 
> > > > > > > to get this
> > > > > > > hosted-engine run.
> > > > > > > 
> > > > > > > Many many thanks!
> > > > > > > 
> > > > > > > Chris
> > > > > > > 
> > > > > > > 
> > > > > > > _______________________________________________
> > > > > > > Users mailing list
> > > > > > > Users@ovirt.org
> > > > > > > http://lists.ovirt.org/mailman/listinfo/users
> > > > > > > 
> > > > > > 
> > > > > > 
> > > > 
> > > > 
> > _______________________________________________
> > Users mailing list
> > Users@ovirt.org
> > http://lists.ovirt.org/mailman/listinfo/users
> 
_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to