Looks normal, hosted engine uses some extra files/leases to track some of it’s stuff.
Looks like you might have a gluster problem though, that IO error appears to be on your hosted engines disk image. Check for split brains and try and initiate a heal on the files, see what you get. -Darrell > On Jul 2, 2015, at 11:33 AM, Christopher Young <[email protected]> wrote: > > If anyone has an experience of the various IDs in hosted-engine.conf > and vm.conf for the Hosted Engine, I believe I need to just verify > everything. I tried a couple of changes, but I feel like I'm just > making this worse, so I've reverted them. > > One thing I do not understand well is how a gluster-based (NFS) storage > domain for the hosted-engine has so many entries: > > ------------- > > [root@orldc-dev-vnode02 ovirt-gluster-nfs:_engine]# find . -type f | > xargs ls -lah > ls: cannot access ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/images/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432 > -4df6-8856-fdb14df260e3: Input/output error > -rw-rw----. 1 vdsm kvm 1.0M Jul 2 11:20 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/dom_md/ids > -rw-rw----. 1 vdsm kvm 16M Jul 1 19:54 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/dom_md/inbox > -rw-rw----. 1 vdsm kvm 2.0M Jul 1 19:50 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/dom_md/leases > -rw-r--r--. 1 vdsm kvm 482 Jul 1 19:49 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/dom_md/metadata > -rw-rw----. 1 vdsm kvm 16M Jul 1 19:49 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/dom_md/outbox > -rw-rw----. 1 vdsm kvm 1.0M Jul 2 11:32 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf/d9bacbf6-c2f4 > -4f74-a91f-3a3a52f255bf > -rw-rw----. 1 vdsm kvm 1.0M Dec 23 2014 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf/d9bacbf6-c2f4 > -4f74-a91f-3a3a52f255bf.lease > -rw-r--r--. 1 vdsm kvm 284 Dec 23 2014 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/images/1d80a60c-8f26-4448-9460-2c7b00ff75bf/d9bacbf6-c2f4 > -4f74-a91f-3a3a52f255bf.meta > -rw-rw----. 1 vdsm kvm 1.0M Dec 23 2014 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/images/23ac8897-b0c7-41d6-a7de-19f46ed78400/dd9f373c-d161 > -4fa0-aab1-3cb52305dba7.lease > -rw-r--r--. 1 vdsm kvm 283 Dec 23 2014 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/images/23ac8897-b0c7-41d6-a7de-19f46ed78400/dd9f373c-d161 > -4fa0-aab1-3cb52305dba7.meta > -rw-rw----. 1 vdsm kvm 25G Dec 23 2014 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/images/3278c444-d92a-4cb9-87d6-9669c6e4993e/1a4b6a5d-9c8f > -4d54-91a7-3dd87377c362 > -rw-rw----. 1 vdsm kvm 1.0M Dec 23 2014 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/images/3278c444-d92a-4cb9-87d6-9669c6e4993e/1a4b6a5d-9c8f > -4d54-91a7-3dd87377c362.lease > -rw-r--r--. 1 vdsm kvm 278 Dec 23 2014 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/images/3278c444-d92a-4cb9-87d6-9669c6e4993e/1a4b6a5d-9c8f > -4d54-91a7-3dd87377c362.meta > -rw-rw----. 1 vdsm kvm 1.0M Jul 1 17:50 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/images/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432 > -4df6-8856-fdb14df260e3.lease > -rw-r--r--. 1 vdsm kvm 278 Dec 23 2014 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/images/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432 > -4df6-8856-fdb14df260e3.meta > -rw-rw----. 1 vdsm kvm 1.0M Dec 23 2014 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/images/6064179f-2720-4db9-a7c4-a97e044c2238/05afaa26-95af > -4226-9a00-5383d8937a83 > -rw-rw----. 1 vdsm kvm 1.0M Dec 23 2014 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/images/6064179f-2720-4db9-a7c4-a97e044c2238/05afaa26-95af > -4226-9a00-5383d8937a83.lease > -rw-r--r--. 1 vdsm kvm 284 Dec 23 2014 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/images/6064179f-2720-4db9-a7c4-a97e044c2238/05afaa26-95af > -4226-9a00-5383d8937a83.meta > -rw-rw----. 1 vdsm kvm 1004K Dec 23 2014 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/images/bb9d9a37-4f91-4973-ba9e-72ee81aed0b6/5acb27b3-62c5 > -46ac-8978-576a8a4a0399 > -rw-rw----. 1 vdsm kvm 1.0M Dec 23 2014 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/images/bb9d9a37-4f91-4973-ba9e-72ee81aed0b6/5acb27b3-62c5 > -46ac-8978-576a8a4a0399.lease > -rw-r--r--. 1 vdsm kvm 283 Dec 23 2014 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/images/bb9d9a37-4f91-4973-ba9e-72ee81aed0b6/5acb27b3-62c5 > -46ac-8978-576a8a4a0399.meta > -rw-r--r--. 1 vdsm kvm 384 Jul 1 19:49 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399 > -092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.job.0 > -rw-r--r--. 1 vdsm kvm 277 Jul 1 19:49 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399 > -092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.recover.0 > -rw-r--r--. 1 vdsm kvm 417 Jul 1 19:49 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399 > -092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.recover.1 > -rw-r--r--. 1 vdsm kvm 107 Jul 1 19:49 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399 > -092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.result > -rw-r--r--. 1 vdsm kvm 299 Jul 1 19:49 ./4e3017eb-d062-4ad1-9df8 > -7057fcee412c/master/tasks/fef13299-0e7f-4c7a-a399 > -092a1235faab/fef13299-0e7f-4c7a-a399-092a1235faab.task > -rwxr-xr-x. 1 vdsm kvm 0 Jul 2 11:22 ./__DIRECT_IO_TEST__ > ------------- > > > > On Thu, 2015-07-02 at 11:05 -0400, Christopher Young wrote: >> That's actually a local NFS implementation (running on Gluster) so I >> wouldn't expect that it would factor in, but I just verified that all >> gluster hosts are in Permissive mode. Like I mentioned, I think I >> may >> have some ids wrong or bad links or something since this had >> previously >> been working fine, though I must be missing it. >> >> On Thu, 2015-07-02 at 16:08 +0200, Sandro Bonazzola wrote: >>> Il 02/07/2015 16:04, Christopher Young ha scritto: >>>> [root@orldc-dev-vnode02 ~]# getenforce >>>> Permissive >>>> >>>> It looks like it isn't SELinux. >>> >>> Checked also on the storage server ovirt-gluster-nfs? >>> >>> >>>> >>>> On Thu, 2015-07-02 at 09:53 +0200, Sandro Bonazzola wrote: >>>>> Il 02/07/2015 02:36, Christopher Young ha scritto: >>>>>> I'm sure I have worked through this before, but I've been >>>>>> banging >>>>>> my head against this one for a while now, and I think I'm too >>>>>> >>>>>> close >>>>>> to the issue. >>>>>> >>>>>> Basically, my hosted-engine won't start anymore. I do recall >>>>>> >>>>>> attempting to migrate it to a new gluster-based NFS share >>>>>> recently, >>>>>> but I could have >>>>>> sworn that was successful and working. I _think_ I have some >>>>>> >>>>>> sort >>>>>> of id issue with storage/volume/whatever id's, but I need >>>>>> some >>>>>> help >>>>>> digging through >>>>>> it if someone would be so kind. >>>>>> >>>>>> I have the following error in >>>>>> /var/log/libvirt/qemu/HostedEngine.log: >>>>>> >>>>>> ---------- >>>>>> 2015-07-02T00:01:13.080952Z qemu-kvm: -drive >>>>>> file=/var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8 >>>>>> -7057fcee412c/5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821 >>>>>> -a432 >>>>>> -4df6-8856-fdb14df260e3,if=none,id=drive-virtio >>>>>> -disk0,format=raw,serial=5ead7b5d-50e8-4d6c-a0e5 >>>>>> -bbe6d93dd836,cache=none,werror=stop,rerror=stop,aio=threads: >>>>>> could not open disk image >>>>>> /var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8 >>>>>> -7057fcee412c/5ead7b5d >>>>>> -50e8-4d6c-a0e5-bbe6d93dd836/eeb2d821-a432-4df6-8856 >>>>>> -fdb14df260e3: >>>>>> Could not >>>>>> refresh total sector count: Input/output error >>>>>> ---------- >>>>> >>>>> >>>>> Please check selinux: ausearch -m avc >>>>> it might be selinux preventing access to the disk image. >>>>> >>>>> >>>>> >>>>>> >>>>>> I also am including a few command outputs that I'm sure might >>>>>> >>>>>> help: >>>>>> >>>>>> ---------- >>>>>> [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8 >>>>>> -7057fcee412c]# >>>>>> ls >>>>>> -la /var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8 >>>>>> -7057fcee412c/ >>>>>> total 8 >>>>>> drwxr-xr-x. 2 vdsm kvm 80 Jul 1 20:04 . >>>>>> drwxr-xr-x. 3 vdsm kvm 60 Jul 1 20:04 .. >>>>>> lrwxrwxrwx. 1 vdsm kvm 128 Jul 1 20:04 1d80a60c-8f26-4448 >>>>>> -9460 >>>>>> -2c7b00ff75bf -> >>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062 >>>>>> -4ad1 >>>>>> -9df8-7057fcee412c/images/1d80a60c-8f26-4448-9460 >>>>>> -2c7b00ff75bf >>>>>> lrwxrwxrwx. 1 vdsm kvm 128 Jul 1 20:04 23ac8897-b0c7-41d6 >>>>>> -a7de >>>>>> -19f46ed78400 -> >>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062 >>>>>> -4ad1 >>>>>> -9df8-7057fcee412c/images/23ac8897-b0c7-41d6-a7de >>>>>> -19f46ed78400 >>>>>> >>>>>> ---------- >>>>>> [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8 >>>>>> -7057fcee412c]# >>>>>> cat >>>>>> /etc/ovirt-hosted-engine/vm.conf >>>>>> vmId=6b7329f9-518a-4488-b1c4-2cd809f2f580 >>>>>> memSize=5120 >>>>>> display=vnc >>>>>> devices={index:2,iface:ide,address:{ controller:0, >>>>>> target:0,unit:0, >>>>>> bus:1, >>>>>> type:drive},specParams:{},readonly:true,deviceId:77924fc2 >>>>>> -c5c9 >>>>>> -408b >>>>>> -97d3-cd0b0d11a62c,path:/home/tmp/CentOS-6.6-x86_64 >>>>>> -minimal.iso,device:cdrom,shared:false,type:disk} >>>>>> devices={index:0,iface:virtio,format:raw,poolID:00000000-0000 >>>>>> -0000 >>>>>> -0000-000000000000,volumeID:eeb2d821-a432-4df6-8856 >>>>>> -fdb14df260e3,imageID:5ead7b5d-50e8-4d6c-a0e5 >>>>>> -bbe6d93dd836,specParams:{},readonly:false,domainID:4e3017eb >>>>>> -d062 >>>>>> -4ad1-9df8-7057fcee412c,optional:false,deviceId:5ead7b5d-50e8 >>>>>> -4d6c >>>>>> -a0e5-bbe6d93dd836,address:{bus:0x00, >>>>>> slot:0x06, domain:0x0000, type:pci, >>>>>> function:0x0},device:disk,shared:exclusive,propagateErrors:of >>>>>> f, >>>>>> type >>>>>> :disk,bootOrder:1} >>>>>> devices={device:scsi,model:virtio-scsi,type:controller} >>>>>> devices={nicModel:pv,macAddr:00:16:3e:0e:d0:68,linkActive:tru >>>>>> e, >>>>>> netw >>>>>> ork:ovirtmgmt,filter:vdsm-no-mac >>>>>> -spoofing,specParams:{},deviceId:f70ba622-6ac8-4c06-a005 >>>>>> -0ebd940a15b2,address:{bus:0x00, >>>>>> slot:0x03, domain:0x0000, type:pci, >>>>>> function:0x0},device:bridge,type:interface} >>>>>> devices={device:console,specParams:{},type:console,deviceId:9 >>>>>> 83 >>>>>> 86e6 >>>>>> c-ae56-4b6d-9bfb-c72bbd299ad1,alias:console0} >>>>>> vmName=HostedEngine >>>>>> spiceSecureChannels=smain,sdisplay,sinputs,scursor,splayback, >>>>>> sr >>>>>> ecor >>>>>> d,ssmartcard,susbredir >>>>>> smp=2 >>>>>> cpuType=Westmere >>>>>> emulatedMachine=pc >>>>>> ---------- >>>>>> >>>>>> [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8 >>>>>> -7057fcee412c]# >>>>>> cat >>>>>> /etc/ovirt-hosted-engine/hosted-engine.conf >>>>>> fqdn=orldc-dev-vengine01.******* >>>>>> vm_disk_id=5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836 >>>>>> vmid=6b7329f9-518a-4488-b1c4-2cd809f2f580 >>>>>> storage=ovirt-gluster-nfs:/engine >>>>>> conf=/etc/ovirt-hosted-engine/vm.conf >>>>>> service_start_time=0 >>>>>> host_id=1 >>>>>> console=vnc >>>>>> domainType=nfs3 >>>>>> spUUID=379cf161-d5b1-4c20-bb71-e3ca5d2ccd6b >>>>>> sdUUID=4e3017eb-d062-4ad1-9df8-7057fcee412c >>>>>> connectionUUID=0d1b50ac-cf3f-4cd7-90df-3c3a6d11a984 >>>>>> ca_cert=/etc/pki/vdsm/libvirt-spice/ca-cert.pem >>>>>> ca_subject="C=EN, L=Test, O=Test, CN=Test" >>>>>> vdsm_use_ssl=true >>>>>> gateway=10.16.3.1 >>>>>> bridge=ovirtmgmt >>>>>> metadata_volume_UUID=dd9f373c-d161-4fa0-aab1-3cb52305dba7 >>>>>> metadata_image_UUID=23ac8897-b0c7-41d6-a7de-19f46ed78400 >>>>>> lockspace_volume_UUID=d9bacbf6-c2f4-4f74-a91f-3a3a52f255bf >>>>>> lockspace_image_UUID=1d80a60c-8f26-4448-9460-2c7b00ff75bf >>>>>> >>>>>> # The following are used only for iSCSI storage >>>>>> iqn= >>>>>> portal= >>>>>> user= >>>>>> password= >>>>>> port= >>>>>> ---------- >>>>>> >>>>>> (mount output for the NFS share this should be running on >>>>>> (gluster >>>>>> -based): >>>>>> >>>>>> ovirt-gluster-nfs:/engine on /rhev/data-center/mnt/ovirt >>>>>> -gluster >>>>>> -nfs:_engine type nfs >>>>>> (rw,relatime,vers=3,rsize=1048576,wsize=1048576,namlen=255,so >>>>>> ft >>>>>> ,nol >>>>>> ock,nosharecache,proto=tcp,timeo=600,retrans=6,sec=sys,mounta >>>>>> dd >>>>>> r=10 >>>>>> .16.3.30,mountvers=3,mountport=38465,mountproto=tcp,local_loc >>>>>> k= >>>>>> all, >>>>>> addr=10.16.3.30) >>>>>> >>>>>> ---------- >>>>>> >>>>>> [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8 >>>>>> -7057fcee412c]# >>>>>> tree >>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs\:_engine/ >>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/ >>>>>> ├── 4e3017eb-d062-4ad1-9df8-7057fcee412c >>>>>> │ ├── dom_md >>>>>> │ │ ├── ids >>>>>> │ │ ├── inbox >>>>>> │ │ ├── leases >>>>>> │ │ ├── metadata >>>>>> │ │ └── outbox >>>>>> │ ├── ha_agent >>>>>> │ │ ├── hosted-engine.lockspace -> >>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062 >>>>>> -4ad1 >>>>>> -9df8-7057fcee412c/images/1d80a60c-8f26-4448-9460 >>>>>> -2c7b00ff75bf/d9bacbf6-c2f4-4f74-a91f-3a3a52f255bf >>>>>> │ │ └── hosted-engine.metadata -> >>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062 >>>>>> -4ad1 >>>>>> -9df8-7057fcee412c/images/23ac8897-b0c7-41d6-a7de >>>>>> -19f46ed78400/dd9f373c-d161-4fa0-aab1-3cb52305dba7 >>>>>> │ ├── images >>>>>> │ │ ├── 1d80a60c-8f26-4448-9460-2c7b00ff75bf >>>>>> │ │ │ ├── d9bacbf6-c2f4-4f74-a91f-3a3a52f255bf >>>>>> │ │ │ ├── d9bacbf6-c2f4-4f74-a91f-3a3a52f255bf.lease >>>>>> │ │ │ └── d9bacbf6-c2f4-4f74-a91f-3a3a52f255bf.meta >>>>>> │ │ ├── 23ac8897-b0c7-41d6-a7de-19f46ed78400 >>>>>> │ │ │ ├── dd9f373c-d161-4fa0-aab1-3cb52305dba7 -> >>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062 >>>>>> -4ad1 >>>>>> -9df8-7057fcee412c/images/23ac8897-b0c7-41d6-a7de >>>>>> -19f46ed78400 >>>>>> │ │ │ ├── dd9f373c-d161-4fa0-aab1-3cb52305dba7.lease >>>>>> │ │ │ └── dd9f373c-d161-4fa0-aab1-3cb52305dba7.meta >>>>>> │ │ ├── 3278c444-d92a-4cb9-87d6-9669c6e4993e >>>>>> │ │ │ ├── 1a4b6a5d-9c8f-4d54-91a7-3dd87377c362 >>>>>> │ │ │ ├── 1a4b6a5d-9c8f-4d54-91a7-3dd87377c362.lease >>>>>> │ │ │ └── 1a4b6a5d-9c8f-4d54-91a7-3dd87377c362.meta >>>>>> │ │ ├── 5ead7b5d-50e8-4d6c-a0e5-bbe6d93dd836 >>>>>> │ │ │ ├── eeb2d821-a432-4df6-8856-fdb14df260e3.lease >>>>>> │ │ │ └── eeb2d821-a432-4df6-8856-fdb14df260e3.meta >>>>>> │ │ ├── 6064179f-2720-4db9-a7c4-a97e044c2238 >>>>>> │ │ │ ├── 05afaa26-95af-4226-9a00-5383d8937a83 >>>>>> │ │ │ ├── 05afaa26-95af-4226-9a00-5383d8937a83.lease >>>>>> │ │ │ └── 05afaa26-95af-4226-9a00-5383d8937a83.meta >>>>>> │ │ └── bb9d9a37-4f91-4973-ba9e-72ee81aed0b6 >>>>>> │ │ ├── 5acb27b3-62c5-46ac-8978-576a8a4a0399 >>>>>> │ │ ├── 5acb27b3-62c5-46ac-8978-576a8a4a0399.lease >>>>>> │ │ └── 5acb27b3-62c5-46ac-8978-576a8a4a0399.meta >>>>>> │ └── master >>>>>> │ ├── tasks >>>>>> │ │ └── fef13299-0e7f-4c7a-a399-092a1235faab >>>>>> │ │ ├── fef13299-0e7f-4c7a-a399 >>>>>> -092a1235faab.job.0 >>>>>> │ │ ├── fef13299-0e7f-4c7a-a399 >>>>>> -092a1235faab.recover.0 >>>>>> │ │ ├── fef13299-0e7f-4c7a-a399 >>>>>> -092a1235faab.recover.1 >>>>>> │ │ ├── fef13299-0e7f-4c7a-a399 >>>>>> -092a1235faab.result >>>>>> │ │ └── fef13299-0e7f-4c7a-a399-092a1235faab.task >>>>>> │ └── vms >>>>>> └── __DIRECT_IO_TEST__ >>>>>> >>>>>> 16 directories, 28 files >>>>>> ---------- >>>>>> >>>>>> [root@orldc-dev-vnode02 4e3017eb-d062-4ad1-9df8 >>>>>> -7057fcee412c]# >>>>>> tree >>>>>> /var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8-7057fcee412c >>>>>> /var/run/vdsm/storage/4e3017eb-d062-4ad1-9df8-7057fcee412c >>>>>> ├── 1d80a60c-8f26-4448-9460-2c7b00ff75bf -> >>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062 >>>>>> -4ad1 >>>>>> -9df8-7057fcee412c/images/1d80a60c-8f26-4448-9460 >>>>>> -2c7b00ff75bf >>>>>> └── 23ac8897-b0c7-41d6-a7de-19f46ed78400 -> >>>>>> /rhev/data-center/mnt/ovirt-gluster-nfs:_engine/4e3017eb-d062 >>>>>> -4ad1 >>>>>> -9df8-7057fcee412c/images/23ac8897-b0c7-41d6-a7de >>>>>> -19f46ed78400 >>>>>> >>>>>> 2 directories, 0 files >>>>>> ---------- >>>>>> >>>>>> >>>>>> Can you anyone point me in the right direction here? >>>>>> Eventually, >>>>>> I'd like to get this hosted-engine running in HA mode (I had >>>>>> started this process >>>>>> with ctdb/gluster/etc. via the well-known guide out there), >>>>>> but >>>>>> I'm >>>>>> thinking that at this point, I should get this up and running >>>>>> >>>>>> again, export >>>>>> everything, and perhaps just build a new hosted-engine and >>>>>> import >>>>>> things into (if that is possible). Right now, though, I just >>>>>> >>>>>> need >>>>>> to get this >>>>>> hosted-engine run. >>>>>> >>>>>> Many many thanks! >>>>>> >>>>>> Chris >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> Users mailing list >>>>>> [email protected] >>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>> >>>>> >>>>> >>> >>> > _______________________________________________ > Users mailing list > [email protected] > http://lists.ovirt.org/mailman/listinfo/users _______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

