Just installed a single host HCI with gluster, with only the engine vm running. Is it expected this situation below?
# virsh -r list Id Name State ---------------------------------------------------- 2 HostedEngine running and # virsh -r dumpxml 2 . . . <disk type='file' device='disk' snapshot='no'> <driver name='qemu' type='raw' cache='none' error_policy='stop' io='native' iothread='1'/> <source file='/var/run/vdsm/storage/e4eb6832-e0f6-40ee-902f-f301e5a3a643/fc34d770-9318-4539-9233-bfb1c5d68d14/b151557e-f1a2-45cb-b5c9-12c1f470467e'> <seclabel model='dac' relabel='no'/> </source> <backingStore/> <target dev='vda' bus='virtio'/> <serial>fc34d770-9318-4539-9233-bfb1c5d68d14</serial> <alias name='ua-fc34d770-9318-4539-9233-bfb1c5d68d14'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x07' function='0x0'/> </disk> . . . where # ll /var/run/vdsm/storage/e4eb6832-e0f6-40ee-902f-f301e5a3a643/ total 24 lrwxrwxrwx. 1 vdsm kvm 133 Jan 16 14:44 39df7b45-4932-4bfe-b69e-4fb2f8872f4f -> /rhev/data-center/mnt/glusterSD/10.10.10.216: _engine/e4eb6832-e0f6-40ee-902f-f301e5a3a643/images/39df7b45-4932-4bfe-b69e-4fb2f8872f4f lrwxrwxrwx. 1 vdsm kvm 133 Jan 16 14:44 5ba6cd9e-b78d-4de4-9b7f-9688365128bf -> /rhev/data-center/mnt/glusterSD/10.10.10.216: _engine/e4eb6832-e0f6-40ee-902f-f301e5a3a643/images/5ba6cd9e-b78d-4de4-9b7f-9688365128bf lrwxrwxrwx. 1 vdsm kvm 133 Jan 16 15:56 8b8e41e0-a875-4204-8ab1-c10214a49f5c -> /rhev/data-center/mnt/glusterSD/10.10.10.216: _engine/e4eb6832-e0f6-40ee-902f-f301e5a3a643/images/8b8e41e0-a875-4204-8ab1-c10214a49f5c lrwxrwxrwx. 1 vdsm kvm 133 Jan 16 15:56 c21a62ba-73d2-4914-940f-cee6a67a1b08 -> /rhev/data-center/mnt/glusterSD/10.10.10.216: _engine/e4eb6832-e0f6-40ee-902f-f301e5a3a643/images/c21a62ba-73d2-4914-940f-cee6a67a1b08 lrwxrwxrwx. 1 vdsm kvm 133 Jan 16 14:44 fc34d770-9318-4539-9233-bfb1c5d68d14 -> /rhev/data-center/mnt/glusterSD/10.10.10.216: _engine/e4eb6832-e0f6-40ee-902f-f301e5a3a643/images/fc34d770-9318-4539-9233-bfb1c5d68d14 lrwxrwxrwx. 1 vdsm kvm 133 Jan 16 14:44 fd73354d-699b-478e-893c-e2a0bd1e6cbb -> /rhev/data-center/mnt/glusterSD/10.10.10.216: _engine/e4eb6832-e0f6-40ee-902f-f301e5a3a643/images/fd73354d-699b-478e-893c-e2a0bd1e6cbb so hosted engine not using libgfapi? Also, on hosted engine [root@hciengine ~]# engine-config -g LibgfApiSupported LibgfApiSupported: false version: 4.1 LibgfApiSupported: false version: 4.2 LibgfApiSupported: false version: 4.3 [root@hciengine ~]# So that if I import a CentOS7 Atomic host image from glance repo as template and create a new vm from it,when running this VM I get # virsh -r dumpxml 3 . . . <disk type='file' device='disk' snapshot='no'> <driver name='qemu' type='qcow2' cache='none' error_policy='stop' io='native' iothread='1'/> <source file='/rhev/data-center/mnt/glusterSD/10.10.10.216: _data/601d725a-1622-4dc8-a24d-2dba72ddf6ae/images/e4f92226-0f56-4822-a622-d1ebff41df9f/c6b2e076-1519-433e-9b37-2005c9ce6d2e'> <seclabel model='dac' relabel='no'/> </source> <backingStore/> <target dev='vda' bus='virtio'/> <serial>e4f92226-0f56-4822-a622-d1ebff41df9f</serial> <boot order='1'/> <alias name='ua-e4f92226-0f56-4822-a622-d1ebff41df9f'/> <address type='pci' domain='0x0000' bus='0x00' slot='0x06' function='0x0'/> </disk> . . . I remember there was an "old" bug opened causing this default of not enabling libgfapi Does this mean it was not solved yet? If I remember correctly the bugzilla was this one related to HA: https://bugzilla.redhat.com/show_bug.cgi?id=1484227 that is still in new status.... since almost 2 years Is this the only one open? Thanks, Gianluca
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/KHWQA72JZVGLVXCIQPU2FDFRS73BW4LY/