Curious do the vms have active snapshots? On Mon, Feb 10, 2020 at 5:59 PM <s.pani...@gmail.com> wrote:
> Hello, all. I have a 3-node Hyperconverged oVirt 4.3.8 cluster running on > CentOS 7.7 hosts. I was investigating poor Gluster performance and heard > about libgfapi, so I thought I'd give it a shot. Looking through the > documentation, followed by lots of threads and BZ reports, I've done the > following to enable it: > > First, I shut down all VMs except the engine. Then... > > On the hosts: > 1. setsebool -P virt_use_glusterfs on > 2. dynamic_ownership=0 in /etc/libvirt/qemu.conf > > On the engine VM: > 1. engine-config -s LibgfApiSupported=true --cver=4.3 > 2. systemctl restart ovirt-engine > > VMs now fail to launch. Am I doing this correctly? I should also note that > the hosts still have the Gluster domain mounted via FUSE. > > Here's a relevant bit from engine.log: > > 2020-02-06T16:38:32.573511Z qemu-kvm: -drive file=gluster:// > node1.fs.trashnet.xyz:24007/vmstore/781717e5-1cff-43a1-b586-9941503544e8/images/a1d56b14-6d72-4f46-a0aa-eb0870c36bc4/a2314816-7970-49ce-a80c-ab0d1cf17c78,file.debug=4,format=qcow2,if=none,id=drive-ua-a1d56b14-6d72-4f46-a0aa-eb0870c36bc4,serial=a1d56b14-6d72-4f46-a0aa-eb0870c36bc4,werror=stop,rerror=stop,cache=none,discard=unmap,aio=native: > Could not read qcow2 header: Invalid argument. > > The full engine.log from one of the attempts: > > 2020-02-06 16:38:24,909Z INFO > [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] > (ForkJoinPool-1-worker-12) [] add VM > 'df9dbac4-35c0-40ee-acd4-a1cfc959aa8b'(yumcache) to rerun treatment > 2020-02-06 16:38:25,010Z ERROR > [org.ovirt.engine.core.vdsbroker.monitoring.VmsMonitoring] > (ForkJoinPool-1-worker-12) [] Rerun VM > 'df9dbac4-35c0-40ee-acd4-a1cfc959aa8b'. Called from VDS ' > node2.ovirt.trashnet.xyz' > 2020-02-06 16:38:25,091Z WARN > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] > (EE-ManagedThreadFactory-engine-Thread-216) [] EVENT_ID: > USER_INITIATED_RUN_VM_FAILED(151), Failed to run VM yumcache on Host > node2.ovirt.trashnet.xyz. > 2020-02-06 16:38:25,166Z INFO [org.ovirt.engine.core.bll.RunVmCommand] > (EE-ManagedThreadFactory-engine-Thread-216) [] Lock Acquired to object > 'EngineLock:{exclusiveLocks='[df9dbac4-35c0-40ee-acd4-a1cfc959aa8b=VM]', > sharedLocks=''}' > 2020-02-06 16:38:25,179Z INFO > [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand] > (EE-ManagedThreadFactory-engine-Thread-216) [] START, > IsVmDuringInitiatingVDSCommand( > IsVmDuringInitiatingVDSCommandParameters:{vmId='df9dbac4-35c0-40ee-acd4-a1cfc959aa8b'}), > log id: 2107f52a > 2020-02-06 16:38:25,181Z INFO > [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand] > (EE-ManagedThreadFactory-engine-Thread-216) [] FINISH, > IsVmDuringInitiatingVDSCommand, return: false, log id: 2107f52a > 2020-02-06 16:38:25,298Z INFO [org.ovirt.engine.core.bll.RunVmCommand] > (EE-ManagedThreadFactory-engine-Thread-216) [] Running command: > RunVmCommand internal: false. Entities affected : ID: > df9dbac4-35c0-40ee-acd4-a1cfc959aa8b Type: VMAction group RUN_VM with role > type USER > 2020-02-06 16:38:25,313Z INFO > [org.ovirt.engine.core.bll.utils.EmulatedMachineUtils] > (EE-ManagedThreadFactory-engine-Thread-216) [] Emulated machine > 'pc-q35-rhel7.6.0' which is different than that of the cluster is set for > 'yumcache'(df9dbac4-35c0-40ee-acd4-a1cfc959aa8b) > 2020-02-06 16:38:25,382Z INFO > [org.ovirt.engine.core.vdsbroker.UpdateVmDynamicDataVDSCommand] > (EE-ManagedThreadFactory-engine-Thread-216) [] START, > UpdateVmDynamicDataVDSCommand( > UpdateVmDynamicDataVDSCommandParameters:{hostId='null', > vmId='df9dbac4-35c0-40ee-acd4-a1cfc959aa8b', > vmDynamic='org.ovirt.engine.core.common.businessentities.VmDynamic@9774a64'}), > log id: 4a83911f > 2020-02-06 16:38:25,417Z INFO > [org.ovirt.engine.core.vdsbroker.UpdateVmDynamicDataVDSCommand] > (EE-ManagedThreadFactory-engine-Thread-216) [] FINISH, > UpdateVmDynamicDataVDSCommand, return: , log id: 4a83911f > 2020-02-06 16:38:25,418Z INFO > [org.ovirt.engine.core.vdsbroker.CreateVDSCommand] > (EE-ManagedThreadFactory-engine-Thread-216) [] START, CreateVDSCommand( > CreateVDSCommandParameters:{hostId='c3465ca2-395e-4c0c-b72e-b5b7153df452', > vmId='df9dbac4-35c0-40ee-acd4-a1cfc959aa8b', vm='VM [yumcache]'}), log id: > 5e07ba66 > 2020-02-06 16:38:25,420Z INFO > [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateBrokerVDSCommand] > (EE-ManagedThreadFactory-engine-Thread-216) [] START, > CreateBrokerVDSCommand(HostName = node1.ovirt.trashnet.xyz, > CreateVDSCommandParameters:{hostId='c3465ca2-395e-4c0c-b72e-b5b7153df452', > vmId='df9dbac4-35c0-40ee-acd4-a1cfc959aa8b', vm='VM [yumcache]'}), log id: > 1bfa03c4 > 2020-02-06 16:38:25,424Z INFO > [org.ovirt.engine.core.vdsbroker.builder.vminfo.VmInfoBuildUtils] > (EE-ManagedThreadFactory-engine-Thread-216) [] Kernel FIPS - Guid: > c3465ca2-395e-4c0c-b72e-b5b7153df452 fips: false > 2020-02-06 16:38:25,435Z INFO > [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateBrokerVDSCommand] > (EE-ManagedThreadFactory-engine-Thread-216) [] VM <?xml version="1.0" > encoding="UTF-8"?><domain type="kvm" xmlns:ovirt-tune=" > http://ovirt.org/vm/tune/1.0" xmlns:ovirt-vm="http://ovirt.org/vm/1.0"> > <name>yumcache</name> > <uuid>df9dbac4-35c0-40ee-acd4-a1cfc959aa8b</uuid> > <memory>1048576</memory> > <currentMemory>1048576</currentMemory> > <iothreads>1</iothreads> > <maxMemory slots="16">4194304</maxMemory> > <vcpu current="1">16</vcpu> > <sysinfo type="smbios"> > <system> > <entry name="manufacturer">oVirt</entry> > <entry name="product">OS-NAME:</entry> > <entry name="version">OS-VERSION:</entry> > <entry name="serial">HOST-SERIAL:</entry> > <entry name="uuid">df9dbac4-35c0-40ee-acd4-a1cfc959aa8b</entry> > </system> > </sysinfo> > <clock offset="variable" adjustment="0"> > <timer name="rtc" tickpolicy="catchup"/> > <timer name="pit" tickpolicy="delay"/> > <timer name="hpet" present="no"/> > </clock> > <features> > <acpi/> > </features> > <cpu match="exact"> > <model>EPYC</model> > <feature name="ibpb" policy="require"/> > <feature name="virt-ssbd" policy="require"/> > <topology cores="1" threads="1" sockets="16"/> > <numa> > <cell id="0" cpus="0" memory="1048576"/> > </numa> > </cpu> > <cputune/> > <devices> > <input type="tablet" bus="usb"/> > <channel type="unix"> > <target type="virtio" name="ovirt-guest-agent.0"/> > <source mode="bind" > path="/var/lib/libvirt/qemu/channels/df9dbac4-35c0-40ee-acd4-a1cfc959aa8b.ovirt-guest-agent.0"/> > </channel> > <channel type="unix"> > <target type="virtio" name="org.qemu.guest_agent.0"/> > <source mode="bind" > path="/var/lib/libvirt/qemu/channels/df9dbac4-35c0-40ee-acd4-a1cfc959aa8b.org.qemu.guest_agent.0"/> > </channel> > <controller type="pci" model="pcie-root-port" index="1"> > <address bus="0x00" domain="0x0000" function="0x0" slot="0x02" > type="pci" multifunction="on"/> > </controller> > <memballoon model="virtio"> > <stats period="5"/> > <alias name="ua-27c77007-3a3c-4431-958d-90fd1c7257dd"/> > <address bus="0x05" domain="0x0000" function="0x0" slot="0x00" > type="pci"/> > </memballoon> > <controller type="pci" model="pcie-root-port" index="2"> > <address bus="0x00" domain="0x0000" function="0x1" slot="0x02" > type="pci"/> > </controller> > <controller type="pci" model="pcie-root-port" index="9"> > <address bus="0x00" domain="0x0000" function="0x0" slot="0x03" > type="pci" multifunction="on"/> > </controller> > <controller type="sata" index="0"> > <address bus="0x00" domain="0x0000" function="0x2" slot="0x1f" > type="pci"/> > </controller> > <rng model="virtio"> > <backend model="random">/dev/urandom</backend> > <alias name="ua-51960005-6b95-47e9-82a7-67d5e0d6cf8a"/> > </rng> > <controller type="pci" model="pcie-root-port" index="6"> > <address bus="0x00" domain="0x0000" function="0x5" slot="0x02" > type="pci"/> > </controller> > <controller type="pci" model="pcie-root-port" index="15"> > <address bus="0x00" domain="0x0000" function="0x6" slot="0x03" > type="pci"/> > </controller> > <controller type="pci" model="pcie-root-port" index="13"> > <address bus="0x00" domain="0x0000" function="0x4" slot="0x03" > type="pci"/> > </controller> > <controller type="pci" model="pcie-root-port" index="7"> > <address bus="0x00" domain="0x0000" function="0x6" slot="0x02" > type="pci"/> > </controller> > <graphics type="vnc" port="-1" autoport="yes" passwd="*****" > passwdValidTo="1970-01-01T00:00:01" keymap="en-us"> > <listen type="network" network="vdsm-ovirtmgmt"/> > </graphics> > <controller type="pci" model="pcie-root-port" index="16"> > <address bus="0x00" domain="0x0000" function="0x7" slot="0x03" > type="pci"/> > </controller> > <controller type="pci" model="pcie-root-port" index="12"> > <address bus="0x00" domain="0x0000" function="0x3" slot="0x03" > type="pci"/> > </controller> > <video> > <model type="qxl" vram="32768" heads="1" ram="65536" vgamem="16384"/> > <alias name="ua-8a295e96-40c3-44de-a3b0-1c4a685a5473"/> > <address bus="0x00" domain="0x0000" function="0x0" slot="0x01" > type="pci"/> > </video> > <graphics type="spice" port="-1" autoport="yes" passwd="*****" > passwdValidTo="1970-01-01T00:00:01" tlsPort="-1"> > <channel name="main" mode="secure"/> > <channel name="inputs" mode="secure"/> > <channel name="cursor" mode="secure"/> > <channel name="playback" mode="secure"/> > <channel name="record" mode="secure"/> > <channel name="display" mode="secure"/> > <channel name="smartcard" mode="secure"/> > <channel name="usbredir" mode="secure"/> > <listen type="network" network="vdsm-ovirtmgmt"/> > </graphics> > <controller type="pci" model="pcie-root-port" index="5"> > <address bus="0x00" domain="0x0000" function="0x4" slot="0x02" > type="pci"/> > </controller> > <controller type="usb" model="qemu-xhci" index="0" ports="8"> > <address bus="0x02" domain="0x0000" function="0x0" slot="0x00" > type="pci"/> > </controller> > <controller type="pci" model="pcie-root-port" index="4"> > <address bus="0x00" domain="0x0000" function="0x3" slot="0x02" > type="pci"/> > </controller> > <controller type="pci" model="pcie-root-port" index="3"> > <address bus="0x00" domain="0x0000" function="0x2" slot="0x02" > type="pci"/> > </controller> > <controller type="pci" model="pcie-root-port" index="11"> > <address bus="0x00" domain="0x0000" function="0x2" slot="0x03" > type="pci"/> > </controller> > <controller type="scsi" model="virtio-scsi" index="0"> > <driver iothread="1"/> > <alias name="ua-d0bf6fcd-7aa2-4658-b7cc-3dac259b7ad2"/> > <address bus="0x03" domain="0x0000" function="0x0" slot="0x00" > type="pci"/> > </controller> > <controller type="pci" model="pcie-root-port" index="8"> > <address bus="0x00" domain="0x0000" function="0x7" slot="0x02" > type="pci"/> > </controller> > <controller type="pci" model="pcie-root-port" index="14"> > <address bus="0x00" domain="0x0000" function="0x5" slot="0x03" > type="pci"/> > </controller> > <controller type="pci" model="pcie-root-port" index="10"> > <address bus="0x00" domain="0x0000" function="0x1" slot="0x03" > type="pci"/> > </controller> > <controller type="virtio-serial" index="0" ports="16"> > <address bus="0x04" domain="0x0000" function="0x0" slot="0x00" > type="pci"/> > </controller> > <channel type="spicevmc"> > <target type="virtio" name="com.redhat.spice.0"/> > </channel> > <controller type="pci" model="pcie-root"/> > <interface type="bridge"> > <model type="virtio"/> > <link state="up"/> > <source bridge="vmnet"/> > <alias name="ua-ceda0ef6-9139-4e5c-8840-86fe344ecbd3"/> > <address bus="0x01" domain="0x0000" function="0x0" slot="0x00" > type="pci"/> > <mac address="56:6f:91:b9:00:05"/> > <mtu size="1500"/> > <filterref filter="vdsm-no-mac-spoofing"/> > <bandwidth/> > </interface> > <disk type="file" device="cdrom" snapshot="no"> > <driver name="qemu" type="raw" error_policy="report"/> > <source file="" startupPolicy="optional"> > <seclabel model="dac" type="none" relabel="no"/> > </source> > <target dev="sdc" bus="sata"/> > <readonly/> > <alias name="ua-bdf99844-2d02-411b-90bb-671ee26764cb"/> > <address bus="0" controller="0" unit="2" type="drive" target="0"/> > </disk> > <disk snapshot="no" type="network" device="disk"> > <target dev="sda" bus="scsi"/> > <source protocol="gluster" > name="vmstore/781717e5-1cff-43a1-b586-9941503544e8/images/a1d56b14-6d72-4f46-a0aa-eb0870c36bc4/a2314816-7970-49ce-a80c-ab0d1cf17c78"> > <host name="node1.fs.trashnet.xyz" port="0"/> > <seclabel model="dac" type="none" relabel="no"/> > </source> > <driver name="qemu" discard="unmap" io="native" type="qcow2" > error_policy="stop" cache="none"/> > <alias name="ua-a1d56b14-6d72-4f46-a0aa-eb0870c36bc4"/> > <address bus="0" controller="0" unit="0" type="drive" target="0"/> > <boot order="1"/> > <serial>a1d56b14-6d72-4f46-a0aa-eb0870c36bc4</serial> > </disk> > <lease> > <key>df9dbac4-35c0-40ee-acd4-a1cfc959aa8b</key> > <lockspace>781717e5-1cff-43a1-b586-9941503544e8</lockspace> > <target offset="6291456" > path="/rhev/data-center/mnt/glusterSD/node1.fs.trashnet.xyz: > _vmstore/781717e5-1cff-43a1-b586-9941503544e8/dom_md/xleases"/> > </lease> > </devices> > <pm> > <suspend-to-disk enabled="no"/> > <suspend-to-mem enabled="no"/> > </pm> > <os> > <type arch="x86_64" machine="pc-q35-rhel7.6.0">hvm</type> > <smbios mode="sysinfo"/> > </os> > <metadata> > <ovirt-tune:qos/> > <ovirt-vm:vm> > <ovirt-vm:minGuaranteedMemoryMb > type="int">512</ovirt-vm:minGuaranteedMemoryMb> > <ovirt-vm:clusterVersion>4.3</ovirt-vm:clusterVersion> > <ovirt-vm:custom/> > <ovirt-vm:device mac_address="56:6f:91:b9:00:05"> > <ovirt-vm:custom/> > </ovirt-vm:device> > <ovirt-vm:device devtype="disk" name="sda"> > > <ovirt-vm:poolID>2ffaec76-462c-11ea-b155-00163e512202</ovirt-vm:poolID> > > <ovirt-vm:volumeID>a2314816-7970-49ce-a80c-ab0d1cf17c78</ovirt-vm:volumeID> > > <ovirt-vm:imageID>a1d56b14-6d72-4f46-a0aa-eb0870c36bc4</ovirt-vm:imageID> > > <ovirt-vm:domainID>781717e5-1cff-43a1-b586-9941503544e8</ovirt-vm:domainID> > </ovirt-vm:device> > <ovirt-vm:launchPaused>false</ovirt-vm:launchPaused> > <ovirt-vm:resumeBehavior>kill</ovirt-vm:resumeBehavior> > </ovirt-vm:vm> > </metadata> > </domain> > > 2020-02-06 16:38:25,455Z INFO > [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateBrokerVDSCommand] > (EE-ManagedThreadFactory-engine-Thread-216) [] FINISH, > CreateBrokerVDSCommand, return: , log id: 1bfa03c4 > 2020-02-06 16:38:25,494Z INFO > [org.ovirt.engine.core.vdsbroker.CreateVDSCommand] > (EE-ManagedThreadFactory-engine-Thread-216) [] FINISH, CreateVDSCommand, > return: WaitForLaunch, log id: 5e07ba66 > 2020-02-06 16:38:25,495Z INFO [org.ovirt.engine.core.bll.RunVmCommand] > (EE-ManagedThreadFactory-engine-Thread-216) [] Lock freed to object > 'EngineLock:{exclusiveLocks='[df9dbac4-35c0-40ee-acd4-a1cfc959aa8b=VM]', > sharedLocks=''}' > 2020-02-06 16:38:25,533Z INFO > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] > (EE-ManagedThreadFactory-engine-Thread-216) [] EVENT_ID: > USER_STARTED_VM(153), VM yumcache was started by admin@internal-authz > (Host: node1.ovirt.trashnet.xyz). > 2020-02-06 16:38:33,300Z INFO > [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] > (ForkJoinPool-1-worker-5) [] VM 'df9dbac4-35c0-40ee-acd4-a1cfc959aa8b' was > reported as Down on VDS 'c3465ca2-395e-4c0c-b72e-b5b7153df452'( > node1.ovirt.trashnet.xyz) > 2020-02-06 16:38:33,301Z INFO > [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand] > (ForkJoinPool-1-worker-5) [] START, DestroyVDSCommand(HostName = > node1.ovirt.trashnet.xyz, > DestroyVmVDSCommandParameters:{hostId='c3465ca2-395e-4c0c-b72e-b5b7153df452', > vmId='df9dbac4-35c0-40ee-acd4-a1cfc959aa8b', secondsToWait='0', > gracefully='false', reason='', ignoreNoVm='true'}), log id: 1f951ea9 > 2020-02-06 16:38:33,478Z INFO > [org.ovirt.engine.core.vdsbroker.monitoring.VmsStatisticsFetcher] > (EE-ManagedThreadFactory-engineScheduled-Thread-8) [] Fetched 2 VMs from > VDS 'c3465ca2-395e-4c0c-b72e-b5b7153df452' > 2020-02-06 16:38:33,545Z INFO > [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand] > (ForkJoinPool-1-worker-5) [] FINISH, DestroyVDSCommand, return: , log id: > 1f951ea9 > 2020-02-06 16:38:33,546Z INFO > [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] > (ForkJoinPool-1-worker-5) [] VM > 'df9dbac4-35c0-40ee-acd4-a1cfc959aa8b'(yumcache) moved from 'WaitForLaunch' > --> 'Down' > 2020-02-06 16:38:33,623Z ERROR > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] > (ForkJoinPool-1-worker-5) [] EVENT_ID: VM_DOWN_ERROR(119), VM yumcache is > down with error. Exit message: internal error: qemu unexpectedly closed the > monitor: [2020-02-06 16:38:31.723977] E [MSGID: 108006] > [afr-common.c:5323:__afr_handle_child_down_event] 0-vmstore-replicate-0: > All subvolumes are down. Going offline until at least one of them comes > back up. > [2020-02-06 16:38:31.724765] I [io-stats.c:4027:fini] 0-vmstore: io-stats > translator unloaded > 2020-02-06T16:38:32.573511Z qemu-kvm: -drive file=gluster:// > node1.fs.trashnet.xyz:24007/vmstore/781717e5-1cff-43a1-b586-9941503544e8/images/a1d56b14-6d72-4f46-a0aa-eb0870c36bc4/a2314816-7970-49ce-a80c-ab0d1cf17c78,file.debug=4,format=qcow2,if=none,id=drive-ua-a1d56b14-6d72-4f46-a0aa-eb0870c36bc4,serial=a1d56b14-6d72-4f46-a0aa-eb0870c36bc4,werror=stop,rerror=stop,cache=none,discard=unmap,aio=native: > Could not read qcow2 header: Invalid argument. > 2020-02-06 16:38:33,624Z INFO > [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] > (ForkJoinPool-1-worker-5) [] add VM > 'df9dbac4-35c0-40ee-acd4-a1cfc959aa8b'(yumcache) to rerun treatment > 2020-02-06 16:38:33,796Z ERROR > [org.ovirt.engine.core.vdsbroker.monitoring.VmsMonitoring] > (ForkJoinPool-1-worker-5) [] Rerun VM > 'df9dbac4-35c0-40ee-acd4-a1cfc959aa8b'. Called from VDS ' > node1.ovirt.trashnet.xyz' > 2020-02-06 16:38:33,899Z WARN > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] > (EE-ManagedThreadFactory-engine-Thread-223) [] EVENT_ID: > USER_INITIATED_RUN_VM_FAILED(151), Failed to run VM yumcache on Host > node1.ovirt.trashnet.xyz. > _______________________________________________ > Users mailing list -- users@ovirt.org > To unsubscribe send an email to users-le...@ovirt.org > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/users@ovirt.org/message/6GTBANZ4R44HJE2BU55GAEBLTETUXTKT/ >
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZKPJITJS4JL6I6GWDIVHWU6XSDM7ZOUZ/