On February 13, 2020 11:51:41 PM GMT+02:00, Stephen Panicho <s.pani...@gmail.com> wrote: >Darrell, would you care to elaborate on your HA workaround? > >As far as I understand, only the primary Gluster host is visible to >libvirt >when using gfapi, so if that host goes down, all VMs break. I imagine >you're using a round-robin DNS entry for the primary Gluster host, but >I'd >like to be sure. > >On Wed, Feb 12, 2020 at 11:01 AM Darrell Budic <bu...@onholyground.com> >wrote: > >> Yes. I’m using libgfapi access on gluster 6.7 with overt 4.3.8 just >fine, >> but I don’t use snapshots. You can work around the HA issue with DNS >and >> backup server entries on the storage domain as well. Worth it to me >for the >> performance, YMMV. >> >> On Feb 12, 2020, at 8:04 AM, Jayme <jay...@gmail.com> wrote: >> >> From my understanding it's not a default option but many users are >still >> using libgfapi successfully. I'm not sure about its status in the >latest >> 4.3.8 release but I know it is/was working for people in previous >versions. >> The libgfapi bugs affect HA and snapshots (on 3 way replica HCI) but >it >> should still be working otherwise, unless like I said something >changed in >> more recent releases of oVirt. >> >> On Wed, Feb 12, 2020 at 9:43 AM Guillaume Pavese < >> guillaume.pav...@interactiv-group.com> wrote: >> >>> Libgfapi is not supported because of an old bug in qemu. That qemu >bug is >>> slowly getting fixed, but the bugs about Libgfapi support in ovirt >have >>> since been closed as WONTFIX and DEFERRED >>> >>> See : >>> https://bugzilla.redhat.com/show_bug.cgi?id=1465810 >>> https://bugzilla.redhat.com/show_bug.cgi?id=1484660 >>> <https://bugzilla.redhat.com/show_bug.cgi?id=1484227> : "No plans to >>> enable libgfapi in RHHI-V for now. Closing this bug" >>> https://bugzilla.redhat.com/show_bug.cgi?id=1484227 : "No plans to >>> enable libgfapi in RHHI-V for now. Closing this bug" >>> https://bugzilla.redhat.com/show_bug.cgi?id=1633642 : "Closing this >as >>> no action taken from long back.Please reopen if required." >>> >>> Would be nice if someone could reopen the closed bugs so this >feature >>> doesn't get forgotten >>> >>> Guillaume Pavese >>> Ingénieur Système et Réseau >>> Interactiv-Group >>> >>> >>> On Tue, Feb 11, 2020 at 9:58 AM Stephen Panicho ><s.pani...@gmail.com> >>> wrote: >>> >>>> I used the cockpit-based hc setup and "option >rpc-auth-allow-insecure" >>>> is absent from /etc/glusterfs/glusterd.vol. >>>> >>>> I'm going to redo the cluster this week and report back. Thanks for >the >>>> tip! >>>> >>>> On Mon, Feb 10, 2020 at 6:01 PM Darrell Budic ><bu...@onholyground.com> >>>> wrote: >>>> >>>>> The hosts will still mount the volume via FUSE, but you might >double >>>>> check you set the storage up as Gluster and not NFS. >>>>> >>>>> Then gluster used to need some config in glusterd.vol to set >>>>> >>>>> option rpc-auth-allow-insecure on >>>>> >>>>> I’m not sure if that got added to a hyper converged setup or not, >but >>>>> I’d check it. >>>>> >>>>> On Feb 10, 2020, at 4:41 PM, Stephen Panicho <s.pani...@gmail.com> >>>>> wrote: >>>>> >>>>> No, this was a relatively new cluster-- only a couple days old. >Just a >>>>> handful of VMs including the engine. >>>>> >>>>> On Mon, Feb 10, 2020 at 5:26 PM Jayme <jay...@gmail.com> wrote: >>>>> >>>>>> Curious do the vms have active snapshots? >>>>>> >>>>>> On Mon, Feb 10, 2020 at 5:59 PM <s.pani...@gmail.com> wrote: >>>>>> >>>>>>> Hello, all. I have a 3-node Hyperconverged oVirt 4.3.8 cluster >>>>>>> running on CentOS 7.7 hosts. I was investigating poor Gluster >performance >>>>>>> and heard about libgfapi, so I thought I'd give it a shot. >Looking through >>>>>>> the documentation, followed by lots of threads and BZ reports, >I've done >>>>>>> the following to enable it: >>>>>>> >>>>>>> First, I shut down all VMs except the engine. Then... >>>>>>> >>>>>>> On the hosts: >>>>>>> 1. setsebool -P virt_use_glusterfs on >>>>>>> 2. dynamic_ownership=0 in /etc/libvirt/qemu.conf >>>>>>> >>>>>>> On the engine VM: >>>>>>> 1. engine-config -s LibgfApiSupported=true --cver=4.3 >>>>>>> 2. systemctl restart ovirt-engine >>>>>>> >>>>>>> VMs now fail to launch. Am I doing this correctly? I should also >note >>>>>>> that the hosts still have the Gluster domain mounted via FUSE. >>>>>>> >>>>>>> Here's a relevant bit from engine.log: >>>>>>> >>>>>>> 2020-02-06T16:38:32.573511Z qemu-kvm: -drive file=gluster:// >>>>>>> >node1.fs.trashnet.xyz:24007/vmstore/781717e5-1cff-43a1-b586-9941503544e8/images/a1d56b14-6d72-4f46-a0aa-eb0870c36bc4/a2314816-7970-49ce-a80c-ab0d1cf17c78,file.debug=4,format=qcow2,if=none,id=drive-ua-a1d56b14-6d72-4f46-a0aa-eb0870c36bc4,serial=a1d56b14-6d72-4f46-a0aa-eb0870c36bc4,werror=stop,rerror=stop,cache=none,discard=unmap,aio=native: >>>>>>> Could not read qcow2 header: Invalid argument. >>>>>>> >>>>>>> The full engine.log from one of the attempts: >>>>>>> >>>>>>> 2020-02-06 16:38:24,909Z INFO >>>>>>> [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] >>>>>>> (ForkJoinPool-1-worker-12) [] add VM >>>>>>> 'df9dbac4-35c0-40ee-acd4-a1cfc959aa8b'(yumcache) to rerun >treatment >>>>>>> 2020-02-06 16:38:25,010Z ERROR >>>>>>> [org.ovirt.engine.core.vdsbroker.monitoring.VmsMonitoring] >>>>>>> (ForkJoinPool-1-worker-12) [] Rerun VM >>>>>>> 'df9dbac4-35c0-40ee-acd4-a1cfc959aa8b'. Called from VDS ' >>>>>>> node2.ovirt.trashnet.xyz' >>>>>>> 2020-02-06 16:38:25,091Z WARN >>>>>>> >[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >>>>>>> (EE-ManagedThreadFactory-engine-Thread-216) [] EVENT_ID: >>>>>>> USER_INITIATED_RUN_VM_FAILED(151), Failed to run VM yumcache on >Host >>>>>>> node2.ovirt.trashnet.xyz. >>>>>>> 2020-02-06 16:38:25,166Z INFO >>>>>>> [org.ovirt.engine.core.bll.RunVmCommand] >>>>>>> (EE-ManagedThreadFactory-engine-Thread-216) [] Lock Acquired to >object >>>>>>> >'EngineLock:{exclusiveLocks='[df9dbac4-35c0-40ee-acd4-a1cfc959aa8b=VM]', >>>>>>> sharedLocks=''}' >>>>>>> 2020-02-06 16:38:25,179Z INFO >>>>>>> [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand] >>>>>>> (EE-ManagedThreadFactory-engine-Thread-216) [] START, >>>>>>> IsVmDuringInitiatingVDSCommand( >>>>>>> >IsVmDuringInitiatingVDSCommandParameters:{vmId='df9dbac4-35c0-40ee-acd4-a1cfc959aa8b'}), >>>>>>> log id: 2107f52a >>>>>>> 2020-02-06 16:38:25,181Z INFO >>>>>>> [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand] >>>>>>> (EE-ManagedThreadFactory-engine-Thread-216) [] FINISH, >>>>>>> IsVmDuringInitiatingVDSCommand, return: false, log id: 2107f52a >>>>>>> 2020-02-06 16:38:25,298Z INFO >>>>>>> [org.ovirt.engine.core.bll.RunVmCommand] >>>>>>> (EE-ManagedThreadFactory-engine-Thread-216) [] Running command: >>>>>>> RunVmCommand internal: false. Entities affected : ID: >>>>>>> df9dbac4-35c0-40ee-acd4-a1cfc959aa8b Type: VMAction group RUN_VM >with role >>>>>>> type USER >>>>>>> 2020-02-06 16:38:25,313Z INFO >>>>>>> [org.ovirt.engine.core.bll.utils.EmulatedMachineUtils] >>>>>>> (EE-ManagedThreadFactory-engine-Thread-216) [] Emulated machine >>>>>>> 'pc-q35-rhel7.6.0' which is different than that of the cluster >is set for >>>>>>> 'yumcache'(df9dbac4-35c0-40ee-acd4-a1cfc959aa8b) >>>>>>> 2020-02-06 16:38:25,382Z INFO >>>>>>> [org.ovirt.engine.core.vdsbroker.UpdateVmDynamicDataVDSCommand] >>>>>>> (EE-ManagedThreadFactory-engine-Thread-216) [] START, >>>>>>> UpdateVmDynamicDataVDSCommand( >>>>>>> UpdateVmDynamicDataVDSCommandParameters:{hostId='null', >>>>>>> vmId='df9dbac4-35c0-40ee-acd4-a1cfc959aa8b', >>>>>>> >vmDynamic='org.ovirt.engine.core.common.businessentities.VmDynamic@9774a64'}), >>>>>>> log id: 4a83911f >>>>>>> 2020-02-06 16:38:25,417Z INFO >>>>>>> [org.ovirt.engine.core.vdsbroker.UpdateVmDynamicDataVDSCommand] >>>>>>> (EE-ManagedThreadFactory-engine-Thread-216) [] FINISH, >>>>>>> UpdateVmDynamicDataVDSCommand, return: , log id: 4a83911f >>>>>>> 2020-02-06 16:38:25,418Z INFO >>>>>>> [org.ovirt.engine.core.vdsbroker.CreateVDSCommand] >>>>>>> (EE-ManagedThreadFactory-engine-Thread-216) [] START, >CreateVDSCommand( >>>>>>> >CreateVDSCommandParameters:{hostId='c3465ca2-395e-4c0c-b72e-b5b7153df452', >>>>>>> vmId='df9dbac4-35c0-40ee-acd4-a1cfc959aa8b', vm='VM >[yumcache]'}), log id: >>>>>>> 5e07ba66 >>>>>>> 2020-02-06 16:38:25,420Z INFO >>>>>>> >[org.ovirt.engine.core.vdsbroker.vdsbroker.CreateBrokerVDSCommand] >>>>>>> (EE-ManagedThreadFactory-engine-Thread-216) [] START, >>>>>>> CreateBrokerVDSCommand(HostName = node1.ovirt.trashnet.xyz, >>>>>>> >CreateVDSCommandParameters:{hostId='c3465ca2-395e-4c0c-b72e-b5b7153df452', >>>>>>> vmId='df9dbac4-35c0-40ee-acd4-a1cfc959aa8b', vm='VM >[yumcache]'}), log id: >>>>>>> 1bfa03c4 >>>>>>> 2020-02-06 16:38:25,424Z INFO >>>>>>> >[org.ovirt.engine.core.vdsbroker.builder.vminfo.VmInfoBuildUtils] >>>>>>> (EE-ManagedThreadFactory-engine-Thread-216) [] Kernel FIPS - >Guid: >>>>>>> c3465ca2-395e-4c0c-b72e-b5b7153df452 fips: false >>>>>>> 2020-02-06 16:38:25,435Z INFO >>>>>>> >[org.ovirt.engine.core.vdsbroker.vdsbroker.CreateBrokerVDSCommand] >>>>>>> (EE-ManagedThreadFactory-engine-Thread-216) [] VM <?xml >version="1.0" >>>>>>> encoding="UTF-8"?><domain type="kvm" xmlns:ovirt-tune=" >>>>>>> http://ovirt.org/vm/tune/1.0" >xmlns:ovirt-vm="http://ovirt.org/vm/1.0 >>>>>>> "> >>>>>>> <name>yumcache</name> >>>>>>> <uuid>df9dbac4-35c0-40ee-acd4-a1cfc959aa8b</uuid> >>>>>>> <memory>1048576</memory> >>>>>>> <currentMemory>1048576</currentMemory> >>>>>>> <iothreads>1</iothreads> >>>>>>> <maxMemory slots="16">4194304</maxMemory> >>>>>>> <vcpu current="1">16</vcpu> >>>>>>> <sysinfo type="smbios"> >>>>>>> <system> >>>>>>> <entry name="manufacturer">oVirt</entry> >>>>>>> <entry name="product">OS-NAME:</entry> >>>>>>> <entry name="version">OS-VERSION:</entry> >>>>>>> <entry name="serial">HOST-SERIAL:</entry> >>>>>>> <entry >name="uuid">df9dbac4-35c0-40ee-acd4-a1cfc959aa8b</entry> >>>>>>> </system> >>>>>>> </sysinfo> >>>>>>> <clock offset="variable" adjustment="0"> >>>>>>> <timer name="rtc" tickpolicy="catchup"/> >>>>>>> <timer name="pit" tickpolicy="delay"/> >>>>>>> <timer name="hpet" present="no"/> >>>>>>> </clock> >>>>>>> <features> >>>>>>> <acpi/> >>>>>>> </features> >>>>>>> <cpu match="exact"> >>>>>>> <model>EPYC</model> >>>>>>> <feature name="ibpb" policy="require"/> >>>>>>> <feature name="virt-ssbd" policy="require"/> >>>>>>> <topology cores="1" threads="1" sockets="16"/> >>>>>>> <numa> >>>>>>> <cell id="0" cpus="0" memory="1048576"/> >>>>>>> </numa> >>>>>>> </cpu> >>>>>>> <cputune/> >>>>>>> <devices> >>>>>>> <input type="tablet" bus="usb"/> >>>>>>> <channel type="unix"> >>>>>>> <target type="virtio" name="ovirt-guest-agent.0"/> >>>>>>> <source mode="bind" >>>>>>> >path="/var/lib/libvirt/qemu/channels/df9dbac4-35c0-40ee-acd4-a1cfc959aa8b.ovirt-guest-agent.0"/> >>>>>>> </channel> >>>>>>> <channel type="unix"> >>>>>>> <target type="virtio" name="org.qemu.guest_agent.0"/> >>>>>>> <source mode="bind" >>>>>>> >path="/var/lib/libvirt/qemu/channels/df9dbac4-35c0-40ee-acd4-a1cfc959aa8b.org.qemu.guest_agent.0"/> >>>>>>> </channel> >>>>>>> <controller type="pci" model="pcie-root-port" index="1"> >>>>>>> <address bus="0x00" domain="0x0000" function="0x0" >slot="0x02" >>>>>>> type="pci" multifunction="on"/> >>>>>>> </controller> >>>>>>> <memballoon model="virtio"> >>>>>>> <stats period="5"/> >>>>>>> <alias name="ua-27c77007-3a3c-4431-958d-90fd1c7257dd"/> >>>>>>> <address bus="0x05" domain="0x0000" function="0x0" >slot="0x00" >>>>>>> type="pci"/> >>>>>>> </memballoon> >>>>>>> <controller type="pci" model="pcie-root-port" index="2"> >>>>>>> <address bus="0x00" domain="0x0000" function="0x1" >slot="0x02" >>>>>>> type="pci"/> >>>>>>> </controller> >>>>>>> <controller type="pci" model="pcie-root-port" index="9"> >>>>>>> <address bus="0x00" domain="0x0000" function="0x0" >slot="0x03" >>>>>>> type="pci" multifunction="on"/> >>>>>>> </controller> >>>>>>> <controller type="sata" index="0"> >>>>>>> <address bus="0x00" domain="0x0000" function="0x2" >slot="0x1f" >>>>>>> type="pci"/> >>>>>>> </controller> >>>>>>> <rng model="virtio"> >>>>>>> <backend model="random">/dev/urandom</backend> >>>>>>> <alias name="ua-51960005-6b95-47e9-82a7-67d5e0d6cf8a"/> >>>>>>> </rng> >>>>>>> <controller type="pci" model="pcie-root-port" index="6"> >>>>>>> <address bus="0x00" domain="0x0000" function="0x5" >slot="0x02" >>>>>>> type="pci"/> >>>>>>> </controller> >>>>>>> <controller type="pci" model="pcie-root-port" index="15"> >>>>>>> <address bus="0x00" domain="0x0000" function="0x6" >slot="0x03" >>>>>>> type="pci"/> >>>>>>> </controller> >>>>>>> <controller type="pci" model="pcie-root-port" index="13"> >>>>>>> <address bus="0x00" domain="0x0000" function="0x4" >slot="0x03" >>>>>>> type="pci"/> >>>>>>> </controller> >>>>>>> <controller type="pci" model="pcie-root-port" index="7"> >>>>>>> <address bus="0x00" domain="0x0000" function="0x6" >slot="0x02" >>>>>>> type="pci"/> >>>>>>> </controller> >>>>>>> <graphics type="vnc" port="-1" autoport="yes" passwd="*****" >>>>>>> passwdValidTo="1970-01-01T00:00:01" keymap="en-us"> >>>>>>> <listen type="network" network="vdsm-ovirtmgmt"/> >>>>>>> </graphics> >>>>>>> <controller type="pci" model="pcie-root-port" index="16"> >>>>>>> <address bus="0x00" domain="0x0000" function="0x7" >slot="0x03" >>>>>>> type="pci"/> >>>>>>> </controller> >>>>>>> <controller type="pci" model="pcie-root-port" index="12"> >>>>>>> <address bus="0x00" domain="0x0000" function="0x3" >slot="0x03" >>>>>>> type="pci"/> >>>>>>> </controller> >>>>>>> <video> >>>>>>> <model type="qxl" vram="32768" heads="1" ram="65536" >>>>>>> vgamem="16384"/> >>>>>>> <alias name="ua-8a295e96-40c3-44de-a3b0-1c4a685a5473"/> >>>>>>> <address bus="0x00" domain="0x0000" function="0x0" >slot="0x01" >>>>>>> type="pci"/> >>>>>>> </video> >>>>>>> <graphics type="spice" port="-1" autoport="yes" >passwd="*****" >>>>>>> passwdValidTo="1970-01-01T00:00:01" tlsPort="-1"> >>>>>>> <channel name="main" mode="secure"/> >>>>>>> <channel name="inputs" mode="secure"/> >>>>>>> <channel name="cursor" mode="secure"/> >>>>>>> <channel name="playback" mode="secure"/> >>>>>>> <channel name="record" mode="secure"/> >>>>>>> <channel name="display" mode="secure"/> >>>>>>> <channel name="smartcard" mode="secure"/> >>>>>>> <channel name="usbredir" mode="secure"/> >>>>>>> <listen type="network" network="vdsm-ovirtmgmt"/> >>>>>>> </graphics> >>>>>>> <controller type="pci" model="pcie-root-port" index="5"> >>>>>>> <address bus="0x00" domain="0x0000" function="0x4" >slot="0x02" >>>>>>> type="pci"/> >>>>>>> </controller> >>>>>>> <controller type="usb" model="qemu-xhci" index="0" >ports="8"> >>>>>>> <address bus="0x02" domain="0x0000" function="0x0" >slot="0x00" >>>>>>> type="pci"/> >>>>>>> </controller> >>>>>>> <controller type="pci" model="pcie-root-port" index="4"> >>>>>>> <address bus="0x00" domain="0x0000" function="0x3" >slot="0x02" >>>>>>> type="pci"/> >>>>>>> </controller> >>>>>>> <controller type="pci" model="pcie-root-port" index="3"> >>>>>>> <address bus="0x00" domain="0x0000" function="0x2" >slot="0x02" >>>>>>> type="pci"/> >>>>>>> </controller> >>>>>>> <controller type="pci" model="pcie-root-port" index="11"> >>>>>>> <address bus="0x00" domain="0x0000" function="0x2" >slot="0x03" >>>>>>> type="pci"/> >>>>>>> </controller> >>>>>>> <controller type="scsi" model="virtio-scsi" index="0"> >>>>>>> <driver iothread="1"/> >>>>>>> <alias name="ua-d0bf6fcd-7aa2-4658-b7cc-3dac259b7ad2"/> >>>>>>> <address bus="0x03" domain="0x0000" function="0x0" >slot="0x00" >>>>>>> type="pci"/> >>>>>>> </controller> >>>>>>> <controller type="pci" model="pcie-root-port" index="8"> >>>>>>> <address bus="0x00" domain="0x0000" function="0x7" >slot="0x02" >>>>>>> type="pci"/> >>>>>>> </controller> >>>>>>> <controller type="pci" model="pcie-root-port" index="14"> >>>>>>> <address bus="0x00" domain="0x0000" function="0x5" >slot="0x03" >>>>>>> type="pci"/> >>>>>>> </controller> >>>>>>> <controller type="pci" model="pcie-root-port" index="10"> >>>>>>> <address bus="0x00" domain="0x0000" function="0x1" >slot="0x03" >>>>>>> type="pci"/> >>>>>>> </controller> >>>>>>> <controller type="virtio-serial" index="0" ports="16"> >>>>>>> <address bus="0x04" domain="0x0000" function="0x0" >slot="0x00" >>>>>>> type="pci"/> >>>>>>> </controller> >>>>>>> <channel type="spicevmc"> >>>>>>> <target type="virtio" name="com.redhat.spice.0"/> >>>>>>> </channel> >>>>>>> <controller type="pci" model="pcie-root"/> >>>>>>> <interface type="bridge"> >>>>>>> <model type="virtio"/> >>>>>>> <link state="up"/> >>>>>>> <source bridge="vmnet"/> >>>>>>> <alias name="ua-ceda0ef6-9139-4e5c-8840-86fe344ecbd3"/> >>>>>>> <address bus="0x01" domain="0x0000" function="0x0" >slot="0x00" >>>>>>> type="pci"/> >>>>>>> <mac address="56:6f:91:b9:00:05"/> >>>>>>> <mtu size="1500"/> >>>>>>> <filterref filter="vdsm-no-mac-spoofing"/> >>>>>>> <bandwidth/> >>>>>>> </interface> >>>>>>> <disk type="file" device="cdrom" snapshot="no"> >>>>>>> <driver name="qemu" type="raw" error_policy="report"/> >>>>>>> <source file="" startupPolicy="optional"> >>>>>>> <seclabel model="dac" type="none" relabel="no"/> >>>>>>> </source> >>>>>>> <target dev="sdc" bus="sata"/> >>>>>>> <readonly/> >>>>>>> <alias name="ua-bdf99844-2d02-411b-90bb-671ee26764cb"/> >>>>>>> <address bus="0" controller="0" unit="2" type="drive" >>>>>>> target="0"/> >>>>>>> </disk> >>>>>>> <disk snapshot="no" type="network" device="disk"> >>>>>>> <target dev="sda" bus="scsi"/> >>>>>>> <source protocol="gluster" >>>>>>> >name="vmstore/781717e5-1cff-43a1-b586-9941503544e8/images/a1d56b14-6d72-4f46-a0aa-eb0870c36bc4/a2314816-7970-49ce-a80c-ab0d1cf17c78"> >>>>>>> <host name="node1.fs.trashnet.xyz" port="0"/> >>>>>>> <seclabel model="dac" type="none" relabel="no"/> >>>>>>> </source> >>>>>>> <driver name="qemu" discard="unmap" io="native" >type="qcow2" >>>>>>> error_policy="stop" cache="none"/> >>>>>>> <alias name="ua-a1d56b14-6d72-4f46-a0aa-eb0870c36bc4"/> >>>>>>> <address bus="0" controller="0" unit="0" type="drive" >>>>>>> target="0"/> >>>>>>> <boot order="1"/> >>>>>>> <serial>a1d56b14-6d72-4f46-a0aa-eb0870c36bc4</serial> >>>>>>> </disk> >>>>>>> <lease> >>>>>>> <key>df9dbac4-35c0-40ee-acd4-a1cfc959aa8b</key> >>>>>>> ><lockspace>781717e5-1cff-43a1-b586-9941503544e8</lockspace> >>>>>>> <target offset="6291456" >path="/rhev/data-center/mnt/glusterSD/ >>>>>>> node1.fs.trashnet.xyz >>>>>>> :_vmstore/781717e5-1cff-43a1-b586-9941503544e8/dom_md/xleases"/> >>>>>>> </lease> >>>>>>> </devices> >>>>>>> <pm> >>>>>>> <suspend-to-disk enabled="no"/> >>>>>>> <suspend-to-mem enabled="no"/> >>>>>>> </pm> >>>>>>> <os> >>>>>>> <type arch="x86_64" machine="pc-q35-rhel7.6.0">hvm</type> >>>>>>> <smbios mode="sysinfo"/> >>>>>>> </os> >>>>>>> <metadata> >>>>>>> <ovirt-tune:qos/> >>>>>>> <ovirt-vm:vm> >>>>>>> <ovirt-vm:minGuaranteedMemoryMb >>>>>>> type="int">512</ovirt-vm:minGuaranteedMemoryMb> >>>>>>> <ovirt-vm:clusterVersion>4.3</ovirt-vm:clusterVersion> >>>>>>> <ovirt-vm:custom/> >>>>>>> <ovirt-vm:device mac_address="56:6f:91:b9:00:05"> >>>>>>> <ovirt-vm:custom/> >>>>>>> </ovirt-vm:device> >>>>>>> <ovirt-vm:device devtype="disk" name="sda"> >>>>>>> >>>>>>> ><ovirt-vm:poolID>2ffaec76-462c-11ea-b155-00163e512202</ovirt-vm:poolID> >>>>>>> >>>>>>> ><ovirt-vm:volumeID>a2314816-7970-49ce-a80c-ab0d1cf17c78</ovirt-vm:volumeID> >>>>>>> >>>>>>> ><ovirt-vm:imageID>a1d56b14-6d72-4f46-a0aa-eb0870c36bc4</ovirt-vm:imageID> >>>>>>> >>>>>>> ><ovirt-vm:domainID>781717e5-1cff-43a1-b586-9941503544e8</ovirt-vm:domainID> >>>>>>> </ovirt-vm:device> >>>>>>> <ovirt-vm:launchPaused>false</ovirt-vm:launchPaused> >>>>>>> <ovirt-vm:resumeBehavior>kill</ovirt-vm:resumeBehavior> >>>>>>> </ovirt-vm:vm> >>>>>>> </metadata> >>>>>>> </domain> >>>>>>> >>>>>>> 2020-02-06 16:38:25,455Z INFO >>>>>>> >[org.ovirt.engine.core.vdsbroker.vdsbroker.CreateBrokerVDSCommand] >>>>>>> (EE-ManagedThreadFactory-engine-Thread-216) [] FINISH, >>>>>>> CreateBrokerVDSCommand, return: , log id: 1bfa03c4 >>>>>>> 2020-02-06 16:38:25,494Z INFO >>>>>>> [org.ovirt.engine.core.vdsbroker.CreateVDSCommand] >>>>>>> (EE-ManagedThreadFactory-engine-Thread-216) [] FINISH, >CreateVDSCommand, >>>>>>> return: WaitForLaunch, log id: 5e07ba66 >>>>>>> 2020-02-06 16:38:25,495Z INFO >>>>>>> [org.ovirt.engine.core.bll.RunVmCommand] >>>>>>> (EE-ManagedThreadFactory-engine-Thread-216) [] Lock freed to >object >>>>>>> >'EngineLock:{exclusiveLocks='[df9dbac4-35c0-40ee-acd4-a1cfc959aa8b=VM]', >>>>>>> sharedLocks=''}' >>>>>>> 2020-02-06 16:38:25,533Z INFO >>>>>>> >[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >>>>>>> (EE-ManagedThreadFactory-engine-Thread-216) [] EVENT_ID: >>>>>>> USER_STARTED_VM(153), VM yumcache was started by >admin@internal-authz >>>>>>> (Host: node1.ovirt.trashnet.xyz). >>>>>>> 2020-02-06 16:38:33,300Z INFO >>>>>>> [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] >>>>>>> (ForkJoinPool-1-worker-5) [] VM >'df9dbac4-35c0-40ee-acd4-a1cfc959aa8b' was >>>>>>> reported as Down on VDS 'c3465ca2-395e-4c0c-b72e-b5b7153df452'( >>>>>>> node1.ovirt.trashnet.xyz) >>>>>>> 2020-02-06 16:38:33,301Z INFO >>>>>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand] >>>>>>> (ForkJoinPool-1-worker-5) [] START, DestroyVDSCommand(HostName = >>>>>>> node1.ovirt.trashnet.xyz, >>>>>>> >DestroyVmVDSCommandParameters:{hostId='c3465ca2-395e-4c0c-b72e-b5b7153df452', >>>>>>> vmId='df9dbac4-35c0-40ee-acd4-a1cfc959aa8b', secondsToWait='0', >>>>>>> gracefully='false', reason='', ignoreNoVm='true'}), log id: >1f951ea9 >>>>>>> 2020-02-06 16:38:33,478Z INFO >>>>>>> >[org.ovirt.engine.core.vdsbroker.monitoring.VmsStatisticsFetcher] >>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-8) [] Fetched 2 >VMs from >>>>>>> VDS 'c3465ca2-395e-4c0c-b72e-b5b7153df452' >>>>>>> 2020-02-06 16:38:33,545Z INFO >>>>>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand] >>>>>>> (ForkJoinPool-1-worker-5) [] FINISH, DestroyVDSCommand, return: >, log id: >>>>>>> 1f951ea9 >>>>>>> 2020-02-06 16:38:33,546Z INFO >>>>>>> [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] >>>>>>> (ForkJoinPool-1-worker-5) [] VM >>>>>>> 'df9dbac4-35c0-40ee-acd4-a1cfc959aa8b'(yumcache) moved from >'WaitForLaunch' >>>>>>> --> 'Down' >>>>>>> 2020-02-06 16:38:33,623Z ERROR >>>>>>> >[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >>>>>>> (ForkJoinPool-1-worker-5) [] EVENT_ID: VM_DOWN_ERROR(119), VM >yumcache is >>>>>>> down with error. Exit message: internal error: qemu unexpectedly >closed the >>>>>>> monitor: [2020-02-06 16:38:31.723977] E [MSGID: 108006] >>>>>>> [afr-common.c:5323:__afr_handle_child_down_event] >0-vmstore-replicate-0: >>>>>>> All subvolumes are down. Going offline until at least one of >them comes >>>>>>> back up. >>>>>>> [2020-02-06 16:38:31.724765] I [io-stats.c:4027:fini] 0-vmstore: >>>>>>> io-stats translator unloaded >>>>>>> 2020-02-06T16:38:32.573511Z qemu-kvm: -drive file=gluster:// >>>>>>> >node1.fs.trashnet.xyz:24007/vmstore/781717e5-1cff-43a1-b586-9941503544e8/images/a1d56b14-6d72-4f46-a0aa-eb0870c36bc4/a2314816-7970-49ce-a80c-ab0d1cf17c78,file.debug=4,format=qcow2,if=none,id=drive-ua-a1d56b14-6d72-4f46-a0aa-eb0870c36bc4,serial=a1d56b14-6d72-4f46-a0aa-eb0870c36bc4,werror=stop,rerror=stop,cache=none,discard=unmap,aio=native: >>>>>>> Could not read qcow2 header: Invalid argument. >>>>>>> 2020-02-06 16:38:33,624Z INFO >>>>>>> [org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer] >>>>>>> (ForkJoinPool-1-worker-5) [] add VM >>>>>>> 'df9dbac4-35c0-40ee-acd4-a1cfc959aa8b'(yumcache) to rerun >treatment >>>>>>> 2020-02-06 16:38:33,796Z ERROR >>>>>>> [org.ovirt.engine.core.vdsbroker.monitoring.VmsMonitoring] >>>>>>> (ForkJoinPool-1-worker-5) [] Rerun VM >>>>>>> 'df9dbac4-35c0-40ee-acd4-a1cfc959aa8b'. Called from VDS ' >>>>>>> node1.ovirt.trashnet.xyz' >>>>>>> 2020-02-06 16:38:33,899Z WARN >>>>>>> >[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] >>>>>>> (EE-ManagedThreadFactory-engine-Thread-223) [] EVENT_ID: >>>>>>> USER_INITIATED_RUN_VM_FAILED(151), Failed to run VM yumcache on >Host >>>>>>> node1.ovirt.trashnet.xyz. >>>>>>> _______________________________________________ >>>>>>> Users mailing list -- users@ovirt.org >>>>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>>>> oVirt Code of Conduct: >>>>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>>>> List Archives: >>>>>>> >https://lists.ovirt.org/archives/list/users@ovirt.org/message/6GTBANZ4R44HJE2BU55GAEBLTETUXTKT/ >>>>>>> >>>>>> _______________________________________________ >>>>> Users mailing list -- users@ovirt.org >>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>>> oVirt Code of Conduct: >>>>> https://www.ovirt.org/community/about/community-guidelines/ >>>>> List Archives: >>>>> >https://lists.ovirt.org/archives/list/users@ovirt.org/message/5EZYM4OADD65RVNDIQVS25EGZOK65PWX/ >>>>> >>>>> >>>>> _______________________________________________ >>>> Users mailing list -- users@ovirt.org >>>> To unsubscribe send an email to users-le...@ovirt.org >>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>>> oVirt Code of Conduct: >>>> https://www.ovirt.org/community/about/community-guidelines/ >>>> List Archives: >>>> >https://lists.ovirt.org/archives/list/users@ovirt.org/message/KKUVN3ZR5QOVKKCOR5XIUYYC6EL7FYXS/ >>>> >>> >>> Ce message et toutes les pièces jointes (ci-après le “message”) sont >>> établis à l’intention exclusive de ses destinataires et sont >confidentiels. >>> Si vous recevez ce message par erreur, merci de le détruire et d’en >avertir >>> immédiatement l’expéditeur. Toute utilisation de ce message non >conforme a >>> sa destination, toute diffusion ou toute publication, totale ou >partielle, >>> est interdite, sauf autorisation expresse. L’internet ne permettant >pas >>> d’assurer l’intégrité de ce message . Interactiv-group (et ses >filiales) >>> décline(nt) toute responsabilité au titre de ce message, dans >l’hypothèse >>> ou il aurait été modifié. IT, ES, UK. >>> <https://interactiv-group.com/disclaimer.html> >>> _______________________________________________ >>> Users mailing list -- users@ovirt.org >>> To unsubscribe send an email to users-le...@ovirt.org >>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >>> oVirt Code of Conduct: >>> https://www.ovirt.org/community/about/community-guidelines/ >>> List Archives: >>> >https://lists.ovirt.org/archives/list/users@ovirt.org/message/S2RYKJLRE533BZB5A3ZDDMIZ573W7HTC/ >>> >> _______________________________________________ >> Users mailing list -- users@ovirt.org >> To unsubscribe send an email to users-le...@ovirt.org >> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ >> oVirt Code of Conduct: >> https://www.ovirt.org/community/about/community-guidelines/ >> List Archives: >> >https://lists.ovirt.org/archives/list/users@ovirt.org/message/23SAGPLPE5MHCGGQJXUK6D7MMO6CFWWJ/ >> >> >>
My guess is different - a real Load Balancer , like HAproxy, can take care when a backend server is down. Still, it will be nice to learn Darrell's solution. Best Regards, Strahil Nikolov _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/NRA6D6GI24UEF362J5MQC2NWGW6F7DJG/