On February 18, 2020 1:01:32 PM GMT+02:00, "Vrgotic, Marko" 
<m.vrgo...@activevideo.com> wrote:
>Hi Strahil,
>
>
>
>We got to meet at next oVirt conf, as all beer rounds will be on me!
>
>
>
>Ok, just to be sure, so upon those 8 hours:
>
>*   Node1 #ssh root@ovirt-engine “shutdown -h now”   <= engine is
>currently running, via virsh start, on this Node1
>  *   Node1 # virsh undefine HostedEngine
>*   Node2 # hosted-engine –vm-start     <= Node3 still needs to be
>updated
>  *   Node2 # hosted-engine --set-maintenance --mode=none
>  *   Node3 # hosted-engine --set-maintenance --mode=local
>  *   Patch Node3
>  *   Test HA
>
>
>
>
>
>
>
>On 18/02/2020, 11:46, "Strahil Nikolov" <hunter86...@yahoo.com> wrote:
>
>
>
>On February 18, 2020 12:03:31 PM GMT+02:00, "Vrgotic, Marko"
><m.vrgo...@activevideo.com> wrote:
>
>    >Dear Strahil,
>
>    >
>
>    >
>
>    >
>
>    >Thank you for all knowledge sharing and support so far.
>
>    >
>
>    >
>
>    >
>
>>The procedure went fine so far and  I have the Engine running on Node1
>
>    >(it was on Node3).
>
>    >
>
>    >
>
>    >
>
>    >However, I see “strange things” :
>
>    >
>
>    >  *   Engine is running and I have access to WebUI as well - good.
>
>>*   None of the HA Nodes actually show who is hosting the Engine atm –
>
>    >all crowns are gray - Strange
>
>>*   If I look at the list of VMs, I see HostedEngine VM as powered off
>
>    >- Strange
>
>    >
>
>    >
>
>    >
>
>  >Can I safely assume procedure went fine and now the Engine conf sync
>
>    >time of 12 hours started or something went wrong?
>
>    >
>
>    >
>
>    >
>
>    >Kindly awaiting your reply.
>
>    >
>
>    >
>
>    >
>
>    >-----
>
>    >kind regards/met vriendelijke groeten
>
>    >
>
>    >Marko Vrgotic
>
>    >Sr. System Engineer @ System Administration
>
>    >
>
>    >ActiveVideo
>
>    >o: +31 (35) 6774131
>
>    >e: m.vrgo...@activevideo.com<mailto:m.vrgo...@activevideo.com>
>
>    >w: www.activevideo.com<http://www.activevideo.com>
>
>    >
>
>    >ActiveVideo Networks BV. Mediacentrum 3745 Joop van den Endeplein
>
>>1.1217 WJ Hilversum, The Netherlands. The information contained in
>this
>
> >message may be legally privileged and confidential. It is intended to
>
>>be read only by the individual or entity to whom it is addressed or by
>
>    >their designee. If the reader of this message is not the intended
>
>>recipient, you are on notice that any distribution of this message, in
>
>>any form, is strictly prohibited.  If you have received this message
>in
>
>    >error, please immediately notify the sender and/or ActiveVideo
>
>>Networks, LLC by telephone at +1 408.931.9200 and delete or destroy
>any
>
>    >copy of this message.
>
>    >
>
>    >
>
>    >
>
>    >
>
>    >
>
>    >
>
>    >
>
>    >
>
>    >
>
>    >
>
>>On 17/02/2020, 15:04, "Strahil Nikolov" <hunter86...@yahoo.com> wrote:
>
>    >
>
>    >
>
>    >
>
>    >On February 17, 2020 1:55:13 PM GMT+02:00, "Vrgotic, Marko"
>
>    ><m.vrgo...@activevideo.com> wrote:
>
>    >
>
>    >    >Good day Strahil,
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >I believe I found the causing link:
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >HostedEngine.log-20200216:-cpu
>
>    >
>
>>>SandyBridge,pcid=on,spec-ctrl=on,ssbd=on,md-clear=on,vme=on,hypervisor=on,arat=on,xsaveopt=on
>
>    >
>
>    >    >\
>
>    >
>
>    >    >
>
>    >
>
>  >    >HostedEngine.log-20200216:2020-02-13T17:58:38.674630Z qemu-kvm:
>
>    >
>
>    >>warning: host doesn't support requested feature:
>
>    >CPUID.07H:EDX.md-clear
>
>    >
>
>    >    >[bit 10]
>
>    >
>
>    >    >
>
>    >
>
>  >    >HostedEngine.log-20200216:2020-02-13T17:58:38.676205Z qemu-kvm:
>
>    >
>
>    >>warning: host doesn't support requested feature:
>
>    >CPUID.07H:EDX.md-clear
>
>    >
>
>    >    >[bit 10]
>
>    >
>
>    >    >
>
>    >
>
>  >    >HostedEngine.log-20200216:2020-02-13T17:58:38.676901Z qemu-kvm:
>
>    >
>
>    >>warning: host doesn't support requested feature:
>
>    >CPUID.07H:EDX.md-clear
>
>    >
>
>    >    >[bit 10]
>
>    >
>
>    >    >
>
>    >
>
>  >    >HostedEngine.log-20200216:2020-02-13T17:58:38.677616Z qemu-kvm:
>
>    >
>
>    >>warning: host doesn't support requested feature:
>
>    >CPUID.07H:EDX.md-clear
>
>    >
>
>    >    >[bit 10]
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>  >    >The "md-clear" CPU seem to be removed as feature due to spectre
>
>    >
>
>    >    >vulnerabilities.
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
> >>However, when I check the CPU Type/flags of the VMs on the same Host
>
>    >as
>
>    >
>
>>>where Engine is currently, as well as on the other hosts, the
>md-clear
>
>    >
>
>    >    >seems to be only present on the HostedEngine:
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >  *   HostedEngine:
>
>    >
>
>    >    >
>
>    >
>
>    >    >FromwebUI:
>
>    >
>
>    >    >Intel SandyBridge IBRS SSBD Family
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >Via virsh:
>
>    >
>
>    >    >
>
>    >
>
>    >    >#virsh dumpxml
>
>    >
>
>    >    >
>
>    >
>
>    >    ><cpu mode='custom' match='exact' check='full'>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    <model fallback='forbid'>SandyBridge</model>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    <topology sockets='16' cores='4' threads='1'/>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    <feature policy='require' name='pcid'/>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    <feature policy='require' name='spec-ctrl'/>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    <feature policy='require' name='ssbd'/>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    <feature policy='require' name='md-clear'/>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    <feature policy='require' name='vme'/>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    <feature policy='require' name='hypervisor'/>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    <feature policy='require' name='arat'/>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    <feature policy='require' name='xsaveopt'/>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    <numa>
>
>    >
>
>    >    >
>
>    >
>
>    >    >      <cell id='0' cpus='0-3' memory='16777216' unit='KiB'/>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    </numa>
>
>    >
>
>    >    >
>
>    >
>
>    >    ></cpu>
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >  *   OtherVMs:
>
>    >
>
>    >    >
>
>    >
>
>    >    >From webUI:
>
>    >
>
>    >    >(SandyBridge,+pcid,+spec-ctrl,+ssbd)
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >Via virsh:
>
>    >
>
>    >    >
>
>    >
>
>    >    >#virsh dumpxml
>
>    >
>
>    >    >
>
>    >
>
>    >    ><cpu mode='custom' match='exact' check='full'>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    <model fallback='forbid'>SandyBridge</model>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    <topology sockets='16' cores='1' threads='1'/>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    <feature policy='require' name='pcid'/>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    <feature policy='require' name='spec-ctrl'/>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    <feature policy='require' name='ssbd'/>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    <feature policy='require' name='vme'/>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    <feature policy='require' name='hypervisor'/>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    <feature policy='require' name='arat'/>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    <feature policy='require' name='xsaveopt'/>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    <numa>
>
>    >
>
>    >    >
>
>    >
>
>    >    >      <cell id='0' cpus='0-3' memory='4194304' unit='KiB'/>
>
>    >
>
>    >    >
>
>    >
>
>    >    >    </numa>
>
>    >
>
>    >    >
>
>    >
>
>    >    >  </cpu>
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>>   >Strahil, knowing this, do you propose different approach or shall
>I
>
>    >
>
>    >    >just proceed with initially suggested workaround?
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >Kindly awaiting your eply.
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >-----
>
>    >
>
>    >    >kind regards/met vriendelijke groeten
>
>    >
>
>    >    >
>
>    >
>
>    >    >Marko Vrgotic
>
>    >
>
>    >    >Sr. System Engineer @ System Administration
>
>    >
>
>    >    >
>
>    >
>
>    >    >ActiveVideo
>
>    >
>
>    >    >o: +31 (35) 6774131
>
>    >
>
>   >    >e: m.vrgo...@activevideo.com<mailto:m.vrgo...@activevideo.com>
>
>    >
>
>    >    >w: www.activevideo.com<http://www.activevideo.com>
>
>    >
>
>    >    >
>
>    >
>
>>    >ActiveVideo Networks BV. Mediacentrum 3745 Joop van den Endeplein
>
>    >
>
>   >>1.1217 WJ Hilversum, The Netherlands. The information contained in
>
>    >this
>
>    >
>
>> >message may be legally privileged and confidential. It is intended
>to
>
>    >
>
>>>be read only by the individual or entity to whom it is addressed or
>by
>
>    >
>
>>    >their designee. If the reader of this message is not the intended
>
>    >
>
>>>recipient, you are on notice that any distribution of this message,
>in
>
>    >
>
> >>any form, is strictly prohibited.  If you have received this message
>
>    >in
>
>    >
>
>   >    >error, please immediately notify the sender and/or ActiveVideo
>
>    >
>
>  >>Networks, LLC by telephone at +1 408.931.9200 and delete or destroy
>
>    >any
>
>    >
>
>    >    >copy of this message.
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>>>On 16/02/2020, 15:28, "Strahil Nikolov" <hunter86...@yahoo.com>
>wrote:
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >    ssh root@engine "poweroff"
>
>    >
>
>    >    >
>
>    >
>
>>  >ssh host-that-holded-engine "virsh undefine HostedEngine; virsh
>list
>
>    >
>
>    >    >--all"
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >    Lot's of virsh - less vdsm :)
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >    Good luck
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >    Best Regards,
>
>    >
>
>    >    >
>
>    >
>
>    >    >    Strahil Nikolov
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>>  >В неделя, 16 февруари 2020 г., 16:01:44 ч. Гринуич+2, Vrgotic,
>Marko
>
>    >
>
>    >    ><m.vrgo...@activevideo.com> написа:
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >    Hi Strahil,
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>   >>   Regarding step 3:  Stop and undefine the VM on the last working
>
>    >host
>
>    >
>
>    >    >
>
>    >
>
>    >    >One question: How do I undefine HostedEngine from last Host?
>
>    >
>
>>  >Hosted-engine command does not provide such option, or it's just
>not
>
>    >
>
>    >    >obvious.
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >   >    Kindly awaiting your reply.
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >    -----
>
>    >
>
>    >    >
>
>    >
>
>    >    >    kind regards/met vriendelijke groeten
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >    Marko Vrgotic
>
>    >
>
>    >    >
>
>    >
>
>    >    >    ActiveVideo
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >> On 14/02/2020, 18:44, "Strahil Nikolov" <hunter86...@yahoo.com>
>
>    >wrote:
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >On February 14, 2020 4:19:53 PM GMT+02:00, "Vrgotic, Marko"
>
>    >
>
>    >    ><m.vrgo...@activevideo.com> wrote:
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >Good answer Strahil,
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >Thank you, I forgot.
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >Libvirt logs are actually showing the reason why:
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >2020-02-14T12:33:51.847970Z qemu-kvm: -drive
>
>    >
>
>    >    >
>
>    >
>
>>>>file=/var/run/vdsm/storage/054c43fc-1924-4106-9f80-0f2ac62b9886/b019c5fa-8fb5-4bfc-8339-f5b7f590a051/f1ce8ba6-2d3b-4309-bca0-e6a00ce74c75,format=raw,if=none,id=drive-ua-b019c5fa-8fb5-4bfc-8339-f5b7f590a051,serial=b019c5fa-8fb5-4bfc-8339-f5b7f590a051,werror=stop,rerror=stop,cache=none,aio=threads:
>
>    >
>
>    >    >
>
>    >
>
> >>        >'serial' is deprecated, please use the corresponding option
>
>    >of
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >'-device' instead
>
>    >
>
>    >    >
>
>    >
>
> >>     >Spice-Message: 04:33:51.856: setting TLS option 'CipherString'
>
>    >to
>
>    >
>
>   >    >
>
>    >
>
>    >    >        >'kECDHE+FIPS:kDHE+FIPS:kRSA+FIPS:!eNULL:!aNULL' from
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >/etc/pki/tls/spice.cnf configuration file
>
>    >
>
>    >    >
>
>    >
>
> >>  >2020-02-14T12:33:51.863449Z qemu-kvm: warning: CPU(s) not present
>
>    >in
>
>    >
>
>    >    >
>
>    >
>
>>> >any NUMA nodes: CPU 4 [socket-id: 1, core-id: 0, thread-id: 0], CPU
>
>    >5
>
>    >
>
>    >    >
>
>    >
>
>  >    >>[socket-id: 1, core-id: 1, thread-id: 0], CPU 6 [socket-id: 1,
>
>    >
>
>    >    >core-id:
>
>    >
>
>    >    >
>
>    >
>
>>>>2, thread-id: 0], CPU 7 [socket-id: 1, core-id: 3, thread-id: 0],
>CPU
>
>    >
>
>    >    >8
>
>    >
>
>    >    >
>
>    >
>
>  >    >>[socket-id: 2, core-id: 0, thread-id: 0], CPU 9 [socket-id: 2,
>
>    >
>
>    >    >core-id:
>
>    >
>
>    >    >
>
>    >
>
>  >>>1, thread-id: 0], CPU 10 [socket-id: 2, core-id: 2, thread-id: 0],
>
>    >CPU
>
>    >
>
>    >    >
>
>    >
>
> >>    >11 [socket-id: 2, core-id: 3, thread-id: 0], CPU 12 [socket-id:
>
>    >3,
>
>    >
>
>    >    >
>
>    >
>
>    >    >>core-id: 0, thread-id: 0], CPU 13 [socket-id: 3, core-id: 1,
>
>    >
>
>    >    >thread-id:
>
>    >
>
>    >    >
>
>    >
>
>    >    >>0], CPU 14 [socket-id: 3, core-id: 2, thread-id: 0], CPU 15
>
>    >
>
>    >    >[socket-id:
>
>    >
>
>    >   >
>
>    >
>
> >>       >3, core-id: 3, thread-id: 0], CPU 16 [socket-id: 4, core-id:
>
>    >0,
>
>    >
>
>    >    >
>
>    >
>
> >>>thread-id: 0], CPU 17 [socket-id: 4, core-id: 1, thread-id: 0], CPU
>
>    >18
>
>    >
>
>    >    >
>
>    >
>
> >>       >[socket-id: 4, core-id: 2, thread-id: 0], CPU 19 [socket-id:
>
>    >4,
>
>    >
>
>    >    >
>
>    >
>
>    >    >>core-id: 3, thread-id: 0], CPU 20 [socket-id: 5, core-id: 0,
>
>    >
>
>    >    >thread-id:
>
>    >
>
>    >    >
>
>    >
>
>    >    >>0], CPU 21 [socket-id: 5, core-id: 1, thread-id: 0], CPU 22
>
>    >
>
>    >    >[socket-id:
>
>    >
>
>    >    >
>
>    >
>
> >>       >5, core-id: 2, thread-id: 0], CPU 23 [socket-id: 5, core-id:
>
>    >3,
>
>    >
>
>    >    >
>
>    >
>
> >>>thread-id: 0], CPU 24 [socket-id: 6, core-id: 0, thread-id: 0], CPU
>
>    >25
>
>    >
>
>    >    >
>
>    >
>
> >>       >[socket-id: 6, core-id: 1, thread-id: 0], CPU 26 [socket-id:
>
>    >6,
>
>    >
>
>    >    >
>
>    >
>
>    >    >>core-id: 2, thread-id: 0], CPU 27 [socket-id: 6, core-id: 3,
>
>    >
>
>    >    >thread-id:
>
>    >
>
>    >    >
>
>    >
>
>    >    >>0], CPU 28 [socket-id: 7, core-id: 0, thread-id: 0], CPU 29
>
>    >
>
>    >    >[socket-id:
>
>    >
>
>    >    >
>
>    >
>
> >>       >7, core-id: 1, thread-id: 0], CPU 30 [socket-id: 7, core-id:
>
>    >2,
>
>    >
>
>    >    >
>
>    >
>
> >>>thread-id: 0], CPU 31 [socket-id: 7, core-id: 3, thread-id: 0], CPU
>
>    >32
>
>    >
>
>    >   >
>
>    >
>
> >>       >[socket-id: 8, core-id: 0, thread-id: 0], CPU 33 [socket-id:
>
>    >8,
>
>    >
>
>    >    >
>
>    >
>
>    >    >>core-id: 1, thread-id: 0], CPU 34 [socket-id: 8, core-id: 2,
>
>   >
>
>    >    >thread-id:
>
>    >
>
>    >    >
>
>    >
>
>    >    >>0], CPU 35 [socket-id: 8, core-id: 3, thread-id: 0], CPU 36
>
>    >
>
>    >    >[socket-id:
>
>    >
>
>    >    >
>
>    >
>
> >>       >9, core-id: 0, thread-id: 0], CPU 37 [socket-id: 9, core-id:
>
>    >1,
>
>    >
>
>    >    >
>
>    >
>
> >>>thread-id: 0], CPU 38 [socket-id: 9, core-id: 2, thread-id: 0], CPU
>
>    >39
>
>    >
>
>    >    >
>
>    >
>
>  >>      >[socket-id: 9, core-id: 3, thread-id: 0], CPU 40 [socket-id:
>
>    >10,
>
>    >
>
>    >    >
>
>    >
>
>>>        >core-id: 0, thread-id: 0], CPU 41 [socket-id: 10, core-id:
>1,
>
>    >
>
>    >    >
>
>    >
>
>> >>thread-id: 0], CPU 42 [socket-id: 10, core-id: 2, thread-id: 0],
>CPU
>
>    >
>
>    >    >43
>
>    >
>
>    >    >
>
>    >
>
>  >>     >[socket-id: 10, core-id: 3, thread-id: 0], CPU 44 [socket-id:
>
>    >11,
>
>    >
>
>    >    >
>
>    >
>
>>>        >core-id: 0, thread-id: 0], CPU 45 [socket-id: 11, core-id:
>1,
>
>    >
>
>    >    >
>
>    >
>
>> >>thread-id: 0], CPU 46 [socket-id: 11, core-id: 2, thread-id: 0],
>CPU
>
>    >
>
>    >    >47
>
>    >
>
>    >    >
>
>    >
>
>  >>     >[socket-id: 11, core-id: 3, thread-id: 0], CPU 48 [socket-id:
>
>    >12,
>
>    >
>
>    >    >
>
>    >
>
>>>        >core-id: 0, thread-id: 0], CPU 49 [socket-id: 12, core-id:
>1,
>
>    >
>
>    >    >
>
>    >
>
>> >>thread-id: 0], CPU 50 [socket-id: 12, core-id: 2, thread-id: 0],
>CPU
>
>    >
>
>    >    >51
>
>    >
>
>    >    >
>
>    >
>
>  >>     >[socket-id: 12, core-id: 3, thread-id: 0], CPU 52 [socket-id:
>
>    >13,
>
>    >
>
>    >    >
>
>    >
>
>>>        >core-id: 0, thread-id: 0], CPU 53 [socket-id: 13, core-id:
>1,
>
>    >
>
>    >    >
>
>    >
>
>> >>thread-id: 0], CPU 54 [socket-id: 13, core-id: 2, thread-id: 0],
>CPU
>
>    >
>
>    >    >55
>
>    >
>
>    >    >
>
>    >
>
>  >>     >[socket-id: 13, core-id: 3, thread-id: 0], CPU 56 [socket-id:
>
>    >14,
>
>    >
>
>    >    >
>
>    >
>
>>>        >core-id: 0, thread-id: 0], CPU 57 [socket-id: 14, core-id:
>1,
>
>    >
>
>    >    >
>
>    >
>
>> >>thread-id: 0], CPU 58 [socket-id: 14, core-id: 2, thread-id: 0],
>CPU
>
>    >
>
>    >    >59
>
>    >
>
>    >    >
>
>    >
>
>  >>     >[socket-id: 14, core-id: 3, thread-id: 0], CPU 60 [socket-id:
>
>    >15,
>
>    >
>
>    >    >
>
>    >
>
>>>        >core-id: 0, thread-id: 0], CPU 61 [socket-id: 15, core-id:
>1,
>
>    >
>
>    >    >
>
>    >
>
>> >>thread-id: 0], CPU 62 [socket-id: 15, core-id: 2, thread-id: 0],
>CPU
>
>    >
>
>    >    >63
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >[socket-id: 15, core-id: 3, thread-id: 0]
>
>    >
>
>    >    >
>
>    >
>
> >    >>2020-02-14T12:33:51.863475Z qemu-kvm: warning: All CPU(s) up to
>
>    >
>
>    >    >maxcpus
>
>    >
>
>    >    >
>
>    >
>
>    >>  >should be described in NUMA config, ability to start up with
>
>    >partial
>
>    >
>
>    >    >
>
>    >
>
>>    >        >NUMA mappings is obsoleted and will be removed in future
>
>    >
>
>    >    >
>
>    >
>
>    >>   >2020-02-14T12:33:51.863973Z qemu-kvm: warning: host doesn't
>
>    >support
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >requested feature: CPUID.07H:EDX.md-clear [bit 10]
>
>    >
>
>    >    >
>
>    >
>
>    >>   >2020-02-14T12:33:51.865066Z qemu-kvm: warning: host doesn't
>
>    >support
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >requested feature: CPUID.07H:EDX.md-clear [bit 10]
>
>    >
>
>    >    >
>
>    >
>
>    >>   >2020-02-14T12:33:51.865547Z qemu-kvm: warning: host doesn't
>
>    >support
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >requested feature: CPUID.07H:EDX.md-clear [bit 10]
>
>    >
>
>    >    >
>
>    >
>
>    >>   >2020-02-14T12:33:51.865996Z qemu-kvm: warning: host doesn't
>
>    >support
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >requested feature: CPUID.07H:EDX.md-clear [bit 10]
>
>    >
>
>    >    >
>
>    >
>
>>   >        >2020-02-14 12:33:51.932+0000: shutting down,
>reason=failed
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >
>
>    >
>
>    >    >
>
>    >
>
>>>        >But then I wonder if the following is related to error
>above:
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >
>
>    >
>
>    >    >
>
>    >
>
>   >>    >Before I started upgrading Host by Host, all Hosts in Cluster
>
>    >were
>
>    >
>
>    >    >
>
>    >
>
>    >>    >showing CPU Family type: " Intel SandyBridge IBRS SSBD MDS
>
>    >Family"
>
>    >
>
>    >    >
>
>    >
>
>>>  >After first Host was upgraded, his CPU Family type was changed to:
>
>    >"
>
>    >
>
>    >    >
>
>    >
>
> >>     >Intel SandyBridge IBRS SSBD Family" and that forced me to have
>
>    >do
>
>    >
>
>    >    >
>
>    >
>
>   >>      >"downgrade" Cluster family type to " Intel SandyBridge IBRS
>
>    >SSBD
>
>    >
>
>    >    >
>
>    >
>
>    >> >Family" in order to be able to Activate the Host back inside
>
>    >Cluster.
>
>    >
>
>    >    >
>
>    >
>
>>   >>Following further, each Host CPU family type changed after
>Upgrade
>
>    >
>
>    >    >from
>
>    >
>
>    >    >
>
>    >
>
>>   >>"" Intel SandyBridge IBRS SSBD MDS Family" to "" Intel
>SandyBridge
>
>    >
>
>    >    >IBRS
>
>    >
>
>    >    >
>
>    >
>
>   >>        >SSBD Family" , except one where HostedEngine is currently
>
>    >one.
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >
>
>    >
>
>    >    >
>
>    >
>
>    >    >>Could this possibly be the reason why I cannot Migrate the
>
>    >
>
>    >    >HostedEngine
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >now and how to solve it?
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >Kindly awaiting your reply.
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >-----
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >kind regards/met vriendelijke groeten
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >Marko Vrgotic
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >Sr. System Engineer @ System Administration
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >ActiveVideo
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >o: +31 (35) 6774131
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >e: m.vrgo...@activevideo.com
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >w: www.activevideo.com <http://www.activevideo.com>
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >
>
>    >
>
>    >    >
>
>    >
>
>    >>     >ActiveVideo Networks BV. Mediacentrum 3745 Joop van den
>
>    >Endeplein
>
>    >
>
>    >    >
>
>    >
>
>>   >>1.1217 WJ Hilversum, The Netherlands. The information contained
>in
>
>    >
>
>    >    >this
>
>    >
>
>    >    >
>
>    >
>
> >> >message may be legally privileged and confidential. It is intended
>
>    >to
>
>    >
>
>    >    >
>
>    >
>
> >>>be read only by the individual or entity to whom it is addressed or
>
>    >by
>
>    >
>
>    >    >
>
>    >
>
>    >>     >their designee. If the reader of this message is not the
>
>    >intended
>
>    >
>
>    >    >
>
>    >
>
> >>>recipient, you are on notice that any distribution of this message,
>
>    >in
>
>    >
>
>    >    >
>
>    >
>
>> >>any form, is strictly prohibited.  If you have received this
>message
>
>    >
>
>    >    >in
>
>    >
>
>    >    >
>
>    >
>
>    >>        >error, please immediately notify the sender and/or
>
>    >ActiveVideo
>
>    >
>
>    >    >
>
>    >
>
>>  >>Networks, LLC by telephone at +1 408.931.9200 and delete or
>destroy
>
>    >
>
>    >    >any
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >copy of this message.
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >
>
>    >
>
>    >    >
>
>    >
>
>    >>>On 14/02/2020, 14:01, "Strahil Nikolov" <hunter86...@yahoo.com>
>
>    >wrote:
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >
>
>    >
>
>    >    >
>
>    >
>
>>  >        >On February 14, 2020 2:47:04 PM GMT+02:00, "Vrgotic,
>Marko"
>
>    >
>
>    >    >
>
>    >
>
>    >    >        ><m.vrgo...@activevideo.com> wrote:
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >Dear oVirt,
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >
>
>    >
>
>    >    >
>
>    >
>
>> >> >I have problem migrating HostedEngine, only HA VM server, to
>other
>
>    >
>
>    >    >HA
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >nodes.
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >Bit of background story:
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >  *  We have oVirt SHE 4.3.5
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >  *  Three Nodes act as HA pool for SHE
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >  *  Node 3 is currently Hosting SHE
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >  *  Actions:
>
>    >
>
>    >    >
>
>    >
>
>    >>>>*  Put Node1 in Maintenance mode, all VMs were successfully
>
>    >migrated,
>
>    >
>
>    >    >
>
>    >
>
>>   >        >    >than Upgrade packages, Activate Host – all looks
>good
>
>    >
>
>    >    >
>
>    >
>
>    >>>>*  Put Node2 in Maintenance mode, all VMs were successfully
>
>    >migrated,
>
>    >
>
>    >    >
>
>    >
>
>>   >        >    >than Upgrade packages, Activate Host – all looks
>good
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >Not the problem:
>
>    >
>
>    >    >
>
>    >
>
>    >> >    >Try to set  Node3 in Maintenance mode, all VMs were
>
>    >successfully
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >migrated, except HostedEngine.
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >
>
>    >
>
>    >    >
>
>    >
>
>   >>  >    >When attempting Migration of the VM HostedEngine, it fails
>
>    >with
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >following error message:
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:49,960Z INFO
>
>    >
>
>    >    >
>
>    >
>
>    >>  >    >[org.ovirt.engine.core.bll.MigrateVmCommand] (default
>
>    >task-265)
>
>    >
>
>    >    >
>
>    >
>
>    >>   >    >[16f4559e-e262-4c9d-80b4-ec81c2cbf950] Lock Acquired to
>
>    >object
>
>    >
>
>    >    >
>
>    >
>
>>>>>'EngineLock:{exclusiveLocks='[66b6d489-ceb8-486a-951a-355e21f13627=VM]',
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >sharedLocks=''}'
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:49,984Z INFO
>
>    >
>
>    >    >
>
>    >
>
>    >>>    >[org.ovirt.engine.core.bll.scheduling.SchedulingManager]
>
>    >(default
>
>    >
>
>    >    >
>
>    >
>
>   >>  >    >task-265) [16f4559e-e262-4c9d-80b4-ec81c2cbf950] Candidate
>
>    >host
>
>    >
>
>    >    >
>
>    >
>
>  >>>  >'ovirt-sj-04.ictv.com' ('d98843da-bd81-46c9-9425-065b196ac59d')
>
>    >was
>
>    >
>
>    >    >
>
>    >
>
>    >    >>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>
>    >
>
>    >    >(correlation
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >id: null)
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:49,984Z INFO
>
>    >
>
>    >    >
>
>    >
>
>    >>>    >[org.ovirt.engine.core.bll.scheduling.SchedulingManager]
>
>    >(default
>
>    >
>
>    >    >
>
>    >
>
>   >>  >    >task-265) [16f4559e-e262-4c9d-80b4-ec81c2cbf950] Candidate
>
>    >host
>
>    >
>
>    >    >
>
>    >
>
>  >>>  >'ovirt-sj-05.ictv.com' ('e3176705-9fb0-41d6-8721-367dfa2e62bd')
>
>    >was
>
>    >
>
>    >    >
>
>    >
>
>    >    >>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>
>    >
>
>    >    >(correlation
>
>    >
>
>    >   >
>
>    >
>
>    >    >        >    >id: null)
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:49,997Z INFO
>
>    >
>
>    >    >
>
>    >
>
>    >>  >    >[org.ovirt.engine.core.bll.MigrateVmCommand] (default
>
>    >task-265)
>
>    >
>
>    >    >
>
>    >
>
>> >        >    >[16f4559e-e262-4c9d-80b4-ec81c2cbf950] Running
>command:
>
>    >
>
>    >    >
>
>    >
>
>   >>       >    >MigrateVmCommand internal: false. Entities affected :
>
>    >ID:
>
>    >
>
>    >    >
>
>    >
>
>    >    >>  >66b6d489-ceb8-486a-951a-355e21f13627 Type: VMAction group
>
>    >
>
>    >    >MIGRATE_VM
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >with role type USER
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:50,008Z INFO
>
>    >
>
>    >    >
>
>    >
>
>    >>>    >[org.ovirt.engine.core.bll.scheduling.SchedulingManager]
>
>    >(default
>
>    >
>
>    >    >
>
>    >
>
>   >>  >    >task-265) [16f4559e-e262-4c9d-80b4-ec81c2cbf950] Candidate
>
>    >host
>
>    >
>
>    >    >
>
>    >
>
>  >>>  >'ovirt-sj-04.ictv.com' ('d98843da-bd81-46c9-9425-065b196ac59d')
>
>    >was
>
>    >
>
>    >    >
>
>    >
>
>    >    >>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>
>    >
>
>    >    >(correlation
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >id: 16f4559e-e262-4c9d-80b4-ec81c2cbf950)
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:50,008Z INFO
>
>    >
>
>    >    >
>
>    >
>
>    >>>    >[org.ovirt.engine.core.bll.scheduling.SchedulingManager]
>
>    >(default
>
>    >
>
>    >    >
>
>    >
>
>   >>  >    >task-265) [16f4559e-e262-4c9d-80b4-ec81c2cbf950] Candidate
>
>    >host
>
>    >
>
>    >    >
>
>    >
>
>  >>>  >'ovirt-sj-05.ictv.com' ('e3176705-9fb0-41d6-8721-367dfa2e62bd')
>
>    >was
>
>    >
>
>    >    >
>
>    >
>
>    >    >>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>
>    >
>
>    >    >(correlation
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >id: 16f4559e-e262-4c9d-80b4-ec81c2cbf950)
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:50,033Z INFO
>
>    >
>
>    >    >
>
>    >
>
>   >    >>>[org.ovirt.engine.core.vdsbroker.MigrateVDSCommand] (default
>
>    >
>
>    >    >task-265)
>
>    >
>
>    >    >
>
>    >
>
>    >> >    >[16f4559e-e262-4c9d-80b4-ec81c2cbf950] START,
>
>    >MigrateVDSCommand(
>
>    >
>
>    >    >
>
>    >
>
>>>>>MigrateVDSCommandParameters:{hostId='f8d27efb-1527-45f0-97d6-d34a86abaaa2',
>
>    >
>
>    >   >
>
>    >
>
>    >    >        >    >vmId='66b6d489-ceb8-486a-951a-355e21f13627',
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >srcHost='ovirt-sj-03.ictv.com',
>
>    >
>
>    >    >
>
>    >
>
>   >    >        >    >dstVdsId='9808f434-5cd4-48b5-8bbc-e639e391c6a5',
>
>    >
>
>    >    >
>
>    >
>
>    >>  >    >dstHost='ovirt-sj-01.ictv.com:54321',
>
>    >migrationMethod='ONLINE',
>
>    >
>
>    >    >
>
>    >
>
>    >    >>  >tunnelMigration='false', migrationDowntime='0',
>
>    >
>
>    >    >autoConverge='true',
>
>    >
>
>    >    >
>
>    >
>
>    >    >>  >migrateCompressed='false', consoleAddress='null',
>
>    >
>
>    >    >maxBandwidth='40',
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >enableGuestEvents='true',
>maxIncomingMigrations='2',
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >maxOutgoingMigrations='2',
>
>    >
>
>    >    >
>
>    >
>
>    >>   >    >convergenceSchedule='[init=[{name=setDowntime,
>
>    >params=[100]}],
>
>    >
>
>    >    >
>
>    >
>
>    >>        >>stalling=[{limit=1, action={name=setDowntime,
>
>    >params=[150]}},
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >{limit=2,
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >action={name=setDowntime, params=[200]}}, {limit=3,
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >action={name=setDowntime, params=[300]}}, {limit=4,
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >action={name=setDowntime, params=[400]}}, {limit=6,
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >action={name=setDowntime, params=[500]}},
>{limit=-1,
>
>    >
>
>    >    >
>
>    >
>
>>  >> >action={name=abort, params=[]}}]]', dstQemu='10.210.13.11'}),
>log
>
>    >
>
>    >    >id:
>
>    >
>
>    >    >
>
>   >
>
>    >   >        >    >5c126a47
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:50,036Z INFO
>
>    >
>
>    >    >
>
>    >
>
>    >>>
>
>  >>[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand]
>
>    >
>
>    >    >
>
>    >
>
>    >> >    >(default task-265) [16f4559e-e262-4c9d-80b4-ec81c2cbf950]
>
>    >START,
>
>    >
>
>    >    >
>
>    >
>
>>>        >    >MigrateBrokerVDSCommand(HostName =
>ovirt-sj-03.ictv.com,
>
>    >
>
>    >    >
>
>    >
>
>>>>>MigrateVDSCommandParameters:{hostId='f8d27efb-1527-45f0-97d6-d34a86abaaa2',
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >vmId='66b6d489-ceb8-486a-951a-355e21f13627',
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >srcHost='ovirt-sj-03.ictv.com',
>
>    >
>
>    >    >
>
>    >
>
>   >    >        >    >dstVdsId='9808f434-5cd4-48b5-8bbc-e639e391c6a5',
>
>    >
>
>    >    >
>
>    >
>
>    >>  >    >dstHost='ovirt-sj-01.ictv.com:54321',
>
>    >migrationMethod='ONLINE',
>
>    >
>
>    >    >
>
>    >
>
>    >    >>  >tunnelMigration='false', migrationDowntime='0',
>
>    >
>
>    >    >autoConverge='true',
>
>    >
>
>    >    >
>
>    >
>
>    >    >>  >migrateCompressed='false', consoleAddress='null',
>
>    >
>
>    >    >maxBandwidth='40',
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >enableGuestEvents='true',
>maxIncomingMigrations='2',
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >maxOutgoingMigrations='2',
>
>    >
>
>    >    >
>
>    >
>
>    >>   >    >convergenceSchedule='[init=[{name=setDowntime,
>
>    >params=[100]}],
>
>    >
>
>    >    >
>
>    >
>
>    >>        >>stalling=[{limit=1, action={name=setDowntime,
>
>    >params=[150]}},
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >{limit=2,
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >action={name=setDowntime, params=[200]}}, {limit=3,
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >action={name=setDowntime, params=[300]}}, {limit=4,
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >action={name=setDowntime, params=[400]}}, {limit=6,
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >action={name=setDowntime, params=[500]}},
>{limit=-1,
>
>    >
>
>    >    >
>
>    >
>
>>  >> >action={name=abort, params=[]}}]]', dstQemu='10.210.13.11'}),
>log
>
>    >
>
>    >    >id:
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >a0f776d
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:50,043Z INFO
>
>    >
>
>    >    >
>
>    >
>
>    >>>
>
>  >>[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand]
>
>    >
>
>    >    >
>
>    >
>
>    >>>    >(default task-265) [16f4559e-e262-4c9d-80b4-ec81c2cbf950]
>
>    >FINISH,
>
>    >
>
>    >    >
>
>    >
>
> >    >        >    >MigrateBrokerVDSCommand, return: , log id: a0f776d
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:50,046Z INFO
>
>    >
>
>    >    >
>
>    >
>
>   >    >>>[org.ovirt.engine.core.vdsbroker.MigrateVDSCommand] (default
>
>    >
>
>    >    >task-265)
>
>    >
>
>    >    >
>
>    >
>
>    >>>    >[16f4559e-e262-4c9d-80b4-ec81c2cbf950] FINISH,
>
>    >MigrateVDSCommand,
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >return: MigratingFrom, log id: 5c126a47
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:50,052Z INFO
>
>    >
>
>    >    >
>
>    >
>
>>>>>[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>
>    >
>
>    >    >
>
>    >
>
>    >>>  >(default task-265) [16f4559e-e262-4c9d-80b4-ec81c2cbf950]
>
>    >EVENT_ID:
>
>    >
>
>    >    >
>
>    >
>
> >    >>  >VM_MIGRATION_START(62), Migration started (VM: HostedEngine,
>
>    >
>
>    >    >Source:
>
>    >
>
>    >    >
>
>    >
>
>    >>   >    >ovirt-sj-03.ictv.com, Destination: ovirt-sj-01.ictv.com,
>
>    >User:
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >mvrgo...@ictv.com@ictv.com-authz).
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:52,893Z INFO
>
>    >
>
>    >    >
>
>    >
>
>> >        >   
>>[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>    >
>
>    >    >
>
>    >
>
>    >    >>>(ForkJoinPool-1-worker-8) [] VM
>
>    >
>
>    >    >'66b6d489-ceb8-486a-951a-355e21f13627'
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >was reported as Down on VDS
>
>    >
>
>    >    >
>
>    >
>
>    >>     >
>
>    >>'9808f434-5cd4-48b5-8bbc-e639e391c6a5'(ovirt-sj-01.ictv.com)
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:52,893Z INFO
>
>    >
>
>    >    >
>
>    >
>
>    >>    >
>
>    >>[org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
>
>    >
>
>    >    >
>
>    >
>
>>> >    >(ForkJoinPool-1-worker-8) [] START, DestroyVDSCommand(HostName
>
>    >=
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >ovirt-sj-01.ictv.com,
>
>    >
>
>    >    >
>
>    >
>
>>>>>DestroyVmVDSCommandParameters:{hostId='9808f434-5cd4-48b5-8bbc-e639e391c6a5',
>
>    >
>
>    >    >
>
>    >
>
>    >>  >    >vmId='66b6d489-ceb8-486a-951a-355e21f13627',
>
>    >secondsToWait='0',
>
>    >
>
>    >    >
>
>    >
>
>  >    >>  >gracefully='false', reason='', ignoreNoVm='true'}), log id:
>
>    >
>
>    >    >7532a8c0
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:53,217Z INFO
>
>    >
>
>    >    >
>
>    >
>
>    >>    >
>
>    >>[org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
>
>    >
>
>    >    >
>
>    >
>
>  >    >        >    >(ForkJoinPool-1-worker-8) [] Failed to destroy VM
>
>    >
>
>    >    >
>
>    >
>
>    >>>    >'66b6d489-ceb8-486a-951a-355e21f13627' because VM does not
>
>    >exist,
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >ignoring
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:53,217Z INFO
>
>    >
>
>    >    >
>
>    >
>
>    >>    >
>
>    >>[org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
>
>    >
>
>    >    >
>
>    >
>
>>  >> >(ForkJoinPool-1-worker-8) [] FINISH, DestroyVDSCommand, return:
>,
>
>    >
>
>    >    >log
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >id: 7532a8c0
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:53,217Z INFO
>
>    >
>
>    >    >
>
>    >
>
>> >        >   
>>[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >(ForkJoinPool-1-worker-8) [] VM
>
>    >
>
>    >    >
>
>    >
>
>    >    >> >'66b6d489-ceb8-486a-951a-355e21f13627'(HostedEngine) was
>
>    >
>
>    >    >unexpectedly
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >detected as 'Down' on VDS
>
>    >
>
>    >    >
>
>    >
>
>   >    >>>'9808f434-5cd4-48b5-8bbc-e639e391c6a5'(ovirt-sj-01.ictv.com)
>
>    >
>
>    >    >(expected
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >on 'f8d27efb-1527-45f0-97d6-d34a86abaaa2')
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:53,217Z ERROR
>
>    >
>
>    >    >
>
>    >
>
>> >        >   
>>[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>    >
>
>    >    >
>
>    >
>
>   >>>  >(ForkJoinPool-1-worker-8) [] Migration of VM 'HostedEngine' to
>
>    >host
>
>    >
>
>    >   >
>
>    >
>
>    >>  >    >'ovirt-sj-01.ictv.com' failed: VM destroyed during the
>
>    >startup.
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:53,219Z INFO
>
>    >
>
>    >    >
>
>    >
>
>> >        >   
>>[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >(ForkJoinPool-1-worker-15) [] VM
>
>    >
>
>    >    >
>
>    >
>
>   >>  >    >'66b6d489-ceb8-486a-951a-355e21f13627'(HostedEngine) moved
>
>    >from
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >'MigratingFrom' --> 'Up'
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:53,219Z INFO
>
>    >
>
>    >    >
>
>    >
>
>> >        >   
>>[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >(ForkJoinPool-1-worker-15) [] Adding VM
>
>    >
>
>    >    >
>
>    >
>
>   >>>  >'66b6d489-ceb8-486a-951a-355e21f13627'(HostedEngine) to re-run
>
>    >list
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:53,221Z ERROR
>
>    >
>
>    >    >
>
>    >
>
>    >>       >
>
>    >>[org.ovirt.engine.core.vdsbroker.monitoring.VmsMonitoring]
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >(ForkJoinPool-1-worker-15) [] Rerun VM
>
>    >
>
>    >    >
>
>    >
>
>> >        >    >'66b6d489-ceb8-486a-951a-355e21f13627'. Called from
>VDS
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >'ovirt-sj-03.ictv.com'
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:53,259Z INFO
>
>    >
>
>    >    >
>
>    >
>
>    >>>
>
>  >>[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand]
>
>    >
>
>    >    >
>
>    >
>
>>>        >    >(EE-ManagedThreadFactory-engine-Thread-377323) []
>START,
>
>    >
>
>    >    >
>
>    >
>
>>>        >    >MigrateStatusVDSCommand(HostName =
>ovirt-sj-03.ictv.com,
>
>    >
>
>    >    >
>
>    >
>
>>>>>MigrateStatusVDSCommandParameters:{hostId='f8d27efb-1527-45f0-97d6-d34a86abaaa2',
>
>    >
>
>    >    >
>
>    >
>
>    >>  >    >vmId='66b6d489-ceb8-486a-951a-355e21f13627'}), log id:
>
>    >62bac076
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:53,265Z INFO
>
>    >
>
>    >    >
>
>    >
>
>    >>>
>
>  >>[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateStatusVDSCommand]
>
>    >
>
>    >    >
>
>    >
>
>    >>        >    >(EE-ManagedThreadFactory-engine-Thread-377323) []
>
>    >FINISH,
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >MigrateStatusVDSCommand, return: , log id: 62bac076
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:53,277Z WARN
>
>    >
>
>    >    >
>
>    >
>
>>>>>[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>
>    >
>
>    >    >
>
>    >
>
>    >>      >    >(EE-ManagedThreadFactory-engine-Thread-377323) []
>
>    >EVENT_ID:
>
>    >
>
>    >    >
>
>    >
>
>> >>  >VM_MIGRATION_TRYING_RERUN(128), Failed to migrate VM
>HostedEngine
>
>    >
>
>    >    >to
>
>    >
>
>    >    >
>
>    >
>
>    >>   >    >Host ovirt-sj-01.ictv.com . Trying to migrate to another
>
>    >Host.
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:53,330Z INFO
>
>    >
>
>    >    >
>
>    >
>
>>>        >   
>>[org.ovirt.engine.core.bll.scheduling.SchedulingManager]
>
>    >
>
>    >    >
>
>    >
>
>   >> >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] Candidate
>
>    >host
>
>    >
>
>    >    >
>
>    >
>
>  >>>  >'ovirt-sj-04.ictv.com' ('d98843da-bd81-46c9-9425-065b196ac59d')
>
>    >was
>
>    >
>
>    >    >
>
>    >
>
>    >    >>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>
>    >
>
>    >    >(correlation
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >id: null)
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:53,330Z INFO
>
>    >
>
>    >    >
>
>    >
>
>>>        >   
>>[org.ovirt.engine.core.bll.scheduling.SchedulingManager]
>
>    >
>
>    >    >
>
>    >
>
>   >> >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] Candidate
>
>    >host
>
>    >
>
>    >    >
>
>    >
>
>  >>>  >'ovirt-sj-05.ictv.com' ('e3176705-9fb0-41d6-8721-367dfa2e62bd')
>
>    >was
>
>    >
>
>    >    >
>
>    >
>
>    >    >>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>
>    >
>
>    >    >(correlation
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >id: null)
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:53,345Z INFO
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >[org.ovirt.engine.core.bll.MigrateVmCommand]
>
>    >
>
>    >    >
>
>    >
>
>  >    >>    >(EE-ManagedThreadFactory-engine-Thread-377323) [] Running
>
>    >
>
>    >    >command:
>
>    >
>
>    >    >
>
>    >
>
>   >>       >    >MigrateVmCommand internal: false. Entities affected :
>
>    >ID:
>
>    >
>
>    >    >
>
>    >
>
>    >    >>  >66b6d489-ceb8-486a-951a-355e21f13627 Type: VMAction group
>
>    >
>
>    >    >MIGRATE_VM
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >with role type USER
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:53,356Z INFO
>
>    >
>
>    >    >
>
>    >
>
>>>        >   
>>[org.ovirt.engine.core.bll.scheduling.SchedulingManager]
>
>    >
>
>    >    >
>
>    >
>
>   >> >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] Candidate
>
>    >host
>
>    >
>
>    >    >
>
>    >
>
>  >>>  >'ovirt-sj-04.ictv.com' ('d98843da-bd81-46c9-9425-065b196ac59d')
>
>    >was
>
>    >
>
>    >    >
>
>    >
>
>    >    >>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>
>    >
>
>    >    >(correlation
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >id: 16f4559e-e262-4c9d-80b4-ec81c2cbf950)
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:53,356Z INFO
>
>    >
>
>    >    >
>
>    >
>
>>>        >   
>>[org.ovirt.engine.core.bll.scheduling.SchedulingManager]
>
>    >
>
>    >    >
>
>    >
>
>   >> >    >(EE-ManagedThreadFactory-engine-Thread-377323) [] Candidate
>
>    >host
>
>    >
>
>    >    >
>
>    >
>
>  >>>  >'ovirt-sj-05.ictv.com' ('e3176705-9fb0-41d6-8721-367dfa2e62bd')
>
>    >was
>
>    >
>
>    >    >
>
>    >
>
>    >    >>  >filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'HA'
>
>    >
>
>    >    >(correlation
>
>    >
>
>    >    >
>
>    >
>
>   >    >        >    >id: 16f4559e-e262-4c9d-80b4-ec81c2cbf950)
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:53,380Z INFO
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >[org.ovirt.engine.core.vdsbroker.MigrateVDSCommand]
>
>    >
>
>    >    >
>
>    >
>
>>>        >    >(EE-ManagedThreadFactory-engine-Thread-377323) []
>START,
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >MigrateVDSCommand(
>
>    >
>
>    >    >
>
>    >
>
>>>>>MigrateVDSCommandParameters:{hostId='f8d27efb-1527-45f0-97d6-d34a86abaaa2',
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >vmId='66b6d489-ceb8-486a-951a-355e21f13627',
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >srcHost='ovirt-sj-03.ictv.com',
>
>    >
>
>    >    >
>
>    >
>
>   >    >        >    >dstVdsId='33e8ff78-e396-4f40-b43c-685bfaaee9af',
>
>    >
>
>    >    >
>
>    >
>
>    >>  >    >dstHost='ovirt-sj-02.ictv.com:54321',
>
>    >migrationMethod='ONLINE',
>
>    >
>
>    >    >
>
>    >
>
>    >    >>  >tunnelMigration='false', migrationDowntime='0',
>
>    >
>
>    >    >autoConverge='true',
>
>    >
>
>    >    >
>
>    >
>
>    >    >>  >migrateCompressed='false', consoleAddress='null',
>
>    >
>
>    >    >maxBandwidth='40',
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >enableGuestEvents='true',
>maxIncomingMigrations='2',
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >maxOutgoingMigrations='2',
>
>    >
>
>    >    >
>
>    >
>
>    >>   >    >convergenceSchedule='[init=[{name=setDowntime,
>
>    >params=[100]}],
>
>    >
>
>    >    >
>
>    >
>
>    >>        >>stalling=[{limit=1, action={name=setDowntime,
>
>    >params=[150]}},
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >{limit=2,
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >action={name=setDowntime, params=[200]}}, {limit=3,
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >action={name=setDowntime, params=[300]}}, {limit=4,
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >action={name=setDowntime, params=[400]}}, {limit=6,
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >action={name=setDowntime, params=[500]}},
>{limit=-1,
>
>    >
>
>    >    >
>
>    >
>
>>  >> >action={name=abort, params=[]}}]]', dstQemu='10.210.13.12'}),
>log
>
>    >
>
>    >    >id:
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >d99059f
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:53,380Z INFO
>
>    >
>
>    >    >
>
>    >
>
>    >>>
>
>  >>[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand]
>
>    >
>
>    >    >
>
>    >
>
>>>        >    >(EE-ManagedThreadFactory-engine-Thread-377323) []
>START,
>
>    >
>
>    >    >
>
>    >
>
>>>        >    >MigrateBrokerVDSCommand(HostName =
>ovirt-sj-03.ictv.com,
>
>    >
>
>    >    >
>
>    >
>
>>>>>MigrateVDSCommandParameters:{hostId='f8d27efb-1527-45f0-97d6-d34a86abaaa2',
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >vmId='66b6d489-ceb8-486a-951a-355e21f13627',
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >srcHost='ovirt-sj-03.ictv.com',
>
>    >
>
>    >    >
>
>    >
>
>   >    >        >    >dstVdsId='33e8ff78-e396-4f40-b43c-685bfaaee9af',
>
>    >
>
>    >    >
>
>    >
>
>    >>  >    >dstHost='ovirt-sj-02.ictv.com:54321',
>
>    >migrationMethod='ONLINE',
>
>    >
>
>    >    >
>
>    >
>
>    >    >>  >tunnelMigration='false', migrationDowntime='0',
>
>    >
>
>    >    >autoConverge='true',
>
>    >
>
>    >    >
>
>    >
>
>    >    >>  >migrateCompressed='false', consoleAddress='null',
>
>    >
>
>    >    >maxBandwidth='40',
>
>    >
>
>    >   >
>
>    >
>
>>    >        >    >enableGuestEvents='true',
>maxIncomingMigrations='2',
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >maxOutgoingMigrations='2',
>
>    >
>
>    >    >
>
>    >
>
>    >>   >    >convergenceSchedule='[init=[{name=setDowntime,
>
>    >params=[100]}],
>
>    >
>
>    >    >
>
>    >
>
>    >>        >>stalling=[{limit=1, action={name=setDowntime,
>
>    >params=[150]}},
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >{limit=2,
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >action={name=setDowntime, params=[200]}}, {limit=3,
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >action={name=setDowntime, params=[300]}}, {limit=4,
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >action={name=setDowntime, params=[400]}}, {limit=6,
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >action={name=setDowntime, params=[500]}},
>{limit=-1,
>
>    >
>
>    >    >
>
>    >
>
>>  >> >action={name=abort, params=[]}}]]', dstQemu='10.210.13.12'}),
>log
>
>    >
>
>    >    >id:
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >6f0483ac
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:53,386Z INFO
>
>    >
>
>    >    >
>
>    >
>
>    >>>
>
>  >>[org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand]
>
>    >
>
>    >   >
>
>    >
>
>    >>        >    >(EE-ManagedThreadFactory-engine-Thread-377323) []
>
>    >FINISH,
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >MigrateBrokerVDSCommand, return: , log id: 6f0483ac
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:53,388Z INFO
>
>    >
>
>    >    >
>
>    >
>
>>    >        >    >[org.ovirt.engine.core.vdsbroker.MigrateVDSCommand]
>
>    >
>
>    >    >
>
>    >
>
>    >>        >    >(EE-ManagedThreadFactory-engine-Thread-377323) []
>
>    >FINISH,
>
>    >
>
>    >    >
>
>    >
>
>    >>        >    >MigrateVDSCommand, return: MigratingFrom, log id:
>
>    >d99059f
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:53,391Z INFO
>
>    >
>
>    >    >
>
>    >
>
>>>>>[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>
>    >
>
>    >    >
>
>    >
>
>    >>      >    >(EE-ManagedThreadFactory-engine-Thread-377323) []
>
>    >EVENT_ID:
>
>    >
>
>    >    >
>
>    >
>
> >    >>  >VM_MIGRATION_START(62), Migration started (VM: HostedEngine,
>
>    >
>
>    >    >Source:
>
>    >
>
>    >    >
>
>    >
>
>    >>   >    >ovirt-sj-03.ictv.com, Destination: ovirt-sj-02.ictv.com,
>
>    >User:
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >mvrgo...@ictv.com@ictv.com-authz).
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:55,108Z INFO
>
>    >
>
>    >    >
>
>    >
>
>    >>>
>
>    >>[org.ovirt.engine.core.vdsbroker.monitoring.VmsStatisticsFetcher]
>
>    >
>
>    >    >
>
>    >
>
>>  >> >(EE-ManagedThreadFactory-engineScheduled-Thread-96) [] Fetched
>10
>
>    >
>
>    >    >VMs
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >from VDS '33e8ff78-e396-4f40-b43c-685bfaaee9af'
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:55,110Z INFO
>
>    >
>
>    >    >
>
>    >
>
>> >        >   
>>[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>    >
>
>    >    >
>
>    >
>
> >>        >    >(EE-ManagedThreadFactory-engineScheduled-Thread-96) []
>
>    >VM
>
>    >
>
>    >    >
>
>    >
>
>  >>       >    >'66b6d489-ceb8-486a-951a-355e21f13627' is migrating to
>
>    >VDS
>
>    >
>
>    >    >
>
>    >
>
>  >    >> >'33e8ff78-e396-4f40-b43c-685bfaaee9af'(ovirt-sj-02.ictv.com)
>
>    >
>
>    >    >ignoring
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >it in the refresh until migration is done
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:57,224Z INFO
>
>    >
>
>    >    >
>
>    >
>
>> >        >   
>>[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >>(ForkJoinPool-1-worker-15) [] VM
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >'66b6d489-ceb8-486a-951a-355e21f13627'
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >was reported as Down on VDS
>
>    >
>
>    >    >
>
>    >
>
>    >>     >
>
>    >>'33e8ff78-e396-4f40-b43c-685bfaaee9af'(ovirt-sj-02.ictv.com)
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:57,225Z INFO
>
>    >
>
>    >    >
>
>    >
>
>    >>    >
>
>    >>[org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
>
>    >
>
>    >    >
>
>    >
>
>>>>    >(ForkJoinPool-1-worker-15) [] START, DestroyVDSCommand(HostName
>
>    >=
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >ovirt-sj-02.ictv.com,
>
>    >
>
>    >    >
>
>    >
>
>>>>>DestroyVmVDSCommandParameters:{hostId='33e8ff78-e396-4f40-b43c-685bfaaee9af',
>
>    >
>
>    >    >
>
>    >
>
>    >>  >    >vmId='66b6d489-ceb8-486a-951a-355e21f13627',
>
>    >secondsToWait='0',
>
>    >
>
>    >    >
>
>    >
>
>  >    >>  >gracefully='false', reason='', ignoreNoVm='true'}), log id:
>
>    >
>
>    >    >1dec553e
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:57,672Z INFO
>
>    >
>
>    >    >
>
>    >
>
>    >>    >
>
>    >>[org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
>
>    >
>
>    >    >
>
>    >
>
> >    >        >    >(ForkJoinPool-1-worker-15) [] Failed to destroy VM
>
>    >
>
>    >    >
>
>    >
>
>    >>>    >'66b6d489-ceb8-486a-951a-355e21f13627' because VM does not
>
>    >exist,
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >ignoring
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:57,672Z INFO
>
>    >
>
>    >    >
>
>    >
>
>    >>    >
>
>    >>[org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
>
>    >
>
>    >    >
>
>    >
>
>>  >>>(ForkJoinPool-1-worker-15) [] FINISH, DestroyVDSCommand, return:
>,
>
>    >
>
>    >    >log
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >id: 1dec553e
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:57,672Z INFO
>
>    >
>
>    >    >
>
>    >
>
>> >        >   
>>[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >(ForkJoinPool-1-worker-15) [] VM
>
>    >
>
>    >    >
>
>    >
>
>    >    >> >'66b6d489-ceb8-486a-951a-355e21f13627'(HostedEngine) was
>
>    >
>
>    >    >unexpectedly
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >detected as 'Down' on VDS
>
>    >
>
>    >    >
>
>    >
>
>   >    >>>'33e8ff78-e396-4f40-b43c-685bfaaee9af'(ovirt-sj-02.ictv.com)
>
>    >
>
>    >    >(expected
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >on 'f8d27efb-1527-45f0-97d6-d34a86abaaa2')
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:57,672Z ERROR
>
>    >
>
>    >    >
>
>    >
>
>> >        >   
>>[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>    >
>
>    >    >
>
>    >
>
>>   >>  >(ForkJoinPool-1-worker-15) [] Migration of VM 'HostedEngine'
>to
>
>   >
>
>    >    >host
>
>    >
>
>    >    >
>
>    >
>
>    >>  >    >'ovirt-sj-02.ictv.com' failed: VM destroyed during the
>
>    >startup.
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:57,674Z INFO
>
>    >
>
>    >    >
>
>    >
>
>> >        >   
>>[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >(ForkJoinPool-1-worker-8) [] VM
>
>    >
>
>    >    >
>
>    >
>
>   >>  >    >'66b6d489-ceb8-486a-951a-355e21f13627'(HostedEngine) moved
>
>    >from
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >'MigratingFrom' --> 'Up'
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:57,674Z INFO
>
>    >
>
>    >    >
>
>    >
>
>> >        >   
>>[org.ovirt.engine.core.vdsbroker.monitoring.VmAnalyzer]
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >(ForkJoinPool-1-worker-8) [] Adding VM
>
>    >
>
>    >    >
>
>    >
>
>   >>>  >'66b6d489-ceb8-486a-951a-355e21f13627'(HostedEngine) to re-run
>
>    >list
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >2020-02-14 12:33:57,676Z ERROR
>
>    >
>
>    >    >
>
>    >
>
>    >>       >
>
>    >>[org.ovirt.engine.core.vdsbroker.monitoring.VmsMonitoring]
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >(ForkJoinPool-1-worker-8) [] Rerun VM
>
>    >
>
>    >    >
>
>    >
>
>    >    >        >    >'66b6d489-ceb8-48
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>   >    >
>
>    >
>
>    >    >        I am afraid that your suspicions  are  right.
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        What is  the host cpu and the HostedEngine's xml?
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>>  >Have  you checked the xml on any working VM ? What cpu flags  do
>the
>
>    >
>
>    >    >working VMs have ?
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>  >>  How to solve - I think I have a solution , but you might not like
>
>    >it.
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        1. Get current VM xml with virsh
>
>    >
>
>    >    >
>
>    >
>
> >    >2. Set all nodes in maintenance 'hosted-engine --set-maintenance
>
>    >
>
>    >    >--mode=global'
>
>    >
>
>    >    >
>
>    >
>
>    >    >        3. Stop and undefine the VM on the last working host
>
>    >
>
>    >    >
>
>    >
>
>   >    >4. Edit the xml from step 1 and add/remove the flags  that are
>
>    >
>
>    >    >different from the other (working) VMs
>
>    >
>
>    >    >
>
>    >
>
>   >    >        5. Define the HostedEngine on any of the updated hosts
>
>    >
>
>    >    >
>
>    >
>
>    >    >        6. Start the HostedEngine via  virsh.
>
>    >
>
>    >    >
>
>    >
>
> >    >        7. Try with different cpu flags until the engine starts.
>
>    >
>
>    >    >
>
>    >
>
>   >>8. Leave the engine for at least 12 hours , so it will have enough
>
>    >time
>
>    >
>
>    >    >to update  it's  own configuration.
>
>    >
>
>    >    >
>
>    >
>
>  >    >9.  Remove the maintenance  and migrate the engine to the other
>
>    >
>
>    >    >upgraded  host
>
>    >
>
>    >    >
>
>    >
>
>    >    >        10.  Patch the last HostedEngine's host
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>>    >I have done this procedure  in order to recover my engine (except
>
>    >
>
>    >    >changing the cpu flags).
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        Note: You may hit some hiccups:
>
>    >
>
>    >    >
>
>    >
>
>    >    >        A) virsh alias
>
>    >
>
>    >    >
>
>    >
>
>    >    >alias virsh='virsh -c
>
>    >
>
>>    >qemu:///system?authfile=/etc/ovirt-hosted-engine/virsh_auth.conf'
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        B) HostedEngine network missing:
>
>    >
>
>    >    >
>
>    >
>
>    >    >        [root@ovirt1 ~]# virsh net-dumpxml vdsm-ovirtmgmt
>
>    >
>
>    >    >
>
>    >
>
>    >    >        <network>
>
>    >
>
>    >    >
>
>    >
>
>    >    >          <name>vdsm-ovirtmgmt</name>
>
>    >
>
>    >    >
>
>    >
>
>    >    >          <uuid>986c27cf-a1ec-44d8-ae61-ee09ce75c886</uuid>
>
>    >
>
>    >    >
>
>    >
>
>    >    >          <forward mode='bridge'/>
>
>    >
>
>    >    >
>
>    >
>
>    >    >          <bridge name='ovirtmgmt'/>
>
>    >
>
>    >    >
>
>    >
>
>    >    >        </network>
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        Define in xml and add  it via:
>
>    >
>
>    >    >
>
>    >
>
>    >    >        virsh net-define somefile.xml
>
>    >
>
>    >    >
>
>    >
>
>    >    >        C) Missing disk
>
>    >
>
>    >    >
>
>    >
>
>    >    >        Vdsm is creating symlinks like these:
>
>    >
>
>    >    >
>
>    >
>
>  >    >        [root@ovirt1 808423f9-8a5c-40cd-bc9f-2568c85b8c74]# pwd
>
>    >
>
>    >    >
>
>    >
>
>>    >       
>/var/run/vdsm/storage/808423f9-8a5c-40cd-bc9f-2568c85b8c74
>
>    >
>
>    >    >
>
>    >
>
>>    >        [root@ovirt1 808423f9-8a5c-40cd-bc9f-2568c85b8c74]# ls -l
>
>    >
>
>    >    >
>
>    >
>
>    >    >        total 20
>
>    >
>
>    >    >
>
>    >
>
>    >    >lrwxrwxrwx. 1 vdsm kvm 129 Feb  2 19:05
>
>    >
>
>    >    >2c74697a-8bd9-4472-8a98-bf624f3462d5 ->
>
>    >
>
>>>/rhev/data-center/mnt/glusterSD/gluster1:_engine/808423f9-8a5c-40cd-bc9f-2568c85b8c74/images/2c74697a-8bd9-4472-8a98-bf624f3462d5
>
>    >
>
>    >    >
>
>    >
>
>    >    >lrwxrwxrwx. 1 vdsm kvm 129 Feb  2 19:09
>
>    >
>
>    >    >3ec27d6d-921c-4348-b799-f50543b6f919 ->
>
>    >
>
>>>/rhev/data-center/mnt/glusterSD/gluster1:_engine/808423f9-8a5c-40cd-bc9f-2568c85b8c74/images/3ec27d6d-921c-4348-b799-f50543b6f919
>
>    >
>
>    >    >
>
>    >
>
>    >    >lrwxrwxrwx. 1 vdsm kvm 129 Feb  2 19:09
>
>    >
>
>    >    >441abdc8-6cb1-49a4-903f-a1ec0ed88429 ->
>
>    >
>
>>>/rhev/data-center/mnt/glusterSD/gluster1:_engine/808423f9-8a5c-40cd-bc9f-2568c85b8c74/images/441abdc8-6cb1-49a4-903f-a1ec0ed88429
>
>    >
>
>    >    >
>
>    >
>
>    >    >lrwxrwxrwx. 1 vdsm kvm 129 Feb  2 19:09
>
>    >
>
>    >    >94ade632-6ecc-4901-8cec-8e39f3d69cb0 ->
>
>    >
>
>>>/rhev/data-center/mnt/glusterSD/gluster1:_engine/808423f9-8a5c-40cd-bc9f-2568c85b8c74/images/94ade632-6ecc-4901-8cec-8e39f3d69cb0
>
>    >
>
>    >    >
>
>    >
>
>    >    >lrwxrwxrwx. 1 vdsm kvm 129 Feb  2 19:05
>
>    >
>
>    >    >fe62a281-51e9-4b23-87b3-2deb52357304 ->
>
>    >
>
>>>/rhev/data-center/mnt/glusterSD/gluster1:_engine/808423f9-8a5c-40cd-bc9f-2568c85b8c74/images/fe62a281-51e9-4b23-87b3-2deb52357304
>
>    >
>
>    >    >
>
>    >
>
>    >    >        [root@ovirt1 808423f9-8a5c-40cd-bc9f-2568c85b8c74]#
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>> >Just create the link,  so it  points to correct  destinationand
>power
>
>    >
>
>    >    >up again.
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        Good  luck !
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >
>
>    >
>
>    >    >        Best Regards,
>
>    >
>
>    >    >
>
>    >
>
>    >    >        Strahil Nikolov
>
>    >
>
>    >
>
>    >
>
>    >    Hi Marko,
>
>    >
>
>    >
>
>    >
>
>    >    If the  other  VMs work without issues  -> it's worth trying.
>
>    >
>
>    >
>
>    >
>
>    >    Best Regards,
>
>    >
>
>    >    Strahil Nikolov
>
>
>
>    Hi Marco,
>
>
>
>
>
>As  the VM was started manually  (and not by the ovirt-ha-agent) - this
>is expected.
>
>
>
>Keep the engine running and it will update it's own OVF . On the safe
>side - 8 hours is an overkill, but will save you from repeating the
>procedure.
>
>
>
>Last step is to shutdown the VM from inside, undefine it 'virsh
>undefine HostedEngine' and last start it manually via vdsm
>'hosted-engine --vm-start' on one of the updated nodes.
>
>
>
>Once you do a migration from one updated to another updated node, you
>can remove the global maintenance.
>
>    'hosted-engine --set-maintenance --mode=none'
>
>
>
>    Next , you can test the HA of the engine by powering it off:
>
>
>
>    A)ssh to not-yet-updated node and set it in local maintenance:
>
>    hosted-engine --set-maintenance  --mode=local
>
>
>
>    B) ssh engine "poweroff"
>
>
>
>Check if the engine is powered up on the last of the 3 nodes (also
>updated).
>
>
>
>Don't forget to patch the last node (and remove the maintenance after
>the reboot).
>
>
>
>The hard part is over - now you just need to verify that the HA is
>working properly.
>
>
>
>    Best Regards,
>
>    Strahil Nikolov

Hey Marko,

Leave the patch for another day.
If something is not OK - you need to have a host where to power up the 
HostedEngine via vdsm. Call it paranoia, but I prefer  to be on the safe side.

Best Regards,
Strahil Nikolov
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/6JJUSEGCL6QHVWOTNNWYZOGWX4EE5LUS/

Reply via email to