So, let's summarize:

- Cannot migrate the HE due to "CPU policy".
- HE's CPU is westmere - just like hosts
- You have enough resources on the second HE host (both CPU + MEMORY)

What is the Cluster's CPU type (you can check in UI) ?

Maybe you should enable debugging on various locations to identify the issue.

Anything interesting in the libvirt's log for the HostedEngine.xml on the 
destination host ?


Best Regards,
Strahil Nikolov






В вторник, 22 септември 2020 г., 05:37:18 Гринуич+3, ddqlo <dd...@126.com> 
написа: 





Yes. I can. The host which does not host the HE could be reinstalled 
sucessfully in web UI. After this is done nothing has changed.






在 2020-09-22 03:08:18,"Strahil Nikolov" <hunter86...@yahoo.com> 写道:
>Can you put 1 host in maintenance and use the "Installation" -> "Reinstall" 
>and enable the HE deployment from one of the tabs ?
>
>Best Regards,
>Strahil Nikolov
>
>
>
>
>
>
>В понеделник, 21 септември 2020 г., 06:38:06 Гринуич+3, ddqlo <dd...@126.com> 
>написа: 
>
>
>
>
>
>so strange! After I set global maintenance, powered off and started H The cpu 
>of HE became 'Westmere'(did not change anything). But HE still could not be 
>migrated.
>
>HE xml:
>  <cpu mode='custom' match='exact' check='full'>
>    <model fallback='forbid'>Westmere</model>
>    <topology sockets='16' cores='4' threads='1'/>
>    <feature policy='require' name='vme'/>
>    <feature policy='require' name='pclmuldq'/>
>    <feature policy='require' name='x2apic'/>
>    <feature policy='require' name='hypervisor'/>
>    <feature policy='require' name='arat'/>
>    <numa>
>      <cell id='0' cpus='0-3' memory='16777216' unit='KiB'/>
>    </numa>
>  </cpu>
>
>host capabilities: 
><model>Westmere</model>
>
>cluster cpu type (UI): 
>
>
>host cpu type (UI):
>
>
>HE cpu type (UI):
>
>
>
>
>
>
>
>在 2020-09-19 13:27:35,"Strahil Nikolov" <hunter86...@yahoo.com> 写道:
>>Hm... interesting.
>>
>>The VM is using 'Haswell-noTSX'  while the host is 'Westmere'.
>>
>>In my case I got no difference:
>>
>>[root@ovirt1 ~]# virsh  dumpxml HostedEngine | grep Opteron
>>   <model fallback='forbid'>Opteron_G5</model>
>>[root@ovirt1 ~]# virsh capabilities | grep Opteron
>>     <model>Opteron_G5</model>
>>
>>Did you update the cluster holding the Hosted Engine ?
>>
>>
>>I guess you can try to:
>>
>>- Set global maintenance
>>- Power off the HostedEngine VM
>>- virsh dumpxml HostedEngine > /root/HE.xml
>>- use virsh edit to change the cpu of the HE (non-permanent) change
>>- try to power on the modified HE
>>
>>If it powers on , you can try to migrate it and if it succeeds - then you 
>>should make it permanent.
>>
>>
>>
>>
>>
>>Best Regards,
>>Strahil Nikolov
>>
>>В петък, 18 септември 2020 г., 04:40:39 Гринуич+3, ddqlo <dd...@126.com> 
>>написа: 
>>
>>
>>
>>
>>
>>HE:
>>
>><domain type='kvm' id='1'>
>>  <name>HostedEngine</name>
>>  <uuid>b4e805ff-556d-42bd-a6df-02f5902fd01c</uuid>
>>  <metadata xmlns:ns0="http://ovirt.org/vm/tune/1.0"; 
>>xmlns:ovirt-vm="http://ovirt.org/vm/1.0";>
>>    <ns0:qos/>
>>    <ovirt-vm:vm xmlns:ovirt-vm="http://ovirt.org/vm/1.0";>
>>    <ovirt-vm:clusterVersion>4.3</ovirt-vm:clusterVersion>
>>    <ovirt-vm:destroy_on_reboot type="bool">False</ovirt-vm:destroy_on_reboot>
>>    <ovirt-vm:launchPaused>false</ovirt-vm:launchPaused>
>>    <ovirt-vm:memGuaranteedSize type="int">1024</ovirt-vm:memGuaranteedSize>
>>    <ovirt-vm:minGuaranteedMemoryMb 
>>type="int">1024</ovirt-vm:minGuaranteedMemoryMb>
>>    <ovirt-vm:resumeBehavior>auto_resume</ovirt-vm:resumeBehavior>
>>    <ovirt-vm:startTime type="float">1600307555.19</ovirt-vm:startTime>
>>    <ovirt-vm:device mac_address="56:6f:9b:b0:00:01">
>>        <ovirt-vm:network>external</ovirt-vm:network>
>>        <ovirt-vm:custom>
>>            <ovirt-vm:queues>4</ovirt-vm:queues>
>>        </ovirt-vm:custom>
>>    </ovirt-vm:device>
>>    <ovirt-vm:device mac_address="00:16:3e:50:c1:97">
>>        <ovirt-vm:network>ovirtmgmt</ovirt-vm:network>
>>        <ovirt-vm:custom>
>>            <ovirt-vm:queues>4</ovirt-vm:queues>
>>        </ovirt-vm:custom>
>>    </ovirt-vm:device>
>>    <ovirt-vm:device devtype="disk" name="vda">
>>        
>><ovirt-vm:domainID>c17c1934-332f-464c-8f89-ad72463c00b3</ovirt-vm:domainID>
>>        <ovirt-vm:guestName>/dev/vda2</ovirt-vm:guestName>
>>        
>><ovirt-vm:imageID>8eca143a-4535-4421-bd35-9f5764d67d70</ovirt-vm:imageID>
>>        
>><ovirt-vm:poolID>00000000-0000-0000-0000-000000000000</ovirt-vm:poolID>
>>        <ovirt-vm:shared>exclusive</ovirt-vm:shared>
>>        
>><ovirt-vm:volumeID>ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</ovirt-vm:volumeID>
>>        <ovirt-vm:specParams>
>>            <ovirt-vm:pinToIoThread>1</ovirt-vm:pinToIoThread>
>>        </ovirt-vm:specParams>
>>        <ovirt-vm:volumeChain>
>>            <ovirt-vm:volumeChainNode>
>>                
>><ovirt-vm:domainID>c17c1934-332f-464c-8f89-ad72463c00b3</ovirt-vm:domainID>
>>                
>><ovirt-vm:imageID>8eca143a-4535-4421-bd35-9f5764d67d70</ovirt-vm:imageID>
>>                <ovirt-vm:leaseOffset 
>>type="int">108003328</ovirt-vm:leaseOffset>
>>                
>><ovirt-vm:leasePath>/dev/c17c1934-332f-464c-8f89-ad72463c00b3/leases</ovirt-vm:leasePath>
>>                
>><ovirt-vm:path>/rhev/data-center/mnt/blockSD/c17c1934-332f-464c-8f89-ad72463c00b3/images/8eca143a-4535-4421-bd35-9f5764d67d70/ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</ovirt-vm:path>
>>                
>><ovirt-vm:volumeID>ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</ovirt-vm:volumeID>
>>            </ovirt-vm:volumeChainNode>
>>        </ovirt-vm:volumeChain>
>>    </ovirt-vm:device>
>>    <ovirt-vm:device devtype="disk" name="hdc"/>
>></ovirt-vm:vm>
>>  </metadata>
>>  <maxMemory slots='16' unit='KiB'>67108864</maxMemory>
>>  <memory unit='KiB'>16777216</memory>
>>  <currentMemory unit='KiB'>16777216</currentMemory>
>>  <vcpu placement='static' current='4'>64</vcpu>
>>  <iothreads>1</iothreads>
>>  <resource>
>>    <partition>/machine</partition>
>>  </resource>
>>  <sysinfo type='smbios'>
>>    <system>
>>      <entry name='manufacturer'>oVirt</entry>
>>      <entry name='product'>oVirt Node</entry>
>>      <entry name='version'>7-5.1804.el7.centos</entry>
>>      <entry name='serial'>00000000-0000-0000-0000-0CC47A6B3160</entry>
>>      <entry name='uuid'>b4e805ff-556d-42bd-a6df-02f5902fd01c</entry>
>>    </system>
>>  </sysinfo>
>>  <os>
>>    <type arch='x86_64' machine='pc-i440fx-rhel7.6.0'>hvm</type>
>>    <boot dev='hd'/>
>>    <bios useserial='yes'/>
>>    <smbios mode='sysinfo'/>
>>  </os>
>>  <features>
>>    <acpi/>
>>  </features>
>>  <cpu mode='custom' match='exact' check='full'>
>>    <model fallback='forbid'>Haswell-noTSX</model>
>>    <topology sockets='16' cores='4' threads='1'/>
>>    <feature policy='require' name='vme'/>
>>    <feature policy='require' name='f16c'/>
>>    <feature policy='require' name='rdrand'/>
>>    <feature policy='require' name='hypervisor'/>
>>    <feature policy='require' name='arat'/>
>>    <feature policy='require' name='xsaveopt'/>
>>    <feature policy='require' name='abm'/>
>>    <numa>
>>      <cell id='0' cpus='0-3' memory='16777216' unit='KiB'/>
>>    </numa>
>>  </cpu>
>>  <clock offset='variable' adjustment='0' basis='utc'>
>>    <timer name='rtc' tickpolicy='catchup'/>
>>    <timer name='pit' tickpolicy='delay'/>
>>    <timer name='hpet' present='no'/>
>>  </clock>
>>  <on_poweroff>destroy</on_poweroff>
>>  <on_reboot>destroy</on_reboot>
>>  <on_crash>destroy</on_crash>
>>  <pm>
>>    <suspend-to-mem enabled='no'/>
>>    <suspend-to-disk enabled='no'/>
>>  </pm>
>>  <devices>
>>    <emulator>/usr/libexec/qemu-kvm</emulator>
>>    <disk type='file' device='cdrom'>
>>      <driver name='qemu' error_policy='report'/>
>>      <source startupPolicy='optional'/>
>>      <target dev='hdc' bus='ide'/>
>>      <readonly/>
>>      <alias name='ua-80fde7d5-ee7f-4201-9118-11bc6c3b8530'/>
>>      <address type='drive' controller='0' bus='1' target='0' unit='0'/>
>>    </disk>
>>    <disk type='block' device='disk' snapshot='no'>
>>      <driver name='qemu' type='raw' cache='none' error_policy='stop' 
>>io='native' iothread='1'/>
>>      <source 
>>dev='/var/run/vdsm/storage/c17c1934-332f-464c-8f89-ad72463c00b3/8eca143a-4535-4421-bd35-9f5764d67d70/ae961104-c3b3-4a43-9f46-7fa6bdc2ac33'>
>>        <seclabel model='dac' relabel='no'/>
>>      </source>
>>      <backingStore/>
>>      <target dev='vda' bus='virtio'/>
>>      <serial>8eca143a-4535-4421-bd35-9f5764d67d70</serial>
>>      <alias name='ua-8eca143a-4535-4421-bd35-9f5764d67d70'/>
>>      <address type='pci' domain='0x0000' bus='0x00' slot='0x07' 
>>function='0x0'/>
>>    </disk>
>>    <controller type='scsi' index='0' model='virtio-scsi'>
>>      <driver iothread='1'/>
>>      <alias name='ua-27331e83-03f4-42a3-9554-c41649c02ba4'/>
>>      <address type='pci' domain='0x0000' bus='0x00' slot='0x05' 
>>function='0x0'/>
>>    </controller>
>>    <controller type='ide' index='0'>
>>      <alias name='ide'/>
>>      <address type='pci' domain='0x0000' bus='0x00' slot='0x01' 
>>function='0x1'/>
>>    </controller>
>>    <controller type='virtio-serial' index='0' ports='16'>
>>      <alias name='ua-8fe74299-b60f-4778-8e80-db05393a9489'/>
>>      <address type='pci' domain='0x0000' bus='0x00' slot='0x06' 
>>function='0x0'/>
>>    </controller>
>>    <controller type='usb' index='0' model='piix3-uhci'>
>>      <alias name='usb'/>
>>      <address type='pci' domain='0x0000' bus='0x00' slot='0x01' 
>>function='0x2'/>
>>    </controller>
>>    <controller type='pci' index='0' model='pci-root'>
>>      <alias name='pci.0'/>
>>    </controller>
>>    <lease>
>>      <lockspace>c17c1934-332f-464c-8f89-ad72463c00b3</lockspace>
>>      <key>ae961104-c3b3-4a43-9f46-7fa6bdc2ac33</key>
>>      <target path='/dev/c17c1934-332f-464c-8f89-ad72463c00b3/leases' 
>>offset='108003328'/>
>>    </lease>
>>    <interface type='bridge'>
>>      <mac address='00:16:3e:50:c1:97'/>
>>      <source bridge='ovirtmgmt'/>
>>      <target dev='vnet0'/>
>>      <model type='virtio'/>
>>      <driver name='vhost' queues='4'/>
>>      <filterref filter='vdsm-no-mac-spoofing'/>
>>      <link state='up'/>
>>      <mtu size='1500'/>
>>      <alias name='ua-fada74ee-2338-4cde-a7ba-43a9a636ad6e'/>
>>      <address type='pci' domain='0x0000' bus='0x00' slot='0x03' 
>>function='0x0'/>
>>    </interface>
>>    <interface type='bridge'>
>>      <mac address='56:6f:9b:b0:00:01'/>
>>      <source bridge='external'/>
>>      <target dev='vnet1'/>
>>      <model type='virtio'/>
>>      <driver name='vhost' queues='4'/>
>>      <filterref filter='vdsm-no-mac-spoofing'/>
>>      <link state='up'/>
>>      <mtu size='1500'/>
>>      <alias name='ua-f7b4c949-1f9f-4355-811d-88428c88ce4e'/>
>>      <address type='pci' domain='0x0000' bus='0x00' slot='0x0a' 
>>function='0x0'/>
>>    </interface>
>>    <serial type='unix'>
>>      <source mode='bind' 
>>path='/var/run/ovirt-vmconsole-console/b4e805ff-556d-42bd-a6df-02f5902fd01c.sock'/>
>>      <target type='isa-serial' port='0'>
>>        <model name='isa-serial'/>
>>      </target>
>>      <alias name='serial0'/>
>>    </serial>
>>    <console type='unix'>
>>      <source mode='bind' 
>>path='/var/run/ovirt-vmconsole-console/b4e805ff-556d-42bd-a6df-02f5902fd01c.sock'/>
>>      <target type='serial' port='0'/>
>>      <alias name='serial0'/>
>>    </console>
>>    <channel type='unix'>
>>      <source mode='bind' 
>>path='/var/lib/libvirt/qemu/channels/b4e805ff-556d-42bd-a6df-02f5902fd01c.ovirt-guest-agent.0'/>
>>      <target type='virtio' name='ovirt-guest-agent.0' state='connected'/>
>>      <alias name='channel0'/>
>>      <address type='virtio-serial' controller='0' bus='0' port='1'/>
>>    </channel>
>>    <channel type='unix'>
>>      <source mode='bind' 
>>path='/var/lib/libvirt/qemu/channels/b4e805ff-556d-42bd-a6df-02f5902fd01c.org.qemu.guest_agent.0'/>
>>      <target type='virtio' name='org.qemu.guest_agent.0' state='connected'/>
>>      <alias name='channel1'/>
>>      <address type='virtio-serial' controller='0' bus='0' port='2'/>
>>    </channel>
>>    <channel type='spicevmc'>
>>      <target type='virtio' name='com.redhat.spice.0' state='disconnected'/>
>>      <alias name='channel2'/>
>>      <address type='virtio-serial' controller='0' bus='0' port='3'/>
>>    </channel>
>>    <channel type='unix'>
>>      <source mode='bind' 
>>path='/var/lib/libvirt/qemu/channels/b4e805ff-556d-42bd-a6df-02f5902fd01c.org.ovirt.hosted-engine-setup.0'/>
>>      <target type='virtio' name='org.ovirt.hosted-engine-setup.0' 
>>state='disconnected'/>
>>      <alias name='channel3'/>
>>      <address type='virtio-serial' controller='0' bus='0' port='4'/>
>>    </channel>
>>    <input type='tablet' bus='usb'>
>>      <alias name='input0'/>
>>      <address type='usb' bus='0' port='1'/>
>>    </input>
>>    <input type='mouse' bus='ps2'>
>>      <alias name='input1'/>
>>    </input>
>>    <input type='keyboard' bus='ps2'>
>>      <alias name='input2'/>
>>    </input>
>>    <graphics type='vnc' port='5900' autoport='yes' listen='192.168.1.22' 
>>keymap='en-us' passwdValidTo='1970-01-01T00:00:01'>
>>      <listen type='network' address='192.168.1.22' network='vdsm-external'/>
>>    </graphics>
>>    <graphics type='spice' port='5901' tlsPort='5902' autoport='yes' 
>>listen='192.168.1.22' passwdValidTo='1970-01-01T00:00:01'>
>>      <listen type='network' address='192.168.1.22' network='vdsm-external'/>
>>      <channel name='main' mode='secure'/>
>>      <channel name='display' mode='secure'/>
>>      <channel name='inputs' mode='secure'/>
>>      <channel name='cursor' mode='secure'/>
>>      <channel name='playback' mode='secure'/>
>>      <channel name='record' mode='secure'/>
>>      <channel name='smartcard' mode='secure'/>
>>      <channel name='usbredir' mode='secure'/>
>>    </graphics>
>>    <sound model='ich6'>
>>      <alias name='ua-bd287767-9b83-4e44-ac6f-8b527f9632b8'/>
>>      <address type='pci' domain='0x0000' bus='0x00' slot='0x04' 
>>function='0x0'/>
>>    </sound>
>>    <video>
>>      <model type='qxl' ram='65536' vram='32768' vgamem='16384' heads='1' 
>>primary='yes'/>
>>      <alias name='ua-bcfb6b4b-0b3c-4d5b-ba2d-8ce40a65facd'/>
>>      <address type='pci' domain='0x0000' bus='0x00' slot='0x02' 
>>function='0x0'/>
>>    </video>
>>    <memballoon model='virtio'>
>>      <stats period='5'/>
>>      <alias name='ua-39d36063-8808-47db-9fef-a0baad9f9661'/>
>>      <address type='pci' domain='0x0000' bus='0x00' slot='0x08' 
>>function='0x0'/>
>>    </memballoon>
>>    <rng model='virtio'>
>>      <backend model='random'>/dev/urandom</backend>
>>      <alias name='ua-75516d34-dd8f-4f0f-8496-e1f222a359a8'/>
>>      <address type='pci' domain='0x0000' bus='0x00' slot='0x09' 
>>function='0x0'/>
>>    </rng>
>>  </devices>
>>  <seclabel type='dynamic' model='selinux' relabel='yes'>
>>    <label>system_u:system_r:svirt_t:s0:c162,c716</label>
>>    <imagelabel>system_u:object_r:svirt_image_t:s0:c162,c716</imagelabel>
>>  </seclabel>
>>  <seclabel type='dynamic' model='dac' relabel='yes'>
>>    <label>+107:+107</label>
>>    <imagelabel>+107:+107</imagelabel>
>>  </seclabel>
>></domain>
>>
>>hosts:
>>
>><capabilities>
>>  <host>
>>    <uuid>b25adcec-eef2-49a5-8663-7cdcfc50891b</uuid>
>>    <cpu>
>>      <arch>x86_64</arch>
>>      <model>Westmere</model>
>>      <vendor>Intel</vendor>
>>      <microcode version='34'/>
>>      <counter name='tsc' frequency='3699996000' scaling='no'/>
>>      <topology sockets='1' cores='2' threads='2'/>
>>      <feature name='vme'/>
>>      <feature name='ds'/>
>>      <feature name='acpi'/>
>>      <feature name='ss'/>
>>      <feature name='ht'/>
>>      <feature name='tm'/>
>>      <feature name='pbe'/>
>>      <feature name='pclmuldq'/>
>>      <feature name='dtes64'/>
>>      <feature name='monitor'/>
>>      <feature name='ds_cpl'/>
>>      <feature name='vmx'/>
>>      <feature name='est'/>
>>      <feature name='tm2'/>
>>      <feature name='fma'/>
>>      <feature name='xtpr'/>
>>      <feature name='pdcm'/>
>>      <feature name='pcid'/>
>>      <feature name='movbe'/>
>>      <feature name='tsc-deadline'/>
>>      <feature name='xsave'/>
>>      <feature name='osxsave'/>
>>      <feature name='avx'/>
>>      <feature name='f16c'/>
>>      <feature name='rdrand'/>
>>      <feature name='arat'/>
>>      <feature name='fsgsbase'/>
>>      <feature name='tsc_adjust'/>
>>      <feature name='bmi1'/>
>>      <feature name='avx2'/>
>>      <feature name='smep'/>
>>      <feature name='bmi2'/>
>>      <feature name='erms'/>
>>      <feature name='invpcid'/>
>>      <feature name='xsaveopt'/>
>>      <feature name='pdpe1gb'/>
>>      <feature name='rdtscp'/>
>>      <feature name='abm'/>
>>      <feature name='invtsc'/>
>>      <pages unit='KiB' size='4'/>
>>      <pages unit='KiB' size='2048'/>
>>      <pages unit='KiB' size='1048576'/>
>>    </cpu>
>>    <power_management>
>>      <suspend_mem/>
>>      <suspend_disk/>
>>      <suspend_hybrid/>
>>    </power_management>
>>    <iommu support='no'/>
>>    <migration_features>
>>      <live/>
>>      <uri_transports>
>>        <uri_transport>tcp</uri_transport>
>>        <uri_transport>rdma</uri_transport>
>>      </uri_transports>
>>    </migration_features>
>>    <topology>
>>      <cells num='1'>
>>        <cell id='0'>
>>          <memory unit='KiB'>32903488</memory>
>>          <pages unit='KiB' size='4'>8225872</pages>
>>          <pages unit='KiB' size='2048'>0</pages>
>>          <pages unit='KiB' size='1048576'>0</pages>
>>          <distances>
>>            <sibling id='0' value='10'/>
>>          </distances>
>>          <cpus num='4'>
>>            <cpu id='0' socket_id='0' core_id='0' siblings='0,2'/>
>>            <cpu id='1' socket_id='0' core_id='1' siblings='1,3'/>
>>            <cpu id='2' socket_id='0' core_id='0' siblings='0,2'/>
>>            <cpu id='3' socket_id='0' core_id='1' siblings='1,3'/>
>>          </cpus>
>>        </cell>
>>      </cells>
>>    </topology>
>>    <cache>
>>      <bank id='0' level='3' type='both' size='3' unit='MiB' cpus='0-3'/>
>>    </cache>
>>    <secmodel>
>>      <model>selinux</model>
>>      <doi>0</doi>
>>      <baselabel type='kvm'>system_u:system_r:svirt_t:s0</baselabel>
>>      <baselabel type='qemu'>system_u:system_r:svirt_tcg_t:s0</baselabel>
>>    </secmodel>
>>    <secmodel>
>>      <model>dac</model>
>>      <doi>0</doi>
>>      <baselabel type='kvm'>+107:+107</baselabel>
>>      <baselabel type='qemu'>+107:+107</baselabel>
>>    </secmodel>
>>  </host>
>>  <guest>
>>    <os_type>hvm</os_type>
>>    <arch name='i686'>
>>      <wordsize>32</wordsize>
>>      <emulator>/usr/libexec/qemu-kvm</emulator>
>>      <machine maxCpus='240'>pc-i440fx-rhel7.6.0</machine>
>>      <machine canonical='pc-i440fx-rhel7.6.0' maxCpus='240'>pc</machine>
>>      <machine maxCpus='240'>pc-i440fx-rhel7.0.0</machine>
>>      <machine maxCpus='384'>pc-q35-rhel7.6.0</machine>
>>      <machine canonical='pc-q35-rhel7.6.0' maxCpus='384'>q35</machine>
>>      <machine maxCpus='240'>rhel6.3.0</machine>
>>      <machine maxCpus='240'>rhel6.4.0</machine>
>>      <machine maxCpus='240'>rhel6.0.0</machine>
>>      <machine maxCpus='240'>pc-i440fx-rhel7.5.0</machine>
>>      <machine maxCpus='240'>pc-i440fx-rhel7.1.0</machine>
>>      <machine maxCpus='240'>pc-i440fx-rhel7.2.0</machine>
>>      <machine maxCpus='255'>pc-q35-rhel7.3.0</machine>
>>      <machine maxCpus='240'>rhel6.5.0</machine>
>>      <machine maxCpus='384'>pc-q35-rhel7.4.0</machine>
>>      <machine maxCpus='240'>rhel6.6.0</machine>
>>      <machine maxCpus='240'>rhel6.1.0</machine>
>>      <machine maxCpus='240'>rhel6.2.0</machine>
>>      <machine maxCpus='240'>pc-i440fx-rhel7.3.0</machine>
>>      <machine maxCpus='240'>pc-i440fx-rhel7.4.0</machine>
>>      <machine maxCpus='384'>pc-q35-rhel7.5.0</machine>
>>      <domain type='qemu'/>
>>      <domain type='kvm'>
>>        <emulator>/usr/libexec/qemu-kvm</emulator>
>>      </domain>
>>    </arch>
>>    <features>
>>      <cpuselection/>
>>      <deviceboot/>
>>      <disksnapshot default='on' toggle='no'/>
>>      <acpi default='on' toggle='yes'/>
>>      <apic default='on' toggle='no'/>
>>      <pae/>
>>      <nonpae/>
>>    </features>
>>  </guest>
>>  <guest>
>>    <os_type>hvm</os_type>
>>    <arch name='x86_64'>
>>      <wordsize>64</wordsize>
>>      <emulator>/usr/libexec/qemu-kvm</emulator>
>>      <machine maxCpus='240'>pc-i440fx-rhel7.6.0</machine>
>>      <machine canonical='pc-i440fx-rhel7.6.0' maxCpus='240'>pc</machine>
>>      <machine maxCpus='240'>pc-i440fx-rhel7.0.0</machine>
>>      <machine maxCpus='384'>pc-q35-rhel7.6.0</machine>
>>      <machine canonical='pc-q35-rhel7.6.0' maxCpus='384'>q35</machine>
>>      <machine maxCpus='240'>rhel6.3.0</machine>
>>      <machine maxCpus='240'>rhel6.4.0</machine>
>>      <machine maxCpus='240'>rhel6.0.0</machine>
>>      <machine maxCpus='240'>pc-i440fx-rhel7.5.0</machine>
>>      <machine maxCpus='240'>pc-i440fx-rhel7.1.0</machine>
>>      <machine maxCpus='240'>pc-i440fx-rhel7.2.0</machine>
>>      <machine maxCpus='255'>pc-q35-rhel7.3.0</machine>
>>      <machine maxCpus='240'>rhel6.5.0</machine>
>>      <machine maxCpus='384'>pc-q35-rhel7.4.0</machine>
>>      <machine maxCpus='240'>rhel6.6.0</machine>
>>      <machine maxCpus='240'>rhel6.1.0</machine>
>>      <machine maxCpus='240'>rhel6.2.0</machine>
>>      <machine maxCpus='240'>pc-i440fx-rhel7.3.0</machine>
>>      <machine maxCpus='240'>pc-i440fx-rhel7.4.0</machine>
>>      <machine maxCpus='384'>pc-q35-rhel7.5.0</machine>
>>      <domain type='qemu'/>
>>      <domain type='kvm'>
>>        <emulator>/usr/libexec/qemu-kvm</emulator>
>>      </domain>
>>    </arch>
>>    <features>
>>      <cpuselection/>
>>      <deviceboot/>
>>      <disksnapshot default='on' toggle='no'/>
>>      <acpi default='on' toggle='yes'/>
>>      <apic default='on' toggle='no'/>
>>    </features>
>>  </guest>
>></capabilities>
>>
>>
>>在 2020-09-17 12:00:19,"Strahil Nikolov" <hunter86...@yahoo.com> 写道:
>>>It would be easier if you posted the whole xml.
>>>
>>>What about the sections (in HE xml) starting with:
>>>feature policy=
>>>
>>>Also the hosts have a section which contains:
>>>
>>><feature name=
>>>
>>>If you can share a VM's xml sections for a good VM.
>>>
>>>Best Regards,
>>>Strahil Nikolov
>>>
>>>
>>>
>>>
>>>
>>>
>>>В четвъртък, 17 септември 2020 г., 05:54:12 Гринуич+3, ddqlo <dd...@126.com> 
>>>написа: 
>>>
>>>
>>>
>>>
>>>
>>>HostedEngine:
>>>......
>>><model fallback='forbid'>Haswell-noTSX</model>
>>>......
>>>
>>>both of the hosts:
>>>......
>>><model>Westmere</model>
>>>......
>>>
>>>others vms which can be migrated:
>>>......
>>><model fallback='forbid'>Haswell-noTSX</model>
>>>......
>>>
>>>
>>>
>>>在 2020-09-17 03:03:24,"Strahil Nikolov" <hunter86...@yahoo.com> 写道:
>>>>Can you verify the HostedEngine's CPU ?
>>>>
>>>>1. ssh to the host hosting the HE
>>>>2. alias virsh='virsh -c 
>>>>qemu:///system?authfile=/etc/ovirt-hosted-engine/virsh_auth.conf'
>>>>3. virsh dumpxml HostedEngine
>>>>
>>>>
>>>>Then set the alias for virsh on all Hosts and 'virsh capabilites' should 
>>>>show the Hosts' <cpu><model> .
>>>>
>>>>Best Regards,
>>>>Strahil Nikolov
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>В сряда, 16 септември 2020 г., 10:16:08 Гринуич+3, ddqlo <dd...@126.com> 
>>>>написа: 
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>My gateway was not pingable. I have fixed this problem and now both nodes 
>>>>have a score(3400).
>>>>Yet, hosted engine could not be migrated. Same log in engine.log:
>>>>host filtered out by 'VAR__FILTERTYPE__INTERNAL' filter 'CPU'
>>>>
>>>>
>>>>在 2020-09-16 02:11:09,"Strahil Nikolov" <hunter86...@yahoo.com> 写道:
>>>>>Both nodes have a lower than the usual score (should be 3400 ).
>>>>>Based on the score you are probably suffering from gateway-score-penalty 
>>>>>[1][2].
>>>>>Check if your gateway is pingable.
>>>>>
>>>>>Best Regards,
>>>>>Strahil Nikolov
>>>>>
>>>>>1 - https://www.ovirt.org/images/Hosted-Engine-4.3-deep-dive.pdf(page 8)
>>>>>2 - /etc/ovirt-hosted-engine-ha/agent.conf 
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>В вторник, 15 септември 2020 г., 04:49:48 Гринуич+3, ddqlo <dd...@126.com> 
>>>>>написа: 
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>--== Host node28 (id: 1) status ==--
>>>>>
>>>>>conf_on_shared_storage             : True
>>>>>Status up-to-date                  : True
>>>>>Hostname                           : node28
>>>>>Host ID                            : 1
>>>>>Engine status                      : {"reason": "vm not running on this 
>>>>>host", "health": "bad", "vm": "down_unexpected", "detail": "unknown"}
>>>>>Score                              : 1800
>>>>>stopped                            : False
>>>>>Local maintenance                  : False
>>>>>crc32                              : 4ac6105b
>>>>>local_conf_timestamp               : 1794597
>>>>>Host timestamp                     : 1794597
>>>>>Extra metadata (valid at timestamp):
>>>>>        metadata_parse_version=1
>>>>>        metadata_feature_version=1
>>>>>        timestamp=1794597 (Tue Sep 15 09:47:17 2020)
>>>>>        host-id=1
>>>>>        score=1800
>>>>>        vm_conf_refresh_time=1794597 (Tue Sep 15 09:47:17 2020)
>>>>>        conf_on_shared_storage=True
>>>>>        maintenance=False
>>>>>        state=EngineDown
>>>>>        stopped=False
>>>>>
>>>>>
>>>>>--== Host node22 (id: 2) status ==--
>>>>>
>>>>>conf_on_shared_storage             : True
>>>>>Status up-to-date                  : True
>>>>>Hostname                           : node22
>>>>>Host ID                            : 2
>>>>>Engine status                      : {"health": "good", "vm": "up", 
>>>>>"detail": "Up"}
>>>>>Score                              : 1800
>>>>>stopped                            : False
>>>>>Local maintenance                  : False
>>>>>crc32                              : ffc41893
>>>>>local_conf_timestamp               : 1877876
>>>>>Host timestamp                     : 1877876
>>>>>Extra metadata (valid at timestamp):
>>>>>        metadata_parse_version=1
>>>>>        metadata_feature_version=1
>>>>>        timestamp=1877876 (Tue Sep 15 09:47:13 2020)
>>>>>        host-id=2
>>>>>        score=1800
>>>>>        vm_conf_refresh_time=1877876 (Tue Sep 15 09:47:13 2020)
>>>>>        conf_on_shared_storage=True
>>>>>        maintenance=False
>>>>>        state=EngineUp
>>>>>        stopped=False
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>在 2020-09-09 01:32:55,"Strahil Nikolov" <hunter86...@yahoo.com> 写道:
>>>>>>What is the output of 'hosted-engine --vm-status' on the node where the 
>>>>>>HostedEngine is running ?
>>>>>>
>>>>>>
>>>>>>Best Regards,
>>>>>>Strahil Nikolov
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>В понеделник, 7 септември 2020 г., 03:53:13 Гринуич+3, ddqlo 
>>>>>><dd...@126.com> написа: 
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>I could not find any logs because the migration button is disabled in the 
>>>>>>web UI. It seems that the engine migration operation is prevented at 
>>>>>>first. Any other ideas? Thanks!
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>在 2020-09-01 00:06:19,"Strahil Nikolov" <hunter86...@yahoo.com> 写道:
>>>>>>>I'm running oVirt 4.3.10 and I can migrate my Engine from node to node.
>>>>>>>I had one similar issue , but powering off and on the HE has fixed it.
>>>>>>>
>>>>>>>You have to check the vdsm log on the source and on destination in order 
>>>>>>>to figure out what is going on.
>>>>>>>Also you might consider checking the libvirt logs on the destination.
>>>>>>>
>>>>>>>Best Regards,
>>>>>>>Strahil Nikolov
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>В понеделник, 31 август 2020 г., 10:47:22 Гринуич+3, ddqlo 
>>>>>>><dd...@126.com> написа: 
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>Thanks! The scores of all nodes are not '0'. I find that someone has 
>>>>>>>already asked a question like this. It seems that  this feature has been 
>>>>>>>disabled in 4.3. I am not sure if it is enabled in 4.4.
>>>>>>>
>>>>>>>
>>>>>>>在 2020-08-29 02:27:03,"Strahil Nikolov" <hunter86...@yahoo.com> :
>>>>>>>>Have you checked under a shell the output of 'hosted-engine 
>>>>>>>>--vm-status' . Check the Score of the hosts. Maybe there is a node with 
>>>>>>>>score of '0' ?
>>>>>>>>
>>>>>>>>Best Regards,
>>>>>>>>Strahil Nikolov
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>В вторник, 25 август 2020 г., 13:46:18 Гринуич+3, 董青龙 <dd...@126.com> 
>>>>>>>>написа: 
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>Hi all,
>>>>>>>>        I have an ovirt4.3.10.4 environment of 2 hosts. Normal vms in 
>>>>>>>>this environment could be migrated, but the hosted engine vm could not 
>>>>>>>>be migrated. Anyone can help? Thanks a lot!
>>>>>>>>
>>>>>>>>hosts status:
>>>>>>>>
>>>>>>>>normal vm migration:
>>>>>>>>
>>>>>>>>hosted engine vm migration:
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> 
>>>>>>>>_______________________________________________
>>>>>>>>Users mailing list -- users@ovirt.org
>>>>>>>>To unsubscribe send an email to users-le...@ovirt.org
>>>>>>>>Privacy Statement: https://www.ovirt.org/privacy-policy.html
>>>>>>>>oVirt Code of Conduct: 
>>>>>>>>https://www.ovirt.org/community/about/community-guidelines/
>>>>>>>>List Archives: 
>>>>>>>>https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZXHE2AJX42HNHOMYHTDCUUIU3VQTQTLF/
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> 
>>>>>>>_______________________________________________
>>>>>>>Users mailing list -- users@ovirt.org
>>>>>>>To unsubscribe send an email to users-le...@ovirt.org
>>>>>>>Privacy Statement: https://www.ovirt.org/privacy-policy.html
>>>>>>>oVirt Code of Conduct: 
>>>>>>>https://www.ovirt.org/community/about/community-guidelines/
>>>>>>>List Archives: 
>>>>>>>https://lists.ovirt.org/archives/list/users@ovirt.org/message/IAYLFLC6K42OUPZSZU3P3ZYAU66LGSCD/
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> 
>>>>>>_______________________________________________
>>>>>>Users mailing list -- users@ovirt.org
>>>>>>To unsubscribe send an email to users-le...@ovirt.org
>>>>>>Privacy Statement: https://www.ovirt.org/privacy-policy.html
>>>>>>oVirt Code of Conduct: 
>>>>>>https://www.ovirt.org/community/about/community-guidelines/
>>>>>>List Archives: 
>>>>>>https://lists.ovirt.org/archives/list/users@ovirt.org/message/23ZMAP5K7N5KKX3HOKUEJNVCBH7CY4ZL/
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> 
>>>>
>>>>
>>>>
>>>>
>>>> 
>>>
>>>
>>>
>>>
>>> 
>>
>>
>>
>>
>> 
>
>_______________________________________________
>Users mailing list -- users@ovirt.org
>To unsubscribe send an email to users-le...@ovirt.org
>Privacy Statement: https://www.ovirt.org/privacy-policy.html
>oVirt Code of Conduct: 
>https://www.ovirt.org/community/about/community-guidelines/
>List Archives: 
>https://lists.ovirt.org/archives/list/users@ovirt.org/message/BM3QAMWVBKUATSMHAOEKNJTXLNKLNDBM/



 
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/N6AT4MP2DJGRZPCTX35RHMCSFK5ZVBM3/
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZIF77IM75LR7AF44IQVXHLQCESGX5673/

Reply via email to