Hi, the stale records are not an issue at all. You can remove them for visually cleaner reports (hoste-engine --clean-metadata command, check the man page), but it makes no difference to the algorithms.
Best regards Martin Sivak On Thu, May 2, 2019 at 11:31 AM Andreas Elvers <[email protected]> wrote: > > I have 5 nodes (node01 to node05). Originally all those nodes were part of > our default datacenter/cluster with a NFS storage domain for vmdisk, engine > and iso-images. All five nodes were engine HA nodes. > Later node01, node02 and node03 were re-installed to have engine HA removed. > Then those nodes were removed from the default cluster. Eventually node01,02 > and 03 were completely re-installed to host our new Ceph/Gluster based > datecenter. The engine is still running on the old default Datacenter. Now I > wish to move it over to our ceph/gluster datacenter. > > when I look at the current output of "hosted-engine --vm-status" I see: > > --== Host node01.infra.solutions.work (id: 1) status ==-- > > conf_on_shared_storage : True > Status up-to-date : False > Hostname : node01.infra.solutions.work > Host ID : 1 > Engine status : unknown stale-data > Score : 0 > stopped : True > Local maintenance : False > crc32 : e437bff4 > local_conf_timestamp : 155627 > Host timestamp : 155877 > Extra metadata (valid at timestamp): > metadata_parse_version=1 > metadata_feature_version=1 > timestamp=155877 (Fri Aug 3 13:09:19 2018) > host-id=1 > score=0 > vm_conf_refresh_time=155627 (Fri Aug 3 13:05:08 2018) > conf_on_shared_storage=True > maintenance=False > state=AgentStopped > stopped=True > > > --== Host node02.infra.solutions.work (id: 2) status ==-- > > conf_on_shared_storage : True > Status up-to-date : False > Hostname : node02.infra.solutions.work > Host ID : 2 > Engine status : unknown stale-data > Score : 0 > stopped : True > Local maintenance : False > crc32 : 11185b04 > local_conf_timestamp : 154757 > Host timestamp : 154856 > Extra metadata (valid at timestamp): > metadata_parse_version=1 > metadata_feature_version=1 > timestamp=154856 (Fri Aug 3 13:22:19 2018) > host-id=2 > score=0 > vm_conf_refresh_time=154757 (Fri Aug 3 13:20:40 2018) > conf_on_shared_storage=True > maintenance=False > state=AgentStopped > stopped=True > > > --== Host node03.infra.solutions.work (id: 3) status ==-- > > conf_on_shared_storage : True > Status up-to-date : False > Hostname : node03.infra.solutions.work > Host ID : 3 > Engine status : unknown stale-data > Score : 0 > stopped : False > Local maintenance : True > crc32 : 9595bed9 > local_conf_timestamp : 14363 > Host timestamp : 14362 > Extra metadata (valid at timestamp): > metadata_parse_version=1 > metadata_feature_version=1 > timestamp=14362 (Thu Aug 2 18:03:25 2018) > host-id=3 > score=0 > vm_conf_refresh_time=14363 (Thu Aug 2 18:03:25 2018) > conf_on_shared_storage=True > maintenance=True > state=LocalMaintenance > stopped=False > > > --== Host node04.infra.solutions.work (id: 4) status ==-- > > conf_on_shared_storage : True > Status up-to-date : True > Hostname : node04.infra.solutions.work > Host ID : 4 > Engine status : {"health": "good", "vm": "up", "detail": > "Up"} > Score : 3400 > stopped : False > Local maintenance : False > crc32 : 245854b1 > local_conf_timestamp : 317498 > Host timestamp : 317498 > Extra metadata (valid at timestamp): > metadata_parse_version=1 > metadata_feature_version=1 > timestamp=317498 (Thu May 2 09:44:47 2019) > host-id=4 > score=3400 > vm_conf_refresh_time=317498 (Thu May 2 09:44:47 2019) > conf_on_shared_storage=True > maintenance=False > state=EngineUp > stopped=False > > > --== Host node05.infra.solutions.work (id: 5) status ==-- > > conf_on_shared_storage : True > Status up-to-date : True > Hostname : node05.infra.solutions.work > Host ID : 5 > Engine status : {"reason": "vm not running on this > host", "health": "bad", "vm": "down", "detail": "unknown"} > Score : 3400 > stopped : False > Local maintenance : False > crc32 : 0711afa0 > local_conf_timestamp : 318044 > Host timestamp : 318044 > Extra metadata (valid at timestamp): > metadata_parse_version=1 > metadata_feature_version=1 > timestamp=318044 (Thu May 2 09:44:45 2019) > host-id=5 > score=3400 > vm_conf_refresh_time=318044 (Thu May 2 09:44:45 2019) > conf_on_shared_storage=True > maintenance=False > state=EngineDown > stopped=False > > > The old node01, node02 and node03 are still present. > > The new incarnations of node01, node02 and node03 will be the destination the > the deployment of the new home of our engine to which I wish to restore the > backup to. But I'm not sure, if (and how) the old date should be removed > first. > _______________________________________________ > Users mailing list -- [email protected] > To unsubscribe send an email to [email protected] > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/[email protected]/message/WF5BNCWZHS2ZFMNCZFFRFCDZHRGSDHT3/ _______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/WRDABZK25VNUOQBSTX3MEHQI5XAOLHSF/

