I have 5 nodes (node01 to node05). Originally all those nodes were part of our 
default datacenter/cluster with a NFS storage domain for vmdisk, engine and 
iso-images. All five nodes were engine HA nodes.
Later node01, node02 and node03 were re-installed to have engine HA removed. 
Then those nodes were removed from the default cluster. Eventually node01,02 
and 03 were completely re-installed to host our new Ceph/Gluster based 
datecenter. The engine is still running on the old default Datacenter. Now I 
wish to move it over to our ceph/gluster datacenter.

when I look at the current output of "hosted-engine --vm-status" I see:

--== Host node01.infra.solutions.work (id: 1) status ==--

conf_on_shared_storage             : True
Status up-to-date                  : False
Hostname                           : node01.infra.solutions.work
Host ID                            : 1
Engine status                      : unknown stale-data
Score                              : 0
stopped                            : True
Local maintenance                  : False
crc32                              : e437bff4
local_conf_timestamp               : 155627
Host timestamp                     : 155877
Extra metadata (valid at timestamp):
        metadata_parse_version=1
        metadata_feature_version=1
        timestamp=155877 (Fri Aug  3 13:09:19 2018)
        host-id=1
        score=0
        vm_conf_refresh_time=155627 (Fri Aug  3 13:05:08 2018)
        conf_on_shared_storage=True
        maintenance=False
        state=AgentStopped
        stopped=True


--== Host node02.infra.solutions.work (id: 2) status ==--

conf_on_shared_storage             : True
Status up-to-date                  : False
Hostname                           : node02.infra.solutions.work
Host ID                            : 2
Engine status                      : unknown stale-data
Score                              : 0
stopped                            : True
Local maintenance                  : False
crc32                              : 11185b04
local_conf_timestamp               : 154757
Host timestamp                     : 154856
Extra metadata (valid at timestamp):
        metadata_parse_version=1
        metadata_feature_version=1
        timestamp=154856 (Fri Aug  3 13:22:19 2018)
        host-id=2
        score=0
        vm_conf_refresh_time=154757 (Fri Aug  3 13:20:40 2018)
        conf_on_shared_storage=True
        maintenance=False
        state=AgentStopped
        stopped=True


--== Host node03.infra.solutions.work (id: 3) status ==--

conf_on_shared_storage             : True
Status up-to-date                  : False
Hostname                           : node03.infra.solutions.work
Host ID                            : 3
Engine status                      : unknown stale-data
Score                              : 0
stopped                            : False
Local maintenance                  : True
crc32                              : 9595bed9
local_conf_timestamp               : 14363
Host timestamp                     : 14362
Extra metadata (valid at timestamp):
        metadata_parse_version=1
        metadata_feature_version=1
        timestamp=14362 (Thu Aug  2 18:03:25 2018)
        host-id=3
        score=0
        vm_conf_refresh_time=14363 (Thu Aug  2 18:03:25 2018)
        conf_on_shared_storage=True
        maintenance=True
        state=LocalMaintenance
        stopped=False


--== Host node04.infra.solutions.work (id: 4) status ==--

conf_on_shared_storage             : True
Status up-to-date                  : True
Hostname                           : node04.infra.solutions.work
Host ID                            : 4
Engine status                      : {"health": "good", "vm": "up", "detail": 
"Up"}
Score                              : 3400
stopped                            : False
Local maintenance                  : False
crc32                              : 245854b1
local_conf_timestamp               : 317498
Host timestamp                     : 317498
Extra metadata (valid at timestamp):
        metadata_parse_version=1
        metadata_feature_version=1
        timestamp=317498 (Thu May  2 09:44:47 2019)
        host-id=4
        score=3400
        vm_conf_refresh_time=317498 (Thu May  2 09:44:47 2019)
        conf_on_shared_storage=True
        maintenance=False
        state=EngineUp
        stopped=False


--== Host node05.infra.solutions.work (id: 5) status ==--

conf_on_shared_storage             : True
Status up-to-date                  : True
Hostname                           : node05.infra.solutions.work
Host ID                            : 5
Engine status                      : {"reason": "vm not running on this host", 
"health": "bad", "vm": "down", "detail": "unknown"}
Score                              : 3400
stopped                            : False
Local maintenance                  : False
crc32                              : 0711afa0
local_conf_timestamp               : 318044
Host timestamp                     : 318044
Extra metadata (valid at timestamp):
        metadata_parse_version=1
        metadata_feature_version=1
        timestamp=318044 (Thu May  2 09:44:45 2019)
        host-id=5
        score=3400
        vm_conf_refresh_time=318044 (Thu May  2 09:44:45 2019)
        conf_on_shared_storage=True
        maintenance=False
        state=EngineDown
        stopped=False 


The old node01, node02 and node03 are still present.

The new incarnations of node01, node02 and node03 will be the destination the 
the deployment of the new home of our engine to which I wish to restore the 
backup to. But I'm not sure, if (and how) the old date should be removed first.
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/WF5BNCWZHS2ZFMNCZFFRFCDZHRGSDHT3/

Reply via email to