Hey folks,

quick (I hope) question: On my 3-node cluster I am swapping out all the SSDs with fewer but higher capacity ones. So I took one node down (maintenance, stop), then removed all SSDs, set up a new RAID, set up lvm and gluster, let it resync. Gluster health status shows no unsynced entries.

Uppon going from maintenance to online from ovirt management It goes into non-operational status, vdsm log on the node shows:

2021-01-17 11:13:29,051+0100 INFO (jsonrpc/6) [api.host] START getAllVmStats() from=::1,48580 (api:48) 2021-01-17 11:13:29,051+0100 INFO (jsonrpc/6) [api.host] FINISH getAllVmStats return={'status': {'message': 'Done', 'code': 0}, 'statsList': (suppressed)} from=::1,48580 (api:54) 2021-01-17 11:13:29,052+0100 INFO (jsonrpc/6) [jsonrpc.JsonRpcServer] RPC call Host.getAllVmStats succeeded in 0.00 seconds (__init__:312) 2021-01-17 11:13:30,420+0100 WARN (monitor/4a62cdb) [storage.LVM] Reloading VGs failed (vgs=[u'4a62cdb4-b314-4c7f-804e-8e7275518a7f'] rc=5 out=[] err=[' Volume group "4a62cdb4-b314-4c7f-804e-8e7275518a7f" not found', ' Cannot process volume group 4a62cdb4-b314-4c7f-804e-8e7275518a7f']) (lvm:470) 2021-01-17 11:13:30,424+0100 ERROR (monitor/4a62cdb) [storage.Monitor] Setting up monitor for 4a62cdb4-b314-4c7f-804e-8e7275518a7f failed (monitor:330)
Traceback (most recent call last):
File "/usr/lib/python2.7/site-packages/vdsm/storage/monitor.py", line 327, in _setupLoop
    self._setupMonitor()
File "/usr/lib/python2.7/site-packages/vdsm/storage/monitor.py", line 349, in _setupMonitor
    self._produceDomain()
File "/usr/lib/python2.7/site-packages/vdsm/utils.py", line 159, in wrapper
    value = meth(self, *a, **kw)
File "/usr/lib/python2.7/site-packages/vdsm/storage/monitor.py", line 367, in _produceDomain
    self.domain = sdCache.produce(self.sdUUID)
File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 110, in produce
    domain.getRealDomain()
File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 51, in getRealDomain
    return self._cache._realProduce(self._sdUUID)
File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 134, in _realProduce
    domain = self._findDomain(sdUUID)
File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 151, in _findDomain
    return findMethod(sdUUID)
File "/usr/lib/python2.7/site-packages/vdsm/storage/sdc.py", line 176, in _findUnfetchedDomain
    raise se.StorageDomainDoesNotExist(sdUUID)


I assume due to the changed LVM UUID it fails, right? Can I someone fix/change the UUID and get the node back up again? It does not seem to be a major issue, to be honest.

I can see the gluster mount (what ovirt mounts when it onlines a node) already, and gluster is happy too.

Any help is appreciated!

-Chris.

--
with kind regards,
mit freundlichen Gruessen,

Christian Reiss

Attachment: OpenPGP_0x44E29126ABCD43C5.asc
Description: application/pgp-keys

<<attachment: email.vcf>>

Attachment: OpenPGP_signature
Description: OpenPGP digital signature

_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/LS4OEXZOE6UA4CDQYXFKC3TZCCO42SU4/

Reply via email to