Hi. When trying to detach an ISO datastore I get the following error. Also fails if I try to switch the datastore from maintenance to active.
2019-04-17 10:07:23,100+0000 INFO (jsonrpc/6) [jsonrpc.JsonRpcServer] RPC call StoragePool.connectStorageServer succeeded in 0.01 seconds (__init__:312) 2019-04-17 10:07:23,104+0000 INFO (jsonrpc/4) [vdsm.api] START detachStorageDomain(sdUUID=u'2722942e-0887-439c-81f0-70b60d860060', spUUID=u'6901a281-c414-42c6-8b41-3cf6e6bcd788', msdUUID=u'00000000-0000-0000-0000-000000000000', masterVersion=14, options=None) from=::fff f:10.5.72.63,58618, flow_id=0420c8d0-fb37-4b22-9fea-9ee8bac7d41d, task_id=9301405a-eaf6-4f1c-a1b6-6537740e2daf (api:48) 2019-04-17 10:07:23,105+0000 INFO (jsonrpc/4) [storage.StoragePool] sdUUID=2722942e-0887-439c-81f0-70b60d860060 spUUID=6901a281-c414-42c6-8b41-3cf6e6bcd788 (sp:1048) 2019-04-17 10:07:23,316+0000 INFO (itmap/0) [IOProcessClient] (/pre1-svm-templates.por-ngcs.lan:_ds__pre1__templates__iso01_ovirt3) Starting client (__init__:308) 2019-04-17 10:07:23,324+0000 INFO (itmap/1) [IOProcessClient] (/pre1-svm-templates.por-ngcs.lan:_ds__pre1__templates__iso01_ovirt) Starting client (__init__:308) 2019-04-17 10:07:23,327+0000 INFO (ioprocess/74639) [IOProcess] (/pre1-svm-templates.por-ngcs.lan:_ds__pre1__templates__iso01_ovirt3) Starting ioprocess (__init__:434) 2019-04-17 10:07:23,333+0000 INFO (ioprocess/74645) [IOProcess] (/pre1-svm-templates.por-ngcs.lan:_ds__pre1__templates__iso01_ovirt) Starting ioprocess (__init__:434) 2019-04-17 10:07:23,335+0000 INFO (jsonrpc/4) [IOProcessClient] (2722942e-0887-439c-81f0-70b60d860060) Starting client (__init__:308) 2019-04-17 10:07:23,344+0000 INFO (ioprocess/74653) [IOProcess] (2722942e-0887-439c-81f0-70b60d860060) Starting ioprocess (__init__:434) 2019-04-17 10:07:23,350+0000 INFO (jsonrpc/4) [storage.StorageDomain] Removing remnants of deleted images [] (fileSD:740) 2019-04-17 10:07:23,352+0000 INFO (jsonrpc/4) [vdsm.api] FINISH detachStorageDomain error=Storage domain not in pool: u'domain=2722942e-0887-439c-81f0-70b60d860060, pool=6901a281-c414-42c6-8b41-3cf6e6bcd788' from=::ffff:10.5.72.63,58618, flow_id=0420c8d0-fb37-4b22-9fea- 9ee8bac7d41d, task_id=9301405a-eaf6-4f1c-a1b6-6537740e2daf (api:52) 2019-04-17 10:07:23,352+0000 ERROR (jsonrpc/4) [storage.TaskManager.Task] (Task='9301405a-eaf6-4f1c-a1b6-6537740e2daf') Unexpected error (task:875) Traceback (most recent call last): File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line 882, in _run return fn(*args, **kargs) File "<string>", line 2, in detachStorageDomain File "/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 50, in method ret = func(*args, **kwargs) File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py", line 891, in detachStorageDomain pool.detachSD(sdUUID) File "/usr/lib/python2.7/site-packages/vdsm/storage/securable.py", line 79, in wrapper return method(self, *args, **kwargs) File "/usr/lib/python2.7/site-packages/vdsm/storage/sp.py", line 1053, in detachSD self.validateAttachedDomain(dom) File "/usr/lib/python2.7/site-packages/vdsm/storage/securable.py", line 79, in wrapper return method(self, *args, **kwargs) File "/usr/lib/python2.7/site-packages/vdsm/storage/sp.py", line 551, in validateAttachedDomain raise se.StorageDomainNotInPool(self.spUUID, dom.sdUUID) StorageDomainNotInPool: Storage domain not in pool: u'domain=2722942e-0887-439c-81f0-70b60d860060, pool=6901a281-c414-42c6-8b41-3cf6e6bcd788' 2019-04-17 10:07:23,352+0000 INFO (jsonrpc/4) [storage.TaskManager.Task] (Task='9301405a-eaf6-4f1c-a1b6-6537740e2daf') aborting: Task is aborted: "Storage domain not in pool: u'domain=2722942e-0887-439c-81f0-70b60d860060, pool=6901a281-c414-42c6-8b41-3cf6e6bcd788'" - co de 353 (task:1181) 2019-04-17 10:07:23,353+0000 ERROR (jsonrpc/4) [storage.Dispatcher] FINISH detachStorageDomain error=Storage domain not in pool: u'domain=2722942e-0887-439c-81f0-70b60d860060, pool=6901a281-c414-42c6-8b41-3cf6e6bcd788' (dispatcher:83) 2019-04-17 10:07:23,353+0000 INFO (jsonrpc/4) [jsonrpc.JsonRpcServer] RPC call StorageDomain.detach failed (error 353) in 0.25 seconds (__init__:312) 2019-04-17 10:07:26,275+0000 INFO (jsonrpc/2) [api.host] START getAllVmStats() from=::ffff:10.5.72.63,58604 (api:48) 2019-04-17 10:07:26,275+0000 INFO (jsonrpc/2) [api.host] FINISH getAllVmStats return={'status': {'message': 'Done', 'code': 0}, 'statsList': (suppressed)} from=::ffff:10.5.72.63,58604 (api:54) 2019-04-17 10:07:26,275+0000 INFO (jsonrpc/2) [jsonrpc.JsonRpcServer] RPC call Host.getAllVmStats succeeded in 0.01 seconds (__init__:312) 2019-04-17 10:07:29,675+0000 INFO (jsonrpc/7) [api.host] START getStats() from=::ffff:10.5.72.63,58604 (api:48) 2019-04-17 10:07:29,739+0000 INFO (jsonrpc/7) [vdsm.api] START repoStats(domains=()) from=::ffff:10.5.72.63,58604, task_id=3ec94f9f-7c98-4c49-8710-ffbfd9910ff2 (api:48) 2019-04-17 10:07:29,739+0000 INFO (jsonrpc/7) [vdsm.api] FINISH repoStats return={u'a97bbda5-33e4-4e78-890c-f7d2a572179e': {'code': 0, 'actual': True, 'version': 4, 'acquired': True, 'delay': '0.000505524', 'lastCheck': '3.6', 'valid': True}, u'4121f681-45c3-413a-9c6e-6 9d154710ace': {'code': 0, 'actual': True, 'version': 4, 'acquired': True, 'delay': '0.000790301', 'lastCheck': '2.7', 'valid': True}} from=::ffff:10.5.72.63,58604, task_id=3ec94f9f-7c98-4c49-8710-ffbfd9910ff2 (api:54) 2019-04-17 10:07:29,740+0000 INFO (jsonrpc/7) [vdsm.api] START multipath_health() from=::ffff:10.5.72.63,58604, task_id=f567e03a-8cf4-49dd-84eb-628acb10e1d3 (api:48) 2019-04-17 10:07:29,740+0000 INFO (jsonrpc/7) [vdsm.api] FINISH multipath_health return={} from=::ffff:10.5.72.63,58604, task_id=f567e03a-8cf4-49dd-84eb-628acb10e1d3 (api:54) Interestingly, I have 2 datacenters, I can attach, detach, set to maintenance or activate this same datastore on the second datacenter, but not on this one. So far I have tried updating OVF, setting master datastore and hosts to maintenance and re-activating them. Started happening on 4.3.2, updating to 4.3.3 makes no difference. This is what I see on the database. Looks OK to me: engine=# select * from storage_pool; id | name | description | storage_pool_type | storage_pool_format_type | status | master_domain_version | spm_vds_id | compatibility_version | _create_date | _update_date | quota_enforcement_type | free_text_comment | is_local --------------------------------------+-------------------+-------------+-------------------+--------------------------+--------+-----------------------+--------------------------------------+-----------------------+-------------------------------+----------------------- --------+------------------------+-------------------+---------- 039aef77-77c8-4f04-8d17-22f4b655df28 | dc-pre1-vc03-cl02 | | | 4 | 1 | 1 | 42c969f3-a2a6-4784-8506-e6fbc731da9d | 4.3 | 2019-03-14 12:13:24.411468+00 | 2019-04-17 08:38:05.53 8042+00 | 0 | | f 6901a281-c414-42c6-8b41-3cf6e6bcd788 | dc-pre1-vc03-cl01 | | | 4 | 1 | 15 | 089cdde4-ecbf-4054-82f9-c106b046d73a | 4.3 | 2019-03-14 12:13:24.365529+00 | 2019-04-17 11:10:46.42 2305+00 | 0 | | f (2 rows) engine=# select * from storage_pool_iso_map ; storage_id | storage_pool_id | status --------------------------------------+--------------------------------------+-------- 2722942e-0887-439c-81f0-70b60d860060 | 039aef77-77c8-4f04-8d17-22f4b655df28 | 6 2722942e-0887-439c-81f0-70b60d860060 | 6901a281-c414-42c6-8b41-3cf6e6bcd788 | 6 5d390949-24ec-45e2-bb6d-43b3f30bd286 | 039aef77-77c8-4f04-8d17-22f4b655df28 | 3 94cb1a4f-a267-479c-88e7-ea9c32e3fd24 | 039aef77-77c8-4f04-8d17-22f4b655df28 | 3 4121f681-45c3-413a-9c6e-69d154710ace | 6901a281-c414-42c6-8b41-3cf6e6bcd788 | 3 a97bbda5-33e4-4e78-890c-f7d2a572179e | 6901a281-c414-42c6-8b41-3cf6e6bcd788 | 3 (6 rows) Apart from root cause. I would be interested in recovering from the error, detaching the ISO storage, even if it is manually. This datacenter is not in production yet, so I have options on the things I can try. -- Eduardo Mayoral Jimeno Systems engineer, platform department. Arsys Internet. emayo...@arsys.es - +34 941 620 105 - ext 2153 _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/SYAMOKGD6DJCK5GRNSDY2OPIEKZBRY4I/