Nevermind, problem was the dom_md/metadata file was bad (Pool uid of the
datacenter from which I could not detach was missing).


On 17/4/19 13:39, Eduardo Mayoral wrote:
> Hi.
>
> When trying to detach an ISO datastore I get the following error. Also
> fails if I try to switch the datastore from maintenance to active.
>
>
> 2019-04-17 10:07:23,100+0000 INFO  (jsonrpc/6) [jsonrpc.JsonRpcServer]
> RPC call StoragePool.connectStorageServer succeeded in 0.01 seconds
> (__init__:312)
> 2019-04-17 10:07:23,104+0000 INFO  (jsonrpc/4) [vdsm.api] START
> detachStorageDomain(sdUUID=u'2722942e-0887-439c-81f0-70b60d860060',
> spUUID=u'6901a281-c414-42c6-8b41-3cf6e6bcd788',
> msdUUID=u'00000000-0000-0000-0000-000000000000', masterVersion=14,
> options=None) from=::fff
> f:10.5.72.63,58618, flow_id=0420c8d0-fb37-4b22-9fea-9ee8bac7d41d,
> task_id=9301405a-eaf6-4f1c-a1b6-6537740e2daf (api:48)
> 2019-04-17 10:07:23,105+0000 INFO  (jsonrpc/4) [storage.StoragePool]
> sdUUID=2722942e-0887-439c-81f0-70b60d860060
> spUUID=6901a281-c414-42c6-8b41-3cf6e6bcd788 (sp:1048)
> 2019-04-17 10:07:23,316+0000 INFO  (itmap/0) [IOProcessClient]
> (/pre1-svm-templates.por-ngcs.lan:_ds__pre1__templates__iso01_ovirt3)
> Starting client (__init__:308)
> 2019-04-17 10:07:23,324+0000 INFO  (itmap/1) [IOProcessClient]
> (/pre1-svm-templates.por-ngcs.lan:_ds__pre1__templates__iso01_ovirt)
> Starting client (__init__:308)
> 2019-04-17 10:07:23,327+0000 INFO  (ioprocess/74639) [IOProcess]
> (/pre1-svm-templates.por-ngcs.lan:_ds__pre1__templates__iso01_ovirt3)
> Starting ioprocess (__init__:434)
> 2019-04-17 10:07:23,333+0000 INFO  (ioprocess/74645) [IOProcess]
> (/pre1-svm-templates.por-ngcs.lan:_ds__pre1__templates__iso01_ovirt)
> Starting ioprocess (__init__:434)
> 2019-04-17 10:07:23,335+0000 INFO  (jsonrpc/4) [IOProcessClient]
> (2722942e-0887-439c-81f0-70b60d860060) Starting client (__init__:308)
> 2019-04-17 10:07:23,344+0000 INFO  (ioprocess/74653) [IOProcess]
> (2722942e-0887-439c-81f0-70b60d860060) Starting ioprocess (__init__:434)
> 2019-04-17 10:07:23,350+0000 INFO  (jsonrpc/4) [storage.StorageDomain]
> Removing remnants of deleted images [] (fileSD:740)
> 2019-04-17 10:07:23,352+0000 INFO  (jsonrpc/4) [vdsm.api] FINISH
> detachStorageDomain error=Storage domain not in pool:
> u'domain=2722942e-0887-439c-81f0-70b60d860060,
> pool=6901a281-c414-42c6-8b41-3cf6e6bcd788' from=::ffff:10.5.72.63,58618,
> flow_id=0420c8d0-fb37-4b22-9fea-
> 9ee8bac7d41d, task_id=9301405a-eaf6-4f1c-a1b6-6537740e2daf (api:52)
> 2019-04-17 10:07:23,352+0000 ERROR (jsonrpc/4)
> [storage.TaskManager.Task] (Task='9301405a-eaf6-4f1c-a1b6-6537740e2daf')
> Unexpected error (task:875)
> Traceback (most recent call last):
>   File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line
> 882, in _run
>     return fn(*args, **kargs)
>   File "<string>", line 2, in detachStorageDomain
>   File "/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 50,
> in method
>     ret = func(*args, **kwargs)
>   File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py", line 891,
> in detachStorageDomain
>     pool.detachSD(sdUUID)
>   File "/usr/lib/python2.7/site-packages/vdsm/storage/securable.py",
> line 79, in wrapper
>     return method(self, *args, **kwargs)
>   File "/usr/lib/python2.7/site-packages/vdsm/storage/sp.py", line 1053,
> in detachSD
>     self.validateAttachedDomain(dom)
>   File "/usr/lib/python2.7/site-packages/vdsm/storage/securable.py",
> line 79, in wrapper
>     return method(self, *args, **kwargs)
>   File "/usr/lib/python2.7/site-packages/vdsm/storage/sp.py", line 551,
> in validateAttachedDomain
>     raise se.StorageDomainNotInPool(self.spUUID, dom.sdUUID)
> StorageDomainNotInPool: Storage domain not in pool:
> u'domain=2722942e-0887-439c-81f0-70b60d860060,
> pool=6901a281-c414-42c6-8b41-3cf6e6bcd788'
> 2019-04-17 10:07:23,352+0000 INFO  (jsonrpc/4)
> [storage.TaskManager.Task] (Task='9301405a-eaf6-4f1c-a1b6-6537740e2daf')
> aborting: Task is aborted: "Storage domain not in pool:
> u'domain=2722942e-0887-439c-81f0-70b60d860060,
> pool=6901a281-c414-42c6-8b41-3cf6e6bcd788'" - co
> de 353 (task:1181)
> 2019-04-17 10:07:23,353+0000 ERROR (jsonrpc/4) [storage.Dispatcher]
> FINISH detachStorageDomain error=Storage domain not in pool:
> u'domain=2722942e-0887-439c-81f0-70b60d860060,
> pool=6901a281-c414-42c6-8b41-3cf6e6bcd788' (dispatcher:83)
> 2019-04-17 10:07:23,353+0000 INFO  (jsonrpc/4) [jsonrpc.JsonRpcServer]
> RPC call StorageDomain.detach failed (error 353) in 0.25 seconds
> (__init__:312)
> 2019-04-17 10:07:26,275+0000 INFO  (jsonrpc/2) [api.host] START
> getAllVmStats() from=::ffff:10.5.72.63,58604 (api:48)
> 2019-04-17 10:07:26,275+0000 INFO  (jsonrpc/2) [api.host] FINISH
> getAllVmStats return={'status': {'message': 'Done', 'code': 0},
> 'statsList': (suppressed)} from=::ffff:10.5.72.63,58604 (api:54)
> 2019-04-17 10:07:26,275+0000 INFO  (jsonrpc/2) [jsonrpc.JsonRpcServer]
> RPC call Host.getAllVmStats succeeded in 0.01 seconds (__init__:312)
> 2019-04-17 10:07:29,675+0000 INFO  (jsonrpc/7) [api.host] START
> getStats() from=::ffff:10.5.72.63,58604 (api:48)
> 2019-04-17 10:07:29,739+0000 INFO  (jsonrpc/7) [vdsm.api] START
> repoStats(domains=()) from=::ffff:10.5.72.63,58604,
> task_id=3ec94f9f-7c98-4c49-8710-ffbfd9910ff2 (api:48)
> 2019-04-17 10:07:29,739+0000 INFO  (jsonrpc/7) [vdsm.api] FINISH
> repoStats return={u'a97bbda5-33e4-4e78-890c-f7d2a572179e': {'code': 0,
> 'actual': True, 'version': 4, 'acquired': True, 'delay': '0.000505524',
> 'lastCheck': '3.6', 'valid': True}, u'4121f681-45c3-413a-9c6e-6
> 9d154710ace': {'code': 0, 'actual': True, 'version': 4, 'acquired':
> True, 'delay': '0.000790301', 'lastCheck': '2.7', 'valid': True}}
> from=::ffff:10.5.72.63,58604,
> task_id=3ec94f9f-7c98-4c49-8710-ffbfd9910ff2 (api:54)
> 2019-04-17 10:07:29,740+0000 INFO  (jsonrpc/7) [vdsm.api] START
> multipath_health() from=::ffff:10.5.72.63,58604,
> task_id=f567e03a-8cf4-49dd-84eb-628acb10e1d3 (api:48)
> 2019-04-17 10:07:29,740+0000 INFO  (jsonrpc/7) [vdsm.api] FINISH
> multipath_health return={} from=::ffff:10.5.72.63,58604,
> task_id=f567e03a-8cf4-49dd-84eb-628acb10e1d3 (api:54)
>
> Interestingly, I have 2 datacenters, I can attach, detach, set to
> maintenance or activate this same datastore on the second datacenter,
> but not on this one.
>
> So far I have tried updating OVF, setting master datastore and hosts to
> maintenance and re-activating them.
>
> Started happening on 4.3.2, updating to 4.3.3 makes no difference.
>
> This is what I see on the database. Looks OK to me:
>
> engine=# select * from storage_pool;
>                   id                  |       name        | description
> | storage_pool_type | storage_pool_format_type | status |
> master_domain_version |              spm_vds_id              |
> compatibility_version |         _create_date          |        
> _update_date 
>         | quota_enforcement_type | free_text_comment | is_local
> --------------------------------------+-------------------+-------------+-------------------+--------------------------+--------+-----------------------+--------------------------------------+-----------------------+-------------------------------+-----------------------
> --------+------------------------+-------------------+----------
>  039aef77-77c8-4f04-8d17-22f4b655df28 | dc-pre1-vc03-cl02 |            
> |                   | 4                        |      1
> |                     1 | 42c969f3-a2a6-4784-8506-e6fbc731da9d |
> 4.3                   | 2019-03-14 12:13:24.411468+00 | 2019-04-17
> 08:38:05.53
> 8042+00 |                      0 |                   | f
>  6901a281-c414-42c6-8b41-3cf6e6bcd788 | dc-pre1-vc03-cl01 |            
> |                   | 4                        |      1
> |                    15 | 089cdde4-ecbf-4054-82f9-c106b046d73a |
> 4.3                   | 2019-03-14 12:13:24.365529+00 | 2019-04-17
> 11:10:46.42
> 2305+00 |                      0 |                   | f
> (2 rows)
>
> engine=# select * from storage_pool_iso_map ;
>               storage_id              |          
> storage_pool_id            | status
> --------------------------------------+--------------------------------------+--------
>  2722942e-0887-439c-81f0-70b60d860060 |
> 039aef77-77c8-4f04-8d17-22f4b655df28 |      6
>  2722942e-0887-439c-81f0-70b60d860060 |
> 6901a281-c414-42c6-8b41-3cf6e6bcd788 |      6
>  5d390949-24ec-45e2-bb6d-43b3f30bd286 |
> 039aef77-77c8-4f04-8d17-22f4b655df28 |      3
>  94cb1a4f-a267-479c-88e7-ea9c32e3fd24 |
> 039aef77-77c8-4f04-8d17-22f4b655df28 |      3
>  4121f681-45c3-413a-9c6e-69d154710ace |
> 6901a281-c414-42c6-8b41-3cf6e6bcd788 |      3
>  a97bbda5-33e4-4e78-890c-f7d2a572179e |
> 6901a281-c414-42c6-8b41-3cf6e6bcd788 |      3
> (6 rows)
>
>
> Apart from root cause. I would be interested in recovering from the
> error, detaching the ISO storage, even if it is manually. This
> datacenter is not in production yet, so I have options on the things I
> can try.
>
>
>
-- 
Eduardo Mayoral Jimeno
Systems engineer, platform department. Arsys Internet.
emayo...@arsys.es - +34 941 620 105 - ext 2153
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/RQOKJXZBDGMQM25QXIN5DBIKHMCWTSRF/

Reply via email to