On Wed, Jun 20, 2018 at 12:17 PM Albl, Oliver <oliver.a...@fabasoft.com> wrote:
> Nir, > > > > thank you! I could successfully recover the corrupt metadata blocks on > all affected storage domains! > > > > Affected storage domains were created between January and March this year, > accessed by 40 hosts in parallel with changing SPMs…unfortunately matching > vdsm logs are not available anymore. I could privide a vsdm.log backup from > SPM showing the first occurrence (+15 minutes prior the error). > These logs may be useful to understand why the metadata was wiped. Please file a bug and attach these logs. Nir > > > All the best, > > Oliver > > > > *Von:* Nir Soffer <nsof...@redhat.com> > *Gesendet:* Mittwoch, 20. Juni 2018 10:01 > *An:* Bruckner, Simone <simone.bruck...@fabasoft.com> > *Cc:* users@ovirt.org > > > *Betreff:* [ovirt-users] Re: Failed to update VMs/Templates OVF data, > cannot change SPM > > > > On Wed, Jun 20, 2018 at 10:33 AM Bruckner, Simone < > simone.bruck...@fabasoft.com> wrote: > > Hi Nir, > > > > I identified the reason for the failing OVF updates on the initial VG – > metadata was affected by blkdiscard tests in scope of > https://bugzilla.redhat.com/show_bug.cgi?id=1562369 > > > > However, the OVF updates are failing on other installations as well (on 2 > out of 40 storage domains). Here is the output of your commands: > > > > # lvs -o vg_name,lv_name,tags | grep 3ad1987a-8b7d-426d-9d51-4a78cb0a888f > > 3ad1987a-8b7d-426d-9d51-4a78cb0a888f > 212d644c-0155-4999-9df9-72bacfc7f141 > IU_0ebefe5e-9053-4bf1-bdfd-fdb26579c179,MD_4,PU_00000000-0000-0000-0000-000000000000 > > 3ad1987a-8b7d-426d-9d51-4a78cb0a888f > 94f519de-bc19-4557-82c4-86bbcfc5dd2f > IU_60d9eec7-951f-4594-ae99-7d31318ee3a9,MD_5,PU_00000000-0000-0000-0000-000000000000 > > 3ad1987a-8b7d-426d-9d51-4a78cb0a888f ids > > 3ad1987a-8b7d-426d-9d51-4a78cb0a888f inbox > > 3ad1987a-8b7d-426d-9d51-4a78cb0a888f leases > > 3ad1987a-8b7d-426d-9d51-4a78cb0a888f master > > 3ad1987a-8b7d-426d-9d51-4a78cb0a888f metadata > > 3ad1987a-8b7d-426d-9d51-4a78cb0a888f outbox > > 3ad1987a-8b7d-426d-9d51-4a78cb0a888f xleases > > > > Looks good > > > > > > # for i in 4 5; do > > dd if=/dev/3ad1987a-8b7d-426d-9d51-4a78cb0a888f/metadata bs=512 count=1 > skip=$i of=metadata.$i > > done > > 1+0 records in > > 1+0 records out > > 512 bytes (512 B) copied, 0.00121297 s, 422 kB/s > > 1+0 records in > > 1+0 records out > > 512 bytes (512 B) copied, 0.000735026 s, 697 kB/s > > > > # file metadata.* > > metadata.4: data > > metadata.5: ASCII text > > > > # cat metadata.5 > > DOMAIN=3ad1987a-8b7d-426d-9d51-4a78cb0a888f > > CTIME=1520597691 > > FORMAT=RAW > > DISKTYPE=OVFS > > LEGALITY=LEGAL > > SIZE=262144 > > VOLTYPE=LEAF > > DESCRIPTION={"Updated":true,"Size":4669440,"Last Updated":"Fri Jun 08 > 09:51:18 CEST 2018","Storage > Domains":[{"uuid":"3ad1987a-8b7d-426d-9d51-4a78cb0a888f"}],"Disk > Description":"OVF_STORE"} > > IMAGE=60d9eec7-951f-4594-ae99-7d31318ee3a9 > > PUUID=00000000-0000-0000-0000-000000000000 > > MTIME=0 > > POOL_UUID= > > TYPE=PREALLOCATED > > GEN=0 > > EOF > > > > Looks good > > > > # od -c metadata.4 > > 0000000 \0 \0 \0 \0 \0 \0 \0 \0 \0 \0 \0 \0 \0 \0 \0 \0 > > * > > 0001000 > > > > There is no metada in this volume. This will fail every time when we try > to store something > > into this volume. > > > > I wonder how the metadata was deleted, maybe there was some error when > the volume > > was created? > > > > Can you file a bug for this, and attach vdsm.log from the point this > storage > domain was created until the error started? > > > > An important date seems Fri Jun 08 09:51:18 CEST 2018 - at this date the > volume > > metadata was updated in the last time. > > For recovering this volume, you can write valid metadata to the same offset > using dd. > > Basically it is the same as the metadata of the other ovf storage > (metadata.5) > the only changes needed are: > > DESCRIPTION={"Updated":false,"Size":0,"Last Updated":"Fri Jun 08 09:51:18 > CEST 2018","Storage > Domains":[{"uuid":"3ad1987a-8b7d-426d-9d51-4a78cb0a888f"}],"Disk > Description":"OVF_STORE"} > > > IMAGE=0ebefe5e-9053-4bf1-bdfd-fdb26579c179 > > I took the image uuid from the IU_ tag in lvs output. > > Make sure the size of the metadata file is less then 512 bytes, otherwise > you will overwrite and corrupt the metadata of the next volume. > > To write the metadata you should use: > > # makes your metadata file aligned to sector size > truncate -s 512 fixed-metadata.txt > > # write exactly one sector > dd if=fixed-metadata.txt of=/dev/vgname/metadata seek=4 count=1 bs=512 > oflag=direct conv=fsync > > > After this change, engine should be able to use this volume again. > > > > vdsm.log from manual OVF update test: > > > > 2018-06-20 09:28:27,840+0200 INFO (jsonrpc/7) [vdsm.api] START > setVolumeDescription(sdUUID=u'3ad1987a-8b7d-426d-9d51-4a78cb0a888f', > spUUID=u'5849b030-626e-47cb-ad90-3ce782d831b3', > imgUUID=u'0ebefe5e-9053-4bf1-bdfd-fdb26579c179', > volUUID=u'212d644c-0155-4999-9df9-72bacfc7f141', > description=u'{"Updated":false,"Last Updated":"Fri Jun 08 09:51:18 CEST > 2018","Storage > Domains":[{"uuid":"3ad1987a-8b7d-426d-9d51-4a78cb0a888f"}],"Disk > Description":"OVF_STORE"}', options=None) from=::ffff:<IP>,51790, > flow_id=7e4edb74, task_id=5f1fda67-a073-419a-bba5-9bf680c0e5d5 (api:46) > > 2018-06-20 09:28:28,072+0200 WARN (jsonrpc/7) [storage.ResourceManager] > Resource factory failed to create resource > '01_img_3ad1987a-8b7d-426d-9d51-4a78cb0a888f.0ebefe5e-9053-4bf1-bdfd-fdb26579c179'. > Canceling request. (resourceManager:543) > > Traceback (most recent call last): > > File "/usr/lib/python2.7/site-packages/vdsm/storage/resourceManager.py", > line 539, in registerResource > > obj = namespaceObj.factory.createResource(name, lockType) > > File > "/usr/lib/python2.7/site-packages/vdsm/storage/resourceFactories.py", line > 193, in createResource > > lockType) > > File > "/usr/lib/python2.7/site-packages/vdsm/storage/resourceFactories.py", line > 122, in __getResourceCandidatesList > > imgUUID=resourceName) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/image.py", line 206, > in getChain > > if len(uuidlist) == 1 and srcVol.isShared(): > > File "/usr/lib/python2.7/site-packages/vdsm/storage/volume.py", line > 1434, in isShared > > return self._manifest.isShared() > > File "/usr/lib/python2.7/site-packages/vdsm/storage/volume.py", line > 141, in isShared > > return self.getVolType() == sc.type2name(sc.SHARED_VOL) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/volume.py", line > 134, in getVolType > > self.voltype = self.getMetaParam(sc.VOLTYPE) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/volume.py", line > 118, in getMetaParam > > meta = self.getMetadata() > > File "/usr/lib/python2.7/site-packages/vdsm/storage/blockVolume.py", > line 112, in getMetadata > > md = VolumeMetadata.from_lines(lines) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/volumemetadata.py", > line 103, in from_lines > > "Missing metadata key: %s: found: %s" % (e, md)) > > MetaDataKeyNotFoundError: Meta Data key not found error: ("Missing > metadata key: 'DOMAIN': found: {}",) > > 2018-06-20 09:28:28,072+0200 WARN (jsonrpc/7) > [storage.ResourceManager.Request] > (ResName='01_img_3ad1987a-8b7d-426d-9d51-4a78cb0a888f.0ebefe5e-9053-4bf1-bdfd-fdb26579c179', > ReqID='10c95223-f349-4ac3-ab2f-7a5f3d1c7749') Tried to cancel a processed > request (resourceManager:187) > > 2018-06-20 09:28:28,073+0200 INFO (jsonrpc/7) [vdsm.api] FINISH > setVolumeDescription error=Could not acquire resource. Probably resource > factory threw an exception.: () from=::ffff:<IP>,51790, flow_id=7e4edb74, > task_id=5f1fda67-a073-419a-bba5-9bf680c0e5d5 (api:50) > > 2018-06-20 09:28:28,073+0200 ERROR (jsonrpc/7) [storage.TaskManager.Task] > (Task='5f1fda67-a073-419a-bba5-9bf680c0e5d5') Unexpected error (task:875) > > Traceback (most recent call last): > > File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line 882, > in _run > > return fn(*args, **kargs) > > File "<string>", line 2, in setVolumeDescription > > File "/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 48, in > method > > ret = func(*args, **kwargs) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py", line 1294, > in setVolumeDescription > > pool.setVolumeDescription(sdUUID, imgUUID, volUUID, description) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/securable.py", line > 79, in wrapper > > return method(self, *args, **kwargs) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/sp.py", line 2011, > in setVolumeDescription > > with rm.acquireResource(img_ns, imgUUID, rm.EXCLUSIVE): > > File "/usr/lib/python2.7/site-packages/vdsm/storage/resourceManager.py", > line 1025, in acquireResource > > return _manager.acquireResource(namespace, name, lockType, > timeout=timeout) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/resourceManager.py", > line 475, in acquireResource > > raise se.ResourceAcqusitionFailed() > > ResourceAcqusitionFailed: Could not acquire resource. Probably resource > factory threw an exception.: () > > 2018-06-20 09:28:28,073+0200 INFO (jsonrpc/7) [storage.TaskManager.Task] > (Task='5f1fda67-a073-419a-bba5-9bf680c0e5d5') aborting: Task is aborted: > u'Could not acquire resource. Probably resource factory threw an > exception.: ()' - code 100 (task:1181) > > 2018-06-20 09:28:28,073+0200 ERROR (jsonrpc/7) [storage.Dispatcher] FINISH > setVolumeDescription error=Could not acquire resource. Probably resource > factory threw an exception.: () (dispatcher:82) > > > > I worked around the SPM problem by running engine-setup (neither > vds-client Host/StoragePool *Task*, async_task_entries or taskcleaner.sh > showed any tasks). > > > > All the best, > > Oliver > > > > *Von:* Nir Soffer <nsof...@redhat.com> > *Gesendet:* Dienstag, 19. Juni 2018 23:39 > *An:* Elad Ben Aharon <ebena...@redhat.com> > *Cc:* users@ovirt.org; Albl, Oliver <oliver.a...@fabasoft.com> > *Betreff:* [ovirt-users] Re: Failed to update VMs/Templates OVF data, > cannot change SPM > > > > On Wed, Jun 20, 2018 at 12:11 AM Elad Ben Aharon <ebena...@redhat.com> > wrote: > > Can you please attach engine.log and vdsm.log? > > > > On Tue, Jun 19, 2018 at 6:58 PM, Albl, Oliver <oliver.a...@fabasoft.com> > wrote: > > Hi all, > > > > does anybody have an idea how to address this? There is also a side > effect that I cannot change SPM (“Error while executing action: Cannot > force select SPM. The Storage Pool has running tasks.“). > > > > All the best, > > Oliver > > > > *Von:* Albl, Oliver > *Gesendet:* Mittwoch, 13. Juni 2018 12:32 > *An:* users@ovirt.org > *Betreff:* Failed to update VMs/Templates OVF data > > > > Hi, > > > > I have a FC storage domain reporting the following messages every hour: > > > > VDSM command SetVolumeDescriptionVDS failed: Could not acquire resource. > Probably resource factory threw an exception.: () > > Failed to update OVF disks cb04b55c-10fb-46fe-b9de-3c133a94e6a5, OVF data > isn't updated on those OVF stores (Data Center VMTEST, Storage Domain > VMHOST_LUN_62). > > Failed to update VMs/Templates OVF data for Storage Domain VMHOST_LUN_62 > in Data Center VMTEST. > > > > Trying to manually update OVF results in “Error while executing action > UpdateOvfStoreForStorageDomain: Internal Engine Error” > > > > I run oVirt 4.2.3.5-1.el7.centos on CentOS 7.5 (3.10.0-862.3.2.el7.x86_64) > with vdsm-4.20.27.1-1.el7.centos.x86_64 > > > > Engine log: > > > > 2018-06-13 12:15:35,649+02 WARN > [org.ovirt.engine.core.dal.job.ExecutionMessageDirector] (default task-12) > [092d8f27-c0a4-4d78-a8cb-f4738aff71e6] The message key > 'UpdateOvfStoreForStorageDomain' is missing from 'bundles/ExecutionMessages' > > 2018-06-13 12:15:35,655+02 INFO > [org.ovirt.engine.core.bll.storage.domain.UpdateOvfStoreForStorageDomainCommand] > (default task-12) [092d8f27-c0a4-4d78-a8cb-f4738aff71e6] Lock Acquired to > object > 'EngineLock:{exclusiveLocks='[3edb5295-3797-4cd0-9b43-f46ec1ee7b14=OVF_UPDATE, > 373efd46-8aea-4d0e-96cc-1da0debf72d0=STORAGE]', sharedLocks=''}' > > 2018-06-13 12:15:35,660+02 INFO > [org.ovirt.engine.core.bll.storage.domain.UpdateOvfStoreForStorageDomainCommand] > (default task-12) [092d8f27-c0a4-4d78-a8cb-f4738aff71e6] Running command: > UpdateOvfStoreForStorageDomainCommand internal: false. Entities affected : > ID: 373efd46-8aea-4d0e-96cc-1da0debf72d0 Type: StorageAction group > MANIPULATE_STORAGE_DOMAIN with role type ADMIN > > 2018-06-13 12:15:35,670+02 INFO > [org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand] > (default task-12) [4fd5b59a] Running command: > ProcessOvfUpdateForStoragePoolCommand internal: true. Entities affected : > ID: 3edb5295-3797-4cd0-9b43-f46ec1ee7b14 Type: StoragePool > > 2018-06-13 12:15:35,674+02 INFO > [org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand] > (default task-12) [4fd5b59a] Attempting to update VM OVFs in Data Center > 'VMTEST' > > 2018-06-13 12:15:35,678+02 INFO > [org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand] > (default task-12) [4fd5b59a] Successfully updated VM OVFs in Data Center > 'VMTEST' > > 2018-06-13 12:15:35,678+02 INFO > [org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand] > (default task-12) [4fd5b59a] Attempting to update template OVFs in Data > Center 'VMTEST' > > 2018-06-13 12:15:35,678+02 INFO > [org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand] > (default task-12) [4fd5b59a] Successfully updated templates OVFs in Data > Center 'VMTEST' > > 2018-06-13 12:15:35,678+02 INFO > [org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand] > (default task-12) [4fd5b59a] Attempting to remove unneeded template/vm OVFs > in Data Center 'VMTEST' > > 2018-06-13 12:15:35,680+02 INFO > [org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand] > (default task-12) [4fd5b59a] Successfully removed unneeded template/vm OVFs > in Data Center 'VMTEST' > > 2018-06-13 12:15:35,684+02 INFO > [org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStoragePoolCommand] > (default task-12) [4fd5b59a] Lock freed to object > 'EngineLock:{exclusiveLocks='[3edb5295-3797-4cd0-9b43-f46ec1ee7b14=OVF_UPDATE, > 373efd46-8aea-4d0e-96cc-1da0debf72d0=STORAGE]', sharedLocks=''}' > > 2018-06-13 12:15:35,704+02 INFO > [org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStorageDomainCommand] > (default task-12) [24485c23] Lock Acquired to object > 'EngineLock:{exclusiveLocks='[]', > sharedLocks='[3edb5295-3797-4cd0-9b43-f46ec1ee7b14=OVF_UPDATE]'}' > > 2018-06-13 12:15:35,714+02 INFO > [org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStorageDomainCommand] > (default task-12) [24485c23] Running command: > ProcessOvfUpdateForStorageDomainCommand internal: true. Entities affected > : ID: 373efd46-8aea-4d0e-96cc-1da0debf72d0 Type: StorageAction group > MANIPULATE_STORAGE_DOMAIN with role type ADMIN > > 2018-06-13 12:15:35,724+02 INFO > [org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeDescriptionVDSCommand] > (default task-12) [24485c23] START, SetVolumeDescriptionVDSCommand( > SetVolumeDescriptionVDSCommandParameters:{storagePoolId='3edb5295-3797-4cd0-9b43-f46ec1ee7b14', > ignoreFailoverLimit='false', > storageDomainId='373efd46-8aea-4d0e-96cc-1da0debf72d0', > imageGroupId='cb04b55c-10fb-46fe-b9de-3c133a94e6a5', > imageId='a1e7554d-530c-4c07-a4b5-459a1c509e39'}), log id: 747d674f > > 2018-06-13 12:15:35,724+02 INFO > [org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeDescriptionVDSCommand] > (default task-12) [24485c23] -- executeIrsBrokerCommand: calling > 'setVolumeDescription', parameters: > > 2018-06-13 12:15:35,724+02 INFO > [org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeDescriptionVDSCommand] > (default task-12) [24485c23] ++ spUUID=3edb5295-3797-4cd0-9b43-f46ec1ee7b14 > > 2018-06-13 12:15:35,724+02 INFO > [org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeDescriptionVDSCommand] > (default task-12) [24485c23] ++ sdUUID=373efd46-8aea-4d0e-96cc-1da0debf72d0 > > 2018-06-13 12:15:35,724+02 INFO > [org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeDescriptionVDSCommand] > (default task-12) [24485c23] ++ > imageGroupGUID=cb04b55c-10fb-46fe-b9de-3c133a94e6a5 > > 2018-06-13 12:15:35,724+02 INFO > [org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeDescriptionVDSCommand] > (default task-12) [24485c23] ++ volUUID=a1e7554d-530c-4c07-a4b5-459a1c509e39 > > 2018-06-13 12:15:35,724+02 INFO > [org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeDescriptionVDSCommand] > (default task-12) [24485c23] ++ description={"Updated":false,"Last > Updated":"Thu May 24 12:02:22 CEST 2018","Storage > Domains":[{"uuid":"373efd46-8aea-4d0e-96cc-1da0debf72d0"}],"Disk > Description":"OVF_STORE"} > > 2018-06-13 12:15:35,827+02 ERROR > [org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeDescriptionVDSCommand] > (default task-12) [24485c23] Failed in 'SetVolumeDescriptionVDS' method > > 2018-06-13 12:15:35,831+02 ERROR > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] > (default task-12) [24485c23] EVENT_ID: IRS_BROKER_COMMAND_FAILURE(10,803), > VDSM command SetVolumeDescriptionVDS failed: Could not acquire resource. > Probably resource factory threw an exception.: () > > 2018-06-13 12:15:35,831+02 ERROR > [org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeDescriptionVDSCommand] > (default task-12) [24485c23] Command 'SetVolumeDescriptionVDSCommand( > SetVolumeDescriptionVDSCommandParameters:{storagePoolId='3edb5295-3797-4cd0-9b43-f46ec1ee7b14', > ignoreFailoverLimit='false', > storageDomainId='373efd46-8aea-4d0e-96cc-1da0debf72d0', > imageGroupId='cb04b55c-10fb-46fe-b9de-3c133a94e6a5', > imageId='a1e7554d-530c-4c07-a4b5-459a1c509e39'})' execution failed: > IRSGenericException: IRSErrorException: Failed to SetVolumeDescriptionVDS, > error = Could not acquire resource. Probably resource factory threw an > exception.: (), code = 855 > > 2018-06-13 12:15:35,831+02 INFO > [org.ovirt.engine.core.vdsbroker.irsbroker.SetVolumeDescriptionVDSCommand] > (default task-12) [24485c23] FINISH, SetVolumeDescriptionVDSCommand, log > id: 747d674f > > 2018-06-13 12:15:35,831+02 WARN > [org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStorageDomainCommand] > (default task-12) [24485c23] failed to update domain > '373efd46-8aea-4d0e-96cc-1da0debf72d0' ovf store disk > 'cb04b55c-10fb-46fe-b9de-3c133a94e6a5' > > 2018-06-13 12:15:35,834+02 WARN > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] > (default task-12) [24485c23] EVENT_ID: UPDATE_FOR_OVF_STORES_FAILED(1,016), > Failed to update OVF disks cb04b55c-10fb-46fe-b9de-3c133a94e6a5, OVF data > isn't updated on those OVF stores (Data Center VMTEST, Storage Domain > HOST_LUN_62). > > 2018-06-13 12:15:35,843+02 WARN > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] > (default task-12) [24485c23] EVENT_ID: > UPDATE_OVF_FOR_STORAGE_DOMAIN_FAILED(190), Failed to update VMs/Templates > OVF data for Storage Domain VMHOST_LUN_62 in Data Center VMTEST. > > 2018-06-13 12:15:35,846+02 INFO > [org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStorageDomainCommand] > (default task-12) [24485c23] Lock freed to object > 'EngineLock:{exclusiveLocks='[]', > sharedLocks='[3edb5295-3797-4cd0-9b43-f46ec1ee7b14=OVF_UPDATE]'}' > > 2018-06-13 12:15:36,031+02 INFO > [org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback] > (EE-ManagedThreadFactory-engineScheduled-Thread-60) [24485c23] Command > 'ProcessOvfUpdateForStorageDomain' id: > 'a887910e-39a1-4120-a29b-76741ade8bf6' child commands '[]' executions were > completed, status 'SUCCEEDED' > > 2018-06-13 12:15:37,052+02 INFO > [org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStorageDomainCommand] > (EE-ManagedThreadFactory-engineScheduled-Thread-31) [24485c23] Ending > command > 'org.ovirt.engine.core.bll.storage.ovfstore.ProcessOvfUpdateForStorageDomainCommand' > successfully. > > 2018-06-13 12:15:37,059+02 INFO > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] > (EE-ManagedThreadFactory-engineScheduled-Thread-31) [24485c23] EVENT_ID: > USER_UPDATE_OVF_STORE(199), OVF_STORE for domain VMHOST_LUN_62 was updated > by <user>@>domain>@<DOMAIN>-authz. > > > > vdsm.log > > > > 2018-06-13 12:15:35,727+0200 INFO (jsonrpc/7) [vdsm.api] START > setVolumeDescription(sdUUID=u'373efd46-8aea-4d0e-96cc-1da0debf72d0', > spUUID=u'3edb5295-3797-4cd0-9b43-f46ec1ee7b14', > imgUUID=u'cb04b55c-10fb-46fe-b9de-3c133a94e6a5', volUU > ID=u'a1e7554d-530c-4c07-a4b5-459a1c509e39', > description=u'{"Updated":false,"Last Updated":"Thu May 24 12:02:22 CEST > 2018","Storage > Domains":[{"uuid":"373efd46-8aea-4d0e-96cc-1da0debf72d0"}],"Disk > Description":"OVF_STORE"}', options=Non e) from=::ffff:<IP>,54686, > flow_id=24485c23, task_id=70941873-0296-4ed0-94c8-b51290cd6963 (api:46) > > 2018-06-13 12:15:35,825+0200 WARN (jsonrpc/7) [storage.ResourceManager] > Resource factory failed to create resource > '01_img_373efd46-8aea-4d0e-96cc-1da0debf72d0.cb04b55c-10fb-46fe-b9de-3c133a94e6a5'. > Canceling request. (resourceManager :543) > > Traceback (most recent call last): > > File "/usr/lib/python2.7/site-packages/vdsm/storage/resourceManager.py", > line 539, in registerResource > > obj = namespaceObj.factory.createResource(name, lockType) > > File > "/usr/lib/python2.7/site-packages/vdsm/storage/resourceFactories.py", line > 193, in createResource > > lockType) > > File > "/usr/lib/python2.7/site-packages/vdsm/storage/resourceFactories.py", line > 122, in __getResourceCandidatesList > > imgUUID=resourceName) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/image.py", line 206, > in getChain > > if len(uuidlist) == 1 and srcVol.isShared(): > > File "/usr/lib/python2.7/site-packages/vdsm/storage/volume.py", line > 1434, in isShared > > return self._manifest.isShared() > > File "/usr/lib/python2.7/site-packages/vdsm/storage/volume.py", line > 141, in isShared > > return self.getVolType() == sc.type2name(sc.SHARED_VOL) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/volume.py", line > 134, in getVolType > > self.voltype = self.getMetaParam(sc.VOLTYPE) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/volume.py", line > 118, in getMetaParam > > meta = self.getMetadata() > > File "/usr/lib/python2.7/site-packages/vdsm/storage/blockVolume.py", > line 112, in getMetadata > > md = VolumeMetadata.from_lines(lines) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/volumemetadata.py", > line 103, in from_lines > > "Missing metadata key: %s: found: %s" % (e, md)) > > MetaDataKeyNotFoundError: Meta Data key not found error: ("Missing > metadata key: 'DOMAIN': found: {}",) > > > > This means the volume metadata is missing a required key: "DOMAIN". > > > > Can you share the output of: > > > > lvs -o vg_name,lv_name,tags | grep a1e7554d-530c-4c07-a4b5-459a1c509e39 > > > > There should a tag named MD_N - N is an integer. Once you find it, > > we would like to see the metadata at offset N * 512 in the metadata volume. > > > > You can extract it like this: > > > > dd if=/dev/vgname/metadata bs=512 skip=N > > > > Nir > > > > 2018-06-13 12:15:35,825+0200 WARN (jsonrpc/7) > [storage.ResourceManager.Request] > (ResName='01_img_373efd46-8aea-4d0e-96cc-1da0debf72d0.cb04b55c-10fb-46fe-b9de-3c133a94e6a5', > ReqID='dc9ebbc2-5cfa-447d-b2be-40ed2cf81992') Tried to cancel a processed > request (resourceManager:187) > > 2018-06-13 12:15:35,825+0200 INFO (jsonrpc/7) [vdsm.api] FINISH > setVolumeDescription error=Could not acquire resource. Probably resource > factory threw an exception.: () from=::ffff:<IP>,54686, flow_id=24485c23, > task_id=70941873 -0296-4ed0-94c8-b51290cd6963 (api:50) > > 2018-06-13 12:15:35,825+0200 ERROR (jsonrpc/7) [storage.TaskManager.Task] > (Task='70941873-0296-4ed0-94c8-b51290cd6963') Unexpected error (task:875) > > Traceback (most recent call last): > > File "/usr/lib/python2.7/site-packages/vdsm/storage/task.py", line 882, > in _run > > return fn(*args, **kargs) > > File "<string>", line 2, in setVolumeDescription > > File "/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 48, in > method > > ret = func(*args, **kwargs) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py", line 1294, > in setVolumeDescription > > pool.setVolumeDescription(sdUUID, imgUUID, volUUID, description) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/securable.py", line > 79, in wrapper > > return method(self, *args, **kwargs) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/sp.py", line 2011, > in setVolumeDescription > > with rm.acquireResource(img_ns, imgUUID, rm.EXCLUSIVE): > > File "/usr/lib/python2.7/site-packages/vdsm/storage/resourceManager.py", > line 1025, in acquireResource > > return _manager.acquireResource(namespace, name, lockType, > timeout=timeout) > > File "/usr/lib/python2.7/site-packages/vdsm/storage/resourceManager.py", > line 475, in acquireResource > > raise se.ResourceAcqusitionFailed() > > ResourceAcqusitionFailed: Could not acquire resource. Probably resource > factory threw an exception.: () > > 2018-06-13 12:15:35,826+0200 INFO (jsonrpc/7) [storage.TaskManager.Task] > (Task='70941873-0296-4ed0-94c8-b51290cd6963') aborting: Task is aborted: > u'Could not acquire resource. Probably resource factory threw an > exception.: ()' - code 100 (task:1181) > > 2018-06-13 12:15:35,826+0200 ERROR (jsonrpc/7) [storage.Dispatcher] FINISH > setVolumeDescription error=Could not acquire resource. Probably resource > factory threw an exception.: () (dispatcher:82) > > 2018-06-13 12:15:35,826+0200 INFO (jsonrpc/7) [jsonrpc.JsonRpcServer] RPC > call Volume.setDescription failed (error 855) in 0.10 seconds (__init__:573) > > 2018-06-13 12:15:38,953+0200 INFO (jsonrpc/5) [api.host] START > getAllVmStats() from=::ffff:<IP>,54666 (api:46) > > 2018-06-13 12:15:38,956+0200 INFO (jsonrpc/5) [api.host] FINISH > getAllVmStats return={'status': {'message': 'Done', 'code': 0}, > 'statsList': (suppressed)} from=::ffff:<IP>,54666 (api:52) > > 2018-06-13 12:15:38,957+0200 INFO (jsonrpc/5) [jsonrpc.JsonRpcServer] RPC > call Host.getAllVmStats succeeded in 0.01 seconds (__init__:573) > > 2018-06-13 12:15:39,406+0200 INFO (jsonrpc/4) [vdsm.api] START > getSpmStatus(spUUID=u'3edb5295-3797-4cd0-9b43-f46ec1ee7b14', options=None) > from=::ffff:<IP>,54666, task_id=eabfe183-dfb0-4982-b7ea-beacca74aeef > (api:46) > > 2018-06-13 12:15:39,410+0200 INFO (jsonrpc/4) [vdsm.api] FINISH > getSpmStatus return={'spm_st': {'spmId': 2, 'spmStatus': 'SPM', 'spmLver': > 20L}} from=::ffff:<IP>,54666, task_id=eabfe183-dfb0-4982-b7ea-beacca74aeef > (api:52) > > 2018-06-13 12:15:39,410+0200 INFO (jsonrpc/4) [jsonrpc.JsonRpcServer] RPC > call StoragePool.getSpmStatus succeeded in 0.01 seconds (__init__:573) > > 2018-06-13 12:15:39,416+0200 INFO (jsonrpc/1) [vdsm.api] START > getStoragePoolInfo(spUUID=u'3edb5295-3797-4cd0-9b43-f46ec1ee7b14', > options=None) from=::ffff:<IP>,54686, > task_id=b2003a6f-dd74-47ab-b4f0-95ffb54dc51d (api:46) > > 2018-06-13 12:15:39,420+0200 INFO (jsonrpc/1) [vdsm.api] FINISH > getStoragePoolInfo return={'info': {'name': 'No Description', 'isoprefix': > '', 'pool_status': 'connected', 'lver': 20L, 'domains': > u'373efd46-8aea-4d0e-96cc-1da0debf72d0: > Active,9ed4b0d2-c555-4b27-8f88-75c47a99ad98:Attached,efd78d36-c395-4e9a-a46e-6059fa53756d:Active,3675435e-851e-4236-81da-fce1cc027238:Active', > 'master_uuid': 'efd78d36-c395-4e9a-a46e-6059fa53756d', 'version': '4', > 'spm_id': 2, 'type': 'FCP', 'master_ver': 12}, 'dominfo': > {u'373efd46-8aea-4d0e-96cc-1da0debf72d0': {'status': u'Active', 'diskfree': > '8722541707264', 'isoprefix': '', 'alerts': [], 'disktotal': > '8795690369024', 'version': 4}, u'9ed4b0d2-c555-4b27-8f88-75c 47a99ad98': > {'status': u'Attached', 'isoprefix': '', 'alerts': []}, > u'efd78d36-c395-4e9a-a46e-6059fa53756d': {'status': u'Active', 'diskfree': > '8718783610880', 'isoprefix': '', 'alerts': [], 'disktotal': > '8795690369024', 'version': 4}, u'3675435e-851e-4236-81da-fce1cc027238': > {'status': u'Active', 'diskfree': '8713280684032', 'isoprefix': '', > 'alerts': [], 'disktotal': '8795690369024', 'version': 4}}} > from=::ffff:<IP>,54686, task_id=b2003a6f-dd74-47ab-b4f0-95 ffb54dc51d > (api:52) > > 2018-06-13 12:15:39,421+0200 INFO (jsonrpc/1) [jsonrpc.JsonRpcServer] RPC > call StoragePool.getInfo succeeded in 0.01 seconds (__init__:573) > > > > All the best, > > Oliver > > > > > _______________________________________________ > Users mailing list -- users@ovirt.org > To unsubscribe send an email to users-le...@ovirt.org > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/users@ovirt.org/message/MRBMUKOZ7VTP52I2SRPMBMCW375CXS5A/ > > > > _______________________________________________ > Users mailing list -- users@ovirt.org > To unsubscribe send an email to users-le...@ovirt.org > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/users@ovirt.org/message/6XG4IFPYXAIDX5IJRBD3SS4FARC74ZCC/ > >
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/QKKNSRWWP2A3SBQISEBQ4KCEOIWG3FI2/