On Tue, Oct 24, 2017 at 8:29 AM, Idan Shaby <[email protected]> wrote:
> Hi Gianluca, > > Can you try to reproduce it with FC? Maybe it's somehow related to that. > Just to make it clear - I know that bug originally occurred on FC, and it will be helpful to have the full logs if we want to understand the root cause for this. > Did you try to lvremove the old lv's, or are they still in use? > Trying to do this from the SPM should be ok. What do you mean synchronize > the information at the other node side? > In any case, this is a disk that you don't need anymore, isn't it? You > said that the copy part of the move operation went well. > > > Thanks, > Idan > > On Thu, Oct 19, 2017 at 10:47 AM, Gianluca Cecchi < > [email protected]> wrote: > >> On Thu, Oct 19, 2017 at 7:36 AM, Idan Shaby <[email protected]> wrote: >> >>> Hi Gianluca, >>> >>> This sounds like a bug to me. >>> Can you please attach engine and vdsm logs? >>> >>> >>> Thanks, >>> Idan >>> >> >> Hi Idan, >> I have not at hand the full files right now, but these were the snippets >> I cut from vdsm log and engine.log yesterday around the time of the first >> error after sending the first e-mail >> It seems vdsm reports that the logical volume is still in use while >> trying to remove it.... >> I have tried to replicate on another environment also based on block >> storage, but now iSCSI and not FC and I was not able to get the error: the >> move disk operation of the powered off vm always completed successfully. >> Both tests were done trying to move preallocated disks if that matters >> >> vdsm.log of host1 >> >> 2017-10-18 14:59:28,907+0200 ERROR (libvirt/events) >> [storage.TaskManager.Task] (Task='6331ec00-b75b-465e-8d9c-04d465c09ea1') >> Unexpected error (task:872) >> Traceback (most recent call last): >> File "/usr/share/vdsm/storage/task.py", line 879, in _run >> return fn(*args, **kargs) >> File "<string>", line 2, in teardownImage >> File "/usr/lib/python2.7/site-packages/vdsm/common/api.py", line 48, >> in method >> ret = func(*args, **kwargs) >> File "/usr/share/vdsm/storage/hsm.py", line 3189, in teardownImage >> dom.deactivateImage(imgUUID) >> File "/usr/share/vdsm/storage/blockSD.py", line 1289, in >> deactivateImage >> lvm.deactivateLVs(self.sdUUID, volUUIDs) >> File "/usr/share/vdsm/storage/lvm.py", line 1304, in deactivateLVs >> _setLVAvailability(vgName, toDeactivate, "n") >> File "/usr/share/vdsm/storage/lvm.py", line 843, in _setLVAvailability >> raise error(str(e)) >> CannotDeactivateLogicalVolume: Cannot deactivate Logical Volume: >> ('General Storage Exception: ("5 [] [\' Logical volume >> c5d360f5-cd90-4b3e-bdef-924649bfbc5f/eff58f10-12bf-412f-a5db-944bfb10904d >> is used by another device.\']\\nc5d360f5-cd90-4b3 >> e-bdef-924649bfbc5f/[\'eff58f10-12bf-412f-a5db-944bfb10904d\']",)',) >> 2017-10-18 14:59:28,907+0200 INFO (libvirt/events) >> [storage.TaskManager.Task] (Task='6331ec00-b75b-465e-8d9c-04d465c09ea1') >> aborting: Task is aborted: 'Cannot deactivate Logical Volume: (\'General >> Storage Exception: ("5 [] [\\\' Logical volume >> c5d360f5-cd90-4b3e-bdef-924649bfbc5f/eff58f10-12bf-412f-a5db-944bfb10904d >> is used by another device.\\\']\\\\nc5d360f5-cd90 >> -4b3e-bdef-924649bfbc5f/[\\\'eff58f10-12bf-412f-a5db-944bfb10904d\\\']",)\',)' >> - code 552 (task:1177) >> 2017-10-18 14:59:28,908+0200 ERROR (libvirt/events) [storage.Dispatcher] >> FINISH teardownImage error=Cannot deactivate Logical Volume: ('General >> Storage Exception: ("5 [] [\' Logical volume c5d360f5-cd90-4b3e-bdef-924649 >> bfbc5f/eff58f10-12bf-412f-a5db-944bfb10904d is used by another >> device.\']\\nc5d360f5-cd90-4b3e-bdef-924649bfbc5f/[\'eff58f1 >> 0-12bf-412f-a5db-944bfb10904d\']",)',) (dispatcher:81) >> >> engine.log >> 2017-10-18 15:01:53,097+02 INFO [org.ovirt.engine.core.dal.dbb >> roker.auditloghandling.AuditLogDirector] (D >> efaultQuartzScheduler6) [] EVENT_ID: USER_MOVED_DISK_FINISHED_SUCCESS(2,010), >> Correlation ID: a8948cbf-80a >> 2-462d-8805-91da94868c6e, Job ID: ffba7504-a324-43ca-ab4c-481610c1098b, >> Call Stack: null, Custom ID: null, >> Custom Event ID: -1, Message: User gcecchi@internal-authz finished >> moving disk padora1_u01 to domain SATA >> _ALLBCK. >> 2017-10-18 15:01:54,029+02 ERROR [org.ovirt.engine.core.dal.dbb >> roker.auditloghandling.AuditLogDirector] (DefaultQuartzScheduler5) >> [394f8568] EVENT_ID: VDS_BROKER_COMMAND_FAILURE(10,802), Correlation ID: >> null, Call Stack: null, Custom ID: null, Custom Event ID: -1, Message: VDSM >> host1 command HSMGetAllTasksStatusesVDS failed: Cannot remove Logical >> Volume: (['Cannot remove Logical Volume: >> (u\'c5d360f5-cd90-4b3e-bdef-924649bfbc5f\', >> "[\'2255c544-165b-4126-86ee-70e6d1edef9a\']")'],) >> 2017-10-18 15:01:54,029+02 INFO >> [org.ovirt.engine.core.bll.tasks.SPMAsyncTask] >> (DefaultQuartzScheduler5) [394f8568] SPMAsyncTask::PollTask: Polling task >> '8446a883-9bb0-40a5-b1eb-8a919defe4c0' (Parent Command 'RemoveImage', >> Parameters Type 'org.ovirt.engine.core.common. >> asynctasks.AsyncTaskParameters') returned status 'finished', result >> 'cleanSuccess'. >> 2017-10-18 15:01:54,029+02 ERROR >> [org.ovirt.engine.core.bll.tasks.SPMAsyncTask] >> (DefaultQuartzScheduler5) [394f8568] BaseAsyncTask::logEndTaskFailure: >> Task '8446a883-9bb0-40a5-b1eb-8a919defe4c0' (Parent Command >> 'RemoveImage', Parameters Type 'org.ovirt.engine.core.common. >> asynctasks.AsyncTaskParameters') ended with failure: >> -- Result: 'cleanSuccess' >> -- Message: 'VDSGenericException: VDSErrorException: Failed in vdscommand >> to HSMGetAllTasksStatusesVDS, error = Cannot remove Logical Volume: >> (['Cannot remove Logical Volume: (u\'c5d360f5-cd90-4b3e-bdef-924649bfbc5f\', >> "[\'2255c544-165b-4126-86ee-70e6d1edef9a\']")'],)', >> -- Exception: 'VDSGenericException: VDSErrorException: Failed in >> vdscommand to HSMGetAllTasksStatusesVDS, error = Cannot remove Logical >> Volume: (['Cannot remove Logical Volume: >> (u\'c5d360f5-cd90-4b3e-bdef-924649bfbc5f\', >> "[\'2255c544-165b-4126-86ee-70e6d1edef9a\']")'],)' >> >> Is there any command I can run to consistently remove the stale LVs? >> As oVirt doesn't use clvmd I don't think it is sufficient to issue the >> remove lv from the SPM node.. How will "synchronize" the information at the >> other node side? >> >> Thanks, >> Gianluca >> >> >
_______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

