[ovirt-users] Re: strange issue: vm lost info on disk
.engine.core.dal.dbb >>>>>>>>>>> roker.auditloghandling.AuditLogDirector] (default task-50) >>>>>>>>>>> [940b051c-8c63-4711-baf9-f3520bb2b825] EVENT_ID: >>>>>>>>>>> USER_MOVED_DISK(2,008), User admin@internal-authz moving disk >>>>>>>>>>> mail02-int_Disk1 to domain 2penLA. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Then, in the last attempt I see the attempt was successful but >>>>>>>>>>> live merge failed: >>>>>>>>>>> 2018-05-11 03:37:59,509-03 ERROR >>>>>>>>>>> [org.ovirt.engine.core.bll.MergeStatusCommand] >>>>>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) >>>>>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Failed to live merge, >>>>>>>>>>> still in volume chain: [5d9d2958-96bc-49fa-9100-2f33a3ba737f, >>>>>>>>>>> 52532d05-970e-4643-9774-96c31796062c] >>>>>>>>>>> 2018-05-11 03:38:01,495-03 INFO [org.ovirt.engine.core.bll.Ser >>>>>>>>>>> ialChildCommandsExecutionCallback] >>>>>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-51) >>>>>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command >>>>>>>>>>> 'LiveMigrateDisk' (id: '115fc375-6018-4d59-b9f2-51ee05ca49f8') >>>>>>>>>>> waiting on child command id: '26bc52a4-4509-4577-b342-44a679bc628f' >>>>>>>>>>> type:'RemoveSnapshot' to complete >>>>>>>>>>> 2018-05-11 03:38:01,501-03 ERROR [org.ovirt.engine.core.bll.sna >>>>>>>>>>> pshots.RemoveSnapshotSingleDiskLiveCommand] >>>>>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-51) >>>>>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command id: >>>>>>>>>>> '4936d196-a891-4484-9cf5-fceaafbf3364 failed child command >>>>>>>>>>> status for step 'MERGE_STATUS' >>>>>>>>>>> 2018-05-11 03:38:01,501-03 INFO [org.ovirt.engine.core.bll.sna >>>>>>>>>>> pshots.RemoveSnapshotSingleDiskLiveCommandCallback] >>>>>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-51) >>>>>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command >>>>>>>>>>> 'RemoveSnapshotSingleDiskLive' id: >>>>>>>>>>> '4936d196-a891-4484-9cf5-fceaafbf3364' >>>>>>>>>>> child commands '[8da5f261-7edd-4930-8d9d-d34f232d84b3, >>>>>>>>>>> 1c320f4b-7296-43c4-a3e6-8a868e23fc35, >>>>>>>>>>> a0e9e70c-cd65-4dfb-bd00-076c4e99556a]' executions were >>>>>>>>>>> completed, status 'FAILED' >>>>>>>>>>> 2018-05-11 03:38:02,513-03 ERROR [org.ovirt.engine.core.bll.sna >>>>>>>>>>> pshots.RemoveSnapshotSingleDiskLiveCommand] >>>>>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-2) >>>>>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Merging of snapshot >>>>>>>>>>> '319e8bbb-9efe-4de4-a9a6-862e3deb891f' images >>>>>>>>>>> '52532d05-970e-4643-9774-96c31796062c'..'5d9d2958-96bc-49fa-9100-2f33a3ba737f' >>>>>>>>>>> failed. Images have been marked illegal and can no longer be >>>>>>>>>>> previewed or >>>>>>>>>>> reverted to. Please retry Live Merge on the snapshot to complete the >>>>>>>>>>> operation. >>>>>>>>>>> 2018-05-11 03:38:02,519-03 ERROR [org.ovirt.engine.core.bll.sna >>>>>>>>>>> pshots.RemoveSnapshotSingleDiskLiveCommand] >>>>>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-2) >>>>>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command >>>>>>>>>>> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand' >>>>>>>>>>> with failure. >>>>>>>>>>> 2018-05-11 03:38:03,530-03 IN
[ovirt-users] Re: strange issue: vm lost info on disk
342-44a679bc628f' >>>>>>>>>> type:'RemoveSnapshot' to complete >>>>>>>>>> 2018-05-11 03:38:01,501-03 ERROR [org.ovirt.engine.core.bll.sna >>>>>>>>>> pshots.RemoveSnapshotSingleDiskLiveCommand] >>>>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-51) >>>>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command id: >>>>>>>>>> '4936d196-a891-4484-9cf5-fceaafbf3364 failed child command >>>>>>>>>> status for step 'MERGE_STATUS' >>>>>>>>>> 2018-05-11 03:38:01,501-03 INFO [org.ovirt.engine.core.bll.sna >>>>>>>>>> pshots.RemoveSnapshotSingleDiskLiveCommandCallback] >>>>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-51) >>>>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command >>>>>>>>>> 'RemoveSnapshotSingleDiskLive' id: >>>>>>>>>> '4936d196-a891-4484-9cf5-fceaafbf3364' >>>>>>>>>> child commands '[8da5f261-7edd-4930-8d9d-d34f232d84b3, >>>>>>>>>> 1c320f4b-7296-43c4-a3e6-8a868e23fc35, >>>>>>>>>> a0e9e70c-cd65-4dfb-bd00-076c4e99556a]' executions were >>>>>>>>>> completed, status 'FAILED' >>>>>>>>>> 2018-05-11 03:38:02,513-03 ERROR [org.ovirt.engine.core.bll.sna >>>>>>>>>> pshots.RemoveSnapshotSingleDiskLiveCommand] >>>>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-2) >>>>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Merging of snapshot >>>>>>>>>> '319e8bbb-9efe-4de4-a9a6-862e3deb891f' images >>>>>>>>>> '52532d05-970e-4643-9774-96c31796062c'..'5d9d2958-96bc-49fa-9100-2f33a3ba737f' >>>>>>>>>> failed. Images have been marked illegal and can no longer be >>>>>>>>>> previewed or >>>>>>>>>> reverted to. Please retry Live Merge on the snapshot to complete the >>>>>>>>>> operation. >>>>>>>>>> 2018-05-11 03:38:02,519-03 ERROR [org.ovirt.engine.core.bll.sna >>>>>>>>>> pshots.RemoveSnapshotSingleDiskLiveCommand] >>>>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-2) >>>>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command >>>>>>>>>> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand' >>>>>>>>>> with failure. >>>>>>>>>> 2018-05-11 03:38:03,530-03 INFO [org.ovirt.engine.core.bll.Con >>>>>>>>>> currentChildCommandsExecutionCallback] >>>>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-37) >>>>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command 'RemoveSnapshot' >>>>>>>>>> id: '26bc52a4-4509-4577-b342-44a679bc628f' child commands >>>>>>>>>> '[4936d196-a891-4484-9cf5-fceaafbf3364]' executions were >>>>>>>>>> completed, status 'FAILED' >>>>>>>>>> 2018-05-11 03:38:04,548-03 ERROR >>>>>>>>>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand] >>>>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-66) >>>>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command >>>>>>>>>> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand' with >>>>>>>>>> failure. >>>>>>>>>> 2018-05-11 03:38:04,557-03 INFO >>>>>>>>>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand] >>>>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-66) >>>>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Lock freed to object >>>>>>>>>> 'EngineLock:{exclusiveLocks='[4808bb70-c9cc-4286-aa39-16b579 >>>>>>>>>> 8213ac=LIVE_STORAGE_MIGRATION]', sharedLocks=''}' >>>>>>>>>> >>>>>>>>>> I do not see the merge attempt in the vdsm.log, so please send >>>>>>>>>> vdsm logs for node02.phy.eze.ampgn.com.ar from that time. >>>>>>>>>> >>>>>>>>>> Al
[ovirt-users] Re: strange issue: vm lost info on disk
;>> 2018-05-11 03:38:02,513-03 ERROR [org.ovirt.engine.core.bll. >>>>>>>>> snapshots.RemoveSnapshotSingleDiskLiveCommand] >>>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-2) >>>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Merging of snapshot >>>>>>>>> '319e8bbb-9efe-4de4-a9a6-862e3deb891f' images >>>>>>>>> '52532d05-970e-4643-9774-96c31796062c'..'5d9d2958-96bc-49fa-9100-2f33a3ba737f' >>>>>>>>> failed. Images have been marked illegal and can no longer be >>>>>>>>> previewed or >>>>>>>>> reverted to. Please retry Live Merge on the snapshot to complete the >>>>>>>>> operation. >>>>>>>>> 2018-05-11 03:38:02,519-03 ERROR [org.ovirt.engine.core.bll. >>>>>>>>> snapshots.RemoveSnapshotSingleDiskLiveCommand] >>>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-2) >>>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command >>>>>>>>> 'org.ovirt.engine.core.bll.snapshots. >>>>>>>>> RemoveSnapshotSingleDiskLiveCommand' with failure. >>>>>>>>> 2018-05-11 03:38:03,530-03 INFO [org.ovirt.engine.core.bll. >>>>>>>>> ConcurrentChildCommandsExecutionCallback] >>>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-37) >>>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command 'RemoveSnapshot' >>>>>>>>> id: '26bc52a4-4509-4577-b342-44a679bc628f' child commands >>>>>>>>> '[4936d196-a891-4484-9cf5-fceaafbf3364]' executions were >>>>>>>>> completed, status 'FAILED' >>>>>>>>> 2018-05-11 03:38:04,548-03 ERROR [org.ovirt.engine.core.bll. >>>>>>>>> snapshots.RemoveSnapshotCommand] >>>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-66) >>>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command >>>>>>>>> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand' with >>>>>>>>> failure. >>>>>>>>> 2018-05-11 03:38:04,557-03 INFO [org.ovirt.engine.core.bll. >>>>>>>>> snapshots.RemoveSnapshotCommand] >>>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-66) >>>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Lock freed to object >>>>>>>>> 'EngineLock:{exclusiveLocks='[4808bb70-c9cc-4286-aa39- >>>>>>>>> 16b5798213ac=LIVE_STORAGE_MIGRATION]', sharedLocks=''}' >>>>>>>>> >>>>>>>>> I do not see the merge attempt in the vdsm.log, so please send >>>>>>>>> vdsm logs for node02.phy.eze.ampgn.com.ar from that time. >>>>>>>>> >>>>>>>>> Also, did you use the auto-generated snapshot to start the vm? >>>>>>>>> >>>>>>>>> >>>>>>>>> On Fri, May 11, 2018 at 6:11 PM, Juan Pablo < >>>>>>>>> pablo.localh...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> after the xfs_repair, it says: sorry I could not find valid >>>>>>>>>> secondary superblock >>>>>>>>>> >>>>>>>>>> 2018-05-11 12:09 GMT-03:00 Juan Pablo <pablo.localh...@gmail.com> >>>>>>>>>> : >>>>>>>>>> >>>>>>>>>>> hi, >>>>>>>>>>> Alias: >>>>>>>>>>> mail02-int_Disk1 >>>>>>>>>>> Description: >>>>>>>>>>> ID: >>>>>>>>>>> 65ec515e-0aae-4fe6-a561-387929c7fb4d >>>>>>>>>>> Alignment: >>>>>>>>>>> Unknown >>>>>>>>>>> Disk Profile: >>>>>>>>>>> Wipe After Delete: >>>>>>>>>>> No >>>>>>>>>>> >>>>>>>>>>> that one >>>>>>>>>>> >>>>>>>>>>> 2018-05-11 11:12 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>: >>>>>>>>>>> >>>>>>>>>>>> I looked at the logs and I see some disks have moved >>>>>>>>>>>> successfully and some failed. Which disk is causing the problems? >>>>>>>>>>>> >>>>>>>>>>>> On Fri, May 11, 2018 at 5:02 PM, Juan Pablo < >>>>>>>>>>>> pablo.localh...@gmail.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Hi, just sent you via drive the files. attaching some extra >>>>>>>>>>>>> info, thanks thanks and thanks : >>>>>>>>>>>>> >>>>>>>>>>>>> from inside the migrated vm I had the following attached dmesg >>>>>>>>>>>>> output before rebooting >>>>>>>>>>>>> >>>>>>>>>>>>> regards and thanks again for the help, >>>>>>>>>>>>> >>>>>>>>>>>>> 2018-05-11 10:45 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com> >>>>>>>>>>>>> : >>>>>>>>>>>>> >>>>>>>>>>>>>> Dropbox or google drive I guess. Also, can you attach >>>>>>>>>>>>>> engine.log? >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Fri, May 11, 2018 at 4:43 PM, Juan Pablo < >>>>>>>>>>>>>> pablo.localh...@gmail.com> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> vdsm is too big for gmail ...any other way I can share it >>>>>>>>>>>>>>> with you? >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -- Forwrded message -- >>>>>>>>>>>>>>> From: Juan Pablo <pablo.localh...@gmail.com> >>>>>>>>>>>>>>> Date: 2018-05-11 10:40 GMT-03:00 >>>>>>>>>>>>>>> Subject: Re: [ovirt-users] strange issue: vm lost info on >>>>>>>>>>>>>>> disk >>>>>>>>>>>>>>> To: Benny Zlotnik <bzlot...@redhat.com> >>>>>>>>>>>>>>> Cc: users <Users@ovirt.org> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Benny, thanks for your reply! it was a Live migration. >>>>>>>>>>>>>>> sorry, it was from nfs to iscsi, not otherwise. I have reboot >>>>>>>>>>>>>>> the vm for >>>>>>>>>>>>>>> rescue and it does not detect any partitions with fdisk, Im >>>>>>>>>>>>>>> running a >>>>>>>>>>>>>>> xfs_repair with -n and found some corrupted primary >>>>>>>>>>>>>>> superblock., its still >>>>>>>>>>>>>>> running... ( so... there's info in the disk maybe?) >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> attaching logs, let me know if those are the ones. >>>>>>>>>>>>>>> thanks again! >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> 2018-05-11 9:45 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com >>>>>>>>>>>>>>> >: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Can you provide the logs? engine and vdsm. >>>>>>>>>>>>>>>> Did you perform a live migration (the VM is running) or >>>>>>>>>>>>>>>> cold? >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On Fri, May 11, 2018 at 2:49 PM, Juan Pablo < >>>>>>>>>>>>>>>> pablo.localh...@gmail.com> wrote: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Hi! , Im strugled about an ongoing problem: >>>>>>>>>>>>>>>>> after migrating a vm's disk from an iscsi domain to a nfs >>>>>>>>>>>>>>>>> and ovirt reporting the migration was successful, I see >>>>>>>>>>>>>>>>> there's no data >>>>>>>>>>>>>>>>> 'inside' the vm's disk. we never had this issues with ovirt >>>>>>>>>>>>>>>>> so Im stranged >>>>>>>>>>>>>>>>> about the root cause and if theres a chance of recovering the >>>>>>>>>>>>>>>>> information. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> can you please help me out troubleshooting this one? I >>>>>>>>>>>>>>>>> would really appreciate it =) >>>>>>>>>>>>>>>>> running ovirt 4.2.1 here! >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> thanks in advance, >>>>>>>>>>>>>>>>> JP >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> ___ >>>>>>>>>>>>>>>>> Users mailing list -- users@ovirt.org >>>>>>>>>>>>>>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> ___ >>>>>>> Users mailing list -- users@ovirt.org >>>>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>>>> >>>>>> >>>>> >>>> >>> >> ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org
[ovirt-users] Re: strange issue: vm lost info on disk
r admin@internal-authz moving disk >>>>>>>> mail02-int_Disk1 to domain 2penLA. >>>>>>>> >>>>>>>> >>>>>>>> Then, in the last attempt I see the attempt was successful but live >>>>>>>> merge failed: >>>>>>>> 2018-05-11 03:37:59,509-03 ERROR >>>>>>>> [org.ovirt.engine.core.bll.MergeStatusCommand] >>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2) >>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Failed to live merge, still in >>>>>>>> volume chain: [5d9d2958-96bc-49fa-9100-2f33a3ba737f, >>>>>>>> 52532d05-970e-4643-9774-96c31796062c] >>>>>>>> 2018-05-11 03:38:01,495-03 INFO >>>>>>>> [org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback] >>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-51) >>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command 'LiveMigrateDisk' (id: >>>>>>>> '115fc375-6018-4d59-b9f2-51ee05ca49f8') waiting on child command id: >>>>>>>> '26bc52a4-4509-4577-b342-44a679bc628f' type:'RemoveSnapshot' to >>>>>>>> complete >>>>>>>> 2018-05-11 03:38:01,501-03 ERROR >>>>>>>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand] >>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-51) >>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command id: >>>>>>>> '4936d196-a891-4484-9cf5-fceaafbf3364 failed child command status for >>>>>>>> step >>>>>>>> 'MERGE_STATUS' >>>>>>>> 2018-05-11 03:38:01,501-03 INFO >>>>>>>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommandCallback] >>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-51) >>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command >>>>>>>> 'RemoveSnapshotSingleDiskLive' id: >>>>>>>> '4936d196-a891-4484-9cf5-fceaafbf3364' >>>>>>>> child commands '[8da5f261-7edd-4930-8d9d-d34f232d84b3, >>>>>>>> 1c320f4b-7296-43c4-a3e6-8a868e23fc35, >>>>>>>> a0e9e70c-cd65-4dfb-bd00-076c4e99556a]' executions were completed, >>>>>>>> status >>>>>>>> 'FAILED' >>>>>>>> 2018-05-11 03:38:02,513-03 ERROR >>>>>>>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand] >>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-2) >>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Merging of snapshot >>>>>>>> '319e8bbb-9efe-4de4-a9a6-862e3deb891f' images >>>>>>>> '52532d05-970e-4643-9774-96c31796062c'..'5d9d2958-96bc-49fa-9100-2f33a3ba737f' >>>>>>>> failed. Images have been marked illegal and can no longer be previewed >>>>>>>> or >>>>>>>> reverted to. Please retry Live Merge on the snapshot to complete the >>>>>>>> operation. >>>>>>>> 2018-05-11 03:38:02,519-03 ERROR >>>>>>>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand] >>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-2) >>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command >>>>>>>> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand' >>>>>>>> with failure. >>>>>>>> 2018-05-11 03:38:03,530-03 INFO >>>>>>>> [org.ovirt.engine.core.bll.ConcurrentChildCommandsExecutionCallback] >>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-37) >>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command 'RemoveSnapshot' id: >>>>>>>> '26bc52a4-4509-4577-b342-44a679bc628f' child commands >>>>>>>> '[4936d196-a891-4484-9cf5-fceaafbf3364]' executions were completed, >>>>>>>> status >>>>>>>> 'FAILED' >>>>>>>> 2018-05-11 03:38:04,548-03 ERROR >>>>>>>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand] >>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-66) >>>>>>>>
[ovirt-users] Re: strange issue: vm lost info on disk
gt;>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-51) >>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command 'LiveMigrateDisk' >>>>>>> (id: '115fc375-6018-4d59-b9f2-51ee05ca49f8') waiting on child >>>>>>> command id: '26bc52a4-4509-4577-b342-44a679bc628f' >>>>>>> type:'RemoveSnapshot' to complete >>>>>>> 2018-05-11 03:38:01,501-03 ERROR [org.ovirt.engine.core.bll.sna >>>>>>> pshots.RemoveSnapshotSingleDiskLiveCommand] >>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-51) >>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command id: >>>>>>> '4936d196-a891-4484-9cf5-fceaafbf3364 failed child command status >>>>>>> for step 'MERGE_STATUS' >>>>>>> 2018-05-11 03:38:01,501-03 INFO [org.ovirt.engine.core.bll.sna >>>>>>> pshots.RemoveSnapshotSingleDiskLiveCommandCallback] >>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-51) >>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command >>>>>>> 'RemoveSnapshotSingleDiskLive' id: >>>>>>> '4936d196-a891-4484-9cf5-fceaafbf3364' >>>>>>> child commands '[8da5f261-7edd-4930-8d9d-d34f232d84b3, >>>>>>> 1c320f4b-7296-43c4-a3e6-8a868e23fc35, >>>>>>> a0e9e70c-cd65-4dfb-bd00-076c4e99556a]' >>>>>>> executions were completed, status 'FAILED' >>>>>>> 2018-05-11 03:38:02,513-03 ERROR [org.ovirt.engine.core.bll.sna >>>>>>> pshots.RemoveSnapshotSingleDiskLiveCommand] >>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-2) >>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Merging of snapshot >>>>>>> '319e8bbb-9efe-4de4-a9a6-862e3deb891f' images >>>>>>> '52532d05-970e-4643-9774-96c31796062c'..'5d9d2958-96bc-49fa-9100-2f33a3ba737f' >>>>>>> failed. Images have been marked illegal and can no longer be previewed >>>>>>> or >>>>>>> reverted to. Please retry Live Merge on the snapshot to complete the >>>>>>> operation. >>>>>>> 2018-05-11 03:38:02,519-03 ERROR [org.ovirt.engine.core.bll.sna >>>>>>> pshots.RemoveSnapshotSingleDiskLiveCommand] >>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-2) >>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command >>>>>>> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand' >>>>>>> with failure. >>>>>>> 2018-05-11 03:38:03,530-03 INFO [org.ovirt.engine.core.bll.Con >>>>>>> currentChildCommandsExecutionCallback] >>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-37) >>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command 'RemoveSnapshot' id: >>>>>>> '26bc52a4-4509-4577-b342-44a679bc628f' child commands >>>>>>> '[4936d196-a891-4484-9cf5-fceaafbf3364]' executions were completed, >>>>>>> status 'FAILED' >>>>>>> 2018-05-11 03:38:04,548-03 ERROR >>>>>>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand] >>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-66) >>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command >>>>>>> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand' with >>>>>>> failure. >>>>>>> 2018-05-11 03:38:04,557-03 INFO >>>>>>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand] >>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-66) >>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Lock freed to object >>>>>>> 'EngineLock:{exclusiveLocks='[4808bb70-c9cc-4286-aa39-16b579 >>>>>>> 8213ac=LIVE_STORAGE_MIGRATION]', sharedLocks=''}' >>>>>>> >>>>>>> I do not see the merge attempt in the vdsm.log, so please send vdsm >>>>>>> logs for node02.phy.eze.ampgn.com.ar from that time. >>>>>>> >>>>>>> Also, did you use the auto-generated snapshot to start the vm? >>>>>>> >>>>>>> >>>>>>> On Fri, May 11, 2018 at 6:11 PM, Juan Pablo < >>>>>>> pablo.localh...@gmail.com> wrote: >>>>>>> >>&
[ovirt-users] Re: strange issue: vm lost info on disk
>>> for step 'MERGE_STATUS' >>>>>> 2018-05-11 03:38:01,501-03 INFO [org.ovirt.engine.core.bll.sna >>>>>> pshots.RemoveSnapshotSingleDiskLiveCommandCallback] >>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-51) >>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command >>>>>> 'RemoveSnapshotSingleDiskLive' id: '4936d196-a891-4484-9cf5-fceaafbf3364' >>>>>> child commands '[8da5f261-7edd-4930-8d9d-d34f232d84b3, >>>>>> 1c320f4b-7296-43c4-a3e6-8a868e23fc35, >>>>>> a0e9e70c-cd65-4dfb-bd00-076c4e99556a]' >>>>>> executions were completed, status 'FAILED' >>>>>> 2018-05-11 03:38:02,513-03 ERROR [org.ovirt.engine.core.bll.sna >>>>>> pshots.RemoveSnapshotSingleDiskLiveCommand] >>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-2) >>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Merging of snapshot >>>>>> '319e8bbb-9efe-4de4-a9a6-862e3deb891f' images >>>>>> '52532d05-970e-4643-9774-96c31796062c'..'5d9d2958-96bc-49fa-9100-2f33a3ba737f' >>>>>> failed. Images have been marked illegal and can no longer be previewed or >>>>>> reverted to. Please retry Live Merge on the snapshot to complete the >>>>>> operation. >>>>>> 2018-05-11 03:38:02,519-03 ERROR [org.ovirt.engine.core.bll.sna >>>>>> pshots.RemoveSnapshotSingleDiskLiveCommand] >>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-2) >>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command >>>>>> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand' >>>>>> with failure. >>>>>> 2018-05-11 03:38:03,530-03 INFO [org.ovirt.engine.core.bll.Con >>>>>> currentChildCommandsExecutionCallback] >>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-37) >>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command 'RemoveSnapshot' id: >>>>>> '26bc52a4-4509-4577-b342-44a679bc628f' child commands >>>>>> '[4936d196-a891-4484-9cf5-fceaafbf3364]' executions were completed, >>>>>> status 'FAILED' >>>>>> 2018-05-11 03:38:04,548-03 ERROR >>>>>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand] >>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-66) >>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command >>>>>> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand' with >>>>>> failure. >>>>>> 2018-05-11 03:38:04,557-03 INFO >>>>>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand] >>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-66) >>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Lock freed to object >>>>>> 'EngineLock:{exclusiveLocks='[4808bb70-c9cc-4286-aa39-16b579 >>>>>> 8213ac=LIVE_STORAGE_MIGRATION]', sharedLocks=''}' >>>>>> >>>>>> I do not see the merge attempt in the vdsm.log, so please send vdsm >>>>>> logs for node02.phy.eze.ampgn.com.ar from that time. >>>>>> >>>>>> Also, did you use the auto-generated snapshot to start the vm? >>>>>> >>>>>> >>>>>> On Fri, May 11, 2018 at 6:11 PM, Juan Pablo < >>>>>> pablo.localh...@gmail.com> wrote: >>>>>> >>>>>>> after the xfs_repair, it says: sorry I could not find valid >>>>>>> secondary superblock >>>>>>> >>>>>>> 2018-05-11 12:09 GMT-03:00 Juan Pablo <pablo.localh...@gmail.com>: >>>>>>> >>>>>>>> hi, >>>>>>>> Alias: >>>>>>>> mail02-int_Disk1 >>>>>>>> Description: >>>>>>>> ID: >>>>>>>> 65ec515e-0aae-4fe6-a561-387929c7fb4d >>>>>>>> Alignment: >>>>>>>> Unknown >>>>>>>> Disk Profile: >>>>>>>> Wipe After Delete: >>>>>>>> No >>>>>>>> >>>>>>>> that one >>>>>>>> >>>>>>>> 2018-05-11 11:12 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>: >>>>>>>> >>>>>>>>> I looked at the logs and I see some disks have moved successfully >&g
[ovirt-users] Re: strange issue: vm lost info on disk
>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Merging of snapshot >>>>> '319e8bbb-9efe-4de4-a9a6-862e3deb891f' images >>>>> '52532d05-970e-4643-9774-96c31796062c'..'5d9d2958-96bc-49fa-9100-2f33a3ba737f' >>>>> failed. Images have been marked illegal and can no longer be previewed or >>>>> reverted to. Please retry Live Merge on the snapshot to complete the >>>>> operation. >>>>> 2018-05-11 03:38:02,519-03 ERROR [org.ovirt.engine.core.bll.sna >>>>> pshots.RemoveSnapshotSingleDiskLiveCommand] >>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-2) >>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command >>>>> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand' >>>>> with failure. >>>>> 2018-05-11 03:38:03,530-03 INFO [org.ovirt.engine.core.bll.Con >>>>> currentChildCommandsExecutionCallback] >>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-37) >>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command 'RemoveSnapshot' id: >>>>> '26bc52a4-4509-4577-b342-44a679bc628f' child commands >>>>> '[4936d196-a891-4484-9cf5-fceaafbf3364]' executions were completed, >>>>> status 'FAILED' >>>>> 2018-05-11 03:38:04,548-03 ERROR >>>>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand] >>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-66) >>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command >>>>> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand' with >>>>> failure. >>>>> 2018-05-11 03:38:04,557-03 INFO >>>>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand] >>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-66) >>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Lock freed to object >>>>> 'EngineLock:{exclusiveLocks='[4808bb70-c9cc-4286-aa39-16b579 >>>>> 8213ac=LIVE_STORAGE_MIGRATION]', sharedLocks=''}' >>>>> >>>>> I do not see the merge attempt in the vdsm.log, so please send vdsm >>>>> logs for node02.phy.eze.ampgn.com.ar from that time. >>>>> >>>>> Also, did you use the auto-generated snapshot to start the vm? >>>>> >>>>> >>>>> On Fri, May 11, 2018 at 6:11 PM, Juan Pablo <pablo.localh...@gmail.com >>>>> > wrote: >>>>> >>>>>> after the xfs_repair, it says: sorry I could not find valid secondary >>>>>> superblock >>>>>> >>>>>> 2018-05-11 12:09 GMT-03:00 Juan Pablo <pablo.localh...@gmail.com>: >>>>>> >>>>>>> hi, >>>>>>> Alias: >>>>>>> mail02-int_Disk1 >>>>>>> Description: >>>>>>> ID: >>>>>>> 65ec515e-0aae-4fe6-a561-387929c7fb4d >>>>>>> Alignment: >>>>>>> Unknown >>>>>>> Disk Profile: >>>>>>> Wipe After Delete: >>>>>>> No >>>>>>> >>>>>>> that one >>>>>>> >>>>>>> 2018-05-11 11:12 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>: >>>>>>> >>>>>>>> I looked at the logs and I see some disks have moved successfully >>>>>>>> and some failed. Which disk is causing the problems? >>>>>>>> >>>>>>>> On Fri, May 11, 2018 at 5:02 PM, Juan Pablo < >>>>>>>> pablo.localh...@gmail.com> wrote: >>>>>>>> >>>>>>>>> Hi, just sent you via drive the files. attaching some extra info, >>>>>>>>> thanks thanks and thanks : >>>>>>>>> >>>>>>>>> from inside the migrated vm I had the following attached dmesg >>>>>>>>> output before rebooting >>>>>>>>> >>>>>>>>> regards and thanks again for the help, >>>>>>>>> >>>>>>>>> 2018-05-11 10:45 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>: >>>>>>>>> >>>>>>>>>> Dropbox or google drive I guess. Also, can you attach engine.log? >>>>>>>>>> >>>>>>>>>> On Fri, May 11, 2018 at 4:43 PM, Juan Pablo < >&g
[ovirt-users] Re: strange issue: vm lost info on disk
RROR [org.ovirt.engine.core.bll.snapshots. >>>> RemoveSnapshotSingleDiskLiveCommand] >>>> (EE-ManagedThreadFactory-engineScheduled-Thread-2) >>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command >>>> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand' >>>> with failure. >>>> 2018-05-11 03:38:03,530-03 INFO [org.ovirt.engine.core.bll. >>>> ConcurrentChildCommandsExecutionCallback] >>>> (EE-ManagedThreadFactory-engineScheduled-Thread-37) >>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command 'RemoveSnapshot' id: >>>> '26bc52a4-4509-4577-b342-44a679bc628f' child commands >>>> '[4936d196-a891-4484-9cf5-fceaafbf3364]' executions were completed, >>>> status 'FAILED' >>>> 2018-05-11 03:38:04,548-03 ERROR >>>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand] >>>> (EE-ManagedThreadFactory-engineScheduled-Thread-66) >>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command >>>> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand' with >>>> failure. >>>> 2018-05-11 03:38:04,557-03 INFO >>>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand] >>>> (EE-ManagedThreadFactory-engineScheduled-Thread-66) >>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Lock freed to object >>>> 'EngineLock:{exclusiveLocks='[4808bb70-c9cc-4286-aa39- >>>> 16b5798213ac=LIVE_STORAGE_MIGRATION]', sharedLocks=''}' >>>> >>>> I do not see the merge attempt in the vdsm.log, so please send vdsm >>>> logs for node02.phy.eze.ampgn.com.ar from that time. >>>> >>>> Also, did you use the auto-generated snapshot to start the vm? >>>> >>>> >>>> On Fri, May 11, 2018 at 6:11 PM, Juan Pablo <pablo.localh...@gmail.com> >>>> wrote: >>>> >>>>> after the xfs_repair, it says: sorry I could not find valid secondary >>>>> superblock >>>>> >>>>> 2018-05-11 12:09 GMT-03:00 Juan Pablo <pablo.localh...@gmail.com>: >>>>> >>>>>> hi, >>>>>> Alias: >>>>>> mail02-int_Disk1 >>>>>> Description: >>>>>> ID: >>>>>> 65ec515e-0aae-4fe6-a561-387929c7fb4d >>>>>> Alignment: >>>>>> Unknown >>>>>> Disk Profile: >>>>>> Wipe After Delete: >>>>>> No >>>>>> >>>>>> that one >>>>>> >>>>>> 2018-05-11 11:12 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>: >>>>>> >>>>>>> I looked at the logs and I see some disks have moved successfully >>>>>>> and some failed. Which disk is causing the problems? >>>>>>> >>>>>>> On Fri, May 11, 2018 at 5:02 PM, Juan Pablo < >>>>>>> pablo.localh...@gmail.com> wrote: >>>>>>> >>>>>>>> Hi, just sent you via drive the files. attaching some extra info, >>>>>>>> thanks thanks and thanks : >>>>>>>> >>>>>>>> from inside the migrated vm I had the following attached dmesg >>>>>>>> output before rebooting >>>>>>>> >>>>>>>> regards and thanks again for the help, >>>>>>>> >>>>>>>> 2018-05-11 10:45 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>: >>>>>>>> >>>>>>>>> Dropbox or google drive I guess. Also, can you attach engine.log? >>>>>>>>> >>>>>>>>> On Fri, May 11, 2018 at 4:43 PM, Juan Pablo < >>>>>>>>> pablo.localh...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> >>>>>>>>>> vdsm is too big for gmail ...any other way I can share it with >>>>>>>>>> you? >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- Forwrded message -- >>>>>>>>>> From: Juan Pablo <pablo.localh...@gmail.com> >>>>>>>>>> Date: 2018-05-11 10:40 GMT-03:00 >>>>>>>>>> Subject: Re: [ovirt-users] strange issue: vm lost info on disk >>>>>>>>>> To: Benny Zlotnik <bzlot...@redhat.com> >>>>>>>>>> Cc: users <Users@ovirt.org> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Benny, thanks for your reply! it was a Live migration. sorry, it >>>>>>>>>> was from nfs to iscsi, not otherwise. I have reboot the vm for >>>>>>>>>> rescue and >>>>>>>>>> it does not detect any partitions with fdisk, Im running a >>>>>>>>>> xfs_repair with >>>>>>>>>> -n and found some corrupted primary superblock., its still >>>>>>>>>> running... ( >>>>>>>>>> so... there's info in the disk maybe?) >>>>>>>>>> >>>>>>>>>> attaching logs, let me know if those are the ones. >>>>>>>>>> thanks again! >>>>>>>>>> >>>>>>>>>> 2018-05-11 9:45 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>: >>>>>>>>>> >>>>>>>>>>> Can you provide the logs? engine and vdsm. >>>>>>>>>>> Did you perform a live migration (the VM is running) or cold? >>>>>>>>>>> >>>>>>>>>>> On Fri, May 11, 2018 at 2:49 PM, Juan Pablo < >>>>>>>>>>> pablo.localh...@gmail.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> Hi! , Im strugled about an ongoing problem: >>>>>>>>>>>> after migrating a vm's disk from an iscsi domain to a nfs and >>>>>>>>>>>> ovirt reporting the migration was successful, I see there's no data >>>>>>>>>>>> 'inside' the vm's disk. we never had this issues with ovirt so Im >>>>>>>>>>>> stranged >>>>>>>>>>>> about the root cause and if theres a chance of recovering the >>>>>>>>>>>> information. >>>>>>>>>>>> >>>>>>>>>>>> can you please help me out troubleshooting this one? I would >>>>>>>>>>>> really appreciate it =) >>>>>>>>>>>> running ovirt 4.2.1 here! >>>>>>>>>>>> >>>>>>>>>>>> thanks in advance, >>>>>>>>>>>> JP >>>>>>>>>>>> >>>>>>>>>>>> ___ >>>>>>>>>>>> Users mailing list -- users@ovirt.org >>>>>>>>>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> ___ >> Users mailing list -- users@ovirt.org >> To unsubscribe send an email to users-le...@ovirt.org >> > ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org
[ovirt-users] Re: strange issue: vm lost info on disk
mand 'RemoveSnapshot' id: >>> '26bc52a4-4509-4577-b342-44a679bc628f' child commands >>> '[4936d196-a891-4484-9cf5-fceaafbf3364]' executions were completed, status >>> 'FAILED' >>> 2018-05-11 03:38:04,548-03 ERROR >>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand] >>> (EE-ManagedThreadFactory-engineScheduled-Thread-66) >>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command >>> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand' with failure. >>> 2018-05-11 03:38:04,557-03 INFO >>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand] >>> (EE-ManagedThreadFactory-engineScheduled-Thread-66) >>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Lock freed to object >>> 'EngineLock:{exclusiveLocks='[4808bb70-c9cc-4286-aa39-16b5798213ac=LIVE_STORAGE_MIGRATION]', >>> sharedLocks=''}' >>> >>> I do not see the merge attempt in the vdsm.log, so please send vdsm logs >>> for node02.phy.eze.ampgn.com.ar from that time. >>> >>> Also, did you use the auto-generated snapshot to start the vm? >>> >>> >>> On Fri, May 11, 2018 at 6:11 PM, Juan Pablo <pablo.localh...@gmail.com> >>> wrote: >>> >>>> after the xfs_repair, it says: sorry I could not find valid secondary >>>> superblock >>>> >>>> 2018-05-11 12:09 GMT-03:00 Juan Pablo <pablo.localh...@gmail.com>: >>>> >>>>> hi, >>>>> Alias: >>>>> mail02-int_Disk1 >>>>> Description: >>>>> ID: >>>>> 65ec515e-0aae-4fe6-a561-387929c7fb4d >>>>> Alignment: >>>>> Unknown >>>>> Disk Profile: >>>>> Wipe After Delete: >>>>> No >>>>> >>>>> that one >>>>> >>>>> 2018-05-11 11:12 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>: >>>>> >>>>>> I looked at the logs and I see some disks have moved successfully and >>>>>> some failed. Which disk is causing the problems? >>>>>> >>>>>> On Fri, May 11, 2018 at 5:02 PM, Juan Pablo < >>>>>> pablo.localh...@gmail.com> wrote: >>>>>> >>>>>>> Hi, just sent you via drive the files. attaching some extra info, >>>>>>> thanks thanks and thanks : >>>>>>> >>>>>>> from inside the migrated vm I had the following attached dmesg >>>>>>> output before rebooting >>>>>>> >>>>>>> regards and thanks again for the help, >>>>>>> >>>>>>> 2018-05-11 10:45 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>: >>>>>>> >>>>>>>> Dropbox or google drive I guess. Also, can you attach engine.log? >>>>>>>> >>>>>>>> On Fri, May 11, 2018 at 4:43 PM, Juan Pablo < >>>>>>>> pablo.localh...@gmail.com> wrote: >>>>>>>> >>>>>>>>> >>>>>>>>> vdsm is too big for gmail ...any other way I can share it with you? >>>>>>>>> >>>>>>>>> >>>>>>>>> -- Forwrded message -- >>>>>>>>> From: Juan Pablo <pablo.localh...@gmail.com> >>>>>>>>> Date: 2018-05-11 10:40 GMT-03:00 >>>>>>>>> Subject: Re: [ovirt-users] strange issue: vm lost info on disk >>>>>>>>> To: Benny Zlotnik <bzlot...@redhat.com> >>>>>>>>> Cc: users <Users@ovirt.org> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> Benny, thanks for your reply! it was a Live migration. sorry, it >>>>>>>>> was from nfs to iscsi, not otherwise. I have reboot the vm for rescue >>>>>>>>> and >>>>>>>>> it does not detect any partitions with fdisk, Im running a xfs_repair >>>>>>>>> with >>>>>>>>> -n and found some corrupted primary superblock., its still running... >>>>>>>>> ( >>>>>>>>> so... there's info in the disk maybe?) >>>>>>>>> >>>>>>>>> attaching logs, let me know if those are the ones. >>>>>>>>> thanks again! >>>>>>>>> >>>>>>>>> 2018-05-11 9:45 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>: >>>>>>>>> >>>>>>>>>> Can you provide the logs? engine and vdsm. >>>>>>>>>> Did you perform a live migration (the VM is running) or cold? >>>>>>>>>> >>>>>>>>>> On Fri, May 11, 2018 at 2:49 PM, Juan Pablo < >>>>>>>>>> pablo.localh...@gmail.com> wrote: >>>>>>>>>> >>>>>>>>>>> Hi! , Im strugled about an ongoing problem: >>>>>>>>>>> after migrating a vm's disk from an iscsi domain to a nfs and >>>>>>>>>>> ovirt reporting the migration was successful, I see there's no data >>>>>>>>>>> 'inside' the vm's disk. we never had this issues with ovirt so Im >>>>>>>>>>> stranged >>>>>>>>>>> about the root cause and if theres a chance of recovering the >>>>>>>>>>> information. >>>>>>>>>>> >>>>>>>>>>> can you please help me out troubleshooting this one? I would >>>>>>>>>>> really appreciate it =) >>>>>>>>>>> running ovirt 4.2.1 here! >>>>>>>>>>> >>>>>>>>>>> thanks in advance, >>>>>>>>>>> JP >>>>>>>>>>> >>>>>>>>>>> ___ >>>>>>>>>>> Users mailing list -- users@ovirt.org >>>>>>>>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> > ___ > Users mailing list -- users@ovirt.org > To unsubscribe send an email to users-le...@ovirt.org > ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org
[ovirt-users] Re: strange issue: vm lost info on disk
'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand' with failure. >> 2018-05-11 03:38:04,557-03 INFO >> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand] >> (EE-ManagedThreadFactory-engineScheduled-Thread-66) >> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Lock freed to object >> 'EngineLock:{exclusiveLocks='[4808bb70-c9cc-4286-aa39-16b579 >> 8213ac=LIVE_STORAGE_MIGRATION]', sharedLocks=''}' >> >> I do not see the merge attempt in the vdsm.log, so please send vdsm logs >> for node02.phy.eze.ampgn.com.ar from that time. >> >> Also, did you use the auto-generated snapshot to start the vm? >> >> >> On Fri, May 11, 2018 at 6:11 PM, Juan Pablo <pablo.localh...@gmail.com> >> wrote: >> >>> after the xfs_repair, it says: sorry I could not find valid secondary >>> superblock >>> >>> 2018-05-11 12:09 GMT-03:00 Juan Pablo <pablo.localh...@gmail.com>: >>> >>>> hi, >>>> Alias: >>>> mail02-int_Disk1 >>>> Description: >>>> ID: >>>> 65ec515e-0aae-4fe6-a561-387929c7fb4d >>>> Alignment: >>>> Unknown >>>> Disk Profile: >>>> Wipe After Delete: >>>> No >>>> >>>> that one >>>> >>>> 2018-05-11 11:12 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>: >>>> >>>>> I looked at the logs and I see some disks have moved successfully and >>>>> some failed. Which disk is causing the problems? >>>>> >>>>> On Fri, May 11, 2018 at 5:02 PM, Juan Pablo <pablo.localh...@gmail.com >>>>> > wrote: >>>>> >>>>>> Hi, just sent you via drive the files. attaching some extra info, >>>>>> thanks thanks and thanks : >>>>>> >>>>>> from inside the migrated vm I had the following attached dmesg output >>>>>> before rebooting >>>>>> >>>>>> regards and thanks again for the help, >>>>>> >>>>>> 2018-05-11 10:45 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>: >>>>>> >>>>>>> Dropbox or google drive I guess. Also, can you attach engine.log? >>>>>>> >>>>>>> On Fri, May 11, 2018 at 4:43 PM, Juan Pablo < >>>>>>> pablo.localh...@gmail.com> wrote: >>>>>>> >>>>>>>> >>>>>>>> vdsm is too big for gmail ...any other way I can share it with you? >>>>>>>> >>>>>>>> >>>>>>>> -- Forwrded message -- >>>>>>>> From: Juan Pablo <pablo.localh...@gmail.com> >>>>>>>> Date: 2018-05-11 10:40 GMT-03:00 >>>>>>>> Subject: Re: [ovirt-users] strange issue: vm lost info on disk >>>>>>>> To: Benny Zlotnik <bzlot...@redhat.com> >>>>>>>> Cc: users <Users@ovirt.org> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> Benny, thanks for your reply! it was a Live migration. sorry, it >>>>>>>> was from nfs to iscsi, not otherwise. I have reboot the vm for rescue >>>>>>>> and >>>>>>>> it does not detect any partitions with fdisk, Im running a xfs_repair >>>>>>>> with >>>>>>>> -n and found some corrupted primary superblock., its still running... ( >>>>>>>> so... there's info in the disk maybe?) >>>>>>>> >>>>>>>> attaching logs, let me know if those are the ones. >>>>>>>> thanks again! >>>>>>>> >>>>>>>> 2018-05-11 9:45 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>: >>>>>>>> >>>>>>>>> Can you provide the logs? engine and vdsm. >>>>>>>>> Did you perform a live migration (the VM is running) or cold? >>>>>>>>> >>>>>>>>> On Fri, May 11, 2018 at 2:49 PM, Juan Pablo < >>>>>>>>> pablo.localh...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> Hi! , Im strugled about an ongoing problem: >>>>>>>>>> after migrating a vm's disk from an iscsi domain to a nfs and >>>>>>>>>> ovirt reporting the migration was successful, I see there's no data >>>>>>>>>> 'inside' the vm's disk. we never had this issues with ovirt so Im >>>>>>>>>> stranged >>>>>>>>>> about the root cause and if theres a chance of recovering the >>>>>>>>>> information. >>>>>>>>>> >>>>>>>>>> can you please help me out troubleshooting this one? I would >>>>>>>>>> really appreciate it =) >>>>>>>>>> running ovirt 4.2.1 here! >>>>>>>>>> >>>>>>>>>> thanks in advance, >>>>>>>>>> JP >>>>>>>>>> >>>>>>>>>> ___ >>>>>>>>>> Users mailing list -- users@ovirt.org >>>>>>>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> > ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org
[ovirt-users] Re: strange issue: vm lost info on disk
lso, did you use the auto-generated snapshot to start the vm? > > > On Fri, May 11, 2018 at 6:11 PM, Juan Pablo <pablo.localh...@gmail.com> > wrote: > >> after the xfs_repair, it says: sorry I could not find valid secondary >> superblock >> >> 2018-05-11 12:09 GMT-03:00 Juan Pablo <pablo.localh...@gmail.com>: >> >>> hi, >>> Alias: >>> mail02-int_Disk1 >>> Description: >>> ID: >>> 65ec515e-0aae-4fe6-a561-387929c7fb4d >>> Alignment: >>> Unknown >>> Disk Profile: >>> Wipe After Delete: >>> No >>> >>> that one >>> >>> 2018-05-11 11:12 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>: >>> >>>> I looked at the logs and I see some disks have moved successfully and >>>> some failed. Which disk is causing the problems? >>>> >>>> On Fri, May 11, 2018 at 5:02 PM, Juan Pablo <pablo.localh...@gmail.com> >>>> wrote: >>>> >>>>> Hi, just sent you via drive the files. attaching some extra info, >>>>> thanks thanks and thanks : >>>>> >>>>> from inside the migrated vm I had the following attached dmesg output >>>>> before rebooting >>>>> >>>>> regards and thanks again for the help, >>>>> >>>>> 2018-05-11 10:45 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>: >>>>> >>>>>> Dropbox or google drive I guess. Also, can you attach engine.log? >>>>>> >>>>>> On Fri, May 11, 2018 at 4:43 PM, Juan Pablo < >>>>>> pablo.localh...@gmail.com> wrote: >>>>>> >>>>>>> >>>>>>> vdsm is too big for gmail ...any other way I can share it with you? >>>>>>> >>>>>>> >>>>>>> -- Forwrded message -- >>>>>>> From: Juan Pablo <pablo.localh...@gmail.com> >>>>>>> Date: 2018-05-11 10:40 GMT-03:00 >>>>>>> Subject: Re: [ovirt-users] strange issue: vm lost info on disk >>>>>>> To: Benny Zlotnik <bzlot...@redhat.com> >>>>>>> Cc: users <Users@ovirt.org> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> Benny, thanks for your reply! it was a Live migration. sorry, it was >>>>>>> from nfs to iscsi, not otherwise. I have reboot the vm for rescue and it >>>>>>> does not detect any partitions with fdisk, Im running a xfs_repair with >>>>>>> -n >>>>>>> and found some corrupted primary superblock., its still running... ( >>>>>>> so... >>>>>>> there's info in the disk maybe?) >>>>>>> >>>>>>> attaching logs, let me know if those are the ones. >>>>>>> thanks again! >>>>>>> >>>>>>> 2018-05-11 9:45 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>: >>>>>>> >>>>>>>> Can you provide the logs? engine and vdsm. >>>>>>>> Did you perform a live migration (the VM is running) or cold? >>>>>>>> >>>>>>>> On Fri, May 11, 2018 at 2:49 PM, Juan Pablo < >>>>>>>> pablo.localh...@gmail.com> wrote: >>>>>>>> >>>>>>>>> Hi! , Im strugled about an ongoing problem: >>>>>>>>> after migrating a vm's disk from an iscsi domain to a nfs and >>>>>>>>> ovirt reporting the migration was successful, I see there's no data >>>>>>>>> 'inside' the vm's disk. we never had this issues with ovirt so Im >>>>>>>>> stranged >>>>>>>>> about the root cause and if theres a chance of recovering the >>>>>>>>> information. >>>>>>>>> >>>>>>>>> can you please help me out troubleshooting this one? I would >>>>>>>>> really appreciate it =) >>>>>>>>> running ovirt 4.2.1 here! >>>>>>>>> >>>>>>>>> thanks in advance, >>>>>>>>> JP >>>>>>>>> >>>>>>>>> ___ >>>>>>>>> Users mailing list -- users@ovirt.org >>>>>>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> > ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org
[ovirt-users] Re: strange issue: vm lost info on disk
k <bzlot...@redhat.com>: >> >>> I looked at the logs and I see some disks have moved successfully and >>> some failed. Which disk is causing the problems? >>> >>> On Fri, May 11, 2018 at 5:02 PM, Juan Pablo <pablo.localh...@gmail.com> >>> wrote: >>> >>>> Hi, just sent you via drive the files. attaching some extra info, >>>> thanks thanks and thanks : >>>> >>>> from inside the migrated vm I had the following attached dmesg output >>>> before rebooting >>>> >>>> regards and thanks again for the help, >>>> >>>> 2018-05-11 10:45 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>: >>>> >>>>> Dropbox or google drive I guess. Also, can you attach engine.log? >>>>> >>>>> On Fri, May 11, 2018 at 4:43 PM, Juan Pablo <pablo.localh...@gmail.com >>>>> > wrote: >>>>> >>>>>> >>>>>> vdsm is too big for gmail ...any other way I can share it with you? >>>>>> >>>>>> >>>>>> -- Forwrded message -- >>>>>> From: Juan Pablo <pablo.localh...@gmail.com> >>>>>> Date: 2018-05-11 10:40 GMT-03:00 >>>>>> Subject: Re: [ovirt-users] strange issue: vm lost info on disk >>>>>> To: Benny Zlotnik <bzlot...@redhat.com> >>>>>> Cc: users <Users@ovirt.org> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> Benny, thanks for your reply! it was a Live migration. sorry, it was >>>>>> from nfs to iscsi, not otherwise. I have reboot the vm for rescue and it >>>>>> does not detect any partitions with fdisk, Im running a xfs_repair with >>>>>> -n >>>>>> and found some corrupted primary superblock., its still running... ( >>>>>> so... >>>>>> there's info in the disk maybe?) >>>>>> >>>>>> attaching logs, let me know if those are the ones. >>>>>> thanks again! >>>>>> >>>>>> 2018-05-11 9:45 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>: >>>>>> >>>>>>> Can you provide the logs? engine and vdsm. >>>>>>> Did you perform a live migration (the VM is running) or cold? >>>>>>> >>>>>>> On Fri, May 11, 2018 at 2:49 PM, Juan Pablo < >>>>>>> pablo.localh...@gmail.com> wrote: >>>>>>> >>>>>>>> Hi! , Im strugled about an ongoing problem: >>>>>>>> after migrating a vm's disk from an iscsi domain to a nfs and >>>>>>>> ovirt reporting the migration was successful, I see there's no data >>>>>>>> 'inside' the vm's disk. we never had this issues with ovirt so Im >>>>>>>> stranged >>>>>>>> about the root cause and if theres a chance of recovering the >>>>>>>> information. >>>>>>>> >>>>>>>> can you please help me out troubleshooting this one? I would really >>>>>>>> appreciate it =) >>>>>>>> running ovirt 4.2.1 here! >>>>>>>> >>>>>>>> thanks in advance, >>>>>>>> JP >>>>>>>> >>>>>>>> ___ >>>>>>>> Users mailing list -- users@ovirt.org >>>>>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>>> >>>>> >>>> >>> >> > ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org
[ovirt-users] Re: strange issue: vm lost info on disk
Can you provide the logs? engine and vdsm. Did you perform a live migration (the VM is running) or cold? On Fri, May 11, 2018 at 2:49 PM, Juan Pablowrote: > Hi! , Im strugled about an ongoing problem: > after migrating a vm's disk from an iscsi domain to a nfs and ovirt > reporting the migration was successful, I see there's no data 'inside' the > vm's disk. we never had this issues with ovirt so Im stranged about the > root cause and if theres a chance of recovering the information. > > can you please help me out troubleshooting this one? I would really > appreciate it =) > running ovirt 4.2.1 here! > > thanks in advance, > JP > > ___ > Users mailing list -- users@ovirt.org > To unsubscribe send an email to users-le...@ovirt.org > > ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org