I see here a failed attempt:
2018-05-09 16:00:20,129-03 ERROR [org.ovirt.engine.core.dal.
dbbroker.auditloghandling.AuditLogDirector]
(EE-ManagedThreadFactory-engineScheduled-Thread-67)
[bd8eeb1d-f49a-4f91-a521-e0f31b4a7cbd] EVENT_ID:
USER_MOVED_DISK_FINISHED_FAILURE(2,011),
User admin@internal-authz have failed to move disk mail02-int_Disk1 to
domain 2penLA.
Then another:
2018-05-09 16:15:06,998-03 ERROR [org.ovirt.engine.core.dal.
dbbroker.auditloghandling.AuditLogDirector]
(EE-ManagedThreadFactory-engineScheduled-Thread-34)
[] EVENT_ID: USER_MOVED_DISK_FINISHED_FAILURE(2,011), User
admin@internal-authz have failed to move disk mail02-int_Disk1 to domain
2penLA.
Here I see a successful attempt:
2018-05-09 21:58:42,628-03 INFO [org.ovirt.engine.core.dal.
dbbroker.auditloghandling.AuditLogDirector] (default task-50)
[940b051c-8c63-4711-baf9-f3520bb2b825] EVENT_ID: USER_MOVED_DISK(2,008),
User admin@internal-authz moving disk mail02-int_Disk1 to domain 2penLA.
Then, in the last attempt I see the attempt was successful but live merge
failed:
2018-05-11 03:37:59,509-03 ERROR
[org.ovirt.engine.core.bll.MergeStatusCommand]
(EE-ManagedThreadFactory-commandCoordinator-Thread-2)
[d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Failed to live merge, still in
volume chain: [5d9d2958-96bc-49fa-9100-2f33a3ba737f,
52532d05-970e-4643-9774-96c31796062c]
2018-05-11 03:38:01,495-03 INFO
[org.ovirt.engine.core.bll.SerialChildCommandsExecutionCallback]
(EE-ManagedThreadFactory-engineScheduled-Thread-51)
[d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command 'LiveMigrateDisk' (id:
'115fc375-6018-4d59-b9f2-51ee05ca49f8') waiting on child command id:
'26bc52a4-4509-4577-b342-44a679bc628f' type:'RemoveSnapshot' to complete
2018-05-11 03:38:01,501-03 ERROR
[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand]
(EE-ManagedThreadFactory-engineScheduled-Thread-51)
[d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command id:
'4936d196-a891-4484-9cf5-fceaafbf3364 failed child command status for step
'MERGE_STATUS'
2018-05-11 03:38:01,501-03 INFO
[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommandCallback]
(EE-ManagedThreadFactory-engineScheduled-Thread-51)
[d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command
'RemoveSnapshotSingleDiskLive' id: '4936d196-a891-4484-9cf5-fceaafbf3364'
child commands '[8da5f261-7edd-4930-8d9d-d34f232d84b3,
1c320f4b-7296-43c4-a3e6-8a868e23fc35,
a0e9e70c-cd65-4dfb-bd00-076c4e99556a]' executions were completed, status
'FAILED'
2018-05-11 03:38:02,513-03 ERROR
[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand]
(EE-ManagedThreadFactory-engineScheduled-Thread-2)
[d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Merging of snapshot
'319e8bbb-9efe-4de4-a9a6-862e3deb891f' images
'52532d05-970e-4643-9774-96c31796062c'..'5d9d2958-96bc-49fa-9100-2f33a3ba737f'
failed. Images have been marked illegal and can no longer be previewed or
reverted to. Please retry Live Merge on the snapshot to complete the
operation.
2018-05-11 03:38:02,519-03 ERROR
[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand]
(EE-ManagedThreadFactory-engineScheduled-Thread-2)
[d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command
'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand'
with failure.
2018-05-11 03:38:03,530-03 INFO
[org.ovirt.engine.core.bll.ConcurrentChildCommandsExecutionCallback]
(EE-ManagedThreadFactory-engineScheduled-Thread-37)
[d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command 'RemoveSnapshot' id:
'26bc52a4-4509-4577-b342-44a679bc628f' child commands
'[4936d196-a891-4484-9cf5-fceaafbf3364]' executions were completed, status
'FAILED'
2018-05-11 03:38:04,548-03 ERROR
[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand]
(EE-ManagedThreadFactory-engineScheduled-Thread-66)
[d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command
'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand' with failure.
2018-05-11 03:38:04,557-03 INFO
[org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand]
(EE-ManagedThreadFactory-engineScheduled-Thread-66)
[d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Lock freed to object
'EngineLock:{exclusiveLocks='[4808bb70-c9cc-4286-aa39-16b5798213ac=LIVE_STORAGE_MIGRATION]',
sharedLocks=''}'
I do not see the merge attempt in the vdsm.log, so please send vdsm logs
for node02.phy.eze.ampgn.com.ar from that time.
Also, did you use the auto-generated snapshot to start the vm?
On Fri, May 11, 2018 at 6:11 PM, Juan Pablo <[email protected]>
wrote:
> after the xfs_repair, it says: sorry I could not find valid secondary
> superblock
>
> 2018-05-11 12:09 GMT-03:00 Juan Pablo <[email protected]>:
>
>> hi,
>> Alias:
>> mail02-int_Disk1
>> Description:
>> ID:
>> 65ec515e-0aae-4fe6-a561-387929c7fb4d
>> Alignment:
>> Unknown
>> Disk Profile:
>> Wipe After Delete:
>> No
>>
>> that one
>>
>> 2018-05-11 11:12 GMT-03:00 Benny Zlotnik <[email protected]>:
>>
>>> I looked at the logs and I see some disks have moved successfully and
>>> some failed. Which disk is causing the problems?
>>>
>>> On Fri, May 11, 2018 at 5:02 PM, Juan Pablo <[email protected]>
>>> wrote:
>>>
>>>> Hi, just sent you via drive the files. attaching some extra info,
>>>> thanks thanks and thanks :
>>>>
>>>> from inside the migrated vm I had the following attached dmesg output
>>>> before rebooting
>>>>
>>>> regards and thanks again for the help,
>>>>
>>>> 2018-05-11 10:45 GMT-03:00 Benny Zlotnik <[email protected]>:
>>>>
>>>>> Dropbox or google drive I guess. Also, can you attach engine.log?
>>>>>
>>>>> On Fri, May 11, 2018 at 4:43 PM, Juan Pablo <[email protected]
>>>>> > wrote:
>>>>>
>>>>>>
>>>>>> vdsm is too big for gmail ...any other way I can share it with you?
>>>>>>
>>>>>>
>>>>>> ---------- Forwrded message ----------
>>>>>> From: Juan Pablo <[email protected]>
>>>>>> Date: 2018-05-11 10:40 GMT-03:00
>>>>>> Subject: Re: [ovirt-users] strange issue: vm lost info on disk
>>>>>> To: Benny Zlotnik <[email protected]>
>>>>>> Cc: users <[email protected]>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> Benny, thanks for your reply! it was a Live migration. sorry, it was
>>>>>> from nfs to iscsi, not otherwise. I have reboot the vm for rescue and it
>>>>>> does not detect any partitions with fdisk, Im running a xfs_repair with
>>>>>> -n
>>>>>> and found some corrupted primary superblock., its still running... (
>>>>>> so...
>>>>>> there's info in the disk maybe?)
>>>>>>
>>>>>> attaching logs, let me know if those are the ones.
>>>>>> thanks again!
>>>>>>
>>>>>> 2018-05-11 9:45 GMT-03:00 Benny Zlotnik <[email protected]>:
>>>>>>
>>>>>>> Can you provide the logs? engine and vdsm.
>>>>>>> Did you perform a live migration (the VM is running) or cold?
>>>>>>>
>>>>>>> On Fri, May 11, 2018 at 2:49 PM, Juan Pablo <
>>>>>>> [email protected]> wrote:
>>>>>>>
>>>>>>>> Hi! , Im strugled about an ongoing problem:
>>>>>>>> after migrating a vm's disk from an iscsi domain to a nfs and
>>>>>>>> ovirt reporting the migration was successful, I see there's no data
>>>>>>>> 'inside' the vm's disk. we never had this issues with ovirt so Im
>>>>>>>> stranged
>>>>>>>> about the root cause and if theres a chance of recovering the
>>>>>>>> information.
>>>>>>>>
>>>>>>>> can you please help me out troubleshooting this one? I would really
>>>>>>>> appreciate it =)
>>>>>>>> running ovirt 4.2.1 here!
>>>>>>>>
>>>>>>>> thanks in advance,
>>>>>>>> JP
>>>>>>>>
>>>>>>>> _______________________________________________
>>>>>>>> Users mailing list -- [email protected]
>>>>>>>> To unsubscribe send an email to [email protected]
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]