ok, so Im confirming that the image is wrong somehow:
with no snapshot, from inside the vm disk size is reporting 750G.
with a snapshot, from inside the vm disk size is reporting 1100G.
both have no partitions on it, so I guess ovirt migrated the structure of
the 750G disk on a 1100 disk, any ideas to troubleshoot this and see if
there's data to recover?

regards,


2018-05-13 15:25 GMT-03:00 Juan Pablo <pablo.localh...@gmail.com>:

> 2 clues:
> -the original size of the disk was 750G and was extended a month ago to
> 1100G. The System rebooted fine several times, and took the new size with
> no problems.
>
> -I run fdisk from a centos 7 rescue cd and '/dev/vda' reported 750G. then,
> I took a snapshot of the disk to play with recovery tools and now fdisk
> reports 1100G...  ¬¬
>
> so my guess is on the extend and later migration to a different storage
> domain caused the issue.
> Im currently running testdisk to see if theres any partition to recover.
>
> regards,
>
> 2018-05-13 12:31 GMT-03:00 Juan Pablo <pablo.localh...@gmail.com>:
>
>> I removed the auto-snapshot and still no lucky. no bootable disk found. =(
>> ideas?
>>
>>
>> 2018-05-13 12:26 GMT-03:00 Juan Pablo <pablo.localh...@gmail.com>:
>>
>>> benny, thanks for your reply:
>>> ok, so the steps are : removing the snapshot on the first place. then
>>> what do you suggest?
>>>
>>>
>>> 2018-05-12 15:23 GMT-03:00 Nir Soffer <nsof...@redhat.com>:
>>>
>>>> On Sat, 12 May 2018, 11:32 Benny Zlotnik, <bzlot...@redhat.com> wrote:
>>>>
>>>>> Using the auto-generated snapshot is generally a bad idea as it's
>>>>> inconsistent,
>>>>>
>>>>
>>>> What do you mean by inconsistant?
>>>>
>>>>
>>>> you should remove it before moving further
>>>>>
>>>>> On Fri, May 11, 2018 at 7:25 PM, Juan Pablo <pablo.localh...@gmail.com
>>>>> > wrote:
>>>>>
>>>>>> I rebooted it with no luck, them I used the auto-gen snapshot , same
>>>>>> luck.
>>>>>> attaching the logs in gdrive
>>>>>>
>>>>>> thanks in advance
>>>>>>
>>>>>> 2018-05-11 12:50 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>:
>>>>>>
>>>>>>> I see here a failed attempt:
>>>>>>> 2018-05-09 16:00:20,129-03 ERROR [org.ovirt.engine.core.dal.dbb
>>>>>>> roker.auditloghandling.AuditLogDirector]
>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-67)
>>>>>>> [bd8eeb1d-f49a-4f91-a521-e0f31b4a7cbd] EVENT_ID:
>>>>>>> USER_MOVED_DISK_FINISHED_FAILURE(2,011), User admin@internal-authz
>>>>>>> have failed to move disk mail02-int_Disk1 to domain 2penLA.
>>>>>>>
>>>>>>> Then another:
>>>>>>> 2018-05-09 16:15:06,998-03 ERROR [org.ovirt.engine.core.dal.dbb
>>>>>>> roker.auditloghandling.AuditLogDirector]
>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-34) [] EVENT_ID:
>>>>>>> USER_MOVED_DISK_FINISHED_FAILURE(2,011), User admin@internal-authz
>>>>>>> have failed to move disk mail02-int_Disk1 to domain 2penLA.
>>>>>>>
>>>>>>> Here I see a successful attempt:
>>>>>>> 2018-05-09 21:58:42,628-03 INFO  [org.ovirt.engine.core.dal.dbb
>>>>>>> roker.auditloghandling.AuditLogDirector] (default task-50)
>>>>>>> [940b051c-8c63-4711-baf9-f3520bb2b825] EVENT_ID:
>>>>>>> USER_MOVED_DISK(2,008), User admin@internal-authz moving disk
>>>>>>> mail02-int_Disk1 to domain 2penLA.
>>>>>>>
>>>>>>>
>>>>>>> Then, in the last attempt I see the attempt was successful but live
>>>>>>> merge failed:
>>>>>>> 2018-05-11 03:37:59,509-03 ERROR 
>>>>>>> [org.ovirt.engine.core.bll.MergeStatusCommand]
>>>>>>> (EE-ManagedThreadFactory-commandCoordinator-Thread-2)
>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Failed to live merge, still
>>>>>>> in volume chain: [5d9d2958-96bc-49fa-9100-2f33a3ba737f,
>>>>>>> 52532d05-970e-4643-9774-96c31796062c]
>>>>>>> 2018-05-11 03:38:01,495-03 INFO  [org.ovirt.engine.core.bll.Ser
>>>>>>> ialChildCommandsExecutionCallback] 
>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-51)
>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command 'LiveMigrateDisk'
>>>>>>> (id: '115fc375-6018-4d59-b9f2-51ee05ca49f8') waiting on child
>>>>>>> command id: '26bc52a4-4509-4577-b342-44a679bc628f'
>>>>>>> type:'RemoveSnapshot' to complete
>>>>>>> 2018-05-11 03:38:01,501-03 ERROR [org.ovirt.engine.core.bll.sna
>>>>>>> pshots.RemoveSnapshotSingleDiskLiveCommand]
>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-51)
>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command id:
>>>>>>> '4936d196-a891-4484-9cf5-fceaafbf3364 failed child command status
>>>>>>> for step 'MERGE_STATUS'
>>>>>>> 2018-05-11 03:38:01,501-03 INFO  [org.ovirt.engine.core.bll.sna
>>>>>>> pshots.RemoveSnapshotSingleDiskLiveCommandCallback]
>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-51)
>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command
>>>>>>> 'RemoveSnapshotSingleDiskLive' id: 
>>>>>>> '4936d196-a891-4484-9cf5-fceaafbf3364'
>>>>>>> child commands '[8da5f261-7edd-4930-8d9d-d34f232d84b3,
>>>>>>> 1c320f4b-7296-43c4-a3e6-8a868e23fc35, 
>>>>>>> a0e9e70c-cd65-4dfb-bd00-076c4e99556a]'
>>>>>>> executions were completed, status 'FAILED'
>>>>>>> 2018-05-11 03:38:02,513-03 ERROR [org.ovirt.engine.core.bll.sna
>>>>>>> pshots.RemoveSnapshotSingleDiskLiveCommand]
>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-2)
>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Merging of snapshot
>>>>>>> '319e8bbb-9efe-4de4-a9a6-862e3deb891f' images
>>>>>>> '52532d05-970e-4643-9774-96c31796062c'..'5d9d2958-96bc-49fa-9100-2f33a3ba737f'
>>>>>>> failed. Images have been marked illegal and can no longer be previewed 
>>>>>>> or
>>>>>>> reverted to. Please retry Live Merge on the snapshot to complete the
>>>>>>> operation.
>>>>>>> 2018-05-11 03:38:02,519-03 ERROR [org.ovirt.engine.core.bll.sna
>>>>>>> pshots.RemoveSnapshotSingleDiskLiveCommand]
>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-2)
>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command
>>>>>>> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotSingleDiskLiveCommand'
>>>>>>> with failure.
>>>>>>> 2018-05-11 03:38:03,530-03 INFO  [org.ovirt.engine.core.bll.Con
>>>>>>> currentChildCommandsExecutionCallback]
>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-37)
>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Command 'RemoveSnapshot' id:
>>>>>>> '26bc52a4-4509-4577-b342-44a679bc628f' child commands
>>>>>>> '[4936d196-a891-4484-9cf5-fceaafbf3364]' executions were completed,
>>>>>>> status 'FAILED'
>>>>>>> 2018-05-11 03:38:04,548-03 ERROR 
>>>>>>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand]
>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-66)
>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Ending command
>>>>>>> 'org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand' with
>>>>>>> failure.
>>>>>>> 2018-05-11 03:38:04,557-03 INFO  
>>>>>>> [org.ovirt.engine.core.bll.snapshots.RemoveSnapshotCommand]
>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-66)
>>>>>>> [d5b7fdf5-9c37-4c1f-8543-a7bc75c993a5] Lock freed to object
>>>>>>> 'EngineLock:{exclusiveLocks='[4808bb70-c9cc-4286-aa39-16b579
>>>>>>> 8213ac=LIVE_STORAGE_MIGRATION]', sharedLocks=''}'
>>>>>>>
>>>>>>> I do not see the merge attempt in the vdsm.log, so please send vdsm
>>>>>>> logs for node02.phy.eze.ampgn.com.ar from that time.
>>>>>>>
>>>>>>> Also, did you use the auto-generated snapshot to start the vm?
>>>>>>>
>>>>>>>
>>>>>>> On Fri, May 11, 2018 at 6:11 PM, Juan Pablo <
>>>>>>> pablo.localh...@gmail.com> wrote:
>>>>>>>
>>>>>>>> after the xfs_repair, it says: sorry I could not find valid
>>>>>>>> secondary superblock
>>>>>>>>
>>>>>>>> 2018-05-11 12:09 GMT-03:00 Juan Pablo <pablo.localh...@gmail.com>:
>>>>>>>>
>>>>>>>>> hi,
>>>>>>>>> Alias:
>>>>>>>>> mail02-int_Disk1
>>>>>>>>> Description:
>>>>>>>>> ID:
>>>>>>>>> 65ec515e-0aae-4fe6-a561-387929c7fb4d
>>>>>>>>> Alignment:
>>>>>>>>> Unknown
>>>>>>>>> Disk Profile:
>>>>>>>>> Wipe After Delete:
>>>>>>>>> No
>>>>>>>>>
>>>>>>>>> that one
>>>>>>>>>
>>>>>>>>> 2018-05-11 11:12 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>:
>>>>>>>>>
>>>>>>>>>> I looked at the logs and I see some disks have moved successfully
>>>>>>>>>> and some failed. Which disk is causing the problems?
>>>>>>>>>>
>>>>>>>>>> On Fri, May 11, 2018 at 5:02 PM, Juan Pablo <
>>>>>>>>>> pablo.localh...@gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi, just sent you via drive the files. attaching some extra
>>>>>>>>>>> info, thanks thanks and thanks :
>>>>>>>>>>>
>>>>>>>>>>> from inside the migrated vm I had the following attached dmesg
>>>>>>>>>>> output before rebooting
>>>>>>>>>>>
>>>>>>>>>>> regards and thanks again for the help,
>>>>>>>>>>>
>>>>>>>>>>> 2018-05-11 10:45 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>:
>>>>>>>>>>>
>>>>>>>>>>>> Dropbox or google drive I guess. Also, can you attach
>>>>>>>>>>>> engine.log?
>>>>>>>>>>>>
>>>>>>>>>>>> On Fri, May 11, 2018 at 4:43 PM, Juan Pablo <
>>>>>>>>>>>> pablo.localh...@gmail.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> vdsm is too big for gmail ...any other way I can share it with
>>>>>>>>>>>>> you?
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> ---------- Forwrded message ----------
>>>>>>>>>>>>> From: Juan Pablo <pablo.localh...@gmail.com>
>>>>>>>>>>>>> Date: 2018-05-11 10:40 GMT-03:00
>>>>>>>>>>>>> Subject: Re: [ovirt-users] strange issue: vm lost info on disk
>>>>>>>>>>>>> To: Benny Zlotnik <bzlot...@redhat.com>
>>>>>>>>>>>>> Cc: users <Users@ovirt.org>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> ​​
>>>>>>>>>>>>> Benny, thanks for your reply! it was a Live migration. sorry,
>>>>>>>>>>>>> it was from nfs to iscsi, not otherwise. I have reboot the vm for 
>>>>>>>>>>>>> rescue
>>>>>>>>>>>>> and it does not detect any partitions with fdisk, Im running a 
>>>>>>>>>>>>> xfs_repair
>>>>>>>>>>>>> with -n and found some corrupted primary superblock., its still 
>>>>>>>>>>>>> running...
>>>>>>>>>>>>> ( so... there's info in the disk maybe?)
>>>>>>>>>>>>>
>>>>>>>>>>>>> attaching logs, let me know if those are the ones.
>>>>>>>>>>>>> thanks again!
>>>>>>>>>>>>>
>>>>>>>>>>>>> 2018-05-11 9:45 GMT-03:00 Benny Zlotnik <bzlot...@redhat.com>:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> Can you provide the logs? engine and vdsm.
>>>>>>>>>>>>>> Did you perform a live migration (the VM is running) or cold?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Fri, May 11, 2018 at 2:49 PM, Juan Pablo <
>>>>>>>>>>>>>> pablo.localh...@gmail.com> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Hi! , Im strugled about an ongoing problem:
>>>>>>>>>>>>>>>  after migrating a vm's disk from an iscsi domain to a nfs
>>>>>>>>>>>>>>> and ovirt reporting the migration was successful, I see there's 
>>>>>>>>>>>>>>> no data
>>>>>>>>>>>>>>> 'inside' the vm's disk. we never had this issues with ovirt so 
>>>>>>>>>>>>>>> Im stranged
>>>>>>>>>>>>>>> about the root cause and if theres a chance of recovering the 
>>>>>>>>>>>>>>> information.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> can you please help me out troubleshooting this one? I would
>>>>>>>>>>>>>>> really appreciate it =)
>>>>>>>>>>>>>>> running ovirt 4.2.1 here!
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> thanks in advance,
>>>>>>>>>>>>>>> JP
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>>> Users mailing list -- users@ovirt.org
>>>>>>>>>>>>>>> To unsubscribe send an email to users-le...@ovirt.org
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>> _______________________________________________
>>>>> Users mailing list -- users@ovirt.org
>>>>> To unsubscribe send an email to users-le...@ovirt.org
>>>>>
>>>>
>>>
>>
>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org

Reply via email to