Hey Johann,

I think this is not an actual bug but a general issue with the migration 
convergence + a general problem with reported error messages 
during migrations (both are being worked on and targeted to 4.0).

In your particular case the root cause seems like this from logs:
"The migration took 40 seconds which is exceeding the configured maximum time 
for migrations of 32 s
econds. The migration will be aborted."

So I'd say the VM was able to trash the memory faster than being moved to 
destination and after 40s it was aborted. 
What you can do is to go to the vdsm.conf on each host and tweak migration 
progress timeout, migration bandwidth (if you have capacity for more than 32 
MBps on the migration network per VM) and maybe max migration downtime (the 
time how long the VM can be paused at the end of migration).

Also from inside of the guest you can try to reduce the load if possible so it 
will not trash the memory so fast...

hope some of this helps,
Tomas

----- Original Message -----
> From: "Johann Kiefel" <i...@dshin.net>
> To: users@ovirt.org
> Sent: Thursday, December 17, 2015 4:32:36 PM
> Subject: [ovirt-users] Migration of VM fails Failed in 'DestroyVDS' method 
> Message: VDSM command failed: Virtual
> machine does not exist
> 
> 
> Dear collegues,
> 
> please help us out. The migration of VMs from host to host fails, while we
> think that we configured everything properly. We searched through the
> mailing list and found nothing that describes/handles our error.
> 
> 
> The Details:
> 
> We have three identical hosts and completed the self_hosted_engine
> installation on all of them (Host OS is the latest CentOS 7). As shared
> storage we use a NFS share hosted by an independent storage server. The
> shares are presented on a dedicated network: (192.168.23.0/24). All of the
> hosts have two networks. The management network is 192.168.10.0\24 and the
> 10Gbit storage network: 192.168.23.0/24.
> 
> Our hosted engine (192.168.10.20) is up and running. All three hosts are
> connected and in status up. We create the second logical network "storage"
> for our physical storage network. Syncing and testing it on every host. This
> "storage" Network get the roles "migration" and "gluster" the Hosted-Engine
> UI.
> We create and mount our Data, Iso and Export domains in the Hosted-Engine UI
> over 192.168.23.0 (storage) network. We successfully install an CentOs 6 or
> 7 VM and can connect to them.
> 
> Now if we try to migrate the VM, the migration fails.
> 
> 
> The Logs:
> 
> 
> 
> ###################################################
> engine.log from the Engine VM
> 
> ###################################################
> 
> 
> 2015-12-17 13:50:09,033 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVDSCommand]
> (DefaultQuartzScheduler_Worker-34) [20ed2774] START,
> FullListVDSCommand(HostName = ,
> FullListVDSCommandParameters:{runAsync='true', hostId='62e30
> d77-fe4b-41d4-a9f7-b66d0a6ad7bd',
> vds='Host[,62e30d77-fe4b-41d4-a9f7-b66d0a6ad7bd]',
> vmIds='[408e7c39-d2a2-4412-8522-6eda216f17e0]'}), log id: 34cea1dd
> 2015-12-17 13:50:09,075 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVDSCommand]
> (DefaultQuartzScheduler_Worker-34) [20ed2774] FINISH, FullListVDSCommand,
> return: [{status=Up, nicModel=rtl8139,pv, emulatedMachine=pc, guestDisk
> Mapping={398398f3-9572-4d67-8={name=/dev/vda}, QEMU_DVD-ROM={name=/dev/sr0}},
> vmId=408e7c39-d2a2-4412-8522-6eda216f17e0, pid=19010,
> devices=[Ljava.lang.Object;@69e4cf40, smp=4, vmType=kvm, displayIp=0,
> display=vnc, displaySecurePort=-1, m
> emSize=4096, displayPort=5900, cpuType=SandyBridge,
> spiceSecureChannels=smain,sdisplay,sinputs,scursor,splayback,srecord,ssmartcard,susbredir,
> statusTime=4455363570, vmName=HostedEngine, clientIp=, pauseCode=NOERR}],
> log id: 34cea1dd
> 2015-12-17 13:50:09,076 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVDSCommand]
> (DefaultQuartzScheduler_Worker-34) [20ed2774] START,
> FullListVDSCommand(HostName = ,
> FullListVDSCommandParameters:{runAsync='true', hostId='62e30
> d77-fe4b-41d4-a9f7-b66d0a6ad7bd',
> vds='Host[,62e30d77-fe4b-41d4-a9f7-b66d0a6ad7bd]',
> vmIds='[0e956372-f4bd-4535-8472-73829277593f]'}), log id: 74f63ccf
> 2015-12-17 13:50:09,105 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVDSCommand]
> (DefaultQuartzScheduler_Worker-34) [20ed2774] FINISH, FullListVDSCommand,
> return: [{acpiEnable=true, emulatedMachine=pc-i440fx-rhel7.2.0, tabletE
> nable=true, pid=30907, memGuaranteedSize=512, transparentHugePages=true,
> displaySecurePort=-1, spiceSslCipherSuite=DEFAULT, cpuType=SandyBridge,
> smp=1, pauseCode=NOERR, guestNumaNodes=[Ljava.lang.Object;@74afaff3,
> smartcardEnable=false, n
> umaTune={nodeset=0,1, mode=interleave}, custom={}, vmType=kvm, memSize=512,
> smpCoresPerSocket=1, vmName=tiny-i1, nice=0, status=Up, maxMemSize=4194304,
> bootMenuEnable=false, vmId=0e956372-f4bd-4535-8472-73829277593f,
> displayIp=0, displayP
> ort=5901, guestDiskMapping={},
> spiceSecureChannels=smain,sinputs,scursor,splayback,srecord,sdisplay,susbredir,ssmartcard,
> nicModel=rtl8139,pv, keyboardLayout=de, kvmEnable=true,
> pitReinjection=false, devices=[Ljava.lang.Object;@3a48b65e,
> display=vnc, timeOffset=3600, maxVCpus=16, clientIp=192.168.1.16,
> statusTime=4455363610, maxMemSlots=16}], log id: 74f63ccf
> 2015-12-17 13:50:09,106 INFO  [org.ovirt.engine.core.vdsbroker.VmsMonitoring]
> (DefaultQuartzScheduler_Worker-34) [20ed2774] Received a memballoon Device
> without an address when processing VM 0e956372-f4bd-4535-8472-73829277593f
> devices, s
> kipping device: {device=memballoon, specParams={model=none},
> deviceType=balloon, type=balloon, alias=balloon0}
> 2015-12-17 13:50:09,106 INFO  [org.ovirt.engine.core.vdsbroker.VmsMonitoring]
> (DefaultQuartzScheduler_Worker-34) [20ed2774] Received a vnc Device without
> an address when processing VM 0e956372-f4bd-4535-8472-73829277593f devices,
> skipping
>  device: {device=vnc, deviceType=graphics, type=graphics,
>  deviceId=c6ebce28-25d5-418c-a46d-0c21d61a62fd, port=5901}
> 2015-12-17 13:50:09,107 ERROR [org.ovirt.engine.core.vdsbroker.VmsMonitoring]
> (DefaultQuartzScheduler_Worker-34) [20ed2774] VM
> '0e956372-f4bd-4535-8472-73829277593f' managed non pluggable device was
> removed unexpectedly from libvirt: 'VmD
> evice:{id='VmDeviceId:{deviceId='62338ecf-adf4-4e12-826b-cc837e75055e',
> vmId='0e956372-f4bd-4535-8472-73829277593f'}', device='vnc',
> type='GRAPHICS', bootOrder='0', specParams='[]', address='', managed='true',
> plugged='false', readOnly='f
> alse', deviceAlias='', customProperties='[]', snapshotId='null',
> logicalName='null', usingScsiReservation='false'}'
> 2015-12-17 13:50:09,107 ERROR [org.ovirt.engine.core.vdsbroker.VmsMonitoring]
> (DefaultQuartzScheduler_Worker-34) [20ed2774] VM
> '0e956372-f4bd-4535-8472-73829277593f' managed non pluggable device was
> removed unexpectedly from libvirt: 'VmD
> evice:{id='VmDeviceId:{deviceId='8c2bad1b-ef19-4f43-9e62-d912f98e3901',
> vmId='0e956372-f4bd-4535-8472-73829277593f'}', device='cdrom', type='DISK',
> bootOrder='0', specParams='[path=]', address='', managed='true',
> plugged='false', readOnly
> ='true', deviceAlias='', customProperties='[]', snapshotId='null',
> logicalName='null', usingScsiReservation='false'}'
> 2015-12-17 13:50:24,185 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVDSCommand]
> (DefaultQuartzScheduler_Worker-74) [50b5a02f] START,
> FullListVDSCommand(HostName = ,
> FullListVDSCommandParameters:{runAsync='true', hostId='62e30
> d77-fe4b-41d4-a9f7-b66d0a6ad7bd',
> vds='Host[,62e30d77-fe4b-41d4-a9f7-b66d0a6ad7bd]',
> vmIds='[408e7c39-d2a2-4412-8522-6eda216f17e0]'}), log id: 30a52aac
> 2015-12-17 13:50:25,103 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVDSCommand]
> (DefaultQuartzScheduler_Worker-74) [50b5a02f] FINISH, FullListVDSCommand,
> return: [{status=Up, nicModel=rtl8139,pv, emulatedMachine=pc, guestDisk
> Mapping={398398f3-9572-4d67-8={name=/dev/vda}, QEMU_DVD-ROM={name=/dev/sr0}},
> vmId=408e7c39-d2a2-4412-8522-6eda216f17e0, pid=19010,
> devices=[Ljava.lang.Object;@793cf9d3, smp=4, vmType=kvm, displayIp=0,
> display=vnc, displaySecurePort=-1, m
> emSize=4096, displayPort=5900, cpuType=SandyBridge,
> spiceSecureChannels=smain,sdisplay,sinputs,scursor,splayback,srecord,ssmartcard,susbredir,
> statusTime=4455378730, vmName=HostedEngine, clientIp=, pauseCode=NOERR}],
> log id: 30a52aac
> 2015-12-17 13:50:40,230 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVDSCommand]
> (DefaultQuartzScheduler_Worker-91) [6b1ec6a9] START,
> FullListVDSCommand(HostName = ,
> FullListVDSCommandParameters:{runAsync='true', hostId='62e30
> d77-fe4b-41d4-a9f7-b66d0a6ad7bd',
> vds='Host[,62e30d77-fe4b-41d4-a9f7-b66d0a6ad7bd]',
> vmIds='[408e7c39-d2a2-4412-8522-6eda216f17e0]'}), log id: 294d3ffe
> 2015-12-17 13:50:40,275 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVDSCommand]
> (DefaultQuartzScheduler_Worker-91) [6b1ec6a9] FINISH, FullListVDSCommand,
> return: [{status=Up, nicModel=rtl8139,pv, emulatedMachine=pc, guestDisk
> Mapping={398398f3-9572-4d67-8={name=/dev/vda}, QEMU_DVD-ROM={name=/dev/sr0}},
> vmId=408e7c39-d2a2-4412-8522-6eda216f17e0, pid=19010,
> devices=[Ljava.lang.Object;@ea451e1, smp=4, vmType=kvm, displayIp=0,
> display=vnc, displaySecurePort=-1, me
> mSize=4096, displayPort=5900, cpuType=SandyBridge,
> spiceSecureChannels=smain,sdisplay,sinputs,scursor,splayback,srecord,ssmartcard,susbredir,
> statusTime=4455394760, vmName=HostedEngine, clientIp=, pauseCode=NOERR}],
> log id: 294d3ffe
> 2015-12-17 13:50:55,641 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVDSCommand]
> (DefaultQuartzScheduler_Worker-5) [] START, FullListVDSCommand(HostName = ,
> FullListVDSCommandParameters:{runAsync='true', hostId='62e30d77-fe4b-
> 41d4-a9f7-b66d0a6ad7bd', vds='Host[,62e30d77-fe4b-41d4-a9f7-b66d0a6ad7bd]',
> vmIds='[408e7c39-d2a2-4412-8522-6eda216f17e0]'}), log id: 2aedaa54
> 2015-12-17 13:50:56,157 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVDSCommand]
> (DefaultQuartzScheduler_Worker-5) [] FINISH, FullListVDSCommand, return:
> [{status=Up, nicModel=rtl8139,pv, emulatedMachine=pc, guestDiskMapping={
> 398398f3-9572-4d67-8={name=/dev/vda}, QEMU_DVD-ROM={name=/dev/sr0}},
> vmId=408e7c39-d2a2-4412-8522-6eda216f17e0, pid=19010,
> devices=[Ljava.lang.Object;@17b907dd, smp=4, vmType=kvm, displayIp=0,
> display=vnc, displaySecurePort=-1, memSize=40
> 96, displayPort=5900, cpuType=SandyBridge,
> spiceSecureChannels=smain,sdisplay,sinputs,scursor,splayback,srecord,ssmartcard,susbredir,
> statusTime=4455410180, vmName=HostedEngine, clientIp=, pauseCode=NOERR}],
> log id: 2aedaa54
> 2015-12-17 13:51:06,563 INFO
> [org.ovirt.engine.core.bll.MigrateVmToServerCommand] (default task-4)
> [5e46a29f] Lock Acquired to object
> 'EngineLock:{exclusiveLocks='[0e956372-f4bd-4535-8472-73829277593f=<VM,
> ACTION_TYPE_FAILED_VM_IS_BEING_
> MIGRATED$VmName tiny-i1>]', sharedLocks='null'}'
> 2015-12-17 13:51:06,595 INFO
> [org.ovirt.engine.core.bll.scheduling.SchedulingManager] (default task-4)
> [5e46a29f] Candidate host 'hosted_engine_1'
> ('62e30d77-fe4b-41d4-a9f7-b66d0a6ad7bd') was filtered out by
> 'VAR__FILTERTYPE__INTERNAL' f
> ilter 'Migration' (correlation id: null)
> 2015-12-17 13:51:06,618 INFO
> [org.ovirt.engine.core.bll.MigrateVmToServerCommand]
> (org.ovirt.thread.pool-8-thread-20) [5e46a29f] Running command:
> MigrateVmToServerCommand internal: false. Entities affected :  ID:
> 0e956372-f4bd-4535-8472-
> 73829277593f Type: VMAction group MIGRATE_VM with role type USER
> 2015-12-17 13:51:06,639 INFO
> [org.ovirt.engine.core.vdsbroker.MigrateVDSCommand]
> (org.ovirt.thread.pool-8-thread-20) [5e46a29f] START, MigrateVDSCommand(
> MigrateVDSCommandParameters:{runAsync='true',
> hostId='62e30d77-fe4b-41d4-a9f7-b66d0
> a6ad7bd', vmId='0e956372-f4bd-4535-8472-73829277593f',
> srcHost='m1.example.de', dstVdsId='e3189b9a-b725-4a7c-987c-6b22bd4abea2',
> dstHost='n1.example.de:54321', migrationMethod='ONLINE',
> tunnelMigration='false', migrationDowntime='0', au
> toConverge='false', migrateCompressed='false', consoleAddress='null'}), log
> id: 25229fa3
> 2015-12-17 13:51:06,640 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand]
> (org.ovirt.thread.pool-8-thread-20) [5e46a29f] START,
> MigrateBrokerVDSCommand(HostName = hosted_engine_1,
> MigrateVDSCommandParameters:{runAs
> ync='true', hostId='62e30d77-fe4b-41d4-a9f7-b66d0a6ad7bd',
> vmId='0e956372-f4bd-4535-8472-73829277593f', srcHost='m1.example.de',
> dstVdsId='e3189b9a-b725-4a7c-987c-6b22bd4abea2',
> dstHost='n1.example.de:54321', migrationMethod='ONLINE', t
> unnelMigration='false', migrationDowntime='0', autoConverge='false',
> migrateCompressed='false', consoleAddress='null'}), log id: 4f18f441
> 2015-12-17 13:51:06,703 ERROR
> [org.ovirt.engine.core.bll.host.provider.foreman.SystemProviderFinder]
> (default task-6) [] Failed to find host on any provider by host name
> 'ov.example.de'
> 2015-12-17 13:51:07,184 ERROR
> [org.ovirt.engine.core.bll.host.provider.foreman.SystemProviderFinder]
> (default task-2) [] Failed to find host on any provider by host name
> 'ov.example.de'
> 2015-12-17 13:51:07,305 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.MigrateBrokerVDSCommand]
> (org.ovirt.thread.pool-8-thread-20) [5e46a29f] FINISH,
> MigrateBrokerVDSCommand, log id: 4f18f441
> 2015-12-17 13:51:07,310 INFO
> [org.ovirt.engine.core.vdsbroker.MigrateVDSCommand]
> (org.ovirt.thread.pool-8-thread-20) [5e46a29f] FINISH, MigrateVDSCommand,
> return: MigratingFrom, log id: 25229fa3
> 2015-12-17 13:51:07,316 INFO
> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> (org.ovirt.thread.pool-8-thread-20) [5e46a29f] Correlation ID: 5e46a29f, Job
> ID: 8c267b6a-3d37-45b5-837c-39bcee545521, Call Stack: null,
> Custom Event ID: -1, Message: Migration started (VM: tiny-i1, Source:
> hosted_engine_1, Destination: hosted_engine_2, User: admin@internal).
> 2015-12-17 13:51:08,022 ERROR
> [org.ovirt.engine.core.bll.host.provider.foreman.SystemProviderFinder]
> (default task-15) [] Failed to find host on any provider by host name
> 'ov.example.de'
> 2015-12-17 13:51:11,224 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVDSCommand]
> (DefaultQuartzScheduler_Worker-79) [] START, FullListVDSCommand(HostName = ,
> FullListVDSCommandParameters:{runAsync='true', hostId='62e30d77-fe4b
> -41d4-a9f7-b66d0a6ad7bd', vds='Host[,62e30d77-fe4b-41d4-a9f7-b66d0a6ad7bd]',
> vmIds='[408e7c39-d2a2-4412-8522-6eda216f17e0]'}), log id: c4b9f52
> 2015-12-17 13:51:11,254 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVDSCommand]
> (DefaultQuartzScheduler_Worker-79) [] FINISH, FullListVDSCommand, return:
> [{status=Up, nicModel=rtl8139,pv, emulatedMachine=pc, guestDiskMapping=
> {398398f3-9572-4d67-8={name=/dev/vda}, QEMU_DVD-ROM={name=/dev/sr0}},
> vmId=408e7c39-d2a2-4412-8522-6eda216f17e0, pid=19010,
> devices=[Ljava.lang.Object;@54d8b81c, smp=4, vmType=kvm, displayIp=0,
> display=vnc, displaySecurePort=-1, memSize=4
> 096, displayPort=5900, cpuType=SandyBridge,
> spiceSecureChannels=smain,sdisplay,sinputs,scursor,splayback,srecord,ssmartcard,susbredir,
> statusTime=4455425760, vmName=HostedEngine, clientIp=, pauseCode=NOERR}],
> log id: c4b9f52
> 2015-12-17 13:51:12,232 INFO  [org.ovirt.engine.core.vdsbroker.VmAnalyzer]
> (DefaultQuartzScheduler_Worker-83) [59659503] RefreshVmList VM id
> '0e956372-f4bd-4535-8472-73829277593f' is migrating to VDS 'hosted_engine_2'
> ignoring it in the r
> efresh until migration is done
> 2015-12-17 13:51:26,290 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVDSCommand]
> (DefaultQuartzScheduler_Worker-21) [2fb3e350] START,
> FullListVDSCommand(HostName = ,
> FullListVDSCommandParameters:{runAsync='true', hostId='62e30
> d77-fe4b-41d4-a9f7-b66d0a6ad7bd',
> vds='Host[,62e30d77-fe4b-41d4-a9f7-b66d0a6ad7bd]',
> vmIds='[408e7c39-d2a2-4412-8522-6eda216f17e0]'}), log id: 29a38cc2
> 2015-12-17 13:51:26,690 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVDSCommand]
> (DefaultQuartzScheduler_Worker-21) [2fb3e350] FINISH, FullListVDSCommand,
> return: [{status=Up, nicModel=rtl8139,pv, emulatedMachine=pc, guestDisk
> Mapping={398398f3-9572-4d67-8={name=/dev/vda}, QEMU_DVD-ROM={name=/dev/sr0}},
> vmId=408e7c39-d2a2-4412-8522-6eda216f17e0, pid=19010,
> devices=[Ljava.lang.Object;@306e29e9, smp=4, vmType=kvm, displayIp=0,
> display=vnc, displaySecurePort=-1, m
> emSize=4096, displayPort=5900, cpuType=SandyBridge,
> spiceSecureChannels=smain,sdisplay,sinputs,scursor,splayback,srecord,ssmartcard,susbredir,
> statusTime=4455440820, vmName=HostedEngine, clientIp=, pauseCode=NOERR}],
> log id: 29a38cc2
> 2015-12-17 13:51:27,295 INFO  [org.ovirt.engine.core.vdsbroker.VmAnalyzer]
> (DefaultQuartzScheduler_Worker-85) [] RefreshVmList VM id
> '0e956372-f4bd-4535-8472-73829277593f' is migrating to VDS 'hosted_engine_2'
> ignoring it in the refresh u
> ntil migration is done
> 2015-12-17 13:51:41,726 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVDSCommand]
> (DefaultQuartzScheduler_Worker-60) [] START, FullListVDSCommand(HostName = ,
> FullListVDSCommandParameters:{runAsync='true', hostId='62e30d77-fe4b
> -41d4-a9f7-b66d0a6ad7bd', vds='Host[,62e30d77-fe4b-41d4-a9f7-b66d0a6ad7bd]',
> vmIds='[408e7c39-d2a2-4412-8522-6eda216f17e0]'}), log id: 224c57c9
> 2015-12-17 13:51:42,263 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.FullListVDSCommand]
> (DefaultQuartzScheduler_Worker-60) [] FINISH, FullListVDSCommand, return:
> [{status=Up, nicModel=rtl8139,pv, emulatedMachine=pc, guestDiskMapping=
> {398398f3-9572-4d67-8={name=/dev/vda}, QEMU_DVD-ROM={name=/dev/sr0}},
> vmId=408e7c39-d2a2-4412-8522-6eda216f17e0, pid=19010,
> devices=[Ljava.lang.Object;@29eb6e5, smp=4, vmType=kvm, displayIp=0,
> display=vnc, displaySecurePort=-1, memSize=40
> 96, displayPort=5900, cpuType=SandyBridge,
> spiceSecureChannels=smain,sdisplay,sinputs,scursor,splayback,srecord,ssmartcard,susbredir,
> statusTime=4455456260, vmName=HostedEngine, clientIp=, pauseCode=NOERR}],
> log id: 224c57c9
> 2015-12-17 13:51:42,311 INFO  [org.ovirt.engine.core.vdsbroker.VmAnalyzer]
> (DefaultQuartzScheduler_Worker-71) [433f9bb6] RefreshVmList VM id
> '0e956372-f4bd-4535-8472-73829277593f' is migrating to VDS 'hosted_engine_2'
> ignoring it in the r
> efresh until migration is done
> 2015-12-17 13:51:47,215 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
> (ForkJoinPool-1-worker-294) [] START, DestroyVDSCommand(HostName =
> hosted_engine_2, DestroyVmVDSCommandParameters:{runAsync='true',
> hostId='e3189b
> 9a-b725-4a7c-987c-6b22bd4abea2', vmId='0e956372-f4bd-4535-8472-73829277593f',
> force='false', secondsToWait='0', gracefully='false', reason=''}), log id:
> 24778267
> 2015-12-17 13:51:47,318 ERROR
> [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
> (ForkJoinPool-1-worker-294) [] Failed in 'DestroyVDS' method
> 2015-12-17 13:51:47,321 ERROR
> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
> (ForkJoinPool-1-worker-294) [] Correlation ID: null, Call Stack: null,
> Custom Event ID: -1, Message: VDSM hosted_engine_2 command failed:
>  Virtual machine does not exist
> 2015-12-17 13:51:47,321 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
> (ForkJoinPool-1-worker-294) [] Command
> 'org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand' return value
> 'StatusOnlyReturnForXmlRpc [stat
> us=StatusForXmlRpc [code=1, message=Virtual machine does not exist]]'
> 2015-12-17 13:51:47,321 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
> (ForkJoinPool-1-worker-294) [] HostName = hosted_engine_2
> 2015-12-17 13:51:47,321 ERROR
> [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
> (ForkJoinPool-1-worker-294) [] Command 'DestroyVDSCommand(HostName =
> hosted_engine_2, DestroyVmVDSCommandParameters:{runAsync='true',
> hostId='e318
> 9b9a-b725-4a7c-987c-6b22bd4abea2',
> vmId='0e956372-f4bd-4535-8472-73829277593f', force='false',
> secondsToWait='0', gracefully='false', reason=''})' execution failed:
> VDSGenericException: VDSErrorException: Failed to DestroyVDS, error = Vir
> tual machine does not exist, code = 1
> 2015-12-17 13:51:47,321 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand]
> (ForkJoinPool-1-worker-294) [] FINISH, DestroyVDSCommand, log id: 24778267
> 2015-12-17 13:51:47,322 INFO  [org.ovirt.engine.core.vdsbroker.VmAnalyzer]
> (ForkJoinPool-1-worker-294) [] RefreshVmList VM id
> '0e956372-f4bd-4535-8472-73829277593f' status = 'Down' on VDS
> 'hosted_engine_2' ignoring it in the refresh until
>  migration is done
> 2015-12-17 13:51:49,203 ERROR
> [org.ovirt.engine.core.bll.host.provider.foreman.SystemProviderFinder]
> (default task-42) [] Failed to find host on any provider by host name
> 'ov.example.de'
> 2015-12-17 13:51:55,416 INFO
> [org.ovirt.vdsm.jsonrpc.client.reactors.ReactorClient] (SSL Stomp Reactor)
> [] Connecting to n1.example.de/192.168.10.12
> 2015-12-17 13:51:57,272 INFO
> [org.ovirt.vdsm.jsonrpc.client.reactors.ReactorClient] (SSL Stomp Reactor)
> [] Connecting to m1.example.de/192.168.10.11
> 2015-12-17 13:51:57,916 INFO  [org.ovirt.engine.core.vdsbroker.VmAnalyzer]
> (DefaultQuartzScheduler_Worker-28) [ff193b8] VM
> '0e956372-f4bd-4535-8472-73829277593f'(tiny-i1) moved from 'MigratingFrom'
> --> 'Up'
> 2015-12-17 13:51:57,916 INFO  [org.ovirt.engine.core.vdsbroker.VmAnalyzer]
> (DefaultQuartzScheduler_Worker-28) [ff193b8] Adding VM
> '0e956372-f4bd-4535-8472-73829277593f' to re-run list
> 
> 
> 
> 
> ###################################################
> The vdsm.log on Source Host:
> 
> ###################################################
> 
> 
> Thread-103967::DEBUG::2015-12-17
> 14:51:40,619::fileSD::173::Storage.Misc.excCmd::(getReadDelay) /usr/bin/dd
> if=/rhev/data-center/mnt/192.168.23.10:_mnt_sdb_nfs_ISO/e793de29-f95f-4512-9a57-ece72268b53f/dom_md/metadata
> iflag=direct of=/dev/
> null bs=4096 count=1 (cwd None)
> Thread-103967::DEBUG::2015-12-17
> 14:51:40,626::fileSD::173::Storage.Misc.excCmd::(getReadDelay) SUCCESS:
> <err> = '0+1 records in\n0+1 records out\n330 bytes (330 B) copied,
> 0.00023453 s, 1.4 MB/s\n'; <rc> = 0
> Reactor thread::INFO::2015-12-17
> 14:51:41,423::protocoldetector::72::ProtocolDetector.AcceptorImpl::(handle_accept)
> Accepting connection from 127.0.0.1:36777
> Reactor thread::DEBUG::2015-12-17
> 14:51:41,427::protocoldetector::82::ProtocolDetector.Detector::(__init__)
> Using required_size=11
> Reactor thread::INFO::2015-12-17
> 14:51:41,428::protocoldetector::118::ProtocolDetector.Detector::(handle_read)
> Detected protocol xml from 127.0.0.1:36777
> Reactor thread::DEBUG::2015-12-17
> 14:51:41,428::bindingxmlrpc::1297::XmlDetector::(handle_socket) xml over
> http detected from ('127.0.0.1', 36777)
> BindingXMLRPC::INFO::2015-12-17
> 14:51:41,428::xmlrpc::73::vds.XMLRPCServer::(handle_request) Starting
> request handler for 127.0.0.1:36777
> Thread-108220::INFO::2015-12-17
> 14:51:41,429::xmlrpc::84::vds.XMLRPCServer::(_process_requests) Request
> handler for 127.0.0.1:36777 started
> Thread-108220::DEBUG::2015-12-17
> 14:51:41,431::task::595::Storage.TaskManager.Task::(_updateState)
> Task=`2cdd058a-ef2c-4eea-a172-33b07f23be33`::moving from state init -> state
> preparing
> Thread-108220::INFO::2015-12-17
> 14:51:41,432::logUtils::48::dispatcher::(wrapper) Run and protect:
> repoStats(options=None)
> Thread-108220::INFO::2015-12-17
> 14:51:41,432::logUtils::51::dispatcher::(wrapper) Run and protect:
> repoStats, Return response: {u'94048698-25fe-4f3e-8f49-34b8633ee996':
> {'code': 0, 'actual': True, 'version': 0, 'acquired': True, 'delay':
> '0.000238152', 'lastCheck': '6.9', 'valid': True},
> u'e793de29-f95f-4512-9a57-ece72268b53f': {'code': 0, 'actual': True,
> 'version': 0, 'acquired': True, 'delay': '0.00023453', 'lastCheck': '0.8',
> 'valid': True}, 'fac2b63e-4359-4ad8-871c-58
> ae92254ba9': {'code': 0, 'actual': True, 'version': 3, 'acquired': True,
> 'delay': '0.000251451', 'lastCheck': '8.3', 'valid': True},
> u'ff24327d-3080-4344-b083-43cc5acd5342': {'code': 0, 'actual': True,
> 'version': 3, 'acquired': True, 'del
> ay': '0.000231135', 'lastCheck': '8.5', 'valid': True}}
> Thread-108220::DEBUG::2015-12-17
> 14:51:41,432::task::1191::Storage.TaskManager.Task::(prepare)
> Task=`2cdd058a-ef2c-4eea-a172-33b07f23be33`::finished:
> {u'94048698-25fe-4f3e-8f49-34b8633ee996': {'code': 0, 'actual': True,
> 'version': 0, 'acq
> uired': True, 'delay': '0.000238152', 'lastCheck': '6.9', 'valid': True},
> u'e793de29-f95f-4512-9a57-ece72268b53f': {'code': 0, 'actual': True,
> 'version': 0, 'acquired': True, 'delay': '0.00023453', 'lastCheck': '0.8',
> 'valid': True}, 'fac
> 2b63e-4359-4ad8-871c-58ae92254ba9': {'code': 0, 'actual': True, 'version': 3,
> 'acquired': True, 'delay': '0.000251451', 'lastCheck': '8.3', 'valid':
> True}, u'ff24327d-3080-4344-b083-43cc5acd5342': {'code': 0, 'actual': True,
> 'version': 3,
>  'acquired': True, 'delay': '0.000231135', 'lastCheck': '8.5', 'valid':
>  True}}
> Thread-108220::DEBUG::2015-12-17
> 14:51:41,432::task::595::Storage.TaskManager.Task::(_updateState)
> Task=`2cdd058a-ef2c-4eea-a172-33b07f23be33`::moving from state preparing ->
> state finished
> Thread-108220::DEBUG::2015-12-17
> 14:51:41,432::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
> Owner.releaseAll requests {} resources {}
> Thread-108220::DEBUG::2015-12-17
> 14:51:41,432::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
> Owner.cancelAll requests {}
> Thread-108220::DEBUG::2015-12-17
> 14:51:41,432::task::993::Storage.TaskManager.Task::(_decref)
> Task=`2cdd058a-ef2c-4eea-a172-33b07f23be33`::ref 0 aborting False
> Thread-108220::INFO::2015-12-17
> 14:51:41,489::xmlrpc::92::vds.XMLRPCServer::(_process_requests) Request
> handler for 127.0.0.1:36777 stopped
> Thread-108162::WARNING::2015-12-17
> 14:51:41,931::migration::526::virt.vm::(monitor_migration)
> vmId=`0e956372-f4bd-4535-8472-73829277593f`::The migration took 40 seconds
> which is exceeding the configured maximum time for migrations of 32 s
> econds. The migration will be aborted.
> Thread-108162::DEBUG::2015-12-17
> 14:51:41,933::migration::558::virt.vm::(stop)
> vmId=`0e956372-f4bd-4535-8472-73829277593f`::stopping migration monitor
> thread
> Thread-108160::DEBUG::2015-12-17
> 14:51:42,372::migration::558::virt.vm::(stop)
> vmId=`0e956372-f4bd-4535-8472-73829277593f`::stopping migration monitor
> thread
> Thread-108160::DEBUG::2015-12-17
> 14:51:42,372::migration::453::virt.vm::(stop)
> vmId=`0e956372-f4bd-4535-8472-73829277593f`::stopping migration downtime
> thread
> Thread-108161::DEBUG::2015-12-17 14:51:42,372::migration::450::virt.vm::(run)
> vmId=`0e956372-f4bd-4535-8472-73829277593f`::migration downtime thread
> exiting
> Thread-108160::ERROR::2015-12-17
> 14:51:42,372::migration::208::virt.vm::(_recover)
> vmId=`0e956372-f4bd-4535-8472-73829277593f`::Operation abgebrochen:
> Migrations-Job: abgebrochen durch Client
> Thread-108160::DEBUG::2015-12-17
> 14:51:42,373::stompreactor::389::jsonrpc.AsyncoreClient::(send) Sending
> response
> Thread-108160::DEBUG::2015-12-17
> 14:51:42,406::__init__::206::jsonrpc.Notification::(emit) Sending event
> {"params": {"notify_time": 4455461740,
> "0e956372-f4bd-4535-8472-73829277593f": {"status": "Migration Source"}},
> "jsonrpc": "2.0", "me
> thod": "|virt|VM_status|0e956372-f4bd-4535-8472-73829277593f"}
> Thread-108160::ERROR::2015-12-17 14:51:42,407::migration::310::virt.vm::(run)
> vmId=`0e956372-f4bd-4535-8472-73829277593f`::Failed to migrate
> Traceback (most recent call last):
>   File "/usr/share/vdsm/virt/migration.py", line 294, in run
>     self._startUnderlyingMigration(time.time())
>   File "/usr/share/vdsm/virt/migration.py", line 364, in
>   _startUnderlyingMigration
>     self._perform_migration(duri, muri)
>   File "/usr/share/vdsm/virt/migration.py", line 403, in _perform_migration
>     self._vm._dom.migrateToURI3(duri, params, flags)
>   File "/usr/share/vdsm/virt/virdomain.py", line 68, in f
>     ret = attr(*args, **kwargs)
>   File "/usr/lib/python2.7/site-packages/vdsm/libvirtconnection.py", line
>   124, in wrapper
>     ret = f(*args, **kwargs)
>   File "/usr/lib64/python2.7/site-packages/libvirt.py", line 1836, in
>   migrateToURI3
>     if ret == -1: raise libvirtError ('virDomainMigrateToURI3() failed',
>     dom=self)
> libvirtError: Operation abgebrochen: Migrations-Job: abgebrochen durch Client
> Thread-103715::DEBUG::2015-12-17
> 14:51:42,895::fileSD::173::Storage.Misc.excCmd::(getReadDelay) /usr/bin/dd
> if=/rhev/data-center/mnt/192.168.23.10:_mnt_sdb_nfs_DATA/ff24327d-3080-4344-b083-43cc5acd5342/dom_md/metadata
> iflag=direct of=/dev
> /null bs=4096 count=1 (cwd None)
> Thread-103715::DEBUG::2015-12-17
> 14:51:42,905::fileSD::173::Storage.Misc.excCmd::(getReadDelay) SUCCESS:
> <err> = '0+1 records in\n0+1 records out\n463 bytes (463 B) copied,
> 0.000402171 s, 1.2 MB/s\n'; <rc> = 0
> Thread-113::DEBUG::2015-12-17
> 14:51:43,093::fileSD::173::Storage.Misc.excCmd::(getReadDelay) /usr/bin/dd
> if=/rhev/data-center/mnt/192.168.23.10:_mnt_sdb_nfs_HOSTED__ENGINE/fac2b63e-4359-4ad8-871c-58ae92254ba9/dom_md/metadata
> iflag=direct
> of=/dev/null bs=4096 count=1 (cwd None)
> Thread-113::DEBUG::2015-12-17
> 14:51:43,100::fileSD::173::Storage.Misc.excCmd::(getReadDelay) SUCCESS:
> <err> = '0+1 records in\n0+1 records out\n316 bytes (316 B) copied,
> 0.000300531 s, 1.1 MB/s\n'; <rc> = 0
> 
> 
> 
> 
> ##################################
> vdsm.log on target
> ##################################
> 
> (Is too long and attached) but below are the significant lines:
> 
> 
> 1:32,193::periodic::254::virt.periodic.VmDispatcher::(__call__) could not run
> <class 'virt.periodic.DriveWatermarkMonitor'> on
> [u'0e956372-f4bd-4535-8472-73829277593f']
> Thread-4556::DEBUG::2015-12-17
> 14:51:32,752::fileSD::173::Storage.Misc.excCmd::(getReadDelay) /usr/bin/dd
> if=/rhev/data-center/mnt/192.168.23.10:_mnt_sdb_nfs_ISO/e793de29-f95f-4512-9a57-ece72268b53f/dom_md/metadata
> iflag=direct of=/dev/nu
> ll bs=4096 count=1 (cwd None)
> Thread-4556::DEBUG::2015-12-17
> 14:51:32,760::fileSD::173::Storage.Misc.excCmd::(getReadDelay) SUCCESS:
> <err> = '0+1 records in\n0+1 records out\n330 bytes (330 B) copied,
> 0.000202573 s, 1.6 MB/s\n'; <rc> = 0
> Thread-4246::DEBUG::2015-12-17
> 14:51:33,185::fileSD::173::Storage.Misc.excCmd::(getReadDelay) /usr/bin/dd
> if=/rhev/data-center/mnt/192.168.23.10:_mnt_sdb_nfs_DATA/ff24327d-3080-4344-b083-43cc5acd5342/dom_md/metadata
> iflag=direct of=/dev/n
> ull bs=4096 count=1 (cwd None)
> Thread-4246::DEBUG::2015-12-17
> 14:51:33,194::fileSD::173::Storage.Misc.excCmd::(getReadDelay) SUCCESS:
> <err> = '0+1 records in\n0+1 records out\n463 bytes (463 B) copied,
> 0.000259942 s, 1.8 MB/s\n'; <rc> = 0
> periodic/4::WARNING::2015-12-17
> 14:51:34,193::periodic::254::virt.periodic.VmDispatcher::(__call__) could
> not run <class 'virt.periodic.DriveWatermarkMonitor'> on
> [u'0e956372-f4bd-4535-8472-73829277593f']
> periodic/2::WARNING::2015-12-17
> 14:51:36,194::periodic::254::virt.periodic.VmDispatcher::(__call__) could
> not run <class 'virt.periodic.DriveWatermarkMonitor'> on
> [u'0e956372-f4bd-4535-8472-73829277593f']
> periodic/1::WARNING::2015-12-17
> 14:51:36,615::periodic::254::virt.periodic.VmDispatcher::(__call__) could
> not run <class 'virt.periodic.BlockjobMonitor'> on
> [u'0e956372-f4bd-4535-8472-73829277593f']
> periodic/4::WARNING::2015-12-17
> 14:51:36,615::periodic::277::virt.vm::(__call__)
> vmId=`0e956372-f4bd-4535-8472-73829277593f`::could not run on
> 0e956372-f4bd-4535-8472-73829277593f: domain not connected
> periodic/2::DEBUG::2015-12-17
> 14:51:36,625::sampling::531::virt.sampling.VMBulkSampler::(__call__)
> sampling 0 domains
> .....
> 
> Thread-8857::INFO::2015-12-17
> 14:51:40,738::xmlrpc::92::vds.XMLRPCServer::(_process_requests) Request
> handler for 127.0.0.1:58492 stopped
> periodic/4::WARNING::2015-12-17
> 14:51:42,197::periodic::254::virt.periodic.VmDispatcher::(__call__) could
> not run <class 'virt.periodic.DriveWatermarkMonitor'> on
> [u'0e956372-f4bd-4535-8472-73829277593f']
> Thread-4556::DEBUG::2015-12-17
> 14:51:42,763::fileSD::173::Storage.Misc.excCmd::(getReadDelay) /usr/bin/dd
> if=/rhev/data-center/mnt/192.168.23.10:_mnt_sdb_nfs_ISO/e793de29-f95f-4512-9a57-ece72268b53f/dom_md/metadata
> iflag=direct of=/dev/nu
> ll bs=4096 count=1 (cwd None)
> Thread-4556::DEBUG::2015-12-17
> 14:51:42,770::fileSD::173::Storage.Misc.excCmd::(getReadDelay) SUCCESS:
> <err> = '0+1 records in\n0+1 records out\n330 bytes (330 B) copied,
> 0.000225483 s, 1.5 MB/s\n'; <rc> = 0
> Thread-4246::DEBUG::2015-12-17
> 14:51:43,199::fileSD::173::Storage.Misc.excCmd::(getReadDelay) /usr/bin/dd
> if=/rhev/data-center/mnt/192.168.23.10:_mnt_sdb_nfs_DATA/ff24327d-3080-4344-b083-43cc5acd5342/dom_md/metadata
> iflag=direct of=/dev/n
> ull bs=4096 count=1 (cwd None)
> Thread-4246::DEBUG::2015-12-17
> 14:51:43,207::fileSD::173::Storage.Misc.excCmd::(getReadDelay) SUCCESS:
> <err> = '0+1 records in\n0+1 records out\n463 bytes (463 B) copied,
> 0.000226464 s, 2.0 MB/s\n'; <rc> = 0
> periodic/2::WARNING::2015-12-17
> 14:51:44,198::periodic::254::virt.periodic.VmDispatcher::(__call__) could
> not run <class 'virt.periodic.DriveWatermarkMonitor'> on
> [u'0e956372-f4bd-4535-8472-73829277593f']
> 
> .....
> 
> Thread-8781::ERROR::2015-12-17
> 14:51:46,636::vm::752::virt.vm::(_startUnderlyingVm)
> vmId=`0e956372-f4bd-4535-8472-73829277593f`::Failed to start a migration
> destination vm
> Traceback (most recent call last):
>   File "/usr/share/vdsm/virt/vm.py", line 721, in _startUnderlyingVm
>     self._completeIncomingMigration()
>   File "/usr/share/vdsm/virt/vm.py", line 2785, in _completeIncomingMigration
>     self._incomingMigrationFinished.isSet(), usedTimeout)
>   File "/usr/share/vdsm/virt/vm.py", line 2838, in
>   _attachLibvirtDomainAfterMigration
>     raise MigrationError(e.get_error_message())
> MigrationError: Domain nicht gefunden: Keine Domain mit ´┐Ż´┐Żbereinstimmender
> UUID '0e956372-f4bd-4535-8472-73829277593f'
> _______________________________________________
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to