Hi Michal,

Thanks for assisting.

I've just done as requested however nothing is logged in the engine.log at
the time I click Migrate, below is the log and I hit the Migrate button
about 4 times between 09:35 and 09:36 and nothing was logged about this...

2019-07-10 09:35:57,967+02 INFO
 [org.ovirt.engine.core.sso.utils.AuthenticationUtils] (default task-14) []
User trouble@internal successfully logged in with scopes: ovirt-app-admin
ovirt-app-api ovirt-app-portal ovirt-ext=auth:sequence-priority=~
ovirt-ext=revoke:revoke-all ovirt-ext=token-info:authz-search
ovirt-ext=token-info:public-authz-search ovirt-ext=token-info:validate
ovirt-ext=token:password-access
2019-07-10 09:35:58,012+02 INFO
 [org.ovirt.engine.core.bll.aaa.CreateUserSessionCommand] (default task-14)
[2997034] Running command: CreateUserSessionCommand internal: false.
2019-07-10 09:35:58,021+02 INFO
 [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
(default task-14) [2997034] EVENT_ID: USER_VDC_LOGIN(30), User
trouble@internal-authz connecting from '160.128.20.85' using session
'bv55G0wZznETUiQwjgjfUNje7wOsG4UDCuFunSslVeAFQkhdY2zzTY7du36ynTF5nW5U7JiPyr7gl9QDHfWuig=='
logged in.
2019-07-10 09:36:58,304+02 INFO
 [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
(EE-ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] Thread pool
'default' is using 0 threads out of 1, 5 threads waiting for tasks.
2019-07-10 09:36:58,305+02 INFO
 [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
(EE-ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] Thread pool
'engine' is using 0 threads out of 500, 16 threads waiting for tasks and 0
tasks in queue.
2019-07-10 09:36:58,305+02 INFO
 [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
(EE-ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] Thread pool
'engineScheduled' is using 0 threads out of 100, 100 threads waiting for
tasks.
2019-07-10 09:36:58,305+02 INFO
 [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
(EE-ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] Thread pool
'engineThreadMonitoring' is using 1 threads out of 1, 0 threads waiting for
tasks.
2019-07-10 09:36:58,305+02 INFO
 [org.ovirt.engine.core.bll.utils.ThreadPoolMonitoringService]
(EE-ManagedThreadFactory-engineThreadMonitoring-Thread-1) [] Thread pool
'hostUpdatesChecker' is using 0 threads out of 5, 2 threads waiting for
tasks.

The same is observed in the vdsm.log too, below is the log during the
attempted migration....

2019-07-10 09:39:57,034+0200 INFO  (jsonrpc/7) [jsonrpc.JsonRpcServer] RPC
call Host.getStats succeeded in 0.01 seconds (__init__:573)
2019-07-10 09:39:57,994+0200 INFO  (jsonrpc/2) [api.host] START getStats()
from=::ffff:10.0.1.1,57934 (api:46)
2019-07-10 09:39:57,994+0200 INFO  (jsonrpc/2) [vdsm.api] START
repoStats(domains=()) from=::ffff:10.0.1.1,57934,
task_id=e2529cfc-4293-42b4-91fa-7f5558e279dd (api:46)
2019-07-10 09:39:57,994+0200 INFO  (jsonrpc/2) [vdsm.api] FINISH repoStats
return={u'8a607f8a-542a-473c-bb18-25c05fe2a3d4': {'code': 0, 'actual':
True, 'version': 4, 'acquired': True, 'delay': '0.000194846', 'lastCheck':
'2.4', 'valid': True}, u'37b1a5d7-4e29-4763-9337-63c51dbc5fc8': {'code': 0,
'actual': True, 'version': 0, 'acquired': True, 'delay': '0.000277154',
'lastCheck': '6.0', 'valid': True},
u'2558679a-2214-466b-8f05-06fdda9146e5': {'code': 0, 'actual': True,
'version': 4, 'acquired': True, 'delay': '0.000421988', 'lastCheck': '2.4',
'valid': True}, u'640a5875-3d82-43c0-860f-7bb3e4a7e6f0': {'code': 0,
'actual': True, 'version': 4, 'acquired': True, 'delay': '0.000228443',
'lastCheck': '2.4', 'valid': True}} from=::ffff:10.0.1.1,57934,
task_id=e2529cfc-4293-42b4-91fa-7f5558e279dd (api:52)
2019-07-10 09:39:57,995+0200 INFO  (jsonrpc/2) [vdsm.api] START
multipath_health() from=::ffff:10.0.1.1,57934,
task_id=fd7ad703-5096-4f09-99fa-54672cb4aad9 (api:46)
2019-07-10 09:39:57,995+0200 INFO  (jsonrpc/2) [vdsm.api] FINISH
multipath_health return={} from=::ffff:10.0.1.1,57934,
task_id=fd7ad703-5096-4f09-99fa-54672cb4aad9 (api:52)
2019-07-10 09:39:58,002+0200 INFO  (jsonrpc/2) [api.host] FINISH getStats
return={'status': {'message': 'Done', 'code': 0}, 'info': {'cpuStatistics':
{'42': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.13', 'cpuIdle':
'99.87'}, '43': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.00',
'cpuIdle': '100.00'}, '24': {'cpuUser': '0.73', 'nodeIndex': 0, 'cpuSys':
'0.07', 'cpuIdle': '99.20'}, '25': {'cpuUser': '0.07', 'nodeIndex': 1,
'cpuSys': '0.00', 'cpuIdle': '99.93'}, '26': {'cpuUser': '5.59',
'nodeIndex': 0, 'cpuSys': '1.20', 'cpuIdle': '93.21'}, '27': {'cpuUser':
'0.87', 'nodeIndex': 1, 'cpuSys': '0.60', 'cpuIdle': '98.53'}, '20':
{'cpuUser': '0.53', 'nodeIndex': 0, 'cpuSys': '0.13', 'cpuIdle': '99.34'},
'21': {'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.07', 'cpuIdle':
'99.93'}, '22': {'cpuUser': '0.40', 'nodeIndex': 0, 'cpuSys': '0.20',
'cpuIdle': '99.40'}, '23': {'cpuUser': '0.07', 'nodeIndex': 1, 'cpuSys':
'0.07', 'cpuIdle': '99.86'}, '46': {'cpuUser': '0.13', 'nodeIndex': 0,
'cpuSys': '0.00', 'cpuIdle': '99.87'}, '47': {'cpuUser': '0.00',
'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '44': {'cpuUser':
'0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '45':
{'cpuUser': '0.00', 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'},
'28': {'cpuUser': '0.60', 'nodeIndex': 0, 'cpuSys': '0.07', 'cpuIdle':
'99.33'}, '29': {'cpuUser': '1.07', 'nodeIndex': 1, 'cpuSys': '0.20',
'cpuIdle': '98.73'}, '40': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys':
'0.00', 'cpuIdle': '100.00'}, '41': {'cpuUser': '0.00', 'nodeIndex': 1,
'cpuSys': '0.00', 'cpuIdle': '100.00'}, '1': {'cpuUser': '1.07',
'nodeIndex': 1, 'cpuSys': '1.13', 'cpuIdle': '97.80'}, '0': {'cpuUser':
'0.60', 'nodeIndex': 0, 'cpuSys': '0.20', 'cpuIdle': '99.20'}, '3':
{'cpuUser': '0.20', 'nodeIndex': 1, 'cpuSys': '0.07', 'cpuIdle': '99.73'},
'2': {'cpuUser': '3.00', 'nodeIndex': 0, 'cpuSys': '0.53', 'cpuIdle':
'96.47'}, '5': {'cpuUser': '0.20', 'nodeIndex': 1, 'cpuSys': '0.13',
'cpuIdle': '99.67'}, '4': {'cpuUser': '0.47', 'nodeIndex': 0, 'cpuSys':
'0.20', 'cpuIdle': '99.33'}, '7': {'cpuUser': '0.40', 'nodeIndex': 1,
'cpuSys': '0.20', 'cpuIdle': '99.40'}, '6': {'cpuUser': '0.67',
'nodeIndex': 0, 'cpuSys': '0.20', 'cpuIdle': '99.13'}, '9': {'cpuUser':
'0.47', 'nodeIndex': 1, 'cpuSys': '0.40', 'cpuIdle': '99.13'}, '8':
{'cpuUser': '0.13', 'nodeIndex': 0, 'cpuSys': '0.07', 'cpuIdle': '99.80'},
'39': {'cpuUser': '0.33', 'nodeIndex': 1, 'cpuSys': '0.13', 'cpuIdle':
'99.54'}, '38': {'cpuUser': '0.07', 'nodeIndex': 0, 'cpuSys': '0.00',
'cpuIdle': '99.93'}, '11': {'cpuUser': '0.67', 'nodeIndex': 1, 'cpuSys':
'0.27', 'cpuIdle': '99.06'}, '10': {'cpuUser': '0.13', 'nodeIndex': 0,
'cpuSys': '0.13', 'cpuIdle': '99.74'}, '13': {'cpuUser': '0.07',
'nodeIndex': 1, 'cpuSys': '0.07', 'cpuIdle': '99.86'}, '12': {'cpuUser':
'0.07', 'nodeIndex': 0, 'cpuSys': '0.27', 'cpuIdle': '99.66'}, '15':
{'cpuUser': '0.27', 'nodeIndex': 1, 'cpuSys': '0.13', 'cpuIdle': '99.60'},
'14': {'cpuUser': '0.27', 'nodeIndex': 0, 'cpuSys': '0.07', 'cpuIdle':
'99.66'}, '17': {'cpuUser': '0.07', 'nodeIndex': 1, 'cpuSys': '0.27',
'cpuIdle': '99.66'}, '16': {'cpuUser': '0.53', 'nodeIndex': 0, 'cpuSys':
'0.07', 'cpuIdle': '99.40'}, '19': {'cpuUser': '0.00', 'nodeIndex': 1,
'cpuSys': '0.00', 'cpuIdle': '100.00'}, '18': {'cpuUser': '1.00',
'nodeIndex': 0, 'cpuSys': '0.27', 'cpuIdle': '98.73'}, '31': {'cpuUser':
'0.00', 'nodeIndex': 1, 'cpuSys': '0.00', 'cpuIdle': '100.00'}, '30':
{'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'},
'37': {'cpuUser': '0.07', 'nodeIndex': 1, 'cpuSys': '0.07', 'cpuIdle':
'99.86'}, '36': {'cpuUser': '0.00', 'nodeIndex': 0, 'cpuSys': '0.00',
'cpuIdle': '100.00'}, '35': {'cpuUser': '0.20', 'nodeIndex': 1, 'cpuSys':
'0.33', 'cpuIdle': '99.47'}, '34': {'cpuUser': '0.00', 'nodeIndex': 0,
'cpuSys': '0.00', 'cpuIdle': '100.00'}, '33': {'cpuUser': '0.07',
'nodeIndex': 1, 'cpuSys': '0.13', 'cpuIdle': '99.80'}, '32': {'cpuUser':
'0.00', 'nodeIndex': 0, 'cpuSys': '0.00', 'cpuIdle': '100.00'}},
'numaNodeMemFree': {'1': {'memPercent': 5, 'memFree': '94165'}, '0':
{'memPercent': 22, 'memFree': '77122'}}, 'memShared': 0, 'haScore': 3400,
'thpState': 'always', 'ksmMergeAcrossNodes': True, 'vmCount': 2, 'memUsed':
'11', 'storageDomains': {u'8a607f8a-542a-473c-bb18-25c05fe2a3d4': {'code':
0, 'actual': True, 'version': 4, 'acquired': True, 'delay': '0.000194846',
'lastCheck': '2.4', 'valid': True},
u'37b1a5d7-4e29-4763-9337-63c51dbc5fc8': {'code': 0, 'actual': True,
'version': 0, 'acquired': True, 'delay': '0.000277154', 'lastCheck': '6.0',
'valid': True}, u'2558679a-2214-466b-8f05-06fdda9146e5': {'code': 0,
'actual': True, 'version': 4, 'acquired': True, 'delay': '0.000421988',
'lastCheck': '2.4', 'valid': True},
u'640a5875-3d82-43c0-860f-7bb3e4a7e6f0': {'code': 0, 'actual': True,
'version': 4, 'acquired': True, 'delay': '0.000228443', 'lastCheck': '2.4',
'valid': True}}, 'incomingVmMigrations': 0, 'network': {'em4': {'txErrors':
'0', 'state': 'up', 'sampleTime': 1562744396.40508, 'name': 'em4', 'tx':
'2160', 'txDropped': '0', 'rx': '261751836', 'rxErrors': '0', 'speed':
'1000', 'rxDropped': '1'}, 'ovirtmgmt': {'txErrors': '0', 'state': 'up',
'sampleTime': 1562744396.40508, 'name': 'ovirtmgmt', 'tx': '193005142',
'txDropped': '0', 'rx': '4300879104', 'rxErrors': '0', 'speed': '1000',
'rxDropped': '478'}, 'restores': {'txErrors': '0', 'state': 'up',
'sampleTime': 1562744396.40508, 'name': 'restores', 'tx': '1362',
'txDropped': '0', 'rx': '226442665', 'rxErrors': '0', 'speed': '1000',
'rxDropped': '478'}, 'em2': {'txErrors': '0', 'state': 'down',
'sampleTime': 1562744396.40508, 'name': 'em2', 'tx': '0', 'txDropped': '0',
'rx': '0', 'rxErrors': '0', 'speed': '1000', 'rxDropped': '0'}, 'vnet0':
{'txErrors': '0', 'state': 'up', 'sampleTime': 1562744396.40508, 'name':
'vnet0', 'tx': '2032610435', 'txDropped': '686', 'rx': '4287479548',
'rxErrors': '0', 'speed': '1000', 'rxDropped': '0'}, ';vdsmdummy;':
{'txErrors': '0', 'state': 'down', 'sampleTime': 1562744396.40508, 'name':
';vdsmdummy;', 'tx': '0', 'txDropped': '0', 'rx': '0', 'rxErrors': '0',
'speed': '1000', 'rxDropped': '0'}, 'em1': {'txErrors': '0', 'state': 'up',
'sampleTime': 1562744396.40508, 'name': 'em1', 'tx': '4548433238',
'txDropped': '0', 'rx': '6476729588', 'rxErrors': '0', 'speed': '1000',
'rxDropped': '1'}, 'em3': {'txErrors': '0', 'state': 'down', 'sampleTime':
1562744396.40508, 'name': 'em3', 'tx': '0', 'txDropped': '0', 'rx': '0',
'rxErrors': '0', 'speed': '1000', 'rxDropped': '0'}, 'lo': {'txErrors':
'0', 'state': 'up', 'sampleTime': 1562744396.40508, 'name': 'lo', 'tx':
'397962377', 'txDropped': '0', 'rx': '397962377', 'rxErrors': '0', 'speed':
'1000', 'rxDropped': '0'}, 'vnet1': {'txErrors': '0', 'state': 'up',
'sampleTime': 1562744396.40508, 'name': 'vnet1', 'tx': '526185708',
'txDropped': '0', 'rx': '118512222', 'rxErrors': '0', 'speed': '1000',
'rxDropped': '0'}}, 'txDropped': '686', 'anonHugePages': '18532',
'ksmPages': 100, 'elapsedTime': '85176.64', 'cpuLoad': '0.06', 'cpuSys':
'0.17', 'diskStats': {'/var/log': {'free': '6850'}, '/var/run/vdsm/':
{'free': '96410'}, '/tmp': {'free': '1825'}}, 'cpuUserVdsmd': '1.07',
'netConfigDirty': 'False', 'memCommitted': 24706, 'ksmState': False,
'vmMigrating': 0, 'ksmCpu': 0, 'memAvailable': 166010, 'bootTime':
'1562659184', 'haStats': {'active': True, 'configured': True, 'score':
3400, 'localMaintenance': False, 'globalMaintenance': False}, 'momStatus':
'active', 'multipathHealth': {}, 'rxDropped': '958',
'outgoingVmMigrations': 0, 'swapTotal': 4095, 'swapFree': 4095,
'hugepages': defaultdict(<type 'dict'>, {1048576: {'resv_hugepages': 0,
'free_hugepages': 0, 'nr_overcommit_hugepages': 0, 'surplus_hugepages': 0,
'vm.free_hugepages': 0, 'nr_hugepages': 0, 'nr_hugepages_mempolicy': 0},
2048: {'resv_hugepages': 0, 'free_hugepages': 0, 'nr_overcommit_hugepages':
0, 'surplus_hugepages': 0, 'vm.free_hugepages': 0, 'nr_hugepages': 0,
'nr_hugepages_mempolicy': 0}}), 'dateTime': '2019-07-10T07:39:57 GMT',
'cpuUser': '0.44', 'memFree': 172451, 'cpuIdle': '99.39', 'vmActive': 2,
'v2vJobs': {}, 'cpuSysVdsmd': '0.60'}} from=::ffff:10.0.1.1,57934 (api:52)
2019-07-10 09:39:58,004+0200 INFO  (jsonrpc/2) [jsonrpc.JsonRpcServer] RPC
call Host.getStats succeeded in 0.01 seconds (__init__:573)

Please let me know if you need further info.

Thank you.

Regards.

Neil Wilson


On Tue, Jul 9, 2019 at 5:52 PM Michal Skrivanek <michal.skriva...@redhat.com>
wrote:

> Can you share the engine.log please? And highlight the exact time when
> you attempt that migrate action
>
> Thanks,
> michal
>
> > On 9 Jul 2019, at 16:42, Neil <nwilson...@gmail.com> wrote:
> >
> > --000000000000166784058d409302
> > Content-Type: text/plain; charset="UTF-8"
> > Content-Transfer-Encoding: quoted-printable
> >
> > I remember seeing the bug earlier but because it was closed thought it
> was
> > unrelated, this appears to be it....
> >
> > https://bugzilla.redhat.com/show_bug.cgi?id=3D1670701
> >
> > Perhaps I'm not understanding your question about the VM guest agent,
> but I
> > don't have any guest agent currently installed on the VM, not sure if the
> > output of my qemu-kvm process maybe answers this question?....
> >
> > /usr/libexec/qemu-kvm -name
> guest=3DHeadoffice.cbl-ho.local,debug-threads=
> > =3Don
> > -S -object
> >
> secret,id=3DmasterKey0,format=3Draw,file=3D/var/lib/libvirt/qemu/domain-1-H=
> > eadoffice.cbl-ho.lo/master-key.aes
> > -machine pc-i440fx-rhel7.3.0,accel=3Dkvm,usb=3Doff,dump-guest-core=3Doff
> -c=
> > pu
> >
> Broadwell,vme=3Don,f16c=3Don,rdrand=3Don,hypervisor=3Don,arat=3Don,xsaveopt=
> > =3Don,abm=3Don,rtm=3Don,hle=3Don
> > -m 8192 -realtime mlock=3Doff -smp
> 8,maxcpus=3D64,sockets=3D16,cores=3D4,th=
> > reads=3D1
> > -numa node,nodeid=3D0,cpus=3D0-7,mem=3D8192 -uuid
> > 9a6561b8-5702-43dc-9e92-1dc5dfed4eef -smbios
> > type=3D1,manufacturer=3DoVirt,product=3DoVirt
> >
> Node,version=3D7-3.1611.el7.centos,serial=3D4C4C4544-0034-5810-8033-C2C04F4=
> > E4B32,uuid=3D9a6561b8-5702-43dc-9e92-1dc5dfed4eef
> > -no-user-config -nodefaults -chardev
> > socket,id=3Dcharmonitor,fd=3D31,server,nowait -mon
> > chardev=3Dcharmonitor,id=3Dmonitor,mode=3Dcontrol -rtc
> > base=3D2019-07-09T10:26:53,driftfix=3Dslew -global
> > kvm-pit.lost_tick_policy=3Ddelay -no-hpet -no-shutdown -boot strict=3Don
> > -device piix3-usb-uhci,id=3Dusb,bus=3Dpci.0,addr=3D0x1.0x2 -device
> > virtio-scsi-pci,id=3Dscsi0,bus=3Dpci.0,addr=3D0x4 -device
> >
> virtio-serial-pci,id=3Dvirtio-serial0,max_ports=3D16,bus=3Dpci.0,addr=3D0x5=
> > -drive
> > if=3Dnone,id=3Ddrive-ide0-1-0,readonly=3Don -device
> > ide-cd,bus=3Dide.1,unit=3D0,drive=3Ddrive-ide0-1-0,id=3Dide0-1-0 -drive
> >
> file=3D/rhev/data-center/59831b91-00a5-01e4-0294-000000000018/8a607f8a-542a=
> >
> -473c-bb18-25c05fe2a3d4/images/56e8240c-a172-4f52-b0c1-2bddc4f34f93/9f24546=
> >
> 7-d31d-4f5a-8037-7c5012a4aa84,format=3Dqcow2,if=3Dnone,id=3Ddrive-virtio-di=
> >
> sk0,serial=3D56e8240c-a172-4f52-b0c1-2bddc4f34f93,werror=3Dstop,rerror=3Dst=
> > op,cache=3Dnone,aio=3Dnative
> > -device
> >
> virtio-blk-pci,scsi=3Doff,bus=3Dpci.0,addr=3D0x7,drive=3Ddrive-virtio-disk0=
> > ,id=3Dvirtio-disk0,bootindex=3D1,write-cache=3Don
> > -netdev tap,fd=3D33,id=3Dhostnet0,vhost=3Don,vhostfd=3D34 -device
> >
> virtio-net-pci,netdev=3Dhostnet0,id=3Dnet0,mac=3D00:1a:4a:16:01:5b,bus=3Dpc=
> > i.0,addr=3D0x3
> > -chardev socket,id=3Dcharchannel0,fd=3D35,server,nowait -device
> >
> virtserialport,bus=3Dvirtio-serial0.0,nr=3D1,chardev=3Dcharchannel0,id=3Dch=
> > annel0,name=3Dcom.redhat.rhevm.vdsm
> > -chardev socket,id=3Dcharchannel1,fd=3D36,server,nowait -device
> >
> virtserialport,bus=3Dvirtio-serial0.0,nr=3D2,chardev=3Dcharchannel1,id=3Dch=
> > annel1,name=3Dorg.qemu.guest_agent.0
> > -chardev spicevmc,id=3Dcharchannel2,name=3Dvdagent -device
> >
> virtserialport,bus=3Dvirtio-serial0.0,nr=3D3,chardev=3Dcharchannel2,id=3Dch=
> > annel2,name=3Dcom.redhat.spice.0
> > -spice
> >
> tls-port=3D5900,addr=3D10.0.1.11,x509-dir=3D/etc/pki/vdsm/libvirt-spice,tls=
> >
> -channel=3Ddefault,tls-channel=3Dmain,tls-channel=3Ddisplay,tls-channel=3Di=
> >
> nputs,tls-channel=3Dcursor,tls-channel=3Dplayback,tls-channel=3Drecord,tls-=
> > channel=3Dsmartcard,tls-channel=3Dusbredir,seamless-migration=3Don
> > -device
> >
> qxl-vga,id=3Dvideo0,ram_size=3D67108864,vram_size=3D8388608,vram64_size_mb=
> > =3D0,vgamem_mb=3D16,max_outputs=3D1,bus=3Dpci.0,addr=3D0x2
> > -incoming defer -device
> virtio-balloon-pci,id=3Dballoon0,bus=3Dpci.0,addr=
> > =3D0x6
> > -object rng-random,id=3Dobjrng0,filename=3D/dev/urandom -device
> > virtio-rng-pci,rng=3Dobjrng0,id=3Drng0,bus=3Dpci.0,addr=3D0x8 -sandbox
> >
> on,obsolete=3Ddeny,elevateprivileges=3Ddeny,spawn=3Ddeny,resourcecontrol=3D=
> > deny
> > -msg timestamp=3Don
> >
> > Please shout if you need further info.
> >
> > Thanks.
> >
> >
> >
> >
> >
> >
> > On Tue, Jul 9, 2019 at 4:17 PM Strahil Nikolov <hunter86...@yahoo.com>
> > wrote:
> >
> >> Shouldn't cause that problem.
> >>
> >> You have to find the bug in bugzilla and report a regression (if it's
> not
> >> closed) , or open a new one and report the regression.
> >> As far as I remember , only the dashboard was affected due to new
> feature=
> > s
> >> about vdo disk savings.
> >>
> >> About the VM - this should be another issue. What agent are you using in
> >> the VMs (ovirt or qemu) ?
> >>
> >> Best Regards,
> >> Strahil Nikolov
> >>
> >> =D0=92 =D0=B2=D1=82=D0=BE=D1=80=D0=BD=D0=B8=D0=BA, 9 =D1=8E=D0=BB=D0=B8
> 2=
> > 019 =D0=B3., 10:09:05 =D1=87.
> =D0=93=D1=80=D0=B8=D0=BD=D1=83=D0=B8=D1=87-4,=
> > Neil <
> >> nwilson...@gmail.com> =D0=BD=D0=B0=D0=BF=D0=B8=D1=81=D0=B0:
> >>
> >>
> >> Hi Strahil,
> >>
> >> Thanks for the quick reply.
> >> I put the cluster into global maintenance, then installed the 4.3 repo,
> >> then "yum update ovirt\*setup\*"  then "engine-upgrade-check",
> >> "engine-setup", then "yum update", once completed, I rebooted the
> >> hosted-engine VM, and took the cluster out of global maintenance.
> >>
> >> Thinking back to the upgrade from 4.1 to 4.2 I don't recall doing a "yum
> >> update" after doing the engine-setup, not sure if this would cause it
> >> perhaps?
> >>
> >> Thank you.
> >> Regards.
> >> Neil Wilson.
> >>
> >> On Tue, Jul 9, 2019 at 3:47 PM Strahil Nikolov <hunter86...@yahoo.com>
> >> wrote:
> >>
> >> Hi Neil,
> >>
> >> for "Could not fetch data needed for VM migrate operation" - there was a
> >> bug and it was fixed.
> >> Are you sure you have fully updated ?
> >> What procedure did you use ?
> >>
> >> Best Regards,
> >> Strahil Nikolov
> >>
> >> =D0=92 =D0=B2=D1=82=D0=BE=D1=80=D0=BD=D0=B8=D0=BA, 9 =D1=8E=D0=BB=D0=B8
> 2=
> > 019 =D0=B3., 7:26:21 =D1=87.
> =D0=93=D1=80=D0=B8=D0=BD=D1=83=D0=B8=D1=87-4, =
> > Neil <nwilson...@gmail.com>
> >> =D0=BD=D0=B0=D0=BF=D0=B8=D1=81=D0=B0:
> >>
> >>
> >> Hi guys.
> >>
> >> I have two problems since upgrading from 4.2.x to 4.3.4
> >>
> >> First issue is I can no longer manually migrate VM's between hosts, I
> get
> >> an error in the ovirt GUI that says "Could not fetch data needed for VM
> >> migrate operation" and nothing gets logged either in my engine.log or my
> >> vdsm.log
> >>
> >> Then the other issue is my Dashboard says the following "Error! Could
> not
> >> fetch dashboard data. Please ensure that data warehouse is properly
> >> installed and configured."
> >>
> >> If I look at my ovirt-engine-dwhd.log I see the following if I try
> restar=
> > t
> >> the dwh service...
> >>
> >> 2019-07-09 11:48:04|ETL Service Started
> >> ovirtEngineDbDriverClass|org.postgresql.Driver
> >>
> >>
> ovirtEngineHistoryDbJdbcConnection|jdbc:postgresql://localhost:5432/ovirt=
> > _engine_history?sslfactory=3Dorg.postgresql.ssl.NonValidatingFactory
> >> hoursToKeepDaily|0
> >> hoursToKeepHourly|720
> >> ovirtEngineDbPassword|**********************
> >> runDeleteTime|3
> >>
> >>
> ovirtEngineDbJdbcConnection|jdbc:postgresql://localhost:5432/engine?sslfa=
> > ctory=3Dorg.postgresql.ssl.NonValidatingFactory
> >> runInterleave|60
> >> limitRows|limit 1000
> >> ovirtEngineHistoryDbUser|ovirt_engine_history
> >> ovirtEngineDbUser|engine
> >> deleteIncrement|10
> >> timeBetweenErrorEvents|300000
> >> hoursToKeepSamples|24
> >> deleteMultiplier|1000
> >> lastErrorSent|2011-07-03 12:46:47.000000
> >> etlVersion|4.3.0
> >> dwhAggregationDebug|false
> >> dwhUuid|dca0ebd3-c58f-4389-a1f8-6aecc20b1316
> >> ovirtEngineHistoryDbDriverClass|org.postgresql.Driver
> >> ovirtEngineHistoryDbPassword|**********************
> >> 2019-07-09 11:48:10|ETL Service Stopped
> >> 2019-07-09 11:49:59|ETL Service Started
> >> ovirtEngineDbDriverClass|org.postgresql.Driver
> >>
> >>
> ovirtEngineHistoryDbJdbcConnection|jdbc:postgresql://localhost:5432/ovirt=
> > _engine_history?sslfactory=3Dorg.postgresql.ssl.NonValidatingFactory
> >> hoursToKeepDaily|0
> >> hoursToKeepHourly|720
> >> ovirtEngineDbPassword|**********************
> >> runDeleteTime|3
> >>
> >>
> ovirtEngineDbJdbcConnection|jdbc:postgresql://localhost:5432/engine?sslfa=
> > ctory=3Dorg.postgresql.ssl.NonValidatingFactory
> >> runInterleave|60
> >> limitRows|limit 1000
> >> ovirtEngineHistoryDbUser|ovirt_engine_history
> >> ovirtEngineDbUser|engine
> >> deleteIncrement|10
> >> timeBetweenErrorEvents|300000
> >> hoursToKeepSamples|24
> >> deleteMultiplier|1000
> >> lastErrorSent|2011-07-03 12:46:47.000000
> >> etlVersion|4.3.0
> >> dwhAggregationDebug|false
> >> dwhUuid|dca0ebd3-c58f-4389-a1f8-6aecc20b1316
> >> ovirtEngineHistoryDbDriverClass|org.postgresql.Driver
> >> ovirtEngineHistoryDbPassword|**********************
> >> 2019-07-09 11:52:56|ETL Service Stopped
> >> 2019-07-09 11:52:57|ETL Service Started
> >> ovirtEngineDbDriverClass|org.postgresql.Driver
> >>
> >>
> ovirtEngineHistoryDbJdbcConnection|jdbc:postgresql://localhost:5432/ovirt=
> > _engine_history?sslfactory=3Dorg.postgresql.ssl.NonValidatingFactory
> >> hoursToKeepDaily|0
> >> hoursToKeepHourly|720
> >> ovirtEngineDbPassword|**********************
> >> runDeleteTime|3
> >>
> >>
> ovirtEngineDbJdbcConnection|jdbc:postgresql://localhost:5432/engine?sslfa=
> > ctory=3Dorg.postgresql.ssl.NonValidatingFactory
> >> runInterleave|60
> >> limitRows|limit 1000
> >> ovirtEngineHistoryDbUser|ovirt_engine_history
> >> ovirtEngineDbUser|engine
> >> deleteIncrement|10
> >> timeBetweenErrorEvents|300000
> >> hoursToKeepSamples|24
> >> deleteMultiplier|1000
> >> lastErrorSent|2011-07-03 12:46:47.000000
> >> etlVersion|4.3.0
> >> dwhAggregationDebug|false
> >> dwhUuid|dca0ebd3-c58f-4389-a1f8-6aecc20b1316
> >> ovirtEngineHistoryDbDriverClass|org.postgresql.Driver
> >> ovirtEngineHistoryDbPassword|**********************
> >> 2019-07-09 12:16:01|ETL Service Stopped
> >> 2019-07-09 12:16:45|ETL Service Started
> >> ovirtEngineDbDriverClass|org.postgresql.Driver
> >>
> >>
> ovirtEngineHistoryDbJdbcConnection|jdbc:postgresql://localhost:5432/ovirt=
> > _engine_history?sslfactory=3Dorg.postgresql.ssl.NonValidatingFactory
> >> hoursToKeepDaily|0
> >> hoursToKeepHourly|720
> >> ovirtEngineDbPassword|**********************
> >> runDeleteTime|3
> >>
> >>
> ovirtEngineDbJdbcConnection|jdbc:postgresql://localhost:5432/engine?sslfa=
> > ctory=3Dorg.postgresql.ssl.NonValidatingFactory
> >> runInterleave|60
> >> limitRows|limit 1000
> >> ovirtEngineHistoryDbUser|ovirt_engine_history
> >> ovirtEngineDbUser|engine
> >> deleteIncrement|10
> >> timeBetweenErrorEvents|300000
> >> hoursToKeepSamples|24
> >> deleteMultiplier|1000
> >> lastErrorSent|2011-07-03 12:46:47.000000
> >> etlVersion|4.3.0
> >> dwhAggregationDebug|false
> >> dwhUuid|dca0ebd3-c58f-4389-a1f8-6aecc20b1316
> >> ovirtEngineHistoryDbDriverClass|org.postgresql.Driver
> >> ovirtEngineHistoryDbPassword|**********************
> >>
> >>
> >>
> >>
> >>
> >> I have a hosted engine, and I have two hosts and my storage is FC based.
> >> The hosts are still running on 4.2 because I'm unable to migrate VM's
> off=
>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/JH4FMOY7NCRTHZ7ZGCGNICKAVSNTS3MH/

Reply via email to