Thanks Alona, Can you please update me once you have a fix? Thanks, Dafna
On Thu, Jun 28, 2018 at 10:28 AM, Alona Kaplan <[email protected]> wrote: > Hi, > I'm aware to the error. Francesco and me are working on it. > > Thank, > Alona. > > On Thu, Jun 28, 2018, 12:23 Dafna Ron <[email protected]> wrote: > >> ovirt-hosted-engine-ha failed on the same issue as well. >> >> On Thu, Jun 28, 2018 at 10:07 AM, Dafna Ron <[email protected]> wrote: >> >>> Hi, >>> >>> We had a failure in test 098_ovirt_provider_ovn.use_ovn_provider. >>> >>> Although CQ is pointing to this change: https://gerrit.ovirt.org/#/c/ >>> 92567/ - packaging: Add python-netaddr requirement I actually think >>> from the error its because of changes made to multiqueues >>> >>> https://gerrit.ovirt.org/#/c/92009/ - engine: Update libvirtVmXml to >>> consider vmBase.multiQueuesEnabled attribute >>> https://gerrit.ovirt.org/#/c/92008/ - engine: Introduce algorithm for >>> calculating how many queues asign per vnic >>> https://gerrit.ovirt.org/#/c/92007/ - engine: Add multiQueuesEnabled to >>> VmBase >>> https://gerrit.ovirt.org/#/c/92318/ - restapi: Add 'Multi Queues >>> Enabled' to the relevant mappers >>> https://gerrit.ovirt.org/#/c/92149/ - webadmin: Add 'Multi Queues >>> Enabled' to vm dialog >>> >>> Alona, can you please take a look? >>> >>> >>> *Link to Job:* >>> >>> >>> >>> >>> *http://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/8375/ >>> <http://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/8375/>Link >>> to all >>> logs:https://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/8375/artifact/exported-artifacts/basic-suit.el7.x86_64/test_logs/basic-suite-master/post-098_ovirt_provider_ovn.py/ >>> <https://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/8375/artifact/exported-artifacts/basic-suit.el7.x86_64/test_logs/basic-suite-master/post-098_ovirt_provider_ovn.py/>(Relevant) >>> error snippet from the log: <error>* >>> *eng**ine: * >>> >>> 2018-06-27 13:59:25,976-04 ERROR [org.ovirt.engine.core. >>> vdsbroker.vdsbroker.GetAllVmStatsVDSCommand] >>> (EE-ManagedThreadFactory-engineScheduled-Thread-80) >>> [] Command 'GetAllVmStatsVDSCommand(HostName = >>> lago-basic-suite-master-host-1, VdsIdVDSCommandParametersBase: >>> {hostId='d9094c95-3275-4616-b4c2-815e753bcfed'})' execution failed: >>> VDSGenericException: VDSNetworkException: Broken pipe >>> 2018-06-27 13:59:25,977-04 DEBUG >>> [org.ovirt.engine.core.utils.threadpool.ThreadPoolUtil] >>> (EE-ManagedThreadFactory-engine-Thread-442) [] Executing task: >>> EE-ManagedThreadFactory-engine-Thread-442 >>> 2018-06-27 13:59:25,977-04 DEBUG [org.ovirt.engine.core.common. >>> di.interceptor.DebugLoggingInterceptor] >>> (EE-ManagedThreadFactory-engine-Thread-442) >>> [] method: getVdsManager, params: [d9094c95-3275-4616-b4c2-815e753bcfed], >>> timeElapsed: 0ms >>> 2018-06-27 13:59:25,977-04 WARN >>> [org.ovirt.engine.core.vdsbroker.VdsManager] >>> (EE-ManagedThreadFactory-engine-Thread-442) [] Host >>> 'lago-basic-suite-master-host-1' is not responding. >>> 2018-06-27 13:59:25,979-04 ERROR [org.ovirt.engine.core.dal. >>> dbbroker.auditloghandling.AuditLogDirector] >>> (EE-ManagedThreadFactory-engineScheduled-Thread-63) >>> [] EVENT_ID: VDS_BROKER_COMMAND_FAILURE(10,802), VDSM >>> lago-basic-suite-master-host-1 command GetStatsAsyncVDS failed: Broken pipe >>> 2018-06-27 13:59:25,976-04 DEBUG [org.ovirt.engine.core. >>> vdsbroker.vdsbroker.GetAllVmStatsVDSCommand] >>> (EE-ManagedThreadFactory-engineScheduled-Thread-80) >>> [] Exception: org.ovirt.engine.core.vdsbroker.vdsbroker.VDSNetworkException: >>> VDSGenericException: VDSNetworkException: Broken pipe >>> at org.ovirt.engine.core.vdsbroker.vdsbroker.BrokerCommandBase. >>> proceedProxyReturnValue(BrokerCommandBase.java:189) [vdsbroker.jar:] >>> at org.ovirt.engine.core.vdsbroker.vdsbroker. >>> GetAllVmStatsVDSCommand.executeVdsBrokerCommand( >>> GetAllVmStatsVDSCommand.java:23) [vdsbroker.jar:] >>> at org.ovirt.engine.core.vdsbroker.vdsbroker.VdsBrokerCommand. >>> executeVdsCommandWithNetworkEvent(VdsBrokerCommand.java:123) >>> [vdsbroker.jar:] >>> at org.ovirt.engine.core.vdsbroker.vdsbroker.VdsBrokerCommand. >>> executeVDSCommand(VdsBrokerCommand.java:111) [vdsbroker.jar:] >>> at org.ovirt.engine.core.vdsbroker.VDSCommandBase. >>> executeCommand(VDSCommandBase.java:65) [vdsbroker.jar:] >>> at >>> org.ovirt.engine.core.dal.VdcCommandBase.execute(VdcCommandBase.java:31) >>> [dal.jar:] >>> at org.ovirt.engine.core.vdsbroker.vdsbroker. >>> DefaultVdsCommandExecutor.execute(DefaultVdsCommandExecutor.java:14) >>> [vdsbroker.jar:] >>> at org.ovirt.engine.core.vdsbroker.ResourceManager. >>> runVdsCommand(ResourceManager.java:399) [vdsbroker.jar:] >>> at org.ovirt.engine.core.vdsbroker.ResourceManager$ >>> Proxy$_$$_WeldSubclass.runVdsCommand$$super(Unknown Source) >>> [vdsbroker.jar:] >>> at sun.reflect.GeneratedMethodAccessor270.invoke(Unknown >>> Source) [:1.8.0_171] >>> at sun.reflect.DelegatingMethodAccessorImpl.invoke( >>> DelegatingMethodAccessorImpl.java:43) [rt.jar:1.8.0_171] >>> at java.lang.reflect.Method.invoke(Method.java:498) >>> [rt.jar:1.8.0_171] >>> at org.jboss.weld.interceptor.proxy. >>> TerminalAroundInvokeInvocationContext.proceedInternal( >>> TerminalAroundInvokeInvocationContext.java:49) >>> [weld-core-impl-2.4.3.Final.jar:2.4.3.Final] >>> at org.jboss.weld.interceptor.proxy. >>> AroundInvokeInvocationContext.proceed(AroundInvokeInvocationContext.java:77) >>> [weld-core-impl-2.4.3.Final.jar:2.4.3.Final] >>> at org.ovirt.engine.core.common.di.interceptor. >>> LoggingInterceptor.apply(LoggingInterceptor.java:12) [common.jar:] >>> at sun.reflect.GeneratedMethodAccessor68.invoke(Unknown Source) >>> [:1.8.0_171] >>> at sun.reflect.DelegatingMethodAccessorImpl.invoke( >>> DelegatingMethodAccessorImpl.java:43) [rt.jar:1.8.0_171] >>> at java.lang.reflect.Method.invoke(Method.java:498) >>> [rt.jar:1.8.0_171] >>> at org.jboss.weld.interceptor.reader. >>> SimpleInterceptorInvocation$SimpleMethodInvocation.invoke( >>> SimpleInterceptorInvocation.java:73) [weld-core-impl-2.4.3.Final. >>> jar:2.4.3.Final] >>> at org.jboss.weld.interceptor.proxy.InterceptorMethodHandler. >>> executeAroundInvoke(InterceptorMethodHandler.java:84) >>> [weld-core-impl-2.4.3.Final.jar:2.4.3.Final] >>> at org.jboss.weld.interceptor.proxy.InterceptorMethodHandler. >>> executeInterception(InterceptorMethodHandler.java:72) >>> [weld-core-impl-2.4.3.Final.jar:2.4.3.Final] >>> at org.jboss.weld.interceptor.proxy.InterceptorMethodHandler. >>> invoke(InterceptorMethodHandler.java:56) [weld-core-impl-2.4.3.Final. >>> jar:2.4.3.Final] >>> at org.jboss.weld.bean.proxy.CombinedInterceptorAndDecorato >>> rStackMethodHandler.invoke(CombinedInterceptorAndDecoratorStackMethodHandler.java:79) >>> [weld-core-impl-2.4.3.Final.jar:2.4.3.Final] >>> at org.jboss.weld.bean.proxy.CombinedInterceptorAndDecorato >>> rStackMethodHandler.invoke(CombinedInterceptorAndDecoratorStackMethodHandler.java:68) >>> [weld-core-impl-2.4.3.Final.jar:2.4.3.Final] >>> at org.ovirt.engine.core.vdsbroker.ResourceManager$ >>> Proxy$_$$_WeldSubclass.runVdsCommand(Unknown Source) [vdsbroker.jar:] >>> at org.ovirt.engine.core.vdsbroker.monitoring. >>> VmsStatisticsFetcher.poll(VmsStatisticsFetcher.java:29) [vdsbroker.jar:] >>> at org.ovirt.engine.core.vdsbroker.monitoring. >>> VmsListFetcher.fetch(VmsListFetcher.java:49) [vdsbroker.jar:] >>> at org.ovirt.engine.core.vdsbroker.monitoring. >>> PollVmStatsRefresher.poll(PollVmStatsRefresher.java:44) [vdsbroker.jar:] >>> at >>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) >>> [rt.jar:1.8.0_171] >>> at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308) >>> [rt.jar:1.8.0_171] >>> at org.glassfish.enterprise.concurrent.internal. >>> ManagedScheduledThreadPoolExecutor$ManagedScheduledFutureTask. >>> access$201(ManagedScheduledThreadPoolExecutor.java:383) >>> [javax.enterprise.concurrent-1.0.jar:] >>> at org.glassfish.enterprise.concurrent.internal. >>> ManagedScheduledThreadPoolExecutor$ManagedScheduledFutureTask.run( >>> ManagedScheduledThreadPoolExecutor.java:534) >>> [javax.enterprise.concurrent-1.0.jar:] >>> at >>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) >>> [rt.jar:1.8.0_171] >>> at >>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) >>> [rt.jar:1.8.0_171] >>> at java.lang.Thread.run(Thread.java:748) [rt.jar:1.8.0_171] >>> at org.glassfish.enterprise.concurrent.ManagedThreadFactoryImpl$ >>> ManagedThread.run(ManagedThreadFactoryImpl.java:250) >>> [javax.enterprise.concurrent-1.0.jar:] >>> at org.jboss.as.ee.concurrent.service. >>> ElytronManagedThreadFactory$ElytronManagedThread.run( >>> ElytronManagedThreadFactory.java:78) >>> >>> 2018-06-27 13:59:25,984-04 DEBUG [org.ovirt.engine.core. >>> vdsbroker.vdsbroker.GetAllVmStatsVDSCommand] >>> (EE-ManagedThreadFactory-engineScheduled-Thread-80) >>> [] FINISH, GetAllVmStatsVDSCommand, return: , log id: 56d99e77 >>> 2018-06-27 13:59:25,984-04 DEBUG [org.ovirt.engine.core.common. >>> di.interceptor.DebugLoggingInterceptor] >>> (EE-ManagedThreadFactory-engineScheduled-Thread-80) >>> [] method: runVdsCommand, params: [GetAllVmStats, >>> VdsIdVDSCommandParametersBase:{hostId='d9094c95-3275-4616-b4c2-815e753bcfed'}], >>> timeElapsed: 1497ms >>> 2018-06-27 13:59:25,984-04 INFO [org.ovirt.engine.core. >>> vdsbroker.monitoring.PollVmStatsRefresher] >>> (EE-ManagedThreadFactory-engineScheduled-Thread-80) >>> [] Failed to fetch vms info for host 'lago-basic-suite-master-host-1' - >>> skipping VMs monitoring. >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> >>> *vdsm: 2018-06-27 14:10:17,314-0400 INFO (jsonrpc/7) [virt.vm] >>> (vmId='b8a11304-07e3-4e64-af35-7421be780d5b') Hotunplug NIC xml: <?xml >>> version='1.0' encoding='utf-8'?><interface type="bridge"> <address >>> bus="0x00" domain="0x0000" function="0x0" slot="0x0b" type="pci" /> <mac >>> address="00:1a:4a:16:01:0e" /> <model type="virtio" /> <source >>> bridge="network_1" /> <link state="up" /> <driver name="vhost" >>> queues="" /> <alias name="ua-3c77476f-f194-476a-8412-d76a9e58d1f9" >>> /></interface> (vm:3321)2018-06-27 14:10:17,328-0400 ERROR (jsonrpc/7) >>> [virt.vm] (vmId='b8a11304-07e3-4e64-af35-7421be780d5b') Hotunplug failed >>> (vm:3353)Traceback (most recent call last): File >>> "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 3343, in >>> hotunplugNic self._dom.detachDevice(nicXml) File >>> "/usr/lib/python2.7/site-packages/vdsm/virt/virdomain.py", line 99, in f >>> ret = attr(*args, **kwargs) File >>> "/usr/lib/python2.7/site-packages/vdsm/common/libvirtconnection.py", line >>> 131, in wrapper ret = f(*args, **kwargs) File >>> "/usr/lib/python2.7/site-packages/vdsm/common/function.py", line 93, in >>> wrapper return func(inst, *args, **kwargs) File >>> "/usr/lib64/python2.7/site-packages/libvirt.py", line 1177, in >>> detachDevice if ret == -1: raise libvirtError ('virDomainDetachDevice() >>> failed', dom=self)libvirtError: 'queues' attribute must be positive number: >>> 2018-06-27 14:10:17,345-0400 DEBUG (jsonrpc/7) [api] FINISH hotunplugNic >>> response={'status': {'message': "'queues' attribute must be positive >>> number: ", 'code': 50}} (api:136)2018-06-27 14:10:17,346-0400 INFO >>> (jsonrpc/7) [api.virt] FINISH hotunplugNic return={'status': {'message': >>> "'queues' attribute must be positive number: ", 'code': 50}} >>> from=::ffff:192.168.201.4,32976, flow_id=ecb6652, >>> vmId=b8a11304-07e3-4e64-af35-7421be780d5b (api:53)2018-06-27 >>> 14:10:17,346-0400 INFO (jsonrpc/7) [jsonrpc.JsonRpcServer] RPC call >>> VM.hotunplugNic failed (error 50) in 0.07 seconds (__init__:311)2018-06-27 >>> 14:10:19,244-0400 DEBUG (qgapoller/2) [vds] Not sending QEMU-GA command >>> 'guest-get-users' to vm_id='b8a11304-07e3-4e64-af35-7421be780d5b', command >>> is not supported (qemuguestagent:192)2018-06-27 14:10:20,038-0400 DEBUG >>> (jsonrpc/1) [jsonrpc.JsonRpcServer] Calling 'Host.getAllVmStats' in bridge >>> with {} (__init__:328)2018-06-27 14:10:20,038-0400 INFO (jsonrpc/1) >>> [api.host] START getAllVmStats() from=::1,48032 (api:47)2018-06-27 >>> 14:10:20,041-0400 INFO (jsonrpc/1) [api.host] FINISH getAllVmStats >>> return={'status': {'message': 'Done', 'code': 0}, 'statsList': >>> (suppressed)} from=::1,48032 (api:53)2018-06-27 14:10:20,043-0400 DEBUG >>> (jsonrpc/1) [jsonrpc.JsonRpcServer] Return 'Host.getAllVmStats' in bridge >>> with (suppressed) (__init__:355)2018-06-27 14:10:20,043-0400 INFO >>> (jsonrpc/1) [jsonrpc.JsonRpcServer] RPC call Host.getAllVmStats succeeded >>> in 0.00 seconds (__init__:311)2018-06-27 14:10:20,057-0400 DEBUG >>> (jsonrpc/6) [jsonrpc.JsonRpcServer] Calling 'Host.getAllVmIoTunePolicies' >>> in bridge with {} (__init__:328)2018-06-27 14:10:20,058-0400 INFO >>> (jsonrpc/6) [api.host] START getAllVmIoTunePolicies() from=::1,48032 >>> (api:47)2018-06-27 14:10:20,058-0400 INFO (jsonrpc/6) [api.host] FINISH >>> getAllVmIoTunePolicies return={'status': {'message': 'Done', 'code': 0}, >>> 'io_tune_policies_dict': {'b8a11304-07e3-4e64-af35-7421be780d5b': >>> {'policy': [], 'current_values': [{'ioTune': {'write_bytes_sec': 0L, >>> 'total_iops_sec': 0L, 'read_iops_sec': 0L, 'read_bytes_sec': 0L, >>> 'write_iops_sec': 0L, 'total_bytes_sec': 0L}, 'path': >>> '/rhev/data-center/mnt/blockSD/cf23ceeb-81a3-4714-85a0-c6ddd1e024da/images/650fe4ae-47a1-4f2d-9cba-1617a8c868c3/03e75c3c-24e7-4e68-a6f1-21728aaaa73e', >>> 'name': 'vda'}]}}} from=::1,48032 (api:53)2018-06-27 14:10:20,059-0400 >>> DEBUG (jsonrpc/6) [jsonrpc.JsonRpcServer] Return >>> 'Host.getAllVmIoTunePolicies' in bridge with >>> {'b8a11304-07e3-4e64-af35-7421be780d5b': {'policy': [], 'current_values': >>> [{'ioTune': {'write_bytes_sec': 0L, 'total_iops_sec': 0L, 'read_iops_sec': >>> 0L, 'read_bytes_sec': 0L, 'write_iops_sec': 0L, 'total_bytes_sec': 0L}, >>> 'path': >>> '/rhev/data-center/mnt/blockSD/cf23ceeb-81a3-4714-85a0-c6ddd1e024da/images/650fe4ae-47a1-4f2d-9cba-1617a8c868c3/03e75c3c-24e7-4e68-a6f1-21728aaaa73e', >>> 'name': 'vda'}]}} (__init__:355)</error>* >>> >> >>
_______________________________________________ Devel mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/ZVMJORGRQ7XSYCF5LJ3N3J3DPR4PBLTA/
