On Sun, Jul 1, 2018 at 12:02 PM, Alona Kaplan <[email protected]> wrote:

> Hi,
>
> Merged an engine side workaround - https://gerrit.ovirt.org/#/c/92656/
>
> It passes system-tests basic and network suits
> https://jenkins.ovirt.org/job/ovirt-system-tests_manual/2872/ ,
> https://jenkins.ovirt.org/job/ovirt-system-tests_manual/2873/
>
> There is also a vdsm side fix - https://gerrit.ovirt.org/#/c/92634/
> (although the engine fix is enough to get the ost to pass).
> I'm not sure it passes OST without the engine fix, verifying it now.
>

It passes OST https://jenkins.ovirt.org/job/ovirt-system-tests_manual/2875


> Thanks,
> Alona.
>
> On Sun, Jul 1, 2018 at 11:21 AM, Ehud Yonasi <[email protected]> wrote:
>
>> Hey Alona,
>> what is the current status of the fix?
>>
>> On Thu, Jun 28, 2018 at 5:57 PM Dafna Ron <[email protected]> wrote:
>>
>>> vdsm also failing on the same issue:
>>>
>>> https://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/8407/
>>>
>>> Thanks,
>>> Dafna
>>>
>>>
>>> On Thu, Jun 28, 2018 at 11:11 AM, Dafna Ron <[email protected]> wrote:
>>>
>>>> Thanks Alona,
>>>> Can you please update me once you have a fix?
>>>>
>>>> Thanks,
>>>> Dafna
>>>>
>>>>
>>>> On Thu, Jun 28, 2018 at 10:28 AM, Alona Kaplan <[email protected]>
>>>> wrote:
>>>>
>>>>> Hi,
>>>>> I'm aware to the error. Francesco and me are working on it.
>>>>>
>>>>> Thank,
>>>>> Alona.
>>>>>
>>>>> On Thu, Jun 28, 2018, 12:23 Dafna Ron <[email protected]> wrote:
>>>>>
>>>>>> ovirt-hosted-engine-ha failed on the same issue as well.
>>>>>>
>>>>>> On Thu, Jun 28, 2018 at 10:07 AM, Dafna Ron <[email protected]> wrote:
>>>>>>
>>>>>>> Hi,
>>>>>>>
>>>>>>> We had a failure in test 098_ovirt_provider_ovn.use_ovn_provider.
>>>>>>>
>>>>>>> Although CQ is pointing to this change:
>>>>>>> https://gerrit.ovirt.org/#/c/92567/ - packaging: Add python-netaddr
>>>>>>> requirement  I actually think from the error its because of changes 
>>>>>>> made to
>>>>>>> multiqueues
>>>>>>>
>>>>>>> https://gerrit.ovirt.org/#/c/92009/ - engine: Update libvirtVmXml
>>>>>>> to consider vmBase.multiQueuesEnabled attribute
>>>>>>> https://gerrit.ovirt.org/#/c/92008/ - engine: Introduce algorithm
>>>>>>> for calculating how many queues asign per vnic
>>>>>>> https://gerrit.ovirt.org/#/c/92007/ - engine: Add
>>>>>>> multiQueuesEnabled to VmBase
>>>>>>> https://gerrit.ovirt.org/#/c/92318/ - restapi: Add 'Multi Queues
>>>>>>> Enabled' to the relevant mappers
>>>>>>> https://gerrit.ovirt.org/#/c/92149/ - webadmin: Add 'Multi Queues
>>>>>>> Enabled' to vm dialog
>>>>>>>
>>>>>>> Alona, can you please take a look?
>>>>>>>
>>>>>>>
>>>>>>> *Link to Job:*
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> *http://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/8375/
>>>>>>> <http://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/8375/>Link
>>>>>>> to all
>>>>>>> logs:https://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/8375/artifact/exported-artifacts/basic-suit.el7.x86_64/test_logs/basic-suite-master/post-098_ovirt_provider_ovn.py/
>>>>>>> <https://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/8375/artifact/exported-artifacts/basic-suit.el7.x86_64/test_logs/basic-suite-master/post-098_ovirt_provider_ovn.py/>(Relevant)
>>>>>>> error snippet from the log: <error>*
>>>>>>> *eng**ine: *
>>>>>>>
>>>>>>> 2018-06-27 13:59:25,976-04 ERROR [org.ovirt.engine.core.vdsbrok
>>>>>>> er.vdsbroker.GetAllVmStatsVDSCommand] 
>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-80)
>>>>>>> [] Command 'GetAllVmStatsVDSCommand(HostName =
>>>>>>> lago-basic-suite-master-host-1, VdsIdVDSCommandParametersBase:
>>>>>>> {hostId='d9094c95-3275-4616-b4c2-815e753bcfed'})' execution failed:
>>>>>>> VDSGenericException: VDSNetworkException: Broken pipe
>>>>>>> 2018-06-27 13:59:25,977-04 DEBUG 
>>>>>>> [org.ovirt.engine.core.utils.threadpool.ThreadPoolUtil]
>>>>>>> (EE-ManagedThreadFactory-engine-Thread-442) [] Executing task:
>>>>>>> EE-ManagedThreadFactory-engine-Thread-442
>>>>>>> 2018-06-27 13:59:25,977-04 DEBUG [org.ovirt.engine.core.common.
>>>>>>> di.interceptor.DebugLoggingInterceptor]
>>>>>>> (EE-ManagedThreadFactory-engine-Thread-442) [] method:
>>>>>>> getVdsManager, params: [d9094c95-3275-4616-b4c2-815e753bcfed],
>>>>>>> timeElapsed: 0ms
>>>>>>> 2018-06-27 13:59:25,977-04 WARN  
>>>>>>> [org.ovirt.engine.core.vdsbroker.VdsManager]
>>>>>>> (EE-ManagedThreadFactory-engine-Thread-442) [] Host
>>>>>>> 'lago-basic-suite-master-host-1' is not responding.
>>>>>>> 2018-06-27 13:59:25,979-04 ERROR [org.ovirt.engine.core.dal.dbb
>>>>>>> roker.auditloghandling.AuditLogDirector]
>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-63) [] EVENT_ID:
>>>>>>> VDS_BROKER_COMMAND_FAILURE(10,802), VDSM
>>>>>>> lago-basic-suite-master-host-1 command GetStatsAsyncVDS failed: Broken 
>>>>>>> pipe
>>>>>>> 2018-06-27 13:59:25,976-04 DEBUG [org.ovirt.engine.core.vdsbrok
>>>>>>> er.vdsbroker.GetAllVmStatsVDSCommand] 
>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-80)
>>>>>>> [] Exception: org.ovirt.engine.core.vdsbroke
>>>>>>> r.vdsbroker.VDSNetworkException: VDSGenericException:
>>>>>>> VDSNetworkException: Broken pipe
>>>>>>>         at org.ovirt.engine.core.vdsbroke
>>>>>>> r.vdsbroker.BrokerCommandBase.proceedProxyReturnValue(BrokerCommandBase.java:189)
>>>>>>> [vdsbroker.jar:]
>>>>>>>         at org.ovirt.engine.core.vdsbroke
>>>>>>> r.vdsbroker.GetAllVmStatsVDSCommand.executeVdsBrokerCommand(
>>>>>>> GetAllVmStatsVDSCommand.java:23) [vdsbroker.jar:]
>>>>>>>         at org.ovirt.engine.core.vdsbroke
>>>>>>> r.vdsbroker.VdsBrokerCommand.executeVdsCommandWithNetworkEvent(VdsBrokerCommand.java:123)
>>>>>>> [vdsbroker.jar:]
>>>>>>>         at org.ovirt.engine.core.vdsbroke
>>>>>>> r.vdsbroker.VdsBrokerCommand.executeVDSCommand(VdsBrokerCommand.java:111)
>>>>>>> [vdsbroker.jar:]
>>>>>>>         at org.ovirt.engine.core.vdsbroke
>>>>>>> r.VDSCommandBase.executeCommand(VDSCommandBase.java:65)
>>>>>>> [vdsbroker.jar:]
>>>>>>>         at org.ovirt.engine.core.dal.VdcC
>>>>>>> ommandBase.execute(VdcCommandBase.java:31) [dal.jar:]
>>>>>>>         at org.ovirt.engine.core.vdsbroke
>>>>>>> r.vdsbroker.DefaultVdsCommandExecutor.execute(DefaultVdsCommandExecutor.java:14)
>>>>>>> [vdsbroker.jar:]
>>>>>>>         at org.ovirt.engine.core.vdsbroke
>>>>>>> r.ResourceManager.runVdsCommand(ResourceManager.java:399)
>>>>>>> [vdsbroker.jar:]
>>>>>>>         at org.ovirt.engine.core.vdsbroke
>>>>>>> r.ResourceManager$Proxy$_$$_WeldSubclass.runVdsCommand$$super(Unknown
>>>>>>> Source) [vdsbroker.jar:]
>>>>>>>         at sun.reflect.GeneratedMethodAccessor270.invoke(Unknown
>>>>>>> Source) [:1.8.0_171]
>>>>>>>         at sun.reflect.DelegatingMethodAc
>>>>>>> cessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>>>>> [rt.jar:1.8.0_171]
>>>>>>>         at java.lang.reflect.Method.invoke(Method.java:498)
>>>>>>> [rt.jar:1.8.0_171]
>>>>>>>         at org.jboss.weld.interceptor.pro
>>>>>>> xy.TerminalAroundInvokeInvocationContext.proceedInternal(Ter
>>>>>>> minalAroundInvokeInvocationContext.java:49)
>>>>>>> [weld-core-impl-2.4.3.Final.jar:2.4.3.Final]
>>>>>>>         at org.jboss.weld.interceptor.pro
>>>>>>> xy.AroundInvokeInvocationContext.proceed(AroundInvokeInvocationContext.java:77)
>>>>>>> [weld-core-impl-2.4.3.Final.jar:2.4.3.Final]
>>>>>>>         at org.ovirt.engine.core.common.d
>>>>>>> i.interceptor.LoggingInterceptor.apply(LoggingInterceptor.java:12)
>>>>>>> [common.jar:]
>>>>>>>         at sun.reflect.GeneratedMethodAccessor68.invoke(Unknown
>>>>>>> Source) [:1.8.0_171]
>>>>>>>         at sun.reflect.DelegatingMethodAc
>>>>>>> cessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>>>>> [rt.jar:1.8.0_171]
>>>>>>>         at java.lang.reflect.Method.invoke(Method.java:498)
>>>>>>> [rt.jar:1.8.0_171]
>>>>>>>         at org.jboss.weld.interceptor.rea
>>>>>>> der.SimpleInterceptorInvocation$SimpleMethodInvocation.
>>>>>>> invoke(SimpleInterceptorInvocation.java:73)
>>>>>>> [weld-core-impl-2.4.3.Final.jar:2.4.3.Final]
>>>>>>>         at org.jboss.weld.interceptor.pro
>>>>>>> xy.InterceptorMethodHandler.executeAroundInvoke(InterceptorMethodHandler.java:84)
>>>>>>> [weld-core-impl-2.4.3.Final.jar:2.4.3.Final]
>>>>>>>         at org.jboss.weld.interceptor.pro
>>>>>>> xy.InterceptorMethodHandler.executeInterception(InterceptorMethodHandler.java:72)
>>>>>>> [weld-core-impl-2.4.3.Final.jar:2.4.3.Final]
>>>>>>>         at org.jboss.weld.interceptor.pro
>>>>>>> xy.InterceptorMethodHandler.invoke(InterceptorMethodHandler.java:56)
>>>>>>> [weld-core-impl-2.4.3.Final.jar:2.4.3.Final]
>>>>>>>         at org.jboss.weld.bean.proxy.Comb
>>>>>>> inedInterceptorAndDecoratorStackMethodHandler.invoke(Combine
>>>>>>> dInterceptorAndDecoratorStackMethodHandler.java:79)
>>>>>>> [weld-core-impl-2.4.3.Final.jar:2.4.3.Final]
>>>>>>>         at org.jboss.weld.bean.proxy.Comb
>>>>>>> inedInterceptorAndDecoratorStackMethodHandler.invoke(Combine
>>>>>>> dInterceptorAndDecoratorStackMethodHandler.java:68)
>>>>>>> [weld-core-impl-2.4.3.Final.jar:2.4.3.Final]
>>>>>>>         at org.ovirt.engine.core.vdsbroke
>>>>>>> r.ResourceManager$Proxy$_$$_WeldSubclass.runVdsCommand(Unknown
>>>>>>> Source) [vdsbroker.jar:]
>>>>>>>         at org.ovirt.engine.core.vdsbroke
>>>>>>> r.monitoring.VmsStatisticsFetcher.poll(VmsStatisticsFetcher.java:29)
>>>>>>> [vdsbroker.jar:]
>>>>>>>         at org.ovirt.engine.core.vdsbroke
>>>>>>> r.monitoring.VmsListFetcher.fetch(VmsListFetcher.java:49)
>>>>>>> [vdsbroker.jar:]
>>>>>>>         at org.ovirt.engine.core.vdsbroke
>>>>>>> r.monitoring.PollVmStatsRefresher.poll(PollVmStatsRefresher.java:44)
>>>>>>> [vdsbroker.jar:]
>>>>>>>         at 
>>>>>>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>>>>>>> [rt.jar:1.8.0_171]
>>>>>>>         at java.util.concurrent.FutureTas
>>>>>>> k.runAndReset(FutureTask.java:308) [rt.jar:1.8.0_171]
>>>>>>>         at org.glassfish.enterprise.concu
>>>>>>> rrent.internal.ManagedScheduledThreadPoolExecutor$ManagedSch
>>>>>>> eduledFutureTask.access$201(ManagedScheduledThreadPoolExecutor.java:383)
>>>>>>> [javax.enterprise.concurrent-1.0.jar:]
>>>>>>>         at org.glassfish.enterprise.concu
>>>>>>> rrent.internal.ManagedScheduledThreadPoolExecutor$ManagedSch
>>>>>>> eduledFutureTask.run(ManagedScheduledThreadPoolExecutor.java:534)
>>>>>>> [javax.enterprise.concurrent-1.0.jar:]
>>>>>>>         at java.util.concurrent.ThreadPoo
>>>>>>> lExecutor.runWorker(ThreadPoolExecutor.java:1149) [rt.jar:1.8.0_171]
>>>>>>>         at java.util.concurrent.ThreadPoo
>>>>>>> lExecutor$Worker.run(ThreadPoolExecutor.java:624) [rt.jar:1.8.0_171]
>>>>>>>         at java.lang.Thread.run(Thread.java:748) [rt.jar:1.8.0_171]
>>>>>>>         at org.glassfish.enterprise.concu
>>>>>>> rrent.ManagedThreadFactoryImpl$ManagedThread.run(ManagedThreadFactoryImpl.java:250)
>>>>>>> [javax.enterprise.concurrent-1.0.jar:]
>>>>>>>         at org.jboss.as.ee.concurrent.ser
>>>>>>> vice.ElytronManagedThreadFactory$ElytronManagedThread.run(El
>>>>>>> ytronManagedThreadFactory.java:78)
>>>>>>>
>>>>>>> 2018-06-27 13:59:25,984-04 DEBUG [org.ovirt.engine.core.vdsbrok
>>>>>>> er.vdsbroker.GetAllVmStatsVDSCommand] 
>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-80)
>>>>>>> [] FINISH, GetAllVmStatsVDSCommand, return: , log id: 56d99e77
>>>>>>> 2018-06-27 13:59:25,984-04 DEBUG [org.ovirt.engine.core.common.
>>>>>>> di.interceptor.DebugLoggingInterceptor]
>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-80) [] method:
>>>>>>> runVdsCommand, params: [GetAllVmStats, VdsIdVDSCommandParametersBase:
>>>>>>> {hostId='d9094c95-3275-4616-b4c2-815e753bcfed'}], timeElapsed:
>>>>>>> 1497ms
>>>>>>> 2018-06-27 13:59:25,984-04 INFO  [org.ovirt.engine.core.vdsbrok
>>>>>>> er.monitoring.PollVmStatsRefresher] 
>>>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-80)
>>>>>>> [] Failed to fetch vms info for host 'lago-basic-suite-master-host-1'
>>>>>>> - skipping VMs monitoring.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> *vdsm: 2018-06-27 14:10:17,314-0400 INFO  (jsonrpc/7) [virt.vm]
>>>>>>> (vmId='b8a11304-07e3-4e64-af35-7421be780d5b') Hotunplug NIC xml: <?xml
>>>>>>> version='1.0' encoding='utf-8'?><interface type="bridge">    <address
>>>>>>> bus="0x00" domain="0x0000" function="0x0" slot="0x0b" type="pci" />    
>>>>>>> <mac
>>>>>>> address="00:1a:4a:16:01:0e" />    <model type="virtio" />    <source
>>>>>>> bridge="network_1" />    <link state="up" />    <driver name="vhost"
>>>>>>> queues="" />    <alias name="ua-3c77476f-f194-476a-8412-d76a9e58d1f9"
>>>>>>> /></interface> (vm:3321)2018-06-27 14:10:17,328-0400 ERROR (jsonrpc/7)
>>>>>>> [virt.vm] (vmId='b8a11304-07e3-4e64-af35-7421be780d5b') Hotunplug failed
>>>>>>> (vm:3353)Traceback (most recent call last):  File
>>>>>>> "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line 3343, in
>>>>>>> hotunplugNic    self._dom.detachDevice(nicXml)  File
>>>>>>> "/usr/lib/python2.7/site-packages/vdsm/virt/virdomain.py", line 99, in f
>>>>>>> ret = attr(*args, **kwargs)  File
>>>>>>> "/usr/lib/python2.7/site-packages/vdsm/common/libvirtconnection.py", 
>>>>>>> line
>>>>>>> 131, in wrapper    ret = f(*args, **kwargs)  File
>>>>>>> "/usr/lib/python2.7/site-packages/vdsm/common/function.py", line 93, in
>>>>>>> wrapper    return func(inst, *args, **kwargs)  File
>>>>>>> "/usr/lib64/python2.7/site-packages/libvirt.py", line 1177, in
>>>>>>> detachDevice    if ret == -1: raise libvirtError 
>>>>>>> ('virDomainDetachDevice()
>>>>>>> failed', dom=self)libvirtError: 'queues' attribute must be positive 
>>>>>>> number:
>>>>>>> 2018-06-27 14:10:17,345-0400 DEBUG (jsonrpc/7) [api] FINISH hotunplugNic
>>>>>>> response={'status': {'message': "'queues' attribute must be positive
>>>>>>> number: ", 'code': 50}} (api:136)2018-06-27 14:10:17,346-0400 INFO
>>>>>>> (jsonrpc/7) [api.virt] FINISH hotunplugNic return={'status': {'message':
>>>>>>> "'queues' attribute must be positive number: ", 'code': 50}}
>>>>>>> from=::ffff:192.168.201.4,32976, flow_id=ecb6652,
>>>>>>> vmId=b8a11304-07e3-4e64-af35-7421be780d5b (api:53)2018-06-27
>>>>>>> 14:10:17,346-0400 INFO  (jsonrpc/7) [jsonrpc.JsonRpcServer] RPC call
>>>>>>> VM.hotunplugNic failed (error 50) in 0.07 seconds 
>>>>>>> (__init__:311)2018-06-27
>>>>>>> 14:10:19,244-0400 DEBUG (qgapoller/2) [vds] Not sending QEMU-GA command
>>>>>>> 'guest-get-users' to vm_id='b8a11304-07e3-4e64-af35-7421be780d5b', 
>>>>>>> command
>>>>>>> is not supported (qemuguestagent:192)2018-06-27 14:10:20,038-0400 DEBUG
>>>>>>> (jsonrpc/1) [jsonrpc.JsonRpcServer] Calling 'Host.getAllVmStats' in 
>>>>>>> bridge
>>>>>>> with {} (__init__:328)2018-06-27 14:10:20,038-0400 INFO  (jsonrpc/1)
>>>>>>> [api.host] START getAllVmStats() from=::1,48032 (api:47)2018-06-27
>>>>>>> 14:10:20,041-0400 INFO  (jsonrpc/1) [api.host] FINISH getAllVmStats
>>>>>>> return={'status': {'message': 'Done', 'code': 0}, 'statsList':
>>>>>>> (suppressed)} from=::1,48032 (api:53)2018-06-27 14:10:20,043-0400 DEBUG
>>>>>>> (jsonrpc/1) [jsonrpc.JsonRpcServer] Return 'Host.getAllVmStats' in 
>>>>>>> bridge
>>>>>>> with (suppressed) (__init__:355)2018-06-27 14:10:20,043-0400 INFO
>>>>>>> (jsonrpc/1) [jsonrpc.JsonRpcServer] RPC call Host.getAllVmStats 
>>>>>>> succeeded
>>>>>>> in 0.00 seconds (__init__:311)2018-06-27 14:10:20,057-0400 DEBUG
>>>>>>> (jsonrpc/6) [jsonrpc.JsonRpcServer] Calling 
>>>>>>> 'Host.getAllVmIoTunePolicies'
>>>>>>> in bridge with {} (__init__:328)2018-06-27 14:10:20,058-0400 INFO
>>>>>>> (jsonrpc/6) [api.host] START getAllVmIoTunePolicies() from=::1,48032
>>>>>>> (api:47)2018-06-27 14:10:20,058-0400 INFO  (jsonrpc/6) [api.host] FINISH
>>>>>>> getAllVmIoTunePolicies return={'status': {'message': 'Done', 'code': 0},
>>>>>>> 'io_tune_policies_dict': {'b8a11304-07e3-4e64-af35-7421be780d5b':
>>>>>>> {'policy': [], 'current_values': [{'ioTune': {'write_bytes_sec': 0L,
>>>>>>> 'total_iops_sec': 0L, 'read_iops_sec': 0L, 'read_bytes_sec': 0L,
>>>>>>> 'write_iops_sec': 0L, 'total_bytes_sec': 0L}, 'path':
>>>>>>> '/rhev/data-center/mnt/blockSD/cf23ceeb-81a3-4714-85a0-c6ddd1e024da/images/650fe4ae-47a1-4f2d-9cba-1617a8c868c3/03e75c3c-24e7-4e68-a6f1-21728aaaa73e',
>>>>>>> 'name': 'vda'}]}}} from=::1,48032 (api:53)2018-06-27 14:10:20,059-0400
>>>>>>> DEBUG (jsonrpc/6) [jsonrpc.JsonRpcServer] Return
>>>>>>> 'Host.getAllVmIoTunePolicies' in bridge with
>>>>>>> {'b8a11304-07e3-4e64-af35-7421be780d5b': {'policy': [], 
>>>>>>> 'current_values':
>>>>>>> [{'ioTune': {'write_bytes_sec': 0L, 'total_iops_sec': 0L, 
>>>>>>> 'read_iops_sec':
>>>>>>> 0L, 'read_bytes_sec': 0L, 'write_iops_sec': 0L, 'total_bytes_sec': 0L},
>>>>>>> 'path':
>>>>>>> '/rhev/data-center/mnt/blockSD/cf23ceeb-81a3-4714-85a0-c6ddd1e024da/images/650fe4ae-47a1-4f2d-9cba-1617a8c868c3/03e75c3c-24e7-4e68-a6f1-21728aaaa73e',
>>>>>>> 'name': 'vda'}]}} (__init__:355)</error>*
>>>>>>>
>>>>>>
>>>>>>
>>>>
>>> _______________________________________________
>>> Infra mailing list -- [email protected]
>>> To unsubscribe send an email to [email protected]
>>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>>> oVirt Code of Conduct: https://www.ovirt.org/communit
>>> y/about/community-guidelines/
>>> List Archives: https://lists.ovirt.org/archiv
>>> es/list/[email protected]/message/52RUJGRYJVGAYTEVHD2PUCVINQRHE5QQ/
>>>
>>
>
_______________________________________________
Devel mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/6RYLEDWJ653ABGUZLEM2AQTXQPTMNPSQ/

Reply via email to