Thanks for reporting this. https://gerrit.ovirt.org/91375 fixes this. I've
re-opened bug https://bugzilla.redhat.com/show_bug.cgi?id=1574508

On Thu, May 17, 2018 at 10:12 PM, Demeter Tibor <tdeme...@itsmart.hu> wrote:

> Hi,
>
> 4.2.4-0.0.master.20180515183442.git00e1340.el7.centos
>
> Firstly, I did a yum update "ovirt-*-setup*"
> second, I have ran engine-setup to upgrade.
>
> I didn't remove the old repos, just installed the nightly repo.
>
> Thank you again,
>
> Regards,
>
> Tibor
>
> ----- 2018. máj.. 17., 15:02, Sahina Bose <sab...@redhat.com> írta:
>
> It doesn't look like the patch was applied. Still see the same error in
> engine.log
> "Error while refreshing brick statuses for volume 'volume1' of cluster
> 'C6220': null"\
>
> Did you use engine-setup to upgrade? What's the version of ovirt-engine
> currently installed?
>
> On Thu, May 17, 2018 at 5:10 PM, Demeter Tibor <tdeme...@itsmart.hu>
> wrote:
>
>> Hi,
>>
>> sure,
>>
>> Thank you for your time!
>>
>> R
>> Tibor
>>
>> ----- 2018. máj.. 17., 12:19, Sahina Bose <sab...@redhat.com> írta:
>>
>> [+users]
>>
>> Can you provide the engine.log to see why the monitoring is not working
>> here. thanks!
>>
>> On Wed, May 16, 2018 at 2:08 PM, Demeter Tibor <tdeme...@itsmart.hu>
>> wrote:
>>
>>> Hi,
>>>
>>> Meanwhile, I did the upgrade engine, but the gluster state is same on my
>>> first node.
>>> I've attached some screenshot of my problem.
>>>
>>> Thanks
>>>
>>> Tibor
>>>
>>>
>>>
>>> ----- 2018. máj.. 16., 10:16, Demeter Tibor <tdeme...@itsmart.hu> írta
>>> Hi,
>>>
>>>
>>> If 4.3.4 will release, i just have to remove the nightly repo and update
>>> to stable?
>>>
>>> I'm sorry for my terrible English, I try to explain what was my problem
>>> with update.
>>> I'm upgraded from 4.1.8.
>>>
>>> I followed up the official hosted-engine update documentation, that was
>>> not clear me, because it has referenced to a lot of old thing (i think).
>>> https://www.ovirt.org/documentation/upgrade-guide/upgrade-guide/
>>> https://www.ovirt.org/documentation/how-to/hosted-
>>> engine/#upgrade-hosted-engine
>>>
>>> Maybe it need to update, because I had a lot of question under upgrade
>>> and I was not sure in all of necessary steps. For example, If I need to
>>> installing the new, 4.2 repo on the hosts, then need to remove the old repo
>>> from that?
>>> Why I need to do a" yum update -y" on hosts, meanwhile there is an
>>> "Updatehost" menu in the GUI? So, maybe it outdated.
>>> Since upgrade hosted engine, and the first node, I have problems with
>>> gluster. It seems to working fine if you check it from console "gluster
>>> volume status, etc" but not on the Gui, because now it yellow, and the
>>> brick reds in the first node.
>>>
>>> Previously I did a mistake with glusterfs, my gluster config was wrong.
>>> I have corrected them, but it did not helped to me,gluster bricks are reds
>>> on my first node yet....
>>>
>>>
>>> Now I try to upgrade to nightly, but I'm affraid, because it a living,
>>> productive system, and I don't have downtime. I hope it will help me.
>>>
>>> Thanks for all,
>>>
>>> Regards,
>>> Tibor Demeter
>>>
>>>
>>>
>>> ----- 2018. máj.. 16., 9:58, Sahina Bose <sab...@redhat.com> írta:
>>>
>>>
>>>
>>> On Wed, May 16, 2018 at 1:19 PM, Demeter Tibor <tdeme...@itsmart.hu>
>>> wrote:
>>>
>>>> Hi,
>>>>
>>>> is it a different, unstable repo? I have a productive cluster, how is
>>>> safe that?
>>>> I don't have any experience with nightly build. How can I use this? It
>>>> have to install to the engine VM or all of my hosts?
>>>> Thanks in advance for help me..
>>>>
>>>
>>> Only on the engine VM.
>>>
>>> Regarding stability - it passes CI so relatively stable, beyond that
>>> there are no guarantees.
>>>
>>> What's the specific problem you're facing with update? Can you elaborate?
>>>
>>>
>>>> Regards,
>>>>
>>>> Tibor
>>>>
>>>> ----- 2018. máj.. 15., 9:58, Demeter Tibor <tdeme...@itsmart.hu> írta:
>>>>
>>>> Hi,
>>>>
>>>> Could you explain how can I use this patch?
>>>>
>>>> R,
>>>> Tibor
>>>>
>>>>
>>>> ----- 2018. máj.. 14., 11:18, Demeter Tibor <tdeme...@itsmart.hu> írta:
>>>>
>>>> Hi,
>>>>
>>>> Sorry for my question, but can you tell me please how can I use this
>>>> patch?
>>>>
>>>> Thanks,
>>>> Regards,
>>>> Tibor
>>>> ----- 2018. máj.. 14., 10:47, Sahina Bose <sab...@redhat.com> írta:
>>>>
>>>>
>>>>
>>>> On Sat, May 12, 2018 at 1:14 PM, Demeter Tibor <tdeme...@itsmart.hu>
>>>> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> Could someone help me please ? I can't finish my upgrade process.
>>>>>
>>>>
>>>> https://gerrit.ovirt.org/91164 should fix the error you're facing.
>>>>
>>>> Can you elaborate why this is affecting the upgrade process?
>>>>
>>>>
>>>>> Thanks
>>>>> R
>>>>> Tibor
>>>>>
>>>>>
>>>>>
>>>>> ----- 2018. máj.. 10., 12:51, Demeter Tibor <tdeme...@itsmart.hu>
>>>>> írta:
>>>>>
>>>>> Hi,
>>>>>
>>>>> I've attached the vdsm and supervdsm logs. But I don't have engine.log
>>>>> here, because that is on hosted engine vm. Should I send that ?
>>>>>
>>>>> Thank you
>>>>>
>>>>> Regards,
>>>>>
>>>>> Tibor
>>>>> ----- 2018. máj.. 10., 12:30, Sahina Bose <sab...@redhat.com> írta:
>>>>>
>>>>> There's a bug here. Can you log one attaching this engine.log and also
>>>>> vdsm.log & supervdsm.log from n3.itsmart.cloud
>>>>>
>>>>> On Thu, May 10, 2018 at 3:35 PM, Demeter Tibor <tdeme...@itsmart.hu>
>>>>> wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> I found this:
>>>>>>
>>>>>>
>>>>>> 2018-05-10 03:24:19,096+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, 
>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand,
>>>>>> return: org.ovirt.engine.core.common.businessentities.gluster.
>>>>>> GlusterVolumeAdvancedDetails@ca97448e, log id: 347435ae
>>>>>> 2018-05-10 03:24:19,097+02 ERROR 
>>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob]
>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Error while refreshing brick 
>>>>>> statuses
>>>>>> for volume 'volume2' of cluster 'C6220': null
>>>>>> 2018-05-10 03:24:19,097+02 INFO  
>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>>>>> (DefaultQuartzScheduler8) [7715ceda] Failed to acquire lock and wait lock
>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>> sharedLocks=''}'
>>>>>> 2018-05-10 03:24:19,104+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, 
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName
>>>>>> = n4.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>>>> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), log id: 6908121d
>>>>>> 2018-05-10 03:24:19,106+02 ERROR [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command '
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
>>>>>> n4.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>>>> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' execution failed:
>>>>>> null
>>>>>> 2018-05-10 03:24:19,106+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, 
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>>>>> log id: 6908121d
>>>>>> 2018-05-10 03:24:19,107+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, 
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName
>>>>>> = n1.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>>>> {hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'}), log id: 735c6a5f
>>>>>> 2018-05-10 03:24:19,109+02 ERROR [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command '
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
>>>>>> n1.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>>>> {hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' execution failed:
>>>>>> null
>>>>>> 2018-05-10 03:24:19,109+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, 
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>>>>> log id: 735c6a5f
>>>>>> 2018-05-10 03:24:19,110+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, 
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName
>>>>>> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 6f9e9f58
>>>>>> 2018-05-10 03:24:19,112+02 ERROR [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command '
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
>>>>>> n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' execution failed:
>>>>>> null
>>>>>> 2018-05-10 03:24:19,112+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, 
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>>>>> log id: 6f9e9f58
>>>>>> 2018-05-10 03:24:19,113+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, 
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName
>>>>>> = n3.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>>>> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), log id: 2ee46967
>>>>>> 2018-05-10 03:24:19,115+02 ERROR [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command '
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
>>>>>> n3.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>>>> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' execution failed:
>>>>>> null
>>>>>> 2018-05-10 03:24:19,116+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, 
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>>>>> log id: 2ee46967
>>>>>> 2018-05-10 03:24:19,117+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, 
>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName
>>>>>> = n1.itsmart.cloud, GlusterVolumeAdvancedDetailsVD
>>>>>> SParameters:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57',
>>>>>> volumeName='volume1'}), log id: 7550e5c
>>>>>> 2018-05-10 03:24:20,748+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, 
>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand,
>>>>>> return: org.ovirt.engine.core.common.businessentities.gluster.
>>>>>> GlusterVolumeAdvancedDetails@4a46066f, log id: 7550e5c
>>>>>> 2018-05-10 03:24:20,749+02 ERROR 
>>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob]
>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Error while refreshing brick 
>>>>>> statuses
>>>>>> for volume 'volume1' of cluster 'C6220': null
>>>>>> 2018-05-10 03:24:20,750+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>>> (DefaultQuartzScheduler8) [7715ceda] START, 
>>>>>> GlusterServersListVDSCommand(HostName
>>>>>> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 120cc68d
>>>>>> 2018-05-10 03:24:20,930+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, 
>>>>>> GlusterServersListVDSCommand,
>>>>>> return: [10.101.0.2/24:CONNECTED, n1.cloudata.local:CONNECTED,
>>>>>> 10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], log id: 120cc68d
>>>>>> 2018-05-10 03:24:20,949+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>>> (DefaultQuartzScheduler8) [7715ceda] START, 
>>>>>> GlusterVolumesListVDSCommand(HostName
>>>>>> = n2.itsmart.cloud, GlusterVolumesListVDSParameter
>>>>>> s:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 118aa264
>>>>>> 2018-05-10 03:24:21,048+02 WARN  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick 
>>>>>> '10.104.0.1:
>>>>>> /gluster/brick/brick1' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27'
>>>>>> with correct network as no gluster network found in cluster
>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>> 2018-05-10 03:24:21,055+02 WARN  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick 
>>>>>> '10.104.0.1:
>>>>>> /gluster/brick/brick2' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27'
>>>>>> with correct network as no gluster network found in cluster
>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>> 2018-05-10 03:24:21,061+02 WARN  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick 
>>>>>> '10.104.0.1:
>>>>>> /gluster/brick/brick3' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27'
>>>>>> with correct network as no gluster network found in cluster
>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>> 2018-05-10 03:24:21,067+02 WARN  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick 
>>>>>> '10.104.0.1:
>>>>>> /gluster2/brick/brick1' of volume '68cfb061-1320-4042-abcd-9228da23c0c8'
>>>>>> with correct network as no gluster network found in cluster
>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>> 2018-05-10 03:24:21,074+02 WARN  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick 
>>>>>> '10.104.0.1:
>>>>>> /gluster2/brick/brick2' of volume '68cfb061-1320-4042-abcd-9228da23c0c8'
>>>>>> with correct network as no gluster network found in cluster
>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>> 2018-05-10 03:24:21,080+02 WARN  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick 
>>>>>> '10.104.0.1:
>>>>>> /gluster2/brick/brick3' of volume '68cfb061-1320-4042-abcd-9228da23c0c8'
>>>>>> with correct network as no gluster network found in cluster
>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>> 2018-05-10 03:24:21,081+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, 
>>>>>> GlusterVolumesListVDSCommand,
>>>>>> return: {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.
>>>>>> core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d,
>>>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.
>>>>>> core.common.businessentities.g
>>>>>> luster.GlusterVolumeEntity@f88c521b}, log id: 118aa264
>>>>>>
>>>>>>
>>>>>>
>>>>>> 2018-05-10 11:59:26,047+02 ERROR [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler4) [400fa486] Command '
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
>>>>>> n1.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>>>> {hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' execution failed:
>>>>>> null
>>>>>> 2018-05-10 11:59:26,047+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, 
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>>>>> log id: 14a71ef0
>>>>>> 2018-05-10 11:59:26,048+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler4) [400fa486] START, 
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName
>>>>>> = n4.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>>>> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), log id: 28d9e255
>>>>>> 2018-05-10 11:59:26,051+02 ERROR [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler4) [400fa486] Command '
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
>>>>>> n4.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>>>> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' execution failed:
>>>>>> null
>>>>>> 2018-05-10 11:59:26,051+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, 
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>>>>> log id: 28d9e255
>>>>>> 2018-05-10 11:59:26,052+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler4) [400fa486] START, 
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName
>>>>>> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 4a7b280e
>>>>>> 2018-05-10 11:59:26,054+02 ERROR [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler4) [400fa486] Command '
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
>>>>>> n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' execution failed:
>>>>>> null
>>>>>> 2018-05-10 11:59:26,054+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, 
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>>>>> log id: 4a7b280e
>>>>>> 2018-05-10 11:59:26,055+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler4) [400fa486] START, 
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName
>>>>>> = n3.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>>>> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), log id: 18adc534
>>>>>> 2018-05-10 11:59:26,057+02 ERROR [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler4) [400fa486] Command '
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
>>>>>> n3.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>>>> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' execution failed:
>>>>>> null
>>>>>> 2018-05-10 11:59:26,057+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, 
>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>>>>> log id: 18adc534
>>>>>> 2018-05-10 11:59:26,058+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>>> (DefaultQuartzScheduler4) [400fa486] START, 
>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName
>>>>>> = n3.itsmart.cloud, GlusterVolumeAdvancedDetailsVD
>>>>>> SParameters:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec',
>>>>>> volumeName='volume1'}), log id: 3451084f
>>>>>> 2018-05-10 11:59:28,050+02 INFO  
>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>>>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>> sharedLocks=''}'
>>>>>> 2018-05-10 11:59:28,060+02 INFO  
>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>>>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>> sharedLocks=''}'
>>>>>> 2018-05-10 11:59:28,062+02 INFO  
>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>>>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>> sharedLocks=''}'
>>>>>> 2018-05-10 11:59:31,054+02 INFO  
>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>>>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>> sharedLocks=''}'
>>>>>> 2018-05-10 11:59:31,054+02 INFO  
>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>>>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>> sharedLocks=''}'
>>>>>> 2018-05-10 11:59:31,062+02 INFO  
>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>>>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>> sharedLocks=''}'
>>>>>> 2018-05-10 11:59:31,064+02 INFO  
>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>>>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>>> sharedLocks=''}'
>>>>>> 2018-05-10 11:59:31,465+02 INFO  [org.ovirt.engine.core.
>>>>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, 
>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand,
>>>>>> return: org.ovirt.engine.core.common.businessentities.gluster.
>>>>>> GlusterVolumeAdvancedDetails@3f1b7f43, log id: 3451084f
>>>>>> 2018-05-10 11:59:31,466+02 ERROR 
>>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob]
>>>>>> (DefaultQuartzScheduler4) [400fa486] Error while refreshing brick 
>>>>>> statuses
>>>>>> for volume 'volume1' of cluster 'C6220': null
>>>>>>
>>>>>>
>>>>>> R
>>>>>> Tibor
>>>>>>
>>>>>> ----- 2018. máj.. 10., 11:43, Sahina Bose <sab...@redhat.com> írta:
>>>>>>
>>>>>> This doesn't affect the monitoring of state.
>>>>>> Any errors in vdsm.log?
>>>>>> Or errors in engine.log of the form "Error while refreshing brick
>>>>>> statuses for volume"
>>>>>>
>>>>>> On Thu, May 10, 2018 at 2:33 PM, Demeter Tibor <tdeme...@itsmart.hu>
>>>>>> wrote:
>>>>>>
>>>>>>> Hi,
>>>>>>>
>>>>>>> Thank you for your fast reply :)
>>>>>>>
>>>>>>>
>>>>>>> 2018-05-10 11:01:51,574+02 INFO  [org.ovirt.engine.core.
>>>>>>> vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START, 
>>>>>>> GlusterServersListVDSCommand(HostName
>>>>>>> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>>>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 39adbbb8
>>>>>>> 2018-05-10 11:01:51,768+02 INFO  [org.ovirt.engine.core.
>>>>>>> vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, 
>>>>>>> GlusterServersListVDSCommand,
>>>>>>> return: [10.101.0.2/24:CONNECTED, n1.cloudata.local:CONNECTED,
>>>>>>> 10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], log id: 39adbbb8
>>>>>>> 2018-05-10 11:01:51,788+02 INFO  [org.ovirt.engine.core.
>>>>>>> vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START, 
>>>>>>> GlusterVolumesListVDSCommand(HostName
>>>>>>> = n2.itsmart.cloud, GlusterVolumesListVDSParameter
>>>>>>> s:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 738a7261
>>>>>>> 2018-05-10 11:01:51,892+02 WARN  [org.ovirt.engine.core.
>>>>>>> vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick 
>>>>>>> '10.104.0.1:
>>>>>>> /gluster/brick/brick1' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27'
>>>>>>> with correct network as no gluster network found in cluster
>>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>>> 2018-05-10 11:01:51,898+02 WARN  [org.ovirt.engine.core.
>>>>>>> vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick 
>>>>>>> '10.104.0.1:
>>>>>>> /gluster/brick/brick2' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27'
>>>>>>> with correct network as no gluster network found in cluster
>>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>>> 2018-05-10 11:01:51,905+02 WARN  [org.ovirt.engine.core.
>>>>>>> vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick 
>>>>>>> '10.104.0.1:
>>>>>>> /gluster/brick/brick3' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27'
>>>>>>> with correct network as no gluster network found in cluster
>>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>>> 2018-05-10 11:01:51,911+02 WARN  [org.ovirt.engine.core.
>>>>>>> vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick 
>>>>>>> '10.104.0.1:
>>>>>>> /gluster2/brick/brick1' of volume '68cfb061-1320-4042-abcd-9228da23c0c8'
>>>>>>> with correct network as no gluster network found in cluster
>>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>>> 2018-05-10 11:01:51,917+02 WARN  [org.ovirt.engine.core.
>>>>>>> vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick 
>>>>>>> '10.104.0.1:
>>>>>>> /gluster2/brick/brick2' of volume '68cfb061-1320-4042-abcd-9228da23c0c8'
>>>>>>> with correct network as no gluster network found in cluster
>>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>>> 2018-05-10 11:01:51,924+02 WARN  [org.ovirt.engine.core.
>>>>>>> vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick 
>>>>>>> '10.104.0.1:
>>>>>>> /gluster2/brick/brick3' of volume '68cfb061-1320-4042-abcd-9228da23c0c8'
>>>>>>> with correct network as no gluster network found in cluster
>>>>>>> '59c10db3-0324-0320-0120-000000000339'
>>>>>>> 2018-05-10 11:01:51,925+02 INFO  [org.ovirt.engine.core.
>>>>>>> vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, 
>>>>>>> GlusterVolumesListVDSCommand,
>>>>>>> return: {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.
>>>>>>> core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d,
>>>>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.
>>>>>>> core.common.businessentities.gluster.GlusterVolumeEntity@f88c521b},
>>>>>>> log id: 738a7261
>>>>>>>
>>>>>>>
>>>>>>> This happening continuously.
>>>>>>>
>>>>>>> Thanks!
>>>>>>> Tibor
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> ----- 2018. máj.. 10., 10:56, Sahina Bose <sab...@redhat.com> írta:
>>>>>>>
>>>>>>> Could you check the engine.log if there are errors related to
>>>>>>> getting GlusterVolumeAdvancedDetails ?
>>>>>>>
>>>>>>> On Thu, May 10, 2018 at 2:02 PM, Demeter Tibor <tdeme...@itsmart.hu>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Dear Ovirt Users,
>>>>>>>> I've followed up the self-hosted-engine upgrade documentation, I
>>>>>>>> upgraded my 4.1 system to 4.2.3.
>>>>>>>> I upgaded the first node with yum upgrade, it seems working now
>>>>>>>> fine. But since upgrade, the gluster informations seems to displayed
>>>>>>>> incorrect on the admin panel. The volume yellow, and there are red 
>>>>>>>> bricks
>>>>>>>> from that node.
>>>>>>>> I've checked in console, I think my gluster is not degraded:
>>>>>>>>
>>>>>>>> root@n1 ~]# gluster volume list
>>>>>>>> volume1
>>>>>>>> volume2
>>>>>>>> [root@n1 ~]# gluster volume info
>>>>>>>>
>>>>>>>> Volume Name: volume1
>>>>>>>> Type: Distributed-Replicate
>>>>>>>> Volume ID: e0f568fa-987c-4f5c-b853-01bce718ee27
>>>>>>>> Status: Started
>>>>>>>> Snapshot Count: 0
>>>>>>>> Number of Bricks: 3 x 3 = 9
>>>>>>>> Transport-type: tcp
>>>>>>>> Bricks:
>>>>>>>> Brick1: 10.104.0.1:/gluster/brick/brick1
>>>>>>>> Brick2: 10.104.0.2:/gluster/brick/brick1
>>>>>>>> Brick3: 10.104.0.3:/gluster/brick/brick1
>>>>>>>> Brick4: 10.104.0.1:/gluster/brick/brick2
>>>>>>>> Brick5: 10.104.0.2:/gluster/brick/brick2
>>>>>>>> Brick6: 10.104.0.3:/gluster/brick/brick2
>>>>>>>> Brick7: 10.104.0.1:/gluster/brick/brick3
>>>>>>>> Brick8: 10.104.0.2:/gluster/brick/brick3
>>>>>>>> Brick9: 10.104.0.3:/gluster/brick/brick3
>>>>>>>> Options Reconfigured:
>>>>>>>> transport.address-family: inet
>>>>>>>> performance.readdir-ahead: on
>>>>>>>> nfs.disable: on
>>>>>>>> storage.owner-uid: 36
>>>>>>>> storage.owner-gid: 36
>>>>>>>> performance.quick-read: off
>>>>>>>> performance.read-ahead: off
>>>>>>>> performance.io-cache: off
>>>>>>>> performance.stat-prefetch: off
>>>>>>>> performance.low-prio-threads: 32
>>>>>>>> network.remote-dio: enable
>>>>>>>> cluster.eager-lock: enable
>>>>>>>> cluster.quorum-type: auto
>>>>>>>> cluster.server-quorum-type: server
>>>>>>>> cluster.data-self-heal-algorithm: full
>>>>>>>> cluster.locking-scheme: granular
>>>>>>>> cluster.shd-max-threads: 8
>>>>>>>> cluster.shd-wait-qlength: 10000
>>>>>>>> features.shard: on
>>>>>>>> user.cifs: off
>>>>>>>> server.allow-insecure: on
>>>>>>>>
>>>>>>>> Volume Name: volume2
>>>>>>>> Type: Distributed-Replicate
>>>>>>>> Volume ID: 68cfb061-1320-4042-abcd-9228da23c0c8
>>>>>>>> Status: Started
>>>>>>>> Snapshot Count: 0
>>>>>>>> Number of Bricks: 3 x 3 = 9
>>>>>>>> Transport-type: tcp
>>>>>>>> Bricks:
>>>>>>>> Brick1: 10.104.0.1:/gluster2/brick/brick1
>>>>>>>> Brick2: 10.104.0.2:/gluster2/brick/brick1
>>>>>>>> Brick3: 10.104.0.3:/gluster2/brick/brick1
>>>>>>>> Brick4: 10.104.0.1:/gluster2/brick/brick2
>>>>>>>> Brick5: 10.104.0.2:/gluster2/brick/brick2
>>>>>>>> Brick6: 10.104.0.3:/gluster2/brick/brick2
>>>>>>>> Brick7: 10.104.0.1:/gluster2/brick/brick3
>>>>>>>> Brick8: 10.104.0.2:/gluster2/brick/brick3
>>>>>>>> Brick9: 10.104.0.3:/gluster2/brick/brick3
>>>>>>>> Options Reconfigured:
>>>>>>>> nfs.disable: on
>>>>>>>> performance.readdir-ahead: on
>>>>>>>> transport.address-family: inet
>>>>>>>> cluster.quorum-type: auto
>>>>>>>> network.ping-timeout: 10
>>>>>>>> auth.allow: *
>>>>>>>> performance.quick-read: off
>>>>>>>> performance.read-ahead: off
>>>>>>>> performance.io-cache: off
>>>>>>>> performance.stat-prefetch: off
>>>>>>>> performance.low-prio-threads: 32
>>>>>>>> network.remote-dio: enable
>>>>>>>> cluster.eager-lock: enable
>>>>>>>> cluster.server-quorum-type: server
>>>>>>>> cluster.data-self-heal-algorithm: full
>>>>>>>> cluster.locking-scheme: granular
>>>>>>>> cluster.shd-max-threads: 8
>>>>>>>> cluster.shd-wait-qlength: 10000
>>>>>>>> features.shard: on
>>>>>>>> user.cifs: off
>>>>>>>> storage.owner-uid: 36
>>>>>>>> storage.owner-gid: 36
>>>>>>>> server.allow-insecure: on
>>>>>>>> [root@n1 ~]# gluster volume status
>>>>>>>> Status of volume: volume1
>>>>>>>> Gluster process                             TCP Port  RDMA Port
>>>>>>>> Online  Pid
>>>>>>>> ------------------------------------------------------------
>>>>>>>> ------------------
>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick1      49152     0
>>>>>>>> Y       3464
>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick1      49152     0
>>>>>>>> Y       68937
>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick1      49161     0
>>>>>>>> Y       94506
>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick2      49153     0
>>>>>>>> Y       3457
>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick2      49153     0
>>>>>>>> Y       68943
>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick2      49162     0
>>>>>>>> Y       94514
>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick3      49154     0
>>>>>>>> Y       3465
>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick3      49154     0
>>>>>>>> Y       68949
>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick3      49163     0
>>>>>>>> Y       94520
>>>>>>>> Self-heal Daemon on localhost               N/A       N/A        Y
>>>>>>>>      54356
>>>>>>>> Self-heal Daemon on 10.104.0.2              N/A       N/A        Y
>>>>>>>>      962
>>>>>>>> Self-heal Daemon on 10.104.0.3              N/A       N/A        Y
>>>>>>>>      108977
>>>>>>>> Self-heal Daemon on 10.104.0.4              N/A       N/A        Y
>>>>>>>>      61603
>>>>>>>>
>>>>>>>> Task Status of Volume volume1
>>>>>>>> ------------------------------------------------------------
>>>>>>>> ------------------
>>>>>>>> There are no active volume tasks
>>>>>>>>
>>>>>>>> Status of volume: volume2
>>>>>>>> Gluster process                             TCP Port  RDMA Port
>>>>>>>> Online  Pid
>>>>>>>> ------------------------------------------------------------
>>>>>>>> ------------------
>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick1     49155     0
>>>>>>>> Y       3852
>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick1     49158     0
>>>>>>>> Y       68955
>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick1     49164     0
>>>>>>>> Y       94527
>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick2     49156     0
>>>>>>>> Y       3851
>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick2     49159     0
>>>>>>>> Y       68961
>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick2     49165     0
>>>>>>>> Y       94533
>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick3     49157     0
>>>>>>>> Y       3883
>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick3     49160     0
>>>>>>>> Y       68968
>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick3     49166     0
>>>>>>>> Y       94541
>>>>>>>> Self-heal Daemon on localhost               N/A       N/A        Y
>>>>>>>>      54356
>>>>>>>> Self-heal Daemon on 10.104.0.2              N/A       N/A        Y
>>>>>>>>      962
>>>>>>>> Self-heal Daemon on 10.104.0.3              N/A       N/A        Y
>>>>>>>>      108977
>>>>>>>> Self-heal Daemon on 10.104.0.4              N/A       N/A        Y
>>>>>>>>      61603
>>>>>>>>
>>>>>>>> Task Status of Volume volume2
>>>>>>>> ------------------------------------------------------------
>>>>>>>> ------------------
>>>>>>>> There are no active volume tasks
>>>>>>>>
>>>>>>>> I think ovirt can't read valid informations about gluster.
>>>>>>>> I can't contiune upgrade of other hosts until this problem exist.
>>>>>>>>
>>>>>>>> Please help me:)
>>>>>>>>
>>>>>>>>
>>>>>>>> Thanks
>>>>>>>>
>>>>>>>> Regards,
>>>>>>>>
>>>>>>>> Tibor
>>>>>>>>
>>>>>>>>
>>>>>>>> _______________________________________________
>>>>>>>> Users mailing list -- users@ovirt.org
>>>>>>>> To unsubscribe send an email to users-le...@ovirt.org
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> Users mailing list -- users@ovirt.org
>>>>> To unsubscribe send an email to users-le...@ovirt.org
>>>>>
>>>>>
>>>>
>>>> _______________________________________________
>>>> Users mailing list -- users@ovirt.org
>>>> To unsubscribe send an email to users-le...@ovirt.org
>>>>
>>>>
>>>> _______________________________________________
>>>> Users mailing list -- users@ovirt.org
>>>> To unsubscribe send an email to users-le...@ovirt.org
>>>> oVirt Code of Conduct: https://www.ovirt.org/community/about/community-
>>>> guidelines/
>>>> List Archives:
>>>>
>>>>
>>>
>>
>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org

Reply via email to