On Tue, May 15, 2018 at 1:28 PM, Demeter Tibor <tdeme...@itsmart.hu> wrote:

> Hi,
>
> Could you explain how can I use this patch?
>

You can use the 4.2 nightly to test it out -
http://resources.ovirt.org/pub/yum-repo/ovirt-release42-snapshot.rpm


> R,
> Tibor
>
>
> ----- 2018. máj.. 14., 11:18, Demeter Tibor <tdeme...@itsmart.hu> írta:
>
> Hi,
>
> Sorry for my question, but can you tell me please how can I use this patch?
>
> Thanks,
> Regards,
> Tibor
> ----- 2018. máj.. 14., 10:47, Sahina Bose <sab...@redhat.com> írta:
>
>
>
> On Sat, May 12, 2018 at 1:14 PM, Demeter Tibor <tdeme...@itsmart.hu>
> wrote:
>
>> Hi,
>>
>> Could someone help me please ? I can't finish my upgrade process.
>>
>
> https://gerrit.ovirt.org/91164 should fix the error you're facing.
>
> Can you elaborate why this is affecting the upgrade process?
>
>
>> Thanks
>> R
>> Tibor
>>
>>
>>
>> ----- 2018. máj.. 10., 12:51, Demeter Tibor <tdeme...@itsmart.hu> írta:
>>
>> Hi,
>>
>> I've attached the vdsm and supervdsm logs. But I don't have engine.log
>> here, because that is on hosted engine vm. Should I send that ?
>>
>> Thank you
>>
>> Regards,
>>
>> Tibor
>> ----- 2018. máj.. 10., 12:30, Sahina Bose <sab...@redhat.com> írta:
>>
>> There's a bug here. Can you log one attaching this engine.log and also
>> vdsm.log & supervdsm.log from n3.itsmart.cloud
>>
>> On Thu, May 10, 2018 at 3:35 PM, Demeter Tibor <tdeme...@itsmart.hu>
>> wrote:
>>
>>> Hi,
>>>
>>> I found this:
>>>
>>>
>>> 2018-05-10 03:24:19,096+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, 
>>> GetGlusterVolumeAdvancedDetailsVDSCommand,
>>> return: org.ovirt.engine.core.common.businessentities.gluster.
>>> GlusterVolumeAdvancedDetails@ca97448e, log id: 347435ae
>>> 2018-05-10 03:24:19,097+02 ERROR 
>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob]
>>> (DefaultQuartzScheduler7) [43f4eaec] Error while refreshing brick statuses
>>> for volume 'volume2' of cluster 'C6220': null
>>> 2018-05-10 03:24:19,097+02 INFO  
>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>> (DefaultQuartzScheduler8) [7715ceda] Failed to acquire lock and wait lock
>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>> sharedLocks=''}'
>>> 2018-05-10 03:24:19,104+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler7) [43f4eaec] START, 
>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName
>>> = n4.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), log id: 6908121d
>>> 2018-05-10 03:24:19,106+02 ERROR [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler7) [43f4eaec] Command '
>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud,
>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})'
>>> execution failed: null
>>> 2018-05-10 03:24:19,106+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, 
>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>> log id: 6908121d
>>> 2018-05-10 03:24:19,107+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler7) [43f4eaec] START, 
>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName
>>> = n1.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>> {hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'}), log id: 735c6a5f
>>> 2018-05-10 03:24:19,109+02 ERROR [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler7) [43f4eaec] Command '
>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n1.itsmart.cloud,
>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})'
>>> execution failed: null
>>> 2018-05-10 03:24:19,109+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, 
>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>> log id: 735c6a5f
>>> 2018-05-10 03:24:19,110+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler7) [43f4eaec] START, 
>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName
>>> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 6f9e9f58
>>> 2018-05-10 03:24:19,112+02 ERROR [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler7) [43f4eaec] Command '
>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud,
>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})'
>>> execution failed: null
>>> 2018-05-10 03:24:19,112+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, 
>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>> log id: 6f9e9f58
>>> 2018-05-10 03:24:19,113+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler7) [43f4eaec] START, 
>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName
>>> = n3.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), log id: 2ee46967
>>> 2018-05-10 03:24:19,115+02 ERROR [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler7) [43f4eaec] Command '
>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud,
>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})'
>>> execution failed: null
>>> 2018-05-10 03:24:19,116+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, 
>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>> log id: 2ee46967
>>> 2018-05-10 03:24:19,117+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>> (DefaultQuartzScheduler7) [43f4eaec] START, 
>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName
>>> = n1.itsmart.cloud, GlusterVolumeAdvancedDetailsVD
>>> SParameters:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57',
>>> volumeName='volume1'}), log id: 7550e5c
>>> 2018-05-10 03:24:20,748+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, 
>>> GetGlusterVolumeAdvancedDetailsVDSCommand,
>>> return: org.ovirt.engine.core.common.businessentities.gluster.
>>> GlusterVolumeAdvancedDetails@4a46066f, log id: 7550e5c
>>> 2018-05-10 03:24:20,749+02 ERROR 
>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob]
>>> (DefaultQuartzScheduler7) [43f4eaec] Error while refreshing brick statuses
>>> for volume 'volume1' of cluster 'C6220': null
>>> 2018-05-10 03:24:20,750+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GlusterServersListVDSCommand]
>>> (DefaultQuartzScheduler8) [7715ceda] START, 
>>> GlusterServersListVDSCommand(HostName
>>> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 120cc68d
>>> 2018-05-10 03:24:20,930+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GlusterServersListVDSCommand]
>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, GlusterServersListVDSCommand,
>>> return: [10.101.0.2/24:CONNECTED, n1.cloudata.local:CONNECTED,
>>> 10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], log id: 120cc68d
>>> 2018-05-10 03:24:20,949+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>> (DefaultQuartzScheduler8) [7715ceda] START, 
>>> GlusterVolumesListVDSCommand(HostName
>>> = n2.itsmart.cloud, GlusterVolumesListVDSParameter
>>> s:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 118aa264
>>> 2018-05-10 03:24:21,048+02 WARN  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8)
>>> [7715ceda] Could not associate brick '10.104.0.1:/gluster/brick/brick1'
>>> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network
>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>> 000000000339'
>>> 2018-05-10 03:24:21,055+02 WARN  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8)
>>> [7715ceda] Could not associate brick '10.104.0.1:/gluster/brick/brick2'
>>> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network
>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>> 000000000339'
>>> 2018-05-10 03:24:21,061+02 WARN  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8)
>>> [7715ceda] Could not associate brick '10.104.0.1:/gluster/brick/brick3'
>>> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network
>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>> 000000000339'
>>> 2018-05-10 03:24:21,067+02 WARN  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8)
>>> [7715ceda] Could not associate brick '10.104.0.1:/gluster2/brick/brick1'
>>> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network
>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>> 000000000339'
>>> 2018-05-10 03:24:21,074+02 WARN  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8)
>>> [7715ceda] Could not associate brick '10.104.0.1:/gluster2/brick/brick2'
>>> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network
>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>> 000000000339'
>>> 2018-05-10 03:24:21,080+02 WARN  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler8)
>>> [7715ceda] Could not associate brick '10.104.0.1:/gluster2/brick/brick3'
>>> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network
>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>> 000000000339'
>>> 2018-05-10 03:24:21,081+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, GlusterVolumesListVDSCommand,
>>> return: {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.
>>> core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d,
>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.
>>> core.common.businessentities.g
>>> luster.GlusterVolumeEntity@f88c521b}, log id: 118aa264
>>>
>>>
>>>
>>> 2018-05-10 11:59:26,047+02 ERROR [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler4) [400fa486] Command '
>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n1.itsmart.cloud,
>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})'
>>> execution failed: null
>>> 2018-05-10 11:59:26,047+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler4) [400fa486] FINISH, 
>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>> log id: 14a71ef0
>>> 2018-05-10 11:59:26,048+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler4) [400fa486] START, 
>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName
>>> = n4.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), log id: 28d9e255
>>> 2018-05-10 11:59:26,051+02 ERROR [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler4) [400fa486] Command '
>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud,
>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})'
>>> execution failed: null
>>> 2018-05-10 11:59:26,051+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler4) [400fa486] FINISH, 
>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>> log id: 28d9e255
>>> 2018-05-10 11:59:26,052+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler4) [400fa486] START, 
>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName
>>> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 4a7b280e
>>> 2018-05-10 11:59:26,054+02 ERROR [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler4) [400fa486] Command '
>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud,
>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})'
>>> execution failed: null
>>> 2018-05-10 11:59:26,054+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler4) [400fa486] FINISH, 
>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>> log id: 4a7b280e
>>> 2018-05-10 11:59:26,055+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler4) [400fa486] START, 
>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName
>>> = n3.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), log id: 18adc534
>>> 2018-05-10 11:59:26,057+02 ERROR [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler4) [400fa486] Command '
>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud,
>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})'
>>> execution failed: null
>>> 2018-05-10 11:59:26,057+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>> (DefaultQuartzScheduler4) [400fa486] FINISH, 
>>> GetGlusterLocalLogicalVolumeListVDSCommand,
>>> log id: 18adc534
>>> 2018-05-10 11:59:26,058+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>> (DefaultQuartzScheduler4) [400fa486] START, 
>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName
>>> = n3.itsmart.cloud, GlusterVolumeAdvancedDetailsVD
>>> SParameters:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec',
>>> volumeName='volume1'}), log id: 3451084f
>>> 2018-05-10 11:59:28,050+02 INFO  
>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>> sharedLocks=''}'
>>> 2018-05-10 11:59:28,060+02 INFO  
>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>> sharedLocks=''}'
>>> 2018-05-10 11:59:28,062+02 INFO  
>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>> sharedLocks=''}'
>>> 2018-05-10 11:59:31,054+02 INFO  
>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>> sharedLocks=''}'
>>> 2018-05-10 11:59:31,054+02 INFO  
>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>> sharedLocks=''}'
>>> 2018-05-10 11:59:31,062+02 INFO  
>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>> sharedLocks=''}'
>>> 2018-05-10 11:59:31,064+02 INFO  
>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>> sharedLocks=''}'
>>> 2018-05-10 11:59:31,465+02 INFO  [org.ovirt.engine.core.
>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>> (DefaultQuartzScheduler4) [400fa486] FINISH, 
>>> GetGlusterVolumeAdvancedDetailsVDSCommand,
>>> return: org.ovirt.engine.core.common.businessentities.gluster.
>>> GlusterVolumeAdvancedDetails@3f1b7f43, log id: 3451084f
>>> 2018-05-10 11:59:31,466+02 ERROR 
>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob]
>>> (DefaultQuartzScheduler4) [400fa486] Error while refreshing brick statuses
>>> for volume 'volume1' of cluster 'C6220': null
>>>
>>>
>>> R
>>> Tibor
>>>
>>> ----- 2018. máj.. 10., 11:43, Sahina Bose <sab...@redhat.com> írta:
>>>
>>> This doesn't affect the monitoring of state.
>>> Any errors in vdsm.log?
>>> Or errors in engine.log of the form "Error while refreshing brick
>>> statuses for volume"
>>>
>>> On Thu, May 10, 2018 at 2:33 PM, Demeter Tibor <tdeme...@itsmart.hu>
>>> wrote:
>>>
>>>> Hi,
>>>>
>>>> Thank you for your fast reply :)
>>>>
>>>>
>>>> 2018-05-10 11:01:51,574+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GlusterServersListVDSCommand]
>>>> (DefaultQuartzScheduler6) [7f01fc2d] START, 
>>>> GlusterServersListVDSCommand(HostName
>>>> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
>>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 39adbbb8
>>>> 2018-05-10 11:01:51,768+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GlusterServersListVDSCommand]
>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, GlusterServersListVDSCommand,
>>>> return: [10.101.0.2/24:CONNECTED, n1.cloudata.local:CONNECTED,
>>>> 10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], log id: 39adbbb8
>>>> 2018-05-10 11:01:51,788+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>> (DefaultQuartzScheduler6) [7f01fc2d] START, 
>>>> GlusterVolumesListVDSCommand(HostName
>>>> = n2.itsmart.cloud, GlusterVolumesListVDSParameter
>>>> s:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 738a7261
>>>> 2018-05-10 11:01:51,892+02 WARN  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6)
>>>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster/brick/brick1'
>>>> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network
>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>>> 000000000339'
>>>> 2018-05-10 11:01:51,898+02 WARN  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6)
>>>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster/brick/brick2'
>>>> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network
>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>>> 000000000339'
>>>> 2018-05-10 11:01:51,905+02 WARN  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6)
>>>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster/brick/brick3'
>>>> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network
>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>>> 000000000339'
>>>> 2018-05-10 11:01:51,911+02 WARN  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6)
>>>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster2/brick/brick1'
>>>> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network
>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>>> 000000000339'
>>>> 2018-05-10 11:01:51,917+02 WARN  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6)
>>>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster2/brick/brick2'
>>>> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network
>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>>> 000000000339'
>>>> 2018-05-10 11:01:51,924+02 WARN  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6)
>>>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster2/brick/brick3'
>>>> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network
>>>> as no gluster network found in cluster '59c10db3-0324-0320-0120-
>>>> 000000000339'
>>>> 2018-05-10 11:01:51,925+02 INFO  [org.ovirt.engine.core.
>>>> vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, GlusterVolumesListVDSCommand,
>>>> return: {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.
>>>> core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d,
>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.
>>>> core.common.businessentities.gluster.GlusterVolumeEntity@f88c521b},
>>>> log id: 738a7261
>>>>
>>>>
>>>> This happening continuously.
>>>>
>>>> Thanks!
>>>> Tibor
>>>>
>>>>
>>>>
>>>> ----- 2018. máj.. 10., 10:56, Sahina Bose <sab...@redhat.com> írta:
>>>>
>>>> Could you check the engine.log if there are errors related to getting
>>>> GlusterVolumeAdvancedDetails ?
>>>>
>>>> On Thu, May 10, 2018 at 2:02 PM, Demeter Tibor <tdeme...@itsmart.hu>
>>>> wrote:
>>>>
>>>>> Dear Ovirt Users,
>>>>> I've followed up the self-hosted-engine upgrade documentation, I
>>>>> upgraded my 4.1 system to 4.2.3.
>>>>> I upgaded the first node with yum upgrade, it seems working now fine.
>>>>> But since upgrade, the gluster informations seems to displayed incorrect 
>>>>> on
>>>>> the admin panel. The volume yellow, and there are red bricks from that 
>>>>> node.
>>>>> I've checked in console, I think my gluster is not degraded:
>>>>>
>>>>> root@n1 ~]# gluster volume list
>>>>> volume1
>>>>> volume2
>>>>> [root@n1 ~]# gluster volume info
>>>>>
>>>>> Volume Name: volume1
>>>>> Type: Distributed-Replicate
>>>>> Volume ID: e0f568fa-987c-4f5c-b853-01bce718ee27
>>>>> Status: Started
>>>>> Snapshot Count: 0
>>>>> Number of Bricks: 3 x 3 = 9
>>>>> Transport-type: tcp
>>>>> Bricks:
>>>>> Brick1: 10.104.0.1:/gluster/brick/brick1
>>>>> Brick2: 10.104.0.2:/gluster/brick/brick1
>>>>> Brick3: 10.104.0.3:/gluster/brick/brick1
>>>>> Brick4: 10.104.0.1:/gluster/brick/brick2
>>>>> Brick5: 10.104.0.2:/gluster/brick/brick2
>>>>> Brick6: 10.104.0.3:/gluster/brick/brick2
>>>>> Brick7: 10.104.0.1:/gluster/brick/brick3
>>>>> Brick8: 10.104.0.2:/gluster/brick/brick3
>>>>> Brick9: 10.104.0.3:/gluster/brick/brick3
>>>>> Options Reconfigured:
>>>>> transport.address-family: inet
>>>>> performance.readdir-ahead: on
>>>>> nfs.disable: on
>>>>> storage.owner-uid: 36
>>>>> storage.owner-gid: 36
>>>>> performance.quick-read: off
>>>>> performance.read-ahead: off
>>>>> performance.io-cache: off
>>>>> performance.stat-prefetch: off
>>>>> performance.low-prio-threads: 32
>>>>> network.remote-dio: enable
>>>>> cluster.eager-lock: enable
>>>>> cluster.quorum-type: auto
>>>>> cluster.server-quorum-type: server
>>>>> cluster.data-self-heal-algorithm: full
>>>>> cluster.locking-scheme: granular
>>>>> cluster.shd-max-threads: 8
>>>>> cluster.shd-wait-qlength: 10000
>>>>> features.shard: on
>>>>> user.cifs: off
>>>>> server.allow-insecure: on
>>>>>
>>>>> Volume Name: volume2
>>>>> Type: Distributed-Replicate
>>>>> Volume ID: 68cfb061-1320-4042-abcd-9228da23c0c8
>>>>> Status: Started
>>>>> Snapshot Count: 0
>>>>> Number of Bricks: 3 x 3 = 9
>>>>> Transport-type: tcp
>>>>> Bricks:
>>>>> Brick1: 10.104.0.1:/gluster2/brick/brick1
>>>>> Brick2: 10.104.0.2:/gluster2/brick/brick1
>>>>> Brick3: 10.104.0.3:/gluster2/brick/brick1
>>>>> Brick4: 10.104.0.1:/gluster2/brick/brick2
>>>>> Brick5: 10.104.0.2:/gluster2/brick/brick2
>>>>> Brick6: 10.104.0.3:/gluster2/brick/brick2
>>>>> Brick7: 10.104.0.1:/gluster2/brick/brick3
>>>>> Brick8: 10.104.0.2:/gluster2/brick/brick3
>>>>> Brick9: 10.104.0.3:/gluster2/brick/brick3
>>>>> Options Reconfigured:
>>>>> nfs.disable: on
>>>>> performance.readdir-ahead: on
>>>>> transport.address-family: inet
>>>>> cluster.quorum-type: auto
>>>>> network.ping-timeout: 10
>>>>> auth.allow: *
>>>>> performance.quick-read: off
>>>>> performance.read-ahead: off
>>>>> performance.io-cache: off
>>>>> performance.stat-prefetch: off
>>>>> performance.low-prio-threads: 32
>>>>> network.remote-dio: enable
>>>>> cluster.eager-lock: enable
>>>>> cluster.server-quorum-type: server
>>>>> cluster.data-self-heal-algorithm: full
>>>>> cluster.locking-scheme: granular
>>>>> cluster.shd-max-threads: 8
>>>>> cluster.shd-wait-qlength: 10000
>>>>> features.shard: on
>>>>> user.cifs: off
>>>>> storage.owner-uid: 36
>>>>> storage.owner-gid: 36
>>>>> server.allow-insecure: on
>>>>> [root@n1 ~]# gluster volume status
>>>>> Status of volume: volume1
>>>>> Gluster process                             TCP Port  RDMA Port
>>>>> Online  Pid
>>>>> ------------------------------------------------------------
>>>>> ------------------
>>>>> Brick 10.104.0.1:/gluster/brick/brick1      49152     0          Y
>>>>>    3464
>>>>> Brick 10.104.0.2:/gluster/brick/brick1      49152     0          Y
>>>>>    68937
>>>>> Brick 10.104.0.3:/gluster/brick/brick1      49161     0          Y
>>>>>    94506
>>>>> Brick 10.104.0.1:/gluster/brick/brick2      49153     0          Y
>>>>>    3457
>>>>> Brick 10.104.0.2:/gluster/brick/brick2      49153     0          Y
>>>>>    68943
>>>>> Brick 10.104.0.3:/gluster/brick/brick2      49162     0          Y
>>>>>    94514
>>>>> Brick 10.104.0.1:/gluster/brick/brick3      49154     0          Y
>>>>>    3465
>>>>> Brick 10.104.0.2:/gluster/brick/brick3      49154     0          Y
>>>>>    68949
>>>>> Brick 10.104.0.3:/gluster/brick/brick3      49163     0          Y
>>>>>    94520
>>>>> Self-heal Daemon on localhost               N/A       N/A        Y
>>>>>    54356
>>>>> Self-heal Daemon on 10.104.0.2              N/A       N/A        Y
>>>>>    962
>>>>> Self-heal Daemon on 10.104.0.3              N/A       N/A        Y
>>>>>    108977
>>>>> Self-heal Daemon on 10.104.0.4              N/A       N/A        Y
>>>>>    61603
>>>>>
>>>>> Task Status of Volume volume1
>>>>> ------------------------------------------------------------
>>>>> ------------------
>>>>> There are no active volume tasks
>>>>>
>>>>> Status of volume: volume2
>>>>> Gluster process                             TCP Port  RDMA Port
>>>>> Online  Pid
>>>>> ------------------------------------------------------------
>>>>> ------------------
>>>>> Brick 10.104.0.1:/gluster2/brick/brick1     49155     0          Y
>>>>>    3852
>>>>> Brick 10.104.0.2:/gluster2/brick/brick1     49158     0          Y
>>>>>    68955
>>>>> Brick 10.104.0.3:/gluster2/brick/brick1     49164     0          Y
>>>>>    94527
>>>>> Brick 10.104.0.1:/gluster2/brick/brick2     49156     0          Y
>>>>>    3851
>>>>> Brick 10.104.0.2:/gluster2/brick/brick2     49159     0          Y
>>>>>    68961
>>>>> Brick 10.104.0.3:/gluster2/brick/brick2     49165     0          Y
>>>>>    94533
>>>>> Brick 10.104.0.1:/gluster2/brick/brick3     49157     0          Y
>>>>>    3883
>>>>> Brick 10.104.0.2:/gluster2/brick/brick3     49160     0          Y
>>>>>    68968
>>>>> Brick 10.104.0.3:/gluster2/brick/brick3     49166     0          Y
>>>>>    94541
>>>>> Self-heal Daemon on localhost               N/A       N/A        Y
>>>>>    54356
>>>>> Self-heal Daemon on 10.104.0.2              N/A       N/A        Y
>>>>>    962
>>>>> Self-heal Daemon on 10.104.0.3              N/A       N/A        Y
>>>>>    108977
>>>>> Self-heal Daemon on 10.104.0.4              N/A       N/A        Y
>>>>>    61603
>>>>>
>>>>> Task Status of Volume volume2
>>>>> ------------------------------------------------------------
>>>>> ------------------
>>>>> There are no active volume tasks
>>>>>
>>>>> I think ovirt can't read valid informations about gluster.
>>>>> I can't contiune upgrade of other hosts until this problem exist.
>>>>>
>>>>> Please help me:)
>>>>>
>>>>>
>>>>> Thanks
>>>>>
>>>>> Regards,
>>>>>
>>>>> Tibor
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> Users mailing list -- users@ovirt.org
>>>>> To unsubscribe send an email to users-le...@ovirt.org
>>>>>
>>>>>
>>>>
>>>
>>
>> _______________________________________________
>> Users mailing list -- users@ovirt.org
>> To unsubscribe send an email to users-le...@ovirt.org
>>
>>
>
> _______________________________________________
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
>
>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives:

Reply via email to