Hi, 

Yes, I have a gluster network, but it's "funny" because that is the 
10.105.0.x/24. :( Also, the n4.itsmart.cloud is mean 10.104.0.4. 

The 10.104.0.x/24 is my ovirtmgmt network. 

However, the 10.104.0.x is accessable from all hosts. 

What should I do? 

Thanks, 

R 

Tibor 

----- 2018. máj.. 12., 17:17, Doug Ingham <[email protected]> írta: 

> The two key errors I'd investigate are these...

>> 2018-05-10 03:24:21,048+02 WARN
>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>> '10.104.0.1:/gluster/brick/brick1' of volume
>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster
>> network found in cluster '59c10db3-0324-0320-0120-000000000339'

>> 2018-05-10 03:24:20,749+02 ERROR
>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] (DefaultQuartzScheduler7)
>> [43f4eaec] Error while refreshing brick statuses for volume 'volume1' of
>> cluster 'C6220': null

>> 2018-05-10 11:59:26,051+02 ERROR
>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>> (DefaultQuartzScheduler4) [400fa486] Command
>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud,
>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})'
>> execution failed: null

> I'd start with that first one. Is the network/interface group of your storage
> layer actually defined as a Gluster & Migration network within oVirt?

> On 12 May 2018 at 03:44, Demeter Tibor < [ mailto:[email protected] |
> [email protected] ] > wrote:

>> Hi,

>> Could someone help me please ? I can't finish my upgrade process.

>> Thanks
>> R
>> Tibor

>> ----- 2018. máj.. 10., 12:51, Demeter Tibor < [ mailto:[email protected] |
>> [email protected] ] > írta:

>>> Hi,

>>> I've attached the vdsm and supervdsm logs. But I don't have engine.log here,
>>> because that is on hosted engine vm. Should I send that ?

>>> Thank you

>>> Regards,

>>> Tibor
>>> ----- 2018. máj.. 10., 12:30, Sahina Bose < [ mailto:[email protected] |
>>> [email protected] ] > írta:

>>>> There's a bug here. Can you log one attaching this engine.log and also 
>>>> vdsm.log
>>>> & supervdsm.log from n3.itsmart.cloud

>>>> On Thu, May 10, 2018 at 3:35 PM, Demeter Tibor < [ 
>>>> mailto:[email protected] |
>>>> [email protected] ] > wrote:

>>>>> Hi,

>>>>> I found this:

>>>>> 2018-05-10 03:24:19,096+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return:
>>>>> org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@ca97448e,
>>>>> log id: 347435ae
>>>>> 2018-05-10 03:24:19,097+02 ERROR
>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] 
>>>>> (DefaultQuartzScheduler7)
>>>>> [43f4eaec] Error while refreshing brick statuses for volume 'volume2' of
>>>>> cluster 'C6220': null
>>>>> 2018-05-10 03:24:19,097+02 INFO
>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>> (DefaultQuartzScheduler8)
>>>>> [7715ceda] Failed to acquire lock and wait lock
>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>> sharedLocks=''}'
>>>>> 2018-05-10 03:24:19,104+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler7) [43f4eaec] START,
>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud,
>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}),
>>>>> log id: 6908121d
>>>>> 2018-05-10 03:24:19,106+02 ERROR
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command
>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud,
>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})'
>>>>> execution failed: null
>>>>> 2018-05-10 03:24:19,106+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6908121d
>>>>> 2018-05-10 03:24:19,107+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler7) [43f4eaec] START,
>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n1.itsmart.cloud,
>>>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'}),
>>>>> log id: 735c6a5f
>>>>> 2018-05-10 03:24:19,109+02 ERROR
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command
>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n1.itsmart.cloud,
>>>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})'
>>>>> execution failed: null
>>>>> 2018-05-10 03:24:19,109+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 735c6a5f
>>>>> 2018-05-10 03:24:19,110+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler7) [43f4eaec] START,
>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud,
>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>> log id: 6f9e9f58
>>>>> 2018-05-10 03:24:19,112+02 ERROR
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command
>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud,
>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})'
>>>>> execution failed: null
>>>>> 2018-05-10 03:24:19,112+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6f9e9f58
>>>>> 2018-05-10 03:24:19,113+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler7) [43f4eaec] START,
>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud,
>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}),
>>>>> log id: 2ee46967
>>>>> 2018-05-10 03:24:19,115+02 ERROR
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command
>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud,
>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})'
>>>>> execution failed: null
>>>>> 2018-05-10 03:24:19,116+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 2ee46967
>>>>> 2018-05-10 03:24:19,117+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>> (DefaultQuartzScheduler7) [43f4eaec] START,
>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = n1.itsmart.cloud,
>>>>> GlusterVolumeAdvancedDetailsVDSParameters:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57',
>>>>> volumeName='volume1'}), log id: 7550e5c
>>>>> 2018-05-10 03:24:20,748+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH,
>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return:
>>>>> org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@4a46066f,
>>>>> log id: 7550e5c
>>>>> 2018-05-10 03:24:20,749+02 ERROR
>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] 
>>>>> (DefaultQuartzScheduler7)
>>>>> [43f4eaec] Error while refreshing brick statuses for volume 'volume1' of
>>>>> cluster 'C6220': null
>>>>> 2018-05-10 03:24:20,750+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>> (DefaultQuartzScheduler8) [7715ceda] START,
>>>>> GlusterServersListVDSCommand(HostName = n2.itsmart.cloud,
>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>> log id: 120cc68d
>>>>> 2018-05-10 03:24:20,930+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, GlusterServersListVDSCommand,
>>>>> return: [ [ http://10.101.0.2/24:CONNECTED | 10.101.0.2/24:CONNECTED ] ,
>>>>> n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], 
>>>>> log
>>>>> id: 120cc68d
>>>>> 2018-05-10 03:24:20,949+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>> (DefaultQuartzScheduler8) [7715ceda] START,
>>>>> GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud,
>>>>> GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>> log id: 118aa264
>>>>> 2018-05-10 03:24:21,048+02 WARN
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>>> '10.104.0.1:/gluster/brick/brick1' of volume
>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster
>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>> 2018-05-10 03:24:21,055+02 WARN
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>>> '10.104.0.1:/gluster/brick/brick2' of volume
>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster
>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>> 2018-05-10 03:24:21,061+02 WARN
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>>> '10.104.0.1:/gluster/brick/brick3' of volume
>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster
>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>> 2018-05-10 03:24:21,067+02 WARN
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>>> '10.104.0.1:/gluster2/brick/brick1' of volume
>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster
>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>> 2018-05-10 03:24:21,074+02 WARN
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>>> '10.104.0.1:/gluster2/brick/brick2' of volume
>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster
>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>> 2018-05-10 03:24:21,080+02 WARN
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick
>>>>> '10.104.0.1:/gluster2/brick/brick3' of volume
>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster
>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>> 2018-05-10 03:24:21,081+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, GlusterVolumesListVDSCommand,
>>>>> return:
>>>>> {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d,
>>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.g
>>>>> luster.GlusterVolumeEntity@f88c521b}, log id: 118aa264

>>>>> 2018-05-10 11:59:26,047+02 ERROR
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler4) [400fa486] Command
>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n1.itsmart.cloud,
>>>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})'
>>>>> execution failed: null
>>>>> 2018-05-10 11:59:26,047+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH,
>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 14a71ef0
>>>>> 2018-05-10 11:59:26,048+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler4) [400fa486] START,
>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud,
>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}),
>>>>> log id: 28d9e255
>>>>> 2018-05-10 11:59:26,051+02 ERROR
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler4) [400fa486] Command
>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud,
>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})'
>>>>> execution failed: null
>>>>> 2018-05-10 11:59:26,051+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH,
>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 28d9e255
>>>>> 2018-05-10 11:59:26,052+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler4) [400fa486] START,
>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud,
>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>> log id: 4a7b280e
>>>>> 2018-05-10 11:59:26,054+02 ERROR
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler4) [400fa486] Command
>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud,
>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})'
>>>>> execution failed: null
>>>>> 2018-05-10 11:59:26,054+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH,
>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 4a7b280e
>>>>> 2018-05-10 11:59:26,055+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler4) [400fa486] START,
>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud,
>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}),
>>>>> log id: 18adc534
>>>>> 2018-05-10 11:59:26,057+02 ERROR
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler4) [400fa486] Command
>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud,
>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})'
>>>>> execution failed: null
>>>>> 2018-05-10 11:59:26,057+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH,
>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 18adc534
>>>>> 2018-05-10 11:59:26,058+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>> (DefaultQuartzScheduler4) [400fa486] START,
>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = n3.itsmart.cloud,
>>>>> GlusterVolumeAdvancedDetailsVDSParameters:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec',
>>>>> volumeName='volume1'}), log id: 3451084f
>>>>> 2018-05-10 11:59:28,050+02 INFO
>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>> (DefaultQuartzScheduler3)
>>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>> sharedLocks=''}'
>>>>> 2018-05-10 11:59:28,060+02 INFO
>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>> (DefaultQuartzScheduler3)
>>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>> sharedLocks=''}'
>>>>> 2018-05-10 11:59:28,062+02 INFO
>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>> (DefaultQuartzScheduler3)
>>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>> sharedLocks=''}'
>>>>> 2018-05-10 11:59:31,054+02 INFO
>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>> (DefaultQuartzScheduler3)
>>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>> sharedLocks=''}'
>>>>> 2018-05-10 11:59:31,054+02 INFO
>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>> (DefaultQuartzScheduler3)
>>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>> sharedLocks=''}'
>>>>> 2018-05-10 11:59:31,062+02 INFO
>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>> (DefaultQuartzScheduler3)
>>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>> sharedLocks=''}'
>>>>> 2018-05-10 11:59:31,064+02 INFO
>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] 
>>>>> (DefaultQuartzScheduler3)
>>>>> [2eb1c389] Failed to acquire lock and wait lock
>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
>>>>> sharedLocks=''}'
>>>>> 2018-05-10 11:59:31,465+02 INFO
>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH,
>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return:
>>>>> org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@3f1b7f43,
>>>>> log id: 3451084f
>>>>> 2018-05-10 11:59:31,466+02 ERROR
>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] 
>>>>> (DefaultQuartzScheduler4)
>>>>> [400fa486] Error while refreshing brick statuses for volume 'volume1' of
>>>>> cluster 'C6220': null

>>>>> R
>>>>> Tibor

>>>>> ----- 2018. máj.. 10., 11:43, Sahina Bose < [ mailto:[email protected] |
>>>>> [email protected] ] > írta:

>>>>>> This doesn't affect the monitoring of state.
>>>>>> Any errors in vdsm.log?
>>>>>> Or errors in engine.log of the form "Error while refreshing brick 
>>>>>> statuses for
>>>>>> volume"

>>>>>> On Thu, May 10, 2018 at 2:33 PM, Demeter Tibor < [ 
>>>>>> mailto:[email protected] |
>>>>>> [email protected] ] > wrote:

>>>>>>> Hi,

>>>>>>> Thank you for your fast reply :)

>>>>>>> 2018-05-10 11:01:51,574+02 INFO
>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START,
>>>>>>> GlusterServersListVDSCommand(HostName = n2.itsmart.cloud,
>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>>> log id: 39adbbb8
>>>>>>> 2018-05-10 11:01:51,768+02 INFO
>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, 
>>>>>>> GlusterServersListVDSCommand,
>>>>>>> return: [ [ http://10.101.0.2/24:CONNECTED | 10.101.0.2/24:CONNECTED ] ,
>>>>>>> n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED, 
>>>>>>> 10.104.0.4:CONNECTED], log
>>>>>>> id: 39adbbb8
>>>>>>> 2018-05-10 11:01:51,788+02 INFO
>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START,
>>>>>>> GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud,
>>>>>>> GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>>>>>>> log id: 738a7261
>>>>>>> 2018-05-10 11:01:51,892+02 WARN
>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>>> '10.104.0.1:/gluster/brick/brick1' of volume
>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no 
>>>>>>> gluster
>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>> 2018-05-10 11:01:51,898+02 WARN
>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>>> '10.104.0.1:/gluster/brick/brick2' of volume
>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no 
>>>>>>> gluster
>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>> 2018-05-10 11:01:51,905+02 WARN
>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>>> '10.104.0.1:/gluster/brick/brick3' of volume
>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no 
>>>>>>> gluster
>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>> 2018-05-10 11:01:51,911+02 WARN
>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>>> '10.104.0.1:/gluster2/brick/brick1' of volume
>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no 
>>>>>>> gluster
>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>> 2018-05-10 11:01:51,917+02 WARN
>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>>> '10.104.0.1:/gluster2/brick/brick2' of volume
>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no 
>>>>>>> gluster
>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>> 2018-05-10 11:01:51,924+02 WARN
>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>>>>>>> '10.104.0.1:/gluster2/brick/brick3' of volume
>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no 
>>>>>>> gluster
>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>>>>>>> 2018-05-10 11:01:51,925+02 INFO
>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, 
>>>>>>> GlusterVolumesListVDSCommand,
>>>>>>> return:
>>>>>>> {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d,
>>>>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@f88c521b},
>>>>>>> log id: 738a7261

>>>>>>> This happening continuously.

>>>>>>> Thanks!
>>>>>>> Tibor

>>>>>>> ----- 2018. máj.. 10., 10:56, Sahina Bose < [ mailto:[email protected] |
>>>>>>> [email protected] ] > írta:

>>>>>>>> Could you check the engine.log if there are errors related to getting
>>>>>>>> GlusterVolumeAdvancedDetails ?

>>>>>>>> On Thu, May 10, 2018 at 2:02 PM, Demeter Tibor < [ 
>>>>>>>> mailto:[email protected] |
>>>>>>>> [email protected] ] > wrote:

>>>>>>>>> Dear Ovirt Users,
>>>>>>>>> I've followed up the self-hosted-engine upgrade documentation, I 
>>>>>>>>> upgraded my 4.1
>>>>>>>>> system to 4.2.3.
>>>>>>>>> I upgaded the first node with yum upgrade, it seems working now fine. 
>>>>>>>>> But since
>>>>>>>>> upgrade, the gluster informations seems to displayed incorrect on the 
>>>>>>>>> admin
>>>>>>>>> panel. The volume yellow, and there are red bricks from that node.
>>>>>>>>> I've checked in console, I think my gluster is not degraded:

>>>>>>>>> root@n1 ~]# gluster volume list
>>>>>>>>> volume1
>>>>>>>>> volume2
>>>>>>>>> [root@n1 ~]# gluster volume info
>>>>>>>>> Volume Name: volume1
>>>>>>>>> Type: Distributed-Replicate
>>>>>>>>> Volume ID: e0f568fa-987c-4f5c-b853-01bce718ee27
>>>>>>>>> Status: Started
>>>>>>>>> Snapshot Count: 0
>>>>>>>>> Number of Bricks: 3 x 3 = 9
>>>>>>>>> Transport-type: tcp
>>>>>>>>> Bricks:
>>>>>>>>> Brick1: 10.104.0.1:/gluster/brick/brick1
>>>>>>>>> Brick2: 10.104.0.2:/gluster/brick/brick1
>>>>>>>>> Brick3: 10.104.0.3:/gluster/brick/brick1
>>>>>>>>> Brick4: 10.104.0.1:/gluster/brick/brick2
>>>>>>>>> Brick5: 10.104.0.2:/gluster/brick/brick2
>>>>>>>>> Brick6: 10.104.0.3:/gluster/brick/brick2
>>>>>>>>> Brick7: 10.104.0.1:/gluster/brick/brick3
>>>>>>>>> Brick8: 10.104.0.2:/gluster/brick/brick3
>>>>>>>>> Brick9: 10.104.0.3:/gluster/brick/brick3
>>>>>>>>> Options Reconfigured:
>>>>>>>>> transport.address-family: inet
>>>>>>>>> performance.readdir-ahead: on
>>>>>>>>> nfs.disable: on
>>>>>>>>> storage.owner-uid: 36
>>>>>>>>> storage.owner-gid: 36
>>>>>>>>> performance.quick-read: off
>>>>>>>>> performance.read-ahead: off
>>>>>>>>> performance.io-cache: off
>>>>>>>>> performance.stat-prefetch: off
>>>>>>>>> performance.low-prio-threads: 32
>>>>>>>>> network.remote-dio: enable
>>>>>>>>> cluster.eager-lock: enable
>>>>>>>>> cluster.quorum-type: auto
>>>>>>>>> cluster.server-quorum-type: server
>>>>>>>>> cluster.data-self-heal-algorithm: full
>>>>>>>>> cluster.locking-scheme: granular
>>>>>>>>> cluster.shd-max-threads: 8
>>>>>>>>> cluster.shd-wait-qlength: 10000
>>>>>>>>> features.shard: on
>>>>>>>>> user.cifs: off
>>>>>>>>> server.allow-insecure: on
>>>>>>>>> Volume Name: volume2
>>>>>>>>> Type: Distributed-Replicate
>>>>>>>>> Volume ID: 68cfb061-1320-4042-abcd-9228da23c0c8
>>>>>>>>> Status: Started
>>>>>>>>> Snapshot Count: 0
>>>>>>>>> Number of Bricks: 3 x 3 = 9
>>>>>>>>> Transport-type: tcp
>>>>>>>>> Bricks:
>>>>>>>>> Brick1: 10.104.0.1:/gluster2/brick/brick1
>>>>>>>>> Brick2: 10.104.0.2:/gluster2/brick/brick1
>>>>>>>>> Brick3: 10.104.0.3:/gluster2/brick/brick1
>>>>>>>>> Brick4: 10.104.0.1:/gluster2/brick/brick2
>>>>>>>>> Brick5: 10.104.0.2:/gluster2/brick/brick2
>>>>>>>>> Brick6: 10.104.0.3:/gluster2/brick/brick2
>>>>>>>>> Brick7: 10.104.0.1:/gluster2/brick/brick3
>>>>>>>>> Brick8: 10.104.0.2:/gluster2/brick/brick3
>>>>>>>>> Brick9: 10.104.0.3:/gluster2/brick/brick3
>>>>>>>>> Options Reconfigured:
>>>>>>>>> nfs.disable: on
>>>>>>>>> performance.readdir-ahead: on
>>>>>>>>> transport.address-family: inet
>>>>>>>>> cluster.quorum-type: auto
>>>>>>>>> network.ping-timeout: 10
>>>>>>>>> auth.allow: *
>>>>>>>>> performance.quick-read: off
>>>>>>>>> performance.read-ahead: off
>>>>>>>>> performance.io-cache: off
>>>>>>>>> performance.stat-prefetch: off
>>>>>>>>> performance.low-prio-threads: 32
>>>>>>>>> network.remote-dio: enable
>>>>>>>>> cluster.eager-lock: enable
>>>>>>>>> cluster.server-quorum-type: server
>>>>>>>>> cluster.data-self-heal-algorithm: full
>>>>>>>>> cluster.locking-scheme: granular
>>>>>>>>> cluster.shd-max-threads: 8
>>>>>>>>> cluster.shd-wait-qlength: 10000
>>>>>>>>> features.shard: on
>>>>>>>>> user.cifs: off
>>>>>>>>> storage.owner-uid: 36
>>>>>>>>> storage.owner-gid: 36
>>>>>>>>> server.allow-insecure: on
>>>>>>>>> [root@n1 ~]# gluster volume status
>>>>>>>>> Status of volume: volume1
>>>>>>>>> Gluster process TCP Port RDMA Port Online Pid
>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick1 49152 0 Y 3464
>>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick1 49152 0 Y 68937
>>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick1 49161 0 Y 94506
>>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick2 49153 0 Y 3457
>>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick2 49153 0 Y 68943
>>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick2 49162 0 Y 94514
>>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick3 49154 0 Y 3465
>>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick3 49154 0 Y 68949
>>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick3 49163 0 Y 94520
>>>>>>>>> Self-heal Daemon on localhost N/A N/A Y 54356
>>>>>>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y 962
>>>>>>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977
>>>>>>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603
>>>>>>>>> Task Status of Volume volume1
>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>> There are no active volume tasks
>>>>>>>>> Status of volume: volume2
>>>>>>>>> Gluster process TCP Port RDMA Port Online Pid
>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick1 49155 0 Y 3852
>>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick1 49158 0 Y 68955
>>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick1 49164 0 Y 94527
>>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick2 49156 0 Y 3851
>>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick2 49159 0 Y 68961
>>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick2 49165 0 Y 94533
>>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick3 49157 0 Y 3883
>>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick3 49160 0 Y 68968
>>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick3 49166 0 Y 94541
>>>>>>>>> Self-heal Daemon on localhost N/A N/A Y 54356
>>>>>>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y 962
>>>>>>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977
>>>>>>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603
>>>>>>>>> Task Status of Volume volume2
>>>>>>>>> ------------------------------------------------------------------------------
>>>>>>>>> There are no active volume tasks
>>>>>>>>> I think ovirt can't read valid informations about gluster.
>>>>>>>>> I can't contiune upgrade of other hosts until this problem exist.

>>>>>>>>> Please help me:)

>>>>>>>>> Thanks

>>>>>>>>> Regards,

>>>>>>>>> Tibor

>>>>>>>>> _______________________________________________
>>>>>>>>> Users mailing list -- [ mailto:[email protected] | [email protected] ]
>>>>>>>>> To unsubscribe send an email to [ mailto:[email protected] |
>>>>>>>>> [email protected] ]

>>> _______________________________________________
>>> Users mailing list -- [ mailto:[email protected] | [email protected] ]
>>> To unsubscribe send an email to [ mailto:[email protected] |
>>> [email protected] ]

>> _______________________________________________
>> Users mailing list -- [ mailto:[email protected] | [email protected] ]
>> To unsubscribe send an email to [ mailto:[email protected] |
>> [email protected] ]

> --
> Doug

> _______________________________________________
> Users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to