There's a bug here. Can you log one attaching this engine.log and also
vdsm.log & supervdsm.log from n3.itsmart.cloud

On Thu, May 10, 2018 at 3:35 PM, Demeter Tibor <tdeme...@itsmart.hu> wrote:

> Hi,
>
> I found this:
>
>
> 2018-05-10 03:24:19,096+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler7)
> [43f4eaec] FINISH, GetGlusterVolumeAdvancedDetailsVDSCommand, return:
> org.ovirt.engine.core.common.businessentities.gluster.
> GlusterVolumeAdvancedDetails@ca97448e, log id: 347435ae
> 2018-05-10 03:24:19,097+02 ERROR 
> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob]
> (DefaultQuartzScheduler7) [43f4eaec] Error while refreshing brick statuses
> for volume 'volume2' of cluster 'C6220': null
> 2018-05-10 03:24:19,097+02 INFO  
> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> (DefaultQuartzScheduler8) [7715ceda] Failed to acquire lock and wait lock
> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
> sharedLocks=''}'
> 2018-05-10 03:24:19,104+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7)
> [43f4eaec] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
> n4.itsmart.cloud, VdsIdVDSCommandParametersBase:
> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), log id: 6908121d
> 2018-05-10 03:24:19,106+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7)
> [43f4eaec] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
> n4.itsmart.cloud, VdsIdVDSCommandParametersBase:
> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' execution failed: null
> 2018-05-10 03:24:19,106+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7)
> [43f4eaec] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id:
> 6908121d
> 2018-05-10 03:24:19,107+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7)
> [43f4eaec] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
> n1.itsmart.cloud, VdsIdVDSCommandParametersBase:
> {hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'}), log id: 735c6a5f
> 2018-05-10 03:24:19,109+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7)
> [43f4eaec] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
> n1.itsmart.cloud, VdsIdVDSCommandParametersBase:
> {hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' execution failed: null
> 2018-05-10 03:24:19,109+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7)
> [43f4eaec] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id:
> 735c6a5f
> 2018-05-10 03:24:19,110+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7)
> [43f4eaec] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
> n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 6f9e9f58
> 2018-05-10 03:24:19,112+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7)
> [43f4eaec] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
> n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' execution failed: null
> 2018-05-10 03:24:19,112+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7)
> [43f4eaec] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id:
> 6f9e9f58
> 2018-05-10 03:24:19,113+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7)
> [43f4eaec] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
> n3.itsmart.cloud, VdsIdVDSCommandParametersBase:
> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), log id: 2ee46967
> 2018-05-10 03:24:19,115+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7)
> [43f4eaec] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
> n3.itsmart.cloud, VdsIdVDSCommandParametersBase:
> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' execution failed: null
> 2018-05-10 03:24:19,116+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7)
> [43f4eaec] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id:
> 2ee46967
> 2018-05-10 03:24:19,117+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler7)
> [43f4eaec] START, GetGlusterVolumeAdvancedDetailsVDSCommand(HostName =
> n1.itsmart.cloud, GlusterVolumeAdvancedDetailsVD
> SParameters:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57',
> volumeName='volume1'}), log id: 7550e5c
> 2018-05-10 03:24:20,748+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler7)
> [43f4eaec] FINISH, GetGlusterVolumeAdvancedDetailsVDSCommand, return:
> org.ovirt.engine.core.common.businessentities.gluster.
> GlusterVolumeAdvancedDetails@4a46066f, log id: 7550e5c
> 2018-05-10 03:24:20,749+02 ERROR 
> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob]
> (DefaultQuartzScheduler7) [43f4eaec] Error while refreshing brick statuses
> for volume 'volume1' of cluster 'C6220': null
> 2018-05-10 03:24:20,750+02 INFO  
> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
> (DefaultQuartzScheduler8) [7715ceda] START, 
> GlusterServersListVDSCommand(HostName
> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 120cc68d
> 2018-05-10 03:24:20,930+02 INFO  
> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
> (DefaultQuartzScheduler8) [7715ceda] FINISH, GlusterServersListVDSCommand,
> return: [10.101.0.2/24:CONNECTED, n1.cloudata.local:CONNECTED, 
> 10.104.0.3:CONNECTED,
> 10.104.0.4:CONNECTED], log id: 120cc68d
> 2018-05-10 03:24:20,949+02 INFO  
> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
> (DefaultQuartzScheduler8) [7715ceda] START, 
> GlusterVolumesListVDSCommand(HostName
> = n2.itsmart.cloud, GlusterVolumesListVDSParameter
> s:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 118aa264
> 2018-05-10 03:24:21,048+02 WARN  
> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1:
> /gluster/brick/brick1' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27'
> with correct network as no gluster network found in cluster
> '59c10db3-0324-0320-0120-000000000339'
> 2018-05-10 03:24:21,055+02 WARN  
> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1:
> /gluster/brick/brick2' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27'
> with correct network as no gluster network found in cluster
> '59c10db3-0324-0320-0120-000000000339'
> 2018-05-10 03:24:21,061+02 WARN  
> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1:
> /gluster/brick/brick3' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27'
> with correct network as no gluster network found in cluster
> '59c10db3-0324-0320-0120-000000000339'
> 2018-05-10 03:24:21,067+02 WARN  
> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1:
> /gluster2/brick/brick1' of volume '68cfb061-1320-4042-abcd-9228da23c0c8'
> with correct network as no gluster network found in cluster
> '59c10db3-0324-0320-0120-000000000339'
> 2018-05-10 03:24:21,074+02 WARN  
> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1:
> /gluster2/brick/brick2' of volume '68cfb061-1320-4042-abcd-9228da23c0c8'
> with correct network as no gluster network found in cluster
> '59c10db3-0324-0320-0120-000000000339'
> 2018-05-10 03:24:21,080+02 WARN  
> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1:
> /gluster2/brick/brick3' of volume '68cfb061-1320-4042-abcd-9228da23c0c8'
> with correct network as no gluster network found in cluster
> '59c10db3-0324-0320-0120-000000000339'
> 2018-05-10 03:24:21,081+02 INFO  
> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
> (DefaultQuartzScheduler8) [7715ceda] FINISH, GlusterVolumesListVDSCommand,
> return: {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.
> core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d,
> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.
> core.common.businessentities.g
> luster.GlusterVolumeEntity@f88c521b}, log id: 118aa264
>
>
>
> 2018-05-10 11:59:26,047+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4)
> [400fa486] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
> n1.itsmart.cloud, VdsIdVDSCommandParametersBase:
> {hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' execution failed: null
> 2018-05-10 11:59:26,047+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4)
> [400fa486] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id:
> 14a71ef0
> 2018-05-10 11:59:26,048+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4)
> [400fa486] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
> n4.itsmart.cloud, VdsIdVDSCommandParametersBase:
> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), log id: 28d9e255
> 2018-05-10 11:59:26,051+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4)
> [400fa486] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
> n4.itsmart.cloud, VdsIdVDSCommandParametersBase:
> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' execution failed: null
> 2018-05-10 11:59:26,051+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4)
> [400fa486] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id:
> 28d9e255
> 2018-05-10 11:59:26,052+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4)
> [400fa486] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
> n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 4a7b280e
> 2018-05-10 11:59:26,054+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4)
> [400fa486] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
> n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' execution failed: null
> 2018-05-10 11:59:26,054+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4)
> [400fa486] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id:
> 4a7b280e
> 2018-05-10 11:59:26,055+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4)
> [400fa486] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
> n3.itsmart.cloud, VdsIdVDSCommandParametersBase:
> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), log id: 18adc534
> 2018-05-10 11:59:26,057+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4)
> [400fa486] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName =
> n3.itsmart.cloud, VdsIdVDSCommandParametersBase:
> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' execution failed: null
> 2018-05-10 11:59:26,057+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4)
> [400fa486] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id:
> 18adc534
> 2018-05-10 11:59:26,058+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler4)
> [400fa486] START, GetGlusterVolumeAdvancedDetailsVDSCommand(HostName =
> n3.itsmart.cloud, GlusterVolumeAdvancedDetailsVD
> SParameters:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec',
> volumeName='volume1'}), log id: 3451084f
> 2018-05-10 11:59:28,050+02 INFO  
> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
> sharedLocks=''}'
> 2018-05-10 11:59:28,060+02 INFO  
> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
> sharedLocks=''}'
> 2018-05-10 11:59:28,062+02 INFO  
> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
> sharedLocks=''}'
> 2018-05-10 11:59:31,054+02 INFO  
> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
> sharedLocks=''}'
> 2018-05-10 11:59:31,054+02 INFO  
> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
> sharedLocks=''}'
> 2018-05-10 11:59:31,062+02 INFO  
> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
> sharedLocks=''}'
> 2018-05-10 11:59:31,064+02 INFO  
> [org.ovirt.engine.core.bll.lock.InMemoryLockManager]
> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock
> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
> sharedLocks=''}'
> 2018-05-10 11:59:31,465+02 INFO  [org.ovirt.engine.core.vdsbroker.gluster.
> GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler4)
> [400fa486] FINISH, GetGlusterVolumeAdvancedDetailsVDSCommand, return:
> org.ovirt.engine.core.common.businessentities.gluster.
> GlusterVolumeAdvancedDetails@3f1b7f43, log id: 3451084f
> 2018-05-10 11:59:31,466+02 ERROR 
> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob]
> (DefaultQuartzScheduler4) [400fa486] Error while refreshing brick statuses
> for volume 'volume1' of cluster 'C6220': null
>
>
> R
> Tibor
>
> ----- 2018. máj.. 10., 11:43, Sahina Bose <sab...@redhat.com> írta:
>
> This doesn't affect the monitoring of state.
> Any errors in vdsm.log?
> Or errors in engine.log of the form "Error while refreshing brick statuses
> for volume"
>
> On Thu, May 10, 2018 at 2:33 PM, Demeter Tibor <tdeme...@itsmart.hu>
> wrote:
>
>> Hi,
>>
>> Thank you for your fast reply :)
>>
>>
>> 2018-05-10 11:01:51,574+02 INFO  [org.ovirt.engine.core.
>> vdsbroker.gluster.GlusterServersListVDSCommand]
>> (DefaultQuartzScheduler6) [7f01fc2d] START, 
>> GlusterServersListVDSCommand(HostName
>> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase:
>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 39adbbb8
>> 2018-05-10 11:01:51,768+02 INFO  [org.ovirt.engine.core.
>> vdsbroker.gluster.GlusterServersListVDSCommand]
>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, GlusterServersListVDSCommand,
>> return: [10.101.0.2/24:CONNECTED, n1.cloudata.local:CONNECTED,
>> 10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], log id: 39adbbb8
>> 2018-05-10 11:01:51,788+02 INFO  [org.ovirt.engine.core.
>> vdsbroker.gluster.GlusterVolumesListVDSCommand]
>> (DefaultQuartzScheduler6) [7f01fc2d] START, 
>> GlusterVolumesListVDSCommand(HostName
>> = n2.itsmart.cloud, GlusterVolumesListVDSParameter
>> s:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 738a7261
>> 2018-05-10 11:01:51,892+02 WARN  [org.ovirt.engine.core.
>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6)
>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster/brick/brick1'
>> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as
>> no gluster network found in cluster '59c10db3-0324-0320-0120-
>> 000000000339'
>> 2018-05-10 11:01:51,898+02 WARN  [org.ovirt.engine.core.
>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6)
>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster/brick/brick2'
>> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as
>> no gluster network found in cluster '59c10db3-0324-0320-0120-
>> 000000000339'
>> 2018-05-10 11:01:51,905+02 WARN  [org.ovirt.engine.core.
>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6)
>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster/brick/brick3'
>> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as
>> no gluster network found in cluster '59c10db3-0324-0320-0120-
>> 000000000339'
>> 2018-05-10 11:01:51,911+02 WARN  [org.ovirt.engine.core.
>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6)
>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster2/brick/brick1'
>> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as
>> no gluster network found in cluster '59c10db3-0324-0320-0120-
>> 000000000339'
>> 2018-05-10 11:01:51,917+02 WARN  [org.ovirt.engine.core.
>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6)
>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster2/brick/brick2'
>> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as
>> no gluster network found in cluster '59c10db3-0324-0320-0120-
>> 000000000339'
>> 2018-05-10 11:01:51,924+02 WARN  [org.ovirt.engine.core.
>> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6)
>> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster2/brick/brick3'
>> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as
>> no gluster network found in cluster '59c10db3-0324-0320-0120-
>> 000000000339'
>> 2018-05-10 11:01:51,925+02 INFO  [org.ovirt.engine.core.
>> vdsbroker.gluster.GlusterVolumesListVDSCommand]
>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, GlusterVolumesListVDSCommand,
>> return: {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.
>> core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d,
>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.
>> core.common.businessentities.gluster.GlusterVolumeEntity@f88c521b}, log
>> id: 738a7261
>>
>>
>> This happening continuously.
>>
>> Thanks!
>> Tibor
>>
>>
>>
>> ----- 2018. máj.. 10., 10:56, Sahina Bose <sab...@redhat.com> írta:
>>
>> Could you check the engine.log if there are errors related to getting
>> GlusterVolumeAdvancedDetails ?
>>
>> On Thu, May 10, 2018 at 2:02 PM, Demeter Tibor <tdeme...@itsmart.hu>
>> wrote:
>>
>>> Dear Ovirt Users,
>>> I've followed up the self-hosted-engine upgrade documentation, I
>>> upgraded my 4.1 system to 4.2.3.
>>> I upgaded the first node with yum upgrade, it seems working now fine.
>>> But since upgrade, the gluster informations seems to displayed incorrect on
>>> the admin panel. The volume yellow, and there are red bricks from that node.
>>> I've checked in console, I think my gluster is not degraded:
>>>
>>> root@n1 ~]# gluster volume list
>>> volume1
>>> volume2
>>> [root@n1 ~]# gluster volume info
>>>
>>> Volume Name: volume1
>>> Type: Distributed-Replicate
>>> Volume ID: e0f568fa-987c-4f5c-b853-01bce718ee27
>>> Status: Started
>>> Snapshot Count: 0
>>> Number of Bricks: 3 x 3 = 9
>>> Transport-type: tcp
>>> Bricks:
>>> Brick1: 10.104.0.1:/gluster/brick/brick1
>>> Brick2: 10.104.0.2:/gluster/brick/brick1
>>> Brick3: 10.104.0.3:/gluster/brick/brick1
>>> Brick4: 10.104.0.1:/gluster/brick/brick2
>>> Brick5: 10.104.0.2:/gluster/brick/brick2
>>> Brick6: 10.104.0.3:/gluster/brick/brick2
>>> Brick7: 10.104.0.1:/gluster/brick/brick3
>>> Brick8: 10.104.0.2:/gluster/brick/brick3
>>> Brick9: 10.104.0.3:/gluster/brick/brick3
>>> Options Reconfigured:
>>> transport.address-family: inet
>>> performance.readdir-ahead: on
>>> nfs.disable: on
>>> storage.owner-uid: 36
>>> storage.owner-gid: 36
>>> performance.quick-read: off
>>> performance.read-ahead: off
>>> performance.io-cache: off
>>> performance.stat-prefetch: off
>>> performance.low-prio-threads: 32
>>> network.remote-dio: enable
>>> cluster.eager-lock: enable
>>> cluster.quorum-type: auto
>>> cluster.server-quorum-type: server
>>> cluster.data-self-heal-algorithm: full
>>> cluster.locking-scheme: granular
>>> cluster.shd-max-threads: 8
>>> cluster.shd-wait-qlength: 10000
>>> features.shard: on
>>> user.cifs: off
>>> server.allow-insecure: on
>>>
>>> Volume Name: volume2
>>> Type: Distributed-Replicate
>>> Volume ID: 68cfb061-1320-4042-abcd-9228da23c0c8
>>> Status: Started
>>> Snapshot Count: 0
>>> Number of Bricks: 3 x 3 = 9
>>> Transport-type: tcp
>>> Bricks:
>>> Brick1: 10.104.0.1:/gluster2/brick/brick1
>>> Brick2: 10.104.0.2:/gluster2/brick/brick1
>>> Brick3: 10.104.0.3:/gluster2/brick/brick1
>>> Brick4: 10.104.0.1:/gluster2/brick/brick2
>>> Brick5: 10.104.0.2:/gluster2/brick/brick2
>>> Brick6: 10.104.0.3:/gluster2/brick/brick2
>>> Brick7: 10.104.0.1:/gluster2/brick/brick3
>>> Brick8: 10.104.0.2:/gluster2/brick/brick3
>>> Brick9: 10.104.0.3:/gluster2/brick/brick3
>>> Options Reconfigured:
>>> nfs.disable: on
>>> performance.readdir-ahead: on
>>> transport.address-family: inet
>>> cluster.quorum-type: auto
>>> network.ping-timeout: 10
>>> auth.allow: *
>>> performance.quick-read: off
>>> performance.read-ahead: off
>>> performance.io-cache: off
>>> performance.stat-prefetch: off
>>> performance.low-prio-threads: 32
>>> network.remote-dio: enable
>>> cluster.eager-lock: enable
>>> cluster.server-quorum-type: server
>>> cluster.data-self-heal-algorithm: full
>>> cluster.locking-scheme: granular
>>> cluster.shd-max-threads: 8
>>> cluster.shd-wait-qlength: 10000
>>> features.shard: on
>>> user.cifs: off
>>> storage.owner-uid: 36
>>> storage.owner-gid: 36
>>> server.allow-insecure: on
>>> [root@n1 ~]# gluster volume status
>>> Status of volume: volume1
>>> Gluster process                             TCP Port  RDMA Port  Online
>>> Pid
>>> ------------------------------------------------------------
>>> ------------------
>>> Brick 10.104.0.1:/gluster/brick/brick1      49152     0          Y
>>>  3464
>>> Brick 10.104.0.2:/gluster/brick/brick1      49152     0          Y
>>>  68937
>>> Brick 10.104.0.3:/gluster/brick/brick1      49161     0          Y
>>>  94506
>>> Brick 10.104.0.1:/gluster/brick/brick2      49153     0          Y
>>>  3457
>>> Brick 10.104.0.2:/gluster/brick/brick2      49153     0          Y
>>>  68943
>>> Brick 10.104.0.3:/gluster/brick/brick2      49162     0          Y
>>>  94514
>>> Brick 10.104.0.1:/gluster/brick/brick3      49154     0          Y
>>>  3465
>>> Brick 10.104.0.2:/gluster/brick/brick3      49154     0          Y
>>>  68949
>>> Brick 10.104.0.3:/gluster/brick/brick3      49163     0          Y
>>>  94520
>>> Self-heal Daemon on localhost               N/A       N/A        Y
>>>  54356
>>> Self-heal Daemon on 10.104.0.2              N/A       N/A        Y
>>>  962
>>> Self-heal Daemon on 10.104.0.3              N/A       N/A        Y
>>>  108977
>>> Self-heal Daemon on 10.104.0.4              N/A       N/A        Y
>>>  61603
>>>
>>> Task Status of Volume volume1
>>> ------------------------------------------------------------
>>> ------------------
>>> There are no active volume tasks
>>>
>>> Status of volume: volume2
>>> Gluster process                             TCP Port  RDMA Port  Online
>>> Pid
>>> ------------------------------------------------------------
>>> ------------------
>>> Brick 10.104.0.1:/gluster2/brick/brick1     49155     0          Y
>>>  3852
>>> Brick 10.104.0.2:/gluster2/brick/brick1     49158     0          Y
>>>  68955
>>> Brick 10.104.0.3:/gluster2/brick/brick1     49164     0          Y
>>>  94527
>>> Brick 10.104.0.1:/gluster2/brick/brick2     49156     0          Y
>>>  3851
>>> Brick 10.104.0.2:/gluster2/brick/brick2     49159     0          Y
>>>  68961
>>> Brick 10.104.0.3:/gluster2/brick/brick2     49165     0          Y
>>>  94533
>>> Brick 10.104.0.1:/gluster2/brick/brick3     49157     0          Y
>>>  3883
>>> Brick 10.104.0.2:/gluster2/brick/brick3     49160     0          Y
>>>  68968
>>> Brick 10.104.0.3:/gluster2/brick/brick3     49166     0          Y
>>>  94541
>>> Self-heal Daemon on localhost               N/A       N/A        Y
>>>  54356
>>> Self-heal Daemon on 10.104.0.2              N/A       N/A        Y
>>>  962
>>> Self-heal Daemon on 10.104.0.3              N/A       N/A        Y
>>>  108977
>>> Self-heal Daemon on 10.104.0.4              N/A       N/A        Y
>>>  61603
>>>
>>> Task Status of Volume volume2
>>> ------------------------------------------------------------
>>> ------------------
>>> There are no active volume tasks
>>>
>>> I think ovirt can't read valid informations about gluster.
>>> I can't contiune upgrade of other hosts until this problem exist.
>>>
>>> Please help me:)
>>>
>>>
>>> Thanks
>>>
>>> Regards,
>>>
>>> Tibor
>>>
>>>
>>> _______________________________________________
>>> Users mailing list -- users@ovirt.org
>>> To unsubscribe send an email to users-le...@ovirt.org
>>>
>>>
>>
>
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org

Reply via email to