Hi,
I found this:
2018-05-10 03:24:19,096+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
(DefaultQuartzScheduler7) [43f4eaec] FINISH,
GetGlusterVolumeAdvancedDetailsVDSCommand, return:
org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@ca97448e,
log id: 347435ae
2018-05-10 03:24:19,097+02 ERROR
[org.ovirt.engine.core.bll.gluster.GlusterSyncJob] (DefaultQuartzScheduler7)
[43f4eaec] Error while refreshing brick statuses for volume 'volume2' of
cluster 'C6220': null
2018-05-10 03:24:19,097+02 INFO
[org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler8)
[7715ceda] Failed to acquire lock and wait lock
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
sharedLocks=''}'
2018-05-10 03:24:19,104+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler7) [43f4eaec] START,
GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud,
VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}),
log id: 6908121d
2018-05-10 03:24:19,106+02 ERROR
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler7) [43f4eaec] Command
'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud,
VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})'
execution failed: null
2018-05-10 03:24:19,106+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler7) [43f4eaec] FINISH,
GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6908121d
2018-05-10 03:24:19,107+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler7) [43f4eaec] START,
GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n1.itsmart.cloud,
VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'}),
log id: 735c6a5f
2018-05-10 03:24:19,109+02 ERROR
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler7) [43f4eaec] Command
'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n1.itsmart.cloud,
VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})'
execution failed: null
2018-05-10 03:24:19,109+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler7) [43f4eaec] FINISH,
GetGlusterLocalLogicalVolumeListVDSCommand, log id: 735c6a5f
2018-05-10 03:24:19,110+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler7) [43f4eaec] START,
GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud,
VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
log id: 6f9e9f58
2018-05-10 03:24:19,112+02 ERROR
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler7) [43f4eaec] Command
'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud,
VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})'
execution failed: null
2018-05-10 03:24:19,112+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler7) [43f4eaec] FINISH,
GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6f9e9f58
2018-05-10 03:24:19,113+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler7) [43f4eaec] START,
GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud,
VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}),
log id: 2ee46967
2018-05-10 03:24:19,115+02 ERROR
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler7) [43f4eaec] Command
'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud,
VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})'
execution failed: null
2018-05-10 03:24:19,116+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler7) [43f4eaec] FINISH,
GetGlusterLocalLogicalVolumeListVDSCommand, log id: 2ee46967
2018-05-10 03:24:19,117+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
(DefaultQuartzScheduler7) [43f4eaec] START,
GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = n1.itsmart.cloud,
GlusterVolumeAdvancedDetailsVDSParameters:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57',
volumeName='volume1'}), log id: 7550e5c
2018-05-10 03:24:20,748+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
(DefaultQuartzScheduler7) [43f4eaec] FINISH,
GetGlusterVolumeAdvancedDetailsVDSCommand, return:
org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@4a46066f,
log id: 7550e5c
2018-05-10 03:24:20,749+02 ERROR
[org.ovirt.engine.core.bll.gluster.GlusterSyncJob] (DefaultQuartzScheduler7)
[43f4eaec] Error while refreshing brick statuses for volume 'volume1' of
cluster 'C6220': null
2018-05-10 03:24:20,750+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler8) [7715ceda] START,
GlusterServersListVDSCommand(HostName = n2.itsmart.cloud,
VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
log id: 120cc68d
2018-05-10 03:24:20,930+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
(DefaultQuartzScheduler8) [7715ceda] FINISH, GlusterServersListVDSCommand,
return: [10.101.0.2/24:CONNECTED, n1.cloudata.local:CONNECTED,
10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], log id: 120cc68d
2018-05-10 03:24:20,949+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler8) [7715ceda] START,
GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud,
GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
log id: 118aa264
2018-05-10 03:24:21,048+02 WARN
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
(DefaultQuartzScheduler8) [7715ceda] Could not associate brick
'10.104.0.1:/gluster/brick/brick1' of volume
'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster
network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 03:24:21,055+02 WARN
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
(DefaultQuartzScheduler8) [7715ceda] Could not associate brick
'10.104.0.1:/gluster/brick/brick2' of volume
'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster
network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 03:24:21,061+02 WARN
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
(DefaultQuartzScheduler8) [7715ceda] Could not associate brick
'10.104.0.1:/gluster/brick/brick3' of volume
'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster
network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 03:24:21,067+02 WARN
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
(DefaultQuartzScheduler8) [7715ceda] Could not associate brick
'10.104.0.1:/gluster2/brick/brick1' of volume
'68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster
network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 03:24:21,074+02 WARN
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
(DefaultQuartzScheduler8) [7715ceda] Could not associate brick
'10.104.0.1:/gluster2/brick/brick2' of volume
'68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster
network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 03:24:21,080+02 WARN
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
(DefaultQuartzScheduler8) [7715ceda] Could not associate brick
'10.104.0.1:/gluster2/brick/brick3' of volume
'68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster
network found in cluster '59c10db3-0324-0320-0120-000000000339'
2018-05-10 03:24:21,081+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
(DefaultQuartzScheduler8) [7715ceda] FINISH, GlusterVolumesListVDSCommand,
return:
{68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d,
e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.g
luster.GlusterVolumeEntity@f88c521b}, log id: 118aa264
2018-05-10 11:59:26,047+02 ERROR
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler4) [400fa486] Command
'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n1.itsmart.cloud,
VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})'
execution failed: null
2018-05-10 11:59:26,047+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler4) [400fa486] FINISH,
GetGlusterLocalLogicalVolumeListVDSCommand, log id: 14a71ef0
2018-05-10 11:59:26,048+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler4) [400fa486] START,
GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud,
VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}),
log id: 28d9e255
2018-05-10 11:59:26,051+02 ERROR
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler4) [400fa486] Command
'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud,
VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})'
execution failed: null
2018-05-10 11:59:26,051+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler4) [400fa486] FINISH,
GetGlusterLocalLogicalVolumeListVDSCommand, log id: 28d9e255
2018-05-10 11:59:26,052+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler4) [400fa486] START,
GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud,
VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
log id: 4a7b280e
2018-05-10 11:59:26,054+02 ERROR
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler4) [400fa486] Command
'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud,
VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})'
execution failed: null
2018-05-10 11:59:26,054+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler4) [400fa486] FINISH,
GetGlusterLocalLogicalVolumeListVDSCommand, log id: 4a7b280e
2018-05-10 11:59:26,055+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler4) [400fa486] START,
GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud,
VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}),
log id: 18adc534
2018-05-10 11:59:26,057+02 ERROR
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler4) [400fa486] Command
'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud,
VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})'
execution failed: null
2018-05-10 11:59:26,057+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand]
(DefaultQuartzScheduler4) [400fa486] FINISH,
GetGlusterLocalLogicalVolumeListVDSCommand, log id: 18adc534
2018-05-10 11:59:26,058+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
(DefaultQuartzScheduler4) [400fa486] START,
GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = n3.itsmart.cloud,
GlusterVolumeAdvancedDetailsVDSParameters:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec',
volumeName='volume1'}), log id: 3451084f
2018-05-10 11:59:28,050+02 INFO
[org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3)
[2eb1c389] Failed to acquire lock and wait lock
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
sharedLocks=''}'
2018-05-10 11:59:28,060+02 INFO
[org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3)
[2eb1c389] Failed to acquire lock and wait lock
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
sharedLocks=''}'
2018-05-10 11:59:28,062+02 INFO
[org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3)
[2eb1c389] Failed to acquire lock and wait lock
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
sharedLocks=''}'
2018-05-10 11:59:31,054+02 INFO
[org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3)
[2eb1c389] Failed to acquire lock and wait lock
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
sharedLocks=''}'
2018-05-10 11:59:31,054+02 INFO
[org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3)
[2eb1c389] Failed to acquire lock and wait lock
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
sharedLocks=''}'
2018-05-10 11:59:31,062+02 INFO
[org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3)
[2eb1c389] Failed to acquire lock and wait lock
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
sharedLocks=''}'
2018-05-10 11:59:31,064+02 INFO
[org.ovirt.engine.core.bll.lock.InMemoryLockManager] (DefaultQuartzScheduler3)
[2eb1c389] Failed to acquire lock and wait lock
'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]',
sharedLocks=''}'
2018-05-10 11:59:31,465+02 INFO
[org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand]
(DefaultQuartzScheduler4) [400fa486] FINISH,
GetGlusterVolumeAdvancedDetailsVDSCommand, return:
org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@3f1b7f43,
log id: 3451084f
2018-05-10 11:59:31,466+02 ERROR
[org.ovirt.engine.core.bll.gluster.GlusterSyncJob] (DefaultQuartzScheduler4)
[400fa486] Error while refreshing brick statuses for volume 'volume1' of
cluster 'C6220': null
R
Tibor
----- 2018. máj.. 10., 11:43, Sahina Bose <[email protected]> írta:
> This doesn't affect the monitoring of state.
> Any errors in vdsm.log?
> Or errors in engine.log of the form "Error while refreshing brick statuses for
> volume"
> On Thu, May 10, 2018 at 2:33 PM, Demeter Tibor < [ mailto:[email protected]
> |
> [email protected] ] > wrote:
>> Hi,
>> Thank you for your fast reply :)
>> 2018-05-10 11:01:51,574+02 INFO
>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>> (DefaultQuartzScheduler6) [7f01fc2d] START,
>> GlusterServersListVDSCommand(HostName = n2.itsmart.cloud,
>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>> log id: 39adbbb8
>> 2018-05-10 11:01:51,768+02 INFO
>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, GlusterServersListVDSCommand,
>> return: [ [ http://10.101.0.2/24:CONNECTED | 10.101.0.2/24:CONNECTED ] ,
>> n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], log
>> id: 39adbbb8
>> 2018-05-10 11:01:51,788+02 INFO
>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>> (DefaultQuartzScheduler6) [7f01fc2d] START,
>> GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud,
>> GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
>> log id: 738a7261
>> 2018-05-10 11:01:51,892+02 WARN
>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>> '10.104.0.1:/gluster/brick/brick1' of volume
>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster
>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>> 2018-05-10 11:01:51,898+02 WARN
>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>> '10.104.0.1:/gluster/brick/brick2' of volume
>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster
>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>> 2018-05-10 11:01:51,905+02 WARN
>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>> '10.104.0.1:/gluster/brick/brick3' of volume
>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster
>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>> 2018-05-10 11:01:51,911+02 WARN
>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>> '10.104.0.1:/gluster2/brick/brick1' of volume
>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster
>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>> 2018-05-10 11:01:51,917+02 WARN
>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>> '10.104.0.1:/gluster2/brick/brick2' of volume
>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster
>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>> 2018-05-10 11:01:51,924+02 WARN
>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn]
>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick
>> '10.104.0.1:/gluster2/brick/brick3' of volume
>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster
>> network found in cluster '59c10db3-0324-0320-0120-000000000339'
>> 2018-05-10 11:01:51,925+02 INFO
>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, GlusterVolumesListVDSCommand,
>> return:
>> {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d,
>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@f88c521b},
>> log id: 738a7261
>> This happening continuously.
>> Thanks!
>> Tibor
>> ----- 2018. máj.. 10., 10:56, Sahina Bose < [ mailto:[email protected] |
>> [email protected] ] > írta:
>>> Could you check the engine.log if there are errors related to getting
>>> GlusterVolumeAdvancedDetails ?
>>> On Thu, May 10, 2018 at 2:02 PM, Demeter Tibor < [
>>> mailto:[email protected] |
>>> [email protected] ] > wrote:
>>>> Dear Ovirt Users,
>>>> I've followed up the self-hosted-engine upgrade documentation, I upgraded
>>>> my 4.1
>>>> system to 4.2.3.
>>>> I upgaded the first node with yum upgrade, it seems working now fine. But
>>>> since
>>>> upgrade, the gluster informations seems to displayed incorrect on the admin
>>>> panel. The volume yellow, and there are red bricks from that node.
>>>> I've checked in console, I think my gluster is not degraded:
>>>> root@n1 ~]# gluster volume list
>>>> volume1
>>>> volume2
>>>> [root@n1 ~]# gluster volume info
>>>> Volume Name: volume1
>>>> Type: Distributed-Replicate
>>>> Volume ID: e0f568fa-987c-4f5c-b853-01bce718ee27
>>>> Status: Started
>>>> Snapshot Count: 0
>>>> Number of Bricks: 3 x 3 = 9
>>>> Transport-type: tcp
>>>> Bricks:
>>>> Brick1: 10.104.0.1:/gluster/brick/brick1
>>>> Brick2: 10.104.0.2:/gluster/brick/brick1
>>>> Brick3: 10.104.0.3:/gluster/brick/brick1
>>>> Brick4: 10.104.0.1:/gluster/brick/brick2
>>>> Brick5: 10.104.0.2:/gluster/brick/brick2
>>>> Brick6: 10.104.0.3:/gluster/brick/brick2
>>>> Brick7: 10.104.0.1:/gluster/brick/brick3
>>>> Brick8: 10.104.0.2:/gluster/brick/brick3
>>>> Brick9: 10.104.0.3:/gluster/brick/brick3
>>>> Options Reconfigured:
>>>> transport.address-family: inet
>>>> performance.readdir-ahead: on
>>>> nfs.disable: on
>>>> storage.owner-uid: 36
>>>> storage.owner-gid: 36
>>>> performance.quick-read: off
>>>> performance.read-ahead: off
>>>> performance.io-cache: off
>>>> performance.stat-prefetch: off
>>>> performance.low-prio-threads: 32
>>>> network.remote-dio: enable
>>>> cluster.eager-lock: enable
>>>> cluster.quorum-type: auto
>>>> cluster.server-quorum-type: server
>>>> cluster.data-self-heal-algorithm: full
>>>> cluster.locking-scheme: granular
>>>> cluster.shd-max-threads: 8
>>>> cluster.shd-wait-qlength: 10000
>>>> features.shard: on
>>>> user.cifs: off
>>>> server.allow-insecure: on
>>>> Volume Name: volume2
>>>> Type: Distributed-Replicate
>>>> Volume ID: 68cfb061-1320-4042-abcd-9228da23c0c8
>>>> Status: Started
>>>> Snapshot Count: 0
>>>> Number of Bricks: 3 x 3 = 9
>>>> Transport-type: tcp
>>>> Bricks:
>>>> Brick1: 10.104.0.1:/gluster2/brick/brick1
>>>> Brick2: 10.104.0.2:/gluster2/brick/brick1
>>>> Brick3: 10.104.0.3:/gluster2/brick/brick1
>>>> Brick4: 10.104.0.1:/gluster2/brick/brick2
>>>> Brick5: 10.104.0.2:/gluster2/brick/brick2
>>>> Brick6: 10.104.0.3:/gluster2/brick/brick2
>>>> Brick7: 10.104.0.1:/gluster2/brick/brick3
>>>> Brick8: 10.104.0.2:/gluster2/brick/brick3
>>>> Brick9: 10.104.0.3:/gluster2/brick/brick3
>>>> Options Reconfigured:
>>>> nfs.disable: on
>>>> performance.readdir-ahead: on
>>>> transport.address-family: inet
>>>> cluster.quorum-type: auto
>>>> network.ping-timeout: 10
>>>> auth.allow: *
>>>> performance.quick-read: off
>>>> performance.read-ahead: off
>>>> performance.io-cache: off
>>>> performance.stat-prefetch: off
>>>> performance.low-prio-threads: 32
>>>> network.remote-dio: enable
>>>> cluster.eager-lock: enable
>>>> cluster.server-quorum-type: server
>>>> cluster.data-self-heal-algorithm: full
>>>> cluster.locking-scheme: granular
>>>> cluster.shd-max-threads: 8
>>>> cluster.shd-wait-qlength: 10000
>>>> features.shard: on
>>>> user.cifs: off
>>>> storage.owner-uid: 36
>>>> storage.owner-gid: 36
>>>> server.allow-insecure: on
>>>> [root@n1 ~]# gluster volume status
>>>> Status of volume: volume1
>>>> Gluster process TCP Port RDMA Port Online Pid
>>>> ------------------------------------------------------------------------------
>>>> Brick 10.104.0.1:/gluster/brick/brick1 49152 0 Y 3464
>>>> Brick 10.104.0.2:/gluster/brick/brick1 49152 0 Y 68937
>>>> Brick 10.104.0.3:/gluster/brick/brick1 49161 0 Y 94506
>>>> Brick 10.104.0.1:/gluster/brick/brick2 49153 0 Y 3457
>>>> Brick 10.104.0.2:/gluster/brick/brick2 49153 0 Y 68943
>>>> Brick 10.104.0.3:/gluster/brick/brick2 49162 0 Y 94514
>>>> Brick 10.104.0.1:/gluster/brick/brick3 49154 0 Y 3465
>>>> Brick 10.104.0.2:/gluster/brick/brick3 49154 0 Y 68949
>>>> Brick 10.104.0.3:/gluster/brick/brick3 49163 0 Y 94520
>>>> Self-heal Daemon on localhost N/A N/A Y 54356
>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y 962
>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977
>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603
>>>> Task Status of Volume volume1
>>>> ------------------------------------------------------------------------------
>>>> There are no active volume tasks
>>>> Status of volume: volume2
>>>> Gluster process TCP Port RDMA Port Online Pid
>>>> ------------------------------------------------------------------------------
>>>> Brick 10.104.0.1:/gluster2/brick/brick1 49155 0 Y 3852
>>>> Brick 10.104.0.2:/gluster2/brick/brick1 49158 0 Y 68955
>>>> Brick 10.104.0.3:/gluster2/brick/brick1 49164 0 Y 94527
>>>> Brick 10.104.0.1:/gluster2/brick/brick2 49156 0 Y 3851
>>>> Brick 10.104.0.2:/gluster2/brick/brick2 49159 0 Y 68961
>>>> Brick 10.104.0.3:/gluster2/brick/brick2 49165 0 Y 94533
>>>> Brick 10.104.0.1:/gluster2/brick/brick3 49157 0 Y 3883
>>>> Brick 10.104.0.2:/gluster2/brick/brick3 49160 0 Y 68968
>>>> Brick 10.104.0.3:/gluster2/brick/brick3 49166 0 Y 94541
>>>> Self-heal Daemon on localhost N/A N/A Y 54356
>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y 962
>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977
>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603
>>>> Task Status of Volume volume2
>>>> ------------------------------------------------------------------------------
>>>> There are no active volume tasks
>>>> I think ovirt can't read valid informations about gluster.
>>>> I can't contiune upgrade of other hosts until this problem exist.
>>>> Please help me:)
>>>> Thanks
>>>> Regards,
>>>> Tibor
>>>> _______________________________________________
>>>> Users mailing list -- [ mailto:[email protected] | [email protected] ]
>>>> To unsubscribe send an email to [ mailto:[email protected] |
>>>> [email protected] ]
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]