There's a bug here. Can you log one attaching this engine.log and also vdsm.log & supervdsm.log from n3.itsmart.cloud
On Thu, May 10, 2018 at 3:35 PM, Demeter Tibor <tdeme...@itsmart.hu> wrote: > Hi, > > I found this: > > > 2018-05-10 03:24:19,096+02 INFO [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler7) > [43f4eaec] FINISH, GetGlusterVolumeAdvancedDetailsVDSCommand, return: > org.ovirt.engine.core.common.businessentities.gluster. > GlusterVolumeAdvancedDetails@ca97448e, log id: 347435ae > 2018-05-10 03:24:19,097+02 ERROR > [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] > (DefaultQuartzScheduler7) [43f4eaec] Error while refreshing brick statuses > for volume 'volume2' of cluster 'C6220': null > 2018-05-10 03:24:19,097+02 INFO > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > (DefaultQuartzScheduler8) [7715ceda] Failed to acquire lock and wait lock > 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', > sharedLocks=''}' > 2018-05-10 03:24:19,104+02 INFO [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) > [43f4eaec] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName = > n4.itsmart.cloud, VdsIdVDSCommandParametersBase: > {hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), log id: 6908121d > 2018-05-10 03:24:19,106+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) > [43f4eaec] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = > n4.itsmart.cloud, VdsIdVDSCommandParametersBase: > {hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' execution failed: null > 2018-05-10 03:24:19,106+02 INFO [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) > [43f4eaec] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: > 6908121d > 2018-05-10 03:24:19,107+02 INFO [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) > [43f4eaec] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName = > n1.itsmart.cloud, VdsIdVDSCommandParametersBase: > {hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'}), log id: 735c6a5f > 2018-05-10 03:24:19,109+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) > [43f4eaec] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = > n1.itsmart.cloud, VdsIdVDSCommandParametersBase: > {hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' execution failed: null > 2018-05-10 03:24:19,109+02 INFO [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) > [43f4eaec] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: > 735c6a5f > 2018-05-10 03:24:19,110+02 INFO [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) > [43f4eaec] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName = > n2.itsmart.cloud, VdsIdVDSCommandParametersBase: > {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 6f9e9f58 > 2018-05-10 03:24:19,112+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) > [43f4eaec] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = > n2.itsmart.cloud, VdsIdVDSCommandParametersBase: > {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' execution failed: null > 2018-05-10 03:24:19,112+02 INFO [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) > [43f4eaec] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: > 6f9e9f58 > 2018-05-10 03:24:19,113+02 INFO [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) > [43f4eaec] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName = > n3.itsmart.cloud, VdsIdVDSCommandParametersBase: > {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), log id: 2ee46967 > 2018-05-10 03:24:19,115+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) > [43f4eaec] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = > n3.itsmart.cloud, VdsIdVDSCommandParametersBase: > {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' execution failed: null > 2018-05-10 03:24:19,116+02 INFO [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler7) > [43f4eaec] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: > 2ee46967 > 2018-05-10 03:24:19,117+02 INFO [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler7) > [43f4eaec] START, GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = > n1.itsmart.cloud, GlusterVolumeAdvancedDetailsVD > SParameters:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57', > volumeName='volume1'}), log id: 7550e5c > 2018-05-10 03:24:20,748+02 INFO [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler7) > [43f4eaec] FINISH, GetGlusterVolumeAdvancedDetailsVDSCommand, return: > org.ovirt.engine.core.common.businessentities.gluster. > GlusterVolumeAdvancedDetails@4a46066f, log id: 7550e5c > 2018-05-10 03:24:20,749+02 ERROR > [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] > (DefaultQuartzScheduler7) [43f4eaec] Error while refreshing brick statuses > for volume 'volume1' of cluster 'C6220': null > 2018-05-10 03:24:20,750+02 INFO > [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] > (DefaultQuartzScheduler8) [7715ceda] START, > GlusterServersListVDSCommand(HostName > = n2.itsmart.cloud, VdsIdVDSCommandParametersBase: > {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 120cc68d > 2018-05-10 03:24:20,930+02 INFO > [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] > (DefaultQuartzScheduler8) [7715ceda] FINISH, GlusterServersListVDSCommand, > return: [10.101.0.2/24:CONNECTED, n1.cloudata.local:CONNECTED, > 10.104.0.3:CONNECTED, > 10.104.0.4:CONNECTED], log id: 120cc68d > 2018-05-10 03:24:20,949+02 INFO > [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] > (DefaultQuartzScheduler8) [7715ceda] START, > GlusterVolumesListVDSCommand(HostName > = n2.itsmart.cloud, GlusterVolumesListVDSParameter > s:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 118aa264 > 2018-05-10 03:24:21,048+02 WARN > [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] > (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1: > /gluster/brick/brick1' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' > with correct network as no gluster network found in cluster > '59c10db3-0324-0320-0120-000000000339' > 2018-05-10 03:24:21,055+02 WARN > [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] > (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1: > /gluster/brick/brick2' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' > with correct network as no gluster network found in cluster > '59c10db3-0324-0320-0120-000000000339' > 2018-05-10 03:24:21,061+02 WARN > [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] > (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1: > /gluster/brick/brick3' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' > with correct network as no gluster network found in cluster > '59c10db3-0324-0320-0120-000000000339' > 2018-05-10 03:24:21,067+02 WARN > [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] > (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1: > /gluster2/brick/brick1' of volume '68cfb061-1320-4042-abcd-9228da23c0c8' > with correct network as no gluster network found in cluster > '59c10db3-0324-0320-0120-000000000339' > 2018-05-10 03:24:21,074+02 WARN > [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] > (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1: > /gluster2/brick/brick2' of volume '68cfb061-1320-4042-abcd-9228da23c0c8' > with correct network as no gluster network found in cluster > '59c10db3-0324-0320-0120-000000000339' > 2018-05-10 03:24:21,080+02 WARN > [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] > (DefaultQuartzScheduler8) [7715ceda] Could not associate brick '10.104.0.1: > /gluster2/brick/brick3' of volume '68cfb061-1320-4042-abcd-9228da23c0c8' > with correct network as no gluster network found in cluster > '59c10db3-0324-0320-0120-000000000339' > 2018-05-10 03:24:21,081+02 INFO > [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] > (DefaultQuartzScheduler8) [7715ceda] FINISH, GlusterVolumesListVDSCommand, > return: {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine. > core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d, > e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine. > core.common.businessentities.g > luster.GlusterVolumeEntity@f88c521b}, log id: 118aa264 > > > > 2018-05-10 11:59:26,047+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) > [400fa486] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = > n1.itsmart.cloud, VdsIdVDSCommandParametersBase: > {hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' execution failed: null > 2018-05-10 11:59:26,047+02 INFO [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) > [400fa486] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: > 14a71ef0 > 2018-05-10 11:59:26,048+02 INFO [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) > [400fa486] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName = > n4.itsmart.cloud, VdsIdVDSCommandParametersBase: > {hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), log id: 28d9e255 > 2018-05-10 11:59:26,051+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) > [400fa486] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = > n4.itsmart.cloud, VdsIdVDSCommandParametersBase: > {hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' execution failed: null > 2018-05-10 11:59:26,051+02 INFO [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) > [400fa486] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: > 28d9e255 > 2018-05-10 11:59:26,052+02 INFO [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) > [400fa486] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName = > n2.itsmart.cloud, VdsIdVDSCommandParametersBase: > {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 4a7b280e > 2018-05-10 11:59:26,054+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) > [400fa486] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = > n2.itsmart.cloud, VdsIdVDSCommandParametersBase: > {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' execution failed: null > 2018-05-10 11:59:26,054+02 INFO [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) > [400fa486] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: > 4a7b280e > 2018-05-10 11:59:26,055+02 INFO [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) > [400fa486] START, GetGlusterLocalLogicalVolumeListVDSCommand(HostName = > n3.itsmart.cloud, VdsIdVDSCommandParametersBase: > {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), log id: 18adc534 > 2018-05-10 11:59:26,057+02 ERROR [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) > [400fa486] Command 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = > n3.itsmart.cloud, VdsIdVDSCommandParametersBase: > {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' execution failed: null > 2018-05-10 11:59:26,057+02 INFO [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterLocalLogicalVolumeListVDSCommand] (DefaultQuartzScheduler4) > [400fa486] FINISH, GetGlusterLocalLogicalVolumeListVDSCommand, log id: > 18adc534 > 2018-05-10 11:59:26,058+02 INFO [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler4) > [400fa486] START, GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = > n3.itsmart.cloud, GlusterVolumeAdvancedDetailsVD > SParameters:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec', > volumeName='volume1'}), log id: 3451084f > 2018-05-10 11:59:28,050+02 INFO > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock > 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', > sharedLocks=''}' > 2018-05-10 11:59:28,060+02 INFO > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock > 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', > sharedLocks=''}' > 2018-05-10 11:59:28,062+02 INFO > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock > 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', > sharedLocks=''}' > 2018-05-10 11:59:31,054+02 INFO > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock > 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', > sharedLocks=''}' > 2018-05-10 11:59:31,054+02 INFO > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock > 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', > sharedLocks=''}' > 2018-05-10 11:59:31,062+02 INFO > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock > 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', > sharedLocks=''}' > 2018-05-10 11:59:31,064+02 INFO > [org.ovirt.engine.core.bll.lock.InMemoryLockManager] > (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock > 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', > sharedLocks=''}' > 2018-05-10 11:59:31,465+02 INFO [org.ovirt.engine.core.vdsbroker.gluster. > GetGlusterVolumeAdvancedDetailsVDSCommand] (DefaultQuartzScheduler4) > [400fa486] FINISH, GetGlusterVolumeAdvancedDetailsVDSCommand, return: > org.ovirt.engine.core.common.businessentities.gluster. > GlusterVolumeAdvancedDetails@3f1b7f43, log id: 3451084f > 2018-05-10 11:59:31,466+02 ERROR > [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] > (DefaultQuartzScheduler4) [400fa486] Error while refreshing brick statuses > for volume 'volume1' of cluster 'C6220': null > > > R > Tibor > > ----- 2018. máj.. 10., 11:43, Sahina Bose <sab...@redhat.com> írta: > > This doesn't affect the monitoring of state. > Any errors in vdsm.log? > Or errors in engine.log of the form "Error while refreshing brick statuses > for volume" > > On Thu, May 10, 2018 at 2:33 PM, Demeter Tibor <tdeme...@itsmart.hu> > wrote: > >> Hi, >> >> Thank you for your fast reply :) >> >> >> 2018-05-10 11:01:51,574+02 INFO [org.ovirt.engine.core. >> vdsbroker.gluster.GlusterServersListVDSCommand] >> (DefaultQuartzScheduler6) [7f01fc2d] START, >> GlusterServersListVDSCommand(HostName >> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase: >> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 39adbbb8 >> 2018-05-10 11:01:51,768+02 INFO [org.ovirt.engine.core. >> vdsbroker.gluster.GlusterServersListVDSCommand] >> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, GlusterServersListVDSCommand, >> return: [10.101.0.2/24:CONNECTED, n1.cloudata.local:CONNECTED, >> 10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], log id: 39adbbb8 >> 2018-05-10 11:01:51,788+02 INFO [org.ovirt.engine.core. >> vdsbroker.gluster.GlusterVolumesListVDSCommand] >> (DefaultQuartzScheduler6) [7f01fc2d] START, >> GlusterVolumesListVDSCommand(HostName >> = n2.itsmart.cloud, GlusterVolumesListVDSParameter >> s:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 738a7261 >> 2018-05-10 11:01:51,892+02 WARN [org.ovirt.engine.core. >> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) >> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster/brick/brick1' >> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as >> no gluster network found in cluster '59c10db3-0324-0320-0120- >> 000000000339' >> 2018-05-10 11:01:51,898+02 WARN [org.ovirt.engine.core. >> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) >> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster/brick/brick2' >> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as >> no gluster network found in cluster '59c10db3-0324-0320-0120- >> 000000000339' >> 2018-05-10 11:01:51,905+02 WARN [org.ovirt.engine.core. >> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) >> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster/brick/brick3' >> of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as >> no gluster network found in cluster '59c10db3-0324-0320-0120- >> 000000000339' >> 2018-05-10 11:01:51,911+02 WARN [org.ovirt.engine.core. >> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) >> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster2/brick/brick1' >> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as >> no gluster network found in cluster '59c10db3-0324-0320-0120- >> 000000000339' >> 2018-05-10 11:01:51,917+02 WARN [org.ovirt.engine.core. >> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) >> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster2/brick/brick2' >> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as >> no gluster network found in cluster '59c10db3-0324-0320-0120- >> 000000000339' >> 2018-05-10 11:01:51,924+02 WARN [org.ovirt.engine.core. >> vdsbroker.gluster.GlusterVolumesListReturn] (DefaultQuartzScheduler6) >> [7f01fc2d] Could not associate brick '10.104.0.1:/gluster2/brick/brick3' >> of volume '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as >> no gluster network found in cluster '59c10db3-0324-0320-0120- >> 000000000339' >> 2018-05-10 11:01:51,925+02 INFO [org.ovirt.engine.core. >> vdsbroker.gluster.GlusterVolumesListVDSCommand] >> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, GlusterVolumesListVDSCommand, >> return: {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine. >> core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d, >> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine. >> core.common.businessentities.gluster.GlusterVolumeEntity@f88c521b}, log >> id: 738a7261 >> >> >> This happening continuously. >> >> Thanks! >> Tibor >> >> >> >> ----- 2018. máj.. 10., 10:56, Sahina Bose <sab...@redhat.com> írta: >> >> Could you check the engine.log if there are errors related to getting >> GlusterVolumeAdvancedDetails ? >> >> On Thu, May 10, 2018 at 2:02 PM, Demeter Tibor <tdeme...@itsmart.hu> >> wrote: >> >>> Dear Ovirt Users, >>> I've followed up the self-hosted-engine upgrade documentation, I >>> upgraded my 4.1 system to 4.2.3. >>> I upgaded the first node with yum upgrade, it seems working now fine. >>> But since upgrade, the gluster informations seems to displayed incorrect on >>> the admin panel. The volume yellow, and there are red bricks from that node. >>> I've checked in console, I think my gluster is not degraded: >>> >>> root@n1 ~]# gluster volume list >>> volume1 >>> volume2 >>> [root@n1 ~]# gluster volume info >>> >>> Volume Name: volume1 >>> Type: Distributed-Replicate >>> Volume ID: e0f568fa-987c-4f5c-b853-01bce718ee27 >>> Status: Started >>> Snapshot Count: 0 >>> Number of Bricks: 3 x 3 = 9 >>> Transport-type: tcp >>> Bricks: >>> Brick1: 10.104.0.1:/gluster/brick/brick1 >>> Brick2: 10.104.0.2:/gluster/brick/brick1 >>> Brick3: 10.104.0.3:/gluster/brick/brick1 >>> Brick4: 10.104.0.1:/gluster/brick/brick2 >>> Brick5: 10.104.0.2:/gluster/brick/brick2 >>> Brick6: 10.104.0.3:/gluster/brick/brick2 >>> Brick7: 10.104.0.1:/gluster/brick/brick3 >>> Brick8: 10.104.0.2:/gluster/brick/brick3 >>> Brick9: 10.104.0.3:/gluster/brick/brick3 >>> Options Reconfigured: >>> transport.address-family: inet >>> performance.readdir-ahead: on >>> nfs.disable: on >>> storage.owner-uid: 36 >>> storage.owner-gid: 36 >>> performance.quick-read: off >>> performance.read-ahead: off >>> performance.io-cache: off >>> performance.stat-prefetch: off >>> performance.low-prio-threads: 32 >>> network.remote-dio: enable >>> cluster.eager-lock: enable >>> cluster.quorum-type: auto >>> cluster.server-quorum-type: server >>> cluster.data-self-heal-algorithm: full >>> cluster.locking-scheme: granular >>> cluster.shd-max-threads: 8 >>> cluster.shd-wait-qlength: 10000 >>> features.shard: on >>> user.cifs: off >>> server.allow-insecure: on >>> >>> Volume Name: volume2 >>> Type: Distributed-Replicate >>> Volume ID: 68cfb061-1320-4042-abcd-9228da23c0c8 >>> Status: Started >>> Snapshot Count: 0 >>> Number of Bricks: 3 x 3 = 9 >>> Transport-type: tcp >>> Bricks: >>> Brick1: 10.104.0.1:/gluster2/brick/brick1 >>> Brick2: 10.104.0.2:/gluster2/brick/brick1 >>> Brick3: 10.104.0.3:/gluster2/brick/brick1 >>> Brick4: 10.104.0.1:/gluster2/brick/brick2 >>> Brick5: 10.104.0.2:/gluster2/brick/brick2 >>> Brick6: 10.104.0.3:/gluster2/brick/brick2 >>> Brick7: 10.104.0.1:/gluster2/brick/brick3 >>> Brick8: 10.104.0.2:/gluster2/brick/brick3 >>> Brick9: 10.104.0.3:/gluster2/brick/brick3 >>> Options Reconfigured: >>> nfs.disable: on >>> performance.readdir-ahead: on >>> transport.address-family: inet >>> cluster.quorum-type: auto >>> network.ping-timeout: 10 >>> auth.allow: * >>> performance.quick-read: off >>> performance.read-ahead: off >>> performance.io-cache: off >>> performance.stat-prefetch: off >>> performance.low-prio-threads: 32 >>> network.remote-dio: enable >>> cluster.eager-lock: enable >>> cluster.server-quorum-type: server >>> cluster.data-self-heal-algorithm: full >>> cluster.locking-scheme: granular >>> cluster.shd-max-threads: 8 >>> cluster.shd-wait-qlength: 10000 >>> features.shard: on >>> user.cifs: off >>> storage.owner-uid: 36 >>> storage.owner-gid: 36 >>> server.allow-insecure: on >>> [root@n1 ~]# gluster volume status >>> Status of volume: volume1 >>> Gluster process TCP Port RDMA Port Online >>> Pid >>> ------------------------------------------------------------ >>> ------------------ >>> Brick 10.104.0.1:/gluster/brick/brick1 49152 0 Y >>> 3464 >>> Brick 10.104.0.2:/gluster/brick/brick1 49152 0 Y >>> 68937 >>> Brick 10.104.0.3:/gluster/brick/brick1 49161 0 Y >>> 94506 >>> Brick 10.104.0.1:/gluster/brick/brick2 49153 0 Y >>> 3457 >>> Brick 10.104.0.2:/gluster/brick/brick2 49153 0 Y >>> 68943 >>> Brick 10.104.0.3:/gluster/brick/brick2 49162 0 Y >>> 94514 >>> Brick 10.104.0.1:/gluster/brick/brick3 49154 0 Y >>> 3465 >>> Brick 10.104.0.2:/gluster/brick/brick3 49154 0 Y >>> 68949 >>> Brick 10.104.0.3:/gluster/brick/brick3 49163 0 Y >>> 94520 >>> Self-heal Daemon on localhost N/A N/A Y >>> 54356 >>> Self-heal Daemon on 10.104.0.2 N/A N/A Y >>> 962 >>> Self-heal Daemon on 10.104.0.3 N/A N/A Y >>> 108977 >>> Self-heal Daemon on 10.104.0.4 N/A N/A Y >>> 61603 >>> >>> Task Status of Volume volume1 >>> ------------------------------------------------------------ >>> ------------------ >>> There are no active volume tasks >>> >>> Status of volume: volume2 >>> Gluster process TCP Port RDMA Port Online >>> Pid >>> ------------------------------------------------------------ >>> ------------------ >>> Brick 10.104.0.1:/gluster2/brick/brick1 49155 0 Y >>> 3852 >>> Brick 10.104.0.2:/gluster2/brick/brick1 49158 0 Y >>> 68955 >>> Brick 10.104.0.3:/gluster2/brick/brick1 49164 0 Y >>> 94527 >>> Brick 10.104.0.1:/gluster2/brick/brick2 49156 0 Y >>> 3851 >>> Brick 10.104.0.2:/gluster2/brick/brick2 49159 0 Y >>> 68961 >>> Brick 10.104.0.3:/gluster2/brick/brick2 49165 0 Y >>> 94533 >>> Brick 10.104.0.1:/gluster2/brick/brick3 49157 0 Y >>> 3883 >>> Brick 10.104.0.2:/gluster2/brick/brick3 49160 0 Y >>> 68968 >>> Brick 10.104.0.3:/gluster2/brick/brick3 49166 0 Y >>> 94541 >>> Self-heal Daemon on localhost N/A N/A Y >>> 54356 >>> Self-heal Daemon on 10.104.0.2 N/A N/A Y >>> 962 >>> Self-heal Daemon on 10.104.0.3 N/A N/A Y >>> 108977 >>> Self-heal Daemon on 10.104.0.4 N/A N/A Y >>> 61603 >>> >>> Task Status of Volume volume2 >>> ------------------------------------------------------------ >>> ------------------ >>> There are no active volume tasks >>> >>> I think ovirt can't read valid informations about gluster. >>> I can't contiune upgrade of other hosts until this problem exist. >>> >>> Please help me:) >>> >>> >>> Thanks >>> >>> Regards, >>> >>> Tibor >>> >>> >>> _______________________________________________ >>> Users mailing list -- users@ovirt.org >>> To unsubscribe send an email to users-le...@ovirt.org >>> >>> >> >
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org