Meanwhile i just changed my gluster network to 10.104.0.0/24 but does not happend anything.
Regards, Tibor ----- 2018. máj.. 14., 9:49, Demeter Tibor <[email protected]> írta: > Hi, > Yes, I have a gluster network, but it's "funny" because that is the > 10.105.0.x/24. :( Also, the n4.itsmart.cloud is mean 10.104.0.4. > The 10.104.0.x/24 is my ovirtmgmt network. > However, the 10.104.0.x is accessable from all hosts. > What should I do? > Thanks, > R > Tibor > ----- 2018. máj.. 12., 17:17, Doug Ingham <[email protected]> írta: >> The two key errors I'd investigate are these... >>> 2018-05-10 03:24:21,048+02 WARN >>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>> '10.104.0.1:/gluster/brick/brick1' of volume >>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster >>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>> 2018-05-10 03:24:20,749+02 ERROR >>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] (DefaultQuartzScheduler7) >>> [43f4eaec] Error while refreshing brick statuses for volume 'volume1' of >>> cluster 'C6220': null >>> 2018-05-10 11:59:26,051+02 ERROR >>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>> (DefaultQuartzScheduler4) [400fa486] Command >>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud, >>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' >>> execution failed: null >> I'd start with that first one. Is the network/interface group of your storage >> layer actually defined as a Gluster & Migration network within oVirt? >> On 12 May 2018 at 03:44, Demeter Tibor < [ mailto:[email protected] | >> [email protected] ] > wrote: >>> Hi, >>> Could someone help me please ? I can't finish my upgrade process. >>> Thanks >>> R >>> Tibor >>> ----- 2018. máj.. 10., 12:51, Demeter Tibor < [ mailto:[email protected] | >>> [email protected] ] > írta: >>>> Hi, >>>> I've attached the vdsm and supervdsm logs. But I don't have engine.log >>>> here, >>>> because that is on hosted engine vm. Should I send that ? >>>> Thank you >>>> Regards, >>>> Tibor >>>> ----- 2018. máj.. 10., 12:30, Sahina Bose < [ mailto:[email protected] | >>>> [email protected] ] > írta: >>>>> There's a bug here. Can you log one attaching this engine.log and also >>>>> vdsm.log >>>>> & supervdsm.log from n3.itsmart.cloud >>>>> On Thu, May 10, 2018 at 3:35 PM, Demeter Tibor < [ >>>>> mailto:[email protected] | >>>>> [email protected] ] > wrote: >>>>>> Hi, >>>>>> I found this: >>>>>> 2018-05-10 03:24:19,096+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return: >>>>>> org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@ca97448e, >>>>>> log id: 347435ae >>>>>> 2018-05-10 03:24:19,097+02 ERROR >>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] >>>>>> (DefaultQuartzScheduler7) >>>>>> [43f4eaec] Error while refreshing brick statuses for volume 'volume2' of >>>>>> cluster 'C6220': null >>>>>> 2018-05-10 03:24:19,097+02 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler8) >>>>>> [7715ceda] Failed to acquire lock and wait lock >>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>> sharedLocks=''}' >>>>>> 2018-05-10 03:24:19,104+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud, >>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), >>>>>> log id: 6908121d >>>>>> 2018-05-10 03:24:19,106+02 ERROR >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command >>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud, >>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' >>>>>> execution failed: null >>>>>> 2018-05-10 03:24:19,106+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6908121d >>>>>> 2018-05-10 03:24:19,107+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n1.itsmart.cloud, >>>>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'}), >>>>>> log id: 735c6a5f >>>>>> 2018-05-10 03:24:19,109+02 ERROR >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command >>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n1.itsmart.cloud, >>>>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' >>>>>> execution failed: null >>>>>> 2018-05-10 03:24:19,109+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 735c6a5f >>>>>> 2018-05-10 03:24:19,110+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud, >>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>> log id: 6f9e9f58 >>>>>> 2018-05-10 03:24:19,112+02 ERROR >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command >>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud, >>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' >>>>>> execution failed: null >>>>>> 2018-05-10 03:24:19,112+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6f9e9f58 >>>>>> 2018-05-10 03:24:19,113+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud, >>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), >>>>>> log id: 2ee46967 >>>>>> 2018-05-10 03:24:19,115+02 ERROR >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command >>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud, >>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' >>>>>> execution failed: null >>>>>> 2018-05-10 03:24:19,116+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 2ee46967 >>>>>> 2018-05-10 03:24:19,117+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = n1.itsmart.cloud, >>>>>> GlusterVolumeAdvancedDetailsVDSParameters:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57', >>>>>> volumeName='volume1'}), log id: 7550e5c >>>>>> 2018-05-10 03:24:20,748+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return: >>>>>> org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@4a46066f, >>>>>> log id: 7550e5c >>>>>> 2018-05-10 03:24:20,749+02 ERROR >>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] >>>>>> (DefaultQuartzScheduler7) >>>>>> [43f4eaec] Error while refreshing brick statuses for volume 'volume1' of >>>>>> cluster 'C6220': null >>>>>> 2018-05-10 03:24:20,750+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] >>>>>> (DefaultQuartzScheduler8) [7715ceda] START, >>>>>> GlusterServersListVDSCommand(HostName = n2.itsmart.cloud, >>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>> log id: 120cc68d >>>>>> 2018-05-10 03:24:20,930+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] >>>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, >>>>>> GlusterServersListVDSCommand, >>>>>> return: [ [ http://10.101.0.2/24:CONNECTED | 10.101.0.2/24:CONNECTED ] , >>>>>> n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED, >>>>>> 10.104.0.4:CONNECTED], log >>>>>> id: 120cc68d >>>>>> 2018-05-10 03:24:20,949+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>>> (DefaultQuartzScheduler8) [7715ceda] START, >>>>>> GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud, >>>>>> GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>> log id: 118aa264 >>>>>> 2018-05-10 03:24:21,048+02 WARN >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>> '10.104.0.1:/gluster/brick/brick1' of volume >>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster >>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>> 2018-05-10 03:24:21,055+02 WARN >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>> '10.104.0.1:/gluster/brick/brick2' of volume >>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster >>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>> 2018-05-10 03:24:21,061+02 WARN >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>> '10.104.0.1:/gluster/brick/brick3' of volume >>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster >>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>> 2018-05-10 03:24:21,067+02 WARN >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>> '10.104.0.1:/gluster2/brick/brick1' of volume >>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster >>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>> 2018-05-10 03:24:21,074+02 WARN >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>> '10.104.0.1:/gluster2/brick/brick2' of volume >>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster >>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>> 2018-05-10 03:24:21,080+02 WARN >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>> '10.104.0.1:/gluster2/brick/brick3' of volume >>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster >>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>> 2018-05-10 03:24:21,081+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, >>>>>> GlusterVolumesListVDSCommand, >>>>>> return: >>>>>> {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d, >>>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.g >>>>>> luster.GlusterVolumeEntity@f88c521b}, log id: 118aa264 >>>>>> 2018-05-10 11:59:26,047+02 ERROR >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] Command >>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n1.itsmart.cloud, >>>>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' >>>>>> execution failed: null >>>>>> 2018-05-10 11:59:26,047+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 14a71ef0 >>>>>> 2018-05-10 11:59:26,048+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] START, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud, >>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), >>>>>> log id: 28d9e255 >>>>>> 2018-05-10 11:59:26,051+02 ERROR >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] Command >>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n4.itsmart.cloud, >>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' >>>>>> execution failed: null >>>>>> 2018-05-10 11:59:26,051+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 28d9e255 >>>>>> 2018-05-10 11:59:26,052+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] START, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud, >>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>> log id: 4a7b280e >>>>>> 2018-05-10 11:59:26,054+02 ERROR >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] Command >>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n2.itsmart.cloud, >>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' >>>>>> execution failed: null >>>>>> 2018-05-10 11:59:26,054+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 4a7b280e >>>>>> 2018-05-10 11:59:26,055+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] START, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud, >>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), >>>>>> log id: 18adc534 >>>>>> 2018-05-10 11:59:26,057+02 ERROR >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] Command >>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = n3.itsmart.cloud, >>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' >>>>>> execution failed: null >>>>>> 2018-05-10 11:59:26,057+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 18adc534 >>>>>> 2018-05-10 11:59:26,058+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] START, >>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = n3.itsmart.cloud, >>>>>> GlusterVolumeAdvancedDetailsVDSParameters:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec', >>>>>> volumeName='volume1'}), log id: 3451084f >>>>>> 2018-05-10 11:59:28,050+02 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler3) >>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>> sharedLocks=''}' >>>>>> 2018-05-10 11:59:28,060+02 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler3) >>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>> sharedLocks=''}' >>>>>> 2018-05-10 11:59:28,062+02 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler3) >>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>> sharedLocks=''}' >>>>>> 2018-05-10 11:59:31,054+02 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler3) >>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>> sharedLocks=''}' >>>>>> 2018-05-10 11:59:31,054+02 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler3) >>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>> sharedLocks=''}' >>>>>> 2018-05-10 11:59:31,062+02 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler3) >>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>> sharedLocks=''}' >>>>>> 2018-05-10 11:59:31,064+02 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler3) >>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>> sharedLocks=''}' >>>>>> 2018-05-10 11:59:31,465+02 INFO >>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return: >>>>>> org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@3f1b7f43, >>>>>> log id: 3451084f >>>>>> 2018-05-10 11:59:31,466+02 ERROR >>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] >>>>>> (DefaultQuartzScheduler4) >>>>>> [400fa486] Error while refreshing brick statuses for volume 'volume1' of >>>>>> cluster 'C6220': null >>>>>> R >>>>>> Tibor >>>>>> ----- 2018. máj.. 10., 11:43, Sahina Bose < [ mailto:[email protected] | >>>>>> [email protected] ] > írta: >>>>>>> This doesn't affect the monitoring of state. >>>>>>> Any errors in vdsm.log? >>>>>>> Or errors in engine.log of the form "Error while refreshing brick >>>>>>> statuses for >>>>>>> volume" >>>>>>> On Thu, May 10, 2018 at 2:33 PM, Demeter Tibor < [ >>>>>>> mailto:[email protected] | >>>>>>> [email protected] ] > wrote: >>>>>>>> Hi, >>>>>>>> Thank you for your fast reply :) >>>>>>>> 2018-05-10 11:01:51,574+02 INFO >>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] >>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START, >>>>>>>> GlusterServersListVDSCommand(HostName = n2.itsmart.cloud, >>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>>>> log id: 39adbbb8 >>>>>>>> 2018-05-10 11:01:51,768+02 INFO >>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] >>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, >>>>>>>> GlusterServersListVDSCommand, >>>>>>>> return: [ [ http://10.101.0.2/24:CONNECTED | 10.101.0.2/24:CONNECTED ] >>>>>>>> , >>>>>>>> n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED, >>>>>>>> 10.104.0.4:CONNECTED], log >>>>>>>> id: 39adbbb8 >>>>>>>> 2018-05-10 11:01:51,788+02 INFO >>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START, >>>>>>>> GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud, >>>>>>>> GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>>>> log id: 738a7261 >>>>>>>> 2018-05-10 11:01:51,892+02 WARN >>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick >>>>>>>> '10.104.0.1:/gluster/brick/brick1' of volume >>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no >>>>>>>> gluster >>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>> 2018-05-10 11:01:51,898+02 WARN >>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick >>>>>>>> '10.104.0.1:/gluster/brick/brick2' of volume >>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no >>>>>>>> gluster >>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>> 2018-05-10 11:01:51,905+02 WARN >>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick >>>>>>>> '10.104.0.1:/gluster/brick/brick3' of volume >>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no >>>>>>>> gluster >>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>> 2018-05-10 11:01:51,911+02 WARN >>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick >>>>>>>> '10.104.0.1:/gluster2/brick/brick1' of volume >>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no >>>>>>>> gluster >>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>> 2018-05-10 11:01:51,917+02 WARN >>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick >>>>>>>> '10.104.0.1:/gluster2/brick/brick2' of volume >>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no >>>>>>>> gluster >>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>> 2018-05-10 11:01:51,924+02 WARN >>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick >>>>>>>> '10.104.0.1:/gluster2/brick/brick3' of volume >>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no >>>>>>>> gluster >>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>> 2018-05-10 11:01:51,925+02 INFO >>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, >>>>>>>> GlusterVolumesListVDSCommand, >>>>>>>> return: >>>>>>>> {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d, >>>>>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@f88c521b}, >>>>>>>> log id: 738a7261 >>>>>>>> This happening continuously. >>>>>>>> Thanks! >>>>>>>> Tibor >>>>>>>> ----- 2018. máj.. 10., 10:56, Sahina Bose < [ mailto:[email protected] >>>>>>>> | >>>>>>>> [email protected] ] > írta: >>>>>>>>> Could you check the engine.log if there are errors related to getting >>>>>>>>> GlusterVolumeAdvancedDetails ? >>>>>>>>> On Thu, May 10, 2018 at 2:02 PM, Demeter Tibor < [ >>>>>>>>> mailto:[email protected] | >>>>>>>>> [email protected] ] > wrote: >>>>>>>>>> Dear Ovirt Users, >>>>>>>>>> I've followed up the self-hosted-engine upgrade documentation, I >>>>>>>>>> upgraded my 4.1 >>>>>>>>>> system to 4.2.3. >>>>>>>>>> I upgaded the first node with yum upgrade, it seems working now >>>>>>>>>> fine. But since >>>>>>>>>> upgrade, the gluster informations seems to displayed incorrect on >>>>>>>>>> the admin >>>>>>>>>> panel. The volume yellow, and there are red bricks from that node. >>>>>>>>>> I've checked in console, I think my gluster is not degraded: >>>>>>>>>> root@n1 ~]# gluster volume list >>>>>>>>>> volume1 >>>>>>>>>> volume2 >>>>>>>>>> [root@n1 ~]# gluster volume info >>>>>>>>>> Volume Name: volume1 >>>>>>>>>> Type: Distributed-Replicate >>>>>>>>>> Volume ID: e0f568fa-987c-4f5c-b853-01bce718ee27 >>>>>>>>>> Status: Started >>>>>>>>>> Snapshot Count: 0 >>>>>>>>>> Number of Bricks: 3 x 3 = 9 >>>>>>>>>> Transport-type: tcp >>>>>>>>>> Bricks: >>>>>>>>>> Brick1: 10.104.0.1:/gluster/brick/brick1 >>>>>>>>>> Brick2: 10.104.0.2:/gluster/brick/brick1 >>>>>>>>>> Brick3: 10.104.0.3:/gluster/brick/brick1 >>>>>>>>>> Brick4: 10.104.0.1:/gluster/brick/brick2 >>>>>>>>>> Brick5: 10.104.0.2:/gluster/brick/brick2 >>>>>>>>>> Brick6: 10.104.0.3:/gluster/brick/brick2 >>>>>>>>>> Brick7: 10.104.0.1:/gluster/brick/brick3 >>>>>>>>>> Brick8: 10.104.0.2:/gluster/brick/brick3 >>>>>>>>>> Brick9: 10.104.0.3:/gluster/brick/brick3 >>>>>>>>>> Options Reconfigured: >>>>>>>>>> transport.address-family: inet >>>>>>>>>> performance.readdir-ahead: on >>>>>>>>>> nfs.disable: on >>>>>>>>>> storage.owner-uid: 36 >>>>>>>>>> storage.owner-gid: 36 >>>>>>>>>> performance.quick-read: off >>>>>>>>>> performance.read-ahead: off >>>>>>>>>> performance.io-cache: off >>>>>>>>>> performance.stat-prefetch: off >>>>>>>>>> performance.low-prio-threads: 32 >>>>>>>>>> network.remote-dio: enable >>>>>>>>>> cluster.eager-lock: enable >>>>>>>>>> cluster.quorum-type: auto >>>>>>>>>> cluster.server-quorum-type: server >>>>>>>>>> cluster.data-self-heal-algorithm: full >>>>>>>>>> cluster.locking-scheme: granular >>>>>>>>>> cluster.shd-max-threads: 8 >>>>>>>>>> cluster.shd-wait-qlength: 10000 >>>>>>>>>> features.shard: on >>>>>>>>>> user.cifs: off >>>>>>>>>> server.allow-insecure: on >>>>>>>>>> Volume Name: volume2 >>>>>>>>>> Type: Distributed-Replicate >>>>>>>>>> Volume ID: 68cfb061-1320-4042-abcd-9228da23c0c8 >>>>>>>>>> Status: Started >>>>>>>>>> Snapshot Count: 0 >>>>>>>>>> Number of Bricks: 3 x 3 = 9 >>>>>>>>>> Transport-type: tcp >>>>>>>>>> Bricks: >>>>>>>>>> Brick1: 10.104.0.1:/gluster2/brick/brick1 >>>>>>>>>> Brick2: 10.104.0.2:/gluster2/brick/brick1 >>>>>>>>>> Brick3: 10.104.0.3:/gluster2/brick/brick1 >>>>>>>>>> Brick4: 10.104.0.1:/gluster2/brick/brick2 >>>>>>>>>> Brick5: 10.104.0.2:/gluster2/brick/brick2 >>>>>>>>>> Brick6: 10.104.0.3:/gluster2/brick/brick2 >>>>>>>>>> Brick7: 10.104.0.1:/gluster2/brick/brick3 >>>>>>>>>> Brick8: 10.104.0.2:/gluster2/brick/brick3 >>>>>>>>>> Brick9: 10.104.0.3:/gluster2/brick/brick3 >>>>>>>>>> Options Reconfigured: >>>>>>>>>> nfs.disable: on >>>>>>>>>> performance.readdir-ahead: on >>>>>>>>>> transport.address-family: inet >>>>>>>>>> cluster.quorum-type: auto >>>>>>>>>> network.ping-timeout: 10 >>>>>>>>>> auth.allow: * >>>>>>>>>> performance.quick-read: off >>>>>>>>>> performance.read-ahead: off >>>>>>>>>> performance.io-cache: off >>>>>>>>>> performance.stat-prefetch: off >>>>>>>>>> performance.low-prio-threads: 32 >>>>>>>>>> network.remote-dio: enable >>>>>>>>>> cluster.eager-lock: enable >>>>>>>>>> cluster.server-quorum-type: server >>>>>>>>>> cluster.data-self-heal-algorithm: full >>>>>>>>>> cluster.locking-scheme: granular >>>>>>>>>> cluster.shd-max-threads: 8 >>>>>>>>>> cluster.shd-wait-qlength: 10000 >>>>>>>>>> features.shard: on >>>>>>>>>> user.cifs: off >>>>>>>>>> storage.owner-uid: 36 >>>>>>>>>> storage.owner-gid: 36 >>>>>>>>>> server.allow-insecure: on >>>>>>>>>> [root@n1 ~]# gluster volume status >>>>>>>>>> Status of volume: volume1 >>>>>>>>>> Gluster process TCP Port RDMA Port Online Pid >>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick1 49152 0 Y 3464 >>>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick1 49152 0 Y 68937 >>>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick1 49161 0 Y 94506 >>>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick2 49153 0 Y 3457 >>>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick2 49153 0 Y 68943 >>>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick2 49162 0 Y 94514 >>>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick3 49154 0 Y 3465 >>>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick3 49154 0 Y 68949 >>>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick3 49163 0 Y 94520 >>>>>>>>>> Self-heal Daemon on localhost N/A N/A Y 54356 >>>>>>>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y 962 >>>>>>>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977 >>>>>>>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603 >>>>>>>>>> Task Status of Volume volume1 >>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>> There are no active volume tasks >>>>>>>>>> Status of volume: volume2 >>>>>>>>>> Gluster process TCP Port RDMA Port Online Pid >>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick1 49155 0 Y 3852 >>>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick1 49158 0 Y 68955 >>>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick1 49164 0 Y 94527 >>>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick2 49156 0 Y 3851 >>>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick2 49159 0 Y 68961 >>>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick2 49165 0 Y 94533 >>>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick3 49157 0 Y 3883 >>>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick3 49160 0 Y 68968 >>>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick3 49166 0 Y 94541 >>>>>>>>>> Self-heal Daemon on localhost N/A N/A Y 54356 >>>>>>>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y 962 >>>>>>>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977 >>>>>>>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603 >>>>>>>>>> Task Status of Volume volume2 >>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>> There are no active volume tasks >>>>>>>>>> I think ovirt can't read valid informations about gluster. >>>>>>>>>> I can't contiune upgrade of other hosts until this problem exist. >>>>>>>>>> Please help me:) >>>>>>>>>> Thanks >>>>>>>>>> Regards, >>>>>>>>>> Tibor >>>>>>>>>> _______________________________________________ >>>>>>>>>> Users mailing list -- [ mailto:[email protected] | [email protected] ] >>>>>>>>>> To unsubscribe send an email to [ mailto:[email protected] | >>>>>>>>>> [email protected] ] >>>> _______________________________________________ >>>> Users mailing list -- [ mailto:[email protected] | [email protected] ] >>>> To unsubscribe send an email to [ mailto:[email protected] | >>>> [email protected] ] >>> _______________________________________________ >>> Users mailing list -- [ mailto:[email protected] | [email protected] ] >>> To unsubscribe send an email to [ mailto:[email protected] | >>> [email protected] ] >> -- >> Doug >> _______________________________________________ >> Users mailing list -- [email protected] >> To unsubscribe send an email to [email protected] > _______________________________________________ > Users mailing list -- [email protected] > To unsubscribe send an email to [email protected]
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected]

