Thanks for reporting this. https://gerrit.ovirt.org/91375 fixes this. I've re-opened bug https://bugzilla.redhat.com/show_bug.cgi?id=1574508
On Thu, May 17, 2018 at 10:12 PM, Demeter Tibor <tdeme...@itsmart.hu> wrote: > Hi, > > 4.2.4-0.0.master.20180515183442.git00e1340.el7.centos > > Firstly, I did a yum update "ovirt-*-setup*" > second, I have ran engine-setup to upgrade. > > I didn't remove the old repos, just installed the nightly repo. > > Thank you again, > > Regards, > > Tibor > > ----- 2018. máj.. 17., 15:02, Sahina Bose <sab...@redhat.com> írta: > > It doesn't look like the patch was applied. Still see the same error in > engine.log > "Error while refreshing brick statuses for volume 'volume1' of cluster > 'C6220': null"\ > > Did you use engine-setup to upgrade? What's the version of ovirt-engine > currently installed? > > On Thu, May 17, 2018 at 5:10 PM, Demeter Tibor <tdeme...@itsmart.hu> > wrote: > >> Hi, >> >> sure, >> >> Thank you for your time! >> >> R >> Tibor >> >> ----- 2018. máj.. 17., 12:19, Sahina Bose <sab...@redhat.com> írta: >> >> [+users] >> >> Can you provide the engine.log to see why the monitoring is not working >> here. thanks! >> >> On Wed, May 16, 2018 at 2:08 PM, Demeter Tibor <tdeme...@itsmart.hu> >> wrote: >> >>> Hi, >>> >>> Meanwhile, I did the upgrade engine, but the gluster state is same on my >>> first node. >>> I've attached some screenshot of my problem. >>> >>> Thanks >>> >>> Tibor >>> >>> >>> >>> ----- 2018. máj.. 16., 10:16, Demeter Tibor <tdeme...@itsmart.hu> írta >>> Hi, >>> >>> >>> If 4.3.4 will release, i just have to remove the nightly repo and update >>> to stable? >>> >>> I'm sorry for my terrible English, I try to explain what was my problem >>> with update. >>> I'm upgraded from 4.1.8. >>> >>> I followed up the official hosted-engine update documentation, that was >>> not clear me, because it has referenced to a lot of old thing (i think). >>> https://www.ovirt.org/documentation/upgrade-guide/upgrade-guide/ >>> https://www.ovirt.org/documentation/how-to/hosted- >>> engine/#upgrade-hosted-engine >>> >>> Maybe it need to update, because I had a lot of question under upgrade >>> and I was not sure in all of necessary steps. For example, If I need to >>> installing the new, 4.2 repo on the hosts, then need to remove the old repo >>> from that? >>> Why I need to do a" yum update -y" on hosts, meanwhile there is an >>> "Updatehost" menu in the GUI? So, maybe it outdated. >>> Since upgrade hosted engine, and the first node, I have problems with >>> gluster. It seems to working fine if you check it from console "gluster >>> volume status, etc" but not on the Gui, because now it yellow, and the >>> brick reds in the first node. >>> >>> Previously I did a mistake with glusterfs, my gluster config was wrong. >>> I have corrected them, but it did not helped to me,gluster bricks are reds >>> on my first node yet.... >>> >>> >>> Now I try to upgrade to nightly, but I'm affraid, because it a living, >>> productive system, and I don't have downtime. I hope it will help me. >>> >>> Thanks for all, >>> >>> Regards, >>> Tibor Demeter >>> >>> >>> >>> ----- 2018. máj.. 16., 9:58, Sahina Bose <sab...@redhat.com> írta: >>> >>> >>> >>> On Wed, May 16, 2018 at 1:19 PM, Demeter Tibor <tdeme...@itsmart.hu> >>> wrote: >>> >>>> Hi, >>>> >>>> is it a different, unstable repo? I have a productive cluster, how is >>>> safe that? >>>> I don't have any experience with nightly build. How can I use this? It >>>> have to install to the engine VM or all of my hosts? >>>> Thanks in advance for help me.. >>>> >>> >>> Only on the engine VM. >>> >>> Regarding stability - it passes CI so relatively stable, beyond that >>> there are no guarantees. >>> >>> What's the specific problem you're facing with update? Can you elaborate? >>> >>> >>>> Regards, >>>> >>>> Tibor >>>> >>>> ----- 2018. máj.. 15., 9:58, Demeter Tibor <tdeme...@itsmart.hu> írta: >>>> >>>> Hi, >>>> >>>> Could you explain how can I use this patch? >>>> >>>> R, >>>> Tibor >>>> >>>> >>>> ----- 2018. máj.. 14., 11:18, Demeter Tibor <tdeme...@itsmart.hu> írta: >>>> >>>> Hi, >>>> >>>> Sorry for my question, but can you tell me please how can I use this >>>> patch? >>>> >>>> Thanks, >>>> Regards, >>>> Tibor >>>> ----- 2018. máj.. 14., 10:47, Sahina Bose <sab...@redhat.com> írta: >>>> >>>> >>>> >>>> On Sat, May 12, 2018 at 1:14 PM, Demeter Tibor <tdeme...@itsmart.hu> >>>> wrote: >>>> >>>>> Hi, >>>>> >>>>> Could someone help me please ? I can't finish my upgrade process. >>>>> >>>> >>>> https://gerrit.ovirt.org/91164 should fix the error you're facing. >>>> >>>> Can you elaborate why this is affecting the upgrade process? >>>> >>>> >>>>> Thanks >>>>> R >>>>> Tibor >>>>> >>>>> >>>>> >>>>> ----- 2018. máj.. 10., 12:51, Demeter Tibor <tdeme...@itsmart.hu> >>>>> írta: >>>>> >>>>> Hi, >>>>> >>>>> I've attached the vdsm and supervdsm logs. But I don't have engine.log >>>>> here, because that is on hosted engine vm. Should I send that ? >>>>> >>>>> Thank you >>>>> >>>>> Regards, >>>>> >>>>> Tibor >>>>> ----- 2018. máj.. 10., 12:30, Sahina Bose <sab...@redhat.com> írta: >>>>> >>>>> There's a bug here. Can you log one attaching this engine.log and also >>>>> vdsm.log & supervdsm.log from n3.itsmart.cloud >>>>> >>>>> On Thu, May 10, 2018 at 3:35 PM, Demeter Tibor <tdeme...@itsmart.hu> >>>>> wrote: >>>>> >>>>>> Hi, >>>>>> >>>>>> I found this: >>>>>> >>>>>> >>>>>> 2018-05-10 03:24:19,096+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, >>>>>> return: org.ovirt.engine.core.common.businessentities.gluster. >>>>>> GlusterVolumeAdvancedDetails@ca97448e, log id: 347435ae >>>>>> 2018-05-10 03:24:19,097+02 ERROR >>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] Error while refreshing brick >>>>>> statuses >>>>>> for volume 'volume2' of cluster 'C6220': null >>>>>> 2018-05-10 03:24:19,097+02 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler8) [7715ceda] Failed to acquire lock and wait lock >>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>> sharedLocks=''}' >>>>>> 2018-05-10 03:24:19,104+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName >>>>>> = n4.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>>>> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), log id: 6908121d >>>>>> 2018-05-10 03:24:19,106+02 ERROR [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command ' >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>> n4.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>>>> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' execution failed: >>>>>> null >>>>>> 2018-05-10 03:24:19,106+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, >>>>>> log id: 6908121d >>>>>> 2018-05-10 03:24:19,107+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName >>>>>> = n1.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>>>> {hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'}), log id: 735c6a5f >>>>>> 2018-05-10 03:24:19,109+02 ERROR [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command ' >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>> n1.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>>>> {hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' execution failed: >>>>>> null >>>>>> 2018-05-10 03:24:19,109+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, >>>>>> log id: 735c6a5f >>>>>> 2018-05-10 03:24:19,110+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName >>>>>> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 6f9e9f58 >>>>>> 2018-05-10 03:24:19,112+02 ERROR [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command ' >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>> n2.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' execution failed: >>>>>> null >>>>>> 2018-05-10 03:24:19,112+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, >>>>>> log id: 6f9e9f58 >>>>>> 2018-05-10 03:24:19,113+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName >>>>>> = n3.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>>>> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), log id: 2ee46967 >>>>>> 2018-05-10 03:24:19,115+02 ERROR [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command ' >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>> n3.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>>>> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' execution failed: >>>>>> null >>>>>> 2018-05-10 03:24:19,116+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, >>>>>> log id: 2ee46967 >>>>>> 2018-05-10 03:24:19,117+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName >>>>>> = n1.itsmart.cloud, GlusterVolumeAdvancedDetailsVD >>>>>> SParameters:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57', >>>>>> volumeName='volume1'}), log id: 7550e5c >>>>>> 2018-05-10 03:24:20,748+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, >>>>>> return: org.ovirt.engine.core.common.businessentities.gluster. >>>>>> GlusterVolumeAdvancedDetails@4a46066f, log id: 7550e5c >>>>>> 2018-05-10 03:24:20,749+02 ERROR >>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] >>>>>> (DefaultQuartzScheduler7) [43f4eaec] Error while refreshing brick >>>>>> statuses >>>>>> for volume 'volume1' of cluster 'C6220': null >>>>>> 2018-05-10 03:24:20,750+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GlusterServersListVDSCommand] >>>>>> (DefaultQuartzScheduler8) [7715ceda] START, >>>>>> GlusterServersListVDSCommand(HostName >>>>>> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 120cc68d >>>>>> 2018-05-10 03:24:20,930+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GlusterServersListVDSCommand] >>>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, >>>>>> GlusterServersListVDSCommand, >>>>>> return: [10.101.0.2/24:CONNECTED, n1.cloudata.local:CONNECTED, >>>>>> 10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], log id: 120cc68d >>>>>> 2018-05-10 03:24:20,949+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>>> (DefaultQuartzScheduler8) [7715ceda] START, >>>>>> GlusterVolumesListVDSCommand(HostName >>>>>> = n2.itsmart.cloud, GlusterVolumesListVDSParameter >>>>>> s:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 118aa264 >>>>>> 2018-05-10 03:24:21,048+02 WARN [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GlusterVolumesListReturn] >>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>> '10.104.0.1: >>>>>> /gluster/brick/brick1' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' >>>>>> with correct network as no gluster network found in cluster >>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>> 2018-05-10 03:24:21,055+02 WARN [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GlusterVolumesListReturn] >>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>> '10.104.0.1: >>>>>> /gluster/brick/brick2' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' >>>>>> with correct network as no gluster network found in cluster >>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>> 2018-05-10 03:24:21,061+02 WARN [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GlusterVolumesListReturn] >>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>> '10.104.0.1: >>>>>> /gluster/brick/brick3' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' >>>>>> with correct network as no gluster network found in cluster >>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>> 2018-05-10 03:24:21,067+02 WARN [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GlusterVolumesListReturn] >>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>> '10.104.0.1: >>>>>> /gluster2/brick/brick1' of volume '68cfb061-1320-4042-abcd-9228da23c0c8' >>>>>> with correct network as no gluster network found in cluster >>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>> 2018-05-10 03:24:21,074+02 WARN [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GlusterVolumesListReturn] >>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>> '10.104.0.1: >>>>>> /gluster2/brick/brick2' of volume '68cfb061-1320-4042-abcd-9228da23c0c8' >>>>>> with correct network as no gluster network found in cluster >>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>> 2018-05-10 03:24:21,080+02 WARN [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GlusterVolumesListReturn] >>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>> '10.104.0.1: >>>>>> /gluster2/brick/brick3' of volume '68cfb061-1320-4042-abcd-9228da23c0c8' >>>>>> with correct network as no gluster network found in cluster >>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>> 2018-05-10 03:24:21,081+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, >>>>>> GlusterVolumesListVDSCommand, >>>>>> return: {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine. >>>>>> core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d, >>>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine. >>>>>> core.common.businessentities.g >>>>>> luster.GlusterVolumeEntity@f88c521b}, log id: 118aa264 >>>>>> >>>>>> >>>>>> >>>>>> 2018-05-10 11:59:26,047+02 ERROR [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] Command ' >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>> n1.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>>>> {hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' execution failed: >>>>>> null >>>>>> 2018-05-10 11:59:26,047+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, >>>>>> log id: 14a71ef0 >>>>>> 2018-05-10 11:59:26,048+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] START, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName >>>>>> = n4.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>>>> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), log id: 28d9e255 >>>>>> 2018-05-10 11:59:26,051+02 ERROR [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] Command ' >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>> n4.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>>>> {hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' execution failed: >>>>>> null >>>>>> 2018-05-10 11:59:26,051+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, >>>>>> log id: 28d9e255 >>>>>> 2018-05-10 11:59:26,052+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] START, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName >>>>>> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 4a7b280e >>>>>> 2018-05-10 11:59:26,054+02 ERROR [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] Command ' >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>> n2.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' execution failed: >>>>>> null >>>>>> 2018-05-10 11:59:26,054+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, >>>>>> log id: 4a7b280e >>>>>> 2018-05-10 11:59:26,055+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] START, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName >>>>>> = n3.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>>>> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), log id: 18adc534 >>>>>> 2018-05-10 11:59:26,057+02 ERROR [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] Command ' >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>> n3.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>>>> {hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' execution failed: >>>>>> null >>>>>> 2018-05-10 11:59:26,057+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, >>>>>> log id: 18adc534 >>>>>> 2018-05-10 11:59:26,058+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] START, >>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName >>>>>> = n3.itsmart.cloud, GlusterVolumeAdvancedDetailsVD >>>>>> SParameters:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec', >>>>>> volumeName='volume1'}), log id: 3451084f >>>>>> 2018-05-10 11:59:28,050+02 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock >>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>> sharedLocks=''}' >>>>>> 2018-05-10 11:59:28,060+02 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock >>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>> sharedLocks=''}' >>>>>> 2018-05-10 11:59:28,062+02 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock >>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>> sharedLocks=''}' >>>>>> 2018-05-10 11:59:31,054+02 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock >>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>> sharedLocks=''}' >>>>>> 2018-05-10 11:59:31,054+02 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock >>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>> sharedLocks=''}' >>>>>> 2018-05-10 11:59:31,062+02 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock >>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>> sharedLocks=''}' >>>>>> 2018-05-10 11:59:31,064+02 INFO >>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>> (DefaultQuartzScheduler3) [2eb1c389] Failed to acquire lock and wait lock >>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>> sharedLocks=''}' >>>>>> 2018-05-10 11:59:31,465+02 INFO [org.ovirt.engine.core. >>>>>> vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, >>>>>> return: org.ovirt.engine.core.common.businessentities.gluster. >>>>>> GlusterVolumeAdvancedDetails@3f1b7f43, log id: 3451084f >>>>>> 2018-05-10 11:59:31,466+02 ERROR >>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] >>>>>> (DefaultQuartzScheduler4) [400fa486] Error while refreshing brick >>>>>> statuses >>>>>> for volume 'volume1' of cluster 'C6220': null >>>>>> >>>>>> >>>>>> R >>>>>> Tibor >>>>>> >>>>>> ----- 2018. máj.. 10., 11:43, Sahina Bose <sab...@redhat.com> írta: >>>>>> >>>>>> This doesn't affect the monitoring of state. >>>>>> Any errors in vdsm.log? >>>>>> Or errors in engine.log of the form "Error while refreshing brick >>>>>> statuses for volume" >>>>>> >>>>>> On Thu, May 10, 2018 at 2:33 PM, Demeter Tibor <tdeme...@itsmart.hu> >>>>>> wrote: >>>>>> >>>>>>> Hi, >>>>>>> >>>>>>> Thank you for your fast reply :) >>>>>>> >>>>>>> >>>>>>> 2018-05-10 11:01:51,574+02 INFO [org.ovirt.engine.core. >>>>>>> vdsbroker.gluster.GlusterServersListVDSCommand] >>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START, >>>>>>> GlusterServersListVDSCommand(HostName >>>>>>> = n2.itsmart.cloud, VdsIdVDSCommandParametersBase: >>>>>>> {hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 39adbbb8 >>>>>>> 2018-05-10 11:01:51,768+02 INFO [org.ovirt.engine.core. >>>>>>> vdsbroker.gluster.GlusterServersListVDSCommand] >>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, >>>>>>> GlusterServersListVDSCommand, >>>>>>> return: [10.101.0.2/24:CONNECTED, n1.cloudata.local:CONNECTED, >>>>>>> 10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], log id: 39adbbb8 >>>>>>> 2018-05-10 11:01:51,788+02 INFO [org.ovirt.engine.core. >>>>>>> vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START, >>>>>>> GlusterVolumesListVDSCommand(HostName >>>>>>> = n2.itsmart.cloud, GlusterVolumesListVDSParameter >>>>>>> s:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), log id: 738a7261 >>>>>>> 2018-05-10 11:01:51,892+02 WARN [org.ovirt.engine.core. >>>>>>> vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick >>>>>>> '10.104.0.1: >>>>>>> /gluster/brick/brick1' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' >>>>>>> with correct network as no gluster network found in cluster >>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>> 2018-05-10 11:01:51,898+02 WARN [org.ovirt.engine.core. >>>>>>> vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick >>>>>>> '10.104.0.1: >>>>>>> /gluster/brick/brick2' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' >>>>>>> with correct network as no gluster network found in cluster >>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>> 2018-05-10 11:01:51,905+02 WARN [org.ovirt.engine.core. >>>>>>> vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick >>>>>>> '10.104.0.1: >>>>>>> /gluster/brick/brick3' of volume 'e0f568fa-987c-4f5c-b853-01bce718ee27' >>>>>>> with correct network as no gluster network found in cluster >>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>> 2018-05-10 11:01:51,911+02 WARN [org.ovirt.engine.core. >>>>>>> vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick >>>>>>> '10.104.0.1: >>>>>>> /gluster2/brick/brick1' of volume '68cfb061-1320-4042-abcd-9228da23c0c8' >>>>>>> with correct network as no gluster network found in cluster >>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>> 2018-05-10 11:01:51,917+02 WARN [org.ovirt.engine.core. >>>>>>> vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick >>>>>>> '10.104.0.1: >>>>>>> /gluster2/brick/brick2' of volume '68cfb061-1320-4042-abcd-9228da23c0c8' >>>>>>> with correct network as no gluster network found in cluster >>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>> 2018-05-10 11:01:51,924+02 WARN [org.ovirt.engine.core. >>>>>>> vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick >>>>>>> '10.104.0.1: >>>>>>> /gluster2/brick/brick3' of volume '68cfb061-1320-4042-abcd-9228da23c0c8' >>>>>>> with correct network as no gluster network found in cluster >>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>> 2018-05-10 11:01:51,925+02 INFO [org.ovirt.engine.core. >>>>>>> vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, >>>>>>> GlusterVolumesListVDSCommand, >>>>>>> return: {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine. >>>>>>> core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d, >>>>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine. >>>>>>> core.common.businessentities.gluster.GlusterVolumeEntity@f88c521b}, >>>>>>> log id: 738a7261 >>>>>>> >>>>>>> >>>>>>> This happening continuously. >>>>>>> >>>>>>> Thanks! >>>>>>> Tibor >>>>>>> >>>>>>> >>>>>>> >>>>>>> ----- 2018. máj.. 10., 10:56, Sahina Bose <sab...@redhat.com> írta: >>>>>>> >>>>>>> Could you check the engine.log if there are errors related to >>>>>>> getting GlusterVolumeAdvancedDetails ? >>>>>>> >>>>>>> On Thu, May 10, 2018 at 2:02 PM, Demeter Tibor <tdeme...@itsmart.hu> >>>>>>> wrote: >>>>>>> >>>>>>>> Dear Ovirt Users, >>>>>>>> I've followed up the self-hosted-engine upgrade documentation, I >>>>>>>> upgraded my 4.1 system to 4.2.3. >>>>>>>> I upgaded the first node with yum upgrade, it seems working now >>>>>>>> fine. But since upgrade, the gluster informations seems to displayed >>>>>>>> incorrect on the admin panel. The volume yellow, and there are red >>>>>>>> bricks >>>>>>>> from that node. >>>>>>>> I've checked in console, I think my gluster is not degraded: >>>>>>>> >>>>>>>> root@n1 ~]# gluster volume list >>>>>>>> volume1 >>>>>>>> volume2 >>>>>>>> [root@n1 ~]# gluster volume info >>>>>>>> >>>>>>>> Volume Name: volume1 >>>>>>>> Type: Distributed-Replicate >>>>>>>> Volume ID: e0f568fa-987c-4f5c-b853-01bce718ee27 >>>>>>>> Status: Started >>>>>>>> Snapshot Count: 0 >>>>>>>> Number of Bricks: 3 x 3 = 9 >>>>>>>> Transport-type: tcp >>>>>>>> Bricks: >>>>>>>> Brick1: 10.104.0.1:/gluster/brick/brick1 >>>>>>>> Brick2: 10.104.0.2:/gluster/brick/brick1 >>>>>>>> Brick3: 10.104.0.3:/gluster/brick/brick1 >>>>>>>> Brick4: 10.104.0.1:/gluster/brick/brick2 >>>>>>>> Brick5: 10.104.0.2:/gluster/brick/brick2 >>>>>>>> Brick6: 10.104.0.3:/gluster/brick/brick2 >>>>>>>> Brick7: 10.104.0.1:/gluster/brick/brick3 >>>>>>>> Brick8: 10.104.0.2:/gluster/brick/brick3 >>>>>>>> Brick9: 10.104.0.3:/gluster/brick/brick3 >>>>>>>> Options Reconfigured: >>>>>>>> transport.address-family: inet >>>>>>>> performance.readdir-ahead: on >>>>>>>> nfs.disable: on >>>>>>>> storage.owner-uid: 36 >>>>>>>> storage.owner-gid: 36 >>>>>>>> performance.quick-read: off >>>>>>>> performance.read-ahead: off >>>>>>>> performance.io-cache: off >>>>>>>> performance.stat-prefetch: off >>>>>>>> performance.low-prio-threads: 32 >>>>>>>> network.remote-dio: enable >>>>>>>> cluster.eager-lock: enable >>>>>>>> cluster.quorum-type: auto >>>>>>>> cluster.server-quorum-type: server >>>>>>>> cluster.data-self-heal-algorithm: full >>>>>>>> cluster.locking-scheme: granular >>>>>>>> cluster.shd-max-threads: 8 >>>>>>>> cluster.shd-wait-qlength: 10000 >>>>>>>> features.shard: on >>>>>>>> user.cifs: off >>>>>>>> server.allow-insecure: on >>>>>>>> >>>>>>>> Volume Name: volume2 >>>>>>>> Type: Distributed-Replicate >>>>>>>> Volume ID: 68cfb061-1320-4042-abcd-9228da23c0c8 >>>>>>>> Status: Started >>>>>>>> Snapshot Count: 0 >>>>>>>> Number of Bricks: 3 x 3 = 9 >>>>>>>> Transport-type: tcp >>>>>>>> Bricks: >>>>>>>> Brick1: 10.104.0.1:/gluster2/brick/brick1 >>>>>>>> Brick2: 10.104.0.2:/gluster2/brick/brick1 >>>>>>>> Brick3: 10.104.0.3:/gluster2/brick/brick1 >>>>>>>> Brick4: 10.104.0.1:/gluster2/brick/brick2 >>>>>>>> Brick5: 10.104.0.2:/gluster2/brick/brick2 >>>>>>>> Brick6: 10.104.0.3:/gluster2/brick/brick2 >>>>>>>> Brick7: 10.104.0.1:/gluster2/brick/brick3 >>>>>>>> Brick8: 10.104.0.2:/gluster2/brick/brick3 >>>>>>>> Brick9: 10.104.0.3:/gluster2/brick/brick3 >>>>>>>> Options Reconfigured: >>>>>>>> nfs.disable: on >>>>>>>> performance.readdir-ahead: on >>>>>>>> transport.address-family: inet >>>>>>>> cluster.quorum-type: auto >>>>>>>> network.ping-timeout: 10 >>>>>>>> auth.allow: * >>>>>>>> performance.quick-read: off >>>>>>>> performance.read-ahead: off >>>>>>>> performance.io-cache: off >>>>>>>> performance.stat-prefetch: off >>>>>>>> performance.low-prio-threads: 32 >>>>>>>> network.remote-dio: enable >>>>>>>> cluster.eager-lock: enable >>>>>>>> cluster.server-quorum-type: server >>>>>>>> cluster.data-self-heal-algorithm: full >>>>>>>> cluster.locking-scheme: granular >>>>>>>> cluster.shd-max-threads: 8 >>>>>>>> cluster.shd-wait-qlength: 10000 >>>>>>>> features.shard: on >>>>>>>> user.cifs: off >>>>>>>> storage.owner-uid: 36 >>>>>>>> storage.owner-gid: 36 >>>>>>>> server.allow-insecure: on >>>>>>>> [root@n1 ~]# gluster volume status >>>>>>>> Status of volume: volume1 >>>>>>>> Gluster process TCP Port RDMA Port >>>>>>>> Online Pid >>>>>>>> ------------------------------------------------------------ >>>>>>>> ------------------ >>>>>>>> Brick 10.104.0.1:/gluster/brick/brick1 49152 0 >>>>>>>> Y 3464 >>>>>>>> Brick 10.104.0.2:/gluster/brick/brick1 49152 0 >>>>>>>> Y 68937 >>>>>>>> Brick 10.104.0.3:/gluster/brick/brick1 49161 0 >>>>>>>> Y 94506 >>>>>>>> Brick 10.104.0.1:/gluster/brick/brick2 49153 0 >>>>>>>> Y 3457 >>>>>>>> Brick 10.104.0.2:/gluster/brick/brick2 49153 0 >>>>>>>> Y 68943 >>>>>>>> Brick 10.104.0.3:/gluster/brick/brick2 49162 0 >>>>>>>> Y 94514 >>>>>>>> Brick 10.104.0.1:/gluster/brick/brick3 49154 0 >>>>>>>> Y 3465 >>>>>>>> Brick 10.104.0.2:/gluster/brick/brick3 49154 0 >>>>>>>> Y 68949 >>>>>>>> Brick 10.104.0.3:/gluster/brick/brick3 49163 0 >>>>>>>> Y 94520 >>>>>>>> Self-heal Daemon on localhost N/A N/A Y >>>>>>>> 54356 >>>>>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y >>>>>>>> 962 >>>>>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y >>>>>>>> 108977 >>>>>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y >>>>>>>> 61603 >>>>>>>> >>>>>>>> Task Status of Volume volume1 >>>>>>>> ------------------------------------------------------------ >>>>>>>> ------------------ >>>>>>>> There are no active volume tasks >>>>>>>> >>>>>>>> Status of volume: volume2 >>>>>>>> Gluster process TCP Port RDMA Port >>>>>>>> Online Pid >>>>>>>> ------------------------------------------------------------ >>>>>>>> ------------------ >>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick1 49155 0 >>>>>>>> Y 3852 >>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick1 49158 0 >>>>>>>> Y 68955 >>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick1 49164 0 >>>>>>>> Y 94527 >>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick2 49156 0 >>>>>>>> Y 3851 >>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick2 49159 0 >>>>>>>> Y 68961 >>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick2 49165 0 >>>>>>>> Y 94533 >>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick3 49157 0 >>>>>>>> Y 3883 >>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick3 49160 0 >>>>>>>> Y 68968 >>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick3 49166 0 >>>>>>>> Y 94541 >>>>>>>> Self-heal Daemon on localhost N/A N/A Y >>>>>>>> 54356 >>>>>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y >>>>>>>> 962 >>>>>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y >>>>>>>> 108977 >>>>>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y >>>>>>>> 61603 >>>>>>>> >>>>>>>> Task Status of Volume volume2 >>>>>>>> ------------------------------------------------------------ >>>>>>>> ------------------ >>>>>>>> There are no active volume tasks >>>>>>>> >>>>>>>> I think ovirt can't read valid informations about gluster. >>>>>>>> I can't contiune upgrade of other hosts until this problem exist. >>>>>>>> >>>>>>>> Please help me:) >>>>>>>> >>>>>>>> >>>>>>>> Thanks >>>>>>>> >>>>>>>> Regards, >>>>>>>> >>>>>>>> Tibor >>>>>>>> >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> Users mailing list -- users@ovirt.org >>>>>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>>> _______________________________________________ >>>>> Users mailing list -- users@ovirt.org >>>>> To unsubscribe send an email to users-le...@ovirt.org >>>>> >>>>> >>>> >>>> _______________________________________________ >>>> Users mailing list -- users@ovirt.org >>>> To unsubscribe send an email to users-le...@ovirt.org >>>> >>>> >>>> _______________________________________________ >>>> Users mailing list -- users@ovirt.org >>>> To unsubscribe send an email to users-le...@ovirt.org >>>> oVirt Code of Conduct: https://www.ovirt.org/community/about/community- >>>> guidelines/ >>>> List Archives: >>>> >>>> >>> >> >
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org