Hi, I have to update the engine again?
Thanks, R Tibor ----- 2018. máj.. 18., 6:47, Sahina Bose <[email protected]> írta: > Thanks for reporting this. [ https://gerrit.ovirt.org/91375 | > https://gerrit.ovirt.org/91375 ] fixes this. I've re-opened bug [ > https://bugzilla.redhat.com/show_bug.cgi?id=1574508 | > https://bugzilla.redhat.com/show_bug.cgi?id=1574508 ] > On Thu, May 17, 2018 at 10:12 PM, Demeter Tibor < [ > mailto:[email protected] | > [email protected] ] > wrote: >> Hi, >> 4.2.4-0.0.master.20180515183442.git00e1340.el7.centos >> Firstly, I did a yum update "ovirt-*-setup*" >> second, I have ran engine-setup to upgrade. >> I didn't remove the old repos, just installed the nightly repo. >> Thank you again, >> Regards, >> Tibor >> ----- 2018. máj.. 17., 15:02, Sahina Bose < [ mailto:[email protected] | >> [email protected] ] > írta: >>> It doesn't look like the patch was applied. Still see the same error in >>> engine.log >>> "Error while refreshing brick statuses for volume 'volume1' of cluster >>> 'C6220': >>> null"\ >>> Did you use engine-setup to upgrade? What's the version of ovirt-engine >>> currently installed? >>> On Thu, May 17, 2018 at 5:10 PM, Demeter Tibor < [ >>> mailto:[email protected] | >>> [email protected] ] > wrote: >>>> Hi, >>>> sure, >>>> Thank you for your time! >>>> R >>>> Tibor >>>> ----- 2018. máj.. 17., 12:19, Sahina Bose < [ mailto:[email protected] | >>>> [email protected] ] > írta: >>>>> [+users] >>>>> Can you provide the engine.log to see why the monitoring is not working >>>>> here. >>>>> thanks! >>>>> On Wed, May 16, 2018 at 2:08 PM, Demeter Tibor < [ >>>>> mailto:[email protected] | >>>>> [email protected] ] > wrote: >>>>>> Hi, >>>>>> Meanwhile, I did the upgrade engine, but the gluster state is same on my >>>>>> first >>>>>> node. >>>>>> I've attached some screenshot of my problem. >>>>>> Thanks >>>>>> Tibor >>>>>> ----- 2018. máj.. 16., 10:16, Demeter Tibor < [ >>>>>> mailto:[email protected] | >>>>>> [email protected] ] > írta Hi, >>>>>>> If 4.3.4 will release, i just have to remove the nightly repo and >>>>>>> update to >>>>>>> stable? >>>>>>> I'm sorry for my terrible English, I try to explain what was my problem >>>>>>> with >>>>>>> update. >>>>>>> I'm upgraded from 4.1.8. >>>>>>> I followed up the official hosted-engine update documentation, that was >>>>>>> not >>>>>>> clear me, because it has referenced to a lot of old thing (i think). >>>>>>> [ https://www.ovirt.org/documentation/upgrade-guide/upgrade-guide/ | >>>>>>> https://www.ovirt.org/documentation/upgrade-guide/upgrade-guide/ ] >>>>>>> [ >>>>>>> https://www.ovirt.org/documentation/how-to/hosted-engine/#upgrade-hosted-engine >>>>>>> | >>>>>>> https://www.ovirt.org/documentation/how-to/hosted-engine/#upgrade-hosted-engine >>>>>>> ] >>>>>>> Maybe it need to update, because I had a lot of question under upgrade >>>>>>> and I was >>>>>>> not sure in all of necessary steps. For example, If I need to >>>>>>> installing the >>>>>>> new, 4.2 repo on the hosts, then need to remove the old repo from that? >>>>>>> Why I need to do a" yum update -y" on hosts, meanwhile there is an >>>>>>> "Updatehost" >>>>>>> menu in the GUI? So, maybe it outdated. >>>>>>> Since upgrade hosted engine, and the first node, I have problems with >>>>>>> gluster. >>>>>>> It seems to working fine if you check it from console "gluster volume >>>>>>> status, >>>>>>> etc" but not on the Gui, because now it yellow, and the brick reds in >>>>>>> the first >>>>>>> node. >>>>>>> Previously I did a mistake with glusterfs, my gluster config was wrong. >>>>>>> I have >>>>>>> corrected them, but it did not helped to me,gluster bricks are reds on >>>>>>> my first >>>>>>> node yet.... >>>>>>> Now I try to upgrade to nightly, but I'm affraid, because it a living, >>>>>>> productive system, and I don't have downtime. I hope it will help me. >>>>>>> Thanks for all, >>>>>>> Regards, >>>>>>> Tibor Demeter >>>>>>> ----- 2018. máj.. 16., 9:58, Sahina Bose < [ mailto:[email protected] | >>>>>>> [email protected] ] > írta: >>>>>>>> On Wed, May 16, 2018 at 1:19 PM, Demeter Tibor < [ >>>>>>>> mailto:[email protected] | >>>>>>>> [email protected] ] > wrote: >>>>>>>>> Hi, >>>>>>>>> is it a different, unstable repo? I have a productive cluster, how is >>>>>>>>> safe that? >>>>>>>>> I don't have any experience with nightly build. How can I use this? >>>>>>>>> It have to >>>>>>>>> install to the engine VM or all of my hosts? >>>>>>>>> Thanks in advance for help me.. >>>>>>>> Only on the engine VM. >>>>>>>> Regarding stability - it passes CI so relatively stable, beyond that >>>>>>>> there are >>>>>>>> no guarantees. >>>>>>>> What's the specific problem you're facing with update? Can you >>>>>>>> elaborate? >>>>>>>>> Regards, >>>>>>>>> Tibor >>>>>>>>> ----- 2018. máj.. 15., 9:58, Demeter Tibor < [ >>>>>>>>> mailto:[email protected] | >>>>>>>>> [email protected] ] > írta: >>>>>>>>>> Hi, >>>>>>>>>> Could you explain how can I use this patch? >>>>>>>>>> R, >>>>>>>>>> Tibor >>>>>>>>>> ----- 2018. máj.. 14., 11:18, Demeter Tibor < [ >>>>>>>>>> mailto:[email protected] | >>>>>>>>>> [email protected] ] > írta: >>>>>>>>>>> Hi, >>>>>>>>>>> Sorry for my question, but can you tell me please how can I use >>>>>>>>>>> this patch? >>>>>>>>>>> Thanks, >>>>>>>>>>> Regards, >>>>>>>>>>> Tibor >>>>>>>>>>> ----- 2018. máj.. 14., 10:47, Sahina Bose < [ >>>>>>>>>>> mailto:[email protected] | >>>>>>>>>>> [email protected] ] > írta: >>>>>>>>>>>> On Sat, May 12, 2018 at 1:14 PM, Demeter Tibor < [ >>>>>>>>>>>> mailto:[email protected] | >>>>>>>>>>>> [email protected] ] > wrote: >>>>>>>>>>>>> Hi, >>>>>>>>>>>>> Could someone help me please ? I can't finish my upgrade process. >>>>>>>>>>>> [ https://gerrit.ovirt.org/91164 | https://gerrit.ovirt.org/91164 >>>>>>>>>>>> ] should fix >>>>>>>>>>>> the error you're facing. >>>>>>>>>>>> Can you elaborate why this is affecting the upgrade process? >>>>>>>>>>>>> Thanks >>>>>>>>>>>>> R >>>>>>>>>>>>> Tibor >>>>>>>>>>>>> ----- 2018. máj.. 10., 12:51, Demeter Tibor < [ >>>>>>>>>>>>> mailto:[email protected] | >>>>>>>>>>>>> [email protected] ] > írta: >>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>> I've attached the vdsm and supervdsm logs. But I don't have >>>>>>>>>>>>>> engine.log here, >>>>>>>>>>>>>> because that is on hosted engine vm. Should I send that ? >>>>>>>>>>>>>> Thank you >>>>>>>>>>>>>> Regards, >>>>>>>>>>>>>> Tibor >>>>>>>>>>>>>> ----- 2018. máj.. 10., 12:30, Sahina Bose < [ >>>>>>>>>>>>>> mailto:[email protected] | >>>>>>>>>>>>>> [email protected] ] > írta: >>>>>>>>>>>>>>> There's a bug here. Can you log one attaching this engine.log >>>>>>>>>>>>>>> and also vdsm.log >>>>>>>>>>>>>>> & supervdsm.log from n3.itsmart.cloud >>>>>>>>>>>>>>> On Thu, May 10, 2018 at 3:35 PM, Demeter Tibor < [ >>>>>>>>>>>>>>> mailto:[email protected] | >>>>>>>>>>>>>>> [email protected] ] > wrote: >>>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>>> I found this: >>>>>>>>>>>>>>>> 2018-05-10 03:24:19,096+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return: >>>>>>>>>>>>>>>> org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@ca97448e, >>>>>>>>>>>>>>>> log id: 347435ae >>>>>>>>>>>>>>>> 2018-05-10 03:24:19,097+02 ERROR >>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) >>>>>>>>>>>>>>>> [43f4eaec] Error while refreshing brick statuses for volume >>>>>>>>>>>>>>>> 'volume2' of >>>>>>>>>>>>>>>> cluster 'C6220': null >>>>>>>>>>>>>>>> 2018-05-10 03:24:19,097+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) >>>>>>>>>>>>>>>> [7715ceda] Failed to acquire lock and wait lock >>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>>>> 2018-05-10 03:24:19,104+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>> n4.itsmart.cloud, >>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), >>>>>>>>>>>>>>>> log id: 6908121d >>>>>>>>>>>>>>>> 2018-05-10 03:24:19,106+02 ERROR >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command >>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>> n4.itsmart.cloud, >>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' >>>>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>>>> 2018-05-10 03:24:19,106+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6908121d >>>>>>>>>>>>>>>> 2018-05-10 03:24:19,107+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>> n1.itsmart.cloud, >>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'}), >>>>>>>>>>>>>>>> log id: 735c6a5f >>>>>>>>>>>>>>>> 2018-05-10 03:24:19,109+02 ERROR >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command >>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>> n1.itsmart.cloud, >>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' >>>>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>>>> 2018-05-10 03:24:19,109+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 735c6a5f >>>>>>>>>>>>>>>> 2018-05-10 03:24:19,110+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>> n2.itsmart.cloud, >>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>>>>>>>>>>>> log id: 6f9e9f58 >>>>>>>>>>>>>>>> 2018-05-10 03:24:19,112+02 ERROR >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command >>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>> n2.itsmart.cloud, >>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' >>>>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>>>> 2018-05-10 03:24:19,112+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 6f9e9f58 >>>>>>>>>>>>>>>> 2018-05-10 03:24:19,113+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>> n3.itsmart.cloud, >>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), >>>>>>>>>>>>>>>> log id: 2ee46967 >>>>>>>>>>>>>>>> 2018-05-10 03:24:19,115+02 ERROR >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] Command >>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>> n3.itsmart.cloud, >>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' >>>>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>>>> 2018-05-10 03:24:19,116+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 2ee46967 >>>>>>>>>>>>>>>> 2018-05-10 03:24:19,117+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] START, >>>>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = >>>>>>>>>>>>>>>> n1.itsmart.cloud, >>>>>>>>>>>>>>>> GlusterVolumeAdvancedDetailsVDSParameters:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57', >>>>>>>>>>>>>>>> volumeName='volume1'}), log id: 7550e5c >>>>>>>>>>>>>>>> 2018-05-10 03:24:20,748+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) [43f4eaec] FINISH, >>>>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return: >>>>>>>>>>>>>>>> org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@4a46066f, >>>>>>>>>>>>>>>> log id: 7550e5c >>>>>>>>>>>>>>>> 2018-05-10 03:24:20,749+02 ERROR >>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler7) >>>>>>>>>>>>>>>> [43f4eaec] Error while refreshing brick statuses for volume >>>>>>>>>>>>>>>> 'volume1' of >>>>>>>>>>>>>>>> cluster 'C6220': null >>>>>>>>>>>>>>>> 2018-05-10 03:24:20,750+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] START, >>>>>>>>>>>>>>>> GlusterServersListVDSCommand(HostName = n2.itsmart.cloud, >>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>>>>>>>>>>>> log id: 120cc68d >>>>>>>>>>>>>>>> 2018-05-10 03:24:20,930+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, >>>>>>>>>>>>>>>> GlusterServersListVDSCommand, >>>>>>>>>>>>>>>> return: [ [ http://10.101.0.2/24:CONNECTED | >>>>>>>>>>>>>>>> 10.101.0.2/24:CONNECTED ] , >>>>>>>>>>>>>>>> n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED, >>>>>>>>>>>>>>>> 10.104.0.4:CONNECTED], log >>>>>>>>>>>>>>>> id: 120cc68d >>>>>>>>>>>>>>>> 2018-05-10 03:24:20,949+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] START, >>>>>>>>>>>>>>>> GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud, >>>>>>>>>>>>>>>> GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>>>>>>>>>>>> log id: 118aa264 >>>>>>>>>>>>>>>> 2018-05-10 03:24:21,048+02 WARN >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick1' of volume >>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as >>>>>>>>>>>>>>>> no gluster >>>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>> 2018-05-10 03:24:21,055+02 WARN >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick2' of volume >>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as >>>>>>>>>>>>>>>> no gluster >>>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>> 2018-05-10 03:24:21,061+02 WARN >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick3' of volume >>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as >>>>>>>>>>>>>>>> no gluster >>>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>> 2018-05-10 03:24:21,067+02 WARN >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick1' of volume >>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as >>>>>>>>>>>>>>>> no gluster >>>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>> 2018-05-10 03:24:21,074+02 WARN >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick2' of volume >>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as >>>>>>>>>>>>>>>> no gluster >>>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>> 2018-05-10 03:24:21,080+02 WARN >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] Could not associate brick >>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick3' of volume >>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as >>>>>>>>>>>>>>>> no gluster >>>>>>>>>>>>>>>> network found in cluster '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>> 2018-05-10 03:24:21,081+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler8) [7715ceda] FINISH, >>>>>>>>>>>>>>>> GlusterVolumesListVDSCommand, >>>>>>>>>>>>>>>> return: >>>>>>>>>>>>>>>> {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d, >>>>>>>>>>>>>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.g >>>>>>>>>>>>>>>> luster.GlusterVolumeEntity@f88c521b}, log id: 118aa264 >>>>>>>>>>>>>>>> 2018-05-10 11:59:26,047+02 ERROR >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] Command >>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>> n1.itsmart.cloud, >>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='8e737bab-e0bb-4f16-ab85-e24e91882f57'})' >>>>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>>>> 2018-05-10 11:59:26,047+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 14a71ef0 >>>>>>>>>>>>>>>> 2018-05-10 11:59:26,048+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] START, >>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>> n4.itsmart.cloud, >>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'}), >>>>>>>>>>>>>>>> log id: 28d9e255 >>>>>>>>>>>>>>>> 2018-05-10 11:59:26,051+02 ERROR >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] Command >>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>> n4.itsmart.cloud, >>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='3ddef95f-158d-407c-a7d8-49641e012755'})' >>>>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>>>> 2018-05-10 11:59:26,051+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 28d9e255 >>>>>>>>>>>>>>>> 2018-05-10 11:59:26,052+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] START, >>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>> n2.itsmart.cloud, >>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>>>>>>>>>>>> log id: 4a7b280e >>>>>>>>>>>>>>>> 2018-05-10 11:59:26,054+02 ERROR >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] Command >>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>> n2.itsmart.cloud, >>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'})' >>>>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>>>> 2018-05-10 11:59:26,054+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 4a7b280e >>>>>>>>>>>>>>>> 2018-05-10 11:59:26,055+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] START, >>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>> n3.itsmart.cloud, >>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'}), >>>>>>>>>>>>>>>> log id: 18adc534 >>>>>>>>>>>>>>>> 2018-05-10 11:59:26,057+02 ERROR >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] Command >>>>>>>>>>>>>>>> 'GetGlusterLocalLogicalVolumeListVDSCommand(HostName = >>>>>>>>>>>>>>>> n3.itsmart.cloud, >>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec'})' >>>>>>>>>>>>>>>> execution failed: null >>>>>>>>>>>>>>>> 2018-05-10 11:59:26,057+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterLocalLogicalVolumeListVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>>>>>>>>>>>> GetGlusterLocalLogicalVolumeListVDSCommand, log id: 18adc534 >>>>>>>>>>>>>>>> 2018-05-10 11:59:26,058+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] START, >>>>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand(HostName = >>>>>>>>>>>>>>>> n3.itsmart.cloud, >>>>>>>>>>>>>>>> GlusterVolumeAdvancedDetailsVDSParameters:{hostId='fd2ee743-f5d4-403b-ba18-377e309169ec', >>>>>>>>>>>>>>>> volumeName='volume1'}), log id: 3451084f >>>>>>>>>>>>>>>> 2018-05-10 11:59:28,050+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler3) >>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>>>> 2018-05-10 11:59:28,060+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler3) >>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>>>> 2018-05-10 11:59:28,062+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler3) >>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>>>> 2018-05-10 11:59:31,054+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler3) >>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>>>> 2018-05-10 11:59:31,054+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler3) >>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>>>> 2018-05-10 11:59:31,062+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler3) >>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>>>> 2018-05-10 11:59:31,064+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.lock.InMemoryLockManager] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler3) >>>>>>>>>>>>>>>> [2eb1c389] Failed to acquire lock and wait lock >>>>>>>>>>>>>>>> 'EngineLock:{exclusiveLocks='[59c10db3-0324-0320-0120-000000000339=GLUSTER]', >>>>>>>>>>>>>>>> sharedLocks=''}' >>>>>>>>>>>>>>>> 2018-05-10 11:59:31,465+02 INFO >>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GetGlusterVolumeAdvancedDetailsVDSCommand] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) [400fa486] FINISH, >>>>>>>>>>>>>>>> GetGlusterVolumeAdvancedDetailsVDSCommand, return: >>>>>>>>>>>>>>>> org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeAdvancedDetails@3f1b7f43, >>>>>>>>>>>>>>>> log id: 3451084f >>>>>>>>>>>>>>>> 2018-05-10 11:59:31,466+02 ERROR >>>>>>>>>>>>>>>> [org.ovirt.engine.core.bll.gluster.GlusterSyncJob] >>>>>>>>>>>>>>>> (DefaultQuartzScheduler4) >>>>>>>>>>>>>>>> [400fa486] Error while refreshing brick statuses for volume >>>>>>>>>>>>>>>> 'volume1' of >>>>>>>>>>>>>>>> cluster 'C6220': null >>>>>>>>>>>>>>>> R >>>>>>>>>>>>>>>> Tibor >>>>>>>>>>>>>>>> ----- 2018. máj.. 10., 11:43, Sahina Bose < [ >>>>>>>>>>>>>>>> mailto:[email protected] | >>>>>>>>>>>>>>>> [email protected] ] > írta: >>>>>>>>>>>>>>>>> This doesn't affect the monitoring of state. >>>>>>>>>>>>>>>>> Any errors in vdsm.log? >>>>>>>>>>>>>>>>> Or errors in engine.log of the form "Error while refreshing >>>>>>>>>>>>>>>>> brick statuses for >>>>>>>>>>>>>>>>> volume" >>>>>>>>>>>>>>>>> On Thu, May 10, 2018 at 2:33 PM, Demeter Tibor < [ >>>>>>>>>>>>>>>>> mailto:[email protected] | >>>>>>>>>>>>>>>>> [email protected] ] > wrote: >>>>>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>>>>> Thank you for your fast reply :) >>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,574+02 INFO >>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] >>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START, >>>>>>>>>>>>>>>>>> GlusterServersListVDSCommand(HostName = n2.itsmart.cloud, >>>>>>>>>>>>>>>>>> VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>>>>>>>>>>>>>> log id: 39adbbb8 >>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,768+02 INFO >>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] >>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, >>>>>>>>>>>>>>>>>> GlusterServersListVDSCommand, >>>>>>>>>>>>>>>>>> return: [ [ http://10.101.0.2/24:CONNECTED | >>>>>>>>>>>>>>>>>> 10.101.0.2/24:CONNECTED ] , >>>>>>>>>>>>>>>>>> n1.cloudata.local:CONNECTED, 10.104.0.3:CONNECTED, >>>>>>>>>>>>>>>>>> 10.104.0.4:CONNECTED], log >>>>>>>>>>>>>>>>>> id: 39adbbb8 >>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,788+02 INFO >>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] START, >>>>>>>>>>>>>>>>>> GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud, >>>>>>>>>>>>>>>>>> GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), >>>>>>>>>>>>>>>>>> log id: 738a7261 >>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,892+02 WARN >>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate >>>>>>>>>>>>>>>>>> brick >>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick1' of volume >>>>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network >>>>>>>>>>>>>>>>>> as no gluster >>>>>>>>>>>>>>>>>> network found in cluster >>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,898+02 WARN >>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate >>>>>>>>>>>>>>>>>> brick >>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick2' of volume >>>>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network >>>>>>>>>>>>>>>>>> as no gluster >>>>>>>>>>>>>>>>>> network found in cluster >>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,905+02 WARN >>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate >>>>>>>>>>>>>>>>>> brick >>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster/brick/brick3' of volume >>>>>>>>>>>>>>>>>> 'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network >>>>>>>>>>>>>>>>>> as no gluster >>>>>>>>>>>>>>>>>> network found in cluster >>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,911+02 WARN >>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate >>>>>>>>>>>>>>>>>> brick >>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick1' of volume >>>>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network >>>>>>>>>>>>>>>>>> as no gluster >>>>>>>>>>>>>>>>>> network found in cluster >>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,917+02 WARN >>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate >>>>>>>>>>>>>>>>>> brick >>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick2' of volume >>>>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network >>>>>>>>>>>>>>>>>> as no gluster >>>>>>>>>>>>>>>>>> network found in cluster >>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,924+02 WARN >>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] >>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] Could not associate >>>>>>>>>>>>>>>>>> brick >>>>>>>>>>>>>>>>>> '10.104.0.1:/gluster2/brick/brick3' of volume >>>>>>>>>>>>>>>>>> '68cfb061-1320-4042-abcd-9228da23c0c8' with correct network >>>>>>>>>>>>>>>>>> as no gluster >>>>>>>>>>>>>>>>>> network found in cluster >>>>>>>>>>>>>>>>>> '59c10db3-0324-0320-0120-000000000339' >>>>>>>>>>>>>>>>>> 2018-05-10 11:01:51,925+02 INFO >>>>>>>>>>>>>>>>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] >>>>>>>>>>>>>>>>>> (DefaultQuartzScheduler6) [7f01fc2d] FINISH, >>>>>>>>>>>>>>>>>> GlusterVolumesListVDSCommand, >>>>>>>>>>>>>>>>>> return: >>>>>>>>>>>>>>>>>> {68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d, >>>>>>>>>>>>>>>>>> e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@f88c521b}, >>>>>>>>>>>>>>>>>> log id: 738a7261 >>>>>>>>>>>>>>>>>> This happening continuously. >>>>>>>>>>>>>>>>>> Thanks! >>>>>>>>>>>>>>>>>> Tibor >>>>>>>>>>>>>>>>>> ----- 2018. máj.. 10., 10:56, Sahina Bose < [ >>>>>>>>>>>>>>>>>> mailto:[email protected] | >>>>>>>>>>>>>>>>>> [email protected] ] > írta: >>>>>>>>>>>>>>>>>>> Could you check the engine.log if there are errors related >>>>>>>>>>>>>>>>>>> to getting >>>>>>>>>>>>>>>>>>> GlusterVolumeAdvancedDetails ? >>>>>>>>>>>>>>>>>>> On Thu, May 10, 2018 at 2:02 PM, Demeter Tibor < [ >>>>>>>>>>>>>>>>>>> mailto:[email protected] | >>>>>>>>>>>>>>>>>>> [email protected] ] > wrote: >>>>>>>>>>>>>>>>>>>> Dear Ovirt Users, >>>>>>>>>>>>>>>>>>>> I've followed up the self-hosted-engine upgrade >>>>>>>>>>>>>>>>>>>> documentation, I upgraded my 4.1 >>>>>>>>>>>>>>>>>>>> system to 4.2.3. >>>>>>>>>>>>>>>>>>>> I upgaded the first node with yum upgrade, it seems >>>>>>>>>>>>>>>>>>>> working now fine. But since >>>>>>>>>>>>>>>>>>>> upgrade, the gluster informations seems to displayed >>>>>>>>>>>>>>>>>>>> incorrect on the admin >>>>>>>>>>>>>>>>>>>> panel. The volume yellow, and there are red bricks from >>>>>>>>>>>>>>>>>>>> that node. >>>>>>>>>>>>>>>>>>>> I've checked in console, I think my gluster is not >>>>>>>>>>>>>>>>>>>> degraded: >>>>>>>>>>>>>>>>>>>> root@n1 ~]# gluster volume list >>>>>>>>>>>>>>>>>>>> volume1 >>>>>>>>>>>>>>>>>>>> volume2 >>>>>>>>>>>>>>>>>>>> [root@n1 ~]# gluster volume info >>>>>>>>>>>>>>>>>>>> Volume Name: volume1 >>>>>>>>>>>>>>>>>>>> Type: Distributed-Replicate >>>>>>>>>>>>>>>>>>>> Volume ID: e0f568fa-987c-4f5c-b853-01bce718ee27 >>>>>>>>>>>>>>>>>>>> Status: Started >>>>>>>>>>>>>>>>>>>> Snapshot Count: 0 >>>>>>>>>>>>>>>>>>>> Number of Bricks: 3 x 3 = 9 >>>>>>>>>>>>>>>>>>>> Transport-type: tcp >>>>>>>>>>>>>>>>>>>> Bricks: >>>>>>>>>>>>>>>>>>>> Brick1: 10.104.0.1:/gluster/brick/brick1 >>>>>>>>>>>>>>>>>>>> Brick2: 10.104.0.2:/gluster/brick/brick1 >>>>>>>>>>>>>>>>>>>> Brick3: 10.104.0.3:/gluster/brick/brick1 >>>>>>>>>>>>>>>>>>>> Brick4: 10.104.0.1:/gluster/brick/brick2 >>>>>>>>>>>>>>>>>>>> Brick5: 10.104.0.2:/gluster/brick/brick2 >>>>>>>>>>>>>>>>>>>> Brick6: 10.104.0.3:/gluster/brick/brick2 >>>>>>>>>>>>>>>>>>>> Brick7: 10.104.0.1:/gluster/brick/brick3 >>>>>>>>>>>>>>>>>>>> Brick8: 10.104.0.2:/gluster/brick/brick3 >>>>>>>>>>>>>>>>>>>> Brick9: 10.104.0.3:/gluster/brick/brick3 >>>>>>>>>>>>>>>>>>>> Options Reconfigured: >>>>>>>>>>>>>>>>>>>> transport.address-family: inet >>>>>>>>>>>>>>>>>>>> performance.readdir-ahead: on >>>>>>>>>>>>>>>>>>>> nfs.disable: on >>>>>>>>>>>>>>>>>>>> storage.owner-uid: 36 >>>>>>>>>>>>>>>>>>>> storage.owner-gid: 36 >>>>>>>>>>>>>>>>>>>> performance.quick-read: off >>>>>>>>>>>>>>>>>>>> performance.read-ahead: off >>>>>>>>>>>>>>>>>>>> performance.io-cache: off >>>>>>>>>>>>>>>>>>>> performance.stat-prefetch: off >>>>>>>>>>>>>>>>>>>> performance.low-prio-threads: 32 >>>>>>>>>>>>>>>>>>>> network.remote-dio: enable >>>>>>>>>>>>>>>>>>>> cluster.eager-lock: enable >>>>>>>>>>>>>>>>>>>> cluster.quorum-type: auto >>>>>>>>>>>>>>>>>>>> cluster.server-quorum-type: server >>>>>>>>>>>>>>>>>>>> cluster.data-self-heal-algorithm: full >>>>>>>>>>>>>>>>>>>> cluster.locking-scheme: granular >>>>>>>>>>>>>>>>>>>> cluster.shd-max-threads: 8 >>>>>>>>>>>>>>>>>>>> cluster.shd-wait-qlength: 10000 >>>>>>>>>>>>>>>>>>>> features.shard: on >>>>>>>>>>>>>>>>>>>> user.cifs: off >>>>>>>>>>>>>>>>>>>> server.allow-insecure: on >>>>>>>>>>>>>>>>>>>> Volume Name: volume2 >>>>>>>>>>>>>>>>>>>> Type: Distributed-Replicate >>>>>>>>>>>>>>>>>>>> Volume ID: 68cfb061-1320-4042-abcd-9228da23c0c8 >>>>>>>>>>>>>>>>>>>> Status: Started >>>>>>>>>>>>>>>>>>>> Snapshot Count: 0 >>>>>>>>>>>>>>>>>>>> Number of Bricks: 3 x 3 = 9 >>>>>>>>>>>>>>>>>>>> Transport-type: tcp >>>>>>>>>>>>>>>>>>>> Bricks: >>>>>>>>>>>>>>>>>>>> Brick1: 10.104.0.1:/gluster2/brick/brick1 >>>>>>>>>>>>>>>>>>>> Brick2: 10.104.0.2:/gluster2/brick/brick1 >>>>>>>>>>>>>>>>>>>> Brick3: 10.104.0.3:/gluster2/brick/brick1 >>>>>>>>>>>>>>>>>>>> Brick4: 10.104.0.1:/gluster2/brick/brick2 >>>>>>>>>>>>>>>>>>>> Brick5: 10.104.0.2:/gluster2/brick/brick2 >>>>>>>>>>>>>>>>>>>> Brick6: 10.104.0.3:/gluster2/brick/brick2 >>>>>>>>>>>>>>>>>>>> Brick7: 10.104.0.1:/gluster2/brick/brick3 >>>>>>>>>>>>>>>>>>>> Brick8: 10.104.0.2:/gluster2/brick/brick3 >>>>>>>>>>>>>>>>>>>> Brick9: 10.104.0.3:/gluster2/brick/brick3 >>>>>>>>>>>>>>>>>>>> Options Reconfigured: >>>>>>>>>>>>>>>>>>>> nfs.disable: on >>>>>>>>>>>>>>>>>>>> performance.readdir-ahead: on >>>>>>>>>>>>>>>>>>>> transport.address-family: inet >>>>>>>>>>>>>>>>>>>> cluster.quorum-type: auto >>>>>>>>>>>>>>>>>>>> network.ping-timeout: 10 >>>>>>>>>>>>>>>>>>>> auth.allow: * >>>>>>>>>>>>>>>>>>>> performance.quick-read: off >>>>>>>>>>>>>>>>>>>> performance.read-ahead: off >>>>>>>>>>>>>>>>>>>> performance.io-cache: off >>>>>>>>>>>>>>>>>>>> performance.stat-prefetch: off >>>>>>>>>>>>>>>>>>>> performance.low-prio-threads: 32 >>>>>>>>>>>>>>>>>>>> network.remote-dio: enable >>>>>>>>>>>>>>>>>>>> cluster.eager-lock: enable >>>>>>>>>>>>>>>>>>>> cluster.server-quorum-type: server >>>>>>>>>>>>>>>>>>>> cluster.data-self-heal-algorithm: full >>>>>>>>>>>>>>>>>>>> cluster.locking-scheme: granular >>>>>>>>>>>>>>>>>>>> cluster.shd-max-threads: 8 >>>>>>>>>>>>>>>>>>>> cluster.shd-wait-qlength: 10000 >>>>>>>>>>>>>>>>>>>> features.shard: on >>>>>>>>>>>>>>>>>>>> user.cifs: off >>>>>>>>>>>>>>>>>>>> storage.owner-uid: 36 >>>>>>>>>>>>>>>>>>>> storage.owner-gid: 36 >>>>>>>>>>>>>>>>>>>> server.allow-insecure: on >>>>>>>>>>>>>>>>>>>> [root@n1 ~]# gluster volume status >>>>>>>>>>>>>>>>>>>> Status of volume: volume1 >>>>>>>>>>>>>>>>>>>> Gluster process TCP Port RDMA Port Online Pid >>>>>>>>>>>>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick1 49152 0 Y 3464 >>>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick1 49152 0 Y 68937 >>>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick1 49161 0 Y 94506 >>>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick2 49153 0 Y 3457 >>>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick2 49153 0 Y 68943 >>>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick2 49162 0 Y 94514 >>>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster/brick/brick3 49154 0 Y 3465 >>>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster/brick/brick3 49154 0 Y 68949 >>>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster/brick/brick3 49163 0 Y 94520 >>>>>>>>>>>>>>>>>>>> Self-heal Daemon on localhost N/A N/A Y 54356 >>>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y 962 >>>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977 >>>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603 >>>>>>>>>>>>>>>>>>>> Task Status of Volume volume1 >>>>>>>>>>>>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>>>>>>>>>>>> There are no active volume tasks >>>>>>>>>>>>>>>>>>>> Status of volume: volume2 >>>>>>>>>>>>>>>>>>>> Gluster process TCP Port RDMA Port Online Pid >>>>>>>>>>>>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick1 49155 0 Y 3852 >>>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick1 49158 0 Y 68955 >>>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick1 49164 0 Y 94527 >>>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick2 49156 0 Y 3851 >>>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick2 49159 0 Y 68961 >>>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick2 49165 0 Y 94533 >>>>>>>>>>>>>>>>>>>> Brick 10.104.0.1:/gluster2/brick/brick3 49157 0 Y 3883 >>>>>>>>>>>>>>>>>>>> Brick 10.104.0.2:/gluster2/brick/brick3 49160 0 Y 68968 >>>>>>>>>>>>>>>>>>>> Brick 10.104.0.3:/gluster2/brick/brick3 49166 0 Y 94541 >>>>>>>>>>>>>>>>>>>> Self-heal Daemon on localhost N/A N/A Y 54356 >>>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.2 N/A N/A Y 962 >>>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977 >>>>>>>>>>>>>>>>>>>> Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603 >>>>>>>>>>>>>>>>>>>> Task Status of Volume volume2 >>>>>>>>>>>>>>>>>>>> ------------------------------------------------------------------------------ >>>>>>>>>>>>>>>>>>>> There are no active volume tasks >>>>>>>>>>>>>>>>>>>> I think ovirt can't read valid informations about gluster. >>>>>>>>>>>>>>>>>>>> I can't contiune upgrade of other hosts until this problem >>>>>>>>>>>>>>>>>>>> exist. >>>>>>>>>>>>>>>>>>>> Please help me:) >>>>>>>>>>>>>>>>>>>> Thanks >>>>>>>>>>>>>>>>>>>> Regards, >>>>>>>>>>>>>>>>>>>> Tibor >>>>>>>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>>>>>>> Users mailing list -- [ mailto:[email protected] | >>>>>>>>>>>>>>>>>>>> [email protected] ] >>>>>>>>>>>>>>>>>>>> To unsubscribe send an email to [ >>>>>>>>>>>>>>>>>>>> mailto:[email protected] | >>>>>>>>>>>>>>>>>>>> [email protected] ] >>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>> Users mailing list -- [ mailto:[email protected] | [email protected] >>>>>>>>>>>>>> ] >>>>>>>>>>>>>> To unsubscribe send an email to [ mailto:[email protected] | >>>>>>>>>>>>>> [email protected] ] >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> Users mailing list -- [ mailto:[email protected] | [email protected] ] >>>>>>>>>>> To unsubscribe send an email to [ mailto:[email protected] | >>>>>>>>>>> [email protected] ] >>>>>>>>>> _______________________________________________ >>>>>>>>>> Users mailing list -- [ mailto:[email protected] | [email protected] ] >>>>>>>>>> To unsubscribe send an email to [ mailto:[email protected] | >>>>>>>>>> [email protected] ] >>>>>>>>>> oVirt Code of Conduct: [ >>>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/ | >>>>>>>>>> https://www.ovirt.org/community/about/community-guidelines/ ] >>>>>>>>>> List Archives:
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected]

