Hi, 

Thank you for your fast reply :) 

2018-05-10 11:01:51,574+02 INFO 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] 
(DefaultQuartzScheduler6) [7f01fc2d] START, 
GlusterServersListVDSCommand(HostName = n2.itsmart.cloud, 
VdsIdVDSCommandParametersBase:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}), 
log id: 39adbbb8 
2018-05-10 11:01:51,768+02 INFO 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand] 
(DefaultQuartzScheduler6) [7f01fc2d] FINISH, GlusterServersListVDSCommand, 
return: [10.101.0.2/24:CONNECTED, n1.cloudata.local:CONNECTED, 
10.104.0.3:CONNECTED, 10.104.0.4:CONNECTED], log id: 39adbbb8 
2018-05-10 11:01:51,788+02 INFO 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] 
(DefaultQuartzScheduler6) [7f01fc2d] START, 
GlusterVolumesListVDSCommand(HostName = n2.itsmart.cloud, 
GlusterVolumesListVDSParameters:{hostId='06e361ef-3361-4eaa-9923-27fa1a0187a4'}),
 log id: 738a7261 
2018-05-10 11:01:51,892+02 WARN 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] 
(DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick 
'10.104.0.1:/gluster/brick/brick1' of volume 
'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster 
network found in cluster '59c10db3-0324-0320-0120-000000000339' 
2018-05-10 11:01:51,898+02 WARN 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] 
(DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick 
'10.104.0.1:/gluster/brick/brick2' of volume 
'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster 
network found in cluster '59c10db3-0324-0320-0120-000000000339' 
2018-05-10 11:01:51,905+02 WARN 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] 
(DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick 
'10.104.0.1:/gluster/brick/brick3' of volume 
'e0f568fa-987c-4f5c-b853-01bce718ee27' with correct network as no gluster 
network found in cluster '59c10db3-0324-0320-0120-000000000339' 
2018-05-10 11:01:51,911+02 WARN 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] 
(DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick 
'10.104.0.1:/gluster2/brick/brick1' of volume 
'68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster 
network found in cluster '59c10db3-0324-0320-0120-000000000339' 
2018-05-10 11:01:51,917+02 WARN 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] 
(DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick 
'10.104.0.1:/gluster2/brick/brick2' of volume 
'68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster 
network found in cluster '59c10db3-0324-0320-0120-000000000339' 
2018-05-10 11:01:51,924+02 WARN 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListReturn] 
(DefaultQuartzScheduler6) [7f01fc2d] Could not associate brick 
'10.104.0.1:/gluster2/brick/brick3' of volume 
'68cfb061-1320-4042-abcd-9228da23c0c8' with correct network as no gluster 
network found in cluster '59c10db3-0324-0320-0120-000000000339' 
2018-05-10 11:01:51,925+02 INFO 
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand] 
(DefaultQuartzScheduler6) [7f01fc2d] FINISH, GlusterVolumesListVDSCommand, 
return: 
{68cfb061-1320-4042-abcd-9228da23c0c8=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@7a6720d,
 
e0f568fa-987c-4f5c-b853-01bce718ee27=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@f88c521b},
 log id: 738a7261 

This happening continuously. 

Thanks! 
Tibor 

----- 2018. máj.. 10., 10:56, Sahina Bose <sab...@redhat.com> írta: 

> Could you check the engine.log if there are errors related to getting
> GlusterVolumeAdvancedDetails ?

> On Thu, May 10, 2018 at 2:02 PM, Demeter Tibor < [ mailto:tdeme...@itsmart.hu 
> |
> tdeme...@itsmart.hu ] > wrote:

>> Dear Ovirt Users,
>> I've followed up the self-hosted-engine upgrade documentation, I upgraded my 
>> 4.1
>> system to 4.2.3.
>> I upgaded the first node with yum upgrade, it seems working now fine. But 
>> since
>> upgrade, the gluster informations seems to displayed incorrect on the admin
>> panel. The volume yellow, and there are red bricks from that node.
>> I've checked in console, I think my gluster is not degraded:

>> root@n1 ~]# gluster volume list
>> volume1
>> volume2
>> [root@n1 ~]# gluster volume info
>> Volume Name: volume1
>> Type: Distributed-Replicate
>> Volume ID: e0f568fa-987c-4f5c-b853-01bce718ee27
>> Status: Started
>> Snapshot Count: 0
>> Number of Bricks: 3 x 3 = 9
>> Transport-type: tcp
>> Bricks:
>> Brick1: 10.104.0.1:/gluster/brick/brick1
>> Brick2: 10.104.0.2:/gluster/brick/brick1
>> Brick3: 10.104.0.3:/gluster/brick/brick1
>> Brick4: 10.104.0.1:/gluster/brick/brick2
>> Brick5: 10.104.0.2:/gluster/brick/brick2
>> Brick6: 10.104.0.3:/gluster/brick/brick2
>> Brick7: 10.104.0.1:/gluster/brick/brick3
>> Brick8: 10.104.0.2:/gluster/brick/brick3
>> Brick9: 10.104.0.3:/gluster/brick/brick3
>> Options Reconfigured:
>> transport.address-family: inet
>> performance.readdir-ahead: on
>> nfs.disable: on
>> storage.owner-uid: 36
>> storage.owner-gid: 36
>> performance.quick-read: off
>> performance.read-ahead: off
>> performance.io-cache: off
>> performance.stat-prefetch: off
>> performance.low-prio-threads: 32
>> network.remote-dio: enable
>> cluster.eager-lock: enable
>> cluster.quorum-type: auto
>> cluster.server-quorum-type: server
>> cluster.data-self-heal-algorithm: full
>> cluster.locking-scheme: granular
>> cluster.shd-max-threads: 8
>> cluster.shd-wait-qlength: 10000
>> features.shard: on
>> user.cifs: off
>> server.allow-insecure: on
>> Volume Name: volume2
>> Type: Distributed-Replicate
>> Volume ID: 68cfb061-1320-4042-abcd-9228da23c0c8
>> Status: Started
>> Snapshot Count: 0
>> Number of Bricks: 3 x 3 = 9
>> Transport-type: tcp
>> Bricks:
>> Brick1: 10.104.0.1:/gluster2/brick/brick1
>> Brick2: 10.104.0.2:/gluster2/brick/brick1
>> Brick3: 10.104.0.3:/gluster2/brick/brick1
>> Brick4: 10.104.0.1:/gluster2/brick/brick2
>> Brick5: 10.104.0.2:/gluster2/brick/brick2
>> Brick6: 10.104.0.3:/gluster2/brick/brick2
>> Brick7: 10.104.0.1:/gluster2/brick/brick3
>> Brick8: 10.104.0.2:/gluster2/brick/brick3
>> Brick9: 10.104.0.3:/gluster2/brick/brick3
>> Options Reconfigured:
>> nfs.disable: on
>> performance.readdir-ahead: on
>> transport.address-family: inet
>> cluster.quorum-type: auto
>> network.ping-timeout: 10
>> auth.allow: *
>> performance.quick-read: off
>> performance.read-ahead: off
>> performance.io-cache: off
>> performance.stat-prefetch: off
>> performance.low-prio-threads: 32
>> network.remote-dio: enable
>> cluster.eager-lock: enable
>> cluster.server-quorum-type: server
>> cluster.data-self-heal-algorithm: full
>> cluster.locking-scheme: granular
>> cluster.shd-max-threads: 8
>> cluster.shd-wait-qlength: 10000
>> features.shard: on
>> user.cifs: off
>> storage.owner-uid: 36
>> storage.owner-gid: 36
>> server.allow-insecure: on
>> [root@n1 ~]# gluster volume status
>> Status of volume: volume1
>> Gluster process TCP Port RDMA Port Online Pid
>> ------------------------------------------------------------------------------
>> Brick 10.104.0.1:/gluster/brick/brick1 49152 0 Y 3464
>> Brick 10.104.0.2:/gluster/brick/brick1 49152 0 Y 68937
>> Brick 10.104.0.3:/gluster/brick/brick1 49161 0 Y 94506
>> Brick 10.104.0.1:/gluster/brick/brick2 49153 0 Y 3457
>> Brick 10.104.0.2:/gluster/brick/brick2 49153 0 Y 68943
>> Brick 10.104.0.3:/gluster/brick/brick2 49162 0 Y 94514
>> Brick 10.104.0.1:/gluster/brick/brick3 49154 0 Y 3465
>> Brick 10.104.0.2:/gluster/brick/brick3 49154 0 Y 68949
>> Brick 10.104.0.3:/gluster/brick/brick3 49163 0 Y 94520
>> Self-heal Daemon on localhost N/A N/A Y 54356
>> Self-heal Daemon on 10.104.0.2 N/A N/A Y 962
>> Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977
>> Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603
>> Task Status of Volume volume1
>> ------------------------------------------------------------------------------
>> There are no active volume tasks
>> Status of volume: volume2
>> Gluster process TCP Port RDMA Port Online Pid
>> ------------------------------------------------------------------------------
>> Brick 10.104.0.1:/gluster2/brick/brick1 49155 0 Y 3852
>> Brick 10.104.0.2:/gluster2/brick/brick1 49158 0 Y 68955
>> Brick 10.104.0.3:/gluster2/brick/brick1 49164 0 Y 94527
>> Brick 10.104.0.1:/gluster2/brick/brick2 49156 0 Y 3851
>> Brick 10.104.0.2:/gluster2/brick/brick2 49159 0 Y 68961
>> Brick 10.104.0.3:/gluster2/brick/brick2 49165 0 Y 94533
>> Brick 10.104.0.1:/gluster2/brick/brick3 49157 0 Y 3883
>> Brick 10.104.0.2:/gluster2/brick/brick3 49160 0 Y 68968
>> Brick 10.104.0.3:/gluster2/brick/brick3 49166 0 Y 94541
>> Self-heal Daemon on localhost N/A N/A Y 54356
>> Self-heal Daemon on 10.104.0.2 N/A N/A Y 962
>> Self-heal Daemon on 10.104.0.3 N/A N/A Y 108977
>> Self-heal Daemon on 10.104.0.4 N/A N/A Y 61603
>> Task Status of Volume volume2
>> ------------------------------------------------------------------------------
>> There are no active volume tasks
>> I think ovirt can't read valid informations about gluster.
>> I can't contiune upgrade of other hosts until this problem exist.

>> Please help me:)

>> Thanks

>> Regards,

>> Tibor

>> _______________________________________________
>> Users mailing list -- [ mailto:users@ovirt.org | users@ovirt.org ]
>> To unsubscribe send an email to [ mailto:users-le...@ovirt.org |
>> users-le...@ovirt.org ]
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org

Reply via email to