Could you check the engine.log if there are errors related to getting GlusterVolumeAdvancedDetails ?
On Thu, May 10, 2018 at 2:02 PM, Demeter Tibor <[email protected]> wrote: > Dear Ovirt Users, > I've followed up the self-hosted-engine upgrade documentation, I upgraded > my 4.1 system to 4.2.3. > I upgaded the first node with yum upgrade, it seems working now fine. But > since upgrade, the gluster informations seems to displayed incorrect on the > admin panel. The volume yellow, and there are red bricks from that node. > I've checked in console, I think my gluster is not degraded: > > root@n1 ~]# gluster volume list > volume1 > volume2 > [root@n1 ~]# gluster volume info > > Volume Name: volume1 > Type: Distributed-Replicate > Volume ID: e0f568fa-987c-4f5c-b853-01bce718ee27 > Status: Started > Snapshot Count: 0 > Number of Bricks: 3 x 3 = 9 > Transport-type: tcp > Bricks: > Brick1: 10.104.0.1:/gluster/brick/brick1 > Brick2: 10.104.0.2:/gluster/brick/brick1 > Brick3: 10.104.0.3:/gluster/brick/brick1 > Brick4: 10.104.0.1:/gluster/brick/brick2 > Brick5: 10.104.0.2:/gluster/brick/brick2 > Brick6: 10.104.0.3:/gluster/brick/brick2 > Brick7: 10.104.0.1:/gluster/brick/brick3 > Brick8: 10.104.0.2:/gluster/brick/brick3 > Brick9: 10.104.0.3:/gluster/brick/brick3 > Options Reconfigured: > transport.address-family: inet > performance.readdir-ahead: on > nfs.disable: on > storage.owner-uid: 36 > storage.owner-gid: 36 > performance.quick-read: off > performance.read-ahead: off > performance.io-cache: off > performance.stat-prefetch: off > performance.low-prio-threads: 32 > network.remote-dio: enable > cluster.eager-lock: enable > cluster.quorum-type: auto > cluster.server-quorum-type: server > cluster.data-self-heal-algorithm: full > cluster.locking-scheme: granular > cluster.shd-max-threads: 8 > cluster.shd-wait-qlength: 10000 > features.shard: on > user.cifs: off > server.allow-insecure: on > > Volume Name: volume2 > Type: Distributed-Replicate > Volume ID: 68cfb061-1320-4042-abcd-9228da23c0c8 > Status: Started > Snapshot Count: 0 > Number of Bricks: 3 x 3 = 9 > Transport-type: tcp > Bricks: > Brick1: 10.104.0.1:/gluster2/brick/brick1 > Brick2: 10.104.0.2:/gluster2/brick/brick1 > Brick3: 10.104.0.3:/gluster2/brick/brick1 > Brick4: 10.104.0.1:/gluster2/brick/brick2 > Brick5: 10.104.0.2:/gluster2/brick/brick2 > Brick6: 10.104.0.3:/gluster2/brick/brick2 > Brick7: 10.104.0.1:/gluster2/brick/brick3 > Brick8: 10.104.0.2:/gluster2/brick/brick3 > Brick9: 10.104.0.3:/gluster2/brick/brick3 > Options Reconfigured: > nfs.disable: on > performance.readdir-ahead: on > transport.address-family: inet > cluster.quorum-type: auto > network.ping-timeout: 10 > auth.allow: * > performance.quick-read: off > performance.read-ahead: off > performance.io-cache: off > performance.stat-prefetch: off > performance.low-prio-threads: 32 > network.remote-dio: enable > cluster.eager-lock: enable > cluster.server-quorum-type: server > cluster.data-self-heal-algorithm: full > cluster.locking-scheme: granular > cluster.shd-max-threads: 8 > cluster.shd-wait-qlength: 10000 > features.shard: on > user.cifs: off > storage.owner-uid: 36 > storage.owner-gid: 36 > server.allow-insecure: on > [root@n1 ~]# gluster volume status > Status of volume: volume1 > Gluster process TCP Port RDMA Port Online > Pid > ------------------------------------------------------------ > ------------------ > Brick 10.104.0.1:/gluster/brick/brick1 49152 0 Y > 3464 > Brick 10.104.0.2:/gluster/brick/brick1 49152 0 Y > 68937 > Brick 10.104.0.3:/gluster/brick/brick1 49161 0 Y > 94506 > Brick 10.104.0.1:/gluster/brick/brick2 49153 0 Y > 3457 > Brick 10.104.0.2:/gluster/brick/brick2 49153 0 Y > 68943 > Brick 10.104.0.3:/gluster/brick/brick2 49162 0 Y > 94514 > Brick 10.104.0.1:/gluster/brick/brick3 49154 0 Y > 3465 > Brick 10.104.0.2:/gluster/brick/brick3 49154 0 Y > 68949 > Brick 10.104.0.3:/gluster/brick/brick3 49163 0 Y > 94520 > Self-heal Daemon on localhost N/A N/A Y > 54356 > Self-heal Daemon on 10.104.0.2 N/A N/A Y > 962 > Self-heal Daemon on 10.104.0.3 N/A N/A Y > 108977 > Self-heal Daemon on 10.104.0.4 N/A N/A Y > 61603 > > Task Status of Volume volume1 > ------------------------------------------------------------ > ------------------ > There are no active volume tasks > > Status of volume: volume2 > Gluster process TCP Port RDMA Port Online > Pid > ------------------------------------------------------------ > ------------------ > Brick 10.104.0.1:/gluster2/brick/brick1 49155 0 Y > 3852 > Brick 10.104.0.2:/gluster2/brick/brick1 49158 0 Y > 68955 > Brick 10.104.0.3:/gluster2/brick/brick1 49164 0 Y > 94527 > Brick 10.104.0.1:/gluster2/brick/brick2 49156 0 Y > 3851 > Brick 10.104.0.2:/gluster2/brick/brick2 49159 0 Y > 68961 > Brick 10.104.0.3:/gluster2/brick/brick2 49165 0 Y > 94533 > Brick 10.104.0.1:/gluster2/brick/brick3 49157 0 Y > 3883 > Brick 10.104.0.2:/gluster2/brick/brick3 49160 0 Y > 68968 > Brick 10.104.0.3:/gluster2/brick/brick3 49166 0 Y > 94541 > Self-heal Daemon on localhost N/A N/A Y > 54356 > Self-heal Daemon on 10.104.0.2 N/A N/A Y > 962 > Self-heal Daemon on 10.104.0.3 N/A N/A Y > 108977 > Self-heal Daemon on 10.104.0.4 N/A N/A Y > 61603 > > Task Status of Volume volume2 > ------------------------------------------------------------ > ------------------ > There are no active volume tasks > > I think ovirt can't read valid informations about gluster. > I can't contiune upgrade of other hosts until this problem exist. > > Please help me:) > > > Thanks > > Regards, > > Tibor > > > _______________________________________________ > Users mailing list -- [email protected] > To unsubscribe send an email to [email protected] > >
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected]

