I did.

Here are all three nodes with restart. I find it odd ... their has been a
set of messages at end (see below) which I don't know enough about what
oVirt laid out to know if it is bad.

#######
[root@thor vmstore]# systemctl status glusterd
● glusterd.service - GlusterFS, a clustered file-system server
   Loaded: loaded (/usr/lib/systemd/system/glusterd.service; enabled;
vendor preset: disabled)
  Drop-In: /etc/systemd/system/glusterd.service.d
           └─99-cpu.conf
   Active: active (running) since Mon 2020-09-21 20:32:26 EDT; 10h ago
     Docs: man:glusterd(8)
  Process: 2001 ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid
--log-level $LOG_LEVEL $GLUSTERD_OPTIONS (code=exited, status=0/SUCCESS)
 Main PID: 2113 (glusterd)
    Tasks: 151 (limit: 1235410)
   Memory: 3.8G
      CPU: 6min 46.050s
   CGroup: /glusterfs.slice/glusterd.service
           ├─ 2113 /usr/sbin/glusterd -p /var/run/glusterd.pid --log-level
INFO
           ├─ 2914 /usr/sbin/glusterfs -s localhost --volfile-id shd/data
-p /var/run/gluster/shd/data/data-shd.pid -l
/var/log/glusterfs/glustershd.log -S
/var/run/gluster/2f41374c2e36bf4d.socket --xlator-option
*replicate*.node-uu>
           ├─ 9342 /usr/sbin/glusterfsd -s thorst.penguinpages.local
--volfile-id data.thorst.penguinpages.local.gluster_bricks-data-data -p
/var/run/gluster/vols/data/thorst.penguinpages.local-gluster_bricks-data-data.pid
-S /var/r>
           ├─ 9433 /usr/sbin/glusterfsd -s thorst.penguinpages.local
--volfile-id engine.thorst.penguinpages.local.gluster_bricks-engine-engine
-p
/var/run/gluster/vols/engine/thorst.penguinpages.local-gluster_bricks-engine-engine.p>
           ├─ 9444 /usr/sbin/glusterfsd -s thorst.penguinpages.local
--volfile-id
vmstore.thorst.penguinpages.local.gluster_bricks-vmstore-vmstore -p
/var/run/gluster/vols/vmstore/thorst.penguinpages.local-gluster_bricks-vmstore-vms>
           └─35639 /usr/sbin/glusterfsd -s thorst.penguinpages.local
--volfile-id iso.thorst.penguinpages.local.gluster_bricks-iso-iso -p
/var/run/gluster/vols/iso/thorst.penguinpages.local-gluster_bricks-iso-iso.pid
-S /var/run/glu>

Sep 21 20:32:24 thor.penguinpages.local systemd[1]: Starting GlusterFS, a
clustered file-system server...
Sep 21 20:32:26 thor.penguinpages.local systemd[1]: Started GlusterFS, a
clustered file-system server.
Sep 21 20:32:28 thor.penguinpages.local glusterd[2113]: [2020-09-22
00:32:28.605674] C [MSGID: 106003]
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action]
0-management: Server quorum regained for volume data. Starting lo>
Sep 21 20:32:28 thor.penguinpages.local glusterd[2113]: [2020-09-22
00:32:28.639490] C [MSGID: 106003]
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action]
0-management: Server quorum regained for volume engine. Starting >
Sep 21 20:32:28 thor.penguinpages.local glusterd[2113]: [2020-09-22
00:32:28.680665] C [MSGID: 106003]
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action]
0-management: Server quorum regained for volume vmstore. Starting>
Sep 21 20:33:24 thor.penguinpages.local glustershd[2914]: [2020-09-22
00:33:24.813409] C [rpc-clnt-ping.c:155:rpc_clnt_ping_timer_expired]
0-data-client-0: server 172.16.101.101:24007 has not responded in the last
30 seconds, discon>
Sep 21 20:33:24 thor.penguinpages.local glustershd[2914]: [2020-09-22
00:33:24.815147] C [rpc-clnt-ping.c:155:rpc_clnt_ping_timer_expired]
2-engine-client-0: server 172.16.101.101:24007 has not responded in the
last 30 seconds, disc>
Sep 21 20:33:24 thor.penguinpages.local glustershd[2914]: [2020-09-22
00:33:24.818735] C [rpc-clnt-ping.c:155:rpc_clnt_ping_timer_expired]
4-vmstore-client-0: server 172.16.101.101:24007 has not responded in the
last 30 seconds, dis>
Sep 21 20:33:36 thor.penguinpages.local glustershd[2914]: [2020-09-22
00:33:36.816978] C [rpc-clnt-ping.c:155:rpc_clnt_ping_timer_expired]
3-iso-client-0: server 172.16.101.101:24007 has not responded in the last
42 seconds, disconn>
[root@thor vmstore]#
[root@thor vmstore]#
[root@thor vmstore]#
[root@thor vmstore]#
[root@thor vmstore]#
[root@thor vmstore]#
[root@thor vmstore]#
[root@thor vmstore]#
[root@thor vmstore]#
[root@thor vmstore]#
[root@thor vmstore]#
[root@thor vmstore]#
[root@thor vmstore]#
[root@thor vmstore]# systemctl restart glusterd
[root@thor vmstore]# systemctl status glusterd
● glusterd.service - GlusterFS, a clustered file-system server
   Loaded: loaded (/usr/lib/systemd/system/glusterd.service; enabled;
vendor preset: disabled)
  Drop-In: /etc/systemd/system/glusterd.service.d
           └─99-cpu.conf
   Active: active (running) since Tue 2020-09-22 07:24:34 EDT; 2s ago
     Docs: man:glusterd(8)
  Process: 245831 ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid
--log-level $LOG_LEVEL $GLUSTERD_OPTIONS (code=exited, status=0/SUCCESS)
 Main PID: 245832 (glusterd)
    Tasks: 151 (limit: 1235410)
   Memory: 3.8G
      CPU: 132ms
   CGroup: /glusterfs.slice/glusterd.service
           ├─  2914 /usr/sbin/glusterfs -s localhost --volfile-id shd/data
-p /var/run/gluster/shd/data/data-shd.pid -l
/var/log/glusterfs/glustershd.log -S
/var/run/gluster/2f41374c2e36bf4d.socket --xlator-option *replicate*.node-u>
           ├─  9342 /usr/sbin/glusterfsd -s thorst.penguinpages.local
--volfile-id data.thorst.penguinpages.local.gluster_bricks-data-data -p
/var/run/gluster/vols/data/thorst.penguinpages.local-gluster_bricks-data-data.pid
-S /var/>
           ├─  9433 /usr/sbin/glusterfsd -s thorst.penguinpages.local
--volfile-id engine.thorst.penguinpages.local.gluster_bricks-engine-engine
-p
/var/run/gluster/vols/engine/thorst.penguinpages.local-gluster_bricks-engine-engine.>
           ├─  9444 /usr/sbin/glusterfsd -s thorst.penguinpages.local
--volfile-id
vmstore.thorst.penguinpages.local.gluster_bricks-vmstore-vmstore -p
/var/run/gluster/vols/vmstore/thorst.penguinpages.local-gluster_bricks-vmstore-vm>
           ├─ 35639 /usr/sbin/glusterfsd -s thorst.penguinpages.local
--volfile-id iso.thorst.penguinpages.local.gluster_bricks-iso-iso -p
/var/run/gluster/vols/iso/thorst.penguinpages.local-gluster_bricks-iso-iso.pid
-S /var/run/gl>
           └─245832 /usr/sbin/glusterd -p /var/run/glusterd.pid --log-level
INFO

Sep 22 07:24:34 thor.penguinpages.local systemd[1]: Starting GlusterFS, a
clustered file-system server...
Sep 22 07:24:34 thor.penguinpages.local systemd[1]: Started GlusterFS, a
clustered file-system server.
[root@thor vmstore]# gluster volume status
Status of volume: data
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick thorst.penguinpages.local:/gluster_br
icks/data/data                              49152     0          Y
9342
Brick odinst.penguinpages.local:/gluster_br
icks/data/data                              49152     0          Y
3231
Brick medusast.penguinpages.local:/gluster_
bricks/data/data                            49152     0          Y
7819
Self-heal Daemon on localhost               N/A       N/A        Y
245870
Self-heal Daemon on odinst.penguinpages.loc
al                                          N/A       N/A        Y
2693
Self-heal Daemon on medusast.penguinpages.l
ocal                                        N/A       N/A        Y
7863

Task Status of Volume data
------------------------------------------------------------------------------
There are no active volume tasks

Status of volume: engine
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick thorst.penguinpages.local:/gluster_br
icks/engine/engine                          49153     0          Y
9433
Brick odinst.penguinpages.local:/gluster_br
icks/engine/engine                          49153     0          Y
3249
Brick medusast.penguinpages.local:/gluster_
bricks/engine/engine                        49153     0          Y
7830
Self-heal Daemon on localhost               N/A       N/A        Y
245870
Self-heal Daemon on odinst.penguinpages.loc
al                                          N/A       N/A        Y
2693
Self-heal Daemon on medusast.penguinpages.l
ocal                                        N/A       N/A        Y
7863

Task Status of Volume engine
------------------------------------------------------------------------------
There are no active volume tasks

Status of volume: iso
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick thorst.penguinpages.local:/gluster_br
icks/iso/iso                                49155     49156      Y
35639
Brick odinst.penguinpages.local:/gluster_br
icks/iso/iso                                49155     49156      Y
21735
Brick medusast.penguinpages.local:/gluster_
bricks/iso/iso                              49155     49156      Y
21228
Self-heal Daemon on localhost               N/A       N/A        Y
245870
Self-heal Daemon on odinst.penguinpages.loc
al                                          N/A       N/A        Y
2693
Self-heal Daemon on medusast.penguinpages.l
ocal                                        N/A       N/A        Y
7863

Task Status of Volume iso
------------------------------------------------------------------------------
There are no active volume tasks

Status of volume: vmstore
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick thorst.penguinpages.local:/gluster_br
icks/vmstore/vmstore                        49154     0          Y
9444
Brick odinst.penguinpages.local:/gluster_br
icks/vmstore/vmstore                        49154     0          Y
3269
Brick medusast.penguinpages.local:/gluster_
bricks/vmstore/vmstore                      49154     0          Y
7841
Self-heal Daemon on localhost               N/A       N/A        Y
245870
Self-heal Daemon on odinst.penguinpages.loc
al                                          N/A       N/A        Y
2693
Self-heal Daemon on medusast.penguinpages.l
ocal                                        N/A       N/A        Y
7863

Task Status of Volume vmstore
------------------------------------------------------------------------------
There are no active volume tasks

[root@thor vmstore]# ls /gluster_bricks/vmstore/vmstore/
example.log  f118dcae-6162-4e9a-89e4-f30ffcfb9ccf  ns02_20200910.tgz
[root@thor vmstore]#


###  VS nodes that are listed as "ok"

[root@odin vmstore]# systemctl status glusterd
● glusterd.service - GlusterFS, a clustered file-system server
   Loaded: loaded (/usr/lib/systemd/system/glusterd.service; enabled;
vendor preset: disabled)
  Drop-In: /etc/systemd/system/glusterd.service.d
           └─99-cpu.conf
   Active: active (running) since Mon 2020-09-21 20:41:31 EDT; 11h ago
     Docs: man:glusterd(8)
  Process: 1792 ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid
--log-level $LOG_LEVEL $GLUSTERD_OPTIONS (code=exited, status=0/SUCCESS)
 Main PID: 1818 (glusterd)
    Tasks: 149 (limit: 409666)
   Memory: 1.0G
      CPU: 7min 13.719s
   CGroup: /glusterfs.slice/glusterd.service
           ├─ 1818 /usr/sbin/glusterd -p /var/run/glusterd.pid --log-level
INFO
           ├─ 2693 /usr/sbin/glusterfs -s localhost --volfile-id shd/data
-p /var/run/gluster/shd/data/data-shd.pid -l
/var/log/glusterfs/glustershd.log -S
/var/run/gluster/3971f0a4d5e2fd53.socket --xlator-option
*replicate*.node-uu>
           ├─ 3231 /usr/sbin/glusterfsd -s odinst.penguinpages.local
--volfile-id data.odinst.penguinpages.local.gluster_bricks-data-data -p
/var/run/gluster/vols/data/odinst.penguinpages.local-gluster_bricks-data-data.pid
-S /var/r>
           ├─ 3249 /usr/sbin/glusterfsd -s odinst.penguinpages.local
--volfile-id engine.odinst.penguinpages.local.gluster_bricks-engine-engine
-p
/var/run/gluster/vols/engine/odinst.penguinpages.local-gluster_bricks-engine-engine.p>
           ├─ 3269 /usr/sbin/glusterfsd -s odinst.penguinpages.local
--volfile-id
vmstore.odinst.penguinpages.local.gluster_bricks-vmstore-vmstore -p
/var/run/gluster/vols/vmstore/odinst.penguinpages.local-gluster_bricks-vmstore-vms>
           └─21735 /usr/sbin/glusterfsd -s odinst.penguinpages.local
--volfile-id iso.odinst.penguinpages.local.gluster_bricks-iso-iso -p
/var/run/gluster/vols/iso/odinst.penguinpages.local-gluster_bricks-iso-iso.pid
-S /var/run/glu>

Sep 21 20:41:28 odin.penguinpages.local systemd[1]: Starting GlusterFS, a
clustered file-system server...
Sep 21 20:41:31 odin.penguinpages.local systemd[1]: Started GlusterFS, a
clustered file-system server.
Sep 21 20:41:34 odin.penguinpages.local glusterd[1818]: [2020-09-22
00:41:34.478890] C [MSGID: 106003]
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action]
0-management: Server quorum regained for volume data. Starting lo>
Sep 21 20:41:34 odin.penguinpages.local glusterd[1818]: [2020-09-22
00:41:34.483375] C [MSGID: 106003]
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action]
0-management: Server quorum regained for volume engine. Starting >
Sep 21 20:41:34 odin.penguinpages.local glusterd[1818]: [2020-09-22
00:41:34.487583] C [MSGID: 106003]
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action]
0-management: Server quorum regained for volume vmstore. Starting>
[root@odin vmstore]# systemctl restart glusterd
[root@odin vmstore]# systemctl status glusterd
● glusterd.service - GlusterFS, a clustered file-system server
   Loaded: loaded (/usr/lib/systemd/system/glusterd.service; enabled;
vendor preset: disabled)
  Drop-In: /etc/systemd/system/glusterd.service.d
           └─99-cpu.conf
   Active: active (running) since Tue 2020-09-22 07:50:52 EDT; 1s ago
     Docs: man:glusterd(8)
  Process: 141691 ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid
--log-level $LOG_LEVEL $GLUSTERD_OPTIONS (code=exited, status=0/SUCCESS)
 Main PID: 141692 (glusterd)
    Tasks: 134 (limit: 409666)
   Memory: 1.0G
      CPU: 2.265s
   CGroup: /glusterfs.slice/glusterd.service
           ├─  3231 /usr/sbin/glusterfsd -s odinst.penguinpages.local
--volfile-id data.odinst.penguinpages.local.gluster_bricks-data-data -p
/var/run/gluster/vols/data/odinst.penguinpages.local-gluster_bricks-data-data.pid
-S /var/>
           ├─  3249 /usr/sbin/glusterfsd -s odinst.penguinpages.local
--volfile-id engine.odinst.penguinpages.local.gluster_bricks-engine-engine
-p
/var/run/gluster/vols/engine/odinst.penguinpages.local-gluster_bricks-engine-engine.>
           ├─  3269 /usr/sbin/glusterfsd -s odinst.penguinpages.local
--volfile-id
vmstore.odinst.penguinpages.local.gluster_bricks-vmstore-vmstore -p
/var/run/gluster/vols/vmstore/odinst.penguinpages.local-gluster_bricks-vmstore-vm>
           ├─ 21735 /usr/sbin/glusterfsd -s odinst.penguinpages.local
--volfile-id iso.odinst.penguinpages.local.gluster_bricks-iso-iso -p
/var/run/gluster/vols/iso/odinst.penguinpages.local-gluster_bricks-iso-iso.pid
-S /var/run/gl>
           └─141692 /usr/sbin/glusterd -p /var/run/glusterd.pid --log-level
INFO

Sep 22 07:50:49 odin.penguinpages.local systemd[1]: Starting GlusterFS, a
clustered file-system server...
Sep 22 07:50:52 odin.penguinpages.local systemd[1]: Started GlusterFS, a
clustered file-system server.
Sep 22 07:50:52 odin.penguinpages.local glusterd[141692]: [2020-09-22
11:50:52.964585] C [MSGID: 106003]
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action]
0-management: Server quorum regained for volume data. Starting >
Sep 22 07:50:52 odin.penguinpages.local glusterd[141692]: [2020-09-22
11:50:52.969084] C [MSGID: 106003]
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action]
0-management: Server quorum regained for volume engine. Startin>
Sep 22 07:50:52 odin.penguinpages.local glusterd[141692]: [2020-09-22
11:50:52.973197] C [MSGID: 106003]
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action]
0-management: Server quorum regained for volume vmstore. Starti>
lines 1-23/23 (END)
[root@odin vmstore]# ls  /gluster_bricks/vmstore/vmstore/
example.log  f118dcae-6162-4e9a-89e4-f30ffcfb9ccf  ns02_20200910.tgz
 ns02.qcow2  ns02_var.qcow2
[root@odin vmstore]#
##################
[root@medusa sw2_usb_A2]#
[root@medusa sw2_usb_A2]# systemctl status glusterd
● glusterd.service - GlusterFS, a clustered file-system server
   Loaded: loaded (/usr/lib/systemd/system/glusterd.service; enabled;
vendor preset: disabled)
  Drop-In: /etc/systemd/system/glusterd.service.d
           └─99-cpu.conf
   Active: active (running) since Mon 2020-09-21 20:31:29 EDT; 11h ago
     Docs: man:glusterd(8)
  Process: 1713 ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid
--log-level $LOG_LEVEL $GLUSTERD_OPTIONS (code=exited, status=0/SUCCESS)
 Main PID: 1718 (glusterd)
    Tasks: 153 (limit: 409064)
   Memory: 265.5M
      CPU: 10min 10.739s
   CGroup: /glusterfs.slice/glusterd.service
           ├─ 1718 /usr/sbin/glusterd -p /var/run/glusterd.pid --log-level
INFO
           ├─ 7819 /usr/sbin/glusterfsd -s medusast.penguinpages.local
--volfile-id data.medusast.penguinpages.local.gluster_bricks-data-data -p
/var/run/gluster/vols/data/medusast.penguinpages.local-gluster_bricks-data-data.pid
-S >
           ├─ 7830 /usr/sbin/glusterfsd -s medusast.penguinpages.local
--volfile-id
engine.medusast.penguinpages.local.gluster_bricks-engine-engine -p
/var/run/gluster/vols/engine/medusast.penguinpages.local-gluster_bricks-engine-en>
           ├─ 7841 /usr/sbin/glusterfsd -s medusast.penguinpages.local
--volfile-id
vmstore.medusast.penguinpages.local.gluster_bricks-vmstore-vmstore -p
/var/run/gluster/vols/vmstore/medusast.penguinpages.local-gluster_bricks-vmsto>
           ├─ 7863 /usr/sbin/glusterfs -s localhost --volfile-id shd/data
-p /var/run/gluster/shd/data/data-shd.pid -l
/var/log/glusterfs/glustershd.log -S
/var/run/gluster/709d753e1e04185a.socket --xlator-option
*replicate*.node-uu>
           └─21228 /usr/sbin/glusterfsd -s medusast.penguinpages.local
--volfile-id iso.medusast.penguinpages.local.gluster_bricks-iso-iso -p
/var/run/gluster/vols/iso/medusast.penguinpages.local-gluster_bricks-iso-iso.pid
-S /var/r>

Sep 21 20:31:29 medusa.penguinpages.local glusterd[1718]: [2020-09-22
00:31:29.352090] C [MSGID: 106002]
[glusterd-server-quorum.c:355:glusterd_do_volume_quorum_action]
0-management: Server quorum lost for volume engine. Stopping lo>
Sep 21 20:31:29 medusa.penguinpages.local systemd[1]: Started GlusterFS, a
clustered file-system server.
Sep 21 20:31:29 medusa.penguinpages.local glusterd[1718]: [2020-09-22
00:31:29.352297] C [MSGID: 106002]
[glusterd-server-quorum.c:355:glusterd_do_volume_quorum_action]
0-management: Server quorum lost for volume vmstore. Stopping l>
Sep 21 20:32:29 medusa.penguinpages.local glusterd[1718]: [2020-09-22
00:32:29.104708] C [MSGID: 106003]
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action]
0-management: Server quorum regained for volume data. Starting >
Sep 21 20:32:29 medusa.penguinpages.local glusterd[1718]: [2020-09-22
00:32:29.125119] C [MSGID: 106003]
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action]
0-management: Server quorum regained for volume engine. Startin>
Sep 21 20:32:29 medusa.penguinpages.local glusterd[1718]: [2020-09-22
00:32:29.145341] C [MSGID: 106003]
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action]
0-management: Server quorum regained for volume vmstore. Starti>
Sep 21 20:33:24 medusa.penguinpages.local glustershd[7863]: [2020-09-22
00:33:24.815657] C [rpc-clnt-ping.c:155:rpc_clnt_ping_timer_expired]
0-data-client-0: server 172.16.101.101:24007 has not responded in the last
30 seconds, disc>
Sep 21 20:33:24 medusa.penguinpages.local glustershd[7863]: [2020-09-22
00:33:24.817641] C [rpc-clnt-ping.c:155:rpc_clnt_ping_timer_expired]
2-engine-client-0: server 172.16.101.101:24007 has not responded in the
last 30 seconds, di>
Sep 21 20:33:24 medusa.penguinpages.local glustershd[7863]: [2020-09-22
00:33:24.821774] C [rpc-clnt-ping.c:155:rpc_clnt_ping_timer_expired]
4-vmstore-client-0: server 172.16.101.101:24007 has not responded in the
last 30 seconds, d>
Sep 21 20:33:36 medusa.penguinpages.local glustershd[7863]: [2020-09-22
00:33:36.819762] C [rpc-clnt-ping.c:155:rpc_clnt_ping_timer_expired]
3-iso-client-0: server 172.16.101.101:24007 has not responded in the last
42 seconds, disco>
[root@medusa sw2_usb_A2]# systemctl restart glusterd
[root@medusa sw2_usb_A2]# systemctl status glusterd
● glusterd.service - GlusterFS, a clustered file-system server
   Loaded: loaded (/usr/lib/systemd/system/glusterd.service; enabled;
vendor preset: disabled)
  Drop-In: /etc/systemd/system/glusterd.service.d
           └─99-cpu.conf
   Active: active (running) since Tue 2020-09-22 07:51:46 EDT; 2s ago
     Docs: man:glusterd(8)
  Process: 80099 ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid
--log-level $LOG_LEVEL $GLUSTERD_OPTIONS (code=exited, status=0/SUCCESS)
 Main PID: 80100 (glusterd)
    Tasks: 146 (limit: 409064)
   Memory: 207.7M
      CPU: 2.705s
   CGroup: /glusterfs.slice/glusterd.service
           ├─ 7819 /usr/sbin/glusterfsd -s medusast.penguinpages.local
--volfile-id data.medusast.penguinpages.local.gluster_bricks-data-data -p
/var/run/gluster/vols/data/medusast.penguinpages.local-gluster_bricks-data-data.pid
-S >
           ├─ 7830 /usr/sbin/glusterfsd -s medusast.penguinpages.local
--volfile-id
engine.medusast.penguinpages.local.gluster_bricks-engine-engine -p
/var/run/gluster/vols/engine/medusast.penguinpages.local-gluster_bricks-engine-en>
           ├─ 7841 /usr/sbin/glusterfsd -s medusast.penguinpages.local
--volfile-id
vmstore.medusast.penguinpages.local.gluster_bricks-vmstore-vmstore -p
/var/run/gluster/vols/vmstore/medusast.penguinpages.local-gluster_bricks-vmsto>
           ├─21228 /usr/sbin/glusterfsd -s medusast.penguinpages.local
--volfile-id iso.medusast.penguinpages.local.gluster_bricks-iso-iso -p
/var/run/gluster/vols/iso/medusast.penguinpages.local-gluster_bricks-iso-iso.pid
-S /var/r>
           ├─80100 /usr/sbin/glusterd -p /var/run/glusterd.pid --log-level
INFO
           └─80152 /usr/sbin/glusterfs -s localhost --volfile-id shd/data
-p /var/run/gluster/shd/data/data-shd.pid -l
/var/log/glusterfs/glustershd.log -S
/var/run/gluster/709d753e1e04185a.socket --xlator-option
*replicate*.node-uu>

Sep 22 07:51:43 medusa.penguinpages.local systemd[1]: This usually
indicates unclean termination of a previous run, or service implementation
deficiencies.
Sep 22 07:51:43 medusa.penguinpages.local systemd[1]: glusterd.service:
Found left-over process 7863 (glusterfs) in control group while starting
unit. Ignoring.
Sep 22 07:51:43 medusa.penguinpages.local systemd[1]: This usually
indicates unclean termination of a previous run, or service implementation
deficiencies.
Sep 22 07:51:43 medusa.penguinpages.local systemd[1]: glusterd.service:
Found left-over process 21228 (glusterfsd) in control group while starting
unit. Ignoring.
Sep 22 07:51:43 medusa.penguinpages.local systemd[1]: This usually
indicates unclean termination of a previous run, or service implementation
deficiencies.
Sep 22 07:51:43 medusa.penguinpages.local systemd[1]: Starting GlusterFS, a
clustered file-system server...
Sep 22 07:51:46 medusa.penguinpages.local systemd[1]: Started GlusterFS, a
clustered file-system server.
Sep 22 07:51:46 medusa.penguinpages.local glusterd[80100]: [2020-09-22
11:51:46.789628] C [MSGID: 106003]
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action]
0-management: Server quorum regained for volume data. Starting>
Sep 22 07:51:46 medusa.penguinpages.local glusterd[80100]: [2020-09-22
11:51:46.807618] C [MSGID: 106003]
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action]
0-management: Server quorum regained for volume engine. Starti>
Sep 22 07:51:46 medusa.penguinpages.local glusterd[80100]: [2020-09-22
11:51:46.825589] C [MSGID: 106003]
[glusterd-server-quorum.c:348:glusterd_do_volume_quorum_action]
0-management: Server quorum regained for volume vmstore. Start>
[root@medusa sw2_usb_A2]# ls /gluster_bricks/vmstore/vmstore/
example.log  f118dcae-6162-4e9a-89e4-f30ffcfb9ccf  isos  media
 ns01_20200910.tgz  ns02_20200910.tgz  ns02.qcow2  ns02_var.qcow2  qemu


As for files... there is replication issues.  Not really sure how bricks
show ok but it is not replicating



On Tue, Sep 22, 2020 at 2:38 AM Strahil Nikolov <hunter86...@yahoo.com>
wrote:

> Have you restarted glusterd.service on the affected node.
> glusterd is just management layer and it won't affect the brick processes.
>
> Best Regards,
> Strahil Nikolov
>
>
>
>
>
>
> В вторник, 22 септември 2020 г., 01:43:36 Гринуич+3, Jeremey Wise <
> jeremey.w...@gmail.com> написа:
>
>
>
>
>
>
> Start is not an option.
>
> It notes two bricks.  but command line denotes three bricks and all present
>
> [root@odin thorst.penguinpages.local:_vmstore]# gluster volume status data
> Status of volume: data
> Gluster process                             TCP Port  RDMA Port  Online
>  Pid
>
> ------------------------------------------------------------------------------
> Brick thorst.penguinpages.local:/gluster_br
> icks/data/data                              49152     0          Y
> 33123
> Brick odinst.penguinpages.local:/gluster_br
> icks/data/data                              49152     0          Y
> 2970
> Brick medusast.penguinpages.local:/gluster_
> bricks/data/data                            49152     0          Y
> 2646
> Self-heal Daemon on localhost               N/A       N/A        Y
> 3004
> Self-heal Daemon on thorst.penguinpages.loc
> al                                          N/A       N/A        Y
> 33230
> Self-heal Daemon on medusast.penguinpages.l
> ocal                                        N/A       N/A        Y
> 2475
>
> Task Status of Volume data
>
> ------------------------------------------------------------------------------
> There are no active volume tasks
>
> [root@odin thorst.penguinpages.local:_vmstore]# gluster peer status
> Number of Peers: 2
>
> Hostname: thorst.penguinpages.local
> Uuid: 7726b514-e7c3-4705-bbc9-5a90c8a966c9
> State: Peer in Cluster (Connected)
>
> Hostname: medusast.penguinpages.local
> Uuid: 977b2c1d-36a8-4852-b953-f75850ac5031
> State: Peer in Cluster (Connected)
> [root@odin thorst.penguinpages.local:_vmstore]#
>
>
>
>
> On Mon, Sep 21, 2020 at 4:32 PM Strahil Nikolov <hunter86...@yahoo.com>
> wrote:
> > Just select the volume and press "start" . It will automatically mark
> "force start" and will fix itself.
> >
> > Best Regards,
> > Strahil Nikolov
> >
> >
> >
> >
> >
> >
> > В понеделник, 21 септември 2020 г., 20:53:15 Гринуич+3, Jeremey Wise <
> jeremey.w...@gmail.com> написа:
> >
> >
> >
> >
> >
> >
> > oVirt engine shows  one of the gluster servers having an issue.  I did a
> graceful shutdown of all three nodes over weekend as I have to move around
> some power connections in prep for UPS.
> >
> > Came back up.. but....
> >
> >
> >
> > And this is reflected in 2 bricks online (should be three for each
> volume)
> >
> >
> > Command line shows gluster should be happy.
> >
> > [root@thor engine]# gluster peer status
> > Number of Peers: 2
> >
> > Hostname: odinst.penguinpages.local
> > Uuid: 83c772aa-33cd-430f-9614-30a99534d10e
> > State: Peer in Cluster (Connected)
> >
> > Hostname: medusast.penguinpages.local
> > Uuid: 977b2c1d-36a8-4852-b953-f75850ac5031
> > State: Peer in Cluster (Connected)
> > [root@thor engine]#
> >
> > # All bricks showing online
> > [root@thor engine]# gluster volume status
> > Status of volume: data
> > Gluster process                             TCP Port  RDMA Port  Online
>  Pid
> >
> ------------------------------------------------------------------------------
> > Brick thorst.penguinpages.local:/gluster_br
> > icks/data/data                              49152     0          Y
> 11001
> > Brick odinst.penguinpages.local:/gluster_br
> > icks/data/data                              49152     0          Y
> 2970
> > Brick medusast.penguinpages.local:/gluster_
> > bricks/data/data                            49152     0          Y
> 2646
> > Self-heal Daemon on localhost               N/A       N/A        Y
> 50560
> > Self-heal Daemon on odinst.penguinpages.loc
> > al                                          N/A       N/A        Y
> 3004
> > Self-heal Daemon on medusast.penguinpages.l
> > ocal                                        N/A       N/A        Y
> 2475
> >
> > Task Status of Volume data
> >
> ------------------------------------------------------------------------------
> > There are no active volume tasks
> >
> > Status of volume: engine
> > Gluster process                             TCP Port  RDMA Port  Online
>  Pid
> >
> ------------------------------------------------------------------------------
> > Brick thorst.penguinpages.local:/gluster_br
> > icks/engine/engine                          49153     0          Y
> 11012
> > Brick odinst.penguinpages.local:/gluster_br
> > icks/engine/engine                          49153     0          Y
> 2982
> > Brick medusast.penguinpages.local:/gluster_
> > bricks/engine/engine                        49153     0          Y
> 2657
> > Self-heal Daemon on localhost               N/A       N/A        Y
> 50560
> > Self-heal Daemon on odinst.penguinpages.loc
> > al                                          N/A       N/A        Y
> 3004
> > Self-heal Daemon on medusast.penguinpages.l
> > ocal                                        N/A       N/A        Y
> 2475
> >
> > Task Status of Volume engine
> >
> ------------------------------------------------------------------------------
> > There are no active volume tasks
> >
> > Status of volume: iso
> > Gluster process                             TCP Port  RDMA Port  Online
>  Pid
> >
> ------------------------------------------------------------------------------
> > Brick thorst.penguinpages.local:/gluster_br
> > icks/iso/iso                                49156     49157      Y
> 151426
> > Brick odinst.penguinpages.local:/gluster_br
> > icks/iso/iso                                49156     49157      Y
> 69225
> > Brick medusast.penguinpages.local:/gluster_
> > bricks/iso/iso                              49156     49157      Y
> 45018
> > Self-heal Daemon on localhost               N/A       N/A        Y
> 50560
> > Self-heal Daemon on odinst.penguinpages.loc
> > al                                          N/A       N/A        Y
> 3004
> > Self-heal Daemon on medusast.penguinpages.l
> > ocal                                        N/A       N/A        Y
> 2475
> >
> > Task Status of Volume iso
> >
> ------------------------------------------------------------------------------
> > There are no active volume tasks
> >
> > Status of volume: vmstore
> > Gluster process                             TCP Port  RDMA Port  Online
>  Pid
> >
> ------------------------------------------------------------------------------
> > Brick thorst.penguinpages.local:/gluster_br
> > icks/vmstore/vmstore                        49154     0          Y
> 11023
> > Brick odinst.penguinpages.local:/gluster_br
> > icks/vmstore/vmstore                        49154     0          Y
> 2993
> > Brick medusast.penguinpages.local:/gluster_
> > bricks/vmstore/vmstore                      49154     0          Y
> 2668
> > Self-heal Daemon on localhost               N/A       N/A        Y
> 50560
> > Self-heal Daemon on medusast.penguinpages.l
> > ocal                                        N/A       N/A        Y
> 2475
> > Self-heal Daemon on odinst.penguinpages.loc
> > al                                          N/A       N/A        Y
> 3004
> >
> > Task Status of Volume vmstore
> >
> ------------------------------------------------------------------------------
> > There are no active volume tasks
> >
> > [root@thor engine]# gluster volume heal
> > data     engine   iso      vmstore
> > [root@thor engine]# gluster volume heal data info
> > Brick thorst.penguinpages.local:/gluster_bricks/data/data
> > Status: Connected
> > Number of entries: 0
> >
> > Brick odinst.penguinpages.local:/gluster_bricks/data/data
> > Status: Connected
> > Number of entries: 0
> >
> > Brick medusast.penguinpages.local:/gluster_bricks/data/data
> > Status: Connected
> > Number of entries: 0
> >
> > [root@thor engine]# gluster volume heal engine
> > Launching heal operation to perform index self heal on volume engine has
> been successful
> > Use heal info commands to check status.
> > [root@thor engine]# gluster volume heal engine info
> > Brick thorst.penguinpages.local:/gluster_bricks/engine/engine
> > Status: Connected
> > Number of entries: 0
> >
> > Brick odinst.penguinpages.local:/gluster_bricks/engine/engine
> > Status: Connected
> > Number of entries: 0
> >
> > Brick medusast.penguinpages.local:/gluster_bricks/engine/engine
> > Status: Connected
> > Number of entries: 0
> >
> > [root@thor engine]# gluster volume heal vmwatore info
> > Volume vmwatore does not exist
> > Volume heal failed.
> > [root@thor engine]#
> >
> > So not sure what to do with oVirt Engine to make it happy again.
> >
> >
> >
> > --
> > penguinpages
> > _______________________________________________
> > Users mailing list -- users@ovirt.org
> > To unsubscribe send an email to users-le...@ovirt.org
> > Privacy Statement: https://www.ovirt.org/privacy-policy.html
> > oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> > List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/RRHEVF2STA5EVJ5OPRPP6FOYDMLGZM5T/
> >
>
>
> --
> jeremey.w...@gmail.com
> _______________________________________________
> Users mailing list -- users@ovirt.org
> To unsubscribe send an email to users-le...@ovirt.org
> Privacy Statement: https://www.ovirt.org/privacy-policy.html
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/users@ovirt.org/message/ZFVGMWRMELRGEEDJTLSZCK3PRPXBXYEF/
>


-- 
jeremey.w...@gmail.com
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/RCKP7L4BM7K5Y3EAU7K6XGEF3EQ3HK73/

Reply via email to