fs-davids-c1-n1:~ # cat /var/run/gluster/vols/volume1/fs-davids-c1-n1-gluster-brick2-glusterbrick.pid
28851
fs-davids-c1-n1:~ # cat /var/run/gluster/vols/test/fs-davids-c1-n1-gluster-brick3-glusterbrick.pid 
2732
fs-davids-c1-n1:~ # cat /var/run/gluster/vols/volume2/fs-davids-c1-n1-gluster-brick1-glusterbrick.pid 
2642

fs-davids-c1-n1:~ # ps aux | grep glusterfsd
root      2617  0.0  0.0 1208024 23604 ?       Ssl  Jun13   0:31 /usr/sbin/glusterfsd -s fs-davids-c1-n1 --volfile-id fs-ss.fs-davids-c1-n1.gluster-fs-ss-brick -p /var/run/gluster/vols/fs-ss/fs-davids-c1-n1-gluster-fs-ss-brick.pid -S /var/run/gluster/22225c20ed452d41.socket --brick-name /gluster/fs-ss-brick -l /var/log/glusterfs/bricks/gluster-fs-ss-brick.log --xlator-option *-posix.glusterd-uuid=36f2a331-e1a2-4d45-ae43-ee2870573274 --process-name brick --brick-port 49152 --xlator-option fs-ss-server.listen-port=49152
root      2642  0.0  0.0 3434868 27788 ?       Ssl  Jun13   0:07 /usr/sbin/glusterfsd -s fs-davids-c1-n1 --volfile-id volume2.fs-davids-c1-n1.gluster-brick1-glusterbrick -p /var/run/gluster/vols/volume2/fs-davids-c1-n1-gluster-brick1-glusterbrick.pid -S /var/run/gluster/1344420eb8f5103a.socket --brick-name /gluster/brick1/glusterbrick -l /var/log/glusterfs/bricks/gluster-brick1-glusterbrick.log --xlator-option *-posix.glusterd-uuid=36f2a331-e1a2-4d45-ae43-ee2870573274 --process-name brick --brick-port 49153 --xlator-option volume2-server.listen-port=49153
root      2732  0.0  0.0 3369332 23620 ?       Ssl  Jun13   0:07 /usr/sbin/glusterfsd -s fs-davids-c1-n1 --volfile-id test.fs-davids-c1-n1.gluster-brick3-glusterbrick -p /var/run/gluster/vols/test/fs-davids-c1-n1-gluster-brick3-glusterbrick.pid -S /var/run/gluster/51825a6c7f2d23d2.socket --brick-name /gluster/brick3/glusterbrick -l /var/log/glusterfs/bricks/gluster-brick3-glusterbrick.log --xlator-option *-posix.glusterd-uuid=36f2a331-e1a2-4d45-ae43-ee2870573274 --process-name brick --brick-port 49155 --xlator-option test-server.listen-port=49155
root      3047  3.7  0.4 3895908 294760 ?      Ssl  Jun13  39:53 /usr/sbin/glusterfsd -s fs-davids-c1-n1 --volfile-id volume2.fs-davids-c1-n1.gluster-brick1-glusterbrick -p /var/run/gluster/vols/volume2/fs-davids-c1-n1-gluster-brick1-glusterbrick.pid -S /var/run/gluster/1344420eb8f5103a.socket --brick-name /gluster/brick1/glusterbrick -l /var/log/glusterfs/bricks/gluster-brick1-glusterbrick.log --xlator-option *-posix.glusterd-uuid=36f2a331-e1a2-4d45-ae43-ee2870573274 --process-name brick --brick-port 49156 --xlator-option volume2-server.listen-port=49156
root      3057 90.6  1.2 7434224 827520 ?      Ssl  Jun13 964:41 /usr/sbin/glusterfsd -s fs-davids-c1-n1 --volfile-id volume1.fs-davids-c1-n1.gluster-brick2-glusterbrick -p /var/run/gluster/vols/volume1/fs-davids-c1-n1-gluster-brick2-glusterbrick.pid -S /var/run/gluster/472e1e4f2582f66d.socket --brick-name /gluster/brick2/glusterbrick -l /var/log/glusterfs/bricks/gluster-brick2-glusterbrick.log --xlator-option *-posix.glusterd-uuid=36f2a331-e1a2-4d45-ae43-ee2870573274 --process-name brick --brick-port 49157 --xlator-option volume1-server.listen-port=49157
root      3071  0.0  0.0 3829592 33540 ?       Ssl  Jun13   0:41 /usr/sbin/glusterfsd -s fs-davids-c1-n1 --volfile-id test.fs-davids-c1-n1.gluster-brick3-glusterbrick -p /var/run/gluster/vols/test/fs-davids-c1-n1-gluster-brick3-glusterbrick.pid -S /var/run/gluster/51825a6c7f2d23d2.socket --brick-name /gluster/brick3/glusterbrick -l /var/log/glusterfs/bricks/gluster-brick3-glusterbrick.log --xlator-option *-posix.glusterd-uuid=36f2a331-e1a2-4d45-ae43-ee2870573274 --process-name brick --brick-port 49158 --xlator-option test-server.listen-port=49158
root     12807  0.0  0.0   7432   824 pts/0    S+   05:43   0:00 grep --color=auto glusterfsd
root     28851  0.0  0.0 3765940 43880 ?       Ssl  Jun13   0:39 /usr/sbin/glusterfsd -s fs-davids-c1-n1 --volfile-id volume1.fs-davids-c1-n1.gluster-brick2-glusterbrick -p /var/run/gluster/vols/volume1/fs-davids-c1-n1-gluster-brick2-glusterbrick.pid -S /var/run/gluster/472e1e4f2582f66d.socket --brick-name /gluster/brick2/glusterbrick -l /var/log/glusterfs/bricks/gluster-brick2-glusterbrick.log --xlator-option *-posix.glusterd-uuid=36f2a331-e1a2-4d45-ae43-ee2870573274 --process-name brick --brick-port 49154 --xlator-option volume1-server.listen-port=49154
fs-davids-c1-n1:~ # ps -o etime= -p 2617
   17:44:07
fs-davids-c1-n1:~ # ps -o etime= -p 2642
   17:44:28
fs-davids-c1-n1:~ # ps -o etime= -p 2732
   17:44:39
fs-davids-c1-n1:~ # ps -o etime= -p 3047
   17:44:57
fs-davids-c1-n1:~ # ps -o etime= -p 3057
   17:52:51
fs-davids-c1-n1:~ # ps -o etime= -p 3071
   17:52:55
fs-davids-c1-n1:~ # ps -o etime= -p 28851
   15:51:48


fs-davids-c1-n1:~ # gluster vo status
Status of volume: fs-ss
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick fs-davids-c1-n1:/gluster/fs-ss-brick    49152     0          Y       2617 
Brick fs-davids-c1-n2:/gluster/fs-ss-brick    49152     0          Y       2621 
Self-heal Daemon on localhost               N/A       N/A        Y       28907
Self-heal Daemon on fs-davids-c1-n2           N/A       N/A        Y       2816 
 
Task Status of Volume fs-ss
------------------------------------------------------------------------------
There are no active volume tasks
 
Status of volume: volume2
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick fs-davids-c1-n1:/gluster/brick1/gluster
brick                                       49156     0          Y       2642 
Brick fs-davids-c1-n2:/gluster/brick1/gluster
brick                                       49153     0          Y       2640 
Self-heal Daemon on localhost               N/A       N/A        Y       28907
Bitrot Daemon on localhost                  N/A       N/A        Y       28968
Scrubber Daemon on localhost                N/A       N/A        Y       29004
Self-heal Daemon on fs-davids-c1-n2           N/A       N/A        Y       2816 
Bitrot Daemon on fs-davids-c1-n2              N/A       N/A        Y       2984 
Scrubber Daemon on fs-davids-c1-n2            N/A       N/A        Y       2997 
 
Task Status of Volume volume2
------------------------------------------------------------------------------
There are no active volume tasks
 
Status of volume: volume1
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick fs-davids-c1-n1:/gluster/brick2/gluster
brick                                       49154     0          Y       28851
Brick fs-davids-c1-n2:/gluster/brick2/gluster
brick                                       49154     0          Y       2723 
Self-heal Daemon on localhost               N/A       N/A        Y       28907
Bitrot Daemon on localhost                  N/A       N/A        Y       28968
Scrubber Daemon on localhost                N/A       N/A        Y       29004
Self-heal Daemon on fs-davids-c1-n2           N/A       N/A        Y       2816 
Bitrot Daemon on fs-davids-c1-n2              N/A       N/A        Y       2984 
Scrubber Daemon on fs-davids-c1-n2            N/A       N/A        Y       2997 
 
Task Status of Volume volume1
------------------------------------------------------------------------------
There are no active volume tasks
 
Status of volume: test
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick fs-davids-c1-n1:/gluster/brick3/gluster
brick                                       49158     0          Y       2732 
Brick fs-davids-c1-n2:/gluster/brick3/gluster
brick                                       49155     0          Y       2791 
Self-heal Daemon on localhost               N/A       N/A        Y       28907
Bitrot Daemon on localhost                  N/A       N/A        Y       28968
Scrubber Daemon on localhost                N/A       N/A        Y       29004
Self-heal Daemon on fs-davids-c1-n2           N/A       N/A        Y       2816 
Bitrot Daemon on fs-davids-c1-n2              N/A       N/A        Y       2984 
Scrubber Daemon on fs-davids-c1-n2            N/A       N/A        Y       2997 
 
Task Status of Volume test
------------------------------------------------------------------------------
There are no active volume tasks

fs-davids-c1-n1:~ # gluster vo info test
 
Volume Name: test
Type: Replicate
Volume ID: 06d45c73-03cb-4ab5-954e-2d0d18aec6e4
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: fs-davids-c1-n1:/gluster/brick3/glusterbrick
Brick2: fs-davids-c1-n2:/gluster/brick3/glusterbrick
Options Reconfigured:
features.scrub-freq: daily
features.scrub: Active
features.bitrot: on
cluster.quorum-count: 1
cluster.quorum-type: fixed
storage.ctime: on
features.utime: on
storage.build-pgfid: on
performance.write-behind: on
performance.write-behind-window-size: 4MB
performance.read-ahead: off
performance.cache-refresh-timeout: 10
performance.cache-size: 512MB
cluster.use-compound-fops: on
performance.io-thread-count: 64
performance.cache-ima-xattrs: on
performance.cache-samba-metadata: on
performance.md-cache-timeout: 600
performance.cache-invalidation: on
performance.stat-prefetch: on
cluster.lookup-optimize: on
server.event-threads: 32
client.event-threads: 32
performance.nl-cache-timeout: 600
performance.nl-cache: on
features.cache-invalidation-timeout: 600
features.cache-invalidation: on
network.ping-timeout: 10
features.default-retention-period: 3600
features.retention-mode: enterprise
features.worm-file-level: on
features.worm: off
features.read-only: off
user.smb: disable
transport.address-family: inet
nfs.disable: on
performance.client-io-threads: off


