Good morning Atin, Thanks for the reply.
I believe that log file is "rhev-data-center-mnt-glusterSD-ovirt-hyp-01.reis.com:_data.log", please correct me if I'm wrong. However, it happens to be empty. See below: ls -lah /var/log/glusterfs/|grep data -rw-------. 1 root root 0 Jun 13 17:09 glfsheal-data.log -rw-------. 1 root root 34K Jun 4 03:06 glfsheal-data.log-20170604.gz -rw-------. 1 root root 563K Jun 7 16:01 glfsheal-data.log-20170613 *-rw-------. 1 root root 0 Jun 13 17:09 rhev-data-center-mnt-glusterSD-ovirt-hyp-01.reis.com:_data.log* -rw-------. 1 root root 61K Jun 4 03:08 rhev-data-center-mnt-glusterSD-ovirt-hyp-01.reis.com:_data.log-20170604.gz -rw-------. 1 root root 164K Jun 8 08:58 rhev-data-center-mnt-glusterSD-ovirt-hyp-01.reis.com:_data.log-20170613 -rw-------. 1 root root 0 Jun 4 03:08 rhev-data-center-mnt-glusterSD-ovirt-hyp-01.reis.com:_engine.log -rw-------. 1 root root 371 Jun 28 03:30 rhev-data-center-mnt-glusterSD-ovirt-hyp-01.reis.com:engine.log -rw-------. 1 root root 16K May 31 14:12 rhev-data-center-mnt-glusterSD-ovirt-hyp-01.reis.com:_engine.log-20170604 -rw-------. 1 root root 4.8K Jun 4 03:08 rhev-data-center-mnt-glusterSD-ovirt-hyp-01.reis.com:engine.log-20170604.gz -rw-------. 1 root root 34K Jun 13 17:09 rhev-data-center-mnt-glusterSD-ovirt-hyp-01.reis.com:engine.log-20170613.gz -rw-------. 1 root root 21K Jun 18 03:10 rhev-data-center-mnt-glusterSD-ovirt-hyp-01.reis.com:engine.log-20170618.gz -rw-------. 1 root root 32K Jun 25 03:26 rhev-data-center-mnt-glusterSD-ovirt-hyp-01.reis.com:engine.log-20170625 [root@ovirt-hyp-01 ~]# cat /var/log/glusterfs/rhev-data-center-mnt-glusterSD-ovirt-hyp-01.reis.com: _data.log [root@ovirt-hyp-01 ~]# Please let me know what other information I can provide. Thank you, Joel On Wed, Jun 28, 2017 at 12:08 AM, Atin Mukherjee <[email protected]> wrote: > The mount log file of the volume would help in debugging the actual cause. > > On Tue, Jun 27, 2017 at 6:33 PM, Joel Diaz <[email protected]> wrote: > >> Good morning Gluster users, >> >> I'm very new to the Gluster file system. My apologies if this is not the >> correct way to seek assistance. However, I would appreciate some insight >> into understanding the issue I have. >> >> I have three nodes running two volumes, engine and data. The third node >> is the arbiter on both volumes. Both volumes were operation fine but one of >> the volumes, data, no longer mounts. >> >> Please see below: >> >> gluster volume info all >> >> Volume Name: data >> Type: Replicate >> Volume ID: 1d6bb110-9be4-4630-ae91-36ec1cf6cc02 >> Status: Started >> Snapshot Count: 0 >> Number of Bricks: 1 x (2 + 1) = 3 >> Transport-type: tcp >> Bricks: >> Brick1: 192.168.170.141:/gluster_bricks/data/data >> Brick2: 192.168.170.143:/gluster_bricks/data/data >> Brick3: 192.168.170.147:/gluster_bricks/data/data (arbiter) >> Options Reconfigured: >> nfs.disable: on >> performance.readdir-ahead: on >> transport.address-family: inet >> performance.quick-read: off >> performance.read-ahead: off >> performance.io-cache: off >> performance.stat-prefetch: off >> performance.low-prio-threads: 32 >> network.remote-dio: off >> cluster.eager-lock: enable >> cluster.quorum-type: auto >> cluster.server-quorum-type: server >> cluster.data-self-heal-algorithm: full >> cluster.locking-scheme: granular >> cluster.shd-max-threads: 8 >> cluster.shd-wait-qlength: 10000 >> features.shard: on >> user.cifs: off >> storage.owner-uid: 36 >> storage.owner-gid: 36 >> network.ping-timeout: 30 >> performance.strict-o-direct: on >> cluster.granular-entry-heal: enable >> >> Volume Name: engine >> Type: Replicate >> Volume ID: b160f0b2-8bd3-4ff2-a07c-134cab1519dd >> Status: Started >> Snapshot Count: 0 >> Number of Bricks: 1 x (2 + 1) = 3 >> Transport-type: tcp >> Bricks: >> Brick1: 192.168.170.141:/gluster_bricks/engine/engine >> Brick2: 192.168.170.143:/gluster_bricks/engine/engine >> Brick3: 192.168.170.147:/gluster_bricks/engine/engine (arbiter) >> Options Reconfigured: >> nfs.disable: on >> performance.readdir-ahead: on >> transport.address-family: inet >> performance.quick-read: off >> performance.read-ahead: off >> performance.io-cache: off >> performance.stat-prefetch: off >> performance.low-prio-threads: 32 >> network.remote-dio: off >> cluster.eager-lock: enable >> cluster.quorum-type: auto >> cluster.server-quorum-type: server >> cluster.data-self-heal-algorithm: full >> cluster.locking-scheme: granular >> cluster.shd-max-threads: 8 >> cluster.shd-wait-qlength: 10000 >> features.shard: on >> user.cifs: off >> storage.owner-uid: 36 >> storage.owner-gid: 36 >> network.ping-timeout: 30 >> performance.strict-o-direct: on >> cluster.granular-entry-heal: enable >> >> df -h >> Filesystem Size Used Avail Use% >> Mounted on >> /dev/mapper/centos_ovirt--hyp--01-root 50G 3.9G 47G 8% / >> devtmpfs 7.7G 0 7.7G 0% /dev >> tmpfs 7.8G 0 7.8G 0% >> /dev/shm >> tmpfs 7.8G 8.7M 7.7G 1% /run >> tmpfs 7.8G 0 7.8G 0% >> /sys/fs/cgroup >> /dev/mapper/centos_ovirt--hyp--01-home 61G 33M 61G 1% /home >> /dev/mapper/gluster_vg_sdb-gluster_lv_engine 50G 8.1G 42G 17% >> /gluster_bricks/engine >> /dev/sda1 497M 173M 325M 35% /boot >> /dev/mapper/gluster_vg_sdb-gluster_lv_data 730G 157G 574G 22% >> /gluster_bricks/data >> tmpfs 1.6G 0 1.6G 0% >> /run/user/0 >> ovirt-hyp-01.reis.com:engine 50G 8.1G 42G 17% >> /rhev/data-center/mnt/glusterSD/ovirt-hyp-01.reis.com:engine >> >> gluster volume status data >> Status of volume: data >> Gluster process TCP Port RDMA Port Online >> Pid >> ------------------------------------------------------------ >> ------------------ >> Brick 192.168.170.141:/gluster_bricks/data/ >> data 49157 0 Y >> 11967 >> Brick 192.168.170.143:/gluster_bricks/data/ >> data 49157 0 Y >> 2901 >> Brick 192.168.170.147:/gluster_bricks/data/ >> data 49158 0 Y >> 2626 >> Self-heal Daemon on localhost N/A N/A Y >> 16211 >> Self-heal Daemon on 192.168.170.147 N/A N/A Y >> 3402 >> Self-heal Daemon on 192.168.170.143 N/A N/A Y >> 20254 >> >> Task Status of Volume data >> ------------------------------------------------------------ >> ------------------ >> There are no active volume tasks >> >> gluster peer status >> Number of Peers: 2 >> >> Hostname: 192.168.170.143 >> Uuid: b2b30d05-cf91-4567-92fd-022575e082f5 >> State: Peer in Cluster (Connected) >> Other names: >> 10.0.0.2 >> >> Hostname: 192.168.170.147 >> Uuid: 4e50acc4-f3cb-422d-b499-fb5796a53529 >> State: Peer in Cluster (Connected) >> Other names: >> 10.0.0.3 >> >> Any assistance in understanding how and why the volume no longer mounts >> and a possible resolution would be greatly appreciated. >> >> Thank you, >> >> Joel >> >> _______________________________________________ >> Gluster-users mailing list >> [email protected] >> http://lists.gluster.org/mailman/listinfo/gluster-users >> > >
_______________________________________________ Gluster-users mailing list [email protected] http://lists.gluster.org/mailman/listinfo/gluster-users
