No firewalls in this case... -- Gene Liverman Systems Administrator Information Technology Services University of West Georgia [email protected]
On Jun 10, 2014 12:57 PM, "Paul Robert Marino" <[email protected]> wrote: > Ive also seen this happen when there is a firewall in the middle and > nfslockd malfunctioned because of it. > > > On Tue, Jun 10, 2014 at 12:20 PM, Gene Liverman <[email protected]> > wrote: > > Thanks! I turned off drc as suggested and will have to wait and see how > that > > works. Here are the packages I have installed via yum: > > # rpm -qa |grep -i gluster > > glusterfs-cli-3.5.0-2.el6.x86_64 > > glusterfs-libs-3.5.0-2.el6.x86_64 > > glusterfs-fuse-3.5.0-2.el6.x86_64 > > glusterfs-server-3.5.0-2.el6.x86_64 > > glusterfs-3.5.0-2.el6.x86_64 > > glusterfs-geo-replication-3.5.0-2.el6.x86_64 > > > > The nfs server service was showing to be running even when stuff wasn't > > working. This is from while it was broken: > > > > # gluster volume status > > Status of volume: gv0 > > Gluster process Port > > Online Pid > > > ------------------------------------------------------------------------------------------------------------ > > Brick eapps-gluster01.my.domain:/export/sdb1/gv0 49152 Y 39593 > > Brick eapps-gluster02.my.domain:/export/sdb1/gv0 49152 Y 2472 > > Brick eapps-gluster03.my.domain:/export/sdb1/gv0 49152 Y 1866 > > NFS Server on localhost 2049 > Y > > 39603 > > Self-heal Daemon on localhost N/A Y > > 39610 > > NFS Server on eapps-gluster03.my.domain 2049 Y > 35125 > > Self-heal Daemon on eapps-gluster03.my.domain N/A Y 35132 > > NFS Server on eapps-gluster02.my.domain 2049 Y > 37103 > > Self-heal Daemon on eapps-gluster02.my.domain N/A Y 37110 > > > > Task Status of Volume gv0 > > > --------------------------------------------------------------------------------------------------------------- > > > > > > Running 'service glusterd restart' on the NFS server made things start > > working again after this. > > > > > > -- Gene > > > > > > > > > > On Tue, Jun 10, 2014 at 12:10 PM, Niels de Vos <[email protected]> > wrote: > >> > >> On Tue, Jun 10, 2014 at 11:32:50AM -0400, Gene Liverman wrote: > >> > Twice now I have had my nfs connection to a replicated gluster volume > >> > stop > >> > responding. On both servers that connect to the system I have the > >> > following > >> > symptoms: > >> > > >> > 1. Accessing the mount with the native client is still working fine > >> > (the > >> > volume is mounted both that way and via nfs. One app requires the > nfs > >> > version) > >> > 2. The logs have messages stating the following: "kernel: nfs: > server > >> > my-servers-name not responding, still trying" > >> > > >> > How can I fix this? > >> > >> You should check if the NFS-server (a glusterfs process) is still > >> running: > >> > >> # gluster volume status > >> > >> If the NFS-server is not running anymore, you can start it with: > >> > >> # gluster volume start $VOLUME force > >> (you only need to do that for one volume) > >> > >> > >> In case this is with GlusterFS 3.5, you may be hitting a memory leak in > >> the DRC (Duplicate Request Cache) implementation of the NFS-server. You > >> can disable DRC with this: > >> > >> # gluster volume set $VOLUME nfs.drc off > >> > >> In glusterfs-3.5.1 DRC will be disabled by default, there have been too > >> many issues with DRC to enable it for everyone. We need to do more tests > >> and fix DRC in the current development (master) branch. > >> > >> HTH, > >> Niels > > > > > > > > _______________________________________________ > > Gluster-users mailing list > > [email protected] > > http://supercolony.gluster.org/mailman/listinfo/gluster-users >
_______________________________________________ Gluster-users mailing list [email protected] http://supercolony.gluster.org/mailman/listinfo/gluster-users
