Hello,

I'm running GlusterFS v3.0.2 with the native FUSE plugin on 2 Rackspace VM's each running 4GB Ram & 160GB HD. (GFS1 and GFS2) Available HD space is approx 57% remaining.

glusterfsd and postfix are the only processes running on these 2 servers, with a total of 6 external clients connected. Each server is a client to each other as well. (8 Total clients)

Upon fresh boot of the server, and processes, the total RAM usage is very minimal, however after a few hours of uptime, the RAM usage is almost completely depleted down to < 100MB on GFS2 and < 20MB on GFS1.

"lsof | grep gfs" reveals 53 connections on GFS1 and 45 on GFS2 from the multiple clients.

This doesn't appear to be client related, since resources are minimal at boot time, with all connections active. However, I'm not completely familiar with the configuration features.

I've just pushed these servers into production, and the websites they serve are receiving approximately 50k hits a day total. Yet, this RAM issue was present before any real traffic existed. Do I have a config error? or am I missing any major performance tuning options?

Any help would be very much appreciated. Thanks,
Chris

TOP:

 PID USER      PR  NI  VIRT  RES  SHR S %CPU %MEM    TIME+  COMMAND
 2421 root      20   0  612m 419m 1092 S    0 10.4   1715:36 glusterfsd

Here is my server config:

r...@lmdc3gfs02:~# cat /etc/glusterfs/glusterfs-server.vol
volume posix
  type storage/posix
  option directory /data/export
end-volume

volume locks
  type features/locks
  subvolumes posix
end-volume

volume brick
  type performance/io-threads
  option thread-count 8
  subvolumes locks
end-volume

volume posix-ns
  type storage/posix
  option directory /data/export-ns
end-volume

volume locks-ns
  type features/locks
  subvolumes posix-ns
end-volume

volume brick-ns
  type performance/io-threads
  option thread-count 8
  subvolumes locks-ns
end-volume

volume server
  type protocol/server
  option transport-type tcp
  option auth.addr.brick.allow *
  option auth.addr.brick-ns.allow *
  subvolumes brick brick-ns
end-volume

Client Config:

r...@lmdc3gfs02:~# cat /etc/glusterfs/glusterfs-client.vol
volume brick1
 type protocol/client
 option transport-type tcp/client
 option remote-host 10.179.122.66   # IP address of the remote brick
 option remote-subvolume brick        # name of the remote volume
 option ping-timeout 2
end-volume

volume brick2
 type protocol/client
 option transport-type tcp/client
 option remote-host 10.179.122.69      # IP address of the remote brick
 option remote-subvolume brick        # name of the remote volume
 option ping-timeout 2
end-volume

volume brick1-ns
 type protocol/client
 option transport-type tcp/client
 option remote-host 10.179.122.66    # IP address of the remote brick
 option remote-subvolume brick-ns        # name of the remote volume
 option ping-timeout 2
end-volume

volume brick2-ns
 type protocol/client
 option transport-type tcp/client
 option remote-host 10.179.122.69      # IP address of the remote brick
 option remote-subvolume brick-ns        # name of the remote volume
 option ping-timeout 2
end-volume

volume afr1
 type cluster/afr
 subvolumes brick1 brick2
end-volume

volume afr-ns
 type cluster/afr
 subvolumes brick1-ns brick2-ns
end-volume


_______________________________________________
Gluster-devel mailing list
[email protected]
http://lists.nongnu.org/mailman/listinfo/gluster-devel

Reply via email to