There is a lot of data missing...
Using FUSE clients:
For every write, there are three writes on the network (one per brick in
your x3 config). So, outside of bandwidth requirements, you have TCP
connection limits which comes with filedescriptor limits.
If all three bricks are on the same server, then you have a max for that
server divided by three. If each brick is on its own server, then you
don't have that 1/3rd problem.
The idea of GlusterFS is horizontal scaling, so you would add more
bricks on more hosts to scale when you reach that arbitrary limit.
Using NFS:
You have the single connection from the client to the NFS server, then
the fan out to the bricks.
On 03/02/17 21:14, Tamal Saha wrote:
Hi,
Anyone has any comments about this issue? Thanks again.
-Tamal
On Mon, Feb 27, 2017 at 8:34 PM, Tamal Saha <[email protected]
<mailto:[email protected]>> wrote:
Hi,
I am running a GlusterFS cluster in Kubernetes. This has a single
1x3 volume. But this volume is mounted by around 30 other docker
containers. Basically each docker container represents a separate
"user" in our multi-tenant application. As a result there is no
conflicting writes among the "user"s. Each user writes to their
own folder in the volume.
My question is how many clients can mount a GlusterFS volume
before it becomes a performance issue?
Thanks,
-Tamal
_______________________________________________
Gluster-users mailing list
[email protected]
http://lists.gluster.org/mailman/listinfo/gluster-users
--
Mr. Flibble
King of the Potato People
http://www.linkedin.com/in/RobertLanning
_______________________________________________
Gluster-users mailing list
[email protected]
http://lists.gluster.org/mailman/listinfo/gluster-users