Hi Carlo,
Very good question: IMO it's a matter of compromise between having lots of small
files to heal and how long it takes to heal each one shard.
With small shards, it increases probability to have to self-heal more files in a
failure, also it means latency in metadata of each file could be significant
versus time to transfer rsync deltas.
With big shards, time to rsync could be very long.
You need to find a size something enough big for avoid latency affects
significantly to self-heal process and so small to avoid rsync delta take too
much time....
We found 256MB shard size for distributed-replicated (replica 2 o 3) volumes is
good enough and also we recommend to use 512MB shard size for volumes
distributed-dispersed 3 redundancy 1. For volumes distributed-dispersed 6
redundancy 2 we use 1024MB as shard size.
Hope this helps!
El 07/09/19 a les 15:04, Cristian Del Carlo ha escrit:
Thank you Ramon for your answer.
The only thing I can't understand is why to use such a big shard?
The default is 64MB. I thought maybe to decrease it, I wanted to do some tests
on it.
Best regards,
Il giorno ven 6 set 2019 alle ore 23:09 Ramon Selga <[email protected]
<mailto:[email protected]>> ha scritto:
Hi Cristian,
Both approaches are correct but they have different usable capacity and
tolerance to node failures.
First one is a full replica 3 meaning you get your total node capacity
divided by 3, because replica 3, and it tolerates a simultaneous of two
nodes and very good for split-brain avoidance.
Second one is a replica 2 with arbiter, also very good for split-brain
avoidance (that's purpose of arbiter bricks). In this case you get your
total capacity divided by two except a little space going to
arbiter-bricks, may be less than 1% of normal storage bricks. It tolerates
one node failure at the same time.
For VM usage remember to enable sharding with a shard size of 256MB at
least before use volume.
If efficiency between total and usable capacity is a concern for you and
you think you could tolerate only one node failure at the same time, may I
suggest you to use a distributed dispersed 3 redundancy 1 volume?. You
will get your total capacity divided by 3 times 2 (that's 2/3 of total
capacity) and this config still tolerates one node failure at the same time.
Hope this helps.
*Ramon Selga*
934 76 69 10
670 25 37 05
DataLab SL <http://www.datalab.es/>
Aviso Legal <http://www.datalab.es/cont_cas/legal.html>
El 06/09/19 a les 17:11, Cristian Del Carlo ha escrit:
Hi,
I have an environment consisting of 4 nodes ( with large disks).
I have to create a volume to contain image of virtual machines.
In documentation i read:
/Hosting virtual machine images requires the consistency of three-way
replication,
which is provided by three-way replicated volumes, three-way distributed
replicated volumes,
arbitrated replicated volumes, and distributed arbitrated replicated
volumes. /
So I'm going to confusion to configure this volume.
I have 4 nodes and I don't want to lose space by dedicating one to the
function of arbiter.
Would it be reasonable to configure the volume as in these two examples?
# gluster volume create test1 replica 3 \
server1:/bricks/brick1 server2:/bricks/brick1 server3:/bricks/brick1 \
server2:/bricks/brick2 server3:/bricks/brick2 server4:/bricks/brick2 \
server3:/bricks/brick3 server4:/bricks/brick3 server1:/bricks/brick3 \
server4:/bricks/brick4 server1:/bricks/brick4 server2:/bricks/brick4
# gluster volume create test1 replica 3 arbiter 1 \
server1:/bricks/brick1 server2:/bricks/brick1
server3:/bricks/arbiter_brick1 \
server2:/bricks/brick2 server3:/bricks/brick2
server4:/bricks/arbiter_brick2 \
server3:/bricks/brick3 server4:/bricks/brick3
server1:/bricks/arbiter_brick3 \
server4:/bricks/brick4 server1:/bricks/brick4 server2:/bricks/arbiter_brick4
Thanks,
--
*/
/*
*/Cristian Del Carlo/*
_______________________________________________
Gluster-users mailing list
[email protected] <mailto:[email protected]>
https://lists.gluster.org/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
[email protected] <mailto:[email protected]>
https://lists.gluster.org/mailman/listinfo/gluster-users
--
*/
/*
*/Cristian Del Carlo/*
/Target Solutions s.r.l.
/
/
/
*T***+39 0583 1905621
*F* +39 0583 1905675
*@* [email protected]
<mailto:[email protected]>
http://www.targetsolutions.it <http://www.targetsolutions.it/>
P.IVA e C.Fiscale: 01815270465 Reg. Imp. di Lucca
Capitale Sociale: €11.000,00 iv - REA n° 173227
Il testo e gli eventuali documenti trasmessi contengono informazioni riservate
al destinatario indicato. La seguente e-mail e' confidenziale e la sua
riservatezza e' tutelata legalmente dal Decreto Legislativo 196 del 30/06/2003
(Codice di tutela della privacy). La lettura, copia o altro uso non
autorizzato o qualsiasi altra azione derivante dalla conoscenza di queste
informazioni sono rigorosamente vietate. Qualora abbiate ricevuto questo
documento per errore siete cortesemente pregati di darne immediata
comunicazione al mittente e di provvedere, immediatamente, alla sua distruzione.
_______________________________________________
Gluster-users mailing list
[email protected]
https://lists.gluster.org/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
[email protected]
https://lists.gluster.org/mailman/listinfo/gluster-users