Well... I do the follow: gluster vol create VMS replica 3 arbiter 1 pve01:/DATA/brick1 pve02:/DATA/brick1.5 pve01:/DATA/arbiter1.5 pve02:/DATA/brick2 pv e01:/DATA/brick2.5 pve02:/DATA/arbiter2.5 force
And now I have: gluster vol info Volume Name: VMS Type: Distributed-Replicate Volume ID: 1bd712f5-ccb9-4322-8275-abe363d1ffdd Status: Started Snapshot Count: 0 Number of Bricks: 2 x (2 + 1) = 6 Transport-type: tcp Bricks: Brick1: pve01:/DATA/brick1 Brick2: pve02:/DATA/brick1.5 Brick3: pve01:/DATA/arbiter1.5 (arbiter) Brick4: pve02:/DATA/brick2 Brick5: pve01:/DATA/brick2.5 Brick6: pve02:/DATA/arbiter2.5 (arbiter) Options Reconfigured: cluster.quorum-count: 1 cluster.quorum-reads: false cluster.self-heal-daemon: enable cluster.heal-timeout: 10 storage.fips-mode-rchecksum: on transport.address-family: inet nfs.disable: on performance.client-io-threads: off This values I have put it myself, in order to see if could improve the time to make the volume available, when pve01 goes down with ifupdown cluster.quorum-count: 1 cluster.quorum-reads: false cluster.self-heal-daemon: enable cluster.heal-timeout: 10 Nevertheless, it took more than 1 minutes to the volume VMS available in the other host (pve02). Is there any trick to reduce this time ? Thanks --- Gilberto Nunes Ferreira Em qua., 5 de ago. de 2020 às 08:57, Gilberto Nunes < [email protected]> escreveu: > hum I see... like this: > [image: image.png] > --- > Gilberto Nunes Ferreira > > (47) 3025-5907 > (47) 99676-7530 - Whatsapp / Telegram > > Skype: gilberto.nunes36 > > > > > > Em qua., 5 de ago. de 2020 às 02:14, Computerisms Corporation < > [email protected]> escreveu: > >> check the example of the chained configuration on this page: >> >> >> https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/administration_guide/creating_arbitrated_replicated_volumes >> >> and apply it to two servers... >> >> On 2020-08-04 8:25 p.m., Gilberto Nunes wrote: >> > Hi Bob! >> > >> > Could you, please, send me more detail about this configuration? >> > I will appreciate that! >> > >> > Thank you >> > --- >> > Gilberto Nunes Ferreira >> > >> > (47) 3025-5907 >> > ** >> > (47) 99676-7530 - Whatsapp / Telegram >> > >> > Skype: gilberto.nunes36 >> > >> > >> > >> > >> > >> > Em ter., 4 de ago. de 2020 às 23:47, Computerisms Corporation >> > <[email protected] <mailto:[email protected]>> escreveu: >> > >> > Hi Gilberto, >> > >> > My understanding is there can only be one arbiter per replicated >> > set. I >> > don't have a lot of practice with gluster, so this could be bad >> advice, >> > but the way I dealt with it on my two servers was to use 6 bricks as >> > distributed-replicated (this is also relatively easy to migrate to 3 >> > servers if that happens for you in the future): >> > >> > Server1 Server2 >> > brick1 brick1.5 >> > arbiter1.5 brick2 >> > brick2.5 arbiter2.5 >> > >> > On 2020-08-04 7:00 p.m., Gilberto Nunes wrote: >> > > Hi there. >> > > I have two physical servers deployed as replica 2 and, obviously, >> > I got >> > > a split-brain. >> > > So I am thinking in use two virtual machines,each one in physical >> > > servers.... >> > > Then this two VMS act as a artiber of gluster set.... >> > > >> > > Is this doable? >> > > >> > > Thanks >> > > >> > > ________ >> > > >> > > >> > > >> > > Community Meeting Calendar: >> > > >> > > Schedule - >> > > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC >> > > Bridge: https://bluejeans.com/441850968 >> > > >> > > Gluster-users mailing list >> > > [email protected] <mailto:[email protected]> >> > > https://lists.gluster.org/mailman/listinfo/gluster-users >> > > >> > ________ >> > >> > >> > >> > Community Meeting Calendar: >> > >> > Schedule - >> > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC >> > Bridge: https://bluejeans.com/441850968 >> > >> > Gluster-users mailing list >> > [email protected] <mailto:[email protected]> >> > https://lists.gluster.org/mailman/listinfo/gluster-users >> > >> >
________ Community Meeting Calendar: Schedule - Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC Bridge: https://bluejeans.com/441850968 Gluster-users mailing list [email protected] https://lists.gluster.org/mailman/listinfo/gluster-users
