Ok...Thanks a lot Strahil This gluster volume set VMS cluster.favorite-child-policy size do the trick to me here!
Cheers --- Gilberto Nunes Ferreira (47) 3025-5907 (47) 99676-7530 - Whatsapp / Telegram Skype: gilberto.nunes36 Em qua., 5 de ago. de 2020 às 18:15, Strahil Nikolov <[email protected]> escreveu: > This could happen if you have pending heals. Did you reboot that node > recently ? > Did you set automatic unsplit-brain ? > > Check for pending heals and files in splitbrain. > > If not, you can check > https://docs.gluster.org/en/latest/Troubleshooting/resolving-splitbrain/ > (look at point 5). > > Best Regards, > Strahil Nikolov > > На 5 август 2020 г. 23:41:57 GMT+03:00, Gilberto Nunes < > [email protected]> написа: > >I'm in trouble here. > >When I shutdown the pve01 server, the shared folder over glusterfs is > >EMPTY! > >It's supposed to be a qcow2 file inside it. > >The content is show right, just after I power on pve01 backup... > > > >Some advice? > > > > > >Thanks > > > >--- > >Gilberto Nunes Ferreira > > > >(47) 3025-5907 > >(47) 99676-7530 - Whatsapp / Telegram > > > >Skype: gilberto.nunes36 > > > > > > > > > > > >Em qua., 5 de ago. de 2020 às 11:07, Gilberto Nunes < > >[email protected]> escreveu: > > > >> Well... > >> I do the follow: > >> > >> gluster vol create VMS replica 3 arbiter 1 pve01:/DATA/brick1 > >> pve02:/DATA/brick1.5 pve01:/DATA/arbiter1.5 pve02:/DATA/brick2 pv > >> e01:/DATA/brick2.5 pve02:/DATA/arbiter2.5 force > >> > >> And now I have: > >> gluster vol info > >> > >> Volume Name: VMS > >> Type: Distributed-Replicate > >> Volume ID: 1bd712f5-ccb9-4322-8275-abe363d1ffdd > >> Status: Started > >> Snapshot Count: 0 > >> Number of Bricks: 2 x (2 + 1) = 6 > >> Transport-type: tcp > >> Bricks: > >> Brick1: pve01:/DATA/brick1 > >> Brick2: pve02:/DATA/brick1.5 > >> Brick3: pve01:/DATA/arbiter1.5 (arbiter) > >> Brick4: pve02:/DATA/brick2 > >> Brick5: pve01:/DATA/brick2.5 > >> Brick6: pve02:/DATA/arbiter2.5 (arbiter) > >> Options Reconfigured: > >> cluster.quorum-count: 1 > >> cluster.quorum-reads: false > >> cluster.self-heal-daemon: enable > >> cluster.heal-timeout: 10 > >> storage.fips-mode-rchecksum: on > >> transport.address-family: inet > >> nfs.disable: on > >> performance.client-io-threads: off > >> > >> This values I have put it myself, in order to see if could improve > >the > >> time to make the volume available, when pve01 goes down with ifupdown > >> cluster.quorum-count: 1 > >> cluster.quorum-reads: false > >> cluster.self-heal-daemon: enable > >> cluster.heal-timeout: 10 > >> > >> Nevertheless, it took more than 1 minutes to the volume VMS available > >in > >> the other host (pve02). > >> Is there any trick to reduce this time ? > >> > >> Thanks > >> > >> --- > >> Gilberto Nunes Ferreira > >> > >> > >> > >> > >> > >> > >> Em qua., 5 de ago. de 2020 às 08:57, Gilberto Nunes < > >> [email protected]> escreveu: > >> > >>> hum I see... like this: > >>> [image: image.png] > >>> --- > >>> Gilberto Nunes Ferreira > >>> > >>> (47) 3025-5907 > >>> (47) 99676-7530 - Whatsapp / Telegram > >>> > >>> Skype: gilberto.nunes36 > >>> > >>> > >>> > >>> > >>> > >>> Em qua., 5 de ago. de 2020 às 02:14, Computerisms Corporation < > >>> [email protected]> escreveu: > >>> > >>>> check the example of the chained configuration on this page: > >>>> > >>>> > >>>> > > > https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/administration_guide/creating_arbitrated_replicated_volumes > >>>> > >>>> and apply it to two servers... > >>>> > >>>> On 2020-08-04 8:25 p.m., Gilberto Nunes wrote: > >>>> > Hi Bob! > >>>> > > >>>> > Could you, please, send me more detail about this configuration? > >>>> > I will appreciate that! > >>>> > > >>>> > Thank you > >>>> > --- > >>>> > Gilberto Nunes Ferreira > >>>> > > >>>> > (47) 3025-5907 > >>>> > ** > >>>> > (47) 99676-7530 - Whatsapp / Telegram > >>>> > > >>>> > Skype: gilberto.nunes36 > >>>> > > >>>> > > >>>> > > >>>> > > >>>> > > >>>> > Em ter., 4 de ago. de 2020 às 23:47, Computerisms Corporation > >>>> > <[email protected] <mailto:[email protected]>> escreveu: > >>>> > > >>>> > Hi Gilberto, > >>>> > > >>>> > My understanding is there can only be one arbiter per > >replicated > >>>> > set. I > >>>> > don't have a lot of practice with gluster, so this could be > >bad > >>>> advice, > >>>> > but the way I dealt with it on my two servers was to use 6 > >bricks > >>>> as > >>>> > distributed-replicated (this is also relatively easy to > >migrate to > >>>> 3 > >>>> > servers if that happens for you in the future): > >>>> > > >>>> > Server1 Server2 > >>>> > brick1 brick1.5 > >>>> > arbiter1.5 brick2 > >>>> > brick2.5 arbiter2.5 > >>>> > > >>>> > On 2020-08-04 7:00 p.m., Gilberto Nunes wrote: > >>>> > > Hi there. > >>>> > > I have two physical servers deployed as replica 2 and, > >>>> obviously, > >>>> > I got > >>>> > > a split-brain. > >>>> > > So I am thinking in use two virtual machines,each one in > >>>> physical > >>>> > > servers.... > >>>> > > Then this two VMS act as a artiber of gluster set.... > >>>> > > > >>>> > > Is this doable? > >>>> > > > >>>> > > Thanks > >>>> > > > >>>> > > ________ > >>>> > > > >>>> > > > >>>> > > > >>>> > > Community Meeting Calendar: > >>>> > > > >>>> > > Schedule - > >>>> > > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC > >>>> > > Bridge: https://bluejeans.com/441850968 > >>>> > > > >>>> > > Gluster-users mailing list > >>>> > > [email protected] > ><mailto:[email protected]> > >>>> > > https://lists.gluster.org/mailman/listinfo/gluster-users > >>>> > > > >>>> > ________ > >>>> > > >>>> > > >>>> > > >>>> > Community Meeting Calendar: > >>>> > > >>>> > Schedule - > >>>> > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC > >>>> > Bridge: https://bluejeans.com/441850968 > >>>> > > >>>> > Gluster-users mailing list > >>>> > [email protected] <mailto:[email protected]> > >>>> > https://lists.gluster.org/mailman/listinfo/gluster-users > >>>> > > >>>> > >>> >
________ Community Meeting Calendar: Schedule - Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC Bridge: https://bluejeans.com/441850968 Gluster-users mailing list [email protected] https://lists.gluster.org/mailman/listinfo/gluster-users
