What do you mean "sharding"? Do you mean sharing folders between two servers to host qcow2 or raw vm images? Here I am using Proxmox which uses qemu but not virsh.
Thanks --- Gilberto Nunes Ferreira (47) 3025-5907 (47) 99676-7530 - Whatsapp / Telegram Skype: gilberto.nunes36 Em qui., 6 de ago. de 2020 às 01:09, Strahil Nikolov <[email protected]> escreveu: > As you mentioned qcow2 files, check the virt group > (/var/lib/glusterfs/group or something like that). It has optimal setttins > for VMs and is used by oVirt. > > WARNING: If you decide to enable the group, which will also enable > sharding, NEVER EVER DISABLE SHARDING -> ONCE ENABLED STAYS ENABLED !!! > Sharding helps reduce loocking during replica heals. > > WARNING2: As virt group uses sharding (fixes the size of file into shard > size), you should consider cluster.favorite-child-policy with value > ctime/mtime. > > Best Regards, > Strahil Nikolov > > На 6 август 2020 г. 1:56:58 GMT+03:00, Gilberto Nunes < > [email protected]> написа: > >Ok...Thanks a lot Strahil > > > >This gluster volume set VMS cluster.favorite-child-policy size do the > >trick > >to me here! > > > >Cheers > >--- > >Gilberto Nunes Ferreira > > > >(47) 3025-5907 > >(47) 99676-7530 - Whatsapp / Telegram > > > >Skype: gilberto.nunes36 > > > > > > > > > > > >Em qua., 5 de ago. de 2020 às 18:15, Strahil Nikolov > ><[email protected]> > >escreveu: > > > >> This could happen if you have pending heals. Did you reboot that node > >> recently ? > >> Did you set automatic unsplit-brain ? > >> > >> Check for pending heals and files in splitbrain. > >> > >> If not, you can check > >> > >https://docs.gluster.org/en/latest/Troubleshooting/resolving-splitbrain/ > >> (look at point 5). > >> > >> Best Regards, > >> Strahil Nikolov > >> > >> На 5 август 2020 г. 23:41:57 GMT+03:00, Gilberto Nunes < > >> [email protected]> написа: > >> >I'm in trouble here. > >> >When I shutdown the pve01 server, the shared folder over glusterfs > >is > >> >EMPTY! > >> >It's supposed to be a qcow2 file inside it. > >> >The content is show right, just after I power on pve01 backup... > >> > > >> >Some advice? > >> > > >> > > >> >Thanks > >> > > >> >--- > >> >Gilberto Nunes Ferreira > >> > > >> >(47) 3025-5907 > >> >(47) 99676-7530 - Whatsapp / Telegram > >> > > >> >Skype: gilberto.nunes36 > >> > > >> > > >> > > >> > > >> > > >> >Em qua., 5 de ago. de 2020 às 11:07, Gilberto Nunes < > >> >[email protected]> escreveu: > >> > > >> >> Well... > >> >> I do the follow: > >> >> > >> >> gluster vol create VMS replica 3 arbiter 1 pve01:/DATA/brick1 > >> >> pve02:/DATA/brick1.5 pve01:/DATA/arbiter1.5 pve02:/DATA/brick2 pv > >> >> e01:/DATA/brick2.5 pve02:/DATA/arbiter2.5 force > >> >> > >> >> And now I have: > >> >> gluster vol info > >> >> > >> >> Volume Name: VMS > >> >> Type: Distributed-Replicate > >> >> Volume ID: 1bd712f5-ccb9-4322-8275-abe363d1ffdd > >> >> Status: Started > >> >> Snapshot Count: 0 > >> >> Number of Bricks: 2 x (2 + 1) = 6 > >> >> Transport-type: tcp > >> >> Bricks: > >> >> Brick1: pve01:/DATA/brick1 > >> >> Brick2: pve02:/DATA/brick1.5 > >> >> Brick3: pve01:/DATA/arbiter1.5 (arbiter) > >> >> Brick4: pve02:/DATA/brick2 > >> >> Brick5: pve01:/DATA/brick2.5 > >> >> Brick6: pve02:/DATA/arbiter2.5 (arbiter) > >> >> Options Reconfigured: > >> >> cluster.quorum-count: 1 > >> >> cluster.quorum-reads: false > >> >> cluster.self-heal-daemon: enable > >> >> cluster.heal-timeout: 10 > >> >> storage.fips-mode-rchecksum: on > >> >> transport.address-family: inet > >> >> nfs.disable: on > >> >> performance.client-io-threads: off > >> >> > >> >> This values I have put it myself, in order to see if could improve > >> >the > >> >> time to make the volume available, when pve01 goes down with > >ifupdown > >> >> cluster.quorum-count: 1 > >> >> cluster.quorum-reads: false > >> >> cluster.self-heal-daemon: enable > >> >> cluster.heal-timeout: 10 > >> >> > >> >> Nevertheless, it took more than 1 minutes to the volume VMS > >available > >> >in > >> >> the other host (pve02). > >> >> Is there any trick to reduce this time ? > >> >> > >> >> Thanks > >> >> > >> >> --- > >> >> Gilberto Nunes Ferreira > >> >> > >> >> > >> >> > >> >> > >> >> > >> >> > >> >> Em qua., 5 de ago. de 2020 às 08:57, Gilberto Nunes < > >> >> [email protected]> escreveu: > >> >> > >> >>> hum I see... like this: > >> >>> [image: image.png] > >> >>> --- > >> >>> Gilberto Nunes Ferreira > >> >>> > >> >>> (47) 3025-5907 > >> >>> (47) 99676-7530 - Whatsapp / Telegram > >> >>> > >> >>> Skype: gilberto.nunes36 > >> >>> > >> >>> > >> >>> > >> >>> > >> >>> > >> >>> Em qua., 5 de ago. de 2020 às 02:14, Computerisms Corporation < > >> >>> [email protected]> escreveu: > >> >>> > >> >>>> check the example of the chained configuration on this page: > >> >>>> > >> >>>> > >> >>>> > >> > > >> > > > https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/administration_guide/creating_arbitrated_replicated_volumes > >> >>>> > >> >>>> and apply it to two servers... > >> >>>> > >> >>>> On 2020-08-04 8:25 p.m., Gilberto Nunes wrote: > >> >>>> > Hi Bob! > >> >>>> > > >> >>>> > Could you, please, send me more detail about this > >configuration? > >> >>>> > I will appreciate that! > >> >>>> > > >> >>>> > Thank you > >> >>>> > --- > >> >>>> > Gilberto Nunes Ferreira > >> >>>> > > >> >>>> > (47) 3025-5907 > >> >>>> > ** > >> >>>> > (47) 99676-7530 - Whatsapp / Telegram > >> >>>> > > >> >>>> > Skype: gilberto.nunes36 > >> >>>> > > >> >>>> > > >> >>>> > > >> >>>> > > >> >>>> > > >> >>>> > Em ter., 4 de ago. de 2020 às 23:47, Computerisms Corporation > >> >>>> > <[email protected] <mailto:[email protected]>> escreveu: > >> >>>> > > >> >>>> > Hi Gilberto, > >> >>>> > > >> >>>> > My understanding is there can only be one arbiter per > >> >replicated > >> >>>> > set. I > >> >>>> > don't have a lot of practice with gluster, so this could > >be > >> >bad > >> >>>> advice, > >> >>>> > but the way I dealt with it on my two servers was to use 6 > >> >bricks > >> >>>> as > >> >>>> > distributed-replicated (this is also relatively easy to > >> >migrate to > >> >>>> 3 > >> >>>> > servers if that happens for you in the future): > >> >>>> > > >> >>>> > Server1 Server2 > >> >>>> > brick1 brick1.5 > >> >>>> > arbiter1.5 brick2 > >> >>>> > brick2.5 arbiter2.5 > >> >>>> > > >> >>>> > On 2020-08-04 7:00 p.m., Gilberto Nunes wrote: > >> >>>> > > Hi there. > >> >>>> > > I have two physical servers deployed as replica 2 and, > >> >>>> obviously, > >> >>>> > I got > >> >>>> > > a split-brain. > >> >>>> > > So I am thinking in use two virtual machines,each one > >in > >> >>>> physical > >> >>>> > > servers.... > >> >>>> > > Then this two VMS act as a artiber of gluster set.... > >> >>>> > > > >> >>>> > > Is this doable? > >> >>>> > > > >> >>>> > > Thanks > >> >>>> > > > >> >>>> > > ________ > >> >>>> > > > >> >>>> > > > >> >>>> > > > >> >>>> > > Community Meeting Calendar: > >> >>>> > > > >> >>>> > > Schedule - > >> >>>> > > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC > >> >>>> > > Bridge: https://bluejeans.com/441850968 > >> >>>> > > > >> >>>> > > Gluster-users mailing list > >> >>>> > > [email protected] > >> ><mailto:[email protected]> > >> >>>> > > > >https://lists.gluster.org/mailman/listinfo/gluster-users > >> >>>> > > > >> >>>> > ________ > >> >>>> > > >> >>>> > > >> >>>> > > >> >>>> > Community Meeting Calendar: > >> >>>> > > >> >>>> > Schedule - > >> >>>> > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC > >> >>>> > Bridge: https://bluejeans.com/441850968 > >> >>>> > > >> >>>> > Gluster-users mailing list > >> >>>> > [email protected] > ><mailto:[email protected]> > >> >>>> > https://lists.gluster.org/mailman/listinfo/gluster-users > >> >>>> > > >> >>>> > >> >>> > >> >
________ Community Meeting Calendar: Schedule - Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC Bridge: https://bluejeans.com/441850968 Gluster-users mailing list [email protected] https://lists.gluster.org/mailman/listinfo/gluster-users
