Yes but I want to add. Is it the same logic? --- Gilberto Nunes Ferreira +55 (47) 99676-7530 Proxmox VE VinChin Backup & Restore
Em ter., 5 de nov. de 2024, 14:09, Aravinda <aravi...@kadalu.tech> escreveu: > Hello Gilberto, > > You can create a Arbiter volume using three bricks. Two of them will be > data bricks and one will be Arbiter brick. > > gluster volume create VMS replica 3 arbiter 1 gluster1:/disco2TB-0/vms > gluster2:/disco2TB-0/vms arbiter:/arbiter1 > > To make this Volume as distributed Arbiter, add more bricks (multiple of > 3, two data bricks and one arbiter brick) similar to above. > > -- > Aravinda > > > ---- On Tue, 05 Nov 2024 22:24:38 +0530 *Gilberto Ferreira > <gilberto.nune...@gmail.com <gilberto.nune...@gmail.com>>* wrote --- > > Clearly I am doing something wrong > > pve01:~# gluster vol info > > Volume Name: VMS > Type: Distributed-Replicate > Volume ID: e1a4f787-3f62-441e-a7ce-c0ae6b111ebf > Status: Started > Snapshot Count: 0 > Number of Bricks: 3 x 2 = 6 > Transport-type: tcp > Bricks: > Brick1: gluster1:/disco2TB-0/vms > Brick2: gluster2:/disco2TB-0/vms > Brick3: gluster1:/disco1TB-0/vms > Brick4: gluster2:/disco1TB-0/vms > Brick5: gluster1:/disco1TB-1/vms > Brick6: gluster2:/disco1TB-1/vms > Options Reconfigured: > cluster.self-heal-daemon: off > cluster.entry-self-heal: off > cluster.metadata-self-heal: off > cluster.data-self-heal: off > cluster.granular-entry-heal: on > storage.fips-mode-rchecksum: on > transport.address-family: inet > performance.client-io-threads: off > pve01:~# gluster vol add-brick VMS replica 3 arbiter 1 > gluster1:/disco2TB-0/vms gluster2:/disco2TB-0/vms gluster1:/disco1TB-0/vms > gluster2:/disco1TB-0/vms gluster1:/disco1TB-1/vms gluster2: > /disco1TB-1/vms arbiter:/arbiter1 > volume add-brick: failed: Operation failed > pve01:~# gluster vol add-brick VMS replica 3 arbiter 1 > gluster1:/disco2TB-0/vms gluster2:/disco2TB-0/vms gluster1:/disco1TB-0/vms > gluster2:/disco1TB-0/vms gluster1:/disco1TB-1/vms gluster2: > /disco1TB-1/vms arbiter:/arbiter1 arbiter:/arbiter2 arbiter:/arbiter3 > arbiter:/arbiter4 > volume add-brick: failed: Operation failed > > --- > > > Gilberto Nunes Ferreira > > > > > > > > Em ter., 5 de nov. de 2024 às 13:39, Andreas Schwibbe <a.schwi...@gmx.net> > escreveu: > > > If you create a volume with replica 2 arbiter 1 > > you create 2 data bricks that are mirrored (makes 2 file copies) > + > you create 1 arbiter that holds metadata of all files on these bricks. > > You "can" create all on the same server, but this makes no sense, because > when the server goes down, no files on these disks are accessible anymore, > hence why bestpractice is to spread out over 3 servers, so when one server > (or disk) goes down, you will still have 1 file copy and 1 arbiter with > metadata online. > Which is also very handy when the down server comes up again, because then > you prevent splitbrain as you have matching file copy + metadata showing > which version of each file is newest, thus self-heal can jump in to get you > back to 2 file copies. > > when you want to add further bricks, you must add pairs i.e. > you will add again 2 bricks 1 arbiter and these bricks and arbiter belong > together and share the same files and metadata. > > Hth. > A. > > > Am Dienstag, dem 05.11.2024 um 13:28 -0300 schrieb Gilberto Ferreira: > > Ok. > I got confused here! > For each brick I will need one arbiter brick, in a different > partition/folder? > And what if in some point in the future I decide to add a new brick in the > main servers? > Do I need to provide another partition/folder in the arbiter and then > adjust the arbiter brick counter? > > --- > > > Gilberto Nunes Ferreira > > > > > > > > Em ter., 5 de nov. de 2024 às 13:22, Andreas Schwibbe <a.schwi...@gmx.net> > escreveu: > > Your add-brick command adds 2 bricks 1 arbiter (even though you name them > all arbiter!) > > The sequence is important: > > gluster v add-brick VMS replica 2 arbiter 1 gluster1:/gv0 gluster2:/gv0 > arbiter1:/arb1 > > adds two data bricks and a corresponding arbiter from 3 different servers > and 3 different disks, > thus you can loose any one server OR any one disk and stay up and > consistent. > > adding more bricks to the volume you can follow the pattern. > > A. > > Am Dienstag, dem 05.11.2024 um 12:51 -0300 schrieb Gilberto Ferreira: > > Hi there. > > In previous emails, I comment with you guys, about 2 node gluster server, > where the bricks lay down in different size and folders in the same server, > like > > gluster vol create VMS replica 2 gluster1:/disco2TB-0/vms > gluster2:/disco2TB-0/vms gluster1:/disco1TB-0/vms gluster2:/disco1TB-0/vms > gluster1:/disco1TB-1/vms gluster2:/disco1TB-1/vms > > So I went ahead and installed a Debian 12 and installed the same gluster > version that the other servers, which is now 11.1 or something like that. > In this new server, I have a small disk like 480G in size. > And I created 3 partitions formatted with XFS using imaxpct=75, as > suggested in previous emails. > > And than in the gluster nodes, I tried to add the brick > gluster vol add-brick VMS replica 3 arbiter 1 arbiter:/arbiter1/arbiter1 > arbiter:/arbiter2/arbiter2 arbiter:/arbiter3/arbiter3 > > But to my surprise (or not!) I got this message: > volume add-brick: failed: Multiple bricks of a replicate volume are > present on the same server. This setup is not optimal. Bricks should be on > different nodes to have best fault tolerant co > nfiguration. Use 'force' at the end of the command if you want to override > this behavior. > > Why is that? > > > > > > > --- > > > Gilberto Nunes Ferreira > > > > > > ________ > > > > Community Meeting Calendar: > > Schedule - > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC > Bridge: https://meet.google.com/cpu-eiue-hvk > Gluster-users mailing list > Gluster-users@gluster.org > https://lists.gluster.org/mailman/listinfo/gluster-users > > > > > > ________ > > > > Community Meeting Calendar: > > Schedule - > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC > Bridge: https://meet.google.com/cpu-eiue-hvk > Gluster-users mailing list > Gluster-users@gluster.org > https://lists.gluster.org/mailman/listinfo/gluster-users > > ________ > > > > Community Meeting Calendar: > > Schedule - > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC > Bridge: https://meet.google.com/cpu-eiue-hvk > Gluster-users mailing list > Gluster-users@gluster.org > https://lists.gluster.org/mailman/listinfo/gluster-users > > > >
________ Community Meeting Calendar: Schedule - Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC Bridge: https://meet.google.com/cpu-eiue-hvk Gluster-users mailing list Gluster-users@gluster.org https://lists.gluster.org/mailman/listinfo/gluster-users