>> When you rebalance vm-disk I honestly don't know what is going to happen, because I doubt they will be writeable during transit. You mean that the VM disk could become read only? And crash the system inside the VM, I guess!
--- Gilberto Nunes Ferreira +55 (47) 99676-7530 - Whatsapp / Telegram Em seg., 9 de jun. de 2025 às 13:23, Andreas Schwibbe <a.schwi...@gmx.net> escreveu: > When adding bricks the new layout is populated, but only for new files. No > automatic rebalance (of old files). > When you rebalance vm-disk I honestly don't know what is going to happen, > because I doubt they will be writeable during transit. > > 09.06.2025 16:57:05 Gilberto Ferreira <gilberto.nune...@gmail.com>: > > Hi Andreas > Thanks for your reply. > Initially I have had 3 directories in both servers, like: > server1|server2 > /data1 > /data2 > /data3 > > Sometime after that, like 6 months after create the glusterfs volumes, I > added more 3 disks in to both servers. > Now that's how it is: > server1|server2 > /data1 > /data2 > /data3 > /data4 > /data5 > /date6 > > I never perform any rebalance or fix layout, because never needed. > Now that 2 directories are nearly full, this situation comes up! > > Is there any impact, if I took too long to perform a rebalance? > Honestly, I never know if a rebalance is needed when adding new bricks. > This should be in the gluster add-brick command output, something like > this: You'll need to rebalance of fix the layout of you volume, after > add-brick. > Anyway, thanks so much for your help. > > Best regards. > > > > --- > > > Gilberto Nunes Ferreira > +55 (47) 99676-7530 - Whatsapp / Telegram > > > > > > > Em seg., 9 de jun. de 2025 às 05:07, Andreas Schwibbe <a.schwi...@gmx.net> > escreveu: > >> Hi Gilberto, >> >> maybe you should set: >> cluster.rebal-throttle: lazy >> >> thus the rebalance is not consuming too many resources. >> This always worked for me, but it does have an impact on responsiveness. >> You however can stop running rebalance if performance impact on >> production environment is too much. >> >> Fix layout will only fix if you added bricks later to the cluster and >> wont fix if you store large files in a single dir. >> >> Best >> A. >> >> Am Sonntag, dem 08.06.2025 um 22:34 -0300 schrieb Gilberto Ferreira: >> >> Hello >> Anybody can help me? >> I need to know if the rebalance has any risk? >> I have 7 SSD KingSton DataCenter DC600M and need to know if I run the >> fix-layout and rebalance will took a long time or if there is any dangerous >> to do that 'online'. >> Thanks >> >> >> --- >> >> >> Gilberto Nunes Ferreira >> +55 (47) 99676-7530 - Whatsapp / Telegram >> >> >> >> >> >> >> Em ter., 3 de jun. de 2025 às 10:25, Gilberto Ferreira < >> gilberto.nune...@gmail.com> escreveu: >> >> Or >> gluster vol rebalance VOL start ?? >> >> --- >> >> >> Gilberto Nunes Ferreira >> +55 (47) 99676-7530 - Whatsapp / Telegram >> >> >> >> >> >> >> Em ter., 3 de jun. de 2025 às 10:24, Gilberto Ferreira < >> gilberto.nune...@gmail.com> escreveu: >> >> Hi there.. >> >> I have two servers with the follow disk layout: >> >> /dev/sdc1 1,8T 1,6T 178G 91% /data1 >> /dev/sdd1 1,8T 1,6T 175G 91% /data2 >> /dev/sdb1 1,8T 225G 1,6T 13% /data3 >> /dev/fuse 128M 44K 128M 1% /etc/pve >> gluster1:stg-vms 8,8T 3,0T 5,8T 34% /stg/vms >> /dev/sdf1 1,8T 108G 1,7T 7% /data4 >> /dev/sdg1 1,8T 108G 1,7T 7% /data5 >> /dev/sdh1 1,8T 105G 1,7T 6% /data6 >> >> As you can see, the /data1 and /data2 are almost full. >> Question is: >> Can I run gluster vol rebalance VOL fix-layout to fix this? >> >> Best regards. >> >> >> --- >> >> >> Gilberto Nunes Ferreira >> +55 (47) 99676-7530 - Whatsapp / Telegram >> >> >> >> >> ________ >> >> >> >> Community Meeting Calendar: >> >> Schedule - >> Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC >> Bridge: https://meet.google.com/cpu-eiue-hvk >> Gluster-users mailing list >> Gluster-users@gluster.org >> https://lists.gluster.org/mailman/listinfo/gluster-users >> >> >>
________ Community Meeting Calendar: Schedule - Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC Bridge: https://meet.google.com/cpu-eiue-hvk Gluster-users mailing list Gluster-users@gluster.org https://lists.gluster.org/mailman/listinfo/gluster-users