I created a distributed replicated volume, set the group virt, enabled sharding, migrated a few VMs to the volume, after that i added more bricks to the volume and started the rebalance, i checked the VMs and it was corrupted.
And yes, what you suggested about Gluster is on point, i think we need to have more bug fixes and performance enhancements. Im going to deploy a test Gluster soon just to test patches and updates and report back bugs and issues. -- Respectfully Mahdi A. Mahdi ________________________________ From: Gandalf Corvotempesta <[email protected]> Sent: Sunday, February 26, 2017 11:07:04 AM To: Mahdi Adnan Cc: [email protected] Subject: Re: [Gluster-users] Volume rebalance issue How did you replicate the issue? Next week I'll spin up a gluster storage and I would like to try the same to see the corruption and to test any patches from gluster Il 25 feb 2017 4:31 PM, "Mahdi Adnan" <[email protected]<mailto:[email protected]>> ha scritto: Hi, We have a volume of 4 servers 8x2 bricks (Distributed-Replicate) hosting VMs for ESXi, i tried expanding the volume with 8 more bricks, and after rebalancing the volume, the VMs got corrupted. Gluster version is 3.8.9 and the volume is using the default parameters of group "virt" plus sharding. I created a new volume without sharding and got the same issue after the rebalance. I checked the reported bugs and the mailing list, and i noticed it's a bug in Gluster. Is it affecting all of Gluster versions ? is there any workaround or a volume setup that is not affected by this issue ? Thank you. -- Respectfully Mahdi A. Mahdi _______________________________________________ Gluster-users mailing list [email protected]<mailto:[email protected]> http://lists.gluster.org/mailman/listinfo/gluster-users
_______________________________________________ Gluster-users mailing list [email protected] http://lists.gluster.org/mailman/listinfo/gluster-users
