Hi, I was under the impression that under Gluster 3.3 the self-heal process would only lock parts of files as required during the self-heal process ??
I have a 3.3 setup running here and earlier rebooted one of the storage nodes. Replication meant that the volume holding around 20 VM's (400G's) kept on running quite happily. However, when Gluster restarted and kicked off it's self heal, it queued and LOCKED all 20 VM's, unlocking images as the healing process finished on each one, over a period of a number of hours (!) I'm using 3.3.0 release from the semiosis PPA on Ubuntu 12.04. Is there a trick to making this work properly, or is there a fix due out that will correct this behaviour ?? tia Gareth. volume enc-client-0 type protocol/client option remote-host 10.1.0.1 option remote-subvolume /srv/enc option transport-type tcp option username *** option password *** end-volume volume enc-client-1 type protocol/client option remote-host 10.2.0.4 option remote-subvolume /srv/enc option transport-type tcp option username *** option password *** end-volume volume enc-client-2 type protocol/client option remote-host 10.2.0.3 option remote-subvolume /srv/enc option transport-type tcp option username *** option password *** end-volume volume encr-client-3 type protocol/client option remote-host 10.1.0.2 option remote-subvolume /srv/enc option transport-type tcp option username *** option password *** end-volume volume enc-replicate-0 type cluster/replicate option background-self-heal-count 0 option metadata-self-heal on option data-self-heal on option entry-self-heal on option self-heal-daemon on option iam-self-heal-daemon yes subvolumes enc-client-0 enc-client-1 end-volume volume enc-replicate-1 type cluster/replicate option background-self-heal-count 0 option metadata-self-heal on option data-self-heal on option entry-self-heal on option self-heal-daemon on option iam-self-heal-daemon yes subvolumes enc-client-2 enc-client-3 end-volume # volume glustershd type debug/io-stats subvolumes enc-replicate-0 enc-replicate-1 end-volume
_______________________________________________ Gluster-devel mailing list [email protected] https://lists.nongnu.org/mailman/listinfo/gluster-devel
