Looks like replace-brick in a corrupted replicate brick with 3.3 doesn't work any more. To rebuild the cluster is exactly what we want to avoid when we choose to use glusterfs at the first place.
I assume to replace a failed replicate disk or node should be a standard procedure, isn't it? I could find anything related to this in the 3.3 manual. On Thu, Jan 10, 2013 at 12:18 PM, Daniel Taylor <[email protected]>wrote: > I've run replace-brick on missing bricks before, it should still work. > > On the other hand, data corruption is the worst case failure mode. > The one time I hit data corruption on a node my final answer ended up > being to rebuild the cluster from scratch and restore the best copy of the > data I had (mix of backups and live data). > > > On 01/10/2013 11:12 AM, Liang Ma wrote: > >> >> Thank you Daniel for you more comments. >> >> Now I can remove the damaged zfs brick after rebooting the system. But >> then what can I do to rejoin a new brick? I can't run gluster volume >> replace-brick because the old brick is gone. I can't even remove the old >> brick because the gluster's replicate count is 2. So what is the right >> procedure to replace a failed brick for replicate gluster volume? >> >> Liang >> >> >> On Thu, Jan 10, 2013 at 11:57 AM, Daniel Taylor <[email protected]<mailto: >> [email protected]>> wrote: >> >> I'm not familiar with zfs in particular, but it should have given >> you a message saying why it won't unmount. >> >> In the worst case you can indeed remove the mount point from >> /etc/fstab and reboot. A hard reboot may be necessary in a case >> like this. >> >> >> On 01/10/2013 10:43 AM, Liang Ma wrote: >> >> >> Yes, I stopped the glusterfs service on the damaged system but >> zfs still won't allow me to umount the filesystem. Maybe I >> should try to shutdown the entire system. >> >> >> On Wed, Jan 9, 2013 at 10:28 AM, Daniel Taylor >> <[email protected] <mailto:[email protected]> >> <mailto:[email protected] <mailto:[email protected]>>**> >> >> wrote: >> >> >> On 01/09/2013 08:31 AM, Liang Ma wrote: >> >> >> Hi Daniel, >> >> Ok, if gluster can't self-heal from this situation, I >> hope at >> least I can manually restore the volume by using the good >> brick available. So would you please tell me how can I >> "simply >> rebuild the filesystem and let gluster attempt to >> restore it >> from a *clean* filesystem"? >> >> >> Trimmed for space. >> >> You could do as Tom Pfaff suggests, but given the odds of data >> corruption carrying forward I'd do the following: >> Shut down gluster on the damaged system. >> Unmount the damaged filesystem. >> Reformat the damaged filesystem as new (throwing away any >> potential corruption that might not get caught on rebuild) >> Mount the new filesystem at the original mount point >> Restart gluster >> >> In the event of corruption due to hardware failure you'd >> be doing >> this on replacement hardware. >> The key is you have to have a functional filesystem for >> gluster to >> work with. >> >> >> -- Daniel Taylor VP Operations Vocal >> Laboratories, Inc >> [email protected] <mailto:[email protected]> >> <mailto:[email protected] <mailto:[email protected]>> >> 612-235-5711 <tel:612-235-5711> >> <tel:612-235-5711 <tel:612-235-5711>> >> >> >> ______________________________**_________________ >> Gluster-users mailing list >> [email protected] >> <mailto:Gluster-users@gluster.**org<[email protected]> >> > >> <mailto:Gluster-users@gluster.**org <[email protected]> >> >> <mailto:Gluster-users@gluster.**org <[email protected]>>> >> >> http://supercolony.gluster.**org/mailman/listinfo/gluster-**users<http://supercolony.gluster.org/mailman/listinfo/gluster-users> >> >> >> >> -- Daniel Taylor VP Operations Vocal >> Laboratories, Inc >> [email protected] <mailto:[email protected]> 612-235-5711 >> <tel:612-235-5711> >> >> ______________________________**_________________ >> Gluster-users mailing list >> [email protected] >> <mailto:Gluster-users@gluster.**org<[email protected]> >> > >> >> http://supercolony.gluster.**org/mailman/listinfo/gluster-**users<http://supercolony.gluster.org/mailman/listinfo/gluster-users> >> >> >> > -- > Daniel Taylor VP Operations Vocal Laboratories, Inc > [email protected] 612-235-5711 > > ______________________________**_________________ > Gluster-users mailing list > [email protected] > http://supercolony.gluster.**org/mailman/listinfo/gluster-**users<http://supercolony.gluster.org/mailman/listinfo/gluster-users> >
_______________________________________________ Gluster-users mailing list [email protected] http://supercolony.gluster.org/mailman/listinfo/gluster-users
