Looks like replace-brick in a corrupted replicate brick with 3.3 doesn't
work any more. To rebuild the cluster is exactly what we want to avoid when
we choose to use glusterfs at the first place.

I assume to replace a failed replicate disk or node should be a standard
procedure, isn't it? I could find anything related to this in the 3.3
manual.



On Thu, Jan 10, 2013 at 12:18 PM, Daniel Taylor <[email protected]>wrote:

> I've run replace-brick on missing bricks before, it should still work.
>
> On the other hand, data corruption is the worst case failure mode.
> The one time I hit data corruption on a node my final answer ended up
> being to rebuild the cluster from scratch and restore the best copy of the
> data I had (mix of backups and live data).
>
>
> On 01/10/2013 11:12 AM, Liang Ma wrote:
>
>>
>> Thank you Daniel for you more comments.
>>
>> Now I can remove the damaged zfs brick after rebooting the system. But
>> then what can I do to rejoin a new brick? I can't run gluster volume
>> replace-brick because the old brick is gone. I can't even remove the old
>> brick because the gluster's replicate count is 2. So what is the right
>> procedure to replace a failed brick for replicate gluster volume?
>>
>> Liang
>>
>>
>> On Thu, Jan 10, 2013 at 11:57 AM, Daniel Taylor <[email protected]<mailto:
>> [email protected]>> wrote:
>>
>>     I'm not familiar with zfs in particular, but it should have given
>>     you a message saying why it won't unmount.
>>
>>     In the worst case you can indeed remove the mount point from
>>     /etc/fstab and reboot. A hard reboot may be necessary in a case
>>     like this.
>>
>>
>>     On 01/10/2013 10:43 AM, Liang Ma wrote:
>>
>>
>>         Yes, I stopped the glusterfs service on the damaged system but
>>         zfs still won't allow me to umount the filesystem. Maybe I
>>         should try to shutdown the entire system.
>>
>>
>>         On Wed, Jan 9, 2013 at 10:28 AM, Daniel Taylor
>>         <[email protected] <mailto:[email protected]>
>>         <mailto:[email protected] <mailto:[email protected]>>**>
>>
>>         wrote:
>>
>>
>>             On 01/09/2013 08:31 AM, Liang Ma wrote:
>>
>>
>>                 Hi Daniel,
>>
>>                 Ok, if gluster can't self-heal from this situation, I
>>         hope at
>>                 least I can manually restore the volume by using the good
>>                 brick available. So would you please tell me how can I
>>         "simply
>>                 rebuild the filesystem and let gluster attempt to
>>         restore it
>>                 from a *clean* filesystem"?
>>
>>
>>             Trimmed for space.
>>
>>             You could do as Tom Pfaff suggests, but given the odds of data
>>             corruption carrying forward I'd do the following:
>>             Shut down gluster on the damaged system.
>>             Unmount the damaged filesystem.
>>             Reformat the damaged filesystem as new (throwing away any
>>             potential corruption that might not get caught on rebuild)
>>             Mount the new filesystem at the original mount point
>>             Restart gluster
>>
>>             In the event of corruption due to hardware failure you'd
>>         be doing
>>             this on replacement hardware.
>>             The key is you have to have a functional filesystem for
>>         gluster to
>>             work with.
>>
>>
>>             --     Daniel Taylor             VP Operations Vocal
>>         Laboratories, Inc
>>         [email protected] <mailto:[email protected]>
>>         <mailto:[email protected] <mailto:[email protected]>>
>>         612-235-5711 <tel:612-235-5711>
>>             <tel:612-235-5711 <tel:612-235-5711>>
>>
>>
>>             ______________________________**_________________
>>             Gluster-users mailing list
>>         [email protected] 
>> <mailto:Gluster-users@gluster.**org<[email protected]>
>> >
>>         <mailto:Gluster-users@gluster.**org <[email protected]>
>>
>>         <mailto:Gluster-users@gluster.**org <[email protected]>>>
>>         
>> http://supercolony.gluster.**org/mailman/listinfo/gluster-**users<http://supercolony.gluster.org/mailman/listinfo/gluster-users>
>>
>>
>>
>>     --     Daniel Taylor             VP Operations       Vocal
>> Laboratories, Inc
>>     [email protected] <mailto:[email protected]> 612-235-5711
>>     <tel:612-235-5711>
>>
>>     ______________________________**_________________
>>     Gluster-users mailing list
>>     [email protected] 
>> <mailto:Gluster-users@gluster.**org<[email protected]>
>> >
>>     
>> http://supercolony.gluster.**org/mailman/listinfo/gluster-**users<http://supercolony.gluster.org/mailman/listinfo/gluster-users>
>>
>>
>>
> --
> Daniel Taylor             VP Operations       Vocal Laboratories, Inc
> [email protected]                                     612-235-5711
>
> ______________________________**_________________
> Gluster-users mailing list
> [email protected]
> http://supercolony.gluster.**org/mailman/listinfo/gluster-**users<http://supercolony.gluster.org/mailman/listinfo/gluster-users>
>
_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Reply via email to