On 01/08/2014 05:57 PM, Alun James wrote:
I have this a go.
/gluster volume add-brick myvol replica 2 server2:/brick1 /
/gluster volume heal myvol full/
It seems to be syncing the files but very slowly. Also the server load
on server01 has risen to 200+ and the gluster clients are no longer
able to access the mounts. Is there a way to do this that is not as
impactful? Could I manually rsync the bricks before adding the second
node back in?
The recommended way to heal is using the command mentioned. The gluster
self heal daemon takes appropriate file locks before healing. Since
clients are accessing the volume, I don't think bypassing that and
rsyncing the bricks is a good idea.
Regards,
Ravi
Alun.
------------------------------------------------------------------------
*From: *"Ravishankar N" <[email protected]>
*To: *"Alun James" <[email protected]>
*Cc: *[email protected]
*Sent: *Wednesday, 8 January, 2014 4:00:44 AM
*Subject: *Re: [Gluster-users] delete brick / format / add empty brick
On 01/07/2014 09:40 PM, Alun James wrote:
Hi folks,
I had a 2 node (1 brick each) replica, some network meltdown
issues seemed to cause problems with second node
(server02). glusterfsd process reaching 200-300% and errors
relating to split brain possibilities and self heal errors.
Original volume info:
/Volume Name: myvol/
/Type: Replicate/
/Status: Started/
/Number of Bricks: 2/
/Transport-type: tcp/
/Bricks:/
/Brick1: server01:/brick1/
/Brick2: server02:/brick1/
I removed the second brick (that was showing server problems).
/gluster volume remove-brick myvol replica 1 server02:/brick1/
/
/
Now the volume status is:
/Volume Name: tsfsvol0/
/Type: Distribute/
/Status: Started/
/Number of Bricks: 1/
/Transport-type: tcp/
/Bricks:/
/Brick1: server01:/brick1/
/
/
All is fine and the data on working server is sound.
The xfs partition for /server02:/brick1/ has been formatted and
therefore the data gone. All other gluster config data has
remained untouched. Can I re-add the second server to the volume
with an empty brick and the data will auto replicate over from the
working server?
/gluster volume add-brick myvol replica 2 server2:/brick1 ??/
Yes this should work fine. You will need to run a `gluster volume
heal myvol full` to manually trigger the replication.
ALUN JAMES
Senior Systems Engineer
Tibus
T: +44 (0)28 9033 1122
E: [email protected]
W: www.tibus.com <http://www.tibus.com>
Follow us on Twitter @tibus
<http://twitter.com/intent/user?screen_name=tibus>
Tibus is a trading name of The Internet Business Ltd, a company
limited by share capital and registered in Northern Ireland,
NI31325. It is part of UTV Media Plc.
This email and any attachment may contain confidential information
for the sole use of the intended recipient. Any review, use,
distribution or disclosure by others is strictly prohibited. If
you are not the intended recipient (or authorised to receive for
the recipient), please contact the sender by reply email and
delete all copies of this message.
_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users