Since the "trusted.glusterfs.volume-id" is not set on the bricks , the volume start will fail.

1) Execute : "getfattr -e hex -n trusted.glusterfs.volume-id /opt/gluster_data/eccp_glance" from any of the node which the brick process running.

You will get the hex value of the "trusted.glusterfs.volume-id" extended attribute . Set this value on the bricks created (On the nodes where you deleted the bricks)

2) setfattr -n trusted.glusterfs.volume-id -v <hex_value_of_the_trusted_glusterfs_volume-id> /opt/gluster_data/eccp_glance

3) gluster volume start <volume_name> force  : To restart the brick process

4) gluster volume status <volume_name> : Check all the brick process are started.

5) gluster volume heal <volume_name> full : To trigger self-heal on to the removed bricks.

Shwetha
On 11/28/2013 02:59 PM, weiyuanke wrote:
hi Shwetha,


command "gluster volume start eccp_glance force" on the other node gives following:
Pasted Graphic.tiff

with cli.log
Pasted Graphic 1.tiff

on the damaged node,gluster volume start eccp_glance force gives:
Pasted Graphic 2.tiff




---------------------------
韦远科
010 5881 3749
中国科学院 计算机网络信息中心





On 2013年11月28日, at 下午4:50, shwetha <[email protected] <mailto:[email protected]>> wrote:

1) create the brick directory "/opt/gluster_data/eccp_glance" on the nodes where you deleted the directories.

2) From any of the storage node execute :

 1. gluster volume start <volume_name> force  : To restart the brick
    process
 2. gluster volume status <volume_name> : Check all the brick process
    are started.
 3. gluster volume heal <volume_name> full : To trigger self-heal on
    to the removed bricks.

-Shwetha

On 11/28/2013 02:09 PM, 韦远科 wrote:
hi all,

I accidentally removed the brick directory of a volume on one node, the replica for this volume is 2.

now the situation is , there is no corresponding glusterfsd process on this node, and 'glusterfs volume status' shows that the brick is offline, like this:
Brick 192.168.64.11:/opt/gluster_data/eccp_glance  N/A    Y    2513
Brick 192.168.64.12:/opt/gluster_data/eccp_glance  49161    Y    2542
Brick 192.168.64.17:/opt/gluster_data/eccp_glance  49164    Y    2537
Brick 192.168.64.18:/opt/gluster_data/eccp_glance  49154    Y    4978
Brick 192.168.64.29:/opt/gluster_data/eccp_glance  N/A    N    N/A
Brick 192.168.64.30:/opt/gluster_data/eccp_glance  49154    Y    4072
Brick 192.168.64.25:/opt/gluster_data/eccp_glance  49155    Y    11975
Brick 192.168.64.26:/opt/gluster_data/eccp_glance  49155    Y    17947
Brick 192.168.64.13:/opt/gluster_data/eccp_glance  49154    Y    26045
Brick 192.168.64.14:/opt/gluster_data/eccp_glance  49154    Y    22143


so are there ways to bring this brick back to normal?

thanks!


-----------------------------------------------------------------
韦远科
中国科学院 计算机网络信息中心



_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users





_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Reply via email to