You couldn't force start the volume because the brick "nas-02-data:/datastore" doesn't have the "trusted.glusterfs.volume-id" .

From nas-02 node execute :

1. setfattr -n trusted.glusterfs.volume-id -v 0xfdff519085ef4cba9056a6bbbd8d6863 nas-02-data:/datastore

2. gluster volume start datastore1 force.


-Shwetha
On 11/27/2013 12:23 PM, gandalf istari wrote:
Hi Shwetha,

[root@nas-01 ~]# getfattr -d -e hex -m . /datastore

getfattr: Removing leading '/' from absolute path names

# file: datastore

trusted.afr.datastore1-client-0=0x000000000000000000000000

trusted.afr.datastore1-client-1=0x000000000000000000000000

trusted.gfid=0x00000000000000000000000000000001

trusted.glusterfs.dht=0x000000010000000000000000ffffffff

trusted.glusterfs.volume-id=0xfdff519085ef4cba9056a6bbbd8d6863



[root@nas-02 ~]# getfattr -d -e hex -m . /datastore

getfattr: Removing leading '/' from absolute path names

# file: datastore

security.selinux=0x73797374656d5f753a6f626a6563745f723a6574635f72756e74696d655f743a733000

I don't understand what happened .


gr

Patrick


Hi Gandalf,

can you run the following command on the brick path?

"getfattr -d -e hex -m . /datastore" on both "nas-01-data" and "nas-02-data" nodes.

This will let us know whether there is "trusted.glusterfs.volume-id" set.

-Shwetha

On 11/26/2013 07:36 PM, gandalf istari wrote:
hi thanks for the quick answer.

I'm running glusterfs 3.4.1

[root@nas-02 datastore]# gluster volume start datastore1 force

volume start: datastore1: failed: Failed to get extended attribute trusted.glusterfs.volume-id for brick dir /datastore. Reason : No data available

It seems that the .gluster directory is missing for some reason.


volume replace-brick datastore1 nas-01-data:/datastorenas-02-data:/datastore
commit force
To rebuild/replace the missing brick ?
I'm quite new with glusterfs
Thanks






On 26/11/13 12:47, gandalf istari wrote:
Hi have setup a two node replication glusterfs. After the initial installation the "master" node was put into the datacenter and after two week we moved the second one also to the datacenter.

But the sync has not started yet.

On the "master"

gluster> volume info all

Volume Name: datastore1

Type: Replicate

Volume ID: fdff5190-85ef-4cba-9056-a6bbbd8d6863

Status: Started

Number of Bricks: 1 x 2 = 2

Transport-type: tcp

Bricks:

Brick1: nas-01-data:/datastore

Brick2: nas-02-data:/datastore

gluster> peer status

Number of Peers: 1


Hostname: nas-02-data

Uuid: 71df9f86-a87b-481d-896c-c0d4ab679cfa

State: Peer in Cluster (Connected)


On the "slave"

gluster> peer status

Number of Peers: 1

Hostname: 192.168.70.6

Uuid: 97ef0154-ad7b-402a-b0cb-22be09134a3c

State: Peer in Cluster (Connected)


gluster> volume status all

Status of volume: datastore1

Gluster processPortOnlinePid

------------------------------------------------------------------------------

Brick nas-01-data:/datastore49152Y2130

Brick nas-02-data:/datastoreN/ANN/A

NFS Server on localhost2049Y8064

Self-heal Daemon on localhostN/AY8073

NFS Server on 192.168.70.62049Y3379

Self-heal Daemon on 192.168.70.6N/AY3384

Which version of glusterfs are you running?

volume status suggests that the second brick (nas-02-data:/datastore) is not running.

Can you run "gluster volume start <volname> force" in any of these two nodes and try again? Then you would also required to run `find . | xargs stat` on the mountpoint of the volume. That should trigger the self heal.

There are no active volume tasks


I would like to run on the "slave" gluster volume sync nas-01-data datastore1

BTW, There is no concept of "master" and "slave" in afr (replication). However there is concept of "master volume" and "slave volume" in gluster geo-replication.

But then the virtual machines hosted will be unavailible is there another way to start the replication ?


Thanks






_______________________________________________
Gluster-users mailing list
[email protected]  <mailto:[email protected]>
http://supercolony.gluster.org/mailman/listinfo/gluster-users



_______________________________________________
Gluster-users mailing list
[email protected]  <mailto:[email protected]>
http://supercolony.gluster.org/mailman/listinfo/gluster-users





_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Reply via email to