1) run peer detach for all the servers

done.

2) from server 1 ->3 and 3->1 make sure ports are open and iptables
are turned off.

done.
by the way, the same test on 3.1.5-1 works fine with same environment.

3) remove config files under /etc/glusterd

please review following logs.

4) run your tests again.

I don't know why but glusterd service failed to start on all 3 hosts.

[root@vhead-010 ~]# date
Tue Aug 16 09:12:53 JST 2011
[root@vhead-010 ~]# cat a.sh
#!/bin/bash
for i in foo-{1..3}-private
do
        ssh ${i} service glusterd stop
        ssh ${i} 'find /etc/glusterd -type f|xargs rm -f'
        ssh ${i} service iptables restart
        ssh ${i} iptables -vL
        ssh ${i} service glusterd start
        ssh ${i} find /etc/glusterd
        ssh ${i} service glusterd status
done
[root@vhead-010 ~]# bash a.sh
Stopping glusterd:[  OK  ]
Flushing firewall rules: [  OK  ]
Setting chains to policy ACCEPT: filter [  OK  ]
Unloading iptables modules: [  OK  ]
Chain INPUT (policy ACCEPT 0 packets, 0 bytes)
 pkts bytes target     prot opt in     out     source               destination

Chain FORWARD (policy ACCEPT 0 packets, 0 bytes)
 pkts bytes target     prot opt in     out     source               destination

Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes)
 pkts bytes target     prot opt in     out     source               destination
Starting glusterd:[  OK  ]
/etc/glusterd
/etc/glusterd/glusterd.info
/etc/glusterd/nfs
/etc/glusterd/nfs/run
/etc/glusterd/peers
/etc/glusterd/vols
/etc/glusterd/vols/foo
/etc/glusterd/vols/foo/bricks
/etc/glusterd/vols/foo/run
glusterd is stopped
Stopping glusterd:[  OK  ]
Flushing firewall rules: [  OK  ]
Setting chains to policy ACCEPT: filter [  OK  ]
Unloading iptables modules: [  OK  ]
Chain INPUT (policy ACCEPT 0 packets, 0 bytes)
 pkts bytes target     prot opt in     out     source               destination

Chain FORWARD (policy ACCEPT 0 packets, 0 bytes)
 pkts bytes target     prot opt in     out     source               destination

Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes)
 pkts bytes target     prot opt in     out     source               destination
Starting glusterd:[  OK  ]
/etc/glusterd
/etc/glusterd/glusterd.info
/etc/glusterd/nfs
/etc/glusterd/nfs/run
/etc/glusterd/peers
/etc/glusterd/vols
/etc/glusterd/vols/foo
/etc/glusterd/vols/foo/bricks
/etc/glusterd/vols/foo/run
glusterd is stopped
Stopping glusterd:[  OK  ]
Flushing firewall rules: [  OK  ]
Setting chains to policy ACCEPT: filter [  OK  ]
Unloading iptables modules: [  OK  ]
Chain INPUT (policy ACCEPT 0 packets, 0 bytes)
 pkts bytes target     prot opt in     out     source               destination

Chain FORWARD (policy ACCEPT 0 packets, 0 bytes)
 pkts bytes target     prot opt in     out     source               destination

Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes)
 pkts bytes target     prot opt in     out     source               destination
Starting glusterd:[  OK  ]
/etc/glusterd
/etc/glusterd/glusterd.info
/etc/glusterd/nfs
/etc/glusterd/nfs/run
/etc/glusterd/peers
/etc/glusterd/vols
/etc/glusterd/vols/foo
/etc/glusterd/vols/foo/bricks
/etc/glusterd/vols/foo/run
glusterd is stopped
[root@vhead-010 ~]# date
Tue Aug 16 09:13:20 JST 2011
[root@vhead-010 ~]# ssh foo-1-private
Last login: Tue Aug 16 09:06:57 2011 from dlp.local.valinux.co.jp
[root@localhost ~]# tail -20 /var/log/glusterfs/etc-glusterfs-glusterd.vol.log
...
[2011-08-16 09:13:01.85858] I [glusterd.c:304:init] 0-management: Using 
/etc/glusterd as working directory
[2011-08-16 09:13:01.87294] E [rpc-transport.c:799:rpc_transport_load] 
0-rpc-transport: 
/opt/glusterfs/3.1.6/lib64/glusterfs/3.1.6/rpc-transport/rdma.so: cannot open 
shared object file: No such file or directory
[2011-08-16 09:13:01.87340] E [rpc-transport.c:803:rpc_transport_load] 
0-rpc-transport: volume 'rdma.management': transport-type 'rdma' is not valid 
or not found on this machine
[2011-08-16 09:13:01.87402] E 
[glusterd-store.c:654:glusterd_store_handle_retrieve] 0-glusterd: Unable to 
retrieve store handle for /etc/glusterd/glusterd.info, error: No such file or 
directory
[2011-08-16 09:13:01.87422] E [glusterd-store.c:761:glusterd_retrieve_uuid] 0-: 
Unable to get store handle!
[2011-08-16 09:13:01.87514] I [glusterd.c:95:glusterd_uuid_init] 0-glusterd: 
generated UUID: c0cef9f9-a79e-4189-8955-d83927db9cee
[2011-08-16 09:13:01.87681] E 
[glusterd-store.c:654:glusterd_store_handle_retrieve] 0-glusterd: Unable to 
retrieve store handle for /etc/glusterd/vols/foo/info, error: No such file or 
directory
[2011-08-16 09:13:01.87704] E 
[glusterd-store.c:1328:glusterd_store_retrieve_volumes] 0-: Unable to restore 
volume: foo
[2011-08-16 09:13:01.87732] E [xlator.c:843:xlator_init] 0-management: 
Initialization of volume 'management' failed, review your volfile again
[2011-08-16 09:13:01.87751] E [graph.c:331:glusterfs_graph_init] 0-management: 
initializing translator failed
[2011-08-16 09:13:01.87818] I [glusterfsd.c:712:cleanup_and_exit] 0-glusterfsd: 
shutting down
[root@localhost ~]# exit

Best,

(2011/08/16 8:52), Mohit Anchlia wrote:
Logs are generally in /var/log/gluster

Since you are playing with it. I would suggest this:

1) run peer detach for all the servers
2) from server 1 ->3 and 3->1 make sure ports are open and iptables
are turned off.
3) remove config files under /etc/glusterd
4) run your tests again.

On Mon, Aug 15, 2011 at 4:28 PM, Tomoaki Sato<[email protected]>  wrote:
Thanks, Mohit

(2011/08/16 8:05), Mohit Anchlia wrote:

What's in your logs?

I can obtain logs needed. could you tell me the instruction to take the
logs?


Did you have  foo-3-private before in your gluster cluster ever or
adding this host for the first time?

It was first time.
All foo-X-private has no entries in /etc/glusterd/peers/ and
/etc/glusterd/vols/.


Try gluster peer detach and then remove any left over configuration in
/etc/glusterd config directory. After that try again and see if that
works.

[root@vhead-010 ~]# date
Tue Aug 16 08:17:49 JST 2011
[root@vhead-010 ~]# cat a.sh
#!/bin/bash
for i in foo-{1..3}-private
do
        ssh ${i} service glusterd stop
        ssh ${i} rm -rf /etc/glusterd/peers/*
        ssh ${i} rm -rf /etc/glusterd/vols/*
        ssh ${i} service glusterd start
        ssh ${i} find /etc/glusterd
done
[root@vhead-010 ~]# bash a.sh
Stopping glusterd:[  OK  ]
Starting glusterd:[  OK  ]
/etc/glusterd
/etc/glusterd/glusterd.info
/etc/glusterd/nfs
/etc/glusterd/nfs/nfs-server.vol
/etc/glusterd/nfs/run
/etc/glusterd/peers
/etc/glusterd/vols
Stopping glusterd:[  OK  ]
Starting glusterd:[  OK  ]
/etc/glusterd
/etc/glusterd/glusterd.info
/etc/glusterd/nfs
/etc/glusterd/nfs/nfs-server.vol
/etc/glusterd/nfs/run
/etc/glusterd/peers
/etc/glusterd/vols
Stopping glusterd:[  OK  ]
Starting glusterd:[  OK  ]
/etc/glusterd
/etc/glusterd/glusterd.info
/etc/glusterd/nfs
/etc/glusterd/nfs/nfs-server.vol
/etc/glusterd/nfs/run
/etc/glusterd/peers
/etc/glusterd/vols
[root@vhead-010 ~]# ssh foo-1-private
[root@localhost ~]# gluster peer probe foo-2-private
Probe successful
[root@localhost ~]# gluster peer status
Number of Peers: 1

Hostname: foo-2-private
Uuid: c2b314ac-6ed1-455a-84d4-ec22041ee2b2
State: Peer in Cluster (Connected)
[root@localhost ~]# gluster volume create foo foo-1-private:/mnt/brick
Creation of volume foo has been successful. Please start the volume to
access da
ta.
[root@localhost ~]# gluster volume start foo
Starting volume foo has been successful
[root@localhost ~]# gluster volume add-brick foo foo-2-private:/mnt/brick
Add Brick successful
[root@localhost ~]# gluster peer probe foo-3-private
Probe successful
[root@localhost ~]# gluster peer status
Number of Peers: 2

Hostname: foo-2-private
Uuid: c2b314ac-6ed1-455a-84d4-ec22041ee2b2
State: Peer in Cluster (Connected)

Hostname: foo-3-private
Uuid: 7fb98dac-fef7-4b33-837c-6483a767ec3e
State: Peer Rejected (Connected)
[root@localhost ~]# cat /var/log/glusterfs/.cmd_log_history
...
[2011-08-16 08:20:28.862619] peer probe :  on host foo-2-private:24007
[2011-08-16 08:20:28.912419] peer probe : on host foo-2-private:24007 FAILED
[2011-08-16 08:20:58.382350] Volume create : on volname: foo attempted
[2011-08-16 08:20:58.382461] Volume create : on volname: foo type:DEFAULT
count:
1 bricks: foo-1-private:/mnt/brick
[2011-08-16 08:20:58.384674] Volume create : on volname: foo SUCCESS
[2011-08-16 08:21:04.831772] volume start : on volname: foo SUCCESS
[2011-08-16 08:21:22.682292] Volume add-brick : on volname: foo attempted
[2011-08-16 08:21:22.682385] Volume add-brick : volname: foo type DEFAULT
count:
1 bricks: foo-2-private:/mnt/brick
[2011-08-16 08:21:22.682499] Volume add-brick : on volname: foo SUCCESS
[2011-08-16 08:21:39.124574] peer probe :  on host foo-3-private:24007
[2011-08-16 08:21:39.135609] peer probe : on host foo-3-private:24007 FAILED

Tomo




On Mon, Aug 15, 2011 at 3:37 PM, Tomoaki Sato<[email protected]>    wrote:

Hi,

following instructions work fine with 3.1.5-1 but with 3.1.6-1.

1. make a new file system without peers. [OK]

foo-1-private# gluster volume create foo foo-1-private:/mnt/brick
foo-1-private# gluster volume start foo
foo-1-private# gluster peer status
No peers present
foo-1-private#

2. add a peer to the file system. [NG]

foo-1-private# gluster peer probe foo-2-private
Probe successful
foo-1-private# gluster peer status
Number of Peers: 1

Hostname: foo-2-private
Uuid: c2b314ac-6ed1-455a-84d4-ec22041ee2b2
State: Peer Rejected (Connected)
foo-1-private# gluster volume add-brick foo foo-2-private:/mnt/brick
Host foo-2-private not connected
foo-1-private#


following instructions work fine even with 3.1.6-1.

1. make a new file system with single peer. [OK]

foo-1-private# gluster peer status
No peer presents
foo-1-private# gluster peer probe foo-2-private
Probe successful
foo-1-private# gluster peer status
Number of Peers: 1

Hostname: foo-2-private
Uuid: c2b314ac-6ed1-455a-84d4-ec22041ee2b2
State: Peer in Cluster (Connected)
foo-1-private# gluster volume create foo foo-1-private:/mnt/brick
Creation of volume foo has been successful. Please start the volume to
access data.
foo-1-private# gluster volume start foo
Starting volume foo has been successful
foo-1-private# gluster volume add-brick foo foo-2-private:/mnt/brick
Add Brick successful
foo-1-private#

But ...

2. add a peer to the file system. [NG]

foo-1-private# gluster peer probe foo-3-private
Probe successful
foo-1-private# gluster peer status
Number of Peers: 2

Hostname: foo-2-private
Uuid: c2b314ac-6ed1-455a-84d4-ec22041ee2b2
State: Peer in Cluster (Connected)

Hostname: foo-3-private
Uuid: 7fb98dac-fef704b33-837c-6483a767ec3e
State: Peer Rejected (Connected)
foo-1-private# gluster volume add-brick foo foo-3-private:/mnt/brick
Host foo-3-private not connected
foo-1-private#

How should I add extra peers to existing file systems ?

Best,
_______________________________________________
Gluster-users mailing list
[email protected]
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users




_______________________________________________
Gluster-users mailing list
[email protected]
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users

Reply via email to