On 11/28/2013 03:12 AM, Pat Haley wrote:

Hi,

We are currently using gluster with 3 bricks.  We just
rebooted one of the bricks (mseas-data, also identified
as gluster-data) which is actually the main server.  After
rebooting this brick, our client machine (mseas) only sees
the files on the other 2 bricks.  Note that if I mount
the gluster filespace (/gdata) on the brick I rebooted,
it sees the entire space.

The last time I had this problem, there was an error in
one of our /etc/hosts file.  This does not seem to be the
case now.

What else can I look at to debug this problem?

Some information I have from the gluster server

[root@mseas-data ~]# gluster --version
glusterfs 3.3.1 built on Oct 11 2012 22:01:05

[root@mseas-data ~]# gluster volume info

Volume Name: gdata
Type: Distribute
Volume ID: eccc3a90-212d-4563-ae8d-10a77758738d
Status: Started
Number of Bricks: 3
Transport-type: tcp
Bricks:
Brick1: gluster-0-0:/mseas-data-0-0
Brick2: gluster-0-1:/mseas-data-0-1
Brick3: gluster-data:/data

[root@mseas-data ~]# ps -ef | grep gluster

root 2781 1 0 15:16 ? 00:00:00 /usr/sbin/glusterd -p /var/run/glusterd.pid root 2897 1 0 15:16 ? 00:00:00 /usr/sbin/glusterfsd -s localhost --volfile-id gdata.gluster-data.data -p /var/lib/glusterd/vols/gdata/run/gluster-data-data.pid -S /tmp/e3eac7ce95e786a3d909b8fc65ed2059.socket --brick-name /data -l /var/log/glusterfs/bricks/data.log --xlator-option *-posix.glusterd-uuid=22f1102a-08e6-482d-ad23-d8e063cf32ed --brick-port 24009 --xlator-option gdata-server.listen-port=24009 root 2903 1 0 15:16 ? 00:00:00 /usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p /var/lib/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log -S /tmp/d5c892de43c28a1ee7481b780245b789.socket root 4258 1 0 15:52 ? 00:00:00 /usr/sbin/glusterfs --volfile-id=/gdata --volfile-server=mseas-data /gdata
root      4475  4033  0 16:35 pts/0    00:00:00 grep gluster
[

From the ps output, the brick process (glusterfsd) doesn't seem to be running on the gluster-data server. Run `gluster volume status` and check if that is indeed the case. If yes, you could either restart glusterd on the brick node (`service glusterd restart`) or restart the entire volume (`gluster volume start gdata force`) which should bring back the brick process online.

I'm not sure why glusterd did not start the brick process when you rebooted the machine in the first place. You could perhaps check the glusterd log for clues).

Hope this helps,
Ravi

-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-
Pat Haley                          Email:  [email protected]
Center for Ocean Engineering       Phone:  (617) 253-6824
Dept. of Mechanical Engineering    Fax:    (617) 253-8125
MIT, Room 5-213                    http://web.mit.edu/phaley/www/
77 Massachusetts Avenue
Cambridge, MA  02139-4301
_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users

_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Reply via email to