Con Zyor,
Could you attach the client and server log files? That would helps us
understand what is causing the hang.
thanks,
kp
On 11/05/2011 03:17 AM, Con Zyor wrote:
Hello;
I have installed glusterfs 3.2.4 to a pair of Red Hat Enterprise Linux
6.1 x86_64 machines with 2GB memory. I am attempting to mirror a
directory full of content between the two servers, which also serve
and update the content through a webapp via Apache. My issue is that
the client mount points hang for 30 minutes or so if either node is
brought down.
The volfile will be at the end of this e-mail.
I setup two bricks, one each on nodes server01 and server02, using
ext4 and acl mount options. The vfstab entries on each server look
like this
/dev/mapper/sysvg-brick01 /brick01 ext4
defaults,nosuid,acl 1 2
From one host, I configure them as a mirror and start the volume:
gluster volume create volume01 replica 2 transport tcp
server01:/brick01 server02:/brick01
gluster volume start volume01
Then server1 and server2 each mount the volume from themselves via
/etc/fstab entry:
localhost:/volume01 /glusterfs/vol01 glusterfs
defaults,_netdev,acl 0 0
This works, modifications inside /glusterfs/vol01 are seen by the
other host. However when I reboot either server01 or server02, the
client mount point on the surviving node (/glusterfs/vol01) hangs
until the node reboots. If the node never boots, the client mount
point on the surviving node hangs for 30 minutes. I have tried
reducing frame-timeout to 10 seconds to no avail.
Also; once the rebooted server comes back online it fails to mount
/glusterfs/vol01, hanging, again for 30 minutes. A subsequent remount
succeeds. Cancelling the hung mount with umount -f /glusterfs/vol01
and then re-mounting succeeds.
Any ideas what I am doing wrong?
Here is the volfile from /var/log/glusterfs/glusterfs-vol01.log
1: volume volume01-client-0
2: type protocol/client
3: option remote-host server01
4: option remote-subvolume /brick01
5: option transport-type tcp
6: option frame-timeout 10
7: end-volume
8:
9: volume volume01-client-1
10: type protocol/client
11: option remote-host server02
12: option remote-subvolume /brick01
13: option transport-type tcp
14: option frame-timeout 10
15: end-volume
16:
17: volume volume01-replicate-0
18: type cluster/replicate
19: subvolumes volume01-client-0 volume01-client-1
20: end-volume
21:
22: volume volume01-write-behind
23: type performance/write-behind
24: subvolumes volume01-replicate-0
25: end-volume
26:
27: volume volume01-read-ahead
28: type performance/read-ahead
29: subvolumes volume01-write-behind
30: end-volume
31:
32: volume volume01-io-cache
33: type performance/io-cache
34: subvolumes volume01-read-ahead
35: end-volume
36:
37: volume volume01-quick-read
38: type performance/quick-read
39: subvolumes volume01-io-cache
40: end-volume
41:
42: volume volume01-stat-prefetch
43: type performance/stat-prefetch
44: subvolumes volume01-quick-read
45: end-volume
46:
47: volume volume01
48: type debug/io-stats
49: option latency-measurement off
50: option count-fop-hits off
51: subvolumes volume01-stat-prefetch
52: end-volume
_______________________________________________
Gluster-users mailing list
[email protected]
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
[email protected]
http://gluster.org/cgi-bin/mailman/listinfo/gluster-users