Hi all!
I'm rather new to glusterfs, trying it out for redundant storage for my very small company. I have a minimal setup of glusterfs, 2 servers (storage1 & storage2) with one brick each, both added to volume "testvol1". I then mount "testvol1" on a third server (app1). This is fine as long as I use any one servers ip and filesystem glusterfs, but when trying to mount the volume using a file "/owndata/conf/glusterfs/storage.vol", it does not work. Also I can't use nfs - don't need it but maybe it's related.

So:
"mount -t glusterfs 192.168.160.21:/testvol1 /mnt/tmp/" works. I write a file to /mnt/tmp/filename, then mount 192.168.12.210:/testvol1 and the newly created file is there.

Trying to mount using config file storage.vol:

   mount -t glusterfs /owndata/conf/glusterfs/storage.vol /mnt/tmp
   Mount failed. Please check the log file for more details.

The main error in the log is

   /E [client-handshake.c:1778:client_query_portmap_cbk] 0-remote1:
   failed to get the port number for remote subvolume. Please run
   'gluster volume status' on server to see if brick process is running/.

There are lots & lots of pages on the 'net about this error message, none of the solutions I've found has worked.

CentOS 6.5 on all servers, they are all kvm hosts under oVirt (this is just the testing stage, will be on real iron in production).
No firewall anywhere, selinux is permissive.

*File storage.vol:*

   volume remote1
      type protocol/client
      option transport-type tcp
      option remote-host 192.168.12.210
      option remote-subvolume testvol1
   end-volume

   volume remote2
      type protocol/client
      option transport-type tcp
      option remote-host 192.168.160.21
      option remote-subvolume testvol1
   end-volume

   volume replicate
      type cluster/replicate
      subvolumes remote1 remote2
   end-volume

   volume writebehind
      type performance/write-behind
      option window-size 1MB
      subvolumes replicate
   end-volume

   volume cache
      type performance/io-cache
      option cache-size 256MB
      subvolumes writebehind
   end-volume

*# gluster volume info*

   Volume Name: testvol1
   Type: Replicate
   Volume ID: bcca4aa2-46c0-44a2-8175-1305faa8b4f9
   Status: Started
   Number of Bricks: 1 x 2 = 2
   Transport-type: tcp
   Bricks:
   Brick1: 192.168.12.210:/export/vdb/brick1
   Brick2: 192.168.160.21:/export/vdb/brick1

*# gluster volume status*

   Status of volume: testvol1
   Gluster process                                         Port Online  Pid
   -------------------------------------------------------------------------
   Brick 192.168.12.210:/export/vdb/brick1                 49152
   Y       1656
   Brick 192.168.160.21:/export/vdb/brick1                 49152
   Y       139090
NFS Server on localhost 2049 Y 1670 Self-heal Daemon on localhost N/A Y 1674
   NFS Server on 192.168.160.21                            2049    Y
          1481
Self-heal Daemon on 192.168.160.21 N/A Y 139105

   Task Status of Volume testvol1
   
------------------------------------------------------------------------------
   There are no active volume tasks

*Complete log after fail:
*

   [2014-10-02 14:38:22.252235] I [glusterfsd.c:2026:main]
   0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version
   3.4.0.57rhs (/usr/sbin/glusterfs
   --fuse-mountopts=allow_other,default_permissions,max_read=131072
   --volfile=/owndata/conf/glusterfs/storage.vol
   --fuse-mountopts=allow_other,default_permissions,max_read=131072
   /mnt/glust)
   [2014-10-02 14:38:22.284438] W [options.c:848:xl_opt_validate]
   0-writebehind: option 'window-size' is deprecated, preferred is
   'cache-size', continuing with correction
   [2014-10-02 14:38:22.284476] W [io-cache.c:1672:init] 0-cache:
   dangling volume. check volfile
   [2014-10-02 14:38:22.294306] I [socket.c:3505:socket_init]
   0-remote2: SSL support is NOT enabled
   [2014-10-02 14:38:22.294339] I [socket.c:3520:socket_init]
   0-remote2: using system polling thread
   [2014-10-02 14:38:22.294832] I [socket.c:3505:socket_init]
   0-remote1: SSL support is NOT enabled
   [2014-10-02 14:38:22.294848] I [socket.c:3520:socket_init]
   0-remote1: using system polling thread
   [2014-10-02 14:38:22.294870] I [client.c:2171:notify] 0-remote1:
   parent translators are ready, attempting connect on transport
   [2014-10-02 14:38:22.306697] I [client.c:2171:notify] 0-remote2:
   parent translators are ready, attempting connect on transport
   Final graph:
   
+------------------------------------------------------------------------------+
      1: volume remote1
      2:     type protocol/client
      3:     option remote-subvolume testvol1
      4:     option remote-host 192.168.12.210
      5:     option transport-type socket
      6: end-volume
      7:
      8: volume remote2
      9:     type protocol/client
     10:     option remote-subvolume testvol1
     11:     option remote-host 192.168.160.21
     12:     option transport-type socket
     13: end-volume
     14:
     15: volume replicate
     16:     type cluster/replicate
     17:     subvolumes remote1 remote2
     18: end-volume
     19:
     20: volume writebehind
     21:     type performance/write-behind
     22:     option cache-size 1MB
     23:     subvolumes replicate
     24: end-volume
     25:
     26: volume cache
     27:     type performance/io-cache
     28:     option cache-size 256MB
     29:     subvolumes writebehind
     30: end-volume
     31:
   
+------------------------------------------------------------------------------+
   [2014-10-02 14:38:22.310830] E
   [client-handshake.c:1778:client_query_portmap_cbk] 0-remote1: failed
   to get the port number for remote subvolume. Please run 'gluster
   volume status' on server to see if brick process is running.
   [2014-10-02 14:38:22.310887] I [client.c:2103:client_rpc_notify]
   0-remote1: disconnected from 192.168.12.210:24007. Client process
   will keep trying to connect to glusterd until brick's port is
   available.
   [2014-10-02 14:38:22.311031] E
   [client-handshake.c:1778:client_query_portmap_cbk] 0-remote2: failed
   to get the port number for remote subvolume. Please run 'gluster
   volume status' on server to see if brick process is running.
   [2014-10-02 14:38:22.311059] I [client.c:2103:client_rpc_notify]
   0-remote2: disconnected from 192.168.160.21:24007. Client process
   will keep trying to connect to glusterd until brick's port is
   available.
   [2014-10-02 14:38:22.311070] E [afr-common.c:4025:afr_notify]
   0-replicate: All subvolumes are down. Going offline until atleast
   one of them comes back up.
   [2014-10-02 14:38:22.314827] I [fuse-bridge.c:5874:fuse_graph_setup]
   0-fuse: switched to graph 0
   [2014-10-02 14:38:22.316140] I [fuse-bridge.c:4811:fuse_init]
   0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.13
   kernel 7.13
   [2014-10-02 14:38:22.321404] W [fuse-bridge.c:1134:fuse_attr_cbk]
   0-glusterfs-fuse: 2: LOOKUP() / => -1 (Transport endpoint is not
   connected)
   [2014-10-02 14:38:22.324731] I [fuse-bridge.c:5715:fuse_thread_proc]
   0-fuse: unmounting /mnt/glust
   [2014-10-02 14:38:22.324931] W [glusterfsd.c:1099:cleanup_and_exit]
   (-->/lib64/libc.so.6(clone+0x6d) [0x7f6e2ec5e86d]
   (-->/lib64/libpthread.so.0(+0x79d1) [0x7f6e2f2f19d1]
   (-->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xcd) [0x4052ad]))) 0-:
   received signum (15), shutting down
   [2014-10-02 14:38:22.324946] I [fuse-bridge.c:6412:fini] 0-fuse:
   Unmounting '/mnt/glust'.


Regards
Peter H

_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users

Reply via email to