It was about halo enabled. When i disabled mounting was succesfull. It seems just enabling is not enough.
So far i am happy with currwnt situation just need to figured out whether all data is replicated. With big and even small files, replication is fast. I tested with GB files. Just need to know the replication latency with multiple volumes usage, one volume and GB files, didnt realized a latency. 31 Eki 2018 Çar, saat 06:46 tarihinde Vlad Kopylov <[email protected]> şunu yazdı: > Try adding routes so it can connect to all. > Also curious if fuse mount does need access to all nodes. Supposedly it > does wright to all at the same time unless you have halo feature enabled. > > v > > > > > On Sun, Oct 28, 2018 at 1:07 AM Oğuz Yarımtepe <[email protected]> > wrote: > >> My two nodes are at another vlan. Should my client have connection to all >> nodes, at replicated mod? >> >> Regards. >> >> On Fri, Oct 26, 2018 at 4:44 AM Poornima Gurusiddaiah < >> [email protected]> wrote: >> >>> Is this a new volume? Has it never been mounted successfully? If so try >>> changing firewall settings to allow gluster ports, also check for selinux >>> settings. >>> >>> Regards, >>> Poornima >>> >>> On Fri, Oct 26, 2018, 1:26 AM Oğuz Yarımtepe <[email protected]> >>> wrote: >>> >>>> One more addition: >>>> >>>> # gluster volume info >>>> >>>> >>>> Volume Name: vol0 >>>> Type: Replicate >>>> Volume ID: 28384e2b-ea7e-407e-83ae-4d4e69a2cc7e >>>> Status: Started >>>> Snapshot Count: 0 >>>> Number of Bricks: 1 x 4 = 4 >>>> Transport-type: tcp >>>> Bricks: >>>> Brick1: aslrplpgls01:/bricks/brick1/vol0 >>>> Brick2: aslrplpgls02:/bricks/brick2/vol0 >>>> Brick3: bslrplpgls01:/bricks/brick3/vol0 >>>> Brick4: bslrplpgls02:/bricks/brick4/vol0 >>>> Options Reconfigured: >>>> cluster.self-heal-daemon: enable >>>> cluster.halo-enabled: True >>>> transport.address-family: inet >>>> nfs.disable: on >>>> performance.client-io-threads: off >>>> >>>> On Thu, Oct 25, 2018 at 10:39 PM Oğuz Yarımtepe < >>>> [email protected]> wrote: >>>> >>>>> I have 4 node GlusterFS cluster. Used Centos SIG 4.1 repo. >>>>> >>>>> # gluster peer status >>>>> Number of Peers: 3 >>>>> >>>>> Hostname: aslrplpgls02 >>>>> Uuid: 0876151a-058e-42ec-91f2-f25f353a0207 >>>>> State: Peer in Cluster (Connected) >>>>> >>>>> Hostname: bslrplpgls01 >>>>> Uuid: 6d73ed2a-2287-4872-9a8f-64d6e833181f >>>>> State: Peer in Cluster (Connected) >>>>> >>>>> Hostname: bslrplpgls02 >>>>> Uuid: 8ab6b61f-f502-44c7-8966-2ab03a6b9f7e >>>>> State: Peer in Cluster (Connected) >>>>> >>>>> # gluster volume status vol0 >>>>> Status of volume: vol0 >>>>> Gluster process TCP Port RDMA Port >>>>> Online Pid >>>>> >>>>> ------------------------------------------------------------------------------ >>>>> Brick aslrplpgls01:/bricks/brick1/vol0 49152 0 Y >>>>> 12991 >>>>> Brick aslrplpgls02:/bricks/brick2/vol0 49152 0 Y >>>>> 9344 >>>>> Brick bslrplpgls01:/bricks/brick3/vol0 49152 0 Y >>>>> 61662 >>>>> Brick bslrplpgls02:/bricks/brick4/vol0 49152 0 Y >>>>> 61843 >>>>> Self-heal Daemon on localhost N/A N/A Y >>>>> 13014 >>>>> Self-heal Daemon on bslrplpgls02 N/A N/A Y >>>>> 61866 >>>>> Self-heal Daemon on bslrplpgls01 N/A N/A Y >>>>> 61685 >>>>> Self-heal Daemon on aslrplpgls02 N/A N/A Y >>>>> 9367 >>>>> >>>>> Task Status of Volume vol0 >>>>> >>>>> ------------------------------------------------------------------------------ >>>>> There are no active volume tasks >>>>> >>>>> This is how volume area is mounted: >>>>> >>>>> /dev/gluster_vg/gluster_lv /bricks/brick1 xfs defaults 1 2 >>>>> >>>>> When i try to mount vol0 on a remote machine below is what i got: >>>>> >>>>> [2018-10-25 19:37:23.033302] D [MSGID: 0] >>>>>>> [write-behind.c:2396:wb_lookup_cbk] 0-stack-trace: stack-address: >>>>>>> 0x7f0d04001038, vol0-write-behind returned -1 error: Transport endpoint >>>>>>> is >>>>>>> not connected [Transport endpoint is not connected] >>>>>> >>>>>> [2018-10-25 19:37:23.033329] D [MSGID: 0] >>>>>>> [io-cache.c:268:ioc_lookup_cbk] 0-stack-trace: stack-address: >>>>>>> 0x7f0d04001038, vol0-io-cache returned -1 error: Transport endpoint is >>>>>>> not >>>>>>> connected [Transport endpoint is not connected] >>>>>> >>>>>> [2018-10-25 19:37:23.033356] D [MSGID: 0] >>>>>>> [quick-read.c:473:qr_lookup_cbk] 0-stack-trace: stack-address: >>>>>>> 0x7f0d04001038, vol0-quick-read returned -1 error: Transport endpoint is >>>>>>> not connected [Transport endpoint is not connected] >>>>>> >>>>>> [2018-10-25 19:37:23.033373] D [MSGID: 0] >>>>>>> [md-cache.c:1130:mdc_lookup_cbk] 0-stack-trace: stack-address: >>>>>>> 0x7f0d04001038, vol0-md-cache returned -1 error: Transport endpoint is >>>>>>> not >>>>>>> connected [Transport endpoint is not connected] >>>>>> >>>>>> [2018-10-25 19:37:23.033389] D [MSGID: 0] >>>>>>> [io-stats.c:2278:io_stats_lookup_cbk] 0-stack-trace: stack-address: >>>>>>> 0x7f0d04001038, vol0 returned -1 error: Transport endpoint is not >>>>>>> connected >>>>>>> [Transport endpoint is not connected] >>>>>> >>>>>> [2018-10-25 19:37:23.033408] W >>>>>>> [fuse-resolve.c:132:fuse_resolve_gfid_cbk] 0-fuse: >>>>>>> 00000000-0000-0000-0000-000000000001: failed to resolve (Transport >>>>>>> endpoint >>>>>>> is not connected) >>>>>> >>>>>> [2018-10-25 19:37:23.033426] E >>>>>>> [fuse-bridge.c:928:fuse_getattr_resume] 0-glusterfs-fuse: 2: GETATTR 1 >>>>>>> (00000000-0000-0000-0000-000000000001) resolution failed >>>>>> >>>>>> [2018-10-25 19:37:23.036511] D [MSGID: 0] >>>>>>> [dht-common.c:3468:dht_lookup] 0-vol0-dht: Calling fresh lookup for / on >>>>>>> vol0-replicate-0 >>>>>> >>>>>> [2018-10-25 19:37:23.037347] D [MSGID: 0] >>>>>>> [afr-common.c:3241:afr_discover_do] 0-stack-trace: stack-address: >>>>>>> 0x7f0d04001038, vol0-replicate-0 returned -1 error: Transport endpoint >>>>>>> is >>>>>>> not connected [Transport endpoint is not connected] >>>>>> >>>>>> [2018-10-25 19:37:23.037375] D [MSGID: 0] >>>>>>> [dht-common.c:3020:dht_lookup_cbk] 0-vol0-dht: fresh_lookup returned >>>>>>> for / >>>>>>> with op_ret -1 [Transport endpoint is not connected] >>>>>> >>>>>> [2018-10-25 19:37:23.037940] D [MSGID: 0] >>>>>>> [afr-common.c:3241:afr_discover_do] 0-stack-trace: stack-address: >>>>>>> 0x7f0d04001038, vol0-replicate-0 returned -1 error: Transport endpoint >>>>>>> is >>>>>>> not connected [Transport endpoint is not connected] >>>>>> >>>>>> [2018-10-25 19:37:23.037963] D [MSGID: 0] >>>>>>> [dht-common.c:1378:dht_lookup_dir_cbk] 0-vol0-dht: lookup of / on >>>>>>> vol0-replicate-0 returned error [Transport endpoint is not connected] >>>>>> >>>>>> [2018-10-25 19:37:23.037979] E [MSGID: 101046] >>>>>>> [dht-common.c:1502:dht_lookup_dir_cbk] 0-vol0-dht: dict is null >>>>>> >>>>>> [2018-10-25 19:37:23.037994] D [MSGID: 0] >>>>>>> [dht-common.c:1505:dht_lookup_dir_cbk] 0-stack-trace: stack-address: >>>>>>> 0x7f0d04001038, vol0-dht returned -1 error: Transport endpoint is not >>>>>>> connected [Transport endpoint is not connected] >>>>>> >>>>>> [2018-10-25 19:37:23.038010] D [MSGID: 0] >>>>>>> [write-behind.c:2396:wb_lookup_cbk] 0-stack-trace: stack-address: >>>>>>> 0x7f0d04001038, vol0-write-behind returned -1 error: Transport endpoint >>>>>>> is >>>>>>> not connected [Transport endpoint is not connected] >>>>>> >>>>>> [2018-10-25 19:37:23.038028] D [MSGID: 0] >>>>>>> [io-cache.c:268:ioc_lookup_cbk] 0-stack-trace: stack-address: >>>>>>> 0x7f0d04001038, vol0-io-cache returned -1 error: Transport endpoint is >>>>>>> not >>>>>>> connected [Transport endpoint is not connected] >>>>>> >>>>>> [2018-10-25 19:37:23.038045] D [MSGID: 0] >>>>>>> [quick-read.c:473:qr_lookup_cbk] 0-stack-trace: stack-address: >>>>>>> 0x7f0d04001038, vol0-quick-read returned -1 error: Transport endpoint is >>>>>>> not connected [Transport endpoint is not connected] >>>>>> >>>>>> [2018-10-25 19:37:23.038061] D [MSGID: 0] >>>>>>> [md-cache.c:1130:mdc_lookup_cbk] 0-stack-trace: stack-address: >>>>>>> 0x7f0d04001038, vol0-md-cache returned -1 error: Transport endpoint is >>>>>>> not >>>>>>> connected [Transport endpoint is not connected] >>>>>> >>>>>> [2018-10-25 19:37:23.038078] D [MSGID: 0] >>>>>>> [io-stats.c:2278:io_stats_lookup_cbk] 0-stack-trace: stack-address: >>>>>>> 0x7f0d04001038, vol0 returned -1 error: Transport endpoint is not >>>>>>> connected >>>>>>> [Transport endpoint is not connected] >>>>>> >>>>>> [2018-10-25 19:37:23.038096] W >>>>>>> [fuse-resolve.c:132:fuse_resolve_gfid_cbk] 0-fuse: >>>>>>> 00000000-0000-0000-0000-000000000001: failed to resolve (Transport >>>>>>> endpoint >>>>>>> is not connected) >>>>>> >>>>>> [2018-10-25 19:37:23.038110] E >>>>>>> [fuse-bridge.c:928:fuse_getattr_resume] 0-glusterfs-fuse: 3: GETATTR 1 >>>>>>> (00000000-0000-0000-0000-000000000001) resolution failed >>>>>> >>>>>> [2018-10-25 19:37:23.041169] D [fuse-bridge.c:5087:fuse_thread_proc] >>>>>>> 0-glusterfs-fuse: terminating upon getting ENODEV when reading /dev/fuse >>>>>> >>>>>> [2018-10-25 19:37:23.041196] I [fuse-bridge.c:5199:fuse_thread_proc] >>>>>>> 0-fuse: initating unmount of /mnt/gluster >>>>>> >>>>>> [2018-10-25 19:37:23.041306] D >>>>>>> [logging.c:1795:gf_log_flush_extra_msgs] 0-logging-infra: Log buffer >>>>>>> size >>>>>>> reduced. About to flush 5 extra log messages >>>>>> >>>>>> [2018-10-25 19:37:23.041331] D >>>>>>> [logging.c:1798:gf_log_flush_extra_msgs] 0-logging-infra: Just flushed 5 >>>>>>> extra log messages >>>>>> >>>>>> [2018-10-25 19:37:23.041398] W [glusterfsd.c:1514:cleanup_and_exit] >>>>>>> (-->/lib64/libpthread.so.0(+0x7e25) [0x7f0d24e0ae25] >>>>>>> -->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xe5) [0x5594b73edd65] >>>>>>> -->/usr/sbin/glusterfs(cleanup_and_exit+0x6b) [0x5594b73edb8b] ) 0-: >>>>>>> received signum (15), shutting down >>>>>> >>>>>> [2018-10-25 19:37:23.041417] D >>>>>>> [mgmt-pmap.c:79:rpc_clnt_mgmt_pmap_signout] 0-fsd-mgmt: portmapper >>>>>>> signout >>>>>>> arguments not given >>>>>> >>>>>> [2018-10-25 19:37:23.041428] I [fuse-bridge.c:5981:fini] 0-fuse: >>>>>>> Unmounting '/mnt/gluster'. >>>>>> >>>>>> [2018-10-25 19:37:23.041441] I [fuse-bridge.c:5986:fini] 0-fuse: >>>>>>> Closing fuse connection to '/mnt/gluster'. >>>>>> >>>>>> >>>>> This is how i added mount point to fstab >>>>> >>>>> 10.35.72.138:/vol0 /mnt/gluster glusterfs >>>>> defaults,_netdev,log-level=DEBUG 0 0 >>>>> >>>>> Any idea what the problem is? I found some bug entries, not sure >>>>> whether this situation is a bug. >>>>> >>>>> >>>>> >>>>> -- >>>>> Oğuz Yarımtepe >>>>> http://about.me/oguzy >>>>> >>>> >>>> >>>> -- >>>> Oğuz Yarımtepe >>>> http://about.me/oguzy >>>> _______________________________________________ >>>> Gluster-users mailing list >>>> [email protected] >>>> https://lists.gluster.org/mailman/listinfo/gluster-users >>> >>> >> >> -- >> Oğuz Yarımtepe >> http://about.me/oguzy >> _______________________________________________ >> Gluster-users mailing list >> [email protected] >> https://lists.gluster.org/mailman/listinfo/gluster-users > > -- Oğuz Yarımtepe http://about.me/oguzy
_______________________________________________ Gluster-users mailing list [email protected] https://lists.gluster.org/mailman/listinfo/gluster-users
