I tried to set the limits as you suggest by running the following command. $ sudo gluster volume quota myvolume limit-usage /directory 200GB volume quota : success
but then when I list the quotas there is still nothing, so nothing really happened. I also tried to run stat on all directories which have a quota but nothing happened either. I will send you tomorrow all the other logfiles as requested. -------- Original Message -------- On February 13, 2018 12:20 PM, Hari Gowtham <[email protected]> wrote: >Were you able to set new limits after seeing this error? > > On Tue, Feb 13, 2018 at 4:19 PM, Hari Gowtham [email protected] wrote: >>Yes, I need the log files in that duration, the log rotated file after >> hitting the >> issue aren't necessary, but the ones before hitting the issues are needed >> (not just when you hit it, the ones even before you hit it). >>Yes, you have to do a stat from the client through fuse mount. >>On Tue, Feb 13, 2018 at 3:56 PM, mabi [email protected] wrote: >>>Thank you for your answer. This problem seem to have started since last >>>week, so should I also send you the same log files but for last week? I >>>think logrotate rotates them on a weekly basis. >>>The only two quota commands we use are the following: >>>gluster volume quota myvolume limit-usage /directory 10GB >>> gluster volume quota myvolume list >>>basically to set a new quota or to list the current quotas. The quota list >>>was working in the past yes but we already had a similar issue where the >>>quotas disappeared last August 2017: >>>http://lists.gluster.org/pipermail/gluster-users/2017-August/031946.html >>>In the mean time the only thing we did is to upgrade from 3.8 to 3.10. >>>There are actually no errors to be seen using any gluster commands. The >>>"quota myvolume list" returns simply nothing. >>>In order to lookup the directories should I run a "stat" on them? and if yes >>>should I do that on a client through the fuse mount? >>>-------- Original Message -------- >>> On February 13, 2018 10:58 AM, Hari Gowtham [email protected] wrote: >>>>The log provided are from 11th, you have seen the issue a while before >>>> that itself. >>>>The logs help us to know if something has actually went wrong. >>>> once something goes wrong the output might get affected and i need to know >>>> what >>>> went wrong. Which means i need the log from the beginning. >>>>and i need to know a few more things, >>>> Was the quota list command was working as expected at the beginning? >>>> If yes, what were the commands issued, before you noticed this problem. >>>> Is there any other error that you see other than this? >>>>And can you try looking up the directories the limits are set on and >>>> check if that fixes the error? >>>>>-------- Original Message -------- >>>>> On February 13, 2018 10:44 AM, mabi [email protected] wrote: >>>>>>Hi Hari, >>>>>> Sure no problem, I will send you in a minute another mail where you can >>>>>> download all the relevant log files including the quota.conf binary >>>>>> file. Let me know if you need anything else. In the mean time here below >>>>>> is the output of a volume status. >>>>>> Best regards, >>>>>> M. >>>>>> Status of volume: myvolume >>>>>> Gluster process TCP Port RDMA Port Online >>>>>> Pid >>>>>> Brick gfs1a.domain.local:/data/myvolume >>>>>> /brick 49153 0 Y 3214 >>>>>> Brick gfs1b.domain.local:/data/myvolume >>>>>> /brick 49154 0 Y 3256 >>>>>> Brick gfs1c.domain.local:/srv/glusterf >>>>>> s/myvolume/brick 49153 0 Y 515 >>>>>> Self-heal Daemon on localhost N/A N/A Y >>>>>> 3186 >>>>>> Quota Daemon on localhost N/A N/A Y >>>>>> 3195 >>>>>> Self-heal Daemon on gfs1b.domain.local N/A N/A Y 3217 >>>>>> Quota Daemon on gfs1b.domain.local N/A N/A Y 3229 >>>>>> Self-heal Daemon on gfs1c.domain.local N/A N/A Y 486 >>>>>> Quota Daemon on gfs1c.domain.local N/A N/A Y 495 >>>>>> Task Status of Volume myvolume >>>>>> There are no active volume tasks >>>>>> -------- Original Message -------- >>>>>> On February 13, 2018 10:09 AM, Hari Gowtham [email protected] wrote: >>>>>>>Hi, >>>>>>> A part of the log won't be enough to debug the issue. >>>>>>> Need the whole log messages till date. >>>>>>> You can send it as attachments. >>>>>>> Yes the quota.conf is a binary file. >>>>>>> And I need the volume status output too. >>>>>>> On Tue, Feb 13, 2018 at 1:56 PM, mabi [email protected] wrote: >>>>>>>>Hi Hari, >>>>>>>> Sorry for not providing you more details from the start. Here below >>>>>>>> you will >>>>>>>> find all the relevant log entries and info. Regarding the quota.conf >>>>>>>> file I >>>>>>>> have found one for my volume but it is a binary file. Is it supposed >>>>>>>> to be >>>>>>>> binary or text? >>>>>>>> Regards, >>>>>>>> M. >>>>>>>> *** gluster volume info myvolume *** >>>>>>>> Volume Name: myvolume >>>>>>>> Type: Replicate >>>>>>>> Volume ID: e7a40a1b-45c9-4d3c-bb19-0c59b4eceec5 >>>>>>>> Status: Started >>>>>>>> Snapshot Count: 0 >>>>>>>> Number of Bricks: 1 x (2 + 1) = 3 >>>>>>>> Transport-type: tcp >>>>>>>> Bricks: >>>>>>>> Brick1: gfs1a.domain.local:/data/myvolume/brick >>>>>>>> Brick2: gfs1b.domain.local:/data/myvolume/brick >>>>>>>> Brick3: gfs1c.domain.local:/srv/glusterfs/myvolume/brick (arbiter) >>>>>>>> Options Reconfigured: >>>>>>>> server.event-threads: 4 >>>>>>>> client.event-threads: 4 >>>>>>>> performance.readdir-ahead: on >>>>>>>> nfs.disable: on >>>>>>>> features.quota: on >>>>>>>> features.inode-quota: on >>>>>>>> features.quota-deem-statfs: on >>>>>>>> transport.address-family: inet >>>>>>>> *** /var/log/glusterfs/glusterd.log *** >>>>>>>> [2018-02-13 08:16:09.929568] E [MSGID: 101042] >>>>>>>> [compat.c:569:gf_umount_lazy] >>>>>>>> 0-management: Lazy unmount of /var/run/gluster/myvolume_quota_list/ >>>>>>>> [2018-02-13 08:16:28.596527] I [MSGID: 106499] >>>>>>>> [glusterd-handler.c:4363:__glusterd_handle_status_volume] 0-management: >>>>>>>> Received status volume req for volume myvolume >>>>>>>> [2018-02-13 08:16:28.601097] I [MSGID: 106419] >>>>>>>> [glusterd-utils.c:6110:glusterd_add_inode_size_to_dict] 0-management: >>>>>>>> the >>>>>>>> brick on data/myvolume (zfs) uses dynamic inode sizes >>>>>>>> *** /var/log/glusterfs/cmd_history.log *** >>>>>>>> [2018-02-13 08:16:28.605478] : volume status myvolume detail : SUCCESS >>>>>>>> *** /var/log/glusterfs/quota-mount-myvolume.log *** >>>>>>>> [2018-02-13 08:16:09.934117] I [MSGID: 100030] [glusterfsd.c:2503:main] >>>>>>>> 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version >>>>>>>> 3.10.7 >>>>>>>> (args: /usr/sbin/glusterfs --volfile-server localhost --volfile-id >>>>>>>> myvolume >>>>>>>> -l /var/log/glusterfs/quota-mount-myvolume.log -p >>>>>>>> /var/run/gluster/myvolume_quota_list.pid --client-pid -5 >>>>>>>> /var/run/gluster/myvolume_quota_list/) >>>>>>>> [2018-02-13 08:16:09.940432] I [MSGID: 101190] >>>>>>>> [event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: Started >>>>>>>> thread with >>>>>>>> index 1 >>>>>>>> [2018-02-13 08:16:09.940491] E [socket.c:2327:socket_connect_finish] >>>>>>>> 0-glusterfs: connection to ::1:24007 failed (Connection refused); >>>>>>>> disconnecting socket >>>>>>>> [2018-02-13 08:16:09.940519] I [glusterfsd-mgmt.c:2134:mgmt_rpc_notify] >>>>>>>> 0-glusterfsd-mgmt: disconnected from remote-host: localhost >>>>>>>> [2018-02-13 08:16:13.943827] I [afr.c:94:fix_quorum_options] >>>>>>>> 0-myvolume-replicate-0: reindeer: incoming qtype = none >>>>>>>> [2018-02-13 08:16:13.943857] I [afr.c:116:fix_quorum_options] >>>>>>>> 0-myvolume-replicate-0: reindeer: quorum_count = 2147483647 >>>>>>>> [2018-02-13 08:16:13.945194] I [MSGID: 101190] >>>>>>>> [event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: Started >>>>>>>> thread with >>>>>>>> index 2 >>>>>>>> [2018-02-13 08:16:13.945237] I [MSGID: 101190] >>>>>>>> [event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: Started >>>>>>>> thread with >>>>>>>> index 4 >>>>>>>> [2018-02-13 08:16:13.945247] I [MSGID: 101190] >>>>>>>> [event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: Started >>>>>>>> thread with >>>>>>>> index 3 >>>>>>>> [2018-02-13 08:16:13.945941] W [MSGID: 101174] >>>>>>>> [graph.c:361:_log_if_unknown_option] 0-myvolume-readdir-ahead: option >>>>>>>> 'parallel-readdir' is not recognized >>>>>>>> [2018-02-13 08:16:13.946342] I [MSGID: 114020] [client.c:2352:notify] >>>>>>>> 0-myvolume-client-0: parent translators are ready, attempting connect >>>>>>>> on >>>>>>>> transport >>>>>>>> [2018-02-13 08:16:13.946789] I [MSGID: 114020] [client.c:2352:notify] >>>>>>>> 0-myvolume-client-1: parent translators are ready, attempting connect >>>>>>>> on >>>>>>>> transport >>>>>>>> [2018-02-13 08:16:13.947151] I [rpc-clnt.c:2000:rpc_clnt_reconfig] >>>>>>>> 0-myvolume-client-0: changing port to 49153 (from 0) >>>>>>>> [2018-02-13 08:16:13.947846] I [MSGID: 114057] >>>>>>>> [client-handshake.c:1451:select_server_supported_programs] >>>>>>>> 0-myvolume-client-0: Using Program GlusterFS 3.3, Num (1298437), >>>>>>>> Version >>>>>>>> (330) >>>>>>>> [2018-02-13 08:16:13.948073] I [MSGID: 114020] [client.c:2352:notify] >>>>>>>> 0-myvolume-client-2: parent translators are ready, attempting connect >>>>>>>> on >>>>>>>> transport >>>>>>>> [2018-02-13 08:16:13.948579] I [rpc-clnt.c:2000:rpc_clnt_reconfig] >>>>>>>> 0-myvolume-client-1: changing port to 49154 (from 0) >>>>>>>> [2018-02-13 08:16:13.948699] I [MSGID: 114046] >>>>>>>> [client-handshake.c:1216:client_setvolume_cbk] 0-myvolume-client-0: >>>>>>>> Connected to myvolume-client-0, attached to remote volume >>>>>>>> '/data/myvolume/brick'. >>>>>>>> [2018-02-13 08:16:13.948747] I [MSGID: 114047] >>>>>>>> [client-handshake.c:1227:client_setvolume_cbk] 0-myvolume-client-0: >>>>>>>> Server >>>>>>>> and Client lk-version numbers are not same, reopening the fds >>>>>>>> [2018-02-13 08:16:13.948842] I [MSGID: 108005] >>>>>>>> [afr-common.c:4813:afr_notify] 0-myvolume-replicate-0: Subvolume >>>>>>>> 'myvolume-client-0' came back up; going online. >>>>>>>> Final graph: >>>>>>>> +------------------------------------------------------------------------------+ >>>>>>>> 1: volume myvolume-client-0 >>>>>>>> 2: type protocol/client >>>>>>>> 3: option opversion 31004 >>>>>>>> 4: option clnt-lk-version 1 >>>>>>>> 5: option volfile-checksum 0 >>>>>>>> 6: option volfile-key myvolume >>>>>>>> 7: option client-version 3.10.7 >>>>>>>> 8: option process-uuid >>>>>>>> gfs1a-14348-2018/02/13-08:16:09:933625-myvolume-client-0-0-0 >>>>>>>> 9: option fops-version 1298437 >>>>>>>> 10: option ping-timeout 42 >>>>>>>> 11: option remote-host gfs1a.domain.local >>>>>>>> 12: option remote-subvolume /data/myvolume/brick >>>>>>>> 13: option transport-type socket >>>>>>>> 14: option transport.address-family inet >>>>>>>> 15: option username bea3e634-e174-4bb3-a1d6-25b09d03b536 >>>>>>>> 16: option password 3a6f98bd-795e-4ec4-adfe-42c61ccfa0a6 >>>>>>>> 17: option event-threads 4 >>>>>>>> 18: option transport.tcp-user-timeout 0 >>>>>>>> 19: option transport.socket.keepalive-time 20 >>>>>>>> 20: option transport.socket.keepalive-interval 2 >>>>>>>> 21: option transport.socket.keepalive-count 9 >>>>>>>> [2018-02-13 08:16:13.949147] I [MSGID: 114035] >>>>>>>> [client-handshake.c:202:client_set_lk_version_cbk] 0-myvolume-client-0: >>>>>>>> Server lk version = 1 >>>>>>>> 22: option send-gids true >>>>>>>> 23: end-volume >>>>>>>> 24: >>>>>>>> 25: volume myvolume-client-1 >>>>>>>> 26: type protocol/client >>>>>>>> 27: option ping-timeout 42 >>>>>>>> 28: option remote-host gfs1b.domain.local >>>>>>>> 29: option remote-subvolume /data/myvolume/brick >>>>>>>> 30: option transport-type socket >>>>>>>> 31: option transport.address-family inet >>>>>>>> 32: option username bea3e634-e174-4bb3-a1d6-25b09d03b536 >>>>>>>> 33: option password 3a6f98bd-795e-4ec4-adfe-42c61ccfa0a6 >>>>>>>> 34: option event-threads 4 >>>>>>>> 35: option transport.tcp-user-timeout 0 >>>>>>>> 36: option transport.socket.keepalive-time 20 >>>>>>>> 37: option transport.socket.keepalive-interval 2 >>>>>>>> 38: option transport.socket.keepalive-count 9 >>>>>>>> 39: option send-gids true >>>>>>>> 40: end-volume >>>>>>>> 41: >>>>>>>> 42: volume myvolume-client-2 >>>>>>>> 43: type protocol/client >>>>>>>> 44: option ping-timeout 42 >>>>>>>> 45: option remote-host gfs1c.domain.local >>>>>>>> 46: option remote-subvolume /srv/glusterfs/myvolume/brick >>>>>>>> 47: option transport-type socket >>>>>>>> 48: option transport.address-family inet >>>>>>>> 49: option username bea3e634-e174-4bb3-a1d6-25b09d03b536 >>>>>>>> 50: option password 3a6f98bd-795e-4ec4-adfe-42c61ccfa0a6 >>>>>>>> 51: option event-threads 4 >>>>>>>> 52: option transport.tcp-user-timeout 0 >>>>>>>> 53: option transport.socket.keepalive-time 20 >>>>>>>> 54: option transport.socket.keepalive-interval 2 >>>>>>>> 55: option transport.socket.keepalive-count 9 >>>>>>>> 56: option send-gids true >>>>>>>> 57: end-volume >>>>>>>> 58: >>>>>>>> 59: volume myvolume-replicate-0 >>>>>>>> 60: type cluster/replicate >>>>>>>> 61: option afr-pending-xattr >>>>>>>> myvolume-client-0,myvolume-client-1,myvolume-client-2 >>>>>>>> 62: option arbiter-count 1 >>>>>>>> 63: option use-compound-fops off >>>>>>>> 64: subvolumes myvolume-client-0 myvolume-client-1 >>>>>>>> myvolume-client-2 >>>>>>>> 65: end-volume >>>>>>>> [2018-02-13 08:16:13.949442] I [rpc-clnt.c:2000:rpc_clnt_reconfig] >>>>>>>> 0-myvolume-client-2: changing port to 49153 (from 0) >>>>>>>> 66: >>>>>>>> 67: volume myvolume-dht >>>>>>>> 68: type cluster/distribute >>>>>>>> 69: option lock-migration off >>>>>>>> 70: subvolumes myvolume-replicate-0 >>>>>>>> 71: end-volume >>>>>>>> 72: >>>>>>>> 73: volume myvolume-write-behind >>>>>>>> 74: type performance/write-behind >>>>>>>> 75: subvolumes myvolume-dht >>>>>>>> 76: end-volume >>>>>>>> 77: >>>>>>>> 78: volume myvolume-read-ahead >>>>>>>> 79: type performance/read-ahead >>>>>>>> 80: subvolumes myvolume-write-behind >>>>>>>> 81: end-volume >>>>>>>> 82: >>>>>>>> 83: volume myvolume-readdir-ahead >>>>>>>> 84: type performance/readdir-ahead >>>>>>>> 85: option parallel-readdir off >>>>>>>> 86: option rda-request-size 131072 >>>>>>>> 87: option rda-cache-limit 10MB >>>>>>>> 88: subvolumes myvolume-read-ahead >>>>>>>> 89: end-volume >>>>>>>> 90: >>>>>>>> 91: volume myvolume-io-cache >>>>>>>> 92: type performance/io-cache >>>>>>>> 93: subvolumes myvolume-readdir-ahead >>>>>>>> 94: end-volume >>>>>>>> 95: >>>>>>>> 96: volume myvolume-quick-read >>>>>>>> 97: type performance/quick-read >>>>>>>> 98: subvolumes myvolume-io-cache >>>>>>>> 99: end-volume >>>>>>>> 100: >>>>>>>> 101: volume myvolume-open-behind >>>>>>>> 102: type performance/open-behind >>>>>>>> 103: subvolumes myvolume-quick-read >>>>>>>> 104: end-volume >>>>>>>> 105: >>>>>>>> 106: volume myvolume-md-cache >>>>>>>> 107: type performance/md-cache >>>>>>>> 108: subvolumes myvolume-open-behind >>>>>>>> 109: end-volume >>>>>>>> 110: >>>>>>>> 111: volume myvolume >>>>>>>> 112: type debug/io-stats >>>>>>>> 113: option log-level INFO >>>>>>>> 114: option latency-measurement off >>>>>>>> 115: option count-fop-hits off >>>>>>>> 116: subvolumes myvolume-md-cache >>>>>>>> 117: end-volume >>>>>>>> 118: >>>>>>>> 119: volume meta-autoload >>>>>>>> 120: type meta >>>>>>>> 121: subvolumes myvolume >>>>>>>> 122: end-volume >>>>>>>> 123: >>>>>>>> +------------------------------------------------------------------------------+ >>>>>>>> [2018-02-13 08:16:13.949813] I [MSGID: 114057] >>>>>>>> [client-handshake.c:1451:select_server_supported_programs] >>>>>>>> 0-myvolume-client-1: Using Program GlusterFS 3.3, Num (1298437), >>>>>>>> Version >>>>>>>> (330) >>>>>>>> [2018-02-13 08:16:13.950686] I [MSGID: 114057] >>>>>>>> [client-handshake.c:1451:select_server_supported_programs] >>>>>>>> 0-myvolume-client-2: Using Program GlusterFS 3.3, Num (1298437), >>>>>>>> Version >>>>>>>> (330) >>>>>>>> [2018-02-13 08:16:13.950698] I [MSGID: 114046] >>>>>>>> [client-handshake.c:1216:client_setvolume_cbk] 0-myvolume-client-1: >>>>>>>> Connected to myvolume-client-1, attached to remote volume >>>>>>>> '/data/myvolume/brick'. >>>>>>>> [2018-02-13 08:16:13.950759] I [MSGID: 114047] >>>>>>>> [client-handshake.c:1227:client_setvolume_cbk] 0-myvolume-client-1: >>>>>>>> Server >>>>>>>> and Client lk-version numbers are not same, reopening the fds >>>>>>>> [2018-02-13 08:16:13.951009] I [MSGID: 114035] >>>>>>>> [client-handshake.c:202:client_set_lk_version_cbk] 0-myvolume-client-1: >>>>>>>> Server lk version = 1 >>>>>>>> [2018-02-13 08:16:13.951320] I [MSGID: 114046] >>>>>>>> [client-handshake.c:1216:client_setvolume_cbk] 0-myvolume-client-2: >>>>>>>> Connected to myvolume-client-2, attached to remote volume >>>>>>>> '/srv/glusterfs/myvolume/brick'. >>>>>>>> [2018-02-13 08:16:13.951348] I [MSGID: 114047] >>>>>>>> [client-handshake.c:1227:client_setvolume_cbk] 0-myvolume-client-2: >>>>>>>> Server >>>>>>>> and Client lk-version numbers are not same, reopening the fds >>>>>>>> [2018-02-13 08:16:13.952626] I [MSGID: 114035] >>>>>>>> [client-handshake.c:202:client_set_lk_version_cbk] 0-myvolume-client-2: >>>>>>>> Server lk version = 1 >>>>>>>> [2018-02-13 08:16:13.952704] I [fuse-bridge.c:4138:fuse_init] >>>>>>>> 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.24 >>>>>>>> kernel >>>>>>>> 7.23 >>>>>>>> [2018-02-13 08:16:13.952742] I [fuse-bridge.c:4823:fuse_graph_sync] >>>>>>>> 0-fuse: >>>>>>>> switched to graph 0 >>>>>>>> [2018-02-13 08:16:13.954337] I [MSGID: 108031] >>>>>>>> [afr-common.c:2357:afr_local_discovery_cbk] 0-myvolume-replicate-0: >>>>>>>> selecting local read_child myvolume-client-0 >>>>>>>> [2018-02-13 08:16:14.020681] I [fuse-bridge.c:5081:fuse_thread_proc] >>>>>>>> 0-fuse: >>>>>>>> unmounting /var/run/gluster/myvolume_quota_list/ >>>>>>>> [2018-02-13 08:16:14.020981] W [glusterfsd.c:1360:cleanup_and_exit] >>>>>>>> (-->/lib/x86_64-linux-gnu/libpthread.so.0(+0x8064) [0x7f92aaeef064] >>>>>>>> -->/usr/sbin/glusterfs(glusterfs_sigwaiter+0xe5) [0x55d8ddb85f75] >>>>>>>> -->/usr/sbin/glusterfs(cleanup_and_exit+0x57) [0x55d8ddb85d97] ) 0-: >>>>>>>> received signum (15), shutting down >>>>>>>> [2018-02-13 08:16:14.021017] I [fuse-bridge.c:5804:fini] 0-fuse: >>>>>>>> Unmounting >>>>>>>> '/var/run/gluster/myvolume_quota_list/'. >>>>>>>> *** /var/log/glusterfs/quotad.log *** >>>>>>>> [2018-02-13 08:16:14.023519] W [MSGID: 108027] >>>>>>>> [afr-common.c:2718:afr_discover_done] 0-myvolume-replicate-0: no read >>>>>>>> subvols for (null) >>>>>>>> [2018-02-13 08:16:14.064410] W [dict.c:599:dict_unref] >>>>>>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.7/xlator/features/quotad.so(+0x1f3d) >>>>>>>> [0x7f63153e7f3d] >>>>>>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.7/xlator/features/quotad.so(+0x2d72) >>>>>>>> [0x7f63153e8d72] >>>>>>>> -->/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(dict_unref+0xc0) >>>>>>>> [0x7f631b9bd870] ) 0-dict: dict is NULL [Invalid argument] >>>>>>>> [2018-02-13 08:16:14.074683] W [dict.c:599:dict_unref] >>>>>>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.7/xlator/features/quotad.so(+0x1f3d) >>>>>>>> [0x7f63153e7f3d] >>>>>>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.7/xlator/features/quotad.so(+0x2d72) >>>>>>>> [0x7f63153e8d72] >>>>>>>> -->/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(dict_unref+0xc0) >>>>>>>> [0x7f631b9bd870] ) 0-dict: dict is NULL [Invalid argument] >>>>>>>> [2018-02-13 08:16:14.082040] W [dict.c:599:dict_unref] >>>>>>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.7/xlator/features/quotad.so(+0x1f3d) >>>>>>>> [0x7f63153e7f3d] >>>>>>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.7/xlator/features/quotad.so(+0x2d72) >>>>>>>> [0x7f63153e8d72] >>>>>>>> -->/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(dict_unref+0xc0) >>>>>>>> [0x7f631b9bd870] ) 0-dict: dict is NULL [Invalid argument] >>>>>>>> [2018-02-13 08:16:14.086950] W [dict.c:599:dict_unref] >>>>>>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.7/xlator/features/quotad.so(+0x1f3d) >>>>>>>> [0x7f63153e7f3d] >>>>>>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.7/xlator/features/quotad.so(+0x2d72) >>>>>>>> [0x7f63153e8d72] >>>>>>>> -->/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(dict_unref+0xc0) >>>>>>>> [0x7f631b9bd870] ) 0-dict: dict is NULL [Invalid argument] >>>>>>>> [2018-02-13 08:16:14.087939] W [dict.c:599:dict_unref] >>>>>>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.7/xlator/features/quotad.so(+0x1f3d) >>>>>>>> [0x7f63153e7f3d] >>>>>>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.7/xlator/features/quotad.so(+0x2d72) >>>>>>>> [0x7f63153e8d72] >>>>>>>> -->/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(dict_unref+0xc0) >>>>>>>> [0x7f631b9bd870] ) 0-dict: dict is NULL [Invalid argument] >>>>>>>> [2018-02-13 08:16:14.089775] W [dict.c:599:dict_unref] >>>>>>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.7/xlator/features/quotad.so(+0x1f3d) >>>>>>>> [0x7f63153e7f3d] >>>>>>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.7/xlator/features/quotad.so(+0x2d72) >>>>>>>> [0x7f63153e8d72] >>>>>>>> -->/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(dict_unref+0xc0) >>>>>>>> [0x7f631b9bd870] ) 0-dict: dict is NULL [Invalid argument] >>>>>>>> [2018-02-13 08:16:14.092937] W [dict.c:599:dict_unref] >>>>>>>> (-->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.7/xlator/features/quotad.so(+0x1f3d) >>>>>>>> [0x7f63153e7f3d] >>>>>>>> -->/usr/lib/x86_64-linux-gnu/glusterfs/3.10.7/xlator/features/quotad.so(+0x2d72) >>>>>>>> [0x7f63153e8d72] >>>>>>>> -->/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(dict_unref+0xc0) >>>>>>>> [0x7f631b9bd870] ) 0-dict: dict is NULL [Invalid argument] >>>>>>>> The message "W [MSGID: 108027] [afr-common.c:2718:afr_discover_done] >>>>>>>> 0-myvolume-replicate-0: no read subvols for (null)" repeated 34 times >>>>>>>> between [2018-02-13 08:16:14.023519] and [2018-02-13 08:17:26.560943] >>>>>>>> *** /var/log/glusterfs/cli.log *** >>>>>>>> [2018-02-13 08:16:14.064404] E >>>>>>>> [cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to >>>>>>>> get >>>>>>>> quota limits for 3df709ee-641d-46a2-bd61-889583e3033c >>>>>>>> [2018-02-13 08:16:14.074693] E >>>>>>>> [cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to >>>>>>>> get >>>>>>>> quota limits for a27818fe-0248-40fe-bb23-d43d61010478 >>>>>>>> [2018-02-13 08:16:14.082067] E >>>>>>>> [cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to >>>>>>>> get >>>>>>>> quota limits for daf97388-bcec-4cc0-a8ef-5b93f05b30f6 >>>>>>>> [2018-02-13 08:16:14.086929] E >>>>>>>> [cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to >>>>>>>> get >>>>>>>> quota limits for 3c768b36-2625-4509-87ef-fe5214cb9b01 >>>>>>>> [2018-02-13 08:16:14.087905] E >>>>>>>> [cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to >>>>>>>> get >>>>>>>> quota limits for f8cf47d4-4f54-43c5-ab0d-75b45b4677a3 >>>>>>>> [2018-02-13 08:16:14.089788] E >>>>>>>> [cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to >>>>>>>> get >>>>>>>> quota limits for b4c81a39-2152-45c5-95d3-b796d88226fe >>>>>>>> [2018-02-13 08:16:14.092919] E >>>>>>>> [cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to >>>>>>>> get >>>>>>>> quota limits for 16ac4cde-a5d4-451f-adcc-422a542fea24 >>>>>>>> [2018-02-13 08:16:14.092980] I [input.c:31:cli_batch] 0-: Exiting >>>>>>>> with: 0 >>>>>>>> *** /var/log/glusterfs/bricks/data-myvolume-brick.log *** >>>>>>>> [2018-02-13 08:16:13.948065] I [addr.c:182:gf_auth] >>>>>>>> 0-/data/myvolume/brick: >>>>>>>> allowed = "*", received addr = "127.0.0.1" >>>>>>>> [2018-02-13 08:16:13.948105] I [login.c:76:gf_auth] 0-auth/login: >>>>>>>> allowed >>>>>>>> user names: bea3e634-e174-4bb3-a1d6-25b09d03b536 >>>>>>>> [2018-02-13 08:16:13.948125] I [MSGID: 115029] >>>>>>>> [server-handshake.c:695:server_setvolume] 0-myvolume-server: accepted >>>>>>>> client >>>>>>>> from gfs1a-14348-2018/02/13-08:16:09:933625-myvolume-client-0-0-0 >>>>>>>> (version: >>>>>>>> 3.10.7) >>>>>>>> [2018-02-13 08:16:14.022257] I [MSGID: 115036] >>>>>>>> [server.c:559:server_rpc_notify] 0-myvolume-server: disconnecting >>>>>>>> connection >>>>>>>> from gfs1a-14348-2018/02/13-08:16:09:933625-myvolume-client-0-0-0 >>>>>>>> [2018-02-13 08:16:14.022465] I [MSGID: 101055] >>>>>>>> [client_t.c:436:gf_client_unref] 0-myvolume-server: Shutting down >>>>>>>> connection >>>>>>>> gfs1a-14348-2018/02/13-08:16:09:933625-myvolume-client-0-0-0 >>>>>>>> -------- Original Message -------- >>>>>>>> On February 13, 2018 12:47 AM, Hari Gowtham [email protected] wrote: >>>>>>>> Hi, >>>>>>>> Can you provide more information like, the volume configuration, >>>>>>>> quota.conf >>>>>>>> file and the log files. >>>>>>>> On Sat, Feb 10, 2018 at 1:05 AM, mabi [email protected] wrote: >>>>>>>>>Hello, >>>>>>>>> I am running GlusterFS 3.10.7 and just noticed by doing a "gluster >>>>>>>>> volume >>>>>>>>> quota <volname> list" that my quotas on that volume are broken. The >>>>>>>>> command >>>>>>>>> returns no output and no errors but by looking in >>>>>>>>> /var/log/glusterfs.cli I >>>>>>>>> found the following errors: >>>>>>>>> [2018-02-09 19:31:24.242324] E >>>>>>>>> [cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed >>>>>>>>> to get >>>>>>>>> quota limits for 3df709ee-641d-46a2-bd61-889583e3033c >>>>>>>>> [2018-02-09 19:31:24.249790] E >>>>>>>>> [cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed >>>>>>>>> to get >>>>>>>>> quota limits for a27818fe-0248-40fe-bb23-d43d61010478 >>>>>>>>> [2018-02-09 19:31:24.252378] E >>>>>>>>> [cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed >>>>>>>>> to get >>>>>>>>> quota limits for daf97388-bcec-4cc0-a8ef-5b93f05b30f6 >>>>>>>>> [2018-02-09 19:31:24.256775] E >>>>>>>>> [cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed >>>>>>>>> to get >>>>>>>>> quota limits for 3c768b36-2625-4509-87ef-fe5214cb9b01 >>>>>>>>> [2018-02-09 19:31:24.257434] E >>>>>>>>> [cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed >>>>>>>>> to get >>>>>>>>> quota limits for f8cf47d4-4f54-43c5-ab0d-75b45b4677a3 >>>>>>>>> [2018-02-09 19:31:24.259126] E >>>>>>>>> [cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed >>>>>>>>> to get >>>>>>>>> quota limits for b4c81a39-2152-45c5-95d3-b796d88226fe >>>>>>>>> [2018-02-09 19:31:24.261664] E >>>>>>>>> [cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed >>>>>>>>> to get >>>>>>>>> quota limits for 16ac4cde-a5d4-451f-adcc-422a542fea24 >>>>>>>>> [2018-02-09 19:31:24.261719] I [input.c:31:cli_batch] 0-: Exiting >>>>>>>>> with: 0 >>>>>>>>> How can I fix my quota on that volume again? I had around 30 quotas >>>>>>>>> set on >>>>>>>>> different directories of that volume. >>>>>>>>> Thanks in advance. >>>>>>>>> Regards, >>>>>>>>> M. >>>>>>>>> Gluster-users mailing list >>>>>>>>>[email protected] >>>>>>>>>http://lists.gluster.org/mailman/listinfo/gluster-users >>>>>>>>> Regards, >>>>>>>>> Hari Gowtham. >>>>>>>>> Regards, >>>>>>>>> Hari Gowtham. >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>>Regards, >>>> Hari Gowtham. >>>> >>>-- >> Regards, >> Hari Gowtham. >> > > > > >Regards, > Hari Gowtham. > _______________________________________________ Gluster-users mailing list [email protected] http://lists.gluster.org/mailman/listinfo/gluster-users
