Re: [Gluster-users] Failed to get quota limits

2018-02-27 Thread Hari Gowtham
Hi,

Yes. This bugs is fixed in 3.12.

On Wed, Feb 28, 2018 at 1:18 AM, mabi  wrote:
> Hi,
>
> Thanks for the link to the bug. We should be hopefully moving soon onto 3.12 
> so I guess this bug is also fixed there.
>
> Best regards,
> M.
>
>
> ‐‐‐ Original Message ‐‐‐
>
> On February 27, 2018 9:38 AM, Hari Gowtham  wrote:
>
>>
>>
>> Hi Mabi,
>>
>> The bugs is fixed from 3.11. For 3.10 it is yet to be backported and
>>
>> made available.
>>
>> The bug is https://bugzilla.redhat.com/show_bug.cgi?id=1418259.
>>
>> On Sat, Feb 24, 2018 at 4:05 PM, mabi m...@protonmail.ch wrote:
>>
>> > Dear Hari,
>> >
>> > Thank you for getting back to me after having analysed the problem.
>> >
>> > As you said I tried to run "gluster volume quota  list " 
>> > for all of my directories which have a quota and found out that there was 
>> > one directory quota which was missing (stale) as you can see below:
>> >
>> > $ gluster volume quota myvolume list /demo.domain.tld
>> >
>> > Path Hard-limit Soft-limit Used Available Soft-limit exceeded? Hard-limit 
>> > exceeded?
>> >
>> >
>> > --
>> >
>> > /demo.domain.tld N/A N/A 8.0MB N/A N/A N/A
>> >
>> > So as you suggest I added again the quota on that directory and now the 
>> > "list" finally works again and show the quotas for every directories as I 
>> > defined them. That did the trick!
>> >
>> > Now do you know if this bug is already corrected in a new release of 
>> > GlusterFS? if not do you know when it will be fixed?
>> >
>> > Again many thanks for your help here!
>> >
>> > Best regards,
>> >
>> > M.
>> >
>> > ‐‐‐ Original Message ‐‐‐
>> >
>> > On February 23, 2018 7:45 AM, Hari Gowtham hgowt...@redhat.com wrote:
>> >
>> > > Hi,
>> > >
>> > > There is a bug in 3.10 which doesn't allow the quota list command to
>> > >
>> > > output, if the last entry on the conf file is a stale entry.
>> > >
>> > > The workaround for this is to remove the stale entry at the end. (If
>> > >
>> > > the last two entries are stale then both have to be removed and so on
>> > >
>> > > until the last entry on the conf file is a valid entry).
>> > >
>> > > This can be avoided by adding a new limit. As the new limit you added
>> > >
>> > > didn't work there is another way to check this.
>> > >
>> > > Try quota list command with a specific limit mentioned in the command.
>> > >
>> > > gluster volume quota  list 
>> > >
>> > > Make sure this path and the limit are set.
>> > >
>> > > If this works then you need to clean up the last stale entry.
>> > >
>> > > If this doesn't work we need to look further.
>> > >
>> > > Thanks Sanoj for the guidance.
>> > >
>> > > On Wed, Feb 14, 2018 at 1:36 AM, mabi m...@protonmail.ch wrote:
>> > >
>> > > > I tried to set the limits as you suggest by running the following 
>> > > > command.
>> > > >
>> > > > $ sudo gluster volume quota myvolume limit-usage /directory 200GB
>> > > >
>> > > > volume quota : success
>> > > >
>> > > > but then when I list the quotas there is still nothing, so nothing 
>> > > > really happened.
>> > > >
>> > > > I also tried to run stat on all directories which have a quota but 
>> > > > nothing happened either.
>> > > >
>> > > > I will send you tomorrow all the other logfiles as requested.
>> > > >
>> > > > \-\-\-\-\-\-\-\- Original Message 
>> > > >
>> > > > On February 13, 2018 12:20 PM, Hari Gowtham hgowt...@redhat.com wrote:
>> > > >
>> > > > > Were you able to set new limits after seeing this error?
>> > > > >
>> > > > > On Tue, Feb 13, 2018 at 4:19 PM, Hari Gowtham hgowt...@redhat.com 
>> > > > > wrote:
>> > > > >
>> > > > > > Yes, I need the log files in that duration, the log rotated file 
>> > > > > > after
>> > > > > >
>> > > > > > hitting the
>> > > > > >
>> > > > > > issue aren't necessary, but the ones before hitting the issues are 
>> > > > > > needed
>> > > > > >
>> > > > > > (not just when you hit it, the ones even before you hit it).
>> > > > > >
>> > > > > > Yes, you have to do a stat from the client through fuse mount.
>> > > > > >
>> > > > > > On Tue, Feb 13, 2018 at 3:56 PM, mabi m...@protonmail.ch wrote:
>> > > > > >
>> > > > > > > Thank you for your answer. This problem seem to have started 
>> > > > > > > since last week, so should I also send you the same log files 
>> > > > > > > but for last week? I think logrotate rotates them on a weekly 
>> > > > > > > basis.
>> > > > > > >
>> > > > > > > The only two quota commands we use are the following:
>> > > > > > >
>> > > > > > > gluster volume quota myvolume limit-usage /directory 10GB
>> > > > > > >
>> > > > > > > gluster volume quota myvolume list
>> > > > > > >
>> > > > > > > basically to set a new quota or to list the current quotas. The 
>> > > > > > > quota list was working in the past yes but we already had a 
>> > > > > > > similar issue where the quotas 

Re: [Gluster-users] Failed to get quota limits

2018-02-27 Thread mabi
Hi,

Thanks for the link to the bug. We should be hopefully moving soon onto 3.12 so 
I guess this bug is also fixed there.

Best regards,
M.
​

‐‐‐ Original Message ‐‐‐

On February 27, 2018 9:38 AM, Hari Gowtham  wrote:

> ​​
> 
> Hi Mabi,
> 
> The bugs is fixed from 3.11. For 3.10 it is yet to be backported and
> 
> made available.
> 
> The bug is https://bugzilla.redhat.com/show_bug.cgi?id=1418259.
> 
> On Sat, Feb 24, 2018 at 4:05 PM, mabi m...@protonmail.ch wrote:
> 
> > Dear Hari,
> > 
> > Thank you for getting back to me after having analysed the problem.
> > 
> > As you said I tried to run "gluster volume quota  list " for 
> > all of my directories which have a quota and found out that there was one 
> > directory quota which was missing (stale) as you can see below:
> > 
> > $ gluster volume quota myvolume list /demo.domain.tld
> > 
> > Path Hard-limit Soft-limit Used Available Soft-limit exceeded? Hard-limit 
> > exceeded?
> > 
> > 
> > --
> > 
> > /demo.domain.tld N/A N/A 8.0MB N/A N/A N/A
> > 
> > So as you suggest I added again the quota on that directory and now the 
> > "list" finally works again and show the quotas for every directories as I 
> > defined them. That did the trick!
> > 
> > Now do you know if this bug is already corrected in a new release of 
> > GlusterFS? if not do you know when it will be fixed?
> > 
> > Again many thanks for your help here!
> > 
> > Best regards,
> > 
> > M.
> > 
> > ‐‐‐ Original Message ‐‐‐
> > 
> > On February 23, 2018 7:45 AM, Hari Gowtham hgowt...@redhat.com wrote:
> > 
> > > Hi,
> > > 
> > > There is a bug in 3.10 which doesn't allow the quota list command to
> > > 
> > > output, if the last entry on the conf file is a stale entry.
> > > 
> > > The workaround for this is to remove the stale entry at the end. (If
> > > 
> > > the last two entries are stale then both have to be removed and so on
> > > 
> > > until the last entry on the conf file is a valid entry).
> > > 
> > > This can be avoided by adding a new limit. As the new limit you added
> > > 
> > > didn't work there is another way to check this.
> > > 
> > > Try quota list command with a specific limit mentioned in the command.
> > > 
> > > gluster volume quota  list 
> > > 
> > > Make sure this path and the limit are set.
> > > 
> > > If this works then you need to clean up the last stale entry.
> > > 
> > > If this doesn't work we need to look further.
> > > 
> > > Thanks Sanoj for the guidance.
> > > 
> > > On Wed, Feb 14, 2018 at 1:36 AM, mabi m...@protonmail.ch wrote:
> > > 
> > > > I tried to set the limits as you suggest by running the following 
> > > > command.
> > > > 
> > > > $ sudo gluster volume quota myvolume limit-usage /directory 200GB
> > > > 
> > > > volume quota : success
> > > > 
> > > > but then when I list the quotas there is still nothing, so nothing 
> > > > really happened.
> > > > 
> > > > I also tried to run stat on all directories which have a quota but 
> > > > nothing happened either.
> > > > 
> > > > I will send you tomorrow all the other logfiles as requested.
> > > > 
> > > > \-\-\-\-\-\-\-\- Original Message 
> > > > 
> > > > On February 13, 2018 12:20 PM, Hari Gowtham hgowt...@redhat.com wrote:
> > > > 
> > > > > Were you able to set new limits after seeing this error?
> > > > > 
> > > > > On Tue, Feb 13, 2018 at 4:19 PM, Hari Gowtham hgowt...@redhat.com 
> > > > > wrote:
> > > > > 
> > > > > > Yes, I need the log files in that duration, the log rotated file 
> > > > > > after
> > > > > > 
> > > > > > hitting the
> > > > > > 
> > > > > > issue aren't necessary, but the ones before hitting the issues are 
> > > > > > needed
> > > > > > 
> > > > > > (not just when you hit it, the ones even before you hit it).
> > > > > > 
> > > > > > Yes, you have to do a stat from the client through fuse mount.
> > > > > > 
> > > > > > On Tue, Feb 13, 2018 at 3:56 PM, mabi m...@protonmail.ch wrote:
> > > > > > 
> > > > > > > Thank you for your answer. This problem seem to have started 
> > > > > > > since last week, so should I also send you the same log files but 
> > > > > > > for last week? I think logrotate rotates them on a weekly basis.
> > > > > > > 
> > > > > > > The only two quota commands we use are the following:
> > > > > > > 
> > > > > > > gluster volume quota myvolume limit-usage /directory 10GB
> > > > > > > 
> > > > > > > gluster volume quota myvolume list
> > > > > > > 
> > > > > > > basically to set a new quota or to list the current quotas. The 
> > > > > > > quota list was working in the past yes but we already had a 
> > > > > > > similar issue where the quotas disappeared last August 2017:
> > > > > > > 
> > > > > > > http://lists.gluster.org/pipermail/gluster-users/2017-August/031946.html
> > > > > > > 
> > > > > > > In the mean time the only thing we did is to upgrade 

Re: [Gluster-users] Failed to get quota limits

2018-02-27 Thread Hari Gowtham
Hi Mabi,

The bugs is fixed from 3.11. For 3.10 it is yet to be backported and
made available.

The bug is https://bugzilla.redhat.com/show_bug.cgi?id=1418259.

On Sat, Feb 24, 2018 at 4:05 PM, mabi  wrote:
> Dear Hari,
>
> Thank you for getting back to me after having analysed the problem.
>
> As you said I tried to run "gluster volume quota  list " for 
> all of my directories which have a quota and found out that there was one 
> directory quota which was missing (stale) as you can see below:
>
> $ gluster volume quota myvolume list /demo.domain.tld
>   Path   Hard-limit  Soft-limit  Used  
> Available  Soft-limit exceeded? Hard-limit exceeded?
> ---
> /demo.domain.tldN/AN/A  8.0MB 
>   N/A N/AN/A
>
> So as you suggest I added again the quota on that directory and now the 
> "list" finally works again and show the quotas for every directories as I 
> defined them. That did the trick!
>
> Now do you know if this bug is already corrected in a new release of 
> GlusterFS? if not do you know when it will be fixed?
>
> Again many thanks for your help here!
>
> Best regards,
> M.
>
> ‐‐‐ Original Message ‐‐‐
>
> On February 23, 2018 7:45 AM, Hari Gowtham  wrote:
>
>>
>>
>> Hi,
>>
>> There is a bug in 3.10 which doesn't allow the quota list command to
>>
>> output, if the last entry on the conf file is a stale entry.
>>
>> The workaround for this is to remove the stale entry at the end. (If
>>
>> the last two entries are stale then both have to be removed and so on
>>
>> until the last entry on the conf file is a valid entry).
>>
>> This can be avoided by adding a new limit. As the new limit you added
>>
>> didn't work there is another way to check this.
>>
>> Try quota list command with a specific limit mentioned in the command.
>>
>> gluster volume quota  list 
>>
>> Make sure this path and the limit are set.
>>
>> If this works then you need to clean up the last stale entry.
>>
>> If this doesn't work we need to look further.
>>
>> Thanks Sanoj for the guidance.
>>
>> On Wed, Feb 14, 2018 at 1:36 AM, mabi m...@protonmail.ch wrote:
>>
>> > I tried to set the limits as you suggest by running the following command.
>> >
>> > $ sudo gluster volume quota myvolume limit-usage /directory 200GB
>> >
>> > volume quota : success
>> >
>> > but then when I list the quotas there is still nothing, so nothing really 
>> > happened.
>> >
>> > I also tried to run stat on all directories which have a quota but nothing 
>> > happened either.
>> >
>> > I will send you tomorrow all the other logfiles as requested.
>> >
>> > \-\-\-\-\-\-\-\- Original Message 
>> >
>> > On February 13, 2018 12:20 PM, Hari Gowtham hgowt...@redhat.com wrote:
>> >
>> > > Were you able to set new limits after seeing this error?
>> > >
>> > > On Tue, Feb 13, 2018 at 4:19 PM, Hari Gowtham hgowt...@redhat.com wrote:
>> > >
>> > > > Yes, I need the log files in that duration, the log rotated file after
>> > > >
>> > > > hitting the
>> > > >
>> > > > issue aren't necessary, but the ones before hitting the issues are 
>> > > > needed
>> > > >
>> > > > (not just when you hit it, the ones even before you hit it).
>> > > >
>> > > > Yes, you have to do a stat from the client through fuse mount.
>> > > >
>> > > > On Tue, Feb 13, 2018 at 3:56 PM, mabi m...@protonmail.ch wrote:
>> > > >
>> > > > > Thank you for your answer. This problem seem to have started since 
>> > > > > last week, so should I also send you the same log files but for last 
>> > > > > week? I think logrotate rotates them on a weekly basis.
>> > > > >
>> > > > > The only two quota commands we use are the following:
>> > > > >
>> > > > > gluster volume quota myvolume limit-usage /directory 10GB
>> > > > >
>> > > > > gluster volume quota myvolume list
>> > > > >
>> > > > > basically to set a new quota or to list the current quotas. The 
>> > > > > quota list was working in the past yes but we already had a similar 
>> > > > > issue where the quotas disappeared last August 2017:
>> > > > >
>> > > > > http://lists.gluster.org/pipermail/gluster-users/2017-August/031946.html
>> > > > >
>> > > > > In the mean time the only thing we did is to upgrade from 3.8 to 
>> > > > > 3.10.
>> > > > >
>> > > > > There are actually no errors to be seen using any gluster commands. 
>> > > > > The "quota myvolume list" returns simply nothing.
>> > > > >
>> > > > > In order to lookup the directories should I run a "stat" on them? 
>> > > > > and if yes should I do that on a client through the fuse mount?
>> > > > >
>> > > > > \-\-\-\-\-\-\-\- Original Message 
>> > > > >
>> > > > > On February 13, 2018 10:58 AM, Hari Gowtham hgowt...@redhat.com 
>> > > > > wrote:
>> > > > >
>> > > > > > The log provided are from 11th, 

Re: [Gluster-users] Failed to get quota limits

2018-02-24 Thread mabi
Dear Hari,

Thank you for getting back to me after having analysed the problem.

As you said I tried to run "gluster volume quota  list " for all 
of my directories which have a quota and found out that there was one directory 
quota which was missing (stale) as you can see below:

$ gluster volume quota myvolume list /demo.domain.tld
  Path   Hard-limit  Soft-limit  Used  
Available  Soft-limit exceeded? Hard-limit exceeded?
---
/demo.domain.tldN/AN/A  8.0MB   
N/A N/AN/A

So as you suggest I added again the quota on that directory and now the "list" 
finally works again and show the quotas for every directories as I defined 
them. That did the trick!

Now do you know if this bug is already corrected in a new release of GlusterFS? 
if not do you know when it will be fixed?

Again many thanks for your help here!

Best regards,
M.

‐‐‐ Original Message ‐‐‐

On February 23, 2018 7:45 AM, Hari Gowtham  wrote:

> ​​
> 
> Hi,
> 
> There is a bug in 3.10 which doesn't allow the quota list command to
> 
> output, if the last entry on the conf file is a stale entry.
> 
> The workaround for this is to remove the stale entry at the end. (If
> 
> the last two entries are stale then both have to be removed and so on
> 
> until the last entry on the conf file is a valid entry).
> 
> This can be avoided by adding a new limit. As the new limit you added
> 
> didn't work there is another way to check this.
> 
> Try quota list command with a specific limit mentioned in the command.
> 
> gluster volume quota  list 
> 
> Make sure this path and the limit are set.
> 
> If this works then you need to clean up the last stale entry.
> 
> If this doesn't work we need to look further.
> 
> Thanks Sanoj for the guidance.
> 
> On Wed, Feb 14, 2018 at 1:36 AM, mabi m...@protonmail.ch wrote:
> 
> > I tried to set the limits as you suggest by running the following command.
> > 
> > $ sudo gluster volume quota myvolume limit-usage /directory 200GB
> > 
> > volume quota : success
> > 
> > but then when I list the quotas there is still nothing, so nothing really 
> > happened.
> > 
> > I also tried to run stat on all directories which have a quota but nothing 
> > happened either.
> > 
> > I will send you tomorrow all the other logfiles as requested.
> > 
> > \-\-\-\-\-\-\-\- Original Message 
> > 
> > On February 13, 2018 12:20 PM, Hari Gowtham hgowt...@redhat.com wrote:
> > 
> > > Were you able to set new limits after seeing this error?
> > > 
> > > On Tue, Feb 13, 2018 at 4:19 PM, Hari Gowtham hgowt...@redhat.com wrote:
> > > 
> > > > Yes, I need the log files in that duration, the log rotated file after
> > > > 
> > > > hitting the
> > > > 
> > > > issue aren't necessary, but the ones before hitting the issues are 
> > > > needed
> > > > 
> > > > (not just when you hit it, the ones even before you hit it).
> > > > 
> > > > Yes, you have to do a stat from the client through fuse mount.
> > > > 
> > > > On Tue, Feb 13, 2018 at 3:56 PM, mabi m...@protonmail.ch wrote:
> > > > 
> > > > > Thank you for your answer. This problem seem to have started since 
> > > > > last week, so should I also send you the same log files but for last 
> > > > > week? I think logrotate rotates them on a weekly basis.
> > > > > 
> > > > > The only two quota commands we use are the following:
> > > > > 
> > > > > gluster volume quota myvolume limit-usage /directory 10GB
> > > > > 
> > > > > gluster volume quota myvolume list
> > > > > 
> > > > > basically to set a new quota or to list the current quotas. The quota 
> > > > > list was working in the past yes but we already had a similar issue 
> > > > > where the quotas disappeared last August 2017:
> > > > > 
> > > > > http://lists.gluster.org/pipermail/gluster-users/2017-August/031946.html
> > > > > 
> > > > > In the mean time the only thing we did is to upgrade from 3.8 to 3.10.
> > > > > 
> > > > > There are actually no errors to be seen using any gluster commands. 
> > > > > The "quota myvolume list" returns simply nothing.
> > > > > 
> > > > > In order to lookup the directories should I run a "stat" on them? and 
> > > > > if yes should I do that on a client through the fuse mount?
> > > > > 
> > > > > \-\-\-\-\-\-\-\- Original Message 
> > > > > 
> > > > > On February 13, 2018 10:58 AM, Hari Gowtham hgowt...@redhat.com wrote:
> > > > > 
> > > > > > The log provided are from 11th, you have seen the issue a while 
> > > > > > before
> > > > > > 
> > > > > > that itself.
> > > > > > 
> > > > > > The logs help us to know if something has actually went wrong.
> > > > > > 
> > > > > > once something goes wrong the output might get affected and i need 
> > > > > > to know what
> > > > > > 
> > > > > > went wrong. Which means i need the 

Re: [Gluster-users] Failed to get quota limits

2018-02-22 Thread Hari Gowtham
Hi,

There is a bug in 3.10 which doesn't allow the quota list command to
output, if the last entry on the conf file is a stale entry.
The workaround for this is to remove the stale entry at the end. (If
the last two entries are stale then both have to be removed and so on
until the last entry on the conf file is a valid entry).
This can be avoided by adding a new limit. As the new limit you added
didn't work there is another way to check this.

Try quota list command with a specific limit mentioned in the command.
gluster volume quota  list 
Make sure this path and the limit are set.

If this works then you need to clean up the last stale entry.
If this doesn't work we need to look further.

Thanks Sanoj for the guidance.

On Wed, Feb 14, 2018 at 1:36 AM, mabi  wrote:
> I tried to set the limits as you suggest by running the following command.
>
> $ sudo gluster volume quota myvolume limit-usage /directory 200GB
> volume quota : success
>
> but then when I list the quotas there is still nothing, so nothing really 
> happened.
>
> I also tried to run stat on all directories which have a quota but nothing 
> happened either.
>
> I will send you tomorrow all the other logfiles as requested.
>
>
>  Original Message 
>  On February 13, 2018 12:20 PM, Hari Gowtham  wrote:
>
>>Were you able to set new limits after seeing this error?
>>
>> On Tue, Feb 13, 2018 at 4:19 PM, Hari Gowtham hgowt...@redhat.com wrote:
>>>Yes, I need the log files in that duration, the log rotated file after
>>> hitting the
>>> issue aren't necessary, but the ones before hitting the issues are needed
>>> (not just when you hit it, the ones even before you hit it).
>>>Yes, you have to do a stat from the client through fuse mount.
>>>On Tue, Feb 13, 2018 at 3:56 PM, mabi m...@protonmail.ch wrote:
Thank you for your answer. This problem seem to have started since last 
week, so should I also send you the same log files but for last week? I 
think logrotate rotates them on a weekly basis.
The only two quota commands we use are the following:
gluster volume quota myvolume limit-usage /directory 10GB
 gluster volume quota myvolume list
basically to set a new quota or to list the current quotas. The quota list 
was working in the past yes but we already had a similar issue where the 
quotas disappeared last August 2017:
http://lists.gluster.org/pipermail/gluster-users/2017-August/031946.html
In the mean time the only thing we did is to upgrade from 3.8 to 3.10.
There are actually no errors to be seen using any gluster commands. The 
"quota myvolume list" returns simply nothing.
In order to lookup the directories should I run a "stat" on them? and if 
yes should I do that on a client through the fuse mount?
 Original Message 
 On February 13, 2018 10:58 AM, Hari Gowtham hgowt...@redhat.com wrote:
>The log provided are from 11th, you have seen the issue a while before
> that itself.
>The logs help us to know if something has actually went wrong.
> once something goes wrong the output might get affected and i need to 
> know what
> went wrong. Which means i need the log from the beginning.
>and i need to know a few more things,
> Was the quota list command was working as expected at the beginning?
> If yes, what were the commands issued, before you noticed this problem.
> Is there any other error that you see other than this?
>And can you try looking up the directories the limits are set on and
> check if that fixes the error?
>> Original Message 
>> On February 13, 2018 10:44 AM, mabi m...@protonmail.ch wrote:
>>>Hi Hari,
>>> Sure no problem, I will send you in a minute another mail where you can 
>>> download all the relevant log files including the quota.conf binary 
>>> file. Let me know if you need anything else. In the mean time here 
>>> below is the output of a volume status.
>>> Best regards,
>>> M.
>>> Status of volume: myvolume
>>> Gluster process TCP Port  RDMA Port  Online 
>>>  Pid
>>> Brick gfs1a.domain.local:/data/myvolume
>>> /brick  49153 0  Y   3214
>>> Brick gfs1b.domain.local:/data/myvolume
>>> /brick  49154 0  Y   3256
>>> Brick gfs1c.domain.local:/srv/glusterf
>>> s/myvolume/brick 49153 0  Y   515
>>> Self-heal Daemon on localhost   N/A   N/AY  
>>>  3186
>>> Quota Daemon on localhost   N/A   N/AY  
>>>  3195
>>> Self-heal Daemon on gfs1b.domain.local N/A   N/AY   3217
>>> Quota Daemon on gfs1b.domain.local N/A   N/AY   3229
>>> Self-heal Daemon on gfs1c.domain.local N/A   N/A  

Re: [Gluster-users] Failed to get quota limits

2018-02-13 Thread mabi
I tried to set the limits as you suggest by running the following command.

$ sudo gluster volume quota myvolume limit-usage /directory 200GB
volume quota : success

but then when I list the quotas there is still nothing, so nothing really 
happened.

I also tried to run stat on all directories which have a quota but nothing 
happened either.

I will send you tomorrow all the other logfiles as requested.
​

 Original Message 
 On February 13, 2018 12:20 PM, Hari Gowtham  wrote:

>Were you able to set new limits after seeing this error?
>
> On Tue, Feb 13, 2018 at 4:19 PM, Hari Gowtham hgowt...@redhat.com wrote:
>>Yes, I need the log files in that duration, the log rotated file after
>> hitting the
>> issue aren't necessary, but the ones before hitting the issues are needed
>> (not just when you hit it, the ones even before you hit it).
>>Yes, you have to do a stat from the client through fuse mount.
>>On Tue, Feb 13, 2018 at 3:56 PM, mabi m...@protonmail.ch wrote:
>>>Thank you for your answer. This problem seem to have started since last 
>>>week, so should I also send you the same log files but for last week? I 
>>>think logrotate rotates them on a weekly basis.
>>>The only two quota commands we use are the following:
>>>gluster volume quota myvolume limit-usage /directory 10GB
>>> gluster volume quota myvolume list
>>>basically to set a new quota or to list the current quotas. The quota list 
>>>was working in the past yes but we already had a similar issue where the 
>>>quotas disappeared last August 2017:
>>>http://lists.gluster.org/pipermail/gluster-users/2017-August/031946.html
>>>In the mean time the only thing we did is to upgrade from 3.8 to 3.10.
>>>There are actually no errors to be seen using any gluster commands. The 
>>>"quota myvolume list" returns simply nothing.
>>>In order to lookup the directories should I run a "stat" on them? and if yes 
>>>should I do that on a client through the fuse mount?
>>> Original Message 
>>> On February 13, 2018 10:58 AM, Hari Gowtham hgowt...@redhat.com wrote:
The log provided are from 11th, you have seen the issue a while before
 that itself.
The logs help us to know if something has actually went wrong.
 once something goes wrong the output might get affected and i need to know 
 what
 went wrong. Which means i need the log from the beginning.
and i need to know a few more things,
 Was the quota list command was working as expected at the beginning?
 If yes, what were the commands issued, before you noticed this problem.
 Is there any other error that you see other than this?
And can you try looking up the directories the limits are set on and
 check if that fixes the error?
> Original Message 
> On February 13, 2018 10:44 AM, mabi m...@protonmail.ch wrote:
>>Hi Hari,
>> Sure no problem, I will send you in a minute another mail where you can 
>> download all the relevant log files including the quota.conf binary 
>> file. Let me know if you need anything else. In the mean time here below 
>> is the output of a volume status.
>> Best regards,
>> M.
>> Status of volume: myvolume
>> Gluster process TCP Port  RDMA Port  Online  
>> Pid
>> Brick gfs1a.domain.local:/data/myvolume
>> /brick  49153 0  Y   3214
>> Brick gfs1b.domain.local:/data/myvolume
>> /brick  49154 0  Y   3256
>> Brick gfs1c.domain.local:/srv/glusterf
>> s/myvolume/brick 49153 0  Y   515
>> Self-heal Daemon on localhost   N/A   N/AY   
>> 3186
>> Quota Daemon on localhost   N/A   N/AY   
>> 3195
>> Self-heal Daemon on gfs1b.domain.local N/A   N/AY   3217
>> Quota Daemon on gfs1b.domain.local N/A   N/AY   3229
>> Self-heal Daemon on gfs1c.domain.local N/A   N/AY   486
>> Quota Daemon on gfs1c.domain.local N/A   N/AY   495
>> Task Status of Volume myvolume
>> There are no active volume tasks
>>  Original Message 
>> On February 13, 2018 10:09 AM, Hari Gowtham hgowt...@redhat.com wrote:
>>>Hi,
>>> A part of the log won't be enough to debug the issue.
>>> Need the whole log messages till date.
>>> You can send it as attachments.
>>> Yes the quota.conf is a binary file.
>>> And I need the volume status output too.
>>> On Tue, Feb 13, 2018 at 1:56 PM, mabi m...@protonmail.ch wrote:
Hi Hari,
 Sorry for not providing you more details from the start. Here below 
 you will
 find all the relevant log entries and info. Regarding the quota.conf 
 file I
 have found one for my volume but it is a binary 

Re: [Gluster-users] Failed to get quota limits

2018-02-13 Thread Hari Gowtham
Were you able to set new limits after seeing this error?

On Tue, Feb 13, 2018 at 4:19 PM, Hari Gowtham  wrote:
> Yes, I need the log files in that duration, the log rotated file after
> hitting the
> issue aren't necessary, but the ones before hitting the issues are needed
> (not just when you hit it, the ones even before you hit it).
>
> Yes, you have to do a stat from the client through fuse mount.
>
> On Tue, Feb 13, 2018 at 3:56 PM, mabi  wrote:
>> Thank you for your answer. This problem seem to have started since last 
>> week, so should I also send you the same log files but for last week? I 
>> think logrotate rotates them on a weekly basis.
>>
>> The only two quota commands we use are the following:
>>
>> gluster volume quota myvolume limit-usage /directory 10GB
>> gluster volume quota myvolume list
>>
>> basically to set a new quota or to list the current quotas. The quota list 
>> was working in the past yes but we already had a similar issue where the 
>> quotas disappeared last August 2017:
>>
>> http://lists.gluster.org/pipermail/gluster-users/2017-August/031946.html
>>
>> In the mean time the only thing we did is to upgrade from 3.8 to 3.10.
>>
>> There are actually no errors to be seen using any gluster commands. The 
>> "quota myvolume list" returns simply nothing.
>>
>> In order to lookup the directories should I run a "stat" on them? and if yes 
>> should I do that on a client through the fuse mount?
>>
>>
>>  Original Message 
>>  On February 13, 2018 10:58 AM, Hari Gowtham  wrote:
>>
>>>The log provided are from 11th, you have seen the issue a while before
>>> that itself.
>>>
>>> The logs help us to know if something has actually went wrong.
>>> once something goes wrong the output might get affected and i need to know 
>>> what
>>> went wrong. Which means i need the log from the beginning.
>>>
>>> and i need to know a few more things,
>>> Was the quota list command was working as expected at the beginning?
>>> If yes, what were the commands issued, before you noticed this problem.
>>> Is there any other error that you see other than this?
>>>
>>> And can you try looking up the directories the limits are set on and
>>> check if that fixes the error?
>>>
 Original Message 
 On February 13, 2018 10:44 AM, mabi m...@protonmail.ch wrote:
>Hi Hari,
>Sure no problem, I will send you in a minute another mail where you can 
>download all the relevant log files including the quota.conf binary file. 
>Let me know if you need anything else. In the mean time here below is the 
>output of a volume status.
>Best regards,
> M.
>Status of volume: myvolume
> Gluster process TCP Port  RDMA Port  Online  
> Pid
>Brick gfs1a.domain.local:/data/myvolume
> /brick  49153 0  Y   3214
> Brick gfs1b.domain.local:/data/myvolume
> /brick  49154 0  Y   3256
> Brick gfs1c.domain.local:/srv/glusterf
> s/myvolume/brick 49153 0  Y   515
> Self-heal Daemon on localhost   N/A   N/AY   
> 3186
> Quota Daemon on localhost   N/A   N/AY   
> 3195
> Self-heal Daemon on gfs1b.domain.local N/A   N/AY   3217
> Quota Daemon on gfs1b.domain.local N/A   N/AY   3229
> Self-heal Daemon on gfs1c.domain.local N/A   N/AY   486
> Quota Daemon on gfs1c.domain.local N/A   N/AY   495
>Task Status of Volume myvolume
>There are no active volume tasks
> Original Message 
> On February 13, 2018 10:09 AM, Hari Gowtham hgowt...@redhat.com wrote:
>>Hi,
>> A part of the log won't be enough to debug the issue.
>> Need the whole log messages till date.
>> You can send it as attachments.
>> Yes the quota.conf is a binary file.
>> And I need the volume status output too.
>> On Tue, Feb 13, 2018 at 1:56 PM, mabi m...@protonmail.ch wrote:
>>>Hi Hari,
>>> Sorry for not providing you more details from the start. Here below you 
>>> will
>>> find all the relevant log entries and info. Regarding the quota.conf 
>>> file I
>>> have found one for my volume but it is a binary file. Is it supposed to 
>>> be
>>> binary or text?
>>> Regards,
>>> M.
>>> *** gluster volume info myvolume ***
>>> Volume Name: myvolume
>>> Type: Replicate
>>> Volume ID: e7a40a1b-45c9-4d3c-bb19-0c59b4eceec5
>>> Status: Started
>>> Snapshot Count: 0
>>> Number of Bricks: 1 x (2 + 1) = 3
>>> Transport-type: tcp
>>> Bricks:
>>> Brick1: gfs1a.domain.local:/data/myvolume/brick
>>> Brick2: gfs1b.domain.local:/data/myvolume/brick
>>> Brick3: 

Re: [Gluster-users] Failed to get quota limits

2018-02-13 Thread Hari Gowtham
Yes, I need the log files in that duration, the log rotated file after
hitting the
issue aren't necessary, but the ones before hitting the issues are needed
(not just when you hit it, the ones even before you hit it).

Yes, you have to do a stat from the client through fuse mount.

On Tue, Feb 13, 2018 at 3:56 PM, mabi  wrote:
> Thank you for your answer. This problem seem to have started since last week, 
> so should I also send you the same log files but for last week? I think 
> logrotate rotates them on a weekly basis.
>
> The only two quota commands we use are the following:
>
> gluster volume quota myvolume limit-usage /directory 10GB
> gluster volume quota myvolume list
>
> basically to set a new quota or to list the current quotas. The quota list 
> was working in the past yes but we already had a similar issue where the 
> quotas disappeared last August 2017:
>
> http://lists.gluster.org/pipermail/gluster-users/2017-August/031946.html
>
> In the mean time the only thing we did is to upgrade from 3.8 to 3.10.
>
> There are actually no errors to be seen using any gluster commands. The 
> "quota myvolume list" returns simply nothing.
>
> In order to lookup the directories should I run a "stat" on them? and if yes 
> should I do that on a client through the fuse mount?
>
>
>  Original Message 
>  On February 13, 2018 10:58 AM, Hari Gowtham  wrote:
>
>>The log provided are from 11th, you have seen the issue a while before
>> that itself.
>>
>> The logs help us to know if something has actually went wrong.
>> once something goes wrong the output might get affected and i need to know 
>> what
>> went wrong. Which means i need the log from the beginning.
>>
>> and i need to know a few more things,
>> Was the quota list command was working as expected at the beginning?
>> If yes, what were the commands issued, before you noticed this problem.
>> Is there any other error that you see other than this?
>>
>> And can you try looking up the directories the limits are set on and
>> check if that fixes the error?
>>
>>> Original Message 
>>> On February 13, 2018 10:44 AM, mabi m...@protonmail.ch wrote:
Hi Hari,
Sure no problem, I will send you in a minute another mail where you can 
download all the relevant log files including the quota.conf binary file. 
Let me know if you need anything else. In the mean time here below is the 
output of a volume status.
Best regards,
 M.
Status of volume: myvolume
 Gluster process TCP Port  RDMA Port  Online  
 Pid
Brick gfs1a.domain.local:/data/myvolume
 /brick  49153 0  Y   3214
 Brick gfs1b.domain.local:/data/myvolume
 /brick  49154 0  Y   3256
 Brick gfs1c.domain.local:/srv/glusterf
 s/myvolume/brick 49153 0  Y   515
 Self-heal Daemon on localhost   N/A   N/AY   
 3186
 Quota Daemon on localhost   N/A   N/AY   
 3195
 Self-heal Daemon on gfs1b.domain.local N/A   N/AY   3217
 Quota Daemon on gfs1b.domain.local N/A   N/AY   3229
 Self-heal Daemon on gfs1c.domain.local N/A   N/AY   486
 Quota Daemon on gfs1c.domain.local N/A   N/AY   495
Task Status of Volume myvolume
There are no active volume tasks
 Original Message 
 On February 13, 2018 10:09 AM, Hari Gowtham hgowt...@redhat.com wrote:
>Hi,
> A part of the log won't be enough to debug the issue.
> Need the whole log messages till date.
> You can send it as attachments.
> Yes the quota.conf is a binary file.
> And I need the volume status output too.
> On Tue, Feb 13, 2018 at 1:56 PM, mabi m...@protonmail.ch wrote:
>>Hi Hari,
>> Sorry for not providing you more details from the start. Here below you 
>> will
>> find all the relevant log entries and info. Regarding the quota.conf 
>> file I
>> have found one for my volume but it is a binary file. Is it supposed to 
>> be
>> binary or text?
>> Regards,
>> M.
>> *** gluster volume info myvolume ***
>> Volume Name: myvolume
>> Type: Replicate
>> Volume ID: e7a40a1b-45c9-4d3c-bb19-0c59b4eceec5
>> Status: Started
>> Snapshot Count: 0
>> Number of Bricks: 1 x (2 + 1) = 3
>> Transport-type: tcp
>> Bricks:
>> Brick1: gfs1a.domain.local:/data/myvolume/brick
>> Brick2: gfs1b.domain.local:/data/myvolume/brick
>> Brick3: gfs1c.domain.local:/srv/glusterfs/myvolume/brick (arbiter)
>> Options Reconfigured:
>> server.event-threads: 4
>> client.event-threads: 4
>> performance.readdir-ahead: on
>> nfs.disable: on
>> features.quota: on
>> features.inode-quota: on
>> 

Re: [Gluster-users] Failed to get quota limits

2018-02-13 Thread mabi
Thank you for your answer. This problem seem to have started since last week, 
so should I also send you the same log files but for last week? I think 
logrotate rotates them on a weekly basis.

The only two quota commands we use are the following:

gluster volume quota myvolume limit-usage /directory 10GB
gluster volume quota myvolume list

basically to set a new quota or to list the current quotas. The quota list was 
working in the past yes but we already had a similar issue where the quotas 
disappeared last August 2017:

http://lists.gluster.org/pipermail/gluster-users/2017-August/031946.html

In the mean time the only thing we did is to upgrade from 3.8 to 3.10.

There are actually no errors to be seen using any gluster commands. The "quota 
myvolume list" returns simply nothing.

In order to lookup the directories should I run a "stat" on them? and if yes 
should I do that on a client through the fuse mount?
​

 Original Message 
 On February 13, 2018 10:58 AM, Hari Gowtham  wrote:

>The log provided are from 11th, you have seen the issue a while before
> that itself.
>
> The logs help us to know if something has actually went wrong.
> once something goes wrong the output might get affected and i need to know 
> what
> went wrong. Which means i need the log from the beginning.
>
> and i need to know a few more things,
> Was the quota list command was working as expected at the beginning?
> If yes, what were the commands issued, before you noticed this problem.
> Is there any other error that you see other than this?
>
> And can you try looking up the directories the limits are set on and
> check if that fixes the error?
>
>> Original Message 
>> On February 13, 2018 10:44 AM, mabi m...@protonmail.ch wrote:
>>>Hi Hari,
>>>Sure no problem, I will send you in a minute another mail where you can 
>>>download all the relevant log files including the quota.conf binary file. 
>>>Let me know if you need anything else. In the mean time here below is the 
>>>output of a volume status.
>>>Best regards,
>>> M.
>>>Status of volume: myvolume
>>> Gluster process TCP Port  RDMA Port  Online  Pid
>>>Brick gfs1a.domain.local:/data/myvolume
>>> /brick  49153 0  Y   3214
>>> Brick gfs1b.domain.local:/data/myvolume
>>> /brick  49154 0  Y   3256
>>> Brick gfs1c.domain.local:/srv/glusterf
>>> s/myvolume/brick 49153 0  Y   515
>>> Self-heal Daemon on localhost   N/A   N/AY   
>>> 3186
>>> Quota Daemon on localhost   N/A   N/AY   
>>> 3195
>>> Self-heal Daemon on gfs1b.domain.local N/A   N/AY   3217
>>> Quota Daemon on gfs1b.domain.local N/A   N/AY   3229
>>> Self-heal Daemon on gfs1c.domain.local N/A   N/AY   486
>>> Quota Daemon on gfs1c.domain.local N/A   N/AY   495
>>>Task Status of Volume myvolume
>>>There are no active volume tasks
>>> Original Message 
>>> On February 13, 2018 10:09 AM, Hari Gowtham hgowt...@redhat.com wrote:
Hi,
 A part of the log won't be enough to debug the issue.
 Need the whole log messages till date.
 You can send it as attachments.
 Yes the quota.conf is a binary file.
 And I need the volume status output too.
 On Tue, Feb 13, 2018 at 1:56 PM, mabi m...@protonmail.ch wrote:
>Hi Hari,
> Sorry for not providing you more details from the start. Here below you 
> will
> find all the relevant log entries and info. Regarding the quota.conf file 
> I
> have found one for my volume but it is a binary file. Is it supposed to be
> binary or text?
> Regards,
> M.
> *** gluster volume info myvolume ***
> Volume Name: myvolume
> Type: Replicate
> Volume ID: e7a40a1b-45c9-4d3c-bb19-0c59b4eceec5
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 1 x (2 + 1) = 3
> Transport-type: tcp
> Bricks:
> Brick1: gfs1a.domain.local:/data/myvolume/brick
> Brick2: gfs1b.domain.local:/data/myvolume/brick
> Brick3: gfs1c.domain.local:/srv/glusterfs/myvolume/brick (arbiter)
> Options Reconfigured:
> server.event-threads: 4
> client.event-threads: 4
> performance.readdir-ahead: on
> nfs.disable: on
> features.quota: on
> features.inode-quota: on
> features.quota-deem-statfs: on
> transport.address-family: inet
> *** /var/log/glusterfs/glusterd.log ***
> [2018-02-13 08:16:09.929568] E [MSGID: 101042] 
> [compat.c:569:gf_umount_lazy]
> 0-management: Lazy unmount of /var/run/gluster/myvolume_quota_list/
> [2018-02-13 08:16:28.596527] I [MSGID: 106499]
> [glusterd-handler.c:4363:__glusterd_handle_status_volume] 0-management:
> Received status volume req for volume myvolume
> [2018-02-13 

Re: [Gluster-users] Failed to get quota limits

2018-02-13 Thread mabi
Hi Hari,

Sure no problem, I will send you in a minute another mail where you can 
download all the relevant log files including the quota.conf binary file. Let 
me know if you need anything else. In the mean time here below is the output of 
a volume status.

Best regards,
M.

Status of volume: myvolume
Gluster process TCP Port  RDMA Port  Online  Pid
--
Brick gfs1a.domain.local:/data/myvolume
/brick  49153 0  Y   3214 
Brick gfs1b.domain.local:/data/myvolume
/brick  49154 0  Y   3256 
Brick gfs1c.domain.local:/srv/glusterf
s/myvolume/brick 49153 0  Y   515  
Self-heal Daemon on localhost   N/A   N/AY   3186 
Quota Daemon on localhost   N/A   N/AY   3195 
Self-heal Daemon on gfs1b.domain.local N/A   N/AY   3217 
Quota Daemon on gfs1b.domain.local N/A   N/AY   3229 
Self-heal Daemon on gfs1c.domain.local N/A   N/AY   486  
Quota Daemon on gfs1c.domain.local N/A   N/AY   495  
 
Task Status of Volume myvolume
--
There are no active volume tasks


​

 Original Message 
 On February 13, 2018 10:09 AM, Hari Gowtham  wrote:

>Hi,
>
> A part of the log won't be enough to debug the issue.
> Need the whole log messages till date.
> You can send it as attachments.
>
> Yes the quota.conf is a binary file.
>
> And I need the volume status output too.
>
> On Tue, Feb 13, 2018 at 1:56 PM, mabi m...@protonmail.ch wrote:
>>Hi Hari,
>> Sorry for not providing you more details from the start. Here below you will
>> find all the relevant log entries and info. Regarding the quota.conf file I
>> have found one for my volume but it is a binary file. Is it supposed to be
>> binary or text?
>> Regards,
>> M.
>>*** gluster volume info myvolume ***
>>Volume Name: myvolume
>> Type: Replicate
>> Volume ID: e7a40a1b-45c9-4d3c-bb19-0c59b4eceec5
>> Status: Started
>> Snapshot Count: 0
>> Number of Bricks: 1 x (2 + 1) = 3
>> Transport-type: tcp
>> Bricks:
>> Brick1: gfs1a.domain.local:/data/myvolume/brick
>> Brick2: gfs1b.domain.local:/data/myvolume/brick
>> Brick3: gfs1c.domain.local:/srv/glusterfs/myvolume/brick (arbiter)
>> Options Reconfigured:
>> server.event-threads: 4
>> client.event-threads: 4
>> performance.readdir-ahead: on
>> nfs.disable: on
>> features.quota: on
>> features.inode-quota: on
>> features.quota-deem-statfs: on
>> transport.address-family: inet
>>*** /var/log/glusterfs/glusterd.log ***
>>[2018-02-13 08:16:09.929568] E [MSGID: 101042] [compat.c:569:gf_umount_lazy]
>> 0-management: Lazy unmount of /var/run/gluster/myvolume_quota_list/
>> [2018-02-13 08:16:28.596527] I [MSGID: 106499]
>> [glusterd-handler.c:4363:__glusterd_handle_status_volume] 0-management:
>> Received status volume req for volume myvolume
>> [2018-02-13 08:16:28.601097] I [MSGID: 106419]
>> [glusterd-utils.c:6110:glusterd_add_inode_size_to_dict] 0-management: the
>> brick on data/myvolume (zfs) uses dynamic inode sizes
>>*** /var/log/glusterfs/cmd_history.log ***
>>[2018-02-13 08:16:28.605478]  : volume status myvolume detail : SUCCESS
>>*** /var/log/glusterfs/quota-mount-myvolume.log ***
>>[2018-02-13 08:16:09.934117] I [MSGID: 100030] [glusterfsd.c:2503:main]
>> 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version 3.10.7
>> (args: /usr/sbin/glusterfs --volfile-server localhost --volfile-id myvolume
>> -l /var/log/glusterfs/quota-mount-myvolume.log -p
>> /var/run/gluster/myvolume_quota_list.pid --client-pid -5
>> /var/run/gluster/myvolume_quota_list/)
>> [2018-02-13 08:16:09.940432] I [MSGID: 101190]
>> [event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: Started thread with
>> index 1
>> [2018-02-13 08:16:09.940491] E [socket.c:2327:socket_connect_finish]
>> 0-glusterfs: connection to ::1:24007 failed (Connection refused);
>> disconnecting socket
>> [2018-02-13 08:16:09.940519] I [glusterfsd-mgmt.c:2134:mgmt_rpc_notify]
>> 0-glusterfsd-mgmt: disconnected from remote-host: localhost
>> [2018-02-13 08:16:13.943827] I [afr.c:94:fix_quorum_options]
>> 0-myvolume-replicate-0: reindeer: incoming qtype = none
>> [2018-02-13 08:16:13.943857] I [afr.c:116:fix_quorum_options]
>> 0-myvolume-replicate-0: reindeer: quorum_count = 2147483647
>> [2018-02-13 08:16:13.945194] I [MSGID: 101190]
>> [event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: Started thread with
>> index 2
>> [2018-02-13 08:16:13.945237] I [MSGID: 101190]
>> [event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: Started thread with
>> index 4
>> [2018-02-13 08:16:13.945247] I [MSGID: 101190]
>> [event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: Started thread with
>> index 3
>> [2018-02-13 

Re: [Gluster-users] Failed to get quota limits

2018-02-13 Thread Hari Gowtham
Hi,

A part of the log won't be enough to debug the issue.
Need the whole log messages till date.
You can send it as attachments.

Yes the quota.conf is a binary file.

And I need the volume status output too.

On Tue, Feb 13, 2018 at 1:56 PM, mabi  wrote:
> Hi Hari,
> Sorry for not providing you more details from the start. Here below you will
> find all the relevant log entries and info. Regarding the quota.conf file I
> have found one for my volume but it is a binary file. Is it supposed to be
> binary or text?
> Regards,
> M.
>
> *** gluster volume info myvolume ***
>
> Volume Name: myvolume
> Type: Replicate
> Volume ID: e7a40a1b-45c9-4d3c-bb19-0c59b4eceec5
> Status: Started
> Snapshot Count: 0
> Number of Bricks: 1 x (2 + 1) = 3
> Transport-type: tcp
> Bricks:
> Brick1: gfs1a.domain.local:/data/myvolume/brick
> Brick2: gfs1b.domain.local:/data/myvolume/brick
> Brick3: gfs1c.domain.local:/srv/glusterfs/myvolume/brick (arbiter)
> Options Reconfigured:
> server.event-threads: 4
> client.event-threads: 4
> performance.readdir-ahead: on
> nfs.disable: on
> features.quota: on
> features.inode-quota: on
> features.quota-deem-statfs: on
> transport.address-family: inet
>
>
> *** /var/log/glusterfs/glusterd.log ***
>
> [2018-02-13 08:16:09.929568] E [MSGID: 101042] [compat.c:569:gf_umount_lazy]
> 0-management: Lazy unmount of /var/run/gluster/myvolume_quota_list/
> [2018-02-13 08:16:28.596527] I [MSGID: 106499]
> [glusterd-handler.c:4363:__glusterd_handle_status_volume] 0-management:
> Received status volume req for volume myvolume
> [2018-02-13 08:16:28.601097] I [MSGID: 106419]
> [glusterd-utils.c:6110:glusterd_add_inode_size_to_dict] 0-management: the
> brick on data/myvolume (zfs) uses dynamic inode sizes
>
>
> *** /var/log/glusterfs/cmd_history.log ***
>
> [2018-02-13 08:16:28.605478]  : volume status myvolume detail : SUCCESS
>
>
> *** /var/log/glusterfs/quota-mount-myvolume.log ***
>
> [2018-02-13 08:16:09.934117] I [MSGID: 100030] [glusterfsd.c:2503:main]
> 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version 3.10.7
> (args: /usr/sbin/glusterfs --volfile-server localhost --volfile-id myvolume
> -l /var/log/glusterfs/quota-mount-myvolume.log -p
> /var/run/gluster/myvolume_quota_list.pid --client-pid -5
> /var/run/gluster/myvolume_quota_list/)
> [2018-02-13 08:16:09.940432] I [MSGID: 101190]
> [event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: Started thread with
> index 1
> [2018-02-13 08:16:09.940491] E [socket.c:2327:socket_connect_finish]
> 0-glusterfs: connection to ::1:24007 failed (Connection refused);
> disconnecting socket
> [2018-02-13 08:16:09.940519] I [glusterfsd-mgmt.c:2134:mgmt_rpc_notify]
> 0-glusterfsd-mgmt: disconnected from remote-host: localhost
> [2018-02-13 08:16:13.943827] I [afr.c:94:fix_quorum_options]
> 0-myvolume-replicate-0: reindeer: incoming qtype = none
> [2018-02-13 08:16:13.943857] I [afr.c:116:fix_quorum_options]
> 0-myvolume-replicate-0: reindeer: quorum_count = 2147483647
> [2018-02-13 08:16:13.945194] I [MSGID: 101190]
> [event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: Started thread with
> index 2
> [2018-02-13 08:16:13.945237] I [MSGID: 101190]
> [event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: Started thread with
> index 4
> [2018-02-13 08:16:13.945247] I [MSGID: 101190]
> [event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: Started thread with
> index 3
> [2018-02-13 08:16:13.945941] W [MSGID: 101174]
> [graph.c:361:_log_if_unknown_option] 0-myvolume-readdir-ahead: option
> 'parallel-readdir' is not recognized
> [2018-02-13 08:16:13.946342] I [MSGID: 114020] [client.c:2352:notify]
> 0-myvolume-client-0: parent translators are ready, attempting connect on
> transport
> [2018-02-13 08:16:13.946789] I [MSGID: 114020] [client.c:2352:notify]
> 0-myvolume-client-1: parent translators are ready, attempting connect on
> transport
> [2018-02-13 08:16:13.947151] I [rpc-clnt.c:2000:rpc_clnt_reconfig]
> 0-myvolume-client-0: changing port to 49153 (from 0)
> [2018-02-13 08:16:13.947846] I [MSGID: 114057]
> [client-handshake.c:1451:select_server_supported_programs]
> 0-myvolume-client-0: Using Program GlusterFS 3.3, Num (1298437), Version
> (330)
> [2018-02-13 08:16:13.948073] I [MSGID: 114020] [client.c:2352:notify]
> 0-myvolume-client-2: parent translators are ready, attempting connect on
> transport
> [2018-02-13 08:16:13.948579] I [rpc-clnt.c:2000:rpc_clnt_reconfig]
> 0-myvolume-client-1: changing port to 49154 (from 0)
> [2018-02-13 08:16:13.948699] I [MSGID: 114046]
> [client-handshake.c:1216:client_setvolume_cbk] 0-myvolume-client-0:
> Connected to myvolume-client-0, attached to remote volume
> '/data/myvolume/brick'.
> [2018-02-13 08:16:13.948747] I [MSGID: 114047]
> [client-handshake.c:1227:client_setvolume_cbk] 0-myvolume-client-0: Server
> and Client lk-version numbers are not same, reopening the fds
> [2018-02-13 08:16:13.948842] I [MSGID: 108005]
> [afr-common.c:4813:afr_notify] 0-myvolume-replicate-0: 

Re: [Gluster-users] Failed to get quota limits

2018-02-13 Thread mabi
Hi Hari,
Sorry for not providing you more details from the start. Here below you will 
find all the relevant log entries and info. Regarding the quota.conf file I 
have found one for my volume but it is a binary file. Is it supposed to be 
binary or text?
Regards,
M.

*** gluster volume info myvolume ***

Volume Name: myvolume
Type: Replicate
Volume ID: e7a40a1b-45c9-4d3c-bb19-0c59b4eceec5
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x (2 + 1) = 3
Transport-type: tcp
Bricks:
Brick1: gfs1a.domain.local:/data/myvolume/brick
Brick2: gfs1b.domain.local:/data/myvolume/brick
Brick3: gfs1c.domain.local:/srv/glusterfs/myvolume/brick (arbiter)
Options Reconfigured:
server.event-threads: 4
client.event-threads: 4
performance.readdir-ahead: on
nfs.disable: on
features.quota: on
features.inode-quota: on
features.quota-deem-statfs: on
transport.address-family: inet

*** /var/log/glusterfs/glusterd.log ***

[2018-02-13 08:16:09.929568] E [MSGID: 101042] [compat.c:569:gf_umount_lazy] 
0-management: Lazy unmount of /var/run/gluster/myvolume_quota_list/
[2018-02-13 08:16:28.596527] I [MSGID: 106499] 
[glusterd-handler.c:4363:__glusterd_handle_status_volume] 0-management: 
Received status volume req for volume myvolume
[2018-02-13 08:16:28.601097] I [MSGID: 106419] 
[glusterd-utils.c:6110:glusterd_add_inode_size_to_dict] 0-management: the brick 
on data/myvolume (zfs) uses dynamic inode sizes

*** /var/log/glusterfs/cmd_history.log ***

[2018-02-13 08:16:28.605478]  : volume status myvolume detail : SUCCESS

*** /var/log/glusterfs/quota-mount-myvolume.log ***

[2018-02-13 08:16:09.934117] I [MSGID: 100030] [glusterfsd.c:2503:main] 
0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version 3.10.7 
(args: /usr/sbin/glusterfs --volfile-server localhost --volfile-id myvolume -l 
/var/log/glusterfs/quota-mount-myvolume.log -p 
/var/run/gluster/myvolume_quota_list.pid --client-pid -5 
/var/run/gluster/myvolume_quota_list/)
[2018-02-13 08:16:09.940432] I [MSGID: 101190] 
[event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: Started thread with 
index 1
[2018-02-13 08:16:09.940491] E [socket.c:2327:socket_connect_finish] 
0-glusterfs: connection to ::1:24007 failed (Connection refused); disconnecting 
socket
[2018-02-13 08:16:09.940519] I [glusterfsd-mgmt.c:2134:mgmt_rpc_notify] 
0-glusterfsd-mgmt: disconnected from remote-host: localhost
[2018-02-13 08:16:13.943827] I [afr.c:94:fix_quorum_options] 
0-myvolume-replicate-0: reindeer: incoming qtype = none
[2018-02-13 08:16:13.943857] I [afr.c:116:fix_quorum_options] 
0-myvolume-replicate-0: reindeer: quorum_count = 2147483647
[2018-02-13 08:16:13.945194] I [MSGID: 101190] 
[event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: Started thread with 
index 2
[2018-02-13 08:16:13.945237] I [MSGID: 101190] 
[event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: Started thread with 
index 4
[2018-02-13 08:16:13.945247] I [MSGID: 101190] 
[event-epoll.c:629:event_dispatch_epoll_worker] 0-epoll: Started thread with 
index 3
[2018-02-13 08:16:13.945941] W [MSGID: 101174] 
[graph.c:361:_log_if_unknown_option] 0-myvolume-readdir-ahead: option 
'parallel-readdir' is not recognized
[2018-02-13 08:16:13.946342] I [MSGID: 114020] [client.c:2352:notify] 
0-myvolume-client-0: parent translators are ready, attempting connect on 
transport
[2018-02-13 08:16:13.946789] I [MSGID: 114020] [client.c:2352:notify] 
0-myvolume-client-1: parent translators are ready, attempting connect on 
transport
[2018-02-13 08:16:13.947151] I [rpc-clnt.c:2000:rpc_clnt_reconfig] 
0-myvolume-client-0: changing port to 49153 (from 0)
[2018-02-13 08:16:13.947846] I [MSGID: 114057] 
[client-handshake.c:1451:select_server_supported_programs] 0-myvolume-client-0: 
Using Program GlusterFS 3.3, Num (1298437), Version (330)
[2018-02-13 08:16:13.948073] I [MSGID: 114020] [client.c:2352:notify] 
0-myvolume-client-2: parent translators are ready, attempting connect on 
transport
[2018-02-13 08:16:13.948579] I [rpc-clnt.c:2000:rpc_clnt_reconfig] 
0-myvolume-client-1: changing port to 49154 (from 0)
[2018-02-13 08:16:13.948699] I [MSGID: 114046] 
[client-handshake.c:1216:client_setvolume_cbk] 0-myvolume-client-0: Connected 
to myvolume-client-0, attached to remote volume '/data/myvolume/brick'.
[2018-02-13 08:16:13.948747] I [MSGID: 114047] 
[client-handshake.c:1227:client_setvolume_cbk] 0-myvolume-client-0: Server and 
Client lk-version numbers are not same, reopening the fds
[2018-02-13 08:16:13.948842] I [MSGID: 108005] [afr-common.c:4813:afr_notify] 
0-myvolume-replicate-0: Subvolume 'myvolume-client-0' came back up; going 
online.
Final graph:
+--+
  1: volume myvolume-client-0
  2: type protocol/client
  3: option opversion 31004
  4: option clnt-lk-version 1
  5: option volfile-checksum 0
  6: option volfile-key myvolume
  7: option client-version 3.10.7
  8: option process-uuid 

Re: [Gluster-users] Failed to get quota limits

2018-02-12 Thread Hari Gowtham
Hi,

Can you provide more information like, the volume configuration, quota.conf
file and the log files.

On Sat, Feb 10, 2018 at 1:05 AM, mabi  wrote:
> Hello,
>
> I am running GlusterFS 3.10.7 and just noticed by doing a "gluster volume
quota  list" that my quotas on that volume are broken. The command
returns no output and no errors but by looking in /var/log/glusterfs.cli I
found the following errors:
>
> [2018-02-09 19:31:24.242324] E
[cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to get
quota limits for 3df709ee-641d-46a2-bd61-889583e3033c
> [2018-02-09 19:31:24.249790] E
[cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to get
quota limits for a27818fe-0248-40fe-bb23-d43d61010478
> [2018-02-09 19:31:24.252378] E
[cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to get
quota limits for daf97388-bcec-4cc0-a8ef-5b93f05b30f6
> [2018-02-09 19:31:24.256775] E
[cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to get
quota limits for 3c768b36-2625-4509-87ef-fe5214cb9b01
> [2018-02-09 19:31:24.257434] E
[cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to get
quota limits for f8cf47d4-4f54-43c5-ab0d-75b45b4677a3
> [2018-02-09 19:31:24.259126] E
[cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to get
quota limits for b4c81a39-2152-45c5-95d3-b796d88226fe
> [2018-02-09 19:31:24.261664] E
[cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to get
quota limits for 16ac4cde-a5d4-451f-adcc-422a542fea24
> [2018-02-09 19:31:24.261719] I [input.c:31:cli_batch] 0-: Exiting with: 0
>
> How can I fix my quota on that volume again? I had around 30 quotas set
on different directories of that volume.
>
> Thanks in advance.
>
> Regards,
> M.
> ___
> Gluster-users mailing list
> Gluster-users@gluster.org
> http://lists.gluster.org/mailman/listinfo/gluster-users



-- 
Regards,
Hari Gowtham.
___
Gluster-users mailing list
Gluster-users@gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users

Re: [Gluster-users] Failed to get quota limits

2018-02-12 Thread mabi
Would anyone be able to help me fix my quotas again?

Thanks
​

 Original Message 
 On February 9, 2018 8:35 PM, mabi  wrote:

>Hello,
>
> I am running GlusterFS 3.10.7 and just noticed by doing a "gluster volume 
> quota  list" that my quotas on that volume are broken. The command 
> returns no output and no errors but by looking in /var/log/glusterfs.cli I 
> found the following errors:
>
> [2018-02-09 19:31:24.242324] E 
> [cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to get 
> quota limits for 3df709ee-641d-46a2-bd61-889583e3033c
> [2018-02-09 19:31:24.249790] E 
> [cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to get 
> quota limits for a27818fe-0248-40fe-bb23-d43d61010478
> [2018-02-09 19:31:24.252378] E 
> [cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to get 
> quota limits for daf97388-bcec-4cc0-a8ef-5b93f05b30f6
> [2018-02-09 19:31:24.256775] E 
> [cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to get 
> quota limits for 3c768b36-2625-4509-87ef-fe5214cb9b01
> [2018-02-09 19:31:24.257434] E 
> [cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to get 
> quota limits for f8cf47d4-4f54-43c5-ab0d-75b45b4677a3
> [2018-02-09 19:31:24.259126] E 
> [cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to get 
> quota limits for b4c81a39-2152-45c5-95d3-b796d88226fe
> [2018-02-09 19:31:24.261664] E 
> [cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to get 
> quota limits for 16ac4cde-a5d4-451f-adcc-422a542fea24
> [2018-02-09 19:31:24.261719] I [input.c:31:cli_batch] 0-: Exiting with: 0
>
> How can I fix my quota on that volume again? I had around 30 quotas set on 
> different directories of that volume.
>
> Thanks in advance.
>
> Regards,
> M.
>
>Gluster-users mailing list
>Gluster-users@gluster.org
>http://lists.gluster.org/mailman/listinfo/gluster-users
>

___
Gluster-users mailing list
Gluster-users@gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users

[Gluster-users] Failed to get quota limits

2018-02-09 Thread mabi
Hello,

I am running GlusterFS 3.10.7 and just noticed by doing a "gluster volume quota 
 list" that my quotas on that volume are broken. The command returns 
no output and no errors but by looking in /var/log/glusterfs.cli I found the 
following errors:

[2018-02-09 19:31:24.242324] E 
[cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to get 
quota limits for 3df709ee-641d-46a2-bd61-889583e3033c
[2018-02-09 19:31:24.249790] E 
[cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to get 
quota limits for a27818fe-0248-40fe-bb23-d43d61010478
[2018-02-09 19:31:24.252378] E 
[cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to get 
quota limits for daf97388-bcec-4cc0-a8ef-5b93f05b30f6
[2018-02-09 19:31:24.256775] E 
[cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to get 
quota limits for 3c768b36-2625-4509-87ef-fe5214cb9b01
[2018-02-09 19:31:24.257434] E 
[cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to get 
quota limits for f8cf47d4-4f54-43c5-ab0d-75b45b4677a3
[2018-02-09 19:31:24.259126] E 
[cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to get 
quota limits for b4c81a39-2152-45c5-95d3-b796d88226fe
[2018-02-09 19:31:24.261664] E 
[cli-cmd-volume.c:1674:cli_cmd_quota_handle_list_all] 0-cli: Failed to get 
quota limits for 16ac4cde-a5d4-451f-adcc-422a542fea24
[2018-02-09 19:31:24.261719] I [input.c:31:cli_batch] 0-: Exiting with: 0

How can I fix my quota on that volume again? I had around 30 quotas set on 
different directories of that volume.

Thanks in advance.

Regards,
M.
___
Gluster-users mailing list
Gluster-users@gluster.org
http://lists.gluster.org/mailman/listinfo/gluster-users