Also, quotas on the OSTS don’t add up to near 3 million files either:

[root@lustreclient scratch]# ssh ossnode0 lfs quota -g somegroup -I 0 /lustre1
Disk quotas for grp somegroup (gid 9544):
     Filesystem  kbytes   quota   limit   grace   files   quota   limit   grace
                1394853459       0 1913344192       -  132863       0       0   
    -
[root@lustreclient scratch]# ssh ossnode0 lfs quota -g somegroup -I 1 /lustre1
Disk quotas for grp somegroup (gid 9544):
     Filesystem  kbytes   quota   limit   grace   files   quota   limit   grace
                1411579601       0 1963246413       -  120643       0       0   
    -
[root@lustreclient scratch]# ssh ossnode1 lfs quota -g somegroup -I 2 /lustre1
Disk quotas for grp somegroup (gid 9544):
     Filesystem  kbytes   quota   limit   grace   files   quota   limit   grace
                1416507527       0 1789950778       -  190687       0       0   
    -
[root@lustreclient scratch]# ssh ossnode1 lfs quota -g somegroup -I 3 /lustre1
Disk quotas for grp somegroup (gid 9544):
     Filesystem  kbytes   quota   limit   grace   files   quota   limit   grace
                1636465724       0 1926578117       -  195034       0       0   
    -
[root@lustreclient scratch]# ssh ossnode2 lfs quota -g somegroup -I 4 /lustre1
Disk quotas for grp somegroup (gid 9544):
     Filesystem  kbytes   quota   limit   grace   files   quota   limit   grace
                2202272244       0 3020159313       -  185097       0       0   
    -
[root@lustreclient scratch]# ssh ossnode2 lfs quota -g somegroup -I 5 /lustre1
Disk quotas for grp somegroup (gid 9544):
     Filesystem  kbytes   quota   limit   grace   files   quota   limit   grace
                1324770165       0 1371244768       -  145347       0       0   
    -
[root@lustreclient scratch]# ssh ossnode3 lfs quota -g somegroup -I 6 /lustre1
Disk quotas for grp somegroup (gid 9544):
     Filesystem  kbytes   quota   limit   grace   files   quota   limit   grace
                2892027349       0 3221225472       -  169386       0       0   
    -
[root@lustreclient scratch]# ssh ossnode3 lfs quota -g somegroup -I 7 /lustre1
Disk quotas for grp somegroup (gid 9544):
     Filesystem  kbytes   quota   limit   grace   files   quota   limit   grace
                2076201636       0 2474853207       -  171552       0       0   
    -


—
Dan Szkola
FNAL

> On Oct 4, 2023, at 8:45 AM, Daniel Szkola via lustre-discuss 
> <lustre-discuss@lists.lustre.org> wrote:
> 
> No combination of ossnodek runs has helped with this.
> 
> Again, robinhood shows 1796104 files for the group, an 'lfs find -G gid' 
> found 1796104 files as well.
> 
> So why is the quota command showing over 3 million inodes used?
> 
> There must be a way to force it to recount or clear all stale quota data and 
> have it regenerate it?
> 
> Anyone?
> 
> —
> Dan Szkola
> FNAL
> 
> 
>> On Sep 27, 2023, at 9:42 AM, Daniel Szkola via lustre-discuss 
>> <lustre-discuss@lists.lustre.org> wrote:
>> 
>> We have a lustre filesystem that we just upgraded to 2.15.3, however this 
>> problem has been going on for some time.
>> 
>> The quota command shows this:
>> 
>> Disk quotas for grp somegroup (gid 9544):
>>    Filesystem    used   quota   limit   grace   files   quota   limit   grace
>>      /lustre1  13.38T     40T     45T       - 3136761* 2621440 3670016 
>> expired
>> 
>> The group is not using nearly that many files. We have robinhood installed 
>> and it show this:
>> 
>> Using config file '/etc/robinhood.d/lustre1.conf'.
>>    group,     type,      count,     volume,   spc_used,   avg_size
>> somegroup,   symlink,      59071,    5.12 MB,  103.16 MB,         91
>> somegroup,       dir,     426619,    5.24 GB,    5.24 GB,   12.87 KB
>> somegroup,      file,    1310414,   16.24 TB,   13.37 TB,   13.00 MB
>> 
>> Total: 1796104 entries, volume: 17866508365925 bytes (16.25 TB), space used: 
>> 14704924899840 bytes (13.37 TB)
>> 
>> Any ideas what is wrong here?
>> 
>> —
>> Dan Szkola
>> FNAL
>> _______________________________________________
>> lustre-discuss mailing list
>> lustre-discuss@lists.lustre.org
>> https://urldefense.proofpoint.com/v2/url?u=http-3A__lists.lustre.org_listinfo.cgi_lustre-2Ddiscuss-2Dlustre.org&d=DwIGaQ&c=gRgGjJ3BkIsb5y6s49QqsA&r=e9DXjyTaQ786Tg7WH7oIVaQOA1YDRqyxHOUaYU2_LQw&m=Nk1MkSBTpT-KnrXzEvOOP5tZoVAKyHfPvB-o8_OhewuwHF6S0KelH_WPMLq8IRnR&s=JzAV0C2_CqaDUOG0wZr0mx5tiblBde6ZRUuIHZ2n9DI&e=
>>  
> 
> _______________________________________________
> lustre-discuss mailing list
> lustre-discuss@lists.lustre.org
> https://urldefense.proofpoint.com/v2/url?u=http-3A__lists.lustre.org_listinfo.cgi_lustre-2Ddiscuss-2Dlustre.org&d=DwIGaQ&c=gRgGjJ3BkIsb5y6s49QqsA&r=e9DXjyTaQ786Tg7WH7oIVaQOA1YDRqyxHOUaYU2_LQw&m=k8TeSgok6MIb-uQMJaquDJS0FQPt0RQxysFNe4d7Rp5TMqGtcqdlezA_TZNuoTJS&s=SRDKhUKQgMW9_OohjyrkzKNYbzTw_M5BJk-bmEi_6w4&e=
>  

_______________________________________________
lustre-discuss mailing list
lustre-discuss@lists.lustre.org
http://lists.lustre.org/listinfo.cgi/lustre-discuss-lustre.org
  • [... Daniel Szkola via lustre-discuss
    • ... Daniel Szkola via lustre-discuss
      • ... Mark Dixon via lustre-discuss
        • ... Daniel Szkola via lustre-discuss
          • ... Mark Dixon via lustre-discuss
      • ... Daniel Szkola via lustre-discuss
        • ... Daniel Szkola via lustre-discuss
          • ... Andreas Dilger via lustre-discuss
            • ... Daniel Szkola via lustre-discuss
              • ... Daniel Szkola via lustre-discuss
                • ... Andreas Dilger via lustre-discuss
                • ... Daniel Szkola via lustre-discuss
                • ... Vicker, Darby J. (JSC-EG111)[Jacobs Technology, Inc.] via lustre-discuss
                • ... Daniel Szkola via lustre-discuss
                • ... Andreas Dilger via lustre-discuss
                • ... Daniel Szkola via lustre-discuss

Reply via email to