Thanks very much for your feedback Chris.

Renata

On Tue, 1 Dec 2020, Christopher Black wrote:

>We tune vm-related sysctl values on our gpfs clients.
>These are values we use for 256GB+ mem hpc nodes:
>vm.min_free_kbytes=2097152
>vm.dirty_bytes = 3435973836
>vm.dirty_background_bytes = 1717986918
>
>The vm.dirty parameters are to prevent NFS from buffering huge amounts of 
>writes and then pushing them over the network all at once flooding out gpfs 
>traffic.
>
>I'd also recommend checking client gpfs parameters pagepool and/or 
>pagepoolMaxPhysMemPct to ensure you have a reasonable and understood limit for 
>how much memory mmfsd will use.
>
>Best,
>Chris
>
>On 12/1/20, 1:32 PM, "[email protected] on behalf of 
>Renata Maria Dart" <[email protected] on behalf of 
>[email protected]> wrote:
>
>    Hi, some of our gpfs clients will get stale file handles for gpfs
>    mounts and it seems to be related to memory depletion.  Even after the
>    memory is freed though gpfs will continue be unavailable and df will
>    hang.  I have read about setting vm.min_free_kbytes as a possible fix
>    for this, but wasn't sure if it was meant for a gpfs server or if a
>    gpfs client would also benefit, and what value should be set.
>
>    Thanks for any insights,
>
>    Renata
>
>
>    _______________________________________________
>    gpfsug-discuss mailing list
>    gpfsug-discuss at spectrumscale.org
>    
> https://urldefense.com/v3/__http://gpfsug.org/mailman/listinfo/gpfsug-discuss__;!!C6sPl7C9qQ!H08HlNmBIkQRBOJKSHohzKHL6r39gAhQ3XTTczWoSmvffRFmQMcpJo8OyjMP7j-g$
>
>________________________________
>
>This message is for the recipient?s use only, and may contain confidential, 
>privileged or protected information. Any unauthorized use or dissemination of 
>this communication is prohibited. If you received this message in error, 
>please immediately notify the sender and destroy all copies of this message. 
>The recipient should check this email and any attachments for the presence of 
>viruses, as we accept no liability for any damage caused by any virus 
>transmitted by this email.
>

_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss

Reply via email to