For „1“ we use the quorum node to do “start disk” or “restripe file system” 
(quorum node without disks).
For “2” we use kernel NFS with cNFS

I used the command “cnfsNFSDprocs 64” to set the NFS threads. Is this correct?

gpfs01:~ # cat /proc/fs/nfsd/threads
64

I will verify the settings in our lab, will use the following configuration:
mmchconfig worker1Threads=128
mmchconfig prefetchThreads=128
mmchconfig nsdMaxWorkerThreads=128
mmchconfig cnfsNFSDprocs=256

daniel



Von: 
[email protected]<mailto:[email protected]> 
[mailto:[email protected]]<mailto:[mailto:[email protected]]>
 Im Auftrag von Sven Oehme
Gesendet: Samstag, 4. Juli 2015 00:49
An: gpfsug main discussion list
Betreff: Re: [gpfsug-discuss] GPFS 4.1.1 without QoS for mmrestripefs?


this triggers a few questions

1. have you tried running it only on a node that doesn't serve NFS data ?
2. what NFS stack are you using ? is this the kernel NFS Server as part of 
linux means you use cNFS ?

if the answer to 2 is yes, have you adjusted the nfsd threads in 
/etc/sysconfig/nfs ? the default is only 8 and if you run with the default you 
have a very low number of threads from the outside competing with a larger 
number of threads doing restripe, increasing the nfsd threads could help. you 
could also reduce the number of internal restripe threads to try out if that 
helps mitigating the impact.

to try an extreme low value set the following :

mmchconfig pitWorkerThreadsPerNode=1 -i

and retry the restripe again, to reset it back to default run

mmchconfig pitWorkerThreadsPerNode=DEFAULT -i

sven

------------------------------------------
Sven Oehme
Scalable Storage Research
email: [email protected]<mailto:[email protected]>
Phone: +1 (408) 824-8904
IBM Almaden Research Lab
------------------------------------------

[Beschreibung: Inactive hide details for Daniel Vogel ---07/02/2015 12:12:46 
AM---Sven, Yes I agree, but “using –N” to reduce the load help]Daniel Vogel 
---07/02/2015 12:12:46 AM---Sven, Yes I agree, but “using –N” to reduce the 
load helps not really. If I use NFS, for example, as

From: Daniel Vogel 
<[email protected]<mailto:[email protected]>>
To: "'gpfsug main discussion list'" 
<[email protected]<mailto:[email protected]>>
Date: 07/02/2015 12:12 AM
Subject: Re: [gpfsug-discuss] GPFS 4.1.1 without QoS for mmrestripefs?
Sent by: 
[email protected]<mailto:[email protected]>

________________________________



Sven,

Yes I agree, but “using –N” to reduce the load helps not really. If I use NFS, 
for example, as a ESX data store, ESX I/O latency for NFS goes very high, the 
VM’s hangs. By the way I use SSD PCIe cards, perfect “mirror speed” but slow 
I/O on NFS.
The GPFS cluster concept I use are different than GSS or traditional FC (shared 
storage). I use shared nothing with IB (no FPO), many GPFS nodes with NSD’s. I 
know the need to resync the FS with mmchdisk / mmrestripe will happen more 
often. The only one feature will help is QoS for the GPFS admin jobs. I hope we 
are not fare away from this.

Thanks,
Daniel


Von: 
[email protected]<mailto:[email protected]> 
[mailto:[email protected]] Im Auftrag von Sven Oehme
Gesendet: Mittwoch, 1. Juli 2015 16:21
An: gpfsug main discussion list
Betreff: Re: [gpfsug-discuss] GPFS 4.1.1 without QoS for mmrestripefs?

Daniel,

as you know, we can't discuss future / confidential items on a mailing list.
what i presented as an outlook to future releases hasn't changed from a 
technical standpoint, we just can't share a release date until we announce it 
official.
there are multiple ways today to limit the impact on restripe and other tasks, 
the best way to do this is to run the task ( using -N) on a node (or very small 
number of nodes) that has no performance critical role. while this is not 
perfect, it should limit the impact significantly. .

sven

------------------------------------------
Sven Oehme
Scalable Storage Research
email: [email protected]<mailto:[email protected]>
Phone: +1 (408) 824-8904
IBM Almaden Research Lab
------------------------------------------

[Beschreibung: Inactive hide details for Daniel Vogel ---07/01/2015 03:29:11 
AM---Hi Years ago, IBM made some plan to do a implementation "QoS]Daniel Vogel 
---07/01/2015 03:29:11 AM---Hi Years ago, IBM made some plan to do a 
implementation "QoS for mmrestripefs, mmdeldisk...". If a "

From: Daniel Vogel 
<[email protected]<mailto:[email protected]>>
To: "'[email protected]'" 
<[email protected]<mailto:[email protected]>>
Date: 07/01/2015 03:29 AM
Subject: [gpfsug-discuss] GPFS 4.1.1 without QoS for mmrestripefs?
Sent by: 
[email protected]<mailto:[email protected]>

________________________________




Hi

Years ago, IBM made some plan to do a implementation “QoS for mmrestripefs, 
mmdeldisk…”. If a “mmfsrestripe” is running, very poor performance for NFS 
access.
I opened a PMR to ask for QoS in version 4.1.1 (Spectrum Scale).

PMR 61309,113,848:
I discussed the question of QOS with the development team. These
command changes that were noticed are not meant to be used as GA code
which is why they are not documented. I cannot provide any further
information from the support perspective.


Anybody knows about QoS? The last hope was at “GPFS Workshop Stuttgart März 
2015” with Sven Oehme as speaker.

Daniel Vogel
IT Consultant

ABC SYSTEMS AG
Hauptsitz Zürich
Rütistrasse 28
CH - 8952 Schlieren
T +41 43 433 6 433
D +41 43 433 6 467
http://www.abcsystems.ch<http://www.abcsystems.ch/>

ABC - Always Better Concepts. Approved By Customers since 1981.
_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at gpfsug.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss

_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at gpfsug.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss


_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at gpfsug.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss

Reply via email to