Hello Olaf, hello Marc,we in Jülich are in the middle of migrating/copying all our old filesystems which were created with filesystem version: 13.23 (3.5.0.7) to new filesystems created with GPFS 5.0.1.
We move to new filesystems mainly for two reasons: 1. We want to use the new increased number of subblocks. 2. We have to change our quota from normal "group-quota per filesystem" to "fileset-quota".
The idea is to create a separate fileset for each group/project. For the users the quota-computation should be much more transparent. From now on all data which is stored inside of their directory (fileset) counts for their quota independent of the ownership.
Right now we have round about 900 groups which means we will create round about 900 filesets per filesystem. In one filesystem we will have about 400million inodes (with rising tendency).
This filesystem we will back up with "mmbackup" so we talked with Dominic Mueller-Wicke and he recommended us to use independent filesets. Because then the policy-runs can be parallelized and we can increase the backup performance. We belive that we require these parallelized policies run to meet our backup performance targets.
But there are even more features we enable by using independet filesets. E.g. "Fileset level snapshots" and "user and group quotas inside of a fileset".
I did not know about performance issues regarding independent filesets... Can you give us some more information about this?
All in all we are strongly supporting the idea of increasing this limit.Do I understand correctly that by opening a PMR IBM allows to increase this limit on special sides? I would rather like to increase the limit and make it official public available and supported.
Regards, Martin Am 10.08.2018 um 14:51 schrieb Olaf Weiser:
Hallo Stephan,the limit is not a hard coded limit - technically spoken, you can raise it easily.But as always, it is a question of test 'n support ..I've seen customer cases, where the use of much smaller amount of independent filesets generates a lot performance issues, hangs ... at least noise and partial trouble .. it might be not the case with your specific workload, because due to the fact, that you 're running already close to 1000 ...I suspect , this number of 1000 file sets - at the time of introducing it - was as also just that one had to pick a number...... turns out.. that a general commitment to support > 1000 ind.fileset is more or less hard.. because what uses cases should we test / support I think , there might be a good chance for you , that for your specific workload, one would allow and support more than 1000do you still have a PMR for your side for this ? - if not - I know .. open PMRs is an additional ...but could you please ..then we can decide .. if raising the limit is an option for you .. Mit freundlichen Grüßen / Kind regards Olaf WeiserEMEA Storage Competence Center Mainz, German / IBM Systems, Storage Platform,------------------------------------------------------------------------------------------------------------------------------------------- IBM Deutschland IBM Allee 1 71139 Ehningen Phone: +49-170-579-44-66 E-Mail: [email protected] ------------------------------------------------------------------------------------------------------------------------------------------- IBM Deutschland GmbH / Vorsitzender des Aufsichtsrats: Martin JetterGeschäftsführung: Martina Koederitz (Vorsitzende), Susanne Peter, Norbert Janzen, Dr. Christian Keller, Ivo Koerner, Markus Koerner Sitz der Gesellschaft: Ehningen / Registergericht: Amtsgericht Stuttgart, HRB 14562 / WEEE-Reg.-Nr. DE 99369940From: "Peinkofer, Stephan" <[email protected]> To: gpfsug main discussion list <[email protected]>Cc: Doris Franke <[email protected]>, Uwe Tron <[email protected]>, Dorian Krause <[email protected]>Date: 08/10/2018 01:29 PM Subject: [gpfsug-discuss] GPFS Independent Fileset Limit Sent by: [email protected] ------------------------------------------------------------------------ Dear IBM and GPFS List,we at the Leibniz Supercomputing Centre and our GCS Partners from the Jülich Supercomputing Centre will soon be hitting the current Independent Fileset Limit of 1000 on a number of our GPFS Filesystems.There are also a number of RFEs from other users open, that target this limitation:_https://www.ibm.com/developerworks/rfe/execute?use_case=viewRfe&CR_ID=56780_ https://www.ibm.com/developerworks/rfe/execute?use_case=viewRfe&CR_ID=120534_ __https://www.ibm.com/developerworks/rfe/execute?use_case=viewRfe&CR_ID=106530_ _https://www.ibm.com/developerworks/rfe/execute?use_case=viewRfe&CR_ID=85282_I know GPFS Development was very busy fulfilling the CORAL requirements but maybe now there is again some time to improve something else.If there are any other users on the list that are approaching the current limitation in independent filesets, please take some time and vote for the RFEs above.Many thanks in advance and have a nice weekend. Best Regards, Stephan Peinkofer _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss
smime.p7s
Description: S/MIME Cryptographic Signature
_______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at spectrumscale.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss
