Hi Lohit, Nathan

Would you be willing to share some more details about your setup? We are just 
getting started here and I would like to hear about what your configuration 
looks like. Direct email to me is fine, thanks.

Bob Oesterlin
Sr Principal Storage Engineer, Nuance


From: <[email protected]> on behalf of 
"[email protected]" <[email protected]>
Reply-To: gpfsug main discussion list <[email protected]>
Date: Thursday, April 26, 2018 at 9:45 AM
To: gpfsug main discussion list <[email protected]>
Subject: [EXTERNAL] Re: [gpfsug-discuss] Singularity + GPFS

We do run Singularity + GPFS, on our production HPC clusters.
Most of the time things are fine without any issues.

However, i do see a significant performance loss when running some applications 
on singularity containers with GPFS.

As of now, the applications that have severe performance issues with 
singularity on GPFS - seem to be because of “mmap io”. (Deep learning 
applications)
When i run the same application on bare metal, they seem to have a huge 
difference in GPFS IO when compared to running on singularity containers.
I am yet to raise a PMR about this with IBM.
I have not seen performance degradation for any other kind of IO, but i am not 
sure.

Regards,
Lohit

On Apr 26, 2018, 10:35 AM -0400, Nathan Harper <[email protected]>, 
wrote:

We are running on a test system at the moment, and haven't run into any issues 
yet, but so far it's only been 'hello world' and running FIO.

I'm interested to hear about experience with MPI-IO within Singularity.

On 26 April 2018 at 15:20, Oesterlin, Robert 
<[email protected]<mailto:[email protected]>> wrote:
Anyone (including IBM) doing any work in this area? I would appreciate hearing 
from you.

Bob Oesterlin
Sr Principal Storage Engineer, Nuance


_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at 
spectrumscale.org<https://urldefense.proofpoint.com/v2/url?u=http-3A__spectrumscale.org&d=DwMFaQ&c=djjh8EKwHtOepW4Bjau0lKhLlu-DxM1dlgP0rrLsOzY&r=LPDewt1Z4o9eKc86MXmhqX-45Cz1yz1ylYELF9olLKU&m=_6XPyvEYCL-PnMjKsOcjeRy1H19CJ1ujTFPrAEwWnLo&s=OTGc6UhyurUT-JrFZxFuLp_VKCTGGknEbzjMSKUa-QQ&e=>
http://gpfsug.org/mailman/listinfo/gpfsug-discuss<https://urldefense.proofpoint.com/v2/url?u=http-3A__gpfsug.org_mailman_listinfo_gpfsug-2Ddiscuss&d=DwMFaQ&c=djjh8EKwHtOepW4Bjau0lKhLlu-DxM1dlgP0rrLsOzY&r=LPDewt1Z4o9eKc86MXmhqX-45Cz1yz1ylYELF9olLKU&m=_6XPyvEYCL-PnMjKsOcjeRy1H19CJ1ujTFPrAEwWnLo&s=hPDughTsSwQUdR8KHv4gxXMumoGw3jXLgGGXG1MQrJM&e=>



--
Nathan Harper // IT Systems Lead

[Image removed by 
sender.]<https://urldefense.proofpoint.com/v2/url?u=https-3A__cfms.org.uk_news-2Devents_events_2018_july_farnborough-2Dinternational-2Dairshow-2D2018_&d=DwMFaQ&c=djjh8EKwHtOepW4Bjau0lKhLlu-DxM1dlgP0rrLsOzY&r=LPDewt1Z4o9eKc86MXmhqX-45Cz1yz1ylYELF9olLKU&m=_6XPyvEYCL-PnMjKsOcjeRy1H19CJ1ujTFPrAEwWnLo&s=SAiH6rxM-6WYWwUBRecbBS9Yr7f4adg4GizgVJXIAqI&e=>

e: [email protected]<mailto:[email protected]>   t: 0117 906 
1104  m:  0787 551 0891  w: 
www.cfms.org.uk<https://urldefense.proofpoint.com/v2/url?u=http-3A__www.cfms.org.uk_&d=DwMFaQ&c=djjh8EKwHtOepW4Bjau0lKhLlu-DxM1dlgP0rrLsOzY&r=LPDewt1Z4o9eKc86MXmhqX-45Cz1yz1ylYELF9olLKU&m=_6XPyvEYCL-PnMjKsOcjeRy1H19CJ1ujTFPrAEwWnLo&s=KAbz_Md-BsCIrN6LDOb6Yc7sriwSq5IeO4aaxRGQfjc&e=>
CFMS Services Ltd // Bristol & Bath Science Park // Dirac Crescent // Emersons 
Green // Bristol // BS16 7FR
[Image removed by sender.]
CFMS Services Ltd is registered in England and Wales No 05742022 - a subsidiary 
of CFMS Ltd
CFMS Services Ltd registered office // 43 Queens Square // Bristol // BS1 4QP
_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss
_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss

Reply via email to