Re: [hwloc-users] AMD EPYC topology

2017-12-29 Thread Matthew Scutter
Following up on this, Indeed with a recent kernel the error message goes away. The poor performance stays though (a few percent difference between 4.13 and 4.15rc5), and I'm at a loss as to whether it's related to MPI or not. I see oddities such as locking the job to the first 12 cores yield 100%

Re: [hwloc-users] AMD EPYC topology

2017-12-29 Thread Brice Goglin
Le 29/12/2017 à 23:15, Bill Broadley a écrit : > > > Very interesting, I was running parallel finite element code and was seeing > great performance compared to Intel in most cases, but on larger runs it was > 20x > slower. This would explain it. > > Do you know which commit, or anything else

Re: [hwloc-users] AMD EPYC topology

2017-12-29 Thread Bill Broadley
Very interesting, I was running parallel finite element code and was seeing great performance compared to Intel in most cases, but on larger runs it was 20x slower. This would explain it. Do you know which commit, or anything else that might help find any related discussion? I tried a few

Re: [hwloc-users] AMD EPYC topology

2017-12-24 Thread Brice Goglin
Hello Make sure you use a very recent Linux kernel. There was a bug regarding L3 caches on 24-core Epyc processors which has been fixed in 4.14 and backported in 4.13.x (and maybe in distro kernels too). However, that would likely not cause huge performance difference unless your application

[hwloc-users] AMD EPYC topology

2017-12-24 Thread Matthew Scutter
I'm getting poor performance on OpenMPI tasks on a new AMD 7401P EPYC server. I suspect hwloc providing a poor topology may have something to do with it as I receive this warning below when creating a job. Requested data files available at http://static.skysight.io/out.tgz Cheers, Matthew