Maybe it is a stupid question, but... why it is not tested and enabled by default at configure time since it is part of the kernel?

Eric


On 02/19/2015 03:53 PM, Nathan Hjelm wrote:
Great! I will add an MCA variable to force CMA and also enable it if 1)
no yama and 2) no PR_SET_PTRACER.

You might also look at using xpmem. You can find a version that supports
3.x @ https://github.com/hjelmn/xpmem . It is a kernel module +
userspace library that can be used by vader as a single-copy mechanism.

In benchmarks it performs better than CMA but it may or may not perform
better with a real application.

See:

http://blogs.cisco.com/performance/the-vader-shared-memory-transport-in-open-mpi-now-featuring-3-flavors-of-zero-copy

-Nathan

On Thu, Feb 19, 2015 at 03:32:43PM -0500, Eric Chamberland wrote:
On 02/19/2015 02:58 PM, Nathan Hjelm wrote:
On Thu, Feb 19, 2015 at 12:16:49PM -0500, Eric Chamberland wrote:
On 02/19/2015 11:56 AM, Nathan Hjelm wrote:
If you have yama installed you can try:
Nope, I do not have it installed... is it absolutely necessary? (and would
it change something when it fails when I am root?)

Other question: In addition to "--with-cma" configure flag, do we have to
pass any options to "mpicc" when compiling/linking an mpi application to use
cma?
No. CMA should work out of the box. You appear to have a setup I haven't
yet tested. It doesn't have yama nor does it have the PR_SET_PTRACER
prctl. Its quite possible there are no restriction on ptrace in this
setup. Can you try changing the following line at
opal/mca/btl/vader/btl_vader_component.c:370 from:

bool cma_happy = false;

to

bool cma_happy = true;

ok! (as of the officiel release, this is line 386.)

and let me know if that works. If it does I will update vader to allow
CMA in this configuration.
Yep!  It now works perfectly.  Testing with
https://computing.llnl.gov/tutorials/mpi/samples/C/mpi_bandwidth.c, on my
own computer (dual Xeon), I have this:

Without CMA:

***Message size:  1000000 *** best  /  avg  / worst (MB/sec)
    task pair:    0 -    1:    8363.52 / 7946.77 / 5391.14

with CMA:
    task pair:    0 -    1:    9137.92 / 8955.98 / 7489.83

Great!

Now I have to bench my real application... ;-)

Thanks!

Eric

_______________________________________________
users mailing list
us...@open-mpi.org
Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
Link to this post: 
http://www.open-mpi.org/community/lists/users/2015/02/26355.php

Reply via email to