Re: [OMPI devel] sm BTL performace of the openmpi-2.0.0

2016-08-07 Thread Nathan Hjelm
Ok, looking at the code I can’t see what the issue is. I will run this on our mlx cluster tomorrow. Should have a working patch then. -Nathan > On Aug 7, 2016, at 9:58 PM, tmish...@jcity.maeda.co.jp wrote: > > Hi, here is the gdb output for additional information: > > (It might be inexact, bec

Re: [OMPI devel] sm BTL performace of the openmpi-2.0.0

2016-08-07 Thread tmishima
Hi, here is the gdb output for additional information: (It might be inexact, because I built openmpi-2.0.0 without debug option) Core was generated by `osu_bw'. Program terminated with signal 11, Segmentation fault. #0 0x0031d9008806 in ?? () from /lib64/libgcc_s.so.1 (gdb) where #0 0x0

Re: [OMPI devel] sm BTL performace of the openmpi-2.0.0

2016-08-07 Thread tmishima
Hi, it caused segfault as below: [manage.cluster:25436] MCW rank 0 bound to socket 0[core 0[hwt 0]], socket 0[core 1[hwt 0]], socket 0[core 2[hwt 0]], so cket 0[core 3[hwt 0]], socket 0[core 4[hwt 0]], socket 0[core 5[hwt 0]]: [B/B/B/B/B/B][./././././.] [manage.cluster:25436] MCW rank 1 bound to s

Re: [OMPI devel] sm BTL performace of the openmpi-2.0.0

2016-08-07 Thread Nathan Hjelm
This patch also modifies the put path. Let me know if this works: diff --git a/ompi/mca/pml/ob1/pml_ob1_rdma.c b/ompi/mca/pml/ob1/pml_ob1_rdma.c index 888e126..a3ec6f8 100644 --- a/ompi/mca/pml/ob1/pml_ob1_rdma.c +++ b/ompi/mca/pml/ob1/pml_ob1_rdma.c @@ -42,6 +42,7 @@ size_t mca_pml_ob1_rdma_btls(

Re: [OMPI devel] sm BTL performace of the openmpi-2.0.0

2016-08-07 Thread Nathan Hjelm
Looks like the put path probably needs a similar patch. Will send another patch soon. > On Aug 7, 2016, at 6:01 PM, tmish...@jcity.maeda.co.jp wrote: > > Hi, > > I applied the patch to the file "pml_ob1_rdma.c" and ran osu_bw again. > Then, I still see the bad performance for larger size(>=209

Re: [OMPI devel] sm BTL performace of the openmpi-2.0.0

2016-08-07 Thread tmishima
Hi, I applied the patch to the file "pml_ob1_rdma.c" and ran osu_bw again. Then, I still see the bad performance for larger size(>=2097152 ). [mishima@manage OMB-3.1.1-openmpi2.0.0]$ mpirun -np 2 -report-bindings osu_bw [manage.cluster:27444] MCW rank 0 bound to socket 0[core 0[hwt 0]], socket 0[

[OMPI devel] PMIx Language Bindings

2016-08-07 Thread r...@open-mpi.org
Hi folks I’m looking for someone(s) interested in writing some simple language bindings (e.g., Python, Java, Fortran) for the PMIx library (which is written in C). There aren’t a lot of APIs, so I don’t envision this as being a monstrous effort. Please let me know if you have any interest - an