I have already sent it On Thu, May 28, 2015 at 10:21 AM.
28 мая 2015 г. 20:07 пользователь Mike Dubman написал:it is fine to recompile OMPI from HPCx to apply site default (choice of job scheduler for example, OMPI from HPCX compiled with ssh support only, etc.).If ssh
015 at 3:39 PM, Mike Dubman <mi...@dev.mellanox.co.il
> <https://e.mail.ru/compose/?mailto=mailto%3ami...@dev.mellanox.co.il>>
> wrote:
>
> Alina - could you please take a look?
> Thx
>
>
> -- Forwarded message --
> From: *Timur Ismagilov* <tisma
R/debug/lib/libmxm.so -x MXM_LOG_LEVEL=data"
>>>
>>>Also, could you please attach the entire output of
>>>"$HPCX_MPI_DIR/bin/ompi_info -a"
>>>
>>>Thank you,
>>>Alina.
>>>
>>>On Tue, May 26, 2015 at 3:39 PM, M
t;>Alina - could you please take a look?
>>Thx
>>
>>
>>-- Forwarded message ------
>>From: Timur Ismagilov < tismagi...@mail.ru >
>>Date: Tue, May 26, 2015 at 12:40 PM
>>Subject: Re[12]: [OMPI users] MXM problem
>>To: Open MPI Users < u
It does not work for single node:
1) host: $ $HPCX_MPI_DIR/bin/mpirun -x MXM_IB_PORTS=mlx4_0:1 -x
MXM_SHM_KCOPY_MODE=off -host node5 -mca pml yalla -x MXM_TLS=ud,self,shm
--prefix $HPCX_MPI_DIR -mca plm_base_verbose 5 -mca oob_base_verbose 10 -mca
rml_base_verbose 10 --debug-daemons -np 1
1. mxm_perf_test - OK.
2. no_tree_spawn - OK.
3. ompi yalla and "--mca pml cm --mca mtl mxm" still does not work (I use
prebuild ompi-1.8.5 from hpcx-v1.3.330)
3.a) host:$ $HPCX_MPI_DIR/bin/mpirun -x MXM_IB_PORTS=mlx4_0:1 -x
MXM_SHM_KCOPY_MODE=off -host node5,node153 --mca pml cm --mca mtl
scif is a OFA device from Intel.
can you please select export MXM_IB_PORTS=mlx4_0:1 explicitly and retry
On Mon, May 25, 2015 at 8:26 PM, Timur Ismagilov wrote:
> Hi, Mike,
> that is what i have:
>
> $ echo $LD_LIBRARY_PATH | tr ":" "\n"
>
Hi, Mike,
that is what i have:
$ echo $LD_LIBRARY_PATH | tr ":" "\n"
/gpfs/NETHOME/oivt1/nicevt/itf/sources/hpcx-v1.3.0-327-icc-OFED-1.5.3-redhat6.2/fca/lib
/gpfs/NETHOME/oivt1/nicevt/itf/sources/hpcx-v1.3.0-327-icc-OFED-1.5.3-redhat6.2/hcoll/lib
Hi Timur,
seems that yalla component was not found in your OMPI tree.
can it be that your mpirun is not from hpcx? Can you please check
LD_LIBRARY_PATH,PATH, LD_PRELOAD and OPAL_PREFIX that it is pointing to the
right mpirun?
Also, could you please check that yalla is present in the ompi_info -l
I can password-less ssh to all nodes:
base$ ssh node1
node1$ssh node2
Last login: Mon May 25 18:41:23
node2$ssh node3
Last login: Mon May 25 16:25:01
node3$ssh node4
Last login: Mon May 25 16:27:04
node4$
Is this correct?
In ompi-1.9 i do not have no-tree-spawn problem.
Понедельник, 25 мая
I can’t speak to the mxm problem, but the no-tree-spawn issue indicates that
you don’t have password-less ssh authorized between the compute nodes
> On May 25, 2015, at 8:55 AM, Timur Ismagilov wrote:
>
> Hello!
>
> I use ompi-v1.8.4 from
Hello!
I use ompi-v1.8.4 from hpcx-v1.3.0-327-icc-OFED-1.5.3-redhat6.2;
OFED-1.5.4.1;
CentOS release 6.2;
infiniband 4x FDR
I have two problems:
1. I can not use mxm :
1.a) $mpirun --mca pml cm --mca mtl mxm -host node5,node14,node28,node29 -mca
plm_rsh_no_tree_spawn 1 -np 4 ./hello
12 matches
Mail list logo