Re: [OMPI users] anybody tried OMPI with gpudirect?

2011-02-28 Thread Pak Lui
Actually, since that GPUDirect is not yet officially released, but you may want to contact h...@mellanox.com to get the needed info and when the drivers will be released. Thanks! - Pak -Original Message- From: users-boun...@open-mpi.org [mailto:users-boun...@open-mpi.org] On Behalf

Re: [OMPI users] anybody tried OMPI with gpudirect?

2011-02-28 Thread Pak Lui
Hi Brice, You will need the MLNX_OFED with the GPUDirect support in order to work. I will check to there's a release of it that supports SLES and let you know. [pak@maia001 ~]$ /sbin/modinfo ib_core filename:

Re: [OMPI users] anybody tried OMPI with gpudirect?

2011-02-28 Thread Rolf vandeVaart
-Original Message- From: users-boun...@open-mpi.org [mailto:users-boun...@open-mpi.org] On Behalf Of Brice Goglin Sent: Monday, February 28, 2011 2:14 PM To: Open MPI Users Subject: Re: [OMPI users] anybody tried OMPI with gpudirect? Le 28/02/2011 19:49, Rolf vandeVaart a écrit : > For

Re: [OMPI users] anybody tried OMPI with gpudirect?

2011-02-28 Thread Brice Goglin
Le 28/02/2011 19:49, Rolf vandeVaart a écrit : > For the GPU Direct to work with Infiniband, you need to get some updated OFED > bits from your Infiniband vendor. > > In terms of checking the driver updates, you can do a grep on the string > get_driver_pages in the file/proc/kallsyms. If it is

Re: [OMPI users] anybody tried OMPI with gpudirect?

2011-02-28 Thread Rolf vandeVaart
For the GPU Direct to work with Infiniband, you need to get some updated OFED bits from your Infiniband vendor. In terms of checking the driver updates, you can do a grep on the string get_driver_pages in the file/proc/kallsyms. If it is there, then the Linux kernel is updated correctly.

[OMPI users] RDMACM Differences

2011-02-28 Thread Jagga Soorma
Hello, I am running into the following issue while trying to run osu_latency: -- -bash-3.2$ mpiexec --mca btl openib,self -mca btl_openib_warn_default_gid_ prefix 0 -np 2 --hostfile mpihosts /home/jagga/osu-micro-benchmarks-3.3/openmpi/ofed-1.5.2/bin/osu_latency # OSU MPI Latency Test v3.3 #

Re: [OMPI users] mpirun error: "error while loading shared libraries: libopen-rte.so.0: cannot open shared object file:"

2011-02-28 Thread Jeff Squyres
More specifically -- ensure that LD_LIBRARY_PATH is set properly *on all nodes where you are running Open MPI processes*. For example, if you're using a hostfile to launch across multiple machines, ensure that your shell startup files (e.g., .bashrc) are setup to set your LD_LIBRARY_PATH

Re: [OMPI users] anybody tried OMPI with gpudirect?

2011-02-28 Thread Brice Goglin
Le 28/02/2011 17:30, Rolf vandeVaart a écrit : > Hi Brice: > Yes, I have tired OMPI 1.5 with gpudirect and it worked for me. You > definitely need the patch or you will see the behavior just as you described, > a hang. One thing you could try is disabling the large message RDMA in OMPI > and

Re: [OMPI users] anybody tried OMPI with gpudirect?

2011-02-28 Thread Rolf vandeVaart
Hi Brice: Yes, I have tired OMPI 1.5 with gpudirect and it worked for me. You definitely need the patch or you will see the behavior just as you described, a hang. One thing you could try is disabling the large message RDMA in OMPI and see if that works. That can be done by adjusting the

[OMPI users] anybody tried OMPI with gpudirect?

2011-02-28 Thread Brice Goglin
Hello, I am trying to play with nvidia's gpudirect. The test program given with the gpudirect tarball just does a basic MPI ping-pong between two process that allocated their buffers with cudaHostMalloc instead of malloc. It seems to work with Intel MPI but Open MPI 1.5 hangs in the first