Hi Lev,

This code worked for me (even after changing line 55 to use
'gpuarray.to_gpu(np.arange...'). I'm on an environment very similar to
yours. Just a couple of suggestions:

1. Insert MPI.Finalize() at the end of your code.
2. If you're not already, pass the parameter '--mca pml ob1' to your
mpiexec command.

Thanks,
Ashwin

On Tue, Dec 23, 2014 at 6:45 PM, Lev Givon <[email protected]> wrote:

> (Not sure if this is more of an mpi4py or a pycuda issue at this point.)
>
> I recently tried running a gist a wrote in the past [1] to test
> communication of
> data stored in GPU memory with pycuda using mpi4py compiled against OpenMPI
> 1.8.* (which contains CUDA support). Using the latest revision (9a70e69)
> compiled against OpenMPI 1.8.4 (which was in turn compiled against CUDA
> 6.5 on
> Ubuntu 14.04.1) and installed in a Python 2.7.6 virtualenv along with
> pycuda
> 2014.1 (also manually compiled against CUDA 6.5), I was able to run the
> gist
> without any problems. However, when I changed line 55 from
>
> x_gpu = gpuarray.arange(100, 200, 10, dtype=np.double)
>
> to
>
> x_gpu = gpuarray.to_gpu(np.arange(100, 200, 10, dtype=np.double))
>
> the data transfer succeeded but was immediately followed by the following
> error:
>
> [avicenna:32494] *** Process received signal ***
> [avicenna:32494] Signal: Segmentation fault (11)
> [avicenna:32494] Signal code: Address not mapped (1)
> [avicenna:32494] Failing at address: (nil)
> [avicenna:32494] [ 0]
> /lib/x86_64-linux-gnu/libpthread.so.0(+0x10340)[0x2ba2e8fe2340]
> [avicenna:32494] [ 1]
> /usr/lib/x86_64-linux-gnu/libcuda.so.1(+0x1f60f5)[0x2ba2fd19b0f5]
> [avicenna:32494] [ 2]
> /usr/lib/x86_64-linux-gnu/libcuda.so.1(+0x20470b)[0x2ba2fd1a970b]
> [avicenna:32494] [ 3]
> /usr/lib/x86_64-linux-gnu/libcuda.so.1(+0x17ac02)[0x2ba2fd11fc02]
> [avicenna:32494] [ 4]
> /usr/lib/x86_64-linux-gnu/libcuda.so.1(cuStreamDestroy_v2+0x52)[0x2ba2fd0eeb32]
> [avicenna:32494] [ 5]
> /opt/openmpi-1.8.4/lib/libmpi.so.1(mca_common_cuda_fini+0x1c3)[0x2ba2f57718a3]
> [avicenna:32494] [ 6]
> /opt/openmpi-1.8.4/lib/libmpi.so.1(+0xf5e3e)[0x2ba2f57aee3e]
> [avicenna:32494] [ 7]
> /opt/openmpi-1.8.4/lib/libopen-pal.so.6(mca_base_component_close+0x19)[0x2ba2f6122099]
> [avicenna:32494] [ 8]
> /opt/openmpi-1.8.4/lib/libopen-pal.so.6(mca_base_components_close+0x42)[0x2ba2f6122112]
> [avicenna:32494] [ 9]
> /opt/openmpi-1.8.4/lib/libmpi.so.1(+0xd7515)[0x2ba2f5790515]
> [avicenna:32494] [10]
> /opt/openmpi-1.8.4/lib/libopen-pal.so.6(mca_base_framework_close+0x63)[0x2ba2f612b3c3]
> [avicenna:32494] [11]
> /opt/openmpi-1.8.4/lib/libopen-pal.so.6(mca_base_framework_close+0x63)[0x2ba2f612b3c3]
> [avicenna:32494] [12]
> /opt/openmpi-1.8.4/lib/libmpi.so.1(ompi_mpi_finalize+0x56d)[0x2ba2f573693d]
> [avicenna:32494] [13]
> /home/lev/Work/virtualenvs/PYTHON/lib/python2.7/site-packages/mpi4py/MPI.so(+0x2e694)[0x2ba2f53b2694]
> [avicenna:32494] [14] python(Py_Finalize+0x1a6)[0x42fb0f]
> [avicenna:32494] [15] python(Py_Main+0xbed)[0x46ac10]
> [avicenna:32494] [16]
> /lib/x86_64-linux-gnu/libc.so.6(__libc_start_main+0xf5)[0x2ba2e9211ec5]
> [avicenna:32494] [17] python[0x57497e]
> [avicenna:32494] *** End of error message ***
>
> I also tried replacing line 55 with
>
> x_gpu = gpuarray.zeros(10, dtype=np.double)
> x_gpu.set(np.arange(100, 200, 10, dtype=np.double))
>
> which resulted in no error and
>
> x_gpu = gpuarray.empty(10, dtype=np.double)
> x_gpu.set(np.arange(100, 200, 10, dtype=np.double))
>
> which resulted in the same error as mentioned earlier.
>
> Any ideas as to what could be going on?
>
> [1] https://gist.github.com/8514d3456a94a6c73e6d
> --
> Lev Givon
> Bionet Group | Neurokernel Project
> http://www.columbia.edu/~lev/
> http://lebedov.github.io/
> http://neurokernel.github.io/
>
>
> _______________________________________________
> PyCUDA mailing list
> [email protected]
> http://lists.tiker.net/listinfo/pycuda
>
_______________________________________________
PyCUDA mailing list
[email protected]
http://lists.tiker.net/listinfo/pycuda

Reply via email to