[OMPI users] Gathering results of MPI_SCAN

2012-05-29 Thread Rohan Deshpande
Hi, I have this array [1 2 3 4 5 6 7 8 9] and i am performing scan operation on that. I have 3 mpi tasks and each task gets 3 elements then each task calculates its scan and returns result to master task task 0 - [1 2 3] => [1 3 6] task 1 - [4 5 6 ] => [4 9 15] task 2 - [7 8 9] => [7 15 24]

Re: [OMPI users] MPI_SCAN

2012-05-19 Thread Rohan Deshpande
PI_Reduce, instead. > >http://www.open-mpi.org/doc/v1.6/man3/MPI_Reduce.3.php > > > On May 15, 2012, at 11:27 PM, Rohan Deshpande wrote: > > > I am performing Prefix scan operation on cluster > > > > I have 3 MPI tasks and master task is responsible for d

Re: [OMPI users] MPI - memory sharing?

2012-05-17 Thread Rohan Deshpande
lloc() that you do > in MCW rank 0 will be totally different than the malloc() that you do in > MCW rank 1. > > Make sense? > > On May 16, 2012, at 8:08 AM, Rohan Deshpande wrote: > > > I have following structure of MPI code - > > > > int main(){ > > >

[OMPI users] MPI - memory sharing?

2012-05-16 Thread Rohan Deshpande
I have following structure of MPI code - int main(){ MPI_INIT. //initialize MPI data = malloc(sizeof(int)*200); //initialize data /*Master-*/ if(taskid == 0){ //send data to each slave MPI_SEND } /*Slaves---*/ if(taskid > 0){

[OMPI users] MPI_SCAN

2012-05-16 Thread Rohan Deshpande
I am performing Prefix scan operation on cluster I have 3 MPI tasks and master task is responsible for distributing the data Now, each task calculates sum of its own part of array using GPUs and returns the results to master task. Master task also calculates its own part of array using GPU.

[OMPI users] MPI books and resources

2012-05-12 Thread Rohan Deshpande
Hi, Can anyone point me to good resources and books to understand the detailed architecture and working of MPI. I would like to know all the minute details. Thanks, --

Re: [OMPI users] GPU and CPU timing - OpenMPI and Thrust

2012-05-08 Thread Rohan Deshpande
your numbers come out better. > > ** ** > > ** ** > > *From:* users-boun...@open-mpi.org [mailto:users-boun...@open-mpi.org] *On > Behalf Of *Rohan Deshpande > *Sent:* Monday, May 07, 2012 9:38 PM > *To:* Open MPI Users > *Subject:* [OMPI users] GPU and CPU timing -

[OMPI users] GPU and CPU timing - OpenMPI and Thrust

2012-05-07 Thread Rohan Deshpande
I am running MPI and Thrust code on a cluster and measuring time for calculations. My MPI code - #include "mpi.h" #include #include #include #include #include #include #define MASTER 0 #define ARRAYSIZE 2000 int

Re: [OMPI users] Segmentation fault on master task

2012-05-03 Thread Rohan Deshpande
* >> -- >> mpirun noticed that process rank 0 with PID 3054 on node < >> http://ecm-c-l-207-004.uniwa.**uwa.edu.au<http://ecm-c-l-207-004.uniwa.uwa.edu.au> >> >ecm-c-l-207-004.**uniwa.uwa.edu.au<http://ecm-c-l-207-004.uniwa.uwa.edu.au>exited >> > on signal 11 (Segmentation fault). >> >> --**--** >> -- >> >> Sadly i cant install memory checker such as valgrind on my machine due to >> some restrictions. I could not spot any error by looking in code. >> >> Can anyone help me ?what is wrong in above code. >> >> Thanks >> > > > -- Best Regards, ROHAN DESHPANDE

[OMPI users] Segmentation fault on master task

2012-05-02 Thread Rohan Deshpande
Hi, I am trying to execute following code on cluster. run_kernel is a cuda call with signature int run_kernel(int array[],int nelements, int taskid, char hostname[]); My MPI code is #include "mpi.h" #include #include #include #define MASTER0 int *masterarray; int *onearray; int

Re: [OMPI users] OpenMPI and CUDA on cluster

2012-05-01 Thread Rohan Deshpande
d on here: > > ** ** > > http://forums.nvidia.com/index.php?showtopic=227854 > > ** ** > > Maybe that could be your issue too. > > ** ** > > ** ** > > ** ** > > *From:* users-boun...@open-mpi.org [mailto:users-boun...@open-mpi.org] *On > Behalf

[OMPI users] OpenMPI and CUDA on cluster

2012-05-01 Thread Rohan Deshpande
Hi, I am trying to execute OpenMPI and CUDA code on a cluster. The code works fine on single machine but when I try to execute it on cluster I get error: error while loading shared libraries: libcudart.so.4: cannot open shared object file: No such file or directory I checked my PATH and LD_PATH

[OMPI users] MPI and CUDA

2012-04-24 Thread Rohan Deshpande
I am combining mpi and cuda. Trying to find out sum of array elements using cuda and using mpi to distribute the array. my cuda code #include __global__ void add(int *devarray, int *devsum) { int index = blockIdx.x * blockDim.x + threadIdx.x; *devsum = *devsum +

Re: [OMPI users] machine exited on signal 11 (Segmentation fault).

2012-04-19 Thread Rohan Deshpande
mysum = mysum + data[i]; } printf("Task %d has sum = %d\n",myid,mysum); return(mysum); } On Thu, Apr 19, 2012 at 5:59 PM, Jeffrey Squyres <jsquy...@cisco.com> wrote: > Send the most recent version of your code. > > Have you tried running it through a memory-checking deb

Re: [OMPI users] machine exited on signal 11 (Segmentation fault).

2012-04-19 Thread Rohan Deshpande
, Jeffrey Squyres <jsquy...@cisco.com> wrote: > As a guess, you're passing in a bad address. > > Double check the buffers that you're sending to MPI_SEND/MPI_RECV/etc. > > > On Apr 17, 2012, at 10:43 PM, Rohan Deshpande wrote: > > > After using malloc i am getting fol

Re: [OMPI users] machine exited on signal 11 (Segmentation fault).

2012-04-17 Thread Rohan Deshpande
creating it statically on the stack. > Something like: > > int *data; > > int main(..) { > { >data = malloc(ARRAYSIZE * sizeof(int)); >if (NULL == data) { >perror("malloc"); >exit(1); >} > // ... > } > > &g

[OMPI users] machine exited on signal 11 (Segmentation fault).

2012-04-17 Thread Rohan Deshpande
--- mpirun noticed that process rank 3 with PID 5695 on node machine4 exited on signal 11 (Segmentation fault). Any idea what could be wrong here? -- Best Regards, ROHAN DESHPANDE

[OMPI users] Open MPI 1.4.5 and CUDA support

2012-04-17 Thread Rohan Deshpande
Hi, I am using Open MPI 1.4.5 and I have CUDA 3.2 installed. Anyone knows whether CUDA 3.2 is supported by OpenMPI? Thanks

Re: [OMPI users] Cluster : received unexpected process identifier

2012-04-05 Thread Rohan Deshpande
I actually had too many Ethernet ports open. When specify which Ethernet to use by btl_tcp_if_include the program works like a charm :) Thanks for inputs. Cheers On Thu, Apr 5, 2012 at 8:11 AM, Jeffrey Squyres <jsquy...@cisco.com> wrote: > On Apr 4, 2012, at 8:04 PM, Rohan Deshpa

Re: [OMPI users] Cluster : received unexpected process identifier

2012-04-04 Thread Rohan Deshpande
on the same IP subnet? (e.g., one IP interface > 192.168.1.x/24 and another on 192.168.1.y/24) > > > On Apr 4, 2012, at 7:55 PM, Rohan Deshpande wrote: > > > Hi, > > > > I am not quite sure what does that mean. Can you please explain more. > > > > Tha

Re: [OMPI users] Cluster : received unexpected process identifier

2012-04-04 Thread Rohan Deshpande
at. :-) Put each of your IP addresses in a different > IP subnet. > > > On Apr 3, 2012, at 10:30 PM, Rohan Deshpande wrote: > > > Hi, > > > > I am running a MPI program using cluster and tcp communication. > > > > To run I am using: mpirun --prefix /

[OMPI users] Cluster : received unexpected process identifier

2012-04-03 Thread Rohan Deshpande
Hi, I am running a MPI program using cluster and tcp communication. To run I am using: *mpirun --prefix /usr/local/ --mca btl tcp,self --hostfile slaves -np 6 scatter* I am getting following output: Process 0 on host host1 has elements 0 1 2 3 Process 1 on host host1 has elements 4 5 6 7

Re: [OMPI users] Error while loading shared libraries

2012-04-02 Thread Rohan Deshpande
Thanks guys. Using absolute path of mpirun fixes my problem. Cheers On Mon, Apr 2, 2012 at 6:24 PM, Reuti <re...@staff.uni-marburg.de> wrote: > Am 02.04.2012 um 09:56 schrieb Rohan Deshpande: > > > Yes, I am trying to run the program using multiple hosts. > > >

Re: [OMPI users] Error while loading shared libraries

2012-04-02 Thread Rohan Deshpande
at 2:52 PM, Rayson Ho <raysonlo...@gmail.com> wrote: > On Sun, Apr 1, 2012 at 11:27 PM, Rohan Deshpande <rohan...@gmail.com> > wrote: > > error while loading shared libraries: libmpi.so.0: cannot open shared > > object file no such object file: No such file or

[OMPI users] Error while loading shared libraries

2012-04-02 Thread Rohan Deshpande
Hi , I have installed mpi successfully and I am able to compile the programs using mpicc But when I run using mpirun, I get following error * error while loading shared libraries: libmpi.so.0: cannot open shared object file no such object file: No such file or directory. * When I run using -

Re: [OMPI users] mpicc command not found - Fedora

2012-03-30 Thread Rohan Deshpande
nds) are split off into that "devel" package. >> >> The rationale is that you don't need mpicc (and friends) to *run* Open >> MPI applications -- you only need mpicc (etc.) to *develop* Open MPI >> applications. >> >> Poke around and see if you can find a de

[OMPI users] mpicc command not found - Fedora

2012-03-29 Thread Rohan Deshpande
Hi, I have installed mpi successfully on fedora using *yum install openmpi openmpi-devel openmpi-libs* * * I have also added */usr/lib/openmpi/bin* to *PATH *and* LD_LIBRARY_PATH*variable. But when I try to complie my program using *mpicc hello.c* or*/usr/lib/openmpi/bin/mpicc hello.c * I get

[OMPI users] Data distribution on different machines

2012-03-25 Thread Rohan Deshpande
Hi, I want to distribute the data on different machines using open mpi. I am a new user. Can some one point me to the resources or atleast functions i would have to use to complete the task? I am using red hat linux. Thanks, --