Hi,
I have this array [1 2 3 4 5 6 7 8 9] and i am performing scan operation on
that.
I have 3 mpi tasks and each task gets 3 elements then each task calculates
its scan and returns result to master task
task 0 - [1 2 3] => [1 3 6]
task 1 - [4 5 6 ] => [4 9 15]
task 2 - [7 8 9] => [7 15 24]
PI_Reduce, instead.
>
>http://www.open-mpi.org/doc/v1.6/man3/MPI_Reduce.3.php
>
>
> On May 15, 2012, at 11:27 PM, Rohan Deshpande wrote:
>
> > I am performing Prefix scan operation on cluster
> >
> > I have 3 MPI tasks and master task is responsible for d
lloc() that you do
> in MCW rank 0 will be totally different than the malloc() that you do in
> MCW rank 1.
>
> Make sense?
>
> On May 16, 2012, at 8:08 AM, Rohan Deshpande wrote:
>
> > I have following structure of MPI code -
> >
> > int main(){
> >
>
I have following structure of MPI code -
int main(){
MPI_INIT.
//initialize MPI
data = malloc(sizeof(int)*200);
//initialize data
/*Master-*/
if(taskid == 0){
//send data to each slave
MPI_SEND
}
/*Slaves---*/
if(taskid > 0){
I am performing Prefix scan operation on cluster
I have 3 MPI tasks and master task is responsible for distributing the data
Now, each task calculates sum of its own part of array using GPUs and
returns the results to master task.
Master task also calculates its own part of array using GPU.
Hi,
Can anyone point me to good resources and books to understand the detailed
architecture and working of MPI.
I would like to know all the minute details.
Thanks,
--
your numbers come out better.
>
> ** **
>
> ** **
>
> *From:* users-boun...@open-mpi.org [mailto:users-boun...@open-mpi.org] *On
> Behalf Of *Rohan Deshpande
> *Sent:* Monday, May 07, 2012 9:38 PM
> *To:* Open MPI Users
> *Subject:* [OMPI users] GPU and CPU timing -
I am running MPI and Thrust code on a cluster and measuring time for
calculations.
My MPI code -
#include "mpi.h"
#include
#include
#include
#include
#include
#include
#define MASTER 0
#define ARRAYSIZE 2000
int
*
>> --
>> mpirun noticed that process rank 0 with PID 3054 on node <
>> http://ecm-c-l-207-004.uniwa.**uwa.edu.au<http://ecm-c-l-207-004.uniwa.uwa.edu.au>
>> >ecm-c-l-207-004.**uniwa.uwa.edu.au<http://ecm-c-l-207-004.uniwa.uwa.edu.au>exited
>> > on signal 11 (Segmentation fault).
>>
>> --**--**
>> --
>>
>> Sadly i cant install memory checker such as valgrind on my machine due to
>> some restrictions. I could not spot any error by looking in code.
>>
>> Can anyone help me ?what is wrong in above code.
>>
>> Thanks
>>
>
>
>
--
Best Regards,
ROHAN DESHPANDE
Hi,
I am trying to execute following code on cluster.
run_kernel is a cuda call with signature int run_kernel(int array[],int
nelements, int taskid, char hostname[]);
My MPI code is
#include "mpi.h"
#include
#include
#include
#define MASTER0
int *masterarray;
int *onearray;
int
d on here:
>
> ** **
>
> http://forums.nvidia.com/index.php?showtopic=227854
>
> ** **
>
> Maybe that could be your issue too.
>
> ** **
>
> ** **
>
> ** **
>
> *From:* users-boun...@open-mpi.org [mailto:users-boun...@open-mpi.org] *On
> Behalf
Hi,
I am trying to execute OpenMPI and CUDA code on a cluster. The code works
fine on single machine but when I try to execute it on cluster I get error:
error while loading shared libraries: libcudart.so.4: cannot open shared
object file: No such file or directory
I checked my PATH and LD_PATH
I am combining mpi and cuda. Trying to find out sum of array elements
using cuda and using mpi to distribute the array.
my cuda code
#include
__global__ void add(int *devarray, int *devsum)
{
int index = blockIdx.x * blockDim.x + threadIdx.x;
*devsum = *devsum +
mysum = mysum + data[i];
}
printf("Task %d has sum = %d\n",myid,mysum);
return(mysum);
}
On Thu, Apr 19, 2012 at 5:59 PM, Jeffrey Squyres <jsquy...@cisco.com> wrote:
> Send the most recent version of your code.
>
> Have you tried running it through a memory-checking deb
, Jeffrey Squyres <jsquy...@cisco.com> wrote:
> As a guess, you're passing in a bad address.
>
> Double check the buffers that you're sending to MPI_SEND/MPI_RECV/etc.
>
>
> On Apr 17, 2012, at 10:43 PM, Rohan Deshpande wrote:
>
> > After using malloc i am getting fol
creating it statically on the stack.
> Something like:
>
> int *data;
>
> int main(..) {
> {
>data = malloc(ARRAYSIZE * sizeof(int));
>if (NULL == data) {
>perror("malloc");
>exit(1);
>}
> // ...
> }
>
>
&g
---
mpirun noticed that process rank 3 with PID 5695 on node machine4 exited on
signal 11 (Segmentation fault).
Any idea what could be wrong here?
--
Best Regards,
ROHAN DESHPANDE
Hi,
I am using Open MPI 1.4.5 and I have CUDA 3.2 installed.
Anyone knows whether CUDA 3.2 is supported by OpenMPI?
Thanks
I actually had too many Ethernet ports open.
When specify which Ethernet to use by btl_tcp_if_include the program works
like a charm :)
Thanks for inputs.
Cheers
On Thu, Apr 5, 2012 at 8:11 AM, Jeffrey Squyres <jsquy...@cisco.com> wrote:
> On Apr 4, 2012, at 8:04 PM, Rohan Deshpa
on the same IP subnet? (e.g., one IP interface
> 192.168.1.x/24 and another on 192.168.1.y/24)
>
>
> On Apr 4, 2012, at 7:55 PM, Rohan Deshpande wrote:
>
> > Hi,
> >
> > I am not quite sure what does that mean. Can you please explain more.
> >
> > Tha
at. :-) Put each of your IP addresses in a different
> IP subnet.
>
>
> On Apr 3, 2012, at 10:30 PM, Rohan Deshpande wrote:
>
> > Hi,
> >
> > I am running a MPI program using cluster and tcp communication.
> >
> > To run I am using: mpirun --prefix /
Hi,
I am running a MPI program using cluster and tcp communication.
To run I am using: *mpirun --prefix /usr/local/ --mca btl tcp,self
--hostfile slaves -np 6 scatter*
I am getting following output:
Process 0 on host host1 has elements 0 1 2 3
Process 1 on host host1 has elements 4 5 6 7
Thanks guys.
Using absolute path of mpirun fixes my problem.
Cheers
On Mon, Apr 2, 2012 at 6:24 PM, Reuti <re...@staff.uni-marburg.de> wrote:
> Am 02.04.2012 um 09:56 schrieb Rohan Deshpande:
>
> > Yes, I am trying to run the program using multiple hosts.
> >
>
at 2:52 PM, Rayson Ho <raysonlo...@gmail.com> wrote:
> On Sun, Apr 1, 2012 at 11:27 PM, Rohan Deshpande <rohan...@gmail.com>
> wrote:
> > error while loading shared libraries: libmpi.so.0: cannot open shared
> > object file no such object file: No such file or
Hi ,
I have installed mpi successfully and I am able to compile the programs
using mpicc
But when I run using mpirun, I get following error
* error while loading shared libraries: libmpi.so.0: cannot open shared
object file no such object file: No such file or directory. *
When I run using -
nds) are split off into that "devel" package.
>>
>> The rationale is that you don't need mpicc (and friends) to *run* Open
>> MPI applications -- you only need mpicc (etc.) to *develop* Open MPI
>> applications.
>>
>> Poke around and see if you can find a de
Hi,
I have installed mpi successfully on fedora using *yum install openmpi
openmpi-devel openmpi-libs*
*
*
I have also added */usr/lib/openmpi/bin* to *PATH *and*
LD_LIBRARY_PATH*variable.
But when I try to complie my program using *mpicc hello.c*
or*/usr/lib/openmpi/bin/mpicc hello.c
* I get
Hi,
I want to distribute the data on different machines using open mpi.
I am a new user. Can some one point me to the resources or atleast
functions i would have to use to complete the task?
I am using red hat linux.
Thanks,
--
28 matches
Mail list logo