Hi,
I want to distribute the data on different machines using open mpi.
I am a new user. Can some one point me to the resources or atleast
functions i would have to use to complete the task?
I am using red hat linux.
Thanks,
--
Hi,
I have installed mpi successfully on fedora using *yum install openmpi
openmpi-devel openmpi-libs*
*
*
I have also added */usr/lib/openmpi/bin* to *PATH *and*
LD_LIBRARY_PATH*variable.
But when I try to complie my program using *mpicc hello.c*
or*/usr/lib/openmpi/bin/mpicc hello.c
* I get
nds) are split off into that "devel" package.
>>
>> The rationale is that you don't need mpicc (and friends) to *run* Open
>> MPI applications -- you only need mpicc (etc.) to *develop* Open MPI
>> applications.
>>
>> Poke around and see if you can find a de
Hi ,
I have installed mpi successfully and I am able to compile the programs
using mpicc
But when I run using mpirun, I get following error
* error while loading shared libraries: libmpi.so.0: cannot open shared
object file no such object file: No such file or directory. *
When I run using -
at 2:52 PM, Rayson Ho <raysonlo...@gmail.com> wrote:
> On Sun, Apr 1, 2012 at 11:27 PM, Rohan Deshpande <rohan...@gmail.com>
> wrote:
> > error while loading shared libraries: libmpi.so.0: cannot open shared
> > object file no such object file: No such file or
Thanks guys.
Using absolute path of mpirun fixes my problem.
Cheers
On Mon, Apr 2, 2012 at 6:24 PM, Reuti <re...@staff.uni-marburg.de> wrote:
> Am 02.04.2012 um 09:56 schrieb Rohan Deshpande:
>
> > Yes, I am trying to run the program using multiple hosts.
> >
>
Hi,
I am running a MPI program using cluster and tcp communication.
To run I am using: *mpirun --prefix /usr/local/ --mca btl tcp,self
--hostfile slaves -np 6 scatter*
I am getting following output:
Process 0 on host host1 has elements 0 1 2 3
Process 1 on host host1 has elements 4 5 6 7
at. :-) Put each of your IP addresses in a different
> IP subnet.
>
>
> On Apr 3, 2012, at 10:30 PM, Rohan Deshpande wrote:
>
> > Hi,
> >
> > I am running a MPI program using cluster and tcp communication.
> >
> > To run I am using: mpirun --prefix /
on the same IP subnet? (e.g., one IP interface
> 192.168.1.x/24 and another on 192.168.1.y/24)
>
>
> On Apr 4, 2012, at 7:55 PM, Rohan Deshpande wrote:
>
> > Hi,
> >
> > I am not quite sure what does that mean. Can you please explain more.
> >
> > Tha
I actually had too many Ethernet ports open.
When specify which Ethernet to use by btl_tcp_if_include the program works
like a charm :)
Thanks for inputs.
Cheers
On Thu, Apr 5, 2012 at 8:11 AM, Jeffrey Squyres <jsquy...@cisco.com> wrote:
> On Apr 4, 2012, at 8:04 PM, Rohan Deshpa
Hi,
I am using Open MPI 1.4.5 and I have CUDA 3.2 installed.
Anyone knows whether CUDA 3.2 is supported by OpenMPI?
Thanks
---
mpirun noticed that process rank 3 with PID 5695 on node machine4 exited on
signal 11 (Segmentation fault).
Any idea what could be wrong here?
--
Best Regards,
ROHAN DESHPANDE
creating it statically on the stack.
> Something like:
>
> int *data;
>
> int main(..) {
> {
>data = malloc(ARRAYSIZE * sizeof(int));
>if (NULL == data) {
>perror("malloc");
>exit(1);
>}
> // ...
> }
>
>
&g
, Jeffrey Squyres <jsquy...@cisco.com> wrote:
> As a guess, you're passing in a bad address.
>
> Double check the buffers that you're sending to MPI_SEND/MPI_RECV/etc.
>
>
> On Apr 17, 2012, at 10:43 PM, Rohan Deshpande wrote:
>
> > After using malloc i am getting fol
mysum = mysum + data[i];
}
printf("Task %d has sum = %d\n",myid,mysum);
return(mysum);
}
On Thu, Apr 19, 2012 at 5:59 PM, Jeffrey Squyres <jsquy...@cisco.com> wrote:
> Send the most recent version of your code.
>
> Have you tried running it through a memory-checking deb
I am combining mpi and cuda. Trying to find out sum of array elements
using cuda and using mpi to distribute the array.
my cuda code
#include
__global__ void add(int *devarray, int *devsum)
{
int index = blockIdx.x * blockDim.x + threadIdx.x;
*devsum = *devsum +
Hi,
I am trying to execute OpenMPI and CUDA code on a cluster. The code works
fine on single machine but when I try to execute it on cluster I get error:
error while loading shared libraries: libcudart.so.4: cannot open shared
object file: No such file or directory
I checked my PATH and LD_PATH
d on here:
>
> ** **
>
> http://forums.nvidia.com/index.php?showtopic=227854
>
> ** **
>
> Maybe that could be your issue too.
>
> ** **
>
> ** **
>
> ** **
>
> *From:* users-boun...@open-mpi.org [mailto:users-boun...@open-mpi.org] *On
> Behalf
Hi,
I am trying to execute following code on cluster.
run_kernel is a cuda call with signature int run_kernel(int array[],int
nelements, int taskid, char hostname[]);
My MPI code is
#include "mpi.h"
#include
#include
#include
#define MASTER0
int *masterarray;
int *onearray;
int
*
>> --
>> mpirun noticed that process rank 0 with PID 3054 on node <
>> http://ecm-c-l-207-004.uniwa.**uwa.edu.au<http://ecm-c-l-207-004.uniwa.uwa.edu.au>
>> >ecm-c-l-207-004.**uniwa.uwa.edu.au<http://ecm-c-l-207-004.uniwa.uwa.edu.au>exited
>> > on signal 11 (Segmentation fault).
>>
>> --**--**
>> --
>>
>> Sadly i cant install memory checker such as valgrind on my machine due to
>> some restrictions. I could not spot any error by looking in code.
>>
>> Can anyone help me ?what is wrong in above code.
>>
>> Thanks
>>
>
>
>
--
Best Regards,
ROHAN DESHPANDE
I am running MPI and Thrust code on a cluster and measuring time for
calculations.
My MPI code -
#include "mpi.h"
#include
#include
#include
#include
#include
#include
#define MASTER 0
#define ARRAYSIZE 2000
int
your numbers come out better.
>
> ** **
>
> ** **
>
> *From:* users-boun...@open-mpi.org [mailto:users-boun...@open-mpi.org] *On
> Behalf Of *Rohan Deshpande
> *Sent:* Monday, May 07, 2012 9:38 PM
> *To:* Open MPI Users
> *Subject:* [OMPI users] GPU and CPU timing -
Hi,
Can anyone point me to good resources and books to understand the detailed
architecture and working of MPI.
I would like to know all the minute details.
Thanks,
--
I am performing Prefix scan operation on cluster
I have 3 MPI tasks and master task is responsible for distributing the data
Now, each task calculates sum of its own part of array using GPUs and
returns the results to master task.
Master task also calculates its own part of array using GPU.
I have following structure of MPI code -
int main(){
MPI_INIT.
//initialize MPI
data = malloc(sizeof(int)*200);
//initialize data
/*Master-*/
if(taskid == 0){
//send data to each slave
MPI_SEND
}
/*Slaves---*/
if(taskid > 0){
lloc() that you do
> in MCW rank 0 will be totally different than the malloc() that you do in
> MCW rank 1.
>
> Make sense?
>
> On May 16, 2012, at 8:08 AM, Rohan Deshpande wrote:
>
> > I have following structure of MPI code -
> >
> > int main(){
> >
>
PI_Reduce, instead.
>
>http://www.open-mpi.org/doc/v1.6/man3/MPI_Reduce.3.php
>
>
> On May 15, 2012, at 11:27 PM, Rohan Deshpande wrote:
>
> > I am performing Prefix scan operation on cluster
> >
> > I have 3 MPI tasks and master task is responsible for d
Hi,
I have this array [1 2 3 4 5 6 7 8 9] and i am performing scan operation on
that.
I have 3 mpi tasks and each task gets 3 elements then each task calculates
its scan and returns result to master task
task 0 - [1 2 3] => [1 3 6]
task 1 - [4 5 6 ] => [4 9 15]
task 2 - [7 8 9] => [7 15 24]
28 matches
Mail list logo