Re: [OMPI users] Granular locks?

2011-01-03 Thread Gijsbert Wiesenekker
On Oct 2, 2010, at 10:54 , Gijsbert Wiesenekker wrote: > > On Oct 1, 2010, at 23:24 , Gijsbert Wiesenekker wrote: > >> I have a large array that is shared between two processes. One process >> updates array elements randomly, the other process reads array elemen

Re: [OMPI users] Granular locks?

2010-10-02 Thread Gijsbert Wiesenekker
On Oct 1, 2010, at 23:24 , Gijsbert Wiesenekker wrote: > I have a large array that is shared between two processes. One process > updates array elements randomly, the other process reads array elements > randomly. Most of the time these writes and reads do not overlap. > The cur

[OMPI users] Granular locks?

2010-10-01 Thread Gijsbert Wiesenekker
I have a large array that is shared between two processes. One process updates array elements randomly, the other process reads array elements randomly. Most of the time these writes and reads do not overlap. The current version of the code uses Linux shared memory with NSEMS semaphores. When

Re: [OMPI users] Problem submitting TORQUE jobs with ppn=1

2010-08-17 Thread Gijsbert Wiesenekker
On Aug 17, 2010, at 11:29 , Gijsbert Wiesenekker wrote: > I have a four-node quad core cluster. I am running OpenMPI (version 1.4.2) > jobs with Torque (version 2.4.8). I can submit jobs using > #PBS -lnodes=4:ppn=4 > And 16 processes are launched. However if I use > #PBS

[OMPI users] Problem submitting TORQUE jobs with ppn=1

2010-08-17 Thread Gijsbert Wiesenekker
I have a four-node quad core cluster. I am running OpenMPI (version 1.4.2) jobs with Torque (version 2.4.8). I can submit jobs using #PBS -lnodes=4:ppn=4 And 16 processes are launched. However if I use #PBS -lnodes=4:ppn=1 or #PBS -lnodes=4 The call to MPI_Init is succesful, but the call to

[OMPI users] Behaviour of MPI_Cancel when using 'large' messages

2010-06-07 Thread Gijsbert Wiesenekker
The following code tries to send a message, but if it takes too long the message is cancelled: #define DEADLOCK_ABORT (30.0) MPI_Isend(message, count, MPI_BYTE, comm_id, MPI_MESSAGE_TAG, MPI_COMM_WORLD, ); t0 = time(NULL); cancelled = FALSE; while(TRUE) { //do some work

Re: [OMPI users] Deadlock question

2010-05-27 Thread Gijsbert Wiesenekker
On May 24, 2010, at 20:27 , Eugene Loh wrote: > Gijsbert Wiesenekker wrote: > >> My MPI program consists of a number of processes that send 0 or more >> messages (using MPI_Isend) to 0 or more other processes. The processes check >> periodically if messages are a

Re: [OMPI users] Deadlock question

2010-05-26 Thread Gijsbert Wiesenekker
On May 24, 2010, at 19:42 , Gijsbert Wiesenekker wrote: > My MPI program consists of a number of processes that send 0 or more messages > (using MPI_Isend) to 0 or more other processes. The processes check > periodically if messages are available to be processed. It was running fine

[OMPI users] Deadlock question

2010-05-24 Thread Gijsbert Wiesenekker
My MPI program consists of a number of processes that send 0 or more messages (using MPI_Isend) to 0 or more other processes. The processes check periodically if messages are available to be processed. It was running fine until I increased the message size, and I got deadlock problems. Googling

Re: [OMPI users] Questions about MPI_Isend

2010-05-11 Thread Gijsbert Wiesenekker
On May 11, 2010, at 9:18 , Gijsbert Wiesenekker wrote: > An OpenMPI program of mine that uses MPI_Isend and MPI_Irecv crashes after > some non-reproducible time my Fedora Linux kernel (invalid opcode), which > makes it hard to debug (there is no trace, even with the debug kernel, an

Re: [OMPI users] Questions about MPI_Isend

2010-05-11 Thread Gijsbert Wiesenekker
do this? > > > you can check number of message pending simply using MPI_Test function. It > return false if the request of message is in pending status. The difference > with MPI_Wait is in the behaviour of these two routines: MPI_Wait is > blocking, MPi_Test, check and returns im

[OMPI users] Questions about MPI_Isend

2010-05-11 Thread Gijsbert Wiesenekker
An OpenMPI program of mine that uses MPI_Isend and MPI_Irecv crashes after some non-reproducible time my Fedora Linux kernel (invalid opcode), which makes it hard to debug (there is no trace, even with the debug kernel, and if I run it under valgrind it does not crash). My guess is that the

[OMPI users] Questions on /tmp/openmpi-sessions-userid directory

2010-03-07 Thread Gijsbert Wiesenekker
I was having non-reproducible hangs in an OpenMPI program. While troubleshooting this problem I found that there were many temporary directories in my /tmp/openmpi-sessions-userid directory (probably the result of MPI_Abort aborted OpenMPI programs). I cleaned those directories up and it looks

[OMPI users] OpenMPI problem on Fedora Core 12

2009-12-13 Thread Gijsbert Wiesenekker
The following routine gives a problem after some (not reproducible) time on Fedora Core 12. The routine is a CPU usage friendly version of MPI_Barrier. The verbose output shows that if the problem occurs one of the (not reproducible) nodes does not receive a message from one of the other (not

[OMPI users] GFS2 and OpenMPI

2009-07-26 Thread Gijsbert Wiesenekker
Maybe this question is not entirely appropriate to this forum, but maybe someone reading this forum has already tried this and knows which method is faster. I am about to hook up a NAS node to my Centos based Linux cluster. NAS storage will be shared amongst the nodes using GFS2. My OpenMPI

Re: [OMPI users] Handling output of processes

2009-01-23 Thread Gijsbert Wiesenekker
jody wrote: Hi I have a small cluster consisting of 9 computers (8x2 CPUs, 1x4 CPUs). I would like to be able to observe the output of the processes separately during an mpirun. What i currently do is to apply the mpirun to a shell script which opens a xterm for each process, which then starts

[OMPI users] MPIAbort and raise(SIGTERM)

2009-01-22 Thread Gijsbert Wiesenekker
I am running OpenMPI on Fedora Core 10. MPI_Abort does not seem to abort the processes on FC10. I found out that when you send a signal all processes seem to abort as expected, so I am now using raise(SIGTERM) instead of MPI_Abort(). Are there any consequences to not using MPI_Abort? Regards,

Re: [OMPI users] How to know which task on which node

2009-01-19 Thread Gijsbert Wiesenekker
gaurav gupta wrote: Hello, I want to know that which task is running on which node. Is there any way to know this. Is there any profiling tool provided along with openmpi to calculate time taken in various steps. -- GAURAV GUPTA B.Tech III Yr. , Department of Computer Science & Engineering

[OMPI users] mpiexec and iptables

2009-01-18 Thread Gijsbert Wiesenekker
I am running mpiexec on Fedora Core 10. Where can I find the list of ports that I have to open if I want to run mpiexec with iptables enabled? Running mpiexec with debug shows that it uses ports around 5. strace does not show a call to getenv, so I am not sure if mpiexec on Fedora supports