Re: [OMPI users] (no subject)

2018-11-01 Thread Jeff Squyres (jsquyres) via users
That's pretty weird. I notice that you're using 3.1.0rc2. Does the same thing happen with Open MPI 3.1.3? > On Oct 31, 2018, at 9:08 PM, Dmitry N. Mikushin wrote: > > Dear all, > > ompi_info reports pml components are available: > > $ /usr/mpi/gcc/openmpi-3.1.0rc2/bin/ompi_info -a | grep

Re: [OMPI users] (no subject)

2017-05-16 Thread Gilles Gouaillardet
mmap closed [se01.grid.tuc.gr:19607] mca: base: close: unloading component mmap jb -Original Message- From: users [mailto:users-boun...@lists.open-mpi.org] On Behalf Of gil...@rist.or.jp Sent: Monday, May 15, 2017 1:47 PM To: Open MPI Users <users@lists.open-mpi.org> Subject: Re:

Re: [OMPI users] (no subject)

2017-05-15 Thread Ioannis Botsis
[se01.grid.tuc.gr:19607] mca: base: close: unloading component mmap jb -Original Message- From: users [mailto:users-boun...@lists.open-mpi.org] On Behalf Of gil...@rist.or.jp Sent: Monday, May 15, 2017 1:47 PM To: Open MPI Users <users@lists.open-mpi.org> Subject: Re: [OMPI

Re: [OMPI users] (no subject)

2017-05-15 Thread gilles
Ioannis, ### What version of Open MPI are you using? (e.g., v1.10.3, v2.1.0, git branch name and hash, etc.) ### Describe how Open MPI was installed (e.g., from a source/ distribution tarball, from a git clone, from an operating system distribution package, etc.) ### Please describe the

Re: [OMPI users] (no subject)

2016-02-22 Thread Mike Dubman
Hi, it seems that your ompi was compiled with ofed ver X but running on ofed ver Y. X and Y are incompatible. On Mon, Feb 22, 2016 at 8:18 PM, Mark Potter wrote: > I am usually able to find the answer to my problems by searching the > archive but I've run up against one

Re: [OMPI users] (no subject)

2013-11-02 Thread Ralph Castain
As I said, the degree of impact depends on the messaging pattern. If rank A typically sends/recvs with rank A+!, then you won't see much difference. However, if rank A typically sends/recvs with rank N-A, where N=#ranks in job, then you'll see a very large difference. You might try simply

Re: [OMPI users] (no subject)

2013-11-02 Thread San B
Yes MM... But here a single node has 16cores not 64 cores. The 1st two jobs were with OMPI-1.4.5. 16 cores of single node - 3692.403 16 cores on two nodes (8 cores per node) - 12338.809 The 1st two jobs were with OMPI-1.6.5. 16 cores of single node - 3547.879 16 cores on

Re: [OMPI users] (no subject)

2013-10-31 Thread Ralph Castain
Yes, though the degree of impact obviously depends on the messaging pattern of the app. On Oct 31, 2013, at 2:50 AM, MM wrote: > Of course, by this you mean, with the same total number of nodes, for e.g. 64 > process on 1 node using shared mem, vs 64 processes spread

Re: [OMPI users] (no subject)

2013-10-31 Thread MM
Of course, by this you mean, with the same total number of nodes, for e.g. 64 process on 1 node using shared mem, vs 64 processes spread over 2 nodes (32 each for e.g.)? On 29 October 2013 14:37, Ralph Castain wrote: > As someone previously noted, apps will always run slower

Re: [OMPI users] (no subject)

2013-10-29 Thread Ralph Castain
I don't think it's a bug in OMPI, but more likely reflects improvements in the default collective algorithms. If you want to further improve performance, you should bind your processes to a core (if your application isn't threaded) or to a socket (if threaded). As someone previously noted,

Re: [OMPI users] (no subject)

2013-10-29 Thread San B
As discussed earlier, the executable which was compiled with OpenMPI-1.4.5 gave very low performance of 12338.809 seconds when job executed on two nodes(8 cores per node). The same job run on single node(all 16cores) got executed in just 3692.403 seconds. Now I compiled the application with

Re: [OMPI users] (no subject)

2013-10-15 Thread San B
Hi, As per your instruction, I did the profiling of the application with mpiP. Following is the difference between the two runs: Run 1: 16 mpi processes on single node @--- MPI Time (seconds) ---

Re: [OMPI users] (no subject)

2013-10-08 Thread Iliev, Hristo
Hi, When all processes run on the same node they communicate via shared memory which delivers both high bandwidth and low latency. InfiniBand is slower and more latent than shared memory. Your parallel algorithm might simply be very latency sensitive and you should profile it with something like

Re: [OMPI users] (no subject)

2013-10-07 Thread Reuti
Hi, Am 07.10.2013 um 08:45 schrieb San B: > I'm facing a performance issue with a scientific application(Fortran). The > issue is, it runs faster on single node but runs very slow on multiple nodes. > For example, a 16 core job on single node finishes in 1hr 2mins, but the same > job on two

Re: [OMPI users] (no subject)

2012-11-30 Thread Josh Hursey
Pramoda, That paper was exploring an application of a proposed extension to the MPI standard for fault tolerance purposes. By default this proposed interface is not provided by Open MPI. We have created a prototype version of Open MPI that includes this extension, and it can be found at the

Re: [OMPI users] (no subject)

2012-03-17 Thread John Hearns
e. > > I hope this helps, > Gus Correa > >> >> From: users-boun...@open-mpi.org [users-boun...@open-mpi.org] On Behalf Of >> jody [jody@gmail.com] >> Sent: Friday, March 16, 2012 4:04 AM >> To: Open MPI Users >>

Re: [OMPI users] (no subject)

2012-03-16 Thread Gustavo Correa
i.org] On Behalf Of > jody [jody@gmail.com] > Sent: Friday, March 16, 2012 4:04 AM > To: Open MPI Users > Subject: Re: [OMPI users] (no subject) > > Hi > > Did you run your program with mpirun? > For example: > mpirun -np 4 ./a.out > > jody > > On F

Re: [OMPI users] (no subject)

2012-03-16 Thread Addepalli, Srirangam V
From: users-boun...@open-mpi.org [users-boun...@open-mpi.org] On Behalf Of jody [jody@gmail.com] Sent: Friday, March 16, 2012 4:04 AM To: Open MPI Users Subject: Re: [OMPI users] (no subject) Hi Did you run your program with mpirun? For example

Re: [OMPI users] (no subject)

2012-03-16 Thread jody
Hi Did you run your program with mpirun? For example: mpirun -np 4 ./a.out jody On Fri, Mar 16, 2012 at 7:24 AM, harini.s .. wrote: > Hi , > > I am very new to openMPI and I just installed openMPI 4.1.5 on Linux > platform. Now am trying to run the examples in the

Re: [OMPI users] (no subject)

2011-02-14 Thread Jeff Squyres
This type of error message *usually* means that you haven't set your LD_LIBRARY_PATH to point to the intel library. Further, this *usually* means that you aren't sourcing the iccvars.sh file in your shell startup file on remote nodes (or iccvars.csh, depending on your shell). Remember that

Re: [OMPI users] (no subject)

2010-06-11 Thread Peter Kjellstrom
On Friday 11 June 2010, asmae.elbahlo...@mpsa.com wrote: > Hello > i have a problem with parFoam, when i type in the terminal parafoam, it > lauches nothing but in the terminal i have : This is the OpenMPI mailling list, not OpenFoam. I suggest you contact the team behind OpenFoam. I also

Re: [OMPI users] (no subject)

2010-05-18 Thread Josh Hursey
The functionality of checkpoint operation is not tied to CPU utilization. Are you running with the C/R thread enabled? If not then the checkpoint might be waiting until the process enters the MPI library. Does the system emit an error message describing the error that it encountered?

Re: [OMPI users] (no subject)

2010-02-22 Thread Eugene Loh
Jeff Squyres wrote: On Feb 21, 2010, at 10:25 AM, Rodolfo Chua wrote: I used openMPI compiled with the GNU (gcc) compiler to run GULP code in parallel. But when I try to input "mpirun -np 2 gulp ", GULP did not run in two processors. Can you give me any suggestion on how to

Re: [OMPI users] (no subject)

2010-02-22 Thread Jeff Squyres
On Feb 21, 2010, at 10:25 AM, Rodolfo Chua wrote: > I used openMPI compiled with the GNU (gcc) compiler to run GULP code in > parallel. > But when I try to input "mpirun -np 2 gulp ", GULP did not run in two > processors. Can you give me any suggestion on how to compile GULP code > exactly

Re: [OMPI users] (no subject)

2009-10-30 Thread Gus Correa
Hi Konstantinos, list If you want "qsub" you need to install the resource manager / queue system in your PC. Assuming your PC is a Linux box, if your resource manager is Torque/PBS on some Linux distributions it can be installed from an rpm through yum (or equivalent mechanism), for instance. I

Re: [OMPI users] (no subject)

2009-10-30 Thread Peter Kjellstrom
On Friday 30 October 2009, Konstantinos Angelopoulos wrote: > good part of the day, > > I am trying to run a parallel program (that used to run in a cluster) in my > double core pc. Could openmpi simulate the distribution of the parallel > jobs to my 2 processors If your program is an MPI

Re: [OMPI users] (no subject)

2009-07-06 Thread Josh Hursey
The MPI standard does not define any functions for taking checkpoints from the application. The checkpoint/restart work in Open MPI is a command line driven, transparent solution. So the application does not have change in any way, and the user (or scheduler) must initiate the checkpoint

Re: [OMPI users] (no subject)

2009-05-14 Thread Camelia Avram
boun...@open-mpi.org] On Behalf Of Jeff Squyres Sent: Thursday, May 14, 2009 3:02 PM To: Open MPI Users Subject: Re: [OMPI users] (no subject) Please send all the information listed here: http://www.open-mpi.org/community/help/ On May 14, 2009, at 1:20 AM, Camelia Avram wrote: > Ni, > I

Re: [OMPI users] (no subject)

2009-05-14 Thread Jeff Squyres
Please send all the information listed here: http://www.open-mpi.org/community/help/ On May 14, 2009, at 1:20 AM, Camelia Avram wrote: Ni, I’m new to MPI. I’m trying to install OpenMPI and I got some errors. I use the command: ./configure –prefix=/usr/local – no problem with this But

Re: [OMPI users] (no subject)

2008-05-27 Thread Jeff Squyres
On May 27, 2008, at 9:33 AM, Gabriele Fatigati wrote: Great, it works! Thank you very very much. But, can you explain me how this parameter works? You might want to have a look at this short video for a little background on some relevant OpenFabrics concepts:

Re: [OMPI users] (no subject)

2008-05-27 Thread Gabriele Fatigati
Great, it works! Thank you very very much. But, can you explain me how this parameter works? On Thu, 15 May 2008 21:40:45 -0400, Jeff Squyres said: > > Sorry this message escaped for so long it got buried in my INBOX. The > problem you're seeing might be related to one we just answered about

Re: [OMPI users] (no subject)

2008-05-15 Thread Jeff Squyres
Sorry this message escaped for so long it got buried in my INBOX. The problem you're seeing might be related to one we just answered about a similar situation: http://www.open-mpi.org/community/lists/users/2008/05/5657.php See if using the pml_ob1_use_early_completion flag works for

Re: [OMPI users] (no subject)

2008-04-23 Thread Ingo Josopait
I can think of several advantages that using blocking or signals to reduce the cpu load would have: - Reduced energy consumption - Running additional background programs could be done far more efficiently - It would be much simpler to examine the load balance. It may depend on the type of

Re: [OMPI users] (no subject)

2008-04-23 Thread Danesh Daroui
Do you really mean that Open-MPI uses busy loop in order to handle incomming calls? It seems to be incorrect since spinning is a very bad and inefficient technique for this purpose. Why don't you use blocking and/or signals instead of that? I think the priority of this task is very high because

Re: [OMPI users] (no subject)

2008-04-23 Thread Jeff Squyres
OMPI doesn't use SYSV shared memory; it uses mmaped files. ompi_info will tell you all about the components installed. If you see a BTL component named "sm", then shared memory support is installed. I do not believe that we conditionally install sm on Linux or OS X systems -- it should

Re: [OMPI users] (no subject)

2008-04-23 Thread Alberto Giannetti
I am running the test program on Darwin 8.11.1, 1.83 Ghz Intel dual core. My Open MPI install is 1.2.4. I can't see any allocated shared memory segment on my system (ipcs - m), although the receiver opens a couple of TCP sockets in listening mode. It looks like my implementation does not use

Re: [OMPI users] (no subject)

2007-04-06 Thread JiaXing Cai
? Thank you. ÔÚÄúµÄÀ´ÐÅÖÐÔø¾­Ìáµ½: >From: "Götz Waschk" <goetz.was...@gmail.com> >Reply-To: Open MPI Users <us...@open-mpi.org> >To: "Open MPI Users" <us...@open-mpi.org> >Subject: Re: [OMPI users] (no subject) >Date:Wed, 4 Apr 2007 13:28:15 +0

Re: [OMPI users] (no subject)

2006-07-06 Thread Jonathan Blocksom
Check out "Windows Compute Cluster Server 2003", http://www.microsoft.com/windowsserver2003/ccs/default.mspx. From the FAQ: "Windows Compute Cluster Server 2003 comes with the Microsoft Message Passing Interface (MS MPI), an MPI stack based on the MPICH2 implementation from Argonne National