[OMPI users] (no subject)

2018-11-11 Thread Andrei Berceanu
Hi all, Running a CUDA+MPI application on a node with 2 K80 GPUs, I get the following warnings: -- WARNING: There is at least non-excluded one OpenFabrics device found, but there are no active ports detected (or Open MPI was

Re: [OMPI users] (no subject)

2018-11-01 Thread Jeff Squyres (jsquyres) via users
That's pretty weird. I notice that you're using 3.1.0rc2. Does the same thing happen with Open MPI 3.1.3? > On Oct 31, 2018, at 9:08 PM, Dmitry N. Mikushin wrote: > > Dear all, > > ompi_info reports pml components are available: > > $ /usr/mpi/gcc/openmpi-3.1.0rc2/bin/ompi_info -a | grep

[OMPI users] (no subject)

2018-10-31 Thread Dmitry N. Mikushin
Dear all, ompi_info reports pml components are available: $ /usr/mpi/gcc/openmpi-3.1.0rc2/bin/ompi_info -a | grep pml MCA pml: v (MCA v2.1.0, API v2.0.0, Component v3.1.0) MCA pml: monitoring (MCA v2.1.0, API v2.0.0, Component v3.1.0) MCA pml:

[OMPI users] (no subject)

2017-08-21 Thread Lizhaogeng
Hi all, I encountered a problem when I tested the performance of OpenMPI over ROCE 100Gbps. I have two servers connected with mellanox 100Gbps Connect-X4 ROCE NICs on them. I used intel mpi benchmark to test the performance of OpenMPI (1.10.3) over RDMA. I found the bandwidth of benchmark

Re: [OMPI users] (no subject)

2017-05-16 Thread Gilles Gouaillardet
mmap closed [se01.grid.tuc.gr:19607] mca: base: close: unloading component mmap jb -Original Message- From: users [mailto:users-boun...@lists.open-mpi.org] On Behalf Of gil...@rist.or.jp Sent: Monday, May 15, 2017 1:47 PM To: Open MPI Users <users@lists.open-mpi.org> Subject: Re:

Re: [OMPI users] (no subject)

2017-05-15 Thread Ioannis Botsis
[se01.grid.tuc.gr:19607] mca: base: close: unloading component mmap jb -Original Message- From: users [mailto:users-boun...@lists.open-mpi.org] On Behalf Of gil...@rist.or.jp Sent: Monday, May 15, 2017 1:47 PM To: Open MPI Users <users@lists.open-mpi.org> Subject: Re: [OMPI

Re: [OMPI users] (no subject)

2017-05-15 Thread gilles
Ioannis, ### What version of Open MPI are you using? (e.g., v1.10.3, v2.1.0, git branch name and hash, etc.) ### Describe how Open MPI was installed (e.g., from a source/ distribution tarball, from a git clone, from an operating system distribution package, etc.) ### Please describe the

[OMPI users] (no subject)

2017-05-15 Thread Ioannis Botsis
Hi I am trying to run the following simple demo to a cluster of two nodes -- #include #include int main(int argc, char** argv) { MPI_Init(NULL, NULL); int world_size;

[OMPI users] (no subject)

2016-10-05 Thread Mahmood Naderan
Hi, Is there any idea about the following error? On that node, there are 15 empty cores. Regards, Mahmood ___ users mailing list users@lists.open-mpi.org https://rfd.newmexicoconsortium.org/mailman/listinfo/users

Re: [OMPI users] (no subject)

2016-02-22 Thread Mike Dubman
Hi, it seems that your ompi was compiled with ofed ver X but running on ofed ver Y. X and Y are incompatible. On Mon, Feb 22, 2016 at 8:18 PM, Mark Potter wrote: > I am usually able to find the answer to my problems by searching the > archive but I've run up against one

[OMPI users] (no subject)

2016-02-22 Thread Mark Potter
I am usually able to find the answer to my problems by searching the archive but I've run up against one that I can't suss out. bison-opt: relocation error: /home/pbme002/opt/gcc-4.8.2-tpls/openmpi-1.8.4/lib/libmpi.so.1: symbol rdma_get_src_port, version RDMACM_1.0 not defined in file

[OMPI users] (no subject)

2016-01-18 Thread ?q???s
??

[OMPI users] (no subject)

2014-09-11 Thread Ahmed Salama
I install openmpi-1.6.5 , decide to install new version to support Java openmpi so i chosse openmpi-1.8.2 , configure it with following : $./configure --enable-mpi-java --with-jdk-bindir=/usr/jdk7/bin --with-jdk-headers=/usr/jdk6/include --prefix=/usr/openmpi-1.8.2 and it's ok with no errors

Re: [OMPI users] (no subject)

2013-11-02 Thread Ralph Castain
As I said, the degree of impact depends on the messaging pattern. If rank A typically sends/recvs with rank A+!, then you won't see much difference. However, if rank A typically sends/recvs with rank N-A, where N=#ranks in job, then you'll see a very large difference. You might try simply

Re: [OMPI users] (no subject)

2013-11-02 Thread San B
Yes MM... But here a single node has 16cores not 64 cores. The 1st two jobs were with OMPI-1.4.5. 16 cores of single node - 3692.403 16 cores on two nodes (8 cores per node) - 12338.809 The 1st two jobs were with OMPI-1.6.5. 16 cores of single node - 3547.879 16 cores on

Re: [OMPI users] (no subject)

2013-10-31 Thread Ralph Castain
Yes, though the degree of impact obviously depends on the messaging pattern of the app. On Oct 31, 2013, at 2:50 AM, MM wrote: > Of course, by this you mean, with the same total number of nodes, for e.g. 64 > process on 1 node using shared mem, vs 64 processes spread

Re: [OMPI users] (no subject)

2013-10-31 Thread MM
Of course, by this you mean, with the same total number of nodes, for e.g. 64 process on 1 node using shared mem, vs 64 processes spread over 2 nodes (32 each for e.g.)? On 29 October 2013 14:37, Ralph Castain wrote: > As someone previously noted, apps will always run slower

Re: [OMPI users] (no subject)

2013-10-29 Thread Ralph Castain
I don't think it's a bug in OMPI, but more likely reflects improvements in the default collective algorithms. If you want to further improve performance, you should bind your processes to a core (if your application isn't threaded) or to a socket (if threaded). As someone previously noted,

Re: [OMPI users] (no subject)

2013-10-29 Thread San B
As discussed earlier, the executable which was compiled with OpenMPI-1.4.5 gave very low performance of 12338.809 seconds when job executed on two nodes(8 cores per node). The same job run on single node(all 16cores) got executed in just 3692.403 seconds. Now I compiled the application with

Re: [OMPI users] (no subject)

2013-10-15 Thread San B
Hi, As per your instruction, I did the profiling of the application with mpiP. Following is the difference between the two runs: Run 1: 16 mpi processes on single node @--- MPI Time (seconds) ---

Re: [OMPI users] (no subject)

2013-10-08 Thread Iliev, Hristo
mpiP or Vampir/VampirTrace in order to find why and only then try to further tune Open MPI. Hope that helps, Hristo From: users [mailto:users-boun...@open-mpi.org] On Behalf Of San B Sent: Monday, October 07, 2013 8:46 AM To: OpenMPI ML Subject: [OMPI users] (no subject) Hi, I'm facing

Re: [OMPI users] (no subject)

2013-10-07 Thread Reuti
Hi, Am 07.10.2013 um 08:45 schrieb San B: > I'm facing a performance issue with a scientific application(Fortran). The > issue is, it runs faster on single node but runs very slow on multiple nodes. > For example, a 16 core job on single node finishes in 1hr 2mins, but the same > job on two

[OMPI users] (no subject)

2013-10-07 Thread San B
Hi, I'm facing a performance issue with a scientific application(Fortran). The issue is, it runs faster on single node but runs very slow on multiple nodes. For example, a 16 core job on single node finishes in 1hr 2mins, but the same job on two nodes (i.e. 8 cores per node & remaining 8 cores

Re: [OMPI users] (no subject)

2012-11-30 Thread Josh Hursey
Pramoda, That paper was exploring an application of a proposed extension to the MPI standard for fault tolerance purposes. By default this proposed interface is not provided by Open MPI. We have created a prototype version of Open MPI that includes this extension, and it can be found at the

[OMPI users] (no subject)

2012-11-19 Thread sri pramoda
 Dear Sir,     I am Pramoda,PG scholar from Jadavpur Univesity,India.  I've gone through a paper "Building a Fault Tolerant MPI Application:  A Ring Communication Example".In this I found MPI_Comm_validate_rank command.  But I didn't found this command in mpi. Hence I request you to

[OMPI users] (no subject)

2012-04-09 Thread ffffaa fffff4 ffffaem ffffb3 fffff3
Hello !  I had some  problems . My environment    BLCR= 0.8.4   , openMPI= 1.5.5  , OS= ubuntu 11.04    I have 2 Node : cuda05(Master ,it have NFS  file system)  , cuda07(slave ,mount Master)    I had also set  ~/.openmpi/mca-params.conf-> crs_base_snapshot_dir=/root/kidd_openMPI/Tmp

[OMPI users] (no subject)

2012-04-02 Thread vladimir marjanovic
http://whatbunny.org/web/app/_cache/02efpk.html;> http://whatbunny.org/web/app/_cache/02efpk.html

Re: [OMPI users] (no subject)

2012-03-17 Thread John Hearns
e. > > I hope this helps, > Gus Correa > >> >> From: users-boun...@open-mpi.org [users-boun...@open-mpi.org] On Behalf Of >> jody [jody@gmail.com] >> Sent: Friday, March 16, 2012 4:04 AM >> To: Open MPI Users >>

Re: [OMPI users] (no subject)

2012-03-16 Thread Gustavo Correa
i.org] On Behalf Of > jody [jody@gmail.com] > Sent: Friday, March 16, 2012 4:04 AM > To: Open MPI Users > Subject: Re: [OMPI users] (no subject) > > Hi > > Did you run your program with mpirun? > For example: > mpirun -np 4 ./a.out > > jody > > On F

Re: [OMPI users] (no subject)

2012-03-16 Thread Addepalli, Srirangam V
From: users-boun...@open-mpi.org [users-boun...@open-mpi.org] On Behalf Of jody [jody@gmail.com] Sent: Friday, March 16, 2012 4:04 AM To: Open MPI Users Subject: Re: [OMPI users] (no subject) Hi Did you run your program with mpirun? For example

Re: [OMPI users] (no subject)

2012-03-16 Thread jody
Hi Did you run your program with mpirun? For example: mpirun -np 4 ./a.out jody On Fri, Mar 16, 2012 at 7:24 AM, harini.s .. wrote: > Hi , > > I am very new to openMPI and I just installed openMPI 4.1.5 on Linux > platform. Now am trying to run the examples in the

[OMPI users] (no subject)

2012-03-16 Thread harini.s ..
Hi , I am very new to openMPI and I just installed openMPI 4.1.5 on Linux platform. Now am trying to run the examples in the folder got downloaded. But when i run , I got this >> a.out: error while loading shared libraries: libmpi.so.0: cannot open shared >> object file: No such file or

Re: [OMPI users] (no subject)

2011-02-14 Thread Jeff Squyres
This type of error message *usually* means that you haven't set your LD_LIBRARY_PATH to point to the intel library. Further, this *usually* means that you aren't sourcing the iccvars.sh file in your shell startup file on remote nodes (or iccvars.csh, depending on your shell). Remember that

[OMPI users] (no subject)

2011-02-13 Thread lagoun brahim
hi every one i need your help i have a dual core machine with os linux opensuse 10.3 64bits i configure openmpi with ifort and icc (icpc) i compiled a wien2k code but when i run the parralel version of it i gut the follow error message /home/wien/lapw1_mpi: symbol lookup error:

Re: [OMPI users] (no subject)

2010-06-11 Thread Peter Kjellstrom
On Friday 11 June 2010, asmae.elbahlo...@mpsa.com wrote: > Hello > i have a problem with parFoam, when i type in the terminal parafoam, it > lauches nothing but in the terminal i have : This is the OpenMPI mailling list, not OpenFoam. I suggest you contact the team behind OpenFoam. I also

[OMPI users] (no subject)

2010-06-09 Thread asmae . elbahlouli
asmae.elbahlo...@mpsa.com

Re: [OMPI users] (no subject)

2010-05-18 Thread Josh Hursey
The functionality of checkpoint operation is not tied to CPU utilization. Are you running with the C/R thread enabled? If not then the checkpoint might be waiting until the process enters the MPI library. Does the system emit an error message describing the error that it encountered?

[OMPI users] (no subject)

2010-05-12 Thread ananda.mudar
Ralph When you say manually, do you mean setting these parameters in the command line while calling mpirun, ompi-restart, and ompi-checkpoint? Or is there another way to set these parameters? Thanks Ananda == Subject: Re: [OMPI users] opal_cr_tmp_dir From: Ralph Castain

Re: [OMPI users] (no subject)

2010-02-22 Thread Eugene Loh
Jeff Squyres wrote: On Feb 21, 2010, at 10:25 AM, Rodolfo Chua wrote: I used openMPI compiled with the GNU (gcc) compiler to run GULP code in parallel. But when I try to input "mpirun -np 2 gulp ", GULP did not run in two processors. Can you give me any suggestion on how to

Re: [OMPI users] (no subject)

2010-02-22 Thread Jeff Squyres
On Feb 21, 2010, at 10:25 AM, Rodolfo Chua wrote: > I used openMPI compiled with the GNU (gcc) compiler to run GULP code in > parallel. > But when I try to input "mpirun -np 2 gulp ", GULP did not run in two > processors. Can you give me any suggestion on how to compile GULP code > exactly

[OMPI users] (no subject)

2010-02-21 Thread Rodolfo Chua
Hi! I used openMPI compiled with the GNU (gcc) compiler to run GULP code in parallel. But when I try to input "mpirun -np 2 gulp ", GULP did not run in two processors. Can you give me any suggestion on how to compile GULP code exactly with openMPI. Below is the instruction from GULP code

Re: [OMPI users] (no subject)

2009-10-30 Thread Gus Correa
Hi Konstantinos, list If you want "qsub" you need to install the resource manager / queue system in your PC. Assuming your PC is a Linux box, if your resource manager is Torque/PBS on some Linux distributions it can be installed from an rpm through yum (or equivalent mechanism), for instance. I

Re: [OMPI users] (no subject)

2009-10-30 Thread Peter Kjellstrom
On Friday 30 October 2009, Konstantinos Angelopoulos wrote: > good part of the day, > > I am trying to run a parallel program (that used to run in a cluster) in my > double core pc. Could openmpi simulate the distribution of the parallel > jobs to my 2 processors If your program is an MPI

[OMPI users] (no subject)

2009-10-30 Thread Konstantinos Angelopoulos
good part of the day, I am trying to run a parallel program (that used to run in a cluster) in my double core pc. Could openmpi simulate the distribution of the parallel jobs to my 2 processors meaning will qsub work even if it is not a real cluster? thank you for reading my message and for

[OMPI users] (no subject)

2009-10-30 Thread basant.lakhotiya
Hi All, I compiled OpenMPI in windows server 2003 through Cygwin and also through CMake and Visual Studio. In both the method I successfully complied in cygwin I configured with following command ./configure --enable-mca-no-build=timer-windows,memory_mallopt,maffinity,paffinity without these

[OMPI users] (no subject)

2009-09-25 Thread Mallikarjuna Shastry
dear sir i am sending the details as follows 1. i am using openmpi-1.3.3 and blcr 0.8.2 2. i have installed blcr 0.8.2 first under /root/MS 3. then i installed openmpi 1.3.3 under /root/MS 4 i have configured and installed open mpi as follows #./configure --with-ft=cr --enable-mpi-threads

[OMPI users] (no subject)

2009-08-18 Thread Julia He
Hi, I found that the subroutine call inside a loop did not return correct value after certain iterations. In order to simplify the problem, the inputs to the subroutine are chosen to be constant, so the output should be the same for every iteration on every computing node. It is a fortran

Re: [OMPI users] (no subject)

2009-07-06 Thread Josh Hursey
The MPI standard does not define any functions for taking checkpoints from the application. The checkpoint/restart work in Open MPI is a command line driven, transparent solution. So the application does not have change in any way, and the user (or scheduler) must initiate the checkpoint

[OMPI users] (no subject)

2009-07-06 Thread Mallikarjuna Shastry
dear sir/madam what are the mpi functins used for taking checkpoint and restart within applicaion in mpi programs and where do i get these functions from ? with regards mallikarjuna shastry

Re: [OMPI users] (no subject)

2009-05-14 Thread Camelia Avram
boun...@open-mpi.org] On Behalf Of Jeff Squyres Sent: Thursday, May 14, 2009 3:02 PM To: Open MPI Users Subject: Re: [OMPI users] (no subject) Please send all the information listed here: http://www.open-mpi.org/community/help/ On May 14, 2009, at 1:20 AM, Camelia Avram wrote: > Ni, > I

Re: [OMPI users] (no subject)

2009-05-14 Thread Jeff Squyres
Please send all the information listed here: http://www.open-mpi.org/community/help/ On May 14, 2009, at 1:20 AM, Camelia Avram wrote: Ni, I’m new to MPI. I’m trying to install OpenMPI and I got some errors. I use the command: ./configure –prefix=/usr/local – no problem with this But

[OMPI users] (no subject)

2009-05-14 Thread Camelia Avram
Ni, I'm new to MPI. I'm trying to install OpenMPI and I got some errors. I use the command: ./configure -prefix=/usr/local - no problem with this But after that: "make all install", I got the next message: "no rule to make target 'VERSION', needed by Makefile.in STOP " What should I do?

[OMPI users] (no subject)

2008-11-25 Thread Максим Чусовлянов
Hello! How i can integrated my collective communication algorithm in openMPI with MCA?

Re: [OMPI users] (no subject)

2008-05-27 Thread Jeff Squyres
On May 27, 2008, at 9:33 AM, Gabriele Fatigati wrote: Great, it works! Thank you very very much. But, can you explain me how this parameter works? You might want to have a look at this short video for a little background on some relevant OpenFabrics concepts:

Re: [OMPI users] (no subject)

2008-05-27 Thread Gabriele Fatigati
Great, it works! Thank you very very much. But, can you explain me how this parameter works? On Thu, 15 May 2008 21:40:45 -0400, Jeff Squyres said: > > Sorry this message escaped for so long it got buried in my INBOX. The > problem you're seeing might be related to one we just answered about

Re: [OMPI users] (no subject)

2008-05-15 Thread Jeff Squyres
Sorry this message escaped for so long it got buried in my INBOX. The problem you're seeing might be related to one we just answered about a similar situation: http://www.open-mpi.org/community/lists/users/2008/05/5657.php See if using the pml_ob1_use_early_completion flag works for

Re: [OMPI users] (no subject)

2008-04-23 Thread Ingo Josopait
I can think of several advantages that using blocking or signals to reduce the cpu load would have: - Reduced energy consumption - Running additional background programs could be done far more efficiently - It would be much simpler to examine the load balance. It may depend on the type of

Re: [OMPI users] (no subject)

2008-04-23 Thread Danesh Daroui
Do you really mean that Open-MPI uses busy loop in order to handle incomming calls? It seems to be incorrect since spinning is a very bad and inefficient technique for this purpose. Why don't you use blocking and/or signals instead of that? I think the priority of this task is very high because

Re: [OMPI users] (no subject)

2008-04-23 Thread Jeff Squyres
OMPI doesn't use SYSV shared memory; it uses mmaped files. ompi_info will tell you all about the components installed. If you see a BTL component named "sm", then shared memory support is installed. I do not believe that we conditionally install sm on Linux or OS X systems -- it should

Re: [OMPI users] (no subject)

2008-04-23 Thread Alberto Giannetti
I am running the test program on Darwin 8.11.1, 1.83 Ghz Intel dual core. My Open MPI install is 1.2.4. I can't see any allocated shared memory segment on my system (ipcs - m), although the receiver opens a couple of TCP sockets in listening mode. It looks like my implementation does not use

[OMPI users] (no subject)

2008-04-23 Thread Alberto Giannetti
I have simple MPI program that sends data to processor rank 0. The communication works well but when I run the program on more than 2 processors (-np 4) the extra receivers waiting for data run on > 90% CPU load. I understand MPI_Recv() is a blocking operation, but why does it consume so

[OMPI users] (no subject)

2008-01-24 Thread Wong, Wayne
I'm having some difficulty geting the Open MPI checkpoint/restart fault tolerance working. I have compiled Open MPI with the "--with-ft=cr" flag, but when I attempt to run my test program (ring), the ompi-checkpoint command fails. I have verified that the test program works fine without the

Re: [OMPI users] (no subject)

2007-04-06 Thread JiaXing Cai
? Thank you. ÔÚÄúµÄÀ´ÐÅÖÐÔø¾­Ìáµ½: >From: "Götz Waschk" <goetz.was...@gmail.com> >Reply-To: Open MPI Users <us...@open-mpi.org> >To: "Open MPI Users" <us...@open-mpi.org> >Subject: Re: [OMPI users] (no subject) >Date:Wed, 4 Apr 2007 13:28:15 +0

[OMPI users] (no subject)

2007-04-04 Thread JiaXing Cai
Help I want to do a parallel job on 2 Apple PowerPCs(Power Ma GS Quad) which run on Mac OS X 10.4.8.How can I make them to communicate with each other using open-mpi? I have tried ,but failed.An error related to daemons has occured.

Re: [OMPI users] (no subject)

2006-07-06 Thread Jonathan Blocksom
Check out "Windows Compute Cluster Server 2003", http://www.microsoft.com/windowsserver2003/ccs/default.mspx. From the FAQ: "Windows Compute Cluster Server 2003 comes with the Microsoft Message Passing Interface (MS MPI), an MPI stack based on the MPICH2 implementation from Argonne National