Did you run the program from your home directory or the /opt/ directory where it lives? For LAM to work properly the program needs to be available on all the compute nodes. By default, /home is the only directory exported by nfs.
I am still confused by the mpi_init error even if that is the problem, did you compile the program using gcc or mpicc? On 4/4/06, Michelle Chu <[EMAIL PROTECTED]> wrote: > > Bernard, > Here are some outputs... > [EMAIL PROTECTED] ~]$ lamnodes > n0 athena.cs.xxx.edu:1:origin,this_node > n1 oscarnode1.cs.xxx.edu:1 : > n2 oscarnode2.cs.xxx.edu:1: > n3 oscarnode3.cs.xxx.edu:1: > n4 oscarnode4.cs.xxx.edu:1 : > n5 oscarnode5.cs.xxx.edu:1: > n6 oscarnode6.cs.xxx.edu:1: > n7 oscarnode7.cs.xxx.edu:1 : > n8 oscarnode8.cs.xxx.edu:1: > *********************************************************************************************** > [EMAIL PROTECTED] ~]$ vi lamtest. output > > wall clock time = 0.000074 > Process 1 of 8 on oscarnode8.cs.xxx.edu > > --> MPI C++ bindings test: > > Hello World! I am 1 of 8 > Hello World! I am 0 of 8 > Hello World! I am 6 of 8 > Hello World! I am 4 of 8 > Hello World! I am 2 of 8 > Hello World! I am 5 of 8 > Hello World! I am 7 of 8 > Hello World! I am 3 of 8 > > > > --> MPI Fortran bindings test: > > Hello World! I am 0 of 8 > Hello World! I am 1 of 8 > Hello World! I am 4 of 8 > Hello World! I am 6 of 8 > Hello World! I am 2 of 8 > Hello World! I am 7 of 8 > Hello World! I am 5 of 8 > Hello World! I am 3 of 8 > > > LAM 7.0.6/MPI 2 C++/ROMIO - Indiana University > > > LAM/MPI test complete > Unless there are errors above, test completed successfully. > > ******************************************************************************************************************** > [EMAIL PROTECTED] examples]$ mpirun N hello++ > > Hello World! I am 0 of 1 > Hello World! I am 0 of 1 > Hello World! I am 0 of 1 > Hello World! I am 0 of 1 > Hello World! I am 0 of 1 > Hello World! I am 0 of 1 > Hello World! I am 0 of 1 > Hello World! I am 0 of 1 > ----------------------------------------------------------------------------- > It seems that [at least] one of the processes that was started with > mpirun did not invoke MPI_INIT before quitting (it is possible that > more than one process did not invoke MPI_INIT -- mpirun was only > notified of the first one, which was on node n0). > > mpirun can *only* be used with MPI programs (i.e., programs that > invoke MPI_INIT and MPI_FINALIZE). You can use the "lamexec" program > to run non-MPI programs over the lambooted nodes. > ********************************************************************************************************************* > Attached is the output file file: lamboot -d hostfile > > What i did was: > 1). lamboot -d hostfile > 2). mpirun N hello++ > hostfile lists host name of headnode and all eight cluster nodes as: > > athena.cs.xxx.edu > oscarnode1.cs.xxx.edu > > ..... > oscarnode8.cs.xxx.edu > > Thanks, > > > Michelle > > On 4/4/06, Bernard Li <[EMAIL PROTECTED]> wrote: > > > > > > > > Hi Michelle: > > > > I just tested your code on a 2 node cluster (including headnode) and got > the following result: > > > > $ mpirun N a.out > > Hello World! I am 1 of 2 > > Hello World! I am 0 of 2 > > > > So it seems fine (you had a space between " and 'mpi.h' but I fixed that) > > > > Can you show us the output of "lamnodes" after you have successfully > booted the nodes? Also, post the output of your LAM/MPI OSCAR tests > (/home/oscartst/lam/lamtest.out). > > > > Cheers, > > > > Bernard > > > > ________________________________ > From: [EMAIL PROTECTED] on behalf of > Michelle Chu > > Sent: Mon 03/04/2006 21:02 > > To: Michael Edwards > > Cc: [email protected] > > Subject: Re: [Oscar-users] MPI ranking and size problem > > > > > > > > > > Michael, > > OSCAR version is 4.2. The OS on the client node is Red Hat and is > installed from the client image file generated during the OSCAR > installation. The cluster testing step at the end of installation passed > except the ganglia part. Thank you very much for your help. Michelle > > > > Here is the code for Hello.cc. > > > ******************************************************************************************* > > #include <iostream.h> > > // modified to reference the master mpi.h file, to meet the MPI standard > spec. > > #include " mpi.h" > > int > > main(int argc, char *argv[]) > > { > > MPI::Init(argc, argv); > > > > int rank = MPI::COMM_WORLD.Get_rank(); > > int size = MPI::COMM_WORLD.Get_size(); > > > > cout << "Hello World! I am " << rank << " of " << size << endl; > > > > MPI::Finalize(); > > } > > > > > ***************************************************************************************** > > > > > > On 4/3/06, Michael Edwards <[EMAIL PROTECTED] > wrote: > > > What version of OSCAR are you using, and on what platform? > > > > > > Also, could you send us a copy of hello++.cpp, it looks like there are > > > some errors there? Also, did all the oscar tests pass? > > > > > > LAM appears to be working correctly at surface anyway. > > > > > > > > > > > > On 4/3/06, Michelle Chu < [EMAIL PROTECTED] > wrote: > > > > Hello, there, > > > > > > > > When I mpirun a simple hello MPI program on all my eight nodes as the > > > > following. I get a sequence of hello world! i am 0 of 1, instead of 1 > of 8, > > > > 2 of 8, 3 of 8. Also, problem with MPI_INIT. Thank you for your help. > > > > > > > > which mpicc > > > > /opt/lam-7.0.6/bin/mpicc > > > > lamboot -v my_hostfile > > > > my_hostfile is: > > > > ************************************** > > > > athena.cs.xxx.edu > > > > oscarnode1.cs.xxx.edu > > > > oscarnode2.cs.xxx.edu > > > > oscarnode3.cs.xxx.edu > > > > oscarnode4.cs.xxx.edu > > > > oscarnode5.cs.xxx.edu > > > > oscarnode6.cs.xxx.edu > > > > oscarnode7.cs.xxx.edu > > > > oscarnode8.cs.xxx.edu > > > > > > > > > ***************************************************************************** > > > > LAM 7.0.6/MPI 2 C++/ROMIO - Indiana University > > > > > > > > n-1<16365> ssi:boot:base:linear: booting n0 ( athena.cs.xxx.edu) > > > > n-1<16365> ssi:boot:base:linear: booting n1 ( oscarnode1.cs.xxx.edu) > > > > n-1<16365> ssi:boot:base:linear: booting n2 (oscarnode2.cs.xxx.edu) > > > > n-1<16365> ssi:boot:base:linear: booting n3 ( oscarnode3.cs.xxx.edu) > > > > n-1<16365> ssi:boot:base:linear: booting n4 ( oscarnode4.cs.xxx.edu) > > > > n-1<16365> ssi:boot:base:linear: booting n5 ( oscarnode5.cs.xxx.edu ) > > > > n-1<16365> ssi:boot:base:linear: booting n6 (oscarnode6.cs.xxx.edu) > > > > n-1<16365> ssi:boot:base:linear: booting n7 ( oscarnode7.cs.xxx.edu) > > > > n-1<16365> ssi:boot:base:linear: booting n8 ( oscarnode8.cs.xxx.edu) > > > > n-1<16365> ssi:boot:base:linear: finished > > > > > > > > mpirun N hello++ > > > > > ***************************************************************** > > > > Hello World! I am 0 of 1 > > > > Hello World! I am 0 of 1 > > > > Hello World! I am 0 of 1 > > > > Hello World! I am 0 of 1 > > > > Hello World! I am 0 of 1 > > > > Hello World! I am 0 of 1 > > > > > ----------------------------------------------------------------------------- > > > > It seems that [at least] one of the processes that was started with > > > > mpirun did not invoke MPI_INIT before quitting (it is possible that > > > > more than one process did not invoke MPI_INIT -- mpirun was only > > > > notified of the first one, which was on node n0). > > > > > > > > mpirun can *only* be used with MPI programs (i.e., programs that > > > > invoke MPI_INIT and MPI_FINALIZE). You can use the "lamexec" program > > > > to run non-MPI programs over the lambooted nodes. > > > > > ----------------------------------------------------------------------------- > > > > Hello World! I am 0 of 1 > > > > > **************************************************************************************************** > > > > > > > > > > > > > ------------------------------------------------------- This SF.Net email is sponsored by xPML, a groundbreaking scripting language that extends applications into web and mobile media. Attend the live webcast and join the prime developer group breaking into this new coding territory! http://sel.as-us.falkag.net/sel?cmd=lnk&kid0944&bid$1720&dat1642 _______________________________________________ Oscar-users mailing list [email protected] https://lists.sourceforge.net/lists/listinfo/oscar-users
