Can you ssh to the other machine without a password? It looks like that may be the problem... You probably have to setup a private/public key pair.
-berk On Fri, Nov 7, 2008 at 3:43 AM, yewyong <[EMAIL PROTECTED]> wrote: > Sorry for the confusion. i'll try to explain more on my situation. > the first example show that the pvserver is "launchable" when it only > involves one machine as the pvserver. > i could connect to it using any binary version of paraview client from any > machine. > > at the second example, i'm trying to run a parallel server using 2 machines. > > added the mpi bin directory PATH in my bashrc file and came out with > following konsole lines: > >> >> [EMAIL PROTECTED]:~/installer/ParaView-3.4.0_build/bin> mpirun -hostfile >> ../../openmpi-1.2.8_build/etc/quadcore_frodo -np 6 ./pvserver >> --use-offscreen-rendering >> Password: >> >> [192.168.0.10][0,1,1][btl_tcp_endpoint.c:572:mca_btl_tcp_endpoint_complete_connect] >> connect() failed with errno=110 >> >> [192.168.0.10][0,1,3][btl_tcp_endpoint.c:572:mca_btl_tcp_endpoint_complete_connect] >> connect() failed with errno=110 > > then the machine just fails to give any response. > when pressed with "control c" the following shows: > >> >> mpirun: killing job... >> >> mpirun noticed that job rank 0 with PID 20731 on node 192.168.0.10 exited >> on signal 15 (Terminated). >> 5 additional processes aborted (not shown) > > Any ideas what going on? > > thanks, > > > > On Fri, Nov 7, 2008 at 12:08 AM, John M. Patchett <[EMAIL PROTECTED]> wrote: >> >> On Nov 6, 2008, at 8:34 AM, Moreland, Kenneth wrote: >> >> I am a bit confused by your question. Are you saying the first example >> works and the second does not? I am also confused by what you are trying to >> set up. Are you trying to run the server on one of your machines and the >> client on the other, or are you trying to run a parallel server using both >> machines? >> >> In the second example, it looks like your OpenMPI install is messed up. >> In the 4th line (1st error), it complains about not finding orted, which is >> a daemon that the OpenMPI version of mpirun uses to launch processes. It >> does not look like pvserver is being run at all. >> >> I think Ken is correct. It would appear that you don't have your bash >> environment set up correctly. You need to have in your .profile or .bashrc >> or .something that is read when OpenMPI launches your job, the path to the >> mpi bin directory. >> the LD_LIBRARY_PATH needs to be set to your OpenMPI lib directory. You >> can test that the PATH component is set correctly for your type of login by >> running a command like: >> ssh 192.168.0.16 which mpirun >> The LD_LIBRARY_PATH will become obvious once your pvserver executable is >> actually launched. >> (The man page for bash on your system should explain what order the dot >> files are read in for interactive and non interactive shells...) >> At the very very end of this mail I'm pasting the show output of a very >> (very) old openmpi environment modules ... >> Hope this helps, >> -John. >> >> >> I cannot claim to have a lot of experience in setting up MPI installs, >> especially ones on heterogeneous nodes like you are setting up. My advice >> would be to first make sure that you can launch a job "locally" on each >> machine, then try to launch a job only one the opposite computer (i.e. >> launch only on the 32 bit machine from the 64 bit machine and vice versa), >> and then finally launch a single job on both. >> >> -Ken >> >> >> On 11/6/08 1:32 AM, "yewyong" <[EMAIL PROTECTED]> wrote: >> >> Hi all, >> >> i managed to compile the paraview 3.4.0 in both a 64bit dual core machine >> and a 32bit quad core machine. >> having trouble starting the pvserver when linking 2 machines. Below are >> the konsole lines.. >> >> [EMAIL PROTECTED]:~/installer/ParaView-3.4.0_build/bin> >> ../../openmpi-1.2.8_build/bin/mpirun -hostfile >> ../../openmpi-1.2.8_build/etc/frodo -np 2 ./pvserver >> --use-offscreen-rendering >> Listen on port: 11111 >> Waiting for client... >> ^Cmpirun: killing job... >> >> mpirun noticed that job rank 0 with PID 27000 on node 192.168.0.16 >> <http://192.168.0.16> exited on signal 15 (Terminated). >> 1 additional process aborted (not shown) >> >> everything looks good when there is no linking of 2 machines. (this is the >> 64bit dual core machine) >> but, >> >> [EMAIL PROTECTED]:~/installer/ParaView-3.4.0_build/bin> >> ../../openmpi-1.2.8_build/bin/mpirun -hostfile >> ../../openmpi-1.2.8_build/etc/frodo_quadcore -np 6 ./pvserver >> --use-offscreen-rendering >> Password: >> bash: orted: command not found >> [frodo:27009] [0,0,0] ORTE_ERROR_LOG: Timeout in file >> base/pls_base_orted_cmds.c at line 275 >> [frodo:27009] [0,0,0] ORTE_ERROR_LOG: Timeout in file pls_rsh_module.c at >> line 1166 >> [frodo:27009] [0,0,0] ORTE_ERROR_LOG: Timeout in file errmgr_hnp.c at line >> 90 >> [frodo:27009] ERROR: A daemon on node 192.168.0.10 <http://192.168.0.10> >> failed to start as expected. >> [frodo:27009] ERROR: There may be more information available from >> [frodo:27009] ERROR: the remote shell (see above). >> [frodo:27009] ERROR: The daemon exited unexpectedly with status 127. >> [frodo:27009] [0,0,0] ORTE_ERROR_LOG: Timeout in file >> base/pls_base_orted_cmds.c at line 188 >> [frodo:27009] [0,0,0] ORTE_ERROR_LOG: Timeout in file pls_rsh_module.c at >> line 1198 >> -------------------------------------------------------------------------- >> mpirun was unable to cleanly terminate the daemons for this job. Returned >> value Timeout instead of ORTE_SUCCESS. >> -------------------------------------------------------------------------- >> >> i can't start the pvserver when involving 2 machines. >> 192.168.0.16 <http://192.168.0.16> = 64bit dual core >> 192.168.0.10 <http://192.168.0.10> = 32bit quad core >> >> Anyone faced this before? >> Thanks in advance for helping. >> >> Cheers, >> yewyong >> VRC UM >> >> >> >> >> **** Kenneth Moreland >> *** Sandia National Laboratories >> *********** >> *** *** *** email: [EMAIL PROTECTED] >> ** *** ** phone: (505) 844-8919 >> *** fax: (505) 845-0833 >> >> _______________________________________________ >> ParaView mailing list >> [email protected] >> http://www.paraview.org/mailman/listinfo/paraview >> >> [dq001 ~]$module show openmpi >> ------------------------------------------------------------------- >> /usr/local/Modules/modulefiles/openmpi/1.0.1curr: >> prepend-path PATH >> /usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/bin >> prepend-path MANPATH >> /usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/man >> prepend-path LD_LIBRARY_PATH >> /usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib >> prepend-path LIBRARY_PATH /usr/mellanox/lib >> prepend-path PATH /usr/mellanox/bin >> prepend-path CPLUS_INCLUDE_PATH >> /usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/include >> prepend-path C_INCLUDE_PATH >> /usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/include >> prepend-path LIBRARY_PATH >> /usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib >> prepend-path LIBRARY_PATH /usr/mellanox/lib >> setenv MPI_ROOT >> /usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install >> setenv LDLIBS -lmpi >> setenv PV_MPI_INCLUDE_PATH >> /usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/include/openmpi/ompi >> setenv PV_MPI_LIBRARY >> /usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib/libmpi.a >> setenv PV_MPI_EXTRA_LIBRARY >> /usr/lib64/libmtl_common.so;/usr/lib64/libvapi.so;/usr/lib64/libmpga.so;/usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib/libmpi_cxx.a;/usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib/libopal.a;/usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib/liborte.a;/usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib/libmpi.a;/usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib/libmpi_cxx.a;/usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib/libopal.a;/usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib/liborte.a;/usr/lib64/libutil.a >> /usr/lib64/libc.a >> setenv PV_MPI_NUMPROC_FLAG -np >> setenv PV_MPIRUN >> /usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/bin/mpirun >> ------------------------------------------------------------------- > > _______________________________________________ > ParaView mailing list > [email protected] > http://www.paraview.org/mailman/listinfo/paraview > > _______________________________________________ ParaView mailing list [email protected] http://www.paraview.org/mailman/listinfo/paraview
