Garrick Staples wrote:
On Thu, Jan 25, 2007 at 05:37:56AM +0530, S Ranjan alleged:
Garrick Staples wrote:
On Wed, Jan 24, 2007 at 08:51:11AM +0530, S Ranjan alleged:
Hi
I have torque pbs_server running on the headnode, which is also the
submit host. There are 32 other compute nodes, mentioned in
/var/spool/torque/server_priv/nodes file. There is a single queue at
present. Sometimes, mpi jobs requesting for 28/30 nodes, land up
running on the head node, though the head node is not a compute node at
all. netstat -anp shows several sockets being openend for the job, and
eventually the head node hangs up.
Appreciate any help/suggestion on this.
Which MPI? MPICH? I'd guess mpirun is using the default machinefile
that is created when mpich is built, and not the hostlist provided by
the PBS job.
Run mpirun with "-machinefile $PBS_NODEFILE" or use OSC's mpiexec
instead of mpirun: http://www.osc.edu/~pw/mpiexec/
_______________________________________________
mauiusers mailing list
[email protected]
http://www.supercluster.org/mailman/listinfo/mauiusers
_____________________________________________________________________
The mail server at Institute for Plasma Research has scanned this
email for Virus using ClamAV 0.88.4
_____________________________________________________________________
We are using Intel mpi 2.0. We are using mpiexec -n 28 ......
inside the pbs script.
However, for mpdboot (executable in the mpi 2.0 binary dir), we are
running it before running the pbs script. The exact syntax being used is
mpdboot -n 32 -f mpd.hosts --rsh=ssh -v
mpd.hosts file, residing in the user's home directory, contains the
names of the 32 compute nodes (excluding the head node).
<>
There is your problem, you want to use the list of nodes assigned to
your job. So you'll want something like this:
np=$(wc -l < $PBS_NODEFILE)
mpdboot -n $np -f $PBS_NODEFILE --rsh=ssh -v
But I still recommend using OSC's mpiexec instead.
<>
Hi
Using OSC's mpiexec, the mpi job starts and then gives the following
errors (when used without any mpdboot) with --comm=pmi option . When
used without --comm=pmi, the job just aborts complaining that it cannot
connect to mpd2.console (the same error that is generated if a mpi job
is launched without starting mpdboot).
I am using Intel MPI 2.0.
Thanks is advance for any help/suggestions
Sutapa
aborting job:
Fatal error in MPI_Barrier: Other MPI error, error stack:
MPI_Barrier(385): MPI_Barrier(MPI_COMM_WORLD) failed
MPIR_Barrier(75):
MPIC_Sendrecv(152):
MPIC_Wait(321):
MPIDI_CH3_Progress_wait(202): an error occurred while handling an event returned by MPIDU_Sock_Wait()
MPIDI_CH3I_Progress_handle_sock_event(1022): [ch3:sock] failed to connnect to
remote process 339.clustserver-spawn-0:2
MPIDU_Socki_handle_connect(780): connection failure
(set=0,sock=1,errno=111:Connection refused)
aborting job:
Fatal error in MPI_Barrier: Other MPI error, error stack:
MPI_Barrier(385): MPI_Barrier(MPI_COMM_WORLD) failed
MPIR_Barrier(75):
MPIC_Sendrecv(152):
MPIC_Wait(321):
MPIDI_CH3_Progress_wait(202): an error occurred while handling an event returned by MPIDU_Sock_Wait()
MPIDI_CH3I_Progress_handle_sock_event(1022): [ch3:sock] failed to connnect to
remote process 339.clustserver-spawn-0:3
MPIDU_Socki_handle_connect(780): connection failure
(set=0,sock=1,errno=111:Connection refused)
aborting job:
Fatal error in MPI_Barrier: Other MPI error, error stack:
MPI_Barrier(385): MPI_Barrier(MPI_COMM_WORLD) failed
MPIR_Barrier(75):
MPIC_Sendrecv(152):
MPIC_Wait(321):
MPIDI_CH3_Progress_wait(202): an error occurred while handling an event returned by MPIDU_Sock_Wait()
MPIDI_CH3I_Progress_handle_sock_event(461):
connection_recv_fail(1685):
MPIDU_Socki_handle_read(627): connection failure (set=0,sock=1,errno=104:Connection reset by peer)
aborting job:
Fatal error in MPI_Barrier: Other MPI error, error stack:
MPI_Barrier(385): MPI_Barrier(MPI_COMM_WORLD) failed
MPIR_Barrier(75):
MPIC_Sendrecv(152):
MPIC_Wait(321):
MPIDI_CH3_Progress_wait(202): an error occurred while handling an event returned by MPIDU_Sock_Wait()
MPIDI_CH3I_Progress_handle_sock_event(1022): [ch3:sock] failed to connnect to
remote process 339.clustserver-spawn-0:1
MPIDU_Socki_handle_connect(780): connection failure
(set=0,sock=1,errno=111:Connection refused)
newmpiexec: Warning: tasks 0-3 exited with status 13.
------------------------------------------------------------------------
_______________________________________________
mauiusers mailing list
[email protected]
http://www.supercluster.org/mailman/listinfo/mauiusers