On 2/7/07 12:07 PM, "Heywood, Todd" <heyw...@cshl.edu> wrote:
> Hi Ralph,
Patience is not an issue since I have a workaround (a locally
> authenticated user), and other users are not running large enough MPI jobs to
> hit this problem.
I'm a bit confused now though. I thought that setting this
> switch would set off 50 ssh sessions at a time, or 50 connections to slapd.
> I.e. a second group of 50 connections wouldn't initiate until the first group
> "closed" their sessions, which should be reflected by a corresponding decrease
> in the number of established connections for slapd. So my conclusion was that
> no sessions are "closing".
The way the rsh launcher works is to fork/exec num_concurrent rsh/ssh
sessions, and watch as each one "closes" the connection back to the HNP.
When that block has cleared, we then begin launching the next one. Note that
we are talking here about closure of stdin/stdout connections - i.e., the
orteds "daemonize" themselves after launch, thus severing their stdin/stdout
relationship back to the HNP.
It is possible that this mechanism isn't actually limiting the launch rate -
e.g., the orteds may daemonize themselves so quickly that the block launch
doesn't help. In a soon-to-come future version, we won't use that mechanism
for determining when to launch the next block - my offer of a patch was to
give you that new version now, modify it to more explicitly limit launch
rate, and see if that helped. I'll try to put that together in the next week
or so.
Since you observed that the pls_rsh_num_concurrent option had no impact on
the *rate* at which we launched, that would indicate that either the slapd
connection isn't bottlenecking - the time to authenticate is showing as
independent of the rate at which we are hitting the slapd - or we are not
rate limiting as we had hoped. Hence my comment that it may not look like a
rate issue.
As I said earlier, we have never tested this with LDAP. From what I
understand of LDAP (which is limited, I admit), the ssh'd process (the orted
in this case) forms an authentication connection back to the slapd. It may
not be possible to sever this connection during the life of that process.
There typically are limits on the number of simultaneous LDAP sessions a
single user can have open - mainly for security reasons - so that could also
be causing the problem.
Given that you also observed that the total number of nodes we could launch
upon was the same regardless of the rate, it could be that we are hitting
the LDAP session limit.
Logging may have a broader impact than just slapd response rate - I honestly
don't know.
Hope that helps - I'll pass along that patch as soon as I can.
Ralph
>
There's also the observation that when slapd is
> slowed down by logging (extensively), things hang with fewer number of
> established connections (open ssh sessions). I don't see how this fitrs with a
> total number of connections limitation.
> Thanks,
Todd
-----Original
> Message-----
From: users-boun...@open-mpi.org
> [mailto:users-boun...@open-mpi.org] On Behalf Of Ralph Castain
Sent:
> Wednesday, February 07, 2007 1:28 PM
To: Open MPI Users
Subject: Re: [OMPI
> users] large jobs hang on startup (deadlock?)
Hi Todd
I truly appreciate
> your patience. If the rate was the same with that switch
set, then that would
> indicate to me that we aren't having trouble getting
through the slapd - it
> probably isn't a problem with how hard we are driving
it, but rather with the
> total number of connections being created.
Basically, we need to establish one
> connection/node to launch the orteds
(the app procs are just fork/exec'd by
> the orteds so they shouldn't see the
slapd).
The issue may have to do with
> limits on the total number of LDAP
authentication connections allowed for one
> user. I believe that is settable,
but will have to look it up and/or ask a few
> friends that might know.
I have not seen an LDAP-based cluster before (though
> authentication onto the
head node of a cluster is frequently handled that
> way), but that doesn't
mean someone hasn't done it.
Again, appreciate the
> patience.
Ralph
On 2/7/07 10:28 AM, "Heywood, Todd" <heyw...@cshl.edu>
> wrote:
> Hi Ralph,
Unfortunately, adding "-mca pls_rsh_num_concurrent 50" to
> mpirun
> (with just -np and -hostfile) has no effect. The number of
> established
> connections for slapd grows to the same number at the same rate
> as without it.
>
BTW, I upgraded from 1.2b2 to
> 1.2b3
Thanks,
TOdd
-----Original
> Message-----
From:
> users-boun...@open-mpi.org
> [mailto:users-boun...@open-mpi.org] On Behalf Of
> Ralph Castain
Sent: Tuesday,
> February 06, 2007 6:48 PM
To: Open MPI
> Users
Subject: Re: [OMPI users] large
> jobs hang on startup (deadlock?)
Hi
> Todd
Just as a thought - you could try
> not using --debug-daemons or -d
> and
instead setting "-mca
> pls_rsh_num_concurrent 50" or some such small
> number.
This will tell the
> system to launch 50 ssh calls at a time, waiting
> for each
group to complete
> before launching the next. You can't use it
> with
--debug-daemons as that
> option prevents the ssh calls from "closing" so
> that
you can get the output
> from the daemons. You can still launch as big a
> job
as you like - we'll just
> do it 50 ssh calls at a time.
If we are truly
> overwhelming the slapd, then
> this should alleviate the
problem.
Let me know
> if you get to try
> it...
Ralph
On 2/6/07 4:05 PM, "Heywood, Todd"
> <heyw...@cshl.edu> wrote:
>
> Hi Ralph,
It looks that way. I created a user
> local to each node, with
> local
> authentication via /etc/passwd and
> /etc/shadow, and OpenMPI scales up
> just
> fine for that.
I know this is an
> OpenMPI list, but does anyone know
> how
> common or uncommon LDAP-based
> clusters are? I would have thought this
> issue
> would have arisen elsewhere,
> but Googling MPI+LDAP (and similar)
> doesn't turn
> up much.
I'd certainly
> be willing to test any patch.
>
> Thanks.
Todd
-----Original
> Message-----
From: users-boun...@open-mpi.org
>
>
> [mailto:users-boun...@open-mpi.org] On Behalf Of Ralph H Castain
Sent:
>
>
> Tuesday, February 06, 2007 9:54 AM
To: Open MPI Users
>
>
> <us...@open-mpi.org>
Subject: Re: [OMPI users] large jobs hang on startup
>
>
> (deadlock?)
It sounds to me like we are probably overwhelming your slapd
> -
>
> your test
would seem to indicate that slowing down the slapd makes us
> fail
>
> even with
smaller jobs, which tends to support that idea.
We
> frankly
> haven't
> encountered that before since our rsh tests have all
> been
done using
> non-LDAP
> authentication (basically, we ask that you setup
> rsh
> to
auto-authenticate on
> each node). It sounds like we need to add an
> ability
> to
slow down so that the
> daemon doesn't "fail" due to
> authentication
> timeout
and/or slapd rejection due
> to the queue being
> full.
This may take a
> little time to fix due to other
> priorities, and
> will almost
certainly have
> to be released in a subsequent
> 1.2.x version.
> Meantime, I'll
let you know
> when I get something to test -
> would you be
> willing to give it
a shot if I
> provide a patch? I don't have
> access to an
> LDAP-based system.
Ralph
On
> 2/6/07 7:44 AM, "Heywood, Todd"
>
> <heyw...@cshl.edu> wrote:
> Hi
> Ralph,
Thanks for the reply. This is a
> tough
> one. It is OpenLDAP. I had
>
> thought that I might be hitting a file
> descriptor
> limit for slapd (LDAP
>
> daemon), which ulimit -n does not
> effect (you have to
> rebuild LDAP with a
>
> different FD_SETSIZE variable).
> However, I simply turned
> on more
> expressive
> logging to /var/log/slapd,
> and that resulted in smaller
> jobs
> (which
> successfully ran before)
> hanging. Go figure. It appears that
>
> daemons are up
> and running (from
> ps), and everything hangs in MPI_Init.
>
> Ctl-C
> gives
[blade1:04524]
> ERROR: A daemon on node blade26 failed to
> start
> as
>
> expected.
[blade1:04524] ERROR: There may be more information
> available
>
>
> from
[blade1:04524] ERROR: the remote shell (see
> above).
[blade1:04524]
>
> ERROR:
> The daemon exited unexpectedly with status
> 255.
I'm interested
> in
> any
> suggestion, semi-fixes, etc. which might help
> get to the bottom
> of this.
> Right
> now: whether the daemons are indeed up
> and running, or if
> there are
> some that
> are not (causing MPI_Init to
>
>
> hang).
Thanks,
Todd
-----Original
> Message-----
From:
>
>
> users-boun...@open-mpi.org
> [mailto:users-boun...@open-mpi.org] On Behalf
>
> Of
> Ralph H Castain
Sent:
> Tuesday, February 06, 2007 8:52 AM
To: Open
> MPI
>
> Users
> <us...@open-mpi.org>
Subject: Re: [OMPI users] large jobs hang
> on
>
> startup
> (deadlock?)
Well, I can't say for sure about LDAP. I did a
> quick
>
> search and
> found two
things:
1. there are limits imposed in LDAP
> that may
>
> apply to your
> situation, and
2. that statement varies
> tremendously
>
> depending upon the
> specific LDAP
implementation you are
> using
I would
>
> suggest you see which LDAP
> you are using and contact
> the
>
> respective
organization to ask if they do have
> such a limit, and if
> so,
> how
> to adjust
it.
It sounds like maybe we are
> hitting the LDAP
> server
> with too
> many requests
too rapidly. Usually, the issue
> is not
> starting
> fast enough,
> so this is a
new one! We don't currently check to
>
> see if
> everything started
> up okay, so
that is why the processes might hang
> -
> we
> hope to fix that soon.
> I'll have
to see if there is something we
> can do to
>
> help alleviate such
> problems -
might not be in time for the
> 1.2 release,
> but
> perhaps it will
> make a
subsequent "fix" or, if you
> are
> willing/interested, I
> could provide
> it to
you as a "patch" you could
> use
> until a later official
>
> release.
Meantime, you might try upgrading
> to
> 1.2b3 or even a nightly
> release
> from
the trunk. There are known
> problems
> with 1.2b2 (which is why
> there is a
> b3
and soon to be an rc1),
> though I
> don't think that will be the
> problem
> here.
At the least, the
> nightly trunk
> has a much better response to
> ctrl-c in
> it.
Ralph
On
> 2/5/07 9:50 AM,
> "Heywood, Todd" <heyw...@cshl.edu>
> wrote:
>
> Hi Ralph,
>
>
> Thanks for
> the reply. The OpenMPI version is 1.2b2
> (because I
> would
> like to
>
> integrate it with SGE).
>
> Here is what is
> happening:
>
>
>
> (1)
> When I run with debug-daemons (but WITHOUT d), I
> get ³Daemon>
>
> [0,0,27]
> checking in as pid 7620 on host blade28² (for example)
> messages
> for
>
> most
> but not all of the daemons that should be started up,
> and
> then it
> hangs.
>
> I also notice ³reconnecting to LDAP server² messages in
>
> various
>
>
> /var/log/secure files, and cannot login while things are hung
>
> (with ³su:>
>
> pam_ldap: ldap_result Can't contact LDAP server² in
>
> /var/log/messages).
> So
>
> apparently LDAP hits some limit to opening ssh
>
> sessions, and I¹m not
> sure
> how
> to address this.
> (2) When I run
> with
> debug-daemons AND
> the debug
> option d, all daemons
> start start
> up and
> check-in, albeit
> slowly (debug
> must slow things down so
> LDAP
> can handle
> all the
> requests??). Then
> apparently, the cpi process is
>
> started for each
> task
> but it then hangs:
>
>
> [blade1:23816] spawn: in
>
> job_state_callback(jobid
> = 1, state = 0x4)
>
> [blade1:23816] Info: Setting
> up
> debugger process table
> for applications
>
> MPIR_being_debugged = 0
>
>
> MPIR_debug_gate = 0
>
> MPIR_debug_state = 1
>
> MPIR_acquired_pre_main =
> 0
>
> MPIR_i_am_starter =
> 0
> MPIR_proctable_size =
> 800
>
> MPIR_proctable:
>
> (i, host, exe, pid) =
> (0, blade1,
>
> /home4/itstaff/heywood/ompi/cpi, 24193)
>
> Š
> Š(i, host, exe,
> pid) =
> (799,
> blade213, /home4/itstaff/heywood/ompi/cpi,
> 4762)
>
> A ³ps²
> on
> the head node
> shows 200 open ssh sessions, and 4 cpi
> processes doing
>
>
> nothing. A ^C gives
> this:
>
> mpirun: killing job...
>
>
>
>
>
>
> --------------------------------------------------------------------------
>
>
>
>
> WARNING: A process refused to die!
>
> Host: blade1
> PID: 24193
>
>
>
>
> This
> process may still be running and/or consuming resources.
>
>
>
>
>
>
>
>
> Still got a ways to go, but any ideas/suggestions are
> welcome!
>
>
>
> Thanks,
>
>
> Todd
>
>
>
>
> From:
> users-boun...@open-mpi.org
>
>
> [mailto:users-boun...@open-mpi.org] On
> Behalf
> Of Ralph Castain
> Sent:
>
>
> Friday, February 02, 2007 5:20 PM
>
> To: Open MPI Users
> Subject: Re:
> [OMPI
>
> users] large jobs hang on
> startup (deadlock?)
>
> Hi Todd
>
> To
> help us
>
> provide advice, could
> you tell us what version of OpenMPI you
> are
> using?
>
>
>
> Meantime, try
> adding ³-mca pls_rsh_num_concurrent 200²
> to your mpirun
>
> command
> line.
> You can up the number of concurrent daemons
> we launch to
>
> anything your
>
> system will support basically, we limit the
> number only
>
> because some
> systems
> have limits on the number of ssh calls
> we can have
>
> active at
> any one time.
> Because we hold stdio open when
> running with
>
>
> ‹debug-daemons, the number of
> concurrent daemons must match
> or exceed
> the
>
> number of nodes you are trying to
> launch on.
>
> I have a
> ³fix²
> in the
>
> works that will help relieve some of that restriction,
> but
> that
> won¹t come
>
> out until a later release.
>
> Hopefully, that will allow
>
> you to obtain
> more
> debug info about why/where
> things are hanging.
>
>
>
> Ralph
>
>
> On
> 2/2/07
> 11:41 AM, "Heywood, Todd" <heyw...@cshl.edu>
>
> wrote:
> I have OpenMPI
> running
> fine for a small/medium number of tasks
>
> (simple hello
> or cpi
> program). But
> when I try 700 or 800 tasks, it
>
> hangs, apparently on
>
> startup. I think this
> might be related to LDAP,
>
> since if I try to log into
> my
> account while the
> job is hung, I get
> told
> my username doesn¹t exist.
> However,
> I tried adding
> debug to the
> mpirun,
> and got the same sequence of
> output as
> for successful
> smaller
> runs,
> until it hung again. So I added
> -debug-daemons
> and got this
>
> (with an
> exit, i.e. no hanging):
> Š
>
> [blade1:31733] [0,0,0] wrote
> setup
> file
>
>
>
>
>
> --------------------------------------------------------------------------
>
>
>
>
> The rsh launcher has been given a number of 128 concurrent daemons
> to
>
>
> launch
> and is in a debug-daemons option. However, the total number
> of
>
>
> daemons to
> launch (200) is greater than this value. This is a
> scenario
>
> that
> will cause
> the system to deadlock.
>
> To avoid
> deadlock, either
>
> increase the number of
> concurrent daemons, or
> remove
> the debug-daemons
>
> flag.
>
>
>
>
> --------------------------------------------------------------------------
>
>
>
>
> [blade1:31733] [0,0,0] ORTE_ERROR_LOG: Fatal in file
>
>
>
>
> ../../../../../orte/mca/rmgr/urm/
> rmgr_urm.c at line 455
>
>
> [blade1:31733]
>
> mpirun: spawn failed with errno=-6
> [blade1:31733]
>
> sess_dir_finalize: proc
>
> session dir not empty - leaving
>
> Any ideas
> or
> suggestions
> appreciated.
>
>
> Todd Heywood
>
>
>
>
>
>
>
>
>
>
> _______________________________________________
> users mailing
> list
>
>
>
> us...@open-mpi.org
>
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>
>
>
>
>
>
> _______________________________________________
> users mailing
>
> list
>
>
> us...@open-mpi.org
>
>
>
>
> http://www.open-mpi.org/mailman/listinfo.cgi/users
_______________________
>
>
>
> ________________________
users mailing
>
>
>
> list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users
__
>
>
>
> _____________________________________________
users mailing
>
>
>
> list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>
> _______________________________________________
users mailing
>
>
> list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users
__
>
>
> _____________________________________________
users mailing
>
>
> list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users
>
> _______________________________________________
users mailing
>
> list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users
__
>
> _____________________________________________
users mailing
>
> list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users
> _______________________________________________
users mailing
> list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users
__
> _____________________________________________
users mailing
> list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users