interface.
Besides, "pineapple" hit a roadblock during the call and is a totally separate
discussion.
On Jun 13, 2012, at 7:03 AM, Richard Graham wrote:
> I would suggest exposing modex at the pineapple level, and not tie it to a
> particular instance of run-time inst
Don't agree here - the only synchronization point is the completion. Ibarrier
can't be completed until all have entered the barrier, but each process can
leave the ibarrier() call as soon as they want to.
Rich
-Original Message-
From: devel-boun...@open-mpi.org
Forget what I just posted - I looked at George's words, and not the code -
wait() is the synchronization point, so George's response is correct.
Rich
-Original Message-
From: devel-boun...@open-mpi.org [mailto:devel-boun...@open-mpi.org] On Behalf
Of George Bosilca
Sent: Friday, July
I do not see any resetting of sequence numbers. It has been a long time since
I have looked at the matching code, so don't know if the out-of-order handling
has been taken out. If not, the sequence number has to be dealt with in some
manner, or else there will be a gap in the arriving
removing a non-matched request
has no impact on the sequence number.
george.
On Jul 26, 2012, at 16:31 , Richard Graham wrote:
> I do not see any resetting of sequence numbers. It has been a long time
> since I have looked at the matching code, so don't know if the out-of-order
>
, at 11:31 PM, Richard Graham <richa...@mellanox.com> wrote:
> Brian,
> First - thanks. I am very happy this is proceeding.
> General question here - do you have any idea how much global state sits
> behind the current implementation ? What I am trying to gauge at what level
This check in was in error - I had not realized that the checkout was from
the 1.3 branch, so we will fix this, and put these into the trunk (1.4). We
are going to bring in some limited multi-cluster support - limited is the
operative word.
Rich
On 9/22/08 12:50 PM, "Jeff Squyres"
and has significant code consequences when we look at abnormal
> terminations, comm_spawn, etc.
>
> Thanks
> Ralph
>
> On Sep 22, 2008, at 11:26 AM, Richard Graham wrote:
>
>> This check in was in error - I had not realized that the checkout
>> was from
>> the 1.3
ney Sr, Kenneth D.
>>>>> >>>> Subject: Re: [OMPI devel] [OMPI svn] svn:open-mpi r19600
>>>>> >>>>
>>>>> >>>> The issue isn't with adding a string. The question is whether or not
>>>>> >>
I plan to attend.
Rich
On 10/31/08 12:03 PM, "Ralph Castain" wrote:
> Hello all
>
> Those of us who participated in the July ORTE meeting had so much fun,
> we decided it was worth doing again! :-)
>
> Seriously, there are design issues that would benefit by a face-to-
> face
Just a few comments:
- not sure what sort of alternative memory approach is being considered.
The current approach was selected for two reasons:
- If something like anonymous memory is being used, one can only inherit
access to the shared files, so one process needs
set up the shared
Agreed.
On 11/14/08 9:56 AM, "Ralph Castain" <r...@lanl.gov> wrote:
>
> On Nov 14, 2008, at 7:41 AM, Richard Graham wrote:
>
>> Just a few comments:
>>- not sure what sort of alternative memory approach is being considered.
>> The cu
Now that 1.3 will be released, we would like to go ahead with the plan to
move the btl¹s out of the MPI layer. Greg Koenig who is doing most of the
work has started a wiki page with details on the plans. Right now details
are sketchy, as Greg is digging through the code, and has only hand
eorge B have been working on some Fastpath code changes that we
>> should make sure neither project obliterates the other.
>>
>> --td
>>
>> Richard Graham wrote:
>>> Now that 1.3 will be released, we would like to go ahead with the
>>> plan to mov
Ls might have added calls to the notifier framework in their
>>> >> error paths.
>>> >> The notifier framework is currently in the ORTE layer... not sure
>>> >> if we could
>>> >> move it down to OPAL. Ralph, any thoughts
come from configure
>> > (i.e., opal/include/opal_config.h) and were not renamed back when we
>> > split the code base into OPAL, ORTE, and OMPI. I don't think we had
>> > a strong reason for not renaming them -- most could probably be
>> > renamed to OPAL_* -- we just didn't do it th
g reason for not renaming them -- most could probably be
>> > renamed to OPAL_* -- we just didn't do it then. Perhaps they can be
>> > changed during the BTL extraction process (I noted this on the wiki).
>> >
>> >
>> >
>> > On Dec 3,
sages (flowing across who knows what transport) is an unlikely thing
>> > to happen.
>> >
>> > Besides, one of the primary reasons for needing to call notifier is a
>> > failure in the btl - so relying on the btl to send the message is
>> > self-defeating.
>
ferate quickly, while having the error reporting
> mechanism right where the error occurs represents the minimal impact and
> maximum flexibility.
>
>>> >> more flexibility is obtained if the data is passed up the call stack, and
>>> handled by the layer that wants to.
t. The
>> > proposed approach contains a number of impacts that may be avoided
>> > with an alternative approach.
>> >
>> > Without such a meeting, I fear we are going to rapidly dissolve into
>> > email hell again.
>> >
>> > Ral
How about if we start on this over e-mail and phone ? A face-to-face
meeting is good, but I am already booked Jan 5-9, maybe 12-13, Jan 16-Feb
6th, and Feb 8-11. I would prefer not to tack on something at the end of
the MPI Forum meeting, as I will have been gone from home for most of the
month
>
> On 12/5/08 6:49 AM, "Terry D. Dontje" <terry.don...@sun.com> wrote:
>
> Richard Graham wrote:
> > Let me start the e-mail conversation, and see how far we get.
> >
> > Goal: The goal several of us have is to be able to use the btl’s
>
n-overlapping groups, and run-time support we want to bring into
OMPI is to support new
functionality. The main point is that this is not STCI vs. OMPI at all.
Rich
>
> My $0.0002 - hope it helps
> Ralph
>
>
> On Dec 4, 2008, at 6:00 PM, Richard Graham wrote:
>
> Let me start the e-mail conv
>> >
>
> On 12/12/08 8:21 PM, "Eugene Loh" <eugene@sun.com> wrote:
>
> Richard Graham wrote:
> Re: [OMPI devel] shared-memory allocations The memory allocation is intended
to take into account that two separate procs may be touching the same
Not needed now. Since we did not want to deal with trying to grow the
shared memory file after it's allocation, with all the required
synchronization, we allocated extra memory up front - for dynamic process
control. Since this has never been enabled, we really don't need this extra
memory.
First, the performance improvements look really nice.
A few questions:
- How much of an abstraction violation does this introduce ? This looks
like the btl needs to start “knowing” about MPI level semantics. Currently,
the btl purposefully is ulp agnostic. I ask for 2 reasons
- you
On 1/20/09 2:08 PM, "Eugene Loh" <eugene@sun.com> wrote:
> Richard Graham wrote:
>> Re: [OMPI devel] RFC: sm Latency First, the performance improvements look
>> really nice.
>> A few questions:
>> - How much of an abstraction violation doe
On 1/20/09 8:53 PM, "Jeff Squyres" wrote:
> This all sounds really great to me. I agree with most of what has
> been said -- e.g., benchmarks *are* important. Improving them can
> even sometimes have the side effect of improving real applications. ;-)
>
> My one big
BTW,
In the recvi function, do you first try to match off the unexpected list
before you try and match data in the fifo¹s ?
Rich
On 1/21/09 8:00 PM, "Eugene Loh" wrote:
> Ron Brightwell wrote:
>>
>>>
>>> If you poll only the queue that correspond to a posted
rules.
Rich
On 1/22/09 12:51 PM, "Eugene Loh" <eugene@sun.com> wrote:
> Richard Graham wrote:
>> Re: [OMPI devel] RFC: sm Latency In the recvi function, do you first try to
>> match off the unexpected list before you try and match data in the fifo¹s?
>
This should really be viewed as a code maintenance RFC. The reason this
came up in the first place is because we are investigating the btl move, but
these are really two very distinct issues. There are two bits of code that
have virtually the same functionality - they do have the same interface
be done.
>
> I think it primarily is a question for the Fortran folks to address -
> can they deal with Fortran limits in some other manner without making
> the code unmanageable and/or taking a performance hit?
>
> Ralph
>
>
> On Jan 30, 2009, at 2:40 PM, Richard Graham wr
Fortran.
>>>> >>>
>>>> >>> So two were created. Then the orte_bitmap_t was blown away at a
>>>> >>> later time when we removed the GPR as George felt it wasn't
>>>> >>> necessary (which was true). It was later reborn
In the pt-2-pt code, the default case should never be hit - it would be a
bug in the code. Don't know about other uses of the sm btl.
Rich
On 2/5/09 12:30 PM, "Eugene Loh" wrote:
> In btl_sm_component.c, mca_btl_sm_component_progress() polls on FIFOs.
> If it gets
I would leave the code alone. The intent was for (A), but it is not used
for that. It is not in the performance critical region, works correctly as
we use it today, and putting it back later on would be a hassle not needed.
Rich
On 2/5/09 2:41 PM, "Eugene Loh" wrote:
>
It really does not matter what one does with the sm sends that can't be
posted to the FIFO, as long as they are posted at some later time. The
current implementation generates does not rely on the ordering memory
provides, but generates a sequence number and uses this in the matching,
just like
I am assuming that by distributed OS you are referring to the changes that
we (not just ORNL) are trying to do. If this is the case, this is a
mischaracterization of the of out intentions. We have two goals
- To be able to use a different run-time than ORTE to drive Open MPI
- To use the
We should not pretend that threads work in the 1.2 code branch. Thread
safety has been
designed in, but we are just kicking off an effort to complete and verify
the thread
safety.
Rich
On 6/11/07 2:49 PM, "Paul H. Hargrove" wrote:
> If Jeff has the resources to run
This is good work, so I am happy to see it come over. My initial
understanding was that
there would be compile time protection for this. In the absence of this, I
think we need
to see performance data on a variety of communication substrates. It seems
like a latency
measurement is, perhaps,
On 8/13/07 12:34 PM, "Galen Shipman" wrote:
>
Ok here is the numbers on my machines:
0 bytes
mvapich with header caching: 1.56
mvapich without header caching: 1.79
ompi 1.2: 1.59
So on zero bytes ompi not so bad. Also we can see
Rolf,
Would it be better to put this parameter in the system configuration file,
rather than change the compile time option ?
Rich
On 8/27/07 3:10 PM, "Rolf vandeVaart" wrote:
> We are running into a problem when running on one of our larger SMPs
> using the latest
Gleb,
Are you looking at this ?
Rich
On 8/29/07 9:56 AM, "Gleb Natapov" wrote:
> On Wed, Aug 29, 2007 at 04:48:07PM +0300, Gleb Natapov wrote:
>> Is this trunk or 1.2?
> Oops. I should read more carefully :) This is trunk.
>
>>
>> On Wed, Aug 29, 2007 at 09:40:30AM
If you are going to look at it, I will not bother with this.
Rich
On 8/29/07 10:47 AM, "Gleb Natapov" <gl...@voltaire.com> wrote:
> On Wed, Aug 29, 2007 at 10:46:06AM -0400, Richard Graham wrote:
>> Gleb,
>> Are you looking at this ?
> Not today. And I
We are looking at making some changes to the ompi free list in ompi/class/
ompi_free_list.[c,h] , and are trying to decide if to go ahead with an
interface change that will allow separate control over alignment of the frag
and payload data structures. We are aware of several implementations of
FYI.
Rich
-- Forwarded Message
From: Richard Graham <rlgra...@ornl.gov>
Reply-To: <mpi...@mpi-forum.org>
List-Post: devel@lists.open-mpi.org
Date: Fri, 5 Oct 2007 03:55:27 -0400
To: <mpi...@mpi-forum.org>
Conversation: Meeting at SC'07
Subject: Re: Meeting at SC'07
I will
One of the assumptions about the MTL¹s is that only a given MTL can handle
the message
matching for communications. This is done to accommodate mpi-like network
stack that
also handle the MPI message matching, which often do not expose their
internal data
structures used for matching. Open
What: Change the mpicc/mpicxx/mpif77/mpif90 from being binaries to being
shell scripts
Why: Our build environment assumes that wrapper compilers will use the same
binary format that the Open MPI libraries do. In cross-compile environment,
the MPI wrapper compilers will run on the front-end and
-- Forwarded Message
From: Richard Graham <rlgra...@ornl.gov>
Reply-To: "mpi...@mpi-forum.org" <mpi...@mpi-forum.org>
List-Post: devel@lists.open-mpi.org
Date: Fri, 12 Oct 2007 12:48:04 -0400
To: "mpi...@mpi-forum.org" <mpi...@mpi-forum.org>
Conversation
-- Forwarded Message
From: Richard Graham <rlgra...@ornl.gov>
List-Post: devel@lists.open-mpi.org
Date: Wed, 12 Sep 2007 19:53:19 -0400
Conversation: [RFC] Proposed changes to ompi_free_list
Subject: [devel-core] [RFC] Proposed changes to ompi_free_list
Proposed changes to the ompi_free_list:
Please c
I have just gone through and re-implemented the changes ompi_free_list_t in
the trunk, and have changed all instances of ompi_free_list_init() to
ompi_free_list_init_new() (keeping the old version around for a while). I
have tested this with ob1 and dr (the system I use for cm is not available),
ry running a nontrivial omb to check?
>
> -jms
> Sent from my PDA
>
> -Original Message-
> From: Richard Graham [mailto:rlgra...@ornl.gov]
> Sent: Friday, November 02, 2007 02:07 PM Eastern Standard Time
> To: Open MPI Developers
> Subject:Re: [OMPI d
On 11/8/07 4:03 AM, "Gleb Natapov" <gl...@voltaire.com> wrote:
> On Wed, Nov 07, 2007 at 11:25:43PM -0500, Patrick Geoffray wrote:
>> Richard Graham wrote:
>>> The real problem, as you and others have pointed out is the lack of
>>> predictable time sl
-- Forwarded Message
From: Richard Graham <rlgra...@ornl.gov>
Reply-To: <mpi...@mpi-forum.org>
List-Post: devel@lists.open-mpi.org
Date: Fri, 16 Nov 2007 23:21:16 -0500
To: <mpi...@mpi-forum.org>
Conversation: Follow up on the MPI Forum meeting
Subject: [mpi-21] Follow u
Any suggestions on ISV's that should be notified about the MPI 2.1+ effort
that is being started ? Any vendors that may have been missed ?
Rich
One question there is a mention a new pml that is essentially CM+matching.
Why is this no just another instance of CM ?
Rich
On 11/26/07 7:54 PM, "Jeff Squyres" wrote:
> OMPI OF Pow Wow Notes
> 26 Nov 2007
>
>
Gleb,
I would suggest that before this is checked in this be tested on a system
that has N-way network parallelism, where N is as large as you can find.
This is a key bit of code for MPI correctness, and out-of-order operations
will break it, so you want to maximize the chance for such
test.
>>> Good Idea I'll try this. BTW I thing the reason for such a high rate of
>>> reordering in UD is that it polls for MCA_BTL_UD_NUM_WC completions
>>> (500) and process them one by one and if progress function is called
>>> recursively next 500 comp
e:
>>
>>> On Wed, 12 Dec 2007, Gleb Natapov wrote:
>>>
>>>> On Wed, Dec 12, 2007 at 03:46:10PM -0500, Richard Graham wrote:
>>>>> This is better than nothing, but really not very helpful for
>>>>> looking at the
>>>>>
4:10 PM, Brian W. Barrett wrote:
>
>> > On Wed, 12 Dec 2007, Gleb Natapov wrote:
>> >
>>> >> On Wed, Dec 12, 2007 at 03:46:10PM -0500, Richard Graham wrote:
>>>> >>> This is better than nothing, but really not very helpful for
>>>> >>>
.
Rich
On 12/14/07 2:20 AM, "Gleb Natapov" <gl...@voltaire.com> wrote:
> On Thu, Dec 13, 2007 at 06:16:49PM -0500, Richard Graham wrote:
>> The situation that needs to be triggered, just as George has mentions, is
>> where we have a lot of unexpected messa
Are the group operations in ORTE (I assume this is what the grpcomm
component does) available to subsets of a job, or do all procs in the
orte_jobid_t need to invoke this ?
Thanks,
Rich
FYI, About six months ago several of us spent some time coming up with a
plan to deal with the latency problems in Open MPI. George went ahead and
has been implementing the send side changes of this optimization over the
last several months, but has not had time to get to the receive side. Galen
I am running into a situation where I am trying to deliver a signal to the
mpi procs (sigusr2). I deliver this to mpirun, which propagates it to the
mpi procs, but then proceeds to kill the children. Is there an easy way
that I can get around this ? I am using this mechanism in a situation
lso get passed to
>> the user processes, and let them decide what to do with the signals
>> themselves.
>>
>> SGE needed this so the job kill or job suspension notification to work
>> properly since they would send a SIGUSR1/2 to mpirun. I believe this is
&g
On 4/8/08 2:19 PM, "Ralph H Castain" <r...@lanl.gov> wrote:
>
>
>
> On 4/8/08 12:10 PM, "Pak Lui" <pak@sun.com> wrote:
>
>> Richard Graham wrote:
>>> What happens if I deliver sigusr2 to mpirun ? What I observe (for both
Ralph,
Thanks for looking into this. I do not think that the behaviour needs to
change - it is correct. However, for some reason this is not how things
were running for me - I wander what the difference is. I worked around
this by getting the pid's of the mpi processes, and delivered the
Is there a way to trick ompi/orte into thinking that a single node is
actually a collection of several smp nodes interconnected with tcp ? If so,
can someone give me a hint how to set this up ? I want to create a hierarchy
on my laptop for testing purposes.
Thanks,
Rich
Brad,
Do you want these for bug fixes too ?
Rich
On 5/20/08 5:53 PM, "Brad Benton" wrote:
> All:
>
> In order to better track changes on the trunk until we branch for 1.3, we (the
> release managers) would like to ask that all trunk checkins have corresponding
>
Thanks,
Rich
On 5/20/08 10:37 PM, "Brad Benton" <bradford.ben...@gmail.com> wrote:
>
>
> 2008/5/20 Richard Graham <rlgra...@ornl.gov>:
>> Brad,
>> Do you want these for bug fixes too ?
>
> I think that it's okay to check in small bug fix
No need for SM2 right now. We need a change to the communicator, before we
can bring this over, and have just gotten around to addressing this yet.
Rich
On 7/21/08 3:40 PM, "Jeff Squyres" wrote:
> Should these 2 components be in v1.3?
>
> --
> Jeff Squyres
> Cisco
70 matches
Mail list logo