Re: [OMPI devel] possible bug in 1.3.2 sm transport

2009-05-18 Thread Bryan Lally
Eugene Loh wrote: Ralph Castain wrote: Hi Bryan I have seen similar issues on LANL clusters when message sizes were fairly large. How big are your buffers when you call Allreduce? Can you send us your Allreduce call params (e.g., the reduce operation, datatype, num elements)? If you

Re: [OMPI devel] possible bug in 1.3.2 sm transport

2009-05-18 Thread Eugene Loh
Ralph Castain wrote: Hi Bryan I have seen similar issues on LANL clusters when message sizes were fairly large. How big are your buffers when you call Allreduce? Can you send us your Allreduce call params (e.g., the reduce operation, datatype, num elements)? If you don't want to send

Re: [OMPI devel] RFC: Warn user about deprecated MPI functionalityand "wrong" compiler usage

2009-05-18 Thread Rainer Keller
Dear Paul & all, On Monday 18 May 2009 03:19:48 pm Paul H. Hargrove wrote: > IMHO there are two distinct issues being entangled here. > 1) Flagging deprecated functionality > 2) Informing the user about a change of compiler (possibly as an > #error or #warning) > > I understand why solving #1

Re: [OMPI devel] RFC: Warn user about deprecated MPI functionalityand "wrong" compiler usage

2009-05-18 Thread Paul H. Hargrove
IMHO there are two distinct issues being entangled here. 1) Flagging deprecated functionality 2) Informing the user about a change of compiler (possibly as an #error or #warning) I understand why solving #1 requires detecting the compiler change to avoid a "bad attribute" (see BACKGROUND,

Re: [OMPI devel] ompi-ps broken or just changed?

2009-05-18 Thread Ralph Castain
Bizarre - it works perfectly for me. Is it possible you have stale libraries around? Or are attempting to connect to older versions of mpirun? You might also try cleaning out any old session dirs just to be safe - my best guess is that you are connecting to an older version of mpirun and

Re: [OMPI devel] ompi-ps broken or just changed?

2009-05-18 Thread Ashley Pittman
Ralph, This patch fixed it, num_nodes was being used initialised and hence the client was getting a bogus value for the number of nodes. Ashley, On Mon, 2009-05-18 at 10:09 +0100, Ashley Pittman wrote: > No joy I'm afraid, now I get errors when I run it. This is a single > node job run with

Re: [OMPI devel] RFC: Warn user about deprecated MPI functionalityand "wrong" compiler usage

2009-05-18 Thread Jeff Squyres
Agreed. Being able to handle such scenarios properly is one of the reasons that Rainer and I are iterating on this in a mercurial branch. On May 18, 2009, at 7:39 AM, Brian Barrett wrote: I think care must be taken on this front. While I know we don't like to admit it, there is no reason

Re: [OMPI devel] RFC: Warn user about deprecated MPI functionality and "wrong" compiler usage

2009-05-18 Thread Brian Barrett
I think care must be taken on this front. While I know we don't like to admit it, there is no reason the C compilers have to match, and indeed good reasons they might not. For example, at LANL, we frequently compiled OMPI with GCC, then fixed up the wrapper compilers to use Icc or

Re: [OMPI devel] ompi-ps broken or just changed?

2009-05-18 Thread Ashley Pittman
No joy I'm afraid, now I get errors when I run it. This is a single node job run with the command line "mpirun -n 3 ./a.out". I've attached the strace output and gzipped /tmp files from the machine. Valgrind on the opmi-ps process doesn't show anything interesting. [alpha:29942]

[OMPI devel] RFC: Warn user about deprecated MPI functionality and "wrong" compiler usage

2009-05-18 Thread Rainer Keller
What: Warn user about deprecated MPI functionality and "wrong" compiler usage Why: Because deprecated MPI functions, are ... deprecated Where: On trunk When: Apply on trunk before branching for v1.5 (it is user-visible) Timeout: 1 weeks - May 26, 2009 after the teleconf.