Re: [OMPI users] Problems building Open MPI 1.4.1 with Pathscale

2010-02-09 Thread Iain Bason
Well, I am by no means an expert on the GNU-style asm directives.  I  
believe someone else (George Bosilca?) tweaked what I had suggested.


That being said, I think the memory "clobber" is harmless.

Iain

On Feb 9, 2010, at 5:51 PM, Jeff Squyres wrote:


Iain did the genius for the new assembly.  Iain -- can you respond?


On Feb 9, 2010, at 5:44 PM, Mostyn Lewis wrote:


The old opal_atomic_cmpset_32 worked:

static inline int opal_atomic_cmpset_32( volatile int32_t *addr,
   unsigned char ret;
   __asm__ __volatile__ (
   SMPLOCK "cmpxchgl %1,%2   \n\t"
   "sete %0  \n\t"
   : "=qm" (ret)
   : "q"(newval), "m"(*addr), "a"(oldval)
   : "memory");

   return (int)ret;
}

The new opal_atomic_cmpset_32 fails:

static inline int opal_atomic_cmpset_32( volatile int32_t *addr,
int32_t oldval, int32_t  
newval)

{
   unsigned char ret;
   __asm__ __volatile__ (
   SMPLOCK "cmpxchgl %3,%4   \n\t"
   "sete %0  \n\t"
   : "=qm" (ret), "=a" (oldval), "=m" (*addr)
   : "q"(newval), "m"(*addr), "1"(oldval)
   return (int)ret;
}

**However** if you put back the "clobber" for memory line (3rd :),  
it works:


static inline int opal_atomic_cmpset_32( volatile int32_t *addr,
int32_t oldval, int32_t  
newval)

{
   unsigned char ret;
   __asm__ __volatile__ (
   SMPLOCK "cmpxchgl %3,%4   \n\t"
   "sete %0  \n\t"
   : "=qm" (ret), "=a" (oldval), "=m" (*addr)
   : "q"(newval), "m"(*addr), "1"(oldval)
   : "memory");

   return (int)ret;
}

This works in a test case for pathcc, gcc, icc, pgcc, SUN studio cc  
and open64 (pathscale

lineage - which also fails with 1.4.1).
Also the SMPLOCK above is defined as "lock; " - the ";" is a GNU as  
statement delimter - is

that right? Seems to work with/without the ";".


Also, a question - I see you generate via perl another "lock" asm  
file which you put into
opal/asm/generated/ and stick  
into libasm - what you
generate there for whatever usage hasn't changed 1.4->1.4.1->svn  
trunk?


DM

On Tue, 9 Feb 2010, Jeff Squyres wrote:

Perhaps someone with a pathscale compiler support contract can  
investigate this with them.


Have them contact us if they want/need help understanding our  
atomics; we're happy to explain, etc. (the atomics are fairly  
localized to a small part of OMPI).




On Feb 9, 2010, at 11:42 AM, Mostyn Lewis wrote:


All,

FWIW, Pathscale is dying in the new atomics in 1.4.1 (and svn  
trunk) - actually looping -


from gdb:

opal_progress_event_users_decrement () at ../.././opal/include/ 
opal/sys/atomic_impl.h:61
61 } while (0 == opal_atomic_cmpset_32(addr, oldval,  
oldval - delta));

Current language:  auto; currently asm
(gdb) where
#0  opal_progress_event_users_decrement () at ../.././opal/ 
include/opal/sys/atomic_impl.h:61

#1  0x0001 in ?? ()
#2  0x2aec4cf6a5e0 in ?? ()
#3  0x00eb in ?? ()
#4  0x2aec4cfb57e0 in ompi_mpi_init () at ../.././ompi/ 
runtime/ompi_mpi_init.c:818

#5  0x7fff5db3bd58 in ?? ()
Backtrace stopped: previous frame inner to this frame (corrupt  
stack?)

(gdb) list
56  {
57 int32_t oldval;
58
59 do {
60oldval = *addr;
61 } while (0 == opal_atomic_cmpset_32(addr, oldval,  
oldval - delta));

62 return (oldval - delta);
63  }
64  #endif  /* OPAL_HAVE_ATOMIC_SUB_32 */
65
(gdb)

DM

On Tue, 9 Feb 2010, Jeff Squyres wrote:

FWIW, I have had terrible luck with the patschale compiler over  
the years.  Repeated attempts to get support from them -- even  
when I was a paying customer -- resulted in no help (e.g., a  
pathCC bug with the OMPI C++ bindings that I filed years ago was  
never resolved).


Is this compiler even supported anymore?  I.e., is there a  
support department somewhere that you have a hope of getting any  
help from?


I can't say for sure, of course, but if MPI hello world hangs,  
it smells like a compiler bug.  You might want to attach to  
"hello world" in a debugger and see where it's hung.  You might  
need to compile OMPI with debugging symbols to get any  
meaningful information.


** NOTE: My personal feelings about the pathscale compiler suite  
do not reflect anyone else's feelings in the Open MPI  
community.  Perhaps someone could change my mind someday, but  
*I* have personally given up on this compiler.  :-(



On Feb 8, 2010, at 2:38 AM, Rafael Arco Arredondo wrote:


Hello,

It does work with version 1.4. This is the hello world that  
hangs with

1.4.1:

#include 
#include 

int main(int argc, char **argv)
{
 int node, size;

 MPI_Init(,);
 

Re: [OMPI users] Problems building Open MPI 1.4.1 with Pathscale

2010-02-09 Thread Jeff Squyres
Iain did the genius for the new assembly.  Iain -- can you respond?


On Feb 9, 2010, at 5:44 PM, Mostyn Lewis wrote:

> The old opal_atomic_cmpset_32 worked:
> 
> static inline int opal_atomic_cmpset_32( volatile int32_t *addr,
> unsigned char ret;
> __asm__ __volatile__ (
> SMPLOCK "cmpxchgl %1,%2   \n\t"
> "sete %0  \n\t"
> : "=qm" (ret)
> : "q"(newval), "m"(*addr), "a"(oldval)
> : "memory");
> 
> return (int)ret;
> }
> 
> The new opal_atomic_cmpset_32 fails:
> 
> static inline int opal_atomic_cmpset_32( volatile int32_t *addr,
>  int32_t oldval, int32_t newval)
> {
> unsigned char ret;
> __asm__ __volatile__ (
> SMPLOCK "cmpxchgl %3,%4   \n\t"
> "sete %0  \n\t"
> : "=qm" (ret), "=a" (oldval), "=m" (*addr)
> : "q"(newval), "m"(*addr), "1"(oldval)
> return (int)ret;
> }
> 
> **However** if you put back the "clobber" for memory line (3rd :), it works:
> 
> static inline int opal_atomic_cmpset_32( volatile int32_t *addr,
>  int32_t oldval, int32_t newval)
> {
> unsigned char ret;
> __asm__ __volatile__ (
> SMPLOCK "cmpxchgl %3,%4   \n\t"
> "sete %0  \n\t"
> : "=qm" (ret), "=a" (oldval), "=m" (*addr)
> : "q"(newval), "m"(*addr), "1"(oldval)
> : "memory");
> 
> return (int)ret;
> }
> 
> This works in a test case for pathcc, gcc, icc, pgcc, SUN studio cc and 
> open64 (pathscale
> lineage - which also fails with 1.4.1).
> Also the SMPLOCK above is defined as "lock; " - the ";" is a GNU as statement 
> delimter - is
> that right? Seems to work with/without the ";".
> 
> 
> Also, a question - I see you generate via perl another "lock" asm file which 
> you put into
> opal/asm/generated/ and stick into 
> libasm - what you
> generate there for whatever usage hasn't changed 1.4->1.4.1->svn trunk?
> 
> DM
> 
> On Tue, 9 Feb 2010, Jeff Squyres wrote:
> 
> > Perhaps someone with a pathscale compiler support contract can investigate 
> > this with them.
> >
> > Have them contact us if they want/need help understanding our atomics; 
> > we're happy to explain, etc. (the atomics are fairly localized to a small 
> > part of OMPI).
> >
> >
> >
> > On Feb 9, 2010, at 11:42 AM, Mostyn Lewis wrote:
> >
> >> All,
> >>
> >> FWIW, Pathscale is dying in the new atomics in 1.4.1 (and svn trunk) - 
> >> actually looping -
> >>
> >> from gdb:
> >>
> >> opal_progress_event_users_decrement () at 
> >> ../.././opal/include/opal/sys/atomic_impl.h:61
> >> 61 } while (0 == opal_atomic_cmpset_32(addr, oldval, oldval - 
> >> delta));
> >> Current language:  auto; currently asm
> >> (gdb) where
> >> #0  opal_progress_event_users_decrement () at 
> >> ../.././opal/include/opal/sys/atomic_impl.h:61
> >> #1  0x0001 in ?? ()
> >> #2  0x2aec4cf6a5e0 in ?? ()
> >> #3  0x00eb in ?? ()
> >> #4  0x2aec4cfb57e0 in ompi_mpi_init () at 
> >> ../.././ompi/runtime/ompi_mpi_init.c:818
> >> #5  0x7fff5db3bd58 in ?? ()
> >> Backtrace stopped: previous frame inner to this frame (corrupt stack?)
> >> (gdb) list
> >> 56  {
> >> 57 int32_t oldval;
> >> 58
> >> 59 do {
> >> 60oldval = *addr;
> >> 61 } while (0 == opal_atomic_cmpset_32(addr, oldval, oldval - 
> >> delta));
> >> 62 return (oldval - delta);
> >> 63  }
> >> 64  #endif  /* OPAL_HAVE_ATOMIC_SUB_32 */
> >> 65
> >> (gdb)
> >>
> >> DM
> >>
> >> On Tue, 9 Feb 2010, Jeff Squyres wrote:
> >>
> >>> FWIW, I have had terrible luck with the patschale compiler over the 
> >>> years.  Repeated attempts to get support from them -- even when I was a 
> >>> paying customer -- resulted in no help (e.g., a pathCC bug with the OMPI 
> >>> C++ bindings that I filed years ago was never resolved).
> >>>
> >>> Is this compiler even supported anymore?  I.e., is there a support 
> >>> department somewhere that you have a hope of getting any help from?
> >>>
> >>> I can't say for sure, of course, but if MPI hello world hangs, it smells 
> >>> like a compiler bug.  You might want to attach to "hello world" in a 
> >>> debugger and see where it's hung.  You might need to compile OMPI with 
> >>> debugging symbols to get any meaningful information.
> >>>
> >>> ** NOTE: My personal feelings about the pathscale compiler suite do not 
> >>> reflect anyone else's feelings in the Open MPI community.  Perhaps 
> >>> someone could change my mind someday, but *I* have personally given up on 
> >>> this compiler.  :-(
> >>>
> >>>
> >>> On Feb 8, 2010, at 2:38 AM, Rafael Arco Arredondo wrote:
> >>>
>  Hello,

Re: [OMPI users] Problems building Open MPI 1.4.1 with Pathscale

2010-02-09 Thread Mostyn Lewis

The old opal_atomic_cmpset_32 worked:

static inline int opal_atomic_cmpset_32( volatile int32_t *addr,
   unsigned char ret;
   __asm__ __volatile__ (
   SMPLOCK "cmpxchgl %1,%2   \n\t"
   "sete %0  \n\t"
   : "=qm" (ret)
   : "q"(newval), "m"(*addr), "a"(oldval)
   : "memory");

   return (int)ret; 
}


The new opal_atomic_cmpset_32 fails:

static inline int opal_atomic_cmpset_32( volatile int32_t *addr,
int32_t oldval, int32_t newval)
{
   unsigned char ret;
   __asm__ __volatile__ (
   SMPLOCK "cmpxchgl %3,%4   \n\t"
   "sete %0  \n\t"
   : "=qm" (ret), "=a" (oldval), "=m" (*addr)
   : "q"(newval), "m"(*addr), "1"(oldval)
   return (int)ret;
}

**However** if you put back the "clobber" for memory line (3rd :), it works:

static inline int opal_atomic_cmpset_32( volatile int32_t *addr,
int32_t oldval, int32_t newval)
{
   unsigned char ret;
   __asm__ __volatile__ (
   SMPLOCK "cmpxchgl %3,%4   \n\t"
   "sete %0  \n\t"
   : "=qm" (ret), "=a" (oldval), "=m" (*addr)
   : "q"(newval), "m"(*addr), "1"(oldval)
   : "memory");

   return (int)ret;
}

This works in a test case for pathcc, gcc, icc, pgcc, SUN studio cc and open64 
(pathscale
lineage - which also fails with 1.4.1).
Also the SMPLOCK above is defined as "lock; " - the ";" is a GNU as statement 
delimter - is
that right? Seems to work with/without the ";".


Also, a question - I see you generate via perl another "lock" asm file which 
you put into
opal/asm/generated/ and stick into libasm 
- what you
generate there for whatever usage hasn't changed 1.4->1.4.1->svn trunk?

DM

On Tue, 9 Feb 2010, Jeff Squyres wrote:


Perhaps someone with a pathscale compiler support contract can investigate this 
with them.

Have them contact us if they want/need help understanding our atomics; we're 
happy to explain, etc. (the atomics are fairly localized to a small part of 
OMPI).



On Feb 9, 2010, at 11:42 AM, Mostyn Lewis wrote:


All,

FWIW, Pathscale is dying in the new atomics in 1.4.1 (and svn trunk) - actually 
looping -

from gdb:

opal_progress_event_users_decrement () at 
../.././opal/include/opal/sys/atomic_impl.h:61
61 } while (0 == opal_atomic_cmpset_32(addr, oldval, oldval - delta));
Current language:  auto; currently asm
(gdb) where
#0  opal_progress_event_users_decrement () at 
../.././opal/include/opal/sys/atomic_impl.h:61
#1  0x0001 in ?? ()
#2  0x2aec4cf6a5e0 in ?? ()
#3  0x00eb in ?? ()
#4  0x2aec4cfb57e0 in ompi_mpi_init () at 
../.././ompi/runtime/ompi_mpi_init.c:818
#5  0x7fff5db3bd58 in ?? ()
Backtrace stopped: previous frame inner to this frame (corrupt stack?)
(gdb) list
56  {
57 int32_t oldval;
58
59 do {
60oldval = *addr;
61 } while (0 == opal_atomic_cmpset_32(addr, oldval, oldval - delta));
62 return (oldval - delta);
63  }
64  #endif  /* OPAL_HAVE_ATOMIC_SUB_32 */
65
(gdb)

DM

On Tue, 9 Feb 2010, Jeff Squyres wrote:


FWIW, I have had terrible luck with the patschale compiler over the years.  
Repeated attempts to get support from them -- even when I was a paying customer 
-- resulted in no help (e.g., a pathCC bug with the OMPI C++ bindings that I 
filed years ago was never resolved).

Is this compiler even supported anymore?  I.e., is there a support department 
somewhere that you have a hope of getting any help from?

I can't say for sure, of course, but if MPI hello world hangs, it smells like a compiler 
bug.  You might want to attach to "hello world" in a debugger and see where 
it's hung.  You might need to compile OMPI with debugging symbols to get any meaningful 
information.

** NOTE: My personal feelings about the pathscale compiler suite do not reflect 
anyone else's feelings in the Open MPI community.  Perhaps someone could change 
my mind someday, but *I* have personally given up on this compiler.  :-(


On Feb 8, 2010, at 2:38 AM, Rafael Arco Arredondo wrote:


Hello,

It does work with version 1.4. This is the hello world that hangs with
1.4.1:

#include 
#include 

int main(int argc, char **argv)
{
  int node, size;

  MPI_Init(,);
  MPI_Comm_rank(MPI_COMM_WORLD, );
  MPI_Comm_size(MPI_COMM_WORLD, );

  printf("Hello World from Node %d of %d.\n", node, size);

  MPI_Finalize();
  return 0;
}

El mar, 26-01-2010 a las 03:57 -0500, ?ke Sandgren escribi?:

1 - Do you have problems with openmpi 1.4 too? (I don't, haven't built
1.4.1 yet)
2 - There is a bug in the pathscale compiler with -fPIC and -g that
generates incorrect dwarf2 data so debuggers get really confused and
will have 

Re: [OMPI users] openmpi errors on ubuntu:no connectivity

2010-02-09 Thread Jeff Squyres
Is there any chance you can upgrade to Open MPI v1.4?  1.2.x. is fairly ancient.

Upgrading to 1.4.x will fix the "unable to find any HCAs..." warning message.

For the a.out message, however, it is generally easiest to have the executable 
available on all nodes in the same filesystem location.  For small clusters, 
using NFS can be convenient here.  E.g., NFS share /home, and have your a.out 
under /home/mpi/a.out.  Then you can mpirun -np 2 --host a,b /home/mpi/a.out.


On Feb 9, 2010, at 2:17 PM, swagat mishra wrote:

> hello,
> we have installed open mpi 1.2 using synaptic package manager in 2 machines 
> running on ubuntu 8.10 and ubuntu 8.04.the hello.c program runs correctly,but 
> connectivity_c.c program included in the openmpi tarball example fails when 
> we it tries to communicate between both computers.also on the 8.04 version we 
> get  an error of "unable to find any HCA's will use another transport 
> instead".on the 8.10 machine we get an error "unable to find executable:a.out 
> host:ip of node".
> however the programs run without errors when we specify  only localhost in 
> hostfile.we have set up PATH and LD_LIBRARY_PATH,but still get same errors. 
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users


-- 
Jeff Squyres
jsquy...@cisco.com

For corporate legal information go to:
http://www.cisco.com/web/about/doing_business/legal/cri/




[OMPI users] openmpi errors on ubuntu:no connectivity

2010-02-09 Thread swagat mishra
hello,
we have installed open mpi 1.2 using synaptic package manager in 2 machines
running on ubuntu 8.10 and ubuntu 8.04.the hello.c program runs
correctly,but connectivity_c.c program included in the openmpi tarball
example fails when we it tries to communicate between both computers.also on
the 8.04 version we get  an error of "unable to find any HCA's will use
another transport instead".on the 8.10 machine we get an error "unable to
find executable:a.out host:ip of node".
however the programs run without errors when we specify  only localhost in
hostfile.we have set up PATH and LD_LIBRARY_PATH,but still get same errors.


Re: [OMPI users] Problems building Open MPI 1.4.1 with Pathscale

2010-02-09 Thread Åke Sandgren
On Tue, 2010-02-09 at 13:42 -0500, Jeff Squyres wrote:
> Perhaps someone with a pathscale compiler support contract can investigate 
> this with them.
> 
> Have them contact us if they want/need help understanding our atomics; we're 
> happy to explain, etc. (the atomics are fairly localized to a small part of 
> OMPI).

I will surely do that.
It will take a few days though due to lots of other work.

-- 
Ake Sandgren, HPC2N, Umea University, S-90187 Umea, Sweden
Internet: a...@hpc2n.umu.se   Phone: +46 90 7866134 Fax: +46 90 7866126
Mobile: +46 70 7716134 WWW: http://www.hpc2n.umu.se



Re: [OMPI users] Problems building Open MPI 1.4.1 with Pathscale

2010-02-09 Thread Jeff Squyres
Perhaps someone with a pathscale compiler support contract can investigate this 
with them.

Have them contact us if they want/need help understanding our atomics; we're 
happy to explain, etc. (the atomics are fairly localized to a small part of 
OMPI).



On Feb 9, 2010, at 11:42 AM, Mostyn Lewis wrote:

> All,
> 
> FWIW, Pathscale is dying in the new atomics in 1.4.1 (and svn trunk) - 
> actually looping -
> 
> from gdb:
> 
> opal_progress_event_users_decrement () at 
> ../.././opal/include/opal/sys/atomic_impl.h:61
> 61 } while (0 == opal_atomic_cmpset_32(addr, oldval, oldval - delta));
> Current language:  auto; currently asm
> (gdb) where
> #0  opal_progress_event_users_decrement () at 
> ../.././opal/include/opal/sys/atomic_impl.h:61
> #1  0x0001 in ?? ()
> #2  0x2aec4cf6a5e0 in ?? ()
> #3  0x00eb in ?? ()
> #4  0x2aec4cfb57e0 in ompi_mpi_init () at 
> ../.././ompi/runtime/ompi_mpi_init.c:818
> #5  0x7fff5db3bd58 in ?? ()
> Backtrace stopped: previous frame inner to this frame (corrupt stack?)
> (gdb) list
> 56  {
> 57 int32_t oldval;
> 58
> 59 do {
> 60oldval = *addr;
> 61 } while (0 == opal_atomic_cmpset_32(addr, oldval, oldval - delta));
> 62 return (oldval - delta);
> 63  }
> 64  #endif  /* OPAL_HAVE_ATOMIC_SUB_32 */
> 65
> (gdb)
> 
> DM
> 
> On Tue, 9 Feb 2010, Jeff Squyres wrote:
> 
> > FWIW, I have had terrible luck with the patschale compiler over the years.  
> > Repeated attempts to get support from them -- even when I was a paying 
> > customer -- resulted in no help (e.g., a pathCC bug with the OMPI C++ 
> > bindings that I filed years ago was never resolved).
> >
> > Is this compiler even supported anymore?  I.e., is there a support 
> > department somewhere that you have a hope of getting any help from?
> >
> > I can't say for sure, of course, but if MPI hello world hangs, it smells 
> > like a compiler bug.  You might want to attach to "hello world" in a 
> > debugger and see where it's hung.  You might need to compile OMPI with 
> > debugging symbols to get any meaningful information.
> >
> > ** NOTE: My personal feelings about the pathscale compiler suite do not 
> > reflect anyone else's feelings in the Open MPI community.  Perhaps someone 
> > could change my mind someday, but *I* have personally given up on this 
> > compiler.  :-(
> >
> >
> > On Feb 8, 2010, at 2:38 AM, Rafael Arco Arredondo wrote:
> >
> >> Hello,
> >>
> >> It does work with version 1.4. This is the hello world that hangs with
> >> 1.4.1:
> >>
> >> #include 
> >> #include 
> >>
> >> int main(int argc, char **argv)
> >> {
> >>   int node, size;
> >>
> >>   MPI_Init(,);
> >>   MPI_Comm_rank(MPI_COMM_WORLD, );
> >>   MPI_Comm_size(MPI_COMM_WORLD, );
> >>
> >>   printf("Hello World from Node %d of %d.\n", node, size);
> >>
> >>   MPI_Finalize();
> >>   return 0;
> >> }
> >>
> >> El mar, 26-01-2010 a las 03:57 -0500, Åke Sandgren escribió:
> >>> 1 - Do you have problems with openmpi 1.4 too? (I don't, haven't built
> >>> 1.4.1 yet)
> >>> 2 - There is a bug in the pathscale compiler with -fPIC and -g that
> >>> generates incorrect dwarf2 data so debuggers get really confused and
> >>> will have BIG problems debugging the code. I'm chasing them to get a
> >>> fix...
> >>> 3 - Do you have an example code that have problems?
> >>
> >> --
> >> Rafael Arco Arredondo
> >> Centro de Servicios de Informática y Redes de Comunicaciones
> >> Universidad de Granada
> >>
> >> ___
> >> users mailing list
> >> us...@open-mpi.org
> >> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>
> >
> >
> > --
> > Jeff Squyres
> > jsquy...@cisco.com
> >
> > For corporate legal information go to:
> > http://www.cisco.com/web/about/doing_business/legal/cri/
> >
> >
> > ___
> > users mailing list
> > us...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
> >
> 
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users


-- 
Jeff Squyres
jsquy...@cisco.com

For corporate legal information go to:
http://www.cisco.com/web/about/doing_business/legal/cri/




Re: [OMPI users] Problems building Open MPI 1.4.1 with Pathscale

2010-02-09 Thread Mostyn Lewis

All,

FWIW, Pathscale is dying in the new atomics in 1.4.1 (and svn trunk) - actually 
looping -

from gdb:

opal_progress_event_users_decrement () at 
../.././opal/include/opal/sys/atomic_impl.h:61
61 } while (0 == opal_atomic_cmpset_32(addr, oldval, oldval - delta));
Current language:  auto; currently asm
(gdb) where
#0  opal_progress_event_users_decrement () at 
../.././opal/include/opal/sys/atomic_impl.h:61
#1  0x0001 in ?? ()
#2  0x2aec4cf6a5e0 in ?? ()
#3  0x00eb in ?? ()
#4  0x2aec4cfb57e0 in ompi_mpi_init () at 
../.././ompi/runtime/ompi_mpi_init.c:818
#5  0x7fff5db3bd58 in ?? ()
Backtrace stopped: previous frame inner to this frame (corrupt stack?)
(gdb) list
56  {
57 int32_t oldval;
58 
59 do {

60oldval = *addr;
61 } while (0 == opal_atomic_cmpset_32(addr, oldval, oldval - delta));
62 return (oldval - delta);
63  }
64  #endif  /* OPAL_HAVE_ATOMIC_SUB_32 */
65
(gdb)

DM

On Tue, 9 Feb 2010, Jeff Squyres wrote:


FWIW, I have had terrible luck with the patschale compiler over the years.  
Repeated attempts to get support from them -- even when I was a paying customer 
-- resulted in no help (e.g., a pathCC bug with the OMPI C++ bindings that I 
filed years ago was never resolved).

Is this compiler even supported anymore?  I.e., is there a support department 
somewhere that you have a hope of getting any help from?

I can't say for sure, of course, but if MPI hello world hangs, it smells like a compiler 
bug.  You might want to attach to "hello world" in a debugger and see where 
it's hung.  You might need to compile OMPI with debugging symbols to get any meaningful 
information.

** NOTE: My personal feelings about the pathscale compiler suite do not reflect 
anyone else's feelings in the Open MPI community.  Perhaps someone could change 
my mind someday, but *I* have personally given up on this compiler.  :-(


On Feb 8, 2010, at 2:38 AM, Rafael Arco Arredondo wrote:


Hello,

It does work with version 1.4. This is the hello world that hangs with
1.4.1:

#include 
#include 

int main(int argc, char **argv)
{
  int node, size;

  MPI_Init(,);
  MPI_Comm_rank(MPI_COMM_WORLD, );
  MPI_Comm_size(MPI_COMM_WORLD, );

  printf("Hello World from Node %d of %d.\n", node, size);

  MPI_Finalize();
  return 0;
}

El mar, 26-01-2010 a las 03:57 -0500, ?ke Sandgren escribi?:

1 - Do you have problems with openmpi 1.4 too? (I don't, haven't built
1.4.1 yet)
2 - There is a bug in the pathscale compiler with -fPIC and -g that
generates incorrect dwarf2 data so debuggers get really confused and
will have BIG problems debugging the code. I'm chasing them to get a
fix...
3 - Do you have an example code that have problems?


--
Rafael Arco Arredondo
Centro de Servicios de Inform?tica y Redes de Comunicaciones
Universidad de Granada

___
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users




--
Jeff Squyres
jsquy...@cisco.com

For corporate legal information go to:
http://www.cisco.com/web/about/doing_business/legal/cri/


___
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users



Re: [OMPI users] ompi_info loop

2010-02-09 Thread Ethan Mallove
On Tue, Feb/09/2010 08:46:53AM, Benjamin Gaudio wrote:
> In trying to track down my default hostfile problem, I found that
> when I run ompi_info, it simply keeps repeating:
> 
> Displaying Open MPI information for 32-bit ...
> Displaying Open MPI information for 32-bit ...
> Displaying Open MPI information for 32-bit ...
> Displaying Open MPI information for 32-bit ...
> Displaying Open MPI information for 32-bit ...
> Displaying Open MPI information for 32-bit ...
> 
> Has anyone seen this before?  I am using Sun HPC ClusterTools 8.2.1
> on Solaris 10.

Hi Benjamin,

I can reproduce this using /opt/SUNWhpc/HPC8.2.1/sun/bin/64/ompi_info.
We'll fix this ASAP (for the next ClusterTools release).  A workaround
is to use /opt/SUNWhpc/HPC8.2.1/sun/bin/32/ompi_info.

Thanks,
Ethan

> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users


Re: [OMPI users] Problems building Open MPI 1.4.1 with Pathscale

2010-02-09 Thread Ake Sandgren
On Tue, 2010-02-09 at 08:49 -0500, Jeff Squyres wrote:
> FWIW, I have had terrible luck with the patschale compiler over the years.  
> Repeated attempts to get support from them -- even when I was a paying 
> customer -- resulted in no help (e.g., a pathCC bug with the OMPI C++ 
> bindings that I filed years ago was never resolved).
> 
> Is this compiler even supported anymore?  I.e., is there a support department 
> somewhere that you have a hope of getting any help from?
> 
> I can't say for sure, of course, but if MPI hello world hangs, it smells like 
> a compiler bug.  You might want to attach to "hello world" in a debugger and 
> see where it's hung.  You might need to compile OMPI with debugging symbols 
> to get any meaningful information.
> 
> ** NOTE: My personal feelings about the pathscale compiler suite do not 
> reflect anyone else's feelings in the Open MPI community.  Perhaps someone 
> could change my mind someday, but *I* have personally given up on this 
> compiler.  :-(

Pathscale is not dead, in fact I'm talking to them more or less daily at
the moment. They have been restructuring since the demise of SciCortex
last year. I hope they will be able to release a new version fairly
soon.

In my opinion (working mostly with Fortran codes, shudder) it is the
best compiler around. Although they have had problems over the years in
coming out with fixes for bugs in a timely fashion.

-- 
Ake Sandgren, HPC2N, Umea University, S-90187 Umea, Sweden
Internet: a...@hpc2n.umu.se   Phone: +46 90 7866134 Fax: +46 90 7866126
Mobile: +46 70 7716134 WWW: http://www.hpc2n.umu.se



Re: [OMPI users] Problems building Open MPI 1.4.1 with Pathscale

2010-02-09 Thread Jeff Squyres
FWIW, I have had terrible luck with the patschale compiler over the years.  
Repeated attempts to get support from them -- even when I was a paying customer 
-- resulted in no help (e.g., a pathCC bug with the OMPI C++ bindings that I 
filed years ago was never resolved).

Is this compiler even supported anymore?  I.e., is there a support department 
somewhere that you have a hope of getting any help from?

I can't say for sure, of course, but if MPI hello world hangs, it smells like a 
compiler bug.  You might want to attach to "hello world" in a debugger and see 
where it's hung.  You might need to compile OMPI with debugging symbols to get 
any meaningful information.

** NOTE: My personal feelings about the pathscale compiler suite do not reflect 
anyone else's feelings in the Open MPI community.  Perhaps someone could change 
my mind someday, but *I* have personally given up on this compiler.  :-(


On Feb 8, 2010, at 2:38 AM, Rafael Arco Arredondo wrote:

> Hello,
> 
> It does work with version 1.4. This is the hello world that hangs with
> 1.4.1:
> 
> #include 
> #include 
> 
> int main(int argc, char **argv)
> {
>   int node, size;
> 
>   MPI_Init(,);
>   MPI_Comm_rank(MPI_COMM_WORLD, );
>   MPI_Comm_size(MPI_COMM_WORLD, );
> 
>   printf("Hello World from Node %d of %d.\n", node, size);
> 
>   MPI_Finalize();
>   return 0;
> }
> 
> El mar, 26-01-2010 a las 03:57 -0500, Åke Sandgren escribió:
> > 1 - Do you have problems with openmpi 1.4 too? (I don't, haven't built
> > 1.4.1 yet)
> > 2 - There is a bug in the pathscale compiler with -fPIC and -g that
> > generates incorrect dwarf2 data so debuggers get really confused and
> > will have BIG problems debugging the code. I'm chasing them to get a
> > fix...
> > 3 - Do you have an example code that have problems?
> 
> --
> Rafael Arco Arredondo
> Centro de Servicios de Informática y Redes de Comunicaciones
> Universidad de Granada
> 
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 


-- 
Jeff Squyres
jsquy...@cisco.com

For corporate legal information go to:
http://www.cisco.com/web/about/doing_business/legal/cri/




[OMPI users] ompi_info loop

2010-02-09 Thread Benjamin Gaudio
In trying to track down my default hostfile problem, I found that
when I run ompi_info, it simply keeps repeating:

Displaying Open MPI information for 32-bit ...
Displaying Open MPI information for 32-bit ...
Displaying Open MPI information for 32-bit ...
Displaying Open MPI information for 32-bit ...
Displaying Open MPI information for 32-bit ...
Displaying Open MPI information for 32-bit ...

Has anyone seen this before?  I am using Sun HPC ClusterTools 8.2.1
on Solaris 10.


Re: [OMPI users] Anybody built a working 1.4.1 on Solaris 8, (Sparc)?

2010-02-09 Thread Terry Dontje


Date: Fri, 05 Feb 2010 16:16:29 -0800 From: "David Mathog" 
 We haven't tried Solaris 8 in quite some time.  However, for your first 
> issue did you include the --enable-heterogeneous option on your 
> configure command?
> 
> Since you are mix IA-32 and SPARC nodes you'll want to include this so 
> the endian issue doesn't bite you.



Added that on the configure, rebuilt, installed, and now the examples work. 
 
  

Glad to hear that.

Any thoughts on the Forte compiler issue?  This is not quite as pressing
now that the gcc version works, and most of the computation will be on
the remote nodes anyway.  Still, the Forte compilers should generate
faster code than gcc, and I would prefer to use them if possible.

  
Are you using the vampir tracing feature?  If not I would add the 
"--enable-contrib-no-build=vt" option to your configure line and see if 
that works.


Beware, I don't think we've tried compiling OMPI with Forte 7.  Also 
note omp.h is for OpenMP not MPI.   It looks like the last error you ran 
into with vt_unify.cc may be an issue with the Forte 7 C++ compiler. 


--td


Thanks,

David Mathog
mat...@caltech.edu
Manager, Sequence Analysis Facility, Biology Division, Caltech