Re: [OMPI users] Problems building Open MPI 1.4.1 with Pathscale
Well, I am by no means an expert on the GNU-style asm directives. I believe someone else (George Bosilca?) tweaked what I had suggested. That being said, I think the memory "clobber" is harmless. Iain On Feb 9, 2010, at 5:51 PM, Jeff Squyres wrote: Iain did the genius for the new assembly. Iain -- can you respond? On Feb 9, 2010, at 5:44 PM, Mostyn Lewis wrote: The old opal_atomic_cmpset_32 worked: static inline int opal_atomic_cmpset_32( volatile int32_t *addr, unsigned char ret; __asm__ __volatile__ ( SMPLOCK "cmpxchgl %1,%2 \n\t" "sete %0 \n\t" : "=qm" (ret) : "q"(newval), "m"(*addr), "a"(oldval) : "memory"); return (int)ret; } The new opal_atomic_cmpset_32 fails: static inline int opal_atomic_cmpset_32( volatile int32_t *addr, int32_t oldval, int32_t newval) { unsigned char ret; __asm__ __volatile__ ( SMPLOCK "cmpxchgl %3,%4 \n\t" "sete %0 \n\t" : "=qm" (ret), "=a" (oldval), "=m" (*addr) : "q"(newval), "m"(*addr), "1"(oldval) return (int)ret; } **However** if you put back the "clobber" for memory line (3rd :), it works: static inline int opal_atomic_cmpset_32( volatile int32_t *addr, int32_t oldval, int32_t newval) { unsigned char ret; __asm__ __volatile__ ( SMPLOCK "cmpxchgl %3,%4 \n\t" "sete %0 \n\t" : "=qm" (ret), "=a" (oldval), "=m" (*addr) : "q"(newval), "m"(*addr), "1"(oldval) : "memory"); return (int)ret; } This works in a test case for pathcc, gcc, icc, pgcc, SUN studio cc and open64 (pathscale lineage - which also fails with 1.4.1). Also the SMPLOCK above is defined as "lock; " - the ";" is a GNU as statement delimter - is that right? Seems to work with/without the ";". Also, a question - I see you generate via perl another "lock" asm file which you put into opal/asm/generated/and stick into libasm - what you generate there for whatever usage hasn't changed 1.4->1.4.1->svn trunk? DM On Tue, 9 Feb 2010, Jeff Squyres wrote: Perhaps someone with a pathscale compiler support contract can investigate this with them. Have them contact us if they want/need help understanding our atomics; we're happy to explain, etc. (the atomics are fairly localized to a small part of OMPI). On Feb 9, 2010, at 11:42 AM, Mostyn Lewis wrote: All, FWIW, Pathscale is dying in the new atomics in 1.4.1 (and svn trunk) - actually looping - from gdb: opal_progress_event_users_decrement () at ../.././opal/include/ opal/sys/atomic_impl.h:61 61 } while (0 == opal_atomic_cmpset_32(addr, oldval, oldval - delta)); Current language: auto; currently asm (gdb) where #0 opal_progress_event_users_decrement () at ../.././opal/ include/opal/sys/atomic_impl.h:61 #1 0x0001 in ?? () #2 0x2aec4cf6a5e0 in ?? () #3 0x00eb in ?? () #4 0x2aec4cfb57e0 in ompi_mpi_init () at ../.././ompi/ runtime/ompi_mpi_init.c:818 #5 0x7fff5db3bd58 in ?? () Backtrace stopped: previous frame inner to this frame (corrupt stack?) (gdb) list 56 { 57 int32_t oldval; 58 59 do { 60oldval = *addr; 61 } while (0 == opal_atomic_cmpset_32(addr, oldval, oldval - delta)); 62 return (oldval - delta); 63 } 64 #endif /* OPAL_HAVE_ATOMIC_SUB_32 */ 65 (gdb) DM On Tue, 9 Feb 2010, Jeff Squyres wrote: FWIW, I have had terrible luck with the patschale compiler over the years. Repeated attempts to get support from them -- even when I was a paying customer -- resulted in no help (e.g., a pathCC bug with the OMPI C++ bindings that I filed years ago was never resolved). Is this compiler even supported anymore? I.e., is there a support department somewhere that you have a hope of getting any help from? I can't say for sure, of course, but if MPI hello world hangs, it smells like a compiler bug. You might want to attach to "hello world" in a debugger and see where it's hung. You might need to compile OMPI with debugging symbols to get any meaningful information. ** NOTE: My personal feelings about the pathscale compiler suite do not reflect anyone else's feelings in the Open MPI community. Perhaps someone could change my mind someday, but *I* have personally given up on this compiler. :-( On Feb 8, 2010, at 2:38 AM, Rafael Arco Arredondo wrote: Hello, It does work with version 1.4. This is the hello world that hangs with 1.4.1: #include #include int main(int argc, char **argv) { int node, size; MPI_Init(,);
Re: [OMPI users] Problems building Open MPI 1.4.1 with Pathscale
Iain did the genius for the new assembly. Iain -- can you respond? On Feb 9, 2010, at 5:44 PM, Mostyn Lewis wrote: > The old opal_atomic_cmpset_32 worked: > > static inline int opal_atomic_cmpset_32( volatile int32_t *addr, > unsigned char ret; > __asm__ __volatile__ ( > SMPLOCK "cmpxchgl %1,%2 \n\t" > "sete %0 \n\t" > : "=qm" (ret) > : "q"(newval), "m"(*addr), "a"(oldval) > : "memory"); > > return (int)ret; > } > > The new opal_atomic_cmpset_32 fails: > > static inline int opal_atomic_cmpset_32( volatile int32_t *addr, > int32_t oldval, int32_t newval) > { > unsigned char ret; > __asm__ __volatile__ ( > SMPLOCK "cmpxchgl %3,%4 \n\t" > "sete %0 \n\t" > : "=qm" (ret), "=a" (oldval), "=m" (*addr) > : "q"(newval), "m"(*addr), "1"(oldval) > return (int)ret; > } > > **However** if you put back the "clobber" for memory line (3rd :), it works: > > static inline int opal_atomic_cmpset_32( volatile int32_t *addr, > int32_t oldval, int32_t newval) > { > unsigned char ret; > __asm__ __volatile__ ( > SMPLOCK "cmpxchgl %3,%4 \n\t" > "sete %0 \n\t" > : "=qm" (ret), "=a" (oldval), "=m" (*addr) > : "q"(newval), "m"(*addr), "1"(oldval) > : "memory"); > > return (int)ret; > } > > This works in a test case for pathcc, gcc, icc, pgcc, SUN studio cc and > open64 (pathscale > lineage - which also fails with 1.4.1). > Also the SMPLOCK above is defined as "lock; " - the ";" is a GNU as statement > delimter - is > that right? Seems to work with/without the ";". > > > Also, a question - I see you generate via perl another "lock" asm file which > you put into > opal/asm/generated/and stick into > libasm - what you > generate there for whatever usage hasn't changed 1.4->1.4.1->svn trunk? > > DM > > On Tue, 9 Feb 2010, Jeff Squyres wrote: > > > Perhaps someone with a pathscale compiler support contract can investigate > > this with them. > > > > Have them contact us if they want/need help understanding our atomics; > > we're happy to explain, etc. (the atomics are fairly localized to a small > > part of OMPI). > > > > > > > > On Feb 9, 2010, at 11:42 AM, Mostyn Lewis wrote: > > > >> All, > >> > >> FWIW, Pathscale is dying in the new atomics in 1.4.1 (and svn trunk) - > >> actually looping - > >> > >> from gdb: > >> > >> opal_progress_event_users_decrement () at > >> ../.././opal/include/opal/sys/atomic_impl.h:61 > >> 61 } while (0 == opal_atomic_cmpset_32(addr, oldval, oldval - > >> delta)); > >> Current language: auto; currently asm > >> (gdb) where > >> #0 opal_progress_event_users_decrement () at > >> ../.././opal/include/opal/sys/atomic_impl.h:61 > >> #1 0x0001 in ?? () > >> #2 0x2aec4cf6a5e0 in ?? () > >> #3 0x00eb in ?? () > >> #4 0x2aec4cfb57e0 in ompi_mpi_init () at > >> ../.././ompi/runtime/ompi_mpi_init.c:818 > >> #5 0x7fff5db3bd58 in ?? () > >> Backtrace stopped: previous frame inner to this frame (corrupt stack?) > >> (gdb) list > >> 56 { > >> 57 int32_t oldval; > >> 58 > >> 59 do { > >> 60oldval = *addr; > >> 61 } while (0 == opal_atomic_cmpset_32(addr, oldval, oldval - > >> delta)); > >> 62 return (oldval - delta); > >> 63 } > >> 64 #endif /* OPAL_HAVE_ATOMIC_SUB_32 */ > >> 65 > >> (gdb) > >> > >> DM > >> > >> On Tue, 9 Feb 2010, Jeff Squyres wrote: > >> > >>> FWIW, I have had terrible luck with the patschale compiler over the > >>> years. Repeated attempts to get support from them -- even when I was a > >>> paying customer -- resulted in no help (e.g., a pathCC bug with the OMPI > >>> C++ bindings that I filed years ago was never resolved). > >>> > >>> Is this compiler even supported anymore? I.e., is there a support > >>> department somewhere that you have a hope of getting any help from? > >>> > >>> I can't say for sure, of course, but if MPI hello world hangs, it smells > >>> like a compiler bug. You might want to attach to "hello world" in a > >>> debugger and see where it's hung. You might need to compile OMPI with > >>> debugging symbols to get any meaningful information. > >>> > >>> ** NOTE: My personal feelings about the pathscale compiler suite do not > >>> reflect anyone else's feelings in the Open MPI community. Perhaps > >>> someone could change my mind someday, but *I* have personally given up on > >>> this compiler. :-( > >>> > >>> > >>> On Feb 8, 2010, at 2:38 AM, Rafael Arco Arredondo wrote: > >>> > Hello,
Re: [OMPI users] Problems building Open MPI 1.4.1 with Pathscale
The old opal_atomic_cmpset_32 worked: static inline int opal_atomic_cmpset_32( volatile int32_t *addr, unsigned char ret; __asm__ __volatile__ ( SMPLOCK "cmpxchgl %1,%2 \n\t" "sete %0 \n\t" : "=qm" (ret) : "q"(newval), "m"(*addr), "a"(oldval) : "memory"); return (int)ret; } The new opal_atomic_cmpset_32 fails: static inline int opal_atomic_cmpset_32( volatile int32_t *addr, int32_t oldval, int32_t newval) { unsigned char ret; __asm__ __volatile__ ( SMPLOCK "cmpxchgl %3,%4 \n\t" "sete %0 \n\t" : "=qm" (ret), "=a" (oldval), "=m" (*addr) : "q"(newval), "m"(*addr), "1"(oldval) return (int)ret; } **However** if you put back the "clobber" for memory line (3rd :), it works: static inline int opal_atomic_cmpset_32( volatile int32_t *addr, int32_t oldval, int32_t newval) { unsigned char ret; __asm__ __volatile__ ( SMPLOCK "cmpxchgl %3,%4 \n\t" "sete %0 \n\t" : "=qm" (ret), "=a" (oldval), "=m" (*addr) : "q"(newval), "m"(*addr), "1"(oldval) : "memory"); return (int)ret; } This works in a test case for pathcc, gcc, icc, pgcc, SUN studio cc and open64 (pathscale lineage - which also fails with 1.4.1). Also the SMPLOCK above is defined as "lock; " - the ";" is a GNU as statement delimter - is that right? Seems to work with/without the ";". Also, a question - I see you generate via perl another "lock" asm file which you put into opal/asm/generated/and stick into libasm - what you generate there for whatever usage hasn't changed 1.4->1.4.1->svn trunk? DM On Tue, 9 Feb 2010, Jeff Squyres wrote: Perhaps someone with a pathscale compiler support contract can investigate this with them. Have them contact us if they want/need help understanding our atomics; we're happy to explain, etc. (the atomics are fairly localized to a small part of OMPI). On Feb 9, 2010, at 11:42 AM, Mostyn Lewis wrote: All, FWIW, Pathscale is dying in the new atomics in 1.4.1 (and svn trunk) - actually looping - from gdb: opal_progress_event_users_decrement () at ../.././opal/include/opal/sys/atomic_impl.h:61 61 } while (0 == opal_atomic_cmpset_32(addr, oldval, oldval - delta)); Current language: auto; currently asm (gdb) where #0 opal_progress_event_users_decrement () at ../.././opal/include/opal/sys/atomic_impl.h:61 #1 0x0001 in ?? () #2 0x2aec4cf6a5e0 in ?? () #3 0x00eb in ?? () #4 0x2aec4cfb57e0 in ompi_mpi_init () at ../.././ompi/runtime/ompi_mpi_init.c:818 #5 0x7fff5db3bd58 in ?? () Backtrace stopped: previous frame inner to this frame (corrupt stack?) (gdb) list 56 { 57 int32_t oldval; 58 59 do { 60oldval = *addr; 61 } while (0 == opal_atomic_cmpset_32(addr, oldval, oldval - delta)); 62 return (oldval - delta); 63 } 64 #endif /* OPAL_HAVE_ATOMIC_SUB_32 */ 65 (gdb) DM On Tue, 9 Feb 2010, Jeff Squyres wrote: FWIW, I have had terrible luck with the patschale compiler over the years. Repeated attempts to get support from them -- even when I was a paying customer -- resulted in no help (e.g., a pathCC bug with the OMPI C++ bindings that I filed years ago was never resolved). Is this compiler even supported anymore? I.e., is there a support department somewhere that you have a hope of getting any help from? I can't say for sure, of course, but if MPI hello world hangs, it smells like a compiler bug. You might want to attach to "hello world" in a debugger and see where it's hung. You might need to compile OMPI with debugging symbols to get any meaningful information. ** NOTE: My personal feelings about the pathscale compiler suite do not reflect anyone else's feelings in the Open MPI community. Perhaps someone could change my mind someday, but *I* have personally given up on this compiler. :-( On Feb 8, 2010, at 2:38 AM, Rafael Arco Arredondo wrote: Hello, It does work with version 1.4. This is the hello world that hangs with 1.4.1: #include #include int main(int argc, char **argv) { int node, size; MPI_Init(,); MPI_Comm_rank(MPI_COMM_WORLD, ); MPI_Comm_size(MPI_COMM_WORLD, ); printf("Hello World from Node %d of %d.\n", node, size); MPI_Finalize(); return 0; } El mar, 26-01-2010 a las 03:57 -0500, ?ke Sandgren escribi?: 1 - Do you have problems with openmpi 1.4 too? (I don't, haven't built 1.4.1 yet) 2 - There is a bug in the pathscale compiler with -fPIC and -g that generates incorrect dwarf2 data so debuggers get really confused and will have
Re: [OMPI users] openmpi errors on ubuntu:no connectivity
Is there any chance you can upgrade to Open MPI v1.4? 1.2.x. is fairly ancient. Upgrading to 1.4.x will fix the "unable to find any HCAs..." warning message. For the a.out message, however, it is generally easiest to have the executable available on all nodes in the same filesystem location. For small clusters, using NFS can be convenient here. E.g., NFS share /home, and have your a.out under /home/mpi/a.out. Then you can mpirun -np 2 --host a,b /home/mpi/a.out. On Feb 9, 2010, at 2:17 PM, swagat mishra wrote: > hello, > we have installed open mpi 1.2 using synaptic package manager in 2 machines > running on ubuntu 8.10 and ubuntu 8.04.the hello.c program runs correctly,but > connectivity_c.c program included in the openmpi tarball example fails when > we it tries to communicate between both computers.also on the 8.04 version we > get an error of "unable to find any HCA's will use another transport > instead".on the 8.10 machine we get an error "unable to find executable:a.out > host:ip of node". > however the programs run without errors when we specify only localhost in > hostfile.we have set up PATH and LD_LIBRARY_PATH,but still get same errors. > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/
[OMPI users] openmpi errors on ubuntu:no connectivity
hello, we have installed open mpi 1.2 using synaptic package manager in 2 machines running on ubuntu 8.10 and ubuntu 8.04.the hello.c program runs correctly,but connectivity_c.c program included in the openmpi tarball example fails when we it tries to communicate between both computers.also on the 8.04 version we get an error of "unable to find any HCA's will use another transport instead".on the 8.10 machine we get an error "unable to find executable:a.out host:ip of node". however the programs run without errors when we specify only localhost in hostfile.we have set up PATH and LD_LIBRARY_PATH,but still get same errors.
Re: [OMPI users] Problems building Open MPI 1.4.1 with Pathscale
On Tue, 2010-02-09 at 13:42 -0500, Jeff Squyres wrote: > Perhaps someone with a pathscale compiler support contract can investigate > this with them. > > Have them contact us if they want/need help understanding our atomics; we're > happy to explain, etc. (the atomics are fairly localized to a small part of > OMPI). I will surely do that. It will take a few days though due to lots of other work. -- Ake Sandgren, HPC2N, Umea University, S-90187 Umea, Sweden Internet: a...@hpc2n.umu.se Phone: +46 90 7866134 Fax: +46 90 7866126 Mobile: +46 70 7716134 WWW: http://www.hpc2n.umu.se
Re: [OMPI users] Problems building Open MPI 1.4.1 with Pathscale
Perhaps someone with a pathscale compiler support contract can investigate this with them. Have them contact us if they want/need help understanding our atomics; we're happy to explain, etc. (the atomics are fairly localized to a small part of OMPI). On Feb 9, 2010, at 11:42 AM, Mostyn Lewis wrote: > All, > > FWIW, Pathscale is dying in the new atomics in 1.4.1 (and svn trunk) - > actually looping - > > from gdb: > > opal_progress_event_users_decrement () at > ../.././opal/include/opal/sys/atomic_impl.h:61 > 61 } while (0 == opal_atomic_cmpset_32(addr, oldval, oldval - delta)); > Current language: auto; currently asm > (gdb) where > #0 opal_progress_event_users_decrement () at > ../.././opal/include/opal/sys/atomic_impl.h:61 > #1 0x0001 in ?? () > #2 0x2aec4cf6a5e0 in ?? () > #3 0x00eb in ?? () > #4 0x2aec4cfb57e0 in ompi_mpi_init () at > ../.././ompi/runtime/ompi_mpi_init.c:818 > #5 0x7fff5db3bd58 in ?? () > Backtrace stopped: previous frame inner to this frame (corrupt stack?) > (gdb) list > 56 { > 57 int32_t oldval; > 58 > 59 do { > 60oldval = *addr; > 61 } while (0 == opal_atomic_cmpset_32(addr, oldval, oldval - delta)); > 62 return (oldval - delta); > 63 } > 64 #endif /* OPAL_HAVE_ATOMIC_SUB_32 */ > 65 > (gdb) > > DM > > On Tue, 9 Feb 2010, Jeff Squyres wrote: > > > FWIW, I have had terrible luck with the patschale compiler over the years. > > Repeated attempts to get support from them -- even when I was a paying > > customer -- resulted in no help (e.g., a pathCC bug with the OMPI C++ > > bindings that I filed years ago was never resolved). > > > > Is this compiler even supported anymore? I.e., is there a support > > department somewhere that you have a hope of getting any help from? > > > > I can't say for sure, of course, but if MPI hello world hangs, it smells > > like a compiler bug. You might want to attach to "hello world" in a > > debugger and see where it's hung. You might need to compile OMPI with > > debugging symbols to get any meaningful information. > > > > ** NOTE: My personal feelings about the pathscale compiler suite do not > > reflect anyone else's feelings in the Open MPI community. Perhaps someone > > could change my mind someday, but *I* have personally given up on this > > compiler. :-( > > > > > > On Feb 8, 2010, at 2:38 AM, Rafael Arco Arredondo wrote: > > > >> Hello, > >> > >> It does work with version 1.4. This is the hello world that hangs with > >> 1.4.1: > >> > >> #include > >> #include > >> > >> int main(int argc, char **argv) > >> { > >> int node, size; > >> > >> MPI_Init(,); > >> MPI_Comm_rank(MPI_COMM_WORLD, ); > >> MPI_Comm_size(MPI_COMM_WORLD, ); > >> > >> printf("Hello World from Node %d of %d.\n", node, size); > >> > >> MPI_Finalize(); > >> return 0; > >> } > >> > >> El mar, 26-01-2010 a las 03:57 -0500, Åke Sandgren escribió: > >>> 1 - Do you have problems with openmpi 1.4 too? (I don't, haven't built > >>> 1.4.1 yet) > >>> 2 - There is a bug in the pathscale compiler with -fPIC and -g that > >>> generates incorrect dwarf2 data so debuggers get really confused and > >>> will have BIG problems debugging the code. I'm chasing them to get a > >>> fix... > >>> 3 - Do you have an example code that have problems? > >> > >> -- > >> Rafael Arco Arredondo > >> Centro de Servicios de Informática y Redes de Comunicaciones > >> Universidad de Granada > >> > >> ___ > >> users mailing list > >> us...@open-mpi.org > >> http://www.open-mpi.org/mailman/listinfo.cgi/users > >> > > > > > > -- > > Jeff Squyres > > jsquy...@cisco.com > > > > For corporate legal information go to: > > http://www.cisco.com/web/about/doing_business/legal/cri/ > > > > > > ___ > > users mailing list > > us...@open-mpi.org > > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/
Re: [OMPI users] Problems building Open MPI 1.4.1 with Pathscale
All, FWIW, Pathscale is dying in the new atomics in 1.4.1 (and svn trunk) - actually looping - from gdb: opal_progress_event_users_decrement () at ../.././opal/include/opal/sys/atomic_impl.h:61 61 } while (0 == opal_atomic_cmpset_32(addr, oldval, oldval - delta)); Current language: auto; currently asm (gdb) where #0 opal_progress_event_users_decrement () at ../.././opal/include/opal/sys/atomic_impl.h:61 #1 0x0001 in ?? () #2 0x2aec4cf6a5e0 in ?? () #3 0x00eb in ?? () #4 0x2aec4cfb57e0 in ompi_mpi_init () at ../.././ompi/runtime/ompi_mpi_init.c:818 #5 0x7fff5db3bd58 in ?? () Backtrace stopped: previous frame inner to this frame (corrupt stack?) (gdb) list 56 { 57 int32_t oldval; 58 59 do { 60oldval = *addr; 61 } while (0 == opal_atomic_cmpset_32(addr, oldval, oldval - delta)); 62 return (oldval - delta); 63 } 64 #endif /* OPAL_HAVE_ATOMIC_SUB_32 */ 65 (gdb) DM On Tue, 9 Feb 2010, Jeff Squyres wrote: FWIW, I have had terrible luck with the patschale compiler over the years. Repeated attempts to get support from them -- even when I was a paying customer -- resulted in no help (e.g., a pathCC bug with the OMPI C++ bindings that I filed years ago was never resolved). Is this compiler even supported anymore? I.e., is there a support department somewhere that you have a hope of getting any help from? I can't say for sure, of course, but if MPI hello world hangs, it smells like a compiler bug. You might want to attach to "hello world" in a debugger and see where it's hung. You might need to compile OMPI with debugging symbols to get any meaningful information. ** NOTE: My personal feelings about the pathscale compiler suite do not reflect anyone else's feelings in the Open MPI community. Perhaps someone could change my mind someday, but *I* have personally given up on this compiler. :-( On Feb 8, 2010, at 2:38 AM, Rafael Arco Arredondo wrote: Hello, It does work with version 1.4. This is the hello world that hangs with 1.4.1: #include #include int main(int argc, char **argv) { int node, size; MPI_Init(,); MPI_Comm_rank(MPI_COMM_WORLD, ); MPI_Comm_size(MPI_COMM_WORLD, ); printf("Hello World from Node %d of %d.\n", node, size); MPI_Finalize(); return 0; } El mar, 26-01-2010 a las 03:57 -0500, ?ke Sandgren escribi?: 1 - Do you have problems with openmpi 1.4 too? (I don't, haven't built 1.4.1 yet) 2 - There is a bug in the pathscale compiler with -fPIC and -g that generates incorrect dwarf2 data so debuggers get really confused and will have BIG problems debugging the code. I'm chasing them to get a fix... 3 - Do you have an example code that have problems? -- Rafael Arco Arredondo Centro de Servicios de Inform?tica y Redes de Comunicaciones Universidad de Granada ___ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/ ___ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users
Re: [OMPI users] ompi_info loop
On Tue, Feb/09/2010 08:46:53AM, Benjamin Gaudio wrote: > In trying to track down my default hostfile problem, I found that > when I run ompi_info, it simply keeps repeating: > > Displaying Open MPI information for 32-bit ... > Displaying Open MPI information for 32-bit ... > Displaying Open MPI information for 32-bit ... > Displaying Open MPI information for 32-bit ... > Displaying Open MPI information for 32-bit ... > Displaying Open MPI information for 32-bit ... > > Has anyone seen this before? I am using Sun HPC ClusterTools 8.2.1 > on Solaris 10. Hi Benjamin, I can reproduce this using /opt/SUNWhpc/HPC8.2.1/sun/bin/64/ompi_info. We'll fix this ASAP (for the next ClusterTools release). A workaround is to use /opt/SUNWhpc/HPC8.2.1/sun/bin/32/ompi_info. Thanks, Ethan > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users
Re: [OMPI users] Problems building Open MPI 1.4.1 with Pathscale
On Tue, 2010-02-09 at 08:49 -0500, Jeff Squyres wrote: > FWIW, I have had terrible luck with the patschale compiler over the years. > Repeated attempts to get support from them -- even when I was a paying > customer -- resulted in no help (e.g., a pathCC bug with the OMPI C++ > bindings that I filed years ago was never resolved). > > Is this compiler even supported anymore? I.e., is there a support department > somewhere that you have a hope of getting any help from? > > I can't say for sure, of course, but if MPI hello world hangs, it smells like > a compiler bug. You might want to attach to "hello world" in a debugger and > see where it's hung. You might need to compile OMPI with debugging symbols > to get any meaningful information. > > ** NOTE: My personal feelings about the pathscale compiler suite do not > reflect anyone else's feelings in the Open MPI community. Perhaps someone > could change my mind someday, but *I* have personally given up on this > compiler. :-( Pathscale is not dead, in fact I'm talking to them more or less daily at the moment. They have been restructuring since the demise of SciCortex last year. I hope they will be able to release a new version fairly soon. In my opinion (working mostly with Fortran codes, shudder) it is the best compiler around. Although they have had problems over the years in coming out with fixes for bugs in a timely fashion. -- Ake Sandgren, HPC2N, Umea University, S-90187 Umea, Sweden Internet: a...@hpc2n.umu.se Phone: +46 90 7866134 Fax: +46 90 7866126 Mobile: +46 70 7716134 WWW: http://www.hpc2n.umu.se
Re: [OMPI users] Problems building Open MPI 1.4.1 with Pathscale
FWIW, I have had terrible luck with the patschale compiler over the years. Repeated attempts to get support from them -- even when I was a paying customer -- resulted in no help (e.g., a pathCC bug with the OMPI C++ bindings that I filed years ago was never resolved). Is this compiler even supported anymore? I.e., is there a support department somewhere that you have a hope of getting any help from? I can't say for sure, of course, but if MPI hello world hangs, it smells like a compiler bug. You might want to attach to "hello world" in a debugger and see where it's hung. You might need to compile OMPI with debugging symbols to get any meaningful information. ** NOTE: My personal feelings about the pathscale compiler suite do not reflect anyone else's feelings in the Open MPI community. Perhaps someone could change my mind someday, but *I* have personally given up on this compiler. :-( On Feb 8, 2010, at 2:38 AM, Rafael Arco Arredondo wrote: > Hello, > > It does work with version 1.4. This is the hello world that hangs with > 1.4.1: > > #include > #include > > int main(int argc, char **argv) > { > int node, size; > > MPI_Init(,); > MPI_Comm_rank(MPI_COMM_WORLD, ); > MPI_Comm_size(MPI_COMM_WORLD, ); > > printf("Hello World from Node %d of %d.\n", node, size); > > MPI_Finalize(); > return 0; > } > > El mar, 26-01-2010 a las 03:57 -0500, Åke Sandgren escribió: > > 1 - Do you have problems with openmpi 1.4 too? (I don't, haven't built > > 1.4.1 yet) > > 2 - There is a bug in the pathscale compiler with -fPIC and -g that > > generates incorrect dwarf2 data so debuggers get really confused and > > will have BIG problems debugging the code. I'm chasing them to get a > > fix... > > 3 - Do you have an example code that have problems? > > -- > Rafael Arco Arredondo > Centro de Servicios de Informática y Redes de Comunicaciones > Universidad de Granada > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/
[OMPI users] ompi_info loop
In trying to track down my default hostfile problem, I found that when I run ompi_info, it simply keeps repeating: Displaying Open MPI information for 32-bit ... Displaying Open MPI information for 32-bit ... Displaying Open MPI information for 32-bit ... Displaying Open MPI information for 32-bit ... Displaying Open MPI information for 32-bit ... Displaying Open MPI information for 32-bit ... Has anyone seen this before? I am using Sun HPC ClusterTools 8.2.1 on Solaris 10.
Re: [OMPI users] Anybody built a working 1.4.1 on Solaris 8, (Sparc)?
Date: Fri, 05 Feb 2010 16:16:29 -0800 From: "David Mathog"We haven't tried Solaris 8 in quite some time. However, for your first > issue did you include the --enable-heterogeneous option on your > configure command? > > Since you are mix IA-32 and SPARC nodes you'll want to include this so > the endian issue doesn't bite you. Added that on the configure, rebuilt, installed, and now the examples work. Glad to hear that. Any thoughts on the Forte compiler issue? This is not quite as pressing now that the gcc version works, and most of the computation will be on the remote nodes anyway. Still, the Forte compilers should generate faster code than gcc, and I would prefer to use them if possible. Are you using the vampir tracing feature? If not I would add the "--enable-contrib-no-build=vt" option to your configure line and see if that works. Beware, I don't think we've tried compiling OMPI with Forte 7. Also note omp.h is for OpenMP not MPI. It looks like the last error you ran into with vt_unify.cc may be an issue with the Forte 7 C++ compiler. --td Thanks, David Mathog mat...@caltech.edu Manager, Sequence Analysis Facility, Biology Division, Caltech