Re: [OMPI users] Deadlock in netcdf tests

2019-10-26 Thread Orion Poplawski via users
to add running the netcdf test cases to my list of standard testsuites, but we didn't used to have any problems with them :-( Thanks for the report, we will be working on them! Edgar -Original Message- From: users [mailto:users-boun...@lists.open-mpi.org] On Behalf Of Orion Poplawski

Re: [OMPI users] Deadlock in netcdf tests

2019-10-25 Thread Orion Poplawski via users
vulcan component. That can be achieved by mpirun --mca fcoll ^vulcan ... or OMPI_MCA_fcoll=^vulcan mpirun ... I also noted the tst_parallel3 program crashes with the ROMIO component. Cheers, Gilles On 10/25/2019 12:55 PM, Orion Poplawski via users wrote: On 10/24/19 9:28 PM, Orion Poplawski via

Re: [OMPI users] Deadlock in netcdf tests

2019-10-24 Thread Orion Poplawski via users
On 10/24/19 9:28 PM, Orion Poplawski via users wrote: Starting with netcdf 4.7.1 (and 4.7.2) in Fedora Rawhide we are seeing a test hang with openmpi 4.0.2.  Backtrace: (gdb) bt #0  0x7f90c197529b in sched_yield () from /lib64/libc.so.6 #1  0x7f90c1ac8a05 in ompi_request_default_wait

[OMPI users] Deadlock in netcdf tests

2019-10-24 Thread Orion Poplawski via users
ons for debugging this would be greatly appreciated. -- Orion Poplawski Manager of NWRA Technical Systems 720-772-5637 NWRA, Boulder/CoRA Office FAX: 303-415-9702 3380 Mitchell Lane or...@nwra.com Boulder, CO 80301 https://www.nwra.com/ smime.

Re: [OMPI users] Hang in mpi on 32-bit

2018-11-27 Thread Orion Poplawski
It does not appear to have any effect, at least not with 2.1.5. Thanks. On 11/26/18 9:17 PM, Nathan Hjelm via users wrote: > Can you try configuring with —disable-builtin-atomics and see if that fixes > the issue for you? > > -Nathan > >> On Nov 26, 2018, at 9:11 PM, O

[OMPI users] Hang in mpi on 32-bit

2018-11-26 Thread Orion Poplawski
df5 1.8.20 and openmpi 3.1.3 on i686 with the "t_cache" test. https://copr-be.cloud.fedoraproject.org/results/@scitech/openmpi3.1/fedora-28-i386/00830432-hdf5/builder-live.log I'm at a loss as to how to debug this further. -- Orion Poplawski Manager of NWRA Technical Systems 720-77

Re: [OMPI users] Issues with different IB adapters and openmpi 2.0.2

2017-02-28 Thread Orion Poplawski
pacting the PML sanity > check. > > Howard > > > Orion Poplawski <or...@cora.nwra.com <mailto:or...@cora.nwra.com>> schrieb am > Mo. 27. Feb. 2017 um 14:50: > > We have a couple nodes with different IB adapters in them: > > font1/var/log/lspci

[OMPI users] Issues with different IB adapters and openmpi 2.0.2

2017-02-27 Thread Orion Poplawski
ve stopped alerting me to the issue. Thoughts? -- Orion Poplawski Technical Manager 720-772-5637 NWRA, Boulder/CoRA Office FAX: 303-415-9702 3380 Mitchell Lane or...@nwra.com Boulder, CO 80301 http://ww

Re: [OMPI users] SGE integration broken in 2.0.0

2016-08-17 Thread Orion Poplawski
This required modifying the mca_component_select function to actually check the return code Also do a little cleanup to avoid bombarding the user with multiple error messages. Thanks to Patrick Begou for reporting the problem -- Orion Poplawski Technical Manager 3

Re: [OMPI users] Hang in MPI_Abort

2016-06-30 Thread Orion Poplawski
On 06/30/2016 02:55 PM, Orion Poplawski wrote: > valgrind output: > > $ valgrind mpiexec -n 6 ./testphdf5 > ==8518== Memcheck, a memory error detector > ==8518== Copyright (C) 2002-2015, and GNU GPL'd, by Julian Seward et al. > ==8518== Using Valgrind-3.11.0 and Lib

Re: [OMPI users] Hang in MPI_Abort

2016-06-30 Thread Orion Poplawski
are all gone, but mpiexec isn’t exiting? I’d suggest > running valgrind, given the corruption. > >> On Jun 30, 2016, at 10:21 AM, Orion Poplawski <or...@cora.nwra.com> wrote: >> >> On 06/30/2016 10:33 AM, Orion Poplawski wrote: >>> No, just mpiexec is running. sin

Re: [OMPI users] Hang in MPI_Abort

2016-06-30 Thread Orion Poplawski
On 06/30/2016 10:33 AM, Orion Poplawski wrote: > No, just mpiexec is running. single node. Only see it when the test is > executed with "make check", not seeing it if I just run mpiexec -n 6 > ./testphdf5 by hand. Hmm, now I'm seeing it running mpiexec by hand. Trying

Re: [OMPI users] Hang in MPI_Abort

2016-06-30 Thread Orion Poplawski
n 30, 2016, at 8:49 AM, Orion Poplawski <or...@cora.nwra.com> wrote: >> >> I'm seeing hangs when MPI_Abort is called. This is with openmpi 1.10.3. >> e.g: >> >> program output: >> >> Testing -- big dataset test (bigdset) >> Proc 3:

Re: [OMPI users] Hang in MPI_Abort

2016-06-30 Thread Orion Poplawski
On 06/30/2016 09:49 AM, Orion Poplawski wrote: > I'm seeing hangs when MPI_Abort is called. This is with openmpi 1.10.3. e.g: I'll also note that I'm seeing this on 32-bit arm, but not i686 or x86_64. -- Orion Poplawski Technical Manager 303-415-9701 x222 NWRA, Boul

[OMPI users] Hang in MPI_Abort

2016-06-30 Thread Orion Poplawski
as a 6 process run on a 4 core machine. Thanks. -- Orion Poplawski Technical Manager 303-415-9701 x222 NWRA, Boulder/CoRA Office FAX: 303-415-9702 3380 Mitchell Lane or...@nwra.com Boulder, CO 80301 http://www.nwra.com

Re: [OMPI users] mpirun command won't run unless the firewalld daemon is disabled

2016-05-10 Thread Orion Poplawski
open your needed ports) or you can remove firewalld and install iptables-services and go back to the old iptables method of configuring the firewall. If you don't want a firewall at all, just remove firewalld. -- Orion Poplawski Technical Manager 303-415-9701 x222 NWRA, Boulder

[OMPI users] 1.10.1 appears to break mpi4py

2015-11-09 Thread Orion Poplawski
mm.pyx", line 815, in mpi4py.MPI.Comm.Iscatterv (src/mpi4py.MPI.c:101257) Exception: MPI_ERR_SIZE: invalid size ====== ERROR: testScatterv3 (test_cco_nb_vec.TestCCOVecWorldDup) ---

[OMPI users] Force mpirun to only run under gridengine

2013-06-04 Thread Orion Poplawski
I'd like to be able to force mpirun to require being run under a gridengine environment. Any ideas on how to achieve this, if possible? -- Orion Poplawski Technical Manager 303-415-9701 x222 NWRA, Boulder/CoRA Office FAX: 303-415-9702 3380 Mitchell Lane

[OMPI users] dmtcp status

2013-05-21 Thread Orion Poplawski
What the status of dmtcp support in opemmpi? Is this still under development? I see some configure checks in 1.7 but they don't really seem to do anything. -- Orion Poplawski Technical Manager 303-415-9701 x222 NWRA, Boulder/CoRA Office FAX: 303-415-9702 3380

[OMPI users] uDAPL status

2013-02-22 Thread Orion Poplawski
value... simple ok (unspecified) configure: WARNING: On Linux and --with-udapl was not specified configure: WARNING: Not building the udapl BTL Should it still emit this warning? -- Orion Poplawski Technical Manager 303-415-9701 x222 NWRA, Boulder Office FAX

Re: [OMPI users] [Open MPI Announce] Open MPI v1.6.3 released

2012-11-05 Thread Orion Poplawski
onfigure seems more straightforward (it's what I'm doing in the Fedora package). -- Orion Poplawski Technical Manager 303-415-9701 x222 NWRA, Boulder Office FAX: 303-415-9702 3380 Mitchell Lane or...@nwra.com Boulder, CO 80301

Re: [OMPI users] make check fails with OpenMPI 1.6.3 and Intel Compilers

2012-11-02 Thread Orion Poplawski
this issue before? Seems to be working okay for me with icc (ICC) 13.0.1 20121010 x86_64, though perhaps it's dependent on compile flags... -- Orion Poplawski Technical Manager 303-415-9701 x222 NWRA, Boulder Office FAX: 303-415-9702 3380 Mitchell Lane

Re: [OMPI users] [Open MPI Announce] Open MPI v1.6.3 released

2012-11-02 Thread Orion Poplawski
bump for libmpi_f90.so from .1 to .3 so it's not strictly ABI compatible at least for Fortran code - unless the bump was a mistake? -- Orion Poplawski Technical Manager 303-415-9701 x222 NWRA, Boulder Office FAX: 303-415-9702 3380 Mitchell Lane

[OMPI users] opal_timer_linux_open() setting inexact floating point exception

2012-06-14 Thread Orion Poplawski
. - Orion -- Orion Poplawski Technical Manager 303-415-9701 x222 NWRA, Boulder Office FAX: 303-415-9702 3380 Mitchell Lane or...@nwra.com Boulder, CO 80301 http://www.nwra.com

Re: [OMPI users] [EXTERNAL] Re: mpicc link shouldn't add -ldl and -lhwloc

2012-05-24 Thread Orion Poplawski
On 05/23/2012 07:29 AM, Barrett, Brian W wrote: On 5/22/12 10:36 PM, "Orion Poplawski"<or...@cora.nwra.com> wrote: On 05/22/2012 10:34 PM, Orion Poplawski wrote: On 05/21/2012 06:15 PM, Jeff Squyres wrote: On May 15, 2012, at 10:37 AM, Orion Poplawski wrote: $ mpicc -show

Re: [OMPI users] mpicc link shouldn't add -ldl and -lhwloc

2012-05-23 Thread Orion Poplawski
On 05/22/2012 10:34 PM, Orion Poplawski wrote: On 05/21/2012 06:15 PM, Jeff Squyres wrote: On May 15, 2012, at 10:37 AM, Orion Poplawski wrote: $ mpicc -showme:link -pthread -m64 -L/usr/lib64/openmpi/lib -lmpi -ldl -lhwloc -ldl and -lhwloc should not be listed. The user should only link

Re: [OMPI users] mpicc link shouldn't add -ldl and -lhwloc

2012-05-23 Thread Orion Poplawski
On 05/21/2012 06:15 PM, Jeff Squyres wrote: On May 15, 2012, at 10:37 AM, Orion Poplawski wrote: $ mpicc -showme:link -pthread -m64 -L/usr/lib64/openmpi/lib -lmpi -ldl -lhwloc -ldl and -lhwloc should not be listed. The user should only link against libraries that they are using directly

Re: [OMPI users] mpicc link shouldn't add -ldl and -lhwloc

2012-05-21 Thread Orion Poplawski
On 05/15/2012 10:37 AM, Orion Poplawski wrote: See https://bugzilla.redhat.com/show_bug.cgi?id=814798 $ mpicc -showme:link -pthread -m64 -L/usr/lib64/openmpi/lib -lmpi -ldl -lhwloc -ldl and -lhwloc should not be listed. The user should only link against libraries that they are using directly

[OMPI users] mpicc link shouldn't add -ldl and -lhwloc

2012-05-15 Thread Orion Poplawski
. This report is with openmpi-1.5.5. Haven't tested with 1.6, but configure looks to do the same in it. Comments? - Orion -- Orion Poplawski Technical Manager 303-415-9701 x222 NWRA, Boulder Office FAX: 303-415-9702 3380 Mitchell Lane

[OMPI users] Trouble using rankfile with gridengine

2010-04-23 Thread Orion Poplawski
plm_rsh_module.c at line 990 -- A daemon (pid unknown) died unexpectedly on signal 1 while attempting to launch so we are aborting. Any ideas? Thanks! - Orion -- Orion Poplawski Technical Manager 303

Re: [OMPI users] PGI Fortran pthread support

2009-04-14 Thread Orion Poplawski
gure tests. -- Orion Poplawski Technical Manager 303-415-9701 x222 NWRA/CoRA DivisionFAX: 303-415-9702 3380 Mitchell Lane or...@cora.nwra.com Boulder, CO 80301 http://www.cora.nwra.com

Re: [OMPI users] PGI Fortran pthread support

2009-04-14 Thread Orion Poplawski
Thanks, that explains the build failures. -- Orion Poplawski Technical Manager 303-415-9701 x222 NWRA/CoRA DivisionFAX: 303-415-9702 3380 Mitchell Lane or...@cora.nwra.com Boulder, CO 80301 http://www.cora.nwra.com

Re: [OMPI users] PGI Fortran pthread support

2009-04-14 Thread Orion Poplawski
Orion Poplawski wrote: Looks like libtool is adding -pthread because it sees that you use -pthread to link C programs and assumes that all linkers use it. Sorry, it inherits it from libmpi.la. I hate libtool. -- Orion Poplawski Technical Manager 303-415-9701 x222 NWRA

Re: [OMPI users] PGI Fortran pthread support

2009-04-14 Thread Orion Poplawski
Orion Poplawski wrote: ./configure LIBS=-lgcc_eh ... did the trick. Spoke too soon. This leads to: /bin/sh ../../../libtool --mode=link pgf90 -I../../../ompi/include -I../../../ompi/include -I. -I. -I../../../ompi/mpi/f90 -fastsse -fPIC -export-dynamic -Wl,-z,noexecstack -o

Re: [OMPI users] PGI Fortran pthread support

2009-04-14 Thread Orion Poplawski
Orion Poplawski wrote: Looks like I need link to -lgcc_eh some how. ./configure LIBS=-lgcc_eh ... did the trick. checking if F77 compiler and POSIX threads work as is... yes checking if C compiler and POSIX threads work with -Kthread... no checking if C compiler and POSIX threads work

Re: [OMPI users] PGI Fortran pthread support

2009-04-14 Thread Orion Poplawski
-fastsse -fPIC conftestf.f conftest.o -o conftest -Wl,-z,noexecstack -lnsl -lutil -lm conftestf.f: conftest.o:(.data.DW.ref.__gcc_personality_v0[DW.ref.__gcc_personality_v0]+0x0): undefined reference to `__gcc_personality_v0' Looks like I need link to -lgcc_eh some how. -- Orion Poplawski

[OMPI users] PGI Fortran pthread support

2009-04-13 Thread Orion Poplawski
threads for openmpi? -- Orion Poplawski Technical Manager 303-415-9701 x222 NWRA/CoRA DivisionFAX: 303-415-9702 3380 Mitchell Lane or...@cora.nwra.com Boulder, CO 80301 http://www.cora.nwra.com