Re: [OMPI users] MPI-IO on Lustre - OMPIO or ROMIO?

2020-11-23 Thread Howard Pritchard via users
HI All, I opened a new issue to track the coll_perf failure in case its not related to the HDF5 problem reported earlier. https://github.com/open-mpi/ompi/issues/8246 Howard Am Mo., 23. Nov. 2020 um 12:14 Uhr schrieb Dave Love via users < users@lists.open-mpi.org>: > Mark Dixon via users

Re: [OMPI users] OMPI 4.0.4 crashes (or hangs) with dynamically processes allocation. OMPI 4.0.1 don't.

2020-08-15 Thread Howard Pritchard via users
* MPI_ERR_UNKNOWN: unknown error* > > *[osboxes:02457] *** MPI_ERRORS_ARE_FATAL (processes in this communicator > will now abort,* > > *[osboxes:02457] ***and potentially your MPI job)* > > *[osboxes:02458] 1 more process has sent help message help-orted.txt / > timedout*

Re: [OMPI users] OMPI 4.0.4 crashes (or hangs) with dynamically processes allocation. OMPI 4.0.1 don't.

2020-08-14 Thread Howard Pritchard via users
> I need spawn only in “worker”. Is there a way or workaround for doing this > without mpirun? > > Thanks a lot for your assistance. > > > > Martín > > > > > > > > > > *From: *Howard Pritchard > *Sent: *lunes, 10 de agosto de 2020 19:13 > *To:

Re: [OMPI users] OMPI 4.0.4 crashes (or hangs) with dynamically processes allocation. OMPI 4.0.1 don't.

2020-08-13 Thread Howard Pritchard via users
und for doing this > without mpirun? > Thanks a lot for your assistance. > > Martín > > > > > *From: *Howard Pritchard > *Sent: *lunes, 10 de agosto de 2020 19:13 > *To: *Martín Morales > *Cc: *Open MPI Users > *Subject: *Re: [OMPI users] OMPI 4.0.4 crashes

Re: [OMPI users] OMPI 4.0.4 crashes (or hangs) with dynamically processes allocation. OMPI 4.0.1 don't.

2020-08-10 Thread Howard Pritchard via users
rtunately the issue persists in OMPI 4.0.5rc1. Do I have > to post this on the bug section? Thanks and regards. > > > > Martín > > > > *From: *Howard Pritchard > *Sent: *lunes, 10 de agosto de 2020 14:44 > *To: *Open MPI Users > *Cc: *Martín Morales >

Re: [OMPI users] OMPI 4.0.4 crashes (or hangs) with dynamically processes allocation. OMPI 4.0.1 don't.

2020-08-10 Thread Howard Pritchard via users
Hello Martin, Between Open MPI 4.0.1 and Open MPI 4.0.4 we upgraded the internal PMIx version that introduced a problem with spawn for the 4.0.2-4.0.4 versions. This is supposed to be fixed in the 4.0.5 release. Could you try the 4.0.5rc1 tarball and see if that addresses the problem you're

Re: [OMPI users] Differences 4.0.3 -> 4.0.4 (Regression?)

2020-08-08 Thread Howard Pritchard via users
Hello Michael, Not sure what could be causing this in terms of delta between v4.0.3 and v4.0.4. Two things to try - add --debug-daemons and --mca pmix_base_verbose 100 to the mpirun line and compare output from the v4.0.3 and v4.0.4 installs - perhaps try using the

Re: [OMPI users] OMPI returns error 63 on AMD 7742 when utilizing 100+ processors per node

2020-01-29 Thread Howard Pritchard via users
itialize with >100 processes per > node. I get the same error message for multiple different codes, so the > error code is mpi related rather than being program specific. > > > > Collin > > > > *From:* Howard Pritchard > *Sent:* Monday, January 27, 2020 11:20

Re: [OMPI users] OMPI returns error 63 on AMD 7742 when utilizing 100+ processors per node

2020-01-27 Thread Howard Pritchard via users
Hello Collen, Could you provide more information about the error. Is there any output from either Open MPI or, maybe, UCX, that could provide more information about the problem you are hitting? Howard Am Mo., 27. Jan. 2020 um 08:38 Uhr schrieb Collin Strassburger via users <

Re: [OMPI users] Do idle MPI threads consume clock cycles?

2019-02-25 Thread Howard Pritchard
Hello Mark, You may want to checkout this package: https://github.com/lanl/libquo Another option would be to do something like use an MPI_Ibarrier in the application with all the MPI processes but rank 0 going into a loop over waiting for completion of the barrier and doing a sleep. Once rank

Re: [OMPI users] OpenMPI v4.0.0 signal 11 (Segmentation fault)

2019-02-20 Thread Howard Pritchard
] IMB-MPI1[0x401d49] >> > [phoebe:07408] *** End of error message *** >> > IMB-MPI1[0x4022ea] >> > [titan:07169] [ 8] >> /usr/lib64/libc.so.6(__libc_start_main+0xf5)[0x7fc025d5a3d5] >> > [titan:07169] [ 9] IMB-MPI1[0x401d49] >> > [titan:07169] *

Re: [OMPI users] OpenMPI v4.0.0 signal 11 (Segmentation fault)

2019-02-20 Thread Howard Pritchard
HI Adam, As a sanity check, if you try to use --mca btl self,vader,tcp do you still see the segmentation fault? Howard Am Mi., 20. Feb. 2019 um 08:50 Uhr schrieb Adam LeBlanc < alebl...@iol.unh.edu>: > Hello, > > When I do a run with OpenMPI v4.0.0 on Infiniband with this command: > mpirun

Re: [OMPI users] Help Getting Started with Open MPI and PMIx and UCX

2019-01-20 Thread Howard Pritchard
Hi Matt Definitely do not include the ucx option for an omnipath cluster. Actually if you accidentally installed ucx in it’s default location use on the system Switch to this config option —with-ucx=no Otherwise you will hit https://github.com/openucx/ucx/issues/750 Howard Gilles

Re: [OMPI users] Segmentation fault using openmpi-master-201901030305-ee26ed9

2019-01-04 Thread Howard Pritchard
Hi Sigmar, I observed this problem yesterday myself and should have a fix in to master later today. Howard Am Fr., 4. Jan. 2019 um 05:30 Uhr schrieb Siegmar Gross < siegmar.gr...@informatik.hs-fulda.de>: > Hi, > > I've installed (tried to install) openmpi-master-201901030305-ee26ed9 on > my

Re: [OMPI users] Unable to build Open MPI with external PMIx library support

2018-12-17 Thread Howard Pritchard
th", is set and doesn't point to any location > that includes at least one usable plugin for this framework. > > Please check your installation and environment. > ------ > > Regards, > Eduardo > &g

Re: [OMPI users] Unable to build Open MPI with external PMIx library support

2018-12-15 Thread Howard Pritchard
Hi Eduardo Could you post the config.log for the build with internal PMIx so we can figure that out first. Howard Eduardo Rothe via users schrieb am Fr. 14. Dez. 2018 um 09:41: > Open MPI: 4.0.0 > PMIx: 3.0.2 > OS: Debian 9 > > I'm building a debian package for Open MPI and either I get the

Re: [OMPI users] [Open MPI Announce] Open MPI 4.0.0 Released

2018-11-14 Thread Howard Pritchard
Hi Bert, If you'd prefer to return to the land of convenience and don't need to mix MPI and OpenSHMEM, then you may want to try the path I outlined in the email archived at the following link https://www.mail-archive.com/users@lists.open-mpi.org/msg32274.html Howard Am Di., 13. Nov. 2018 um

Re: [OMPI users] [Open MPI Announce] Open MPI 4.0.0 Released

2018-11-13 Thread Howard Pritchard
Hello Bert, What OS are you running on your notebook? If you are running Linux, and you have root access to your system, then you should be able to resolve the Open SHMEM support issue by installing the XPMEM device driver on your system, and rebuilding UCX so it picks up XPMEM support. The

Re: [OMPI users] [EXTERNAL] Re: OpenMPI 3.1.0 Lock Up on POWER9 w/ CUDA9.2

2018-07-02 Thread Howard Pritchard
HI Si, Could you add --disable-builtin-atomics to the configure options and see if the hang goes away? Howard 2018-07-02 8:48 GMT-06:00 Jeff Squyres (jsquyres) via users < users@lists.open-mpi.org>: > Simon -- > > You don't currently have another Open MPI installation in your PATH / >

Re: [OMPI users] A couple of general questions

2018-06-14 Thread Howard Pritchard
Hello Charles You are heading in the right direction. First you might want to run the libfabric fi_info command to see what capabilities you picked up from the libfabric RPMs. Next you may well not actually be using the OFI mtl. Could you run your app with export

Re: [OMPI users] Problem running with UCX/oshmem on single node?

2018-05-09 Thread Howard Pritchard
Hi Craig, You are experiencing problems because you don't have a transport installed that UCX can use for oshmem. You either need to go and buy a connectx4/5 HCA from mellanox (and maybe a switch), and install that on your system, or else install xpmem (https://github.com/hjelmn/xpmem). Note

Re: [OMPI users] Debug build of v3.0.1 tarball

2018-05-04 Thread Howard Pritchard
i.org> > *Date: *Friday, May 4, 2018 at 7:46 AM > *To: *Open MPI Users <users@lists.open-mpi.org> > *Subject: *Re: [OMPI users] Debug build of v3.0.1 tarball > > > > HI Adam, > > > > Sorry didn't notice you did try the --enable-debug flag. That should not

Re: [OMPI users] Debug build of v3.0.1 tarball

2018-05-04 Thread Howard Pritchard
HI Adam, Sorry didn't notice you did try the --enable-debug flag. That should not have led to the link error building the opal dso. Did you do a make clean after rerunning configure? Howard 2018-05-04 8:22 GMT-06:00 Howard Pritchard <hpprit...@gmail.com>: > Hi Adam, > > Di

Re: [OMPI users] Debug build of v3.0.1 tarball

2018-05-04 Thread Howard Pritchard
Hi Adam, Did you try using the --enable-debug configure option along with your CFLAGS options? You may want to see if that simplifies your build. In any case, we'll fix the problems you found. Howard 2018-05-03 15:00 GMT-06:00 Moody, Adam T. : > Hello Open MPI team, > > I'm

Re: [OMPI users] Eager RDMA causing slow osu_bibw with 3.0.0

2018-04-05 Thread Howard Pritchard
Hello Ben, Thanks for the info. You would probably be better off installing UCX on your cluster and rebuilding your Open MPI with the --with-ucx configure option. Here's what I'm seeing with Open MPI 3.0.1 on a ConnectX5 based cluster using ob1/openib BTL: mpirun -map-by ppr:1:node -np 2

Re: [OMPI users] OpenMPI with Portals4 transport

2018-02-08 Thread Howard Pritchard
27.10 > 262144 46.97 > 524288 87.55 > 1048576 168.89 > 2097152 331.40 > 4194304 654.08 > > > On Feb 7, 2018, at 9:04 PM, Howard Pritchard <hpprit...@gmail.com> wrote: > > HI Brian, > > A

Re: [OMPI users] Using OpenSHMEM with Shared Memory

2018-02-07 Thread Howard Pritchard
HI Ben, I'm afraid this is bad news for using UCX. The problem is that when UCX was configured/built, it did not find a transport for doing one sided put/get transfers. If you're feeling lucky, you may want to install xpmem (https://github.com/hjelmn/xpmem) and rebuild UCX. This requires

Re: [OMPI users] OpenMPI with Portals4 transport

2018-02-07 Thread Howard Pritchard
HI Brian, As a sanity check, can you see if the ob1 pml works okay, i.e. mpirun -n 2 --mca pml ob1 --mca btl self,vader,openib ./osu_latency Howard 2018-02-07 11:03 GMT-07:00 brian larkins : > Hello, > > I’m doing some work with Portals4 and am trying to run some MPI

Re: [OMPI users] Using OpenSHMEM with Shared Memory

2018-02-07 Thread Howard Pritchard
HI Ben, Could you set these environment variables and post the output ? export OMPI_MCA_spml=ucx export OMPI_MCA_spml_base_verbose=100 then run your test? Also, what OS are you using? Howard 2018-02-06 20:10 GMT-07:00 Jeff Hammond : > > On Tue, Feb 6, 2018 at 3:58

Re: [OMPI users] About my GPU performance using Openmpi-2.0.4

2017-12-13 Thread Howard Pritchard
Hi Phanikumar It’s unlikely the warning message you are seeing is related to GPU performance. Have you tried adding —with-verbs=no to your config line? That should quash openib complaint. Howard Phanikumar Pentyala schrieb am Mo. 11. Dez. 2017 um 22:43: > Dear

Re: [OMPI users] [EXTERNAL] Re: Using shmem_int_fadd() in OpenMPI\'s SHMEM

2017-11-22 Thread Howard Pritchard
Hi Ben, Actually I did some checking about the brew install for OFi libfabric. It looks like if your brew is up to date, it will pick up libfabric 1.5.2. Howard 2017-11-22 15:21 GMT-07:00 Howard Pritchard <hpprit...@gmail.com>: > HI Ben, > > Even on one box, the yoda compone

Re: [OMPI users] [EXTERNAL] Re: Using shmem_int_fadd() in OpenMPI\'s SHMEM

2017-11-22 Thread Howard Pritchard
HI Ben, Even on one box, the yoda component doesn't work any more. If you want to do OpenSHMEM programming on you Macbook pro (like I do) and you don't want to set up a VM to use UCX, then you can use Sandia OpenSHMEM implementation. https://github.com/Sandia-OpenSHMEM/SOS You will need to

Re: [OMPI users] [EXTERNAL] Re: Using shmem_int_fadd() in OpenMPI's SHMEM

2017-11-22 Thread Howard Pritchard
found > > [shepard-lsm1:49499] Set MCA parameter "orte_base_help_aggregate" to 0 to > see all help / error messages > > > > -- > > Si Hammond > > Scalable Computer Architectures > > Sandia National Laboratories, NM, USA > > > > > > *From

Re: [OMPI users] Using shmem_int_fadd() in OpenMPI's SHMEM

2017-11-20 Thread Howard Pritchard
HI Ben, What version of Open MPI are you trying to use? Also, could you describe something about your system. If its a cluster what sort of interconnect is being used. Howard 2017-11-20 14:13 GMT-07:00 Benjamin Brock : > What's the proper way to use shmem_int_fadd()

Re: [OMPI users] Problems building OpenMPI 2.1.1 on Intel KNL

2017-11-20 Thread Howard Pritchard
Hello Ake, Would you mind opening an issue on Github so we can track this? https://github.com/open-mpi/ompi/issues There's a template to show what info we need to fix this. Thanks very much for reporting this, Howard 2017-11-20 3:26 GMT-07:00 Åke Sandgren : > Hi!

Re: [OMPI users] OMPI 2.1.2 and SLURM compatibility

2017-11-17 Thread Howard Pritchard
Hello Bennet, What you are trying to do using srun as the job launcher should work. Could you post the contents of /etc/slurm/slurm.conf for your system? Could you also post the output of the following command: ompi_info --all | grep pmix to the mail list. the config.log from your build would

Re: [OMPI users] [OMPI devel] Open MPI 2.0.4rc2 available for testing

2017-11-02 Thread Howard Pritchard
ort2/src/openmpi-2.0.4/op > enmpi-2.0.4rc2-Linux.x86_64.64_cc/opal/mca/pmix/pmix112' > Makefile:2301: recipe for target 'all-recursive' failed > make[1]: *** [all-recursive] Error 1 > make[1]: Leaving directory '/export2/src/openmpi-2.0.4/op > enmpi-2.0.4rc2-Linux.x86_64.64_cc/opal' >

Re: [OMPI users] Strange benchmarks at large message sizes

2017-09-19 Thread Howard Pritchard
Hello Cooper Could you rerun your test with the following env. variable set export OMPI_MCA_coll=self,basic,libnbc and see if that helps? Also, what type of interconnect are you using - ethernet, IB, ...? Howard 2017-09-19 8:56 GMT-06:00 Cooper Burns : >

Re: [OMPI users] openmpi-2.1.2rc2: warnings from "make" and "make check"

2017-08-30 Thread Howard Pritchard
Hi Siegmar, Opened issue 4151 to track this. Thanks, Howard 2017-08-21 7:13 GMT-06:00 Siegmar Gross < siegmar.gr...@informatik.hs-fulda.de>: > Hi, > > I've installed openmpi-2.1.2rc2 on my "SUSE Linux Enterprise Server 12.2 > (x86_64)" with Sun C 5.15 (Oracle Developer Studio 12.6) and

Re: [OMPI users] openmpi-master-201708190239-9d3f451: warnings from "make" and "make check"

2017-08-30 Thread Howard Pritchard
Hi Siegmar, I opened issue 4151 to track this. This is relevant to a project to get open mpi to build with -Werror. Thanks very much, Howard 2017-08-21 7:27 GMT-06:00 Siegmar Gross < siegmar.gr...@informatik.hs-fulda.de>: > Hi, > > I've installed openmpi-master-201708190239-9d3f451 on my

Re: [OMPI users] pmix, lxc, hpcx

2017-05-26 Thread Howard Pritchard
Hi John, In the 2.1.x release stream a shared memory capability was introduced into the PMIx component. I know nothing about LXC containers, but it looks to me like there's some issue when PMIx tries to create these shared memory segments. I'd check to see if there's something about your

Re: [OMPI users] Openmpi 1.10.4 crashes with 1024 processes

2017-03-22 Thread Howard Pritchard
Forgot you probably need an equal sign after btl arg Howard Pritchard <hpprit...@gmail.com> schrieb am Mi. 22. März 2017 um 18:11: > Hi Goetz > > Thanks for trying these other versions. Looks like a bug. Could you post > the config.log output from your build of the 2.1.0 to

Re: [OMPI users] Openmpi 1.10.4 crashes with 1024 processes

2017-03-22 Thread Howard Pritchard
) Howard Götz Waschk <goetz.was...@gmail.com> schrieb am Mi. 22. März 2017 um 13:09: On Wed, Mar 22, 2017 at 7:46 PM, Howard Pritchard <hpprit...@gmail.com> wrote: > Hi Goetz, > > Would you mind testing against the 2.1.0 release or the latest from the > 1.10.x series (1.1

Re: [OMPI users] Openmpi 1.10.4 crashes with 1024 processes

2017-03-22 Thread Howard Pritchard
Hi Goetz, Would you mind testing against the 2.1.0 release or the latest from the 1.10.x series (1.10.6)? Thanks, Howard 2017-03-22 6:25 GMT-06:00 Götz Waschk : > Hi everyone, > > I'm testing a new machine with 32 nodes of 32 cores each using the IMB > benchmark. It

Re: [OMPI users] Shared Windows and MPI_Accumulate

2017-03-03 Thread Howard Pritchard
ut of ompi_info of master on my laptop. Please let me > know if I can help with anything else. > > Thanks, > Joseph > > On 03/01/2017 11:24 PM, Howard Pritchard wrote: > > Hi Joseph, > > I built this test with craypich (Cray MPI) and it passed. I also tried > with Ope

Re: [OMPI users] sharedfp/lockedfile collision between multiple program instances

2017-03-03 Thread Howard Pritchard
Hi Edgar Please open an issue too so we can track the fix. Howard Edgar Gabriel schrieb am Fr. 3. März 2017 um 07:45: > Nicolas, > > thank you for the bug report, I can confirm the behavior. I will work on > a patch and will try to get that into the next release,

Re: [OMPI users] Shared Windows and MPI_Accumulate

2017-03-01 Thread Howard Pritchard
Hi Joseph, I built this test with craypich (Cray MPI) and it passed. I also tried with Open MPI master and the test passed. I also tried with 2.0.2 and can't seem to reproduce on my system. Could you post the output of config.log? Also, how intermittent is the problem? Thanks, Howard

Re: [OMPI users] Issues with different IB adapters and openmpi 2.0.2

2017-02-27 Thread Howard Pritchard
Hi Orion Does the problem occur if you only use font2 and 3? Do you have MXM installed on the font1 node? The 2.x series is using PMIX and it could be that is impacting the PML sanity check. Howard Orion Poplawski schrieb am Mo. 27. Feb. 2017 um 14:50: > We have a

Re: [OMPI users] MPI_THREAD_MULTIPLE: Fatal error on MPI_Win_create

2017-02-18 Thread Howard Pritchard
Hi Joseph What OS are you using when running the test? Could you try running with export OMPI_mca_osc=^pt2pt and export OMPI_mca_osc_base_verbose=10 This error message was put in to this OMPI release because this part of the code has known problems when used multi threaded. Joseph Schuchart

Re: [OMPI users] Problem with MPI_Comm_spawn using openmpi 2.0.x + sbatch

2017-02-15 Thread Howard Pritchard
Hi Anastasia, Definitely check the mpirun when in batch environment but you may also want to upgrade to Open MPI 2.0.2. Howard r...@open-mpi.org schrieb am Mi. 15. Feb. 2017 um 07:49: > Nothing immediate comes to mind - all sbatch does is create an allocation > and then run

Re: [OMPI users] OpenMPI not running any job on Mac OS X 10.12

2017-02-06 Thread Howard Pritchard
value Bad parameter (-5) instead of ORTE_SUCCESS -- On Thu, Feb 2, 2017 at 12:29 PM, Howard Pritchard <hpprit...@gmail.com> wrote: Hi Michel Try adding --enable-static to the configure. That fixed the problem for me.

Re: [OMPI users] Open MPI over RoCE using breakout cable and switch

2017-02-03 Thread Howard Pritchard
t; > Brendan > > > > *From:* users [mailto:users-boun...@lists.open-mpi.org > <users-boun...@lists.open-mpi.org>] *On Behalf Of *Howard Pritchard > *Sent:* Tuesday, January 24, 2017 8:21 AM > *To:* Open MPI Users <users@lists.open-mpi.org> > *Subject:* Re: [OMP

Re: [OMPI users] OpenMPI not running any job on Mac OS X 10.12

2017-02-02 Thread Howard Pritchard
Hi Michel Try adding --enable-static to the configure. That fixed the problem for me. Howard Michel Lesoinne schrieb am Mi. 1. Feb. 2017 um 19:07: > I have compiled OpenMPI 2.0.2 on a new Macbook running OS X 10.12 and have > been trying to run simple program. > I

Re: [OMPI users] OpenMPI not running any job on Mac OS X 10.12

2017-02-02 Thread Howard Pritchard
Hi Michael, I reproduced this problem on my Mac too: pn1249323:~/ompi/examples (v2.0.x *)$ mpirun -np 2 ./ring_c [pn1249323.lanl.gov:94283] mca_base_component_repository_open: unable to open mca_patcher_overwrite: File not found (ignored) [pn1249323.lanl.gov:94283]

Re: [OMPI users] OpenMPI not running any job on Mac OS X 10.12

2017-02-02 Thread Howard Pritchard
Hi Michel It's somewhat unusual to use the disable-shared configure option. That may be causing this. Could you try to build without using this option and see if you still see the problem? Thanks, Howard Michel Lesoinne schrieb am Mi. 1. Feb. 2017 um 21:07: > I

Re: [OMPI users] Error using hpcc benchmark

2017-01-31 Thread Howard Pritchard
Hi Wodel Randomaccess part of HPCC is probably causing this. Perhaps set PSM env. variable - Export PSM_MQ_REVCREQ_MAX=1000 or something like that. Alternatively launch the job using mpirun --mca plm ob1 --host to avoid use of psm. Performance will probably suffer with this option

Re: [OMPI users] Open MPI over RoCE using breakout cable and switch

2017-01-24 Thread Howard Pritchard
* users [mailto:users-boun...@lists.open-mpi.org] *On Behalf Of *Howard > Pritchard > *Sent:* Friday, January 20, 2017 6:35 PM > *To:* Open MPI Users <users@lists.open-mpi.org> > *Subject:* Re: [OMPI users] Open MPI over RoCE using breakout cable and > switch > > > > Hi Bre

Re: [OMPI users] Open MPI over RoCE using breakout cable and switch

2017-01-20 Thread Howard Pritchard
Hi Brendan I doubt this kind of config has gotten any testing with OMPI. Could you rerun with --mca btl_base_verbose 100 added to the command line and post the output to the list? Howard Brendan Myers schrieb am Fr. 20. Jan. 2017 um 15:04: > Hello, > > I am

Re: [OMPI users] still segmentation fault with openmpi-2.0.2rc3 on Linux

2017-01-09 Thread Howard Pritchard
> [loki:13586] *** reported by process [2873294849,0] > [loki:13586] *** on communicator MPI_COMM_WORLD > [loki:13586] *** MPI_ERR_UNKNOWN: unknown error > [loki:13586] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will > now abort, > [loki:13586] ***and potenti

Re: [OMPI users] still segmentation fault with openmpi-2.0.2rc3 on Linux

2017-01-08 Thread Howard Pritchard
HI Siegmar, Could you post the configury options you use when building the 2.0.2rc3? Maybe that will help in trying to reproduce the segfault you are observing. Howard 2017-01-07 2:30 GMT-07:00 Siegmar Gross < siegmar.gr...@informatik.hs-fulda.de>: > Hi, > > I have installed openmpi-2.0.2rc3

Re: [OMPI users] segmentation fault with openmpi-2.0.2rc2 on Linux

2017-01-03 Thread Howard Pritchard
2] mca: base: close: component self closed > [loki:05572] mca: base: close: unloading component self > [loki:05572] mca: base: close: component tcp closed > [loki:05572] mca: base: close: unloading component tcp > loki spawn 125 > > > Kind regards and thank you very muc

Re: [OMPI users] segmentation fault with openmpi-2.0.2rc2 on Linux

2017-01-02 Thread Howard Pritchard
HI Siegmar, I've attempted to reproduce this using gnu compilers and the version of this test program(s) you posted earlier in 2016 but am unable to reproduce the problem. Could you double check that the slave program can be successfully run when launched directly by mpirun/mpiexec? It might

Re: [OMPI users] Segmentation Fault (Core Dumped) on mpif90 -v

2016-12-23 Thread Howard Pritchard
Hi Paul, Thanks very much Christmas present. The Open MPI README has been updated to include a note about issues with the Intel 16.0.3-4 compiler suites. Enjoy the holidays, Howard 2016-12-23 3:41 GMT-07:00 Paul Kapinos : > Hi all, > > we discussed this issue

Re: [OMPI users] device failed to appear .. Connection timed out

2016-12-08 Thread Howard Pritchard
;Anyway, * /dev/hfi1_0* doesn't exist. >> >> Make sure you have the hfi1 module/driver loaded. >> >> In addition, please confirm the links are in active state on all the >> nodes `opainfo` >> >> >> >> _MAC >> >> >> >&

Re: [OMPI users] device failed to appear .. Connection timed out

2016-12-08 Thread Howard Pritchard
hello Daniele, Could you post the output from ompi_info command? I'm noticing on the RPMS that came with the rhel7.2 distro on one of our systems that it was built to support psm2/hfi-1. Two things, could you try running applications with mpirun --mca pml ob1 (all the rest of your args) and

Re: [OMPI users] Follow-up to Open MPI SC'16 BOF

2016-11-22 Thread Howard Pritchard
Hi Jeff, I don't think it was the use of memkind itself, but a need to refactor the way Open MPI is using info objects that was the issue. I don't recall the details. Howard 2016-11-22 16:27 GMT-07:00 Jeff Hammond : > >> >>1. MPI_ALLOC_MEM integration with memkind

[OMPI users] Follow-up to Open MPI SC'16 BOF

2016-11-22 Thread Howard Pritchard
on whether to go with a v2.2.x release next year or to go from v2.1.x to v3.x in late 2017 or early 2018 at the link below: https://www.open-mpi.org/sc16/ Thanks very much, Howard -- Howard Pritchard HPC-DES Los Alamos National Laboratory ___ users mailing

Re: [OMPI users] ScaLapack tester fails with 2.0.1, works with 1.10.4; Intel Omni-Path

2016-11-18 Thread Howard Pritchard
Hi Christof, Thanks for trying out 2.0.1. Sorry that you're hitting problems. Could you try to run the tests using the 'ob1' PML in order to bypass PSM2? mpirun --mca pml ob1 (all the rest of the args) and see if you still observe the failures? Howard 2016-11-18 9:32 GMT-07:00 Christof

Re: [OMPI users] How to verify RDMA traffic (RoCE) is being sent over a fabric when running OpenMPI

2016-11-08 Thread Howard Pritchard
HI Brenda, I should clarify as my response may confuse folks. We had configured the connectx4 cards to use ethernet/RoCE rather than IB transport for these measurements. Howard 2016-11-08 16:08 GMT-07:00 Howard Pritchard <hpprit...@gmail.com>: > Hi Brenda, > > What type of

Re: [OMPI users] How to verify RDMA traffic (RoCE) is being sent over a fabric when running OpenMPI

2016-11-08 Thread Howard Pritchard
Hi Brenda, What type of ethernet device (is this a Mellanox HCA?) and ethernet switch are you using? The mpirun configure options look correct to me. Is it possible that you have all the mpi processes on a single node? It should be pretty obvious from the SendRecv IMB test if you're using RoCE.

Re: [OMPI users] how to tell if pmi or pmi2 is being used?

2016-10-13 Thread Howard Pritchard
HI David, If you are using srun, you can export OMPI_MCA_pmix_base_verbose=10 and there will be output to show which SLURM pmi library you are using. Howard 2016-10-13 12:55 GMT-06:00 David Shrader : > That is really good to know. Thanks! > David > > > On 10/13/2016 12:27

Re: [OMPI users] Regression: multiple memory regions in dynamic windows

2016-08-25 Thread Howard Pritchard
Hi Joseph, Thanks for reporting this problem. There's an issue now (#2012) https://github.com/open-mpi/ompi/issues/2012 to track this. Howard 2016-08-25 7:44 GMT-06:00 Christoph Niethammer : > Hello, > > The Error is not 100% reproducible for me every time but seems to >

Re: [OMPI users] Java-OpenMPI returns with SIGSEGV

2016-07-08 Thread Howard Pritchard
Hi Gundram Could you configure without the disable dlopen option and retry? Howard Am Freitag, 8. Juli 2016 schrieb Gilles Gouaillardet : > the JVM sets its own signal handlers, and it is important openmpi dones > not override them. > this is what previously happened with PSM (infinipath) but

Re: [OMPI users] problem with exceptions in Java interface

2016-05-24 Thread Howard Pritchard
Hi Siegmar, Sorry for the delay, I seem to have missed this one. It looks like there's an error in the way the native methods are processing java exceptions. The code correctly builds up an exception message for cases where MPI 'c' returns non-success but, not if the problem occured in one of

Re: [OMPI users] mpirun java

2016-05-23 Thread Howard Pritchard
601ae65dc34a347c* > > * Env[3]: OMPI_MCA_orte_peer_modex_id=0* > > * Env[4]: OMPI_MCA_orte_peer_init_barrier_id=1* > > * Env[5]: OMPI_MCA_orte_peer_fini_barrier_id=2* > > * Env[6]: TMPDIR=/var/folders/5t/6tqp003x4fn09fzgtx46tjdhgn/T/* > > * Env[7]: __CF_USER_TEXT_ENCODING=0

Re: [OMPI users] mpirun java

2016-05-23 Thread Howard Pritchard
Hello Claudio, mpirun should be combining your java.library.path option with the one needed to add the Open MPI's java bindings as well. Which version of Open MPI are you using? Could you first try to compile the Ring.java code in ompi/examples and run it with the following additional mpirun

Re: [OMPI users] libfabric verb provider for iWARP RNIC

2016-04-04 Thread Howard Pritchard
Hi Durga, I'd suggest reposting this to the libfabric-users mail list. You can join that list at http://lists.openfabrics.org/mailman/listinfo/libfabric-users I'd suggest including the output of config.log. If you installed ofed in non-canonical location, you may need to give an explicit path

Re: [OMPI users] Java MPI Code for NAS Benchmarks

2016-03-11 Thread Howard Pritchard
Hello Saliya, Sorry i did not see this email earlier. There are a bunch of java test codes including performance tests like used in the paper at https://github.com/open-mpi/ompi-java-test Howard 2016-02-27 23:01 GMT-07:00 Saliya Ekanayake : > Hi, > > I see this paper from

Re: [OMPI users] Issues Building Open MPI static with Intel Fortran 16

2016-01-22 Thread Howard Pritchard
HI Matt, If you don't need oshmem, you could try again with --disable-oshmem added to the config line Howard 2016-01-22 12:15 GMT-07:00 Matt Thompson : > All, > > I'm trying to duplicate an issue I had with ESMF long ago (not sure if I > reported it here or at ESMF,

Re: [OMPI users] How to allocate more memory to java OpenMPI

2016-01-19 Thread Howard Pritchard
HI Ibrahim, Are you using a 32bit or 64bit JVM? I don't think this is an Open MPI issue, but likely something owing to your app or your java setup. You may want to checkout http://javaeesupportpatterns.blogspot.com/2012/09/outofmemoryerror-unable-to-create-new.html If you'd like to post the

Re: [OMPI users] problem with execstack and openmpi-v1.10.1-140-g31ff573

2016-01-14 Thread Howard Pritchard
HI Sigmar, Would you mind posting your MsgSendRecvMain to the mail list? I'd like to see if I can reproduce it on my linux box. Thanks, Howard 2016-01-14 7:30 GMT-07:00 Siegmar Gross < siegmar.gr...@informatik.hs-fulda.de>: > Hi, > > I've successfully built openmpi-v1.10.1-140-g31ff573 on

Re: [OMPI users] RMA operations with java buffers

2016-01-13 Thread Howard Pritchard
Hi Marko, You can probably find examples of what you'd like to do on github: https://github.com/open-mpi/ompi-java-test There are numerous MPI-2 RMA examples in the one-sided subdirectory. If you've never used github before, jus click on the download as zip button in the upper right hand

Re: [OMPI users] help understand unhelpful ORTE error message

2015-11-19 Thread Howard Pritchard
either on edison or cori. Howard 2015-11-19 17:11 GMT-07:00 Howard Pritchard <hpprit...@gmail.com>: > Hi Jeff H. > > Why don't you just try configuring with > > ./configure --prefix=my_favorite_install_dir > --with-libfabric=install_dir_for_libfabric > make -j 8 insta

Re: [OMPI users] mpijavac doesn't compile any thing

2015-11-19 Thread Howard Pritchard
Hi Ibrahim, If you just try to compile with the javac do you at least see a "error: package mpi..." does not exist? Adding the "-verbose" option may also help with diagnosing the problem. If the javac doesn't get that far then your problem is with the java install. Howard 2015-11-19 6:45

Re: [OMPI users] mpijavac doesn't compile any thing

2015-11-18 Thread Howard Pritchard
Hello Ibrahim As a sanity check, could you try to compile the Hello.java in examples? mpijavac --verbose Hello.java you should see something like: /usr/bin/javac -cp /global/homes/h/hpp/ompi_install/lib/mpi.jar:/global/homes/h/hpp/ompi_install/lib/shmem.jar Hello.java You may also want to

Re: [OMPI users] libfabric/usnic does not compile in 2.x

2015-09-30 Thread Howard Pritchard
the answer is yes, although the default is for Open MPI to use mtl/psm on that network. > > Please forgive my ignorance, the amount of different options is rather > overwhelming.. > > Marcin > > > > On 09/30/2015 04:26 PM, Howard Pritchard wrote: > > Hello Marcin > &

Re: [OMPI users] libfabric/usnic does not compile in 2.x

2015-09-30 Thread Howard Pritchard
Hello Marcin What configure options are you using besides with-libfabric? Could you post your config.log file tp the list? Looks like you only install fi_ext_usnic.h if you could build the usnic libfab provider. When you configured libfabric what providers were listed at the end of configure

Re: [OMPI users] segfault on java binding from MPI.init()

2015-08-15 Thread Howard Pritchard
> > Cheers, > > Gilles > > On Saturday, August 15, 2015, Howard Pritchard <hpprit...@gmail.com> > wrote: > >> Hi Jeff, >> >> I don't know why Gilles keeps picking on the persistent request problem >> and mixing >> it up with this user bug. I

Re: [OMPI users] segfault on java binding from MPI.init()

2015-08-14 Thread Howard Pritchard
, but I'd be greatly surprised if > he had InfiniPath on his systems where he ran into this segv issue...? > > > > On Aug 14, 2015, at 1:08 PM, Howard Pritchard <hpprit...@gmail.com> > wrote: > > > > Hi Gilles, > > > > Good catch! Nate we hadn't been testing o

Re: [OMPI users] segfault on java binding from MPI.init()

2015-08-14 Thread Howard Pritchard
hare/info --with-bugurl= >>> <http://bugzilla.redhat.com/bugzilla>http://bugzilla.redhat.com/bugzilla >>> --enable-bootstrap --enable-shared --enable-threads=posix >>> --enable-checking=release --with-system-zlib --enable-__cxa_atexit >>> --di

Re: [OMPI users] segfault on java binding from MPI.init()

2015-08-13 Thread Howard Pritchard
ext file of what I see when I run it with mpirun and your > debug flag. Lots of debug lines.* > > > Nate > > > > > > On Wed, Aug 12, 2015 at 11:09 AM, Howard Pritchard <hpprit...@gmail.com> > wrote: > >> Hi Nate, >> >> Sorry for the delay in g

Re: [OMPI users] segfault on java binding from MPI.init()

2015-08-12 Thread Howard Pritchard
-1.5.0.0/jre > --enable-libgcj-multifile --enable-java-maintainer-mode > --with-ecj-jar=/usr/share/java/eclipse-ecj.jar --disable-libjava-multilib > --with-ppl --with-cloog --with-tune=generic --with-arch_32=i686 > --build=x86_64-redhat-linux > Thread model: posix > gcc version 4

Re: [OMPI users] segfault on java binding from MPI.init()

2015-08-05 Thread Howard Pritchard
can uncompress to have an input > directory. The text file is just the same line over and over again. Run it > as: > > *java MPITestBroke tweets/* > > > Nate > > > > > > On Wed, Aug 5, 2015 at 8:29 AM, Howard Pritchard <hpprit...@gmail.com> > wrote: >

Re: [OMPI users] segfault on java binding from MPI.init()

2015-08-05 Thread Howard Pritchard
t; > On Tue, Aug 4, 2015 at 12:26 PM, Howard Pritchard <hpprit...@gmail.com> > wrote: > >> Hello Nate, >> >> As a sanity check of your installation, could you try to compile the >> examples/*.java codes using the mpijavac you've installed and see that >>

Re: [OMPI users] segfault on java binding from MPI.init()

2015-08-04 Thread Howard Pritchard
n calling MPI.init(). > > I'm not great at debugging this sort of stuff, but happy to try things out > if you need me to. > > Nate > > > On Tue, Aug 4, 2015 at 5:09 AM, Howard Pritchard <hpprit...@gmail.com> > wrote: > >> Hello Nate, >> >> As a firs

Re: [OMPI users] segfault on java binding from MPI.init()

2015-08-04 Thread Howard Pritchard
Hello Nate, As a first step to addressing this, could you please try using gcc rather than the Intel compilers to build Open MPI? We've been doing a lot of work recently on the java bindings, etc. but have never tried using any compilers other than gcc when working with the java bindings.

Re: [OMPI users] Running with native ugni on a Cray XC

2015-06-30 Thread Howard Pritchard
; Software Engineer > Cray, Inc. > ------ > *From:* users [users-boun...@open-mpi.org] on behalf of Howard Pritchard [ > hpprit...@gmail.com] > *Sent:* Thursday, June 25, 2015 11:00 PM > *To:* Open MPI Users > *Subject:* Re: [OMPI users] Running with native

Re: [OMPI users] Running with native ugni on a Cray XC

2015-06-26 Thread Howard Pritchard
[users-boun...@open-mpi.org] on behalf of Howard Pritchard [ > hpprit...@gmail.com] > *Sent:* Thursday, June 25, 2015 5:11 PM > *To:* Open MPI Users > *Subject:* Re: [OMPI users] Running with native ugni on a Cray XC > > Hi Nick > > use master not 1.8.x. for cray xc. also for

Re: [OMPI users] Running with native ugni on a Cray XC

2015-06-25 Thread Howard Pritchard
Hi Nick use master not 1.8.x. for cray xc. also for config do not pay attention to cray/lanl platform files. just do config. also if using nativized slurm launch with srun not mpirun. howard -- sent from my smart phonr so no good type. Howard On Jun 25, 2015 2:56 PM, "Nick

Re: [OMPI users] hybrid programming and OpenMPI compilation

2015-06-25 Thread Howard Pritchard
Hello Fedele, Would it be possible to build the open mpi package to use gfortran and run the test again? Do you observe this problem if you build a Open MP (<-> MP not MPI) only version of the test case? I can't reproduce this problem using gfortran. I don't have access to an Intel compiler at

  1   2   >