Re: [OMPI devel] Question about new component submission

2025-01-23 Thread Howard Pritchard
example NCCL/RCCL. Is is also easier to optimize the communication > protocols regarding BXI hardware. We also can benefit shared memory under > some conditions, which seems to be harder using the portals4 mtl. > Ah that makes sense, esp. the shared memory part. > Regards, &g

Re: [OMPI devel] Question about new component submission

2025-01-22 Thread Howard Pritchard
Hello Florent, Would you be opening a PR for both PML and OSC components? The reason I ask is if you don’t have the OSC component ready to include in a PR how will you be supporting one sided ops? Would your PRs touch code outside of the new components? Also it would be nice to add some documen

Re: [OMPI devel] MPI ABI effort

2023-08-29 Thread Howard Pritchard via devel
LANL would be interested in supporting this feature as well. Howard On Mon, Aug 28, 2023 at 9:58 AM Jeff Squyres (jsquyres) via devel < devel@lists.open-mpi.org> wrote: > We got a presentation from the ABI WG (proxied via Quincey from AWS) a few > months ago. > > The proposal looked reasonable.

[OMPI devel] Open MPI 4.0.6rc4

2021-06-04 Thread Howard Pritchard via devel
Hi All, Open MPI v4.0.6rc4 (we messed up and had to skip rc3) is now available at https://www.open-mpi.org/software/ompi/v4.0/ Changes since the 4.0.5 release include: - Update embedded PMIx to 3.2.3. This update addresses several MPI_COMM_SPAWN problems. - Fix an issue with MPI_FILE_GET_B

[OMPI devel] Open MPI 4.0.4rc3 available for testing

2020-06-09 Thread Howard Pritchard via devel
Open MPI v4.0.4rc3 has been posted to https://www.open-mpi.org/software/ompi/v4.0/ This rc includes a fix for a problem discovered with the memory patcher code. As described in the README: - Open MPI v4.0.4 fixed an issue with the memory patcher's ability to intercept shmat and shmdt that coul

[OMPI devel] Please test Open MPI v4.0.4rc1

2020-05-09 Thread Howard Pritchard via devel
Open MPI v4.0.4rc1 has been posted to https://www.open-mpi.org/software/ompi/v4.0/ 4.0.4 -- May, 2020 --- - Fix an ABI compatibility issue with the Fortran 2008 bindings. Thanks to Alastair McKinstry for reporting. - Fix an issue with rpath of /usr/lib64 when building OMP

[OMPI devel] Open MPI 4.0.1rc3 available for testing

2019-03-22 Thread Howard Pritchard
A third release candidate for the Open MPI v4.0.1 release is posted at https://www.open-mpi.org/software/ompi/v4.0/ Fixes since 4.0.1rc2 include - Add acquire semantics to an Open MPI internal lock acquire function. Our goal is to release 4.0.1 by the end of March, so any testing is appreciated.

[OMPI devel] Open MPI 4.0.1rc2 available for testing

2019-03-19 Thread Howard Pritchard
A second release candidate for the Open MPI v4.0.1 release is posted at https://www.open-mpi.org/software/ompi/v4.0/ Fixes since 4.0.1rc1 include - Fix an issue with Vader (shared-memory) transport on OS-X. Thanks to Daniel Vollmer for reporting. - Fix a problem with the usNIC BTL Makefile. Th

[OMPI devel] Open MPI 4.0.1rc1 available for testing

2019-03-01 Thread Howard Pritchard
The first release candidate for the Open MPI v4.0.1 release is posted at https://www.open-mpi.org/software/ompi/v4.0/ Major changes include: - Update embedded PMIx to 3.1.2. - Fix an issue when using --enable-visibility configure option and older versions of hwloc. Thanks to Ben Menadue fo

Re: [OMPI devel] Entry in mca-btl-openib-device-params.ini

2018-10-15 Thread Howard Pritchard
Hello Sindhu, Open a github PR with your changes. See https://github.com/open-mpi/ompi/wiki/SubmittingPullRequests Howard Am Mo., 15. Okt. 2018 um 13:26 Uhr schrieb Devale, Sindhu < sindhu.dev...@intel.com>: > Hi, > > > > I need to add an entry to the *mca-btl-openib-device-params.ini **file

[OMPI devel] testing again (EOM)

2018-09-01 Thread Howard Pritchard
___ devel mailing list devel@lists.open-mpi.org https://lists.open-mpi.org/mailman/listinfo/devel

[OMPI devel] Open MPI website borked up?

2018-09-01 Thread Howard Pritchard
Hi Folks, Something seems to be borked up about the OMPI website. Got to website and you'll get some odd parsing error appearing. Howard ___ devel mailing list devel@lists.open-mpi.org https://lists.open-mpi.org/mailman/listinfo/devel

[OMPI devel] testing if NMC mail server working again

2018-08-28 Thread Howard Pritchard
___ devel mailing list devel@lists.open-mpi.org https://lists.open-mpi.org/mailman/listinfo/devel

[OMPI devel] Open MPI 2.1.3rc3 available for testing

2018-03-14 Thread Howard Pritchard
HI Folks, A few MPI I/O (both in OMPI I/O and ROMIO glue layer) bugs were found in the rc2 so we're doing an rc3. Open MPI 2.1.3rc3 tarballs are available for testing at the usual place: https://www.open-mpi.org/software/ompi/v2.1/ This is a bug fix release for the Open MPI 2.1.x release strea

[OMPI devel] Open MPI 2.1.3 rc2 available for testing

2018-02-22 Thread Howard Pritchard
Hello Folks, We discovered a bug in the osc/rdma component that we wanted to fix in this release,hence an rc2. Open MPI 2.1.3rc2 tarballs are available for testing at the usual place: https://www.open-mpi.org/software/ompi/v2.1/ This is a bug fix release for the Open MPI 2.1.x release stream. I

[OMPI devel] Open MPI 2.1.3 rc1 available for testing

2018-02-15 Thread Howard Pritchard
Hello Folks, Open MPI 2.1.3rc1 tarballs are available for testing at the usual place: https://www.open-mpi.org/software/ompi/v2.1/ This is a bug fix release for the Open MPI 2.1.x release stream. Items fixed in this release include the following: - Update internal PMIx version to 1.2.5. - Fix a

Re: [OMPI devel] hwloc2 and cuda and non-default cudatoolkit install location

2017-12-22 Thread Howard Pritchard
. > > > On Dec 20, 2017, at 1:45 PM, Brice Goglin wrote: > > Le 20/12/2017 à 22:01, Howard Pritchard a écrit : > > I can think of several ways to fix it. Easiest would be to modify the > opal/mca/hwloc/hwloc2a/configure.m4 > to not set --enable-cuda if --with-cuda is eva

[OMPI devel] hwloc2 and cuda and non-default cudatoolkit install location

2017-12-20 Thread Howard Pritchard
Hi Folks, I've a question about where to fix a problem I'm having building Open MPI master and its embedded hwloc2a on a cluster we have that sports Nvidia GPUs. Here's the problem, I want for Open MPI to be smart about cuda, so I try to configure with: ./configure --prefix=/usr/projects/hpctool

[OMPI devel] 2.0.4rc3 is available for testing

2017-11-07 Thread Howard Pritchard
Hi Folks, We fixed one more thing for the 2.0.4 release, so there's another rc, now rc3. The fixed item was a problem with neighbor collectives. Thanks to Lisandro Dalcin for reporting. Tarballs are at the usual place, https://www.open-mpi.org/software/ompi/v2.0/ Thanks, Open MPI release team

[OMPI devel] Open MPI 2.0.4rc2 available for testing

2017-11-01 Thread Howard Pritchard
HI Folks, We decided to roll an rc2 to pick up a PMIx fix: - Fix an issue with visibility of functions defined in the built-in PMIx. Thanks to Siegmar Gross for reporting this issue. Tarballs can be found at the usual place https://www.open-mpi.org/software/ompi/v2.0/ Thanks, Your Open MPI

[OMPI devel] Open MPI 2.0.4rc1 available for testing

2017-10-29 Thread Howard Pritchard
HI Folks, Open MPI 2.0.4rc1 is available for download and testing at https://www.open-mpi.org/software/ompi/v2.0/ Fixes in this release include: 2.0.4 -- October, 2017 -- Bug fixes/minor improvements: - Add configure check to prevent trying to build this release of Open

[OMPI devel] MTT database

2017-10-12 Thread Howard Pritchard
is anyone seeing issues with MTT today? When I go to the website and click on summary I get this back in my browser window: MTTDatabase abort: Could not connect to the ompidb database; submit this run later. Howard ___ devel mailing list devel@lists.ope

[OMPI devel] Open MPI 2.1.2rc4 available for testing

2017-09-13 Thread Howard Pritchard
Hello Folks, Open MPI 2.1.2 rc4 is uploaded to the usual place: https://www.open-mpi.org/software/ompi/v2.1/ Issues addressed since the last release candidate: - Fix a configury problem with the embedded PMIx 1.2.3 package. - Add an option when using SLURM to launch processes on the head node

[OMPI devel] KNL/hwloc funny message question

2017-09-01 Thread Howard Pritchard
Hi Folks, I just now subscribed to the hwloc user mail list, but I suspect that requires human intervention to get on, and that might not mean something happening till next week. Alas google has failed me in helping to understand the message. So, I decided to post to Open MPI devel list and see

[OMPI devel] Open MPI 2.1.2rc3 available for testing

2017-08-30 Thread Howard Pritchard
Hi Folks, Open MPI 2.1.2rc3 tarballs are available for testing at the usual place: https://www.open-mpi.org/software/ompi/v2.1/ Fixes since rc2: Issue #4122: CMA compilation error in SM BTL.Thanks to Paul Hargrove for catching this. Issue #4034: NAG Fortran compiler -rpath configuration err

[OMPI devel] Open MPI 2.1.2rc2 available

2017-08-17 Thread Howard Pritchard
Hello Folks, Open MPI 2.1.2rc2 tarballs are available for testing: https://www.open-mpi.org/software/ompi/v2.1/ Fixes since rc1: Issue #4069 - PMIx visibility problem. Thanks to Siegmar Gross for reporting. Issue #2324 - Fix a problem with neighbor collectives. Thanks to Lisandro Dalcin for r

[OMPI devel] Open MPI v2.1.2rc1 available

2017-08-10 Thread Howard Pritchard
Hi Folks, Open MPI v2.1.2rc1 tarballs are available for testing at the usual place: https://www.open-mpi.org/software/ompi/v2.1/ There is an outstanding issue which will be fixed before the final release: https://github.com/open-mpi/ompi/issues/4069 but we wanted to get an rc1 out to see

[OMPI devel] hwloc 2 thing

2017-07-20 Thread Howard Pritchard
Hi Folks, I'm noticing that if I pull a recent version of master with hwloc 2 support into my local repo, that my autogen,pl run fails unless I do the following: mkdir $PWD/opal/mca/hwloc/hwloc2x/hwloc/include/private/autogen where PWD is the top level of my work area. I did a git clean -df

Re: [OMPI devel] Open MPI 3.0.0 first release candidate posted

2017-06-29 Thread Howard Pritchard
ub.com/open-mpi/ompi/pull/3794 > > On Jun 29, 2017, at 3:14 PM, Howard Pritchard wrote: > > Hi Brian, > > I tested this rc using both srun native launch and mpirun on the following > systems: > - LANL CTS-1 systems (haswell + Intel OPA/PSM2) > - LANL network testbed sys

Re: [OMPI devel] Open MPI 3.0.0 first release candidate posted

2017-06-29 Thread Howard Pritchard
Hi Brian, I tested this rc using both srun native launch and mpirun on the following systems: - LANL CTS-1 systems (haswell + Intel OPA/PSM2) - LANL network testbed system (haswell + connectX5/UCX and OB1) - LANL Cray XC I am finding some problems with mpirun on the network testbed system. For

[OMPI devel] libtool guru help needed (Fortran problem)

2017-06-22 Thread Howard Pritchard
Hi Folks, I'm trying to do some experiments with clang/llvm and its openmp runtime. To add to this mix, the application I'm wanting to use for testing is written in F08, so I'm having to also use flang: https://github.com/flang-compiler/flang Now when I try to build Open MPI, as long as I disabl

Re: [OMPI devel] orte-clean not cleaning left over temporary I/O files in /tmp

2017-06-22 Thread Howard Pritchard
en a pull request? > > Best > Christoph > > - Original Message - > From: "Howard Pritchard" > To: "Open MPI Developers" > Sent: Wednesday, June 21, 2017 5:57:05 PM > Subject: Re: [OMPI devel] orte-clean not cleaning left over temporary I/O > files i

Re: [OMPI devel] orte-clean not cleaning left over temporary I/O files in /tmp

2017-06-21 Thread Howard Pritchard
Hi Chris, Sorry for being a bit picky, but could you add a sign-off to the commit message? I'm not suppose to manually add it for you. Thanks, Howard 2017-06-21 9:45 GMT-06:00 Howard Pritchard : > Hi Chris, > > Thanks very much for the patch! > > Howard > > &

Re: [OMPI devel] orte-clean not cleaning left over temporary I/O files in /tmp

2017-06-21 Thread Howard Pritchard
Hi Chris, Thanks very much for the patch! Howard 2017-06-21 9:43 GMT-06:00 Christoph Niethammer : > Hello Ralph, > > Thanks for the update on this issue. > > I used the latest master (c38866eb3929339147259a3a46c6fc815720afdb). > > The behaviour is still the same: aborting before MPI_File_close

Re: [OMPI devel] SLURM 17.02 support

2017-06-19 Thread Howard Pritchard
quyres (jsquyres) > wrote: > > > > +1 on the error message. > > > > > > > >> On Jun 16, 2017, at 10:06 AM, Howard Pritchard > wrote: > >> > >> Hi Ralph > >> > >> I think a helpful error message would suffice. > &g

Re: [OMPI devel] SLURM 17.02 support

2017-06-16 Thread Howard Pritchard
Hi Ralph I think a helpful error message would suffice. Howard r...@open-mpi.org schrieb am Di. 13. Juni 2017 um 11:15: > Hey folks > > Brian brought this up today on the call, so I spent a little time > investigating. After installing SLURM 17.02 (with just --prefix as config > args), I conf

Re: [OMPI devel] Time to remove Travis?

2017-06-01 Thread Howard Pritchard
I vote for removal too. Howard r...@open-mpi.org schrieb am Do. 1. Juni 2017 um 08:10: > I’d vote to remove it - it’s too unreliable anyway > > > On Jun 1, 2017, at 6:30 AM, Jeff Squyres (jsquyres) > wrote: > > > > Is it time to remove Travis? > > > > I believe that the Open MPI PRB now covers

[OMPI devel] Open MPI v2.0.3rc1 available for testing

2017-05-26 Thread Howard Pritchard
Hi Folks, Open MPI v2.0.3rc1 tarballs are available on the download site for testing: https://www.open-mpi.org/software/ompi/v2.0/ Fixes in this bug fix release include: - Fix a problem with MPI_IALLTOALLW when zero size messages are present. Thanks to @mathbird for reporting. - Add missing

[OMPI devel] Open MPI 2.1.1rc1 is up

2017-04-27 Thread Howard Pritchard
Hi Open MPI developers, Open MPI 2.1.1rc1 is available for testing at the usual place: https://www.open-mpi.org/software/ompi/v2.1/ Bug fixes in this release: - Add missing MPI_AINT_ADD/MPI_AINT_DIFF function definitions to mpif.h. Thanks to Aboorva Devarajan for reporting. - Fix the error r

[OMPI devel] Open MPI v2.1.1 release reminder - public service announcement

2017-04-19 Thread Howard Pritchard
HI Folks, Reminder that we are planning to do a v2.1.1 bug release next Tuesday (4/25/17) as discussed in yesterday's con-call. If you have bug fixes you'd like to get in to v2.1.1 please open PRs this week so there will be time for review and testing in MTT. Thanks, Howard and Jeff ___

[OMPI devel] OS-X specific jenkins/PR retest

2017-04-07 Thread Howard Pritchard
Hi Folks, I added an OS-X specific bot retest command for jenkins CI: bot:osx:retest Also added a blurb to the related wiki page: https://github.com/open-mpi/ompi/wiki/PRJenkins Hope this helps folks who encounter os-x specific problems with their PRs. Howard _

Re: [OMPI devel] Pull request: LANL-XXX tests failing

2017-03-30 Thread Howard Pritchard
well not sure what's going on. there was an upgrade of jenkins a bunch of functionality seems to have gotten lost. 2017-03-30 9:37 GMT-06:00 Howard Pritchard : > Actually it looks like we're running out of disk space at AWS. > > > 2017-03-30 9:28 GMT-06:00 r...@open-mpi

Re: [OMPI devel] Pull request: LANL-XXX tests failing

2017-03-30 Thread Howard Pritchard
Actually it looks like we're running out of disk space at AWS. 2017-03-30 9:28 GMT-06:00 r...@open-mpi.org : > You didn’t do anything wrong - the Jenkins test server at LANL is having a > problem. > > On Mar 30, 2017, at 8:22 AM, DERBEY, NADIA wrote: > > Hi, > > I just created a pull request an

Re: [OMPI devel] Segfault during a free in reduce_scatter using basic component

2017-03-28 Thread Howard Pritchard
Hello Emmanuel, Which version of Open MPI are you using? Howard 2017-03-28 3:38 GMT-06:00 BRELLE, EMMANUEL : > Hi, > > We are working on a portals4 components and we have found a bug (causing > a segmentation fault ) which must be related to the coll/basic component. > Due to a lack of time

Re: [OMPI devel] [2.1.0rc2] stupid run failure on Mac OS X Sierra

2017-03-07 Thread Howard Pritchard
Hi Paul There is an entry 8 under OS-X FAQ which describes this problem. Adding max allowable len is a good idea. Howard Paul Hargrove schrieb am Di. 7. März 2017 um 08:04: > The following is fairly annoying (though I understand the problem is real): > > $ [full-path-to]/mpirun -mca btl sm,se

Re: [OMPI devel] No Preset Parameters found

2017-02-20 Thread Howard Pritchard
Hello Amit which version of Open MPI are you using? Howard -- sent from my smart phonr so no good type. Howard On Feb 20, 2017 12:09 PM, "Kumar, Amit" wrote: > Dear OpenMPI, > > > > Wondering what preset parameters are this warning is indicating? > > > > Thank you, > > Amit > > > >

Re: [OMPI devel] [2.0.2rc4] "make install" failure on NetBSD/i386 (libtool?)

2017-01-28 Thread Howard Pritchard
HI Paul, This might be a result of building the tarball on a new system. Would you mind trying the rc3 tarball and see if that builds on the system? Howard 2017-01-27 15:12 GMT-07:00 Paul Hargrove : > I had no problem with 2.0.2rc3 on NetBSD, but with 2.0.2rc4 I am seeing a > "make install"

Re: [OMPI devel] [OMPI users] still segmentation fault with openmpi-2.0.2rc3 on Linux

2017-01-13 Thread Howard Pritchard
t; Slave process 3 of 4 running on loki > spawn_slave 3: argv[0]: spawn_slave > spawn_slave 0: argv[0]: spawn_slave > spawn_slave 1: argv[0]: spawn_slave > Parent process 0: tasks in MPI_COMM_WORLD: 1 > tasks in COMM_CHILD_PROCESSES local group: 1

Re: [OMPI devel] Fwd: Re: [OMPI users] still segmentation fault with openmpi-2.0.2rc3 on Linux

2017-01-12 Thread Howard Pritchard
r here) >>> >>> I will keep digging tomorrow. >>> Note that if you specify an incorrect slot list, MPI_Comm_spawn fails >>> with a very unfriendly error message. >>> Right now, the 4th spawn'ed task crashes, so this is a different issue >>>

Re: [OMPI devel] [2.0.2rc3] build failure ppc64/-m32 and bultin-atomics

2017-01-06 Thread Howard Pritchard
HI Paul, https://github.com/open-mpi/ompi/issues/2677 It seems we have a bunch of problems with PPC64 atomics and I'd like to see if we can get at least some of these issues resolved for 2.0.2, so I've set this as a blocker along with 2610. Howard 2017-01-06 9:48 GMT-07:00 Howard

Re: [OMPI devel] [2.0.2rc3] build failure ppc64/-m32 and bultin-atomics

2017-01-06 Thread Howard Pritchard
Hi Paul, Sorry for the confusion. This is a different problem. I'll open an issue for this one too. Howard 2017-01-06 9:18 GMT-07:00 Howard Pritchard : > Hi Paul, > > Thanks for checking this. > > This problem was previously reporting and there's an issue: > &

Re: [OMPI devel] [2.0.2rc3] build failure ppc64/-m32 and bultin-atomics

2017-01-06 Thread Howard Pritchard
Hi Paul, Thanks for checking this. This problem was previously reporting and there's an issue: https://github.com/open-mpi/ompi/issues/2610 tracking it. Howard 2017-01-05 21:19 GMT-07:00 Paul Hargrove : > I have a standard Linux/ppc64 system with gcc-4.8.3 > I have configured the 2.0.2rc3 t

Re: [OMPI devel] [2.0.2rc2] opal_fifo hang w/ --enable-osx-builtin-atomics

2017-01-05 Thread Howard Pritchard
Hi Paul, I opened issue 2666 to track this. Howard 2017-01-05 0:23 GMT-07:00 Paul Hargrove : > On Macs running Yosemite (OS X 10.10 w/ Xcode 7.1) and El Capitan (OS X > 10.11 w/ Xcode 8.1) I have configured with > CC=cc CXX=c++ FC=/sw/bin/gfort

Re: [OMPI devel] rdmacm and udcm for 2.0.1 and RoCE

2017-01-05 Thread Howard Pritchard
Hi Dave, Sorry for the delayed response. Anyway, you have to use rdmacm for connection management when using ROCE. However, with 2.0.1 and later, you have to specify per peer QP info manually on the mpirun command line. Could you try rerunning with mpirun --mca btl_openib_receive_queues P,128,6

Re: [OMPI devel] [2.0.2rc2] FreeBSD-11 run failure

2017-01-05 Thread Howard Pritchard
HI Paul, I opened https://github.com/open-mpi/ompi/issues/2665 to track this. Thanks for reporting this. Howard 2017-01-04 14:43 GMT-07:00 Paul Hargrove : > With the 2.0.2rc2 tarball on FreeBSD-11 (i386 or amd64) I am configuring > with: > --prefix=... CC=clang CXX=clang++ --disable-m

[OMPI devel] Open MPI 2.0.2rc2 is up

2016-12-23 Thread Howard Pritchard
. Thanks, Howard -- Howard Pritchard HPC-DES Los Alamos National Laboratory ___ devel mailing list devel@lists.open-mpi.org https://rfd.newmexicoconsortium.org/mailman/listinfo/devel

Re: [OMPI devel] Open MPI v2.0.2rc1 is up

2016-12-20 Thread Howard Pritchard
HI Orion, Opened issue 2610 <https://github.com/open-mpi/ompi/issues/2610>. Thanks, Howard 2016-12-20 11:27 GMT-07:00 Howard Pritchard : > Hi Orion, > > Thanks for trying out the rc. Which compiler/version of compiler are you > using? > > Howard > > >

Re: [OMPI devel] Open MPI v2.0.2rc1 is up

2016-12-20 Thread Howard Pritchard
Hi Orion, Thanks for trying out the rc. Which compiler/version of compiler are you using? Howard 2016-12-20 10:50 GMT-07:00 Orion Poplawski : > On 12/14/2016 07:58 PM, Jeff Squyres (jsquyres) wrote: > > Please test! > > > > https://www.open-mpi.org/software/ompi/v2.0/ > > I appear to e ge

Re: [OMPI devel] Open MPI v2.0.2rc1 is up

2016-12-19 Thread Howard Pritchard
HI Paul, Would you mind resending the "runtime error w/ PGI usempif08 on OpenPOWER" email without the config.log attached? Thanks, Howard 2016-12-16 12:17 GMT-07:00 Howard Pritchard : > HI Paul, > > Thanks for the checking the rc out. And for noting the grammar &g

Re: [OMPI devel] Open MPI v2.0.2rc1 is up

2016-12-16 Thread Howard Pritchard
HI Paul, Thanks for the checking the rc out. And for noting the grammar mistake. Howard 2016-12-16 1:00 GMT-07:00 Paul Hargrove : > My testing is complete. > > The only problems not already known are related to PGI's recent "Community > Edition" compilers and have been reported in three separ

Re: [OMPI devel] heads up about OMPI/master

2016-12-01 Thread Howard Pritchard
41c6c >>>> >>>> Cheers, >>>> >>>> Gilles >>>> >>>> On Friday, December 2, 2016, r...@open-mpi.org wrote: >>>> >>>>> I should add, FWIW: I’m working with the HEAD of master right now, and >>>

Re: [OMPI devel] heads up about OMPI/master

2016-12-01 Thread Howard Pritchard
g wrote: >> >>> I should add, FWIW: I’m working with the HEAD of master right now, and >>> not seeing any problems. >>> >>> On Dec 1, 2016, at 2:10 PM, r...@open-mpi.org wrote: >>> >>> ?? I see a bunch of commits that were all collected in

[OMPI devel] heads up about OMPI/master

2016-12-01 Thread Howard Pritchard
Hi Folks, Just an FYI it looks like a bunch of commits may have been accidentally pushed to master sometime in the past day. You may not want to merge origin/master (if origin is how you reference https://github.com/open-mpi/omp) into your master or rebase off of it until we get some clarity on w

Re: [OMPI devel] sm BTL performace of the openmpi-2.0.0

2016-08-05 Thread Howard Pritchard
Hello Christoph The rdmacm messages while annoying are not causing the problem. If you specify tcp BTL does the BW drop disappear? Also could you post your configure options to the mail list? Thanks Howard Am Freitag, 5. August 2016 schrieb Christoph Niethammer : > Hello, > > We see the sam

[OMPI devel] tcp btl rendezvous performance question

2016-07-18 Thread Howard Pritchard
Hi Folks, I have a cluster with some 100 Gb ethernet cards installed. What we are noticing if we force Open MPI 1.10.3 to go through the TCP BTL (rather than yalla) is that the performance of osu_bw once the TCP BTL switches from eager to rendezvous (> 32KB) falls off a cliff, going from about 1

Re: [OMPI devel] 2.0.0rc4 Crash in MPI_File_write_all_end

2016-07-13 Thread Howard Pritchard
Hi Eric, Thanks very much for finding this problem. We decided in order to have a reasonably timely release, that we'd triage issues and turn around a new RC if something drastic appeared. We want to fix this issue (and it will be fixed), but we've decided to defer the fix for this issue to a 2

Re: [OMPI devel] [2.0.0rc4] non-critical faulres report

2016-07-12 Thread Howard Pritchard
Paul, Could you narrow down the versions of the PGCC where you get the ICE when using the -m32 option? Thanks, Howard 2016-07-06 15:29 GMT-06:00 Paul Hargrove : > The following are previously reported issues that I am *not* expecting to > be resolved in 2.0.0. > However, I am listing them her

Re: [OMPI devel] Issue with 2.0.0rc3, singleton init

2016-06-16 Thread Howard Pritchard
Hi Lisandro, Thanks for giving the rc3 a try. Could you post the output of ompi_info from your install to the list? Thanks, Howard 2016-06-16 7:55 GMT-06:00 Lisandro Dalcin : > ./configure --prefix=/home/devel/mpi/openmpi/2.0.0rc3 --enable-debug > --enable-mem-debug > > https://bitbucket.org

[OMPI devel] Open MPI v2.0.0rc3 now available

2016-06-15 Thread Howard Pritchard
priority callbacks are made during Open MPI's main progress loop. - Disable backtrace support by default in the PSM/PSM2 libraries to prevent unintentional conflicting behavior. Thanks, Howard -- Howard Pritchard HPC-DES Los Alamos National Laboratory

Re: [OMPI devel] Jenkins testing - what purpose are we striving to achieve?

2016-06-07 Thread Howard Pritchard
HI Ralph, We briefly discussed this some today. I would like to avoid the mini-MTT approach for PR checking. At the same time, one can also see why it might be useful from time to time to make changes to the script a given jenkins project runs on PRs. An idea we discussed was to have jenkins fol

[OMPI devel] NERSC down today so lanl-bot getting time off

2016-05-24 Thread Howard Pritchard
Hi Folks, NERSC is doing some major maintenance and both the edison and cori systems are offline. As a consequence, the lanl-bot can't run some of its checks on OMPI PRs today. So you can either ignore the lack of status update for PRs for today and go ahead and merge, or wait till tomorrow when

[OMPI devel] updating the users migration guide request

2016-05-16 Thread Howard Pritchard
Hi Folks, The last known blocker for 2.0.0 will hopefully be resolved this week, which means its time to be filling in the users' migration guide. If you have a feature that went in to the 2.0.x release stream that's important, please add a short description of the feature to the migration guid

Re: [OMPI devel] Open MPI v2.0.0rc2

2016-04-30 Thread Howard Pritchard
Hi Jeff, Let's just update the MPI_THREAD_MULTIPLE comment to say that enable-mpi-thread-multiple is still required as part of config. Howard 2016-04-29 22:20 GMT-06:00 Orion Poplawski : > On 04/28/2016 05:01 PM, Jeff Squyres (jsquyres) wrote: > >> At long last, here's the next v2.0.0 release c

Re: [OMPI devel] 2.0.0 is coming: what do we need to communicate to users?

2016-04-29 Thread Howard Pritchard
Hi Jeff, checkpoint/restart is not supported in this release. Does this release work with totalview? I recall we had some problems, and do not remember if they were resolved. We may also want to clarify if any PML/MTLs are experimental in this release. MPI_THREAD_MULTIPLE support. Howard 2

Re: [OMPI devel] PSM2 Intel folks question

2016-04-21 Thread Howard Pritchard
et epid > 1) case but I still see the hello_c running with stock OMPI 1.10.2. > > > Would you please try forcing different jobid and share the results? > > > > Thanks, > > > > _MAC > > > > > > *From:* devel [mailto:devel-boun...@open-

Re: [OMPI devel] Common symbol warnings in tarballs (was: make install warns about 'common symbols')

2016-04-21 Thread Howard Pritchard
N in all RC tarballs, but OFF > in the release tarballs. > > Personally, the only thing I feel strongly about is not producing > developer-oriented warnings for the end-user who uses the normal configure > options. > > -Paul > > On Wed, Apr 20, 2016 at 2:44 PM, Howard Pritc

Re: [OMPI devel] Common symbol warnings in tarballs (was: make install warns about 'common symbols')

2016-04-20 Thread Howard Pritchard
I also think this symbol checker should not be in the tarball. Howard 2016-04-20 13:08 GMT-06:00 Jeff Squyres (jsquyres) : > On Apr 20, 2016, at 2:08 PM, dpchoudh . wrote: > > > > Just to clarify, I was doing a build (after adding code to support a new > transport) from code pulled from git (a

Re: [OMPI devel] PSM2 Intel folks question

2016-04-20 Thread Howard Pritchard
t; > Thanks, > > > > _MAC > > > > *From:* devel [mailto:devel-boun...@open-mpi.org] *On Behalf Of *Howard > Pritchard > *Sent:* Tuesday, April 19, 2016 12:21 PM > *To:* Open MPI Developers > *Subject:* Re: [OMPI devel] PSM2 Intel folks question > > >

Re: [OMPI devel] PSM2 Intel folks question

2016-04-19 Thread Howard Pritchard
> > > Thanks, > > _MAC > > > > *From:* devel [mailto:devel-boun...@open-mpi.org] *On Behalf Of *Howard > Pritchard > *Sent:* Tuesday, April 19, 2016 11:04 AM > *To:* Open MPI Developers List > *Subject:* [OMPI devel] PSM2 Intel folks question > > >

[OMPI devel] PSM2 Intel folks question

2016-04-19 Thread Howard Pritchard
Hi Folks, I'm making progress with issue #1559 (patches on the mail list didn't help), and I'll open a PR to help the PSM2 MTL work on a single node, but I'm noticing something more troublesome. If I run on just one node, and I use more than one process, process zero consistently hangs in psm2_ep

Re: [OMPI devel] psm2 and psm2_ep_open problems

2016-04-18 Thread Howard Pritchard
5, 2016, at 7:32 AM, Ralph Castain wrote: > > Actually, it did come across the developer list :-) > > Why don’t I resolve this by just ensuring that the key we create is > properly filled? It’s a trivial fix in the PMI ess component > > > On Apr 15, 2016, at 7:26

[OMPI devel] Fwd: psm2 and psm2_ep_open problems

2016-04-15 Thread Howard Pritchard
I didn't copy dev on this. -- Weitergeleitete Nachricht -- Von: *Howard Pritchard* Datum: Donnerstag, 14. April 2016 Betreff: psm2 and psm2_ep_open problems An: Open MPI Developers Hi Matias Actually I triaged this further. Open mpi PMI subsystem is actually doing t

Re: [OMPI devel] psm2 and psm2_ep_open problems

2016-04-14 Thread Howard Pritchard
e passed key is 0, epid will be 0. > > > > SOLUTION: set > > Set in the environment OMPI_MCA_orte_precondition_transports with a value > different than 0. > > > > Thanks, > > > > _MAC > > > > *From:* devel [mailto:devel-boun...@open-mpi.org &g

[OMPI devel] psm2 and psm2_ep_open problems

2016-04-14 Thread Howard Pritchard
Hi Folks, So we have this brand-new omnipath cluster here at work, but people are having problem using it on a single node using srun as the job launcher. The customer wants to use srun to launch jobs not the open mpi mpirun. The customer installed 1.10.1, but I can reproduce the problem with v2

Re: [OMPI devel] RFC: RML change to multi-select

2016-03-17 Thread Howard Pritchard
Okay, i'll bring this up at the workshop. There's been talking but no one's working on it. 2016-03-17 8:20 GMT-06:00 Ralph Castain : > We are also targeting RDM for now, but I agree that the two may diverge at > some point, and so flexibility makes sense. Only wish that libfabric had a > decent

[OMPI devel] trying to view ciso-community master results and mtt has issue

2016-02-18 Thread Howard Pritchard
Hi Folks, I noticed cisco-community MTT results are really red/pink today. If I try to view some of the ibm test results though, something goes south with mtt and this is what get's posted back to my browser: Fatal error: Allowed memory size of 67108864 bytes exhausted (tried to allocate 71 bytes

Re: [OMPI devel] Trunk is broken

2016-02-17 Thread Howard Pritchard
Hi Folks, Should we revert PR 1351 till there is a fix? Howard 2016-02-17 11:34 GMT-07:00 Ralph Castain : > FWIW: I wouldn’t have seen that because I don’t have IB on my system. > > On Feb 17, 2016, at 10:11 AM, Nysal Jan K A wrote: > > So this seems to be still broken. > > mca_btl_openib.so:

[OMPI devel] MTT error?

2016-02-11 Thread Howard Pritchard
Hi Folks When I go to https://mtt.open-mpi.org/ and then click the summary button I get some kind of DNS lookup error. Howard

Re: [OMPI devel] Porting the underlying fabric interface

2016-02-04 Thread Howard Pritchard
Hi Durga as an alternative you could implement a libfabric provider for your network. In theory, if you can implement the reliable datagram endpoint type on your network and a tag matching mechanism, you could then just use the ofi mtl and not have to do much if anything in open mpi or mpich etc

Re: [OMPI devel] tm-less tm module

2016-01-25 Thread Howard Pritchard
HI Folks, I like Paul's suggestion for configury summary output a lot. It would have helped me when I was trying to deal with an oddball one-off install of the moab/torque software on one of the non-standard front ends at LANL. The libfabric configury has such a summary output at the end of conf

Re: [OMPI devel] tm-less tm module

2016-01-25 Thread Howard Pritchard
Hi Gilles I would prefer improving the faq rather than adding yet more complexity in this area. The way things go you would add this feature then someone else with a different use case would complain we had broken something for them. Then we would add another mca param to disable the new tm less

[OMPI devel] UH jenkins node seems out for the holidays

2015-12-30 Thread Howard Pritchard
Hi Folks, As those of you working on OMPI PRs this week have noticed, it appears that Univ. Houston's CS department may have shut a number of systems down for the holidays. So, for now, ignore the status of the LANL-distcheck and LANL-dlopen jenkins tests. Hopefully the UH server(s) will be back

Re: [OMPI devel] Proposal on RFCs

2015-11-23 Thread Howard Pritchard
HI Ralph, Let's definitely discuss on 12/1. Unless it's something like code deletion or large extern package update (like the PMIx 1.1 PR or hwloc refresh), just opening a PR that touches like 100+ files across a range of the code base needs more than the current github PR interface provides. I

[OMPI devel] master build fails

2015-10-27 Thread Howard Pritchard
Hi Folks, Looks like master can't build any more, at least not on cray with --enable-picky option: -- make all -j 8 result_stderr --- keyval_lex.c: In function 'yy_get_next_buffer': keyval_lex.c:751:18: warning: comparison between signed and unsigned integer expressions [-Wsign-compare] for (

[OMPI devel] Fwd: mtt-submit, etc.

2015-10-22 Thread Howard Pritchard
Hi Folks, I don't seem to have gotten subscribed yet to mtt-users mail list so forwarding to the dev team. Howard -- Forwarded message -- From: Howard Pritchard List-Post: devel@lists.open-mpi.org Date: 2015-10-22 10:18 GMT-06:00 Subject: mtt-submit, etc. To: mtt-us...

[OMPI devel] HPX?

2015-10-19 Thread Howard Pritchard
Hi Folks, I got some kind of strange email from a jenkins.crest.iu.edu concerning an HPX project. It looks like there's some code on some private repo on crest. Does anyone know anything about this? Howard

Re: [OMPI devel] Access to old users@ and devel@ Open MPI mails

2015-10-02 Thread Howard Pritchard
I'm okay with it as long as they use an MPI based mapreduce to do the analytics. Howard 2015-10-02 9:32 GMT-06:00 Jeff Squyres (jsquyres) : > I've received a request from a researcher at Kansas State University to > get a copy of all old us...@open-mpi.org and de...@open-mpi.org emails. > In th

[OMPI devel] new compiler warning with --enable-picky using UH --disable-dlopen jenkins project

2015-09-25 Thread Howard Pritchard
HI Folks, First, the --disable-dlopen/--distcheck projects do not run on anything having anything to do with Cray. So if you see failures with the disable-dlopen or distcheck projects and chose to ignore them please remember they: 1) run on a vanilla linux (Open Suse 13.1) x86_64 box 2) use gnu

[OMPI devel] busted build

2015-09-25 Thread Howard Pritchard
Hi Folks, The UH distcheck is now failing with this compile error: CC pml_ob1_rdma.lo pml_ob1_irecv.c: In function 'mca_pml_ob1_recv': pml_ob1_irecv.c:138:28: error: called object 'mca_pml_ob1_recvreq' is not a function or function pointer mca_pml_ob1_recvreq(recvreq);

[OMPI devel] PR 595 busted build of mpi_f08

2015-09-25 Thread Howard Pritchard
Hi Folks, Well, jenkins doesn't lie. http://jenkins.open-mpi.org/job/ompi_master_cle5.2up02/595/console Looks like the commits associated with PR 595 busted mpi_f08 build. Its a bit frustrating to set all this jenkins stuff up then have it ignored. Howard

  1   2   3   >