HI All,
I opened a new issue to track the coll_perf failure in case its not related
to the HDF5 problem reported earlier.
https://github.com/open-mpi/ompi/issues/8246
Howard
Am Mo., 23. Nov. 2020 um 12:14 Uhr schrieb Dave Love via users <
users@lists.open-mpi.org>:
> Mark Dixon via users
* MPI_ERR_UNKNOWN: unknown error*
>
> *[osboxes:02457] *** MPI_ERRORS_ARE_FATAL (processes in this communicator
> will now abort,*
>
> *[osboxes:02457] ***and potentially your MPI job)*
>
> *[osboxes:02458] 1 more process has sent help message help-orted.txt /
> timedout*
> I need spawn only in “worker”. Is there a way or workaround for doing this
> without mpirun?
>
> Thanks a lot for your assistance.
>
>
>
> Martín
>
>
>
>
>
>
>
>
>
> *From: *Howard Pritchard
> *Sent: *lunes, 10 de agosto de 2020 19:13
> *To:
und for doing this
> without mpirun?
> Thanks a lot for your assistance.
>
> Martín
>
>
>
>
> *From: *Howard Pritchard
> *Sent: *lunes, 10 de agosto de 2020 19:13
> *To: *Martín Morales
> *Cc: *Open MPI Users
> *Subject: *Re: [OMPI users] OMPI 4.0.4 crashes
rtunately the issue persists in OMPI 4.0.5rc1. Do I have
> to post this on the bug section? Thanks and regards.
>
>
>
> Martín
>
>
>
> *From: *Howard Pritchard
> *Sent: *lunes, 10 de agosto de 2020 14:44
> *To: *Open MPI Users
> *Cc: *Martín Morales
>
Hello Martin,
Between Open MPI 4.0.1 and Open MPI 4.0.4 we upgraded the internal PMIx
version that introduced a problem with spawn for the 4.0.2-4.0.4 versions.
This is supposed to be fixed in the 4.0.5 release. Could you try the
4.0.5rc1 tarball and see if that addresses the problem you're
Hello Michael,
Not sure what could be causing this in terms of delta between v4.0.3 and
v4.0.4.
Two things to try
- add --debug-daemons and --mca pmix_base_verbose 100 to the mpirun line
and compare output from the v4.0.3 and v4.0.4 installs
- perhaps try using the
itialize with >100 processes per
> node. I get the same error message for multiple different codes, so the
> error code is mpi related rather than being program specific.
>
>
>
> Collin
>
>
>
> *From:* Howard Pritchard
> *Sent:* Monday, January 27, 2020 11:20
Hello Collen,
Could you provide more information about the error. Is there any output
from either Open MPI or, maybe, UCX, that could provide more information
about the problem you are hitting?
Howard
Am Mo., 27. Jan. 2020 um 08:38 Uhr schrieb Collin Strassburger via users <
Hello Mark,
You may want to checkout this package:
https://github.com/lanl/libquo
Another option would be to do something like use an MPI_Ibarrier in the
application
with all the MPI processes but rank 0 going into a loop over waiting for
completion of the barrier
and doing a sleep. Once rank
] IMB-MPI1[0x401d49]
>> > [phoebe:07408] *** End of error message ***
>> > IMB-MPI1[0x4022ea]
>> > [titan:07169] [ 8]
>> /usr/lib64/libc.so.6(__libc_start_main+0xf5)[0x7fc025d5a3d5]
>> > [titan:07169] [ 9] IMB-MPI1[0x401d49]
>> > [titan:07169] *
HI Adam,
As a sanity check, if you try to use --mca btl self,vader,tcp
do you still see the segmentation fault?
Howard
Am Mi., 20. Feb. 2019 um 08:50 Uhr schrieb Adam LeBlanc <
alebl...@iol.unh.edu>:
> Hello,
>
> When I do a run with OpenMPI v4.0.0 on Infiniband with this command:
> mpirun
Hi Matt
Definitely do not include the ucx option for an omnipath cluster. Actually
if you accidentally installed ucx in it’s default location use on the
system Switch to this config option
—with-ucx=no
Otherwise you will hit
https://github.com/openucx/ucx/issues/750
Howard
Gilles
Hi Sigmar,
I observed this problem yesterday myself and should have a fix in to master
later today.
Howard
Am Fr., 4. Jan. 2019 um 05:30 Uhr schrieb Siegmar Gross <
siegmar.gr...@informatik.hs-fulda.de>:
> Hi,
>
> I've installed (tried to install) openmpi-master-201901030305-ee26ed9 on
> my
th", is set and doesn't point to any location
> that includes at least one usable plugin for this framework.
>
> Please check your installation and environment.
> ------
>
> Regards,
> Eduardo
>
&g
Hi Eduardo
Could you post the config.log for the build with internal PMIx so we can
figure that out first.
Howard
Eduardo Rothe via users schrieb am Fr. 14. Dez.
2018 um 09:41:
> Open MPI: 4.0.0
> PMIx: 3.0.2
> OS: Debian 9
>
> I'm building a debian package for Open MPI and either I get the
Hi Bert,
If you'd prefer to return to the land of convenience and don't need to mix
MPI
and OpenSHMEM, then you may want to try the path I outlined in the email
archived at the following link
https://www.mail-archive.com/users@lists.open-mpi.org/msg32274.html
Howard
Am Di., 13. Nov. 2018 um
Hello Bert,
What OS are you running on your notebook?
If you are running Linux, and you have root access to your system, then
you should be able to resolve the Open SHMEM support issue by installing
the XPMEM device driver on your system, and rebuilding UCX so it picks
up XPMEM support.
The
HI Si,
Could you add --disable-builtin-atomics
to the configure options and see if the hang goes away?
Howard
2018-07-02 8:48 GMT-06:00 Jeff Squyres (jsquyres) via users <
users@lists.open-mpi.org>:
> Simon --
>
> You don't currently have another Open MPI installation in your PATH /
>
Hello Charles
You are heading in the right direction.
First you might want to run the libfabric fi_info command to see what
capabilities you picked up from the libfabric RPMs.
Next you may well not actually be using the OFI mtl.
Could you run your app with
export
Hi Craig,
You are experiencing problems because you don't have a transport installed
that UCX can use for oshmem.
You either need to go and buy a connectx4/5 HCA from mellanox (and maybe a
switch), and install that
on your system, or else install xpmem (https://github.com/hjelmn/xpmem).
Note
i.org>
> *Date: *Friday, May 4, 2018 at 7:46 AM
> *To: *Open MPI Users <users@lists.open-mpi.org>
> *Subject: *Re: [OMPI users] Debug build of v3.0.1 tarball
>
>
>
> HI Adam,
>
>
>
> Sorry didn't notice you did try the --enable-debug flag. That should not
HI Adam,
Sorry didn't notice you did try the --enable-debug flag. That should not
have
led to the link error building the opal dso. Did you do a make clean after
rerunning configure?
Howard
2018-05-04 8:22 GMT-06:00 Howard Pritchard <hpprit...@gmail.com>:
> Hi Adam,
>
> Di
Hi Adam,
Did you try using the --enable-debug configure option along with your
CFLAGS options?
You may want to see if that simplifies your build.
In any case, we'll fix the problems you found.
Howard
2018-05-03 15:00 GMT-06:00 Moody, Adam T. :
> Hello Open MPI team,
>
> I'm
Hello Ben,
Thanks for the info. You would probably be better off installing UCX on
your cluster and rebuilding your Open MPI with the
--with-ucx
configure option.
Here's what I'm seeing with Open MPI 3.0.1 on a ConnectX5 based cluster
using ob1/openib BTL:
mpirun -map-by ppr:1:node -np 2
27.10
> 262144 46.97
> 524288 87.55
> 1048576 168.89
> 2097152 331.40
> 4194304 654.08
>
>
> On Feb 7, 2018, at 9:04 PM, Howard Pritchard <hpprit...@gmail.com> wrote:
>
> HI Brian,
>
> A
HI Ben,
I'm afraid this is bad news for using UCX. The problem is that when UCX
was configured/built, it did not
find a transport for doing one sided put/get transfers. If you're feeling
lucky, you may want to
install xpmem (https://github.com/hjelmn/xpmem) and rebuild UCX. This
requires
HI Brian,
As a sanity check, can you see if the ob1 pml works okay, i.e.
mpirun -n 2 --mca pml ob1 --mca btl self,vader,openib ./osu_latency
Howard
2018-02-07 11:03 GMT-07:00 brian larkins :
> Hello,
>
> I’m doing some work with Portals4 and am trying to run some MPI
HI Ben,
Could you set these environment variables and post the output ?
export OMPI_MCA_spml=ucx
export OMPI_MCA_spml_base_verbose=100
then run your test?
Also, what OS are you using?
Howard
2018-02-06 20:10 GMT-07:00 Jeff Hammond :
>
> On Tue, Feb 6, 2018 at 3:58
Hi Phanikumar
It’s unlikely the warning message you are seeing is related to GPU
performance. Have you tried adding
—with-verbs=no
to your config line? That should quash openib complaint.
Howard
Phanikumar Pentyala schrieb am Mo. 11. Dez. 2017
um 22:43:
> Dear
Hi Ben,
Actually I did some checking about the brew install for OFi libfabric.
It looks like if your brew is up to date, it will pick up libfabric 1.5.2.
Howard
2017-11-22 15:21 GMT-07:00 Howard Pritchard <hpprit...@gmail.com>:
> HI Ben,
>
> Even on one box, the yoda compone
HI Ben,
Even on one box, the yoda component doesn't work any more.
If you want to do OpenSHMEM programming on you Macbook pro (like I do)
and you don't want to set up a VM to use UCX, then you can use
Sandia OpenSHMEM implementation.
https://github.com/Sandia-OpenSHMEM/SOS
You will need to
found
>
> [shepard-lsm1:49499] Set MCA parameter "orte_base_help_aggregate" to 0 to
> see all help / error messages
>
>
>
> --
>
> Si Hammond
>
> Scalable Computer Architectures
>
> Sandia National Laboratories, NM, USA
>
>
>
>
>
> *From
HI Ben,
What version of Open MPI are you trying to use?
Also, could you describe something about your system. If its a cluster
what sort of interconnect is being used.
Howard
2017-11-20 14:13 GMT-07:00 Benjamin Brock :
> What's the proper way to use shmem_int_fadd()
Hello Ake,
Would you mind opening an issue on Github so we can track this?
https://github.com/open-mpi/ompi/issues
There's a template to show what info we need to fix this.
Thanks very much for reporting this,
Howard
2017-11-20 3:26 GMT-07:00 Åke Sandgren :
> Hi!
Hello Bennet,
What you are trying to do using srun as the job launcher should work.
Could you post the contents
of /etc/slurm/slurm.conf for your system?
Could you also post the output of the following command:
ompi_info --all | grep pmix
to the mail list.
the config.log from your build would
ort2/src/openmpi-2.0.4/op
> enmpi-2.0.4rc2-Linux.x86_64.64_cc/opal/mca/pmix/pmix112'
> Makefile:2301: recipe for target 'all-recursive' failed
> make[1]: *** [all-recursive] Error 1
> make[1]: Leaving directory '/export2/src/openmpi-2.0.4/op
> enmpi-2.0.4rc2-Linux.x86_64.64_cc/opal'
>
Hello Cooper
Could you rerun your test with the following env. variable set
export OMPI_MCA_coll=self,basic,libnbc
and see if that helps?
Also, what type of interconnect are you using - ethernet, IB, ...?
Howard
2017-09-19 8:56 GMT-06:00 Cooper Burns :
>
Hi Siegmar,
Opened issue 4151 to track this.
Thanks,
Howard
2017-08-21 7:13 GMT-06:00 Siegmar Gross <
siegmar.gr...@informatik.hs-fulda.de>:
> Hi,
>
> I've installed openmpi-2.1.2rc2 on my "SUSE Linux Enterprise Server 12.2
> (x86_64)" with Sun C 5.15 (Oracle Developer Studio 12.6) and
Hi Siegmar,
I opened issue 4151 to track this. This is relevant to a project to get
open mpi to build with -Werror.
Thanks very much,
Howard
2017-08-21 7:27 GMT-06:00 Siegmar Gross <
siegmar.gr...@informatik.hs-fulda.de>:
> Hi,
>
> I've installed openmpi-master-201708190239-9d3f451 on my
Hi John,
In the 2.1.x release stream a shared memory capability was introduced into
the PMIx component.
I know nothing about LXC containers, but it looks to me like there's some
issue when PMIx tries
to create these shared memory segments. I'd check to see if there's
something about your
Forgot you probably need an equal sign after btl arg
Howard Pritchard <hpprit...@gmail.com> schrieb am Mi. 22. März 2017 um
18:11:
> Hi Goetz
>
> Thanks for trying these other versions. Looks like a bug. Could you post
> the config.log output from your build of the 2.1.0 to
)
Howard
Götz Waschk <goetz.was...@gmail.com> schrieb am Mi. 22. März 2017 um 13:09:
On Wed, Mar 22, 2017 at 7:46 PM, Howard Pritchard <hpprit...@gmail.com>
wrote:
> Hi Goetz,
>
> Would you mind testing against the 2.1.0 release or the latest from the
> 1.10.x series (1.1
Hi Goetz,
Would you mind testing against the 2.1.0 release or the latest from the
1.10.x series (1.10.6)?
Thanks,
Howard
2017-03-22 6:25 GMT-06:00 Götz Waschk :
> Hi everyone,
>
> I'm testing a new machine with 32 nodes of 32 cores each using the IMB
> benchmark. It
ut of ompi_info of master on my laptop. Please let me
> know if I can help with anything else.
>
> Thanks,
> Joseph
>
> On 03/01/2017 11:24 PM, Howard Pritchard wrote:
>
> Hi Joseph,
>
> I built this test with craypich (Cray MPI) and it passed. I also tried
> with Ope
Hi Edgar
Please open an issue too so we can track the fix.
Howard
Edgar Gabriel schrieb am Fr. 3. März 2017 um
07:45:
> Nicolas,
>
> thank you for the bug report, I can confirm the behavior. I will work on
> a patch and will try to get that into the next release,
Hi Joseph,
I built this test with craypich (Cray MPI) and it passed. I also tried
with Open MPI master and the test passed. I also tried with 2.0.2
and can't seem to reproduce on my system.
Could you post the output of config.log?
Also, how intermittent is the problem?
Thanks,
Howard
Hi Orion
Does the problem occur if you only use font2 and 3? Do you have MXM
installed on the font1 node?
The 2.x series is using PMIX and it could be that is impacting the PML
sanity check.
Howard
Orion Poplawski schrieb am Mo. 27. Feb. 2017 um 14:50:
> We have a
Hi Joseph
What OS are you using when running the test?
Could you try running with
export OMPI_mca_osc=^pt2pt
and
export OMPI_mca_osc_base_verbose=10
This error message was put in to this OMPI release because this part of the
code has known problems when used multi threaded.
Joseph Schuchart
Hi Anastasia,
Definitely check the mpirun when in batch environment but you may also want
to upgrade to Open MPI 2.0.2.
Howard
r...@open-mpi.org schrieb am Mi. 15. Feb. 2017 um 07:49:
> Nothing immediate comes to mind - all sbatch does is create an allocation
> and then run
value Bad parameter (-5) instead of ORTE_SUCCESS
--
On Thu, Feb 2, 2017 at 12:29 PM, Howard Pritchard <hpprit...@gmail.com>
wrote:
Hi Michel
Try adding --enable-static to the configure.
That fixed the problem for me.
t;
> Brendan
>
>
>
> *From:* users [mailto:users-boun...@lists.open-mpi.org
> <users-boun...@lists.open-mpi.org>] *On Behalf Of *Howard Pritchard
> *Sent:* Tuesday, January 24, 2017 8:21 AM
> *To:* Open MPI Users <users@lists.open-mpi.org>
> *Subject:* Re: [OMP
Hi Michel
Try adding --enable-static to the configure.
That fixed the problem for me.
Howard
Michel Lesoinne schrieb am Mi. 1. Feb. 2017 um
19:07:
> I have compiled OpenMPI 2.0.2 on a new Macbook running OS X 10.12 and have
> been trying to run simple program.
> I
Hi Michael,
I reproduced this problem on my Mac too:
pn1249323:~/ompi/examples (v2.0.x *)$ mpirun -np 2 ./ring_c
[pn1249323.lanl.gov:94283] mca_base_component_repository_open: unable to
open mca_patcher_overwrite: File not found (ignored)
[pn1249323.lanl.gov:94283]
Hi Michel
It's somewhat unusual to use the disable-shared configure option. That
may be causing this. Could you try to build without using this option and
see if you still see the problem?
Thanks,
Howard
Michel Lesoinne schrieb am Mi. 1. Feb. 2017 um
21:07:
> I
Hi Wodel
Randomaccess part of HPCC is probably causing this.
Perhaps set PSM env. variable -
Export PSM_MQ_REVCREQ_MAX=1000
or something like that.
Alternatively launch the job using
mpirun --mca plm ob1 --host
to avoid use of psm. Performance will probably suffer with this option
* users [mailto:users-boun...@lists.open-mpi.org] *On Behalf Of *Howard
> Pritchard
> *Sent:* Friday, January 20, 2017 6:35 PM
> *To:* Open MPI Users <users@lists.open-mpi.org>
> *Subject:* Re: [OMPI users] Open MPI over RoCE using breakout cable and
> switch
>
>
>
> Hi Bre
Hi Brendan
I doubt this kind of config has gotten any testing with OMPI. Could you
rerun with
--mca btl_base_verbose 100
added to the command line and post the output to the list?
Howard
Brendan Myers schrieb am Fr. 20. Jan. 2017
um 15:04:
> Hello,
>
> I am
> [loki:13586] *** reported by process [2873294849,0]
> [loki:13586] *** on communicator MPI_COMM_WORLD
> [loki:13586] *** MPI_ERR_UNKNOWN: unknown error
> [loki:13586] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will
> now abort,
> [loki:13586] ***and potenti
HI Siegmar,
Could you post the configury options you use when building the 2.0.2rc3?
Maybe that will help in trying to reproduce the segfault you are observing.
Howard
2017-01-07 2:30 GMT-07:00 Siegmar Gross <
siegmar.gr...@informatik.hs-fulda.de>:
> Hi,
>
> I have installed openmpi-2.0.2rc3
2] mca: base: close: component self closed
> [loki:05572] mca: base: close: unloading component self
> [loki:05572] mca: base: close: component tcp closed
> [loki:05572] mca: base: close: unloading component tcp
> loki spawn 125
>
>
> Kind regards and thank you very muc
HI Siegmar,
I've attempted to reproduce this using gnu compilers and
the version of this test program(s) you posted earlier in 2016
but am unable to reproduce the problem.
Could you double check that the slave program can be
successfully run when launched directly by mpirun/mpiexec?
It might
Hi Paul,
Thanks very much Christmas present.
The Open MPI README has been updated
to include a note about issues with the Intel 16.0.3-4 compiler suites.
Enjoy the holidays,
Howard
2016-12-23 3:41 GMT-07:00 Paul Kapinos :
> Hi all,
>
> we discussed this issue
;Anyway, * /dev/hfi1_0* doesn't exist.
>>
>> Make sure you have the hfi1 module/driver loaded.
>>
>> In addition, please confirm the links are in active state on all the
>> nodes `opainfo`
>>
>>
>>
>> _MAC
>>
>>
>>
>&
hello Daniele,
Could you post the output from ompi_info command? I'm noticing on the RPMS
that came with the rhel7.2 distro on
one of our systems that it was built to support psm2/hfi-1.
Two things, could you try running applications with
mpirun --mca pml ob1 (all the rest of your args)
and
Hi Jeff,
I don't think it was the use of memkind itself, but a need to refactor the
way Open MPI is using info objects
that was the issue. I don't recall the details.
Howard
2016-11-22 16:27 GMT-07:00 Jeff Hammond :
>
>>
>>1. MPI_ALLOC_MEM integration with memkind
on whether to go with a v2.2.x release next
year or to go from v2.1.x to v3.x in late 2017 or early 2018 at the link
below:
https://www.open-mpi.org/sc16/
Thanks very much,
Howard
--
Howard Pritchard
HPC-DES
Los Alamos National Laboratory
___
users mailing
Hi Christof,
Thanks for trying out 2.0.1. Sorry that you're hitting problems.
Could you try to run the tests using the 'ob1' PML in order to
bypass PSM2?
mpirun --mca pml ob1 (all the rest of the args)
and see if you still observe the failures?
Howard
2016-11-18 9:32 GMT-07:00 Christof
HI Brenda,
I should clarify as my response may confuse folks. We had configured the
connectx4 cards to use
ethernet/RoCE rather than IB transport for these measurements.
Howard
2016-11-08 16:08 GMT-07:00 Howard Pritchard <hpprit...@gmail.com>:
> Hi Brenda,
>
> What type of
Hi Brenda,
What type of ethernet device (is this a Mellanox HCA?) and ethernet switch
are you using? The mpirun configure
options look correct to me. Is it possible that you have all the mpi
processes on a single node?
It should be pretty obvious from the SendRecv IMB test if you're using
RoCE.
HI David,
If you are using srun, you can
export OMPI_MCA_pmix_base_verbose=10
and there will be output to show which SLURM pmi library you are using.
Howard
2016-10-13 12:55 GMT-06:00 David Shrader :
> That is really good to know. Thanks!
> David
>
>
> On 10/13/2016 12:27
Hi Joseph,
Thanks for reporting this problem.
There's an issue now (#2012)
https://github.com/open-mpi/ompi/issues/2012
to track this.
Howard
2016-08-25 7:44 GMT-06:00 Christoph Niethammer :
> Hello,
>
> The Error is not 100% reproducible for me every time but seems to
>
Hi Gundram
Could you configure without the disable dlopen option and retry?
Howard
Am Freitag, 8. Juli 2016 schrieb Gilles Gouaillardet :
> the JVM sets its own signal handlers, and it is important openmpi dones
> not override them.
> this is what previously happened with PSM (infinipath) but
Hi Siegmar,
Sorry for the delay, I seem to have missed this one.
It looks like there's an error in the way the native methods are processing
java exceptions. The code correctly builds up an exception message for
cases where MPI 'c' returns non-success but, not if the problem occured
in one of
601ae65dc34a347c*
>
> * Env[3]: OMPI_MCA_orte_peer_modex_id=0*
>
> * Env[4]: OMPI_MCA_orte_peer_init_barrier_id=1*
>
> * Env[5]: OMPI_MCA_orte_peer_fini_barrier_id=2*
>
> * Env[6]: TMPDIR=/var/folders/5t/6tqp003x4fn09fzgtx46tjdhgn/T/*
>
> * Env[7]: __CF_USER_TEXT_ENCODING=0
Hello Claudio,
mpirun should be combining your java.library.path option with the one
needed to add
the Open MPI's java bindings as well.
Which version of Open MPI are you using?
Could you first try to compile the Ring.java code in ompi/examples and run
it with the
following additional mpirun
Hi Durga,
I'd suggest reposting this to the libfabric-users mail list.
You can join that list at
http://lists.openfabrics.org/mailman/listinfo/libfabric-users
I'd suggest including the output of config.log. If you installed
ofed in non-canonical location, you may need to give an explicit
path
Hello Saliya,
Sorry i did not see this email earlier. There are a bunch of java test
codes including performance tests like used in the paper at
https://github.com/open-mpi/ompi-java-test
Howard
2016-02-27 23:01 GMT-07:00 Saliya Ekanayake :
> Hi,
>
> I see this paper from
HI Matt,
If you don't need oshmem, you could try again with --disable-oshmem added
to the config line
Howard
2016-01-22 12:15 GMT-07:00 Matt Thompson :
> All,
>
> I'm trying to duplicate an issue I had with ESMF long ago (not sure if I
> reported it here or at ESMF,
HI Ibrahim,
Are you using a 32bit or 64bit JVM?
I don't think this is an Open MPI issue, but likely something owing to your
app or your java setup.
You may want to checkout
http://javaeesupportpatterns.blogspot.com/2012/09/outofmemoryerror-unable-to-create-new.html
If you'd like to post the
HI Sigmar,
Would you mind posting your MsgSendRecvMain to the mail list? I'd like to
see if I can
reproduce it on my linux box.
Thanks,
Howard
2016-01-14 7:30 GMT-07:00 Siegmar Gross <
siegmar.gr...@informatik.hs-fulda.de>:
> Hi,
>
> I've successfully built openmpi-v1.10.1-140-g31ff573 on
Hi Marko,
You can probably find examples of what you'd like to do on github:
https://github.com/open-mpi/ompi-java-test
There are numerous MPI-2 RMA examples in the one-sided subdirectory.
If you've never used github before, jus click on the download as zip
button in the upper right hand
either on edison or cori.
Howard
2015-11-19 17:11 GMT-07:00 Howard Pritchard <hpprit...@gmail.com>:
> Hi Jeff H.
>
> Why don't you just try configuring with
>
> ./configure --prefix=my_favorite_install_dir
> --with-libfabric=install_dir_for_libfabric
> make -j 8 insta
Hi Ibrahim,
If you just try to compile with the javac do you at least see a "error:
package mpi..." does not exist?
Adding the "-verbose" option may also help with diagnosing the problem.
If the javac doesn't get that far then your problem is with the java
install.
Howard
2015-11-19 6:45
Hello Ibrahim
As a sanity check, could you try to compile the Hello.java in examples?
mpijavac --verbose Hello.java
you should see something like:
/usr/bin/javac -cp
/global/homes/h/hpp/ompi_install/lib/mpi.jar:/global/homes/h/hpp/ompi_install/lib/shmem.jar
Hello.java
You may also want to
the answer is yes,
although the default is for Open MPI to use mtl/psm on that network.
>
> Please forgive my ignorance, the amount of different options is rather
> overwhelming..
>
> Marcin
>
>
>
> On 09/30/2015 04:26 PM, Howard Pritchard wrote:
>
> Hello Marcin
>
&
Hello Marcin
What configure options are you using besides with-libfabric?
Could you post your config.log file tp the list?
Looks like you only install fi_ext_usnic.h if you could build the usnic
libfab provider. When you configured libfabric what providers were listed
at the end of configure
>
> Cheers,
>
> Gilles
>
> On Saturday, August 15, 2015, Howard Pritchard <hpprit...@gmail.com>
> wrote:
>
>> Hi Jeff,
>>
>> I don't know why Gilles keeps picking on the persistent request problem
>> and mixing
>> it up with this user bug. I
, but I'd be greatly surprised if
> he had InfiniPath on his systems where he ran into this segv issue...?
>
>
> > On Aug 14, 2015, at 1:08 PM, Howard Pritchard <hpprit...@gmail.com>
> wrote:
> >
> > Hi Gilles,
> >
> > Good catch! Nate we hadn't been testing o
hare/info --with-bugurl=
>>> <http://bugzilla.redhat.com/bugzilla>http://bugzilla.redhat.com/bugzilla
>>> --enable-bootstrap --enable-shared --enable-threads=posix
>>> --enable-checking=release --with-system-zlib --enable-__cxa_atexit
>>> --di
ext file of what I see when I run it with mpirun and your
> debug flag. Lots of debug lines.*
>
>
> Nate
>
>
>
>
>
> On Wed, Aug 12, 2015 at 11:09 AM, Howard Pritchard <hpprit...@gmail.com>
> wrote:
>
>> Hi Nate,
>>
>> Sorry for the delay in g
-1.5.0.0/jre
> --enable-libgcj-multifile --enable-java-maintainer-mode
> --with-ecj-jar=/usr/share/java/eclipse-ecj.jar --disable-libjava-multilib
> --with-ppl --with-cloog --with-tune=generic --with-arch_32=i686
> --build=x86_64-redhat-linux
> Thread model: posix
> gcc version 4
can uncompress to have an input
> directory. The text file is just the same line over and over again. Run it
> as:
>
> *java MPITestBroke tweets/*
>
>
> Nate
>
>
>
>
>
> On Wed, Aug 5, 2015 at 8:29 AM, Howard Pritchard <hpprit...@gmail.com>
> wrote:
>
t;
> On Tue, Aug 4, 2015 at 12:26 PM, Howard Pritchard <hpprit...@gmail.com>
> wrote:
>
>> Hello Nate,
>>
>> As a sanity check of your installation, could you try to compile the
>> examples/*.java codes using the mpijavac you've installed and see that
>>
n calling MPI.init().
>
> I'm not great at debugging this sort of stuff, but happy to try things out
> if you need me to.
>
> Nate
>
>
> On Tue, Aug 4, 2015 at 5:09 AM, Howard Pritchard <hpprit...@gmail.com>
> wrote:
>
>> Hello Nate,
>>
>> As a firs
Hello Nate,
As a first step to addressing this, could you please try using gcc rather
than the Intel compilers to build Open MPI?
We've been doing a lot of work recently on the java bindings, etc. but have
never tried using any compilers other
than gcc when working with the java bindings.
; Software Engineer
> Cray, Inc.
> ------
> *From:* users [users-boun...@open-mpi.org] on behalf of Howard Pritchard [
> hpprit...@gmail.com]
> *Sent:* Thursday, June 25, 2015 11:00 PM
> *To:* Open MPI Users
> *Subject:* Re: [OMPI users] Running with native
[users-boun...@open-mpi.org] on behalf of Howard Pritchard [
> hpprit...@gmail.com]
> *Sent:* Thursday, June 25, 2015 5:11 PM
> *To:* Open MPI Users
> *Subject:* Re: [OMPI users] Running with native ugni on a Cray XC
>
> Hi Nick
>
> use master not 1.8.x. for cray xc. also for
Hi Nick
use master not 1.8.x. for cray xc. also for config do not pay attention to
cray/lanl platform files. just do config. also if using nativized slurm
launch with srun not mpirun.
howard
--
sent from my smart phonr so no good type.
Howard
On Jun 25, 2015 2:56 PM, "Nick
Hello Fedele,
Would it be possible to build the open mpi package to use gfortran
and run the test again?
Do you observe this problem if you build a Open MP (<-> MP not MPI)
only version of the test case?
I can't reproduce this problem using gfortran. I don't have access to an
Intel compiler at
1 - 100 of 127 matches
Mail list logo