FYI, Just noticed this post from the hdf group:
https://forum.hdfgroup.org/t/hdf5-and-openmpi/5437
/Peter K
pgpmcS_mBlpzB.pgp
Description: OpenPGP digital signature
___
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/lis
> On Feb 20, 2019, at 7:14 PM, Gilles Gouaillardet wrote:
>
> Ryan,
>
> That being said, the "Alarm clock" message looks a bit suspicious.
>
> Does it always occur at 20+ minutes elapsed ?
>
> Is there some mechanism that automatically kills a job if it does not write
> anything to stdout for
;> From: users [mailto:users-boun...@lists.open-mpi.org] On Behalf Of Ryan
>> Novosielski
>> Sent: Thursday, February 21, 2019 1:59 PM
>> To: Open MPI Users
>> Subject: Re: [OMPI users] HDF5 1.10.4 "make check" problems w/OpenMPI
>> 3.1.3
>>
nt: Thursday, February 21, 2019 1:59 PM
> To: Open MPI Users
> Subject: Re: [OMPI users] HDF5 1.10.4 "make check" problems w/OpenMPI
> 3.1.3
>
>
> > On Feb 21, 2019, at 2:52 PM, Gabriel, Edgar
> wrote:
> >
> >> -Original Message-
> >
> On Feb 21, 2019, at 2:52 PM, Gabriel, Edgar wrote:
>
>> -Original Message-
>>> Does it always occur at 20+ minutes elapsed ?
>>
>> Aha! Yes, you are right: every time it fails, it’s at the 20 minute and a
>> couple
>> of seconds mark. For comparison, every time it runs, it runs for 2
that indicates that MPI_Abort has been
> called, for both ompio and romio. I will try to investigate this test to see
> what
> is going on.
> >>>>
> >>>> That being said, your report shows an issue in t_mpi, which passes
> without problems f
t; passed, but then throws message that indicates that MPI_Abort has been
>>>> called, for both ompio and romio. I will try to investigate this test to
>>>> see what is going on.
>>>>
>>>> That being said, your report shows an issue in t_mpi, whic
s-boun...@lists.open-mpi.org] On Behalf Of
Gabriel, Edgar
Sent: Sunday, February 17, 2019 10:34 AM
To: Open MPI Users
Subject: Re: [OMPI users] HDF5 1.10.4 "make check" problems w/OpenMPI
3.1.3
I will also run our testsuite and the HDF5 testsuite on GPFS, I have access to a
GPFS file s
is:
>
> btl = tcp,self
>
>
>
> Thanks
> Edgar
>
>
>
>> -Original Message-
>> From: users [mailto:users-boun...@lists.open-mpi.org] On Behalf Of Ryan
>> Novosielski
>> Sent: Wednesday, February 20, 2019 12:02 PM
>> To: Open MPI Users
>> S
s.open-mpi.org] On Behalf Of Ryan
> Novosielski
> Sent: Wednesday, February 20, 2019 12:02 PM
> To: Open MPI Users
> Subject: Re: [OMPI users] HDF5 1.10.4 "make check" problems w/OpenMPI
> 3.1.3
>
> Does it make any sense that it seems to work fine when OpenMPI and HDF5
; Thanks
>> Edgar
>>
>>
>>
>>> -Original Message-----
>>> From: users [mailto:users-boun...@lists.open-mpi.org] On Behalf Of
>>> Gabriel, Edgar
>>> Sent: Sunday, February 17, 2019 10:34 AM
>>> To: Open MPI Users
>&
: Sunday, February 17, 2019 10:34 AM
To: Open MPI Users
Subject: Re: [OMPI users] HDF5 1.10.4 "make check" problems w/OpenMPI
3.1.3
I will also run our testsuite and the HDF5 testsuite on GPFS, I have access to a
GPFS file system since recently, and will report back on that, but it will take
s
> Edgar
>
>
>
>> -Original Message-
>> From: users [mailto:users-boun...@lists.open-mpi.org] On Behalf Of
>> Gabriel, Edgar
>> Sent: Sunday, February 17, 2019 10:34 AM
>> To: Open MPI Users
>> Subject: Re: [OMPI users] HDF5 1.10.4 "m
pi.org] On Behalf Of
> Gabriel, Edgar
> Sent: Sunday, February 17, 2019 10:34 AM
> To: Open MPI Users
> Subject: Re: [OMPI users] HDF5 1.10.4 "make check" problems w/OpenMPI
> 3.1.3
>
> I will also run our testsuite and the HDF5 testsuite on GPFS, I have access
>
f Ryan
> Novosielski
> Sent: Sunday, February 17, 2019 2:37 AM
> To: users@lists.open-mpi.org
> Subject: Re: [OMPI users] HDF5 1.10.4 "make check" problems w/OpenMPI
> 3.1.3
>
> -BEGIN PGP SIGNED MESSAGE-
> Hash: SHA1
>
> This is on GPFS. I'll try it
ake test'
>>> commands. Will have to check which tests are being run by
>>> that.
>>>
>>> Edgar
>>>
>>>> -Original Message- From: users
>>>> [mailto:users-boun...@lists.open-mpi.org] On Behalf Of
>>>> Gille
ve to check which tests are being run by that.
> >
> > Edgar
> >
> >> -Original Message-
> >> From: users [mailto:users-boun...@lists.open-mpi.org] On Behalf Of Gilles
> >> Gouaillardet
> >> Sent: Saturday, February 16, 2019 1:49 AM
>> Gouaillardet
>> Sent: Saturday, February 16, 2019 1:49 AM
>> To: Open MPI Users
>> Subject: Re: [OMPI users] HDF5 1.10.4 "make check" problems w/OpenMPI
>> 3.1.3
>>
>> Ryan,
>>
>> Can you
>>
>> export OMPI_M
1:49 AM
> To: Open MPI Users
> Subject: Re: [OMPI users] HDF5 1.10.4 "make check" problems w/OpenMPI
> 3.1.3
>
> Ryan,
>
> Can you
>
> export OMPI_MCA_io=^ompio
>
> and try again after you made sure this environment variable is passed by srun
>
Ryan,
Can you
export OMPI_MCA_io=^ompio
and try again after you made sure this environment variable is passed by srun
to the MPI tasks ?
We have identified and fixed several issues specific to the (default) ompio
component, so that could be a valid workaround until the next release.
Cheers,
Hi there,
Honestly don’t know which piece of this puzzle to look at or how to get more
information for troubleshooting. I successfully built HDF5 1.10.4 with RHEL
system GCC 4.8.5 and OpenMPI 3.1.3. Running the “make check” in HDF5 is failing
at the below point; I am using a value of RUNPARALLE
21 matches
Mail list logo