Eric,

OpenMPI 2.0.0 has been released, so the fix should land into the v2.x branch shortly.

If i understand correctly, you script download/compile OpenMPI and then download/compile PETSc.

In this is correct, and for the time being, feel free to patch Open MPI v2.x before compiling it, the fix can be

downloaded ad https://patch-diff.githubusercontent.com/raw/open-mpi/ompi-release/pull/1263.patch


Cheers,


Gilles


On 7/14/2016 3:37 AM, Eric Chamberland wrote:
Hi Howard,

ok, I will wait for 2.0.1rcX... ;)

I've put in place a script to download/compile OpenMPI+PETSc(3.7.2) and our code from the git repos.

Now I am in a somewhat uncomfortable situation where neither the ompi-release.git or ompi.git repos are working for me.

The first gives me the errors with MPI_File_write_all_end I reported, but the former gives me errors like these:

[lorien:106919] [[INVALID],INVALID] ORTE_ERROR_LOG: Bad parameter in file ess_singleton_module.c at line 167
*** An error occurred in MPI_Init_thread
*** on a NULL communicator
*** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
***    and potentially your MPI job)
[lorien:106919] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!

So, for my continuous integration of OpenMPI I am in a no man's land... :(

Thanks anyway for the follow-up!

Eric

On 13/07/16 07:49 AM, Howard Pritchard wrote:
Hi Eric,

Thanks very much for finding this problem.   We decided in order to have
a reasonably timely
release, that we'd triage issues and turn around a new RC if something
drastic
appeared.  We want to fix this issue (and it will be fixed), but we've
decided to
defer the fix for this issue to a 2.0.1 bug fix release.

Howard



2016-07-12 13:51 GMT-06:00 Eric Chamberland
<eric.chamberl...@giref.ulaval.ca
<mailto:eric.chamberl...@giref.ulaval.ca>>:

    Hi Edgard,

    I just saw that your patch got into ompi/master... any chances it
    goes into ompi-release/v2.x before rc5?

    thanks,

    Eric


    On 08/07/16 03:14 PM, Edgar Gabriel wrote:

        I think I found the problem, I filed a pr towards master, and if
        that
        passes I will file a pr for the 2.x branch.

        Thanks!
        Edgar


        On 7/8/2016 1:14 PM, Eric Chamberland wrote:


            On 08/07/16 01:44 PM, Edgar Gabriel wrote:

                ok, but just to be able to construct a test case,
                basically what you are
                doing is

                MPI_File_write_all_begin (fh, NULL, 0, some datatype);

                MPI_File_write_all_end (fh, NULL, &status),

                is this correct?

            Yes, but with 2 processes:

            rank 0 writes something, but not rank 1...

            other info: rank 0 didn't wait for rank1 after
            MPI_File_write_all_end so
            it continued to the next MPI_File_write_all_begin with a
            different
            datatype but on the same file...

            thanks!

            Eric
            _______________________________________________
            devel mailing list
            de...@open-mpi.org <mailto:de...@open-mpi.org>
            Subscription:
            https://www.open-mpi.org/mailman/listinfo.cgi/devel
            Link to this post:
http://www.open-mpi.org/community/lists/devel/2016/07/19173.php


    _______________________________________________
    devel mailing list
    de...@open-mpi.org <mailto:de...@open-mpi.org>
    Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/devel
    Link to this post:
http://www.open-mpi.org/community/lists/devel/2016/07/19192.php


_______________________________________________
devel mailing list
de...@open-mpi.org
Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/devel
Link to this post: http://www.open-mpi.org/community/lists/devel/2016/07/19201.php


Reply via email to