Dear Brice, the BIOS is the most latest. However, i wonder if this could be  a 
hardware error, as openmpi sources claim.  Is there any way to find out if this 
is a hardware error?

Thanks


> From: users-requ...@open-mpi.org
> Subject: users Digest, Vol 3074, Issue 1
> To: us...@open-mpi.org
> Date: Sat, 20 Dec 2014 12:00:02 -0500
> 
> Send users mailing list submissions to
>       us...@open-mpi.org
> 
> To subscribe or unsubscribe via the World Wide Web, visit
>       http://www.open-mpi.org/mailman/listinfo.cgi/users
> or, via email, send a message with subject or body 'help' to
>       users-requ...@open-mpi.org
> 
> You can reach the person managing the list at
>       users-ow...@open-mpi.org
> 
> When replying, please edit your Subject line so it is more specific
> than "Re: Contents of users digest..."
> 
> 
> Today's Topics:
> 
>    1. Re: Deadlock in OpenMPI 1.8.3 and PETSc 3.4.5
>       (Jeff Squyres (jsquyres))
>    2. Hwloc error with Openmpi 1.8.3 on AMD 64 (Sergio Manzetti)
>    3. Re: Hwloc error with Openmpi 1.8.3 on AMD 64 (Brice Goglin)
>    4. best function to send data (Diego Avesani)
> 
> 
> ----------------------------------------------------------------------
> 
> Message: 1
> Date: Fri, 19 Dec 2014 19:26:58 +0000
> From: "Jeff Squyres (jsquyres)" <jsquy...@cisco.com>
> To: "Open MPI User's List" <us...@open-mpi.org>
> Cc: "petsc-ma...@mcs.anl.gov" <petsc-ma...@mcs.anl.gov>
> Subject: Re: [OMPI users] Deadlock in OpenMPI 1.8.3 and PETSc 3.4.5
> Message-ID: <027ab453-de85-4f08-bdd7-a676ca90e...@cisco.com>
> Content-Type: text/plain; charset="us-ascii"
> 
> On Dec 19, 2014, at 10:44 AM, George Bosilca <bosi...@icl.utk.edu> wrote:
> 
> > Regarding your second point, while I do tend to agree that such issue is 
> > better addressed in the MPI Forum, the last attempt to fix this was 
> > certainly not a resounding success.
> 
> Yeah, fair enough -- but it wasn't a failure, either.  It could definitely be 
> moved forward, but it will take time/effort, which I unfortunately don't 
> have. I would be willing, however, to spin up someone who *does* have 
> time/effort available to move the proposal forward.
> 
> > Indeed, there is a slight window of opportunity for inconsistencies in the 
> > recursive behavior.
> 
> You're right; it's a small window in the threading case, but a) that's the 
> worst kind :-), and b) the non-threaded case is actually worse (because the 
> global state can change from underneath the loop).
> 
> > But the inconsistencies were already in the code, especially in the single 
> > threaded case. As we never received any complaints related to this topic I 
> > did not deemed interesting to address them with my last commit. Moreover, 
> > the specific behavior needed by PETSc is available in Open MPI when 
> > compiled without thread support, as the only thing that "protects" the 
> > attributes is that global mutex.
> 
> Mmmm.  Ok, I see your point.  But this is a (very) slippery slope.
> 
> > For example, in ompi_attr_delete_all(), it gets the count of all attributes 
> > and then loops <count> times to delete each attribute.  But each attribute 
> > callback can now insert or delete attributes on that entity.  This can mean 
> > that the loop can either fail to delete an attribute (because some 
> > attribute callback already deleted it) or fail to delete *all* attributes 
> > (because some attribute callback added more).
> > 
> > To be extremely precise the deletion part is always correct
> 
> ...as long as the hash map is not altered from the application (e.g., by 
> adding or deleting another attribute during a callback).
> 
> I understand that you mention above that you're not worried about this case.  
> I'm just picking here because there is quite definitely a case where the loop 
> is *not* correct.  PETSc apparently doesn't trigger this badness, but... like 
> I said above, it's a (very) slippery slope.
> 
> > as it copies the values to be deleted into a temporary array before calling 
> > any callbacks (and before releasing the mutex), so we only remove what was 
> > in the object attribute hash when the function was called. Don't 
> > misunderstand we have an extremely good reason to do it this way, we need 
> > to call the callbacks in the order in which they were created (mandated by 
> > the MPI standard).
> >  
> > ompi_attr_copy_all() has similar problems -- in general, the hash that it 
> > is looping over can change underneath it.
> > 
> > For the copy it is a little bit more tricky, as the calling order is not 
> > imposed. Our peculiar implementation of the hash table (with array) makes 
> > the code work, with a single (possible minor) exception when the hash table 
> > itself is grown between 2 calls. However, as stated before this issue was 
> > already present in the code in single threaded cases for years. Addressing 
> > it is another 2 line patch, but I leave this exercise to an interested 
> > reader.
> 
> Yeah, thanks for that.  :-)
> 
> To be clear: both the copy and the delete code could be made thread safe.  I 
> just don't think we should be encouraging users to be exercising undefined / 
> probably not-portable MPI code.
> 
> -- 
> Jeff Squyres
> jsquy...@cisco.com
> For corporate legal information go to: 
> http://www.cisco.com/web/about/doing_business/legal/cri/
> 
> 
> 
> ------------------------------
> 
> Message: 2
> Date: Fri, 19 Dec 2014 20:58:46 +0100
> From: Sergio Manzetti <sergio.manze...@outlook.com>
> To: "us...@open-mpi.org" <us...@open-mpi.org>
> Subject: [OMPI users] Hwloc error with Openmpi 1.8.3 on AMD 64
> Message-ID: <dub126-w2190e22e21596a1b834cf4e3...@phx.gbl>
> Content-Type: text/plain; charset="iso-8859-1"
> 
> 
> 
> 
> 
> 
> Dear all, when trying to run NWchem with openmpi, I get this error.
>  
>  
>  
> ****************************************************************************
> * Hwloc has encountered what looks like an error from the operating system.
> *
> * object intersection without inclusion!
> * Error occurred in topology.c line 594
> *
> * Please report this error message to the hwloc user's mailing list,
> * along with the output from the hwloc-gather-topology.sh script.
>  
> Is there any rationale for solving this?
>  
> Thanks
>                                                                               
>   
> -------------- next part --------------
> HTML attachment scrubbed and removed
> 
> ------------------------------
> 
> Message: 3
> Date: Fri, 19 Dec 2014 21:13:19 +0100
> From: Brice Goglin <brice.gog...@inria.fr>
> To: Open MPI Users <us...@open-mpi.org>
> Subject: Re: [OMPI users] Hwloc error with Openmpi 1.8.3 on AMD 64
> Message-ID: <549486df.50...@inria.fr>
> Content-Type: text/plain; charset="windows-1252"
> 
> Hello,
> 
> The rationale is to read the message and do what it says :)
> 
> Have a look at
>     www.open-mpi.org/projects/hwloc/doc/v1.10.0/a00028.php#faq_os_error
> Try upgrading your BIOS and kernel.
> 
> Otherwise install hwloc and send the output (tarball) of
> hwloc-gather-topology to hwloc-users (not to OMPI users).
> 
> thanks
> Brice
> 
> 
> 
> Le 19/12/2014 20:58, Sergio Manzetti a ?crit :
> >
> >
> > Dear all, when trying to run NWchem with openmpi, I get this error.
> >  
> >  
> >  
> > ****************************************************************************
> > * Hwloc has encountered what looks like an error from the operating
> > system.
> > *
> > * object intersection without inclusion!
> > * Error occurred in topology.c line 594
> > *
> > * Please report this error message to the hwloc user's mailing list,
> > * along with the output from the hwloc-gather-topology.sh script.
> >  
> > Is there any rationale for solving this?
> >  
> > Thanks
> >
> >
> > _______________________________________________
> > users mailing list
> > us...@open-mpi.org
> > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> > Link to this post: 
> > http://www.open-mpi.org/community/lists/users/2014/12/26045.php
> 
> -------------- next part --------------
> HTML attachment scrubbed and removed
> 
> ------------------------------
> 
> Message: 4
> Date: Fri, 19 Dec 2014 23:56:36 +0100
> From: Diego Avesani <diego.aves...@gmail.com>
> To: Open MPI Users <us...@open-mpi.org>
> Subject: [OMPI users] best function to send data
> Message-ID:
>       <cag8o1y4b0uwydtrb+swdbra4tbk6ih5toeypga8b6vs-oty...@mail.gmail.com>
> Content-Type: text/plain; charset="utf-8"
> 
> dear all users,
> I am new in MPI world.
> I would like to know what is the best choice and meaning between different
> function.
> 
> In my program I would like that each process send a vector of data to all
> the other process. What do you suggest?
> Is it correct MPI_Bcast or I am missing something?
> 
> Thanks a lot
> 
> Diego
> -------------- next part --------------
> HTML attachment scrubbed and removed
> 
> ------------------------------
> 
> Subject: Digest Footer
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> ------------------------------
> 
> End of users Digest, Vol 3074, Issue 1
> **************************************
                                          

Reply via email to