This case was a very simple 6 process test on a single node which ran to completion.
I'm installing mpi 1.8.1 now to see if I see the same issue. I just installed and ran hwloc. What am I looking for? I see basic information PCI (ib0, ib1, mix4_0) PCI(eth0) PCI(eth1) PCI() PCI(sda) and others... When I launch the mpi process I'm using mpirun --mca btl self,sm,openib I have not explicitly specified in mpirun to use processor affinity. When running top (1) I do see that the processes migrate from core to core from time to time. Am I using processor affinity and if so shouldn't the process(s) remain on each individual core throughout execution? Hyperthreading is off. I am not using a rank file nor specifying the mpirun command to explicitly use processor affinity. skl 860-557-2895 CONFIDENTIALITY WARNING: This email may contain privileged or confidential information and is for the sole use of the intended recipients. Unauthorized disclosure or use of this communication is prohibited. If you believe that you have received this email in error, please notify the sender immediately and delete it from your system. -----Original Message----- From: devel [mailto:devel-boun...@open-mpi.org] On Behalf Of Jeff Squyres (jsquyres) Sent: Saturday, May 31, 2014 8:13 AM To: Open MPI Developers Subject: [External] Re: [OMPI devel] Open MPI 1.6.5 opal_paffinity_base_get_physical_socket_id The super short answer is: 1.6.x is old and deprecated; can you upgrade to the 1.8.x series? The short answer is "no" -- paffinity calls should never block, but it depends on how and what you're measuring. The more detailed answer is: your trace below looks like it includes a call to MPI_Abort. Did your process hang during the abort, perchance, and (somehow) get stuck in a process affinity call? Are you able to download and run the lstopo command from the hwloc suite? (http://www.open-mpi.org/software/hwloc/v1.9/) On May 30, 2014, at 2:47 PM, "Lecrenski, Stephen K PW" <stephen.lecren...@pw.utc.com> wrote: > I am running some performance tests (Open SpeedShop) with a program which > uses Open MPI and Infiniband. > > I see a line of code which appears to be taking a considerable amount of cpu > time in relation to other pieces of the code. > > opal_paffinity_base_get_physical_socket_id (libmpi.so.1.0.8: > paffinity_base_wrappers.c,118) > > Exclusive CPU time in seconds. > % of CPU Time > Statement Location (Line Number) > 19031.94 > 38.339796 > paffinity_base_wrappers.c(118) > 14188.66 > 28.583021 > paffinity_base_wrappers.c(113) > 10934.38 > 22.027282 > paffinity_base_wrappers.c(129) > 2185.16 > 4.401999 > paffinity_base_wrappers.c(121) > 1081.96 > 2.179606 > paffinity_base_wrappers.c(130) > 546.93 > 1.101789 > paffinity_base_wrappers.c(114) > 546.17 > 1.100258 > paffinity_base_wrappers.c(65) > 541.67 > 1.091193 > paffinity_base_wrappers.c(126) > 540.52 > 1.088876 > ompi_mpi_abort.c(80) > 2.23 > 0.004492 > ompi_mpi_abort.c(101) > > > Is this normal behavior? > > Thanks, > > Stephen Lecrenski > High Performance Technical Computing > > Pratt & Whitney > 400 Main Street > East Hartford,CT 06108 > Telephone: 860 - 557 - 2895 > Email: stephen.lecren...@pw.utc.com > P Please consider the environment before printing this e-mail > CONFIDENTIALITY WARNING: This email may contain privileged or confidential > information and is for the sole use of the intended recipients. Unauthorized > disclosure or use of this communication is prohibited. If you believe that > you have received this email in error, please notify the sender immediately > and delete it from your system. > > _______________________________________________ > devel mailing list > de...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel > Link to this post: > http://www.open-mpi.org/community/lists/devel/2014/05/14915.php -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/ _______________________________________________ devel mailing list de...@open-mpi.org Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel Link to this post: http://www.open-mpi.org/community/lists/devel/2014/05/14916.php