[OMPI users] Question on process and memory affinity with 1.8.1

2014-07-21 Thread Blosch, Edwin L
In making the leap from 1.6 to 1.8, how can I check whether or not process/memory affinity is supported? I've built OpenMPI on a system where the numactl-devel package was not installed, and another where it was, but I can't see anything in the output of ompi_info that suggests any difference

Re: [OMPI users] MPIIO and derived data types

2014-07-21 Thread Tom Rosmond
Rob, Well, if I understand what you are saying, maybe my concerns about sub-optimality are unfounded. Yes, the 'buffer' array is contiguous in memory for each subdomain, and that is what the successful call to 'mpi_file_read_at_all' sees. The call to 'mpi_type_indexed' creates my 'datatype' and

Re: [OMPI users] MPIIO and derived data types

2014-07-21 Thread Rob Latham
On 07/20/2014 04:23 PM, Tom Rosmond wrote: Hello, For several years I have successfully used MPIIO in a Fortran global atmospheric ensemble data assimilation system. However, I always wondered if I was fully exploiting the power of MPIIO, specifically by using derived data types to better

Re: [OMPI users] Help with multirail configuration

2014-07-21 Thread Tobias Kloeffel
Hi Rolf, yes this is exactly what I was looking for, I just hoped that there is also a way to manually control this behavior. But in most cases that would be the best setting. Thanks! Tobias On 07/21/2014 05:01 PM, Rolf vandeVaart wrote: With Open MPI 1.8.1, the library will use the NIC

Re: [OMPI users] Help with multirail configuration

2014-07-21 Thread Rolf vandeVaart
With Open MPI 1.8.1, the library will use the NIC that is "closest" to the CPU. There was a bug in earlier versions of Open MPI 1.8 so that did not happen. You can see this by running with some verbosity using the "btl_base_verbose" flag. For example, this is what I observed on a two node

Re: [OMPI users] Errors for openib, mpirun fails

2014-07-21 Thread Shamis, Pavel
You have to check the ports states on *all* nodes in the run/job/submission. Checking on a single node is not enough. My guess is the 01-00 tries to connect 01-01 and the ports are down on 01-01. You may disable support for infiniband by adding --mca btl ^openib. Best, Pavel (Pasha) Shamis ---

Re: [OMPI users] Incorrect escaping of OMPI_MCA environment variables with spaces (for rsh?)

2014-07-21 Thread Dirk Schubert
Hello Ralph, thanks for your answer. > I can look to see if there is something generic we can do (perhaps enclosing each param in quotes to avoid any special character issues) - will see if something like that might help. Best that will happen, however, is that we launch the app and then

[OMPI users] Errors for openib, mpirun fails

2014-07-21 Thread Syed Ahsan Ali
Dear All I need your help to solve this cluster related issue causing mpirun malfunction. I get following warning for some of the nodes and then the route failure message comes causing failure to mpirun. *WARNING: There is at least one OpenFabrics device found but there are no active ports

Re: [OMPI users] Mpirun 1.5.4 problems when request > 28 slots

2014-07-21 Thread Lane, William
Please see: http://bugs.centos.org/view.php?id=5812 From: users [users-boun...@open-mpi.org] on behalf of Ralph Castain [r...@open-mpi.org] Sent: Sunday, July 20, 2014 9:30 AM To: Open MPI Users Subject: Re: [OMPI users] Mpirun 1.5.4 problems when request > 28

Re: [OMPI users] Fwd: Re[4]: Salloc and mpirun problem

2014-07-21 Thread Timur Ismagilov
NIC = network interface controller?  There is QDR Infiniband 4x/10G Ethernet/Gigabit Ethernet. I want to use  QDR Infiniband. Here is a new output: $ mpirun -mca mca_base_env_list 'LD_PRELOAD' --debug-daemons --mca plm_base_verbose 5 -mca oob_base_verbose 10 -mca rml_base_verbose 10 -np 2