Re: [OMPI devel] PATCH: Wrong event_type value passed in to show_help when getting xrc async events

2010-07-15 Thread nadia.derbey
On Thu, 2010-07-15 at 08:21 -0400, Jeff Squyres wrote: > On Jul 15, 2010, at 8:22 AM, nadia.derbey wrote: > > > So the solution is: > > 1. leave the intermediate event_type declared as an int. > > 2. then: > > . either cast it to ibv_event_type when c

[OMPI devel] mpirun options in the MPMD model

2010-11-09 Thread nadia.derbey
--output-filename is global to the job: even if it is given on any single line of an application context, with different values, the last value is the one that is actually taken as an output file prefix. Regards, Nadia -- nadia.derbey <nadia.der...@bull.net>

[OMPI devel] Fix a hang in carto_base_select() if carto_module_init() fails

2011-06-28 Thread nadia.derbey
Hi, When using the carto/file module with a syntactically incorrect carto file, we get stuck into opal_carto_base_select(). The attached trivial patch fixes the issue. Regards, Nadia -- nadia.derbey <nadia.der...@bull.net> Fix a hang in carto_base_select if carto_module_init fails d

[OMPI devel] known limitation or bug in hwloc?

2011-08-29 Thread nadia.derbey
Hi list, I'm hitting a limitation with paffinity/hwloc with cpu numbers >= 64. In opal/mca/paffinity/hwloc/paffinity_hwloc_module.c, module_set() is the routine that sets the calling process affinity to the mask given as parameter. Note that "mask" is a opal_paffinity_base_cpu_set_t (so we allow

[OMPI devel] Segfault in odls_fork_local_procs() for some values of npersocket

2011-11-08 Thread nadia.derbey
Hi, In v1.5, when mpirun is called with both the "-bind-to-core" and "-npersocket" options, and the npersocket value leads to less procs than sockets allocated on one node, we get a segfault Testing environment: openmpi v1.5 2 nodes with 4 8-cores sockets each mpirun -n 10 -bind-to-core

[OMPI devel] btl/openib: get_ib_dev_distance doesn't see processes as bound if the job has been launched by srun

2012-01-27 Thread nadia.derbey
Hi, If a job is launched using "srun --resv-ports --cpu_bind:..." and slurm is configured with: TaskPlugin=task/affinity TaskPluginParam=Cpusets each rank of that job is in a cpuset that contains a single CPU. Now, if we use carto on top of this, the following happens in