Is there a way to assign one thread to one core? Also from code, not
necessary with OpenMPI option.

Thanks.

2008/11/19 Stephen Wornom <stephen.wor...@sophia.inria.fr>:
> Gabriele Fatigati wrote:
>>
>> Ok,
>> but in Ompi 1.3 how can i enable it?
>>
>
> This may not be relevant, but I could not get a hybrid mpi+OpenMP code to
> work correctly.
> Would my problem be related to Gabriele's and perhaps fixed in openmpi 1.3?
> Stephen
>>
>> 2008/11/18 Ralph Castain <r...@lanl.gov>:
>>
>>>
>>> I am afraid it is only available in 1.3 - we didn't backport it to the
>>> 1.2
>>> series
>>>
>>>
>>> On Nov 18, 2008, at 10:06 AM, Gabriele Fatigati wrote:
>>>
>>>
>>>>
>>>> Hi,
>>>> how can i set "slot mapping" as you told me? With TASK GEOMETRY? Or is
>>>> a new 1.3 OpenMPI feature?
>>>>
>>>> Thanks.
>>>>
>>>> 2008/11/18 Ralph Castain <r...@lanl.gov>:
>>>>
>>>>>
>>>>> Unfortunately, paffinity doesn't know anything about assigning threads
>>>>> to
>>>>> cores. This is actually a behavior of Linux, which only allows
>>>>> paffinity
>>>>> to
>>>>> be set at the process level. So, when you set paffinity on a process,
>>>>> you
>>>>> bind all threads of that process to the specified core(s). You cannot
>>>>> specify that a thread be given a specific core.
>>>>>
>>>>> In this case, your two threads/process are sharing the same core and
>>>>> thus
>>>>> contending for it. As you'd expect in that situation, one thread gets
>>>>> the
>>>>> vast majority of the attention, while the other thread is mostly idle.
>>>>>
>>>>> If you can upgrade to the beta 1.3 release, try using the slot mapping
>>>>> to
>>>>> assign multiple cores to each process. This will ensure that the
>>>>> threads
>>>>> for
>>>>> that process have exclusive access to those cores, but will not bind a
>>>>> particular thread to one core - the threads can "move around" across
>>>>> the
>>>>> specified set of cores. Your threads will then be able to run without
>>>>> interfering with each other.
>>>>>
>>>>> Ralph
>>>>>
>>>>>
>>>>> On Nov 18, 2008, at 9:18 AM, Gabriele Fatigati wrote:
>>>>>
>>>>>
>>>>>>
>>>>>> Dear OpenMPI developers,
>>>>>> i have a strange problem with mixed program MPI+OPENMP over OpenMPI
>>>>>> 1.2.6. I'm using PJL TASK  GEOMETRY in LSF Scheduler, setting 2 MPI
>>>>>> process every compute node, and 2 OMP threads per process. Using
>>>>>> paffinity and maffinity, i've noted that over every node, i have 2
>>>>>> thread that works 100%, and 2 threads doesn't works, or works very
>>>>>> few.
>>>>>>
>>>>>> If i disable paffinity and maffinity, 4 threads works well, without
>>>>>> load imbalance.
>>>>>> I don't understand this issue: paffinity and maffinity should map
>>>>>> every thread over a specific core, optimizing the cache flow, but i
>>>>>> have this without settings there!
>>>>>>
>>>>>> Can i use paffinity and maffinity in mixed MPI+OpenMP program? Or it
>>>>>> works only over MPI thread?
>>>>>>
>>>>>> Thanks in advance.
>>>>>>
>>>>>>
>>>>>> --
>>>>>> Ing. Gabriele Fatigati
>>>>>>
>>>>>> CINECA Systems & Tecnologies Department
>>>>>>
>>>>>> Supercomputing  Group
>>>>>>
>>>>>> Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy
>>>>>>
>>>>>> www.cineca.it                    Tel:   +39 051 6171722
>>>>>>
>>>>>> g.fatig...@cineca.it
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> us...@open-mpi.org
>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>
>>>>>
>>>>>
>>>>
>>>> --
>>>> Ing. Gabriele Fatigati
>>>>
>>>> CINECA Systems & Tecnologies Department
>>>>
>>>> Supercomputing  Group
>>>>
>>>> Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy
>>>>
>>>> www.cineca.it                    Tel:   +39 051 6171722
>>>>
>>>> g.fatig...@cineca.it
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>
>>>
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>
>>>
>>>
>>
>>
>>
>>
>
>
> --
> stephen.wor...@sophia.inria.fr
> 2004 route des lucioles - BP93
> Sophia Antipolis
> 06902 CEDEX
>
> Tel: 04 92 38 50 54
> Fax: 04 97 15 53 51
>
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>



-- 
Ing. Gabriele Fatigati

CINECA Systems & Tecnologies Department

Supercomputing  Group

Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy

www.cineca.it                    Tel:   +39 051 6171722

g.fatig...@cineca.it

Reply via email to