And in the hybrid program MPi+OpenMP?
Are these considerations still good?

2008/11/20 Edgar Gabriel <gabr...@cs.uh.edu>:
> I don't think that they conflict with our paffinity module and setting. My
> understanding is that if you set a new affinity mask, it simply overwrites
> the previous setting. So in the worst case it voids the setting made by Open
> MPI, but I don't think that it should cause 'problems'. Admittedly, I
> haven't tried the library and the function calls yet, I just learned
> relatively recently about them...
>
> Thanks
> Edga
>
> Ralph Castain wrote:
>>
>> Interesting - learn something new every day! :-)
>>
>> How does this interact with OMPI's paffinity/maffinity assignments? With
>> the rank/slot mapping and binding system?
>>
>> Should users -not- set paffinity if they include these numa calls in their
>> code?
>>
>> Can we detect any potential conflict in OMPI and avoid setting
>> paffinity_alone? Reason I ask: many systems set paffinity_alone in the
>> default mca param file because they always assign dedicated nodes to users.
>> While users can be told to be sure to turn it "off" when using these calls,
>> it seems inevitable that they will forget - and complaints will appear.
>>
>> Thanks
>> Ralph
>>
>>
>>
>> On Nov 20, 2008, at 7:34 AM, Edgar Gabriel wrote:
>>
>>> if you look at recent versions of libnuma, there are two functions called
>>> numa_run_on_node() and numa_run_on_node_mask(), which allow thread-based
>>> assignments to CPUs....
>>>
>>> Thanks
>>> Edgar
>>>
>>> Gabriele Fatigati wrote:
>>>>
>>>> Is there a way to assign one thread to one core? Also from code, not
>>>> necessary with OpenMPI option.
>>>> Thanks.
>>>> 2008/11/19 Stephen Wornom <stephen.wor...@sophia.inria.fr>:
>>>>>
>>>>> Gabriele Fatigati wrote:
>>>>>>
>>>>>> Ok,
>>>>>> but in Ompi 1.3 how can i enable it?
>>>>>>
>>>>> This may not be relevant, but I could not get a hybrid mpi+OpenMP code
>>>>> to
>>>>> work correctly.
>>>>> Would my problem be related to Gabriele's and perhaps fixed in openmpi
>>>>> 1.3?
>>>>> Stephen
>>>>>>
>>>>>> 2008/11/18 Ralph Castain <r...@lanl.gov>:
>>>>>>
>>>>>>> I am afraid it is only available in 1.3 - we didn't backport it to
>>>>>>> the
>>>>>>> 1.2
>>>>>>> series
>>>>>>>
>>>>>>>
>>>>>>> On Nov 18, 2008, at 10:06 AM, Gabriele Fatigati wrote:
>>>>>>>
>>>>>>>
>>>>>>>> Hi,
>>>>>>>> how can i set "slot mapping" as you told me? With TASK GEOMETRY? Or
>>>>>>>> is
>>>>>>>> a new 1.3 OpenMPI feature?
>>>>>>>>
>>>>>>>> Thanks.
>>>>>>>>
>>>>>>>> 2008/11/18 Ralph Castain <r...@lanl.gov>:
>>>>>>>>
>>>>>>>>> Unfortunately, paffinity doesn't know anything about assigning
>>>>>>>>> threads
>>>>>>>>> to
>>>>>>>>> cores. This is actually a behavior of Linux, which only allows
>>>>>>>>> paffinity
>>>>>>>>> to
>>>>>>>>> be set at the process level. So, when you set paffinity on a
>>>>>>>>> process,
>>>>>>>>> you
>>>>>>>>> bind all threads of that process to the specified core(s). You
>>>>>>>>> cannot
>>>>>>>>> specify that a thread be given a specific core.
>>>>>>>>>
>>>>>>>>> In this case, your two threads/process are sharing the same core
>>>>>>>>> and
>>>>>>>>> thus
>>>>>>>>> contending for it. As you'd expect in that situation, one thread
>>>>>>>>> gets
>>>>>>>>> the
>>>>>>>>> vast majority of the attention, while the other thread is mostly
>>>>>>>>> idle.
>>>>>>>>>
>>>>>>>>> If you can upgrade to the beta 1.3 release, try using the slot
>>>>>>>>> mapping
>>>>>>>>> to
>>>>>>>>> assign multiple cores to each process. This will ensure that the
>>>>>>>>> threads
>>>>>>>>> for
>>>>>>>>> that process have exclusive access to those cores, but will not
>>>>>>>>> bind a
>>>>>>>>> particular thread to one core - the threads can "move around"
>>>>>>>>> across
>>>>>>>>> the
>>>>>>>>> specified set of cores. Your threads will then be able to run
>>>>>>>>> without
>>>>>>>>> interfering with each other.
>>>>>>>>>
>>>>>>>>> Ralph
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Nov 18, 2008, at 9:18 AM, Gabriele Fatigati wrote:
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>> Dear OpenMPI developers,
>>>>>>>>>> i have a strange problem with mixed program MPI+OPENMP over
>>>>>>>>>> OpenMPI
>>>>>>>>>> 1.2.6. I'm using PJL TASK  GEOMETRY in LSF Scheduler, setting 2
>>>>>>>>>> MPI
>>>>>>>>>> process every compute node, and 2 OMP threads per process. Using
>>>>>>>>>> paffinity and maffinity, i've noted that over every node, i have 2
>>>>>>>>>> thread that works 100%, and 2 threads doesn't works, or works very
>>>>>>>>>> few.
>>>>>>>>>>
>>>>>>>>>> If i disable paffinity and maffinity, 4 threads works well,
>>>>>>>>>> without
>>>>>>>>>> load imbalance.
>>>>>>>>>> I don't understand this issue: paffinity and maffinity should map
>>>>>>>>>> every thread over a specific core, optimizing the cache flow, but
>>>>>>>>>> i
>>>>>>>>>> have this without settings there!
>>>>>>>>>>
>>>>>>>>>> Can i use paffinity and maffinity in mixed MPI+OpenMP program? Or
>>>>>>>>>> it
>>>>>>>>>> works only over MPI thread?
>>>>>>>>>>
>>>>>>>>>> Thanks in advance.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> Ing. Gabriele Fatigati
>>>>>>>>>>
>>>>>>>>>> CINECA Systems & Tecnologies Department
>>>>>>>>>>
>>>>>>>>>> Supercomputing  Group
>>>>>>>>>>
>>>>>>>>>> Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy
>>>>>>>>>>
>>>>>>>>>> www.cineca.it                    Tel:   +39 051 6171722
>>>>>>>>>>
>>>>>>>>>> g.fatig...@cineca.it
>>>>>>>>>> _______________________________________________
>>>>>>>>>> users mailing list
>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>>
>>>>>>>>> _______________________________________________
>>>>>>>>> users mailing list
>>>>>>>>> us...@open-mpi.org
>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>> --
>>>>>>>> Ing. Gabriele Fatigati
>>>>>>>>
>>>>>>>> CINECA Systems & Tecnologies Department
>>>>>>>>
>>>>>>>> Supercomputing  Group
>>>>>>>>
>>>>>>>> Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy
>>>>>>>>
>>>>>>>> www.cineca.it                    Tel:   +39 051 6171722
>>>>>>>>
>>>>>>>> g.fatig...@cineca.it
>>>>>>>> _______________________________________________
>>>>>>>> users mailing list
>>>>>>>> us...@open-mpi.org
>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> users mailing list
>>>>>>> us...@open-mpi.org
>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>> --
>>>>> stephen.wor...@sophia.inria.fr
>>>>> 2004 route des lucioles - BP93
>>>>> Sophia Antipolis
>>>>> 06902 CEDEX
>>>>>
>>>>> Tel: 04 92 38 50 54
>>>>> Fax: 04 97 15 53 51
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
> --
> Edgar Gabriel
> Assistant Professor
> Parallel Software Technologies Lab      http://pstl.cs.uh.edu
> Department of Computer Science          University of Houston
> Philip G. Hoffman Hall, Room 524        Houston, TX-77204, USA
> Tel: +1 (713) 743-3857                  Fax: +1 (713) 743-3335
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>



-- 
Ing. Gabriele Fatigati

CINECA Systems & Tecnologies Department

Supercomputing  Group

Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy

www.cineca.it                    Tel:   +39 051 6171722

g.fatig...@cineca.it

Reply via email to