In that case, why not just directly launch the processes without the orted? We 
do it with slurm and even have the ability to do it with torque - so it could 
be done.

See the orte/mca/ess/slurmd component for an example of how to do so.


On May 4, 2011, at 4:55 PM, Tony Lam wrote:

> Hi Thomas,
> 
> We need to track job resource usage in our resource manager for
> accounting and resource policy enforcement, sharing single orted
> process in multiple jobs makes the tracking much complicated. We don't
> enforce other restrictions, and I'll appreciate any suggestion on how
> to resolve this or work around this.
> 
> We have thought about mapping all processes from a mpirun into a
> single job to simplify job resource tracking, but that will require much 
> spread changes in our software.
> 
> Thanks.
> 
> Tony
> 
> 
> On 05/04/11 15:34, Thomas Herault wrote:
>> Hi,
>> 
>> Could you explain why you would like one orted on top of each MPI process?
>> There are some situations, like resource usage limitation / accounting, that 
>> are possible to solve without changing the one daemon per node deployment.
>> Or do you enforce other kinds of restrictions on the orted process? Why 
>> wouldn't it be able to launch more than one MPI process / why would not that 
>> be desirable?
>> 
>> Bests,
>> Thomas
>> 
>> Le 4 mai 2011 à 15:51, Tony Lam a écrit :
>> 
>>> Hi,
>>> 
>>> I understand a single orted is shared by all MPI processes from the same 
>>> communicator on each execution host, does anyone see any problem that 
>>> MPI/OMPI may have problem with each process has its owner orted? My guess 
>>> it is less efficient in terms of MPI communication and memory foot print, 
>>> but for simplification of our integration with OMPI, launching one orted 
>>> for each MPI process is much easier to do.
>>> 
>>> I will appreciate if someone can confirm this setup will or will not work.
>>> 
>>> Thanks.
>>> 
>>> Tony
>>> 
>>> _______________________________________________
>>> devel mailing list
>>> de...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>> 
>> 
>> _______________________________________________
>> devel mailing list
>> de...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
> _______________________________________________
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel


Reply via email to