Cool - I was just proposing that we do this from within hwloc instead of 
implementing it separately everywhere.

On Nov 2, 2012, at 2:54 PM, Rayson Ho <raysonlo...@gmail.com> wrote:

> Ralph,
> 
> We added cgroups integration support into Grid Engine a few months
> ago, and we ended up implementing routines that write values to
> "memory.memsw.limit_in_bytes", "memory.limit_in_bytes",
> "memory.soft_limit_in_bytes", etc... We just simply write the values
> out to the cgroups files.
> 
> http://blogs.scalablelogic.com/2012/05/grid-engine-cgroups-integration.html
> 
> I am interested in seeing how Greenplum/EMC implements cgroups limits.

I think my other note explained this - hopefully it helped :-)

> 
> 
> Brice - cgroups allow system administrators to set the resource limit
> of a process (or a group of processes) by interacting with the cgroups
> virtual fs. Once the processes are added to a cgroup, the memory usage
> limit can be added or changed as simply as just just by doing:
> 
> # echo 4M > /sys/fs/cgroup/memory/0/memory.limit_in_bytes
> 
> For details, see "3. User Interface":
> http://www.kernel.org/doc/Documentation/cgroups/memory.txt
> 
> Rayson
> 
> ==================================================
> Open Grid Scheduler - The Official Open Source Grid Engine
> http://gridscheduler.sourceforge.net/
> 
> 
> On Fri, Nov 2, 2012 at 5:18 PM, Brice Goglin <brice.gog...@inria.fr> wrote:
>> Hello Ralph,
>> 
>> I am not very familiar with these features. What system mechanism do you
>> currently use for this? Linux cgroups? Any concrete example of what you
>> would like to do?
>> 
>> Brice
>> 
>> 
>> 
>> Le 02/11/2012 22:12, Ralph Castain a écrit :
>>> Hi folks
>>> 
>>> We (Greenplum) have a need to support resource limits (e.g., memory and cpu 
>>> usage) on processes running under Open MPI's RTE. OMPI uses hwloc for 
>>> processor and memory affinity, so this seems a likely place to add the 
>>> required support. Jeff tells me that it doesn't yet exist in hwloc - I'm 
>>> wondering if you would welcome and/or be willing to consider contributions 
>>> from our engineers towards adding this capability?
>>> 
>>> Obviously, we'd need to discuss how and where to do the extension. Just 
>>> wanted to first see if this is an option, or if we should do it directly in 
>>> OMPI.
>>> Ralph
>>> 
>>> 
>>> _______________________________________________
>>> hwloc-devel mailing list
>>> hwloc-de...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/hwloc-devel
>> 
>> _______________________________________________
>> hwloc-devel mailing list
>> hwloc-de...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/hwloc-devel
> 
> _______________________________________________
> hwloc-devel mailing list
> hwloc-de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/hwloc-devel


Reply via email to