Re: CPU affinity

2017-04-06 Thread Jie Yu
Hi Dmitry,

Thank you for sending this out for comments! I'd love to participate in the
discussion!

Are you available for some f2f meetings (or remote hangout) for this?
Thought it'll be more efficient if we can discuss that way than over
emails. We can invite others in the community that are interested in this
into the meeting.

Also, I'd suggest you do some related work research (e.g., how k8s or
Docker solves this problem). It'll always be useful to learn from what
other people are doing.

Let me know!
- Jie

On Thu, Apr 6, 2017 at 3:23 AM, Dmitry Zhuk  wrote:

> Hi Vikram,
>
> Thank you for the reply.
> I understand that hardware information is required for frameworks to make a
> decision in some cases, but for the case I'm interested in, hardware
> information is not that much important. In particular, I have a memory
> bound task, which achieves 10-20% performance gain if it's pinned to a
> single NUMA node due to eliminated foreign memory access. So basically I
> need to be able to specify that task needs 4 CPUs, all on one NUMA node.
> This needs some basic knowledge about hardware, like number of CPUs per
> node. So to me it makes sense to start with affinity support design, and
> then make sure that hardware topology information provides enough details
> for frameworks to specify affinity constraints.
>
> ---
> Dmitry
>
> On Thu, Mar 23, 2017 at 12:28 AM, Vikrama Ditya  wrote:
>
> > Hi Dmitry
> >
> > This problem needs to be addressed with topology information so that
> > scheduler framework can utilize it and request affinity constraints.
> >
> > We started to look into this when we are required to expose GPU HW
> > information. It would be good to introduce generic topology structure so
> > that generic interconnects and associated resource topology can be
> > expressed.
> >
> > Please have a look at https://issues.apache.org/jira/browse/MESOS-7080
> >
> > --
> > Vikram
> >
> > -Original Message-
> > From: Dmitry Zhuk [mailto:dz...@twopensource.com]
> > Sent: Wednesday, March 22, 2017 6:49 AM
> > To: dev@mesos.apache.org
> > Subject: CPU affinity
> >
> > Hi
> >
> > Is anyone working on MESOS-314
> >  “Support the cgroups
> > 'cpusets' subsystem” or related functionality? I found other related
> > tickets in JIRA, but there seems to be no recent progress on them:
> > MESOS-5342 ,
> MESOS-5358
> > . There’s also a
> mention
> > of idea of exposing cpusets similar to network ports.
> >
> > I’d like propose an alternative approach for adding CPU affinity support
> > and would be interested in any feedback on it. If the community is
> > interested in this approach, I can work on design document and
> > implementation.
> >
> > The basic idea is to let frameworks specify affinity requirements in
> > ContainerInfo using the following structure:
> > message AffinityInfo {
> >   enum ProcessingUnit {
> > THREAD = 1;
> > CORE = 2;
> > SOCKET = 3;
> > NUMA_NODE = 4;
> >   }
> >
> >   // Indicates that container should be bound to the units of specified
> > type.
> >   // For example: bind = NUMA_NODE indicates, that process
> >   // can run on any thread from some NUMA node.
> >   required ProcessingUnit bind = 1;
> >
> >   // Indicates that assigned processing units must not be shared with
> >   // other containers.
> >   optional bool exclusive = 2 [default = false];
> > }
> >
> >
> > message ContainerInfo {
> >   …
> >   optional AffinityInfo affinity_info = …;
> > }
> >
> > In future this can be extended to require exclusive NUMA node memory
> > access, proximity to devices, etc.
> > This also requires exposing hardware topology information (such as number
> > of cpus per node) to frameworks to evaluate offer suitability, and
> > providing visibility to frameworks on failures to assign CPUs per
> > requirements, but this can be left out of scope of the MVP.
> >
> > Thanks
> >
> > 
> > ---
> > This email message is for the sole use of the intended recipient(s) and
> > may contain
> > confidential information.  Any unauthorized review, use, disclosure or
> > distribution
> > is prohibited.  If you are not the intended recipient, please contact the
> > sender by
> > reply email and destroy all copies of the original message.
> > 
> > ---
> >
>


Mesos Developer Community Meeting (April 6, 2017)

2017-04-06 Thread Michael Park
We're meeting today at 3pm PST!

If you'd like to join in person, please come to the Mesosphere HQ
at 88 Stevenson St., and check in through the reception on the 2nd floor.

Neil Conway will be giving a tech talk about hierarchical roles in Mesos!

Please add topics you want to discuss to the meeting log

if
you have anything.

Thanks,

MPark


Re: CPU affinity

2017-04-06 Thread Dmitry Zhuk
Hi Vikram,

Thank you for the reply.
I understand that hardware information is required for frameworks to make a
decision in some cases, but for the case I'm interested in, hardware
information is not that much important. In particular, I have a memory
bound task, which achieves 10-20% performance gain if it's pinned to a
single NUMA node due to eliminated foreign memory access. So basically I
need to be able to specify that task needs 4 CPUs, all on one NUMA node.
This needs some basic knowledge about hardware, like number of CPUs per
node. So to me it makes sense to start with affinity support design, and
then make sure that hardware topology information provides enough details
for frameworks to specify affinity constraints.

---
Dmitry

On Thu, Mar 23, 2017 at 12:28 AM, Vikrama Ditya  wrote:

> Hi Dmitry
>
> This problem needs to be addressed with topology information so that
> scheduler framework can utilize it and request affinity constraints.
>
> We started to look into this when we are required to expose GPU HW
> information. It would be good to introduce generic topology structure so
> that generic interconnects and associated resource topology can be
> expressed.
>
> Please have a look at https://issues.apache.org/jira/browse/MESOS-7080
>
> --
> Vikram
>
> -Original Message-
> From: Dmitry Zhuk [mailto:dz...@twopensource.com]
> Sent: Wednesday, March 22, 2017 6:49 AM
> To: dev@mesos.apache.org
> Subject: CPU affinity
>
> Hi
>
> Is anyone working on MESOS-314
>  “Support the cgroups
> 'cpusets' subsystem” or related functionality? I found other related
> tickets in JIRA, but there seems to be no recent progress on them:
> MESOS-5342 , MESOS-5358
> . There’s also a mention
> of idea of exposing cpusets similar to network ports.
>
> I’d like propose an alternative approach for adding CPU affinity support
> and would be interested in any feedback on it. If the community is
> interested in this approach, I can work on design document and
> implementation.
>
> The basic idea is to let frameworks specify affinity requirements in
> ContainerInfo using the following structure:
> message AffinityInfo {
>   enum ProcessingUnit {
> THREAD = 1;
> CORE = 2;
> SOCKET = 3;
> NUMA_NODE = 4;
>   }
>
>   // Indicates that container should be bound to the units of specified
> type.
>   // For example: bind = NUMA_NODE indicates, that process
>   // can run on any thread from some NUMA node.
>   required ProcessingUnit bind = 1;
>
>   // Indicates that assigned processing units must not be shared with
>   // other containers.
>   optional bool exclusive = 2 [default = false];
> }
>
>
> message ContainerInfo {
>   …
>   optional AffinityInfo affinity_info = …;
> }
>
> In future this can be extended to require exclusive NUMA node memory
> access, proximity to devices, etc.
> This also requires exposing hardware topology information (such as number
> of cpus per node) to frameworks to evaluate offer suitability, and
> providing visibility to frameworks on failures to assign CPUs per
> requirements, but this can be left out of scope of the MVP.
>
> Thanks
>
> 
> ---
> This email message is for the sole use of the intended recipient(s) and
> may contain
> confidential information.  Any unauthorized review, use, disclosure or
> distribution
> is prohibited.  If you are not the intended recipient, please contact the
> sender by
> reply email and destroy all copies of the original message.
> 
> ---
>