Thanks Guozhang for the great questions. Answers are inlined:

1. I'm still not sure if it's worthwhile to add a new type of "learner
task" in addition to "standby task": if the only difference is that for the
latter, we would consider workload balance while for the former we would
not, I think we can just adjust the logic of StickyTaskAssignor a bit to
break that difference. Adding a new type of task would be adding a lot of
code complexity, so if we can still piggy-back the logic on a standby-task
I would prefer to do so.
In the proposal we stated that we are not adding a new type of task 
implementation. The
learner task shall share the same implementation with normal standby task, only 
that we
shall tag the standby task with learner and prioritize the learner tasks replay 
effort.
2. One thing that's still not clear from the KIP wiki itself is which layer
would the logic be implemented at. Although for most KIPs we would not
require internal implementation details but only public facing API updates,
for a KIP like this I think it still requires to flesh out details on the
implementation design. More specifically: today Streams embed a full
fledged Consumer client, which hard-code a ConsumerCoordinator inside,
Streams then injects a StreamsPartitionAssignor to its pluggable
PartitionAssignor interface and inside the StreamsPartitionAssignor we also
have a TaskAssignor interface whose default implementation is
StickyPartitionAssignor. Streams partition assignor logic today sites in
the latter two classes. Hence the hierarchy today is:

KafkaConsumer -> ConsumerCoordinator -> StreamsPartitionAssignor ->
StickyTaskAssignor.

We need to think about where the proposed implementation would take place
at, and personally I think it is not the best option to inject all of them
into the StreamsPartitionAssignor / StickyTaskAssignor since the logic of
"triggering another rebalance" etc would require some coordinator logic
which is hard to mimic at PartitionAssignor level. On the other hand, since
we are embedding a KafkaConsumer client as a whole we cannot just replace
ConsumerCoordinator with a specialized StreamsCoordinator like Connect does
in KIP-415. So I'd like to maybe split the current proposal in both
consumer layer and streams-assignor layer like we did in KIP-98/KIP-129.
And then the key thing to consider is how to cut off the boundary so that
the modifications we push to ConsumerCoordinator would be beneficial
universally for any consumers, while keep the Streams-specific logic at the
assignor level.
Yes, that's also my ideal plan. The details for the implementation are depicted
in this 
doc<https://docs.google.com/document/d/1me2a5wvxAZT1QE6HkwyDl7C2TiBQlKN3Dpw_I1ro91U/edit#heading=h.qix74qdmekae>,
 and I have explained the reasoning on why we want to push a
global change of replacing ConsumerCoordinator with StreamCoordinator. The 
motivation
is that KIP space is usually used for public & algorithm level change, not for 
internal
implementation details.

3. Depending on which design direction we choose, our migration plan would
also be quite different. For example, if we stay with ConsumerCoordinator
whose protocol type is "consumer" still, and we can manage to make all
changes agnostic to brokers as well as to old versioned consumers, then our
migration plan could be much easier.
Yes, the upgrade plan was designed to take the new StreamCoordinator approach
which means we shall define a new protocol type. For existing application we 
could only
maintain the same `consumer` protocol type is because current broker only allows
change of protocol type when the consumer group is empty. It is of course 
user-unfriendly to force
a wipe-out for the entire application, and I don't think maintaining old 
protocol type would greatly
impact ongoing services using new stream coordinator. WDYT?

4. I think one major issue related to this KIP is that today, in the
StickyPartitionAssignor, we always try to honor stickiness over workload
balance, and hence "learner task" is needed to break this priority, but I'm
wondering if we can have a better solution within sticky task assignor that
accommodate this?
Great question! That's what I explained in the proposal, which is that we 
should breakdown our
delivery into different stages. At very beginning, our goal is to trigger 
learner task assignment only on
`new` hosts, where we shall leverage leader's knowledge of previous round of 
rebalance to figure out. After
stage one, our goal is to have a smooth scaling up experience, but the task 
balance problem is kind of orthogonal.
The load balance problem is a much broader topic than auto scaling, which I 
figure worth discussing within
this KIP's context since it's a naturally next-step, but wouldn't be the main 
topic.
Learner task or auto scaling support should be treated as `a helpful mechanism 
to reach load balance`, but not `an algorithm defining load balance`. It would 
be great if you could share some insights of the stream task balance, which 
eventually helps us to break out of the KIP-429's scope and even define a 
separate KIP to focus on task weight & assignment logic improvement.

Also thank you for making improvement on the KIP context and organization!

Best,
Boyang
________________________________
From: Guozhang Wang <wangg...@gmail.com>
Sent: Saturday, March 2, 2019 6:00 AM
To: dev
Subject: Re: [DISCUSS] KIP-429 : Smooth Auto-Scaling for Kafka Streams

Hello Boyang,

I've just made a quick pass on the KIP and here are some thoughts.

Meta:

1. I'm still not sure if it's worthwhile to add a new type of "learner
task" in addition to "standby task": if the only difference is that for the
latter, we would consider workload balance while for the former we would
not, I think we can just adjust the logic of StickyTaskAssignor a bit to
break that difference. Adding a new type of task would be adding a lot of
code complexity, so if we can still piggy-back the logic on a standby-task
I would prefer to do so.

2. One thing that's still not clear from the KIP wiki itself is which layer
would the logic be implemented at. Although for most KIPs we would not
require internal implementation details but only public facing API updates,
for a KIP like this I think it still requires to flesh out details on the
implementation design. More specifically: today Streams embed a full
fledged Consumer client, which hard-code a ConsumerCoordinator inside,
Streams then injects a StreamsPartitionAssignor to its plugable
PartitionAssignor interface and inside the StreamsPartitionAssignor we also
have a TaskAssignor interface whose default implementation is
StickyPartitionAssignor. Streams partition assignor logic today sites in
the latter two classes. Hence the hierarchy today is:

KafkaConsumer -> ConsumerCoordinator -> StreamsPartitionAssignor ->
StickyTaskAssignor.

We need to think about where the proposed implementation would take place
at, and personally I think it is not the best option to inject all of them
into the StreamsPartitionAssignor / StickyTaskAssignor since the logic of
"triggering another rebalance" etc would require some coordinator logic
which is hard to mimic at PartitionAssignor level. On the other hand, since
we are embedding a KafkaConsumer client as a whole we cannot just replace
ConsumerCoordinator with a specialized StreamsCoordinator like Connect does
in KIP-415. So I'd like to maybe split the current proposal in both
consumer layer and streams-assignor layer like we did in KIP-98/KIP-129.
And then the key thing to consider is how to cut off the boundary so that
the modifications we push to ConsumerCoordinator would be beneficial
universally for any consumers, while keep the Streams-specific logic at the
assignor level.

3. Depending on which design direction we choose, our migration plan would
also be quite different. For example, if we stay with ConsumerCoordinator
whose protocol type is "consumer" still, and we can manage to make all
changes agnostic to brokers as well as to old versioned consumers, then our
migration plan could be much easier.

4. I think one major issue related to this KIP is that today, in the
StickyPartitionAssignor, we always try to honor stickiness over workload
balance, and hence "learner task" is needed to break this priority, but I'm
wondering if we can have a better solution within sticky task assignor that
accommodate this?

Minor:

1. The idea of two rebalances have also been discussed in
https://issues.apache.org/jira/browse/KAFKA-6145. So we should add the
reference on the wiki page as well.
2. Could you also add a section describing how the subscription /
assignment metadata will be re-formatted? Without this information it is
hard to get to the bottom of your idea. For example in the "Leader Transfer
Before Scaling" section, I'm not sure why "S2 doesn't know S4 is new member"
and hence would blindly obey stickiness over workload balance requirement.

Guozhang


On Thu, Feb 28, 2019 at 11:05 AM Boyang Chen <bche...@outlook.com> wrote:

> Hey community friends,
>
> I'm gladly inviting you to have a look at the proposal to add incremental
> rebalancing to Kafka Streams, A.K.A auto-scaling support.
>
>
> https://cwiki.apache.org/confluence/display/KAFKA/KIP-429%3A+Smooth+Auto-Scaling+for+Kafka+Streams
>
> Special thanks to Guozhang for giving great guidances and important
> feedbacks while making this KIP!
>
> Best,
> Boyang
>


--
-- Guozhang

Reply via email to