Hi Jonah,

Thanks for the KIP.
For now I have one question.

For  *internal.max.size.bytes, *you mentioned removing this internal
config. Can you clarify this config and how it is used? I couldn't find it
in the code.

Justine



On Tue, Oct 7, 2025 at 11:07 AM Jun Rao <[email protected]> wrote:

> Hi, Jonah,
>
> Thanks for the KIP. Just a couple of minor comments.
>
> controller.quorum.fetch.timeout.ms defaults to 2 secs. Is that a bit too
> low?
>
> Orthogonally, in the case when the network latency is long, one can
> typically tune socket.receive.buffer.bytes to improve the throughput.
>
> Jun
>
>
>
> On Fri, Sep 26, 2025 at 10:29 AM José Armando García Sancio
> <[email protected]> wrote:
>
> > Thanks for the changes Jonah. Looking forward to the implementation of
> > this feature.
> >
> > On Fri, Sep 26, 2025 at 12:33 PM Jonah Hooper
> > <[email protected]> wrote:
> > >
> > > Thanks for the feedback José! Did some work to apply the KIP to a
> correct
> > > format.
> > >
> > > The motivation section starts with the solution without motivating the
> > > > problem. I think you want to delete the first paragraph in the
> > > > motivation section.
> > >
> > >
> > > Done.
> > >
> > > In the motivation section you have "non-terminating loop where they
> > > > cannot join the cluster." They are technically not joining the
> > > > cluster. The issue is a liveness issue. Because the FETCH_SNAPSHOT
> RPC
> > > > doesn't complete within the fetch timeout configuration the fetching
> > > > replica cannot make progress. In the worst case, if it is a
> > > > voter/controller, it further disrupts the leader by becoming a
> > > > candidate and increasing the epoch.
> > >
> > >
> > > Updated the motivation to reflect this.
> > >
> > >
> > >
> > > >  In the public interface section, please make it clear that the
> changes
> > > > are the addition of two new configurations. Let's also add
> > descriptions for
> > > > these configurations.
> > >
> > >
> > > Added descriptions.
> > >
> > > Please fill out "Compatibility, Deprecation, and Migration Plan". Why
> > > > is this safe and backward compatible?
> > > >
> > >
> > > Done.
> > >
> > > In the test section you have "We can perform once off tests which
> > > > evaluate the effect of degraded networking on snapshot. It's better
> to
> > > > design system-tests which focus on correctness by using
> pathologically
> > > > low values for controller.quorum.fetch.snapshot.max.bytes and
> > > > controller.quorum.fetch.max.bytes. An example scenario would be a
> > > > controller joining a quorum with a snapshot with a known size and
> then
> > > > attempting to fetch the snapshot in small (say 64kb) or even
> > > > pathologically small values like 1 byte." I don't fully understand
> > > > this recommendation but if my understanding is correct this sounds
> > > > very complicated and flakey to implement. Can't we just add protocol
> > > > tests to KafkaRaftClient*Test that show these configurations being
> > > > used when handling FETC and FETCH_SNAPSHOT requests?
> > >
> > >
> > > Good point; that section is too specific. I've left it a bit more
> vague,
> > > unit tests are sufficient provided we can "mock" a situation where we
> > > retrieve less data than is available.
> > >
> > > Best,
> > > Jonah
> > >
> > > On Mon, Sep 22, 2025 at 12:31 PM José Armando García Sancio
> > > <[email protected]> wrote:
> > >
> > > > Hi Jonah,
> > > >
> > > > Thanks for the changes. The proposed solution looks good to me at a
> > > > high-level. I just have some minor comments.
> > > >
> > > > The motivation section starts with the solution without motivating
> the
> > > > problem. I think you want to delete the first paragraph in the
> > > > motivation section.
> > > >
> > > > In the motivation section you have "non-terminating loop where they
> > > > cannot join the cluster." They are technically not joining the
> > > > cluster. The issue is a liveness issue. Because the FETCH_SNAPSHOT
> RPC
> > > > doesn't complete within the fetch timeout configuration the fetching
> > > > replica cannot make progress. In the worst case, if it is a
> > > > voter/controller, it further disrupts the leader by becoming a
> > > > candidate and increasing the epoch.
> > > >
> > > > In the motivation section you have "The proposed default
> > > > configurations tradeoff a reduction in the throughput of data
> > > > transmitted by Fetch and FetchSnapshot requests with a liveness and
> > > > correctness improvement." This is also the solution and not part of
> > > > the motivation.
> > > >
> > > > In the public interface section, please make it clear that the
> changes
> > > > are the addition of two new configurations. Let's also add
> > > > descriptions for these configurations.
> > > >
> > > > Please fill out "Compatibility, Deprecation, and Migration Plan". Why
> > > > is this safe and backward compatible?
> > > >
> > > > In the test section you have "We can perform once off tests which
> > > > evaluate the effect of degraded networking on snapshot. It's better
> to
> > > > design system-tests which focus on correctness by using
> pathologically
> > > > low values for controller.quorum.fetch.snapshot.max.bytes and
> > > > controller.quorum.fetch.max.bytes. An example scenario would be a
> > > > controller joining a quorum with a snapshot with a known size and
> then
> > > > attempting to fetch the snapshot in small (say 64kb) or even
> > > > pathologically small values like 1 byte." I don't fully understand
> > > > this recommendation but if my understanding is correct this sounds
> > > > very complicated and flakey to implement. Can't we just add protocol
> > > > tests to KafkaRaftClient*Test that show these configurations being
> > > > used when handling FETC and FETCH_SNAPSHOT requests?
> > > >
> > > > I didn't comment on the proposed changes section. I'll comment after
> > > > the comments above are addressed.
> > > >
> > > > Thanks
> > > > --
> > > > -José
> > > >
> >
> >
> >
> > --
> > -José
> >
>

Reply via email to