[
https://issues.apache.org/jira/browse/CAMEL-16181?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Javier Holguera updated CAMEL-16181:
------------------------------------
Description:
The `KafkaIdempotentRepository` initialises its cache off the back of the
pre-existing Kafka topic with previous entries, with the following code:
{code:java}
log.debug("Subscribing consumer to {}", topic);
consumer.subscribe(Collections.singleton(topic));
log.debug("Seeking to beginning");
consumer.seekToBeginning(consumer.assignment());
POLL_LOOP: while (running.get()) {
log.trace("Polling");
ConsumerRecords<String, String> consumerRecords =
consumer.poll(pollDurationMs);
if (consumerRecords.isEmpty()) {
// the first time this happens, we can assume that we have consumed all
messages up to this point
log.trace("0 messages fetched on poll");
if (cacheReadyLatch.getCount() > 0) {
log.debug("Cache warmed up");
cacheReadyLatch.countDown();
}
}{code}
The problem with this code is:
# `consumer.subscribe` doesn't instantaneously assign partitions to the
consumer
# When `consumer.seekToBeginning` is called, the operation doesn't do anything
because it has no partitions yet (see [seekToBeginning doesn't work without
auto.offset.reset
(apache.org)|https://mail-archives.apache.org/mod_mbox/kafka-users/201603.mbox/%3ccakwx9vumpliqtu9o0mpepaupszapw9lm91mwexvafwktgd3...@mail.gmail.com%3e]
# When later the first `consumer.poll` is issued, it returns nothing,
triggering the sequence to *confirm the cache as ready when it isn't yet*. That
can cause upstream messages not been correctly de-duplicated.
The solution is:
# Use a different overload of `consumer.subscribe` that accepts an
implementation of the `ConsumerRebalanceListener`.
# When partitions are assigned to the `consumer` instance, call
`seekToBeginning` there.
# Doing an initial `poll(0)` that will never return records but will force the
partition assignment process
was:
The `KafkaIdempotentRepository` initialises its cache off the back of the
pre-existing Kafka topic with previous entries, with the following code:
{code:java}
log.debug("Subscribing consumer to {}", topic);
consumer.subscribe(Collections.singleton(topic));
log.debug("Seeking to beginning");
consumer.seekToBeginning(consumer.assignment());
POLL_LOOP: while (running.get()) {
log.trace("Polling");
ConsumerRecords<String, String> consumerRecords =
consumer.poll(pollDurationMs);
if (consumerRecords.isEmpty()) {
// the first time this happens, we can assume that we have consumed all
messages up to this point
log.trace("0 messages fetched on poll");
if (cacheReadyLatch.getCount() > 0) {
log.debug("Cache warmed up");
cacheReadyLatch.countDown();
}
}{code}
\{code} The problem with this code is: # `consumer.subscribe`
doesn't instantaneously assign partitions to the consumer # When
`consumer.seekToBeginning` is called, the operation doesn't do anything because
it has no partitions yet (see [seekToBeginning doesn't work without
auto.offset.reset
(apache.org)|https://mail-archives.apache.org/mod_mbox/kafka-users/201603.mbox/%3ccakwx9vumpliqtu9o0mpepaupszapw9lm91mwexvafwktgd3...@mail.gmail.com%3e]
# When later the first `consumer.poll` is issued, it returns nothing,
triggering the sequence to *confirm the cache as ready when it isn't yet*. That
can cause upstream messages not been correctly de-duplicated. The solution
is: # Use a different overload of `consumer.subscribe` that accepts an
implementation of the `ConsumerRebalanceListener`. When partitions are assigned
to the `consumer` instance, call `seekToBeginning` there. # Doing an initial
`poll(0)` that will never return records but will force the partition
assignment process
> KafkaIdempotentRepository cache incorrectly flagged as ready
> ------------------------------------------------------------
>
> Key: CAMEL-16181
> URL: https://issues.apache.org/jira/browse/CAMEL-16181
> Project: Camel
> Issue Type: Improvement
> Components: camel-kafka
> Affects Versions: 3.7.2
> Reporter: Javier Holguera
> Priority: Major
> Fix For: 3.8.0
>
>
> The `KafkaIdempotentRepository` initialises its cache off the back of the
> pre-existing Kafka topic with previous entries, with the following code:
>
> {code:java}
> log.debug("Subscribing consumer to {}", topic);
> consumer.subscribe(Collections.singleton(topic));
> log.debug("Seeking to beginning");
> consumer.seekToBeginning(consumer.assignment());
>
> POLL_LOOP: while (running.get()) {
> log.trace("Polling");
> ConsumerRecords<String, String> consumerRecords =
> consumer.poll(pollDurationMs);
> if (consumerRecords.isEmpty()) {
> // the first time this happens, we can assume that we have consumed all
> messages up to this point
> log.trace("0 messages fetched on poll");
> if (cacheReadyLatch.getCount() > 0) {
> log.debug("Cache warmed up");
> cacheReadyLatch.countDown();
> }
> }{code}
>
> The problem with this code is:
> # `consumer.subscribe` doesn't instantaneously assign partitions to the
> consumer
> # When `consumer.seekToBeginning` is called, the operation doesn't do
> anything because it has no partitions yet (see [seekToBeginning doesn't work
> without auto.offset.reset
> (apache.org)|https://mail-archives.apache.org/mod_mbox/kafka-users/201603.mbox/%3ccakwx9vumpliqtu9o0mpepaupszapw9lm91mwexvafwktgd3...@mail.gmail.com%3e]
>
> # When later the first `consumer.poll` is issued, it returns nothing,
> triggering the sequence to *confirm the cache as ready when it isn't yet*.
> That can cause upstream messages not been correctly de-duplicated.
> The solution is:
> # Use a different overload of `consumer.subscribe` that accepts an
> implementation of the `ConsumerRebalanceListener`.
> # When partitions are assigned to the `consumer` instance, call
> `seekToBeginning` there.
> # Doing an initial `poll(0)` that will never return records but will force
> the partition assignment process
--
This message was sent by Atlassian Jira
(v8.3.4#803005)