Are you seeing 1) duplicate messages stored in a Kafka topic partition or 2) 
duplicate consumption and processing of a single message stored in a Kafka 
topic?

If it’s #1 then you can turn on the idempotent producer feature to get Exactly 
Once Semantics (EOS) while publishing.

If it’s #2 then you can examine more closely how your consumer is doing offset 
commits.
If you are committing offsets automatically by time then there is always a 
possibility that the last time window of messages your consumer did not yet 
commit will be received again when the consumer restarts. 

You can instead manually commit, possibly even after each message which will 
shrink the window of possible duplicate messages to 1, but at the cost of some 
performance. 

What many of the Kafka Sink Connectors do for exactly once processing is to 
store their offsets atomically with the data they write external to Kafka. For 
example a database connector would write the message data and the offsets to a 
database in one atomic write operation. Upon restart of the app it then rereads 
the offset from the database and resumes consumption from Kafka from the last 
offset point using seek() to reposition the Kafka offset for the consumer 
before the first call to poll()

These are the techniques most people use to get end to end exactly once 
processing with no duplicates even in the event of a failure.


-hans

> On May 28, 2018, at 12:17 AM, Karthick Kumar <kku...@apptivo.co.in> wrote:
> 
> Hi,
> 
> Facing Duplication inconsistently while bouncing Kafka producer and
> consumer in tomcat node. any help will be appreciated to find out the root
> cause.
> 
> -- 
> With Regards,
> Karthick.K

Reply via email to