Hi.

I am developing a simple log counting application using Kafka Streams 0.10.1.1.
Its implementation is almost the same as the WordCountProcessor in the 
confluent document 
[http://docs.confluent.io/3.1.1/streams/developer-guide.html#processor-api].
I am using in-memory state store, 
its key is the ID of log category, value is count.
All the changelogs are written to a broker by context.commit() for fault 
tolerance,
but since the data I handle is large and the size of key is large, it takes a 
long time to process.
Even if it is compacted with a broker, this will put a load on the broker.
I would like to write only the latest records for each key on the broker 
instead of all changelogs at context.commit(). 
This will reduce the load on the broker and I do not think 
there will be any negative impact on fault tolerance.
If I use the persistent state store, I can do this by enabling caching, 
but I couldn't find how to accomplish this with the in-memory state store.
Can I do this?

Thank you,
--
Daisuke



Reply via email to