Hi, This is for first time we are getting a weird exception. After this the streams caches.
Only work around is to manually seek and commit offset to a greater number and we are needing this manual intervention again and again. Any idea what is causing it and how can we circumvent this. Note this error happens in both cases when 10.2 client or 10.1.1 client connect to kafka server 10.1.1 So this does not looks like version issue. Also we have following setting message.max.bytes=5000013 ConsumerConfig.MAX_PARTITION_FETCH_BYTES_CONFIG, "5048576" ProducerConfig.MAX_REQUEST_SIZE_CONFIG, "5048576" Rest is all default and also increasing the value for ConsumerConfig.MAX_PARTITION_FETCH_BYTES_CONFIG did not help. Stack trace below. Thanks Sachin org.apache.kafka.common.errors.SerializationException: Error deserializing key/value for partition advice-stream-6 at offset 45153795 java.lang.IllegalArgumentException: null at java.nio.Buffer.limit(Buffer.java:275) ~[na:1.8.0_122-ea] at org.apache.kafka.common.utils.Utils.sizeDelimited(Utils.java:791) ~[kafka-clients-0.10.2.0.jar:na] at org.apache.kafka.common.record.Record.value(Record.java:268) ~[kafka-clients-0.10.2.0.jar:na] at org.apache.kafka.clients.consumer.internals.Fetcher.parseRecord(Fetcher.java:867) ~[kafka-clients-0.10.2.0.jar:na] at org.apache.kafka.clients.consumer.internals.Fetcher. parseCompletedFetch(Fetcher.java:775) ~[kafka-clients-0.10.2.0.jar:na] at org.apache.kafka.clients.consumer.internals.Fetcher. fetchedRecords(Fetcher.java:473) ~[kafka-clients-0.10.2.0.jar:na] at org.apache.kafka.clients.consumer.KafkaConsumer. pollOnce(KafkaConsumer.java:1062) ~[kafka-clients-0.10.2.0.jar:na] at org.apache.kafka.clients.consumer.KafkaConsumer.poll(KafkaConsumer.java:995) ~[kafka-clients-0.10.2.0.jar:na] at org.apache.kafka.streams.processor.internals.StreamThread.runLoop(StreamThread.java:592) ~[kafka-streams-0.10.2.1-SNAPSHOT.jar:na] at org.apache.kafka.streams.processor.internals. StreamThread.run(StreamThread.java:378) ~[kafka-streams-0.10.2.1- SNAPSHOT.jar:na]