[jira] [Updated] (KAFKA-6292) KafkaConsumer ran into Unknown error fetching data for topic-partition caused by integer overflow in FileLogInputStream
[ https://issues.apache.org/jira/browse/KAFKA-6292?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jason Gustafson updated KAFKA-6292: --- Fix Version/s: 1.1.2 1.0.3 > KafkaConsumer ran into Unknown error fetching data for topic-partition caused > by integer overflow in FileLogInputStream > > > Key: KAFKA-6292 > URL: https://issues.apache.org/jira/browse/KAFKA-6292 > Project: Kafka > Issue Type: Bug > Components: log >Affects Versions: 0.11.0.0, 0.11.0.1, 1.0.0, 0.11.0.2 > Environment: OS:Red Hat Enterprise Linux Server release 7.3 (Maipo) > Kafka: kafka_2.12-0.11.0.0 > JDK: jdk1.8.0_121 >Reporter: Terence Yi >Priority: Major > Labels: easyfix, reliability > Fix For: 2.0.0, 1.0.3, 1.1.2 > > > Steps to reproduce: > * Broker config to reproduce this bug: > {code:java} > # The maximum size of a log segment file. When this size is reached a new > log segment will be created. > #2G > log.segment.bytes=2147483647 > {code} > * Setups: > producer sends messages constantly. > consumer polling > topic has 1 partitions and replication factor 1. > min.insync.replicas=1 > producer has "acks=all" > consumer has default "enable.auto.commit=false" > consumer manually commitSync offsets after handling messages. > kafka in standalone > * Observe log in consumer side(for me running 12 hours) > {code:java} > 2017-12-18 07:11:01.013 WARN sep105v1 > [app-consumer-subscription-pool-4-thread-20] > org.apache.kafka.clients.consumer.internals.Fetcher {} Unknown error fetching > data for topic-partition DDI.DISPATCHER.P_TVIN.W_SL.P_appx.P_ul.P_pos-0 > {code} > * Observe server.log in Kafka/logs > {code:java} > [2017-12-14 04:52:21,144] ERROR [Replica Manager on Broker 3]: Error > processing fetch operation on partition > DDI.DISPATCHER.P_TVIN.W_SL.P_appx.P_ul.P_pos-0, offset 4043314339 > (kafka.server.ReplicaManager) > org.apache.kafka.common.KafkaException: java.io.EOFException: Failed to read > `log header` from file channel `sun.nio.ch.FileChannelImpl@5604ea91`. > Expected to read 17 bytes, but reached end of file after reading 0 bytes. > Started read from position 2147483643. > at > org.apache.kafka.common.record.RecordBatchIterator.makeNext(RecordBatchIterator.java:40) > at > org.apache.kafka.common.record.RecordBatchIterator.makeNext(RecordBatchIterator.java:24) > at > org.apache.kafka.common.utils.AbstractIterator.maybeComputeNext(AbstractIterator.java:79) > at > org.apache.kafka.common.utils.AbstractIterator.hasNext(AbstractIterator.java:45) > at > org.apache.kafka.common.record.FileRecords.searchForOffsetWithSize(FileRecords.java:279) > at kafka.log.LogSegment.translateOffset(LogSegment.scala:176) > at kafka.log.LogSegment.read(LogSegment.scala:228) > at kafka.log.Log.read(Log.scala:938) > at kafka.server.ReplicaManager.read$1(ReplicaManager.scala:719) > at > kafka.server.ReplicaManager.$anonfun$readFromLocalLog$6(ReplicaManager.scala:780) > at > scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:59) > at > scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:52) > at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) > at > kafka.server.ReplicaManager.readFromLocalLog(ReplicaManager.scala:779) > at kafka.server.ReplicaManager.fetchMessages(ReplicaManager.scala:617) > at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:615) > at kafka.server.KafkaApis.handle(KafkaApis.scala:98) > at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:66) > at java.lang.Thread.run(Thread.java:745) > Caused by: java.io.EOFException: Failed to read `log header` from file > channel `sun.nio.ch.FileChannelImpl@5604ea91`. Expected to read 17 bytes, but > reached end of file after reading 0 bytes. Started read from position > 2147483643. > at org.apache.kafka.common.utils.Utils.readFullyOrFail(Utils.java:751) > at > org.apache.kafka.common.record.FileLogInputStream.nextBatch(FileLogInputStream.java:66) > at > org.apache.kafka.common.record.FileLogInputStream.nextBatch(FileLogInputStream.java:40) > at > org.apache.kafka.common.record.RecordBatchIterator.makeNext(RecordBatchIterator.java:35) > ... 18 more > {code} > * Impact: > # After EOF exception occurs, the consumer will failed to consume the remain > message > # After the segments log files which cause the EOF exception has been deleted > by the log Cleaner thread. Consumer recovered to consumer message. > # Have no impact from the view of producer >
[jira] [Updated] (KAFKA-6292) KafkaConsumer ran into Unknown error fetching data for topic-partition caused by integer overflow in FileLogInputStream
[ https://issues.apache.org/jira/browse/KAFKA-6292?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Damian Guy updated KAFKA-6292: -- Fix Version/s: (was: 1.1.0) 1.2.0 > KafkaConsumer ran into Unknown error fetching data for topic-partition caused > by integer overflow in FileLogInputStream > > > Key: KAFKA-6292 > URL: https://issues.apache.org/jira/browse/KAFKA-6292 > Project: Kafka > Issue Type: Bug > Components: log >Affects Versions: 0.11.0.0, 0.11.0.1, 1.0.0, 0.11.0.2 > Environment: OS:Red Hat Enterprise Linux Server release 7.3 (Maipo) > Kafka: kafka_2.12-0.11.0.0 > JDK: jdk1.8.0_121 >Reporter: Terence Yi >Priority: Major > Labels: easyfix, reliability > Fix For: 1.2.0 > > > Steps to reproduce: > * Broker config to reproduce this bug: > {code:java} > # The maximum size of a log segment file. When this size is reached a new > log segment will be created. > #2G > log.segment.bytes=2147483647 > {code} > * Setups: > producer sends messages constantly. > consumer polling > topic has 1 partitions and replication factor 1. > min.insync.replicas=1 > producer has "acks=all" > consumer has default "enable.auto.commit=false" > consumer manually commitSync offsets after handling messages. > kafka in standalone > * Observe log in consumer side(for me running 12 hours) > {code:java} > 2017-12-18 07:11:01.013 WARN sep105v1 > [app-consumer-subscription-pool-4-thread-20] > org.apache.kafka.clients.consumer.internals.Fetcher {} Unknown error fetching > data for topic-partition DDI.DISPATCHER.P_TVIN.W_SL.P_appx.P_ul.P_pos-0 > {code} > * Observe server.log in Kafka/logs > {code:java} > [2017-12-14 04:52:21,144] ERROR [Replica Manager on Broker 3]: Error > processing fetch operation on partition > DDI.DISPATCHER.P_TVIN.W_SL.P_appx.P_ul.P_pos-0, offset 4043314339 > (kafka.server.ReplicaManager) > org.apache.kafka.common.KafkaException: java.io.EOFException: Failed to read > `log header` from file channel `sun.nio.ch.FileChannelImpl@5604ea91`. > Expected to read 17 bytes, but reached end of file after reading 0 bytes. > Started read from position 2147483643. > at > org.apache.kafka.common.record.RecordBatchIterator.makeNext(RecordBatchIterator.java:40) > at > org.apache.kafka.common.record.RecordBatchIterator.makeNext(RecordBatchIterator.java:24) > at > org.apache.kafka.common.utils.AbstractIterator.maybeComputeNext(AbstractIterator.java:79) > at > org.apache.kafka.common.utils.AbstractIterator.hasNext(AbstractIterator.java:45) > at > org.apache.kafka.common.record.FileRecords.searchForOffsetWithSize(FileRecords.java:279) > at kafka.log.LogSegment.translateOffset(LogSegment.scala:176) > at kafka.log.LogSegment.read(LogSegment.scala:228) > at kafka.log.Log.read(Log.scala:938) > at kafka.server.ReplicaManager.read$1(ReplicaManager.scala:719) > at > kafka.server.ReplicaManager.$anonfun$readFromLocalLog$6(ReplicaManager.scala:780) > at > scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:59) > at > scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:52) > at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) > at > kafka.server.ReplicaManager.readFromLocalLog(ReplicaManager.scala:779) > at kafka.server.ReplicaManager.fetchMessages(ReplicaManager.scala:617) > at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:615) > at kafka.server.KafkaApis.handle(KafkaApis.scala:98) > at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:66) > at java.lang.Thread.run(Thread.java:745) > Caused by: java.io.EOFException: Failed to read `log header` from file > channel `sun.nio.ch.FileChannelImpl@5604ea91`. Expected to read 17 bytes, but > reached end of file after reading 0 bytes. Started read from position > 2147483643. > at org.apache.kafka.common.utils.Utils.readFullyOrFail(Utils.java:751) > at > org.apache.kafka.common.record.FileLogInputStream.nextBatch(FileLogInputStream.java:66) > at > org.apache.kafka.common.record.FileLogInputStream.nextBatch(FileLogInputStream.java:40) > at > org.apache.kafka.common.record.RecordBatchIterator.makeNext(RecordBatchIterator.java:35) > ... 18 more > {code} > * Impact: > # After EOF exception occurs, the consumer will failed to consume the remain > message > # After the segments log files which cause the EOF exception has been deleted > by the log Cleaner thread. Consumer recovered to consumer message. > # Have no impact from the view of producer > * Analysis: >
[jira] [Updated] (KAFKA-6292) KafkaConsumer ran into Unknown error fetching data for topic-partition caused by integer overflow in FileLogInputStream
[ https://issues.apache.org/jira/browse/KAFKA-6292?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Terence Yi updated KAFKA-6292: -- Description: Steps to reproduce: * Broker config to reproduce this bug: {code:java} # The maximum size of a log segment file. When this size is reached a new log segment will be created. #2G log.segment.bytes=2147483647 {code} * Setups: producer sends messages constantly. consumer polling topic has 1 partitions and replication factor 1. min.insync.replicas=1 producer has "acks=all" consumer has default "enable.auto.commit=false" consumer manually commitSync offsets after handling messages. kafka in standalone * Observe log in consumer side(for me running 12 hours) {code:java} 2017-12-18 07:11:01.013 WARN sep105v1 [app-consumer-subscription-pool-4-thread-20] org.apache.kafka.clients.consumer.internals.Fetcher {} Unknown error fetching data for topic-partition DDI.DISPATCHER.P_TVIN.W_SL.P_appx.P_ul.P_pos-0 {code} * Observe server.log in Kafka/logs {code:java} [2017-12-14 04:52:21,144] ERROR [Replica Manager on Broker 3]: Error processing fetch operation on partition DDI.DISPATCHER.P_TVIN.W_SL.P_appx.P_ul.P_pos-0, offset 4043314339 (kafka.server.ReplicaManager) org.apache.kafka.common.KafkaException: java.io.EOFException: Failed to read `log header` from file channel `sun.nio.ch.FileChannelImpl@5604ea91`. Expected to read 17 bytes, but reached end of file after reading 0 bytes. Started read from position 2147483643. at org.apache.kafka.common.record.RecordBatchIterator.makeNext(RecordBatchIterator.java:40) at org.apache.kafka.common.record.RecordBatchIterator.makeNext(RecordBatchIterator.java:24) at org.apache.kafka.common.utils.AbstractIterator.maybeComputeNext(AbstractIterator.java:79) at org.apache.kafka.common.utils.AbstractIterator.hasNext(AbstractIterator.java:45) at org.apache.kafka.common.record.FileRecords.searchForOffsetWithSize(FileRecords.java:279) at kafka.log.LogSegment.translateOffset(LogSegment.scala:176) at kafka.log.LogSegment.read(LogSegment.scala:228) at kafka.log.Log.read(Log.scala:938) at kafka.server.ReplicaManager.read$1(ReplicaManager.scala:719) at kafka.server.ReplicaManager.$anonfun$readFromLocalLog$6(ReplicaManager.scala:780) at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:59) at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:52) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) at kafka.server.ReplicaManager.readFromLocalLog(ReplicaManager.scala:779) at kafka.server.ReplicaManager.fetchMessages(ReplicaManager.scala:617) at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:615) at kafka.server.KafkaApis.handle(KafkaApis.scala:98) at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:66) at java.lang.Thread.run(Thread.java:745) Caused by: java.io.EOFException: Failed to read `log header` from file channel `sun.nio.ch.FileChannelImpl@5604ea91`. Expected to read 17 bytes, but reached end of file after reading 0 bytes. Started read from position 2147483643. at org.apache.kafka.common.utils.Utils.readFullyOrFail(Utils.java:751) at org.apache.kafka.common.record.FileLogInputStream.nextBatch(FileLogInputStream.java:66) at org.apache.kafka.common.record.FileLogInputStream.nextBatch(FileLogInputStream.java:40) at org.apache.kafka.common.record.RecordBatchIterator.makeNext(RecordBatchIterator.java:35) ... 18 more {code} * Impact: # After EOF exception occurs, the consumer will failed to consume the remain message # After the segments log files which cause the EOF exception has been deleted by the log Cleaner thread. Consumer recovered to consumer message. # Have no impact from the view of producer * Analysis: # Kafka log file list: {code:java} ls /ssd/kafka-logs/DDI.DISPATCHER.P_TVIN.W_SL.P_appx.P_ul.P_pos-0/*.log -rw-r--r-- 1 root root 2147481479 Dec 27 14:15 0004043314387.log -rw-r--r-- 1 root root 2147483647 Dec 27 14:15 0004039884900.log {code} # use tools to check the log: /bin/kafka-run-class.sh kafka.tools.DumpLogSegments --deep-iteration --print-data-log --files /ssd/kafka-logs/DDI.DISPATCHER.P_TVIN.W_SL.P_appx.P_ul.P_pos-0/004039884900.log Exception print in console is the same in the server.log: {code:java} bin/kafka-run-class.sh kafka.tools.DumpLogSegments --deep-iteration --print-data-log --files 004039884900.log > tmp.log Exception in thread "main" org.apache.kafka.common.KafkaException: java.io.EOFException: Failed to read `log header` from file channel `sun.nio.ch.FileChannelImpl@4e41089d`. Expected to read 17 bytes, but reached end of file after reading 0 bytes. Started read from position 2147483637. a
[jira] [Updated] (KAFKA-6292) KafkaConsumer ran into Unknown error fetching data for topic-partition caused by integer overflow in FileLogInputStream
[ https://issues.apache.org/jira/browse/KAFKA-6292?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Terence Yi updated KAFKA-6292: -- Labels: easyfix reliability (was: reliability) > KafkaConsumer ran into Unknown error fetching data for topic-partition caused > by integer overflow in FileLogInputStream > > > Key: KAFKA-6292 > URL: https://issues.apache.org/jira/browse/KAFKA-6292 > Project: Kafka > Issue Type: Bug > Components: log >Affects Versions: 0.11.0.0, 0.11.0.1, 1.0.0, 0.11.0.2 > Environment: OS:Red Hat Enterprise Linux Server release 7.3 (Maipo) > Kafka: kafka_2.12-0.11.0.0 > JDK: jdk1.8.0_121 >Reporter: Terence Yi > Labels: easyfix, reliability > Fix For: 1.1.0 > > > Steps to reproduce: > * Broker config to reproduce this bug: > {code:java} > # The maximum size of a log segment file. When this size is reached a new > log segment will be created. > #2G > log.segment.bytes=2147483647 > {code} > * Setups: > producer sends messages constantly. > consumer polling > topic has 1 partitions and replication factor 1. > min.insync.replicas=1 > producer has "acks=all" > consumer has default "enable.auto.commit=false" > consumer manually commitSync offsets after handling messages. > kafka in standalone > * Observe log in consumer side(for me running 12 hours) > {code:java} > 2017-12-18 07:11:01.013 WARN sep105v1 > [app-consumer-subscription-pool-4-thread-20] > org.apache.kafka.clients.consumer.internals.Fetcher {} Unknown error fetching > data for topic-partition DDI.DISPATCHER.P_TVIN.W_SL.P_appx.P_ul.P_pos-13 > {code} > * Observe server.log in Kafka/logs > {code:java} > [2017-12-14 04:52:21,144] ERROR [Replica Manager on Broker 3]: Error > processing fetch operation on partition > DDI.DISPATCHER.P_TVIN.W_SL.P_appx.P_ul.P_pos-6, offset 4043314339 > (kafka.server.ReplicaManager) > org.apache.kafka.common.KafkaException: java.io.EOFException: Failed to read > `log header` from file channel `sun.nio.ch.FileChannelImpl@5604ea91`. > Expected to read 17 bytes, but reached end of file after reading 0 bytes. > Started read from position 2147483643. > at > org.apache.kafka.common.record.RecordBatchIterator.makeNext(RecordBatchIterator.java:40) > at > org.apache.kafka.common.record.RecordBatchIterator.makeNext(RecordBatchIterator.java:24) > at > org.apache.kafka.common.utils.AbstractIterator.maybeComputeNext(AbstractIterator.java:79) > at > org.apache.kafka.common.utils.AbstractIterator.hasNext(AbstractIterator.java:45) > at > org.apache.kafka.common.record.FileRecords.searchForOffsetWithSize(FileRecords.java:279) > at kafka.log.LogSegment.translateOffset(LogSegment.scala:176) > at kafka.log.LogSegment.read(LogSegment.scala:228) > at kafka.log.Log.read(Log.scala:938) > at kafka.server.ReplicaManager.read$1(ReplicaManager.scala:719) > at > kafka.server.ReplicaManager.$anonfun$readFromLocalLog$6(ReplicaManager.scala:780) > at > scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:59) > at > scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:52) > at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) > at > kafka.server.ReplicaManager.readFromLocalLog(ReplicaManager.scala:779) > at kafka.server.ReplicaManager.fetchMessages(ReplicaManager.scala:617) > at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:615) > at kafka.server.KafkaApis.handle(KafkaApis.scala:98) > at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:66) > at java.lang.Thread.run(Thread.java:745) > Caused by: java.io.EOFException: Failed to read `log header` from file > channel `sun.nio.ch.FileChannelImpl@5604ea91`. Expected to read 17 bytes, but > reached end of file after reading 0 bytes. Started read from position > 2147483643. > at org.apache.kafka.common.utils.Utils.readFullyOrFail(Utils.java:751) > at > org.apache.kafka.common.record.FileLogInputStream.nextBatch(FileLogInputStream.java:66) > at > org.apache.kafka.common.record.FileLogInputStream.nextBatch(FileLogInputStream.java:40) > at > org.apache.kafka.common.record.RecordBatchIterator.makeNext(RecordBatchIterator.java:35) > ... 18 more > {code} > * Impact: > # After EOF exception occurs, the consumer will failed to consume the remain > message > # After the segments log files which cause the EOF exception has been deleted > by the log Cleaner thread. Consumer recovered to consumer message. > # Have no impact from the view of producer > * Analysis: > # Kafka log file list: > {code:java}
[jira] [Updated] (KAFKA-6292) KafkaConsumer ran into Unknown error fetching data for topic-partition caused by integer overflow in FileLogInputStream
[ https://issues.apache.org/jira/browse/KAFKA-6292?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Terence Yi updated KAFKA-6292: -- Affects Version/s: 0.11.0.1 1.0.0 0.11.0.2 Description: Steps to reproduce: * Broker config to reproduce this bug: {code:java} # The maximum size of a log segment file. When this size is reached a new log segment will be created. #2G log.segment.bytes=2147483647 {code} * Setups: producer sends messages constantly. consumer polling topic has 1 partitions and replication factor 1. min.insync.replicas=1 producer has "acks=all" consumer has default "enable.auto.commit=false" consumer manually commitSync offsets after handling messages. kafka in standalone * Observe log in consumer side(for me running 12 hours) {code:java} 2017-12-18 07:11:01.013 WARN sep105v1 [app-consumer-subscription-pool-4-thread-20] org.apache.kafka.clients.consumer.internals.Fetcher {} Unknown error fetching data for topic-partition DDI.DISPATCHER.P_TVIN.W_SL.P_appx.P_ul.P_pos-13 {code} * Observe server.log in Kafka/logs {code:java} [2017-12-14 04:52:21,144] ERROR [Replica Manager on Broker 3]: Error processing fetch operation on partition DDI.DISPATCHER.P_TVIN.W_SL.P_appx.P_ul.P_pos-6, offset 4043314339 (kafka.server.ReplicaManager) org.apache.kafka.common.KafkaException: java.io.EOFException: Failed to read `log header` from file channel `sun.nio.ch.FileChannelImpl@5604ea91`. Expected to read 17 bytes, but reached end of file after reading 0 bytes. Started read from position 2147483643. at org.apache.kafka.common.record.RecordBatchIterator.makeNext(RecordBatchIterator.java:40) at org.apache.kafka.common.record.RecordBatchIterator.makeNext(RecordBatchIterator.java:24) at org.apache.kafka.common.utils.AbstractIterator.maybeComputeNext(AbstractIterator.java:79) at org.apache.kafka.common.utils.AbstractIterator.hasNext(AbstractIterator.java:45) at org.apache.kafka.common.record.FileRecords.searchForOffsetWithSize(FileRecords.java:279) at kafka.log.LogSegment.translateOffset(LogSegment.scala:176) at kafka.log.LogSegment.read(LogSegment.scala:228) at kafka.log.Log.read(Log.scala:938) at kafka.server.ReplicaManager.read$1(ReplicaManager.scala:719) at kafka.server.ReplicaManager.$anonfun$readFromLocalLog$6(ReplicaManager.scala:780) at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:59) at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:52) at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) at kafka.server.ReplicaManager.readFromLocalLog(ReplicaManager.scala:779) at kafka.server.ReplicaManager.fetchMessages(ReplicaManager.scala:617) at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:615) at kafka.server.KafkaApis.handle(KafkaApis.scala:98) at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:66) at java.lang.Thread.run(Thread.java:745) Caused by: java.io.EOFException: Failed to read `log header` from file channel `sun.nio.ch.FileChannelImpl@5604ea91`. Expected to read 17 bytes, but reached end of file after reading 0 bytes. Started read from position 2147483643. at org.apache.kafka.common.utils.Utils.readFullyOrFail(Utils.java:751) at org.apache.kafka.common.record.FileLogInputStream.nextBatch(FileLogInputStream.java:66) at org.apache.kafka.common.record.FileLogInputStream.nextBatch(FileLogInputStream.java:40) at org.apache.kafka.common.record.RecordBatchIterator.makeNext(RecordBatchIterator.java:35) ... 18 more {code} * Impact: # After EOF exception occurs, the consumer will failed to consume the remain message # After the segments log files which cause the EOF exception has been deleted by the log Cleaner thread. Consumer recovered to consumer message. # Have no impact from the view of producer * Analysis: # Kafka log file list: {code:java} ls /ssd/kafka-logs/DDI.DISPATCHER.P_TVIN.W_SL.P_appx.P_ul.P_pos-0/*.log -rw-r--r-- 1 root root 2147481479 Dec 27 14:15 0004043314387.log -rw-r--r-- 1 root root 2147483647 Dec 27 14:15 0004039884900.log {code} # use tools to check the log: /bin/kafka-run-class.sh kafka.tools.DumpLogSegments --deep-iteration --print-data-log --files /ssd/kafka-logs/DDI.DISPATCHER.P_TVIN.W_SL.P_appx.P_ul.P_pos-0/004039884900.log Exception print in console is the same in the server.log: {code:java} bin/kafka-run-class.sh kafka.tools.DumpLogSegments --deep-iteration --print-data-log --files 004039884900.log > tmp.log Exception in thread "main" org.apache.kafka.common.KafkaException: java.io.EOFException: Failed to read `log header` from file channel `sun.nio.ch.FileChannelImpl@4e41089d`. Expected to read