[ 
https://issues.apache.org/jira/browse/KAFKA-5628?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17887876#comment-17887876
 ] 

Charu Agarwal commented on KAFKA-5628:
--------------------------------------

Hi , 

Is there any update on this ticket ? When can we expect a fix on this disk full 
issue ?

Thanks,

Charu

> Kafka Startup fails on corrupted index files
> --------------------------------------------
>
>                 Key: KAFKA-5628
>                 URL: https://issues.apache.org/jira/browse/KAFKA-5628
>             Project: Kafka
>          Issue Type: Bug
>    Affects Versions: 0.10.2.0
>         Environment: Ubuntu 14.04, Java 8(1.8.0_65)
>            Reporter: Prasanna Gautam
>            Assignee: Jun Rao
>            Priority: Minor
>
> One of our kafka brokers shut down after a load test and while there are some 
> corrupted index files , the broker is failing to start with a unsafe memory 
> access error
> {code:java}
> [2017-07-23 15:52:32,019] FATAL Fatal error during KafkaServerStartable 
> startup. Prepare to shutdown (kafka.server.KafkaServerStartable)
> java.lang.InternalError: a fault occurred in a recent unsafe memory access 
> operation in compiled Java code
>         at sun.nio.ch.FileChannelImpl.read(FileChannelImpl.java:53)
>         at org.apache.kafka.common.utils.Utils.readFully(Utils.java:854)
>         at org.apache.kafka.common.utils.Utils.readFullyOrFail(Utils.java:827)
>         at 
> org.apache.kafka.common.record.FileLogInputStream$FileChannelLogEntry.loadRecord(FileLogInputStream.java:136)
>         at 
> org.apache.kafka.common.record.FileLogInputStream$FileChannelLogEntry.record(FileLogInputStream.java:149)
>         at kafka.log.LogSegment$$anonfun$recover$1.apply(LogSegment.scala:225)
>         at kafka.log.LogSegment$$anonfun$recover$1.apply(LogSegment.scala:224)
>         at scala.collection.Iterator$class.foreach(Iterator.scala:893)
>         at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
>         at scala.collection.IterableLike$class.foreach(IterableLike.scala:72)
>         at scala.collection.AbstractIterable.foreach(Iterable.scala:54)
>         at kafka.log.LogSegment.recover(LogSegment.scala:224)
>         at kafka.log.Log$$anonfun$loadSegments$4.apply(Log.scala:231)
>         at kafka.log.Log$$anonfun$loadSegments$4.apply(Log.scala:188)
>         at 
> scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
>         at 
> scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
>         at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:186)
>         at 
> scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)
>         at kafka.log.Log.loadSegments(Log.scala:188)
>         at kafka.log.Log.<init>(Log.scala:116)
>         at 
> kafka.log.LogManager$$anonfun$loadLogs$2$$anonfun$3$$anonfun$apply$10$$anonfun$apply$1.apply$mcV$sp(LogManager.scala:157)
>         at kafka.utils.CoreUtils$$anon$1.run(CoreUtils.scala:57)
>         at 
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>         at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>         at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>         at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>         at java.lang.Thread.run(Thread.java:745)
> {code}
> This doesn't seem to be same as 
> https://issues.apache.org/jira/browse/KAFKA-1554 because these topics are 
> actively in use and the other empty indices are recovered fine..
> Kafka on the machine had died because the disk was full. 
> It seems to have resolved after the disk issue. Should kafka just check disk 
> at startup and refuse to continue starting up? 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to