[jira] [Commented] (KAFKA-6901) Kafka crashes when trying to delete segment when retetention time is reached
[ https://issues.apache.org/jira/browse/KAFKA-6901?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16695985#comment-16695985 ] Grégory R. commented on KAFKA-6901: --- It was an antivirus issue. Closing the request. > Kafka crashes when trying to delete segment when retetention time is reached > - > > Key: KAFKA-6901 > URL: https://issues.apache.org/jira/browse/KAFKA-6901 > Project: Kafka > Issue Type: Bug > Components: core, log >Affects Versions: 1.0.0, 1.1.0 > Environment: OS: Windows Server 2012 R2 >Reporter: Grégory R. >Priority: Major > Labels: windows > Attachments: 20180517 - ProcessExplorer after the crash.png, 20180517 > - ProcessExplorer before the crash.png > > > Following the parameter > {code:java} > log.retention.hours = 16{code} > kafka tries to delete segments. > This action crashes the server with following log: > > {code:java} > [2018-05-11 15:17:58,036] INFO Found deletable segments with base offsets [0] > due to retention time 60480ms breach (kafka.log.Log) > [2018-05-11 15:17:58,068] INFO Rolled new log segment for 'event-0' in 12 ms. > (kafka.log.Log) > [2018-05-11 15:17:58,068] INFO Scheduling log segment 0 for log event-0 for > deletion. (kafka.log.Log) > [2018-05-11 15:17:58,068] ERROR Error while deleting segments for event-0 in > dir C:\App\VISBridge\kafka_2.12-1.0.0\kafka-log (kafka.server.L > ogDirFailureChannel) > java.nio.file.FileSystemException: > C:\App\VISBridge\kafka_2.12-1.0.0\kafka-log\event-0\.log > -> C:\App\VISBridge\kafka_2. > 12-1.0.0\kafka-log\event-0\.log.deleted: Le processus ne > peut pas accÚder au fichier car ce fichier est utilisÚ par un a > utre processus. > at > sun.nio.fs.WindowsException.translateToIOException(WindowsException.java:86) > at > sun.nio.fs.WindowsException.rethrowAsIOException(WindowsException.java:97) > at sun.nio.fs.WindowsFileCopy.move(WindowsFileCopy.java:387) > at > sun.nio.fs.WindowsFileSystemProvider.move(WindowsFileSystemProvider.java:287) > at java.nio.file.Files.move(Files.java:1395) > at > org.apache.kafka.common.utils.Utils.atomicMoveWithFallback(Utils.java:682) > at > org.apache.kafka.common.record.FileRecords.renameTo(FileRecords.java:212) > at kafka.log.LogSegment.changeFileSuffixes(LogSegment.scala:398) > at kafka.log.Log.asyncDeleteSegment(Log.scala:1592) > at kafka.log.Log.deleteSegment(Log.scala:1579) > at kafka.log.Log.$anonfun$deleteSegments$3(Log.scala:1152) > at kafka.log.Log.$anonfun$deleteSegments$3$adapted(Log.scala:1152) > at > scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:59) > at > scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:52) > at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) > at kafka.log.Log.$anonfun$deleteSegments$2(Log.scala:1152) > at > scala.runtime.java8.JFunction0$mcI$sp.apply(JFunction0$mcI$sp.java:12) > at kafka.log.Log.maybeHandleIOException(Log.scala:1669) > at kafka.log.Log.deleteSegments(Log.scala:1143) > at kafka.log.Log.deleteOldSegments(Log.scala:1138) > at kafka.log.Log.deleteRetentionMsBreachedSegments(Log.scala:1211) > at kafka.log.Log.deleteOldSegments(Log.scala:1204) > at kafka.log.LogManager.$anonfun$cleanupLogs$3(LogManager.scala:715) > at > kafka.log.LogManager.$anonfun$cleanupLogs$3$adapted(LogManager.scala:713) > at > scala.collection.TraversableLike$WithFilter.$anonfun$foreach$1(TraversableLike.scala:789) > at scala.collection.Iterator.foreach(Iterator.scala:929) > at scala.collection.Iterator.foreach$(Iterator.scala:929) > at scala.collection.AbstractIterator.foreach(Iterator.scala:1417) > at scala.collection.IterableLike.foreach(IterableLike.scala:71) > at scala.collection.IterableLike.foreach$(IterableLike.scala:70) > at scala.collection.AbstractIterable.foreach(Iterable.scala:54) > at > scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:788) > at kafka.log.LogManager.cleanupLogs(LogManager.scala:713) > at kafka.log.LogManager.$anonfun$startup$2(LogManager.scala:341) > at > kafka.utils.KafkaScheduler.$anonfun$schedule$2(KafkaScheduler.scala:110) > at kafka.utils.CoreUtils$$anon$1.run(CoreUtils.scala:61) > at > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308) > at > java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(Sch
[jira] [Commented] (KAFKA-6901) Kafka crashes when trying to delete segment when retetention time is reached
[ https://issues.apache.org/jira/browse/KAFKA-6901?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16479048#comment-16479048 ] Grégory R. commented on KAFKA-6901: --- Seems to be a duplicate of https://issues.apache.org/jira/browse/KAFKA-1194 > Kafka crashes when trying to delete segment when retetention time is reached > - > > Key: KAFKA-6901 > URL: https://issues.apache.org/jira/browse/KAFKA-6901 > Project: Kafka > Issue Type: Bug > Components: core, log >Affects Versions: 1.0.0, 1.1.0 > Environment: OS: Windows Server 2012 R2 >Reporter: Grégory R. >Priority: Major > Labels: windows > Attachments: 20180517 - ProcessExplorer after the crash.png, 20180517 > - ProcessExplorer before the crash.png > > > Following the parameter > {code:java} > log.retention.hours = 16{code} > kafka tries to delete segments. > This action crashes the server with following log: > > {code:java} > [2018-05-11 15:17:58,036] INFO Found deletable segments with base offsets [0] > due to retention time 60480ms breach (kafka.log.Log) > [2018-05-11 15:17:58,068] INFO Rolled new log segment for 'event-0' in 12 ms. > (kafka.log.Log) > [2018-05-11 15:17:58,068] INFO Scheduling log segment 0 for log event-0 for > deletion. (kafka.log.Log) > [2018-05-11 15:17:58,068] ERROR Error while deleting segments for event-0 in > dir C:\App\VISBridge\kafka_2.12-1.0.0\kafka-log (kafka.server.L > ogDirFailureChannel) > java.nio.file.FileSystemException: > C:\App\VISBridge\kafka_2.12-1.0.0\kafka-log\event-0\.log > -> C:\App\VISBridge\kafka_2. > 12-1.0.0\kafka-log\event-0\.log.deleted: Le processus ne > peut pas accÚder au fichier car ce fichier est utilisÚ par un a > utre processus. > at > sun.nio.fs.WindowsException.translateToIOException(WindowsException.java:86) > at > sun.nio.fs.WindowsException.rethrowAsIOException(WindowsException.java:97) > at sun.nio.fs.WindowsFileCopy.move(WindowsFileCopy.java:387) > at > sun.nio.fs.WindowsFileSystemProvider.move(WindowsFileSystemProvider.java:287) > at java.nio.file.Files.move(Files.java:1395) > at > org.apache.kafka.common.utils.Utils.atomicMoveWithFallback(Utils.java:682) > at > org.apache.kafka.common.record.FileRecords.renameTo(FileRecords.java:212) > at kafka.log.LogSegment.changeFileSuffixes(LogSegment.scala:398) > at kafka.log.Log.asyncDeleteSegment(Log.scala:1592) > at kafka.log.Log.deleteSegment(Log.scala:1579) > at kafka.log.Log.$anonfun$deleteSegments$3(Log.scala:1152) > at kafka.log.Log.$anonfun$deleteSegments$3$adapted(Log.scala:1152) > at > scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:59) > at > scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:52) > at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) > at kafka.log.Log.$anonfun$deleteSegments$2(Log.scala:1152) > at > scala.runtime.java8.JFunction0$mcI$sp.apply(JFunction0$mcI$sp.java:12) > at kafka.log.Log.maybeHandleIOException(Log.scala:1669) > at kafka.log.Log.deleteSegments(Log.scala:1143) > at kafka.log.Log.deleteOldSegments(Log.scala:1138) > at kafka.log.Log.deleteRetentionMsBreachedSegments(Log.scala:1211) > at kafka.log.Log.deleteOldSegments(Log.scala:1204) > at kafka.log.LogManager.$anonfun$cleanupLogs$3(LogManager.scala:715) > at > kafka.log.LogManager.$anonfun$cleanupLogs$3$adapted(LogManager.scala:713) > at > scala.collection.TraversableLike$WithFilter.$anonfun$foreach$1(TraversableLike.scala:789) > at scala.collection.Iterator.foreach(Iterator.scala:929) > at scala.collection.Iterator.foreach$(Iterator.scala:929) > at scala.collection.AbstractIterator.foreach(Iterator.scala:1417) > at scala.collection.IterableLike.foreach(IterableLike.scala:71) > at scala.collection.IterableLike.foreach$(IterableLike.scala:70) > at scala.collection.AbstractIterable.foreach(Iterable.scala:54) > at > scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:788) > at kafka.log.LogManager.cleanupLogs(LogManager.scala:713) > at kafka.log.LogManager.$anonfun$startup$2(LogManager.scala:341) > at > kafka.utils.KafkaScheduler.$anonfun$schedule$2(KafkaScheduler.scala:110) > at kafka.utils.CoreUtils$$anon$1.run(CoreUtils.scala:61) > at > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308) > at > java.util.concurrent.ScheduledThreadPoolExecutor$Sched
[jira] [Commented] (KAFKA-6901) Kafka crashes when trying to delete segment when retetention time is reached
[ https://issues.apache.org/jira/browse/KAFKA-6901?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16479017#comment-16479017 ] Grégory R. commented on KAFKA-6901: --- * I used ProcessExplorer to check the process is locking the file but nothing. Before the crash: !20180517 - ProcessExplorer before the crash.png|width=573,height=287! After the crash: !20180517 - ProcessExplorer after the crash.png|width=605,height=317! * I updated kafka to the last version (kafka_2.11-1.1.0) but bug still here. > Kafka crashes when trying to delete segment when retetention time is reached > - > > Key: KAFKA-6901 > URL: https://issues.apache.org/jira/browse/KAFKA-6901 > Project: Kafka > Issue Type: Bug > Components: core, log >Affects Versions: 1.0.0 > Environment: OS: Windows Server 2012 R2 >Reporter: Grégory R. >Priority: Critical > Labels: windows > Attachments: 20180517 - ProcessExplorer after the crash.png, 20180517 > - ProcessExplorer before the crash.png > > > Following the parameter > {code:java} > log.retention.hours = 16{code} > kafka tries to delete segments. > This action crashes the server with following log: > > {code:java} > [2018-05-11 15:17:58,036] INFO Found deletable segments with base offsets [0] > due to retention time 60480ms breach (kafka.log.Log) > [2018-05-11 15:17:58,068] INFO Rolled new log segment for 'event-0' in 12 ms. > (kafka.log.Log) > [2018-05-11 15:17:58,068] INFO Scheduling log segment 0 for log event-0 for > deletion. (kafka.log.Log) > [2018-05-11 15:17:58,068] ERROR Error while deleting segments for event-0 in > dir C:\App\VISBridge\kafka_2.12-1.0.0\kafka-log (kafka.server.L > ogDirFailureChannel) > java.nio.file.FileSystemException: > C:\App\VISBridge\kafka_2.12-1.0.0\kafka-log\event-0\.log > -> C:\App\VISBridge\kafka_2. > 12-1.0.0\kafka-log\event-0\.log.deleted: Le processus ne > peut pas accÚder au fichier car ce fichier est utilisÚ par un a > utre processus. > at > sun.nio.fs.WindowsException.translateToIOException(WindowsException.java:86) > at > sun.nio.fs.WindowsException.rethrowAsIOException(WindowsException.java:97) > at sun.nio.fs.WindowsFileCopy.move(WindowsFileCopy.java:387) > at > sun.nio.fs.WindowsFileSystemProvider.move(WindowsFileSystemProvider.java:287) > at java.nio.file.Files.move(Files.java:1395) > at > org.apache.kafka.common.utils.Utils.atomicMoveWithFallback(Utils.java:682) > at > org.apache.kafka.common.record.FileRecords.renameTo(FileRecords.java:212) > at kafka.log.LogSegment.changeFileSuffixes(LogSegment.scala:398) > at kafka.log.Log.asyncDeleteSegment(Log.scala:1592) > at kafka.log.Log.deleteSegment(Log.scala:1579) > at kafka.log.Log.$anonfun$deleteSegments$3(Log.scala:1152) > at kafka.log.Log.$anonfun$deleteSegments$3$adapted(Log.scala:1152) > at > scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:59) > at > scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:52) > at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) > at kafka.log.Log.$anonfun$deleteSegments$2(Log.scala:1152) > at > scala.runtime.java8.JFunction0$mcI$sp.apply(JFunction0$mcI$sp.java:12) > at kafka.log.Log.maybeHandleIOException(Log.scala:1669) > at kafka.log.Log.deleteSegments(Log.scala:1143) > at kafka.log.Log.deleteOldSegments(Log.scala:1138) > at kafka.log.Log.deleteRetentionMsBreachedSegments(Log.scala:1211) > at kafka.log.Log.deleteOldSegments(Log.scala:1204) > at kafka.log.LogManager.$anonfun$cleanupLogs$3(LogManager.scala:715) > at > kafka.log.LogManager.$anonfun$cleanupLogs$3$adapted(LogManager.scala:713) > at > scala.collection.TraversableLike$WithFilter.$anonfun$foreach$1(TraversableLike.scala:789) > at scala.collection.Iterator.foreach(Iterator.scala:929) > at scala.collection.Iterator.foreach$(Iterator.scala:929) > at scala.collection.AbstractIterator.foreach(Iterator.scala:1417) > at scala.collection.IterableLike.foreach(IterableLike.scala:71) > at scala.collection.IterableLike.foreach$(IterableLike.scala:70) > at scala.collection.AbstractIterable.foreach(Iterable.scala:54) > at > scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:788) > at kafka.log.LogManager.cleanupLogs(LogManager.scala:713) > at kafka.log.LogManager.$anonfun$startup$2(LogManager.scala:341) > at > kafka.utils.KafkaScheduler.$anonfun$schedule$2(KafkaScheduler.scala:110) > at kafka.utils.CoreUtils$$anon$1.run(Co