Hello all,

My Stream job is throwing below exception at every interval. It is first
deleting the the checkpoint file and then it's trying to checkpoint, is
this normal behaviour? I'm using Spark 1.3.0. Do you know what may cause
this issue?

15/09/24 16:35:55 INFO scheduler.TaskSetManager: Finished task 1.0 in stage
84.0 (TID 799) in 12 ms on itrac1511.cern.ch (1/8)
*15/09/24 16:35:55 INFO streaming.CheckpointWriter:
Deleting 
hdfs://p01001532067275/user/wdtmon/wdt-dstream-44446/checkpoint-1443104220000*
*15/09/24 16:35:55 INFO streaming.CheckpointWriter: Checkpoint for time
1443104220000 ms saved to file
'hdfs://p01001532067275/user/wdtmon/wdt-dstream-44446/*
checkpoint-1443104220000', took 10696 bytes and 108 ms
15/09/24 16:35:55 INFO streaming.DStreamGraph: Clearing checkpoint data for
time 1443104220000 ms
15/09/24 16:35:55 INFO streaming.DStreamGraph: Cleared checkpoint data for
time 1443104220000 ms
15/09/24 16:35:55 ERROR actor.OneForOneStrategy:
java.lang.NullPointerException
        at
org.apache.spark.streaming.DStreamGraph$$anonfun$getMaxInputStreamRememberDuration$2.apply(DStreamGraph.scala:168)
        at
org.apache.spark.streaming.DStreamGraph$$anonfun$getMaxInputStreamRememberDuration$2.apply(DStreamGraph.scala:168)
        at
scala.collection.TraversableOnce$$anonfun$maxBy$1.apply(TraversableOnce.scala:225)
        at
scala.collection.IndexedSeqOptimized$class.foldl(IndexedSeqOptimized.scala:51)
        at
scala.collection.IndexedSeqOptimized$class.reduceLeft(IndexedSeqOptimized.scala:68)
        at
scala.collection.mutable.ArrayBuffer.reduceLeft(ArrayBuffer.scala:47)
        at
scala.collection.TraversableOnce$class.maxBy(TraversableOnce.scala:225)
        at scala.collection.AbstractTraversable.maxBy(Traversable.scala:105)
        at
org.apache.spark.streaming.DStreamGraph.getMaxInputStreamRememberDuration(DStreamGraph.scala:168)
        at
org.apache.spark.streaming.scheduler.JobGenerator.clearCheckpointData(JobGenerator.scala:279)
        at org.apache.spark.streaming.scheduler.JobGenerator.org
$apache$spark$streaming$scheduler$JobGenerator$$processEvent(JobGenerator.scala:181)
        at
org.apache.spark.streaming.scheduler.JobGenerator$$anonfun$start$1$$anon$1$$anonfun$receive$1.applyOrElse(JobGenerator.scala:86)
        at akka.actor.ActorCell.receiveMessage(ActorCell.scala:498)
        at akka.actor.ActorCell.invoke(ActorCell.scala:456)
        at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:237)
        at akka.dispatch.Mailbox.run(Mailbox.scala:219)
        at
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386)
        at
scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
        at
scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
        at
scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
        at
scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)


Cheers,

Uthay

Reply via email to