[
https://issues.apache.org/jira/browse/SPARK-711?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Tathagata Das closed SPARK-711.
-------------------------------
Resolution: Duplicate
Duplicate issue: https://issues.apache.org/jira/browse/SPARK-703
> Spark Streaming 0.7.0: ArrayIndexOutOfBoundsException in KafkaWordCount
> Example
> -------------------------------------------------------------------------------
>
> Key: SPARK-711
> URL: https://issues.apache.org/jira/browse/SPARK-711
> Project: Spark
> Issue Type: Bug
> Components: Streaming
> Affects Versions: 0.7.0
> Reporter: Craig A. Vanderborgh
>
> The unmodified KafkaWordCount example is crashing when run under Mesos. It
> works fine when the master is "local". The KafkaWordCount job works for
> about 5 iterations, then the exceptions start. This problem is related to
> windowing. Here is the stack trace:
> 13/03/07 15:43:46 ERROR streaming.JobManager: Running streaming job 5 @
> 1362696226000 ms failed
> java.lang.ArrayIndexOutOfBoundsException: 0
> at
> spark.rdd.CoGroupedRDD$$anonfun$getPartitions$1$$anonfun$apply$mcVI$sp$1.apply(CoGroupedRDD.scala:71)
> at
> spark.rdd.CoGroupedRDD$$anonfun$getPartitions$1$$anonfun$apply$mcVI$sp$1.apply(CoGroupedRDD.scala:65)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:233)
> at
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:233)
> at
> scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:60)
> at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
> at
> scala.collection.TraversableLike$class.map(TraversableLike.scala:233)
> at scala.collection.mutable.ArrayBuffer.map(ArrayBuffer.scala:47)
> at spark.rdd.CoGroupedRDD.getPartitions(CoGroupedRDD.scala:63)
> at spark.RDD.partitions(RDD.scala:168)
> at spark.MappedValuesRDD.getPartitions(PairRDDFunctions.scala:646)
> at spark.RDD.partitions(RDD.scala:168)
> at spark.RDD.take(RDD.scala:579)
> at
> spark.streaming.DStream$$anonfun$foreachFunc$2$1.apply(DStream.scala:495)
> at
> spark.streaming.DStream$$anonfun$foreachFunc$2$1.apply(DStream.scala:494)
> at
> spark.streaming.dstream.ForEachDStream$$anonfun$1.apply$mcV$sp(ForEachDStream.scala:22)
> at
> spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:21)
> at
> spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:21)
> at spark.streaming.Job.run(Job.scala:10)
> at spark.streaming.JobManager$JobHandler.run(JobManager.scala:17)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
> at java.lang.Thread.run(Thread.java:662)
> Please let me know if I can help or provide additional information.
> Craig
--
This message was sent by Atlassian JIRA
(v6.2#6252)