[ 
https://issues.apache.org/jira/browse/SPARK-711?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tathagata Das closed SPARK-711.
-------------------------------

    Resolution: Duplicate

Duplicate issue: https://issues.apache.org/jira/browse/SPARK-703

> Spark Streaming 0.7.0: ArrayIndexOutOfBoundsException in KafkaWordCount 
> Example
> -------------------------------------------------------------------------------
>
>                 Key: SPARK-711
>                 URL: https://issues.apache.org/jira/browse/SPARK-711
>             Project: Spark
>          Issue Type: Bug
>          Components: Streaming
>    Affects Versions: 0.7.0
>            Reporter: Craig A. Vanderborgh
>
> The unmodified KafkaWordCount example is crashing when run under Mesos.  It 
> works fine when the master is "local".  The KafkaWordCount job works for 
> about 5 iterations, then the exceptions start.  This problem is related to 
> windowing.  Here is the stack trace:
> 13/03/07 15:43:46 ERROR streaming.JobManager: Running streaming job 5 @ 
> 1362696226000 ms failed
> java.lang.ArrayIndexOutOfBoundsException: 0
>         at 
> spark.rdd.CoGroupedRDD$$anonfun$getPartitions$1$$anonfun$apply$mcVI$sp$1.apply(CoGroupedRDD.scala:71)
>         at 
> spark.rdd.CoGroupedRDD$$anonfun$getPartitions$1$$anonfun$apply$mcVI$sp$1.apply(CoGroupedRDD.scala:65)
>         at 
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:233)
>         at 
> scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:233)
>         at 
> scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:60)
>         at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
>         at 
> scala.collection.TraversableLike$class.map(TraversableLike.scala:233)
>         at scala.collection.mutable.ArrayBuffer.map(ArrayBuffer.scala:47)
>         at spark.rdd.CoGroupedRDD.getPartitions(CoGroupedRDD.scala:63)
>         at spark.RDD.partitions(RDD.scala:168)
>         at spark.MappedValuesRDD.getPartitions(PairRDDFunctions.scala:646)
>         at spark.RDD.partitions(RDD.scala:168)
>         at spark.RDD.take(RDD.scala:579)
>         at 
> spark.streaming.DStream$$anonfun$foreachFunc$2$1.apply(DStream.scala:495)
>         at 
> spark.streaming.DStream$$anonfun$foreachFunc$2$1.apply(DStream.scala:494)
>         at 
> spark.streaming.dstream.ForEachDStream$$anonfun$1.apply$mcV$sp(ForEachDStream.scala:22)
>         at 
> spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:21)
>         at 
> spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:21)
>         at spark.streaming.Job.run(Job.scala:10)
>         at spark.streaming.JobManager$JobHandler.run(JobManager.scala:17)
>         at 
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>         at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>         at java.lang.Thread.run(Thread.java:662)
> Please let me know if I can help or provide additional information.
> Craig



--
This message was sent by Atlassian JIRA
(v6.2#6252)

Reply via email to