I fixed this in https://github.com/apache/incubator-spark/pull/100, which was also included in the new 0.8.1 release ( https://spark.incubator.apache.org/releases/spark-release-0-8-1.html)
- Josh On Thu, Dec 19, 2013 at 5:44 AM, Jie Deng <[email protected]> wrote: > This error is happens when you use eclipse compile the .java file, and > fixed in 0.9.0 (in git). > > > > > 2013/12/19 [email protected] <[email protected]> > > I write a example MyWordCount , just set spark.akka.frameSize larger >> than default . but when I run this jar , there is a problem : >> >> 13/12/19 18:53:48 INFO ClusterTaskSetManager: Lost TID 0 (task 0.0:0) >> 13/12/19 18:53:48 INFO ClusterTaskSetManager: Loss was due to >> java.lang.AbstractMethodError >> java.lang.AbstractMethodError: >> org.apache.spark.api.java.function.WrappedFunction1.call(Ljava/lang/Object;)Ljava/lang/Object; >> at >> org.apache.spark.api.java.function.WrappedFunction1.apply(WrappedFunction1.scala:31) >> at >> org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$1$1.apply(JavaRDDLike.scala:90) >> at >> org.apache.spark.api.java.JavaRDDLike$$anonfun$fn$1$1.apply(JavaRDDLike.scala:90) >> at scala.collection.Iterator$$anon$21.hasNext(Iterator.scala:440) >> at scala.collection.Iterator$class.foreach(Iterator.scala:772) >> at scala.collection.Iterator$$anon$21.foreach(Iterator.scala:437) >> at >> scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:48) >> at >> scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:102) >> at >> scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:250) >> at scala.collection.Iterator$$anon$21.toBuffer(Iterator.scala:437) >> at >> scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:237) >> at scala.collection.Iterator$$anon$21.toArray(Iterator.scala:437) >> at org.apache.spark.rdd.RDD$$anonfun$1.apply(RDD.scala:560) >> at org.apache.spark.rdd.RDD$$anonfun$1.apply(RDD.scala:560) >> at >> org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:758) >> at >> org.apache.spark.SparkContext$$anonfun$runJob$4.apply(SparkContext.scala:758) >> >> it caused by this code : >> JavaRDD<String> words = lines.flatMap(new FlatMapFunction<String, >> String>() { >> public Iterable<String> call(String s) { >> return Arrays.asList(s.split(" ")); >> } }); >> >> there is the parent class: >> >> private[spark] abstract class WrappedFunction1[T, R] extends >> AbstractFunction1[T, R] { >> @throws(classOf[Exception]) >> def call(t: T): R >> >> final def apply(t: T): R = call(t) >> } >> >> the code is same as the JavaWordCount , I don't know what's the error . >> >> Thanks >> >> Leo >> >> ------------------------------ >> [email protected] >> > >
