Hi Niko, I execute the script in 0.9/CDH5 using spark-shell , and it does not generate ClassCastException. Which version are you using and can you give more stack trace ?
Cheers, a. On Tue, Mar 25, 2014 at 7:55 PM, Niko Stahl <r.niko.st...@gmail.com> wrote: > Ok, so I've been able to narrow down the problem to this specific case: > > def toCsv(userTuple: String) = {"a,b,c"} > val dataTemp = Array("line1", "line2") > val dataTempDist = sc.parallelize(dataTemp) > val usersFormatted = dataTempDist.map(toCsv) > usersFormatted.saveAsTextFile("hdfs://" + masterDomain + > ":9000/user/root/" + "test_dir") > > Even this simple mapping give me a java.lang.ClassCastException. Sorry, > my knowledge of Scala is very rudimentary. > > Thanks, > Niko > > > On Tue, Mar 25, 2014 at 5:55 PM, Niko Stahl <r.niko.st...@gmail.com> > wrote: > >> Hi, >> >> I'm trying to save an RDD to HDFS with the saveAsTextFile method on my >> ec2 cluster and am encountering the following exception (the app is called >> GraphTest): >> >> Exception failure: java.lang.ClassCastException: cannot assign instance >> of GraphTest$$anonfun$3 to field org.apache.spark.rdd.MappedRDD.f of type >> scala.Function1 in instance of org.apache.spark.rdd.MappedRDD >> >> The RDD is simply a list of strings. Strangely enough the same sequence >> of commands when executed in the Spark shell does not cause the above >> error. Any thoughts on what might be going on here? >> >> Thanks, >> Niko >> > >