Actually I get the same error even when I do something dead simple like below. I ran the same commands in sparkR directly and it worked. Is livy just does not work with sparkR this is with 2.3.1? It is a bit concerning that nothing really works via livy while works dierctly via sparkR would point to a livy issue?
%sparkr df <- createDataFrame(sqlContext, faithful) head(df) On Sun, Dec 23, 2018 at 6:59 PM Jeff Zhang <zjf...@gmail.com> wrote: > This is due to livy 0.5 doesn't support spark 2.4. Because spark 2.4 > changes its SparkR related method signature. I am afraid you have to > downgrade to spark 2.3.x > > > andrew shved <andrewshved.w...@gmail.com> 于2018年12月24日周一 上午7:48写道: > >> Spark 2.4.0 Sorry >> Zeppelin 0.8.0 >> Livy 0.5 >> >> regular livy.sparkr commands like >> 1+1 work the issue when spark comes into play >> >> On Sun, Dec 23, 2018 at 6:44 PM andrew shved <andrewshved.w...@gmail.com> >> wrote: >> >>> 0.5 with spark 2.4.9 on AWS EMR >>> >>> On Sun., Dec. 23, 2018, 6:41 p.m. Jeff Zhang <zjf...@gmail.com wrote: >>> >>>> Which version of livy do you use ? >>>> >>>> andrew shved <andrewshved.w...@gmail.com> 于2018年12月23日周日 下午11:49写道: >>>> >>>>> >>>>> been struggling wiht zeppelin + livy + sparkR integration for days. I >>>>> got livy.pyspark and livy.spark work no issues. with livy.sparkr I get >>>>> >>>>> 18/12/23 15:05:24 INFO SparkEntries: Created Spark session (with Hive >>>>> support). >>>>> Exception in thread "SparkR backend" java.lang.ClassCastException: >>>>> scala.Tuple2 cannot be cast to java.lang.Integer >>>>> at scala.runtime.BoxesRunTime.unboxToInt(BoxesRunTime.java:101) >>>>> at >>>>> org.apache.livy.repl.SparkRInterpreter$$anon$1.run(SparkRInterpreter.scala:83) >>>>> all the time :disappointed: running out of things to try >>>>> simple spark.R works >>>>> >>>>> Any ideas or advice would be appreciated. Thank you! >>>>> >>>> >>>> >>>> -- >>>> Best Regards >>>> >>>> Jeff Zhang >>>> >>> > > -- > Best Regards > > Jeff Zhang >