Launching from eclipse (scala-ide) as a scala process gives such error, but as a java process (a java main class) works fine.
Launching as a scala process from Intellij works fine. There is something wrong at eclipse side, not in Spark. On 03/13/2015 11:47 AM, Jianshi Huang wrote: > Liancheng also found out that the Spark jars are not included in the > classpath of URLClassLoader. > > Hmm... we're very close to the truth now. > > Jianshi > > On Fri, Mar 13, 2015 at 6:03 PM, Jianshi Huang <jianshi.hu...@gmail.com > <mailto:jianshi.hu...@gmail.com>> wrote: > > I'm almost certain the problem is the ClassLoader. > > So adding > > fork := true > > solves problems for test and run. > > The problem is how can I fork a JVM for sbt console? fork in console > := true seems not working... > > Jianshi > > > On Fri, Mar 13, 2015 at 4:35 PM, Jianshi Huang > <jianshi.hu...@gmail.com <mailto:jianshi.hu...@gmail.com>> wrote: > > I guess it's a ClassLoader issue. But I have no idea how to > debug it. Any hints? > > Jianshi > > On Fri, Mar 13, 2015 at 3:00 PM, Eric Charles <e...@apache.org > <mailto:e...@apache.org>> wrote: > > i have the same issue running spark sql code from eclipse > workspace. If you run your code from the command line (with > a packaged jar) or from Intellij, I bet it should work. > > IMHO This is some how related to eclipse env, but would love > to know how to fix it (whether via eclipse conf, or via a > patch in spark). > > > > On 03/01/2015 02:32 AM, Michael Armbrust wrote: >> I think its possible that the problem is that the scala >> compiler is not being loaded by the primordial classloader >> (but instead by some child classloader) and thus the scala >> reflection mirror is failing to initialize when it can't >> find it. Unfortunately, the only solution that I know of >> is to load all required jars when the JVM starts. >> >> On Sat, Feb 28, 2015 at 5:26 PM, Ashish Nigam >> <ashnigamt...@gmail.com <mailto:ashnigamt...@gmail.com>> >> wrote: >> >> Also, can scala version play any role here? >> I am using scala 2.11.5 but all spark packages have >> dependency to scala 2.11.2 >> Just wanted to make sure that scala version is not an >> issue here. >> >> On Sat, Feb 28, 2015 at 9:18 AM, Ashish Nigam >> <ashnigamt...@gmail.com >> <mailto:ashnigamt...@gmail.com>> wrote: >> >> Hi, >> I wrote a very simple program in scala to convert >> an existing RDD to SchemaRDD. >> But createSchemaRDD function is throwing exception >> >> Exception in thread "main" >> scala.ScalaReflectionException: class >> org.apache.spark.sql.catalyst.ScalaReflection in >> JavaMirror with primordial classloader with boot >> classpath [.....] not found >> >> >> Here's more info on the versions I am using - >> >> <scala.binary.version>2.11</scala.binary.version> >> <spark.version>1.2.1</spark.version> >> <scala.version>2.11.5</scala.version> >> >> Please let me know how can I resolve this problem. >> >> Thanks >> Ashish >> >> >> > > > > -- > Jianshi Huang > > LinkedIn: jianshi > Twitter: @jshuang > Github & Blog: http://huangjs.github.com/ > > > > > -- > Jianshi Huang > > LinkedIn: jianshi > Twitter: @jshuang > Github & Blog: http://huangjs.github.com/ > > > > > -- > Jianshi Huang > > LinkedIn: jianshi > Twitter: @jshuang > Github & Blog: http://huangjs.github.com/ --------------------------------------------------------------------- To unsubscribe, e-mail: user-unsubscr...@spark.apache.org For additional commands, e-mail: user-h...@spark.apache.org