Alright, just fixed it.
Found out that I had old ENV variables set for HADOOP_HOME and SPARK_HOME
in my .bash_profile.

On Tue, Sep 8, 2015 at 12:17 PM, Amos B. Elberg <amos.elb...@gmail.com>
wrote:

> Did you use a prebuilt spark or compile it yourself?
>
> > On Sep 7, 2015, at 10:34 PM, Corneau Damien <cornead...@gmail.com>
> wrote:
> >
> > I made a git clean -dxf to have a clean repository, I even deleted the
> .m2
> > folder.
> > I'm not using any configuration, and also tried with spark 1.4 but still
> > getting the same error.
> >
> > This is the files I have in the interpreter/spark/dep directory:
> >
> > datanucleus-api-jdo-3.2.6.jar
> >
> > datanucleus-rdbms-3.2.9.jar
> >
> > datanucleus-core-3.2.10.jar
> >
> > zeppelin-spark-dependencies-0.6.0-incubating-SNAPSHOT.jar
> >
> > On Tue, Sep 8, 2015 at 12:22 AM, Amos B. Elberg <amos.elb...@gmail.com>
> > wrote:
> >
> >> Guys this is very similar to the issues I was having since we changed to
> >> the "spark dependencies" module configuration.
> >>
> >> To solve it, I had to switch to the spark-submit system. I also had to
> >> recompile spark, since apparently some build configurations (and those
> with
> >> Hadoop provided) don't put some classes where Zeppelin now expects.
> >> Ultimately, I had to wipe out the spark and Zeppelin directories
> completely
> >> and rebuild from scratch.
> >>
> >> That said, I continue to feel like there's something going on with
> either
> >> the build process or class loader or both with that spark-dependency
> >> change, that we just don't understand yet.
> >>
> >>> On Sep 7, 2015, at 11:00 AM, moon soo Lee <m...@apache.org> wrote:
> >>>
> >>> I just tried the current master branch
> >>> (0b47cde4410bd52b0a9ec070fda15a22da96980d)
> >>> with your build command and i couldn't see that error.
> >>>
> >>> Can you verify you have files under interpreter/spark/dep directory
> after
> >>> the build?
> >>>
> >>> Best,
> >>> moon
> >>>
> >>>> On Mon, Sep 7, 2015 at 2:32 AM Corneau Damien <cornead...@gmail.com>
> >> wrote:
> >>>>
> >>>> I'm always building using: mvn clean package -Pspark-1.3
> >>>> -Dhadoop.version=2.2.0 -Phadoop-2.2 -DskipTests
> >>>>
> >>>> However today, I haven't been able to run anything with the spark
> >>>> interpreter.
> >>>>
> >>>>
> >>>> Exception in thread "pool-1-thread-2" java.lang.NoClassDefFoundError:
> >>>> scala/tools/nsc/settings/AbsSettings$AbsSetting
> >>>>
> >>>> at java.lang.Class.forName0(Native Method)
> >>>>
> >>>> at java.lang.Class.forName(Class.java:190)
> >>>>
> >>>> at
> >>
> org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer.createInterpreter(RemoteInterpreterServer.java:136)
> >>>>
> >>>> at
> >>
> org.apache.zeppelin.interpreter.thrift.RemoteInterpreterService$Processor$createInterpreter.getResult(RemoteInterpreterService.java:990)
> >>>>
> >>>> at
> >>
> org.apache.zeppelin.interpreter.thrift.RemoteInterpreterService$Processor$createInterpreter.getResult(RemoteInterpreterService.java:975)
> >>>>
> >>>> at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
> >>>>
> >>>> at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
> >>>>
> >>>> at
> >>
> org.apache.thrift.server.TThreadPoolServer$WorkerProcess.run(TThreadPoolServer.java:285)
> >>>>
> >>>> at
> >>
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
> >>>>
> >>>> at
> >>
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
> >>>>
> >>>> at java.lang.Thread.run(Thread.java:745)
> >>>>
> >>>> Caused by: java.lang.ClassNotFoundException:
> >>>> scala.tools.nsc.settings.AbsSettings$AbsSetting
> >>>>
> >>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
> >>>>
> >>>> at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
> >>>>
> >>>> at java.security.AccessController.doPrivileged(Native Method)
> >>>>
> >>>> at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
> >>>>
> >>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
> >>>>
> >>>> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
> >>>>
> >>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
> >>>>
> >>>> ... 11 more
> >>
>

Reply via email to