Find the issue, this is due to the inconsistency between spark.master in
spark-defaults.conf spark.master in zeppelin side.

On Tue, Aug 2, 2016 at 9:20 AM, Jeff Zhang <zjf...@gmail.com> wrote:

>
> I follow the the instruction in this PR to build zeppelin and try to run
> in yarn-client mode, but hit very weird issue. It seems the SPARK_HOME is
> messed up. Does anyone hit this issue or can run zeppelin on spark 2.0 in
> yarn-client mode successfully ?
>
> https://github.com/apache/zeppelin/pull/1195
>
> ERROR [2016-08-02 09:05:27,771] ({pool-4-thread-5}
> Logging.scala[logError]:91) - Error initializing SparkContext.
> java.lang.IllegalStateException: Library directory
> '/Users/jzhang/Temp/hadoop_tmp/nm-local-dir/usercache/jzhang/appcache/application_1470097474471_0011/container_1470097474471_0011_01_000001/assembly/target/scala-2.11/jars'
> does not exist; make sure Spark is built.
>     at
> org.apache.spark.launcher.CommandBuilderUtils.checkState(CommandBuilderUtils.java:248)
>     at
> org.apache.spark.launcher.CommandBuilderUtils.findJarsDir(CommandBuilderUtils.java:368)
>     at
> org.apache.spark.launcher.YarnCommandBuilderUtils$.findJarsDir(YarnCommandBuilderUtils.scala:38)
>     at
> org.apache.spark.deploy.yarn.Client.prepareLocalResources(Client.scala:500)
>     at
> org.apache.spark.deploy.yarn.Client.createContainerLaunchContext(Client.scala:834)
>     at
> org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:167)
>     at
> org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:56)
>     at
> org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:149)
>     at org.apache.spark.SparkContext.<init>(SparkContext.scala:500)
>     at org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2256)
>     at
> org.apache.spark.sql.SparkSession$Builder$$anonfun$8.apply(SparkSession.scala:831)
>     at
> org.apache.spark.sql.SparkSession$Builder$$anonfun$8.apply(SparkSession.scala:823)
>     at scala.Option.getOrElse(Option.scala:121)
>     at
> org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:823)
>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>     at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>     at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>     at java.lang.reflect.Method.invoke(Method.java:497)
>     at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38)
>     at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33)
>
>
> --
> Best Regards
>
> Jeff Zhang
>



-- 
Best Regards

Jeff Zhang

Reply via email to