For some cases, zeppelin side's configuration will override spark-defaults.conf, but if spark.master is yarn-cluster in spark-defaults.conf, then this kind of inconsistency will happen. I will solve this issue in https://issues.apache.org/jira/browse/ZEPPELIN-1263
On Wed, Aug 3, 2016 at 9:38 PM, Jongyoul Lee <jongy...@gmail.com> wrote: > In that case, zeppelin side's one will overrider spark-defaults.conf > > On Tue, Aug 2, 2016 at 11:17 AM, Jeff Zhang <zjf...@gmail.com> wrote: > >> Find the issue, this is due to the inconsistency between spark.master in >> spark-defaults.conf spark.master in zeppelin side. >> >> On Tue, Aug 2, 2016 at 9:20 AM, Jeff Zhang <zjf...@gmail.com> wrote: >> >>> >>> I follow the the instruction in this PR to build zeppelin and try to run >>> in yarn-client mode, but hit very weird issue. It seems the SPARK_HOME is >>> messed up. Does anyone hit this issue or can run zeppelin on spark 2.0 in >>> yarn-client mode successfully ? >>> >>> https://github.com/apache/zeppelin/pull/1195 >>> >>> ERROR [2016-08-02 09:05:27,771] ({pool-4-thread-5} >>> Logging.scala[logError]:91) - Error initializing SparkContext. >>> java.lang.IllegalStateException: Library directory >>> '/Users/jzhang/Temp/hadoop_tmp/nm-local-dir/usercache/jzhang/appcache/application_1470097474471_0011/container_1470097474471_0011_01_000001/assembly/target/scala-2.11/jars' >>> does not exist; make sure Spark is built. >>> at >>> org.apache.spark.launcher.CommandBuilderUtils.checkState(CommandBuilderUtils.java:248) >>> at >>> org.apache.spark.launcher.CommandBuilderUtils.findJarsDir(CommandBuilderUtils.java:368) >>> at >>> org.apache.spark.launcher.YarnCommandBuilderUtils$.findJarsDir(YarnCommandBuilderUtils.scala:38) >>> at >>> org.apache.spark.deploy.yarn.Client.prepareLocalResources(Client.scala:500) >>> at >>> org.apache.spark.deploy.yarn.Client.createContainerLaunchContext(Client.scala:834) >>> at >>> org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:167) >>> at >>> org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:56) >>> at >>> org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:149) >>> at org.apache.spark.SparkContext.<init>(SparkContext.scala:500) >>> at >>> org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2256) >>> at >>> org.apache.spark.sql.SparkSession$Builder$$anonfun$8.apply(SparkSession.scala:831) >>> at >>> org.apache.spark.sql.SparkSession$Builder$$anonfun$8.apply(SparkSession.scala:823) >>> at scala.Option.getOrElse(Option.scala:121) >>> at >>> org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:823) >>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >>> at >>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >>> at >>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >>> at java.lang.reflect.Method.invoke(Method.java:497) >>> at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38) >>> at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33) >>> >>> >>> -- >>> Best Regards >>> >>> Jeff Zhang >>> >> >> >> >> -- >> Best Regards >> >> Jeff Zhang >> > > > > -- > 이종열, Jongyoul Lee, 李宗烈 > http://madeng.net > -- Best Regards Jeff Zhang