For some cases, zeppelin side's configuration will override
spark-defaults.conf, but if spark.master is yarn-cluster in
spark-defaults.conf, then this kind of inconsistency will happen.  I  will
solve this issue in https://issues.apache.org/jira/browse/ZEPPELIN-1263

On Wed, Aug 3, 2016 at 9:38 PM, Jongyoul Lee <jongy...@gmail.com> wrote:

> In that case, zeppelin side's one will overrider spark-defaults.conf
>
> On Tue, Aug 2, 2016 at 11:17 AM, Jeff Zhang <zjf...@gmail.com> wrote:
>
>> Find the issue, this is due to the inconsistency between spark.master in
>> spark-defaults.conf spark.master in zeppelin side.
>>
>> On Tue, Aug 2, 2016 at 9:20 AM, Jeff Zhang <zjf...@gmail.com> wrote:
>>
>>>
>>> I follow the the instruction in this PR to build zeppelin and try to run
>>> in yarn-client mode, but hit very weird issue. It seems the SPARK_HOME is
>>> messed up. Does anyone hit this issue or can run zeppelin on spark 2.0 in
>>> yarn-client mode successfully ?
>>>
>>> https://github.com/apache/zeppelin/pull/1195
>>>
>>> ERROR [2016-08-02 09:05:27,771] ({pool-4-thread-5}
>>> Logging.scala[logError]:91) - Error initializing SparkContext.
>>> java.lang.IllegalStateException: Library directory
>>> '/Users/jzhang/Temp/hadoop_tmp/nm-local-dir/usercache/jzhang/appcache/application_1470097474471_0011/container_1470097474471_0011_01_000001/assembly/target/scala-2.11/jars'
>>> does not exist; make sure Spark is built.
>>>     at
>>> org.apache.spark.launcher.CommandBuilderUtils.checkState(CommandBuilderUtils.java:248)
>>>     at
>>> org.apache.spark.launcher.CommandBuilderUtils.findJarsDir(CommandBuilderUtils.java:368)
>>>     at
>>> org.apache.spark.launcher.YarnCommandBuilderUtils$.findJarsDir(YarnCommandBuilderUtils.scala:38)
>>>     at
>>> org.apache.spark.deploy.yarn.Client.prepareLocalResources(Client.scala:500)
>>>     at
>>> org.apache.spark.deploy.yarn.Client.createContainerLaunchContext(Client.scala:834)
>>>     at
>>> org.apache.spark.deploy.yarn.Client.submitApplication(Client.scala:167)
>>>     at
>>> org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:56)
>>>     at
>>> org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:149)
>>>     at org.apache.spark.SparkContext.<init>(SparkContext.scala:500)
>>>     at
>>> org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2256)
>>>     at
>>> org.apache.spark.sql.SparkSession$Builder$$anonfun$8.apply(SparkSession.scala:831)
>>>     at
>>> org.apache.spark.sql.SparkSession$Builder$$anonfun$8.apply(SparkSession.scala:823)
>>>     at scala.Option.getOrElse(Option.scala:121)
>>>     at
>>> org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:823)
>>>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>>>     at
>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>>     at
>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>>     at java.lang.reflect.Method.invoke(Method.java:497)
>>>     at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:38)
>>>     at org.apache.zeppelin.spark.Utils.invokeMethod(Utils.java:33)
>>>
>>>
>>> --
>>> Best Regards
>>>
>>> Jeff Zhang
>>>
>>
>>
>>
>> --
>> Best Regards
>>
>> Jeff Zhang
>>
>
>
>
> --
> 이종열, Jongyoul Lee, 李宗烈
> http://madeng.net
>



-- 
Best Regards

Jeff Zhang

Reply via email to