Thanks for sharing a problem and solution! Best, moon On 2015년 6월 25일 (목) at 오전 2:00 김동경 <style9...@gmail.com> wrote:
> I figured out the problem. > Once Spark job is accepted, it is trying to access driver which is running > on the machine which in zeppelin resides. > That means, yarn cluster should be able to access to zeppelin machine. > > I configured hostname of my pc in /etc/hosts of yarn clusters and it > worked. > > I hope this help anyone who suffers from the same thing. > > Thanks > Regards > Dongkyoung. > > > > 2015-06-25 17:12 GMT+09:00 김동경 <style9...@gmail.com>: > >> I want to make zeppelin work with remote Yarn cluster. >> >> When I run any spark paragraph, spark job is submitted to Yarn cluster >> and accepted. >> But it does not change to running mode. >> >> Please refer to my configuration. >> >> 1. zeppelin-env.sh >> >> export MASTER=yarn-client >> export >> SPARK_YARN_JAR=/home/style95/lib/spark-assembly-1.2.1-hadoop2.4.0.jar >> export HADOOP_CONF_DIR=/home/style95/hadoop-conf >> >> >> 2. hadoop-conf/core-site.xml >> >> <configuration> >> <property> >> <name>fs.default.name</name> >> <value>hdfs://10.251.53.42:8020</value> >> <final>true</final> >> </property> >> </configuration> >> >> >> 3. hadoop-conf/yarn-site.xml >> >> <configuration> >> <property> >> <name>yarn.nodemanager.aux-services</name> >> <value>mapreduce_shuffle</value> >> </property> >> >> <property> >> <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name> >> <value>org.apache.hadoop.mapred.ShuffleHandler</value> >> </property> >> <property> >> <name>yarn.resourcemanager.hostname</name> >> <value>10.251.53.42</value> >> </property> >> </configuration> >> >> >> There are only above two files in hadoop-conf directory. >> >> Do I miss anything? >> Do I have to declare any other environment variables? >> >> There is no permission configured on yarn cluster. >> The version of yarn cluster is "hadoop-2.6.0" >> >> I already tried with various version of hadoop and spark profiles and got >> nothing. >> Tried versions: >> - spark 1.2 with hadoop 2.4 >> - spark 1.3 with hadoop 2.4 >> - spark 1.3 with hadoop 2.6 >> >> >> Since my pc is behind of NAT, my pc is not reachable from yarn cluster. >> Could it affect? >> >> Please let me get out of it >> >> Thanks >> Regards >> Dongkyoung. >> >> >