Good!! On Tue, Apr 14, 2015 at 11:14 PM, Charmee Patel <[email protected]> wrote:
> Hi, > > Thanks. > > I had pulled the code from https://github.com/NFLabs/zeppelin repository > when it had not moved to apache/incubator-zeppelin yet and I kept pulling > from same repo. Synced up my version with latest on > apache/incubator-zeppelin and it is working now. > > -Charmee > > On Mon, Apr 13, 2015 at 9:12 PM Jongyoul Lee <[email protected]> wrote: > >> Hi, >> >> I don't know what version you use exactly because com.nflabs.zeppelin >> moves to org.apache.zeppelin. You don't need to set SPARK_YARN_JAR at the >> latest master. Could you please check this out? And I've test yarn-client >> mode on 2.5.0-cdh5.3.0 with spark 1.3 but didn't do it with spark 1.2. My >> build script is >> >> mvn clean package -Pspark-1.3 -Dhadoop.version=2.5.0-cdh5.3.0 >> -Phadoop-2.4 -DskipTests -Pyarn -Pbuild-distr >> >> Regards, >> Jongyoul Lee >> >> On Mon, Apr 13, 2015 at 11:40 PM, Charmee Patel <[email protected]> >> wrote: >> >>> Hi, >>> >>> I am trying to get Zeppelin work in yarn-client and yarn-submit >>> mode. There are some conflicting notes in the email distros about how to >>> get Zeppelin to work on yarn-client mode. I have tried a few different >>> things but none have worked for me so far. >>> >>> I have CDH 5.3 . Here is what I have so far >>> >>> 1. Built Zeppelin on local (OSX) using >>> 1. mvn clean package -Pspark-1.2 -Dhadoop.version=2.5.0-cdh5.3.0 >>> -Phadoop-2.4 -DskipTests -Pyarn >>> 2. Generated a distribution package and deployed to edge node of my >>> cluster using >>> 1. mvn clean package -P build-distr -DskipTests >>> 3. At this point local mode works fine >>> 4. To get Yarn mode working >>> 1. I set Hadoop conf dir in zeppelin-env.sh and set Yarn >>> Master=yarn-client >>> I get an exception in my logs >>> >>> ERROR ({pool-2-thread-5} Job.java[run]:165) - Job failed >>> com.nflabs.zeppelin.interpreter.InterpreterException: >>> org.apache.thrift.TApplicationException: Internal error processing >>> open >>> >>> 2. Set Spark_Yarn_Jar but got same exception as above >>> 3. Copied my spark assembly jar in interpreter/spark directory >>> but that did not work either >>> 4. Also set Spark Yarn Jar and Spark Home in interpreter UI but >>> that did not work >>> 5. I took Spark Yarn Jar out and let Zeppelin use Spark that >>> comes bundled with it >>> 1. This seemed to work initially, but as soon as I call Spark >>> Action (collect/count etc) I get this exception >>> >>> java.lang.RuntimeException: Error in configuring object at >>> org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:93) >>> at org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:64) >>> at >>> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:117) >>> at org.apache.spark.rdd.HadoopRDD.getInputFormat(HadoopRDD.scala:182) >>> >>> Any pointers on I might have missed in configuring zeppelin to work with >>> yarn-client mode? >>> >>> Thanks, >>> Charmee >>> >> >> >> >> -- >> 이종열, Jongyoul Lee, 李宗烈 >> http://madeng.net >> > -- 이종열, Jongyoul Lee, 李宗烈 http://madeng.net
