According the log, spark interpreter is started in yarn-client as application_1455038611898_0015. Could you check the yarn application of this app ?
On Thu, Feb 18, 2016 at 1:26 AM, Abhi Basu <9000r...@gmail.com> wrote: > Additional info, installed Zeppelin 0.56 using binaries on cdh5.1 /spark > 1.5.0 > > Any help is appreciated. > > Thanks, > > Abhi > > On Wed, Feb 17, 2016 at 9:07 AM, Abhi Basu <9000r...@gmail.com> wrote: > >> Logs attached. Am I supposed to edit the spark location in the zeppelin >> config file? All I have changed is the hadoop conf folder. >> >> Thanks, >> >> Abhi >> >> On Tue, Feb 16, 2016 at 5:29 PM, Jeff Zhang <zjf...@gmail.com> wrote: >> >>> Can you check zeppelin log to confirm whether it is running in >>> yarn-client mode ? I suspect it is still in local mode. Spark require >>> python version of driver and executor to be the same. In your case it >>> should fail if driver is python2.7 while executor is python 2.6 >>> >>> On Wed, Feb 17, 2016 at 9:03 AM, Abhi Basu <9000r...@gmail.com> wrote: >>> >>>> I have a 6 node cluster and 1 edge node to access. The edge node has >>>> Python 2.7 + NLTK + other libraries + hadoop client and Zeppelin installed. >>>> All hadoop nodes have Python 2.6 and no other additional libraries. >>>> >>>> Running Zeppelin and my python code (with NLTK) is running under >>>> pyspark interpreter fine. It must be running locally as I have not >>>> distributed the python libraries to the other nodes yet. I dont see any >>>> errors in my Yarn logs either. >>>> >>>> This is my interpreter setup. Can you please tell me how this is >>>> working? >>>> >>>> Also, if it is working locally, how to distribute over multiple nodes? >>>> >>>> >>>> Thanks, >>>> >>>> Abhi >>>> >>>> spark %spark (default), %pyspark, %sql, %dep edit restart remove >>>> Properties >>>> namevalue >>>> args >>>> master yarn-client >>>> spark.app.name Zeppelin-App >>>> spark.cores.max 4 >>>> spark.executor.memory 1024m >>>> zeppelin.dep.additionalRemoteRepository spark-packages, >>>> http://dl.bintray.com/spark-packages/maven,false; >>>> zeppelin.dep.localrepo local-repo >>>> zeppelin.pyspark.python /usr/local/bin/python2.7 >>>> zeppelin.spark.concurrentSQL true >>>> zeppelin.spark.maxResult 1000 >>>> zeppelin.spark.useHiveContext true >>>> >>>> >>>> -- >>>> Abhi Basu >>>> >>> >>> >>> >>> -- >>> Best Regards >>> >>> Jeff Zhang >>> >> >> >> >> -- >> Abhi Basu >> > > > > -- > Abhi Basu > -- Best Regards Jeff Zhang