According the log, spark interpreter is started in yarn-client as
application_1455038611898_0015. Could you check the yarn application of
this app ?

On Thu, Feb 18, 2016 at 1:26 AM, Abhi Basu <9000r...@gmail.com> wrote:

> Additional info, installed Zeppelin 0.56 using binaries on cdh5.1 /spark
> 1.5.0
>
> Any help is appreciated.
>
> Thanks,
>
> Abhi
>
> On Wed, Feb 17, 2016 at 9:07 AM, Abhi Basu <9000r...@gmail.com> wrote:
>
>> Logs attached. Am I supposed to edit the spark location in the zeppelin
>> config file? All I have changed is the hadoop conf folder.
>>
>> Thanks,
>>
>> Abhi
>>
>> On Tue, Feb 16, 2016 at 5:29 PM, Jeff Zhang <zjf...@gmail.com> wrote:
>>
>>> Can you check zeppelin log to confirm whether it is running in
>>> yarn-client mode ? I suspect it is still in local mode. Spark require
>>> python version of driver and executor to be the same. In your case it
>>> should fail if driver is python2.7 while executor is python 2.6
>>>
>>> On Wed, Feb 17, 2016 at 9:03 AM, Abhi Basu <9000r...@gmail.com> wrote:
>>>
>>>> I have a 6 node cluster and 1 edge node to access. The edge node has
>>>> Python 2.7 + NLTK + other libraries + hadoop client and Zeppelin installed.
>>>> All hadoop nodes have Python 2.6 and no other additional libraries.
>>>>
>>>> Running Zeppelin and my python code (with NLTK) is running under
>>>> pyspark interpreter fine. It must be running locally as I have not
>>>> distributed the python libraries to the other nodes yet. I dont see any
>>>> errors in my Yarn logs either.
>>>>
>>>> This is my interpreter setup. Can you please tell  me how this is
>>>> working?
>>>>
>>>> Also, if it is working locally, how to distribute over multiple nodes?
>>>>
>>>>
>>>> Thanks,
>>>>
>>>> Abhi
>>>>
>>>> spark %spark (default), %pyspark, %sql, %dep edit  restart  remove
>>>> Properties
>>>> namevalue
>>>> args
>>>> master yarn-client
>>>> spark.app.name Zeppelin-App
>>>> spark.cores.max 4
>>>> spark.executor.memory 1024m
>>>> zeppelin.dep.additionalRemoteRepository spark-packages,
>>>> http://dl.bintray.com/spark-packages/maven,false;
>>>> zeppelin.dep.localrepo local-repo
>>>> zeppelin.pyspark.python /usr/local/bin/python2.7
>>>> zeppelin.spark.concurrentSQL true
>>>> zeppelin.spark.maxResult 1000
>>>> zeppelin.spark.useHiveContext true
>>>>
>>>>
>>>> --
>>>> Abhi Basu
>>>>
>>>
>>>
>>>
>>> --
>>> Best Regards
>>>
>>> Jeff Zhang
>>>
>>
>>
>>
>> --
>> Abhi Basu
>>
>
>
>
> --
> Abhi Basu
>



-- 
Best Regards

Jeff Zhang

Reply via email to