Thanks a lot for clarifying that.

Thanks, Alex

On Oct 20, 2018 7:15 AM, "Jhon Anderson Cardenas Diaz" <
jhonderson2...@gmail.com> wrote:

> Hi, You can specify it in the zeppelin-env.sh, or in the Dockerfile.
>
> Zeppelin will look for that variable first in the interpreter settings,
> and if it does not find it, it will look for it on zeppelin environment
> variables; so you can specify it in both sides, but as it does not change
> frenquently it is better on zeppelin environment variable.
>
> El sáb., 20 oct. 2018 a las 0:25, Alex Dzhagriev (<dzh...@gmail.com>)
> escribió:
>
>> Thanks for the quick reply. Should I specify it to the Zeppelin process
>> or the Spark interpreter?
>>
>> Thanks, Alex.
>>
>> On Fri, Oct 19, 2018 at 4:53 PM Jeff Zhang <zjf...@gmail.com> wrote:
>>
>>> You need to specify SPARK_HOME which is where spark installed.
>>>
>>>
>>> Alex Dzhagriev <dzh...@gmail.com>于2018年10月20日周六 上午3:12写道:
>>>
>>>> Hello,
>>>>
>>>> I have a remote Spark cluster and I'm trying to use it by setting the
>>>> spark interpreter property:
>>>>
>>>> master spark://spark-cluster-master:7077, however I'm getting the
>>>> following error:
>>>>
>>>> java.lang.RuntimeException: SPARK_HOME is not specified in
>>>> interpreter-setting for non-local mode, if you specify it in
>>>> zeppelin-env.sh, please move that into interpreter setting
>>>>
>>>> version: Docker Image 0.8.0
>>>>
>>>> Thanks, Alex.
>>>>
>>>

Reply via email to