Does the spark.files option only work in 0.6?  Any workaround for earlier
versions to load external jars?

On Tue, Oct 20, 2015 at 6:52 PM, moon soo Lee <m...@apache.org> wrote:

> Sourav,
>
> There're couple of ways to add external jar when Zeppelin (0.6.0-SNAPSHOT)
> uses spark-submit command.
>
> 1. Using %dep interpreter, as Vinay mentioned.
>     eg)
>        %dep
>         z.load("group:artifact:version")
>
>         %spark
>         import ....
>
> 2. By adding spark.files property at SPARK_HOME/conf/spark-defaults.conf
>      eg)
>         spark.files  /path/to/my.jar
>
> 3. By exporting SPARK_SUBMIT_OPTIONS env variable in
> ZEPPELIN_HOME/conf/zeppelin-env.sh
>      eg)
>         export SPARK_SUBMIT_OPTIONS="--packages group:artifact:version"
>      note) does not work for pyspark yet.
> https://issues.apache.org/jira/browse/ZEPPELIN-339
>
> Hope this helps.
>
> Best,
> moon
>
> On Wed, Oct 21, 2015 at 12:55 AM Vinay Shukla <vinayshu...@gmail.com>
> wrote:
>
>> Saurav,
>>
>>
>> Agree this would be a useful feature. Right now Zeppelin can import
>> dependencies from Maven with %dep interpreter. But this does not understand
>> spark-packages.
>>
>>
>> -Vinay
>>
>>
>> On Tue, Oct 20, 2015 at 6:54 AM, Sourav Mazumder <
>> sourav.mazumde...@gmail.com> wrote:
>>
>>> Looks like right now there is no way one can pass additional jar files
>>> to spark_submit from Zeppelin.The same works fine if I am not using
>>> spark_submit option (by not specifying spark_home).
>>>
>>> When I checked the code interpreter.sh I found that for the class path
>>> it only passes the zeppelin-spark*.jar available in
>>> zeppelin_home/interpreter/spark directory.
>>>
>>> I suggest to put this as a bug/enhancement. The solution should be
>>> pretty easy by making some small changes in interpreter.sh (I've done the
>>> same and could make it work with some external_lib folder under
>>> zeppelin_home/interpreter/spark directory).
>>>
>>> Regards,
>>> Sourav
>>>
>>
>>

Reply via email to