Does the spark.files option only work in 0.6? Any workaround for earlier versions to load external jars?
On Tue, Oct 20, 2015 at 6:52 PM, moon soo Lee <m...@apache.org> wrote: > Sourav, > > There're couple of ways to add external jar when Zeppelin (0.6.0-SNAPSHOT) > uses spark-submit command. > > 1. Using %dep interpreter, as Vinay mentioned. > eg) > %dep > z.load("group:artifact:version") > > %spark > import .... > > 2. By adding spark.files property at SPARK_HOME/conf/spark-defaults.conf > eg) > spark.files /path/to/my.jar > > 3. By exporting SPARK_SUBMIT_OPTIONS env variable in > ZEPPELIN_HOME/conf/zeppelin-env.sh > eg) > export SPARK_SUBMIT_OPTIONS="--packages group:artifact:version" > note) does not work for pyspark yet. > https://issues.apache.org/jira/browse/ZEPPELIN-339 > > Hope this helps. > > Best, > moon > > On Wed, Oct 21, 2015 at 12:55 AM Vinay Shukla <vinayshu...@gmail.com> > wrote: > >> Saurav, >> >> >> Agree this would be a useful feature. Right now Zeppelin can import >> dependencies from Maven with %dep interpreter. But this does not understand >> spark-packages. >> >> >> -Vinay >> >> >> On Tue, Oct 20, 2015 at 6:54 AM, Sourav Mazumder < >> sourav.mazumde...@gmail.com> wrote: >> >>> Looks like right now there is no way one can pass additional jar files >>> to spark_submit from Zeppelin.The same works fine if I am not using >>> spark_submit option (by not specifying spark_home). >>> >>> When I checked the code interpreter.sh I found that for the class path >>> it only passes the zeppelin-spark*.jar available in >>> zeppelin_home/interpreter/spark directory. >>> >>> I suggest to put this as a bug/enhancement. The solution should be >>> pretty easy by making some small changes in interpreter.sh (I've done the >>> same and could make it work with some external_lib folder under >>> zeppelin_home/interpreter/spark directory). >>> >>> Regards, >>> Sourav >>> >> >>