In zeppelin, what is the equivalent to adding jars in a pyspark call? Such as running pyspark with the elasticsearch-hadoop jar
./bin/pyspark --master local[2] --jars jars/elasticsearch-hadoop-2.1.0.Beta2.jar My assumption is that loading something like this inside a %dep is pointless, since those dependencies would only live in the %spark scala world (the spark jvm). In zeppelin - pyspark spawns a separate process. Also how is the interpreters “spark.home” used? How is it different that the “SPARK_HOME” zeppelin-env.sh And finally – how are args used in the interpreter? (what uses them)? Thank you. Jeff