Jerzy J. Gangi created ZEPPELIN-1741:
----------------------------------------

             Summary: JAR's specified with spark.jars in spark-defaults.conf 
does not affect %pyspark interpreter
                 Key: ZEPPELIN-1741
                 URL: https://issues.apache.org/jira/browse/ZEPPELIN-1741
             Project: Zeppelin
          Issue Type: Bug
          Components: Interpreters
    Affects Versions: 0.6.2
         Environment: Zeppelin 0.6.2
Spark 1.6.2
            Reporter: Jerzy J. Gangi


If you specify JAR's with `spark.jars` in your `spark-defaults.conf`, the 
`%pyspark` interpreter will not load these JARs.

Currently, a note in the Spark interpreter documentation says, "Note that 
adding jar to pyspark is only availabe via `%dep` interpreter at the moment."

This is undesirable for two reasons:

1) `%dep` is deprecated
2) Sysadmins configure `spark-defaults.conf`, and expect these settings to be 
honored however Spark is executed.

As a Zeppelin user, I expect that if I configure JAR's in 
`spark-defaults.conf`, these JAR's will be available when the %pyspark 
interpreter is run.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to