Github user holdenk commented on a diff in the pull request:

    https://github.com/apache/spark/pull/19643#discussion_r151839304
  
    --- Diff: python/pyspark/context.py ---
    @@ -860,6 +860,23 @@ def addPyFile(self, path):
                 import importlib
                 importlib.invalidate_caches()
     
    +    def addJar(self, path, addToCurrentClassLoader=False):
    +        """
    +        Adds a JAR dependency for Spark tasks to be executed in the future.
    +        The `path` passed can be either a local file, a file in HDFS (or 
other Hadoop-supported
    +        filesystems), an HTTP, HTTPS or FTP URI, or local:/path for a file 
on every worker node.
    +        If `addToCurrentClassLoader` is true, add the jar to the current 
threads' class loader
    +        in the backing JVM. In general adding to the current threads' 
class loader will impact all
    +        other application threads unless they have explicitly changed 
their class loader.
    --- End diff --
    
    So we currently use `.. note:: DeveloperApi` to indicate it's a developer 
API (see ml/pipeline and friends for an example).


---

---------------------------------------------------------------------
To unsubscribe, e-mail: reviews-unsubscr...@spark.apache.org
For additional commands, e-mail: reviews-h...@spark.apache.org

Reply via email to