Hyukjin Kwon commented on SPARK-23965:

That's easy to work around in any case. We can simply make a map for Py4J 
version to Spark's version in application side and use it? 
Also, off the record, it usually works properly with different Py4J in my 

> make python/py4j-src-0.x.y.zip file name Spark version-independent
> ------------------------------------------------------------------
>                 Key: SPARK-23965
>                 URL: https://issues.apache.org/jira/browse/SPARK-23965
>             Project: Spark
>          Issue Type: Improvement
>          Components: PySpark
>    Affects Versions: 2.2.1, 2.3.0, 2.4.0
>            Reporter: Ruslan Dautkhanov
>            Priority: Major
> After each Spark release (that's normally packaged with slightly newer 
> version of py4j), we have to adjust our PySpark applications PYTHONPATH to 
> point to correct version of python/py4j-src-0.9.2.zip. 
> Change to python/py4j-src-0.9.2.zip to python/py4j-src-0.9.6.zip, next 
> release to something else etc. 
> Possible solutions. Would be great to either
>  - rename `python/py4j-src-0.x.y.zip` to `python/py4j-src-latest.zip` or 
> `python/py4j-src-current.zip`
>  - or make a symlink in Spark distributed `py4j-src-current.zip` to whatever 
> version Spark is shipped with.
> In either case, if this would be solved, we wouldn't have to adjust 
> PYTHONPATH during upgrades like Spark 2.2 to 2.3.. 
> Thanks.

This message was sent by Atlassian JIRA

To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to