Hi, Mridul. > I am currently on Python 3.11.6, java 8.
For the above, I added `Python 3.11 support` at Apache Spark 3.4.0. That's exactly one of my reasons why I wanted to do the EOL release of Apache Spark 3.3.4. https://issues.apache.org/jira/browse/SPARK-41454 (Support Python 3.11) Thanks, Dongjoon. On Mon, Dec 11, 2023 at 12:22 PM Mridul Muralidharan <mri...@gmail.com> wrote: > > I am seeing a bunch of python related (43) failures in the sql module (for > example [1]) ... I am currently on Python 3.11.6, java 8. > Not sure if ubuntu modified anything from under me, thoughts ? > > I am currently testing this against an older branch to make sure it is not > an issue with my desktop. > > Regards, > Mridul > > > [1] > > > org.apache.spark.sql.IntegratedUDFTestUtils.shouldTestGroupedAggPandasUDFs > was false (QueryCompilationErrorsSuite.scala:112) > Traceback (most recent call last): > File > "/home/mridul/work/apache/vote/spark/python/pyspark/serializers.py", line > 458, in dumps > return cloudpickle.dumps(obj, pickle_protocol) > ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 73, in dumps > cp.dump(obj) > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 602, in dump > return Pickler.dump(self, obj) > ^^^^^^^^^^^^^^^^^^^^^^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 692, in reducer_override > return self._function_reduce(obj) > ^^^^^^^^^^^^^^^^^^^^^^^^^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 565, in _function_reduce > return self._dynamic_function_reduce(obj) > ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 546, in _dynamic_function_reduce > state = _function_getstate(func) > ^^^^^^^^^^^^^^^^^^^^^^^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 157, in _function_getstate > f_globals_ref = _extract_code_globals(func.__code__) > ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle.py", > line 334, in _extract_code_globals > out_names = {names[oparg]: None for _, oparg in _walk_global_ops(co)} > ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle.py", > line 334, in <dictcomp> > out_names = {names[oparg]: None for _, oparg in _walk_global_ops(co)} > ~~~~~^^^^^^^ > IndexError: tuple index out of range > Traceback (most recent call last): > File > "/home/mridul/work/apache/vote/spark/python/pyspark/serializers.py", line > 458, in dumps > return cloudpickle.dumps(obj, pickle_protocol) > ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 73, in dumps > cp.dump(obj) > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 602, in dump > return Pickler.dump(self, obj) > ^^^^^^^^^^^^^^^^^^^^^^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 692, in reducer_override > return self._function_reduce(obj) > ^^^^^^^^^^^^^^^^^^^^^^^^^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 565, in _function_reduce > return self._dynamic_function_reduce(obj) > ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 546, in _dynamic_function_reduce > state = _function_getstate(func) > ^^^^^^^^^^^^^^^^^^^^^^^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 157, in _function_getstate > f_globals_ref = _extract_code_globals(func.__code__) > ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle.py", > line 334, in _extract_code_globals > out_names = {names[oparg]: None for _, oparg in _walk_global_ops(co)} > ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle.py", > line 334, in <dictcomp> > out_names = {names[oparg]: None for _, oparg in _walk_global_ops(co)} > ~~~~~^^^^^^^ > IndexError: tuple index out of range > > During handling of the above exception, another exception occurred: > > Traceback (most recent call last): > File "<string>", line 1, in <module> > File > "/home/mridul/work/apache/vote/spark/python/pyspark/serializers.py", line > 468, in dumps > raise pickle.PicklingError(msg) > _pickle.PicklingError: Could not serialize object: IndexError: tuple index > out of range > - UNSUPPORTED_FEATURE: Using Python UDF with unsupported join condition > *** FAILED *** > > > > On Sun, Dec 10, 2023 at 9:05 PM L. C. Hsieh <vii...@gmail.com> wrote: > >> +1 >> >> On Sun, Dec 10, 2023 at 6:15 PM Kent Yao <y...@apache.org> wrote: >> > >> > +1(non-binding >> > >> > Kent Yao >> > >> > Yuming Wang <yumw...@apache.org> 于2023年12月11日周一 09:33写道: >> > > >> > > +1 >> > > >> > > On Mon, Dec 11, 2023 at 5:55 AM Dongjoon Hyun <dongj...@apache.org> >> wrote: >> > >> >> > >> +1 >> > >> >> > >> Dongjoon >> > >> >> > >> On 2023/12/08 21:41:00 Dongjoon Hyun wrote: >> > >> > Please vote on releasing the following candidate as Apache Spark >> version >> > >> > 3.3.4. >> > >> > >> > >> > The vote is open until December 15th 1AM (PST) and passes if a >> majority +1 >> > >> > PMC votes are cast, with a minimum of 3 +1 votes. >> > >> > >> > >> > [ ] +1 Release this package as Apache Spark 3.3.4 >> > >> > [ ] -1 Do not release this package because ... >> > >> > >> > >> > To learn more about Apache Spark, please see >> https://spark.apache.org/ >> > >> > >> > >> > The tag to be voted on is v3.3.4-rc1 (commit >> > >> > 18db204995b32e87a650f2f09f9bcf047ddafa90) >> > >> > https://github.com/apache/spark/tree/v3.3.4-rc1 >> > >> > >> > >> > The release files, including signatures, digests, etc. can be >> found at: >> > >> > >> > >> > https://dist.apache.org/repos/dist/dev/spark/v3.3.4-rc1-bin/ >> > >> > >> > >> > >> > >> > Signatures used for Spark RCs can be found in this file: >> > >> > >> > >> > https://dist.apache.org/repos/dist/dev/spark/KEYS >> > >> > >> > >> > >> > >> > The staging repository for this release can be found at: >> > >> > >> > >> > >> https://repository.apache.org/content/repositories/orgapachespark-1451/ >> > >> > >> > >> > >> > >> > The documentation corresponding to this release can be found at: >> > >> > >> > >> > https://dist.apache.org/repos/dist/dev/spark/v3.3.4-rc1-docs/ >> > >> > >> > >> > >> > >> > The list of bug fixes going into 3.3.4 can be found at the >> following URL: >> > >> > >> > >> > https://issues.apache.org/jira/projects/SPARK/versions/12353505 >> > >> > >> > >> > >> > >> > This release is using the release script of the tag v3.3.4-rc1. >> > >> > >> > >> > >> > >> > FAQ >> > >> > >> > >> > >> > >> > ========================= >> > >> > >> > >> > How can I help test this release? >> > >> > >> > >> > ========================= >> > >> > >> > >> > >> > >> > >> > >> > If you are a Spark user, you can help us test this release by >> taking >> > >> > >> > >> > an existing Spark workload and running on this release candidate, >> then >> > >> > >> > >> > reporting any regressions. >> > >> > >> > >> > >> > >> > >> > >> > If you're working in PySpark you can set up a virtual env and >> install >> > >> > >> > >> > the current RC and see if anything important breaks, in the >> Java/Scala >> > >> > >> > >> > you can add the staging repository to your projects resolvers and >> test >> > >> > >> > >> > with the RC (make sure to clean up the artifact cache before/after >> so >> > >> > >> > >> > you don't end up building with a out of date RC going forward). >> > >> > >> > >> > >> > >> > >> > >> > =========================================== >> > >> > >> > >> > What should happen to JIRA tickets still targeting 3.3.4? >> > >> > >> > >> > =========================================== >> > >> > >> > >> > >> > >> > >> > >> > The current list of open tickets targeted at 3.3.4 can be found at: >> > >> > >> > >> > https://issues.apache.org/jira/projects/SPARK and search for >> "Target >> > >> > Version/s" = 3.3.4 >> > >> > >> > >> > >> > >> > Committers should look at those and triage. Extremely important bug >> > >> > >> > >> > fixes, documentation, and API tweaks that impact compatibility >> should >> > >> > >> > >> > be worked on immediately. Everything else please retarget to an >> > >> > >> > >> > appropriate release. >> > >> > >> > >> > >> > >> > >> > >> > ================== >> > >> > >> > >> > But my bug isn't fixed? >> > >> > >> > >> > ================== >> > >> > >> > >> > >> > >> > >> > >> > In order to make timely releases, we will typically not hold the >> > >> > >> > >> > release unless the bug in question is a regression from the >> previous >> > >> > >> > >> > release. That being said, if there is something which is a >> regression >> > >> > >> > >> > that has not been correctly targeted please ping me or a committer >> to >> > >> > >> > >> > help target the issue. >> > >> > >> > >> >> > >> --------------------------------------------------------------------- >> > >> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >> > >> >> > >> > --------------------------------------------------------------------- >> > To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >> > >> >> --------------------------------------------------------------------- >> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >> >>