Re: [VOTE] Release Spark 3.3.4 (RC1)
Thank you all. This vote passed. Let me conclude. Dongjoon On 2023/12/11 23:58:28 Malcolm Decuire wrote: > +1 > > On Mon, Dec 11, 2023 at 6:21 PM Yang Jie wrote: > > > +1 > > > > On 2023/12/11 03:03:39 "L. C. Hsieh" wrote: > > > +1 > > > > > > On Sun, Dec 10, 2023 at 6:15 PM Kent Yao wrote: > > > > > > > > +1(non-binding > > > > > > > > Kent Yao > > > > > > > > Yuming Wang 于2023年12月11日周一 09:33写道: > > > > > > > > > > +1 > > > > > > > > > > On Mon, Dec 11, 2023 at 5:55 AM Dongjoon Hyun > > wrote: > > > > >> > > > > >> +1 > > > > >> > > > > >> Dongjoon > > > > >> > > > > >> On 2023/12/08 21:41:00 Dongjoon Hyun wrote: > > > > >> > Please vote on releasing the following candidate as Apache Spark > > version > > > > >> > 3.3.4. > > > > >> > > > > > >> > The vote is open until December 15th 1AM (PST) and passes if a > > majority +1 > > > > >> > PMC votes are cast, with a minimum of 3 +1 votes. > > > > >> > > > > > >> > [ ] +1 Release this package as Apache Spark 3.3.4 > > > > >> > [ ] -1 Do not release this package because ... > > > > >> > > > > > >> > To learn more about Apache Spark, please see > > https://spark.apache.org/ > > > > >> > > > > > >> > The tag to be voted on is v3.3.4-rc1 (commit > > > > >> > 18db204995b32e87a650f2f09f9bcf047ddafa90) > > > > >> > https://github.com/apache/spark/tree/v3.3.4-rc1 > > > > >> > > > > > >> > The release files, including signatures, digests, etc. can be > > found at: > > > > >> > > > > > >> > https://dist.apache.org/repos/dist/dev/spark/v3.3.4-rc1-bin/ > > > > >> > > > > > >> > > > > > >> > Signatures used for Spark RCs can be found in this file: > > > > >> > > > > > >> > https://dist.apache.org/repos/dist/dev/spark/KEYS > > > > >> > > > > > >> > > > > > >> > The staging repository for this release can be found at: > > > > >> > > > > > >> > > > https://repository.apache.org/content/repositories/orgapachespark-1451/ > > > > >> > > > > > >> > > > > > >> > The documentation corresponding to this release can be found at: > > > > >> > > > > > >> > https://dist.apache.org/repos/dist/dev/spark/v3.3.4-rc1-docs/ > > > > >> > > > > > >> > > > > > >> > The list of bug fixes going into 3.3.4 can be found at the > > following URL: > > > > >> > > > > > >> > https://issues.apache.org/jira/projects/SPARK/versions/12353505 > > > > >> > > > > > >> > > > > > >> > This release is using the release script of the tag v3.3.4-rc1. > > > > >> > > > > > >> > > > > > >> > FAQ > > > > >> > > > > > >> > > > > > >> > = > > > > >> > > > > > >> > How can I help test this release? > > > > >> > > > > > >> > = > > > > >> > > > > > >> > > > > > >> > > > > > >> > If you are a Spark user, you can help us test this release by > > taking > > > > >> > > > > > >> > an existing Spark workload and running on this release candidate, > > then > > > > >> > > > > > >> > reporting any regressions. > > > > >> > > > > > >> > > > > > >> > > > > > >> > If you're working in PySpark you can set up a virtual env and > > install > > > > >> > > > > > >> > the current RC and see if anything important breaks, in the > > Java/Scala > > > > >> > > > > > >> > you can add the staging repository to your projects resolvers and > > test > > > > >> > > > > > >> > with the RC (make sure to clean up the artifact cache > > before/after so > > > > >> > > > > > >> > you don't end up building with a out of date RC going forward). > > > > >> > > > > > >> > > > > > >> > > > > > >> > === > > > > >> > > > > > >> > What should happen to JIRA tickets still targeting 3.3.4? > > > > >> > > > > > >> > === > > > > >> > > > > > >> > > > > > >> > > > > > >> > The current list of open tickets targeted at 3.3.4 can be found > > at: > > > > >> > > > > > >> > https://issues.apache.org/jira/projects/SPARK and search for > > "Target > > > > >> > Version/s" = 3.3.4 > > > > >> > > > > > >> > > > > > >> > Committers should look at those and triage. Extremely important > > bug > > > > >> > > > > > >> > fixes, documentation, and API tweaks that impact compatibility > > should > > > > >> > > > > > >> > be worked on immediately. Everything else please retarget to an > > > > >> > > > > > >> > appropriate release. > > > > >> > > > > > >> > > > > > >> > > > > > >> > == > > > > >> > > > > > >> > But my bug isn't fixed? > > > > >> > > > > > >> > == > > > > >> > > > > > >> > > > > > >> > > > > > >> > In order to make timely releases, we will typically not hold the > > > > >> > > > > > >> > release unless the bug in question is a regression from the > > previous > > > > >> > > > > > >> > release. That being said, if there is something which is a > > regression > > > > >> > > > > > >> > that has not been correctly targeted please ping me or a > > committer to > > > > >> > > > > > >> > help target the issue. > > > > >> > > > > > >> > > > > >> > >
Re: [VOTE] Release Spark 3.3.4 (RC1)
+1 On Mon, Dec 11, 2023 at 6:21 PM Yang Jie wrote: > +1 > > On 2023/12/11 03:03:39 "L. C. Hsieh" wrote: > > +1 > > > > On Sun, Dec 10, 2023 at 6:15 PM Kent Yao wrote: > > > > > > +1(non-binding > > > > > > Kent Yao > > > > > > Yuming Wang 于2023年12月11日周一 09:33写道: > > > > > > > > +1 > > > > > > > > On Mon, Dec 11, 2023 at 5:55 AM Dongjoon Hyun > wrote: > > > >> > > > >> +1 > > > >> > > > >> Dongjoon > > > >> > > > >> On 2023/12/08 21:41:00 Dongjoon Hyun wrote: > > > >> > Please vote on releasing the following candidate as Apache Spark > version > > > >> > 3.3.4. > > > >> > > > > >> > The vote is open until December 15th 1AM (PST) and passes if a > majority +1 > > > >> > PMC votes are cast, with a minimum of 3 +1 votes. > > > >> > > > > >> > [ ] +1 Release this package as Apache Spark 3.3.4 > > > >> > [ ] -1 Do not release this package because ... > > > >> > > > > >> > To learn more about Apache Spark, please see > https://spark.apache.org/ > > > >> > > > > >> > The tag to be voted on is v3.3.4-rc1 (commit > > > >> > 18db204995b32e87a650f2f09f9bcf047ddafa90) > > > >> > https://github.com/apache/spark/tree/v3.3.4-rc1 > > > >> > > > > >> > The release files, including signatures, digests, etc. can be > found at: > > > >> > > > > >> > https://dist.apache.org/repos/dist/dev/spark/v3.3.4-rc1-bin/ > > > >> > > > > >> > > > > >> > Signatures used for Spark RCs can be found in this file: > > > >> > > > > >> > https://dist.apache.org/repos/dist/dev/spark/KEYS > > > >> > > > > >> > > > > >> > The staging repository for this release can be found at: > > > >> > > > > >> > > https://repository.apache.org/content/repositories/orgapachespark-1451/ > > > >> > > > > >> > > > > >> > The documentation corresponding to this release can be found at: > > > >> > > > > >> > https://dist.apache.org/repos/dist/dev/spark/v3.3.4-rc1-docs/ > > > >> > > > > >> > > > > >> > The list of bug fixes going into 3.3.4 can be found at the > following URL: > > > >> > > > > >> > https://issues.apache.org/jira/projects/SPARK/versions/12353505 > > > >> > > > > >> > > > > >> > This release is using the release script of the tag v3.3.4-rc1. > > > >> > > > > >> > > > > >> > FAQ > > > >> > > > > >> > > > > >> > = > > > >> > > > > >> > How can I help test this release? > > > >> > > > > >> > = > > > >> > > > > >> > > > > >> > > > > >> > If you are a Spark user, you can help us test this release by > taking > > > >> > > > > >> > an existing Spark workload and running on this release candidate, > then > > > >> > > > > >> > reporting any regressions. > > > >> > > > > >> > > > > >> > > > > >> > If you're working in PySpark you can set up a virtual env and > install > > > >> > > > > >> > the current RC and see if anything important breaks, in the > Java/Scala > > > >> > > > > >> > you can add the staging repository to your projects resolvers and > test > > > >> > > > > >> > with the RC (make sure to clean up the artifact cache > before/after so > > > >> > > > > >> > you don't end up building with a out of date RC going forward). > > > >> > > > > >> > > > > >> > > > > >> > === > > > >> > > > > >> > What should happen to JIRA tickets still targeting 3.3.4? > > > >> > > > > >> > === > > > >> > > > > >> > > > > >> > > > > >> > The current list of open tickets targeted at 3.3.4 can be found > at: > > > >> > > > > >> > https://issues.apache.org/jira/projects/SPARK and search for > "Target > > > >> > Version/s" = 3.3.4 > > > >> > > > > >> > > > > >> > Committers should look at those and triage. Extremely important > bug > > > >> > > > > >> > fixes, documentation, and API tweaks that impact compatibility > should > > > >> > > > > >> > be worked on immediately. Everything else please retarget to an > > > >> > > > > >> > appropriate release. > > > >> > > > > >> > > > > >> > > > > >> > == > > > >> > > > > >> > But my bug isn't fixed? > > > >> > > > > >> > == > > > >> > > > > >> > > > > >> > > > > >> > In order to make timely releases, we will typically not hold the > > > >> > > > > >> > release unless the bug in question is a regression from the > previous > > > >> > > > > >> > release. That being said, if there is something which is a > regression > > > >> > > > > >> > that has not been correctly targeted please ping me or a > committer to > > > >> > > > > >> > help target the issue. > > > >> > > > > >> > > > >> > - > > > >> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org > > > >> > > > > > > - > > > To unsubscribe e-mail: dev-unsubscr...@spark.apache.org > > > > > > > - > > To unsubscribe e-mail: dev-unsubscr...@spark.apache.org > > > > > >
Re: [VOTE] Release Spark 3.3.4 (RC1)
+1 On 2023/12/11 03:03:39 "L. C. Hsieh" wrote: > +1 > > On Sun, Dec 10, 2023 at 6:15 PM Kent Yao wrote: > > > > +1(non-binding > > > > Kent Yao > > > > Yuming Wang 于2023年12月11日周一 09:33写道: > > > > > > +1 > > > > > > On Mon, Dec 11, 2023 at 5:55 AM Dongjoon Hyun wrote: > > >> > > >> +1 > > >> > > >> Dongjoon > > >> > > >> On 2023/12/08 21:41:00 Dongjoon Hyun wrote: > > >> > Please vote on releasing the following candidate as Apache Spark > > >> > version > > >> > 3.3.4. > > >> > > > >> > The vote is open until December 15th 1AM (PST) and passes if a > > >> > majority +1 > > >> > PMC votes are cast, with a minimum of 3 +1 votes. > > >> > > > >> > [ ] +1 Release this package as Apache Spark 3.3.4 > > >> > [ ] -1 Do not release this package because ... > > >> > > > >> > To learn more about Apache Spark, please see https://spark.apache.org/ > > >> > > > >> > The tag to be voted on is v3.3.4-rc1 (commit > > >> > 18db204995b32e87a650f2f09f9bcf047ddafa90) > > >> > https://github.com/apache/spark/tree/v3.3.4-rc1 > > >> > > > >> > The release files, including signatures, digests, etc. can be found at: > > >> > > > >> > https://dist.apache.org/repos/dist/dev/spark/v3.3.4-rc1-bin/ > > >> > > > >> > > > >> > Signatures used for Spark RCs can be found in this file: > > >> > > > >> > https://dist.apache.org/repos/dist/dev/spark/KEYS > > >> > > > >> > > > >> > The staging repository for this release can be found at: > > >> > > > >> > https://repository.apache.org/content/repositories/orgapachespark-1451/ > > >> > > > >> > > > >> > The documentation corresponding to this release can be found at: > > >> > > > >> > https://dist.apache.org/repos/dist/dev/spark/v3.3.4-rc1-docs/ > > >> > > > >> > > > >> > The list of bug fixes going into 3.3.4 can be found at the following > > >> > URL: > > >> > > > >> > https://issues.apache.org/jira/projects/SPARK/versions/12353505 > > >> > > > >> > > > >> > This release is using the release script of the tag v3.3.4-rc1. > > >> > > > >> > > > >> > FAQ > > >> > > > >> > > > >> > = > > >> > > > >> > How can I help test this release? > > >> > > > >> > = > > >> > > > >> > > > >> > > > >> > If you are a Spark user, you can help us test this release by taking > > >> > > > >> > an existing Spark workload and running on this release candidate, then > > >> > > > >> > reporting any regressions. > > >> > > > >> > > > >> > > > >> > If you're working in PySpark you can set up a virtual env and install > > >> > > > >> > the current RC and see if anything important breaks, in the Java/Scala > > >> > > > >> > you can add the staging repository to your projects resolvers and test > > >> > > > >> > with the RC (make sure to clean up the artifact cache before/after so > > >> > > > >> > you don't end up building with a out of date RC going forward). > > >> > > > >> > > > >> > > > >> > === > > >> > > > >> > What should happen to JIRA tickets still targeting 3.3.4? > > >> > > > >> > === > > >> > > > >> > > > >> > > > >> > The current list of open tickets targeted at 3.3.4 can be found at: > > >> > > > >> > https://issues.apache.org/jira/projects/SPARK and search for "Target > > >> > Version/s" = 3.3.4 > > >> > > > >> > > > >> > Committers should look at those and triage. Extremely important bug > > >> > > > >> > fixes, documentation, and API tweaks that impact compatibility should > > >> > > > >> > be worked on immediately. Everything else please retarget to an > > >> > > > >> > appropriate release. > > >> > > > >> > > > >> > > > >> > == > > >> > > > >> > But my bug isn't fixed? > > >> > > > >> > == > > >> > > > >> > > > >> > > > >> > In order to make timely releases, we will typically not hold the > > >> > > > >> > release unless the bug in question is a regression from the previous > > >> > > > >> > release. That being said, if there is something which is a regression > > >> > > > >> > that has not been correctly targeted please ping me or a committer to > > >> > > > >> > help target the issue. > > >> > > > >> > > >> - > > >> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org > > >> > > > > - > > To unsubscribe e-mail: dev-unsubscr...@spark.apache.org > > > > - > To unsubscribe e-mail: dev-unsubscr...@spark.apache.org > > - To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
Re: [VOTE] Release Spark 3.3.4 (RC1)
Hi, Mridul. > I am currently on Python 3.11.6, java 8. For the above, I added `Python 3.11 support` at Apache Spark 3.4.0. That's exactly one of my reasons why I wanted to do the EOL release of Apache Spark 3.3.4. https://issues.apache.org/jira/browse/SPARK-41454 (Support Python 3.11) Thanks, Dongjoon. On Mon, Dec 11, 2023 at 12:22 PM Mridul Muralidharan wrote: > > I am seeing a bunch of python related (43) failures in the sql module (for > example [1]) ... I am currently on Python 3.11.6, java 8. > Not sure if ubuntu modified anything from under me, thoughts ? > > I am currently testing this against an older branch to make sure it is not > an issue with my desktop. > > Regards, > Mridul > > > [1] > > > org.apache.spark.sql.IntegratedUDFTestUtils.shouldTestGroupedAggPandasUDFs > was false (QueryCompilationErrorsSuite.scala:112) > Traceback (most recent call last): > File > "/home/mridul/work/apache/vote/spark/python/pyspark/serializers.py", line > 458, in dumps > return cloudpickle.dumps(obj, pickle_protocol) >^^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 73, in dumps > cp.dump(obj) > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 602, in dump > return Pickler.dump(self, obj) >^^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 692, in reducer_override > return self._function_reduce(obj) >^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 565, in _function_reduce > return self._dynamic_function_reduce(obj) >^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 546, in _dynamic_function_reduce > state = _function_getstate(func) > > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 157, in _function_getstate > f_globals_ref = _extract_code_globals(func.__code__) > > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle.py", > line 334, in _extract_code_globals > out_names = {names[oparg]: None for _, oparg in _walk_global_ops(co)} > ^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle.py", > line 334, in > out_names = {names[oparg]: None for _, oparg in _walk_global_ops(co)} > ~^^^ > IndexError: tuple index out of range > Traceback (most recent call last): > File > "/home/mridul/work/apache/vote/spark/python/pyspark/serializers.py", line > 458, in dumps > return cloudpickle.dumps(obj, pickle_protocol) >^^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 73, in dumps > cp.dump(obj) > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 602, in dump > return Pickler.dump(self, obj) >^^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 692, in reducer_override > return self._function_reduce(obj) >^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 565, in _function_reduce > return self._dynamic_function_reduce(obj) >^^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 546, in _dynamic_function_reduce > state = _function_getstate(func) > > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", > line 157, in _function_getstate > f_globals_ref = _extract_code_globals(func.__code__) > > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle.py", > line 334, in _extract_code_globals > out_names = {names[oparg]: None for _, oparg in _walk_global_ops(co)} > ^ > File > "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle.py", > line 334, in > out_names = {names[oparg]: None for _, oparg in _walk_global_ops(co)} > ~^^^ > IndexError: tuple index out of range > > During handling of the above exception, another exception occurred: > > Traceback (most recent call last): > File "", line 1, in >
Re: [VOTE] Release Spark 3.3.4 (RC1)
I am seeing a bunch of python related (43) failures in the sql module (for example [1]) ... I am currently on Python 3.11.6, java 8. Not sure if ubuntu modified anything from under me, thoughts ? I am currently testing this against an older branch to make sure it is not an issue with my desktop. Regards, Mridul [1] org.apache.spark.sql.IntegratedUDFTestUtils.shouldTestGroupedAggPandasUDFs was false (QueryCompilationErrorsSuite.scala:112) Traceback (most recent call last): File "/home/mridul/work/apache/vote/spark/python/pyspark/serializers.py", line 458, in dumps return cloudpickle.dumps(obj, pickle_protocol) ^^^ File "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", line 73, in dumps cp.dump(obj) File "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", line 602, in dump return Pickler.dump(self, obj) ^^^ File "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", line 692, in reducer_override return self._function_reduce(obj) ^^ File "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", line 565, in _function_reduce return self._dynamic_function_reduce(obj) ^^ File "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", line 546, in _dynamic_function_reduce state = _function_getstate(func) File "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", line 157, in _function_getstate f_globals_ref = _extract_code_globals(func.__code__) File "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle.py", line 334, in _extract_code_globals out_names = {names[oparg]: None for _, oparg in _walk_global_ops(co)} ^ File "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle.py", line 334, in out_names = {names[oparg]: None for _, oparg in _walk_global_ops(co)} ~^^^ IndexError: tuple index out of range Traceback (most recent call last): File "/home/mridul/work/apache/vote/spark/python/pyspark/serializers.py", line 458, in dumps return cloudpickle.dumps(obj, pickle_protocol) ^^^ File "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", line 73, in dumps cp.dump(obj) File "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", line 602, in dump return Pickler.dump(self, obj) ^^^ File "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", line 692, in reducer_override return self._function_reduce(obj) ^^ File "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", line 565, in _function_reduce return self._dynamic_function_reduce(obj) ^^ File "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", line 546, in _dynamic_function_reduce state = _function_getstate(func) File "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle_fast.py", line 157, in _function_getstate f_globals_ref = _extract_code_globals(func.__code__) File "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle.py", line 334, in _extract_code_globals out_names = {names[oparg]: None for _, oparg in _walk_global_ops(co)} ^ File "/home/mridul/work/apache/vote/spark/python/pyspark/cloudpickle/cloudpickle.py", line 334, in out_names = {names[oparg]: None for _, oparg in _walk_global_ops(co)} ~^^^ IndexError: tuple index out of range During handling of the above exception, another exception occurred: Traceback (most recent call last): File "", line 1, in File "/home/mridul/work/apache/vote/spark/python/pyspark/serializers.py", line 468, in dumps raise pickle.PicklingError(msg) _pickle.PicklingError: Could not serialize object: IndexError: tuple index out of range - UNSUPPORTED_FEATURE: Using Python UDF with unsupported join condition *** FAILED *** On Sun, Dec 10, 2023 at 9:05 PM L. C. Hsieh wrote: > +1 > > On Sun, Dec 10, 2023 at 6:15 PM Kent Yao wrote: > > > > +1(non-binding > > > > Kent Yao > > > > Yuming Wang 于2023年12月11日周一 09:33写道: > > > > > > +1 > > > > > > On Mon, Dec 11, 2023 at 5:55 AM Dongjoon Hyun > wrote: > > >> >
Re: [VOTE] Release Spark 3.3.4 (RC1)
+1 On Sun, Dec 10, 2023 at 6:15 PM Kent Yao wrote: > > +1(non-binding > > Kent Yao > > Yuming Wang 于2023年12月11日周一 09:33写道: > > > > +1 > > > > On Mon, Dec 11, 2023 at 5:55 AM Dongjoon Hyun wrote: > >> > >> +1 > >> > >> Dongjoon > >> > >> On 2023/12/08 21:41:00 Dongjoon Hyun wrote: > >> > Please vote on releasing the following candidate as Apache Spark version > >> > 3.3.4. > >> > > >> > The vote is open until December 15th 1AM (PST) and passes if a majority > >> > +1 > >> > PMC votes are cast, with a minimum of 3 +1 votes. > >> > > >> > [ ] +1 Release this package as Apache Spark 3.3.4 > >> > [ ] -1 Do not release this package because ... > >> > > >> > To learn more about Apache Spark, please see https://spark.apache.org/ > >> > > >> > The tag to be voted on is v3.3.4-rc1 (commit > >> > 18db204995b32e87a650f2f09f9bcf047ddafa90) > >> > https://github.com/apache/spark/tree/v3.3.4-rc1 > >> > > >> > The release files, including signatures, digests, etc. can be found at: > >> > > >> > https://dist.apache.org/repos/dist/dev/spark/v3.3.4-rc1-bin/ > >> > > >> > > >> > Signatures used for Spark RCs can be found in this file: > >> > > >> > https://dist.apache.org/repos/dist/dev/spark/KEYS > >> > > >> > > >> > The staging repository for this release can be found at: > >> > > >> > https://repository.apache.org/content/repositories/orgapachespark-1451/ > >> > > >> > > >> > The documentation corresponding to this release can be found at: > >> > > >> > https://dist.apache.org/repos/dist/dev/spark/v3.3.4-rc1-docs/ > >> > > >> > > >> > The list of bug fixes going into 3.3.4 can be found at the following URL: > >> > > >> > https://issues.apache.org/jira/projects/SPARK/versions/12353505 > >> > > >> > > >> > This release is using the release script of the tag v3.3.4-rc1. > >> > > >> > > >> > FAQ > >> > > >> > > >> > = > >> > > >> > How can I help test this release? > >> > > >> > = > >> > > >> > > >> > > >> > If you are a Spark user, you can help us test this release by taking > >> > > >> > an existing Spark workload and running on this release candidate, then > >> > > >> > reporting any regressions. > >> > > >> > > >> > > >> > If you're working in PySpark you can set up a virtual env and install > >> > > >> > the current RC and see if anything important breaks, in the Java/Scala > >> > > >> > you can add the staging repository to your projects resolvers and test > >> > > >> > with the RC (make sure to clean up the artifact cache before/after so > >> > > >> > you don't end up building with a out of date RC going forward). > >> > > >> > > >> > > >> > === > >> > > >> > What should happen to JIRA tickets still targeting 3.3.4? > >> > > >> > === > >> > > >> > > >> > > >> > The current list of open tickets targeted at 3.3.4 can be found at: > >> > > >> > https://issues.apache.org/jira/projects/SPARK and search for "Target > >> > Version/s" = 3.3.4 > >> > > >> > > >> > Committers should look at those and triage. Extremely important bug > >> > > >> > fixes, documentation, and API tweaks that impact compatibility should > >> > > >> > be worked on immediately. Everything else please retarget to an > >> > > >> > appropriate release. > >> > > >> > > >> > > >> > == > >> > > >> > But my bug isn't fixed? > >> > > >> > == > >> > > >> > > >> > > >> > In order to make timely releases, we will typically not hold the > >> > > >> > release unless the bug in question is a regression from the previous > >> > > >> > release. That being said, if there is something which is a regression > >> > > >> > that has not been correctly targeted please ping me or a committer to > >> > > >> > help target the issue. > >> > > >> > >> - > >> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org > >> > > - > To unsubscribe e-mail: dev-unsubscr...@spark.apache.org > - To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
Re: [VOTE] Release Spark 3.3.4 (RC1)
+1(non-binding Kent Yao Yuming Wang 于2023年12月11日周一 09:33写道: > > +1 > > On Mon, Dec 11, 2023 at 5:55 AM Dongjoon Hyun wrote: >> >> +1 >> >> Dongjoon >> >> On 2023/12/08 21:41:00 Dongjoon Hyun wrote: >> > Please vote on releasing the following candidate as Apache Spark version >> > 3.3.4. >> > >> > The vote is open until December 15th 1AM (PST) and passes if a majority +1 >> > PMC votes are cast, with a minimum of 3 +1 votes. >> > >> > [ ] +1 Release this package as Apache Spark 3.3.4 >> > [ ] -1 Do not release this package because ... >> > >> > To learn more about Apache Spark, please see https://spark.apache.org/ >> > >> > The tag to be voted on is v3.3.4-rc1 (commit >> > 18db204995b32e87a650f2f09f9bcf047ddafa90) >> > https://github.com/apache/spark/tree/v3.3.4-rc1 >> > >> > The release files, including signatures, digests, etc. can be found at: >> > >> > https://dist.apache.org/repos/dist/dev/spark/v3.3.4-rc1-bin/ >> > >> > >> > Signatures used for Spark RCs can be found in this file: >> > >> > https://dist.apache.org/repos/dist/dev/spark/KEYS >> > >> > >> > The staging repository for this release can be found at: >> > >> > https://repository.apache.org/content/repositories/orgapachespark-1451/ >> > >> > >> > The documentation corresponding to this release can be found at: >> > >> > https://dist.apache.org/repos/dist/dev/spark/v3.3.4-rc1-docs/ >> > >> > >> > The list of bug fixes going into 3.3.4 can be found at the following URL: >> > >> > https://issues.apache.org/jira/projects/SPARK/versions/12353505 >> > >> > >> > This release is using the release script of the tag v3.3.4-rc1. >> > >> > >> > FAQ >> > >> > >> > = >> > >> > How can I help test this release? >> > >> > = >> > >> > >> > >> > If you are a Spark user, you can help us test this release by taking >> > >> > an existing Spark workload and running on this release candidate, then >> > >> > reporting any regressions. >> > >> > >> > >> > If you're working in PySpark you can set up a virtual env and install >> > >> > the current RC and see if anything important breaks, in the Java/Scala >> > >> > you can add the staging repository to your projects resolvers and test >> > >> > with the RC (make sure to clean up the artifact cache before/after so >> > >> > you don't end up building with a out of date RC going forward). >> > >> > >> > >> > === >> > >> > What should happen to JIRA tickets still targeting 3.3.4? >> > >> > === >> > >> > >> > >> > The current list of open tickets targeted at 3.3.4 can be found at: >> > >> > https://issues.apache.org/jira/projects/SPARK and search for "Target >> > Version/s" = 3.3.4 >> > >> > >> > Committers should look at those and triage. Extremely important bug >> > >> > fixes, documentation, and API tweaks that impact compatibility should >> > >> > be worked on immediately. Everything else please retarget to an >> > >> > appropriate release. >> > >> > >> > >> > == >> > >> > But my bug isn't fixed? >> > >> > == >> > >> > >> > >> > In order to make timely releases, we will typically not hold the >> > >> > release unless the bug in question is a regression from the previous >> > >> > release. That being said, if there is something which is a regression >> > >> > that has not been correctly targeted please ping me or a committer to >> > >> > help target the issue. >> > >> >> - >> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org >> - To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
Re: [VOTE] Release Spark 3.3.4 (RC1)
+1 On Mon, Dec 11, 2023 at 5:55 AM Dongjoon Hyun wrote: > +1 > > Dongjoon > > On 2023/12/08 21:41:00 Dongjoon Hyun wrote: > > Please vote on releasing the following candidate as Apache Spark version > > 3.3.4. > > > > The vote is open until December 15th 1AM (PST) and passes if a majority > +1 > > PMC votes are cast, with a minimum of 3 +1 votes. > > > > [ ] +1 Release this package as Apache Spark 3.3.4 > > [ ] -1 Do not release this package because ... > > > > To learn more about Apache Spark, please see https://spark.apache.org/ > > > > The tag to be voted on is v3.3.4-rc1 (commit > > 18db204995b32e87a650f2f09f9bcf047ddafa90) > > https://github.com/apache/spark/tree/v3.3.4-rc1 > > > > The release files, including signatures, digests, etc. can be found at: > > > > https://dist.apache.org/repos/dist/dev/spark/v3.3.4-rc1-bin/ > > > > > > Signatures used for Spark RCs can be found in this file: > > > > https://dist.apache.org/repos/dist/dev/spark/KEYS > > > > > > The staging repository for this release can be found at: > > > > https://repository.apache.org/content/repositories/orgapachespark-1451/ > > > > > > The documentation corresponding to this release can be found at: > > > > https://dist.apache.org/repos/dist/dev/spark/v3.3.4-rc1-docs/ > > > > > > The list of bug fixes going into 3.3.4 can be found at the following URL: > > > > https://issues.apache.org/jira/projects/SPARK/versions/12353505 > > > > > > This release is using the release script of the tag v3.3.4-rc1. > > > > > > FAQ > > > > > > = > > > > How can I help test this release? > > > > = > > > > > > > > If you are a Spark user, you can help us test this release by taking > > > > an existing Spark workload and running on this release candidate, then > > > > reporting any regressions. > > > > > > > > If you're working in PySpark you can set up a virtual env and install > > > > the current RC and see if anything important breaks, in the Java/Scala > > > > you can add the staging repository to your projects resolvers and test > > > > with the RC (make sure to clean up the artifact cache before/after so > > > > you don't end up building with a out of date RC going forward). > > > > > > > > === > > > > What should happen to JIRA tickets still targeting 3.3.4? > > > > === > > > > > > > > The current list of open tickets targeted at 3.3.4 can be found at: > > > > https://issues.apache.org/jira/projects/SPARK and search for "Target > > Version/s" = 3.3.4 > > > > > > Committers should look at those and triage. Extremely important bug > > > > fixes, documentation, and API tweaks that impact compatibility should > > > > be worked on immediately. Everything else please retarget to an > > > > appropriate release. > > > > > > > > == > > > > But my bug isn't fixed? > > > > == > > > > > > > > In order to make timely releases, we will typically not hold the > > > > release unless the bug in question is a regression from the previous > > > > release. That being said, if there is something which is a regression > > > > that has not been correctly targeted please ping me or a committer to > > > > help target the issue. > > > > - > To unsubscribe e-mail: dev-unsubscr...@spark.apache.org > >
Re: [VOTE] Release Spark 3.3.4 (RC1)
+1 Dongjoon On 2023/12/08 21:41:00 Dongjoon Hyun wrote: > Please vote on releasing the following candidate as Apache Spark version > 3.3.4. > > The vote is open until December 15th 1AM (PST) and passes if a majority +1 > PMC votes are cast, with a minimum of 3 +1 votes. > > [ ] +1 Release this package as Apache Spark 3.3.4 > [ ] -1 Do not release this package because ... > > To learn more about Apache Spark, please see https://spark.apache.org/ > > The tag to be voted on is v3.3.4-rc1 (commit > 18db204995b32e87a650f2f09f9bcf047ddafa90) > https://github.com/apache/spark/tree/v3.3.4-rc1 > > The release files, including signatures, digests, etc. can be found at: > > https://dist.apache.org/repos/dist/dev/spark/v3.3.4-rc1-bin/ > > > Signatures used for Spark RCs can be found in this file: > > https://dist.apache.org/repos/dist/dev/spark/KEYS > > > The staging repository for this release can be found at: > > https://repository.apache.org/content/repositories/orgapachespark-1451/ > > > The documentation corresponding to this release can be found at: > > https://dist.apache.org/repos/dist/dev/spark/v3.3.4-rc1-docs/ > > > The list of bug fixes going into 3.3.4 can be found at the following URL: > > https://issues.apache.org/jira/projects/SPARK/versions/12353505 > > > This release is using the release script of the tag v3.3.4-rc1. > > > FAQ > > > = > > How can I help test this release? > > = > > > > If you are a Spark user, you can help us test this release by taking > > an existing Spark workload and running on this release candidate, then > > reporting any regressions. > > > > If you're working in PySpark you can set up a virtual env and install > > the current RC and see if anything important breaks, in the Java/Scala > > you can add the staging repository to your projects resolvers and test > > with the RC (make sure to clean up the artifact cache before/after so > > you don't end up building with a out of date RC going forward). > > > > === > > What should happen to JIRA tickets still targeting 3.3.4? > > === > > > > The current list of open tickets targeted at 3.3.4 can be found at: > > https://issues.apache.org/jira/projects/SPARK and search for "Target > Version/s" = 3.3.4 > > > Committers should look at those and triage. Extremely important bug > > fixes, documentation, and API tweaks that impact compatibility should > > be worked on immediately. Everything else please retarget to an > > appropriate release. > > > > == > > But my bug isn't fixed? > > == > > > > In order to make timely releases, we will typically not hold the > > release unless the bug in question is a regression from the previous > > release. That being said, if there is something which is a regression > > that has not been correctly targeted please ping me or a committer to > > help target the issue. > - To unsubscribe e-mail: dev-unsubscr...@spark.apache.org