Thanks for letting us know. So this vote is cancelled in favor of RC2.


On Sun, Aug 9, 2020 at 8:31 AM Takeshi Yamamuro <linguin....@gmail.com>
wrote:

> Thanks for letting us know about the two issues above, Dongjoon.
>
> ----
> I've checked the release materials (signatures, tag, ...) and it looks
> fine, too.
> Also, I run the tests on my local Mac (java 1.8.0) with the options
> `-Pyarn -Phadoop-2.7 -Phive -Phive-thriftserver -Pmesos -Pkubernetes
> -Psparkr`
> and they passed.
>
> Bests,
> Takeshi
>
>
>
> On Sun, Aug 9, 2020 at 11:06 AM Dongjoon Hyun <dongjoon.h...@gmail.com>
> wrote:
>
>> Another instance is SPARK-31703 which filed on May 13th and the PR
>> arrived two days ago.
>>
>>     [SPARK-31703][SQL] Parquet RLE float/double are read incorrectly on
>> big endian platforms
>>     https://github.com/apache/spark/pull/29383
>>
>> It seems that the patch is already ready in this case.
>> I raised the priority of SPARK-31703 to `Blocker` for both Apache Spark
>> 2.4.7 and 3.0.1.
>>
>> Bests,
>> Dongjoon.
>>
>>
>> On Sat, Aug 8, 2020 at 6:10 AM Holden Karau <hol...@pigscanfly.ca> wrote:
>>
>>> I'm going to go ahead and vote -0 then based on that then.
>>>
>>> On Fri, Aug 7, 2020 at 11:36 PM Dongjoon Hyun <dongjoon.h...@gmail.com>
>>> wrote:
>>>
>>>> Hi, All.
>>>>
>>>> Unfortunately, there is an on-going discussion about the new decimal
>>>> correctness.
>>>>
>>>> Although we fixed one correctness issue at master and backported it
>>>> partially to 3.0/2.4, it turns out that it needs more patched to be
>>>> complete.
>>>>
>>>> Please see https://github.com/apache/spark/pull/29125 for on-going
>>>> discussion for both 3.0/2.4.
>>>>
>>>>     [SPARK-32018][SQL][3.0] UnsafeRow.setDecimal should set null with
>>>> overflowed value
>>>>
>>>> I also confirmed that 2.4.7 RC1 is affected.
>>>>
>>>> Bests,
>>>> Dongjoon.
>>>>
>>>>
>>>> On Thu, Aug 6, 2020 at 2:48 PM Sean Owen <sro...@apache.org> wrote:
>>>>
>>>>> +1 from me. The same as usual. Licenses and sigs look OK, builds and
>>>>> passes tests on a standard selection of profiles.
>>>>>
>>>>> On Thu, Aug 6, 2020 at 7:07 AM Prashant Sharma <scrapco...@gmail.com>
>>>>> wrote:
>>>>> >
>>>>> > Please vote on releasing the following candidate as Apache Spark
>>>>> version 2.4.7.
>>>>> >
>>>>> > The vote is open until Aug 9th at 9AM PST and passes if a majority
>>>>> +1 PMC votes are cast, with a minimum of 3 +1 votes.
>>>>> >
>>>>> > [ ] +1 Release this package as Apache Spark 2.4.7
>>>>> > [ ] -1 Do not release this package because ...
>>>>> >
>>>>> > To learn more about Apache Spark, please see
>>>>> http://spark.apache.org/
>>>>> >
>>>>> > There are currently no issues targeting 2.4.7 (try project = SPARK
>>>>> AND "Target Version/s" = "2.4.7" AND status in (Open, Reopened, "In
>>>>> Progress"))
>>>>> >
>>>>> > The tag to be voted on is v2.4.7-rc1 (commit
>>>>> dc04bf53fe821b7a07f817966c6c173f3b3788c6):
>>>>> > https://github.com/apache/spark/tree/v2.4.7-rc1
>>>>> >
>>>>> > The release files, including signatures, digests, etc. can be found
>>>>> at:
>>>>> > https://dist.apache.org/repos/dist/dev/spark/v2.4.7-rc1-bin/
>>>>> >
>>>>> > Signatures used for Spark RCs can be found in this file:
>>>>> > https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>>> >
>>>>> > The staging repository for this release can be found at:
>>>>> >
>>>>> https://repository.apache.org/content/repositories/orgapachespark-1352/
>>>>> >
>>>>> > The documentation corresponding to this release can be found at:
>>>>> > https://dist.apache.org/repos/dist/dev/spark/v2.4.7-rc1-docs/
>>>>> >
>>>>> > The list of bug fixes going into 2.4.7 can be found at the following
>>>>> URL:
>>>>> > https://s.apache.org/spark-v2.4.7-rc1
>>>>> >
>>>>> > This release is using the release script of the tag v2.4.7-rc1.
>>>>> >
>>>>> > FAQ
>>>>> >
>>>>> >
>>>>> > =========================
>>>>> > How can I help test this release?
>>>>> > =========================
>>>>> >
>>>>> > If you are a Spark user, you can help us test this release by taking
>>>>> > an existing Spark workload and running on this release candidate,
>>>>> then
>>>>> > reporting any regressions.
>>>>> >
>>>>> > If you're working in PySpark you can set up a virtual env and install
>>>>> > the current RC and see if anything important breaks, in the
>>>>> Java/Scala
>>>>> > you can add the staging repository to your projects resolvers and
>>>>> test
>>>>> > with the RC (make sure to clean up the artifact cache before/after so
>>>>> > you don't end up building with an out of date RC going forward).
>>>>> >
>>>>> > ===========================================
>>>>> > What should happen to JIRA tickets still targeting 2.4.7?
>>>>> > ===========================================
>>>>> >
>>>>> > The current list of open tickets targeted at 2.4.7 can be found at:
>>>>> > https://issues.apache.org/jira/projects/SPARK and search for
>>>>> "Target Version/s" = 2.4.7
>>>>> >
>>>>> > Committers should look at those and triage. Extremely important bug
>>>>> > fixes, documentation, and API tweaks that impact compatibility should
>>>>> > be worked on immediately. Everything else please retarget to an
>>>>> > appropriate release.
>>>>> >
>>>>> > ==================
>>>>> > But my bug isn't fixed?
>>>>> > ==================
>>>>> >
>>>>> > In order to make timely releases, we will typically not hold the
>>>>> > release unless the bug in question is a regression from the previous
>>>>> > release. That being said, if there is something which is a regression
>>>>> > that has not been correctly targeted please ping me or a committer to
>>>>> > help target the issue.
>>>>>
>>>>> ---------------------------------------------------------------------
>>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>>>
>>>>>
>>>
>>> --
>>> Twitter: https://twitter.com/holdenkarau
>>> Books (Learning Spark, High Performance Spark, etc.):
>>> https://amzn.to/2MaRAG9  <https://amzn.to/2MaRAG9>
>>> YouTube Live Streams: https://www.youtube.com/user/holdenkarau
>>>
>>
>
> --
> ---
> Takeshi Yamamuro
>

Reply via email to