If we go for RC2, we should include both:

https://github.com/apache/spark/pull/27210
https://github.com/apache/spark/pull/27184

just for the sake of being complete and making the maintenance simple.


2020년 1월 16일 (목) 오후 12:38, Wenchen Fan <cloud0...@gmail.com>님이 작성:

> Recently we merged several fixes to 2.4:
> https://issues.apache.org/jira/browse/SPARK-30325   a driver hang issue
> https://issues.apache.org/jira/browse/SPARK-30246   a memory leak issue
> https://issues.apache.org/jira/browse/SPARK-29708   a correctness
> issue(for a rarely used feature, so not merged to 2.4 yet)
>
> Shall we include them?
>
>
> On Wed, Jan 15, 2020 at 9:51 PM Hyukjin Kwon <gurwls...@gmail.com> wrote:
>
>> +1
>>
>> On Wed, 15 Jan 2020, 08:24 Takeshi Yamamuro, <linguin....@gmail.com>
>> wrote:
>>
>>> +1;
>>>
>>> I checked the links and materials, then I run the tests with
>>> `-Pyarn -Phadoop-2.7 -Phive -Phive-thriftserver -Pmesos -Pkubernetes
>>> -Psparkr`
>>> on macOS (Java 8).
>>> All the things look fine and I didn't see the error on my env
>>> that Sean said above.
>>>
>>> Thanks, Dongjoon!
>>>
>>> Bests,
>>> Takeshi
>>>
>>> On Wed, Jan 15, 2020 at 4:09 AM DB Tsai <dbt...@dbtsai.com> wrote:
>>>
>>>> +1 Thanks.
>>>>
>>>> Sincerely,
>>>>
>>>> DB Tsai
>>>> ----------------------------------------------------------
>>>> Web: https://www.dbtsai.com
>>>> PGP Key ID: 42E5B25A8F7A82C1
>>>>
>>>> On Tue, Jan 14, 2020 at 11:08 AM Sean Owen <sro...@apache.org> wrote:
>>>> >
>>>> > Yeah it's something about the env I spun up, but I don't know what. It
>>>> > happens frequently when I test, but not on Jenkins.
>>>> > The Kafka error comes up every now and then and a clean rebuild fixes
>>>> > it, but not in my case. I don't know why.
>>>> > But if nobody else sees it, I'm pretty sure it's just an artifact of
>>>> > the local VM.
>>>> >
>>>> > On Tue, Jan 14, 2020 at 12:57 PM Dongjoon Hyun <
>>>> dongjoon.h...@gmail.com> wrote:
>>>> > >
>>>> > > Thank you, Sean.
>>>> > >
>>>> > > First of all, the `Ubuntu` job on Amplab Jenkins farm is green.
>>>> > >
>>>> > >
>>>> https://amplab.cs.berkeley.edu/jenkins/view/Spark%20QA%20Test/job/spark-branch-2.4-test-sbt-hadoop-2.7-ubuntu-testing/
>>>> > >
>>>> > > For the failures,
>>>> > >    1. Yes, the `HiveExternalCatalogVersionsSuite` flakiness is a
>>>> known one.
>>>> > >    2. For `HDFSMetadataLogSuite` failure, I also observed a few
>>>> time before in CentOS too.
>>>> > >    3. Kafka build error is new to me. Does it happen on `Maven`
>>>> clean build?
>>>> > >
>>>> > > Bests,
>>>> > > Dongjoon.
>>>> > >
>>>> > >
>>>> > > On Tue, Jan 14, 2020 at 6:40 AM Sean Owen <sro...@apache.org>
>>>> wrote:
>>>> > >>
>>>> > >> +1 from me. I checked sigs/licenses, and built/tested from source
>>>> on
>>>> > >> Java 8 + Ubuntu 18.04 with " -Pyarn -Phive -Phive-thriftserver
>>>> > >> -Phadoop-2.7 -Pmesos -Pkubernetes -Psparkr -Pkinesis-asl". I do get
>>>> > >> test failures, but, these are some I have always seen on Ubuntu,
>>>> and I
>>>> > >> do not know why they happen. They don't seem to affect others, but,
>>>> > >> let me know if anyone else sees these?
>>>> > >>
>>>> > >>
>>>> > >> Always happens for me:
>>>> > >>
>>>> > >> - HDFSMetadataLog: metadata directory collision *** FAILED ***
>>>> > >>   The await method on Waiter timed out.
>>>> (HDFSMetadataLogSuite.scala:178)
>>>> > >>
>>>> > >> This one has been flaky at times due to external dependencies:
>>>> > >>
>>>> > >> org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite ***
>>>> ABORTED ***
>>>> > >>   Exception encountered when invoking run on a nested suite -
>>>> > >> spark-submit returned with exit code 1.
>>>> > >>   Command line: './bin/spark-submit' '--name' 'prepare testing
>>>> tables'
>>>> > >> '--master' 'local[2]' '--conf' 'spark.ui.enabled=false' '--conf'
>>>> > >> 'spark.master.rest.enabled=false' '--conf'
>>>> > >>
>>>> 'spark.sql.warehouse.dir=/data/spark-2.4.5/sql/hive/target/tmp/warehouse-c2f762fd-688e-42b7-a822-06823a6bbd98'
>>>> > >> '--conf' 'spark.sql.test.version.index=0' '--driver-java-options'
>>>> > >>
>>>> '-Dderby.system.home=/data/spark-2.4.5/sql/hive/target/tmp/warehouse-c2f762fd-688e-42b7-a822-06823a6bbd98'
>>>> > >> '/data/spark-2.4.5/sql/hive/target/tmp/test7297526474581770293.py'
>>>> > >>
>>>> > >> Kafka doesn't build with this weird error. I tried a clean build. I
>>>> > >> think we've seen this before.
>>>> > >>
>>>> > >> [error] This symbol is required by 'method
>>>> > >> org.apache.spark.metrics.MetricsSystem.getServletHandlers'.
>>>> > >> [error] Make sure that term eclipse is in your classpath and check
>>>> for
>>>> > >> conflicting dependencies with `-Ylog-classpath`.
>>>> > >> [error] A full rebuild may help if 'MetricsSystem.class' was
>>>> compiled
>>>> > >> against an incompatible version of org.
>>>> > >> [error]     testUtils.sendMessages(topic, data.toArray)
>>>> > >> [error]
>>>> > >>
>>>> > >> On Mon, Jan 13, 2020 at 6:28 AM Dongjoon Hyun <
>>>> dongjoon.h...@gmail.com> wrote:
>>>> > >> >
>>>> > >> > Please vote on releasing the following candidate as Apache Spark
>>>> version 2.4.5.
>>>> > >> >
>>>> > >> > The vote is open until January 16th 5AM PST and passes if a
>>>> majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
>>>> > >> >
>>>> > >> > [ ] +1 Release this package as Apache Spark 2.4.5
>>>> > >> > [ ] -1 Do not release this package because ...
>>>> > >> >
>>>> > >> > To learn more about Apache Spark, please see
>>>> http://spark.apache.org/
>>>> > >> >
>>>> > >> > The tag to be voted on is v2.4.5-rc1 (commit
>>>> 33bd2beee5e3772a9af1d782f195e6a678c54cf0):
>>>> > >> > https://github.com/apache/spark/tree/v2.4.5-rc1
>>>> > >> >
>>>> > >> > The release files, including signatures, digests, etc. can be
>>>> found at:
>>>> > >> > https://dist.apache.org/repos/dist/dev/spark/v2.4.5-rc1-bin/
>>>> > >> >
>>>> > >> > Signatures used for Spark RCs can be found in this file:
>>>> > >> > https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>> > >> >
>>>> > >> > The staging repository for this release can be found at:
>>>> > >> >
>>>> https://repository.apache.org/content/repositories/orgapachespark-1339/
>>>> > >> >
>>>> > >> > The documentation corresponding to this release can be found at:
>>>> > >> > https://dist.apache.org/repos/dist/dev/spark/v2.4.5-rc1-docs/
>>>> > >> >
>>>> > >> > The list of bug fixes going into 2.4.5 can be found at the
>>>> following URL:
>>>> > >> > https://issues.apache.org/jira/projects/SPARK/versions/12346042
>>>> > >> >
>>>> > >> > This release is using the release script of the tag v2.4.5-rc1.
>>>> > >> >
>>>> > >> > FAQ
>>>> > >> >
>>>> > >> > =========================
>>>> > >> > How can I help test this release?
>>>> > >> > =========================
>>>> > >> >
>>>> > >> > If you are a Spark user, you can help us test this release by
>>>> taking
>>>> > >> > an existing Spark workload and running on this release
>>>> candidate, then
>>>> > >> > reporting any regressions.
>>>> > >> >
>>>> > >> > If you're working in PySpark you can set up a virtual env and
>>>> install
>>>> > >> > the current RC and see if anything important breaks, in the
>>>> Java/Scala
>>>> > >> > you can add the staging repository to your projects resolvers
>>>> and test
>>>> > >> > with the RC (make sure to clean up the artifact cache
>>>> before/after so
>>>> > >> > you don't end up building with a out of date RC going forward).
>>>> > >> >
>>>> > >> > ===========================================
>>>> > >> > What should happen to JIRA tickets still targeting 2.4.5?
>>>> > >> > ===========================================
>>>> > >> >
>>>> > >> > The current list of open tickets targeted at 2.4.5 can be found
>>>> at:
>>>> > >> > https://issues.apache.org/jira/projects/SPARK and search for
>>>> "Target Version/s" = 2.4.5
>>>> > >> >
>>>> > >> > Committers should look at those and triage. Extremely important
>>>> bug
>>>> > >> > fixes, documentation, and API tweaks that impact compatibility
>>>> should
>>>> > >> > be worked on immediately. Everything else please retarget to an
>>>> > >> > appropriate release.
>>>> > >> >
>>>> > >> > ==================
>>>> > >> > But my bug isn't fixed?
>>>> > >> > ==================
>>>> > >> >
>>>> > >> > In order to make timely releases, we will typically not hold the
>>>> > >> > release unless the bug in question is a regression from the
>>>> previous
>>>> > >> > release. That being said, if there is something which is a
>>>> regression
>>>> > >> > that has not been correctly targeted please ping me or a
>>>> committer to
>>>> > >> > help target the issue.
>>>> >
>>>> > ---------------------------------------------------------------------
>>>> > To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>> >
>>>>
>>>> ---------------------------------------------------------------------
>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>>
>>>>
>>>
>>> --
>>> ---
>>> Takeshi Yamamuro
>>>
>>

Reply via email to