-1

Let us include the correctness fix:
https://github.com/apache/spark/pull/27229

Thanks,

Xiao

On Thu, Jan 16, 2020 at 8:46 AM Dongjoon Hyun <dongjoon.h...@gmail.com>
wrote:

> Thank you, Jungtaek!
>
> Bests,
> Dongjoon.
>
>
> On Wed, Jan 15, 2020 at 8:57 PM Jungtaek Lim <kabhwan.opensou...@gmail.com>
> wrote:
>
>> Once we decided to cancel the RC1, what about including SPARK-29450 (
>> https://github.com/apache/spark/pull/27209) into RC2?
>>
>> SPARK-29450 was merged into master, and Xiao figured out it fixed a
>> regression, long lasting one (broken at 2.3.0). The link refers the PR for
>> 2.4 branch.
>>
>> Thanks,
>> Jungtaek Lim (HeartSaVioR)
>>
>> On Thu, Jan 16, 2020 at 12:56 PM Dongjoon Hyun <dongjoon.h...@gmail.com>
>> wrote:
>>
>>> Sure. Wenchen and Hyukjin.
>>>
>>> I observed all of the above reported issues and have been waiting to
>>> collect more information before cancelling RC1 vote.
>>>
>>> The other stuff I've observed is that Marcelo and Sean also requested
>>> reverting the existing commit.
>>> - https://github.com/apache/spark/pull/24732 (spark.shuffle.io.backLog
>>> change)
>>>
>>> To All.
>>> We want your explicit feedbacks. Please reply on this thread.
>>>
>>> Although we get enough positive feedbacks here, I'll cancel this RC1.
>>> I want to address at least the above negative feedbacks and roll RC2
>>> next Monday.
>>>
>>> Bests,
>>> Dongjoon.
>>>
>>>
>>> On Wed, Jan 15, 2020 at 7:47 PM Hyukjin Kwon <gurwls...@gmail.com>
>>> wrote:
>>>
>>>> If we go for RC2, we should include both:
>>>>
>>>> https://github.com/apache/spark/pull/27210
>>>> https://github.com/apache/spark/pull/27184
>>>>
>>>> just for the sake of being complete and making the maintenance simple.
>>>>
>>>>
>>>> 2020년 1월 16일 (목) 오후 12:38, Wenchen Fan <cloud0...@gmail.com>님이 작성:
>>>>
>>>>> Recently we merged several fixes to 2.4:
>>>>> https://issues.apache.org/jira/browse/SPARK-30325   a driver hang
>>>>> issue
>>>>> https://issues.apache.org/jira/browse/SPARK-30246   a memory leak
>>>>> issue
>>>>> https://issues.apache.org/jira/browse/SPARK-29708   a correctness
>>>>> issue(for a rarely used feature, so not merged to 2.4 yet)
>>>>>
>>>>> Shall we include them?
>>>>>
>>>>>
>>>>> On Wed, Jan 15, 2020 at 9:51 PM Hyukjin Kwon <gurwls...@gmail.com>
>>>>> wrote:
>>>>>
>>>>>> +1
>>>>>>
>>>>>> On Wed, 15 Jan 2020, 08:24 Takeshi Yamamuro, <linguin....@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> +1;
>>>>>>>
>>>>>>> I checked the links and materials, then I run the tests with
>>>>>>> `-Pyarn -Phadoop-2.7 -Phive -Phive-thriftserver -Pmesos -Pkubernetes
>>>>>>> -Psparkr`
>>>>>>> on macOS (Java 8).
>>>>>>> All the things look fine and I didn't see the error on my env
>>>>>>> that Sean said above.
>>>>>>>
>>>>>>> Thanks, Dongjoon!
>>>>>>>
>>>>>>> Bests,
>>>>>>> Takeshi
>>>>>>>
>>>>>>> On Wed, Jan 15, 2020 at 4:09 AM DB Tsai <dbt...@dbtsai.com> wrote:
>>>>>>>
>>>>>>>> +1 Thanks.
>>>>>>>>
>>>>>>>> Sincerely,
>>>>>>>>
>>>>>>>> DB Tsai
>>>>>>>> ----------------------------------------------------------
>>>>>>>> Web: https://www.dbtsai.com
>>>>>>>> PGP Key ID: 42E5B25A8F7A82C1
>>>>>>>>
>>>>>>>> On Tue, Jan 14, 2020 at 11:08 AM Sean Owen <sro...@apache.org>
>>>>>>>> wrote:
>>>>>>>> >
>>>>>>>> > Yeah it's something about the env I spun up, but I don't know
>>>>>>>> what. It
>>>>>>>> > happens frequently when I test, but not on Jenkins.
>>>>>>>> > The Kafka error comes up every now and then and a clean rebuild
>>>>>>>> fixes
>>>>>>>> > it, but not in my case. I don't know why.
>>>>>>>> > But if nobody else sees it, I'm pretty sure it's just an artifact
>>>>>>>> of
>>>>>>>> > the local VM.
>>>>>>>> >
>>>>>>>> > On Tue, Jan 14, 2020 at 12:57 PM Dongjoon Hyun <
>>>>>>>> dongjoon.h...@gmail.com> wrote:
>>>>>>>> > >
>>>>>>>> > > Thank you, Sean.
>>>>>>>> > >
>>>>>>>> > > First of all, the `Ubuntu` job on Amplab Jenkins farm is green.
>>>>>>>> > >
>>>>>>>> > >
>>>>>>>> https://amplab.cs.berkeley.edu/jenkins/view/Spark%20QA%20Test/job/spark-branch-2.4-test-sbt-hadoop-2.7-ubuntu-testing/
>>>>>>>> > >
>>>>>>>> > > For the failures,
>>>>>>>> > >    1. Yes, the `HiveExternalCatalogVersionsSuite` flakiness is
>>>>>>>> a known one.
>>>>>>>> > >    2. For `HDFSMetadataLogSuite` failure, I also observed a few
>>>>>>>> time before in CentOS too.
>>>>>>>> > >    3. Kafka build error is new to me. Does it happen on `Maven`
>>>>>>>> clean build?
>>>>>>>> > >
>>>>>>>> > > Bests,
>>>>>>>> > > Dongjoon.
>>>>>>>> > >
>>>>>>>> > >
>>>>>>>> > > On Tue, Jan 14, 2020 at 6:40 AM Sean Owen <sro...@apache.org>
>>>>>>>> wrote:
>>>>>>>> > >>
>>>>>>>> > >> +1 from me. I checked sigs/licenses, and built/tested from
>>>>>>>> source on
>>>>>>>> > >> Java 8 + Ubuntu 18.04 with " -Pyarn -Phive -Phive-thriftserver
>>>>>>>> > >> -Phadoop-2.7 -Pmesos -Pkubernetes -Psparkr -Pkinesis-asl". I
>>>>>>>> do get
>>>>>>>> > >> test failures, but, these are some I have always seen on
>>>>>>>> Ubuntu, and I
>>>>>>>> > >> do not know why they happen. They don't seem to affect others,
>>>>>>>> but,
>>>>>>>> > >> let me know if anyone else sees these?
>>>>>>>> > >>
>>>>>>>> > >>
>>>>>>>> > >> Always happens for me:
>>>>>>>> > >>
>>>>>>>> > >> - HDFSMetadataLog: metadata directory collision *** FAILED ***
>>>>>>>> > >>   The await method on Waiter timed out.
>>>>>>>> (HDFSMetadataLogSuite.scala:178)
>>>>>>>> > >>
>>>>>>>> > >> This one has been flaky at times due to external dependencies:
>>>>>>>> > >>
>>>>>>>> > >> org.apache.spark.sql.hive.HiveExternalCatalogVersionsSuite ***
>>>>>>>> ABORTED ***
>>>>>>>> > >>   Exception encountered when invoking run on a nested suite -
>>>>>>>> > >> spark-submit returned with exit code 1.
>>>>>>>> > >>   Command line: './bin/spark-submit' '--name' 'prepare testing
>>>>>>>> tables'
>>>>>>>> > >> '--master' 'local[2]' '--conf' 'spark.ui.enabled=false'
>>>>>>>> '--conf'
>>>>>>>> > >> 'spark.master.rest.enabled=false' '--conf'
>>>>>>>> > >>
>>>>>>>> 'spark.sql.warehouse.dir=/data/spark-2.4.5/sql/hive/target/tmp/warehouse-c2f762fd-688e-42b7-a822-06823a6bbd98'
>>>>>>>> > >> '--conf' 'spark.sql.test.version.index=0'
>>>>>>>> '--driver-java-options'
>>>>>>>> > >>
>>>>>>>> '-Dderby.system.home=/data/spark-2.4.5/sql/hive/target/tmp/warehouse-c2f762fd-688e-42b7-a822-06823a6bbd98'
>>>>>>>> > >>
>>>>>>>> '/data/spark-2.4.5/sql/hive/target/tmp/test7297526474581770293.py'
>>>>>>>> > >>
>>>>>>>> > >> Kafka doesn't build with this weird error. I tried a clean
>>>>>>>> build. I
>>>>>>>> > >> think we've seen this before.
>>>>>>>> > >>
>>>>>>>> > >> [error] This symbol is required by 'method
>>>>>>>> > >> org.apache.spark.metrics.MetricsSystem.getServletHandlers'.
>>>>>>>> > >> [error] Make sure that term eclipse is in your classpath and
>>>>>>>> check for
>>>>>>>> > >> conflicting dependencies with `-Ylog-classpath`.
>>>>>>>> > >> [error] A full rebuild may help if 'MetricsSystem.class' was
>>>>>>>> compiled
>>>>>>>> > >> against an incompatible version of org.
>>>>>>>> > >> [error]     testUtils.sendMessages(topic, data.toArray)
>>>>>>>> > >> [error]
>>>>>>>> > >>
>>>>>>>> > >> On Mon, Jan 13, 2020 at 6:28 AM Dongjoon Hyun <
>>>>>>>> dongjoon.h...@gmail.com> wrote:
>>>>>>>> > >> >
>>>>>>>> > >> > Please vote on releasing the following candidate as Apache
>>>>>>>> Spark version 2.4.5.
>>>>>>>> > >> >
>>>>>>>> > >> > The vote is open until January 16th 5AM PST and passes if a
>>>>>>>> majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
>>>>>>>> > >> >
>>>>>>>> > >> > [ ] +1 Release this package as Apache Spark 2.4.5
>>>>>>>> > >> > [ ] -1 Do not release this package because ...
>>>>>>>> > >> >
>>>>>>>> > >> > To learn more about Apache Spark, please see
>>>>>>>> http://spark.apache.org/
>>>>>>>> > >> >
>>>>>>>> > >> > The tag to be voted on is v2.4.5-rc1 (commit
>>>>>>>> 33bd2beee5e3772a9af1d782f195e6a678c54cf0):
>>>>>>>> > >> > https://github.com/apache/spark/tree/v2.4.5-rc1
>>>>>>>> > >> >
>>>>>>>> > >> > The release files, including signatures, digests, etc. can
>>>>>>>> be found at:
>>>>>>>> > >> > https://dist.apache.org/repos/dist/dev/spark/v2.4.5-rc1-bin/
>>>>>>>> > >> >
>>>>>>>> > >> > Signatures used for Spark RCs can be found in this file:
>>>>>>>> > >> > https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>>>>>> > >> >
>>>>>>>> > >> > The staging repository for this release can be found at:
>>>>>>>> > >> >
>>>>>>>> https://repository.apache.org/content/repositories/orgapachespark-1339/
>>>>>>>> > >> >
>>>>>>>> > >> > The documentation corresponding to this release can be found
>>>>>>>> at:
>>>>>>>> > >> >
>>>>>>>> https://dist.apache.org/repos/dist/dev/spark/v2.4.5-rc1-docs/
>>>>>>>> > >> >
>>>>>>>> > >> > The list of bug fixes going into 2.4.5 can be found at the
>>>>>>>> following URL:
>>>>>>>> > >> >
>>>>>>>> https://issues.apache.org/jira/projects/SPARK/versions/12346042
>>>>>>>> > >> >
>>>>>>>> > >> > This release is using the release script of the tag
>>>>>>>> v2.4.5-rc1.
>>>>>>>> > >> >
>>>>>>>> > >> > FAQ
>>>>>>>> > >> >
>>>>>>>> > >> > =========================
>>>>>>>> > >> > How can I help test this release?
>>>>>>>> > >> > =========================
>>>>>>>> > >> >
>>>>>>>> > >> > If you are a Spark user, you can help us test this release
>>>>>>>> by taking
>>>>>>>> > >> > an existing Spark workload and running on this release
>>>>>>>> candidate, then
>>>>>>>> > >> > reporting any regressions.
>>>>>>>> > >> >
>>>>>>>> > >> > If you're working in PySpark you can set up a virtual env
>>>>>>>> and install
>>>>>>>> > >> > the current RC and see if anything important breaks, in the
>>>>>>>> Java/Scala
>>>>>>>> > >> > you can add the staging repository to your projects
>>>>>>>> resolvers and test
>>>>>>>> > >> > with the RC (make sure to clean up the artifact cache
>>>>>>>> before/after so
>>>>>>>> > >> > you don't end up building with a out of date RC going
>>>>>>>> forward).
>>>>>>>> > >> >
>>>>>>>> > >> > ===========================================
>>>>>>>> > >> > What should happen to JIRA tickets still targeting 2.4.5?
>>>>>>>> > >> > ===========================================
>>>>>>>> > >> >
>>>>>>>> > >> > The current list of open tickets targeted at 2.4.5 can be
>>>>>>>> found at:
>>>>>>>> > >> > https://issues.apache.org/jira/projects/SPARK and search
>>>>>>>> for "Target Version/s" = 2.4.5
>>>>>>>> > >> >
>>>>>>>> > >> > Committers should look at those and triage. Extremely
>>>>>>>> important bug
>>>>>>>> > >> > fixes, documentation, and API tweaks that impact
>>>>>>>> compatibility should
>>>>>>>> > >> > be worked on immediately. Everything else please retarget to
>>>>>>>> an
>>>>>>>> > >> > appropriate release.
>>>>>>>> > >> >
>>>>>>>> > >> > ==================
>>>>>>>> > >> > But my bug isn't fixed?
>>>>>>>> > >> > ==================
>>>>>>>> > >> >
>>>>>>>> > >> > In order to make timely releases, we will typically not hold
>>>>>>>> the
>>>>>>>> > >> > release unless the bug in question is a regression from the
>>>>>>>> previous
>>>>>>>> > >> > release. That being said, if there is something which is a
>>>>>>>> regression
>>>>>>>> > >> > that has not been correctly targeted please ping me or a
>>>>>>>> committer to
>>>>>>>> > >> > help target the issue.
>>>>>>>> >
>>>>>>>> >
>>>>>>>> ---------------------------------------------------------------------
>>>>>>>> > To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>>>>>> >
>>>>>>>>
>>>>>>>>
>>>>>>>> ---------------------------------------------------------------------
>>>>>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> ---
>>>>>>> Takeshi Yamamuro
>>>>>>>
>>>>>>

-- 
[image: Databricks Summit - Watch the talks]
<https://databricks.com/sparkaisummit/north-america>

Reply via email to