Xiao If it doesn't apply/you've changed your mind if you can re-vote that would be rad.
On Fri, Sep 15, 2017 at 2:22 PM, Felix Cheung <felixcheun...@hotmail.com> wrote: > Yes ;) > > ------------------------------ > *From:* Xiao Li <gatorsm...@gmail.com> > *Sent:* Friday, September 15, 2017 2:22:03 PM > *To:* Holden Karau > *Cc:* Ryan Blue; Denny Lee; Felix Cheung; Sean Owen; dev@spark.apache.org > > *Subject:* Re: [VOTE] Spark 2.1.2 (RC1) > > Sorry, this release candidate is 2.1.2. The issue is in 2.2.1. > > 2017-09-15 14:21 GMT-07:00 Xiao Li <gatorsm...@gmail.com>: > >> -1 >> >> See the discussion in https://github.com/apache/spark/pull/19074 >> >> Xiao >> >> >> >> 2017-09-15 12:28 GMT-07:00 Holden Karau <hol...@pigscanfly.ca>: >> >>> That's a good question, I built the release candidate however the >>> Jenkins scripts don't take a parameter for configuring who signs them >>> rather it always signs them with Patrick's key. You can see this from >>> previous releases which were managed by other folks but still signed by >>> Patrick. >>> >>> On Fri, Sep 15, 2017 at 12:16 PM, Ryan Blue <rb...@netflix.com> wrote: >>> >>>> The signature is valid, but why was the release signed with Patrick >>>> Wendell's private key? Did Patrick build the release candidate? >>>> >>>> rb >>>> >>>> On Fri, Sep 15, 2017 at 6:36 AM, Denny Lee <denny.g....@gmail.com> >>>> wrote: >>>> >>>>> +1 (non-binding) >>>>> >>>>> On Thu, Sep 14, 2017 at 10:57 PM Felix Cheung < >>>>> felixcheun...@hotmail.com> wrote: >>>>> >>>>>> +1 tested SparkR package on Windows, r-hub, Ubuntu. >>>>>> >>>>>> _____________________________ >>>>>> From: Sean Owen <so...@cloudera.com> >>>>>> Sent: Thursday, September 14, 2017 3:12 PM >>>>>> Subject: Re: [VOTE] Spark 2.1.2 (RC1) >>>>>> To: Holden Karau <hol...@pigscanfly.ca>, <dev@spark.apache.org> >>>>>> >>>>>> >>>>>> >>>>>> +1 >>>>>> Very nice. The sigs and hashes look fine, it builds fine for me on >>>>>> Debian Stretch with Java 8, yarn/hive/hadoop-2.7 profiles, and passes >>>>>> tests. >>>>>> >>>>>> Yes as you say, no outstanding issues except for this which doesn't >>>>>> look critical, as it's not a regression. >>>>>> >>>>>> SPARK-21985 PySpark PairDeserializer is broken for double-zipped RDDs >>>>>> >>>>>> >>>>>> On Thu, Sep 14, 2017 at 7:47 PM Holden Karau <hol...@pigscanfly.ca> >>>>>> wrote: >>>>>> >>>>>>> Please vote on releasing the following candidate as Apache Spark >>>>>>> version 2.1.2. The vote is open until Friday September 22nd at >>>>>>> 18:00 PST and passes if a majority of at least 3 +1 PMC votes are >>>>>>> cast. >>>>>>> >>>>>>> [ ] +1 Release this package as Apache Spark 2.1.2 >>>>>>> [ ] -1 Do not release this package because ... >>>>>>> >>>>>>> >>>>>>> To learn more about Apache Spark, please see >>>>>>> https://spark.apache.org/ >>>>>>> >>>>>>> The tag to be voted on is v2.1.2-rc1 >>>>>>> <https://github.com/apache/spark/tree/v2.1.2-rc1> (6f470323a036365 >>>>>>> 6999dd36cb33f528afe627c12) >>>>>>> >>>>>>> List of JIRA tickets resolved in this release can be found with >>>>>>> this filter. >>>>>>> <https://issues.apache.org/jira/browse/SPARK-20134?jql=project%20%3D%20SPARK%20AND%20fixVersion%20%3D%202.1.2> >>>>>>> >>>>>>> The release files, including signatures, digests, etc. can be found >>>>>>> at: >>>>>>> https://home.apache.org/~pwendell/spark-releases/spark-2.1.2 >>>>>>> -rc1-bin/ >>>>>>> >>>>>>> Release artifacts are signed with the following key: >>>>>>> https://people.apache.org/keys/committer/pwendell.asc >>>>>>> >>>>>>> The staging repository for this release can be found at: >>>>>>> https://repository.apache.org/content/repositories/orgapache >>>>>>> spark-1248/ >>>>>>> >>>>>>> The documentation corresponding to this release can be found at: >>>>>>> https://people.apache.org/~pwendell/spark-releases/spark-2.1 >>>>>>> .2-rc1-docs/ >>>>>>> >>>>>>> >>>>>>> *FAQ* >>>>>>> >>>>>>> *How can I help test this release?* >>>>>>> >>>>>>> If you are a Spark user, you can help us test this release by taking >>>>>>> an existing Spark workload and running on this release candidate, then >>>>>>> reporting any regressions. >>>>>>> >>>>>>> If you're working in PySpark you can set up a virtual env and >>>>>>> install the current RC and see if anything important breaks, in the >>>>>>> Java/Scala you can add the staging repository to your projects resolvers >>>>>>> and test with the RC (make sure to clean up the artifact cache >>>>>>> before/after >>>>>>> so you don't end up building with a out of date RC going forward). >>>>>>> >>>>>>> *What should happen to JIRA tickets still targeting 2.1.2?* >>>>>>> >>>>>>> Committers should look at those and triage. Extremely important bug >>>>>>> fixes, documentation, and API tweaks that impact compatibility should be >>>>>>> worked on immediately. Everything else please retarget to 2.1.3. >>>>>>> >>>>>>> *But my bug isn't fixed!??!* >>>>>>> >>>>>>> In order to make timely releases, we will typically not hold the >>>>>>> release unless the bug in question is a regression from 2.1.1. That >>>>>>> being >>>>>>> said if there is something which is a regression form 2.1.1 that has not >>>>>>> been correctly targeted please ping a committer to help target the issue >>>>>>> (you can see the open issues listed as impacting Spark 2.1.1 & 2.1.2 >>>>>>> <https://issues.apache.org/jira/browse/SPARK-21985?jql=project%20%3D%20SPARK%20AND%20status%20%3D%20OPEN%20AND%20(affectedVersion%20%3D%202.1.2%20OR%20affectedVersion%20%3D%202.1.1)> >>>>>>> ) >>>>>>> >>>>>>> *What are the unresolved* issues targeted for 2.1.2 >>>>>>> <https://issues.apache.org/jira/browse/SPARK-21985?jql=project%20%3D%20SPARK%20AND%20status%20in%20(Open%2C%20%22In%20Progress%22%2C%20Reopened)%20AND%20%22Target%20Version%2Fs%22%20%3D%202.1.2> >>>>>>> ? >>>>>>> >>>>>>> At the time of the writing, there is one in progress major issue >>>>>>> SPARK-21985 <https://issues.apache.org/jira/browse/SPARK-21985>, I >>>>>>> believe Andrew Ray & HyukjinKwon are looking into this one. >>>>>>> >>>>>>> -- >>>>>>> Twitter: https://twitter.com/holdenkarau >>>>>>> >>>>>> >>>>>> >>>>>> >>>> >>>> >>>> -- >>>> Ryan Blue >>>> Software Engineer >>>> Netflix >>>> >>> >>> >>> >>> -- >>> Twitter: https://twitter.com/holdenkarau >>> >> >> > -- Twitter: https://twitter.com/holdenkarau