+1

On Tue, Feb 9, 2021 at 1:39 PM Sean Owen <sro...@apache.org> wrote:

> Same result as last time for me, +1. Tested with Java 11.
> I fixed the two issues without assignee; one was WontFix though.
>
> On Mon, Feb 8, 2021 at 7:43 PM Hyukjin Kwon <gurwls...@gmail.com> wrote:
>
>> Let's set the assignees properly then. Shouldn't be a problem for the
>> release.
>>
>> On Tue, 9 Feb 2021, 10:40 Yuming Wang, <wgy...@gmail.com> wrote:
>>
>>>
>>> Many tickets do not have correct assignee:
>>>
>>> https://issues.apache.org/jira/issues/?jql=project%20%3D%20SPARK%20AND%20status%20in%20(Resolved%2C%20Closed)%20AND%20fixVersion%20in%20(3.1.0%2C%203.1.1)%20AND%20(assignee%20is%20EMPTY%20or%20assignee%20%3D%20apachespark)
>>>
>>>
>>> On Tue, Feb 9, 2021 at 9:05 AM Hyukjin Kwon <gurwls...@gmail.com> wrote:
>>>
>>>> +1 (binding) from myself too.
>>>>
>>>> 2021년 2월 9일 (화) 오전 9:28, Kent Yao <yaooq...@qq.com>님이 작성:
>>>>
>>>>>
>>>>> +1
>>>>>
>>>>> *Kent Yao *
>>>>> @ Data Science Center, Hangzhou Research Institute, NetEase Corp.
>>>>> *a spark enthusiast*
>>>>> *kyuubi <https://github.com/yaooqinn/kyuubi>is a
>>>>> unified multi-tenant JDBC interface for large-scale data processing and
>>>>> analytics, built on top of Apache Spark <http://spark.apache.org/>.*
>>>>> *spark-authorizer <https://github.com/yaooqinn/spark-authorizer>A
>>>>> Spark SQL extension which provides SQL Standard Authorization for **Apache
>>>>> Spark <http://spark.apache.org/>.*
>>>>> *spark-postgres <https://github.com/yaooqinn/spark-postgres> A library
>>>>> for reading data from and transferring data to Postgres / Greenplum with
>>>>> Spark SQL and DataFrames, 10~100x faster.*
>>>>> *spark-func-extras <https://github.com/yaooqinn/spark-func-extras>A
>>>>> library that brings excellent and useful functions from various modern
>>>>> database management systems to Apache Spark <http://spark.apache.org/>.*
>>>>>
>>>>>
>>>>>
>>>>> On 02/9/2021 08:24,Hyukjin Kwon<gurwls...@gmail.com>
>>>>> <gurwls...@gmail.com> wrote:
>>>>>
>>>>> Please vote on releasing the following candidate as Apache Spark
>>>>> version 3.1.1.
>>>>>
>>>>> The vote is open until February 15th 5PM PST and passes if a majority
>>>>> +1 PMC votes are cast, with a minimum of 3 +1 votes.
>>>>>
>>>>> Note that it is 7 days this time because it is a holiday season in
>>>>> several countries including South Korea (where I live), China etc., and I
>>>>> would like to make sure people do not miss it because it is a holiday
>>>>> season.
>>>>>
>>>>> [ ] +1 Release this package as Apache Spark 3.1.1
>>>>> [ ] -1 Do not release this package because ...
>>>>>
>>>>> To learn more about Apache Spark, please see http://spark.apache.org/
>>>>>
>>>>> The tag to be voted on is v3.1.1-rc2 (commit
>>>>> cf0115ac2d60070399af481b14566f33d22ec45e):
>>>>> https://github.com/apache/spark/tree/v3.1.1-rc2
>>>>>
>>>>> The release files, including signatures, digests, etc. can be found at:
>>>>> <https://dist.apache.org/repos/dist/dev/spark/v3.1.1-rc1-bin/>
>>>>> https://dist.apache.org/repos/dist/dev/spark/v3.1.1-rc2-bin/
>>>>>
>>>>> Signatures used for Spark RCs can be found in this file:
>>>>> https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>>>
>>>>> The staging repository for this release can be found at:
>>>>> https://repository.apache.org/content/repositories/orgapachespark-1365
>>>>>
>>>>> The documentation corresponding to this release can be found at:
>>>>> https://dist.apache.org/repos/dist/dev/spark/v3.1.1-rc2-docs/
>>>>>
>>>>> The list of bug fixes going into 3.1.1 can be found at the following
>>>>> URL:
>>>>> https://s.apache.org/41kf2
>>>>>
>>>>> This release is using the release script of the tag v3.1.1-rc2.
>>>>>
>>>>> FAQ
>>>>>
>>>>> ===================
>>>>> What happened to 3.1.0?
>>>>> ===================
>>>>>
>>>>> There was a technical issue during Apache Spark 3.1.0 preparation, and
>>>>> it was discussed and decided to skip 3.1.0.
>>>>> Please see
>>>>> https://spark.apache.org/news/next-official-release-spark-3.1.1.html for
>>>>> more details.
>>>>>
>>>>> =========================
>>>>> How can I help test this release?
>>>>> =========================
>>>>>
>>>>> If you are a Spark user, you can help us test this release by taking
>>>>> an existing Spark workload and running on this release candidate, then
>>>>> reporting any regressions.
>>>>>
>>>>> If you're working in PySpark you can set up a virtual env and install
>>>>> the current RC via "pip install
>>>>> https://dist.apache.org/repos/dist/dev/spark/v3.1.1-rc2-bin/pyspark-3.1.1.tar.gz
>>>>> "
>>>>> and see if anything important breaks.
>>>>> In the Java/Scala, you can add the staging repository to your projects
>>>>> resolvers and test
>>>>> with the RC (make sure to clean up the artifact cache before/after so
>>>>> you don't end up building with an out of date RC going forward).
>>>>>
>>>>> ===========================================
>>>>> What should happen to JIRA tickets still targeting 3.1.1?
>>>>> ===========================================
>>>>>
>>>>> The current list of open tickets targeted at 3.1.1 can be found at:
>>>>> https://issues.apache.org/jira/projects/SPARK and search for "Target
>>>>> Version/s" = 3.1.1
>>>>>
>>>>> Committers should look at those and triage. Extremely important bug
>>>>> fixes, documentation, and API tweaks that impact compatibility should
>>>>> be worked on immediately. Everything else please retarget to an
>>>>> appropriate release.
>>>>>
>>>>> ==================
>>>>> But my bug isn't fixed?
>>>>> ==================
>>>>>
>>>>> In order to make timely releases, we will typically not hold the
>>>>> release unless the bug in question is a regression from the previous
>>>>> release. That being said, if there is something which is a regression
>>>>> that has not been correctly targeted please ping me or a committer to
>>>>> help target the issue.
>>>>>
>>>>>
>>>>>

Reply via email to