I agree with Jungtaek,  -1 from me because of the issue of Kafka source
throwing an error with an incorrect error message that was introduced
recently.  This may mislead users and cause unnecessary confusion.

On Wed, Jun 8, 2022 at 12:04 AM Jungtaek Lim <kabhwan.opensou...@gmail.com>
wrote:

> Apologize for late participation.
>
> I'm sorry, but -1 (non-binding) from me.
>
> Unfortunately I found a major user-facing issue which hurts UX seriously
> on Kafka data source usage.
>
> In some cases, Kafka data source can throw IllegalStateException for the
> case of failOnDataLoss=true which condition is bound to the state of Kafka
> topic (not Spark's issue). With the recent change of Spark,
> IllegalStateException is now bound to the "internal error", and Spark gives
> incorrect guidance to the end users, telling to end users that Spark has a
> bug and they are encouraged to file a JIRA ticket which is simply wrong.
>
> Previously, Kafka data source provided the error message with the context
> why it failed, and how to workaround it. I feel this is a serious
> regression on UX.
>
> Please look into https://issues.apache.org/jira/browse/SPARK-39412 for
> more details.
>
>
> On Wed, Jun 8, 2022 at 3:40 PM Hyukjin Kwon <gurwls...@gmail.com> wrote:
>
>> Okay. Thankfully the binary release is fine per
>> https://github.com/apache/spark/blob/v3.3.0-rc5/dev/create-release/release-build.sh#L268
>> .
>> The source package (and GitHub tag) has 3.3.0.dev0, and the binary
>> package has 3.3.0. Technically this is not a blocker now because PyPI
>> upload will be able to be made correctly.
>> I lowered the priority to critical. I switch my -1 to 0.
>>
>> On Wed, 8 Jun 2022 at 15:17, Hyukjin Kwon <gurwls...@gmail.com> wrote:
>>
>>> Arrrgh  .. I am very sorry that I found this problem late.
>>> RC 5 does not have the correct version of PySpark, see
>>> https://github.com/apache/spark/blob/v3.3.0-rc5/python/pyspark/version.py#L19
>>> I think the release script was broken because the version now has 'str'
>>> type, see
>>> https://github.com/apache/spark/blob/v3.3.0-rc5/dev/create-release/release-tag.sh#L88
>>> I filed a JIRA at https://issues.apache.org/jira/browse/SPARK-39411
>>>
>>> -1 from me
>>>
>>>
>>>
>>> On Wed, 8 Jun 2022 at 13:16, Cheng Pan <pan3...@gmail.com> wrote:
>>>
>>>> +1 (non-binding)
>>>>
>>>> * Verified SPARK-39313 has been address[1]
>>>> * Passed integration test w/ Apache Kyuubi (Incubating)[2]
>>>>
>>>> [1] https://github.com/housepower/spark-clickhouse-connector/pull/123
>>>> [2] https://github.com/apache/incubator-kyuubi/pull/2817
>>>>
>>>> Thanks,
>>>> Cheng Pan
>>>>
>>>> On Wed, Jun 8, 2022 at 7:04 AM Chris Nauroth <cnaur...@apache.org>
>>>> wrote:
>>>> >
>>>> > +1 (non-binding)
>>>> >
>>>> > * Verified all checksums.
>>>> > * Verified all signatures.
>>>> > * Built from source, with multiple profiles, to full success, for
>>>> Java 11 and Scala 2.13:
>>>> >     * build/mvn -Phadoop-3 -Phadoop-cloud -Phive-thriftserver
>>>> -Pkubernetes -Pscala-2.13 -Psparkr -Pyarn -DskipTests clean package
>>>> > * Tests passed.
>>>> > * Ran several examples successfully:
>>>> >     * bin/spark-submit --class org.apache.spark.examples.SparkPi
>>>> examples/jars/spark-examples_2.12-3.3.0.jar
>>>> >     * bin/spark-submit --class
>>>> org.apache.spark.examples.sql.hive.SparkHiveExample
>>>> examples/jars/spark-examples_2.12-3.3.0.jar
>>>> >     * bin/spark-submit
>>>> examples/src/main/python/streaming/network_wordcount.py localhost 9999
>>>> > * Tested some of the issues that blocked prior release candidates:
>>>> >     * bin/spark-sql -e 'SELECT (SELECT IF(x, 1, 0)) AS a FROM (SELECT
>>>> true) t(x) UNION SELECT 1 AS a;'
>>>> >     * bin/spark-sql -e "select date '2018-11-17' > 1"
>>>> >     * SPARK-39293 ArrayAggregate fix
>>>> >
>>>> > Chris Nauroth
>>>> >
>>>> >
>>>> > On Tue, Jun 7, 2022 at 1:30 PM Cheng Su <chen...@fb.com.invalid>
>>>> wrote:
>>>> >>
>>>> >> +1 (non-binding). Built and ran some internal test for Spark SQL.
>>>> >>
>>>> >>
>>>> >>
>>>> >> Thanks,
>>>> >>
>>>> >> Cheng Su
>>>> >>
>>>> >>
>>>> >>
>>>> >> From: L. C. Hsieh <vii...@gmail.com>
>>>> >> Date: Tuesday, June 7, 2022 at 1:23 PM
>>>> >> To: dev <dev@spark.apache.org>
>>>> >> Subject: Re: [VOTE] Release Spark 3.3.0 (RC5)
>>>> >>
>>>> >> +1
>>>> >>
>>>> >> Liang-Chi
>>>> >>
>>>> >> On Tue, Jun 7, 2022 at 1:03 PM Gengliang Wang <ltn...@gmail.com>
>>>> wrote:
>>>> >> >
>>>> >> > +1 (non-binding)
>>>> >> >
>>>> >> > Gengliang
>>>> >> >
>>>> >> > On Tue, Jun 7, 2022 at 12:24 PM Thomas Graves <
>>>> tgraves...@gmail.com> wrote:
>>>> >> >>
>>>> >> >> +1
>>>> >> >>
>>>> >> >> Tom Graves
>>>> >> >>
>>>> >> >> On Sat, Jun 4, 2022 at 9:50 AM Maxim Gekk
>>>> >> >> <maxim.g...@databricks.com.invalid> wrote:
>>>> >> >> >
>>>> >> >> > Please vote on releasing the following candidate as Apache
>>>> Spark version 3.3.0.
>>>> >> >> >
>>>> >> >> > The vote is open until 11:59pm Pacific time June 8th and passes
>>>> if a majority +1 PMC votes are cast, with a minimum of 3 +1 votes.
>>>> >> >> >
>>>> >> >> > [ ] +1 Release this package as Apache Spark 3.3.0
>>>> >> >> > [ ] -1 Do not release this package because ...
>>>> >> >> >
>>>> >> >> > To learn more about Apache Spark, please see
>>>> http://spark.apache.org/
>>>> >> >> >
>>>> >> >> > The tag to be voted on is v3.3.0-rc5 (commit
>>>> 7cf29705272ab8e8c70e8885a3664ad8ae3cd5e9):
>>>> >> >> > https://github.com/apache/spark/tree/v3.3.0-rc5
>>>> >> >> >
>>>> >> >> > The release files, including signatures, digests, etc. can be
>>>> found at:
>>>> >> >> > https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc5-bin/
>>>> >> >> >
>>>> >> >> > Signatures used for Spark RCs can be found in this file:
>>>> >> >> > https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>> >> >> >
>>>> >> >> > The staging repository for this release can be found at:
>>>> >> >> >
>>>> https://repository.apache.org/content/repositories/orgapachespark-1406
>>>> >> >> >
>>>> >> >> > The documentation corresponding to this release can be found at:
>>>> >> >> > https://dist.apache.org/repos/dist/dev/spark/v3.3.0-rc5-docs/
>>>> >> >> >
>>>> >> >> > The list of bug fixes going into 3.3.0 can be found at the
>>>> following URL:
>>>> >> >> > https://issues.apache.org/jira/projects/SPARK/versions/12350369
>>>> >> >> >
>>>> >> >> > This release is using the release script of the tag v3.3.0-rc5.
>>>> >> >> >
>>>> >> >> >
>>>> >> >> > FAQ
>>>> >> >> >
>>>> >> >> > =========================
>>>> >> >> > How can I help test this release?
>>>> >> >> > =========================
>>>> >> >> > If you are a Spark user, you can help us test this release by
>>>> taking
>>>> >> >> > an existing Spark workload and running on this release
>>>> candidate, then
>>>> >> >> > reporting any regressions.
>>>> >> >> >
>>>> >> >> > If you're working in PySpark you can set up a virtual env and
>>>> install
>>>> >> >> > the current RC and see if anything important breaks, in the
>>>> Java/Scala
>>>> >> >> > you can add the staging repository to your projects resolvers
>>>> and test
>>>> >> >> > with the RC (make sure to clean up the artifact cache
>>>> before/after so
>>>> >> >> > you don't end up building with a out of date RC going forward).
>>>> >> >> >
>>>> >> >> > ===========================================
>>>> >> >> > What should happen to JIRA tickets still targeting 3.3.0?
>>>> >> >> > ===========================================
>>>> >> >> > The current list of open tickets targeted at 3.3.0 can be found
>>>> at:
>>>> >> >> > https://issues.apache.org/jira/projects/SPARK  and search for
>>>> "Target Version/s" = 3.3.0
>>>> >> >> >
>>>> >> >> > Committers should look at those and triage. Extremely important
>>>> bug
>>>> >> >> > fixes, documentation, and API tweaks that impact compatibility
>>>> should
>>>> >> >> > be worked on immediately. Everything else please retarget to an
>>>> >> >> > appropriate release.
>>>> >> >> >
>>>> >> >> > ==================
>>>> >> >> > But my bug isn't fixed?
>>>> >> >> > ==================
>>>> >> >> > In order to make timely releases, we will typically not hold the
>>>> >> >> > release unless the bug in question is a regression from the
>>>> previous
>>>> >> >> > release. That being said, if there is something which is a
>>>> regression
>>>> >> >> > that has not been correctly targeted please ping me or a
>>>> committer to
>>>> >> >> > help target the issue.
>>>> >> >> >
>>>> >> >> > Maxim Gekk
>>>> >> >> >
>>>> >> >> > Software Engineer
>>>> >> >> >
>>>> >> >> > Databricks, Inc.
>>>> >> >>
>>>> >> >>
>>>> ---------------------------------------------------------------------
>>>> >> >> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>> >> >>
>>>> >>
>>>> >> ---------------------------------------------------------------------
>>>> >> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>>
>>>> ---------------------------------------------------------------------
>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>>
>>>>

Reply via email to