RC2 fails and I'll cut RC3 next week. Thanks for the feedback!

On Thu, Mar 6, 2025 at 6:44 AM Chris Nauroth <cnaur...@apache.org> wrote:

> Here is one more problem I found during RC2 verification:
>
> https://github.com/apache/spark/pull/50173
>
> This one is just a test issue.
>
> Chris Nauroth
>
>
> On Tue, Mar 4, 2025 at 2:55 PM Jules Damji <jules.da...@gmail.com> wrote:
>
>> - 1 (non-binding)
>>
>> A ran into number of installation and launching problems. May be it’s my
>> enviornment, even though I removed any old binaries and packages.
>>
>> 1. Pip installing pyspark4.0.0 and pyspark-connect-4.0 from .tz file
>> workedl, launching pyspark results into
>>
>> 25/03/04 14:00:26 ERROR SparkContext: Error initializing SparkContext.
>>
>> java.lang.ClassNotFoundException:
>> org.apache.spark.sql.connect.SparkConnectPlugin
>>
>>
>> 2. Similary installing the tar balls of either distribution and launch
>> spark-shell goes into a loop and terminated by the shutdown hook.
>>
>>
>> Thank you Wenchen for leading these release onerous manager efforts, but
>> slowly we should be able to install and launch seamlessly.
>>
>>
>> Keep up the good work & tireless effort for the Spark community!
>>
>>
>> cheers
>>
>> Jules
>>
>>
>> WARNING: Using incubator modules: jdk.incubator.vector
>>
>> 25/03/04 14:49:35 INFO BaseAllocator: Debug mode disabled. Enable with
>> the VM option -Darrow.memory.debug.allocator=true.
>>
>> 25/03/04 14:49:35 INFO DefaultAllocationManagerOption: allocation manager
>> type not specified, using netty as the default type
>>
>> 25/03/04 14:49:35 INFO CheckAllocator: Using DefaultAllocationManager at
>> memory/netty/DefaultAllocationManagerFactory.class
>>
>> Using Spark's default log4j profile:
>> org/apache/spark/log4j2-defaults.properties
>>
>> 25/03/04 14:49:35 WARN GrpcRetryHandler: Non-Fatal error during RPC
>> execution: org.sparkproject.io.grpc.StatusRuntimeException: UNAVAILABLE: io
>> exception, retrying (wait=50 ms, currentRetryNum=1, policy=DefaultPolicy).
>>
>> 25/03/04 14:49:35 WARN GrpcRetryHandler: Non-Fatal error during RPC
>> execution: org.sparkproject.io.grpc.StatusRuntimeException: UNAVAILABLE: io
>> exception, retrying (wait=200 ms, currentRetryNum=2, policy=DefaultPolicy).
>>
>> 25/03/04 14:49:35 WARN GrpcRetryHandler: Non-Fatal error during RPC
>> execution: org.sparkproject.io.grpc.StatusRuntimeException: UNAVAILABLE: io
>> exception, retrying (wait=800 ms, currentRetryNum=3, policy=DefaultPolicy).
>>
>> 25/03/04 14:49:36 WARN GrpcRetryHandler: Non-Fatal error during RPC
>> execution: org.sparkproject.io.grpc.StatusRuntimeException: UNAVAILABLE: io
>> exception, retrying (wait=3275 ms, currentRetryNum=4, policy=DefaultPolicy).
>>
>> 25/03/04 14:49:39 WARN GrpcRetryHandler: Non-Fatal error during RPC
>> execution: org.sparkproject.io.grpc.StatusRuntimeException: UNAVAILABLE: io
>> exception, retrying (wait=12995 ms, currentRetryNum=5,
>> policy=DefaultPolicy).
>>
>> ^C25/03/04 14:49:40 INFO ShutdownHookManager: Shutdown hook called
>>
>>
>>
>> On Mar 4, 2025, at 2:24 PM, Chris Nauroth <cnaur...@apache.org> wrote:
>>
>> -1 (non-binding)
>>
>> I think I found some missing license information in the binary
>> distribution. We may want to include this in the next RC:
>>
>> https://github.com/apache/spark/pull/50158
>>
>> Thank you for putting together this RC, Wenchen.
>>
>> Chris Nauroth
>>
>>
>> On Mon, Mar 3, 2025 at 6:10 AM Wenchen Fan <cloud0...@gmail.com> wrote:
>>
>>> Thanks for bringing up these blockers! I know RC2 isn’t fully ready yet,
>>> but with over 70 commits since RC1, it’s time to have a new RC so people
>>> can start testing the latest changes. Please continue testing and keep the
>>> feedback coming!
>>>
>>> On Mon, Mar 3, 2025 at 6:06 PM beliefer <belie...@163.com> wrote:
>>>
>>>> -1
>>>> https://github.com/apache/spark/pull/50112 should be merged before
>>>> release.
>>>>
>>>>
>>>> At 2025-03-01 15:25:06, "Wenchen Fan" <cloud0...@gmail.com> wrote:
>>>>
>>>> Please vote on releasing the following candidate as Apache Spark
>>>> version 4.0.0.
>>>>
>>>> The vote is open until March 5 (PST) and passes if a majority +1 PMC
>>>> votes are cast, with a minimum of 3 +1 votes.
>>>>
>>>> [ ] +1 Release this package as Apache Spark 4.0.0
>>>> [ ] -1 Do not release this package because ...
>>>>
>>>> To learn more about Apache Spark, please see https://spark.apache.org/
>>>>
>>>> The tag to be voted on is v4.0.0-rc2 (commit
>>>> 85188c07519ea809012db24421714bb75b45ab1b)
>>>> https://github.com/apache/spark/tree/v4.0.0-rc2
>>>>
>>>> The release files, including signatures, digests, etc. can be found at:
>>>> https://dist.apache.org/repos/dist/dev/spark/v4.0.0-rc2-bin/
>>>>
>>>> Signatures used for Spark RCs can be found in this file:
>>>> https://dist.apache.org/repos/dist/dev/spark/KEYS
>>>>
>>>> The staging repository for this release can be found at:
>>>> https://repository.apache.org/content/repositories/orgapachespark-1478/
>>>>
>>>> The documentation corresponding to this release can be found at:
>>>> https://dist.apache.org/repos/dist/dev/spark/v4.0.0-rc2-docs/
>>>>
>>>> The list of bug fixes going into 4.0.0 can be found at the following
>>>> URL:
>>>> https://issues.apache.org/jira/projects/SPARK/versions/12353359
>>>>
>>>> This release is using the release script of the tag v4.0.0-rc2.
>>>>
>>>> FAQ
>>>>
>>>> =========================
>>>> How can I help test this release?
>>>> =========================
>>>>
>>>> If you are a Spark user, you can help us test this release by taking
>>>> an existing Spark workload and running on this release candidate, then
>>>> reporting any regressions.
>>>>
>>>> If you're working in PySpark you can set up a virtual env and install
>>>> the current RC and see if anything important breaks, in the Java/Scala
>>>> you can add the staging repository to your projects resolvers and test
>>>> with the RC (make sure to clean up the artifact cache before/after so
>>>> you don't end up building with a out of date RC going forward).
>>>>
>>>>
>>

Reply via email to