Thank you, Sean.

Ya, exactly, we can release 2.4.8 as a normal release first and use 2.4.9
as the EOL release.

Since 2.4.7 was released almost 6 months ago, 2.4.8 is a little late in
terms of the cadence.

Bests,
Dongjoon.


On Wed, Mar 3, 2021 at 10:55 AM Sean Owen <sro...@gmail.com> wrote:

> For reference, 2.3.x was maintained from February 2018 (2.3.0) to Sep 2019
> (2.3.4), or about 19 months. The 2.4 branch should probably be maintained
> longer than that, as the final 2.x branch. 2.4.0 was released in Nov 2018.
> A final release in, say, April 2021 would be about 30 months. That feels
> about right timing-wise.
>
> We should in any event release 2.4.8, yes. We can of course choose to
> release a 2.4.9 if some critical issue is found, later.
>
> But yeah based on the velocity of back-ports to 2.4.x, it seems about time
> to call it EOL.
>
> Sean
>
>
> On Wed, Mar 3, 2021 at 12:05 PM Dongjoon Hyun <dongjoon.h...@gmail.com>
> wrote:
>
>> Hi, All.
>>
>> We successfully completed Apache Spark 3.1.1 and 3.0.2 releases and
>> started 3.2.0 discussion already.
>>
>> Let's talk about branch-2.4 because there exists some discussions on JIRA
>> and GitHub about skipping backporting to 2.4.
>>
>> Since `branch-2.4` has been maintained well as LTS, I'd like to suggest
>> having Apache Spark 2.4.8 release as the official EOL release of 2.4 line
>> in order to focus on 3.x more from now. Please note that `branch-2.4` will
>> be frozen officially like `branch-2.3` after EOL release.
>>
>> - Apache Spark 2.4.0 was released on November 2, 2018.
>> - Apache Spark 2.4.7 was released on September 12, 2020.
>> - Since v2.4.7 tag, `branch-2.4` has 134 commits including the following
>> 12 correctness issues.
>>
>> ## CORRECTNESS ISSUE
>> SPARK-30201 HiveOutputWriter standardOI should use
>> ObjectInspectorCopyOption.DEFAULT
>> SPARK-30228 Update zstd-jni to 1.4.4-3
>> SPARK-30894 The nullability of Size function should not depend on
>> SQLConf.get
>> SPARK-32635 When pyspark.sql.functions.lit() function is used with
>> dataframe cache, it returns wrong result
>> SPARK-32908 percentile_approx() returns incorrect results
>> SPARK-33183 Bug in optimizer rule EliminateSorts
>> SPARK-33290 REFRESH TABLE should invalidate cache even though the table
>> itself may not be cached
>> SPARK-33593 Vector reader got incorrect data with binary partition value
>> SPARK-33726 Duplicate field names causes wrong answers during aggregation
>> SPARK-34187 Use available offset range obtained during polling when
>> checking offset validation
>> SPARK-34212 For parquet table, after changing the precision and scale of
>> decimal type in hive, spark reads incorrect value
>> SPARK-34229 Avro should read decimal values with the file schema
>>
>> ## SECURITY ISSUE
>> SPARK-33333 Upgrade Jetty to 9.4.28.v20200408
>> SPARK-33831 Update to jetty 9.4.34
>> SPARK-34449 Upgrade Jetty to fix CVE-2020-27218
>>
>> What do you think about this?
>>
>> Bests,
>> Dongjoon.
>>
>

Reply via email to