Unfortunately it's for hadoop 3 only.

ср, 14 сент. 2022 г., 19:04 Dongjoon Hyun <dongjoon.h...@gmail.com>:

> Hi, Denis.
>
> Apache Spark community already provides both Scala 2.12 and 2.13 pre-built
> distributions.
> Please check the distribution site and Apache Spark download page.
>
> https://dlcdn.apache.org/spark/spark-3.3.0/
>
>     spark-3.3.0-bin-hadoop3-scala2.13.tgz
>     spark-3.3.0-bin-hadoop3.tgz
>
> [image: Screenshot 2022-09-14 at 9.03.27 AM.png]
>
> Dongjoon.
>
> On Wed, Sep 14, 2022 at 12:31 AM Denis Bolshakov <
> bolshakov.de...@gmail.com> wrote:
>
>> Hello,
>>
>> It would be great if it's possible to provide a spark distro for both
>> scala 2.12 and scala 2.13.
>>
>> It will encourage spark users to switch to scala 2.13.
>>
>> I know that spark jar artifacts available for both scala versions, but it
>> does not make sense to migrate to scala 2.13 while there is no spark distro
>> for this version.
>>
>> Kind regards,
>> Denis
>>
>> On Tue, 13 Sept 2022 at 17:38, Yuming Wang <wgy...@gmail.com> wrote:
>>
>>> Thank you all.
>>>
>>> I will be preparing 3.3.1 RC1 soon.
>>>
>>> On Tue, Sep 13, 2022 at 12:09 PM John Zhuge <jzh...@apache.org> wrote:
>>>
>>>> +1
>>>>
>>>> On Mon, Sep 12, 2022 at 9:08 PM Yang,Jie(INF) <yangji...@baidu.com>
>>>> wrote:
>>>>
>>>>> +1
>>>>>
>>>>>
>>>>>
>>>>> Thanks Yuming ~
>>>>>
>>>>>
>>>>>
>>>>> *发件人**: *Hyukjin Kwon <gurwls...@gmail.com>
>>>>> *日期**: *2022年9月13日 星期二 08:19
>>>>> *收件人**: *Gengliang Wang <ltn...@gmail.com>
>>>>> *抄送**: *"L. C. Hsieh" <vii...@gmail.com>, Dongjoon Hyun <
>>>>> dongjoon.h...@gmail.com>, Yuming Wang <wgy...@gmail.com>, dev <
>>>>> dev@spark.apache.org>
>>>>> *主题**: *Re: Time for Spark 3.3.1 release?
>>>>>
>>>>>
>>>>>
>>>>> +1
>>>>>
>>>>>
>>>>>
>>>>> On Tue, 13 Sept 2022 at 06:45, Gengliang Wang <ltn...@gmail.com>
>>>>> wrote:
>>>>>
>>>>> +1.
>>>>>
>>>>> Thank you, Yuming!
>>>>>
>>>>>
>>>>>
>>>>> On Mon, Sep 12, 2022 at 12:10 PM L. C. Hsieh <vii...@gmail.com> wrote:
>>>>>
>>>>> +1
>>>>>
>>>>> Thanks Yuming!
>>>>>
>>>>> On Mon, Sep 12, 2022 at 11:50 AM Dongjoon Hyun <
>>>>> dongjoon.h...@gmail.com> wrote:
>>>>> >
>>>>> > +1
>>>>> >
>>>>> > Thanks,
>>>>> > Dongjoon.
>>>>> >
>>>>> > On Mon, Sep 12, 2022 at 6:38 AM Yuming Wang <wgy...@gmail.com>
>>>>> wrote:
>>>>> >>
>>>>> >> Hi, All.
>>>>> >>
>>>>> >>
>>>>> >>
>>>>> >> Since Apache Spark 3.3.0 tag creation (Jun 10), new 138 patches
>>>>> including 7 correctness patches arrived at branch-3.3.
>>>>> >>
>>>>> >>
>>>>> >>
>>>>> >> Shall we make a new release, Apache Spark 3.3.1, as the second
>>>>> release at branch-3.3? I'd like to volunteer as the release manager for
>>>>> Apache Spark 3.3.1.
>>>>> >>
>>>>> >>
>>>>> >>
>>>>> >> All changes:
>>>>> >>
>>>>> >> https://github.com/apache/spark/compare/v3.3.0...branch-3.3
>>>>> <https://mailshield.baidu.com/check?q=WzRnV6InLAPdBDRyJZecGtPwF02t%2bnFNwOI8oAyGcb60kX%2bRCS6N3SUnFjTdf47bb94KmZHTTKE%2bBHbIT27Rog%3d%3d>
>>>>> >>
>>>>> >>
>>>>> >>
>>>>> >> Correctness issues:
>>>>> >>
>>>>> >> SPARK-40149: Propagate metadata columns through Project
>>>>> >>
>>>>> >> SPARK-40002: Don't push down limit through window using ntile
>>>>> >>
>>>>> >> SPARK-39976: ArrayIntersect should handle null in left expression
>>>>> correctly
>>>>> >>
>>>>> >> SPARK-39833: Disable Parquet column index in DSv1 to fix a
>>>>> correctness issue in the case of overlapping partition and data columns
>>>>> >>
>>>>> >> SPARK-39061: Set nullable correctly for Inline output attributes
>>>>> >>
>>>>> >> SPARK-39887: RemoveRedundantAliases should keep aliases that make
>>>>> the output of projection nodes unique
>>>>> >>
>>>>> >> SPARK-38614: Don't push down limit through window that's using
>>>>> percent_rank
>>>>>
>>>>> ---------------------------------------------------------------------
>>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>>>
>>>>> --
>>>> John Zhuge
>>>>
>>>
>>
>> --
>> //with Best Regards
>> --Denis Bolshakov
>> e-mail: bolshakov.de...@gmail.com
>>
>

Reply via email to