Yeah we're not going to make convenience binaries for all possible
combinations. It's a pretty good assumption that anyone moving to later
Scala versions is also off old Hadoop versions.
You can of course build the combo you like.

On Wed, Sep 14, 2022 at 11:26 AM Denis Bolshakov <bolshakov.de...@gmail.com>
wrote:

> Unfortunately it's for hadoop 3 only.
>
> ср, 14 сент. 2022 г., 19:04 Dongjoon Hyun <dongjoon.h...@gmail.com>:
>
>> Hi, Denis.
>>
>> Apache Spark community already provides both Scala 2.12 and 2.13
>> pre-built distributions.
>> Please check the distribution site and Apache Spark download page.
>>
>> https://dlcdn.apache.org/spark/spark-3.3.0/
>>
>>     spark-3.3.0-bin-hadoop3-scala2.13.tgz
>>     spark-3.3.0-bin-hadoop3.tgz
>>
>> [image: Screenshot 2022-09-14 at 9.03.27 AM.png]
>>
>> Dongjoon.
>>
>> On Wed, Sep 14, 2022 at 12:31 AM Denis Bolshakov <
>> bolshakov.de...@gmail.com> wrote:
>>
>>> Hello,
>>>
>>> It would be great if it's possible to provide a spark distro for both
>>> scala 2.12 and scala 2.13.
>>>
>>> It will encourage spark users to switch to scala 2.13.
>>>
>>> I know that spark jar artifacts available for both scala versions, but
>>> it does not make sense to migrate to scala 2.13 while there is no spark
>>> distro for this version.
>>>
>>> Kind regards,
>>> Denis
>>>
>>> On Tue, 13 Sept 2022 at 17:38, Yuming Wang <wgy...@gmail.com> wrote:
>>>
>>>> Thank you all.
>>>>
>>>> I will be preparing 3.3.1 RC1 soon.
>>>>
>>>> On Tue, Sep 13, 2022 at 12:09 PM John Zhuge <jzh...@apache.org> wrote:
>>>>
>>>>> +1
>>>>>
>>>>> On Mon, Sep 12, 2022 at 9:08 PM Yang,Jie(INF) <yangji...@baidu.com>
>>>>> wrote:
>>>>>
>>>>>> +1
>>>>>>
>>>>>>
>>>>>>
>>>>>> Thanks Yuming ~
>>>>>>
>>>>>>
>>>>>>
>>>>>> *发件人**: *Hyukjin Kwon <gurwls...@gmail.com>
>>>>>> *日期**: *2022年9月13日 星期二 08:19
>>>>>> *收件人**: *Gengliang Wang <ltn...@gmail.com>
>>>>>> *抄送**: *"L. C. Hsieh" <vii...@gmail.com>, Dongjoon Hyun <
>>>>>> dongjoon.h...@gmail.com>, Yuming Wang <wgy...@gmail.com>, dev <
>>>>>> dev@spark.apache.org>
>>>>>> *主题**: *Re: Time for Spark 3.3.1 release?
>>>>>>
>>>>>>
>>>>>>
>>>>>> +1
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Tue, 13 Sept 2022 at 06:45, Gengliang Wang <ltn...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>> +1.
>>>>>>
>>>>>> Thank you, Yuming!
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Mon, Sep 12, 2022 at 12:10 PM L. C. Hsieh <vii...@gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>> +1
>>>>>>
>>>>>> Thanks Yuming!
>>>>>>
>>>>>> On Mon, Sep 12, 2022 at 11:50 AM Dongjoon Hyun <
>>>>>> dongjoon.h...@gmail.com> wrote:
>>>>>> >
>>>>>> > +1
>>>>>> >
>>>>>> > Thanks,
>>>>>> > Dongjoon.
>>>>>> >
>>>>>> > On Mon, Sep 12, 2022 at 6:38 AM Yuming Wang <wgy...@gmail.com>
>>>>>> wrote:
>>>>>> >>
>>>>>> >> Hi, All.
>>>>>> >>
>>>>>> >>
>>>>>> >>
>>>>>> >> Since Apache Spark 3.3.0 tag creation (Jun 10), new 138 patches
>>>>>> including 7 correctness patches arrived at branch-3.3.
>>>>>> >>
>>>>>> >>
>>>>>> >>
>>>>>> >> Shall we make a new release, Apache Spark 3.3.1, as the second
>>>>>> release at branch-3.3? I'd like to volunteer as the release manager for
>>>>>> Apache Spark 3.3.1.
>>>>>> >>
>>>>>> >>
>>>>>> >>
>>>>>> >> All changes:
>>>>>> >>
>>>>>> >> https://github.com/apache/spark/compare/v3.3.0...branch-3.3
>>>>>> <https://mailshield.baidu.com/check?q=WzRnV6InLAPdBDRyJZecGtPwF02t%2bnFNwOI8oAyGcb60kX%2bRCS6N3SUnFjTdf47bb94KmZHTTKE%2bBHbIT27Rog%3d%3d>
>>>>>> >>
>>>>>> >>
>>>>>> >>
>>>>>> >> Correctness issues:
>>>>>> >>
>>>>>> >> SPARK-40149: Propagate metadata columns through Project
>>>>>> >>
>>>>>> >> SPARK-40002: Don't push down limit through window using ntile
>>>>>> >>
>>>>>> >> SPARK-39976: ArrayIntersect should handle null in left expression
>>>>>> correctly
>>>>>> >>
>>>>>> >> SPARK-39833: Disable Parquet column index in DSv1 to fix a
>>>>>> correctness issue in the case of overlapping partition and data columns
>>>>>> >>
>>>>>> >> SPARK-39061: Set nullable correctly for Inline output attributes
>>>>>> >>
>>>>>> >> SPARK-39887: RemoveRedundantAliases should keep aliases that make
>>>>>> the output of projection nodes unique
>>>>>> >>
>>>>>> >> SPARK-38614: Don't push down limit through window that's using
>>>>>> percent_rank
>>>>>>
>>>>>> ---------------------------------------------------------------------
>>>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>>>>
>>>>>> --
>>>>> John Zhuge
>>>>>
>>>>
>>>
>>> --
>>> //with Best Regards
>>> --Denis Bolshakov
>>> e-mail: bolshakov.de...@gmail.com
>>>
>>

Reply via email to