Ya, +1 for Sean's comment.

In addition, all Apache Spark's Maven artifacts are depending on Hadoop
3.3.x already.


https://mvnrepository.com/artifact/org.apache.spark/spark-core_2.12/3.3.0

https://mvnrepository.com/artifact/org.apache.spark/spark-core_2.13/3.3.0

Apache Spark has been moving away from Hadoop 2 due to many many reasons.

Dongjoon.


On Wed, Sep 14, 2022 at 10:54 AM Sean Owen <sro...@gmail.com> wrote:

> Yeah we're not going to make convenience binaries for all possible
> combinations. It's a pretty good assumption that anyone moving to later
> Scala versions is also off old Hadoop versions.
> You can of course build the combo you like.
>
> On Wed, Sep 14, 2022 at 11:26 AM Denis Bolshakov <
> bolshakov.de...@gmail.com> wrote:
>
>> Unfortunately it's for hadoop 3 only.
>>
>> ср, 14 сент. 2022 г., 19:04 Dongjoon Hyun <dongjoon.h...@gmail.com>:
>>
>>> Hi, Denis.
>>>
>>> Apache Spark community already provides both Scala 2.12 and 2.13
>>> pre-built distributions.
>>> Please check the distribution site and Apache Spark download page.
>>>
>>> https://dlcdn.apache.org/spark/spark-3.3.0/
>>>
>>>     spark-3.3.0-bin-hadoop3-scala2.13.tgz
>>>     spark-3.3.0-bin-hadoop3.tgz
>>>
>>> [image: Screenshot 2022-09-14 at 9.03.27 AM.png]
>>>
>>> Dongjoon.
>>>
>>> On Wed, Sep 14, 2022 at 12:31 AM Denis Bolshakov <
>>> bolshakov.de...@gmail.com> wrote:
>>>
>>>> Hello,
>>>>
>>>> It would be great if it's possible to provide a spark distro for both
>>>> scala 2.12 and scala 2.13.
>>>>
>>>> It will encourage spark users to switch to scala 2.13.
>>>>
>>>> I know that spark jar artifacts available for both scala versions, but
>>>> it does not make sense to migrate to scala 2.13 while there is no spark
>>>> distro for this version.
>>>>
>>>> Kind regards,
>>>> Denis
>>>>
>>>> On Tue, 13 Sept 2022 at 17:38, Yuming Wang <wgy...@gmail.com> wrote:
>>>>
>>>>> Thank you all.
>>>>>
>>>>> I will be preparing 3.3.1 RC1 soon.
>>>>>
>>>>> On Tue, Sep 13, 2022 at 12:09 PM John Zhuge <jzh...@apache.org> wrote:
>>>>>
>>>>>> +1
>>>>>>
>>>>>> On Mon, Sep 12, 2022 at 9:08 PM Yang,Jie(INF) <yangji...@baidu.com>
>>>>>> wrote:
>>>>>>
>>>>>>> +1
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> Thanks Yuming ~
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> *发件人**: *Hyukjin Kwon <gurwls...@gmail.com>
>>>>>>> *日期**: *2022年9月13日 星期二 08:19
>>>>>>> *收件人**: *Gengliang Wang <ltn...@gmail.com>
>>>>>>> *抄送**: *"L. C. Hsieh" <vii...@gmail.com>, Dongjoon Hyun <
>>>>>>> dongjoon.h...@gmail.com>, Yuming Wang <wgy...@gmail.com>, dev <
>>>>>>> dev@spark.apache.org>
>>>>>>> *主题**: *Re: Time for Spark 3.3.1 release?
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> +1
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Tue, 13 Sept 2022 at 06:45, Gengliang Wang <ltn...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>> +1.
>>>>>>>
>>>>>>> Thank you, Yuming!
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Mon, Sep 12, 2022 at 12:10 PM L. C. Hsieh <vii...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>> +1
>>>>>>>
>>>>>>> Thanks Yuming!
>>>>>>>
>>>>>>> On Mon, Sep 12, 2022 at 11:50 AM Dongjoon Hyun <
>>>>>>> dongjoon.h...@gmail.com> wrote:
>>>>>>> >
>>>>>>> > +1
>>>>>>> >
>>>>>>> > Thanks,
>>>>>>> > Dongjoon.
>>>>>>> >
>>>>>>> > On Mon, Sep 12, 2022 at 6:38 AM Yuming Wang <wgy...@gmail.com>
>>>>>>> wrote:
>>>>>>> >>
>>>>>>> >> Hi, All.
>>>>>>> >>
>>>>>>> >>
>>>>>>> >>
>>>>>>> >> Since Apache Spark 3.3.0 tag creation (Jun 10), new 138 patches
>>>>>>> including 7 correctness patches arrived at branch-3.3.
>>>>>>> >>
>>>>>>> >>
>>>>>>> >>
>>>>>>> >> Shall we make a new release, Apache Spark 3.3.1, as the second
>>>>>>> release at branch-3.3? I'd like to volunteer as the release manager for
>>>>>>> Apache Spark 3.3.1.
>>>>>>> >>
>>>>>>> >>
>>>>>>> >>
>>>>>>> >> All changes:
>>>>>>> >>
>>>>>>> >> https://github.com/apache/spark/compare/v3.3.0...branch-3.3
>>>>>>> <https://mailshield.baidu.com/check?q=WzRnV6InLAPdBDRyJZecGtPwF02t%2bnFNwOI8oAyGcb60kX%2bRCS6N3SUnFjTdf47bb94KmZHTTKE%2bBHbIT27Rog%3d%3d>
>>>>>>> >>
>>>>>>> >>
>>>>>>> >>
>>>>>>> >> Correctness issues:
>>>>>>> >>
>>>>>>> >> SPARK-40149: Propagate metadata columns through Project
>>>>>>> >>
>>>>>>> >> SPARK-40002: Don't push down limit through window using ntile
>>>>>>> >>
>>>>>>> >> SPARK-39976: ArrayIntersect should handle null in left expression
>>>>>>> correctly
>>>>>>> >>
>>>>>>> >> SPARK-39833: Disable Parquet column index in DSv1 to fix a
>>>>>>> correctness issue in the case of overlapping partition and data columns
>>>>>>> >>
>>>>>>> >> SPARK-39061: Set nullable correctly for Inline output attributes
>>>>>>> >>
>>>>>>> >> SPARK-39887: RemoveRedundantAliases should keep aliases that make
>>>>>>> the output of projection nodes unique
>>>>>>> >>
>>>>>>> >> SPARK-38614: Don't push down limit through window that's using
>>>>>>> percent_rank
>>>>>>>
>>>>>>> ---------------------------------------------------------------------
>>>>>>> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>>>>>>>
>>>>>>> --
>>>>>> John Zhuge
>>>>>>
>>>>>
>>>>
>>>> --
>>>> //with Best Regards
>>>> --Denis Bolshakov
>>>> e-mail: bolshakov.de...@gmail.com
>>>>
>>>

Reply via email to