There is a new thread here.

https://lists.apache.org/thread.html/ra2418b75ac276861a598e7ec943750e2b038c2f8ba49f41db57e5ae9%40%3Cdev.spark.apache.org%3E

Could you share your use case of Hive 1.2 here, Koert?

Bests,
Dongjoon.

On Wed, Oct 7, 2020 at 1:04 PM Koert Kuipers <ko...@tresata.com> wrote:

> i am a little confused about this. i assumed spark would no longer make a
> distribution with hive 1.x, but the hive-1.2 profile remains.
>
> yet i see the hive-1.2 profile has been removed from pom.xml?
>
> On Wed, Sep 23, 2020 at 6:58 PM Dongjoon Hyun <dongjoon.h...@gmail.com>
> wrote:
>
>> Hi, All.
>>
>> Since Apache Spark 3.0.0, Apache Hive 2.3.7 is the default
>> Hive execution library. The forked Hive 1.2.1 library is not
>> recommended because it's not maintained properly.
>>
>> In Apache Spark 3.1 on December 2020, we are going to
>> remove it from our official distribution.
>>
>>     https://github.com/apache/spark/pull/29856
>>     SPARK-32981 Remove hive-1.2/hadoop-2.7 from Apache Spark 3.1
>> distribution
>>
>> Of course, the users still can build it from the source because the
>> profile `hive-1.2` is still alive.
>>
>> Please let us know if you are going to build with the forked unofficial
>> Hive 1.2.1 library still in Apache Spark 3.1. We want to listen to your
>> pain-points before moving forward in this area. Eventually we will remove
>> Hive 1.2 as a last piece of migration to Hive 2.3/Hadoop3/Java11+.
>>
>> Bests,
>> Dongjoon.
>>
>

Reply via email to