Thanks for your feedbacks!

Working with Yuming to reduce the risk of stability and quality. Will keep
you posted when the proposal is ready.

Cheers,

Xiao

Ryan Blue <rb...@netflix.com> 于2019年1月16日周三 上午9:27写道:

> +1 for what Marcelo and Hyukjin said.
>
> In particular, I agree that we can't expect Hive to release a version that
> is now more than 3 years old just to solve a problem for Spark. Maybe that
> would have been a reasonable ask instead of publishing a fork years ago,
> but I think this is now Spark's problem.
>
> On Tue, Jan 15, 2019 at 9:02 PM Marcelo Vanzin <van...@cloudera.com>
> wrote:
>
>> +1 to that. HIVE-16391 by itself means we're giving up things like
>> Hadoop 3, and we're also putting the burden on the Hive folks to fix a
>> problem that we created.
>>
>> The current PR is basically a Spark-side fix for that bug. It does
>> mean also upgrading Hive (which gives us Hadoop 3, yay!), but I think
>> it's really the right path to take here.
>>
>> On Tue, Jan 15, 2019 at 6:32 PM Hyukjin Kwon <gurwls...@gmail.com> wrote:
>> >
>> > Resolving HIVE-16391 means Hive to release 1.2.x that contains the
>> fixes of our Hive fork (correct me if I am mistaken).
>> >
>> > Just to be honest by myself and as a personal opinion, that basically
>> says Hive to take care of Spark's dependency.
>> > Hive looks going ahead for 3.1.x and no one would use the newer release
>> of 1.2.x. In practice, Spark doesn't make a release 1.6.x anymore for
>> instance,
>> >
>> > Frankly, my impression was that it's, honestly, our mistake to fix.
>> Since Spark community is big enough, I was thinking we should try to fix it
>> by ourselves first.
>> > I am not saying upgrading is the only way to get through this but I
>> think we should at least try first, and see what's next.
>> >
>> > It does, yes, sound more risky to upgrade it in our side but I think
>> it's worth to check and try it and see if it's possible.
>> > I think this is a standard approach to upgrade the dependency than
>> using the fork or letting Hive side to release another 1.2.x.
>> >
>> > If we fail to upgrade it for critical or inevitable reasons somehow,
>> yes, we could find an alternative but that basically means
>> > we're going to stay in 1.2.x for, at least, a long time (say .. until
>> Spark 4.0.0?).
>> >
>> > I know somehow it happened to be sensitive but to be just literally
>> honest to myself, I think we should make a try.
>> >
>>
>>
>> --
>> Marcelo
>>
>
>
> --
> Ryan Blue
> Software Engineer
> Netflix
>

Reply via email to