Yes, it sounds good to me. We can upgrade both Parquet 1.8.2 to 1.8.3 and
ORC 1.4.1 to 1.4.3 in our upcoming Spark 2.3.1 release.

Thanks for your efforts! @Henry and @Dongjoon

Xiao

2018-04-16 14:41 GMT-07:00 Henry Robinson <he...@apache.org>:

> Seems like there aren't any objections. I'll pick this thread back up when
> a Parquet maintenance release has happened.
>
> Henry
>
> On 11 April 2018 at 14:00, Dongjoon Hyun <dongjoon.h...@gmail.com> wrote:
>
>> Great.
>>
>> If we can upgrade the parquet dependency from 1.8.2 to 1.8.3 in Apache
>> Spark 2.3.1, let's upgrade orc dependency from 1.4.1 to 1.4.3 together.
>>
>> Currently, the patch is only merged into master branch now. 1.4.1 has the
>> following issue.
>>
>> https://issues.apache.org/jira/browse/SPARK-23340
>>
>> Bests,
>> Dongjoon.
>>
>>
>>
>> On Wed, Apr 11, 2018 at 1:23 PM, Reynold Xin <r...@databricks.com> wrote:
>>
>>> Seems like this would make sense... we usually make maintenance releases
>>> for bug fixes after a month anyway.
>>>
>>>
>>> On Wed, Apr 11, 2018 at 12:52 PM, Henry Robinson <he...@apache.org>
>>> wrote:
>>>
>>>>
>>>>
>>>> On 11 April 2018 at 12:47, Ryan Blue <rb...@netflix.com.invalid> wrote:
>>>>
>>>>> I think a 1.8.3 Parquet release makes sense for the 2.3.x releases of
>>>>> Spark.
>>>>>
>>>>> To be clear though, this only affects Spark when reading data written
>>>>> by Impala, right? Or does Parquet CPP also produce data like this?
>>>>>
>>>>
>>>> I don't know about parquet-cpp, but yeah, the only implementation I've
>>>> seen writing the half-completed stats is Impala. (as you know, that's
>>>> compliant with the spec, just an unusual choice).
>>>>
>>>>
>>>>>
>>>>> On Wed, Apr 11, 2018 at 12:35 PM, Henry Robinson <he...@apache.org>
>>>>> wrote:
>>>>>
>>>>>> Hi all -
>>>>>>
>>>>>> SPARK-23852 (where a query can silently give wrong results thanks to
>>>>>> a predicate pushdown bug in Parquet) is a fairly bad bug. In other 
>>>>>> projects
>>>>>> I've been involved with, we've released maintenance releases for bugs of
>>>>>> this severity.
>>>>>>
>>>>>> Since Spark 2.4.0 is probably a while away, I wanted to see if there
>>>>>> was any consensus over whether we should consider (at least) a 2.3.1.
>>>>>>
>>>>>> The reason this particular issue is a bit tricky is that the Parquet
>>>>>> community haven't yet produced a maintenance release that fixes the
>>>>>> underlying bug, but they are in the process of releasing a new minor
>>>>>> version, 1.10, which includes a fix. Having spoken to a couple of Parquet
>>>>>> developers, they'd be willing to consider a maintenance release, but 
>>>>>> would
>>>>>> probably only bother if we (or another affected project) asked them to.
>>>>>>
>>>>>> My guess is that we wouldn't want to upgrade to a new minor version
>>>>>> of Parquet for a Spark maintenance release, so asking for a Parquet
>>>>>> maintenance release makes sense.
>>>>>>
>>>>>> What does everyone think?
>>>>>>
>>>>>> Best,
>>>>>> Henry
>>>>>>
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Ryan Blue
>>>>> Software Engineer
>>>>> Netflix
>>>>>
>>>>
>>>>
>>>
>>
>

Reply via email to