GitHub user sameeragarwal opened a pull request:

    https://github.com/apache/spark/pull/11869

    [SPARK-14016][SQL] Support high-precision decimals in vectorized parquet 
reader

    ## What changes were proposed in this pull request?
    
    This patch adds support for reading `DecimalTypes` with high (> 18) 
precision in `VectorizedColumnReader`
    
    ## How was this patch tested?
    
    1. `VectorizedColumnReader` initially had a gating condition on 
`primitiveType.getDecimalMetadata().getPrecision() > Decimal.MAX_LONG_DIGITS()` 
that made us fall back on parquet-mr for handling high-precision decimals. This 
condition is now removed.
    2. In particular, the `ParquetHadoopFsRelationSuite` (that tests for all 
supported hive types -- including `DecimalType(25, 5)`) fails when the gating 
condition is removed (https://github.com/apache/spark/pull/11808) and should 
now pass with this change.

You can merge this pull request into a Git repository by running:

    $ git pull https://github.com/sameeragarwal/spark bigdecimal-parquet

Alternatively you can review and apply these changes as the patch at:

    https://github.com/apache/spark/pull/11869.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

    This closes #11869
    
----
commit cbccee7951ed0401efc80469902bf6380ea9190b
Author: Sameer Agarwal <[email protected]>
Date:   2016-03-21T23:19:11Z

    Support high-precision decimals

----


---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to