[ 
https://issues.apache.org/jira/browse/HIVE-18810?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thejas M Nair resolved HIVE-18810.
----------------------------------
    Resolution: Invalid

> Parquet Or ORC
> --------------
>
>                 Key: HIVE-18810
>                 URL: https://issues.apache.org/jira/browse/HIVE-18810
>             Project: Hive
>          Issue Type: Test
>          Components: Hive
>    Affects Versions: 1.1.0
>         Environment: Hadoop 1.2.1
> Hive 1.1
>            Reporter: Suddhasatwa Bhaumik
>            Priority: Major
>
> Hello Experts, 
> I would like to know for which data types (based on size and complexity of 
> data) should one be using Parquet or ORC tables in Hive. E.g., On Hadoop 
> 0.20.0 with hive 0.13, the performance of ORC tables in Hive is very good 
> when accessed even by 3rd party BI systems like SAP Business Objects or 
> Tableau; performing the same tests on Hadoop 1.2.1 with Hive 1.1 does not 
> yield such reliability in queries, although ETL or insert/update of tables 
> are taking nominal time the read performance is not within acceptable limits. 
> In case of any queries, kindly advise. 
> Thanks
> [~suddhasatwa_bhaumik]



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to