HIVE-11096: Bump the parquet version to 1.7.0 (Sergio Pena and Ferdinand Xu, reviewed by Ferdinand Xu)
Project: http://git-wip-us.apache.org/repos/asf/hive/repo Commit: http://git-wip-us.apache.org/repos/asf/hive/commit/11020ae5 Tree: http://git-wip-us.apache.org/repos/asf/hive/tree/11020ae5 Diff: http://git-wip-us.apache.org/repos/asf/hive/diff/11020ae5 Branch: refs/heads/beeline-cli Commit: 11020ae562aba4c4097556f2e06c6783d211c041 Parents: ebd294a Author: Ferdinand Xu <[email protected]> Authored: Wed Jun 24 13:43:35 2015 -0400 Committer: Ferdinand Xu <[email protected]> Committed: Wed Jun 24 13:43:35 2015 -0400 ---------------------------------------------------------------------- pom.xml | 8 ++--- ql/pom.xml | 7 +++-- .../io/parquet/FilterPredicateLeafBuilder.java | 8 ++--- .../hive/ql/io/parquet/LeafFilterFactory.java | 20 ++++++------- .../ql/io/parquet/MapredParquetInputFormat.java | 3 +- .../io/parquet/MapredParquetOutputFormat.java | 2 +- .../parquet/VectorizedParquetInputFormat.java | 2 +- .../convert/DataWritableRecordConverter.java | 8 ++--- .../ql/io/parquet/convert/ETypeConverter.java | 10 +++---- .../convert/HiveCollectionConverter.java | 6 ++-- .../io/parquet/convert/HiveGroupConverter.java | 14 ++++----- .../io/parquet/convert/HiveSchemaConverter.java | 18 ++++++------ .../io/parquet/convert/HiveStructConverter.java | 6 ++-- .../hive/ql/io/parquet/convert/Repeated.java | 12 ++++---- .../parquet/read/DataWritableReadSupport.java | 20 ++++++------- .../read/ParquetRecordReaderWrapper.java | 31 ++++++++------------ .../ql/io/parquet/serde/ParquetHiveSerDe.java | 4 +-- .../hive/ql/io/parquet/timestamp/NanoTime.java | 6 ++-- .../parquet/write/DataWritableWriteSupport.java | 8 ++--- .../ql/io/parquet/write/DataWritableWriter.java | 10 +++---- .../write/ParquetRecordWriterWrapper.java | 6 ++-- .../hive/ql/io/sarg/SearchArgumentImpl.java | 4 +-- .../io/parquet/AbstractTestParquetDirect.java | 8 ++--- .../ql/io/parquet/TestArrayCompatibility.java | 18 ++++++------ .../ql/io/parquet/TestDataWritableWriter.java | 8 ++--- .../ql/io/parquet/TestHiveSchemaConverter.java | 16 +++++----- .../hive/ql/io/parquet/TestMapStructures.java | 10 +++---- .../parquet/TestMapredParquetInputFormat.java | 2 +- .../parquet/TestMapredParquetOutputFormat.java | 2 +- .../io/parquet/TestParquetRowGroupFilter.java | 6 ++-- .../hive/ql/io/sarg/TestSearchArgumentImpl.java | 2 +- serde/pom.xml | 2 +- .../hadoop/hive/ql/io/sarg/SearchArgument.java | 2 +- 33 files changed, 142 insertions(+), 147 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/pom.xml ---------------------------------------------------------------------- diff --git a/pom.xml b/pom.xml index 9014d70..1c9b02f 100644 --- a/pom.xml +++ b/pom.xml @@ -152,7 +152,7 @@ <mockito-all.version>1.9.5</mockito-all.version> <mina.version>2.0.0-M5</mina.version> <netty.version>4.0.23.Final</netty.version> - <parquet.version>1.6.0</parquet.version> + <parquet.version>1.7.0</parquet.version> <pig.version>0.12.0</pig.version> <protobuf.version>2.5.0</protobuf.version> <stax.version>1.0.1</stax.version> @@ -257,12 +257,12 @@ <version>${bonecp.version}</version> </dependency> <dependency> - <groupId>com.twitter</groupId> - <artifactId>parquet-hadoop-bundle</artifactId> + <groupId>org.apache.parquet</groupId> + <artifactId>parquet</artifactId> <version>${parquet.version}</version> </dependency> <dependency> - <groupId>com.twitter</groupId> + <groupId>org.apache.parquet</groupId> <artifactId>parquet-column</artifactId> <version>${parquet.version}</version> <classifier>tests</classifier> http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/pom.xml ---------------------------------------------------------------------- diff --git a/ql/pom.xml b/ql/pom.xml index b076c26..b86b85d 100644 --- a/ql/pom.xml +++ b/ql/pom.xml @@ -72,8 +72,9 @@ <version>${kryo.version}</version> </dependency> <dependency> - <groupId>com.twitter</groupId> + <groupId>org.apache.parquet</groupId> <artifactId>parquet-hadoop-bundle</artifactId> + <version>${parquet.version}</version> </dependency> <dependency> <groupId>commons-codec</groupId> @@ -309,7 +310,7 @@ <!-- test intra-project --> <!-- test inter-project --> <dependency> - <groupId>com.twitter</groupId> + <groupId>org.apache.parquet</groupId> <artifactId>parquet-column</artifactId> <classifier>tests</classifier> <scope>test</scope> @@ -689,7 +690,7 @@ <include>org.apache.hive:hive-exec</include> <include>org.apache.hive:hive-serde</include> <include>com.esotericsoftware.kryo:kryo</include> - <include>com.twitter:parquet-hadoop-bundle</include> + <include>org.apache.parquet:parquet-hadoop-bundle</include> <include>org.apache.thrift:libthrift</include> <include>commons-lang:commons-lang</include> <include>org.apache.commons:commons-lang3</include> http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/FilterPredicateLeafBuilder.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/FilterPredicateLeafBuilder.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/FilterPredicateLeafBuilder.java index 2797654..fe2094a 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/FilterPredicateLeafBuilder.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/FilterPredicateLeafBuilder.java @@ -16,11 +16,11 @@ package org.apache.hadoop.hive.ql.io.parquet; import java.util.List; import org.apache.hadoop.hive.ql.io.sarg.PredicateLeaf; -import parquet.filter2.predicate.FilterApi; -import parquet.filter2.predicate.FilterPredicate; +import org.apache.parquet.filter2.predicate.FilterApi; +import org.apache.parquet.filter2.predicate.FilterPredicate; -import static parquet.filter2.predicate.FilterApi.not; -import static parquet.filter2.predicate.FilterApi.or; +import static org.apache.parquet.filter2.predicate.FilterApi.not; +import static org.apache.parquet.filter2.predicate.FilterApi.or; /** * The base class for building parquet supported filter predicate in primary types. http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/LeafFilterFactory.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/LeafFilterFactory.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/LeafFilterFactory.java index 83865e8..a1dbc1a 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/LeafFilterFactory.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/LeafFilterFactory.java @@ -19,17 +19,17 @@ import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hive.ql.io.sarg.PredicateLeaf; import org.apache.hadoop.hive.ql.io.sarg.PredicateLeaf.Operator; -import parquet.filter2.predicate.FilterApi; -import parquet.filter2.predicate.FilterPredicate; -import parquet.io.api.Binary; +import org.apache.parquet.filter2.predicate.FilterApi; +import org.apache.parquet.filter2.predicate.FilterPredicate; +import org.apache.parquet.io.api.Binary; -import static parquet.filter2.predicate.FilterApi.eq; -import static parquet.filter2.predicate.FilterApi.lt; -import static parquet.filter2.predicate.FilterApi.ltEq; -import static parquet.filter2.predicate.FilterApi.binaryColumn; -import static parquet.filter2.predicate.FilterApi.booleanColumn; -import static parquet.filter2.predicate.FilterApi.doubleColumn; -import static parquet.filter2.predicate.FilterApi.intColumn; +import static org.apache.parquet.filter2.predicate.FilterApi.eq; +import static org.apache.parquet.filter2.predicate.FilterApi.lt; +import static org.apache.parquet.filter2.predicate.FilterApi.ltEq; +import static org.apache.parquet.filter2.predicate.FilterApi.binaryColumn; +import static org.apache.parquet.filter2.predicate.FilterApi.booleanColumn; +import static org.apache.parquet.filter2.predicate.FilterApi.doubleColumn; +import static org.apache.parquet.filter2.predicate.FilterApi.intColumn; public class LeafFilterFactory { private static final Log LOG = LogFactory.getLog(LeafFilterFactory.class); http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/MapredParquetInputFormat.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/MapredParquetInputFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/MapredParquetInputFormat.java index f4f0f07..d82e93c 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/MapredParquetInputFormat.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/MapredParquetInputFormat.java @@ -17,7 +17,6 @@ import java.io.IOException; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.hive.ql.exec.Utilities; -import org.apache.hadoop.hive.ql.exec.vector.VectorizedInputFormatInterface; import org.apache.hadoop.hive.ql.io.parquet.read.DataWritableReadSupport; import org.apache.hadoop.hive.ql.io.parquet.read.ParquetRecordReaderWrapper; import org.apache.hadoop.io.ArrayWritable; @@ -25,7 +24,7 @@ import org.apache.hadoop.io.NullWritable; import org.apache.hadoop.mapred.FileInputFormat; import org.apache.hadoop.mapred.RecordReader; -import parquet.hadoop.ParquetInputFormat; +import org.apache.parquet.hadoop.ParquetInputFormat; /** http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/MapredParquetOutputFormat.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/MapredParquetOutputFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/MapredParquetOutputFormat.java index c6fb26c..5e71df9 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/MapredParquetOutputFormat.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/MapredParquetOutputFormat.java @@ -40,7 +40,7 @@ import org.apache.hadoop.mapred.RecordWriter; import org.apache.hadoop.mapreduce.OutputFormat; import org.apache.hadoop.util.Progressable; -import parquet.hadoop.ParquetOutputFormat; +import org.apache.parquet.hadoop.ParquetOutputFormat; /** * http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/VectorizedParquetInputFormat.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/VectorizedParquetInputFormat.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/VectorizedParquetInputFormat.java index 843e079..ed99615 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/VectorizedParquetInputFormat.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/VectorizedParquetInputFormat.java @@ -33,7 +33,7 @@ import org.apache.hadoop.mapred.JobConf; import org.apache.hadoop.mapred.RecordReader; import org.apache.hadoop.mapred.Reporter; -import parquet.hadoop.ParquetInputFormat; +import org.apache.parquet.hadoop.ParquetInputFormat; /** * Vectorized input format for Parquet files http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/DataWritableRecordConverter.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/DataWritableRecordConverter.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/DataWritableRecordConverter.java index e9d1131..3261e4b 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/DataWritableRecordConverter.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/DataWritableRecordConverter.java @@ -15,10 +15,10 @@ package org.apache.hadoop.hive.ql.io.parquet.convert; import org.apache.hadoop.hive.ql.io.parquet.read.DataWritableReadSupport; import org.apache.hadoop.io.ArrayWritable; -import parquet.io.api.GroupConverter; -import parquet.io.api.RecordMaterializer; -import parquet.schema.GroupType; -import parquet.schema.MessageTypeParser; +import org.apache.parquet.io.api.GroupConverter; +import org.apache.parquet.io.api.RecordMaterializer; +import org.apache.parquet.schema.GroupType; +import org.apache.parquet.schema.MessageTypeParser; import java.util.Map; http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/ETypeConverter.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/ETypeConverter.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/ETypeConverter.java index 3fc0129..17849fa 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/ETypeConverter.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/ETypeConverter.java @@ -33,11 +33,11 @@ import org.apache.hadoop.io.LongWritable; import org.apache.hadoop.io.Text; import org.apache.hadoop.io.Writable; -import parquet.column.Dictionary; -import parquet.io.api.Binary; -import parquet.io.api.PrimitiveConverter; -import parquet.schema.OriginalType; -import parquet.schema.PrimitiveType; +import org.apache.parquet.column.Dictionary; +import org.apache.parquet.io.api.Binary; +import org.apache.parquet.io.api.PrimitiveConverter; +import org.apache.parquet.schema.OriginalType; +import org.apache.parquet.schema.PrimitiveType; /** * http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveCollectionConverter.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveCollectionConverter.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveCollectionConverter.java index 80405bd..06f3d32 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveCollectionConverter.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveCollectionConverter.java @@ -24,9 +24,9 @@ import java.util.List; import org.apache.hadoop.io.ArrayWritable; import org.apache.hadoop.io.Writable; -import parquet.io.api.Converter; -import parquet.schema.GroupType; -import parquet.schema.Type; +import org.apache.parquet.io.api.Converter; +import org.apache.parquet.schema.GroupType; +import org.apache.parquet.schema.Type; public class HiveCollectionConverter extends HiveGroupConverter { private final GroupType collectionType; http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveGroupConverter.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveGroupConverter.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveGroupConverter.java index c6d03a1..b1ca85a 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveGroupConverter.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveGroupConverter.java @@ -14,13 +14,13 @@ package org.apache.hadoop.hive.ql.io.parquet.convert; import org.apache.hadoop.io.Writable; -import parquet.io.api.Converter; -import parquet.io.api.GroupConverter; -import parquet.io.api.PrimitiveConverter; -import parquet.schema.GroupType; -import parquet.schema.OriginalType; -import parquet.schema.PrimitiveType; -import parquet.schema.Type; +import org.apache.parquet.io.api.Converter; +import org.apache.parquet.io.api.GroupConverter; +import org.apache.parquet.io.api.PrimitiveConverter; +import org.apache.parquet.schema.GroupType; +import org.apache.parquet.schema.OriginalType; +import org.apache.parquet.schema.PrimitiveType; +import org.apache.parquet.schema.Type; import java.util.Map; http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveSchemaConverter.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveSchemaConverter.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveSchemaConverter.java index 43c772f..b01f21f 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveSchemaConverter.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveSchemaConverter.java @@ -25,15 +25,15 @@ import org.apache.hadoop.hive.serde2.typeinfo.StructTypeInfo; import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo; import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoFactory; -import parquet.schema.ConversionPatterns; -import parquet.schema.GroupType; -import parquet.schema.MessageType; -import parquet.schema.OriginalType; -import parquet.schema.PrimitiveType; -import parquet.schema.PrimitiveType.PrimitiveTypeName; -import parquet.schema.Type; -import parquet.schema.Type.Repetition; -import parquet.schema.Types; +import org.apache.parquet.schema.ConversionPatterns; +import org.apache.parquet.schema.GroupType; +import org.apache.parquet.schema.MessageType; +import org.apache.parquet.schema.OriginalType; +import org.apache.parquet.schema.PrimitiveType; +import org.apache.parquet.schema.PrimitiveType.PrimitiveTypeName; +import org.apache.parquet.schema.Type; +import org.apache.parquet.schema.Type.Repetition; +import org.apache.parquet.schema.Types; public class HiveSchemaConverter { http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveStructConverter.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveStructConverter.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveStructConverter.java index f95d15e..9c35a9f 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveStructConverter.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/HiveStructConverter.java @@ -19,9 +19,9 @@ import java.util.Map; import org.apache.hadoop.io.ArrayWritable; import org.apache.hadoop.io.Writable; -import parquet.io.api.Converter; -import parquet.schema.GroupType; -import parquet.schema.Type; +import org.apache.parquet.io.api.Converter; +import org.apache.parquet.schema.GroupType; +import org.apache.parquet.schema.Type; /** * http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/Repeated.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/Repeated.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/Repeated.java index 4becd20..c0af291 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/Repeated.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/convert/Repeated.java @@ -25,12 +25,12 @@ import java.util.Map; import org.apache.hadoop.io.ArrayWritable; import org.apache.hadoop.io.Writable; -import parquet.column.Dictionary; -import parquet.io.api.Binary; -import parquet.io.api.Converter; -import parquet.io.api.PrimitiveConverter; -import parquet.schema.GroupType; -import parquet.schema.PrimitiveType; +import org.apache.parquet.column.Dictionary; +import org.apache.parquet.io.api.Binary; +import org.apache.parquet.io.api.Converter; +import org.apache.parquet.io.api.PrimitiveConverter; +import org.apache.parquet.schema.GroupType; +import org.apache.parquet.schema.PrimitiveType; /** * Converters for repeated fields need to know when the parent field starts and http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/DataWritableReadSupport.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/DataWritableReadSupport.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/DataWritableReadSupport.java index dcd46bd..8b8b92d 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/DataWritableReadSupport.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/DataWritableReadSupport.java @@ -32,14 +32,14 @@ import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoUtils; import org.apache.hadoop.io.ArrayWritable; import org.apache.hadoop.util.StringUtils; -import parquet.hadoop.api.InitContext; -import parquet.hadoop.api.ReadSupport; -import parquet.io.api.RecordMaterializer; -import parquet.schema.GroupType; -import parquet.schema.MessageType; -import parquet.schema.Type; -import parquet.schema.Types; -import parquet.schema.PrimitiveType.PrimitiveTypeName; +import org.apache.parquet.hadoop.api.InitContext; +import org.apache.parquet.hadoop.api.ReadSupport; +import org.apache.parquet.io.api.RecordMaterializer; +import org.apache.parquet.schema.GroupType; +import org.apache.parquet.schema.MessageType; +import org.apache.parquet.schema.Type; +import org.apache.parquet.schema.Types; +import org.apache.parquet.schema.PrimitiveType.PrimitiveTypeName; /** * @@ -195,7 +195,7 @@ public class DataWritableReadSupport extends ReadSupport<ArrayWritable> { * @return the parquet ReadContext */ @Override - public parquet.hadoop.api.ReadSupport.ReadContext init(InitContext context) { + public org.apache.parquet.hadoop.api.ReadSupport.ReadContext init(InitContext context) { Configuration configuration = context.getConfiguration(); MessageType fileSchema = context.getFileSchema(); String columnNames = configuration.get(IOConstants.COLUMNS); @@ -247,7 +247,7 @@ public class DataWritableReadSupport extends ReadSupport<ArrayWritable> { @Override public RecordMaterializer<ArrayWritable> prepareForRead(final Configuration configuration, final Map<String, String> keyValueMetaData, final MessageType fileSchema, - final parquet.hadoop.api.ReadSupport.ReadContext readContext) { + final org.apache.parquet.hadoop.api.ReadSupport.ReadContext readContext) { final Map<String, String> metadata = readContext.getReadSupportMetadata(); if (metadata == null) { throw new IllegalStateException("ReadContext not initialized properly. " + http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/ParquetRecordReaderWrapper.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/ParquetRecordReaderWrapper.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/ParquetRecordReaderWrapper.java index 5c36564..80a7301 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/ParquetRecordReaderWrapper.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/read/ParquetRecordReaderWrapper.java @@ -15,12 +15,7 @@ package org.apache.hadoop.hive.ql.io.parquet.read; import java.io.IOException; import java.util.ArrayList; -import java.util.Collections; -import java.util.HashMap; -import java.util.HashSet; import java.util.List; -import java.util.Map; -import java.util.Set; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; @@ -44,19 +39,19 @@ import org.apache.hadoop.mapred.Reporter; import org.apache.hadoop.mapreduce.TaskAttemptContext; import org.apache.hadoop.mapreduce.TaskAttemptID; -import parquet.filter2.compat.FilterCompat; -import parquet.filter2.compat.RowGroupFilter; -import parquet.filter2.predicate.FilterPredicate; -import parquet.hadoop.ParquetFileReader; -import parquet.hadoop.ParquetInputFormat; -import parquet.hadoop.ParquetInputSplit; -import parquet.hadoop.api.InitContext; -import parquet.hadoop.api.ReadSupport.ReadContext; -import parquet.hadoop.metadata.BlockMetaData; -import parquet.hadoop.metadata.FileMetaData; -import parquet.hadoop.metadata.ParquetMetadata; -import parquet.hadoop.util.ContextUtil; -import parquet.schema.MessageTypeParser; +import org.apache.parquet.filter2.compat.FilterCompat; +import org.apache.parquet.filter2.compat.RowGroupFilter; +import org.apache.parquet.filter2.predicate.FilterPredicate; +import org.apache.parquet.hadoop.ParquetFileReader; +import org.apache.parquet.hadoop.ParquetInputFormat; +import org.apache.parquet.hadoop.ParquetInputSplit; +import org.apache.parquet.hadoop.api.InitContext; +import org.apache.parquet.hadoop.api.ReadSupport.ReadContext; +import org.apache.parquet.hadoop.metadata.BlockMetaData; +import org.apache.parquet.hadoop.metadata.FileMetaData; +import org.apache.parquet.hadoop.metadata.ParquetMetadata; +import org.apache.parquet.hadoop.util.ContextUtil; +import org.apache.parquet.schema.MessageTypeParser; import com.google.common.base.Strings; http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/serde/ParquetHiveSerDe.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/serde/ParquetHiveSerDe.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/serde/ParquetHiveSerDe.java index 7fd5e96..e1bf8e2 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/serde/ParquetHiveSerDe.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/serde/ParquetHiveSerDe.java @@ -36,8 +36,8 @@ import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoUtils; import org.apache.hadoop.io.ArrayWritable; import org.apache.hadoop.io.Text; import org.apache.hadoop.io.Writable; -import parquet.hadoop.ParquetOutputFormat; -import parquet.hadoop.ParquetWriter; +import org.apache.parquet.hadoop.ParquetOutputFormat; +import org.apache.parquet.hadoop.ParquetWriter; /** * http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/timestamp/NanoTime.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/timestamp/NanoTime.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/timestamp/NanoTime.java index fb2a6c2..b024cbb 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/timestamp/NanoTime.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/timestamp/NanoTime.java @@ -16,9 +16,9 @@ package org.apache.hadoop.hive.ql.io.parquet.timestamp; import java.nio.ByteBuffer; import java.nio.ByteOrder; -import parquet.Preconditions; -import parquet.io.api.Binary; -import parquet.io.api.RecordConsumer; +import org.apache.parquet.Preconditions; +import org.apache.parquet.io.api.Binary; +import org.apache.parquet.io.api.RecordConsumer; /** * Provides a wrapper representing a parquet-timestamp, with methods to * convert to and from binary. http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/DataWritableWriteSupport.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/DataWritableWriteSupport.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/DataWritableWriteSupport.java index 71653bb..f4621e5 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/DataWritableWriteSupport.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/DataWritableWriteSupport.java @@ -18,10 +18,10 @@ import java.util.HashMap; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hive.serde2.io.ParquetHiveRecord; -import parquet.hadoop.api.WriteSupport; -import parquet.io.api.RecordConsumer; -import parquet.schema.MessageType; -import parquet.schema.MessageTypeParser; +import org.apache.parquet.hadoop.api.WriteSupport; +import org.apache.parquet.io.api.RecordConsumer; +import org.apache.parquet.schema.MessageType; +import org.apache.parquet.schema.MessageTypeParser; /** * http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/DataWritableWriter.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/DataWritableWriter.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/DataWritableWriter.java index 5bcb270..c195c3e 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/DataWritableWriter.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/DataWritableWriter.java @@ -23,11 +23,11 @@ import org.apache.hadoop.hive.serde2.io.ParquetHiveRecord; import org.apache.hadoop.hive.serde2.objectinspector.*; import org.apache.hadoop.hive.serde2.objectinspector.primitive.*; import org.apache.hadoop.hive.serde2.typeinfo.DecimalTypeInfo; -import parquet.io.api.Binary; -import parquet.io.api.RecordConsumer; -import parquet.schema.GroupType; -import parquet.schema.OriginalType; -import parquet.schema.Type; +import org.apache.parquet.io.api.Binary; +import org.apache.parquet.io.api.RecordConsumer; +import org.apache.parquet.schema.GroupType; +import org.apache.parquet.schema.OriginalType; +import org.apache.parquet.schema.Type; import java.sql.Date; import java.sql.Timestamp; http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/ParquetRecordWriterWrapper.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/ParquetRecordWriterWrapper.java b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/ParquetRecordWriterWrapper.java index b7987a3..9e2a9e1 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/ParquetRecordWriterWrapper.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/parquet/write/ParquetRecordWriterWrapper.java @@ -32,9 +32,9 @@ import org.apache.hadoop.mapreduce.TaskAttemptID; import org.apache.hadoop.hive.serde2.io.ParquetHiveRecord; import org.apache.hadoop.util.Progressable; -import parquet.hadoop.ParquetOutputFormat; -import parquet.hadoop.metadata.CompressionCodecName; -import parquet.hadoop.util.ContextUtil; +import org.apache.parquet.hadoop.ParquetOutputFormat; +import org.apache.parquet.hadoop.metadata.CompressionCodecName; +import org.apache.parquet.hadoop.util.ContextUtil; public class ParquetRecordWriterWrapper implements RecordWriter<NullWritable, ParquetHiveRecord>, org.apache.hadoop.hive.ql.exec.FileSinkOperator.RecordWriter { http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/java/org/apache/hadoop/hive/ql/io/sarg/SearchArgumentImpl.java ---------------------------------------------------------------------- diff --git a/ql/src/java/org/apache/hadoop/hive/ql/io/sarg/SearchArgumentImpl.java b/ql/src/java/org/apache/hadoop/hive/ql/io/sarg/SearchArgumentImpl.java index 63b3ee9..782b5f8 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/io/sarg/SearchArgumentImpl.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/io/sarg/SearchArgumentImpl.java @@ -64,8 +64,8 @@ import com.esotericsoftware.kryo.Kryo; import com.esotericsoftware.kryo.io.Input; import com.esotericsoftware.kryo.io.Output; -import parquet.filter2.predicate.FilterApi; -import parquet.filter2.predicate.FilterPredicate; +import org.apache.parquet.filter2.predicate.FilterApi; +import org.apache.parquet.filter2.predicate.FilterPredicate; /** * The implementation of SearchArguments. http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/AbstractTestParquetDirect.java ---------------------------------------------------------------------- diff --git a/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/AbstractTestParquetDirect.java b/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/AbstractTestParquetDirect.java index 94a780d..c81499a 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/AbstractTestParquetDirect.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/AbstractTestParquetDirect.java @@ -42,10 +42,10 @@ import org.junit.Assert; import org.junit.BeforeClass; import org.junit.Rule; import org.junit.rules.TemporaryFolder; -import parquet.hadoop.ParquetWriter; -import parquet.hadoop.api.WriteSupport; -import parquet.io.api.RecordConsumer; -import parquet.schema.MessageType; +import org.apache.parquet.hadoop.ParquetWriter; +import org.apache.parquet.hadoop.api.WriteSupport; +import org.apache.parquet.io.api.RecordConsumer; +import org.apache.parquet.schema.MessageType; public abstract class AbstractTestParquetDirect { http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestArrayCompatibility.java ---------------------------------------------------------------------- diff --git a/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestArrayCompatibility.java b/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestArrayCompatibility.java index e0ea262..4b69642 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestArrayCompatibility.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestArrayCompatibility.java @@ -27,15 +27,15 @@ import org.apache.hadoop.io.IntWritable; import org.apache.hadoop.io.LongWritable; import org.junit.Assert; import org.junit.Test; -import parquet.io.api.RecordConsumer; -import parquet.schema.MessageType; -import parquet.schema.Types; - -import static parquet.schema.OriginalType.LIST; -import static parquet.schema.PrimitiveType.PrimitiveTypeName.DOUBLE; -import static parquet.schema.PrimitiveType.PrimitiveTypeName.FLOAT; -import static parquet.schema.PrimitiveType.PrimitiveTypeName.INT32; -import static parquet.schema.PrimitiveType.PrimitiveTypeName.INT64; +import org.apache.parquet.io.api.RecordConsumer; +import org.apache.parquet.schema.MessageType; +import org.apache.parquet.schema.Types; + +import static org.apache.parquet.schema.OriginalType.LIST; +import static org.apache.parquet.schema.PrimitiveType.PrimitiveTypeName.DOUBLE; +import static org.apache.parquet.schema.PrimitiveType.PrimitiveTypeName.FLOAT; +import static org.apache.parquet.schema.PrimitiveType.PrimitiveTypeName.INT32; +import static org.apache.parquet.schema.PrimitiveType.PrimitiveTypeName.INT64; public class TestArrayCompatibility extends AbstractTestParquetDirect { http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestDataWritableWriter.java ---------------------------------------------------------------------- diff --git a/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestDataWritableWriter.java b/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestDataWritableWriter.java index 73425e3..7049139 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestDataWritableWriter.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestDataWritableWriter.java @@ -39,10 +39,10 @@ import org.junit.Test; import org.mockito.InOrder; import org.mockito.Mock; import org.mockito.MockitoAnnotations; -import parquet.io.api.Binary; -import parquet.io.api.RecordConsumer; -import parquet.schema.MessageType; -import parquet.schema.MessageTypeParser; +import org.apache.parquet.io.api.Binary; +import org.apache.parquet.io.api.RecordConsumer; +import org.apache.parquet.schema.MessageType; +import org.apache.parquet.schema.MessageTypeParser; import java.io.UnsupportedEncodingException; import java.util.ArrayList; http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestHiveSchemaConverter.java ---------------------------------------------------------------------- diff --git a/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestHiveSchemaConverter.java b/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestHiveSchemaConverter.java index c1baec1..589b5b5 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestHiveSchemaConverter.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestHiveSchemaConverter.java @@ -25,12 +25,12 @@ import org.apache.hadoop.hive.serde2.typeinfo.TypeInfo; import org.apache.hadoop.hive.serde2.typeinfo.TypeInfoUtils; import org.junit.Test; -import parquet.schema.MessageType; -import parquet.schema.MessageTypeParser; -import parquet.schema.OriginalType; -import parquet.schema.Types; -import parquet.schema.PrimitiveType.PrimitiveTypeName; -import parquet.schema.Type.Repetition; +import org.apache.parquet.schema.MessageType; +import org.apache.parquet.schema.MessageTypeParser; +import org.apache.parquet.schema.OriginalType; +import org.apache.parquet.schema.Types; +import org.apache.parquet.schema.PrimitiveType.PrimitiveTypeName; +import org.apache.parquet.schema.Type.Repetition; public class TestHiveSchemaConverter { @@ -194,13 +194,13 @@ public class TestHiveSchemaConverter { final MessageType messageTypeFound = HiveSchemaConverter.convert(columnNames, columnTypes); // this messageType only has one optional field, whose name is mapCol, original Type is MAP assertEquals(1, messageTypeFound.getFieldCount()); - parquet.schema.Type topLevel = messageTypeFound.getFields().get(0); + org.apache.parquet.schema.Type topLevel = messageTypeFound.getFields().get(0); assertEquals("mapCol",topLevel.getName()); assertEquals(OriginalType.MAP, topLevel.getOriginalType()); assertEquals(Repetition.OPTIONAL, topLevel.getRepetition()); assertEquals(1, topLevel.asGroupType().getFieldCount()); - parquet.schema.Type secondLevel = topLevel.asGroupType().getFields().get(0); + org.apache.parquet.schema.Type secondLevel = topLevel.asGroupType().getFields().get(0); //there is one repeated field for mapCol, the field name is "map" and its original Type is MAP_KEY_VALUE; assertEquals("map", secondLevel.getName()); assertEquals(OriginalType.MAP_KEY_VALUE, secondLevel.getOriginalType()); http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestMapStructures.java ---------------------------------------------------------------------- diff --git a/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestMapStructures.java b/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestMapStructures.java index 32264a4..b2facaa 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestMapStructures.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestMapStructures.java @@ -27,12 +27,12 @@ import org.apache.hadoop.io.IntWritable; import org.apache.hadoop.io.Text; import org.junit.Assert; import org.junit.Test; -import parquet.io.api.Binary; -import parquet.io.api.RecordConsumer; -import parquet.schema.Types; +import org.apache.parquet.io.api.Binary; +import org.apache.parquet.io.api.RecordConsumer; +import org.apache.parquet.schema.Types; -import static parquet.schema.OriginalType.*; -import static parquet.schema.PrimitiveType.PrimitiveTypeName.*; +import static org.apache.parquet.schema.OriginalType.*; +import static org.apache.parquet.schema.PrimitiveType.PrimitiveTypeName.*; public class TestMapStructures extends AbstractTestParquetDirect { http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestMapredParquetInputFormat.java ---------------------------------------------------------------------- diff --git a/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestMapredParquetInputFormat.java b/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestMapredParquetInputFormat.java index 1a54bf5..518eefd 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestMapredParquetInputFormat.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestMapredParquetInputFormat.java @@ -18,7 +18,7 @@ import static org.mockito.Mockito.mock; import org.apache.hadoop.io.ArrayWritable; import org.junit.Test; -import parquet.hadoop.ParquetInputFormat; +import org.apache.parquet.hadoop.ParquetInputFormat; public class TestMapredParquetInputFormat { @Test http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestMapredParquetOutputFormat.java ---------------------------------------------------------------------- diff --git a/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestMapredParquetOutputFormat.java b/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestMapredParquetOutputFormat.java index e93aa9a..ec85b5d 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestMapredParquetOutputFormat.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestMapredParquetOutputFormat.java @@ -29,7 +29,7 @@ import org.apache.hadoop.mapred.JobConf; import org.apache.hadoop.util.Progressable; import org.junit.Test; -import parquet.hadoop.ParquetOutputFormat; +import org.apache.parquet.hadoop.ParquetOutputFormat; public class TestMapredParquetOutputFormat { http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestParquetRowGroupFilter.java ---------------------------------------------------------------------- diff --git a/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestParquetRowGroupFilter.java b/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestParquetRowGroupFilter.java index 4ccb207..7204521 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestParquetRowGroupFilter.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/io/parquet/TestParquetRowGroupFilter.java @@ -37,9 +37,9 @@ import org.apache.hadoop.mapred.JobConf; import org.junit.Assert; import org.junit.Before; import org.junit.Test; -import parquet.io.api.RecordConsumer; -import parquet.schema.MessageType; -import parquet.schema.MessageTypeParser; +import org.apache.parquet.io.api.RecordConsumer; +import org.apache.parquet.schema.MessageType; +import org.apache.parquet.schema.MessageTypeParser; import java.util.ArrayList; import java.util.Arrays; http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/ql/src/test/org/apache/hadoop/hive/ql/io/sarg/TestSearchArgumentImpl.java ---------------------------------------------------------------------- diff --git a/ql/src/test/org/apache/hadoop/hive/ql/io/sarg/TestSearchArgumentImpl.java b/ql/src/test/org/apache/hadoop/hive/ql/io/sarg/TestSearchArgumentImpl.java index 51f7f47..9e6adef 100644 --- a/ql/src/test/org/apache/hadoop/hive/ql/io/sarg/TestSearchArgumentImpl.java +++ b/ql/src/test/org/apache/hadoop/hive/ql/io/sarg/TestSearchArgumentImpl.java @@ -44,7 +44,7 @@ import java.sql.Timestamp; import java.util.List; import java.util.Set; -import parquet.filter2.predicate.FilterPredicate; +import org.apache.parquet.filter2.predicate.FilterPredicate; /** * These test the SARG implementation. http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/serde/pom.xml ---------------------------------------------------------------------- diff --git a/serde/pom.xml b/serde/pom.xml index 7a9a12c..b6c0d0c 100644 --- a/serde/pom.xml +++ b/serde/pom.xml @@ -81,7 +81,7 @@ <version>${opencsv.version}</version> </dependency> <dependency> - <groupId>com.twitter</groupId> + <groupId>org.apache.parquet</groupId> <artifactId>parquet-hadoop-bundle</artifactId> <version>${parquet.version}</version> </dependency> http://git-wip-us.apache.org/repos/asf/hive/blob/11020ae5/serde/src/java/org/apache/hadoop/hive/ql/io/sarg/SearchArgument.java ---------------------------------------------------------------------- diff --git a/serde/src/java/org/apache/hadoop/hive/ql/io/sarg/SearchArgument.java b/serde/src/java/org/apache/hadoop/hive/ql/io/sarg/SearchArgument.java index 9be54da..df208d4 100644 --- a/serde/src/java/org/apache/hadoop/hive/ql/io/sarg/SearchArgument.java +++ b/serde/src/java/org/apache/hadoop/hive/ql/io/sarg/SearchArgument.java @@ -18,7 +18,7 @@ package org.apache.hadoop.hive.ql.io.sarg; -import parquet.filter2.predicate.FilterPredicate; +import org.apache.parquet.filter2.predicate.FilterPredicate; import java.util.List;
