[ 
https://issues.apache.org/jira/browse/HAWQ-1637?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16533364#comment-16533364
 ] 

Ruilong Huo commented on HAWQ-1637:
-----------------------------------

Check the JDK you use and make sure it is no higher than 1.8

> Compile apache hawq failure due to Failed to execute goal 
> org.apache.maven.plugins:maven-javadoc-plugin:2.9.1:aggregate-jar on osx 10.11
> ----------------------------------------------------------------------------------------------------------------------------------------
>
>                 Key: HAWQ-1637
>                 URL: https://issues.apache.org/jira/browse/HAWQ-1637
>             Project: Apache HAWQ
>          Issue Type: Bug
>          Components: Build
>            Reporter: Oushu_WangZiming
>            Assignee: Radar Lei
>            Priority: Major
>
> Follow instruction ([https://cwiki.apache.org/confluence/disp
>  /usr/local/bin/mvn package 
> -DskipTestslay/HAWQ/Build+and+Install)|https://cwiki.apache.org/confluence/display/HAWQ/Build+and+Install)]
>  to build apache hawq on osx 10.11, it fails due to Failed to execute goal 
> org.apache.maven.plugins:maven-javadoc-plugin:2.9.1:aggregate-jar
>  
> {code:java}
> /usr/local/bin/mvn package -DskipTests
> [INFO] Scanning for projects...
> [INFO] 
> ------------------------------------------------------------------------
> [INFO] Reactor Build Order:
> [INFO]
> [INFO] hawq-hadoop [pom]
> [INFO] hawq-mapreduce-common [jar]
> [INFO] hawq-mapreduce-ao [jar]
> [INFO] hawq-mapreduce-parquet [jar]
> [INFO] hawq-mapreduce-tool [jar]
> [INFO]
> [INFO] --------------------< com.pivotal.hawq:hawq-hadoop 
> >--------------------
> [INFO] Building hawq-hadoop 1.1.0 [1/5]
> [INFO] --------------------------------[ pom 
> ]---------------------------------
> [INFO]
> [INFO] --- maven-jar-plugin:2.4:test-jar (default) @ hawq-hadoop ---
> [WARNING] JAR will be empty - no content was marked for inclusion!
> [WARNING] The following dependencies could not be resolved at this point of 
> the build but seem to be part of the reactor:
> [WARNING] o com.pivotal.hawq:hawq-mapreduce-common:jar:1.1.0 (compile)
> [WARNING] Try running the build up to the lifecycle phase "package"
> [WARNING] The following dependencies could not be resolved at this point of 
> the build but seem to be part of the reactor:
> [WARNING] o com.pivotal.hawq:hawq-mapreduce-common:jar:1.1.0 (compile)
> [WARNING] Try running the build up to the lifecycle phase "package"
> [WARNING] The following dependencies could not be resolved at this point of 
> the build but seem to be part of the reactor:
> [WARNING] o com.pivotal.hawq:hawq-mapreduce-ao:jar:1.1.0 (compile)
> [WARNING] o com.pivotal.hawq:hawq-mapreduce-common:jar:1.1.0 (compile)
> [WARNING] o com.pivotal.hawq:hawq-mapreduce-parquet:jar:1.1.0 (compile)
> [WARNING] Try running the build up to the lifecycle phase "package"
> [INFO]
> [INFO] --- maven-javadoc-plugin:2.9.1:aggregate-jar (default) @ hawq-hadoop 
> ---
> [WARNING] The dependency: [com.pivotal.hawq:hawq-mapreduce-common:jar:1.1.0] 
> can't be resolved but has been found in the reactor (probably snapshots).
> This dependency has been excluded from the Javadoc classpath. You should 
> rerun javadoc after executing mvn install.
> [WARNING] IGNORED to add some artifacts in the classpath. See above.
> [WARNING] The dependency: [com.pivotal.hawq:hawq-mapreduce-common:jar:1.1.0] 
> can't be resolved but has been found in the reactor (probably snapshots).
> This dependency has been excluded from the Javadoc classpath. You should 
> rerun javadoc after executing mvn install.
> [WARNING] IGNORED to add some artifacts in the classpath. See above.
> [WARNING] The dependency: [com.pivotal.hawq:hawq-mapreduce-common:jar:1.1.0] 
> can't be resolved but has been found in the reactor (probably snapshots).
> This dependency has been excluded from the Javadoc classpath. You should 
> rerun javadoc after executing mvn install.
> [WARNING] The dependency: [com.pivotal.hawq:hawq-mapreduce-ao:jar:1.1.0] 
> can't be resolved but has been found in the reactor (probably snapshots).
> This dependency has been excluded from the Javadoc classpath. You should 
> rerun javadoc after executing mvn install.
> [WARNING] The dependency: [com.pivotal.hawq:hawq-mapreduce-parquet:jar:1.1.0] 
> can't be resolved but has been found in the reactor (probably snapshots).
> This dependency has been excluded from the Javadoc classpath. You should 
> rerun javadoc after executing mvn install.
> [WARNING] IGNORED to add some artifacts in the classpath. See above.
> [INFO]
> 正在加载程序包com.pivotal.hawq.mapreduce.conf的源文件...
> 正在加载程序包com.pivotal.hawq.mapreduce.datatype的源文件...
> 正在加载程序包com.pivotal.hawq.mapreduce.file的源文件...
> 正在加载程序包com.pivotal.hawq.mapreduce的源文件...
> 正在加载程序包com.pivotal.hawq.mapreduce.metadata的源文件...
> 正在加载程序包com.pivotal.hawq.mapreduce.schema的源文件...
> 正在加载程序包com.pivotal.hawq.mapreduce.util的源文件...
> 正在加载程序包com.pivotal.hawq.mapreduce.ao.file的源文件...
> 正在加载程序包com.pivotal.hawq.mapreduce.ao的源文件...
> 正在加载程序包com.pivotal.hawq.mapreduce.ao.io的源文件...
> 正在加载程序包com.pivotal.hawq.mapreduce.ao.util的源文件...
> 正在加载程序包com.pivotal.hawq.mapreduce.parquet.convert的源文件...
> 正在加载程序包com.pivotal.hawq.mapreduce.parquet的源文件...
> 正在加载程序包com.pivotal.hawq.mapreduce.parquet.support的源文件...
> 正在加载程序包com.pivotal.hawq.mapreduce.parquet.util的源文件...
> 正在构造 Javadoc 信息...
> 100 个错误
> [INFO] 
> ------------------------------------------------------------------------
> [INFO] Reactor Summary:
> [INFO]
> [INFO] hawq-hadoop 1.1.0 .................................. FAILURE [ 7.035 s]
> [INFO] hawq-mapreduce-common .............................. SKIPPED
> [INFO] hawq-mapreduce-ao .................................. SKIPPED
> [INFO] hawq-mapreduce-parquet ............................. SKIPPED
> [INFO] hawq-mapreduce-tool 1.1.0 .......................... SKIPPED
> [INFO] 
> ------------------------------------------------------------------------
> [INFO] BUILD FAILURE
> [INFO] 
> ------------------------------------------------------------------------
> [INFO] Total time: 7.235 s
> [INFO] Finished at: 2018-07-05T13:50:16+08:00
> [INFO] 
> ------------------------------------------------------------------------
> [ERROR] Failed to execute goal 
> org.apache.maven.plugins:maven-javadoc-plugin:2.9.1:aggregate-jar (default) 
> on project hawq-hadoop: MavenReportException: Error while creating archive:
> [ERROR] Exit code: 1 - 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/HAWQParquetInputFormat.java:34:
>  错误: 程序包parquet.hadoop不存在
> [ERROR] import parquet.hadoop.ParquetInputFormat;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/HAWQParquetInputFormat.java:42:
>  错误: 找不到符号
> [ERROR] public class HAWQParquetInputFormat extends 
> ParquetInputFormat<HAWQRecord> {
> [ERROR] ^
> [ERROR] 符号: 类 ParquetInputFormat
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQBoxConverter.java:25:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.Converter;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQBoxConverter.java:26:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.GroupConverter;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQBoxConverter.java:27:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.PrimitiveConverter;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQBoxConverter.java:38:
>  错误: 找不到符号
> [ERROR] public class HAWQBoxConverter extends GroupConverter {
> [ERROR] ^
> [ERROR] 符号: 类 GroupConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQBoxConverter.java:41:
>  错误: 找不到符号
> [ERROR] private Converter[] converters;
> [ERROR] ^
> [ERROR] 符号: 类 Converter
> [ERROR] 位置: 类 HAWQBoxConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQBoxConverter.java:78:
>  错误: 找不到符号
> [ERROR] public Converter getConverter(int fieldIndex) {
> [ERROR] ^
> [ERROR] 符号: 类 Converter
> [ERROR] 位置: 类 HAWQBoxConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQCircleConverter.java:25:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.Converter;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQCircleConverter.java:26:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.GroupConverter;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQCircleConverter.java:27:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.PrimitiveConverter;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQCircleConverter.java:37:
>  错误: 找不到符号
> [ERROR] public class HAWQCircleConverter extends GroupConverter {
> [ERROR] ^
> [ERROR] 符号: 类 GroupConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQCircleConverter.java:40:
>  错误: 找不到符号
> [ERROR] private Converter[] converters;
> [ERROR] ^
> [ERROR] 符号: 类 Converter
> [ERROR] 位置: 类 HAWQCircleConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQCircleConverter.java:70:
>  错误: 找不到符号
> [ERROR] public Converter getConverter(int fieldIndex) {
> [ERROR] ^
> [ERROR] 符号: 类 Converter
> [ERROR] 位置: 类 HAWQCircleConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQLineSegmentConverter.java:25:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.Converter;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQLineSegmentConverter.java:26:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.GroupConverter;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQLineSegmentConverter.java:27:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.PrimitiveConverter;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQLineSegmentConverter.java:38:
>  错误: 找不到符号
> [ERROR] public class HAWQLineSegmentConverter extends GroupConverter {
> [ERROR] ^
> [ERROR] 符号: 类 GroupConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQLineSegmentConverter.java:41:
>  错误: 找不到符号
> [ERROR] private Converter[] converters;
> [ERROR] ^
> [ERROR] 符号: 类 Converter
> [ERROR] 位置: 类 HAWQLineSegmentConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQLineSegmentConverter.java:78:
>  错误: 找不到符号
> [ERROR] public Converter getConverter(int fieldIndex) {
> [ERROR] ^
> [ERROR] 符号: 类 Converter
> [ERROR] 位置: 类 HAWQLineSegmentConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPathConverter.java:26:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.Converter;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPathConverter.java:27:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.GroupConverter;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPathConverter.java:42:
>  错误: 找不到符号
> [ERROR] public class HAWQPathConverter extends GroupConverter {
> [ERROR] ^
> [ERROR] 符号: 类 GroupConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPathConverter.java:45:
>  错误: 找不到符号
> [ERROR] private Converter[] converters;
> [ERROR] ^
> [ERROR] 符号: 类 Converter
> [ERROR] 位置: 类 HAWQPathConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPathConverter.java:70:
>  错误: 找不到符号
> [ERROR] public Converter getConverter(int fieldIndex) {
> [ERROR] ^
> [ERROR] 符号: 类 Converter
> [ERROR] 位置: 类 HAWQPathConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPointConverter.java:25:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.Converter;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPointConverter.java:26:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.GroupConverter;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPointConverter.java:27:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.PrimitiveConverter;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPointConverter.java:36:
>  错误: 找不到符号
> [ERROR] public class HAWQPointConverter extends GroupConverter {
> [ERROR] ^
> [ERROR] 符号: 类 GroupConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPointConverter.java:39:
>  错误: 找不到符号
> [ERROR] private Converter[] converters;
> [ERROR] ^
> [ERROR] 符号: 类 Converter
> [ERROR] 位置: 类 HAWQPointConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPointConverter.java:62:
>  错误: 找不到符号
> [ERROR] public Converter getConverter(int fieldIndex) {
> [ERROR] ^
> [ERROR] 符号: 类 Converter
> [ERROR] 位置: 类 HAWQPointConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPolygonConverter.java:27:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.Converter;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPolygonConverter.java:28:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.GroupConverter;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPolygonConverter.java:48:
>  错误: 找不到符号
> [ERROR] public class HAWQPolygonConverter extends GroupConverter {
> [ERROR] ^
> [ERROR] 符号: 类 GroupConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPolygonConverter.java:51:
>  错误: 找不到符号
> [ERROR] private Converter[] converters;
> [ERROR] ^
> [ERROR] 符号: 类 Converter
> [ERROR] 位置: 类 HAWQPolygonConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPolygonConverter.java:77:
>  错误: 找不到符号
> [ERROR] public Converter getConverter(int fieldIndex) {
> [ERROR] ^
> [ERROR] 符号: 类 Converter
> [ERROR] 位置: 类 HAWQPolygonConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:29:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.Binary;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:30:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.Converter;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:31:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.GroupConverter;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:32:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.PrimitiveConverter;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:33:
>  错误: 程序包parquet.schema不存在
> [ERROR] import parquet.schema.MessageType;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:44:
>  错误: 找不到符号
> [ERROR] public class HAWQRecordConverter extends GroupConverter {
> [ERROR] ^
> [ERROR] 符号: 类 GroupConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:48:
>  错误: 找不到符号
> [ERROR] private final Converter[] converters;
> [ERROR] ^
> [ERROR] 符号: 类 Converter
> [ERROR] 位置: 类 HAWQRecordConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:52:
>  错误: 找不到符号
> [ERROR] public HAWQRecordConverter(MessageType requestedSchema, HAWQSchema 
> hawqSchema) {
> [ERROR] ^
> [ERROR] 符号: 类 MessageType
> [ERROR] 位置: 类 HAWQRecordConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:56:
>  错误: 找不到符号
> [ERROR] public HAWQRecordConverter(ParentValueContainer parent, MessageType 
> requestedSchema, HAWQSchema hawqSchema) {
> [ERROR] ^
> [ERROR] 符号: 类 MessageType
> [ERROR] 位置: 类 HAWQRecordConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:201:
>  错误: 找不到符号
> [ERROR] private Converter newConverter(HAWQField hawqType, 
> ParentValueContainer parent) {
> [ERROR] ^
> [ERROR] 符号: 类 Converter
> [ERROR] 位置: 类 HAWQRecordConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:259:
>  错误: 找不到符号
> [ERROR] public Converter getConverter(int fieldIndex) {
> [ERROR] ^
> [ERROR] 符号: 类 Converter
> [ERROR] 位置: 类 HAWQRecordConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:285:
>  错误: 找不到符号
> [ERROR] static class HAWQPrimitiveConverter extends PrimitiveConverter {
> [ERROR] ^
> [ERROR] 符号: 类 PrimitiveConverter
> [ERROR] 位置: 类 HAWQRecordConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:328:
>  错误: 找不到符号
> [ERROR] static class HAWQShortConverter extends PrimitiveConverter {
> [ERROR] ^
> [ERROR] 符号: 类 PrimitiveConverter
> [ERROR] 位置: 类 HAWQRecordConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:343:
>  错误: 找不到符号
> [ERROR] static class HAWQBigDecimalConverter extends PrimitiveConverter {
> [ERROR] ^
> [ERROR] 符号: 类 PrimitiveConverter
> [ERROR] 位置: 类 HAWQRecordConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:351:
>  错误: 找不到符号
> [ERROR] public void addBinary(Binary value) {
> [ERROR] ^
> [ERROR] 符号: 类 Binary
> [ERROR] 位置: 类 HAWQBigDecimalConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:361:
>  错误: 找不到符号
> [ERROR] static class HAWQStringConverter extends PrimitiveConverter {
> [ERROR] ^
> [ERROR] 符号: 类 PrimitiveConverter
> [ERROR] 位置: 类 HAWQRecordConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:369:
>  错误: 找不到符号
> [ERROR] public void addBinary(Binary value) {
> [ERROR] ^
> [ERROR] 符号: 类 Binary
> [ERROR] 位置: 类 HAWQStringConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:378:
>  错误: 找不到符号
> [ERROR] static class HAWQBitsConverter extends PrimitiveConverter {
> [ERROR] ^
> [ERROR] 符号: 类 PrimitiveConverter
> [ERROR] 位置: 类 HAWQRecordConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:386:
>  错误: 找不到符号
> [ERROR] public void addBinary(Binary value) {
> [ERROR] ^
> [ERROR] 符号: 类 Binary
> [ERROR] 位置: 类 HAWQBitsConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:395:
>  错误: 找不到符号
> [ERROR] static class HAWQByteArrayConverter extends PrimitiveConverter {
> [ERROR] ^
> [ERROR] 符号: 类 PrimitiveConverter
> [ERROR] 位置: 类 HAWQRecordConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:403:
>  错误: 找不到符号
> [ERROR] public void addBinary(Binary value) {
> [ERROR] ^
> [ERROR] 符号: 类 Binary
> [ERROR] 位置: 类 HAWQByteArrayConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:411:
>  错误: 找不到符号
> [ERROR] static class HAWQDateConverter extends PrimitiveConverter {
> [ERROR] ^
> [ERROR] 符号: 类 PrimitiveConverter
> [ERROR] 位置: 类 HAWQRecordConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:430:
>  错误: 找不到符号
> [ERROR] static class HAWQTimeConverter extends PrimitiveConverter {
> [ERROR] ^
> [ERROR] 符号: 类 PrimitiveConverter
> [ERROR] 位置: 类 HAWQRecordConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:448:
>  错误: 找不到符号
> [ERROR] static class HAWQTimeTZConverter extends PrimitiveConverter {
> [ERROR] ^
> [ERROR] 符号: 类 PrimitiveConverter
> [ERROR] 位置: 类 HAWQRecordConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:456:
>  错误: 找不到符号
> [ERROR] public void addBinary(Binary value) {
> [ERROR] ^
> [ERROR] 符号: 类 Binary
> [ERROR] 位置: 类 HAWQTimeTZConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:466:
>  错误: 找不到符号
> [ERROR] static class HAWQTimestampConverter extends PrimitiveConverter {
> [ERROR] ^
> [ERROR] 符号: 类 PrimitiveConverter
> [ERROR] 位置: 类 HAWQRecordConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:484:
>  错误: 找不到符号
> [ERROR] static class HAWQTimestampTZConverter extends PrimitiveConverter {
> [ERROR] ^
> [ERROR] 符号: 类 PrimitiveConverter
> [ERROR] 位置: 类 HAWQRecordConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:502:
>  错误: 找不到符号
> [ERROR] static class HAWQIntervalConverter extends PrimitiveConverter {
> [ERROR] ^
> [ERROR] 符号: 类 PrimitiveConverter
> [ERROR] 位置: 类 HAWQRecordConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:510:
>  错误: 找不到符号
> [ERROR] public void addBinary(Binary value) {
> [ERROR] ^
> [ERROR] 符号: 类 Binary
> [ERROR] 位置: 类 HAWQIntervalConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:520:
>  错误: 找不到符号
> [ERROR] static class HAWQMacaddrConverter extends PrimitiveConverter {
> [ERROR] ^
> [ERROR] 符号: 类 PrimitiveConverter
> [ERROR] 位置: 类 HAWQRecordConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:528:
>  错误: 找不到符号
> [ERROR] public void addBinary(Binary value) {
> [ERROR] ^
> [ERROR] 符号: 类 Binary
> [ERROR] 位置: 类 HAWQMacaddrConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:538:
>  错误: 找不到符号
> [ERROR] static class HAWQInetConverter extends PrimitiveConverter {
> [ERROR] ^
> [ERROR] 符号: 类 PrimitiveConverter
> [ERROR] 位置: 类 HAWQRecordConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:546:
>  错误: 找不到符号
> [ERROR] public void addBinary(Binary value) {
> [ERROR] ^
> [ERROR] 符号: 类 Binary
> [ERROR] 位置: 类 HAWQInetConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:556:
>  错误: 找不到符号
> [ERROR] static class HAWQCidrConverter extends PrimitiveConverter {
> [ERROR] ^
> [ERROR] 符号: 类 PrimitiveConverter
> [ERROR] 位置: 类 HAWQRecordConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:564:
>  错误: 找不到符号
> [ERROR] public void addBinary(Binary value) {
> [ERROR] ^
> [ERROR] 符号: 类 Binary
> [ERROR] 位置: 类 HAWQCidrConverter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordMaterializer.java:25:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.GroupConverter;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordMaterializer.java:26:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.RecordMaterializer;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordMaterializer.java:27:
>  错误: 程序包parquet.schema不存在
> [ERROR] import parquet.schema.MessageType;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordMaterializer.java:36:
>  错误: 找不到符号
> [ERROR] public class HAWQRecordMaterializer extends 
> RecordMaterializer<HAWQRecord> {
> [ERROR] ^
> [ERROR] 符号: 类 RecordMaterializer
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordMaterializer.java:40:
>  错误: 找不到符号
> [ERROR] public HAWQRecordMaterializer(MessageType requestedSchema, HAWQSchema 
> hawqSchema) {
> [ERROR] ^
> [ERROR] 符号: 类 MessageType
> [ERROR] 位置: 类 HAWQRecordMaterializer
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordMaterializer.java:50:
>  错误: 找不到符号
> [ERROR] public GroupConverter getRootConverter() {
> [ERROR] ^
> [ERROR] 符号: 类 GroupConverter
> [ERROR] 位置: 类 HAWQRecordMaterializer
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordWriter.java:29:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.Binary;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordWriter.java:30:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.RecordConsumer;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordWriter.java:39:
>  错误: 找不到符号
> [ERROR] private RecordConsumer consumer;
> [ERROR] ^
> [ERROR] 符号: 类 RecordConsumer
> [ERROR] 位置: 类 HAWQRecordWriter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordWriter.java:42:
>  错误: 找不到符号
> [ERROR] public HAWQRecordWriter(RecordConsumer consumer, HAWQSchema schema) {
> [ERROR] ^
> [ERROR] 符号: 类 RecordConsumer
> [ERROR] 位置: 类 HAWQRecordWriter
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/HAWQParquetOutputFormat.java:27:
>  错误: 程序包parquet.hadoop不存在
> [ERROR] import parquet.hadoop.ParquetOutputFormat;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/HAWQParquetOutputFormat.java:28:
>  错误: 程序包parquet.hadoop.util不存在
> [ERROR] import parquet.hadoop.util.ContextUtil;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/HAWQParquetOutputFormat.java:30:
>  错误: 找不到符号
> [ERROR] class HAWQParquetOutputFormat extends ParquetOutputFormat<HAWQRecord> 
> {
> [ERROR] ^
> [ERROR] 符号: 类 ParquetOutputFormat
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQReadSupport.java:27:
>  错误: 程序包parquet.hadoop.api不存在
> [ERROR] import parquet.hadoop.api.ReadSupport;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQReadSupport.java:28:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.RecordMaterializer;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQReadSupport.java:29:
>  错误: 程序包parquet.schema不存在
> [ERROR] import parquet.schema.MessageType;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQReadSupport.java:37:
>  错误: 找不到符号
> [ERROR] public class HAWQReadSupport extends ReadSupport<HAWQRecord> {
> [ERROR] ^
> [ERROR] 符号: 类 ReadSupport
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQReadSupport.java:45:
>  错误: 找不到符号
> [ERROR] MessageType fileSchema) {
> [ERROR] ^
> [ERROR] 符号: 类 MessageType
> [ERROR] 位置: 类 HAWQReadSupport
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQReadSupport.java:43:
>  错误: 找不到符号
> [ERROR] public ReadContext init(Configuration configuration,
> [ERROR] ^
> [ERROR] 符号: 类 ReadContext
> [ERROR] 位置: 类 HAWQReadSupport
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQReadSupport.java:63:
>  错误: 找不到符号
> [ERROR] MessageType fileSchema, ReadContext readContext) {
> [ERROR] ^
> [ERROR] 符号: 类 MessageType
> [ERROR] 位置: 类 HAWQReadSupport
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQReadSupport.java:63:
>  错误: 找不到符号
> [ERROR] MessageType fileSchema, ReadContext readContext) {
> [ERROR] ^
> [ERROR] 符号: 类 ReadContext
> [ERROR] 位置: 类 HAWQReadSupport
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQReadSupport.java:61:
>  错误: 找不到符号
> [ERROR] public RecordMaterializer<HAWQRecord> prepareForRead(Configuration 
> configuration,
> [ERROR] ^
> [ERROR] 符号: 类 RecordMaterializer
> [ERROR] 位置: 类 HAWQReadSupport
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQWriteSupport.java:29:
>  错误: 程序包parquet.hadoop.api不存在
> [ERROR] import parquet.hadoop.api.WriteSupport;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQWriteSupport.java:30:
>  错误: 程序包parquet.io.api不存在
> [ERROR] import parquet.io.api.RecordConsumer;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQWriteSupport.java:31:
>  错误: 程序包parquet.schema不存在
> [ERROR] import parquet.schema.MessageType;
> [ERROR] ^
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQWriteSupport.java:40:
>  错误: 找不到符号
> [ERROR] public class HAWQWriteSupport extends WriteSupport<HAWQRecord> {
> [ERROR] ^
> [ERROR] 符号: 类 WriteSupport
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQWriteSupport.java:46:
>  错误: 找不到符号
> [ERROR] private MessageType parquetSchema;
> [ERROR] ^
> [ERROR] 符号: 类 MessageType
> [ERROR] 位置: 类 HAWQWriteSupport
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQWriteSupport.java:54:
>  错误: 找不到符号
> [ERROR] public WriteContext init(Configuration configuration) {
> [ERROR] ^
> [ERROR] 符号: 类 WriteContext
> [ERROR] 位置: 类 HAWQWriteSupport
> [ERROR] 
> /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQWriteSupport.java:65:
>  错误: 找不到符号
> [ERROR] public void prepareForWrite(RecordConsumer recordConsumer) {
> [ERROR] ^
> [ERROR] 符号: 类 RecordConsumer
> [ERROR] 位置: 类 HAWQWriteSupport
> [ERROR]
> [ERROR] Command line was: 
> /Library/Java/JavaVirtualMachines/jdk-10.0.1.jdk/Contents/Home/bin/javadoc 
> @options @packages
> [ERROR]
> [ERROR] Refer to the generated Javadoc files in 
> '/Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/target/apidocs'
>  dir.
> [ERROR]
> [ERROR] -> [Help 1]
> [ERROR]
> [ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
> switch.
> [ERROR] Re-run Maven using the -X switch to enable full debug logging.
> [ERROR]
> [ERROR] For more information about the errors and possible solutions, please 
> read the following articles:
> [ERROR] [Help 1] 
> http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
> make[2]: *** [hawq-mapreduce-tool/target/hawq-mapreduce-tool-1.1.0.jar] Error 
> 1
> make[1]: *** [all] Error 2
> make: *** [all] Error 2{code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to