[ https://issues.apache.org/jira/browse/HAWQ-1637?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16533364#comment-16533364 ]
Ruilong Huo edited comment on HAWQ-1637 at 7/5/18 7:52 AM: ----------------------------------------------------------- Check the JDK you use and make sure it is <= 1.8 was (Author: huor): Check the JDK you use and make sure it is no higher than 1.8 > Compile apache hawq failure due to Failed to execute goal > org.apache.maven.plugins:maven-javadoc-plugin:2.9.1:aggregate-jar on osx 10.11 > ---------------------------------------------------------------------------------------------------------------------------------------- > > Key: HAWQ-1637 > URL: https://issues.apache.org/jira/browse/HAWQ-1637 > Project: Apache HAWQ > Issue Type: Bug > Components: Build > Reporter: Oushu_WangZiming > Assignee: Radar Lei > Priority: Major > > Follow instruction ([https://cwiki.apache.org/confluence/disp > /usr/local/bin/mvn package > -DskipTestslay/HAWQ/Build+and+Install)|https://cwiki.apache.org/confluence/display/HAWQ/Build+and+Install)] > to build apache hawq on osx 10.11, it fails due to Failed to execute goal > org.apache.maven.plugins:maven-javadoc-plugin:2.9.1:aggregate-jar > > {code:java} > /usr/local/bin/mvn package -DskipTests > [INFO] Scanning for projects... > [INFO] > ------------------------------------------------------------------------ > [INFO] Reactor Build Order: > [INFO] > [INFO] hawq-hadoop [pom] > [INFO] hawq-mapreduce-common [jar] > [INFO] hawq-mapreduce-ao [jar] > [INFO] hawq-mapreduce-parquet [jar] > [INFO] hawq-mapreduce-tool [jar] > [INFO] > [INFO] --------------------< com.pivotal.hawq:hawq-hadoop > >-------------------- > [INFO] Building hawq-hadoop 1.1.0 [1/5] > [INFO] --------------------------------[ pom > ]--------------------------------- > [INFO] > [INFO] --- maven-jar-plugin:2.4:test-jar (default) @ hawq-hadoop --- > [WARNING] JAR will be empty - no content was marked for inclusion! > [WARNING] The following dependencies could not be resolved at this point of > the build but seem to be part of the reactor: > [WARNING] o com.pivotal.hawq:hawq-mapreduce-common:jar:1.1.0 (compile) > [WARNING] Try running the build up to the lifecycle phase "package" > [WARNING] The following dependencies could not be resolved at this point of > the build but seem to be part of the reactor: > [WARNING] o com.pivotal.hawq:hawq-mapreduce-common:jar:1.1.0 (compile) > [WARNING] Try running the build up to the lifecycle phase "package" > [WARNING] The following dependencies could not be resolved at this point of > the build but seem to be part of the reactor: > [WARNING] o com.pivotal.hawq:hawq-mapreduce-ao:jar:1.1.0 (compile) > [WARNING] o com.pivotal.hawq:hawq-mapreduce-common:jar:1.1.0 (compile) > [WARNING] o com.pivotal.hawq:hawq-mapreduce-parquet:jar:1.1.0 (compile) > [WARNING] Try running the build up to the lifecycle phase "package" > [INFO] > [INFO] --- maven-javadoc-plugin:2.9.1:aggregate-jar (default) @ hawq-hadoop > --- > [WARNING] The dependency: [com.pivotal.hawq:hawq-mapreduce-common:jar:1.1.0] > can't be resolved but has been found in the reactor (probably snapshots). > This dependency has been excluded from the Javadoc classpath. You should > rerun javadoc after executing mvn install. > [WARNING] IGNORED to add some artifacts in the classpath. See above. > [WARNING] The dependency: [com.pivotal.hawq:hawq-mapreduce-common:jar:1.1.0] > can't be resolved but has been found in the reactor (probably snapshots). > This dependency has been excluded from the Javadoc classpath. You should > rerun javadoc after executing mvn install. > [WARNING] IGNORED to add some artifacts in the classpath. See above. > [WARNING] The dependency: [com.pivotal.hawq:hawq-mapreduce-common:jar:1.1.0] > can't be resolved but has been found in the reactor (probably snapshots). > This dependency has been excluded from the Javadoc classpath. You should > rerun javadoc after executing mvn install. > [WARNING] The dependency: [com.pivotal.hawq:hawq-mapreduce-ao:jar:1.1.0] > can't be resolved but has been found in the reactor (probably snapshots). > This dependency has been excluded from the Javadoc classpath. You should > rerun javadoc after executing mvn install. > [WARNING] The dependency: [com.pivotal.hawq:hawq-mapreduce-parquet:jar:1.1.0] > can't be resolved but has been found in the reactor (probably snapshots). > This dependency has been excluded from the Javadoc classpath. You should > rerun javadoc after executing mvn install. > [WARNING] IGNORED to add some artifacts in the classpath. See above. > [INFO] > 正在加载程序包com.pivotal.hawq.mapreduce.conf的源文件... > 正在加载程序包com.pivotal.hawq.mapreduce.datatype的源文件... > 正在加载程序包com.pivotal.hawq.mapreduce.file的源文件... > 正在加载程序包com.pivotal.hawq.mapreduce的源文件... > 正在加载程序包com.pivotal.hawq.mapreduce.metadata的源文件... > 正在加载程序包com.pivotal.hawq.mapreduce.schema的源文件... > 正在加载程序包com.pivotal.hawq.mapreduce.util的源文件... > 正在加载程序包com.pivotal.hawq.mapreduce.ao.file的源文件... > 正在加载程序包com.pivotal.hawq.mapreduce.ao的源文件... > 正在加载程序包com.pivotal.hawq.mapreduce.ao.io的源文件... > 正在加载程序包com.pivotal.hawq.mapreduce.ao.util的源文件... > 正在加载程序包com.pivotal.hawq.mapreduce.parquet.convert的源文件... > 正在加载程序包com.pivotal.hawq.mapreduce.parquet的源文件... > 正在加载程序包com.pivotal.hawq.mapreduce.parquet.support的源文件... > 正在加载程序包com.pivotal.hawq.mapreduce.parquet.util的源文件... > 正在构造 Javadoc 信息... > 100 个错误 > [INFO] > ------------------------------------------------------------------------ > [INFO] Reactor Summary: > [INFO] > [INFO] hawq-hadoop 1.1.0 .................................. FAILURE [ 7.035 s] > [INFO] hawq-mapreduce-common .............................. SKIPPED > [INFO] hawq-mapreduce-ao .................................. SKIPPED > [INFO] hawq-mapreduce-parquet ............................. SKIPPED > [INFO] hawq-mapreduce-tool 1.1.0 .......................... SKIPPED > [INFO] > ------------------------------------------------------------------------ > [INFO] BUILD FAILURE > [INFO] > ------------------------------------------------------------------------ > [INFO] Total time: 7.235 s > [INFO] Finished at: 2018-07-05T13:50:16+08:00 > [INFO] > ------------------------------------------------------------------------ > [ERROR] Failed to execute goal > org.apache.maven.plugins:maven-javadoc-plugin:2.9.1:aggregate-jar (default) > on project hawq-hadoop: MavenReportException: Error while creating archive: > [ERROR] Exit code: 1 - > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/HAWQParquetInputFormat.java:34: > 错误: 程序包parquet.hadoop不存在 > [ERROR] import parquet.hadoop.ParquetInputFormat; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/HAWQParquetInputFormat.java:42: > 错误: 找不到符号 > [ERROR] public class HAWQParquetInputFormat extends > ParquetInputFormat<HAWQRecord> { > [ERROR] ^ > [ERROR] 符号: 类 ParquetInputFormat > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQBoxConverter.java:25: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.Converter; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQBoxConverter.java:26: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.GroupConverter; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQBoxConverter.java:27: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.PrimitiveConverter; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQBoxConverter.java:38: > 错误: 找不到符号 > [ERROR] public class HAWQBoxConverter extends GroupConverter { > [ERROR] ^ > [ERROR] 符号: 类 GroupConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQBoxConverter.java:41: > 错误: 找不到符号 > [ERROR] private Converter[] converters; > [ERROR] ^ > [ERROR] 符号: 类 Converter > [ERROR] 位置: 类 HAWQBoxConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQBoxConverter.java:78: > 错误: 找不到符号 > [ERROR] public Converter getConverter(int fieldIndex) { > [ERROR] ^ > [ERROR] 符号: 类 Converter > [ERROR] 位置: 类 HAWQBoxConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQCircleConverter.java:25: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.Converter; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQCircleConverter.java:26: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.GroupConverter; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQCircleConverter.java:27: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.PrimitiveConverter; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQCircleConverter.java:37: > 错误: 找不到符号 > [ERROR] public class HAWQCircleConverter extends GroupConverter { > [ERROR] ^ > [ERROR] 符号: 类 GroupConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQCircleConverter.java:40: > 错误: 找不到符号 > [ERROR] private Converter[] converters; > [ERROR] ^ > [ERROR] 符号: 类 Converter > [ERROR] 位置: 类 HAWQCircleConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQCircleConverter.java:70: > 错误: 找不到符号 > [ERROR] public Converter getConverter(int fieldIndex) { > [ERROR] ^ > [ERROR] 符号: 类 Converter > [ERROR] 位置: 类 HAWQCircleConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQLineSegmentConverter.java:25: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.Converter; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQLineSegmentConverter.java:26: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.GroupConverter; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQLineSegmentConverter.java:27: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.PrimitiveConverter; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQLineSegmentConverter.java:38: > 错误: 找不到符号 > [ERROR] public class HAWQLineSegmentConverter extends GroupConverter { > [ERROR] ^ > [ERROR] 符号: 类 GroupConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQLineSegmentConverter.java:41: > 错误: 找不到符号 > [ERROR] private Converter[] converters; > [ERROR] ^ > [ERROR] 符号: 类 Converter > [ERROR] 位置: 类 HAWQLineSegmentConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQLineSegmentConverter.java:78: > 错误: 找不到符号 > [ERROR] public Converter getConverter(int fieldIndex) { > [ERROR] ^ > [ERROR] 符号: 类 Converter > [ERROR] 位置: 类 HAWQLineSegmentConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPathConverter.java:26: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.Converter; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPathConverter.java:27: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.GroupConverter; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPathConverter.java:42: > 错误: 找不到符号 > [ERROR] public class HAWQPathConverter extends GroupConverter { > [ERROR] ^ > [ERROR] 符号: 类 GroupConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPathConverter.java:45: > 错误: 找不到符号 > [ERROR] private Converter[] converters; > [ERROR] ^ > [ERROR] 符号: 类 Converter > [ERROR] 位置: 类 HAWQPathConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPathConverter.java:70: > 错误: 找不到符号 > [ERROR] public Converter getConverter(int fieldIndex) { > [ERROR] ^ > [ERROR] 符号: 类 Converter > [ERROR] 位置: 类 HAWQPathConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPointConverter.java:25: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.Converter; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPointConverter.java:26: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.GroupConverter; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPointConverter.java:27: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.PrimitiveConverter; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPointConverter.java:36: > 错误: 找不到符号 > [ERROR] public class HAWQPointConverter extends GroupConverter { > [ERROR] ^ > [ERROR] 符号: 类 GroupConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPointConverter.java:39: > 错误: 找不到符号 > [ERROR] private Converter[] converters; > [ERROR] ^ > [ERROR] 符号: 类 Converter > [ERROR] 位置: 类 HAWQPointConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPointConverter.java:62: > 错误: 找不到符号 > [ERROR] public Converter getConverter(int fieldIndex) { > [ERROR] ^ > [ERROR] 符号: 类 Converter > [ERROR] 位置: 类 HAWQPointConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPolygonConverter.java:27: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.Converter; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPolygonConverter.java:28: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.GroupConverter; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPolygonConverter.java:48: > 错误: 找不到符号 > [ERROR] public class HAWQPolygonConverter extends GroupConverter { > [ERROR] ^ > [ERROR] 符号: 类 GroupConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPolygonConverter.java:51: > 错误: 找不到符号 > [ERROR] private Converter[] converters; > [ERROR] ^ > [ERROR] 符号: 类 Converter > [ERROR] 位置: 类 HAWQPolygonConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQPolygonConverter.java:77: > 错误: 找不到符号 > [ERROR] public Converter getConverter(int fieldIndex) { > [ERROR] ^ > [ERROR] 符号: 类 Converter > [ERROR] 位置: 类 HAWQPolygonConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:29: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.Binary; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:30: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.Converter; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:31: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.GroupConverter; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:32: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.PrimitiveConverter; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:33: > 错误: 程序包parquet.schema不存在 > [ERROR] import parquet.schema.MessageType; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:44: > 错误: 找不到符号 > [ERROR] public class HAWQRecordConverter extends GroupConverter { > [ERROR] ^ > [ERROR] 符号: 类 GroupConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:48: > 错误: 找不到符号 > [ERROR] private final Converter[] converters; > [ERROR] ^ > [ERROR] 符号: 类 Converter > [ERROR] 位置: 类 HAWQRecordConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:52: > 错误: 找不到符号 > [ERROR] public HAWQRecordConverter(MessageType requestedSchema, HAWQSchema > hawqSchema) { > [ERROR] ^ > [ERROR] 符号: 类 MessageType > [ERROR] 位置: 类 HAWQRecordConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:56: > 错误: 找不到符号 > [ERROR] public HAWQRecordConverter(ParentValueContainer parent, MessageType > requestedSchema, HAWQSchema hawqSchema) { > [ERROR] ^ > [ERROR] 符号: 类 MessageType > [ERROR] 位置: 类 HAWQRecordConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:201: > 错误: 找不到符号 > [ERROR] private Converter newConverter(HAWQField hawqType, > ParentValueContainer parent) { > [ERROR] ^ > [ERROR] 符号: 类 Converter > [ERROR] 位置: 类 HAWQRecordConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:259: > 错误: 找不到符号 > [ERROR] public Converter getConverter(int fieldIndex) { > [ERROR] ^ > [ERROR] 符号: 类 Converter > [ERROR] 位置: 类 HAWQRecordConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:285: > 错误: 找不到符号 > [ERROR] static class HAWQPrimitiveConverter extends PrimitiveConverter { > [ERROR] ^ > [ERROR] 符号: 类 PrimitiveConverter > [ERROR] 位置: 类 HAWQRecordConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:328: > 错误: 找不到符号 > [ERROR] static class HAWQShortConverter extends PrimitiveConverter { > [ERROR] ^ > [ERROR] 符号: 类 PrimitiveConverter > [ERROR] 位置: 类 HAWQRecordConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:343: > 错误: 找不到符号 > [ERROR] static class HAWQBigDecimalConverter extends PrimitiveConverter { > [ERROR] ^ > [ERROR] 符号: 类 PrimitiveConverter > [ERROR] 位置: 类 HAWQRecordConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:351: > 错误: 找不到符号 > [ERROR] public void addBinary(Binary value) { > [ERROR] ^ > [ERROR] 符号: 类 Binary > [ERROR] 位置: 类 HAWQBigDecimalConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:361: > 错误: 找不到符号 > [ERROR] static class HAWQStringConverter extends PrimitiveConverter { > [ERROR] ^ > [ERROR] 符号: 类 PrimitiveConverter > [ERROR] 位置: 类 HAWQRecordConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:369: > 错误: 找不到符号 > [ERROR] public void addBinary(Binary value) { > [ERROR] ^ > [ERROR] 符号: 类 Binary > [ERROR] 位置: 类 HAWQStringConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:378: > 错误: 找不到符号 > [ERROR] static class HAWQBitsConverter extends PrimitiveConverter { > [ERROR] ^ > [ERROR] 符号: 类 PrimitiveConverter > [ERROR] 位置: 类 HAWQRecordConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:386: > 错误: 找不到符号 > [ERROR] public void addBinary(Binary value) { > [ERROR] ^ > [ERROR] 符号: 类 Binary > [ERROR] 位置: 类 HAWQBitsConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:395: > 错误: 找不到符号 > [ERROR] static class HAWQByteArrayConverter extends PrimitiveConverter { > [ERROR] ^ > [ERROR] 符号: 类 PrimitiveConverter > [ERROR] 位置: 类 HAWQRecordConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:403: > 错误: 找不到符号 > [ERROR] public void addBinary(Binary value) { > [ERROR] ^ > [ERROR] 符号: 类 Binary > [ERROR] 位置: 类 HAWQByteArrayConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:411: > 错误: 找不到符号 > [ERROR] static class HAWQDateConverter extends PrimitiveConverter { > [ERROR] ^ > [ERROR] 符号: 类 PrimitiveConverter > [ERROR] 位置: 类 HAWQRecordConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:430: > 错误: 找不到符号 > [ERROR] static class HAWQTimeConverter extends PrimitiveConverter { > [ERROR] ^ > [ERROR] 符号: 类 PrimitiveConverter > [ERROR] 位置: 类 HAWQRecordConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:448: > 错误: 找不到符号 > [ERROR] static class HAWQTimeTZConverter extends PrimitiveConverter { > [ERROR] ^ > [ERROR] 符号: 类 PrimitiveConverter > [ERROR] 位置: 类 HAWQRecordConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:456: > 错误: 找不到符号 > [ERROR] public void addBinary(Binary value) { > [ERROR] ^ > [ERROR] 符号: 类 Binary > [ERROR] 位置: 类 HAWQTimeTZConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:466: > 错误: 找不到符号 > [ERROR] static class HAWQTimestampConverter extends PrimitiveConverter { > [ERROR] ^ > [ERROR] 符号: 类 PrimitiveConverter > [ERROR] 位置: 类 HAWQRecordConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:484: > 错误: 找不到符号 > [ERROR] static class HAWQTimestampTZConverter extends PrimitiveConverter { > [ERROR] ^ > [ERROR] 符号: 类 PrimitiveConverter > [ERROR] 位置: 类 HAWQRecordConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:502: > 错误: 找不到符号 > [ERROR] static class HAWQIntervalConverter extends PrimitiveConverter { > [ERROR] ^ > [ERROR] 符号: 类 PrimitiveConverter > [ERROR] 位置: 类 HAWQRecordConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:510: > 错误: 找不到符号 > [ERROR] public void addBinary(Binary value) { > [ERROR] ^ > [ERROR] 符号: 类 Binary > [ERROR] 位置: 类 HAWQIntervalConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:520: > 错误: 找不到符号 > [ERROR] static class HAWQMacaddrConverter extends PrimitiveConverter { > [ERROR] ^ > [ERROR] 符号: 类 PrimitiveConverter > [ERROR] 位置: 类 HAWQRecordConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:528: > 错误: 找不到符号 > [ERROR] public void addBinary(Binary value) { > [ERROR] ^ > [ERROR] 符号: 类 Binary > [ERROR] 位置: 类 HAWQMacaddrConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:538: > 错误: 找不到符号 > [ERROR] static class HAWQInetConverter extends PrimitiveConverter { > [ERROR] ^ > [ERROR] 符号: 类 PrimitiveConverter > [ERROR] 位置: 类 HAWQRecordConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:546: > 错误: 找不到符号 > [ERROR] public void addBinary(Binary value) { > [ERROR] ^ > [ERROR] 符号: 类 Binary > [ERROR] 位置: 类 HAWQInetConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:556: > 错误: 找不到符号 > [ERROR] static class HAWQCidrConverter extends PrimitiveConverter { > [ERROR] ^ > [ERROR] 符号: 类 PrimitiveConverter > [ERROR] 位置: 类 HAWQRecordConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordConverter.java:564: > 错误: 找不到符号 > [ERROR] public void addBinary(Binary value) { > [ERROR] ^ > [ERROR] 符号: 类 Binary > [ERROR] 位置: 类 HAWQCidrConverter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordMaterializer.java:25: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.GroupConverter; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordMaterializer.java:26: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.RecordMaterializer; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordMaterializer.java:27: > 错误: 程序包parquet.schema不存在 > [ERROR] import parquet.schema.MessageType; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordMaterializer.java:36: > 错误: 找不到符号 > [ERROR] public class HAWQRecordMaterializer extends > RecordMaterializer<HAWQRecord> { > [ERROR] ^ > [ERROR] 符号: 类 RecordMaterializer > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordMaterializer.java:40: > 错误: 找不到符号 > [ERROR] public HAWQRecordMaterializer(MessageType requestedSchema, HAWQSchema > hawqSchema) { > [ERROR] ^ > [ERROR] 符号: 类 MessageType > [ERROR] 位置: 类 HAWQRecordMaterializer > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordMaterializer.java:50: > 错误: 找不到符号 > [ERROR] public GroupConverter getRootConverter() { > [ERROR] ^ > [ERROR] 符号: 类 GroupConverter > [ERROR] 位置: 类 HAWQRecordMaterializer > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordWriter.java:29: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.Binary; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordWriter.java:30: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.RecordConsumer; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordWriter.java:39: > 错误: 找不到符号 > [ERROR] private RecordConsumer consumer; > [ERROR] ^ > [ERROR] 符号: 类 RecordConsumer > [ERROR] 位置: 类 HAWQRecordWriter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/convert/HAWQRecordWriter.java:42: > 错误: 找不到符号 > [ERROR] public HAWQRecordWriter(RecordConsumer consumer, HAWQSchema schema) { > [ERROR] ^ > [ERROR] 符号: 类 RecordConsumer > [ERROR] 位置: 类 HAWQRecordWriter > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/HAWQParquetOutputFormat.java:27: > 错误: 程序包parquet.hadoop不存在 > [ERROR] import parquet.hadoop.ParquetOutputFormat; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/HAWQParquetOutputFormat.java:28: > 错误: 程序包parquet.hadoop.util不存在 > [ERROR] import parquet.hadoop.util.ContextUtil; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/HAWQParquetOutputFormat.java:30: > 错误: 找不到符号 > [ERROR] class HAWQParquetOutputFormat extends ParquetOutputFormat<HAWQRecord> > { > [ERROR] ^ > [ERROR] 符号: 类 ParquetOutputFormat > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQReadSupport.java:27: > 错误: 程序包parquet.hadoop.api不存在 > [ERROR] import parquet.hadoop.api.ReadSupport; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQReadSupport.java:28: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.RecordMaterializer; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQReadSupport.java:29: > 错误: 程序包parquet.schema不存在 > [ERROR] import parquet.schema.MessageType; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQReadSupport.java:37: > 错误: 找不到符号 > [ERROR] public class HAWQReadSupport extends ReadSupport<HAWQRecord> { > [ERROR] ^ > [ERROR] 符号: 类 ReadSupport > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQReadSupport.java:45: > 错误: 找不到符号 > [ERROR] MessageType fileSchema) { > [ERROR] ^ > [ERROR] 符号: 类 MessageType > [ERROR] 位置: 类 HAWQReadSupport > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQReadSupport.java:43: > 错误: 找不到符号 > [ERROR] public ReadContext init(Configuration configuration, > [ERROR] ^ > [ERROR] 符号: 类 ReadContext > [ERROR] 位置: 类 HAWQReadSupport > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQReadSupport.java:63: > 错误: 找不到符号 > [ERROR] MessageType fileSchema, ReadContext readContext) { > [ERROR] ^ > [ERROR] 符号: 类 MessageType > [ERROR] 位置: 类 HAWQReadSupport > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQReadSupport.java:63: > 错误: 找不到符号 > [ERROR] MessageType fileSchema, ReadContext readContext) { > [ERROR] ^ > [ERROR] 符号: 类 ReadContext > [ERROR] 位置: 类 HAWQReadSupport > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQReadSupport.java:61: > 错误: 找不到符号 > [ERROR] public RecordMaterializer<HAWQRecord> prepareForRead(Configuration > configuration, > [ERROR] ^ > [ERROR] 符号: 类 RecordMaterializer > [ERROR] 位置: 类 HAWQReadSupport > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQWriteSupport.java:29: > 错误: 程序包parquet.hadoop.api不存在 > [ERROR] import parquet.hadoop.api.WriteSupport; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQWriteSupport.java:30: > 错误: 程序包parquet.io.api不存在 > [ERROR] import parquet.io.api.RecordConsumer; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQWriteSupport.java:31: > 错误: 程序包parquet.schema不存在 > [ERROR] import parquet.schema.MessageType; > [ERROR] ^ > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQWriteSupport.java:40: > 错误: 找不到符号 > [ERROR] public class HAWQWriteSupport extends WriteSupport<HAWQRecord> { > [ERROR] ^ > [ERROR] 符号: 类 WriteSupport > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQWriteSupport.java:46: > 错误: 找不到符号 > [ERROR] private MessageType parquetSchema; > [ERROR] ^ > [ERROR] 符号: 类 MessageType > [ERROR] 位置: 类 HAWQWriteSupport > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQWriteSupport.java:54: > 错误: 找不到符号 > [ERROR] public WriteContext init(Configuration configuration) { > [ERROR] ^ > [ERROR] 符号: 类 WriteContext > [ERROR] 位置: 类 HAWQWriteSupport > [ERROR] > /Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/hawq-mapreduce-parquet/src/main/java/com/pivotal/hawq/mapreduce/parquet/support/HAWQWriteSupport.java:65: > 错误: 找不到符号 > [ERROR] public void prepareForWrite(RecordConsumer recordConsumer) { > [ERROR] ^ > [ERROR] 符号: 类 RecordConsumer > [ERROR] 位置: 类 HAWQWriteSupport > [ERROR] > [ERROR] Command line was: > /Library/Java/JavaVirtualMachines/jdk-10.0.1.jdk/Contents/Home/bin/javadoc > @options @packages > [ERROR] > [ERROR] Refer to the generated Javadoc files in > '/Users/wangziming/workplace/incubator-hawq/contrib/hawq-hadoop/target/apidocs' > dir. > [ERROR] > [ERROR] -> [Help 1] > [ERROR] > [ERROR] To see the full stack trace of the errors, re-run Maven with the -e > switch. > [ERROR] Re-run Maven using the -X switch to enable full debug logging. > [ERROR] > [ERROR] For more information about the errors and possible solutions, please > read the following articles: > [ERROR] [Help 1] > http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException > make[2]: *** [hawq-mapreduce-tool/target/hawq-mapreduce-tool-1.1.0.jar] Error > 1 > make[1]: *** [all] Error 2 > make: *** [all] Error 2{code} -- This message was sent by Atlassian JIRA (v7.6.3#76005)