See <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/591/changes>
Changes: [waltersu4549] HDFS-9401. Fix findbugs warnings in BlockRecoveryWorker. Contributed by [lei] HADOOP-12482. Race condition in JMX cache update. (Tony Wu via lei) ------------------------------------------ [...truncated 11654 lines...] [WARNING] ^ [WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/target/generated-sources/java/org/apache/hadoop/contrib/bkjournal/BKJournalProtos.java>:522: warning: no @param for value [WARNING] public Builder setLayoutVersion(int value) { [WARNING] ^ [WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/target/generated-sources/java/org/apache/hadoop/contrib/bkjournal/BKJournalProtos.java>:522: warning: no @return [WARNING] public Builder setLayoutVersion(int value) { [WARNING] ^ [WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/target/generated-sources/java/org/apache/hadoop/contrib/bkjournal/BKJournalProtos.java>:531: warning: no @return [WARNING] public Builder clearLayoutVersion() { [WARNING] ^ [WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/target/generated-sources/java/org/apache/hadoop/contrib/bkjournal/BKJournalProtos.java>:561: warning: no @param for value [WARNING] public Builder setNamespaceInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto value) { [WARNING] ^ [WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/target/generated-sources/java/org/apache/hadoop/contrib/bkjournal/BKJournalProtos.java>:561: warning: no @return [WARNING] public Builder setNamespaceInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto value) { [WARNING] ^ [WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/target/generated-sources/java/org/apache/hadoop/contrib/bkjournal/BKJournalProtos.java>:577: warning: no @param for builderForValue [WARNING] public Builder setNamespaceInfo( [WARNING] ^ [WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/target/generated-sources/java/org/apache/hadoop/contrib/bkjournal/BKJournalProtos.java>:577: warning: no @return [WARNING] public Builder setNamespaceInfo( [WARNING] ^ [WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/target/generated-sources/java/org/apache/hadoop/contrib/bkjournal/BKJournalProtos.java>:591: warning: no @param for value [WARNING] public Builder mergeNamespaceInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto value) { [WARNING] ^ [WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/target/generated-sources/java/org/apache/hadoop/contrib/bkjournal/BKJournalProtos.java>:591: warning: no @return [WARNING] public Builder mergeNamespaceInfo(org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto value) { [WARNING] ^ [WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/target/generated-sources/java/org/apache/hadoop/contrib/bkjournal/BKJournalProtos.java>:610: warning: no @return [WARNING] public Builder clearNamespaceInfo() { [WARNING] ^ [WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/target/generated-sources/java/org/apache/hadoop/contrib/bkjournal/BKJournalProtos.java>:623: warning: no @return [WARNING] public org.apache.hadoop.hdfs.protocol.proto.HdfsServerProtos.NamespaceInfoProto.Builder getNamespaceInfoBuilder() { [WARNING] ^ [WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/src/main/java/org/apache/hadoop/contrib/bkjournal/BookKeeperJournalManager.java>:185: warning: no @param for conf [WARNING] public BookKeeperJournalManager(Configuration conf, URI uri, [WARNING] ^ [WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/src/main/java/org/apache/hadoop/contrib/bkjournal/BookKeeperJournalManager.java>:185: warning: no @param for uri [WARNING] public BookKeeperJournalManager(Configuration conf, URI uri, [WARNING] ^ [WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/src/main/java/org/apache/hadoop/contrib/bkjournal/BookKeeperJournalManager.java>:185: warning: no @param for nsInfo [WARNING] public BookKeeperJournalManager(Configuration conf, URI uri, [WARNING] ^ [WARNING] <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/src/main/java/org/apache/hadoop/contrib/bkjournal/BookKeeperJournalManager.java>:185: warning: no @throws for java.io.IOException [WARNING] public BookKeeperJournalManager(Configuration conf, URI uri, [WARNING] ^ [INFO] Building jar: <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/target/hadoop-hdfs-bkjournal-3.0.0-SNAPSHOT-javadoc.jar> [INFO] [INFO] --- maven-dependency-plugin:2.8:copy (dist) @ hadoop-hdfs-bkjournal --- [INFO] Configured Artifact: org.apache.bookkeeper:bookkeeper-server:?:jar [INFO] Copying bookkeeper-server-4.2.3.jar to <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs/src/contrib/bkjournal/target/lib/bookkeeper-server-4.2.3.jar> [INFO] [INFO] --- maven-enforcer-plugin:1.3.1:enforce (depcheck) @ hadoop-hdfs-bkjournal --- [INFO] [INFO] --- maven-checkstyle-plugin:2.15:checkstyle (default-cli) @ hadoop-hdfs-bkjournal --- [INFO] There are 613 errors reported by Checkstyle 6.6 with checkstyle/checkstyle.xml ruleset. [WARNING] Unable to locate Source XRef to link to - DISABLED [INFO] [INFO] --- findbugs-maven-plugin:3.0.0:findbugs (default-cli) @ hadoop-hdfs-bkjournal --- [INFO] Fork Value is true [INFO] Done FindBugs Analysis.... [INFO] [INFO] ------------------------------------------------------------------------ [INFO] Building Apache Hadoop HDFS-NFS 3.0.0-SNAPSHOT [INFO] ------------------------------------------------------------------------ [INFO] [INFO] --- maven-clean-plugin:2.5:clean (default-clean) @ hadoop-hdfs-nfs --- [INFO] Deleting <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs-nfs/target> [INFO] [INFO] --- maven-antrun-plugin:1.7:run (create-testdirs) @ hadoop-hdfs-nfs --- [INFO] Executing tasks main: [mkdir] Created dir: <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs-nfs/target/test-dir> [INFO] Executed tasks [INFO] [INFO] --- maven-resources-plugin:2.6:resources (default-resources) @ hadoop-hdfs-nfs --- [INFO] Using 'UTF-8' encoding to copy filtered resources. [INFO] skip non existing resourceDirectory <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs-nfs/src/main/resources> [INFO] [INFO] --- maven-compiler-plugin:3.1:compile (default-compile) @ hadoop-hdfs-nfs --- [INFO] Changes detected - recompiling the module! [INFO] Compiling 17 source files to <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs-nfs/target/classes> [INFO] [INFO] --- maven-resources-plugin:2.6:testResources (default-testResources) @ hadoop-hdfs-nfs --- [INFO] Using 'UTF-8' encoding to copy filtered resources. [INFO] Copying 1 resource [INFO] [INFO] --- maven-compiler-plugin:3.1:testCompile (default-testCompile) @ hadoop-hdfs-nfs --- [INFO] Changes detected - recompiling the module! [INFO] Compiling 13 source files to <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs-nfs/target/test-classes> [INFO] [INFO] --- maven-surefire-plugin:2.17:test (default-test) @ hadoop-hdfs-nfs --- [INFO] Surefire report directory: <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs-nfs/target/surefire-reports> ------------------------------------------------------- T E S T S ------------------------------------------------------- ------------------------------------------------------- T E S T S ------------------------------------------------------- Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0 Running org.apache.hadoop.hdfs.nfs.TestMountd Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.387 sec - in org.apache.hadoop.hdfs.nfs.TestMountd Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0 Running org.apache.hadoop.hdfs.nfs.nfs3.TestDFSClientCache Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.24 sec - in org.apache.hadoop.hdfs.nfs.nfs3.TestDFSClientCache Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0 Running org.apache.hadoop.hdfs.nfs.nfs3.TestOffsetRange Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.098 sec - in org.apache.hadoop.hdfs.nfs.nfs3.TestOffsetRange Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0 Running org.apache.hadoop.hdfs.nfs.nfs3.TestNfs3HttpServer Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 3.66 sec <<< FAILURE! - in org.apache.hadoop.hdfs.nfs.nfs3.TestNfs3HttpServer org.apache.hadoop.hdfs.nfs.nfs3.TestNfs3HttpServer Time elapsed: 3.66 sec <<< ERROR! java.net.BindException: Problem binding to [0.0.0.0:50475] java.net.BindException: Address already in use; For more details see: http://wiki.apache.org/hadoop/BindException at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:408) at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:792) at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:721) at org.apache.hadoop.hdfs.server.datanode.web.DatanodeHttpServer.start(DatanodeHttpServer.java:218) at org.apache.hadoop.hdfs.server.datanode.DataNode.startInfoServer(DataNode.java:838) at org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1197) at org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:467) at org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2486) at org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2374) at org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1592) at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:844) at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:482) at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:441) at org.apache.hadoop.hdfs.nfs.nfs3.TestNfs3HttpServer.setUp(TestNfs3HttpServer.java:62) Caused by: java.net.BindException: Address already in use at sun.nio.ch.Net.bind0(Native Method) at sun.nio.ch.Net.bind(Net.java:414) at sun.nio.ch.Net.bind(Net.java:406) at sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:214) at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74) at io.netty.channel.socket.nio.NioServerSocketChannel.doBind(NioServerSocketChannel.java:125) at io.netty.channel.AbstractChannel$AbstractUnsafe.bind(AbstractChannel.java:522) at io.netty.channel.DefaultChannelPipeline$HeadContext.bind(DefaultChannelPipeline.java:1196) at io.netty.channel.ChannelHandlerInvokerUtil.invokeBindNow(ChannelHandlerInvokerUtil.java:108) at io.netty.channel.DefaultChannelHandlerInvoker.invokeBind(DefaultChannelHandlerInvoker.java:214) at io.netty.channel.AbstractChannelHandlerContext.bind(AbstractChannelHandlerContext.java:208) at io.netty.channel.DefaultChannelPipeline.bind(DefaultChannelPipeline.java:1003) at io.netty.channel.AbstractChannel.bind(AbstractChannel.java:216) at io.netty.bootstrap.AbstractBootstrap$2.run(AbstractBootstrap.java:357) at io.netty.util.concurrent.SingleThreadEventExecutor.runAllTasks(SingleThreadEventExecutor.java:322) at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:356) at io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:703) at io.netty.util.concurrent.DefaultThreadFactory$DefaultRunnableDecorator.run(DefaultThreadFactory.java:137) at java.lang.Thread.run(Thread.java:744) Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0 Running org.apache.hadoop.hdfs.nfs.nfs3.TestClientAccessPrivilege Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.566 sec - in org.apache.hadoop.hdfs.nfs.nfs3.TestClientAccessPrivilege Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0 Running org.apache.hadoop.hdfs.nfs.nfs3.TestReaddir Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.814 sec - in org.apache.hadoop.hdfs.nfs.nfs3.TestReaddir Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0 Running org.apache.hadoop.hdfs.nfs.nfs3.TestOpenFileCtxCache Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 32.047 sec - in org.apache.hadoop.hdfs.nfs.nfs3.TestOpenFileCtxCache Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0 Running org.apache.hadoop.hdfs.nfs.nfs3.TestNfs3Utils Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 0.256 sec - in org.apache.hadoop.hdfs.nfs.nfs3.TestNfs3Utils Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0 Running org.apache.hadoop.hdfs.nfs.nfs3.TestExportsTable Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.277 sec - in org.apache.hadoop.hdfs.nfs.nfs3.TestExportsTable Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0 Running org.apache.hadoop.hdfs.nfs.nfs3.TestWrites Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 19.406 sec - in org.apache.hadoop.hdfs.nfs.nfs3.TestWrites Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=768m; support was removed in 8.0 Running org.apache.hadoop.hdfs.nfs.nfs3.TestRpcProgramNfs3 Tests run: 22, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 6.735 sec - in org.apache.hadoop.hdfs.nfs.nfs3.TestRpcProgramNfs3 Results : Tests in error: TestNfs3HttpServer.setUp:62 ยป Bind Problem binding to [0.0.0.0:50475] java.net... Tests run: 49, Failures: 0, Errors: 1, Skipped: 0 [INFO] [INFO] ------------------------------------------------------------------------ [INFO] Building Apache Hadoop HDFS Project 3.0.0-SNAPSHOT [INFO] ------------------------------------------------------------------------ [INFO] [INFO] --- maven-clean-plugin:2.5:clean (default-clean) @ hadoop-hdfs-project --- [INFO] Deleting <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/target> [INFO] [INFO] --- maven-antrun-plugin:1.7:run (create-testdirs) @ hadoop-hdfs-project --- [INFO] Executing tasks main: [mkdir] Created dir: <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/target/test-dir> [INFO] Executed tasks [INFO] [INFO] --- maven-source-plugin:2.3:jar-no-fork (hadoop-java-sources) @ hadoop-hdfs-project --- [INFO] [INFO] --- maven-source-plugin:2.3:test-jar-no-fork (hadoop-java-sources) @ hadoop-hdfs-project --- [INFO] [INFO] --- maven-enforcer-plugin:1.3.1:enforce (dist-enforce) @ hadoop-hdfs-project --- [INFO] [INFO] --- maven-site-plugin:3.4:attach-descriptor (attach-descriptor) @ hadoop-hdfs-project --- [INFO] [INFO] --- maven-javadoc-plugin:2.8.1:jar (module-javadocs) @ hadoop-hdfs-project --- [INFO] Not executing Javadoc as the project is not a Java classpath-capable package [INFO] [INFO] --- maven-enforcer-plugin:1.3.1:enforce (depcheck) @ hadoop-hdfs-project --- [INFO] [INFO] --- maven-checkstyle-plugin:2.15:checkstyle (default-cli) @ hadoop-hdfs-project --- [INFO] [INFO] --- findbugs-maven-plugin:3.0.0:findbugs (default-cli) @ hadoop-hdfs-project --- [INFO] ------------------------------------------------------------------------ [INFO] Reactor Summary: [INFO] [INFO] Apache Hadoop HDFS Client ......................... SUCCESS [04:00 min] [INFO] Apache Hadoop HDFS ................................ SUCCESS [ 03:22 h] [INFO] Apache Hadoop HDFS Native Client .................. SUCCESS [ 31.178 s] [INFO] Apache Hadoop HttpFS .............................. SUCCESS [04:33 min] [INFO] Apache Hadoop HDFS BookKeeper Journal ............. SUCCESS [03:51 min] [INFO] Apache Hadoop HDFS-NFS ............................ FAILURE [01:27 min] [INFO] Apache Hadoop HDFS Project ........................ SUCCESS [ 0.039 s] [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 03:36 h [INFO] Finished at: 2015-11-10T11:32:33+00:00 [INFO] Final Memory: 104M/1182M [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project hadoop-hdfs-nfs: There are test failures. [ERROR] [ERROR] Please refer to <https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/ws/hadoop-hdfs-project/hadoop-hdfs-nfs/target/surefire-reports> for the individual test results. [ERROR] -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException [ERROR] [ERROR] After correcting the problems, you can resume the build with the command [ERROR] mvn <goals> -rf :hadoop-hdfs-nfs Build step 'Execute shell' marked build as failure Archiving artifacts Recording test results Updating HDFS-9401 Updating HADOOP-12482