See https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/176/
################################################################################### ########################## LAST 60 LINES OF THE CONSOLE ########################### [...truncated 7368 lines...] [INFO] [INFO] --- maven-checkstyle-plugin:2.15:checkstyle (default-cli) @ hadoop-hdfs-project --- [INFO] [INFO] --- findbugs-maven-plugin:3.0.0:findbugs (default-cli) @ hadoop-hdfs-project --- [INFO] ------------------------------------------------------------------------ [INFO] Reactor Summary: [INFO] [INFO] Apache Hadoop HDFS Client ......................... SUCCESS [01:03 min] [INFO] Apache Hadoop HDFS ................................ FAILURE [ 03:09 h] [INFO] Apache Hadoop HttpFS .............................. SKIPPED [INFO] Apache Hadoop HDFS BookKeeper Journal ............. SKIPPED [INFO] Apache Hadoop HDFS-NFS ............................ SKIPPED [INFO] Apache Hadoop HDFS Project ........................ SUCCESS [ 0.049 s] [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 03:10 h [INFO] Finished at: 2015-05-06T15:35:48+00:00 [INFO] Final Memory: 52M/283M [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project hadoop-hdfs: There was a timeout or other error in the fork -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException [ERROR] [ERROR] After correcting the problems, you can resume the build with the command [ERROR] mvn <goals> -rf :hadoop-hdfs Build step 'Execute shell' marked build as failure Archiving artifacts Sending artifact delta relative to Hadoop-Hdfs-trunk-Java8 #175 Archived 1 artifacts Archive block size is 32768 Received 0 blocks and 810640 bytes Compression is 0.0% Took 27 sec Recording test results Updating YARN-3582 Updating YARN-3552 Updating HADOOP-11120 Updating HADOOP-11904 Updating HADOOP-11912 Updating HDFS-8314 Updating HADOOP-11911 Updating HADOOP-11926 Updating HADOOP-11917 Updating YARN-3396 Updating MAPREDUCE-6192 Updating HDFS-8305 Updating HDFS-7758 Updating HDFS-7847 Updating HDFS-8219 Updating YARN-2123 Updating YARN-3343 Sending e-mails to: hdfs-dev@hadoop.apache.org Email was triggered for: Failure Sending email for trigger: Failure ################################################################################### ############################## FAILED TESTS (if any) ############################## 2 tests failed. REGRESSION: org.apache.hadoop.hdfs.TestAppendSnapshotTruncate.testAST Error Message: dir has ERROR Stack Trace: java.lang.IllegalStateException: dir has ERROR at org.apache.hadoop.hdfs.TestAppendSnapshotTruncate$Worker.checkErrorState(TestAppendSnapshotTruncate.java:429) at org.apache.hadoop.hdfs.TestAppendSnapshotTruncate$Worker.stop(TestAppendSnapshotTruncate.java:483) at org.apache.hadoop.hdfs.TestAppendSnapshotTruncate.testAST(TestAppendSnapshotTruncate.java:127) Caused by: java.lang.IllegalStateException: null at com.google.common.base.Preconditions.checkState(Preconditions.java:129) at org.apache.hadoop.hdfs.TestAppendSnapshotTruncate$Worker.pause(TestAppendSnapshotTruncate.java:479) at org.apache.hadoop.hdfs.TestAppendSnapshotTruncate$DirWorker.pauseAllFiles(TestAppendSnapshotTruncate.java:246) at org.apache.hadoop.hdfs.TestAppendSnapshotTruncate$DirWorker.call(TestAppendSnapshotTruncate.java:220) at org.apache.hadoop.hdfs.TestAppendSnapshotTruncate$DirWorker.call(TestAppendSnapshotTruncate.java:139) at org.apache.hadoop.hdfs.TestAppendSnapshotTruncate$Worker$1.run(TestAppendSnapshotTruncate.java:454) at java.lang.Thread.run(Thread.java:744) REGRESSION: org.apache.hadoop.tracing.TestTraceAdmin.testCreateAndDestroySpanReceiver Error Message: Failed to load SpanReceiver org.apache.htrace.impl.LocalFileSpanReceiver at org.apache.hadoop.tracing.SpanReceiverHost.loadInstance(SpanReceiverHost.java:171) at org.apache.hadoop.tracing.SpanReceiverHost.addSpanReceiver(SpanReceiverHost.java:216) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addSpanReceiver(NameNodeRpcServer.java:2029) at org.apache.hadoop.tracing.TraceAdminProtocolServerSideTranslatorPB.addSpanReceiver(TraceAdminProtocolServerSideTranslatorPB.java:81) at org.apache.hadoop.tracing.TraceAdminPB$TraceAdminService$2.callBlockingMethod(TraceAdminPB.java:4580) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:636) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:976) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2174) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2170) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1669) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2168) Stack Trace: org.apache.hadoop.ipc.RemoteException: Failed to load SpanReceiver org.apache.htrace.impl.LocalFileSpanReceiver at org.apache.hadoop.tracing.SpanReceiverHost.loadInstance(SpanReceiverHost.java:171) at org.apache.hadoop.tracing.SpanReceiverHost.addSpanReceiver(SpanReceiverHost.java:216) at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.addSpanReceiver(NameNodeRpcServer.java:2029) at org.apache.hadoop.tracing.TraceAdminProtocolServerSideTranslatorPB.addSpanReceiver(TraceAdminProtocolServerSideTranslatorPB.java:81) at org.apache.hadoop.tracing.TraceAdminPB$TraceAdminService$2.callBlockingMethod(TraceAdminPB.java:4580) at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:636) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:976) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2174) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2170) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1669) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2168) at org.apache.hadoop.ipc.Client.call(Client.java:1492) at org.apache.hadoop.ipc.Client.call(Client.java:1423) at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:229) at com.sun.proxy.$Proxy20.addSpanReceiver(Unknown Source) at org.apache.hadoop.tracing.TraceAdminProtocolTranslatorPB.addSpanReceiver(TraceAdminProtocolTranslatorPB.java:81) at org.apache.hadoop.tracing.TraceAdmin.addSpanReceiver(TraceAdmin.java:120) at org.apache.hadoop.tracing.TraceAdmin.run(TraceAdmin.java:182) at org.apache.hadoop.tracing.TestTraceAdmin.runTraceCommand(TestTraceAdmin.java:44) at org.apache.hadoop.tracing.TestTraceAdmin.testCreateAndDestroySpanReceiver(TestTraceAdmin.java:76)