See https://builds.apache.org/job/Hadoop-Mapreduce-trunk/2965/
################################################################################### ########################## LAST 60 LINES OF THE CONSOLE ########################### [...truncated 31958 lines...] Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 74.283 sec - in org.apache.hadoop.mapred.TestReduceFetchFromPartialMem Running org.apache.hadoop.mapred.TestSpecialCharactersInOutputPath Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 26.61 sec - in org.apache.hadoop.mapred.TestSpecialCharactersInOutputPath Running org.apache.hadoop.mapred.TestSequenceFileAsBinaryOutputFormat Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.303 sec - in org.apache.hadoop.mapred.TestSequenceFileAsBinaryOutputFormat Results : Failed tests: TestNetworkedJob.testNetworkedJob:174 expected:<[[Mon Feb 15 12:48:18 +0000 2016] Application is Activated, waiting for resources to be assigned for AM. Details : AM Partition = <DEFAULT_PARTITION> ; Partition Resource = <memory:8192, vCores:16> ; Queue's Absolute capacity = 100.0 % ; Queue's Absolute used capacity = 0.0 % ; Queue's Absolute max capacity = 100.0 % ; ]> but was:<[]> Tests in error: TestMRCJCSocketFactory.testSocketFactory:85->initAndStartMiniMRYarnCluster:112 » YarnRuntime TestFileSystem.testFsCache:513->runTestCache:527 » IO Unable to initialize Web... TestMiniMRClientCluster.setup:80 » NoClassDefFound org/apache/hadoop/yarn/serv... TestMiniMRClientCluster.cleanup:93 NullPointer TestReduceFetchFromPartialMem$1>TestSetup.run:27->setUp:61 » NoClassDefFound o... Tests run: 524, Failures: 1, Errors: 5, Skipped: 11 [INFO] ------------------------------------------------------------------------ [INFO] Reactor Summary: [INFO] [INFO] Apache Hadoop MapReduce Client .................... SUCCESS [ 2.811 s] [INFO] Apache Hadoop MapReduce Core ...................... SUCCESS [01:51 min] [INFO] Apache Hadoop MapReduce Common .................... SUCCESS [ 27.927 s] [INFO] Apache Hadoop MapReduce Shuffle ................... SUCCESS [ 5.158 s] [INFO] Apache Hadoop MapReduce App ....................... SUCCESS [09:23 min] [INFO] Apache Hadoop MapReduce HistoryServer ............. SUCCESS [05:39 min] [INFO] Apache Hadoop MapReduce JobClient ................. FAILURE [ 01:40 h] [INFO] Apache Hadoop MapReduce HistoryServer Plugins ..... SKIPPED [INFO] Apache Hadoop MapReduce NativeTask ................ SKIPPED [INFO] Apache Hadoop MapReduce Examples .................. SKIPPED [INFO] Apache Hadoop MapReduce ........................... SKIPPED [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 01:58 h [INFO] Finished at: 2016-02-15T12:55:35+00:00 [INFO] Final Memory: 33M/603M [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project hadoop-mapreduce-client-jobclient: ExecutionException: java.lang.RuntimeException: The forked VM terminated without properly saying goodbye. VM crash or System.exit called? [ERROR] Command was /bin/sh -c cd /home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient && /home/jenkins/tools/java/jdk1.7.0_55/jre/bin/java -Xmx2048m -XX:MaxPermSize=768m -XX:+HeapDumpOnOutOfMemoryError -jar /home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/target/surefire/surefirebooter5408370019631875991.jar /home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/target/surefire/surefire9056680930216046359tmp /home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/target/surefire/surefire_2123588810134663681855tmp [ERROR] -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException [ERROR] [ERROR] After correcting the problems, you can resume the build with the command [ERROR] mvn <goals> -rf :hadoop-mapreduce-client-jobclient Build step 'Execute shell' marked build as failure [FINDBUGS] Skipping publisher since build result is FAILURE Archiving artifacts Recording test results Email was triggered for: Failure - Any Sending email for trigger: Failure - Any ################################################################################### ############################## FAILED TESTS (if any) ############################## 6 tests failed. FAILED: org.apache.hadoop.fs.TestFileSystem.testFsCache Error Message: Unable to initialize WebAppContext Stack Trace: java.io.IOException: Unable to initialize WebAppContext at java.util.zip.InflaterInputStream.fill(InflaterInputStream.java:240) at java.util.zip.InflaterInputStream.read(InflaterInputStream.java:158) at java.util.zip.ZipInputStream.read(ZipInputStream.java:193) at java.util.jar.JarInputStream.read(JarInputStream.java:207) at java.util.zip.ZipInputStream.closeEntry(ZipInputStream.java:139) at java.util.zip.ZipInputStream.getNextEntry(ZipInputStream.java:117) at java.util.jar.JarInputStream.getNextEntry(JarInputStream.java:142) at java.util.jar.JarInputStream.getNextJarEntry(JarInputStream.java:179) at org.mortbay.resource.JarResource.extract(JarResource.java:144) at org.mortbay.jetty.webapp.WebAppContext.resolveWebApp(WebAppContext.java:974) at org.mortbay.jetty.webapp.WebAppContext.getWebInf(WebAppContext.java:832) at org.mortbay.jetty.webapp.WebInfConfiguration.configureClassLoader(WebInfConfiguration.java:62) at org.mortbay.jetty.webapp.WebAppContext.doStart(WebAppContext.java:489) at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50) at org.mortbay.jetty.handler.HandlerCollection.doStart(HandlerCollection.java:152) at org.mortbay.jetty.handler.ContextHandlerCollection.doStart(ContextHandlerCollection.java:156) at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50) at org.mortbay.jetty.handler.HandlerWrapper.doStart(HandlerWrapper.java:130) at org.mortbay.jetty.Server.doStart(Server.java:224) at org.mortbay.component.AbstractLifeCycle.start(AbstractLifeCycle.java:50) at org.apache.hadoop.http.HttpServer2.start(HttpServer2.java:851) at org.apache.hadoop.hdfs.server.datanode.web.DatanodeHttpServer.<init>(DatanodeHttpServer.java:103) at org.apache.hadoop.hdfs.server.datanode.DataNode.startInfoServer(DataNode.java:891) at org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1261) at org.apache.hadoop.hdfs.server.datanode.DataNode.<init>(DataNode.java:479) at org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2551) at org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2439) at org.apache.hadoop.hdfs.MiniDFSCluster.startDataNodes(MiniDFSCluster.java:1592) at org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:844) at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:482) at org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:441) at org.apache.hadoop.fs.TestFileSystem.runTestCache(TestFileSystem.java:527) at org.apache.hadoop.fs.TestFileSystem.testFsCache(TestFileSystem.java:513) FAILED: org.apache.hadoop.ipc.TestMRCJCSocketFactory.testSocketFactory Error Message: java.lang.NoClassDefFoundError: org/apache/hadoop/hdfs/protocol/proto/ClientNamenodeProtocolProtos$GetListingRequestProto$Builder Stack Trace: org.apache.hadoop.yarn.exceptions.YarnRuntimeException: java.lang.NoClassDefFoundError: org/apache/hadoop/hdfs/protocol/proto/ClientNamenodeProtocolProtos$GetListingRequestProto$Builder at java.net.URLClassLoader$1.run(URLClassLoader.java:366) at java.net.URLClassLoader$1.run(URLClassLoader.java:355) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:354) at java.lang.ClassLoader.loadClass(ClassLoader.java:425) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) at java.lang.ClassLoader.loadClass(ClassLoader.java:358) at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$GetListingRequestProto.newBuilder(ClientNamenodeProtocolProtos.java:29103) at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolTranslatorPB.getListing(ClientNamenodeProtocolTranslatorPB.java:544) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:606) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:255) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:103) at com.sun.proxy.$Proxy22.getListing(Unknown Source) at org.apache.hadoop.hdfs.DFSClient.listPaths(DFSClient.java:1603) at org.apache.hadoop.fs.Hdfs$DirListingIterator.<init>(Hdfs.java:211) at org.apache.hadoop.fs.Hdfs$DirListingIterator.<init>(Hdfs.java:198) at org.apache.hadoop.fs.Hdfs$2.<init>(Hdfs.java:180) at org.apache.hadoop.fs.Hdfs.listStatusIterator(Hdfs.java:180) at org.apache.hadoop.fs.FileContext$22.next(FileContext.java:1492) at org.apache.hadoop.fs.FileContext$22.next(FileContext.java:1487) at org.apache.hadoop.fs.FSLinkResolver.resolve(FSLinkResolver.java:90) at org.apache.hadoop.fs.FileContext.listStatus(FileContext.java:1487) at org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils.localGlobber(JobHistoryUtils.java:456) at org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils.localGlobber(JobHistoryUtils.java:444) at org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils.localGlobber(JobHistoryUtils.java:439) at org.apache.hadoop.mapreduce.v2.hs.HistoryFileManager.findTimestampedDirectories(HistoryFileManager.java:809) at org.apache.hadoop.mapreduce.v2.hs.HistoryFileManager.initExisting(HistoryFileManager.java:703) at org.apache.hadoop.mapreduce.v2.hs.JobHistory.serviceInit(JobHistory.java:97) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107) at org.apache.hadoop.mapreduce.v2.hs.JobHistoryServer.serviceInit(JobHistoryServer.java:152) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.mapreduce.v2.MiniMRYarnCluster$JobHistoryServerWrapper.serviceStart(MiniMRYarnCluster.java:232) at org.apache.hadoop.service.AbstractService.start(AbstractService.java:193) at org.apache.hadoop.service.CompositeService.serviceStart(CompositeService.java:120) at org.apache.hadoop.mapreduce.v2.MiniMRYarnCluster.serviceStart(MiniMRYarnCluster.java:191) at org.apache.hadoop.service.AbstractService.start(AbstractService.java:193) at org.apache.hadoop.ipc.TestMRCJCSocketFactory.initAndStartMiniMRYarnCluster(TestMRCJCSocketFactory.java:112) at org.apache.hadoop.ipc.TestMRCJCSocketFactory.testSocketFactory(TestMRCJCSocketFactory.java:85) FAILED: org.apache.hadoop.mapred.TestMiniMRClientCluster.org.apache.hadoop.mapred.TestMiniMRClientCluster Error Message: org/apache/hadoop/yarn/server/applicationhistoryservice/ApplicationHistoryWriter Stack Trace: java.lang.NoClassDefFoundError: org/apache/hadoop/yarn/server/applicationhistoryservice/ApplicationHistoryWriter at java.net.URLClassLoader$1.run(URLClassLoader.java:366) at java.net.URLClassLoader$1.run(URLClassLoader.java:355) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:354) at java.lang.ClassLoader.loadClass(ClassLoader.java:425) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) at java.lang.ClassLoader.loadClass(ClassLoader.java:358) at org.apache.hadoop.yarn.server.resourcemanager.ResourceManager.createRMApplicationHistoryWriter(ResourceManager.java:445) at org.apache.hadoop.yarn.server.resourcemanager.ResourceManager.serviceInit(ResourceManager.java:292) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.yarn.server.MiniYARNCluster.initResourceManager(MiniYARNCluster.java:300) at org.apache.hadoop.yarn.server.MiniYARNCluster.access$300(MiniYARNCluster.java:100) at org.apache.hadoop.yarn.server.MiniYARNCluster$ResourceManagerWrapper.serviceInit(MiniYARNCluster.java:452) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107) at org.apache.hadoop.yarn.server.MiniYARNCluster.serviceInit(MiniYARNCluster.java:274) at org.apache.hadoop.mapreduce.v2.MiniMRYarnCluster.serviceInit(MiniMRYarnCluster.java:186) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.mapred.MiniMRClientClusterFactory.create(MiniMRClientClusterFactory.java:79) at org.apache.hadoop.mapred.MiniMRClientClusterFactory.create(MiniMRClientClusterFactory.java:41) at org.apache.hadoop.mapred.TestMiniMRClientCluster.setup(TestMiniMRClientCluster.java:80) FAILED: org.apache.hadoop.mapred.TestMiniMRClientCluster.org.apache.hadoop.mapred.TestMiniMRClientCluster Error Message: null Stack Trace: java.lang.NullPointerException: null at org.apache.hadoop.mapred.TestMiniMRClientCluster.cleanup(TestMiniMRClientCluster.java:93) FAILED: org.apache.hadoop.mapred.TestReduceFetchFromPartialMem$1.org.apache.hadoop.mapred.TestReduceFetchFromPartialMem Error Message: org/apache/hadoop/yarn/server/MiniYARNCluster Stack Trace: java.lang.NoClassDefFoundError: org/apache/hadoop/yarn/server/MiniYARNCluster at java.net.URLClassLoader$1.run(URLClassLoader.java:366) at java.net.URLClassLoader$1.run(URLClassLoader.java:355) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:354) at java.lang.ClassLoader.loadClass(ClassLoader.java:425) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) at java.lang.ClassLoader.loadClass(ClassLoader.java:358) at java.lang.ClassLoader.defineClass1(Native Method) at java.lang.ClassLoader.defineClass(ClassLoader.java:800) at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) at java.net.URLClassLoader.defineClass(URLClassLoader.java:449) at java.net.URLClassLoader.access$100(URLClassLoader.java:71) at java.net.URLClassLoader$1.run(URLClassLoader.java:361) at java.net.URLClassLoader$1.run(URLClassLoader.java:355) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:354) at java.lang.ClassLoader.loadClass(ClassLoader.java:425) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) at java.lang.ClassLoader.loadClass(ClassLoader.java:358) at org.apache.hadoop.mapred.MiniMRClientClusterFactory.create(MiniMRClientClusterFactory.java:58) at org.apache.hadoop.mapred.MiniMRCluster.<init>(MiniMRCluster.java:187) at org.apache.hadoop.mapred.MiniMRCluster.<init>(MiniMRCluster.java:175) at org.apache.hadoop.mapred.MiniMRCluster.<init>(MiniMRCluster.java:167) at org.apache.hadoop.mapred.MiniMRCluster.<init>(MiniMRCluster.java:159) at org.apache.hadoop.mapred.MiniMRCluster.<init>(MiniMRCluster.java:152) at org.apache.hadoop.mapred.MiniMRCluster.<init>(MiniMRCluster.java:145) at org.apache.hadoop.mapred.MiniMRCluster.<init>(MiniMRCluster.java:138) at org.apache.hadoop.mapred.MiniMRCluster.<init>(MiniMRCluster.java:133) at org.apache.hadoop.mapred.TestReduceFetchFromPartialMem$1.setUp(TestReduceFetchFromPartialMem.java:61) FAILED: org.apache.hadoop.mapred.TestNetworkedJob.testNetworkedJob Error Message: expected:<[[Mon Feb 15 12:48:18 +0000 2016] Application is Activated, waiting for resources to be assigned for AM. Details : AM Partition = <DEFAULT_PARTITION> ; Partition Resource = <memory:8192, vCores:16> ; Queue's Absolute capacity = 100.0 % ; Queue's Absolute used capacity = 0.0 % ; Queue's Absolute max capacity = 100.0 % ; ]> but was:<[]> Stack Trace: org.junit.ComparisonFailure: expected:<[[Mon Feb 15 12:48:18 +0000 2016] Application is Activated, waiting for resources to be assigned for AM. Details : AM Partition = <DEFAULT_PARTITION> ; Partition Resource = <memory:8192, vCores:16> ; Queue's Absolute capacity = 100.0 % ; Queue's Absolute used capacity = 0.0 % ; Queue's Absolute max capacity = 100.0 % ; ]> but was:<[]> at org.junit.Assert.assertEquals(Assert.java:115) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapred.TestNetworkedJob.testNetworkedJob(TestNetworkedJob.java:174)