See https://builds.apache.org/job/Hadoop-Mapreduce-trunk/2337/
################################################################################### ########################## LAST 60 LINES OF THE CONSOLE ########################### [...truncated 33512 lines...] TestLocalRunner.testMultiMaps:320 » NullPointer TestLocalRunner.testMultiMapMultiReduce:569->doMultiReducerTest:546 » NullPointer TestLocalRunner.testOneMapMultiParallelReduce:559->doMultiReducerTest:546 » NullPointer TestLocalRunner.testMultiMapOneReduce:564->doMultiReducerTest:546 » NullPointer TestMapCollection.testRandomCompress:537->runTest:328 » Runtime org.xml.sax.SA... TestMapCollection.testSpillPer2B:348->runTest:301 » Runtime org.xml.sax.SAXPar... TestMapCollection.testValLastByte:334->runTest:301 » Runtime org.xml.sax.SAXPa... TestMapCollection.testRandom:500 » Runtime org.xml.sax.SAXParseException; syst... TestMRAppWithCombiner.testCombinerShouldUpdateTheReporter:124->runJob:146 » NoClassDefFound Tests run: 387, Failures: 2, Errors: 14, Skipped: 9 [INFO] ------------------------------------------------------------------------ [INFO] Reactor Summary: [INFO] [INFO] Apache Hadoop MapReduce Client .................... SUCCESS [ 3.076 s] [INFO] Apache Hadoop MapReduce Core ...................... SUCCESS [01:35 min] [INFO] Apache Hadoop MapReduce Common .................... SUCCESS [ 29.449 s] [INFO] Apache Hadoop MapReduce Shuffle ................... SUCCESS [ 5.106 s] [INFO] Apache Hadoop MapReduce App ....................... SUCCESS [08:59 min] [INFO] Apache Hadoop MapReduce HistoryServer ............. SUCCESS [05:37 min] [INFO] Apache Hadoop MapReduce JobClient ................. FAILURE [ 01:11 h] [INFO] Apache Hadoop MapReduce HistoryServer Plugins ..... SKIPPED [INFO] Apache Hadoop MapReduce NativeTask ................ SKIPPED [INFO] Apache Hadoop MapReduce Examples .................. SKIPPED [INFO] Apache Hadoop MapReduce ........................... SKIPPED [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 01:28 h [INFO] Finished at: 2015-09-15T20:35:34+00:00 [INFO] Final Memory: 44M/737M [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project hadoop-mapreduce-client-jobclient: ExecutionException: java.lang.RuntimeException: The forked VM terminated without properly saying goodbye. VM crash or System.exit called? [ERROR] Command was /bin/sh -c cd /home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient && /home/jenkins/tools/java/jdk1.7.0_55/jre/bin/java -Xmx4096m -XX:MaxPermSize=768m -XX:+HeapDumpOnOutOfMemoryError -jar /home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/target/surefire/surefirebooter2874754138681265931.jar /home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/target/surefire/surefire4151895669068123060tmp /home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/target/surefire/surefire_1504602903375118102902tmp [ERROR] -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException [ERROR] [ERROR] After correcting the problems, you can resume the build with the command [ERROR] mvn <goals> -rf :hadoop-mapreduce-client-jobclient Build step 'Execute shell' marked build as failure [FINDBUGS] Skipping publisher since build result is FAILURE Archiving artifacts Sending artifact delta relative to Hadoop-Mapreduce-trunk #2330 Archived 1 artifacts Archive block size is 32768 Received 0 blocks and 20467504 bytes Compression is 0.0% Took 5.3 sec Recording test results Updating HADOOP-12413 Updating HDFS-9008 Email was triggered for: Failure Sending email for trigger: Failure ################################################################################### ############################## FAILED TESTS (if any) ############################## 16 tests failed. FAILED: org.apache.hadoop.mapred.TestTextInputFormat.testSplitableCodecs Error Message: Key in multiple partitions. Stack Trace: java.lang.AssertionError: Key in multiple partitions. at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.assertTrue(Assert.java:41) at org.junit.Assert.assertFalse(Assert.java:64) at org.apache.hadoop.mapred.TestTextInputFormat.testSplitableCodecs(TestTextInputFormat.java:223) FAILED: org.apache.hadoop.mapreduce.TestLocalRunner.testGcCounter Error Message: org/apache/hadoop/io/FastByteComparisons Stack Trace: java.lang.NoClassDefFoundError: org/apache/hadoop/io/FastByteComparisons at java.net.URLClassLoader$1.run(URLClassLoader.java:366) at java.net.URLClassLoader$1.run(URLClassLoader.java:355) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:354) at java.lang.ClassLoader.loadClass(ClassLoader.java:425) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) at java.lang.ClassLoader.loadClass(ClassLoader.java:358) at org.apache.hadoop.io.WritableComparator.compareBytes(WritableComparator.java:188) at org.apache.hadoop.io.BinaryComparable.compareTo(BinaryComparable.java:50) at org.apache.hadoop.io.BinaryComparable.equals(BinaryComparable.java:72) at org.apache.hadoop.io.Text.equals(Text.java:348) at org.apache.hadoop.mapreduce.JobID.equals(JobID.java:89) at java.util.HashMap.getEntry(HashMap.java:471) at java.util.HashMap.get(HashMap.java:421) at org.apache.hadoop.mapred.LocalJobRunner.getJobCounters(LocalJobRunner.java:826) at org.apache.hadoop.mapreduce.Job$7.run(Job.java:761) at org.apache.hadoop.mapreduce.Job$7.run(Job.java:758) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667) at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:758) at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1383) at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1311) at org.apache.hadoop.mapreduce.TestLocalRunner.testGcCounter(TestLocalRunner.java:271) FAILED: org.apache.hadoop.mapreduce.TestLocalRunner.testInvalidMultiMapParallelism Error Message: null Stack Trace: java.lang.NullPointerException: null at org.apache.hadoop.io.serializer.SerializationFactory.<init>(SerializationFactory.java:58) at org.apache.hadoop.mapreduce.split.JobSplitWriter.writeNewSplits(JobSplitWriter.java:122) at org.apache.hadoop.mapreduce.split.JobSplitWriter.createSplitFiles(JobSplitWriter.java:79) at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:308) at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:319) at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:196) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667) at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287) at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1308) at org.apache.hadoop.mapreduce.TestLocalRunner.testInvalidMultiMapParallelism(TestLocalRunner.java:372) FAILED: org.apache.hadoop.mapreduce.TestLocalRunner.testOneMapMultiReduce Error Message: null Stack Trace: java.lang.NullPointerException: null at org.apache.hadoop.io.serializer.SerializationFactory.<init>(SerializationFactory.java:58) at org.apache.hadoop.mapreduce.split.JobSplitWriter.writeNewSplits(JobSplitWriter.java:122) at org.apache.hadoop.mapreduce.split.JobSplitWriter.createSplitFiles(JobSplitWriter.java:79) at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:308) at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:319) at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:196) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667) at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287) at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1308) at org.apache.hadoop.mapreduce.TestLocalRunner.doMultiReducerTest(TestLocalRunner.java:546) at org.apache.hadoop.mapreduce.TestLocalRunner.testOneMapMultiReduce(TestLocalRunner.java:554) FAILED: org.apache.hadoop.mapreduce.TestLocalRunner.testEmptyMaps Error Message: fs.AbstractFileSystem.file.impl=null: No AbstractFileSystem configured for scheme: file Stack Trace: org.apache.hadoop.fs.UnsupportedFileSystemException: fs.AbstractFileSystem.file.impl=null: No AbstractFileSystem configured for scheme: file at org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161) at org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:326) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667) at org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:326) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:443) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:420) at org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:406) at org.apache.hadoop.mapred.LocalDistributedCacheManager.setup(LocalDistributedCacheManager.java:116) at org.apache.hadoop.mapred.LocalJobRunner$Job.<init>(LocalJobRunner.java:172) at org.apache.hadoop.mapred.LocalJobRunner.submitJob(LocalJobRunner.java:786) at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:241) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667) at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287) at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1308) at org.apache.hadoop.mapreduce.TestLocalRunner.testEmptyMaps(TestLocalRunner.java:429) FAILED: org.apache.hadoop.mapreduce.TestLocalRunner.testMultiMaps Error Message: null Stack Trace: java.lang.NullPointerException: null at org.apache.hadoop.io.serializer.SerializationFactory.<init>(SerializationFactory.java:58) at org.apache.hadoop.mapreduce.split.JobSplitWriter.writeNewSplits(JobSplitWriter.java:122) at org.apache.hadoop.mapreduce.split.JobSplitWriter.createSplitFiles(JobSplitWriter.java:79) at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:308) at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:319) at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:196) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667) at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287) at org.apache.hadoop.mapreduce.TestLocalRunner.testMultiMaps(TestLocalRunner.java:320) FAILED: org.apache.hadoop.mapreduce.TestLocalRunner.testMultiMapMultiReduce Error Message: null Stack Trace: java.lang.NullPointerException: null at org.apache.hadoop.io.serializer.SerializationFactory.<init>(SerializationFactory.java:58) at org.apache.hadoop.mapreduce.split.JobSplitWriter.writeNewSplits(JobSplitWriter.java:122) at org.apache.hadoop.mapreduce.split.JobSplitWriter.createSplitFiles(JobSplitWriter.java:79) at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:308) at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:319) at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:196) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667) at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287) at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1308) at org.apache.hadoop.mapreduce.TestLocalRunner.doMultiReducerTest(TestLocalRunner.java:546) at org.apache.hadoop.mapreduce.TestLocalRunner.testMultiMapMultiReduce(TestLocalRunner.java:569) FAILED: org.apache.hadoop.mapreduce.TestLocalRunner.testOneMapMultiParallelReduce Error Message: null Stack Trace: java.lang.NullPointerException: null at org.apache.hadoop.io.serializer.SerializationFactory.<init>(SerializationFactory.java:58) at org.apache.hadoop.mapreduce.split.JobSplitWriter.writeNewSplits(JobSplitWriter.java:122) at org.apache.hadoop.mapreduce.split.JobSplitWriter.createSplitFiles(JobSplitWriter.java:79) at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:308) at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:319) at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:196) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667) at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287) at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1308) at org.apache.hadoop.mapreduce.TestLocalRunner.doMultiReducerTest(TestLocalRunner.java:546) at org.apache.hadoop.mapreduce.TestLocalRunner.testOneMapMultiParallelReduce(TestLocalRunner.java:559) FAILED: org.apache.hadoop.mapreduce.TestLocalRunner.testMultiMapOneReduce Error Message: null Stack Trace: java.lang.NullPointerException: null at org.apache.hadoop.io.serializer.SerializationFactory.<init>(SerializationFactory.java:58) at org.apache.hadoop.mapreduce.split.JobSplitWriter.writeNewSplits(JobSplitWriter.java:122) at org.apache.hadoop.mapreduce.split.JobSplitWriter.createSplitFiles(JobSplitWriter.java:79) at org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:308) at org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:319) at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:196) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667) at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287) at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1308) at org.apache.hadoop.mapreduce.TestLocalRunner.doMultiReducerTest(TestLocalRunner.java:546) at org.apache.hadoop.mapreduce.TestLocalRunner.testMultiMapOneReduce(TestLocalRunner.java:564) FAILED: org.apache.hadoop.mapreduce.TestMapCollection.testRandomCompress Error Message: org.xml.sax.SAXParseException; systemId: jar:file:/home/jenkins/.m2/repository/org/apache/hadoop/hadoop-hdfs/3.0.0-SNAPSHOT/hadoop-hdfs-3.0.0-SNAPSHOT.jar!/hdfs-default.xml; lineNumber: 2451; columnNumber: 55; The element type "name" must be terminated by the matching end-tag "</name>". Stack Trace: java.lang.RuntimeException: org.xml.sax.SAXParseException; systemId: jar:file:/home/jenkins/.m2/repository/org/apache/hadoop/hadoop-hdfs/3.0.0-SNAPSHOT/hadoop-hdfs-3.0.0-SNAPSHOT.jar!/hdfs-default.xml; lineNumber: 2451; columnNumber: 55; The element type "name" must be terminated by the matching end-tag "</name>". at org.apache.xerces.parsers.DOMParser.parse(Unknown Source) at org.apache.xerces.jaxp.DocumentBuilderImpl.parse(Unknown Source) at javax.xml.parsers.DocumentBuilder.parse(DocumentBuilder.java:150) at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2546) at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2534) at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2605) at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2558) at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2469) at org.apache.hadoop.conf.Configuration.get(Configuration.java:1043) at org.apache.hadoop.conf.Configuration.getTrimmed(Configuration.java:1093) at org.apache.hadoop.conf.Configuration.getClass(Configuration.java:2253) at org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:159) at org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:326) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667) at org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:326) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:443) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:420) at org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:406) at org.apache.hadoop.mapred.LocalDistributedCacheManager.setup(LocalDistributedCacheManager.java:116) at org.apache.hadoop.mapred.LocalJobRunner$Job.<init>(LocalJobRunner.java:172) at org.apache.hadoop.mapred.LocalJobRunner.submitJob(LocalJobRunner.java:786) at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:241) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290) at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667) at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287) at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1308) at org.apache.hadoop.mapreduce.TestMapCollection.runTest(TestMapCollection.java:328) at org.apache.hadoop.mapreduce.TestMapCollection.testRandomCompress(TestMapCollection.java:537) FAILED: org.apache.hadoop.mapreduce.TestMapCollection.testSpillPer2B Error Message: org.xml.sax.SAXParseException; systemId: jar:file:/home/jenkins/.m2/repository/org/apache/hadoop/hadoop-hdfs/3.0.0-SNAPSHOT/hadoop-hdfs-3.0.0-SNAPSHOT.jar!/hdfs-default.xml; lineNumber: 2451; columnNumber: 55; The element type "name" must be terminated by the matching end-tag "</name>". Stack Trace: java.lang.RuntimeException: org.xml.sax.SAXParseException; systemId: jar:file:/home/jenkins/.m2/repository/org/apache/hadoop/hadoop-hdfs/3.0.0-SNAPSHOT/hadoop-hdfs-3.0.0-SNAPSHOT.jar!/hdfs-default.xml; lineNumber: 2451; columnNumber: 55; The element type "name" must be terminated by the matching end-tag "</name>". at org.apache.xerces.parsers.DOMParser.parse(Unknown Source) at org.apache.xerces.jaxp.DocumentBuilderImpl.parse(Unknown Source) at javax.xml.parsers.DocumentBuilder.parse(DocumentBuilder.java:150) at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2546) at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2534) at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2605) at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2558) at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2469) at org.apache.hadoop.conf.Configuration.set(Configuration.java:1205) at org.apache.hadoop.conf.Configuration.set(Configuration.java:1177) at org.apache.hadoop.conf.Configuration.setInt(Configuration.java:1349) at org.apache.hadoop.mapreduce.TestMapCollection.runTest(TestMapCollection.java:301) at org.apache.hadoop.mapreduce.TestMapCollection.testSpillPer2B(TestMapCollection.java:348) FAILED: org.apache.hadoop.mapreduce.TestMapCollection.testValLastByte Error Message: org.xml.sax.SAXParseException; systemId: jar:file:/home/jenkins/.m2/repository/org/apache/hadoop/hadoop-hdfs/3.0.0-SNAPSHOT/hadoop-hdfs-3.0.0-SNAPSHOT.jar!/hdfs-default.xml; lineNumber: 2451; columnNumber: 55; The element type "name" must be terminated by the matching end-tag "</name>". Stack Trace: java.lang.RuntimeException: org.xml.sax.SAXParseException; systemId: jar:file:/home/jenkins/.m2/repository/org/apache/hadoop/hadoop-hdfs/3.0.0-SNAPSHOT/hadoop-hdfs-3.0.0-SNAPSHOT.jar!/hdfs-default.xml; lineNumber: 2451; columnNumber: 55; The element type "name" must be terminated by the matching end-tag "</name>". at org.apache.xerces.parsers.DOMParser.parse(Unknown Source) at org.apache.xerces.jaxp.DocumentBuilderImpl.parse(Unknown Source) at javax.xml.parsers.DocumentBuilder.parse(DocumentBuilder.java:150) at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2546) at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2534) at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2605) at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2558) at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2469) at org.apache.hadoop.conf.Configuration.set(Configuration.java:1205) at org.apache.hadoop.conf.Configuration.set(Configuration.java:1177) at org.apache.hadoop.conf.Configuration.setInt(Configuration.java:1349) at org.apache.hadoop.mapreduce.TestMapCollection.runTest(TestMapCollection.java:301) at org.apache.hadoop.mapreduce.TestMapCollection.testValLastByte(TestMapCollection.java:334) FAILED: org.apache.hadoop.mapreduce.TestMapCollection.testRandom Error Message: org.xml.sax.SAXParseException; systemId: jar:file:/home/jenkins/.m2/repository/org/apache/hadoop/hadoop-hdfs/3.0.0-SNAPSHOT/hadoop-hdfs-3.0.0-SNAPSHOT.jar!/hdfs-default.xml; lineNumber: 2451; columnNumber: 55; The element type "name" must be terminated by the matching end-tag "</name>". Stack Trace: java.lang.RuntimeException: org.xml.sax.SAXParseException; systemId: jar:file:/home/jenkins/.m2/repository/org/apache/hadoop/hadoop-hdfs/3.0.0-SNAPSHOT/hadoop-hdfs-3.0.0-SNAPSHOT.jar!/hdfs-default.xml; lineNumber: 2451; columnNumber: 55; The element type "name" must be terminated by the matching end-tag "</name>". at org.apache.xerces.parsers.DOMParser.parse(Unknown Source) at org.apache.xerces.jaxp.DocumentBuilderImpl.parse(Unknown Source) at javax.xml.parsers.DocumentBuilder.parse(DocumentBuilder.java:150) at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2546) at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2534) at org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2605) at org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2558) at org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2469) at org.apache.hadoop.conf.Configuration.set(Configuration.java:1205) at org.apache.hadoop.conf.Configuration.set(Configuration.java:1177) at org.apache.hadoop.conf.Configuration.setInt(Configuration.java:1349) at org.apache.hadoop.mapreduce.TestMapCollection.testRandom(TestMapCollection.java:500) FAILED: org.apache.hadoop.mapreduce.security.TestBinaryTokenFile.org.apache.hadoop.mapreduce.security.TestBinaryTokenFile Error Message: java.io.IOException: ResourceManager failed to start. Final state is STOPPED Stack Trace: org.apache.hadoop.yarn.exceptions.YarnRuntimeException: java.io.IOException: ResourceManager failed to start. Final state is STOPPED at org.apache.hadoop.yarn.server.MiniYARNCluster.startResourceManager(MiniYARNCluster.java:329) at org.apache.hadoop.yarn.server.MiniYARNCluster.access$500(MiniYARNCluster.java:98) at org.apache.hadoop.yarn.server.MiniYARNCluster$ResourceManagerWrapper.serviceStart(MiniYARNCluster.java:455) at org.apache.hadoop.service.AbstractService.start(AbstractService.java:193) at org.apache.hadoop.service.CompositeService.serviceStart(CompositeService.java:120) at org.apache.hadoop.service.AbstractService.start(AbstractService.java:193) at org.apache.hadoop.mapreduce.security.TestBinaryTokenFile.setUp(TestBinaryTokenFile.java:193) FAILED: org.apache.hadoop.mapreduce.v2.TestMRAppWithCombiner.testCombinerShouldUpdateTheReporter Error Message: org/apache/hadoop/yarn/exceptions/ApplicationIdNotProvidedException Stack Trace: java.lang.NoClassDefFoundError: org/apache/hadoop/yarn/exceptions/ApplicationIdNotProvidedException at java.net.URLClassLoader$1.run(URLClassLoader.java:366) at java.net.URLClassLoader$1.run(URLClassLoader.java:355) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:354) at java.lang.ClassLoader.loadClass(ClassLoader.java:425) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) at java.lang.ClassLoader.loadClass(ClassLoader.java:358) at org.apache.hadoop.yarn.client.api.YarnClient.createYarnClient(YarnClient.java:79) at org.apache.hadoop.mapred.ResourceMgrDelegate.<init>(ResourceMgrDelegate.java:97) at org.apache.hadoop.mapred.YARNRunner.<init>(YARNRunner.java:122) at org.apache.hadoop.mapred.YarnClientProtocolProvider.create(YarnClientProtocolProvider.java:34) at org.apache.hadoop.mapreduce.Cluster.initialize(Cluster.java:95) at org.apache.hadoop.mapreduce.Cluster.<init>(Cluster.java:82) at org.apache.hadoop.mapreduce.Cluster.<init>(Cluster.java:75) at org.apache.hadoop.mapred.JobClient.init(JobClient.java:475) at org.apache.hadoop.mapred.JobClient.<init>(JobClient.java:454) at org.apache.hadoop.mapreduce.v2.TestMRAppWithCombiner.runJob(TestMRAppWithCombiner.java:146) at org.apache.hadoop.mapreduce.v2.TestMRAppWithCombiner.testCombinerShouldUpdateTheReporter(TestMRAppWithCombiner.java:124) FAILED: org.apache.hadoop.mapreduce.v2.TestSpeculativeExecution.testSpeculativeExecution Error Message: expected:<3> but was:<2> Stack Trace: java.lang.AssertionError: expected:<3> but was:<2> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:555) at org.junit.Assert.assertEquals(Assert.java:542) at org.apache.hadoop.mapreduce.v2.TestSpeculativeExecution.testSpeculativeExecution(TestSpeculativeExecution.java:233)