See https://builds.apache.org/job/Hadoop-Mapreduce-trunk/2337/

###################################################################################
########################## LAST 60 LINES OF THE CONSOLE 
###########################
[...truncated 33512 lines...]
  TestLocalRunner.testMultiMaps:320 » NullPointer
  TestLocalRunner.testMultiMapMultiReduce:569->doMultiReducerTest:546 » 
NullPointer
  TestLocalRunner.testOneMapMultiParallelReduce:559->doMultiReducerTest:546 » 
NullPointer
  TestLocalRunner.testMultiMapOneReduce:564->doMultiReducerTest:546 » 
NullPointer
  TestMapCollection.testRandomCompress:537->runTest:328 » Runtime 
org.xml.sax.SA...
  TestMapCollection.testSpillPer2B:348->runTest:301 » Runtime 
org.xml.sax.SAXPar...
  TestMapCollection.testValLastByte:334->runTest:301 » Runtime 
org.xml.sax.SAXPa...
  TestMapCollection.testRandom:500 » Runtime org.xml.sax.SAXParseException; 
syst...
  TestMRAppWithCombiner.testCombinerShouldUpdateTheReporter:124->runJob:146 » 
NoClassDefFound

Tests run: 387, Failures: 2, Errors: 14, Skipped: 9

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] Apache Hadoop MapReduce Client .................... SUCCESS [  3.076 s]
[INFO] Apache Hadoop MapReduce Core ...................... SUCCESS [01:35 min]
[INFO] Apache Hadoop MapReduce Common .................... SUCCESS [ 29.449 s]
[INFO] Apache Hadoop MapReduce Shuffle ................... SUCCESS [  5.106 s]
[INFO] Apache Hadoop MapReduce App ....................... SUCCESS [08:59 min]
[INFO] Apache Hadoop MapReduce HistoryServer ............. SUCCESS [05:37 min]
[INFO] Apache Hadoop MapReduce JobClient ................. FAILURE [  01:11 h]
[INFO] Apache Hadoop MapReduce HistoryServer Plugins ..... SKIPPED
[INFO] Apache Hadoop MapReduce NativeTask ................ SKIPPED
[INFO] Apache Hadoop MapReduce Examples .................. SKIPPED
[INFO] Apache Hadoop MapReduce ........................... SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 01:28 h
[INFO] Finished at: 2015-09-15T20:35:34+00:00
[INFO] Final Memory: 44M/737M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal 
org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on 
project hadoop-mapreduce-client-jobclient: ExecutionException: 
java.lang.RuntimeException: The forked VM terminated without properly saying 
goodbye. VM crash or System.exit called?
[ERROR] Command was /bin/sh -c cd 
/home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient
 && /home/jenkins/tools/java/jdk1.7.0_55/jre/bin/java -Xmx4096m 
-XX:MaxPermSize=768m -XX:+HeapDumpOnOutOfMemoryError -jar 
/home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/target/surefire/surefirebooter2874754138681265931.jar
 
/home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/target/surefire/surefire4151895669068123060tmp
 
/home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/target/surefire/surefire_1504602903375118102902tmp
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please 
read the following articles:
[ERROR] [Help 1] 
http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hadoop-mapreduce-client-jobclient
Build step 'Execute shell' marked build as failure
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Sending artifact delta relative to Hadoop-Mapreduce-trunk #2330
Archived 1 artifacts
Archive block size is 32768
Received 0 blocks and 20467504 bytes
Compression is 0.0%
Took 5.3 sec
Recording test results
Updating HADOOP-12413
Updating HDFS-9008
Email was triggered for: Failure
Sending email for trigger: Failure



###################################################################################
############################## FAILED TESTS (if any) 
##############################
16 tests failed.
FAILED:  org.apache.hadoop.mapred.TestTextInputFormat.testSplitableCodecs

Error Message:
Key in multiple partitions.

Stack Trace:
java.lang.AssertionError: Key in multiple partitions.
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.assertTrue(Assert.java:41)
        at org.junit.Assert.assertFalse(Assert.java:64)
        at 
org.apache.hadoop.mapred.TestTextInputFormat.testSplitableCodecs(TestTextInputFormat.java:223)


FAILED:  org.apache.hadoop.mapreduce.TestLocalRunner.testGcCounter

Error Message:
org/apache/hadoop/io/FastByteComparisons

Stack Trace:
java.lang.NoClassDefFoundError: org/apache/hadoop/io/FastByteComparisons
        at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
        at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
        at java.security.AccessController.doPrivileged(Native Method)
        at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
        at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
        at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
        at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
        at 
org.apache.hadoop.io.WritableComparator.compareBytes(WritableComparator.java:188)
        at 
org.apache.hadoop.io.BinaryComparable.compareTo(BinaryComparable.java:50)
        at 
org.apache.hadoop.io.BinaryComparable.equals(BinaryComparable.java:72)
        at org.apache.hadoop.io.Text.equals(Text.java:348)
        at org.apache.hadoop.mapreduce.JobID.equals(JobID.java:89)
        at java.util.HashMap.getEntry(HashMap.java:471)
        at java.util.HashMap.get(HashMap.java:421)
        at 
org.apache.hadoop.mapred.LocalJobRunner.getJobCounters(LocalJobRunner.java:826)
        at org.apache.hadoop.mapreduce.Job$7.run(Job.java:761)
        at org.apache.hadoop.mapreduce.Job$7.run(Job.java:758)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667)
        at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:758)
        at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1383)
        at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1311)
        at 
org.apache.hadoop.mapreduce.TestLocalRunner.testGcCounter(TestLocalRunner.java:271)


FAILED:  
org.apache.hadoop.mapreduce.TestLocalRunner.testInvalidMultiMapParallelism

Error Message:
null

Stack Trace:
java.lang.NullPointerException: null
        at 
org.apache.hadoop.io.serializer.SerializationFactory.<init>(SerializationFactory.java:58)
        at 
org.apache.hadoop.mapreduce.split.JobSplitWriter.writeNewSplits(JobSplitWriter.java:122)
        at 
org.apache.hadoop.mapreduce.split.JobSplitWriter.createSplitFiles(JobSplitWriter.java:79)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:308)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:319)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:196)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667)
        at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287)
        at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1308)
        at 
org.apache.hadoop.mapreduce.TestLocalRunner.testInvalidMultiMapParallelism(TestLocalRunner.java:372)


FAILED:  org.apache.hadoop.mapreduce.TestLocalRunner.testOneMapMultiReduce

Error Message:
null

Stack Trace:
java.lang.NullPointerException: null
        at 
org.apache.hadoop.io.serializer.SerializationFactory.<init>(SerializationFactory.java:58)
        at 
org.apache.hadoop.mapreduce.split.JobSplitWriter.writeNewSplits(JobSplitWriter.java:122)
        at 
org.apache.hadoop.mapreduce.split.JobSplitWriter.createSplitFiles(JobSplitWriter.java:79)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:308)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:319)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:196)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667)
        at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287)
        at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1308)
        at 
org.apache.hadoop.mapreduce.TestLocalRunner.doMultiReducerTest(TestLocalRunner.java:546)
        at 
org.apache.hadoop.mapreduce.TestLocalRunner.testOneMapMultiReduce(TestLocalRunner.java:554)


FAILED:  org.apache.hadoop.mapreduce.TestLocalRunner.testEmptyMaps

Error Message:
fs.AbstractFileSystem.file.impl=null: No AbstractFileSystem configured for 
scheme: file

Stack Trace:
org.apache.hadoop.fs.UnsupportedFileSystemException: 
fs.AbstractFileSystem.file.impl=null: No AbstractFileSystem configured for 
scheme: file
        at 
org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161)
        at 
org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:326)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667)
        at 
org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:326)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:443)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:420)
        at 
org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:406)
        at 
org.apache.hadoop.mapred.LocalDistributedCacheManager.setup(LocalDistributedCacheManager.java:116)
        at 
org.apache.hadoop.mapred.LocalJobRunner$Job.<init>(LocalJobRunner.java:172)
        at 
org.apache.hadoop.mapred.LocalJobRunner.submitJob(LocalJobRunner.java:786)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:241)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667)
        at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287)
        at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1308)
        at 
org.apache.hadoop.mapreduce.TestLocalRunner.testEmptyMaps(TestLocalRunner.java:429)


FAILED:  org.apache.hadoop.mapreduce.TestLocalRunner.testMultiMaps

Error Message:
null

Stack Trace:
java.lang.NullPointerException: null
        at 
org.apache.hadoop.io.serializer.SerializationFactory.<init>(SerializationFactory.java:58)
        at 
org.apache.hadoop.mapreduce.split.JobSplitWriter.writeNewSplits(JobSplitWriter.java:122)
        at 
org.apache.hadoop.mapreduce.split.JobSplitWriter.createSplitFiles(JobSplitWriter.java:79)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:308)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:319)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:196)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667)
        at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287)
        at 
org.apache.hadoop.mapreduce.TestLocalRunner.testMultiMaps(TestLocalRunner.java:320)


FAILED:  org.apache.hadoop.mapreduce.TestLocalRunner.testMultiMapMultiReduce

Error Message:
null

Stack Trace:
java.lang.NullPointerException: null
        at 
org.apache.hadoop.io.serializer.SerializationFactory.<init>(SerializationFactory.java:58)
        at 
org.apache.hadoop.mapreduce.split.JobSplitWriter.writeNewSplits(JobSplitWriter.java:122)
        at 
org.apache.hadoop.mapreduce.split.JobSplitWriter.createSplitFiles(JobSplitWriter.java:79)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:308)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:319)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:196)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667)
        at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287)
        at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1308)
        at 
org.apache.hadoop.mapreduce.TestLocalRunner.doMultiReducerTest(TestLocalRunner.java:546)
        at 
org.apache.hadoop.mapreduce.TestLocalRunner.testMultiMapMultiReduce(TestLocalRunner.java:569)


FAILED:  
org.apache.hadoop.mapreduce.TestLocalRunner.testOneMapMultiParallelReduce

Error Message:
null

Stack Trace:
java.lang.NullPointerException: null
        at 
org.apache.hadoop.io.serializer.SerializationFactory.<init>(SerializationFactory.java:58)
        at 
org.apache.hadoop.mapreduce.split.JobSplitWriter.writeNewSplits(JobSplitWriter.java:122)
        at 
org.apache.hadoop.mapreduce.split.JobSplitWriter.createSplitFiles(JobSplitWriter.java:79)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:308)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:319)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:196)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667)
        at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287)
        at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1308)
        at 
org.apache.hadoop.mapreduce.TestLocalRunner.doMultiReducerTest(TestLocalRunner.java:546)
        at 
org.apache.hadoop.mapreduce.TestLocalRunner.testOneMapMultiParallelReduce(TestLocalRunner.java:559)


FAILED:  org.apache.hadoop.mapreduce.TestLocalRunner.testMultiMapOneReduce

Error Message:
null

Stack Trace:
java.lang.NullPointerException: null
        at 
org.apache.hadoop.io.serializer.SerializationFactory.<init>(SerializationFactory.java:58)
        at 
org.apache.hadoop.mapreduce.split.JobSplitWriter.writeNewSplits(JobSplitWriter.java:122)
        at 
org.apache.hadoop.mapreduce.split.JobSplitWriter.createSplitFiles(JobSplitWriter.java:79)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:308)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:319)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:196)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667)
        at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287)
        at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1308)
        at 
org.apache.hadoop.mapreduce.TestLocalRunner.doMultiReducerTest(TestLocalRunner.java:546)
        at 
org.apache.hadoop.mapreduce.TestLocalRunner.testMultiMapOneReduce(TestLocalRunner.java:564)


FAILED:  org.apache.hadoop.mapreduce.TestMapCollection.testRandomCompress

Error Message:
org.xml.sax.SAXParseException; systemId: 
jar:file:/home/jenkins/.m2/repository/org/apache/hadoop/hadoop-hdfs/3.0.0-SNAPSHOT/hadoop-hdfs-3.0.0-SNAPSHOT.jar!/hdfs-default.xml;
 lineNumber: 2451; columnNumber: 55; The element type "name" must be terminated 
by the matching end-tag "</name>".

Stack Trace:
java.lang.RuntimeException: org.xml.sax.SAXParseException; systemId: 
jar:file:/home/jenkins/.m2/repository/org/apache/hadoop/hadoop-hdfs/3.0.0-SNAPSHOT/hadoop-hdfs-3.0.0-SNAPSHOT.jar!/hdfs-default.xml;
 lineNumber: 2451; columnNumber: 55; The element type "name" must be terminated 
by the matching end-tag "</name>".
        at org.apache.xerces.parsers.DOMParser.parse(Unknown Source)
        at org.apache.xerces.jaxp.DocumentBuilderImpl.parse(Unknown Source)
        at javax.xml.parsers.DocumentBuilder.parse(DocumentBuilder.java:150)
        at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2546)
        at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2534)
        at 
org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2605)
        at 
org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2558)
        at 
org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2469)
        at org.apache.hadoop.conf.Configuration.get(Configuration.java:1043)
        at 
org.apache.hadoop.conf.Configuration.getTrimmed(Configuration.java:1093)
        at 
org.apache.hadoop.conf.Configuration.getClass(Configuration.java:2253)
        at 
org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:159)
        at 
org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:326)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667)
        at 
org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:326)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:443)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:420)
        at 
org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:406)
        at 
org.apache.hadoop.mapred.LocalDistributedCacheManager.setup(LocalDistributedCacheManager.java:116)
        at 
org.apache.hadoop.mapred.LocalJobRunner$Job.<init>(LocalJobRunner.java:172)
        at 
org.apache.hadoop.mapred.LocalJobRunner.submitJob(LocalJobRunner.java:786)
        at 
org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:241)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290)
        at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667)
        at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287)
        at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1308)
        at 
org.apache.hadoop.mapreduce.TestMapCollection.runTest(TestMapCollection.java:328)
        at 
org.apache.hadoop.mapreduce.TestMapCollection.testRandomCompress(TestMapCollection.java:537)


FAILED:  org.apache.hadoop.mapreduce.TestMapCollection.testSpillPer2B

Error Message:
org.xml.sax.SAXParseException; systemId: 
jar:file:/home/jenkins/.m2/repository/org/apache/hadoop/hadoop-hdfs/3.0.0-SNAPSHOT/hadoop-hdfs-3.0.0-SNAPSHOT.jar!/hdfs-default.xml;
 lineNumber: 2451; columnNumber: 55; The element type "name" must be terminated 
by the matching end-tag "</name>".

Stack Trace:
java.lang.RuntimeException: org.xml.sax.SAXParseException; systemId: 
jar:file:/home/jenkins/.m2/repository/org/apache/hadoop/hadoop-hdfs/3.0.0-SNAPSHOT/hadoop-hdfs-3.0.0-SNAPSHOT.jar!/hdfs-default.xml;
 lineNumber: 2451; columnNumber: 55; The element type "name" must be terminated 
by the matching end-tag "</name>".
        at org.apache.xerces.parsers.DOMParser.parse(Unknown Source)
        at org.apache.xerces.jaxp.DocumentBuilderImpl.parse(Unknown Source)
        at javax.xml.parsers.DocumentBuilder.parse(DocumentBuilder.java:150)
        at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2546)
        at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2534)
        at 
org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2605)
        at 
org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2558)
        at 
org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2469)
        at org.apache.hadoop.conf.Configuration.set(Configuration.java:1205)
        at org.apache.hadoop.conf.Configuration.set(Configuration.java:1177)
        at org.apache.hadoop.conf.Configuration.setInt(Configuration.java:1349)
        at 
org.apache.hadoop.mapreduce.TestMapCollection.runTest(TestMapCollection.java:301)
        at 
org.apache.hadoop.mapreduce.TestMapCollection.testSpillPer2B(TestMapCollection.java:348)


FAILED:  org.apache.hadoop.mapreduce.TestMapCollection.testValLastByte

Error Message:
org.xml.sax.SAXParseException; systemId: 
jar:file:/home/jenkins/.m2/repository/org/apache/hadoop/hadoop-hdfs/3.0.0-SNAPSHOT/hadoop-hdfs-3.0.0-SNAPSHOT.jar!/hdfs-default.xml;
 lineNumber: 2451; columnNumber: 55; The element type "name" must be terminated 
by the matching end-tag "</name>".

Stack Trace:
java.lang.RuntimeException: org.xml.sax.SAXParseException; systemId: 
jar:file:/home/jenkins/.m2/repository/org/apache/hadoop/hadoop-hdfs/3.0.0-SNAPSHOT/hadoop-hdfs-3.0.0-SNAPSHOT.jar!/hdfs-default.xml;
 lineNumber: 2451; columnNumber: 55; The element type "name" must be terminated 
by the matching end-tag "</name>".
        at org.apache.xerces.parsers.DOMParser.parse(Unknown Source)
        at org.apache.xerces.jaxp.DocumentBuilderImpl.parse(Unknown Source)
        at javax.xml.parsers.DocumentBuilder.parse(DocumentBuilder.java:150)
        at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2546)
        at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2534)
        at 
org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2605)
        at 
org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2558)
        at 
org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2469)
        at org.apache.hadoop.conf.Configuration.set(Configuration.java:1205)
        at org.apache.hadoop.conf.Configuration.set(Configuration.java:1177)
        at org.apache.hadoop.conf.Configuration.setInt(Configuration.java:1349)
        at 
org.apache.hadoop.mapreduce.TestMapCollection.runTest(TestMapCollection.java:301)
        at 
org.apache.hadoop.mapreduce.TestMapCollection.testValLastByte(TestMapCollection.java:334)


FAILED:  org.apache.hadoop.mapreduce.TestMapCollection.testRandom

Error Message:
org.xml.sax.SAXParseException; systemId: 
jar:file:/home/jenkins/.m2/repository/org/apache/hadoop/hadoop-hdfs/3.0.0-SNAPSHOT/hadoop-hdfs-3.0.0-SNAPSHOT.jar!/hdfs-default.xml;
 lineNumber: 2451; columnNumber: 55; The element type "name" must be terminated 
by the matching end-tag "</name>".

Stack Trace:
java.lang.RuntimeException: org.xml.sax.SAXParseException; systemId: 
jar:file:/home/jenkins/.m2/repository/org/apache/hadoop/hadoop-hdfs/3.0.0-SNAPSHOT/hadoop-hdfs-3.0.0-SNAPSHOT.jar!/hdfs-default.xml;
 lineNumber: 2451; columnNumber: 55; The element type "name" must be terminated 
by the matching end-tag "</name>".
        at org.apache.xerces.parsers.DOMParser.parse(Unknown Source)
        at org.apache.xerces.jaxp.DocumentBuilderImpl.parse(Unknown Source)
        at javax.xml.parsers.DocumentBuilder.parse(DocumentBuilder.java:150)
        at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2546)
        at org.apache.hadoop.conf.Configuration.parse(Configuration.java:2534)
        at 
org.apache.hadoop.conf.Configuration.loadResource(Configuration.java:2605)
        at 
org.apache.hadoop.conf.Configuration.loadResources(Configuration.java:2558)
        at 
org.apache.hadoop.conf.Configuration.getProps(Configuration.java:2469)
        at org.apache.hadoop.conf.Configuration.set(Configuration.java:1205)
        at org.apache.hadoop.conf.Configuration.set(Configuration.java:1177)
        at org.apache.hadoop.conf.Configuration.setInt(Configuration.java:1349)
        at 
org.apache.hadoop.mapreduce.TestMapCollection.testRandom(TestMapCollection.java:500)


FAILED:  
org.apache.hadoop.mapreduce.security.TestBinaryTokenFile.org.apache.hadoop.mapreduce.security.TestBinaryTokenFile

Error Message:
java.io.IOException: ResourceManager failed to start. Final state is STOPPED

Stack Trace:
org.apache.hadoop.yarn.exceptions.YarnRuntimeException: java.io.IOException: 
ResourceManager failed to start. Final state is STOPPED
        at 
org.apache.hadoop.yarn.server.MiniYARNCluster.startResourceManager(MiniYARNCluster.java:329)
        at 
org.apache.hadoop.yarn.server.MiniYARNCluster.access$500(MiniYARNCluster.java:98)
        at 
org.apache.hadoop.yarn.server.MiniYARNCluster$ResourceManagerWrapper.serviceStart(MiniYARNCluster.java:455)
        at 
org.apache.hadoop.service.AbstractService.start(AbstractService.java:193)
        at 
org.apache.hadoop.service.CompositeService.serviceStart(CompositeService.java:120)
        at 
org.apache.hadoop.service.AbstractService.start(AbstractService.java:193)
        at 
org.apache.hadoop.mapreduce.security.TestBinaryTokenFile.setUp(TestBinaryTokenFile.java:193)


FAILED:  
org.apache.hadoop.mapreduce.v2.TestMRAppWithCombiner.testCombinerShouldUpdateTheReporter

Error Message:
org/apache/hadoop/yarn/exceptions/ApplicationIdNotProvidedException

Stack Trace:
java.lang.NoClassDefFoundError: 
org/apache/hadoop/yarn/exceptions/ApplicationIdNotProvidedException
        at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
        at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
        at java.security.AccessController.doPrivileged(Native Method)
        at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
        at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
        at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
        at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
        at 
org.apache.hadoop.yarn.client.api.YarnClient.createYarnClient(YarnClient.java:79)
        at 
org.apache.hadoop.mapred.ResourceMgrDelegate.<init>(ResourceMgrDelegate.java:97)
        at org.apache.hadoop.mapred.YARNRunner.<init>(YARNRunner.java:122)
        at 
org.apache.hadoop.mapred.YarnClientProtocolProvider.create(YarnClientProtocolProvider.java:34)
        at org.apache.hadoop.mapreduce.Cluster.initialize(Cluster.java:95)
        at org.apache.hadoop.mapreduce.Cluster.<init>(Cluster.java:82)
        at org.apache.hadoop.mapreduce.Cluster.<init>(Cluster.java:75)
        at org.apache.hadoop.mapred.JobClient.init(JobClient.java:475)
        at org.apache.hadoop.mapred.JobClient.<init>(JobClient.java:454)
        at 
org.apache.hadoop.mapreduce.v2.TestMRAppWithCombiner.runJob(TestMRAppWithCombiner.java:146)
        at 
org.apache.hadoop.mapreduce.v2.TestMRAppWithCombiner.testCombinerShouldUpdateTheReporter(TestMRAppWithCombiner.java:124)


FAILED:  
org.apache.hadoop.mapreduce.v2.TestSpeculativeExecution.testSpeculativeExecution

Error Message:
expected:<3> but was:<2>

Stack Trace:
java.lang.AssertionError: expected:<3> but was:<2>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:555)
        at org.junit.Assert.assertEquals(Assert.java:542)
        at 
org.apache.hadoop.mapreduce.v2.TestSpeculativeExecution.testSpeculativeExecution(TestSpeculativeExecution.java:233)


Reply via email to