See https://builds.apache.org/job/Hadoop-Mapreduce-trunk/2472/

###################################################################################
########################## LAST 60 LINES OF THE CONSOLE 
###########################
[...truncated 33380 lines...]
  TestRMContainerAllocator.testReportedAppProgressWithOnlyMaps:1118 » 
NullPointer
  TestRMContainerAllocator.testUpdatedNodes:1213 » NullPointer
  TestRMContainerAllocator.testSimple:231 » NullPointer
  TestRMContainerAllocator.testIgnoreBlacklisting:1393->getContainerOnHost:1511 
» NullPointer
  
TestRMContainerAllocator.testRMContainerAllocatorResendsRequestsOnRMRestart:2509
 » NullPointer
  TestRecovery.testMultipleCrashes:792 » YarnRuntime 
org.apache.hadoop.fs.Unsupp...
  TestRecovery.testRecoveryFailsUsingCustomOutputCommitter:572 » YarnRuntime 
cou...
  TestRecovery.testCrashOfMapsOnlyJob:327 » YarnRuntime could not cleanup test 
d...
  TestRecovery.testOutputRecovery:821 » YarnRuntime could not cleanup test dir
  TestRecovery.testOutputRecoveryMapsOnly:930 » YarnRuntime could not cleanup 
te...
  TestRecovery.testRecoverySuccessUsingCustomOutputCommitter:459 » YarnRuntime 
c...
  TestRecovery.testCrashed:130 » YarnRuntime could not cleanup test dir
  TestRecovery.testSpeculative:1164 » YarnRuntime could not cleanup test dir
  TestRecovery.testRecoveryWithoutShuffleSecret:1316 » YarnRuntime could not 
cle...
  TestRecovery.testRecoveryWithOldCommiter:1045 » YarnRuntime could not cleanup 
...

Tests run: 339, Failures: 21, Errors: 19, Skipped: 0

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] Apache Hadoop MapReduce Client .................... SUCCESS [  2.942 s]
[INFO] Apache Hadoop MapReduce Core ...................... SUCCESS [01:36 min]
[INFO] Apache Hadoop MapReduce Common .................... SUCCESS [ 29.307 s]
[INFO] Apache Hadoop MapReduce Shuffle ................... SUCCESS [  4.859 s]
[INFO] Apache Hadoop MapReduce App ....................... FAILURE [07:58 min]
[INFO] Apache Hadoop MapReduce HistoryServer ............. SKIPPED
[INFO] Apache Hadoop MapReduce JobClient ................. SKIPPED
[INFO] Apache Hadoop MapReduce HistoryServer Plugins ..... SKIPPED
[INFO] Apache Hadoop MapReduce NativeTask ................ SKIPPED
[INFO] Apache Hadoop MapReduce Examples .................. SKIPPED
[INFO] Apache Hadoop MapReduce ........................... SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 10:13 min
[INFO] Finished at: 2015-10-13T23:55:50+00:00
[INFO] Final Memory: 39M/825M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal 
org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on 
project hadoop-mapreduce-client-app: There are test failures.
[ERROR] 
[ERROR] Please refer to 
/home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/target/surefire-reports
 for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please 
read the following articles:
[ERROR] [Help 1] 
http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hadoop-mapreduce-client-app
Build step 'Execute shell' marked build as failure
[FINDBUGS] Skipping publisher since build result is FAILURE
Archiving artifacts
Recording test results
Updating HDFS-9167
Email was triggered for: Failure - Any
Sending email for trigger: Failure - Any



###################################################################################
############################## FAILED TESTS (if any) 
##############################
40 tests failed.
FAILED:  org.apache.hadoop.mapreduce.v2.app.TestRecovery.testMultipleCrashes

Error Message:
org.apache.hadoop.fs.UnsupportedFileSystemException: 
fs.AbstractFileSystem.file.impl=null: No AbstractFileSystem configured for 
scheme: file

Stack Trace:
org.apache.hadoop.yarn.exceptions.YarnRuntimeException: 
org.apache.hadoop.fs.UnsupportedFileSystemException: 
fs.AbstractFileSystem.file.impl=null: No AbstractFileSystem configured for 
scheme: file
        at 
org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161)
        at 
org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:332)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667)
        at 
org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:329)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:446)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:472)
        at 
org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler.serviceInit(JobHistoryEventHandler.java:165)
        at 
org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
        at 
org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107)
        at 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster.serviceInit(MRAppMaster.java:482)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.serviceInit(MRApp.java:267)
        at 
org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.submit(MRApp.java:295)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.submit(MRApp.java:280)
        at 
org.apache.hadoop.mapreduce.v2.app.TestRecovery.testMultipleCrashes(TestRecovery.java:792)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.TestRecovery.testRecoveryFailsUsingCustomOutputCommitter

Error Message:
could not cleanup test dir

Stack Trace:
org.apache.hadoop.yarn.exceptions.YarnRuntimeException: could not cleanup test 
dir
        at 
org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161)
        at 
org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:332)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667)
        at 
org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:329)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:446)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:423)
        at 
org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:409)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:239)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:208)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:189)
        at 
org.apache.hadoop.mapreduce.v2.app.TestRecovery$MRAppWithHistory.<init>(TestRecovery.java:1928)
        at 
org.apache.hadoop.mapreduce.v2.app.TestRecovery.testRecoveryFailsUsingCustomOutputCommitter(TestRecovery.java:572)


FAILED:  org.apache.hadoop.mapreduce.v2.app.TestRecovery.testCrashOfMapsOnlyJob

Error Message:
could not cleanup test dir

Stack Trace:
org.apache.hadoop.yarn.exceptions.YarnRuntimeException: could not cleanup test 
dir
        at 
org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161)
        at 
org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:332)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667)
        at 
org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:329)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:446)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:423)
        at 
org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:409)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:239)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:208)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:189)
        at 
org.apache.hadoop.mapreduce.v2.app.TestRecovery$MRAppWithHistory.<init>(TestRecovery.java:1928)
        at 
org.apache.hadoop.mapreduce.v2.app.TestRecovery.testCrashOfMapsOnlyJob(TestRecovery.java:327)


FAILED:  org.apache.hadoop.mapreduce.v2.app.TestRecovery.testOutputRecovery

Error Message:
could not cleanup test dir

Stack Trace:
org.apache.hadoop.yarn.exceptions.YarnRuntimeException: could not cleanup test 
dir
        at 
org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161)
        at 
org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:332)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667)
        at 
org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:329)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:446)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:423)
        at 
org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:409)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:239)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:208)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:189)
        at 
org.apache.hadoop.mapreduce.v2.app.TestRecovery$MRAppWithHistory.<init>(TestRecovery.java:1928)
        at 
org.apache.hadoop.mapreduce.v2.app.TestRecovery.testOutputRecovery(TestRecovery.java:821)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.TestRecovery.testOutputRecoveryMapsOnly

Error Message:
could not cleanup test dir

Stack Trace:
org.apache.hadoop.yarn.exceptions.YarnRuntimeException: could not cleanup test 
dir
        at 
org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161)
        at 
org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:332)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667)
        at 
org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:329)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:446)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:423)
        at 
org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:409)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:239)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:208)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:189)
        at 
org.apache.hadoop.mapreduce.v2.app.TestRecovery$MRAppWithHistory.<init>(TestRecovery.java:1928)
        at 
org.apache.hadoop.mapreduce.v2.app.TestRecovery.testOutputRecoveryMapsOnly(TestRecovery.java:930)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.TestRecovery.testRecoverySuccessUsingCustomOutputCommitter

Error Message:
could not cleanup test dir

Stack Trace:
org.apache.hadoop.yarn.exceptions.YarnRuntimeException: could not cleanup test 
dir
        at 
org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161)
        at 
org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:332)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667)
        at 
org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:329)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:446)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:423)
        at 
org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:409)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:239)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:208)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:189)
        at 
org.apache.hadoop.mapreduce.v2.app.TestRecovery$MRAppWithHistory.<init>(TestRecovery.java:1928)
        at 
org.apache.hadoop.mapreduce.v2.app.TestRecovery.testRecoverySuccessUsingCustomOutputCommitter(TestRecovery.java:459)


FAILED:  org.apache.hadoop.mapreduce.v2.app.TestRecovery.testCrashed

Error Message:
could not cleanup test dir

Stack Trace:
org.apache.hadoop.yarn.exceptions.YarnRuntimeException: could not cleanup test 
dir
        at 
org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161)
        at 
org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:332)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667)
        at 
org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:329)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:446)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:423)
        at 
org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:409)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:239)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:208)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:189)
        at 
org.apache.hadoop.mapreduce.v2.app.TestRecovery$MRAppWithHistory.<init>(TestRecovery.java:1928)
        at 
org.apache.hadoop.mapreduce.v2.app.TestRecovery.testCrashed(TestRecovery.java:130)


FAILED:  org.apache.hadoop.mapreduce.v2.app.TestRecovery.testSpeculative

Error Message:
could not cleanup test dir

Stack Trace:
org.apache.hadoop.yarn.exceptions.YarnRuntimeException: could not cleanup test 
dir
        at 
org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161)
        at 
org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:332)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667)
        at 
org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:329)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:446)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:423)
        at 
org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:409)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:239)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:208)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:189)
        at 
org.apache.hadoop.mapreduce.v2.app.TestRecovery$MRAppWithHistory.<init>(TestRecovery.java:1928)
        at 
org.apache.hadoop.mapreduce.v2.app.TestRecovery.testSpeculative(TestRecovery.java:1164)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.TestRecovery.testRecoveryWithoutShuffleSecret

Error Message:
could not cleanup test dir

Stack Trace:
org.apache.hadoop.yarn.exceptions.YarnRuntimeException: could not cleanup test 
dir
        at 
org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161)
        at 
org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:332)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667)
        at 
org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:329)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:446)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:423)
        at 
org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:409)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:239)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:208)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:189)
        at 
org.apache.hadoop.mapreduce.v2.app.TestRecovery$MRAppWithHistory.<init>(TestRecovery.java:1928)
        at 
org.apache.hadoop.mapreduce.v2.app.TestRecovery$MRAppNoShuffleSecret.<init>(TestRecovery.java:1960)
        at 
org.apache.hadoop.mapreduce.v2.app.TestRecovery.testRecoveryWithoutShuffleSecret(TestRecovery.java:1316)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.TestRecovery.testRecoveryWithOldCommiter

Error Message:
could not cleanup test dir

Stack Trace:
org.apache.hadoop.yarn.exceptions.YarnRuntimeException: could not cleanup test 
dir
        at 
org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161)
        at 
org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:332)
        at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329)
        at java.security.AccessController.doPrivileged(Native Method)
        at javax.security.auth.Subject.doAs(Subject.java:415)
        at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667)
        at 
org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:329)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:446)
        at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:423)
        at 
org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:409)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:239)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:208)
        at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:189)
        at 
org.apache.hadoop.mapreduce.v2.app.TestRecovery$MRAppWithHistory.<init>(TestRecovery.java:1928)
        at 
org.apache.hadoop.mapreduce.v2.app.TestRecovery.testRecoveryWithOldCommiter(TestRecovery.java:1045)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobCountersXML

Error Message:
expected:<application/xml> but was:<text/html; charset=ISO-8859-1>

Stack Trace:
java.lang.AssertionError: expected:<application/xml> but was:<text/html; 
charset=ISO-8859-1>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobCountersXML(TestAMWebServicesJobs.java:688)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobAttemptsXML

Error Message:
expected:<application/xml> but was:<text/html; charset=ISO-8859-1>

Stack Trace:
java.lang.AssertionError: expected:<application/xml> but was:<text/html; 
charset=ISO-8859-1>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobAttemptsXML(TestAMWebServicesJobs.java:841)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobsXML

Error Message:
expected:<application/xml> but was:<text/html; charset=ISO-8859-1>

Stack Trace:
java.lang.AssertionError: expected:<application/xml> but was:<text/html; 
charset=ISO-8859-1>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobsXML(TestAMWebServicesJobs.java:179)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobsDefault

Error Message:
expected:<application/json> but was:<text/html; charset=ISO-8859-1>

Stack Trace:
java.lang.AssertionError: expected:<application/json> but was:<text/html; 
charset=ISO-8859-1>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobsDefault(TestAMWebServicesJobs.java:162)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobsSlash

Error Message:
expected:<application/json> but was:<text/html; charset=ISO-8859-1>

Stack Trace:
java.lang.AssertionError: expected:<application/json> but was:<text/html; 
charset=ISO-8859-1>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobsSlash(TestAMWebServicesJobs.java:146)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobId

Error Message:
expected:<application/json> but was:<text/html; charset=ISO-8859-1>

Stack Trace:
java.lang.AssertionError: expected:<application/json> but was:<text/html; 
charset=ISO-8859-1>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobId(TestAMWebServicesJobs.java:204)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobs

Error Message:
expected:<application/json> but was:<text/html; charset=ISO-8859-1>

Stack Trace:
java.lang.AssertionError: expected:<application/json> but was:<text/html; 
charset=ISO-8859-1>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobs(TestAMWebServicesJobs.java:129)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdXML

Error Message:
expected:<application/xml> but was:<text/html; charset=ISO-8859-1>

Stack Trace:
java.lang.AssertionError: expected:<application/xml> but was:<text/html; 
charset=ISO-8859-1>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdXML(TestAMWebServicesJobs.java:401)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdInvalidXML

Error Message:
expected:<Not Found> but was:<Internal Server Error>

Stack Trace:
java.lang.AssertionError: expected:<Not Found> but was:<Internal Server Error>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdInvalidXML(TestAMWebServicesJobs.java:332)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdInvalidBogus

Error Message:
expected:<Not Found> but was:<Internal Server Error>

Stack Trace:
java.lang.AssertionError: expected:<Not Found> but was:<Internal Server Error>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdInvalidBogus(TestAMWebServicesJobs.java:371)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobAttemptsSlash

Error Message:
expected:<application/json> but was:<text/html; charset=ISO-8859-1>

Stack Trace:
java.lang.AssertionError: expected:<application/json> but was:<text/html; 
charset=ISO-8859-1>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobAttemptsSlash(TestAMWebServicesJobs.java:805)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdSlash

Error Message:
expected:<application/json> but was:<text/html; charset=ISO-8859-1>

Stack Trace:
java.lang.AssertionError: expected:<application/json> but was:<text/html; 
charset=ISO-8859-1>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdSlash(TestAMWebServicesJobs.java:223)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobCountersSlash

Error Message:
expected:<application/json> but was:<text/html; charset=ISO-8859-1>

Stack Trace:
java.lang.AssertionError: expected:<application/json> but was:<text/html; 
charset=ISO-8859-1>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobCountersSlash(TestAMWebServicesJobs.java:653)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdDefault

Error Message:
expected:<application/json> but was:<text/html; charset=ISO-8859-1>

Stack Trace:
java.lang.AssertionError: expected:<application/json> but was:<text/html; 
charset=ISO-8859-1>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdDefault(TestAMWebServicesJobs.java:240)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobCounters

Error Message:
expected:<application/json> but was:<text/html; charset=ISO-8859-1>

Stack Trace:
java.lang.AssertionError: expected:<application/json> but was:<text/html; 
charset=ISO-8859-1>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobCounters(TestAMWebServicesJobs.java:635)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdInvalid

Error Message:
expected:<Not Found> but was:<Internal Server Error>

Stack Trace:
java.lang.AssertionError: expected:<Not Found> but was:<Internal Server Error>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdInvalid(TestAMWebServicesJobs.java:286)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobAttempts

Error Message:
expected:<application/json> but was:<text/html; charset=ISO-8859-1>

Stack Trace:
java.lang.AssertionError: expected:<application/json> but was:<text/html; 
charset=ISO-8859-1>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobAttempts(TestAMWebServicesJobs.java:787)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdInvalidDefault

Error Message:
expected:<Not Found> but was:<Internal Server Error>

Stack Trace:
java.lang.AssertionError: expected:<Not Found> but was:<Internal Server Error>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdInvalidDefault(TestAMWebServicesJobs.java:309)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobCountersDefault

Error Message:
expected:<application/json> but was:<text/html; charset=ISO-8859-1>

Stack Trace:
java.lang.AssertionError: expected:<application/json> but was:<text/html; 
charset=ISO-8859-1>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobCountersDefault(TestAMWebServicesJobs.java:670)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdNonExist

Error Message:
expected:<Not Found> but was:<Internal Server Error>

Stack Trace:
java.lang.AssertionError: expected:<Not Found> but was:<Internal Server Error>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdNonExist(TestAMWebServicesJobs.java:259)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobAttemptsDefault

Error Message:
expected:<application/json> but was:<text/html; charset=ISO-8859-1>

Stack Trace:
java.lang.AssertionError: expected:<application/json> but was:<text/html; 
charset=ISO-8859-1>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:144)
        at 
org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobAttemptsDefault(TestAMWebServicesJobs.java:823)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testReportedAppProgress

Error Message:
null

Stack Trace:
java.lang.NullPointerException: null
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.isRequestLabelChanged(AppSchedulingInfo.java:420)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.updateResourceRequests(AppSchedulingInfo.java:341)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplicationAttempt.updateResourceRequests(SchedulerApplicationAttempt.java:300)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoScheduler.allocate(FifoScheduler.java:350)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyFifoScheduler.allocate(TestRMContainerAllocator.java:1719)
        at 
org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:506)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor.makeRemoteRequest(RMContainerRequestor.java:204)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.makeRemoteRequest(TestRMContainerAllocator.java:2060)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.getResources(RMContainerAllocator.java:724)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.heartbeat(RMContainerAllocator.java:268)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.schedule(TestRMContainerAllocator.java:2023)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testReportedAppProgress(TestRMContainerAllocator.java:970)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testBlackListedNodesWithSchedulingToThatNode

Error Message:
null

Stack Trace:
java.lang.NullPointerException: null
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.isRequestLabelChanged(AppSchedulingInfo.java:420)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.updateResourceRequests(AppSchedulingInfo.java:341)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplicationAttempt.updateResourceRequests(SchedulerApplicationAttempt.java:300)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoScheduler.allocate(FifoScheduler.java:350)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyFifoScheduler.allocate(TestRMContainerAllocator.java:1719)
        at 
org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:506)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor.makeRemoteRequest(RMContainerRequestor.java:204)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.makeRemoteRequest(TestRMContainerAllocator.java:2060)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.getResources(RMContainerAllocator.java:724)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.heartbeat(RMContainerAllocator.java:268)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.schedule(TestRMContainerAllocator.java:2023)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testBlackListedNodesWithSchedulingToThatNode(TestRMContainerAllocator.java:1605)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testBlackListedNodes

Error Message:
null

Stack Trace:
java.lang.NullPointerException: null
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.isRequestLabelChanged(AppSchedulingInfo.java:420)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.updateResourceRequests(AppSchedulingInfo.java:341)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplicationAttempt.updateResourceRequests(SchedulerApplicationAttempt.java:300)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoScheduler.allocate(FifoScheduler.java:350)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyFifoScheduler.allocate(TestRMContainerAllocator.java:1719)
        at 
org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:506)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor.makeRemoteRequest(RMContainerRequestor.java:204)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.makeRemoteRequest(TestRMContainerAllocator.java:2060)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.getResources(RMContainerAllocator.java:724)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.heartbeat(RMContainerAllocator.java:268)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.schedule(TestRMContainerAllocator.java:2023)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testBlackListedNodes(TestRMContainerAllocator.java:1318)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testReducerRampdownDiagnostics

Error Message:
null

Stack Trace:
java.lang.NullPointerException: null
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.isRequestLabelChanged(AppSchedulingInfo.java:420)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.updateResourceRequests(AppSchedulingInfo.java:341)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplicationAttempt.updateResourceRequests(SchedulerApplicationAttempt.java:300)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoScheduler.allocate(FifoScheduler.java:350)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyFifoScheduler.allocate(TestRMContainerAllocator.java:1719)
        at 
org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:506)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor.makeRemoteRequest(RMContainerRequestor.java:204)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.makeRemoteRequest(TestRMContainerAllocator.java:2060)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.getResources(RMContainerAllocator.java:724)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.heartbeat(RMContainerAllocator.java:268)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.schedule(TestRMContainerAllocator.java:2023)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testReducerRampdownDiagnostics(TestRMContainerAllocator.java:450)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testReportedAppProgressWithOnlyMaps

Error Message:
null

Stack Trace:
java.lang.NullPointerException: null
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.isRequestLabelChanged(AppSchedulingInfo.java:420)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.updateResourceRequests(AppSchedulingInfo.java:341)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplicationAttempt.updateResourceRequests(SchedulerApplicationAttempt.java:300)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoScheduler.allocate(FifoScheduler.java:350)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyFifoScheduler.allocate(TestRMContainerAllocator.java:1719)
        at 
org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:506)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor.makeRemoteRequest(RMContainerRequestor.java:204)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.makeRemoteRequest(TestRMContainerAllocator.java:2060)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.getResources(RMContainerAllocator.java:724)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.heartbeat(RMContainerAllocator.java:268)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.schedule(TestRMContainerAllocator.java:2023)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testReportedAppProgressWithOnlyMaps(TestRMContainerAllocator.java:1118)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testUpdatedNodes

Error Message:
null

Stack Trace:
java.lang.NullPointerException: null
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.isRequestLabelChanged(AppSchedulingInfo.java:420)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.updateResourceRequests(AppSchedulingInfo.java:341)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplicationAttempt.updateResourceRequests(SchedulerApplicationAttempt.java:300)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoScheduler.allocate(FifoScheduler.java:350)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyFifoScheduler.allocate(TestRMContainerAllocator.java:1719)
        at 
org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:506)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor.makeRemoteRequest(RMContainerRequestor.java:204)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.makeRemoteRequest(TestRMContainerAllocator.java:2060)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.getResources(RMContainerAllocator.java:724)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.heartbeat(RMContainerAllocator.java:268)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.schedule(TestRMContainerAllocator.java:2023)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testUpdatedNodes(TestRMContainerAllocator.java:1213)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testSimple

Error Message:
null

Stack Trace:
java.lang.NullPointerException: null
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.isRequestLabelChanged(AppSchedulingInfo.java:420)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.updateResourceRequests(AppSchedulingInfo.java:341)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplicationAttempt.updateResourceRequests(SchedulerApplicationAttempt.java:300)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoScheduler.allocate(FifoScheduler.java:350)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyFifoScheduler.allocate(TestRMContainerAllocator.java:1719)
        at 
org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:506)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor.makeRemoteRequest(RMContainerRequestor.java:204)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.makeRemoteRequest(TestRMContainerAllocator.java:2060)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.getResources(RMContainerAllocator.java:724)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.heartbeat(RMContainerAllocator.java:268)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.schedule(TestRMContainerAllocator.java:2023)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testSimple(TestRMContainerAllocator.java:231)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testIgnoreBlacklisting

Error Message:
null

Stack Trace:
java.lang.NullPointerException: null
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.isRequestLabelChanged(AppSchedulingInfo.java:420)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.updateResourceRequests(AppSchedulingInfo.java:341)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplicationAttempt.updateResourceRequests(SchedulerApplicationAttempt.java:300)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoScheduler.allocate(FifoScheduler.java:350)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyFifoScheduler.allocate(TestRMContainerAllocator.java:1719)
        at 
org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:506)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor.makeRemoteRequest(RMContainerRequestor.java:204)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.makeRemoteRequest(TestRMContainerAllocator.java:2060)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.getResources(RMContainerAllocator.java:724)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.heartbeat(RMContainerAllocator.java:268)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.schedule(TestRMContainerAllocator.java:2023)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.getContainerOnHost(TestRMContainerAllocator.java:1511)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testIgnoreBlacklisting(TestRMContainerAllocator.java:1393)


FAILED:  
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testRMContainerAllocatorResendsRequestsOnRMRestart

Error Message:
null

Stack Trace:
java.lang.NullPointerException: null
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.isRequestLabelChanged(AppSchedulingInfo.java:420)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.updateResourceRequests(AppSchedulingInfo.java:341)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplicationAttempt.updateResourceRequests(SchedulerApplicationAttempt.java:300)
        at 
org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoScheduler.allocate(FifoScheduler.java:350)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyFifoScheduler.allocate(TestRMContainerAllocator.java:1719)
        at 
org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:506)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor.makeRemoteRequest(RMContainerRequestor.java:204)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.makeRemoteRequest(TestRMContainerAllocator.java:2060)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.getResources(RMContainerAllocator.java:724)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.heartbeat(RMContainerAllocator.java:268)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.schedule(TestRMContainerAllocator.java:2023)
        at 
org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testRMContainerAllocatorResendsRequestsOnRMRestart(TestRMContainerAllocator.java:2509)


Reply via email to