See https://builds.apache.org/job/Hadoop-Mapreduce-trunk/2472/
################################################################################### ########################## LAST 60 LINES OF THE CONSOLE ########################### [...truncated 33380 lines...] TestRMContainerAllocator.testReportedAppProgressWithOnlyMaps:1118 » NullPointer TestRMContainerAllocator.testUpdatedNodes:1213 » NullPointer TestRMContainerAllocator.testSimple:231 » NullPointer TestRMContainerAllocator.testIgnoreBlacklisting:1393->getContainerOnHost:1511 » NullPointer TestRMContainerAllocator.testRMContainerAllocatorResendsRequestsOnRMRestart:2509 » NullPointer TestRecovery.testMultipleCrashes:792 » YarnRuntime org.apache.hadoop.fs.Unsupp... TestRecovery.testRecoveryFailsUsingCustomOutputCommitter:572 » YarnRuntime cou... TestRecovery.testCrashOfMapsOnlyJob:327 » YarnRuntime could not cleanup test d... TestRecovery.testOutputRecovery:821 » YarnRuntime could not cleanup test dir TestRecovery.testOutputRecoveryMapsOnly:930 » YarnRuntime could not cleanup te... TestRecovery.testRecoverySuccessUsingCustomOutputCommitter:459 » YarnRuntime c... TestRecovery.testCrashed:130 » YarnRuntime could not cleanup test dir TestRecovery.testSpeculative:1164 » YarnRuntime could not cleanup test dir TestRecovery.testRecoveryWithoutShuffleSecret:1316 » YarnRuntime could not cle... TestRecovery.testRecoveryWithOldCommiter:1045 » YarnRuntime could not cleanup ... Tests run: 339, Failures: 21, Errors: 19, Skipped: 0 [INFO] ------------------------------------------------------------------------ [INFO] Reactor Summary: [INFO] [INFO] Apache Hadoop MapReduce Client .................... SUCCESS [ 2.942 s] [INFO] Apache Hadoop MapReduce Core ...................... SUCCESS [01:36 min] [INFO] Apache Hadoop MapReduce Common .................... SUCCESS [ 29.307 s] [INFO] Apache Hadoop MapReduce Shuffle ................... SUCCESS [ 4.859 s] [INFO] Apache Hadoop MapReduce App ....................... FAILURE [07:58 min] [INFO] Apache Hadoop MapReduce HistoryServer ............. SKIPPED [INFO] Apache Hadoop MapReduce JobClient ................. SKIPPED [INFO] Apache Hadoop MapReduce HistoryServer Plugins ..... SKIPPED [INFO] Apache Hadoop MapReduce NativeTask ................ SKIPPED [INFO] Apache Hadoop MapReduce Examples .................. SKIPPED [INFO] Apache Hadoop MapReduce ........................... SKIPPED [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 10:13 min [INFO] Finished at: 2015-10-13T23:55:50+00:00 [INFO] Final Memory: 39M/825M [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.apache.maven.plugins:maven-surefire-plugin:2.17:test (default-test) on project hadoop-mapreduce-client-app: There are test failures. [ERROR] [ERROR] Please refer to /home/jenkins/jenkins-slave/workspace/Hadoop-Mapreduce-trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app/target/surefire-reports for the individual test results. [ERROR] -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException [ERROR] [ERROR] After correcting the problems, you can resume the build with the command [ERROR] mvn <goals> -rf :hadoop-mapreduce-client-app Build step 'Execute shell' marked build as failure [FINDBUGS] Skipping publisher since build result is FAILURE Archiving artifacts Recording test results Updating HDFS-9167 Email was triggered for: Failure - Any Sending email for trigger: Failure - Any ################################################################################### ############################## FAILED TESTS (if any) ############################## 40 tests failed. FAILED: org.apache.hadoop.mapreduce.v2.app.TestRecovery.testMultipleCrashes Error Message: org.apache.hadoop.fs.UnsupportedFileSystemException: fs.AbstractFileSystem.file.impl=null: No AbstractFileSystem configured for scheme: file Stack Trace: org.apache.hadoop.yarn.exceptions.YarnRuntimeException: org.apache.hadoop.fs.UnsupportedFileSystemException: fs.AbstractFileSystem.file.impl=null: No AbstractFileSystem configured for scheme: file at org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161) at org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:332) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667) at org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:329) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:446) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:472) at org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler.serviceInit(JobHistoryEventHandler.java:165) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107) at org.apache.hadoop.mapreduce.v2.app.MRAppMaster.serviceInit(MRAppMaster.java:482) at org.apache.hadoop.mapreduce.v2.app.MRApp.serviceInit(MRApp.java:267) at org.apache.hadoop.service.AbstractService.init(AbstractService.java:163) at org.apache.hadoop.mapreduce.v2.app.MRApp.submit(MRApp.java:295) at org.apache.hadoop.mapreduce.v2.app.MRApp.submit(MRApp.java:280) at org.apache.hadoop.mapreduce.v2.app.TestRecovery.testMultipleCrashes(TestRecovery.java:792) FAILED: org.apache.hadoop.mapreduce.v2.app.TestRecovery.testRecoveryFailsUsingCustomOutputCommitter Error Message: could not cleanup test dir Stack Trace: org.apache.hadoop.yarn.exceptions.YarnRuntimeException: could not cleanup test dir at org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161) at org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:332) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667) at org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:329) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:446) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:423) at org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:409) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:239) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:208) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:189) at org.apache.hadoop.mapreduce.v2.app.TestRecovery$MRAppWithHistory.<init>(TestRecovery.java:1928) at org.apache.hadoop.mapreduce.v2.app.TestRecovery.testRecoveryFailsUsingCustomOutputCommitter(TestRecovery.java:572) FAILED: org.apache.hadoop.mapreduce.v2.app.TestRecovery.testCrashOfMapsOnlyJob Error Message: could not cleanup test dir Stack Trace: org.apache.hadoop.yarn.exceptions.YarnRuntimeException: could not cleanup test dir at org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161) at org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:332) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667) at org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:329) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:446) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:423) at org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:409) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:239) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:208) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:189) at org.apache.hadoop.mapreduce.v2.app.TestRecovery$MRAppWithHistory.<init>(TestRecovery.java:1928) at org.apache.hadoop.mapreduce.v2.app.TestRecovery.testCrashOfMapsOnlyJob(TestRecovery.java:327) FAILED: org.apache.hadoop.mapreduce.v2.app.TestRecovery.testOutputRecovery Error Message: could not cleanup test dir Stack Trace: org.apache.hadoop.yarn.exceptions.YarnRuntimeException: could not cleanup test dir at org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161) at org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:332) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667) at org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:329) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:446) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:423) at org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:409) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:239) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:208) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:189) at org.apache.hadoop.mapreduce.v2.app.TestRecovery$MRAppWithHistory.<init>(TestRecovery.java:1928) at org.apache.hadoop.mapreduce.v2.app.TestRecovery.testOutputRecovery(TestRecovery.java:821) FAILED: org.apache.hadoop.mapreduce.v2.app.TestRecovery.testOutputRecoveryMapsOnly Error Message: could not cleanup test dir Stack Trace: org.apache.hadoop.yarn.exceptions.YarnRuntimeException: could not cleanup test dir at org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161) at org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:332) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667) at org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:329) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:446) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:423) at org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:409) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:239) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:208) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:189) at org.apache.hadoop.mapreduce.v2.app.TestRecovery$MRAppWithHistory.<init>(TestRecovery.java:1928) at org.apache.hadoop.mapreduce.v2.app.TestRecovery.testOutputRecoveryMapsOnly(TestRecovery.java:930) FAILED: org.apache.hadoop.mapreduce.v2.app.TestRecovery.testRecoverySuccessUsingCustomOutputCommitter Error Message: could not cleanup test dir Stack Trace: org.apache.hadoop.yarn.exceptions.YarnRuntimeException: could not cleanup test dir at org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161) at org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:332) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667) at org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:329) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:446) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:423) at org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:409) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:239) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:208) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:189) at org.apache.hadoop.mapreduce.v2.app.TestRecovery$MRAppWithHistory.<init>(TestRecovery.java:1928) at org.apache.hadoop.mapreduce.v2.app.TestRecovery.testRecoverySuccessUsingCustomOutputCommitter(TestRecovery.java:459) FAILED: org.apache.hadoop.mapreduce.v2.app.TestRecovery.testCrashed Error Message: could not cleanup test dir Stack Trace: org.apache.hadoop.yarn.exceptions.YarnRuntimeException: could not cleanup test dir at org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161) at org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:332) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667) at org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:329) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:446) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:423) at org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:409) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:239) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:208) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:189) at org.apache.hadoop.mapreduce.v2.app.TestRecovery$MRAppWithHistory.<init>(TestRecovery.java:1928) at org.apache.hadoop.mapreduce.v2.app.TestRecovery.testCrashed(TestRecovery.java:130) FAILED: org.apache.hadoop.mapreduce.v2.app.TestRecovery.testSpeculative Error Message: could not cleanup test dir Stack Trace: org.apache.hadoop.yarn.exceptions.YarnRuntimeException: could not cleanup test dir at org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161) at org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:332) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667) at org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:329) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:446) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:423) at org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:409) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:239) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:208) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:189) at org.apache.hadoop.mapreduce.v2.app.TestRecovery$MRAppWithHistory.<init>(TestRecovery.java:1928) at org.apache.hadoop.mapreduce.v2.app.TestRecovery.testSpeculative(TestRecovery.java:1164) FAILED: org.apache.hadoop.mapreduce.v2.app.TestRecovery.testRecoveryWithoutShuffleSecret Error Message: could not cleanup test dir Stack Trace: org.apache.hadoop.yarn.exceptions.YarnRuntimeException: could not cleanup test dir at org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161) at org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:332) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667) at org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:329) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:446) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:423) at org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:409) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:239) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:208) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:189) at org.apache.hadoop.mapreduce.v2.app.TestRecovery$MRAppWithHistory.<init>(TestRecovery.java:1928) at org.apache.hadoop.mapreduce.v2.app.TestRecovery$MRAppNoShuffleSecret.<init>(TestRecovery.java:1960) at org.apache.hadoop.mapreduce.v2.app.TestRecovery.testRecoveryWithoutShuffleSecret(TestRecovery.java:1316) FAILED: org.apache.hadoop.mapreduce.v2.app.TestRecovery.testRecoveryWithOldCommiter Error Message: could not cleanup test dir Stack Trace: org.apache.hadoop.yarn.exceptions.YarnRuntimeException: could not cleanup test dir at org.apache.hadoop.fs.AbstractFileSystem.createFileSystem(AbstractFileSystem.java:161) at org.apache.hadoop.fs.AbstractFileSystem.get(AbstractFileSystem.java:250) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:332) at org.apache.hadoop.fs.FileContext$2.run(FileContext.java:329) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:415) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1667) at org.apache.hadoop.fs.FileContext.getAbstractFileSystem(FileContext.java:329) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:446) at org.apache.hadoop.fs.FileContext.getFileContext(FileContext.java:423) at org.apache.hadoop.fs.FileContext.getLocalFSFileContext(FileContext.java:409) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:239) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:208) at org.apache.hadoop.mapreduce.v2.app.MRApp.<init>(MRApp.java:189) at org.apache.hadoop.mapreduce.v2.app.TestRecovery$MRAppWithHistory.<init>(TestRecovery.java:1928) at org.apache.hadoop.mapreduce.v2.app.TestRecovery.testRecoveryWithOldCommiter(TestRecovery.java:1045) FAILED: org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobCountersXML Error Message: expected:<application/xml> but was:<text/html; charset=ISO-8859-1> Stack Trace: java.lang.AssertionError: expected:<application/xml> but was:<text/html; charset=ISO-8859-1> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobCountersXML(TestAMWebServicesJobs.java:688) FAILED: org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobAttemptsXML Error Message: expected:<application/xml> but was:<text/html; charset=ISO-8859-1> Stack Trace: java.lang.AssertionError: expected:<application/xml> but was:<text/html; charset=ISO-8859-1> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobAttemptsXML(TestAMWebServicesJobs.java:841) FAILED: org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobsXML Error Message: expected:<application/xml> but was:<text/html; charset=ISO-8859-1> Stack Trace: java.lang.AssertionError: expected:<application/xml> but was:<text/html; charset=ISO-8859-1> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobsXML(TestAMWebServicesJobs.java:179) FAILED: org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobsDefault Error Message: expected:<application/json> but was:<text/html; charset=ISO-8859-1> Stack Trace: java.lang.AssertionError: expected:<application/json> but was:<text/html; charset=ISO-8859-1> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobsDefault(TestAMWebServicesJobs.java:162) FAILED: org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobsSlash Error Message: expected:<application/json> but was:<text/html; charset=ISO-8859-1> Stack Trace: java.lang.AssertionError: expected:<application/json> but was:<text/html; charset=ISO-8859-1> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobsSlash(TestAMWebServicesJobs.java:146) FAILED: org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobId Error Message: expected:<application/json> but was:<text/html; charset=ISO-8859-1> Stack Trace: java.lang.AssertionError: expected:<application/json> but was:<text/html; charset=ISO-8859-1> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobId(TestAMWebServicesJobs.java:204) FAILED: org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobs Error Message: expected:<application/json> but was:<text/html; charset=ISO-8859-1> Stack Trace: java.lang.AssertionError: expected:<application/json> but was:<text/html; charset=ISO-8859-1> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobs(TestAMWebServicesJobs.java:129) FAILED: org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdXML Error Message: expected:<application/xml> but was:<text/html; charset=ISO-8859-1> Stack Trace: java.lang.AssertionError: expected:<application/xml> but was:<text/html; charset=ISO-8859-1> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdXML(TestAMWebServicesJobs.java:401) FAILED: org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdInvalidXML Error Message: expected:<Not Found> but was:<Internal Server Error> Stack Trace: java.lang.AssertionError: expected:<Not Found> but was:<Internal Server Error> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdInvalidXML(TestAMWebServicesJobs.java:332) FAILED: org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdInvalidBogus Error Message: expected:<Not Found> but was:<Internal Server Error> Stack Trace: java.lang.AssertionError: expected:<Not Found> but was:<Internal Server Error> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdInvalidBogus(TestAMWebServicesJobs.java:371) FAILED: org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobAttemptsSlash Error Message: expected:<application/json> but was:<text/html; charset=ISO-8859-1> Stack Trace: java.lang.AssertionError: expected:<application/json> but was:<text/html; charset=ISO-8859-1> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobAttemptsSlash(TestAMWebServicesJobs.java:805) FAILED: org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdSlash Error Message: expected:<application/json> but was:<text/html; charset=ISO-8859-1> Stack Trace: java.lang.AssertionError: expected:<application/json> but was:<text/html; charset=ISO-8859-1> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdSlash(TestAMWebServicesJobs.java:223) FAILED: org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobCountersSlash Error Message: expected:<application/json> but was:<text/html; charset=ISO-8859-1> Stack Trace: java.lang.AssertionError: expected:<application/json> but was:<text/html; charset=ISO-8859-1> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobCountersSlash(TestAMWebServicesJobs.java:653) FAILED: org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdDefault Error Message: expected:<application/json> but was:<text/html; charset=ISO-8859-1> Stack Trace: java.lang.AssertionError: expected:<application/json> but was:<text/html; charset=ISO-8859-1> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdDefault(TestAMWebServicesJobs.java:240) FAILED: org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobCounters Error Message: expected:<application/json> but was:<text/html; charset=ISO-8859-1> Stack Trace: java.lang.AssertionError: expected:<application/json> but was:<text/html; charset=ISO-8859-1> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobCounters(TestAMWebServicesJobs.java:635) FAILED: org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdInvalid Error Message: expected:<Not Found> but was:<Internal Server Error> Stack Trace: java.lang.AssertionError: expected:<Not Found> but was:<Internal Server Error> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdInvalid(TestAMWebServicesJobs.java:286) FAILED: org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobAttempts Error Message: expected:<application/json> but was:<text/html; charset=ISO-8859-1> Stack Trace: java.lang.AssertionError: expected:<application/json> but was:<text/html; charset=ISO-8859-1> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobAttempts(TestAMWebServicesJobs.java:787) FAILED: org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdInvalidDefault Error Message: expected:<Not Found> but was:<Internal Server Error> Stack Trace: java.lang.AssertionError: expected:<Not Found> but was:<Internal Server Error> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdInvalidDefault(TestAMWebServicesJobs.java:309) FAILED: org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobCountersDefault Error Message: expected:<application/json> but was:<text/html; charset=ISO-8859-1> Stack Trace: java.lang.AssertionError: expected:<application/json> but was:<text/html; charset=ISO-8859-1> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobCountersDefault(TestAMWebServicesJobs.java:670) FAILED: org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdNonExist Error Message: expected:<Not Found> but was:<Internal Server Error> Stack Trace: java.lang.AssertionError: expected:<Not Found> but was:<Internal Server Error> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobIdNonExist(TestAMWebServicesJobs.java:259) FAILED: org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobAttemptsDefault Error Message: expected:<application/json> but was:<text/html; charset=ISO-8859-1> Stack Trace: java.lang.AssertionError: expected:<application/json> but was:<text/html; charset=ISO-8859-1> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:144) at org.apache.hadoop.mapreduce.v2.app.webapp.TestAMWebServicesJobs.testJobAttemptsDefault(TestAMWebServicesJobs.java:823) FAILED: org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testReportedAppProgress Error Message: null Stack Trace: java.lang.NullPointerException: null at org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.isRequestLabelChanged(AppSchedulingInfo.java:420) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.updateResourceRequests(AppSchedulingInfo.java:341) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplicationAttempt.updateResourceRequests(SchedulerApplicationAttempt.java:300) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoScheduler.allocate(FifoScheduler.java:350) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyFifoScheduler.allocate(TestRMContainerAllocator.java:1719) at org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:506) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor.makeRemoteRequest(RMContainerRequestor.java:204) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.makeRemoteRequest(TestRMContainerAllocator.java:2060) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.getResources(RMContainerAllocator.java:724) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.heartbeat(RMContainerAllocator.java:268) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.schedule(TestRMContainerAllocator.java:2023) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testReportedAppProgress(TestRMContainerAllocator.java:970) FAILED: org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testBlackListedNodesWithSchedulingToThatNode Error Message: null Stack Trace: java.lang.NullPointerException: null at org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.isRequestLabelChanged(AppSchedulingInfo.java:420) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.updateResourceRequests(AppSchedulingInfo.java:341) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplicationAttempt.updateResourceRequests(SchedulerApplicationAttempt.java:300) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoScheduler.allocate(FifoScheduler.java:350) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyFifoScheduler.allocate(TestRMContainerAllocator.java:1719) at org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:506) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor.makeRemoteRequest(RMContainerRequestor.java:204) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.makeRemoteRequest(TestRMContainerAllocator.java:2060) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.getResources(RMContainerAllocator.java:724) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.heartbeat(RMContainerAllocator.java:268) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.schedule(TestRMContainerAllocator.java:2023) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testBlackListedNodesWithSchedulingToThatNode(TestRMContainerAllocator.java:1605) FAILED: org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testBlackListedNodes Error Message: null Stack Trace: java.lang.NullPointerException: null at org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.isRequestLabelChanged(AppSchedulingInfo.java:420) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.updateResourceRequests(AppSchedulingInfo.java:341) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplicationAttempt.updateResourceRequests(SchedulerApplicationAttempt.java:300) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoScheduler.allocate(FifoScheduler.java:350) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyFifoScheduler.allocate(TestRMContainerAllocator.java:1719) at org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:506) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor.makeRemoteRequest(RMContainerRequestor.java:204) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.makeRemoteRequest(TestRMContainerAllocator.java:2060) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.getResources(RMContainerAllocator.java:724) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.heartbeat(RMContainerAllocator.java:268) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.schedule(TestRMContainerAllocator.java:2023) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testBlackListedNodes(TestRMContainerAllocator.java:1318) FAILED: org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testReducerRampdownDiagnostics Error Message: null Stack Trace: java.lang.NullPointerException: null at org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.isRequestLabelChanged(AppSchedulingInfo.java:420) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.updateResourceRequests(AppSchedulingInfo.java:341) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplicationAttempt.updateResourceRequests(SchedulerApplicationAttempt.java:300) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoScheduler.allocate(FifoScheduler.java:350) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyFifoScheduler.allocate(TestRMContainerAllocator.java:1719) at org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:506) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor.makeRemoteRequest(RMContainerRequestor.java:204) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.makeRemoteRequest(TestRMContainerAllocator.java:2060) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.getResources(RMContainerAllocator.java:724) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.heartbeat(RMContainerAllocator.java:268) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.schedule(TestRMContainerAllocator.java:2023) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testReducerRampdownDiagnostics(TestRMContainerAllocator.java:450) FAILED: org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testReportedAppProgressWithOnlyMaps Error Message: null Stack Trace: java.lang.NullPointerException: null at org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.isRequestLabelChanged(AppSchedulingInfo.java:420) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.updateResourceRequests(AppSchedulingInfo.java:341) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplicationAttempt.updateResourceRequests(SchedulerApplicationAttempt.java:300) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoScheduler.allocate(FifoScheduler.java:350) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyFifoScheduler.allocate(TestRMContainerAllocator.java:1719) at org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:506) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor.makeRemoteRequest(RMContainerRequestor.java:204) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.makeRemoteRequest(TestRMContainerAllocator.java:2060) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.getResources(RMContainerAllocator.java:724) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.heartbeat(RMContainerAllocator.java:268) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.schedule(TestRMContainerAllocator.java:2023) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testReportedAppProgressWithOnlyMaps(TestRMContainerAllocator.java:1118) FAILED: org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testUpdatedNodes Error Message: null Stack Trace: java.lang.NullPointerException: null at org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.isRequestLabelChanged(AppSchedulingInfo.java:420) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.updateResourceRequests(AppSchedulingInfo.java:341) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplicationAttempt.updateResourceRequests(SchedulerApplicationAttempt.java:300) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoScheduler.allocate(FifoScheduler.java:350) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyFifoScheduler.allocate(TestRMContainerAllocator.java:1719) at org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:506) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor.makeRemoteRequest(RMContainerRequestor.java:204) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.makeRemoteRequest(TestRMContainerAllocator.java:2060) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.getResources(RMContainerAllocator.java:724) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.heartbeat(RMContainerAllocator.java:268) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.schedule(TestRMContainerAllocator.java:2023) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testUpdatedNodes(TestRMContainerAllocator.java:1213) FAILED: org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testSimple Error Message: null Stack Trace: java.lang.NullPointerException: null at org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.isRequestLabelChanged(AppSchedulingInfo.java:420) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.updateResourceRequests(AppSchedulingInfo.java:341) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplicationAttempt.updateResourceRequests(SchedulerApplicationAttempt.java:300) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoScheduler.allocate(FifoScheduler.java:350) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyFifoScheduler.allocate(TestRMContainerAllocator.java:1719) at org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:506) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor.makeRemoteRequest(RMContainerRequestor.java:204) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.makeRemoteRequest(TestRMContainerAllocator.java:2060) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.getResources(RMContainerAllocator.java:724) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.heartbeat(RMContainerAllocator.java:268) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.schedule(TestRMContainerAllocator.java:2023) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testSimple(TestRMContainerAllocator.java:231) FAILED: org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testIgnoreBlacklisting Error Message: null Stack Trace: java.lang.NullPointerException: null at org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.isRequestLabelChanged(AppSchedulingInfo.java:420) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.updateResourceRequests(AppSchedulingInfo.java:341) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplicationAttempt.updateResourceRequests(SchedulerApplicationAttempt.java:300) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoScheduler.allocate(FifoScheduler.java:350) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyFifoScheduler.allocate(TestRMContainerAllocator.java:1719) at org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:506) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor.makeRemoteRequest(RMContainerRequestor.java:204) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.makeRemoteRequest(TestRMContainerAllocator.java:2060) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.getResources(RMContainerAllocator.java:724) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.heartbeat(RMContainerAllocator.java:268) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.schedule(TestRMContainerAllocator.java:2023) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.getContainerOnHost(TestRMContainerAllocator.java:1511) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testIgnoreBlacklisting(TestRMContainerAllocator.java:1393) FAILED: org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testRMContainerAllocatorResendsRequestsOnRMRestart Error Message: null Stack Trace: java.lang.NullPointerException: null at org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.isRequestLabelChanged(AppSchedulingInfo.java:420) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.AppSchedulingInfo.updateResourceRequests(AppSchedulingInfo.java:341) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplicationAttempt.updateResourceRequests(SchedulerApplicationAttempt.java:300) at org.apache.hadoop.yarn.server.resourcemanager.scheduler.fifo.FifoScheduler.allocate(FifoScheduler.java:350) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyFifoScheduler.allocate(TestRMContainerAllocator.java:1719) at org.apache.hadoop.yarn.server.resourcemanager.ApplicationMasterService.allocate(ApplicationMasterService.java:506) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor.makeRemoteRequest(RMContainerRequestor.java:204) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.makeRemoteRequest(TestRMContainerAllocator.java:2060) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.getResources(RMContainerAllocator.java:724) at org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.heartbeat(RMContainerAllocator.java:268) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator$MyContainerAllocator.schedule(TestRMContainerAllocator.java:2023) at org.apache.hadoop.mapreduce.v2.app.rm.TestRMContainerAllocator.testRMContainerAllocatorResendsRequestsOnRMRestart(TestRMContainerAllocator.java:2509)
