See 
<https://builds.apache.org/job/HBase-1.5/jdk=JDK_1_7,label=Hadoop&&!H13/56/display/redirect?page=changes>

Changes:

[tedyu] HBASE-18789 Displays the reporting interval of each RS on the Master

------------------------------------------
[...truncated 2.42 MB...]
  Run 1: 
TestRSKilledWhenInitializing.testRSTerminationAfterRegisteringToMasterBeforeCreatingEphemeralNode:142
 null
  Run 2: 
TestRSKilledWhenInitializing.testRSTerminationAfterRegisteringToMasterBeforeCreatingEphemeralNode:115
 » TestTimedOut
  Run 3: 
TestRSKilledWhenInitializing.testRSTerminationAfterRegisteringToMasterBeforeCreatingEphemeralNode
 » 
  Run 4: 
TestRSKilledWhenInitializing.testRSTerminationAfterRegisteringToMasterBeforeCreatingEphemeralNode:115
 » TestTimedOut
  Run 5: 
TestRSKilledWhenInitializing.testRSTerminationAfterRegisteringToMasterBeforeCreatingEphemeralNode
 » 

org.apache.hadoop.hbase.regionserver.TestSCVFWithMiniCluster.org.apache.hadoop.hbase.regionserver.TestSCVFWithMiniCluster
  Run 1: TestSCVFWithMiniCluster.setUp:83->create:229 » RetriesExhausted Failed 
after a...
  Run 2: TestSCVFWithMiniCluster.tearDown:128 NullPointer

Flaked tests: 
org.apache.hadoop.hbase.client.TestClientScannerRPCTimeout.testScannerNextRPCTimesout(org.apache.hadoop.hbase.client.TestClientScannerRPCTimeout)
  Run 1: TestClientScannerRPCTimeout.testScannerNextRPCTimesout:87 » 
TableExists testSc...
  Run 2: PASS

org.apache.hadoop.hbase.client.TestReplicasClient.testCancelOfMultiGet(org.apache.hadoop.hbase.client.TestReplicasClient)
  Run 1: TestReplicasClient.testCancelOfMultiGet:585 null
  Run 2: PASS

org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFilesSplitRecovery.testGroupOrSplitFailure(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFilesSplitRecovery)
  Run 1: 
TestSecureLoadIncrementalHFilesSplitRecovery>TestLoadIncrementalHFilesSplitRecovery.testGroupOrSplitFailure:462->TestLoadIncrementalHFilesSplitRecovery.setupTable:136->Object.wait:-2
 » TestTimedOut
  Run 2: PASS

org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFilesSplitRecovery.testGroupOrSplitPresplit(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFilesSplitRecovery)
  Run 1: 
TestSecureLoadIncrementalHFilesSplitRecovery>TestLoadIncrementalHFilesSplitRecovery.testGroupOrSplitPresplit:389->TestLoadIncrementalHFilesSplitRecovery.setupTable:136->Object.wait:-2
 » TestTimedOut
  Run 2: PASS

org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFilesSplitRecovery.testGroupOrSplitWhenRegionHoleExistsInMeta(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFilesSplitRecovery)
  Run 1: 
TestSecureLoadIncrementalHFilesSplitRecovery>TestLoadIncrementalHFilesSplitRecovery.testGroupOrSplitWhenRegionHoleExistsInMeta:502->TestLoadIncrementalHFilesSplitRecovery.setupTableWithSplitkeys:159->Object.wait:-2
 » TestTimedOut
  Run 2: PASS

org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFilesSplitRecovery.testSplitTmpFileCleanUp(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFilesSplitRecovery)
  Run 1: 
TestSecureLoadIncrementalHFilesSplitRecovery>TestLoadIncrementalHFilesSplitRecovery.testSplitTmpFileCleanUp:431->TestLoadIncrementalHFilesSplitRecovery.setupTableWithSplitkeys:159->Object.wait:-2
 » TestTimedOut
  Run 2: PASS

org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFilesSplitRecovery.testSplitWhileBulkLoadPhase(org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFilesSplitRecovery)
  Run 1: 
TestSecureLoadIncrementalHFilesSplitRecovery>TestLoadIncrementalHFilesSplitRecovery.testSplitWhileBulkLoadPhase:346->TestLoadIncrementalHFilesSplitRecovery.setupTable:136->Object.wait:-2
 » TestTimedOut
  Run 2: PASS

org.apache.hadoop.hbase.regionserver.TestCompactionInDeadRegionServer.test(org.apache.hadoop.hbase.regionserver.TestCompactionInDeadRegionServer)
  Run 1: TestCompactionInDeadRegionServer.test:132 Should fail as our wal file 
has already been closed, and walDir has also been renamed
  Run 2: TestCompactionInDeadRegionServer.test:132 Should fail as our wal file 
has already been closed, and walDir has also been renamed
  Run 3: PASS

org.apache.hadoop.hbase.security.access.TestAccessControlFilter.testQualifierAccess(org.apache.hadoop.hbase.security.access.TestAccessControlFilter)
  Run 1: 
TestAccessControlFilter.testQualifierAccess:97->SecureTestUtil.createTable:710->SecureTestUtil.createTable:726->SecureTestUtil.createTable:739->Object.wait:-2
 » TestTimedOut
  Run 2: PASS


Tests run: 2844, Failures: 0, Errors: 2, Skipped: 57, Flakes: 9

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] Apache HBase ...................................... SUCCESS [1:45.437s]
[INFO] Apache HBase - Checkstyle ......................... SUCCESS [15.954s]
[INFO] Apache HBase - Resource Bundle .................... SUCCESS [0.321s]
[INFO] Apache HBase - Annotations ........................ SUCCESS [1.369s]
[INFO] Apache HBase - Protocol ........................... SUCCESS [17.505s]
[INFO] Apache HBase - Common ............................. SUCCESS [3:01.029s]
[INFO] Apache HBase - Procedure .......................... SUCCESS [4:05.762s]
[INFO] Apache HBase - Metrics API ........................ SUCCESS [2.224s]
[INFO] Apache HBase - Hadoop Compatibility ............... SUCCESS [13.194s]
[INFO] Apache HBase - Metrics Implementation ............. SUCCESS [7.175s]
[INFO] Apache HBase - Hadoop Two Compatibility ........... SUCCESS [24.780s]
[INFO] Apache HBase - Client ............................. SUCCESS [6:12.601s]
[INFO] Apache HBase - Prefix Tree ........................ SUCCESS [9.970s]
[INFO] Apache HBase - Server ............................. FAILURE 
[2:05:46.605s]
[INFO] Apache HBase - Testing Util ....................... SKIPPED
[INFO] Apache HBase - Thrift ............................. SKIPPED
[INFO] Apache HBase - Rest ............................... SKIPPED
[INFO] Apache HBase - Shell .............................. SKIPPED
[INFO] Apache HBase - Integration Tests .................. SKIPPED
[INFO] Apache HBase - Examples ........................... SKIPPED
[INFO] Apache HBase - External Block Cache ............... SKIPPED
[INFO] Apache HBase - Assembly ........................... SKIPPED
[INFO] Apache HBase - Shaded ............................. SKIPPED
[INFO] Apache HBase - Shaded - Client .................... SKIPPED
[INFO] Apache HBase - Shaded - Server .................... SKIPPED
[INFO] Apache HBase Shaded Packaging Invariants .......... SKIPPED
[INFO] Apache HBase - Archetypes ......................... SKIPPED
[INFO] Apache HBase - Exemplar for hbase-client archetype  SKIPPED
[INFO] Apache HBase - Exemplar for hbase-shaded-client archetype  SKIPPED
[INFO] Apache HBase - Archetype builder .................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 2:22:57.009s
[INFO] Finished at: Wed Sep 13 19:03:07 UTC 2017
[INFO] Final Memory: 525M/1306M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal 
org.apache.maven.plugins:maven-surefire-plugin:2.18.1:test 
(secondPartTestsExecution) on project hbase-server: ExecutionException: 
java.lang.RuntimeException: There was an error in the forked process
[ERROR] java.lang.ArrayIndexOutOfBoundsException: 1
[ERROR] at 
org.apache.maven.surefire.common.junit4.JUnit4ProviderUtil.generateFailingTests(JUnit4ProviderUtil.java:64)
[ERROR] at 
org.apache.maven.surefire.junitcore.JUnitCoreProvider.invoke(JUnitCoreProvider.java:151)
[ERROR] at 
org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:203)
[ERROR] at 
org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:155)
[ERROR] at 
org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:103)
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please 
read the following articles:
[ERROR] [Help 1] 
http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hbase-server
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  : # Post-build task script. TODO: Check this in and have all 
builds reference check-in.
pwd && ls
# NOTE!!!! The below code has been copied and pasted from 
./dev-tools/run-test.sh
# Do not change here without syncing there and vice-versa.
ZOMBIE_TESTS_COUNT=`jps -v | grep surefirebooter | grep -e '-Dhbase.test' | wc 
-l`
if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
 echo "Suspicious java process found - waiting 30s to see if there are just 
slow to stop"
 sleep 30
 ZOMBIE_TESTS_COUNT=`jps -v | grep surefirebooter | grep -e '-Dhbase.test' | wc 
-l`
 if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
   echo " {color:red}There appear to be $ZOMBIE_TESTS_COUNT zombie 
tests{color}, they should have been killed by surefire but survived"
   jps -v | grep surefirebooter | grep -e '-Dhbase.test'
   jps -v | grep surefirebooter | grep -e '-Dhbase.test' | cut -d ' ' -f 1 | 
xargs -n 1 jstack
   # Exit with error
   exit 1
 else
   echo "We're ok: there is no zombie test, but some tests took some time to 
stop"
 fi
else
  echo "We're ok: there is no zombie test"
fi
Setting JDK_1_7_LATEST__HOME=/home/jenkins/tools/java/latest1.7
Setting JDK_1_8_LATEST__HOME=/home/jenkins/tools/java/latest1.8
[19ccb43d] $ /bin/bash -xe /tmp/jenkins6694159300253699934.sh
+ pwd
<https://builds.apache.org/job/HBase-1.5/jdk=JDK_1_7,label=Hadoop&&!H13/ws/>
+ ls
bin
CHANGES.txt
conf
dev-support
extra_env_var
hbase-annotations
hbase-archetypes
hbase-assembly
hbase-checkstyle
hbase-client
hbase-common
hbase-examples
hbase-external-blockcache
hbase-hadoop2-compat
hbase-hadoop-compat
hbase-it
hbase-metrics
hbase-metrics-api
hbase-native-client
hbase-prefix-tree
hbase-procedure
hbase-protocol
hbase-resource-bundle
hbase-rest
hbase-server
hbase-shaded
hbase-shell
hbase-testing-util
hbase-thrift
LICENSE.txt
NOTICE.txt
pom.xml
README.txt
src
target
++ jps -v
++ grep surefirebooter
++ grep -e -Dhbase.test
++ wc -l
+ ZOMBIE_TESTS_COUNT=0
+ [[ 0 != 0 ]]
+ echo 'We'\''re ok: there is no zombie test'
We're ok: there is no zombie test
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Setting JDK_1_7_LATEST__HOME=/home/jenkins/tools/java/latest1.7
Setting JDK_1_8_LATEST__HOME=/home/jenkins/tools/java/latest1.8
[Fast Archiver] No prior successful build to compare, so performing full copy 
of artifacts
Recording test results
Setting JDK_1_7_LATEST__HOME=/home/jenkins/tools/java/latest1.7
Setting JDK_1_8_LATEST__HOME=/home/jenkins/tools/java/latest1.8
ERROR: Step ?Publish JUnit test result report? aborted due to exception: 
java.lang.OutOfMemoryError: Java heap space
        at 
com.sun.org.apache.xerces.internal.util.XMLStringBuffer.append(XMLStringBuffer.java:208)
        at 
com.sun.org.apache.xerces.internal.impl.XMLEntityScanner.scanData(XMLEntityScanner.java:1515)
        at 
com.sun.org.apache.xerces.internal.impl.XMLDocumentFragmentScannerImpl.scanCDATASection(XMLDocumentFragmentScannerImpl.java:1654)
        at 
com.sun.org.apache.xerces.internal.impl.XMLDocumentFragmentScannerImpl$FragmentContentDriver.next(XMLDocumentFragmentScannerImpl.java:3014)
        at 
com.sun.org.apache.xerces.internal.impl.XMLDocumentScannerImpl.next(XMLDocumentScannerImpl.java:602)
        at 
com.sun.org.apache.xerces.internal.impl.XMLNSDocumentScannerImpl.next(XMLNSDocumentScannerImpl.java:112)
        at 
com.sun.org.apache.xerces.internal.impl.XMLDocumentFragmentScannerImpl.scanDocument(XMLDocumentFragmentScannerImpl.java:505)
        at 
com.sun.org.apache.xerces.internal.parsers.XML11Configuration.parse(XML11Configuration.java:841)
        at 
com.sun.org.apache.xerces.internal.parsers.XML11Configuration.parse(XML11Configuration.java:770)
        at 
com.sun.org.apache.xerces.internal.parsers.XMLParser.parse(XMLParser.java:141)
        at 
com.sun.org.apache.xerces.internal.parsers.AbstractSAXParser.parse(AbstractSAXParser.java:1213)
        at 
com.sun.org.apache.xerces.internal.jaxp.SAXParserImpl$JAXPSAXParser.parse(SAXParserImpl.java:643)
        at org.dom4j.io.SAXReader.read(SAXReader.java:465)
        at org.dom4j.io.SAXReader.read(SAXReader.java:343)
        at hudson.tasks.junit.SuiteResult.parse(SuiteResult.java:132)
        at hudson.tasks.junit.TestResult.parse(TestResult.java:302)
        at hudson.tasks.junit.TestResult.parsePossiblyEmpty(TestResult.java:244)
        at hudson.tasks.junit.TestResult.parse(TestResult.java:175)
        at hudson.tasks.junit.TestResult.parse(TestResult.java:154)
        at hudson.tasks.junit.TestResult.<init>(TestResult.java:126)
        at 
hudson.tasks.junit.JUnitParser$ParseResultCallable.invoke(JUnitParser.java:132)
        at 
hudson.tasks.junit.JUnitParser$ParseResultCallable.invoke(JUnitParser.java:107)
        at hudson.FilePath$FileCallableWrapper.call(FilePath.java:2739)
        at hudson.remoting.UserRequest.perform(UserRequest.java:153)
        at hudson.remoting.UserRequest.perform(UserRequest.java:50)
        at hudson.remoting.Request$2.run(Request.java:336)
        at 
hudson.remoting.InterceptingExecutorService$1.call(InterceptingExecutorService.java:68)
        at java.util.concurrent.FutureTask.run(FutureTask.java:266)
        at 
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
        at 
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
        at java.lang.Thread.run(Thread.java:748)
        at ......remote call to H5(Native Method)
        at hudson.remoting.Channel.attachCallSiteStackTrace(Channel.java:1545)
        at hudson.remoting.UserResponse.retrieve(UserRequest.java:253)
        at hudson.remoting.Channel.call(Channel.java:830)
Caused: java.io.IOException: Remote call on H5 failed
        at hudson.remoting.Channel.call(Channel.java:838)
        at hudson.FilePath.act(FilePath.java:986)
Caused: java.io.IOException: remote file operation failed: 
<https://builds.apache.org/job/HBase-1.5/jdk=JDK_1_7,label=Hadoop&&!H13/ws/> at 
hudson.remoting.Channel@9608f0d:H5
        at hudson.FilePath.act(FilePath.java:993)
        at hudson.FilePath.act(FilePath.java:975)
        at hudson.tasks.junit.JUnitParser.parseResult(JUnitParser.java:103)
        at 
hudson.tasks.junit.JUnitResultArchiver.parse(JUnitResultArchiver.java:128)
        at 
hudson.tasks.junit.JUnitResultArchiver.perform(JUnitResultArchiver.java:149)
        at 
hudson.tasks.BuildStepCompatibilityLayer.perform(BuildStepCompatibilityLayer.java:81)
        at hudson.tasks.BuildStepMonitor$1.perform(BuildStepMonitor.java:20)
        at 
hudson.model.AbstractBuild$AbstractBuildExecution.perform(AbstractBuild.java:735)
        at 
hudson.model.AbstractBuild$AbstractBuildExecution.performAllBuildSteps(AbstractBuild.java:676)
        at hudson.model.Build$BuildExecution.post2(Build.java:186)
        at 
hudson.model.AbstractBuild$AbstractBuildExecution.post(AbstractBuild.java:621)
        at hudson.model.Run.execute(Run.java:1760)
        at hudson.matrix.MatrixRun.run(MatrixRun.java:146)
        at hudson.model.ResourceController.execute(ResourceController.java:97)
        at hudson.model.Executor.run(Executor.java:405)
Setting JDK_1_7_LATEST__HOME=/home/jenkins/tools/java/latest1.7
Setting JDK_1_8_LATEST__HOME=/home/jenkins/tools/java/latest1.8

Reply via email to