See
<https://builds.apache.org/job/HBase-1.2-IT/it.test=IntegrationTestAcidGuarantees,jdk=latest1.7,label=yahoo-not-h2/538/changes>
Changes:
[tedyu] HBASE-16062 Improper error handling in WAL Reader/Writer creation
------------------------------------------
[...truncated 51747 lines...]
bool UsePPCLWSYNC = true {product}
bool UsePSAdaptiveSurvivorSizePolicy = true {product}
bool UseParNewGC = false {product}
bool UseParallelGC := true {product}
bool UseParallelOldGC = true {product}
bool UsePerfData = true {product}
bool UsePopCountInstruction = true {product}
bool UseRDPCForConstantTableBase = false {C2
product}
bool UseSHM = false {product}
intx UseSSE = 4 {product}
bool UseSSE42Intrinsics = true {product}
bool UseSerialGC = false {product}
bool UseSharedSpaces = false {product}
bool UseSignalChaining = true {product}
bool UseSplitVerifier = true {product}
bool UseStoreImmI16 = false {ARCH
product}
bool UseStringCache = false {product}
bool UseSuperWord = true {C2
product}
bool UseTLAB = true {pd
product}
bool UseThreadPriorities = true {pd
product}
bool UseTransparentHugePages = false {product}
bool UseTypeProfile = true {product}
bool UseUnalignedLoadStores = true {ARCH
product}
bool UseVMInterruptibleIO = false {product}
bool UseVectoredExceptions = false {pd
product}
bool UseXMMForArrayCopy = true {product}
bool UseXmmI2D = false {ARCH
product}
bool UseXmmI2F = false {ARCH
product}
bool UseXmmLoadAndClearUpper = true {ARCH
product}
bool UseXmmRegToRegMoveAll = true {ARCH
product}
bool VMThreadHintNoPreempt = false {product}
intx VMThreadPriority = -1 {product}
intx VMThreadStackSize = 1024 {pd
product}
intx ValueMapInitialSize = 11 {C1
product}
intx ValueMapMaxLoopSize = 8 {C1
product}
intx ValueSearchLimit = 1000 {C2
product}
bool VerifyMergedCPBytecodes = true {product}
intx WorkAroundNPTLTimedWaitHang = 1 {product}
uintx YoungGenerationSizeIncrement = 20 {product}
uintx YoungGenerationSizeSupplement = 80 {product}
uintx YoungGenerationSizeSupplementDecay = 8 {product}
uintx YoungPLABSize = 4096 {product}
bool ZeroTLAB = false {product}
intx hashCode = 0 {product}
Running org.apache.hadoop.hbase.IntegrationTestAcidGuarantees
SUREFIRE-859: [GC 194048K->13057K(740864K), 0.0243980 secs]
SUREFIRE-859: [GC 199001K->66859K(740864K), 0.1472020 secs]
Tests run: 3, Failures: 0, Errors: 3, Skipped: 0, Time elapsed: 5.11 sec <<<
FAILURE! - in org.apache.hadoop.hbase.IntegrationTestAcidGuarantees
testMixedAtomicity(org.apache.hadoop.hbase.IntegrationTestAcidGuarantees) Time
elapsed: 4.245 sec <<< ERROR!
java.lang.OutOfMemoryError: unable to create new native thread
at java.lang.Thread.start0(Native Method)
at java.lang.Thread.start(Thread.java:714)
at java.util.Timer.<init>(Timer.java:176)
at
org.apache.hadoop.ipc.Server$ConnectionManager.<init>(Server.java:2613)
at org.apache.hadoop.ipc.Server.<init>(Server.java:2168)
at org.apache.hadoop.ipc.RPC$Server.<init>(RPC.java:897)
at
org.apache.hadoop.ipc.ProtobufRpcEngine$Server.<init>(ProtobufRpcEngine.java:505)
at
org.apache.hadoop.ipc.ProtobufRpcEngine.getServer(ProtobufRpcEngine.java:480)
at org.apache.hadoop.ipc.RPC$Builder.build(RPC.java:742)
at
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.<init>(NameNodeRpcServer.java:311)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.createRpcServer(NameNode.java:614)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:587)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:751)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameNode.java:735)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1407)
at
org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:998)
at
org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:869)
at
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:704)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:642)
at
org.apache.hadoop.hbase.HBaseTestingUtility.startMiniDFSCluster(HBaseTestingUtility.java:585)
at
org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:982)
at
org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:863)
at
org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:845)
at
org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:832)
at
org.apache.hadoop.hbase.IntegrationTestingUtility.initializeCluster(IntegrationTestingUtility.java:78)
at
org.apache.hadoop.hbase.IntegrationTestAcidGuarantees.setUpCluster(IntegrationTestAcidGuarantees.java:66)
testGetAtomicity(org.apache.hadoop.hbase.IntegrationTestAcidGuarantees) Time
elapsed: 0.319 sec <<< ERROR!
java.lang.OutOfMemoryError: unable to create new native thread
at java.lang.Thread.start0(Native Method)
at java.lang.Thread.start(Thread.java:714)
at
org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:1100)
at
org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:1075)
at
org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:150)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:926)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:287)
at
org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:143)
at
org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:860)
at
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:704)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:642)
at
org.apache.hadoop.hbase.HBaseTestingUtility.startMiniDFSCluster(HBaseTestingUtility.java:585)
at
org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:982)
at
org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:863)
at
org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:845)
at
org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:832)
at
org.apache.hadoop.hbase.IntegrationTestingUtility.initializeCluster(IntegrationTestingUtility.java:78)
at
org.apache.hadoop.hbase.IntegrationTestAcidGuarantees.setUpCluster(IntegrationTestAcidGuarantees.java:66)
testScanAtomicity(org.apache.hadoop.hbase.IntegrationTestAcidGuarantees) Time
elapsed: 0.385 sec <<< ERROR!
java.lang.OutOfMemoryError: unable to create new native thread
at java.lang.Thread.start0(Native Method)
at java.lang.Thread.start(Thread.java:714)
at
org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:1100)
at
org.apache.hadoop.hdfs.server.namenode.FSImage.saveFSImageInAllDirs(FSImage.java:1075)
at
org.apache.hadoop.hdfs.server.namenode.FSImage.format(FSImage.java:150)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:926)
at
org.apache.hadoop.hdfs.server.namenode.NameNode.format(NameNode.java:287)
at
org.apache.hadoop.hdfs.DFSTestUtil.formatNameNode(DFSTestUtil.java:143)
at
org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:860)
at
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:704)
at org.apache.hadoop.hdfs.MiniDFSCluster.<init>(MiniDFSCluster.java:642)
at
org.apache.hadoop.hbase.HBaseTestingUtility.startMiniDFSCluster(HBaseTestingUtility.java:585)
at
org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:982)
at
org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:863)
at
org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:845)
at
org.apache.hadoop.hbase.HBaseTestingUtility.startMiniCluster(HBaseTestingUtility.java:832)
at
org.apache.hadoop.hbase.IntegrationTestingUtility.initializeCluster(IntegrationTestingUtility.java:78)
at
org.apache.hadoop.hbase.IntegrationTestAcidGuarantees.setUpCluster(IntegrationTestAcidGuarantees.java:66)
Results :
Tests in error:
IntegrationTestAcidGuarantees>IntegrationTestBase.setUp:133->setUpCluster:66
» OutOfMemory
IntegrationTestAcidGuarantees>IntegrationTestBase.setUp:133->setUpCluster:66
» OutOfMemory
IntegrationTestAcidGuarantees>IntegrationTestBase.setUp:133->setUpCluster:66
» OutOfMemory
Tests run: 3, Failures: 0, Errors: 3, Skipped: 0
[WARNING] File encoding has not been set, using platform encoding ISO-8859-1,
i.e. build is platform dependent! The file encoding for reports output files
should be provided by the POM property ${project.reporting.outputEncoding}.
[INFO]
[INFO] --- maven-failsafe-plugin:2.18.1:verify (verify) @ hbase-it ---
[INFO] Failsafe report directory:
<https://builds.apache.org/job/HBase-1.2-IT/it.test=IntegrationTestAcidGuarantees,jdk=latest1.7,label=yahoo-not-h2/ws/hbase-it/target/failsafe-reports>
[WARNING] File encoding has not been set, using platform encoding ISO-8859-1,
i.e. build is platform dependent! The file encoding for reports output files
should be provided by the POM property ${project.reporting.outputEncoding}.
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Apache HBase ...................................... SUCCESS [1:13.768s]
[INFO] Apache HBase - Checkstyle ......................... SUCCESS [5.740s]
[INFO] Apache HBase - Resource Bundle .................... SUCCESS [0.391s]
[INFO] Apache HBase - Annotations ........................ SUCCESS [1.546s]
[INFO] Apache HBase - Protocol ........................... SUCCESS [24.007s]
[INFO] Apache HBase - Common ............................. SUCCESS [1:18.335s]
[INFO] Apache HBase - Procedure .......................... SUCCESS [2.263s]
[INFO] Apache HBase - Client ............................. SUCCESS [11.859s]
[INFO] Apache HBase - Hadoop Compatibility ............... SUCCESS [1.667s]
[INFO] Apache HBase - Hadoop Two Compatibility ........... SUCCESS [3.328s]
[INFO] Apache HBase - Prefix Tree ........................ SUCCESS [2.058s]
[INFO] Apache HBase - Server ............................. SUCCESS [1:10.968s]
[INFO] Apache HBase - Testing Util ....................... SUCCESS [5.215s]
[INFO] Apache HBase - Thrift ............................. SUCCESS [9.170s]
[INFO] Apache HBase - Rest ............................... SUCCESS [4.084s]
[INFO] Apache HBase - Shell .............................. SUCCESS [14.169s]
[INFO] Apache HBase - Integration Tests .................. FAILURE [30:04.394s]
[INFO] Apache HBase - Examples ........................... SKIPPED
[INFO] Apache HBase - External Block Cache ............... SKIPPED
[INFO] Apache HBase - Assembly ........................... SKIPPED
[INFO] Apache HBase - Shaded ............................. SKIPPED
[INFO] Apache HBase - Shaded - Client .................... SKIPPED
[INFO] Apache HBase - Shaded - Server .................... SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 35:34.308s
[INFO] Finished at: Wed Jun 22 08:02:58 UTC 2016
[INFO] Final Memory: 153M/1918M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal
org.apache.maven.plugins:maven-failsafe-plugin:2.18.1:verify (verify) on
project hbase-it: There was a timeout or other error in the fork -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please
read the following articles:
[ERROR] [Help 1]
http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :hbase-it
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script : # Post-build task script. TODO: Check this in and have all
builds reference check-in.
pwd && ls
# NOTE!!!! The below code has been copied and pasted from
./dev-tools/run-test.sh
# Do not change here without syncing there and vice-versa.
ZOMBIE_TESTS_COUNT=`jps -v | grep surefirebooter | grep -e '-Dhbase.test' | wc
-l`
if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
echo "Suspicious java process found - waiting 30s to see if there are just
slow to stop"
sleep 30
ZOMBIE_TESTS_COUNT=`jps -v | grep surefirebooter | grep -e '-Dhbase.test' | wc
-l`
if [[ $ZOMBIE_TESTS_COUNT != 0 ]] ; then
echo " {color:red}There appear to be $ZOMBIE_TESTS_COUNT zombie
tests{color}, they should have been killed by surefire but survived"
jps -v | grep surefirebooter | grep -e '-Dhbase.test'
jps -v | grep surefirebooter | grep -e '-Dhbase.test' | cut -d ' ' -f 1 |
xargs -n 1 jstack
# Exit with error
exit 1
else
echo "We're ok: there is no zombie test, but some tests took some time to
stop"
fi
else
echo "We're ok: there is no zombie test"
fi
[yahoo-not-h2] $ /bin/bash -xe /tmp/hudson6612910674342418043.sh
+ pwd
<https://builds.apache.org/job/HBase-1.2-IT/it.test=IntegrationTestAcidGuarantees,jdk=latest1.7,label=yahoo-not-h2/ws/>
+ ls
bin
CHANGES.txt
conf
dev-support
hbase-annotations
hbase-assembly
hbase-checkstyle
hbase-client
hbase-common
hbase-examples
hbase-external-blockcache
hbase-hadoop2-compat
hbase-hadoop-compat
hbase-it
hbase-native-client
hbase-prefix-tree
hbase-procedure
hbase-protocol
hbase-resource-bundle
hbase-rest
hbase-server
hbase-shaded
hbase-shell
hbase-testing-util
hbase-thrift
hs_err_pid18239.log
LICENSE.txt
NOTICE.txt
pom.xml
README.txt
src
target
++ jps -v
++ grep surefirebooter
++ grep -e -Dhbase.test
++ wc -l
+ ZOMBIE_TESTS_COUNT=0
+ [[ 0 != 0 ]]
+ echo 'We'\''re ok: there is no zombie test'
We're ok: there is no zombie test
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts