See
<https://builds.apache.org/job/HBase-Trunk_matrix/jdk=latest1.7,label=yahoo-not-h2/1122/>
------------------------------------------
[...truncated 5765 lines...]
at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
at
org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
at
org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:95)
at org.apache.hadoop.hbase.ipc.AsyncCall.setFailed(AsyncCall.java:159)
at
org.apache.hadoop.hbase.ipc.AsyncServerResponseHandler.channelRead0(AsyncServerResponseHandler.java:81)
at
org.apache.hadoop.hbase.ipc.AsyncServerResponseHandler.channelRead0(AsyncServerResponseHandler.java:38)
at
io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105)
at
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:334)
at
io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:326)
at
io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:293)
at
io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:267)
at
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:334)
at
io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:326)
at
io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1320)
at
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:334)
at
io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:905)
at
io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:123)
at
io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:563)
at
io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:504)
at
io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:418)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:390)
at
io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:742)
at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException:
org.apache.hadoop.hbase.UnknownScannerException: Unknown scanner '48'. This can
happen due to any of the following reasons: a) Scanner id given is wrong, b)
Scanner lease expired because of long wait between consecutive client checkins,
c) Server may be closing down, d) RegionServer restart during upgrade.
If the issue is due to reason (b), a possible fix would be increasing the value
of'hbase.client.scanner.timeout.period' configuration.
at
org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2546)
at
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:34818)
at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2212)
at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:118)
at
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:189)
at
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:169)
at
org.apache.hadoop.hbase.ipc.AsyncServerResponseHandler.createRemoteException(AsyncServerResponseHandler.java:120)
at
org.apache.hadoop.hbase.ipc.AsyncServerResponseHandler.channelRead0(AsyncServerResponseHandler.java:76)
at
org.apache.hadoop.hbase.ipc.AsyncServerResponseHandler.channelRead0(AsyncServerResponseHandler.java:38)
at
io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105)
at
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:334)
at
io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:326)
at
io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:293)
at
io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:267)
at
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:334)
at
io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:326)
at
io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1320)
at
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
at
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:334)
at
io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:905)
at
io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:123)
at
io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:563)
at
io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:504)
at
io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:418)
at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:390)
at
io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:742)
at java.lang.Thread.run(Thread.java:745)
Running org.apache.hadoop.hbase.mapreduce.TestLoadIncrementalHFiles
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 24.442 sec - in
org.apache.hadoop.hbase.zookeeper.lock.TestZKInterProcessReadWriteLock
Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 162.541 sec -
in org.apache.hadoop.hbase.TestZooKeeper
Running org.apache.hadoop.hbase.mapreduce.TestCellCounter
Running org.apache.hadoop.hbase.mapreduce.TestImportTSVWithOperationAttributes
Running
org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFilesSplitRecovery
Tests run: 12, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 130.712 sec -
in org.apache.hadoop.hbase.namespace.TestNamespaceAuditor
Running org.apache.hadoop.hbase.mapreduce.TestTableInputFormat
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 22.799 sec - in
org.apache.hadoop.hbase.mapreduce.TestImportTSVWithOperationAttributes
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 36.33 sec - in
org.apache.hadoop.hbase.mapreduce.TestCellCounter
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 30.398 sec - in
org.apache.hadoop.hbase.mapreduce.TestSecureLoadIncrementalHFilesSplitRecovery
Running org.apache.hadoop.hbase.mapreduce.TestLoadIncrementalHFilesSplitRecovery
Running org.apache.hadoop.hbase.mapreduce.TestHLogRecordReader
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.489 sec - in
org.apache.hadoop.hbase.mapreduce.TestHLogRecordReader
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 23.314 sec - in
org.apache.hadoop.hbase.mapreduce.TestTableInputFormat
Running org.apache.hadoop.hbase.mapreduce.TestHRegionPartitioner
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.749 sec - in
org.apache.hadoop.hbase.mapreduce.TestHRegionPartitioner
Running org.apache.hadoop.hbase.mapreduce.TestImportTsv
Running org.apache.hadoop.hbase.mapreduce.TestTableInputFormatScan2
Running org.apache.hadoop.hbase.mapreduce.TestTableMapReduce
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 57.711 sec - in
org.apache.hadoop.hbase.mapreduce.TestLoadIncrementalHFilesSplitRecovery
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 39.334 sec - in
org.apache.hadoop.hbase.mapreduce.TestTableInputFormatScan2
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 34.468 sec - in
org.apache.hadoop.hbase.mapreduce.TestTableMapReduce
Running org.apache.hadoop.hbase.mapreduce.TestWALPlayer
Running org.apache.hadoop.hbase.mapreduce.TestImportExport
Running org.apache.hadoop.hbase.mapreduce.TestRowCounter
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 11.652 sec - in
org.apache.hadoop.hbase.mapreduce.TestWALPlayer
Running org.apache.hadoop.hbase.mapreduce.TestImportTSVWithVisibilityLabels
Tests run: 22, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 146.238 sec -
in org.apache.hadoop.hbase.mapreduce.TestLoadIncrementalHFiles
Running org.apache.hadoop.hbase.mapreduce.TestHashTable
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 41.981 sec - in
org.apache.hadoop.hbase.mapreduce.TestRowCounter
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 14.614 sec - in
org.apache.hadoop.hbase.mapreduce.TestHashTable
Running org.apache.hadoop.hbase.mapreduce.TestHFileOutputFormat2
Running org.apache.hadoop.hbase.mapreduce.TestImportTSVWithTTLs
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 12.521 sec - in
org.apache.hadoop.hbase.mapreduce.TestImportTSVWithTTLs
Running org.apache.hadoop.hbase.mapreduce.TestWALRecordReader
Tests run: 18, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 139.054 sec -
in org.apache.hadoop.hbase.mapreduce.TestImportTsv
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 1.786 sec - in
org.apache.hadoop.hbase.mapreduce.TestWALRecordReader
Tests run: 16, Failures: 0, Errors: 0, Skipped: 14, Time elapsed: 34.171 sec -
in org.apache.hadoop.hbase.mapreduce.TestHFileOutputFormat2
Tests run: 12, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 93.789 sec -
in org.apache.hadoop.hbase.mapreduce.TestImportExport
Running org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 78.714 sec - in
org.apache.hadoop.hbase.mapreduce.TestImportTSVWithVisibilityLabels
Running org.apache.hadoop.hbase.mapreduce.TestMultithreadedTableMapper
Running org.apache.hadoop.hbase.mapreduce.TestTableInputFormatScan1
Running org.apache.hadoop.hbase.mapreduce.TestTimeRangeMapRed
Running org.apache.hadoop.hbase.mapreduce.TestMultiTableSnapshotInputFormat
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 8.585 sec - in
org.apache.hadoop.hbase.mapreduce.TestTimeRangeMapRed
Running org.apache.hadoop.hbase.mapreduce.TestSyncTable
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 39.813 sec - in
org.apache.hadoop.hbase.mapreduce.TestTableInputFormatScan1
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 36.967 sec - in
org.apache.hadoop.hbase.mapreduce.TestMultiTableSnapshotInputFormat
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 47.195 sec - in
org.apache.hadoop.hbase.mapreduce.TestMultithreadedTableMapper
Running org.apache.hadoop.hbase.mapreduce.TestCopyTable
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 29.594 sec - in
org.apache.hadoop.hbase.mapreduce.TestSyncTable
Running org.apache.hadoop.hbase.io.TestFileLink
Running org.apache.hadoop.hbase.TestMetaTableAccessor
Running org.apache.hadoop.hbase.mapreduce.TestMultiTableInputFormat
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 15.036 sec - in
org.apache.hadoop.hbase.io.TestFileLink
Running org.apache.hadoop.hbase.io.hfile.TestHFileBlockIndex
Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 7.765 sec - in
org.apache.hadoop.hbase.io.hfile.TestHFileBlockIndex
Running org.apache.hadoop.hbase.io.hfile.TestSeekBeforeWithInlineBlocks
Tests run: 15, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 25.807 sec -
in org.apache.hadoop.hbase.TestMetaTableAccessor
Running org.apache.hadoop.hbase.io.hfile.TestScannerSelectionUsingTTL
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 13.421 sec - in
org.apache.hadoop.hbase.io.hfile.TestSeekBeforeWithInlineBlocks
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 34.964 sec - in
org.apache.hadoop.hbase.mapreduce.TestMultiTableInputFormat
Running org.apache.hadoop.hbase.io.hfile.TestHFileSeek
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.486 sec - in
org.apache.hadoop.hbase.io.hfile.TestHFileSeek
Running org.apache.hadoop.hbase.io.hfile.TestHFileBlock
Running org.apache.hadoop.hbase.io.hfile.TestCacheConfig
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 121.846 sec -
in org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.403 sec - in
org.apache.hadoop.hbase.io.hfile.TestCacheConfig
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 57.248 sec - in
org.apache.hadoop.hbase.mapreduce.TestCopyTable
Running org.apache.hadoop.hbase.io.hfile.TestScannerFromBucketCache
Running org.apache.hadoop.hbase.io.hfile.TestForceCacheImportantBlocks
Running org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 5.891 sec - in
org.apache.hadoop.hbase.io.hfile.TestScannerFromBucketCache
Running org.apache.hadoop.hbase.io.encoding.TestLoadAndSwitchEncodeOnDisk
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 29.633 sec - in
org.apache.hadoop.hbase.io.hfile.TestForceCacheImportantBlocks
Running org.apache.hadoop.hbase.io.encoding.TestChangingEncoding
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 63.949 sec - in
org.apache.hadoop.hbase.io.hfile.TestScannerSelectionUsingTTL
Running org.apache.hadoop.hbase.io.encoding.TestEncodedSeekers
Tests run: 28, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 91.468 sec -
in org.apache.hadoop.hbase.io.hfile.TestHFileBlock
Running org.apache.hadoop.hbase.io.encoding.TestDataBlockEncoders
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 88.365 sec - in
org.apache.hadoop.hbase.io.encoding.TestLoadAndSwitchEncodeOnDisk
Running org.apache.hadoop.hbase.io.encoding.TestBufferedDataBlockEncoder
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 14.148 sec - in
org.apache.hadoop.hbase.io.encoding.TestBufferedDataBlockEncoder
Running org.apache.hadoop.hbase.io.asyncfs.TestFanOutOneBlockAsyncDFSOutput
Tests run: 56, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 49.659 sec -
in org.apache.hadoop.hbase.io.encoding.TestDataBlockEncoders
Running org.apache.hadoop.hbase.io.asyncfs.TestSaslFanOutOneBlockAsyncDFSOutput
Tests run: 20, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 108.003 sec -
in org.apache.hadoop.hbase.io.encoding.TestEncodedSeekers
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 39.474 sec - in
org.apache.hadoop.hbase.io.asyncfs.TestFanOutOneBlockAsyncDFSOutput
Tests run: 72, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 163.411 sec -
in org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite
Running org.apache.hadoop.hbase.filter.TestScanRowPrefix
Running org.apache.hadoop.hbase.filter.TestFilterWithScanLimits
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 5.806 sec - in
org.apache.hadoop.hbase.filter.TestScanRowPrefix
Running org.apache.hadoop.hbase.filter.TestFilterWrapper
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 145.54 sec - in
org.apache.hadoop.hbase.io.encoding.TestChangingEncoding
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.248 sec - in
org.apache.hadoop.hbase.filter.TestFilterWrapper
Running org.apache.hadoop.hbase.filter.TestFuzzyRowFilterEndToEnd
Running org.apache.hadoop.hbase.filter.TestFilterListOrOperatorWithBlkCnt
Running org.apache.hadoop.hbase.filter.TestColumnRangeFilter
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 7.417 sec - in
org.apache.hadoop.hbase.filter.TestColumnRangeFilter
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 16.27 sec - in
org.apache.hadoop.hbase.filter.TestFilterListOrOperatorWithBlkCnt
Running org.apache.hadoop.hbase.filter.TestFuzzyRowAndColumnRangeFilter
Running org.apache.hadoop.hbase.filter.TestMultiRowRangeFilter
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.903 sec - in
org.apache.hadoop.hbase.filter.TestFuzzyRowAndColumnRangeFilter
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 53.599 sec - in
org.apache.hadoop.hbase.filter.TestFilterWithScanLimits
Tests run: 36, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 102.985 sec -
in org.apache.hadoop.hbase.io.asyncfs.TestSaslFanOutOneBlockAsyncDFSOutput
Tests run: 20, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 16.637 sec -
in org.apache.hadoop.hbase.filter.TestMultiRowRangeFilter
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 287.093 sec -
in org.apache.hadoop.hbase.filter.TestFuzzyRowFilterEndToEnd
Results :
Tests in error:
TestFromClientSide.testWeirdCacheBehaviour:342->deleteColumns:359 » Runtime
or...
TestMobCloneSnapshotFromClient>TestCloneSnapshotFromClient.testCloneLinksAfterDelete:203->verifyRowCount:92
» Runtime
Tests run: 2567, Failures: 0, Errors: 2, Skipped: 49
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Apache HBase ...................................... SUCCESS [ 2.579 s]
[INFO] Apache HBase - Checkstyle ......................... SUCCESS [ 0.662 s]
[INFO] Apache HBase - Resource Bundle .................... SUCCESS [ 0.201 s]
[INFO] Apache HBase - Annotations ........................ SUCCESS [ 0.179 s]
[INFO] Apache HBase - Protocol ........................... SUCCESS [ 3.252 s]
[INFO] Apache HBase - Common ............................. SUCCESS [01:30 min]
[INFO] Apache HBase - Procedure .......................... SUCCESS [01:35 min]
[INFO] Apache HBase - Client ............................. SUCCESS [ 46.828 s]
[INFO] Apache HBase - Hadoop Compatibility ............... SUCCESS [ 7.435 s]
[INFO] Apache HBase - Hadoop Two Compatibility ........... SUCCESS [ 9.387 s]
[INFO] Apache HBase - Prefix Tree ........................ SUCCESS [ 10.527 s]
[INFO] Apache HBase - Server ............................. FAILURE [ 01:29 h]
[INFO] Apache HBase - Testing Util ....................... SKIPPED
[INFO] Apache HBase - Thrift ............................. SKIPPED
[INFO] Apache HBase - RSGroup ............................ SKIPPED
[INFO] Apache HBase - Shell .............................. SKIPPED
[INFO] Apache HBase - Integration Tests .................. SKIPPED
[INFO] Apache HBase - Examples ........................... SKIPPED
[INFO] Apache HBase - Rest ............................... SKIPPED
[INFO] Apache HBase - External Block Cache ............... SKIPPED
[INFO] Apache HBase - Spark .............................. SKIPPED
[INFO] Apache HBase - Assembly ........................... SKIPPED
[INFO] Apache HBase - Shaded ............................. SKIPPED
[INFO] Apache HBase - Shaded - Client .................... SKIPPED
[INFO] Apache HBase - Shaded - Server .................... SKIPPED
[INFO] Apache HBase - Archetypes ......................... SKIPPED
[INFO] Apache HBase - Exemplar for hbase-client archetype SKIPPED
[INFO] Apache HBase - Exemplar for hbase-shaded-client archetype SKIPPED
[INFO] Apache HBase - Archetype builder .................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 01:34 h
[INFO] Finished at: 2016-06-27T07:15:32+00:00
[INFO] Final Memory: 70M/651M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal
org.apache.maven.plugins:maven-surefire-plugin:2.18.1:test
(secondPartTestsExecution) on project hbase-server: There are test failures.
[ERROR]
[ERROR] Please refer to
<https://builds.apache.org/job/HBase-Trunk_matrix/jdk=latest1.7,label=yahoo-not-h2/ws/hbase-server/target/surefire-reports>
for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please
read the following articles:
[ERROR] [Help 1]
http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :hbase-server
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script : # Run zombie detector script
./dev-support/zombie-detector.sh --jenkins ${BUILD_ID}
[yahoo-not-h2] $ /bin/bash -xe /tmp/hudson1335894813615554673.sh
+ ./dev-support/zombie-detector.sh --jenkins 1122
Mon Jun 27 07:15:34 UTC 2016 We're ok: there is no zombie test
{color:green}+1 zombies{color}. No zombie tests found running at the end of
the build.
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
[FINDBUGS] Skipping publisher since build result is FAILURE
[CHECKSTYLE] Skipping publisher since build result is FAILURE