See 
<https://builds.apache.org/job/HBase-Trunk_matrix/jdk=latest1.7,label=yahoo-not-h2/1111/>

------------------------------------------
[...truncated 5720 lines...]
Running org.apache.hadoop.hbase.mapreduce.TestTableInputFormatScan1
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 72.794 sec - in 
org.apache.hadoop.hbase.mapreduce.TestImportTSVWithVisibilityLabels
Running org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat
Tests run: 22, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 205.977 sec 
<<< FAILURE! - in 
org.apache.hadoop.hbase.mapreduce.TestLoadIncrementalHFilesUseSecurityEndPoint
testRegionCrossingHFileSplitRowBloom(org.apache.hadoop.hbase.mapreduce.TestLoadIncrementalHFilesUseSecurityEndPoint)
  Time elapsed: 70.598 sec  <<< ERROR!
org.apache.hadoop.hbase.client.ScannerTimeoutException: 60017ms passed since 
the last invocation, timeout is currently set to 60000
        at 
org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:433)
        at 
org.apache.hadoop.hbase.client.ClientScanner.nextWithSyncCache(ClientScanner.java:363)
        at 
org.apache.hadoop.hbase.client.ClientSimpleScanner.next(ClientSimpleScanner.java:51)
        at 
org.apache.hadoop.hbase.MetaTableAccessor.scanMeta(MetaTableAccessor.java:776)
        at 
org.apache.hadoop.hbase.MetaTableAccessor.scanMeta(MetaTableAccessor.java:695)
        at 
org.apache.hadoop.hbase.MetaTableAccessor.scanMetaForTableRegions(MetaTableAccessor.java:690)
        at 
org.apache.hadoop.hbase.client.HRegionLocator.listRegionLocations(HRegionLocator.java:155)
        at 
org.apache.hadoop.hbase.client.HRegionLocator.getStartEndKeys(HRegionLocator.java:121)
        at 
org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.doBulkLoad(LoadIncrementalHFiles.java:360)
        at 
org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.run(LoadIncrementalHFiles.java:1065)
        at 
org.apache.hadoop.hbase.mapreduce.TestLoadIncrementalHFiles.runTest(TestLoadIncrementalHFiles.java:295)
        at 
org.apache.hadoop.hbase.mapreduce.TestLoadIncrementalHFiles.runTest(TestLoadIncrementalHFiles.java:269)
        at 
org.apache.hadoop.hbase.mapreduce.TestLoadIncrementalHFiles.runTest(TestLoadIncrementalHFiles.java:263)
        at 
org.apache.hadoop.hbase.mapreduce.TestLoadIncrementalHFiles.testRegionCrossingHFileSplit(TestLoadIncrementalHFiles.java:227)
        at 
org.apache.hadoop.hbase.mapreduce.TestLoadIncrementalHFiles.testRegionCrossingHFileSplitRowBloom(TestLoadIncrementalHFiles.java:192)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
        at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
        at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
        at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
        at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:298)
        at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:292)
        at java.util.concurrent.FutureTask.run(FutureTask.java:262)
        at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.UnknownScannerException: 
org.apache.hadoop.hbase.UnknownScannerException: Unknown scanner '405'. This 
can happen due to any of the following reasons: a) Scanner id given is wrong, 
b) Scanner lease expired because of long wait between consecutive client 
checkins, c) Server may be closing down, d) RegionServer restart during upgrade.
If the issue is due to reason (b), a possible fix would be increasing the value 
of'hbase.client.scanner.timeout.period' configuration.
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2546)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:34818)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2212)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:116)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:189)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:169)

        at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
        at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
        at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
        at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
        at 
org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
        at 
org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:95)
        at org.apache.hadoop.hbase.ipc.AsyncCall.setFailed(AsyncCall.java:159)
        at 
org.apache.hadoop.hbase.ipc.AsyncServerResponseHandler.channelRead0(AsyncServerResponseHandler.java:81)
        at 
org.apache.hadoop.hbase.ipc.AsyncServerResponseHandler.channelRead0(AsyncServerResponseHandler.java:38)
        at 
io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105)
        at 
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
        at 
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:334)
        at 
io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:326)
        at 
io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:293)
        at 
io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:267)
        at 
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
        at 
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:334)
        at 
io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:326)
        at 
io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1320)
        at 
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
        at 
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:334)
        at 
io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:905)
        at 
io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:123)
        at 
io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:563)
        at 
io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:504)
        at 
io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:418)
        at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:390)
        at 
io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:742)
        at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException: 
org.apache.hadoop.hbase.UnknownScannerException: Unknown scanner '405'. This 
can happen due to any of the following reasons: a) Scanner id given is wrong, 
b) Scanner lease expired because of long wait between consecutive client 
checkins, c) Server may be closing down, d) RegionServer restart during upgrade.
If the issue is due to reason (b), a possible fix would be increasing the value 
of'hbase.client.scanner.timeout.period' configuration.
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2546)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:34818)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2212)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:116)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:189)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor$Handler.run(RpcExecutor.java:169)

        at 
org.apache.hadoop.hbase.ipc.AsyncServerResponseHandler.createRemoteException(AsyncServerResponseHandler.java:120)
        at 
org.apache.hadoop.hbase.ipc.AsyncServerResponseHandler.channelRead0(AsyncServerResponseHandler.java:76)
        at 
org.apache.hadoop.hbase.ipc.AsyncServerResponseHandler.channelRead0(AsyncServerResponseHandler.java:38)
        at 
io.netty.channel.SimpleChannelInboundHandler.channelRead(SimpleChannelInboundHandler.java:105)
        at 
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
        at 
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:334)
        at 
io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:326)
        at 
io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:293)
        at 
io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:267)
        at 
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
        at 
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:334)
        at 
io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:326)
        at 
io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1320)
        at 
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:348)
        at 
io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:334)
        at 
io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:905)
        at 
io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:123)
        at 
io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:563)
        at 
io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:504)
        at 
io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:418)
        at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:390)
        at 
io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:742)
        at java.lang.Thread.run(Thread.java:745)

Running org.apache.hadoop.hbase.mapreduce.TestMultithreadedTableMapper
Running org.apache.hadoop.hbase.mapreduce.TestTimeRangeMapRed
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 7.437 sec - in 
org.apache.hadoop.hbase.mapreduce.TestTimeRangeMapRed
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 38.718 sec - in 
org.apache.hadoop.hbase.mapreduce.TestTableInputFormatScan1
Running org.apache.hadoop.hbase.mapreduce.TestSyncTable
Running org.apache.hadoop.hbase.mapreduce.TestCopyTable
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 45.022 sec - in 
org.apache.hadoop.hbase.mapreduce.TestMultithreadedTableMapper
Running org.apache.hadoop.hbase.mapreduce.TestMultiTableSnapshotInputFormat
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 27.701 sec - in 
org.apache.hadoop.hbase.mapreduce.TestSyncTable
Running org.apache.hadoop.hbase.mapreduce.TestMultiTableInputFormat
Running org.apache.hadoop.hbase.TestMetaTableAccessor
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 38.757 sec - in 
org.apache.hadoop.hbase.mapreduce.TestMultiTableSnapshotInputFormat
Running org.apache.hadoop.hbase.io.TestFileLink
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 58.536 sec - in 
org.apache.hadoop.hbase.mapreduce.TestCopyTable
Tests run: 15, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 26.531 sec - 
in org.apache.hadoop.hbase.TestMetaTableAccessor
Running org.apache.hadoop.hbase.io.hfile.TestHFileBlockIndex
Running org.apache.hadoop.hbase.io.hfile.TestSeekBeforeWithInlineBlocks
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 36.123 sec - in 
org.apache.hadoop.hbase.mapreduce.TestMultiTableInputFormat
Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 7.251 sec - in 
org.apache.hadoop.hbase.io.hfile.TestHFileBlockIndex
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 15.774 sec - in 
org.apache.hadoop.hbase.io.TestFileLink
Running org.apache.hadoop.hbase.io.hfile.TestHFileSeek
Running org.apache.hadoop.hbase.io.hfile.TestScannerSelectionUsingTTL
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.131 sec - in 
org.apache.hadoop.hbase.io.hfile.TestHFileSeek
Running org.apache.hadoop.hbase.io.hfile.TestHFileBlock
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 13.342 sec - in 
org.apache.hadoop.hbase.io.hfile.TestSeekBeforeWithInlineBlocks
Tests run: 8, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 123.5 sec - in 
org.apache.hadoop.hbase.mapreduce.TestTableSnapshotInputFormat
Running org.apache.hadoop.hbase.io.hfile.TestCacheConfig
Running org.apache.hadoop.hbase.io.hfile.TestScannerFromBucketCache
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.928 sec - in 
org.apache.hadoop.hbase.io.hfile.TestCacheConfig
Running org.apache.hadoop.hbase.io.hfile.TestForceCacheImportantBlocks
Running org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 10.515 sec - in 
org.apache.hadoop.hbase.io.hfile.TestScannerFromBucketCache
Running org.apache.hadoop.hbase.io.encoding.TestLoadAndSwitchEncodeOnDisk
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 33.533 sec - in 
org.apache.hadoop.hbase.io.hfile.TestForceCacheImportantBlocks
Running org.apache.hadoop.hbase.io.encoding.TestChangingEncoding
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 34.953 sec - in 
org.apache.hadoop.hbase.io.encoding.TestLoadAndSwitchEncodeOnDisk
Running org.apache.hadoop.hbase.io.encoding.TestEncodedSeekers
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 65.364 sec - in 
org.apache.hadoop.hbase.io.hfile.TestScannerSelectionUsingTTL
Running org.apache.hadoop.hbase.io.encoding.TestDataBlockEncoders
Tests run: 28, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 94.074 sec - 
in org.apache.hadoop.hbase.io.hfile.TestHFileBlock
Running org.apache.hadoop.hbase.io.encoding.TestBufferedDataBlockEncoder
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 12.7 sec - in 
org.apache.hadoop.hbase.io.encoding.TestBufferedDataBlockEncoder
Running org.apache.hadoop.hbase.io.asyncfs.TestFanOutOneBlockAsyncDFSOutput
Tests run: 56, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 54.086 sec - 
in org.apache.hadoop.hbase.io.encoding.TestDataBlockEncoders
Running org.apache.hadoop.hbase.io.asyncfs.TestSaslFanOutOneBlockAsyncDFSOutput
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 38.252 sec - in 
org.apache.hadoop.hbase.io.asyncfs.TestFanOutOneBlockAsyncDFSOutput
Tests run: 20, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 104.041 sec - 
in org.apache.hadoop.hbase.io.encoding.TestEncodedSeekers
Running org.apache.hadoop.hbase.filter.TestScanRowPrefix
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 5.726 sec - in 
org.apache.hadoop.hbase.filter.TestScanRowPrefix
Running org.apache.hadoop.hbase.filter.TestFilterWithScanLimits
Tests run: 72, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 164.052 sec - 
in org.apache.hadoop.hbase.io.hfile.TestCacheOnWrite
Running org.apache.hadoop.hbase.filter.TestFilterWrapper
Running org.apache.hadoop.hbase.filter.TestFuzzyRowFilterEndToEnd
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.235 sec - in 
org.apache.hadoop.hbase.filter.TestFilterWrapper
Running org.apache.hadoop.hbase.filter.TestFilterListOrOperatorWithBlkCnt
Tests run: 36, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 86.854 sec - 
in org.apache.hadoop.hbase.io.asyncfs.TestSaslFanOutOneBlockAsyncDFSOutput
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 15.788 sec - in 
org.apache.hadoop.hbase.filter.TestFilterListOrOperatorWithBlkCnt
Running org.apache.hadoop.hbase.filter.TestColumnRangeFilter
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 52.277 sec - in 
org.apache.hadoop.hbase.filter.TestFilterWithScanLimits
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 8.706 sec - in 
org.apache.hadoop.hbase.filter.TestColumnRangeFilter
Running org.apache.hadoop.hbase.filter.TestFuzzyRowAndColumnRangeFilter
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.652 sec - in 
org.apache.hadoop.hbase.filter.TestFuzzyRowAndColumnRangeFilter
Running org.apache.hadoop.hbase.filter.TestMultiRowRangeFilter
Tests run: 20, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 16.02 sec - in 
org.apache.hadoop.hbase.filter.TestMultiRowRangeFilter
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 207.872 sec - 
in org.apache.hadoop.hbase.io.encoding.TestChangingEncoding
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 282.687 sec - 
in org.apache.hadoop.hbase.filter.TestFuzzyRowFilterEndToEnd

Results :

Tests in error: 
  
TestLoadIncrementalHFilesUseSecurityEndPoint>TestLoadIncrementalHFiles.testRegionCrossingHFileSplitRowBloom:192->TestLoadIncrementalHFiles.testRegionCrossingHFileSplit:227->TestLoadIncrementalHFiles.runTest:263->TestLoadIncrementalHFiles.runTest:269->TestLoadIncrementalHFiles.runTest:295
 » ScannerTimeout
  
TestSimpleRegionNormalizerOnCluster.testRegionNormalizationMergeOnCluster:185->Object.wait:461->Object.wait:-2
 » TestTimedOut

Tests run: 2721, Failures: 0, Errors: 2, Skipped: 49

[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] Apache HBase ...................................... SUCCESS [  2.605 s]
[INFO] Apache HBase - Checkstyle ......................... SUCCESS [  0.693 s]
[INFO] Apache HBase - Resource Bundle .................... SUCCESS [  0.226 s]
[INFO] Apache HBase - Annotations ........................ SUCCESS [  0.241 s]
[INFO] Apache HBase - Protocol ........................... SUCCESS [  3.077 s]
[INFO] Apache HBase - Common ............................. SUCCESS [01:30 min]
[INFO] Apache HBase - Procedure .......................... SUCCESS [01:36 min]
[INFO] Apache HBase - Client ............................. SUCCESS [ 42.219 s]
[INFO] Apache HBase - Hadoop Compatibility ............... SUCCESS [  7.460 s]
[INFO] Apache HBase - Hadoop Two Compatibility ........... SUCCESS [  9.371 s]
[INFO] Apache HBase - Prefix Tree ........................ SUCCESS [ 10.688 s]
[INFO] Apache HBase - Server ............................. FAILURE [  01:26 h]
[INFO] Apache HBase - Testing Util ....................... SKIPPED
[INFO] Apache HBase - Thrift ............................. SKIPPED
[INFO] Apache HBase - RSGroup ............................ SKIPPED
[INFO] Apache HBase - Shell .............................. SKIPPED
[INFO] Apache HBase - Integration Tests .................. SKIPPED
[INFO] Apache HBase - Examples ........................... SKIPPED
[INFO] Apache HBase - Rest ............................... SKIPPED
[INFO] Apache HBase - External Block Cache ............... SKIPPED
[INFO] Apache HBase - Spark .............................. SKIPPED
[INFO] Apache HBase - Assembly ........................... SKIPPED
[INFO] Apache HBase - Shaded ............................. SKIPPED
[INFO] Apache HBase - Shaded - Client .................... SKIPPED
[INFO] Apache HBase - Shaded - Server .................... SKIPPED
[INFO] Apache HBase - Archetypes ......................... SKIPPED
[INFO] Apache HBase - Exemplar for hbase-client archetype  SKIPPED
[INFO] Apache HBase - Exemplar for hbase-shaded-client archetype  SKIPPED
[INFO] Apache HBase - Archetype builder .................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 01:30 h
[INFO] Finished at: 2016-06-25T11:12:22+00:00
[INFO] Final Memory: 71M/743M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal 
org.apache.maven.plugins:maven-surefire-plugin:2.18.1:test 
(secondPartTestsExecution) on project hbase-server: There are test failures.
[ERROR] 
[ERROR] Please refer to 
<https://builds.apache.org/job/HBase-Trunk_matrix/jdk=latest1.7,label=yahoo-not-h2/ws/hbase-server/target/surefire-reports>
 for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please 
read the following articles:
[ERROR] [Help 1] 
http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :hbase-server
Build step 'Invoke top-level Maven targets' marked build as failure
Performing Post build task...
Match found for :.* : True
Logical operation result is TRUE
Running script  : # Run zombie detector script
./dev-support/zombie-detector.sh --jenkins ${BUILD_ID}
[yahoo-not-h2] $ /bin/bash -xe /tmp/hudson241646143255635973.sh
+ ./dev-support/zombie-detector.sh --jenkins 1111
Sat Jun 25 11:12:23 UTC 2016 We're ok: there is no zombie test


    {color:green}+1 zombies{color}. No zombie tests found running at the end of 
the build.
POST BUILD TASK : SUCCESS
END OF POST BUILD TASK : 0
Archiving artifacts
Recording test results
[FINDBUGS] Skipping publisher since build result is FAILURE
[CHECKSTYLE] Skipping publisher since build result is FAILURE

Reply via email to