See <https://builds.apache.org/job/Phoenix-4.0-hadoop1/203/changes>
Changes: [mujtaba] PHOENIX-1090 - Fix HTable leak in ServerCacheClient (SamarthJain) ------------------------------------------ [...truncated 615 lines...] at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService$Stub.createTable(MetaDataProtos.java:6715) at org.apache.phoenix.query.ConnectionQueryServicesImpl$4.call(ConnectionQueryServicesImpl.java:1113) at org.apache.phoenix.query.ConnectionQueryServicesImpl$4.call(ConnectionQueryServicesImpl.java:1102) at org.apache.hadoop.hbase.client.HTable$16.call(HTable.java:1527) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) at java.util.concurrent.FutureTask.run(FutureTask.java:166) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:724) Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException: org.apache.hadoop.hbase.DoNotRetryIOException: ATABLE: Didn't find expected key values for table row in metadata row at org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:77) at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:833) at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:6581) at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:5493) at org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3300) at org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3282) at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:29501) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2012) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:98) at org.apache.hadoop.hbase.ipc.SimpleRpcScheduler.consumerLoop(SimpleRpcScheduler.java:160) at org.apache.hadoop.hbase.ipc.SimpleRpcScheduler.access$000(SimpleRpcScheduler.java:38) at org.apache.hadoop.hbase.ipc.SimpleRpcScheduler$1.run(SimpleRpcScheduler.java:110) at java.lang.Thread.run(Thread.java:724) Caused by: java.lang.IllegalStateException: Didn't find expected key values for table row in metadata row at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTable(MetaDataEndpointImpl.java:579) at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildTable(MetaDataEndpointImpl.java:415) at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.loadTable(MetaDataEndpointImpl.java:704) at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:782) ... 11 more at org.apache.hadoop.hbase.ipc.RpcClient.call(RpcClient.java:1450) at org.apache.hadoop.hbase.ipc.RpcClient.callBlockingMethod(RpcClient.java:1654) at org.apache.hadoop.hbase.ipc.RpcClient$BlockingRpcChannelImplementation.callBlockingMethod(RpcClient.java:1712) at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$BlockingStub.execService(ClientProtos.java:29924) at org.apache.hadoop.hbase.protobuf.ProtobufUtil.execService(ProtobufUtil.java:1508) at org.apache.hadoop.hbase.ipc.RegionCoprocessorRpcChannel$1.call(RegionCoprocessorRpcChannel.java:87) at org.apache.hadoop.hbase.ipc.RegionCoprocessorRpcChannel$1.call(RegionCoprocessorRpcChannel.java:84) at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:114) at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:90) at org.apache.hadoop.hbase.ipc.RegionCoprocessorRpcChannel.callExecService(RegionCoprocessorRpcChannel.java:90) at org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel.callMethod(CoprocessorRpcChannel.java:51) at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService$Stub.createTable(MetaDataProtos.java:6715) at org.apache.phoenix.query.ConnectionQueryServicesImpl$4.call(ConnectionQueryServicesImpl.java:1113) at org.apache.phoenix.query.ConnectionQueryServicesImpl$4.call(ConnectionQueryServicesImpl.java:1102) at org.apache.hadoop.hbase.client.HTable$16.call(HTable.java:1527) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) at java.util.concurrent.FutureTask.run(FutureTask.java:166) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:724) Running org.apache.phoenix.end2end.TenantSpecificViewIndexIT Tests run: 13, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 11.614 sec - in org.apache.phoenix.end2end.CSVCommonsLoaderIT Running org.apache.phoenix.end2end.StatementHintsIT Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.79 sec - in org.apache.phoenix.end2end.StatementHintsIT Running org.apache.phoenix.end2end.TenantSpecificViewIndexSaltedIT Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 10.93 sec - in org.apache.phoenix.end2end.TenantSpecificViewIndexIT Running org.apache.phoenix.end2end.TimezoneOffsetFunctionIT Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.695 sec - in org.apache.phoenix.end2end.TimezoneOffsetFunctionIT Running org.apache.phoenix.end2end.QueryExecWithoutSCNIT Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 6.195 sec - in org.apache.phoenix.end2end.TenantSpecificViewIndexSaltedIT Running org.apache.phoenix.end2end.LpadFunctionIT Tests run: 60, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 49.675 sec - in org.apache.phoenix.end2end.HashJoinIT Running org.apache.phoenix.end2end.LastValueFunctionIT Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 2.648 sec <<< FAILURE! - in org.apache.phoenix.end2end.QueryExecWithoutSCNIT testScanNoSCN(org.apache.phoenix.end2end.QueryExecWithoutSCNIT) Time elapsed: 1.245 sec <<< ERROR! org.apache.phoenix.exception.PhoenixIOException: org.apache.hadoop.hbase.DoNotRetryIOException: ATABLE: Didn't find expected key values for table row in metadata row at org.apache.phoenix.util.ServerUtil.parseServerException(ServerUtil.java:101) at org.apache.phoenix.query.ConnectionQueryServicesImpl.metaDataCoprocessorExec(ConnectionQueryServicesImpl.java:983) at org.apache.phoenix.query.ConnectionQueryServicesImpl.createTable(ConnectionQueryServicesImpl.java:1101) at org.apache.phoenix.query.DelegateConnectionQueryServices.createTable(DelegateConnectionQueryServices.java:114) at org.apache.phoenix.schema.MetaDataClient.createTableInternal(MetaDataClient.java:1170) at org.apache.phoenix.schema.MetaDataClient.createTable(MetaDataClient.java:422) at org.apache.phoenix.compile.CreateTableCompiler$2.execute(CreateTableCompiler.java:183) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:246) at org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:237) at org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:54) at org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:236) at org.apache.phoenix.jdbc.PhoenixStatement.execute(PhoenixStatement.java:952) at org.apache.phoenix.query.BaseTest.createTestTable(BaseTest.java:636) at org.apache.phoenix.query.BaseTest.createTestTable(BaseTest.java:610) at org.apache.phoenix.query.BaseTest.ensureTableCreated(BaseTest.java:602) at org.apache.phoenix.query.BaseTest.initATableValues(BaseTest.java:799) at org.apache.phoenix.end2end.QueryExecWithoutSCNIT.testScanNoSCN(QueryExecWithoutSCNIT.java:41) Caused by: org.apache.hadoop.hbase.DoNotRetryIOException: org.apache.hadoop.hbase.DoNotRetryIOException: ATABLE: Didn't find expected key values for table row in metadata row at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:526) at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:95) at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:79) at org.apache.hadoop.hbase.protobuf.ProtobufUtil.getRemoteException(ProtobufUtil.java:284) at org.apache.hadoop.hbase.protobuf.ProtobufUtil.execService(ProtobufUtil.java:1512) at org.apache.hadoop.hbase.ipc.RegionCoprocessorRpcChannel$1.call(RegionCoprocessorRpcChannel.java:87) at org.apache.hadoop.hbase.ipc.RegionCoprocessorRpcChannel$1.call(RegionCoprocessorRpcChannel.java:84) at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:114) at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:90) at org.apache.hadoop.hbase.ipc.RegionCoprocessorRpcChannel.callExecService(RegionCoprocessorRpcChannel.java:90) at org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel.callMethod(CoprocessorRpcChannel.java:51) at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService$Stub.createTable(MetaDataProtos.java:6715) at org.apache.phoenix.query.ConnectionQueryServicesImpl$4.call(ConnectionQueryServicesImpl.java:1113) at org.apache.phoenix.query.ConnectionQueryServicesImpl$4.call(ConnectionQueryServicesImpl.java:1102) at org.apache.hadoop.hbase.client.HTable$16.call(HTable.java:1527) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) at java.util.concurrent.FutureTask.run(FutureTask.java:166) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:724) Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException: org.apache.hadoop.hbase.DoNotRetryIOException: ATABLE: Didn't find expected key values for table row in metadata row at org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:77) at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:833) at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:6581) at org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:5493) at org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3300) at org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3282) at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:29501) at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2012) at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:98) at org.apache.hadoop.hbase.ipc.SimpleRpcScheduler.consumerLoop(SimpleRpcScheduler.java:160) at org.apache.hadoop.hbase.ipc.SimpleRpcScheduler.access$000(SimpleRpcScheduler.java:38) at org.apache.hadoop.hbase.ipc.SimpleRpcScheduler$1.run(SimpleRpcScheduler.java:110) at java.lang.Thread.run(Thread.java:724) Caused by: java.lang.IllegalStateException: Didn't find expected key values for table row in metadata row at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.getTable(MetaDataEndpointImpl.java:579) at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.buildTable(MetaDataEndpointImpl.java:415) at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.loadTable(MetaDataEndpointImpl.java:704) at org.apache.phoenix.coprocessor.MetaDataEndpointImpl.createTable(MetaDataEndpointImpl.java:782) ... 11 more at org.apache.hadoop.hbase.ipc.RpcClient.call(RpcClient.java:1450) at org.apache.hadoop.hbase.ipc.RpcClient.callBlockingMethod(RpcClient.java:1654) at org.apache.hadoop.hbase.ipc.RpcClient$BlockingRpcChannelImplementation.callBlockingMethod(RpcClient.java:1712) at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$BlockingStub.execService(ClientProtos.java:29924) at org.apache.hadoop.hbase.protobuf.ProtobufUtil.execService(ProtobufUtil.java:1508) at org.apache.hadoop.hbase.ipc.RegionCoprocessorRpcChannel$1.call(RegionCoprocessorRpcChannel.java:87) at org.apache.hadoop.hbase.ipc.RegionCoprocessorRpcChannel$1.call(RegionCoprocessorRpcChannel.java:84) at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:114) at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:90) at org.apache.hadoop.hbase.ipc.RegionCoprocessorRpcChannel.callExecService(RegionCoprocessorRpcChannel.java:90) at org.apache.hadoop.hbase.ipc.CoprocessorRpcChannel.callMethod(CoprocessorRpcChannel.java:51) at org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService$Stub.createTable(MetaDataProtos.java:6715) at org.apache.phoenix.query.ConnectionQueryServicesImpl$4.call(ConnectionQueryServicesImpl.java:1113) at org.apache.phoenix.query.ConnectionQueryServicesImpl$4.call(ConnectionQueryServicesImpl.java:1102) at org.apache.hadoop.hbase.client.HTable$16.call(HTable.java:1527) at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) at java.util.concurrent.FutureTask.run(FutureTask.java:166) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) at java.lang.Thread.run(Thread.java:724) Running org.apache.phoenix.end2end.salted.SaltedTableUpsertSelectIT Tests run: 11, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 5.818 sec - in org.apache.phoenix.end2end.LastValueFunctionIT Running org.apache.phoenix.end2end.DynamicFamilyIT Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 6.543 sec - in org.apache.phoenix.end2end.LpadFunctionIT Running org.apache.phoenix.end2end.SaltedViewIT Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 6.764 sec - in org.apache.phoenix.end2end.salted.SaltedTableUpsertSelectIT Running org.apache.phoenix.end2end.AlterTableIT Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 4.964 sec - in org.apache.phoenix.end2end.DynamicFamilyIT Running org.apache.phoenix.end2end.SpillableGroupByIT Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 5.308 sec - in org.apache.phoenix.end2end.SaltedViewIT Running org.apache.phoenix.end2end.ConvertTimezoneFunctionIT Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.258 sec - in org.apache.phoenix.end2end.SpillableGroupByIT Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.141 sec - in org.apache.phoenix.end2end.ConvertTimezoneFunctionIT Tests run: 13, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 29.182 sec - in org.apache.phoenix.end2end.AlterTableIT Results : Tests in error: ExecuteStatementsIT.testCharPadding:201->BaseTest.initATableValues:995 ? PhoenixIO ExecuteStatementsIT.testExecuteStatements:57->BaseTest.initATableValues:786->BaseTest.initATableValues:790->BaseTest.initATableValues:797->BaseTest.ensureTableCreated:593->BaseTest.ensureTableCreated:602->BaseTest.createTestTable:610->BaseTest.createTestTable:636 ? PhoenixIO QueryExecWithoutSCNIT.testScanNoSCN:41->BaseTest.initATableValues:799->BaseTest.ensureTableCreated:602->BaseTest.createTestTable:610->BaseTest.createTestTable:636 ? PhoenixIO Tests run: 334, Failures: 0, Errors: 3, Skipped: 0 [INFO] [INFO] --- maven-failsafe-plugin:2.17:integration-test (NeedTheirOwnClusterTests) @ phoenix-core --- [INFO] Failsafe report directory: <https://builds.apache.org/job/Phoenix-4.0-hadoop1/ws/phoenix-core/target/failsafe-reports> [INFO] parallel='none', perCoreThreadCount=true, threadCount=0, useUnlimitedThreads=false, threadCountSuites=0, threadCountClasses=0, threadCountMethods=0, parallelOptimized=true ------------------------------------------------------- T E S T S ------------------------------------------------------- ------------------------------------------------------- T E S T S ------------------------------------------------------- Running org.apache.phoenix.end2end.index.MutableIndexFailureIT Running org.apache.phoenix.end2end.ContextClassloaderIT Running org.apache.phoenix.mapreduce.CsvBulkLoadToolIT Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 8.936 sec - in org.apache.phoenix.end2end.index.MutableIndexFailureIT Running org.apache.phoenix.hbase.index.FailForUnsupportedHBaseVersionsIT Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 7.394 sec - in org.apache.phoenix.end2end.ContextClassloaderIT Running org.apache.phoenix.hbase.index.covered.example.EndToEndCoveredIndexingIT Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 18.632 sec - in org.apache.phoenix.hbase.index.FailForUnsupportedHBaseVersionsIT Running org.apache.phoenix.hbase.index.covered.example.EndtoEndIndexingWithCompressionIT Tests run: 11, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 59.149 sec - in org.apache.phoenix.hbase.index.covered.example.EndToEndCoveredIndexingIT Running org.apache.phoenix.hbase.index.covered.example.FailWithoutRetriesIT Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 9.986 sec - in org.apache.phoenix.hbase.index.covered.example.FailWithoutRetriesIT Tests run: 11, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 59.113 sec - in org.apache.phoenix.hbase.index.covered.example.EndtoEndIndexingWithCompressionIT Running org.apache.phoenix.hbase.index.covered.EndToEndCoveredColumnsIndexBuilderIT Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 9.177 sec - in org.apache.phoenix.hbase.index.covered.EndToEndCoveredColumnsIndexBuilderIT Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 173.969 sec - in org.apache.phoenix.mapreduce.CsvBulkLoadToolIT Results : Tests run: 35, Failures: 0, Errors: 0, Skipped: 0 [INFO] [INFO] --- maven-failsafe-plugin:2.17:verify (ClientManagedTimeTests) @ phoenix-core --- [INFO] Failsafe report directory: <https://builds.apache.org/job/Phoenix-4.0-hadoop1/ws/phoenix-core/target/failsafe-reports> [INFO] ------------------------------------------------------------------------ [INFO] Reactor Summary: [INFO] [INFO] Apache Phoenix .................................... SUCCESS [1.922s] [INFO] Phoenix Hadoop Compatibility ...................... SUCCESS [0.477s] [INFO] Phoenix Core ...................................... FAILURE [13:09.159s] [INFO] Phoenix - Flume ................................... SKIPPED [INFO] Phoenix - Pig ..................................... SKIPPED [INFO] Phoenix Hadoop1 Compatibility ..................... SKIPPED [INFO] Phoenix Assembly .................................. SKIPPED [INFO] ------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO] ------------------------------------------------------------------------ [INFO] Total time: 13:11.969s [INFO] Finished at: Tue Jul 15 21:59:33 UTC 2014 [INFO] Final Memory: 33M/389M [INFO] ------------------------------------------------------------------------ [ERROR] Failed to execute goal org.apache.maven.plugins:maven-failsafe-plugin:2.17:verify (ClientManagedTimeTests) on project phoenix-core: There are test failures. [ERROR] [ERROR] Please refer to <https://builds.apache.org/job/Phoenix-4.0-hadoop1/ws/phoenix-core/target/failsafe-reports> for the individual test results. [ERROR] -> [Help 1] [ERROR] [ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch. [ERROR] Re-run Maven using the -X switch to enable full debug logging. [ERROR] [ERROR] For more information about the errors and possible solutions, please read the following articles: [ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException [ERROR] [ERROR] After correcting the problems, you can resume the build with the command [ERROR] mvn <goals> -rf :phoenix-core Build step 'Invoke top-level Maven targets' marked build as failure Archiving artifacts Sending artifact delta relative to Phoenix | 4.0 | Hadoop1 #202 Archived 482 artifacts Archive block size is 32768 Received 2 blocks and 119452117 bytes Compression is 0.1% Took 39 sec Recording test results