See <https://builds.apache.org/job/Phoenix-master-hadoop1/383/changes>
Changes:
[larsh] PHOENIX-1281 Each MultiKeyValueTuple.setKeyValues creates a new
immutable list object.
------------------------------------------
[...truncated 653 lines...]
at
org.apache.phoenix.util.ServerUtil.parseServerException(ServerUtil.java:101)
at
org.apache.phoenix.iterate.TableResultIterator.<init>(TableResultIterator.java:57)
at
org.apache.phoenix.iterate.ParallelIterators$3.call(ParallelIterators.java:362)
at
org.apache.phoenix.iterate.ParallelIterators$3.call(ParallelIterators.java:357)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
at java.util.concurrent.FutureTask.run(FutureTask.java:166)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:724)
Caused by: org.apache.hadoop.hbase.DoNotRetryIOException:
org.apache.hadoop.hbase.DoNotRetryIOException:
_LOCAL_IDX_T,e\x00\x00\x00\x00\x00\x00\x00\x00\x00,1411660419823.d785014a3005655ccdd68d85075dd73e.:
Requested memory of 21196 bytes could not be allocated from remaining memory
of 21196 bytes from global pool of 40000 bytes after waiting for 0ms.
at
org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:77)
at
org.apache.phoenix.util.ServerUtil.throwIOException(ServerUtil.java:45)
at
org.apache.phoenix.coprocessor.BaseScannerRegionObserver.postScannerOpen(BaseScannerRegionObserver.java:152)
at
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.postScannerOpen(RegionCoprocessorHost.java:1845)
at
org.apache.hadoop.hbase.regionserver.HRegionServer.scan(HRegionServer.java:3092)
at
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:29497)
at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2027)
at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:98)
at
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:114)
at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:94)
at java.lang.Thread.run(Thread.java:724)
Caused by: org.apache.phoenix.memory.InsufficientMemoryException: Requested
memory of 21196 bytes could not be allocated from remaining memory of 21196
bytes from global pool of 40000 bytes after waiting for 0ms.
at
org.apache.phoenix.memory.GlobalMemoryManager.allocateBytes(GlobalMemoryManager.java:81)
at
org.apache.phoenix.memory.GlobalMemoryManager.allocate(GlobalMemoryManager.java:100)
at
org.apache.phoenix.memory.GlobalMemoryManager.allocate(GlobalMemoryManager.java:106)
at
org.apache.phoenix.cache.aggcache.SpillableGroupByCache.<init>(SpillableGroupByCache.java:150)
at
org.apache.phoenix.coprocessor.GroupedAggregateRegionObserver$GroupByCacheFactory.newCache(GroupedAggregateRegionObserver.java:362)
at
org.apache.phoenix.coprocessor.GroupedAggregateRegionObserver.scanUnordered(GroupedAggregateRegionObserver.java:397)
at
org.apache.phoenix.coprocessor.GroupedAggregateRegionObserver.doPostScannerOpen(GroupedAggregateRegionObserver.java:160)
at
org.apache.phoenix.coprocessor.BaseScannerRegionObserver.postScannerOpen(BaseScannerRegionObserver.java:134)
... 8 more
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
at
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
at
org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
at
org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:95)
at
org.apache.hadoop.hbase.protobuf.ProtobufUtil.getRemoteException(ProtobufUtil.java:285)
at
org.apache.hadoop.hbase.client.ScannerCallable.openScanner(ScannerCallable.java:316)
at
org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:164)
at
org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:59)
at
org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:114)
at
org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:90)
at
org.apache.hadoop.hbase.client.ClientScanner.nextScanner(ClientScanner.java:282)
at
org.apache.hadoop.hbase.client.ClientScanner.initializeScannerInConstruction(ClientScanner.java:187)
at
org.apache.hadoop.hbase.client.ClientScanner.<init>(ClientScanner.java:182)
at
org.apache.hadoop.hbase.client.ClientScanner.<init>(ClientScanner.java:109)
at org.apache.hadoop.hbase.client.HTable.getScanner(HTable.java:738)
at
org.apache.phoenix.iterate.TableResultIterator.<init>(TableResultIterator.java:54)
at
org.apache.phoenix.iterate.ParallelIterators$3.call(ParallelIterators.java:362)
at
org.apache.phoenix.iterate.ParallelIterators$3.call(ParallelIterators.java:357)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
at java.util.concurrent.FutureTask.run(FutureTask.java:166)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:724)
Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException:
org.apache.hadoop.hbase.DoNotRetryIOException:
_LOCAL_IDX_T,e\x00\x00\x00\x00\x00\x00\x00\x00\x00,1411660419823.d785014a3005655ccdd68d85075dd73e.:
Requested memory of 21196 bytes could not be allocated from remaining memory
of 21196 bytes from global pool of 40000 bytes after waiting for 0ms.
at
org.apache.phoenix.util.ServerUtil.createIOException(ServerUtil.java:77)
at
org.apache.phoenix.util.ServerUtil.throwIOException(ServerUtil.java:45)
at
org.apache.phoenix.coprocessor.BaseScannerRegionObserver.postScannerOpen(BaseScannerRegionObserver.java:152)
at
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.postScannerOpen(RegionCoprocessorHost.java:1845)
at
org.apache.hadoop.hbase.regionserver.HRegionServer.scan(HRegionServer.java:3092)
at
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:29497)
at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2027)
at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:98)
at
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:114)
at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:94)
at java.lang.Thread.run(Thread.java:724)
Caused by: org.apache.phoenix.memory.InsufficientMemoryException: Requested
memory of 21196 bytes could not be allocated from remaining memory of 21196
bytes from global pool of 40000 bytes after waiting for 0ms.
at
org.apache.phoenix.memory.GlobalMemoryManager.allocateBytes(GlobalMemoryManager.java:81)
at
org.apache.phoenix.memory.GlobalMemoryManager.allocate(GlobalMemoryManager.java:100)
at
org.apache.phoenix.memory.GlobalMemoryManager.allocate(GlobalMemoryManager.java:106)
at
org.apache.phoenix.cache.aggcache.SpillableGroupByCache.<init>(SpillableGroupByCache.java:150)
at
org.apache.phoenix.coprocessor.GroupedAggregateRegionObserver$GroupByCacheFactory.newCache(GroupedAggregateRegionObserver.java:362)
at
org.apache.phoenix.coprocessor.GroupedAggregateRegionObserver.scanUnordered(GroupedAggregateRegionObserver.java:397)
at
org.apache.phoenix.coprocessor.GroupedAggregateRegionObserver.doPostScannerOpen(GroupedAggregateRegionObserver.java:160)
at
org.apache.phoenix.coprocessor.BaseScannerRegionObserver.postScannerOpen(BaseScannerRegionObserver.java:134)
... 8 more
at org.apache.hadoop.hbase.ipc.RpcClient.call(RpcClient.java:1452)
at
org.apache.hadoop.hbase.ipc.RpcClient.callBlockingMethod(RpcClient.java:1656)
at
org.apache.hadoop.hbase.ipc.RpcClient$BlockingRpcChannelImplementation.callBlockingMethod(RpcClient.java:1714)
at
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$BlockingStub.scan(ClientProtos.java:29900)
at
org.apache.hadoop.hbase.client.ScannerCallable.openScanner(ScannerCallable.java:308)
at
org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:164)
at
org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:59)
at
org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:114)
at
org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:90)
at
org.apache.hadoop.hbase.client.ClientScanner.nextScanner(ClientScanner.java:282)
at
org.apache.hadoop.hbase.client.ClientScanner.initializeScannerInConstruction(ClientScanner.java:187)
at
org.apache.hadoop.hbase.client.ClientScanner.<init>(ClientScanner.java:182)
at
org.apache.hadoop.hbase.client.ClientScanner.<init>(ClientScanner.java:109)
at org.apache.hadoop.hbase.client.HTable.getScanner(HTable.java:738)
at
org.apache.phoenix.iterate.TableResultIterator.<init>(TableResultIterator.java:54)
at
org.apache.phoenix.iterate.ParallelIterators$3.call(ParallelIterators.java:362)
at
org.apache.phoenix.iterate.ParallelIterators$3.call(ParallelIterators.java:357)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
at java.util.concurrent.FutureTask.run(FutureTask.java:166)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:724)
Running org.apache.phoenix.end2end.DecodeFunctionIT
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.223 sec - in
org.apache.phoenix.end2end.GuidePostsLifeCycleIT
Running org.apache.phoenix.end2end.MD5FunctionIT
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 7.375 sec - in
org.apache.phoenix.end2end.PhoenixEncodeDecodeIT
Running org.apache.phoenix.end2end.UpsertSelectAutoCommitIT
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.378 sec - in
org.apache.phoenix.end2end.DecodeFunctionIT
Running org.apache.phoenix.end2end.ExecuteStatementsIT
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.728 sec - in
org.apache.phoenix.end2end.MD5FunctionIT
Running org.apache.phoenix.end2end.HashJoinIT
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 8.753 sec - in
org.apache.phoenix.end2end.UpsertSelectAutoCommitIT
Running org.apache.phoenix.end2end.TenantSpecificViewIndexIT
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 9.352 sec - in
org.apache.phoenix.end2end.ExecuteStatementsIT
Running org.apache.phoenix.end2end.CoalesceFunctionIT
Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 7.045 sec - in
org.apache.phoenix.end2end.CoalesceFunctionIT
Running org.apache.phoenix.end2end.TimezoneOffsetFunctionIT
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.74 sec - in
org.apache.phoenix.end2end.TimezoneOffsetFunctionIT
Running org.apache.phoenix.end2end.ArithmeticQueryIT
Tests run: 26, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 117.821 sec -
in org.apache.phoenix.end2end.index.MutableIndexIT
Running org.apache.phoenix.end2end.ReverseFunctionIT
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 13.684 sec - in
org.apache.phoenix.end2end.TenantSpecificViewIndexIT
Running org.apache.phoenix.end2end.RegexpSplitFunctionIT
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.743 sec - in
org.apache.phoenix.end2end.ReverseFunctionIT
Running org.apache.phoenix.end2end.SaltedViewIT
Tests run: 11, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.831 sec - in
org.apache.phoenix.end2end.RegexpSplitFunctionIT
Running org.apache.phoenix.end2end.FirstValueFunctionIT
Tests run: 7, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.796 sec - in
org.apache.phoenix.end2end.FirstValueFunctionIT
Running org.apache.phoenix.end2end.QueryPlanIT
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 8.61 sec - in
org.apache.phoenix.end2end.SaltedViewIT
Running org.apache.phoenix.end2end.QueryExecWithoutSCNIT
Tests run: 20, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 14.02 sec - in
org.apache.phoenix.end2end.ArithmeticQueryIT
Running org.apache.phoenix.end2end.ConvertTimezoneFunctionIT
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.361 sec - in
org.apache.phoenix.end2end.QueryExecWithoutSCNIT
Running org.apache.phoenix.end2end.StatementHintsIT
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 8.148 sec - in
org.apache.phoenix.end2end.QueryPlanIT
Running org.apache.phoenix.end2end.UpsertBigValuesIT
Tests run: 6, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.019 sec - in
org.apache.phoenix.end2end.ConvertTimezoneFunctionIT
Running org.apache.phoenix.end2end.salted.SaltedTableUpsertSelectIT
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.879 sec - in
org.apache.phoenix.end2end.StatementHintsIT
Tests run: 25, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 83.763 sec -
in org.apache.phoenix.end2end.InListIT
Running org.apache.phoenix.end2end.SortOrderFIT
Running org.apache.phoenix.end2end.QueryMoreIT
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 9.141 sec - in
org.apache.phoenix.end2end.UpsertBigValuesIT
Running org.apache.phoenix.end2end.ReverseScanIT
Tests run: 30, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 7.781 sec - in
org.apache.phoenix.end2end.SortOrderFIT
Running org.apache.phoenix.end2end.RegexpSubstrFunctionIT
Tests run: 5, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 10.161 sec - in
org.apache.phoenix.end2end.salted.SaltedTableUpsertSelectIT
Running org.apache.phoenix.end2end.ServerExceptionIT
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 3.193 sec - in
org.apache.phoenix.end2end.ReverseScanIT
Running org.apache.phoenix.end2end.AutoCommitIT
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.529 sec - in
org.apache.phoenix.end2end.RegexpSubstrFunctionIT
Running org.apache.phoenix.end2end.LastValueFunctionIT
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.421 sec - in
org.apache.phoenix.end2end.ServerExceptionIT
Running org.apache.phoenix.end2end.RoundFloorCeilFunctionsEnd2EndIT
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 2.908 sec - in
org.apache.phoenix.end2end.AutoCommitIT
Running org.apache.phoenix.end2end.LpadFunctionIT
Tests run: 11, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 6.07 sec - in
org.apache.phoenix.end2end.LastValueFunctionIT
Tests run: 30, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 6.583 sec - in
org.apache.phoenix.end2end.RoundFloorCeilFunctionsEnd2EndIT
Tests run: 10, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 8.323 sec - in
org.apache.phoenix.end2end.LpadFunctionIT
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 35.572 sec - in
org.apache.phoenix.end2end.QueryMoreIT
Tests run: 96, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 95.024 sec -
in org.apache.phoenix.end2end.HashJoinIT
Results :
Tests in error:
LocalIndexIT.testLocalIndexScanJoinColumnsFromDataTable:439 ? PhoenixIO
org.ap...
Tests run: 490, Failures: 0, Errors: 1, Skipped: 0
[INFO]
[INFO] --- maven-failsafe-plugin:2.17:integration-test
(NeedTheirOwnClusterTests) @ phoenix-core ---
[INFO] Failsafe report directory:
<https://builds.apache.org/job/Phoenix-master-hadoop1/ws/phoenix-core/target/failsafe-reports>
[INFO] parallel='none', perCoreThreadCount=true, threadCount=0,
useUnlimitedThreads=false, threadCountSuites=0, threadCountClasses=0,
threadCountMethods=0, parallelOptimized=true
-------------------------------------------------------
T E S T S
-------------------------------------------------------
-------------------------------------------------------
T E S T S
-------------------------------------------------------
Running
org.apache.phoenix.hbase.index.covered.example.EndtoEndIndexingWithCompressionIT
Running org.apache.phoenix.hbase.index.covered.example.EndToEndCoveredIndexingIT
Running org.apache.phoenix.hbase.index.covered.example.FailWithoutRetriesIT
Running
org.apache.phoenix.hbase.index.covered.EndToEndCoveredColumnsIndexBuilderIT
Running org.apache.phoenix.hbase.index.balancer.IndexLoadBalancerIT
Tests run: 1, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 7.653 sec - in
org.apache.phoenix.hbase.index.covered.example.FailWithoutRetriesIT
Running org.apache.phoenix.hbase.index.FailForUnsupportedHBaseVersionsIT
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 7.694 sec - in
org.apache.phoenix.hbase.index.covered.EndToEndCoveredColumnsIndexBuilderIT
Running org.apache.phoenix.end2end.index.MutableIndexFailureIT
Tests run: 2, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 16.255 sec - in
org.apache.phoenix.hbase.index.FailForUnsupportedHBaseVersionsIT
Running org.apache.phoenix.end2end.ContextClassloaderIT
Tests run: 11, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 48.137 sec -
in
org.apache.phoenix.hbase.index.covered.example.EndtoEndIndexingWithCompressionIT
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 6.699 sec - in
org.apache.phoenix.end2end.ContextClassloaderIT
Tests run: 11, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 57.985 sec -
in org.apache.phoenix.hbase.index.covered.example.EndToEndCoveredIndexingIT
Running org.apache.phoenix.mapreduce.CsvBulkLoadToolIT
Tests run: 9, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 132.574 sec -
in org.apache.phoenix.hbase.index.balancer.IndexLoadBalancerIT
Tests run: 3, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 165.72 sec - in
org.apache.phoenix.end2end.index.MutableIndexFailureIT
Tests run: 4, Failures: 0, Errors: 0, Skipped: 0, Time elapsed: 181.118 sec -
in org.apache.phoenix.mapreduce.CsvBulkLoadToolIT
Results :
Tests run: 46, Failures: 0, Errors: 0, Skipped: 0
[INFO]
[INFO] --- maven-failsafe-plugin:2.17:verify (ClientManagedTimeTests) @
phoenix-core ---
[INFO] Failsafe report directory:
<https://builds.apache.org/job/Phoenix-master-hadoop1/ws/phoenix-core/target/failsafe-reports>
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Apache Phoenix .................................... SUCCESS [4.017s]
[INFO] Phoenix Hadoop Compatibility ...................... SUCCESS [6.230s]
[INFO] Phoenix Hadoop2 Compatibility ..................... SUCCESS [4.816s]
[INFO] Phoenix Core ...................................... FAILURE [14:41.953s]
[INFO] Phoenix - Flume ................................... SKIPPED
[INFO] Phoenix - Pig ..................................... SKIPPED
[INFO] Phoenix Assembly .................................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 14:58.052s
[INFO] Finished at: Thu Sep 25 16:00:49 UTC 2014
[INFO] Final Memory: 53M/968M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal
org.apache.maven.plugins:maven-failsafe-plugin:2.17:verify
(ClientManagedTimeTests) on project phoenix-core: There are test failures.
[ERROR]
[ERROR] Please refer to
<https://builds.apache.org/job/Phoenix-master-hadoop1/ws/phoenix-core/target/failsafe-reports>
for the individual test results.
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please
read the following articles:
[ERROR] [Help 1]
http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :phoenix-core
Build step 'Invoke top-level Maven targets' marked build as failure
Archiving artifacts
Sending artifact delta relative to Phoenix | Master | Hadoop1 #381
Archived 706 artifacts
Archive block size is 32768
Received 5008 blocks and 258971391 bytes
Compression is 38.8%
Took 2 min 11 sec
Updating PHOENIX-1281
Recording test results