See <https://builds.apache.org/job/Phoenix-encode-columns/23/changes>

Changes:

[samarth] Fix test failure

------------------------------------------
[...truncated 1072 lines...]
Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE6
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1604)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1746)
        at 
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
        at java.lang.Thread.run(Thread.java:745)


testImportOneLocalIndexTable(org.apache.phoenix.end2end.CsvBulkLoadToolIT)  
Time elapsed: 2,407.411 sec  <<< ERROR!
org.apache.phoenix.exception.PhoenixIOException: callTimeout=1200000, 
callDuration=1202402: row '  TABLE5_IDX' on table 'SYSTEM.CATALOG' at 
region=SYSTEM.CATALOG,,1478639182330.f6c4adc266ec47261b7f9c5ca35b8af1., 
hostname=asf914.gq1.ygridcore.net,35948,1478639176615, seqNum=1
        at 
org.apache.phoenix.end2end.CsvBulkLoadToolIT.testImportOneIndexTable(CsvBulkLoadToolIT.java:309)
        at 
org.apache.phoenix.end2end.CsvBulkLoadToolIT.testImportOneLocalIndexTable(CsvBulkLoadToolIT.java:297)
Caused by: java.net.SocketTimeoutException: callTimeout=1200000, 
callDuration=1202402: row '  TABLE5_IDX' on table 'SYSTEM.CATALOG' at 
region=SYSTEM.CATALOG,,1478639182330.f6c4adc266ec47261b7f9c5ca35b8af1., 
hostname=asf914.gq1.ygridcore.net,35948,1478639176615, seqNum=1
Caused by: java.net.SocketTimeoutException: Call to 
asf914.gq1.ygridcore.net/67.195.81.158:35948 failed because 
java.net.SocketTimeoutException: 1200000 millis timeout while waiting for 
channel to be ready for read. ch : java.nio.channels.SocketChannel[connected 
local=/67.195.81.158:56306 remote=asf914.gq1.ygridcore.net/67.195.81.158:35948]
Caused by: java.net.SocketTimeoutException: 1200000 millis timeout while 
waiting for channel to be ready for read. ch : 
java.nio.channels.SocketChannel[connected local=/67.195.81.158:56306 
remote=asf914.gq1.ygridcore.net/67.195.81.158:35948]

testImportOneLocalIndexTable(org.apache.phoenix.end2end.CsvBulkLoadToolIT)  
Time elapsed: 2,407.412 sec  <<< ERROR!
org.apache.phoenix.exception.PhoenixIOException: callTimeout=1200000, 
callDuration=1222555: row '  TABLE5' on table 'SYSTEM.CATALOG' at 
region=SYSTEM.CATALOG,,1478639182330.f6c4adc266ec47261b7f9c5ca35b8af1., 
hostname=asf914.gq1.ygridcore.net,35948,1478639176615, seqNum=1
Caused by: java.net.SocketTimeoutException: callTimeout=1200000, 
callDuration=1222555: row '  TABLE5' on table 'SYSTEM.CATALOG' at 
region=SYSTEM.CATALOG,,1478639182330.f6c4adc266ec47261b7f9c5ca35b8af1., 
hostname=asf914.gq1.ygridcore.net,35948,1478639176615, seqNum=1
Caused by: java.io.IOException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1604)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1746)
        at 
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
        at java.lang.Thread.run(Thread.java:745)

Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1604)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1746)
        at 
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
        at java.lang.Thread.run(Thread.java:745)


testInvalidArguments(org.apache.phoenix.end2end.CsvBulkLoadToolIT)  Time 
elapsed: 1,199.97 sec  <<< ERROR!
org.apache.phoenix.exception.PhoenixIOException: callTimeout=1200000, 
callDuration=1221909: row '  TABLE5' on table 'SYSTEM.CATALOG' at 
region=SYSTEM.CATALOG,,1478639182330.f6c4adc266ec47261b7f9c5ca35b8af1., 
hostname=asf914.gq1.ygridcore.net,35948,1478639176615, seqNum=1
Caused by: java.net.SocketTimeoutException: callTimeout=1200000, 
callDuration=1221909: row '  TABLE5' on table 'SYSTEM.CATALOG' at 
region=SYSTEM.CATALOG,,1478639182330.f6c4adc266ec47261b7f9c5ca35b8af1., 
hostname=asf914.gq1.ygridcore.net,35948,1478639176615, seqNum=1
Caused by: java.io.IOException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1604)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1746)
        at 
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
        at java.lang.Thread.run(Thread.java:745)

Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1604)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1746)
        at 
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
        at java.lang.Thread.run(Thread.java:745)


testImportWithIndex(org.apache.phoenix.end2end.CsvBulkLoadToolIT)  Time 
elapsed: 1,205.306 sec  <<< ERROR!
org.apache.phoenix.exception.PhoenixIOException: callTimeout=1200000, 
callDuration=1222121: row '  TABLE5' on table 'SYSTEM.CATALOG' at 
region=SYSTEM.CATALOG,,1478639182330.f6c4adc266ec47261b7f9c5ca35b8af1., 
hostname=asf914.gq1.ygridcore.net,35948,1478639176615, seqNum=1
Caused by: java.net.SocketTimeoutException: callTimeout=1200000, 
callDuration=1222121: row '  TABLE5' on table 'SYSTEM.CATALOG' at 
region=SYSTEM.CATALOG,,1478639182330.f6c4adc266ec47261b7f9c5ca35b8af1., 
hostname=asf914.gq1.ygridcore.net,35948,1478639176615, seqNum=1
Caused by: java.io.IOException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1604)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1746)
        at 
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
        at java.lang.Thread.run(Thread.java:745)

Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1604)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1746)
        at 
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
        at java.lang.Thread.run(Thread.java:745)


testImportOneIndexTable(org.apache.phoenix.end2end.CsvBulkLoadToolIT)  Time 
elapsed: 1,203.85 sec  <<< ERROR!
org.apache.phoenix.exception.PhoenixIOException: callTimeout=1200000, 
callDuration=1221970: row '  TABLE5' on table 'SYSTEM.CATALOG' at 
region=SYSTEM.CATALOG,,1478639182330.f6c4adc266ec47261b7f9c5ca35b8af1., 
hostname=asf914.gq1.ygridcore.net,35948,1478639176615, seqNum=1
Caused by: java.net.SocketTimeoutException: callTimeout=1200000, 
callDuration=1221970: row '  TABLE5' on table 'SYSTEM.CATALOG' at 
region=SYSTEM.CATALOG,,1478639182330.f6c4adc266ec47261b7f9c5ca35b8af1., 
hostname=asf914.gq1.ygridcore.net,35948,1478639176615, seqNum=1
Caused by: java.io.IOException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1604)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1746)
        at 
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
        at java.lang.Thread.run(Thread.java:745)

Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException: 
java.io.IOException: Timed out waiting for lock for row: \x00\x00TABLE5
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLockInternal(HRegion.java:3804)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3766)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.getRowLock(HRegion.java:3830)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.acquireLock(MetaDataEndpointImpl.java:1604)
        at 
org.apache.phoenix.coprocessor.MetaDataEndpointImpl.dropTable(MetaDataEndpointImpl.java:1746)
        at 
org.apache.phoenix.coprocessor.generated.MetaDataProtos$MetaDataService.callMethod(MetaDataProtos.java:16297)
        at 
org.apache.hadoop.hbase.regionserver.HRegion.execService(HRegion.java:6041)
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.execServiceOnRegion(HRegionServer.java:3520)
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.execService(HRegionServer.java:3502)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31194)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2149)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:104)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
        at java.lang.Thread.run(Thread.java:745)



Results :

Tests in error: 
  
CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921
 » PhoenixIO
  
CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921
 » PhoenixIO
  
CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921
 » PhoenixIO
org.apache.phoenix.end2end.CsvBulkLoadToolIT.testImportOneLocalIndexTable(org.apache.phoenix.end2end.CsvBulkLoadToolIT)
  Run 1: 
CsvBulkLoadToolIT.testImportOneLocalIndexTable:297->testImportOneIndexTable:309 
» PhoenixIO
  Run 2: 
CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921
 » PhoenixIO

  
CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921
 » PhoenixIO
org.apache.phoenix.end2end.CsvBulkLoadToolIT.testImportWithLocalIndex(org.apache.phoenix.end2end.CsvBulkLoadToolIT)
  Run 1: CsvBulkLoadToolIT.testImportWithLocalIndex:258 » PhoenixIO 
callTimeout=1200000...
  Run 2: 
CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921
 » PhoenixIO

  
CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921
 » PhoenixIO
  
CsvBulkLoadToolIT>BaseOwnClusterIT.cleanUpAfterTest:35->BaseTest.deletePriorMetaData:857->BaseTest.deletePriorTables:865->BaseTest.deletePriorTables:876->BaseTest.deletePriorTables:921
 » PhoenixIO

Tests run: 207, Failures: 0, Errors: 8, Skipped: 17

[INFO] 
[INFO] --- maven-failsafe-plugin:2.19.1:verify (ParallelStatsEnabledTest) @ 
phoenix-core ---
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO] 
[INFO] Apache Phoenix ..................................... SUCCESS [  1.821 s]
[INFO] Phoenix Core ....................................... FAILURE [  03:58 h]
[INFO] Phoenix - Flume .................................... SKIPPED
[INFO] Phoenix - Pig ...................................... SKIPPED
[INFO] Phoenix Query Server Client ........................ SKIPPED
[INFO] Phoenix Query Server ............................... SKIPPED
[INFO] Phoenix - Pherf .................................... SKIPPED
[INFO] Phoenix - Spark .................................... SKIPPED
[INFO] Phoenix - Hive ..................................... SKIPPED
[INFO] Phoenix Client ..................................... SKIPPED
[INFO] Phoenix Server ..................................... SKIPPED
[INFO] Phoenix Assembly ................................... SKIPPED
[INFO] Phoenix - Tracing Web Application .................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 03:58 h
[INFO] Finished at: 2016-11-09T00:27:33+00:00
[INFO] Final Memory: 59M/1165M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal 
org.apache.maven.plugins:maven-failsafe-plugin:2.19.1:verify 
(ParallelStatsEnabledTest) on project phoenix-core: There are test failures.
[ERROR] 
[ERROR] Please refer to 
<https://builds.apache.org/job/Phoenix-encode-columns/ws/phoenix-core/target/failsafe-reports>
 for the individual test results.
[ERROR] -> [Help 1]
[ERROR] 
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e 
switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR] 
[ERROR] For more information about the errors and possible solutions, please 
read the following articles:
[ERROR] [Help 1] 
http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] 
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR]   mvn <goals> -rf :phoenix-core
Build step 'Invoke top-level Maven targets' marked build as failure
Archiving artifacts
Recording test results

Reply via email to