[ 
https://issues.apache.org/jira/browse/PHOENIX-3163?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16473653#comment-16473653
 ] 

Pedro Boado commented on PHOENIX-3163:
--------------------------------------

Hi [~sergey.soldatov] 

I've just noticed that SkipScanAfterManualSplitIT.testManualSplit started 
failing in branch 4.x-HBase-1.2 .
{code:java}
2018-05-13 23:41:03,729 DEBUG [B.defaultRpcServer.handler=1,queue=0,port=43819] 
org.apache.hadoop.hbase.ipc.CallRunner(115): 
B.defaultRpcServer.handler=1,queue=0,port=43819: callId: 465 service: 
ClientService methodName: Scan size: 595 connection: 127.0.0.1:57462
org.apache.hadoop.hbase.NotServingRegionException: Region 
T000002,\x01,1526251248024.34b289cddcb2b99d8e776602b796d731. is not online on 
xps,43819,1526251221783
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.getRegionByEncodedName(HRegionServer.java:2942)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.getRegion(RSRpcServices.java:1072)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2410)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33648)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2188)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
        at java.lang.Thread.run(Thread.java:745)
2018-05-13 23:41:03,729 DEBUG [B.defaultRpcServer.handler=2,queue=0,port=43819] 
org.apache.hadoop.hbase.ipc.CallRunner(115): 
B.defaultRpcServer.handler=2,queue=0,port=43819: callId: 467 service: 
ClientService methodName: Scan size: 586 connection: 127.0.0.1:57462
org.apache.hadoop.hbase.NotServingRegionException: Region 
T000002,\x02,1526251248024.9201bdc4f44225f390edb40ab1548a82. is not online on 
xps,43819,1526251221783
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.getRegionByEncodedName(HRegionServer.java:2942)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.getRegion(RSRpcServices.java:1072)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2410)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33648)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2188)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
        at java.lang.Thread.run(Thread.java:745)
2018-05-13 23:41:03,729 DEBUG [B.defaultRpcServer.handler=0,queue=0,port=43819] 
org.apache.hadoop.hbase.ipc.CallRunner(115): 
B.defaultRpcServer.handler=0,queue=0,port=43819: callId: 466 service: 
ClientService methodName: Scan size: 585 connection: 127.0.0.1:57462
org.apache.hadoop.hbase.NotServingRegionException: Region 
T000002,,1526251248024.9bb19fa73f91248dd407192c4ce512fe. is not online on 
xps,43819,1526251221783
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.getRegionByEncodedName(HRegionServer.java:2942)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.getRegion(RSRpcServices.java:1072)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2410)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33648)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2188)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
        at java.lang.Thread.run(Thread.java:745)
2018-05-13 23:41:03,729 DEBUG [B.defaultRpcServer.handler=3,queue=0,port=43819] 
org.apache.hadoop.hbase.ipc.CallRunner(115): 
B.defaultRpcServer.handler=3,queue=0,port=43819: callId: 468 service: 
ClientService methodName: Scan size: 595 connection: 127.0.0.1:57462
org.apache.hadoop.hbase.NotServingRegionException: Region 
T000002,\x03,1526251248024.6944b7e5e33cdcbdcc674c745ad8c1a5. is not online on 
xps,43819,1526251221783
        at 
org.apache.hadoop.hbase.regionserver.HRegionServer.getRegionByEncodedName(HRegionServer.java:2942)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.getRegion(RSRpcServices.java:1072)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2410)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33648)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2188)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
        at java.lang.Thread.run(Thread.java:745)
2018-05-13 23:41:03,790 DEBUG [RS:0;xps:43819-splits-1526251252838] 
org.apache.hadoop.hbase.coprocessor.CoprocessorHost(182): Loading coprocessor 
class org.apache.phoenix.hbase.index.Indexer with path null and priority 
805306366
2018-05-13 23:41:03,826 DEBUG [RS:0;xps:43819-splits-1526251252838] 
org.apache.phoenix.hbase.index.Indexer(244): Setting up recovery writter with 
failure policy: class 
org.apache.phoenix.hbase.index.write.recovery.StoreFailuresInCachePolicy
2018-05-13 23:41:03,826 INFO  [RS:0;xps:43819-splits-1526251252838] 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost(367): Loaded 
coprocessor org.apache.phoenix.hbase.index.Indexer from HTD of T000002 
successfully.
2018-05-13 23:41:03,826 DEBUG [RS:0;xps:43819-splits-1526251252838] 
org.apache.hadoop.hbase.coprocessor.CoprocessorHost(182): Loading coprocessor 
class org.apache.phoenix.coprocessor.ServerCachingEndpointImpl with path null 
and priority 805306366
2018-05-13 23:41:03,826 DEBUG [RS:0;xps:43819-splits-1526251252838] 
org.apache.hadoop.hbase.regionserver.HRegion(7836): Registered coprocessor 
service: region=T000002,\x01sp,1526251261881.bff1498b104381411dbcc5c1b104e0e8. 
service=ServerCachingService
2018-05-13 23:41:03,826 INFO  [RS:0;xps:43819-splits-1526251252838] 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost(367): Loaded 
coprocessor org.apache.phoenix.coprocessor.ServerCachingEndpointImpl from HTD 
of T000002 successfully.
2018-05-13 23:41:03,826 DEBUG [RS:0;xps:43819-splits-1526251252838] 
org.apache.hadoop.hbase.coprocessor.CoprocessorHost(182): Loading coprocessor 
class org.apache.phoenix.coprocessor.GroupedAggregateRegionObserver with path 
null and priority 805306366
2018-05-13 23:41:03,827 INFO  [RS:0;xps:43819-splits-1526251252838] 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost(367): Loaded 
coprocessor org.apache.phoenix.coprocessor.GroupedAggregateRegionObserver from 
HTD of T000002 successfully.
2018-05-13 23:41:03,827 DEBUG [RS:0;xps:43819-splits-1526251252838] 
org.apache.hadoop.hbase.coprocessor.CoprocessorHost(182): Loading coprocessor 
class org.apache.phoenix.coprocessor.UngroupedAggregateRegionObserver with path 
null and priority 805306366
2018-05-13 23:41:03,847 DEBUG [B.defaultRpcServer.handler=4,queue=0,port=43819] 
org.apache.hadoop.hbase.ipc.CallRunner(115): 
B.defaultRpcServer.handler=4,queue=0,port=43819: callId: 472 service: 
ClientService methodName: Scan size: 595 connection: 127.0.0.1:57462
org.apache.hadoop.hbase.DoNotRetryIOException: ERROR 1108 (XCL08): Cache of 
region boundaries are out of date. tableName=T000002
        at 
org.apache.phoenix.coprocessor.BaseScannerRegionObserver.throwIfScanOutOfRegion(BaseScannerRegionObserver.java:175)
        at 
org.apache.phoenix.coprocessor.BaseScannerRegionObserver.preScannerOpen(BaseScannerRegionObserver.java:203)
        at 
org.apache.phoenix.coprocessor.UngroupedAggregateRegionObserver.preScannerOpen(UngroupedAggregateRegionObserver.java:327)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$50.call(RegionCoprocessorHost.java:1300)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$RegionOperation.call(RegionCoprocessorHost.java:1673)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1749)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperationWithResult(RegionCoprocessorHost.java:1722)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preScannerOpen(RegionCoprocessorHost.java:1295)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2428)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33648)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2188)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
        at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.phoenix.schema.StaleRegionBoundaryCacheException: ERROR 
1108 (XCL08): Cache of region boundaries are out of date. tableName=T000002
        at 
org.apache.phoenix.coprocessor.BaseScannerRegionObserver.throwIfScanOutOfRegion(BaseScannerRegionObserver.java:174)
        ... 14 more
2018-05-13 23:41:03,851 DEBUG [B.defaultRpcServer.handler=0,queue=0,port=43819] 
org.apache.phoenix.coprocessor.UngroupedAggregateRegionObserver(528): Starting 
ungrouped coprocessor scan 
{"timeRange":[0,9223372036854775807],"batch":-1,"startRow":"\\x00 
c","stopRow":"\\x00 
c\\x00","loadColumnFamiliesOnDemand":true,"totalColumns":1,"cacheBlocks":true,"families":{"0":["ALL"]},"maxResultSize":2097152,"maxVersions":1,"filter":"FilterList
 AND (2/2): [FirstKeyOnlyFilter, SkipScanFilter [[\\x00 
c]]]","caching":2147483647} {ENCODED => ed285ee58edd90670656b3753dd6a3fd, NAME 
=> 'T000002,,1526251252848.ed285ee58edd90670656b3753dd6a3fd.', STARTKEY => '', 
ENDKEY => '\x00jj'}
2018-05-13 23:41:03,851 DEBUG [B.defaultRpcServer.handler=0,queue=0,port=43819] 
org.apache.phoenix.coprocessor.UngroupedAggregateRegionObserver(822): Finished 
scanning 0 rows for ungrouped coprocessor scan 
{"timeRange":[0,9223372036854775807],"batch":-1,"startRow":"\\x00 
c","stopRow":"\\x00 
c\\x00","loadColumnFamiliesOnDemand":true,"totalColumns":1,"cacheBlocks":true,"families":{"0":["ALL"]},"maxResultSize":2097152,"maxVersions":1,"filter":"FilterList
 AND (2/2): [FirstKeyOnlyFilter, SkipScanFilter [[\\x00 
c]]]","caching":2147483647}
2018-05-13 23:41:03,852 DEBUG [B.defaultRpcServer.handler=2,queue=0,port=43819] 
org.apache.hadoop.hbase.ipc.CallRunner(115): 
B.defaultRpcServer.handler=2,queue=0,port=43819: callId: 475 service: 
ClientService methodName: Scan size: 595 connection: 127.0.0.1:57462
org.apache.hadoop.hbase.DoNotRetryIOException: ERROR 1108 (XCL08): Cache of 
region boundaries are out of date. tableName=T000002
        at 
org.apache.phoenix.coprocessor.BaseScannerRegionObserver.throwIfScanOutOfRegion(BaseScannerRegionObserver.java:175)
        at 
org.apache.phoenix.coprocessor.BaseScannerRegionObserver.preScannerOpen(BaseScannerRegionObserver.java:203)
        at 
org.apache.phoenix.coprocessor.UngroupedAggregateRegionObserver.preScannerOpen(UngroupedAggregateRegionObserver.java:327)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$50.call(RegionCoprocessorHost.java:1300)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$RegionOperation.call(RegionCoprocessorHost.java:1673)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1749)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperationWithResult(RegionCoprocessorHost.java:1722)
        at 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preScannerOpen(RegionCoprocessorHost.java:1295)
        at 
org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2428)
        at 
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33648)
        at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2188)
        at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
        at 
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
        at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
        at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.phoenix.schema.StaleRegionBoundaryCacheException: ERROR 
1108 (XCL08): Cache of region boundaries are out of date. tableName=T000002
        at 
org.apache.phoenix.coprocessor.BaseScannerRegionObserver.throwIfScanOutOfRegion(BaseScannerRegionObserver.java:174)
        ... 14 more
2018-05-13 23:41:03,853 DEBUG [B.defaultRpcServer.handler=3,queue=0,port=43819] 
org.apache.phoenix.coprocessor.UngroupedAggregateRegionObserver(528): Starting 
ungrouped coprocessor scan 
{"timeRange":[0,9223372036854775807],"batch":-1,"startRow":"\\x02 
a","stopRow":"\\x02 
a\\x00","loadColumnFamiliesOnDemand":true,"totalColumns":1,"cacheBlocks":true,"families":{"0":["ALL"]},"maxResultSize":2097152,"maxVersions":1,"filter":"FilterList
 AND (2/2): [FirstKeyOnlyFilter, SkipScanFilter [[\\x02 
a]]]","caching":2147483647} {ENCODED => c5c9386b8657bed3abfba259adc54cf5, NAME 
=> 'T000002,\x02,1526251257604.c5c9386b8657bed3abfba259adc54cf5.', STARTKEY => 
'\x02', ENDKEY => '\x02jh'}
2018-05-13 23:41:03,854 DEBUG [B.defaultRpcServer.handler=3,queue=0,port=43819] 
org.apache.phoenix.coprocessor.UngroupedAggregateRegionObserver(822): Finished 
scanning 0 rows for ungrouped coprocessor scan 
{"timeRange":[0,9223372036854775807],"batch":-1,"startRow":"\\x02 
a","stopRow":"\\x02 
a\\x00","loadColumnFamiliesOnDemand":true,"totalColumns":1,"cacheBlocks":true,"families":{"0":["ALL"]},"maxResultSize":2097152,"maxVersions":1,"filter":"FilterList
 AND (2/2): [FirstKeyOnlyFilter, SkipScanFilter [[\\x02 
a]]]","caching":2147483647}
2018-05-13 23:41:03,866 INFO  [RS:0;xps:43819-splits-1526251252838] 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost(367): Loaded 
coprocessor org.apache.phoenix.coprocessor.UngroupedAggregateRegionObserver 
from HTD of T000002 successfully.
2018-05-13 23:41:03,867 DEBUG [RS:0;xps:43819-splits-1526251252838] 
org.apache.hadoop.hbase.coprocessor.CoprocessorHost(182): Loading coprocessor 
class org.apache.phoenix.coprocessor.ScanRegionObserver with path null and 
priority 805306366
2018-05-13 23:41:03,867 INFO  [RS:0;xps:43819-splits-1526251252838] 
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost(367): Loaded 
coprocessor org.apache.phoenix.coprocessor.ScanRegionObserver from HTD of 
T000002 successfully.
2018-05-13 23:41:03,867 DEBUG [RS:0;xps:43819-splits-1526251252838] 
org.apache.hadoop.hbase.regionserver.MetricsRegionSourceImpl(70): Creating new 
MetricsRegionSourceImpl for table T000002 bff1498b104381411dbcc5c1b104e0e8
2018-05-13 23:41:03,867 DEBUG [RS:0;xps:43819-splits-1526251252838] 
org.apache.hadoop.hbase.regionserver.HRegion(738): Instantiated 
T000002,\x01sp,1526251261881.bff1498b104381411dbcc5c1b104e0e8.
2018-05-13 23:41:03,877 DEBUG [main] 
org.apache.phoenix.util.ReadOnlyProps(317): Creating new ReadOnlyProps due to 
phoenix.query.force.rowkeyorder with true!=false
2018-05-13 23:41:03,877 INFO  [StoreOpener-730c649272e9520771a3b38973dea323-1] 
org.apache.hadoop.hbase.io.hfile.CacheConfig(236): Created cacheConfig for 0: 
blockCache=LruBlockCache{blockCount=0, currentSize=3061960, 
freeSize=2981814584, maxSize=2984876544, heapSize=3061960, minSize=2835632640, 
minFactor=0.95, multiSize=1417816320, multiFactor=0.5, singleSize=708908160, 
singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, 
cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, 
cacheDataCompressed=false, prefetchOnOpen=false
2018-05-13 23:41:03,877 INFO  [StoreOpener-730c649272e9520771a3b38973dea323-1] 
org.apache.hadoop.hbase.regionserver.compactions.CompactionConfiguration(104): 
size [134217728, 9223372036854775807, 9223372036854775807); files [3, 10); 
ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major 
period 604800000, major jitter 0.500000, min locality to compact 0.000000
2018-05-13 23:41:03,878 INFO  [StoreOpener-bff1498b104381411dbcc5c1b104e0e8-1] 
org.apache.hadoop.hbase.io.hfile.CacheConfig(236): Created cacheConfig for 0: 
blockCache=LruBlockCache{blockCount=0, currentSize=3061960, 
freeSize=2981814584, maxSize=2984876544, heapSize=3061960, minSize=2835632640, 
minFactor=0.95, multiSize=1417816320, multiFactor=0.5, singleSize=708908160, 
singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, 
cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false, 
cacheDataCompressed=false, prefetchOnOpen=false
2018-05-13 23:41:03,878 INFO  [StoreOpener-bff1498b104381411dbcc5c1b104e0e8-1] 
org.apache.hadoop.hbase.regionserver.compactions.CompactionConfiguration(104): 
size [134217728, 9223372036854775807, 9223372036854775807); files [3, 10); 
ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major 
period 604800000, major jitter 0.500000, min locality to compact 0.000000

java.lang.AssertionError: 
Expected :2
Actual   :0
 {code}

Thanks!

> Split during global index creation may cause ERROR 201 error
> ------------------------------------------------------------
>
>                 Key: PHOENIX-3163
>                 URL: https://issues.apache.org/jira/browse/PHOENIX-3163
>             Project: Phoenix
>          Issue Type: Bug
>    Affects Versions: 4.8.0
>            Reporter: Sergey Soldatov
>            Assignee: Sergey Soldatov
>            Priority: Major
>             Fix For: 4.14.0, 5.0.0
>
>         Attachments: PHOENIX-3163_v1.patch, PHOENIX-3163_v3.patch, 
> PHOENIX-3163_v4.patch, PHOENIX-3163_v5.patch, PHOENIX-3163_v6.patch
>
>
> When we create global index and split happen meanwhile there is a chance to 
> fail with ERROR 201:
> {noformat}
> 2016-08-08 15:55:17,248 INFO  [Thread-6] 
> org.apache.phoenix.iterate.BaseResultIterators(878): Failed to execute task 
> during cancel
> java.util.concurrent.ExecutionException: java.sql.SQLException: ERROR 201 
> (22000): Illegal data.
>       at java.util.concurrent.FutureTask.report(FutureTask.java:122)
>       at java.util.concurrent.FutureTask.get(FutureTask.java:192)
>       at 
> org.apache.phoenix.iterate.BaseResultIterators.close(BaseResultIterators.java:872)
>       at 
> org.apache.phoenix.iterate.BaseResultIterators.getIterators(BaseResultIterators.java:809)
>       at 
> org.apache.phoenix.iterate.BaseResultIterators.getIterators(BaseResultIterators.java:713)
>       at 
> org.apache.phoenix.iterate.RoundRobinResultIterator.getIterators(RoundRobinResultIterator.java:176)
>       at 
> org.apache.phoenix.iterate.RoundRobinResultIterator.next(RoundRobinResultIterator.java:91)
>       at 
> org.apache.phoenix.compile.UpsertCompiler$2.execute(UpsertCompiler.java:815)
>       at 
> org.apache.phoenix.compile.DelegateMutationPlan.execute(DelegateMutationPlan.java:31)
>       at 
> org.apache.phoenix.compile.PostIndexDDLCompiler$1.execute(PostIndexDDLCompiler.java:124)
>       at 
> org.apache.phoenix.query.ConnectionQueryServicesImpl.updateData(ConnectionQueryServicesImpl.java:2823)
>       at 
> org.apache.phoenix.schema.MetaDataClient.buildIndex(MetaDataClient.java:1079)
>       at 
> org.apache.phoenix.schema.MetaDataClient.createIndex(MetaDataClient.java:1382)
>       at 
> org.apache.phoenix.compile.CreateIndexCompiler$1.execute(CreateIndexCompiler.java:85)
>       at 
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:343)
>       at 
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:331)
>       at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53)
>       at 
> org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:330)
>       at 
> org.apache.phoenix.jdbc.PhoenixStatement.execute(PhoenixStatement.java:1440)
>       at 
> org.apache.phoenix.hbase.index.write.TestIndexWriter$1.run(TestIndexWriter.java:93)
> Caused by: java.sql.SQLException: ERROR 201 (22000): Illegal data.
>       at 
> org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:441)
>       at 
> org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:145)
>       at 
> org.apache.phoenix.schema.types.PDataType.newIllegalDataException(PDataType.java:287)
>       at 
> org.apache.phoenix.schema.types.PUnsignedSmallint$UnsignedShortCodec.decodeShort(PUnsignedSmallint.java:146)
>       at 
> org.apache.phoenix.schema.types.PSmallint.toObject(PSmallint.java:104)
>       at org.apache.phoenix.schema.types.PSmallint.toObject(PSmallint.java:28)
>       at 
> org.apache.phoenix.schema.types.PDataType.toObject(PDataType.java:980)
>       at 
> org.apache.phoenix.schema.types.PUnsignedSmallint.toObject(PUnsignedSmallint.java:102)
>       at 
> org.apache.phoenix.schema.types.PDataType.toObject(PDataType.java:980)
>       at 
> org.apache.phoenix.schema.types.PDataType.toObject(PDataType.java:992)
>       at 
> org.apache.phoenix.schema.types.PDataType.coerceBytes(PDataType.java:830)
>       at 
> org.apache.phoenix.schema.types.PDecimal.coerceBytes(PDecimal.java:342)
>       at 
> org.apache.phoenix.schema.types.PDataType.coerceBytes(PDataType.java:810)
>       at 
> org.apache.phoenix.expression.CoerceExpression.evaluate(CoerceExpression.java:149)
>       at 
> org.apache.phoenix.compile.ExpressionProjector.getValue(ExpressionProjector.java:69)
>       at 
> org.apache.phoenix.jdbc.PhoenixResultSet.getBytes(PhoenixResultSet.java:308)
>       at 
> org.apache.phoenix.compile.UpsertCompiler.upsertSelect(UpsertCompiler.java:197)
>       at 
> org.apache.phoenix.compile.UpsertCompiler.access$000(UpsertCompiler.java:115)
>       at 
> org.apache.phoenix.compile.UpsertCompiler$UpsertingParallelIteratorFactory.mutate(UpsertCompiler.java:259)
>       at 
> org.apache.phoenix.compile.MutatingParallelIteratorFactory.newIterator(MutatingParallelIteratorFactory.java:59)
>       at 
> org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:112)
>       at 
> org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:103)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>       at 
> org.apache.phoenix.job.JobManager$InstrumentedJobFutureTask.run(JobManager.java:183)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>       at java.lang.Thread.run(Thread.java:745)
> java.sql.SQLException: ERROR 201 (22000): Illegal data. ERROR 201 (22000): 
> Illegal data.
>       at 
> org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:441)
>       at 
> org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:145)
>       at 
> org.apache.phoenix.util.ServerUtil.parseRemoteException(ServerUtil.java:129)
>       at 
> org.apache.phoenix.util.ServerUtil.parseServerExceptionOrNull(ServerUtil.java:118)
>       at 
> org.apache.phoenix.util.ServerUtil.parseServerException(ServerUtil.java:107)
>       at 
> org.apache.phoenix.iterate.BaseResultIterators.getIterators(BaseResultIterators.java:768)
>       at 
> org.apache.phoenix.iterate.BaseResultIterators.getIterators(BaseResultIterators.java:713)
>       at 
> org.apache.phoenix.iterate.RoundRobinResultIterator.getIterators(RoundRobinResultIterator.java:176)
>       at 
> org.apache.phoenix.iterate.RoundRobinResultIterator.next(RoundRobinResultIterator.java:91)
>       at 
> org.apache.phoenix.compile.UpsertCompiler$2.execute(UpsertCompiler.java:815)
>       at 
> org.apache.phoenix.compile.DelegateMutationPlan.execute(DelegateMutationPlan.java:31)
>       at 
> org.apache.phoenix.compile.PostIndexDDLCompiler$1.execute(PostIndexDDLCompiler.java:124)
>       at 
> org.apache.phoenix.query.ConnectionQueryServicesImpl.updateData(ConnectionQueryServicesImpl.java:2823)
>       at 
> org.apache.phoenix.schema.MetaDataClient.buildIndex(MetaDataClient.java:1079)
>       at 
> org.apache.phoenix.schema.MetaDataClient.createIndex(MetaDataClient.java:1382)
>       at 
> org.apache.phoenix.compile.CreateIndexCompiler$1.execute(CreateIndexCompiler.java:85)
>       at 
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:343)
>       at 
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:331)
>       at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53)
>       at 
> org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:330)
>       at 
> org.apache.phoenix.jdbc.PhoenixStatement.execute(PhoenixStatement.java:1440)
>       at 
> org.apache.phoenix.hbase.index.write.TestIndexWriter$1.run(TestIndexWriter.java:93)
> Caused by: java.sql.SQLException: ERROR 201 (22000): Illegal data.
>       at 
> org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:441)
>       at 
> org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:145)
>       at 
> org.apache.phoenix.schema.types.PDataType.newIllegalDataException(PDataType.java:287)
>       at 
> org.apache.phoenix.schema.types.PUnsignedSmallint$UnsignedShortCodec.decodeShort(PUnsignedSmallint.java:146)
>       at 
> org.apache.phoenix.schema.types.PSmallint.toObject(PSmallint.java:104)
>       at org.apache.phoenix.schema.types.PSmallint.toObject(PSmallint.java:28)
>       at 
> org.apache.phoenix.schema.types.PDataType.toObject(PDataType.java:980)
>       at 
> org.apache.phoenix.schema.types.PUnsignedSmallint.toObject(PUnsignedSmallint.java:102)
>       at 
> org.apache.phoenix.schema.types.PDataType.toObject(PDataType.java:980)
>       at 
> org.apache.phoenix.schema.types.PDataType.toObject(PDataType.java:992)
>       at 
> org.apache.phoenix.schema.types.PDataType.coerceBytes(PDataType.java:830)
>       at 
> org.apache.phoenix.schema.types.PDecimal.coerceBytes(PDecimal.java:342)
>       at 
> org.apache.phoenix.schema.types.PDataType.coerceBytes(PDataType.java:810)
>       at 
> org.apache.phoenix.expression.CoerceExpression.evaluate(CoerceExpression.java:149)
>       at 
> org.apache.phoenix.compile.ExpressionProjector.getValue(ExpressionProjector.java:69)
>       at 
> org.apache.phoenix.jdbc.PhoenixResultSet.getBytes(PhoenixResultSet.java:308)
>       at 
> org.apache.phoenix.compile.UpsertCompiler.upsertSelect(UpsertCompiler.java:197)
>       at 
> org.apache.phoenix.compile.UpsertCompiler.access$000(UpsertCompiler.java:115)
>       at 
> org.apache.phoenix.compile.UpsertCompiler$UpsertingParallelIteratorFactory.mutate(UpsertCompiler.java:259)
>       at 
> org.apache.phoenix.compile.MutatingParallelIteratorFactory.newIterator(MutatingParallelIteratorFactory.java:59)
>       at 
> org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:112)
>       at 
> org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:103)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>       at 
> org.apache.phoenix.job.JobManager$InstrumentedJobFutureTask.run(JobManager.java:183)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>       at java.lang.Thread.run(Thread.java:745)
> {noformat}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to