[
https://issues.apache.org/jira/browse/PHOENIX-3163?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16473653#comment-16473653
]
Pedro Boado commented on PHOENIX-3163:
--------------------------------------
Hi [~sergey.soldatov]
I've just noticed that SkipScanAfterManualSplitIT.testManualSplit started
failing in branch 4.x-HBase-1.2 .
{code:java}
2018-05-13 23:41:03,729 DEBUG [B.defaultRpcServer.handler=1,queue=0,port=43819]
org.apache.hadoop.hbase.ipc.CallRunner(115):
B.defaultRpcServer.handler=1,queue=0,port=43819: callId: 465 service:
ClientService methodName: Scan size: 595 connection: 127.0.0.1:57462
org.apache.hadoop.hbase.NotServingRegionException: Region
T000002,\x01,1526251248024.34b289cddcb2b99d8e776602b796d731. is not online on
xps,43819,1526251221783
at
org.apache.hadoop.hbase.regionserver.HRegionServer.getRegionByEncodedName(HRegionServer.java:2942)
at
org.apache.hadoop.hbase.regionserver.RSRpcServices.getRegion(RSRpcServices.java:1072)
at
org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2410)
at
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33648)
at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2188)
at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
at
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
at java.lang.Thread.run(Thread.java:745)
2018-05-13 23:41:03,729 DEBUG [B.defaultRpcServer.handler=2,queue=0,port=43819]
org.apache.hadoop.hbase.ipc.CallRunner(115):
B.defaultRpcServer.handler=2,queue=0,port=43819: callId: 467 service:
ClientService methodName: Scan size: 586 connection: 127.0.0.1:57462
org.apache.hadoop.hbase.NotServingRegionException: Region
T000002,\x02,1526251248024.9201bdc4f44225f390edb40ab1548a82. is not online on
xps,43819,1526251221783
at
org.apache.hadoop.hbase.regionserver.HRegionServer.getRegionByEncodedName(HRegionServer.java:2942)
at
org.apache.hadoop.hbase.regionserver.RSRpcServices.getRegion(RSRpcServices.java:1072)
at
org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2410)
at
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33648)
at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2188)
at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
at
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
at java.lang.Thread.run(Thread.java:745)
2018-05-13 23:41:03,729 DEBUG [B.defaultRpcServer.handler=0,queue=0,port=43819]
org.apache.hadoop.hbase.ipc.CallRunner(115):
B.defaultRpcServer.handler=0,queue=0,port=43819: callId: 466 service:
ClientService methodName: Scan size: 585 connection: 127.0.0.1:57462
org.apache.hadoop.hbase.NotServingRegionException: Region
T000002,,1526251248024.9bb19fa73f91248dd407192c4ce512fe. is not online on
xps,43819,1526251221783
at
org.apache.hadoop.hbase.regionserver.HRegionServer.getRegionByEncodedName(HRegionServer.java:2942)
at
org.apache.hadoop.hbase.regionserver.RSRpcServices.getRegion(RSRpcServices.java:1072)
at
org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2410)
at
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33648)
at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2188)
at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
at
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
at java.lang.Thread.run(Thread.java:745)
2018-05-13 23:41:03,729 DEBUG [B.defaultRpcServer.handler=3,queue=0,port=43819]
org.apache.hadoop.hbase.ipc.CallRunner(115):
B.defaultRpcServer.handler=3,queue=0,port=43819: callId: 468 service:
ClientService methodName: Scan size: 595 connection: 127.0.0.1:57462
org.apache.hadoop.hbase.NotServingRegionException: Region
T000002,\x03,1526251248024.6944b7e5e33cdcbdcc674c745ad8c1a5. is not online on
xps,43819,1526251221783
at
org.apache.hadoop.hbase.regionserver.HRegionServer.getRegionByEncodedName(HRegionServer.java:2942)
at
org.apache.hadoop.hbase.regionserver.RSRpcServices.getRegion(RSRpcServices.java:1072)
at
org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2410)
at
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33648)
at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2188)
at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
at
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
at java.lang.Thread.run(Thread.java:745)
2018-05-13 23:41:03,790 DEBUG [RS:0;xps:43819-splits-1526251252838]
org.apache.hadoop.hbase.coprocessor.CoprocessorHost(182): Loading coprocessor
class org.apache.phoenix.hbase.index.Indexer with path null and priority
805306366
2018-05-13 23:41:03,826 DEBUG [RS:0;xps:43819-splits-1526251252838]
org.apache.phoenix.hbase.index.Indexer(244): Setting up recovery writter with
failure policy: class
org.apache.phoenix.hbase.index.write.recovery.StoreFailuresInCachePolicy
2018-05-13 23:41:03,826 INFO [RS:0;xps:43819-splits-1526251252838]
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost(367): Loaded
coprocessor org.apache.phoenix.hbase.index.Indexer from HTD of T000002
successfully.
2018-05-13 23:41:03,826 DEBUG [RS:0;xps:43819-splits-1526251252838]
org.apache.hadoop.hbase.coprocessor.CoprocessorHost(182): Loading coprocessor
class org.apache.phoenix.coprocessor.ServerCachingEndpointImpl with path null
and priority 805306366
2018-05-13 23:41:03,826 DEBUG [RS:0;xps:43819-splits-1526251252838]
org.apache.hadoop.hbase.regionserver.HRegion(7836): Registered coprocessor
service: region=T000002,\x01sp,1526251261881.bff1498b104381411dbcc5c1b104e0e8.
service=ServerCachingService
2018-05-13 23:41:03,826 INFO [RS:0;xps:43819-splits-1526251252838]
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost(367): Loaded
coprocessor org.apache.phoenix.coprocessor.ServerCachingEndpointImpl from HTD
of T000002 successfully.
2018-05-13 23:41:03,826 DEBUG [RS:0;xps:43819-splits-1526251252838]
org.apache.hadoop.hbase.coprocessor.CoprocessorHost(182): Loading coprocessor
class org.apache.phoenix.coprocessor.GroupedAggregateRegionObserver with path
null and priority 805306366
2018-05-13 23:41:03,827 INFO [RS:0;xps:43819-splits-1526251252838]
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost(367): Loaded
coprocessor org.apache.phoenix.coprocessor.GroupedAggregateRegionObserver from
HTD of T000002 successfully.
2018-05-13 23:41:03,827 DEBUG [RS:0;xps:43819-splits-1526251252838]
org.apache.hadoop.hbase.coprocessor.CoprocessorHost(182): Loading coprocessor
class org.apache.phoenix.coprocessor.UngroupedAggregateRegionObserver with path
null and priority 805306366
2018-05-13 23:41:03,847 DEBUG [B.defaultRpcServer.handler=4,queue=0,port=43819]
org.apache.hadoop.hbase.ipc.CallRunner(115):
B.defaultRpcServer.handler=4,queue=0,port=43819: callId: 472 service:
ClientService methodName: Scan size: 595 connection: 127.0.0.1:57462
org.apache.hadoop.hbase.DoNotRetryIOException: ERROR 1108 (XCL08): Cache of
region boundaries are out of date. tableName=T000002
at
org.apache.phoenix.coprocessor.BaseScannerRegionObserver.throwIfScanOutOfRegion(BaseScannerRegionObserver.java:175)
at
org.apache.phoenix.coprocessor.BaseScannerRegionObserver.preScannerOpen(BaseScannerRegionObserver.java:203)
at
org.apache.phoenix.coprocessor.UngroupedAggregateRegionObserver.preScannerOpen(UngroupedAggregateRegionObserver.java:327)
at
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$50.call(RegionCoprocessorHost.java:1300)
at
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$RegionOperation.call(RegionCoprocessorHost.java:1673)
at
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1749)
at
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperationWithResult(RegionCoprocessorHost.java:1722)
at
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preScannerOpen(RegionCoprocessorHost.java:1295)
at
org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2428)
at
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33648)
at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2188)
at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
at
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.phoenix.schema.StaleRegionBoundaryCacheException: ERROR
1108 (XCL08): Cache of region boundaries are out of date. tableName=T000002
at
org.apache.phoenix.coprocessor.BaseScannerRegionObserver.throwIfScanOutOfRegion(BaseScannerRegionObserver.java:174)
... 14 more
2018-05-13 23:41:03,851 DEBUG [B.defaultRpcServer.handler=0,queue=0,port=43819]
org.apache.phoenix.coprocessor.UngroupedAggregateRegionObserver(528): Starting
ungrouped coprocessor scan
{"timeRange":[0,9223372036854775807],"batch":-1,"startRow":"\\x00
c","stopRow":"\\x00
c\\x00","loadColumnFamiliesOnDemand":true,"totalColumns":1,"cacheBlocks":true,"families":{"0":["ALL"]},"maxResultSize":2097152,"maxVersions":1,"filter":"FilterList
AND (2/2): [FirstKeyOnlyFilter, SkipScanFilter [[\\x00
c]]]","caching":2147483647} {ENCODED => ed285ee58edd90670656b3753dd6a3fd, NAME
=> 'T000002,,1526251252848.ed285ee58edd90670656b3753dd6a3fd.', STARTKEY => '',
ENDKEY => '\x00jj'}
2018-05-13 23:41:03,851 DEBUG [B.defaultRpcServer.handler=0,queue=0,port=43819]
org.apache.phoenix.coprocessor.UngroupedAggregateRegionObserver(822): Finished
scanning 0 rows for ungrouped coprocessor scan
{"timeRange":[0,9223372036854775807],"batch":-1,"startRow":"\\x00
c","stopRow":"\\x00
c\\x00","loadColumnFamiliesOnDemand":true,"totalColumns":1,"cacheBlocks":true,"families":{"0":["ALL"]},"maxResultSize":2097152,"maxVersions":1,"filter":"FilterList
AND (2/2): [FirstKeyOnlyFilter, SkipScanFilter [[\\x00
c]]]","caching":2147483647}
2018-05-13 23:41:03,852 DEBUG [B.defaultRpcServer.handler=2,queue=0,port=43819]
org.apache.hadoop.hbase.ipc.CallRunner(115):
B.defaultRpcServer.handler=2,queue=0,port=43819: callId: 475 service:
ClientService methodName: Scan size: 595 connection: 127.0.0.1:57462
org.apache.hadoop.hbase.DoNotRetryIOException: ERROR 1108 (XCL08): Cache of
region boundaries are out of date. tableName=T000002
at
org.apache.phoenix.coprocessor.BaseScannerRegionObserver.throwIfScanOutOfRegion(BaseScannerRegionObserver.java:175)
at
org.apache.phoenix.coprocessor.BaseScannerRegionObserver.preScannerOpen(BaseScannerRegionObserver.java:203)
at
org.apache.phoenix.coprocessor.UngroupedAggregateRegionObserver.preScannerOpen(UngroupedAggregateRegionObserver.java:327)
at
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$50.call(RegionCoprocessorHost.java:1300)
at
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$RegionOperation.call(RegionCoprocessorHost.java:1673)
at
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1749)
at
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperationWithResult(RegionCoprocessorHost.java:1722)
at
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preScannerOpen(RegionCoprocessorHost.java:1295)
at
org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2428)
at
org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33648)
at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2188)
at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112)
at
org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133)
at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108)
at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.phoenix.schema.StaleRegionBoundaryCacheException: ERROR
1108 (XCL08): Cache of region boundaries are out of date. tableName=T000002
at
org.apache.phoenix.coprocessor.BaseScannerRegionObserver.throwIfScanOutOfRegion(BaseScannerRegionObserver.java:174)
... 14 more
2018-05-13 23:41:03,853 DEBUG [B.defaultRpcServer.handler=3,queue=0,port=43819]
org.apache.phoenix.coprocessor.UngroupedAggregateRegionObserver(528): Starting
ungrouped coprocessor scan
{"timeRange":[0,9223372036854775807],"batch":-1,"startRow":"\\x02
a","stopRow":"\\x02
a\\x00","loadColumnFamiliesOnDemand":true,"totalColumns":1,"cacheBlocks":true,"families":{"0":["ALL"]},"maxResultSize":2097152,"maxVersions":1,"filter":"FilterList
AND (2/2): [FirstKeyOnlyFilter, SkipScanFilter [[\\x02
a]]]","caching":2147483647} {ENCODED => c5c9386b8657bed3abfba259adc54cf5, NAME
=> 'T000002,\x02,1526251257604.c5c9386b8657bed3abfba259adc54cf5.', STARTKEY =>
'\x02', ENDKEY => '\x02jh'}
2018-05-13 23:41:03,854 DEBUG [B.defaultRpcServer.handler=3,queue=0,port=43819]
org.apache.phoenix.coprocessor.UngroupedAggregateRegionObserver(822): Finished
scanning 0 rows for ungrouped coprocessor scan
{"timeRange":[0,9223372036854775807],"batch":-1,"startRow":"\\x02
a","stopRow":"\\x02
a\\x00","loadColumnFamiliesOnDemand":true,"totalColumns":1,"cacheBlocks":true,"families":{"0":["ALL"]},"maxResultSize":2097152,"maxVersions":1,"filter":"FilterList
AND (2/2): [FirstKeyOnlyFilter, SkipScanFilter [[\\x02
a]]]","caching":2147483647}
2018-05-13 23:41:03,866 INFO [RS:0;xps:43819-splits-1526251252838]
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost(367): Loaded
coprocessor org.apache.phoenix.coprocessor.UngroupedAggregateRegionObserver
from HTD of T000002 successfully.
2018-05-13 23:41:03,867 DEBUG [RS:0;xps:43819-splits-1526251252838]
org.apache.hadoop.hbase.coprocessor.CoprocessorHost(182): Loading coprocessor
class org.apache.phoenix.coprocessor.ScanRegionObserver with path null and
priority 805306366
2018-05-13 23:41:03,867 INFO [RS:0;xps:43819-splits-1526251252838]
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost(367): Loaded
coprocessor org.apache.phoenix.coprocessor.ScanRegionObserver from HTD of
T000002 successfully.
2018-05-13 23:41:03,867 DEBUG [RS:0;xps:43819-splits-1526251252838]
org.apache.hadoop.hbase.regionserver.MetricsRegionSourceImpl(70): Creating new
MetricsRegionSourceImpl for table T000002 bff1498b104381411dbcc5c1b104e0e8
2018-05-13 23:41:03,867 DEBUG [RS:0;xps:43819-splits-1526251252838]
org.apache.hadoop.hbase.regionserver.HRegion(738): Instantiated
T000002,\x01sp,1526251261881.bff1498b104381411dbcc5c1b104e0e8.
2018-05-13 23:41:03,877 DEBUG [main]
org.apache.phoenix.util.ReadOnlyProps(317): Creating new ReadOnlyProps due to
phoenix.query.force.rowkeyorder with true!=false
2018-05-13 23:41:03,877 INFO [StoreOpener-730c649272e9520771a3b38973dea323-1]
org.apache.hadoop.hbase.io.hfile.CacheConfig(236): Created cacheConfig for 0:
blockCache=LruBlockCache{blockCount=0, currentSize=3061960,
freeSize=2981814584, maxSize=2984876544, heapSize=3061960, minSize=2835632640,
minFactor=0.95, multiSize=1417816320, multiFactor=0.5, singleSize=708908160,
singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false,
cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false,
cacheDataCompressed=false, prefetchOnOpen=false
2018-05-13 23:41:03,877 INFO [StoreOpener-730c649272e9520771a3b38973dea323-1]
org.apache.hadoop.hbase.regionserver.compactions.CompactionConfiguration(104):
size [134217728, 9223372036854775807, 9223372036854775807); files [3, 10);
ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major
period 604800000, major jitter 0.500000, min locality to compact 0.000000
2018-05-13 23:41:03,878 INFO [StoreOpener-bff1498b104381411dbcc5c1b104e0e8-1]
org.apache.hadoop.hbase.io.hfile.CacheConfig(236): Created cacheConfig for 0:
blockCache=LruBlockCache{blockCount=0, currentSize=3061960,
freeSize=2981814584, maxSize=2984876544, heapSize=3061960, minSize=2835632640,
minFactor=0.95, multiSize=1417816320, multiFactor=0.5, singleSize=708908160,
singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false,
cacheIndexesOnWrite=false, cacheBloomsOnWrite=false, cacheEvictOnClose=false,
cacheDataCompressed=false, prefetchOnOpen=false
2018-05-13 23:41:03,878 INFO [StoreOpener-bff1498b104381411dbcc5c1b104e0e8-1]
org.apache.hadoop.hbase.regionserver.compactions.CompactionConfiguration(104):
size [134217728, 9223372036854775807, 9223372036854775807); files [3, 10);
ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major
period 604800000, major jitter 0.500000, min locality to compact 0.000000
java.lang.AssertionError:
Expected :2
Actual :0
{code}
Thanks!
> Split during global index creation may cause ERROR 201 error
> ------------------------------------------------------------
>
> Key: PHOENIX-3163
> URL: https://issues.apache.org/jira/browse/PHOENIX-3163
> Project: Phoenix
> Issue Type: Bug
> Affects Versions: 4.8.0
> Reporter: Sergey Soldatov
> Assignee: Sergey Soldatov
> Priority: Major
> Fix For: 4.14.0, 5.0.0
>
> Attachments: PHOENIX-3163_v1.patch, PHOENIX-3163_v3.patch,
> PHOENIX-3163_v4.patch, PHOENIX-3163_v5.patch, PHOENIX-3163_v6.patch
>
>
> When we create global index and split happen meanwhile there is a chance to
> fail with ERROR 201:
> {noformat}
> 2016-08-08 15:55:17,248 INFO [Thread-6]
> org.apache.phoenix.iterate.BaseResultIterators(878): Failed to execute task
> during cancel
> java.util.concurrent.ExecutionException: java.sql.SQLException: ERROR 201
> (22000): Illegal data.
> at java.util.concurrent.FutureTask.report(FutureTask.java:122)
> at java.util.concurrent.FutureTask.get(FutureTask.java:192)
> at
> org.apache.phoenix.iterate.BaseResultIterators.close(BaseResultIterators.java:872)
> at
> org.apache.phoenix.iterate.BaseResultIterators.getIterators(BaseResultIterators.java:809)
> at
> org.apache.phoenix.iterate.BaseResultIterators.getIterators(BaseResultIterators.java:713)
> at
> org.apache.phoenix.iterate.RoundRobinResultIterator.getIterators(RoundRobinResultIterator.java:176)
> at
> org.apache.phoenix.iterate.RoundRobinResultIterator.next(RoundRobinResultIterator.java:91)
> at
> org.apache.phoenix.compile.UpsertCompiler$2.execute(UpsertCompiler.java:815)
> at
> org.apache.phoenix.compile.DelegateMutationPlan.execute(DelegateMutationPlan.java:31)
> at
> org.apache.phoenix.compile.PostIndexDDLCompiler$1.execute(PostIndexDDLCompiler.java:124)
> at
> org.apache.phoenix.query.ConnectionQueryServicesImpl.updateData(ConnectionQueryServicesImpl.java:2823)
> at
> org.apache.phoenix.schema.MetaDataClient.buildIndex(MetaDataClient.java:1079)
> at
> org.apache.phoenix.schema.MetaDataClient.createIndex(MetaDataClient.java:1382)
> at
> org.apache.phoenix.compile.CreateIndexCompiler$1.execute(CreateIndexCompiler.java:85)
> at
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:343)
> at
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:331)
> at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53)
> at
> org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:330)
> at
> org.apache.phoenix.jdbc.PhoenixStatement.execute(PhoenixStatement.java:1440)
> at
> org.apache.phoenix.hbase.index.write.TestIndexWriter$1.run(TestIndexWriter.java:93)
> Caused by: java.sql.SQLException: ERROR 201 (22000): Illegal data.
> at
> org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:441)
> at
> org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:145)
> at
> org.apache.phoenix.schema.types.PDataType.newIllegalDataException(PDataType.java:287)
> at
> org.apache.phoenix.schema.types.PUnsignedSmallint$UnsignedShortCodec.decodeShort(PUnsignedSmallint.java:146)
> at
> org.apache.phoenix.schema.types.PSmallint.toObject(PSmallint.java:104)
> at org.apache.phoenix.schema.types.PSmallint.toObject(PSmallint.java:28)
> at
> org.apache.phoenix.schema.types.PDataType.toObject(PDataType.java:980)
> at
> org.apache.phoenix.schema.types.PUnsignedSmallint.toObject(PUnsignedSmallint.java:102)
> at
> org.apache.phoenix.schema.types.PDataType.toObject(PDataType.java:980)
> at
> org.apache.phoenix.schema.types.PDataType.toObject(PDataType.java:992)
> at
> org.apache.phoenix.schema.types.PDataType.coerceBytes(PDataType.java:830)
> at
> org.apache.phoenix.schema.types.PDecimal.coerceBytes(PDecimal.java:342)
> at
> org.apache.phoenix.schema.types.PDataType.coerceBytes(PDataType.java:810)
> at
> org.apache.phoenix.expression.CoerceExpression.evaluate(CoerceExpression.java:149)
> at
> org.apache.phoenix.compile.ExpressionProjector.getValue(ExpressionProjector.java:69)
> at
> org.apache.phoenix.jdbc.PhoenixResultSet.getBytes(PhoenixResultSet.java:308)
> at
> org.apache.phoenix.compile.UpsertCompiler.upsertSelect(UpsertCompiler.java:197)
> at
> org.apache.phoenix.compile.UpsertCompiler.access$000(UpsertCompiler.java:115)
> at
> org.apache.phoenix.compile.UpsertCompiler$UpsertingParallelIteratorFactory.mutate(UpsertCompiler.java:259)
> at
> org.apache.phoenix.compile.MutatingParallelIteratorFactory.newIterator(MutatingParallelIteratorFactory.java:59)
> at
> org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:112)
> at
> org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:103)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at
> org.apache.phoenix.job.JobManager$InstrumentedJobFutureTask.run(JobManager.java:183)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> java.sql.SQLException: ERROR 201 (22000): Illegal data. ERROR 201 (22000):
> Illegal data.
> at
> org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:441)
> at
> org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:145)
> at
> org.apache.phoenix.util.ServerUtil.parseRemoteException(ServerUtil.java:129)
> at
> org.apache.phoenix.util.ServerUtil.parseServerExceptionOrNull(ServerUtil.java:118)
> at
> org.apache.phoenix.util.ServerUtil.parseServerException(ServerUtil.java:107)
> at
> org.apache.phoenix.iterate.BaseResultIterators.getIterators(BaseResultIterators.java:768)
> at
> org.apache.phoenix.iterate.BaseResultIterators.getIterators(BaseResultIterators.java:713)
> at
> org.apache.phoenix.iterate.RoundRobinResultIterator.getIterators(RoundRobinResultIterator.java:176)
> at
> org.apache.phoenix.iterate.RoundRobinResultIterator.next(RoundRobinResultIterator.java:91)
> at
> org.apache.phoenix.compile.UpsertCompiler$2.execute(UpsertCompiler.java:815)
> at
> org.apache.phoenix.compile.DelegateMutationPlan.execute(DelegateMutationPlan.java:31)
> at
> org.apache.phoenix.compile.PostIndexDDLCompiler$1.execute(PostIndexDDLCompiler.java:124)
> at
> org.apache.phoenix.query.ConnectionQueryServicesImpl.updateData(ConnectionQueryServicesImpl.java:2823)
> at
> org.apache.phoenix.schema.MetaDataClient.buildIndex(MetaDataClient.java:1079)
> at
> org.apache.phoenix.schema.MetaDataClient.createIndex(MetaDataClient.java:1382)
> at
> org.apache.phoenix.compile.CreateIndexCompiler$1.execute(CreateIndexCompiler.java:85)
> at
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:343)
> at
> org.apache.phoenix.jdbc.PhoenixStatement$2.call(PhoenixStatement.java:331)
> at org.apache.phoenix.call.CallRunner.run(CallRunner.java:53)
> at
> org.apache.phoenix.jdbc.PhoenixStatement.executeMutation(PhoenixStatement.java:330)
> at
> org.apache.phoenix.jdbc.PhoenixStatement.execute(PhoenixStatement.java:1440)
> at
> org.apache.phoenix.hbase.index.write.TestIndexWriter$1.run(TestIndexWriter.java:93)
> Caused by: java.sql.SQLException: ERROR 201 (22000): Illegal data.
> at
> org.apache.phoenix.exception.SQLExceptionCode$Factory$1.newException(SQLExceptionCode.java:441)
> at
> org.apache.phoenix.exception.SQLExceptionInfo.buildException(SQLExceptionInfo.java:145)
> at
> org.apache.phoenix.schema.types.PDataType.newIllegalDataException(PDataType.java:287)
> at
> org.apache.phoenix.schema.types.PUnsignedSmallint$UnsignedShortCodec.decodeShort(PUnsignedSmallint.java:146)
> at
> org.apache.phoenix.schema.types.PSmallint.toObject(PSmallint.java:104)
> at org.apache.phoenix.schema.types.PSmallint.toObject(PSmallint.java:28)
> at
> org.apache.phoenix.schema.types.PDataType.toObject(PDataType.java:980)
> at
> org.apache.phoenix.schema.types.PUnsignedSmallint.toObject(PUnsignedSmallint.java:102)
> at
> org.apache.phoenix.schema.types.PDataType.toObject(PDataType.java:980)
> at
> org.apache.phoenix.schema.types.PDataType.toObject(PDataType.java:992)
> at
> org.apache.phoenix.schema.types.PDataType.coerceBytes(PDataType.java:830)
> at
> org.apache.phoenix.schema.types.PDecimal.coerceBytes(PDecimal.java:342)
> at
> org.apache.phoenix.schema.types.PDataType.coerceBytes(PDataType.java:810)
> at
> org.apache.phoenix.expression.CoerceExpression.evaluate(CoerceExpression.java:149)
> at
> org.apache.phoenix.compile.ExpressionProjector.getValue(ExpressionProjector.java:69)
> at
> org.apache.phoenix.jdbc.PhoenixResultSet.getBytes(PhoenixResultSet.java:308)
> at
> org.apache.phoenix.compile.UpsertCompiler.upsertSelect(UpsertCompiler.java:197)
> at
> org.apache.phoenix.compile.UpsertCompiler.access$000(UpsertCompiler.java:115)
> at
> org.apache.phoenix.compile.UpsertCompiler$UpsertingParallelIteratorFactory.mutate(UpsertCompiler.java:259)
> at
> org.apache.phoenix.compile.MutatingParallelIteratorFactory.newIterator(MutatingParallelIteratorFactory.java:59)
> at
> org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:112)
> at
> org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:103)
> at java.util.concurrent.FutureTask.run(FutureTask.java:266)
> at
> org.apache.phoenix.job.JobManager$InstrumentedJobFutureTask.run(JobManager.java:183)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
> at java.lang.Thread.run(Thread.java:745)
> {noformat}
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)