[ 
https://issues.apache.org/jira/browse/PHOENIX-1800?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rajeshbabu Chintaguntla updated PHOENIX-1800:
---------------------------------------------
    Attachment: PHOENIX-1800.patch

Here is the patch.

> Fix failing test case SaltedViewIT#testSaltedUpdatableViewWithLocalIndex
> ------------------------------------------------------------------------
>
>                 Key: PHOENIX-1800
>                 URL: https://issues.apache.org/jira/browse/PHOENIX-1800
>             Project: Phoenix
>          Issue Type: Bug
>            Reporter: Rajeshbabu Chintaguntla
>            Assignee: Rajeshbabu Chintaguntla
>         Attachments: PHOENIX-1800.patch
>
>
> The test case failing with 
> {noformat}
> org.apache.phoenix.exception.PhoenixIOException: 
> org.apache.phoenix.exception.PhoenixIOException: 
> org.apache.hadoop.hbase.TableNotFoundException: _LOCAL_IDX__LOCAL_IDX_T
>       at 
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegionInMeta(HConnectionManager.java:1279)
>       at 
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1150)
>       at 
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1107)
>       at 
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.getRegionLocation(HConnectionManager.java:948)
>       at 
> org.apache.hadoop.hbase.client.RegionServerCallable.prepare(RegionServerCallable.java:72)
>       at 
> org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:116)
>       at org.apache.hadoop.hbase.client.HTable.get(HTable.java:813)
>       at 
> org.apache.hadoop.hbase.coprocessor.CoprocessorHost$Environment$HTableWrapper.get(CoprocessorHost.java:452)
>       at 
> org.apache.phoenix.util.IndexUtil.wrapResultUsingOffset(IndexUtil.java:490)
>       at 
> org.apache.phoenix.coprocessor.BaseScannerRegionObserver$2.nextRaw(BaseScannerRegionObserver.java:310)
>       at 
> org.apache.phoenix.coprocessor.DelegateRegionScanner.nextRaw(DelegateRegionScanner.java:76)
>       at 
> org.apache.hadoop.hbase.regionserver.HRegionServer.scan(HRegionServer.java:3249)
>       at 
> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:29941)
>       at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2029)
>       at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:108)
>       at 
> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:112)
>       at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:92)
>       at java.lang.Thread.run(Thread.java:724)
>       at java.util.concurrent.FutureTask$Sync.innerGet(FutureTask.java:262)
>       at java.util.concurrent.FutureTask.get(FutureTask.java:119)
>       at 
> org.apache.phoenix.iterate.BaseResultIterators.getIterators(BaseResultIterators.java:536)
>       at 
> org.apache.phoenix.iterate.MergeSortResultIterator.getIterators(MergeSortResultIterator.java:48)
>       at 
> org.apache.phoenix.iterate.MergeSortResultIterator.minIterator(MergeSortResultIterator.java:84)
>       at 
> org.apache.phoenix.iterate.MergeSortResultIterator.next(MergeSortResultIterator.java:111)
>       at 
> org.apache.phoenix.jdbc.PhoenixResultSet.next(PhoenixResultSet.java:756)
>       at 
> org.apache.phoenix.end2end.BaseViewIT.testUpdatableViewIndex(BaseViewIT.java:125)
>       at 
> org.apache.phoenix.end2end.BaseViewIT.testUpdatableViewWithIndex(BaseViewIT.java:54)
>       at 
> org.apache.phoenix.end2end.SaltedViewIT.testSaltedUpdatableViewWithLocalIndex(SaltedViewIT.java:39)
> Caused by: org.apache.phoenix.exception.PhoenixIOException: 
> org.apache.hadoop.hbase.TableNotFoundException: _LOCAL_IDX__LOCAL_IDX_T
>       at 
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegionInMeta(HConnectionManager.java:1279)
>       at 
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1150)
>       at 
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1107)
>       at 
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.getRegionLocation(HConnectionManager.java:948)
>       at 
> org.apache.hadoop.hbase.client.RegionServerCallable.prepare(RegionServerCallable.java:72)
>       at 
> org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:116)
>       at org.apache.hadoop.hbase.client.HTable.get(HTable.java:813)
>       at 
> org.apache.hadoop.hbase.coprocessor.CoprocessorHost$Environment$HTableWrapper.get(CoprocessorHost.java:452)
>       at 
> org.apache.phoenix.util.IndexUtil.wrapResultUsingOffset(IndexUtil.java:490)
>       at 
> org.apache.phoenix.coprocessor.BaseScannerRegionObserver$2.nextRaw(BaseScannerRegionObserver.java:310)
>       at 
> org.apache.phoenix.coprocessor.DelegateRegionScanner.nextRaw(DelegateRegionScanner.java:76)
>       at 
> org.apache.hadoop.hbase.regionserver.HRegionServer.scan(HRegionServer.java:3249)
>       at 
> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:29941)
>       at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2029)
>       at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:108)
>       at 
> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:112)
>       at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:92)
>       at java.lang.Thread.run(Thread.java:724)
>       at 
> org.apache.phoenix.util.ServerUtil.parseServerException(ServerUtil.java:108)
>       at 
> org.apache.phoenix.iterate.ScanningResultIterator.next(ScanningResultIterator.java:56)
>       at 
> org.apache.phoenix.iterate.TableResultIterator.next(TableResultIterator.java:104)
>       at 
> org.apache.phoenix.iterate.ChunkedResultIterator$SingleChunkResultIterator.next(ChunkedResultIterator.java:149)
>       at 
> org.apache.phoenix.iterate.SpoolingResultIterator.<init>(SpoolingResultIterator.java:108)
>       at 
> org.apache.phoenix.iterate.SpoolingResultIterator.<init>(SpoolingResultIterator.java:75)
>       at 
> org.apache.phoenix.iterate.SpoolingResultIterator$SpoolingResultIteratorFactory.newIterator(SpoolingResultIterator.java:69)
>       at 
> org.apache.phoenix.iterate.ChunkedResultIterator.<init>(ChunkedResultIterator.java:92)
>       at 
> org.apache.phoenix.iterate.ChunkedResultIterator$ChunkedResultIteratorFactory.newIterator(ChunkedResultIterator.java:72)
>       at 
> org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:93)
>       at 
> org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:84)
>       at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:166)
>       at 
> org.apache.phoenix.job.JobManager$InstrumentedJobFutureTask.run(JobManager.java:172)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:724)
> Caused by: org.apache.hadoop.hbase.TableNotFoundException: 
> org.apache.hadoop.hbase.TableNotFoundException: _LOCAL_IDX__LOCAL_IDX_T
>       at 
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegionInMeta(HConnectionManager.java:1279)
>       at 
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1150)
>       at 
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1107)
>       at 
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.getRegionLocation(HConnectionManager.java:948)
>       at 
> org.apache.hadoop.hbase.client.RegionServerCallable.prepare(RegionServerCallable.java:72)
>       at 
> org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:116)
>       at org.apache.hadoop.hbase.client.HTable.get(HTable.java:813)
>       at 
> org.apache.hadoop.hbase.coprocessor.CoprocessorHost$Environment$HTableWrapper.get(CoprocessorHost.java:452)
>       at 
> org.apache.phoenix.util.IndexUtil.wrapResultUsingOffset(IndexUtil.java:490)
>       at 
> org.apache.phoenix.coprocessor.BaseScannerRegionObserver$2.nextRaw(BaseScannerRegionObserver.java:310)
>       at 
> org.apache.phoenix.coprocessor.DelegateRegionScanner.nextRaw(DelegateRegionScanner.java:76)
>       at 
> org.apache.hadoop.hbase.regionserver.HRegionServer.scan(HRegionServer.java:3249)
>       at 
> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:29941)
>       at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2029)
>       at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:108)
>       at 
> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:112)
>       at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:92)
>       at java.lang.Thread.run(Thread.java:724)
>       at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
>       at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
>       at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>       at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
>       at 
> org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
>       at 
> org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:95)
>       at 
> org.apache.hadoop.hbase.protobuf.ProtobufUtil.getRemoteException(ProtobufUtil.java:284)
>       at 
> org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:204)
>       at 
> org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:59)
>       at 
> org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:117)
>       at 
> org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:93)
>       at 
> org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:355)
>       at 
> org.apache.phoenix.iterate.ScanningResultIterator.next(ScanningResultIterator.java:50)
>       at 
> org.apache.phoenix.iterate.TableResultIterator.next(TableResultIterator.java:104)
>       at 
> org.apache.phoenix.iterate.ChunkedResultIterator$SingleChunkResultIterator.next(ChunkedResultIterator.java:149)
>       at 
> org.apache.phoenix.iterate.SpoolingResultIterator.<init>(SpoolingResultIterator.java:108)
>       at 
> org.apache.phoenix.iterate.SpoolingResultIterator.<init>(SpoolingResultIterator.java:75)
>       at 
> org.apache.phoenix.iterate.SpoolingResultIterator$SpoolingResultIteratorFactory.newIterator(SpoolingResultIterator.java:69)
>       at 
> org.apache.phoenix.iterate.ChunkedResultIterator.<init>(ChunkedResultIterator.java:92)
>       at 
> org.apache.phoenix.iterate.ChunkedResultIterator$ChunkedResultIteratorFactory.newIterator(ChunkedResultIterator.java:72)
>       at 
> org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:93)
>       at 
> org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:84)
>       at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:166)
>       at 
> org.apache.phoenix.job.JobManager$InstrumentedJobFutureTask.run(JobManager.java:172)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:724)
> Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException: 
> org.apache.hadoop.hbase.TableNotFoundException: _LOCAL_IDX__LOCAL_IDX_T
>       at 
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegionInMeta(HConnectionManager.java:1279)
>       at 
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1150)
>       at 
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1107)
>       at 
> org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.getRegionLocation(HConnectionManager.java:948)
>       at 
> org.apache.hadoop.hbase.client.RegionServerCallable.prepare(RegionServerCallable.java:72)
>       at 
> org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:116)
>       at org.apache.hadoop.hbase.client.HTable.get(HTable.java:813)
>       at 
> org.apache.hadoop.hbase.coprocessor.CoprocessorHost$Environment$HTableWrapper.get(CoprocessorHost.java:452)
>       at 
> org.apache.phoenix.util.IndexUtil.wrapResultUsingOffset(IndexUtil.java:490)
>       at 
> org.apache.phoenix.coprocessor.BaseScannerRegionObserver$2.nextRaw(BaseScannerRegionObserver.java:310)
>       at 
> org.apache.phoenix.coprocessor.DelegateRegionScanner.nextRaw(DelegateRegionScanner.java:76)
>       at 
> org.apache.hadoop.hbase.regionserver.HRegionServer.scan(HRegionServer.java:3249)
>       at 
> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:29941)
>       at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2029)
>       at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:108)
>       at 
> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:112)
>       at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:92)
>       at java.lang.Thread.run(Thread.java:724)
>       at org.apache.hadoop.hbase.ipc.RpcClient.call(RpcClient.java:1457)
>       at 
> org.apache.hadoop.hbase.ipc.RpcClient.callBlockingMethod(RpcClient.java:1661)
>       at 
> org.apache.hadoop.hbase.ipc.RpcClient$BlockingRpcChannelImplementation.callBlockingMethod(RpcClient.java:1719)
>       at 
> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$BlockingStub.scan(ClientProtos.java:30387)
>       at 
> org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:174)
>       at 
> org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:59)
>       at 
> org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:117)
>       at 
> org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:93)
>       at 
> org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:355)
>       at 
> org.apache.phoenix.iterate.ScanningResultIterator.next(ScanningResultIterator.java:50)
>       at 
> org.apache.phoenix.iterate.TableResultIterator.next(TableResultIterator.java:104)
>       at 
> org.apache.phoenix.iterate.ChunkedResultIterator$SingleChunkResultIterator.next(ChunkedResultIterator.java:149)
>       at 
> org.apache.phoenix.iterate.SpoolingResultIterator.<init>(SpoolingResultIterator.java:108)
>       at 
> org.apache.phoenix.iterate.SpoolingResultIterator.<init>(SpoolingResultIterator.java:75)
>       at 
> org.apache.phoenix.iterate.SpoolingResultIterator$SpoolingResultIteratorFactory.newIterator(SpoolingResultIterator.java:69)
>       at 
> org.apache.phoenix.iterate.ChunkedResultIterator.<init>(ChunkedResultIterator.java:92)
>       at 
> org.apache.phoenix.iterate.ChunkedResultIterator$ChunkedResultIteratorFactory.newIterator(ChunkedResultIterator.java:72)
>       at 
> org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:93)
>       at 
> org.apache.phoenix.iterate.ParallelIterators$1.call(ParallelIterators.java:84)
>       at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334)
>       at java.util.concurrent.FutureTask.run(FutureTask.java:166)
>       at 
> org.apache.phoenix.job.JobManager$InstrumentedJobFutureTask.run(JobManager.java:172)
>       at 
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
>       at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
>       at java.lang.Thread.run(Thread.java:724)
> {noformat}
> To join the missing columns from the data table it should get data table from 
> local index table name. This will happen when there is no colocation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

Reply via email to