I tried to set hbase.client.scanner.caching to 1 but still failed.
On Wed, Sep 9, 2015 at 2:45 PM, Stack <[email protected]> wrote: > Yes. > > <property> > <name>hbase.client.scanner.caching</name> > <value>100</value> > <description>Number of rows that will be fetched when calling next > on a scanner if it is not served from (local, client) memory. Higher > caching values will enable faster scanners but will eat up more memory > and some calls of next may take longer and longer times when the cache > is empty. > Do not set this value such that the time between invocations is greater > than the scanner timeout; i.e. > hbase.client.scanner.timeout.period</description> > </property> > > St.Ack > > On Tue, Sep 8, 2015 at 11:18 PM, Li Li <[email protected]> wrote: > >> is it possible setting it using hbase-site.xml? >> I can't modify titan codes. it only read hbase configuration file. >> >> On Wed, Sep 9, 2015 at 1:38 AM, Stack <[email protected]> wrote: >> > Are you seeing hiccups in the scans ahead of this exception -- scan next >> > retrying? Are the rows large? Try fetching in smaller batches, smaller >> than >> > 100. >> > St.Ack >> > >> > On Mon, Sep 7, 2015 at 3:56 AM, Li Li <[email protected]> wrote: >> > >> >> I am using titan which use hbase as it's storage engine. The hbase >> >> version is 1.0.0-cdh5.4.4. >> >> it's a full table scan over a large table. Is there any configuration >> >> I can change to tackle this problem. >> >> The exception stack is: >> >> >> >> Exception in thread "main" java.lang.RuntimeException: >> >> org.apache.hadoop.hbase.DoNotRetryIOException: Failed after retry of >> >> OutOfOrderScannerNextException: was there a rpc timeout? >> >> >> >> at >> >> >> org.apache.hadoop.hbase.client.AbstractClientScanner$1.hasNext(AbstractClientScanner.java:94) >> >> >> >> at >> >> com.google.common.collect.Iterators$8.computeNext(Iterators.java:686) >> >> >> >> at >> >> >> com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) >> >> >> >> at >> >> >> com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) >> >> >> >> at >> >> >> com.thinkaurelius.titan.diskstorage.hbase.HBaseKeyColumnValueStore$RowIterator.hasNext(HBaseKeyColumnValueStore.java:289) >> >> >> >> at >> >> >> com.thinkaurelius.titan.graphdb.database.StandardTitanGraph$2.hasNext(StandardTitanGraph.java:343) >> >> >> >> at >> >> >> com.thinkaurelius.titan.graphdb.transaction.VertexIterable$1.nextVertex(VertexIterable.java:34) >> >> >> >> at >> >> >> com.thinkaurelius.titan.graphdb.transaction.VertexIterable$1.next(VertexIterable.java:55) >> >> >> >> at >> >> >> com.thinkaurelius.titan.graphdb.transaction.VertexIterable$1.next(VertexIterable.java:27) >> >> >> >> at >> >> com.google.common.collect.Iterators$8.computeNext(Iterators.java:687) >> >> >> >> at >> >> >> com.google.common.collect.AbstractIterator.tryToComputeNext(AbstractIterator.java:143) >> >> >> >> at >> >> >> com.google.common.collect.AbstractIterator.hasNext(AbstractIterator.java:138) >> >> >> >> at >> >> >> com.mobvoi.knowledgegraph.storage.v2.music.DumpAllNodesAndEdges.main(DumpAllNodesAndEdges.java:34) >> >> >> >> Caused by: org.apache.hadoop.hbase.DoNotRetryIOException: Failed after >> >> retry of OutOfOrderScannerNextException: was there a rpc timeout? >> >> >> >> at >> >> >> org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:384) >> >> >> >> at >> >> >> org.apache.hadoop.hbase.client.AbstractClientScanner$1.hasNext(AbstractClientScanner.java:91) >> >> >> >> ... 12 more >> >> >> >> Caused by: >> >> org.apache.hadoop.hbase.exceptions.OutOfOrderScannerNextException: >> >> org.apache.hadoop.hbase.exceptions.OutOfOrderScannerNextException: >> >> Expected nextCallSeq: 2 But the nextCallSeq >> >> >> >> got from client: 1; request=scanner_id: 11996 number_of_rows: 100 >> >> close_scanner: false next_call_seq: 1 >> >> >> >> at >> >> >> org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2144) >> >> >> >> at >> >> >> org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:31443) >> >> >> >> at >> org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2035) >> >> >> >> at >> org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:107) >> >> >> >> at >> >> >> org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130) >> >> >> >> at >> >> org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107) >> >> >> >> at java.lang.Thread.run(Thread.java:745) >> >> >> >> >> >> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native >> >> Method) >> >> >> >> at >> >> >> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) >> >> >> >> at >> >> >> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) >> >> >> >> at >> java.lang.reflect.Constructor.newInstance(Constructor.java:526) >> >> >> >> at >> >> >> org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106) >> >> >> >> at >> >> >> org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:95) >> >> >> >> at >> >> >> org.apache.hadoop.hbase.protobuf.ProtobufUtil.getRemoteException(ProtobufUtil.java:284) >> >> >> >> at >> >> >> org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:198) >> >> >> >> at >> >> >> org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:57) >> >> >> >> at >> >> >> org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:114) >> >> >> >> at >> >> >> org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:90) >> >> >> >> at >> >> >> org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:336) >> >> >>
