[ https://issues.apache.org/jira/browse/HBASE-16345?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15475327#comment-15475327 ]
huaxiang sun commented on HBASE-16345: -------------------------------------- A side note, we found that the current scan does not honor the following {code} scan.setConsistency(Consistency.TIMELINE); scan.setReplicaId(1); {code} The scan actually scans all replicas. This is not consistent with the Get. If this is an issue, I can create a new jira to address it. The following testing case shows the issue: {code} @Test public void testReplicaScanWithReplica() throws IOException { // Create table then get the single region for our new table. HTableDescriptor hdt = HTU.createTableDescriptor("testReplicaScanTable"); hdt.setRegionReplication(NB_SERVERS); try { Table table = HTU.createTable(hdt, new byte[][] { f }, null); Put p = new Put(row); p.addColumn(f, row, row); table.put(p); // Flush so it can be picked by the replica refresher thread HTU.flush(table.getName()); // Sleep for some time until data is picked up by replicas try { Thread.sleep(2 * REFRESH_PERIOD); } catch (InterruptedException e1) { LOG.error(e1); } // But if we ask for stale we will get it // Instantiating the Scan class Scan scan = new Scan(); // Scanning the required columns scan.addFamily(f); scan.setConsistency(Consistency.TIMELINE); scan.setReplicaId(1); // Getting the scan result ResultScanner scanner = table.getScanner(scan); Result r = scanner.next(); Assert.assertTrue(r.isStale()); } finally { HTU.getHBaseAdmin().disableTable(hdt.getTableName()); HTU.deleteTable(hdt.getTableName()); } } {code} > RpcRetryingCallerWithReadReplicas#call() should catch some RegionServer > Exceptions > ---------------------------------------------------------------------------------- > > Key: HBASE-16345 > URL: https://issues.apache.org/jira/browse/HBASE-16345 > Project: HBase > Issue Type: Bug > Components: Client > Affects Versions: 2.0.0 > Reporter: huaxiang sun > Assignee: huaxiang sun > Attachments: HBASE-16345-v001.patch, HBASE-16345.master.001.patch, > HBASE-16345.master.002.patch, HBASE-16345.master.003.patch, > HBASE-16345.master.004.patch, HBASE-16345.master.005.patch > > > Update for the description. Debugged more at this front based on the comments > from Enis. > The cause is that for the primary replica, if its retry is exhausted too > fast, f.get() [1] returns ExecutionException. This Exception needs to be > ignored and continue with the replicas. > The other issue is that after adding calls for the replicas, if the first > completed task gets ExecutionException (due to the retry exhausted), it > throws the exception to the client[2]. > In this case, it needs to loop through these tasks, waiting for the success > one. If no one succeeds, throw exception. > Similar for the scan as well > [1] > https://github.com/apache/hbase/blob/master/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerWithReadReplicas.java#L197 > [2] > https://github.com/apache/hbase/blob/master/hbase-client/src/main/java/org/apache/hadoop/hbase/client/RpcRetryingCallerWithReadReplicas.java#L219 -- This message was sent by Atlassian JIRA (v6.3.4#6332)