Thanks! It is already fixed in HBASE-17170. ________________________________________ 보낸 사람: Ankit Singhal <ankitsingha...@gmail.com> 보낸 날짜: 2019년 5월 8일 수요일 02:50 받는 사람: user@hbase.apache.org 제목: Re: Why HBase client retry even though AccessDeniedException
Yes, you also might be hitting https://issues.apache.org/jira/browse/HBASE-17170 On Tue, May 7, 2019 at 10:33 AM Josh Elser <els...@apache.org> wrote: > Sounds like a bug to me. > > On 5/7/19 5:52 AM, Kang Minwoo wrote: > > Why do not use "doNotRetry" value in RemoteWithExtrasException? > > > > ________________________________________ > > 보낸 사람: Kang Minwoo <minwoo.k...@outlook.com> > > 보낸 날짜: 2019년 5월 7일 화요일 18:23 > > 받는 사람: user@hbase.apache.org > > 제목: Why HBase client retry even though AccessDeniedException > > > > Hello User. > > > > (HBase version: 1.2.9) > > > > Recently, I am testing about DoNotRetryIOException. > > > > I expected when RegionServer send a DoNotRetryIOException (or > AccessDeniedException), Client does not retry. > > But, In Spark or MR, Client retries even though they receive > AccessDeniedException. > > > > Here is a call stack. > > > > Caused by: org.apache.hadoop.hbase.client.RetriesExhaustedException: > Failed after attempts={}, exceptions: {time}, null, > java.net.SocketTimeoutException: > {detail info} > > ... > > Caused by: > org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.security.AccessDeniedException): > org.apache.hadoop.hbase.security.AccessDeniedException: the client is not > authorized > > at (... coprocessor throw AccessDeniedException) > > at > org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$50.call(RegionCoprocessorHost.java:1300) > > at > org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost$RegionOperation.call(RegionCoprocessorHost.java:1673) > > at > org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperation(RegionCoprocessorHost.java:1749) > > at > org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.execOperationWithResult(RegionCoprocessorHost.java:1722) > > at > org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preScannerOpen(RegionCoprocessorHost.java:1295) > > at > org.apache.hadoop.hbase.regionserver.RSRpcServices.scan(RSRpcServices.java:2468) > > at > org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$2.callBlockingMethod(ClientProtos.java:33770) > > at > org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2216) > > at > org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:112) > > at > org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:133) > > at > org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:108) > > at java.lang.Thread.run(Thread.java:748) > > > > at > org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1272) > > at > org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:227) > > at > org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:336) > > at > org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$BlockingStub.scan(ClientProtos.java:34216) > > at > org.apache.hadoop.hbase.client.ScannerCallable.openScanner(ScannerCallable.java:400) > > ... 10 more > > > > The client can not aware of AccessDeniedException because the exception > is RemoteWithExtrasException. > > I wonder it is a bug. > > > > Best regards, > > Minwoo Kang > > >