[ https://issues.apache.org/jira/browse/HDDS-4186?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Glen Geng updated HDDS-4186: ---------------------------- Description: Current RetryPolicy of Datanode for SCM is retryForeverWithFixedSleep: {code:java} RetryPolicy retryPolicy = RetryPolicies.retryForeverWithFixedSleep( 1000, TimeUnit.MILLISECONDS); StorageContainerDatanodeProtocolPB rpcProxy = RPC.getProtocolProxy( StorageContainerDatanodeProtocolPB.class, version, address, UserGroupInformation.getCurrentUser(), hadoopConfig, NetUtils.getDefaultSocketFactory(hadoopConfig), getRpcTimeout(), retryPolicy).getProxy();{code} that for Recon is retryUpToMaximumCountWithFixedSleep: {code:java} RetryPolicy retryPolicy = RetryPolicies.retryUpToMaximumCountWithFixedSleep(10, 60000, TimeUnit.MILLISECONDS); ReconDatanodeProtocolPB rpcProxy = RPC.getProtocolProxy( ReconDatanodeProtocolPB.class, version, address, UserGroupInformation.getCurrentUser(), hadoopConfig, NetUtils.getDefaultSocketFactory(hadoopConfig), getRpcTimeout(), retryPolicy).getProxy(); {code} The executorService in DatanodeStateMachine is Executors.newFixedThreadPool(...), whose default pool size is 2, one for Recon, another for SCM. When encounter rpc failure, call() of RegisterEndpointTask, VersionEndpointTask, HeartbeatEndpointTask will retry while holding the rpcEndpoint.lock(). For example: {code:java} public EndpointStateMachine.EndPointStates call() throws Exception { rpcEndpoint.lock(); try { .... SCMHeartbeatResponseProto reponse = rpcEndpoint.getEndPoint() .sendHeartbeat(request); .... } finally { rpcEndpoint.unlock(); } return rpcEndpoint.getState(); } {code} *The problem is:* If setup one Recon and one SCM, then shutdown the Recon server, all Datanodes will be stale/dead very soon at SCM side. *The root cause is:* The thread running Recon task will retry due to rpc failure, meanwhile holds the lock of EndpointStateMachine for Recon. When DatanodeStateMachine schedule the next round of SCM/Recon task, the only left thread will be assigned to run Recon task, and blocked at waiting for the lock of EndpointStateMachine for Recon. {code:java} public EndpointStateMachine.EndPointStates call() throws Exception { rpcEndpoint.lock(); ...{code} *The solution is:* Since DatanodeStateMachine will periodically schedule SCM/Recon tasks, we may adjust RetryPolicy so that won't retry for longer that 1min. The change has no side effect: 1) VersionEndpointTask.call() is fine 2) RegisterEndpointTask.call() will query containerReport, nodeReport, pipelineReports from OzoneContainer, which is fine. 3) HeartbeatEndpointTask.call() will putBackReports(), which is fine. was: Current RetryPolicy of Datanode for SCM is retryForeverWithFixedSleep: {code:java} RetryPolicy retryPolicy = RetryPolicies.retryForeverWithFixedSleep( 1000, TimeUnit.MILLISECONDS); StorageContainerDatanodeProtocolPB rpcProxy = RPC.getProtocolProxy( StorageContainerDatanodeProtocolPB.class, version, address, UserGroupInformation.getCurrentUser(), hadoopConfig, NetUtils.getDefaultSocketFactory(hadoopConfig), getRpcTimeout(), retryPolicy).getProxy();{code} for Recon is retryUpToMaximumCountWithFixedSleep: {code:java} RetryPolicy retryPolicy = RetryPolicies.retryUpToMaximumCountWithFixedSleep(10, 60000, TimeUnit.MILLISECONDS); ReconDatanodeProtocolPB rpcProxy = RPC.getProtocolProxy( ReconDatanodeProtocolPB.class, version, address, UserGroupInformation.getCurrentUser(), hadoopConfig, NetUtils.getDefaultSocketFactory(hadoopConfig), getRpcTimeout(), retryPolicy).getProxy(); {code} The executorService in DatanodeStateMachine is now Executors.newFixedThreadPool(...), whose pool size is 2, one for Recon, another for SCM. When encounter rpc failure, call() of RegisterEndpointTask, VersionEndpointTask, HeartbeatEndpointTask will retry while holding the rpcEndpoint.lock(). Here is the problem: if setup one Recon and one SCM, then shutdown the Recon server, all Datanodes will be stale/dead very soon. The root cause is that, the thread working for Recon will retry while holding the lock of EndpointStateMachine for Recon, when DatanodeStateMachine schedule the next round of task, the other thread is blocked by waiting for the lock of EndpointStateMachine for Recon. Since DatanodeStateMachine will periodically schedule tasks, we may adjust RetryPolicy so that the execution of tasks no need to be longer than 1min. > Adjust RetryPolicy of SCMConnectionManager > ------------------------------------------ > > Key: HDDS-4186 > URL: https://issues.apache.org/jira/browse/HDDS-4186 > Project: Hadoop Distributed Data Store > Issue Type: Improvement > Reporter: Glen Geng > Assignee: Glen Geng > Priority: Blocker > > Current RetryPolicy of Datanode for SCM is retryForeverWithFixedSleep: > {code:java} > RetryPolicy retryPolicy = > RetryPolicies.retryForeverWithFixedSleep( > 1000, TimeUnit.MILLISECONDS); > StorageContainerDatanodeProtocolPB rpcProxy = RPC.getProtocolProxy( > StorageContainerDatanodeProtocolPB.class, version, > address, UserGroupInformation.getCurrentUser(), hadoopConfig, > NetUtils.getDefaultSocketFactory(hadoopConfig), getRpcTimeout(), > retryPolicy).getProxy();{code} > that for Recon is retryUpToMaximumCountWithFixedSleep: > {code:java} > RetryPolicy retryPolicy = > RetryPolicies.retryUpToMaximumCountWithFixedSleep(10, > 60000, TimeUnit.MILLISECONDS); > ReconDatanodeProtocolPB rpcProxy = RPC.getProtocolProxy( > ReconDatanodeProtocolPB.class, version, > address, UserGroupInformation.getCurrentUser(), hadoopConfig, > NetUtils.getDefaultSocketFactory(hadoopConfig), getRpcTimeout(), > retryPolicy).getProxy(); > {code} > > The executorService in DatanodeStateMachine is > Executors.newFixedThreadPool(...), whose default pool size is 2, one for > Recon, another for SCM. > > When encounter rpc failure, call() of RegisterEndpointTask, > VersionEndpointTask, HeartbeatEndpointTask will retry while holding the > rpcEndpoint.lock(). For example: > > {code:java} > public EndpointStateMachine.EndPointStates call() throws Exception { > rpcEndpoint.lock(); > try { > .... > SCMHeartbeatResponseProto reponse = rpcEndpoint.getEndPoint() > .sendHeartbeat(request); > .... > } finally { > rpcEndpoint.unlock(); > } > return rpcEndpoint.getState(); > } > {code} > > > *The problem is:* > If setup one Recon and one SCM, then shutdown the Recon server, all Datanodes > will be stale/dead very soon at SCM side. > > *The root cause is:* > The thread running Recon task will retry due to rpc failure, meanwhile holds > the lock of EndpointStateMachine for Recon. When DatanodeStateMachine > schedule the next round of SCM/Recon task, the only left thread will be > assigned to run Recon task, and blocked at waiting for the lock of > EndpointStateMachine for Recon. > > {code:java} > public EndpointStateMachine.EndPointStates call() throws Exception { > rpcEndpoint.lock(); > ...{code} > > > *The solution is:* > Since DatanodeStateMachine will periodically schedule SCM/Recon tasks, we may > adjust RetryPolicy so that won't retry for longer that 1min. > The change has no side effect: > 1) VersionEndpointTask.call() is fine > 2) RegisterEndpointTask.call() will query containerReport, nodeReport, > pipelineReports from OzoneContainer, which is fine. > 3) HeartbeatEndpointTask.call() will putBackReports(), which is fine. > -- This message was sent by Atlassian Jira (v8.3.4#803005) --------------------------------------------------------------------- To unsubscribe, e-mail: ozone-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: ozone-issues-h...@hadoop.apache.org