[ 
https://issues.apache.org/jira/browse/HDDS-5413?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Attila Doroszlai updated HDDS-5413:
-----------------------------------
    Labels:   (was: pull-request-available)

> Limit num of containers to process per round for ReplicationManager.
> --------------------------------------------------------------------
>
>                 Key: HDDS-5413
>                 URL: https://issues.apache.org/jira/browse/HDDS-5413
>             Project: Apache Ozone
>          Issue Type: Sub-task
>            Reporter: Mark Gui
>            Assignee: Mark Gui
>            Priority: Major
>
> For now, ReplicationManager process all containers at once, this will 
> potentially bring healthy load to datanodes if there are a lot of containers 
> to be replicated/deleted/closed.
> So it is nice to have a bound for each round, HDFS has similar settings, and 
> this issue tries to implement sth like 
> 'dfs.block.misreplication.processing.limit: 10000' in HDFS.
>  
> This is just a limit on the number of containers to be processed, note that 
> ReplicationManager count each container as processed no mater it is under 
> replicated or over replicated or good. And all the cmds are queued and will 
> be sent with heartbeats.
> So this limit does not directly limit the data to be replicated, just to have 
> a basic throttling for ReplicationManager in-memory processing, so 10000 
> should be good for ozone as well.
> And we should add more throttling limits in the future.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to