[
https://issues.apache.org/jira/browse/HDFS-7411?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Andrew Wang updated HDFS-7411:
------------------------------
Attachment: hdfs-7411.007.patch
Here's a new patch addressing Colin's review feedback:
* Got rid of the nodes.per.interval config key and instead use
blocks.per.interval. I didn't bother with a DeprecationDelta because the keys
are semantically different and can't just be aliased. Not sure if there's more
to do here.
* Added support for "0 means unlimited" for the max.concurrent.tracked.nodes
property
* Fixed typos, spacing, added additional debug logging requested. The "every 30
minute print" since we print out the first insufficiently block encountered
while checking each node.
* Fixed a small OBO in the rate limiting, and added a test for the limiting
> Refactor and improve decommissioning logic into DecommissionManager
> -------------------------------------------------------------------
>
> Key: HDFS-7411
> URL: https://issues.apache.org/jira/browse/HDFS-7411
> Project: Hadoop HDFS
> Issue Type: Improvement
> Affects Versions: 2.5.1
> Reporter: Andrew Wang
> Assignee: Andrew Wang
> Attachments: hdfs-7411.001.patch, hdfs-7411.002.patch,
> hdfs-7411.003.patch, hdfs-7411.004.patch, hdfs-7411.005.patch,
> hdfs-7411.006.patch, hdfs-7411.007.patch
>
>
> Would be nice to split out decommission logic from DatanodeManager to
> DecommissionManager.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)