[
https://issues.apache.org/jira/browse/HDFS-7411?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14236549#comment-14236549
]
Ming Ma commented on HDFS-7411:
-------------------------------
Andrew, nice work. It appears I don't need to continue the work on
https://issues.apache.org/jira/browse/HDFS-7442. Some initial comments.
1. NN Memory impact on the additional decomNodeBlocks. It shouldn't be an issue
given admins won't decommission lots of nodes at the same time. But it might be
worth calling out some limit here. 100 nodes * 400k blocks per node * 8 bytes
per blockInfo reference, 320MB extra at the start of the decommission process?
2. It appears "dfs.namenode.decommission.blocks.per.node" description should
refer to "dfs.namenode.decommission.nodes.per.interval" instead.
3. It appears this patch also fixed
https://issues.apache.org/jira/browse/HDFS-5757 by calling decomNodeBlocks.put
during refreshNodes.
> Refactor and improve decommissioning logic into DecommissionManager
> -------------------------------------------------------------------
>
> Key: HDFS-7411
> URL: https://issues.apache.org/jira/browse/HDFS-7411
> Project: Hadoop HDFS
> Issue Type: Improvement
> Affects Versions: 2.5.1
> Reporter: Andrew Wang
> Assignee: Andrew Wang
> Attachments: hdfs-7411.001.patch, hdfs-7411.002.patch,
> hdfs-7411.003.patch
>
>
> Would be nice to split out decommission logic from DatanodeManager to
> DecommissionManager.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)