[ https://issues.apache.org/jira/browse/HDFS-6088?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Kihwal Lee resolved HDFS-6088. ------------------------------ Resolution: Won't Fix Assignee: Kihwal Lee > Add configurable maximum block count for datanode > ------------------------------------------------- > > Key: HDFS-6088 > URL: https://issues.apache.org/jira/browse/HDFS-6088 > Project: Hadoop HDFS > Issue Type: Bug > Reporter: Kihwal Lee > Assignee: Kihwal Lee > > Currently datanode resources are protected by the free space check and the > balancer. But datanodes can run out of memory simply storing too many > blocks. If the sizes of blocks are small, datanodes will appear to have > plenty of space to put more blocks. > I propose adding a configurable max block count to datanode. Since datanodes > can have different heap configurations, it will make sense to make it > datanode-level, rather than something enforced by namenode. -- This message was sent by Atlassian JIRA (v6.3.4#6332)