[ https://issues.apache.org/jira/browse/HADOOP-1138?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12524814 ]
dhruba borthakur commented on HADOOP-1138: ------------------------------------------ The code looks good. A few comments: 1. FSnamesystem.getDatanodeListForReport excludes nodes that are listed in dfs.hosts.exclude. Maybe a better option wold be to show them with a status of "Excluded". Currently, it shows "Decommisioned" or "In Service". 2. The comment in FSnamesystem.getDatanodeListForReport talks about "dfs.report.datanode.timeout.day" but it should be "dfs.report.datanode.timeout.hours". 3. Maybe a unit test case that tests this functionality would be really nice. > Datanodes that are dead for a long long time should not show up in the UI > ------------------------------------------------------------------------- > > Key: HADOOP-1138 > URL: https://issues.apache.org/jira/browse/HADOOP-1138 > Project: Hadoop > Issue Type: Bug > Components: dfs > Reporter: dhruba borthakur > Assignee: Raghu Angadi > Fix For: 0.15.0 > > Attachments: HADOOP-1138.patch > > > Proposal 1: > If a include files is used, then show all nodes (dead/alive) that are listed > in the includes file. If there isn't an include file, then display only nodes > that have pinged this instance of the namenode. > Proposal2: > A config variable specifies the time duration. The namenode, on a restart, > purges all datanodes that have not pinged for that time duration. The default > value of this config variable can be 1 week. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.