[ https://issues.apache.org/jira/browse/HDFS-1111?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12904801#action_12904801 ]
dhruba borthakur commented on HDFS-1111: ---------------------------------------- I am -1 for committing this change. I think we all agree that this functionality is going to be useful for a variety of tools. One of the tools is the existing fsck process. The reason fsck uses a servlet is because it is long running, otherwise Hadop RPCs will timeout. But there is no such requirement that this new option added to fsck should also use the servlet, "fsck -listcoprruptfiles" is not a long-running process. In fact, i would argue that "fsck -listcorruptfiles", should itself use the DistributedFileSystem API. > getCorruptFiles() should give some hint that the list is not complete > --------------------------------------------------------------------- > > Key: HDFS-1111 > URL: https://issues.apache.org/jira/browse/HDFS-1111 > Project: Hadoop HDFS > Issue Type: New Feature > Affects Versions: 0.22.0 > Reporter: Rodrigo Schmidt > Assignee: Sriram Rao > Fix For: 0.22.0 > > Attachments: HADFS-1111.0.patch, HDFS-1111-y20.1.patch, > HDFS-1111-y20.2.patch, HDFS-1111.trunk.1.patch, HDFS-1111.trunk.patch > > > If the list of corruptfiles returned by the namenode doesn't say anything if > the number of corrupted files is larger than the call output limit (which > means the list is not complete). There should be a way to hint incompleteness > to clients. > A simple hack would be to add an extra entry to the array returned with the > value null. Clients could interpret this as a sign that there are other > corrupt files in the system. > We should also do some rephrasing of the fsck output to make it more > confident when the list is not complete and less confident when the list is > known to be incomplete. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.