[ https://issues.apache.org/jira/browse/HDFS-729?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12805520#action_12805520 ]
dhruba borthakur commented on HDFS-729: --------------------------------------- Hi Hairong and Raghu, I had a discussion with Konstantin and he suggested simplifying the API by making this API always return the first 500 corrupted files. Even if there are more than 500 corrupted files in the system, successive calls to this API will always return only 500 files. There is no relationship between the files returned by one invocation of this API with the ones returned by the next invocation of this API. The above proposal address Hairongs (1) above. It also is in line with Raghu's suggestion "t is probably good enough to always return up to (100 or 500 files". Regarding Hairong's (2) above, this should be addressed and will fixed in the next patch (if not already there). Does this sound like a reasonable approach to follow? > fsck option to list only corrupted files > ---------------------------------------- > > Key: HDFS-729 > URL: https://issues.apache.org/jira/browse/HDFS-729 > Project: Hadoop HDFS > Issue Type: Improvement > Reporter: dhruba borthakur > Assignee: dhruba borthakur > Attachments: badFiles.txt, badFiles2.txt, corruptFiles.txt > > > An option to fsck to list only corrupted files will be very helpful for > frequent monitoring. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.