[ 
https://issues.apache.org/jira/browse/HDFS-729?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12805520#action_12805520
 ] 

dhruba borthakur commented on HDFS-729:
---------------------------------------

Hi Hairong and Raghu, I had a discussion with Konstantin and he suggested 
simplifying the API by making this API always return the first 500 corrupted 
files. Even if there are more than 500 corrupted files in the system,  
successive calls to this API will always  return only 500 files. There is no 
relationship between the files returned by one invocation of this API with the 
ones returned by the next invocation of this API.

The above proposal address Hairongs (1) above. It also is in line with Raghu's 
suggestion "t is probably good enough to always return up to (100 or 500 
files". Regarding Hairong's (2) above, this should be addressed and will fixed 
in the next patch (if not already there).

Does this sound like a reasonable approach to follow?

> fsck option to list only corrupted files
> ----------------------------------------
>
>                 Key: HDFS-729
>                 URL: https://issues.apache.org/jira/browse/HDFS-729
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>            Reporter: dhruba borthakur
>            Assignee: dhruba borthakur
>         Attachments: badFiles.txt, badFiles2.txt, corruptFiles.txt
>
>
> An option to fsck to list only corrupted files will be very helpful for 
> frequent monitoring.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to