[ 
https://issues.apache.org/jira/browse/HDFS-11060?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16356726#comment-16356726
 ] 

Lantao Jin commented on HDFS-11060:
-----------------------------------

Thank you [~jojochuang], new patch file uploaded.

> make DEFAULT_MAX_CORRUPT_FILEBLOCKS_RETURNED configurable
> ---------------------------------------------------------
>
>                 Key: HDFS-11060
>                 URL: https://issues.apache.org/jira/browse/HDFS-11060
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>          Components: hdfs
>    Affects Versions: 2.7.3, 2.8.1, 3.0.0-alpha3
>            Reporter: Lantao Jin
>            Assignee: Lantao Jin
>            Priority: Minor
>         Attachments: HDFS-11060.1.patch, HDFS-11060.2.patch
>
>
> Current, the easiest way to determine which blocks is missing is using NN web 
> UI or JMX. Unfortunately, because the 
> DEFAULT_MAX_CORRUPT_FILEBLOCKS_RETURNED=100 is hard code in FSNamesystem, 
> only 100 missing blocks can be return by UI and JMX. Even the result of URL 
> "https://nn:50070/fsck?listcorruptfileblocks=1&path=%2F"; is limited by this 
> hard code value too.
> I did know FSCK can return more than 100 result but due to the security 
> reason(with kerberos), it is very hard to involve to costumer programs and 
> scripts.
> So I think it should add a configurable var "maxCorruptFileBlocksReturned" to 
> fix above case.
> If community also think it's worth to do, I will patch this. If not, please 
> feel free to tell me the reason. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org

Reply via email to