[ 
https://issues.apache.org/jira/browse/HDFS-1954?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13041888#comment-13041888
 ] 

Konstantin Shvachko commented on HDFS-1954:
-------------------------------------------

I agree with Suresh on this. It should be not the message on the web ui, which 
most ops will find annoying. It should just be addressed in FAQ, so that 
newbies could learn from there about mis-configuration.
I would change the message though that way:
{code}
 - Please check the log or run fsck.
 + Please check the logs or run fsck in order to identify the missing blocks.
{code}


> Improve corrupt files warning message
> -------------------------------------
>
>                 Key: HDFS-1954
>                 URL: https://issues.apache.org/jira/browse/HDFS-1954
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>            Reporter: philo vivero
>            Assignee: Patrick Hunt
>             Fix For: 0.22.0
>
>         Attachments: HDFS-1954.patch, HDFS-1954.patch
>
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> On NameNode web interface, you may get this warning:
>   WARNING : There are about 32 missing blocks. Please check the log or run 
> fsck.
> If the cluster was started less than 14 days before, it would be great to 
> add: "Is dfs.data.dir defined?"
> If at the point of that error message, that parameter could be checked, and 
> error made "OMG dfs.data.dir isn't defined!" that'd be even better. As is, 
> troubleshooting undefined parameters is a difficult proposition.
> I suspect this is an easy fix.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira

Reply via email to