[ 
https://issues.apache.org/jira/browse/HDFS-1954?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13041437#comment-13041437
 ] 

Patrick Hunt commented on HDFS-1954:
------------------------------------

Hey Suresh, if everyone is ok with adding external links I'll be happy to 
create a second patch. I was unable to find a FAQ entry for this, if you 
provide a link to one I'll give it a shot. (also if you have any preferences 
for the page/link text lmk and I'll incorporate it)

> Improve corrupt files warning message
> -------------------------------------
>
>                 Key: HDFS-1954
>                 URL: https://issues.apache.org/jira/browse/HDFS-1954
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>            Reporter: philo vivero
>            Assignee: Patrick Hunt
>             Fix For: 0.22.0
>
>         Attachments: HDFS-1954.patch, HDFS-1954.patch
>
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> On NameNode web interface, you may get this warning:
>   WARNING : There are about 32 missing blocks. Please check the log or run 
> fsck.
> If the cluster was started less than 14 days before, it would be great to 
> add: "Is dfs.data.dir defined?"
> If at the point of that error message, that parameter could be checked, and 
> error made "OMG dfs.data.dir isn't defined!" that'd be even better. As is, 
> troubleshooting undefined parameters is a difficult proposition.
> I suspect this is an easy fix.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira

Reply via email to