hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Todd Lipcon (JIRA)" <j...@apache.org>
Subject [jira] [Reopened] (HDFS-1954) Improve corrupt files warning message
Date Thu, 02 Jun 2011 17:55:48 GMT

     [ https://issues.apache.org/jira/browse/HDFS-1954?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]

Todd Lipcon reopened HDFS-1954:
-------------------------------


> Improve corrupt files warning message
> -------------------------------------
>
>                 Key: HDFS-1954
>                 URL: https://issues.apache.org/jira/browse/HDFS-1954
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>            Reporter: philo vivero
>            Assignee: Patrick Hunt
>             Fix For: 0.22.0
>
>         Attachments: HDFS-1954.patch, HDFS-1954.patch
>
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> On NameNode web interface, you may get this warning:
>   WARNING : There are about 32 missing blocks. Please check the log or run fsck.
> If the cluster was started less than 14 days before, it would be great to add: "Is dfs.data.dir
defined?"
> If at the point of that error message, that parameter could be checked, and error made
"OMG dfs.data.dir isn't defined!" that'd be even better. As is, troubleshooting undefined
parameters is a difficult proposition.
> I suspect this is an easy fix.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira

Mime
View raw message