hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Suresh Srinivas (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HDFS-1954) Improve corrupt files warning message
Date Sat, 28 May 2011 21:52:47 GMT

    [ https://issues.apache.org/jira/browse/HDFS-1954?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13040680#comment-13040680
] 

Suresh Srinivas commented on HDFS-1954:
---------------------------------------

My preference would be not to have this change. What are we basing "this is the most common
cause of the error" other than the opinion expressed by Philo?

>From what I have seen in our clusters, the main reason this happens is when a bunch of
datanodes do not come up or when replication factor 1 is used on files and some datanodes
do not come up.

> Improve corrupt files warning message
> -------------------------------------
>
>                 Key: HDFS-1954
>                 URL: https://issues.apache.org/jira/browse/HDFS-1954
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>            Reporter: philo vivero
>            Assignee: Patrick Hunt
>             Fix For: 0.22.0
>
>         Attachments: HDFS-1954.patch, HDFS-1954.patch
>
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> On NameNode web interface, you may get this warning:
>   WARNING : There are about 32 missing blocks. Please check the log or run fsck.
> If the cluster was started less than 14 days before, it would be great to add: "Is dfs.data.dir
defined?"
> If at the point of that error message, that parameter could be checked, and error made
"OMG dfs.data.dir isn't defined!" that'd be even better. As is, troubleshooting undefined
parameters is a difficult proposition.
> I suspect this is an easy fix.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira

Mime
View raw message