hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Patrick Kling (JIRA)" <j...@apache.org>
Subject [jira] Commented: (HDFS-1171) RaidNode should fix missing blocks directly on Data Node
Date Thu, 14 Oct 2010 18:08:33 GMT

    [ https://issues.apache.org/jira/browse/HDFS-1171?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12921054#action_12921054

Patrick Kling commented on HDFS-1171:

In order to be able to handle a larger number of simultaneous block losses, we should have
the option of dispatching map reduce jobs to fix the affected files in a distributed fashion.
For greater flexibility, we propose a configuration option that specifies whether to use the
local or the distributed implementation of the block fixer.

> RaidNode should fix missing blocks directly on Data Node
> --------------------------------------------------------
>                 Key: HDFS-1171
>                 URL: https://issues.apache.org/jira/browse/HDFS-1171
>             Project: Hadoop HDFS
>          Issue Type: Task
>          Components: contrib/raid
>    Affects Versions: 0.20.1
>            Reporter: Ramkumar Vadali
> RaidNode currently does not fix missing blocks. The missing blocks have to be fixed manually.
> This task proposes that recovery be more automated:
> 1. RaidNode periodically fetches a list of corrupt files from the NameNode
> 2. If the corrupt files has a RAID parity file, RaidNode identifies missing block(s)
in the file and recomputes the block(s) using the parity file and other good blocks
> 3. RaidNode sends the generated block contents to a DataNode
>    a. RaidNode chooses a DataNode with the most available space to send the block. 

This message is automatically generated by JIRA.
You can reply to this email to add a comment to the issue online.

View raw message