hadoop-common-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Runping Qi (JIRA)" <j...@apache.org>
Subject [jira] Commented: (HADOOP-4480) data node process should not die if one dir goes bad
Date Sun, 26 Oct 2008 16:32:44 GMT

    [ https://issues.apache.org/jira/browse/HADOOP-4480?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12642801#action_12642801

Runping Qi commented on HADOOP-4480:

I think the map/reduce framework has to handling similar problems.
If a drive of a machine goes bad, the tasks on that machine tend to become stragglers.
The overall performance will be impacted.
Overall, Hadoop is much better at handling total failure than partial failure of nodes, I
think it is better to decommission a bad node at a drive failure.
The admin may later choose to remove the drive from the configuration file and restart the
node, if he does not want to take away the node for repair.

> data node process should not die if one dir goes bad
> ----------------------------------------------------
>                 Key: HADOOP-4480
>                 URL: https://issues.apache.org/jira/browse/HADOOP-4480
>             Project: Hadoop Core
>          Issue Type: Bug
>          Components: dfs
>    Affects Versions: 0.18.1
>            Reporter: Allen Wittenauer
> When multiple directories are configured for the data node process to use to store blocks,
it currently exits when one of them is not writable.   Instead, it should either completely
ignore that directory or attempt to continue reading and then marking it unusable if reads

This message is automatically generated by JIRA.
You can reply to this email to add a comment to the issue online.

View raw message