hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "dhruba borthakur (JIRA)" <j...@apache.org>
Subject [jira] Commented: (HDFS-457) better handling of volume failure in Data Node storage
Date Wed, 01 Jul 2009 13:04:47 GMT

    [ https://issues.apache.org/jira/browse/HDFS-457?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12726044#action_12726044
] 

dhruba borthakur commented on HDFS-457:
---------------------------------------

Another big issue is "reporting". In the current code, an adminstrator sees the dead Datanode
in the webui and knows that he/she has to fix it.

In your proposal, a datanode could stop using a portion of its disk and it would be very worthwhile
to report this to the adminstrator. Otherwise, a certain percentage of disk space could leak
out of HDFS usage without anybody knowing about it.

> better handling of volume failure in Data Node storage
> ------------------------------------------------------
>
>                 Key: HDFS-457
>                 URL: https://issues.apache.org/jira/browse/HDFS-457
>             Project: Hadoop HDFS
>          Issue Type: Bug
>          Components: data-node
>            Reporter: Boris Shkolnik
>            Assignee: Boris Shkolnik
>
> Current implementation shuts DataNode down completely when one of the configured volumes
of the storage fails.
> This is rather wasteful behavior because it  decreases utilization (good storage becomes
unavailable) and imposes extra load on the system (replication of the blocks from the good
volumes). These problems will become even more prominent when we move to mixed (heterogeneous)
clusters with many more volumes per Data Node.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


Mime
View raw message