hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Todd Lipcon (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HDFS-1960) dfs.*.dir should not default to /tmp (or other typically volatile storage)
Date Thu, 19 May 2011 01:24:47 GMT

    [ https://issues.apache.org/jira/browse/HDFS-1960?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13035910#comment-13035910
] 

Todd Lipcon commented on HDFS-1960:
-----------------------------------

This will likely require some discussion. But, I agree.

Another piece of rationale is that HDFS will not start up without defining fs.defaultFS (nee
fs.default.name). So, we already require the user to make some configuration changes before
running HDFS - we may as well require them to configure the persistent storage locations.

> dfs.*.dir should not default to /tmp (or other typically volatile storage)
> --------------------------------------------------------------------------
>
>                 Key: HDFS-1960
>                 URL: https://issues.apache.org/jira/browse/HDFS-1960
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>          Components: data-node
>    Affects Versions: 0.20.2
>         Environment: *nix systems
>            Reporter: philo vivero
>            Priority: Critical
>
> The hdfs-site.xml file possibly will not define one or both of:
> dfs.name.dir
> dfs.data.dir
> If they are not specified, data is stored in /tmp. This is extremely dangerous. Rationale:
the cluster will work fine for days, possibly even weeks, before blocks will start to go missing.
Rebooting a datanode on common Linux systems will clear all the data from that node. There
is no documented way (that I'm aware of) to recover the situation. The cluster must be completely
obliterated and rebuilt from scratch.
> Better reactions to the missing configuration parameters:
> 1. DataNode dies on startup and asks that these parameters be defined.
> 2. Default is /var/db/hadoop (or some other non-volatile storage location). Naturally,
inability to write into that directory leads to DataNode to die on startup, logging error.
> The first solution would be most likely preferred by typical enterprise sysadmins. The
second solution is suboptimal (since /var/db/hadoop might not be the optimal location for
the data) but is still preferable to the current implementation, since it will less often
lead to an irretrievably corrupt cluster.

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira

Mime
View raw message