hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "stack (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HDFS-1539) prevent data loss when a cluster suffers a power loss
Date Fri, 04 May 2012 18:34:49 GMT

    [ https://issues.apache.org/jira/browse/HDFS-1539?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13268609#comment-13268609

stack commented on HDFS-1539:

Should we pull this into 1.0.3?  Or 1.1.0?
> prevent data loss when a cluster suffers a power loss
> -----------------------------------------------------
>                 Key: HDFS-1539
>                 URL: https://issues.apache.org/jira/browse/HDFS-1539
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>          Components: data-node, hdfs client, name-node
>            Reporter: dhruba borthakur
>            Assignee: dhruba borthakur
>             Fix For: 0.23.0
>         Attachments: syncOnClose1.txt, syncOnClose2.txt
> we have seen an instance where a external outage caused many datanodes to reboot at around
the same time.  This resulted in many corrupted blocks. These were recently written blocks;
the current implementation of HDFS Datanodes do not sync the data of a block file when the
block is closed.
> 1. Have a cluster-wide config setting that causes the datanode to sync a block file when
a block is finalized.
> 2. Introduce a new parameter to the FileSystem.create() to trigger the new behaviour,
i.e. cause the datanode to sync a block-file when it is finalized.
> 3. Implement the FSDataOutputStream.hsync() to cause all data written to the specified
file to be written to stable storage.

This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira


View raw message