hadoop-hdfs-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Hudson (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (HDFS-10178) Permanent write failures can happen if pipeline recoveries occur for the first packet
Date Mon, 04 Apr 2016 21:59:25 GMT

    [ https://issues.apache.org/jira/browse/HDFS-10178?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15225148#comment-15225148

Hudson commented on HDFS-10178:

FAILURE: Integrated in Hadoop-trunk-Commit #9552 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/9552/])
HDFS-10178. Permanent write failures can happen if pipeline recoveries (kihwal: rev a7d1fb0cd2fdbf830602eb4dbbd9bbe62f4d5584)
* hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestClientProtocolForPipelineRecovery.java
* hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockSender.java
* hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/BlockReceiver.java
* hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/datanode/DataNodeFaultInjector.java

> Permanent write failures can happen if pipeline recoveries occur for the first packet
> -------------------------------------------------------------------------------------
>                 Key: HDFS-10178
>                 URL: https://issues.apache.org/jira/browse/HDFS-10178
>             Project: Hadoop HDFS
>          Issue Type: Bug
>            Reporter: Kihwal Lee
>            Assignee: Kihwal Lee
>            Priority: Critical
>             Fix For: 2.7.3
>         Attachments: HDFS-10178.patch, HDFS-10178.v2.patch, HDFS-10178.v3.patch, HDFS-10178.v4.patch,
> We have observed that write fails permanently if the first packet doesn't go through
properly and pipeline recovery happens. If the write op creates a pipeline, but the actual
data packet does not reach one or more datanodes in time, the pipeline recovery will be done
against the 0-byte partial block.  
> If additional datanodes are added, the block is transferred to the new nodes.  After
the transfer, each node will have a meta file containing the header and 0-length data block
file. The pipeline recovery seems to work correctly up to this point, but write fails when
actual data packet is resent. 

This message was sent by Atlassian JIRA

View raw message