Return-Path: Delivered-To: apmail-hadoop-core-dev-archive@www.apache.org Received: (qmail 7907 invoked from network); 5 May 2008 20:30:22 -0000 Received: from hermes.apache.org (HELO mail.apache.org) (140.211.11.2) by minotaur.apache.org with SMTP; 5 May 2008 20:30:22 -0000 Received: (qmail 83696 invoked by uid 500); 5 May 2008 20:30:19 -0000 Delivered-To: apmail-hadoop-core-dev-archive@hadoop.apache.org Received: (qmail 83674 invoked by uid 500); 5 May 2008 20:30:19 -0000 Mailing-List: contact core-dev-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: core-dev@hadoop.apache.org Delivered-To: mailing list core-dev@hadoop.apache.org Received: (qmail 83660 invoked by uid 99); 5 May 2008 20:30:19 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 05 May 2008 13:30:19 -0700 X-ASF-Spam-Status: No, hits=-2000.0 required=10.0 tests=ALL_TRUSTED X-Spam-Check-By: apache.org Received: from [140.211.11.140] (HELO brutus.apache.org) (140.211.11.140) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 05 May 2008 20:29:41 +0000 Received: from brutus (localhost [127.0.0.1]) by brutus.apache.org (Postfix) with ESMTP id 96D64234C100 for ; Mon, 5 May 2008 13:29:55 -0700 (PDT) Message-ID: <25331898.1210019395613.JavaMail.jira@brutus> Date: Mon, 5 May 2008 13:29:55 -0700 (PDT) From: "Nigel Daley (JIRA)" To: core-dev@hadoop.apache.org Subject: [jira] Updated: (HADOOP-3339) DFS Write pipeline does not detect defective datanode correctly if it times out. In-Reply-To: <1051639654.1209751135597.JavaMail.jira@brutus> MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 7bit X-Virus-Checked: Checked by ClamAV on apache.org [ https://issues.apache.org/jira/browse/HADOOP-3339?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Nigel Daley updated HADOOP-3339: -------------------------------- Fix Version/s: 0.18.0 Assigning to 0.18. This isn't a blocker for 0.17 > DFS Write pipeline does not detect defective datanode correctly if it times out. > -------------------------------------------------------------------------------- > > Key: HADOOP-3339 > URL: https://issues.apache.org/jira/browse/HADOOP-3339 > Project: Hadoop Core > Issue Type: Bug > Components: dfs > Affects Versions: 0.16.0 > Reporter: Raghu Angadi > Assignee: Raghu Angadi > Fix For: 0.18.0 > > > When DFSClient is writing to DFS, it does not correctly detect the culprit datanode (rather datanodes do not inform) properly if the bad node times out. Say, the last datanode in in 3 node pipeline is is too slow or defective. In this case, pipeline removes the first two datanodes in first two attempts. The third attempt has only the 3rd datanode in the pipeline and it will fail too. If the pipeline detects the bad 3rd node when the first failure occurs, the write will succeed in the second attempt. > I will attach example logs of such cases. I think this should be fixed in 0.17.x. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.