Return-Path: X-Original-To: apmail-hadoop-mapreduce-user-archive@minotaur.apache.org Delivered-To: apmail-hadoop-mapreduce-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 0AB6D11EAD for ; Mon, 28 Jul 2014 17:22:59 +0000 (UTC) Received: (qmail 74797 invoked by uid 500); 28 Jul 2014 17:22:49 -0000 Delivered-To: apmail-hadoop-mapreduce-user-archive@hadoop.apache.org Received: (qmail 74685 invoked by uid 500); 28 Jul 2014 17:22:49 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 74674 invoked by uid 99); 28 Jul 2014 17:22:49 -0000 Received: from athena.apache.org (HELO athena.apache.org) (140.211.11.136) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 28 Jul 2014 17:22:49 +0000 X-ASF-Spam-Status: No, hits=1.5 required=5.0 tests=HTML_MESSAGE,RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (athena.apache.org: domain of hadoophive@gmail.com designates 209.85.212.176 as permitted sender) Received: from [209.85.212.176] (HELO mail-wi0-f176.google.com) (209.85.212.176) by apache.org (qpsmtpd/0.29) with ESMTP; Mon, 28 Jul 2014 17:22:44 +0000 Received: by mail-wi0-f176.google.com with SMTP id bs8so4817850wib.15 for ; Mon, 28 Jul 2014 10:22:21 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=mime-version:in-reply-to:references:date:message-id:subject:from:to :content-type; bh=7sMFYUeqQDDfbsdXB5pVNRUEQea2UhRO5mCszrlUHFI=; b=VutkJM+0dUVfJTXBrtuFuPq33jym/4oFQ3wArm6HsHESQVrTqY1BE3JdVQ4ri9I9ij 66ZeovlOi7uJg4eX98eJ0SJAmMc0zgsfFqRfAO+gyJiuo2T2RGgSdDn5RDxPLtIXUPqS pQ5Io8g2n1gyIx6zvU1TJhKSOtUdotbCCaMiacy7QsX8HRmzhdL2atC2ZqsBLg/J5jj7 GCsZvgd/BB4Jy7h+WkeeoNuhXPCo9jw/MMGY7GCWvsZ/SZc2Y5l46hyqdNYSG7Yd7n2f ih+VTySsYnFzsyTqWvD34eVyn7zipPVwgoRURqTAHxKDMVbgucAq4jM5r47eDELcIwD8 ldUg== MIME-Version: 1.0 X-Received: by 10.180.93.8 with SMTP id cq8mr32878093wib.17.1406568140071; Mon, 28 Jul 2014 10:22:20 -0700 (PDT) Received: by 10.217.56.200 with HTTP; Mon, 28 Jul 2014 10:22:20 -0700 (PDT) Received: by 10.217.56.200 with HTTP; Mon, 28 Jul 2014 10:22:20 -0700 (PDT) In-Reply-To: <53D67DEE.6000806@ericsson.com> References: <0N9F00JYYKMRH020@phonehome.asia.apple.com> <53D67DEE.6000806@ericsson.com> Date: Mon, 28 Jul 2014 22:52:20 +0530 Message-ID: Subject: Re: One datanode is down then write/read starts failing From: hadoop hive To: user@hadoop.apache.org Content-Type: multipart/alternative; boundary=f46d043892b1576a4304ff442a97 X-Virus-Checked: Checked by ClamAV on apache.org --f46d043892b1576a4304ff442a97 Content-Type: text/plain; charset=UTF-8 If you have 2 an live initially and rep set to 2 which is perfectly fine but you killed one dn... There is no place to put another replica of new files as well as old files... Which causing issue in writing blocks. On Jul 28, 2014 10:15 PM, "Satyam Singh" wrote: > @vikas i have initially set 2 but after that i have make one DN down. So > you are saying from initial i have to make replication factor as 1 even i > have DN 2 active initially. If so then what is the reason? > > On 07/28/2014 10:02 PM, Vikas Srivastava wrote: > >> What replication have you set for cluster. >> >> Its should be 1 in your case. >> >> On Jul 28, 2014 9:26 PM, Satyam Singh wrote: >> >>> Hello, >>> >>> >>> I have hadoop cluster setup of one namenode and two datanodes. >>> And i continuously write/read/delete through hdfs on namenode through >>> hadoop client. >>> >>> Then i kill one of the datanode, still one is working but writing on >>> datanode is getting failed for all write requests. >>> >>> I want to overcome this scenario because at live traffic scenario any of >>> datanode might get down then how do we handle those cases. >>> >>> Can anybody face this issue or i am doing something wrong in my setup. >>> >>> Thanx in advance. >>> >>> >>> Warm Regards, >>> Satyam >>> >> > --f46d043892b1576a4304ff442a97 Content-Type: text/html; charset=UTF-8 Content-Transfer-Encoding: quoted-printable

If you have 2 an live initially and rep set to 2 which is pe= rfectly fine but you killed one dn... There is no place to put another repl= ica of new files as well as old files... Which causing issue in writing blo= cks.

On Jul 28, 2014 10:15 PM, "Satyam Singh&quo= t; <satyam.singh@ericsson.c= om> wrote:
@vikas i have initially set 2 but after that i have make one DN down. So yo= u are saying from initial i have to make replication factor as 1 even i hav= e DN 2 active initially. If so then what is the reason?

On 07/28/2014 10:02 PM, Vikas Srivastava wrote:
What replication have you set for cluster.

Its should be 1 in your case.

On Jul 28, 2014 9:26 PM, Satyam Singh <satyam.singh@ericsson.com> wrote:
Hello,


I have hadoop cluster setup of one namenode and two datanodes.
And i continuously write/read/delete through hdfs on namenode through
hadoop client.

Then i kill one of the datanode, still one is working but writing on
datanode is getting failed for all write requests.

I want to overcome this scenario because at live traffic scenario any of datanode might get down then how do we handle those cases.

Can anybody face this issue or i am doing something wrong in my setup.

Thanx in advance.


Warm Regards,
Satyam

--f46d043892b1576a4304ff442a97--