Return-Path: X-Original-To: apmail-hadoop-hdfs-user-archive@minotaur.apache.org Delivered-To: apmail-hadoop-hdfs-user-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 95182FDC4 for ; Tue, 2 Apr 2013 08:17:35 +0000 (UTC) Received: (qmail 72600 invoked by uid 500); 2 Apr 2013 08:17:30 -0000 Delivered-To: apmail-hadoop-hdfs-user-archive@hadoop.apache.org Received: (qmail 72337 invoked by uid 500); 2 Apr 2013 08:17:30 -0000 Mailing-List: contact user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: user@hadoop.apache.org Delivered-To: mailing list user@hadoop.apache.org Received: (qmail 72313 invoked by uid 99); 2 Apr 2013 08:17:29 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 02 Apr 2013 08:17:29 +0000 X-ASF-Spam-Status: No, hits=-0.7 required=5.0 tests=RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of henry.jykim@gmail.com designates 209.85.220.44 as permitted sender) Received: from [209.85.220.44] (HELO mail-pa0-f44.google.com) (209.85.220.44) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 02 Apr 2013 08:17:23 +0000 Received: by mail-pa0-f44.google.com with SMTP id bi5so165688pad.17 for ; Tue, 02 Apr 2013 01:17:01 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=x-received:content-type:mime-version:subject:from:in-reply-to:date :content-transfer-encoding:message-id:references:to:x-mailer; bh=47exsVY6U8l0qBq741U+WO2UP8CB+ezhmiv2Uf+oASk=; b=pTa/j3RWV3YKR0Mv+V0W5IfADK9SK6UGIO6rg8TATpa/9wFLKRLR0iJ93+yZqhWmai bt6t55QX5JkmTXQrT9WakNL0lu8DpQOYvOewUuZLYr9tx+0vHJqtfty5Ro3ba2++d0aB I9Xjc29akf2EXnj77On2RDQTHp6fGLOlAX/HCVqnafanObp5JKv/KPEvflkreICBc3NR gn6BbNRhUYVYaUvd7nbXeMlxu+ex+EaGj0M7ojMJ3lUbpJBZATD9WOK6HIm5rcmEZKVy 52d5C88sCbEvUNmC42fku1VIjPc19hcv6nFfuKOmzkpRaG7Rf2Lp7Fyz8e6kz25hv+TN C16g== X-Received: by 10.68.241.225 with SMTP id wl1mr22778114pbc.170.1364890621841; Tue, 02 Apr 2013 01:17:01 -0700 (PDT) Received: from [192.168.162.80] ([182.162.60.66]) by mx.google.com with ESMTPS id cy4sm839603pbc.13.2013.04.02.01.17.00 (version=TLSv1 cipher=ECDHE-RSA-RC4-SHA bits=128/128); Tue, 02 Apr 2013 01:17:01 -0700 (PDT) Content-Type: text/plain; charset=euc-kr Mime-Version: 1.0 (Mac OS X Mail 6.2 \(1499\)) Subject: Re: are we able to decommission multi nodes at one time? From: Henry Junyoung Kim In-Reply-To: Date: Tue, 2 Apr 2013 17:16:58 +0900 Content-Transfer-Encoding: quoted-printable Message-Id: <9E45CED1-1D86-4116-8892-7C514EF64342@gmail.com> References: <57310067-9C98-41EA-A674-82940D584EED@gmail.com> To: user@hadoop.apache.org X-Mailer: Apple Mail (2.1499) X-Virus-Checked: Checked by ClamAV on apache.org the rest of nodes to be alive has enough size to store. for this one that you've mentioned. > its easier to do so in a rolling manner without need of a > decommission. to check my understanding, just shutting down 2 of them and then 2 more = and then 2 more without decommissions. is this correct? 2013. 4. 2., =BF=C0=C8=C4 4:54, Harsh J =C0=DB=BC=BA:= > Note though that its only possible to decommission 7 nodes at the same > time and expect it to finish iff the remaining 8 nodes have adequate > free space for the excess replicas. >=20 > If you're just going to take them down for a short while (few mins > each), its easier to do so in a rolling manner without need of a > decommission. You can take upto two down at a time on a replication > average of 3 or 3+, and put it back in later without too much data > movement impact. >=20 > On Tue, Apr 2, 2013 at 1:06 PM, Yanbo Liang = wrote: >> It's reasonable to decommission 7 nodes at the same time. >> But may be it also takes long time to finish it. >> Because all the replicas in these 7 nodes need to be copied to = remaining 8 >> nodes. >> The size of transfer from these nodes to the remaining nodes is = equal. >>=20 >>=20 >> 2013/4/2 Henry Junyoung Kim >>>=20 >>> :) >>>=20 >>> currently, I have 15 data nodes. >>> for some tests, I am trying to decommission until 8 nodes. >>>=20 >>> Now, the total dfs used size is 52 TB which is including all = replicated >>> blocks. >>> from 15 to 8, total spent time is almost 4 days long. ;( >>>=20 >>> someone mentioned that I don't need to decommission node by node. >>> for this case, is there no problems if I decommissioned 7 nodes at = the >>> same time? >>>=20 >>>=20 >>> 2013. 4. 2., =BF=C0=C8=C4 12:14, Azuryy Yu = =C0=DB=BC=BA: >>>=20 >>> I can translate it to native English: how many nodes you want to >>> decommission? >>>=20 >>>=20 >>> On Tue, Apr 2, 2013 at 11:01 AM, Yanbo Liang = wrote: >>>>=20 >>>> You want to decommission how many nodes? >>>>=20 >>>>=20 >>>> 2013/4/2 Henry JunYoung KIM >>>>>=20 >>>>> 15 for datanodes and 3 for replication factor. >>>>>=20 >>>>> 2013. 4. 1., =BF=C0=C8=C4 3:23, varun kumar = =C0=DB=BC=BA: >>>>>=20 >>>>>> How many nodes do you have and replication factor for it. >>>>>=20 >>>>=20 >>>=20 >>>=20 >>=20 >=20 >=20 >=20 > --=20 > Harsh J