Return-Path: X-Original-To: apmail-hadoop-common-user-archive@www.apache.org Delivered-To: apmail-hadoop-common-user-archive@www.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id 746E67B13 for ; Tue, 2 Aug 2011 00:29:44 +0000 (UTC) Received: (qmail 22223 invoked by uid 500); 2 Aug 2011 00:29:41 -0000 Delivered-To: apmail-hadoop-common-user-archive@hadoop.apache.org Received: (qmail 22014 invoked by uid 500); 2 Aug 2011 00:29:40 -0000 Mailing-List: contact common-user-help@hadoop.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: common-user@hadoop.apache.org Delivered-To: mailing list common-user@hadoop.apache.org Received: (qmail 22006 invoked by uid 99); 2 Aug 2011 00:29:40 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 02 Aug 2011 00:29:40 +0000 X-ASF-Spam-Status: No, hits=-0.7 required=5.0 tests=FREEMAIL_FROM,MIME_QP_LONG_LINE,RCVD_IN_DNSWL_LOW,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: domain of mohitanchlia@gmail.com designates 209.85.210.53 as permitted sender) Received: from [209.85.210.53] (HELO mail-pz0-f53.google.com) (209.85.210.53) by apache.org (qpsmtpd/0.29) with ESMTP; Tue, 02 Aug 2011 00:29:32 +0000 Received: by pzk6 with SMTP id 6so16661436pzk.40 for ; Mon, 01 Aug 2011 17:29:11 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=subject:references:from:content-type:x-mailer:in-reply-to :message-id:date:to:content-transfer-encoding:mime-version; bh=IPEkH0fvO6Jy4c8ZHnuwQJ63I7aVO7RKeXwnPKuXRyg=; b=Q2teVJPEaVumkhUvyqTd7pUWImVOCes5RC4Xs7HVTaJto8I3HzEpx63kJtQtQqMZjR zg06cYLb2mHrRBQ89VBYpCe3hIqTEau290NqpqqTAisj2FGF73aGDDz3u4bcXoRXa0T0 vm+a5kYzS4hltrrn67y9CG9tEaQALQlt4nulI= Received: by 10.68.49.70 with SMTP id s6mr8543562pbn.233.1312244951286; Mon, 01 Aug 2011 17:29:11 -0700 (PDT) Received: from [172.20.55.50] ([12.69.234.130]) by mx.google.com with ESMTPS id l7sm6096112pbh.74.2011.08.01.17.29.10 (version=TLSv1/SSLv3 cipher=OTHER); Mon, 01 Aug 2011 17:29:10 -0700 (PDT) Subject: Re: Hadoop cluster network requirement References: <000b01cc4ff2$ed066140$c71323c0$@margallacomm.com> <1548924D-41E6-45D3-A7EB-F324966A0E56@apache.org> From: Mohit Anchlia Content-Type: text/plain; charset=us-ascii X-Mailer: iPad Mail (8C148) In-Reply-To: Message-Id: <6025B1DC-575B-4641-B40A-AC97F10252FE@gmail.com> Date: Mon, 1 Aug 2011 17:29:06 -0700 To: "common-user@hadoop.apache.org" Content-Transfer-Encoding: quoted-printable Mime-Version: 1.0 (iPad Mail 8C148) X-Virus-Checked: Checked by ClamAV on apache.org Assuming everything is up this solution still will not scale given the laten= cy, tcpip buffers, sliding window etc. See BDP Sent from my iPad On Aug 1, 2011, at 4:57 PM, Michael Segel wrote:= >=20 > Yeah what he said. > Its never a good idea. > Forget about losing a NN or a Rack, but just losing connectivity between d= ata centers. (It happens more than you think.) > Your entire cluster in both data centers go down. Boom! >=20 > Its a bad design.=20 >=20 > You're better off doing two different clusters. >=20 > Is anyone really trying to sell this as a design? That's even more scary. >=20 >=20 >> Subject: Re: Hadoop cluster network requirement >> From: aw@apache.org >> Date: Sun, 31 Jul 2011 20:28:53 -0700 >> To: common-user@hadoop.apache.org; saqibj@margallacomm.com >>=20 >>=20 >> On Jul 31, 2011, at 7:30 PM, Saqib Jang -- Margalla Communications wrote:= >>=20 >>> Thanks, I'm independently doing some digging into Hadoop networking >>> requirements and=20 >>> had a couple of quick follow-ups. Could I have some specific info on why= >>> different data centers=20 >>> cannot be supported for master node and data node comms? >>> Also, what=20 >>> may be the benefits/use cases for such a scenario? >>=20 >> Most people who try to put the NN and DNs in different data centers ar= e trying to achieve disaster recovery: one file system in multiple location= s. That isn't the way HDFS is designed and it will end in tears. There are m= ultiple problems: >>=20 >> 1) no guarantee that one block replica will be each data center (thereby d= efeating the whole purpose!) >> 2) assuming one can work out problem 1, during a network break, the NN wi= ll lose contact from one half of the DNs, causing a massive network replica= tion storm >> 3) if one using MR on top of this HDFS, the shuffle will likely kill the n= etwork in between (making MR performance pretty dreadful) is going to cause d= elays for the DN heartbeats >> 4) I don't even want to think about rebalancing. >>=20 >> ... and I'm sure a lot of other problems I'm forgetting at the moment.= So don't do it. >>=20 >> If you want disaster recovery, set up two completely separate HDFSes a= nd run everything in parallel. > =20